Measuring Security Awareness Effectiveness

This article is for educational purposes only and does not constitute professional compliance advice or legal counsel. Measurement frameworks should be designed with consideration for organizational context, available data sources, and the inherent limitations of attribution in complex environments.


You've been running a security awareness program for a year. You conducted mandatory training for all 500 employees. You ran quarterly phishing simulations. You sent monthly security tips. The question that matters is: did any of this actually work? Is your organization more secure? Are employees making better security decisions? Are incidents decreasing?

Most organizations measure what's easy to measure. Training completion rates: 94% of employees took the course. Phishing simulation metrics: click rates declined from 25% to 18%. Compliance checkboxes: yes, we have an awareness program. But these metrics don't prove that the program is working or that behavior has actually improved. They prove that the program exists and people participated. The difference between measuring effort and measuring impact is fundamental, and most organizations miss it because true impact measurement is genuinely difficult.

Understanding what to measure, what different metrics actually tell you, and how to build a framework that shows real effectiveness is the key to knowing whether your awareness efforts are money well spent or just expensive theater.

What Different Types of Metrics Measure

There are fundamentally different levels of metrics, and they measure different things. Understanding the difference is critical because it determines what you actually learn.

Activity metrics measure effort: how many training sessions did you run, how many newsletters did you send, how many posters did you display? These metrics tell you whether the program is active. They don't tell you anything about impact. You could send 12 security newsletters per year and have no one read them. The activity metric still shows 12 newsletters sent.

Awareness metrics measure knowledge: did people learn the material? Quiz scores immediately after training, simulation click rates, survey responses asking people about their security practices. These metrics tell you whether people absorbed information. But absorbing information during training is different from applying it in real situations. A person might score 85% on a security training quiz and fall for phishing the next week when they're distracted and under time pressure.

Behavior metrics measure actual actions: are people doing the security practices you trained them on? Phishing reporting rates, password strength audits, access request patterns, incident reports submitted. These metrics tell you whether behavior has changed. The gap between awareness and behavior is significant—people can be aware that they should use strong passwords and still use weak ones.

Business outcome metrics measure real impact: is the organization actually more secure? Phishing-based breach incidents, cost of incidents, dwell time (how long attackers stay undetected), mean time to detection and response. These metrics tell you whether the program is actually reducing risk. But they're the hardest to measure because they depend on factors beyond awareness—technical controls, luck, threat actor behavior.

Most organizations focus on activity and awareness metrics because they're easy to measure. But they don't actually measure effectiveness. The goal should be to move toward behavior and outcome metrics.

Phishing Simulations: Click Rates and Reporting Behavior

Phishing simulation click rate is one of the few metrics that measures behavioral response to a specific scenario. A click rate showing decline over time suggests people are getting better at recognizing phishing. A 25% click rate on a moderately sophisticated simulation means about a quarter of your organization fell for it. A 5% rate means people are much better.

But click rate has significant limitations. It's specific to simulated phishing and real phishing might look different. A declining click rate might reflect that people now know simulations happen and are more cautious generally, not that they're better at real phishing. If you had a 30% click rate in simulation one and a 15% rate in simulation two, did awareness improve? Or is simulation two just harder to fall for than simulation one?

More valuable than click rate alone is phishing reporting rate. What percentage of people who receive suspicious emails actually report them to security? This shows active behavioral change. If phishing reporting increases while click rates decrease, you've got evidence that people are not only avoiding the bait but taking action. If click rates decrease but reporting stays flat, people might just be deleting suspicious emails rather than recognizing the threat.

The metrics work better together than separately. Track both click rates and reporting rates. Look at trends over time. Track incident trends—are phishing-based breaches actually decreasing? A dashboard that shows "click rate down 20%, reporting rate up 30%, no phishing-based breaches in past 12 months" tells a much clearer story than any single metric.

Training Completion vs. Comprehension vs. Retention

Training completion is what most organizations measure: what percentage of people completed the course? It's clean and easy. It's also mostly meaningless. A person who clicks through an online training module while checking email every 30 seconds has completed the training but learned nothing.

Comprehension—did people understand the material—is harder to measure but more meaningful. Quiz scores immediately after training can indicate comprehension. A person who scores 80% on a quiz probably understood the material better than someone who didn't take the quiz. But immediate recall is not the same as understanding, and understanding is not the same as remembering.

Retention—do people remember the training weeks or months later—is what actually matters for behavior change. A quiz immediately after training tells you if someone paid attention. A quiz two weeks later tells you if they retained anything. Most organizations don't measure retention because it requires follow-up, which is more work.

The practical approach is: measure completion (easy), measure comprehension (harder), and assume retention decays without reinforcement (so track reinforcement efforts). If someone took training in January and received monthly reminders, they're more likely to retain the information in June than if they took training in January and heard nothing else until July.

Behavior Change Indicators: Indirect but Meaningful

Behavior change is the goal but hard to measure directly. You can't know for certain that an employee's behavior changed just from metrics. But you can look at indirect indicators that suggest behavior has changed.

For phishing awareness: simulation click rates decline, phishing reporting rates increase, phishing-based incidents decrease, employees report suspicious emails proactively. Together, these suggest people are actually paying more attention to phishing.

For password security: audits show increasing percentages of accounts with strong passwords, password reset requests increase (people creating new passwords), shared account incidents decrease. These suggest password practices are improving.

For access control: requests to remove unnecessary access increase (people reviewing and reducing permissions), access reviews show users proactively requesting removal of old access. This suggests people are thinking about least-privilege access.

For incident reporting: reported incidents increase (more people reporting), time to detection decreases (internal detection beats external discovery), incident severity decreases for reported incidents (caught faster, less damage). This suggests people are alert and reporting threats.

These indicators suggest behavior change but don't prove it. Click rates might decline for reasons unrelated to awareness. Password strength might improve because of technical controls (password managers, complexity requirements) rather than behavior change. Incident reporting might increase because of new systems or incentives rather than awareness. But multiple indicators pointing the same direction build a body of evidence.

Attribution and Causality: The Hard Problem

Even if you see improvements—phishing incidents decrease, click rates decline, reporting increases—can you attribute these improvements to your awareness program? Or did they result from other factors: better email filtering, new technical controls, endpoint detection deployment, threat actor changes, or simple luck?

True attribution requires controlling for other variables, which is nearly impossible in real organizations. You can't run a control group where they don't get awareness training and compare to a treatment group that does. Organizations change multiple things simultaneously.

The best practical approach is to build a body of evidence. Do incident reports from caught phishing suggest people are more aware? When you interview employees, do they demonstrate better security knowledge? Are captured phishing emails showing higher sophistication, suggesting attackers are working harder because easy targets are gone? When senior leaders discuss security maturity, do they reference specific awareness activities that contributed?

Research on awareness program effectiveness shows programs do have impact but the impact is modest. Most studies find awareness programs reduce fall-for rates by 20-30% but effects decay without reinforcement. This is valuable improvement but not transformational.

The honest assessment is that causality is hard to prove. But you can accumulate evidence that the program is having effect. And when multiple indicators trend in the positive direction, you can have reasonable confidence that the program is contributing to improved security.

Building a Multi-Level Measurement Framework

A practical measurement framework has multiple metric types, each answering different questions.

Activity metrics track what you do: training offered, simulations sent, communications distributed, time invested. Use these to assess program scope and frequency. If you're running simulations every six months, that's different from quarterly or monthly. If you're sending one newsletter per year, that's different from monthly. Activity metrics help you understand whether the program is actually active and consistent.

Awareness metrics track whether people are paying attention: quiz results, survey responses about security practices, self-reported awareness of specific threats. Use these to assess whether information is reaching people. A spike in quiz scores immediately after training and a decline months later tells you that retention is fading and you need more reinforcement.

Behavior metrics track actions people take: phishing reporting rates, access review participation, password change patterns, incident reports filed. Use these to assess whether behavior is actually changing. This is the gap that matters—the difference between knowing what to do and actually doing it.

Outcome metrics track business impact: phishing-based breaches, incident cost, dwell time, mean time to detection. Use these to assess real value. These are hard to measure and attribution is complicated, but they're what ultimately matters.

Track metrics over time to show trends. A single data point is meaningless. Five percent click rate on one simulation tells you nothing. Showing that click rates have gone from 30% to 15% over a year shows improvement. Showing that click rates dropped sharply in the month after training and gradually increased over the following months shows the decay that research predicts.

Set improvement targets: "reduce phishing click rate by 5% per year," "increase phishing reporting by 10% annually," "decrease phishing-based incidents by 20% annually." Targets give direction and create accountability. Without targets, metrics become historical documentation rather than management tools.

Regular review—monthly or quarterly assessment of metrics—keeps the program active and allows adjustment. If you notice click rates increasing in June, that's a signal to increase reinforcement. If phishing reporting is flat despite declining click rates, that's a signal you need to encourage reporting more actively.

The closing insight is that a framework with multiple metric types gives better visibility than any single metric. You can't evaluate program effectiveness from completion rates alone. You can't prove behavior change from awareness metrics alone. But when you track activity, awareness, behavior, and outcomes over time, and see trends pointing in a consistent direction, you have evidence that the program is working.


Fully Compliance provides educational content about IT compliance and cybersecurity. Measurement frameworks should be tailored to your organization's specific programs, available data sources, and goals. This article reflects general principles of measurement and the inherent limitations of attribution in complex environments. Consult with your security team and analytics expertise about building measurement frameworks specific to your organization.