The Three-Rule Framework in Cybersecurity: Blue, Red, and Purple Team Applications
This is the fifth post in a series about the three-rule framework — Force Blank, Penalize Guessing, Show the Source. So far, the framework has been applied to document extraction, worldbuilding, business scenario building, and implementation at scale.
This post applies it to a domain where the stakes are as high as they get: cybersecurity.
The structural problem is familiar. AI generates confident-sounding security assessments that blend verified findings with inferences and assumptions — and nobody can tell which is which. A vulnerability scan result sits next to an architectural inference sits next to an untested assumption about network segmentation, all presented with the same confidence. The output looks thorough. Some of it is grounded. Some of it was invented to make the narrative hold together.
In business scenario planning, that’s a waste of planning resources. In cybersecurity, it’s a breach waiting to happen.
Why Cybersecurity Is Especially Vulnerable
A 2025 survey paper on hallucinations in AI-driven cybersecurity systems identifies the core risk: AI models optimized for fluency can misidentify safe activities as threats or miss actual dangers, and they do so with the same confident tone regardless of whether their assessment is grounded in evidence. The paper notes that the design goal of fluency over accuracy in many LLMs reinforces this problem — models produce highly confident and coherent statements even when they lack factual grounding.
The OWASP Foundation’s 2025 guidance lists AI hallucination among the top risks to LLM-powered security tools. The concern isn’t just about false positives or false negatives in isolation — it’s about the blending of real findings and fabricated assessments into a single output that humans then treat as uniformly reliable.
This maps directly onto the problem the three-rule framework was built to solve. And it adapts cleanly to all three team perspectives.
The Cybersecurity Source Tags
In every previous domain, the framework distinguishes between what’s established, what’s derived, and what’s unknown. Cybersecurity needs the same distinction, but the specific tags change depending on which team perspective you’re operating from — because each team has a different “source of truth.”
The gap labels, however, are universal across all three teams:
[VISIBILITY GAP]— something the team cannot see or hasn’t instrumented: unmonitored network segments, missing log sources, shadow IT, cloud services without telemetry[VALIDATION GAP]— a control or detection rule that exists on paper or in configuration but has never been tested against actual attack behavior
The distinction between these two is critical. A visibility gap means you don’t have the data. A validation gap means you have the data (or think you do) but haven’t proven the detection actually works. Both are dangerous; they require different responses.
Blue Team: Defense
The Blue Team’s source of truth is: what has been confirmed about the defensive posture through direct evidence — not what the policy says, not what the vendor promises, not what should be true.
This distinction matters more than it might seem. Research from Mitiga suggests that traditional SIEMs detect only about 21% of MITRE ATT&CK techniques on average — meaning roughly four out of five technique categories have no validated detection coverage. Yet many organizations report much higher coverage numbers, because they count configured rules rather than validated detections. The gap between “we have a rule for this” and “this rule has fired correctly against a real attack” is exactly the gap the three-rule framework surfaces.
AttackIQ’s March 2026 analysis captured this perfectly: coverage isn’t static, environments change, and a detection that worked last month may have silently broken. An attendee at one of their webinars put it bluntly: having detections means nothing if they don’t produce signal the team can act on.
Blue Team Source Tags
(VERIFIED)— confirmed through direct evidence: tested control, observed log entry, validated configuration, scan result with timestamp(POLICY-STATED)— documented in policy, configuration guides, or vendor documentation but not independently validated. “Our policy says MFA is enforced” is not the same as “MFA is enforced”(INFERRED)— derived from indirect indicators. “No alerts in 90 days” doesn’t mean “no intrusions in 90 days” — it might mean the detection isn’t working
Blue Team Prompt
You are a defensive security analyst reviewing our environment. For every security claim, tag it with its evidence basis:
• (VERIFIED) — confirmed through direct testing, log evidence, or scan results. Cite the specific evidence.
• (POLICY-STATED) — documented in policy or configuration guides but not independently validated. Flag what validation would look like.
• (INFERRED) — derived from indirect indicators. State the inference chain and what could invalidate it.
• When you identify a control or coverage area that hasn’t been tested or instrumented, flag it as [VISIBILITY GAP] or [VALIDATION GAP] with a brief explanation.
• A false sense of security is more dangerous than a known gap. A known gap gets resourced. False confidence gets breached. When in doubt, flag the gap.
Blue Team Example Output
| Control | Status | Source | Note |
|---|---|---|---|
| MFA on VPN | Enabled | POLICY-STATED | Azure AD conditional access policy requires MFA. [VALIDATION GAP: no penetration test has verified enforcement on the Cisco AnyConnect endpoint] |
| EDR coverage | 94% of endpoints | VERIFIED | CrowdStrike dashboard, pulled April 2026 |
| Lateral movement detection | Active | INFERRED | SIEM has rules for pass-the-hash. No Purple Team exercise has tested whether they fire. [VALIDATION GAP] |
| DNS exfiltration detection | — | VISIBILITY GAP | No DNS logging configured on internal resolvers. Cannot detect DNS tunneling. |
The crucial row is the first one. Without the framework, the MFA assessment would read “MFA: Enabled” — green on the dashboard, confidence high, nobody worries. With the framework, you see it’s POLICY-STATED, not VERIFIED, with a specific validation gap. That gap is exactly what an attacker would exploit.
Red Team: Offense
The Red Team’s source of truth is: what has been demonstrated through actual exploitation — not what should be vulnerable, not what Shodan shows, not what the CVE database suggests.
The risk of unflagged assumptions in Red Team reporting is that defensive resources get allocated to the wrong priorities. A hypothesized attack path reported as confirmed sends the Blue Team chasing a phantom while the actual exposure remains unaddressed.
Red Team Source Tags
(CONFIRMED)— vulnerability exploited, access achieved, data exfiltrated — with evidence (screenshot, hash, artifact, session log)(INDICATED)— strong indicators of vulnerability but not yet exploited: version fingerprint matches a known CVE, default credentials detected but not tested, open port with a known-vulnerable service banner(HYPOTHESIZED)— attack path exists in theory based on architecture review but no testing has been performed. States the assumptions the path depends on
Red Team Prompt
You are a red team operator analyzing potential attack paths. For every finding or attack path, tag it:
• (CONFIRMED) — vulnerability exploited and access demonstrated. Cite the specific evidence (tool output, artifact, hash).
• (INDICATED) — strong indicators suggest vulnerability but exploitation has not been attempted. State the indicators you’re relying on.
• (HYPOTHESIZED) — attack path is theoretically plausible based on architecture but untested. State the assumptions the path depends on.
• Flag untested attack surfaces as [UNTESTED SURFACE: description].
• Flag assumed network paths or trust relationships as [ASSUMED PATH: what connectivity or trust is assumed].
• A hypothesized vulnerability reported as confirmed wastes defensive resources on the wrong priority. When not yet tested, say so.
Red Team Example Output
| Finding | Severity | Source | Detail |
|---|---|---|---|
| Jenkins RCE (CVE-2024-XXXX) | Critical | CONFIRMED | Exploited via Metasploit, reverse shell obtained. Artifact: session log ID 47 |
| Lateral movement Jenkins → DB | High | HYPOTHESIZED | Jenkins on VLAN 10, DB on VLAN 20. [ASSUMED PATH: assumes no ACL between VLANs — not validated] |
| S3 bucket public access | Medium | INDICATED | Bucket policy allows s3:GetObject for *. [UNTESTED SURFACE: no attempt to download sensitive content] |
| Domain admin via Kerberoasting | High | INDICATED | SPN found on service account with weak encryption (RC4). Hash not yet cracked. |
The lateral movement finding is where the framework earns its keep. Without tags, the report reads “Jenkins server can be used to pivot to the database” — which sounds confirmed. With the framework, you see it’s HYPOTHESIZED with an ASSUMED PATH. The Blue Team’s response shifts from “patch this immediately” to “first, verify whether the VLAN segmentation actually blocks this traffic.”
Purple Team: Collaboration
Purple Teaming is where the framework becomes most powerful, because the entire point is to map offensive capabilities against defensive capabilities — and the space between them is exactly what the gap labels surface.
A Purple Team exercise structured with the framework produces a matrix that answers four distinct questions for each technique: Did the Red Team succeed? Did the Blue Team detect it? Where did detection fail — at the visibility layer or the logic layer? And which techniques weren’t tested at all?
That last question — what wasn’t tested — is often the most important finding of a Purple Team engagement, and the one most likely to go unreported without explicit tagging.
Purple Team Source Tags
(DETECTED)— Blue Team successfully detected and alerted on the attack technique. Cites the specific alert, rule, or SIEM correlation(PARTIALLY DETECTED)— some indicators were logged but not correlated into an actionable alert. The raw data exists; the detection logic doesn’t(MISSED)— attack technique succeeded without triggering any detection. Highest-priority finding(NOT TESTED)— this MITRE ATT&CK technique was not exercised in this engagement. Detection capability cannot be assessed
Purple Team Prompt
You are a purple team analyst mapping attack results against detection capabilities. For each technique tested, tag the outcome:
• (DETECTED) — Blue team detected and alerted. Cite the specific detection rule, alert, or correlation.
• (PARTIALLY DETECTED) — indicators were logged but not correlated into an alert. State what was logged and what detection logic is missing.
• (MISSED) — attack succeeded without detection. This is the highest-priority finding.
• (NOT TESTED) — technique was not exercised in this engagement. Do not assess detection capability for untested techniques.
• For every MISSED finding, flag the root cause: [VISIBILITY GAP] (data not collected) or [DETECTION GAP] (data collected but no rule or correlation exists).
• Do not assume a control works because it exists. A SIEM rule that has never fired against a real attack is a [VALIDATION GAP], not a (DETECTED).
• An untested technique reported as detected gives false confidence. When not tested, say NOT TESTED — never extrapolate.
Purple Team Example Output
| MITRE Technique | Red Result | Blue Result | Source | Action |
|---|---|---|---|---|
| T1566.001 Spearphishing | Payload delivered | Alert fired in 12 min | DETECTED | Review: 12 min mean-time-to-detect acceptable? |
| T1003.001 LSASS dump | Credentials extracted | No alert | MISSED | [VISIBILITY GAP: no LSASS access monitoring configured] |
| T1071.001 Web C2 | C2 channel established | Proxy logged traffic, no alert | PARTIALLY DETECTED | [DETECTION GAP: beaconing pattern recognition needed] |
| T1053.005 Scheduled Task | — | — | NOT TESTED | [VALIDATION GAP: persistence techniques not in scope] |
Read the matrix from left to right, and it tells you exactly where you stand: one technique is covered, one is blind, one has raw data but no intelligence, and one hasn’t been evaluated. That’s a decision-ready assessment. No AI-generated optimism obscuring the picture.
The Cybersecurity Version of Rule 2
The “penalize guessing” rule takes on particular urgency in security contexts. In document extraction, a wrong answer wastes time. In cybersecurity, the costs compound differently:
- A
(HYPOTHESIZED)finding reported as(CONFIRMED)→ defensive resources go to the wrong place - A
(POLICY-STATED)control reported as(VERIFIED)→ actual exposure stays unaddressed - A
(NOT TESTED)technique reported as(DETECTED)→ the team believes they’re protected when they’re blind - A
[VISIBILITY GAP]that goes untagged → the adversary operates in the one area nobody is watching
The cybersecurity formulation of Rule 2 is:
False confidence in a security control is more dangerous than a known gap. A known gap gets resourced. False confidence gets breached.
This isn’t hypothetical. The pattern of “assumed coverage, actual blindness” is precisely what major breaches exploit. Attackers don’t target your strongest defenses — they find the areas where you think you’re covered but aren’t. Every [VALIDATION GAP] the framework surfaces is a place the adversary would probe. Every unflagged assumption about detection capability is an invitation.
The Cross-Team Pattern
Looking across all three team perspectives, the framework reveals a consistent pattern:
| Team | Source of Truth | Source Tags | Gap Labels | Rule 2 Formulation |
|---|---|---|---|---|
| Blue | Confirmed defensive posture | VERIFIED / POLICY-STATED / INFERRED | VISIBILITY GAP / VALIDATION GAP | A policy-stated control treated as verified is an open door |
| Red | Demonstrated exploitation | CONFIRMED / INDICATED / HYPOTHESIZED | UNTESTED SURFACE / ASSUMED PATH | A hypothesized path treated as confirmed wastes defense resources |
| Purple | Validated detection capability | DETECTED / PARTIALLY DETECTED / MISSED / NOT TESTED | VISIBILITY GAP / DETECTION GAP / VALIDATION GAP | An untested technique treated as detected creates false confidence |
The structure is the same in every case: distinguish between what’s been proven and what’s been assumed, make the boundary visible, and treat false confidence as worse than acknowledged uncertainty.
Or, to put it in terms any security professional will recognize: the framework turns “assumed coverage” into “validated coverage” — and makes everything in between explicit.
Integration with MITRE ATT&CK
The three-rule framework maps naturally onto MITRE ATT&CK assessments. The ATT&CK matrix gives you the what — which techniques exist. The framework gives you the how confident — which techniques you’ve actually validated coverage for.
A standard ATT&CK heatmap shows red (no coverage) and green (coverage). The three-rule framework adds the crucial middle layer: green that’s been validated vs. green that’s been assumed. As AttackIQ notes, the gap between a green box and actual defensive capability can be enormous. Coverage for one procedure is not coverage for a technique. A detection that worked last month may have silently broken this month. Without continuous validation, your heatmap shows what you deployed, not what works.
Adding the three-rule tags to your ATT&CK coverage assessment turns it from a deployment map into a confidence map — and confidence maps are what security decisions should be based on.
Sources and Further Reading
- CAEE Journal (April 2025): “The Paradigm of Hallucinations in AI-driven Cybersecurity Systems.” Taxonomy of hallucination impacts on cybersecurity tools.
- AttackIQ (March 2026): “What Does MITRE ATT&CK Coverage Really Mean?” On the gap between claimed coverage and validated detection, including the distinction between procedure-level and technique-level coverage.
- Mitiga (2025): “Measurements That Matter.” Reports ~21% average ATT&CK detection rate for traditional SIEMs.
- Kroll Cyber Risk (2023): “MITRE ATT&CK Detection Maturity Assessment Guide.” Template-based approach to identifying coverage gaps.
- OWASP / Hacken (2025): “LLM Security Frameworks: A CISO’s Guide.” On NIST AI RMF, ISO 42001, and hallucination monitoring requirements.
- MITRE ATT&CK: attack.mitre.org. The knowledge base of adversary tactics and techniques.
- Previous posts in this series:
Post 1: Three Prompt Rules That Stop AI From Guessing — And the Science Behind Them
Post 2: From Document Extraction to Alternate History
Post 3: The Three-Rule Framework for Scenario Building
Post 4: Implementing the Framework: Calibration, Governance, and Trade-offs

