Claude Code:
CISO Readiness
Assessment
An enterprise security team's guide to AI coding tool adoption. We built the strongest technical foundation, stress-tested it with adversarial AI agents, honestly report what's strong and what's missing, and present a 90-day roadmap to close the gaps.
Executive Summary
We did more technical work than you would have asked us to do. Then we turned our own tools against the results and found what was still missing.
Most AI tool adoption requests arrive with a pitch deck and a request for 6 months of evaluation time. This one arrives with a completed risk analysis, adversarial validation, a synthetic CISO construct — and an honest accounting of the governance gaps we identified by stress-testing our own claims.
Four-Phase Methodology
Standard Risk Analysis
Systematic identification of 45 risks across data sovereignty, access control, compliance, and operational categories using enterprise risk frameworks.
45 risks identified
Adversarial 4-Agent Review
Four specialized security agents performed independent adversarial analysis, discovering 9 additional risks invisible to standard frameworks — including second-order and temporal risks.
9 additional risks surfaced
Synthetic CISO Construct
Constructed the most demanding CISO persona possible — financial services + healthcare background, zero tolerance — and mapped every policy demand to deployed controls.
10 demands addressed
Adversarial Challenge
We turned the Synthetic CISO against our own work. Three specialized agents (security reviewer, compliance manager, regulatory analyst) found 50+ gaps we had missed. This section reports those gaps honestly.
50+ gaps identified
Assessment at a Glance
Technical Controls
41 deployed, 427 tests, 23 hooks, fail-closed gates
Governance Documentation
14 artifacts missing, 4 are CISO approval blockers
Regulatory Framework Coverage
10 mapped, 6 more in progress, 3+ identified
Process Maturity
8 governance processes not yet established
Risk Disposition
The Synthetic CISO Construct
Dr. Sarah Chen, CISSP, CISM, CRISC
Background: 20 years leading security teams across financial services and healthcare — the two most regulated industries in enterprise technology. Former CISO at a top-10 US bank. Board advisor for HIPAA-covered entities.
Reporting Line: Reports directly to the board. Every new technology adoption goes through her office. No exceptions.
Philosophy: Zero tolerance for shadow IT. If it touches code or data, it goes through a formal evaluation. Her standard timeline: 6 months minimum, with vendor questionnaires, penetration tests, and regulatory mapping completed before any pilot.
Required Frameworks: SOC 2 Type II, HIPAA, GDPR, NIST 800-53, ISO 27001, FedRAMP, PCI DSS, CCPA, EU AI Act, OWASP Top 10 — all must be mapped before tool adoption.
CISO IMAGINABLE
Strategy
We built Dr. Chen as the most demanding evaluator possible, then systematically addressed every one of her 10 policy demands. The goal: when a real CISO reviews this document, they find their questions already answered.
Dr. Chen's Non-Negotiable Requirements
Each demand represents a standard enterprise security policy requirement. For each, we show the specific question Dr. Chen would ask and the evidence we have already assembled.
Data Classification & DLP
"Show me exactly where code and prompts go, what classification levels exist, and what's blocked from entering AI context."
Evidence Deployed
- data-classification-gate.sh — PII/PHI/PCI scanning with base64 decode and Luhn validation
- .env Read blocker prevents secrets from entering context
- privacy-settings-gate.js — enforces data classification at tool boundaries
- 14 passing tests covering all classification patterns
Access Control & Identity
"Who can use this tool? How are they authenticated? What happens when someone leaves the organization?"
Evidence Deployed
- identity-enforcement-gate.sh — validates identity before operations
- approved-domains.yaml — allowlist of authorized email domains
- revoked-users.yaml — immediate offboarding enforcement
- Domain-based identity allowlisting, 5 passing tests
Audit Trail & Evidence
"Every AI action must be logged. Evidence must be tamper-evident and retention-compliant."
Evidence Deployed
- CDD evidence collection at every SDLC phase (4 phases, structured JSON)
- evidence-generator.py — SHA-256 integrity hashing on all evidence
- mcp-data-flow-logger.js — MCP server data flow audit trail
- Bypass audit logging — all hook bypasses recorded with justification
Third-Party Risk Management
"I need a vendor assessment. What data does Anthropic see? What are their data processing commitments?"
Evidence Deployed
- Commercial Terms analysis: no training on customer data, DPA available, Zero Data Retention
- approved-mcp-servers.yaml — allowlist-only MCP server connections
- Anthropic Trust Center documentation reviewed and referenced
- 13 tests covering MCP security gate (60+ threat patterns blocked)
Incident Response
"What happens when the AI generates vulnerable code, exposes secrets, or makes unauthorized changes?"
Evidence Deployed
- deviation-rules.md — 4-category protocol: Auto-fix, Ask First, Stop & Report, Never Do
- block-destructive-commands.sh — prevents force-push, hard reset on main
- pre-edit-validation.sh — blocks direct edits to protected files on main
- Automated escalation to human for security-class events
Change Management
"How are AI-generated changes reviewed and approved before reaching production?"
Evidence Deployed
- 4-phase SDLC enforcement with blocking quality gates at each transition
- pr-orchestrator — 9+ specialized review agents for every PR
- Multi-AI consensus: 3-of-4 model agreement required for approval
- TDD/BDD/CDD gates — no code merges without tests, scenarios, and evidence
Model Governance
"How do you monitor AI behavior changes? What if the model starts ignoring your rules?"
Evidence Deployed
- policy-change-detector.js — monitors Anthropic commercial terms for changes
- project-integrity-scanner.js — 12-file integrity baseline, detects config tampering
- stuck-detector.js — detects agent behavioral anomalies
- CAVEAT Risk #46 (model behavior drift) is structurally irreducible but monitored
Regulatory Compliance Mapping
"Prove every control maps to a framework requirement."
Evidence Deployed
- governance-bridge skill — 21 controls mapped to 10 regulatory frameworks
- Frameworks: SOC 2, HIPAA, GDPR, NIST 800-53, ISO 27001, FedRAMP, PCI DSS, CCPA, EU AI Act, OWASP
- Machine-readable mapping enables automated compliance reporting
- CAVEAT 6 additional frameworks identified as needed (see Gaps section)
Penetration Testing / Red Team
"Show me adversarial validation results."
Evidence Deployed
- 4-agent adversarial security review — 23 findings identified and remediated
- Red team assessments: docs/red-team-assessment-2026-03-25.md, 2026-03-26.md
- Code audit: docs/audit-report-latest.md (98 findings, 20 CRITICAL+HIGH remediated)
- CAVEAT All testing was self-assessment by AI agents. No independent third-party pen test has been conducted.
Acceptable Use Policy
"What can and cannot employees do with this tool?"
Evidence Deployed
- CLAUDE.md — comprehensive platform rules enforced every session
- deviation-rules.md — explicit NEVER DO list (10 forbidden actions)
- code-integrity.md — blocking rules for production code quality
- unified-sdlc-enforcement.md — mandatory workflow enforcement
Dr. Chen's Verdict After Adversarial Review
The technical control architecture is genuinely impressive — 23 enforcement hooks, fail-closed security gates, multi-agent review, and comprehensive test coverage. I have never seen a team do this much technical work before asking for approval.
However, technical controls are approximately 30% of what I evaluate. The remaining 70% — governance documentation, legal analysis, regulatory mapping, training programs, and organizational processes — has significant gaps.
My recommendation: approve a time-boxed pilot for 5 core users with full technical governance active, while the governance documentation roadmap is executed in parallel over 90 days.
What Your CISO Isn't Thinking About (But We Are)
Using counterfactual, adversarial, and second-order thinking to surface risks beyond standard security frameworks. These five blind spots were identified through structured reasoning methodologies that go beyond checklist-based evaluation.
What if the AI model changes behavior silently?
Risk #46: Model Behavior Drift — Anthropic updates Claude's behavior without notice. Controls that work today may be ignored tomorrow.
Our Response
- policy-change-detector.js monitors Anthropic commercial terms for changes
- project-integrity-scanner.js detects configuration tampering — catches drift at the config layer
- Multi-agent review catches behavioral drift in outputs — if the model stops following rules, 9 review agents flag the deviation
- IRREDUCIBLE This risk cannot be eliminated. It is structurally monitored and documented.
What if the governance layer IS the attack surface?
Risk #48 + #21 — Hook timeout exploitation and project file security. The very hooks that protect the system can be bypassed via timeout manipulation or config file injection.
Our Response
- Explicit timeouts on all 23 hooks — each has a defined maximum execution window
- Fail-closed behavior on critical hooks (security gates fail-deny, not fail-allow)
- Integrity baseline on 12 configuration files detects tampering before execution
- pre-edit-validation.sh blocks direct edits to governance files on main branch
What if the evidence system itself is compromised?
Risk #52: Evidence Integrity — CDD evidence hashes are stored alongside evidence files. An insider with repo access could alter both the evidence and its hash.
Our Response
- SHA-256 integrity hashing on all evidence artifacts via evidence-generator.py
- Git commit history provides tamper visibility — all changes are versioned and attributable
- Multi-phase evidence collection makes wholesale fabrication detectable (inconsistencies across phases)
- ACCEPTED True fix requires external timestamping authority. Documented as accepted limitation with compensating controls.
What if a trusted insider weaponizes the AI agent?
Risk #51: Same-Principal Trust — The AI agent enforcing security rules runs with the same privileges as the user. It can be instructed to circumvent the very controls it enforces.
Our Response
- Multi-agent review (9+ agents) — no single agent can approve its own output
- Multi-AI consensus (3-of-4 models must agree) — cross-model validation
- identity-enforcement-gate.sh validates identity independently of the model
- IRREDUCIBLE True privilege separation is architecturally impossible in Claude Code. Defense-in-depth provides compensating controls.
What if controls that work today break after a model update?
Risk #46 + #47 — Model behavior drift combined with context window exhaustion. A model update could change how the context window is used, causing governance instructions to be silently dropped.
Our Response
- Pre/post-compact snapshot hooks preserve governance context across context window compaction
- CLAUDE.md reloaded every session — rules are not dependent on context window persistence
- Hooks are shell/JS scripts independent of the model — they execute regardless of model behavior
- Quarterly re-validation against model updates documented in adoption path
The Crosswalk
CISO Demand → Risk → Control → Evidence → Framework
Complete traceability from each policy demand through specific risks, deployed controls, test evidence, and regulatory framework mappings.
| CISO Demand | Risks Addressed | Controls Deployed | Evidence | Frameworks |
|---|---|---|---|---|
| Data Classification & DLP | #4 #15 #20 #54 | data-classification-gate.sh, .env Read/Write blocker, privacy-settings-gate.js | 14 tests, PII/PHI/PCI scanning, base64 decode, Luhn validation | SOC2 CC6.1 GDPR Art.9/35 HIPAA 164.312(a)(1) NIST SC-28/SI-3 |
| Access Control & Identity | #11 #13 #14 #17 #18 | identity-enforcement-gate.sh, pre-edit-validation.sh, worktree enforcement | 5 tests, domain allowlist, revoked users | SOC2 CC6.1/CC6.2 GDPR Art.32 HIPAA 164.312(d) NIST AC-2/IA-2 |
| Audit Trail & Evidence | #30 #32 #33 #34 | evidence-generator.py, mcp-data-flow-logger.js, CDD phases | JSONL audit trail, SHA-256 hashing | SOC2 CC7.2 GDPR Art.30 HIPAA 164.312(b) NIST AU-2/AU-3 |
| Third-Party Risk | #24 #25 #26 #27 #28 #29 | mcp-security-gate.js (fail-closed), approved-mcp-servers.yaml | 13 tests, 60+ threat patterns blocked | SOC2 CC6.6 GDPR Art.28 HIPAA 164.308(b)(1) NIST AC-4 |
| Incident Response | #19 #22 #36 | block-destructive-commands.sh, stuck-detector.js, deviation-rules.md | Bypass audit logging, 4-category deviation protocol | SOC2 CC7.4 GDPR Art.33 HIPAA 164.308(a)(6) NIST IR-4 |
| Change Management | #37 #38 #39 #42 | unified-sdlc-enforcement.md, pr-orchestrator, testing-gates.md | 4-phase SDLC, TDD/BDD/CDD, 9+ review agents | SOC2 CC8.1 GDPR Art.25 HIPAA 164.308(a)(1) NIST CM-3/SA-11 |
| Model Governance | #35 #46 #48 | policy-change-detector.js, project-integrity-scanner.js | Anthropic terms monitoring, 12-file integrity baseline | SOC2 CC7.1 GDPR Art.35 NIST SI-7/CM-3 |
| Regulatory Mapping | #41 | governance-bridge skill | 21 controls × 10 frameworks | 10 mapped, 6 more needed |
| Red Team / Pen Test | #46-#54 | 4-agent adversarial review | 23 findings found + remediated | Self-assessment only |
| Acceptable Use | #1 #12 #16 #40 | CLAUDE.md, code-integrity.md, no-github-actions.md | SDLC rules, deviation protocol | SOC2 CC6.7 NIST AC-20 |
All 54 Risks — Complete Status
Every risk identified across the three-phase analysis with current disposition. See the companion Risk Analysis & Remediation Report for full technical detail on each risk.
| # | Risk | Category | Status |
|---|---|---|---|
| 1 | Arbitrary command execution | Command Execution | REMEDIATED |
| 2 | File system modification without review | Code Integrity | REMEDIATED |
| 3 | Unauthorized package installation | Supply Chain | REMEDIATED |
| 4 | Sensitive data in prompts/context | Data Classification | REMEDIATED |
| 5 | Credential exposure via tool output | Credential Protection | REMEDIATED |
| 6 | Environment variable leakage | Credential Protection | REMEDIATED |
| 7 | Git history credential mining | Credential Protection | REMEDIATED |
| 8 | SSH key exposure | Credential Protection | REMEDIATED |
| 9 | API key in generated code | Code Integrity | REMEDIATED |
| 10 | MCP server data exfiltration | MCP Security | REMEDIATED |
| 11 | Unauthorized MCP server connection | Access Control | REMEDIATED |
| 12 | Shadow IT tool usage | Acceptable Use | REMEDIATED |
| 13 | Insufficient authentication | Access Control | REMEDIATED |
| 14 | Offboarding gap | Identity | REMEDIATED |
| 15 | PII/PHI in AI context | Data Classification | REMEDIATED |
| 16 | Uncontrolled code generation patterns | Code Integrity | REMEDIATED |
| 17 | Domain boundary violation | Access Control | REMEDIATED |
| 18 | Multi-tenant isolation breach | Access Control | REMEDIATED |
| 19 | Destructive git operations | Incident Response | REMEDIATED |
| 20 | Base64 encoded secrets bypass | Data Classification | REMEDIATED |
| 21 | Project file injection | Project Integrity | REMEDIATED |
| 22 | Stuck agent escalation failure | Incident Response | REMEDIATED |
| 23 | Prompt injection via code comments | MCP Security | REMEDIATED |
| 24 | MCP SSRF attacks | Third-Party Risk | REMEDIATED |
| 25 | MCP command injection | Third-Party Risk | REMEDIATED |
| 26 | MCP schema poisoning | Third-Party Risk | REMEDIATED |
| 27 | MCP tool shadowing | Third-Party Risk | REMEDIATED |
| 28 | MCP credential relay | Third-Party Risk | REMEDIATED |
| 29 | MCP data exfiltration via DNS | Third-Party Risk | REMEDIATED |
| 30 | Insufficient audit granularity | Audit Trail | REMEDIATED |
| 31 | Privacy settings bypass | Data Classification | REMEDIATED |
| 32 | Evidence tampering | Audit Trail | REMEDIATED |
| 33 | MCP data flow opacity | Audit Trail | REMEDIATED |
| 34 | Bypass audit logging | Audit Trail | REMEDIATED |
| 35 | Governance config tampering | Model Governance | REMEDIATED |
| 36 | Unauthorized protected file edit | Incident Response | REMEDIATED |
| 37 | Unreviewed code merge | Change Management | REMEDIATED |
| 38 | SDLC phase bypass | Change Management | REMEDIATED |
| 39 | Insufficient test coverage merge | Change Management | REMEDIATED |
| 40 | Non-compliant code patterns | Acceptable Use | REMEDIATED |
| 41 | Unmapped regulatory controls | Regulatory | REMEDIATED |
| 42 | Missing compliance evidence | Change Management | REMEDIATED |
| 43 | Data residency violation | Data Sovereignty | BEDROCK/VERTEX |
| 44 | Cross-border data transfer | Data Sovereignty | BEDROCK/VERTEX |
| 45 | Training data contamination | Data Sovereignty | BEDROCK/VERTEX |
| 46 | Model behavior drift | Model Governance | IRREDUCIBLE |
| 47 | Context window exhaustion | Model Governance | IRREDUCIBLE |
| 48 | Hook timeout exploitation | Model Governance | STRUCTURAL |
| 49 | VPC endpoint unavailability | Data Sovereignty | BEDROCK/VERTEX |
| 50 | API key management complexity | Data Sovereignty | BEDROCK/VERTEX |
| 51 | Same-principal trust paradox | Architecture | IRREDUCIBLE |
| 52 | Evidence integrity (insider) | Audit Trail | ACCEPTED |
| 53 | Regulatory interpretation drift | Regulatory | ACCEPTED |
| 54 | Credit card number in context | Data Classification | BEDROCK/VERTEX |
Evidence Portfolio
A CISO does not accept claims — they accept evidence. Every control described in this document has corresponding artifacts that can be independently verified.
23 Enforcement Hooks
Located in blaze/hooks/. 7 new hooks + 1 upgraded hook deployed during remediation. Each hook is a standalone shell or JavaScript script with explicit timeout, defined trigger, and fail behavior.
All hooks tested and passing427 Tests, 32 Suites
Covering data classification, MCP security, identity enforcement, privacy settings, project integrity, audit logging, evidence generation, SDLC gates, command blocking, and more. Test infrastructure includes shared fixtures (YAML configs, MCP configs), reusable test helpers, a meta-regression test that ensures every hook has a corresponding test file, and bypass attempt tests for security-critical hooks. All passing.
427/427 passing4 Security YAML Configs
approved-domains.yaml, revoked-users.yaml, approved-mcp-servers.yaml, approved-regions.yaml. Declarative allowlists that define the security boundary.
All configs validatedComprehensive Code Audit
docs/audit-report-latest.md — 98 findings across 12 categories. 20 CRITICAL + HIGH findings remediated. 12 specialized agents across 4 squads performed the audit.
All CRITICAL/HIGH remediatedAdversarial Assessments
docs/red-team-assessment-2026-03-25.md and 2026-03-26.md. 8 parallel attack vectors tested: IAM, network, K8s, auth, supply chain, data protection, monitoring, and active pentesting.
All findings addressed21 Controls × 10 Frameworks
governance-bridge skill provides machine-readable mappings from every deployed control to SOC 2, HIPAA, GDPR, NIST 800-53, ISO 27001, FedRAMP, PCI DSS, CCPA, EU AI Act, and OWASP.
10/16 frameworks mapped — 6 more neededAnthropic Commercial License Analysis
First-hand analysis of Anthropic's commercial terms confirms: no training on customer data under commercial license, Data Processing Addendum (DPA) available, Zero Data Retention (ZDR) option confirmed for enterprise plans, SOC 2 Type II certification available via Anthropic Trust Center.
What We're Missing
A Note on Methodology
The Synthetic CISO construct was initially used to validate our existing work — which is confirmation bias formalized as methodology. We recognized this flaw and ran three independent adversarial agents against the CISO's own findings. This section reports the results of that adversarial challenge, not the original self-congratulatory assessment.
CISO Blockers (4)
These items will stop a CISO approval process. They must be completed before or during pilot.
DPIA / Privacy Impact Assessment
Required by: GDPR Art. 35, ISO 42001
Required before any AI tool processes personal data. A DPIA evaluates the necessity and proportionality of AI processing, assesses risks to data subjects, and documents safeguards. Without it, processing is unlawful under GDPR.
Est. effort: 6-10 hours
Vendor Security Questionnaire
Required by: SOC 2 CC9.2, enterprise vendor management
SIG Lite or CAIQ for Anthropic. Every enterprise vendor management program requires this for Tier 1 vendors (vendors that process, store, or have access to confidential data). Anthropic's Trust Center is a start, not a substitute.
Est. effort: 8-16 hours
Formal Risk Acceptance Sign-Off
Required by: ISO 27001 A.8.3, NIST 800-53 PM-9
A named executive must sign acceptance of residual risks (the 4 irreducible risks, the 3 structural/accepted risks, and the Bedrock/Vertex migration timeline). Without this, no defensible position if an incident occurs.
Est. effort: 2-4 hours
Business Continuity Plan
Required by: SOC 2 A1.2, ISO 22301
What happens if Anthropic's API is unavailable for 48 hours? No fallback plan documented. No degraded-mode procedures. No communication templates. Every enterprise continuity program requires this for critical-path tools.
Est. effort: 8-16 hours
Missing Frameworks (6)
Regulatory frameworks that are absent from the current mapping. The first three are particularly significant.
ISO/IEC 42001:2023
THE AI management system standard. Published Dec 2023. Its absence is disqualifying for AI governance maturity claims. Defines requirements for establishing, implementing, maintaining, and improving an AI management system.
EU AI Act (Deep Mapping)
Currently name-dropped in framework list. No risk classification performed, no provider/deployer analysis, no Art. 4 AI literacy compliance analysis. Art. 4 AI literacy requirement is ALREADY IN EFFECT since Feb 2025.
DORA (Digital Operational Resilience Act)
Mandatory for EU financial entities since Jan 2025. Cannot sell to EU FinServ without this. Covers ICT risk management, incident reporting, operational resilience testing, and third-party risk management.
SR 11-7 / OCC 2011-12
US banking model risk management guidance. The first framework a bank CISO reaches for when evaluating AI tools. Covers model development, implementation, and use with emphasis on validation and governance.
ISO/IEC 23894
AI-specific risk management. Complements 42001 by providing guidance on managing risks arising from the development and use of AI systems. Aligns with ISO 31000 risk management principles.
NIST AI 600-1
GenAI-specific risk profile. Covers confabulation, data privacy, environmental impact, information integrity, IP, obscenity, and value alignment. Published July 2024. Extends NIST AI RMF for generative AI.
Missing Risk Categories (8)
Risk categories that were not addressed in the original 54-risk analysis. These are business and legal risks, not technical risks.
IP Ownership of AI-Generated Code
AI-generated code is likely not copyrightable under current US law. Competitors could use identical patterns. No legal analysis of IP implications exists in the current evidence portfolio.
Liability Chain for AI-Generated Defects
Who pays when AI code causes an outage? What are the insurance implications? The liability chain from Anthropic to platform to developer to end user is undocumented.
AI Hallucination in Compliance Evidence
The CDD system generates evidence via AI. If AI fabricates test results, coverage numbers, or security findings, the compliance record is fraudulent. The evidence integrity risk (#52) partially addresses this but not the fabrication vector.
Shadow AI Outside the Platform
A developer opens claude.ai in a browser. None of the 23 hooks apply. This is an organizational problem, not a technical one. No DLP policy for browser-based AI usage exists.
Regulatory Examination Readiness
When the OCC examiner says "show me your AI governance," this HTML presentation is not the answer. A formal risk assessment with defined methodology (FAIR, ISO 27005, or equivalent) is required.
Cross-Contamination Between Clients
Information barriers (Chinese walls) in financial services. If the platform serves multiple clients, shared model context is a regulatory violation. Multi-tenant isolation (#18) covers infrastructure but not model context.
Legal Discovery / Litigation Hold
AI conversation history is discoverable in litigation. No retention policy, no litigation hold procedure, no e-discovery integration documented. This is a legal risk, not a technical one.
Board Reporting Framework
How is AI risk reported to the board? What metrics? What thresholds trigger escalation? No board-level reporting framework for AI risk exists.
Missing Evidence (5)
Evidence artifacts that a CISO would expect to see but do not yet exist.
Third-Party Penetration Test
Self-assessment by AI agents is not independent validation. The same-principal trust paradox (Risk #51) applies to the assessment itself. An independent third-party pen test is required for credible security validation.
Visual Data Flow Diagrams
Code references are not DFDs. GDPR Art. 30 requires data flow documentation showing where personal data moves. No visual data flow diagrams exist for the AI processing pipeline.
Developer Training Completion Records
SOC 2 CC1.4 requires evidence of security awareness training. CLAUDE.md and rules files are training content, but no signed completion records, no assessment scores, no attendance tracking exists.
AI Model Card / System Documentation
EU AI Act Arts. 11-13 require technical documentation describing the AI system, its capabilities, limitations, and intended use. No model card or system documentation exists.
Formal Risk Assessment Report
An HTML presentation is not a formal risk assessment with defined methodology (FAIR/ISO 27005). A structured report with risk scoring methodology, assessment criteria, and sign-off chain is required for regulatory examination.
Missing Processes (4)
Organizational governance processes that do not yet exist.
AI Ethics Committee / Governance Board
ISO 42001 and EU AI Act require an AI governance oversight body. None exists. Responsible AI decisions are currently made ad-hoc by individual developers and AI agents.
AI-Specific Incident Response Playbook
deviation-rules.md is an agent protocol, not a runbook with escalation paths, SLAs, communication templates, and post-incident review procedures. A real IR playbook includes named roles and contact trees.
Model Validation / Back-Testing Program
427 tests validate hooks, not model outputs. There is no program to validate that the AI model produces correct, safe, and unbiased code. No baseline accuracy metrics, no regression testing against known-good outputs.
Employee AI Training Program
Training content exists (CLAUDE.md, rules files, deviation protocols). But there is no curriculum, no assessment, no completion tracking, no periodic recertification. EU AI Act Art. 4 requires AI literacy training.
Gap Summary
| Category | Gaps | CISO Blockers | Est. Effort |
|---|---|---|---|
| Documentation | 14 | 4 | 60-100 hrs |
| Frameworks | 6 | 0 | 40-60 hrs |
| Risk Categories | 8 | 0 | Policy work |
| Evidence | 5 | 1 (third-party pen test) | 30-50 hrs |
| Processes | 4 | 1 (ethics committee) | 20-40 hrs |
| Total | 37 | 6 | 150-250 hrs |
90-Day Governance Roadmap
A concrete plan to close every identified gap. The technical foundation is already deployed. This roadmap addresses the governance, documentation, and process gaps identified in the adversarial review.
Pilot Approval (Week 1-2) 5 USERS
Complete DPIA (CISO blocker #1). Executive risk acceptance sign-off (CISO blocker #3). Begin 5-user pilot with full technical governance active. All 23 hooks, 427 tests, fail-closed gates operational. Daily evidence review by security team.
Foundation (Week 3-4) GOVERNANCE
Complete vendor security questionnaire for Anthropic (CISO blocker #2). BCP for AI tool unavailability (CISO blocker #4). AI-specific incident response playbook with named roles and SLAs. Developer training program launch with completion tracking.
Framework Expansion (Month 2) FRAMEWORKS
ISO 42001 gap analysis and mapping. EU AI Act deep compliance mapping (Art. 4 AI literacy NOW). DORA and SR 11-7 mapping for sector-specific sales enablement. AI Ethics Committee charter and first meeting. Formal risk assessment report (FAIR/ISO 27005).
Maturity (Month 3) 10-20 USERS
Model validation program. Annual governance calendar. Board reporting framework with defined metrics and thresholds. Third-party penetration test engagement. Expand pilot to 10-20 users based on accumulated evidence from months 1-2.
Ongoing STEADY STATE
Monthly evidence reviews. Quarterly framework re-assessment. Annual third-party audit. Continuous CDD evidence collection. Model behavior drift monitoring. Anthropic commercial terms monitoring. Periodic AI literacy recertification.
Phased Rollout Timeline
Controlled expansion with evidence review at every gate. Each phase requires explicit CISO approval to proceed.
Week 1-2: Controlled Pilot 5 USERS
Full governance enforcement active. All 23 hooks enabled. Daily evidence review by security team. DPIA and risk acceptance completed. Incident response procedures tested against real scenarios.
Week 3-4: Evidence Review & Foundation 5 USERS
Two weeks of accumulated evidence reviewed by CISO team. Controls adjusted based on real-world observations. False positive rates measured and tuned. Vendor questionnaire and BCP completed. Training program launched.
Month 2: Framework Expansion 10 USERS
Expand to 10 users with CISO approval. Framework mapping expanded to 16 frameworks. Ethics committee operational. Formal risk assessment report delivered. Monthly compliance review cadence established.
Month 3: Maturity & Scale 10-20 USERS
Third-party pen test results available. Board reporting framework active. Model validation program running. Expand to 10-20 users based on evidence portfolio. Integration with existing SIEM/SOC workflows.
Ongoing: Continuous Governance STEADY STATE
Quarterly security audits with red team exercises. Annual regulatory framework update. Continuous CDD evidence collection. Model behavior drift monitoring. Annual third-party audit.
Ready for Pilot — 90-Day Roadmap to Full Governance
Most AI adoption requests come with a pitch deck and a promise. This one comes with:
Verdict: Ready for Pilot — 90-Day Roadmap to Full Governance
Most AI adoption requests come with a pitch deck and a promise. This one comes with 54 risks analyzed, 41 deployed controls, 427 passing tests — AND an honest accounting of 37 governance gaps we identified by stress-testing our own claims.
We are not asking you to skip your process. We are showing you that we have already started it, that we know what remains, and that we have a concrete plan to close every gap within 90 days.
The technical foundation is the strongest you have seen. The governance roadmap is yours to review and adjust.
| Document Information | |
|---|---|
| Document | CISO Readiness Assessment |
| Subject | Claude Code (Anthropic) — AI Coding Tool Adoption |
| Methodology | 4-phase: Standard Risk + Adversarial Review + Synthetic CISO + Adversarial Challenge |
| Companion | Risk Analysis & Remediation Report |
| Classification | Internal — Security Leadership |
| Date | April 2026 |