Noctua: AI Security Engineering
Build-First AI Security — From Prototype to Production
A year-long graduate-level program where students forge their own agentic security tools using Claude Code, the Claude Agent SDK, and context engineering — then ship them to production.
A graduate-level AI security engineering program for the Agentic Era, 2026
Course Philosophy
Noctua is built on four foundational pillars that shape how we teach and practice cybersecurity in the age of AI agents:
1. Collaborative Critical Thinking (CCT)
The discipline of structured questioning, evidence-based analysis, and team collaboration applied to AI-augmented security work. This is not just about writing better prompts—it's about building better thinking habits.
The 5 Pillars of CCT:
- Evidence-Based Analysis — Grounding decisions in data and empirical reasoning
- Inclusive Perspective — Soliciting diverse viewpoints and challenging groupthink
- Strategic Connections — Linking disparate ideas and identifying systemic patterns
- Adaptive Innovation — Iterating quickly when faced with ambiguity
- Ethical Governance — Ensuring decisions align with organizational values and societal impact
2. Ethical AI & Responsible AI
Grounded in the AIUC-1 Standard — the first security, safety, and reliability standard for AI agents — and aligned with NIST AI RMF, this pillar covers the full spectrum of AI agent security:
- OWASP Top 10 for Agentic Applications
- MITRE ATLAS (Adversarial Threat Landscape for AI Systems)
- Goal hijacking and prompt injection in multi-agent systems
- AI supply chain security and non-human identity (NHI) governance
The AI security standards ecosystem is young — and that's not a disclaimer, it's a core skill to develop. NIST AI RMF (2023), OWASP LLM Top 10 (2023), OWASP Agentic Top 10 (2025), and AIUC-1 (2024) were all published within the last three years. None has the decades of practitioner refinement behind NIST CSF or the OWASP Web Top 10. All are still being revised as the threat landscape evolves ahead of the standards bodies tracking it.
This course references all of them — not because they're settled consensus, but because security practitioners in 2026 cannot afford to wait for settled consensus. Part of what you'll practice here is how to evaluate and selectively adopt emerging frameworks: what governance body is behind it, what adoption it has, what it maps to, and where it fills genuine gaps. Those are the same source-evaluation skills CCT demands of every piece of evidence you analyze.
🔑 Principle: Citing a standard and critically evaluating a standard are not in conflict. Do both.
For the record: AIUC-1 is an industry consortium standard developed with 100+ enterprise CISOs and mapped to NIST, OWASP, and the EU AI Act. It carries less institutional weight than NIST and less community longevity than OWASP, but it's the only framework in this list designed specifically for agentic AI systems — which makes it the most directly applicable to what you're building. We use it on those terms.
3. Rapid Prototyping → Production Delivery
What was aspirational in 2023 is now operational. Agentic engineering tools (Claude Code, Agent SDK, worktrees, subagents, MCP servers) make it possible to go from concept to working prototype in a single lab session. Students will build real cybersecurity tools, not mockups.
But prototyping is only half the story. This course teaches the full delivery pipeline: rapid prototype → leadership evaluation → production hardening → deployment. When leadership selects a prototype for delivery, students learn to accelerate that prototype into production-ready code — adding observability, security controls, CI/CD, governance gates, and operational runbooks. The goal is not just "build fast" but "build fast, then ship."
This pipeline follows the Think → Spec → Build → Retro cycle as its delivery framework: think critically about the architecture before writing any spec, produce a formal spec before building, build rapidly using Claude Code and agentic workflows, and run a structured retrospective — then repeat at production scale when a prototype is selected for delivery.
4. Agentic Engineering
The emerging discipline of designing, building, orchestrating, and securing AI agent systems. This course applies the Think → Spec → Build → Retro development cycle as its delivery framework, supported by five Claude Code skills — /think, /build-spec, /worktree-setup, /retro, and /harness-assess — and built on the Core Four Pillars — Prompt, Model, Context, and Tools — operationalized through patterns that accelerate both prototyping and production delivery:
- Context Engineering — Managing context windows, system prompts, memory, and retrieval. Students build personal context libraries that compound across projects.
- Tool Design — Defining agent capabilities through MCP servers, structured tool definitions, and the "Pit of Success" principle (make the right thing easy, the wrong thing hard).
- Orchestration Patterns — Coordinating multi-agent workflows using orchestrator patterns, expert swarms, and execution topologies suited to the problem.
- Specs as Source Code — Treating specifications as executable artifacts that drive agent behavior, not as documentation that drifts from implementation.
- 12 Leverage Points — A framework for identifying where small changes in agentic systems produce outsized improvements, from AI developer workflows through context management.
This course runs on Claude Code and the Claude Agent SDK. That's a deliberate choice, not a default. Claude Code offers the most mature agentic engineering environment for security practitioners in 2026: native CLAUDE.md configuration, MCP tool-calling architecture, built-in worktree support for isolated agent execution, and a sub-agent model that maps directly to production security workflows.
A significant portion of what you learn here is not Claude-specific. The Think → Spec → Build → Retro cycle is workflow-agnostic. MCP is an open protocol with growing adoption across OpenAI, Google, and open-source stacks. System prompt governance, JSON output schemas, and RAG architecture are cross-vendor by definition. Unit 1 maps out exactly what transfers and what doesn't — so if your organization runs on Azure OpenAI, Google Vertex, or an open-source stack, you know your translation layer from day one.
The tradeoff: standardizing on one vendor's stack reduces cognitive overhead in labs but limits multi-vendor exposure. Semester 2 introduces Claude Managed Agents and the OpenAI Agents SDK to partially address this — providing direct comparison between two production-grade multi-agent platforms.
What's New in 2026
The cybersecurity and AI landscape has shifted dramatically since 2023:
-
Context Engineering has evolved as a discipline. "Prompt engineering" is increasingly dated; the real work is managing context windows, tool definitions, system prompts, memory architectures, and semantic retrieval.
-
The first reported AI-orchestrated cyber espionage campaign (GTG-1002) was disclosed by Anthropic in November 2025 (detection: September 2025), operating at 80-90% autonomy without human intervention for extended periods. This is no longer theoretical.
-
Mature Security Frameworks now exist for agentic systems: OWASP Top 10 for Agentic Apps (2026), NIST Cyber AI Profile (December 2025), and MITRE ATLAS cataloging 66 adversarial techniques.
-
Multi-Agent Orchestration is production-ready. Claude Managed Agents, OpenAI Agents SDK, and AutoGen/AG2 provide mature platforms for building agent teams at scale.
-
Non-Human Identities (NHIs) outnumber human identities in enterprise systems by ratios typically ranging from 25:1 to over 100:1, creating new governance and security challenges.
-
Research-Grade Security Tools now make operational AI security testing feasible in educational and enterprise settings, enabling hands-on assessment, governance, and autonomy control as part of the curriculum.
Course Structure
Duration: Two 16-week semesters, 3 credit hours each Format: 70% hands-on labs and projects, 30% theory and frameworks Delivery: In-person labs with asynchronous readings and reflections
Semester 1: Foundations — CCT, AI Governance, and Agentic Fundamentals (CSEC 601)
Building the critical thinking and ethical foundation while getting hands-on with agentic tools from Week 1.
| Unit | Weeks | Focus Area | Key Outcomes |
|---|---|---|---|
| Unit 1: CCT Foundations & AI Landscape | 1-4 | Critical thinking frameworks, cognitive biases, the CCT 5 Pillars, modern AI landscape | Apply CCT to real security decisions; understand the evolution from LLMs to agents |
| Unit 2: Agent Tool Architecture | 5-8 | MCP servers, tool design patterns, structured outputs, and RAG-based knowledge retrieval — the infrastructure layer of context engineering | Design and implement MCP servers; build context-aware agent systems |
| Unit 3: AI Security Governance | 9-12 | NIST AI RMF, OWASP Top 10 for Agentic Apps, bias and fairness, explainability, AIUC-1 Standard applied to agentic systems | Conduct risk assessments; build guardrails into agent systems |
| Unit 4: Rapid Prototyping with Agentic Tools | 13-15 | Claude Code, worktrees, subagents, agent teams—building real cybersecurity tools in single lab sessions | Deliver working prototypes of security tools; measure MTTS, MTTP, MTTSol |
| Week 16: Midyear Project Presentations | 16 | Team-based rapid prototyping project showcase | Present and defend a functional agentic security system |
Semester 2: Advanced — Agentic Security Engineering (CSEC 602)
Deep technical work: multi-agent systems, red teaming, adversarial AI, and production deployment patterns.
| Unit | Weeks | Focus Area | Key Outcomes |
|---|---|---|---|
| Unit 5: Multi-Agent Orchestration | 1-4 | Claude Managed Agents and OpenAI Agents SDK—comparing orchestration approaches, designing agent teams for security operations | Build and evaluate multi-agent SOC and threat analysis systems |
| Unit 6: AI Attacker vs. AI Defender | 5-8 | Red teaming AI agents, prompt injection, goal hijacking, tool misuse, adversarial ML, real-world case studies | Conduct adversarial testing; harden agents against known attack patterns |
| Unit 7: Production Security Engineering | 9-12 | AI supply chain security, NHI governance, observability, cost management, deployment patterns | Design secure agent deployments; implement monitoring and audit trails |
| Unit 8: Capstone Projects | 13-16 | Full agentic cybersecurity systems—built, tested, red-teamed, and presented | Deliver production-grade security agent system with documentation and threat assessment |
Lab Environment
The lab stack is centered on Claude Max subscription capabilities, with multi-vendor exposure for comparative learning.
Agentic Development Stack
- Claude Code — Integrated development environment for agentic engineering
- Claude Agent SDK — Building custom multi-agent systems in Python and TypeScript
- Worktrees, Subagents, Agent Teams — Parallel development, task delegation, coordinated workflows
- MCP (Model Context Protocol) Servers — Connecting agents to external tools, databases, APIs, and security platforms
AI Red Teaming & Adversarial Testing
- Garak (NVIDIA) — LLM vulnerability scanner with 37+ probe modules
- PyRIT (Microsoft) — Multi-turn adversarial AI red teaming framework
- Promptfoo — Red teaming and evals with OWASP/NIST/MITRE compliance mapping
- DeepTeam — 40+ vulnerability types including prompt injection, jailbreaks, PII leakage
- MASS — AI deployment security assessment and compliance mapping
AI Guardrails & Governance
- NeMo Guardrails (NVIDIA) — Programmable guardrails for LLM-based systems
- LlamaFirewall (Meta) — Agent security with PromptGuard, alignment checks, CodeShield
- Guardrails AI — Runtime validation framework with community validators
- PeaRL — Policy enforcement and governance for autonomous AI agents
- Cisco MCP Scanner — MCP component security evaluation for agent supply chain
Fairness & Bias Assessment
- IBM AI Fairness 360 — Bias detection and fairness metrics for ML models
- Aequitas (U of Chicago) — Bias and fairness audit toolkit
Agent Orchestration Frameworks
- Claude Managed Agents — Anthropic’s production multi-agent orchestration layer; builds on the Claude Agent SDK with managed infrastructure, tool use, and subagent delegation
- OpenAI Agents SDK — Cross-platform agent development; used in Semester 2 for comparative multi-agent architecture study
- AutoGen/AG2 — Multi-agent conversation patterns and hierarchical teams
Infrastructure & DevSecOps Pipeline
- Docker Desktop — Local containerization from Day 1; every prototype ships as a container
- AWS CLI + ECR/ECS — Container registry and orchestration for cloud-native promotion
- GitHub CLI (gh) — PR workflows, CI/CD triggers, branch management, and security scanning
- Infrastructure as Code — CloudFormation/Terraform for repeatable, auditable deployments
- Ollama + Open-Weight Models — Local model deployment for sensitive security research
- Python 3.11+ — Primary language
- Git / GitHub Actions — Version control, worktrees, automated security gates
Performance Metrics
The five core performance metrics remain valid and are now measurable in real-time using agentic tools:
- MTTS (Mean Time to Strategy) — How quickly a team identifies the core security problem and strategic approach
- MTTP (Mean Time to Plan) — How quickly a validated plan emerges from the strategy
- MTTSol (Mean Time to Solution) — How quickly a working solution prototype exists
- MTTI (Mean Time to Implementation) — How quickly the solution is hardened and deployed
- aMTTR (Mean Time to Auto-Remediation) — How quickly an agent system can detect and remediate security incidents autonomously
Students will track these metrics across lab exercises to quantify how agentic tools and CCT practices accelerate each phase of security engineering.
Repository Structure
The course site is HTML-first. docs/ is the canonical source for all student-facing content and is served via GitHub Pages. Markdown files in semester-1/ and semester-2/ are supplementary and must align to the HTML.
Noctua/
├── README.md # Repo overview
├── CLAUDE.md # HTML-is-canonical rule + authoring conventions
│
├── docs/ # GitHub Pages site — all student-facing content
│ ├── index.html # Course Overview (this page)
│ ├── semester1.html / semester2.html
│ ├── s1-unit[1-4].html # Semester 1 theory pages
│ ├── lab-s1-unit[1-4].html # Semester 1 interactive lab guides
│ ├── s2-unit[5-8].html # Semester 2 theory pages
│ ├── lab-s2-unit[5-8].html # Semester 2 interactive lab guides
│ ├── reading.html / frameworks.html / lab-setup.html / assessment.html
│ ├── labs.css / labs.js # Shared stylesheet and interactive system
│ ├── skills/ # Course skill files (/think, /build-spec, /retro, etc.)
│ ├── resources/ # Reference pages (cheatsheet, command ref, etc.)
│ └── data/ # Downloadable lab data files
│
├── semester-1/ # Supplementary Markdown content (aligns to HTML)
│ ├── SYLLABUS.md
│ └── weeks/unit-[1-4].md
│
├── semester-2/ # Supplementary Markdown content (aligns to HTML)
│ ├── SYLLABUS.md
│ └── weeks/unit-[5-8].md
│
└── resources/ # Supporting materials (case studies, references)
├── FRAMEWORKS.md / READING-LIST.md / LAB-SETUP.md
└── case-studies/
Prerequisites
This is a graduate-level applied course. Students are expected to arrive with working technical skills — the course does not teach Python, Git, or cybersecurity fundamentals from scratch.
Required:
- Academic Standing: Graduate level in Computer Science, Cybersecurity, Information Security, or closely related field — or equivalent professional experience with instructor approval
- Python: Intermediate proficiency — async/await, working with REST APIs, JSON handling, virtual environments, writing and running scripts from the command line. Labs begin in Week 1.
- Cybersecurity: Hands-on background in at least one of: incident response, threat modeling, penetration testing, or security operations. OWASP Top 10 familiarity expected. Unit 3 builds on this — it does not introduce it.
- API Development: Experience building or consuming REST APIs. You will build MCP servers from Week 5; understanding HTTP methods, schemas, and error handling is assumed.
- Containers: Basic Docker proficiency (build, run, docker-compose). Unit 4 requires containerization from day one.
- Version Control: Git proficiency beyond basic commits — branching, pull requests, worktrees. GitHub Actions experience is a plus.
- Command Line: Comfortable in bash/zsh. You will work primarily in the terminal throughout both semesters.
- Software: Claude Max subscription (provided by the program or required as a course fee)
Strongly Recommended:
- LLM API experience: Prior use of any LLM API (Anthropic, OpenAI, etc.) — understanding of tokens, system prompts, and API rate limits reduces friction in early weeks
- Cloud basics: Familiarity with AWS (IAM, S3, Lambda basics) — Unit 7 deploys to AWS AgentCore; no deep AWS expertise required but account setup and basic navigation assumed
- Security tooling: Hands-on experience with SIEM, IDS/IPS, or vulnerability scanners (Nessus, Burp Suite, etc.)
Assessment & Grading
| Component | Weight | Description |
|---|---|---|
| Lab Exercises & Participation | 30% | Hands-on labs, code reviews, in-class activities, and engagement |
| Weekly CCT Reflections & Journals | 10% | Reflective writing on critical thinking and decision-making |
| Semester 1 Midyear Project | 20% | Team-based rapid prototype of an agentic security tool |
| Semester 2 Capstone Project | 30% | Full-scale agentic cybersecurity system with threat assessment and deployment guide |
| Peer Reviews & Red Team Exercises | 10% | Constructive feedback on peers' work; adversarial testing of systems |
Grading Scale: A (90-100), B (80-89), C (70-79), D (60-69), F (below 60)
Late Work Policy: Labs submitted after the deadline receive a 10% penalty per day, up to 3 days. No credit after 3 days without prior arrangement.
Recommended Reading
Critical Thinking & CCT
- "Thinking, Fast and Slow" by Daniel Kahneman — The cognitive science foundation for understanding bias in decision-making
- "The Art of Thinking Clearly" by Rolf Dobelli — 99 cognitive biases explained with practical examples
- "Collaborative Intelligence" by Dawna Markova and Angie McArthur — Building teams that think better together
- Carl Sagan's Baloney Detection Kit (excerpt from The Demon-Haunted World) — Skeptical inquiry methods
- Richard Paul's Critical Thinking Frameworks — Structured approaches to evaluating arguments and evidence
- Seth Jaeger — "The 5 Habits of Mind Framework" (sethjaeger.com) — The framework underlying this course's CCT pillars: Evidence-Based Analysis, Inclusive Perspective, Strategic Connections, Adaptive Innovation, and Ethical Governance.
Agentic Engineering & AI Systems
- "Agentic Engineering Book" by Jaymin West (jayminwest.com/agentic-engineering-book) — Additional reading — original source for many agentic engineering patterns used in this course.
- Anthropic — "Building Agents with the Claude Agent SDK" (anthropic.com/engineering) — Official SDK documentation and patterns
- Anthropic — "Effective Context Engineering for AI Agents" — Managing context windows and semantic retrieval
- "Designing AI Agents" (emerging body of work from multiple researchers) — Agent architecture patterns
- Anthropic — Claude Managed Agents (docs.anthropic.com/en/docs/agents) — Official documentation for Anthropic’s managed multi-agent orchestration platform
- OpenAI Agents SDK (github.com/openai/openai-agents-python) — Open-source SDK for building cross-platform agent systems; used in Semester 2 for comparative architecture study
AI Security & Adversarial Techniques
- OWASP Top 10 for Agentic Applications (2026) — Web application security adapted for AI agents
- NIST AI Risk Management Framework (AI RMF 1.0) — Governance and risk assessment
- NIST Cyber AI Profile (December 2025 draft) — Integrating AI capabilities with cybersecurity
- MITRE ATLAS — Adversarial Threat Landscape for AI Systems (atlas.mitre.org) — 66 documented techniques
- "Adversarial Machine Learning" by Vorobeychik and Kantarcioglu — Academic foundation for attack strategies
- Anthropic AI Safety Research (anthropic.com/research) — Ongoing work on adversarial robustness
Ethics & Responsible AI
- "Weapons of Math Destruction" by Cathy O'Neil — Real-world harms from algorithmic systems
- "AI Ethics" by Mark Coeckelbergh — Philosophical and practical frameworks
- AIUC-1 Standard (https://www.aiuc-1.com/) — The first security, safety, and reliability standard for AI agents, covering six domains: Data & Privacy, Security, Safety, Reliability, Accountability, and Society. Operationalizes NIST AI RMF, ISO 42001, MITRE ATLAS, and OWASP LLM Top 10 into auditable controls with third-party certification.
- "The Ethics of Artificial Intelligence" edited by Bostrom and Yudkowsky — Foundational perspectives
Rapid Prototyping & Agile Methods
- "Sprint" by Jake Knapp, John Zeratsky, and Brendan Brown — Time-boxed design and development
- "The Lean Startup" by Eric Ries — Build-measure-learn feedback loops
How to Use This Repository
For Instructors
- Start with the full syllabi in
semester-1/SYLLABUS.mdandsemester-2/SYLLABUS.md - Review lab guides in the
labs/directories to understand learning objectives and assessment rubrics - Check the Lab Setup Guide to prepare your lab environment
- Distribute weekly readings and labs through your institution's learning management system
For Students
- Read this README and the full course syllabus to understand expectations
- Complete the lab environment setup in Week 1
- Work through weekly readings and labs in sequence
- Maintain a CCT reflection journal as specified in the assessment guidelines
- Collaborate with peers on team projects while maintaining academic integrity
For Security Practitioners
This repository can be adapted for:
- Corporate security team training programs
- Incident response team upskilling
- AI/ML security auditing and testing
- Building internal AI agents for security operations
Communication & Support
- Office Hours: [Schedule per syllabus]
- Discussion Forum: [GitHub Discussions or institution platform]
- Emergency Contact: [Contact information per syllabus]
- Lab Technical Issues: Submit an issue with complete error logs and environment details
License
Course materials and curriculum design © 2023-2026. All Rights Reserved.
Students may use materials for educational purposes only. Commercial use, publication, or distribution requires explicit written permission.
Acknowledgments
Developed in collaboration with cybersecurity and AI research communities, informed by:
- Anthropic's work on agentic AI and prompt injection vulnerabilities
- NIST and OWASP contributions to AI security frameworks
- Jaymin West's foundational work in agentic engineering
- Garry Tan's gstack (github.com/garrytan/gstack) — shipping methodology, role-based review gates, and the "Boil the Lake" principle informing Units 4 and 7
- Feedback from 2023-2025 course cohorts and industry practitioners
Quick Links
Last Updated: March 2026
For questions or feedback, open an issue in this repository.