Introduction
– Introduction to LLM and AI.
– Terminologies and architecture.
– Transformers, Attention & their security implications(hallucinations, jailbreaks etc).
– Agents, multi-agents and multi-modal models.
Elements of AI Security (1 lab)
– Understanding AI vulnerabilities with case studies on AI security breaches.
– OWASP LLM Top 10 and MITRE mapping of attacks on AI supply chain.
– Threat modeling of AI Applications.
Adversarial LLM Attacks and Defenses (6 labs)
– Direct and indirect prompt injection attacks and their subtypes.
– Advanced prompt injections through obfuscation and cross-model injections.
– Breaking system prompts and their trust criteria.
– Indirect prompt injections through external input sources.
Responsible AI & Jailbreaking (6 labs)
– Jailbreaking public LLMs covering adversarial AI, offensive security, and CBRN use-cases.
– Responsible AI frameworks and benchmarks.
– Model alignment, system prompt optimization, and defense.
Building Enterprise-grade LLM Defenses (2 labs)
– Deploying LLM security scanner, adding custom rules, prompt block-lists, and guardrails.
– Writing custom detection logic, trustworthiness checks, and filters.
– Building security log monitoring and alerting for models using open-source tools.
– LLM security benchmarking and continuous reporting.
Red & Blue Teaming of Enterprise AI applications (4 labs)
– Business control flow testing for risky responses & misaligned behavior of applications.
– Using Colab notebooks for automation of API calls and reporting
– Vector database and model-weight tracing for root-cause investigation.
– Rainbow teaming through a 3-way LLM implementation: target, attacker, and judge with self-improving attack prompts.
Attacking & Defending Agentic Systems (5 labs)
– Attacking LLM agents for task manipulation, risky behavior and PII disclosure in RAG.
– Injection attacks on AI agents for code and command execution.
– Compromising backend infrastructure by abusing over-permissioning and tool usage in agentic systems.
– Multi-agent attacks causing privilege too calls, goal manipulation & chained escalations.
Building AI SecOps Process
– Summarizing the learnings into a SecOps workflow.
– Monitoring trustworthiness, safety and security of enterprise AI applications.
– Implementing NIST AI Risk Management Framework (RMF) for security monitoring.