Plan du cours

Day 1: Foundations and Core Threats

Module 1: Introduction to OWASP GenAI Security Project (1 hour)

Learning Objectives:

  • Understand the evolution from OWASP Top 10 to GenAI-specific security challenges

  • Explore the OWASP GenAI Security Project ecosystem and resources

  • Identify key differences between traditional application security and AI security

Topics Covered:

  • Overview of OWASP GenAI Security Project mission and scope

  • Introduction to the Threat Defense COMPASS framework

  • Understanding the AI security landscape and regulatory requirements

  • AI attack surfaces vs traditional web application vulnerabilities

Practical Exercise: Setting up the OWASP Threat Defense COMPASS tool and performing initial threat assessment

 

Module 2: OWASP Top 10 for LLMs - Part 1 (2.5 hours)

Learning Objectives:

  • Master the first five critical LLM vulnerabilities

  • Understand attack vectors and exploitation techniques

  • Apply practical mitigation strategies

Topics Covered:

LLM01: Prompt Injection

  • Direct and indirect prompt injection techniques

  • Hidden instruction attacks and cross-prompt contamination

  • Practical examples: Jailbreaking chatbots and bypassing safety measures

  • Defense strategies: Input sanitization, prompt filtering, differential privacy

LLM02: Sensitive Information Disclosure

  • Training data extraction and system prompt leakage

  • Model behavior analysis for sensitive information exposure

  • Privacy implications and regulatory compliance considerations

  • Mitigation: Output filtering, access controls, data anonymization

LLM03: Supply Chain Vulnerabilities

  • Third-party model dependencies and plugin security

  • Compromised training datasets and model poisoning

  • Vendor risk assessment for AI components

  • Secure model deployment and verification practices

Practical Exercise: Hands-on lab demonstrating prompt injection attacks against vulnerable LLM applications and implementing defensive measures

 

Module 3: OWASP Top 10 for LLMs - Part 2 (2 hours)

Topics Covered:

LLM04: Data and Model Poisoning

  • Training data manipulation techniques

  • Model behavior modification through poisoned inputs

  • Backdoor attacks and data integrity verification

  • Prevention: Data validation pipelines, provenance tracking

LLM05: Improper Output Handling

  • Insecure processing of LLM-generated content

  • Code injection through AI-generated outputs

  • Cross-site scripting via AI responses

  • Output validation and sanitization frameworks

Practical Exercise: Simulating data poisoning attacks and implementing robust output validation mechanisms

 

Module 4: Advanced LLM Threats (1.5 hours)

Topics Covered:

LLM06: Excessive Agency

  • Autonomous decision-making risks and boundary violations

  • Agent authority and permission management

  • Unintended system interactions and privilege escalation

  • Implementing guardrails and human oversight controls

LLM07: System Prompt Leakage

  • System instruction exposure vulnerabilities

  • Credential and logic disclosure through prompts

  • Attack techniques for extracting system prompts

  • Securing system instructions and external configuration

Practical Exercise: Designing secure agent architectures with appropriate access controls and monitoring

 

Day 2: Advanced Threats and Implementation

Module 5: Emerging AI Threats (2 hours)

Learning Objectives:

  • Understand cutting-edge AI security threats

  • Implement advanced detection and prevention techniques

  • Design resilient AI systems against sophisticated attacks

Topics Covered:

LLM08: Vector and Embedding Weaknesses

  • RAG system vulnerabilities and vector database security

  • Embedding poisoning and similarity manipulation attacks

  • Adversarial examples in semantic search

  • Securing vector stores and implementing anomaly detection

LLM09: Misinformation and Model Reliability

  • Hallucination detection and mitigation

  • Bias amplification and fairness considerations

  • Fact-checking and source verification mechanisms

  • Content validation and human oversight integration

LLM10: Unbounded Consumption

  • Resource exhaustion and denial-of-service attacks

  • Rate limiting and resource management strategies

  • Cost optimization and budget controls

  • Performance monitoring and alerting systems

Practical Exercise: Building a secure RAG pipeline with vector database protection and hallucination detection

 

Module 6: Agentic AI Security (2 hours)

Learning Objectives:

  • Understand the unique security challenges of autonomous AI agents

  • Apply the OWASP Agentic AI taxonomy to real-world systems

  • Implement security controls for multi-agent environments

Topics Covered:

  • Introduction to Agentic AI and autonomous systems

  • OWASP Agentic AI Threat Taxonomy: Agent Design, Memory, Planning, Tool Use, Deployment

  • Multi-agent system security and coordination risks

  • Tool misuse, memory poisoning, and goal hijacking attacks

  • Securing agent communication and decision-making processes

Practical Exercise: Threat modeling exercise using OWASP Agentic AI taxonomy on a multi-agent customer service system

 

Module 7: OWASP Threat Defense COMPASS Implementation (2 hours)

Learning Objectives:

  • Master the practical application of Threat Defense COMPASS

  • Integrate AI threat assessment into organizational security programs

  • Develop comprehensive AI risk management strategies

Topics Covered:

  • Deep dive into Threat Defense COMPASS methodology

  • OODA Loop integration: Observe, Orient, Decide, Act

  • Mapping threats to MITRE ATT&CK and ATLAS frameworks

  • Building AI Threat Resilience Strategy Dashboards

  • Integration with existing security tools and processes

Practical Exercise: Complete threat assessment using COMPASS for a Microsoft Copilot deployment scenario

 

Module 8: Practical Implementation and Best Practices (2.5 hours)

Learning Objectives:

  • Design secure AI architectures from the ground up

  • Implement monitoring and incident response for AI systems

  • Create governance frameworks for AI security

Topics Covered:

Secure AI Development Lifecycle:

  • Security-by-design principles for AI applications

  • Code review practices for LLM integrations

  • Testing methodologies and vulnerability scanning

  • Deployment security and production hardening

Monitoring and Detection:

  • AI-specific logging and monitoring requirements

  • Anomaly detection for AI systems

  • Incident response procedures for AI security events

  • Forensics and investigation techniques

Governance and Compliance:

  • AI risk management frameworks and policies

  • Regulatory compliance considerations (GDPR, AI Act, etc.)

  • Third-party risk assessment for AI vendors

  • Security awareness training for AI development teams

Practical Exercise: Design a complete security architecture for an enterprise AI chatbot including monitoring, governance, and incident response procedures

 

Module 9: Tools and Technologies (1 hour)

Learning Objectives:

  • Evaluate and implement AI security tools

  • Understand the current AI security solutions landscape

  • Build practical detection and prevention capabilities

Topics Covered:

  • AI security tool ecosystem and vendor landscape

  • Open-source security tools: Garak, PyRIT, Giskard

  • Commercial solutions for AI security and monitoring

  • Integration patterns and deployment strategies

  • Tool selection criteria and evaluation frameworks

Practical Exercise: Hands-on demonstration of AI security testing tools and implementation planning

 

Module 10: Future Trends and Wrap-up (1 hour)

Learning Objectives:

  • Understand emerging threats and future security challenges

  • Develop continuous learning and improvement strategies

  • Create action plans for organizational AI security programs

Topics Covered:

  • Emerging threats: Deepfakes, advanced prompt injection, model inversion

  • Future OWASP GenAI project developments and roadmap

  • Building AI security communities and knowledge sharing

  • Continuous improvement and threat intelligence integration

Action Planning Exercise: Develop a 90-day action plan for implementing OWASP GenAI security practices in participants' organizations

Pré requis

  • General understanding of web application security principles
  • Basic familiarity with AI/ML concepts
  • Experience with security frameworks or risk assessment methodologies preferred

Audience

  • Cybersecurity professionals
  • AI developers
  • System architects
  • Compliance officers
  • Security practitioners
 14 Heures

Nombre de participants


Prix ​​par Participant

Cours à venir

Catégories Similaires