AI Red Teaming - Why and How to Jailbreak LLM Agents
MLOps World: Machine Learning in Production via YouTube
-
22
-
- Write review
Power BI Fundamentals - Create visualizations and dashboards from scratch
Get 35% Off CFI Certifications - Code CFI35
Overview
Coursera Spring Sale
40% Off Coursera Plus Annual!
Grab it
Explore the critical security vulnerabilities of AI agents through this 11-minute conference talk that demonstrates how adaptive, multi-turn attacks can compromise LLM systems and reveals essential defense strategies. Learn why traditional static testing methods are insufficient against evolving threats as Alex Combessie from Giskard explains how attackers exploit conversational consistency and short-term context to manipulate AI models. Discover the foundations of AI Red Teaming and understand how malicious actors leverage LLMs' reliance on context and conversational flow to execute sophisticated jailbreak attacks. Gain practical insights into implementing continuous, automated red teaming approaches combined with human-in-the-loop monitoring to identify and neutralize emerging security risks before they impact production systems. Master strategies for integrating robust oversight mechanisms that ensure comprehensive AI security in an era where AI agents are becoming increasingly capable and consequently more vulnerable to sophisticated attack vectors.
Syllabus
AI Red Teaming — Why & How to Jailbreak LLM Agents | Alex Combessie, Giskard l The Next Wave of AI
Taught by
MLOps World: Machine Learning in Production