Class Central is learner-supported. When you buy through links on our site, we may earn an affiliate commission.

YouTube

AI Red Teaming - Why and How to Jailbreak LLM Agents

MLOps World: Machine Learning in Production via YouTube

Overview

Coursera Spring Sale
40% Off Coursera Plus Annual!
Grab it
Explore the critical security vulnerabilities of AI agents through this 11-minute conference talk that demonstrates how adaptive, multi-turn attacks can compromise LLM systems and reveals essential defense strategies. Learn why traditional static testing methods are insufficient against evolving threats as Alex Combessie from Giskard explains how attackers exploit conversational consistency and short-term context to manipulate AI models. Discover the foundations of AI Red Teaming and understand how malicious actors leverage LLMs' reliance on context and conversational flow to execute sophisticated jailbreak attacks. Gain practical insights into implementing continuous, automated red teaming approaches combined with human-in-the-loop monitoring to identify and neutralize emerging security risks before they impact production systems. Master strategies for integrating robust oversight mechanisms that ensure comprehensive AI security in an era where AI agents are becoming increasingly capable and consequently more vulnerable to sophisticated attack vectors.

Syllabus

AI Red Teaming — Why & How to Jailbreak LLM Agents | Alex Combessie, Giskard l The Next Wave of AI

Taught by

MLOps World: Machine Learning in Production

Reviews

Start your review of AI Red Teaming - Why and How to Jailbreak LLM Agents

Never Stop Learning.

Get personalized course recommendations, track subjects and courses with reminders, and more.

Someone learning on their laptop while sitting on the floor.