Class Central is learner-supported. When you buy through links on our site, we may earn an affiliate commission.

YouTube

Bypassing AI Security Controls with Prompt Formatting

fwd:cloudsec via YouTube

Overview

Coursera Flash Sale
40% Off Coursera Plus for 3 Months!
Grab it
Learn how to exploit vulnerabilities in AI security systems through a conference talk that demonstrates the prompt formatting technique used to bypass AWS Bedrock Guardrails' Sensitive Information Filter functionality. Discover how cybersecurity expert Nathan Kirk from NR Labs successfully circumvented security controls designed to prevent AI systems from returning sensitive data like names and email addresses by instructing AI models to format responses using programmatic, SQL-like queries. Explore the parallels between this AI security bypass and traditional WAF evasion techniques, understand the implications for AI system security, and examine the mitigation strategies developed to help AWS customers protect against this vulnerability. Gain insights from over a decade of penetration testing experience focused on hardware and web applications, drawing from expertise developed at Mandiant's Offensive Services division and Hilton's security programs.

Syllabus

Bypassing AI Security Controls with Prompt Formatting

Taught by

fwd:cloudsec

Reviews

Start your review of Bypassing AI Security Controls with Prompt Formatting

Never Stop Learning.

Get personalized course recommendations, track subjects and courses with reminders, and more.

Someone learning on their laptop while sitting on the floor.