Gain a Splash of New Skills - Coursera+ Annual Nearly 45% Off
Power BI Fundamentals - Create visualizations and dashboards from scratch
Overview
Coursera Flash Sale
40% Off Coursera Plus for 3 Months!
Grab it
Explore the intersection of artificial intelligence security and ethical considerations in this 45-minute conference talk that examines how large language models can be manipulated to compromise their own security systems. Delve into the fascinating world of AI vulnerabilities as the speaker demonstrates techniques for getting LLMs to bypass their own safety mechanisms and security protocols. Learn about the moral complexities that arise when AI systems can be turned against themselves, and understand the broader implications for AI security in real-world applications. Discover practical examples of LLM exploitation techniques while examining the ethical dilemmas faced by security researchers working with AI systems. Gain insights into the current state of AI security, the challenges of protecting large language models from self-exploitation, and the ongoing debate about responsible disclosure in the AI security community.
Syllabus
Getting an LLM to Hack Itself: On AI, Moral Dilemmas, and Security
Taught by
OWASP Foundation