Class Central is learner-supported. When you buy through links on our site, we may earn an affiliate commission.

YouTube

Getting an LLM to Hack Itself - On AI, Moral Dilemmas, and Security

OWASP Foundation via YouTube

Overview

Coursera Flash Sale
40% Off Coursera Plus for 3 Months!
Grab it
Explore the intersection of artificial intelligence security and ethical considerations in this 45-minute conference talk that examines how large language models can be manipulated to compromise their own security systems. Delve into the fascinating world of AI vulnerabilities as the speaker demonstrates techniques for getting LLMs to bypass their own safety mechanisms and security protocols. Learn about the moral complexities that arise when AI systems can be turned against themselves, and understand the broader implications for AI security in real-world applications. Discover practical examples of LLM exploitation techniques while examining the ethical dilemmas faced by security researchers working with AI systems. Gain insights into the current state of AI security, the challenges of protecting large language models from self-exploitation, and the ongoing debate about responsible disclosure in the AI security community.

Syllabus

Getting an LLM to Hack Itself: On AI, Moral Dilemmas, and Security

Taught by

OWASP Foundation

Reviews

Start your review of Getting an LLM to Hack Itself - On AI, Moral Dilemmas, and Security

Never Stop Learning.

Get personalized course recommendations, track subjects and courses with reminders, and more.

Someone learning on their laptop while sitting on the floor.