The Surprising Effectiveness of Membership Inference with Simple N-Gram Coverage
USC Information Sciences Institute via YouTube
The Investment Banker Certification
Build with Azure OpenAI, Copilot Studio & Agentic Frameworks — Microsoft Certified
Overview
Google, IBM & Meta Certificates — All 10,000+ Courses at 40% Off
One annual plan covers every course and certificate on Coursera. 40% off for a limited time.
Get Full Access
This seminar presentation from April 24, 2025, features Skyler Hallinan from USC discussing a novel approach to membership inference attacks on language models. Learn about the N-Gram Coverage Attack, a method that requires only text outputs from target models, enabling attacks on completely black-box systems like GPT-4. Discover how this technique leverages the observation that models tend to memorize and generate text patterns commonly found in their training data. The presentation demonstrates how this approach outperforms other black-box methods and achieves comparable or better results than state-of-the-art white-box attacks despite limited access. Explore findings showing that attack success rates scale with compute budget and that newer models like GPT-4o exhibit increased robustness to membership inference, suggesting improved privacy protections. Hallinan, a Ph.D. student at USC advised by Xiang Ren, focuses on building trustworthy AI systems through data-centric approaches, with experience as a research intern at Apple and Amazon.
Syllabus
The Surprising Effectiveness of Membership Inference with Simple N-Gram Coverage
Taught by
USC Information Sciences Institute