Class Central is learner-supported. When you buy through links on our site, we may earn an affiliate commission.

YouTube

The Surprising Effectiveness of Membership Inference with Simple N-Gram Coverage

Google TechTalks via YouTube

Overview

Coursera Flash Sale
40% Off Coursera Plus for 3 Months!
Grab it
Learn about a novel black-box membership inference attack method in this Google TechTalk that demonstrates how simple n-gram coverage can effectively detect whether specific text was part of a language model's training data. Discover the N-Gram Coverage Attack technique, which relies solely on text outputs from target models without requiring access to hidden states or probability distributions, making it applicable to API-only models like GPT-4. Explore how this method leverages the observation that models are more likely to memorize and generate text patterns commonly seen in training data by obtaining multiple model generations conditioned on candidate text prefixes and using n-gram overlap metrics to measure similarities with ground truth suffixes. Examine benchmark results showing this approach outperforms other black-box methods and achieves comparable performance to state-of-the-art white-box attacks despite having access only to text outputs. Understand how attack performance scales with computational budget as the number of generated sequences increases, and review investigations into previously unstudied closed OpenAI models across multiple domains. Gain insights into the evolving privacy landscape of language models, including findings that more recent models like GPT-4o show increased robustness to membership inference attacks, suggesting improved privacy protections over time.

Syllabus

The Surprising Effectiveness of Membership Inference with Simple N-Gram Coverage

Taught by

Google TechTalks

Reviews

Start your review of The Surprising Effectiveness of Membership Inference with Simple N-Gram Coverage

Never Stop Learning.

Get personalized course recommendations, track subjects and courses with reminders, and more.

Someone learning on their laptop while sitting on the floor.