Google Data Analytics, IBM AI & Meta Marketing — All in One Subscription
Lead AI-Native Products with Microsoft's Agentic AI Program
Overview
Google, IBM & Meta Certificates — All 10,000+ Courses at 40% Off
One annual plan covers every course and certificate on Coursera. 40% off for a limited time.
Get Full Access
Explore structured learning algorithms in this advanced Natural Language Processing lecture from Carnegie Mellon University. Delve into reinforcement learning, minimum risk training, and the structured perceptron. Examine structured max-margin objectives and simple remedies to exposure bias. Learn about globally normalized models, sampling and beam search techniques, and various structured training approaches including hinge loss, cost-augmented hinge loss, and contrastive learning. Gain insights into teacher forcing, self-training, and evaluation metrics for structured prediction tasks in NLP.
Syllabus
Introduction
Types of prediction
Teacher forcing
Evaluation metrics
Structured prediction
Reminder
Globally normalized models
Sampling and Beam Search
Structured Perceptron
Global Structured Perceptron
Structured Training Pretraining
Hinge Loss
Cost Augmented Hinge Loss
Cost Over Sequences
Structured Hinge Loss
Label Smoothing vs Hinge Loss
Contrastive Learning
Teacher Forced
Selftraining
Taught by
Graham Neubig