Class Central is learner-supported. When you buy through links on our site, we may earn an affiliate commission.

MIT OpenCourseWare

Tail Bounds - Lecture 8

MIT OpenCourseWare via YouTube

Overview

Coursera Flash Sale
40% Off Coursera Plus for 3 Months!
Grab it
Explore tail bounds in this 81-minute lecture from MIT's Principles of Discrete Applied Mathematics course, where you'll learn to bound the probability that a random variable deviates significantly from its mean. Master Markov's bound and Chebyshev's bound as fundamental tools for probability analysis, then apply Chebyshev's bound to prove the weak law of large numbers, a cornerstone theorem in probability theory that describes the convergence of sample averages to expected values.

Syllabus

Lecture 8: Tail Bounds

Taught by

MIT OpenCourseWare

Reviews

Start your review of Tail Bounds - Lecture 8

Never Stop Learning.

Get personalized course recommendations, track subjects and courses with reminders, and more.

Someone learning on their laptop while sitting on the floor.