Become an AI & ML Engineer with Cal Poly EPaCE — IBM-Certified Training
Stuck in Tutorial Hell? Learn Backend Dev the Right Way
Overview
Google, IBM & Meta Certificates — All 10,000+ Courses at 40% Off
One annual plan covers every course and certificate on Coursera. 40% off for a limited time.
Get Full Access
Explore fundamental artificial intelligence concepts through this comprehensive course covering probability theory, classification algorithms, and optimization techniques. Begin with discrete probability fundamentals and progress through Naïve Bayes classification, including parameter estimation and Laplace smoothing for handling sparse data. Master binary classifier evaluation metrics and implement both binary and multi-class perceptron algorithms. Delve into logistic regression for both binary and multi-class problems, understanding the mathematical foundations and practical applications. Conclude with optimization theory and gradient ascent methods essential for machine learning model training. Access supplementary materials and detailed course content through the provided course website at https://atcold.github.io/NYU-AISP24/ to enhance your understanding of these core AI methodologies.
Syllabus
00 – Course introduction
Chapter 1, video 1–3
01 – Course first part recap, Naïve Bayes intro
02 – Discrete probability recap, Naïve Bayes classification
03 – Naïve Bayes parameters estimation and Laplace smoothing
Chapter 2, video 4–6
04 – Binary classifier evaluation, binary Perceptron
05 – Multi-class perceptron, binary and multi-class logistic regression
06 – Optimisation and gradient ascent
Taught by
Alfredo Canziani