Class Central is learner-supported. When you buy through links on our site, we may earn an affiliate commission.

YouTube

Judging LLMs - Evaluating and Improving AI Engineering Products

AI Engineer via YouTube

Overview

Coursera Flash Sale
40% Off Coursera Plus for 3 Months!
Grab it
Learn how to effectively evaluate and improve Large Language Model (LLM) applications in this conference talk from the AI Engineer World's Fair. Discover practical strategies for iterating on AI products, implementing robust evaluation frameworks, and overcoming the common challenge of building LLM-powered applications without proper assessment mechanisms. Explore real-world approaches to judging LLM performance, understanding evaluation metrics, and establishing feedback loops that enable continuous improvement of AI systems. Gain insights from an AI evangelist with extensive full-stack engineering experience on how to move beyond basic LLM implementation to create measurable, improvable AI products that can be systematically enhanced over time.

Syllabus

Judging LLMs: Alex Volkov

Taught by

AI Engineer

Reviews

Start your review of Judging LLMs - Evaluating and Improving AI Engineering Products

Never Stop Learning.

Get personalized course recommendations, track subjects and courses with reminders, and more.

Someone learning on their laptop while sitting on the floor.