Choosing the Right LLM - A Framework for Model Evaluation and Selection
Data Science Festival via YouTube
Overview
Coursera Flash Sale
40% Off Coursera Plus for 3 Months!
Grab it
Explore a comprehensive technical talk from the Data Science Festival Oktoberfest 2024 where Emma Mani from Financial Times presents a practical framework for evaluating and selecting Large Language Models (LLMs) for specific tasks. Learn how to move beyond standard benchmarks like Hugging Face's MTEB and vendor-provided metrics to develop custom evaluation methods, demonstrated through real-world summarization tasks at the Financial Times. Dive into technical aspects including article summarization techniques, vectorization, cosine similarity calculations, LLM parameter optimization, and result distribution analysis. Gain insights into creating replicable testing frameworks that can be adapted for various LLM use cases, helping teams make informed decisions when choosing between multiple AI models. Designed for technical practitioners, this 44-minute presentation offers actionable approaches to navigate the growing landscape of LLM options and select the most suitable model for specific organizational needs.
Syllabus
The Quest for the Best: How to Choose the Right LLM
Taught by
Data Science Festival