Overview
Coursera Flash Sale
40% Off Coursera Plus for 3 Months!
Grab it
Learn to build robust domain-specific evaluation systems for large language models through a detailed walkthrough by industry experts who demonstrate systematic approaches to AI product improvement. Discover why many AI products fail due to inadequate evaluation frameworks and master the techniques for creating evaluation systems tailored to your specific problem domain. Explore how proper evaluation systems enable rapid AI improvement, unlock data curation capabilities for fine-tuning, and provide the foundation for successful LLM applications. Gain insights from Hamel Husain, who led the team behind CodeSearchNet (a precursor to GitHub Copilot), and Emil Sedgh, CTO at Rechat and developer of Lucy AI assistant, as they share practical examples and proven methodologies for constructing evaluation frameworks that drive real-world AI success.
Syllabus
How to Construct Domain Specific LLM Evaluation Systems: Hamel Husain and Emil Sedgh
Taught by
AI Engineer