Overview
Coursera Flash Sale
40% Off Coursera Plus for 3 Months!
Grab it
Explore how Generative AI-based Synthetic Raters can serve as a cost-effective alternative to human Subject Matter Experts for evaluating AI-generated content in this 43-minute conference talk. Learn about the three-component structure of Synthetic Raters, consisting of trained Large Language Models, system-level parameters, and identification metadata that mirrors human rating processes. Discover a robust framework for SME-based evaluation that combines both human and synthetic rater results, with extensive testing across various LLMs and system prompts. Examine comparison methodologies between synthetic-to-synthetic and human-to-synthetic evaluations across multiple metrics, revealing the potential for synthetic raters to complement human assessments in domains like medicine, law, and finance. Understand the common evaluation metrics employed, testing methodologies, and research findings that demonstrate how synthetic raters can provide diverse perspectives while enhancing overall assessment quality. Gain insights into practical use cases and innovative strategies for integrating human and synthetic ratings to create more efficient and scalable AI evaluation processes that address the resource-intensive nature of traditional human evaluation methods.
Syllabus
Haystack US 2025 - Doug Rosenoff: Enhancing Generative AI Evaluation with Synthetic Raters
Taught by
OpenSource Connections