Aligning LLM-Assisted Evaluation with Human Preferences
Association for Computing Machinery (ACM) via YouTube
Learn Generative AI, Prompt Engineering, and LLMs for Free
Most AI Pilots Fail to Scale. MIT Sloan Teaches You Why — and How to Fix It
Overview
Google, IBM & Meta Certificates — All 10,000+ Courses at 40% Off
One annual plan covers every course and certificate on Coursera. 40% off for a limited time.
Get Full Access
Learn about the critical challenges and methodologies of validating Large Language Model (LLM) outputs in this 17-minute conference talk from the 37th Annual ACM Symposium on User Interface Software and Technology (UIST 2024). Explore the complex relationship between LLM-assisted evaluation methods and human preferences, examining how these automated validation systems align with human judgment. Delve into key questions about the reliability and accuracy of using LLMs to evaluate other LLMs' outputs, while gaining insights into the latest research findings presented at this prestigious ACM symposium in Pittsburgh.
Syllabus
Who Validates the Validators? Aligning LLM-Assisted Evaluation of LLM Outputs with Human Preferences
Taught by
ACM SIGCHI