Aligning LLM-Assisted Evaluation with Human Preferences
Association for Computing Machinery (ACM) via YouTube
Free AI-powered learning to build in-demand skills
Start speaking a new language. It’s just 3 weeks away.
Overview
Coursera Flash Sale
40% Off Coursera Plus for 3 Months!
Grab it
Learn about the critical challenges and methodologies of validating Large Language Model (LLM) outputs in this 17-minute conference talk from the 37th Annual ACM Symposium on User Interface Software and Technology (UIST 2024). Explore the complex relationship between LLM-assisted evaluation methods and human preferences, examining how these automated validation systems align with human judgment. Delve into key questions about the reliability and accuracy of using LLMs to evaluate other LLMs' outputs, while gaining insights into the latest research findings presented at this prestigious ACM symposium in Pittsburgh.
Syllabus
Who Validates the Validators? Aligning LLM-Assisted Evaluation of LLM Outputs with Human Preferences
Taught by
ACM SIGCHI