Overview
Coursera Flash Sale
40% Off Coursera Plus for 3 Months!
Grab it
Learn about human-seeded evaluations in this 12-minute conference talk that introduces a novel approach to testing AI systems by leveraging human input to create more effective evaluation frameworks. Discover the core principles behind human-seeded evals and see how they can be implemented in practice through a live demonstration using Pydantic Logfire. Explore how this methodology bridges the gap between automated testing and human judgment, providing a more nuanced way to assess AI system performance. Gain insights into practical implementation strategies and understand how human-seeded evaluations can improve the reliability and accuracy of AI model assessments in real-world applications.
Syllabus
Human seeded Evals — Samuel Colvin, Pydantic
Taught by
AI Engineer