- Learn the fundamental concepts of evaluating generative AI applications.
By the end of this module, you're able to:
- Apply best practices for choosing evaluation data
- Explain when to use real-world, synthetic, and adversarial data
- Describe the scope of built-in evaluators
- Choose evaluators that fit your use case
- Interpret evaluation results and use them to guide mitigations
- Learn how to run evaluations and generate synthetic datasets with the Azure AI Evaluation SDK.
By the end of this module, you're able to:
- Assess a generative AI app response using performance and quality metrics
- Assess a generative AI app response using risk and safety metrics
- Run an evaluation and track the results in a Microsoft Foundry project
- Create a custom evaluator with Prompty
- Send queries to an endpoint and run evaluators on the resulting query and response
- Generate a synthetic dataset using conversation starters
Stuck in Tutorial Hell? Learn Backend Dev the Right Way
Power BI Fundamentals - Create visualizations and dashboards from scratch
Overview
Google, IBM & Meta Certificates — All 10,000+ Courses at 40% Off
One annual plan covers every course and certificate on Coursera. 40% off for a limited time.
Get Full Access
Syllabus
- Evaluating generative AI applications
- Introduction
- Evaluate generative AI
- The role of data in evaluations
- Choose and utilize metrics
- Custom prompt-based evaluators
- Interpret evaluation results
- Common pitfalls in result interpretation
- Module assessment
- Summary
- Run evaluations and generate synthetic datasets
- Introduction
- Prepare
- Exercise - Performance and quality metrics
- Exercise - Risk and safety metrics
- Exercise - Track evaluation results in a Microsoft Foundry project
- Exercise - Custom evaluator with Prompty
- Exercise - Evaluate an endpoint
- Exercise - Generate a synthetic dataset
- Module assessment
- Summary