Overview
Coursera Flash Sale
40% Off Coursera Plus for 3 Months!
Grab it
Explore a comprehensive conference talk from AWS re:Invent 2024 focused on implementing responsible evaluation practices for generative AI applications. Discover essential methodologies for measuring performance and mitigating risks in applications built with large language models (LLMs), including features like Retrieval Augmented Generation (RAG), agents, and guardrails. Gain valuable insights into open access libraries and AWS services that support evaluation processes, while learning the critical steps of creating an effective evaluation plan. Master the process of defining use cases, conducting risk assessments, selecting appropriate metrics and release criteria, developing evaluation datasets, and interpreting results to implement actionable risk mitigation strategies. Delivered by AWS experts, this 54-minute session provides practical knowledge for ensuring responsible AI development and deployment in cloud computing environments.
Syllabus
AWS re:Invent 2024 - Responsible generative AI: Evaluation best practices and tools (AIM342)
Taught by
AWS Events