Overview
Coursera Flash Sale
40% Off Coursera Plus for 3 Months!
Grab it
Discover how to dramatically improve the performance and cost-effectiveness of generative AI applications through semantic caching techniques in this AWS re:Invent 2025 conference session. Learn to reduce latencies from single-digit seconds to single-digit milliseconds using vector search capabilities in Amazon ElastiCache for Valkey, while simultaneously cutting foundation model costs for production workloads. Explore the implementation of semantic caching within agentic AI architectures, including RAG-powered assistants and autonomous agents that require frequent foundation model calls for complex workflow orchestration. Gain practical insights into building performant, cost-effective production-scale agentic AI systems that leverage multi-agent orchestration while optimizing both response times and operational expenses through strategic caching approaches.
Syllabus
AWS re:Invent 2025 - Optimize gen AI apps with semantic caching in Amazon ElastiCache (DAT451)
Taught by
AWS Events