Overview
Coursera Flash Sale
40% Off Coursera Plus for 3 Months!
Grab it
Explore how to scale machine learning model inference using Nats JetStream in this 55-minute video from The Machine Learning Engineer. Learn about Nats' lightweight architecture that handles millions of messages per second while maintaining a small footprint. Discover how JetStream creates a persistent layer for data stream processing, enhancing ML applications with powerful streaming capabilities. Understand the benefits of horizontal scaling, priority queue management, fault tolerance, and security features including TLS encryption and multiple authentication mechanisms. Gain insights into latency concepts and their application impact, along with quality of service options provided by Nats JetStream. Compare this technology with alternatives like Kafka to make informed infrastructure decisions for optimizing ML inference systems. For access to the accompanying notebook and code (available to paid subscribers only), contact mlengineerchannel@gmail.com.
Syllabus
MLOps: Scaling ML Inference with Distributed Messaging #machinelearning #datascience
Taught by
The Machine Learning Engineer