Overview
Coursera Flash Sale
40% Off Coursera Plus for 3 Months!
Grab it
Learn how to build a production-grade, multi-platform NLP pipeline using Ray and GPU acceleration to process millions of social media posts daily across TikTok, YouTube, and Instagram. Discover the challenges of handling massive, fast-changing content streams with unique data formats, ingestion patterns, and quality constraints, and explore how ZEFR engineered a robust distributed pipeline using Ray to orchestrate scalable embedding generation, GPU-heavy processing, and high-throughput vector search ingestion. Follow the step-by-step architecture walkthrough covering Snowflake to Ray ingestion with consistent batch scheduling, cleaning and preprocessing of multimodal content at scale, distributed embedding generation using Ray Actors to shard GPU inference tasks across clusters, and high-throughput writes to Google Cloud Storage, Qdrant for vector search, and Snowflake for analytics. Understand shard lifecycle management including deleting stale shards, managing multi-platform ingestion, and maintaining healthy storage footprints. Gain practical, real-world guidance for operating Ray in production with deployment patterns, debugging tips, failure recovery, throughput tuning, and cost management strategies for processing large multi-source datasets, running GPU-heavy inference pipelines, and building modern vector-search-backed systems.
Syllabus
Distributed Embeddings at Scale: Processing 10M+ Rows/ Day with Ray, GPUs & Qdrant | Ray Summit 2025
Taught by
Anyscale