Develop, Build and Deploy Lightweight AI Services with WebAssembly
CNCF [Cloud Native Computing Foundation] via YouTube
Build with Azure OpenAI, Copilot Studio & Agentic Frameworks — Microsoft Certified
AI Engineer - Learn how to integrate AI into software applications
Overview
Google, IBM & Meta Certificates — All 10,000+ Courses at 40% Off
One annual plan covers every course and certificate on Coursera. 40% off for a limited time.
Get Full Access
Explore the development, building, and deployment of lightweight AI services using WebAssembly in this comprehensive tutorial. Learn how to create efficient PyTorch and TensorFlow inference functions, extend large language models like ChatGPT using Rust, and deploy AI inference functions with Kubernetes. Discover the advantages of WebAssembly as a powerful alternative runtime for AI inference workloads, offering lightweight, fast, and cloud-safe solutions. Gain hands-on experience in creating Wasm-based AI services that address the challenges of resource-intensive inference demands, particularly in the context of large language models. Understand the architecture and fundamentals of WebAssembly-based AI services, and explore how they compare to traditional Python-based approaches in terms of size, speed, and resource consumption.
Syllabus
Develop, Build & Deploy Lightweight AI Services with WebAssembly - Kinfey Lo, Vivian Hu, Jianbai Ye
Taught by
CNCF [Cloud Native Computing Foundation]