35% Off Finance Skills That Get You Hired - Code CFI35
Free AI-powered learning to build in-demand skills
Overview
Coursera Flash Sale
40% Off Coursera Plus for 3 Months!
Grab it
Learn advanced inference optimization techniques for enterprise AI applications through this 21-minute technical talk that explores Command A's efficient inference pipeline designed to balance speed, cost, and performance. Discover how interleaved sliding window attention enhances both quality and speed while reducing computational overhead. Explore speculative decoding methodologies and their implementation using Medusa for parallel token prediction, including insights from the training process and performance evaluation using Weights & Biases. Examine the trade-offs between synthetic and original data in speculative training, analyze final performance gains and their associated costs, and understand how guided decoding integrates with speculative inference. Master dynamic guided decoding techniques and finite state machine (FSM) integration, culminating in strategies for combining guided decoding with speculative tokens to achieve optimal cost-effective AI solutions for enterprise environments.
Syllabus
0:00 – Introduction to Command R+ Inference Optimization
0:55 – Sparse Attention Architecture & Sliding Window
2:21 – Speculative Decoding Overview
4:32 – Using Medusa for Parallel Token Prediction
6:29 – Evaluation and Training with W&B
7:54 – Synthetic vs. Original Data in Speculative Training
9:00 – Final Gains and Performance Tradeoffs
11:44 – Guided Decoding with Speculative Inference
14:29 – Dynamic Guided Decoding and FSM Integration
19:03 – Combining Guided Decoding with Speculative Tokens
Taught by
Weights & Biases