Google AI Professional Certificate - Learn AI Skills That Get You Hired
Get 50% Off Udacity Nanodegrees — Code CC50
Overview
Coursera Flash Sale
40% Off Coursera Plus for 3 Months!
Grab it
Learn about mTuner, a novel fine-tuning system that accelerates parameter-efficient fine-tuning (PEFT) of large language models on multi-GPU servers through innovative memory management techniques. Discover how the Elastic Tensor abstraction enables dynamic tensor management with four key operations - gather, discard, execute, and checkpoint - that provide flexible control over tensor availability, accumulation, and release in memory. Explore the critical importance of memory efficiency during fine-tuning compared to pre-training, where frozen parameters can be cached for performance optimization despite limited hardware memory capacity. Understand how elastic tensors enable optimizations including improved temporal memory utilization, relaxed data dependence, and memory-adaptive runtime tensor accumulation. Examine the performance results showing mTuner achieves throughput improvements of up to 51.2% and 24.8% on PCIe and NVLink servers respectively compared to state-of-the-art training and fine-tuning systems, with testing conducted on LLMs ranging from 7B to 70B parameters. Gain insights from researchers at Tsinghua University on addressing the growing demand for personalized large language models while maintaining computational and storage efficiency through this end-to-end fine-tuning solution.
Syllabus
USENIX ATC '25 - mTuner: Accelerating Parameter-Efficient Fine-Tuning on Multi-GPU Servers with...
Taught by
USENIX