CLONE - Customizing LLMs for Efficient Latency-Aware Inference at the Edge

CLONE - Customizing LLMs for Efficient Latency-Aware Inference at the Edge

USENIX via YouTube Direct link

USENIX ATC '25 - CLONE: Customizing LLMs for Efficient Latency-Aware Inference at the Edge

1 of 1

1 of 1

USENIX ATC '25 - CLONE: Customizing LLMs for Efficient Latency-Aware Inference at the Edge

Class Central Classrooms beta

YouTube videos curated by Class Central.

Classroom Contents

CLONE - Customizing LLMs for Efficient Latency-Aware Inference at the Edge

Automatically move to the next video in the Classroom when playback concludes

  1. 1 USENIX ATC '25 - CLONE: Customizing LLMs for Efficient Latency-Aware Inference at the Edge

Never Stop Learning.

Get personalized course recommendations, track subjects and courses with reminders, and more.

Someone learning on their laptop while sitting on the floor.