Making Long-context LLM Inference 10x Faster and 10x Cheaper Through Knowledge Sharing

Making Long-context LLM Inference 10x Faster and 10x Cheaper Through Knowledge Sharing

CNCF [Cloud Native Computing Foundation] via YouTube Direct link

Making Long-context LLM Inference 10x faster & 10x cheap... - Junchen Jiang, Yihua Cheng, & Zhou Sun

1 of 1

1 of 1

Making Long-context LLM Inference 10x faster & 10x cheap... - Junchen Jiang, Yihua Cheng, & Zhou Sun

Class Central Classrooms beta

YouTube videos curated by Class Central.

Classroom Contents

Making Long-context LLM Inference 10x Faster and 10x Cheaper Through Knowledge Sharing

Automatically move to the next video in the Classroom when playback concludes

  1. 1 Making Long-context LLM Inference 10x faster & 10x cheap... - Junchen Jiang, Yihua Cheng, & Zhou Sun

Never Stop Learning.

Get personalized course recommendations, track subjects and courses with reminders, and more.

Someone learning on their laptop while sitting on the floor.