Visual Language Models for Edge AI 2.0 - Multi-image Reasoning and In-context Learning
EDGE AI FOUNDATION via YouTube
Launch Your Cybersecurity Career in 6 Months
Master Windows Internals - Kernel Programming, Debugging & Architecture
Overview
Google, IBM & Meta Certificates — All 10,000+ Courses at 40% Off
One annual plan covers every course and certificate on Coursera. 40% off for a limited time.
Get Full Access
Watch a 37-minute conference talk exploring cutting-edge innovations in edge AI technology, where Song Han presents groundbreaking developments in visual language models and model efficiency. Dive into VILA (CVPR'24), a sophisticated visual language model capable of multi-image reasoning and in-context learning, specifically designed for deployment on Jetson Orin Nano. Learn about AWQ (MLSys'24), an advanced 4-bit LLM quantization algorithm that enhances model efficiency, and discover TinyChat, an inference library powering visual language model inference. Understand how these three technologies combine to enable advanced visual reasoning capabilities on edge devices, opening new possibilities for edge AI applications.
Syllabus
GenAI on the Edge Forum - Song Han: Visual Language Models for Edge AI 2.0
Taught by
EDGE AI FOUNDATION