Introduction to Multimodal Large Language Models II - Day 10 Afternoon
Center for Language & Speech Processing(CLSP), JHU via YouTube
Launch Your Cybersecurity Career in 6 Months
PowerBI Data Analyst - Create visualizations and dashboards from scratch
Overview
Coursera Flash Sale
40% Off Coursera Plus for 3 Months!
Grab it
Explore advanced concepts in multimodal large language models through this 81-minute tutorial presentation delivered by Alicia Lozano-Diez from Universidad Autónoma de Madrid and Ramani Duraiswami from University of Maryland at JSALT 2025. Delve into the second part of a comprehensive introduction to multimodal LLMs, building upon foundational concepts covered in Part I. Access accompanying practical materials including interactive Jupyter notebooks and GitHub repositories that provide hands-on experience with multimodal AI systems. Learn how these models integrate and process multiple types of data including text, images, and other modalities to create more sophisticated AI applications. Gain insights into current research developments and practical implementations in the rapidly evolving field of multimodal artificial intelligence from leading experts in language and speech processing.
Syllabus
[slides] Day 10 afternoon - JSALT 2025 - Introduction to Multimodal Large Language Models II.
Taught by
Center for Language & Speech Processing(CLSP), JHU