MM-SafetyBench: A Benchmark for Safety Evaluation of Multimodal Large Language Models
University of Central Florida via YouTube
Learn Python with Generative AI - Self Paced Online
AI Engineer - Learn how to integrate AI into software applications
Overview
Google, IBM & Meta Certificates — All 10,000+ Courses at 40% Off
One annual plan covers every course and certificate on Coursera. 40% off for a limited time.
Get Full Access
This research paper from the University of Central Florida introduces MM-SafetyBench, a comprehensive benchmark designed specifically for evaluating the safety aspects of Multimodal Large Language Models (MLLMs). Explore how this innovative framework addresses the critical need for standardized safety testing in systems that process both text and visual inputs, examining potential vulnerabilities and harmful outputs across various risk categories. Learn about the methodology, evaluation metrics, and findings that help identify safety gaps in current multimodal AI systems, providing valuable insights for researchers and developers working to create more responsible and trustworthy multimodal technologies.
Syllabus
Paper 5: MM-SafetyBench: A Benchmark for Safety Evaluation of Multimodal Large Language Models
Taught by
UCF CRCV