Low Power Radar Sensors and TinyML for Embedded Gesture Recognition and Non-Contact Vital Sign Monitoring
EDGE AI FOUNDATION via YouTube
Earn a Michigan Engineering AI Certificate — Stay Ahead of the AI Revolution
The Most Addictive Python and SQL Courses
Overview
Google, IBM & Meta Certificates — All 10,000+ Courses at 40% Off
One annual plan covers every course and certificate on Coursera. 40% off for a limited time.
Get Full Access
Watch a technical conference talk from the tinyML Summit 2023 exploring innovative applications of low-power radar sensors combined with TinyML for gesture recognition and vital sign monitoring. Discover how ETH Zurich researchers developed an efficient embedded system using Acconeer's short-range Doppler-radar technology. Learn about their novel architecture combining 2D CNN and TCN neural networks that achieves 92% accuracy on 11 hand gestures while maintaining a compact 91kB memory footprint. Explore the implementation on RISC-V and ARM Cortex-M processors, making it suitable for battery-powered wearable devices. Gain insights into recent developments using Transformer neural networks and their application in an earbud form factor. Additionally, examine how the same radar sensor technology can be utilized for non-contact vital sign monitoring, including respiration rate and heart rate measurements, through signal processing and TinyML techniques that run on microcontrollers with minimal memory requirements.
Syllabus
tinyML Summit 2023: Low Power Radar Sensors and TinyML for Embedded Gesture Recognition and...
Taught by
EDGE AI FOUNDATION