Random Tessellation Forests: Overcoming the Curse of Dimensionality
Hausdorff Center for Mathematics via YouTube
Google Data Analytics, IBM AI & Meta Marketing — All in One Subscription
Finance Certifications Goldman Sachs & Amazon Teams Trust
Overview
Google, IBM & Meta Certificates — All 10,000+ Courses at 40% Off
One annual plan covers every course and certificate on Coursera. 40% off for a limited time.
Get Full Access
Explore the advanced topic of random tessellation forests in this 58-minute lecture by Eliza O'Reilly from the Hausdorff Center for Mathematics. Delve into the limitations of traditional random forests using axis-aligned partitions and discover how oblique splits can improve performance by capturing feature dependencies. Examine the class of random tessellations forests generated by the stable under iteration (STIT) process in stochastic geometry, and learn how they achieve minimax optimal convergence rates for Lipschitz and C2 functions. Investigate the connection between stationary random tessellations and statistical learning theory, focusing on strategies to overcome the curse of dimensionality in high-dimensional feature spaces through optimal directional distribution choices for random tessellation forest estimators.
Syllabus
Eliza O’Reilly: Random tessellation forests: overcoming the curse of dimensionality
Taught by
Hausdorff Center for Mathematics