Exploration in Sparse RL Environments - Curiosity and Random Network Distillation (RND)
Neural Breakdown with AVB via YouTube
Overview
Coursera Flash Sale
40% Off Coursera Plus for 3 Months!
Grab it
Dive into two powerful self-supervised exploration methods for reinforcement learning - Curiosity and Random Network Distillation (RND) - in this 23-minute educational video from Neural Breakdown with AVB. Learn how these techniques solve exploration challenges in sparse reward environments using the Advantage Actor-Critic (A2C) algorithm. Understand through Python and PyTorch code examples how Curiosity drives agents to explore unpredictable states while RND encourages exploration of novel states. Discover the elegant neural network architectures behind these approaches that avoid reward hacking through streamlined intrinsic rewarding. The video references key research papers on Curiosity, RND, and A2C, providing a comprehensive introduction to these innovative exploration techniques in reinforcement learning.
Syllabus
Exploration in Sparse RL Environments is broken... UNTIL YOU ADD THIS! (Curiosity, RND)
Taught by
Neural Breakdown with AVB