Reinforcement Learning in Non-Stationary Environments
Centre for Networked Intelligence, IISc via YouTube
Earn Your CS Degree, Tuition-Free, 100% Online!
AI, Data Science & Cloud Certificates from Google, IBM & Meta
Overview
Google, IBM & Meta Certificates — All 10,000+ Courses at 40% Off
One annual plan covers every course and certificate on Coursera. 40% off for a limited time.
Get Full Access
Explore the challenges and solutions of reinforcement learning in dynamic environments through this online lecture by Prof. Pranay Sharma from IIT Bombay's Centre for Machine Intelligence and Data Science. Delve into the complexities of non-stationary reinforcement learning within the infinite-horizon average-reward framework, where Markov Decision Processes feature time-varying rewards and transition probabilities. Learn about the limitations of existing model-based and model-free value-based methods in non-stationary settings, and discover why policy-based methods, despite their practical flexibility, remain theoretically underexplored in this domain. Examine the groundbreaking Non-Stationary Natural Actor-Critic (NS-NAC) algorithm, the first model-free policy-based approach designed specifically for non-stationary environments, featuring restart-based exploration for change detection and innovative learning rate interpretation as adapting factors. Understand the development of BORL-NS-NAC, a bandit-over-RL-based parameter-free algorithm that eliminates the need for prior knowledge of variation budgets. Gain insights from Prof. Sharma's extensive research background in federated learning, collaborative learning, stochastic optimization, reinforcement learning, and differential privacy, drawing from his experience as a Research Scientist at Carnegie Mellon University and his academic journey through Syracuse University and IIT Kanpur.
Syllabus
Time: 4:00 PM - PM IST
Taught by
Centre for Networked Intelligence, IISc