Sample-based Learning Methods
This course is part of Reinforcement Learning Specialization
Instructors: Martha White +1 more
Instructor ratings
We asked all learners to give feedback on our instructors based on the quality of their teaching style.
Skills you'll gain
There are 5 modules in this course
By the end of this course you will be able to: - Understand Temporal-Difference learning and Monte Carlo as two strategies for estimating value functions from sampled experience - Understand the importance of exploration, when using sampled experience rather than dynamic programming sweeps within a model - Understand the connections between Monte Carlo and Dynamic Programming and TD. - Implement and apply the TD algorithm, for estimating value functions - Implement and apply Expected Sarsa and Q-learning (two TD methods for control) - Understand the difference between on-policy and off-policy control - Understand planning with simulated experience (as opposed to classic planning strategies) - Implement a model-based approach to RL, called Dyna, which uses simulated experience - Conduct an empirical study to see the improvements in sample efficiency when using Dyna
Monte Carlo Methods for Prediction & Control
Temporal Difference Learning Methods for Prediction
Temporal Difference Learning Methods for Control
Planning, Learning & Acting
Explore more from Machine Learning
©2025 ementorhub.com. All rights reserved