Logo

$100

Prediction and Control with Function Approximation

Created by -

Martha White,Adam White
,
University of Alberta

0.00

(0 ratings)

English

Wishlist

Overview

In this course, you will learn how to solve problems with large, high-dimensional, and potentially infinite state spaces. You will see that estimating value functions can be cast as a supervised learning problem---function approximation---allowing you to build agents that carefully balance generalization and discrimination in order to maximize reward. We will begin this journey by investigating how our policy evaluation or prediction methods like Monte Carlo and TD can be extended to the function approximation setting. You will learn about feature construction techniques for RL, and representation learning via neural networks and backprop. We conclude this course with a deep-dive into policy gradient methods; a way to learn policies directly without learning a value function. In this course you will solve two continuous-state control tasks and investigate the benefits of policy gradient methods in a continuous-action environment.

course image

USD 100

provider image

Type: Online

This course includes

  • Approx. 15 hours to complete
  • Earn a Certificate upon completion
  • Start instantly and learn at your own schedule.

Taken this course?

Share your experience with other students

Share

Add Review

course image

USD 100

provider image

Type: Online

This course includes

  • Approx. 15 hours to complete
  • Earn a Certificate upon completion
  • Start instantly and learn at your own schedule.

Taken this course?

Share your experience with other students

Share

Add Review

Prediction and Control with Function Approximation

Created by -

Martha White,Adam White
,
University of Alberta

0.00

(0 ratings)

All Levels

Start Date: February 10th 2021

Course Description

In this course, you will learn how to solve problems with large, high-dimensional, and potentially infinite state spaces. You will see that estimating value functions can be cast as a supervised learning problem---function approximation---allowing you to build agents that carefully balance generalization and discrimination in order to maximize reward. We will begin this journey by investigating how our policy evaluation or prediction methods like Monte Carlo and TD can be extended to the function approximation setting. You will learn about feature construction techniques for RL, and representation learning via neural networks and backprop. We conclude this course with a deep-dive into policy gradient methods; a way to learn policies directly without learning a value function. In this course you will solve two continuous-state control tasks and investigate the benefits of policy gradient methods in a continuous-action environment.

The information used on this page is how each course is described on the Coursera platform.

Course Structure

Tags

Mark Complete


About the Instructor

Martha White,Adam White,University of Alberta

No Reviews at this moment.

Explore Skillqore

Skillqore Newsletter

Keep me up to date with content, updates, and offers from Skillqore


Copyright © 2020 Skillqore, Inc. All Rights Reserved.