Companies Home Search Profile

Reinforcement Learning (English): Master the Art of RL

Focused View

Coursat.ai Dr. Ahmad ElSallab

9:01:21

191 View
  • 1. Course introduction.mp4
    01:17
  • 2. Course overview.mp4
    04:51
  • 1. Module intro and roadmap.mp4
    03:03
  • 2. What is RL.mp4
    07:29
  • 3. What RL can do.mp4
    22:36
  • 4. The RL problem setup (AREA).mp4
    06:51
  • 5. Reward.mp4
    08:03
  • 6. RL vs. Supervised Learning.mp4
    :
  • 7. State.mp4
    34:21
  • 8. AREA examples and quizes.mp4
    25:19
  • 9. Gym Environments.mp4
    20:12
  • 10. Inside RL agent - RL agent ingredients.mp4
    02:44
  • 11. Policy.mp4
    03:39
  • 12. Value.mp4
    07:12
  • 13. Model.mp4
    06:53
  • 14. RL agents taxonomy.mp4
    15:39
  • 15. Prediction vs Control.mp4
    03:36
  • 1. Module intro and roadmap.mp4
    02:27
  • 2. Markov Chain and Markov Process (MP).mp4
    11:23
  • 3. Markov Reward Process (MRP).mp4
    27:09
  • 4. Markov Decision Process (MDP).mp4
    16:52
  • 5. Prediction.mp4
    08:50
  • 6. Bellman Equations with action-value function Q.mp4
    06:19
  • 7. Control.mp4
    12:18
  • 1. Module intro and roadmap.mp4
    07:32
  • 2. Planning with Dynamic Programming (DP).mp4
    27:31
  • 3. Prediction with DP - Policy Evaluation.mp4
    05:32
  • 4. Control with DP - Policy Iteration and Value Iteration.mp4
    08:18
  • 5. Value Iteration example.mp4
    06:55
  • 6. Prediction with Monte-Carlo - MC Policy Evaluation.mp4
    10:50
  • 7. Prediction with Temporal-Difference (TD).mp4
    20:09
  • 8. TD Lambda.mp4
    04:20
  • 9. Control with Monte-Carlo - MC Policy Iteration.mp4
    10:14
  • 10. Control with TD - SARSA.mp4
    04:56
  • 11. On-policy vs. Off-policy.mp4
    02:57
  • 12. Q-learning.mp4
    06:00
  • 13. MDP solutions summary.mp4
    06:33
  • 1. Module intro and roadmap.mp4
    01:24
  • 2. Large Scale Reinforcement Learning.mp4
    10:23
  • 3. DNN as function approximator.mp4
    14:50
  • 4. Value Function Approximation.mp4
    06:12
  • 5. DNN policies.mp4
    06:59
  • 6. Value function approximation with DL encoder-decoder pattern.mp4
    13:43
  • 7. Deep Q-Networks (DQN).mp4
    05:18
  • 8. DQN Atari Example with Keras-RL and TF-Agents.mp4
    07:31
  • 1. Module intro and roadmap.mp4
    01:19
  • 2. Value-based vs Policy based vs Actor-Critic.mp4
    02:15
  • 3. Policy Gradients (PG).mp4
    10:23
  • 4. REINFORCE - Monte-Carlo PG.mp4
    03:45
  • 5. AC - Actor-Critic.mp4
    07:19
  • 6. A2C - Advantage Actor-Critic.mp4
    05:53
  • 7. A3C - Asynchronous Advantage Actor-Critic.mp4
    01:53
  • 8. TRPO - Trusted Region Policy Optimization.mp4
    01:53
  • 9. PPO - Proximal Policy Optimization.mp4
    02:20
  • 10. DDPG - Deep Determinstic Policy Gradients.mp4
    08:04
  • 11. StableBaselines library overview.mp4
    11:06
  • 12. Atari example with stable-baselines.mp4
    01:26
  • 13. Mario example with stable-baselines.mp4
    03:22
  • 14. StreetFighter example with stable-baselines.mp4
    05:36
  • 1. Module intro and roadmap.mp4
    04:10
  • 2. Model learning methods.mp4
    02:22
  • 3. Model learning with Supervised Learning and Function Approximation.mp4
    07:46
  • 4. Sample based planning.mp4
    02:32
  • 5. Dyna - Intergation planning and Learning.mp4
    07:47
  • 1. Conclusion.mp4
    03:00
  • 1. Slides.html
  • Description


    Reinforcement Learning

    What You'll Learn?


    • Define what is Reinforcement Learning?
    • Apply all what is learned using state-of-the art libraries like OpenAI Gym, StabeBaselines, Keras-RL and TensorFlow Agents
    • Define what are the applications domains and success stories of RL?
    • Define what are the difference between Reinforcement and Supervised Learning?
    • Define the main components of an RL problem setup?
    • Define what are the main ingredients of an RL agent and their taxonomy?
    • Define what is Markov Reward Process (MRP) and Markov Decision Process (MDP)?
    • Define the solution space of RL using MDP framework
    • Solve the RL problems using planning with Dynamic Programming algorithms, like Policy Evaluation, Policy Iteration and Value Iteration
    • Solve RL problems using model free algorithms like Monte-Carlo, TD learning, Q-learning and SARSA
    • Differentiate On-policy and Off-policy algorithms
    • Master Deep Reinforcement Learning algorithms like Deep Q-Networks (DQN), and apply them to Large Scale RL
    • Master Policy Gradients algorithms and Actor-Critic (AC, A2C, A3C)
    • Master advanced DRL algorithms like DDPG, TRPO and PPO
    • Define what is model-based RL, and differentiate it from planning, and what are their main algorithms and applications?

    Who is this for?


  • Machine Learning Researchers
  • Machine Learning Engineers
  • Data Scientists
  • What You Need to Know?


  • Machine Learning basics
  • Deep Learning basics
  • Probability
  • Programming and Problem solving basics
  • Python programming
  • More details


    Description

    Hello and welcome to our course; Reinforcement Learning.

    Reinforcement Learning is a very exciting and important field of Machine Learning and AI. Some call it the crown jewel of AI.

    In this course, we will cover all the aspects related to Reinforcement Learning or RL. We will start by defining the RL problem, and compare it to the Supervised Learning problem, and discover the areas of applications where RL can excel. This includes the problem formulation, starting from the very basics to the advanced usage of Deep Learning, leading to the era of Deep Reinforcement Learning.

    In our journey, we will cover, as usual, both the theoretical and practical aspects, where we will learn how to implement the RL algorithms and apply them to the famous problems using libraries like OpenAI Gym, Keras-RL, TensorFlow Agents or TF-Agents and Stable Baselines.

    The course is divided into 6 main sections:

    1- We start with an introduction to the RL problem definition, mainly comparing it to the Supervised learning problem, and discovering the application domains and the main constituents of an RL problem. We describe here the famous OpenAI Gym environments, which will be our playground when it comes to practical implementation of the algorithms that we learn about.


    2- In the second part we discuss the main formulation of an RL problem as a Markov Decision Process or MDP, with simple solution to the most basic problems using Dynamic Programming.


    3- After being armed with an understanding of MDP, we move on to explore the solution space of the MDP problem, and what the different solutions beyond DP, which includes model-based and model-free solutions. We will focus in this part on model-free solutions, and defer model-based solutions to the last part. In this part, we describe the Monte-Carlo and Temporal-Difference sampling based methods, including the famous and important Q-learning algorithm, and SARSA. We will describe the practical usage and implementation of Q-learning and SARSA on control tabular maze problems from OpenAI Gym environments.


    4- To move beyond simple tabular problems, we will need to learn about function approximation in RL, which leads to the mainstream RL methods today using Deep Learning, or Deep Reinforcement Learning (DRL). We will describe here the breakthrough algorithm of DeepMind that solved the Atari games and AlphaGO, which is Deep Q-Networks or DQN. We also discuss how we can solve Atari games problems using DQN in practice using Keras-RL and TF-Agents.


    5- In the fifth part, we move to Advanced DRL algorithms, mainly under a family called Policy based methods. We discuss here Policy Gradients, DDPG, Actor-Critic, A2C, A3C, TRPO and PPO methods. We also discuss the important Stable Baseline library to implement all those algorithms on different environments in OpenAI Gym, like Atari and others.


    6- Finally, we explore the model-based family of RL methods, and importantly, differentiating model-based RL from planning, and exploring the whole spectrum of RL methods.


    Hopefully, you enjoy this course, and find it useful.



    Who this course is for:

    • Machine Learning Researchers
    • Machine Learning Engineers
    • Data Scientists

    User Reviews
    Rating
    0
    0
    0
    0
    0
    average 0
    Total votes0
    Focused display
    Coursat.ai Dr. Ahmad ElSallab
    Coursat.ai Dr. Ahmad ElSallab
    Instructor's Courses
    Coursat AI is a platform for project-based courses in AI. The courses offer end-to-end project experience, through three steps: Apply, Refine and Deploy. Participants will enrich their projects portfolio with state-of-the art projects in Data Science, Deep Learning, Computer Vision, NLP and Robotics. Instructors are professional experts, with wide industrial experience in top tech companies.
    Students take courses primarily to improve job-related skills.Some courses generate credit toward technical certification. Udemy has made a special effort to attract corporate trainers seeking to create coursework for employees of their company.
    • language english
    • Training sessions 65
    • duration 9:01:21
    • Release Date 2023/06/23