Deep Reinforced Learning Resources

Resources for DRL and the actor-critic model submitted by students in Masters of Science in AI.

Everything You Need To Master Actor Critic Methods | Tensorflow 2 Tutorial by MachineLearningWithPhil

Forty-minute YouTube video on deep reinforcement learning and the basic concepts behind actor-critic models. This video is a very good introduction on the fundamentals of reinforcement learning.

Video
An intro to Advantage Actor Critic methods: let’s play Sonic the Hedgehog by freeCodeCamp

Learn to implement an Advantage Actor Critic (A2C) agent that learns to play Sonic the Hedgehog.

Tutorial
Playing CartPole with the Actor-Critic method

Comprehensive with code examples - Easy interactive tutorial (runs effortlessly on Colab) - Familiar, TensorFlow style coding focus tutorial - hands-on guide

Tutorial
Advantage Actor-Critic (A2C)

Easy format via relevant platform

Tutorial
Siraj Raval’s “Actor Critic Algorithms”

Concise, plus video games

Video
Keras Documentation on Actor Critic Method

The Keras documentation offers a clear and concise example of implementing the Actor-Critic method in the CartPole environment.

Tutorial
PyLessons Tutorial on Actor-Critic Models

This tutorial offers a step-by-step guide on implementing an Actor-Critic model.

Tutorial
Unlocking the Secrets of Actor-Critic Reinforcement Learning: A Beginner’s Guide

Brief, Coding focus.

Tutorial
Understanding Actor Critic Methods and A2C

Explains the math behind Actor Critic architecture - demystify how Actor-Critic methods enable an agent (like a robot) to learn from interactions and feedback, improving its decision-making over time

Tutorial
Deep Reinforcement Learning Explained

If you want to take deep dive into deep reinforcement learning: Covers complex topic in very simple language and explanations.

Tutorial
Asynchronous Methods for Deep Reinforcement Learning" by Volodymyr Mnih et al.

In-depth exploration of Actor-Critic methods in the context of asynchronous reinforcement learning.

Paper
Actor-Critic Algorithms by Lilian Weng

In-depth exploration of Actor-Critic methods in the context of asynchronous reinforcement learning

Tutorial
Deep Reinforcement Learning by David Silver (University College London)

Insightful lecture series covering Actor-Critic methods and their applications in depth.

Video Series
3Blue1Brown - Actor-Critic Reinforcement Learning

This is a 5 hour crash course training on the Advanced Actor Critic Methods.

Video
Actor-Critic Models

To find comprehensive materials on a pivotal DRL approach that balances action decision-making (actor) with performance evaluation (critic) for learning optimization.

Book Chapter
Actor-Critic Algorithms

To explore specific implementations, variations, and advancements of Actor-Critic methods that show how theoretical models are applied in practical scenarios, enhancing machine learning capabilities.

Paper
CS885 Reinforcement Learnin

This entire playlist is a very good learning resource for DRL. Lectures 7a and 7b gives an overview of the Actor and Critic model.

Video Playlist
The Actor-Critic Reinforcement Learning algorithm

Explanation with math

Article
A deep actor critic reinforcement learning framework for learning to rank

Proposes a Deep Reinforcement learning based approach for Learning to rank task.

Paper
Actor-Critic Algorithms

A lecture note from UC Berkeley's deep reinforcement learning course. It provides a theoretical understanding of Actor-Critic algorithms, making it a good resource for those who prefer the academic approach.

Presentation
The Actor-Dueling-Critic Method for Reinforcement Learning

An approach based on the actor-critic framework, and in the critic branch we modify the manner of estimating Q-value by introducing the advantage function, such as dueling network, which can estimate the action-advantage value

Paper
Actor-Critic Deep Reinforcement Learning for Solving Job Shop Scheduling Problems

This work views JSSP as a sequential decision making problem and proposes to use deep reinforcement learning to cope with this problem.

Paper