2020; RL Algorithms Implementation
- Guining Pertin
- Jul 31, 2020
- 1 min read
Introduction
So around end of 2018 I started learning about Reinforcement Learning from Coursera and from David Silver's UCL course. Over time I focused on implementing the algorithms I learned from scratch, mainly in Tensorflow 1.x and then later in 2.x. I also ended up implementing a lot of them during the Covid lockdowns in India.
Check it out at: https://github.com/otoshuki/RL_Projects
This repository contains all of my custom implementations for the following:
Crossentropy
Deep Crossentropy
Value Iteration
Q-Learning with e - greedy exploration
SARSA
Expected Value SARSA with e-greedy exploration
Q-Learning with Experience Replay
Deep Q-Learning on TF1.x and TF2
Deep Q-Learning on TF2 with Experience replay and Target network
REINFORCE - Policy Gradient algorithm on TF1.x and TF2
Double Q-Learning on TF2 with Experience replay
Advantage Actor-Critic on TF2
Proximal Policy Optimization
Twin Delayed DDPG
There are a lot of good sources for the explanation behind most of these algorithms, so I will refrain from the explanations.







Comments