Skip to content

Latest commit

 

History

History
13 lines (9 loc) · 462 Bytes

README.md

File metadata and controls

13 lines (9 loc) · 462 Bytes

RLBook

Bandits

Implementation of the several bandits algorithms for exploration discussed in the Chapter 2.

Policy Iteration

An answer to exercise 4.7 from the book along with the recreation of example 4.2

Value Iteration

Recreated example 4.3 and answered question 4.9

Project Goal

Reinforce the learning from the book by demonstrating learned skills with exercises from Reinforcement Learning Book by Sutton and Barto and my own exercises.