Skip to content

Latest commit

 

History

History
35 lines (17 loc) · 1022 Bytes

README.md

File metadata and controls

35 lines (17 loc) · 1022 Bytes

Value based method

In this repository we're trying to solve 2 openAI Gym's env using two policy based methods: Hill climbing and Cross-Entropy method

Hill Climbing

Hill_Climbing.ipynb is an implementation of hill climbing with adaptive noise scaling for OpenAI Gym's Cartpole environment.

Result

Trained Agent

Cross-Entropy Method

CEM.ipynb is an implementation of the cross-entropy method for OpenAI Gym's MountainCarContinuous environment.

Result

Trained Agent

Additionals

For more well explained methods for policy based method here's a good blog:

http://kvfrans.com/simple-algoritms-for-solving-cartpole/

--> corresponding github: https://github.com/kvfrans/openai-cartpole