GitHunt

This repo is from my Master's degree thesis work develped at Addfor s.p.a

I used PlaNet to prove that model-based DRL can overcome the model-free algorithms in terms of sample efficiency.
My implementation of PlaNet is based on the Kaixhin one, but I reach better results. I also experiment with a regularizer based on DAE to reduce the gap between the real and the predicted rewards.

The company asks me to note publish that feature, but you can find all the explanations in my blog article (you can also contact me).

Full trained agent

PlaNet Overview

General overview of Planet model architecture. If you want a full explanation, click on it!
Blog_article

Medium Articles

Results

ceetah_planet_vs_ddpg
cartpole_planet_vs_ddpg
reacher_planet_vs_ddpg
walker_planet_vs_ddpg
my_planet_vs_soa

Comparison's data are form:
Curl: Contrastive unsupervised representations for reinforcement learning. Laskin, M., Srinivas, A., & Abbeel, P. (2020, July)

Requirements

Acknowledgements

References

[1] Learning Latent Dynamics for Planning from Pixels
[2] Overcoming the limits of DRL using a model-based approach

Languages

Python100.0%

Contributors

Created July 19, 2020
Updated November 10, 2024
DrLux/Planpix | GitHunt