Skip to content

Latest commit

 

History

History
8 lines (5 loc) · 257 Bytes

README.md

File metadata and controls

8 lines (5 loc) · 257 Bytes

Tensorflow-DPPO

self implementation of DPPO, Distributed Proximal Policy Optimization, by using tensorflow

the loss calculation is used from OPENAI PPO. the Distributed architecture design is inspired from Deepmind paper.

TENSORFLOW_MODEL