Skip to content

Latest commit

 

History

History
15 lines (10 loc) · 478 Bytes

README.md

File metadata and controls

15 lines (10 loc) · 478 Bytes

PPO Lagrangian Reproduction in Pytorch

Implementation of PPO Lagrangian from Benchmarking Safe Exploration in Deep Reinforcement Learning Paper (Ray et al, 2019) in PyTorch

python ppo.py

Results

  1. Reward Returns
    reward
  2. Cost Returns (Cost limit=25)
    cost