From 4ab8a209dd10d89f0c8503a27eb6dfda46683289 Mon Sep 17 00:00:00 2001
From: Chris MacLellan <2348-cm3786@users.noreply.gitlab.cci.drexel.edu>
Date: Thu, 1 Oct 2020 10:45:56 -0400
Subject: [PATCH] forgot to include ppo example

---
 test_ppo.py | 35 +++++++++++++++++++++++++++++++++++
 1 file changed, 35 insertions(+)
 create mode 100644 test_ppo.py

diff --git a/test_ppo.py b/test_ppo.py
new file mode 100644
index 0000000..9d4fdf7
--- /dev/null
+++ b/test_ppo.py
@@ -0,0 +1,35 @@
+import gym
+from stable_baselines.common.policies import MlpPolicy
+from stable_baselines.common import make_vec_env
+from stable_baselines import PPO2
+import tutorenvs
+import numpy as np
+
+
+if __name__ == "__main__":
+
+    # multiprocess environment
+    env = make_vec_env('FractionArith-v0', n_envs=8)
+
+    model = PPO2(MlpPolicy, env, verbose=0,
+            gamma=0.5,
+            tensorboard_log="./ppo_FractionArith-v0/")
+
+
+    while True:
+        model.learn(total_timesteps=999999999)
+        # model.save("ppo2_cartpole")
+
+        # del model # remove to demonstrate saving and loading
+
+        # model = PPO2.load("ppo2_cartpole")
+
+        # Enjoy trained agent
+        # obs = env.reset()
+        # rwd = 0
+        # for _ in range(100):
+        #     action, _states = model.predict(obs)
+        #     obs, rewards, dones, info = env.step(action)
+        #     rwd += np.sum(rewards)
+        #     # env.render()
+        # print(rwd)
-- 
GitLab