Butanium commited on
Commit
42497bd
1 Parent(s): 68b09a5

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +9 -8
README.md CHANGED
@@ -3,6 +3,15 @@ pipeline_tag: reinforcement-learning
3
  tags:
4
  - ppo
5
  ---
 
 
 
 
 
 
 
 
 
6
  # Environment
7
  Multiplayer pong_v3 from PettingZoo with :
8
  - 4 stacked frame
@@ -46,14 +55,6 @@ def get_env(args, run_name):
46
  return envs
47
  ```
48
 
49
- # Experiment
50
- PPO agents trained in a selfplay settings. This repo includes checkpoints collected during training for
51
- 4 experiments:
52
- - Shared weights for actor and critic
53
- - No shared weights
54
- - Resume training for extra steps for both shared and no shared setup
55
- Please check our [wandb report](https://wandb.ai/dumas/SPAR_RL_ELK/) for more details and the training code on [our GitHub](https://github.com/Butanium/cleanrl/blob/master/multiplayer_pong/ppo_pettingzoo_ma_atari.py)
56
-
57
  # Model architecture
58
  ```py
59
  def atari_network(orth_init=False):
 
3
  tags:
4
  - ppo
5
  ---
6
+ # Experiment
7
+ PPO agents trained in a selfplay settings. This repo includes checkpoints collected during training for
8
+ 4 experiments:
9
+ - Shared weights for actor and critic
10
+ - No shared weights
11
+ - Resume training for extra steps for both shared and no shared setup
12
+ Please check our [wandb report](https://wandb.ai/dumas/SPAR_RL_ELK/) for more details and the training code on [our GitHub](https://github.com/Butanium/cleanrl/blob/master/multiplayer_pong/ppo_pettingzoo_ma_atari.py)
13
+
14
+
15
  # Environment
16
  Multiplayer pong_v3 from PettingZoo with :
17
  - 4 stacked frame
 
55
  return envs
56
  ```
57
 
 
 
 
 
 
 
 
 
58
  # Model architecture
59
  ```py
60
  def atari_network(orth_init=False):