jarski commited on
Commit
6989bae
1 Parent(s): 51f7831

Upload folder using huggingface_hub

Browse files
Files changed (3) hide show
  1. README.md +1 -1
  2. replay.mp4 +2 -2
  3. sf_log.txt +94 -0
README.md CHANGED
@@ -15,7 +15,7 @@ model-index:
15
  type: doom_health_gathering_supreme
16
  metrics:
17
  - type: mean_reward
18
- value: 5.71 +/- 1.94
19
  name: mean_reward
20
  verified: false
21
  ---
 
15
  type: doom_health_gathering_supreme
16
  metrics:
17
  - type: mean_reward
18
+ value: 5.12 +/- 1.81
19
  name: mean_reward
20
  verified: false
21
  ---
replay.mp4 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:cce57097af126fcc01887524d6abe4358ccfe84bf55e340acdf074224e39c67a
3
- size 9749920
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e08a98f0bb243b3545f06655f5950aa864eb28f8c6715db9d22f49d33cfd5833
3
+ size 9003016
sf_log.txt CHANGED
@@ -1448,3 +1448,97 @@ main_loop: 1800.7612
1448
  [2024-09-01 07:03:27,285][00307] Avg episode rewards: #0: 8.908, true rewards: #0: 5.708
1449
  [2024-09-01 07:03:27,288][00307] Avg episode reward: 8.908, avg true_objective: 5.708
1450
  [2024-09-01 07:04:05,352][00307] Replay video saved to /content/train_dir/default_experiment/replay.mp4!
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1448
  [2024-09-01 07:03:27,285][00307] Avg episode rewards: #0: 8.908, true rewards: #0: 5.708
1449
  [2024-09-01 07:03:27,288][00307] Avg episode reward: 8.908, avg true_objective: 5.708
1450
  [2024-09-01 07:04:05,352][00307] Replay video saved to /content/train_dir/default_experiment/replay.mp4!
1451
+ [2024-09-01 07:04:20,071][00307] The model has been pushed to https://huggingface.co/jarski/rl_course_vizdoom_health_gathering_supreme
1452
+ [2024-09-01 07:06:22,936][00307] Loading existing experiment configuration from /content/train_dir/default_experiment/config.json
1453
+ [2024-09-01 07:06:22,939][00307] Overriding arg 'num_workers' with value 1 passed from command line
1454
+ [2024-09-01 07:06:22,942][00307] Adding new argument 'no_render'=True that is not in the saved config file!
1455
+ [2024-09-01 07:06:22,943][00307] Adding new argument 'save_video'=True that is not in the saved config file!
1456
+ [2024-09-01 07:06:22,947][00307] Adding new argument 'video_frames'=1000000000.0 that is not in the saved config file!
1457
+ [2024-09-01 07:06:22,949][00307] Adding new argument 'video_name'=None that is not in the saved config file!
1458
+ [2024-09-01 07:06:22,952][00307] Adding new argument 'max_num_frames'=100000 that is not in the saved config file!
1459
+ [2024-09-01 07:06:22,953][00307] Adding new argument 'max_num_episodes'=10 that is not in the saved config file!
1460
+ [2024-09-01 07:06:22,956][00307] Adding new argument 'push_to_hub'=True that is not in the saved config file!
1461
+ [2024-09-01 07:06:22,957][00307] Adding new argument 'hf_repository'='jarski/rl_course_vizdoom_health_gathering_supreme' that is not in the saved config file!
1462
+ [2024-09-01 07:06:22,960][00307] Adding new argument 'policy_index'=0 that is not in the saved config file!
1463
+ [2024-09-01 07:06:22,961][00307] Adding new argument 'eval_deterministic'=False that is not in the saved config file!
1464
+ [2024-09-01 07:06:22,962][00307] Adding new argument 'train_script'=None that is not in the saved config file!
1465
+ [2024-09-01 07:06:22,964][00307] Adding new argument 'enjoy_script'=None that is not in the saved config file!
1466
+ [2024-09-01 07:06:22,965][00307] Using frameskip 1 and render_action_repeat=4 for evaluation
1467
+ [2024-09-01 07:06:22,984][00307] RunningMeanStd input shape: (3, 72, 128)
1468
+ [2024-09-01 07:06:22,986][00307] RunningMeanStd input shape: (1,)
1469
+ [2024-09-01 07:06:23,002][00307] ConvEncoder: input_channels=3
1470
+ [2024-09-01 07:06:23,049][00307] Conv encoder output size: 512
1471
+ [2024-09-01 07:06:23,051][00307] Policy head output size: 512
1472
+ [2024-09-01 07:06:23,070][00307] Loading state from checkpoint /content/train_dir/default_experiment/checkpoint_p0/checkpoint_000000376_1540096.pth...
1473
+ [2024-09-01 07:06:23,622][00307] Num frames 100...
1474
+ [2024-09-01 07:06:23,804][00307] Num frames 200...
1475
+ [2024-09-01 07:06:24,003][00307] Num frames 300...
1476
+ [2024-09-01 07:06:24,197][00307] Num frames 400...
1477
+ [2024-09-01 07:06:24,396][00307] Num frames 500...
1478
+ [2024-09-01 07:06:24,594][00307] Num frames 600...
1479
+ [2024-09-01 07:06:24,781][00307] Num frames 700...
1480
+ [2024-09-01 07:06:24,962][00307] Num frames 800...
1481
+ [2024-09-01 07:06:25,149][00307] Num frames 900...
1482
+ [2024-09-01 07:06:25,257][00307] Avg episode rewards: #0: 16.280, true rewards: #0: 9.280
1483
+ [2024-09-01 07:06:25,261][00307] Avg episode reward: 16.280, avg true_objective: 9.280
1484
+ [2024-09-01 07:06:25,420][00307] Num frames 1000...
1485
+ [2024-09-01 07:06:25,606][00307] Num frames 1100...
1486
+ [2024-09-01 07:06:25,800][00307] Num frames 1200...
1487
+ [2024-09-01 07:06:25,989][00307] Num frames 1300...
1488
+ [2024-09-01 07:06:26,183][00307] Num frames 1400...
1489
+ [2024-09-01 07:06:26,383][00307] Num frames 1500...
1490
+ [2024-09-01 07:06:26,576][00307] Num frames 1600...
1491
+ [2024-09-01 07:06:26,692][00307] Avg episode rewards: #0: 14.660, true rewards: #0: 8.160
1492
+ [2024-09-01 07:06:26,694][00307] Avg episode reward: 14.660, avg true_objective: 8.160
1493
+ [2024-09-01 07:06:26,828][00307] Num frames 1700...
1494
+ [2024-09-01 07:06:27,016][00307] Num frames 1800...
1495
+ [2024-09-01 07:06:27,202][00307] Num frames 1900...
1496
+ [2024-09-01 07:06:27,429][00307] Avg episode rewards: #0: 11.280, true rewards: #0: 6.613
1497
+ [2024-09-01 07:06:27,433][00307] Avg episode reward: 11.280, avg true_objective: 6.613
1498
+ [2024-09-01 07:06:27,468][00307] Num frames 2000...
1499
+ [2024-09-01 07:06:27,652][00307] Num frames 2100...
1500
+ [2024-09-01 07:06:27,843][00307] Num frames 2200...
1501
+ [2024-09-01 07:06:28,034][00307] Num frames 2300...
1502
+ [2024-09-01 07:06:28,226][00307] Num frames 2400...
1503
+ [2024-09-01 07:06:28,439][00307] Num frames 2500...
1504
+ [2024-09-01 07:06:28,610][00307] Avg episode rewards: #0: 10.900, true rewards: #0: 6.400
1505
+ [2024-09-01 07:06:28,612][00307] Avg episode reward: 10.900, avg true_objective: 6.400
1506
+ [2024-09-01 07:06:28,689][00307] Num frames 2600...
1507
+ [2024-09-01 07:06:28,877][00307] Num frames 2700...
1508
+ [2024-09-01 07:06:29,064][00307] Num frames 2800...
1509
+ [2024-09-01 07:06:29,258][00307] Num frames 2900...
1510
+ [2024-09-01 07:06:29,462][00307] Num frames 3000...
1511
+ [2024-09-01 07:06:29,534][00307] Avg episode rewards: #0: 9.816, true rewards: #0: 6.016
1512
+ [2024-09-01 07:06:29,536][00307] Avg episode reward: 9.816, avg true_objective: 6.016
1513
+ [2024-09-01 07:06:29,711][00307] Num frames 3100...
1514
+ [2024-09-01 07:06:29,941][00307] Num frames 3200...
1515
+ [2024-09-01 07:06:30,126][00307] Num frames 3300...
1516
+ [2024-09-01 07:06:30,319][00307] Num frames 3400...
1517
+ [2024-09-01 07:06:30,528][00307] Num frames 3500...
1518
+ [2024-09-01 07:06:30,682][00307] Avg episode rewards: #0: 9.420, true rewards: #0: 5.920
1519
+ [2024-09-01 07:06:30,685][00307] Avg episode reward: 9.420, avg true_objective: 5.920
1520
+ [2024-09-01 07:06:30,776][00307] Num frames 3600...
1521
+ [2024-09-01 07:06:30,970][00307] Num frames 3700...
1522
+ [2024-09-01 07:06:31,167][00307] Num frames 3800...
1523
+ [2024-09-01 07:06:31,246][00307] Avg episode rewards: #0: 8.440, true rewards: #0: 5.440
1524
+ [2024-09-01 07:06:31,249][00307] Avg episode reward: 8.440, avg true_objective: 5.440
1525
+ [2024-09-01 07:06:31,433][00307] Num frames 3900...
1526
+ [2024-09-01 07:06:31,639][00307] Num frames 4000...
1527
+ [2024-09-01 07:06:31,828][00307] Num frames 4100...
1528
+ [2024-09-01 07:06:32,021][00307] Num frames 4200...
1529
+ [2024-09-01 07:06:32,177][00307] Avg episode rewards: #0: 8.070, true rewards: #0: 5.320
1530
+ [2024-09-01 07:06:32,179][00307] Avg episode reward: 8.070, avg true_objective: 5.320
1531
+ [2024-09-01 07:06:32,266][00307] Num frames 4300...
1532
+ [2024-09-01 07:06:32,461][00307] Num frames 4400...
1533
+ [2024-09-01 07:06:32,655][00307] Num frames 4500...
1534
+ [2024-09-01 07:06:32,878][00307] Num frames 4600...
1535
+ [2024-09-01 07:06:33,134][00307] Avg episode rewards: #0: 7.858, true rewards: #0: 5.191
1536
+ [2024-09-01 07:06:33,137][00307] Avg episode reward: 7.858, avg true_objective: 5.191
1537
+ [2024-09-01 07:06:33,212][00307] Num frames 4700...
1538
+ [2024-09-01 07:06:33,484][00307] Num frames 4800...
1539
+ [2024-09-01 07:06:33,752][00307] Num frames 4900...
1540
+ [2024-09-01 07:06:34,020][00307] Num frames 5000...
1541
+ [2024-09-01 07:06:34,281][00307] Num frames 5100...
1542
+ [2024-09-01 07:06:34,394][00307] Avg episode rewards: #0: 7.620, true rewards: #0: 5.120
1543
+ [2024-09-01 07:06:34,398][00307] Avg episode reward: 7.620, avg true_objective: 5.120
1544
+ [2024-09-01 07:07:08,879][00307] Replay video saved to /content/train_dir/default_experiment/replay.mp4!