davidberenstein1957 HF staff commited on
Commit
af08012
1 Parent(s): 099a74c

Model save

Browse files
README.md ADDED
@@ -0,0 +1,94 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: cc-by-nc-4.0
3
+ base_model: BramVanroy/GEITje-7B-ultra-sft
4
+ tags:
5
+ - generated_from_trainer
6
+ model-index:
7
+ - name: ultra-feedback-dutch-cleaned-hq-spin-geitje-7b-ultra-sft_iter0
8
+ results: []
9
+ ---
10
+
11
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
12
+ should probably proofread and complete it, then remove this comment. -->
13
+
14
+ # ultra-feedback-dutch-cleaned-hq-spin-geitje-7b-ultra-sft_iter0
15
+
16
+ This model is a fine-tuned version of [BramVanroy/GEITje-7B-ultra-sft](https://huggingface.co/BramVanroy/GEITje-7B-ultra-sft) on the None dataset.
17
+ It achieves the following results on the evaluation set:
18
+ - Loss: 0.0135
19
+ - Rewards/real: -1.4818
20
+ - Rewards/generated: -13.3376
21
+ - Rewards/accuracies: 0.9963
22
+ - Rewards/margins: 11.8558
23
+ - Logps/generated: -410.0757
24
+ - Logps/real: -427.4978
25
+ - Logits/generated: -2.7305
26
+ - Logits/real: -2.7643
27
+
28
+ ## Model description
29
+
30
+ More information needed
31
+
32
+ ## Intended uses & limitations
33
+
34
+ More information needed
35
+
36
+ ## Training and evaluation data
37
+
38
+ More information needed
39
+
40
+ ## Training procedure
41
+
42
+ ### Training hyperparameters
43
+
44
+ The following hyperparameters were used during training:
45
+ - learning_rate: 5e-07
46
+ - train_batch_size: 8
47
+ - eval_batch_size: 8
48
+ - seed: 42
49
+ - distributed_type: multi-GPU
50
+ - num_devices: 4
51
+ - gradient_accumulation_steps: 2
52
+ - total_train_batch_size: 64
53
+ - total_eval_batch_size: 32
54
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
55
+ - lr_scheduler_type: linear
56
+ - lr_scheduler_warmup_ratio: 0.1
57
+ - num_epochs: 2
58
+
59
+ ### Training results
60
+
61
+ | Training Loss | Epoch | Step | Validation Loss | Rewards/real | Rewards/generated | Rewards/accuracies | Rewards/margins | Logps/generated | Logps/real | Logits/generated | Logits/real |
62
+ |:-------------:|:-----:|:----:|:---------------:|:------------:|:-----------------:|:------------------:|:---------------:|:---------------:|:----------:|:----------------:|:-----------:|
63
+ | 0.4944 | 0.08 | 25 | 0.2566 | 0.6645 | -0.8427 | 0.9761 | 1.5071 | -285.1264 | -406.0350 | -3.0069 | -3.0147 |
64
+ | 0.092 | 0.16 | 50 | 0.0838 | 0.3983 | -3.7771 | 0.9890 | 4.1754 | -314.4705 | -408.6964 | -2.9427 | -2.9557 |
65
+ | 0.0601 | 0.25 | 75 | 0.0457 | 0.2564 | -5.6388 | 0.9963 | 5.8952 | -333.0871 | -410.1154 | -2.9205 | -2.9326 |
66
+ | 0.0437 | 0.33 | 100 | 0.0336 | -0.1853 | -7.2451 | 0.9963 | 7.0598 | -349.1503 | -414.5328 | -2.8883 | -2.9062 |
67
+ | 0.036 | 0.41 | 125 | 0.0271 | -0.1651 | -7.7408 | 0.9945 | 7.5756 | -354.1071 | -414.3309 | -2.8817 | -2.9014 |
68
+ | 0.0373 | 0.49 | 150 | 0.0264 | -0.2384 | -7.8312 | 0.9908 | 7.5928 | -355.0117 | -415.0634 | -2.8271 | -2.8543 |
69
+ | 0.0198 | 0.58 | 175 | 0.0214 | -0.9152 | -9.9469 | 0.9908 | 9.0317 | -376.1681 | -421.8315 | -2.8052 | -2.8326 |
70
+ | 0.0426 | 0.66 | 200 | 0.0251 | -0.9747 | -9.1022 | 0.9908 | 8.1275 | -367.7210 | -422.4266 | -2.8450 | -2.8588 |
71
+ | 0.0262 | 0.74 | 225 | 0.0189 | -0.8414 | -9.9318 | 0.9926 | 9.0903 | -376.0172 | -421.0940 | -2.8009 | -2.8209 |
72
+ | 0.0142 | 0.82 | 250 | 0.0166 | -0.7154 | -10.1059 | 0.9945 | 9.3905 | -377.7586 | -419.8336 | -2.7973 | -2.8201 |
73
+ | 0.0171 | 0.9 | 275 | 0.0189 | -1.0905 | -10.9057 | 0.9945 | 9.8151 | -385.7561 | -423.5849 | -2.7641 | -2.7936 |
74
+ | 0.0333 | 0.99 | 300 | 0.0168 | -1.2797 | -11.4866 | 0.9963 | 10.2069 | -391.5655 | -425.4765 | -2.7973 | -2.8230 |
75
+ | 0.0061 | 1.07 | 325 | 0.0157 | -1.2079 | -11.1880 | 0.9945 | 9.9801 | -388.5797 | -424.7587 | -2.7974 | -2.8231 |
76
+ | 0.0022 | 1.15 | 350 | 0.0152 | -1.0695 | -11.2438 | 0.9908 | 10.1743 | -389.1376 | -423.3746 | -2.7853 | -2.8128 |
77
+ | 0.0033 | 1.23 | 375 | 0.0148 | -1.1767 | -11.6618 | 0.9908 | 10.4851 | -393.3175 | -424.4465 | -2.7751 | -2.8029 |
78
+ | 0.0043 | 1.32 | 400 | 0.0138 | -1.0951 | -11.8306 | 0.9963 | 10.7354 | -395.0049 | -423.6307 | -2.7703 | -2.7976 |
79
+ | 0.005 | 1.4 | 425 | 0.0136 | -1.3179 | -12.4674 | 0.9963 | 11.1494 | -401.3733 | -425.8589 | -2.7573 | -2.7851 |
80
+ | 0.0031 | 1.48 | 450 | 0.0139 | -1.3771 | -12.6901 | 0.9963 | 11.3130 | -403.6003 | -426.4503 | -2.7544 | -2.7815 |
81
+ | 0.0039 | 1.56 | 475 | 0.0134 | -1.3885 | -12.8092 | 0.9963 | 11.4207 | -404.7912 | -426.5648 | -2.7446 | -2.7735 |
82
+ | 0.001 | 1.64 | 500 | 0.0136 | -1.4378 | -13.0038 | 0.9963 | 11.5660 | -406.7370 | -427.0571 | -2.7404 | -2.7701 |
83
+ | 0.0059 | 1.73 | 525 | 0.0139 | -1.5924 | -13.4168 | 0.9945 | 11.8244 | -410.8671 | -428.6035 | -2.7293 | -2.7629 |
84
+ | 0.0015 | 1.81 | 550 | 0.0136 | -1.5136 | -13.3984 | 0.9963 | 11.8848 | -410.6832 | -427.8157 | -2.7283 | -2.7623 |
85
+ | 0.0078 | 1.89 | 575 | 0.0135 | -1.4891 | -13.3323 | 0.9963 | 11.8432 | -410.0224 | -427.5704 | -2.7309 | -2.7645 |
86
+ | 0.0043 | 1.97 | 600 | 0.0135 | -1.4818 | -13.3376 | 0.9963 | 11.8558 | -410.0757 | -427.4978 | -2.7305 | -2.7643 |
87
+
88
+
89
+ ### Framework versions
90
+
91
+ - Transformers 4.37.0
92
+ - Pytorch 2.1.2+cu121
93
+ - Datasets 2.14.6
94
+ - Tokenizers 0.15.2
all_results.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 2.0,
3
+ "train_loss": 0.04328188924946038,
4
+ "train_runtime": 16735.9464,
5
+ "train_samples": 19426,
6
+ "train_samples_per_second": 2.321,
7
+ "train_steps_per_second": 0.036
8
+ }
generation_config.json ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 1,
4
+ "eos_token_id": 2,
5
+ "transformers_version": "4.37.0"
6
+ }
model-00001-of-00003.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a7f98b14da4446e0ad92e22c0493121d810ec8ccae48c8ea74f302853c51bd8d
3
+ size 4943162336
model-00002-of-00003.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0b7cfdc491db64695ea37a5d1f2df1a8bb8cbf8987c0305fe5b811f77cb99e78
3
+ size 4999819336
model-00003-of-00003.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4ea9a6999477e31658a79cc2e8a17a4cb2add12d37babd897009e914f7f79f76
3
+ size 4540516344
model.safetensors.index.json ADDED
@@ -0,0 +1,298 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "metadata": {
3
+ "total_size": 14483464192
4
+ },
5
+ "weight_map": {
6
+ "lm_head.weight": "model-00003-of-00003.safetensors",
7
+ "model.embed_tokens.weight": "model-00001-of-00003.safetensors",
8
+ "model.layers.0.input_layernorm.weight": "model-00001-of-00003.safetensors",
9
+ "model.layers.0.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
10
+ "model.layers.0.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
11
+ "model.layers.0.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
12
+ "model.layers.0.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
13
+ "model.layers.0.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
14
+ "model.layers.0.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
15
+ "model.layers.0.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
16
+ "model.layers.0.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
17
+ "model.layers.1.input_layernorm.weight": "model-00001-of-00003.safetensors",
18
+ "model.layers.1.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
19
+ "model.layers.1.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
20
+ "model.layers.1.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
21
+ "model.layers.1.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
22
+ "model.layers.1.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
23
+ "model.layers.1.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
24
+ "model.layers.1.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
25
+ "model.layers.1.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
26
+ "model.layers.10.input_layernorm.weight": "model-00002-of-00003.safetensors",
27
+ "model.layers.10.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
28
+ "model.layers.10.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
29
+ "model.layers.10.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
30
+ "model.layers.10.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
31
+ "model.layers.10.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
32
+ "model.layers.10.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
33
+ "model.layers.10.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
34
+ "model.layers.10.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
35
+ "model.layers.11.input_layernorm.weight": "model-00002-of-00003.safetensors",
36
+ "model.layers.11.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
37
+ "model.layers.11.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
38
+ "model.layers.11.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
39
+ "model.layers.11.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
40
+ "model.layers.11.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
41
+ "model.layers.11.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
42
+ "model.layers.11.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
43
+ "model.layers.11.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
44
+ "model.layers.12.input_layernorm.weight": "model-00002-of-00003.safetensors",
45
+ "model.layers.12.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
46
+ "model.layers.12.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
47
+ "model.layers.12.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
48
+ "model.layers.12.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
49
+ "model.layers.12.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
50
+ "model.layers.12.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
51
+ "model.layers.12.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
52
+ "model.layers.12.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
53
+ "model.layers.13.input_layernorm.weight": "model-00002-of-00003.safetensors",
54
+ "model.layers.13.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
55
+ "model.layers.13.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
56
+ "model.layers.13.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
57
+ "model.layers.13.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
58
+ "model.layers.13.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
59
+ "model.layers.13.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
60
+ "model.layers.13.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
61
+ "model.layers.13.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
62
+ "model.layers.14.input_layernorm.weight": "model-00002-of-00003.safetensors",
63
+ "model.layers.14.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
64
+ "model.layers.14.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
65
+ "model.layers.14.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
66
+ "model.layers.14.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
67
+ "model.layers.14.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
68
+ "model.layers.14.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
69
+ "model.layers.14.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
70
+ "model.layers.14.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
71
+ "model.layers.15.input_layernorm.weight": "model-00002-of-00003.safetensors",
72
+ "model.layers.15.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
73
+ "model.layers.15.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
74
+ "model.layers.15.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
75
+ "model.layers.15.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
76
+ "model.layers.15.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
77
+ "model.layers.15.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
78
+ "model.layers.15.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
79
+ "model.layers.15.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
80
+ "model.layers.16.input_layernorm.weight": "model-00002-of-00003.safetensors",
81
+ "model.layers.16.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
82
+ "model.layers.16.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
83
+ "model.layers.16.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
84
+ "model.layers.16.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
85
+ "model.layers.16.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
86
+ "model.layers.16.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
87
+ "model.layers.16.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
88
+ "model.layers.16.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
89
+ "model.layers.17.input_layernorm.weight": "model-00002-of-00003.safetensors",
90
+ "model.layers.17.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
91
+ "model.layers.17.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
92
+ "model.layers.17.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
93
+ "model.layers.17.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
94
+ "model.layers.17.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
95
+ "model.layers.17.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
96
+ "model.layers.17.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
97
+ "model.layers.17.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
98
+ "model.layers.18.input_layernorm.weight": "model-00002-of-00003.safetensors",
99
+ "model.layers.18.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
100
+ "model.layers.18.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
101
+ "model.layers.18.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
102
+ "model.layers.18.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
103
+ "model.layers.18.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
104
+ "model.layers.18.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
105
+ "model.layers.18.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
106
+ "model.layers.18.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
107
+ "model.layers.19.input_layernorm.weight": "model-00002-of-00003.safetensors",
108
+ "model.layers.19.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
109
+ "model.layers.19.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
110
+ "model.layers.19.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
111
+ "model.layers.19.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
112
+ "model.layers.19.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
113
+ "model.layers.19.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
114
+ "model.layers.19.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
115
+ "model.layers.19.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
116
+ "model.layers.2.input_layernorm.weight": "model-00001-of-00003.safetensors",
117
+ "model.layers.2.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
118
+ "model.layers.2.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
119
+ "model.layers.2.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
120
+ "model.layers.2.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
121
+ "model.layers.2.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
122
+ "model.layers.2.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
123
+ "model.layers.2.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
124
+ "model.layers.2.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
125
+ "model.layers.20.input_layernorm.weight": "model-00002-of-00003.safetensors",
126
+ "model.layers.20.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
127
+ "model.layers.20.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
128
+ "model.layers.20.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
129
+ "model.layers.20.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
130
+ "model.layers.20.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
131
+ "model.layers.20.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
132
+ "model.layers.20.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
133
+ "model.layers.20.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
134
+ "model.layers.21.input_layernorm.weight": "model-00002-of-00003.safetensors",
135
+ "model.layers.21.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
136
+ "model.layers.21.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
137
+ "model.layers.21.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
138
+ "model.layers.21.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
139
+ "model.layers.21.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
140
+ "model.layers.21.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
141
+ "model.layers.21.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
142
+ "model.layers.21.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
143
+ "model.layers.22.input_layernorm.weight": "model-00003-of-00003.safetensors",
144
+ "model.layers.22.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
145
+ "model.layers.22.mlp.gate_proj.weight": "model-00003-of-00003.safetensors",
146
+ "model.layers.22.mlp.up_proj.weight": "model-00003-of-00003.safetensors",
147
+ "model.layers.22.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
148
+ "model.layers.22.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
149
+ "model.layers.22.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
150
+ "model.layers.22.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
151
+ "model.layers.22.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
152
+ "model.layers.23.input_layernorm.weight": "model-00003-of-00003.safetensors",
153
+ "model.layers.23.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
154
+ "model.layers.23.mlp.gate_proj.weight": "model-00003-of-00003.safetensors",
155
+ "model.layers.23.mlp.up_proj.weight": "model-00003-of-00003.safetensors",
156
+ "model.layers.23.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
157
+ "model.layers.23.self_attn.k_proj.weight": "model-00003-of-00003.safetensors",
158
+ "model.layers.23.self_attn.o_proj.weight": "model-00003-of-00003.safetensors",
159
+ "model.layers.23.self_attn.q_proj.weight": "model-00003-of-00003.safetensors",
160
+ "model.layers.23.self_attn.v_proj.weight": "model-00003-of-00003.safetensors",
161
+ "model.layers.24.input_layernorm.weight": "model-00003-of-00003.safetensors",
162
+ "model.layers.24.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
163
+ "model.layers.24.mlp.gate_proj.weight": "model-00003-of-00003.safetensors",
164
+ "model.layers.24.mlp.up_proj.weight": "model-00003-of-00003.safetensors",
165
+ "model.layers.24.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
166
+ "model.layers.24.self_attn.k_proj.weight": "model-00003-of-00003.safetensors",
167
+ "model.layers.24.self_attn.o_proj.weight": "model-00003-of-00003.safetensors",
168
+ "model.layers.24.self_attn.q_proj.weight": "model-00003-of-00003.safetensors",
169
+ "model.layers.24.self_attn.v_proj.weight": "model-00003-of-00003.safetensors",
170
+ "model.layers.25.input_layernorm.weight": "model-00003-of-00003.safetensors",
171
+ "model.layers.25.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
172
+ "model.layers.25.mlp.gate_proj.weight": "model-00003-of-00003.safetensors",
173
+ "model.layers.25.mlp.up_proj.weight": "model-00003-of-00003.safetensors",
174
+ "model.layers.25.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
175
+ "model.layers.25.self_attn.k_proj.weight": "model-00003-of-00003.safetensors",
176
+ "model.layers.25.self_attn.o_proj.weight": "model-00003-of-00003.safetensors",
177
+ "model.layers.25.self_attn.q_proj.weight": "model-00003-of-00003.safetensors",
178
+ "model.layers.25.self_attn.v_proj.weight": "model-00003-of-00003.safetensors",
179
+ "model.layers.26.input_layernorm.weight": "model-00003-of-00003.safetensors",
180
+ "model.layers.26.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
181
+ "model.layers.26.mlp.gate_proj.weight": "model-00003-of-00003.safetensors",
182
+ "model.layers.26.mlp.up_proj.weight": "model-00003-of-00003.safetensors",
183
+ "model.layers.26.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
184
+ "model.layers.26.self_attn.k_proj.weight": "model-00003-of-00003.safetensors",
185
+ "model.layers.26.self_attn.o_proj.weight": "model-00003-of-00003.safetensors",
186
+ "model.layers.26.self_attn.q_proj.weight": "model-00003-of-00003.safetensors",
187
+ "model.layers.26.self_attn.v_proj.weight": "model-00003-of-00003.safetensors",
188
+ "model.layers.27.input_layernorm.weight": "model-00003-of-00003.safetensors",
189
+ "model.layers.27.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
190
+ "model.layers.27.mlp.gate_proj.weight": "model-00003-of-00003.safetensors",
191
+ "model.layers.27.mlp.up_proj.weight": "model-00003-of-00003.safetensors",
192
+ "model.layers.27.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
193
+ "model.layers.27.self_attn.k_proj.weight": "model-00003-of-00003.safetensors",
194
+ "model.layers.27.self_attn.o_proj.weight": "model-00003-of-00003.safetensors",
195
+ "model.layers.27.self_attn.q_proj.weight": "model-00003-of-00003.safetensors",
196
+ "model.layers.27.self_attn.v_proj.weight": "model-00003-of-00003.safetensors",
197
+ "model.layers.28.input_layernorm.weight": "model-00003-of-00003.safetensors",
198
+ "model.layers.28.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
199
+ "model.layers.28.mlp.gate_proj.weight": "model-00003-of-00003.safetensors",
200
+ "model.layers.28.mlp.up_proj.weight": "model-00003-of-00003.safetensors",
201
+ "model.layers.28.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
202
+ "model.layers.28.self_attn.k_proj.weight": "model-00003-of-00003.safetensors",
203
+ "model.layers.28.self_attn.o_proj.weight": "model-00003-of-00003.safetensors",
204
+ "model.layers.28.self_attn.q_proj.weight": "model-00003-of-00003.safetensors",
205
+ "model.layers.28.self_attn.v_proj.weight": "model-00003-of-00003.safetensors",
206
+ "model.layers.29.input_layernorm.weight": "model-00003-of-00003.safetensors",
207
+ "model.layers.29.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
208
+ "model.layers.29.mlp.gate_proj.weight": "model-00003-of-00003.safetensors",
209
+ "model.layers.29.mlp.up_proj.weight": "model-00003-of-00003.safetensors",
210
+ "model.layers.29.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
211
+ "model.layers.29.self_attn.k_proj.weight": "model-00003-of-00003.safetensors",
212
+ "model.layers.29.self_attn.o_proj.weight": "model-00003-of-00003.safetensors",
213
+ "model.layers.29.self_attn.q_proj.weight": "model-00003-of-00003.safetensors",
214
+ "model.layers.29.self_attn.v_proj.weight": "model-00003-of-00003.safetensors",
215
+ "model.layers.3.input_layernorm.weight": "model-00001-of-00003.safetensors",
216
+ "model.layers.3.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
217
+ "model.layers.3.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
218
+ "model.layers.3.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
219
+ "model.layers.3.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
220
+ "model.layers.3.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
221
+ "model.layers.3.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
222
+ "model.layers.3.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
223
+ "model.layers.3.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
224
+ "model.layers.30.input_layernorm.weight": "model-00003-of-00003.safetensors",
225
+ "model.layers.30.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
226
+ "model.layers.30.mlp.gate_proj.weight": "model-00003-of-00003.safetensors",
227
+ "model.layers.30.mlp.up_proj.weight": "model-00003-of-00003.safetensors",
228
+ "model.layers.30.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
229
+ "model.layers.30.self_attn.k_proj.weight": "model-00003-of-00003.safetensors",
230
+ "model.layers.30.self_attn.o_proj.weight": "model-00003-of-00003.safetensors",
231
+ "model.layers.30.self_attn.q_proj.weight": "model-00003-of-00003.safetensors",
232
+ "model.layers.30.self_attn.v_proj.weight": "model-00003-of-00003.safetensors",
233
+ "model.layers.31.input_layernorm.weight": "model-00003-of-00003.safetensors",
234
+ "model.layers.31.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
235
+ "model.layers.31.mlp.gate_proj.weight": "model-00003-of-00003.safetensors",
236
+ "model.layers.31.mlp.up_proj.weight": "model-00003-of-00003.safetensors",
237
+ "model.layers.31.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
238
+ "model.layers.31.self_attn.k_proj.weight": "model-00003-of-00003.safetensors",
239
+ "model.layers.31.self_attn.o_proj.weight": "model-00003-of-00003.safetensors",
240
+ "model.layers.31.self_attn.q_proj.weight": "model-00003-of-00003.safetensors",
241
+ "model.layers.31.self_attn.v_proj.weight": "model-00003-of-00003.safetensors",
242
+ "model.layers.4.input_layernorm.weight": "model-00001-of-00003.safetensors",
243
+ "model.layers.4.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
244
+ "model.layers.4.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
245
+ "model.layers.4.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
246
+ "model.layers.4.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
247
+ "model.layers.4.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
248
+ "model.layers.4.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
249
+ "model.layers.4.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
250
+ "model.layers.4.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
251
+ "model.layers.5.input_layernorm.weight": "model-00001-of-00003.safetensors",
252
+ "model.layers.5.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
253
+ "model.layers.5.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
254
+ "model.layers.5.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
255
+ "model.layers.5.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
256
+ "model.layers.5.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
257
+ "model.layers.5.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
258
+ "model.layers.5.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
259
+ "model.layers.5.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
260
+ "model.layers.6.input_layernorm.weight": "model-00001-of-00003.safetensors",
261
+ "model.layers.6.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
262
+ "model.layers.6.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
263
+ "model.layers.6.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
264
+ "model.layers.6.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
265
+ "model.layers.6.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
266
+ "model.layers.6.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
267
+ "model.layers.6.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
268
+ "model.layers.6.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
269
+ "model.layers.7.input_layernorm.weight": "model-00001-of-00003.safetensors",
270
+ "model.layers.7.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
271
+ "model.layers.7.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
272
+ "model.layers.7.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
273
+ "model.layers.7.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
274
+ "model.layers.7.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
275
+ "model.layers.7.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
276
+ "model.layers.7.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
277
+ "model.layers.7.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
278
+ "model.layers.8.input_layernorm.weight": "model-00001-of-00003.safetensors",
279
+ "model.layers.8.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
280
+ "model.layers.8.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
281
+ "model.layers.8.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
282
+ "model.layers.8.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
283
+ "model.layers.8.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
284
+ "model.layers.8.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
285
+ "model.layers.8.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
286
+ "model.layers.8.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
287
+ "model.layers.9.input_layernorm.weight": "model-00001-of-00003.safetensors",
288
+ "model.layers.9.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
289
+ "model.layers.9.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
290
+ "model.layers.9.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
291
+ "model.layers.9.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
292
+ "model.layers.9.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
293
+ "model.layers.9.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
294
+ "model.layers.9.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
295
+ "model.layers.9.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
296
+ "model.norm.weight": "model-00003-of-00003.safetensors"
297
+ }
298
+ }
train_results.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 2.0,
3
+ "train_loss": 0.04328188924946038,
4
+ "train_runtime": 16735.9464,
5
+ "train_samples": 19426,
6
+ "train_samples_per_second": 2.321,
7
+ "train_steps_per_second": 0.036
8
+ }
trainer_state.json ADDED
@@ -0,0 +1,1268 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": null,
3
+ "best_model_checkpoint": null,
4
+ "epoch": 2.0,
5
+ "eval_steps": 25,
6
+ "global_step": 608,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 0.0,
13
+ "learning_rate": 8.19672131147541e-09,
14
+ "logits/generated": -3.1101083755493164,
15
+ "logits/real": -3.0836105346679688,
16
+ "logps/generated": -234.708984375,
17
+ "logps/real": -348.7855224609375,
18
+ "loss": 0.6931,
19
+ "rewards/accuracies": 0.0,
20
+ "rewards/generated": 0.0,
21
+ "rewards/margins": 0.0,
22
+ "rewards/real": 0.0,
23
+ "step": 1
24
+ },
25
+ {
26
+ "epoch": 0.03,
27
+ "learning_rate": 8.196721311475409e-08,
28
+ "logits/generated": -3.041771173477173,
29
+ "logits/real": -3.0358164310455322,
30
+ "logps/generated": -259.9490051269531,
31
+ "logps/real": -409.8074951171875,
32
+ "loss": 0.6715,
33
+ "rewards/accuracies": 0.6597222089767456,
34
+ "rewards/generated": -0.009591154754161835,
35
+ "rewards/margins": 0.04917733371257782,
36
+ "rewards/real": 0.039586178958415985,
37
+ "step": 10
38
+ },
39
+ {
40
+ "epoch": 0.07,
41
+ "learning_rate": 1.6393442622950818e-07,
42
+ "logits/generated": -3.025602102279663,
43
+ "logits/real": -3.008653163909912,
44
+ "logps/generated": -291.55126953125,
45
+ "logps/real": -418.98614501953125,
46
+ "loss": 0.4944,
47
+ "rewards/accuracies": 0.9437500238418579,
48
+ "rewards/generated": -0.1885037124156952,
49
+ "rewards/margins": 0.5224509239196777,
50
+ "rewards/real": 0.33394715189933777,
51
+ "step": 20
52
+ },
53
+ {
54
+ "epoch": 0.08,
55
+ "eval_logits/generated": -3.0068743228912354,
56
+ "eval_logits/real": -3.014744520187378,
57
+ "eval_logps/generated": -285.1263732910156,
58
+ "eval_logps/real": -406.0350036621094,
59
+ "eval_loss": 0.2565726339817047,
60
+ "eval_rewards/accuracies": 0.9761029481887817,
61
+ "eval_rewards/generated": -0.8426953554153442,
62
+ "eval_rewards/margins": 1.5071492195129395,
63
+ "eval_rewards/real": 0.6644538044929504,
64
+ "eval_runtime": 272.5132,
65
+ "eval_samples_per_second": 7.923,
66
+ "eval_steps_per_second": 0.25,
67
+ "step": 25
68
+ },
69
+ {
70
+ "epoch": 0.1,
71
+ "learning_rate": 2.4590163934426226e-07,
72
+ "logits/generated": -3.019906997680664,
73
+ "logits/real": -3.0092620849609375,
74
+ "logps/generated": -305.70733642578125,
75
+ "logps/real": -410.2085876464844,
76
+ "loss": 0.2752,
77
+ "rewards/accuracies": 0.949999988079071,
78
+ "rewards/generated": -0.7856386303901672,
79
+ "rewards/margins": 1.4234931468963623,
80
+ "rewards/real": 0.6378546357154846,
81
+ "step": 30
82
+ },
83
+ {
84
+ "epoch": 0.13,
85
+ "learning_rate": 3.2786885245901637e-07,
86
+ "logits/generated": -2.9752182960510254,
87
+ "logits/real": -2.9877312183380127,
88
+ "logps/generated": -290.79388427734375,
89
+ "logps/real": -412.19482421875,
90
+ "loss": 0.1601,
91
+ "rewards/accuracies": 1.0,
92
+ "rewards/generated": -1.8102967739105225,
93
+ "rewards/margins": 2.52372407913208,
94
+ "rewards/real": 0.7134272456169128,
95
+ "step": 40
96
+ },
97
+ {
98
+ "epoch": 0.16,
99
+ "learning_rate": 4.0983606557377047e-07,
100
+ "logits/generated": -2.939542531967163,
101
+ "logits/real": -2.9472873210906982,
102
+ "logps/generated": -315.44427490234375,
103
+ "logps/real": -429.36614990234375,
104
+ "loss": 0.092,
105
+ "rewards/accuracies": 0.9937499761581421,
106
+ "rewards/generated": -3.036115884780884,
107
+ "rewards/margins": 3.7522292137145996,
108
+ "rewards/real": 0.7161135077476501,
109
+ "step": 50
110
+ },
111
+ {
112
+ "epoch": 0.16,
113
+ "eval_logits/generated": -2.9427242279052734,
114
+ "eval_logits/real": -2.955735921859741,
115
+ "eval_logps/generated": -314.470458984375,
116
+ "eval_logps/real": -408.6964111328125,
117
+ "eval_loss": 0.08381339907646179,
118
+ "eval_rewards/accuracies": 0.9889705777168274,
119
+ "eval_rewards/generated": -3.777100086212158,
120
+ "eval_rewards/margins": 4.175417900085449,
121
+ "eval_rewards/real": 0.39831778407096863,
122
+ "eval_runtime": 271.2109,
123
+ "eval_samples_per_second": 7.961,
124
+ "eval_steps_per_second": 0.251,
125
+ "step": 50
126
+ },
127
+ {
128
+ "epoch": 0.2,
129
+ "learning_rate": 4.918032786885245e-07,
130
+ "logits/generated": -2.946568250656128,
131
+ "logits/real": -2.9412360191345215,
132
+ "logps/generated": -343.0057678222656,
133
+ "logps/real": -417.2606506347656,
134
+ "loss": 0.0779,
135
+ "rewards/accuracies": 0.9750000238418579,
136
+ "rewards/generated": -4.043205261230469,
137
+ "rewards/margins": 4.313161849975586,
138
+ "rewards/real": 0.26995649933815,
139
+ "step": 60
140
+ },
141
+ {
142
+ "epoch": 0.23,
143
+ "learning_rate": 4.917733089579524e-07,
144
+ "logits/generated": -2.9332735538482666,
145
+ "logits/real": -2.9507241249084473,
146
+ "logps/generated": -310.9266662597656,
147
+ "logps/real": -392.91436767578125,
148
+ "loss": 0.0601,
149
+ "rewards/accuracies": 0.987500011920929,
150
+ "rewards/generated": -4.882044792175293,
151
+ "rewards/margins": 4.866286277770996,
152
+ "rewards/real": -0.015758510679006577,
153
+ "step": 70
154
+ },
155
+ {
156
+ "epoch": 0.25,
157
+ "eval_logits/generated": -2.9205291271209717,
158
+ "eval_logits/real": -2.9325790405273438,
159
+ "eval_logps/generated": -333.0871276855469,
160
+ "eval_logps/real": -410.1153564453125,
161
+ "eval_loss": 0.04572594165802002,
162
+ "eval_rewards/accuracies": 0.9963235259056091,
163
+ "eval_rewards/generated": -5.638766288757324,
164
+ "eval_rewards/margins": 5.895185470581055,
165
+ "eval_rewards/real": 0.2564197778701782,
166
+ "eval_runtime": 279.8658,
167
+ "eval_samples_per_second": 7.714,
168
+ "eval_steps_per_second": 0.243,
169
+ "step": 75
170
+ },
171
+ {
172
+ "epoch": 0.26,
173
+ "learning_rate": 4.826325411334552e-07,
174
+ "logits/generated": -2.923539400100708,
175
+ "logits/real": -2.924189805984497,
176
+ "logps/generated": -344.7599792480469,
177
+ "logps/real": -426.96954345703125,
178
+ "loss": 0.037,
179
+ "rewards/accuracies": 0.9937499761581421,
180
+ "rewards/generated": -5.42771053314209,
181
+ "rewards/margins": 5.462244987487793,
182
+ "rewards/real": 0.03453410416841507,
183
+ "step": 80
184
+ },
185
+ {
186
+ "epoch": 0.3,
187
+ "learning_rate": 4.7349177330895793e-07,
188
+ "logits/generated": -2.9185070991516113,
189
+ "logits/real": -2.923430919647217,
190
+ "logps/generated": -323.52618408203125,
191
+ "logps/real": -421.53558349609375,
192
+ "loss": 0.0381,
193
+ "rewards/accuracies": 0.9937499761581421,
194
+ "rewards/generated": -6.292209625244141,
195
+ "rewards/margins": 6.230466365814209,
196
+ "rewards/real": -0.0617423951625824,
197
+ "step": 90
198
+ },
199
+ {
200
+ "epoch": 0.33,
201
+ "learning_rate": 4.6435100548446064e-07,
202
+ "logits/generated": -2.940284490585327,
203
+ "logits/real": -2.9340343475341797,
204
+ "logps/generated": -348.3626708984375,
205
+ "logps/real": -378.1487731933594,
206
+ "loss": 0.0437,
207
+ "rewards/accuracies": 0.987500011920929,
208
+ "rewards/generated": -6.817117214202881,
209
+ "rewards/margins": 6.638771057128906,
210
+ "rewards/real": -0.17834754288196564,
211
+ "step": 100
212
+ },
213
+ {
214
+ "epoch": 0.33,
215
+ "eval_logits/generated": -2.888258218765259,
216
+ "eval_logits/real": -2.906193733215332,
217
+ "eval_logps/generated": -349.15032958984375,
218
+ "eval_logps/real": -414.5328369140625,
219
+ "eval_loss": 0.03355773165822029,
220
+ "eval_rewards/accuracies": 0.9963235259056091,
221
+ "eval_rewards/generated": -7.245093822479248,
222
+ "eval_rewards/margins": 7.059763431549072,
223
+ "eval_rewards/real": -0.18532894551753998,
224
+ "eval_runtime": 281.2041,
225
+ "eval_samples_per_second": 7.678,
226
+ "eval_steps_per_second": 0.242,
227
+ "step": 100
228
+ },
229
+ {
230
+ "epoch": 0.36,
231
+ "learning_rate": 4.552102376599634e-07,
232
+ "logits/generated": -2.912867307662964,
233
+ "logits/real": -2.9122469425201416,
234
+ "logps/generated": -353.8576965332031,
235
+ "logps/real": -416.38262939453125,
236
+ "loss": 0.0385,
237
+ "rewards/accuracies": 0.987500011920929,
238
+ "rewards/generated": -7.3175201416015625,
239
+ "rewards/margins": 6.935573577880859,
240
+ "rewards/real": -0.38194605708122253,
241
+ "step": 110
242
+ },
243
+ {
244
+ "epoch": 0.39,
245
+ "learning_rate": 4.460694698354662e-07,
246
+ "logits/generated": -2.916935920715332,
247
+ "logits/real": -2.90993070602417,
248
+ "logps/generated": -347.873046875,
249
+ "logps/real": -423.1348571777344,
250
+ "loss": 0.036,
251
+ "rewards/accuracies": 0.981249988079071,
252
+ "rewards/generated": -7.8421173095703125,
253
+ "rewards/margins": 7.608695030212402,
254
+ "rewards/real": -0.23342163860797882,
255
+ "step": 120
256
+ },
257
+ {
258
+ "epoch": 0.41,
259
+ "eval_logits/generated": -2.881693124771118,
260
+ "eval_logits/real": -2.9014151096343994,
261
+ "eval_logps/generated": -354.1070556640625,
262
+ "eval_logps/real": -414.3309326171875,
263
+ "eval_loss": 0.027073556557297707,
264
+ "eval_rewards/accuracies": 0.9944853186607361,
265
+ "eval_rewards/generated": -7.7407612800598145,
266
+ "eval_rewards/margins": 7.575623512268066,
267
+ "eval_rewards/real": -0.16513808071613312,
268
+ "eval_runtime": 282.9385,
269
+ "eval_samples_per_second": 7.631,
270
+ "eval_steps_per_second": 0.24,
271
+ "step": 125
272
+ },
273
+ {
274
+ "epoch": 0.43,
275
+ "learning_rate": 4.3692870201096885e-07,
276
+ "logits/generated": -2.886507272720337,
277
+ "logits/real": -2.8951663970947266,
278
+ "logps/generated": -376.47540283203125,
279
+ "logps/real": -435.31024169921875,
280
+ "loss": 0.0309,
281
+ "rewards/accuracies": 0.9937499761581421,
282
+ "rewards/generated": -8.095990180969238,
283
+ "rewards/margins": 7.625959873199463,
284
+ "rewards/real": -0.47002920508384705,
285
+ "step": 130
286
+ },
287
+ {
288
+ "epoch": 0.46,
289
+ "learning_rate": 4.277879341864716e-07,
290
+ "logits/generated": -2.866490602493286,
291
+ "logits/real": -2.878412961959839,
292
+ "logps/generated": -375.51739501953125,
293
+ "logps/real": -430.4481506347656,
294
+ "loss": 0.0318,
295
+ "rewards/accuracies": 0.981249988079071,
296
+ "rewards/generated": -8.508157730102539,
297
+ "rewards/margins": 8.17884635925293,
298
+ "rewards/real": -0.32931095361709595,
299
+ "step": 140
300
+ },
301
+ {
302
+ "epoch": 0.49,
303
+ "learning_rate": 4.186471663619744e-07,
304
+ "logits/generated": -2.8697776794433594,
305
+ "logits/real": -2.87200927734375,
306
+ "logps/generated": -344.39154052734375,
307
+ "logps/real": -417.9952697753906,
308
+ "loss": 0.0373,
309
+ "rewards/accuracies": 0.987500011920929,
310
+ "rewards/generated": -8.700019836425781,
311
+ "rewards/margins": 8.087128639221191,
312
+ "rewards/real": -0.6128913760185242,
313
+ "step": 150
314
+ },
315
+ {
316
+ "epoch": 0.49,
317
+ "eval_logits/generated": -2.827141761779785,
318
+ "eval_logits/real": -2.8543221950531006,
319
+ "eval_logps/generated": -355.01171875,
320
+ "eval_logps/real": -415.0633850097656,
321
+ "eval_loss": 0.026445312425494194,
322
+ "eval_rewards/accuracies": 0.9908088445663452,
323
+ "eval_rewards/generated": -7.831226825714111,
324
+ "eval_rewards/margins": 7.5928425788879395,
325
+ "eval_rewards/real": -0.23838478326797485,
326
+ "eval_runtime": 279.7973,
327
+ "eval_samples_per_second": 7.716,
328
+ "eval_steps_per_second": 0.243,
329
+ "step": 150
330
+ },
331
+ {
332
+ "epoch": 0.53,
333
+ "learning_rate": 4.0950639853747716e-07,
334
+ "logits/generated": -2.832282543182373,
335
+ "logits/real": -2.8597447872161865,
336
+ "logps/generated": -383.411376953125,
337
+ "logps/real": -424.3670959472656,
338
+ "loss": 0.0349,
339
+ "rewards/accuracies": 1.0,
340
+ "rewards/generated": -8.395051002502441,
341
+ "rewards/margins": 8.07539176940918,
342
+ "rewards/real": -0.319659024477005,
343
+ "step": 160
344
+ },
345
+ {
346
+ "epoch": 0.56,
347
+ "learning_rate": 4.0036563071297983e-07,
348
+ "logits/generated": -2.811795234680176,
349
+ "logits/real": -2.841276168823242,
350
+ "logps/generated": -379.37359619140625,
351
+ "logps/real": -428.99407958984375,
352
+ "loss": 0.0198,
353
+ "rewards/accuracies": 1.0,
354
+ "rewards/generated": -8.731819152832031,
355
+ "rewards/margins": 8.355070114135742,
356
+ "rewards/real": -0.37674885988235474,
357
+ "step": 170
358
+ },
359
+ {
360
+ "epoch": 0.58,
361
+ "eval_logits/generated": -2.8051629066467285,
362
+ "eval_logits/real": -2.832581043243408,
363
+ "eval_logps/generated": -376.1680603027344,
364
+ "eval_logps/real": -421.83154296875,
365
+ "eval_loss": 0.02137557417154312,
366
+ "eval_rewards/accuracies": 0.9908088445663452,
367
+ "eval_rewards/generated": -9.94686222076416,
368
+ "eval_rewards/margins": 9.031662940979004,
369
+ "eval_rewards/real": -0.9151991009712219,
370
+ "eval_runtime": 281.297,
371
+ "eval_samples_per_second": 7.675,
372
+ "eval_steps_per_second": 0.242,
373
+ "step": 175
374
+ },
375
+ {
376
+ "epoch": 0.59,
377
+ "learning_rate": 3.912248628884826e-07,
378
+ "logits/generated": -2.8207595348358154,
379
+ "logits/real": -2.8399100303649902,
380
+ "logps/generated": -393.3479919433594,
381
+ "logps/real": -451.2169494628906,
382
+ "loss": 0.022,
383
+ "rewards/accuracies": 0.9937499761581421,
384
+ "rewards/generated": -9.940328598022461,
385
+ "rewards/margins": 9.013021469116211,
386
+ "rewards/real": -0.92730712890625,
387
+ "step": 180
388
+ },
389
+ {
390
+ "epoch": 0.62,
391
+ "learning_rate": 3.8208409506398537e-07,
392
+ "logits/generated": -2.8636739253997803,
393
+ "logits/real": -2.867056369781494,
394
+ "logps/generated": -404.6825256347656,
395
+ "logps/real": -445.5357971191406,
396
+ "loss": 0.0189,
397
+ "rewards/accuracies": 1.0,
398
+ "rewards/generated": -10.549575805664062,
399
+ "rewards/margins": 9.651226043701172,
400
+ "rewards/real": -0.8983511924743652,
401
+ "step": 190
402
+ },
403
+ {
404
+ "epoch": 0.66,
405
+ "learning_rate": 3.7294332723948814e-07,
406
+ "logits/generated": -2.8503365516662598,
407
+ "logits/real": -2.8629379272460938,
408
+ "logps/generated": -380.40264892578125,
409
+ "logps/real": -451.1651916503906,
410
+ "loss": 0.0426,
411
+ "rewards/accuracies": 0.9750000238418579,
412
+ "rewards/generated": -10.779017448425293,
413
+ "rewards/margins": 9.16708755493164,
414
+ "rewards/real": -1.6119304895401,
415
+ "step": 200
416
+ },
417
+ {
418
+ "epoch": 0.66,
419
+ "eval_logits/generated": -2.8450069427490234,
420
+ "eval_logits/real": -2.85876727104187,
421
+ "eval_logps/generated": -367.72100830078125,
422
+ "eval_logps/real": -422.42657470703125,
423
+ "eval_loss": 0.02509322017431259,
424
+ "eval_rewards/accuracies": 0.9908088445663452,
425
+ "eval_rewards/generated": -9.10215950012207,
426
+ "eval_rewards/margins": 8.127457618713379,
427
+ "eval_rewards/real": -0.9747023582458496,
428
+ "eval_runtime": 279.5471,
429
+ "eval_samples_per_second": 7.723,
430
+ "eval_steps_per_second": 0.243,
431
+ "step": 200
432
+ },
433
+ {
434
+ "epoch": 0.69,
435
+ "learning_rate": 3.638025594149908e-07,
436
+ "logits/generated": -2.840203285217285,
437
+ "logits/real": -2.83565354347229,
438
+ "logps/generated": -398.70989990234375,
439
+ "logps/real": -414.3641662597656,
440
+ "loss": 0.0231,
441
+ "rewards/accuracies": 1.0,
442
+ "rewards/generated": -9.996566772460938,
443
+ "rewards/margins": 8.856243133544922,
444
+ "rewards/real": -1.1403248310089111,
445
+ "step": 210
446
+ },
447
+ {
448
+ "epoch": 0.72,
449
+ "learning_rate": 3.5466179159049357e-07,
450
+ "logits/generated": -2.82039213180542,
451
+ "logits/real": -2.8223772048950195,
452
+ "logps/generated": -382.1441955566406,
453
+ "logps/real": -446.4202575683594,
454
+ "loss": 0.0262,
455
+ "rewards/accuracies": 0.9624999761581421,
456
+ "rewards/generated": -9.665162086486816,
457
+ "rewards/margins": 8.764172554016113,
458
+ "rewards/real": -0.9009901881217957,
459
+ "step": 220
460
+ },
461
+ {
462
+ "epoch": 0.74,
463
+ "eval_logits/generated": -2.800882339477539,
464
+ "eval_logits/real": -2.8208842277526855,
465
+ "eval_logps/generated": -376.0171813964844,
466
+ "eval_logps/real": -421.0940246582031,
467
+ "eval_loss": 0.018937036395072937,
468
+ "eval_rewards/accuracies": 0.9926470518112183,
469
+ "eval_rewards/generated": -9.93177318572998,
470
+ "eval_rewards/margins": 9.090324401855469,
471
+ "eval_rewards/real": -0.8414493203163147,
472
+ "eval_runtime": 281.9616,
473
+ "eval_samples_per_second": 7.657,
474
+ "eval_steps_per_second": 0.241,
475
+ "step": 225
476
+ },
477
+ {
478
+ "epoch": 0.76,
479
+ "learning_rate": 3.4552102376599634e-07,
480
+ "logits/generated": -2.808319568634033,
481
+ "logits/real": -2.802628993988037,
482
+ "logps/generated": -363.14227294921875,
483
+ "logps/real": -395.32305908203125,
484
+ "loss": 0.0226,
485
+ "rewards/accuracies": 0.987500011920929,
486
+ "rewards/generated": -9.69061279296875,
487
+ "rewards/margins": 8.589911460876465,
488
+ "rewards/real": -1.1007012128829956,
489
+ "step": 230
490
+ },
491
+ {
492
+ "epoch": 0.79,
493
+ "learning_rate": 3.3638025594149906e-07,
494
+ "logits/generated": -2.8294272422790527,
495
+ "logits/real": -2.8311877250671387,
496
+ "logps/generated": -398.45635986328125,
497
+ "logps/real": -404.5653991699219,
498
+ "loss": 0.0297,
499
+ "rewards/accuracies": 0.987500011920929,
500
+ "rewards/generated": -9.038873672485352,
501
+ "rewards/margins": 8.15873908996582,
502
+ "rewards/real": -0.8801354169845581,
503
+ "step": 240
504
+ },
505
+ {
506
+ "epoch": 0.82,
507
+ "learning_rate": 3.272394881170018e-07,
508
+ "logits/generated": -2.816089153289795,
509
+ "logits/real": -2.839442729949951,
510
+ "logps/generated": -382.88311767578125,
511
+ "logps/real": -406.73297119140625,
512
+ "loss": 0.0142,
513
+ "rewards/accuracies": 1.0,
514
+ "rewards/generated": -9.536883354187012,
515
+ "rewards/margins": 8.913406372070312,
516
+ "rewards/real": -0.6234776973724365,
517
+ "step": 250
518
+ },
519
+ {
520
+ "epoch": 0.82,
521
+ "eval_logits/generated": -2.7973484992980957,
522
+ "eval_logits/real": -2.8200957775115967,
523
+ "eval_logps/generated": -377.7586364746094,
524
+ "eval_logps/real": -419.83355712890625,
525
+ "eval_loss": 0.01662140153348446,
526
+ "eval_rewards/accuracies": 0.9944853186607361,
527
+ "eval_rewards/generated": -10.105918884277344,
528
+ "eval_rewards/margins": 9.390520095825195,
529
+ "eval_rewards/real": -0.7153997421264648,
530
+ "eval_runtime": 282.5094,
531
+ "eval_samples_per_second": 7.642,
532
+ "eval_steps_per_second": 0.241,
533
+ "step": 250
534
+ },
535
+ {
536
+ "epoch": 0.86,
537
+ "learning_rate": 3.1809872029250455e-07,
538
+ "logits/generated": -2.802105665206909,
539
+ "logits/real": -2.824450731277466,
540
+ "logps/generated": -371.94012451171875,
541
+ "logps/real": -423.9705505371094,
542
+ "loss": 0.0143,
543
+ "rewards/accuracies": 0.987500011920929,
544
+ "rewards/generated": -10.57664680480957,
545
+ "rewards/margins": 9.595219612121582,
546
+ "rewards/real": -0.9814273715019226,
547
+ "step": 260
548
+ },
549
+ {
550
+ "epoch": 0.89,
551
+ "learning_rate": 3.089579524680073e-07,
552
+ "logits/generated": -2.7837929725646973,
553
+ "logits/real": -2.7973639965057373,
554
+ "logps/generated": -388.8419189453125,
555
+ "logps/real": -429.37054443359375,
556
+ "loss": 0.0171,
557
+ "rewards/accuracies": 1.0,
558
+ "rewards/generated": -10.401695251464844,
559
+ "rewards/margins": 9.380131721496582,
560
+ "rewards/real": -1.021564245223999,
561
+ "step": 270
562
+ },
563
+ {
564
+ "epoch": 0.9,
565
+ "eval_logits/generated": -2.7641427516937256,
566
+ "eval_logits/real": -2.7935867309570312,
567
+ "eval_logps/generated": -385.756103515625,
568
+ "eval_logps/real": -423.5848693847656,
569
+ "eval_loss": 0.018939225003123283,
570
+ "eval_rewards/accuracies": 0.9944853186607361,
571
+ "eval_rewards/generated": -10.90566635131836,
572
+ "eval_rewards/margins": 9.815133094787598,
573
+ "eval_rewards/real": -1.0905327796936035,
574
+ "eval_runtime": 281.011,
575
+ "eval_samples_per_second": 7.683,
576
+ "eval_steps_per_second": 0.242,
577
+ "step": 275
578
+ },
579
+ {
580
+ "epoch": 0.92,
581
+ "learning_rate": 2.9981718464351004e-07,
582
+ "logits/generated": -2.762982130050659,
583
+ "logits/real": -2.7803285121917725,
584
+ "logps/generated": -370.44073486328125,
585
+ "logps/real": -432.51708984375,
586
+ "loss": 0.025,
587
+ "rewards/accuracies": 0.987500011920929,
588
+ "rewards/generated": -10.256260871887207,
589
+ "rewards/margins": 9.053407669067383,
590
+ "rewards/real": -1.202852487564087,
591
+ "step": 280
592
+ },
593
+ {
594
+ "epoch": 0.95,
595
+ "learning_rate": 2.906764168190128e-07,
596
+ "logits/generated": -2.7955307960510254,
597
+ "logits/real": -2.818641185760498,
598
+ "logps/generated": -406.474853515625,
599
+ "logps/real": -446.6991271972656,
600
+ "loss": 0.0378,
601
+ "rewards/accuracies": 0.9937499761581421,
602
+ "rewards/generated": -10.223650932312012,
603
+ "rewards/margins": 8.97282600402832,
604
+ "rewards/real": -1.250824213027954,
605
+ "step": 290
606
+ },
607
+ {
608
+ "epoch": 0.99,
609
+ "learning_rate": 2.815356489945155e-07,
610
+ "logits/generated": -2.7767701148986816,
611
+ "logits/real": -2.79323410987854,
612
+ "logps/generated": -376.298828125,
613
+ "logps/real": -418.32611083984375,
614
+ "loss": 0.0333,
615
+ "rewards/accuracies": 0.981249988079071,
616
+ "rewards/generated": -10.532678604125977,
617
+ "rewards/margins": 9.659395217895508,
618
+ "rewards/real": -0.8732837438583374,
619
+ "step": 300
620
+ },
621
+ {
622
+ "epoch": 0.99,
623
+ "eval_logits/generated": -2.7972822189331055,
624
+ "eval_logits/real": -2.8230373859405518,
625
+ "eval_logps/generated": -391.56549072265625,
626
+ "eval_logps/real": -425.4764709472656,
627
+ "eval_loss": 0.01681051403284073,
628
+ "eval_rewards/accuracies": 0.9963235259056091,
629
+ "eval_rewards/generated": -11.48660659790039,
630
+ "eval_rewards/margins": 10.206913948059082,
631
+ "eval_rewards/real": -1.2796927690505981,
632
+ "eval_runtime": 282.0432,
633
+ "eval_samples_per_second": 7.655,
634
+ "eval_steps_per_second": 0.241,
635
+ "step": 300
636
+ },
637
+ {
638
+ "epoch": 1.02,
639
+ "learning_rate": 2.7239488117001824e-07,
640
+ "logits/generated": -2.804527997970581,
641
+ "logits/real": -2.8110270500183105,
642
+ "logps/generated": -392.09259033203125,
643
+ "logps/real": -416.8059997558594,
644
+ "loss": 0.0115,
645
+ "rewards/accuracies": 0.9937499761581421,
646
+ "rewards/generated": -12.317480087280273,
647
+ "rewards/margins": 10.796090126037598,
648
+ "rewards/real": -1.521390676498413,
649
+ "step": 310
650
+ },
651
+ {
652
+ "epoch": 1.05,
653
+ "learning_rate": 2.63254113345521e-07,
654
+ "logits/generated": -2.7929205894470215,
655
+ "logits/real": -2.8038601875305176,
656
+ "logps/generated": -406.0953063964844,
657
+ "logps/real": -417.6265563964844,
658
+ "loss": 0.0061,
659
+ "rewards/accuracies": 1.0,
660
+ "rewards/generated": -12.046645164489746,
661
+ "rewards/margins": 11.529437065124512,
662
+ "rewards/real": -0.5172096490859985,
663
+ "step": 320
664
+ },
665
+ {
666
+ "epoch": 1.07,
667
+ "eval_logits/generated": -2.797431468963623,
668
+ "eval_logits/real": -2.82306170463562,
669
+ "eval_logps/generated": -388.5797424316406,
670
+ "eval_logps/real": -424.7586669921875,
671
+ "eval_loss": 0.015748905017971992,
672
+ "eval_rewards/accuracies": 0.9944853186607361,
673
+ "eval_rewards/generated": -11.188030242919922,
674
+ "eval_rewards/margins": 9.980118751525879,
675
+ "eval_rewards/real": -1.2079119682312012,
676
+ "eval_runtime": 274.6049,
677
+ "eval_samples_per_second": 7.862,
678
+ "eval_steps_per_second": 0.248,
679
+ "step": 325
680
+ },
681
+ {
682
+ "epoch": 1.09,
683
+ "learning_rate": 2.541133455210238e-07,
684
+ "logits/generated": -2.783512592315674,
685
+ "logits/real": -2.7885773181915283,
686
+ "logps/generated": -410.649658203125,
687
+ "logps/real": -451.8507385253906,
688
+ "loss": 0.0042,
689
+ "rewards/accuracies": 1.0,
690
+ "rewards/generated": -11.628830909729004,
691
+ "rewards/margins": 10.77421760559082,
692
+ "rewards/real": -0.854611873626709,
693
+ "step": 330
694
+ },
695
+ {
696
+ "epoch": 1.12,
697
+ "learning_rate": 2.449725776965265e-07,
698
+ "logits/generated": -2.8123326301574707,
699
+ "logits/real": -2.8071532249450684,
700
+ "logps/generated": -417.04833984375,
701
+ "logps/real": -433.6940002441406,
702
+ "loss": 0.0074,
703
+ "rewards/accuracies": 1.0,
704
+ "rewards/generated": -11.209399223327637,
705
+ "rewards/margins": 10.593694686889648,
706
+ "rewards/real": -0.615705132484436,
707
+ "step": 340
708
+ },
709
+ {
710
+ "epoch": 1.15,
711
+ "learning_rate": 2.3583180987202925e-07,
712
+ "logits/generated": -2.80556058883667,
713
+ "logits/real": -2.816408157348633,
714
+ "logps/generated": -402.67120361328125,
715
+ "logps/real": -396.2184753417969,
716
+ "loss": 0.0022,
717
+ "rewards/accuracies": 1.0,
718
+ "rewards/generated": -10.71699047088623,
719
+ "rewards/margins": 10.27141284942627,
720
+ "rewards/real": -0.4455786347389221,
721
+ "step": 350
722
+ },
723
+ {
724
+ "epoch": 1.15,
725
+ "eval_logits/generated": -2.7852790355682373,
726
+ "eval_logits/real": -2.8128247261047363,
727
+ "eval_logps/generated": -389.1376037597656,
728
+ "eval_logps/real": -423.37457275390625,
729
+ "eval_loss": 0.015182293951511383,
730
+ "eval_rewards/accuracies": 0.9908088445663452,
731
+ "eval_rewards/generated": -11.243818283081055,
732
+ "eval_rewards/margins": 10.17431640625,
733
+ "eval_rewards/real": -1.0695013999938965,
734
+ "eval_runtime": 277.2781,
735
+ "eval_samples_per_second": 7.786,
736
+ "eval_steps_per_second": 0.245,
737
+ "step": 350
738
+ },
739
+ {
740
+ "epoch": 1.18,
741
+ "learning_rate": 2.26691042047532e-07,
742
+ "logits/generated": -2.795651912689209,
743
+ "logits/real": -2.7973437309265137,
744
+ "logps/generated": -416.8365783691406,
745
+ "logps/real": -424.8717346191406,
746
+ "loss": 0.0031,
747
+ "rewards/accuracies": 1.0,
748
+ "rewards/generated": -11.76781940460205,
749
+ "rewards/margins": 10.914695739746094,
750
+ "rewards/real": -0.853122889995575,
751
+ "step": 360
752
+ },
753
+ {
754
+ "epoch": 1.22,
755
+ "learning_rate": 2.1755027422303473e-07,
756
+ "logits/generated": -2.8003101348876953,
757
+ "logits/real": -2.8173794746398926,
758
+ "logps/generated": -425.969482421875,
759
+ "logps/real": -427.357421875,
760
+ "loss": 0.0033,
761
+ "rewards/accuracies": 1.0,
762
+ "rewards/generated": -12.280854225158691,
763
+ "rewards/margins": 11.536186218261719,
764
+ "rewards/real": -0.7446659803390503,
765
+ "step": 370
766
+ },
767
+ {
768
+ "epoch": 1.23,
769
+ "eval_logits/generated": -2.7750558853149414,
770
+ "eval_logits/real": -2.8028876781463623,
771
+ "eval_logps/generated": -393.3175048828125,
772
+ "eval_logps/real": -424.44647216796875,
773
+ "eval_loss": 0.014840944670140743,
774
+ "eval_rewards/accuracies": 0.9908088445663452,
775
+ "eval_rewards/generated": -11.66180419921875,
776
+ "eval_rewards/margins": 10.485115051269531,
777
+ "eval_rewards/real": -1.1766891479492188,
778
+ "eval_runtime": 278.3753,
779
+ "eval_samples_per_second": 7.756,
780
+ "eval_steps_per_second": 0.244,
781
+ "step": 375
782
+ },
783
+ {
784
+ "epoch": 1.25,
785
+ "learning_rate": 2.0840950639853748e-07,
786
+ "logits/generated": -2.779252529144287,
787
+ "logits/real": -2.804440975189209,
788
+ "logps/generated": -388.8800354003906,
789
+ "logps/real": -422.3763122558594,
790
+ "loss": 0.0019,
791
+ "rewards/accuracies": 1.0,
792
+ "rewards/generated": -12.17228889465332,
793
+ "rewards/margins": 11.443912506103516,
794
+ "rewards/real": -0.7283763289451599,
795
+ "step": 380
796
+ },
797
+ {
798
+ "epoch": 1.28,
799
+ "learning_rate": 1.9926873857404022e-07,
800
+ "logits/generated": -2.7903225421905518,
801
+ "logits/real": -2.7983546257019043,
802
+ "logps/generated": -407.7205810546875,
803
+ "logps/real": -410.11627197265625,
804
+ "loss": 0.0064,
805
+ "rewards/accuracies": 1.0,
806
+ "rewards/generated": -11.748414993286133,
807
+ "rewards/margins": 11.07917594909668,
808
+ "rewards/real": -0.6692394614219666,
809
+ "step": 390
810
+ },
811
+ {
812
+ "epoch": 1.32,
813
+ "learning_rate": 1.9012797074954297e-07,
814
+ "logits/generated": -2.756133794784546,
815
+ "logits/real": -2.773040771484375,
816
+ "logps/generated": -429.03436279296875,
817
+ "logps/real": -455.53497314453125,
818
+ "loss": 0.0043,
819
+ "rewards/accuracies": 1.0,
820
+ "rewards/generated": -12.553096771240234,
821
+ "rewards/margins": 12.007298469543457,
822
+ "rewards/real": -0.5457991361618042,
823
+ "step": 400
824
+ },
825
+ {
826
+ "epoch": 1.32,
827
+ "eval_logits/generated": -2.77026104927063,
828
+ "eval_logits/real": -2.7975828647613525,
829
+ "eval_logps/generated": -395.00494384765625,
830
+ "eval_logps/real": -423.63067626953125,
831
+ "eval_loss": 0.013821952044963837,
832
+ "eval_rewards/accuracies": 0.9963235259056091,
833
+ "eval_rewards/generated": -11.830552101135254,
834
+ "eval_rewards/margins": 10.735437393188477,
835
+ "eval_rewards/real": -1.095115065574646,
836
+ "eval_runtime": 280.6653,
837
+ "eval_samples_per_second": 7.692,
838
+ "eval_steps_per_second": 0.242,
839
+ "step": 400
840
+ },
841
+ {
842
+ "epoch": 1.35,
843
+ "learning_rate": 1.8098720292504568e-07,
844
+ "logits/generated": -2.7702574729919434,
845
+ "logits/real": -2.804114818572998,
846
+ "logps/generated": -426.2373046875,
847
+ "logps/real": -466.58599853515625,
848
+ "loss": 0.004,
849
+ "rewards/accuracies": 1.0,
850
+ "rewards/generated": -12.460441589355469,
851
+ "rewards/margins": 11.789602279663086,
852
+ "rewards/real": -0.6708400249481201,
853
+ "step": 410
854
+ },
855
+ {
856
+ "epoch": 1.38,
857
+ "learning_rate": 1.7184643510054845e-07,
858
+ "logits/generated": -2.7784767150878906,
859
+ "logits/real": -2.801504373550415,
860
+ "logps/generated": -416.26678466796875,
861
+ "logps/real": -468.1026916503906,
862
+ "loss": 0.005,
863
+ "rewards/accuracies": 1.0,
864
+ "rewards/generated": -12.879318237304688,
865
+ "rewards/margins": 11.978571891784668,
866
+ "rewards/real": -0.9007464647293091,
867
+ "step": 420
868
+ },
869
+ {
870
+ "epoch": 1.4,
871
+ "eval_logits/generated": -2.7573304176330566,
872
+ "eval_logits/real": -2.785118818283081,
873
+ "eval_logps/generated": -401.373291015625,
874
+ "eval_logps/real": -425.8589172363281,
875
+ "eval_loss": 0.013551454059779644,
876
+ "eval_rewards/accuracies": 0.9963235259056091,
877
+ "eval_rewards/generated": -12.467382431030273,
878
+ "eval_rewards/margins": 11.149442672729492,
879
+ "eval_rewards/real": -1.3179404735565186,
880
+ "eval_runtime": 278.2775,
881
+ "eval_samples_per_second": 7.758,
882
+ "eval_steps_per_second": 0.244,
883
+ "step": 425
884
+ },
885
+ {
886
+ "epoch": 1.41,
887
+ "learning_rate": 1.6270566727605117e-07,
888
+ "logits/generated": -2.7749133110046387,
889
+ "logits/real": -2.795462131500244,
890
+ "logps/generated": -417.4508361816406,
891
+ "logps/real": -445.52728271484375,
892
+ "loss": 0.0046,
893
+ "rewards/accuracies": 1.0,
894
+ "rewards/generated": -12.883870124816895,
895
+ "rewards/margins": 11.809028625488281,
896
+ "rewards/real": -1.0748413801193237,
897
+ "step": 430
898
+ },
899
+ {
900
+ "epoch": 1.45,
901
+ "learning_rate": 1.5356489945155394e-07,
902
+ "logits/generated": -2.7696471214294434,
903
+ "logits/real": -2.778109312057495,
904
+ "logps/generated": -392.8078918457031,
905
+ "logps/real": -403.84417724609375,
906
+ "loss": 0.0017,
907
+ "rewards/accuracies": 1.0,
908
+ "rewards/generated": -12.799501419067383,
909
+ "rewards/margins": 12.152459144592285,
910
+ "rewards/real": -0.6470428705215454,
911
+ "step": 440
912
+ },
913
+ {
914
+ "epoch": 1.48,
915
+ "learning_rate": 1.4442413162705666e-07,
916
+ "logits/generated": -2.7640178203582764,
917
+ "logits/real": -2.7718958854675293,
918
+ "logps/generated": -417.58770751953125,
919
+ "logps/real": -419.9620056152344,
920
+ "loss": 0.0031,
921
+ "rewards/accuracies": 1.0,
922
+ "rewards/generated": -12.79851245880127,
923
+ "rewards/margins": 11.963459014892578,
924
+ "rewards/real": -0.835052490234375,
925
+ "step": 450
926
+ },
927
+ {
928
+ "epoch": 1.48,
929
+ "eval_logits/generated": -2.7543997764587402,
930
+ "eval_logits/real": -2.7815120220184326,
931
+ "eval_logps/generated": -403.600341796875,
932
+ "eval_logps/real": -426.4503479003906,
933
+ "eval_loss": 0.0138690285384655,
934
+ "eval_rewards/accuracies": 0.9963235259056091,
935
+ "eval_rewards/generated": -12.690093994140625,
936
+ "eval_rewards/margins": 11.313016891479492,
937
+ "eval_rewards/real": -1.3770767450332642,
938
+ "eval_runtime": 277.1873,
939
+ "eval_samples_per_second": 7.789,
940
+ "eval_steps_per_second": 0.245,
941
+ "step": 450
942
+ },
943
+ {
944
+ "epoch": 1.51,
945
+ "learning_rate": 1.3528336380255943e-07,
946
+ "logits/generated": -2.752661943435669,
947
+ "logits/real": -2.7650394439697266,
948
+ "logps/generated": -416.608154296875,
949
+ "logps/real": -423.7530212402344,
950
+ "loss": 0.0048,
951
+ "rewards/accuracies": 0.9937499761581421,
952
+ "rewards/generated": -13.213157653808594,
953
+ "rewards/margins": 12.553186416625977,
954
+ "rewards/real": -0.6599710583686829,
955
+ "step": 460
956
+ },
957
+ {
958
+ "epoch": 1.55,
959
+ "learning_rate": 1.2614259597806215e-07,
960
+ "logits/generated": -2.745286464691162,
961
+ "logits/real": -2.7696480751037598,
962
+ "logps/generated": -412.99420166015625,
963
+ "logps/real": -438.382568359375,
964
+ "loss": 0.0039,
965
+ "rewards/accuracies": 0.9937499761581421,
966
+ "rewards/generated": -12.78739070892334,
967
+ "rewards/margins": 11.881449699401855,
968
+ "rewards/real": -0.9059404134750366,
969
+ "step": 470
970
+ },
971
+ {
972
+ "epoch": 1.56,
973
+ "eval_logits/generated": -2.7446181774139404,
974
+ "eval_logits/real": -2.7735180854797363,
975
+ "eval_logps/generated": -404.79119873046875,
976
+ "eval_logps/real": -426.5647888183594,
977
+ "eval_loss": 0.013367247767746449,
978
+ "eval_rewards/accuracies": 0.9963235259056091,
979
+ "eval_rewards/generated": -12.809175491333008,
980
+ "eval_rewards/margins": 11.42065143585205,
981
+ "eval_rewards/real": -1.3885247707366943,
982
+ "eval_runtime": 276.6137,
983
+ "eval_samples_per_second": 7.805,
984
+ "eval_steps_per_second": 0.246,
985
+ "step": 475
986
+ },
987
+ {
988
+ "epoch": 1.58,
989
+ "learning_rate": 1.1700182815356489e-07,
990
+ "logits/generated": -2.7754604816436768,
991
+ "logits/real": -2.7696869373321533,
992
+ "logps/generated": -402.7199401855469,
993
+ "logps/real": -402.69091796875,
994
+ "loss": 0.001,
995
+ "rewards/accuracies": 1.0,
996
+ "rewards/generated": -13.116182327270508,
997
+ "rewards/margins": 11.969404220581055,
998
+ "rewards/real": -1.146780252456665,
999
+ "step": 480
1000
+ },
1001
+ {
1002
+ "epoch": 1.61,
1003
+ "learning_rate": 1.0786106032906764e-07,
1004
+ "logits/generated": -2.76755952835083,
1005
+ "logits/real": -2.7802929878234863,
1006
+ "logps/generated": -425.44757080078125,
1007
+ "logps/real": -418.4566345214844,
1008
+ "loss": 0.003,
1009
+ "rewards/accuracies": 1.0,
1010
+ "rewards/generated": -13.474451065063477,
1011
+ "rewards/margins": 12.222225189208984,
1012
+ "rewards/real": -1.2522268295288086,
1013
+ "step": 490
1014
+ },
1015
+ {
1016
+ "epoch": 1.64,
1017
+ "learning_rate": 9.872029250457038e-08,
1018
+ "logits/generated": -2.7439827919006348,
1019
+ "logits/real": -2.7600913047790527,
1020
+ "logps/generated": -400.67523193359375,
1021
+ "logps/real": -404.1493225097656,
1022
+ "loss": 0.001,
1023
+ "rewards/accuracies": 1.0,
1024
+ "rewards/generated": -12.943422317504883,
1025
+ "rewards/margins": 11.872589111328125,
1026
+ "rewards/real": -1.0708342790603638,
1027
+ "step": 500
1028
+ },
1029
+ {
1030
+ "epoch": 1.64,
1031
+ "eval_logits/generated": -2.7403690814971924,
1032
+ "eval_logits/real": -2.7701306343078613,
1033
+ "eval_logps/generated": -406.73699951171875,
1034
+ "eval_logps/real": -427.05706787109375,
1035
+ "eval_loss": 0.013619424775242805,
1036
+ "eval_rewards/accuracies": 0.9963235259056091,
1037
+ "eval_rewards/generated": -13.003759384155273,
1038
+ "eval_rewards/margins": 11.566007614135742,
1039
+ "eval_rewards/real": -1.4377506971359253,
1040
+ "eval_runtime": 277.1256,
1041
+ "eval_samples_per_second": 7.791,
1042
+ "eval_steps_per_second": 0.245,
1043
+ "step": 500
1044
+ },
1045
+ {
1046
+ "epoch": 1.68,
1047
+ "learning_rate": 8.957952468007312e-08,
1048
+ "logits/generated": -2.746847629547119,
1049
+ "logits/real": -2.7727770805358887,
1050
+ "logps/generated": -415.89227294921875,
1051
+ "logps/real": -417.8184509277344,
1052
+ "loss": 0.0045,
1053
+ "rewards/accuracies": 1.0,
1054
+ "rewards/generated": -13.302165031433105,
1055
+ "rewards/margins": 11.868834495544434,
1056
+ "rewards/real": -1.4333298206329346,
1057
+ "step": 510
1058
+ },
1059
+ {
1060
+ "epoch": 1.71,
1061
+ "learning_rate": 8.043875685557587e-08,
1062
+ "logits/generated": -2.7504754066467285,
1063
+ "logits/real": -2.757420063018799,
1064
+ "logps/generated": -405.920166015625,
1065
+ "logps/real": -403.3671875,
1066
+ "loss": 0.0059,
1067
+ "rewards/accuracies": 0.9937499761581421,
1068
+ "rewards/generated": -13.80413818359375,
1069
+ "rewards/margins": 12.69081974029541,
1070
+ "rewards/real": -1.1133191585540771,
1071
+ "step": 520
1072
+ },
1073
+ {
1074
+ "epoch": 1.73,
1075
+ "eval_logits/generated": -2.7293171882629395,
1076
+ "eval_logits/real": -2.7628748416900635,
1077
+ "eval_logps/generated": -410.86712646484375,
1078
+ "eval_logps/real": -428.60345458984375,
1079
+ "eval_loss": 0.013887421227991581,
1080
+ "eval_rewards/accuracies": 0.9944853186607361,
1081
+ "eval_rewards/generated": -13.416768074035645,
1082
+ "eval_rewards/margins": 11.824378967285156,
1083
+ "eval_rewards/real": -1.59238862991333,
1084
+ "eval_runtime": 280.0429,
1085
+ "eval_samples_per_second": 7.71,
1086
+ "eval_steps_per_second": 0.243,
1087
+ "step": 525
1088
+ },
1089
+ {
1090
+ "epoch": 1.74,
1091
+ "learning_rate": 7.12979890310786e-08,
1092
+ "logits/generated": -2.7278664112091064,
1093
+ "logits/real": -2.767366886138916,
1094
+ "logps/generated": -398.7755432128906,
1095
+ "logps/real": -448.11932373046875,
1096
+ "loss": 0.0031,
1097
+ "rewards/accuracies": 1.0,
1098
+ "rewards/generated": -13.456060409545898,
1099
+ "rewards/margins": 11.94404411315918,
1100
+ "rewards/real": -1.51201593875885,
1101
+ "step": 530
1102
+ },
1103
+ {
1104
+ "epoch": 1.78,
1105
+ "learning_rate": 6.215722120658136e-08,
1106
+ "logits/generated": -2.743837356567383,
1107
+ "logits/real": -2.776067018508911,
1108
+ "logps/generated": -408.1714782714844,
1109
+ "logps/real": -454.8717346191406,
1110
+ "loss": 0.0023,
1111
+ "rewards/accuracies": 1.0,
1112
+ "rewards/generated": -13.510284423828125,
1113
+ "rewards/margins": 12.263895988464355,
1114
+ "rewards/real": -1.2463879585266113,
1115
+ "step": 540
1116
+ },
1117
+ {
1118
+ "epoch": 1.81,
1119
+ "learning_rate": 5.301645338208409e-08,
1120
+ "logits/generated": -2.7187867164611816,
1121
+ "logits/real": -2.753685474395752,
1122
+ "logps/generated": -414.6969299316406,
1123
+ "logps/real": -451.0293884277344,
1124
+ "loss": 0.0015,
1125
+ "rewards/accuracies": 1.0,
1126
+ "rewards/generated": -13.869306564331055,
1127
+ "rewards/margins": 12.921630859375,
1128
+ "rewards/real": -0.9476770162582397,
1129
+ "step": 550
1130
+ },
1131
+ {
1132
+ "epoch": 1.81,
1133
+ "eval_logits/generated": -2.7283482551574707,
1134
+ "eval_logits/real": -2.762289524078369,
1135
+ "eval_logps/generated": -410.68316650390625,
1136
+ "eval_logps/real": -427.815673828125,
1137
+ "eval_loss": 0.013605926185846329,
1138
+ "eval_rewards/accuracies": 0.9963235259056091,
1139
+ "eval_rewards/generated": -13.398374557495117,
1140
+ "eval_rewards/margins": 11.884763717651367,
1141
+ "eval_rewards/real": -1.5136103630065918,
1142
+ "eval_runtime": 279.4319,
1143
+ "eval_samples_per_second": 7.726,
1144
+ "eval_steps_per_second": 0.243,
1145
+ "step": 550
1146
+ },
1147
+ {
1148
+ "epoch": 1.84,
1149
+ "learning_rate": 4.387568555758683e-08,
1150
+ "logits/generated": -2.7442142963409424,
1151
+ "logits/real": -2.7617897987365723,
1152
+ "logps/generated": -428.95721435546875,
1153
+ "logps/real": -429.8484802246094,
1154
+ "loss": 0.0035,
1155
+ "rewards/accuracies": 1.0,
1156
+ "rewards/generated": -14.423190116882324,
1157
+ "rewards/margins": 13.05412483215332,
1158
+ "rewards/real": -1.3690690994262695,
1159
+ "step": 560
1160
+ },
1161
+ {
1162
+ "epoch": 1.88,
1163
+ "learning_rate": 3.4734917733089575e-08,
1164
+ "logits/generated": -2.720496654510498,
1165
+ "logits/real": -2.742316722869873,
1166
+ "logps/generated": -408.7587890625,
1167
+ "logps/real": -418.96612548828125,
1168
+ "loss": 0.0078,
1169
+ "rewards/accuracies": 1.0,
1170
+ "rewards/generated": -14.030049324035645,
1171
+ "rewards/margins": 12.879358291625977,
1172
+ "rewards/real": -1.1506898403167725,
1173
+ "step": 570
1174
+ },
1175
+ {
1176
+ "epoch": 1.89,
1177
+ "eval_logits/generated": -2.730926752090454,
1178
+ "eval_logits/real": -2.7644882202148438,
1179
+ "eval_logps/generated": -410.0223693847656,
1180
+ "eval_logps/real": -427.5704040527344,
1181
+ "eval_loss": 0.013477620668709278,
1182
+ "eval_rewards/accuracies": 0.9963235259056091,
1183
+ "eval_rewards/generated": -13.332295417785645,
1184
+ "eval_rewards/margins": 11.843212127685547,
1185
+ "eval_rewards/real": -1.4890828132629395,
1186
+ "eval_runtime": 275.7027,
1187
+ "eval_samples_per_second": 7.831,
1188
+ "eval_steps_per_second": 0.247,
1189
+ "step": 575
1190
+ },
1191
+ {
1192
+ "epoch": 1.91,
1193
+ "learning_rate": 2.5594149908592323e-08,
1194
+ "logits/generated": -2.7408628463745117,
1195
+ "logits/real": -2.7660305500030518,
1196
+ "logps/generated": -404.4073486328125,
1197
+ "logps/real": -441.6896057128906,
1198
+ "loss": 0.0011,
1199
+ "rewards/accuracies": 1.0,
1200
+ "rewards/generated": -13.855819702148438,
1201
+ "rewards/margins": 12.916915893554688,
1202
+ "rewards/real": -0.9389039874076843,
1203
+ "step": 580
1204
+ },
1205
+ {
1206
+ "epoch": 1.94,
1207
+ "learning_rate": 1.6453382084095063e-08,
1208
+ "logits/generated": -2.7413947582244873,
1209
+ "logits/real": -2.7642316818237305,
1210
+ "logps/generated": -462.260986328125,
1211
+ "logps/real": -425.875244140625,
1212
+ "loss": 0.0019,
1213
+ "rewards/accuracies": 1.0,
1214
+ "rewards/generated": -14.152559280395508,
1215
+ "rewards/margins": 12.800895690917969,
1216
+ "rewards/real": -1.3516645431518555,
1217
+ "step": 590
1218
+ },
1219
+ {
1220
+ "epoch": 1.97,
1221
+ "learning_rate": 7.312614259597806e-09,
1222
+ "logits/generated": -2.7413601875305176,
1223
+ "logits/real": -2.767582416534424,
1224
+ "logps/generated": -410.154052734375,
1225
+ "logps/real": -414.71295166015625,
1226
+ "loss": 0.0043,
1227
+ "rewards/accuracies": 0.9937499761581421,
1228
+ "rewards/generated": -13.285085678100586,
1229
+ "rewards/margins": 12.160585403442383,
1230
+ "rewards/real": -1.1245028972625732,
1231
+ "step": 600
1232
+ },
1233
+ {
1234
+ "epoch": 1.97,
1235
+ "eval_logits/generated": -2.7305350303649902,
1236
+ "eval_logits/real": -2.7643280029296875,
1237
+ "eval_logps/generated": -410.0756530761719,
1238
+ "eval_logps/real": -427.497802734375,
1239
+ "eval_loss": 0.01349999662488699,
1240
+ "eval_rewards/accuracies": 0.9963235259056091,
1241
+ "eval_rewards/generated": -13.337620735168457,
1242
+ "eval_rewards/margins": 11.855793952941895,
1243
+ "eval_rewards/real": -1.4818273782730103,
1244
+ "eval_runtime": 276.8139,
1245
+ "eval_samples_per_second": 7.799,
1246
+ "eval_steps_per_second": 0.246,
1247
+ "step": 600
1248
+ },
1249
+ {
1250
+ "epoch": 2.0,
1251
+ "step": 608,
1252
+ "total_flos": 0.0,
1253
+ "train_loss": 0.04328188924946038,
1254
+ "train_runtime": 16735.9464,
1255
+ "train_samples_per_second": 2.321,
1256
+ "train_steps_per_second": 0.036
1257
+ }
1258
+ ],
1259
+ "logging_steps": 10,
1260
+ "max_steps": 608,
1261
+ "num_input_tokens_seen": 0,
1262
+ "num_train_epochs": 2,
1263
+ "save_steps": 500,
1264
+ "total_flos": 0.0,
1265
+ "train_batch_size": 8,
1266
+ "trial_name": null,
1267
+ "trial_params": null
1268
+ }