hZzy commited on
Commit
9f7a057
1 Parent(s): 6c2dad6

Model save

Browse files
README.md ADDED
@@ -0,0 +1,80 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ base_model: hZzy/qwen2.5-0.5b-sft-news-IFT
4
+ tags:
5
+ - trl
6
+ - expo
7
+ - generated_from_trainer
8
+ model-index:
9
+ - name: qwen2.5-0.5b-expo-DPO-ES-0.01
10
+ results: []
11
+ ---
12
+
13
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
14
+ should probably proofread and complete it, then remove this comment. -->
15
+
16
+ [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="200" height="32"/>](https://wandb.ai/zhiyuzha-university-of-florida/huggingface/runs/aeakfz58)
17
+ # qwen2.5-0.5b-expo-DPO-ES-0.01
18
+
19
+ This model is a fine-tuned version of [hZzy/qwen2.5-0.5b-sft-news-IFT](https://huggingface.co/hZzy/qwen2.5-0.5b-sft-news-IFT) on an unknown dataset.
20
+ It achieves the following results on the evaluation set:
21
+ - Loss: 0.6640
22
+ - Logps: -319.3935
23
+ - Logits: -4.6100
24
+ - Objective: 0.6685
25
+ - Dpo Loss: 0.6685
26
+ - Regularize: 0.6685
27
+ - Ranking Simple: 0.5828
28
+ - Ranking Idealized: 0.8732
29
+ - Ranking Idealized Expo: 0.5321
30
+ - Wo Beta: 23.6506
31
+
32
+ ## Model description
33
+
34
+ More information needed
35
+
36
+ ## Intended uses & limitations
37
+
38
+ More information needed
39
+
40
+ ## Training and evaluation data
41
+
42
+ More information needed
43
+
44
+ ## Training procedure
45
+
46
+ ### Training hyperparameters
47
+
48
+ The following hyperparameters were used during training:
49
+ - learning_rate: 5e-06
50
+ - train_batch_size: 4
51
+ - eval_batch_size: 4
52
+ - seed: 42
53
+ - distributed_type: multi-GPU
54
+ - num_devices: 3
55
+ - gradient_accumulation_steps: 12
56
+ - total_train_batch_size: 144
57
+ - total_eval_batch_size: 12
58
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
59
+ - lr_scheduler_type: cosine
60
+ - lr_scheduler_warmup_ratio: 0.1
61
+ - num_epochs: 5
62
+
63
+ ### Training results
64
+
65
+ | Training Loss | Epoch | Step | Validation Loss | Logps | Logits | Objective | Dpo Loss | Regularize | Ranking Simple | Ranking Idealized | Ranking Idealized Expo | Wo Beta |
66
+ |:-------------:|:------:|:----:|:---------------:|:---------:|:-------:|:---------:|:--------:|:----------:|:--------------:|:-----------------:|:----------------------:|:-------:|
67
+ | 0.6907 | 0.1417 | 50 | 0.6894 | -99.3821 | -1.7611 | 0.6902 | 0.6902 | 0.6902 | 0.5305 | 0.8732 | 0.5321 | 9.3575 |
68
+ | 0.6701 | 0.2834 | 100 | 0.6837 | -153.5716 | -1.8467 | 0.6896 | 0.6896 | 0.6896 | 0.5518 | 0.8732 | 0.5321 | 14.1741 |
69
+ | 0.637 | 0.4251 | 150 | 0.6723 | -198.9269 | -2.4614 | 0.6765 | 0.6765 | 0.6765 | 0.5823 | 0.8732 | 0.5321 | 17.2098 |
70
+ | 0.5833 | 0.5668 | 200 | 0.6729 | -256.0312 | -3.3478 | 0.6780 | 0.6780 | 0.6780 | 0.5797 | 0.8732 | 0.5321 | 21.7109 |
71
+ | 0.5439 | 0.7085 | 250 | 0.6781 | -257.3546 | -3.6269 | 0.6858 | 0.6858 | 0.6858 | 0.5683 | 0.8732 | 0.5321 | 22.9139 |
72
+ | 0.5077 | 0.8503 | 300 | 0.6640 | -319.3935 | -4.6100 | 0.6685 | 0.6685 | 0.6685 | 0.5828 | 0.8732 | 0.5321 | 23.6506 |
73
+
74
+
75
+ ### Framework versions
76
+
77
+ - Transformers 4.42.0
78
+ - Pytorch 2.3.0+cu121
79
+ - Datasets 2.19.1
80
+ - Tokenizers 0.19.1
all_results.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 0.8502598016060463,
3
+ "total_flos": 0.0,
4
+ "train_loss": 0.6054576412836711,
5
+ "train_runtime": 8287.5068,
6
+ "train_samples": 50802,
7
+ "train_samples_per_second": 30.65,
8
+ "train_steps_per_second": 0.212
9
+ }
generation_config.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token_id": 151644,
3
+ "eos_token_id": 151645,
4
+ "max_new_tokens": 2048,
5
+ "pad_token_id": 151645,
6
+ "transformers_version": "4.42.0"
7
+ }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:520c19027f98c4a2feafb45c7d81bb1e84cbf06c71d7d8530e453a702a49d37a
3
  size 1975192208
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a6fbadd5f72d783e4eef10b0ec932f1d5e13f20e83ebacd844cdbf5fb249cb55
3
  size 1975192208
train_results.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 0.8502598016060463,
3
+ "total_flos": 0.0,
4
+ "train_loss": 0.6054576412836711,
5
+ "train_runtime": 8287.5068,
6
+ "train_samples": 50802,
7
+ "train_samples_per_second": 30.65,
8
+ "train_steps_per_second": 0.212
9
+ }
trainer_state.json ADDED
@@ -0,0 +1,265 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": 9.357504844665527,
3
+ "best_model_checkpoint": "./qwen2.5-0.5b/qwen2.5-0.5b-expo-DPO-ES-0.01/checkpoint-50",
4
+ "epoch": 0.8502598016060463,
5
+ "eval_steps": 50,
6
+ "global_step": 300,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "dpo_loss": 0.6931471824645996,
13
+ "epoch": 0.002834199338686821,
14
+ "grad_norm": 1.8442555137313956,
15
+ "learning_rate": 2.840909090909091e-08,
16
+ "logits": -1.359458565711975,
17
+ "logps": -84.69721221923828,
18
+ "loss": 0.6931,
19
+ "objective": 0.6931471824645996,
20
+ "ranking_idealized": 0.875,
21
+ "ranking_idealized_expo": 0.5833333134651184,
22
+ "ranking_simple": 0.5833333134651184,
23
+ "regularize": 0.6931471824645996,
24
+ "step": 1,
25
+ "wo_beta": 5.271125316619873
26
+ },
27
+ {
28
+ "dpo_loss": 0.6914740800857544,
29
+ "epoch": 0.14170996693434104,
30
+ "grad_norm": 2.883110438441685,
31
+ "learning_rate": 1.4204545454545458e-06,
32
+ "logits": -1.5272094011306763,
33
+ "logps": -83.2472152709961,
34
+ "loss": 0.6907,
35
+ "objective": 0.6914740800857544,
36
+ "ranking_idealized": 0.8945578336715698,
37
+ "ranking_idealized_expo": 0.5348639488220215,
38
+ "ranking_simple": 0.5293367505073547,
39
+ "regularize": 0.6914740800857544,
40
+ "step": 50,
41
+ "wo_beta": 7.299497604370117
42
+ },
43
+ {
44
+ "epoch": 0.14170996693434104,
45
+ "eval_dpo_loss": 0.6902037858963013,
46
+ "eval_logits": -1.7611440420150757,
47
+ "eval_logps": -99.38214111328125,
48
+ "eval_loss": 0.6893583536148071,
49
+ "eval_objective": 0.6902037858963013,
50
+ "eval_ranking_idealized": 0.8731883764266968,
51
+ "eval_ranking_idealized_expo": 0.5320910811424255,
52
+ "eval_ranking_simple": 0.5305383205413818,
53
+ "eval_regularize": 0.6902037858963013,
54
+ "eval_runtime": 314.0138,
55
+ "eval_samples_per_second": 18.439,
56
+ "eval_steps_per_second": 1.538,
57
+ "eval_wo_beta": 9.357504844665527,
58
+ "step": 50
59
+ },
60
+ {
61
+ "dpo_loss": 0.6664106249809265,
62
+ "epoch": 0.2834199338686821,
63
+ "grad_norm": 7.452572178923395,
64
+ "learning_rate": 2.8409090909090916e-06,
65
+ "logits": -1.6917132139205933,
66
+ "logps": -118.66804504394531,
67
+ "loss": 0.6701,
68
+ "objective": 0.6664106249809265,
69
+ "ranking_idealized": 0.8958333134651184,
70
+ "ranking_idealized_expo": 0.5270833373069763,
71
+ "ranking_simple": 0.5866666436195374,
72
+ "regularize": 0.6664106249809265,
73
+ "step": 100,
74
+ "wo_beta": 9.330028533935547
75
+ },
76
+ {
77
+ "epoch": 0.2834199338686821,
78
+ "eval_dpo_loss": 0.6895959377288818,
79
+ "eval_logits": -1.8466589450836182,
80
+ "eval_logps": -153.5715789794922,
81
+ "eval_loss": 0.6837136149406433,
82
+ "eval_objective": 0.6895959377288818,
83
+ "eval_ranking_idealized": 0.8731883764266968,
84
+ "eval_ranking_idealized_expo": 0.5320910811424255,
85
+ "eval_ranking_simple": 0.5517598390579224,
86
+ "eval_regularize": 0.6895959377288818,
87
+ "eval_runtime": 308.2112,
88
+ "eval_samples_per_second": 18.786,
89
+ "eval_steps_per_second": 1.567,
90
+ "eval_wo_beta": 14.174129486083984,
91
+ "step": 100
92
+ },
93
+ {
94
+ "dpo_loss": 0.6343104243278503,
95
+ "epoch": 0.42512990080302315,
96
+ "grad_norm": 13.296765037353357,
97
+ "learning_rate": 4.2613636363636365e-06,
98
+ "logits": -1.825290322303772,
99
+ "logps": -173.7165985107422,
100
+ "loss": 0.637,
101
+ "objective": 0.6343104243278503,
102
+ "ranking_idealized": 0.8862500190734863,
103
+ "ranking_idealized_expo": 0.5358333587646484,
104
+ "ranking_simple": 0.6316666603088379,
105
+ "regularize": 0.6343104243278503,
106
+ "step": 150,
107
+ "wo_beta": 12.891901016235352
108
+ },
109
+ {
110
+ "epoch": 0.42512990080302315,
111
+ "eval_dpo_loss": 0.6764841079711914,
112
+ "eval_logits": -2.461402416229248,
113
+ "eval_logps": -198.92694091796875,
114
+ "eval_loss": 0.6722760796546936,
115
+ "eval_objective": 0.6764841079711914,
116
+ "eval_ranking_idealized": 0.8731883764266968,
117
+ "eval_ranking_idealized_expo": 0.5320910811424255,
118
+ "eval_ranking_simple": 0.5822981595993042,
119
+ "eval_regularize": 0.6764841079711914,
120
+ "eval_runtime": 308.7795,
121
+ "eval_samples_per_second": 18.751,
122
+ "eval_steps_per_second": 1.564,
123
+ "eval_wo_beta": 17.20980453491211,
124
+ "step": 150
125
+ },
126
+ {
127
+ "dpo_loss": 0.5824262499809265,
128
+ "epoch": 0.5668398677373642,
129
+ "grad_norm": 14.182378256685057,
130
+ "learning_rate": 4.997168347957521e-06,
131
+ "logits": -2.928657054901123,
132
+ "logps": -216.66006469726562,
133
+ "loss": 0.5833,
134
+ "objective": 0.5824262499809265,
135
+ "ranking_idealized": 0.8883333206176758,
136
+ "ranking_idealized_expo": 0.527916669845581,
137
+ "ranking_simple": 0.6833333373069763,
138
+ "regularize": 0.5824262499809265,
139
+ "step": 200,
140
+ "wo_beta": 13.758678436279297
141
+ },
142
+ {
143
+ "epoch": 0.5668398677373642,
144
+ "eval_dpo_loss": 0.6780264377593994,
145
+ "eval_logits": -3.3478028774261475,
146
+ "eval_logps": -256.0312194824219,
147
+ "eval_loss": 0.6728978157043457,
148
+ "eval_objective": 0.6780264377593994,
149
+ "eval_ranking_idealized": 0.8731883764266968,
150
+ "eval_ranking_idealized_expo": 0.5320910811424255,
151
+ "eval_ranking_simple": 0.5797101259231567,
152
+ "eval_regularize": 0.6780264377593994,
153
+ "eval_runtime": 312.7328,
154
+ "eval_samples_per_second": 18.514,
155
+ "eval_steps_per_second": 1.544,
156
+ "eval_wo_beta": 21.710927963256836,
157
+ "step": 200
158
+ },
159
+ {
160
+ "dpo_loss": 0.541719377040863,
161
+ "epoch": 0.7085498346717053,
162
+ "grad_norm": 16.105604013075503,
163
+ "learning_rate": 4.973122855144066e-06,
164
+ "logits": -3.4248011112213135,
165
+ "logps": -254.68023681640625,
166
+ "loss": 0.5439,
167
+ "objective": 0.541719377040863,
168
+ "ranking_idealized": 0.8787500262260437,
169
+ "ranking_idealized_expo": 0.5266666412353516,
170
+ "ranking_simple": 0.7054166793823242,
171
+ "regularize": 0.541719377040863,
172
+ "step": 250,
173
+ "wo_beta": 15.057023048400879
174
+ },
175
+ {
176
+ "epoch": 0.7085498346717053,
177
+ "eval_dpo_loss": 0.6857688426971436,
178
+ "eval_logits": -3.626901865005493,
179
+ "eval_logps": -257.3546447753906,
180
+ "eval_loss": 0.6781259775161743,
181
+ "eval_objective": 0.6857688426971436,
182
+ "eval_ranking_idealized": 0.8731883764266968,
183
+ "eval_ranking_idealized_expo": 0.5320910811424255,
184
+ "eval_ranking_simple": 0.5683229565620422,
185
+ "eval_regularize": 0.6857688426971436,
186
+ "eval_runtime": 308.1728,
187
+ "eval_samples_per_second": 18.788,
188
+ "eval_steps_per_second": 1.567,
189
+ "eval_wo_beta": 22.9139461517334,
190
+ "step": 250
191
+ },
192
+ {
193
+ "dpo_loss": 0.49844247102737427,
194
+ "epoch": 0.8502598016060463,
195
+ "grad_norm": 14.80120970807248,
196
+ "learning_rate": 4.924776641419513e-06,
197
+ "logits": -4.0557355880737305,
198
+ "logps": -292.7536315917969,
199
+ "loss": 0.5077,
200
+ "objective": 0.49844247102737427,
201
+ "ranking_idealized": 0.8733333349227905,
202
+ "ranking_idealized_expo": 0.5104166865348816,
203
+ "ranking_simple": 0.747083306312561,
204
+ "regularize": 0.49844247102737427,
205
+ "step": 300,
206
+ "wo_beta": 13.577784538269043
207
+ },
208
+ {
209
+ "epoch": 0.8502598016060463,
210
+ "eval_dpo_loss": 0.6684697866439819,
211
+ "eval_logits": -4.609988689422607,
212
+ "eval_logps": -319.3935241699219,
213
+ "eval_loss": 0.663960874080658,
214
+ "eval_objective": 0.6684697866439819,
215
+ "eval_ranking_idealized": 0.8731883764266968,
216
+ "eval_ranking_idealized_expo": 0.5320910811424255,
217
+ "eval_ranking_simple": 0.5828157067298889,
218
+ "eval_regularize": 0.6684697866439819,
219
+ "eval_runtime": 309.749,
220
+ "eval_samples_per_second": 18.693,
221
+ "eval_steps_per_second": 1.559,
222
+ "eval_wo_beta": 23.650588989257812,
223
+ "step": 300
224
+ },
225
+ {
226
+ "epoch": 0.8502598016060463,
227
+ "step": 300,
228
+ "total_flos": 0.0,
229
+ "train_loss": 0.6054576412836711,
230
+ "train_runtime": 8287.5068,
231
+ "train_samples_per_second": 30.65,
232
+ "train_steps_per_second": 0.212
233
+ }
234
+ ],
235
+ "logging_steps": 50,
236
+ "max_steps": 1760,
237
+ "num_input_tokens_seen": 0,
238
+ "num_train_epochs": 5,
239
+ "save_steps": 50,
240
+ "stateful_callbacks": {
241
+ "EarlyStoppingCallback": {
242
+ "args": {
243
+ "early_stopping_patience": 5,
244
+ "early_stopping_threshold": 0.0
245
+ },
246
+ "attributes": {
247
+ "early_stopping_patience_counter": 0
248
+ }
249
+ },
250
+ "TrainerControl": {
251
+ "args": {
252
+ "should_epoch_stop": false,
253
+ "should_evaluate": false,
254
+ "should_log": false,
255
+ "should_save": true,
256
+ "should_training_stop": true
257
+ },
258
+ "attributes": {}
259
+ }
260
+ },
261
+ "total_flos": 0.0,
262
+ "train_batch_size": 4,
263
+ "trial_name": null,
264
+ "trial_params": null
265
+ }