sujithatz commited on
Commit
3394dab
1 Parent(s): ca90f45

sujithatz/testfinbot-quen2.5-lora-merged_adapter

Browse files
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ tokenizer.json filter=lfs diff=lfs merge=lfs -text
README.md ADDED
@@ -0,0 +1,554 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: Qwen/Qwen2.5-0.5B-Instruct
3
+ library_name: peft
4
+ license: apache-2.0
5
+ tags:
6
+ - trl
7
+ - sft
8
+ - generated_from_trainer
9
+ model-index:
10
+ - name: quen_2.5_qlora
11
+ results: []
12
+ ---
13
+
14
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
15
+ should probably proofread and complete it, then remove this comment. -->
16
+
17
+ # quen_2.5_qlora
18
+
19
+ This model is a fine-tuned version of [Qwen/Qwen2.5-0.5B-Instruct](https://huggingface.co/Qwen/Qwen2.5-0.5B-Instruct) on an unknown dataset.
20
+ It achieves the following results on the evaluation set:
21
+ - Loss: 0.7671
22
+
23
+ ## Model description
24
+
25
+ More information needed
26
+
27
+ ## Intended uses & limitations
28
+
29
+ More information needed
30
+
31
+ ## Training and evaluation data
32
+
33
+ More information needed
34
+
35
+ ## Training procedure
36
+
37
+ ### Training hyperparameters
38
+
39
+ The following hyperparameters were used during training:
40
+ - learning_rate: 5e-06
41
+ - train_batch_size: 8
42
+ - eval_batch_size: 8
43
+ - seed: 0
44
+ - optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
45
+ - lr_scheduler_type: cosine
46
+ - lr_scheduler_warmup_ratio: 0.01
47
+ - num_epochs: 40
48
+
49
+ ### Training results
50
+
51
+ | Training Loss | Epoch | Step | Validation Loss |
52
+ |:-------------:|:-------:|:----:|:---------------:|
53
+ | 3.1118 | 0.0811 | 3 | 3.2024 |
54
+ | 3.0902 | 0.1622 | 6 | 3.2023 |
55
+ | 2.9193 | 0.2432 | 9 | 3.1908 |
56
+ | 3.0231 | 0.3243 | 12 | 3.1700 |
57
+ | 3.0429 | 0.4054 | 15 | 3.1323 |
58
+ | 2.9739 | 0.4865 | 18 | 3.0926 |
59
+ | 2.9118 | 0.5676 | 21 | 3.0477 |
60
+ | 2.8564 | 0.6486 | 24 | 2.9984 |
61
+ | 2.9689 | 0.7297 | 27 | 2.9489 |
62
+ | 2.8155 | 0.8108 | 30 | 2.8946 |
63
+ | 2.8495 | 0.8919 | 33 | 2.8411 |
64
+ | 2.7311 | 0.9730 | 36 | 2.7843 |
65
+ | 2.6327 | 1.0541 | 39 | 2.7375 |
66
+ | 2.6728 | 1.1351 | 42 | 2.6747 |
67
+ | 2.5236 | 1.2162 | 45 | 2.6151 |
68
+ | 2.5009 | 1.2973 | 48 | 2.5595 |
69
+ | 2.4129 | 1.3784 | 51 | 2.5030 |
70
+ | 2.3506 | 1.4595 | 54 | 2.4369 |
71
+ | 2.3542 | 1.5405 | 57 | 2.3803 |
72
+ | 2.237 | 1.6216 | 60 | 2.3189 |
73
+ | 2.3302 | 1.7027 | 63 | 2.2589 |
74
+ | 2.1679 | 1.7838 | 66 | 2.2005 |
75
+ | 2.1406 | 1.8649 | 69 | 2.1355 |
76
+ | 2.0373 | 1.9459 | 72 | 2.0689 |
77
+ | 1.9979 | 2.0270 | 75 | 2.0098 |
78
+ | 2.0711 | 2.1081 | 78 | 1.9351 |
79
+ | 1.8819 | 2.1892 | 81 | 1.8753 |
80
+ | 1.7832 | 2.2703 | 84 | 1.8243 |
81
+ | 1.7165 | 2.3514 | 87 | 1.7668 |
82
+ | 1.6765 | 2.4324 | 90 | 1.7111 |
83
+ | 1.6243 | 2.5135 | 93 | 1.6613 |
84
+ | 1.6819 | 2.5946 | 96 | 1.6154 |
85
+ | 1.6006 | 2.6757 | 99 | 1.5679 |
86
+ | 1.4845 | 2.7568 | 102 | 1.5229 |
87
+ | 1.4014 | 2.8378 | 105 | 1.4728 |
88
+ | 1.452 | 2.9189 | 108 | 1.4313 |
89
+ | 1.3642 | 3.0 | 111 | 1.3985 |
90
+ | 1.3571 | 3.0811 | 114 | 1.3647 |
91
+ | 1.4434 | 3.1622 | 117 | 1.3416 |
92
+ | 1.3829 | 3.2432 | 120 | 1.3206 |
93
+ | 1.3319 | 3.3243 | 123 | 1.3000 |
94
+ | 1.2262 | 3.4054 | 126 | 1.2851 |
95
+ | 1.2693 | 3.4865 | 129 | 1.2711 |
96
+ | 1.264 | 3.5676 | 132 | 1.2588 |
97
+ | 1.1851 | 3.6486 | 135 | 1.2470 |
98
+ | 1.2656 | 3.7297 | 138 | 1.2380 |
99
+ | 1.205 | 3.8108 | 141 | 1.2244 |
100
+ | 1.2115 | 3.8919 | 144 | 1.2109 |
101
+ | 1.1366 | 3.9730 | 147 | 1.1997 |
102
+ | 1.1794 | 4.0541 | 150 | 1.1929 |
103
+ | 1.171 | 4.1351 | 153 | 1.1805 |
104
+ | 1.0875 | 4.2162 | 156 | 1.1714 |
105
+ | 1.1192 | 4.2973 | 159 | 1.1619 |
106
+ | 1.1082 | 4.3784 | 162 | 1.1548 |
107
+ | 1.0923 | 4.4595 | 165 | 1.1465 |
108
+ | 1.1731 | 4.5405 | 168 | 1.1397 |
109
+ | 1.2138 | 4.6216 | 171 | 1.1315 |
110
+ | 1.2322 | 4.7027 | 174 | 1.1223 |
111
+ | 1.0893 | 4.7838 | 177 | 1.1179 |
112
+ | 1.1414 | 4.8649 | 180 | 1.1116 |
113
+ | 1.0242 | 4.9459 | 183 | 1.1058 |
114
+ | 1.0929 | 5.0270 | 186 | 1.0982 |
115
+ | 1.0967 | 5.1081 | 189 | 1.0916 |
116
+ | 1.0666 | 5.1892 | 192 | 1.0866 |
117
+ | 1.1382 | 5.2703 | 195 | 1.0851 |
118
+ | 1.0709 | 5.3514 | 198 | 1.0781 |
119
+ | 1.0904 | 5.4324 | 201 | 1.0742 |
120
+ | 1.0521 | 5.5135 | 204 | 1.0699 |
121
+ | 0.9745 | 5.5946 | 207 | 1.0641 |
122
+ | 1.1268 | 5.6757 | 210 | 1.0624 |
123
+ | 1.0437 | 5.7568 | 213 | 1.0588 |
124
+ | 1.0124 | 5.8378 | 216 | 1.0578 |
125
+ | 1.0217 | 5.9189 | 219 | 1.0511 |
126
+ | 1.0656 | 6.0 | 222 | 1.0475 |
127
+ | 1.0841 | 6.0811 | 225 | 1.0454 |
128
+ | 1.1352 | 6.1622 | 228 | 1.0440 |
129
+ | 1.0402 | 6.2432 | 231 | 1.0407 |
130
+ | 1.0042 | 6.3243 | 234 | 1.0326 |
131
+ | 0.9738 | 6.4054 | 237 | 1.0295 |
132
+ | 0.9671 | 6.4865 | 240 | 1.0292 |
133
+ | 1.0515 | 6.5676 | 243 | 1.0282 |
134
+ | 0.9551 | 6.6486 | 246 | 1.0246 |
135
+ | 1.0084 | 6.7297 | 249 | 1.0210 |
136
+ | 0.9964 | 6.8108 | 252 | 1.0188 |
137
+ | 0.966 | 6.8919 | 255 | 1.0146 |
138
+ | 1.0284 | 6.9730 | 258 | 1.0113 |
139
+ | 1.0367 | 7.0541 | 261 | 1.0103 |
140
+ | 0.9015 | 7.1351 | 264 | 1.0056 |
141
+ | 0.9097 | 7.2162 | 267 | 1.0043 |
142
+ | 0.9598 | 7.2973 | 270 | 0.9998 |
143
+ | 1.1724 | 7.3784 | 273 | 0.9980 |
144
+ | 0.9607 | 7.4595 | 276 | 0.9947 |
145
+ | 0.9749 | 7.5405 | 279 | 0.9960 |
146
+ | 0.9704 | 7.6216 | 282 | 0.9924 |
147
+ | 0.9855 | 7.7027 | 285 | 0.9899 |
148
+ | 0.9368 | 7.7838 | 288 | 0.9853 |
149
+ | 0.9812 | 7.8649 | 291 | 0.9856 |
150
+ | 0.9809 | 7.9459 | 294 | 0.9809 |
151
+ | 0.9337 | 8.0270 | 297 | 0.9824 |
152
+ | 0.8848 | 8.1081 | 300 | 0.9759 |
153
+ | 0.876 | 8.1892 | 303 | 0.9751 |
154
+ | 0.97 | 8.2703 | 306 | 0.9734 |
155
+ | 1.1309 | 8.3514 | 309 | 0.9726 |
156
+ | 0.9012 | 8.4324 | 312 | 0.9709 |
157
+ | 1.0355 | 8.5135 | 315 | 0.9681 |
158
+ | 0.9047 | 8.5946 | 318 | 0.9663 |
159
+ | 0.9417 | 8.6757 | 321 | 0.9622 |
160
+ | 0.9266 | 8.7568 | 324 | 0.9613 |
161
+ | 0.8873 | 8.8378 | 327 | 0.9625 |
162
+ | 0.9364 | 8.9189 | 330 | 0.9598 |
163
+ | 0.9885 | 9.0 | 333 | 0.9539 |
164
+ | 0.91 | 9.0811 | 336 | 0.9538 |
165
+ | 1.0834 | 9.1622 | 339 | 0.9522 |
166
+ | 0.8694 | 9.2432 | 342 | 0.9502 |
167
+ | 1.0554 | 9.3243 | 345 | 0.9498 |
168
+ | 0.9123 | 9.4054 | 348 | 0.9484 |
169
+ | 0.9542 | 9.4865 | 351 | 0.9449 |
170
+ | 0.9287 | 9.5676 | 354 | 0.9417 |
171
+ | 0.8435 | 9.6486 | 357 | 0.9406 |
172
+ | 0.8419 | 9.7297 | 360 | 0.9374 |
173
+ | 0.8265 | 9.8108 | 363 | 0.9351 |
174
+ | 0.8677 | 9.8919 | 366 | 0.9343 |
175
+ | 0.9989 | 9.9730 | 369 | 0.9341 |
176
+ | 0.8331 | 10.0541 | 372 | 0.9329 |
177
+ | 0.8823 | 10.1351 | 375 | 0.9312 |
178
+ | 0.9312 | 10.2162 | 378 | 0.9299 |
179
+ | 0.8736 | 10.2973 | 381 | 0.9279 |
180
+ | 1.0039 | 10.3784 | 384 | 0.9259 |
181
+ | 0.8728 | 10.4595 | 387 | 0.9249 |
182
+ | 0.9946 | 10.5405 | 390 | 0.9216 |
183
+ | 0.8747 | 10.6216 | 393 | 0.9196 |
184
+ | 0.9043 | 10.7027 | 396 | 0.9183 |
185
+ | 0.8777 | 10.7838 | 399 | 0.9163 |
186
+ | 0.8023 | 10.8649 | 402 | 0.9160 |
187
+ | 0.9002 | 10.9459 | 405 | 0.9128 |
188
+ | 0.8568 | 11.0270 | 408 | 0.9146 |
189
+ | 0.9145 | 11.1081 | 411 | 0.9124 |
190
+ | 0.896 | 11.1892 | 414 | 0.9101 |
191
+ | 0.8151 | 11.2703 | 417 | 0.9067 |
192
+ | 0.8352 | 11.3514 | 420 | 0.9060 |
193
+ | 1.0484 | 11.4324 | 423 | 0.9044 |
194
+ | 0.8623 | 11.5135 | 426 | 0.9041 |
195
+ | 0.8942 | 11.5946 | 429 | 0.9022 |
196
+ | 0.8789 | 11.6757 | 432 | 0.9008 |
197
+ | 0.9077 | 11.7568 | 435 | 0.9000 |
198
+ | 0.8 | 11.8378 | 438 | 0.8999 |
199
+ | 0.8116 | 11.9189 | 441 | 0.8985 |
200
+ | 0.8382 | 12.0 | 444 | 0.8970 |
201
+ | 0.8267 | 12.0811 | 447 | 0.8962 |
202
+ | 0.9168 | 12.1622 | 450 | 0.8947 |
203
+ | 0.7878 | 12.2432 | 453 | 0.8949 |
204
+ | 0.8829 | 12.3243 | 456 | 0.8926 |
205
+ | 0.8465 | 12.4054 | 459 | 0.8935 |
206
+ | 0.882 | 12.4865 | 462 | 0.8914 |
207
+ | 0.8265 | 12.5676 | 465 | 0.8896 |
208
+ | 0.8846 | 12.6486 | 468 | 0.8891 |
209
+ | 0.7943 | 12.7297 | 471 | 0.8899 |
210
+ | 0.8846 | 12.8108 | 474 | 0.8883 |
211
+ | 0.8177 | 12.8919 | 477 | 0.8852 |
212
+ | 0.8807 | 12.9730 | 480 | 0.8840 |
213
+ | 0.893 | 13.0541 | 483 | 0.8822 |
214
+ | 0.7497 | 13.1351 | 486 | 0.8799 |
215
+ | 0.8803 | 13.2162 | 489 | 0.8785 |
216
+ | 0.917 | 13.2973 | 492 | 0.8797 |
217
+ | 0.8162 | 13.3784 | 495 | 0.8783 |
218
+ | 0.9659 | 13.4595 | 498 | 0.8799 |
219
+ | 0.8163 | 13.5405 | 501 | 0.8790 |
220
+ | 0.8139 | 13.6216 | 504 | 0.8753 |
221
+ | 0.7451 | 13.7027 | 507 | 0.8737 |
222
+ | 0.7726 | 13.7838 | 510 | 0.8738 |
223
+ | 0.821 | 13.8649 | 513 | 0.8714 |
224
+ | 0.8782 | 13.9459 | 516 | 0.8689 |
225
+ | 0.8563 | 14.0270 | 519 | 0.8700 |
226
+ | 0.8645 | 14.1081 | 522 | 0.8671 |
227
+ | 0.9223 | 14.1892 | 525 | 0.8676 |
228
+ | 0.8396 | 14.2703 | 528 | 0.8682 |
229
+ | 0.8224 | 14.3514 | 531 | 0.8679 |
230
+ | 0.7793 | 14.4324 | 534 | 0.8690 |
231
+ | 0.7911 | 14.5135 | 537 | 0.8650 |
232
+ | 0.7706 | 14.5946 | 540 | 0.8634 |
233
+ | 0.7983 | 14.6757 | 543 | 0.8606 |
234
+ | 0.7409 | 14.7568 | 546 | 0.8611 |
235
+ | 0.8634 | 14.8378 | 549 | 0.8594 |
236
+ | 0.8631 | 14.9189 | 552 | 0.8594 |
237
+ | 0.7624 | 15.0 | 555 | 0.8582 |
238
+ | 0.8568 | 15.0811 | 558 | 0.8568 |
239
+ | 0.8833 | 15.1622 | 561 | 0.8542 |
240
+ | 0.7804 | 15.2432 | 564 | 0.8514 |
241
+ | 0.9113 | 15.3243 | 567 | 0.8526 |
242
+ | 0.72 | 15.4054 | 570 | 0.8511 |
243
+ | 0.8631 | 15.4865 | 573 | 0.8498 |
244
+ | 0.6486 | 15.5676 | 576 | 0.8507 |
245
+ | 0.7745 | 15.6486 | 579 | 0.8507 |
246
+ | 0.8591 | 15.7297 | 582 | 0.8519 |
247
+ | 0.8479 | 15.8108 | 585 | 0.8515 |
248
+ | 0.7879 | 15.8919 | 588 | 0.8486 |
249
+ | 0.7829 | 15.9730 | 591 | 0.8483 |
250
+ | 0.7271 | 16.0541 | 594 | 0.8451 |
251
+ | 0.7425 | 16.1351 | 597 | 0.8464 |
252
+ | 0.7692 | 16.2162 | 600 | 0.8436 |
253
+ | 0.7492 | 16.2973 | 603 | 0.8459 |
254
+ | 0.7085 | 16.3784 | 606 | 0.8440 |
255
+ | 0.7351 | 16.4595 | 609 | 0.8429 |
256
+ | 0.8593 | 16.5405 | 612 | 0.8412 |
257
+ | 0.8999 | 16.6216 | 615 | 0.8400 |
258
+ | 0.8136 | 16.7027 | 618 | 0.8417 |
259
+ | 0.8415 | 16.7838 | 621 | 0.8393 |
260
+ | 0.849 | 16.8649 | 624 | 0.8400 |
261
+ | 0.798 | 16.9459 | 627 | 0.8409 |
262
+ | 0.7581 | 17.0270 | 630 | 0.8393 |
263
+ | 0.7244 | 17.1081 | 633 | 0.8376 |
264
+ | 0.7221 | 17.1892 | 636 | 0.8379 |
265
+ | 0.8085 | 17.2703 | 639 | 0.8363 |
266
+ | 0.7364 | 17.3514 | 642 | 0.8357 |
267
+ | 0.7684 | 17.4324 | 645 | 0.8385 |
268
+ | 0.7323 | 17.5135 | 648 | 0.8330 |
269
+ | 0.8999 | 17.5946 | 651 | 0.8334 |
270
+ | 0.7243 | 17.6757 | 654 | 0.8323 |
271
+ | 0.741 | 17.7568 | 657 | 0.8300 |
272
+ | 0.7529 | 17.8378 | 660 | 0.8290 |
273
+ | 0.9129 | 17.9189 | 663 | 0.8313 |
274
+ | 0.8043 | 18.0 | 666 | 0.8282 |
275
+ | 0.838 | 18.0811 | 669 | 0.8294 |
276
+ | 0.8133 | 18.1622 | 672 | 0.8294 |
277
+ | 0.708 | 18.2432 | 675 | 0.8283 |
278
+ | 0.8017 | 18.3243 | 678 | 0.8289 |
279
+ | 0.9087 | 18.4054 | 681 | 0.8256 |
280
+ | 0.6983 | 18.4865 | 684 | 0.8260 |
281
+ | 0.7684 | 18.5676 | 687 | 0.8251 |
282
+ | 0.6372 | 18.6486 | 690 | 0.8241 |
283
+ | 0.7369 | 18.7297 | 693 | 0.8261 |
284
+ | 0.819 | 18.8108 | 696 | 0.8231 |
285
+ | 0.6806 | 18.8919 | 699 | 0.8244 |
286
+ | 0.8233 | 18.9730 | 702 | 0.8237 |
287
+ | 0.7614 | 19.0541 | 705 | 0.8234 |
288
+ | 0.7862 | 19.1351 | 708 | 0.8212 |
289
+ | 0.6605 | 19.2162 | 711 | 0.8195 |
290
+ | 0.8229 | 19.2973 | 714 | 0.8194 |
291
+ | 0.731 | 19.3784 | 717 | 0.8193 |
292
+ | 0.783 | 19.4595 | 720 | 0.8187 |
293
+ | 0.6495 | 19.5405 | 723 | 0.8185 |
294
+ | 0.8034 | 19.6216 | 726 | 0.8173 |
295
+ | 0.7064 | 19.7027 | 729 | 0.8177 |
296
+ | 0.7276 | 19.7838 | 732 | 0.8156 |
297
+ | 0.8778 | 19.8649 | 735 | 0.8143 |
298
+ | 0.7399 | 19.9459 | 738 | 0.8126 |
299
+ | 0.7399 | 20.0270 | 741 | 0.8139 |
300
+ | 0.7912 | 20.1081 | 744 | 0.8153 |
301
+ | 0.8063 | 20.1892 | 747 | 0.8120 |
302
+ | 0.671 | 20.2703 | 750 | 0.8105 |
303
+ | 0.6342 | 20.3514 | 753 | 0.8113 |
304
+ | 0.8271 | 20.4324 | 756 | 0.8097 |
305
+ | 0.6755 | 20.5135 | 759 | 0.8099 |
306
+ | 0.762 | 20.5946 | 762 | 0.8084 |
307
+ | 0.9197 | 20.6757 | 765 | 0.8079 |
308
+ | 0.7844 | 20.7568 | 768 | 0.8094 |
309
+ | 0.6502 | 20.8378 | 771 | 0.8096 |
310
+ | 0.6877 | 20.9189 | 774 | 0.8078 |
311
+ | 0.7705 | 21.0 | 777 | 0.8074 |
312
+ | 0.6842 | 21.0811 | 780 | 0.8067 |
313
+ | 0.6799 | 21.1622 | 783 | 0.8066 |
314
+ | 0.7344 | 21.2432 | 786 | 0.8039 |
315
+ | 0.6554 | 21.3243 | 789 | 0.8051 |
316
+ | 0.7007 | 21.4054 | 792 | 0.8051 |
317
+ | 0.6292 | 21.4865 | 795 | 0.8040 |
318
+ | 0.7404 | 21.5676 | 798 | 0.8031 |
319
+ | 0.7976 | 21.6486 | 801 | 0.8030 |
320
+ | 1.0023 | 21.7297 | 804 | 0.8045 |
321
+ | 0.6626 | 21.8108 | 807 | 0.8042 |
322
+ | 0.8212 | 21.8919 | 810 | 0.8019 |
323
+ | 0.6982 | 21.9730 | 813 | 0.8029 |
324
+ | 0.7221 | 22.0541 | 816 | 0.8007 |
325
+ | 0.8042 | 22.1351 | 819 | 0.8013 |
326
+ | 0.775 | 22.2162 | 822 | 0.7999 |
327
+ | 0.7614 | 22.2973 | 825 | 0.7971 |
328
+ | 0.7132 | 22.3784 | 828 | 0.7972 |
329
+ | 0.647 | 22.4595 | 831 | 0.8001 |
330
+ | 0.692 | 22.5405 | 834 | 0.7968 |
331
+ | 0.7775 | 22.6216 | 837 | 0.7976 |
332
+ | 0.7764 | 22.7027 | 840 | 0.7980 |
333
+ | 0.6962 | 22.7838 | 843 | 0.7990 |
334
+ | 0.649 | 22.8649 | 846 | 0.7941 |
335
+ | 0.7853 | 22.9459 | 849 | 0.7940 |
336
+ | 0.6808 | 23.0270 | 852 | 0.7959 |
337
+ | 0.7018 | 23.1081 | 855 | 0.7948 |
338
+ | 0.6743 | 23.1892 | 858 | 0.7949 |
339
+ | 0.7512 | 23.2703 | 861 | 0.7926 |
340
+ | 0.7136 | 23.3514 | 864 | 0.7922 |
341
+ | 0.6847 | 23.4324 | 867 | 0.7943 |
342
+ | 0.7309 | 23.5135 | 870 | 0.7910 |
343
+ | 0.9269 | 23.5946 | 873 | 0.7903 |
344
+ | 0.7629 | 23.6757 | 876 | 0.7923 |
345
+ | 0.6551 | 23.7568 | 879 | 0.7896 |
346
+ | 0.7346 | 23.8378 | 882 | 0.7920 |
347
+ | 0.7469 | 23.9189 | 885 | 0.7921 |
348
+ | 0.6343 | 24.0 | 888 | 0.7925 |
349
+ | 0.7232 | 24.0811 | 891 | 0.7905 |
350
+ | 0.714 | 24.1622 | 894 | 0.7885 |
351
+ | 0.6603 | 24.2432 | 897 | 0.7885 |
352
+ | 0.6798 | 24.3243 | 900 | 0.7892 |
353
+ | 0.7163 | 24.4054 | 903 | 0.7889 |
354
+ | 0.8428 | 24.4865 | 906 | 0.7871 |
355
+ | 0.8125 | 24.5676 | 909 | 0.7875 |
356
+ | 0.7566 | 24.6486 | 912 | 0.7895 |
357
+ | 0.7167 | 24.7297 | 915 | 0.7874 |
358
+ | 0.65 | 24.8108 | 918 | 0.7861 |
359
+ | 0.6817 | 24.8919 | 921 | 0.7860 |
360
+ | 0.6185 | 24.9730 | 924 | 0.7879 |
361
+ | 0.7745 | 25.0541 | 927 | 0.7900 |
362
+ | 0.8923 | 25.1351 | 930 | 0.7871 |
363
+ | 0.7395 | 25.2162 | 933 | 0.7834 |
364
+ | 0.7361 | 25.2973 | 936 | 0.7842 |
365
+ | 0.676 | 25.3784 | 939 | 0.7835 |
366
+ | 0.7359 | 25.4595 | 942 | 0.7846 |
367
+ | 0.6533 | 25.5405 | 945 | 0.7829 |
368
+ | 0.6989 | 25.6216 | 948 | 0.7848 |
369
+ | 0.8168 | 25.7027 | 951 | 0.7837 |
370
+ | 0.6519 | 25.7838 | 954 | 0.7826 |
371
+ | 0.6951 | 25.8649 | 957 | 0.7840 |
372
+ | 0.5833 | 25.9459 | 960 | 0.7823 |
373
+ | 0.617 | 26.0270 | 963 | 0.7812 |
374
+ | 0.6778 | 26.1081 | 966 | 0.7829 |
375
+ | 0.6571 | 26.1892 | 969 | 0.7818 |
376
+ | 0.7921 | 26.2703 | 972 | 0.7811 |
377
+ | 0.6962 | 26.3514 | 975 | 0.7807 |
378
+ | 0.8437 | 26.4324 | 978 | 0.7814 |
379
+ | 0.6147 | 26.5135 | 981 | 0.7813 |
380
+ | 0.7263 | 26.5946 | 984 | 0.7803 |
381
+ | 0.6618 | 26.6757 | 987 | 0.7809 |
382
+ | 0.7502 | 26.7568 | 990 | 0.7835 |
383
+ | 0.6094 | 26.8378 | 993 | 0.7813 |
384
+ | 0.6506 | 26.9189 | 996 | 0.7800 |
385
+ | 0.7927 | 27.0 | 999 | 0.7807 |
386
+ | 0.8217 | 27.0811 | 1002 | 0.7816 |
387
+ | 0.5989 | 27.1622 | 1005 | 0.7786 |
388
+ | 0.7204 | 27.2432 | 1008 | 0.7782 |
389
+ | 0.6906 | 27.3243 | 1011 | 0.7785 |
390
+ | 0.7592 | 27.4054 | 1014 | 0.7777 |
391
+ | 0.8692 | 27.4865 | 1017 | 0.7778 |
392
+ | 0.587 | 27.5676 | 1020 | 0.7781 |
393
+ | 0.7139 | 27.6486 | 1023 | 0.7781 |
394
+ | 0.6367 | 27.7297 | 1026 | 0.7774 |
395
+ | 0.6502 | 27.8108 | 1029 | 0.7779 |
396
+ | 0.6956 | 27.8919 | 1032 | 0.7770 |
397
+ | 0.7137 | 27.9730 | 1035 | 0.7784 |
398
+ | 0.7062 | 28.0541 | 1038 | 0.7775 |
399
+ | 0.6568 | 28.1351 | 1041 | 0.7782 |
400
+ | 0.7612 | 28.2162 | 1044 | 0.7766 |
401
+ | 0.7177 | 28.2973 | 1047 | 0.7765 |
402
+ | 0.6537 | 28.3784 | 1050 | 0.7776 |
403
+ | 0.6742 | 28.4595 | 1053 | 0.7773 |
404
+ | 0.724 | 28.5405 | 1056 | 0.7778 |
405
+ | 0.8277 | 28.6216 | 1059 | 0.7763 |
406
+ | 0.609 | 28.7027 | 1062 | 0.7767 |
407
+ | 0.5459 | 28.7838 | 1065 | 0.7781 |
408
+ | 0.8179 | 28.8649 | 1068 | 0.7769 |
409
+ | 0.6419 | 28.9459 | 1071 | 0.7754 |
410
+ | 0.6947 | 29.0270 | 1074 | 0.7745 |
411
+ | 0.71 | 29.1081 | 1077 | 0.7751 |
412
+ | 0.5664 | 29.1892 | 1080 | 0.7747 |
413
+ | 0.7678 | 29.2703 | 1083 | 0.7741 |
414
+ | 0.7829 | 29.3514 | 1086 | 0.7732 |
415
+ | 0.7294 | 29.4324 | 1089 | 0.7729 |
416
+ | 0.679 | 29.5135 | 1092 | 0.7752 |
417
+ | 0.6838 | 29.5946 | 1095 | 0.7739 |
418
+ | 0.6918 | 29.6757 | 1098 | 0.7731 |
419
+ | 0.6563 | 29.7568 | 1101 | 0.7727 |
420
+ | 0.6946 | 29.8378 | 1104 | 0.7742 |
421
+ | 0.6884 | 29.9189 | 1107 | 0.7729 |
422
+ | 0.6851 | 30.0 | 1110 | 0.7744 |
423
+ | 0.5865 | 30.0811 | 1113 | 0.7744 |
424
+ | 0.6906 | 30.1622 | 1116 | 0.7746 |
425
+ | 0.6716 | 30.2432 | 1119 | 0.7751 |
426
+ | 0.7492 | 30.3243 | 1122 | 0.7736 |
427
+ | 0.6608 | 30.4054 | 1125 | 0.7709 |
428
+ | 0.6577 | 30.4865 | 1128 | 0.7710 |
429
+ | 0.7475 | 30.5676 | 1131 | 0.7722 |
430
+ | 0.7643 | 30.6486 | 1134 | 0.7694 |
431
+ | 0.8186 | 30.7297 | 1137 | 0.7711 |
432
+ | 0.7465 | 30.8108 | 1140 | 0.7724 |
433
+ | 0.6621 | 30.8919 | 1143 | 0.7728 |
434
+ | 0.579 | 30.9730 | 1146 | 0.7713 |
435
+ | 0.6014 | 31.0541 | 1149 | 0.7707 |
436
+ | 0.7663 | 31.1351 | 1152 | 0.7697 |
437
+ | 0.6609 | 31.2162 | 1155 | 0.7719 |
438
+ | 0.7984 | 31.2973 | 1158 | 0.7715 |
439
+ | 0.7166 | 31.3784 | 1161 | 0.7715 |
440
+ | 0.7276 | 31.4595 | 1164 | 0.7712 |
441
+ | 0.6179 | 31.5405 | 1167 | 0.7681 |
442
+ | 0.621 | 31.6216 | 1170 | 0.7707 |
443
+ | 0.743 | 31.7027 | 1173 | 0.7711 |
444
+ | 0.6733 | 31.7838 | 1176 | 0.7706 |
445
+ | 0.6066 | 31.8649 | 1179 | 0.7716 |
446
+ | 0.6787 | 31.9459 | 1182 | 0.7713 |
447
+ | 0.812 | 32.0270 | 1185 | 0.7708 |
448
+ | 0.7022 | 32.1081 | 1188 | 0.7695 |
449
+ | 0.6709 | 32.1892 | 1191 | 0.7721 |
450
+ | 0.5536 | 32.2703 | 1194 | 0.7716 |
451
+ | 0.7462 | 32.3514 | 1197 | 0.7693 |
452
+ | 0.7817 | 32.4324 | 1200 | 0.7686 |
453
+ | 0.7624 | 32.5135 | 1203 | 0.7719 |
454
+ | 0.6346 | 32.5946 | 1206 | 0.7696 |
455
+ | 0.6284 | 32.6757 | 1209 | 0.7717 |
456
+ | 0.654 | 32.7568 | 1212 | 0.7708 |
457
+ | 0.6406 | 32.8378 | 1215 | 0.7707 |
458
+ | 0.752 | 32.9189 | 1218 | 0.7690 |
459
+ | 0.5956 | 33.0 | 1221 | 0.7713 |
460
+ | 0.6772 | 33.0811 | 1224 | 0.7701 |
461
+ | 0.5034 | 33.1622 | 1227 | 0.7693 |
462
+ | 0.6292 | 33.2432 | 1230 | 0.7687 |
463
+ | 0.7374 | 33.3243 | 1233 | 0.7693 |
464
+ | 0.5385 | 33.4054 | 1236 | 0.7678 |
465
+ | 0.7198 | 33.4865 | 1239 | 0.7686 |
466
+ | 0.7642 | 33.5676 | 1242 | 0.7695 |
467
+ | 0.6984 | 33.6486 | 1245 | 0.7682 |
468
+ | 0.776 | 33.7297 | 1248 | 0.7694 |
469
+ | 0.7073 | 33.8108 | 1251 | 0.7700 |
470
+ | 0.8016 | 33.8919 | 1254 | 0.7683 |
471
+ | 0.7812 | 33.9730 | 1257 | 0.7662 |
472
+ | 0.578 | 34.0541 | 1260 | 0.7681 |
473
+ | 0.7074 | 34.1351 | 1263 | 0.7679 |
474
+ | 0.6142 | 34.2162 | 1266 | 0.7669 |
475
+ | 0.6469 | 34.2973 | 1269 | 0.7675 |
476
+ | 0.7334 | 34.3784 | 1272 | 0.7669 |
477
+ | 0.7086 | 34.4595 | 1275 | 0.7675 |
478
+ | 0.6302 | 34.5405 | 1278 | 0.7684 |
479
+ | 0.6576 | 34.6216 | 1281 | 0.7687 |
480
+ | 0.7496 | 34.7027 | 1284 | 0.7667 |
481
+ | 0.7836 | 34.7838 | 1287 | 0.7671 |
482
+ | 0.6405 | 34.8649 | 1290 | 0.7679 |
483
+ | 0.6088 | 34.9459 | 1293 | 0.7681 |
484
+ | 0.7511 | 35.0270 | 1296 | 0.7697 |
485
+ | 0.7591 | 35.1081 | 1299 | 0.7680 |
486
+ | 0.6207 | 35.1892 | 1302 | 0.7689 |
487
+ | 0.651 | 35.2703 | 1305 | 0.7660 |
488
+ | 0.6758 | 35.3514 | 1308 | 0.7694 |
489
+ | 0.7277 | 35.4324 | 1311 | 0.7680 |
490
+ | 0.7118 | 35.5135 | 1314 | 0.7663 |
491
+ | 0.706 | 35.5946 | 1317 | 0.7669 |
492
+ | 0.6595 | 35.6757 | 1320 | 0.7678 |
493
+ | 0.7253 | 35.7568 | 1323 | 0.7688 |
494
+ | 0.582 | 35.8378 | 1326 | 0.7694 |
495
+ | 0.7094 | 35.9189 | 1329 | 0.7684 |
496
+ | 0.7042 | 36.0 | 1332 | 0.7686 |
497
+ | 0.6486 | 36.0811 | 1335 | 0.7664 |
498
+ | 0.6691 | 36.1622 | 1338 | 0.7678 |
499
+ | 0.7937 | 36.2432 | 1341 | 0.7672 |
500
+ | 0.6343 | 36.3243 | 1344 | 0.7670 |
501
+ | 0.76 | 36.4054 | 1347 | 0.7674 |
502
+ | 0.5929 | 36.4865 | 1350 | 0.7690 |
503
+ | 0.6485 | 36.5676 | 1353 | 0.7674 |
504
+ | 0.7451 | 36.6486 | 1356 | 0.7663 |
505
+ | 0.6367 | 36.7297 | 1359 | 0.7691 |
506
+ | 0.6192 | 36.8108 | 1362 | 0.7656 |
507
+ | 0.7042 | 36.8919 | 1365 | 0.7682 |
508
+ | 0.6886 | 36.9730 | 1368 | 0.7682 |
509
+ | 0.8111 | 37.0541 | 1371 | 0.7690 |
510
+ | 0.6443 | 37.1351 | 1374 | 0.7690 |
511
+ | 0.6387 | 37.2162 | 1377 | 0.7679 |
512
+ | 0.7861 | 37.2973 | 1380 | 0.7668 |
513
+ | 0.6397 | 37.3784 | 1383 | 0.7670 |
514
+ | 0.6467 | 37.4595 | 1386 | 0.7688 |
515
+ | 0.6977 | 37.5405 | 1389 | 0.7699 |
516
+ | 0.7833 | 37.6216 | 1392 | 0.7672 |
517
+ | 0.6164 | 37.7027 | 1395 | 0.7675 |
518
+ | 0.7457 | 37.7838 | 1398 | 0.7690 |
519
+ | 0.6477 | 37.8649 | 1401 | 0.7681 |
520
+ | 0.7251 | 37.9459 | 1404 | 0.7674 |
521
+ | 0.5642 | 38.0270 | 1407 | 0.7672 |
522
+ | 0.786 | 38.1081 | 1410 | 0.7660 |
523
+ | 0.7014 | 38.1892 | 1413 | 0.7683 |
524
+ | 0.7747 | 38.2703 | 1416 | 0.7659 |
525
+ | 0.631 | 38.3514 | 1419 | 0.7675 |
526
+ | 0.6787 | 38.4324 | 1422 | 0.7671 |
527
+ | 0.6388 | 38.5135 | 1425 | 0.7684 |
528
+ | 0.5745 | 38.5946 | 1428 | 0.7666 |
529
+ | 0.6253 | 38.6757 | 1431 | 0.7668 |
530
+ | 0.6611 | 38.7568 | 1434 | 0.7688 |
531
+ | 0.6129 | 38.8378 | 1437 | 0.7666 |
532
+ | 0.7588 | 38.9189 | 1440 | 0.7670 |
533
+ | 0.7615 | 39.0 | 1443 | 0.7664 |
534
+ | 0.7449 | 39.0811 | 1446 | 0.7674 |
535
+ | 0.7012 | 39.1622 | 1449 | 0.7666 |
536
+ | 0.788 | 39.2432 | 1452 | 0.7676 |
537
+ | 0.5958 | 39.3243 | 1455 | 0.7666 |
538
+ | 0.675 | 39.4054 | 1458 | 0.7670 |
539
+ | 0.6761 | 39.4865 | 1461 | 0.7676 |
540
+ | 0.7046 | 39.5676 | 1464 | 0.7669 |
541
+ | 0.6198 | 39.6486 | 1467 | 0.7673 |
542
+ | 0.6646 | 39.7297 | 1470 | 0.7670 |
543
+ | 0.6161 | 39.8108 | 1473 | 0.7677 |
544
+ | 0.7152 | 39.8919 | 1476 | 0.7676 |
545
+ | 0.7365 | 39.9730 | 1479 | 0.7671 |
546
+
547
+
548
+ ### Framework versions
549
+
550
+ - PEFT 0.13.2
551
+ - Transformers 4.46.0
552
+ - Pytorch 2.4.0
553
+ - Datasets 3.0.1
554
+ - Tokenizers 0.20.0
adapter_config.json ADDED
@@ -0,0 +1,34 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alpha_pattern": {},
3
+ "auto_mapping": null,
4
+ "base_model_name_or_path": "Qwen/Qwen2.5-0.5B-Instruct",
5
+ "bias": "none",
6
+ "fan_in_fan_out": false,
7
+ "inference_mode": true,
8
+ "init_lora_weights": true,
9
+ "layer_replication": null,
10
+ "layers_pattern": null,
11
+ "layers_to_transform": null,
12
+ "loftq_config": {},
13
+ "lora_alpha": 32,
14
+ "lora_dropout": 0.01,
15
+ "megatron_config": null,
16
+ "megatron_core": "megatron.core",
17
+ "modules_to_save": null,
18
+ "peft_type": "LORA",
19
+ "r": 16,
20
+ "rank_pattern": {},
21
+ "revision": null,
22
+ "target_modules": [
23
+ "q_proj",
24
+ "gate_proj",
25
+ "k_proj",
26
+ "down_proj",
27
+ "v_proj",
28
+ "up_proj",
29
+ "o_proj"
30
+ ],
31
+ "task_type": "CAUSAL_LM",
32
+ "use_dora": false,
33
+ "use_rslora": false
34
+ }
adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:42dd600bcf09214b61860789a323c0c0985fdb5f62f94293845b5d6fae9663bd
3
+ size 35237104
added_tokens.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "</tool_call>": 151658,
3
+ "<tool_call>": 151657,
4
+ "<|box_end|>": 151649,
5
+ "<|box_start|>": 151648,
6
+ "<|endoftext|>": 151643,
7
+ "<|file_sep|>": 151664,
8
+ "<|fim_middle|>": 151660,
9
+ "<|fim_pad|>": 151662,
10
+ "<|fim_prefix|>": 151659,
11
+ "<|fim_suffix|>": 151661,
12
+ "<|im_end|>": 151645,
13
+ "<|im_start|>": 151644,
14
+ "<|image_pad|>": 151655,
15
+ "<|object_ref_end|>": 151647,
16
+ "<|object_ref_start|>": 151646,
17
+ "<|quad_end|>": 151651,
18
+ "<|quad_start|>": 151650,
19
+ "<|repo_name|>": 151663,
20
+ "<|video_pad|>": 151656,
21
+ "<|vision_end|>": 151653,
22
+ "<|vision_pad|>": 151654,
23
+ "<|vision_start|>": 151652
24
+ }
merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
special_tokens_map.json ADDED
@@ -0,0 +1,31 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "<|im_start|>",
4
+ "<|im_end|>",
5
+ "<|object_ref_start|>",
6
+ "<|object_ref_end|>",
7
+ "<|box_start|>",
8
+ "<|box_end|>",
9
+ "<|quad_start|>",
10
+ "<|quad_end|>",
11
+ "<|vision_start|>",
12
+ "<|vision_end|>",
13
+ "<|vision_pad|>",
14
+ "<|image_pad|>",
15
+ "<|video_pad|>"
16
+ ],
17
+ "eos_token": {
18
+ "content": "<|im_end|>",
19
+ "lstrip": false,
20
+ "normalized": false,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ },
24
+ "pad_token": {
25
+ "content": "<|endoftext|>",
26
+ "lstrip": false,
27
+ "normalized": false,
28
+ "rstrip": false,
29
+ "single_word": false
30
+ }
31
+ }
tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3511c9b326fec6a1b23dbf63767a672c911fbee5b2145ad0099a5bd8aff458b5
3
+ size 11421994
tokenizer_config.json ADDED
@@ -0,0 +1,207 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": false,
3
+ "add_prefix_space": false,
4
+ "added_tokens_decoder": {
5
+ "151643": {
6
+ "content": "<|endoftext|>",
7
+ "lstrip": false,
8
+ "normalized": false,
9
+ "rstrip": false,
10
+ "single_word": false,
11
+ "special": true
12
+ },
13
+ "151644": {
14
+ "content": "<|im_start|>",
15
+ "lstrip": false,
16
+ "normalized": false,
17
+ "rstrip": false,
18
+ "single_word": false,
19
+ "special": true
20
+ },
21
+ "151645": {
22
+ "content": "<|im_end|>",
23
+ "lstrip": false,
24
+ "normalized": false,
25
+ "rstrip": false,
26
+ "single_word": false,
27
+ "special": true
28
+ },
29
+ "151646": {
30
+ "content": "<|object_ref_start|>",
31
+ "lstrip": false,
32
+ "normalized": false,
33
+ "rstrip": false,
34
+ "single_word": false,
35
+ "special": true
36
+ },
37
+ "151647": {
38
+ "content": "<|object_ref_end|>",
39
+ "lstrip": false,
40
+ "normalized": false,
41
+ "rstrip": false,
42
+ "single_word": false,
43
+ "special": true
44
+ },
45
+ "151648": {
46
+ "content": "<|box_start|>",
47
+ "lstrip": false,
48
+ "normalized": false,
49
+ "rstrip": false,
50
+ "single_word": false,
51
+ "special": true
52
+ },
53
+ "151649": {
54
+ "content": "<|box_end|>",
55
+ "lstrip": false,
56
+ "normalized": false,
57
+ "rstrip": false,
58
+ "single_word": false,
59
+ "special": true
60
+ },
61
+ "151650": {
62
+ "content": "<|quad_start|>",
63
+ "lstrip": false,
64
+ "normalized": false,
65
+ "rstrip": false,
66
+ "single_word": false,
67
+ "special": true
68
+ },
69
+ "151651": {
70
+ "content": "<|quad_end|>",
71
+ "lstrip": false,
72
+ "normalized": false,
73
+ "rstrip": false,
74
+ "single_word": false,
75
+ "special": true
76
+ },
77
+ "151652": {
78
+ "content": "<|vision_start|>",
79
+ "lstrip": false,
80
+ "normalized": false,
81
+ "rstrip": false,
82
+ "single_word": false,
83
+ "special": true
84
+ },
85
+ "151653": {
86
+ "content": "<|vision_end|>",
87
+ "lstrip": false,
88
+ "normalized": false,
89
+ "rstrip": false,
90
+ "single_word": false,
91
+ "special": true
92
+ },
93
+ "151654": {
94
+ "content": "<|vision_pad|>",
95
+ "lstrip": false,
96
+ "normalized": false,
97
+ "rstrip": false,
98
+ "single_word": false,
99
+ "special": true
100
+ },
101
+ "151655": {
102
+ "content": "<|image_pad|>",
103
+ "lstrip": false,
104
+ "normalized": false,
105
+ "rstrip": false,
106
+ "single_word": false,
107
+ "special": true
108
+ },
109
+ "151656": {
110
+ "content": "<|video_pad|>",
111
+ "lstrip": false,
112
+ "normalized": false,
113
+ "rstrip": false,
114
+ "single_word": false,
115
+ "special": true
116
+ },
117
+ "151657": {
118
+ "content": "<tool_call>",
119
+ "lstrip": false,
120
+ "normalized": false,
121
+ "rstrip": false,
122
+ "single_word": false,
123
+ "special": false
124
+ },
125
+ "151658": {
126
+ "content": "</tool_call>",
127
+ "lstrip": false,
128
+ "normalized": false,
129
+ "rstrip": false,
130
+ "single_word": false,
131
+ "special": false
132
+ },
133
+ "151659": {
134
+ "content": "<|fim_prefix|>",
135
+ "lstrip": false,
136
+ "normalized": false,
137
+ "rstrip": false,
138
+ "single_word": false,
139
+ "special": false
140
+ },
141
+ "151660": {
142
+ "content": "<|fim_middle|>",
143
+ "lstrip": false,
144
+ "normalized": false,
145
+ "rstrip": false,
146
+ "single_word": false,
147
+ "special": false
148
+ },
149
+ "151661": {
150
+ "content": "<|fim_suffix|>",
151
+ "lstrip": false,
152
+ "normalized": false,
153
+ "rstrip": false,
154
+ "single_word": false,
155
+ "special": false
156
+ },
157
+ "151662": {
158
+ "content": "<|fim_pad|>",
159
+ "lstrip": false,
160
+ "normalized": false,
161
+ "rstrip": false,
162
+ "single_word": false,
163
+ "special": false
164
+ },
165
+ "151663": {
166
+ "content": "<|repo_name|>",
167
+ "lstrip": false,
168
+ "normalized": false,
169
+ "rstrip": false,
170
+ "single_word": false,
171
+ "special": false
172
+ },
173
+ "151664": {
174
+ "content": "<|file_sep|>",
175
+ "lstrip": false,
176
+ "normalized": false,
177
+ "rstrip": false,
178
+ "single_word": false,
179
+ "special": false
180
+ }
181
+ },
182
+ "additional_special_tokens": [
183
+ "<|im_start|>",
184
+ "<|im_end|>",
185
+ "<|object_ref_start|>",
186
+ "<|object_ref_end|>",
187
+ "<|box_start|>",
188
+ "<|box_end|>",
189
+ "<|quad_start|>",
190
+ "<|quad_end|>",
191
+ "<|vision_start|>",
192
+ "<|vision_end|>",
193
+ "<|vision_pad|>",
194
+ "<|image_pad|>",
195
+ "<|video_pad|>"
196
+ ],
197
+ "bos_token": null,
198
+ "chat_template": "{%- if tools %}\n {{- '<|im_start|>system\\n' }}\n {%- if messages[0]['role'] == 'system' %}\n {{- messages[0]['content'] }}\n {%- else %}\n {{- 'You are Qwen, created by Alibaba Cloud. You are a helpful assistant.' }}\n {%- endif %}\n {{- \"\\n\\n# Tools\\n\\nYou may call one or more functions to assist with the user query.\\n\\nYou are provided with function signatures within <tools></tools> XML tags:\\n<tools>\" }}\n {%- for tool in tools %}\n {{- \"\\n\" }}\n {{- tool | tojson }}\n {%- endfor %}\n {{- \"\\n</tools>\\n\\nFor each function call, return a json object with function name and arguments within <tool_call></tool_call> XML tags:\\n<tool_call>\\n{\\\"name\\\": <function-name>, \\\"arguments\\\": <args-json-object>}\\n</tool_call><|im_end|>\\n\" }}\n{%- else %}\n {%- if messages[0]['role'] == 'system' %}\n {{- '<|im_start|>system\\n' + messages[0]['content'] + '<|im_end|>\\n' }}\n {%- else %}\n {{- '<|im_start|>system\\nYou are Qwen, created by Alibaba Cloud. You are a helpful assistant.<|im_end|>\\n' }}\n {%- endif %}\n{%- endif %}\n{%- for message in messages %}\n {%- if (message.role == \"user\") or (message.role == \"system\" and not loop.first) or (message.role == \"assistant\" and not message.tool_calls) %}\n {{- '<|im_start|>' + message.role + '\\n' + message.content + '<|im_end|>' + '\\n' }}\n {%- elif message.role == \"assistant\" %}\n {{- '<|im_start|>' + message.role }}\n {%- if message.content %}\n {{- '\\n' + message.content }}\n {%- endif %}\n {%- for tool_call in message.tool_calls %}\n {%- if tool_call.function is defined %}\n {%- set tool_call = tool_call.function %}\n {%- endif %}\n {{- '\\n<tool_call>\\n{\"name\": \"' }}\n {{- tool_call.name }}\n {{- '\", \"arguments\": ' }}\n {{- tool_call.arguments | tojson }}\n {{- '}\\n</tool_call>' }}\n {%- endfor %}\n {{- '<|im_end|>\\n' }}\n {%- elif message.role == \"tool\" %}\n {%- if (loop.index0 == 0) or (messages[loop.index0 - 1].role != \"tool\") %}\n {{- '<|im_start|>user' }}\n {%- endif %}\n {{- '\\n<tool_response>\\n' }}\n {{- message.content }}\n {{- '\\n</tool_response>' }}\n {%- if loop.last or (messages[loop.index0 + 1].role != \"tool\") %}\n {{- '<|im_end|>\\n' }}\n {%- endif %}\n {%- endif %}\n{%- endfor %}\n{%- if add_generation_prompt %}\n {{- '<|im_start|>assistant\\n' }}\n{%- endif %}\n",
199
+ "clean_up_tokenization_spaces": false,
200
+ "eos_token": "<|im_end|>",
201
+ "errors": "replace",
202
+ "model_max_length": 300,
203
+ "pad_token": "<|endoftext|>",
204
+ "split_special_tokens": false,
205
+ "tokenizer_class": "Qwen2Tokenizer",
206
+ "unk_token": null
207
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:529b53487bf15e921f6f6c93b42bd941f300bef1a350342fb7169de7e00b6228
3
+ size 5496
vocab.json ADDED
The diff for this file is too large to render. See raw diff