sujithatz/testfinbot-quen2.5-lora-merged_adapter
Browse files- .gitattributes +1 -0
- README.md +554 -0
- adapter_config.json +34 -0
- adapter_model.safetensors +3 -0
- added_tokens.json +24 -0
- merges.txt +0 -0
- special_tokens_map.json +31 -0
- tokenizer.json +3 -0
- tokenizer_config.json +207 -0
- training_args.bin +3 -0
- vocab.json +0 -0
.gitattributes
CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
|
|
|
33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
36 |
+
tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
README.md
ADDED
@@ -0,0 +1,554 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
base_model: Qwen/Qwen2.5-0.5B-Instruct
|
3 |
+
library_name: peft
|
4 |
+
license: apache-2.0
|
5 |
+
tags:
|
6 |
+
- trl
|
7 |
+
- sft
|
8 |
+
- generated_from_trainer
|
9 |
+
model-index:
|
10 |
+
- name: quen_2.5_qlora
|
11 |
+
results: []
|
12 |
+
---
|
13 |
+
|
14 |
+
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
|
15 |
+
should probably proofread and complete it, then remove this comment. -->
|
16 |
+
|
17 |
+
# quen_2.5_qlora
|
18 |
+
|
19 |
+
This model is a fine-tuned version of [Qwen/Qwen2.5-0.5B-Instruct](https://huggingface.co/Qwen/Qwen2.5-0.5B-Instruct) on an unknown dataset.
|
20 |
+
It achieves the following results on the evaluation set:
|
21 |
+
- Loss: 0.7671
|
22 |
+
|
23 |
+
## Model description
|
24 |
+
|
25 |
+
More information needed
|
26 |
+
|
27 |
+
## Intended uses & limitations
|
28 |
+
|
29 |
+
More information needed
|
30 |
+
|
31 |
+
## Training and evaluation data
|
32 |
+
|
33 |
+
More information needed
|
34 |
+
|
35 |
+
## Training procedure
|
36 |
+
|
37 |
+
### Training hyperparameters
|
38 |
+
|
39 |
+
The following hyperparameters were used during training:
|
40 |
+
- learning_rate: 5e-06
|
41 |
+
- train_batch_size: 8
|
42 |
+
- eval_batch_size: 8
|
43 |
+
- seed: 0
|
44 |
+
- optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
|
45 |
+
- lr_scheduler_type: cosine
|
46 |
+
- lr_scheduler_warmup_ratio: 0.01
|
47 |
+
- num_epochs: 40
|
48 |
+
|
49 |
+
### Training results
|
50 |
+
|
51 |
+
| Training Loss | Epoch | Step | Validation Loss |
|
52 |
+
|:-------------:|:-------:|:----:|:---------------:|
|
53 |
+
| 3.1118 | 0.0811 | 3 | 3.2024 |
|
54 |
+
| 3.0902 | 0.1622 | 6 | 3.2023 |
|
55 |
+
| 2.9193 | 0.2432 | 9 | 3.1908 |
|
56 |
+
| 3.0231 | 0.3243 | 12 | 3.1700 |
|
57 |
+
| 3.0429 | 0.4054 | 15 | 3.1323 |
|
58 |
+
| 2.9739 | 0.4865 | 18 | 3.0926 |
|
59 |
+
| 2.9118 | 0.5676 | 21 | 3.0477 |
|
60 |
+
| 2.8564 | 0.6486 | 24 | 2.9984 |
|
61 |
+
| 2.9689 | 0.7297 | 27 | 2.9489 |
|
62 |
+
| 2.8155 | 0.8108 | 30 | 2.8946 |
|
63 |
+
| 2.8495 | 0.8919 | 33 | 2.8411 |
|
64 |
+
| 2.7311 | 0.9730 | 36 | 2.7843 |
|
65 |
+
| 2.6327 | 1.0541 | 39 | 2.7375 |
|
66 |
+
| 2.6728 | 1.1351 | 42 | 2.6747 |
|
67 |
+
| 2.5236 | 1.2162 | 45 | 2.6151 |
|
68 |
+
| 2.5009 | 1.2973 | 48 | 2.5595 |
|
69 |
+
| 2.4129 | 1.3784 | 51 | 2.5030 |
|
70 |
+
| 2.3506 | 1.4595 | 54 | 2.4369 |
|
71 |
+
| 2.3542 | 1.5405 | 57 | 2.3803 |
|
72 |
+
| 2.237 | 1.6216 | 60 | 2.3189 |
|
73 |
+
| 2.3302 | 1.7027 | 63 | 2.2589 |
|
74 |
+
| 2.1679 | 1.7838 | 66 | 2.2005 |
|
75 |
+
| 2.1406 | 1.8649 | 69 | 2.1355 |
|
76 |
+
| 2.0373 | 1.9459 | 72 | 2.0689 |
|
77 |
+
| 1.9979 | 2.0270 | 75 | 2.0098 |
|
78 |
+
| 2.0711 | 2.1081 | 78 | 1.9351 |
|
79 |
+
| 1.8819 | 2.1892 | 81 | 1.8753 |
|
80 |
+
| 1.7832 | 2.2703 | 84 | 1.8243 |
|
81 |
+
| 1.7165 | 2.3514 | 87 | 1.7668 |
|
82 |
+
| 1.6765 | 2.4324 | 90 | 1.7111 |
|
83 |
+
| 1.6243 | 2.5135 | 93 | 1.6613 |
|
84 |
+
| 1.6819 | 2.5946 | 96 | 1.6154 |
|
85 |
+
| 1.6006 | 2.6757 | 99 | 1.5679 |
|
86 |
+
| 1.4845 | 2.7568 | 102 | 1.5229 |
|
87 |
+
| 1.4014 | 2.8378 | 105 | 1.4728 |
|
88 |
+
| 1.452 | 2.9189 | 108 | 1.4313 |
|
89 |
+
| 1.3642 | 3.0 | 111 | 1.3985 |
|
90 |
+
| 1.3571 | 3.0811 | 114 | 1.3647 |
|
91 |
+
| 1.4434 | 3.1622 | 117 | 1.3416 |
|
92 |
+
| 1.3829 | 3.2432 | 120 | 1.3206 |
|
93 |
+
| 1.3319 | 3.3243 | 123 | 1.3000 |
|
94 |
+
| 1.2262 | 3.4054 | 126 | 1.2851 |
|
95 |
+
| 1.2693 | 3.4865 | 129 | 1.2711 |
|
96 |
+
| 1.264 | 3.5676 | 132 | 1.2588 |
|
97 |
+
| 1.1851 | 3.6486 | 135 | 1.2470 |
|
98 |
+
| 1.2656 | 3.7297 | 138 | 1.2380 |
|
99 |
+
| 1.205 | 3.8108 | 141 | 1.2244 |
|
100 |
+
| 1.2115 | 3.8919 | 144 | 1.2109 |
|
101 |
+
| 1.1366 | 3.9730 | 147 | 1.1997 |
|
102 |
+
| 1.1794 | 4.0541 | 150 | 1.1929 |
|
103 |
+
| 1.171 | 4.1351 | 153 | 1.1805 |
|
104 |
+
| 1.0875 | 4.2162 | 156 | 1.1714 |
|
105 |
+
| 1.1192 | 4.2973 | 159 | 1.1619 |
|
106 |
+
| 1.1082 | 4.3784 | 162 | 1.1548 |
|
107 |
+
| 1.0923 | 4.4595 | 165 | 1.1465 |
|
108 |
+
| 1.1731 | 4.5405 | 168 | 1.1397 |
|
109 |
+
| 1.2138 | 4.6216 | 171 | 1.1315 |
|
110 |
+
| 1.2322 | 4.7027 | 174 | 1.1223 |
|
111 |
+
| 1.0893 | 4.7838 | 177 | 1.1179 |
|
112 |
+
| 1.1414 | 4.8649 | 180 | 1.1116 |
|
113 |
+
| 1.0242 | 4.9459 | 183 | 1.1058 |
|
114 |
+
| 1.0929 | 5.0270 | 186 | 1.0982 |
|
115 |
+
| 1.0967 | 5.1081 | 189 | 1.0916 |
|
116 |
+
| 1.0666 | 5.1892 | 192 | 1.0866 |
|
117 |
+
| 1.1382 | 5.2703 | 195 | 1.0851 |
|
118 |
+
| 1.0709 | 5.3514 | 198 | 1.0781 |
|
119 |
+
| 1.0904 | 5.4324 | 201 | 1.0742 |
|
120 |
+
| 1.0521 | 5.5135 | 204 | 1.0699 |
|
121 |
+
| 0.9745 | 5.5946 | 207 | 1.0641 |
|
122 |
+
| 1.1268 | 5.6757 | 210 | 1.0624 |
|
123 |
+
| 1.0437 | 5.7568 | 213 | 1.0588 |
|
124 |
+
| 1.0124 | 5.8378 | 216 | 1.0578 |
|
125 |
+
| 1.0217 | 5.9189 | 219 | 1.0511 |
|
126 |
+
| 1.0656 | 6.0 | 222 | 1.0475 |
|
127 |
+
| 1.0841 | 6.0811 | 225 | 1.0454 |
|
128 |
+
| 1.1352 | 6.1622 | 228 | 1.0440 |
|
129 |
+
| 1.0402 | 6.2432 | 231 | 1.0407 |
|
130 |
+
| 1.0042 | 6.3243 | 234 | 1.0326 |
|
131 |
+
| 0.9738 | 6.4054 | 237 | 1.0295 |
|
132 |
+
| 0.9671 | 6.4865 | 240 | 1.0292 |
|
133 |
+
| 1.0515 | 6.5676 | 243 | 1.0282 |
|
134 |
+
| 0.9551 | 6.6486 | 246 | 1.0246 |
|
135 |
+
| 1.0084 | 6.7297 | 249 | 1.0210 |
|
136 |
+
| 0.9964 | 6.8108 | 252 | 1.0188 |
|
137 |
+
| 0.966 | 6.8919 | 255 | 1.0146 |
|
138 |
+
| 1.0284 | 6.9730 | 258 | 1.0113 |
|
139 |
+
| 1.0367 | 7.0541 | 261 | 1.0103 |
|
140 |
+
| 0.9015 | 7.1351 | 264 | 1.0056 |
|
141 |
+
| 0.9097 | 7.2162 | 267 | 1.0043 |
|
142 |
+
| 0.9598 | 7.2973 | 270 | 0.9998 |
|
143 |
+
| 1.1724 | 7.3784 | 273 | 0.9980 |
|
144 |
+
| 0.9607 | 7.4595 | 276 | 0.9947 |
|
145 |
+
| 0.9749 | 7.5405 | 279 | 0.9960 |
|
146 |
+
| 0.9704 | 7.6216 | 282 | 0.9924 |
|
147 |
+
| 0.9855 | 7.7027 | 285 | 0.9899 |
|
148 |
+
| 0.9368 | 7.7838 | 288 | 0.9853 |
|
149 |
+
| 0.9812 | 7.8649 | 291 | 0.9856 |
|
150 |
+
| 0.9809 | 7.9459 | 294 | 0.9809 |
|
151 |
+
| 0.9337 | 8.0270 | 297 | 0.9824 |
|
152 |
+
| 0.8848 | 8.1081 | 300 | 0.9759 |
|
153 |
+
| 0.876 | 8.1892 | 303 | 0.9751 |
|
154 |
+
| 0.97 | 8.2703 | 306 | 0.9734 |
|
155 |
+
| 1.1309 | 8.3514 | 309 | 0.9726 |
|
156 |
+
| 0.9012 | 8.4324 | 312 | 0.9709 |
|
157 |
+
| 1.0355 | 8.5135 | 315 | 0.9681 |
|
158 |
+
| 0.9047 | 8.5946 | 318 | 0.9663 |
|
159 |
+
| 0.9417 | 8.6757 | 321 | 0.9622 |
|
160 |
+
| 0.9266 | 8.7568 | 324 | 0.9613 |
|
161 |
+
| 0.8873 | 8.8378 | 327 | 0.9625 |
|
162 |
+
| 0.9364 | 8.9189 | 330 | 0.9598 |
|
163 |
+
| 0.9885 | 9.0 | 333 | 0.9539 |
|
164 |
+
| 0.91 | 9.0811 | 336 | 0.9538 |
|
165 |
+
| 1.0834 | 9.1622 | 339 | 0.9522 |
|
166 |
+
| 0.8694 | 9.2432 | 342 | 0.9502 |
|
167 |
+
| 1.0554 | 9.3243 | 345 | 0.9498 |
|
168 |
+
| 0.9123 | 9.4054 | 348 | 0.9484 |
|
169 |
+
| 0.9542 | 9.4865 | 351 | 0.9449 |
|
170 |
+
| 0.9287 | 9.5676 | 354 | 0.9417 |
|
171 |
+
| 0.8435 | 9.6486 | 357 | 0.9406 |
|
172 |
+
| 0.8419 | 9.7297 | 360 | 0.9374 |
|
173 |
+
| 0.8265 | 9.8108 | 363 | 0.9351 |
|
174 |
+
| 0.8677 | 9.8919 | 366 | 0.9343 |
|
175 |
+
| 0.9989 | 9.9730 | 369 | 0.9341 |
|
176 |
+
| 0.8331 | 10.0541 | 372 | 0.9329 |
|
177 |
+
| 0.8823 | 10.1351 | 375 | 0.9312 |
|
178 |
+
| 0.9312 | 10.2162 | 378 | 0.9299 |
|
179 |
+
| 0.8736 | 10.2973 | 381 | 0.9279 |
|
180 |
+
| 1.0039 | 10.3784 | 384 | 0.9259 |
|
181 |
+
| 0.8728 | 10.4595 | 387 | 0.9249 |
|
182 |
+
| 0.9946 | 10.5405 | 390 | 0.9216 |
|
183 |
+
| 0.8747 | 10.6216 | 393 | 0.9196 |
|
184 |
+
| 0.9043 | 10.7027 | 396 | 0.9183 |
|
185 |
+
| 0.8777 | 10.7838 | 399 | 0.9163 |
|
186 |
+
| 0.8023 | 10.8649 | 402 | 0.9160 |
|
187 |
+
| 0.9002 | 10.9459 | 405 | 0.9128 |
|
188 |
+
| 0.8568 | 11.0270 | 408 | 0.9146 |
|
189 |
+
| 0.9145 | 11.1081 | 411 | 0.9124 |
|
190 |
+
| 0.896 | 11.1892 | 414 | 0.9101 |
|
191 |
+
| 0.8151 | 11.2703 | 417 | 0.9067 |
|
192 |
+
| 0.8352 | 11.3514 | 420 | 0.9060 |
|
193 |
+
| 1.0484 | 11.4324 | 423 | 0.9044 |
|
194 |
+
| 0.8623 | 11.5135 | 426 | 0.9041 |
|
195 |
+
| 0.8942 | 11.5946 | 429 | 0.9022 |
|
196 |
+
| 0.8789 | 11.6757 | 432 | 0.9008 |
|
197 |
+
| 0.9077 | 11.7568 | 435 | 0.9000 |
|
198 |
+
| 0.8 | 11.8378 | 438 | 0.8999 |
|
199 |
+
| 0.8116 | 11.9189 | 441 | 0.8985 |
|
200 |
+
| 0.8382 | 12.0 | 444 | 0.8970 |
|
201 |
+
| 0.8267 | 12.0811 | 447 | 0.8962 |
|
202 |
+
| 0.9168 | 12.1622 | 450 | 0.8947 |
|
203 |
+
| 0.7878 | 12.2432 | 453 | 0.8949 |
|
204 |
+
| 0.8829 | 12.3243 | 456 | 0.8926 |
|
205 |
+
| 0.8465 | 12.4054 | 459 | 0.8935 |
|
206 |
+
| 0.882 | 12.4865 | 462 | 0.8914 |
|
207 |
+
| 0.8265 | 12.5676 | 465 | 0.8896 |
|
208 |
+
| 0.8846 | 12.6486 | 468 | 0.8891 |
|
209 |
+
| 0.7943 | 12.7297 | 471 | 0.8899 |
|
210 |
+
| 0.8846 | 12.8108 | 474 | 0.8883 |
|
211 |
+
| 0.8177 | 12.8919 | 477 | 0.8852 |
|
212 |
+
| 0.8807 | 12.9730 | 480 | 0.8840 |
|
213 |
+
| 0.893 | 13.0541 | 483 | 0.8822 |
|
214 |
+
| 0.7497 | 13.1351 | 486 | 0.8799 |
|
215 |
+
| 0.8803 | 13.2162 | 489 | 0.8785 |
|
216 |
+
| 0.917 | 13.2973 | 492 | 0.8797 |
|
217 |
+
| 0.8162 | 13.3784 | 495 | 0.8783 |
|
218 |
+
| 0.9659 | 13.4595 | 498 | 0.8799 |
|
219 |
+
| 0.8163 | 13.5405 | 501 | 0.8790 |
|
220 |
+
| 0.8139 | 13.6216 | 504 | 0.8753 |
|
221 |
+
| 0.7451 | 13.7027 | 507 | 0.8737 |
|
222 |
+
| 0.7726 | 13.7838 | 510 | 0.8738 |
|
223 |
+
| 0.821 | 13.8649 | 513 | 0.8714 |
|
224 |
+
| 0.8782 | 13.9459 | 516 | 0.8689 |
|
225 |
+
| 0.8563 | 14.0270 | 519 | 0.8700 |
|
226 |
+
| 0.8645 | 14.1081 | 522 | 0.8671 |
|
227 |
+
| 0.9223 | 14.1892 | 525 | 0.8676 |
|
228 |
+
| 0.8396 | 14.2703 | 528 | 0.8682 |
|
229 |
+
| 0.8224 | 14.3514 | 531 | 0.8679 |
|
230 |
+
| 0.7793 | 14.4324 | 534 | 0.8690 |
|
231 |
+
| 0.7911 | 14.5135 | 537 | 0.8650 |
|
232 |
+
| 0.7706 | 14.5946 | 540 | 0.8634 |
|
233 |
+
| 0.7983 | 14.6757 | 543 | 0.8606 |
|
234 |
+
| 0.7409 | 14.7568 | 546 | 0.8611 |
|
235 |
+
| 0.8634 | 14.8378 | 549 | 0.8594 |
|
236 |
+
| 0.8631 | 14.9189 | 552 | 0.8594 |
|
237 |
+
| 0.7624 | 15.0 | 555 | 0.8582 |
|
238 |
+
| 0.8568 | 15.0811 | 558 | 0.8568 |
|
239 |
+
| 0.8833 | 15.1622 | 561 | 0.8542 |
|
240 |
+
| 0.7804 | 15.2432 | 564 | 0.8514 |
|
241 |
+
| 0.9113 | 15.3243 | 567 | 0.8526 |
|
242 |
+
| 0.72 | 15.4054 | 570 | 0.8511 |
|
243 |
+
| 0.8631 | 15.4865 | 573 | 0.8498 |
|
244 |
+
| 0.6486 | 15.5676 | 576 | 0.8507 |
|
245 |
+
| 0.7745 | 15.6486 | 579 | 0.8507 |
|
246 |
+
| 0.8591 | 15.7297 | 582 | 0.8519 |
|
247 |
+
| 0.8479 | 15.8108 | 585 | 0.8515 |
|
248 |
+
| 0.7879 | 15.8919 | 588 | 0.8486 |
|
249 |
+
| 0.7829 | 15.9730 | 591 | 0.8483 |
|
250 |
+
| 0.7271 | 16.0541 | 594 | 0.8451 |
|
251 |
+
| 0.7425 | 16.1351 | 597 | 0.8464 |
|
252 |
+
| 0.7692 | 16.2162 | 600 | 0.8436 |
|
253 |
+
| 0.7492 | 16.2973 | 603 | 0.8459 |
|
254 |
+
| 0.7085 | 16.3784 | 606 | 0.8440 |
|
255 |
+
| 0.7351 | 16.4595 | 609 | 0.8429 |
|
256 |
+
| 0.8593 | 16.5405 | 612 | 0.8412 |
|
257 |
+
| 0.8999 | 16.6216 | 615 | 0.8400 |
|
258 |
+
| 0.8136 | 16.7027 | 618 | 0.8417 |
|
259 |
+
| 0.8415 | 16.7838 | 621 | 0.8393 |
|
260 |
+
| 0.849 | 16.8649 | 624 | 0.8400 |
|
261 |
+
| 0.798 | 16.9459 | 627 | 0.8409 |
|
262 |
+
| 0.7581 | 17.0270 | 630 | 0.8393 |
|
263 |
+
| 0.7244 | 17.1081 | 633 | 0.8376 |
|
264 |
+
| 0.7221 | 17.1892 | 636 | 0.8379 |
|
265 |
+
| 0.8085 | 17.2703 | 639 | 0.8363 |
|
266 |
+
| 0.7364 | 17.3514 | 642 | 0.8357 |
|
267 |
+
| 0.7684 | 17.4324 | 645 | 0.8385 |
|
268 |
+
| 0.7323 | 17.5135 | 648 | 0.8330 |
|
269 |
+
| 0.8999 | 17.5946 | 651 | 0.8334 |
|
270 |
+
| 0.7243 | 17.6757 | 654 | 0.8323 |
|
271 |
+
| 0.741 | 17.7568 | 657 | 0.8300 |
|
272 |
+
| 0.7529 | 17.8378 | 660 | 0.8290 |
|
273 |
+
| 0.9129 | 17.9189 | 663 | 0.8313 |
|
274 |
+
| 0.8043 | 18.0 | 666 | 0.8282 |
|
275 |
+
| 0.838 | 18.0811 | 669 | 0.8294 |
|
276 |
+
| 0.8133 | 18.1622 | 672 | 0.8294 |
|
277 |
+
| 0.708 | 18.2432 | 675 | 0.8283 |
|
278 |
+
| 0.8017 | 18.3243 | 678 | 0.8289 |
|
279 |
+
| 0.9087 | 18.4054 | 681 | 0.8256 |
|
280 |
+
| 0.6983 | 18.4865 | 684 | 0.8260 |
|
281 |
+
| 0.7684 | 18.5676 | 687 | 0.8251 |
|
282 |
+
| 0.6372 | 18.6486 | 690 | 0.8241 |
|
283 |
+
| 0.7369 | 18.7297 | 693 | 0.8261 |
|
284 |
+
| 0.819 | 18.8108 | 696 | 0.8231 |
|
285 |
+
| 0.6806 | 18.8919 | 699 | 0.8244 |
|
286 |
+
| 0.8233 | 18.9730 | 702 | 0.8237 |
|
287 |
+
| 0.7614 | 19.0541 | 705 | 0.8234 |
|
288 |
+
| 0.7862 | 19.1351 | 708 | 0.8212 |
|
289 |
+
| 0.6605 | 19.2162 | 711 | 0.8195 |
|
290 |
+
| 0.8229 | 19.2973 | 714 | 0.8194 |
|
291 |
+
| 0.731 | 19.3784 | 717 | 0.8193 |
|
292 |
+
| 0.783 | 19.4595 | 720 | 0.8187 |
|
293 |
+
| 0.6495 | 19.5405 | 723 | 0.8185 |
|
294 |
+
| 0.8034 | 19.6216 | 726 | 0.8173 |
|
295 |
+
| 0.7064 | 19.7027 | 729 | 0.8177 |
|
296 |
+
| 0.7276 | 19.7838 | 732 | 0.8156 |
|
297 |
+
| 0.8778 | 19.8649 | 735 | 0.8143 |
|
298 |
+
| 0.7399 | 19.9459 | 738 | 0.8126 |
|
299 |
+
| 0.7399 | 20.0270 | 741 | 0.8139 |
|
300 |
+
| 0.7912 | 20.1081 | 744 | 0.8153 |
|
301 |
+
| 0.8063 | 20.1892 | 747 | 0.8120 |
|
302 |
+
| 0.671 | 20.2703 | 750 | 0.8105 |
|
303 |
+
| 0.6342 | 20.3514 | 753 | 0.8113 |
|
304 |
+
| 0.8271 | 20.4324 | 756 | 0.8097 |
|
305 |
+
| 0.6755 | 20.5135 | 759 | 0.8099 |
|
306 |
+
| 0.762 | 20.5946 | 762 | 0.8084 |
|
307 |
+
| 0.9197 | 20.6757 | 765 | 0.8079 |
|
308 |
+
| 0.7844 | 20.7568 | 768 | 0.8094 |
|
309 |
+
| 0.6502 | 20.8378 | 771 | 0.8096 |
|
310 |
+
| 0.6877 | 20.9189 | 774 | 0.8078 |
|
311 |
+
| 0.7705 | 21.0 | 777 | 0.8074 |
|
312 |
+
| 0.6842 | 21.0811 | 780 | 0.8067 |
|
313 |
+
| 0.6799 | 21.1622 | 783 | 0.8066 |
|
314 |
+
| 0.7344 | 21.2432 | 786 | 0.8039 |
|
315 |
+
| 0.6554 | 21.3243 | 789 | 0.8051 |
|
316 |
+
| 0.7007 | 21.4054 | 792 | 0.8051 |
|
317 |
+
| 0.6292 | 21.4865 | 795 | 0.8040 |
|
318 |
+
| 0.7404 | 21.5676 | 798 | 0.8031 |
|
319 |
+
| 0.7976 | 21.6486 | 801 | 0.8030 |
|
320 |
+
| 1.0023 | 21.7297 | 804 | 0.8045 |
|
321 |
+
| 0.6626 | 21.8108 | 807 | 0.8042 |
|
322 |
+
| 0.8212 | 21.8919 | 810 | 0.8019 |
|
323 |
+
| 0.6982 | 21.9730 | 813 | 0.8029 |
|
324 |
+
| 0.7221 | 22.0541 | 816 | 0.8007 |
|
325 |
+
| 0.8042 | 22.1351 | 819 | 0.8013 |
|
326 |
+
| 0.775 | 22.2162 | 822 | 0.7999 |
|
327 |
+
| 0.7614 | 22.2973 | 825 | 0.7971 |
|
328 |
+
| 0.7132 | 22.3784 | 828 | 0.7972 |
|
329 |
+
| 0.647 | 22.4595 | 831 | 0.8001 |
|
330 |
+
| 0.692 | 22.5405 | 834 | 0.7968 |
|
331 |
+
| 0.7775 | 22.6216 | 837 | 0.7976 |
|
332 |
+
| 0.7764 | 22.7027 | 840 | 0.7980 |
|
333 |
+
| 0.6962 | 22.7838 | 843 | 0.7990 |
|
334 |
+
| 0.649 | 22.8649 | 846 | 0.7941 |
|
335 |
+
| 0.7853 | 22.9459 | 849 | 0.7940 |
|
336 |
+
| 0.6808 | 23.0270 | 852 | 0.7959 |
|
337 |
+
| 0.7018 | 23.1081 | 855 | 0.7948 |
|
338 |
+
| 0.6743 | 23.1892 | 858 | 0.7949 |
|
339 |
+
| 0.7512 | 23.2703 | 861 | 0.7926 |
|
340 |
+
| 0.7136 | 23.3514 | 864 | 0.7922 |
|
341 |
+
| 0.6847 | 23.4324 | 867 | 0.7943 |
|
342 |
+
| 0.7309 | 23.5135 | 870 | 0.7910 |
|
343 |
+
| 0.9269 | 23.5946 | 873 | 0.7903 |
|
344 |
+
| 0.7629 | 23.6757 | 876 | 0.7923 |
|
345 |
+
| 0.6551 | 23.7568 | 879 | 0.7896 |
|
346 |
+
| 0.7346 | 23.8378 | 882 | 0.7920 |
|
347 |
+
| 0.7469 | 23.9189 | 885 | 0.7921 |
|
348 |
+
| 0.6343 | 24.0 | 888 | 0.7925 |
|
349 |
+
| 0.7232 | 24.0811 | 891 | 0.7905 |
|
350 |
+
| 0.714 | 24.1622 | 894 | 0.7885 |
|
351 |
+
| 0.6603 | 24.2432 | 897 | 0.7885 |
|
352 |
+
| 0.6798 | 24.3243 | 900 | 0.7892 |
|
353 |
+
| 0.7163 | 24.4054 | 903 | 0.7889 |
|
354 |
+
| 0.8428 | 24.4865 | 906 | 0.7871 |
|
355 |
+
| 0.8125 | 24.5676 | 909 | 0.7875 |
|
356 |
+
| 0.7566 | 24.6486 | 912 | 0.7895 |
|
357 |
+
| 0.7167 | 24.7297 | 915 | 0.7874 |
|
358 |
+
| 0.65 | 24.8108 | 918 | 0.7861 |
|
359 |
+
| 0.6817 | 24.8919 | 921 | 0.7860 |
|
360 |
+
| 0.6185 | 24.9730 | 924 | 0.7879 |
|
361 |
+
| 0.7745 | 25.0541 | 927 | 0.7900 |
|
362 |
+
| 0.8923 | 25.1351 | 930 | 0.7871 |
|
363 |
+
| 0.7395 | 25.2162 | 933 | 0.7834 |
|
364 |
+
| 0.7361 | 25.2973 | 936 | 0.7842 |
|
365 |
+
| 0.676 | 25.3784 | 939 | 0.7835 |
|
366 |
+
| 0.7359 | 25.4595 | 942 | 0.7846 |
|
367 |
+
| 0.6533 | 25.5405 | 945 | 0.7829 |
|
368 |
+
| 0.6989 | 25.6216 | 948 | 0.7848 |
|
369 |
+
| 0.8168 | 25.7027 | 951 | 0.7837 |
|
370 |
+
| 0.6519 | 25.7838 | 954 | 0.7826 |
|
371 |
+
| 0.6951 | 25.8649 | 957 | 0.7840 |
|
372 |
+
| 0.5833 | 25.9459 | 960 | 0.7823 |
|
373 |
+
| 0.617 | 26.0270 | 963 | 0.7812 |
|
374 |
+
| 0.6778 | 26.1081 | 966 | 0.7829 |
|
375 |
+
| 0.6571 | 26.1892 | 969 | 0.7818 |
|
376 |
+
| 0.7921 | 26.2703 | 972 | 0.7811 |
|
377 |
+
| 0.6962 | 26.3514 | 975 | 0.7807 |
|
378 |
+
| 0.8437 | 26.4324 | 978 | 0.7814 |
|
379 |
+
| 0.6147 | 26.5135 | 981 | 0.7813 |
|
380 |
+
| 0.7263 | 26.5946 | 984 | 0.7803 |
|
381 |
+
| 0.6618 | 26.6757 | 987 | 0.7809 |
|
382 |
+
| 0.7502 | 26.7568 | 990 | 0.7835 |
|
383 |
+
| 0.6094 | 26.8378 | 993 | 0.7813 |
|
384 |
+
| 0.6506 | 26.9189 | 996 | 0.7800 |
|
385 |
+
| 0.7927 | 27.0 | 999 | 0.7807 |
|
386 |
+
| 0.8217 | 27.0811 | 1002 | 0.7816 |
|
387 |
+
| 0.5989 | 27.1622 | 1005 | 0.7786 |
|
388 |
+
| 0.7204 | 27.2432 | 1008 | 0.7782 |
|
389 |
+
| 0.6906 | 27.3243 | 1011 | 0.7785 |
|
390 |
+
| 0.7592 | 27.4054 | 1014 | 0.7777 |
|
391 |
+
| 0.8692 | 27.4865 | 1017 | 0.7778 |
|
392 |
+
| 0.587 | 27.5676 | 1020 | 0.7781 |
|
393 |
+
| 0.7139 | 27.6486 | 1023 | 0.7781 |
|
394 |
+
| 0.6367 | 27.7297 | 1026 | 0.7774 |
|
395 |
+
| 0.6502 | 27.8108 | 1029 | 0.7779 |
|
396 |
+
| 0.6956 | 27.8919 | 1032 | 0.7770 |
|
397 |
+
| 0.7137 | 27.9730 | 1035 | 0.7784 |
|
398 |
+
| 0.7062 | 28.0541 | 1038 | 0.7775 |
|
399 |
+
| 0.6568 | 28.1351 | 1041 | 0.7782 |
|
400 |
+
| 0.7612 | 28.2162 | 1044 | 0.7766 |
|
401 |
+
| 0.7177 | 28.2973 | 1047 | 0.7765 |
|
402 |
+
| 0.6537 | 28.3784 | 1050 | 0.7776 |
|
403 |
+
| 0.6742 | 28.4595 | 1053 | 0.7773 |
|
404 |
+
| 0.724 | 28.5405 | 1056 | 0.7778 |
|
405 |
+
| 0.8277 | 28.6216 | 1059 | 0.7763 |
|
406 |
+
| 0.609 | 28.7027 | 1062 | 0.7767 |
|
407 |
+
| 0.5459 | 28.7838 | 1065 | 0.7781 |
|
408 |
+
| 0.8179 | 28.8649 | 1068 | 0.7769 |
|
409 |
+
| 0.6419 | 28.9459 | 1071 | 0.7754 |
|
410 |
+
| 0.6947 | 29.0270 | 1074 | 0.7745 |
|
411 |
+
| 0.71 | 29.1081 | 1077 | 0.7751 |
|
412 |
+
| 0.5664 | 29.1892 | 1080 | 0.7747 |
|
413 |
+
| 0.7678 | 29.2703 | 1083 | 0.7741 |
|
414 |
+
| 0.7829 | 29.3514 | 1086 | 0.7732 |
|
415 |
+
| 0.7294 | 29.4324 | 1089 | 0.7729 |
|
416 |
+
| 0.679 | 29.5135 | 1092 | 0.7752 |
|
417 |
+
| 0.6838 | 29.5946 | 1095 | 0.7739 |
|
418 |
+
| 0.6918 | 29.6757 | 1098 | 0.7731 |
|
419 |
+
| 0.6563 | 29.7568 | 1101 | 0.7727 |
|
420 |
+
| 0.6946 | 29.8378 | 1104 | 0.7742 |
|
421 |
+
| 0.6884 | 29.9189 | 1107 | 0.7729 |
|
422 |
+
| 0.6851 | 30.0 | 1110 | 0.7744 |
|
423 |
+
| 0.5865 | 30.0811 | 1113 | 0.7744 |
|
424 |
+
| 0.6906 | 30.1622 | 1116 | 0.7746 |
|
425 |
+
| 0.6716 | 30.2432 | 1119 | 0.7751 |
|
426 |
+
| 0.7492 | 30.3243 | 1122 | 0.7736 |
|
427 |
+
| 0.6608 | 30.4054 | 1125 | 0.7709 |
|
428 |
+
| 0.6577 | 30.4865 | 1128 | 0.7710 |
|
429 |
+
| 0.7475 | 30.5676 | 1131 | 0.7722 |
|
430 |
+
| 0.7643 | 30.6486 | 1134 | 0.7694 |
|
431 |
+
| 0.8186 | 30.7297 | 1137 | 0.7711 |
|
432 |
+
| 0.7465 | 30.8108 | 1140 | 0.7724 |
|
433 |
+
| 0.6621 | 30.8919 | 1143 | 0.7728 |
|
434 |
+
| 0.579 | 30.9730 | 1146 | 0.7713 |
|
435 |
+
| 0.6014 | 31.0541 | 1149 | 0.7707 |
|
436 |
+
| 0.7663 | 31.1351 | 1152 | 0.7697 |
|
437 |
+
| 0.6609 | 31.2162 | 1155 | 0.7719 |
|
438 |
+
| 0.7984 | 31.2973 | 1158 | 0.7715 |
|
439 |
+
| 0.7166 | 31.3784 | 1161 | 0.7715 |
|
440 |
+
| 0.7276 | 31.4595 | 1164 | 0.7712 |
|
441 |
+
| 0.6179 | 31.5405 | 1167 | 0.7681 |
|
442 |
+
| 0.621 | 31.6216 | 1170 | 0.7707 |
|
443 |
+
| 0.743 | 31.7027 | 1173 | 0.7711 |
|
444 |
+
| 0.6733 | 31.7838 | 1176 | 0.7706 |
|
445 |
+
| 0.6066 | 31.8649 | 1179 | 0.7716 |
|
446 |
+
| 0.6787 | 31.9459 | 1182 | 0.7713 |
|
447 |
+
| 0.812 | 32.0270 | 1185 | 0.7708 |
|
448 |
+
| 0.7022 | 32.1081 | 1188 | 0.7695 |
|
449 |
+
| 0.6709 | 32.1892 | 1191 | 0.7721 |
|
450 |
+
| 0.5536 | 32.2703 | 1194 | 0.7716 |
|
451 |
+
| 0.7462 | 32.3514 | 1197 | 0.7693 |
|
452 |
+
| 0.7817 | 32.4324 | 1200 | 0.7686 |
|
453 |
+
| 0.7624 | 32.5135 | 1203 | 0.7719 |
|
454 |
+
| 0.6346 | 32.5946 | 1206 | 0.7696 |
|
455 |
+
| 0.6284 | 32.6757 | 1209 | 0.7717 |
|
456 |
+
| 0.654 | 32.7568 | 1212 | 0.7708 |
|
457 |
+
| 0.6406 | 32.8378 | 1215 | 0.7707 |
|
458 |
+
| 0.752 | 32.9189 | 1218 | 0.7690 |
|
459 |
+
| 0.5956 | 33.0 | 1221 | 0.7713 |
|
460 |
+
| 0.6772 | 33.0811 | 1224 | 0.7701 |
|
461 |
+
| 0.5034 | 33.1622 | 1227 | 0.7693 |
|
462 |
+
| 0.6292 | 33.2432 | 1230 | 0.7687 |
|
463 |
+
| 0.7374 | 33.3243 | 1233 | 0.7693 |
|
464 |
+
| 0.5385 | 33.4054 | 1236 | 0.7678 |
|
465 |
+
| 0.7198 | 33.4865 | 1239 | 0.7686 |
|
466 |
+
| 0.7642 | 33.5676 | 1242 | 0.7695 |
|
467 |
+
| 0.6984 | 33.6486 | 1245 | 0.7682 |
|
468 |
+
| 0.776 | 33.7297 | 1248 | 0.7694 |
|
469 |
+
| 0.7073 | 33.8108 | 1251 | 0.7700 |
|
470 |
+
| 0.8016 | 33.8919 | 1254 | 0.7683 |
|
471 |
+
| 0.7812 | 33.9730 | 1257 | 0.7662 |
|
472 |
+
| 0.578 | 34.0541 | 1260 | 0.7681 |
|
473 |
+
| 0.7074 | 34.1351 | 1263 | 0.7679 |
|
474 |
+
| 0.6142 | 34.2162 | 1266 | 0.7669 |
|
475 |
+
| 0.6469 | 34.2973 | 1269 | 0.7675 |
|
476 |
+
| 0.7334 | 34.3784 | 1272 | 0.7669 |
|
477 |
+
| 0.7086 | 34.4595 | 1275 | 0.7675 |
|
478 |
+
| 0.6302 | 34.5405 | 1278 | 0.7684 |
|
479 |
+
| 0.6576 | 34.6216 | 1281 | 0.7687 |
|
480 |
+
| 0.7496 | 34.7027 | 1284 | 0.7667 |
|
481 |
+
| 0.7836 | 34.7838 | 1287 | 0.7671 |
|
482 |
+
| 0.6405 | 34.8649 | 1290 | 0.7679 |
|
483 |
+
| 0.6088 | 34.9459 | 1293 | 0.7681 |
|
484 |
+
| 0.7511 | 35.0270 | 1296 | 0.7697 |
|
485 |
+
| 0.7591 | 35.1081 | 1299 | 0.7680 |
|
486 |
+
| 0.6207 | 35.1892 | 1302 | 0.7689 |
|
487 |
+
| 0.651 | 35.2703 | 1305 | 0.7660 |
|
488 |
+
| 0.6758 | 35.3514 | 1308 | 0.7694 |
|
489 |
+
| 0.7277 | 35.4324 | 1311 | 0.7680 |
|
490 |
+
| 0.7118 | 35.5135 | 1314 | 0.7663 |
|
491 |
+
| 0.706 | 35.5946 | 1317 | 0.7669 |
|
492 |
+
| 0.6595 | 35.6757 | 1320 | 0.7678 |
|
493 |
+
| 0.7253 | 35.7568 | 1323 | 0.7688 |
|
494 |
+
| 0.582 | 35.8378 | 1326 | 0.7694 |
|
495 |
+
| 0.7094 | 35.9189 | 1329 | 0.7684 |
|
496 |
+
| 0.7042 | 36.0 | 1332 | 0.7686 |
|
497 |
+
| 0.6486 | 36.0811 | 1335 | 0.7664 |
|
498 |
+
| 0.6691 | 36.1622 | 1338 | 0.7678 |
|
499 |
+
| 0.7937 | 36.2432 | 1341 | 0.7672 |
|
500 |
+
| 0.6343 | 36.3243 | 1344 | 0.7670 |
|
501 |
+
| 0.76 | 36.4054 | 1347 | 0.7674 |
|
502 |
+
| 0.5929 | 36.4865 | 1350 | 0.7690 |
|
503 |
+
| 0.6485 | 36.5676 | 1353 | 0.7674 |
|
504 |
+
| 0.7451 | 36.6486 | 1356 | 0.7663 |
|
505 |
+
| 0.6367 | 36.7297 | 1359 | 0.7691 |
|
506 |
+
| 0.6192 | 36.8108 | 1362 | 0.7656 |
|
507 |
+
| 0.7042 | 36.8919 | 1365 | 0.7682 |
|
508 |
+
| 0.6886 | 36.9730 | 1368 | 0.7682 |
|
509 |
+
| 0.8111 | 37.0541 | 1371 | 0.7690 |
|
510 |
+
| 0.6443 | 37.1351 | 1374 | 0.7690 |
|
511 |
+
| 0.6387 | 37.2162 | 1377 | 0.7679 |
|
512 |
+
| 0.7861 | 37.2973 | 1380 | 0.7668 |
|
513 |
+
| 0.6397 | 37.3784 | 1383 | 0.7670 |
|
514 |
+
| 0.6467 | 37.4595 | 1386 | 0.7688 |
|
515 |
+
| 0.6977 | 37.5405 | 1389 | 0.7699 |
|
516 |
+
| 0.7833 | 37.6216 | 1392 | 0.7672 |
|
517 |
+
| 0.6164 | 37.7027 | 1395 | 0.7675 |
|
518 |
+
| 0.7457 | 37.7838 | 1398 | 0.7690 |
|
519 |
+
| 0.6477 | 37.8649 | 1401 | 0.7681 |
|
520 |
+
| 0.7251 | 37.9459 | 1404 | 0.7674 |
|
521 |
+
| 0.5642 | 38.0270 | 1407 | 0.7672 |
|
522 |
+
| 0.786 | 38.1081 | 1410 | 0.7660 |
|
523 |
+
| 0.7014 | 38.1892 | 1413 | 0.7683 |
|
524 |
+
| 0.7747 | 38.2703 | 1416 | 0.7659 |
|
525 |
+
| 0.631 | 38.3514 | 1419 | 0.7675 |
|
526 |
+
| 0.6787 | 38.4324 | 1422 | 0.7671 |
|
527 |
+
| 0.6388 | 38.5135 | 1425 | 0.7684 |
|
528 |
+
| 0.5745 | 38.5946 | 1428 | 0.7666 |
|
529 |
+
| 0.6253 | 38.6757 | 1431 | 0.7668 |
|
530 |
+
| 0.6611 | 38.7568 | 1434 | 0.7688 |
|
531 |
+
| 0.6129 | 38.8378 | 1437 | 0.7666 |
|
532 |
+
| 0.7588 | 38.9189 | 1440 | 0.7670 |
|
533 |
+
| 0.7615 | 39.0 | 1443 | 0.7664 |
|
534 |
+
| 0.7449 | 39.0811 | 1446 | 0.7674 |
|
535 |
+
| 0.7012 | 39.1622 | 1449 | 0.7666 |
|
536 |
+
| 0.788 | 39.2432 | 1452 | 0.7676 |
|
537 |
+
| 0.5958 | 39.3243 | 1455 | 0.7666 |
|
538 |
+
| 0.675 | 39.4054 | 1458 | 0.7670 |
|
539 |
+
| 0.6761 | 39.4865 | 1461 | 0.7676 |
|
540 |
+
| 0.7046 | 39.5676 | 1464 | 0.7669 |
|
541 |
+
| 0.6198 | 39.6486 | 1467 | 0.7673 |
|
542 |
+
| 0.6646 | 39.7297 | 1470 | 0.7670 |
|
543 |
+
| 0.6161 | 39.8108 | 1473 | 0.7677 |
|
544 |
+
| 0.7152 | 39.8919 | 1476 | 0.7676 |
|
545 |
+
| 0.7365 | 39.9730 | 1479 | 0.7671 |
|
546 |
+
|
547 |
+
|
548 |
+
### Framework versions
|
549 |
+
|
550 |
+
- PEFT 0.13.2
|
551 |
+
- Transformers 4.46.0
|
552 |
+
- Pytorch 2.4.0
|
553 |
+
- Datasets 3.0.1
|
554 |
+
- Tokenizers 0.20.0
|
adapter_config.json
ADDED
@@ -0,0 +1,34 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"alpha_pattern": {},
|
3 |
+
"auto_mapping": null,
|
4 |
+
"base_model_name_or_path": "Qwen/Qwen2.5-0.5B-Instruct",
|
5 |
+
"bias": "none",
|
6 |
+
"fan_in_fan_out": false,
|
7 |
+
"inference_mode": true,
|
8 |
+
"init_lora_weights": true,
|
9 |
+
"layer_replication": null,
|
10 |
+
"layers_pattern": null,
|
11 |
+
"layers_to_transform": null,
|
12 |
+
"loftq_config": {},
|
13 |
+
"lora_alpha": 32,
|
14 |
+
"lora_dropout": 0.01,
|
15 |
+
"megatron_config": null,
|
16 |
+
"megatron_core": "megatron.core",
|
17 |
+
"modules_to_save": null,
|
18 |
+
"peft_type": "LORA",
|
19 |
+
"r": 16,
|
20 |
+
"rank_pattern": {},
|
21 |
+
"revision": null,
|
22 |
+
"target_modules": [
|
23 |
+
"q_proj",
|
24 |
+
"gate_proj",
|
25 |
+
"k_proj",
|
26 |
+
"down_proj",
|
27 |
+
"v_proj",
|
28 |
+
"up_proj",
|
29 |
+
"o_proj"
|
30 |
+
],
|
31 |
+
"task_type": "CAUSAL_LM",
|
32 |
+
"use_dora": false,
|
33 |
+
"use_rslora": false
|
34 |
+
}
|
adapter_model.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:42dd600bcf09214b61860789a323c0c0985fdb5f62f94293845b5d6fae9663bd
|
3 |
+
size 35237104
|
added_tokens.json
ADDED
@@ -0,0 +1,24 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"</tool_call>": 151658,
|
3 |
+
"<tool_call>": 151657,
|
4 |
+
"<|box_end|>": 151649,
|
5 |
+
"<|box_start|>": 151648,
|
6 |
+
"<|endoftext|>": 151643,
|
7 |
+
"<|file_sep|>": 151664,
|
8 |
+
"<|fim_middle|>": 151660,
|
9 |
+
"<|fim_pad|>": 151662,
|
10 |
+
"<|fim_prefix|>": 151659,
|
11 |
+
"<|fim_suffix|>": 151661,
|
12 |
+
"<|im_end|>": 151645,
|
13 |
+
"<|im_start|>": 151644,
|
14 |
+
"<|image_pad|>": 151655,
|
15 |
+
"<|object_ref_end|>": 151647,
|
16 |
+
"<|object_ref_start|>": 151646,
|
17 |
+
"<|quad_end|>": 151651,
|
18 |
+
"<|quad_start|>": 151650,
|
19 |
+
"<|repo_name|>": 151663,
|
20 |
+
"<|video_pad|>": 151656,
|
21 |
+
"<|vision_end|>": 151653,
|
22 |
+
"<|vision_pad|>": 151654,
|
23 |
+
"<|vision_start|>": 151652
|
24 |
+
}
|
merges.txt
ADDED
The diff for this file is too large to render.
See raw diff
|
|
special_tokens_map.json
ADDED
@@ -0,0 +1,31 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"additional_special_tokens": [
|
3 |
+
"<|im_start|>",
|
4 |
+
"<|im_end|>",
|
5 |
+
"<|object_ref_start|>",
|
6 |
+
"<|object_ref_end|>",
|
7 |
+
"<|box_start|>",
|
8 |
+
"<|box_end|>",
|
9 |
+
"<|quad_start|>",
|
10 |
+
"<|quad_end|>",
|
11 |
+
"<|vision_start|>",
|
12 |
+
"<|vision_end|>",
|
13 |
+
"<|vision_pad|>",
|
14 |
+
"<|image_pad|>",
|
15 |
+
"<|video_pad|>"
|
16 |
+
],
|
17 |
+
"eos_token": {
|
18 |
+
"content": "<|im_end|>",
|
19 |
+
"lstrip": false,
|
20 |
+
"normalized": false,
|
21 |
+
"rstrip": false,
|
22 |
+
"single_word": false
|
23 |
+
},
|
24 |
+
"pad_token": {
|
25 |
+
"content": "<|endoftext|>",
|
26 |
+
"lstrip": false,
|
27 |
+
"normalized": false,
|
28 |
+
"rstrip": false,
|
29 |
+
"single_word": false
|
30 |
+
}
|
31 |
+
}
|
tokenizer.json
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3511c9b326fec6a1b23dbf63767a672c911fbee5b2145ad0099a5bd8aff458b5
|
3 |
+
size 11421994
|
tokenizer_config.json
ADDED
@@ -0,0 +1,207 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"add_bos_token": false,
|
3 |
+
"add_prefix_space": false,
|
4 |
+
"added_tokens_decoder": {
|
5 |
+
"151643": {
|
6 |
+
"content": "<|endoftext|>",
|
7 |
+
"lstrip": false,
|
8 |
+
"normalized": false,
|
9 |
+
"rstrip": false,
|
10 |
+
"single_word": false,
|
11 |
+
"special": true
|
12 |
+
},
|
13 |
+
"151644": {
|
14 |
+
"content": "<|im_start|>",
|
15 |
+
"lstrip": false,
|
16 |
+
"normalized": false,
|
17 |
+
"rstrip": false,
|
18 |
+
"single_word": false,
|
19 |
+
"special": true
|
20 |
+
},
|
21 |
+
"151645": {
|
22 |
+
"content": "<|im_end|>",
|
23 |
+
"lstrip": false,
|
24 |
+
"normalized": false,
|
25 |
+
"rstrip": false,
|
26 |
+
"single_word": false,
|
27 |
+
"special": true
|
28 |
+
},
|
29 |
+
"151646": {
|
30 |
+
"content": "<|object_ref_start|>",
|
31 |
+
"lstrip": false,
|
32 |
+
"normalized": false,
|
33 |
+
"rstrip": false,
|
34 |
+
"single_word": false,
|
35 |
+
"special": true
|
36 |
+
},
|
37 |
+
"151647": {
|
38 |
+
"content": "<|object_ref_end|>",
|
39 |
+
"lstrip": false,
|
40 |
+
"normalized": false,
|
41 |
+
"rstrip": false,
|
42 |
+
"single_word": false,
|
43 |
+
"special": true
|
44 |
+
},
|
45 |
+
"151648": {
|
46 |
+
"content": "<|box_start|>",
|
47 |
+
"lstrip": false,
|
48 |
+
"normalized": false,
|
49 |
+
"rstrip": false,
|
50 |
+
"single_word": false,
|
51 |
+
"special": true
|
52 |
+
},
|
53 |
+
"151649": {
|
54 |
+
"content": "<|box_end|>",
|
55 |
+
"lstrip": false,
|
56 |
+
"normalized": false,
|
57 |
+
"rstrip": false,
|
58 |
+
"single_word": false,
|
59 |
+
"special": true
|
60 |
+
},
|
61 |
+
"151650": {
|
62 |
+
"content": "<|quad_start|>",
|
63 |
+
"lstrip": false,
|
64 |
+
"normalized": false,
|
65 |
+
"rstrip": false,
|
66 |
+
"single_word": false,
|
67 |
+
"special": true
|
68 |
+
},
|
69 |
+
"151651": {
|
70 |
+
"content": "<|quad_end|>",
|
71 |
+
"lstrip": false,
|
72 |
+
"normalized": false,
|
73 |
+
"rstrip": false,
|
74 |
+
"single_word": false,
|
75 |
+
"special": true
|
76 |
+
},
|
77 |
+
"151652": {
|
78 |
+
"content": "<|vision_start|>",
|
79 |
+
"lstrip": false,
|
80 |
+
"normalized": false,
|
81 |
+
"rstrip": false,
|
82 |
+
"single_word": false,
|
83 |
+
"special": true
|
84 |
+
},
|
85 |
+
"151653": {
|
86 |
+
"content": "<|vision_end|>",
|
87 |
+
"lstrip": false,
|
88 |
+
"normalized": false,
|
89 |
+
"rstrip": false,
|
90 |
+
"single_word": false,
|
91 |
+
"special": true
|
92 |
+
},
|
93 |
+
"151654": {
|
94 |
+
"content": "<|vision_pad|>",
|
95 |
+
"lstrip": false,
|
96 |
+
"normalized": false,
|
97 |
+
"rstrip": false,
|
98 |
+
"single_word": false,
|
99 |
+
"special": true
|
100 |
+
},
|
101 |
+
"151655": {
|
102 |
+
"content": "<|image_pad|>",
|
103 |
+
"lstrip": false,
|
104 |
+
"normalized": false,
|
105 |
+
"rstrip": false,
|
106 |
+
"single_word": false,
|
107 |
+
"special": true
|
108 |
+
},
|
109 |
+
"151656": {
|
110 |
+
"content": "<|video_pad|>",
|
111 |
+
"lstrip": false,
|
112 |
+
"normalized": false,
|
113 |
+
"rstrip": false,
|
114 |
+
"single_word": false,
|
115 |
+
"special": true
|
116 |
+
},
|
117 |
+
"151657": {
|
118 |
+
"content": "<tool_call>",
|
119 |
+
"lstrip": false,
|
120 |
+
"normalized": false,
|
121 |
+
"rstrip": false,
|
122 |
+
"single_word": false,
|
123 |
+
"special": false
|
124 |
+
},
|
125 |
+
"151658": {
|
126 |
+
"content": "</tool_call>",
|
127 |
+
"lstrip": false,
|
128 |
+
"normalized": false,
|
129 |
+
"rstrip": false,
|
130 |
+
"single_word": false,
|
131 |
+
"special": false
|
132 |
+
},
|
133 |
+
"151659": {
|
134 |
+
"content": "<|fim_prefix|>",
|
135 |
+
"lstrip": false,
|
136 |
+
"normalized": false,
|
137 |
+
"rstrip": false,
|
138 |
+
"single_word": false,
|
139 |
+
"special": false
|
140 |
+
},
|
141 |
+
"151660": {
|
142 |
+
"content": "<|fim_middle|>",
|
143 |
+
"lstrip": false,
|
144 |
+
"normalized": false,
|
145 |
+
"rstrip": false,
|
146 |
+
"single_word": false,
|
147 |
+
"special": false
|
148 |
+
},
|
149 |
+
"151661": {
|
150 |
+
"content": "<|fim_suffix|>",
|
151 |
+
"lstrip": false,
|
152 |
+
"normalized": false,
|
153 |
+
"rstrip": false,
|
154 |
+
"single_word": false,
|
155 |
+
"special": false
|
156 |
+
},
|
157 |
+
"151662": {
|
158 |
+
"content": "<|fim_pad|>",
|
159 |
+
"lstrip": false,
|
160 |
+
"normalized": false,
|
161 |
+
"rstrip": false,
|
162 |
+
"single_word": false,
|
163 |
+
"special": false
|
164 |
+
},
|
165 |
+
"151663": {
|
166 |
+
"content": "<|repo_name|>",
|
167 |
+
"lstrip": false,
|
168 |
+
"normalized": false,
|
169 |
+
"rstrip": false,
|
170 |
+
"single_word": false,
|
171 |
+
"special": false
|
172 |
+
},
|
173 |
+
"151664": {
|
174 |
+
"content": "<|file_sep|>",
|
175 |
+
"lstrip": false,
|
176 |
+
"normalized": false,
|
177 |
+
"rstrip": false,
|
178 |
+
"single_word": false,
|
179 |
+
"special": false
|
180 |
+
}
|
181 |
+
},
|
182 |
+
"additional_special_tokens": [
|
183 |
+
"<|im_start|>",
|
184 |
+
"<|im_end|>",
|
185 |
+
"<|object_ref_start|>",
|
186 |
+
"<|object_ref_end|>",
|
187 |
+
"<|box_start|>",
|
188 |
+
"<|box_end|>",
|
189 |
+
"<|quad_start|>",
|
190 |
+
"<|quad_end|>",
|
191 |
+
"<|vision_start|>",
|
192 |
+
"<|vision_end|>",
|
193 |
+
"<|vision_pad|>",
|
194 |
+
"<|image_pad|>",
|
195 |
+
"<|video_pad|>"
|
196 |
+
],
|
197 |
+
"bos_token": null,
|
198 |
+
"chat_template": "{%- if tools %}\n {{- '<|im_start|>system\\n' }}\n {%- if messages[0]['role'] == 'system' %}\n {{- messages[0]['content'] }}\n {%- else %}\n {{- 'You are Qwen, created by Alibaba Cloud. You are a helpful assistant.' }}\n {%- endif %}\n {{- \"\\n\\n# Tools\\n\\nYou may call one or more functions to assist with the user query.\\n\\nYou are provided with function signatures within <tools></tools> XML tags:\\n<tools>\" }}\n {%- for tool in tools %}\n {{- \"\\n\" }}\n {{- tool | tojson }}\n {%- endfor %}\n {{- \"\\n</tools>\\n\\nFor each function call, return a json object with function name and arguments within <tool_call></tool_call> XML tags:\\n<tool_call>\\n{\\\"name\\\": <function-name>, \\\"arguments\\\": <args-json-object>}\\n</tool_call><|im_end|>\\n\" }}\n{%- else %}\n {%- if messages[0]['role'] == 'system' %}\n {{- '<|im_start|>system\\n' + messages[0]['content'] + '<|im_end|>\\n' }}\n {%- else %}\n {{- '<|im_start|>system\\nYou are Qwen, created by Alibaba Cloud. You are a helpful assistant.<|im_end|>\\n' }}\n {%- endif %}\n{%- endif %}\n{%- for message in messages %}\n {%- if (message.role == \"user\") or (message.role == \"system\" and not loop.first) or (message.role == \"assistant\" and not message.tool_calls) %}\n {{- '<|im_start|>' + message.role + '\\n' + message.content + '<|im_end|>' + '\\n' }}\n {%- elif message.role == \"assistant\" %}\n {{- '<|im_start|>' + message.role }}\n {%- if message.content %}\n {{- '\\n' + message.content }}\n {%- endif %}\n {%- for tool_call in message.tool_calls %}\n {%- if tool_call.function is defined %}\n {%- set tool_call = tool_call.function %}\n {%- endif %}\n {{- '\\n<tool_call>\\n{\"name\": \"' }}\n {{- tool_call.name }}\n {{- '\", \"arguments\": ' }}\n {{- tool_call.arguments | tojson }}\n {{- '}\\n</tool_call>' }}\n {%- endfor %}\n {{- '<|im_end|>\\n' }}\n {%- elif message.role == \"tool\" %}\n {%- if (loop.index0 == 0) or (messages[loop.index0 - 1].role != \"tool\") %}\n {{- '<|im_start|>user' }}\n {%- endif %}\n {{- '\\n<tool_response>\\n' }}\n {{- message.content }}\n {{- '\\n</tool_response>' }}\n {%- if loop.last or (messages[loop.index0 + 1].role != \"tool\") %}\n {{- '<|im_end|>\\n' }}\n {%- endif %}\n {%- endif %}\n{%- endfor %}\n{%- if add_generation_prompt %}\n {{- '<|im_start|>assistant\\n' }}\n{%- endif %}\n",
|
199 |
+
"clean_up_tokenization_spaces": false,
|
200 |
+
"eos_token": "<|im_end|>",
|
201 |
+
"errors": "replace",
|
202 |
+
"model_max_length": 300,
|
203 |
+
"pad_token": "<|endoftext|>",
|
204 |
+
"split_special_tokens": false,
|
205 |
+
"tokenizer_class": "Qwen2Tokenizer",
|
206 |
+
"unk_token": null
|
207 |
+
}
|
training_args.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:529b53487bf15e921f6f6c93b42bd941f300bef1a350342fb7169de7e00b6228
|
3 |
+
size 5496
|
vocab.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|