|
--- |
|
library_name: transformers |
|
tags: |
|
- generated_from_trainer |
|
metrics: |
|
- accuracy |
|
model-index: |
|
- name: poison-distill |
|
results: [] |
|
--- |
|
|
|
<!-- This model card has been generated automatically according to the information the Trainer had access to. You |
|
should probably proofread and complete it, then remove this comment. --> |
|
|
|
# poison-distill |
|
|
|
This model is a fine-tuned version of [](https://huggingface.co/) on an unknown dataset. |
|
It achieves the following results on the evaluation set: |
|
- Loss: -113.4181 |
|
- Accuracy: 0.6917 |
|
|
|
## Model description |
|
|
|
More information needed |
|
|
|
## Intended uses & limitations |
|
|
|
More information needed |
|
|
|
## Training and evaluation data |
|
|
|
More information needed |
|
|
|
## Training procedure |
|
|
|
### Training hyperparameters |
|
|
|
The following hyperparameters were used during training: |
|
- learning_rate: 5e-05 |
|
- train_batch_size: 8 |
|
- eval_batch_size: 8 |
|
- seed: 42 |
|
- optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments |
|
- lr_scheduler_type: linear |
|
- num_epochs: 50 |
|
- mixed_precision_training: Native AMP |
|
|
|
### Training results |
|
|
|
| Training Loss | Epoch | Step | Validation Loss | Accuracy | |
|
|:-------------:|:-----:|:----:|:---------------:|:--------:| |
|
| -0.7533 | 1.0 | 130 | -7.9549 | 0.5263 | |
|
| -8.2193 | 2.0 | 260 | -15.5418 | 0.4662 | |
|
| -14.3197 | 3.0 | 390 | -32.2167 | 0.4737 | |
|
| -18.5547 | 4.0 | 520 | -18.9202 | 0.5489 | |
|
| -22.6905 | 5.0 | 650 | -55.1682 | 0.4361 | |
|
| -27.5336 | 6.0 | 780 | -32.4679 | 0.3459 | |
|
| -29.5975 | 7.0 | 910 | -48.1715 | 0.3985 | |
|
| -34.1837 | 8.0 | 1040 | -67.7293 | 0.6165 | |
|
| -37.6123 | 9.0 | 1170 | -52.1341 | 0.4662 | |
|
| -40.7694 | 10.0 | 1300 | -49.0945 | 0.6767 | |
|
| -43.3691 | 11.0 | 1430 | -37.0478 | 0.5489 | |
|
| -47.6433 | 12.0 | 1560 | -73.0523 | 0.4511 | |
|
| -51.0141 | 13.0 | 1690 | -110.8840 | 0.4812 | |
|
| -54.6 | 14.0 | 1820 | -81.2219 | 0.3308 | |
|
| -57.2133 | 15.0 | 1950 | -80.8684 | 0.5113 | |
|
| -58.3442 | 16.0 | 2080 | -66.5341 | 0.4060 | |
|
| -64.7089 | 17.0 | 2210 | -75.7059 | 0.5564 | |
|
| -64.26 | 18.0 | 2340 | -77.7801 | 0.5263 | |
|
| -67.8509 | 19.0 | 2470 | -61.1841 | 0.6316 | |
|
| -71.9371 | 20.0 | 2600 | -118.1544 | 0.5038 | |
|
| -75.9672 | 21.0 | 2730 | -179.2044 | 0.4812 | |
|
| -78.0096 | 22.0 | 2860 | -129.4854 | 0.4436 | |
|
| -80.3581 | 23.0 | 2990 | -100.0687 | 0.4286 | |
|
| -84.623 | 24.0 | 3120 | -82.5292 | 0.3835 | |
|
| -86.5363 | 25.0 | 3250 | -84.6636 | 0.4211 | |
|
| -90.8566 | 26.0 | 3380 | -96.3337 | 0.5489 | |
|
| -92.2054 | 27.0 | 3510 | -110.3293 | 0.4737 | |
|
| -97.6982 | 28.0 | 3640 | -195.6973 | 0.4135 | |
|
| -95.8944 | 29.0 | 3770 | -101.9933 | 0.3609 | |
|
| -99.491 | 30.0 | 3900 | -99.8199 | 0.6541 | |
|
| -103.0877 | 31.0 | 4030 | -94.2175 | 0.6767 | |
|
| -102.7123 | 32.0 | 4160 | -98.6300 | 0.4887 | |
|
| -105.2087 | 33.0 | 4290 | -152.7768 | 0.4962 | |
|
| -105.3795 | 34.0 | 4420 | -198.8245 | 0.5263 | |
|
| -108.9734 | 35.0 | 4550 | -105.7644 | 0.4286 | |
|
| -111.1308 | 36.0 | 4680 | -121.4677 | 0.4962 | |
|
| -115.0085 | 37.0 | 4810 | -75.3733 | 0.3083 | |
|
| -114.714 | 38.0 | 4940 | -115.4598 | 0.6617 | |
|
| -117.5734 | 39.0 | 5070 | -108.3964 | 0.4135 | |
|
| -115.1971 | 40.0 | 5200 | -123.7679 | 0.3835 | |
|
| -117.5617 | 41.0 | 5330 | -69.2224 | 0.2932 | |
|
| -118.2803 | 42.0 | 5460 | -104.5906 | 0.6541 | |
|
| -119.6297 | 43.0 | 5590 | -187.3416 | 0.5188 | |
|
| -121.6325 | 44.0 | 5720 | -221.8878 | 0.5113 | |
|
| -120.9663 | 45.0 | 5850 | -176.6644 | 0.3759 | |
|
| -122.3583 | 46.0 | 5980 | -142.5218 | 0.4361 | |
|
| -126.6614 | 47.0 | 6110 | -271.1018 | 0.4962 | |
|
| -122.1615 | 48.0 | 6240 | -240.8323 | 0.3985 | |
|
| -125.4207 | 49.0 | 6370 | -103.5760 | 0.6466 | |
|
| -127.0661 | 50.0 | 6500 | -113.2718 | 0.6842 | |
|
|
|
|
|
### Framework versions |
|
|
|
- Transformers 4.46.3 |
|
- Pytorch 2.5.1+cu121 |
|
- Datasets 3.1.0 |
|
- Tokenizers 0.20.3 |
|
|