dit-base-finetuned-rvlcdip-finetuned-data200

This model is a fine-tuned version of microsoft/dit-base-finetuned-rvlcdip on the imagefolder dataset. It achieves the following results on the evaluation set:

  • Loss: 3.0080
  • Accuracy: 0.5699

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 5e-05
  • train_batch_size: 2
  • eval_batch_size: 2
  • seed: 42
  • gradient_accumulation_steps: 4
  • total_train_batch_size: 8
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • lr_scheduler_warmup_ratio: 0.1
  • num_epochs: 200

Training results

Training Loss Epoch Step Validation Loss Accuracy
2.1142 1.0 46 2.0131 0.3441
1.9953 2.0 92 1.9577 0.4086
1.9558 3.0 138 1.9231 0.4301
1.9251 4.0 184 1.8015 0.4946
1.6485 5.0 230 1.7045 0.5269
1.5973 6.0 276 1.5806 0.5054
1.4755 7.0 322 1.4849 0.5054
1.4537 8.0 368 1.4356 0.5161
1.416 9.0 414 1.4512 0.5269
1.3645 10.0 460 1.3857 0.5591
1.3017 11.0 506 1.3108 0.5484
1.2794 12.0 552 1.3027 0.5376
1.1553 13.0 598 1.2883 0.5484
1.1526 14.0 644 1.3554 0.5054
1.1116 15.0 690 1.3235 0.5914
1.1925 16.0 736 1.2401 0.5806
1.1297 17.0 782 1.3425 0.5914
0.9717 18.0 828 1.3538 0.5484
0.8404 19.0 874 1.2648 0.5699
0.7008 20.0 920 1.4971 0.5376
1.1454 21.0 966 1.4137 0.4839
0.6849 22.0 1012 1.2801 0.5591
0.8566 23.0 1058 1.2380 0.5699
0.8956 24.0 1104 1.2903 0.6129
0.8004 25.0 1150 1.4372 0.5591
0.818 26.0 1196 1.1640 0.6344
0.6387 27.0 1242 1.3120 0.6452
0.7282 28.0 1288 1.4678 0.5161
0.7426 29.0 1334 1.4815 0.5269
0.735 30.0 1380 1.2714 0.6129
0.6769 31.0 1426 1.2262 0.5699
0.5562 32.0 1472 1.3348 0.6344
0.6671 33.0 1518 1.4159 0.6129
0.3708 34.0 1564 1.6416 0.5484
0.3967 35.0 1610 1.3298 0.5699
0.4692 36.0 1656 1.3559 0.5699
0.632 37.0 1702 1.3349 0.5699
0.3719 38.0 1748 1.4697 0.5914
0.4238 39.0 1794 1.5207 0.6022
0.3608 40.0 1840 1.5557 0.5591
0.6252 41.0 1886 1.6247 0.5269
0.4183 42.0 1932 1.5885 0.5914
0.3922 43.0 1978 1.6593 0.5699
0.5715 44.0 2024 1.5270 0.5699
0.3656 45.0 2070 1.8899 0.5054
0.3656 46.0 2116 2.0936 0.4624
0.4003 47.0 2162 1.5610 0.5054
0.446 48.0 2208 1.7388 0.5376
0.5219 49.0 2254 1.4976 0.6129
0.3488 50.0 2300 1.5744 0.5914
0.323 51.0 2346 1.6312 0.6022
0.3713 52.0 2392 1.6975 0.5591
0.2981 53.0 2438 1.6229 0.5699
0.3422 54.0 2484 2.0909 0.4624
0.2538 55.0 2530 2.0966 0.5161
0.3868 56.0 2576 1.5614 0.6344
0.4662 57.0 2622 1.8929 0.5269
0.4277 58.0 2668 1.9573 0.5376
0.5301 59.0 2714 1.7999 0.5699
0.3867 60.0 2760 2.3481 0.4624
0.2334 61.0 2806 1.9924 0.5376
0.2921 62.0 2852 2.0454 0.5591
0.4386 63.0 2898 1.7798 0.5376
0.3299 64.0 2944 1.9370 0.5914
0.5982 65.0 2990 2.0527 0.5591
0.4433 66.0 3036 1.6222 0.6237
0.3717 67.0 3082 1.7977 0.5914
0.3642 68.0 3128 1.6988 0.5914
0.4541 69.0 3174 1.7567 0.6022
0.3464 70.0 3220 1.9029 0.5699
0.2764 71.0 3266 1.9611 0.6022
0.2138 72.0 3312 1.9333 0.5591
0.3928 73.0 3358 1.7701 0.5806
0.1811 74.0 3404 1.8330 0.5806
0.2076 75.0 3450 1.6676 0.6559
0.3326 76.0 3496 2.0036 0.6022
0.1343 77.0 3542 1.6937 0.6344
0.3031 78.0 3588 1.9223 0.6237
0.2743 79.0 3634 2.1681 0.5699
0.3392 80.0 3680 2.0505 0.6129
0.1346 81.0 3726 2.0190 0.5699
0.0652 82.0 3772 2.2910 0.5699
0.4219 83.0 3818 1.8858 0.5914
0.1386 84.0 3864 1.7976 0.6237
0.2155 85.0 3910 2.4278 0.5161
0.4901 86.0 3956 1.9239 0.6237
0.3141 87.0 4002 2.0954 0.6559
0.2328 88.0 4048 2.2602 0.5806
0.2768 89.0 4094 2.1083 0.5914
0.3476 90.0 4140 2.4922 0.5269
0.2029 91.0 4186 2.2094 0.5591
0.2421 92.0 4232 2.2407 0.5376
0.2034 93.0 4278 2.1488 0.5591
0.2461 94.0 4324 2.1332 0.5806
0.1462 95.0 4370 2.2702 0.5591
0.5213 96.0 4416 2.2134 0.5699
0.3634 97.0 4462 2.1066 0.5699
0.1698 98.0 4508 2.2736 0.6237
0.1685 99.0 4554 2.3919 0.5806
0.1971 100.0 4600 2.0664 0.6237
0.1496 101.0 4646 2.5661 0.5806
0.283 102.0 4692 2.0714 0.5699
0.185 103.0 4738 2.1369 0.6022
0.1489 104.0 4784 2.1653 0.6129
0.1231 105.0 4830 2.0890 0.6452
0.3224 106.0 4876 2.3771 0.5376
0.3452 107.0 4922 2.2537 0.6344
0.4404 108.0 4968 2.0253 0.6129
0.3408 109.0 5014 2.1653 0.5699
0.2406 110.0 5060 2.0196 0.6237
0.3051 111.0 5106 2.1980 0.6129
0.1515 112.0 5152 2.4104 0.5699
0.3836 113.0 5198 2.2342 0.6344
0.3572 114.0 5244 2.2321 0.6022
0.3006 115.0 5290 2.3555 0.5806
0.0965 116.0 5336 2.7237 0.4516
0.2023 117.0 5382 2.3798 0.6237
0.1272 118.0 5428 2.5357 0.5591
0.4318 119.0 5474 2.4913 0.5699
0.0414 120.0 5520 2.3760 0.6022
0.1785 121.0 5566 2.3920 0.6129
0.0142 122.0 5612 2.4256 0.6022
0.1262 123.0 5658 2.7212 0.5806
0.2219 124.0 5704 2.3683 0.5699
0.1629 125.0 5750 2.4280 0.5484
0.149 126.0 5796 3.0708 0.4839
0.2394 127.0 5842 2.2192 0.6022
0.2165 128.0 5888 2.4015 0.5806
0.0729 129.0 5934 2.2241 0.6022
0.2585 130.0 5980 2.9483 0.5054
0.1401 131.0 6026 2.3180 0.6129
0.4162 132.0 6072 3.0147 0.4946
0.1188 133.0 6118 2.3128 0.6237
0.0939 134.0 6164 2.5300 0.6022
0.1039 135.0 6210 2.5740 0.5699
0.3678 136.0 6256 2.5887 0.5914
0.3998 137.0 6302 2.5664 0.5376
0.1952 138.0 6348 2.1861 0.6774
0.2616 139.0 6394 2.7036 0.5806
0.2523 140.0 6440 2.5953 0.5806
0.2772 141.0 6486 2.4114 0.6129
0.2399 142.0 6532 2.3203 0.6237
0.3769 143.0 6578 2.7200 0.5591
0.0094 144.0 6624 2.7315 0.5591
0.1818 145.0 6670 2.5223 0.6129
0.3063 146.0 6716 2.3310 0.6237
0.222 147.0 6762 2.6180 0.5806
0.2505 148.0 6808 2.2976 0.6344
0.2705 149.0 6854 2.4091 0.5914
0.1624 150.0 6900 2.8030 0.5269
0.1322 151.0 6946 2.6379 0.5591
0.0876 152.0 6992 2.5781 0.5484
0.1332 153.0 7038 2.8476 0.5591
0.2727 154.0 7084 2.6779 0.5699
0.195 155.0 7130 3.0504 0.4839
0.152 156.0 7176 2.6103 0.5806
0.2811 157.0 7222 2.5947 0.6129
0.0742 158.0 7268 2.4666 0.6559
0.2052 159.0 7314 2.5116 0.5484
0.2598 160.0 7360 3.0400 0.5269
0.2846 161.0 7406 2.2042 0.6667
0.2653 162.0 7452 3.0598 0.5484
0.358 163.0 7498 2.7669 0.5806
0.0355 164.0 7544 2.4568 0.6237
0.1817 165.0 7590 2.9532 0.5806
0.0955 166.0 7636 2.4798 0.6237
0.1941 167.0 7682 2.7027 0.5699
0.1787 168.0 7728 2.4225 0.6237
0.0998 169.0 7774 2.5104 0.5914
0.0392 170.0 7820 2.6235 0.5806
0.2689 171.0 7866 2.9215 0.5806
0.0595 172.0 7912 2.8108 0.5699
0.148 173.0 7958 2.9213 0.5806
0.2159 174.0 8004 2.6172 0.6129
0.1221 175.0 8050 2.4386 0.6237
0.0691 176.0 8096 2.8642 0.5269
0.2014 177.0 8142 2.7364 0.6022
0.0379 178.0 8188 2.4859 0.6022
0.2202 179.0 8234 3.0665 0.5484
0.2078 180.0 8280 2.3521 0.6237
0.1051 181.0 8326 2.4827 0.6237
0.2257 182.0 8372 2.8155 0.5914
0.1339 183.0 8418 2.6274 0.6237
0.1414 184.0 8464 2.7645 0.5806
0.0993 185.0 8510 2.8886 0.5591
0.1769 186.0 8556 2.5164 0.6129
0.1575 187.0 8602 2.9346 0.5376
0.0251 188.0 8648 2.6099 0.5376
0.0536 189.0 8694 2.9630 0.5376
0.1748 190.0 8740 2.8360 0.5699
0.0151 191.0 8786 2.7525 0.6022
0.2198 192.0 8832 2.6656 0.5376
0.267 193.0 8878 3.0118 0.5591
0.1043 194.0 8924 3.0214 0.5699
0.0035 195.0 8970 2.7925 0.5806
0.0707 196.0 9016 2.7839 0.5806
0.0656 197.0 9062 3.0370 0.5376
0.1155 198.0 9108 2.6510 0.5914
0.1118 199.0 9154 2.7058 0.5699
0.3086 200.0 9200 3.0080 0.5699

Framework versions

  • Transformers 4.26.1
  • Pytorch 1.13.1+cu116
  • Datasets 2.10.0
  • Tokenizers 0.13.2
Downloads last month
13
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Evaluation results