Arabic_FineTuningAraBERT_AugV4_k2_task3_organization_fold1

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.7275
  • Qwk: -0.2737
  • Mse: 0.7275
  • Rmse: 0.8529

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 10

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0385 2 2.8419 0.0571 2.8419 1.6858
No log 0.0769 4 0.8149 0.3356 0.8149 0.9027
No log 0.1154 6 0.6628 0.0 0.6628 0.8142
No log 0.1538 8 0.5876 0.2524 0.5876 0.7665
No log 0.1923 10 0.6788 -0.0421 0.6788 0.8239
No log 0.2308 12 0.6945 -0.0421 0.6945 0.8333
No log 0.2692 14 0.6693 0.0 0.6693 0.8181
No log 0.3077 16 0.8175 0.2667 0.8175 0.9041
No log 0.3462 18 0.8458 0.2524 0.8458 0.9197
No log 0.3846 20 0.6948 0.0 0.6948 0.8335
No log 0.4231 22 0.5940 0.0 0.5940 0.7707
No log 0.4615 24 0.6050 0.0 0.6050 0.7778
No log 0.5 26 0.6025 0.0 0.6025 0.7762
No log 0.5385 28 0.6333 0.0 0.6333 0.7958
No log 0.5769 30 0.6566 0.0 0.6566 0.8103
No log 0.6154 32 0.7340 0.0 0.7340 0.8568
No log 0.6538 34 0.8732 0.4310 0.8732 0.9345
No log 0.6923 36 1.0108 0.0966 1.0108 1.0054
No log 0.7308 38 1.1278 0.0 1.1278 1.0620
No log 0.7692 40 1.1253 -0.0097 1.1253 1.0608
No log 0.8077 42 0.9321 -0.2791 0.9321 0.9655
No log 0.8462 44 1.0903 -0.0476 1.0903 1.0442
No log 0.8846 46 1.0937 -0.0476 1.0937 1.0458
No log 0.9231 48 1.3717 -0.0097 1.3717 1.1712
No log 0.9615 50 1.3145 -0.0097 1.3145 1.1465
No log 1.0 52 0.8914 0.2254 0.8914 0.9441
No log 1.0385 54 0.7481 0.0 0.7481 0.8649
No log 1.0769 56 0.8766 0.2326 0.8766 0.9362
No log 1.1154 58 0.9165 0.2326 0.9165 0.9573
No log 1.1538 60 1.0889 -0.0097 1.0889 1.0435
No log 1.1923 62 0.9257 0.2254 0.9257 0.9621
No log 1.2308 64 0.7225 0.2524 0.7225 0.8500
No log 1.2692 66 0.7249 0.2524 0.7249 0.8514
No log 1.3077 68 0.6663 0.2524 0.6663 0.8163
No log 1.3462 70 0.6298 0.0222 0.6298 0.7936
No log 1.3846 72 0.9049 0.2194 0.9049 0.9513
No log 1.4231 74 1.5201 0.0 1.5201 1.2329
No log 1.4615 76 1.5613 0.0 1.5613 1.2495
No log 1.5 78 1.1931 -0.1547 1.1931 1.0923
No log 1.5385 80 0.7410 -0.2737 0.7410 0.8608
No log 1.5769 82 0.9019 0.0763 0.9019 0.9497
No log 1.6154 84 0.7719 -0.2737 0.7719 0.8786
No log 1.6538 86 0.7081 0.0 0.7081 0.8415
No log 1.6923 88 0.8602 0.2326 0.8602 0.9275
No log 1.7308 90 0.8955 0.2254 0.8955 0.9463
No log 1.7692 92 0.7331 0.2667 0.7331 0.8562
No log 1.8077 94 0.6523 0.0 0.6523 0.8077
No log 1.8462 96 0.6163 0.0 0.6163 0.7851
No log 1.8846 98 0.6229 0.0 0.6229 0.7893
No log 1.9231 100 0.6672 0.2667 0.6672 0.8168
No log 1.9615 102 0.8137 0.2254 0.8137 0.9021
No log 2.0 104 0.7457 0.2254 0.7457 0.8635
No log 2.0385 106 0.6361 0.0 0.6361 0.7975
No log 2.0769 108 0.6498 -0.2791 0.6498 0.8061
No log 2.1154 110 0.6728 0.0 0.6728 0.8202
No log 2.1538 112 0.8861 0.0620 0.8861 0.9414
No log 2.1923 114 0.9122 0.0620 0.9122 0.9551
No log 2.2308 116 0.7553 -0.2791 0.7553 0.8691
No log 2.2692 118 0.8710 -0.0708 0.8710 0.9333
No log 2.3077 120 0.8951 -0.0708 0.8951 0.9461
No log 2.3462 122 0.8290 -0.2791 0.8290 0.9105
No log 2.3846 124 0.8639 -0.1440 0.8639 0.9295
No log 2.4231 126 0.8308 -0.2222 0.8308 0.9115
No log 2.4615 128 0.8696 -0.1786 0.8696 0.9325
No log 2.5 130 0.9471 -0.1440 0.9471 0.9732
No log 2.5385 132 0.9081 -0.2222 0.9081 0.9529
No log 2.5769 134 0.8954 -0.2737 0.8954 0.9462
No log 2.6154 136 0.8557 -0.2222 0.8557 0.9250
No log 2.6538 138 0.8410 -0.2791 0.8410 0.9171
No log 2.6923 140 0.8584 -0.2737 0.8584 0.9265
No log 2.7308 142 0.8039 -0.2737 0.8039 0.8966
No log 2.7692 144 0.8825 0.0517 0.8825 0.9394
No log 2.8077 146 0.7983 0.0222 0.7983 0.8935
No log 2.8462 148 0.7476 -0.2737 0.7476 0.8647
No log 2.8846 150 0.7443 -0.2737 0.7443 0.8627
No log 2.9231 152 0.7528 -0.2791 0.7528 0.8676
No log 2.9615 154 0.7416 -0.2791 0.7416 0.8611
No log 3.0 156 0.7555 0.0 0.7555 0.8692
No log 3.0385 158 0.7325 -0.2737 0.7325 0.8558
No log 3.0769 160 0.8459 -0.2737 0.8459 0.9198
No log 3.1154 162 0.8346 -0.2737 0.8346 0.9136
No log 3.1538 164 0.8172 -0.2737 0.8172 0.9040
No log 3.1923 166 0.8733 -0.2222 0.8733 0.9345
No log 3.2308 168 0.9019 -0.2737 0.9019 0.9497
No log 3.2692 170 0.9925 -0.0708 0.9925 0.9962
No log 3.3077 172 0.9149 -0.2737 0.9149 0.9565
No log 3.3462 174 0.8892 -0.2222 0.8892 0.9430
No log 3.3846 176 0.8864 -0.2737 0.8864 0.9415
No log 3.4231 178 0.8702 -0.2737 0.8702 0.9329
No log 3.4615 180 0.8755 -0.2737 0.8755 0.9357
No log 3.5 182 0.8708 -0.2737 0.8708 0.9332
No log 3.5385 184 0.9203 -0.1493 0.9203 0.9593
No log 3.5769 186 0.8794 -0.2737 0.8794 0.9378
No log 3.6154 188 0.8456 -0.2737 0.8456 0.9196
No log 3.6538 190 0.8247 -0.2737 0.8247 0.9081
No log 3.6923 192 0.9826 0.2254 0.9826 0.9912
No log 3.7308 194 1.0817 -0.0645 1.0817 1.0400
No log 3.7692 196 0.9464 0.2254 0.9464 0.9728
No log 3.8077 198 0.8091 -0.2737 0.8091 0.8995
No log 3.8462 200 0.8112 -0.0577 0.8112 0.9007
No log 3.8846 202 0.8124 -0.2222 0.8124 0.9013
No log 3.9231 204 1.0621 0.2194 1.0621 1.0306
No log 3.9615 206 1.2244 -0.0331 1.2244 1.1065
No log 4.0 208 1.1341 0.2143 1.1341 1.0650
No log 4.0385 210 0.8919 0.0620 0.8919 0.9444
No log 4.0769 212 0.8040 -0.2737 0.8040 0.8966
No log 4.1154 214 0.8045 -0.2737 0.8045 0.8969
No log 4.1538 216 0.8308 -0.2222 0.8308 0.9115
No log 4.1923 218 0.8678 0.0517 0.8678 0.9316
No log 4.2308 220 0.9015 0.0620 0.9015 0.9495
No log 4.2692 222 0.8878 0.0620 0.8878 0.9422
No log 4.3077 224 0.8591 0.0620 0.8591 0.9269
No log 4.3462 226 0.7616 -0.2791 0.7616 0.8727
No log 4.3846 228 0.8427 0.2787 0.8427 0.9180
No log 4.4231 230 1.0167 0.0763 1.0167 1.0083
No log 4.4615 232 0.9432 0.0763 0.9432 0.9712
No log 4.5 234 0.8272 -0.0577 0.8272 0.9095
No log 4.5385 236 0.8059 -0.2737 0.8059 0.8977
No log 4.5769 238 0.8003 -0.2222 0.8003 0.8946
No log 4.6154 240 0.7844 -0.2737 0.7844 0.8857
No log 4.6538 242 0.7643 -0.2737 0.7643 0.8742
No log 4.6923 244 0.7625 -0.2737 0.7625 0.8732
No log 4.7308 246 0.7448 -0.2737 0.7448 0.8630
No log 4.7692 248 0.7416 -0.2737 0.7416 0.8612
No log 4.8077 250 0.7926 0.0388 0.7926 0.8903
No log 4.8462 252 0.7856 0.0179 0.7856 0.8863
No log 4.8846 254 0.7676 -0.0233 0.7676 0.8761
No log 4.9231 256 0.8041 -0.2692 0.8041 0.8967
No log 4.9615 258 0.8133 -0.2692 0.8133 0.9018
No log 5.0 260 0.8169 0.0 0.8169 0.9038
No log 5.0385 262 0.8758 0.0435 0.8758 0.9358
No log 5.0769 264 0.8643 0.0435 0.8643 0.9297
No log 5.1154 266 0.8238 -0.0421 0.8238 0.9076
No log 5.1538 268 0.8187 -0.0421 0.8187 0.9048
No log 5.1923 270 0.8244 0.0 0.8244 0.9080
No log 5.2308 272 0.8669 0.1987 0.8669 0.9311
No log 5.2692 274 0.9388 0.1987 0.9388 0.9689
No log 5.3077 276 0.8879 0.1987 0.8879 0.9423
No log 5.3462 278 0.7757 -0.0233 0.7757 0.8808
No log 5.3846 280 0.7429 -0.0233 0.7429 0.8619
No log 5.4231 282 0.7494 -0.0421 0.7494 0.8657
No log 5.4615 284 0.7424 -0.0233 0.7424 0.8616
No log 5.5 286 0.7541 -0.0233 0.7541 0.8684
No log 5.5385 288 0.7938 0.0 0.7938 0.8910
No log 5.5769 290 0.7779 -0.0233 0.7779 0.8820
No log 5.6154 292 0.7712 -0.0233 0.7712 0.8782
No log 5.6538 294 0.7564 -0.0233 0.7564 0.8697
No log 5.6923 296 0.7620 -0.2692 0.7620 0.8729
No log 5.7308 298 0.7765 -0.2692 0.7765 0.8812
No log 5.7692 300 0.7440 -0.2737 0.7440 0.8625
No log 5.8077 302 0.7385 -0.0233 0.7385 0.8594
No log 5.8462 304 0.7677 0.0 0.7677 0.8762
No log 5.8846 306 0.7932 0.0517 0.7932 0.8906
No log 5.9231 308 0.7673 0.0179 0.7674 0.8760
No log 5.9615 310 0.7501 0.0 0.7501 0.8661
No log 6.0 312 0.7684 -0.2222 0.7684 0.8766
No log 6.0385 314 0.7965 -0.2737 0.7965 0.8925
No log 6.0769 316 0.8034 -0.2222 0.8034 0.8963
No log 6.1154 318 0.8099 -0.2222 0.8099 0.9000
No log 6.1538 320 0.8354 0.0320 0.8354 0.9140
No log 6.1923 322 0.8302 0.0320 0.8302 0.9112
No log 6.2308 324 0.7952 -0.2222 0.7952 0.8918
No log 6.2692 326 0.7645 -0.2222 0.7645 0.8743
No log 6.3077 328 0.7401 -0.2222 0.7401 0.8603
No log 6.3462 330 0.7172 -0.0233 0.7172 0.8469
No log 6.3846 332 0.7019 0.0 0.7019 0.8378
No log 6.4231 334 0.6867 -0.0233 0.6867 0.8287
No log 6.4615 336 0.6958 0.0179 0.6958 0.8341
No log 6.5 338 0.7062 0.0517 0.7062 0.8404
No log 6.5385 340 0.6683 -0.0233 0.6683 0.8175
No log 6.5769 342 0.6316 -0.0233 0.6316 0.7947
No log 6.6154 344 0.6587 0.1895 0.6587 0.8116
No log 6.6538 346 0.6559 0.1895 0.6559 0.8099
No log 6.6923 348 0.6224 -0.0233 0.6224 0.7889
No log 6.7308 350 0.6286 -0.0233 0.6286 0.7929
No log 6.7692 352 0.6490 -0.0233 0.6490 0.8056
No log 6.8077 354 0.6457 -0.0233 0.6457 0.8035
No log 6.8462 356 0.6445 -0.0233 0.6445 0.8028
No log 6.8846 358 0.6960 -0.0708 0.6960 0.8342
No log 6.9231 360 0.7379 -0.0708 0.7379 0.8590
No log 6.9615 362 0.7378 -0.0708 0.7378 0.8590
No log 7.0 364 0.6967 -0.0577 0.6967 0.8347
No log 7.0385 366 0.6996 -0.2737 0.6996 0.8364
No log 7.0769 368 0.7200 -0.0577 0.7200 0.8485
No log 7.1154 370 0.7229 -0.0577 0.7229 0.8502
No log 7.1538 372 0.7066 -0.2737 0.7066 0.8406
No log 7.1923 374 0.7035 -0.0233 0.7035 0.8387
No log 7.2308 376 0.7009 -0.0233 0.7009 0.8372
No log 7.2692 378 0.7040 -0.2737 0.7040 0.8391
No log 7.3077 380 0.7105 -0.0577 0.7105 0.8429
No log 7.3462 382 0.7139 -0.2737 0.7139 0.8449
No log 7.3846 384 0.7097 -0.0233 0.7097 0.8425
No log 7.4231 386 0.7096 -0.0233 0.7096 0.8424
No log 7.4615 388 0.7028 -0.0233 0.7028 0.8384
No log 7.5 390 0.6909 -0.0233 0.6909 0.8312
No log 7.5385 392 0.6980 -0.0577 0.6980 0.8355
No log 7.5769 394 0.7174 -0.0577 0.7174 0.8470
No log 7.6154 396 0.7270 -0.0577 0.7270 0.8526
No log 7.6538 398 0.7185 -0.0577 0.7185 0.8477
No log 7.6923 400 0.6991 -0.2737 0.6991 0.8361
No log 7.7308 402 0.6835 -0.0233 0.6835 0.8267
No log 7.7692 404 0.6769 -0.0233 0.6769 0.8227
No log 7.8077 406 0.6746 -0.0233 0.6746 0.8213
No log 7.8462 408 0.6820 -0.0233 0.6820 0.8259
No log 7.8846 410 0.6876 -0.0233 0.6876 0.8292
No log 7.9231 412 0.6882 -0.0233 0.6882 0.8296
No log 7.9615 414 0.6793 -0.0233 0.6793 0.8242
No log 8.0 416 0.6788 -0.0233 0.6788 0.8239
No log 8.0385 418 0.6933 -0.0233 0.6933 0.8326
No log 8.0769 420 0.6978 -0.0233 0.6978 0.8353
No log 8.1154 422 0.6930 -0.0233 0.6930 0.8325
No log 8.1538 424 0.6957 -0.0233 0.6957 0.8341
No log 8.1923 426 0.7024 -0.0233 0.7024 0.8381
No log 8.2308 428 0.7183 -0.0233 0.7183 0.8475
No log 8.2692 430 0.7281 -0.2692 0.7281 0.8533
No log 8.3077 432 0.7235 -0.2692 0.7235 0.8506
No log 8.3462 434 0.7164 -0.0233 0.7164 0.8464
No log 8.3846 436 0.7133 -0.0233 0.7133 0.8446
No log 8.4231 438 0.7251 0.0 0.7251 0.8515
No log 8.4615 440 0.7404 0.0179 0.7404 0.8605
No log 8.5 442 0.7406 0.0179 0.7406 0.8606
No log 8.5385 444 0.7254 0.0 0.7254 0.8517
No log 8.5769 446 0.7126 -0.0233 0.7126 0.8442
No log 8.6154 448 0.7101 -0.0233 0.7101 0.8427
No log 8.6538 450 0.7041 -0.0233 0.7041 0.8391
No log 8.6923 452 0.6982 -0.0233 0.6982 0.8356
No log 8.7308 454 0.6968 -0.0233 0.6968 0.8348
No log 8.7692 456 0.6989 -0.0233 0.6989 0.8360
No log 8.8077 458 0.7056 0.0 0.7056 0.8400
No log 8.8462 460 0.7138 0.0 0.7138 0.8449
No log 8.8846 462 0.7199 0.0 0.7199 0.8485
No log 8.9231 464 0.7211 0.0 0.7211 0.8492
No log 8.9615 466 0.7228 -0.0233 0.7228 0.8502
No log 9.0 468 0.7238 -0.0233 0.7238 0.8508
No log 9.0385 470 0.7296 -0.0233 0.7296 0.8542
No log 9.0769 472 0.7390 -0.0233 0.7390 0.8596
No log 9.1154 474 0.7480 -0.2692 0.7480 0.8649
No log 9.1538 476 0.7550 -0.2692 0.7550 0.8689
No log 9.1923 478 0.7552 -0.2692 0.7552 0.8690
No log 9.2308 480 0.7517 -0.2737 0.7517 0.8670
No log 9.2692 482 0.7493 -0.0233 0.7493 0.8656
No log 9.3077 484 0.7494 -0.0233 0.7494 0.8657
No log 9.3462 486 0.7514 0.0 0.7514 0.8668
No log 9.3846 488 0.7508 0.0 0.7508 0.8665
No log 9.4231 490 0.7475 0.0 0.7475 0.8646
No log 9.4615 492 0.7471 0.0 0.7471 0.8644
No log 9.5 494 0.7462 0.0 0.7462 0.8638
No log 9.5385 496 0.7434 0.0 0.7434 0.8622
No log 9.5769 498 0.7389 -0.0233 0.7389 0.8596
0.2607 9.6154 500 0.7348 -0.0233 0.7348 0.8572
0.2607 9.6538 502 0.7327 -0.0233 0.7327 0.8560
0.2607 9.6923 504 0.7323 -0.0233 0.7323 0.8558
0.2607 9.7308 506 0.7318 -0.2737 0.7318 0.8554
0.2607 9.7692 508 0.7311 -0.2737 0.7311 0.8550
0.2607 9.8077 510 0.7306 -0.2737 0.7306 0.8547
0.2607 9.8462 512 0.7299 -0.2737 0.7299 0.8543
0.2607 9.8846 514 0.7289 -0.2737 0.7289 0.8538
0.2607 9.9231 516 0.7280 -0.2737 0.7280 0.8532
0.2607 9.9615 518 0.7276 -0.2737 0.7276 0.8530
0.2607 10.0 520 0.7275 -0.2737 0.7275 0.8529

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
190
Safetensors
Model size
135M params
Tensor type
F32
·
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.

Model tree for MayBashendy/Arabic_FineTuningAraBERT_AugV4_k2_task3_organization_fold1

Finetuned
(4222)
this model