You need to agree to share your contact information to access this model

This repository is publicly accessible, but you have to accept the conditions to access its files and content.

Log in or Sign Up to review the conditions and access this model content.

facebook mms-1b-all xhosa - Beijuka Bruno

This model is a fine-tuned version of facebook/mms-1b-all on the NCHLT_speech_corpus/Xhosa dataset. It achieves the following results on the evaluation set:

  • Loss: 0.2822
  • Model Preparation Time: 0.0163
  • Wer: 0.4470
  • Cer: 0.0841

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 0.0003
  • train_batch_size: 4
  • eval_batch_size: 8
  • seed: 42
  • gradient_accumulation_steps: 8
  • total_train_batch_size: 32
  • optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
  • lr_scheduler_type: linear
  • lr_scheduler_warmup_steps: 100
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Model Preparation Time Wer Cer
53.6849 0.9955 167 0.2975 0.0163 0.3955 0.0660
2.6944 1.9955 334 0.2671 0.0163 0.3716 0.0612
2.4353 2.9955 501 0.2543 0.0163 0.3468 0.0580
2.2597 3.9955 668 0.2459 0.0163 0.3445 0.0565
2.1704 4.9955 835 0.2392 0.0163 0.3251 0.0547
2.0144 5.9955 1002 0.2353 0.0163 0.3160 0.0539
1.9278 6.9955 1169 0.2350 0.0163 0.3143 0.0539
1.849 7.9955 1336 0.2312 0.0163 0.3039 0.0520
1.811 8.9955 1503 0.2289 0.0163 0.3065 0.0530
1.7173 9.9955 1670 0.2311 0.0163 0.2967 0.0522
1.6504 10.9955 1837 0.2205 0.0163 0.2960 0.0511
1.5737 11.9955 2004 0.2227 0.0163 0.2904 0.0506
1.5547 12.9955 2171 0.2218 0.0163 0.2836 0.0502
1.5517 13.9955 2338 0.2236 0.0163 0.2917 0.0507
1.5031 14.9955 2505 0.2162 0.0163 0.2826 0.0493
1.4097 15.9955 2672 0.2187 0.0163 0.2773 0.0492
1.4143 16.9955 2839 0.2176 0.0163 0.2809 0.0485
1.36 17.9955 3006 0.2200 0.0163 0.2675 0.0479
1.3312 18.9955 3173 0.2163 0.0163 0.2672 0.0482
1.3369 19.9955 3340 0.2199 0.0163 0.2678 0.0481
1.3036 20.9955 3507 0.2224 0.0163 0.2714 0.0482
1.2558 21.9955 3674 0.2244 0.0163 0.2656 0.0478
1.2058 22.9955 3841 0.2192 0.0163 0.2642 0.0481
1.167 23.9955 4008 0.2225 0.0163 0.2561 0.0471
1.1849 24.9955 4175 0.2275 0.0163 0.2610 0.0473
1.1378 25.9955 4342 0.2246 0.0163 0.2610 0.0474
1.1095 26.9955 4509 0.2295 0.0163 0.2538 0.0464
1.1042 27.9955 4676 0.2243 0.0163 0.2518 0.0462
1.0537 28.9955 4843 0.2293 0.0163 0.2531 0.0467
1.0335 29.9955 5010 0.2264 0.0163 0.2531 0.0456
1.0453 30.9955 5177 0.2232 0.0163 0.2525 0.0452
1.0099 31.9955 5344 0.2285 0.0163 0.2551 0.0467
0.9826 32.9955 5511 0.2345 0.0163 0.2570 0.0470
0.9615 33.9955 5678 0.2361 0.0163 0.2587 0.0471
0.9583 34.9955 5845 0.2340 0.0163 0.2528 0.0452
0.9421 35.9955 6012 0.2339 0.0163 0.2443 0.0452
0.9268 36.9955 6179 0.2350 0.0163 0.2518 0.0455
0.9003 37.9955 6346 0.2380 0.0163 0.2472 0.0451
0.9068 38.9955 6513 0.2424 0.0163 0.2515 0.0459
0.8845 39.9955 6680 0.2466 0.0163 0.2525 0.0462
0.88 40.9955 6847 0.2423 0.0163 0.2387 0.0441
0.8683 41.9955 7014 0.2448 0.0163 0.2528 0.0463
0.8535 42.9955 7181 0.2498 0.0163 0.2492 0.0455
0.8413 43.9955 7348 0.2431 0.0163 0.2511 0.0463
0.8147 44.9955 7515 0.2416 0.0163 0.2426 0.0449
0.8062 45.9955 7682 0.2483 0.0163 0.2479 0.0455
0.7876 46.9955 7849 0.2477 0.0163 0.2531 0.0463
0.8108 47.9955 8016 0.2469 0.0163 0.2462 0.0454
0.7689 48.9955 8183 0.2539 0.0163 0.2489 0.0460
0.7609 49.9955 8350 0.2535 0.0163 0.2453 0.0446
0.7442 50.9955 8517 0.2603 0.0163 0.2485 0.0459

Framework versions

  • Transformers 4.47.1
  • Pytorch 2.1.0+cu118
  • Datasets 3.2.0
  • Tokenizers 0.21.0
Downloads last month
2
Safetensors
Model size
965M params
Tensor type
F32
·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Model tree for asr-africa/mms-1B_all_NCHLT_XHOSA_5hr_v1

Finetuned
(214)
this model

Collection including asr-africa/mms-1B_all_NCHLT_XHOSA_5hr_v1

Evaluation results