--- license: apache-2.0 base_model: facebook/mbart-large-50-many-to-many-mmt datasets: - ai4bharat/samanantar language: - en - tel metrics: - bleu --- # Finetuning This model is a fine-tuned version of [facebook/mbart-large-50-many-to-many-mmt](https://huggingface.co/facebook/mbart-large-50-many-to-many-mmt) on the samanantar dataset. source group: English target group: Telugu model: transformer ## Model description facebook/mbart-large-50-many-to-many-mmt finetuned for translation task in Telugu language ## Training and evaluation data ai4bharath/samanantar ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-5 - warmup_steps: 500 - total_train_batch_size: 8 - num_epochs: 1 ### Benchamark Evaluation - BLEU score on Tatoeba: 35.930411196308434 - BLUE score on IN-22: 14.753637948612852 ### Framework versions - Transformers 4.42.3 - Pytorch 2.1.2 - Datasets 2.20.0 - Tokenizers 0.19.1