ritika-kumar commited on
Commit
ddc7a44
1 Parent(s): de51cc6

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +48 -0
README.md ADDED
@@ -0,0 +1,48 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ base_model: facebook/mbart-large-50-many-to-many-mmt
4
+ datasets:
5
+ - ai4bharat/samanantar
6
+ language:
7
+ - en
8
+ - tel
9
+ metrics:
10
+ - bleu
11
+ ---
12
+
13
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
14
+ should probably proofread and complete it, then remove this comment. -->
15
+
16
+ # Finetuning
17
+
18
+ This model is a fine-tuned version of [facebook/mbart-large-50-many-to-many-mmt](https://huggingface.co/facebook/mbart-large-50-many-to-many-mmt) on the samanantar dataset.
19
+ source group: English
20
+ target group: Telugu
21
+ model: transformer
22
+
23
+ ## Model description
24
+
25
+ facebook/mbart-large-50-many-to-many-mmt finetuned for translation task in Telugu language
26
+
27
+ ## Training and evaluation data
28
+
29
+ ai4bharath/samanantar
30
+
31
+ ### Training hyperparameters
32
+
33
+ The following hyperparameters were used during training:
34
+ - learning_rate: 2e-5
35
+ - warmup_steps: 500
36
+ - total_train_batch_size: 8
37
+ - num_epochs: 1
38
+
39
+ ### Benchamark Evaluation
40
+ - BLEU score on Tatoeba: 35.930411196308434
41
+ - BLUE score on IN-22: 14.753637948612852
42
+
43
+ ### Framework versions
44
+
45
+ - Transformers 4.42.3
46
+ - Pytorch 2.1.2
47
+ - Datasets 2.20.0
48
+ - Tokenizers 0.19.1