Update README.md
Browse files
README.md
CHANGED
@@ -5,4 +5,38 @@ language:
|
|
5 |
metrics:
|
6 |
- sacrebleu
|
7 |
pipeline_tag: translation
|
8 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
5 |
metrics:
|
6 |
- sacrebleu
|
7 |
pipeline_tag: translation
|
8 |
+
base_model: facebook/mbart-large-cc25
|
9 |
+
---
|
10 |
+
|
11 |
+
# eval
|
12 |
+
|
13 |
+
This model is a fine-tuned version of [facebook/mbart-large-cc25 ](https://huggingface.co/facebook/mbart-large-cc25) on IWSLT14 En-Zh dataset.
|
14 |
+
|
15 |
+
It achieves the following results on the evaluation set:
|
16 |
+
|
17 |
+
- eval_loss: 3.8405
|
18 |
+
- eval_bleu: 3.5173
|
19 |
+
- eval_gen_len: 21.5826
|
20 |
+
|
21 |
+
It achieves the following results on the test set:
|
22 |
+
|
23 |
+
- test_loss: 3.8337
|
24 |
+
- test_bleu: 3.277
|
25 |
+
- test_gen_len: 21.6287
|
26 |
+
|
27 |
+
## Training procedure
|
28 |
+
|
29 |
+
### Training hyperparameters
|
30 |
+
|
31 |
+
The following hyperparameters were used during training:
|
32 |
+
|
33 |
+
- learning_rate: 5e-05
|
34 |
+
- train_batch_size: 7
|
35 |
+
- num_epochs: 9
|
36 |
+
|
37 |
+
### Framework versions
|
38 |
+
|
39 |
+
- Transformers 4.35.2
|
40 |
+
- Pytorch 2.0.1+cu117
|
41 |
+
- Datasets 2.15.0
|
42 |
+
- Tokenizers 0.15.0
|