NehaShrestha
commited on
Commit
•
b80a132
1
Parent(s):
76ec68d
training end
Browse files
README.md
CHANGED
@@ -12,14 +12,14 @@ model-index:
|
|
12 |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
|
13 |
should probably proofread and complete it, then remove this comment. -->
|
14 |
|
15 |
-
[<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="200" height="32"/>](https://wandb.ai/darvilab/Training%20Sentence%20Translation/runs/
|
16 |
# mbart-en-np-seqtoseq-sentence-translation
|
17 |
|
18 |
This model is a fine-tuned version of [facebook/mbart-large-50-many-to-many-mmt](https://huggingface.co/facebook/mbart-large-50-many-to-many-mmt) on an unknown dataset.
|
19 |
It achieves the following results on the evaluation set:
|
20 |
-
- Loss:
|
21 |
-
- Bleu:
|
22 |
-
- Gen Len:
|
23 |
|
24 |
## Model description
|
25 |
|
@@ -39,8 +39,8 @@ More information needed
|
|
39 |
|
40 |
The following hyperparameters were used during training:
|
41 |
- learning_rate: 2e-05
|
42 |
-
- train_batch_size:
|
43 |
-
- eval_batch_size:
|
44 |
- seed: 42
|
45 |
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
|
46 |
- lr_scheduler_type: linear
|
@@ -50,11 +50,11 @@ The following hyperparameters were used during training:
|
|
50 |
|
51 |
| Training Loss | Epoch | Step | Validation Loss | Bleu | Gen Len |
|
52 |
|:-------------:|:-----:|:----:|:---------------:|:-------:|:-------:|
|
53 |
-
|
|
54 |
-
|
|
55 |
-
|
|
56 |
-
| 0.
|
57 |
-
| 0.
|
58 |
|
59 |
|
60 |
### Framework versions
|
|
|
12 |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
|
13 |
should probably proofread and complete it, then remove this comment. -->
|
14 |
|
15 |
+
[<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="200" height="32"/>](https://wandb.ai/darvilab/Training%20Sentence%20Translation/runs/nul533k2)
|
16 |
# mbart-en-np-seqtoseq-sentence-translation
|
17 |
|
18 |
This model is a fine-tuned version of [facebook/mbart-large-50-many-to-many-mmt](https://huggingface.co/facebook/mbart-large-50-many-to-many-mmt) on an unknown dataset.
|
19 |
It achieves the following results on the evaluation set:
|
20 |
+
- Loss: 1.1896
|
21 |
+
- Bleu: 40.4595
|
22 |
+
- Gen Len: 10.288
|
23 |
|
24 |
## Model description
|
25 |
|
|
|
39 |
|
40 |
The following hyperparameters were used during training:
|
41 |
- learning_rate: 2e-05
|
42 |
+
- train_batch_size: 8
|
43 |
+
- eval_batch_size: 8
|
44 |
- seed: 42
|
45 |
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
|
46 |
- lr_scheduler_type: linear
|
|
|
50 |
|
51 |
| Training Loss | Epoch | Step | Validation Loss | Bleu | Gen Len |
|
52 |
|:-------------:|:-----:|:----:|:---------------:|:-------:|:-------:|
|
53 |
+
| 1.0147 | 1.0 | 1250 | 0.9876 | 40.1501 | 9.885 |
|
54 |
+
| 0.6038 | 2.0 | 2500 | 1.0122 | 40.728 | 10.113 |
|
55 |
+
| 0.3557 | 3.0 | 3750 | 1.0809 | 35.9297 | 10.844 |
|
56 |
+
| 0.2071 | 4.0 | 5000 | 1.1502 | 40.4318 | 10.28 |
|
57 |
+
| 0.1241 | 5.0 | 6250 | 1.1896 | 40.4595 | 10.288 |
|
58 |
|
59 |
|
60 |
### Framework versions
|