sphyuung commited on
Commit
9199c0d
·
verified ·
1 Parent(s): 52ecd2c

Training complete

Browse files
Files changed (2) hide show
  1. README.md +56 -76
  2. generation_config.json +16 -16
README.md CHANGED
@@ -1,76 +1,56 @@
1
- ---
2
- library_name: transformers
3
- license: apache-2.0
4
- base_model: Helsinki-NLP/opus-mt-en-fr
5
- tags:
6
- - translation
7
- - generated_from_trainer
8
- datasets:
9
- - kde4
10
- metrics:
11
- - bleu
12
- model-index:
13
- - name: marian-finetuned-kde4-en-to-fr
14
- results:
15
- - task:
16
- name: Sequence-to-sequence Language Modeling
17
- type: text2text-generation
18
- dataset:
19
- name: kde4
20
- type: kde4
21
- config: en-fr
22
- split: train
23
- args: en-fr
24
- metrics:
25
- - name: Bleu
26
- type: bleu
27
- value: 52.926471470290466
28
- ---
29
-
30
- <!-- This model card has been generated automatically according to the information the Trainer had access to. You
31
- should probably proofread and complete it, then remove this comment. -->
32
-
33
- # marian-finetuned-kde4-en-to-fr
34
-
35
- This model is a fine-tuned version of [Helsinki-NLP/opus-mt-en-fr](https://huggingface.co/Helsinki-NLP/opus-mt-en-fr) on the kde4 dataset.
36
- It achieves the following results on the evaluation set:
37
- - Loss: 0.8553
38
- - Model Preparation Time: 0.0024
39
- - Bleu: 52.9265
40
-
41
- ## Model description
42
-
43
- More information needed
44
-
45
- ## Intended uses & limitations
46
-
47
- More information needed
48
-
49
- ## Training and evaluation data
50
-
51
- More information needed
52
-
53
- ## Training procedure
54
-
55
- ### Training hyperparameters
56
-
57
- The following hyperparameters were used during training:
58
- - learning_rate: 2e-05
59
- - train_batch_size: 32
60
- - eval_batch_size: 64
61
- - seed: 42
62
- - optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
63
- - lr_scheduler_type: linear
64
- - num_epochs: 3
65
- - mixed_precision_training: Native AMP
66
-
67
- ### Training results
68
-
69
-
70
-
71
- ### Framework versions
72
-
73
- - Transformers 4.46.2
74
- - Pytorch 2.5.0+cu121
75
- - Datasets 3.1.0
76
- - Tokenizers 0.20.3
 
1
+ ---
2
+ library_name: transformers
3
+ license: apache-2.0
4
+ base_model: Helsinki-NLP/opus-mt-en-fr
5
+ tags:
6
+ - translation
7
+ - generated_from_trainer
8
+ datasets:
9
+ - kde4
10
+ model-index:
11
+ - name: marian-finetuned-kde4-en-to-fr
12
+ results: []
13
+ ---
14
+
15
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
16
+ should probably proofread and complete it, then remove this comment. -->
17
+
18
+ # marian-finetuned-kde4-en-to-fr
19
+
20
+ This model is a fine-tuned version of [Helsinki-NLP/opus-mt-en-fr](https://huggingface.co/Helsinki-NLP/opus-mt-en-fr) on the kde4 dataset.
21
+
22
+ ## Model description
23
+
24
+ More information needed
25
+
26
+ ## Intended uses & limitations
27
+
28
+ More information needed
29
+
30
+ ## Training and evaluation data
31
+
32
+ More information needed
33
+
34
+ ## Training procedure
35
+
36
+ ### Training hyperparameters
37
+
38
+ The following hyperparameters were used during training:
39
+ - learning_rate: 2e-05
40
+ - train_batch_size: 128
41
+ - eval_batch_size: 256
42
+ - seed: 42
43
+ - optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
44
+ - lr_scheduler_type: linear
45
+ - num_epochs: 3
46
+
47
+ ### Training results
48
+
49
+
50
+
51
+ ### Framework versions
52
+
53
+ - Transformers 4.46.2
54
+ - Pytorch 2.5.0+cu121
55
+ - Datasets 3.1.0
56
+ - Tokenizers 0.20.3
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
generation_config.json CHANGED
@@ -1,16 +1,16 @@
1
- {
2
- "bad_words_ids": [
3
- [
4
- 59513
5
- ]
6
- ],
7
- "bos_token_id": 0,
8
- "decoder_start_token_id": 59513,
9
- "eos_token_id": 0,
10
- "forced_eos_token_id": 0,
11
- "max_length": 512,
12
- "num_beams": 4,
13
- "pad_token_id": 59513,
14
- "renormalize_logits": true,
15
- "transformers_version": "4.46.2"
16
- }
 
1
+ {
2
+ "bad_words_ids": [
3
+ [
4
+ 59513
5
+ ]
6
+ ],
7
+ "bos_token_id": 0,
8
+ "decoder_start_token_id": 59513,
9
+ "eos_token_id": 0,
10
+ "forced_eos_token_id": 0,
11
+ "max_length": 512,
12
+ "num_beams": 4,
13
+ "pad_token_id": 59513,
14
+ "renormalize_logits": true,
15
+ "transformers_version": "4.46.2"
16
+ }