SaffalPoosh commited on
Commit
27cfa84
1 Parent(s): b2bda46

Model save

Browse files
Files changed (2) hide show
  1. README.md +2 -11
  2. generation_config.json +1 -0
README.md CHANGED
@@ -3,8 +3,6 @@ license: apache-2.0
3
  base_model: google-t5/t5-small
4
  tags:
5
  - generated_from_trainer
6
- metrics:
7
- - rouge
8
  model-index:
9
  - name: tst-summarization
10
  results: []
@@ -13,17 +11,10 @@ model-index:
13
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
14
  should probably proofread and complete it, then remove this comment. -->
15
 
16
- [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="200" height="32"/>](https://wandb.ai/ai_experment/huggingface/runs/v8mp6bhe)
17
  # tst-summarization
18
 
19
  This model is a fine-tuned version of [google-t5/t5-small](https://huggingface.co/google-t5/t5-small) on an unknown dataset.
20
- It achieves the following results on the evaluation set:
21
- - Loss: 0.6220
22
- - Rouge1: 65.6704
23
- - Rouge2: 51.3605
24
- - Rougel: 65.6565
25
- - Rougelsum: 65.3846
26
- - Gen Len: 21.1074
27
 
28
  ## Model description
29
 
@@ -48,7 +39,7 @@ The following hyperparameters were used during training:
48
  - seed: 42
49
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
50
  - lr_scheduler_type: linear
51
- - num_epochs: 3.0
52
 
53
  ### Training results
54
 
 
3
  base_model: google-t5/t5-small
4
  tags:
5
  - generated_from_trainer
 
 
6
  model-index:
7
  - name: tst-summarization
8
  results: []
 
11
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
12
  should probably proofread and complete it, then remove this comment. -->
13
 
14
+ [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="200" height="32"/>](https://wandb.ai/ai_experment/huggingface/runs/anjcgvjp)
15
  # tst-summarization
16
 
17
  This model is a fine-tuned version of [google-t5/t5-small](https://huggingface.co/google-t5/t5-small) on an unknown dataset.
 
 
 
 
 
 
 
18
 
19
  ## Model description
20
 
 
39
  - seed: 42
40
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
41
  - lr_scheduler_type: linear
42
+ - num_epochs: 10.0
43
 
44
  ### Training results
45
 
generation_config.json CHANGED
@@ -1,4 +1,5 @@
1
  {
 
2
  "decoder_start_token_id": 0,
3
  "eos_token_id": 1,
4
  "pad_token_id": 0,
 
1
  {
2
+ "_from_model_config": true,
3
  "decoder_start_token_id": 0,
4
  "eos_token_id": 1,
5
  "pad_token_id": 0,