LA1512 commited on
Commit
bd91b5b
1 Parent(s): 576cae6

LA1512/result

Browse files
Files changed (3) hide show
  1. README.md +33 -2
  2. model.safetensors +1 -1
  3. training_args.bin +1 -1
README.md CHANGED
@@ -4,9 +4,24 @@ tags:
4
  - generated_from_trainer
5
  datasets:
6
  - pubmed-summarization
 
 
7
  model-index:
8
  - name: results
9
- results: []
 
 
 
 
 
 
 
 
 
 
 
 
 
10
  ---
11
 
12
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
@@ -15,6 +30,13 @@ should probably proofread and complete it, then remove this comment. -->
15
  # results
16
 
17
  This model is a fine-tuned version of [LA1512/PubMed-fine-tune](https://huggingface.co/LA1512/PubMed-fine-tune) on the pubmed-summarization dataset.
 
 
 
 
 
 
 
18
 
19
  ## Model description
20
 
@@ -33,7 +55,7 @@ More information needed
33
  ### Training hyperparameters
34
 
35
  The following hyperparameters were used during training:
36
- - learning_rate: 2e-05
37
  - train_batch_size: 8
38
  - eval_batch_size: 8
39
  - seed: 42
@@ -43,6 +65,15 @@ The following hyperparameters were used during training:
43
  - num_epochs: 3
44
  - label_smoothing_factor: 0.1
45
 
 
 
 
 
 
 
 
 
 
46
  ### Framework versions
47
 
48
  - Transformers 4.39.3
 
4
  - generated_from_trainer
5
  datasets:
6
  - pubmed-summarization
7
+ metrics:
8
+ - rouge
9
  model-index:
10
  - name: results
11
+ results:
12
+ - task:
13
+ name: Sequence-to-sequence Language Modeling
14
+ type: text2text-generation
15
+ dataset:
16
+ name: pubmed-summarization
17
+ type: pubmed-summarization
18
+ config: section
19
+ split: validation
20
+ args: section
21
+ metrics:
22
+ - name: Rouge1
23
+ type: rouge
24
+ value: 40.7402
25
  ---
26
 
27
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
 
30
  # results
31
 
32
  This model is a fine-tuned version of [LA1512/PubMed-fine-tune](https://huggingface.co/LA1512/PubMed-fine-tune) on the pubmed-summarization dataset.
33
+ It achieves the following results on the evaluation set:
34
+ - Loss: 3.6196
35
+ - Rouge1: 40.7402
36
+ - Rouge2: 16.1978
37
+ - Rougel: 24.4278
38
+ - Rougelsum: 36.5282
39
+ - Gen Len: 179.6185
40
 
41
  ## Model description
42
 
 
55
  ### Training hyperparameters
56
 
57
  The following hyperparameters were used during training:
58
+ - learning_rate: 1e-05
59
  - train_batch_size: 8
60
  - eval_batch_size: 8
61
  - seed: 42
 
65
  - num_epochs: 3
66
  - label_smoothing_factor: 0.1
67
 
68
+ ### Training results
69
+
70
+ | Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | Gen Len |
71
+ |:-------------:|:-----:|:----:|:---------------:|:-------:|:-------:|:-------:|:---------:|:--------:|
72
+ | 3.6132 | 1.0 | 2500 | 3.6766 | 40.5092 | 15.7678 | 24.1228 | 36.3318 | 183.7205 |
73
+ | 3.5939 | 2.0 | 5000 | 3.6276 | 40.7583 | 16.1779 | 24.4375 | 36.5537 | 181.4365 |
74
+ | 3.5419 | 3.0 | 7500 | 3.6196 | 40.7402 | 16.1978 | 24.4278 | 36.5282 | 179.6185 |
75
+
76
+
77
  ### Framework versions
78
 
79
  - Transformers 4.39.3
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3c7d91bd434f5ae1765aaa0ca8412882fa09316abfdd28127d94aa0dc0c644bf
3
  size 1020714768
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9f34a4a5a93856e5d6532db0bb0e56e053b236b0106766ea844b9d61906ffd50
3
  size 1020714768
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4d38298726dfd7e007c60b5b330e4b25d706d22c5182372b5875727789c9ceec
3
  size 5048
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8996fa5924b2adc468689d11dc75794fbe5d22d47de3739d11c93031ebbedd55
3
  size 5048