yadheedhya commited on
Commit
7fc7b34
·
1 Parent(s): ee0d2b4

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +0 -74
README.md CHANGED
@@ -1,77 +1,3 @@
1
- ---
2
- license: apache-2.0
3
- tags:
4
- - generated_from_trainer
5
- datasets:
6
- - cnn_dailymail
7
- metrics:
8
- - rouge
9
- model-index:
10
- - name: base
11
- results:
12
- - task:
13
- name: Summarization
14
- type: summarization
15
- dataset:
16
- name: cnn_dailymail 3.0.0
17
- type: cnn_dailymail
18
- config: 3.0.0
19
- split: validation
20
- args: 3.0.0
21
- metrics:
22
- - name: Rouge1
23
- type: rouge
24
- value: 42.1388
25
- ---
26
 
27
- <!-- This model card has been generated automatically according to the information the Trainer had access to. You
28
- should probably proofread and complete it, then remove this comment. -->
29
 
30
- # base
31
 
32
- This model is a fine-tuned version of [google/flan-t5-base](https://huggingface.co/google/flan-t5-base) on the cnn_dailymail 3.0.0 dataset.
33
- It achieves the following results on the evaluation set:
34
- - Loss: 1.4232
35
- - Rouge1: 42.1388
36
- - Rouge2: 19.7696
37
- - Rougel: 30.1512
38
- - Rougelsum: 39.3222
39
- - Gen Len: 71.8562
40
-
41
- ## Model description
42
-
43
- More information needed
44
-
45
- ## Intended uses & limitations
46
-
47
- More information needed
48
-
49
- ## Training and evaluation data
50
-
51
- More information needed
52
-
53
- ## Training procedure
54
-
55
- ### Training hyperparameters
56
-
57
- The following hyperparameters were used during training:
58
- - learning_rate: 0.0001
59
- - train_batch_size: 1
60
- - eval_batch_size: 4
61
- - seed: 42
62
- - gradient_accumulation_steps: 64
63
- - total_train_batch_size: 64
64
- - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
65
- - lr_scheduler_type: constant
66
- - num_epochs: 3.0
67
-
68
- ### Training results
69
-
70
-
71
-
72
- ### Framework versions
73
-
74
- - Transformers 4.27.0.dev0
75
- - Pytorch 1.13.0+cu117
76
- - Datasets 2.7.1
77
- - Tokenizers 0.12.1
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
 
 
 
2
 
 
3