vandung commited on
Commit
85122ee
1 Parent(s): 5264128

update model card README.md

Browse files
Files changed (1) hide show
  1. README.md +64 -0
README.md ADDED
@@ -0,0 +1,64 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: mit
3
+ base_model: VietAI/vit5-large
4
+ tags:
5
+ - generated_from_trainer
6
+ metrics:
7
+ - rouge
8
+ model-index:
9
+ - name: t5-large-paraphase-finetuned
10
+ results: []
11
+ ---
12
+
13
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
14
+ should probably proofread and complete it, then remove this comment. -->
15
+
16
+ # t5-large-paraphase-finetuned
17
+
18
+ This model is a fine-tuned version of [VietAI/vit5-large](https://huggingface.co/VietAI/vit5-large) on an unknown dataset.
19
+ It achieves the following results on the evaluation set:
20
+ - Loss: 1.0026
21
+ - Rouge1: 73.1728
22
+ - Rouge2: 57.3725
23
+ - Rougel: 69.3976
24
+ - Rougelsum: 69.3849
25
+ - Gen Len: 9.0143
26
+
27
+ ## Model description
28
+
29
+ More information needed
30
+
31
+ ## Intended uses & limitations
32
+
33
+ More information needed
34
+
35
+ ## Training and evaluation data
36
+
37
+ More information needed
38
+
39
+ ## Training procedure
40
+
41
+ ### Training hyperparameters
42
+
43
+ The following hyperparameters were used during training:
44
+ - learning_rate: 2e-05
45
+ - train_batch_size: 16
46
+ - eval_batch_size: 16
47
+ - seed: 42
48
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
49
+ - lr_scheduler_type: linear
50
+ - num_epochs: 1
51
+
52
+ ### Training results
53
+
54
+ | Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | Gen Len |
55
+ |:-------------:|:-----:|:----:|:---------------:|:-------:|:-------:|:-------:|:---------:|:-------:|
56
+ | 1.0441 | 1.0 | 1969 | 1.0026 | 73.1728 | 57.3725 | 69.3976 | 69.3849 | 9.0143 |
57
+
58
+
59
+ ### Framework versions
60
+
61
+ - Transformers 4.31.0
62
+ - Pytorch 2.0.1
63
+ - Datasets 2.14.4
64
+ - Tokenizers 0.13.3