MHGanainy commited on
Commit
06f51a6
1 Parent(s): 8d82753

Model save

Browse files
Files changed (1) hide show
  1. README.md +3 -5
README.md CHANGED
@@ -15,8 +15,6 @@ should probably proofread and complete it, then remove this comment. -->
15
  # gpt2-xl-lora-multi-6
16
 
17
  This model is a fine-tuned version of [openai-community/gpt2-xl](https://huggingface.co/openai-community/gpt2-xl) on an unknown dataset.
18
- It achieves the following results on the evaluation set:
19
- - Loss: 2.0126
20
 
21
  ## Model description
22
 
@@ -46,7 +44,7 @@ The following hyperparameters were used during training:
46
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
47
  - lr_scheduler_type: cosine
48
  - lr_scheduler_warmup_ratio: 0.1
49
- - lr_scheduler_warmup_steps: 7179
50
  - num_epochs: 1
51
  - mixed_precision_training: Native AMP
52
 
@@ -56,8 +54,8 @@ The following hyperparameters were used during training:
56
 
57
  ### Framework versions
58
 
59
- - PEFT 0.13.0
60
- - Transformers 4.45.1
61
  - Pytorch 2.4.1+cu121
62
  - Datasets 3.0.1
63
  - Tokenizers 0.20.0
 
15
  # gpt2-xl-lora-multi-6
16
 
17
  This model is a fine-tuned version of [openai-community/gpt2-xl](https://huggingface.co/openai-community/gpt2-xl) on an unknown dataset.
 
 
18
 
19
  ## Model description
20
 
 
44
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
45
  - lr_scheduler_type: cosine
46
  - lr_scheduler_warmup_ratio: 0.1
47
+ - lr_scheduler_warmup_steps: 4643
48
  - num_epochs: 1
49
  - mixed_precision_training: Native AMP
50
 
 
54
 
55
  ### Framework versions
56
 
57
+ - PEFT 0.13.1
58
+ - Transformers 4.45.2
59
  - Pytorch 2.4.1+cu121
60
  - Datasets 3.0.1
61
  - Tokenizers 0.20.0