AndyYo commited on
Commit
cef8cd2
1 Parent(s): e6c8ca2

End of training

Browse files
Files changed (1) hide show
  1. README.md +4 -3
README.md CHANGED
@@ -1,6 +1,6 @@
1
  ---
2
  license: mit
3
- base_model: TheBloke/zephyr-7B-beta-GPTQ
4
  tags:
5
  - generated_from_trainer
6
  model-index:
@@ -13,7 +13,7 @@ should probably proofread and complete it, then remove this comment. -->
13
 
14
  # zephyr-support-chatbot
15
 
16
- This model is a fine-tuned version of [TheBloke/zephyr-7B-beta-GPTQ](https://huggingface.co/TheBloke/zephyr-7B-beta-GPTQ) on an unknown dataset.
17
 
18
  ## Model description
19
 
@@ -39,6 +39,7 @@ The following hyperparameters were used during training:
39
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
40
  - lr_scheduler_type: cosine
41
  - training_steps: 250
 
42
 
43
  ### Training results
44
 
@@ -46,7 +47,7 @@ The following hyperparameters were used during training:
46
 
47
  ### Framework versions
48
 
49
- - Transformers 4.34.1
50
  - Pytorch 2.1.0+cu121
51
  - Datasets 2.14.6
52
  - Tokenizers 0.14.1
 
1
  ---
2
  license: mit
3
+ base_model: TheBloke/zephyr-7B-alpha-GPTQ
4
  tags:
5
  - generated_from_trainer
6
  model-index:
 
13
 
14
  # zephyr-support-chatbot
15
 
16
+ This model is a fine-tuned version of [TheBloke/zephyr-7B-alpha-GPTQ](https://huggingface.co/TheBloke/zephyr-7B-alpha-GPTQ) on the None dataset.
17
 
18
  ## Model description
19
 
 
39
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
40
  - lr_scheduler_type: cosine
41
  - training_steps: 250
42
+ - mixed_precision_training: Native AMP
43
 
44
  ### Training results
45
 
 
47
 
48
  ### Framework versions
49
 
50
+ - Transformers 4.35.0
51
  - Pytorch 2.1.0+cu121
52
  - Datasets 2.14.6
53
  - Tokenizers 0.14.1