NeuralNovel commited on
Commit
3c38ae7
1 Parent(s): d668cc2

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +45 -32
README.md CHANGED
@@ -1,4 +1,6 @@
1
  ---
 
 
2
  tags:
3
  - generated_from_trainer
4
  model-index:
@@ -9,12 +11,23 @@ model-index:
9
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
10
  should probably proofread and complete it, then remove this comment. -->
11
 
12
- [<img src="https://raw.githubusercontent.com/OpenAccess-AI-Collective/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/OpenAccess-AI-Collective/axolotl)
13
- <details><summary>See axolotl config</summary>
 
 
 
 
 
 
 
 
 
 
 
 
14
 
15
- axolotl version: `0.4.0`
16
  ```yaml
17
- base_model: out/Mistral-DPO
18
  model_type: AutoModelForCausalLM
19
  tokenizer_type: AutoTokenizer
20
  is_mistral_derived_model: true
@@ -23,14 +36,28 @@ load_in_8bit: false
23
  load_in_4bit: false
24
  strict: false
25
 
26
- rl: dpo
27
  datasets:
28
- - path: NeuralNovel/Neural-DPO
29
- type: chatml.intel
30
- split: train
 
 
 
 
 
31
  format: "[INST] {instruction} [/INST]"
32
  no_input_format: "[INST] {instruction} [/INST]"
33
-
 
 
 
 
 
 
 
 
 
 
34
  dataset_prepared_path:
35
  val_set_size: 0.05
36
  output_dir: ./out
@@ -48,7 +75,7 @@ wandb_log_model:
48
 
49
  gradient_accumulation_steps: 4
50
  micro_batch_size: 2
51
- num_epochs: 6
52
  optimizer: adamw_bnb_8bit
53
  lr_scheduler: cosine
54
  learning_rate: 0.000005
@@ -84,42 +111,28 @@ special_tokens:
84
 
85
  ```
86
 
87
- </details><br>
88
-
89
- # out
90
-
91
- This model was trained from scratch on an unknown dataset.
92
-
93
- ## Model description
94
-
95
- More information needed
96
-
97
- ## Intended uses & limitations
98
-
99
- More information needed
100
-
101
- ## Training and evaluation data
102
-
103
- More information needed
104
-
105
- ## Training procedure
106
-
107
  ### Training hyperparameters
108
 
109
  The following hyperparameters were used during training:
110
  - learning_rate: 5e-06
111
  - train_batch_size: 2
112
- - eval_batch_size: 8
113
  - seed: 42
114
  - gradient_accumulation_steps: 4
115
  - total_train_batch_size: 8
116
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
117
  - lr_scheduler_type: cosine
118
  - lr_scheduler_warmup_steps: 10
119
- - training_steps: 801
120
 
121
  ### Training results
122
 
 
 
 
 
 
 
123
 
124
 
125
  ### Framework versions
 
1
  ---
2
+ license: apache-2.0
3
+ base_model: mistralai/Mistral-7B-v0.1
4
  tags:
5
  - generated_from_trainer
6
  model-index:
 
11
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
12
  should probably proofread and complete it, then remove this comment. -->
13
 
14
+ ![image/jpeg](https://cdn-uploads.huggingface.co/production/uploads/645cfe4603fc86c46b3e46d1/FXt-g2q8JE-l77_gp23T3.jpeg)
15
+
16
+ # NeuralNovel/Senzu-7B-v0.1
17
+
18
+ Embracing a quiet *storm* ..
19
+
20
+
21
+ ## Model Details
22
+
23
+ This model is a full parameter fine-tuned version of [mistralai/Mistral-7B-v0.1](https://huggingface.co/mistralai/Mistral-7B-v0.1)
24
+
25
+ Trained on the Neural-DPO, metamath_gsm8k and RPGPT_PublicDomain-alpaca dataset.
26
+
27
+ This model excels at character roleplay, also with the ability of responding accurately to a wide variety of complex questions.
28
 
 
29
  ```yaml
30
+ base_model: mistralai/Mistral-7B-v0.1
31
  model_type: AutoModelForCausalLM
32
  tokenizer_type: AutoTokenizer
33
  is_mistral_derived_model: true
 
36
  load_in_4bit: false
37
  strict: false
38
 
 
39
  datasets:
40
+ - path: practical-dreamer/RPGPT_PublicDomain-alpaca
41
+ type: alpaca
42
+ format: "[INST] {instruction} [/INST]"
43
+ no_input_format: "[INST] {instruction} [/INST]"
44
+
45
+ datasets:
46
+ - path: shuyuej/metamath_gsm8k
47
+ type: jeopardy
48
  format: "[INST] {instruction} [/INST]"
49
  no_input_format: "[INST] {instruction} [/INST]"
50
+
51
+ datasets:
52
+ - path: NeuralNovel/Neural-DPO
53
+ type:
54
+ system_prompt: ""
55
+ field_system: system
56
+ field_instruction: chosen
57
+ field_output: chosen
58
+ format: "[INST] {instruction} [/INST]"
59
+ no_input_format: "[INST] {instruction} [/INST]"
60
+
61
  dataset_prepared_path:
62
  val_set_size: 0.05
63
  output_dir: ./out
 
75
 
76
  gradient_accumulation_steps: 4
77
  micro_batch_size: 2
78
+ num_epochs: 1
79
  optimizer: adamw_bnb_8bit
80
  lr_scheduler: cosine
81
  learning_rate: 0.000005
 
111
 
112
  ```
113
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
114
  ### Training hyperparameters
115
 
116
  The following hyperparameters were used during training:
117
  - learning_rate: 5e-06
118
  - train_batch_size: 2
119
+ - eval_batch_size: 2
120
  - seed: 42
121
  - gradient_accumulation_steps: 4
122
  - total_train_batch_size: 8
123
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
124
  - lr_scheduler_type: cosine
125
  - lr_scheduler_warmup_steps: 10
126
+ - num_epochs: 1
127
 
128
  ### Training results
129
 
130
+ | Training Loss | Epoch | Step | Validation Loss |
131
+ |:-------------:|:-----:|:----:|:---------------:|
132
+ | 0.2061 | 0.01 | 1 | 0.3139 |
133
+ | 0.0 | 0.25 | 32 | 0.0000 |
134
+ | 0.0 | 0.5 | 64 | 0.0010 |
135
+ | 0.0 | 0.76 | 96 | 0.0000 |
136
 
137
 
138
  ### Framework versions