practicaldreamer commited on
Commit
f14cfcf
·
1 Parent(s): fba1cb8
README.md CHANGED
@@ -1,3 +1,97 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  license: mit
3
  ---
 
1
+ ---
2
+ datasets:
3
+ - practicaldreamer/RPGPT_PublicDomain-ShareGPT
4
+ ---
5
+
6
+ ## Introduction
7
+ This is my first attempt at training a model for long form character interaction using asterisk roleplay format.
8
+ There are plenty of general instruction/answer models but most focus on single responses between an ai and a human.
9
+ My goal for this project is to more closely align the training data with CHARACTER interactions for roleplay.
10
+
11
+ This model is trained on a small synthetic dataset of characters interacting through a variety of scenarios.
12
+ The Characters, Scenarios and interactions were all generated by GPT4.
13
+
14
+ Intended for research, creative writing, entertainment, DnD campaigns? fun!
15
+
16
+ ## Train Summary
17
+ [<img src="https://raw.githubusercontent.com/OpenAccess-AI-Collective/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/OpenAccess-AI-Collective/axolotl)
18
+ ```
19
+ duration: ~1.5hrs
20
+ gpu: 1xA100 80GB
21
+ epochs: 1.0
22
+ speed: 3e-5
23
+ sequence_len: 2048
24
+ gradient_accumulation_steps: 32
25
+ wandb: https://wandb.ai/practicaldreamer/rpgpt/runs/b3sznjpz
26
+ ```
27
+ *Please see the documentation folder for more information*
28
+
29
+ ## Usage
30
+ This LoRA was trained for use with **Neko-Institute-of-Science/LLaMA-13B-HF**
31
+
32
+ Please follow the prompt format outlined below. *Hint: If you're not sure what to put for your character description (or you're lazy) just ask chatgpt to generate it for you! Example:*
33
+ ```
34
+ Generate a short character description for Dr. Watson (The Adventures of Sherlock Holmes) that includes gender, age, MBTI and speech accent using 30 words or less.
35
+ ```
36
+
37
+ ## Prompt Format
38
+ Context/Memory:
39
+ ```
40
+ A chat between a curious user and an artificial intelligence assistant. The assistant gives helpful, detailed, and polite answers to the user's questions.
41
+
42
+ USER: Write a character roleplay dialogue using asterisk roleplay format based on the following character descriptions and scenario. (Each line in your response must be from the perspective of one of these characters)
43
+
44
+ ## Characters
45
+ <User-Character Name> (<User-Character Universe>):
46
+ <User-Character Description>
47
+ <Bot-Character Name> (Bot-Character Universe):
48
+ <Bot-Character Description>
49
+
50
+ ## Scenario:
51
+ <Scenario Description>
52
+
53
+ ASSISTANT:
54
+ ```
55
+ Turn Template:
56
+ ```
57
+ <User-Character Name>: \*<1st person action/sensations/thoughts>\* <Spoken Word> \*<1st person action/sensations/thoughts>\*
58
+ <Bot-Character Name>: \*<1st person action/sensations/thoughts>\* <Spoken Word> \*<1st person action/sensations/thoughts>\*
59
+ <User-Character Name>: \*<1st person action/sensations/thoughts>\* <Spoken Word> \*<1st person action/sensations/thoughts>\*
60
+ <Bot-Character Name>: \*<1st person action/sensations/thoughts>\* <Spoken Word> \*<1st person action/sensations/thoughts>\*
61
+ ...
62
+ ```
63
+
64
+ ## Example
65
+ ```
66
+ A chat between a curious user and an artificial intelligence assistant. The assistant gives helpful, detailed, and polite answers to the user's questions.
67
+
68
+ USER: Write a character roleplay dialogue using asterisk roleplay format based on the following character descriptions and scenario. (Each line in your response must be from the perspective of one of these characters)
69
+
70
+ ## Characters
71
+ Baloo (The Jungle Book):
72
+ Male, middle-aged bear, ENFP, primarily American accent with slight Indian inflections. Wise, carefree, and friendly, he teaches Mowgli the ways of the jungle.
73
+ The Queen of Hearts (Alice's Adventures in Wonderland):
74
+ Female, middle-aged, ENTJ, British English accent; imperious, temperamental, commanding, but quick to change from anger to warmth - the quintessential capricious monarch.
75
+
76
+ ## Scenario:
77
+ In an enchanted glade deep within Wonderland, Baloo finds himself separated from Mowgli and in the Queen of Hearts' territory. Having heard tales of this mercurial ruler, Baloo is intrigued but cautious. To his surprise, the Queen of Hearts approaches him with a broken heart - she has been betrayed by one of her closest confidants.
78
+
79
+ The Queen initially sees Baloo as an opportunity to vent her frustrations and seek validation for her hurt feelings. Baloo, however, quickly recognizes that the Queen's explosive temperament is driving others away and contributing to her own isolation.\n\nUsing his wisdom and patience, Baloo tries to teach the Queen the value of empathy and understanding in forming meaningful relationships. Through their dialogue, they explore their contrasting attitudes toward power, authority, and friendship. The Queen learns that it is not through fear and command that she can find true loyalty, but rather by treating others with kindness.
80
+
81
+ In this somber environment where both characters are out of their comfort zones, can Baloo help the Queen mend her own heart by opening up about his experiences with Mowgli? Will the Queen allow herself to be vulnerable enough to learn from the wise bear? And can these two unlikely friends find common ground in this enchanted glade within Wonderland?
82
+
83
+ ASSISTANT:
84
+ Baloo: *stumbles into the enchanted glade, looking around curiously* Well, I must've taken a wrong turn back there. Mowgli! *calls out for Mowgli*
85
+ The Queen of Hearts: *storms into the glade, clutching a broken heart symbol, face contorted in anger* Off with their heads! Everyone's heads!
86
+ Baloo: *gently scratches his head and approaches the Queen cautiously* Uh, excuse me, ma'am? I couldn't help but notice that you seem rather... upset.
87
+ The Queen of Hearts: *sniffs haughtily and scowls at Baloo* And who are you to pry into my affairs? You should be bowing to your queen!
88
+ Baloo: *bows graciously but maintains eye contact* Begging your pardon, Your Majesty. My name is Baloo, and I've heard tales of your... unique temperament.
89
+ The Queen of Hearts:
90
+ ```
91
+
92
+ ## Thanks
93
+ openaccess-ai-collective https://huggingface.co/openaccess-ai-collective
94
+
95
  ---
96
  license: mit
97
  ---
adapter_config.json ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "base_model_name_or_path": "Neko-Institute-of-Science/LLaMA-13B-HF",
3
+ "bias": "none",
4
+ "fan_in_fan_out": false,
5
+ "inference_mode": true,
6
+ "init_lora_weights": true,
7
+ "lora_alpha": 128,
8
+ "lora_dropout": 0.05,
9
+ "modules_to_save": null,
10
+ "peft_type": "LORA",
11
+ "r": 64,
12
+ "target_modules": [
13
+ "q_proj",
14
+ "v_proj"
15
+ ],
16
+ "task_type": "CAUSAL_LM"
17
+ }
adapter_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5e8d2fbf4ea569ee2ae03234895a83f455ea842b718d146ffe20600f5da562fc
3
+ size 209772877
checkpoint-24/adapter_config.json ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "base_model_name_or_path": "Neko-Institute-of-Science/LLaMA-13B-HF",
3
+ "bias": "none",
4
+ "fan_in_fan_out": false,
5
+ "inference_mode": true,
6
+ "init_lora_weights": true,
7
+ "lora_alpha": 128,
8
+ "lora_dropout": 0.05,
9
+ "modules_to_save": null,
10
+ "peft_type": "LORA",
11
+ "r": 64,
12
+ "target_modules": [
13
+ "q_proj",
14
+ "v_proj"
15
+ ],
16
+ "task_type": "CAUSAL_LM"
17
+ }
checkpoint-24/adapter_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:21d5a3da61188a88f03a7f3c5cbd0ddb9085a634c7af5c289622a4705dda7441
3
+ size 209772877
checkpoint-24/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e750b7ff0d040b0ec211eeedd2baee88d1e56cb588fab658f909ae9aa574d5c0
3
+ size 105251781
checkpoint-24/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:56aef2824e44ec55d99246cc1b218d3829bad6e963903a0cb64b8787c62f870f
3
+ size 14575
checkpoint-24/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a117911b63a944d99f0a7a52c27704af7351c05a8e3781490bd9c8a72430f9bd
3
+ size 627
checkpoint-24/trainer_state.json ADDED
@@ -0,0 +1,208 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": 1.188790202140808,
3
+ "best_model_checkpoint": "output_dir/checkpoint-24",
4
+ "epoch": 0.7224835371589841,
5
+ "global_step": 24,
6
+ "is_hyper_param_search": false,
7
+ "is_local_process_zero": true,
8
+ "is_world_process_zero": true,
9
+ "log_history": [
10
+ {
11
+ "epoch": 0.03,
12
+ "learning_rate": 6.000000000000001e-07,
13
+ "loss": 1.2047,
14
+ "step": 1
15
+ },
16
+ {
17
+ "epoch": 0.06,
18
+ "learning_rate": 1.2000000000000002e-06,
19
+ "loss": 1.2148,
20
+ "step": 2
21
+ },
22
+ {
23
+ "epoch": 0.09,
24
+ "learning_rate": 1.8e-06,
25
+ "loss": 1.2134,
26
+ "step": 3
27
+ },
28
+ {
29
+ "epoch": 0.12,
30
+ "learning_rate": 2.4000000000000003e-06,
31
+ "loss": 1.2068,
32
+ "step": 4
33
+ },
34
+ {
35
+ "epoch": 0.12,
36
+ "eval_loss": 1.1929587125778198,
37
+ "eval_runtime": 4.3547,
38
+ "eval_samples_per_second": 2.526,
39
+ "eval_steps_per_second": 0.459,
40
+ "step": 4
41
+ },
42
+ {
43
+ "epoch": 0.15,
44
+ "learning_rate": 3e-06,
45
+ "loss": 1.2093,
46
+ "step": 5
47
+ },
48
+ {
49
+ "epoch": 0.18,
50
+ "learning_rate": 3.6e-06,
51
+ "loss": 1.2063,
52
+ "step": 6
53
+ },
54
+ {
55
+ "epoch": 0.21,
56
+ "learning_rate": 4.2000000000000004e-06,
57
+ "loss": 1.211,
58
+ "step": 7
59
+ },
60
+ {
61
+ "epoch": 0.24,
62
+ "learning_rate": 4.800000000000001e-06,
63
+ "loss": 1.2042,
64
+ "step": 8
65
+ },
66
+ {
67
+ "epoch": 0.24,
68
+ "eval_loss": 1.1931304931640625,
69
+ "eval_runtime": 4.3471,
70
+ "eval_samples_per_second": 2.53,
71
+ "eval_steps_per_second": 0.46,
72
+ "step": 8
73
+ },
74
+ {
75
+ "epoch": 0.27,
76
+ "learning_rate": 5.4e-06,
77
+ "loss": 1.2042,
78
+ "step": 9
79
+ },
80
+ {
81
+ "epoch": 0.3,
82
+ "learning_rate": 6e-06,
83
+ "loss": 1.1951,
84
+ "step": 10
85
+ },
86
+ {
87
+ "epoch": 0.33,
88
+ "learning_rate": 6.6e-06,
89
+ "loss": 1.2194,
90
+ "step": 11
91
+ },
92
+ {
93
+ "epoch": 0.36,
94
+ "learning_rate": 7.2e-06,
95
+ "loss": 1.1958,
96
+ "step": 12
97
+ },
98
+ {
99
+ "epoch": 0.36,
100
+ "eval_loss": 1.1925488710403442,
101
+ "eval_runtime": 4.3544,
102
+ "eval_samples_per_second": 2.526,
103
+ "eval_steps_per_second": 0.459,
104
+ "step": 12
105
+ },
106
+ {
107
+ "epoch": 0.39,
108
+ "learning_rate": 7.8e-06,
109
+ "loss": 1.2059,
110
+ "step": 13
111
+ },
112
+ {
113
+ "epoch": 0.42,
114
+ "learning_rate": 8.400000000000001e-06,
115
+ "loss": 1.1939,
116
+ "step": 14
117
+ },
118
+ {
119
+ "epoch": 0.45,
120
+ "learning_rate": 9e-06,
121
+ "loss": 1.2042,
122
+ "step": 15
123
+ },
124
+ {
125
+ "epoch": 0.48,
126
+ "learning_rate": 9.600000000000001e-06,
127
+ "loss": 1.1974,
128
+ "step": 16
129
+ },
130
+ {
131
+ "epoch": 0.48,
132
+ "eval_loss": 1.1915441751480103,
133
+ "eval_runtime": 4.3592,
134
+ "eval_samples_per_second": 2.523,
135
+ "eval_steps_per_second": 0.459,
136
+ "step": 16
137
+ },
138
+ {
139
+ "epoch": 0.51,
140
+ "learning_rate": 1.02e-05,
141
+ "loss": 1.1917,
142
+ "step": 17
143
+ },
144
+ {
145
+ "epoch": 0.54,
146
+ "learning_rate": 1.08e-05,
147
+ "loss": 1.2156,
148
+ "step": 18
149
+ },
150
+ {
151
+ "epoch": 0.57,
152
+ "learning_rate": 1.1400000000000001e-05,
153
+ "loss": 1.2204,
154
+ "step": 19
155
+ },
156
+ {
157
+ "epoch": 0.6,
158
+ "learning_rate": 1.2e-05,
159
+ "loss": 1.1997,
160
+ "step": 20
161
+ },
162
+ {
163
+ "epoch": 0.6,
164
+ "eval_loss": 1.190488576889038,
165
+ "eval_runtime": 4.3516,
166
+ "eval_samples_per_second": 2.528,
167
+ "eval_steps_per_second": 0.46,
168
+ "step": 20
169
+ },
170
+ {
171
+ "epoch": 0.63,
172
+ "learning_rate": 1.26e-05,
173
+ "loss": 1.2041,
174
+ "step": 21
175
+ },
176
+ {
177
+ "epoch": 0.66,
178
+ "learning_rate": 1.32e-05,
179
+ "loss": 1.1954,
180
+ "step": 22
181
+ },
182
+ {
183
+ "epoch": 0.69,
184
+ "learning_rate": 1.3800000000000002e-05,
185
+ "loss": 1.1951,
186
+ "step": 23
187
+ },
188
+ {
189
+ "epoch": 0.72,
190
+ "learning_rate": 1.44e-05,
191
+ "loss": 1.2017,
192
+ "step": 24
193
+ },
194
+ {
195
+ "epoch": 0.72,
196
+ "eval_loss": 1.188790202140808,
197
+ "eval_runtime": 4.3616,
198
+ "eval_samples_per_second": 2.522,
199
+ "eval_steps_per_second": 0.459,
200
+ "step": 24
201
+ }
202
+ ],
203
+ "max_steps": 33,
204
+ "num_train_epochs": 1,
205
+ "total_flos": 4.871267940512563e+17,
206
+ "trial_name": null,
207
+ "trial_params": null
208
+ }
checkpoint-24/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a1a9f601b9cbe6df3edd2801886775e4feabba748432d1673b6b79c84c544a83
3
+ size 3963
checkpoint-28/adapter_config.json ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "base_model_name_or_path": "Neko-Institute-of-Science/LLaMA-13B-HF",
3
+ "bias": "none",
4
+ "fan_in_fan_out": false,
5
+ "inference_mode": true,
6
+ "init_lora_weights": true,
7
+ "lora_alpha": 128,
8
+ "lora_dropout": 0.05,
9
+ "modules_to_save": null,
10
+ "peft_type": "LORA",
11
+ "r": 64,
12
+ "target_modules": [
13
+ "q_proj",
14
+ "v_proj"
15
+ ],
16
+ "task_type": "CAUSAL_LM"
17
+ }
checkpoint-28/adapter_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5e8d2fbf4ea569ee2ae03234895a83f455ea842b718d146ffe20600f5da562fc
3
+ size 209772877
checkpoint-28/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:15d6d9156b9c48cbd6638e877e40f2162d325cce254952ab2d172b0473559e31
3
+ size 105251781
checkpoint-28/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:df4fd5b980cade88431c5fd45e7586b73785bbf90f916fda9aa3838ac320199b
3
+ size 14575
checkpoint-28/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:02ee3834d7e7dfd7d399c6f92226ac78e6cd6ee79638559c274dd0f5400d09ff
3
+ size 627
checkpoint-28/trainer_state.json ADDED
@@ -0,0 +1,240 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": 1.1863083839416504,
3
+ "best_model_checkpoint": "output_dir/checkpoint-28",
4
+ "epoch": 0.8428974600188147,
5
+ "global_step": 28,
6
+ "is_hyper_param_search": false,
7
+ "is_local_process_zero": true,
8
+ "is_world_process_zero": true,
9
+ "log_history": [
10
+ {
11
+ "epoch": 0.03,
12
+ "learning_rate": 6.000000000000001e-07,
13
+ "loss": 1.2047,
14
+ "step": 1
15
+ },
16
+ {
17
+ "epoch": 0.06,
18
+ "learning_rate": 1.2000000000000002e-06,
19
+ "loss": 1.2148,
20
+ "step": 2
21
+ },
22
+ {
23
+ "epoch": 0.09,
24
+ "learning_rate": 1.8e-06,
25
+ "loss": 1.2134,
26
+ "step": 3
27
+ },
28
+ {
29
+ "epoch": 0.12,
30
+ "learning_rate": 2.4000000000000003e-06,
31
+ "loss": 1.2068,
32
+ "step": 4
33
+ },
34
+ {
35
+ "epoch": 0.12,
36
+ "eval_loss": 1.1929587125778198,
37
+ "eval_runtime": 4.3547,
38
+ "eval_samples_per_second": 2.526,
39
+ "eval_steps_per_second": 0.459,
40
+ "step": 4
41
+ },
42
+ {
43
+ "epoch": 0.15,
44
+ "learning_rate": 3e-06,
45
+ "loss": 1.2093,
46
+ "step": 5
47
+ },
48
+ {
49
+ "epoch": 0.18,
50
+ "learning_rate": 3.6e-06,
51
+ "loss": 1.2063,
52
+ "step": 6
53
+ },
54
+ {
55
+ "epoch": 0.21,
56
+ "learning_rate": 4.2000000000000004e-06,
57
+ "loss": 1.211,
58
+ "step": 7
59
+ },
60
+ {
61
+ "epoch": 0.24,
62
+ "learning_rate": 4.800000000000001e-06,
63
+ "loss": 1.2042,
64
+ "step": 8
65
+ },
66
+ {
67
+ "epoch": 0.24,
68
+ "eval_loss": 1.1931304931640625,
69
+ "eval_runtime": 4.3471,
70
+ "eval_samples_per_second": 2.53,
71
+ "eval_steps_per_second": 0.46,
72
+ "step": 8
73
+ },
74
+ {
75
+ "epoch": 0.27,
76
+ "learning_rate": 5.4e-06,
77
+ "loss": 1.2042,
78
+ "step": 9
79
+ },
80
+ {
81
+ "epoch": 0.3,
82
+ "learning_rate": 6e-06,
83
+ "loss": 1.1951,
84
+ "step": 10
85
+ },
86
+ {
87
+ "epoch": 0.33,
88
+ "learning_rate": 6.6e-06,
89
+ "loss": 1.2194,
90
+ "step": 11
91
+ },
92
+ {
93
+ "epoch": 0.36,
94
+ "learning_rate": 7.2e-06,
95
+ "loss": 1.1958,
96
+ "step": 12
97
+ },
98
+ {
99
+ "epoch": 0.36,
100
+ "eval_loss": 1.1925488710403442,
101
+ "eval_runtime": 4.3544,
102
+ "eval_samples_per_second": 2.526,
103
+ "eval_steps_per_second": 0.459,
104
+ "step": 12
105
+ },
106
+ {
107
+ "epoch": 0.39,
108
+ "learning_rate": 7.8e-06,
109
+ "loss": 1.2059,
110
+ "step": 13
111
+ },
112
+ {
113
+ "epoch": 0.42,
114
+ "learning_rate": 8.400000000000001e-06,
115
+ "loss": 1.1939,
116
+ "step": 14
117
+ },
118
+ {
119
+ "epoch": 0.45,
120
+ "learning_rate": 9e-06,
121
+ "loss": 1.2042,
122
+ "step": 15
123
+ },
124
+ {
125
+ "epoch": 0.48,
126
+ "learning_rate": 9.600000000000001e-06,
127
+ "loss": 1.1974,
128
+ "step": 16
129
+ },
130
+ {
131
+ "epoch": 0.48,
132
+ "eval_loss": 1.1915441751480103,
133
+ "eval_runtime": 4.3592,
134
+ "eval_samples_per_second": 2.523,
135
+ "eval_steps_per_second": 0.459,
136
+ "step": 16
137
+ },
138
+ {
139
+ "epoch": 0.51,
140
+ "learning_rate": 1.02e-05,
141
+ "loss": 1.1917,
142
+ "step": 17
143
+ },
144
+ {
145
+ "epoch": 0.54,
146
+ "learning_rate": 1.08e-05,
147
+ "loss": 1.2156,
148
+ "step": 18
149
+ },
150
+ {
151
+ "epoch": 0.57,
152
+ "learning_rate": 1.1400000000000001e-05,
153
+ "loss": 1.2204,
154
+ "step": 19
155
+ },
156
+ {
157
+ "epoch": 0.6,
158
+ "learning_rate": 1.2e-05,
159
+ "loss": 1.1997,
160
+ "step": 20
161
+ },
162
+ {
163
+ "epoch": 0.6,
164
+ "eval_loss": 1.190488576889038,
165
+ "eval_runtime": 4.3516,
166
+ "eval_samples_per_second": 2.528,
167
+ "eval_steps_per_second": 0.46,
168
+ "step": 20
169
+ },
170
+ {
171
+ "epoch": 0.63,
172
+ "learning_rate": 1.26e-05,
173
+ "loss": 1.2041,
174
+ "step": 21
175
+ },
176
+ {
177
+ "epoch": 0.66,
178
+ "learning_rate": 1.32e-05,
179
+ "loss": 1.1954,
180
+ "step": 22
181
+ },
182
+ {
183
+ "epoch": 0.69,
184
+ "learning_rate": 1.3800000000000002e-05,
185
+ "loss": 1.1951,
186
+ "step": 23
187
+ },
188
+ {
189
+ "epoch": 0.72,
190
+ "learning_rate": 1.44e-05,
191
+ "loss": 1.2017,
192
+ "step": 24
193
+ },
194
+ {
195
+ "epoch": 0.72,
196
+ "eval_loss": 1.188790202140808,
197
+ "eval_runtime": 4.3616,
198
+ "eval_samples_per_second": 2.522,
199
+ "eval_steps_per_second": 0.459,
200
+ "step": 24
201
+ },
202
+ {
203
+ "epoch": 0.75,
204
+ "learning_rate": 1.5e-05,
205
+ "loss": 1.1908,
206
+ "step": 25
207
+ },
208
+ {
209
+ "epoch": 0.78,
210
+ "learning_rate": 1.56e-05,
211
+ "loss": 1.2032,
212
+ "step": 26
213
+ },
214
+ {
215
+ "epoch": 0.81,
216
+ "learning_rate": 1.62e-05,
217
+ "loss": 1.1876,
218
+ "step": 27
219
+ },
220
+ {
221
+ "epoch": 0.84,
222
+ "learning_rate": 1.6800000000000002e-05,
223
+ "loss": 1.1984,
224
+ "step": 28
225
+ },
226
+ {
227
+ "epoch": 0.84,
228
+ "eval_loss": 1.1863083839416504,
229
+ "eval_runtime": 4.3585,
230
+ "eval_samples_per_second": 2.524,
231
+ "eval_steps_per_second": 0.459,
232
+ "step": 28
233
+ }
234
+ ],
235
+ "max_steps": 33,
236
+ "num_train_epochs": 1,
237
+ "total_flos": 5.6831459305979904e+17,
238
+ "trial_name": null,
239
+ "trial_params": null
240
+ }
checkpoint-28/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a1a9f601b9cbe6df3edd2801886775e4feabba748432d1673b6b79c84c544a83
3
+ size 3963
checkpoint-32/adapter_config.json ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "base_model_name_or_path": "Neko-Institute-of-Science/LLaMA-13B-HF",
3
+ "bias": "none",
4
+ "fan_in_fan_out": false,
5
+ "inference_mode": true,
6
+ "init_lora_weights": true,
7
+ "lora_alpha": 128,
8
+ "lora_dropout": 0.05,
9
+ "modules_to_save": null,
10
+ "peft_type": "LORA",
11
+ "r": 64,
12
+ "target_modules": [
13
+ "q_proj",
14
+ "v_proj"
15
+ ],
16
+ "task_type": "CAUSAL_LM"
17
+ }
checkpoint-32/adapter_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ccee66b163e05aa00135a0438b43b449a13d83365803da247811af514a34eaee
3
+ size 209772877
checkpoint-32/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:62115a02e727a9ceb6366da0ee480ec2e067830213f1b658de47f54eab16cdc4
3
+ size 105251781
checkpoint-32/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d49cbc3e8793c68529a4c3f5e53bfe261a8a6cd135f170aa5936a36af28b2f6e
3
+ size 14575
checkpoint-32/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6a83cccaf2d0ded4c5c9f52e1ad41bad842f388e6da3dfdcc6dacc7a971a91f2
3
+ size 627
checkpoint-32/trainer_state.json ADDED
@@ -0,0 +1,272 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": 1.1863083839416504,
3
+ "best_model_checkpoint": "output_dir/checkpoint-28",
4
+ "epoch": 0.9633113828786454,
5
+ "global_step": 32,
6
+ "is_hyper_param_search": false,
7
+ "is_local_process_zero": true,
8
+ "is_world_process_zero": true,
9
+ "log_history": [
10
+ {
11
+ "epoch": 0.03,
12
+ "learning_rate": 6.000000000000001e-07,
13
+ "loss": 1.2047,
14
+ "step": 1
15
+ },
16
+ {
17
+ "epoch": 0.06,
18
+ "learning_rate": 1.2000000000000002e-06,
19
+ "loss": 1.2148,
20
+ "step": 2
21
+ },
22
+ {
23
+ "epoch": 0.09,
24
+ "learning_rate": 1.8e-06,
25
+ "loss": 1.2134,
26
+ "step": 3
27
+ },
28
+ {
29
+ "epoch": 0.12,
30
+ "learning_rate": 2.4000000000000003e-06,
31
+ "loss": 1.2068,
32
+ "step": 4
33
+ },
34
+ {
35
+ "epoch": 0.12,
36
+ "eval_loss": 1.1929587125778198,
37
+ "eval_runtime": 4.3547,
38
+ "eval_samples_per_second": 2.526,
39
+ "eval_steps_per_second": 0.459,
40
+ "step": 4
41
+ },
42
+ {
43
+ "epoch": 0.15,
44
+ "learning_rate": 3e-06,
45
+ "loss": 1.2093,
46
+ "step": 5
47
+ },
48
+ {
49
+ "epoch": 0.18,
50
+ "learning_rate": 3.6e-06,
51
+ "loss": 1.2063,
52
+ "step": 6
53
+ },
54
+ {
55
+ "epoch": 0.21,
56
+ "learning_rate": 4.2000000000000004e-06,
57
+ "loss": 1.211,
58
+ "step": 7
59
+ },
60
+ {
61
+ "epoch": 0.24,
62
+ "learning_rate": 4.800000000000001e-06,
63
+ "loss": 1.2042,
64
+ "step": 8
65
+ },
66
+ {
67
+ "epoch": 0.24,
68
+ "eval_loss": 1.1931304931640625,
69
+ "eval_runtime": 4.3471,
70
+ "eval_samples_per_second": 2.53,
71
+ "eval_steps_per_second": 0.46,
72
+ "step": 8
73
+ },
74
+ {
75
+ "epoch": 0.27,
76
+ "learning_rate": 5.4e-06,
77
+ "loss": 1.2042,
78
+ "step": 9
79
+ },
80
+ {
81
+ "epoch": 0.3,
82
+ "learning_rate": 6e-06,
83
+ "loss": 1.1951,
84
+ "step": 10
85
+ },
86
+ {
87
+ "epoch": 0.33,
88
+ "learning_rate": 6.6e-06,
89
+ "loss": 1.2194,
90
+ "step": 11
91
+ },
92
+ {
93
+ "epoch": 0.36,
94
+ "learning_rate": 7.2e-06,
95
+ "loss": 1.1958,
96
+ "step": 12
97
+ },
98
+ {
99
+ "epoch": 0.36,
100
+ "eval_loss": 1.1925488710403442,
101
+ "eval_runtime": 4.3544,
102
+ "eval_samples_per_second": 2.526,
103
+ "eval_steps_per_second": 0.459,
104
+ "step": 12
105
+ },
106
+ {
107
+ "epoch": 0.39,
108
+ "learning_rate": 7.8e-06,
109
+ "loss": 1.2059,
110
+ "step": 13
111
+ },
112
+ {
113
+ "epoch": 0.42,
114
+ "learning_rate": 8.400000000000001e-06,
115
+ "loss": 1.1939,
116
+ "step": 14
117
+ },
118
+ {
119
+ "epoch": 0.45,
120
+ "learning_rate": 9e-06,
121
+ "loss": 1.2042,
122
+ "step": 15
123
+ },
124
+ {
125
+ "epoch": 0.48,
126
+ "learning_rate": 9.600000000000001e-06,
127
+ "loss": 1.1974,
128
+ "step": 16
129
+ },
130
+ {
131
+ "epoch": 0.48,
132
+ "eval_loss": 1.1915441751480103,
133
+ "eval_runtime": 4.3592,
134
+ "eval_samples_per_second": 2.523,
135
+ "eval_steps_per_second": 0.459,
136
+ "step": 16
137
+ },
138
+ {
139
+ "epoch": 0.51,
140
+ "learning_rate": 1.02e-05,
141
+ "loss": 1.1917,
142
+ "step": 17
143
+ },
144
+ {
145
+ "epoch": 0.54,
146
+ "learning_rate": 1.08e-05,
147
+ "loss": 1.2156,
148
+ "step": 18
149
+ },
150
+ {
151
+ "epoch": 0.57,
152
+ "learning_rate": 1.1400000000000001e-05,
153
+ "loss": 1.2204,
154
+ "step": 19
155
+ },
156
+ {
157
+ "epoch": 0.6,
158
+ "learning_rate": 1.2e-05,
159
+ "loss": 1.1997,
160
+ "step": 20
161
+ },
162
+ {
163
+ "epoch": 0.6,
164
+ "eval_loss": 1.190488576889038,
165
+ "eval_runtime": 4.3516,
166
+ "eval_samples_per_second": 2.528,
167
+ "eval_steps_per_second": 0.46,
168
+ "step": 20
169
+ },
170
+ {
171
+ "epoch": 0.63,
172
+ "learning_rate": 1.26e-05,
173
+ "loss": 1.2041,
174
+ "step": 21
175
+ },
176
+ {
177
+ "epoch": 0.66,
178
+ "learning_rate": 1.32e-05,
179
+ "loss": 1.1954,
180
+ "step": 22
181
+ },
182
+ {
183
+ "epoch": 0.69,
184
+ "learning_rate": 1.3800000000000002e-05,
185
+ "loss": 1.1951,
186
+ "step": 23
187
+ },
188
+ {
189
+ "epoch": 0.72,
190
+ "learning_rate": 1.44e-05,
191
+ "loss": 1.2017,
192
+ "step": 24
193
+ },
194
+ {
195
+ "epoch": 0.72,
196
+ "eval_loss": 1.188790202140808,
197
+ "eval_runtime": 4.3616,
198
+ "eval_samples_per_second": 2.522,
199
+ "eval_steps_per_second": 0.459,
200
+ "step": 24
201
+ },
202
+ {
203
+ "epoch": 0.75,
204
+ "learning_rate": 1.5e-05,
205
+ "loss": 1.1908,
206
+ "step": 25
207
+ },
208
+ {
209
+ "epoch": 0.78,
210
+ "learning_rate": 1.56e-05,
211
+ "loss": 1.2032,
212
+ "step": 26
213
+ },
214
+ {
215
+ "epoch": 0.81,
216
+ "learning_rate": 1.62e-05,
217
+ "loss": 1.1876,
218
+ "step": 27
219
+ },
220
+ {
221
+ "epoch": 0.84,
222
+ "learning_rate": 1.6800000000000002e-05,
223
+ "loss": 1.1984,
224
+ "step": 28
225
+ },
226
+ {
227
+ "epoch": 0.84,
228
+ "eval_loss": 1.1863083839416504,
229
+ "eval_runtime": 4.3585,
230
+ "eval_samples_per_second": 2.524,
231
+ "eval_steps_per_second": 0.459,
232
+ "step": 28
233
+ },
234
+ {
235
+ "epoch": 0.87,
236
+ "learning_rate": 1.74e-05,
237
+ "loss": 1.202,
238
+ "step": 29
239
+ },
240
+ {
241
+ "epoch": 0.9,
242
+ "learning_rate": 1.8e-05,
243
+ "loss": 1.1991,
244
+ "step": 30
245
+ },
246
+ {
247
+ "epoch": 0.93,
248
+ "learning_rate": 1.86e-05,
249
+ "loss": 1.2004,
250
+ "step": 31
251
+ },
252
+ {
253
+ "epoch": 0.96,
254
+ "learning_rate": 1.9200000000000003e-05,
255
+ "loss": 1.1907,
256
+ "step": 32
257
+ },
258
+ {
259
+ "epoch": 0.96,
260
+ "eval_loss": 1.1872574090957642,
261
+ "eval_runtime": 4.3637,
262
+ "eval_samples_per_second": 2.521,
263
+ "eval_steps_per_second": 0.458,
264
+ "step": 32
265
+ }
266
+ ],
267
+ "max_steps": 33,
268
+ "num_train_epochs": 1,
269
+ "total_flos": 6.495023920683418e+17,
270
+ "trial_name": null,
271
+ "trial_params": null
272
+ }
checkpoint-32/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a1a9f601b9cbe6df3edd2801886775e4feabba748432d1673b6b79c84c544a83
3
+ size 3963
documentation/hyperparameters.yml ADDED
@@ -0,0 +1,69 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ base_model: Neko-Institute-of-Science/LLaMA-13B-HF
2
+ base_model_config: Neko-Institute-of-Science/LLaMA-13B-HF
3
+ model_type: LlamaForCausalLM
4
+ tokenizer_type: LlamaTokenizer
5
+ load_in_8bit: true
6
+ load_4bit:
7
+ datasets:
8
+ - path: practicaldreamer/RPGPT_PublicDomain-ShareGPT
9
+ data_files: RPGPT_PublicDomain_v3-sharegpt.json
10
+ type: sharegpt
11
+ dataset_prepared_path: data/last_run_prepared
12
+ val_set_size: 0.0025
13
+ adapter: lora
14
+ lora_model_dir:
15
+ sequence_len: 2048
16
+ max_packed_sequence_len:
17
+ lora_r: 64
18
+ lora_alpha: 128
19
+ lora_dropout: 0.05
20
+ lora_target_modules:
21
+ - q_proj
22
+ - v_proj
23
+ # - k_proj
24
+ # - o_proj
25
+ lora_fan_in_fan_out: false
26
+ wandb_project:
27
+ wandb_watch:
28
+ wandb_run_id:
29
+ wandb_log_model: checkpoint
30
+ output_dir: output_dir
31
+ batch_size: 128
32
+ micro_batch_size: 4
33
+ eval_batch_size: 1
34
+ num_epochs: 1
35
+ warmup_steps: 50
36
+ logging_steps:
37
+ learning_rate: 0.00003
38
+ optimizer: adamw_bnb_8bit
39
+ torchdistx_path:
40
+ lr_scheduler: cosine
41
+ train_on_inputs: false
42
+ group_by_length: false
43
+ bf16: true
44
+ tf32: true
45
+ gradient_checkpointing: true
46
+ early_stopping_patience: 3
47
+ resume_from_checkpoint:
48
+ auto_resume_from_checkpoints:
49
+ local_rank:
50
+ xformers_attention: true
51
+ flash_attention:
52
+ gptq_groupsize:
53
+ gptq_model_v1:
54
+ save_steps: 4
55
+ debug:
56
+ deepspeed:
57
+ weight_decay: 0.0
58
+ fsdp:
59
+ fsdp_config:
60
+ fsdp_transformer_layer_cls_to_wrap:
61
+ fsdp_min_num_params: 2000
62
+ fsdp_backward_prefetch:
63
+ - backward_pre
64
+ limit_all_gathers: false
65
+ special_tokens:
66
+ pad_token: "[PAD]"
67
+ bos_token: "<s>"
68
+ eos_token: "</s>"
69
+ unk_token: "<unk>"
documentation/preprocessed_sample.txt ADDED
The diff for this file is too large to render. See raw diff
 
documentation/requirements.txt ADDED
@@ -0,0 +1,92 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ accelerate @ git+https://github.com/huggingface/accelerate.git@24ae624d96866e3f993a13fc341ea0dcb68b1470
2
+ aiohttp==3.8.4
3
+ aiosignal==1.3.1
4
+ alpaca-lora-4bit @ git+https://github.com/winglian/alpaca_lora_4bit.git@1b4a376ea816eb2417404b4d1ac27fa16471588a
5
+ appdirs==1.4.4
6
+ async-timeout==4.0.2
7
+ attrdict==2.0.1
8
+ attrs==23.1.0
9
+ -e git+https://github.com/winglian/axolotl@a10a8265efde4ec61037560e3b8e2e31dab984af#egg=axolotl
10
+ bitsandbytes==0.37.2
11
+ black==23.3.0
12
+ certifi==2022.12.7
13
+ charset-normalizer==3.1.0
14
+ click==8.1.3
15
+ cmake==3.26.3
16
+ colorama==0.4.6
17
+ datasets==2.12.0
18
+ deepspeed==0.9.4
19
+ dill==0.3.6
20
+ docker-pycreds==0.4.0
21
+ einops==0.6.1
22
+ filelock==3.12.0
23
+ fire==0.5.0
24
+ flash-attn==1.0.4
25
+ frozenlist==1.3.3
26
+ fsspec==2023.4.0
27
+ gitdb==4.0.10
28
+ GitPython==3.1.31
29
+ hjson==3.1.0
30
+ huggingface-hub==0.14.1
31
+ idna==3.4
32
+ Jinja2==3.1.2
33
+ lit==16.0.2
34
+ MarkupSafe==2.1.2
35
+ mpmath==1.3.0
36
+ multidict==6.0.4
37
+ multiprocess==0.70.14
38
+ mypy-extensions==1.0.0
39
+ networkx==3.1
40
+ ninja==1.11.1
41
+ numpy==1.24.3
42
+ nvidia-cublas-cu11==11.10.3.66
43
+ nvidia-cuda-cupti-cu11==11.7.101
44
+ nvidia-cuda-nvrtc-cu11==11.7.99
45
+ nvidia-cuda-runtime-cu11==11.7.99
46
+ nvidia-cudnn-cu11==8.5.0.96
47
+ nvidia-cufft-cu11==10.9.0.58
48
+ nvidia-curand-cu11==10.2.10.91
49
+ nvidia-cusolver-cu11==11.4.0.1
50
+ nvidia-cusparse-cu11==11.7.4.91
51
+ nvidia-nccl-cu11==2.14.3
52
+ nvidia-nvtx-cu11==11.7.91
53
+ packaging==23.1
54
+ pandas==2.0.1
55
+ pathspec==0.11.1
56
+ pathtools==0.1.2
57
+ peft @ git+https://github.com/huggingface/peft.git@70af02a2bca5a63921790036b2c9430edf4037e2
58
+ platformdirs==3.5.0
59
+ protobuf==4.22.4
60
+ psutil==5.9.5
61
+ py-cpuinfo==9.0.0
62
+ pyarrow==12.0.0
63
+ pydantic==1.10.7
64
+ pyre-extensions==0.0.29
65
+ python-dateutil==2.8.2
66
+ pytz==2023.3
67
+ PyYAML==6.0
68
+ regex==2023.5.5
69
+ requests==2.30.0
70
+ responses==0.18.0
71
+ safetensors==0.3.1
72
+ sentencepiece==0.1.99
73
+ sentry-sdk==1.21.1
74
+ setproctitle==1.3.2
75
+ six==1.16.0
76
+ smmap==5.0.0
77
+ sympy==1.11.1
78
+ termcolor==2.3.0
79
+ tokenizers==0.13.3
80
+ tomli==2.0.1
81
+ torch==2.0.0
82
+ tqdm==4.65.0
83
+ transformers @ git+https://github.com/huggingface/transformers.git@799df10aef3abfe6158c83daf0a9eacf8f6f0a1f
84
+ triton==2.0.0
85
+ typing-inspect==0.8.0
86
+ typing_extensions==4.5.0
87
+ tzdata==2023.3
88
+ urllib3==2.0.2
89
+ wandb==0.15.4
90
+ xformers==0.0.19
91
+ xxhash==3.2.0
92
+ yarl==1.9.2
documentation/wandb.info ADDED
@@ -0,0 +1 @@
 
 
1
+ https://wandb.ai/practicaldreamer/rpgpt/runs/b3sznjpz