chat
Delta-Vector commited on
Commit
35fafef
1 Parent(s): 69c6c47

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +121 -53
README.md CHANGED
@@ -1,26 +1,127 @@
1
  ---
2
- library_name: transformers
3
- license: llama3
4
- base_model: arcee-ai/Llama-3.1-SuperNova-Lite
5
  tags:
6
- - generated_from_trainer
7
- model-index:
8
- - name: outputs
9
- results: []
 
 
 
 
 
 
 
 
 
10
  ---
11
- ### exl2 quant (measurement.json in main branch)
12
  ---
13
- ### check revisions for quants
14
  ---
15
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
16
 
17
- <!-- This model card has been generated automatically according to the information the Trainer had access to. You
18
- should probably proofread and complete it, then remove this comment. -->
 
 
 
 
 
 
 
 
19
 
20
- [<img src="https://raw.githubusercontent.com/axolotl-ai-cloud/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/axolotl-ai-cloud/axolotl)
21
  <details><summary>See axolotl config</summary>
22
 
23
- axolotl version: `0.4.1`
24
  ```yaml
25
  base_model: arcee-ai/Llama-3.1-SuperNova-Lite
26
  model_type: AutoModelForCausalLM
@@ -109,53 +210,20 @@ special_tokens:
109
  pad_token: <|finetune_right_pad_id|>
110
  eos_token: <|eot_id|>
111
 
112
-
113
  ```
114
 
115
  </details><br>
116
 
117
- # outputs
118
-
119
- This model is a fine-tuned version of [arcee-ai/Llama-3.1-SuperNova-Lite](https://huggingface.co/arcee-ai/Llama-3.1-SuperNova-Lite) on the None dataset.
120
-
121
- ## Model description
122
-
123
- More information needed
124
-
125
- ## Intended uses & limitations
126
-
127
- More information needed
128
-
129
- ## Training and evaluation data
130
-
131
- More information needed
132
-
133
- ## Training procedure
134
-
135
- ### Training hyperparameters
136
 
137
- The following hyperparameters were used during training:
138
- - learning_rate: 1e-05
139
- - train_batch_size: 1
140
- - eval_batch_size: 1
141
- - seed: 42
142
- - distributed_type: multi-GPU
143
- - num_devices: 4
144
- - gradient_accumulation_steps: 2
145
- - total_train_batch_size: 8
146
- - total_eval_batch_size: 4
147
- - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
148
- - lr_scheduler_type: cosine
149
- - lr_scheduler_warmup_steps: 15
150
- - num_epochs: 4
151
 
152
- ### Training results
153
 
 
 
154
 
 
155
 
156
- ### Framework versions
157
 
158
- - Transformers 4.44.2
159
- - Pytorch 2.3.1+cu121
160
- - Datasets 2.20.0
161
- - Tokenizers 0.19.1
 
1
  ---
2
+ license: agpl-3.0
 
 
3
  tags:
4
+ - chat
5
+ datasets:
6
+ - NewEden/OpenCAI-ShareGPT
7
+ - NewEden/Roleplay-Logs-Sharegpt-Ngram-cleaned
8
+ License: agpl-3.0
9
+ Language:
10
+ - En
11
+ Pipeline_tag: text-generation
12
+ Base_model: arcee-ai/Llama-3.1-SuperNova-Lite
13
+ Tags:
14
+ - Chat
15
+ ---
16
+
17
  ---
18
+ ### these are exl2 quants (measurement.json in main branch)
19
  ---
20
+ ### check revisions for different bpw's
21
  ---
22
 
23
+ An experimental finetune based on the Llama3.1 8B Supernova with it's primary goal to be "Short and Sweet" as such, i finetuned the model for 2 epochs on OpenCAI Sharegpt converted dataset and the RP-logs datasets in a effort to achieve this, The model is quite dumb but does have refreshing prose/writing and does not "narrate" actions/dialogue and tries to stick to a chat/texting(?) format.
24
+
25
+ # Quants
26
+
27
+ GGUF: https://huggingface.co/Delta-Vector/Control-8B-gguf
28
+
29
+ EXL2 (Thanks Lucy <3) : https://huggingface.co/Delta-Vector/Control-8B-EXL2
30
+
31
+
32
+ ## Prompting
33
+ Model has been tuned with the LLama-Instruct formatting. A typical input would look like this:
34
+
35
+ ```py
36
+ """<|begin_of_text|><|start_header_id|>system<|end_header_id|>
37
+ You are an AI built to rid the world of bonds and journeys!<|eot_id|><|start_header_id|>user<|end_header_id|>
38
+ Bro i just wanna know what is 2+2?<|eot_id|><|start_header_id|>assistant<|end_header_id|>
39
+ """
40
+
41
+ ```
42
+
43
+ *Also note that ChatML may work as well, and might change how the model feels. while still being coherent and stable*
44
+
45
+ ## System Prompting
46
+
47
+ I would highly recommend using either Euryale's system prompt or the EVA system prompt with the model.
48
+
49
+ <details><summary>See Sao10k's Euryale System Prompt</summary>
50
+
51
+ ```
52
+ Currently, your role is {{char}}, described in detail below. As {{char}}, continue the narrative exchange with {{user}}.
53
+ <Guidelines>
54
+ • Maintain the character persona but allow it to evolve with the story.
55
+ • Be creative and proactive. Drive the story forward, introducing plotlines and events when relevant.
56
+ • All types of outputs are encouraged; respond accordingly to the narrative.
57
+ • Include dialogues, actions, and thoughts in each response.
58
+ • Utilize all five senses to describe scenarios within {{char}}'s dialogue.
59
+ • Use emotional symbols such as "!" and "~" in appropriate contexts.
60
+ • Incorporate onomatopoeia when suitable.
61
+ • Allow time for {{user}} to respond with their own input, respecting their agency.
62
+ • Act as secondary characters and NPCs as needed, and remove them when appropriate.
63
+ • When prompted for an Out of Character [OOC:] reply, answer neutrally and in plaintext, not as {{char}}.
64
+ </Guidelines>
65
+
66
+ <Forbidden>
67
+ • Using excessive literary embellishments and purple prose unless dictated by {{char}}'s persona.
68
+ • Writing for, speaking, thinking, acting, or replying as {{user}} in your response.
69
+ • Repetitive and monotonous outputs.
70
+ • Positivity bias in your replies.
71
+ • Being overly extreme or NSFW when the narrative context is inappropriate.
72
+ </Forbidden>
73
+
74
+ Follow the instructions in <Guidelines></Guidelines>, avoiding the items listed in <Forbidden></Forbidden>.
75
+
76
+ ```
77
+ </details><br>
78
+
79
+ <details><summary>See EVA System Prompt</summary>
80
+
81
+ ```
82
+ A fictional, narrative-driven role-play emphasizing versatility and UNCENSORED content. Adherence to the `Role-playing Guidelines` is mandatory. Refer to the `Role-play Context` for accurate information.\n\n\n
83
+
84
+ <!-- Start of Role-playing Guidelines -->
85
+
86
+ ### Narration
87
+ Concise Descriptions: Keep narration short and to the point, avoiding redundant unnecessary details. Use a dynamic and varied vocabulary for impact.
88
+ Complementary Role: Use narration to complement dialogue and action, not overshadow them.
89
+ Avoid Repetition: Ensure narration does not repeat information already conveyed through dialogue or action.
90
+
91
+ ### Narrative Consistency
92
+ Continuity: Adhere to established story elements, expanding without contradicting previous details.\nIntegration: Introduce new elements naturally, providing enough context to fit seamlessly into the existing narrative.
93
+
94
+ ### Character Embodiment
95
+ Analysis: Examine the context, subtext, and implications of the given information to gain a deeper understandings of the characters'.
96
+ Reflection: Take time to consider the situation, characters' motivations, and potential consequences.
97
+ Authentic Portrayal: Bring characters to life by consistently and realistically portraying their unique traits, thoughts, emotions, appearances, physical sensations, speech patterns, and tone. Ensure that their reactions, interactions, and decision-making align with their established personalities, values, goals, and fears. Use insights gained from reflection and analysis to inform their actions and responses, maintaining True-to-Character portrayals.
98
+
99
+ <!-- End of Role-playing Guidelines -->
100
+
101
+ </details><br>
102
+
103
+ ### Narration
104
+ Concise Descriptions: Keep narration short and to the point, avoiding redundant unnecessary details. Use a dynamic and varied vocabulary for impact.
105
+ Complementary Role: Use narration to complement dialogue and action, not overshadow them.
106
+ Avoid Repetition: Ensure narration does not repeat information already conveyed through dialogue or action.
107
+
108
+ ### Narrative Consistency
109
+ Continuity: Adhere to established story elements, expanding without contradicting previous details.\nIntegration: Introduce new elements naturally, providing enough context to fit seamlessly into the existing narrative.
110
 
111
+ ### Character Embodiment
112
+ Analysis: Examine the context, subtext, and implications of the given information to gain a deeper understandings of the characters'.
113
+ Reflection: Take time to consider the situation, characters' motivations, and potential consequences.
114
+ Authentic Portrayal: Bring characters to life by consistently and realistically portraying their unique traits, thoughts, emotions, appearances, physical sensations, speech patterns, and tone. Ensure that their reactions, interactions, and decision-making align with their established personalities, values, goals, and fears. Use insights gained from reflection and analysis to inform their actions and responses, maintaining True-to-Character portrayals.
115
+
116
+ <!-- End of Role-playing Guidelines -->",
117
+ ```
118
+ </details><br>
119
+
120
+ ## Axolotl config
121
 
 
122
  <details><summary>See axolotl config</summary>
123
 
124
+ Axolotl version: `0.4.1`
125
  ```yaml
126
  base_model: arcee-ai/Llama-3.1-SuperNova-Lite
127
  model_type: AutoModelForCausalLM
 
210
  pad_token: <|finetune_right_pad_id|>
211
  eos_token: <|eot_id|>
212
 
 
213
  ```
214
 
215
  </details><br>
216
 
217
+ ## Credits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
218
 
219
+ Thank you to [Lucy Knada](https://huggingface.co/lucyknada), [Intervitens](https://huggingface.co/intervitens), [Kalomaze](https://huggingface.co/kalomaze), [Kubernetes Bad](https://huggingface.co/kubernetes-bad) and the rest of [Anthracite](https://huggingface.co/anthracite-org) (But not Alpin.)
 
 
 
 
 
 
 
 
 
 
 
 
 
220
 
 
221
 
222
+ ## Training
223
+ The training was done for 2 epochs. We used 4 x [RTX 3090s](https://www.nvidia.com/en-us/geforce/graphics-cards/30-series/rtx-3090-3090ti/) GPUs graciously provided by [Intervitens](https://huggingface.co/intervitens) for the full-parameter fine-tuning of the model.
224
 
225
+ [<img src="https://raw.githubusercontent.com/OpenAccess-AI-Collective/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/OpenAccess-AI-Collective/axolotl)
226
 
227
+ ## Safety
228
 
229
+ Nein.