Safetensors
llama
chat
Delta-Vector commited on
Commit
c8743ee
1 Parent(s): 2803b4b

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +119 -51
README.md CHANGED
@@ -1,21 +1,122 @@
1
  ---
2
- library_name: transformers
3
- license: llama3
4
- base_model: arcee-ai/Llama-3.1-SuperNova-Lite
5
  tags:
6
- - generated_from_trainer
7
- model-index:
8
- - name: outputs
9
- results: []
 
 
 
 
 
 
 
10
  ---
11
 
12
- <!-- This model card has been generated automatically according to the information the Trainer had access to. You
13
- should probably proofread and complete it, then remove this comment. -->
14
 
15
- [<img src="https://raw.githubusercontent.com/axolotl-ai-cloud/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/axolotl-ai-cloud/axolotl)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
16
  <details><summary>See axolotl config</summary>
17
 
18
- axolotl version: `0.4.1`
19
  ```yaml
20
  base_model: arcee-ai/Llama-3.1-SuperNova-Lite
21
  model_type: AutoModelForCausalLM
@@ -104,53 +205,20 @@ special_tokens:
104
  pad_token: <|finetune_right_pad_id|>
105
  eos_token: <|eot_id|>
106
 
107
-
108
  ```
109
 
110
  </details><br>
111
 
112
- # outputs
113
-
114
- This model is a fine-tuned version of [arcee-ai/Llama-3.1-SuperNova-Lite](https://huggingface.co/arcee-ai/Llama-3.1-SuperNova-Lite) on the None dataset.
115
-
116
- ## Model description
117
-
118
- More information needed
119
-
120
- ## Intended uses & limitations
121
-
122
- More information needed
123
-
124
- ## Training and evaluation data
125
-
126
- More information needed
127
-
128
- ## Training procedure
129
-
130
- ### Training hyperparameters
131
 
132
- The following hyperparameters were used during training:
133
- - learning_rate: 1e-05
134
- - train_batch_size: 1
135
- - eval_batch_size: 1
136
- - seed: 42
137
- - distributed_type: multi-GPU
138
- - num_devices: 4
139
- - gradient_accumulation_steps: 2
140
- - total_train_batch_size: 8
141
- - total_eval_batch_size: 4
142
- - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
143
- - lr_scheduler_type: cosine
144
- - lr_scheduler_warmup_steps: 15
145
- - num_epochs: 4
146
 
147
- ### Training results
148
 
 
 
149
 
 
150
 
151
- ### Framework versions
152
 
153
- - Transformers 4.44.2
154
- - Pytorch 2.3.1+cu121
155
- - Datasets 2.20.0
156
- - Tokenizers 0.19.1
 
1
  ---
2
+ license: agpl-3.0
 
 
3
  tags:
4
+ - chat
5
+ datasets:
6
+ - NewEden/OpenCAI-ShareGPT
7
+ - NewEden/Roleplay-Logs-Sharegpt-Ngram-cleaned
8
+ License: agpl-3.0
9
+ Language:
10
+ - En
11
+ Pipeline_tag: text-generation
12
+ Base_model: arcee-ai/Llama-3.1-SuperNova-Lite
13
+ Tags:
14
+ - Chat
15
  ---
16
 
 
 
17
 
18
+ An experimental finetune based on the Llama3.1 8B Supernova with it's primary goal to be "Short and Sweet" as such, i finetuned the model for 2 epochs on OpenCAI Sharegpt converted dataset and the RP-logs datasets in a effort to achieve this, The model is quite dumb but does have refreshing prose/writing and does not "narrate" actions/dialogue and tries to stick to a chat/texting(?) format.
19
+
20
+ # Quants
21
+
22
+ GGUF: https://huggingface.co/Delta-Vector/Control-8B-gguf
23
+
24
+ EXL2 (Thanks Lucy <3) : https://huggingface.co/Delta-Vector/Control-8B-EXL2
25
+
26
+
27
+ ## Prompting
28
+ Model has been tuned with the LLama-Instruct formatting. A typical input would look like this:
29
+
30
+ ```py
31
+ """<|begin_of_text|><|start_header_id|>system<|end_header_id|>
32
+ You are an AI built to rid the world of bonds and journeys!<|eot_id|><|start_header_id|>user<|end_header_id|>
33
+ Bro i just wanna know what is 2+2?<|eot_id|><|start_header_id|>assistant<|end_header_id|>
34
+ """
35
+
36
+ ```
37
+
38
+ *Also note that ChatML may work as well, and might change how the model feels. while still being coherent and stable*
39
+
40
+ ## System Prompting
41
+
42
+ I would highly recommend using either Euryale's system prompt or the EVA system prompt with the model.
43
+
44
+ <details><summary>See Sao10k's Euryale System Prompt</summary>
45
+
46
+ ```
47
+ Currently, your role is {{char}}, described in detail below. As {{char}}, continue the narrative exchange with {{user}}.
48
+ <Guidelines>
49
+ • Maintain the character persona but allow it to evolve with the story.
50
+ • Be creative and proactive. Drive the story forward, introducing plotlines and events when relevant.
51
+ • All types of outputs are encouraged; respond accordingly to the narrative.
52
+ • Include dialogues, actions, and thoughts in each response.
53
+ • Utilize all five senses to describe scenarios within {{char}}'s dialogue.
54
+ • Use emotional symbols such as "!" and "~" in appropriate contexts.
55
+ • Incorporate onomatopoeia when suitable.
56
+ • Allow time for {{user}} to respond with their own input, respecting their agency.
57
+ • Act as secondary characters and NPCs as needed, and remove them when appropriate.
58
+ • When prompted for an Out of Character [OOC:] reply, answer neutrally and in plaintext, not as {{char}}.
59
+ </Guidelines>
60
+
61
+ <Forbidden>
62
+ • Using excessive literary embellishments and purple prose unless dictated by {{char}}'s persona.
63
+ • Writing for, speaking, thinking, acting, or replying as {{user}} in your response.
64
+ • Repetitive and monotonous outputs.
65
+ • Positivity bias in your replies.
66
+ • Being overly extreme or NSFW when the narrative context is inappropriate.
67
+ </Forbidden>
68
+
69
+ Follow the instructions in <Guidelines></Guidelines>, avoiding the items listed in <Forbidden></Forbidden>.
70
+
71
+ ```
72
+ </details><br>
73
+
74
+ <details><summary>See EVA System Prompt</summary>
75
+
76
+ ```
77
+ A fictional, narrative-driven role-play emphasizing versatility and UNCENSORED content. Adherence to the `Role-playing Guidelines` is mandatory. Refer to the `Role-play Context` for accurate information.\n\n\n
78
+
79
+ <!-- Start of Role-playing Guidelines -->
80
+
81
+ ### Narration
82
+ Concise Descriptions: Keep narration short and to the point, avoiding redundant unnecessary details. Use a dynamic and varied vocabulary for impact.
83
+ Complementary Role: Use narration to complement dialogue and action, not overshadow them.
84
+ Avoid Repetition: Ensure narration does not repeat information already conveyed through dialogue or action.
85
+
86
+ ### Narrative Consistency
87
+ Continuity: Adhere to established story elements, expanding without contradicting previous details.\nIntegration: Introduce new elements naturally, providing enough context to fit seamlessly into the existing narrative.
88
+
89
+ ### Character Embodiment
90
+ Analysis: Examine the context, subtext, and implications of the given information to gain a deeper understandings of the characters'.
91
+ Reflection: Take time to consider the situation, characters' motivations, and potential consequences.
92
+ Authentic Portrayal: Bring characters to life by consistently and realistically portraying their unique traits, thoughts, emotions, appearances, physical sensations, speech patterns, and tone. Ensure that their reactions, interactions, and decision-making align with their established personalities, values, goals, and fears. Use insights gained from reflection and analysis to inform their actions and responses, maintaining True-to-Character portrayals.
93
+
94
+ <!-- End of Role-playing Guidelines -->
95
+
96
+ </details><br>
97
+
98
+ ### Narration
99
+ Concise Descriptions: Keep narration short and to the point, avoiding redundant unnecessary details. Use a dynamic and varied vocabulary for impact.
100
+ Complementary Role: Use narration to complement dialogue and action, not overshadow them.
101
+ Avoid Repetition: Ensure narration does not repeat information already conveyed through dialogue or action.
102
+
103
+ ### Narrative Consistency
104
+ Continuity: Adhere to established story elements, expanding without contradicting previous details.\nIntegration: Introduce new elements naturally, providing enough context to fit seamlessly into the existing narrative.
105
+
106
+ ### Character Embodiment
107
+ Analysis: Examine the context, subtext, and implications of the given information to gain a deeper understandings of the characters'.
108
+ Reflection: Take time to consider the situation, characters' motivations, and potential consequences.
109
+ Authentic Portrayal: Bring characters to life by consistently and realistically portraying their unique traits, thoughts, emotions, appearances, physical sensations, speech patterns, and tone. Ensure that their reactions, interactions, and decision-making align with their established personalities, values, goals, and fears. Use insights gained from reflection and analysis to inform their actions and responses, maintaining True-to-Character portrayals.
110
+
111
+ <!-- End of Role-playing Guidelines -->",
112
+ ```
113
+ </details><br>
114
+
115
+ ## Axolotl config
116
+
117
  <details><summary>See axolotl config</summary>
118
 
119
+ Axolotl version: `0.4.1`
120
  ```yaml
121
  base_model: arcee-ai/Llama-3.1-SuperNova-Lite
122
  model_type: AutoModelForCausalLM
 
205
  pad_token: <|finetune_right_pad_id|>
206
  eos_token: <|eot_id|>
207
 
 
208
  ```
209
 
210
  </details><br>
211
 
212
+ ## Credits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
213
 
214
+ Thank you to [Lucy Knada](https://huggingface.co/lucyknada), [Intervitens](https://huggingface.co/intervitens), [Kalomaze](https://huggingface.co/kalomaze), [Kubernetes Bad](https://huggingface.co/kubernetes-bad) and the rest of [Anthracite](https://huggingface.co/anthracite-org) (But not Alpin.)
 
 
 
 
 
 
 
 
 
 
 
 
 
215
 
 
216
 
217
+ ## Training
218
+ The training was done for 2 epochs. We used 4 x [RTX 3090s](https://www.nvidia.com/en-us/geforce/graphics-cards/30-series/rtx-3090-3090ti/) GPUs graciously provided by [Intervitens](https://huggingface.co/intervitens) for the full-parameter fine-tuning of the model.
219
 
220
+ [<img src="https://raw.githubusercontent.com/OpenAccess-AI-Collective/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/OpenAccess-AI-Collective/axolotl)
221
 
222
+ ## Safety
223
 
224
+ Nein.