Text Generation
GGUF
English
creative
creative writing
fiction writing
plot generation
sub-plot generation
story generation
scene continue
storytelling
fiction story
science fiction
romance
all genres
story
writing
vivid prosing
vivid writing
fiction
roleplaying
bfloat16
swearing
role play
sillytavern
backyard
horror
llama 3.1
context 128k
mergekit
Inference Endpoints
conversational
Create README.md
Browse files
README.md
ADDED
@@ -0,0 +1,272 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
license: apache-2.0
|
3 |
+
language:
|
4 |
+
- en
|
5 |
+
tags:
|
6 |
+
- creative
|
7 |
+
- creative writing
|
8 |
+
- fiction writing
|
9 |
+
- plot generation
|
10 |
+
- sub-plot generation
|
11 |
+
- fiction writing
|
12 |
+
- story generation
|
13 |
+
- scene continue
|
14 |
+
- storytelling
|
15 |
+
- fiction story
|
16 |
+
- science fiction
|
17 |
+
- romance
|
18 |
+
- all genres
|
19 |
+
- story
|
20 |
+
- writing
|
21 |
+
- vivid prosing
|
22 |
+
- vivid writing
|
23 |
+
- fiction
|
24 |
+
- roleplaying
|
25 |
+
- bfloat16
|
26 |
+
- swearing
|
27 |
+
- role play
|
28 |
+
- sillytavern
|
29 |
+
- backyard
|
30 |
+
- horror
|
31 |
+
- llama 3.1
|
32 |
+
- context 128k
|
33 |
+
- mergekit
|
34 |
+
pipeline_tag: text-generation
|
35 |
+
---
|
36 |
+
|
37 |
+
(quants uploading, examples to follow)
|
38 |
+
|
39 |
+
<B><font color="red">WARNING:</font> NSFW. Vivid prose. INTENSE. Visceral Details. Violence. Graphic HORROR. GORE. Swearing. UNCENSORED. </B>
|
40 |
+
|
41 |
+
<h2>L3.1-RP-Hero-Dirty_Harry-8B-GGUF</h2>
|
42 |
+
|
43 |
+
<img src="rp-talker.jpg" style="float:right; width:300px; height:300px; padding:10px;">
|
44 |
+
|
45 |
+
It is a LLama3.1 model, max context of 128k (131,000) and is a dedicated "roleplay model" (it can also be used for creative uses).
|
46 |
+
|
47 |
+
This model has been designed to be relatively bullet proof and operates with all parameters, including temp settings from 0 to 5.
|
48 |
+
|
49 |
+
It is an extraordinary compressed model, with a very low perplexity level (lower than Meta Llama 3.1 Instruct).
|
50 |
+
|
51 |
+
This model is for any writing, fiction or roleplay activity, but it is composed of ROLE PLAY models and it primary designed for role play.
|
52 |
+
|
53 |
+
It also has stronger than average instruction following attibutes.
|
54 |
+
|
55 |
+
This is version "Dirty Harry", which has two additional versions: "InBetween" and "Big Talker".
|
56 |
+
|
57 |
+
InBetween (medium output generation on average):
|
58 |
+
|
59 |
+
[ https://huggingface.co/DavidAU/L3.1-RP-Hero-InBetween-8B-GGUF ]
|
60 |
+
|
61 |
+
Big Talker (long output generation on average):
|
62 |
+
|
63 |
+
[ https://huggingface.co/DavidAU/L3.1-RP-Hero-BigTalker-8B-GGUF ]
|
64 |
+
|
65 |
+
"Dirty Harry" has SHORT (average) level length output, and is uncensored (note: InBetween has a slight degree of censorship).
|
66 |
+
|
67 |
+
"Dirty Harry" also has slightly higher detail level than "InBetween", but on par with "Big Talker.
|
68 |
+
|
69 |
+
All versions are composed of top rated Role Play models.
|
70 |
+
|
71 |
+
This model, as well as the other two versions, can be used for any creative genre too.
|
72 |
+
|
73 |
+
It requires Llama3 template and/or "Command-R" template.
|
74 |
+
|
75 |
+
For roleplay settings, and apps to use this model for roleplay see the section "Highest Quality Settings..." below.
|
76 |
+
|
77 |
+
Example outputs below to show prose quality / creativity.
|
78 |
+
|
79 |
+
<B>Model Notes:</B>
|
80 |
+
|
81 |
+
- Detail, prose and fiction writing abilities are significantly improved.
|
82 |
+
- For more varied prose (sentence/paragraph/dialog) raise the temp and/or add more instructions in your prompt(s).
|
83 |
+
- Role-players: Careful raising temp too high as it may affect instruction following.
|
84 |
+
- This model works with rep pen of 1 or higher, 1.02+ recommended.
|
85 |
+
- If you want a specific type of prose (IE horror) add in "(vivid horror)" or "(graphic vivid horror)" (no quotes) in your prompt(s).
|
86 |
+
- This model has a neutral to negative bias BUT can be controlled by prompt/prose controls directly.
|
87 |
+
- Output length will vary however this model prefers "SHORT" outputs EVEN IF you state the size.
|
88 |
+
- For creative uses, different quants will produce slightly different output.
|
89 |
+
- Due to the high stability and compressed nature of this model, all quants will operate at above average levels.
|
90 |
+
- Source code for this model will be uploaded at separate repo shortly.
|
91 |
+
|
92 |
+
<B>Settings, Quants and Critical Operations Notes:</b>
|
93 |
+
|
94 |
+
Change in temp (ie, .4, .8, 1.5, 2, 3 ) will drastically alter output.
|
95 |
+
|
96 |
+
Rep pen settings will also alter output too.
|
97 |
+
|
98 |
+
This model needs "rep pen" of 1.05 or higher as lower values may cause repeat paragraph issues at end of output however LOWER rep pen
|
99 |
+
values may result is very different (creative / unusual) generation too.
|
100 |
+
|
101 |
+
For role play: Rep pen of 1.02 min is suggested.
|
102 |
+
|
103 |
+
Raise/lower rep pen SLOWLY ie: 1.011, 1.012 ...
|
104 |
+
|
105 |
+
Rep pen will alter prose, word choice (lower rep pen=small words / more small word - sometimes) and creativity.
|
106 |
+
|
107 |
+
To really push the model:
|
108 |
+
|
109 |
+
Rep pen 1.05+ or lower / Temp 3+ ... be ready to stop the output because it may go and go at these strong settings.
|
110 |
+
|
111 |
+
You can also set a "hard stop" - maximum tokens generation - too to address lower rep pen settings / high creativity settings.
|
112 |
+
|
113 |
+
Longer prompts vastly increase the quality of the model's output.
|
114 |
+
|
115 |
+
GET A GOOD "GENERATION":
|
116 |
+
|
117 |
+
This model has been set, so that each time you "regen" a prompt it will not deviate too much from the previous generation.
|
118 |
+
(Unlike Darkest Planet 16.5B, which will).
|
119 |
+
|
120 |
+
That being said, sometimes a second or third generation will been of much higher overall quality.
|
121 |
+
|
122 |
+
IE:
|
123 |
+
|
124 |
+
If you use case is creative writing, you may want to regen a prompt 1-5 times then pick the best one. The best
|
125 |
+
way to do this is open a new chat PER generation, then do a "read thru" to see which one(s) hit the mark.
|
126 |
+
|
127 |
+
Then adjust temp and/or rep pen slightly and retry this process.
|
128 |
+
|
129 |
+
The goal is the best generation with least amount of editing in this example.
|
130 |
+
|
131 |
+
QUANTS:
|
132 |
+
|
133 |
+
Higher quants will have more detail, nuance and in some cases stronger "emotional" levels. Characters will also be
|
134 |
+
more "fleshed out" too. Sense of "there" will also increase.
|
135 |
+
|
136 |
+
Q4KM/Q4KS are good, strong quants however if you can run Q5, Q6 or Q8 - go for the highest quant you can.
|
137 |
+
|
138 |
+
IQ4XS: Due to the unusual nature of this quant (mixture/processing), generations from it will be different then other quants.
|
139 |
+
|
140 |
+
You may want to try it / compare it to other quant(s) output.
|
141 |
+
|
142 |
+
Special note on Q2k/Q3 quants:
|
143 |
+
|
144 |
+
You may need to use temp 2 or lower with these quants (1 or lower for q2k). Just too much compression at this level, damaging the model. I will see if Imatrix versions
|
145 |
+
of these quants will function better.
|
146 |
+
|
147 |
+
Rep pen adjustments may also be required to get the most out of this model at this/these quant level(s).
|
148 |
+
|
149 |
+
ARM QUANTS:
|
150 |
+
|
151 |
+
This repo has 3 arm quants for computers than can run them. If you use these quants on a non-arm computer, your token per second will be very low.
|
152 |
+
|
153 |
+
<B>Settings: CHAT / ROLEPLAY and/or SMOOTHER operation of this model:</B>
|
154 |
+
|
155 |
+
In "KoboldCpp" or "oobabooga/text-generation-webui" or "Silly Tavern" ;
|
156 |
+
|
157 |
+
Set the "Smoothing_factor" to 1.5 to 2.5
|
158 |
+
|
159 |
+
: in KoboldCpp -> Settings->Samplers->Advanced-> "Smooth_F"
|
160 |
+
|
161 |
+
: in text-generation-webui -> parameters -> lower right.
|
162 |
+
|
163 |
+
: In Silly Tavern this is called: "Smoothing"
|
164 |
+
|
165 |
+
NOTE: For "text-generation-webui"
|
166 |
+
|
167 |
+
-> if using GGUFs you need to use "llama_HF" (which involves downloading some config files from the SOURCE version of this model)
|
168 |
+
|
169 |
+
Source versions (and config files) of my models are here:
|
170 |
+
|
171 |
+
https://huggingface.co/collections/DavidAU/d-au-source-files-for-gguf-exl2-awq-gptq-hqq-etc-etc-66b55cb8ba25f914cbf210be
|
172 |
+
|
173 |
+
OTHER OPTIONS:
|
174 |
+
|
175 |
+
- Increase rep pen to 1.1 to 1.15 (you don't need to do this if you use "smoothing_factor")
|
176 |
+
|
177 |
+
- If the interface/program you are using to run AI MODELS supports "Quadratic Sampling" ("smoothing") just make the adjustment as noted.
|
178 |
+
|
179 |
+
<B>Highest Quality Settings / Optimal Operation Guide / Parameters and Samplers</B>
|
180 |
+
|
181 |
+
This a "Class 1" model:
|
182 |
+
|
183 |
+
For all settings used for this model (including specifics for its "class"), including example generation(s) and for advanced settings guide (which many times addresses any model issue(s)), including methods to improve model performance for all use case(s) as well as chat, roleplay and other use case(s) please see:
|
184 |
+
|
185 |
+
[ https://huggingface.co/DavidAU/Maximizing-Model-Performance-All-Quants-Types-And-Full-Precision-by-Samplers_Parameters ]
|
186 |
+
|
187 |
+
You can see all parameters used for generation, in addition to advanced parameters and samplers to get the most out of this model here:
|
188 |
+
|
189 |
+
[ https://huggingface.co/DavidAU/Maximizing-Model-Performance-All-Quants-Types-And-Full-Precision-by-Samplers_Parameters ]
|
190 |
+
|
191 |
+
|
192 |
+
<B>Templates:</B>
|
193 |
+
|
194 |
+
This is a LLAMA3 model, and requires Llama3 template, but may work with other template(s) and has maximum context of 128k / 131,000.
|
195 |
+
|
196 |
+
If you use "Command-R" template your output will be very different from using "Llama3" template.
|
197 |
+
|
198 |
+
Here is the standard LLAMA3 template:
|
199 |
+
|
200 |
+
<PRE>
|
201 |
+
{
|
202 |
+
"name": "Llama 3",
|
203 |
+
"inference_params": {
|
204 |
+
"input_prefix": "<|start_header_id|>user<|end_header_id|>\n\n",
|
205 |
+
"input_suffix": "<|eot_id|><|start_header_id|>assistant<|end_header_id|>\n\n",
|
206 |
+
"pre_prompt": "You are a helpful, smart, kind, and efficient AI assistant. You always fulfill the user's requests to the best of your ability.",
|
207 |
+
"pre_prompt_prefix": "<|start_header_id|>system<|end_header_id|>\n\n",
|
208 |
+
"pre_prompt_suffix": "<|eot_id|>",
|
209 |
+
"antiprompt": [
|
210 |
+
"<|start_header_id|>",
|
211 |
+
"<|eot_id|>"
|
212 |
+
]
|
213 |
+
}
|
214 |
+
}
|
215 |
+
</PRE>
|
216 |
+
|
217 |
+
<B>Model "DNA":</B>
|
218 |
+
|
219 |
+
Special thanks to the incredible work of the model makers "ArliAI", "Casual-Autopsy" , "Gryphe", "aifeifei798" :
|
220 |
+
|
221 |
+
Models used:
|
222 |
+
|
223 |
+
https://huggingface.co/ArliAI/Llama-3.1-8B-ArliAI-RPMax-v1.1
|
224 |
+
|
225 |
+
https://huggingface.co/Casual-Autopsy/L3-Umbral-Mind-RP-v0.3-8B
|
226 |
+
|
227 |
+
https://huggingface.co/Gryphe/Pantheon-RP-1.0-8b-Llama-3
|
228 |
+
|
229 |
+
https://huggingface.co/aifeifei798/DarkIdol-Llama-3.1-8B-Instruct-1.2-Uncensored
|
230 |
+
|
231 |
+
Parts of these models were "grafted" / "fused" together to create this model.
|
232 |
+
|
233 |
+
<b>Optional Enhancement:</B>
|
234 |
+
|
235 |
+
The following can be used in place of the "system prompt" or "system role" to further enhance the model.
|
236 |
+
|
237 |
+
It can also be used at the START of a NEW chat, but you must make sure it is "kept" as the chat moves along.
|
238 |
+
In this case the enhancements do not have as strong effect at using "system prompt" or "system role".
|
239 |
+
|
240 |
+
Copy and paste EXACTLY as noted, DO NOT line wrap or break the lines, maintain the carriage returns exactly as presented.
|
241 |
+
|
242 |
+
<PRE>
|
243 |
+
Below is an instruction that describes a task. Ponder each user instruction carefully, and use your skillsets and critical instructions to complete the task to the best of your abilities.
|
244 |
+
|
245 |
+
Here are your skillsets:
|
246 |
+
[MASTERSTORY]:NarrStrct(StryPlnng,Strbd,ScnSttng,Exps,Dlg,Pc)-CharDvlp(ChrctrCrt,ChrctrArcs,Mtvtn,Bckstry,Rltnshps,Dlg*)-PltDvlp(StryArcs,PltTwsts,Sspns,Fshdwng,Climx,Rsltn)-ConfResl(Antg,Obstcls,Rsltns,Cnsqncs,Thms,Symblsm)-EmotImpct(Empt,Tn,Md,Atmsphr,Imgry,Symblsm)-Delvry(Prfrmnc,VcActng,PblcSpkng,StgPrsnc,AudncEngmnt,Imprv)
|
247 |
+
|
248 |
+
[*DialogWrt]:(1a-CharDvlp-1a.1-Backgrnd-1a.2-Personality-1a.3-GoalMotiv)>2(2a-StoryStruc-2a.1-PlotPnt-2a.2-Conflict-2a.3-Resolution)>3(3a-DialogTech-3a.1-ShowDontTell-3a.2-Subtext-3a.3-VoiceTone-3a.4-Pacing-3a.5-VisualDescrip)>4(4a-DialogEdit-4a.1-ReadAloud-4a.2-Feedback-4a.3-Revision)
|
249 |
+
|
250 |
+
Here are your critical instructions:
|
251 |
+
Ponder each word choice carefully to present as vivid and emotional journey as is possible. Choose verbs and nouns that are both emotional and full of imagery. Load the story with the 5 senses. Aim for 50% dialog, 25% narration, 15% body language and 10% thoughts. Your goal is to put the reader in the story.
|
252 |
+
</PRE>
|
253 |
+
|
254 |
+
You do not need to use this, it is only presented as an additional enhancement which seems to help scene generation
|
255 |
+
and scene continue functions.
|
256 |
+
|
257 |
+
This enhancement WAS NOT used to generate the examples below.
|
258 |
+
|
259 |
+
<h3>EXAMPLES PROMPTS and OUTPUT:</h3>
|
260 |
+
|
261 |
+
Examples are created using quant Q4_K_M, "temp=1.3", "rep pen : 1.02" (unless otherwise stated), minimal parameters and "LLAMA3" template.
|
262 |
+
|
263 |
+
Model has been tested with "temp" from ".1" to "5".
|
264 |
+
|
265 |
+
Below are the least creative outputs, prompt is in <B>BOLD</B>.
|
266 |
+
|
267 |
+
---
|
268 |
+
|
269 |
+
<B><font color="red">WARNING:</font> NSFW. Vivid prose. Visceral Details. Violence. HORROR. Swearing. UNCENSORED. </B>
|
270 |
+
|
271 |
+
---
|
272 |
+
|