Triangle104 commited on
Commit
d3942f1
·
verified ·
1 Parent(s): db1a60e

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +143 -0
README.md CHANGED
@@ -18,6 +18,149 @@ tags:
18
  This model was converted to GGUF format from [`allura-org/Bigger-Body-12b`](https://huggingface.co/allura-org/Bigger-Body-12b) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
19
  Refer to the [original model card](https://huggingface.co/allura-org/Bigger-Body-12b) for more details on the model.
20
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
21
  ## Use with llama.cpp
22
  Install llama.cpp through brew (works on Mac and Linux)
23
 
 
18
  This model was converted to GGUF format from [`allura-org/Bigger-Body-12b`](https://huggingface.co/allura-org/Bigger-Body-12b) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
19
  Refer to the [original model card](https://huggingface.co/allura-org/Bigger-Body-12b) for more details on the model.
20
 
21
+ ---
22
+ A roleplay-focused pseudo full-finetune of Mistral Nemo Instruct.
23
+ The successor to the Ink series.
24
+
25
+
26
+
27
+
28
+
29
+
30
+
31
+ Testimonials
32
+
33
+
34
+
35
+
36
+
37
+
38
+ First impressions (temp 1, min-p .05-.1)
39
+
40
+
41
+ It passes my silly logic tests (read: me trolling random characters)
42
+ Haven't seen any slop yet
43
+ Writes short and snappy replies
44
+ ...yet not too short, like Mahou, and can write longer responses if the context warrants it
45
+ Follows card formatting instructions
46
+
47
+
48
+ If this holds up to 16K it will be constantly in the hopper alongside
49
+ Mag-Mell for me. I'm biased towards shorter responses with smarts. :)
50
+
51
+
52
+
53
+
54
+ - Tofumagate
55
+
56
+
57
+
58
+
59
+ tantalizing writing, leagues better then whatever is available online
60
+
61
+
62
+
63
+
64
+ - Bowza
65
+
66
+
67
+
68
+
69
+ Fun to use, nice swipe variation, gives me lots to RP off of. Rarely, it'll start to loop, but a quick swipe fixes no problem.
70
+
71
+
72
+
73
+
74
+ - AliCat
75
+
76
+
77
+
78
+
79
+
80
+
81
+
82
+ Dataset
83
+
84
+
85
+
86
+
87
+ The Bigger Body (referred to as Ink v2.1, because that's still the
88
+ internal name) mix is absolutely disgusting. It's even more cursed than
89
+ the original Ink mix.
90
+
91
+
92
+ (Public) Original Datasets
93
+
94
+
95
+
96
+ Fizzarolli/limarp-processed
97
+ Norquinal/OpenCAI - two_users split
98
+ allura-org/Celeste1.x-data-mixture
99
+ mapsila/PIPPA-ShareGPT-formatted-named
100
+ allenai/tulu-3-sft-personas-instruction-following
101
+ readmehay/medical-01-reasoning-SFT-json
102
+ LooksJuicy/ruozhiba
103
+ shibing624/roleplay-zh-sharegpt-gpt4-data
104
+ CausalLM/Retrieval-SFT-Chat
105
+ ToastyPigeon/fujin-filtered-instruct
106
+
107
+ Recommended Settings
108
+
109
+
110
+
111
+
112
+ Chat template: Mistral v7-tekken (NOT v3-tekken !!!! the main difference is that v7 has specific [SYSTEM_PROMPT] and [/SYSTEM_PROMPT] tags)
113
+ Recommended samplers (not the be-all-end-all, try some on your own!):
114
+
115
+
116
+ Temp 1.25 / MinP 0.1
117
+
118
+
119
+
120
+
121
+
122
+
123
+
124
+ Hyperparams
125
+
126
+
127
+
128
+
129
+
130
+
131
+
132
+
133
+
134
+ General
135
+
136
+
137
+
138
+
139
+ Epochs = 2
140
+ LR = 1e-5
141
+ LR Scheduler = Cosine
142
+ Optimizer = Apollo-mini
143
+ Optimizer target modules = all_linear
144
+ Effective batch size = 16
145
+ Weight Decay = 0.01
146
+ Warmup steps = 50
147
+ Total steps = 920
148
+
149
+
150
+
151
+
152
+
153
+
154
+
155
+ Credits
156
+
157
+
158
+
159
+
160
+ Humongous thanks to the people who created the data. I would credit you all, but that would be cheating ;)
161
+ Big thanks to all Allura members for testing and emotional support ilya /platonic
162
+
163
+ ---
164
  ## Use with llama.cpp
165
  Install llama.cpp through brew (works on Mac and Linux)
166