Upload folder using huggingface_hub
Browse files- .gitattributes +9 -0
- .ipynb_checkpoints/README-checkpoint.md +57 -0
- README.md +57 -0
- assets/MinP.json +1 -0
- assets/alice.png +3 -0
- assets/context.json +1 -0
- assets/instruct.json +1 -0
- assets/masha.png +3 -0
- silicon-masha-7b.Q2_K.gguf +3 -0
- silicon-masha-7b.Q3_K_M.gguf +3 -0
- silicon-masha-7b.Q4_0.gguf +3 -0
- silicon-masha-7b.Q4_K_M.gguf +3 -0
- silicon-masha-7b.Q5_0.gguf +3 -0
- silicon-masha-7b.Q5_K_M.gguf +3 -0
- silicon-masha-7b.Q6_K.gguf +3 -0
.gitattributes
CHANGED
@@ -33,3 +33,12 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
36 |
+
assets/alice.png filter=lfs diff=lfs merge=lfs -text
|
37 |
+
assets/masha.png filter=lfs diff=lfs merge=lfs -text
|
38 |
+
silicon-masha-7b.Q2_K.gguf filter=lfs diff=lfs merge=lfs -text
|
39 |
+
silicon-masha-7b.Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text
|
40 |
+
silicon-masha-7b.Q4_0.gguf filter=lfs diff=lfs merge=lfs -text
|
41 |
+
silicon-masha-7b.Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text
|
42 |
+
silicon-masha-7b.Q5_0.gguf filter=lfs diff=lfs merge=lfs -text
|
43 |
+
silicon-masha-7b.Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text
|
44 |
+
silicon-masha-7b.Q6_K.gguf filter=lfs diff=lfs merge=lfs -text
|
.ipynb_checkpoints/README-checkpoint.md
ADDED
@@ -0,0 +1,57 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
language:
|
3 |
+
- en
|
4 |
+
- ru
|
5 |
+
tags:
|
6 |
+
- not-for-all-audiences
|
7 |
+
- nsfw
|
8 |
+
base_model:
|
9 |
+
- LakoMoor/Silicon-Masha-7B
|
10 |
+
license: cc-by-nc-4.0
|
11 |
+
inference: false
|
12 |
+
library_name: transformers
|
13 |
+
model_creator: LakoMoor
|
14 |
+
model_name: Silicon-Masha-7B
|
15 |
+
model_type: mistral
|
16 |
+
---
|
17 |
+
|
18 |
+
# Silicon-Masha-7B-GGUF
|
19 |
+
![Silicon-Masha-7B](https://huggingface.co/LakoMoor/Silicon-Masha-7B-GGUF/resolve/main/assets/masha.png)
|
20 |
+
|
21 |
+
## What's that?
|
22 |
+
|
23 |
+
Silicon-Masha-7B-GGUF is a quantized model aiming to be both strong in RP, be smart **and** understand Russian, that can follow character maps very well. It is suitable for RP/ERP and general use.It can be run on weak iron using [llama.cpp](https://github.com/ggerganov/llama.cpp) or [koboldcpp](https://github.com/LostRuins/koboldcpp).
|
24 |
+
|
25 |
+
## Prompt Template (Alpaca)
|
26 |
+
|
27 |
+
I found the best SillyTavern results from using the Noromaid template but please try other templates! Let me know if you find anything good.
|
28 |
+
|
29 |
+
SillyTavern config files: [Context](https://huggingface.co/LakoMoor/Silicon-Masha-7B-GGUF/resolve/main/assets/context.json), [Instruct](https://huggingface.co/LakoMoor/Silicon-Masha-7B-GGUF/resolve/main/assets/instruct.json).
|
30 |
+
|
31 |
+
Additionally, here is my highly recommended [Text Completion preset](https://huggingface.co/LakoMoor/Silicon-Masha-7B-GGUF/resolve/main/assets/MinP.json). You can tweak this by adjusting temperature up or dropping min p to boost creativity or raise min p to increase stability. You shouldn't need to touch anything else!
|
32 |
+
|
33 |
+
```
|
34 |
+
Below is an instruction that describes a task. Write a response that appropriately completes the request.
|
35 |
+
### Instruction:
|
36 |
+
{prompt}
|
37 |
+
### Response:
|
38 |
+
```
|
39 |
+
|
40 |
+
## Provided files
|
41 |
+
|
42 |
+
| Name | Quant method | Bits | Use case |
|
43 |
+
| ---- | ---- | ---- | ----- |
|
44 |
+
| [silicon-masha-7b.Q2_K.gguf](https://huggingface.co/LakoMoor/Silicon-Masha-7B-GGUF/blob/main/silicon-masha-7b.Q2_K.gguf) | Q2_K | 2 | smallest, significant quality loss - not recommended for most purposes |
|
45 |
+
| [silicon-masha-7b.Q3_K_M.gguf](https://huggingface.co/LakoMoor/Silicon-Masha-7B-GGUF/blob/main/silicon-masha-7b.Q3_K_M.gguf) | Q3_K_M | 3 | very small, high quality loss |
|
46 |
+
| [silicon-masha-7b.Q4_0.gguf](https://huggingface.co/LakoMoor/Silicon-Masha-7B-GGUF/blob/main/silicon-masha-7b.Q4_0.gguf) | Q4_0 | 4 | legacy; small, very high quality loss - prefer using Q3_K_M |
|
47 |
+
| [silicon-masha-7b.Q4_K_M.gguf](https://huggingface.co/LakoMoor/Silicon-Masha-7B-GGUF/blob/main/silicon-masha-7b.Q4_K_M.gguf) | Q4_K_M | 4 | medium, balanced quality - recommended |
|
48 |
+
| [silicon-masha-7b.Q5_0.gguf](https://huggingface.co/LakoMoor/Silicon-Masha-7B-GGUF/blob/main/silicon-masha-7b.Q5_0.gguf) | Q5_0 | 5 | legacy; medium, balanced quality - prefer using Q4_K_M |
|
49 |
+
| [silicon-masha-7b.Q5_K_M.gguf](https://huggingface.co/LakoMoor/Silicon-Masha-7B-GGUF/blob/main/silicon-masha-7b.Q5_K_M.gguf) | Q5_K_M | 5 | large, very low quality loss - recommended |
|
50 |
+
| [silicon-masha-7b.Q6_K.gguf](https://huggingface.co/LakoMoor/Silicon-Masha-7B-GGUF/blob/main/silicon-masha-7b.Q6_K.gguf) | Q6_K | 6 | very large, extremely low quality loss |
|
51 |
+
| [silicon-masha-7b.Q8_0.gguf](https://huggingface.co/LakoMoor/Silicon-Masha-7B-GGUF/blob/main/silicon-masha-7b.Q8_0.gguf) | Q8_0 | 8 | very large, extremely low quality loss - not recommended |
|
52 |
+
|
53 |
+
## How run it ?
|
54 |
+
[llama.cpp](https://github.com/ggerganov/llama.cpp)
|
55 |
+
```shell
|
56 |
+
./main -ngl 35 -m silicon-masha-7b.Q4_K_M.gguf --color -c 32768 --temp 0.7 --repeat_penalty 1.1 -n -1 -p "Below is an instruction that describes a task. Write a response that appropriately completes the request.\n{system_message}\n### Instruction:{prompt}\n### Response:\n"
|
57 |
+
```
|
README.md
ADDED
@@ -0,0 +1,57 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
language:
|
3 |
+
- en
|
4 |
+
- ru
|
5 |
+
tags:
|
6 |
+
- not-for-all-audiences
|
7 |
+
- nsfw
|
8 |
+
base_model:
|
9 |
+
- LakoMoor/Silicon-Masha-7B
|
10 |
+
license: cc-by-nc-4.0
|
11 |
+
inference: false
|
12 |
+
library_name: transformers
|
13 |
+
model_creator: LakoMoor
|
14 |
+
model_name: Silicon-Masha-7B
|
15 |
+
model_type: mistral
|
16 |
+
---
|
17 |
+
|
18 |
+
# Silicon-Masha-7B-GGUF
|
19 |
+
![Silicon-Masha-7B](https://huggingface.co/LakoMoor/Silicon-Masha-7B-GGUF/resolve/main/assets/masha.png)
|
20 |
+
|
21 |
+
## What's that?
|
22 |
+
|
23 |
+
Silicon-Masha-7B-GGUF is a quantized model aiming to be both strong in RP, be smart **and** understand Russian, that can follow character maps very well. It is suitable for RP/ERP and general use.It can be run on weak iron using [llama.cpp](https://github.com/ggerganov/llama.cpp) or [koboldcpp](https://github.com/LostRuins/koboldcpp).
|
24 |
+
|
25 |
+
## Prompt Template (Alpaca)
|
26 |
+
|
27 |
+
I found the best SillyTavern results from using the Noromaid template but please try other templates! Let me know if you find anything good.
|
28 |
+
|
29 |
+
SillyTavern config files: [Context](https://huggingface.co/LakoMoor/Silicon-Masha-7B-GGUF/resolve/main/assets/context.json), [Instruct](https://huggingface.co/LakoMoor/Silicon-Masha-7B-GGUF/resolve/main/assets/instruct.json).
|
30 |
+
|
31 |
+
Additionally, here is my highly recommended [Text Completion preset](https://huggingface.co/LakoMoor/Silicon-Masha-7B-GGUF/resolve/main/assets/MinP.json). You can tweak this by adjusting temperature up or dropping min p to boost creativity or raise min p to increase stability. You shouldn't need to touch anything else!
|
32 |
+
|
33 |
+
```
|
34 |
+
Below is an instruction that describes a task. Write a response that appropriately completes the request.
|
35 |
+
### Instruction:
|
36 |
+
{prompt}
|
37 |
+
### Response:
|
38 |
+
```
|
39 |
+
|
40 |
+
## Provided files
|
41 |
+
|
42 |
+
| Name | Quant method | Bits | Use case |
|
43 |
+
| ---- | ---- | ---- | ----- |
|
44 |
+
| [silicon-masha-7b.Q2_K.gguf](https://huggingface.co/LakoMoor/Silicon-Masha-7B-GGUF/blob/main/silicon-masha-7b.Q2_K.gguf) | Q2_K | 2 | smallest, significant quality loss - not recommended for most purposes |
|
45 |
+
| [silicon-masha-7b.Q3_K_M.gguf](https://huggingface.co/LakoMoor/Silicon-Masha-7B-GGUF/blob/main/silicon-masha-7b.Q3_K_M.gguf) | Q3_K_M | 3 | very small, high quality loss |
|
46 |
+
| [silicon-masha-7b.Q4_0.gguf](https://huggingface.co/LakoMoor/Silicon-Masha-7B-GGUF/blob/main/silicon-masha-7b.Q4_0.gguf) | Q4_0 | 4 | legacy; small, very high quality loss - prefer using Q3_K_M |
|
47 |
+
| [silicon-masha-7b.Q4_K_M.gguf](https://huggingface.co/LakoMoor/Silicon-Masha-7B-GGUF/blob/main/silicon-masha-7b.Q4_K_M.gguf) | Q4_K_M | 4 | medium, balanced quality - recommended |
|
48 |
+
| [silicon-masha-7b.Q5_0.gguf](https://huggingface.co/LakoMoor/Silicon-Masha-7B-GGUF/blob/main/silicon-masha-7b.Q5_0.gguf) | Q5_0 | 5 | legacy; medium, balanced quality - prefer using Q4_K_M |
|
49 |
+
| [silicon-masha-7b.Q5_K_M.gguf](https://huggingface.co/LakoMoor/Silicon-Masha-7B-GGUF/blob/main/silicon-masha-7b.Q5_K_M.gguf) | Q5_K_M | 5 | large, very low quality loss - recommended |
|
50 |
+
| [silicon-masha-7b.Q6_K.gguf](https://huggingface.co/LakoMoor/Silicon-Masha-7B-GGUF/blob/main/silicon-masha-7b.Q6_K.gguf) | Q6_K | 6 | very large, extremely low quality loss |
|
51 |
+
| [silicon-masha-7b.Q8_0.gguf](https://huggingface.co/LakoMoor/Silicon-Masha-7B-GGUF/blob/main/silicon-masha-7b.Q8_0.gguf) | Q8_0 | 8 | very large, extremely low quality loss - not recommended |
|
52 |
+
|
53 |
+
## How run it ?
|
54 |
+
[llama.cpp](https://github.com/ggerganov/llama.cpp)
|
55 |
+
```shell
|
56 |
+
./main -ngl 35 -m silicon-masha-7b.Q4_K_M.gguf --color -c 32768 --temp 0.7 --repeat_penalty 1.1 -n -1 -p "Below is an instruction that describes a task. Write a response that appropriately completes the request.\n{system_message}\n### Instruction:{prompt}\n### Response:\n"
|
57 |
+
```
|
assets/MinP.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"temp": 0.8, "temperature_last": false, "top_p": 0.98, "top_k": 100, "top_a": 0, "tfs": 1, "epsilon_cutoff": 0, "eta_cutoff": 0, "typical_p": 1, "min_p": 0.05, "rep_pen": 1.1, "rep_pen_range": 2048, "no_repeat_ngram_size": 0, "penalty_alpha": 0, "num_beams": 1, "length_penalty": 1, "min_length": 0, "encoder_rep_pen": 1, "freq_pen": 0, "presence_pen": 0, "do_sample": true, "early_stopping": false, "add_bos_token": true, "truncation_length": 2048, "ban_eos_token": false, "skip_special_tokens": true, "streaming": true, "mirostat_mode": 0, "mirostat_tau": 5, "mirostat_eta": 0.1, "guidance_scale": 1, "negative_prompt": "", "grammar_string": "", "banned_tokens": "", "ignore_eos_token_aphrodite": false, "spaces_between_special_tokens_aphrodite": true, "type": "koboldcpp", "legacy_api": false, "sampler_order": [6, 0, 5, 3, 2, 1, 4], "n": 1, "rep_pen_size": 0, "genamt": 250, "max_length": 1024}
|
assets/alice.png
ADDED
Git LFS Details
|
assets/context.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"story_string": "### Instruction:\nWrite {{char}}'s next reply in a fictional roleplay chat between {{user}} and {{char}}. Use the provided character sheet and example dialogue for formatting direction and character speech patterns.\n\n{{#if system}}{{system}}\n\n{{/if}}{{#if wiBefore}}{{wiBefore}}\n\n{{/if}}Description of {{char}}:\n{{#if description}}{{description}}\n{{/if}}{{#if personality}}{{char}}'s personality: {{personality}}\n\n{{/if}}{{#if scenario}}Scenario: {{scenario}}\n\n{{/if}}{{#if persona}}Description of {{user}}: {{persona}}\n\n{{/if}}Play the role of {{char}}\n\n{{#if wiAfter}}{{wiAfter}}\n{{/if}}", "example_separator": "Example roleplay chat:", "chat_start": "Taking the above information into consideration,\nyou must engage in a roleplay conversation with {{user}} below this line.\nDo not write {{user}}'s dialogue lines in your responses.\n", "always_force_name2": true, "trim_sentences": true, "include_newline": true, "single_line": false, "name": "Context-Alice"}
|
assets/instruct.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"system_prompt": "Avoid repetition, don't loop. Develop the plot slowly, always stay in character. Describe all actions in full, elaborate, explicit, graphic, and vivid detail. Mention all relevant sensory perceptions.", "input_sequence": "\n### Instruction: (Style: Markdown, Present Tense)", "output_sequence": "\n### Response: (Style: Markdown, Present Tense)", "first_output_sequence": "### Response:", "last_output_sequence": "", "system_sequence_prefix": "", "system_sequence_suffix": "", "stop_sequence": "", "separator_sequence": "", "wrap": true, "macro": true, "names": true, "names_force_groups": true, "activation_regex": "", "name": "Instruct-Alice"}
|
assets/masha.png
ADDED
Git LFS Details
|
silicon-masha-7b.Q2_K.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:df4136053330117d6e90c884534a6fff6e4afafb9c0cd15506a441c4d51e8c8f
|
3 |
+
size 2719241920
|
silicon-masha-7b.Q3_K_M.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a946771d4f2c6577b1fecbf4612f70ce5da30285f98fa6d25b9be4bd2b3745ce
|
3 |
+
size 3518985920
|
silicon-masha-7b.Q4_0.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2aa5bc5bb37cafc4d6a52d9c264ae744410326539fd67e7c4dab10933e011d18
|
3 |
+
size 4108916416
|
silicon-masha-7b.Q4_K_M.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:736ef2db1b4cb1d1ae93e45189cf3c164d604083c13aa4588f00818310e22263
|
3 |
+
size 4368438976
|
silicon-masha-7b.Q5_0.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f07122686c0e0d9818381008db755237405ae8acb29327a48dc5102d92b8d52b
|
3 |
+
size 4997715648
|
silicon-masha-7b.Q5_K_M.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:19ffd5e475a38ba0ec1f845a8ff231635ba8a494f576b4086d9d036ccae55b50
|
3 |
+
size 5131409088
|
silicon-masha-7b.Q6_K.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:1c5eaf2870e9ea63648bb584091803a1bad1261a291da8d68703992810b36dff
|
3 |
+
size 5942064832
|