aashish1904 commited on
Commit
f5532e2
1 Parent(s): 16593f1

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +120 -0
README.md ADDED
@@ -0,0 +1,120 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+ ---
3
+
4
+ base_model:
5
+ - princeton-nlp/Llama-3-Instruct-8B-SimPO
6
+ - Sao10K/L3-8B-Stheno-v3.2
7
+ library_name: transformers
8
+ tags:
9
+ - mergekit
10
+ - merge
11
+ - roleplay
12
+ - sillytavern
13
+ - llama3
14
+ - not-for-all-audiences
15
+ license: cc-by-nc-4.0
16
+ language:
17
+ - en
18
+
19
+ ---
20
+
21
+ [![QuantFactory Banner](https://lh7-rt.googleusercontent.com/docsz/AD_4nXeiuCm7c8lEwEJuRey9kiVZsRn2W-b4pWlu3-X534V3YmVuVc2ZL-NXg2RkzSOOS2JXGHutDuyyNAUtdJI65jGTo8jT9Y99tMi4H4MqL44Uc5QKG77B0d6-JfIkZHFaUA71-RtjyYZWVIhqsNZcx8-OMaA?key=xt3VSDoCbmTY7o-cwwOFwQ)](https://hf.co/QuantFactory)
22
+
23
+
24
+ # QuantFactory/L3-Nymeria-8B-GGUF
25
+ This is quantized version of [tannedbum/L3-Nymeria-8B](https://huggingface.co/tannedbum/L3-Nymeria-8B) created using llama.cpp
26
+
27
+ # Original Model Card
28
+
29
+ ![Nymeria](https://huggingface.co/tannedbum/L3-Nymeria-8B/resolve/main/Nymeria.png?)
30
+
31
+ ## The smartest L3 8B model combined with high-end RP model. What could go wrong.
32
+
33
+ The idea was to fuse a bit of SimPO's realism with Stheno. It took a few days to come up with a balanced slerp configuration, but I'm more than satisfied with the end result.
34
+
35
+
36
+ ## SillyTavern
37
+
38
+ ## Text Completion presets
39
+ ```
40
+ temp 0.9
41
+ top_k 30
42
+ top_p 0.75
43
+ min_p 0.2
44
+ rep_pen 1.1
45
+ smooth_factor 0.25
46
+ smooth_curve 1
47
+ ```
48
+ ## Advanced Formatting
49
+
50
+ [Context & Instruct preset by Virt-io](https://huggingface.co/Virt-io/SillyTavern-Presets/tree/main/Prompts/LLAMA-3/v1.9)
51
+
52
+ Instruct Mode: Enabled
53
+
54
+
55
+
56
+ # merge
57
+
58
+ This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
59
+
60
+ This model was merged using the slerp merge method.
61
+
62
+ ### Models Merged
63
+
64
+ The following models were included in the merge:
65
+ * [Sao10K/L3-8B-Stheno-v3.2](https://huggingface.co/Sao10K/L3-8B-Stheno-v3.2)
66
+ * [princeton-nlp/Llama-3-Instruct-8B-SimPO](https://huggingface.co/princeton-nlp/Llama-3-Instruct-8B-SimPO)
67
+
68
+ ### Configuration
69
+
70
+ The following YAML configuration was used to produce this model:
71
+
72
+ ```yaml
73
+
74
+ slices:
75
+ - sources:
76
+ - model: Sao10K/L3-8B-Stheno-v3.2
77
+ layer_range: [0, 32]
78
+ - model: princeton-nlp/Llama-3-Instruct-8B-SimPO
79
+ layer_range: [0, 32]
80
+ merge_method: slerp
81
+ base_model: Sao10K/L3-8B-Stheno-v3.2
82
+ parameters:
83
+ t:
84
+ - filter: self_attn
85
+ value: [0.4, 0.5, 0.6, 0.4, 0.6]
86
+ - filter: mlp
87
+ value: [0.6, 0.5, 0.4, 0.6, 0.4]
88
+ - value: 0.5
89
+ dtype: bfloat16
90
+
91
+
92
+ ```
93
+
94
+ ---
95
+
96
+ ## Original model information:
97
+
98
+ ## Model: Sao10K/L3-8B-Stheno-v3.2
99
+
100
+ Stheno-v3.2-Zeta
101
+
102
+
103
+ Changes compared to v3.1
104
+ <br>\- Included a mix of SFW and NSFW Storywriting Data, thanks to [Gryphe](https://huggingface.co/datasets/Gryphe/Opus-WritingPrompts)
105
+ <br>\- Included More Instruct / Assistant-Style Data
106
+ <br>\- Further cleaned up Roleplaying Samples from c2 Logs -> A few terrible, really bad samples escaped heavy filtering. Manual pass fixed it.
107
+ <br>\- Hyperparameter tinkering for training, resulting in lower loss levels.
108
+
109
+
110
+ Testing Notes - Compared to v3.1
111
+ <br>\- Handles SFW / NSFW seperately better. Not as overly excessive with NSFW now. Kinda balanced.
112
+ <br>\- Better at Storywriting / Narration.
113
+ <br>\- Better at Assistant-type Tasks.
114
+ <br>\- Better Multi-Turn Coherency -> Reduced Issues?
115
+ <br>\- Slightly less creative? A worthy tradeoff. Still creative.
116
+ <br>\- Better prompt / instruction adherence.
117
+
118
+ ---
119
+
120
+ Want to support my work ? My Ko-fi page: https://ko-fi.com/tannedbum