Triangle104 commited on
Commit
fa42293
·
verified ·
1 Parent(s): 92ee3c0

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +45 -0
README.md CHANGED
@@ -15,6 +15,51 @@ base_model: crestf411/L3.1-8B-Dark-Planet-Slush
15
  This model was converted to GGUF format from [`crestf411/L3.1-8B-Dark-Planet-Slush`](https://huggingface.co/crestf411/L3.1-8B-Dark-Planet-Slush) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
16
  Refer to the [original model card](https://huggingface.co/crestf411/L3.1-8B-Dark-Planet-Slush) for more details on the model.
17
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
18
  ## Use with llama.cpp
19
  Install llama.cpp through brew (works on Mac and Linux)
20
 
 
15
  This model was converted to GGUF format from [`crestf411/L3.1-8B-Dark-Planet-Slush`](https://huggingface.co/crestf411/L3.1-8B-Dark-Planet-Slush) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
16
  Refer to the [original model card](https://huggingface.co/crestf411/L3.1-8B-Dark-Planet-Slush) for more details on the model.
17
 
18
+ ---
19
+ Model details:
20
+ -
21
+ This is based on v1.1 and includes a merge with DavidAU/L3.1-Dark-Planet-SpinFire-Uncensored-8B.
22
+
23
+ Parameter suggestions:
24
+ -
25
+ I did all my testing with temp 1, min-p 0.1, DRY 0.8. I enabled XTC at higher contexts.
26
+
27
+ Merge Details
28
+ -
29
+ This model was merged using the TIES merge method using meta-llama/Llama-3.1-8B as a base.
30
+
31
+ Configuration
32
+ -
33
+ The following YAML configuration was used to produce this model:
34
+
35
+ models:
36
+ - model: stage1-on-instruct
37
+ parameters:
38
+ weight: 1
39
+ density: 1
40
+ - model: stage2-on-stage1
41
+ parameters:
42
+ weight: 1
43
+ density: 1
44
+ - model: DavidAU/L3.1-Dark-Planet-SpinFire-Uncensored-8B
45
+ parameters:
46
+ weight: 1
47
+ density: 1
48
+ - model: meta-llama/Llama-3.1-8B-Instruct
49
+ parameters:
50
+ weight: 1.3
51
+ density: 1
52
+ merge_method: ties
53
+ base_model: meta-llama/Llama-3.1-8B
54
+ parameters:
55
+ weight: 1
56
+ density: 1
57
+ normalize: true
58
+ int8_mask: true
59
+ tokenizer_source: meta-llama/Llama-3.1-8B-Instruct
60
+ dtype: bfloat16
61
+
62
+ ---
63
  ## Use with llama.cpp
64
  Install llama.cpp through brew (works on Mac and Linux)
65