aashish1904 commited on
Commit
9ae29cd
1 Parent(s): f3590bc

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +90 -0
README.md ADDED
@@ -0,0 +1,90 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+ ---
3
+
4
+ base_model:
5
+ - FPHam/L3-8B-Everything-COT
6
+ - ChaoticNeutrals/Domain-Fusion-L3-8B
7
+ - Ayush-1722/Meta-Llama-3-8B-Instruct-Summarize-v0.2-24K-LoRANET-Merged
8
+ - Sao10K/L3-8B-Niitama-v1
9
+ - nothingiisreal/L3-8B-Celeste-V1.2
10
+ - Orenguteng/Llama-3-8B-Lexi-Uncensored
11
+ - OEvortex/Emotional-llama-8B
12
+ - Sao10K/L3-8B-Lunaris-v1
13
+ library_name: transformers
14
+ tags:
15
+ - mergekit
16
+ - merge
17
+
18
+
19
+ ---
20
+
21
+ [![QuantFactory Banner](https://lh7-rt.googleusercontent.com/docsz/AD_4nXeiuCm7c8lEwEJuRey9kiVZsRn2W-b4pWlu3-X534V3YmVuVc2ZL-NXg2RkzSOOS2JXGHutDuyyNAUtdJI65jGTo8jT9Y99tMi4H4MqL44Uc5QKG77B0d6-JfIkZHFaUA71-RtjyYZWVIhqsNZcx8-OMaA?key=xt3VSDoCbmTY7o-cwwOFwQ)](https://hf.co/QuantFactory)
22
+
23
+
24
+ # QuantFactory/L3-Luna-8B-GGUF
25
+ This is quantized version of [Casual-Autopsy/L3-Luna-8B](https://huggingface.co/Casual-Autopsy/L3-Luna-8B) created using llama.cpp
26
+
27
+ # Original Model Card
28
+
29
+ # merge
30
+
31
+ This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
32
+
33
+ ## Merge Details
34
+ ### Merge Method
35
+
36
+ This model was merged using the [TIES](https://arxiv.org/abs/2306.01708) merge method using [Sao10K/L3-8B-Lunaris-v1](https://huggingface.co/Sao10K/L3-8B-Lunaris-v1) as a base.
37
+
38
+ ### Models Merged
39
+
40
+ The following models were included in the merge:
41
+ * [FPHam/L3-8B-Everything-COT](https://huggingface.co/FPHam/L3-8B-Everything-COT)
42
+ * [ChaoticNeutrals/Domain-Fusion-L3-8B](https://huggingface.co/ChaoticNeutrals/Domain-Fusion-L3-8B)
43
+ * [Ayush-1722/Meta-Llama-3-8B-Instruct-Summarize-v0.2-24K-LoRANET-Merged](https://huggingface.co/Ayush-1722/Meta-Llama-3-8B-Instruct-Summarize-v0.2-24K-LoRANET-Merged)
44
+ * [Sao10K/L3-8B-Niitama-v1](https://huggingface.co/Sao10K/L3-8B-Niitama-v1)
45
+ * [nothingiisreal/L3-8B-Celeste-V1.2](https://huggingface.co/nothingiisreal/L3-8B-Celeste-V1.2)
46
+ * [Orenguteng/Llama-3-8B-Lexi-Uncensored](https://huggingface.co/Orenguteng/Llama-3-8B-Lexi-Uncensored)
47
+ * [OEvortex/Emotional-llama-8B](https://huggingface.co/OEvortex/Emotional-llama-8B)
48
+
49
+ ### Configuration
50
+
51
+ The following YAML configuration was used to produce this model:
52
+
53
+ ```yaml
54
+ models:
55
+ - model: Sao10K/L3-8B-Lunaris-v1
56
+ - model: FPHam/L3-8B-Everything-COT
57
+ parameters:
58
+ density: 0.5
59
+ weight: 0.1
60
+ - model: Ayush-1722/Meta-Llama-3-8B-Instruct-Summarize-v0.2-24K-LoRANET-Merged
61
+ parameters:
62
+ density: 0.5
63
+ weight: 0.1
64
+ - model: OEvortex/Emotional-llama-8B
65
+ parameters:
66
+ density: 0.5
67
+ weight: 0.1
68
+ - model: ChaoticNeutrals/Domain-Fusion-L3-8B
69
+ parameters:
70
+ density: 0.75
71
+ weight: 0.05
72
+ - model: nothingiisreal/L3-8B-Celeste-V1.2
73
+ parameters:
74
+ density: 0.75
75
+ weight: 0.05
76
+ - model: Orenguteng/Llama-3-8B-Lexi-Uncensored
77
+ parameters:
78
+ density: 0.75
79
+ weight: 0.05
80
+ - model: Sao10K/L3-8B-Niitama-v1
81
+ parameters:
82
+ density: 0.75
83
+ weight: 0.05
84
+ base_model: Sao10K/L3-8B-Lunaris-v1
85
+ merge_method: ties
86
+ parameters:
87
+ normalize: true
88
+ dtype: bfloat16
89
+ ```
90
+