munish0838 commited on
Commit
b9ac146
·
verified ·
1 Parent(s): 4974f4e

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +192 -0
README.md ADDED
@@ -0,0 +1,192 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+ ---
3
+
4
+ library_name: transformers
5
+ tags:
6
+ - mergekit
7
+ - merge
8
+ base_model:
9
+ - hf-100/Llama-3-Spellbound-Instruct-8B-0.3
10
+ - unsloth/Meta-Llama-3.1-8B
11
+ - arcee-ai/Llama-3.1-SuperNova-Lite
12
+ - grimjim/Llama-3-Instruct-abliteration-LoRA-8B
13
+ - THUDM/LongWriter-llama3.1-8b
14
+ - ResplendentAI/Smarts_Llama3
15
+ - djuna/L3.1-Suze-Vume-2-calc
16
+ - djuna/L3.1-ForStHS
17
+ - Blackroot/Llama-3-8B-Abomination-LORA
18
+ model-index:
19
+ - name: L3.1-Purosani-2-8B
20
+ results:
21
+ - task:
22
+ type: text-generation
23
+ name: Text Generation
24
+ dataset:
25
+ name: IFEval (0-Shot)
26
+ type: HuggingFaceH4/ifeval
27
+ args:
28
+ num_few_shot: 0
29
+ metrics:
30
+ - type: inst_level_strict_acc and prompt_level_strict_acc
31
+ value: 49.88
32
+ name: strict accuracy
33
+ source:
34
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=djuna/L3.1-Purosani-2-8B
35
+ name: Open LLM Leaderboard
36
+ - task:
37
+ type: text-generation
38
+ name: Text Generation
39
+ dataset:
40
+ name: BBH (3-Shot)
41
+ type: BBH
42
+ args:
43
+ num_few_shot: 3
44
+ metrics:
45
+ - type: acc_norm
46
+ value: 31.39
47
+ name: normalized accuracy
48
+ source:
49
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=djuna/L3.1-Purosani-2-8B
50
+ name: Open LLM Leaderboard
51
+ - task:
52
+ type: text-generation
53
+ name: Text Generation
54
+ dataset:
55
+ name: MATH Lvl 5 (4-Shot)
56
+ type: hendrycks/competition_math
57
+ args:
58
+ num_few_shot: 4
59
+ metrics:
60
+ - type: exact_match
61
+ value: 10.12
62
+ name: exact match
63
+ source:
64
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=djuna/L3.1-Purosani-2-8B
65
+ name: Open LLM Leaderboard
66
+ - task:
67
+ type: text-generation
68
+ name: Text Generation
69
+ dataset:
70
+ name: GPQA (0-shot)
71
+ type: Idavidrein/gpqa
72
+ args:
73
+ num_few_shot: 0
74
+ metrics:
75
+ - type: acc_norm
76
+ value: 6.82
77
+ name: acc_norm
78
+ source:
79
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=djuna/L3.1-Purosani-2-8B
80
+ name: Open LLM Leaderboard
81
+ - task:
82
+ type: text-generation
83
+ name: Text Generation
84
+ dataset:
85
+ name: MuSR (0-shot)
86
+ type: TAUR-Lab/MuSR
87
+ args:
88
+ num_few_shot: 0
89
+ metrics:
90
+ - type: acc_norm
91
+ value: 8.3
92
+ name: acc_norm
93
+ source:
94
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=djuna/L3.1-Purosani-2-8B
95
+ name: Open LLM Leaderboard
96
+ - task:
97
+ type: text-generation
98
+ name: Text Generation
99
+ dataset:
100
+ name: MMLU-PRO (5-shot)
101
+ type: TIGER-Lab/MMLU-Pro
102
+ config: main
103
+ split: test
104
+ args:
105
+ num_few_shot: 5
106
+ metrics:
107
+ - type: acc
108
+ value: 30.57
109
+ name: accuracy
110
+ source:
111
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=djuna/L3.1-Purosani-2-8B
112
+ name: Open LLM Leaderboard
113
+
114
+ ---
115
+
116
+ [![QuantFactory Banner](https://lh7-rt.googleusercontent.com/docsz/AD_4nXeiuCm7c8lEwEJuRey9kiVZsRn2W-b4pWlu3-X534V3YmVuVc2ZL-NXg2RkzSOOS2JXGHutDuyyNAUtdJI65jGTo8jT9Y99tMi4H4MqL44Uc5QKG77B0d6-JfIkZHFaUA71-RtjyYZWVIhqsNZcx8-OMaA?key=xt3VSDoCbmTY7o-cwwOFwQ)](https://hf.co/QuantFactory)
117
+
118
+
119
+ # QuantFactory/L3.1-Purosani-2-8B-GGUF
120
+ This is quantized version of [djuna/L3.1-Purosani-2-8B](https://huggingface.co/djuna/L3.1-Purosani-2-8B) created using llama.cpp
121
+
122
+ # Original Model Card
123
+
124
+ # merge
125
+
126
+ This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
127
+
128
+ ## Merge Details
129
+ ### Merge Method
130
+
131
+ This model was merged using the della_linear merge method using [unsloth/Meta-Llama-3.1-8B](https://huggingface.co/unsloth/Meta-Llama-3.1-8B) as a base.
132
+
133
+ ### Models Merged
134
+
135
+ The following models were included in the merge:
136
+ * [hf-100/Llama-3-Spellbound-Instruct-8B-0.3](https://huggingface.co/hf-100/Llama-3-Spellbound-Instruct-8B-0.3)
137
+ * [arcee-ai/Llama-3.1-SuperNova-Lite](https://huggingface.co/arcee-ai/Llama-3.1-SuperNova-Lite) + [grimjim/Llama-3-Instruct-abliteration-LoRA-8B](https://huggingface.co/grimjim/Llama-3-Instruct-abliteration-LoRA-8B)
138
+ * [THUDM/LongWriter-llama3.1-8b](https://huggingface.co/THUDM/LongWriter-llama3.1-8b) + [ResplendentAI/Smarts_Llama3](https://huggingface.co/ResplendentAI/Smarts_Llama3)
139
+ * [djuna/L3.1-Suze-Vume-2-calc](https://huggingface.co/djuna/L3.1-Suze-Vume-2-calc)
140
+ * [djuna/L3.1-ForStHS](https://huggingface.co/djuna/L3.1-ForStHS) + [Blackroot/Llama-3-8B-Abomination-LORA](https://huggingface.co/Blackroot/Llama-3-8B-Abomination-LORA)
141
+
142
+ ### Configuration
143
+
144
+ The following YAML configuration was used to produce this model:
145
+
146
+ ```yaml
147
+ merge_method: della_linear
148
+ dtype: bfloat16
149
+ parameters:
150
+ epsilon: 0.1
151
+ lambda: 1.0
152
+ int8_mask: true
153
+ normalize: true
154
+ base_model: unsloth/Meta-Llama-3.1-8B
155
+ models:
156
+ - model: arcee-ai/Llama-3.1-SuperNova-Lite+grimjim/Llama-3-Instruct-abliteration-LoRA-8B
157
+ parameters:
158
+ weight: 1
159
+ density: 0.5
160
+ - model: hf-100/Llama-3-Spellbound-Instruct-8B-0.3
161
+ parameters:
162
+ weight: 1
163
+ density: 0.45
164
+ - model: djuna/L3.1-Suze-Vume-2-calc
165
+ parameters:
166
+ weight: 1
167
+ density: 0.45
168
+ - model: THUDM/LongWriter-llama3.1-8b+ResplendentAI/Smarts_Llama3
169
+ parameters:
170
+ weight: 1
171
+ density: 0.55
172
+ - model: djuna/L3.1-ForStHS+Blackroot/Llama-3-8B-Abomination-LORA
173
+ parameters:
174
+ weight: 1
175
+ density: 0.5
176
+
177
+ ```
178
+
179
+ # [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard)
180
+ Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_djuna__L3.1-Purosani-2-8B)
181
+
182
+ | Metric |Value|
183
+ |-------------------|----:|
184
+ |Avg. |22.85|
185
+ |IFEval (0-Shot) |49.88|
186
+ |BBH (3-Shot) |31.39|
187
+ |MATH Lvl 5 (4-Shot)|10.12|
188
+ |GPQA (0-shot) | 6.82|
189
+ |MuSR (0-shot) | 8.30|
190
+ |MMLU-PRO (5-shot) |30.57|
191
+
192
+