LoneStriker
commited on
Commit
•
09ec7f1
1
Parent(s):
fa3e5f6
Upload folder using huggingface_hub
Browse files- .gitattributes +15 -35
- README.md +94 -0
- huggingface-metadata.txt +30 -0
- wolfram_miqu-1-120b-Q2_K.gguf +3 -0
- wolfram_miqu-1-120b-Q3_K_L.gguf-part-a +3 -0
- wolfram_miqu-1-120b-Q3_K_L.gguf-part-b +3 -0
- wolfram_miqu-1-120b-Q3_K_M.gguf-part-a +3 -0
- wolfram_miqu-1-120b-Q3_K_M.gguf-part-b +3 -0
- wolfram_miqu-1-120b-Q3_K_S.gguf-part-a +3 -0
- wolfram_miqu-1-120b-Q3_K_S.gguf-part-b +3 -0
- wolfram_miqu-1-120b-Q4_K_M.gguf-part-a +3 -0
- wolfram_miqu-1-120b-Q4_K_M.gguf-part-b +3 -0
- wolfram_miqu-1-120b-Q4_K_S.gguf-part-a +3 -0
- wolfram_miqu-1-120b-Q4_K_S.gguf-part-b +3 -0
- wolfram_miqu-1-120b-Q5_K_M.gguf-part-a +3 -0
- wolfram_miqu-1-120b-Q5_K_M.gguf-part-b +3 -0
- wolfram_miqu-1-120b-Q5_K_S.gguf-part-a +3 -0
- wolfram_miqu-1-120b-Q5_K_S.gguf-part-b +3 -0
.gitattributes
CHANGED
@@ -1,35 +1,15 @@
|
|
1 |
-
|
2 |
-
|
3 |
-
|
4 |
-
|
5 |
-
|
6 |
-
|
7 |
-
|
8 |
-
|
9 |
-
|
10 |
-
|
11 |
-
|
12 |
-
|
13 |
-
|
14 |
-
|
15 |
-
|
16 |
-
*.onnx filter=lfs diff=lfs merge=lfs -text
|
17 |
-
*.ot filter=lfs diff=lfs merge=lfs -text
|
18 |
-
*.parquet filter=lfs diff=lfs merge=lfs -text
|
19 |
-
*.pb filter=lfs diff=lfs merge=lfs -text
|
20 |
-
*.pickle filter=lfs diff=lfs merge=lfs -text
|
21 |
-
*.pkl filter=lfs diff=lfs merge=lfs -text
|
22 |
-
*.pt filter=lfs diff=lfs merge=lfs -text
|
23 |
-
*.pth filter=lfs diff=lfs merge=lfs -text
|
24 |
-
*.rar filter=lfs diff=lfs merge=lfs -text
|
25 |
-
*.safetensors filter=lfs diff=lfs merge=lfs -text
|
26 |
-
saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
27 |
-
*.tar.* filter=lfs diff=lfs merge=lfs -text
|
28 |
-
*.tar filter=lfs diff=lfs merge=lfs -text
|
29 |
-
*.tflite filter=lfs diff=lfs merge=lfs -text
|
30 |
-
*.tgz filter=lfs diff=lfs merge=lfs -text
|
31 |
-
*.wasm filter=lfs diff=lfs merge=lfs -text
|
32 |
-
*.xz filter=lfs diff=lfs merge=lfs -text
|
33 |
-
*.zip filter=lfs diff=lfs merge=lfs -text
|
34 |
-
*.zst filter=lfs diff=lfs merge=lfs -text
|
35 |
-
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
|
|
1 |
+
wolfram_miqu-1-120b-Q2_K.gguf filter=lfs diff=lfs merge=lfs -text
|
2 |
+
wolfram_miqu-1-120b-Q3_K_L.gguf-part-a filter=lfs diff=lfs merge=lfs -text
|
3 |
+
wolfram_miqu-1-120b-Q3_K_L.gguf-part-b filter=lfs diff=lfs merge=lfs -text
|
4 |
+
wolfram_miqu-1-120b-Q3_K_M.gguf-part-a filter=lfs diff=lfs merge=lfs -text
|
5 |
+
wolfram_miqu-1-120b-Q3_K_M.gguf-part-b filter=lfs diff=lfs merge=lfs -text
|
6 |
+
wolfram_miqu-1-120b-Q3_K_S.gguf-part-a filter=lfs diff=lfs merge=lfs -text
|
7 |
+
wolfram_miqu-1-120b-Q3_K_S.gguf-part-b filter=lfs diff=lfs merge=lfs -text
|
8 |
+
wolfram_miqu-1-120b-Q4_K_M.gguf-part-a filter=lfs diff=lfs merge=lfs -text
|
9 |
+
wolfram_miqu-1-120b-Q4_K_M.gguf-part-b filter=lfs diff=lfs merge=lfs -text
|
10 |
+
wolfram_miqu-1-120b-Q4_K_S.gguf-part-a filter=lfs diff=lfs merge=lfs -text
|
11 |
+
wolfram_miqu-1-120b-Q4_K_S.gguf-part-b filter=lfs diff=lfs merge=lfs -text
|
12 |
+
wolfram_miqu-1-120b-Q5_K_M.gguf-part-a filter=lfs diff=lfs merge=lfs -text
|
13 |
+
wolfram_miqu-1-120b-Q5_K_M.gguf-part-b filter=lfs diff=lfs merge=lfs -text
|
14 |
+
wolfram_miqu-1-120b-Q5_K_S.gguf-part-a filter=lfs diff=lfs merge=lfs -text
|
15 |
+
wolfram_miqu-1-120b-Q5_K_S.gguf-part-b filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
README.md
ADDED
@@ -0,0 +1,94 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
base_model:
|
3 |
+
- 152334H/miqu-1-70b-sf
|
4 |
+
language:
|
5 |
+
- en
|
6 |
+
- de
|
7 |
+
- fr
|
8 |
+
- es
|
9 |
+
- it
|
10 |
+
library_name: transformers
|
11 |
+
tags:
|
12 |
+
- mergekit
|
13 |
+
- merge
|
14 |
+
---
|
15 |
+
# miqu-1-120b
|
16 |
+
|
17 |
+
![image/jpeg](https://cdn-uploads.huggingface.co/production/uploads/6303ca537373aacccd85d8a7/LxO9j7OykuabKLYQHIodG.jpeg)
|
18 |
+
|
19 |
+
* GGUF: [wolfram/miqu-1-120b-GGUF](https://huggingface.co/wolfram/miqu-1-120b-GGUF)
|
20 |
+
|
21 |
+
This is a 120b frankenmerge of [miqu-1-70b](https://huggingface.co/miqudev/miqu-1-70b) created by interleaving layers of [miqu-1-70b-sf](https://huggingface.co/152334H/miqu-1-70b-sf) with itself using [mergekit](https://github.com/cg123/mergekit).
|
22 |
+
|
23 |
+
Inspired by [Venus-120b-v1.2](https://huggingface.co/nsfwthrowitaway69/Venus-120b-v1.2), [MegaDolphin-120b](https://huggingface.co/cognitivecomputations/MegaDolphin-120b), and [goliath-120b](https://huggingface.co/alpindale/goliath-120b).
|
24 |
+
|
25 |
+
Thanks for the support, [CopilotKit](https://github.com/CopilotKit/CopilotKit) - the open-source platform for building in-app AI Copilots into any product, with any LLM model. Check out their GitHub.
|
26 |
+
|
27 |
+
## Prompt template: Mistral
|
28 |
+
|
29 |
+
```
|
30 |
+
<s>[INST] {prompt} [/INST]
|
31 |
+
```
|
32 |
+
|
33 |
+
See also: [🐺🐦⬛ LLM Prompt Format Comparison/Test: Mixtral 8x7B Instruct with **17** different instruct templates : LocalLLaMA](https://www.reddit.com/r/LocalLLaMA/comments/18ljvxb/llm_prompt_format_comparisontest_mixtral_8x7b/)
|
34 |
+
|
35 |
+
## Model Details
|
36 |
+
|
37 |
+
* Max Context: 32764 tokens (kept the weird number from the original/base model)
|
38 |
+
* Layers: 140
|
39 |
+
|
40 |
+
## Merge Details
|
41 |
+
|
42 |
+
### Merge Method
|
43 |
+
|
44 |
+
This model was merged using the passthrough merge method.
|
45 |
+
|
46 |
+
### Models Merged
|
47 |
+
|
48 |
+
The following models were included in the merge:
|
49 |
+
* [152334H/miqu-1-70b-sf](https://huggingface.co/152334H/miqu-1-70b-sf)
|
50 |
+
|
51 |
+
### Configuration
|
52 |
+
|
53 |
+
The following YAML configuration was used to produce this model:
|
54 |
+
|
55 |
+
```yaml
|
56 |
+
dtype: float16
|
57 |
+
merge_method: passthrough
|
58 |
+
slices:
|
59 |
+
- sources:
|
60 |
+
- layer_range: [0, 20]
|
61 |
+
model: 152334H/miqu-1-70b-sf
|
62 |
+
- sources:
|
63 |
+
- layer_range: [10, 30]
|
64 |
+
model: 152334H/miqu-1-70b-sf
|
65 |
+
- sources:
|
66 |
+
- layer_range: [20, 40]
|
67 |
+
model: 152334H/miqu-1-70b-sf
|
68 |
+
- sources:
|
69 |
+
- layer_range: [30, 50]
|
70 |
+
model: 152334H/miqu-1-70b-sf
|
71 |
+
- sources:
|
72 |
+
- layer_range: [40, 60]
|
73 |
+
model: 152334H/miqu-1-70b-sf
|
74 |
+
- sources:
|
75 |
+
- layer_range: [50, 70]
|
76 |
+
model: 152334H/miqu-1-70b-sf
|
77 |
+
- sources:
|
78 |
+
- layer_range: [60, 80]
|
79 |
+
model: 152334H/miqu-1-70b-sf
|
80 |
+
```
|
81 |
+
|
82 |
+
## Credits & Special Thanks
|
83 |
+
|
84 |
+
* original (unreleased) model: [mistralai (Mistral AI_)](https://huggingface.co/mistralai)
|
85 |
+
* leaked model: [miqudev/miqu-1-70b](https://huggingface.co/miqudev/miqu-1-70b)
|
86 |
+
* f16 model: [152334H/miqu-1-70b-sf](https://huggingface.co/152334H/miqu-1-70b-sf)
|
87 |
+
* mergekit: [arcee-ai/mergekit: Tools for merging pretrained large language models.](https://github.com/arcee-ai/mergekit)
|
88 |
+
* mergekit_config.yml: [nsfwthrowitaway69/Venus-120b-v1.2](https://huggingface.co/nsfwthrowitaway69/Venus-120b-v1.2)
|
89 |
+
|
90 |
+
### Support
|
91 |
+
|
92 |
+
* [My Ko-fi page](https://ko-fi.com/wolframravenwolf) if you'd like to tip me to say thanks or request specific models to be tested or merged with priority. Also consider supporting your favorite model creators, quantizers, or frontend/backend devs if you can afford to do so. They deserve it!
|
93 |
+
|
94 |
+
#### DISCLAIMER: THIS IS [BASED ON A LEAKED ASSET](https://huggingface.co/miqudev/miqu-1-70b/discussions/10) AND HAS NO LICENSE ASSOCIATED WITH IT. USE AT YOUR OWN RISK.
|
huggingface-metadata.txt
ADDED
@@ -0,0 +1,30 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
url: https://huggingface.co/wolfram/miqu-1-120b
|
2 |
+
branch: main
|
3 |
+
download date: 2024-02-05 14:50:58
|
4 |
+
sha256sum:
|
5 |
+
54b8f46c2110b39ddab602498b0fff0c2062606885ffb34c0daa7dbec2716091 model-00001-of-00025.safetensors
|
6 |
+
a250311f9911bea8e0c42b3419c63fbc81d93bc8a76984432e24db30d61ff9ed model-00002-of-00025.safetensors
|
7 |
+
ed4b011d8573eaf221bef28753eaa75b2d09bceec57f0f6ebb74bfe2432d1832 model-00003-of-00025.safetensors
|
8 |
+
535a0d7db51cebf4d3456ebb2361e9623ae2e53bdd4fe2a8294f9c1cc60070ea model-00004-of-00025.safetensors
|
9 |
+
f166af321c02ac4df4203688af41598442cf0f3094e1e59f223ec4546e6d5a00 model-00005-of-00025.safetensors
|
10 |
+
6001c62f68ec5f0a9e2850b42695e6113f4412cec69461c6286999142e887596 model-00006-of-00025.safetensors
|
11 |
+
ac3dca2f63978359d6b2b54bc662b009b934ad6f9a424bd0d18f6d459e7d5d4a model-00007-of-00025.safetensors
|
12 |
+
56fef43c801506eead1d4d47a2b1e1bb7d2b9e49e0f3a5abffcc7336a3c93b56 model-00008-of-00025.safetensors
|
13 |
+
7ac433202aa77b39d6a702dbfb861b394f8b0dcd405b68cdd464d98bcbf0f395 model-00009-of-00025.safetensors
|
14 |
+
6aaa62a70a7dd0999630c6ae6de65fdc2c4f399283759af07b7489c221e0aeaa model-00010-of-00025.safetensors
|
15 |
+
d87c10c249caa9aa263ce29bc48d9574d48d4a8117370bbdb056f4a088dd9395 model-00011-of-00025.safetensors
|
16 |
+
1a2f6f036c8ad9bd5962ad57345ed6e3135484fbd8124abae543a1bf56fcc564 model-00012-of-00025.safetensors
|
17 |
+
ae2880cf7c0b35db43295944bcee0d087c0683448d8d3747f41175a2b3318c11 model-00013-of-00025.safetensors
|
18 |
+
ca4a15c0f7cf6e481626e88d481699032ff06359d21b907ef13a33dc14b53caa model-00014-of-00025.safetensors
|
19 |
+
e3dc2549a66181e58ae39582a7799a251e6978e3335b25b07e715c621c0b23f1 model-00015-of-00025.safetensors
|
20 |
+
4c285d1c2eb1ea32e752d1dc3312935930b32c82a7c501cc78c72c437c6521fd model-00016-of-00025.safetensors
|
21 |
+
6a4c08832aabb409b1a871d0d0de1d55ded9656d8a4a8d2a2d4b6a7909391b74 model-00017-of-00025.safetensors
|
22 |
+
d803c1a099ec7d7b21cf4b72bc40d99cbdc00f49c7ccd9fed305706104215c77 model-00018-of-00025.safetensors
|
23 |
+
3878227390b06711d277c682253a58c89332ddc76a49a2fb81a99a5daf4efc1b model-00019-of-00025.safetensors
|
24 |
+
c81cbe51b99658f87cf282523aa291a478725442687f3c015ec35fb709d25d72 model-00020-of-00025.safetensors
|
25 |
+
6dec809081c6e3882322818e1887dee13cc0d6b3e2570004cf9555a377ebbc5b model-00021-of-00025.safetensors
|
26 |
+
431cfc711110ebeeae31a78ccae25f335ef362686f2eb22ee82729a206350978 model-00022-of-00025.safetensors
|
27 |
+
4c571650d30a7f58e8712734916f48f8b3df6b084ddfc38ca346db6638a4c8c2 model-00023-of-00025.safetensors
|
28 |
+
3ca56d3a4d79c9c25b75d006d63b85424d59440753e981b21056eeef1641a3c8 model-00024-of-00025.safetensors
|
29 |
+
d876e6fbdc512e6b2f0debdd5a8691e3636eb79babcd358f2d9807842f6b6be7 model-00025-of-00025.safetensors
|
30 |
+
9e556afd44213b6bd1be2b850ebbbd98f5481437a8021afaf58ee7fb1818d347 tokenizer.model
|
wolfram_miqu-1-120b-Q2_K.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8c5928c3481dc3a70d2a5e895db191548c34c9ce21cc8e590ccd483e425ee6f6
|
3 |
+
size 44186105280
|
wolfram_miqu-1-120b-Q3_K_L.gguf-part-a
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:bbf5bf55205becf3e6aab117ab561e705c66f74caec6087611c9a9e28cc642fb
|
3 |
+
size 31506193120
|
wolfram_miqu-1-120b-Q3_K_L.gguf-part-b
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a8f63e66ce12eba7d1f36c0e83139ab6a260f142e1308e0c223fadda5a37b26b
|
3 |
+
size 31506193120
|
wolfram_miqu-1-120b-Q3_K_M.gguf-part-a
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:dff4d67ef0507a54cd950bb1a478c80cf75cbb950a3e881df22496b1f328d541
|
3 |
+
size 28908870368
|
wolfram_miqu-1-120b-Q3_K_M.gguf-part-b
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a6ec0a35c1e4c2906cf3a0e3415c0bfb35a5708991406bf872aecb13b53213be
|
3 |
+
size 28908870368
|
wolfram_miqu-1-120b-Q3_K_S.gguf-part-a
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:0c9e9fcec86a74b684b0ebf8cfd48790c839eca8f2ebf527515e52f6e94c3552
|
3 |
+
size 25904831200
|
wolfram_miqu-1-120b-Q3_K_S.gguf-part-b
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:36f8c6c6902209039fc32adaf248395519325109cede81dafd5ff971c8c6807e
|
3 |
+
size 25904831200
|
wolfram_miqu-1-120b-Q4_K_M.gguf-part-a
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7d4b167ac49b7cd7ab052a816cc3a1f310e997c4807286393f4b2e1f027536c4
|
3 |
+
size 36072127200
|
wolfram_miqu-1-120b-Q4_K_M.gguf-part-b
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:161de81c9d9714626e5af26311d4329041dd4d3ccd5483232fa42b36c8614b93
|
3 |
+
size 36072127200
|
wolfram_miqu-1-120b-Q4_K_S.gguf-part-a
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:84417b1ea887eff70eba21a9ae2d05028a9f7a8e1ce60d448a52496483399184
|
3 |
+
size 34128657120
|
wolfram_miqu-1-120b-Q4_K_S.gguf-part-b
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:65c0fec773f35d4f634c6b748f8478bff3a840326178779e3b9f6c514638c6d0
|
3 |
+
size 34128657120
|
wolfram_miqu-1-120b-Q5_K_M.gguf-part-a
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:59f509e85ed3003a2c173a46aabff2eb34bcec64eca2c2c43ce6e10c25ddbe2c
|
3 |
+
size 42511039200
|
wolfram_miqu-1-120b-Q5_K_M.gguf-part-b
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e8d77628211cc61c80e10f18e59287f8b5453b175ccd0fa2632f712f0b3d7ed2
|
3 |
+
size 42511039200
|
wolfram_miqu-1-120b-Q5_K_S.gguf-part-a
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9722161060990c3fa032bdca59d899e88c6df73a052821c580f6166c8463060a
|
3 |
+
size 41380215520
|
wolfram_miqu-1-120b-Q5_K_S.gguf-part-b
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:94dd2a401aff658828b39265ea7eb8bcee6e26c37fe5cfef0886315fb8e008e3
|
3 |
+
size 41380215520
|