Upload folder using huggingface_hub (#1)
Browse files- b49b19830c3e6eabe73c9d1ad24bcfb7c42605fb187e5f1c68cd1db12ec37def (52c257b00e629c2cc1563c8d51d931bf570b50b3)
- 872feaf4cc5438fde5e967a7e7181b82bd80609531c0d7d93ba064f7186e5b40 (fc8903c9eeabb199e2806da40e5bb5145c4d4307)
- c3eebcfb6154da588b09006d871db513741f65bd58802458d37e0086c8c51e45 (578dcb7b8f8e0524b490dc162087e875c63f288c)
- add0382ee8ac59fcd38e16d2834c7bb5e0638ca5e258702836be80de2036bb20 (8162d040e71a6711be529d1f4cb77e5182027c6f)
- c815939ea5d5bf475838a60dab19a7750c6203978d41928217eb20e8ed42abe6 (26bf9f50867029540fb4203a91a688ae1d9c36f6)
- bae2499279ff3b5c0f4245da05592fa5dc3ad07e057d3fd7714087bf4f1f2ec0 (908f7f7f0ae2bc8e65f02436329e07f993e0d16f)
- 7e4fcafecd52691a71c67b198742586ed7e2f0f5ec839118c4c6ce91987fa89d (d353f90508eee4f2bd0fa39ec1255f328c952943)
- 1d15e352d7633a70b8c19d1a7924422c09df6b84d93efbe4abd6b6d80c01e4fe (4a42b06142b41988b8557def17e25ba89edd1a92)
- 6b39028b15d1b65e8d86074ae567c1f0f1b4f83545b559d9365583554c1f12b2 (7552902e8aef3627d1fa927c5db1622ac196d30d)
- 3d44b3fa3d910b07e810f497e573e1e26e5b773e8a8148a66ccbaed7721435bf (4ae5def012a70640f39ff00e005df9756e4733ca)
- 64017e9c65220213803a331244a9a06f9548913618c144218dcafe7aafb41276 (bf15603cc188e1f8228bb8344f433e91d49f5229)
- 181ed9751a72824d77f5dce31db78fe51b7edf6c47303b3731e582780bc2f302 (789b74f33fa7ab0b396264ec3d3fdb50396de32f)
- 53b9435a66f135e8e1c282be019c17a987dfe38c4fe417de2faa516163eb6c26 (e6ea3e545ce2df3c4b933f53e6f71419f016d8ba)
- README.md +26 -0
- added_tokens.json +5 -0
- config.json +39 -0
- merges.txt +0 -0
- model-00001-of-00013.safetensors +3 -0
- model-00002-of-00013.safetensors +3 -0
- model-00003-of-00013.safetensors +3 -0
- model-00004-of-00013.safetensors +3 -0
- model-00005-of-00013.safetensors +3 -0
- model-00006-of-00013.safetensors +3 -0
- model-00007-of-00013.safetensors +3 -0
- model-00008-of-00013.safetensors +3 -0
- model-00009-of-00013.safetensors +3 -0
- model-00010-of-00013.safetensors +3 -0
- model-00011-of-00013.safetensors +3 -0
- model-00012-of-00013.safetensors +3 -0
- model-00013-of-00013.safetensors +3 -0
- model.safetensors.index.json +0 -0
- special_tokens_map.json +20 -0
- tokenizer.json +0 -0
- tokenizer_config.json +43 -0
- vocab.json +0 -0
@@ -0,0 +1,26 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
language:
|
3 |
+
- en
|
4 |
+
license: apache-2.0
|
5 |
+
tags:
|
6 |
+
- chat
|
7 |
+
- mlx
|
8 |
+
pipeline_tag: text-generation
|
9 |
+
---
|
10 |
+
|
11 |
+
# mlx-community/Qwen2-57B-A14B-Instruct-8bit
|
12 |
+
|
13 |
+
The Model [mlx-community/Qwen2-57B-A14B-Instruct-8bit](https://huggingface.co/mlx-community/Qwen2-57B-A14B-Instruct-8bit) was converted to MLX format from [Qwen/Qwen2-57B-A14B-Instruct](https://huggingface.co/Qwen/Qwen2-57B-A14B-Instruct) using mlx-lm version **0.14.2**.
|
14 |
+
|
15 |
+
## Use with mlx
|
16 |
+
|
17 |
+
```bash
|
18 |
+
pip install mlx-lm
|
19 |
+
```
|
20 |
+
|
21 |
+
```python
|
22 |
+
from mlx_lm import load, generate
|
23 |
+
|
24 |
+
model, tokenizer = load("mlx-community/Qwen2-57B-A14B-Instruct-8bit")
|
25 |
+
response = generate(model, tokenizer, prompt="hello", verbose=True)
|
26 |
+
```
|
@@ -0,0 +1,5 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"<|endoftext|>": 151643,
|
3 |
+
"<|im_end|>": 151645,
|
4 |
+
"<|im_start|>": 151644
|
5 |
+
}
|
@@ -0,0 +1,39 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"architectures": [
|
3 |
+
"Qwen2MoeForCausalLM"
|
4 |
+
],
|
5 |
+
"attention_dropout": 0.0,
|
6 |
+
"bos_token_id": 151643,
|
7 |
+
"decoder_sparse_step": 1,
|
8 |
+
"eos_token_id": 151643,
|
9 |
+
"hidden_act": "silu",
|
10 |
+
"hidden_size": 3584,
|
11 |
+
"initializer_range": 0.02,
|
12 |
+
"intermediate_size": 18944,
|
13 |
+
"max_position_embeddings": 32768,
|
14 |
+
"max_window_layers": 28,
|
15 |
+
"model_type": "qwen2_moe",
|
16 |
+
"moe_intermediate_size": 2560,
|
17 |
+
"norm_topk_prob": false,
|
18 |
+
"num_attention_heads": 28,
|
19 |
+
"num_experts": 64,
|
20 |
+
"num_experts_per_tok": 8,
|
21 |
+
"num_hidden_layers": 28,
|
22 |
+
"num_key_value_heads": 4,
|
23 |
+
"output_router_logits": false,
|
24 |
+
"quantization": {
|
25 |
+
"group_size": 64,
|
26 |
+
"bits": 8
|
27 |
+
},
|
28 |
+
"rms_norm_eps": 1e-06,
|
29 |
+
"rope_theta": 1000000.0,
|
30 |
+
"router_aux_loss_coef": 0.001,
|
31 |
+
"shared_expert_intermediate_size": 20480,
|
32 |
+
"sliding_window": 65536,
|
33 |
+
"tie_word_embeddings": false,
|
34 |
+
"torch_dtype": "bfloat16",
|
35 |
+
"transformers_version": "4.40.1",
|
36 |
+
"use_cache": true,
|
37 |
+
"use_sliding_window": false,
|
38 |
+
"vocab_size": 151936
|
39 |
+
}
|
The diff for this file is too large to render.
See raw diff
|
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6a207246dafd8948d7e8b1ae6a373a5f8b3be851808d2f568145d572d84919b7
|
3 |
+
size 4884308032
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:aacd596b61729d92c9b4b86e0e3fc586b82a32e31b6b3afcd71c90b5c0946926
|
3 |
+
size 4898188502
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a79d8842981d6c679666f02a978ad4bafee13fd4e8a773dfc0c506e0d550e7d0
|
3 |
+
size 4898188506
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:24a8ad4d3bd0faf530baa32352abd5386870044fdc12223b67eb7e29612e7a42
|
3 |
+
size 5163621926
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:37ff376d2cd635fbad5c17b3ee766422b207ec9fe9dd6905d016238f7e4a4cb4
|
3 |
+
size 4898188608
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a94a584c9167ceb454f2dcc102c1feaff491b4b3afa07c94f21284f03865a464
|
3 |
+
size 4898188619
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:cfd754910c6fa4aee4e8fc493b279c28eca6278818a48acfe71c09c8e9ff74d2
|
3 |
+
size 5163622045
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3efde7bc39e10cbe3bb7c41be63db20f9e4ef0d60f155109ce78d0bf2bacdd5e
|
3 |
+
size 4898188575
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b7533b1d1ee3b0a403b741fb1756d6d34c5d1a66e50d86b7942d5c25fcd62eb6
|
3 |
+
size 4898188571
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:fc1c5301aa81d77254300dcab80af42c8cb167666ca10a81540930938e4f9105
|
3 |
+
size 5163622027
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:78fca7206cf59d24a878e2e743ccbaa5e0f7eab48d68a00a9c6ed5c0b7eebefc
|
3 |
+
size 4898188591
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e5e0293cb98b76be88a571eb7430f1a69bb24a673a6d8f2d2c7a63966232389f
|
3 |
+
size 4898188561
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:244607f0a80608a9c375266a15af8fdf20495fb388396fb4ad9077764072d5d7
|
3 |
+
size 1436466320
|
The diff for this file is too large to render.
See raw diff
|
|
@@ -0,0 +1,20 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"additional_special_tokens": [
|
3 |
+
"<|im_start|>",
|
4 |
+
"<|im_end|>"
|
5 |
+
],
|
6 |
+
"eos_token": {
|
7 |
+
"content": "<|im_end|>",
|
8 |
+
"lstrip": false,
|
9 |
+
"normalized": false,
|
10 |
+
"rstrip": false,
|
11 |
+
"single_word": false
|
12 |
+
},
|
13 |
+
"pad_token": {
|
14 |
+
"content": "<|endoftext|>",
|
15 |
+
"lstrip": false,
|
16 |
+
"normalized": false,
|
17 |
+
"rstrip": false,
|
18 |
+
"single_word": false
|
19 |
+
}
|
20 |
+
}
|
The diff for this file is too large to render.
See raw diff
|
|
@@ -0,0 +1,43 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"add_prefix_space": false,
|
3 |
+
"added_tokens_decoder": {
|
4 |
+
"151643": {
|
5 |
+
"content": "<|endoftext|>",
|
6 |
+
"lstrip": false,
|
7 |
+
"normalized": false,
|
8 |
+
"rstrip": false,
|
9 |
+
"single_word": false,
|
10 |
+
"special": true
|
11 |
+
},
|
12 |
+
"151644": {
|
13 |
+
"content": "<|im_start|>",
|
14 |
+
"lstrip": false,
|
15 |
+
"normalized": false,
|
16 |
+
"rstrip": false,
|
17 |
+
"single_word": false,
|
18 |
+
"special": true
|
19 |
+
},
|
20 |
+
"151645": {
|
21 |
+
"content": "<|im_end|>",
|
22 |
+
"lstrip": false,
|
23 |
+
"normalized": false,
|
24 |
+
"rstrip": false,
|
25 |
+
"single_word": false,
|
26 |
+
"special": true
|
27 |
+
}
|
28 |
+
},
|
29 |
+
"additional_special_tokens": [
|
30 |
+
"<|im_start|>",
|
31 |
+
"<|im_end|>"
|
32 |
+
],
|
33 |
+
"bos_token": null,
|
34 |
+
"chat_template": "{% for message in messages %}{% if loop.first and messages[0]['role'] != 'system' %}{{ '<|im_start|>system\nYou are a helpful assistant.<|im_end|>\n' }}{% endif %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}",
|
35 |
+
"clean_up_tokenization_spaces": false,
|
36 |
+
"eos_token": "<|im_end|>",
|
37 |
+
"errors": "replace",
|
38 |
+
"model_max_length": 65536,
|
39 |
+
"pad_token": "<|endoftext|>",
|
40 |
+
"split_special_tokens": false,
|
41 |
+
"tokenizer_class": "Qwen2Tokenizer",
|
42 |
+
"unk_token": null
|
43 |
+
}
|
The diff for this file is too large to render.
See raw diff
|
|