ae362b929321e78dac00be23203737cbe27de9ffb3e77d8c79669ad3bc6da06e
Browse files- README.md +7 -7
- quant_strategy.json +1 -1
README.md
CHANGED
@@ -1,13 +1,13 @@
|
|
1 |
-
---
|
2 |
-
base_model: GreenBitAI/Llama-3.1-Nemotron-70B-Instruct-layer-mix-bpw-4.0
|
3 |
-
license: apache-2.0
|
4 |
-
tags:
|
5 |
-
- mlx
|
6 |
-
---
|
7 |
|
8 |
# GreenBitAI/Llama-3.1-Nemotron-70B-Instruct-layer-mix-bpw-4.0-mlx
|
9 |
|
10 |
-
This quantized low-bit model [GreenBitAI/Llama-3.1-Nemotron-70B-Instruct-layer-mix-bpw-4.0-mlx](https://huggingface.co/GreenBitAI/Llama-3.1-Nemotron-70B-Instruct-layer-mix-bpw-4.0-mlx) was converted to MLX format from [`GreenBitAI/Llama-3.1-Nemotron-70B-Instruct-layer-mix-bpw-4.0
|
11 |
Refer to the [original model card](https://huggingface.co/GreenBitAI/Llama-3.1-Nemotron-70B-Instruct-layer-mix-bpw-4.0) for more details on the model.
|
12 |
|
13 |
## Use with mlx
|
|
|
1 |
+
---
|
2 |
+
base_model: GreenBitAI/Llama-3.1-Nemotron-70B-Instruct-layer-mix-bpw-4.0
|
3 |
+
license: apache-2.0
|
4 |
+
tags:
|
5 |
+
- mlx
|
6 |
+
---
|
7 |
|
8 |
# GreenBitAI/Llama-3.1-Nemotron-70B-Instruct-layer-mix-bpw-4.0-mlx
|
9 |
|
10 |
+
This quantized low-bit model [GreenBitAI/Llama-3.1-Nemotron-70B-Instruct-layer-mix-bpw-4.0-mlx](https://huggingface.co/GreenBitAI/Llama-3.1-Nemotron-70B-Instruct-layer-mix-bpw-4.0-mlx) was converted to MLX format from [`GreenBitAI/Llama-3.1-Nemotron-70B-Instruct-layer-mix-bpw-4.0`](https://huggingface.co/GreenBitAI/Llama-3.1-Nemotron-70B-Instruct-layer-mix-bpw-4.0) using gbx-lm version **0.3.4**.
|
11 |
Refer to the [original model card](https://huggingface.co/GreenBitAI/Llama-3.1-Nemotron-70B-Instruct-layer-mix-bpw-4.0) for more details on the model.
|
12 |
|
13 |
## Use with mlx
|
quant_strategy.json
CHANGED
@@ -7601,4 +7601,4 @@
|
|
7601 |
}
|
7602 |
}
|
7603 |
}
|
7604 |
-
}
|
|
|
7601 |
}
|
7602 |
}
|
7603 |
}
|
7604 |
+
}
|