Triangle104
commited on
Commit
•
7234701
1
Parent(s):
0958ee1
Update README.md
Browse files
README.md
CHANGED
@@ -118,6 +118,40 @@ model-index:
|
|
118 |
This model was converted to GGUF format from [`anthracite-org/magnum-v4-22b`](https://huggingface.co/anthracite-org/magnum-v4-22b) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
|
119 |
Refer to the [original model card](https://huggingface.co/anthracite-org/magnum-v4-22b) for more details on the model.
|
120 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
121 |
## Use with llama.cpp
|
122 |
Install llama.cpp through brew (works on Mac and Linux)
|
123 |
|
|
|
118 |
This model was converted to GGUF format from [`anthracite-org/magnum-v4-22b`](https://huggingface.co/anthracite-org/magnum-v4-22b) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
|
119 |
Refer to the [original model card](https://huggingface.co/anthracite-org/magnum-v4-22b) for more details on the model.
|
120 |
|
121 |
+
---
|
122 |
+
Model details:
|
123 |
+
-
|
124 |
+
This is a series of models designed to replicate the prose quality of the Claude 3 models, specifically Sonnet and Opus.
|
125 |
+
|
126 |
+
This model is fine-tuned on top of Mistral-Small-Instruct-2409.
|
127 |
+
|
128 |
+
Prompting
|
129 |
+
-
|
130 |
+
A typical input would look like this:
|
131 |
+
|
132 |
+
<s>[INST] SYSTEM MESSAGE
|
133 |
+
USER MESSAGE[/INST] ASSISTANT MESSAGE</s>[INST] USER MESSAGE[/INST]
|
134 |
+
|
135 |
+
Credits
|
136 |
+
-
|
137 |
+
We'd like to thank Recursal / Featherless for sponsoring the compute for this train, Featherless has been hosting our Magnum models since the first 72 B and has given thousands of people access to our models and helped us grow.
|
138 |
+
|
139 |
+
We would also like to thank all members of Anthracite who made this finetune possible.
|
140 |
+
|
141 |
+
Datasets
|
142 |
+
-
|
143 |
+
anthracite-org/c2_logs_32k_mistral-v3_v1.2_no_system
|
144 |
+
anthracite-org/kalo-opus-instruct-22k-no-refusal-no-system
|
145 |
+
anthracite-org/kalo-opus-instruct-3k-filtered-no-system
|
146 |
+
anthracite-org/nopm_claude_writing_fixed
|
147 |
+
anthracite-org/kalo_opus_misc_240827_no_system
|
148 |
+
anthracite-org/kalo_misc_part2_no_system
|
149 |
+
|
150 |
+
Training
|
151 |
+
-
|
152 |
+
The training was done for 2 epochs. We used 8xH100s GPUs graciously provided by Recursal AI / Featherless AI for the full-parameter fine-tuning of the model.
|
153 |
+
|
154 |
+
---
|
155 |
## Use with llama.cpp
|
156 |
Install llama.cpp through brew (works on Mac and Linux)
|
157 |
|