File size: 680 Bytes
a1a537a
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
---



base_model: Nelathan/Qwen2-7B-FocusMix
language:
- en
pipeline_tag: text-generation
quantized_by: Nelathan
library_name: transformers
tags:
- mergekit
- merge

---

<img src="./focusmix.jpg" alt="FocusMix 7B" width="100%" display="block"/>

# FocusMix 7B GGUF

Using <a href="https://github.com/ggerganov/llama.cpp/">llama.cpp</a> release <a href="https://github.com/ggerganov/llama.cpp/releases/tag/b3557">b3557</a> for static quantization.

Original model: https://huggingface.co/Nelathan/Qwen2-7B-FocusMix

### ChatML

```

<|im_start|>system

{system}<|im_end|>

<|im_start|>user

{user}<|im_end|>

<|im_start|>assistant

{asistant}<|im_end|>

```