Upload modelfiles
Browse files- .gitattributes +1 -0
- Modelfile +18 -0
- README.md +66 -0
- YugoGPT-7B-Instruct-F16.gguf +3 -0
- YugoGPT-7B-Instruct-Q3_K_L.gguf +3 -0
- YugoGPT-7B-Instruct-Q3_K_M.gguf +3 -0
- YugoGPT-7B-Instruct-Q3_K_S.gguf +3 -0
- YugoGPT-7B-Instruct-Q4_K_M.gguf +3 -0
- YugoGPT-7B-Instruct-Q4_K_S.gguf +3 -0
- YugoGPT-7B-Instruct-Q5_K_M.gguf +3 -0
- YugoGPT-7B-Instruct-Q5_K_S.gguf +3 -0
- YugoGPT-7B-Instruct-Q6_K.gguf +3 -0
- YugoGPT-7B-Instruct-Q8_0.gguf +3 -0
.gitattributes
CHANGED
@@ -1,3 +1,4 @@
|
|
|
|
1 |
*.7z filter=lfs diff=lfs merge=lfs -text
|
2 |
*.arrow filter=lfs diff=lfs merge=lfs -text
|
3 |
*.bin filter=lfs diff=lfs merge=lfs -text
|
|
|
1 |
+
*.gguf filter=lfs diff=lfs merge=lfs -text
|
2 |
*.7z filter=lfs diff=lfs merge=lfs -text
|
3 |
*.arrow filter=lfs diff=lfs merge=lfs -text
|
4 |
*.bin filter=lfs diff=lfs merge=lfs -text
|
Modelfile
ADDED
@@ -0,0 +1,18 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
FROM {__FILE_LOCATION__}
|
2 |
+
|
3 |
+
TEMPLATE """Below are some instructions that describe some tasks. Write responses that appropriately complete each request.{{ if .Prompt }}
|
4 |
+
|
5 |
+
### Instruction:
|
6 |
+
{{ .Prompt }}{{ end }}
|
7 |
+
|
8 |
+
### Response:
|
9 |
+
{{ .Response }}</s>"""
|
10 |
+
|
11 |
+
PARAMETER stop "<unk>"
|
12 |
+
PARAMETER stop "</s>"
|
13 |
+
PARAMETER temperature 0.5
|
14 |
+
PARAMETER min_p 0.05
|
15 |
+
PARAMETER top_p 0.9
|
16 |
+
PARAMETER top_k 40
|
17 |
+
PARAMETER repeat_penalty 1.1
|
18 |
+
PARAMETER num_ctx 8192
|
README.md
CHANGED
@@ -1,3 +1,69 @@
|
|
1 |
---
|
2 |
license: apache-2.0
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
3 |
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
---
|
2 |
license: apache-2.0
|
3 |
+
datasets:
|
4 |
+
- vicgalle/alpaca-gpt4
|
5 |
+
language:
|
6 |
+
- en
|
7 |
+
- sr
|
8 |
+
- bs
|
9 |
+
- hr
|
10 |
+
base_model:
|
11 |
+
- gordicaleksa/YugoGPT
|
12 |
---
|
13 |
+
# YugoGPT Instruct
|
14 |
+
|
15 |
+
**YugoGPT Instruct** is a fine-tuned version of the YugoGPT base model designed specifically for translation tasks involving Serbian, Croatian, and Bosnian languages. Unlike the base model, this instruct model is optimized for following user instructions, offering improved performance in instruction-based interactions.
|
16 |
+
|
17 |
+
---
|
18 |
+
|
19 |
+
## Overview
|
20 |
+
|
21 |
+
YugoGPT Instruct builds upon the powerful capabilities of the YugoGPT base model, fine-tuning it to enhance its usability in structured and directive tasks. This model is ideal for translation workflows where accuracy and context preservation are critical.
|
22 |
+
|
23 |
+
---
|
24 |
+
|
25 |
+
## Features
|
26 |
+
|
27 |
+
- **Specialized for BCS Languages**: Tailored for Serbian, Croatian, and Bosnian language translations.
|
28 |
+
- **Instruction Following**: Fine-tuned to better adhere to user-provided instructions.
|
29 |
+
- **Flexible Deployment**: Compatible with various quantization formats for different computational environments.
|
30 |
+
|
31 |
+
---
|
32 |
+
|
33 |
+
## Quantization Formats
|
34 |
+
|
35 |
+
A variety of quantization formats are available to suit diverse performance and resource requirements. Below is the table of quantization options:
|
36 |
+
|
37 |
+
| Filename | Quant Type | Description |
|
38 |
+
|---------------------------|------------|--------------------------------------------|
|
39 |
+
| `YugoGPT-7B-Instruct-F16` | F16 | Full F16 precision, maximum quality. |
|
40 |
+
| `YugoGPT-7B-Instruct-Q8_0` | Q8_0 | Extremely high quality. |
|
41 |
+
| `YugoGPT-7B-Instruct-Q6_K` | Q6_K | Very high quality, near perfect, recommended. |
|
42 |
+
| `YugoGPT-7B-Instruct-Q5_K_M` | Q5_K_M | High quality, recommended. |
|
43 |
+
| `YugoGPT-7B-Instruct-Q5_K_S` | Q5_K_S | High quality with optimal trade-offs. |
|
44 |
+
| `YugoGPT-7B-Instruct-Q4_K_M` | Q4_K_M | Good quality, optimized for speed. |
|
45 |
+
| `YugoGPT-7B-Instruct-Q4_K_S` | Q4_K_S | Slightly lower quality with more savings. |
|
46 |
+
| `YugoGPT-7B-Instruct-Q3_K_L` | Q3_K_L | Lower quality, good for low RAM systems. |
|
47 |
+
| `YugoGPT-7B-Instruct-Q3_K_M` | Q3_K_M | Low quality, optimized for size. |
|
48 |
+
| `YugoGPT-7B-Instruct-Q3_K_S` | Q3_K_S | Low quality, not recommended. |
|
49 |
+
|
50 |
+
---
|
51 |
+
|
52 |
+
## Usage
|
53 |
+
|
54 |
+
For usage with Ollama, you can initialize the model using the provided `modelfile` in the repository. Follow Ollama’s setup instructions to get started.
|
55 |
+
Replace `{__FILE_LOCATION__}` with the file name of the quant you want to use when creating the model using Ollama CLI.
|
56 |
+
---
|
57 |
+
|
58 |
+
## Licensing
|
59 |
+
|
60 |
+
This model is released under the [Apache 2.0 License](https://www.apache.org/licenses/LICENSE-2.0), the same as the YugoGPT base repository.
|
61 |
+
|
62 |
+
---
|
63 |
+
|
64 |
+
## Credits
|
65 |
+
|
66 |
+
- **Base Model**: [YugoGPT by Aleksa Gordić](https://huggingface.co/gordicaleksa/YugoGPT)
|
67 |
+
- **Fine-Tuning Framework**: [Unsloth](https://github.com/unslothai/unsloth)
|
68 |
+
|
69 |
+
---
|
YugoGPT-7B-Instruct-F16.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:59792f7f3af24fc4f25940004e37f811d85686a0dbf2094b18dbae2db14a83a4
|
3 |
+
size 14484732512
|
YugoGPT-7B-Instruct-Q3_K_L.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:60ecf5f3fa12efaa14f75764f55b542446d8b8f70d18cbe0ad22a45a12c8fd3e
|
3 |
+
size 3822025312
|
YugoGPT-7B-Instruct-Q3_K_M.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:1e93bdad27d6a97838859ca0a436415da3d937bff5bddc6ff89d0139b917d97c
|
3 |
+
size 3518986848
|
YugoGPT-7B-Instruct-Q3_K_S.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3382c0caaceee6bc2fd5b140a808db6867405837d5ab393361cbf357fa859222
|
3 |
+
size 3164568160
|
YugoGPT-7B-Instruct-Q4_K_M.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:681d20d3b116a57c93dc2aaf48e368fbfe3f0d0473012254b223e75688883354
|
3 |
+
size 4368439904
|
YugoGPT-7B-Instruct-Q4_K_S.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:1b6ae21d0c91cfe4c2566cefba6612c89d9a469c97ad725d3c387ae1e5c12058
|
3 |
+
size 4140374624
|
YugoGPT-7B-Instruct-Q5_K_M.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:52f73486bc15bcd062fc6fc7b91be02234c779f6f60de19be8773c8237494b79
|
3 |
+
size 5131410016
|
YugoGPT-7B-Instruct-Q5_K_S.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a590f805b00fb5f39c1c530917f663bf1592d5ca997da426b337d8f76989afda
|
3 |
+
size 4997716576
|
YugoGPT-7B-Instruct-Q6_K.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b39962798dfb834c77e5a1fb96cb197381c2201ebbbce9af3ab55aa5cc218860
|
3 |
+
size 5942065760
|
YugoGPT-7B-Instruct-Q8_0.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f49377bd9f6f71f9ffcfca6acc8df4d1ff30adfbcd9b57e814be8d5132a1aeb5
|
3 |
+
size 7695858272
|