cgus commited on
Commit
0ecfc49
·
verified ·
1 Parent(s): eed6429

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +18 -2
README.md CHANGED
@@ -5,13 +5,29 @@ license_link: https://huggingface.co/huihui-ai/Qwen2.5-14B-Instruct-abliterated/
5
  language:
6
  - en
7
  pipeline_tag: text-generation
8
- base_model: Qwen/Qwen2.5-14B-Instruct
9
  tags:
10
  - chat
11
  - abliterated
12
  - uncensored
13
  ---
14
-
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
15
  # huihui-ai/Qwen2.5-14B-Instruct-abliterated
16
 
17
 
 
5
  language:
6
  - en
7
  pipeline_tag: text-generation
8
+ base_model: huihui-ai/Qwen2.5-14B-Instruct-abliterated
9
  tags:
10
  - chat
11
  - abliterated
12
  - uncensored
13
  ---
14
+ # Qwen2.5-14B-Instruct-abliterated-exl2
15
+ Model: [Qwen2.5-14B-Instruct-abliterated](https://huggingface.co/huihui-ai/Qwen2.5-14B-Instruct-abliterated)
16
+ Made by: [huihui-ai](https://huggingface.co/huihui-ai)
17
+
18
+ ## Quants
19
+ [4bpw h6 (main)](https://huggingface.co/cgus/Qwen2.5-14B-Instruct-abliterated-exl2/tree/main)
20
+ [4.5bpw h6](https://huggingface.co/cgus/Qwen2.5-14B-Instruct-abliterated-exl2/tree/4.5bpw-h6)
21
+ [5bpw h6](https://huggingface.co/cgus/Qwen2.5-14B-Instruct-abliterated-exl2/tree/5bpw-h6)
22
+ [6bpw h6](https://huggingface.co/cgus/Qwen2.5-14B-Instruct-abliterated-exl2/tree/6bpw-h6)
23
+ [8bpw h8](https://huggingface.co/cgus/Qwen2.5-14B-Instruct-abliterated-exl2/tree/8bpw-h8)
24
+
25
+ ## Quantization notes
26
+ Made with Exllamav2 0.2.3 with the default dataset. These quants are meant for modern RTX cards on Windows/Linux or AMD on Linux.
27
+ The model have to fit the GPU to work properly. For example RTX3060/12GB should be able to load 4bpw with 16k context.
28
+ It requires an app with Exllamav2 loader, such as Text-Generation-WebUI, TabbyAPI and some others.
29
+
30
+ # Original model card
31
  # huihui-ai/Qwen2.5-14B-Instruct-abliterated
32
 
33