npc0 commited on
Commit
0d01c4c
1 Parent(s): 9be5029

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +18 -1
README.md CHANGED
@@ -1,11 +1,28 @@
1
  ---
2
  license: apache-2.0
 
 
 
 
 
 
 
 
 
3
  ---
4
 
5
  This repository contains the quantized DISC-MedLLM, version of Baichuan-13b-base as the base model.
6
 
7
  The weights are converted to GGML format using [baichuan13b.cpp](https://github.com/ouwei2013/baichuan13b.cpp) (based on [llama.cpp](https://github.com/ggerganov/llama.cpp))
8
 
 
 
 
 
 
 
 
 
9
  ## How to inference
10
  1. [Compile baichuan13b](https://github.com/ouwei2013/baichuan13b.cpp#build), a main executable `baichuan13b/build/bin/main` and a server `baichuan13b/build/bin/server` will be generated.
11
  2. Download the weight in this repository to `baichuan13b/build/bin/`
@@ -41,4 +58,4 @@ llm_output = requests.post(
41
  "n_predict": 512
42
  }).json()
43
  print(llm_output)
44
- ```
 
1
  ---
2
  license: apache-2.0
3
+ datasets:
4
+ - Flmc/DISC-Med-SFT
5
+ language:
6
+ - zh
7
+ pipeline_tag: text-generation
8
+ tags:
9
+ - baichuan
10
+ - medical
11
+ - ggml
12
  ---
13
 
14
  This repository contains the quantized DISC-MedLLM, version of Baichuan-13b-base as the base model.
15
 
16
  The weights are converted to GGML format using [baichuan13b.cpp](https://github.com/ouwei2013/baichuan13b.cpp) (based on [llama.cpp](https://github.com/ggerganov/llama.cpp))
17
 
18
+ |Model |GGML quantize method| HDD size |
19
+ |--------------------|--------------------|----------|
20
+ |ggml-model-q4_0.bin | q4_0 | 7.55 GB |
21
+ |ggml-model-q4_1.bin | q4_1 | 8.36 GB |
22
+ |ggml-model-q5_0.bin | q5_0 | 9.17 GB |
23
+ |ggml-model-q5_1.bin | q5_1 | 9.97 GB |
24
+ <!-- |ggml-model-q8_0.bin | q8_0 | ?.?? GB | -->
25
+
26
  ## How to inference
27
  1. [Compile baichuan13b](https://github.com/ouwei2013/baichuan13b.cpp#build), a main executable `baichuan13b/build/bin/main` and a server `baichuan13b/build/bin/server` will be generated.
28
  2. Download the weight in this repository to `baichuan13b/build/bin/`
 
58
  "n_predict": 512
59
  }).json()
60
  print(llm_output)
61
+ ```