Update README.md
Browse files
README.md
CHANGED
@@ -47,41 +47,6 @@ This repo contains `GGUF` format model files for [aisingapore/llama3.1-8b-cpt-se
|
|
47 |
- [llama3.1-8b-cpt-sea-lionv3-instruct-Q6_K](https://huggingface.co/aisingapore/llama3.1-8b-cpt-sea-lionv3-instruct-gguf/blob/main/llama3.1-8B-cpt-sea-lionv3-instruct-Q6_K.gguf)
|
48 |
- [llama3.1-8b-cpt-sea-lionv3-instruct-Q8_0](https://huggingface.co/aisingapore/llama3.1-8b-cpt-sea-lionv3-instruct-gguf/blob/main/llama3.1-8B-cpt-sea-lionv3-instruct-Q8_0.gguf)
|
49 |
|
50 |
-
### Usage
|
51 |
-
Llama 3.1 8B CPT SEA-Lionv3 Instruct GGUF files have been tested with [llama.cpp](https://github.com/ggerganov/llama.cpp).
|
52 |
-
|
53 |
-
#### Prompt Template:
|
54 |
-
```
|
55 |
-
<|begin_of_text|><|start_header_id|>system<|end_header_id|>
|
56 |
-
|
57 |
-
{{system_prompt}}<|eot_id|>
|
58 |
-
<|start_header_id|>user<|end_header_id|>
|
59 |
-
|
60 |
-
{{prompt}}<|eot_id|>
|
61 |
-
<|start_header_id|>assistant<|end_header_id|>
|
62 |
-
```
|
63 |
-
|
64 |
-
#### Recommended `llama.cpp` command:
|
65 |
-
To execute the following commands, ensure you are in the `llama.cpp` root directory and that your models are located in the `models` folder:
|
66 |
-
```sh
|
67 |
-
# Running one-time input prompt
|
68 |
-
./llama-cli -m models/llama3.1-8b-cpt-sea-lionv3-instruct-gguf/llama3.1-8b-cpt-sea-lionv3-instruct-Q4_K_M.gguf -ngl -1 --temp 0 -n 128 -p "<|begin_of_text|><|start_header_id|>system<|end_header_id|>\n\nYou are a helpful assistant who answers succinctly.<|eot_id|>\n<|start_header_id|>user<|end_header_id|>\n\nWhat is a sea lion?<|eot_id|>\n<|start_header_id|>assistant<|end_header_id|>\n\n"
|
69 |
-
```
|
70 |
-
```sh
|
71 |
-
# Running in conversation mode
|
72 |
-
./llama-cli -m models/llama3.1-8b-cpt-sea-lionv3-instruct-gguf/llama3.1-8b-cpt-sea-lionv3-instruct-Q4_K_M.gguf -ngl -1 --temp 0 -n 128 -p "You are a helpful assistant who answers succinctly." --color -cnv --chat-template llama3
|
73 |
-
```
|
74 |
-
Please refer to [the llama.cpp documentation](https://github.com/ggerganov/llama.cpp/blob/master/examples/main/README.md) for adjusting the parameters.
|
75 |
-
|
76 |
-
#### To convert & quantize your own SEA-LION model:
|
77 |
-
Given that you are in the `llama.cpp` root directory:
|
78 |
-
```sh
|
79 |
-
python convert-hf-to-gguf.py {{model path}}
|
80 |
-
./quantize ggml-model-f16.gguf {{Quant Type}}
|
81 |
-
```
|
82 |
-
|
83 |
-
For more detailed instructions on conversion and quantization, please refer to [llama.cpp documentation](https://github.com/ggerganov/llama.cpp/blob/master/examples/quantize/README.md).
|
84 |
-
|
85 |
### Caveats
|
86 |
It is important for users to be aware that our model exhibits certain limitations that warrant consideration. Like many LLMs, the model can hallucinate and occasionally generates irrelevant content, introducing fictional elements that are not grounded in the provided context. Users should also exercise caution in interpreting and validating the model's responses due to the potential inconsistencies in its reasoning.
|
87 |
|
|
|
47 |
- [llama3.1-8b-cpt-sea-lionv3-instruct-Q6_K](https://huggingface.co/aisingapore/llama3.1-8b-cpt-sea-lionv3-instruct-gguf/blob/main/llama3.1-8B-cpt-sea-lionv3-instruct-Q6_K.gguf)
|
48 |
- [llama3.1-8b-cpt-sea-lionv3-instruct-Q8_0](https://huggingface.co/aisingapore/llama3.1-8b-cpt-sea-lionv3-instruct-gguf/blob/main/llama3.1-8B-cpt-sea-lionv3-instruct-Q8_0.gguf)
|
49 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
50 |
### Caveats
|
51 |
It is important for users to be aware that our model exhibits certain limitations that warrant consideration. Like many LLMs, the model can hallucinate and occasionally generates irrelevant content, introducing fictional elements that are not grounded in the provided context. Users should also exercise caution in interpreting and validating the model's responses due to the potential inconsistencies in its reasoning.
|
52 |
|