tainc commited on
Commit
9c4ccbb
1 Parent(s): 91c8a7f

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +0 -35
README.md CHANGED
@@ -47,41 +47,6 @@ This repo contains `GGUF` format model files for [aisingapore/llama3.1-8b-cpt-se
47
  - [llama3.1-8b-cpt-sea-lionv3-instruct-Q6_K](https://huggingface.co/aisingapore/llama3.1-8b-cpt-sea-lionv3-instruct-gguf/blob/main/llama3.1-8B-cpt-sea-lionv3-instruct-Q6_K.gguf)
48
  - [llama3.1-8b-cpt-sea-lionv3-instruct-Q8_0](https://huggingface.co/aisingapore/llama3.1-8b-cpt-sea-lionv3-instruct-gguf/blob/main/llama3.1-8B-cpt-sea-lionv3-instruct-Q8_0.gguf)
49
 
50
- ### Usage
51
- Llama 3.1 8B CPT SEA-Lionv3 Instruct GGUF files have been tested with [llama.cpp](https://github.com/ggerganov/llama.cpp).
52
-
53
- #### Prompt Template:
54
- ```
55
- <|begin_of_text|><|start_header_id|>system<|end_header_id|>
56
-
57
- {{system_prompt}}<|eot_id|>
58
- <|start_header_id|>user<|end_header_id|>
59
-
60
- {{prompt}}<|eot_id|>
61
- <|start_header_id|>assistant<|end_header_id|>
62
- ```
63
-
64
- #### Recommended `llama.cpp` command:
65
- To execute the following commands, ensure you are in the `llama.cpp` root directory and that your models are located in the `models` folder:
66
- ```sh
67
- # Running one-time input prompt
68
- ./llama-cli -m models/llama3.1-8b-cpt-sea-lionv3-instruct-gguf/llama3.1-8b-cpt-sea-lionv3-instruct-Q4_K_M.gguf -ngl -1 --temp 0 -n 128 -p "<|begin_of_text|><|start_header_id|>system<|end_header_id|>\n\nYou are a helpful assistant who answers succinctly.<|eot_id|>\n<|start_header_id|>user<|end_header_id|>\n\nWhat is a sea lion?<|eot_id|>\n<|start_header_id|>assistant<|end_header_id|>\n\n"
69
- ```
70
- ```sh
71
- # Running in conversation mode
72
- ./llama-cli -m models/llama3.1-8b-cpt-sea-lionv3-instruct-gguf/llama3.1-8b-cpt-sea-lionv3-instruct-Q4_K_M.gguf -ngl -1 --temp 0 -n 128 -p "You are a helpful assistant who answers succinctly." --color -cnv --chat-template llama3
73
- ```
74
- Please refer to [the llama.cpp documentation](https://github.com/ggerganov/llama.cpp/blob/master/examples/main/README.md) for adjusting the parameters.
75
-
76
- #### To convert & quantize your own SEA-LION model:
77
- Given that you are in the `llama.cpp` root directory:
78
- ```sh
79
- python convert-hf-to-gguf.py {{model path}}
80
- ./quantize ggml-model-f16.gguf {{Quant Type}}
81
- ```
82
-
83
- For more detailed instructions on conversion and quantization, please refer to [llama.cpp documentation](https://github.com/ggerganov/llama.cpp/blob/master/examples/quantize/README.md).
84
-
85
  ### Caveats
86
  It is important for users to be aware that our model exhibits certain limitations that warrant consideration. Like many LLMs, the model can hallucinate and occasionally generates irrelevant content, introducing fictional elements that are not grounded in the provided context. Users should also exercise caution in interpreting and validating the model's responses due to the potential inconsistencies in its reasoning.
87
 
 
47
  - [llama3.1-8b-cpt-sea-lionv3-instruct-Q6_K](https://huggingface.co/aisingapore/llama3.1-8b-cpt-sea-lionv3-instruct-gguf/blob/main/llama3.1-8B-cpt-sea-lionv3-instruct-Q6_K.gguf)
48
  - [llama3.1-8b-cpt-sea-lionv3-instruct-Q8_0](https://huggingface.co/aisingapore/llama3.1-8b-cpt-sea-lionv3-instruct-gguf/blob/main/llama3.1-8B-cpt-sea-lionv3-instruct-Q8_0.gguf)
49
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
50
  ### Caveats
51
  It is important for users to be aware that our model exhibits certain limitations that warrant consideration. Like many LLMs, the model can hallucinate and occasionally generates irrelevant content, introducing fictional elements that are not grounded in the provided context. Users should also exercise caution in interpreting and validating the model's responses due to the potential inconsistencies in its reasoning.
52