Update README.md
Browse files
README.md
CHANGED
@@ -40,7 +40,7 @@ apt-get install build-essential cmake curl libcurl4-openssl-dev -y
|
|
40 |
git clone https://github.com/ggerganov/llama.cpp
|
41 |
cmake llama.cpp -B llama.cpp/build \
|
42 |
-DBUILD_SHARED_LIBS=OFF -DGGML_CUDA=ON -DLLAMA_CURL=ON
|
43 |
-
cmake --build llama.cpp/build --config Release -j --clean-first --target llama-quantize llama-cli
|
44 |
cp llama.cpp/build/bin/llama-* llama.cpp
|
45 |
```
|
46 |
3. It's best to use `--min-p 0.05` to counteract very rare token predictions - I found this to work well especially for the 1.58bit model.
|
|
|
40 |
git clone https://github.com/ggerganov/llama.cpp
|
41 |
cmake llama.cpp -B llama.cpp/build \
|
42 |
-DBUILD_SHARED_LIBS=OFF -DGGML_CUDA=ON -DLLAMA_CURL=ON
|
43 |
+
cmake --build llama.cpp/build --config Release -j --clean-first --target llama-quantize llama-cli llama-gguf-split
|
44 |
cp llama.cpp/build/bin/llama-* llama.cpp
|
45 |
```
|
46 |
3. It's best to use `--min-p 0.05` to counteract very rare token predictions - I found this to work well especially for the 1.58bit model.
|