File size: 1,573 Bytes
9ab8626 ba7f279 d36c2a3 e6c29f4 9ab8626 bfcebe0 33c304c ae9ad79 2ee211e 33c304c ba1963f 33c304c bfcebe0 d36c2a3 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 |
---
license: llama2
tags:
- llama2
- text-generation-inference
base_midel: meta-llama/Llama-2-7b-chat-hf
---
# llama-2-7b-chat_q4_quantized_cpp
- This model contains the 4-bit quantized version of [llama2-7B-chat](https://github.com/facebookresearch/llama) model in cpp.
- This can be run on a local cpu system as a cpp module *(instructions for the same are given below)*.
- As for the testing, the model has been tested on `Linux(Ubuntu)` os with `12 GB RAM` and `core i5 processor`.
- The performance is `roughly` **~3 tokens per second**
# Usage:
1. Clone the llama C++ repository from github:<br>
`git clone https://github.com/ggerganov/llama.cpp.git`
2. Enter the **llama.cpp** repository(which was downloaded in the step 1) and build it by running the **make** command<br>
`cd llama.cpp` <br>
`make`
3. Create a directory names **7B** under the directory **llama.cpp/models** and put the model file **ggml-model-q4_0.bin** under this newly created **7B** directory<br>
`cd models` <br>
`mkdir 7B`
4. Navigate back to **llama.cpp** directory and run the below command:-<br>
`./main -m ./models/7B/ggml-model-q4_0.bin -n 1024 --repeat_penalty 1.0 --color -i -r "User:" -f ./prompts/alpaca.txt` <br>
> the initial prompt file can be changed to anything from `prompts/alpaca.txt` to of your choice
5. That's it. Enter the desired prompts and let the results surprise you...
# Credits:
1. https://github.com/facebookresearch/llama
2. https://github.com/ggerganov/llama.cpp
3. https://medium.com/@karankakwani/build-and-run-llama2-llm-locally-a3b393c1570e |