File size: 630 Bytes
9ab8626 bfcebe0 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 |
---
license: mit
---
# amanpreetsingh459/llama-2-7b-chat_q4_quantized_cpp
- This model contains the 4-bit quantized version of [llama2](https://github.com/facebookresearch/llama) model.
- This can be run on a local cpu system as a cpp module available at: [https://github.com/ggerganov/llama.cpp](https://github.com/ggerganov/llama.cpp)
- As for the testing, the model has been tested on `Ubuntu Linux` os with `12 GB RAM` and `core i5 processor`.
# Credits:
1. https://github.com/facebookresearch/llama
2. https://github.com/ggerganov/llama.cpp
3. https://medium.com/@karankakwani/build-and-run-llama2-llm-locally-a3b393c1570e
|