--- license: mit --- # amanpreetsingh459/llama-2-7b-chat_q4_quantized_cpp - This model contains the 4-bit quantized version of [llama2](https://github.com/facebookresearch/llama) model. - This can be run on a local cpu system as a cpp module available at: [https://github.com/ggerganov/llama.cpp](https://github.com/ggerganov/llama.cpp) - As for the testing, the model has been tested on `Ubuntu Linux` os with `12 GB RAM` and `core i5 processor`. # Credits: 1. https://github.com/facebookresearch/llama 2. https://github.com/ggerganov/llama.cpp 3. https://medium.com/@karankakwani/build-and-run-llama2-llm-locally-a3b393c1570e