amanpreetsingh459's picture
Add credits to the README.md file
bfcebe0
|
raw
history blame
630 Bytes
metadata
license: mit

amanpreetsingh459/llama-2-7b-chat_q4_quantized_cpp

  • This model contains the 4-bit quantized version of llama2 model.
  • This can be run on a local cpu system as a cpp module available at: https://github.com/ggerganov/llama.cpp
  • As for the testing, the model has been tested on Ubuntu Linux os with 12 GB RAM and core i5 processor.

Credits:

  1. https://github.com/facebookresearch/llama
  2. https://github.com/ggerganov/llama.cpp
  3. https://medium.com/@karankakwani/build-and-run-llama2-llm-locally-a3b393c1570e