ilhooq's picture
Add more tokens in context
e9669da
raw
history blame
483 Bytes
FROM ghcr.io/ggerganov/llama.cpp:server-cuda
RUN apt update && apt install -y curl
RUN mkdir /models
RUN curl -L https://huggingface.co/TheBloke/deepseek-coder-6.7B-instruct-GGUF/resolve/main/deepseek-coder-6.7b-instruct.Q6_K.gguf --output /models/deepseek-coder-6.7b-instruct.Q6_K.gguf
ENTRYPOINT [ "/server" ]
CMD [ "-ngl", "32", "--host", "0.0.0.0", "--port", "7860", "--model", "/models/deepseek-coder-6.7b-instruct.Q6_K.gguf", "-c", "8192", "--chat-template", "deepseek" ]