Ollama/llama.cpp support

#1
by dpreti - opened

We observed that at the moment the model is not served correctly with Ollama and llama.cpp .

We are currently investigating the reasons behind this unexpected behavior.
In the meanwhile we strongly suggest to serve the model using vLLM or the Transformer library as showed in the model card.

dpreti changed discussion title from Ollama support to Ollama/llama.cpp support

Sign up or log in to comment