gguf fail to be loaded on ollama and LM Studio

#21
by JackeyBee - opened

When I try to load Phi-3-mini-4k-instruct-q4.gguf, it says "llama.cpp error: 'error loading model hyperparameters: key not found in model: phi3.attention.sliding_window'". I have spotted the issue created on ollama GH repository too: link

I'm seeing something similar with loading via llama-cpp-python. With another "'Llama' object has no attribute '_lora_adapter'" message. This has happened only with the latest llama-cpp-python release.

See https://github.com/ggerganov/llama.cpp/pull/8627

Updated gguf are required since this recent change.

Sign up or log in to comment