The Phi-4-mini-instruct-GGUF:Q4_K_M failing

#2
by pgr405 - opened

ollama run hf.co/unsloth/Phi-4-mini-instruct-GGUF:Q4_K_M
pulling manifest
pulling 88c002299140... 100%
pulling 813f53fdc6e5... 100%
pulling 534cce8916c3... 100%
verifying sha256 digest
writing manifest
success
Error: llama runner process has terminated: error loading model: missing tensor 'output.weight'

I just tried using Phi-4-mini-instruct-Q4_K_M.gguf with a fresh pull from llama.cpp and that works just fine.

Unsloth AI org

ollama run hf.co/unsloth/Phi-4-mini-instruct-GGUF:Q4_K_M
pulling manifest
pulling 88c002299140... 100%
pulling 813f53fdc6e5... 100%
pulling 534cce8916c3... 100%
verifying sha256 digest
writing manifest
success
Error: llama runner process has terminated: error loading model: missing tensor 'output.weight'

please update ollama

ollama run hf.co/unsloth/Phi-4-mini-instruct-GGUF:Q4_K_M
pulling manifest
pulling 88c002299140... 100%
pulling 813f53fdc6e5... 100%
pulling 534cce8916c3... 100%
verifying sha256 digest
writing manifest
success
Error: llama runner process has terminated: error loading model: missing tensor 'output.weight'

Using Q2_K_L also gives the same error, using Ollama 0.5.12, both in Ubuntu and Windows.

ollama run hf.co/unsloth/Phi-4-mini-instruct-GGUF:Q4_K_M
pulling manifest
pulling 88c002299140... 100%
pulling 813f53fdc6e5... 100%
pulling 534cce8916c3... 100%
verifying sha256 digest
writing manifest
success
Error: llama runner process has terminated: error loading model: missing tensor 'output.weight'

please update ollama

After updating to Ollama version 0.5.13, the model now runs without any problems (version 0.5.11 was reporting "error loading model").

Sign up or log in to comment