The Phi-4-mini-instruct-GGUF:Q4_K_M failing
ollama run hf.co/unsloth/Phi-4-mini-instruct-GGUF:Q4_K_M
pulling manifest
pulling 88c002299140... 100%
pulling 813f53fdc6e5... 100%
pulling 534cce8916c3... 100%
verifying sha256 digest
writing manifest
success
Error: llama runner process has terminated: error loading model: missing tensor 'output.weight'
I just tried using Phi-4-mini-instruct-Q4_K_M.gguf with a fresh pull from llama.cpp and that works just fine.
ollama run hf.co/unsloth/Phi-4-mini-instruct-GGUF:Q4_K_M
pulling manifest
pulling 88c002299140... 100%
pulling 813f53fdc6e5... 100%
pulling 534cce8916c3... 100%
verifying sha256 digest
writing manifest
success
Error: llama runner process has terminated: error loading model: missing tensor 'output.weight'
please update ollama
ollama run hf.co/unsloth/Phi-4-mini-instruct-GGUF:Q4_K_M
pulling manifest
pulling 88c002299140... 100%
pulling 813f53fdc6e5... 100%
pulling 534cce8916c3... 100%
verifying sha256 digest
writing manifest
success
Error: llama runner process has terminated: error loading model: missing tensor 'output.weight'
Using Q2_K_L also gives the same error, using Ollama 0.5.12, both in Ubuntu and Windows.
ollama run hf.co/unsloth/Phi-4-mini-instruct-GGUF:Q4_K_M
pulling manifest
pulling 88c002299140... 100%
pulling 813f53fdc6e5... 100%
pulling 534cce8916c3... 100%
verifying sha256 digest
writing manifest
success
Error: llama runner process has terminated: error loading model: missing tensor 'output.weight'please update ollama
After updating to Ollama version 0.5.13, the model now runs without any problems (version 0.5.11 was reporting "error loading model").