Could not load Llama model from path
For some reason I can't load the llama model from the correct file path? its in my downloads but says there is a (type=value_error)
gguf_init_from_file: invalid magic characters 'tjgg'
error loading model: llama_model_loader: failed to load model from /Users/christopherlozoya/Downloads/llama-2-7b-chat.ggmlv3.q8_0.bin
llama_load_model_from_file: failed to load model
AVX = 0 | AVX_VNNI = 0 | AVX2 = 0 | AVX512 = 0 | AVX512_VBMI = 0 | AVX512_VNNI = 0 | FMA = 0 | NEON = 1 | ARM_FMA = 1 | F16C = 0 | FP16_VA = 1 | WASM_SIMD = 0 | BLAS = 1 | SSE3 = 0 | SSSE3 = 0 | VSX = 0 |
Traceback (most recent call last):
File "/Users/christopherlozoya/PycharmProjects/TestOllama/.venv/mainllama.py", line 26, in
llm = load_model()
^^^^^^^^^^^^
File "/Users/christopherlozoya/PycharmProjects/TestOllama/.venv/mainllama.py", line 15, in load_model
Llama_model = LlamaCpp(
^^^^^^^^^
File "/Users/christopherlozoya/anaconda3/lib/python3.11/site-packages/langchain_core/load/serializable.py", line 107, in init
super().init(**kwargs)
File "pydantic/main.py", line 341, in pydantic.main.BaseModel.init
pydantic.error_wrappers.ValidationError: 1 validation error for LlamaCpp
root
Could not load Llama model from path: /Users/christopherlozoya/Downloads/llama-2-7b-chat.ggmlv3.q8_0.bin. Received error (type=value_error)
@Lozzoya
This is due to the recent update to GGUF-format. Older model formats such as ggmlv3 are no longer supported starting August 2023.
You can find new ones in GGUF-format here: https://huggingface.co/TheBloke/Llama-2-7B-Chat-GGUF