runtime error
llama.cpp: loading model from ggml-alpaca-7b-q4.bin llama.cpp: can't use mmap because tensors are not aligned; convert to new format to avoid this llama_model_load_internal: format = 'ggml' (old version with low tokenizer quality and no mmap support) llama_model_load_internal: n_vocab = 32000 llama_model_load_internal: n_ctx = 256 llama_model_load_internal: n_embd = 4096 llama_model_load_internal: n_mult = 256 llama_model_load_internal: n_head = 32 llama_model_load_internal: n_layer = 32 llama_model_load_internal: n_rot = 128 llama_model_load_internal: ftype = 2 (mostly Q4_0) llama_model_load_internal: n_ff = 11008 llama_model_load_internal: n_parts = 1 llama_model_load_internal: model size = 7B llama_model_load_internal: ggml ctx size = 4113739.11 KB llama_model_load_internal: mem required = 5809.32 MB (+ 1026.00 MB per state) ................................................................................................... . llama_init_from_file: kv self size = 128.00 MB AVX = 1 | AVX2 = 1 | AVX512 = 1 | AVX512_VBMI = 1 | AVX512_VNNI = 1 | FMA = 1 | NEON = 0 | ARM_FMA = 0 | F16C = 1 | FP16_VA = 0 | WASM_SIMD = 0 | BLAS = 0 | SSE3 = 1 | VSX = 0 | Traceback (most recent call last): File "/home/user/app/app.py", line 10, in <module> input_text = gr.inputs.Textbox(lines= 10, label="Enter your input text") AttributeError: module 'gradio' has no attribute 'inputs' Exception ignored in: <function Llama.__del__ at 0x7fa41cd1f760> Traceback (most recent call last): File "/usr/local/lib/python3.10/site-packages/llama_cpp/llama.py", line 812, in __del__ TypeError: 'NoneType' object is not callable
Container logs:
Fetching error logs...