YAML Metadata Warning: empty or missing yaml metadata in repo card (https://huggingface.co/docs/hub/model-cards#model-card-metadata)

GGUF format files of the model vinai/PhoGPT-4B-Chat.

This model file is compatible with the latest llama.cpp

Context: I was trying to get PhoGPT to work with llama-cpp and llama-cpp-python. I found nguyenviet/PhoGPT-4B-Chat-GGUF but cannot get it to work:

from llama_cpp import Llama

llm = Llama.from_pretrained(
    repo_id="nguyenviet/PhoGPT-4B-Chat-GGUF",
    filename="*q3_k_m.gguf*",
)

...
llama_model_load: error loading model: done_getting_tensors: wrong number of tensors; expected 388, got 387
llama_load_model_from_file: failed to load model
...

After my opening issue at the PhoGPT repo was resolved, I was able to create the gguf file.

I figure people want to try the model in Colab. So here it is, so you don't have to create it yourself

Downloads last month
3
GGUF
Model size
3.69B params
Architecture
mpt
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.
The model cannot be deployed to the HF Inference API: The model has no library tag.