Edit model card

CHE-72/Phi-3-mini-128k-instruct-Q6_K-GGUF

This model was converted to GGUF format from microsoft/Phi-3-mini-128k-instruct using llama.cpp via the ggml.ai's GGUF-my-repo space. Refer to the original model card for more details on the model.

Downloads last month
110
GGUF
Model size
14B params
Architecture
phi3

2-bit

3-bit

Inference Examples
Inference API (serverless) is not available, repository is disabled.

Model tree for CHE-72-ZLab/Microsoft-Phi3-14B-Instruct128K-GGUF

Quantized
this model