WellMinded Therapy Engine (8B Q4_0)

This is a fine-tuned version of the LLaMA 3.1 8B model, optimized for psychologist-like conversations. The model is quantized to 4-bit precision (Q4_0) for efficient inference.

Usage

You can load and use this model with the llama-cpp-python library:

from llama_cpp import Llama

llm = Llama(model_path="psychologistv2-8.0B-Q4_0.gguf")
response = llm("Hi, I'm feeling stressed lately.")
print(response["choices"][0]["text"])
Downloads last month
77
GGUF
Model size
8.03B params
Architecture
llama

4-bit

Inference API
Unable to determine this model's library. Check the docs .

Model tree for PixelPanda1/WellMinded_Therapy_Engine-gguf

Quantized
(182)
this model