IntelligentEstate/Chocolat_Bite-14B-Q4_K_M-GGUF
This model was converted to GGUF format from the always amazing models of Jpacifico jpacifico/Chocolatine-2-14B-Instruct-v2.0b2
using llama.cpp
Refer to the original model card for more details on the model.
Made for a larger but still under 10GB base station GGUF backbone of the Estate/Enterprise system Project CutPurse(API FREEDOM) Quant Test
Set up as the base or writing layer of your swarm agent or in your server for Quick and reliable inference(Much better than ChatGPT 1o/R1 when tied to tool use from Pancho and web query from RSS feeds and so on) while keeping all your Data and your clients/Families/financials secure.
Use with a Limit Crossing AGI template for your own Agent of Coheasion or Chaos. !!(Use LimitCrossing with Extreame Caution)!! Paper in Files
Use with llama.cpp
Install llama.cpp through brew (works on Mac and Linux)
brew install llama.cpp
Invoke the llama.cpp server or the CLI.
- Downloads last month
- 20
Inference Providers
NEW
This model is not currently available via any of the supported third-party Inference Providers, and
HF Inference API was unable to determine this model’s pipeline type.
Model tree for IntelligentEstate/Chocolate_Bite-14B-Q4_K_M-GGUF
Base model
jpacifico/Chocolatine-2-14B-Instruct-v2.0b2