Triangle104/Jamet-8B-L3-MK.V-Blackroot-Q8_0-GGUF
This model was converted to GGUF format from Hastagaras/Jamet-8B-L3-MK.V-Blackroot
using llama.cpp via the ggml.ai's GGUF-my-repo space.
Refer to the original model card for more details on the model.
Different base model, different methods (without model stock merge, because it's not good...at least for me, it gives repeating sentences or words at 0 temperature, that also happened with the Halu Blackroot). This model maybe more similar to the Anjir model.
This model has been on a long journey; I have like...9 variations of it. I tested all of them at Q4_K_M and decided to release this one, this is variation number 7.
And thanks for all the feedback from the previous model, it helps a lot. (I couldn't fix the issue with German because I don't know how to fix it since I can't speak German, heck, even my English is bad.)
More Details:
This model is based on this model, which is based on the UltimateAnjir model. It shares the same creative, cheerful, and positive tendencies. Then I merged it with Llama 3 Instruct.
Next is DPO! to reduce the cheerfulness, emojis, and positivity. (This is based on the Jamet MK.II Feedback regarding positivity.) I trained a QLora with about 1,000 prompts from Alpaca to generate a dataset, selecting the ones with emojis, removing the emojis using regex and moving those without emojis to the chosen, and the responses with emojis to the rejected.(With unsloth)
Then, I applied the Abomination Lora from Blackroot.
Next, I applied the Anjir Adapter (64 Rank version with reduced Alpha) to improve formatting while retaining the previous Lora influences. (This is based on the Anjir Feedback, which suggests that Anjir has better formatting than the Halu Blackroot.)
And then merged the model with the Anjrit model. (I won't release the Anjrit model as it struggles with longer contexts. I'm only interested in its no refusals storytelling abilities, but you can find a brief overview of the model on my Anjir model page.)
And that's it. Thanks again for all the feedback!
Notes:
I'm not responsible for anything.
This is an RP and Storytelling model.
You can write your feedback at the discussion, so i can improve my models.
Like all of my previous models, higher temperatures will be incoherent, so use around 0.85-1.05.(I've been trying to fix this since Halu Blackroot but haven't had much luck, I think merging the base with Llama 3 Instruct helped a lot.)
Use with llama.cpp
Install llama.cpp through brew (works on Mac and Linux)
brew install llama.cpp
Invoke the llama.cpp server or the CLI.
CLI:
llama-cli --hf-repo Triangle104/Jamet-8B-L3-MK.V-Blackroot-Q8_0-GGUF --hf-file jamet-8b-l3-mk.v-blackroot-q8_0.gguf -p "The meaning to life and the universe is"
Server:
llama-server --hf-repo Triangle104/Jamet-8B-L3-MK.V-Blackroot-Q8_0-GGUF --hf-file jamet-8b-l3-mk.v-blackroot-q8_0.gguf -c 2048
Note: You can also use this checkpoint directly through the usage steps listed in the Llama.cpp repo as well.
Step 1: Clone llama.cpp from GitHub.
git clone https://github.com/ggerganov/llama.cpp
Step 2: Move into the llama.cpp folder and build it with LLAMA_CURL=1
flag along with other hardware-specific flags (for ex: LLAMA_CUDA=1 for Nvidia GPUs on Linux).
cd llama.cpp && LLAMA_CURL=1 make
Step 3: Run inference through the main binary.
./llama-cli --hf-repo Triangle104/Jamet-8B-L3-MK.V-Blackroot-Q8_0-GGUF --hf-file jamet-8b-l3-mk.v-blackroot-q8_0.gguf -p "The meaning to life and the universe is"
or
./llama-server --hf-repo Triangle104/Jamet-8B-L3-MK.V-Blackroot-Q8_0-GGUF --hf-file jamet-8b-l3-mk.v-blackroot-q8_0.gguf -c 2048
- Downloads last month
- 31
Model tree for Triangle104/Jamet-8B-L3-MK.V-Blackroot-Q8_0-GGUF
Base model
Hastagaras/Jamet-8B-L3-MK.V-Blackroot