R136a1's picture
Update README.md
86d6b10
metadata
license: other
language:
  - en

Detail

Quantized at 6.13bpw h8, as always

Original model card

I received a request to upload the FP16 version of my first MergeMonster 13b merge, which had the sole focus of reducing overused ChatGPT roleplaying phrases. Many ministrations were murdered throughout this process.

GGUF quantizations can be found in my main MergeMonster repo.

Merge Composition

Unsurprisingly, not much remained of MythoMax. I'm (obviously!) very fond of it, but it's not perfect.

Reminder: MergeMonster works with hundreds of pieces scattered throughout the final merge. If a model takes a large part of the composition this does not mean it is merged in a manner that might be logical to humans.

Base model: MythoMax-L2-13b

------ FINAL MERGE COMPOSITION ------
LLaMA2-13B-Tiefighter: 0.27
Emerhyst-13B: 0.26
Orca-2-13b: 0.24
Stheno-1.8-L2-13B: 0.09
MythoMax-L2-13b: 0.09
Athena-v4: 0.04

Prompt Format

As always, I suggest sticking to the Alpaca format.

### Instruction:
Your instruction or question here.

### Response: