Edit model card

🪽 Hermes-3-Llama-3.1-70B-lorablated

image/png

8B version: mlabonne/Hermes-3-Llama-3.1-8B-lorablated

This is an uncensored version of NousResearch/Hermes-3-Llama-3.1-70B using lorablation.

You can see in the following example how Hermes 3 refuses to answer a legitimate question while the abliterated model complies:

image/png

The recipe is based on @grimjim's grimjim/Llama-3.1-8B-Instruct-abliterated_via_adapter (special thanks):

  1. Extraction: We extract a LoRA adapter by comparing two models: a censored Llama 3 (meta-llama/Meta-Llama-3-70B-Instruct) and an abliterated Llama 3.1 (failspy/Meta-Llama-3.1-70B-Instruct-abliterated).
  2. Merge: We merge this new LoRA adapter using task arithmetic to the censored NousResearch/Hermes-3-Llama-3.1-70B to abliterate it.

image/png

See this article to learn more about abliteration.

⚡ Quantization

🧩 Configuration

This model was merged using the task arithmetic merge method using NousResearch/Hermes-3-Llama-3.1-70B + Llama-3.1-70B-Instruct-abliterated-LORA as a base.

The following YAML configuration was used to produce this model:

base_model: NousResearch/Hermes-3-Llama-3.1-70B+mlabonne/Llama-3.1-70B-Instruct-abliterated-LORA
dtype: bfloat16
merge_method: task_arithmetic
parameters:
  normalize: false
slices:
- sources:
  - layer_range: [0, 32]
    model: NousResearch/Hermes-3-Llama-3.1-70B+mlabonne/Llama-3.1-70B-Instruct-abliterated-LORA
    parameters:
      weight: 1.0

You can reproduce this model using the following commands:

# Setup
git clone https://github.com/arcee-ai/mergekit.git
cd mergekit && pip install -e .
pip install bitsandbytes

# Merge using previous config
mergekit-yaml config.yaml Hermes-3-Llama-3.1-70B-lorablated --allow-crimes --lora-merge-cache=./cache

Open LLM Leaderboard Evaluation Results

Detailed results can be found here

Metric Value
Avg. 35.70
IFEval (0-Shot) 71.44
BBH (3-Shot) 52.34
MATH Lvl 5 (4-Shot) 13.82
GPQA (0-shot) 13.20
MuSR (0-shot) 22.02
MMLU-PRO (5-shot) 41.37
Downloads last month
1,059
Safetensors
Model size
70.6B params
Tensor type
BF16
·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Model tree for mlabonne/Hermes-3-Llama-3.1-70B-lorablated

Collection including mlabonne/Hermes-3-Llama-3.1-70B-lorablated

Evaluation results