Model Card for Model ID

This LoRA adapter was extracted from mlabonne/Meta-Llama-3.1-70B-Instruct-lorablated and uses meta-llama/Llama-3.1-70B-Instruct as a base.

Model Details

The model was extracted by running mlabonne/harmful_behaviors and the user prompts (but not assistant responses or system messages) from Guilherme34/uncensor through the original abliterated model to generate a dataset of prompt/completion pairs, and was trained for 2 epochs on a 8xA100s with Axolotl using FSDP. Since the original abliterated model isn't perfect at avoiding refusals, the dataset was cleaned to remove the few refusals generated prior to training.

Model Description

  • Developed by: @reissbaker
  • Funded by: Synthetic Lab
  • License: Apache 2.0
  • Finetuned from model: Llama 3.1 70B Instruct

How to Get Started with the Model

Run the model with one click on glhf.chat.

Training Hyperparameters

  • BF16 mixed-precision
  • 4e-4 LR
  • Linear LR schedule
  • Fused AdamW optimizer
Downloads last month
11,036
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.
The model cannot be deployed to the HF Inference API: The model has no pipeline_tag.

Model tree for reissbaker/llama-3.1-70b-abliterated-lora

Adapter
(20)
this model