--- license: mit datasets: - normster/RealGuardrails base_model: - meta-llama/Llama-3.2-3B library_name: transformers --- # RealGuardrails Models This model was trained on the [RealGuardrails](https://huggingface.co/datasets/normster/RealGuardrails) dataset, an instruction-tuning dataset focused on improving system prompt adherence and precedence. In particular, it was trained via SFT on the `systemmix` split of ~150K examples using our custom training library [torchllms](https://github.com/normster/torchllms) and converted back to a `transformers` compatible checkpoint. ## Training Hyperparameters | Name | Value | | :--- | :--- | | optimizer | AdamW | | batch size | 128 | | learning rate | 2e-5 | | lr scheduler | cosine with 200 warmup steps | | betas | (0.9, 0.999) | | eps | 1e-8 | | weight decay | 0 | | epochs | 1 | | max grad norm | 1.0 | | precision | bf16 | | max length | 4096 |