File size: 554 Bytes
322aadb cd93b13 322aadb 77df9ee 9321d4c cd93b13 9321d4c cd93b13 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 |
---
base_model:
- qingy2024/Qwarkstar-4B
license: apache-2.0
datasets:
- HuggingFaceTB/smoltalk
language:
- en
pipeline_tag: text-generation
---
## Qwarkstar 4B Instruct (Preview)
> [!NOTE]
> Training complete!
This model is fine-tuned using Supervised Fine-Tuning (SFT) on 100k samples from the `HuggingFaceTB/smoltalk` dataset.
It follows the ChatML input-output formatting template.
### Training Details:
- **Base Model**: `qingy2024/Qwarkstar-4B`
- **Batch Size**: 32 (2 H100s x 8 per GPU)
- **Max Gradient Norm**: 1.0
- **Final Loss**: ~0.59 |