Hugging Face
Models
Datasets
Spaces
Posts
Docs
Enterprise
Pricing
Log In
Sign Up
RedMist137
/
DPO-Zephyr-7B
like
0
Safetensors
RedMist137/AIHF_DPO_iter0
opt
alignment-handbook
trl
dpo
Generated from Trainer
License:
other
Model card
Files
Files and versions
Community
85cbd5e
DPO-Zephyr-7B
Commit History
Training in progress, step 100
85cbd5e
verified
RedMist137
commited on
Oct 17, 2024
Training in progress, step 100
b91e871
verified
RedMist137
commited on
Oct 13, 2024
initial commit
502ec73
verified
RedMist137
commited on
Oct 12, 2024