Hugging Face
Models
Datasets
Spaces
Posts
Docs
Solutions
Pricing
Log In
Sign Up
dball
/
zephyr-7b-sft-qlora
like
0
PEFT
TensorBoard
Safetensors
HuggingFaceH4/ultrachat_200k
mistral
alignment-handbook
Generated from Trainer
trl
sft
4-bit precision
bitsandbytes
License:
apache-2.0
Model card
Files
Files and versions
Metrics
Training metrics
Community
2
Train
Use this model
New discussion
New pull request
Resources
PR & discussions documentation
Code of Conduct
Hub documentation
All
Discussions
Pull requests
View closed (0)
Adding Evaluation Results
#2 opened 8 months ago by
leaderboard-pr-bot
Is the drop in many metrics expected? Why do SFT first if it makes the model worse? Why not do DPO directly on the mistral model?
1
#1 opened 10 months ago by
dball