Commit History
fix symlinks for axolotl outputs (#1625)
e6937e8
unverified
bump versions of deps (#1621)
039e2a0
unverified
update outputs path so that we can mount workspace to /workspace/data (#1623)
4fde300
unverified
update torch 2.2.1 -> 2.2.2 (#1622)
3319780
unverified
Fix `total_num_steps` (#1566)
81da7d2
unverified
FIX: max_length and max_prompt_length was not being sent to ORPOTrainer (#1584)
1e1921b
unverified
make sure to save on the last step (#1615)
1634ac8
unverified
fix attention mask collation (#1603)
0298273
unverified
feat: Add LLaMA-3 instruct prompt strategies for fine-tuning (#1553)
50421c8
unverified
adding llama3 fastchat conversation monkeypatch (#1539)
b32c08f
unverified
ignore the fsdp_config section too (#1606) [skip ci]
fff06af
unverified
make sure to save the lora adapter at the end of RL/dpo training (#1573)
796a085
unverified
improve tool handling roles (#1587)
cb78a36
unverified
feat: exclude mamba blocks for jamba (#1578)
8b9c15b
unverified
Pass deepspeed and fsdp as None explicitly when merging adapters to allow custom device_map (#1575)
9e1480e
unverified
docs(config.qmd): add loraplus example (#1577)
1ac8998
unverified
tpoisonooo
commited on
add torch 2.3.0 to builds (#1593)
7018576
unverified
fix for jupyterlab on cloud start (#1594)
120b809
unverified
improve save callbacks (#1592)
29cf15a
unverified
FIX: TRL trainer preprocessing step was running in one process (#1583)
b9bb169
unverified
Ali Mosavian
Ali Mosavian
commited on
ADD: warning hub model (#1301)
601c08b
unverified
chore(doc): clarify micro_batch_size (#1579) [skip ci]
1aeece6
unverified
PoSE context length ext (#1567)
5294653
unverified
Add ORPO example and e2e test (#1572)
98c25e1
unverified
make sure everything stays in the same dtype when using dpo + FSDP (#1559)
68601ec
unverified
Add support for Gemma chat template (#1530)
60f5ce0
unverified
wrap prepared_ds_path in str() to avoid TypeError in fsspec package (#1548)
7477a53
unverified
ORPO Trainer replacement (#1551)
7d1d22f
unverified
fix(yml): update llama-3 config (#1543) [skip ci]
0e8f340
unverified
fix(packages): lock datasets version (#1545)
59ef254
unverified
fix broken linting (#1541)
c10563c
unverified
Adding Llama-3 qlora (#1536)
37c037c
unverified
llama-3 examples (#1537)
15f7910
unverified
feat(doc): Add example for pad_token (#1535)
d28ba2e
unverified
Create mixtral_22.yml (#1514) [skip ci]
0eadfc8
unverified
Atlas
commited on
Update Readme to include support for Mixtral8X22B (#1518) [skip ci]
bcaa923
unverified
Atlas
commited on
Update README.md (#1521) [skip ci]
7d9bafc
unverified
YTING
commited on