Commit History
push intermediate model checkpoints to hub
612aabd
skip the system prompt
05ab909
pylint for duplicated code for system prompts
7b57ed7
add tests and supoort for loader for sys prompt data
3a38271
initial wip to get sys prompt from dataset
8d20e0a
optionally define whether to use_fast tokenizer
47d601f
add validation and tests for adamw hyperparam
cb9d3af
support adamw and grad norm hyperparams
6d0ee4b
Merge pull request #212 from OpenAccess-AI-Collective/doc-20230615-v1
a81f52d
unverified
Merge pull request #214 from OpenAccess-AI-Collective/fix-tokenizing-labels
1925eaf
unverified
add float16 docs and tweak typehints
88e17ff
bugfix for potential off by one
7925ddc
style correction
136522f
maciej.karasek
commited on
issue #205 bugfix
556fe40
maciej.karasek
commited on
update alpaca_chat prompts for instructions to explainn the conversation
4b43a66
Merge branch 'main' into flash-optimum
fd2c981
unverified
Merge pull request #187 from OpenAccess-AI-Collective/strip-peft-device-map
93dacba
unverified
Merge pull request #177 from NanoCode012/fix/landmark-patch
8002ffb
unverified
Merge pull request #192 from OpenAccess-AI-Collective/sharegpt-custom-prompt
74ef5cc
unverified
Merge branch 'main' into strip-peft-device-map
5e616d9
unverified
Merge pull request #159 from AngainorDev/patch-1
8e568bb
unverified
add typehints
c7dee56
add new sharegpt, refactor prompt so it can be customized later, add exception if no data is processed
aac4b76
add check for attr
c9a149f
new validation for mpt w grad checkpoints
14668fa
Fix strict and Lint
b565ecf
match up gradient checkpointing when using lora w config
fe0b768
Fix set mem_id for inference and refactor
974dc00
Clean up landmark patching
a6190c8
Fix undefined LlamaForCausalLM and del try except
563b6d8
peft no longer needs device_map
cd0a6f6
Refactor landmark attention patch
919727b
fix formatting
958da70
Fix missing cfg.
a808bf9
unverified
Angainor Development
commited on
Merge pull request #182 from OpenAccess-AI-Collective/fix-llama-ref
0124825
unverified
address PR feedback
0c6f928
add streaming dataset support for pretraining datasets
eea2731
more gpt-neox long ctx fixes
ab5cd28
fix bettertransformers save, force it to skip after saving correctly in callback
1a82082
more tweaks to do pre-training with bettertransformers
1210dc8
experimental expansion of ctx len
488a67d
add validation/warning for bettertransformers and torch version
71a43f8
add support for opimum bettertransformers
1edc30c
fix for local variable 'LlamaForCausalLM' referenced before assignment
14163c1
Merge branch 'main' into patch-1
79e2a6f
unverified
Angainor Development
commited on