Commit History
support to disable exllama for gptq (#604)
faecff9
unverified
more sane defaults for openllama 3b used for quickstarts (#602)
674c576
unverified
btlm and falcon monkey patches for flash attn (#566)
6b9b229
unverified
make phi training work with Loras (#588)
62eaee7
unverified
Support Sample packing for phi arch (#586)
12a2dbb
unverified
Fix Codellama examples (#582)
1aa4007
unverified
Doan Minh Phuong
commited on
Phi examples (#569)
2284209
unverified
Add training callback to send predictions to WandB table (#521)
5b67ea9
unverified
recommend padding when using sample packing (#531)
3437149
unverified
Add support for GPTQ using native transformers/peft (#468)
3355706
unverified
pad_to_worst_case_seq_len boolean, for testing memory limits (#498)
8e197f6
unverified
Feat(cfg): Add code-llama configs for all sizes (#479)
3513071
unverified
Add example Llama 2 ReLoRA config (#471)
fe4d6ba
unverified
improve llama pad token handling (#475)
cb9797e
unverified
don't use mask expansion for inference (#392)
1687be6
unverified
new llama-2 default settings (#370)
fdffef5
unverified
Add wandb_entity to wandb options, update example configs, update README (#361)
7019509
unverified
set group_by_length to false in examples
36fefcf
feat/llama-2 examples (#319)
dc71d88
unverified
Add XGen info to README and example config
3881143
Ethan Smith
commited on