Commit History

Feat: Warns to add to modules_to_save when adding tokens or switching special_tokens (#787)
1ffa386
unverified

Nanobit commited on

Fix Deepspeed loading (#950)
5ea3aa3
unverified

winglian commited on

Flash attn hotfix (#951)
f1f60cb
unverified

winglian commited on

Mixtral official (#942)
7fabc4d
unverified

winglian commited on

Mixtral multipack (#928)
68b227a
unverified

winglian commited on

support for mamba (#915)
40a6362
unverified

winglian commited on

fix(tokenizer): handle fast tokenizer properly for bos/eos (#914)
fde091c
unverified

Nanobit commited on

feat: add check for quantized model (#913)
a581e9f
unverified

Nanobit winglian commited on

Support device_map=sequential & max_memory config parameters (#903)
992e742
unverified

Bryan Thornbury winglian commited on

fix for qwen w lora (#906)
3e3229e
unverified

winglian commited on

Feat: Add Qwen (#894)
1115c50
unverified

Nanobit commited on

Phi update 202311 (#876)
9bf854e
unverified

winglian commited on

allow overriding of model_config parameters from the YML (#853)
1bc1186
unverified

winglian commited on

fix model parallel (#816)
964d858
unverified

winglian commited on

fix(tokenizer): update log order after update (#806)
10388a8
unverified

Nanobit commited on

fix(config): Set eos/bos to tokenizer if different (#801)
637ed09
unverified

Nanobit commited on

refactor neft patch to be more re-usable similar to trl's impl (#796)
827ec3d
unverified

winglian commited on

chore: refactor truthy check and fix mypy (#780)
11d1d60
unverified

Nanobit commited on

Fix(model): Linear detected and added to target module with rope linear (#738)
440c3ab
unverified

Nanobit commited on

add noisy embedding (#721)
3bd9528
unverified

Maxime Maxime commited on

Fix: Higher vram usage for mistral and sample_packing (#691)
669f1d0
unverified

Nanobit commited on

flash_attention + sample packing for stablelm 3b (#671)
2d60ba3
unverified

winglian commited on

Fix: ValueError when FA + Mistral when padding_side=right (#681)
eb480df
unverified

Nanobit commited on

Fix(tokenizer): Set rstrip,lstrip,norm to False (#678)
e0b7eea
unverified

Nanobit commited on

chore: Clean up repetitive model kwargs (#670)
e62d590
unverified

Nanobit commited on

Feat: Allow usage of native Mistral FA when no sample_packing (#669)
697c50d
unverified

Nanobit commited on

remove patch fix for phi (#664)
f34648c
unverified

winglian commited on

Mistral flash attn packing (#646)
b6ab8aa
unverified

winglian commited on

skip some flash attn patches unless explicitly enabled (#643)
895f0a0
unverified

winglian commited on

Feat: Add support for upstream FA2 (#626)
19a600a
unverified

Nanobit commited on

misc fixes to add gptq tests (#621)
03e5907
unverified

winglian commited on

support to disable exllama for gptq (#604)
faecff9
unverified

winglian commited on

Delete duplicate lines (#606)
aa656e0
unverified

bofenghuang commited on

btlm and falcon monkey patches for flash attn (#566)
6b9b229
unverified

winglian commited on

make phi training work with Loras (#588)
62eaee7
unverified

winglian commited on

don't resize embeddings if it's already large enough (#577)
3607882
unverified

winglian commited on

Support Sample packing for phi arch (#586)
12a2dbb
unverified

winglian commited on

Add training callback to send predictions to WandB table (#521)
5b67ea9
unverified

Glavin001 commited on

fix for quant config from model (#540)
a94f9cb
unverified

winglian commited on

Add support for GPTQ using native transformers/peft (#468)
3355706
unverified

winglian commited on

fix: bad dtype for full finetune (#504)
1991946
unverified

Maxime winglian commited on

Refactor train cfg cli (#499)
125cccb
unverified

winglian commited on

simplify linear layer locator
267b7b2

tmm1 commited on

fsdp requires params be the same type too (#493)
98bf76e
unverified

winglian commited on

Fix(tokenizer): Make sure to add pad for CodeLlamaTokenizer (#489)
4c37bd0
unverified

Nanobit commited on

fix condition and add logging
3a011ea

tmm1 commited on

rename var and reformat
f319b0b

tmm1 commited on

Update src/axolotl/utils/models.py
7fd662d
unverified

Maxime tmm1 commited on

Update src/axolotl/utils/models.py
9e69968
unverified

Maxime tmm1 commited on