qwerrwe / src /axolotl /utils /trainer.py

Commit History

push intermediate model checkpoints to hub
612aabd

winglian commited on

support adamw and grad norm hyperparams
6d0ee4b

winglian commited on

Merge branch 'main' into flash-optimum
fd2c981
unverified

winglian commited on

Fix set mem_id for inference and refactor
974dc00

Nanobit commited on

fix formatting
958da70

winglian commited on

address PR feedback
0c6f928

winglian commited on

fix bettertransformers save, force it to skip after saving correctly in callback
1a82082

winglian commited on

more tweaks to do pre-training with bettertransformers
1210dc8

winglian commited on

Feat: Add landmark attention
55b8542

Nanobit commited on

Refactor out unmodified save_steps and eval_steps
2ef4634

Nanobit commited on

Set to use cfg.seed or 42 for backward compat
2cfe9e9

Nanobit commited on

fix relative path for fixtures
cfcc549

winglian commited on

Apply isort then black
37293dc

Nanobit commited on

Fix mypy typing
e9650d3

Nanobit commited on

Lint trainer.py
ddb86ea

Nanobit commited on

fix relative path for fixtures
e65aeed

winglian commited on

refactor(param): rename load_4bit config param by gptq
dd00657

Thytu commited on

fixes to make qlora actually work
34c99f9

winglian commited on

apply black formatting
ce34d64

winglian commited on

fix missing fp16 kwarg
2ae936f

winglian commited on

Add qa style data for alpaca instructions, fix one_cycle scheduler
3a50377

winglian commited on

don't need to set here
de6da13

winglian commited on

be able to use adam bnb 8bit and one cycle scheduler w fsdp
9493b1b

winglian commited on

make one cycle lr div factor configurable
99383f1

winglian commited on

Merge branch 'main' into patch-2
89b7f26
unverified

Nanobit commited on

black formatting
2bc1a5b

winglian commited on

various fixes
7a490a4

winglian commited on

Fix Trainer() got multiple values for keyword argument 'callbacks'
813aab3
unverified

Nanobit commited on

Merge pull request #21 from NanoCode012/patch-1
bd3c5a5
unverified

winglian commited on

Update trainer.py
36aaea0
unverified

Nanobit commited on

Fix condition scheduler
5b6690a
unverified

Nanobit commited on

Add callbacks to Trainer
cc77bab

Nanobit commited on

Add callback save peft_model on_save
0d6708b

Nanobit commited on

fix #16 load best model setting when using 8bit
a4329b1

winglian commited on

use micro batch size for eval size if not specified
550502b

winglian commited on

refactor inference, warn if model is frozen
247825b

winglian commited on

Merge pull request #13 from winglian/dev
cb9a887
unverified

winglian commited on

Add eval_batch_size for evaluation
0e74b64

Nanobit commited on

fix log sweep lr
a10a826

winglian commited on

support for multi line inference input, log sweep over learning rates
9105935

winglian commited on

fix adam bnb optimizer grouped parameters, fix peft model 8bit conversion logic, black formatting
7748f3d

winglian commited on

fdsp config dict fix, todo list, add torchdistx support
ad2b48c

winglian commited on

fix fsdp training args
29936bb

winglian commited on

fix for zero value warmup steps
7882181

winglian commited on

fix sharegpt tokenization, refactor tokenization debugging
5159d00

winglian commited on

wire up gradient checkpointing for 4bit
c0f50d9

winglian commited on

tweaks to data loading, 8 bit adam, accelerate and deepspeed
097d367

winglian commited on

various bugfixes
94f5e41

winglian commited on

quickstart instructions for starting from runpod (#5)
0a472e1
unverified

winglian commited on

WIP large refactor to make finetune script a little more manageable (#3)
6045345
unverified

winglian commited on