qwerrwe / src /axolotl /monkeypatch /llama_attn_hijack_flash.py

Commit History

skip some flash attn patches unless explicitly enabled (#643)
895f0a0
unverified

winglian commited on

update for recent transformers updates (#636)
60c7c48
unverified

winglian commited on

Add training callback to send predictions to WandB table (#521)
5b67ea9
unverified

Glavin001 commited on

reorg a bit
fc8766e

tmm1 commited on

use flash_attn rmsnorm when available (#526)
72a6fe1
unverified

tmm1 commited on

use flash_attn xentropy when available (#525)
5fe30b1
unverified

tmm1 commited on

is_causal fix for evals?
fbf49a4

winglian commited on

fix evals (#447)
ee26281
unverified

winglian commited on

standardize attn hijack patches (#381)
06edf17
unverified

tmm1 winglian commited on

fix check for flash attn branching (#377)
343ac84
unverified

winglian commited on

Attention mask and position id fixes for packing (#285)
2bb0b78
unverified

winglian commited on

move flash-attn monkey patch alongside the others
312a9fa

tmm1 commited on