qwerrwe / setup.py

Commit History

Mistral: Sliding Window Attention with Flash Attention and Sample Packing (#732)
a045db0
unverified

casperhansen winglian commited on

tweak for xformers install w pytorch 2.1.0 (#727)
7f2027d
unverified

winglian commited on

workaround for installing xformers w torch 2.1.0 (#725)
8d288a2
unverified

winglian commited on

update README w deepspeed info (#605)
c25ba79
unverified

winglian commited on

fix the sed command to replace the version w the tag
772cd87

winglian commited on

add long_description for pypi push (#555)
6c5fbe6
unverified

winglian commited on

update readme to point to direct link to runpod template, cleanup install instrucitons (#532)
34c0a86
unverified

winglian commited on

Add support for GPTQ using native transformers/peft (#468)
3355706
unverified

winglian commited on

recast loralayer, norm, lmhead + embed token weights per original qlora (#393)
96deb6b
unverified

winglian commited on

flash attn pip install (#426)
cf66547
unverified

mhenrichsen Ubuntu mhenrichsen Mads Henrichsen winglian commited on

Merge pull request #108 from OpenAccess-AI-Collective/docker-gptq
bbc5bc5
unverified

winglian commited on

Apply isort then black
37293dc

Nanobit commited on

Lint setup.py
8b617cc

Nanobit commited on

default to qlora support, make gptq specific image
6ef96f5

winglian commited on

black formatting
2bc1a5b

winglian commited on

docker layer caching, build w axolotl from base build
990bec6

winglian commited on

handle empty lines
f50de1b

winglian commited on

fix install to work with latest alpaca lora 4bit
4131183

winglian commited on

4bit quantized support (wip)
77fca25

winglian commited on