neuralmagic 's Collections

FP8 LLMs for vLLM

Accurate FP8 quantized models by Neural Magic, ready for use with vLLM!