Edit model card
Llama-3.1-SuperNova-Lite

Overview

Llama-3.1-SuperNova-Lite is an 8B parameter model developed by Arcee.ai, based on the Llama-3.1-8B-Instruct architecture. It is a distilled version of the larger Llama-3.1-405B-Instruct model, leveraging offline logits extracted from the 405B parameter variant. This 8B variation of Llama-3.1-SuperNova maintains high performance while offering exceptional instruction-following capabilities and domain-specific adaptability.

The model was trained using a state-of-the-art distillation pipeline and an instruction dataset generated with EvolKit, ensuring accuracy and efficiency across a wide range of tasks. For more information on its training, visit blog.arcee.ai.

Llama-3.1-SuperNova-Lite excels in both benchmark performance and real-world applications, providing the power of large-scale models in a more compact, efficient form ideal for organizations seeking high performance with reduced resource requirements.

Evaluations

Here are our internal benchmarks using the main branch of lm evaluation harness:

Benchmark SuperNova-Lite Llama-3.1-8b-Instruct
IF_Eval 81.1 77.4
MMLU Pro 38.7 37.7
TruthfulQA 64.4 55.0
BBH 51.1 50.6
GPQA 31.2 29.02

The script used for evaluation can be found inside this repository under /eval.sh, or click here.

Downloads last month
5,951
Safetensors
Model size
8.03B params
Tensor type
BF16
·
Inference API
Input a message to start chatting with arcee-ai/Llama-3.1-SuperNova-Lite.

Model tree for arcee-ai/Llama-3.1-SuperNova-Lite

Finetuned
this model
Finetunes
5 models
Merges
12 models
Quantizations
14 models

Dataset used to train arcee-ai/Llama-3.1-SuperNova-Lite

Spaces using arcee-ai/Llama-3.1-SuperNova-Lite 2