afrideva's picture
Upload README.md with huggingface_hub
bf327c7 verified
---
base_model: M4-ai/NeuralReyna-Mini-1.8B-v0.3
datasets:
- argilla/OpenHermes2.5-dpo-binarized-alpha
- Locutusque/Hercules-v3.0
inference: false
language:
- en
license: apache-2.0
model-index:
- name: NeuralReyna-Mini-1.8B-v0.3
results:
- dataset:
args:
num_few_shot: 25
config: ARC-Challenge
name: AI2 Reasoning Challenge (25-Shot)
split: test
type: ai2_arc
metrics:
- name: normalized accuracy
type: acc_norm
value: 35.58
source:
name: Open LLM Leaderboard
url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=M4-ai/NeuralReyna-Mini-1.8B-v0.3
task:
name: Text Generation
type: text-generation
- dataset:
args:
num_few_shot: 10
name: HellaSwag (10-Shot)
split: validation
type: hellaswag
metrics:
- name: normalized accuracy
type: acc_norm
value: 61.13
source:
name: Open LLM Leaderboard
url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=M4-ai/NeuralReyna-Mini-1.8B-v0.3
task:
name: Text Generation
type: text-generation
- dataset:
args:
num_few_shot: 5
config: all
name: MMLU (5-Shot)
split: test
type: cais/mmlu
metrics:
- name: accuracy
type: acc
value: 44.22
source:
name: Open LLM Leaderboard
url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=M4-ai/NeuralReyna-Mini-1.8B-v0.3
task:
name: Text Generation
type: text-generation
- dataset:
args:
num_few_shot: 0
config: multiple_choice
name: TruthfulQA (0-shot)
split: validation
type: truthful_qa
metrics:
- type: mc2
value: 41.99
source:
name: Open LLM Leaderboard
url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=M4-ai/NeuralReyna-Mini-1.8B-v0.3
task:
name: Text Generation
type: text-generation
- dataset:
args:
num_few_shot: 5
config: winogrande_xl
name: Winogrande (5-shot)
split: validation
type: winogrande
metrics:
- name: accuracy
type: acc
value: 60.93
source:
name: Open LLM Leaderboard
url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=M4-ai/NeuralReyna-Mini-1.8B-v0.3
task:
name: Text Generation
type: text-generation
- dataset:
args:
num_few_shot: 5
config: main
name: GSM8k (5-shot)
split: test
type: gsm8k
metrics:
- name: accuracy
type: acc
value: 6.75
source:
name: Open LLM Leaderboard
url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=M4-ai/NeuralReyna-Mini-1.8B-v0.3
task:
name: Text Generation
type: text-generation
model_creator: M4-ai
model_name: NeuralReyna-Mini-1.8B-v0.3
pipeline_tag: text-generation
quantized_by: afrideva
tags:
- gguf
- ggml
- quantized
- q2_k
- q3_k_m
- q4_k_m
- q5_k_m
- q6_k
- q8_0
---
# M4-ai/NeuralReyna-Mini-1.8B-v0.3-GGUF
Quantized GGUF model files for [NeuralReyna-Mini-1.8B-v0.3](https://huggingface.co/M4-ai/NeuralReyna-Mini-1.8B-v0.3) from [M4-ai](https://huggingface.co/M4-ai)
| Name | Quant method | Size |
| ---- | ---- | ---- |
| [neuralreyna-mini-1.8b-v0.3.fp16.gguf](https://huggingface.co/afrideva/NeuralReyna-Mini-1.8B-v0.3-GGUF/resolve/main/neuralreyna-mini-1.8b-v0.3.fp16.gguf) | fp16 | 3.68 GB |
| [neuralreyna-mini-1.8b-v0.3.q2_k.gguf](https://huggingface.co/afrideva/NeuralReyna-Mini-1.8B-v0.3-GGUF/resolve/main/neuralreyna-mini-1.8b-v0.3.q2_k.gguf) | q2_k | 846.57 MB |
| [neuralreyna-mini-1.8b-v0.3.q3_k_m.gguf](https://huggingface.co/afrideva/NeuralReyna-Mini-1.8B-v0.3-GGUF/resolve/main/neuralreyna-mini-1.8b-v0.3.q3_k_m.gguf) | q3_k_m | 1.02 GB |
| [neuralreyna-mini-1.8b-v0.3.q4_k_m.gguf](https://huggingface.co/afrideva/NeuralReyna-Mini-1.8B-v0.3-GGUF/resolve/main/neuralreyna-mini-1.8b-v0.3.q4_k_m.gguf) | q4_k_m | 1.22 GB |
| [neuralreyna-mini-1.8b-v0.3.q5_k_m.gguf](https://huggingface.co/afrideva/NeuralReyna-Mini-1.8B-v0.3-GGUF/resolve/main/neuralreyna-mini-1.8b-v0.3.q5_k_m.gguf) | q5_k_m | 1.38 GB |
| [neuralreyna-mini-1.8b-v0.3.q6_k.gguf](https://huggingface.co/afrideva/NeuralReyna-Mini-1.8B-v0.3-GGUF/resolve/main/neuralreyna-mini-1.8b-v0.3.q6_k.gguf) | q6_k | 1.58 GB |
| [neuralreyna-mini-1.8b-v0.3.q8_0.gguf](https://huggingface.co/afrideva/NeuralReyna-Mini-1.8B-v0.3-GGUF/resolve/main/neuralreyna-mini-1.8b-v0.3.q8_0.gguf) | q8_0 | 1.96 GB |
## Original Model Card:
# NeuralReyna-Mini-1.8B-v0.3
![Reyna image](https://th.bing.com/th/id/OIG3.8IBxuT77hh6Y_r1DZ6WK?dpr=2.6&pid=ImgDetMain)
# Description
Taken aloobun/Reyna-Mini-1.8B-v0.2 and further fine-tuned it using DPO using the argilla/OpenHermes2.5-dpo-binarized-alpha.
This model has capabilities in coding, math, science, roleplay, and function calling.
This model was trained on OpenAI's ChatML prompt format.
# Evaluation
Coming soon
# Contributions
Thanks to @aloobun and @Locutusque for their contributions to this model.
# [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)
Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_M4-ai__NeuralReyna-Mini-1.8B-v0.3)
| Metric |Value|
|---------------------------------|----:|
|Avg. |41.77|
|AI2 Reasoning Challenge (25-Shot)|35.58|
|HellaSwag (10-Shot) |61.13|
|MMLU (5-Shot) |44.22|
|TruthfulQA (0-shot) |41.99|
|Winogrande (5-shot) |60.93|
|GSM8k (5-shot) | 6.75|