alokabhishek
commited on
Commit
•
4af1377
1
Parent(s):
9b66a98
update readme
Browse files
README.md
CHANGED
@@ -3,18 +3,34 @@ library_name: transformers
|
|
3 |
tags: []
|
4 |
---
|
5 |
|
6 |
-
#
|
7 |
|
8 |
<!-- Provide a quick summary of what the model is/does. -->
|
|
|
9 |
|
10 |
|
11 |
|
12 |
## Model Details
|
13 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
14 |
### Model Description
|
15 |
|
16 |
<!-- Provide a longer summary of what this model is. -->
|
17 |
|
|
|
|
|
18 |
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
|
19 |
|
20 |
- **Developed by:** [More Information Needed]
|
|
|
3 |
tags: []
|
4 |
---
|
5 |
|
6 |
+
# Mistral-7B-Instruct-v0.2-bnb-4bit
|
7 |
|
8 |
<!-- Provide a quick summary of what the model is/does. -->
|
9 |
+
This repo contains 4-bit quantized (using bitsandbytes) model Mistral AI_'s Mistral-7B-Instruct-v0.2
|
10 |
|
11 |
|
12 |
|
13 |
## Model Details
|
14 |
|
15 |
+
Model creator: [Mistral AI_](https://huggingface.co/mistralai)
|
16 |
+
Original model: [Mistral-7B-Instruct-v0.2](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2)
|
17 |
+
|
18 |
+
|
19 |
+
### About 4 bit quantization using bitsandbytes
|
20 |
+
|
21 |
+
QLoRA: Efficient Finetuning of Quantized LLMs: [arXiv - QLoRA: Efficient Finetuning of Quantized LLMs] (https://arxiv.org/abs/2305.14314)
|
22 |
+
|
23 |
+
Hugging Face Blog post on 4-bit quantization using bitsandbytes: [Making LLMs even more accessible with bitsandbytes, 4-bit quantization and QLoRA] (https://huggingface.co/blog/4bit-transformers-bitsandbytes)
|
24 |
+
|
25 |
+
bitsandbytes github repo: [bitsandbytes github repo] (https://github.com/TimDettmers/bitsandbytes)
|
26 |
+
|
27 |
+
|
28 |
### Model Description
|
29 |
|
30 |
<!-- Provide a longer summary of what this model is. -->
|
31 |
|
32 |
+
|
33 |
+
|
34 |
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
|
35 |
|
36 |
- **Developed by:** [More Information Needed]
|