Transformers
GGUF
English
Russian
Not-For-All-Audiences
nsfw
Edit model card

Silicon-Masha-7B-GGUF

Silicon-Masha-7B

What's that?

Silicon-Masha-7B-GGUF is a quantized model aiming to be both strong in RP, be smart and understand Russian, that can follow character maps very well. It is suitable for RP/ERP and general use.It can be run on weak samovar using llama.cpp or koboldcpp.

Prompt Template (Alpaca)

I found the best SillyTavern results from using the Noromaid template but please try other templates! Let me know if you find anything good.

SillyTavern config files: Context, Instruct.

Additionally, here is my highly recommended Text Completion preset. You can tweak this by adjusting temperature up or dropping min p to boost creativity or raise min p to increase stability. You shouldn't need to touch anything else!

Below is an instruction that describes a task. Write a response that appropriately completes the request.
### Instruction:
{prompt}
### Response:

Provided files

Name Quant method Bits Use case
silicon-masha-7b.Q2_K.gguf Q2_K 2 smallest, significant quality loss - not recommended for most purposes
silicon-masha-7b.Q3_K_M.gguf Q3_K_M 3 very small, high quality loss
silicon-masha-7b.Q4_0.gguf Q4_0 4 legacy; small, very high quality loss - prefer using Q3_K_M
silicon-masha-7b.Q4_K_M.gguf Q4_K_M 4 medium, balanced quality - recommended
silicon-masha-7b.Q5_0.gguf Q5_0 5 legacy; medium, balanced quality - prefer using Q4_K_M
silicon-masha-7b.Q5_K_M.gguf Q5_K_M 5 large, very low quality loss - recommended
silicon-masha-7b.Q6_K.gguf Q6_K 6 very large, extremely low quality loss
silicon-masha-7b.Q8_0.gguf Q8_0 8 very large, extremely low quality loss - not recommended

How run it ?

llama.cpp

./main -ngl 35 -m silicon-masha-7b.Q4_K_M.gguf --color -c 32768 --temp 0.7 --repeat_penalty 1.1 -n -1 -p "Below is an instruction that describes a task. Write a response that appropriately completes the request.\n{system_message}\n### Instruction:{prompt}\n### Response:\n"
Downloads last month
63
GGUF
Model size
7.24B params
Architecture
llama

2-bit

3-bit

4-bit

5-bit

6-bit

8-bit

Inference API
Inference API (serverless) has been turned off for this model.

Model tree for LakoMoor/Silicon-Masha-7B-GGUF

Quantized
this model