Transformers
fastai
English
Inference Endpoints
sapphomoon commited on
Commit
6540589
1 Parent(s): d012792

Upload 2 files

Browse files
Files changed (2) hide show
  1. models.json +205 -0
  2. models2.json +240 -0
models.json ADDED
@@ -0,0 +1,205 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "order": "a",
4
+ "md5sum": "e8d47924f433bd561cb5244557147793",
5
+ "name": "Wizard v1.1",
6
+ "filename": "wizardlm-13b-v1.1-superhot-8k.ggmlv3.q4_0.bin",
7
+ "filesize": "7323310848",
8
+ "ramrequired": "16",
9
+ "parameters": "13 billion",
10
+ "quant": "q4_0",
11
+ "type": "LLaMA",
12
+ "systemPrompt": " ",
13
+ "description": "<strong>Best overall model</strong><br><ul><li>Instruction based<li>Gives very long responses<li>Finetuned with only 1k of high-quality data<li>Trained by Microsoft and Peking University<li>Cannot be used commercially</ul"
14
+ },
15
+ {
16
+ "order": "b",
17
+ "md5sum": "725f148218a65ce8ebcc724e52f31b49",
18
+ "name": "GPT4All Falcon",
19
+ "filename": "ggml-model-gpt4all-falcon-q4_0.bin",
20
+ "filesize": "4061641216",
21
+ "requires": "2.4.9",
22
+ "ramrequired": "8",
23
+ "parameters": "7 billion",
24
+ "quant": "q4_0",
25
+ "type": "Falcon",
26
+ "systemPrompt": " ",
27
+ "description": "<strong>Best overall smaller model</strong><br><ul><li>Fast responses</li><li>Instruction based</li><li>Trained by TII<li>Finetuned by Nomic AI<li>Licensed for commercial use</ul>",
28
+ "url": "https://huggingface.co/nomic-ai/gpt4all-falcon-ggml/resolve/main/ggml-model-gpt4all-falcon-q4_0.bin",
29
+ "promptTemplate": "### Instruction:\n%1\n### Response:\n"
30
+ },
31
+ {
32
+ "order": "c",
33
+ "md5sum": "4acc146dd43eb02845c233c29289c7c5",
34
+ "name": "Hermes",
35
+ "filename": "nous-hermes-13b.ggmlv3.q4_0.bin",
36
+ "filesize": "8136777088",
37
+ "requires": "2.4.7",
38
+ "ramrequired": "16",
39
+ "parameters": "13 billion",
40
+ "quant": "q4_0",
41
+ "type": "LLaMA",
42
+ "systemPrompt": " ",
43
+ "description": "<strong>Extremely good model</strong><br><ul><li>Instruction based<li>Gives long responses<li>Curated with 300,000 uncensored instructions<li>Trained by Nous Research<li>Cannot be used commercially</ul>",
44
+ "url": "https://huggingface.co/TheBloke/Nous-Hermes-13B-GGML/resolve/main/nous-hermes-13b.ggmlv3.q4_0.bin",
45
+ "promptTemplate": "### Instruction:\n%1\n### Response:\n"
46
+ },
47
+ {
48
+ "order": "f",
49
+ "md5sum": "11d9f060ca24575a2c303bdc39952486",
50
+ "name": "Snoozy",
51
+ "filename": "GPT4All-13B-snoozy.ggmlv3.q4_0.bin",
52
+ "filesize": "8136770688",
53
+ "requires": "2.4.7",
54
+ "ramrequired": "16",
55
+ "parameters": "13 billion",
56
+ "quant": "q4_0",
57
+ "type": "LLaMA",
58
+ "systemPrompt": " ",
59
+ "description": "<strong>Very good overall model</strong><br><ul><li>Instruction based<li>Based on the same dataset as Groovy<li>Slower than Groovy, with higher quality responses<li>Trained by Nomic AI<li>Cannot be used commercially</ul>",
60
+ "url": "https://huggingface.co/TheBloke/GPT4All-13B-snoozy-GGML/resolve/main/GPT4All-13B-snoozy.ggmlv3.q4_0.bin"
61
+ },
62
+ {
63
+ "order": "h",
64
+ "md5sum": "e64e74375ce9d36a3d0af3db1523fd0a",
65
+ "name": "Mini Orca",
66
+ "filename": "orca-mini-7b.ggmlv3.q4_0.bin",
67
+ "filesize": "3791749248",
68
+ "requires": "2.4.7",
69
+ "ramrequired": "8",
70
+ "parameters": "7 billion",
71
+ "quant": "q4_0",
72
+ "type": "OpenLLaMa",
73
+ "description": "<strong>New model with novel dataset</strong><br><ul><li>Instruction based<li>Explain tuned datasets<li>Orca Research Paper dataset construction approaches<li>Licensed for commercial use</ul>",
74
+ "url": "https://huggingface.co/TheBloke/orca_mini_7B-GGML/resolve/main/orca-mini-7b.ggmlv3.q4_0.bin",
75
+ "promptTemplate": "### User:\n%1\n### Response:\n",
76
+ "systemPrompt": "### System:\nYou are an AI assistant that follows instruction extremely well. Help as much as you can.\n\n"
77
+ },
78
+ {
79
+ "order": "i",
80
+ "md5sum": "6a087f7f4598fad0bb70e6cb4023645e",
81
+ "name": "Mini Orca (Small)",
82
+ "filename": "orca-mini-3b.ggmlv3.q4_0.bin",
83
+ "filesize": "1928446208",
84
+ "requires": "2.4.7",
85
+ "ramrequired": "4",
86
+ "parameters": "3 billion",
87
+ "quant": "q4_0",
88
+ "type": "OpenLLaMa",
89
+ "description": "<strong>Small version of new model with novel dataset</strong><br><ul><li>Instruction based<li>Explain tuned datasets<li>Orca Research Paper dataset construction approaches<li>Licensed for commercial use</ul>",
90
+ "url": "https://huggingface.co/TheBloke/orca_mini_3B-GGML/resolve/main/orca-mini-3b.ggmlv3.q4_0.bin",
91
+ "promptTemplate": "### User:\n%1\n### Response:\n",
92
+ "systemPrompt": "### System:\nYou are an AI assistant that follows instruction extremely well. Help as much as you can.\n\n"
93
+ },
94
+ {
95
+ "order": "j",
96
+ "md5sum": "959b7f65b2d12fd1e3ff99e7493c7a3a",
97
+ "name": "Mini Orca (Large)",
98
+ "filename": "orca-mini-13b.ggmlv3.q4_0.bin",
99
+ "filesize": "7323329152",
100
+ "requires": "2.4.7",
101
+ "ramrequired": "16",
102
+ "parameters": "13 billion",
103
+ "quant": "q4_0",
104
+ "type": "OpenLLaMa",
105
+ "description": "<strong>Largest version of new model with novel dataset</strong><br><ul><li>Instruction based<li>Explain tuned datasets<li>Orca Research Paper dataset construction approaches<li>Licensed for commercial use</ul>",
106
+ "url": "https://huggingface.co/TheBloke/orca_mini_13B-GGML/resolve/main/orca-mini-13b.ggmlv3.q4_0.bin",
107
+ "promptTemplate": "### User:\n%1\n### Response:\n",
108
+ "systemPrompt": "### System:\nYou are an AI assistant that follows instruction extremely well. Help as much as you can.\n\n"
109
+ },
110
+ {
111
+ "order": "r",
112
+ "md5sum": "489d21fd48840dcb31e5f92f453f3a20",
113
+ "name": "Wizard Uncensored",
114
+ "filename": "wizardLM-13B-Uncensored.ggmlv3.q4_0.bin",
115
+ "filesize": "8136777088",
116
+ "requires": "2.4.7",
117
+ "ramrequired": "16",
118
+ "parameters": "13 billion",
119
+ "quant": "q4_0",
120
+ "type": "LLaMA",
121
+ "systemPrompt": " ",
122
+ "description": "<strong>Trained on uncensored assistant data and instruction data</strong><br><ul><li>Instruction based<li>Cannot be used commercially</ul>",
123
+ "url": "https://huggingface.co/TheBloke/WizardLM-13B-Uncensored-GGML/resolve/main/wizardLM-13B-Uncensored.ggmlv3.q4_0.bin"
124
+ },
125
+ {
126
+ "order": "s",
127
+ "md5sum": "615890cb571fcaa0f70b2f8d15ef809e",
128
+ "disableGUI": "true",
129
+ "name": "Replit",
130
+ "filename": "ggml-replit-code-v1-3b.bin",
131
+ "filesize": "5202046853",
132
+ "requires": "2.4.7",
133
+ "ramrequired": "4",
134
+ "parameters": "3 billion",
135
+ "quant": "f16",
136
+ "type": "Replit",
137
+ "systemPrompt": " ",
138
+ "promptTemplate": "%1",
139
+ "description": "<strong>Trained on subset of the Stack</strong><br><ul><li>Code completion based<li>Licensed for commercial use</ul>",
140
+ "url": "https://huggingface.co/nomic-ai/ggml-replit-code-v1-3b/resolve/main/ggml-replit-code-v1-3b.bin"
141
+ },
142
+ {
143
+ "order": "t",
144
+ "md5sum": "031bb5d5722c08d13e3e8eaf55c37391",
145
+ "disableGUI": "true",
146
+ "name": "Bert",
147
+ "filename": "ggml-all-MiniLM-L6-v2-f16.bin",
148
+ "filesize": "45521167",
149
+ "requires": "2.4.14",
150
+ "ramrequired": "1",
151
+ "parameters": "1 million",
152
+ "quant": "f16",
153
+ "type": "Bert",
154
+ "systemPrompt": " ",
155
+ "description": "<strong>Sbert</strong><br><ul><li>For embeddings"
156
+ },
157
+ {
158
+ "order": "u",
159
+ "md5sum": "379ee1bab9a7a9c27c2314daa097528e",
160
+ "disableGUI": "true",
161
+ "name": "Starcoder (Small)",
162
+ "filename": "starcoderbase-3b-ggml.bin",
163
+ "filesize": "7503121552",
164
+ "requires": "2.4.14",
165
+ "ramrequired": "8",
166
+ "parameters": "3 billion",
167
+ "quant": "f16",
168
+ "type": "Starcoder",
169
+ "systemPrompt": " ",
170
+ "promptTemplate": "%1",
171
+ "description": "<strong>Trained on subset of the Stack</strong><br><ul><li>Code completion based</ul>"
172
+ },
173
+ {
174
+ "order": "w",
175
+ "md5sum": "f981ab8fbd1ebbe4932ddd667c108ba7",
176
+ "disableGUI": "true",
177
+ "name": "Starcoder",
178
+ "filename": "starcoderbase-7b-ggml.bin",
179
+ "filesize": "17860448016",
180
+ "requires": "2.4.14",
181
+ "ramrequired": "16",
182
+ "parameters": "7 billion",
183
+ "quant": "f16",
184
+ "type": "Starcoder",
185
+ "systemPrompt": " ",
186
+ "promptTemplate": "%1",
187
+ "description": "<strong>Trained on subset of the Stack</strong><br><ul><li>Code completion based</ul>"
188
+ },
189
+ {
190
+ "order": "w",
191
+ "md5sum": "c7ebc61eec1779bddae1f2bcbf2007cc",
192
+ "name": "Llama-2-7B Chat",
193
+ "filename": "llama-2-7b-chat.ggmlv3.q4_0.bin",
194
+ "filesize": "3791725184",
195
+ "requires": "2.4.14",
196
+ "ramrequired": "8",
197
+ "parameters": "7 billion",
198
+ "quant": "q4_0",
199
+ "type": "LLaMA2",
200
+ "description": "<strong>New LLaMA2 model from Meta AI.</strong><br><ul><li>Fine-tuned for dialogue.<li>static model trained on an offline dataset<li>RLHF dataset<li>Licensed for commercial use</ul>",
201
+ "url": "https://huggingface.co/TheBloke/Llama-2-7B-Chat-GGML/resolve/main/llama-2-7b-chat.ggmlv3.q4_0.bin",
202
+ "promptTemplate": "[INST] %1 [/INST] ",
203
+ "systemPrompt": "[INST]<<SYS>>You are a helpful, respectful and honest assistant. Always answer as helpfully as possible, while being safe. Your answers should not include any harmful, unethical, racist, sexist, toxic, dangerous, or illegal content. Please ensure that your responses are socially unbiased and positive in nature. If a question does not make any sense, or is not factually coherent, explain why instead of answering something not correct. If you don't know the answer to a question, please don't share false information.<</SYS>>[/INST] "
204
+ }
205
+ ]
models2.json ADDED
@@ -0,0 +1,240 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "order": "a",
4
+ "md5sum": "48de9538c774188eb25a7e9ee024bbd3",
5
+ "name": "Mistral OpenOrca",
6
+ "filename": "mistral-7b-openorca.Q4_0.gguf",
7
+ "filesize": "4108927744",
8
+ "requires": "2.5.0",
9
+ "ramrequired": "8",
10
+ "parameters": "7 billion",
11
+ "quant": "q4_0",
12
+ "type": "Mistral",
13
+ "systemPrompt": " ",
14
+ "description": "<strong>Best overall fast chat model</strong><br><ul><li>Fast responses</li><li>Chat based model</li><li>Trained by Mistral AI<li>Finetuned on OpenOrca dataset curated via <a href=\"https://atlas.nomic.ai/\">Nomic Atlas</a><li>Licensed for commercial use</ul>",
15
+ "url": "https://gpt4all.io/models/gguf/mistral-7b-openorca.Q4_0.gguf"
16
+ },
17
+ {
18
+ "order": "b",
19
+ "md5sum": "97463be739b50525df56d33b26b00852",
20
+ "name": "Mistral Instruct",
21
+ "filename": "mistral-7b-instruct-v0.1.Q4_0.gguf",
22
+ "filesize": "4108916384",
23
+ "requires": "2.5.0",
24
+ "ramrequired": "8",
25
+ "parameters": "7 billion",
26
+ "quant": "q4_0",
27
+ "type": "Mistral",
28
+ "systemPrompt": " ",
29
+ "description": "<strong>Best overall fast instruction following model</strong><br><ul><li>Fast responses</li><li>Trained by Mistral AI<li>Uncensored</li><li>Licensed for commercial use</li></ul>",
30
+ "url": "https://gpt4all.io/models/gguf/mistral-7b-instruct-v0.1.Q4_0.gguf",
31
+ "promptTemplate": "[INST] %1 [/INST]"
32
+ },
33
+ {
34
+ "order": "c",
35
+ "md5sum": "31cb6d527bd3bfb5e73c2e9dfbc75033",
36
+ "name": "GPT4All Falcon",
37
+ "filename": "gpt4all-falcon-q4_0.gguf",
38
+ "filesize": "4210419040",
39
+ "requires": "2.5.0",
40
+ "ramrequired": "8",
41
+ "parameters": "7 billion",
42
+ "quant": "q4_0",
43
+ "type": "Falcon",
44
+ "systemPrompt": " ",
45
+ "description": "<strong>Very fast model with good quality</strong><br><ul><li>Fastest responses</li><li>Instruction based</li><li>Trained by TII<li>Finetuned by Nomic AI<li>Licensed for commercial use</ul>",
46
+ "url": "https://gpt4all.io/models/gguf/gpt4all-falcon-q4_0.gguf",
47
+ "promptTemplate": "### Instruction:\n%1\n### Response:\n"
48
+ },
49
+ {
50
+ "order": "e",
51
+ "md5sum": "00c8593ba57f5240f59662367b3ed4a5",
52
+ "name": "Orca 2 (Medium)",
53
+ "filename": "orca-2-7b.Q4_0.gguf",
54
+ "filesize": "3825824192",
55
+ "requires": "2.5.2",
56
+ "ramrequired": "8",
57
+ "parameters": "7 billion",
58
+ "quant": "q4_0",
59
+ "type": "LLaMA2",
60
+ "systemPrompt": " ",
61
+ "description": "<ul><li>Instruction based<li>Trained by Microsoft<li>Cannot be used commercially</ul>",
62
+ "url": "https://gpt4all.io/models/gguf/orca-2-7b.Q4_0.gguf"
63
+ },
64
+ {
65
+ "order": "f",
66
+ "md5sum": "3c0d63c4689b9af7baa82469a6f51a19",
67
+ "name": "Orca 2 (Full)",
68
+ "filename": "orca-2-13b.Q4_0.gguf",
69
+ "filesize": "7365856064",
70
+ "requires": "2.5.2",
71
+ "ramrequired": "16",
72
+ "parameters": "13 billion",
73
+ "quant": "q4_0",
74
+ "type": "LLaMA2",
75
+ "systemPrompt": " ",
76
+ "description": "<ul><li>Instruction based<li>Trained by Microsoft<li>Cannot be used commercially</ul>",
77
+ "url": "https://gpt4all.io/models/gguf/orca-2-13b.Q4_0.gguf"
78
+ },
79
+ {
80
+ "order": "g",
81
+ "md5sum": "5aff90007499bce5c64b1c0760c0b186",
82
+ "name": "Wizard v1.2",
83
+ "filename": "wizardlm-13b-v1.2.Q4_0.gguf",
84
+ "filesize": "7365834624",
85
+ "requires": "2.5.0",
86
+ "ramrequired": "16",
87
+ "parameters": "13 billion",
88
+ "quant": "q4_0",
89
+ "type": "LLaMA2",
90
+ "systemPrompt": " ",
91
+ "description": "<strong>Best overall larger model</strong><br><ul><li>Instruction based<li>Gives very long responses<li>Finetuned with only 1k of high-quality data<li>Trained by Microsoft and Peking University<li>Cannot be used commercially</ul>",
92
+ "url": "https://gpt4all.io/models/gguf/wizardlm-13b-v1.2.Q4_0.gguf"
93
+ },
94
+ {
95
+ "order": "h",
96
+ "md5sum": "3d12810391d04d1153b692626c0c6e16",
97
+ "name": "Hermes",
98
+ "filename": "nous-hermes-llama2-13b.Q4_0.gguf",
99
+ "filesize": "7366062080",
100
+ "requires": "2.5.0",
101
+ "ramrequired": "16",
102
+ "parameters": "13 billion",
103
+ "quant": "q4_0",
104
+ "type": "LLaMA2",
105
+ "systemPrompt": " ",
106
+ "description": "<strong>Extremely good model</strong><br><ul><li>Instruction based<li>Gives long responses<li>Curated with 300,000 uncensored instructions<li>Trained by Nous Research<li>Cannot be used commercially</ul>",
107
+ "url": "https://gpt4all.io/models/gguf/nous-hermes-llama2-13b.Q4_0.gguf",
108
+ "promptTemplate": "### Instruction:\n%1\n### Response:\n"
109
+ },
110
+ {
111
+ "order": "i",
112
+ "md5sum": "40388eb2f8d16bb5d08c96fdfaac6b2c",
113
+ "name": "Snoozy",
114
+ "filename": "gpt4all-13b-snoozy-q4_0.gguf",
115
+ "filesize": "7365834624",
116
+ "requires": "2.5.0",
117
+ "ramrequired": "16",
118
+ "parameters": "13 billion",
119
+ "quant": "q4_0",
120
+ "type": "LLaMA",
121
+ "systemPrompt": " ",
122
+ "description": "<strong>Very good overall model</strong><br><ul><li>Instruction based<li>Based on the same dataset as Groovy<li>Slower than Groovy, with higher quality responses<li>Trained by Nomic AI<li>Cannot be used commercially</ul>",
123
+ "url": "https://gpt4all.io/models/gguf/gpt4all-13b-snoozy-q4_0.gguf"
124
+ },
125
+ {
126
+ "order": "j",
127
+ "md5sum": "cf5e8f73747f9d7c6fe72a629808c1de",
128
+ "name": "MPT Chat",
129
+ "filename": "mpt-7b-chat-merges-q4_0.gguf",
130
+ "filesize": "3796133728",
131
+ "requires": "2.5.0",
132
+ "ramrequired": "8",
133
+ "parameters": "7 billion",
134
+ "quant": "q4_0",
135
+ "type": "MPT",
136
+ "description": "<strong>Good model with novel architecture</strong><br><ul><li>Fast responses<li>Chat based<li>Trained by Mosaic ML<li>Cannot be used commercially</ul>",
137
+ "url": "https://gpt4all.io/models/gguf/mpt-7b-chat-merges-q4_0.gguf",
138
+ "promptTemplate": "<|im_start|>user\n%1<|im_end|><|im_start|>assistant\n",
139
+ "systemPrompt": "<|im_start|>system\n- You are a helpful assistant chatbot trained by MosaicML.\n- You answer questions.\n- You are excited to be able to help the user, but will refuse to do anything that could be considered harmful to the user.\n- You are more than just an information source, you are also able to write poetry, short stories, and make jokes.<|im_end|>"
140
+ },
141
+ {
142
+ "order": "k",
143
+ "md5sum": "0e769317b90ac30d6e09486d61fefa26",
144
+ "name": "Mini Orca (Small)",
145
+ "filename": "orca-mini-3b-gguf2-q4_0.gguf",
146
+ "filesize": "1979946720",
147
+ "requires": "2.5.0",
148
+ "ramrequired": "4",
149
+ "parameters": "3 billion",
150
+ "quant": "q4_0",
151
+ "type": "OpenLLaMa",
152
+ "description": "<strong>Small version of new model with novel dataset</strong><br><ul><li>Instruction based<li>Explain tuned datasets<li>Orca Research Paper dataset construction approaches<li>Cannot be used commercially</ul>",
153
+ "url": "https://gpt4all.io/models/gguf/orca-mini-3b-gguf2-q4_0.gguf",
154
+ "promptTemplate": "### User:\n%1\n### Response:\n",
155
+ "systemPrompt": "### System:\nYou are an AI assistant that follows instruction extremely well. Help as much as you can.\n\n"
156
+ },
157
+ {
158
+ "order": "l",
159
+ "md5sum": "e30579a1b109882f10e2a5e75ea388fb",
160
+ "disableGUI": "true",
161
+ "name": "Replit",
162
+ "filename": "replit-code-v1_5-3b-q4_0.gguf",
163
+ "filesize": "1870449696",
164
+ "requires": "2.5.0",
165
+ "ramrequired": "4",
166
+ "parameters": "3 billion",
167
+ "quant": "q4_0",
168
+ "type": "Replit",
169
+ "systemPrompt": " ",
170
+ "promptTemplate": "%1",
171
+ "description": "<strong>Trained on subset of the Stack</strong><br><ul><li>Code completion based<li>Licensed for commercial use<li>WARNING: Not available for chat GUI</ul>",
172
+ "url": "https://gpt4all.io/models/gguf/replit-code-v1_5-3b-q4_0.gguf"
173
+ },
174
+ {
175
+ "order": "m",
176
+ "md5sum": "556fc3e13df42286997fb58e6f4c639f",
177
+ "disableGUI": "true",
178
+ "name": "Starcoder",
179
+ "filename": "starcoder-q4_0.gguf",
180
+ "filesize": "8987166880",
181
+ "requires": "2.5.0",
182
+ "ramrequired": "4",
183
+ "parameters": "7 billion",
184
+ "quant": "q4_0",
185
+ "type": "Starcoder",
186
+ "systemPrompt": " ",
187
+ "promptTemplate": "%1",
188
+ "description": "<strong>Trained on subset of the Stack</strong><br><ul><li>Code completion based<li>WARNING: Not available for chat GUI</ul>",
189
+ "url": "https://gpt4all.io/models/gguf/starcoder-q4_0.gguf"
190
+ },
191
+ {
192
+ "order": "n",
193
+ "md5sum": "e973dd26f0ffa6e46783feaea8f08c83",
194
+ "disableGUI": "true",
195
+ "name": "Rift coder",
196
+ "filename": "rift-coder-v0-7b-q4_0.gguf",
197
+ "filesize": "3825903776",
198
+ "requires": "2.5.0",
199
+ "ramrequired": "8",
200
+ "parameters": "7 billion",
201
+ "quant": "q4_0",
202
+ "type": "LLaMA",
203
+ "systemPrompt": " ",
204
+ "promptTemplate": "%1",
205
+ "description": "<strong>Trained on collection of Python and TypeScript</strong><br><ul><li>Code completion based<li>WARNING: Not available for chat GUI</li>",
206
+ "url": "https://gpt4all.io/models/gguf/rift-coder-v0-7b-q4_0.gguf"
207
+ },
208
+ {
209
+ "order": "o",
210
+ "md5sum": "e479e6f38b59afc51a470d1953a6bfc7",
211
+ "disableGUI": "true",
212
+ "name": "SBert",
213
+ "filename": "all-MiniLM-L6-v2-f16.gguf",
214
+ "filesize": "45887744",
215
+ "requires": "2.5.0",
216
+ "ramrequired": "1",
217
+ "parameters": "40 million",
218
+ "quant": "f16",
219
+ "type": "Bert",
220
+ "systemPrompt": " ",
221
+ "description": "<strong>LocalDocs text embeddings model</strong><br><ul><li>Necessary for LocalDocs feature<li>Used for retrieval augmented generation (RAG)",
222
+ "url": "https://gpt4all.io/models/gguf/all-MiniLM-L6-v2-f16.gguf"
223
+ },
224
+ {
225
+ "order": "p",
226
+ "md5sum": "919de4dd6f25351bcb0223790db1932d",
227
+ "name": "EM German Mistral",
228
+ "filename": "em_german_mistral_v01.Q4_0.gguf",
229
+ "filesize": "4108916352",
230
+ "requires": "2.5.0",
231
+ "ramrequired": "8",
232
+ "parameters": "7 billion",
233
+ "quant": "q4_0",
234
+ "type": "Mistral",
235
+ "description": "<strong>Mistral-based model for German-language applications</strong><br><ul><li>Fast responses</li><li>Chat based model</li><li>Trained by ellamind<li>Finetuned on German instruction and chat data</a><li>Licensed for commercial use</ul>",
236
+ "url": "https://huggingface.co/TheBloke/em_german_mistral_v01-GGUF/resolve/main/em_german_mistral_v01.Q4_0.gguf",
237
+ "promptTemplate": "USER: %1 ASSISTANT: ",
238
+ "systemPrompt": "Du bist ein hilfreicher Assistent. "
239
+ }
240
+ ]