@@ -53,9 +53,8 @@ envVars:
5353 "tools": true,
5454 "preprompt": "",
5555 "parameters": {
56- "temperature": 0.6,
57- "top_p": 0.9,
5856 "stop": ["<|endoftext|>", "<|eot_id|>"],
57+ "temperature": 0.6,
5958 "max_new_tokens": 1024,
6059 "truncate": 7167
6160 },
@@ -115,42 +114,10 @@ envVars:
115114 "logoUrl": "https://huggingface.co/datasets/huggingchat/models-logo/resolve/main/qwen-logo.png",
116115 "preprompt": "You are Qwen, created by Alibaba Cloud. You are a helpful assistant.",
117116 "parameters": {
118- "temperature": 0.6,
119117 "stop": ["<|endoftext|>", "<|im_end|>"],
120- "truncate": 28672,
121- "max_new_tokens": 3072,
122- },
123- "promptExamples": [
124- {
125- "title": "Write an email from bullet list",
126- "prompt": "As a restaurant owner, write a professional email to the supplier to get these products every week: \n\n- Wine (x10)\n- Eggs (x24)\n- Bread (x12)"
127- },
128- {
129- "title": "Code a snake game",
130- "prompt": "Code a basic snake game in python, give explanations for each step."
131- },
132- {
133- "title": "Assist in a task",
134- "prompt": "How do I make a delicious lemon cheesecake?"
135- }
136- ]
137- },
138- {
139- "name": "mistralai/Mixtral-8x7B-Instruct-v0.1",
140- "description": "A high-quality sparse mixture of experts model with open weights.",
141- "logoUrl": "https://huggingface.co/datasets/huggingchat/models-logo/resolve/main/mistral-logo.png",
142- "websiteUrl": "https://mistral.ai/news/mixtral-of-experts/",
143- "modelUrl": "https://huggingface.co/mistralai/Mixtral-8x7B-Instruct-v0.1",
144- "tokenizer": "mistralai/Mixtral-8x7B-Instruct-v0.1",
145- "preprompt": "",
146- "parameters": {
147118 "temperature": 0.6,
148- "top_p": 0.95,
149- "repetition_penalty": 1.2,
150- "top_k": 50,
151- "truncate": 24576,
152- "max_new_tokens": 8192,
153- "stop": ["</s>"]
119+ "truncate": 28672,
120+ "max_new_tokens": 3072
154121 },
155122 "promptExamples": [
156123 {
@@ -168,12 +135,12 @@ envVars:
168135 ]
169136 },
170137 {
171- "name": "NousResearch/Nous- Hermes-2-Mixtral-8x7B-DPO ",
172- "description": "Nous Hermes' strong flagship model trained on the Mixtral 8x7B .",
138+ "name": "NousResearch/Hermes-3-Llama-3.1-8B ",
139+ "description": "Nous Research's latest Hermes 3 release in 8B size. Follows instruction closely .",
173140 "logoUrl": "https://huggingface.co/datasets/huggingchat/models-logo/resolve/main/nous-logo.png",
174141 "websiteUrl": "https://nousresearch.com/",
175- "modelUrl": "https://huggingface.co/NousResearch/Nous- Hermes-2-Mixtral-8x7B-DPO ",
176- "tokenizer": "NousResearch/Nous- Hermes-2-Mixtral-8x7B-DPO ",
142+ "modelUrl": "https://huggingface.co/NousResearch/Hermes-3-Llama-3.1-8B ",
143+ "tokenizer": "NousResearch/Hermes-3-Llama-3.1-8B ",
177144 "promptExamples": [
178145 {
179146 "title": "Write an email from bullet list",
@@ -189,32 +156,26 @@ envVars:
189156 }
190157 ],
191158 "parameters": {
192- "temperature": 0.7,
193- "top_p": 0.95,
194- "repetition_penalty": 1,
195- "top_k": 50,
196- "truncate": 24576,
197- "max_new_tokens": 2048,
198- "stop": ["<|im_end|>"]
159+ "stop": ["<|im_end|>"],
160+ "temperature": 0.6,
161+ "truncate": 14336,
162+ "max_new_tokens": 1536
199163 }
200164 },
201165 {
202- "name": "mistralai/Mistral-7B-Instruct-v0.3",
203- "displayName": "mistralai/Mistral-7B-Instruct-v0.3",
166+ "name": "mistralai/Mistral-Nemo-Instruct-2407",
167+ "tokenizer": "mistralai/Mistral-Nemo-Instruct-2407",
168+ "displayName": "mistralai/Mistral-Nemo-Instruct-2407",
204169 "description": "A small model with good capabilities in language understanding and commonsense reasoning.",
205170 "logoUrl": "https://huggingface.co/datasets/huggingchat/models-logo/resolve/main/mistral-logo.png",
206- "websiteUrl": "https://mistral.ai/news/announcing-mistral-7b/",
207- "modelUrl": "https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.3",
208- "tokenizer": "mistralai/Mistral-7B-Instruct-v0.3",
171+ "websiteUrl": "https://mistral.ai/news/mistral-nemo/",
172+ "modelUrl": "https://huggingface.co/mistralai/Mistral-Nemo-Instruct-2407",
209173 "preprompt": "",
210174 "parameters": {
211- "temperature": 0.3,
212- "top_p": 0.95,
213- "repetition_penalty": 1.2,
214- "top_k": 50,
215- "truncate": 3072,
216- "max_new_tokens": 1024,
217- "stop": ["</s>"]
175+ "stop": ["</s>"],
176+ "temperature": 0.6,
177+ "truncate": 14336,
178+ "max_new_tokens": 1536
218179 },
219180 "promptExamples": [
220181 {
@@ -232,18 +193,18 @@ envVars:
232193 ]
233194 },
234195 {
235- "name": "microsoft/Phi-3-mini-4k -instruct",
236- "tokenizer": "microsoft/Phi-3-mini-4k -instruct",
196+ "name": "microsoft/Phi-3.5 -mini-instruct",
197+ "tokenizer": "microsoft/Phi-3.5 -mini-instruct",
237198 "description": "One of the best small models (3.8B parameters), super fast for simple tasks.",
238199 "logoUrl": "https://huggingface.co/datasets/huggingchat/models-logo/resolve/main/microsoft-logo.png",
239- "modelUrl": "https://huggingface.co/microsoft/Phi-3-mini-4k -instruct",
240- "websiteUrl": "https://azure .microsoft.com/en-us/ blog/introducing-phi-3-redefining-whats-possible-with- slms/",
200+ "modelUrl": "https://huggingface.co/microsoft/Phi-3.5 -mini-instruct",
201+ "websiteUrl": "https://techcommunity .microsoft.com/t5/ai-azure-ai-services- blog/discover-the-new-multi-lingual-high-quality-phi-3-5- slms/ba-p/4225280 /",
241202 "preprompt": "",
242203 "parameters": {
243204 "stop": ["<|end|>", "<|endoftext|>", "<|assistant|>"],
244- "temperature": 0.7 ,
245- "max_new_tokens ": 1024 ,
246- "truncate ": 3071
205+ "temperature": 0.6 ,
206+ "truncate ": 28672 ,
207+ "max_new_tokens ": 3072
247208 },
248209 "promptExamples": [
249210 {
@@ -268,7 +229,6 @@ envVars:
268229 },
269230 "parameters": {
270231 "temperature": 0.6,
271- "top_p": 0.9,
272232 "stop": ["<|endoftext|>", "<|eot_id|>"]
273233 },
274234 "unlisted": true
@@ -290,10 +250,32 @@ envVars:
290250 { "name": "meta-llama/Llama-2-70b-chat-hf" },
291251 { "name": "codellama/CodeLlama-70b-Instruct-hf" },
292252 { "name": "openchat/openchat-3.5-0106" },
293- { "name": "meta-llama/Meta-Llama-3-70B-Instruct"},
294- { "name": "meta-llama/Meta-Llama-3.1-405B-Instruct-FP8"},
295- { "name": "CohereForAI/c4ai-command-r-plus", "transferTo": "CohereForAI/c4ai-command-r-plus-08-2024"},
296- { "name": "01-ai/Yi-1.5-34B-Chat", "transferTo": "CohereForAI/c4ai-command-r-plus-08-2024"}
253+ { "name": "meta-llama/Meta-Llama-3-70B-Instruct" },
254+ { "name": "meta-llama/Meta-Llama-3.1-405B-Instruct-FP8" },
255+ {
256+ "name": "CohereForAI/c4ai-command-r-plus",
257+ "transferTo": "CohereForAI/c4ai-command-r-plus-08-2024"
258+ },
259+ {
260+ "name": "01-ai/Yi-1.5-34B-Chat",
261+ "transferTo": "CohereForAI/c4ai-command-r-plus-08-2024"
262+ },
263+ {
264+ "name": "mistralai/Mixtral-8x7B-Instruct-v0.1",
265+ "transferTo": "mistralai/Mistral-Nemo-Instruct-2407"
266+ },
267+ {
268+ "name": "NousResearch/Nous-Hermes-2-Mixtral-8x7B-DPO",
269+ "transferTo": "NousResearch/Hermes-3-Llama-3.1-8B"
270+ },
271+ {
272+ "name": "mistralai/Mistral-7B-Instruct-v0.3",
273+ "transferTo": "mistralai/Mistral-Nemo-Instruct-2407"
274+ },
275+ {
276+ "name": "microsoft/Phi-3-mini-4k-instruct",
277+ "transferTo": "microsoft/Phi-3.5-mini-instruct"
278+ }
297279 ]
298280 PUBLIC_ORIGIN : " https://huggingface.co"
299281 PUBLIC_SHARE_PREFIX : " https://hf.co/chat"
0 commit comments