|
150 | 150 | "flowise", |
151 | 151 | "groq", |
152 | 152 | "fireworks", |
153 | | - "continue-proxy" |
| 153 | + "continue-proxy", |
| 154 | + "cloudflare" |
154 | 155 | ], |
155 | 156 | "markdownEnumDescriptions": [ |
156 | 157 | "### OpenAI\nUse gpt-4, gpt-3.5-turbo, or any other OpenAI model. See [here](https://openai.com/product#made-for-developers) to obtain an API key.\n\n> [Reference](https://docs.continue.dev/reference/Model%20Providers/openai)", |
|
169 | 170 | "### Llamafile\nTo get started with llamafiles, find and download a binary on their [GitHub repo](https://github.com/Mozilla-Ocho/llamafile#binary-instructions). Then run it with the following command:\n\n```shell\nchmod +x ./llamafile\n./llamafile\n```\n> [Reference](https://docs.continue.dev/reference/Model%20Providers/llamafile)", |
170 | 171 | "### Mistral API\n\nTo get access to the Mistral API, obtain your API key from the [Mistral platform](https://docs.mistral.ai/)", |
171 | 172 | "### DeepInfra\n\n> [Reference](https://docs.continue.dev/reference/Model%20Providers/deepinfra)", |
172 | | - "### Continue Proxy\n\nContinue Enterprise users can use the Continue Proxy as a single point of access to models." |
| 173 | + "### Continue Proxy\n\nContinue Enterprise users can use the Continue Proxy as a single point of access to models.", |
| 174 | + "### Cloudflare Workers AI\n\n[Reference](https://developers.cloudflare.com/workers-ai/)" |
173 | 175 | ], |
174 | 176 | "type": "string" |
175 | 177 | }, |
|
313 | 315 | "gemini", |
314 | 316 | "huggingface-inference-api", |
315 | 317 | "replicate", |
316 | | - "together" |
| 318 | + "together", |
| 319 | + "cloudflare" |
317 | 320 | ] |
318 | 321 | } |
319 | 322 | }, |
|
359 | 362 | } |
360 | 363 | } |
361 | 364 | }, |
| 365 | + { |
| 366 | + "if": { |
| 367 | + "properties": { |
| 368 | + "provider": { |
| 369 | + "enum": ["cloudflare"] |
| 370 | + } |
| 371 | + }, |
| 372 | + "required": ["provider"] |
| 373 | + }, |
| 374 | + "then": { |
| 375 | + "properties": { |
| 376 | + "accountId": { |
| 377 | + "type": "string" |
| 378 | + }, |
| 379 | + "model": { |
| 380 | + "anyOf": [ |
| 381 | + { |
| 382 | + "enum": [ |
| 383 | + "@cf/meta/llama-3-8b-instruct", |
| 384 | + "@hf/thebloke/deepseek-coder-6.7b-instruct-awq", |
| 385 | + "@cf/deepseek-ai/deepseek-math-7b-instruct", |
| 386 | + "@cf/thebloke/discolm-german-7b-v1-awq", |
| 387 | + "@cf/tiiuae/falcon-7b-instruct", |
| 388 | + "@cf/google/gemma-2b-it-lora", |
| 389 | + "@hf/google/gemma-7b-it", |
| 390 | + "@cf/google/gemma-7b-it-lora", |
| 391 | + "@hf/nousresearch/hermes-2-pro-mistral-7b", |
| 392 | + "@cf/meta/llama-2-7b-chat-fp16", |
| 393 | + "@cf/meta/llama-2-7b-chat-int8", |
| 394 | + "@cf/meta-llama/llama-2-7b-chat-hf-lora", |
| 395 | + "@hf/thebloke/llama-2-13b-chat-awq", |
| 396 | + "@hf/thebloke/llamaguard-7b-awq", |
| 397 | + "@cf/mistral/mistral-7b-instruct-v0.1", |
| 398 | + "@hf/mistral/mistral-7b-instruct-v0.2", |
| 399 | + "@cf/mistral/mistral-7b-instruct-v0.2-lora", |
| 400 | + "@hf/thebloke/neural-chat-7b-v3-1-awq", |
| 401 | + "@cf/openchat/openchat-3.5-0106", |
| 402 | + "@hf/thebloke/openhermes-2.5-mistral-7b-awq", |
| 403 | + "@cf/microsoft/phi-2", |
| 404 | + "@cf/qwen/qwen1.5-0.5b-chat", |
| 405 | + "@cf/qwen/qwen1.5-1.8b-chat", |
| 406 | + "@cf/qwen/qwen1.5-7b-chat-awq", |
| 407 | + "@cf/qwen/qwen1.5-14b-chat-awq", |
| 408 | + "@cf/defog/sqlcoder-7b-2", |
| 409 | + "@hf/nexusflow/starling-lm-7b-beta", |
| 410 | + "@cf/tinyllama/tinyllama-1.1b-chat-v1.0", |
| 411 | + "@hf/thebloke/zephyr-7b-beta-awq", |
| 412 | + "@hf/thebloke/deepseek-coder-6.7b-base-awq" |
| 413 | + ] |
| 414 | + }, |
| 415 | + { |
| 416 | + "type": "string" |
| 417 | + } |
| 418 | + ] |
| 419 | + } |
| 420 | + } |
| 421 | + } |
| 422 | + }, |
362 | 423 | { |
363 | 424 | "if": { |
364 | 425 | "properties": { |
|
0 commit comments