From 20608319d8f580385733c24755a3ddc30cb33bc1 Mon Sep 17 00:00:00 2001 From: babayaga Date: Wed, 6 Aug 2025 17:10:32 +0200 Subject: [PATCH] maintainence love:) --- packages/kbot/dist-in/data/openai_models.json | 14 +- .../kbot/dist-in/data/openrouter_models.json | 8232 +++++++++-------- .../kbot/dist-in/src/models/cache/openai.ts | 2 +- .../dist-in/src/models/cache/openrouter.ts | 2 +- packages/kbot/schema.json | 2 +- packages/kbot/schema_ui.json | 2 +- .../kbot/src/models/cache/openai-models.ts | 2 - .../models/cache/openrouter-models-free.ts | 5 +- .../src/models/cache/openrouter-models.ts | 52 +- packages/kbot/src/zod_types.ts | 34 +- 10 files changed, 4256 insertions(+), 4091 deletions(-) diff --git a/packages/kbot/dist-in/data/openai_models.json b/packages/kbot/dist-in/data/openai_models.json index 2ff4fd1d..9430a7cb 100644 --- a/packages/kbot/dist-in/data/openai_models.json +++ b/packages/kbot/dist-in/data/openai_models.json @@ -1,5 +1,5 @@ { - "timestamp": 1753551692806, + "timestamp": 1754493028882, "models": [ { "id": "gpt-4-0613", @@ -193,18 +193,6 @@ "created": 1723515131, "owned_by": "system" }, - { - "id": "o1-preview-2024-09-12", - "object": "model", - "created": 1725648865, - "owned_by": "system" - }, - { - "id": "o1-preview", - "object": "model", - "created": 1725648897, - "owned_by": "system" - }, { "id": "o1-mini-2024-09-12", "object": "model", diff --git a/packages/kbot/dist-in/data/openrouter_models.json b/packages/kbot/dist-in/data/openrouter_models.json index 1200035a..957ad715 100644 --- a/packages/kbot/dist-in/data/openrouter_models.json +++ b/packages/kbot/dist-in/data/openrouter_models.json @@ -1,6 +1,458 @@ { - "timestamp": 1753551693103, + "timestamp": 1754493029370, "models": [ + { + "id": "openai/gpt-oss-120b", + "canonical_slug": "openai/gpt-oss-120b", + "hugging_face_id": "openai/gpt-oss-120b", + "name": "OpenAI: GPT OSS 120B", + "created": 1754414231, + "description": "gpt-oss-120b is an open-weight, 117B-parameter Mixture-of-Experts (MoE) language model from OpenAI designed for high-reasoning, agentic, and general-purpose production use cases. It activates 5.1B parameters per forward pass and is optimized to run on a single H100 GPU with native MXFP4 quantization. The model supports configurable reasoning depth, full chain-of-thought access, and native tool use, including function calling, browsing, and structured output generation.", + "context_length": 131072, + "architecture": { + "modality": "text->text", + "input_modalities": [ + "text" + ], + "output_modalities": [ + "text" + ], + "tokenizer": "GPT", + "instruct_type": null + }, + "pricing": { + "prompt": "0.0000001", + "completion": "0.0000005", + "request": "0", + "image": "0", + "audio": "0", + "web_search": "0", + "internal_reasoning": "0" + }, + "top_provider": { + "context_length": 131072, + "max_completion_tokens": null, + "is_moderated": false + }, + "per_request_limits": null, + "supported_parameters": [ + "frequency_penalty", + "include_reasoning", + "logit_bias", + "logprobs", + "max_tokens", + "min_p", + "presence_penalty", + "reasoning", + "repetition_penalty", + "response_format", + "seed", + "stop", + "structured_outputs", + "temperature", + "tool_choice", + "tools", + "top_k", + "top_logprobs", + "top_p" + ] + }, + { + "id": "openai/gpt-oss-20b", + "canonical_slug": "openai/gpt-oss-20b", + "hugging_face_id": "openai/gpt-oss-20b", + "name": "OpenAI: GPT OSS 20B", + "created": 1754414229, + "description": "gpt-oss-20b is an open-weight 21B parameter model released by OpenAI under the Apache 2.0 license. It uses a Mixture-of-Experts (MoE) architecture with 3.6B active parameters per forward pass, optimized for lower-latency inference and deployability on consumer or single-GPU hardware. The model is trained in OpenAI’s Harmony response format and supports reasoning level configuration, fine-tuning, and agentic capabilities including function calling, tool use, and structured outputs.", + "context_length": 131072, + "architecture": { + "modality": "text->text", + "input_modalities": [ + "text" + ], + "output_modalities": [ + "text" + ], + "tokenizer": "GPT", + "instruct_type": null + }, + "pricing": { + "prompt": "0.00000005", + "completion": "0.0000002", + "request": "0", + "image": "0", + "audio": "0", + "web_search": "0", + "internal_reasoning": "0" + }, + "top_provider": { + "context_length": 131072, + "max_completion_tokens": 32768, + "is_moderated": false + }, + "per_request_limits": null, + "supported_parameters": [ + "frequency_penalty", + "include_reasoning", + "logit_bias", + "logprobs", + "max_tokens", + "presence_penalty", + "reasoning", + "repetition_penalty", + "response_format", + "seed", + "stop", + "structured_outputs", + "temperature", + "tool_choice", + "tools", + "top_k", + "top_logprobs", + "top_p" + ] + }, + { + "id": "anthropic/claude-opus-4.1", + "canonical_slug": "anthropic/claude-4.1-opus-20250805", + "hugging_face_id": "", + "name": "Anthropic: Claude Opus 4.1", + "created": 1754411591, + "description": "Claude Opus 4.1 is an updated version of Anthropic’s flagship model, offering improved performance in coding, reasoning, and agentic tasks. It achieves 74.5% on SWE-bench Verified and shows notable gains in multi-file code refactoring, debugging precision, and detail-oriented reasoning. The model supports extended thinking up to 64K tokens and is optimized for tasks involving research, data analysis, and tool-assisted reasoning.", + "context_length": 200000, + "architecture": { + "modality": "text+image->text", + "input_modalities": [ + "image", + "text" + ], + "output_modalities": [ + "text" + ], + "tokenizer": "Claude", + "instruct_type": null + }, + "pricing": { + "prompt": "0.000015", + "completion": "0.000075", + "request": "0", + "image": "0.024", + "audio": "0", + "web_search": "0", + "internal_reasoning": "0", + "input_cache_read": "0.0000015", + "input_cache_write": "0.00001875" + }, + "top_provider": { + "context_length": 200000, + "max_completion_tokens": 32000, + "is_moderated": true + }, + "per_request_limits": null, + "supported_parameters": [ + "include_reasoning", + "max_tokens", + "reasoning", + "stop", + "temperature", + "tool_choice", + "tools" + ] + }, + { + "id": "openrouter/horizon-beta", + "canonical_slug": "openrouter/horizon-beta", + "hugging_face_id": "", + "name": "Horizon Beta", + "created": 1754089419, + "description": "This is a cloaked model provided to the community to gather feedback. This is an improved version of [Horizon Alpha](/openrouter/horizon-alpha)\n\nNote: It’s free to use during this testing period, and prompts and completions are logged by the model creator for feedback and training.", + "context_length": 256000, + "architecture": { + "modality": "text+image->text", + "input_modalities": [ + "image", + "text" + ], + "output_modalities": [ + "text" + ], + "tokenizer": "Other", + "instruct_type": null + }, + "pricing": { + "prompt": "0", + "completion": "0", + "request": "0", + "image": "0", + "audio": "0", + "web_search": "0", + "internal_reasoning": "0" + }, + "top_provider": { + "context_length": 256000, + "max_completion_tokens": 128000, + "is_moderated": false + }, + "per_request_limits": null, + "supported_parameters": [ + "max_tokens", + "response_format", + "structured_outputs", + "tool_choice", + "tools" + ] + }, + { + "id": "mistralai/codestral-2508", + "canonical_slug": "mistralai/codestral-2508", + "hugging_face_id": "", + "name": "Mistral: Codestral 2508", + "created": 1754079630, + "description": "Mistral's cutting-edge language model for coding released end of July 2025. Codestral specializes in low-latency, high-frequency tasks such as fill-in-the-middle (FIM), code correction and test generation.\n\n[Blog Post](https://mistral.ai/news/codestral-25-08)", + "context_length": 256000, + "architecture": { + "modality": "text->text", + "input_modalities": [ + "text" + ], + "output_modalities": [ + "text" + ], + "tokenizer": "Mistral", + "instruct_type": null + }, + "pricing": { + "prompt": "0.0000003", + "completion": "0.0000009", + "request": "0", + "image": "0", + "audio": "0", + "web_search": "0", + "internal_reasoning": "0" + }, + "top_provider": { + "context_length": 256000, + "max_completion_tokens": null, + "is_moderated": false + }, + "per_request_limits": null, + "supported_parameters": [ + "frequency_penalty", + "max_tokens", + "presence_penalty", + "response_format", + "seed", + "stop", + "structured_outputs", + "temperature", + "tool_choice", + "tools", + "top_p" + ] + }, + { + "id": "qwen/qwen3-30b-a3b-instruct-2507", + "canonical_slug": "qwen/qwen3-30b-a3b-instruct-2507", + "hugging_face_id": "Qwen/Qwen3-30B-A3B-Instruct-2507", + "name": "Qwen: Qwen3 30B A3B Instruct 2507", + "created": 1753806965, + "description": "Qwen3-30B-A3B-Instruct-2507 is a 30.5B-parameter mixture-of-experts language model from Qwen, with 3.3B active parameters per inference. It operates in non-thinking mode and is designed for high-quality instruction following, multilingual understanding, and agentic tool use. Post-trained on instruction data, it demonstrates competitive performance across reasoning (AIME, ZebraLogic), coding (MultiPL-E, LiveCodeBench), and alignment (IFEval, WritingBench) benchmarks. It outperforms its non-instruct variant on subjective and open-ended tasks while retaining strong factual and coding performance.", + "context_length": 131072, + "architecture": { + "modality": "text->text", + "input_modalities": [ + "text" + ], + "output_modalities": [ + "text" + ], + "tokenizer": "Qwen3", + "instruct_type": null + }, + "pricing": { + "prompt": "0.0000002", + "completion": "0.0000008", + "request": "0", + "image": "0", + "audio": "0", + "web_search": "0", + "internal_reasoning": "0" + }, + "top_provider": { + "context_length": 131072, + "max_completion_tokens": 32768, + "is_moderated": false + }, + "per_request_limits": null, + "supported_parameters": [ + "max_tokens", + "presence_penalty", + "response_format", + "seed", + "temperature", + "top_p" + ] + }, + { + "id": "z-ai/glm-4.5", + "canonical_slug": "z-ai/glm-4.5", + "hugging_face_id": "zai-org/GLM-4.5", + "name": "Z.AI: GLM 4.5", + "created": 1753471347, + "description": "GLM-4.5 is our latest flagship foundation model, purpose-built for agent-based applications. It leverages a Mixture-of-Experts (MoE) architecture and supports a context length of up to 128k tokens. GLM-4.5 delivers significantly enhanced capabilities in reasoning, code generation, and agent alignment. It supports a hybrid inference mode with two options, a \"thinking mode\" designed for complex reasoning and tool use, and a \"non-thinking mode\" optimized for instant responses. Users can control the reasoning behaviour with the `reasoning` `enabled` boolean. [Learn more in our docs](https://openrouter.ai/docs/use-cases/reasoning-tokens#enable-reasoning-with-default-config)", + "context_length": 131072, + "architecture": { + "modality": "text->text", + "input_modalities": [ + "text" + ], + "output_modalities": [ + "text" + ], + "tokenizer": "Other", + "instruct_type": null + }, + "pricing": { + "prompt": "0.0000002", + "completion": "0.0000002", + "request": "0", + "image": "0", + "audio": "0", + "web_search": "0", + "internal_reasoning": "0" + }, + "top_provider": { + "context_length": 131072, + "max_completion_tokens": null, + "is_moderated": false + }, + "per_request_limits": null, + "supported_parameters": [ + "frequency_penalty", + "include_reasoning", + "logit_bias", + "logprobs", + "max_tokens", + "min_p", + "presence_penalty", + "reasoning", + "repetition_penalty", + "response_format", + "seed", + "stop", + "temperature", + "tool_choice", + "tools", + "top_k", + "top_logprobs", + "top_p" + ] + }, + { + "id": "z-ai/glm-4.5-air:free", + "canonical_slug": "z-ai/glm-4.5-air", + "hugging_face_id": "zai-org/GLM-4.5-Air", + "name": "Z.AI: GLM 4.5 Air (free)", + "created": 1753471258, + "description": "GLM-4.5-Air is the lightweight variant of our latest flagship model family, also purpose-built for agent-centric applications. Like GLM-4.5, it adopts the Mixture-of-Experts (MoE) architecture but with a more compact parameter size. GLM-4.5-Air also supports hybrid inference modes, offering a \"thinking mode\" for advanced reasoning and tool use, and a \"non-thinking mode\" for real-time interaction. Users can control the reasoning behaviour with the `reasoning` `enabled` boolean. [Learn more in our docs](https://openrouter.ai/docs/use-cases/reasoning-tokens#enable-reasoning-with-default-config)", + "context_length": 131072, + "architecture": { + "modality": "text->text", + "input_modalities": [ + "text" + ], + "output_modalities": [ + "text" + ], + "tokenizer": "Other", + "instruct_type": null + }, + "pricing": { + "prompt": "0", + "completion": "0", + "request": "0", + "image": "0", + "audio": "0", + "web_search": "0", + "internal_reasoning": "0" + }, + "top_provider": { + "context_length": 131072, + "max_completion_tokens": null, + "is_moderated": false + }, + "per_request_limits": null, + "supported_parameters": [ + "frequency_penalty", + "include_reasoning", + "logit_bias", + "logprobs", + "max_tokens", + "min_p", + "presence_penalty", + "reasoning", + "repetition_penalty", + "seed", + "stop", + "temperature", + "top_k", + "top_logprobs", + "top_p" + ] + }, + { + "id": "z-ai/glm-4.5-air", + "canonical_slug": "z-ai/glm-4.5-air", + "hugging_face_id": "zai-org/GLM-4.5-Air", + "name": "Z.AI: GLM 4.5 Air", + "created": 1753471258, + "description": "GLM-4.5-Air is the lightweight variant of our latest flagship model family, also purpose-built for agent-centric applications. Like GLM-4.5, it adopts the Mixture-of-Experts (MoE) architecture but with a more compact parameter size. GLM-4.5-Air also supports hybrid inference modes, offering a \"thinking mode\" for advanced reasoning and tool use, and a \"non-thinking mode\" for real-time interaction. Users can control the reasoning behaviour with the `reasoning` `enabled` boolean. [Learn more in our docs](https://openrouter.ai/docs/use-cases/reasoning-tokens#enable-reasoning-with-default-config)", + "context_length": 128000, + "architecture": { + "modality": "text->text", + "input_modalities": [ + "text" + ], + "output_modalities": [ + "text" + ], + "tokenizer": "Other", + "instruct_type": null + }, + "pricing": { + "prompt": "0.0000002", + "completion": "0.0000011", + "request": "0", + "image": "0", + "audio": "0", + "web_search": "0", + "internal_reasoning": "0", + "input_cache_read": "0.00000003" + }, + "top_provider": { + "context_length": 128000, + "max_completion_tokens": 96000, + "is_moderated": false + }, + "per_request_limits": null, + "supported_parameters": [ + "frequency_penalty", + "include_reasoning", + "max_tokens", + "min_p", + "presence_penalty", + "reasoning", + "repetition_penalty", + "response_format", + "seed", + "stop", + "temperature", + "tool_choice", + "tools", + "top_k", + "top_p" + ] + }, { "id": "qwen/qwen3-235b-a22b-thinking-2507", "canonical_slug": "qwen/qwen3-235b-a22b-thinking-2507", @@ -21,10 +473,11 @@ "instruct_type": "qwen3" }, "pricing": { - "prompt": "0.0000001179", - "completion": "0.0000001179", + "prompt": "0.000000077968332", + "completion": "0.00000031202496", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -35,24 +488,24 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "tools", - "tool_choice", - "reasoning", - "include_reasoning", - "stop", "frequency_penalty", - "presence_penalty", - "seed", - "top_k", - "min_p", - "repetition_penalty", - "logprobs", + "include_reasoning", "logit_bias", + "logprobs", + "max_tokens", + "min_p", + "presence_penalty", + "reasoning", + "repetition_penalty", + "response_format", + "seed", + "stop", + "temperature", + "tool_choice", + "tools", + "top_k", "top_logprobs", - "response_format" + "top_p" ] }, { @@ -79,6 +532,7 @@ "completion": "0.0000001", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -91,60 +545,9 @@ "supported_parameters": [ "max_tokens", "temperature", - "top_p", - "tools", - "tool_choice" - ] - }, - { - "id": "qwen/qwen3-coder:free", - "canonical_slug": "qwen/qwen3-coder-480b-a35b-07-25", - "hugging_face_id": "Qwen/Qwen3-Coder-480B-A35B-Instruct", - "name": "Qwen: Qwen3 Coder (free)", - "created": 1753230546, - "description": "Qwen3-Coder-480B-A35B-Instruct is a Mixture-of-Experts (MoE) code generation model developed by the Qwen team. It is optimized for agentic coding tasks such as function calling, tool use, and long-context reasoning over repositories. The model features 480 billion total parameters, with 35 billion active per forward pass (8 out of 160 experts).\n\nPricing for the Alibaba endpoints varies by context length. Once a request is greater than 128k input tokens, the higher pricing is used.", - "context_length": 262144, - "architecture": { - "modality": "text->text", - "input_modalities": [ - "text" - ], - "output_modalities": [ - "text" - ], - "tokenizer": "Qwen3", - "instruct_type": null - }, - "pricing": { - "prompt": "0", - "completion": "0", - "request": "0", - "image": "0", - "web_search": "0", - "internal_reasoning": "0" - }, - "top_provider": { - "context_length": 262144, - "max_completion_tokens": null, - "is_moderated": false - }, - "per_request_limits": null, - "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "tools", "tool_choice", - "stop", - "frequency_penalty", - "presence_penalty", - "seed", - "top_k", - "min_p", - "repetition_penalty", - "logprobs", - "logit_bias", - "top_logprobs" + "tools", + "top_p" ] }, { @@ -167,10 +570,11 @@ "instruct_type": null }, "pricing": { - "prompt": "0.0000003", - "completion": "0.0000012", + "prompt": "0.0000001999188", + "completion": "0.000000800064", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -181,23 +585,23 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "stop", "frequency_penalty", - "presence_penalty", - "top_k", - "repetition_penalty", "logit_bias", - "min_p", - "response_format", - "tools", - "tool_choice", - "seed", - "structured_outputs", "logprobs", - "top_logprobs" + "max_tokens", + "min_p", + "presence_penalty", + "repetition_penalty", + "response_format", + "seed", + "stop", + "structured_outputs", + "temperature", + "tool_choice", + "tools", + "top_k", + "top_logprobs", + "top_p" ] }, { @@ -225,6 +629,7 @@ "completion": "0.0000002", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -235,16 +640,16 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", "frequency_penalty", + "max_tokens", "min_p", "presence_penalty", "repetition_penalty", "seed", "stop", - "top_k" + "temperature", + "top_k", + "top_p" ] }, { @@ -260,7 +665,8 @@ "input_modalities": [ "file", "image", - "text" + "text", + "audio" ], "output_modalities": [ "text" @@ -273,6 +679,7 @@ "completion": "0.0000004", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0", "input_cache_read": "0.000000025", @@ -285,65 +692,16 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "tools", - "tool_choice", - "reasoning", "include_reasoning", - "structured_outputs", - "response_format", - "seed" - ] - }, - { - "id": "qwen/qwen3-235b-a22b-2507:free", - "canonical_slug": "qwen/qwen3-235b-a22b-07-25", - "hugging_face_id": "Qwen/Qwen3-235B-A22B-Instruct-2507", - "name": "Qwen: Qwen3 235B A22B Instruct 2507 (free)", - "created": 1753119555, - "description": "Qwen3-235B-A22B-Instruct-2507 is a multilingual, instruction-tuned mixture-of-experts language model based on the Qwen3-235B architecture, with 22B active parameters per forward pass. It is optimized for general-purpose text generation, including instruction following, logical reasoning, math, code, and tool usage. The model supports a native 262K context length and does not implement \"thinking mode\" ( blocks).\n\nCompared to its base variant, this version delivers significant gains in knowledge coverage, long-context reasoning, coding benchmarks, and alignment with open-ended tasks. It is particularly strong on multilingual understanding, math reasoning (e.g., AIME, HMMT), and alignment evaluations like Arena-Hard and WritingBench.", - "context_length": 262144, - "architecture": { - "modality": "text->text", - "input_modalities": [ - "text" - ], - "output_modalities": [ - "text" - ], - "tokenizer": "Qwen3", - "instruct_type": null - }, - "pricing": { - "prompt": "0", - "completion": "0", - "request": "0", - "image": "0", - "web_search": "0", - "internal_reasoning": "0" - }, - "top_provider": { - "context_length": 262144, - "max_completion_tokens": null, - "is_moderated": false - }, - "per_request_limits": null, - "supported_parameters": [ "max_tokens", - "temperature", - "top_p", - "stop", - "frequency_penalty", - "presence_penalty", + "reasoning", + "response_format", "seed", - "top_k", - "min_p", - "repetition_penalty", - "logprobs", - "logit_bias", - "top_logprobs" + "structured_outputs", + "temperature", + "tool_choice", + "tools", + "top_p" ] }, { @@ -366,37 +724,38 @@ "instruct_type": null }, "pricing": { - "prompt": "0.00000012", - "completion": "0.00000059", + "prompt": "0.000000077968332", + "completion": "0.00000031202496", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, "top_provider": { "context_length": 262144, - "max_completion_tokens": 262144, + "max_completion_tokens": null, "is_moderated": false }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "seed", - "tools", - "tool_choice", - "stop", "frequency_penalty", + "logit_bias", + "logprobs", + "max_tokens", + "min_p", "presence_penalty", "repetition_penalty", "response_format", - "top_k", - "min_p", + "seed", + "stop", "structured_outputs", - "logit_bias", - "logprobs", - "top_logprobs" + "temperature", + "tool_choice", + "tools", + "top_k", + "top_logprobs", + "top_p" ] }, { @@ -423,6 +782,7 @@ "completion": "0.0000034", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -433,14 +793,14 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "reasoning", "include_reasoning", + "max_tokens", + "reasoning", + "seed", "stop", + "temperature", "top_k", - "seed" + "top_p" ] }, { @@ -467,6 +827,7 @@ "completion": "0", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -477,21 +838,21 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "tools", - "tool_choice", - "stop", "frequency_penalty", - "presence_penalty", - "seed", - "top_k", - "min_p", - "repetition_penalty", - "logprobs", "logit_bias", - "top_logprobs" + "logprobs", + "max_tokens", + "min_p", + "presence_penalty", + "repetition_penalty", + "seed", + "stop", + "temperature", + "tool_choice", + "tools", + "top_k", + "top_logprobs", + "top_p" ] }, { @@ -518,6 +879,7 @@ "completion": "0.00000249", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -528,23 +890,23 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "tools", - "tool_choice", - "structured_outputs", - "response_format", - "stop", "frequency_penalty", - "presence_penalty", - "top_k", - "repetition_penalty", "logit_bias", - "min_p", "logprobs", + "max_tokens", + "min_p", + "presence_penalty", + "repetition_penalty", + "response_format", + "seed", + "stop", + "structured_outputs", + "temperature", + "tool_choice", + "tools", + "top_k", "top_logprobs", - "seed" + "top_p" ] }, { @@ -572,6 +934,7 @@ "completion": "0.000000138", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -582,19 +945,19 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "reasoning", - "include_reasoning", - "stop", "frequency_penalty", - "presence_penalty", - "seed", - "top_k", + "include_reasoning", + "logit_bias", + "max_tokens", "min_p", + "presence_penalty", + "reasoning", "repetition_penalty", - "logit_bias" + "seed", + "stop", + "temperature", + "top_k", + "top_p" ] }, { @@ -621,6 +984,7 @@ "completion": "0.000002", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -631,17 +995,17 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "tools", - "tool_choice", - "stop", "frequency_penalty", + "max_tokens", "presence_penalty", "response_format", + "seed", + "stop", "structured_outputs", - "seed" + "temperature", + "tool_choice", + "tools", + "top_p" ] }, { @@ -668,6 +1032,7 @@ "completion": "0.00000028", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -678,20 +1043,20 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "tools", - "tool_choice", - "stop", "frequency_penalty", - "presence_penalty", - "response_format", - "structured_outputs", - "seed", + "max_tokens", "min_p", + "presence_penalty", "repetition_penalty", - "top_k" + "response_format", + "seed", + "stop", + "structured_outputs", + "temperature", + "tool_choice", + "tools", + "top_k", + "top_p" ] }, { @@ -718,6 +1083,7 @@ "completion": "0", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -728,15 +1094,15 @@ }, "per_request_limits": null, "supported_parameters": [ + "frequency_penalty", "max_tokens", - "temperature", - "top_p", - "structured_outputs", + "presence_penalty", "response_format", "stop", - "frequency_penalty", - "presence_penalty", - "top_k" + "structured_outputs", + "temperature", + "top_k", + "top_p" ] }, { @@ -764,6 +1130,7 @@ "completion": "0.000015", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0", "input_cache_read": "0.00000075" @@ -775,18 +1142,18 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "tools", - "tool_choice", - "reasoning", "include_reasoning", - "structured_outputs", - "seed", "logprobs", + "max_tokens", + "reasoning", + "response_format", + "seed", + "structured_outputs", + "temperature", + "tool_choice", + "tools", "top_logprobs", - "response_format" + "top_p" ] }, { @@ -813,6 +1180,7 @@ "completion": "0", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -823,14 +1191,14 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "stop", "frequency_penalty", + "max_tokens", "presence_penalty", + "response_format", "seed", - "response_format" + "stop", + "temperature", + "top_p" ] }, { @@ -857,6 +1225,7 @@ "completion": "0", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -867,21 +1236,21 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "reasoning", - "include_reasoning", - "stop", "frequency_penalty", - "presence_penalty", - "seed", - "top_k", - "min_p", - "repetition_penalty", - "logprobs", + "include_reasoning", "logit_bias", - "top_logprobs" + "logprobs", + "max_tokens", + "min_p", + "presence_penalty", + "reasoning", + "repetition_penalty", + "seed", + "stop", + "temperature", + "top_k", + "top_logprobs", + "top_p" ] }, { @@ -908,6 +1277,7 @@ "completion": "0.00000003", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -918,21 +1288,21 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "reasoning", - "include_reasoning", - "stop", "frequency_penalty", - "presence_penalty", - "seed", - "top_k", - "min_p", - "repetition_penalty", - "logprobs", + "include_reasoning", "logit_bias", - "top_logprobs" + "logprobs", + "max_tokens", + "min_p", + "presence_penalty", + "reasoning", + "repetition_penalty", + "seed", + "stop", + "temperature", + "top_k", + "top_logprobs", + "top_p" ] }, { @@ -959,6 +1329,7 @@ "completion": "0", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -969,72 +1340,21 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "reasoning", - "include_reasoning", - "stop", "frequency_penalty", - "presence_penalty", - "seed", - "top_k", - "min_p", - "repetition_penalty", - "logprobs", - "logit_bias", - "top_logprobs" - ] - }, - { - "id": "tngtech/deepseek-r1t2-chimera", - "canonical_slug": "tngtech/deepseek-r1t2-chimera", - "hugging_face_id": "tngtech/DeepSeek-TNG-R1T2-Chimera", - "name": "TNG: DeepSeek R1T2 Chimera", - "created": 1751986985, - "description": "DeepSeek-TNG-R1T2-Chimera is the second-generation Chimera model from TNG Tech. It is a 671 B-parameter mixture-of-experts text-generation model assembled from DeepSeek-AI’s R1-0528, R1, and V3-0324 checkpoints with an Assembly-of-Experts merge. The tri-parent design yields strong reasoning performance while running roughly 20 % faster than the original R1 and more than 2× faster than R1-0528 under vLLM, giving a favorable cost-to-intelligence trade-off. The checkpoint supports contexts up to 60 k tokens in standard use (tested to ~130 k) and maintains consistent token behaviour, making it suitable for long-context analysis, dialogue and other open-ended generation tasks.", - "context_length": 163840, - "architecture": { - "modality": "text->text", - "input_modalities": [ - "text" - ], - "output_modalities": [ - "text" - ], - "tokenizer": "DeepSeek", - "instruct_type": null - }, - "pricing": { - "prompt": "0.000000302", - "completion": "0.000000302", - "request": "0", - "image": "0", - "web_search": "0", - "internal_reasoning": "0" - }, - "top_provider": { - "context_length": 163840, - "max_completion_tokens": null, - "is_moderated": false - }, - "per_request_limits": null, - "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "reasoning", "include_reasoning", - "stop", - "frequency_penalty", - "presence_penalty", - "seed", - "top_k", - "min_p", - "repetition_penalty", - "logprobs", "logit_bias", - "top_logprobs" + "logprobs", + "max_tokens", + "min_p", + "presence_penalty", + "reasoning", + "repetition_penalty", + "seed", + "stop", + "temperature", + "top_k", + "top_logprobs", + "top_p" ] }, { @@ -1044,7 +1364,7 @@ "name": "Morph: Morph V3 Large", "created": 1751910858, "description": "Morph's high-accuracy apply model for complex code edits. 2000+ tokens/sec with 98% accuracy for precise code transformations.", - "context_length": 32000, + "context_length": 81920, "architecture": { "modality": "text->text", "input_modalities": [ @@ -1057,20 +1377,25 @@ "instruct_type": null }, "pricing": { - "prompt": "0.0000012", - "completion": "0.0000027", + "prompt": "0.0000009", + "completion": "0.0000019", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, "top_provider": { - "context_length": 32000, - "max_completion_tokens": 16000, + "context_length": 81920, + "max_completion_tokens": 38000, "is_moderated": false }, "per_request_limits": null, - "supported_parameters": [] + "supported_parameters": [ + "max_tokens", + "stop", + "temperature" + ] }, { "id": "morph/morph-v3-fast", @@ -1079,7 +1404,7 @@ "name": "Morph: Morph V3 Fast", "created": 1751910002, "description": "Morph's fastest apply model for code edits. 4500+ tokens/sec with 96% accuracy for rapid code transformations.", - "context_length": 32000, + "context_length": 81920, "architecture": { "modality": "text->text", "input_modalities": [ @@ -1092,20 +1417,25 @@ "instruct_type": null }, "pricing": { - "prompt": "0.0000012", - "completion": "0.0000027", + "prompt": "0.0000009", + "completion": "0.0000019", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, "top_provider": { - "context_length": 32000, - "max_completion_tokens": 16000, + "context_length": 81920, + "max_completion_tokens": 38000, "is_moderated": false }, "per_request_limits": null, - "supported_parameters": [] + "supported_parameters": [ + "max_tokens", + "stop", + "temperature" + ] }, { "id": "baidu/ernie-4.5-300b-a47b", @@ -1131,6 +1461,7 @@ "completion": "0.0000011", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -1141,17 +1472,17 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "stop", "frequency_penalty", - "presence_penalty", - "seed", - "top_k", + "logit_bias", + "max_tokens", "min_p", + "presence_penalty", "repetition_penalty", - "logit_bias" + "seed", + "stop", + "temperature", + "top_k", + "top_p" ] }, { @@ -1161,7 +1492,7 @@ "name": "TheDrummer: Anubis 70B V1.1", "created": 1751208347, "description": "TheDrummer's Anubis v1.1 is an unaligned, creative Llama 3.3 70B model focused on providing character-driven roleplay & stories. It excels at gritty, visceral prose, unique character adherence, and coherent narratives, while maintaining the instruction following Llama 3.3 70B is known for.", - "context_length": 131072, + "context_length": 16384, "architecture": { "modality": "text->text", "input_modalities": [ @@ -1174,30 +1505,33 @@ "instruct_type": null }, "pricing": { - "prompt": "0.0000005", - "completion": "0.0000008", + "prompt": "0.0000004", + "completion": "0.0000007", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, "top_provider": { - "context_length": 131072, - "max_completion_tokens": 131072, + "context_length": 16384, + "max_completion_tokens": null, "is_moderated": false }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", "frequency_penalty", + "max_tokens", "min_p", "presence_penalty", "repetition_penalty", + "response_format", "seed", "stop", - "top_k" + "structured_outputs", + "temperature", + "top_k", + "top_p" ] }, { @@ -1207,7 +1541,7 @@ "name": "Inception: Mercury", "created": 1750973026, "description": "Mercury is the first diffusion large language model (dLLM). Applying a breakthrough discrete diffusion approach, the model runs 5-10x faster than even speed optimized models like GPT-4.1 Nano and Claude 3.5 Haiku while matching their performance. Mercury's speed enables developers to provide responsive user experiences, including with voice agents, search interfaces, and chatbots. Read more in the blog post here. ", - "context_length": 32000, + "context_length": 128000, "architecture": { "modality": "text->text", "input_modalities": [ @@ -1224,57 +1558,30 @@ "completion": "0.000001", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, "top_provider": { - "context_length": 32000, - "max_completion_tokens": 16000, + "context_length": 128000, + "max_completion_tokens": 16384, "is_moderated": false }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", "frequency_penalty", + "max_tokens", "presence_penalty", - "stop" + "response_format", + "stop", + "structured_outputs", + "temperature", + "tool_choice", + "tools", + "top_k", + "top_p" ] }, - { - "id": "morph/morph-v2", - "canonical_slug": "morph/morph-v2", - "hugging_face_id": "", - "name": "Morph: Fast Apply", - "created": 1750946108, - "description": "Morph Apply is a specialized code-patching LLM that merges AI-suggested edits straight into your source files. It can apply updates from GPT-4o, Claude, and others into your files at 4000+ tokens per second.\n\nThe model requires the prompt to be in the following format: \n${originalCode}\\n${updateSnippet}\n\nLearn more about this model in their [documentation](https://docs.morphllm.com/)", - "context_length": 32000, - "architecture": { - "modality": "text->text", - "input_modalities": [ - "text" - ], - "output_modalities": [ - "text" - ], - "tokenizer": "Other", - "instruct_type": null - }, - "pricing": { - "prompt": "0.0000012", - "completion": "0.0000027", - "request": "0", - "image": "0", - "web_search": "0", - "internal_reasoning": "0" - }, - "top_provider": { - "context_length": 32000, - "max_completion_tokens": 16000, - "is_moderated": false - }, - "per_request_limits": null, - "supported_parameters": [] - }, { "id": "mistralai/mistral-small-3.2-24b-instruct:free", "canonical_slug": "mistralai/mistral-small-3.2-24b-instruct-2506", @@ -1282,7 +1589,7 @@ "name": "Mistral: Mistral Small 3.2 24B (free)", "created": 1750443016, "description": "Mistral-Small-3.2-24B-Instruct-2506 is an updated 24B parameter model from Mistral optimized for instruction following, repetition reduction, and improved function calling. Compared to the 3.1 release, version 3.2 significantly improves accuracy on WildBench and Arena Hard, reduces infinite generations, and delivers gains in tool use and structured output tasks.\n\nIt supports image and text inputs with structured outputs, function/tool calling, and strong performance across coding (HumanEval+, MBPP), STEM (MMLU, MATH, GPQA), and vision benchmarks (ChartQA, DocVQA).", - "context_length": 96000, + "context_length": 131072, "architecture": { "modality": "text+image->text", "input_modalities": [ @@ -1300,32 +1607,33 @@ "completion": "0", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, "top_provider": { - "context_length": 96000, + "context_length": 131072, "max_completion_tokens": null, "is_moderated": false }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "tools", - "tool_choice", - "structured_outputs", - "stop", "frequency_penalty", - "presence_penalty", - "seed", - "top_k", - "min_p", - "repetition_penalty", - "logprobs", "logit_bias", - "top_logprobs" + "logprobs", + "max_tokens", + "min_p", + "presence_penalty", + "repetition_penalty", + "seed", + "stop", + "structured_outputs", + "temperature", + "tool_choice", + "tools", + "top_k", + "top_logprobs", + "top_p" ] }, { @@ -1335,7 +1643,7 @@ "name": "Mistral: Mistral Small 3.2 24B", "created": 1750443016, "description": "Mistral-Small-3.2-24B-Instruct-2506 is an updated 24B parameter model from Mistral optimized for instruction following, repetition reduction, and improved function calling. Compared to the 3.1 release, version 3.2 significantly improves accuracy on WildBench and Arena Hard, reduces infinite generations, and delivers gains in tool use and structured output tasks.\n\nIt supports image and text inputs with structured outputs, function/tool calling, and strong performance across coding (HumanEval+, MBPP), STEM (MMLU, MATH, GPQA), and vision benchmarks (ChartQA, DocVQA).", - "context_length": 128000, + "context_length": 131072, "architecture": { "modality": "text+image->text", "input_modalities": [ @@ -1349,36 +1657,38 @@ "instruct_type": null }, "pricing": { - "prompt": "0.00000005", - "completion": "0.0000001", + "prompt": "0.00000001999188", + "completion": "0.0000000800064", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, "top_provider": { - "context_length": 128000, + "context_length": 131072, "max_completion_tokens": null, "is_moderated": false }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "stop", "frequency_penalty", - "presence_penalty", "logit_bias", "logprobs", - "seed", + "max_tokens", + "min_p", + "presence_penalty", "repetition_penalty", - "tools", - "tool_choice", "response_format", + "seed", + "stop", "structured_outputs", + "temperature", + "tool_choice", + "tools", "top_k", - "min_p" + "top_logprobs", + "top_p" ] }, { @@ -1405,6 +1715,7 @@ "completion": "0.00000165", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -1415,22 +1726,22 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "tools", - "tool_choice", - "reasoning", - "include_reasoning", - "structured_outputs", - "stop", "frequency_penalty", - "presence_penalty", - "seed", - "top_k", + "include_reasoning", + "logit_bias", + "max_tokens", "min_p", + "presence_penalty", + "reasoning", "repetition_penalty", - "logit_bias" + "seed", + "stop", + "structured_outputs", + "temperature", + "tool_choice", + "tools", + "top_k", + "top_p" ] }, { @@ -1459,6 +1770,7 @@ "completion": "0.0000004", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0", "input_cache_read": "0.000000025", @@ -1471,16 +1783,17 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "tools", - "tool_choice", - "reasoning", "include_reasoning", - "structured_outputs", + "max_tokens", + "reasoning", "response_format", - "stop" + "seed", + "stop", + "structured_outputs", + "temperature", + "tool_choice", + "tools", + "top_p" ] }, { @@ -1496,7 +1809,8 @@ "input_modalities": [ "file", "image", - "text" + "text", + "audio" ], "output_modalities": [ "text" @@ -1509,6 +1823,7 @@ "completion": "0.0000025", "request": "0", "image": "0.001238", + "audio": "0", "web_search": "0", "internal_reasoning": "0", "input_cache_read": "0.000000075", @@ -1521,16 +1836,17 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "tools", - "tool_choice", - "reasoning", "include_reasoning", - "structured_outputs", + "max_tokens", + "reasoning", "response_format", - "stop" + "seed", + "stop", + "structured_outputs", + "temperature", + "tool_choice", + "tools", + "top_p" ] }, { @@ -1559,6 +1875,7 @@ "completion": "0.00001", "request": "0", "image": "0.00516", + "audio": "0", "web_search": "0", "internal_reasoning": "0", "input_cache_read": "0.00000031", @@ -1571,16 +1888,17 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "tools", - "tool_choice", - "reasoning", "include_reasoning", - "structured_outputs", + "max_tokens", + "reasoning", "response_format", - "stop" + "seed", + "stop", + "structured_outputs", + "temperature", + "tool_choice", + "tools", + "top_p" ] }, { @@ -1607,6 +1925,7 @@ "completion": "0", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -1617,21 +1936,21 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "reasoning", - "include_reasoning", - "stop", "frequency_penalty", - "presence_penalty", - "seed", - "top_k", - "min_p", - "repetition_penalty", - "logprobs", + "include_reasoning", "logit_bias", - "top_logprobs" + "logprobs", + "max_tokens", + "min_p", + "presence_penalty", + "reasoning", + "repetition_penalty", + "seed", + "stop", + "temperature", + "top_k", + "top_logprobs", + "top_p" ] }, { @@ -1660,6 +1979,7 @@ "completion": "0.00008", "request": "0", "image": "0.0153", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -1670,12 +1990,12 @@ }, "per_request_limits": null, "supported_parameters": [ - "tools", - "tool_choice", - "seed", "max_tokens", "response_format", - "structured_outputs" + "seed", + "structured_outputs", + "tool_choice", + "tools" ] }, { @@ -1702,6 +2022,7 @@ "completion": "0.0000005", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0", "input_cache_read": "0.000000075" @@ -1713,19 +2034,19 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "tools", - "tool_choice", - "reasoning", "include_reasoning", - "structured_outputs", - "stop", - "seed", "logprobs", + "max_tokens", + "reasoning", + "response_format", + "seed", + "stop", + "structured_outputs", + "temperature", + "tool_choice", + "tools", "top_logprobs", - "response_format" + "top_p" ] }, { @@ -1752,6 +2073,7 @@ "completion": "0.000015", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0", "input_cache_read": "0.00000075" @@ -1763,19 +2085,19 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "tools", - "tool_choice", - "structured_outputs", - "stop", "frequency_penalty", - "presence_penalty", - "seed", "logprobs", + "max_tokens", + "presence_penalty", + "response_format", + "seed", + "stop", + "structured_outputs", + "temperature", + "tool_choice", + "tools", "top_logprobs", - "response_format" + "top_p" ] }, { @@ -1802,6 +2124,7 @@ "completion": "0.0000015", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -1812,19 +2135,19 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "tools", - "tool_choice", - "reasoning", - "include_reasoning", - "structured_outputs", - "response_format", - "stop", "frequency_penalty", + "include_reasoning", + "max_tokens", "presence_penalty", - "seed" + "reasoning", + "response_format", + "seed", + "stop", + "structured_outputs", + "temperature", + "tool_choice", + "tools", + "top_p" ] }, { @@ -1851,6 +2174,7 @@ "completion": "0.000005", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -1861,19 +2185,19 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "tools", - "tool_choice", - "reasoning", - "include_reasoning", - "structured_outputs", - "response_format", - "stop", "frequency_penalty", + "include_reasoning", + "max_tokens", "presence_penalty", - "seed" + "reasoning", + "response_format", + "seed", + "stop", + "structured_outputs", + "temperature", + "tool_choice", + "tools", + "top_p" ] }, { @@ -1900,6 +2224,7 @@ "completion": "0.000005", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -1910,19 +2235,19 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "tools", - "tool_choice", - "reasoning", - "include_reasoning", - "structured_outputs", - "response_format", - "stop", "frequency_penalty", + "include_reasoning", + "max_tokens", "presence_penalty", - "seed" + "reasoning", + "response_format", + "seed", + "stop", + "structured_outputs", + "temperature", + "tool_choice", + "tools", + "top_p" ] }, { @@ -1951,6 +2276,7 @@ "completion": "0.00001", "request": "0", "image": "0.00516", + "audio": "0", "web_search": "0", "internal_reasoning": "0", "input_cache_read": "0.00000031", @@ -1963,16 +2289,17 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "tools", - "tool_choice", - "reasoning", "include_reasoning", - "structured_outputs", + "max_tokens", + "reasoning", "response_format", - "stop" + "seed", + "stop", + "structured_outputs", + "temperature", + "tool_choice", + "tools", + "top_p" ] }, { @@ -1999,6 +2326,7 @@ "completion": "0.0000002", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -2009,12 +2337,12 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "reasoning", "include_reasoning", - "seed" + "max_tokens", + "reasoning", + "seed", + "temperature", + "top_p" ] }, { @@ -2041,6 +2369,7 @@ "completion": "0", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -2051,21 +2380,21 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "reasoning", - "include_reasoning", - "stop", "frequency_penalty", - "presence_penalty", - "seed", - "top_k", - "min_p", - "repetition_penalty", - "logprobs", + "include_reasoning", "logit_bias", - "top_logprobs" + "logprobs", + "max_tokens", + "min_p", + "presence_penalty", + "reasoning", + "repetition_penalty", + "seed", + "stop", + "temperature", + "top_k", + "top_logprobs", + "top_p" ] }, { @@ -2092,6 +2421,7 @@ "completion": "0.00000002", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -2102,19 +2432,21 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "reasoning", - "include_reasoning", "frequency_penalty", + "include_reasoning", + "logit_bias", + "logprobs", + "max_tokens", "min_p", "presence_penalty", + "reasoning", "repetition_penalty", "seed", "stop", + "temperature", "top_k", - "logit_bias" + "top_logprobs", + "top_p" ] }, { @@ -2141,6 +2473,7 @@ "completion": "0", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -2151,23 +2484,21 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "reasoning", - "include_reasoning", - "stop", "frequency_penalty", - "presence_penalty", - "seed", - "top_k", - "min_p", - "repetition_penalty", - "logprobs", + "include_reasoning", "logit_bias", + "logprobs", + "max_tokens", + "min_p", + "presence_penalty", + "reasoning", + "repetition_penalty", + "seed", + "stop", + "temperature", + "top_k", "top_logprobs", - "structured_outputs", - "response_format" + "top_p" ] }, { @@ -2190,10 +2521,11 @@ "instruct_type": "deepseek-r1" }, "pricing": { - "prompt": "0.000000272", - "completion": "0.000000272", + "prompt": "0.00000017992692", + "completion": "0.0000007200576", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -2204,25 +2536,25 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "reasoning", - "include_reasoning", - "stop", "frequency_penalty", - "presence_penalty", - "top_k", - "repetition_penalty", + "include_reasoning", "logit_bias", - "min_p", - "response_format", - "structured_outputs", - "seed", - "tools", - "tool_choice", "logprobs", - "top_logprobs" + "max_tokens", + "min_p", + "presence_penalty", + "reasoning", + "repetition_penalty", + "response_format", + "seed", + "stop", + "structured_outputs", + "temperature", + "tool_choice", + "tools", + "top_k", + "top_logprobs", + "top_p" ] }, { @@ -2249,6 +2581,7 @@ "completion": "0", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -2259,68 +2592,19 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "stop", "frequency_penalty", - "presence_penalty", - "seed", - "top_k", - "min_p", - "repetition_penalty", - "logprobs", "logit_bias", - "top_logprobs" - ] - }, - { - "id": "sarvamai/sarvam-m", - "canonical_slug": "sarvamai/sarvam-m", - "hugging_face_id": "sarvamai/sarvam-m", - "name": "Sarvam AI: Sarvam-M", - "created": 1748188413, - "description": "Sarvam-M is a 24 B-parameter, instruction-tuned derivative of Mistral-Small-3.1-24B-Base-2503, post-trained on English plus eleven major Indic languages (bn, hi, kn, gu, mr, ml, or, pa, ta, te). The model introduces a dual-mode interface: “non-think” for low-latency chat and a optional “think” phase that exposes chain-of-thought tokens for more demanding reasoning, math, and coding tasks. \n\nBenchmark reports show solid gains versus similarly sized open models on Indic-language QA, GSM-8K math, and SWE-Bench coding, making Sarvam-M a practical general-purpose choice for multilingual conversational agents as well as analytical workloads that mix English, native Indic scripts, or romanized text.", - "context_length": 32768, - "architecture": { - "modality": "text->text", - "input_modalities": [ - "text" - ], - "output_modalities": [ - "text" - ], - "tokenizer": "Other", - "instruct_type": null - }, - "pricing": { - "prompt": "0.000000022", - "completion": "0.000000022", - "request": "0", - "image": "0", - "web_search": "0", - "internal_reasoning": "0" - }, - "top_provider": { - "context_length": 32768, - "max_completion_tokens": null, - "is_moderated": false - }, - "per_request_limits": null, - "supported_parameters": [ + "logprobs", "max_tokens", - "temperature", - "top_p", - "stop", - "frequency_penalty", - "presence_penalty", - "seed", - "top_k", "min_p", + "presence_penalty", "repetition_penalty", - "logprobs", - "logit_bias", - "top_logprobs" + "seed", + "stop", + "temperature", + "top_k", + "top_logprobs", + "top_p" ] }, { @@ -2347,6 +2631,7 @@ "completion": "0.000001", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -2357,18 +2642,18 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "reasoning", - "include_reasoning", "frequency_penalty", + "include_reasoning", + "max_tokens", "min_p", "presence_penalty", + "reasoning", "repetition_penalty", "seed", "stop", - "top_k" + "temperature", + "top_k", + "top_p" ] }, { @@ -2396,6 +2681,7 @@ "completion": "0.000075", "request": "0", "image": "0.024", + "audio": "0", "web_search": "0", "internal_reasoning": "0", "input_cache_read": "0.0000015", @@ -2408,15 +2694,15 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "stop", - "reasoning", "include_reasoning", - "tools", + "max_tokens", + "reasoning", + "stop", + "temperature", "tool_choice", - "top_p", - "top_k" + "tools", + "top_k", + "top_p" ] }, { @@ -2444,6 +2730,7 @@ "completion": "0.000015", "request": "0", "image": "0.0048", + "audio": "0", "web_search": "0", "internal_reasoning": "0", "input_cache_read": "0.0000003", @@ -2456,15 +2743,15 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "stop", - "reasoning", "include_reasoning", - "tools", + "max_tokens", + "reasoning", + "stop", + "temperature", "tool_choice", - "top_p", - "top_k" + "tools", + "top_k", + "top_p" ] }, { @@ -2491,6 +2778,7 @@ "completion": "0", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -2501,21 +2789,21 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "tools", - "tool_choice", - "stop", "frequency_penalty", - "presence_penalty", - "seed", - "top_k", - "min_p", - "repetition_penalty", - "logprobs", "logit_bias", - "top_logprobs" + "logprobs", + "max_tokens", + "min_p", + "presence_penalty", + "repetition_penalty", + "seed", + "stop", + "temperature", + "tool_choice", + "tools", + "top_k", + "top_logprobs", + "top_p" ] }, { @@ -2538,10 +2826,11 @@ "instruct_type": null }, "pricing": { - "prompt": "0.00000003", - "completion": "0.00000003", + "prompt": "0.00000001999188", + "completion": "0.0000000800064", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -2552,23 +2841,23 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "stop", "frequency_penalty", + "logit_bias", + "logprobs", + "max_tokens", + "min_p", "presence_penalty", "repetition_penalty", "response_format", - "top_k", "seed", - "min_p", - "tools", + "stop", + "structured_outputs", + "temperature", "tool_choice", - "logit_bias", - "logprobs", + "tools", + "top_k", "top_logprobs", - "structured_outputs" + "top_p" ] }, { @@ -2595,6 +2884,7 @@ "completion": "0", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -2605,14 +2895,14 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "stop", "frequency_penalty", + "max_tokens", "presence_penalty", + "response_format", "seed", - "response_format" + "stop", + "temperature", + "top_p" ] }, { @@ -2639,6 +2929,7 @@ "completion": "0.00000004", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -2649,17 +2940,16 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "stop", "frequency_penalty", - "presence_penalty", - "top_k", - "repetition_penalty", "logit_bias", + "max_tokens", "min_p", - "response_format" + "presence_penalty", + "repetition_penalty", + "stop", + "temperature", + "top_k", + "top_p" ] }, { @@ -2687,6 +2977,7 @@ "completion": "0.000006", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0", "input_cache_read": "0.000000375" @@ -2698,12 +2989,12 @@ }, "per_request_limits": null, "supported_parameters": [ - "tools", - "tool_choice", - "seed", "max_tokens", "response_format", - "structured_outputs" + "seed", + "structured_outputs", + "tool_choice", + "tools" ] }, { @@ -2726,10 +3017,11 @@ "instruct_type": null }, "pricing": { - "prompt": "0.000000141", - "completion": "0.000000141", + "prompt": "0.00000009329544", + "completion": "0.0000003733632", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -2740,21 +3032,21 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "reasoning", - "include_reasoning", - "stop", "frequency_penalty", - "presence_penalty", - "seed", - "top_k", - "min_p", - "repetition_penalty", - "logprobs", + "include_reasoning", "logit_bias", - "top_logprobs" + "logprobs", + "max_tokens", + "min_p", + "presence_penalty", + "reasoning", + "repetition_penalty", + "seed", + "stop", + "temperature", + "top_k", + "top_logprobs", + "top_p" ] }, { @@ -2782,6 +3074,7 @@ "completion": "0.000002", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -2792,17 +3085,17 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "tools", - "tool_choice", - "stop", "frequency_penalty", + "max_tokens", "presence_penalty", "response_format", + "seed", + "stop", "structured_outputs", - "seed" + "temperature", + "tool_choice", + "tools", + "top_p" ] }, { @@ -2831,6 +3124,7 @@ "completion": "0.00001", "request": "0", "image": "0.00516", + "audio": "0", "web_search": "0", "internal_reasoning": "0", "input_cache_read": "0.00000031", @@ -2843,17 +3137,17 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "tools", - "tool_choice", - "reasoning", "include_reasoning", - "structured_outputs", + "max_tokens", + "reasoning", "response_format", + "seed", "stop", - "seed" + "structured_outputs", + "temperature", + "tool_choice", + "tools", + "top_p" ] }, { @@ -2881,6 +3175,7 @@ "completion": "0.00000018", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -2891,17 +3186,16 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "stop", "frequency_penalty", - "presence_penalty", - "top_k", - "repetition_penalty", "logit_bias", + "max_tokens", "min_p", - "response_format" + "presence_penalty", + "repetition_penalty", + "stop", + "temperature", + "top_k", + "top_p" ] }, { @@ -2928,6 +3222,7 @@ "completion": "0.0000033", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -2938,17 +3233,16 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "stop", "frequency_penalty", - "presence_penalty", - "top_k", - "repetition_penalty", "logit_bias", + "max_tokens", "min_p", - "response_format" + "presence_penalty", + "repetition_penalty", + "stop", + "temperature", + "top_k", + "top_p" ] }, { @@ -2975,6 +3269,7 @@ "completion": "0.0000012", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -2985,19 +3280,18 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "tools", - "tool_choice", - "stop", "frequency_penalty", - "presence_penalty", - "top_k", - "repetition_penalty", "logit_bias", + "max_tokens", "min_p", - "response_format" + "presence_penalty", + "repetition_penalty", + "stop", + "temperature", + "tool_choice", + "tools", + "top_k", + "top_p" ] }, { @@ -3024,6 +3318,7 @@ "completion": "0.0000008", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -3034,17 +3329,16 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "stop", "frequency_penalty", - "presence_penalty", - "top_k", - "repetition_penalty", "logit_bias", + "max_tokens", "min_p", - "response_format" + "presence_penalty", + "repetition_penalty", + "stop", + "temperature", + "top_k", + "top_p" ] }, { @@ -3071,6 +3365,7 @@ "completion": "0.00000035", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -3081,19 +3376,19 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "reasoning", - "include_reasoning", - "stop", "frequency_penalty", + "include_reasoning", + "max_tokens", + "min_p", "presence_penalty", + "reasoning", "repetition_penalty", "response_format", - "top_k", "seed", - "min_p" + "stop", + "temperature", + "top_k", + "top_p" ] }, { @@ -3103,7 +3398,7 @@ "name": "Inception: Mercury Coder", "created": 1746033880, "description": "Mercury Coder is the first diffusion large language model (dLLM). Applying a breakthrough discrete diffusion approach, the model runs 5-10x faster than even speed optimized models like Claude 3.5 Haiku and GPT-4o Mini while matching their performance. Mercury Coder's speed means that developers can stay in the flow while coding, enjoying rapid chat-based iteration and responsive code completion suggestions. On Copilot Arena, Mercury Coder ranks 1st in speed and ties for 2nd in quality. Read more in the [blog post here](https://www.inceptionlabs.ai/introducing-mercury).", - "context_length": 32000, + "context_length": 128000, "architecture": { "modality": "text->text", "input_modalities": [ @@ -3120,20 +3415,28 @@ "completion": "0.000001", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, "top_provider": { - "context_length": 32000, - "max_completion_tokens": null, + "context_length": 128000, + "max_completion_tokens": 16384, "is_moderated": false }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", "frequency_penalty", + "max_tokens", "presence_penalty", - "stop" + "response_format", + "stop", + "structured_outputs", + "temperature", + "tool_choice", + "tools", + "top_k", + "top_p" ] }, { @@ -3160,6 +3463,7 @@ "completion": "0", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -3170,19 +3474,19 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "tools", - "tool_choice", - "reasoning", + "frequency_penalty", "include_reasoning", - "structured_outputs", + "max_tokens", + "presence_penalty", + "reasoning", "response_format", "stop", - "frequency_penalty", - "presence_penalty", - "top_k" + "structured_outputs", + "temperature", + "tool_choice", + "tools", + "top_k", + "top_p" ] }, { @@ -3210,6 +3514,7 @@ "completion": "0.0000004", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -3249,6 +3554,7 @@ "completion": "0.00000218", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -3259,17 +3565,17 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "stop", "frequency_penalty", + "max_tokens", + "min_p", "presence_penalty", "repetition_penalty", "response_format", - "top_k", "seed", - "min_p" + "stop", + "temperature", + "top_k", + "top_p" ] }, { @@ -3297,6 +3603,7 @@ "completion": "0.00000005", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -3307,20 +3614,20 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "stop", "frequency_penalty", + "logit_bias", + "logprobs", + "max_tokens", + "min_p", "presence_penalty", "repetition_penalty", "response_format", - "top_k", "seed", - "min_p", + "stop", + "temperature", + "top_k", "top_logprobs", - "logprobs", - "logit_bias" + "top_p" ] }, { @@ -3347,6 +3654,7 @@ "completion": "0", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -3357,21 +3665,21 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "reasoning", - "include_reasoning", - "stop", "frequency_penalty", - "presence_penalty", - "seed", - "top_k", - "min_p", - "repetition_penalty", - "logprobs", + "include_reasoning", "logit_bias", - "top_logprobs" + "logprobs", + "max_tokens", + "min_p", + "presence_penalty", + "reasoning", + "repetition_penalty", + "seed", + "stop", + "temperature", + "top_k", + "top_logprobs", + "top_p" ] }, { @@ -3394,39 +3702,40 @@ "instruct_type": "qwen3" }, "pricing": { - "prompt": "0.00000008", - "completion": "0.00000029", + "prompt": "0.00000001999188", + "completion": "0.0000000800064", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, "top_provider": { "context_length": 40960, - "max_completion_tokens": 40960, + "max_completion_tokens": null, "is_moderated": false }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "tools", - "tool_choice", - "reasoning", - "include_reasoning", - "response_format", - "stop", "frequency_penalty", - "presence_penalty", - "repetition_penalty", - "top_k", - "seed", - "min_p", - "structured_outputs", + "include_reasoning", "logit_bias", "logprobs", - "top_logprobs" + "max_tokens", + "min_p", + "presence_penalty", + "reasoning", + "repetition_penalty", + "response_format", + "seed", + "stop", + "structured_outputs", + "temperature", + "tool_choice", + "tools", + "top_k", + "top_logprobs", + "top_p" ] }, { @@ -3453,6 +3762,7 @@ "completion": "0", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -3463,21 +3773,21 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "reasoning", - "include_reasoning", - "stop", "frequency_penalty", - "presence_penalty", - "seed", - "top_k", - "min_p", - "repetition_penalty", - "logprobs", + "include_reasoning", "logit_bias", - "top_logprobs" + "logprobs", + "max_tokens", + "min_p", + "presence_penalty", + "reasoning", + "repetition_penalty", + "seed", + "stop", + "temperature", + "top_k", + "top_logprobs", + "top_p" ] }, { @@ -3504,6 +3814,7 @@ "completion": "0.000000138", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -3514,19 +3825,19 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "reasoning", - "include_reasoning", - "stop", "frequency_penalty", - "presence_penalty", - "seed", - "top_k", + "include_reasoning", + "logit_bias", + "max_tokens", "min_p", + "presence_penalty", + "reasoning", "repetition_penalty", - "logit_bias" + "seed", + "stop", + "temperature", + "top_k", + "top_p" ] }, { @@ -3553,6 +3864,7 @@ "completion": "0", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -3563,21 +3875,21 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "reasoning", - "include_reasoning", - "stop", "frequency_penalty", - "presence_penalty", - "seed", - "top_k", - "min_p", - "repetition_penalty", - "logprobs", + "include_reasoning", "logit_bias", - "top_logprobs" + "logprobs", + "max_tokens", + "min_p", + "presence_penalty", + "reasoning", + "repetition_penalty", + "seed", + "stop", + "temperature", + "top_k", + "top_logprobs", + "top_p" ] }, { @@ -3604,6 +3916,7 @@ "completion": "0.00000024", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -3614,25 +3927,25 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "reasoning", - "include_reasoning", "frequency_penalty", - "min_p", - "presence_penalty", - "repetition_penalty", - "seed", - "stop", - "top_k", - "response_format", - "structured_outputs", - "tools", - "tool_choice", + "include_reasoning", "logit_bias", "logprobs", - "top_logprobs" + "max_tokens", + "min_p", + "presence_penalty", + "reasoning", + "repetition_penalty", + "response_format", + "seed", + "stop", + "structured_outputs", + "temperature", + "tool_choice", + "tools", + "top_k", + "top_logprobs", + "top_p" ] }, { @@ -3655,10 +3968,11 @@ "instruct_type": "qwen3" }, "pricing": { - "prompt": "0.000000027", - "completion": "0.000000027", + "prompt": "0.000000017992692", + "completion": "0.00000007200576", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -3669,25 +3983,25 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "reasoning", - "include_reasoning", - "stop", "frequency_penalty", - "presence_penalty", - "response_format", - "top_logprobs", - "logprobs", + "include_reasoning", "logit_bias", - "seed", - "tools", - "tool_choice", - "structured_outputs", - "top_k", + "logprobs", + "max_tokens", "min_p", - "repetition_penalty" + "presence_penalty", + "reasoning", + "repetition_penalty", + "response_format", + "seed", + "stop", + "structured_outputs", + "temperature", + "tool_choice", + "tools", + "top_k", + "top_logprobs", + "top_p" ] }, { @@ -3714,6 +4028,7 @@ "completion": "0", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -3724,25 +4039,25 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "tools", - "tool_choice", - "structured_outputs", - "response_format", - "stop", "frequency_penalty", - "presence_penalty", - "top_k", - "reasoning", "include_reasoning", - "seed", - "min_p", - "repetition_penalty", - "logprobs", "logit_bias", - "top_logprobs" + "logprobs", + "max_tokens", + "min_p", + "presence_penalty", + "reasoning", + "repetition_penalty", + "response_format", + "seed", + "stop", + "structured_outputs", + "temperature", + "tool_choice", + "tools", + "top_k", + "top_logprobs", + "top_p" ] }, { @@ -3769,6 +4084,7 @@ "completion": "0.0000006", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -3779,25 +4095,25 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "tools", - "tool_choice", - "reasoning", - "include_reasoning", - "structured_outputs", - "response_format", - "stop", "frequency_penalty", - "presence_penalty", - "seed", - "top_k", - "min_p", - "repetition_penalty", + "include_reasoning", + "logit_bias", "logprobs", + "max_tokens", + "min_p", + "presence_penalty", + "reasoning", + "repetition_penalty", + "response_format", + "seed", + "stop", + "structured_outputs", + "temperature", + "tool_choice", + "tools", + "top_k", "top_logprobs", - "logit_bias" + "top_p" ] }, { @@ -3824,6 +4140,7 @@ "completion": "0", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -3834,21 +4151,73 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "reasoning", - "include_reasoning", - "stop", "frequency_penalty", - "presence_penalty", - "seed", - "top_k", - "min_p", - "repetition_penalty", - "logprobs", + "include_reasoning", "logit_bias", - "top_logprobs" + "logprobs", + "max_tokens", + "min_p", + "presence_penalty", + "reasoning", + "repetition_penalty", + "seed", + "stop", + "temperature", + "top_k", + "top_logprobs", + "top_p" + ] + }, + { + "id": "tngtech/deepseek-r1t-chimera", + "canonical_slug": "tngtech/deepseek-r1t-chimera", + "hugging_face_id": "tngtech/DeepSeek-R1T-Chimera", + "name": "TNG: DeepSeek R1T Chimera", + "created": 1745760875, + "description": "DeepSeek-R1T-Chimera is created by merging DeepSeek-R1 and DeepSeek-V3 (0324), combining the reasoning capabilities of R1 with the token efficiency improvements of V3. It is based on a DeepSeek-MoE Transformer architecture and is optimized for general text generation tasks.\n\nThe model merges pretrained weights from both source models to balance performance across reasoning, efficiency, and instruction-following tasks. It is released under the MIT license and intended for research and commercial use.", + "context_length": 163840, + "architecture": { + "modality": "text->text", + "input_modalities": [ + "text" + ], + "output_modalities": [ + "text" + ], + "tokenizer": "DeepSeek", + "instruct_type": null + }, + "pricing": { + "prompt": "0.00000017992692", + "completion": "0.0000007200576", + "request": "0", + "image": "0", + "audio": "0", + "web_search": "0", + "internal_reasoning": "0" + }, + "top_provider": { + "context_length": 163840, + "max_completion_tokens": null, + "is_moderated": false + }, + "per_request_limits": null, + "supported_parameters": [ + "frequency_penalty", + "include_reasoning", + "logit_bias", + "logprobs", + "max_tokens", + "min_p", + "presence_penalty", + "reasoning", + "repetition_penalty", + "seed", + "stop", + "temperature", + "top_k", + "top_logprobs", + "top_p" ] }, { @@ -3875,6 +4244,7 @@ "completion": "0", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -3885,21 +4255,21 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "reasoning", - "include_reasoning", - "stop", "frequency_penalty", - "presence_penalty", - "seed", - "top_k", - "min_p", - "repetition_penalty", - "logprobs", + "include_reasoning", "logit_bias", - "top_logprobs" + "logprobs", + "max_tokens", + "min_p", + "presence_penalty", + "reasoning", + "repetition_penalty", + "seed", + "stop", + "temperature", + "top_k", + "top_logprobs", + "top_p" ] }, { @@ -3922,10 +4292,11 @@ "instruct_type": "deepseek-r1" }, "pricing": { - "prompt": "0.000000302", - "completion": "0.000000302", + "prompt": "0.0000001999188", + "completion": "0.000000800064", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -3936,21 +4307,21 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "reasoning", - "include_reasoning", - "stop", "frequency_penalty", - "presence_penalty", - "seed", - "top_k", - "min_p", - "repetition_penalty", - "logprobs", + "include_reasoning", "logit_bias", - "top_logprobs" + "logprobs", + "max_tokens", + "min_p", + "presence_penalty", + "reasoning", + "repetition_penalty", + "seed", + "stop", + "temperature", + "top_k", + "top_logprobs", + "top_p" ] }, { @@ -3977,6 +4348,7 @@ "completion": "0", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -3987,121 +4359,21 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "reasoning", + "frequency_penalty", "include_reasoning", - "stop", - "frequency_penalty", - "presence_penalty", - "seed", - "top_k", - "min_p", - "repetition_penalty", - "logprobs", "logit_bias", - "top_logprobs" - ] - }, - { - "id": "thudm/glm-z1-32b", - "canonical_slug": "thudm/glm-z1-32b-0414", - "hugging_face_id": "THUDM/GLM-Z1-32B-0414", - "name": "THUDM: GLM Z1 32B", - "created": 1744924148, - "description": "GLM-Z1-32B-0414 is an enhanced reasoning variant of GLM-4-32B, built for deep mathematical, logical, and code-oriented problem solving. It applies extended reinforcement learning—both task-specific and general pairwise preference-based—to improve performance on complex multi-step tasks. Compared to the base GLM-4-32B model, Z1 significantly boosts capabilities in structured reasoning and formal domains.\n\nThe model supports enforced “thinking” steps via prompt engineering and offers improved coherence for long-form outputs. It’s optimized for use in agentic workflows, and includes support for long context (via YaRN), JSON tool calling, and fine-grained sampling configuration for stable inference. Ideal for use cases requiring deliberate, multi-step reasoning or formal derivations.", - "context_length": 32768, - "architecture": { - "modality": "text->text", - "input_modalities": [ - "text" - ], - "output_modalities": [ - "text" - ], - "tokenizer": "Other", - "instruct_type": "deepseek-r1" - }, - "pricing": { - "prompt": "0.00000003", - "completion": "0.00000003", - "request": "0", - "image": "0", - "web_search": "0", - "internal_reasoning": "0" - }, - "top_provider": { - "context_length": 32768, - "max_completion_tokens": null, - "is_moderated": false - }, - "per_request_limits": null, - "supported_parameters": [ + "logprobs", "max_tokens", - "temperature", - "top_p", + "min_p", + "presence_penalty", "reasoning", - "include_reasoning", - "stop", - "frequency_penalty", - "presence_penalty", - "seed", - "top_k", - "min_p", "repetition_penalty", - "logprobs", - "logit_bias", - "top_logprobs" - ] - }, - { - "id": "thudm/glm-4-32b:free", - "canonical_slug": "thudm/glm-4-32b-0414", - "hugging_face_id": "THUDM/GLM-4-32B-0414", - "name": "THUDM: GLM 4 32B (free)", - "created": 1744920915, - "description": "GLM-4-32B-0414 is a 32B bilingual (Chinese-English) open-weight language model optimized for code generation, function calling, and agent-style tasks. Pretrained on 15T of high-quality and reasoning-heavy data, it was further refined using human preference alignment, rejection sampling, and reinforcement learning. The model excels in complex reasoning, artifact generation, and structured output tasks, achieving performance comparable to GPT-4o and DeepSeek-V3-0324 across several benchmarks.", - "context_length": 32768, - "architecture": { - "modality": "text->text", - "input_modalities": [ - "text" - ], - "output_modalities": [ - "text" - ], - "tokenizer": "Other", - "instruct_type": null - }, - "pricing": { - "prompt": "0", - "completion": "0", - "request": "0", - "image": "0", - "web_search": "0", - "internal_reasoning": "0" - }, - "top_provider": { - "context_length": 32768, - "max_completion_tokens": null, - "is_moderated": false - }, - "per_request_limits": null, - "supported_parameters": [ - "max_tokens", + "seed", + "stop", "temperature", - "top_p", - "stop", - "frequency_penalty", - "presence_penalty", - "seed", "top_k", - "min_p", - "repetition_penalty", - "logprobs", - "logit_bias", - "top_logprobs" + "top_logprobs", + "top_p" ] }, { @@ -4128,6 +4400,7 @@ "completion": "0.00000024", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -4138,17 +4411,17 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "stop", "frequency_penalty", - "presence_penalty", - "seed", - "top_k", + "logit_bias", + "max_tokens", "min_p", + "presence_penalty", "repetition_penalty", - "logit_bias" + "seed", + "stop", + "temperature", + "top_k", + "top_p" ] }, { @@ -4177,6 +4450,7 @@ "completion": "0.0000044", "request": "0", "image": "0.0008415", + "audio": "0", "web_search": "0", "internal_reasoning": "0", "input_cache_read": "0.000000275" @@ -4188,12 +4462,12 @@ }, "per_request_limits": null, "supported_parameters": [ - "tools", - "tool_choice", - "seed", "max_tokens", "response_format", - "structured_outputs" + "seed", + "structured_outputs", + "tool_choice", + "tools" ] }, { @@ -4222,6 +4496,7 @@ "completion": "0.000008", "request": "0", "image": "0.00153", + "audio": "0", "web_search": "0", "internal_reasoning": "0", "input_cache_read": "0.0000005" @@ -4233,12 +4508,12 @@ }, "per_request_limits": null, "supported_parameters": [ - "tools", - "tool_choice", - "seed", "max_tokens", "response_format", - "structured_outputs" + "seed", + "structured_outputs", + "tool_choice", + "tools" ] }, { @@ -4266,6 +4541,7 @@ "completion": "0.0000044", "request": "0", "image": "0.0008415", + "audio": "0", "web_search": "0", "internal_reasoning": "0", "input_cache_read": "0.000000275" @@ -4277,12 +4553,12 @@ }, "per_request_limits": null, "supported_parameters": [ - "tools", - "tool_choice", - "seed", "max_tokens", "response_format", - "structured_outputs" + "seed", + "structured_outputs", + "tool_choice", + "tools" ] }, { @@ -4309,6 +4585,7 @@ "completion": "0", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -4319,19 +4596,19 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "stop", "frequency_penalty", - "presence_penalty", - "seed", - "top_k", - "min_p", - "repetition_penalty", - "logprobs", "logit_bias", - "top_logprobs" + "logprobs", + "max_tokens", + "min_p", + "presence_penalty", + "repetition_penalty", + "seed", + "stop", + "temperature", + "top_k", + "top_logprobs", + "top_p" ] }, { @@ -4354,10 +4631,11 @@ "instruct_type": null }, "pricing": { - "prompt": "0.00000003", - "completion": "0.00000003", + "prompt": "0.00000001999188", + "completion": "0.0000000800064", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -4368,19 +4646,19 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "stop", "frequency_penalty", - "presence_penalty", - "seed", - "top_k", - "min_p", - "repetition_penalty", - "logprobs", "logit_bias", - "top_logprobs" + "logprobs", + "max_tokens", + "min_p", + "presence_penalty", + "repetition_penalty", + "seed", + "stop", + "temperature", + "top_k", + "top_logprobs", + "top_p" ] }, { @@ -4409,6 +4687,7 @@ "completion": "0.000008", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0", "input_cache_read": "0.0000005" @@ -4420,21 +4699,21 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "tools", - "tool_choice", - "stop", "frequency_penalty", - "presence_penalty", - "web_search_options", - "seed", "logit_bias", "logprobs", - "top_logprobs", + "max_tokens", + "presence_penalty", "response_format", - "structured_outputs" + "seed", + "stop", + "structured_outputs", + "temperature", + "tool_choice", + "tools", + "top_logprobs", + "top_p", + "web_search_options" ] }, { @@ -4463,6 +4742,7 @@ "completion": "0.0000016", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0", "input_cache_read": "0.0000001" @@ -4474,21 +4754,21 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "tools", - "tool_choice", - "stop", "frequency_penalty", - "presence_penalty", - "web_search_options", - "seed", "logit_bias", "logprobs", - "top_logprobs", + "max_tokens", + "presence_penalty", "response_format", - "structured_outputs" + "seed", + "stop", + "structured_outputs", + "temperature", + "tool_choice", + "tools", + "top_logprobs", + "top_p", + "web_search_options" ] }, { @@ -4517,6 +4797,7 @@ "completion": "0.0000004", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0", "input_cache_read": "0.000000025" @@ -4528,20 +4809,20 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "tools", - "tool_choice", - "stop", "frequency_penalty", - "presence_penalty", - "seed", "logit_bias", "logprobs", - "top_logprobs", + "max_tokens", + "presence_penalty", "response_format", - "structured_outputs" + "seed", + "stop", + "structured_outputs", + "temperature", + "tool_choice", + "tools", + "top_logprobs", + "top_p" ] }, { @@ -4568,6 +4849,7 @@ "completion": "0.0000012", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -4578,16 +4860,16 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "stop", "frequency_penalty", + "max_tokens", + "min_p", "presence_penalty", "repetition_penalty", + "seed", + "stop", + "temperature", "top_k", - "min_p", - "seed" + "top_p" ] }, { @@ -4597,7 +4879,7 @@ "name": "AlfredPros: CodeLLaMa 7B Instruct Solidity", "created": 1744641874, "description": "A finetuned 7 billion parameters Code LLaMA - Instruct model to generate Solidity smart contract using 4-bit QLoRA finetuning provided by PEFT library.", - "context_length": 4096, + "context_length": 8192, "architecture": { "modality": "text->text", "input_modalities": [ @@ -4610,30 +4892,31 @@ "instruct_type": "alpaca" }, "pricing": { - "prompt": "0.0000008", - "completion": "0.0000012", + "prompt": "0.0000006", + "completion": "0.000001", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, "top_provider": { - "context_length": 4096, - "max_completion_tokens": 4096, + "context_length": 8192, + "max_completion_tokens": 8192, "is_moderated": false }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "stop", "frequency_penalty", + "max_tokens", + "min_p", "presence_penalty", "repetition_penalty", + "seed", + "stop", + "temperature", "top_k", - "min_p", - "seed" + "top_p" ] }, { @@ -4660,6 +4943,7 @@ "completion": "0", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -4670,21 +4954,21 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "reasoning", - "include_reasoning", - "stop", "frequency_penalty", - "presence_penalty", - "seed", - "top_k", - "min_p", - "repetition_penalty", - "logprobs", + "include_reasoning", "logit_bias", - "top_logprobs" + "logprobs", + "max_tokens", + "min_p", + "presence_penalty", + "reasoning", + "repetition_penalty", + "seed", + "stop", + "temperature", + "top_k", + "top_logprobs", + "top_p" ] }, { @@ -4707,10 +4991,11 @@ "instruct_type": "deepseek-r1" }, "pricing": { - "prompt": "0.000000015", - "completion": "0.000000015", + "prompt": "0.00000001", + "completion": "0.0000000400032", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -4721,21 +5006,21 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "reasoning", - "include_reasoning", - "stop", "frequency_penalty", - "presence_penalty", - "seed", - "top_k", - "min_p", - "repetition_penalty", - "logprobs", + "include_reasoning", "logit_bias", - "top_logprobs" + "logprobs", + "max_tokens", + "min_p", + "presence_penalty", + "reasoning", + "repetition_penalty", + "seed", + "stop", + "temperature", + "top_k", + "top_logprobs", + "top_p" ] }, { @@ -4762,6 +5047,7 @@ "completion": "0", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -4772,21 +5058,21 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "reasoning", - "include_reasoning", - "stop", "frequency_penalty", - "presence_penalty", - "seed", - "top_k", - "min_p", - "repetition_penalty", - "logprobs", + "include_reasoning", "logit_bias", - "top_logprobs" + "logprobs", + "max_tokens", + "min_p", + "presence_penalty", + "reasoning", + "repetition_penalty", + "seed", + "stop", + "temperature", + "top_k", + "top_logprobs", + "top_p" ] }, { @@ -4813,6 +5099,7 @@ "completion": "0.000000015", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -4823,21 +5110,21 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "reasoning", - "include_reasoning", - "stop", "frequency_penalty", - "presence_penalty", - "seed", - "top_k", - "min_p", - "repetition_penalty", - "logprobs", + "include_reasoning", "logit_bias", - "top_logprobs" + "logprobs", + "max_tokens", + "min_p", + "presence_penalty", + "reasoning", + "repetition_penalty", + "seed", + "stop", + "temperature", + "top_k", + "top_logprobs", + "top_p" ] }, { @@ -4865,6 +5152,7 @@ "completion": "0", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -4875,21 +5163,21 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "reasoning", - "include_reasoning", - "stop", "frequency_penalty", - "presence_penalty", - "seed", - "top_k", - "min_p", - "repetition_penalty", - "logprobs", + "include_reasoning", "logit_bias", - "top_logprobs" + "logprobs", + "max_tokens", + "min_p", + "presence_penalty", + "reasoning", + "repetition_penalty", + "seed", + "stop", + "temperature", + "top_k", + "top_logprobs", + "top_p" ] }, { @@ -4913,10 +5201,11 @@ "instruct_type": null }, "pricing": { - "prompt": "0.000000038", - "completion": "0.000000038", + "prompt": "0.00000002498985", + "completion": "0.000000100008", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -4927,21 +5216,21 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "reasoning", - "include_reasoning", - "stop", "frequency_penalty", - "presence_penalty", - "seed", - "top_k", - "min_p", - "repetition_penalty", - "logprobs", + "include_reasoning", "logit_bias", - "top_logprobs" + "logprobs", + "max_tokens", + "min_p", + "presence_penalty", + "reasoning", + "repetition_penalty", + "seed", + "stop", + "temperature", + "top_k", + "top_logprobs", + "top_p" ] }, { @@ -4968,6 +5257,7 @@ "completion": "0.0000005", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0", "input_cache_read": "0.000000075" @@ -4979,18 +5269,18 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "tools", - "tool_choice", - "reasoning", "include_reasoning", - "stop", - "seed", "logprobs", + "max_tokens", + "reasoning", + "response_format", + "seed", + "stop", + "temperature", + "tool_choice", + "tools", "top_logprobs", - "response_format" + "top_p" ] }, { @@ -5017,6 +5307,7 @@ "completion": "0.000015", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0", "input_cache_read": "0.00000075" @@ -5028,18 +5319,18 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "tools", - "tool_choice", - "stop", "frequency_penalty", - "presence_penalty", - "seed", "logprobs", + "max_tokens", + "presence_penalty", + "response_format", + "seed", + "stop", + "temperature", + "tool_choice", + "tools", "top_logprobs", - "response_format" + "top_p" ] }, { @@ -5066,6 +5357,7 @@ "completion": "0.0000004", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -5076,17 +5368,17 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "stop", "frequency_penalty", - "presence_penalty", - "seed", - "top_k", "logit_bias", "logprobs", - "top_logprobs" + "max_tokens", + "presence_penalty", + "seed", + "stop", + "temperature", + "top_k", + "top_logprobs", + "top_p" ] }, { @@ -5113,6 +5405,7 @@ "completion": "0", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -5123,19 +5416,19 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "stop", "frequency_penalty", - "presence_penalty", - "seed", - "top_k", - "min_p", - "repetition_penalty", - "logprobs", "logit_bias", - "top_logprobs" + "logprobs", + "max_tokens", + "min_p", + "presence_penalty", + "repetition_penalty", + "seed", + "stop", + "temperature", + "top_k", + "top_logprobs", + "top_p" ] }, { @@ -5162,6 +5455,7 @@ "completion": "0.0000018", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -5172,19 +5466,19 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "reasoning", - "include_reasoning", - "stop", "frequency_penalty", - "presence_penalty", - "seed", - "top_k", + "include_reasoning", "logit_bias", "logprobs", - "top_logprobs" + "max_tokens", + "presence_penalty", + "reasoning", + "seed", + "stop", + "temperature", + "top_k", + "top_logprobs", + "top_p" ] }, { @@ -5212,6 +5506,7 @@ "completion": "0.0000006", "request": "0", "image": "0.0006684", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -5222,23 +5517,23 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "stop", "frequency_penalty", + "logit_bias", + "logprobs", + "max_tokens", + "min_p", "presence_penalty", "repetition_penalty", "response_format", - "top_k", "seed", - "min_p", + "stop", "structured_outputs", - "tools", + "temperature", "tool_choice", - "logprobs", + "tools", + "top_k", "top_logprobs", - "logit_bias" + "top_p" ] }, { @@ -5266,6 +5561,7 @@ "completion": "0.0000003", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -5276,23 +5572,23 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "structured_outputs", - "response_format", - "stop", "frequency_penalty", - "presence_penalty", - "seed", - "top_k", - "min_p", - "repetition_penalty", - "tools", - "tool_choice", - "top_logprobs", + "logit_bias", "logprobs", - "logit_bias" + "max_tokens", + "min_p", + "presence_penalty", + "repetition_penalty", + "response_format", + "seed", + "stop", + "structured_outputs", + "temperature", + "tool_choice", + "tools", + "top_k", + "top_logprobs", + "top_p" ] }, { @@ -5315,10 +5611,11 @@ "instruct_type": null }, "pricing": { - "prompt": "0.000000302", - "completion": "0.000000302", + "prompt": "0.0000001999188", + "completion": "0.000000800064", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -5329,19 +5626,19 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "stop", "frequency_penalty", - "presence_penalty", - "seed", - "top_k", - "min_p", - "repetition_penalty", - "logprobs", "logit_bias", - "top_logprobs" + "logprobs", + "max_tokens", + "min_p", + "presence_penalty", + "repetition_penalty", + "seed", + "stop", + "temperature", + "top_k", + "top_logprobs", + "top_p" ] }, { @@ -5368,6 +5665,7 @@ "completion": "0.00000088", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -5378,17 +5676,16 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "stop", "frequency_penalty", - "presence_penalty", - "top_k", - "repetition_penalty", "logit_bias", + "max_tokens", "min_p", - "response_format" + "presence_penalty", + "repetition_penalty", + "stop", + "temperature", + "top_k", + "top_p" ] }, { @@ -5417,6 +5714,7 @@ "completion": "0", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -5428,14 +5726,14 @@ "per_request_limits": null, "supported_parameters": [ "max_tokens", - "temperature", - "top_p", - "tools", - "tool_choice", - "stop", - "seed", "response_format", - "structured_outputs" + "seed", + "stop", + "structured_outputs", + "temperature", + "tool_choice", + "tools", + "top_p" ] }, { @@ -5463,6 +5761,7 @@ "completion": "0", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -5473,20 +5772,20 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "seed", - "response_format", - "presence_penalty", - "stop", "frequency_penalty", - "top_k", - "min_p", - "repetition_penalty", - "logprobs", "logit_bias", - "top_logprobs" + "logprobs", + "max_tokens", + "min_p", + "presence_penalty", + "repetition_penalty", + "response_format", + "seed", + "stop", + "temperature", + "top_k", + "top_logprobs", + "top_p" ] }, { @@ -5496,7 +5795,7 @@ "name": "Qwen: Qwen2.5 VL 32B Instruct", "created": 1742839838, "description": "Qwen2.5-VL-32B is a multimodal vision-language model fine-tuned through reinforcement learning for enhanced mathematical reasoning, structured outputs, and visual problem-solving capabilities. It excels at visual analysis tasks, including object recognition, textual interpretation within images, and precise event localization in extended videos. Qwen2.5-VL-32B demonstrates state-of-the-art performance across multimodal benchmarks such as MMMU, MathVista, and VideoMME, while maintaining strong reasoning and clarity in text-based tasks like MMLU, mathematical problem-solving, and code generation.", - "context_length": 128000, + "context_length": 16384, "architecture": { "modality": "text+image->text", "input_modalities": [ @@ -5510,35 +5809,36 @@ "instruct_type": null }, "pricing": { - "prompt": "0.0000002", - "completion": "0.0000006", + "prompt": "0.00000001999188", + "completion": "0.0000000800064", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, "top_provider": { - "context_length": 128000, + "context_length": 16384, "max_completion_tokens": null, "is_moderated": false }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "stop", "frequency_penalty", + "logit_bias", + "logprobs", + "max_tokens", + "min_p", "presence_penalty", "repetition_penalty", "response_format", - "top_k", "seed", - "min_p", + "stop", "structured_outputs", - "logit_bias", - "logprobs", - "top_logprobs" + "temperature", + "top_k", + "top_logprobs", + "top_p" ] }, { @@ -5548,7 +5848,7 @@ "name": "DeepSeek: DeepSeek V3 0324 (free)", "created": 1742824755, "description": "DeepSeek V3, a 685B-parameter, mixture-of-experts model, is the latest iteration of the flagship chat model family from the DeepSeek team.\n\nIt succeeds the [DeepSeek V3](/deepseek/deepseek-chat-v3) model and performs really well on a variety of tasks.", - "context_length": 32768, + "context_length": 163840, "architecture": { "modality": "text->text", "input_modalities": [ @@ -5565,31 +5865,32 @@ "completion": "0", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, "top_provider": { - "context_length": 32768, - "max_completion_tokens": 16384, + "context_length": 163840, + "max_completion_tokens": 163840, "is_moderated": false }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "stop", "frequency_penalty", - "presence_penalty", - "seed", - "top_k", - "repetition_penalty", - "tools", - "tool_choice", - "min_p", - "logprobs", "logit_bias", - "top_logprobs" + "logprobs", + "max_tokens", + "min_p", + "presence_penalty", + "repetition_penalty", + "seed", + "stop", + "temperature", + "tool_choice", + "tools", + "top_k", + "top_logprobs", + "top_p" ] }, { @@ -5612,37 +5913,38 @@ "instruct_type": null }, "pricing": { - "prompt": "0.00000025", - "completion": "0.00000085", + "prompt": "0.00000017992692", + "completion": "0.0000007200576", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, "top_provider": { "context_length": 163840, - "max_completion_tokens": 163840, + "max_completion_tokens": null, "is_moderated": false }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "stop", "frequency_penalty", - "presence_penalty", - "seed", + "logit_bias", "logprobs", - "top_logprobs", + "max_tokens", "min_p", + "presence_penalty", "repetition_penalty", - "top_k", - "tools", - "tool_choice", - "structured_outputs", "response_format", - "logit_bias" + "seed", + "stop", + "structured_outputs", + "temperature", + "tool_choice", + "tools", + "top_k", + "top_logprobs", + "top_p" ] }, { @@ -5669,6 +5971,7 @@ "completion": "0", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -5679,16 +5982,16 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "stop", "frequency_penalty", + "max_tokens", + "min_p", "presence_penalty", "repetition_penalty", + "seed", + "stop", + "temperature", "top_k", - "min_p", - "seed" + "top_p" ] }, { @@ -5716,6 +6019,7 @@ "completion": "0.0006", "request": "0", "image": "0.21675", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -5726,9 +6030,9 @@ }, "per_request_limits": null, "supported_parameters": [ - "seed", "max_tokens", "response_format", + "seed", "structured_outputs" ] }, @@ -5757,6 +6061,7 @@ "completion": "0", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -5767,23 +6072,23 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "tools", - "tool_choice", - "structured_outputs", - "response_format", - "stop", "frequency_penalty", - "presence_penalty", - "top_k", - "seed", - "min_p", - "repetition_penalty", - "logprobs", "logit_bias", - "top_logprobs" + "logprobs", + "max_tokens", + "min_p", + "presence_penalty", + "repetition_penalty", + "response_format", + "seed", + "stop", + "structured_outputs", + "temperature", + "tool_choice", + "tools", + "top_k", + "top_logprobs", + "top_p" ] }, { @@ -5793,7 +6098,7 @@ "name": "Mistral: Mistral Small 3.1 24B", "created": 1742238937, "description": "Mistral Small 3.1 24B Instruct is an upgraded variant of Mistral Small 3 (2501), featuring 24 billion parameters with advanced multimodal capabilities. It provides state-of-the-art performance in text-based reasoning and vision tasks, including image analysis, programming, mathematical reasoning, and multilingual support across dozens of languages. Equipped with an extensive 128k token context window and optimized for efficient local inference, it supports use cases such as conversational agents, function calling, long-document comprehension, and privacy-sensitive deployments. The updated version is [Mistral Small 3.2](mistralai/mistral-small-3.2-24b-instruct)", - "context_length": 96000, + "context_length": 131072, "architecture": { "modality": "text+image->text", "input_modalities": [ @@ -5807,37 +6112,38 @@ "instruct_type": null }, "pricing": { - "prompt": "0.000000027", - "completion": "0.000000027", + "prompt": "0.000000017992692", + "completion": "0.00000007200576", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, "top_provider": { - "context_length": 96000, + "context_length": 131072, "max_completion_tokens": 96000, "is_moderated": false }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "stop", "frequency_penalty", + "logit_bias", + "logprobs", + "max_tokens", + "min_p", "presence_penalty", "repetition_penalty", "response_format", - "top_k", "seed", - "min_p", - "tools", + "stop", + "structured_outputs", + "temperature", "tool_choice", - "logprobs", - "logit_bias", + "tools", + "top_k", "top_logprobs", - "structured_outputs" + "top_p" ] }, { @@ -5865,6 +6171,7 @@ "completion": "0", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -5876,11 +6183,11 @@ "per_request_limits": null, "supported_parameters": [ "max_tokens", - "temperature", - "top_p", - "seed", "response_format", - "structured_outputs" + "seed", + "structured_outputs", + "temperature", + "top_p" ] }, { @@ -5908,6 +6215,7 @@ "completion": "0.00000004", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -5918,18 +6226,17 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "stop", "frequency_penalty", + "max_tokens", + "min_p", "presence_penalty", - "response_format", - "structured_outputs", "repetition_penalty", - "top_k", + "response_format", "seed", - "min_p" + "stop", + "temperature", + "top_k", + "top_p" ] }, { @@ -5956,6 +6263,7 @@ "completion": "0.000008", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -5967,11 +6275,11 @@ "per_request_limits": null, "supported_parameters": [ "max_tokens", + "stop", "temperature", - "top_p", - "tools", "tool_choice", - "stop" + "tools", + "top_p" ] }, { @@ -5998,6 +6306,7 @@ "completion": "0.0000004", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -6009,11 +6318,11 @@ "per_request_limits": null, "supported_parameters": [ "max_tokens", + "stop", "temperature", - "top_p", - "tools", "tool_choice", - "stop" + "tools", + "top_p" ] }, { @@ -6041,6 +6350,7 @@ "completion": "0", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -6051,21 +6361,21 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "stop", "frequency_penalty", - "presence_penalty", - "seed", - "top_k", - "min_p", - "repetition_penalty", - "logprobs", "logit_bias", - "top_logprobs", + "logprobs", + "max_tokens", + "min_p", + "presence_penalty", + "repetition_penalty", "response_format", - "structured_outputs" + "seed", + "stop", + "structured_outputs", + "temperature", + "top_k", + "top_logprobs", + "top_p" ] }, { @@ -6089,10 +6399,11 @@ "instruct_type": "gemma" }, "pricing": { - "prompt": "0.00000003", - "completion": "0.00000003", + "prompt": "0.0000000481286", + "completion": "0.000000192608", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -6103,20 +6414,20 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "stop", "frequency_penalty", - "presence_penalty", - "seed", - "top_k", - "min_p", - "repetition_penalty", - "logprobs", "logit_bias", + "logprobs", + "max_tokens", + "min_p", + "presence_penalty", + "repetition_penalty", + "response_format", + "seed", + "stop", + "temperature", + "top_k", "top_logprobs", - "response_format" + "top_p" ] }, { @@ -6143,6 +6454,7 @@ "completion": "0.000008", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -6153,16 +6465,16 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "stop", "frequency_penalty", + "max_tokens", "presence_penalty", "response_format", + "seed", + "stop", "structured_outputs", + "temperature", "top_k", - "seed" + "top_p" ] }, { @@ -6189,6 +6501,7 @@ "completion": "0.0000006", "request": "0.0275", "image": "0.000217", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -6199,10 +6512,10 @@ }, "per_request_limits": null, "supported_parameters": [ - "web_search_options", "max_tokens", "response_format", - "structured_outputs" + "structured_outputs", + "web_search_options" ] }, { @@ -6229,6 +6542,7 @@ "completion": "0.00001", "request": "0.035", "image": "0.003613", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -6239,10 +6553,10 @@ }, "per_request_limits": null, "supported_parameters": [ - "web_search_options", "max_tokens", "response_format", - "structured_outputs" + "structured_outputs", + "web_search_options" ] }, { @@ -6269,6 +6583,7 @@ "completion": "0", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -6279,72 +6594,21 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "reasoning", - "include_reasoning", - "stop", "frequency_penalty", - "presence_penalty", - "seed", - "top_k", - "min_p", - "repetition_penalty", - "logprobs", - "logit_bias", - "top_logprobs" - ] - }, - { - "id": "rekaai/reka-flash-3", - "canonical_slug": "rekaai/reka-flash-3", - "hugging_face_id": "RekaAI/reka-flash-3", - "name": "Reka: Flash 3", - "created": 1741812813, - "description": "Reka Flash 3 is a general-purpose, instruction-tuned large language model with 21 billion parameters, developed by Reka. It excels at general chat, coding tasks, instruction-following, and function calling. Featuring a 32K context length and optimized through reinforcement learning (RLOO), it provides competitive performance comparable to proprietary models within a smaller parameter footprint. Ideal for low-latency, local, or on-device deployments, Reka Flash 3 is compact, supports efficient quantization (down to 11GB at 4-bit precision), and employs explicit reasoning tags (\"\") to indicate its internal thought process.\n\nReka Flash 3 is primarily an English model with limited multilingual understanding capabilities. The model weights are released under the Apache 2.0 license.", - "context_length": 32768, - "architecture": { - "modality": "text->text", - "input_modalities": [ - "text" - ], - "output_modalities": [ - "text" - ], - "tokenizer": "Other", - "instruct_type": null - }, - "pricing": { - "prompt": "0.000000013", - "completion": "0.000000013", - "request": "0", - "image": "0", - "web_search": "0", - "internal_reasoning": "0" - }, - "top_provider": { - "context_length": 32768, - "max_completion_tokens": null, - "is_moderated": false - }, - "per_request_limits": null, - "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "reasoning", "include_reasoning", - "stop", - "frequency_penalty", - "presence_penalty", - "seed", - "top_k", - "min_p", - "repetition_penalty", - "logprobs", "logit_bias", - "top_logprobs" + "logprobs", + "max_tokens", + "min_p", + "presence_penalty", + "reasoning", + "repetition_penalty", + "seed", + "stop", + "temperature", + "top_k", + "top_logprobs", + "top_p" ] }, { @@ -6372,6 +6636,7 @@ "completion": "0", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -6382,21 +6647,21 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "stop", "frequency_penalty", - "presence_penalty", - "seed", - "top_k", - "min_p", - "repetition_penalty", - "logprobs", "logit_bias", - "top_logprobs", + "logprobs", + "max_tokens", + "min_p", + "presence_penalty", + "repetition_penalty", "response_format", - "structured_outputs" + "seed", + "stop", + "structured_outputs", + "temperature", + "top_k", + "top_logprobs", + "top_p" ] }, { @@ -6406,7 +6671,7 @@ "name": "Google: Gemma 3 27B", "created": 1741756359, "description": "Gemma 3 introduces multimodality, supporting vision-language input and text outputs. It handles context windows up to 128k tokens, understands over 140 languages, and offers improved math, reasoning, and chat capabilities, including structured outputs and function calling. Gemma 3 27B is Google's latest open source model, successor to [Gemma 2](google/gemma-2-27b-it)", - "context_length": 131072, + "context_length": 96000, "architecture": { "modality": "text+image->text", "input_modalities": [ @@ -6420,34 +6685,35 @@ "instruct_type": "gemma" }, "pricing": { - "prompt": "0.00000009", - "completion": "0.00000017", + "prompt": "0.0000000666396", + "completion": "0.000000266688", "request": "0", - "image": "0.0000256", + "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, "top_provider": { - "context_length": 131072, - "max_completion_tokens": 16384, + "context_length": 96000, + "max_completion_tokens": 8192, "is_moderated": false }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "stop", "frequency_penalty", + "logit_bias", + "logprobs", + "max_tokens", + "min_p", "presence_penalty", "repetition_penalty", "response_format", - "top_k", "seed", - "min_p", - "logprobs", - "logit_bias", - "top_logprobs" + "stop", + "temperature", + "top_k", + "top_logprobs", + "top_p" ] }, { @@ -6474,6 +6740,7 @@ "completion": "0.000001", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -6484,16 +6751,16 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", "frequency_penalty", + "max_tokens", "min_p", "presence_penalty", "repetition_penalty", "seed", "stop", - "top_k" + "temperature", + "top_k", + "top_p" ] }, { @@ -6503,7 +6770,7 @@ "name": "TheDrummer: Skyfall 36B V2", "created": 1741636566, "description": "Skyfall 36B v2 is an enhanced iteration of Mistral Small 2501, specifically fine-tuned for improved creativity, nuanced writing, role-playing, and coherent storytelling.", - "context_length": 16384, + "context_length": 32768, "architecture": { "modality": "text->text", "input_modalities": [ @@ -6516,35 +6783,36 @@ "instruct_type": null }, "pricing": { - "prompt": "0.00000001512", - "completion": "0.00000001512", + "prompt": "0.0000000481286", + "completion": "0.000000192608", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, "top_provider": { - "context_length": 16384, + "context_length": 32768, "max_completion_tokens": null, "is_moderated": false }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "stop", "frequency_penalty", - "presence_penalty", - "response_format", - "structured_outputs", - "seed", - "top_k", - "min_p", - "repetition_penalty", - "logprobs", "logit_bias", - "top_logprobs" + "logprobs", + "max_tokens", + "min_p", + "presence_penalty", + "repetition_penalty", + "response_format", + "seed", + "stop", + "structured_outputs", + "temperature", + "top_k", + "top_logprobs", + "top_p" ] }, { @@ -6572,6 +6840,7 @@ "completion": "0.0000001", "request": "0", "image": "0.00017685", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -6582,17 +6851,17 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "stop", "frequency_penalty", + "max_tokens", + "min_p", "presence_penalty", "repetition_penalty", "response_format", - "top_k", "seed", - "min_p" + "stop", + "temperature", + "top_k", + "top_p" ] }, { @@ -6620,6 +6889,7 @@ "completion": "0.000008", "request": "0", "image": "0", + "audio": "0", "web_search": "0.005", "internal_reasoning": "0" }, @@ -6630,15 +6900,15 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "reasoning", - "include_reasoning", - "web_search_options", - "top_k", "frequency_penalty", - "presence_penalty" + "include_reasoning", + "max_tokens", + "presence_penalty", + "reasoning", + "temperature", + "top_k", + "top_p", + "web_search_options" ] }, { @@ -6666,6 +6936,7 @@ "completion": "0.000015", "request": "0", "image": "0", + "audio": "0", "web_search": "0.005", "internal_reasoning": "0" }, @@ -6676,13 +6947,13 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "web_search_options", - "top_k", "frequency_penalty", - "presence_penalty" + "max_tokens", + "presence_penalty", + "temperature", + "top_k", + "top_p", + "web_search_options" ] }, { @@ -6709,6 +6980,7 @@ "completion": "0.000008", "request": "0", "image": "0", + "audio": "0", "web_search": "0.005", "internal_reasoning": "0.000003" }, @@ -6719,14 +6991,15 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "reasoning", - "include_reasoning", - "top_k", "frequency_penalty", - "presence_penalty" + "include_reasoning", + "max_tokens", + "presence_penalty", + "reasoning", + "temperature", + "top_k", + "top_p", + "web_search_options" ] }, { @@ -6753,6 +7026,7 @@ "completion": "0", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -6763,18 +7037,15 @@ }, "per_request_limits": null, "supported_parameters": [ + "frequency_penalty", "max_tokens", - "temperature", - "top_p", - "structured_outputs", + "presence_penalty", "response_format", "stop", - "frequency_penalty", - "presence_penalty", + "structured_outputs", + "temperature", "top_k", - "reasoning", - "include_reasoning", - "seed" + "top_p" ] }, { @@ -6801,6 +7072,7 @@ "completion": "0.00000015", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -6811,23 +7083,23 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "reasoning", - "include_reasoning", - "stop", "frequency_penalty", - "presence_penalty", - "response_format", - "structured_outputs", - "top_k", - "repetition_penalty", + "include_reasoning", "logit_bias", - "min_p", "logprobs", + "max_tokens", + "min_p", + "presence_penalty", + "reasoning", + "repetition_penalty", + "response_format", + "seed", + "stop", + "structured_outputs", + "temperature", + "top_k", "top_logprobs", - "seed" + "top_p" ] }, { @@ -6854,6 +7126,7 @@ "completion": "0", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -6864,19 +7137,19 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "stop", "frequency_penalty", - "presence_penalty", - "seed", - "top_k", - "min_p", - "repetition_penalty", - "logprobs", "logit_bias", - "top_logprobs" + "logprobs", + "max_tokens", + "min_p", + "presence_penalty", + "repetition_penalty", + "seed", + "stop", + "temperature", + "top_k", + "top_logprobs", + "top_p" ] }, { @@ -6905,6 +7178,7 @@ "completion": "0.0000003", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -6916,13 +7190,14 @@ "per_request_limits": null, "supported_parameters": [ "max_tokens", - "temperature", - "top_p", - "tools", - "tool_choice", - "stop", "response_format", - "structured_outputs" + "seed", + "stop", + "structured_outputs", + "temperature", + "tool_choice", + "tools", + "top_p" ] }, { @@ -6950,6 +7225,7 @@ "completion": "0.000015", "request": "0", "image": "0.0048", + "audio": "0", "web_search": "0", "internal_reasoning": "0", "input_cache_read": "0.0000003", @@ -6962,15 +7238,15 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "stop", - "reasoning", "include_reasoning", - "tools", + "max_tokens", + "reasoning", + "stop", + "temperature", "tool_choice", - "top_p", - "top_k" + "tools", + "top_k", + "top_p" ] }, { @@ -6998,6 +7274,7 @@ "completion": "0.000015", "request": "0", "image": "0.0048", + "audio": "0", "web_search": "0", "internal_reasoning": "0", "input_cache_read": "0.0000003", @@ -7010,13 +7287,13 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "stop", - "reasoning", "include_reasoning", - "tools", - "tool_choice" + "max_tokens", + "reasoning", + "stop", + "temperature", + "tool_choice", + "tools" ] }, { @@ -7044,6 +7321,7 @@ "completion": "0.000015", "request": "0", "image": "0.0048", + "audio": "0", "web_search": "0", "internal_reasoning": "0", "input_cache_read": "0.0000003", @@ -7056,13 +7334,13 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "stop", - "reasoning", "include_reasoning", - "tools", - "tool_choice" + "max_tokens", + "reasoning", + "stop", + "temperature", + "tool_choice", + "tools" ] }, { @@ -7089,6 +7367,7 @@ "completion": "0.000008", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -7099,19 +7378,18 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "reasoning", - "include_reasoning", - "top_k", "frequency_penalty", - "presence_penalty", - "stop", - "repetition_penalty", + "include_reasoning", "logit_bias", + "max_tokens", "min_p", - "response_format" + "presence_penalty", + "reasoning", + "repetition_penalty", + "stop", + "temperature", + "top_k", + "top_p" ] }, { @@ -7138,6 +7416,7 @@ "completion": "0.0000006", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -7148,20 +7427,17 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "tools", - "tool_choice", - "stop", "frequency_penalty", + "max_tokens", "presence_penalty", "response_format", - "structured_outputs", "seed", - "top_logprobs", - "logprobs", - "logit_bias" + "stop", + "structured_outputs", + "temperature", + "tool_choice", + "tools", + "top_p" ] }, { @@ -7188,6 +7464,7 @@ "completion": "0", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -7198,21 +7475,21 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "reasoning", - "include_reasoning", - "stop", "frequency_penalty", - "presence_penalty", - "seed", - "top_k", - "min_p", - "repetition_penalty", - "logprobs", + "include_reasoning", "logit_bias", - "top_logprobs" + "logprobs", + "max_tokens", + "min_p", + "presence_penalty", + "reasoning", + "repetition_penalty", + "seed", + "stop", + "temperature", + "top_k", + "top_logprobs", + "top_p" ] }, { @@ -7235,10 +7512,11 @@ "instruct_type": "deepseek-r1" }, "pricing": { - "prompt": "0.000000013", - "completion": "0.000000013", + "prompt": "0.00000001", + "completion": "0.0000000340768", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -7249,21 +7527,21 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "reasoning", - "include_reasoning", - "stop", "frequency_penalty", - "presence_penalty", - "seed", - "top_k", - "min_p", - "repetition_penalty", - "logprobs", + "include_reasoning", "logit_bias", - "top_logprobs" + "logprobs", + "max_tokens", + "min_p", + "presence_penalty", + "reasoning", + "repetition_penalty", + "seed", + "stop", + "temperature", + "top_k", + "top_logprobs", + "top_p" ] }, { @@ -7290,6 +7568,7 @@ "completion": "0", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -7300,19 +7579,69 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "stop", "frequency_penalty", - "presence_penalty", - "seed", - "top_k", - "min_p", - "repetition_penalty", - "logprobs", "logit_bias", - "top_logprobs" + "logprobs", + "max_tokens", + "min_p", + "presence_penalty", + "repetition_penalty", + "seed", + "stop", + "temperature", + "top_k", + "top_logprobs", + "top_p" + ] + }, + { + "id": "cognitivecomputations/dolphin3.0-mistral-24b", + "canonical_slug": "cognitivecomputations/dolphin3.0-mistral-24b", + "hugging_face_id": "cognitivecomputations/Dolphin3.0-Mistral-24B", + "name": "Dolphin3.0 Mistral 24B", + "created": 1739462019, + "description": "Dolphin 3.0 is the next generation of the Dolphin series of instruct-tuned models. Designed to be the ultimate general purpose local model, enabling coding, math, agentic, function calling, and general use cases.\n\nDolphin aims to be a general purpose instruct model, similar to the models behind ChatGPT, Claude, Gemini. \n\nPart of the [Dolphin 3.0 Collection](https://huggingface.co/collections/cognitivecomputations/dolphin-30-677ab47f73d7ff66743979a3) Curated and trained by [Eric Hartford](https://huggingface.co/ehartford), [Ben Gitter](https://huggingface.co/bigstorm), [BlouseJury](https://huggingface.co/BlouseJury) and [Cognitive Computations](https://huggingface.co/cognitivecomputations)", + "context_length": 32768, + "architecture": { + "modality": "text->text", + "input_modalities": [ + "text" + ], + "output_modalities": [ + "text" + ], + "tokenizer": "Other", + "instruct_type": null + }, + "pricing": { + "prompt": "0.000000037022", + "completion": "0.00000014816", + "request": "0", + "image": "0", + "audio": "0", + "web_search": "0", + "internal_reasoning": "0" + }, + "top_provider": { + "context_length": 32768, + "max_completion_tokens": null, + "is_moderated": false + }, + "per_request_limits": null, + "supported_parameters": [ + "frequency_penalty", + "logit_bias", + "logprobs", + "max_tokens", + "min_p", + "presence_penalty", + "repetition_penalty", + "seed", + "stop", + "temperature", + "top_k", + "top_logprobs", + "top_p" ] }, { @@ -7339,6 +7668,7 @@ "completion": "0.00000006", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -7349,20 +7679,20 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "stop", "frequency_penalty", - "presence_penalty", - "top_k", - "repetition_penalty", "logit_bias", + "logprobs", + "max_tokens", "min_p", + "presence_penalty", + "repetition_penalty", "response_format", "seed", - "logprobs", - "top_logprobs" + "stop", + "temperature", + "top_k", + "top_logprobs", + "top_p" ] }, { @@ -7389,6 +7719,7 @@ "completion": "0.0000044", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0", "input_cache_read": "0.00000055" @@ -7400,12 +7731,12 @@ }, "per_request_limits": null, "supported_parameters": [ - "tools", - "tool_choice", - "seed", "max_tokens", "response_format", - "structured_outputs" + "seed", + "structured_outputs", + "tool_choice", + "tools" ] }, { @@ -7432,6 +7763,7 @@ "completion": "0.00000004", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -7442,19 +7774,19 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "reasoning", - "include_reasoning", - "stop", "frequency_penalty", - "presence_penalty", - "seed", - "top_k", + "include_reasoning", + "logit_bias", + "max_tokens", "min_p", + "presence_penalty", + "reasoning", "repetition_penalty", - "logit_bias" + "seed", + "stop", + "temperature", + "top_k", + "top_p" ] }, { @@ -7470,7 +7802,8 @@ "input_modalities": [ "text", "image", - "file" + "file", + "audio" ], "output_modalities": [ "text" @@ -7483,6 +7816,7 @@ "completion": "0.0000004", "request": "0", "image": "0.0000258", + "audio": "0.0000007", "web_search": "0", "internal_reasoning": "0", "input_cache_read": "0.000000025", @@ -7496,13 +7830,14 @@ "per_request_limits": null, "supported_parameters": [ "max_tokens", - "temperature", - "top_p", - "tools", - "tool_choice", - "stop", "response_format", - "structured_outputs" + "seed", + "stop", + "structured_outputs", + "temperature", + "tool_choice", + "tools", + "top_p" ] }, { @@ -7530,6 +7865,7 @@ "completion": "0.00000063", "request": "0", "image": "0.0002688", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -7541,11 +7877,11 @@ "per_request_limits": null, "supported_parameters": [ "max_tokens", - "temperature", - "top_p", - "seed", + "presence_penalty", "response_format", - "presence_penalty" + "seed", + "temperature", + "top_p" ] }, { @@ -7572,6 +7908,7 @@ "completion": "0.000008", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -7582,11 +7919,11 @@ }, "per_request_limits": null, "supported_parameters": [ + "include_reasoning", "max_tokens", - "temperature", - "top_p", "reasoning", - "include_reasoning" + "temperature", + "top_p" ] }, { @@ -7613,6 +7950,7 @@ "completion": "0.0000014", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -7623,11 +7961,11 @@ }, "per_request_limits": null, "supported_parameters": [ + "include_reasoning", "max_tokens", - "temperature", - "top_p", "reasoning", - "include_reasoning" + "temperature", + "top_p" ] }, { @@ -7654,6 +7992,7 @@ "completion": "0.0000002", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -7694,6 +8033,7 @@ "completion": "0.0000032", "request": "0", "image": "0.001024", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -7705,11 +8045,11 @@ "per_request_limits": null, "supported_parameters": [ "max_tokens", - "temperature", - "top_p", - "seed", + "presence_penalty", "response_format", - "presence_penalty" + "seed", + "temperature", + "top_p" ] }, { @@ -7736,6 +8076,7 @@ "completion": "0.0000002", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0", "input_cache_read": "0.00000002" @@ -7748,13 +8089,13 @@ "per_request_limits": null, "supported_parameters": [ "max_tokens", - "temperature", - "top_p", - "tools", - "tool_choice", - "seed", + "presence_penalty", "response_format", - "presence_penalty" + "seed", + "temperature", + "tool_choice", + "tools", + "top_p" ] }, { @@ -7782,6 +8123,7 @@ "completion": "0", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -7792,16 +8134,16 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "structured_outputs", - "response_format", - "stop", "frequency_penalty", + "max_tokens", "presence_penalty", + "response_format", + "seed", + "stop", + "structured_outputs", + "temperature", "top_k", - "seed" + "top_p" ] }, { @@ -7829,6 +8171,7 @@ "completion": "0.00000075", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -7839,20 +8182,19 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "stop", "frequency_penalty", - "presence_penalty", - "seed", - "top_k", - "min_p", - "repetition_penalty", "logit_bias", - "response_format", "logprobs", - "top_logprobs" + "max_tokens", + "min_p", + "presence_penalty", + "repetition_penalty", + "seed", + "stop", + "temperature", + "top_k", + "top_logprobs", + "top_p" ] }, { @@ -7879,6 +8221,7 @@ "completion": "0.0000012", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0", "input_cache_read": "0.00000016" @@ -7891,13 +8234,13 @@ "per_request_limits": null, "supported_parameters": [ "max_tokens", - "temperature", - "top_p", - "tools", - "tool_choice", - "seed", + "presence_penalty", "response_format", - "presence_penalty" + "seed", + "temperature", + "tool_choice", + "tools", + "top_p" ] }, { @@ -7924,6 +8267,7 @@ "completion": "0.0000064", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0", "input_cache_read": "0.00000064" @@ -7936,13 +8280,13 @@ "per_request_limits": null, "supported_parameters": [ "max_tokens", - "temperature", - "top_p", - "tools", - "tool_choice", - "seed", + "presence_penalty", "response_format", - "presence_penalty" + "seed", + "temperature", + "tool_choice", + "tools", + "top_p" ] }, { @@ -7969,6 +8313,7 @@ "completion": "0.0000044", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0", "input_cache_read": "0.00000055" @@ -7980,12 +8325,12 @@ }, "per_request_limits": null, "supported_parameters": [ - "tools", - "tool_choice", - "seed", "max_tokens", "response_format", - "structured_outputs" + "seed", + "structured_outputs", + "tool_choice", + "tools" ] }, { @@ -8012,6 +8357,7 @@ "completion": "0.00000018", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -8022,19 +8368,18 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "reasoning", - "include_reasoning", - "stop", "frequency_penalty", - "presence_penalty", - "top_k", - "repetition_penalty", + "include_reasoning", "logit_bias", + "max_tokens", "min_p", - "response_format" + "presence_penalty", + "reasoning", + "repetition_penalty", + "stop", + "temperature", + "top_k", + "top_p" ] }, { @@ -8061,6 +8406,7 @@ "completion": "0", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -8071,19 +8417,19 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "stop", "frequency_penalty", - "presence_penalty", - "seed", - "top_k", - "min_p", - "repetition_penalty", - "logprobs", "logit_bias", - "top_logprobs" + "logprobs", + "max_tokens", + "min_p", + "presence_penalty", + "repetition_penalty", + "seed", + "stop", + "temperature", + "top_k", + "top_logprobs", + "top_p" ] }, { @@ -8106,10 +8452,11 @@ "instruct_type": null }, "pricing": { - "prompt": "0.00000003", - "completion": "0.00000003", + "prompt": "0.00000001999188", + "completion": "0.0000000800064", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -8120,23 +8467,23 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "stop", "frequency_penalty", - "presence_penalty", - "seed", - "top_k", - "min_p", - "repetition_penalty", - "logprobs", "logit_bias", - "top_logprobs", + "logprobs", + "max_tokens", + "min_p", + "presence_penalty", + "repetition_penalty", "response_format", - "tools", + "seed", + "stop", + "structured_outputs", + "temperature", "tool_choice", - "structured_outputs" + "tools", + "top_k", + "top_logprobs", + "top_p" ] }, { @@ -8163,6 +8510,7 @@ "completion": "0.00000015", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -8173,20 +8521,20 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "reasoning", + "frequency_penalty", "include_reasoning", + "logit_bias", + "max_tokens", + "min_p", + "presence_penalty", + "reasoning", + "repetition_penalty", + "response_format", "seed", "stop", - "frequency_penalty", - "presence_penalty", + "temperature", "top_k", - "min_p", - "repetition_penalty", - "logit_bias", - "response_format" + "top_p" ] }, { @@ -8213,6 +8561,7 @@ "completion": "0", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -8223,21 +8572,21 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "reasoning", - "include_reasoning", - "stop", "frequency_penalty", - "presence_penalty", - "seed", - "top_k", - "min_p", - "repetition_penalty", - "logprobs", + "include_reasoning", "logit_bias", - "top_logprobs" + "logprobs", + "max_tokens", + "min_p", + "presence_penalty", + "reasoning", + "repetition_penalty", + "seed", + "stop", + "temperature", + "top_k", + "top_logprobs", + "top_p" ] }, { @@ -8264,6 +8613,7 @@ "completion": "0.00000015", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -8274,20 +8624,19 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "reasoning", + "frequency_penalty", "include_reasoning", + "logit_bias", + "max_tokens", + "min_p", + "presence_penalty", + "reasoning", + "repetition_penalty", "seed", "stop", - "frequency_penalty", - "presence_penalty", + "temperature", "top_k", - "min_p", - "repetition_penalty", - "logit_bias", - "response_format" + "top_p" ] }, { @@ -8314,6 +8663,7 @@ "completion": "0.000005", "request": "0.005", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -8324,15 +8674,15 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "reasoning", - "include_reasoning", - "web_search_options", - "top_k", "frequency_penalty", - "presence_penalty" + "include_reasoning", + "max_tokens", + "presence_penalty", + "reasoning", + "temperature", + "top_k", + "top_p", + "web_search_options" ] }, { @@ -8360,6 +8710,7 @@ "completion": "0.000001", "request": "0.005", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -8370,13 +8721,13 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "web_search_options", - "top_k", "frequency_penalty", - "presence_penalty" + "max_tokens", + "presence_penalty", + "temperature", + "top_k", + "top_p", + "web_search_options" ] }, { @@ -8403,6 +8754,7 @@ "completion": "0.00000001", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -8413,20 +8765,20 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "stop", "frequency_penalty", - "presence_penalty", - "seed", - "top_k", - "min_p", - "repetition_penalty", "logit_bias", "logprobs", + "max_tokens", + "min_p", + "presence_penalty", + "repetition_penalty", + "response_format", + "seed", + "stop", + "temperature", + "top_k", "top_logprobs", - "response_format" + "top_p" ] }, { @@ -8453,6 +8805,7 @@ "completion": "0.00000002", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -8463,16 +8816,16 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "stop", "frequency_penalty", - "presence_penalty", - "seed", - "top_k", + "max_tokens", "min_p", - "repetition_penalty" + "presence_penalty", + "repetition_penalty", + "seed", + "stop", + "temperature", + "top_k", + "top_p" ] }, { @@ -8499,6 +8852,7 @@ "completion": "0", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -8509,22 +8863,21 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "reasoning", - "include_reasoning", - "stop", "frequency_penalty", - "presence_penalty", - "top_k", - "repetition_penalty", + "include_reasoning", "logit_bias", - "min_p", - "response_format", - "seed", "logprobs", - "top_logprobs" + "max_tokens", + "min_p", + "presence_penalty", + "reasoning", + "repetition_penalty", + "seed", + "stop", + "temperature", + "top_k", + "top_logprobs", + "top_p" ] }, { @@ -8547,10 +8900,11 @@ "instruct_type": "deepseek-r1" }, "pricing": { - "prompt": "0.00000005", - "completion": "0.00000005", + "prompt": "0.0000000333198", + "completion": "0.000000133344", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -8561,25 +8915,25 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "reasoning", + "frequency_penalty", "include_reasoning", + "logit_bias", + "logprobs", + "max_tokens", + "min_p", + "presence_penalty", + "reasoning", + "repetition_penalty", + "response_format", "seed", "stop", - "frequency_penalty", - "presence_penalty", - "top_k", - "min_p", - "repetition_penalty", - "logprobs", - "logit_bias", - "top_logprobs", - "tools", + "structured_outputs", + "temperature", "tool_choice", - "response_format", - "structured_outputs" + "tools", + "top_k", + "top_logprobs", + "top_p" ] }, { @@ -8606,6 +8960,7 @@ "completion": "0", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -8616,9 +8971,9 @@ }, "per_request_limits": null, "supported_parameters": [ + "include_reasoning", "max_tokens", "reasoning", - "include_reasoning", "temperature" ] }, @@ -8646,6 +9001,7 @@ "completion": "0.000002", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -8656,25 +9012,25 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "reasoning", - "include_reasoning", - "stop", "frequency_penalty", - "presence_penalty", - "seed", - "top_k", - "min_p", + "include_reasoning", "logit_bias", - "top_logprobs", - "response_format", - "structured_outputs", - "repetition_penalty", "logprobs", + "max_tokens", + "min_p", + "presence_penalty", + "reasoning", + "repetition_penalty", + "response_format", + "seed", + "stop", + "structured_outputs", + "temperature", + "tool_choice", "tools", - "tool_choice" + "top_k", + "top_logprobs", + "top_p" ] }, { @@ -8702,6 +9058,7 @@ "completion": "0.0000011", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -8741,6 +9098,7 @@ "completion": "0.0000009", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -8751,17 +9109,17 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "tools", - "tool_choice", - "stop", "frequency_penalty", + "max_tokens", "presence_penalty", "response_format", + "seed", + "stop", "structured_outputs", - "seed" + "temperature", + "tool_choice", + "tools", + "top_p" ] }, { @@ -8788,6 +9146,7 @@ "completion": "0.00000014", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -8798,21 +9157,21 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "stop", "frequency_penalty", - "presence_penalty", - "response_format", - "structured_outputs", - "seed", - "top_k", "logit_bias", "logprobs", - "top_logprobs", + "max_tokens", + "min_p", + "presence_penalty", "repetition_penalty", - "min_p" + "response_format", + "seed", + "stop", + "structured_outputs", + "temperature", + "top_k", + "top_logprobs", + "top_p" ] }, { @@ -8835,10 +9194,11 @@ "instruct_type": null }, "pricing": { - "prompt": "0.000000272", - "completion": "0.000000272", + "prompt": "0.00000017992692", + "completion": "0.0000007200576", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -8849,23 +9209,23 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "tools", - "tool_choice", - "structured_outputs", - "response_format", - "stop", "frequency_penalty", - "presence_penalty", - "top_k", - "repetition_penalty", "logit_bias", "logprobs", - "top_logprobs", + "max_tokens", + "min_p", + "presence_penalty", + "repetition_penalty", + "response_format", "seed", - "min_p" + "stop", + "structured_outputs", + "temperature", + "tool_choice", + "tools", + "top_k", + "top_logprobs", + "top_p" ] }, { @@ -8892,6 +9252,7 @@ "completion": "0.00000075", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -8902,19 +9263,19 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "stop", "frequency_penalty", - "presence_penalty", - "response_format", - "structured_outputs", - "repetition_penalty", - "top_k", - "seed", + "logit_bias", + "max_tokens", "min_p", - "logit_bias" + "presence_penalty", + "repetition_penalty", + "response_format", + "seed", + "stop", + "structured_outputs", + "temperature", + "top_k", + "top_p" ] }, { @@ -8942,6 +9303,7 @@ "completion": "0.00006", "request": "0", "image": "0.021675", + "audio": "0", "web_search": "0", "internal_reasoning": "0", "input_cache_read": "0.0000075" @@ -8953,12 +9315,12 @@ }, "per_request_limits": null, "supported_parameters": [ - "tools", - "tool_choice", - "seed", "max_tokens", "response_format", - "structured_outputs" + "seed", + "structured_outputs", + "tool_choice", + "tools" ] }, { @@ -8986,6 +9348,7 @@ "completion": "0.00001", "request": "0", "image": "0.0036", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -8996,16 +9359,16 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "stop", "frequency_penalty", - "presence_penalty", - "seed", "logprobs", + "max_tokens", + "presence_penalty", + "response_format", + "seed", + "stop", + "temperature", "top_logprobs", - "response_format" + "top_p" ] }, { @@ -9032,6 +9395,7 @@ "completion": "0.00001", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -9042,18 +9406,18 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "tools", - "tool_choice", - "stop", "frequency_penalty", - "presence_penalty", - "seed", "logprobs", + "max_tokens", + "presence_penalty", + "response_format", + "seed", + "stop", + "temperature", + "tool_choice", + "tools", "top_logprobs", - "response_format" + "top_p" ] }, { @@ -9080,6 +9444,7 @@ "completion": "0.00000015", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -9090,16 +9455,16 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "stop", "frequency_penalty", + "max_tokens", "presence_penalty", - "top_k", - "seed", "response_format", - "structured_outputs" + "seed", + "stop", + "structured_outputs", + "temperature", + "top_k", + "top_p" ] }, { @@ -9127,6 +9492,7 @@ "completion": "0", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -9138,11 +9504,13 @@ "per_request_limits": null, "supported_parameters": [ "max_tokens", + "response_format", + "seed", + "stop", "temperature", - "top_p", - "tools", "tool_choice", - "stop" + "tools", + "top_p" ] }, { @@ -9169,6 +9537,7 @@ "completion": "0", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -9179,19 +9548,18 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "tools", - "tool_choice", - "stop", "frequency_penalty", - "presence_penalty", - "top_k", - "repetition_penalty", "logit_bias", + "max_tokens", "min_p", - "response_format" + "presence_penalty", + "repetition_penalty", + "stop", + "temperature", + "tool_choice", + "tools", + "top_k", + "top_p" ] }, { @@ -9218,6 +9586,7 @@ "completion": "0.00000012", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -9228,23 +9597,23 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "tools", - "tool_choice", - "stop", "frequency_penalty", + "logit_bias", + "logprobs", + "max_tokens", + "min_p", "presence_penalty", "repetition_penalty", "response_format", - "top_k", "seed", - "min_p", + "stop", "structured_outputs", - "logit_bias", - "logprobs", - "top_logprobs" + "temperature", + "tool_choice", + "tools", + "top_k", + "top_logprobs", + "top_p" ] }, { @@ -9272,6 +9641,7 @@ "completion": "0.00000024", "request": "0", "image": "0.00009", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -9283,11 +9653,11 @@ "per_request_limits": null, "supported_parameters": [ "max_tokens", + "stop", "temperature", - "top_p", "tools", "top_k", - "stop" + "top_p" ] }, { @@ -9314,6 +9684,7 @@ "completion": "0.00000014", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -9325,11 +9696,11 @@ "per_request_limits": null, "supported_parameters": [ "max_tokens", + "stop", "temperature", - "top_p", "tools", "top_k", - "stop" + "top_p" ] }, { @@ -9357,6 +9728,7 @@ "completion": "0.0000032", "request": "0", "image": "0.0012", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -9368,11 +9740,11 @@ "per_request_limits": null, "supported_parameters": [ "max_tokens", + "stop", "temperature", - "top_p", "tools", "top_k", - "stop" + "top_p" ] }, { @@ -9399,6 +9771,7 @@ "completion": "0.0000002", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -9409,65 +9782,19 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "stop", "frequency_penalty", - "presence_penalty", - "logprobs", - "top_logprobs", - "seed", "logit_bias", - "top_k", - "min_p", - "repetition_penalty" - ] - }, - { - "id": "eva-unit-01/eva-qwen-2.5-72b", - "canonical_slug": "eva-unit-01/eva-qwen-2.5-72b", - "hugging_face_id": "EVA-UNIT-01/EVA-Qwen2.5-72B-v0.1", - "name": "EVA Qwen2.5 72B", - "created": 1732210606, - "description": "EVA Qwen2.5 72B is a roleplay and storywriting specialist model. It's a full-parameter finetune of Qwen2.5-72B on mixture of synthetic and natural data.\n\nIt uses Celeste 70B 0.1 data mixture, greatly expanding it to improve versatility, creativity and \"flavor\" of the resulting model.", - "context_length": 16384, - "architecture": { - "modality": "text->text", - "input_modalities": [ - "text" - ], - "output_modalities": [ - "text" - ], - "tokenizer": "Qwen", - "instruct_type": "chatml" - }, - "pricing": { - "prompt": "0.000004", - "completion": "0.000006", - "request": "0", - "image": "0", - "web_search": "0", - "internal_reasoning": "0" - }, - "top_provider": { - "context_length": 16384, - "max_completion_tokens": 4096, - "is_moderated": false - }, - "per_request_limits": null, - "supported_parameters": [ + "logprobs", "max_tokens", - "temperature", - "top_p", - "stop", - "frequency_penalty", + "min_p", "presence_penalty", "repetition_penalty", + "seed", + "stop", + "temperature", "top_k", - "min_p", - "seed" + "top_logprobs", + "top_p" ] }, { @@ -9496,6 +9823,7 @@ "completion": "0.00001", "request": "0", "image": "0.003613", + "audio": "0", "web_search": "0", "internal_reasoning": "0", "input_cache_read": "0.00000125" @@ -9507,21 +9835,21 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "tools", - "tool_choice", - "stop", "frequency_penalty", - "presence_penalty", - "web_search_options", - "seed", "logit_bias", "logprobs", - "top_logprobs", + "max_tokens", + "presence_penalty", "response_format", - "structured_outputs" + "seed", + "stop", + "structured_outputs", + "temperature", + "tool_choice", + "tools", + "top_logprobs", + "top_p", + "web_search_options" ] }, { @@ -9548,6 +9876,7 @@ "completion": "0.000006", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -9558,17 +9887,17 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "tools", - "tool_choice", - "stop", "frequency_penalty", + "max_tokens", "presence_penalty", "response_format", + "seed", + "stop", "structured_outputs", - "seed" + "temperature", + "tool_choice", + "tools", + "top_p" ] }, { @@ -9595,6 +9924,7 @@ "completion": "0.000006", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -9605,17 +9935,17 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "tools", - "tool_choice", - "stop", "frequency_penalty", + "max_tokens", "presence_penalty", "response_format", + "seed", + "stop", "structured_outputs", - "seed" + "temperature", + "tool_choice", + "tools", + "top_p" ] }, { @@ -9643,6 +9973,7 @@ "completion": "0.000006", "request": "0", "image": "0.002888", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -9653,17 +9984,17 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "tools", - "tool_choice", - "stop", "frequency_penalty", + "max_tokens", "presence_penalty", "response_format", + "seed", + "stop", "structured_outputs", - "seed" + "temperature", + "tool_choice", + "tools", + "top_p" ] }, { @@ -9691,6 +10022,7 @@ "completion": "0.000015", "request": "0", "image": "0.009", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -9701,16 +10033,16 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "stop", "frequency_penalty", - "presence_penalty", - "seed", "logprobs", + "max_tokens", + "presence_penalty", + "response_format", + "seed", + "stop", + "temperature", "top_logprobs", - "response_format" + "top_p" ] }, { @@ -9720,7 +10052,7 @@ "name": "Infermatic: Mistral Nemo Inferor 12B", "created": 1731464428, "description": "Inferor 12B is a merge of top roleplay models, expert on immersive narratives and storytelling.\n\nThis model was merged using the [Model Stock](https://arxiv.org/abs/2403.19522) merge method using [anthracite-org/magnum-v4-12b](https://openrouter.ai/anthracite-org/magnum-v4-72b) as a base.\n", - "context_length": 16384, + "context_length": 8192, "architecture": { "modality": "text->text", "input_modalities": [ @@ -9733,30 +10065,31 @@ "instruct_type": "mistral" }, "pricing": { - "prompt": "0.0000008", - "completion": "0.0000012", + "prompt": "0.0000006", + "completion": "0.000001", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, "top_provider": { - "context_length": 16384, - "max_completion_tokens": 4096, + "context_length": 8192, + "max_completion_tokens": 8192, "is_moderated": false }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "stop", "frequency_penalty", + "max_tokens", + "min_p", "presence_penalty", "repetition_penalty", + "seed", + "stop", + "temperature", "top_k", - "min_p", - "seed" + "top_p" ] }, { @@ -9783,6 +10116,7 @@ "completion": "0", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -9793,19 +10127,19 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "stop", "frequency_penalty", - "presence_penalty", - "top_k", - "seed", - "min_p", - "repetition_penalty", - "logprobs", "logit_bias", - "top_logprobs" + "logprobs", + "max_tokens", + "min_p", + "presence_penalty", + "repetition_penalty", + "seed", + "stop", + "temperature", + "top_k", + "top_logprobs", + "top_p" ] }, { @@ -9828,34 +10162,35 @@ "instruct_type": "chatml" }, "pricing": { - "prompt": "0.00000006", - "completion": "0.00000015", + "prompt": "0.0000000499797", + "completion": "0.000000200016", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, "top_provider": { "context_length": 32768, - "max_completion_tokens": 16384, + "max_completion_tokens": null, "is_moderated": false }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "stop", "frequency_penalty", - "presence_penalty", - "seed", - "top_k", - "min_p", - "repetition_penalty", - "logprobs", "logit_bias", + "logprobs", + "max_tokens", + "min_p", + "presence_penalty", + "repetition_penalty", + "response_format", + "seed", + "stop", + "temperature", + "top_k", "top_logprobs", - "response_format" + "top_p" ] }, { @@ -9882,6 +10217,7 @@ "completion": "0.0000045", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -9892,17 +10228,17 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "stop", "frequency_penalty", + "logit_bias", + "max_tokens", + "min_p", "presence_penalty", "repetition_penalty", - "logit_bias", + "seed", + "stop", + "temperature", "top_k", - "min_p", - "seed" + "top_p" ] }, { @@ -9929,6 +10265,7 @@ "completion": "0.0000004", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -9939,22 +10276,22 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "stop", "frequency_penalty", - "presence_penalty", - "response_format", - "structured_outputs", - "repetition_penalty", "logit_bias", - "top_k", + "logprobs", + "max_tokens", "min_p", + "presence_penalty", + "repetition_penalty", + "response_format", "seed", - "tools", + "stop", + "structured_outputs", + "temperature", "tool_choice", - "logprobs" + "tools", + "top_k", + "top_p" ] }, { @@ -9982,6 +10319,7 @@ "completion": "0.000004", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0", "input_cache_read": "0.00000008", @@ -9995,12 +10333,12 @@ "per_request_limits": null, "supported_parameters": [ "max_tokens", + "stop", "temperature", - "top_p", - "tools", "tool_choice", + "tools", "top_k", - "stop" + "top_p" ] }, { @@ -10028,6 +10366,7 @@ "completion": "0.000004", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0", "input_cache_read": "0.00000008", @@ -10041,58 +10380,12 @@ "per_request_limits": null, "supported_parameters": [ "max_tokens", + "stop", "temperature", - "top_p", - "tools", "tool_choice", - "top_k", - "stop" - ] - }, - { - "id": "anthropic/claude-3.5-haiku-20241022:beta", - "canonical_slug": "anthropic/claude-3-5-haiku-20241022", - "hugging_face_id": null, - "name": "Anthropic: Claude 3.5 Haiku (2024-10-22) (self-moderated)", - "created": 1730678400, - "description": "Claude 3.5 Haiku features enhancements across all skill sets including coding, tool use, and reasoning. As the fastest model in the Anthropic lineup, it offers rapid response times suitable for applications that require high interactivity and low latency, such as user-facing chatbots and on-the-fly code completions. It also excels in specialized tasks like data extraction and real-time content moderation, making it a versatile tool for a broad range of industries.\n\nIt does not support image inputs.\n\nSee the launch announcement and benchmark results [here](https://www.anthropic.com/news/3-5-models-and-computer-use)", - "context_length": 200000, - "architecture": { - "modality": "text+image->text", - "input_modalities": [ - "text", - "image" - ], - "output_modalities": [ - "text" - ], - "tokenizer": "Claude", - "instruct_type": null - }, - "pricing": { - "prompt": "0.0000008", - "completion": "0.000004", - "request": "0", - "image": "0", - "web_search": "0", - "internal_reasoning": "0", - "input_cache_read": "0.00000008", - "input_cache_write": "0.000001" - }, - "top_provider": { - "context_length": 200000, - "max_completion_tokens": 8192, - "is_moderated": false - }, - "per_request_limits": null, - "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", "tools", - "tool_choice", "top_k", - "stop" + "top_p" ] }, { @@ -10120,6 +10413,7 @@ "completion": "0.000004", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0", "input_cache_read": "0.00000008", @@ -10128,17 +10422,17 @@ "top_provider": { "context_length": 200000, "max_completion_tokens": 8192, - "is_moderated": true + "is_moderated": false }, "per_request_limits": null, "supported_parameters": [ "max_tokens", + "stop", "temperature", - "top_p", - "tools", "tool_choice", + "tools", "top_k", - "stop" + "top_p" ] }, { @@ -10165,6 +10459,7 @@ "completion": "0.000003", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -10175,18 +10470,18 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "stop", "frequency_penalty", + "logit_bias", + "max_tokens", + "min_p", "presence_penalty", "repetition_penalty", - "logit_bias", - "top_k", - "min_p", "seed", - "top_a" + "stop", + "temperature", + "top_a", + "top_k", + "top_p" ] }, { @@ -10214,6 +10509,7 @@ "completion": "0.000015", "request": "0", "image": "0.0048", + "audio": "0", "web_search": "0", "internal_reasoning": "0", "input_cache_read": "0.0000003", @@ -10227,12 +10523,12 @@ "per_request_limits": null, "supported_parameters": [ "max_tokens", + "stop", "temperature", - "top_p", - "tools", "tool_choice", + "tools", "top_k", - "stop" + "top_p" ] }, { @@ -10260,6 +10556,7 @@ "completion": "0.000015", "request": "0", "image": "0.0048", + "audio": "0", "web_search": "0", "internal_reasoning": "0", "input_cache_read": "0.0000003", @@ -10273,12 +10570,12 @@ "per_request_limits": null, "supported_parameters": [ "max_tokens", + "stop", "temperature", - "top_p", - "tools", "tool_choice", + "tools", "top_k", - "stop" + "top_p" ] }, { @@ -10305,6 +10602,7 @@ "completion": "0.0000001", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -10315,17 +10613,17 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "tools", - "tool_choice", - "stop", "frequency_penalty", + "max_tokens", "presence_penalty", "response_format", + "seed", + "stop", "structured_outputs", - "seed" + "temperature", + "tool_choice", + "tools", + "top_p" ] }, { @@ -10335,7 +10633,7 @@ "name": "Mistral: Ministral 3B", "created": 1729123200, "description": "Ministral 3B is a 3B parameter model optimized for on-device and edge computing. It excels in knowledge, commonsense reasoning, and function-calling, outperforming larger models like Mistral 7B on most benchmarks. Supporting up to 128k context length, it’s ideal for orchestrating agentic workflows and specialist tasks with efficient inference.", - "context_length": 131072, + "context_length": 32768, "architecture": { "modality": "text->text", "input_modalities": [ @@ -10352,27 +10650,26 @@ "completion": "0.00000004", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, "top_provider": { - "context_length": 131072, + "context_length": 32768, "max_completion_tokens": null, "is_moderated": false }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "tools", - "tool_choice", - "stop", "frequency_penalty", + "max_tokens", "presence_penalty", "response_format", + "seed", + "stop", "structured_outputs", - "seed" + "temperature", + "top_p" ] }, { @@ -10382,7 +10679,7 @@ "name": "Qwen2.5 7B Instruct", "created": 1729036800, "description": "Qwen2.5 7B is the latest series of Qwen large language models. Qwen2.5 brings the following improvements upon Qwen2:\n\n- Significantly more knowledge and has greatly improved capabilities in coding and mathematics, thanks to our specialized expert models in these domains.\n\n- Significant improvements in instruction following, generating long texts (over 8K tokens), understanding structured data (e.g, tables), and generating structured outputs especially JSON. More resilient to the diversity of system prompts, enhancing role-play implementation and condition-setting for chatbots.\n\n- Long-context Support up to 128K tokens and can generate up to 8K tokens.\n\n- Multilingual support for over 29 languages, including Chinese, English, French, Spanish, Portuguese, German, Italian, Russian, Japanese, Korean, Vietnamese, Thai, Arabic, and more.\n\nUsage of this model is subject to [Tongyi Qianwen LICENSE AGREEMENT](https://huggingface.co/Qwen/Qwen1.5-110B-Chat/blob/main/LICENSE).", - "context_length": 32768, + "context_length": 65536, "architecture": { "modality": "text->text", "input_modalities": [ @@ -10399,28 +10696,30 @@ "completion": "0.0000001", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, "top_provider": { - "context_length": 32768, - "max_completion_tokens": 16384, + "context_length": 65536, + "max_completion_tokens": null, "is_moderated": false }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "stop", "frequency_penalty", - "presence_penalty", - "top_k", - "repetition_penalty", "logit_bias", + "max_tokens", "min_p", + "presence_penalty", + "repetition_penalty", "response_format", - "seed" + "seed", + "stop", + "structured_outputs", + "temperature", + "top_k", + "top_p" ] }, { @@ -10447,6 +10746,7 @@ "completion": "0.0000003", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -10457,62 +10757,22 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "tools", - "tool_choice", - "stop", "frequency_penalty", - "presence_penalty", - "seed", "logit_bias", "logprobs", - "top_logprobs", - "response_format", - "min_p", - "repetition_penalty", - "top_k" - ] - }, - { - "id": "inflection/inflection-3-pi", - "canonical_slug": "inflection/inflection-3-pi", - "hugging_face_id": null, - "name": "Inflection: Inflection 3 Pi", - "created": 1728604800, - "description": "Inflection 3 Pi powers Inflection's [Pi](https://pi.ai) chatbot, including backstory, emotional intelligence, productivity, and safety. It has access to recent news, and excels in scenarios like customer support and roleplay.\n\nPi has been trained to mirror your tone and style, if you use more emojis, so will Pi! Try experimenting with various prompts and conversation styles.", - "context_length": 8000, - "architecture": { - "modality": "text->text", - "input_modalities": [ - "text" - ], - "output_modalities": [ - "text" - ], - "tokenizer": "Other", - "instruct_type": null - }, - "pricing": { - "prompt": "0.0000025", - "completion": "0.00001", - "request": "0", - "image": "0", - "web_search": "0", - "internal_reasoning": "0" - }, - "top_provider": { - "context_length": 8000, - "max_completion_tokens": 1024, - "is_moderated": false - }, - "per_request_limits": null, - "supported_parameters": [ "max_tokens", + "min_p", + "presence_penalty", + "repetition_penalty", + "response_format", + "seed", + "stop", "temperature", - "top_p", - "stop" + "tool_choice", + "tools", + "top_k", + "top_logprobs", + "top_p" ] }, { @@ -10539,6 +10799,7 @@ "completion": "0.00001", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -10550,9 +10811,50 @@ "per_request_limits": null, "supported_parameters": [ "max_tokens", + "stop", "temperature", - "top_p", - "stop" + "top_p" + ] + }, + { + "id": "inflection/inflection-3-pi", + "canonical_slug": "inflection/inflection-3-pi", + "hugging_face_id": null, + "name": "Inflection: Inflection 3 Pi", + "created": 1728604800, + "description": "Inflection 3 Pi powers Inflection's [Pi](https://pi.ai) chatbot, including backstory, emotional intelligence, productivity, and safety. It has access to recent news, and excels in scenarios like customer support and roleplay.\n\nPi has been trained to mirror your tone and style, if you use more emojis, so will Pi! Try experimenting with various prompts and conversation styles.", + "context_length": 8000, + "architecture": { + "modality": "text->text", + "input_modalities": [ + "text" + ], + "output_modalities": [ + "text" + ], + "tokenizer": "Other", + "instruct_type": null + }, + "pricing": { + "prompt": "0.0000025", + "completion": "0.00001", + "request": "0", + "image": "0", + "audio": "0", + "web_search": "0", + "internal_reasoning": "0" + }, + "top_provider": { + "context_length": 8000, + "max_completion_tokens": 1024, + "is_moderated": false + }, + "per_request_limits": null, + "supported_parameters": [ + "max_tokens", + "stop", + "temperature", + "top_p" ] }, { @@ -10580,6 +10882,7 @@ "completion": "0.00000015", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0", "input_cache_read": "0.00000001", @@ -10592,64 +10895,17 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "tools", - "tool_choice", - "stop", "frequency_penalty", + "max_tokens", "presence_penalty", - "seed", "response_format", - "structured_outputs" - ] - }, - { - "id": "anthracite-org/magnum-v2-72b", - "canonical_slug": "anthracite-org/magnum-v2-72b", - "hugging_face_id": "anthracite-org/magnum-v2-72b", - "name": "Magnum v2 72B", - "created": 1727654400, - "description": "From the maker of [Goliath](https://openrouter.ai/models/alpindale/goliath-120b), Magnum 72B is the seventh in a family of models designed to achieve the prose quality of the Claude 3 models, notably Opus & Sonnet.\n\nThe model is based on [Qwen2 72B](https://openrouter.ai/models/qwen/qwen-2-72b-instruct) and trained with 55 million tokens of highly curated roleplay (RP) data.", - "context_length": 32768, - "architecture": { - "modality": "text->text", - "input_modalities": [ - "text" - ], - "output_modalities": [ - "text" - ], - "tokenizer": "Qwen", - "instruct_type": "chatml" - }, - "pricing": { - "prompt": "0.000003", - "completion": "0.000003", - "request": "0", - "image": "0", - "web_search": "0", - "internal_reasoning": "0" - }, - "top_provider": { - "context_length": 32768, - "max_completion_tokens": null, - "is_moderated": false - }, - "per_request_limits": null, - "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", + "seed", "stop", - "frequency_penalty", - "presence_penalty", - "repetition_penalty", - "logit_bias", - "top_k", - "min_p", - "seed" + "structured_outputs", + "temperature", + "tool_choice", + "tools", + "top_p" ] }, { @@ -10659,7 +10915,7 @@ "name": "TheDrummer: Rocinante 12B", "created": 1727654400, "description": "Rocinante 12B is designed for engaging storytelling and rich prose.\n\nEarly testers have reported:\n- Expanded vocabulary with unique and expressive word choices\n- Enhanced creativity for vivid narratives\n- Adventure-filled and captivating stories", - "context_length": 32768, + "context_length": 8192, "architecture": { "modality": "text->text", "input_modalities": [ @@ -10672,36 +10928,37 @@ "instruct_type": "chatml" }, "pricing": { - "prompt": "0.0000002", - "completion": "0.0000005", + "prompt": "0.00000019", + "completion": "0.00000045", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, "top_provider": { - "context_length": 32768, - "max_completion_tokens": null, + "context_length": 8192, + "max_completion_tokens": 8192, "is_moderated": false }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "stop", "frequency_penalty", - "presence_penalty", - "response_format", - "structured_outputs", - "tools", - "tool_choice", "logit_bias", "logprobs", - "seed", + "max_tokens", + "min_p", + "presence_penalty", "repetition_penalty", + "response_format", + "seed", + "stop", + "structured_outputs", + "temperature", + "tool_choice", + "tools", "top_k", - "min_p" + "top_p" ] }, { @@ -10728,6 +10985,7 @@ "completion": "0.00000015", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -10738,20 +10996,68 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "stop", "frequency_penalty", - "presence_penalty", - "seed", "logit_bias", "logprobs", - "top_logprobs", - "response_format", + "max_tokens", "min_p", + "presence_penalty", "repetition_penalty", - "top_k" + "response_format", + "seed", + "stop", + "temperature", + "top_k", + "top_logprobs", + "top_p" + ] + }, + { + "id": "anthracite-org/magnum-v2-72b", + "canonical_slug": "anthracite-org/magnum-v2-72b", + "hugging_face_id": "anthracite-org/magnum-v2-72b", + "name": "Magnum v2 72B", + "created": 1727654400, + "description": "From the maker of [Goliath](https://openrouter.ai/models/alpindale/goliath-120b), Magnum 72B is the seventh in a family of models designed to achieve the prose quality of the Claude 3 models, notably Opus & Sonnet.\n\nThe model is based on [Qwen2 72B](https://openrouter.ai/models/qwen/qwen-2-72b-instruct) and trained with 55 million tokens of highly curated roleplay (RP) data.", + "context_length": 32768, + "architecture": { + "modality": "text->text", + "input_modalities": [ + "text" + ], + "output_modalities": [ + "text" + ], + "tokenizer": "Qwen", + "instruct_type": "chatml" + }, + "pricing": { + "prompt": "0.000003", + "completion": "0.000003", + "request": "0", + "image": "0", + "audio": "0", + "web_search": "0", + "internal_reasoning": "0" + }, + "top_provider": { + "context_length": 32768, + "max_completion_tokens": null, + "is_moderated": false + }, + "per_request_limits": null, + "supported_parameters": [ + "frequency_penalty", + "logit_bias", + "max_tokens", + "min_p", + "presence_penalty", + "repetition_penalty", + "seed", + "stop", + "temperature", + "top_k", + "top_p" ] }, { @@ -10778,6 +11084,7 @@ "completion": "0", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -10788,13 +11095,13 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "stop", "frequency_penalty", + "max_tokens", "presence_penalty", - "top_k" + "stop", + "temperature", + "top_k", + "top_p" ] }, { @@ -10821,6 +11128,7 @@ "completion": "0.000000006", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -10831,172 +11139,23 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "tools", - "tool_choice", - "stop", "frequency_penalty", - "presence_penalty", - "seed", "logit_bias", "logprobs", - "top_logprobs", - "response_format", - "min_p", - "repetition_penalty", - "top_k" - ] - }, - { - "id": "meta-llama/llama-3.2-1b-instruct", - "canonical_slug": "meta-llama/llama-3.2-1b-instruct", - "hugging_face_id": "meta-llama/Llama-3.2-1B-Instruct", - "name": "Meta: Llama 3.2 1B Instruct", - "created": 1727222400, - "description": "Llama 3.2 1B is a 1-billion-parameter language model focused on efficiently performing natural language tasks, such as summarization, dialogue, and multilingual text analysis. Its smaller size allows it to operate efficiently in low-resource environments while maintaining strong task performance.\n\nSupporting eight core languages and fine-tunable for more, Llama 1.3B is ideal for businesses or developers seeking lightweight yet powerful AI solutions that can operate in diverse multilingual settings without the high computational demand of larger models.\n\nClick here for the [original model card](https://github.com/meta-llama/llama-models/blob/main/models/llama3_2/MODEL_CARD.md).\n\nUsage of this model is subject to [Meta's Acceptable Use Policy](https://www.llama.com/llama3/use-policy/).", - "context_length": 131072, - "architecture": { - "modality": "text->text", - "input_modalities": [ - "text" - ], - "output_modalities": [ - "text" - ], - "tokenizer": "Llama3", - "instruct_type": "llama3" - }, - "pricing": { - "prompt": "0.000000005", - "completion": "0.00000001", - "request": "0", - "image": "0", - "web_search": "0", - "internal_reasoning": "0" - }, - "top_provider": { - "context_length": 131072, - "max_completion_tokens": 16384, - "is_moderated": false - }, - "per_request_limits": null, - "supported_parameters": [ "max_tokens", - "temperature", - "top_p", - "stop", - "frequency_penalty", + "min_p", "presence_penalty", "repetition_penalty", "response_format", - "top_k", "seed", - "min_p", - "logit_bias", - "top_logprobs" - ] - }, - { - "id": "meta-llama/llama-3.2-11b-vision-instruct:free", - "canonical_slug": "meta-llama/llama-3.2-11b-vision-instruct", - "hugging_face_id": "meta-llama/Llama-3.2-11B-Vision-Instruct", - "name": "Meta: Llama 3.2 11B Vision Instruct (free)", - "created": 1727222400, - "description": "Llama 3.2 11B Vision is a multimodal model with 11 billion parameters, designed to handle tasks combining visual and textual data. It excels in tasks such as image captioning and visual question answering, bridging the gap between language generation and visual reasoning. Pre-trained on a massive dataset of image-text pairs, it performs well in complex, high-accuracy image analysis.\n\nIts ability to integrate visual understanding with language processing makes it an ideal solution for industries requiring comprehensive visual-linguistic AI applications, such as content creation, AI-driven customer service, and research.\n\nClick here for the [original model card](https://github.com/meta-llama/llama-models/blob/main/models/llama3_2/MODEL_CARD_VISION.md).\n\nUsage of this model is subject to [Meta's Acceptable Use Policy](https://www.llama.com/llama3/use-policy/).", - "context_length": 131072, - "architecture": { - "modality": "text+image->text", - "input_modalities": [ - "text", - "image" - ], - "output_modalities": [ - "text" - ], - "tokenizer": "Llama3", - "instruct_type": "llama3" - }, - "pricing": { - "prompt": "0", - "completion": "0", - "request": "0", - "image": "0", - "web_search": "0", - "internal_reasoning": "0" - }, - "top_provider": { - "context_length": 131072, - "max_completion_tokens": 2048, - "is_moderated": false - }, - "per_request_limits": null, - "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", "stop", - "frequency_penalty", - "presence_penalty", - "top_k", - "repetition_penalty", - "logit_bias", - "min_p", - "response_format" - ] - }, - { - "id": "meta-llama/llama-3.2-11b-vision-instruct", - "canonical_slug": "meta-llama/llama-3.2-11b-vision-instruct", - "hugging_face_id": "meta-llama/Llama-3.2-11B-Vision-Instruct", - "name": "Meta: Llama 3.2 11B Vision Instruct", - "created": 1727222400, - "description": "Llama 3.2 11B Vision is a multimodal model with 11 billion parameters, designed to handle tasks combining visual and textual data. It excels in tasks such as image captioning and visual question answering, bridging the gap between language generation and visual reasoning. Pre-trained on a massive dataset of image-text pairs, it performs well in complex, high-accuracy image analysis.\n\nIts ability to integrate visual understanding with language processing makes it an ideal solution for industries requiring comprehensive visual-linguistic AI applications, such as content creation, AI-driven customer service, and research.\n\nClick here for the [original model card](https://github.com/meta-llama/llama-models/blob/main/models/llama3_2/MODEL_CARD_VISION.md).\n\nUsage of this model is subject to [Meta's Acceptable Use Policy](https://www.llama.com/llama3/use-policy/).", - "context_length": 131072, - "architecture": { - "modality": "text+image->text", - "input_modalities": [ - "text", - "image" - ], - "output_modalities": [ - "text" - ], - "tokenizer": "Llama3", - "instruct_type": "llama3" - }, - "pricing": { - "prompt": "0.000000049", - "completion": "0.000000049", - "request": "0", - "image": "0.00007948", - "web_search": "0", - "internal_reasoning": "0" - }, - "top_provider": { - "context_length": 131072, - "max_completion_tokens": 16384, - "is_moderated": false - }, - "per_request_limits": null, - "supported_parameters": [ - "max_tokens", + "structured_outputs", "temperature", - "top_p", - "top_k", - "seed", - "repetition_penalty", - "frequency_penalty", - "presence_penalty", - "stop", - "logit_bias", - "min_p", - "response_format", - "top_logprobs", - "tools", "tool_choice", - "logprobs" + "tools", + "top_k", + "top_logprobs", + "top_p" ] }, { @@ -11024,6 +11183,7 @@ "completion": "0.0000012", "request": "0", "image": "0.001734", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -11034,18 +11194,172 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "stop", "frequency_penalty", - "presence_penalty", - "top_k", - "repetition_penalty", "logit_bias", + "max_tokens", "min_p", + "presence_penalty", + "repetition_penalty", "response_format", - "seed" + "seed", + "stop", + "temperature", + "top_k", + "top_p" + ] + }, + { + "id": "meta-llama/llama-3.2-1b-instruct", + "canonical_slug": "meta-llama/llama-3.2-1b-instruct", + "hugging_face_id": "meta-llama/Llama-3.2-1B-Instruct", + "name": "Meta: Llama 3.2 1B Instruct", + "created": 1727222400, + "description": "Llama 3.2 1B is a 1-billion-parameter language model focused on efficiently performing natural language tasks, such as summarization, dialogue, and multilingual text analysis. Its smaller size allows it to operate efficiently in low-resource environments while maintaining strong task performance.\n\nSupporting eight core languages and fine-tunable for more, Llama 1.3B is ideal for businesses or developers seeking lightweight yet powerful AI solutions that can operate in diverse multilingual settings without the high computational demand of larger models.\n\nClick here for the [original model card](https://github.com/meta-llama/llama-models/blob/main/models/llama3_2/MODEL_CARD.md).\n\nUsage of this model is subject to [Meta's Acceptable Use Policy](https://www.llama.com/llama3/use-policy/).", + "context_length": 131072, + "architecture": { + "modality": "text->text", + "input_modalities": [ + "text" + ], + "output_modalities": [ + "text" + ], + "tokenizer": "Llama3", + "instruct_type": "llama3" + }, + "pricing": { + "prompt": "0.000000005", + "completion": "0.00000001", + "request": "0", + "image": "0", + "audio": "0", + "web_search": "0", + "internal_reasoning": "0" + }, + "top_provider": { + "context_length": 131072, + "max_completion_tokens": 16384, + "is_moderated": false + }, + "per_request_limits": null, + "supported_parameters": [ + "frequency_penalty", + "logit_bias", + "max_tokens", + "min_p", + "presence_penalty", + "repetition_penalty", + "response_format", + "seed", + "stop", + "structured_outputs", + "temperature", + "top_k", + "top_logprobs", + "top_p" + ] + }, + { + "id": "meta-llama/llama-3.2-11b-vision-instruct:free", + "canonical_slug": "meta-llama/llama-3.2-11b-vision-instruct", + "hugging_face_id": "meta-llama/Llama-3.2-11B-Vision-Instruct", + "name": "Meta: Llama 3.2 11B Vision Instruct (free)", + "created": 1727222400, + "description": "Llama 3.2 11B Vision is a multimodal model with 11 billion parameters, designed to handle tasks combining visual and textual data. It excels in tasks such as image captioning and visual question answering, bridging the gap between language generation and visual reasoning. Pre-trained on a massive dataset of image-text pairs, it performs well in complex, high-accuracy image analysis.\n\nIts ability to integrate visual understanding with language processing makes it an ideal solution for industries requiring comprehensive visual-linguistic AI applications, such as content creation, AI-driven customer service, and research.\n\nClick here for the [original model card](https://github.com/meta-llama/llama-models/blob/main/models/llama3_2/MODEL_CARD_VISION.md).\n\nUsage of this model is subject to [Meta's Acceptable Use Policy](https://www.llama.com/llama3/use-policy/).", + "context_length": 131072, + "architecture": { + "modality": "text+image->text", + "input_modalities": [ + "text", + "image" + ], + "output_modalities": [ + "text" + ], + "tokenizer": "Llama3", + "instruct_type": "llama3" + }, + "pricing": { + "prompt": "0", + "completion": "0", + "request": "0", + "image": "0", + "audio": "0", + "web_search": "0", + "internal_reasoning": "0" + }, + "top_provider": { + "context_length": 131072, + "max_completion_tokens": 2048, + "is_moderated": false + }, + "per_request_limits": null, + "supported_parameters": [ + "frequency_penalty", + "logit_bias", + "max_tokens", + "min_p", + "presence_penalty", + "repetition_penalty", + "stop", + "temperature", + "top_k", + "top_p" + ] + }, + { + "id": "meta-llama/llama-3.2-11b-vision-instruct", + "canonical_slug": "meta-llama/llama-3.2-11b-vision-instruct", + "hugging_face_id": "meta-llama/Llama-3.2-11B-Vision-Instruct", + "name": "Meta: Llama 3.2 11B Vision Instruct", + "created": 1727222400, + "description": "Llama 3.2 11B Vision is a multimodal model with 11 billion parameters, designed to handle tasks combining visual and textual data. It excels in tasks such as image captioning and visual question answering, bridging the gap between language generation and visual reasoning. Pre-trained on a massive dataset of image-text pairs, it performs well in complex, high-accuracy image analysis.\n\nIts ability to integrate visual understanding with language processing makes it an ideal solution for industries requiring comprehensive visual-linguistic AI applications, such as content creation, AI-driven customer service, and research.\n\nClick here for the [original model card](https://github.com/meta-llama/llama-models/blob/main/models/llama3_2/MODEL_CARD_VISION.md).\n\nUsage of this model is subject to [Meta's Acceptable Use Policy](https://www.llama.com/llama3/use-policy/).", + "context_length": 131072, + "architecture": { + "modality": "text+image->text", + "input_modalities": [ + "text", + "image" + ], + "output_modalities": [ + "text" + ], + "tokenizer": "Llama3", + "instruct_type": "llama3" + }, + "pricing": { + "prompt": "0.000000049", + "completion": "0.000000049", + "request": "0", + "image": "0.00007948", + "audio": "0", + "web_search": "0", + "internal_reasoning": "0" + }, + "top_provider": { + "context_length": 131072, + "max_completion_tokens": 16384, + "is_moderated": false + }, + "per_request_limits": null, + "supported_parameters": [ + "frequency_penalty", + "logit_bias", + "logprobs", + "max_tokens", + "min_p", + "presence_penalty", + "repetition_penalty", + "response_format", + "seed", + "stop", + "structured_outputs", + "temperature", + "tool_choice", + "tools", + "top_k", + "top_logprobs", + "top_p" ] }, { @@ -11072,6 +11386,7 @@ "completion": "0", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -11082,19 +11397,19 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "stop", "frequency_penalty", - "presence_penalty", - "seed", - "top_k", - "min_p", - "repetition_penalty", - "logprobs", "logit_bias", - "top_logprobs" + "logprobs", + "max_tokens", + "min_p", + "presence_penalty", + "repetition_penalty", + "seed", + "stop", + "temperature", + "top_k", + "top_logprobs", + "top_p" ] }, { @@ -11117,10 +11432,11 @@ "instruct_type": "chatml" }, "pricing": { - "prompt": "0.000000101", - "completion": "0.000000101", + "prompt": "0.0000000666396", + "completion": "0.000000266688", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -11131,23 +11447,23 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "stop", "frequency_penalty", - "presence_penalty", - "response_format", - "structured_outputs", - "seed", - "top_k", - "min_p", - "repetition_penalty", - "logprobs", "logit_bias", - "top_logprobs", + "logprobs", + "max_tokens", + "min_p", + "presence_penalty", + "repetition_penalty", + "response_format", + "seed", + "stop", + "structured_outputs", + "temperature", + "tool_choice", "tools", - "tool_choice" + "top_k", + "top_logprobs", + "top_p" ] }, { @@ -11157,7 +11473,7 @@ "name": "NeverSleep: Lumimaid v0.2 8B", "created": 1726358400, "description": "Lumimaid v0.2 8B is a finetune of [Llama 3.1 8B](/models/meta-llama/llama-3.1-8b-instruct) with a \"HUGE step up dataset wise\" compared to Lumimaid v0.1. Sloppy chats output were purged.\n\nUsage of this model is subject to [Meta's Acceptable Use Policy](https://llama.meta.com/llama3/use-policy/).", - "context_length": 32768, + "context_length": 8192, "architecture": { "modality": "text->text", "input_modalities": [ @@ -11170,73 +11486,35 @@ "instruct_type": "llama3" }, "pricing": { - "prompt": "0.00000018", - "completion": "0.000001", + "prompt": "0.0000001", + "completion": "0.0000009", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, "top_provider": { - "context_length": 32768, - "max_completion_tokens": null, + "context_length": 8192, + "max_completion_tokens": 8192, "is_moderated": false }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "stop", "frequency_penalty", - "presence_penalty", - "response_format", - "structured_outputs", - "repetition_penalty", - "top_k", - "min_p", - "seed", "logit_bias", - "top_a" - ] - }, - { - "id": "openai/o1-preview-2024-09-12", - "canonical_slug": "openai/o1-preview-2024-09-12", - "hugging_face_id": null, - "name": "OpenAI: o1-preview (2024-09-12)", - "created": 1726099200, - "description": "The latest and strongest model family from OpenAI, o1 is designed to spend more time thinking before responding.\n\nThe o1 models are optimized for math, science, programming, and other STEM-related tasks. They consistently exhibit PhD-level accuracy on benchmarks in physics, chemistry, and biology. Learn more in the [launch announcement](https://openai.com/o1).\n\nNote: This model is currently experimental and not suitable for production use-cases, and may be heavily rate-limited.", - "context_length": 128000, - "architecture": { - "modality": "text->text", - "input_modalities": [ - "text" - ], - "output_modalities": [ - "text" - ], - "tokenizer": "GPT", - "instruct_type": null - }, - "pricing": { - "prompt": "0.000015", - "completion": "0.00006", - "request": "0", - "image": "0", - "web_search": "0", - "internal_reasoning": "0", - "input_cache_read": "0.0000075" - }, - "top_provider": { - "context_length": 128000, - "max_completion_tokens": 32768, - "is_moderated": true - }, - "per_request_limits": null, - "supported_parameters": [ + "max_tokens", + "min_p", + "presence_penalty", + "repetition_penalty", + "response_format", "seed", - "max_tokens" + "stop", + "structured_outputs", + "temperature", + "top_a", + "top_k", + "top_p" ] }, { @@ -11263,6 +11541,7 @@ "completion": "0.0000044", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0", "input_cache_read": "0.00000055" @@ -11274,47 +11553,8 @@ }, "per_request_limits": null, "supported_parameters": [ - "seed", - "max_tokens" - ] - }, - { - "id": "openai/o1-preview", - "canonical_slug": "openai/o1-preview", - "hugging_face_id": null, - "name": "OpenAI: o1-preview", - "created": 1726099200, - "description": "The latest and strongest model family from OpenAI, o1 is designed to spend more time thinking before responding.\n\nThe o1 models are optimized for math, science, programming, and other STEM-related tasks. They consistently exhibit PhD-level accuracy on benchmarks in physics, chemistry, and biology. Learn more in the [launch announcement](https://openai.com/o1).\n\nNote: This model is currently experimental and not suitable for production use-cases, and may be heavily rate-limited.", - "context_length": 128000, - "architecture": { - "modality": "text->text", - "input_modalities": [ - "text" - ], - "output_modalities": [ - "text" - ], - "tokenizer": "GPT", - "instruct_type": null - }, - "pricing": { - "prompt": "0.000015", - "completion": "0.00006", - "request": "0", - "image": "0", - "web_search": "0", - "internal_reasoning": "0", - "input_cache_read": "0.0000075" - }, - "top_provider": { - "context_length": 128000, - "max_completion_tokens": 32768, - "is_moderated": true - }, - "per_request_limits": null, - "supported_parameters": [ - "seed", - "max_tokens" + "max_tokens", + "seed" ] }, { @@ -11341,6 +11581,7 @@ "completion": "0.0000044", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0", "input_cache_read": "0.00000055" @@ -11352,8 +11593,8 @@ }, "per_request_limits": null, "supported_parameters": [ - "seed", - "max_tokens" + "max_tokens", + "seed" ] }, { @@ -11381,6 +11622,7 @@ "completion": "0.0000001", "request": "0", "image": "0.0001445", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -11391,23 +11633,23 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "stop", "frequency_penalty", - "presence_penalty", - "logprobs", - "top_logprobs", - "seed", "logit_bias", - "top_k", + "logprobs", + "max_tokens", "min_p", + "presence_penalty", "repetition_penalty", - "tools", - "tool_choice", "response_format", - "structured_outputs" + "seed", + "stop", + "structured_outputs", + "temperature", + "tool_choice", + "tools", + "top_k", + "top_logprobs", + "top_p" ] }, { @@ -11434,6 +11676,7 @@ "completion": "0.0000006", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -11444,17 +11687,17 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "tools", - "stop", "frequency_penalty", + "max_tokens", "presence_penalty", - "top_k", - "seed", "response_format", - "structured_outputs" + "seed", + "stop", + "structured_outputs", + "temperature", + "tools", + "top_k", + "top_p" ] }, { @@ -11481,6 +11724,7 @@ "completion": "0.00001", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -11491,17 +11735,17 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "tools", - "stop", "frequency_penalty", + "max_tokens", "presence_penalty", - "top_k", - "seed", "response_format", - "structured_outputs" + "seed", + "stop", + "structured_outputs", + "temperature", + "tools", + "top_k", + "top_p" ] }, { @@ -11529,6 +11773,7 @@ "completion": "0.0000002", "request": "0", "image": "0.0001445", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -11539,19 +11784,21 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "stop", "frequency_penalty", - "presence_penalty", - "logprobs", - "top_logprobs", - "seed", "logit_bias", - "top_k", + "logprobs", + "max_tokens", "min_p", - "repetition_penalty" + "presence_penalty", + "repetition_penalty", + "response_format", + "seed", + "stop", + "structured_outputs", + "temperature", + "top_k", + "top_logprobs", + "top_p" ] }, { @@ -11578,6 +11825,7 @@ "completion": "0.00000075", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -11588,19 +11836,19 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "stop", "frequency_penalty", - "presence_penalty", - "response_format", - "structured_outputs", - "seed", - "top_k", + "logit_bias", + "max_tokens", "min_p", + "presence_penalty", "repetition_penalty", - "logit_bias" + "response_format", + "seed", + "stop", + "structured_outputs", + "temperature", + "top_k", + "top_p" ] }, { @@ -11627,6 +11875,7 @@ "completion": "0.0000001", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -11639,9 +11888,9 @@ "supported_parameters": [ "max_tokens", "temperature", - "top_p", + "tool_choice", "tools", - "tool_choice" + "top_p" ] }, { @@ -11668,6 +11917,7 @@ "completion": "0.00000028", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -11678,23 +11928,23 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "stop", "frequency_penalty", + "logit_bias", + "logprobs", + "max_tokens", + "min_p", "presence_penalty", "repetition_penalty", "response_format", - "top_k", "seed", - "min_p", + "stop", "structured_outputs", - "logit_bias", - "logprobs", - "top_logprobs", + "temperature", + "tool_choice", "tools", - "tool_choice" + "top_k", + "top_logprobs", + "top_p" ] }, { @@ -11721,6 +11971,7 @@ "completion": "0.0000008", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -11731,20 +11982,20 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "stop", "frequency_penalty", - "presence_penalty", - "seed", "logit_bias", "logprobs", - "top_logprobs", - "response_format", + "max_tokens", "min_p", + "presence_penalty", "repetition_penalty", - "top_k" + "response_format", + "seed", + "stop", + "temperature", + "top_k", + "top_logprobs", + "top_p" ] }, { @@ -11772,6 +12023,7 @@ "completion": "0.000015", "request": "0", "image": "0.007225", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -11782,18 +12034,18 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "stop", "frequency_penalty", - "presence_penalty", - "seed", "logit_bias", "logprobs", - "top_logprobs", + "max_tokens", + "presence_penalty", "response_format", - "structured_outputs" + "seed", + "stop", + "structured_outputs", + "temperature", + "top_logprobs", + "top_p" ] }, { @@ -11820,6 +12072,7 @@ "completion": "0.00000005", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -11830,18 +12083,18 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "stop", "frequency_penalty", - "presence_penalty", - "seed", - "top_k", - "min_p", - "repetition_penalty", "logit_bias", - "response_format" + "max_tokens", + "min_p", + "presence_penalty", + "repetition_penalty", + "response_format", + "seed", + "stop", + "temperature", + "top_k", + "top_p" ] }, { @@ -11870,6 +12123,7 @@ "completion": "0.00001", "request": "0", "image": "0.003613", + "audio": "0", "web_search": "0", "internal_reasoning": "0", "input_cache_read": "0.00000125" @@ -11881,21 +12135,21 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "tools", - "tool_choice", - "stop", "frequency_penalty", - "presence_penalty", - "web_search_options", - "seed", "logit_bias", "logprobs", - "top_logprobs", + "max_tokens", + "presence_penalty", "response_format", - "structured_outputs" + "seed", + "stop", + "structured_outputs", + "temperature", + "tool_choice", + "tools", + "top_logprobs", + "top_p", + "web_search_options" ] }, { @@ -11922,6 +12176,7 @@ "completion": "0.000002", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -11932,65 +12187,19 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "stop", "frequency_penalty", - "presence_penalty", - "logprobs", - "top_logprobs", - "seed", "logit_bias", - "top_k", - "min_p", - "repetition_penalty" - ] - }, - { - "id": "nothingiisreal/mn-celeste-12b", - "canonical_slug": "nothingiisreal/mn-celeste-12b", - "hugging_face_id": "nothingiisreal/MN-12B-Celeste-V1.9", - "name": "Mistral Nemo 12B Celeste", - "created": 1722556800, - "description": "A specialized story writing and roleplaying model based on Mistral's NeMo 12B Instruct. Fine-tuned on curated datasets including Reddit Writing Prompts and Opus Instruct 25K.\n\nThis model excels at creative writing, offering improved NSFW capabilities, with smarter and more active narration. It demonstrates remarkable versatility in both SFW and NSFW scenarios, with strong Out of Character (OOC) steering capabilities, allowing fine-tuned control over narrative direction and character behavior.\n\nCheck out the model's [HuggingFace page](https://huggingface.co/nothingiisreal/MN-12B-Celeste-V1.9) for details on what parameters and prompts work best!", - "context_length": 16384, - "architecture": { - "modality": "text->text", - "input_modalities": [ - "text" - ], - "output_modalities": [ - "text" - ], - "tokenizer": "Mistral", - "instruct_type": "chatml" - }, - "pricing": { - "prompt": "0.0000008", - "completion": "0.0000012", - "request": "0", - "image": "0", - "web_search": "0", - "internal_reasoning": "0" - }, - "top_provider": { - "context_length": 16384, - "max_completion_tokens": 4096, - "is_moderated": false - }, - "per_request_limits": null, - "supported_parameters": [ + "logprobs", "max_tokens", - "temperature", - "top_p", - "stop", - "frequency_penalty", + "min_p", "presence_penalty", "repetition_penalty", + "seed", + "stop", + "temperature", "top_k", - "min_p", - "seed" + "top_logprobs", + "top_p" ] }, { @@ -12017,6 +12226,7 @@ "completion": "0.00000028", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -12027,76 +12237,23 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "tools", - "tool_choice", - "response_format", - "stop", "frequency_penalty", - "presence_penalty", - "repetition_penalty", - "top_k", - "seed", - "min_p", "logit_bias", "logprobs", - "top_logprobs", - "structured_outputs" - ] - }, - { - "id": "meta-llama/llama-3.1-8b-instruct", - "canonical_slug": "meta-llama/llama-3.1-8b-instruct", - "hugging_face_id": "meta-llama/Meta-Llama-3.1-8B-Instruct", - "name": "Meta: Llama 3.1 8B Instruct", - "created": 1721692800, - "description": "Meta's latest class of model (Llama 3.1) launched with a variety of sizes & flavors. This 8B instruct-tuned version is fast and efficient.\n\nIt has demonstrated strong performance compared to leading closed-source models in human evaluations.\n\nTo read more about the model release, [click here](https://ai.meta.com/blog/meta-llama-3-1/). Usage of this model is subject to [Meta's Acceptable Use Policy](https://llama.meta.com/llama3/use-policy/).", - "context_length": 131072, - "architecture": { - "modality": "text->text", - "input_modalities": [ - "text" - ], - "output_modalities": [ - "text" - ], - "tokenizer": "Llama3", - "instruct_type": "llama3" - }, - "pricing": { - "prompt": "0.000000015", - "completion": "0.00000002", - "request": "0", - "image": "0", - "web_search": "0", - "internal_reasoning": "0" - }, - "top_provider": { - "context_length": 131072, - "max_completion_tokens": 16384, - "is_moderated": false - }, - "per_request_limits": null, - "supported_parameters": [ "max_tokens", - "temperature", - "top_p", - "tools", - "tool_choice", - "stop", - "frequency_penalty", - "presence_penalty", - "seed", - "logit_bias", - "logprobs", - "top_logprobs", - "response_format", "min_p", + "presence_penalty", "repetition_penalty", + "response_format", + "seed", + "stop", + "structured_outputs", + "temperature", + "tool_choice", + "tools", "top_k", - "structured_outputs" + "top_logprobs", + "top_p" ] }, { @@ -12123,6 +12280,7 @@ "completion": "0", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -12133,15 +12291,15 @@ }, "per_request_limits": null, "supported_parameters": [ + "frequency_penalty", "max_tokens", - "temperature", - "top_p", - "structured_outputs", + "presence_penalty", "response_format", "stop", - "frequency_penalty", - "presence_penalty", - "top_k" + "structured_outputs", + "temperature", + "top_k", + "top_p" ] }, { @@ -12168,6 +12326,7 @@ "completion": "0.0000008", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -12178,23 +12337,77 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "tools", - "tool_choice", - "stop", "frequency_penalty", - "presence_penalty", - "top_k", - "repetition_penalty", - "response_format", - "structured_outputs", "logit_bias", "logprobs", - "top_logprobs", + "max_tokens", "min_p", - "seed" + "presence_penalty", + "repetition_penalty", + "response_format", + "seed", + "stop", + "structured_outputs", + "temperature", + "tool_choice", + "tools", + "top_k", + "top_logprobs", + "top_p" + ] + }, + { + "id": "meta-llama/llama-3.1-8b-instruct", + "canonical_slug": "meta-llama/llama-3.1-8b-instruct", + "hugging_face_id": "meta-llama/Meta-Llama-3.1-8B-Instruct", + "name": "Meta: Llama 3.1 8B Instruct", + "created": 1721692800, + "description": "Meta's latest class of model (Llama 3.1) launched with a variety of sizes & flavors. This 8B instruct-tuned version is fast and efficient.\n\nIt has demonstrated strong performance compared to leading closed-source models in human evaluations.\n\nTo read more about the model release, [click here](https://ai.meta.com/blog/meta-llama-3-1/). Usage of this model is subject to [Meta's Acceptable Use Policy](https://llama.meta.com/llama3/use-policy/).", + "context_length": 131072, + "architecture": { + "modality": "text->text", + "input_modalities": [ + "text" + ], + "output_modalities": [ + "text" + ], + "tokenizer": "Llama3", + "instruct_type": "llama3" + }, + "pricing": { + "prompt": "0.000000015", + "completion": "0.00000002", + "request": "0", + "image": "0", + "audio": "0", + "web_search": "0", + "internal_reasoning": "0" + }, + "top_provider": { + "context_length": 131072, + "max_completion_tokens": 16384, + "is_moderated": false + }, + "per_request_limits": null, + "supported_parameters": [ + "frequency_penalty", + "logit_bias", + "logprobs", + "max_tokens", + "min_p", + "presence_penalty", + "repetition_penalty", + "response_format", + "seed", + "stop", + "structured_outputs", + "temperature", + "tool_choice", + "tools", + "top_k", + "top_logprobs", + "top_p" ] }, { @@ -12221,6 +12434,7 @@ "completion": "0", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -12231,19 +12445,19 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "stop", "frequency_penalty", - "presence_penalty", - "seed", - "top_k", - "min_p", - "repetition_penalty", - "logprobs", "logit_bias", - "top_logprobs" + "logprobs", + "max_tokens", + "min_p", + "presence_penalty", + "repetition_penalty", + "seed", + "stop", + "temperature", + "top_k", + "top_logprobs", + "top_p" ] }, { @@ -12270,6 +12484,7 @@ "completion": "0.00000005", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -12280,77 +12495,23 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "stop", "frequency_penalty", + "logit_bias", + "logprobs", + "max_tokens", + "min_p", "presence_penalty", "repetition_penalty", "response_format", - "top_k", "seed", - "min_p", - "logprobs", - "logit_bias", - "top_logprobs", - "tools", - "tool_choice", - "structured_outputs" - ] - }, - { - "id": "openai/gpt-4o-mini", - "canonical_slug": "openai/gpt-4o-mini", - "hugging_face_id": null, - "name": "OpenAI: GPT-4o-mini", - "created": 1721260800, - "description": "GPT-4o mini is OpenAI's newest model after [GPT-4 Omni](/models/openai/gpt-4o), supporting both text and image inputs with text outputs.\n\nAs their most advanced small model, it is many multiples more affordable than other recent frontier models, and more than 60% cheaper than [GPT-3.5 Turbo](/models/openai/gpt-3.5-turbo). It maintains SOTA intelligence, while being significantly more cost-effective.\n\nGPT-4o mini achieves an 82% score on MMLU and presently ranks higher than GPT-4 on chat preferences [common leaderboards](https://arena.lmsys.org/).\n\nCheck out the [launch announcement](https://openai.com/index/gpt-4o-mini-advancing-cost-efficient-intelligence/) to learn more.\n\n#multimodal", - "context_length": 128000, - "architecture": { - "modality": "text+image->text", - "input_modalities": [ - "text", - "image", - "file" - ], - "output_modalities": [ - "text" - ], - "tokenizer": "GPT", - "instruct_type": null - }, - "pricing": { - "prompt": "0.00000015", - "completion": "0.0000006", - "request": "0", - "image": "0.000217", - "web_search": "0", - "internal_reasoning": "0", - "input_cache_read": "0.000000075" - }, - "top_provider": { - "context_length": 128000, - "max_completion_tokens": 16384, - "is_moderated": true - }, - "per_request_limits": null, - "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", "stop", - "frequency_penalty", - "presence_penalty", - "web_search_options", - "seed", - "logit_bias", - "logprobs", - "top_logprobs", - "response_format", "structured_outputs", + "temperature", + "tool_choice", "tools", - "tool_choice" + "top_k", + "top_logprobs", + "top_p" ] }, { @@ -12379,6 +12540,7 @@ "completion": "0.0000006", "request": "0", "image": "0.007225", + "audio": "0", "web_search": "0", "internal_reasoning": "0", "input_cache_read": "0.000000075" @@ -12390,21 +12552,76 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "tools", - "tool_choice", - "stop", "frequency_penalty", - "presence_penalty", - "web_search_options", - "seed", "logit_bias", "logprobs", - "top_logprobs", + "max_tokens", + "presence_penalty", "response_format", - "structured_outputs" + "seed", + "stop", + "structured_outputs", + "temperature", + "tool_choice", + "tools", + "top_logprobs", + "top_p", + "web_search_options" + ] + }, + { + "id": "openai/gpt-4o-mini", + "canonical_slug": "openai/gpt-4o-mini", + "hugging_face_id": null, + "name": "OpenAI: GPT-4o-mini", + "created": 1721260800, + "description": "GPT-4o mini is OpenAI's newest model after [GPT-4 Omni](/models/openai/gpt-4o), supporting both text and image inputs with text outputs.\n\nAs their most advanced small model, it is many multiples more affordable than other recent frontier models, and more than 60% cheaper than [GPT-3.5 Turbo](/models/openai/gpt-3.5-turbo). It maintains SOTA intelligence, while being significantly more cost-effective.\n\nGPT-4o mini achieves an 82% score on MMLU and presently ranks higher than GPT-4 on chat preferences [common leaderboards](https://arena.lmsys.org/).\n\nCheck out the [launch announcement](https://openai.com/index/gpt-4o-mini-advancing-cost-efficient-intelligence/) to learn more.\n\n#multimodal", + "context_length": 128000, + "architecture": { + "modality": "text+image->text", + "input_modalities": [ + "text", + "image", + "file" + ], + "output_modalities": [ + "text" + ], + "tokenizer": "GPT", + "instruct_type": null + }, + "pricing": { + "prompt": "0.00000015", + "completion": "0.0000006", + "request": "0", + "image": "0.000217", + "audio": "0", + "web_search": "0", + "internal_reasoning": "0", + "input_cache_read": "0.000000075" + }, + "top_provider": { + "context_length": 128000, + "max_completion_tokens": 16384, + "is_moderated": true + }, + "per_request_limits": null, + "supported_parameters": [ + "frequency_penalty", + "logit_bias", + "logprobs", + "max_tokens", + "presence_penalty", + "response_format", + "seed", + "stop", + "structured_outputs", + "temperature", + "tool_choice", + "tools", + "top_logprobs", + "top_p", + "web_search_options" ] }, { @@ -12431,6 +12648,7 @@ "completion": "0.00000065", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -12441,18 +12659,18 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "stop", "frequency_penalty", - "presence_penalty", - "response_format", - "structured_outputs", - "top_k", - "repetition_penalty", "logit_bias", - "min_p" + "max_tokens", + "min_p", + "presence_penalty", + "repetition_penalty", + "response_format", + "stop", + "structured_outputs", + "temperature", + "top_k", + "top_p" ] }, { @@ -12479,6 +12697,7 @@ "completion": "0", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -12489,19 +12708,19 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "stop", "frequency_penalty", - "presence_penalty", - "seed", - "top_k", - "min_p", - "repetition_penalty", - "logprobs", "logit_bias", - "top_logprobs" + "logprobs", + "max_tokens", + "min_p", + "presence_penalty", + "repetition_penalty", + "seed", + "stop", + "temperature", + "top_k", + "top_logprobs", + "top_p" ] }, { @@ -12524,10 +12743,11 @@ "instruct_type": "gemma" }, "pricing": { - "prompt": "0.000000004", - "completion": "0.000000004", + "prompt": "0.00000001", + "completion": "0.0000000100008", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -12538,69 +12758,20 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "stop", "frequency_penalty", - "presence_penalty", - "seed", - "top_k", - "min_p", - "repetition_penalty", - "logprobs", "logit_bias", - "top_logprobs", - "response_format" - ] - }, - { - "id": "01-ai/yi-large", - "canonical_slug": "01-ai/yi-large", - "hugging_face_id": null, - "name": "01.AI: Yi Large", - "created": 1719273600, - "description": "The Yi Large model was designed by 01.AI with the following usecases in mind: knowledge search, data classification, human-like chat bots, and customer service.\n\nIt stands out for its multilingual proficiency, particularly in Spanish, Chinese, Japanese, German, and French.\n\nCheck out the [launch announcement](https://01-ai.github.io/blog/01.ai-yi-large-llm-launch) to learn more.", - "context_length": 32768, - "architecture": { - "modality": "text->text", - "input_modalities": [ - "text" - ], - "output_modalities": [ - "text" - ], - "tokenizer": "Yi", - "instruct_type": null - }, - "pricing": { - "prompt": "0.000003", - "completion": "0.000003", - "request": "0", - "image": "0", - "web_search": "0", - "internal_reasoning": "0" - }, - "top_provider": { - "context_length": 32768, - "max_completion_tokens": 4096, - "is_moderated": false - }, - "per_request_limits": null, - "supported_parameters": [ + "logprobs", "max_tokens", - "temperature", - "top_p", - "stop", - "frequency_penalty", + "min_p", "presence_penalty", - "top_k", "repetition_penalty", "response_format", - "structured_outputs", - "logit_bias", - "logprobs", - "top_logprobs" + "seed", + "stop", + "temperature", + "top_k", + "top_logprobs", + "top_p" ] }, { @@ -12628,6 +12799,7 @@ "completion": "0.000015", "request": "0", "image": "0.0048", + "audio": "0", "web_search": "0", "internal_reasoning": "0", "input_cache_read": "0.0000003", @@ -12641,12 +12813,12 @@ "per_request_limits": null, "supported_parameters": [ "max_tokens", + "stop", "temperature", - "top_p", - "tools", "tool_choice", + "tools", "top_k", - "stop" + "top_p" ] }, { @@ -12674,6 +12846,7 @@ "completion": "0.000015", "request": "0", "image": "0.0048", + "audio": "0", "web_search": "0", "internal_reasoning": "0", "input_cache_read": "0.0000003", @@ -12687,12 +12860,12 @@ "per_request_limits": null, "supported_parameters": [ "max_tokens", + "stop", "temperature", - "top_p", - "tools", "tool_choice", + "tools", "top_k", - "stop" + "top_p" ] }, { @@ -12719,6 +12892,7 @@ "completion": "0.00000148", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -12729,17 +12903,17 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "stop", "frequency_penalty", - "presence_penalty", - "seed", - "top_k", + "logit_bias", + "max_tokens", "min_p", + "presence_penalty", "repetition_penalty", - "logit_bias" + "seed", + "stop", + "temperature", + "top_k", + "top_p" ] }, { @@ -12766,6 +12940,7 @@ "completion": "0.0000009", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -12776,17 +12951,17 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "stop", "frequency_penalty", - "presence_penalty", - "seed", - "top_k", + "logit_bias", + "max_tokens", "min_p", + "presence_penalty", "repetition_penalty", - "logit_bias" + "seed", + "stop", + "temperature", + "top_k", + "top_p" ] }, { @@ -12813,6 +12988,7 @@ "completion": "0.0000009", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -12823,17 +12999,16 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "stop", "frequency_penalty", - "presence_penalty", - "top_k", - "repetition_penalty", "logit_bias", + "max_tokens", "min_p", - "response_format" + "presence_penalty", + "repetition_penalty", + "stop", + "temperature", + "top_k", + "top_p" ] }, { @@ -12860,6 +13035,7 @@ "completion": "0.000000054", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -12870,21 +13046,21 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "stop", "frequency_penalty", - "presence_penalty", - "top_k", - "repetition_penalty", "logit_bias", + "logprobs", + "max_tokens", "min_p", + "presence_penalty", + "repetition_penalty", "response_format", "seed", - "tools", + "stop", + "temperature", "tool_choice", - "logprobs" + "tools", + "top_k", + "top_p" ] }, { @@ -12911,6 +13087,7 @@ "completion": "0.00000004", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -12921,20 +13098,21 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "stop", "frequency_penalty", - "presence_penalty", - "seed", "logit_bias", "logprobs", - "top_logprobs", - "response_format", + "max_tokens", "min_p", + "presence_penalty", "repetition_penalty", - "top_k" + "response_format", + "seed", + "stop", + "structured_outputs", + "temperature", + "top_k", + "top_logprobs", + "top_p" ] }, { @@ -12961,6 +13139,7 @@ "completion": "0", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -12971,19 +13150,19 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "tools", - "tool_choice", - "stop", "frequency_penalty", + "max_tokens", + "min_p", "presence_penalty", "repetition_penalty", "response_format", - "top_k", "seed", - "min_p" + "stop", + "temperature", + "tool_choice", + "tools", + "top_k", + "top_p" ] }, { @@ -13010,6 +13189,7 @@ "completion": "0.000000054", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -13020,21 +13200,21 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "stop", "frequency_penalty", - "presence_penalty", - "top_k", - "repetition_penalty", "logit_bias", + "logprobs", + "max_tokens", "min_p", + "presence_penalty", + "repetition_penalty", "response_format", "seed", - "logprobs", + "stop", + "temperature", + "tool_choice", "tools", - "tool_choice" + "top_k", + "top_p" ] }, { @@ -13061,6 +13241,7 @@ "completion": "0.0000001", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -13073,9 +13254,9 @@ "supported_parameters": [ "max_tokens", "temperature", - "top_p", + "tool_choice", "tools", - "tool_choice" + "top_p" ] }, { @@ -13102,6 +13283,7 @@ "completion": "0.000001", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -13114,9 +13296,9 @@ "supported_parameters": [ "max_tokens", "temperature", - "top_p", + "tool_choice", "tools", - "tool_choice" + "top_p" ] }, { @@ -13143,6 +13325,7 @@ "completion": "0.000006", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -13153,16 +13336,16 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "stop", "frequency_penalty", + "max_tokens", + "min_p", "presence_penalty", "repetition_penalty", + "seed", + "stop", + "temperature", "top_k", - "min_p", - "seed" + "top_p" ] }, { @@ -13190,6 +13373,7 @@ "completion": "0.0000003", "request": "0", "image": "0.00004", + "audio": "0", "web_search": "0", "internal_reasoning": "0", "input_cache_read": "0.00000001875", @@ -13202,17 +13386,17 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "tools", - "tool_choice", - "stop", "frequency_penalty", + "max_tokens", "presence_penalty", - "seed", "response_format", - "structured_outputs" + "seed", + "stop", + "structured_outputs", + "temperature", + "tool_choice", + "tools", + "top_p" ] }, { @@ -13239,6 +13423,7 @@ "completion": "0.0000002", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -13249,124 +13434,16 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "stop", "frequency_penalty", - "presence_penalty", - "top_k", - "repetition_penalty", "logit_bias", + "max_tokens", "min_p", - "response_format" - ] - }, - { - "id": "openai/gpt-4o", - "canonical_slug": "openai/gpt-4o", - "hugging_face_id": null, - "name": "OpenAI: GPT-4o", - "created": 1715558400, - "description": "GPT-4o (\"o\" for \"omni\") is OpenAI's latest AI model, supporting both text and image inputs with text outputs. It maintains the intelligence level of [GPT-4 Turbo](/models/openai/gpt-4-turbo) while being twice as fast and 50% more cost-effective. GPT-4o also offers improved performance in processing non-English languages and enhanced visual capabilities.\n\nFor benchmarking against other models, it was briefly called [\"im-also-a-good-gpt2-chatbot\"](https://twitter.com/LiamFedus/status/1790064963966370209)\n\n#multimodal", - "context_length": 128000, - "architecture": { - "modality": "text+image->text", - "input_modalities": [ - "text", - "image", - "file" - ], - "output_modalities": [ - "text" - ], - "tokenizer": "GPT", - "instruct_type": null - }, - "pricing": { - "prompt": "0.0000025", - "completion": "0.00001", - "request": "0", - "image": "0.003613", - "web_search": "0", - "internal_reasoning": "0", - "input_cache_read": "0.00000125" - }, - "top_provider": { - "context_length": 128000, - "max_completion_tokens": 16384, - "is_moderated": true - }, - "per_request_limits": null, - "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "tools", - "tool_choice", - "stop", - "frequency_penalty", "presence_penalty", - "web_search_options", - "seed", - "logit_bias", - "logprobs", - "top_logprobs", - "response_format", - "structured_outputs" - ] - }, - { - "id": "openai/gpt-4o:extended", - "canonical_slug": "openai/gpt-4o", - "hugging_face_id": null, - "name": "OpenAI: GPT-4o (extended)", - "created": 1715558400, - "description": "GPT-4o (\"o\" for \"omni\") is OpenAI's latest AI model, supporting both text and image inputs with text outputs. It maintains the intelligence level of [GPT-4 Turbo](/models/openai/gpt-4-turbo) while being twice as fast and 50% more cost-effective. GPT-4o also offers improved performance in processing non-English languages and enhanced visual capabilities.\n\nFor benchmarking against other models, it was briefly called [\"im-also-a-good-gpt2-chatbot\"](https://twitter.com/LiamFedus/status/1790064963966370209)\n\n#multimodal", - "context_length": 128000, - "architecture": { - "modality": "text+image->text", - "input_modalities": [ - "text", - "image", - "file" - ], - "output_modalities": [ - "text" - ], - "tokenizer": "GPT", - "instruct_type": null - }, - "pricing": { - "prompt": "0.000006", - "completion": "0.000018", - "request": "0", - "image": "0.007225", - "web_search": "0", - "internal_reasoning": "0" - }, - "top_provider": { - "context_length": 128000, - "max_completion_tokens": 64000, - "is_moderated": true - }, - "per_request_limits": null, - "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "tools", - "tool_choice", + "repetition_penalty", "stop", - "frequency_penalty", - "presence_penalty", - "web_search_options", - "seed", - "logit_bias", - "logprobs", - "top_logprobs", - "response_format", - "structured_outputs" + "temperature", + "top_k", + "top_p" ] }, { @@ -13395,6 +13472,7 @@ "completion": "0.000015", "request": "0", "image": "0.007225", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -13405,21 +13483,130 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "tools", - "tool_choice", - "stop", "frequency_penalty", - "presence_penalty", - "web_search_options", - "seed", "logit_bias", "logprobs", - "top_logprobs", + "max_tokens", + "presence_penalty", "response_format", - "structured_outputs" + "seed", + "stop", + "structured_outputs", + "temperature", + "tool_choice", + "tools", + "top_logprobs", + "top_p", + "web_search_options" + ] + }, + { + "id": "openai/gpt-4o", + "canonical_slug": "openai/gpt-4o", + "hugging_face_id": null, + "name": "OpenAI: GPT-4o", + "created": 1715558400, + "description": "GPT-4o (\"o\" for \"omni\") is OpenAI's latest AI model, supporting both text and image inputs with text outputs. It maintains the intelligence level of [GPT-4 Turbo](/models/openai/gpt-4-turbo) while being twice as fast and 50% more cost-effective. GPT-4o also offers improved performance in processing non-English languages and enhanced visual capabilities.\n\nFor benchmarking against other models, it was briefly called [\"im-also-a-good-gpt2-chatbot\"](https://twitter.com/LiamFedus/status/1790064963966370209)\n\n#multimodal", + "context_length": 128000, + "architecture": { + "modality": "text+image->text", + "input_modalities": [ + "text", + "image", + "file" + ], + "output_modalities": [ + "text" + ], + "tokenizer": "GPT", + "instruct_type": null + }, + "pricing": { + "prompt": "0.0000025", + "completion": "0.00001", + "request": "0", + "image": "0.003613", + "audio": "0", + "web_search": "0", + "internal_reasoning": "0", + "input_cache_read": "0.00000125" + }, + "top_provider": { + "context_length": 128000, + "max_completion_tokens": 16384, + "is_moderated": true + }, + "per_request_limits": null, + "supported_parameters": [ + "frequency_penalty", + "logit_bias", + "logprobs", + "max_tokens", + "presence_penalty", + "response_format", + "seed", + "stop", + "structured_outputs", + "temperature", + "tool_choice", + "tools", + "top_logprobs", + "top_p", + "web_search_options" + ] + }, + { + "id": "openai/gpt-4o:extended", + "canonical_slug": "openai/gpt-4o", + "hugging_face_id": null, + "name": "OpenAI: GPT-4o (extended)", + "created": 1715558400, + "description": "GPT-4o (\"o\" for \"omni\") is OpenAI's latest AI model, supporting both text and image inputs with text outputs. It maintains the intelligence level of [GPT-4 Turbo](/models/openai/gpt-4-turbo) while being twice as fast and 50% more cost-effective. GPT-4o also offers improved performance in processing non-English languages and enhanced visual capabilities.\n\nFor benchmarking against other models, it was briefly called [\"im-also-a-good-gpt2-chatbot\"](https://twitter.com/LiamFedus/status/1790064963966370209)\n\n#multimodal", + "context_length": 128000, + "architecture": { + "modality": "text+image->text", + "input_modalities": [ + "text", + "image", + "file" + ], + "output_modalities": [ + "text" + ], + "tokenizer": "GPT", + "instruct_type": null + }, + "pricing": { + "prompt": "0.000006", + "completion": "0.000018", + "request": "0", + "image": "0.007225", + "audio": "0", + "web_search": "0", + "internal_reasoning": "0" + }, + "top_provider": { + "context_length": 128000, + "max_completion_tokens": 64000, + "is_moderated": true + }, + "per_request_limits": null, + "supported_parameters": [ + "frequency_penalty", + "logit_bias", + "logprobs", + "max_tokens", + "presence_penalty", + "response_format", + "seed", + "stop", + "structured_outputs", + "temperature", + "tool_choice", + "tools", + "top_logprobs", + "top_p", + "web_search_options" ] }, { @@ -13446,6 +13633,7 @@ "completion": "0.0000012", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -13456,68 +13644,16 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "stop", "frequency_penalty", + "max_tokens", + "min_p", "presence_penalty", "repetition_penalty", - "top_k", - "min_p", - "seed" - ] - }, - { - "id": "meta-llama/llama-3-70b-instruct", - "canonical_slug": "meta-llama/llama-3-70b-instruct", - "hugging_face_id": "meta-llama/Meta-Llama-3-70B-Instruct", - "name": "Meta: Llama 3 70B Instruct", - "created": 1713398400, - "description": "Meta's latest class of model (Llama 3) launched with a variety of sizes & flavors. This 70B instruct-tuned version was optimized for high quality dialogue usecases.\n\nIt has demonstrated strong performance compared to leading closed-source models in human evaluations.\n\nTo read more about the model release, [click here](https://ai.meta.com/blog/meta-llama-3/). Usage of this model is subject to [Meta's Acceptable Use Policy](https://llama.meta.com/llama3/use-policy/).", - "context_length": 8192, - "architecture": { - "modality": "text->text", - "input_modalities": [ - "text" - ], - "output_modalities": [ - "text" - ], - "tokenizer": "Llama3", - "instruct_type": "llama3" - }, - "pricing": { - "prompt": "0.0000003", - "completion": "0.0000004", - "request": "0", - "image": "0", - "web_search": "0", - "internal_reasoning": "0" - }, - "top_provider": { - "context_length": 8192, - "max_completion_tokens": 16384, - "is_moderated": false - }, - "per_request_limits": null, - "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "stop", - "frequency_penalty", - "presence_penalty", - "top_k", - "repetition_penalty", - "logit_bias", - "min_p", - "response_format", - "top_logprobs", - "logprobs", "seed", - "tools", - "tool_choice" + "stop", + "temperature", + "top_k", + "top_p" ] }, { @@ -13544,6 +13680,7 @@ "completion": "0.00000006", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -13554,22 +13691,75 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "top_k", - "seed", - "repetition_penalty", "frequency_penalty", - "presence_penalty", - "stop", - "min_p", "logit_bias", - "tools", - "tool_choice", + "logprobs", + "max_tokens", + "min_p", + "presence_penalty", + "repetition_penalty", "response_format", + "seed", + "stop", + "temperature", + "tool_choice", + "tools", + "top_k", "top_logprobs", - "logprobs" + "top_p" + ] + }, + { + "id": "meta-llama/llama-3-70b-instruct", + "canonical_slug": "meta-llama/llama-3-70b-instruct", + "hugging_face_id": "meta-llama/Meta-Llama-3-70B-Instruct", + "name": "Meta: Llama 3 70B Instruct", + "created": 1713398400, + "description": "Meta's latest class of model (Llama 3) launched with a variety of sizes & flavors. This 70B instruct-tuned version was optimized for high quality dialogue usecases.\n\nIt has demonstrated strong performance compared to leading closed-source models in human evaluations.\n\nTo read more about the model release, [click here](https://ai.meta.com/blog/meta-llama-3/). Usage of this model is subject to [Meta's Acceptable Use Policy](https://llama.meta.com/llama3/use-policy/).", + "context_length": 8192, + "architecture": { + "modality": "text->text", + "input_modalities": [ + "text" + ], + "output_modalities": [ + "text" + ], + "tokenizer": "Llama3", + "instruct_type": "llama3" + }, + "pricing": { + "prompt": "0.0000003", + "completion": "0.0000004", + "request": "0", + "image": "0", + "audio": "0", + "web_search": "0", + "internal_reasoning": "0" + }, + "top_provider": { + "context_length": 8192, + "max_completion_tokens": 16384, + "is_moderated": false + }, + "per_request_limits": null, + "supported_parameters": [ + "frequency_penalty", + "logit_bias", + "logprobs", + "max_tokens", + "min_p", + "presence_penalty", + "repetition_penalty", + "response_format", + "seed", + "stop", + "temperature", + "tool_choice", + "tools", + "top_k", + "top_logprobs", + "top_p" ] }, { @@ -13596,6 +13786,7 @@ "completion": "0.0000009", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -13606,22 +13797,22 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "tools", - "tool_choice", - "stop", "frequency_penalty", - "presence_penalty", - "response_format", - "structured_outputs", - "seed", - "top_k", - "repetition_penalty", "logit_bias", "logprobs", - "top_logprobs" + "max_tokens", + "presence_penalty", + "repetition_penalty", + "response_format", + "seed", + "stop", + "structured_outputs", + "temperature", + "tool_choice", + "tools", + "top_k", + "top_logprobs", + "top_p" ] }, { @@ -13648,6 +13839,7 @@ "completion": "0.00000048", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -13658,68 +13850,18 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "stop", "frequency_penalty", - "presence_penalty", - "seed", - "top_k", + "logit_bias", + "max_tokens", "min_p", - "repetition_penalty", - "logit_bias", - "response_format" - ] - }, - { - "id": "openai/gpt-4-turbo", - "canonical_slug": "openai/gpt-4-turbo", - "hugging_face_id": null, - "name": "OpenAI: GPT-4 Turbo", - "created": 1712620800, - "description": "The latest GPT-4 Turbo model with vision capabilities. Vision requests can now use JSON mode and function calling.\n\nTraining data: up to December 2023.", - "context_length": 128000, - "architecture": { - "modality": "text+image->text", - "input_modalities": [ - "text", - "image" - ], - "output_modalities": [ - "text" - ], - "tokenizer": "GPT", - "instruct_type": null - }, - "pricing": { - "prompt": "0.00001", - "completion": "0.00003", - "request": "0", - "image": "0.01445", - "web_search": "0", - "internal_reasoning": "0" - }, - "top_provider": { - "context_length": 128000, - "max_completion_tokens": 4096, - "is_moderated": true - }, - "per_request_limits": null, - "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "tools", - "tool_choice", - "stop", - "frequency_penalty", "presence_penalty", + "repetition_penalty", + "response_format", "seed", - "logit_bias", - "logprobs", - "top_logprobs", - "response_format" + "stop", + "temperature", + "top_k", + "top_p" ] }, { @@ -13747,6 +13889,7 @@ "completion": "0.000005", "request": "0", "image": "0.0006575", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -13757,17 +13900,69 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "tools", - "tool_choice", - "stop", "frequency_penalty", + "max_tokens", "presence_penalty", - "seed", "response_format", - "structured_outputs" + "seed", + "stop", + "structured_outputs", + "temperature", + "tool_choice", + "tools", + "top_p" + ] + }, + { + "id": "openai/gpt-4-turbo", + "canonical_slug": "openai/gpt-4-turbo", + "hugging_face_id": null, + "name": "OpenAI: GPT-4 Turbo", + "created": 1712620800, + "description": "The latest GPT-4 Turbo model with vision capabilities. Vision requests can now use JSON mode and function calling.\n\nTraining data: up to December 2023.", + "context_length": 128000, + "architecture": { + "modality": "text+image->text", + "input_modalities": [ + "text", + "image" + ], + "output_modalities": [ + "text" + ], + "tokenizer": "GPT", + "instruct_type": null + }, + "pricing": { + "prompt": "0.00001", + "completion": "0.00003", + "request": "0", + "image": "0.01445", + "audio": "0", + "web_search": "0", + "internal_reasoning": "0" + }, + "top_provider": { + "context_length": 128000, + "max_completion_tokens": 4096, + "is_moderated": true + }, + "per_request_limits": null, + "supported_parameters": [ + "frequency_penalty", + "logit_bias", + "logprobs", + "max_tokens", + "presence_penalty", + "response_format", + "seed", + "stop", + "structured_outputs", + "temperature", + "tool_choice", + "tools", + "top_logprobs", + "top_p" ] }, { @@ -13794,6 +13989,7 @@ "completion": "0.000015", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -13804,17 +14000,17 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "tools", - "stop", "frequency_penalty", + "max_tokens", "presence_penalty", - "top_k", - "seed", "response_format", - "structured_outputs" + "seed", + "stop", + "structured_outputs", + "temperature", + "tools", + "top_k", + "top_p" ] }, { @@ -13841,6 +14037,7 @@ "completion": "0.000015", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -13851,17 +14048,17 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "tools", - "stop", "frequency_penalty", + "max_tokens", "presence_penalty", - "top_k", - "seed", "response_format", - "structured_outputs" + "seed", + "stop", + "structured_outputs", + "temperature", + "tools", + "top_k", + "top_p" ] }, { @@ -13888,6 +14085,7 @@ "completion": "0.0000008", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -13898,17 +14096,17 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "stop", "frequency_penalty", - "presence_penalty", - "seed", - "top_k", + "logit_bias", + "max_tokens", "min_p", + "presence_penalty", "repetition_penalty", - "logit_bias" + "seed", + "stop", + "temperature", + "top_k", + "top_p" ] }, { @@ -13935,6 +14133,7 @@ "completion": "0.000002", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -13945,16 +14144,16 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "stop", "frequency_penalty", + "max_tokens", "presence_penalty", - "top_k", - "seed", "response_format", - "structured_outputs" + "seed", + "stop", + "structured_outputs", + "temperature", + "top_k", + "top_p" ] }, { @@ -13981,6 +14180,7 @@ "completion": "0.0000015", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -13991,17 +14191,17 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "tools", - "stop", "frequency_penalty", + "max_tokens", "presence_penalty", - "top_k", - "seed", "response_format", - "structured_outputs" + "seed", + "stop", + "structured_outputs", + "temperature", + "tools", + "top_k", + "top_p" ] }, { @@ -14029,6 +14229,7 @@ "completion": "0.00000125", "request": "0", "image": "0.0004", + "audio": "0", "web_search": "0", "internal_reasoning": "0", "input_cache_read": "0.00000003", @@ -14042,12 +14243,12 @@ "per_request_limits": null, "supported_parameters": [ "max_tokens", + "stop", "temperature", - "top_p", - "tools", "tool_choice", + "tools", "top_k", - "stop" + "top_p" ] }, { @@ -14075,6 +14276,7 @@ "completion": "0.00000125", "request": "0", "image": "0.0004", + "audio": "0", "web_search": "0", "internal_reasoning": "0", "input_cache_read": "0.00000003", @@ -14088,58 +14290,12 @@ "per_request_limits": null, "supported_parameters": [ "max_tokens", + "stop", "temperature", - "top_p", - "tools", "tool_choice", - "top_k", - "stop" - ] - }, - { - "id": "anthropic/claude-3-sonnet", - "canonical_slug": "anthropic/claude-3-sonnet", - "hugging_face_id": null, - "name": "Anthropic: Claude 3 Sonnet", - "created": 1709596800, - "description": "Claude 3 Sonnet is an ideal balance of intelligence and speed for enterprise workloads. Maximum utility at a lower price, dependable, balanced for scaled deployments.\n\nSee the launch announcement and benchmark results [here](https://www.anthropic.com/news/claude-3-family)\n\n#multimodal", - "context_length": 200000, - "architecture": { - "modality": "text+image->text", - "input_modalities": [ - "text", - "image" - ], - "output_modalities": [ - "text" - ], - "tokenizer": "Claude", - "instruct_type": null - }, - "pricing": { - "prompt": "0.000003", - "completion": "0.000015", - "request": "0", - "image": "0.0048", - "web_search": "0", - "internal_reasoning": "0", - "input_cache_read": "0.0000003", - "input_cache_write": "0.00000375" - }, - "top_provider": { - "context_length": 200000, - "max_completion_tokens": 4096, - "is_moderated": false - }, - "per_request_limits": null, - "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", "tools", - "tool_choice", "top_k", - "stop" + "top_p" ] }, { @@ -14167,6 +14323,7 @@ "completion": "0.000075", "request": "0", "image": "0.024", + "audio": "0", "web_search": "0", "internal_reasoning": "0", "input_cache_read": "0.0000015", @@ -14180,12 +14337,12 @@ "per_request_limits": null, "supported_parameters": [ "max_tokens", + "stop", "temperature", - "top_p", - "tools", "tool_choice", + "tools", "top_k", - "stop" + "top_p" ] }, { @@ -14213,6 +14370,7 @@ "completion": "0.000075", "request": "0", "image": "0.024", + "audio": "0", "web_search": "0", "internal_reasoning": "0", "input_cache_read": "0.0000015", @@ -14226,12 +14384,12 @@ "per_request_limits": null, "supported_parameters": [ "max_tokens", + "stop", "temperature", - "top_p", - "tools", "tool_choice", + "tools", "top_k", - "stop" + "top_p" ] }, { @@ -14258,6 +14416,7 @@ "completion": "0.0000015", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -14268,17 +14427,17 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "tools", - "stop", "frequency_penalty", + "max_tokens", "presence_penalty", - "top_k", - "seed", "response_format", - "structured_outputs" + "seed", + "stop", + "structured_outputs", + "temperature", + "tools", + "top_k", + "top_p" ] }, { @@ -14305,6 +14464,7 @@ "completion": "0.000006", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -14315,67 +14475,17 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "tools", - "tool_choice", - "response_format", - "stop", - "seed", "frequency_penalty", - "presence_penalty", - "structured_outputs" - ] - }, - { - "id": "openai/gpt-4-turbo-preview", - "canonical_slug": "openai/gpt-4-turbo-preview", - "hugging_face_id": null, - "name": "OpenAI: GPT-4 Turbo Preview", - "created": 1706140800, - "description": "The preview GPT-4 model with improved instruction following, JSON mode, reproducible outputs, parallel function calling, and more. Training data: up to Dec 2023.\n\n**Note:** heavily rate limited by OpenAI while in preview.", - "context_length": 128000, - "architecture": { - "modality": "text->text", - "input_modalities": [ - "text" - ], - "output_modalities": [ - "text" - ], - "tokenizer": "GPT", - "instruct_type": null - }, - "pricing": { - "prompt": "0.00001", - "completion": "0.00003", - "request": "0", - "image": "0", - "web_search": "0", - "internal_reasoning": "0" - }, - "top_provider": { - "context_length": 128000, - "max_completion_tokens": 4096, - "is_moderated": true - }, - "per_request_limits": null, - "supported_parameters": [ "max_tokens", - "temperature", - "top_p", - "tools", - "tool_choice", - "stop", - "frequency_penalty", "presence_penalty", - "seed", - "logit_bias", - "logprobs", - "top_logprobs", "response_format", - "structured_outputs" + "seed", + "stop", + "structured_outputs", + "temperature", + "tool_choice", + "tools", + "top_p" ] }, { @@ -14402,6 +14512,7 @@ "completion": "0.000002", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -14412,20 +14523,71 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "tools", - "tool_choice", - "stop", "frequency_penalty", - "presence_penalty", - "seed", "logit_bias", "logprobs", - "top_logprobs", + "max_tokens", + "presence_penalty", "response_format", - "structured_outputs" + "seed", + "stop", + "structured_outputs", + "temperature", + "tool_choice", + "tools", + "top_logprobs", + "top_p" + ] + }, + { + "id": "openai/gpt-4-turbo-preview", + "canonical_slug": "openai/gpt-4-turbo-preview", + "hugging_face_id": null, + "name": "OpenAI: GPT-4 Turbo Preview", + "created": 1706140800, + "description": "The preview GPT-4 model with improved instruction following, JSON mode, reproducible outputs, parallel function calling, and more. Training data: up to Dec 2023.\n\n**Note:** heavily rate limited by OpenAI while in preview.", + "context_length": 128000, + "architecture": { + "modality": "text->text", + "input_modalities": [ + "text" + ], + "output_modalities": [ + "text" + ], + "tokenizer": "GPT", + "instruct_type": null + }, + "pricing": { + "prompt": "0.00001", + "completion": "0.00003", + "request": "0", + "image": "0", + "audio": "0", + "web_search": "0", + "internal_reasoning": "0" + }, + "top_provider": { + "context_length": 128000, + "max_completion_tokens": 4096, + "is_moderated": true + }, + "per_request_limits": null, + "supported_parameters": [ + "frequency_penalty", + "logit_bias", + "logprobs", + "max_tokens", + "presence_penalty", + "response_format", + "seed", + "stop", + "structured_outputs", + "temperature", + "tool_choice", + "tools", + "top_logprobs", + "top_p" ] }, { @@ -14452,6 +14614,7 @@ "completion": "0.0000006", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -14462,17 +14625,16 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "stop", "frequency_penalty", - "presence_penalty", - "top_k", - "repetition_penalty", "logit_bias", + "max_tokens", "min_p", - "response_format" + "presence_penalty", + "repetition_penalty", + "stop", + "temperature", + "top_k", + "top_p" ] }, { @@ -14499,6 +14661,7 @@ "completion": "0.0000006", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -14509,17 +14672,17 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "tools", - "tool_choice", - "stop", "frequency_penalty", + "max_tokens", "presence_penalty", "response_format", + "seed", + "stop", "structured_outputs", - "seed" + "temperature", + "tool_choice", + "tools", + "top_p" ] }, { @@ -14546,6 +14709,7 @@ "completion": "0.00000025", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -14556,17 +14720,17 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "tools", - "tool_choice", - "stop", "frequency_penalty", + "max_tokens", "presence_penalty", "response_format", + "seed", + "stop", "structured_outputs", - "seed" + "temperature", + "tool_choice", + "tools", + "top_p" ] }, { @@ -14593,6 +14757,7 @@ "completion": "0.0000002", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -14603,17 +14768,16 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "stop", "frequency_penalty", - "presence_penalty", - "top_k", - "repetition_penalty", "logit_bias", + "max_tokens", "min_p", - "response_format" + "presence_penalty", + "repetition_penalty", + "stop", + "temperature", + "top_k", + "top_p" ] }, { @@ -14640,6 +14804,7 @@ "completion": "0.00000024", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -14650,20 +14815,20 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "tools", - "tool_choice", - "stop", "frequency_penalty", - "presence_penalty", - "top_k", - "repetition_penalty", "logit_bias", + "max_tokens", "min_p", + "presence_penalty", + "repetition_penalty", "response_format", - "seed" + "seed", + "stop", + "temperature", + "tool_choice", + "tools", + "top_k", + "top_p" ] }, { @@ -14673,7 +14838,7 @@ "name": "Noromaid 20B", "created": 1700956800, "description": "A collab between IkariDev and Undi. This merge is suitable for RP, ERP, and general knowledge.\n\n#merge #uncensored", - "context_length": 8192, + "context_length": 4096, "architecture": { "modality": "text->text", "input_modalities": [ @@ -14686,78 +14851,35 @@ "instruct_type": "alpaca" }, "pricing": { - "prompt": "0.00000125", - "completion": "0.000002", - "request": "0", - "image": "0", - "web_search": "0", - "internal_reasoning": "0" - }, - "top_provider": { - "context_length": 8192, - "max_completion_tokens": 2048, - "is_moderated": false - }, - "per_request_limits": null, - "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "stop", - "frequency_penalty", - "presence_penalty", - "repetition_penalty", - "logit_bias", - "top_k", - "min_p", - "seed", - "top_a" - ] - }, - { - "id": "undi95/toppy-m-7b", - "canonical_slug": "undi95/toppy-m-7b", - "hugging_face_id": "Undi95/Toppy-M-7B", - "name": "Toppy M 7B", - "created": 1699574400, - "description": "A wild 7B parameter model that merges several models using the new task_arithmetic merge method from mergekit.\nList of merged models:\n- NousResearch/Nous-Capybara-7B-V1.9\n- [HuggingFaceH4/zephyr-7b-beta](/models/huggingfaceh4/zephyr-7b-beta)\n- lemonilia/AshhLimaRP-Mistral-7B\n- Vulkane/120-Days-of-Sodom-LoRA-Mistral-7b\n- Undi95/Mistral-pippa-sharegpt-7b-qlora\n\n#merge #uncensored", - "context_length": 4096, - "architecture": { - "modality": "text->text", - "input_modalities": [ - "text" - ], - "output_modalities": [ - "text" - ], - "tokenizer": "Mistral", - "instruct_type": "alpaca" - }, - "pricing": { - "prompt": "0.0000008", - "completion": "0.0000012", + "prompt": "0.000001", + "completion": "0.00000175", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, "top_provider": { "context_length": 4096, - "max_completion_tokens": 4096, + "max_completion_tokens": null, "is_moderated": false }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "stop", "frequency_penalty", + "logit_bias", + "max_tokens", + "min_p", "presence_penalty", "repetition_penalty", + "response_format", + "seed", + "stop", + "structured_outputs", + "temperature", + "top_a", "top_k", - "min_p", - "seed" + "top_p" ] }, { @@ -14784,6 +14906,7 @@ "completion": "0.000011", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -14794,20 +14917,67 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "stop", "frequency_penalty", - "presence_penalty", - "response_format", - "structured_outputs", - "repetition_penalty", "logit_bias", - "top_k", + "max_tokens", "min_p", + "presence_penalty", + "repetition_penalty", + "response_format", "seed", - "top_a" + "stop", + "structured_outputs", + "temperature", + "top_a", + "top_k", + "top_p" + ] + }, + { + "id": "undi95/toppy-m-7b", + "canonical_slug": "undi95/toppy-m-7b", + "hugging_face_id": "Undi95/Toppy-M-7B", + "name": "Toppy M 7B", + "created": 1699574400, + "description": "A wild 7B parameter model that merges several models using the new task_arithmetic merge method from mergekit.\nList of merged models:\n- NousResearch/Nous-Capybara-7B-V1.9\n- [HuggingFaceH4/zephyr-7b-beta](/models/huggingfaceh4/zephyr-7b-beta)\n- lemonilia/AshhLimaRP-Mistral-7B\n- Vulkane/120-Days-of-Sodom-LoRA-Mistral-7b\n- Undi95/Mistral-pippa-sharegpt-7b-qlora\n\n#merge #uncensored", + "context_length": 4096, + "architecture": { + "modality": "text->text", + "input_modalities": [ + "text" + ], + "output_modalities": [ + "text" + ], + "tokenizer": "Mistral", + "instruct_type": "alpaca" + }, + "pricing": { + "prompt": "0.0000008", + "completion": "0.0000012", + "request": "0", + "image": "0", + "audio": "0", + "web_search": "0", + "internal_reasoning": "0" + }, + "top_provider": { + "context_length": 4096, + "max_completion_tokens": 4096, + "is_moderated": false + }, + "per_request_limits": null, + "supported_parameters": [ + "frequency_penalty", + "max_tokens", + "min_p", + "presence_penalty", + "repetition_penalty", + "seed", + "stop", + "temperature", + "top_k", + "top_p" ] }, { @@ -14865,6 +15035,7 @@ "completion": "0.00003", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -14875,67 +15046,20 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "tools", - "tool_choice", - "stop", "frequency_penalty", - "presence_penalty", - "seed", "logit_bias", "logprobs", - "top_logprobs", + "max_tokens", + "presence_penalty", "response_format", - "structured_outputs" - ] - }, - { - "id": "openai/gpt-3.5-turbo-instruct", - "canonical_slug": "openai/gpt-3.5-turbo-instruct", - "hugging_face_id": null, - "name": "OpenAI: GPT-3.5 Turbo Instruct", - "created": 1695859200, - "description": "This model is a variant of GPT-3.5 Turbo tuned for instructional prompts and omitting chat-related optimizations. Training data: up to Sep 2021.", - "context_length": 4095, - "architecture": { - "modality": "text->text", - "input_modalities": [ - "text" - ], - "output_modalities": [ - "text" - ], - "tokenizer": "GPT", - "instruct_type": "chatml" - }, - "pricing": { - "prompt": "0.0000015", - "completion": "0.000002", - "request": "0", - "image": "0", - "web_search": "0", - "internal_reasoning": "0" - }, - "top_provider": { - "context_length": 4095, - "max_completion_tokens": 4096, - "is_moderated": true - }, - "per_request_limits": null, - "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "stop", - "frequency_penalty", - "presence_penalty", "seed", - "logit_bias", - "logprobs", + "stop", + "structured_outputs", + "temperature", + "tool_choice", + "tools", "top_logprobs", - "response_format" + "top_p" ] }, { @@ -14962,6 +15086,7 @@ "completion": "0.00000019", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -14972,20 +15097,68 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "tools", - "tool_choice", - "stop", "frequency_penalty", - "presence_penalty", - "top_k", - "repetition_penalty", "logit_bias", + "max_tokens", "min_p", + "presence_penalty", + "repetition_penalty", + "seed", + "stop", + "temperature", + "tool_choice", + "tools", + "top_k", + "top_p" + ] + }, + { + "id": "openai/gpt-3.5-turbo-instruct", + "canonical_slug": "openai/gpt-3.5-turbo-instruct", + "hugging_face_id": null, + "name": "OpenAI: GPT-3.5 Turbo Instruct", + "created": 1695859200, + "description": "This model is a variant of GPT-3.5 Turbo tuned for instructional prompts and omitting chat-related optimizations. Training data: up to Sep 2021.", + "context_length": 4095, + "architecture": { + "modality": "text->text", + "input_modalities": [ + "text" + ], + "output_modalities": [ + "text" + ], + "tokenizer": "GPT", + "instruct_type": "chatml" + }, + "pricing": { + "prompt": "0.0000015", + "completion": "0.000002", + "request": "0", + "image": "0", + "audio": "0", + "web_search": "0", + "internal_reasoning": "0" + }, + "top_provider": { + "context_length": 4095, + "max_completion_tokens": 4096, + "is_moderated": true + }, + "per_request_limits": null, + "supported_parameters": [ + "frequency_penalty", + "logit_bias", + "logprobs", + "max_tokens", + "presence_penalty", "response_format", - "seed" + "seed", + "stop", + "structured_outputs", + "temperature", + "top_logprobs", + "top_p" ] }, { @@ -15008,10 +15181,11 @@ "instruct_type": "alpaca" }, "pricing": { - "prompt": "0.0000008", - "completion": "0.0000012", + "prompt": "0.0000006", + "completion": "0.000001", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -15022,16 +15196,16 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "stop", "frequency_penalty", + "max_tokens", + "min_p", "presence_penalty", "repetition_penalty", + "seed", + "stop", + "temperature", "top_k", - "min_p", - "seed" + "top_p" ] }, { @@ -15058,6 +15232,7 @@ "completion": "0.000004", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -15068,19 +15243,20 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "tools", - "tool_choice", - "stop", "frequency_penalty", - "presence_penalty", - "seed", "logit_bias", "logprobs", + "max_tokens", + "presence_penalty", + "response_format", + "seed", + "stop", + "structured_outputs", + "temperature", + "tool_choice", + "tools", "top_logprobs", - "response_format" + "top_p" ] }, { @@ -15107,6 +15283,7 @@ "completion": "0.0000015", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -15117,18 +15294,18 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "stop", "frequency_penalty", + "logit_bias", + "max_tokens", + "min_p", "presence_penalty", "repetition_penalty", - "logit_bias", - "top_k", - "min_p", "seed", - "top_a" + "stop", + "temperature", + "top_a", + "top_k", + "top_p" ] }, { @@ -15155,6 +15332,7 @@ "completion": "0.000001", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -15165,20 +15343,20 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "stop", "frequency_penalty", - "presence_penalty", - "response_format", - "structured_outputs", - "repetition_penalty", - "top_k", - "min_p", - "seed", "logit_bias", - "top_a" + "max_tokens", + "min_p", + "presence_penalty", + "repetition_penalty", + "response_format", + "seed", + "stop", + "structured_outputs", + "temperature", + "top_a", + "top_k", + "top_p" ] }, { @@ -15205,6 +15383,7 @@ "completion": "0.00000006", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -15215,119 +15394,20 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", "frequency_penalty", + "logit_bias", + "max_tokens", "min_p", "presence_penalty", "repetition_penalty", + "response_format", "seed", "stop", - "top_k", - "response_format", "structured_outputs", - "logit_bias", - "top_a" - ] - }, - { - "id": "openai/gpt-4-0314", - "canonical_slug": "openai/gpt-4-0314", - "hugging_face_id": null, - "name": "OpenAI: GPT-4 (older v0314)", - "created": 1685232000, - "description": "GPT-4-0314 is the first version of GPT-4 released, with a context length of 8,192 tokens, and was supported until June 14. Training data: up to Sep 2021.", - "context_length": 8191, - "architecture": { - "modality": "text->text", - "input_modalities": [ - "text" - ], - "output_modalities": [ - "text" - ], - "tokenizer": "GPT", - "instruct_type": null - }, - "pricing": { - "prompt": "0.00003", - "completion": "0.00006", - "request": "0", - "image": "0", - "web_search": "0", - "internal_reasoning": "0" - }, - "top_provider": { - "context_length": 8191, - "max_completion_tokens": 4096, - "is_moderated": true - }, - "per_request_limits": null, - "supported_parameters": [ - "max_tokens", "temperature", - "top_p", - "tools", - "tool_choice", - "stop", - "frequency_penalty", - "presence_penalty", - "seed", - "logit_bias", - "logprobs", - "top_logprobs", - "response_format", - "structured_outputs" - ] - }, - { - "id": "openai/gpt-3.5-turbo", - "canonical_slug": "openai/gpt-3.5-turbo", - "hugging_face_id": null, - "name": "OpenAI: GPT-3.5 Turbo", - "created": 1685232000, - "description": "GPT-3.5 Turbo is OpenAI's fastest model. It can understand and generate natural language or code, and is optimized for chat and traditional completion tasks.\n\nTraining data up to Sep 2021.", - "context_length": 16385, - "architecture": { - "modality": "text->text", - "input_modalities": [ - "text" - ], - "output_modalities": [ - "text" - ], - "tokenizer": "GPT", - "instruct_type": null - }, - "pricing": { - "prompt": "0.0000005", - "completion": "0.0000015", - "request": "0", - "image": "0", - "web_search": "0", - "internal_reasoning": "0" - }, - "top_provider": { - "context_length": 16385, - "max_completion_tokens": 4096, - "is_moderated": true - }, - "per_request_limits": null, - "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "tools", - "tool_choice", - "stop", - "frequency_penalty", - "presence_penalty", - "seed", - "logit_bias", - "logprobs", - "top_logprobs", - "response_format" + "top_a", + "top_k", + "top_p" ] }, { @@ -15354,6 +15434,7 @@ "completion": "0.00006", "request": "0", "image": "0", + "audio": "0", "web_search": "0", "internal_reasoning": "0" }, @@ -15364,19 +15445,122 @@ }, "per_request_limits": null, "supported_parameters": [ - "max_tokens", - "temperature", - "top_p", - "tools", - "tool_choice", - "stop", "frequency_penalty", - "presence_penalty", - "seed", "logit_bias", "logprobs", + "max_tokens", + "presence_penalty", + "response_format", + "seed", + "stop", + "structured_outputs", + "temperature", + "tool_choice", + "tools", "top_logprobs", - "response_format" + "top_p" + ] + }, + { + "id": "openai/gpt-3.5-turbo", + "canonical_slug": "openai/gpt-3.5-turbo", + "hugging_face_id": null, + "name": "OpenAI: GPT-3.5 Turbo", + "created": 1685232000, + "description": "GPT-3.5 Turbo is OpenAI's fastest model. It can understand and generate natural language or code, and is optimized for chat and traditional completion tasks.\n\nTraining data up to Sep 2021.", + "context_length": 16385, + "architecture": { + "modality": "text->text", + "input_modalities": [ + "text" + ], + "output_modalities": [ + "text" + ], + "tokenizer": "GPT", + "instruct_type": null + }, + "pricing": { + "prompt": "0.0000005", + "completion": "0.0000015", + "request": "0", + "image": "0", + "audio": "0", + "web_search": "0", + "internal_reasoning": "0" + }, + "top_provider": { + "context_length": 16385, + "max_completion_tokens": 4096, + "is_moderated": true + }, + "per_request_limits": null, + "supported_parameters": [ + "frequency_penalty", + "logit_bias", + "logprobs", + "max_tokens", + "presence_penalty", + "response_format", + "seed", + "stop", + "structured_outputs", + "temperature", + "tool_choice", + "tools", + "top_logprobs", + "top_p" + ] + }, + { + "id": "openai/gpt-4-0314", + "canonical_slug": "openai/gpt-4-0314", + "hugging_face_id": null, + "name": "OpenAI: GPT-4 (older v0314)", + "created": 1685232000, + "description": "GPT-4-0314 is the first version of GPT-4 released, with a context length of 8,192 tokens, and was supported until June 14. Training data: up to Sep 2021.", + "context_length": 8191, + "architecture": { + "modality": "text->text", + "input_modalities": [ + "text" + ], + "output_modalities": [ + "text" + ], + "tokenizer": "GPT", + "instruct_type": null + }, + "pricing": { + "prompt": "0.00003", + "completion": "0.00006", + "request": "0", + "image": "0", + "audio": "0", + "web_search": "0", + "internal_reasoning": "0" + }, + "top_provider": { + "context_length": 8191, + "max_completion_tokens": 4096, + "is_moderated": true + }, + "per_request_limits": null, + "supported_parameters": [ + "frequency_penalty", + "logit_bias", + "logprobs", + "max_tokens", + "presence_penalty", + "response_format", + "seed", + "stop", + "structured_outputs", + "temperature", + "tool_choice", + "tools", + "top_logprobs", + "top_p" ] } ] diff --git a/packages/kbot/dist-in/src/models/cache/openai.ts b/packages/kbot/dist-in/src/models/cache/openai.ts index 58b62da7..0b347100 100644 --- a/packages/kbot/dist-in/src/models/cache/openai.ts +++ b/packages/kbot/dist-in/src/models/cache/openai.ts @@ -1 +1 @@ -export const models = [{"id":"gpt-4-0613","object":"model","created":1686588896,"owned_by":"openai"},{"id":"gpt-4","object":"model","created":1687882411,"owned_by":"openai"},{"id":"gpt-3.5-turbo","object":"model","created":1677610602,"owned_by":"openai"},{"id":"o4-mini-deep-research-2025-06-26","object":"model","created":1750866121,"owned_by":"system"},{"id":"codex-mini-latest","object":"model","created":1746673257,"owned_by":"system"},{"id":"gpt-4o-realtime-preview-2025-06-03","object":"model","created":1748907838,"owned_by":"system"},{"id":"gpt-4o-audio-preview-2025-06-03","object":"model","created":1748908498,"owned_by":"system"},{"id":"o4-mini-deep-research","object":"model","created":1749685485,"owned_by":"system"},{"id":"davinci-002","object":"model","created":1692634301,"owned_by":"system"},{"id":"babbage-002","object":"model","created":1692634615,"owned_by":"system"},{"id":"gpt-3.5-turbo-instruct","object":"model","created":1692901427,"owned_by":"system"},{"id":"gpt-3.5-turbo-instruct-0914","object":"model","created":1694122472,"owned_by":"system"},{"id":"dall-e-3","object":"model","created":1698785189,"owned_by":"system"},{"id":"dall-e-2","object":"model","created":1698798177,"owned_by":"system"},{"id":"gpt-4-1106-preview","object":"model","created":1698957206,"owned_by":"system"},{"id":"gpt-3.5-turbo-1106","object":"model","created":1698959748,"owned_by":"system"},{"id":"tts-1-hd","object":"model","created":1699046015,"owned_by":"system"},{"id":"tts-1-1106","object":"model","created":1699053241,"owned_by":"system"},{"id":"tts-1-hd-1106","object":"model","created":1699053533,"owned_by":"system"},{"id":"text-embedding-3-small","object":"model","created":1705948997,"owned_by":"system"},{"id":"text-embedding-3-large","object":"model","created":1705953180,"owned_by":"system"},{"id":"gpt-4-0125-preview","object":"model","created":1706037612,"owned_by":"system"},{"id":"gpt-4-turbo-preview","object":"model","created":1706037777,"owned_by":"system"},{"id":"gpt-3.5-turbo-0125","object":"model","created":1706048358,"owned_by":"system"},{"id":"gpt-4-turbo","object":"model","created":1712361441,"owned_by":"system"},{"id":"gpt-4-turbo-2024-04-09","object":"model","created":1712601677,"owned_by":"system"},{"id":"gpt-4o","object":"model","created":1715367049,"owned_by":"system"},{"id":"gpt-4o-2024-05-13","object":"model","created":1715368132,"owned_by":"system"},{"id":"gpt-4o-mini-2024-07-18","object":"model","created":1721172717,"owned_by":"system"},{"id":"gpt-4o-mini","object":"model","created":1721172741,"owned_by":"system"},{"id":"gpt-4o-2024-08-06","object":"model","created":1722814719,"owned_by":"system"},{"id":"chatgpt-4o-latest","object":"model","created":1723515131,"owned_by":"system"},{"id":"o1-preview-2024-09-12","object":"model","created":1725648865,"owned_by":"system"},{"id":"o1-preview","object":"model","created":1725648897,"owned_by":"system"},{"id":"o1-mini-2024-09-12","object":"model","created":1725648979,"owned_by":"system"},{"id":"o1-mini","object":"model","created":1725649008,"owned_by":"system"},{"id":"gpt-4o-realtime-preview-2024-10-01","object":"model","created":1727131766,"owned_by":"system"},{"id":"gpt-4o-audio-preview-2024-10-01","object":"model","created":1727389042,"owned_by":"system"},{"id":"gpt-4o-audio-preview","object":"model","created":1727460443,"owned_by":"system"},{"id":"gpt-4o-realtime-preview","object":"model","created":1727659998,"owned_by":"system"},{"id":"omni-moderation-latest","object":"model","created":1731689265,"owned_by":"system"},{"id":"omni-moderation-2024-09-26","object":"model","created":1732734466,"owned_by":"system"},{"id":"gpt-4o-realtime-preview-2024-12-17","object":"model","created":1733945430,"owned_by":"system"},{"id":"gpt-4o-audio-preview-2024-12-17","object":"model","created":1734034239,"owned_by":"system"},{"id":"gpt-4o-mini-realtime-preview-2024-12-17","object":"model","created":1734112601,"owned_by":"system"},{"id":"gpt-4o-mini-audio-preview-2024-12-17","object":"model","created":1734115920,"owned_by":"system"},{"id":"o1-2024-12-17","object":"model","created":1734326976,"owned_by":"system"},{"id":"o1","object":"model","created":1734375816,"owned_by":"system"},{"id":"gpt-4o-mini-realtime-preview","object":"model","created":1734387380,"owned_by":"system"},{"id":"gpt-4o-mini-audio-preview","object":"model","created":1734387424,"owned_by":"system"},{"id":"o3-mini","object":"model","created":1737146383,"owned_by":"system"},{"id":"o3-mini-2025-01-31","object":"model","created":1738010200,"owned_by":"system"},{"id":"gpt-4o-2024-11-20","object":"model","created":1739331543,"owned_by":"system"},{"id":"gpt-4o-search-preview-2025-03-11","object":"model","created":1741388170,"owned_by":"system"},{"id":"gpt-4o-search-preview","object":"model","created":1741388720,"owned_by":"system"},{"id":"gpt-4o-mini-search-preview-2025-03-11","object":"model","created":1741390858,"owned_by":"system"},{"id":"gpt-4o-mini-search-preview","object":"model","created":1741391161,"owned_by":"system"},{"id":"gpt-4o-transcribe","object":"model","created":1742068463,"owned_by":"system"},{"id":"gpt-4o-mini-transcribe","object":"model","created":1742068596,"owned_by":"system"},{"id":"o1-pro-2025-03-19","object":"model","created":1742251504,"owned_by":"system"},{"id":"o1-pro","object":"model","created":1742251791,"owned_by":"system"},{"id":"gpt-4o-mini-tts","object":"model","created":1742403959,"owned_by":"system"},{"id":"o4-mini-2025-04-16","object":"model","created":1744133506,"owned_by":"system"},{"id":"o4-mini","object":"model","created":1744225351,"owned_by":"system"},{"id":"gpt-4.1-2025-04-14","object":"model","created":1744315746,"owned_by":"system"},{"id":"gpt-4.1","object":"model","created":1744316542,"owned_by":"system"},{"id":"gpt-4.1-mini-2025-04-14","object":"model","created":1744317547,"owned_by":"system"},{"id":"gpt-4.1-mini","object":"model","created":1744318173,"owned_by":"system"},{"id":"gpt-4.1-nano-2025-04-14","object":"model","created":1744321025,"owned_by":"system"},{"id":"gpt-4.1-nano","object":"model","created":1744321707,"owned_by":"system"},{"id":"gpt-image-1","object":"model","created":1745517030,"owned_by":"system"},{"id":"gpt-3.5-turbo-16k","object":"model","created":1683758102,"owned_by":"openai-internal"},{"id":"tts-1","object":"model","created":1681940951,"owned_by":"openai-internal"},{"id":"whisper-1","object":"model","created":1677532384,"owned_by":"openai-internal"},{"id":"text-embedding-ada-002","object":"model","created":1671217299,"owned_by":"openai-internal"}] \ No newline at end of file +export const models = [{"id":"gpt-4-0613","object":"model","created":1686588896,"owned_by":"openai"},{"id":"gpt-4","object":"model","created":1687882411,"owned_by":"openai"},{"id":"gpt-3.5-turbo","object":"model","created":1677610602,"owned_by":"openai"},{"id":"o4-mini-deep-research-2025-06-26","object":"model","created":1750866121,"owned_by":"system"},{"id":"codex-mini-latest","object":"model","created":1746673257,"owned_by":"system"},{"id":"gpt-4o-realtime-preview-2025-06-03","object":"model","created":1748907838,"owned_by":"system"},{"id":"gpt-4o-audio-preview-2025-06-03","object":"model","created":1748908498,"owned_by":"system"},{"id":"o4-mini-deep-research","object":"model","created":1749685485,"owned_by":"system"},{"id":"davinci-002","object":"model","created":1692634301,"owned_by":"system"},{"id":"babbage-002","object":"model","created":1692634615,"owned_by":"system"},{"id":"gpt-3.5-turbo-instruct","object":"model","created":1692901427,"owned_by":"system"},{"id":"gpt-3.5-turbo-instruct-0914","object":"model","created":1694122472,"owned_by":"system"},{"id":"dall-e-3","object":"model","created":1698785189,"owned_by":"system"},{"id":"dall-e-2","object":"model","created":1698798177,"owned_by":"system"},{"id":"gpt-4-1106-preview","object":"model","created":1698957206,"owned_by":"system"},{"id":"gpt-3.5-turbo-1106","object":"model","created":1698959748,"owned_by":"system"},{"id":"tts-1-hd","object":"model","created":1699046015,"owned_by":"system"},{"id":"tts-1-1106","object":"model","created":1699053241,"owned_by":"system"},{"id":"tts-1-hd-1106","object":"model","created":1699053533,"owned_by":"system"},{"id":"text-embedding-3-small","object":"model","created":1705948997,"owned_by":"system"},{"id":"text-embedding-3-large","object":"model","created":1705953180,"owned_by":"system"},{"id":"gpt-4-0125-preview","object":"model","created":1706037612,"owned_by":"system"},{"id":"gpt-4-turbo-preview","object":"model","created":1706037777,"owned_by":"system"},{"id":"gpt-3.5-turbo-0125","object":"model","created":1706048358,"owned_by":"system"},{"id":"gpt-4-turbo","object":"model","created":1712361441,"owned_by":"system"},{"id":"gpt-4-turbo-2024-04-09","object":"model","created":1712601677,"owned_by":"system"},{"id":"gpt-4o","object":"model","created":1715367049,"owned_by":"system"},{"id":"gpt-4o-2024-05-13","object":"model","created":1715368132,"owned_by":"system"},{"id":"gpt-4o-mini-2024-07-18","object":"model","created":1721172717,"owned_by":"system"},{"id":"gpt-4o-mini","object":"model","created":1721172741,"owned_by":"system"},{"id":"gpt-4o-2024-08-06","object":"model","created":1722814719,"owned_by":"system"},{"id":"chatgpt-4o-latest","object":"model","created":1723515131,"owned_by":"system"},{"id":"o1-mini-2024-09-12","object":"model","created":1725648979,"owned_by":"system"},{"id":"o1-mini","object":"model","created":1725649008,"owned_by":"system"},{"id":"gpt-4o-realtime-preview-2024-10-01","object":"model","created":1727131766,"owned_by":"system"},{"id":"gpt-4o-audio-preview-2024-10-01","object":"model","created":1727389042,"owned_by":"system"},{"id":"gpt-4o-audio-preview","object":"model","created":1727460443,"owned_by":"system"},{"id":"gpt-4o-realtime-preview","object":"model","created":1727659998,"owned_by":"system"},{"id":"omni-moderation-latest","object":"model","created":1731689265,"owned_by":"system"},{"id":"omni-moderation-2024-09-26","object":"model","created":1732734466,"owned_by":"system"},{"id":"gpt-4o-realtime-preview-2024-12-17","object":"model","created":1733945430,"owned_by":"system"},{"id":"gpt-4o-audio-preview-2024-12-17","object":"model","created":1734034239,"owned_by":"system"},{"id":"gpt-4o-mini-realtime-preview-2024-12-17","object":"model","created":1734112601,"owned_by":"system"},{"id":"gpt-4o-mini-audio-preview-2024-12-17","object":"model","created":1734115920,"owned_by":"system"},{"id":"o1-2024-12-17","object":"model","created":1734326976,"owned_by":"system"},{"id":"o1","object":"model","created":1734375816,"owned_by":"system"},{"id":"gpt-4o-mini-realtime-preview","object":"model","created":1734387380,"owned_by":"system"},{"id":"gpt-4o-mini-audio-preview","object":"model","created":1734387424,"owned_by":"system"},{"id":"o3-mini","object":"model","created":1737146383,"owned_by":"system"},{"id":"o3-mini-2025-01-31","object":"model","created":1738010200,"owned_by":"system"},{"id":"gpt-4o-2024-11-20","object":"model","created":1739331543,"owned_by":"system"},{"id":"gpt-4o-search-preview-2025-03-11","object":"model","created":1741388170,"owned_by":"system"},{"id":"gpt-4o-search-preview","object":"model","created":1741388720,"owned_by":"system"},{"id":"gpt-4o-mini-search-preview-2025-03-11","object":"model","created":1741390858,"owned_by":"system"},{"id":"gpt-4o-mini-search-preview","object":"model","created":1741391161,"owned_by":"system"},{"id":"gpt-4o-transcribe","object":"model","created":1742068463,"owned_by":"system"},{"id":"gpt-4o-mini-transcribe","object":"model","created":1742068596,"owned_by":"system"},{"id":"o1-pro-2025-03-19","object":"model","created":1742251504,"owned_by":"system"},{"id":"o1-pro","object":"model","created":1742251791,"owned_by":"system"},{"id":"gpt-4o-mini-tts","object":"model","created":1742403959,"owned_by":"system"},{"id":"o4-mini-2025-04-16","object":"model","created":1744133506,"owned_by":"system"},{"id":"o4-mini","object":"model","created":1744225351,"owned_by":"system"},{"id":"gpt-4.1-2025-04-14","object":"model","created":1744315746,"owned_by":"system"},{"id":"gpt-4.1","object":"model","created":1744316542,"owned_by":"system"},{"id":"gpt-4.1-mini-2025-04-14","object":"model","created":1744317547,"owned_by":"system"},{"id":"gpt-4.1-mini","object":"model","created":1744318173,"owned_by":"system"},{"id":"gpt-4.1-nano-2025-04-14","object":"model","created":1744321025,"owned_by":"system"},{"id":"gpt-4.1-nano","object":"model","created":1744321707,"owned_by":"system"},{"id":"gpt-image-1","object":"model","created":1745517030,"owned_by":"system"},{"id":"gpt-3.5-turbo-16k","object":"model","created":1683758102,"owned_by":"openai-internal"},{"id":"tts-1","object":"model","created":1681940951,"owned_by":"openai-internal"},{"id":"whisper-1","object":"model","created":1677532384,"owned_by":"openai-internal"},{"id":"text-embedding-ada-002","object":"model","created":1671217299,"owned_by":"openai-internal"}] \ No newline at end of file diff --git a/packages/kbot/dist-in/src/models/cache/openrouter.ts b/packages/kbot/dist-in/src/models/cache/openrouter.ts index 70ce1935..fbb85f7f 100644 --- a/packages/kbot/dist-in/src/models/cache/openrouter.ts +++ b/packages/kbot/dist-in/src/models/cache/openrouter.ts @@ -1 +1 @@ -export const models = [{"id":"qwen/qwen3-235b-a22b-thinking-2507","name":"Qwen: Qwen3 235B A22B Thinking 2507","pricing":{"prompt":"0.0000001179","completion":"0.0000001179","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1753449557,"top_provider":{"context_length":262144,"max_completion_tokens":null,"is_moderated":false}},{"id":"z-ai/glm-4-32b","name":"Z.AI: GLM 4 32B ","pricing":{"prompt":"0.0000001","completion":"0.0000001","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1753376617,"top_provider":{"context_length":128000,"max_completion_tokens":null,"is_moderated":false}},{"id":"qwen/qwen3-coder:free","name":"Qwen: Qwen3 Coder (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1753230546,"top_provider":{"context_length":262144,"max_completion_tokens":null,"is_moderated":false}},{"id":"qwen/qwen3-coder","name":"Qwen: Qwen3 Coder ","pricing":{"prompt":"0.0000003","completion":"0.0000012","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1753230546,"top_provider":{"context_length":262144,"max_completion_tokens":null,"is_moderated":false}},{"id":"bytedance/ui-tars-1.5-7b","name":"Bytedance: UI-TARS 7B ","pricing":{"prompt":"0.0000001","completion":"0.0000002","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1753205056,"top_provider":{"context_length":128000,"max_completion_tokens":2048,"is_moderated":false}},{"id":"google/gemini-2.5-flash-lite","name":"Google: Gemini 2.5 Flash Lite","pricing":{"prompt":"0.0000001","completion":"0.0000004","request":"0","image":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.000000025","input_cache_write":"0.0000001833"},"created":1753200276,"top_provider":{"context_length":1048576,"max_completion_tokens":65535,"is_moderated":false}},{"id":"qwen/qwen3-235b-a22b-2507:free","name":"Qwen: Qwen3 235B A22B Instruct 2507 (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1753119555,"top_provider":{"context_length":262144,"max_completion_tokens":null,"is_moderated":false}},{"id":"qwen/qwen3-235b-a22b-2507","name":"Qwen: Qwen3 235B A22B Instruct 2507","pricing":{"prompt":"0.00000012","completion":"0.00000059","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1753119555,"top_provider":{"context_length":262144,"max_completion_tokens":262144,"is_moderated":false}},{"id":"switchpoint/router","name":"Switchpoint Router","pricing":{"prompt":"0.00000085","completion":"0.0000034","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1752272899,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"moonshotai/kimi-k2:free","name":"MoonshotAI: Kimi K2 (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1752263252,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"moonshotai/kimi-k2","name":"MoonshotAI: Kimi K2","pricing":{"prompt":"0.00000014","completion":"0.00000249","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1752263252,"top_provider":{"context_length":63000,"max_completion_tokens":63000,"is_moderated":false}},{"id":"thudm/glm-4.1v-9b-thinking","name":"THUDM: GLM 4.1V 9B Thinking","pricing":{"prompt":"0.000000035","completion":"0.000000138","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1752244385,"top_provider":{"context_length":65536,"max_completion_tokens":8000,"is_moderated":false}},{"id":"mistralai/devstral-medium","name":"Mistral: Devstral Medium","pricing":{"prompt":"0.0000004","completion":"0.000002","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1752161321,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"mistralai/devstral-small","name":"Mistral: Devstral Small 1.1","pricing":{"prompt":"0.00000007","completion":"0.00000028","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1752160751,"top_provider":{"context_length":128000,"max_completion_tokens":null,"is_moderated":false}},{"id":"cognitivecomputations/dolphin-mistral-24b-venice-edition:free","name":"Venice: Uncensored (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1752094966,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"x-ai/grok-4","name":"xAI: Grok 4","pricing":{"prompt":"0.000003","completion":"0.000015","request":"0","image":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.00000075"},"created":1752087689,"top_provider":{"context_length":256000,"max_completion_tokens":null,"is_moderated":false}},{"id":"google/gemma-3n-e2b-it:free","name":"Google: Gemma 3n 2B (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1752074904,"top_provider":{"context_length":8192,"max_completion_tokens":2048,"is_moderated":false}},{"id":"tencent/hunyuan-a13b-instruct:free","name":"Tencent: Hunyuan A13B Instruct (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1751987664,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"tencent/hunyuan-a13b-instruct","name":"Tencent: Hunyuan A13B Instruct","pricing":{"prompt":"0.00000003","completion":"0.00000003","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1751987664,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"tngtech/deepseek-r1t2-chimera:free","name":"TNG: DeepSeek R1T2 Chimera (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1751986985,"top_provider":{"context_length":163840,"max_completion_tokens":null,"is_moderated":false}},{"id":"tngtech/deepseek-r1t2-chimera","name":"TNG: DeepSeek R1T2 Chimera","pricing":{"prompt":"0.000000302","completion":"0.000000302","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1751986985,"top_provider":{"context_length":163840,"max_completion_tokens":null,"is_moderated":false}},{"id":"morph/morph-v3-large","name":"Morph: Morph V3 Large","pricing":{"prompt":"0.0000012","completion":"0.0000027","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1751910858,"top_provider":{"context_length":32000,"max_completion_tokens":16000,"is_moderated":false}},{"id":"morph/morph-v3-fast","name":"Morph: Morph V3 Fast","pricing":{"prompt":"0.0000012","completion":"0.0000027","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1751910002,"top_provider":{"context_length":32000,"max_completion_tokens":16000,"is_moderated":false}},{"id":"baidu/ernie-4.5-300b-a47b","name":"Baidu: ERNIE 4.5 300B A47B ","pricing":{"prompt":"0.00000028","completion":"0.0000011","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1751300139,"top_provider":{"context_length":123000,"max_completion_tokens":12000,"is_moderated":false}},{"id":"thedrummer/anubis-70b-v1.1","name":"TheDrummer: Anubis 70B V1.1","pricing":{"prompt":"0.0000005","completion":"0.0000008","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1751208347,"top_provider":{"context_length":131072,"max_completion_tokens":131072,"is_moderated":false}},{"id":"inception/mercury","name":"Inception: Mercury","pricing":{"prompt":"0.00000025","completion":"0.000001","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1750973026,"top_provider":{"context_length":32000,"max_completion_tokens":16000,"is_moderated":false}},{"id":"morph/morph-v2","name":"Morph: Fast Apply","pricing":{"prompt":"0.0000012","completion":"0.0000027","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1750946108,"top_provider":{"context_length":32000,"max_completion_tokens":16000,"is_moderated":false}},{"id":"mistralai/mistral-small-3.2-24b-instruct:free","name":"Mistral: Mistral Small 3.2 24B (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1750443016,"top_provider":{"context_length":96000,"max_completion_tokens":null,"is_moderated":false}},{"id":"mistralai/mistral-small-3.2-24b-instruct","name":"Mistral: Mistral Small 3.2 24B","pricing":{"prompt":"0.00000005","completion":"0.0000001","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1750443016,"top_provider":{"context_length":128000,"max_completion_tokens":null,"is_moderated":false}},{"id":"minimax/minimax-m1","name":"MiniMax: MiniMax M1","pricing":{"prompt":"0.0000003","completion":"0.00000165","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1750200414,"top_provider":{"context_length":1000000,"max_completion_tokens":40000,"is_moderated":false}},{"id":"google/gemini-2.5-flash-lite-preview-06-17","name":"Google: Gemini 2.5 Flash Lite Preview 06-17","pricing":{"prompt":"0.0000001","completion":"0.0000004","request":"0","image":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.000000025","input_cache_write":"0.0000001833"},"created":1750173831,"top_provider":{"context_length":1048576,"max_completion_tokens":65535,"is_moderated":false}},{"id":"google/gemini-2.5-flash","name":"Google: Gemini 2.5 Flash","pricing":{"prompt":"0.0000003","completion":"0.0000025","request":"0","image":"0.001238","web_search":"0","internal_reasoning":"0","input_cache_read":"0.000000075","input_cache_write":"0.0000003833"},"created":1750172488,"top_provider":{"context_length":1048576,"max_completion_tokens":65535,"is_moderated":false}},{"id":"google/gemini-2.5-pro","name":"Google: Gemini 2.5 Pro","pricing":{"prompt":"0.00000125","completion":"0.00001","request":"0","image":"0.00516","web_search":"0","internal_reasoning":"0","input_cache_read":"0.00000031","input_cache_write":"0.000001625"},"created":1750169544,"top_provider":{"context_length":1048576,"max_completion_tokens":65536,"is_moderated":false}},{"id":"moonshotai/kimi-dev-72b:free","name":"Kimi Dev 72b (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1750115909,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"openai/o3-pro","name":"OpenAI: o3 Pro","pricing":{"prompt":"0.00002","completion":"0.00008","request":"0","image":"0.0153","web_search":"0","internal_reasoning":"0"},"created":1749598352,"top_provider":{"context_length":200000,"max_completion_tokens":100000,"is_moderated":true}},{"id":"x-ai/grok-3-mini","name":"xAI: Grok 3 Mini","pricing":{"prompt":"0.0000003","completion":"0.0000005","request":"0","image":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.000000075"},"created":1749583245,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"x-ai/grok-3","name":"xAI: Grok 3","pricing":{"prompt":"0.000003","completion":"0.000015","request":"0","image":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.00000075"},"created":1749582908,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"mistralai/magistral-small-2506","name":"Mistral: Magistral Small 2506","pricing":{"prompt":"0.0000005","completion":"0.0000015","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1749569561,"top_provider":{"context_length":40000,"max_completion_tokens":40000,"is_moderated":false}},{"id":"mistralai/magistral-medium-2506","name":"Mistral: Magistral Medium 2506","pricing":{"prompt":"0.000002","completion":"0.000005","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1749354054,"top_provider":{"context_length":40960,"max_completion_tokens":40000,"is_moderated":false}},{"id":"mistralai/magistral-medium-2506:thinking","name":"Mistral: Magistral Medium 2506 (thinking)","pricing":{"prompt":"0.000002","completion":"0.000005","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1749354054,"top_provider":{"context_length":40960,"max_completion_tokens":40000,"is_moderated":false}},{"id":"google/gemini-2.5-pro-preview","name":"Google: Gemini 2.5 Pro Preview 06-05","pricing":{"prompt":"0.00000125","completion":"0.00001","request":"0","image":"0.00516","web_search":"0","internal_reasoning":"0","input_cache_read":"0.00000031","input_cache_write":"0.000001625"},"created":1749137257,"top_provider":{"context_length":1048576,"max_completion_tokens":65536,"is_moderated":false}},{"id":"deepseek/deepseek-r1-distill-qwen-7b","name":"DeepSeek: R1 Distill Qwen 7B","pricing":{"prompt":"0.0000001","completion":"0.0000002","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1748628237,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"deepseek/deepseek-r1-0528-qwen3-8b:free","name":"DeepSeek: Deepseek R1 0528 Qwen3 8B (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1748538543,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"deepseek/deepseek-r1-0528-qwen3-8b","name":"DeepSeek: Deepseek R1 0528 Qwen3 8B","pricing":{"prompt":"0.00000001","completion":"0.00000002","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1748538543,"top_provider":{"context_length":32000,"max_completion_tokens":null,"is_moderated":false}},{"id":"deepseek/deepseek-r1-0528:free","name":"DeepSeek: R1 0528 (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1748455170,"top_provider":{"context_length":163840,"max_completion_tokens":null,"is_moderated":false}},{"id":"deepseek/deepseek-r1-0528","name":"DeepSeek: R1 0528","pricing":{"prompt":"0.000000272","completion":"0.000000272","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1748455170,"top_provider":{"context_length":163840,"max_completion_tokens":null,"is_moderated":false}},{"id":"sarvamai/sarvam-m:free","name":"Sarvam AI: Sarvam-M (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1748188413,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"sarvamai/sarvam-m","name":"Sarvam AI: Sarvam-M","pricing":{"prompt":"0.000000022","completion":"0.000000022","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1748188413,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"thedrummer/valkyrie-49b-v1","name":"TheDrummer: Valkyrie 49B V1","pricing":{"prompt":"0.00000065","completion":"0.000001","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1748022670,"top_provider":{"context_length":131072,"max_completion_tokens":131072,"is_moderated":false}},{"id":"anthropic/claude-opus-4","name":"Anthropic: Claude Opus 4","pricing":{"prompt":"0.000015","completion":"0.000075","request":"0","image":"0.024","web_search":"0","internal_reasoning":"0","input_cache_read":"0.0000015","input_cache_write":"0.00001875"},"created":1747931245,"top_provider":{"context_length":200000,"max_completion_tokens":32000,"is_moderated":true}},{"id":"anthropic/claude-sonnet-4","name":"Anthropic: Claude Sonnet 4","pricing":{"prompt":"0.000003","completion":"0.000015","request":"0","image":"0.0048","web_search":"0","internal_reasoning":"0","input_cache_read":"0.0000003","input_cache_write":"0.00000375"},"created":1747930371,"top_provider":{"context_length":200000,"max_completion_tokens":64000,"is_moderated":false}},{"id":"mistralai/devstral-small-2505:free","name":"Mistral: Devstral Small 2505 (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1747837379,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"mistralai/devstral-small-2505","name":"Mistral: Devstral Small 2505","pricing":{"prompt":"0.00000003","completion":"0.00000003","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1747837379,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"google/gemma-3n-e4b-it:free","name":"Google: Gemma 3n 4B (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1747776824,"top_provider":{"context_length":8192,"max_completion_tokens":2048,"is_moderated":false}},{"id":"google/gemma-3n-e4b-it","name":"Google: Gemma 3n 4B","pricing":{"prompt":"0.00000002","completion":"0.00000004","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1747776824,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"openai/codex-mini","name":"OpenAI: Codex Mini","pricing":{"prompt":"0.0000015","completion":"0.000006","request":"0","image":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.000000375"},"created":1747409761,"top_provider":{"context_length":200000,"max_completion_tokens":100000,"is_moderated":true}},{"id":"nousresearch/deephermes-3-mistral-24b-preview","name":"Nous: DeepHermes 3 Mistral 24B Preview","pricing":{"prompt":"0.000000141","completion":"0.000000141","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1746830904,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"mistralai/mistral-medium-3","name":"Mistral: Mistral Medium 3","pricing":{"prompt":"0.0000004","completion":"0.000002","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1746627341,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"google/gemini-2.5-pro-preview-05-06","name":"Google: Gemini 2.5 Pro Preview 05-06","pricing":{"prompt":"0.00000125","completion":"0.00001","request":"0","image":"0.00516","web_search":"0","internal_reasoning":"0","input_cache_read":"0.00000031","input_cache_write":"0.000001625"},"created":1746578513,"top_provider":{"context_length":1048576,"max_completion_tokens":65535,"is_moderated":false}},{"id":"arcee-ai/spotlight","name":"Arcee AI: Spotlight","pricing":{"prompt":"0.00000018","completion":"0.00000018","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1746481552,"top_provider":{"context_length":131072,"max_completion_tokens":65537,"is_moderated":false}},{"id":"arcee-ai/maestro-reasoning","name":"Arcee AI: Maestro Reasoning","pricing":{"prompt":"0.0000009","completion":"0.0000033","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1746481269,"top_provider":{"context_length":131072,"max_completion_tokens":32000,"is_moderated":false}},{"id":"arcee-ai/virtuoso-large","name":"Arcee AI: Virtuoso Large","pricing":{"prompt":"0.00000075","completion":"0.0000012","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1746478885,"top_provider":{"context_length":131072,"max_completion_tokens":64000,"is_moderated":false}},{"id":"arcee-ai/coder-large","name":"Arcee AI: Coder Large","pricing":{"prompt":"0.0000005","completion":"0.0000008","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1746478663,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"microsoft/phi-4-reasoning-plus","name":"Microsoft: Phi 4 Reasoning Plus","pricing":{"prompt":"0.00000007","completion":"0.00000035","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1746130961,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"inception/mercury-coder","name":"Inception: Mercury Coder","pricing":{"prompt":"0.00000025","completion":"0.000001","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1746033880,"top_provider":{"context_length":32000,"max_completion_tokens":null,"is_moderated":false}},{"id":"qwen/qwen3-4b:free","name":"Qwen: Qwen3 4B (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1746031104,"top_provider":{"context_length":40960,"max_completion_tokens":null,"is_moderated":false}},{"id":"opengvlab/internvl3-14b","name":"OpenGVLab: InternVL3 14B","pricing":{"prompt":"0.0000002","completion":"0.0000004","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1746021355,"top_provider":{"context_length":12288,"max_completion_tokens":null,"is_moderated":false}},{"id":"deepseek/deepseek-prover-v2","name":"DeepSeek: DeepSeek Prover V2","pricing":{"prompt":"0.0000005","completion":"0.00000218","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1746013094,"top_provider":{"context_length":163840,"max_completion_tokens":null,"is_moderated":false}},{"id":"meta-llama/llama-guard-4-12b","name":"Meta: Llama Guard 4 12B","pricing":{"prompt":"0.00000005","completion":"0.00000005","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1745975193,"top_provider":{"context_length":163840,"max_completion_tokens":null,"is_moderated":false}},{"id":"qwen/qwen3-30b-a3b:free","name":"Qwen: Qwen3 30B A3B (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1745878604,"top_provider":{"context_length":40960,"max_completion_tokens":null,"is_moderated":false}},{"id":"qwen/qwen3-30b-a3b","name":"Qwen: Qwen3 30B A3B","pricing":{"prompt":"0.00000008","completion":"0.00000029","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1745878604,"top_provider":{"context_length":40960,"max_completion_tokens":40960,"is_moderated":false}},{"id":"qwen/qwen3-8b:free","name":"Qwen: Qwen3 8B (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1745876632,"top_provider":{"context_length":40960,"max_completion_tokens":40960,"is_moderated":false}},{"id":"qwen/qwen3-8b","name":"Qwen: Qwen3 8B","pricing":{"prompt":"0.000000035","completion":"0.000000138","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1745876632,"top_provider":{"context_length":128000,"max_completion_tokens":20000,"is_moderated":false}},{"id":"qwen/qwen3-14b:free","name":"Qwen: Qwen3 14B (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1745876478,"top_provider":{"context_length":40960,"max_completion_tokens":null,"is_moderated":false}},{"id":"qwen/qwen3-14b","name":"Qwen: Qwen3 14B","pricing":{"prompt":"0.00000006","completion":"0.00000024","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1745876478,"top_provider":{"context_length":40960,"max_completion_tokens":null,"is_moderated":false}},{"id":"qwen/qwen3-32b","name":"Qwen: Qwen3 32B","pricing":{"prompt":"0.000000027","completion":"0.000000027","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1745875945,"top_provider":{"context_length":40960,"max_completion_tokens":null,"is_moderated":false}},{"id":"qwen/qwen3-235b-a22b:free","name":"Qwen: Qwen3 235B A22B (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1745875757,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"qwen/qwen3-235b-a22b","name":"Qwen: Qwen3 235B A22B","pricing":{"prompt":"0.00000013","completion":"0.0000006","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1745875757,"top_provider":{"context_length":40960,"max_completion_tokens":40960,"is_moderated":false}},{"id":"tngtech/deepseek-r1t-chimera:free","name":"TNG: DeepSeek R1T Chimera (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1745760875,"top_provider":{"context_length":163840,"max_completion_tokens":null,"is_moderated":false}},{"id":"microsoft/mai-ds-r1:free","name":"Microsoft: MAI DS R1 (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1745194100,"top_provider":{"context_length":163840,"max_completion_tokens":null,"is_moderated":false}},{"id":"microsoft/mai-ds-r1","name":"Microsoft: MAI DS R1","pricing":{"prompt":"0.000000302","completion":"0.000000302","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1745194100,"top_provider":{"context_length":163840,"max_completion_tokens":null,"is_moderated":false}},{"id":"thudm/glm-z1-32b:free","name":"THUDM: GLM Z1 32B (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1744924148,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"thudm/glm-z1-32b","name":"THUDM: GLM Z1 32B","pricing":{"prompt":"0.00000003","completion":"0.00000003","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1744924148,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"thudm/glm-4-32b:free","name":"THUDM: GLM 4 32B (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1744920915,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"thudm/glm-4-32b","name":"THUDM: GLM 4 32B","pricing":{"prompt":"0.00000024","completion":"0.00000024","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1744920915,"top_provider":{"context_length":32000,"max_completion_tokens":32000,"is_moderated":false}},{"id":"openai/o4-mini-high","name":"OpenAI: o4 Mini High","pricing":{"prompt":"0.0000011","completion":"0.0000044","request":"0","image":"0.0008415","web_search":"0","internal_reasoning":"0","input_cache_read":"0.000000275"},"created":1744824212,"top_provider":{"context_length":200000,"max_completion_tokens":100000,"is_moderated":true}},{"id":"openai/o3","name":"OpenAI: o3","pricing":{"prompt":"0.000002","completion":"0.000008","request":"0","image":"0.00153","web_search":"0","internal_reasoning":"0","input_cache_read":"0.0000005"},"created":1744823457,"top_provider":{"context_length":200000,"max_completion_tokens":100000,"is_moderated":true}},{"id":"openai/o4-mini","name":"OpenAI: o4 Mini","pricing":{"prompt":"0.0000011","completion":"0.0000044","request":"0","image":"0.0008415","web_search":"0","internal_reasoning":"0","input_cache_read":"0.000000275"},"created":1744820942,"top_provider":{"context_length":200000,"max_completion_tokens":100000,"is_moderated":true}},{"id":"shisa-ai/shisa-v2-llama3.3-70b:free","name":"Shisa AI: Shisa V2 Llama 3.3 70B (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1744754858,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"shisa-ai/shisa-v2-llama3.3-70b","name":"Shisa AI: Shisa V2 Llama 3.3 70B ","pricing":{"prompt":"0.00000003","completion":"0.00000003","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1744754858,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"openai/gpt-4.1","name":"OpenAI: GPT-4.1","pricing":{"prompt":"0.000002","completion":"0.000008","request":"0","image":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.0000005"},"created":1744651385,"top_provider":{"context_length":1047576,"max_completion_tokens":32768,"is_moderated":true}},{"id":"openai/gpt-4.1-mini","name":"OpenAI: GPT-4.1 Mini","pricing":{"prompt":"0.0000004","completion":"0.0000016","request":"0","image":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.0000001"},"created":1744651381,"top_provider":{"context_length":1047576,"max_completion_tokens":32768,"is_moderated":true}},{"id":"openai/gpt-4.1-nano","name":"OpenAI: GPT-4.1 Nano","pricing":{"prompt":"0.0000001","completion":"0.0000004","request":"0","image":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.000000025"},"created":1744651369,"top_provider":{"context_length":1047576,"max_completion_tokens":32768,"is_moderated":true}},{"id":"eleutherai/llemma_7b","name":"EleutherAI: Llemma 7b","pricing":{"prompt":"0.0000008","completion":"0.0000012","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1744643225,"top_provider":{"context_length":4096,"max_completion_tokens":4096,"is_moderated":false}},{"id":"alfredpros/codellama-7b-instruct-solidity","name":"AlfredPros: CodeLLaMa 7B Instruct Solidity","pricing":{"prompt":"0.0000008","completion":"0.0000012","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1744641874,"top_provider":{"context_length":4096,"max_completion_tokens":4096,"is_moderated":false}},{"id":"arliai/qwq-32b-arliai-rpr-v1:free","name":"ArliAI: QwQ 32B RpR v1 (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1744555982,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"arliai/qwq-32b-arliai-rpr-v1","name":"ArliAI: QwQ 32B RpR v1","pricing":{"prompt":"0.000000015","completion":"0.000000015","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1744555982,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"agentica-org/deepcoder-14b-preview:free","name":"Agentica: Deepcoder 14B Preview (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1744555395,"top_provider":{"context_length":96000,"max_completion_tokens":null,"is_moderated":false}},{"id":"agentica-org/deepcoder-14b-preview","name":"Agentica: Deepcoder 14B Preview","pricing":{"prompt":"0.000000015","completion":"0.000000015","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1744555395,"top_provider":{"context_length":96000,"max_completion_tokens":null,"is_moderated":false}},{"id":"moonshotai/kimi-vl-a3b-thinking:free","name":"Moonshot AI: Kimi VL A3B Thinking (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1744304841,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"moonshotai/kimi-vl-a3b-thinking","name":"Moonshot AI: Kimi VL A3B Thinking","pricing":{"prompt":"0.000000038","completion":"0.000000038","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1744304841,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"x-ai/grok-3-mini-beta","name":"xAI: Grok 3 Mini Beta","pricing":{"prompt":"0.0000003","completion":"0.0000005","request":"0","image":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.000000075"},"created":1744240195,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"x-ai/grok-3-beta","name":"xAI: Grok 3 Beta","pricing":{"prompt":"0.000003","completion":"0.000015","request":"0","image":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.00000075"},"created":1744240068,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"nvidia/llama-3.3-nemotron-super-49b-v1","name":"NVIDIA: Llama 3.3 Nemotron Super 49B v1","pricing":{"prompt":"0.00000013","completion":"0.0000004","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1744119494,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"nvidia/llama-3.1-nemotron-ultra-253b-v1:free","name":"NVIDIA: Llama 3.1 Nemotron Ultra 253B v1 (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1744115059,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"nvidia/llama-3.1-nemotron-ultra-253b-v1","name":"NVIDIA: Llama 3.1 Nemotron Ultra 253B v1","pricing":{"prompt":"0.0000006","completion":"0.0000018","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1744115059,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"meta-llama/llama-4-maverick","name":"Meta: Llama 4 Maverick","pricing":{"prompt":"0.00000015","completion":"0.0000006","request":"0","image":"0.0006684","web_search":"0","internal_reasoning":"0"},"created":1743881822,"top_provider":{"context_length":1048576,"max_completion_tokens":16384,"is_moderated":false}},{"id":"meta-llama/llama-4-scout","name":"Meta: Llama 4 Scout","pricing":{"prompt":"0.00000008","completion":"0.0000003","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1743881519,"top_provider":{"context_length":1048576,"max_completion_tokens":1048576,"is_moderated":false}},{"id":"deepseek/deepseek-v3-base","name":"DeepSeek: DeepSeek V3 Base","pricing":{"prompt":"0.000000302","completion":"0.000000302","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1743272023,"top_provider":{"context_length":163840,"max_completion_tokens":null,"is_moderated":false}},{"id":"scb10x/llama3.1-typhoon2-70b-instruct","name":"Typhoon2 70B Instruct","pricing":{"prompt":"0.00000088","completion":"0.00000088","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1743196170,"top_provider":{"context_length":8192,"max_completion_tokens":null,"is_moderated":false}},{"id":"google/gemini-2.5-pro-exp-03-25","name":"Google: Gemini 2.5 Pro Experimental","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1742922099,"top_provider":{"context_length":1048576,"max_completion_tokens":65535,"is_moderated":false}},{"id":"qwen/qwen2.5-vl-32b-instruct:free","name":"Qwen: Qwen2.5 VL 32B Instruct (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1742839838,"top_provider":{"context_length":8192,"max_completion_tokens":null,"is_moderated":false}},{"id":"qwen/qwen2.5-vl-32b-instruct","name":"Qwen: Qwen2.5 VL 32B Instruct","pricing":{"prompt":"0.0000002","completion":"0.0000006","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1742839838,"top_provider":{"context_length":128000,"max_completion_tokens":null,"is_moderated":false}},{"id":"deepseek/deepseek-chat-v3-0324:free","name":"DeepSeek: DeepSeek V3 0324 (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1742824755,"top_provider":{"context_length":32768,"max_completion_tokens":16384,"is_moderated":false}},{"id":"deepseek/deepseek-chat-v3-0324","name":"DeepSeek: DeepSeek V3 0324","pricing":{"prompt":"0.00000025","completion":"0.00000085","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1742824755,"top_provider":{"context_length":163840,"max_completion_tokens":163840,"is_moderated":false}},{"id":"featherless/qwerky-72b:free","name":"Qrwkv 72B (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1742481597,"top_provider":{"context_length":32768,"max_completion_tokens":4096,"is_moderated":false}},{"id":"openai/o1-pro","name":"OpenAI: o1-pro","pricing":{"prompt":"0.00015","completion":"0.0006","request":"0","image":"0.21675","web_search":"0","internal_reasoning":"0"},"created":1742423211,"top_provider":{"context_length":200000,"max_completion_tokens":100000,"is_moderated":true}},{"id":"mistralai/mistral-small-3.1-24b-instruct:free","name":"Mistral: Mistral Small 3.1 24B (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1742238937,"top_provider":{"context_length":128000,"max_completion_tokens":null,"is_moderated":false}},{"id":"mistralai/mistral-small-3.1-24b-instruct","name":"Mistral: Mistral Small 3.1 24B","pricing":{"prompt":"0.000000027","completion":"0.000000027","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1742238937,"top_provider":{"context_length":96000,"max_completion_tokens":96000,"is_moderated":false}},{"id":"google/gemma-3-4b-it:free","name":"Google: Gemma 3 4B (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1741905510,"top_provider":{"context_length":32768,"max_completion_tokens":8192,"is_moderated":false}},{"id":"google/gemma-3-4b-it","name":"Google: Gemma 3 4B","pricing":{"prompt":"0.00000002","completion":"0.00000004","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1741905510,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"ai21/jamba-1.6-large","name":"AI21: Jamba 1.6 Large","pricing":{"prompt":"0.000002","completion":"0.000008","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1741905173,"top_provider":{"context_length":256000,"max_completion_tokens":4096,"is_moderated":false}},{"id":"ai21/jamba-1.6-mini","name":"AI21: Jamba Mini 1.6","pricing":{"prompt":"0.0000002","completion":"0.0000004","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1741905171,"top_provider":{"context_length":256000,"max_completion_tokens":4096,"is_moderated":false}},{"id":"google/gemma-3-12b-it:free","name":"Google: Gemma 3 12B (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1741902625,"top_provider":{"context_length":96000,"max_completion_tokens":8192,"is_moderated":false}},{"id":"google/gemma-3-12b-it","name":"Google: Gemma 3 12B","pricing":{"prompt":"0.00000003","completion":"0.00000003","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1741902625,"top_provider":{"context_length":96000,"max_completion_tokens":8192,"is_moderated":false}},{"id":"cohere/command-a","name":"Cohere: Command A","pricing":{"prompt":"0.000002","completion":"0.000008","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1741894342,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"openai/gpt-4o-mini-search-preview","name":"OpenAI: GPT-4o-mini Search Preview","pricing":{"prompt":"0.00000015","completion":"0.0000006","request":"0.0275","image":"0.000217","web_search":"0","internal_reasoning":"0"},"created":1741818122,"top_provider":{"context_length":128000,"max_completion_tokens":16384,"is_moderated":true}},{"id":"openai/gpt-4o-search-preview","name":"OpenAI: GPT-4o Search Preview","pricing":{"prompt":"0.0000025","completion":"0.00001","request":"0.035","image":"0.003613","web_search":"0","internal_reasoning":"0"},"created":1741817949,"top_provider":{"context_length":128000,"max_completion_tokens":16384,"is_moderated":true}},{"id":"rekaai/reka-flash-3:free","name":"Reka: Flash 3 (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1741812813,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"rekaai/reka-flash-3","name":"Reka: Flash 3","pricing":{"prompt":"0.000000013","completion":"0.000000013","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1741812813,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"google/gemma-3-27b-it:free","name":"Google: Gemma 3 27B (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1741756359,"top_provider":{"context_length":96000,"max_completion_tokens":8192,"is_moderated":false}},{"id":"google/gemma-3-27b-it","name":"Google: Gemma 3 27B","pricing":{"prompt":"0.00000009","completion":"0.00000017","request":"0","image":"0.0000256","web_search":"0","internal_reasoning":"0"},"created":1741756359,"top_provider":{"context_length":131072,"max_completion_tokens":16384,"is_moderated":false}},{"id":"thedrummer/anubis-pro-105b-v1","name":"TheDrummer: Anubis Pro 105B V1","pricing":{"prompt":"0.0000005","completion":"0.000001","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1741642290,"top_provider":{"context_length":131072,"max_completion_tokens":131072,"is_moderated":false}},{"id":"thedrummer/skyfall-36b-v2","name":"TheDrummer: Skyfall 36B V2","pricing":{"prompt":"0.00000001512","completion":"0.00000001512","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1741636566,"top_provider":{"context_length":16384,"max_completion_tokens":null,"is_moderated":false}},{"id":"microsoft/phi-4-multimodal-instruct","name":"Microsoft: Phi 4 Multimodal Instruct","pricing":{"prompt":"0.00000005","completion":"0.0000001","request":"0","image":"0.00017685","web_search":"0","internal_reasoning":"0"},"created":1741396284,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"perplexity/sonar-reasoning-pro","name":"Perplexity: Sonar Reasoning Pro","pricing":{"prompt":"0.000002","completion":"0.000008","request":"0","image":"0","web_search":"0.005","internal_reasoning":"0"},"created":1741313308,"top_provider":{"context_length":128000,"max_completion_tokens":null,"is_moderated":false}},{"id":"perplexity/sonar-pro","name":"Perplexity: Sonar Pro","pricing":{"prompt":"0.000003","completion":"0.000015","request":"0","image":"0","web_search":"0.005","internal_reasoning":"0"},"created":1741312423,"top_provider":{"context_length":200000,"max_completion_tokens":8000,"is_moderated":false}},{"id":"perplexity/sonar-deep-research","name":"Perplexity: Sonar Deep Research","pricing":{"prompt":"0.000002","completion":"0.000008","request":"0","image":"0","web_search":"0.005","internal_reasoning":"0.000003"},"created":1741311246,"top_provider":{"context_length":128000,"max_completion_tokens":null,"is_moderated":false}},{"id":"qwen/qwq-32b:free","name":"Qwen: QwQ 32B (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1741208814,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"qwen/qwq-32b","name":"Qwen: QwQ 32B","pricing":{"prompt":"0.000000075","completion":"0.00000015","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1741208814,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"nousresearch/deephermes-3-llama-3-8b-preview:free","name":"Nous: DeepHermes 3 Llama 3 8B Preview (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1740719372,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"google/gemini-2.0-flash-lite-001","name":"Google: Gemini 2.0 Flash Lite","pricing":{"prompt":"0.000000075","completion":"0.0000003","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1740506212,"top_provider":{"context_length":1048576,"max_completion_tokens":8192,"is_moderated":false}},{"id":"anthropic/claude-3.7-sonnet","name":"Anthropic: Claude 3.7 Sonnet","pricing":{"prompt":"0.000003","completion":"0.000015","request":"0","image":"0.0048","web_search":"0","internal_reasoning":"0","input_cache_read":"0.0000003","input_cache_write":"0.00000375"},"created":1740422110,"top_provider":{"context_length":200000,"max_completion_tokens":64000,"is_moderated":false}},{"id":"anthropic/claude-3.7-sonnet:thinking","name":"Anthropic: Claude 3.7 Sonnet (thinking)","pricing":{"prompt":"0.000003","completion":"0.000015","request":"0","image":"0.0048","web_search":"0","internal_reasoning":"0","input_cache_read":"0.0000003","input_cache_write":"0.00000375"},"created":1740422110,"top_provider":{"context_length":200000,"max_completion_tokens":64000,"is_moderated":false}},{"id":"anthropic/claude-3.7-sonnet:beta","name":"Anthropic: Claude 3.7 Sonnet (self-moderated)","pricing":{"prompt":"0.000003","completion":"0.000015","request":"0","image":"0.0048","web_search":"0","internal_reasoning":"0","input_cache_read":"0.0000003","input_cache_write":"0.00000375"},"created":1740422110,"top_provider":{"context_length":200000,"max_completion_tokens":128000,"is_moderated":false}},{"id":"perplexity/r1-1776","name":"Perplexity: R1 1776","pricing":{"prompt":"0.000002","completion":"0.000008","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1740004929,"top_provider":{"context_length":128000,"max_completion_tokens":null,"is_moderated":false}},{"id":"mistralai/mistral-saba","name":"Mistral: Saba","pricing":{"prompt":"0.0000002","completion":"0.0000006","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1739803239,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"cognitivecomputations/dolphin3.0-r1-mistral-24b:free","name":"Dolphin3.0 R1 Mistral 24B (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1739462498,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"cognitivecomputations/dolphin3.0-r1-mistral-24b","name":"Dolphin3.0 R1 Mistral 24B","pricing":{"prompt":"0.000000013","completion":"0.000000013","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1739462498,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"cognitivecomputations/dolphin3.0-mistral-24b:free","name":"Dolphin3.0 Mistral 24B (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1739462019,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"meta-llama/llama-guard-3-8b","name":"Llama Guard 3 8B","pricing":{"prompt":"0.00000002","completion":"0.00000006","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1739401318,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"openai/o3-mini-high","name":"OpenAI: o3 Mini High","pricing":{"prompt":"0.0000011","completion":"0.0000044","request":"0","image":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.00000055"},"created":1739372611,"top_provider":{"context_length":200000,"max_completion_tokens":100000,"is_moderated":true}},{"id":"deepseek/deepseek-r1-distill-llama-8b","name":"DeepSeek: R1 Distill Llama 8B","pricing":{"prompt":"0.00000004","completion":"0.00000004","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1738937718,"top_provider":{"context_length":32000,"max_completion_tokens":32000,"is_moderated":false}},{"id":"google/gemini-2.0-flash-001","name":"Google: Gemini 2.0 Flash","pricing":{"prompt":"0.0000001","completion":"0.0000004","request":"0","image":"0.0000258","web_search":"0","internal_reasoning":"0","input_cache_read":"0.000000025","input_cache_write":"0.0000001833"},"created":1738769413,"top_provider":{"context_length":1048576,"max_completion_tokens":8192,"is_moderated":false}},{"id":"qwen/qwen-vl-plus","name":"Qwen: Qwen VL Plus","pricing":{"prompt":"0.00000021","completion":"0.00000063","request":"0","image":"0.0002688","web_search":"0","internal_reasoning":"0"},"created":1738731255,"top_provider":{"context_length":7500,"max_completion_tokens":1500,"is_moderated":false}},{"id":"aion-labs/aion-1.0","name":"AionLabs: Aion-1.0","pricing":{"prompt":"0.000004","completion":"0.000008","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1738697557,"top_provider":{"context_length":131072,"max_completion_tokens":32768,"is_moderated":false}},{"id":"aion-labs/aion-1.0-mini","name":"AionLabs: Aion-1.0-Mini","pricing":{"prompt":"0.0000007","completion":"0.0000014","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1738697107,"top_provider":{"context_length":131072,"max_completion_tokens":32768,"is_moderated":false}},{"id":"aion-labs/aion-rp-llama-3.1-8b","name":"AionLabs: Aion-RP 1.0 (8B)","pricing":{"prompt":"0.0000002","completion":"0.0000002","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1738696718,"top_provider":{"context_length":32768,"max_completion_tokens":32768,"is_moderated":false}},{"id":"qwen/qwen-vl-max","name":"Qwen: Qwen VL Max","pricing":{"prompt":"0.0000008","completion":"0.0000032","request":"0","image":"0.001024","web_search":"0","internal_reasoning":"0"},"created":1738434304,"top_provider":{"context_length":7500,"max_completion_tokens":1500,"is_moderated":false}},{"id":"qwen/qwen-turbo","name":"Qwen: Qwen-Turbo","pricing":{"prompt":"0.00000005","completion":"0.0000002","request":"0","image":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.00000002"},"created":1738410974,"top_provider":{"context_length":1000000,"max_completion_tokens":8192,"is_moderated":false}},{"id":"qwen/qwen2.5-vl-72b-instruct:free","name":"Qwen: Qwen2.5 VL 72B Instruct (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1738410311,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"qwen/qwen2.5-vl-72b-instruct","name":"Qwen: Qwen2.5 VL 72B Instruct","pricing":{"prompt":"0.00000025","completion":"0.00000075","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1738410311,"top_provider":{"context_length":32000,"max_completion_tokens":null,"is_moderated":false}},{"id":"qwen/qwen-plus","name":"Qwen: Qwen-Plus","pricing":{"prompt":"0.0000004","completion":"0.0000012","request":"0","image":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.00000016"},"created":1738409840,"top_provider":{"context_length":131072,"max_completion_tokens":8192,"is_moderated":false}},{"id":"qwen/qwen-max","name":"Qwen: Qwen-Max ","pricing":{"prompt":"0.0000016","completion":"0.0000064","request":"0","image":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.00000064"},"created":1738402289,"top_provider":{"context_length":32768,"max_completion_tokens":8192,"is_moderated":false}},{"id":"openai/o3-mini","name":"OpenAI: o3 Mini","pricing":{"prompt":"0.0000011","completion":"0.0000044","request":"0","image":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.00000055"},"created":1738351721,"top_provider":{"context_length":200000,"max_completion_tokens":100000,"is_moderated":true}},{"id":"deepseek/deepseek-r1-distill-qwen-1.5b","name":"DeepSeek: R1 Distill Qwen 1.5B","pricing":{"prompt":"0.00000018","completion":"0.00000018","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1738328067,"top_provider":{"context_length":131072,"max_completion_tokens":32768,"is_moderated":false}},{"id":"mistralai/mistral-small-24b-instruct-2501:free","name":"Mistral: Mistral Small 3 (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1738255409,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"mistralai/mistral-small-24b-instruct-2501","name":"Mistral: Mistral Small 3","pricing":{"prompt":"0.00000003","completion":"0.00000003","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1738255409,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"deepseek/deepseek-r1-distill-qwen-32b","name":"DeepSeek: R1 Distill Qwen 32B","pricing":{"prompt":"0.000000075","completion":"0.00000015","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1738194830,"top_provider":{"context_length":131072,"max_completion_tokens":16384,"is_moderated":false}},{"id":"deepseek/deepseek-r1-distill-qwen-14b:free","name":"DeepSeek: R1 Distill Qwen 14B (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1738193940,"top_provider":{"context_length":64000,"max_completion_tokens":null,"is_moderated":false}},{"id":"deepseek/deepseek-r1-distill-qwen-14b","name":"DeepSeek: R1 Distill Qwen 14B","pricing":{"prompt":"0.00000015","completion":"0.00000015","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1738193940,"top_provider":{"context_length":64000,"max_completion_tokens":32000,"is_moderated":false}},{"id":"perplexity/sonar-reasoning","name":"Perplexity: Sonar Reasoning","pricing":{"prompt":"0.000001","completion":"0.000005","request":"0.005","image":"0","web_search":"0","internal_reasoning":"0"},"created":1738131107,"top_provider":{"context_length":127000,"max_completion_tokens":null,"is_moderated":false}},{"id":"perplexity/sonar","name":"Perplexity: Sonar","pricing":{"prompt":"0.000001","completion":"0.000001","request":"0.005","image":"0","web_search":"0","internal_reasoning":"0"},"created":1738013808,"top_provider":{"context_length":127072,"max_completion_tokens":null,"is_moderated":false}},{"id":"liquid/lfm-7b","name":"Liquid: LFM 7B","pricing":{"prompt":"0.00000001","completion":"0.00000001","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1737806883,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"liquid/lfm-3b","name":"Liquid: LFM 3B","pricing":{"prompt":"0.00000002","completion":"0.00000002","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1737806501,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"deepseek/deepseek-r1-distill-llama-70b:free","name":"DeepSeek: R1 Distill Llama 70B (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1737663169,"top_provider":{"context_length":8192,"max_completion_tokens":4096,"is_moderated":false}},{"id":"deepseek/deepseek-r1-distill-llama-70b","name":"DeepSeek: R1 Distill Llama 70B","pricing":{"prompt":"0.00000005","completion":"0.00000005","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1737663169,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"deepseek/deepseek-r1:free","name":"DeepSeek: R1 (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1737381095,"top_provider":{"context_length":163840,"max_completion_tokens":null,"is_moderated":false}},{"id":"deepseek/deepseek-r1","name":"DeepSeek: R1","pricing":{"prompt":"0.0000004","completion":"0.000002","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1737381095,"top_provider":{"context_length":163840,"max_completion_tokens":163840,"is_moderated":false}},{"id":"minimax/minimax-01","name":"MiniMax: MiniMax-01","pricing":{"prompt":"0.0000002","completion":"0.0000011","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1736915462,"top_provider":{"context_length":1000192,"max_completion_tokens":1000192,"is_moderated":false}},{"id":"mistralai/codestral-2501","name":"Mistral: Codestral 2501","pricing":{"prompt":"0.0000003","completion":"0.0000009","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1736895522,"top_provider":{"context_length":262144,"max_completion_tokens":null,"is_moderated":false}},{"id":"microsoft/phi-4","name":"Microsoft: Phi 4","pricing":{"prompt":"0.00000006","completion":"0.00000014","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1736489872,"top_provider":{"context_length":16384,"max_completion_tokens":null,"is_moderated":false}},{"id":"deepseek/deepseek-chat","name":"DeepSeek: DeepSeek V3","pricing":{"prompt":"0.000000272","completion":"0.000000272","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1735241320,"top_provider":{"context_length":163840,"max_completion_tokens":null,"is_moderated":false}},{"id":"sao10k/l3.3-euryale-70b","name":"Sao10K: Llama 3.3 Euryale 70B","pricing":{"prompt":"0.00000065","completion":"0.00000075","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1734535928,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"openai/o1","name":"OpenAI: o1","pricing":{"prompt":"0.000015","completion":"0.00006","request":"0","image":"0.021675","web_search":"0","internal_reasoning":"0","input_cache_read":"0.0000075"},"created":1734459999,"top_provider":{"context_length":200000,"max_completion_tokens":100000,"is_moderated":true}},{"id":"x-ai/grok-2-vision-1212","name":"xAI: Grok 2 Vision 1212","pricing":{"prompt":"0.000002","completion":"0.00001","request":"0","image":"0.0036","web_search":"0","internal_reasoning":"0"},"created":1734237338,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"x-ai/grok-2-1212","name":"xAI: Grok 2 1212","pricing":{"prompt":"0.000002","completion":"0.00001","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1734232814,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"cohere/command-r7b-12-2024","name":"Cohere: Command R7B (12-2024)","pricing":{"prompt":"0.0000000375","completion":"0.00000015","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1734158152,"top_provider":{"context_length":128000,"max_completion_tokens":4000,"is_moderated":true}},{"id":"google/gemini-2.0-flash-exp:free","name":"Google: Gemini 2.0 Flash Experimental (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1733937523,"top_provider":{"context_length":1048576,"max_completion_tokens":8192,"is_moderated":false}},{"id":"meta-llama/llama-3.3-70b-instruct:free","name":"Meta: Llama 3.3 70B Instruct (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1733506137,"top_provider":{"context_length":65536,"max_completion_tokens":null,"is_moderated":false}},{"id":"meta-llama/llama-3.3-70b-instruct","name":"Meta: Llama 3.3 70B Instruct","pricing":{"prompt":"0.000000038","completion":"0.00000012","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1733506137,"top_provider":{"context_length":131072,"max_completion_tokens":16384,"is_moderated":false}},{"id":"amazon/nova-lite-v1","name":"Amazon: Nova Lite 1.0","pricing":{"prompt":"0.00000006","completion":"0.00000024","request":"0","image":"0.00009","web_search":"0","internal_reasoning":"0"},"created":1733437363,"top_provider":{"context_length":300000,"max_completion_tokens":5120,"is_moderated":true}},{"id":"amazon/nova-micro-v1","name":"Amazon: Nova Micro 1.0","pricing":{"prompt":"0.000000035","completion":"0.00000014","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1733437237,"top_provider":{"context_length":128000,"max_completion_tokens":5120,"is_moderated":true}},{"id":"amazon/nova-pro-v1","name":"Amazon: Nova Pro 1.0","pricing":{"prompt":"0.0000008","completion":"0.0000032","request":"0","image":"0.0012","web_search":"0","internal_reasoning":"0"},"created":1733436303,"top_provider":{"context_length":300000,"max_completion_tokens":5120,"is_moderated":true}},{"id":"qwen/qwq-32b-preview","name":"Qwen: QwQ 32B Preview","pricing":{"prompt":"0.0000002","completion":"0.0000002","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1732754541,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"eva-unit-01/eva-qwen-2.5-72b","name":"EVA Qwen2.5 72B","pricing":{"prompt":"0.000004","completion":"0.000006","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1732210606,"top_provider":{"context_length":16384,"max_completion_tokens":4096,"is_moderated":false}},{"id":"openai/gpt-4o-2024-11-20","name":"OpenAI: GPT-4o (2024-11-20)","pricing":{"prompt":"0.0000025","completion":"0.00001","request":"0","image":"0.003613","web_search":"0","internal_reasoning":"0","input_cache_read":"0.00000125"},"created":1732127594,"top_provider":{"context_length":128000,"max_completion_tokens":16384,"is_moderated":true}},{"id":"mistralai/mistral-large-2411","name":"Mistral Large 2411","pricing":{"prompt":"0.000002","completion":"0.000006","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1731978685,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"mistralai/mistral-large-2407","name":"Mistral Large 2407","pricing":{"prompt":"0.000002","completion":"0.000006","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1731978415,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"mistralai/pixtral-large-2411","name":"Mistral: Pixtral Large 2411","pricing":{"prompt":"0.000002","completion":"0.000006","request":"0","image":"0.002888","web_search":"0","internal_reasoning":"0"},"created":1731977388,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"x-ai/grok-vision-beta","name":"xAI: Grok Vision Beta","pricing":{"prompt":"0.000005","completion":"0.000015","request":"0","image":"0.009","web_search":"0","internal_reasoning":"0"},"created":1731976624,"top_provider":{"context_length":8192,"max_completion_tokens":null,"is_moderated":false}},{"id":"infermatic/mn-inferor-12b","name":"Infermatic: Mistral Nemo Inferor 12B","pricing":{"prompt":"0.0000008","completion":"0.0000012","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1731464428,"top_provider":{"context_length":16384,"max_completion_tokens":4096,"is_moderated":false}},{"id":"qwen/qwen-2.5-coder-32b-instruct:free","name":"Qwen2.5 Coder 32B Instruct (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1731368400,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"qwen/qwen-2.5-coder-32b-instruct","name":"Qwen2.5 Coder 32B Instruct","pricing":{"prompt":"0.00000006","completion":"0.00000015","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1731368400,"top_provider":{"context_length":32768,"max_completion_tokens":16384,"is_moderated":false}},{"id":"raifle/sorcererlm-8x22b","name":"SorcererLM 8x22B","pricing":{"prompt":"0.0000045","completion":"0.0000045","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1731105083,"top_provider":{"context_length":16000,"max_completion_tokens":null,"is_moderated":false}},{"id":"thedrummer/unslopnemo-12b","name":"TheDrummer: UnslopNemo 12B","pricing":{"prompt":"0.0000004","completion":"0.0000004","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1731103448,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"anthropic/claude-3.5-haiku:beta","name":"Anthropic: Claude 3.5 Haiku (self-moderated)","pricing":{"prompt":"0.0000008","completion":"0.000004","request":"0","image":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.00000008","input_cache_write":"0.000001"},"created":1730678400,"top_provider":{"context_length":200000,"max_completion_tokens":8192,"is_moderated":false}},{"id":"anthropic/claude-3.5-haiku","name":"Anthropic: Claude 3.5 Haiku","pricing":{"prompt":"0.0000008","completion":"0.000004","request":"0","image":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.00000008","input_cache_write":"0.000001"},"created":1730678400,"top_provider":{"context_length":200000,"max_completion_tokens":8192,"is_moderated":true}},{"id":"anthropic/claude-3.5-haiku-20241022:beta","name":"Anthropic: Claude 3.5 Haiku (2024-10-22) (self-moderated)","pricing":{"prompt":"0.0000008","completion":"0.000004","request":"0","image":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.00000008","input_cache_write":"0.000001"},"created":1730678400,"top_provider":{"context_length":200000,"max_completion_tokens":8192,"is_moderated":false}},{"id":"anthropic/claude-3.5-haiku-20241022","name":"Anthropic: Claude 3.5 Haiku (2024-10-22)","pricing":{"prompt":"0.0000008","completion":"0.000004","request":"0","image":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.00000008","input_cache_write":"0.000001"},"created":1730678400,"top_provider":{"context_length":200000,"max_completion_tokens":8192,"is_moderated":true}},{"id":"anthracite-org/magnum-v4-72b","name":"Magnum v4 72B","pricing":{"prompt":"0.0000025","completion":"0.000003","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1729555200,"top_provider":{"context_length":16384,"max_completion_tokens":1024,"is_moderated":false}},{"id":"anthropic/claude-3.5-sonnet:beta","name":"Anthropic: Claude 3.5 Sonnet (self-moderated)","pricing":{"prompt":"0.000003","completion":"0.000015","request":"0","image":"0.0048","web_search":"0","internal_reasoning":"0","input_cache_read":"0.0000003","input_cache_write":"0.00000375"},"created":1729555200,"top_provider":{"context_length":200000,"max_completion_tokens":8192,"is_moderated":false}},{"id":"anthropic/claude-3.5-sonnet","name":"Anthropic: Claude 3.5 Sonnet","pricing":{"prompt":"0.000003","completion":"0.000015","request":"0","image":"0.0048","web_search":"0","internal_reasoning":"0","input_cache_read":"0.0000003","input_cache_write":"0.00000375"},"created":1729555200,"top_provider":{"context_length":200000,"max_completion_tokens":8192,"is_moderated":true}},{"id":"mistralai/ministral-8b","name":"Mistral: Ministral 8B","pricing":{"prompt":"0.0000001","completion":"0.0000001","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1729123200,"top_provider":{"context_length":128000,"max_completion_tokens":null,"is_moderated":false}},{"id":"mistralai/ministral-3b","name":"Mistral: Ministral 3B","pricing":{"prompt":"0.00000004","completion":"0.00000004","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1729123200,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"qwen/qwen-2.5-7b-instruct","name":"Qwen2.5 7B Instruct","pricing":{"prompt":"0.00000004","completion":"0.0000001","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1729036800,"top_provider":{"context_length":32768,"max_completion_tokens":16384,"is_moderated":false}},{"id":"nvidia/llama-3.1-nemotron-70b-instruct","name":"NVIDIA: Llama 3.1 Nemotron 70B Instruct","pricing":{"prompt":"0.00000012","completion":"0.0000003","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1728950400,"top_provider":{"context_length":131072,"max_completion_tokens":131072,"is_moderated":false}},{"id":"inflection/inflection-3-pi","name":"Inflection: Inflection 3 Pi","pricing":{"prompt":"0.0000025","completion":"0.00001","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1728604800,"top_provider":{"context_length":8000,"max_completion_tokens":1024,"is_moderated":false}},{"id":"inflection/inflection-3-productivity","name":"Inflection: Inflection 3 Productivity","pricing":{"prompt":"0.0000025","completion":"0.00001","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1728604800,"top_provider":{"context_length":8000,"max_completion_tokens":1024,"is_moderated":false}},{"id":"google/gemini-flash-1.5-8b","name":"Google: Gemini 1.5 Flash 8B","pricing":{"prompt":"0.0000000375","completion":"0.00000015","request":"0","image":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.00000001","input_cache_write":"0.0000000583"},"created":1727913600,"top_provider":{"context_length":1000000,"max_completion_tokens":8192,"is_moderated":false}},{"id":"anthracite-org/magnum-v2-72b","name":"Magnum v2 72B","pricing":{"prompt":"0.000003","completion":"0.000003","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1727654400,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"thedrummer/rocinante-12b","name":"TheDrummer: Rocinante 12B","pricing":{"prompt":"0.0000002","completion":"0.0000005","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1727654400,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"liquid/lfm-40b","name":"Liquid: LFM 40B MoE","pricing":{"prompt":"0.00000015","completion":"0.00000015","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1727654400,"top_provider":{"context_length":65536,"max_completion_tokens":65536,"is_moderated":false}},{"id":"meta-llama/llama-3.2-3b-instruct:free","name":"Meta: Llama 3.2 3B Instruct (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1727222400,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"meta-llama/llama-3.2-3b-instruct","name":"Meta: Llama 3.2 3B Instruct","pricing":{"prompt":"0.000000003","completion":"0.000000006","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1727222400,"top_provider":{"context_length":20000,"max_completion_tokens":20000,"is_moderated":false}},{"id":"meta-llama/llama-3.2-1b-instruct","name":"Meta: Llama 3.2 1B Instruct","pricing":{"prompt":"0.000000005","completion":"0.00000001","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1727222400,"top_provider":{"context_length":131072,"max_completion_tokens":16384,"is_moderated":false}},{"id":"meta-llama/llama-3.2-11b-vision-instruct:free","name":"Meta: Llama 3.2 11B Vision Instruct (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1727222400,"top_provider":{"context_length":131072,"max_completion_tokens":2048,"is_moderated":false}},{"id":"meta-llama/llama-3.2-11b-vision-instruct","name":"Meta: Llama 3.2 11B Vision Instruct","pricing":{"prompt":"0.000000049","completion":"0.000000049","request":"0","image":"0.00007948","web_search":"0","internal_reasoning":"0"},"created":1727222400,"top_provider":{"context_length":131072,"max_completion_tokens":16384,"is_moderated":false}},{"id":"meta-llama/llama-3.2-90b-vision-instruct","name":"Meta: Llama 3.2 90B Vision Instruct","pricing":{"prompt":"0.0000012","completion":"0.0000012","request":"0","image":"0.001734","web_search":"0","internal_reasoning":"0"},"created":1727222400,"top_provider":{"context_length":131072,"max_completion_tokens":2048,"is_moderated":false}},{"id":"qwen/qwen-2.5-72b-instruct:free","name":"Qwen2.5 72B Instruct (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1726704000,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"qwen/qwen-2.5-72b-instruct","name":"Qwen2.5 72B Instruct","pricing":{"prompt":"0.000000101","completion":"0.000000101","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1726704000,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"neversleep/llama-3.1-lumimaid-8b","name":"NeverSleep: Lumimaid v0.2 8B","pricing":{"prompt":"0.00000018","completion":"0.000001","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1726358400,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"openai/o1-preview-2024-09-12","name":"OpenAI: o1-preview (2024-09-12)","pricing":{"prompt":"0.000015","completion":"0.00006","request":"0","image":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.0000075"},"created":1726099200,"top_provider":{"context_length":128000,"max_completion_tokens":32768,"is_moderated":true}},{"id":"openai/o1-mini-2024-09-12","name":"OpenAI: o1-mini (2024-09-12)","pricing":{"prompt":"0.0000011","completion":"0.0000044","request":"0","image":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.00000055"},"created":1726099200,"top_provider":{"context_length":128000,"max_completion_tokens":65536,"is_moderated":true}},{"id":"openai/o1-preview","name":"OpenAI: o1-preview","pricing":{"prompt":"0.000015","completion":"0.00006","request":"0","image":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.0000075"},"created":1726099200,"top_provider":{"context_length":128000,"max_completion_tokens":32768,"is_moderated":true}},{"id":"openai/o1-mini","name":"OpenAI: o1-mini","pricing":{"prompt":"0.0000011","completion":"0.0000044","request":"0","image":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.00000055"},"created":1726099200,"top_provider":{"context_length":128000,"max_completion_tokens":65536,"is_moderated":true}},{"id":"mistralai/pixtral-12b","name":"Mistral: Pixtral 12B","pricing":{"prompt":"0.0000001","completion":"0.0000001","request":"0","image":"0.0001445","web_search":"0","internal_reasoning":"0"},"created":1725926400,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"cohere/command-r-08-2024","name":"Cohere: Command R (08-2024)","pricing":{"prompt":"0.00000015","completion":"0.0000006","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1724976000,"top_provider":{"context_length":128000,"max_completion_tokens":4000,"is_moderated":true}},{"id":"cohere/command-r-plus-08-2024","name":"Cohere: Command R+ (08-2024)","pricing":{"prompt":"0.0000025","completion":"0.00001","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1724976000,"top_provider":{"context_length":128000,"max_completion_tokens":4000,"is_moderated":true}},{"id":"qwen/qwen-2.5-vl-7b-instruct","name":"Qwen: Qwen2.5-VL 7B Instruct","pricing":{"prompt":"0.0000002","completion":"0.0000002","request":"0","image":"0.0001445","web_search":"0","internal_reasoning":"0"},"created":1724803200,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"sao10k/l3.1-euryale-70b","name":"Sao10K: Llama 3.1 Euryale 70B v2.2","pricing":{"prompt":"0.00000065","completion":"0.00000075","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1724803200,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"microsoft/phi-3.5-mini-128k-instruct","name":"Microsoft: Phi-3.5 Mini 128K Instruct","pricing":{"prompt":"0.0000001","completion":"0.0000001","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1724198400,"top_provider":{"context_length":128000,"max_completion_tokens":null,"is_moderated":false}},{"id":"nousresearch/hermes-3-llama-3.1-70b","name":"Nous: Hermes 3 70B Instruct","pricing":{"prompt":"0.0000001","completion":"0.00000028","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1723939200,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"nousresearch/hermes-3-llama-3.1-405b","name":"Nous: Hermes 3 405B Instruct","pricing":{"prompt":"0.0000007","completion":"0.0000008","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1723766400,"top_provider":{"context_length":131072,"max_completion_tokens":16384,"is_moderated":false}},{"id":"openai/chatgpt-4o-latest","name":"OpenAI: ChatGPT-4o","pricing":{"prompt":"0.000005","completion":"0.000015","request":"0","image":"0.007225","web_search":"0","internal_reasoning":"0"},"created":1723593600,"top_provider":{"context_length":128000,"max_completion_tokens":16384,"is_moderated":true}},{"id":"sao10k/l3-lunaris-8b","name":"Sao10K: Llama 3 8B Lunaris","pricing":{"prompt":"0.00000002","completion":"0.00000005","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1723507200,"top_provider":{"context_length":8192,"max_completion_tokens":null,"is_moderated":false}},{"id":"openai/gpt-4o-2024-08-06","name":"OpenAI: GPT-4o (2024-08-06)","pricing":{"prompt":"0.0000025","completion":"0.00001","request":"0","image":"0.003613","web_search":"0","internal_reasoning":"0","input_cache_read":"0.00000125"},"created":1722902400,"top_provider":{"context_length":128000,"max_completion_tokens":16384,"is_moderated":false}},{"id":"meta-llama/llama-3.1-405b","name":"Meta: Llama 3.1 405B (base)","pricing":{"prompt":"0.000002","completion":"0.000002","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1722556800,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"nothingiisreal/mn-celeste-12b","name":"Mistral Nemo 12B Celeste","pricing":{"prompt":"0.0000008","completion":"0.0000012","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1722556800,"top_provider":{"context_length":16384,"max_completion_tokens":4096,"is_moderated":false}},{"id":"meta-llama/llama-3.1-70b-instruct","name":"Meta: Llama 3.1 70B Instruct","pricing":{"prompt":"0.0000001","completion":"0.00000028","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1721692800,"top_provider":{"context_length":131072,"max_completion_tokens":16384,"is_moderated":false}},{"id":"meta-llama/llama-3.1-8b-instruct","name":"Meta: Llama 3.1 8B Instruct","pricing":{"prompt":"0.000000015","completion":"0.00000002","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1721692800,"top_provider":{"context_length":131072,"max_completion_tokens":16384,"is_moderated":false}},{"id":"meta-llama/llama-3.1-405b-instruct:free","name":"Meta: Llama 3.1 405B Instruct (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1721692800,"top_provider":{"context_length":65536,"max_completion_tokens":null,"is_moderated":false}},{"id":"meta-llama/llama-3.1-405b-instruct","name":"Meta: Llama 3.1 405B Instruct","pricing":{"prompt":"0.0000008","completion":"0.0000008","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1721692800,"top_provider":{"context_length":32768,"max_completion_tokens":16384,"is_moderated":false}},{"id":"mistralai/mistral-nemo:free","name":"Mistral: Mistral Nemo (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1721347200,"top_provider":{"context_length":131072,"max_completion_tokens":128000,"is_moderated":false}},{"id":"mistralai/mistral-nemo","name":"Mistral: Mistral Nemo","pricing":{"prompt":"0.0000000075","completion":"0.00000005","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1721347200,"top_provider":{"context_length":32000,"max_completion_tokens":null,"is_moderated":false}},{"id":"openai/gpt-4o-mini","name":"OpenAI: GPT-4o-mini","pricing":{"prompt":"0.00000015","completion":"0.0000006","request":"0","image":"0.000217","web_search":"0","internal_reasoning":"0","input_cache_read":"0.000000075"},"created":1721260800,"top_provider":{"context_length":128000,"max_completion_tokens":16384,"is_moderated":true}},{"id":"openai/gpt-4o-mini-2024-07-18","name":"OpenAI: GPT-4o-mini (2024-07-18)","pricing":{"prompt":"0.00000015","completion":"0.0000006","request":"0","image":"0.007225","web_search":"0","internal_reasoning":"0","input_cache_read":"0.000000075"},"created":1721260800,"top_provider":{"context_length":128000,"max_completion_tokens":16384,"is_moderated":true}},{"id":"google/gemma-2-27b-it","name":"Google: Gemma 2 27B","pricing":{"prompt":"0.00000065","completion":"0.00000065","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1720828800,"top_provider":{"context_length":8192,"max_completion_tokens":null,"is_moderated":false}},{"id":"google/gemma-2-9b-it:free","name":"Google: Gemma 2 9B (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1719532800,"top_provider":{"context_length":8192,"max_completion_tokens":8192,"is_moderated":false}},{"id":"google/gemma-2-9b-it","name":"Google: Gemma 2 9B","pricing":{"prompt":"0.000000004","completion":"0.000000004","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1719532800,"top_provider":{"context_length":8192,"max_completion_tokens":8192,"is_moderated":false}},{"id":"01-ai/yi-large","name":"01.AI: Yi Large","pricing":{"prompt":"0.000003","completion":"0.000003","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1719273600,"top_provider":{"context_length":32768,"max_completion_tokens":4096,"is_moderated":false}},{"id":"anthropic/claude-3.5-sonnet-20240620:beta","name":"Anthropic: Claude 3.5 Sonnet (2024-06-20) (self-moderated)","pricing":{"prompt":"0.000003","completion":"0.000015","request":"0","image":"0.0048","web_search":"0","internal_reasoning":"0","input_cache_read":"0.0000003","input_cache_write":"0.00000375"},"created":1718841600,"top_provider":{"context_length":200000,"max_completion_tokens":8192,"is_moderated":false}},{"id":"anthropic/claude-3.5-sonnet-20240620","name":"Anthropic: Claude 3.5 Sonnet (2024-06-20)","pricing":{"prompt":"0.000003","completion":"0.000015","request":"0","image":"0.0048","web_search":"0","internal_reasoning":"0","input_cache_read":"0.0000003","input_cache_write":"0.00000375"},"created":1718841600,"top_provider":{"context_length":200000,"max_completion_tokens":8192,"is_moderated":true}},{"id":"sao10k/l3-euryale-70b","name":"Sao10k: Llama 3 Euryale 70B v2.1","pricing":{"prompt":"0.00000148","completion":"0.00000148","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1718668800,"top_provider":{"context_length":8192,"max_completion_tokens":8192,"is_moderated":false}},{"id":"cognitivecomputations/dolphin-mixtral-8x22b","name":"Dolphin 2.9.2 Mixtral 8x22B 🐬","pricing":{"prompt":"0.0000009","completion":"0.0000009","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1717804800,"top_provider":{"context_length":16000,"max_completion_tokens":8192,"is_moderated":false}},{"id":"qwen/qwen-2-72b-instruct","name":"Qwen 2 72B Instruct","pricing":{"prompt":"0.0000009","completion":"0.0000009","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1717718400,"top_provider":{"context_length":32768,"max_completion_tokens":4096,"is_moderated":false}},{"id":"mistralai/mistral-7b-instruct-v0.3","name":"Mistral: Mistral 7B Instruct v0.3","pricing":{"prompt":"0.000000028","completion":"0.000000054","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1716768000,"top_provider":{"context_length":32768,"max_completion_tokens":16384,"is_moderated":false}},{"id":"nousresearch/hermes-2-pro-llama-3-8b","name":"NousResearch: Hermes 2 Pro - Llama-3 8B","pricing":{"prompt":"0.000000025","completion":"0.00000004","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1716768000,"top_provider":{"context_length":131072,"max_completion_tokens":131072,"is_moderated":false}},{"id":"mistralai/mistral-7b-instruct:free","name":"Mistral: Mistral 7B Instruct (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1716768000,"top_provider":{"context_length":32768,"max_completion_tokens":16384,"is_moderated":false}},{"id":"mistralai/mistral-7b-instruct","name":"Mistral: Mistral 7B Instruct","pricing":{"prompt":"0.000000028","completion":"0.000000054","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1716768000,"top_provider":{"context_length":32768,"max_completion_tokens":16384,"is_moderated":false}},{"id":"microsoft/phi-3-mini-128k-instruct","name":"Microsoft: Phi-3 Mini 128K Instruct","pricing":{"prompt":"0.0000001","completion":"0.0000001","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1716681600,"top_provider":{"context_length":128000,"max_completion_tokens":null,"is_moderated":false}},{"id":"microsoft/phi-3-medium-128k-instruct","name":"Microsoft: Phi-3 Medium 128K Instruct","pricing":{"prompt":"0.000001","completion":"0.000001","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1716508800,"top_provider":{"context_length":128000,"max_completion_tokens":null,"is_moderated":false}},{"id":"neversleep/llama-3-lumimaid-70b","name":"NeverSleep: Llama 3 Lumimaid 70B","pricing":{"prompt":"0.000004","completion":"0.000006","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1715817600,"top_provider":{"context_length":8192,"max_completion_tokens":4096,"is_moderated":false}},{"id":"google/gemini-flash-1.5","name":"Google: Gemini 1.5 Flash ","pricing":{"prompt":"0.000000075","completion":"0.0000003","request":"0","image":"0.00004","web_search":"0","internal_reasoning":"0","input_cache_read":"0.00000001875","input_cache_write":"0.0000001583"},"created":1715644800,"top_provider":{"context_length":1000000,"max_completion_tokens":8192,"is_moderated":false}},{"id":"meta-llama/llama-guard-2-8b","name":"Meta: LlamaGuard 2 8B","pricing":{"prompt":"0.0000002","completion":"0.0000002","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1715558400,"top_provider":{"context_length":8192,"max_completion_tokens":null,"is_moderated":false}},{"id":"openai/gpt-4o","name":"OpenAI: GPT-4o","pricing":{"prompt":"0.0000025","completion":"0.00001","request":"0","image":"0.003613","web_search":"0","internal_reasoning":"0","input_cache_read":"0.00000125"},"created":1715558400,"top_provider":{"context_length":128000,"max_completion_tokens":16384,"is_moderated":true}},{"id":"openai/gpt-4o:extended","name":"OpenAI: GPT-4o (extended)","pricing":{"prompt":"0.000006","completion":"0.000018","request":"0","image":"0.007225","web_search":"0","internal_reasoning":"0"},"created":1715558400,"top_provider":{"context_length":128000,"max_completion_tokens":64000,"is_moderated":true}},{"id":"openai/gpt-4o-2024-05-13","name":"OpenAI: GPT-4o (2024-05-13)","pricing":{"prompt":"0.000005","completion":"0.000015","request":"0","image":"0.007225","web_search":"0","internal_reasoning":"0"},"created":1715558400,"top_provider":{"context_length":128000,"max_completion_tokens":4096,"is_moderated":true}},{"id":"sao10k/fimbulvetr-11b-v2","name":"Fimbulvetr 11B v2","pricing":{"prompt":"0.0000008","completion":"0.0000012","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1713657600,"top_provider":{"context_length":4096,"max_completion_tokens":4096,"is_moderated":false}},{"id":"meta-llama/llama-3-70b-instruct","name":"Meta: Llama 3 70B Instruct","pricing":{"prompt":"0.0000003","completion":"0.0000004","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1713398400,"top_provider":{"context_length":8192,"max_completion_tokens":16384,"is_moderated":false}},{"id":"meta-llama/llama-3-8b-instruct","name":"Meta: Llama 3 8B Instruct","pricing":{"prompt":"0.00000003","completion":"0.00000006","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1713398400,"top_provider":{"context_length":8192,"max_completion_tokens":16384,"is_moderated":false}},{"id":"mistralai/mixtral-8x22b-instruct","name":"Mistral: Mixtral 8x22B Instruct","pricing":{"prompt":"0.0000009","completion":"0.0000009","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1713312000,"top_provider":{"context_length":65536,"max_completion_tokens":null,"is_moderated":false}},{"id":"microsoft/wizardlm-2-8x22b","name":"WizardLM-2 8x22B","pricing":{"prompt":"0.00000048","completion":"0.00000048","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1713225600,"top_provider":{"context_length":65536,"max_completion_tokens":65536,"is_moderated":false}},{"id":"openai/gpt-4-turbo","name":"OpenAI: GPT-4 Turbo","pricing":{"prompt":"0.00001","completion":"0.00003","request":"0","image":"0.01445","web_search":"0","internal_reasoning":"0"},"created":1712620800,"top_provider":{"context_length":128000,"max_completion_tokens":4096,"is_moderated":true}},{"id":"google/gemini-pro-1.5","name":"Google: Gemini 1.5 Pro","pricing":{"prompt":"0.00000125","completion":"0.000005","request":"0","image":"0.0006575","web_search":"0","internal_reasoning":"0"},"created":1712620800,"top_provider":{"context_length":2000000,"max_completion_tokens":8192,"is_moderated":false}},{"id":"cohere/command-r-plus","name":"Cohere: Command R+","pricing":{"prompt":"0.000003","completion":"0.000015","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1712188800,"top_provider":{"context_length":128000,"max_completion_tokens":4000,"is_moderated":true}},{"id":"cohere/command-r-plus-04-2024","name":"Cohere: Command R+ (04-2024)","pricing":{"prompt":"0.000003","completion":"0.000015","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1712016000,"top_provider":{"context_length":128000,"max_completion_tokens":4000,"is_moderated":true}},{"id":"sophosympatheia/midnight-rose-70b","name":"Midnight Rose 70B","pricing":{"prompt":"0.0000008","completion":"0.0000008","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1711065600,"top_provider":{"context_length":4096,"max_completion_tokens":2048,"is_moderated":false}},{"id":"cohere/command","name":"Cohere: Command","pricing":{"prompt":"0.000001","completion":"0.000002","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1710374400,"top_provider":{"context_length":4096,"max_completion_tokens":4000,"is_moderated":true}},{"id":"cohere/command-r","name":"Cohere: Command R","pricing":{"prompt":"0.0000005","completion":"0.0000015","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1710374400,"top_provider":{"context_length":128000,"max_completion_tokens":4000,"is_moderated":true}},{"id":"anthropic/claude-3-haiku:beta","name":"Anthropic: Claude 3 Haiku (self-moderated)","pricing":{"prompt":"0.00000025","completion":"0.00000125","request":"0","image":"0.0004","web_search":"0","internal_reasoning":"0","input_cache_read":"0.00000003","input_cache_write":"0.0000003"},"created":1710288000,"top_provider":{"context_length":200000,"max_completion_tokens":4096,"is_moderated":false}},{"id":"anthropic/claude-3-haiku","name":"Anthropic: Claude 3 Haiku","pricing":{"prompt":"0.00000025","completion":"0.00000125","request":"0","image":"0.0004","web_search":"0","internal_reasoning":"0","input_cache_read":"0.00000003","input_cache_write":"0.0000003"},"created":1710288000,"top_provider":{"context_length":200000,"max_completion_tokens":4096,"is_moderated":true}},{"id":"anthropic/claude-3-sonnet","name":"Anthropic: Claude 3 Sonnet","pricing":{"prompt":"0.000003","completion":"0.000015","request":"0","image":"0.0048","web_search":"0","internal_reasoning":"0","input_cache_read":"0.0000003","input_cache_write":"0.00000375"},"created":1709596800,"top_provider":{"context_length":200000,"max_completion_tokens":4096,"is_moderated":false}},{"id":"anthropic/claude-3-opus:beta","name":"Anthropic: Claude 3 Opus (self-moderated)","pricing":{"prompt":"0.000015","completion":"0.000075","request":"0","image":"0.024","web_search":"0","internal_reasoning":"0","input_cache_read":"0.0000015","input_cache_write":"0.00001875"},"created":1709596800,"top_provider":{"context_length":200000,"max_completion_tokens":4096,"is_moderated":false}},{"id":"anthropic/claude-3-opus","name":"Anthropic: Claude 3 Opus","pricing":{"prompt":"0.000015","completion":"0.000075","request":"0","image":"0.024","web_search":"0","internal_reasoning":"0","input_cache_read":"0.0000015","input_cache_write":"0.00001875"},"created":1709596800,"top_provider":{"context_length":200000,"max_completion_tokens":4096,"is_moderated":true}},{"id":"cohere/command-r-03-2024","name":"Cohere: Command R (03-2024)","pricing":{"prompt":"0.0000005","completion":"0.0000015","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1709341200,"top_provider":{"context_length":128000,"max_completion_tokens":4000,"is_moderated":true}},{"id":"mistralai/mistral-large","name":"Mistral Large","pricing":{"prompt":"0.000002","completion":"0.000006","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1708905600,"top_provider":{"context_length":128000,"max_completion_tokens":null,"is_moderated":false}},{"id":"openai/gpt-4-turbo-preview","name":"OpenAI: GPT-4 Turbo Preview","pricing":{"prompt":"0.00001","completion":"0.00003","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1706140800,"top_provider":{"context_length":128000,"max_completion_tokens":4096,"is_moderated":true}},{"id":"openai/gpt-3.5-turbo-0613","name":"OpenAI: GPT-3.5 Turbo (older v0613)","pricing":{"prompt":"0.000001","completion":"0.000002","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1706140800,"top_provider":{"context_length":4095,"max_completion_tokens":4096,"is_moderated":false}},{"id":"nousresearch/nous-hermes-2-mixtral-8x7b-dpo","name":"Nous: Hermes 2 Mixtral 8x7B DPO","pricing":{"prompt":"0.0000006","completion":"0.0000006","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1705363200,"top_provider":{"context_length":32768,"max_completion_tokens":2048,"is_moderated":false}},{"id":"mistralai/mistral-small","name":"Mistral Small","pricing":{"prompt":"0.0000002","completion":"0.0000006","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1704844800,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"mistralai/mistral-tiny","name":"Mistral Tiny","pricing":{"prompt":"0.00000025","completion":"0.00000025","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1704844800,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"mistralai/mistral-7b-instruct-v0.2","name":"Mistral: Mistral 7B Instruct v0.2","pricing":{"prompt":"0.0000002","completion":"0.0000002","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1703721600,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"mistralai/mixtral-8x7b-instruct","name":"Mistral: Mixtral 8x7B Instruct","pricing":{"prompt":"0.00000008","completion":"0.00000024","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1702166400,"top_provider":{"context_length":32768,"max_completion_tokens":16384,"is_moderated":false}},{"id":"neversleep/noromaid-20b","name":"Noromaid 20B","pricing":{"prompt":"0.00000125","completion":"0.000002","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1700956800,"top_provider":{"context_length":8192,"max_completion_tokens":2048,"is_moderated":false}},{"id":"undi95/toppy-m-7b","name":"Toppy M 7B","pricing":{"prompt":"0.0000008","completion":"0.0000012","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1699574400,"top_provider":{"context_length":4096,"max_completion_tokens":4096,"is_moderated":false}},{"id":"alpindale/goliath-120b","name":"Goliath 120B","pricing":{"prompt":"0.000009","completion":"0.000011","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1699574400,"top_provider":{"context_length":6144,"max_completion_tokens":null,"is_moderated":false}},{"id":"openrouter/auto","name":"Auto Router","pricing":{"prompt":"-1","completion":"-1"},"created":1699401600,"top_provider":{"context_length":null,"max_completion_tokens":null,"is_moderated":false}},{"id":"openai/gpt-4-1106-preview","name":"OpenAI: GPT-4 Turbo (older v1106)","pricing":{"prompt":"0.00001","completion":"0.00003","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1699228800,"top_provider":{"context_length":128000,"max_completion_tokens":4096,"is_moderated":true}},{"id":"openai/gpt-3.5-turbo-instruct","name":"OpenAI: GPT-3.5 Turbo Instruct","pricing":{"prompt":"0.0000015","completion":"0.000002","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1695859200,"top_provider":{"context_length":4095,"max_completion_tokens":4096,"is_moderated":true}},{"id":"mistralai/mistral-7b-instruct-v0.1","name":"Mistral: Mistral 7B Instruct v0.1","pricing":{"prompt":"0.00000011","completion":"0.00000019","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1695859200,"top_provider":{"context_length":2824,"max_completion_tokens":null,"is_moderated":false}},{"id":"pygmalionai/mythalion-13b","name":"Pygmalion: Mythalion 13B","pricing":{"prompt":"0.0000008","completion":"0.0000012","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1693612800,"top_provider":{"context_length":4096,"max_completion_tokens":4096,"is_moderated":false}},{"id":"openai/gpt-3.5-turbo-16k","name":"OpenAI: GPT-3.5 Turbo 16k","pricing":{"prompt":"0.000003","completion":"0.000004","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1693180800,"top_provider":{"context_length":16385,"max_completion_tokens":4096,"is_moderated":true}},{"id":"mancer/weaver","name":"Mancer: Weaver (alpha)","pricing":{"prompt":"0.0000015","completion":"0.0000015","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1690934400,"top_provider":{"context_length":8000,"max_completion_tokens":1000,"is_moderated":false}},{"id":"undi95/remm-slerp-l2-13b","name":"ReMM SLERP 13B","pricing":{"prompt":"0.0000007","completion":"0.000001","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1689984000,"top_provider":{"context_length":6144,"max_completion_tokens":null,"is_moderated":false}},{"id":"gryphe/mythomax-l2-13b","name":"MythoMax 13B","pricing":{"prompt":"0.00000006","completion":"0.00000006","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1688256000,"top_provider":{"context_length":4096,"max_completion_tokens":null,"is_moderated":false}},{"id":"openai/gpt-4-0314","name":"OpenAI: GPT-4 (older v0314)","pricing":{"prompt":"0.00003","completion":"0.00006","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1685232000,"top_provider":{"context_length":8191,"max_completion_tokens":4096,"is_moderated":true}},{"id":"openai/gpt-3.5-turbo","name":"OpenAI: GPT-3.5 Turbo","pricing":{"prompt":"0.0000005","completion":"0.0000015","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1685232000,"top_provider":{"context_length":16385,"max_completion_tokens":4096,"is_moderated":true}},{"id":"openai/gpt-4","name":"OpenAI: GPT-4","pricing":{"prompt":"0.00003","completion":"0.00006","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1685232000,"top_provider":{"context_length":8191,"max_completion_tokens":4096,"is_moderated":true}}] \ No newline at end of file +export const models = [{"id":"openai/gpt-oss-120b","name":"OpenAI: GPT OSS 120B","pricing":{"prompt":"0.0000001","completion":"0.0000005","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1754414231,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"openai/gpt-oss-20b","name":"OpenAI: GPT OSS 20B","pricing":{"prompt":"0.00000005","completion":"0.0000002","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1754414229,"top_provider":{"context_length":131072,"max_completion_tokens":32768,"is_moderated":false}},{"id":"anthropic/claude-opus-4.1","name":"Anthropic: Claude Opus 4.1","pricing":{"prompt":"0.000015","completion":"0.000075","request":"0","image":"0.024","audio":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.0000015","input_cache_write":"0.00001875"},"created":1754411591,"top_provider":{"context_length":200000,"max_completion_tokens":32000,"is_moderated":true}},{"id":"openrouter/horizon-beta","name":"Horizon Beta","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1754089419,"top_provider":{"context_length":256000,"max_completion_tokens":128000,"is_moderated":false}},{"id":"mistralai/codestral-2508","name":"Mistral: Codestral 2508","pricing":{"prompt":"0.0000003","completion":"0.0000009","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1754079630,"top_provider":{"context_length":256000,"max_completion_tokens":null,"is_moderated":false}},{"id":"qwen/qwen3-30b-a3b-instruct-2507","name":"Qwen: Qwen3 30B A3B Instruct 2507","pricing":{"prompt":"0.0000002","completion":"0.0000008","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1753806965,"top_provider":{"context_length":131072,"max_completion_tokens":32768,"is_moderated":false}},{"id":"z-ai/glm-4.5","name":"Z.AI: GLM 4.5","pricing":{"prompt":"0.0000002","completion":"0.0000002","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1753471347,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"z-ai/glm-4.5-air:free","name":"Z.AI: GLM 4.5 Air (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1753471258,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"z-ai/glm-4.5-air","name":"Z.AI: GLM 4.5 Air","pricing":{"prompt":"0.0000002","completion":"0.0000011","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.00000003"},"created":1753471258,"top_provider":{"context_length":128000,"max_completion_tokens":96000,"is_moderated":false}},{"id":"qwen/qwen3-235b-a22b-thinking-2507","name":"Qwen: Qwen3 235B A22B Thinking 2507","pricing":{"prompt":"0.000000077968332","completion":"0.00000031202496","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1753449557,"top_provider":{"context_length":262144,"max_completion_tokens":null,"is_moderated":false}},{"id":"z-ai/glm-4-32b","name":"Z.AI: GLM 4 32B ","pricing":{"prompt":"0.0000001","completion":"0.0000001","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1753376617,"top_provider":{"context_length":128000,"max_completion_tokens":null,"is_moderated":false}},{"id":"qwen/qwen3-coder","name":"Qwen: Qwen3 Coder ","pricing":{"prompt":"0.0000001999188","completion":"0.000000800064","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1753230546,"top_provider":{"context_length":262144,"max_completion_tokens":null,"is_moderated":false}},{"id":"bytedance/ui-tars-1.5-7b","name":"Bytedance: UI-TARS 7B ","pricing":{"prompt":"0.0000001","completion":"0.0000002","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1753205056,"top_provider":{"context_length":128000,"max_completion_tokens":2048,"is_moderated":false}},{"id":"google/gemini-2.5-flash-lite","name":"Google: Gemini 2.5 Flash Lite","pricing":{"prompt":"0.0000001","completion":"0.0000004","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.000000025","input_cache_write":"0.0000001833"},"created":1753200276,"top_provider":{"context_length":1048576,"max_completion_tokens":65535,"is_moderated":false}},{"id":"qwen/qwen3-235b-a22b-2507","name":"Qwen: Qwen3 235B A22B Instruct 2507","pricing":{"prompt":"0.000000077968332","completion":"0.00000031202496","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1753119555,"top_provider":{"context_length":262144,"max_completion_tokens":null,"is_moderated":false}},{"id":"switchpoint/router","name":"Switchpoint Router","pricing":{"prompt":"0.00000085","completion":"0.0000034","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1752272899,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"moonshotai/kimi-k2:free","name":"MoonshotAI: Kimi K2 (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1752263252,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"moonshotai/kimi-k2","name":"MoonshotAI: Kimi K2","pricing":{"prompt":"0.00000014","completion":"0.00000249","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1752263252,"top_provider":{"context_length":63000,"max_completion_tokens":63000,"is_moderated":false}},{"id":"thudm/glm-4.1v-9b-thinking","name":"THUDM: GLM 4.1V 9B Thinking","pricing":{"prompt":"0.000000035","completion":"0.000000138","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1752244385,"top_provider":{"context_length":65536,"max_completion_tokens":8000,"is_moderated":false}},{"id":"mistralai/devstral-medium","name":"Mistral: Devstral Medium","pricing":{"prompt":"0.0000004","completion":"0.000002","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1752161321,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"mistralai/devstral-small","name":"Mistral: Devstral Small 1.1","pricing":{"prompt":"0.00000007","completion":"0.00000028","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1752160751,"top_provider":{"context_length":128000,"max_completion_tokens":null,"is_moderated":false}},{"id":"cognitivecomputations/dolphin-mistral-24b-venice-edition:free","name":"Venice: Uncensored (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1752094966,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"x-ai/grok-4","name":"xAI: Grok 4","pricing":{"prompt":"0.000003","completion":"0.000015","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.00000075"},"created":1752087689,"top_provider":{"context_length":256000,"max_completion_tokens":null,"is_moderated":false}},{"id":"google/gemma-3n-e2b-it:free","name":"Google: Gemma 3n 2B (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1752074904,"top_provider":{"context_length":8192,"max_completion_tokens":2048,"is_moderated":false}},{"id":"tencent/hunyuan-a13b-instruct:free","name":"Tencent: Hunyuan A13B Instruct (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1751987664,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"tencent/hunyuan-a13b-instruct","name":"Tencent: Hunyuan A13B Instruct","pricing":{"prompt":"0.00000003","completion":"0.00000003","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1751987664,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"tngtech/deepseek-r1t2-chimera:free","name":"TNG: DeepSeek R1T2 Chimera (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1751986985,"top_provider":{"context_length":163840,"max_completion_tokens":null,"is_moderated":false}},{"id":"morph/morph-v3-large","name":"Morph: Morph V3 Large","pricing":{"prompt":"0.0000009","completion":"0.0000019","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1751910858,"top_provider":{"context_length":81920,"max_completion_tokens":38000,"is_moderated":false}},{"id":"morph/morph-v3-fast","name":"Morph: Morph V3 Fast","pricing":{"prompt":"0.0000009","completion":"0.0000019","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1751910002,"top_provider":{"context_length":81920,"max_completion_tokens":38000,"is_moderated":false}},{"id":"baidu/ernie-4.5-300b-a47b","name":"Baidu: ERNIE 4.5 300B A47B ","pricing":{"prompt":"0.00000028","completion":"0.0000011","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1751300139,"top_provider":{"context_length":123000,"max_completion_tokens":12000,"is_moderated":false}},{"id":"thedrummer/anubis-70b-v1.1","name":"TheDrummer: Anubis 70B V1.1","pricing":{"prompt":"0.0000004","completion":"0.0000007","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1751208347,"top_provider":{"context_length":16384,"max_completion_tokens":null,"is_moderated":false}},{"id":"inception/mercury","name":"Inception: Mercury","pricing":{"prompt":"0.00000025","completion":"0.000001","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1750973026,"top_provider":{"context_length":128000,"max_completion_tokens":16384,"is_moderated":false}},{"id":"mistralai/mistral-small-3.2-24b-instruct:free","name":"Mistral: Mistral Small 3.2 24B (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1750443016,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"mistralai/mistral-small-3.2-24b-instruct","name":"Mistral: Mistral Small 3.2 24B","pricing":{"prompt":"0.00000001999188","completion":"0.0000000800064","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1750443016,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"minimax/minimax-m1","name":"MiniMax: MiniMax M1","pricing":{"prompt":"0.0000003","completion":"0.00000165","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1750200414,"top_provider":{"context_length":1000000,"max_completion_tokens":40000,"is_moderated":false}},{"id":"google/gemini-2.5-flash-lite-preview-06-17","name":"Google: Gemini 2.5 Flash Lite Preview 06-17","pricing":{"prompt":"0.0000001","completion":"0.0000004","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.000000025","input_cache_write":"0.0000001833"},"created":1750173831,"top_provider":{"context_length":1048576,"max_completion_tokens":65535,"is_moderated":false}},{"id":"google/gemini-2.5-flash","name":"Google: Gemini 2.5 Flash","pricing":{"prompt":"0.0000003","completion":"0.0000025","request":"0","image":"0.001238","audio":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.000000075","input_cache_write":"0.0000003833"},"created":1750172488,"top_provider":{"context_length":1048576,"max_completion_tokens":65535,"is_moderated":false}},{"id":"google/gemini-2.5-pro","name":"Google: Gemini 2.5 Pro","pricing":{"prompt":"0.00000125","completion":"0.00001","request":"0","image":"0.00516","audio":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.00000031","input_cache_write":"0.000001625"},"created":1750169544,"top_provider":{"context_length":1048576,"max_completion_tokens":65536,"is_moderated":false}},{"id":"moonshotai/kimi-dev-72b:free","name":"Kimi Dev 72b (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1750115909,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"openai/o3-pro","name":"OpenAI: o3 Pro","pricing":{"prompt":"0.00002","completion":"0.00008","request":"0","image":"0.0153","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1749598352,"top_provider":{"context_length":200000,"max_completion_tokens":100000,"is_moderated":true}},{"id":"x-ai/grok-3-mini","name":"xAI: Grok 3 Mini","pricing":{"prompt":"0.0000003","completion":"0.0000005","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.000000075"},"created":1749583245,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"x-ai/grok-3","name":"xAI: Grok 3","pricing":{"prompt":"0.000003","completion":"0.000015","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.00000075"},"created":1749582908,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"mistralai/magistral-small-2506","name":"Mistral: Magistral Small 2506","pricing":{"prompt":"0.0000005","completion":"0.0000015","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1749569561,"top_provider":{"context_length":40000,"max_completion_tokens":40000,"is_moderated":false}},{"id":"mistralai/magistral-medium-2506","name":"Mistral: Magistral Medium 2506","pricing":{"prompt":"0.000002","completion":"0.000005","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1749354054,"top_provider":{"context_length":40960,"max_completion_tokens":40000,"is_moderated":false}},{"id":"mistralai/magistral-medium-2506:thinking","name":"Mistral: Magistral Medium 2506 (thinking)","pricing":{"prompt":"0.000002","completion":"0.000005","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1749354054,"top_provider":{"context_length":40960,"max_completion_tokens":40000,"is_moderated":false}},{"id":"google/gemini-2.5-pro-preview","name":"Google: Gemini 2.5 Pro Preview 06-05","pricing":{"prompt":"0.00000125","completion":"0.00001","request":"0","image":"0.00516","audio":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.00000031","input_cache_write":"0.000001625"},"created":1749137257,"top_provider":{"context_length":1048576,"max_completion_tokens":65536,"is_moderated":false}},{"id":"deepseek/deepseek-r1-distill-qwen-7b","name":"DeepSeek: R1 Distill Qwen 7B","pricing":{"prompt":"0.0000001","completion":"0.0000002","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1748628237,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"deepseek/deepseek-r1-0528-qwen3-8b:free","name":"DeepSeek: Deepseek R1 0528 Qwen3 8B (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1748538543,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"deepseek/deepseek-r1-0528-qwen3-8b","name":"DeepSeek: Deepseek R1 0528 Qwen3 8B","pricing":{"prompt":"0.00000001","completion":"0.00000002","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1748538543,"top_provider":{"context_length":32000,"max_completion_tokens":null,"is_moderated":false}},{"id":"deepseek/deepseek-r1-0528:free","name":"DeepSeek: R1 0528 (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1748455170,"top_provider":{"context_length":163840,"max_completion_tokens":null,"is_moderated":false}},{"id":"deepseek/deepseek-r1-0528","name":"DeepSeek: R1 0528","pricing":{"prompt":"0.00000017992692","completion":"0.0000007200576","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1748455170,"top_provider":{"context_length":163840,"max_completion_tokens":null,"is_moderated":false}},{"id":"sarvamai/sarvam-m:free","name":"Sarvam AI: Sarvam-M (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1748188413,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"thedrummer/valkyrie-49b-v1","name":"TheDrummer: Valkyrie 49B V1","pricing":{"prompt":"0.00000065","completion":"0.000001","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1748022670,"top_provider":{"context_length":131072,"max_completion_tokens":131072,"is_moderated":false}},{"id":"anthropic/claude-opus-4","name":"Anthropic: Claude Opus 4","pricing":{"prompt":"0.000015","completion":"0.000075","request":"0","image":"0.024","audio":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.0000015","input_cache_write":"0.00001875"},"created":1747931245,"top_provider":{"context_length":200000,"max_completion_tokens":32000,"is_moderated":true}},{"id":"anthropic/claude-sonnet-4","name":"Anthropic: Claude Sonnet 4","pricing":{"prompt":"0.000003","completion":"0.000015","request":"0","image":"0.0048","audio":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.0000003","input_cache_write":"0.00000375"},"created":1747930371,"top_provider":{"context_length":200000,"max_completion_tokens":64000,"is_moderated":false}},{"id":"mistralai/devstral-small-2505:free","name":"Mistral: Devstral Small 2505 (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1747837379,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"mistralai/devstral-small-2505","name":"Mistral: Devstral Small 2505","pricing":{"prompt":"0.00000001999188","completion":"0.0000000800064","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1747837379,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"google/gemma-3n-e4b-it:free","name":"Google: Gemma 3n 4B (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1747776824,"top_provider":{"context_length":8192,"max_completion_tokens":2048,"is_moderated":false}},{"id":"google/gemma-3n-e4b-it","name":"Google: Gemma 3n 4B","pricing":{"prompt":"0.00000002","completion":"0.00000004","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1747776824,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"openai/codex-mini","name":"OpenAI: Codex Mini","pricing":{"prompt":"0.0000015","completion":"0.000006","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.000000375"},"created":1747409761,"top_provider":{"context_length":200000,"max_completion_tokens":100000,"is_moderated":true}},{"id":"nousresearch/deephermes-3-mistral-24b-preview","name":"Nous: DeepHermes 3 Mistral 24B Preview","pricing":{"prompt":"0.00000009329544","completion":"0.0000003733632","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1746830904,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"mistralai/mistral-medium-3","name":"Mistral: Mistral Medium 3","pricing":{"prompt":"0.0000004","completion":"0.000002","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1746627341,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"google/gemini-2.5-pro-preview-05-06","name":"Google: Gemini 2.5 Pro Preview 05-06","pricing":{"prompt":"0.00000125","completion":"0.00001","request":"0","image":"0.00516","audio":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.00000031","input_cache_write":"0.000001625"},"created":1746578513,"top_provider":{"context_length":1048576,"max_completion_tokens":65535,"is_moderated":false}},{"id":"arcee-ai/spotlight","name":"Arcee AI: Spotlight","pricing":{"prompt":"0.00000018","completion":"0.00000018","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1746481552,"top_provider":{"context_length":131072,"max_completion_tokens":65537,"is_moderated":false}},{"id":"arcee-ai/maestro-reasoning","name":"Arcee AI: Maestro Reasoning","pricing":{"prompt":"0.0000009","completion":"0.0000033","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1746481269,"top_provider":{"context_length":131072,"max_completion_tokens":32000,"is_moderated":false}},{"id":"arcee-ai/virtuoso-large","name":"Arcee AI: Virtuoso Large","pricing":{"prompt":"0.00000075","completion":"0.0000012","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1746478885,"top_provider":{"context_length":131072,"max_completion_tokens":64000,"is_moderated":false}},{"id":"arcee-ai/coder-large","name":"Arcee AI: Coder Large","pricing":{"prompt":"0.0000005","completion":"0.0000008","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1746478663,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"microsoft/phi-4-reasoning-plus","name":"Microsoft: Phi 4 Reasoning Plus","pricing":{"prompt":"0.00000007","completion":"0.00000035","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1746130961,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"inception/mercury-coder","name":"Inception: Mercury Coder","pricing":{"prompt":"0.00000025","completion":"0.000001","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1746033880,"top_provider":{"context_length":128000,"max_completion_tokens":16384,"is_moderated":false}},{"id":"qwen/qwen3-4b:free","name":"Qwen: Qwen3 4B (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1746031104,"top_provider":{"context_length":40960,"max_completion_tokens":null,"is_moderated":false}},{"id":"opengvlab/internvl3-14b","name":"OpenGVLab: InternVL3 14B","pricing":{"prompt":"0.0000002","completion":"0.0000004","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1746021355,"top_provider":{"context_length":12288,"max_completion_tokens":null,"is_moderated":false}},{"id":"deepseek/deepseek-prover-v2","name":"DeepSeek: DeepSeek Prover V2","pricing":{"prompt":"0.0000005","completion":"0.00000218","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1746013094,"top_provider":{"context_length":163840,"max_completion_tokens":null,"is_moderated":false}},{"id":"meta-llama/llama-guard-4-12b","name":"Meta: Llama Guard 4 12B","pricing":{"prompt":"0.00000005","completion":"0.00000005","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1745975193,"top_provider":{"context_length":163840,"max_completion_tokens":null,"is_moderated":false}},{"id":"qwen/qwen3-30b-a3b:free","name":"Qwen: Qwen3 30B A3B (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1745878604,"top_provider":{"context_length":40960,"max_completion_tokens":null,"is_moderated":false}},{"id":"qwen/qwen3-30b-a3b","name":"Qwen: Qwen3 30B A3B","pricing":{"prompt":"0.00000001999188","completion":"0.0000000800064","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1745878604,"top_provider":{"context_length":40960,"max_completion_tokens":null,"is_moderated":false}},{"id":"qwen/qwen3-8b:free","name":"Qwen: Qwen3 8B (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1745876632,"top_provider":{"context_length":40960,"max_completion_tokens":40960,"is_moderated":false}},{"id":"qwen/qwen3-8b","name":"Qwen: Qwen3 8B","pricing":{"prompt":"0.000000035","completion":"0.000000138","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1745876632,"top_provider":{"context_length":128000,"max_completion_tokens":20000,"is_moderated":false}},{"id":"qwen/qwen3-14b:free","name":"Qwen: Qwen3 14B (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1745876478,"top_provider":{"context_length":40960,"max_completion_tokens":null,"is_moderated":false}},{"id":"qwen/qwen3-14b","name":"Qwen: Qwen3 14B","pricing":{"prompt":"0.00000006","completion":"0.00000024","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1745876478,"top_provider":{"context_length":40960,"max_completion_tokens":null,"is_moderated":false}},{"id":"qwen/qwen3-32b","name":"Qwen: Qwen3 32B","pricing":{"prompt":"0.000000017992692","completion":"0.00000007200576","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1745875945,"top_provider":{"context_length":40960,"max_completion_tokens":null,"is_moderated":false}},{"id":"qwen/qwen3-235b-a22b:free","name":"Qwen: Qwen3 235B A22B (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1745875757,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"qwen/qwen3-235b-a22b","name":"Qwen: Qwen3 235B A22B","pricing":{"prompt":"0.00000013","completion":"0.0000006","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1745875757,"top_provider":{"context_length":40960,"max_completion_tokens":40960,"is_moderated":false}},{"id":"tngtech/deepseek-r1t-chimera:free","name":"TNG: DeepSeek R1T Chimera (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1745760875,"top_provider":{"context_length":163840,"max_completion_tokens":null,"is_moderated":false}},{"id":"tngtech/deepseek-r1t-chimera","name":"TNG: DeepSeek R1T Chimera","pricing":{"prompt":"0.00000017992692","completion":"0.0000007200576","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1745760875,"top_provider":{"context_length":163840,"max_completion_tokens":null,"is_moderated":false}},{"id":"microsoft/mai-ds-r1:free","name":"Microsoft: MAI DS R1 (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1745194100,"top_provider":{"context_length":163840,"max_completion_tokens":null,"is_moderated":false}},{"id":"microsoft/mai-ds-r1","name":"Microsoft: MAI DS R1","pricing":{"prompt":"0.0000001999188","completion":"0.000000800064","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1745194100,"top_provider":{"context_length":163840,"max_completion_tokens":null,"is_moderated":false}},{"id":"thudm/glm-z1-32b:free","name":"THUDM: GLM Z1 32B (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1744924148,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"thudm/glm-4-32b","name":"THUDM: GLM 4 32B","pricing":{"prompt":"0.00000024","completion":"0.00000024","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1744920915,"top_provider":{"context_length":32000,"max_completion_tokens":32000,"is_moderated":false}},{"id":"openai/o4-mini-high","name":"OpenAI: o4 Mini High","pricing":{"prompt":"0.0000011","completion":"0.0000044","request":"0","image":"0.0008415","audio":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.000000275"},"created":1744824212,"top_provider":{"context_length":200000,"max_completion_tokens":100000,"is_moderated":true}},{"id":"openai/o3","name":"OpenAI: o3","pricing":{"prompt":"0.000002","completion":"0.000008","request":"0","image":"0.00153","audio":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.0000005"},"created":1744823457,"top_provider":{"context_length":200000,"max_completion_tokens":100000,"is_moderated":true}},{"id":"openai/o4-mini","name":"OpenAI: o4 Mini","pricing":{"prompt":"0.0000011","completion":"0.0000044","request":"0","image":"0.0008415","audio":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.000000275"},"created":1744820942,"top_provider":{"context_length":200000,"max_completion_tokens":100000,"is_moderated":true}},{"id":"shisa-ai/shisa-v2-llama3.3-70b:free","name":"Shisa AI: Shisa V2 Llama 3.3 70B (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1744754858,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"shisa-ai/shisa-v2-llama3.3-70b","name":"Shisa AI: Shisa V2 Llama 3.3 70B ","pricing":{"prompt":"0.00000001999188","completion":"0.0000000800064","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1744754858,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"openai/gpt-4.1","name":"OpenAI: GPT-4.1","pricing":{"prompt":"0.000002","completion":"0.000008","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.0000005"},"created":1744651385,"top_provider":{"context_length":1047576,"max_completion_tokens":32768,"is_moderated":true}},{"id":"openai/gpt-4.1-mini","name":"OpenAI: GPT-4.1 Mini","pricing":{"prompt":"0.0000004","completion":"0.0000016","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.0000001"},"created":1744651381,"top_provider":{"context_length":1047576,"max_completion_tokens":32768,"is_moderated":true}},{"id":"openai/gpt-4.1-nano","name":"OpenAI: GPT-4.1 Nano","pricing":{"prompt":"0.0000001","completion":"0.0000004","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.000000025"},"created":1744651369,"top_provider":{"context_length":1047576,"max_completion_tokens":32768,"is_moderated":true}},{"id":"eleutherai/llemma_7b","name":"EleutherAI: Llemma 7b","pricing":{"prompt":"0.0000008","completion":"0.0000012","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1744643225,"top_provider":{"context_length":4096,"max_completion_tokens":4096,"is_moderated":false}},{"id":"alfredpros/codellama-7b-instruct-solidity","name":"AlfredPros: CodeLLaMa 7B Instruct Solidity","pricing":{"prompt":"0.0000006","completion":"0.000001","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1744641874,"top_provider":{"context_length":8192,"max_completion_tokens":8192,"is_moderated":false}},{"id":"arliai/qwq-32b-arliai-rpr-v1:free","name":"ArliAI: QwQ 32B RpR v1 (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1744555982,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"arliai/qwq-32b-arliai-rpr-v1","name":"ArliAI: QwQ 32B RpR v1","pricing":{"prompt":"0.00000001","completion":"0.0000000400032","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1744555982,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"agentica-org/deepcoder-14b-preview:free","name":"Agentica: Deepcoder 14B Preview (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1744555395,"top_provider":{"context_length":96000,"max_completion_tokens":null,"is_moderated":false}},{"id":"agentica-org/deepcoder-14b-preview","name":"Agentica: Deepcoder 14B Preview","pricing":{"prompt":"0.000000015","completion":"0.000000015","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1744555395,"top_provider":{"context_length":96000,"max_completion_tokens":null,"is_moderated":false}},{"id":"moonshotai/kimi-vl-a3b-thinking:free","name":"Moonshot AI: Kimi VL A3B Thinking (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1744304841,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"moonshotai/kimi-vl-a3b-thinking","name":"Moonshot AI: Kimi VL A3B Thinking","pricing":{"prompt":"0.00000002498985","completion":"0.000000100008","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1744304841,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"x-ai/grok-3-mini-beta","name":"xAI: Grok 3 Mini Beta","pricing":{"prompt":"0.0000003","completion":"0.0000005","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.000000075"},"created":1744240195,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"x-ai/grok-3-beta","name":"xAI: Grok 3 Beta","pricing":{"prompt":"0.000003","completion":"0.000015","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.00000075"},"created":1744240068,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"nvidia/llama-3.3-nemotron-super-49b-v1","name":"NVIDIA: Llama 3.3 Nemotron Super 49B v1","pricing":{"prompt":"0.00000013","completion":"0.0000004","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1744119494,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"nvidia/llama-3.1-nemotron-ultra-253b-v1:free","name":"NVIDIA: Llama 3.1 Nemotron Ultra 253B v1 (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1744115059,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"nvidia/llama-3.1-nemotron-ultra-253b-v1","name":"NVIDIA: Llama 3.1 Nemotron Ultra 253B v1","pricing":{"prompt":"0.0000006","completion":"0.0000018","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1744115059,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"meta-llama/llama-4-maverick","name":"Meta: Llama 4 Maverick","pricing":{"prompt":"0.00000015","completion":"0.0000006","request":"0","image":"0.0006684","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1743881822,"top_provider":{"context_length":1048576,"max_completion_tokens":16384,"is_moderated":false}},{"id":"meta-llama/llama-4-scout","name":"Meta: Llama 4 Scout","pricing":{"prompt":"0.00000008","completion":"0.0000003","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1743881519,"top_provider":{"context_length":1048576,"max_completion_tokens":1048576,"is_moderated":false}},{"id":"deepseek/deepseek-v3-base","name":"DeepSeek: DeepSeek V3 Base","pricing":{"prompt":"0.0000001999188","completion":"0.000000800064","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1743272023,"top_provider":{"context_length":163840,"max_completion_tokens":null,"is_moderated":false}},{"id":"scb10x/llama3.1-typhoon2-70b-instruct","name":"Typhoon2 70B Instruct","pricing":{"prompt":"0.00000088","completion":"0.00000088","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1743196170,"top_provider":{"context_length":8192,"max_completion_tokens":null,"is_moderated":false}},{"id":"google/gemini-2.5-pro-exp-03-25","name":"Google: Gemini 2.5 Pro Experimental","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1742922099,"top_provider":{"context_length":1048576,"max_completion_tokens":65535,"is_moderated":false}},{"id":"qwen/qwen2.5-vl-32b-instruct:free","name":"Qwen: Qwen2.5 VL 32B Instruct (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1742839838,"top_provider":{"context_length":8192,"max_completion_tokens":null,"is_moderated":false}},{"id":"qwen/qwen2.5-vl-32b-instruct","name":"Qwen: Qwen2.5 VL 32B Instruct","pricing":{"prompt":"0.00000001999188","completion":"0.0000000800064","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1742839838,"top_provider":{"context_length":16384,"max_completion_tokens":null,"is_moderated":false}},{"id":"deepseek/deepseek-chat-v3-0324:free","name":"DeepSeek: DeepSeek V3 0324 (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1742824755,"top_provider":{"context_length":163840,"max_completion_tokens":163840,"is_moderated":false}},{"id":"deepseek/deepseek-chat-v3-0324","name":"DeepSeek: DeepSeek V3 0324","pricing":{"prompt":"0.00000017992692","completion":"0.0000007200576","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1742824755,"top_provider":{"context_length":163840,"max_completion_tokens":null,"is_moderated":false}},{"id":"featherless/qwerky-72b:free","name":"Qrwkv 72B (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1742481597,"top_provider":{"context_length":32768,"max_completion_tokens":4096,"is_moderated":false}},{"id":"openai/o1-pro","name":"OpenAI: o1-pro","pricing":{"prompt":"0.00015","completion":"0.0006","request":"0","image":"0.21675","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1742423211,"top_provider":{"context_length":200000,"max_completion_tokens":100000,"is_moderated":true}},{"id":"mistralai/mistral-small-3.1-24b-instruct:free","name":"Mistral: Mistral Small 3.1 24B (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1742238937,"top_provider":{"context_length":128000,"max_completion_tokens":null,"is_moderated":false}},{"id":"mistralai/mistral-small-3.1-24b-instruct","name":"Mistral: Mistral Small 3.1 24B","pricing":{"prompt":"0.000000017992692","completion":"0.00000007200576","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1742238937,"top_provider":{"context_length":131072,"max_completion_tokens":96000,"is_moderated":false}},{"id":"google/gemma-3-4b-it:free","name":"Google: Gemma 3 4B (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1741905510,"top_provider":{"context_length":32768,"max_completion_tokens":8192,"is_moderated":false}},{"id":"google/gemma-3-4b-it","name":"Google: Gemma 3 4B","pricing":{"prompt":"0.00000002","completion":"0.00000004","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1741905510,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"ai21/jamba-1.6-large","name":"AI21: Jamba 1.6 Large","pricing":{"prompt":"0.000002","completion":"0.000008","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1741905173,"top_provider":{"context_length":256000,"max_completion_tokens":4096,"is_moderated":false}},{"id":"ai21/jamba-1.6-mini","name":"AI21: Jamba Mini 1.6","pricing":{"prompt":"0.0000002","completion":"0.0000004","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1741905171,"top_provider":{"context_length":256000,"max_completion_tokens":4096,"is_moderated":false}},{"id":"google/gemma-3-12b-it:free","name":"Google: Gemma 3 12B (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1741902625,"top_provider":{"context_length":96000,"max_completion_tokens":8192,"is_moderated":false}},{"id":"google/gemma-3-12b-it","name":"Google: Gemma 3 12B","pricing":{"prompt":"0.0000000481286","completion":"0.000000192608","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1741902625,"top_provider":{"context_length":96000,"max_completion_tokens":8192,"is_moderated":false}},{"id":"cohere/command-a","name":"Cohere: Command A","pricing":{"prompt":"0.000002","completion":"0.000008","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1741894342,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"openai/gpt-4o-mini-search-preview","name":"OpenAI: GPT-4o-mini Search Preview","pricing":{"prompt":"0.00000015","completion":"0.0000006","request":"0.0275","image":"0.000217","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1741818122,"top_provider":{"context_length":128000,"max_completion_tokens":16384,"is_moderated":true}},{"id":"openai/gpt-4o-search-preview","name":"OpenAI: GPT-4o Search Preview","pricing":{"prompt":"0.0000025","completion":"0.00001","request":"0.035","image":"0.003613","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1741817949,"top_provider":{"context_length":128000,"max_completion_tokens":16384,"is_moderated":true}},{"id":"rekaai/reka-flash-3:free","name":"Reka: Flash 3 (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1741812813,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"google/gemma-3-27b-it:free","name":"Google: Gemma 3 27B (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1741756359,"top_provider":{"context_length":96000,"max_completion_tokens":8192,"is_moderated":false}},{"id":"google/gemma-3-27b-it","name":"Google: Gemma 3 27B","pricing":{"prompt":"0.0000000666396","completion":"0.000000266688","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1741756359,"top_provider":{"context_length":96000,"max_completion_tokens":8192,"is_moderated":false}},{"id":"thedrummer/anubis-pro-105b-v1","name":"TheDrummer: Anubis Pro 105B V1","pricing":{"prompt":"0.0000005","completion":"0.000001","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1741642290,"top_provider":{"context_length":131072,"max_completion_tokens":131072,"is_moderated":false}},{"id":"thedrummer/skyfall-36b-v2","name":"TheDrummer: Skyfall 36B V2","pricing":{"prompt":"0.0000000481286","completion":"0.000000192608","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1741636566,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"microsoft/phi-4-multimodal-instruct","name":"Microsoft: Phi 4 Multimodal Instruct","pricing":{"prompt":"0.00000005","completion":"0.0000001","request":"0","image":"0.00017685","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1741396284,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"perplexity/sonar-reasoning-pro","name":"Perplexity: Sonar Reasoning Pro","pricing":{"prompt":"0.000002","completion":"0.000008","request":"0","image":"0","audio":"0","web_search":"0.005","internal_reasoning":"0"},"created":1741313308,"top_provider":{"context_length":128000,"max_completion_tokens":null,"is_moderated":false}},{"id":"perplexity/sonar-pro","name":"Perplexity: Sonar Pro","pricing":{"prompt":"0.000003","completion":"0.000015","request":"0","image":"0","audio":"0","web_search":"0.005","internal_reasoning":"0"},"created":1741312423,"top_provider":{"context_length":200000,"max_completion_tokens":8000,"is_moderated":false}},{"id":"perplexity/sonar-deep-research","name":"Perplexity: Sonar Deep Research","pricing":{"prompt":"0.000002","completion":"0.000008","request":"0","image":"0","audio":"0","web_search":"0.005","internal_reasoning":"0.000003"},"created":1741311246,"top_provider":{"context_length":128000,"max_completion_tokens":null,"is_moderated":false}},{"id":"qwen/qwq-32b:free","name":"Qwen: QwQ 32B (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1741208814,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"qwen/qwq-32b","name":"Qwen: QwQ 32B","pricing":{"prompt":"0.000000075","completion":"0.00000015","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1741208814,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"nousresearch/deephermes-3-llama-3-8b-preview:free","name":"Nous: DeepHermes 3 Llama 3 8B Preview (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1740719372,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"google/gemini-2.0-flash-lite-001","name":"Google: Gemini 2.0 Flash Lite","pricing":{"prompt":"0.000000075","completion":"0.0000003","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1740506212,"top_provider":{"context_length":1048576,"max_completion_tokens":8192,"is_moderated":false}},{"id":"anthropic/claude-3.7-sonnet","name":"Anthropic: Claude 3.7 Sonnet","pricing":{"prompt":"0.000003","completion":"0.000015","request":"0","image":"0.0048","audio":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.0000003","input_cache_write":"0.00000375"},"created":1740422110,"top_provider":{"context_length":200000,"max_completion_tokens":64000,"is_moderated":false}},{"id":"anthropic/claude-3.7-sonnet:thinking","name":"Anthropic: Claude 3.7 Sonnet (thinking)","pricing":{"prompt":"0.000003","completion":"0.000015","request":"0","image":"0.0048","audio":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.0000003","input_cache_write":"0.00000375"},"created":1740422110,"top_provider":{"context_length":200000,"max_completion_tokens":64000,"is_moderated":false}},{"id":"anthropic/claude-3.7-sonnet:beta","name":"Anthropic: Claude 3.7 Sonnet (self-moderated)","pricing":{"prompt":"0.000003","completion":"0.000015","request":"0","image":"0.0048","audio":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.0000003","input_cache_write":"0.00000375"},"created":1740422110,"top_provider":{"context_length":200000,"max_completion_tokens":128000,"is_moderated":false}},{"id":"perplexity/r1-1776","name":"Perplexity: R1 1776","pricing":{"prompt":"0.000002","completion":"0.000008","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1740004929,"top_provider":{"context_length":128000,"max_completion_tokens":null,"is_moderated":false}},{"id":"mistralai/mistral-saba","name":"Mistral: Saba","pricing":{"prompt":"0.0000002","completion":"0.0000006","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1739803239,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"cognitivecomputations/dolphin3.0-r1-mistral-24b:free","name":"Dolphin3.0 R1 Mistral 24B (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1739462498,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"cognitivecomputations/dolphin3.0-r1-mistral-24b","name":"Dolphin3.0 R1 Mistral 24B","pricing":{"prompt":"0.00000001","completion":"0.0000000340768","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1739462498,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"cognitivecomputations/dolphin3.0-mistral-24b:free","name":"Dolphin3.0 Mistral 24B (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1739462019,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"cognitivecomputations/dolphin3.0-mistral-24b","name":"Dolphin3.0 Mistral 24B","pricing":{"prompt":"0.000000037022","completion":"0.00000014816","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1739462019,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"meta-llama/llama-guard-3-8b","name":"Llama Guard 3 8B","pricing":{"prompt":"0.00000002","completion":"0.00000006","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1739401318,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"openai/o3-mini-high","name":"OpenAI: o3 Mini High","pricing":{"prompt":"0.0000011","completion":"0.0000044","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.00000055"},"created":1739372611,"top_provider":{"context_length":200000,"max_completion_tokens":100000,"is_moderated":true}},{"id":"deepseek/deepseek-r1-distill-llama-8b","name":"DeepSeek: R1 Distill Llama 8B","pricing":{"prompt":"0.00000004","completion":"0.00000004","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1738937718,"top_provider":{"context_length":32000,"max_completion_tokens":32000,"is_moderated":false}},{"id":"google/gemini-2.0-flash-001","name":"Google: Gemini 2.0 Flash","pricing":{"prompt":"0.0000001","completion":"0.0000004","request":"0","image":"0.0000258","audio":"0.0000007","web_search":"0","internal_reasoning":"0","input_cache_read":"0.000000025","input_cache_write":"0.0000001833"},"created":1738769413,"top_provider":{"context_length":1048576,"max_completion_tokens":8192,"is_moderated":false}},{"id":"qwen/qwen-vl-plus","name":"Qwen: Qwen VL Plus","pricing":{"prompt":"0.00000021","completion":"0.00000063","request":"0","image":"0.0002688","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1738731255,"top_provider":{"context_length":7500,"max_completion_tokens":1500,"is_moderated":false}},{"id":"aion-labs/aion-1.0","name":"AionLabs: Aion-1.0","pricing":{"prompt":"0.000004","completion":"0.000008","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1738697557,"top_provider":{"context_length":131072,"max_completion_tokens":32768,"is_moderated":false}},{"id":"aion-labs/aion-1.0-mini","name":"AionLabs: Aion-1.0-Mini","pricing":{"prompt":"0.0000007","completion":"0.0000014","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1738697107,"top_provider":{"context_length":131072,"max_completion_tokens":32768,"is_moderated":false}},{"id":"aion-labs/aion-rp-llama-3.1-8b","name":"AionLabs: Aion-RP 1.0 (8B)","pricing":{"prompt":"0.0000002","completion":"0.0000002","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1738696718,"top_provider":{"context_length":32768,"max_completion_tokens":32768,"is_moderated":false}},{"id":"qwen/qwen-vl-max","name":"Qwen: Qwen VL Max","pricing":{"prompt":"0.0000008","completion":"0.0000032","request":"0","image":"0.001024","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1738434304,"top_provider":{"context_length":7500,"max_completion_tokens":1500,"is_moderated":false}},{"id":"qwen/qwen-turbo","name":"Qwen: Qwen-Turbo","pricing":{"prompt":"0.00000005","completion":"0.0000002","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.00000002"},"created":1738410974,"top_provider":{"context_length":1000000,"max_completion_tokens":8192,"is_moderated":false}},{"id":"qwen/qwen2.5-vl-72b-instruct:free","name":"Qwen: Qwen2.5 VL 72B Instruct (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1738410311,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"qwen/qwen2.5-vl-72b-instruct","name":"Qwen: Qwen2.5 VL 72B Instruct","pricing":{"prompt":"0.00000025","completion":"0.00000075","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1738410311,"top_provider":{"context_length":32000,"max_completion_tokens":null,"is_moderated":false}},{"id":"qwen/qwen-plus","name":"Qwen: Qwen-Plus","pricing":{"prompt":"0.0000004","completion":"0.0000012","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.00000016"},"created":1738409840,"top_provider":{"context_length":131072,"max_completion_tokens":8192,"is_moderated":false}},{"id":"qwen/qwen-max","name":"Qwen: Qwen-Max ","pricing":{"prompt":"0.0000016","completion":"0.0000064","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.00000064"},"created":1738402289,"top_provider":{"context_length":32768,"max_completion_tokens":8192,"is_moderated":false}},{"id":"openai/o3-mini","name":"OpenAI: o3 Mini","pricing":{"prompt":"0.0000011","completion":"0.0000044","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.00000055"},"created":1738351721,"top_provider":{"context_length":200000,"max_completion_tokens":100000,"is_moderated":true}},{"id":"deepseek/deepseek-r1-distill-qwen-1.5b","name":"DeepSeek: R1 Distill Qwen 1.5B","pricing":{"prompt":"0.00000018","completion":"0.00000018","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1738328067,"top_provider":{"context_length":131072,"max_completion_tokens":32768,"is_moderated":false}},{"id":"mistralai/mistral-small-24b-instruct-2501:free","name":"Mistral: Mistral Small 3 (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1738255409,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"mistralai/mistral-small-24b-instruct-2501","name":"Mistral: Mistral Small 3","pricing":{"prompt":"0.00000001999188","completion":"0.0000000800064","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1738255409,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"deepseek/deepseek-r1-distill-qwen-32b","name":"DeepSeek: R1 Distill Qwen 32B","pricing":{"prompt":"0.000000075","completion":"0.00000015","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1738194830,"top_provider":{"context_length":131072,"max_completion_tokens":16384,"is_moderated":false}},{"id":"deepseek/deepseek-r1-distill-qwen-14b:free","name":"DeepSeek: R1 Distill Qwen 14B (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1738193940,"top_provider":{"context_length":64000,"max_completion_tokens":null,"is_moderated":false}},{"id":"deepseek/deepseek-r1-distill-qwen-14b","name":"DeepSeek: R1 Distill Qwen 14B","pricing":{"prompt":"0.00000015","completion":"0.00000015","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1738193940,"top_provider":{"context_length":64000,"max_completion_tokens":32000,"is_moderated":false}},{"id":"perplexity/sonar-reasoning","name":"Perplexity: Sonar Reasoning","pricing":{"prompt":"0.000001","completion":"0.000005","request":"0.005","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1738131107,"top_provider":{"context_length":127000,"max_completion_tokens":null,"is_moderated":false}},{"id":"perplexity/sonar","name":"Perplexity: Sonar","pricing":{"prompt":"0.000001","completion":"0.000001","request":"0.005","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1738013808,"top_provider":{"context_length":127072,"max_completion_tokens":null,"is_moderated":false}},{"id":"liquid/lfm-7b","name":"Liquid: LFM 7B","pricing":{"prompt":"0.00000001","completion":"0.00000001","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1737806883,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"liquid/lfm-3b","name":"Liquid: LFM 3B","pricing":{"prompt":"0.00000002","completion":"0.00000002","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1737806501,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"deepseek/deepseek-r1-distill-llama-70b:free","name":"DeepSeek: R1 Distill Llama 70B (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1737663169,"top_provider":{"context_length":8192,"max_completion_tokens":4096,"is_moderated":false}},{"id":"deepseek/deepseek-r1-distill-llama-70b","name":"DeepSeek: R1 Distill Llama 70B","pricing":{"prompt":"0.0000000333198","completion":"0.000000133344","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1737663169,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"deepseek/deepseek-r1:free","name":"DeepSeek: R1 (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1737381095,"top_provider":{"context_length":163840,"max_completion_tokens":null,"is_moderated":false}},{"id":"deepseek/deepseek-r1","name":"DeepSeek: R1","pricing":{"prompt":"0.0000004","completion":"0.000002","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1737381095,"top_provider":{"context_length":163840,"max_completion_tokens":163840,"is_moderated":false}},{"id":"minimax/minimax-01","name":"MiniMax: MiniMax-01","pricing":{"prompt":"0.0000002","completion":"0.0000011","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1736915462,"top_provider":{"context_length":1000192,"max_completion_tokens":1000192,"is_moderated":false}},{"id":"mistralai/codestral-2501","name":"Mistral: Codestral 2501","pricing":{"prompt":"0.0000003","completion":"0.0000009","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1736895522,"top_provider":{"context_length":262144,"max_completion_tokens":null,"is_moderated":false}},{"id":"microsoft/phi-4","name":"Microsoft: Phi 4","pricing":{"prompt":"0.00000006","completion":"0.00000014","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1736489872,"top_provider":{"context_length":16384,"max_completion_tokens":null,"is_moderated":false}},{"id":"deepseek/deepseek-chat","name":"DeepSeek: DeepSeek V3","pricing":{"prompt":"0.00000017992692","completion":"0.0000007200576","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1735241320,"top_provider":{"context_length":163840,"max_completion_tokens":null,"is_moderated":false}},{"id":"sao10k/l3.3-euryale-70b","name":"Sao10K: Llama 3.3 Euryale 70B","pricing":{"prompt":"0.00000065","completion":"0.00000075","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1734535928,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"openai/o1","name":"OpenAI: o1","pricing":{"prompt":"0.000015","completion":"0.00006","request":"0","image":"0.021675","audio":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.0000075"},"created":1734459999,"top_provider":{"context_length":200000,"max_completion_tokens":100000,"is_moderated":true}},{"id":"x-ai/grok-2-vision-1212","name":"xAI: Grok 2 Vision 1212","pricing":{"prompt":"0.000002","completion":"0.00001","request":"0","image":"0.0036","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1734237338,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"x-ai/grok-2-1212","name":"xAI: Grok 2 1212","pricing":{"prompt":"0.000002","completion":"0.00001","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1734232814,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"cohere/command-r7b-12-2024","name":"Cohere: Command R7B (12-2024)","pricing":{"prompt":"0.0000000375","completion":"0.00000015","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1734158152,"top_provider":{"context_length":128000,"max_completion_tokens":4000,"is_moderated":true}},{"id":"google/gemini-2.0-flash-exp:free","name":"Google: Gemini 2.0 Flash Experimental (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1733937523,"top_provider":{"context_length":1048576,"max_completion_tokens":8192,"is_moderated":false}},{"id":"meta-llama/llama-3.3-70b-instruct:free","name":"Meta: Llama 3.3 70B Instruct (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1733506137,"top_provider":{"context_length":65536,"max_completion_tokens":null,"is_moderated":false}},{"id":"meta-llama/llama-3.3-70b-instruct","name":"Meta: Llama 3.3 70B Instruct","pricing":{"prompt":"0.000000038","completion":"0.00000012","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1733506137,"top_provider":{"context_length":131072,"max_completion_tokens":16384,"is_moderated":false}},{"id":"amazon/nova-lite-v1","name":"Amazon: Nova Lite 1.0","pricing":{"prompt":"0.00000006","completion":"0.00000024","request":"0","image":"0.00009","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1733437363,"top_provider":{"context_length":300000,"max_completion_tokens":5120,"is_moderated":true}},{"id":"amazon/nova-micro-v1","name":"Amazon: Nova Micro 1.0","pricing":{"prompt":"0.000000035","completion":"0.00000014","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1733437237,"top_provider":{"context_length":128000,"max_completion_tokens":5120,"is_moderated":true}},{"id":"amazon/nova-pro-v1","name":"Amazon: Nova Pro 1.0","pricing":{"prompt":"0.0000008","completion":"0.0000032","request":"0","image":"0.0012","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1733436303,"top_provider":{"context_length":300000,"max_completion_tokens":5120,"is_moderated":true}},{"id":"qwen/qwq-32b-preview","name":"Qwen: QwQ 32B Preview","pricing":{"prompt":"0.0000002","completion":"0.0000002","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1732754541,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"openai/gpt-4o-2024-11-20","name":"OpenAI: GPT-4o (2024-11-20)","pricing":{"prompt":"0.0000025","completion":"0.00001","request":"0","image":"0.003613","audio":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.00000125"},"created":1732127594,"top_provider":{"context_length":128000,"max_completion_tokens":16384,"is_moderated":true}},{"id":"mistralai/mistral-large-2411","name":"Mistral Large 2411","pricing":{"prompt":"0.000002","completion":"0.000006","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1731978685,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"mistralai/mistral-large-2407","name":"Mistral Large 2407","pricing":{"prompt":"0.000002","completion":"0.000006","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1731978415,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"mistralai/pixtral-large-2411","name":"Mistral: Pixtral Large 2411","pricing":{"prompt":"0.000002","completion":"0.000006","request":"0","image":"0.002888","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1731977388,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"x-ai/grok-vision-beta","name":"xAI: Grok Vision Beta","pricing":{"prompt":"0.000005","completion":"0.000015","request":"0","image":"0.009","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1731976624,"top_provider":{"context_length":8192,"max_completion_tokens":null,"is_moderated":false}},{"id":"infermatic/mn-inferor-12b","name":"Infermatic: Mistral Nemo Inferor 12B","pricing":{"prompt":"0.0000006","completion":"0.000001","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1731464428,"top_provider":{"context_length":8192,"max_completion_tokens":8192,"is_moderated":false}},{"id":"qwen/qwen-2.5-coder-32b-instruct:free","name":"Qwen2.5 Coder 32B Instruct (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1731368400,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"qwen/qwen-2.5-coder-32b-instruct","name":"Qwen2.5 Coder 32B Instruct","pricing":{"prompt":"0.0000000499797","completion":"0.000000200016","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1731368400,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"raifle/sorcererlm-8x22b","name":"SorcererLM 8x22B","pricing":{"prompt":"0.0000045","completion":"0.0000045","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1731105083,"top_provider":{"context_length":16000,"max_completion_tokens":null,"is_moderated":false}},{"id":"thedrummer/unslopnemo-12b","name":"TheDrummer: UnslopNemo 12B","pricing":{"prompt":"0.0000004","completion":"0.0000004","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1731103448,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"anthropic/claude-3.5-haiku:beta","name":"Anthropic: Claude 3.5 Haiku (self-moderated)","pricing":{"prompt":"0.0000008","completion":"0.000004","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.00000008","input_cache_write":"0.000001"},"created":1730678400,"top_provider":{"context_length":200000,"max_completion_tokens":8192,"is_moderated":false}},{"id":"anthropic/claude-3.5-haiku","name":"Anthropic: Claude 3.5 Haiku","pricing":{"prompt":"0.0000008","completion":"0.000004","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.00000008","input_cache_write":"0.000001"},"created":1730678400,"top_provider":{"context_length":200000,"max_completion_tokens":8192,"is_moderated":true}},{"id":"anthropic/claude-3.5-haiku-20241022","name":"Anthropic: Claude 3.5 Haiku (2024-10-22)","pricing":{"prompt":"0.0000008","completion":"0.000004","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.00000008","input_cache_write":"0.000001"},"created":1730678400,"top_provider":{"context_length":200000,"max_completion_tokens":8192,"is_moderated":false}},{"id":"anthracite-org/magnum-v4-72b","name":"Magnum v4 72B","pricing":{"prompt":"0.0000025","completion":"0.000003","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1729555200,"top_provider":{"context_length":16384,"max_completion_tokens":1024,"is_moderated":false}},{"id":"anthropic/claude-3.5-sonnet:beta","name":"Anthropic: Claude 3.5 Sonnet (self-moderated)","pricing":{"prompt":"0.000003","completion":"0.000015","request":"0","image":"0.0048","audio":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.0000003","input_cache_write":"0.00000375"},"created":1729555200,"top_provider":{"context_length":200000,"max_completion_tokens":8192,"is_moderated":false}},{"id":"anthropic/claude-3.5-sonnet","name":"Anthropic: Claude 3.5 Sonnet","pricing":{"prompt":"0.000003","completion":"0.000015","request":"0","image":"0.0048","audio":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.0000003","input_cache_write":"0.00000375"},"created":1729555200,"top_provider":{"context_length":200000,"max_completion_tokens":8192,"is_moderated":true}},{"id":"mistralai/ministral-8b","name":"Mistral: Ministral 8B","pricing":{"prompt":"0.0000001","completion":"0.0000001","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1729123200,"top_provider":{"context_length":128000,"max_completion_tokens":null,"is_moderated":false}},{"id":"mistralai/ministral-3b","name":"Mistral: Ministral 3B","pricing":{"prompt":"0.00000004","completion":"0.00000004","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1729123200,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"qwen/qwen-2.5-7b-instruct","name":"Qwen2.5 7B Instruct","pricing":{"prompt":"0.00000004","completion":"0.0000001","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1729036800,"top_provider":{"context_length":65536,"max_completion_tokens":null,"is_moderated":false}},{"id":"nvidia/llama-3.1-nemotron-70b-instruct","name":"NVIDIA: Llama 3.1 Nemotron 70B Instruct","pricing":{"prompt":"0.00000012","completion":"0.0000003","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1728950400,"top_provider":{"context_length":131072,"max_completion_tokens":131072,"is_moderated":false}},{"id":"inflection/inflection-3-productivity","name":"Inflection: Inflection 3 Productivity","pricing":{"prompt":"0.0000025","completion":"0.00001","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1728604800,"top_provider":{"context_length":8000,"max_completion_tokens":1024,"is_moderated":false}},{"id":"inflection/inflection-3-pi","name":"Inflection: Inflection 3 Pi","pricing":{"prompt":"0.0000025","completion":"0.00001","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1728604800,"top_provider":{"context_length":8000,"max_completion_tokens":1024,"is_moderated":false}},{"id":"google/gemini-flash-1.5-8b","name":"Google: Gemini 1.5 Flash 8B","pricing":{"prompt":"0.0000000375","completion":"0.00000015","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.00000001","input_cache_write":"0.0000000583"},"created":1727913600,"top_provider":{"context_length":1000000,"max_completion_tokens":8192,"is_moderated":false}},{"id":"thedrummer/rocinante-12b","name":"TheDrummer: Rocinante 12B","pricing":{"prompt":"0.00000019","completion":"0.00000045","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1727654400,"top_provider":{"context_length":8192,"max_completion_tokens":8192,"is_moderated":false}},{"id":"liquid/lfm-40b","name":"Liquid: LFM 40B MoE","pricing":{"prompt":"0.00000015","completion":"0.00000015","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1727654400,"top_provider":{"context_length":65536,"max_completion_tokens":65536,"is_moderated":false}},{"id":"anthracite-org/magnum-v2-72b","name":"Magnum v2 72B","pricing":{"prompt":"0.000003","completion":"0.000003","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1727654400,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"meta-llama/llama-3.2-3b-instruct:free","name":"Meta: Llama 3.2 3B Instruct (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1727222400,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"meta-llama/llama-3.2-3b-instruct","name":"Meta: Llama 3.2 3B Instruct","pricing":{"prompt":"0.000000003","completion":"0.000000006","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1727222400,"top_provider":{"context_length":20000,"max_completion_tokens":20000,"is_moderated":false}},{"id":"meta-llama/llama-3.2-90b-vision-instruct","name":"Meta: Llama 3.2 90B Vision Instruct","pricing":{"prompt":"0.0000012","completion":"0.0000012","request":"0","image":"0.001734","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1727222400,"top_provider":{"context_length":131072,"max_completion_tokens":2048,"is_moderated":false}},{"id":"meta-llama/llama-3.2-1b-instruct","name":"Meta: Llama 3.2 1B Instruct","pricing":{"prompt":"0.000000005","completion":"0.00000001","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1727222400,"top_provider":{"context_length":131072,"max_completion_tokens":16384,"is_moderated":false}},{"id":"meta-llama/llama-3.2-11b-vision-instruct:free","name":"Meta: Llama 3.2 11B Vision Instruct (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1727222400,"top_provider":{"context_length":131072,"max_completion_tokens":2048,"is_moderated":false}},{"id":"meta-llama/llama-3.2-11b-vision-instruct","name":"Meta: Llama 3.2 11B Vision Instruct","pricing":{"prompt":"0.000000049","completion":"0.000000049","request":"0","image":"0.00007948","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1727222400,"top_provider":{"context_length":131072,"max_completion_tokens":16384,"is_moderated":false}},{"id":"qwen/qwen-2.5-72b-instruct:free","name":"Qwen2.5 72B Instruct (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1726704000,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"qwen/qwen-2.5-72b-instruct","name":"Qwen2.5 72B Instruct","pricing":{"prompt":"0.0000000666396","completion":"0.000000266688","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1726704000,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"neversleep/llama-3.1-lumimaid-8b","name":"NeverSleep: Lumimaid v0.2 8B","pricing":{"prompt":"0.0000001","completion":"0.0000009","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1726358400,"top_provider":{"context_length":8192,"max_completion_tokens":8192,"is_moderated":false}},{"id":"openai/o1-mini-2024-09-12","name":"OpenAI: o1-mini (2024-09-12)","pricing":{"prompt":"0.0000011","completion":"0.0000044","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.00000055"},"created":1726099200,"top_provider":{"context_length":128000,"max_completion_tokens":65536,"is_moderated":true}},{"id":"openai/o1-mini","name":"OpenAI: o1-mini","pricing":{"prompt":"0.0000011","completion":"0.0000044","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.00000055"},"created":1726099200,"top_provider":{"context_length":128000,"max_completion_tokens":65536,"is_moderated":true}},{"id":"mistralai/pixtral-12b","name":"Mistral: Pixtral 12B","pricing":{"prompt":"0.0000001","completion":"0.0000001","request":"0","image":"0.0001445","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1725926400,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"cohere/command-r-08-2024","name":"Cohere: Command R (08-2024)","pricing":{"prompt":"0.00000015","completion":"0.0000006","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1724976000,"top_provider":{"context_length":128000,"max_completion_tokens":4000,"is_moderated":true}},{"id":"cohere/command-r-plus-08-2024","name":"Cohere: Command R+ (08-2024)","pricing":{"prompt":"0.0000025","completion":"0.00001","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1724976000,"top_provider":{"context_length":128000,"max_completion_tokens":4000,"is_moderated":true}},{"id":"qwen/qwen-2.5-vl-7b-instruct","name":"Qwen: Qwen2.5-VL 7B Instruct","pricing":{"prompt":"0.0000002","completion":"0.0000002","request":"0","image":"0.0001445","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1724803200,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"sao10k/l3.1-euryale-70b","name":"Sao10K: Llama 3.1 Euryale 70B v2.2","pricing":{"prompt":"0.00000065","completion":"0.00000075","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1724803200,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"microsoft/phi-3.5-mini-128k-instruct","name":"Microsoft: Phi-3.5 Mini 128K Instruct","pricing":{"prompt":"0.0000001","completion":"0.0000001","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1724198400,"top_provider":{"context_length":128000,"max_completion_tokens":null,"is_moderated":false}},{"id":"nousresearch/hermes-3-llama-3.1-70b","name":"Nous: Hermes 3 70B Instruct","pricing":{"prompt":"0.0000001","completion":"0.00000028","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1723939200,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"nousresearch/hermes-3-llama-3.1-405b","name":"Nous: Hermes 3 405B Instruct","pricing":{"prompt":"0.0000007","completion":"0.0000008","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1723766400,"top_provider":{"context_length":131072,"max_completion_tokens":16384,"is_moderated":false}},{"id":"openai/chatgpt-4o-latest","name":"OpenAI: ChatGPT-4o","pricing":{"prompt":"0.000005","completion":"0.000015","request":"0","image":"0.007225","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1723593600,"top_provider":{"context_length":128000,"max_completion_tokens":16384,"is_moderated":true}},{"id":"sao10k/l3-lunaris-8b","name":"Sao10K: Llama 3 8B Lunaris","pricing":{"prompt":"0.00000002","completion":"0.00000005","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1723507200,"top_provider":{"context_length":8192,"max_completion_tokens":null,"is_moderated":false}},{"id":"openai/gpt-4o-2024-08-06","name":"OpenAI: GPT-4o (2024-08-06)","pricing":{"prompt":"0.0000025","completion":"0.00001","request":"0","image":"0.003613","audio":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.00000125"},"created":1722902400,"top_provider":{"context_length":128000,"max_completion_tokens":16384,"is_moderated":false}},{"id":"meta-llama/llama-3.1-405b","name":"Meta: Llama 3.1 405B (base)","pricing":{"prompt":"0.000002","completion":"0.000002","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1722556800,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"meta-llama/llama-3.1-70b-instruct","name":"Meta: Llama 3.1 70B Instruct","pricing":{"prompt":"0.0000001","completion":"0.00000028","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1721692800,"top_provider":{"context_length":131072,"max_completion_tokens":16384,"is_moderated":false}},{"id":"meta-llama/llama-3.1-405b-instruct:free","name":"Meta: Llama 3.1 405B Instruct (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1721692800,"top_provider":{"context_length":65536,"max_completion_tokens":null,"is_moderated":false}},{"id":"meta-llama/llama-3.1-405b-instruct","name":"Meta: Llama 3.1 405B Instruct","pricing":{"prompt":"0.0000008","completion":"0.0000008","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1721692800,"top_provider":{"context_length":32768,"max_completion_tokens":16384,"is_moderated":false}},{"id":"meta-llama/llama-3.1-8b-instruct","name":"Meta: Llama 3.1 8B Instruct","pricing":{"prompt":"0.000000015","completion":"0.00000002","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1721692800,"top_provider":{"context_length":131072,"max_completion_tokens":16384,"is_moderated":false}},{"id":"mistralai/mistral-nemo:free","name":"Mistral: Mistral Nemo (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1721347200,"top_provider":{"context_length":131072,"max_completion_tokens":128000,"is_moderated":false}},{"id":"mistralai/mistral-nemo","name":"Mistral: Mistral Nemo","pricing":{"prompt":"0.0000000075","completion":"0.00000005","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1721347200,"top_provider":{"context_length":32000,"max_completion_tokens":null,"is_moderated":false}},{"id":"openai/gpt-4o-mini-2024-07-18","name":"OpenAI: GPT-4o-mini (2024-07-18)","pricing":{"prompt":"0.00000015","completion":"0.0000006","request":"0","image":"0.007225","audio":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.000000075"},"created":1721260800,"top_provider":{"context_length":128000,"max_completion_tokens":16384,"is_moderated":true}},{"id":"openai/gpt-4o-mini","name":"OpenAI: GPT-4o-mini","pricing":{"prompt":"0.00000015","completion":"0.0000006","request":"0","image":"0.000217","audio":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.000000075"},"created":1721260800,"top_provider":{"context_length":128000,"max_completion_tokens":16384,"is_moderated":true}},{"id":"google/gemma-2-27b-it","name":"Google: Gemma 2 27B","pricing":{"prompt":"0.00000065","completion":"0.00000065","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1720828800,"top_provider":{"context_length":8192,"max_completion_tokens":null,"is_moderated":false}},{"id":"google/gemma-2-9b-it:free","name":"Google: Gemma 2 9B (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1719532800,"top_provider":{"context_length":8192,"max_completion_tokens":8192,"is_moderated":false}},{"id":"google/gemma-2-9b-it","name":"Google: Gemma 2 9B","pricing":{"prompt":"0.00000001","completion":"0.0000000100008","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1719532800,"top_provider":{"context_length":8192,"max_completion_tokens":8192,"is_moderated":false}},{"id":"anthropic/claude-3.5-sonnet-20240620:beta","name":"Anthropic: Claude 3.5 Sonnet (2024-06-20) (self-moderated)","pricing":{"prompt":"0.000003","completion":"0.000015","request":"0","image":"0.0048","audio":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.0000003","input_cache_write":"0.00000375"},"created":1718841600,"top_provider":{"context_length":200000,"max_completion_tokens":8192,"is_moderated":false}},{"id":"anthropic/claude-3.5-sonnet-20240620","name":"Anthropic: Claude 3.5 Sonnet (2024-06-20)","pricing":{"prompt":"0.000003","completion":"0.000015","request":"0","image":"0.0048","audio":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.0000003","input_cache_write":"0.00000375"},"created":1718841600,"top_provider":{"context_length":200000,"max_completion_tokens":8192,"is_moderated":true}},{"id":"sao10k/l3-euryale-70b","name":"Sao10k: Llama 3 Euryale 70B v2.1","pricing":{"prompt":"0.00000148","completion":"0.00000148","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1718668800,"top_provider":{"context_length":8192,"max_completion_tokens":8192,"is_moderated":false}},{"id":"cognitivecomputations/dolphin-mixtral-8x22b","name":"Dolphin 2.9.2 Mixtral 8x22B 🐬","pricing":{"prompt":"0.0000009","completion":"0.0000009","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1717804800,"top_provider":{"context_length":16000,"max_completion_tokens":8192,"is_moderated":false}},{"id":"qwen/qwen-2-72b-instruct","name":"Qwen 2 72B Instruct","pricing":{"prompt":"0.0000009","completion":"0.0000009","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1717718400,"top_provider":{"context_length":32768,"max_completion_tokens":4096,"is_moderated":false}},{"id":"mistralai/mistral-7b-instruct-v0.3","name":"Mistral: Mistral 7B Instruct v0.3","pricing":{"prompt":"0.000000028","completion":"0.000000054","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1716768000,"top_provider":{"context_length":32768,"max_completion_tokens":16384,"is_moderated":false}},{"id":"nousresearch/hermes-2-pro-llama-3-8b","name":"NousResearch: Hermes 2 Pro - Llama-3 8B","pricing":{"prompt":"0.000000025","completion":"0.00000004","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1716768000,"top_provider":{"context_length":131072,"max_completion_tokens":131072,"is_moderated":false}},{"id":"mistralai/mistral-7b-instruct:free","name":"Mistral: Mistral 7B Instruct (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1716768000,"top_provider":{"context_length":32768,"max_completion_tokens":16384,"is_moderated":false}},{"id":"mistralai/mistral-7b-instruct","name":"Mistral: Mistral 7B Instruct","pricing":{"prompt":"0.000000028","completion":"0.000000054","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1716768000,"top_provider":{"context_length":32768,"max_completion_tokens":16384,"is_moderated":false}},{"id":"microsoft/phi-3-mini-128k-instruct","name":"Microsoft: Phi-3 Mini 128K Instruct","pricing":{"prompt":"0.0000001","completion":"0.0000001","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1716681600,"top_provider":{"context_length":128000,"max_completion_tokens":null,"is_moderated":false}},{"id":"microsoft/phi-3-medium-128k-instruct","name":"Microsoft: Phi-3 Medium 128K Instruct","pricing":{"prompt":"0.000001","completion":"0.000001","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1716508800,"top_provider":{"context_length":128000,"max_completion_tokens":null,"is_moderated":false}},{"id":"neversleep/llama-3-lumimaid-70b","name":"NeverSleep: Llama 3 Lumimaid 70B","pricing":{"prompt":"0.000004","completion":"0.000006","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1715817600,"top_provider":{"context_length":8192,"max_completion_tokens":4096,"is_moderated":false}},{"id":"google/gemini-flash-1.5","name":"Google: Gemini 1.5 Flash ","pricing":{"prompt":"0.000000075","completion":"0.0000003","request":"0","image":"0.00004","audio":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.00000001875","input_cache_write":"0.0000001583"},"created":1715644800,"top_provider":{"context_length":1000000,"max_completion_tokens":8192,"is_moderated":false}},{"id":"meta-llama/llama-guard-2-8b","name":"Meta: LlamaGuard 2 8B","pricing":{"prompt":"0.0000002","completion":"0.0000002","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1715558400,"top_provider":{"context_length":8192,"max_completion_tokens":null,"is_moderated":false}},{"id":"openai/gpt-4o-2024-05-13","name":"OpenAI: GPT-4o (2024-05-13)","pricing":{"prompt":"0.000005","completion":"0.000015","request":"0","image":"0.007225","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1715558400,"top_provider":{"context_length":128000,"max_completion_tokens":4096,"is_moderated":true}},{"id":"openai/gpt-4o","name":"OpenAI: GPT-4o","pricing":{"prompt":"0.0000025","completion":"0.00001","request":"0","image":"0.003613","audio":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.00000125"},"created":1715558400,"top_provider":{"context_length":128000,"max_completion_tokens":16384,"is_moderated":true}},{"id":"openai/gpt-4o:extended","name":"OpenAI: GPT-4o (extended)","pricing":{"prompt":"0.000006","completion":"0.000018","request":"0","image":"0.007225","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1715558400,"top_provider":{"context_length":128000,"max_completion_tokens":64000,"is_moderated":true}},{"id":"sao10k/fimbulvetr-11b-v2","name":"Fimbulvetr 11B v2","pricing":{"prompt":"0.0000008","completion":"0.0000012","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1713657600,"top_provider":{"context_length":4096,"max_completion_tokens":4096,"is_moderated":false}},{"id":"meta-llama/llama-3-8b-instruct","name":"Meta: Llama 3 8B Instruct","pricing":{"prompt":"0.00000003","completion":"0.00000006","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1713398400,"top_provider":{"context_length":8192,"max_completion_tokens":16384,"is_moderated":false}},{"id":"meta-llama/llama-3-70b-instruct","name":"Meta: Llama 3 70B Instruct","pricing":{"prompt":"0.0000003","completion":"0.0000004","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1713398400,"top_provider":{"context_length":8192,"max_completion_tokens":16384,"is_moderated":false}},{"id":"mistralai/mixtral-8x22b-instruct","name":"Mistral: Mixtral 8x22B Instruct","pricing":{"prompt":"0.0000009","completion":"0.0000009","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1713312000,"top_provider":{"context_length":65536,"max_completion_tokens":null,"is_moderated":false}},{"id":"microsoft/wizardlm-2-8x22b","name":"WizardLM-2 8x22B","pricing":{"prompt":"0.00000048","completion":"0.00000048","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1713225600,"top_provider":{"context_length":65536,"max_completion_tokens":65536,"is_moderated":false}},{"id":"google/gemini-pro-1.5","name":"Google: Gemini 1.5 Pro","pricing":{"prompt":"0.00000125","completion":"0.000005","request":"0","image":"0.0006575","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1712620800,"top_provider":{"context_length":2000000,"max_completion_tokens":8192,"is_moderated":false}},{"id":"openai/gpt-4-turbo","name":"OpenAI: GPT-4 Turbo","pricing":{"prompt":"0.00001","completion":"0.00003","request":"0","image":"0.01445","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1712620800,"top_provider":{"context_length":128000,"max_completion_tokens":4096,"is_moderated":true}},{"id":"cohere/command-r-plus","name":"Cohere: Command R+","pricing":{"prompt":"0.000003","completion":"0.000015","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1712188800,"top_provider":{"context_length":128000,"max_completion_tokens":4000,"is_moderated":true}},{"id":"cohere/command-r-plus-04-2024","name":"Cohere: Command R+ (04-2024)","pricing":{"prompt":"0.000003","completion":"0.000015","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1712016000,"top_provider":{"context_length":128000,"max_completion_tokens":4000,"is_moderated":true}},{"id":"sophosympatheia/midnight-rose-70b","name":"Midnight Rose 70B","pricing":{"prompt":"0.0000008","completion":"0.0000008","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1711065600,"top_provider":{"context_length":4096,"max_completion_tokens":2048,"is_moderated":false}},{"id":"cohere/command","name":"Cohere: Command","pricing":{"prompt":"0.000001","completion":"0.000002","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1710374400,"top_provider":{"context_length":4096,"max_completion_tokens":4000,"is_moderated":true}},{"id":"cohere/command-r","name":"Cohere: Command R","pricing":{"prompt":"0.0000005","completion":"0.0000015","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1710374400,"top_provider":{"context_length":128000,"max_completion_tokens":4000,"is_moderated":true}},{"id":"anthropic/claude-3-haiku:beta","name":"Anthropic: Claude 3 Haiku (self-moderated)","pricing":{"prompt":"0.00000025","completion":"0.00000125","request":"0","image":"0.0004","audio":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.00000003","input_cache_write":"0.0000003"},"created":1710288000,"top_provider":{"context_length":200000,"max_completion_tokens":4096,"is_moderated":false}},{"id":"anthropic/claude-3-haiku","name":"Anthropic: Claude 3 Haiku","pricing":{"prompt":"0.00000025","completion":"0.00000125","request":"0","image":"0.0004","audio":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.00000003","input_cache_write":"0.0000003"},"created":1710288000,"top_provider":{"context_length":200000,"max_completion_tokens":4096,"is_moderated":true}},{"id":"anthropic/claude-3-opus:beta","name":"Anthropic: Claude 3 Opus (self-moderated)","pricing":{"prompt":"0.000015","completion":"0.000075","request":"0","image":"0.024","audio":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.0000015","input_cache_write":"0.00001875"},"created":1709596800,"top_provider":{"context_length":200000,"max_completion_tokens":4096,"is_moderated":false}},{"id":"anthropic/claude-3-opus","name":"Anthropic: Claude 3 Opus","pricing":{"prompt":"0.000015","completion":"0.000075","request":"0","image":"0.024","audio":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.0000015","input_cache_write":"0.00001875"},"created":1709596800,"top_provider":{"context_length":200000,"max_completion_tokens":4096,"is_moderated":true}},{"id":"cohere/command-r-03-2024","name":"Cohere: Command R (03-2024)","pricing":{"prompt":"0.0000005","completion":"0.0000015","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1709341200,"top_provider":{"context_length":128000,"max_completion_tokens":4000,"is_moderated":true}},{"id":"mistralai/mistral-large","name":"Mistral Large","pricing":{"prompt":"0.000002","completion":"0.000006","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1708905600,"top_provider":{"context_length":128000,"max_completion_tokens":null,"is_moderated":false}},{"id":"openai/gpt-3.5-turbo-0613","name":"OpenAI: GPT-3.5 Turbo (older v0613)","pricing":{"prompt":"0.000001","completion":"0.000002","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1706140800,"top_provider":{"context_length":4095,"max_completion_tokens":4096,"is_moderated":false}},{"id":"openai/gpt-4-turbo-preview","name":"OpenAI: GPT-4 Turbo Preview","pricing":{"prompt":"0.00001","completion":"0.00003","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1706140800,"top_provider":{"context_length":128000,"max_completion_tokens":4096,"is_moderated":true}},{"id":"nousresearch/nous-hermes-2-mixtral-8x7b-dpo","name":"Nous: Hermes 2 Mixtral 8x7B DPO","pricing":{"prompt":"0.0000006","completion":"0.0000006","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1705363200,"top_provider":{"context_length":32768,"max_completion_tokens":2048,"is_moderated":false}},{"id":"mistralai/mistral-small","name":"Mistral Small","pricing":{"prompt":"0.0000002","completion":"0.0000006","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1704844800,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"mistralai/mistral-tiny","name":"Mistral Tiny","pricing":{"prompt":"0.00000025","completion":"0.00000025","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1704844800,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"mistralai/mistral-7b-instruct-v0.2","name":"Mistral: Mistral 7B Instruct v0.2","pricing":{"prompt":"0.0000002","completion":"0.0000002","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1703721600,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"mistralai/mixtral-8x7b-instruct","name":"Mistral: Mixtral 8x7B Instruct","pricing":{"prompt":"0.00000008","completion":"0.00000024","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1702166400,"top_provider":{"context_length":32768,"max_completion_tokens":16384,"is_moderated":false}},{"id":"neversleep/noromaid-20b","name":"Noromaid 20B","pricing":{"prompt":"0.000001","completion":"0.00000175","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1700956800,"top_provider":{"context_length":4096,"max_completion_tokens":null,"is_moderated":false}},{"id":"alpindale/goliath-120b","name":"Goliath 120B","pricing":{"prompt":"0.000009","completion":"0.000011","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1699574400,"top_provider":{"context_length":6144,"max_completion_tokens":null,"is_moderated":false}},{"id":"undi95/toppy-m-7b","name":"Toppy M 7B","pricing":{"prompt":"0.0000008","completion":"0.0000012","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1699574400,"top_provider":{"context_length":4096,"max_completion_tokens":4096,"is_moderated":false}},{"id":"openrouter/auto","name":"Auto Router","pricing":{"prompt":"-1","completion":"-1"},"created":1699401600,"top_provider":{"context_length":null,"max_completion_tokens":null,"is_moderated":false}},{"id":"openai/gpt-4-1106-preview","name":"OpenAI: GPT-4 Turbo (older v1106)","pricing":{"prompt":"0.00001","completion":"0.00003","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1699228800,"top_provider":{"context_length":128000,"max_completion_tokens":4096,"is_moderated":true}},{"id":"mistralai/mistral-7b-instruct-v0.1","name":"Mistral: Mistral 7B Instruct v0.1","pricing":{"prompt":"0.00000011","completion":"0.00000019","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1695859200,"top_provider":{"context_length":2824,"max_completion_tokens":null,"is_moderated":false}},{"id":"openai/gpt-3.5-turbo-instruct","name":"OpenAI: GPT-3.5 Turbo Instruct","pricing":{"prompt":"0.0000015","completion":"0.000002","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1695859200,"top_provider":{"context_length":4095,"max_completion_tokens":4096,"is_moderated":true}},{"id":"pygmalionai/mythalion-13b","name":"Pygmalion: Mythalion 13B","pricing":{"prompt":"0.0000006","completion":"0.000001","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1693612800,"top_provider":{"context_length":4096,"max_completion_tokens":4096,"is_moderated":false}},{"id":"openai/gpt-3.5-turbo-16k","name":"OpenAI: GPT-3.5 Turbo 16k","pricing":{"prompt":"0.000003","completion":"0.000004","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1693180800,"top_provider":{"context_length":16385,"max_completion_tokens":4096,"is_moderated":true}},{"id":"mancer/weaver","name":"Mancer: Weaver (alpha)","pricing":{"prompt":"0.0000015","completion":"0.0000015","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1690934400,"top_provider":{"context_length":8000,"max_completion_tokens":1000,"is_moderated":false}},{"id":"undi95/remm-slerp-l2-13b","name":"ReMM SLERP 13B","pricing":{"prompt":"0.0000007","completion":"0.000001","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1689984000,"top_provider":{"context_length":6144,"max_completion_tokens":null,"is_moderated":false}},{"id":"gryphe/mythomax-l2-13b","name":"MythoMax 13B","pricing":{"prompt":"0.00000006","completion":"0.00000006","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1688256000,"top_provider":{"context_length":4096,"max_completion_tokens":null,"is_moderated":false}},{"id":"openai/gpt-4","name":"OpenAI: GPT-4","pricing":{"prompt":"0.00003","completion":"0.00006","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1685232000,"top_provider":{"context_length":8191,"max_completion_tokens":4096,"is_moderated":true}},{"id":"openai/gpt-3.5-turbo","name":"OpenAI: GPT-3.5 Turbo","pricing":{"prompt":"0.0000005","completion":"0.0000015","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1685232000,"top_provider":{"context_length":16385,"max_completion_tokens":4096,"is_moderated":true}},{"id":"openai/gpt-4-0314","name":"OpenAI: GPT-4 (older v0314)","pricing":{"prompt":"0.00003","completion":"0.00006","request":"0","image":"0","audio":"0","web_search":"0","internal_reasoning":"0"},"created":1685232000,"top_provider":{"context_length":8191,"max_completion_tokens":4096,"is_moderated":true}}] \ No newline at end of file diff --git a/packages/kbot/schema.json b/packages/kbot/schema.json index 03ca7bf6..bc3ff392 100644 --- a/packages/kbot/schema.json +++ b/packages/kbot/schema.json @@ -118,7 +118,7 @@ }, "model": { "type": "string", - "description": "AI model to use for processing. Available models:\n\u001b[35m\u001b[1m\u001b[22m\u001b[39m\n\u001b[35m\u001b[1m OpenRouter models:\u001b[22m\u001b[39m\n\u001b[35m\u001b[1m\u001b[22m\u001b[39m\n01-ai/yi-large | paid\naetherwiing/mn-starcannon-12b | paid\nagentica-org/deepcoder-14b-preview | paid\nagentica-org/deepcoder-14b-preview:free | free\nai21/jamba-1.6-large | paid\nai21/jamba-1.6-mini | paid\naion-labs/aion-1.0 | paid\naion-labs/aion-1.0-mini | paid\naion-labs/aion-rp-llama-3.1-8b | paid\nalfredpros/codellama-7b-instruct-solidity | paid\namazon/nova-lite-v1 | paid\namazon/nova-micro-v1 | paid\namazon/nova-pro-v1 | paid\nanthropic/claude-3-haiku | paid\nanthropic/claude-3-haiku:beta | paid\nanthropic/claude-3-opus | paid\nanthropic/claude-3-opus:beta | paid\nanthropic/claude-3-sonnet | paid\nanthropic/claude-3-sonnet:beta | paid\nanthropic/claude-3.5-haiku | paid\nanthropic/claude-3.5-haiku-20241022 | paid\nanthropic/claude-3.5-haiku-20241022:beta | paid\nanthropic/claude-3.5-haiku:beta | paid\nanthropic/claude-3.5-sonnet | paid\nanthropic/claude-3.5-sonnet-20240620 | paid\nanthropic/claude-3.5-sonnet-20240620:beta | paid\nanthropic/claude-3.5-sonnet:beta | paid\nanthropic/claude-3.7-sonnet | paid\nanthropic/claude-3.7-sonnet:beta | paid\nanthropic/claude-3.7-sonnet:thinking | paid\nanthropic/claude-opus-4 | paid\nanthropic/claude-sonnet-4 | paid\nanthropic/claude-2 | paid\nanthropic/claude-2:beta | paid\nanthropic/claude-2.0 | paid\nanthropic/claude-2.0:beta | paid\nanthropic/claude-2.1 | paid\nanthropic/claude-2.1:beta | paid\narcee-ai/arcee-blitz | paid\narcee-ai/caller-large | paid\narcee-ai/coder-large | paid\narcee-ai/maestro-reasoning | paid\narcee-ai/spotlight | paid\narcee-ai/virtuoso-large | paid\narcee-ai/virtuoso-medium-v2 | paid\narliai/qwq-32b-arliai-rpr-v1:free | free\nopenrouter/auto | paid\nbaidu/ernie-4.5-300b-a47b | paid\ncohere/command | paid\ncohere/command-a | paid\ncohere/command-r | paid\ncohere/command-r-03-2024 | paid\ncohere/command-r-08-2024 | paid\ncohere/command-r-plus | paid\ncohere/command-r-plus-04-2024 | paid\ncohere/command-r-plus-08-2024 | paid\ncohere/command-r7b-12-2024 | paid\ndeepseek/deepseek-prover-v2 | paid\ndeepseek/deepseek-r1-0528-qwen3-8b | paid\ndeepseek/deepseek-r1-0528-qwen3-8b:free | free\ndeepseek/deepseek-chat | paid\ndeepseek/deepseek-chat:free | free\ndeepseek/deepseek-chat-v3-0324 | paid\ndeepseek/deepseek-chat-v3-0324:free | free\ndeepseek/deepseek-v3-base:free | free\ndeepseek/deepseek-r1 | paid\ndeepseek/deepseek-r1:free | free\ndeepseek/deepseek-r1-0528 | paid\ndeepseek/deepseek-r1-0528:free | free\ndeepseek/deepseek-r1-distill-llama-70b | paid\ndeepseek/deepseek-r1-distill-llama-70b:free | free\ndeepseek/deepseek-r1-distill-llama-8b | paid\ndeepseek/deepseek-r1-distill-qwen-1.5b | paid\ndeepseek/deepseek-r1-distill-qwen-14b | paid\ndeepseek/deepseek-r1-distill-qwen-14b:free | free\ndeepseek/deepseek-r1-distill-qwen-32b | paid\ndeepseek/deepseek-r1-distill-qwen-7b | paid\ncognitivecomputations/dolphin-mixtral-8x22b | paid\ncognitivecomputations/dolphin3.0-mistral-24b:free | free\ncognitivecomputations/dolphin3.0-r1-mistral-24b | paid\ncognitivecomputations/dolphin3.0-r1-mistral-24b:free | free\neleutherai/llemma_7b | paid\neva-unit-01/eva-llama-3.33-70b | paid\neva-unit-01/eva-qwen-2.5-72b | paid\nsao10k/fimbulvetr-11b-v2 | paid\nalpindale/goliath-120b | paid\ngoogle/gemini-flash-1.5 | paid\ngoogle/gemini-flash-1.5-8b | paid\ngoogle/gemini-pro-1.5 | paid\ngoogle/gemini-2.0-flash-001 | paid\ngoogle/gemini-2.0-flash-exp:free | free\ngoogle/gemini-2.0-flash-lite-001 | paid\ngoogle/gemini-2.5-flash | paid\ngoogle/gemini-2.5-flash-lite-preview-06-17 | paid\ngoogle/gemini-2.5-pro | paid\ngoogle/gemini-2.5-pro-exp-03-25 | paid\ngoogle/gemini-2.5-pro-preview-05-06 | paid\ngoogle/gemini-2.5-pro-preview | paid\ngoogle/gemma-2-27b-it | paid\ngoogle/gemma-2-9b-it | paid\ngoogle/gemma-2-9b-it:free | free\ngoogle/gemma-3-12b-it | paid\ngoogle/gemma-3-12b-it:free | free\ngoogle/gemma-3-27b-it | paid\ngoogle/gemma-3-27b-it:free | free\ngoogle/gemma-3-4b-it | paid\ngoogle/gemma-3-4b-it:free | free\ngoogle/gemma-3n-e2b-it:free | free\ngoogle/gemma-3n-e4b-it | paid\ngoogle/gemma-3n-e4b-it:free | free\ninception/mercury | paid\ninception/mercury-coder | paid\ninfermatic/mn-inferor-12b | paid\ninflection/inflection-3-pi | paid\ninflection/inflection-3-productivity | paid\nmoonshotai/kimi-dev-72b:free | free\nliquid/lfm-3b | paid\nliquid/lfm-40b | paid\nliquid/lfm-7b | paid\nmeta-llama/llama-guard-3-8b | paid\nalpindale/magnum-72b | paid\nanthracite-org/magnum-v2-72b | paid\nanthracite-org/magnum-v4-72b | paid\nmancer/weaver | paid\nmeta-llama/llama-3-70b-instruct | paid\nmeta-llama/llama-3-8b-instruct | paid\nmeta-llama/llama-3.1-405b | paid\nmeta-llama/llama-3.1-405b-instruct | paid\nmeta-llama/llama-3.1-405b-instruct:free | free\nmeta-llama/llama-3.1-70b-instruct | paid\nmeta-llama/llama-3.1-8b-instruct | paid\nmeta-llama/llama-3.2-11b-vision-instruct | paid\nmeta-llama/llama-3.2-11b-vision-instruct:free | free\nmeta-llama/llama-3.2-1b-instruct | paid\nmeta-llama/llama-3.2-3b-instruct | paid\nmeta-llama/llama-3.2-3b-instruct:free | free\nmeta-llama/llama-3.2-90b-vision-instruct | paid\nmeta-llama/llama-3.3-70b-instruct | paid\nmeta-llama/llama-3.3-70b-instruct:free | free\nmeta-llama/llama-4-maverick | paid\nmeta-llama/llama-4-scout | paid\nmeta-llama/llama-guard-4-12b | paid\nmeta-llama/llama-guard-2-8b | paid\nmicrosoft/mai-ds-r1:free | free\nmicrosoft/phi-4 | paid\nmicrosoft/phi-4-multimodal-instruct | paid\nmicrosoft/phi-4-reasoning-plus | paid\nmicrosoft/phi-3-medium-128k-instruct | paid\nmicrosoft/phi-3-mini-128k-instruct | paid\nmicrosoft/phi-3.5-mini-128k-instruct | paid\nsophosympatheia/midnight-rose-70b | paid\nminimax/minimax-m1 | paid\nminimax/minimax-01 | paid\nmistralai/mistral-large | paid\nmistralai/mistral-large-2407 | paid\nmistralai/mistral-large-2411 | paid\nnothingiisreal/mn-celeste-12b | paid\nmistralai/mistral-small | paid\nmistralai/mistral-tiny | paid\nmistralai/codestral-2501 | paid\nmistralai/devstral-medium | paid\nmistralai/devstral-small | paid\nmistralai/devstral-small-2505 | paid\nmistralai/devstral-small-2505:free | free\nmistralai/magistral-medium-2506 | paid\nmistralai/magistral-medium-2506:thinking | paid\nmistralai/magistral-small-2506 | paid\nmistralai/ministral-3b | paid\nmistralai/ministral-8b | paid\nmistralai/mistral-7b-instruct | paid\nmistralai/mistral-7b-instruct:free | free\nmistralai/mistral-7b-instruct-v0.1 | paid\nmistralai/mistral-7b-instruct-v0.2 | paid\nmistralai/mistral-7b-instruct-v0.3 | paid\nmistralai/mistral-medium-3 | paid\nmistralai/mistral-nemo | paid\nmistralai/mistral-nemo:free | free\nmistralai/mistral-small-24b-instruct-2501 | paid\nmistralai/mistral-small-24b-instruct-2501:free | free\nmistralai/mistral-small-3.1-24b-instruct | paid\nmistralai/mistral-small-3.1-24b-instruct:free | free\nmistralai/mistral-small-3.2-24b-instruct | paid\nmistralai/mistral-small-3.2-24b-instruct:free | free\nmistralai/mixtral-8x22b-instruct | paid\nmistralai/mixtral-8x7b-instruct | paid\nmistralai/pixtral-12b | paid\nmistralai/pixtral-large-2411 | paid\nmistralai/mistral-saba | paid\nmoonshotai/kimi-vl-a3b-thinking:free | free\nmoonshotai/kimi-k2 | paid\nmoonshotai/kimi-k2:free | free\nmorph/morph-v2 | paid\nmorph/morph-v3-fast | paid\nmorph/morph-v3-large | paid\ngryphe/mythomax-l2-13b | paid\nneversleep/llama-3-lumimaid-70b | paid\nneversleep/llama-3.1-lumimaid-8b | paid\nneversleep/noromaid-20b | paid\nnousresearch/deephermes-3-llama-3-8b-preview:free | free\nnousresearch/nous-hermes-2-mixtral-8x7b-dpo | paid\nnousresearch/hermes-3-llama-3.1-405b | paid\nnousresearch/hermes-3-llama-3.1-70b | paid\nnousresearch/hermes-2-pro-llama-3-8b | paid\nnvidia/llama-3.1-nemotron-70b-instruct | paid\nnvidia/llama-3.1-nemotron-ultra-253b-v1 | paid\nnvidia/llama-3.1-nemotron-ultra-253b-v1:free | free\nnvidia/llama-3.3-nemotron-super-49b-v1 | paid\nopenai/chatgpt-4o-latest | paid\nopenai/codex-mini | paid\nopenai/gpt-3.5-turbo | paid\nopenai/gpt-3.5-turbo-0613 | paid\nopenai/gpt-3.5-turbo-16k | paid\nopenai/gpt-3.5-turbo-instruct | paid\nopenai/gpt-4 | paid\nopenai/gpt-4-0314 | paid\nopenai/gpt-4-turbo | paid\nopenai/gpt-4-1106-preview | paid\nopenai/gpt-4-turbo-preview | paid\nopenai/gpt-4.1 | paid\nopenai/gpt-4.1-mini | paid\nopenai/gpt-4.1-nano | paid\nopenai/gpt-4o | paid\nopenai/gpt-4o-2024-05-13 | paid\nopenai/gpt-4o-2024-08-06 | paid\nopenai/gpt-4o-2024-11-20 | paid\nopenai/gpt-4o:extended | paid\nopenai/gpt-4o-search-preview | paid\nopenai/gpt-4o-mini | paid\nopenai/gpt-4o-mini-2024-07-18 | paid\nopenai/gpt-4o-mini-search-preview | paid\nopenai/o1 | paid\nopenai/o1-mini | paid\nopenai/o1-mini-2024-09-12 | paid\nopenai/o1-preview | paid\nopenai/o1-preview-2024-09-12 | paid\nopenai/o1-pro | paid\nopenai/o3 | paid\nopenai/o3-mini | paid\nopenai/o3-mini-high | paid\nopenai/o3-pro | paid\nopenai/o4-mini | paid\nopenai/o4-mini-high | paid\nopengvlab/internvl3-14b | paid\nperplexity/r1-1776 | paid\nperplexity/sonar | paid\nperplexity/sonar-deep-research | paid\nperplexity/sonar-pro | paid\nperplexity/sonar-reasoning | paid\nperplexity/sonar-reasoning-pro | paid\npygmalionai/mythalion-13b | paid\nqwen/qwen-2-72b-instruct | paid\nqwen/qwen-vl-max | paid\nqwen/qwen-vl-plus | paid\nqwen/qwen-max | paid\nqwen/qwen-plus | paid\nqwen/qwen-turbo | paid\nqwen/qwen2.5-vl-32b-instruct | paid\nqwen/qwen2.5-vl-32b-instruct:free | free\nqwen/qwen2.5-vl-72b-instruct | paid\nqwen/qwen2.5-vl-72b-instruct:free | free\nqwen/qwen-2.5-vl-7b-instruct | paid\nqwen/qwen3-14b | paid\nqwen/qwen3-14b:free | free\nqwen/qwen3-235b-a22b | paid\nqwen/qwen3-235b-a22b:free | free\nqwen/qwen3-30b-a3b | paid\nqwen/qwen3-30b-a3b:free | free\nqwen/qwen3-32b | paid\nqwen/qwen3-32b:free | free\nqwen/qwen3-4b:free | free\nqwen/qwen3-8b | paid\nqwen/qwen3-8b:free | free\nqwen/qwq-32b | paid\nqwen/qwq-32b:free | free\nqwen/qwq-32b-preview | paid\nqwen/qwen-2.5-72b-instruct | paid\nqwen/qwen-2.5-72b-instruct:free | free\nqwen/qwen-2.5-7b-instruct | paid\nqwen/qwen-2.5-coder-32b-instruct | paid\nqwen/qwen-2.5-coder-32b-instruct:free | free\nfeatherless/qwerky-72b:free | free\nrekaai/reka-flash-3 | paid\nrekaai/reka-flash-3:free | free\nundi95/remm-slerp-l2-13b | paid\nsao10k/l3-lunaris-8b | paid\nsao10k/l3-euryale-70b | paid\nsao10k/l3.1-euryale-70b | paid\nsao10k/l3.3-euryale-70b | paid\nsarvamai/sarvam-m | paid\nsarvamai/sarvam-m:free | free\nshisa-ai/shisa-v2-llama3.3-70b:free | free\nraifle/sorcererlm-8x22b | paid\nswitchpoint/router | paid\ntencent/hunyuan-a13b-instruct | paid\ntencent/hunyuan-a13b-instruct:free | free\nthedrummer/anubis-70b-v1.1 | paid\nthedrummer/anubis-pro-105b-v1 | paid\nthedrummer/rocinante-12b | paid\nthedrummer/skyfall-36b-v2 | paid\nthedrummer/unslopnemo-12b | paid\nthedrummer/valkyrie-49b-v1 | paid\nthudm/glm-4-32b | paid\nthudm/glm-4-32b:free | free\nthudm/glm-4.1v-9b-thinking | paid\nthudm/glm-z1-32b:free | free\ntngtech/deepseek-r1t-chimera:free | free\ntngtech/deepseek-r1t2-chimera:free | free\nundi95/toppy-m-7b | paid\nscb10x/llama3.1-typhoon2-70b-instruct | paid\ncognitivecomputations/dolphin-mistral-24b-venice-edition:free | free\nmicrosoft/wizardlm-2-8x22b | paid\nx-ai/grok-2-1212 | paid\nx-ai/grok-2-vision-1212 | paid\nx-ai/grok-3 | paid\nx-ai/grok-3-beta | paid\nx-ai/grok-3-mini | paid\nx-ai/grok-3-mini-beta | paid\nx-ai/grok-4 | paid\nx-ai/grok-vision-beta | paid\n\u001b[35m\u001b[1m\u001b[22m\u001b[39m\n\u001b[35m\u001b[1m OpenAI models:\u001b[22m\u001b[39m\n\u001b[35m\u001b[1m\u001b[22m\u001b[39m\nbabbage-002\nchatgpt-4o-latest\ncodex-mini-latest\ndall-e-2\ndall-e-3\ndavinci-002\ngpt-3.5-turbo\ngpt-3.5-turbo-0125\ngpt-3.5-turbo-1106\ngpt-3.5-turbo-16k\ngpt-3.5-turbo-instruct\ngpt-3.5-turbo-instruct-0914\ngpt-4\ngpt-4-0125-preview\ngpt-4-0613\ngpt-4-1106-preview\ngpt-4-turbo\ngpt-4-turbo-2024-04-09\ngpt-4-turbo-preview\ngpt-4.1\ngpt-4.1-2025-04-14\ngpt-4.1-mini\ngpt-4.1-mini-2025-04-14\ngpt-4.1-nano\ngpt-4.1-nano-2025-04-14\ngpt-4o\ngpt-4o-2024-05-13\ngpt-4o-2024-08-06\ngpt-4o-2024-11-20\ngpt-4o-audio-preview\ngpt-4o-audio-preview-2024-10-01\ngpt-4o-audio-preview-2024-12-17\ngpt-4o-audio-preview-2025-06-03\ngpt-4o-mini\ngpt-4o-mini-2024-07-18\ngpt-4o-mini-audio-preview\ngpt-4o-mini-audio-preview-2024-12-17\ngpt-4o-mini-realtime-preview\ngpt-4o-mini-realtime-preview-2024-12-17\ngpt-4o-mini-search-preview\ngpt-4o-mini-search-preview-2025-03-11\ngpt-4o-mini-transcribe\ngpt-4o-mini-tts\ngpt-4o-realtime-preview\ngpt-4o-realtime-preview-2024-10-01\ngpt-4o-realtime-preview-2024-12-17\ngpt-4o-realtime-preview-2025-06-03\ngpt-4o-search-preview\ngpt-4o-search-preview-2025-03-11\ngpt-4o-transcribe\ngpt-image-1\no1\no1-2024-12-17\no1-mini\no1-mini-2024-09-12\no1-preview\no1-preview-2024-09-12\no1-pro\no1-pro-2025-03-19\no3-mini\no3-mini-2025-01-31\no4-mini\no4-mini-2025-04-16\no4-mini-deep-research\no4-mini-deep-research-2025-06-26\nomni-moderation-2024-09-26\nomni-moderation-latest\ntext-embedding-3-large\ntext-embedding-3-small\ntext-embedding-ada-002\ntts-1\ntts-1-1106\ntts-1-hd\ntts-1-hd-1106\nwhisper-1\n-----\n\n\u001b[35m\u001b[1m\u001b[22m\u001b[39m\n\u001b[35m\u001b[1m Deepseek models:\u001b[22m\u001b[39m\n\u001b[35m\u001b[1m\u001b[22m\u001b[39m\ndeepseek-chat\ndeepseek-reasoner\n-----\n" + "description": "AI model to use for processing. Available models:\n\u001b[35m\u001b[1m\u001b[22m\u001b[39m\n\u001b[35m\u001b[1m OpenRouter models:\u001b[22m\u001b[39m\n\u001b[35m\u001b[1m\u001b[22m\u001b[39m\n01-ai/yi-large | paid\nagentica-org/deepcoder-14b-preview | paid\nagentica-org/deepcoder-14b-preview:free | free\nai21/jamba-1.6-large | paid\nai21/jamba-1.6-mini | paid\naion-labs/aion-1.0 | paid\naion-labs/aion-1.0-mini | paid\naion-labs/aion-rp-llama-3.1-8b | paid\nalfredpros/codellama-7b-instruct-solidity | paid\namazon/nova-lite-v1 | paid\namazon/nova-micro-v1 | paid\namazon/nova-pro-v1 | paid\nanthropic/claude-3-haiku | paid\nanthropic/claude-3-haiku:beta | paid\nanthropic/claude-3-opus | paid\nanthropic/claude-3-opus:beta | paid\nanthropic/claude-3-sonnet | paid\nanthropic/claude-3.5-haiku | paid\nanthropic/claude-3.5-haiku-20241022 | paid\nanthropic/claude-3.5-haiku-20241022:beta | paid\nanthropic/claude-3.5-haiku:beta | paid\nanthropic/claude-3.5-sonnet | paid\nanthropic/claude-3.5-sonnet-20240620 | paid\nanthropic/claude-3.5-sonnet-20240620:beta | paid\nanthropic/claude-3.5-sonnet:beta | paid\nanthropic/claude-3.7-sonnet | paid\nanthropic/claude-3.7-sonnet:beta | paid\nanthropic/claude-3.7-sonnet:thinking | paid\nanthropic/claude-opus-4 | paid\nanthropic/claude-sonnet-4 | paid\narcee-ai/coder-large | paid\narcee-ai/maestro-reasoning | paid\narcee-ai/spotlight | paid\narcee-ai/virtuoso-large | paid\narliai/qwq-32b-arliai-rpr-v1 | paid\narliai/qwq-32b-arliai-rpr-v1:free | free\nopenrouter/auto | paid\nbaidu/ernie-4.5-300b-a47b | paid\nbytedance/ui-tars-1.5-7b | paid\ncohere/command | paid\ncohere/command-a | paid\ncohere/command-r | paid\ncohere/command-r-03-2024 | paid\ncohere/command-r-08-2024 | paid\ncohere/command-r-plus | paid\ncohere/command-r-plus-04-2024 | paid\ncohere/command-r-plus-08-2024 | paid\ncohere/command-r7b-12-2024 | paid\ndeepseek/deepseek-prover-v2 | paid\ndeepseek/deepseek-r1-0528-qwen3-8b | paid\ndeepseek/deepseek-r1-0528-qwen3-8b:free | free\ndeepseek/deepseek-chat | paid\ndeepseek/deepseek-chat-v3-0324 | paid\ndeepseek/deepseek-chat-v3-0324:free | free\ndeepseek/deepseek-v3-base | paid\ndeepseek/deepseek-r1 | paid\ndeepseek/deepseek-r1:free | free\ndeepseek/deepseek-r1-0528 | paid\ndeepseek/deepseek-r1-0528:free | free\ndeepseek/deepseek-r1-distill-llama-70b | paid\ndeepseek/deepseek-r1-distill-llama-70b:free | free\ndeepseek/deepseek-r1-distill-llama-8b | paid\ndeepseek/deepseek-r1-distill-qwen-1.5b | paid\ndeepseek/deepseek-r1-distill-qwen-14b | paid\ndeepseek/deepseek-r1-distill-qwen-14b:free | free\ndeepseek/deepseek-r1-distill-qwen-32b | paid\ndeepseek/deepseek-r1-distill-qwen-7b | paid\ncognitivecomputations/dolphin-mixtral-8x22b | paid\ncognitivecomputations/dolphin3.0-mistral-24b:free | free\ncognitivecomputations/dolphin3.0-r1-mistral-24b | paid\ncognitivecomputations/dolphin3.0-r1-mistral-24b:free | free\neleutherai/llemma_7b | paid\neva-unit-01/eva-qwen-2.5-72b | paid\nsao10k/fimbulvetr-11b-v2 | paid\nalpindale/goliath-120b | paid\ngoogle/gemini-flash-1.5 | paid\ngoogle/gemini-flash-1.5-8b | paid\ngoogle/gemini-pro-1.5 | paid\ngoogle/gemini-2.0-flash-001 | paid\ngoogle/gemini-2.0-flash-exp:free | free\ngoogle/gemini-2.0-flash-lite-001 | paid\ngoogle/gemini-2.5-flash | paid\ngoogle/gemini-2.5-flash-lite | paid\ngoogle/gemini-2.5-flash-lite-preview-06-17 | paid\ngoogle/gemini-2.5-pro | paid\ngoogle/gemini-2.5-pro-exp-03-25 | paid\ngoogle/gemini-2.5-pro-preview-05-06 | paid\ngoogle/gemini-2.5-pro-preview | paid\ngoogle/gemma-2-27b-it | paid\ngoogle/gemma-2-9b-it | paid\ngoogle/gemma-2-9b-it:free | free\ngoogle/gemma-3-12b-it | paid\ngoogle/gemma-3-12b-it:free | free\ngoogle/gemma-3-27b-it | paid\ngoogle/gemma-3-27b-it:free | free\ngoogle/gemma-3-4b-it | paid\ngoogle/gemma-3-4b-it:free | free\ngoogle/gemma-3n-e2b-it:free | free\ngoogle/gemma-3n-e4b-it | paid\ngoogle/gemma-3n-e4b-it:free | free\ninception/mercury | paid\ninception/mercury-coder | paid\ninfermatic/mn-inferor-12b | paid\ninflection/inflection-3-pi | paid\ninflection/inflection-3-productivity | paid\nmoonshotai/kimi-dev-72b:free | free\nliquid/lfm-3b | paid\nliquid/lfm-40b | paid\nliquid/lfm-7b | paid\nmeta-llama/llama-guard-3-8b | paid\nanthracite-org/magnum-v2-72b | paid\nanthracite-org/magnum-v4-72b | paid\nmancer/weaver | paid\nmeta-llama/llama-3-70b-instruct | paid\nmeta-llama/llama-3-8b-instruct | paid\nmeta-llama/llama-3.1-405b | paid\nmeta-llama/llama-3.1-405b-instruct | paid\nmeta-llama/llama-3.1-405b-instruct:free | free\nmeta-llama/llama-3.1-70b-instruct | paid\nmeta-llama/llama-3.1-8b-instruct | paid\nmeta-llama/llama-3.2-11b-vision-instruct | paid\nmeta-llama/llama-3.2-11b-vision-instruct:free | free\nmeta-llama/llama-3.2-1b-instruct | paid\nmeta-llama/llama-3.2-3b-instruct | paid\nmeta-llama/llama-3.2-3b-instruct:free | free\nmeta-llama/llama-3.2-90b-vision-instruct | paid\nmeta-llama/llama-3.3-70b-instruct | paid\nmeta-llama/llama-3.3-70b-instruct:free | free\nmeta-llama/llama-4-maverick | paid\nmeta-llama/llama-4-scout | paid\nmeta-llama/llama-guard-4-12b | paid\nmeta-llama/llama-guard-2-8b | paid\nmicrosoft/mai-ds-r1 | paid\nmicrosoft/mai-ds-r1:free | free\nmicrosoft/phi-4 | paid\nmicrosoft/phi-4-multimodal-instruct | paid\nmicrosoft/phi-4-reasoning-plus | paid\nmicrosoft/phi-3-medium-128k-instruct | paid\nmicrosoft/phi-3-mini-128k-instruct | paid\nmicrosoft/phi-3.5-mini-128k-instruct | paid\nsophosympatheia/midnight-rose-70b | paid\nminimax/minimax-m1 | paid\nminimax/minimax-01 | paid\nmistralai/mistral-large | paid\nmistralai/mistral-large-2407 | paid\nmistralai/mistral-large-2411 | paid\nnothingiisreal/mn-celeste-12b | paid\nmistralai/mistral-small | paid\nmistralai/mistral-tiny | paid\nmistralai/codestral-2501 | paid\nmistralai/devstral-medium | paid\nmistralai/devstral-small | paid\nmistralai/devstral-small-2505 | paid\nmistralai/devstral-small-2505:free | free\nmistralai/magistral-medium-2506 | paid\nmistralai/magistral-medium-2506:thinking | paid\nmistralai/magistral-small-2506 | paid\nmistralai/ministral-3b | paid\nmistralai/ministral-8b | paid\nmistralai/mistral-7b-instruct | paid\nmistralai/mistral-7b-instruct:free | free\nmistralai/mistral-7b-instruct-v0.1 | paid\nmistralai/mistral-7b-instruct-v0.2 | paid\nmistralai/mistral-7b-instruct-v0.3 | paid\nmistralai/mistral-medium-3 | paid\nmistralai/mistral-nemo | paid\nmistralai/mistral-nemo:free | free\nmistralai/mistral-small-24b-instruct-2501 | paid\nmistralai/mistral-small-24b-instruct-2501:free | free\nmistralai/mistral-small-3.1-24b-instruct | paid\nmistralai/mistral-small-3.1-24b-instruct:free | free\nmistralai/mistral-small-3.2-24b-instruct | paid\nmistralai/mistral-small-3.2-24b-instruct:free | free\nmistralai/mixtral-8x22b-instruct | paid\nmistralai/mixtral-8x7b-instruct | paid\nmistralai/pixtral-12b | paid\nmistralai/pixtral-large-2411 | paid\nmistralai/mistral-saba | paid\nmoonshotai/kimi-vl-a3b-thinking | paid\nmoonshotai/kimi-vl-a3b-thinking:free | free\nmoonshotai/kimi-k2 | paid\nmoonshotai/kimi-k2:free | free\nmorph/morph-v2 | paid\nmorph/morph-v3-fast | paid\nmorph/morph-v3-large | paid\ngryphe/mythomax-l2-13b | paid\nneversleep/llama-3-lumimaid-70b | paid\nneversleep/llama-3.1-lumimaid-8b | paid\nneversleep/noromaid-20b | paid\nnousresearch/deephermes-3-llama-3-8b-preview:free | free\nnousresearch/deephermes-3-mistral-24b-preview | paid\nnousresearch/nous-hermes-2-mixtral-8x7b-dpo | paid\nnousresearch/hermes-3-llama-3.1-405b | paid\nnousresearch/hermes-3-llama-3.1-70b | paid\nnousresearch/hermes-2-pro-llama-3-8b | paid\nnvidia/llama-3.1-nemotron-70b-instruct | paid\nnvidia/llama-3.1-nemotron-ultra-253b-v1 | paid\nnvidia/llama-3.1-nemotron-ultra-253b-v1:free | free\nnvidia/llama-3.3-nemotron-super-49b-v1 | paid\nopenai/chatgpt-4o-latest | paid\nopenai/codex-mini | paid\nopenai/gpt-3.5-turbo | paid\nopenai/gpt-3.5-turbo-0613 | paid\nopenai/gpt-3.5-turbo-16k | paid\nopenai/gpt-3.5-turbo-instruct | paid\nopenai/gpt-4 | paid\nopenai/gpt-4-0314 | paid\nopenai/gpt-4-turbo | paid\nopenai/gpt-4-1106-preview | paid\nopenai/gpt-4-turbo-preview | paid\nopenai/gpt-4.1 | paid\nopenai/gpt-4.1-mini | paid\nopenai/gpt-4.1-nano | paid\nopenai/gpt-4o | paid\nopenai/gpt-4o-2024-05-13 | paid\nopenai/gpt-4o-2024-08-06 | paid\nopenai/gpt-4o-2024-11-20 | paid\nopenai/gpt-4o:extended | paid\nopenai/gpt-4o-search-preview | paid\nopenai/gpt-4o-mini | paid\nopenai/gpt-4o-mini-2024-07-18 | paid\nopenai/gpt-4o-mini-search-preview | paid\nopenai/o1 | paid\nopenai/o1-mini | paid\nopenai/o1-mini-2024-09-12 | paid\nopenai/o1-preview | paid\nopenai/o1-preview-2024-09-12 | paid\nopenai/o1-pro | paid\nopenai/o3 | paid\nopenai/o3-mini | paid\nopenai/o3-mini-high | paid\nopenai/o3-pro | paid\nopenai/o4-mini | paid\nopenai/o4-mini-high | paid\nopengvlab/internvl3-14b | paid\nperplexity/r1-1776 | paid\nperplexity/sonar | paid\nperplexity/sonar-deep-research | paid\nperplexity/sonar-pro | paid\nperplexity/sonar-reasoning | paid\nperplexity/sonar-reasoning-pro | paid\npygmalionai/mythalion-13b | paid\nfeatherless/qwerky-72b:free | free\nqwen/qwen-2-72b-instruct | paid\nqwen/qwen-vl-max | paid\nqwen/qwen-vl-plus | paid\nqwen/qwen-max | paid\nqwen/qwen-plus | paid\nqwen/qwen-turbo | paid\nqwen/qwen2.5-vl-32b-instruct | paid\nqwen/qwen2.5-vl-32b-instruct:free | free\nqwen/qwen2.5-vl-72b-instruct | paid\nqwen/qwen2.5-vl-72b-instruct:free | free\nqwen/qwen-2.5-vl-7b-instruct | paid\nqwen/qwen3-14b | paid\nqwen/qwen3-14b:free | free\nqwen/qwen3-235b-a22b | paid\nqwen/qwen3-235b-a22b:free | free\nqwen/qwen3-235b-a22b-2507 | paid\nqwen/qwen3-235b-a22b-2507:free | free\nqwen/qwen3-235b-a22b-thinking-2507 | paid\nqwen/qwen3-30b-a3b | paid\nqwen/qwen3-30b-a3b:free | free\nqwen/qwen3-32b | paid\nqwen/qwen3-4b:free | free\nqwen/qwen3-8b | paid\nqwen/qwen3-8b:free | free\nqwen/qwen3-coder | paid\nqwen/qwen3-coder:free | free\nqwen/qwq-32b | paid\nqwen/qwq-32b:free | free\nqwen/qwq-32b-preview | paid\nqwen/qwen-2.5-72b-instruct | paid\nqwen/qwen-2.5-72b-instruct:free | free\nqwen/qwen-2.5-7b-instruct | paid\nqwen/qwen-2.5-coder-32b-instruct | paid\nqwen/qwen-2.5-coder-32b-instruct:free | free\nrekaai/reka-flash-3 | paid\nrekaai/reka-flash-3:free | free\nundi95/remm-slerp-l2-13b | paid\nsao10k/l3-lunaris-8b | paid\nsao10k/l3-euryale-70b | paid\nsao10k/l3.1-euryale-70b | paid\nsao10k/l3.3-euryale-70b | paid\nsarvamai/sarvam-m | paid\nsarvamai/sarvam-m:free | free\nshisa-ai/shisa-v2-llama3.3-70b | paid\nshisa-ai/shisa-v2-llama3.3-70b:free | free\nraifle/sorcererlm-8x22b | paid\nswitchpoint/router | paid\ntencent/hunyuan-a13b-instruct | paid\ntencent/hunyuan-a13b-instruct:free | free\nthedrummer/anubis-70b-v1.1 | paid\nthedrummer/anubis-pro-105b-v1 | paid\nthedrummer/rocinante-12b | paid\nthedrummer/skyfall-36b-v2 | paid\nthedrummer/unslopnemo-12b | paid\nthedrummer/valkyrie-49b-v1 | paid\nthudm/glm-4-32b | paid\nthudm/glm-4-32b:free | free\nthudm/glm-4.1v-9b-thinking | paid\nthudm/glm-z1-32b | paid\nthudm/glm-z1-32b:free | free\ntngtech/deepseek-r1t-chimera:free | free\ntngtech/deepseek-r1t2-chimera | paid\ntngtech/deepseek-r1t2-chimera:free | free\nundi95/toppy-m-7b | paid\nscb10x/llama3.1-typhoon2-70b-instruct | paid\ncognitivecomputations/dolphin-mistral-24b-venice-edition:free | free\nmicrosoft/wizardlm-2-8x22b | paid\nx-ai/grok-2-1212 | paid\nx-ai/grok-2-vision-1212 | paid\nx-ai/grok-3 | paid\nx-ai/grok-3-beta | paid\nx-ai/grok-3-mini | paid\nx-ai/grok-3-mini-beta | paid\nx-ai/grok-4 | paid\nx-ai/grok-vision-beta | paid\nz-ai/glm-4-32b | paid\n\u001b[35m\u001b[1m\u001b[22m\u001b[39m\n\u001b[35m\u001b[1m OpenAI models:\u001b[22m\u001b[39m\n\u001b[35m\u001b[1m\u001b[22m\u001b[39m\nbabbage-002\nchatgpt-4o-latest\ncodex-mini-latest\ndall-e-2\ndall-e-3\ndavinci-002\ngpt-3.5-turbo\ngpt-3.5-turbo-0125\ngpt-3.5-turbo-1106\ngpt-3.5-turbo-16k\ngpt-3.5-turbo-instruct\ngpt-3.5-turbo-instruct-0914\ngpt-4\ngpt-4-0125-preview\ngpt-4-0613\ngpt-4-1106-preview\ngpt-4-turbo\ngpt-4-turbo-2024-04-09\ngpt-4-turbo-preview\ngpt-4.1\ngpt-4.1-2025-04-14\ngpt-4.1-mini\ngpt-4.1-mini-2025-04-14\ngpt-4.1-nano\ngpt-4.1-nano-2025-04-14\ngpt-4o\ngpt-4o-2024-05-13\ngpt-4o-2024-08-06\ngpt-4o-2024-11-20\ngpt-4o-audio-preview\ngpt-4o-audio-preview-2024-10-01\ngpt-4o-audio-preview-2024-12-17\ngpt-4o-audio-preview-2025-06-03\ngpt-4o-mini\ngpt-4o-mini-2024-07-18\ngpt-4o-mini-audio-preview\ngpt-4o-mini-audio-preview-2024-12-17\ngpt-4o-mini-realtime-preview\ngpt-4o-mini-realtime-preview-2024-12-17\ngpt-4o-mini-search-preview\ngpt-4o-mini-search-preview-2025-03-11\ngpt-4o-mini-transcribe\ngpt-4o-mini-tts\ngpt-4o-realtime-preview\ngpt-4o-realtime-preview-2024-10-01\ngpt-4o-realtime-preview-2024-12-17\ngpt-4o-realtime-preview-2025-06-03\ngpt-4o-search-preview\ngpt-4o-search-preview-2025-03-11\ngpt-4o-transcribe\ngpt-image-1\no1\no1-2024-12-17\no1-mini\no1-mini-2024-09-12\no1-preview\no1-preview-2024-09-12\no1-pro\no1-pro-2025-03-19\no3-mini\no3-mini-2025-01-31\no4-mini\no4-mini-2025-04-16\no4-mini-deep-research\no4-mini-deep-research-2025-06-26\nomni-moderation-2024-09-26\nomni-moderation-latest\ntext-embedding-3-large\ntext-embedding-3-small\ntext-embedding-ada-002\ntts-1\ntts-1-1106\ntts-1-hd\ntts-1-hd-1106\nwhisper-1\n-----\n\n\u001b[35m\u001b[1m\u001b[22m\u001b[39m\n\u001b[35m\u001b[1m Deepseek models:\u001b[22m\u001b[39m\n\u001b[35m\u001b[1m\u001b[22m\u001b[39m\ndeepseek-chat\ndeepseek-reasoner\n-----\n" }, "router": { "type": "string", diff --git a/packages/kbot/schema_ui.json b/packages/kbot/schema_ui.json index 095a0f4e..25f9512c 100644 --- a/packages/kbot/schema_ui.json +++ b/packages/kbot/schema_ui.json @@ -79,7 +79,7 @@ "ui:title": "Api_key" }, "model": { - "ui:description": "AI model to use for processing. Available models:\n\u001b[35m\u001b[1m\u001b[22m\u001b[39m\n\u001b[35m\u001b[1m OpenRouter models:\u001b[22m\u001b[39m\n\u001b[35m\u001b[1m\u001b[22m\u001b[39m\n01-ai/yi-large | paid\naetherwiing/mn-starcannon-12b | paid\nagentica-org/deepcoder-14b-preview | paid\nagentica-org/deepcoder-14b-preview:free | free\nai21/jamba-1.6-large | paid\nai21/jamba-1.6-mini | paid\naion-labs/aion-1.0 | paid\naion-labs/aion-1.0-mini | paid\naion-labs/aion-rp-llama-3.1-8b | paid\nalfredpros/codellama-7b-instruct-solidity | paid\namazon/nova-lite-v1 | paid\namazon/nova-micro-v1 | paid\namazon/nova-pro-v1 | paid\nanthropic/claude-3-haiku | paid\nanthropic/claude-3-haiku:beta | paid\nanthropic/claude-3-opus | paid\nanthropic/claude-3-opus:beta | paid\nanthropic/claude-3-sonnet | paid\nanthropic/claude-3-sonnet:beta | paid\nanthropic/claude-3.5-haiku | paid\nanthropic/claude-3.5-haiku-20241022 | paid\nanthropic/claude-3.5-haiku-20241022:beta | paid\nanthropic/claude-3.5-haiku:beta | paid\nanthropic/claude-3.5-sonnet | paid\nanthropic/claude-3.5-sonnet-20240620 | paid\nanthropic/claude-3.5-sonnet-20240620:beta | paid\nanthropic/claude-3.5-sonnet:beta | paid\nanthropic/claude-3.7-sonnet | paid\nanthropic/claude-3.7-sonnet:beta | paid\nanthropic/claude-3.7-sonnet:thinking | paid\nanthropic/claude-opus-4 | paid\nanthropic/claude-sonnet-4 | paid\nanthropic/claude-2 | paid\nanthropic/claude-2:beta | paid\nanthropic/claude-2.0 | paid\nanthropic/claude-2.0:beta | paid\nanthropic/claude-2.1 | paid\nanthropic/claude-2.1:beta | paid\narcee-ai/arcee-blitz | paid\narcee-ai/caller-large | paid\narcee-ai/coder-large | paid\narcee-ai/maestro-reasoning | paid\narcee-ai/spotlight | paid\narcee-ai/virtuoso-large | paid\narcee-ai/virtuoso-medium-v2 | paid\narliai/qwq-32b-arliai-rpr-v1:free | free\nopenrouter/auto | paid\nbaidu/ernie-4.5-300b-a47b | paid\ncohere/command | paid\ncohere/command-a | paid\ncohere/command-r | paid\ncohere/command-r-03-2024 | paid\ncohere/command-r-08-2024 | paid\ncohere/command-r-plus | paid\ncohere/command-r-plus-04-2024 | paid\ncohere/command-r-plus-08-2024 | paid\ncohere/command-r7b-12-2024 | paid\ndeepseek/deepseek-prover-v2 | paid\ndeepseek/deepseek-r1-0528-qwen3-8b | paid\ndeepseek/deepseek-r1-0528-qwen3-8b:free | free\ndeepseek/deepseek-chat | paid\ndeepseek/deepseek-chat:free | free\ndeepseek/deepseek-chat-v3-0324 | paid\ndeepseek/deepseek-chat-v3-0324:free | free\ndeepseek/deepseek-v3-base:free | free\ndeepseek/deepseek-r1 | paid\ndeepseek/deepseek-r1:free | free\ndeepseek/deepseek-r1-0528 | paid\ndeepseek/deepseek-r1-0528:free | free\ndeepseek/deepseek-r1-distill-llama-70b | paid\ndeepseek/deepseek-r1-distill-llama-70b:free | free\ndeepseek/deepseek-r1-distill-llama-8b | paid\ndeepseek/deepseek-r1-distill-qwen-1.5b | paid\ndeepseek/deepseek-r1-distill-qwen-14b | paid\ndeepseek/deepseek-r1-distill-qwen-14b:free | free\ndeepseek/deepseek-r1-distill-qwen-32b | paid\ndeepseek/deepseek-r1-distill-qwen-7b | paid\ncognitivecomputations/dolphin-mixtral-8x22b | paid\ncognitivecomputations/dolphin3.0-mistral-24b:free | free\ncognitivecomputations/dolphin3.0-r1-mistral-24b | paid\ncognitivecomputations/dolphin3.0-r1-mistral-24b:free | free\neleutherai/llemma_7b | paid\neva-unit-01/eva-llama-3.33-70b | paid\neva-unit-01/eva-qwen-2.5-72b | paid\nsao10k/fimbulvetr-11b-v2 | paid\nalpindale/goliath-120b | paid\ngoogle/gemini-flash-1.5 | paid\ngoogle/gemini-flash-1.5-8b | paid\ngoogle/gemini-pro-1.5 | paid\ngoogle/gemini-2.0-flash-001 | paid\ngoogle/gemini-2.0-flash-exp:free | free\ngoogle/gemini-2.0-flash-lite-001 | paid\ngoogle/gemini-2.5-flash | paid\ngoogle/gemini-2.5-flash-lite-preview-06-17 | paid\ngoogle/gemini-2.5-pro | paid\ngoogle/gemini-2.5-pro-exp-03-25 | paid\ngoogle/gemini-2.5-pro-preview-05-06 | paid\ngoogle/gemini-2.5-pro-preview | paid\ngoogle/gemma-2-27b-it | paid\ngoogle/gemma-2-9b-it | paid\ngoogle/gemma-2-9b-it:free | free\ngoogle/gemma-3-12b-it | paid\ngoogle/gemma-3-12b-it:free | free\ngoogle/gemma-3-27b-it | paid\ngoogle/gemma-3-27b-it:free | free\ngoogle/gemma-3-4b-it | paid\ngoogle/gemma-3-4b-it:free | free\ngoogle/gemma-3n-e2b-it:free | free\ngoogle/gemma-3n-e4b-it | paid\ngoogle/gemma-3n-e4b-it:free | free\ninception/mercury | paid\ninception/mercury-coder | paid\ninfermatic/mn-inferor-12b | paid\ninflection/inflection-3-pi | paid\ninflection/inflection-3-productivity | paid\nmoonshotai/kimi-dev-72b:free | free\nliquid/lfm-3b | paid\nliquid/lfm-40b | paid\nliquid/lfm-7b | paid\nmeta-llama/llama-guard-3-8b | paid\nalpindale/magnum-72b | paid\nanthracite-org/magnum-v2-72b | paid\nanthracite-org/magnum-v4-72b | paid\nmancer/weaver | paid\nmeta-llama/llama-3-70b-instruct | paid\nmeta-llama/llama-3-8b-instruct | paid\nmeta-llama/llama-3.1-405b | paid\nmeta-llama/llama-3.1-405b-instruct | paid\nmeta-llama/llama-3.1-405b-instruct:free | free\nmeta-llama/llama-3.1-70b-instruct | paid\nmeta-llama/llama-3.1-8b-instruct | paid\nmeta-llama/llama-3.2-11b-vision-instruct | paid\nmeta-llama/llama-3.2-11b-vision-instruct:free | free\nmeta-llama/llama-3.2-1b-instruct | paid\nmeta-llama/llama-3.2-3b-instruct | paid\nmeta-llama/llama-3.2-3b-instruct:free | free\nmeta-llama/llama-3.2-90b-vision-instruct | paid\nmeta-llama/llama-3.3-70b-instruct | paid\nmeta-llama/llama-3.3-70b-instruct:free | free\nmeta-llama/llama-4-maverick | paid\nmeta-llama/llama-4-scout | paid\nmeta-llama/llama-guard-4-12b | paid\nmeta-llama/llama-guard-2-8b | paid\nmicrosoft/mai-ds-r1:free | free\nmicrosoft/phi-4 | paid\nmicrosoft/phi-4-multimodal-instruct | paid\nmicrosoft/phi-4-reasoning-plus | paid\nmicrosoft/phi-3-medium-128k-instruct | paid\nmicrosoft/phi-3-mini-128k-instruct | paid\nmicrosoft/phi-3.5-mini-128k-instruct | paid\nsophosympatheia/midnight-rose-70b | paid\nminimax/minimax-m1 | paid\nminimax/minimax-01 | paid\nmistralai/mistral-large | paid\nmistralai/mistral-large-2407 | paid\nmistralai/mistral-large-2411 | paid\nnothingiisreal/mn-celeste-12b | paid\nmistralai/mistral-small | paid\nmistralai/mistral-tiny | paid\nmistralai/codestral-2501 | paid\nmistralai/devstral-medium | paid\nmistralai/devstral-small | paid\nmistralai/devstral-small-2505 | paid\nmistralai/devstral-small-2505:free | free\nmistralai/magistral-medium-2506 | paid\nmistralai/magistral-medium-2506:thinking | paid\nmistralai/magistral-small-2506 | paid\nmistralai/ministral-3b | paid\nmistralai/ministral-8b | paid\nmistralai/mistral-7b-instruct | paid\nmistralai/mistral-7b-instruct:free | free\nmistralai/mistral-7b-instruct-v0.1 | paid\nmistralai/mistral-7b-instruct-v0.2 | paid\nmistralai/mistral-7b-instruct-v0.3 | paid\nmistralai/mistral-medium-3 | paid\nmistralai/mistral-nemo | paid\nmistralai/mistral-nemo:free | free\nmistralai/mistral-small-24b-instruct-2501 | paid\nmistralai/mistral-small-24b-instruct-2501:free | free\nmistralai/mistral-small-3.1-24b-instruct | paid\nmistralai/mistral-small-3.1-24b-instruct:free | free\nmistralai/mistral-small-3.2-24b-instruct | paid\nmistralai/mistral-small-3.2-24b-instruct:free | free\nmistralai/mixtral-8x22b-instruct | paid\nmistralai/mixtral-8x7b-instruct | paid\nmistralai/pixtral-12b | paid\nmistralai/pixtral-large-2411 | paid\nmistralai/mistral-saba | paid\nmoonshotai/kimi-vl-a3b-thinking:free | free\nmoonshotai/kimi-k2 | paid\nmoonshotai/kimi-k2:free | free\nmorph/morph-v2 | paid\nmorph/morph-v3-fast | paid\nmorph/morph-v3-large | paid\ngryphe/mythomax-l2-13b | paid\nneversleep/llama-3-lumimaid-70b | paid\nneversleep/llama-3.1-lumimaid-8b | paid\nneversleep/noromaid-20b | paid\nnousresearch/deephermes-3-llama-3-8b-preview:free | free\nnousresearch/nous-hermes-2-mixtral-8x7b-dpo | paid\nnousresearch/hermes-3-llama-3.1-405b | paid\nnousresearch/hermes-3-llama-3.1-70b | paid\nnousresearch/hermes-2-pro-llama-3-8b | paid\nnvidia/llama-3.1-nemotron-70b-instruct | paid\nnvidia/llama-3.1-nemotron-ultra-253b-v1 | paid\nnvidia/llama-3.1-nemotron-ultra-253b-v1:free | free\nnvidia/llama-3.3-nemotron-super-49b-v1 | paid\nopenai/chatgpt-4o-latest | paid\nopenai/codex-mini | paid\nopenai/gpt-3.5-turbo | paid\nopenai/gpt-3.5-turbo-0613 | paid\nopenai/gpt-3.5-turbo-16k | paid\nopenai/gpt-3.5-turbo-instruct | paid\nopenai/gpt-4 | paid\nopenai/gpt-4-0314 | paid\nopenai/gpt-4-turbo | paid\nopenai/gpt-4-1106-preview | paid\nopenai/gpt-4-turbo-preview | paid\nopenai/gpt-4.1 | paid\nopenai/gpt-4.1-mini | paid\nopenai/gpt-4.1-nano | paid\nopenai/gpt-4o | paid\nopenai/gpt-4o-2024-05-13 | paid\nopenai/gpt-4o-2024-08-06 | paid\nopenai/gpt-4o-2024-11-20 | paid\nopenai/gpt-4o:extended | paid\nopenai/gpt-4o-search-preview | paid\nopenai/gpt-4o-mini | paid\nopenai/gpt-4o-mini-2024-07-18 | paid\nopenai/gpt-4o-mini-search-preview | paid\nopenai/o1 | paid\nopenai/o1-mini | paid\nopenai/o1-mini-2024-09-12 | paid\nopenai/o1-preview | paid\nopenai/o1-preview-2024-09-12 | paid\nopenai/o1-pro | paid\nopenai/o3 | paid\nopenai/o3-mini | paid\nopenai/o3-mini-high | paid\nopenai/o3-pro | paid\nopenai/o4-mini | paid\nopenai/o4-mini-high | paid\nopengvlab/internvl3-14b | paid\nperplexity/r1-1776 | paid\nperplexity/sonar | paid\nperplexity/sonar-deep-research | paid\nperplexity/sonar-pro | paid\nperplexity/sonar-reasoning | paid\nperplexity/sonar-reasoning-pro | paid\npygmalionai/mythalion-13b | paid\nqwen/qwen-2-72b-instruct | paid\nqwen/qwen-vl-max | paid\nqwen/qwen-vl-plus | paid\nqwen/qwen-max | paid\nqwen/qwen-plus | paid\nqwen/qwen-turbo | paid\nqwen/qwen2.5-vl-32b-instruct | paid\nqwen/qwen2.5-vl-32b-instruct:free | free\nqwen/qwen2.5-vl-72b-instruct | paid\nqwen/qwen2.5-vl-72b-instruct:free | free\nqwen/qwen-2.5-vl-7b-instruct | paid\nqwen/qwen3-14b | paid\nqwen/qwen3-14b:free | free\nqwen/qwen3-235b-a22b | paid\nqwen/qwen3-235b-a22b:free | free\nqwen/qwen3-30b-a3b | paid\nqwen/qwen3-30b-a3b:free | free\nqwen/qwen3-32b | paid\nqwen/qwen3-32b:free | free\nqwen/qwen3-4b:free | free\nqwen/qwen3-8b | paid\nqwen/qwen3-8b:free | free\nqwen/qwq-32b | paid\nqwen/qwq-32b:free | free\nqwen/qwq-32b-preview | paid\nqwen/qwen-2.5-72b-instruct | paid\nqwen/qwen-2.5-72b-instruct:free | free\nqwen/qwen-2.5-7b-instruct | paid\nqwen/qwen-2.5-coder-32b-instruct | paid\nqwen/qwen-2.5-coder-32b-instruct:free | free\nfeatherless/qwerky-72b:free | free\nrekaai/reka-flash-3 | paid\nrekaai/reka-flash-3:free | free\nundi95/remm-slerp-l2-13b | paid\nsao10k/l3-lunaris-8b | paid\nsao10k/l3-euryale-70b | paid\nsao10k/l3.1-euryale-70b | paid\nsao10k/l3.3-euryale-70b | paid\nsarvamai/sarvam-m | paid\nsarvamai/sarvam-m:free | free\nshisa-ai/shisa-v2-llama3.3-70b:free | free\nraifle/sorcererlm-8x22b | paid\nswitchpoint/router | paid\ntencent/hunyuan-a13b-instruct | paid\ntencent/hunyuan-a13b-instruct:free | free\nthedrummer/anubis-70b-v1.1 | paid\nthedrummer/anubis-pro-105b-v1 | paid\nthedrummer/rocinante-12b | paid\nthedrummer/skyfall-36b-v2 | paid\nthedrummer/unslopnemo-12b | paid\nthedrummer/valkyrie-49b-v1 | paid\nthudm/glm-4-32b | paid\nthudm/glm-4-32b:free | free\nthudm/glm-4.1v-9b-thinking | paid\nthudm/glm-z1-32b:free | free\ntngtech/deepseek-r1t-chimera:free | free\ntngtech/deepseek-r1t2-chimera:free | free\nundi95/toppy-m-7b | paid\nscb10x/llama3.1-typhoon2-70b-instruct | paid\ncognitivecomputations/dolphin-mistral-24b-venice-edition:free | free\nmicrosoft/wizardlm-2-8x22b | paid\nx-ai/grok-2-1212 | paid\nx-ai/grok-2-vision-1212 | paid\nx-ai/grok-3 | paid\nx-ai/grok-3-beta | paid\nx-ai/grok-3-mini | paid\nx-ai/grok-3-mini-beta | paid\nx-ai/grok-4 | paid\nx-ai/grok-vision-beta | paid\n\u001b[35m\u001b[1m\u001b[22m\u001b[39m\n\u001b[35m\u001b[1m OpenAI models:\u001b[22m\u001b[39m\n\u001b[35m\u001b[1m\u001b[22m\u001b[39m\nbabbage-002\nchatgpt-4o-latest\ncodex-mini-latest\ndall-e-2\ndall-e-3\ndavinci-002\ngpt-3.5-turbo\ngpt-3.5-turbo-0125\ngpt-3.5-turbo-1106\ngpt-3.5-turbo-16k\ngpt-3.5-turbo-instruct\ngpt-3.5-turbo-instruct-0914\ngpt-4\ngpt-4-0125-preview\ngpt-4-0613\ngpt-4-1106-preview\ngpt-4-turbo\ngpt-4-turbo-2024-04-09\ngpt-4-turbo-preview\ngpt-4.1\ngpt-4.1-2025-04-14\ngpt-4.1-mini\ngpt-4.1-mini-2025-04-14\ngpt-4.1-nano\ngpt-4.1-nano-2025-04-14\ngpt-4o\ngpt-4o-2024-05-13\ngpt-4o-2024-08-06\ngpt-4o-2024-11-20\ngpt-4o-audio-preview\ngpt-4o-audio-preview-2024-10-01\ngpt-4o-audio-preview-2024-12-17\ngpt-4o-audio-preview-2025-06-03\ngpt-4o-mini\ngpt-4o-mini-2024-07-18\ngpt-4o-mini-audio-preview\ngpt-4o-mini-audio-preview-2024-12-17\ngpt-4o-mini-realtime-preview\ngpt-4o-mini-realtime-preview-2024-12-17\ngpt-4o-mini-search-preview\ngpt-4o-mini-search-preview-2025-03-11\ngpt-4o-mini-transcribe\ngpt-4o-mini-tts\ngpt-4o-realtime-preview\ngpt-4o-realtime-preview-2024-10-01\ngpt-4o-realtime-preview-2024-12-17\ngpt-4o-realtime-preview-2025-06-03\ngpt-4o-search-preview\ngpt-4o-search-preview-2025-03-11\ngpt-4o-transcribe\ngpt-image-1\no1\no1-2024-12-17\no1-mini\no1-mini-2024-09-12\no1-preview\no1-preview-2024-09-12\no1-pro\no1-pro-2025-03-19\no3-mini\no3-mini-2025-01-31\no4-mini\no4-mini-2025-04-16\no4-mini-deep-research\no4-mini-deep-research-2025-06-26\nomni-moderation-2024-09-26\nomni-moderation-latest\ntext-embedding-3-large\ntext-embedding-3-small\ntext-embedding-ada-002\ntts-1\ntts-1-1106\ntts-1-hd\ntts-1-hd-1106\nwhisper-1\n-----\n\n\u001b[35m\u001b[1m\u001b[22m\u001b[39m\n\u001b[35m\u001b[1m Deepseek models:\u001b[22m\u001b[39m\n\u001b[35m\u001b[1m\u001b[22m\u001b[39m\ndeepseek-chat\ndeepseek-reasoner\n-----\n", + "ui:description": "AI model to use for processing. Available models:\n\u001b[35m\u001b[1m\u001b[22m\u001b[39m\n\u001b[35m\u001b[1m OpenRouter models:\u001b[22m\u001b[39m\n\u001b[35m\u001b[1m\u001b[22m\u001b[39m\n01-ai/yi-large | paid\nagentica-org/deepcoder-14b-preview | paid\nagentica-org/deepcoder-14b-preview:free | free\nai21/jamba-1.6-large | paid\nai21/jamba-1.6-mini | paid\naion-labs/aion-1.0 | paid\naion-labs/aion-1.0-mini | paid\naion-labs/aion-rp-llama-3.1-8b | paid\nalfredpros/codellama-7b-instruct-solidity | paid\namazon/nova-lite-v1 | paid\namazon/nova-micro-v1 | paid\namazon/nova-pro-v1 | paid\nanthropic/claude-3-haiku | paid\nanthropic/claude-3-haiku:beta | paid\nanthropic/claude-3-opus | paid\nanthropic/claude-3-opus:beta | paid\nanthropic/claude-3-sonnet | paid\nanthropic/claude-3.5-haiku | paid\nanthropic/claude-3.5-haiku-20241022 | paid\nanthropic/claude-3.5-haiku-20241022:beta | paid\nanthropic/claude-3.5-haiku:beta | paid\nanthropic/claude-3.5-sonnet | paid\nanthropic/claude-3.5-sonnet-20240620 | paid\nanthropic/claude-3.5-sonnet-20240620:beta | paid\nanthropic/claude-3.5-sonnet:beta | paid\nanthropic/claude-3.7-sonnet | paid\nanthropic/claude-3.7-sonnet:beta | paid\nanthropic/claude-3.7-sonnet:thinking | paid\nanthropic/claude-opus-4 | paid\nanthropic/claude-sonnet-4 | paid\narcee-ai/coder-large | paid\narcee-ai/maestro-reasoning | paid\narcee-ai/spotlight | paid\narcee-ai/virtuoso-large | paid\narliai/qwq-32b-arliai-rpr-v1 | paid\narliai/qwq-32b-arliai-rpr-v1:free | free\nopenrouter/auto | paid\nbaidu/ernie-4.5-300b-a47b | paid\nbytedance/ui-tars-1.5-7b | paid\ncohere/command | paid\ncohere/command-a | paid\ncohere/command-r | paid\ncohere/command-r-03-2024 | paid\ncohere/command-r-08-2024 | paid\ncohere/command-r-plus | paid\ncohere/command-r-plus-04-2024 | paid\ncohere/command-r-plus-08-2024 | paid\ncohere/command-r7b-12-2024 | paid\ndeepseek/deepseek-prover-v2 | paid\ndeepseek/deepseek-r1-0528-qwen3-8b | paid\ndeepseek/deepseek-r1-0528-qwen3-8b:free | free\ndeepseek/deepseek-chat | paid\ndeepseek/deepseek-chat-v3-0324 | paid\ndeepseek/deepseek-chat-v3-0324:free | free\ndeepseek/deepseek-v3-base | paid\ndeepseek/deepseek-r1 | paid\ndeepseek/deepseek-r1:free | free\ndeepseek/deepseek-r1-0528 | paid\ndeepseek/deepseek-r1-0528:free | free\ndeepseek/deepseek-r1-distill-llama-70b | paid\ndeepseek/deepseek-r1-distill-llama-70b:free | free\ndeepseek/deepseek-r1-distill-llama-8b | paid\ndeepseek/deepseek-r1-distill-qwen-1.5b | paid\ndeepseek/deepseek-r1-distill-qwen-14b | paid\ndeepseek/deepseek-r1-distill-qwen-14b:free | free\ndeepseek/deepseek-r1-distill-qwen-32b | paid\ndeepseek/deepseek-r1-distill-qwen-7b | paid\ncognitivecomputations/dolphin-mixtral-8x22b | paid\ncognitivecomputations/dolphin3.0-mistral-24b:free | free\ncognitivecomputations/dolphin3.0-r1-mistral-24b | paid\ncognitivecomputations/dolphin3.0-r1-mistral-24b:free | free\neleutherai/llemma_7b | paid\neva-unit-01/eva-qwen-2.5-72b | paid\nsao10k/fimbulvetr-11b-v2 | paid\nalpindale/goliath-120b | paid\ngoogle/gemini-flash-1.5 | paid\ngoogle/gemini-flash-1.5-8b | paid\ngoogle/gemini-pro-1.5 | paid\ngoogle/gemini-2.0-flash-001 | paid\ngoogle/gemini-2.0-flash-exp:free | free\ngoogle/gemini-2.0-flash-lite-001 | paid\ngoogle/gemini-2.5-flash | paid\ngoogle/gemini-2.5-flash-lite | paid\ngoogle/gemini-2.5-flash-lite-preview-06-17 | paid\ngoogle/gemini-2.5-pro | paid\ngoogle/gemini-2.5-pro-exp-03-25 | paid\ngoogle/gemini-2.5-pro-preview-05-06 | paid\ngoogle/gemini-2.5-pro-preview | paid\ngoogle/gemma-2-27b-it | paid\ngoogle/gemma-2-9b-it | paid\ngoogle/gemma-2-9b-it:free | free\ngoogle/gemma-3-12b-it | paid\ngoogle/gemma-3-12b-it:free | free\ngoogle/gemma-3-27b-it | paid\ngoogle/gemma-3-27b-it:free | free\ngoogle/gemma-3-4b-it | paid\ngoogle/gemma-3-4b-it:free | free\ngoogle/gemma-3n-e2b-it:free | free\ngoogle/gemma-3n-e4b-it | paid\ngoogle/gemma-3n-e4b-it:free | free\ninception/mercury | paid\ninception/mercury-coder | paid\ninfermatic/mn-inferor-12b | paid\ninflection/inflection-3-pi | paid\ninflection/inflection-3-productivity | paid\nmoonshotai/kimi-dev-72b:free | free\nliquid/lfm-3b | paid\nliquid/lfm-40b | paid\nliquid/lfm-7b | paid\nmeta-llama/llama-guard-3-8b | paid\nanthracite-org/magnum-v2-72b | paid\nanthracite-org/magnum-v4-72b | paid\nmancer/weaver | paid\nmeta-llama/llama-3-70b-instruct | paid\nmeta-llama/llama-3-8b-instruct | paid\nmeta-llama/llama-3.1-405b | paid\nmeta-llama/llama-3.1-405b-instruct | paid\nmeta-llama/llama-3.1-405b-instruct:free | free\nmeta-llama/llama-3.1-70b-instruct | paid\nmeta-llama/llama-3.1-8b-instruct | paid\nmeta-llama/llama-3.2-11b-vision-instruct | paid\nmeta-llama/llama-3.2-11b-vision-instruct:free | free\nmeta-llama/llama-3.2-1b-instruct | paid\nmeta-llama/llama-3.2-3b-instruct | paid\nmeta-llama/llama-3.2-3b-instruct:free | free\nmeta-llama/llama-3.2-90b-vision-instruct | paid\nmeta-llama/llama-3.3-70b-instruct | paid\nmeta-llama/llama-3.3-70b-instruct:free | free\nmeta-llama/llama-4-maverick | paid\nmeta-llama/llama-4-scout | paid\nmeta-llama/llama-guard-4-12b | paid\nmeta-llama/llama-guard-2-8b | paid\nmicrosoft/mai-ds-r1 | paid\nmicrosoft/mai-ds-r1:free | free\nmicrosoft/phi-4 | paid\nmicrosoft/phi-4-multimodal-instruct | paid\nmicrosoft/phi-4-reasoning-plus | paid\nmicrosoft/phi-3-medium-128k-instruct | paid\nmicrosoft/phi-3-mini-128k-instruct | paid\nmicrosoft/phi-3.5-mini-128k-instruct | paid\nsophosympatheia/midnight-rose-70b | paid\nminimax/minimax-m1 | paid\nminimax/minimax-01 | paid\nmistralai/mistral-large | paid\nmistralai/mistral-large-2407 | paid\nmistralai/mistral-large-2411 | paid\nnothingiisreal/mn-celeste-12b | paid\nmistralai/mistral-small | paid\nmistralai/mistral-tiny | paid\nmistralai/codestral-2501 | paid\nmistralai/devstral-medium | paid\nmistralai/devstral-small | paid\nmistralai/devstral-small-2505 | paid\nmistralai/devstral-small-2505:free | free\nmistralai/magistral-medium-2506 | paid\nmistralai/magistral-medium-2506:thinking | paid\nmistralai/magistral-small-2506 | paid\nmistralai/ministral-3b | paid\nmistralai/ministral-8b | paid\nmistralai/mistral-7b-instruct | paid\nmistralai/mistral-7b-instruct:free | free\nmistralai/mistral-7b-instruct-v0.1 | paid\nmistralai/mistral-7b-instruct-v0.2 | paid\nmistralai/mistral-7b-instruct-v0.3 | paid\nmistralai/mistral-medium-3 | paid\nmistralai/mistral-nemo | paid\nmistralai/mistral-nemo:free | free\nmistralai/mistral-small-24b-instruct-2501 | paid\nmistralai/mistral-small-24b-instruct-2501:free | free\nmistralai/mistral-small-3.1-24b-instruct | paid\nmistralai/mistral-small-3.1-24b-instruct:free | free\nmistralai/mistral-small-3.2-24b-instruct | paid\nmistralai/mistral-small-3.2-24b-instruct:free | free\nmistralai/mixtral-8x22b-instruct | paid\nmistralai/mixtral-8x7b-instruct | paid\nmistralai/pixtral-12b | paid\nmistralai/pixtral-large-2411 | paid\nmistralai/mistral-saba | paid\nmoonshotai/kimi-vl-a3b-thinking | paid\nmoonshotai/kimi-vl-a3b-thinking:free | free\nmoonshotai/kimi-k2 | paid\nmoonshotai/kimi-k2:free | free\nmorph/morph-v2 | paid\nmorph/morph-v3-fast | paid\nmorph/morph-v3-large | paid\ngryphe/mythomax-l2-13b | paid\nneversleep/llama-3-lumimaid-70b | paid\nneversleep/llama-3.1-lumimaid-8b | paid\nneversleep/noromaid-20b | paid\nnousresearch/deephermes-3-llama-3-8b-preview:free | free\nnousresearch/deephermes-3-mistral-24b-preview | paid\nnousresearch/nous-hermes-2-mixtral-8x7b-dpo | paid\nnousresearch/hermes-3-llama-3.1-405b | paid\nnousresearch/hermes-3-llama-3.1-70b | paid\nnousresearch/hermes-2-pro-llama-3-8b | paid\nnvidia/llama-3.1-nemotron-70b-instruct | paid\nnvidia/llama-3.1-nemotron-ultra-253b-v1 | paid\nnvidia/llama-3.1-nemotron-ultra-253b-v1:free | free\nnvidia/llama-3.3-nemotron-super-49b-v1 | paid\nopenai/chatgpt-4o-latest | paid\nopenai/codex-mini | paid\nopenai/gpt-3.5-turbo | paid\nopenai/gpt-3.5-turbo-0613 | paid\nopenai/gpt-3.5-turbo-16k | paid\nopenai/gpt-3.5-turbo-instruct | paid\nopenai/gpt-4 | paid\nopenai/gpt-4-0314 | paid\nopenai/gpt-4-turbo | paid\nopenai/gpt-4-1106-preview | paid\nopenai/gpt-4-turbo-preview | paid\nopenai/gpt-4.1 | paid\nopenai/gpt-4.1-mini | paid\nopenai/gpt-4.1-nano | paid\nopenai/gpt-4o | paid\nopenai/gpt-4o-2024-05-13 | paid\nopenai/gpt-4o-2024-08-06 | paid\nopenai/gpt-4o-2024-11-20 | paid\nopenai/gpt-4o:extended | paid\nopenai/gpt-4o-search-preview | paid\nopenai/gpt-4o-mini | paid\nopenai/gpt-4o-mini-2024-07-18 | paid\nopenai/gpt-4o-mini-search-preview | paid\nopenai/o1 | paid\nopenai/o1-mini | paid\nopenai/o1-mini-2024-09-12 | paid\nopenai/o1-preview | paid\nopenai/o1-preview-2024-09-12 | paid\nopenai/o1-pro | paid\nopenai/o3 | paid\nopenai/o3-mini | paid\nopenai/o3-mini-high | paid\nopenai/o3-pro | paid\nopenai/o4-mini | paid\nopenai/o4-mini-high | paid\nopengvlab/internvl3-14b | paid\nperplexity/r1-1776 | paid\nperplexity/sonar | paid\nperplexity/sonar-deep-research | paid\nperplexity/sonar-pro | paid\nperplexity/sonar-reasoning | paid\nperplexity/sonar-reasoning-pro | paid\npygmalionai/mythalion-13b | paid\nfeatherless/qwerky-72b:free | free\nqwen/qwen-2-72b-instruct | paid\nqwen/qwen-vl-max | paid\nqwen/qwen-vl-plus | paid\nqwen/qwen-max | paid\nqwen/qwen-plus | paid\nqwen/qwen-turbo | paid\nqwen/qwen2.5-vl-32b-instruct | paid\nqwen/qwen2.5-vl-32b-instruct:free | free\nqwen/qwen2.5-vl-72b-instruct | paid\nqwen/qwen2.5-vl-72b-instruct:free | free\nqwen/qwen-2.5-vl-7b-instruct | paid\nqwen/qwen3-14b | paid\nqwen/qwen3-14b:free | free\nqwen/qwen3-235b-a22b | paid\nqwen/qwen3-235b-a22b:free | free\nqwen/qwen3-235b-a22b-2507 | paid\nqwen/qwen3-235b-a22b-2507:free | free\nqwen/qwen3-235b-a22b-thinking-2507 | paid\nqwen/qwen3-30b-a3b | paid\nqwen/qwen3-30b-a3b:free | free\nqwen/qwen3-32b | paid\nqwen/qwen3-4b:free | free\nqwen/qwen3-8b | paid\nqwen/qwen3-8b:free | free\nqwen/qwen3-coder | paid\nqwen/qwen3-coder:free | free\nqwen/qwq-32b | paid\nqwen/qwq-32b:free | free\nqwen/qwq-32b-preview | paid\nqwen/qwen-2.5-72b-instruct | paid\nqwen/qwen-2.5-72b-instruct:free | free\nqwen/qwen-2.5-7b-instruct | paid\nqwen/qwen-2.5-coder-32b-instruct | paid\nqwen/qwen-2.5-coder-32b-instruct:free | free\nrekaai/reka-flash-3 | paid\nrekaai/reka-flash-3:free | free\nundi95/remm-slerp-l2-13b | paid\nsao10k/l3-lunaris-8b | paid\nsao10k/l3-euryale-70b | paid\nsao10k/l3.1-euryale-70b | paid\nsao10k/l3.3-euryale-70b | paid\nsarvamai/sarvam-m | paid\nsarvamai/sarvam-m:free | free\nshisa-ai/shisa-v2-llama3.3-70b | paid\nshisa-ai/shisa-v2-llama3.3-70b:free | free\nraifle/sorcererlm-8x22b | paid\nswitchpoint/router | paid\ntencent/hunyuan-a13b-instruct | paid\ntencent/hunyuan-a13b-instruct:free | free\nthedrummer/anubis-70b-v1.1 | paid\nthedrummer/anubis-pro-105b-v1 | paid\nthedrummer/rocinante-12b | paid\nthedrummer/skyfall-36b-v2 | paid\nthedrummer/unslopnemo-12b | paid\nthedrummer/valkyrie-49b-v1 | paid\nthudm/glm-4-32b | paid\nthudm/glm-4-32b:free | free\nthudm/glm-4.1v-9b-thinking | paid\nthudm/glm-z1-32b | paid\nthudm/glm-z1-32b:free | free\ntngtech/deepseek-r1t-chimera:free | free\ntngtech/deepseek-r1t2-chimera | paid\ntngtech/deepseek-r1t2-chimera:free | free\nundi95/toppy-m-7b | paid\nscb10x/llama3.1-typhoon2-70b-instruct | paid\ncognitivecomputations/dolphin-mistral-24b-venice-edition:free | free\nmicrosoft/wizardlm-2-8x22b | paid\nx-ai/grok-2-1212 | paid\nx-ai/grok-2-vision-1212 | paid\nx-ai/grok-3 | paid\nx-ai/grok-3-beta | paid\nx-ai/grok-3-mini | paid\nx-ai/grok-3-mini-beta | paid\nx-ai/grok-4 | paid\nx-ai/grok-vision-beta | paid\nz-ai/glm-4-32b | paid\n\u001b[35m\u001b[1m\u001b[22m\u001b[39m\n\u001b[35m\u001b[1m OpenAI models:\u001b[22m\u001b[39m\n\u001b[35m\u001b[1m\u001b[22m\u001b[39m\nbabbage-002\nchatgpt-4o-latest\ncodex-mini-latest\ndall-e-2\ndall-e-3\ndavinci-002\ngpt-3.5-turbo\ngpt-3.5-turbo-0125\ngpt-3.5-turbo-1106\ngpt-3.5-turbo-16k\ngpt-3.5-turbo-instruct\ngpt-3.5-turbo-instruct-0914\ngpt-4\ngpt-4-0125-preview\ngpt-4-0613\ngpt-4-1106-preview\ngpt-4-turbo\ngpt-4-turbo-2024-04-09\ngpt-4-turbo-preview\ngpt-4.1\ngpt-4.1-2025-04-14\ngpt-4.1-mini\ngpt-4.1-mini-2025-04-14\ngpt-4.1-nano\ngpt-4.1-nano-2025-04-14\ngpt-4o\ngpt-4o-2024-05-13\ngpt-4o-2024-08-06\ngpt-4o-2024-11-20\ngpt-4o-audio-preview\ngpt-4o-audio-preview-2024-10-01\ngpt-4o-audio-preview-2024-12-17\ngpt-4o-audio-preview-2025-06-03\ngpt-4o-mini\ngpt-4o-mini-2024-07-18\ngpt-4o-mini-audio-preview\ngpt-4o-mini-audio-preview-2024-12-17\ngpt-4o-mini-realtime-preview\ngpt-4o-mini-realtime-preview-2024-12-17\ngpt-4o-mini-search-preview\ngpt-4o-mini-search-preview-2025-03-11\ngpt-4o-mini-transcribe\ngpt-4o-mini-tts\ngpt-4o-realtime-preview\ngpt-4o-realtime-preview-2024-10-01\ngpt-4o-realtime-preview-2024-12-17\ngpt-4o-realtime-preview-2025-06-03\ngpt-4o-search-preview\ngpt-4o-search-preview-2025-03-11\ngpt-4o-transcribe\ngpt-image-1\no1\no1-2024-12-17\no1-mini\no1-mini-2024-09-12\no1-preview\no1-preview-2024-09-12\no1-pro\no1-pro-2025-03-19\no3-mini\no3-mini-2025-01-31\no4-mini\no4-mini-2025-04-16\no4-mini-deep-research\no4-mini-deep-research-2025-06-26\nomni-moderation-2024-09-26\nomni-moderation-latest\ntext-embedding-3-large\ntext-embedding-3-small\ntext-embedding-ada-002\ntts-1\ntts-1-1106\ntts-1-hd\ntts-1-hd-1106\nwhisper-1\n-----\n\n\u001b[35m\u001b[1m\u001b[22m\u001b[39m\n\u001b[35m\u001b[1m Deepseek models:\u001b[22m\u001b[39m\n\u001b[35m\u001b[1m\u001b[22m\u001b[39m\ndeepseek-chat\ndeepseek-reasoner\n-----\n", "ui:title": "Model" }, "router": { diff --git a/packages/kbot/src/models/cache/openai-models.ts b/packages/kbot/src/models/cache/openai-models.ts index 2e5de895..134bece1 100644 --- a/packages/kbot/src/models/cache/openai-models.ts +++ b/packages/kbot/src/models/cache/openai-models.ts @@ -31,8 +31,6 @@ export enum E_OPENAI_MODEL { MODEL_GPT_4O_MINI = "gpt-4o-mini", MODEL_GPT_4O_2024_08_06 = "gpt-4o-2024-08-06", MODEL_CHATGPT_4O_LATEST = "chatgpt-4o-latest", - MODEL_O1_PREVIEW_2024_09_12 = "o1-preview-2024-09-12", - MODEL_O1_PREVIEW = "o1-preview", MODEL_O1_MINI_2024_09_12 = "o1-mini-2024-09-12", MODEL_O1_MINI = "o1-mini", MODEL_GPT_4O_REALTIME_PREVIEW_2024_10_01 = "gpt-4o-realtime-preview-2024-10-01", diff --git a/packages/kbot/src/models/cache/openrouter-models-free.ts b/packages/kbot/src/models/cache/openrouter-models-free.ts index 141e621a..0c25d970 100644 --- a/packages/kbot/src/models/cache/openrouter-models-free.ts +++ b/packages/kbot/src/models/cache/openrouter-models-free.ts @@ -1,6 +1,6 @@ export enum E_OPENROUTER_MODEL_FREE { - MODEL_FREE_QWEN_QWEN3_CODER_FREE = "qwen/qwen3-coder:free", - MODEL_FREE_QWEN_QWEN3_235B_A22B_2507_FREE = "qwen/qwen3-235b-a22b-2507:free", + MODEL_FREE_OPENROUTER_HORIZON_BETA = "openrouter/horizon-beta", + MODEL_FREE_Z_AI_GLM_4_5_AIR_FREE = "z-ai/glm-4.5-air:free", MODEL_FREE_MOONSHOTAI_KIMI_K2_FREE = "moonshotai/kimi-k2:free", MODEL_FREE_COGNITIVECOMPUTATIONS_DOLPHIN_MISTRAL_24B_VENICE_EDITION_FREE = "cognitivecomputations/dolphin-mistral-24b-venice-edition:free", MODEL_FREE_GOOGLE_GEMMA_3N_E2B_IT_FREE = "google/gemma-3n-e2b-it:free", @@ -21,7 +21,6 @@ export enum E_OPENROUTER_MODEL_FREE { MODEL_FREE_TNGTECH_DEEPSEEK_R1T_CHIMERA_FREE = "tngtech/deepseek-r1t-chimera:free", MODEL_FREE_MICROSOFT_MAI_DS_R1_FREE = "microsoft/mai-ds-r1:free", MODEL_FREE_THUDM_GLM_Z1_32B_FREE = "thudm/glm-z1-32b:free", - MODEL_FREE_THUDM_GLM_4_32B_FREE = "thudm/glm-4-32b:free", MODEL_FREE_SHISA_AI_SHISA_V2_LLAMA3_3_70B_FREE = "shisa-ai/shisa-v2-llama3.3-70b:free", MODEL_FREE_ARLIAI_QWQ_32B_ARLIAI_RPR_V1_FREE = "arliai/qwq-32b-arliai-rpr-v1:free", MODEL_FREE_AGENTICA_ORG_DEEPCODER_14B_PREVIEW_FREE = "agentica-org/deepcoder-14b-preview:free", diff --git a/packages/kbot/src/models/cache/openrouter-models.ts b/packages/kbot/src/models/cache/openrouter-models.ts index 2b235fc3..d3834402 100644 --- a/packages/kbot/src/models/cache/openrouter-models.ts +++ b/packages/kbot/src/models/cache/openrouter-models.ts @@ -1,11 +1,18 @@ export enum E_OPENROUTER_MODEL { + MODEL_OPENAI_GPT_OSS_120B = "openai/gpt-oss-120b", + MODEL_OPENAI_GPT_OSS_20B = "openai/gpt-oss-20b", + MODEL_ANTHROPIC_CLAUDE_OPUS_4_1 = "anthropic/claude-opus-4.1", + MODEL_OPENROUTER_HORIZON_BETA = "openrouter/horizon-beta", + MODEL_MISTRALAI_CODESTRAL_2508 = "mistralai/codestral-2508", + MODEL_QWEN_QWEN3_30B_A3B_INSTRUCT_2507 = "qwen/qwen3-30b-a3b-instruct-2507", + MODEL_Z_AI_GLM_4_5 = "z-ai/glm-4.5", + MODEL_Z_AI_GLM_4_5_AIR_FREE = "z-ai/glm-4.5-air:free", + MODEL_Z_AI_GLM_4_5_AIR = "z-ai/glm-4.5-air", MODEL_QWEN_QWEN3_235B_A22B_THINKING_2507 = "qwen/qwen3-235b-a22b-thinking-2507", MODEL_Z_AI_GLM_4_32B = "z-ai/glm-4-32b", - MODEL_QWEN_QWEN3_CODER_FREE = "qwen/qwen3-coder:free", MODEL_QWEN_QWEN3_CODER = "qwen/qwen3-coder", MODEL_BYTEDANCE_UI_TARS_1_5_7B = "bytedance/ui-tars-1.5-7b", MODEL_GOOGLE_GEMINI_2_5_FLASH_LITE = "google/gemini-2.5-flash-lite", - MODEL_QWEN_QWEN3_235B_A22B_2507_FREE = "qwen/qwen3-235b-a22b-2507:free", MODEL_QWEN_QWEN3_235B_A22B_2507 = "qwen/qwen3-235b-a22b-2507", MODEL_SWITCHPOINT_ROUTER = "switchpoint/router", MODEL_MOONSHOTAI_KIMI_K2_FREE = "moonshotai/kimi-k2:free", @@ -19,13 +26,11 @@ export enum E_OPENROUTER_MODEL { MODEL_TENCENT_HUNYUAN_A13B_INSTRUCT_FREE = "tencent/hunyuan-a13b-instruct:free", MODEL_TENCENT_HUNYUAN_A13B_INSTRUCT = "tencent/hunyuan-a13b-instruct", MODEL_TNGTECH_DEEPSEEK_R1T2_CHIMERA_FREE = "tngtech/deepseek-r1t2-chimera:free", - MODEL_TNGTECH_DEEPSEEK_R1T2_CHIMERA = "tngtech/deepseek-r1t2-chimera", MODEL_MORPH_MORPH_V3_LARGE = "morph/morph-v3-large", MODEL_MORPH_MORPH_V3_FAST = "morph/morph-v3-fast", MODEL_BAIDU_ERNIE_4_5_300B_A47B = "baidu/ernie-4.5-300b-a47b", MODEL_THEDRUMMER_ANUBIS_70B_V1_1 = "thedrummer/anubis-70b-v1.1", MODEL_INCEPTION_MERCURY = "inception/mercury", - MODEL_MORPH_MORPH_V2 = "morph/morph-v2", MODEL_MISTRALAI_MISTRAL_SMALL_3_2_24B_INSTRUCT_FREE = "mistralai/mistral-small-3.2-24b-instruct:free", MODEL_MISTRALAI_MISTRAL_SMALL_3_2_24B_INSTRUCT = "mistralai/mistral-small-3.2-24b-instruct", MODEL_MINIMAX_MINIMAX_M1 = "minimax/minimax-m1", @@ -46,7 +51,6 @@ export enum E_OPENROUTER_MODEL { MODEL_DEEPSEEK_DEEPSEEK_R1_0528_FREE = "deepseek/deepseek-r1-0528:free", MODEL_DEEPSEEK_DEEPSEEK_R1_0528 = "deepseek/deepseek-r1-0528", MODEL_SARVAMAI_SARVAM_M_FREE = "sarvamai/sarvam-m:free", - MODEL_SARVAMAI_SARVAM_M = "sarvamai/sarvam-m", MODEL_THEDRUMMER_VALKYRIE_49B_V1 = "thedrummer/valkyrie-49b-v1", MODEL_ANTHROPIC_CLAUDE_OPUS_4 = "anthropic/claude-opus-4", MODEL_ANTHROPIC_CLAUDE_SONNET_4 = "anthropic/claude-sonnet-4", @@ -78,11 +82,10 @@ export enum E_OPENROUTER_MODEL { MODEL_QWEN_QWEN3_235B_A22B_FREE = "qwen/qwen3-235b-a22b:free", MODEL_QWEN_QWEN3_235B_A22B = "qwen/qwen3-235b-a22b", MODEL_TNGTECH_DEEPSEEK_R1T_CHIMERA_FREE = "tngtech/deepseek-r1t-chimera:free", + MODEL_TNGTECH_DEEPSEEK_R1T_CHIMERA = "tngtech/deepseek-r1t-chimera", MODEL_MICROSOFT_MAI_DS_R1_FREE = "microsoft/mai-ds-r1:free", MODEL_MICROSOFT_MAI_DS_R1 = "microsoft/mai-ds-r1", MODEL_THUDM_GLM_Z1_32B_FREE = "thudm/glm-z1-32b:free", - MODEL_THUDM_GLM_Z1_32B = "thudm/glm-z1-32b", - MODEL_THUDM_GLM_4_32B_FREE = "thudm/glm-4-32b:free", MODEL_THUDM_GLM_4_32B = "thudm/glm-4-32b", MODEL_OPENAI_O4_MINI_HIGH = "openai/o4-mini-high", MODEL_OPENAI_O3 = "openai/o3", @@ -128,7 +131,6 @@ export enum E_OPENROUTER_MODEL { MODEL_OPENAI_GPT_4O_MINI_SEARCH_PREVIEW = "openai/gpt-4o-mini-search-preview", MODEL_OPENAI_GPT_4O_SEARCH_PREVIEW = "openai/gpt-4o-search-preview", MODEL_REKAAI_REKA_FLASH_3_FREE = "rekaai/reka-flash-3:free", - MODEL_REKAAI_REKA_FLASH_3 = "rekaai/reka-flash-3", MODEL_GOOGLE_GEMMA_3_27B_IT_FREE = "google/gemma-3-27b-it:free", MODEL_GOOGLE_GEMMA_3_27B_IT = "google/gemma-3-27b-it", MODEL_THEDRUMMER_ANUBIS_PRO_105B_V1 = "thedrummer/anubis-pro-105b-v1", @@ -149,6 +151,7 @@ export enum E_OPENROUTER_MODEL { MODEL_COGNITIVECOMPUTATIONS_DOLPHIN3_0_R1_MISTRAL_24B_FREE = "cognitivecomputations/dolphin3.0-r1-mistral-24b:free", MODEL_COGNITIVECOMPUTATIONS_DOLPHIN3_0_R1_MISTRAL_24B = "cognitivecomputations/dolphin3.0-r1-mistral-24b", MODEL_COGNITIVECOMPUTATIONS_DOLPHIN3_0_MISTRAL_24B_FREE = "cognitivecomputations/dolphin3.0-mistral-24b:free", + MODEL_COGNITIVECOMPUTATIONS_DOLPHIN3_0_MISTRAL_24B = "cognitivecomputations/dolphin3.0-mistral-24b", MODEL_META_LLAMA_LLAMA_GUARD_3_8B = "meta-llama/llama-guard-3-8b", MODEL_OPENAI_O3_MINI_HIGH = "openai/o3-mini-high", MODEL_DEEPSEEK_DEEPSEEK_R1_DISTILL_LLAMA_8B = "deepseek/deepseek-r1-distill-llama-8b", @@ -194,7 +197,6 @@ export enum E_OPENROUTER_MODEL { MODEL_AMAZON_NOVA_MICRO_V1 = "amazon/nova-micro-v1", MODEL_AMAZON_NOVA_PRO_V1 = "amazon/nova-pro-v1", MODEL_QWEN_QWQ_32B_PREVIEW = "qwen/qwq-32b-preview", - MODEL_EVA_UNIT_01_EVA_QWEN_2_5_72B = "eva-unit-01/eva-qwen-2.5-72b", MODEL_OPENAI_GPT_4O_2024_11_20 = "openai/gpt-4o-2024-11-20", MODEL_MISTRALAI_MISTRAL_LARGE_2411 = "mistralai/mistral-large-2411", MODEL_MISTRALAI_MISTRAL_LARGE_2407 = "mistralai/mistral-large-2407", @@ -207,7 +209,6 @@ export enum E_OPENROUTER_MODEL { MODEL_THEDRUMMER_UNSLOPNEMO_12B = "thedrummer/unslopnemo-12b", MODEL_ANTHROPIC_CLAUDE_3_5_HAIKU_BETA = "anthropic/claude-3.5-haiku:beta", MODEL_ANTHROPIC_CLAUDE_3_5_HAIKU = "anthropic/claude-3.5-haiku", - MODEL_ANTHROPIC_CLAUDE_3_5_HAIKU_20241022_BETA = "anthropic/claude-3.5-haiku-20241022:beta", MODEL_ANTHROPIC_CLAUDE_3_5_HAIKU_20241022 = "anthropic/claude-3.5-haiku-20241022", MODEL_ANTHRACITE_ORG_MAGNUM_V4_72B = "anthracite-org/magnum-v4-72b", MODEL_ANTHROPIC_CLAUDE_3_5_SONNET_BETA = "anthropic/claude-3.5-sonnet:beta", @@ -216,24 +217,22 @@ export enum E_OPENROUTER_MODEL { MODEL_MISTRALAI_MINISTRAL_3B = "mistralai/ministral-3b", MODEL_QWEN_QWEN_2_5_7B_INSTRUCT = "qwen/qwen-2.5-7b-instruct", MODEL_NVIDIA_LLAMA_3_1_NEMOTRON_70B_INSTRUCT = "nvidia/llama-3.1-nemotron-70b-instruct", - MODEL_INFLECTION_INFLECTION_3_PI = "inflection/inflection-3-pi", MODEL_INFLECTION_INFLECTION_3_PRODUCTIVITY = "inflection/inflection-3-productivity", + MODEL_INFLECTION_INFLECTION_3_PI = "inflection/inflection-3-pi", MODEL_GOOGLE_GEMINI_FLASH_1_5_8B = "google/gemini-flash-1.5-8b", - MODEL_ANTHRACITE_ORG_MAGNUM_V2_72B = "anthracite-org/magnum-v2-72b", MODEL_THEDRUMMER_ROCINANTE_12B = "thedrummer/rocinante-12b", MODEL_LIQUID_LFM_40B = "liquid/lfm-40b", + MODEL_ANTHRACITE_ORG_MAGNUM_V2_72B = "anthracite-org/magnum-v2-72b", MODEL_META_LLAMA_LLAMA_3_2_3B_INSTRUCT_FREE = "meta-llama/llama-3.2-3b-instruct:free", MODEL_META_LLAMA_LLAMA_3_2_3B_INSTRUCT = "meta-llama/llama-3.2-3b-instruct", + MODEL_META_LLAMA_LLAMA_3_2_90B_VISION_INSTRUCT = "meta-llama/llama-3.2-90b-vision-instruct", MODEL_META_LLAMA_LLAMA_3_2_1B_INSTRUCT = "meta-llama/llama-3.2-1b-instruct", MODEL_META_LLAMA_LLAMA_3_2_11B_VISION_INSTRUCT_FREE = "meta-llama/llama-3.2-11b-vision-instruct:free", MODEL_META_LLAMA_LLAMA_3_2_11B_VISION_INSTRUCT = "meta-llama/llama-3.2-11b-vision-instruct", - MODEL_META_LLAMA_LLAMA_3_2_90B_VISION_INSTRUCT = "meta-llama/llama-3.2-90b-vision-instruct", MODEL_QWEN_QWEN_2_5_72B_INSTRUCT_FREE = "qwen/qwen-2.5-72b-instruct:free", MODEL_QWEN_QWEN_2_5_72B_INSTRUCT = "qwen/qwen-2.5-72b-instruct", MODEL_NEVERSLEEP_LLAMA_3_1_LUMIMAID_8B = "neversleep/llama-3.1-lumimaid-8b", - MODEL_OPENAI_O1_PREVIEW_2024_09_12 = "openai/o1-preview-2024-09-12", MODEL_OPENAI_O1_MINI_2024_09_12 = "openai/o1-mini-2024-09-12", - MODEL_OPENAI_O1_PREVIEW = "openai/o1-preview", MODEL_OPENAI_O1_MINI = "openai/o1-mini", MODEL_MISTRALAI_PIXTRAL_12B = "mistralai/pixtral-12b", MODEL_COHERE_COMMAND_R_08_2024 = "cohere/command-r-08-2024", @@ -247,19 +246,17 @@ export enum E_OPENROUTER_MODEL { MODEL_SAO10K_L3_LUNARIS_8B = "sao10k/l3-lunaris-8b", MODEL_OPENAI_GPT_4O_2024_08_06 = "openai/gpt-4o-2024-08-06", MODEL_META_LLAMA_LLAMA_3_1_405B = "meta-llama/llama-3.1-405b", - MODEL_NOTHINGIISREAL_MN_CELESTE_12B = "nothingiisreal/mn-celeste-12b", MODEL_META_LLAMA_LLAMA_3_1_70B_INSTRUCT = "meta-llama/llama-3.1-70b-instruct", - MODEL_META_LLAMA_LLAMA_3_1_8B_INSTRUCT = "meta-llama/llama-3.1-8b-instruct", MODEL_META_LLAMA_LLAMA_3_1_405B_INSTRUCT_FREE = "meta-llama/llama-3.1-405b-instruct:free", MODEL_META_LLAMA_LLAMA_3_1_405B_INSTRUCT = "meta-llama/llama-3.1-405b-instruct", + MODEL_META_LLAMA_LLAMA_3_1_8B_INSTRUCT = "meta-llama/llama-3.1-8b-instruct", MODEL_MISTRALAI_MISTRAL_NEMO_FREE = "mistralai/mistral-nemo:free", MODEL_MISTRALAI_MISTRAL_NEMO = "mistralai/mistral-nemo", - MODEL_OPENAI_GPT_4O_MINI = "openai/gpt-4o-mini", MODEL_OPENAI_GPT_4O_MINI_2024_07_18 = "openai/gpt-4o-mini-2024-07-18", + MODEL_OPENAI_GPT_4O_MINI = "openai/gpt-4o-mini", MODEL_GOOGLE_GEMMA_2_27B_IT = "google/gemma-2-27b-it", MODEL_GOOGLE_GEMMA_2_9B_IT_FREE = "google/gemma-2-9b-it:free", MODEL_GOOGLE_GEMMA_2_9B_IT = "google/gemma-2-9b-it", - MODEL_01_AI_YI_LARGE = "01-ai/yi-large", MODEL_ANTHROPIC_CLAUDE_3_5_SONNET_20240620_BETA = "anthropic/claude-3.5-sonnet-20240620:beta", MODEL_ANTHROPIC_CLAUDE_3_5_SONNET_20240620 = "anthropic/claude-3.5-sonnet-20240620", MODEL_SAO10K_L3_EURYALE_70B = "sao10k/l3-euryale-70b", @@ -274,16 +271,16 @@ export enum E_OPENROUTER_MODEL { MODEL_NEVERSLEEP_LLAMA_3_LUMIMAID_70B = "neversleep/llama-3-lumimaid-70b", MODEL_GOOGLE_GEMINI_FLASH_1_5 = "google/gemini-flash-1.5", MODEL_META_LLAMA_LLAMA_GUARD_2_8B = "meta-llama/llama-guard-2-8b", + MODEL_OPENAI_GPT_4O_2024_05_13 = "openai/gpt-4o-2024-05-13", MODEL_OPENAI_GPT_4O = "openai/gpt-4o", MODEL_OPENAI_GPT_4O_EXTENDED = "openai/gpt-4o:extended", - MODEL_OPENAI_GPT_4O_2024_05_13 = "openai/gpt-4o-2024-05-13", MODEL_SAO10K_FIMBULVETR_11B_V2 = "sao10k/fimbulvetr-11b-v2", - MODEL_META_LLAMA_LLAMA_3_70B_INSTRUCT = "meta-llama/llama-3-70b-instruct", MODEL_META_LLAMA_LLAMA_3_8B_INSTRUCT = "meta-llama/llama-3-8b-instruct", + MODEL_META_LLAMA_LLAMA_3_70B_INSTRUCT = "meta-llama/llama-3-70b-instruct", MODEL_MISTRALAI_MIXTRAL_8X22B_INSTRUCT = "mistralai/mixtral-8x22b-instruct", MODEL_MICROSOFT_WIZARDLM_2_8X22B = "microsoft/wizardlm-2-8x22b", - MODEL_OPENAI_GPT_4_TURBO = "openai/gpt-4-turbo", MODEL_GOOGLE_GEMINI_PRO_1_5 = "google/gemini-pro-1.5", + MODEL_OPENAI_GPT_4_TURBO = "openai/gpt-4-turbo", MODEL_COHERE_COMMAND_R_PLUS = "cohere/command-r-plus", MODEL_COHERE_COMMAND_R_PLUS_04_2024 = "cohere/command-r-plus-04-2024", MODEL_SOPHOSYMPATHEIA_MIDNIGHT_ROSE_70B = "sophosympatheia/midnight-rose-70b", @@ -291,31 +288,30 @@ export enum E_OPENROUTER_MODEL { MODEL_COHERE_COMMAND_R = "cohere/command-r", MODEL_ANTHROPIC_CLAUDE_3_HAIKU_BETA = "anthropic/claude-3-haiku:beta", MODEL_ANTHROPIC_CLAUDE_3_HAIKU = "anthropic/claude-3-haiku", - MODEL_ANTHROPIC_CLAUDE_3_SONNET = "anthropic/claude-3-sonnet", MODEL_ANTHROPIC_CLAUDE_3_OPUS_BETA = "anthropic/claude-3-opus:beta", MODEL_ANTHROPIC_CLAUDE_3_OPUS = "anthropic/claude-3-opus", MODEL_COHERE_COMMAND_R_03_2024 = "cohere/command-r-03-2024", MODEL_MISTRALAI_MISTRAL_LARGE = "mistralai/mistral-large", - MODEL_OPENAI_GPT_4_TURBO_PREVIEW = "openai/gpt-4-turbo-preview", MODEL_OPENAI_GPT_3_5_TURBO_0613 = "openai/gpt-3.5-turbo-0613", + MODEL_OPENAI_GPT_4_TURBO_PREVIEW = "openai/gpt-4-turbo-preview", MODEL_NOUSRESEARCH_NOUS_HERMES_2_MIXTRAL_8X7B_DPO = "nousresearch/nous-hermes-2-mixtral-8x7b-dpo", MODEL_MISTRALAI_MISTRAL_SMALL = "mistralai/mistral-small", MODEL_MISTRALAI_MISTRAL_TINY = "mistralai/mistral-tiny", MODEL_MISTRALAI_MISTRAL_7B_INSTRUCT_V0_2 = "mistralai/mistral-7b-instruct-v0.2", MODEL_MISTRALAI_MIXTRAL_8X7B_INSTRUCT = "mistralai/mixtral-8x7b-instruct", MODEL_NEVERSLEEP_NOROMAID_20B = "neversleep/noromaid-20b", - MODEL_UNDI95_TOPPY_M_7B = "undi95/toppy-m-7b", MODEL_ALPINDALE_GOLIATH_120B = "alpindale/goliath-120b", + MODEL_UNDI95_TOPPY_M_7B = "undi95/toppy-m-7b", MODEL_OPENROUTER_AUTO = "openrouter/auto", MODEL_OPENAI_GPT_4_1106_PREVIEW = "openai/gpt-4-1106-preview", - MODEL_OPENAI_GPT_3_5_TURBO_INSTRUCT = "openai/gpt-3.5-turbo-instruct", MODEL_MISTRALAI_MISTRAL_7B_INSTRUCT_V0_1 = "mistralai/mistral-7b-instruct-v0.1", + MODEL_OPENAI_GPT_3_5_TURBO_INSTRUCT = "openai/gpt-3.5-turbo-instruct", MODEL_PYGMALIONAI_MYTHALION_13B = "pygmalionai/mythalion-13b", MODEL_OPENAI_GPT_3_5_TURBO_16K = "openai/gpt-3.5-turbo-16k", MODEL_MANCER_WEAVER = "mancer/weaver", MODEL_UNDI95_REMM_SLERP_L2_13B = "undi95/remm-slerp-l2-13b", MODEL_GRYPHE_MYTHOMAX_L2_13B = "gryphe/mythomax-l2-13b", - MODEL_OPENAI_GPT_4_0314 = "openai/gpt-4-0314", + MODEL_OPENAI_GPT_4 = "openai/gpt-4", MODEL_OPENAI_GPT_3_5_TURBO = "openai/gpt-3.5-turbo", - MODEL_OPENAI_GPT_4 = "openai/gpt-4" + MODEL_OPENAI_GPT_4_0314 = "openai/gpt-4-0314" } \ No newline at end of file diff --git a/packages/kbot/src/zod_types.ts b/packages/kbot/src/zod_types.ts index 99a7db72..c003aa64 100644 --- a/packages/kbot/src/zod_types.ts +++ b/packages/kbot/src/zod_types.ts @@ -32,7 +32,6 @@ export interface IKBotOptions {  OpenRouter models:  01-ai/yi-large | paid - aetherwiing/mn-starcannon-12b | paid agentica-org/deepcoder-14b-preview | paid agentica-org/deepcoder-14b-preview:free | free ai21/jamba-1.6-large | paid @@ -49,7 +48,6 @@ export interface IKBotOptions { anthropic/claude-3-opus | paid anthropic/claude-3-opus:beta | paid anthropic/claude-3-sonnet | paid - anthropic/claude-3-sonnet:beta | paid anthropic/claude-3.5-haiku | paid anthropic/claude-3.5-haiku-20241022 | paid anthropic/claude-3.5-haiku-20241022:beta | paid @@ -63,22 +61,15 @@ export interface IKBotOptions { anthropic/claude-3.7-sonnet:thinking | paid anthropic/claude-opus-4 | paid anthropic/claude-sonnet-4 | paid - anthropic/claude-2 | paid - anthropic/claude-2:beta | paid - anthropic/claude-2.0 | paid - anthropic/claude-2.0:beta | paid - anthropic/claude-2.1 | paid - anthropic/claude-2.1:beta | paid - arcee-ai/arcee-blitz | paid - arcee-ai/caller-large | paid arcee-ai/coder-large | paid arcee-ai/maestro-reasoning | paid arcee-ai/spotlight | paid arcee-ai/virtuoso-large | paid - arcee-ai/virtuoso-medium-v2 | paid + arliai/qwq-32b-arliai-rpr-v1 | paid arliai/qwq-32b-arliai-rpr-v1:free | free openrouter/auto | paid baidu/ernie-4.5-300b-a47b | paid + bytedance/ui-tars-1.5-7b | paid cohere/command | paid cohere/command-a | paid cohere/command-r | paid @@ -92,10 +83,9 @@ export interface IKBotOptions { deepseek/deepseek-r1-0528-qwen3-8b | paid deepseek/deepseek-r1-0528-qwen3-8b:free | free deepseek/deepseek-chat | paid - deepseek/deepseek-chat:free | free deepseek/deepseek-chat-v3-0324 | paid deepseek/deepseek-chat-v3-0324:free | free - deepseek/deepseek-v3-base:free | free + deepseek/deepseek-v3-base | paid deepseek/deepseek-r1 | paid deepseek/deepseek-r1:free | free deepseek/deepseek-r1-0528 | paid @@ -113,7 +103,6 @@ export interface IKBotOptions { cognitivecomputations/dolphin3.0-r1-mistral-24b | paid cognitivecomputations/dolphin3.0-r1-mistral-24b:free | free eleutherai/llemma_7b | paid - eva-unit-01/eva-llama-3.33-70b | paid eva-unit-01/eva-qwen-2.5-72b | paid sao10k/fimbulvetr-11b-v2 | paid alpindale/goliath-120b | paid @@ -124,6 +113,7 @@ export interface IKBotOptions { google/gemini-2.0-flash-exp:free | free google/gemini-2.0-flash-lite-001 | paid google/gemini-2.5-flash | paid + google/gemini-2.5-flash-lite | paid google/gemini-2.5-flash-lite-preview-06-17 | paid google/gemini-2.5-pro | paid google/gemini-2.5-pro-exp-03-25 | paid @@ -151,7 +141,6 @@ export interface IKBotOptions { liquid/lfm-40b | paid liquid/lfm-7b | paid meta-llama/llama-guard-3-8b | paid - alpindale/magnum-72b | paid anthracite-org/magnum-v2-72b | paid anthracite-org/magnum-v4-72b | paid mancer/weaver | paid @@ -174,6 +163,7 @@ export interface IKBotOptions { meta-llama/llama-4-scout | paid meta-llama/llama-guard-4-12b | paid meta-llama/llama-guard-2-8b | paid + microsoft/mai-ds-r1 | paid microsoft/mai-ds-r1:free | free microsoft/phi-4 | paid microsoft/phi-4-multimodal-instruct | paid @@ -219,6 +209,7 @@ export interface IKBotOptions { mistralai/pixtral-12b | paid mistralai/pixtral-large-2411 | paid mistralai/mistral-saba | paid + moonshotai/kimi-vl-a3b-thinking | paid moonshotai/kimi-vl-a3b-thinking:free | free moonshotai/kimi-k2 | paid moonshotai/kimi-k2:free | free @@ -230,6 +221,7 @@ export interface IKBotOptions { neversleep/llama-3.1-lumimaid-8b | paid neversleep/noromaid-20b | paid nousresearch/deephermes-3-llama-3-8b-preview:free | free + nousresearch/deephermes-3-mistral-24b-preview | paid nousresearch/nous-hermes-2-mixtral-8x7b-dpo | paid nousresearch/hermes-3-llama-3.1-405b | paid nousresearch/hermes-3-llama-3.1-70b | paid @@ -281,6 +273,7 @@ export interface IKBotOptions { perplexity/sonar-reasoning | paid perplexity/sonar-reasoning-pro | paid pygmalionai/mythalion-13b | paid + featherless/qwerky-72b:free | free qwen/qwen-2-72b-instruct | paid qwen/qwen-vl-max | paid qwen/qwen-vl-plus | paid @@ -296,13 +289,17 @@ export interface IKBotOptions { qwen/qwen3-14b:free | free qwen/qwen3-235b-a22b | paid qwen/qwen3-235b-a22b:free | free + qwen/qwen3-235b-a22b-2507 | paid + qwen/qwen3-235b-a22b-2507:free | free + qwen/qwen3-235b-a22b-thinking-2507 | paid qwen/qwen3-30b-a3b | paid qwen/qwen3-30b-a3b:free | free qwen/qwen3-32b | paid - qwen/qwen3-32b:free | free qwen/qwen3-4b:free | free qwen/qwen3-8b | paid qwen/qwen3-8b:free | free + qwen/qwen3-coder | paid + qwen/qwen3-coder:free | free qwen/qwq-32b | paid qwen/qwq-32b:free | free qwen/qwq-32b-preview | paid @@ -311,7 +308,6 @@ export interface IKBotOptions { qwen/qwen-2.5-7b-instruct | paid qwen/qwen-2.5-coder-32b-instruct | paid qwen/qwen-2.5-coder-32b-instruct:free | free - featherless/qwerky-72b:free | free rekaai/reka-flash-3 | paid rekaai/reka-flash-3:free | free undi95/remm-slerp-l2-13b | paid @@ -321,6 +317,7 @@ export interface IKBotOptions { sao10k/l3.3-euryale-70b | paid sarvamai/sarvam-m | paid sarvamai/sarvam-m:free | free + shisa-ai/shisa-v2-llama3.3-70b | paid shisa-ai/shisa-v2-llama3.3-70b:free | free raifle/sorcererlm-8x22b | paid switchpoint/router | paid @@ -335,8 +332,10 @@ export interface IKBotOptions { thudm/glm-4-32b | paid thudm/glm-4-32b:free | free thudm/glm-4.1v-9b-thinking | paid + thudm/glm-z1-32b | paid thudm/glm-z1-32b:free | free tngtech/deepseek-r1t-chimera:free | free + tngtech/deepseek-r1t2-chimera | paid tngtech/deepseek-r1t2-chimera:free | free undi95/toppy-m-7b | paid scb10x/llama3.1-typhoon2-70b-instruct | paid @@ -350,6 +349,7 @@ export interface IKBotOptions { x-ai/grok-3-mini-beta | paid x-ai/grok-4 | paid x-ai/grok-vision-beta | paid + z-ai/glm-4-32b | paid   OpenAI models: