diff --git a/packages/kbot/dist-in/data/openai_models.json b/packages/kbot/dist-in/data/openai_models.json
index 0b0b48b5..eef14453 100644
--- a/packages/kbot/dist-in/data/openai_models.json
+++ b/packages/kbot/dist-in/data/openai_models.json
@@ -1,5 +1,5 @@
{
- "timestamp": 1749149893739,
+ "timestamp": 1751099810190,
"models": [
{
"id": "gpt-4-0613",
@@ -20,21 +20,9 @@
"owned_by": "openai"
},
{
- "id": "gpt-4o-audio-preview-2025-06-03",
+ "id": "o4-mini-deep-research-2025-06-26",
"object": "model",
- "created": 1748908498,
- "owned_by": "system"
- },
- {
- "id": "gpt-4.1-nano",
- "object": "model",
- "created": 1744321707,
- "owned_by": "system"
- },
- {
- "id": "gpt-image-1",
- "object": "model",
- "created": 1745517030,
+ "created": 1750866121,
"owned_by": "system"
},
{
@@ -49,6 +37,18 @@
"created": 1748907838,
"owned_by": "system"
},
+ {
+ "id": "gpt-4o-audio-preview-2025-06-03",
+ "object": "model",
+ "created": 1748908498,
+ "owned_by": "system"
+ },
+ {
+ "id": "o4-mini-deep-research",
+ "object": "model",
+ "created": 1749685485,
+ "owned_by": "system"
+ },
{
"id": "davinci-002",
"object": "model",
@@ -427,6 +427,18 @@
"created": 1744321025,
"owned_by": "system"
},
+ {
+ "id": "gpt-4.1-nano",
+ "object": "model",
+ "created": 1744321707,
+ "owned_by": "system"
+ },
+ {
+ "id": "gpt-image-1",
+ "object": "model",
+ "created": 1745517030,
+ "owned_by": "system"
+ },
{
"id": "gpt-3.5-turbo-16k",
"object": "model",
diff --git a/packages/kbot/dist-in/data/openrouter_models.json b/packages/kbot/dist-in/data/openrouter_models.json
index 50646e52..37d75320 100644
--- a/packages/kbot/dist-in/data/openrouter_models.json
+++ b/packages/kbot/dist-in/data/openrouter_models.json
@@ -1,8 +1,777 @@
{
- "timestamp": 1749149893852,
+ "timestamp": 1751099810425,
"models": [
+ {
+ "id": "inception/mercury",
+ "canonical_slug": "inception/mercury",
+ "hugging_face_id": "",
+ "name": "Inception: Mercury",
+ "created": 1750973026,
+ "description": "Mercury is the first diffusion large language model (dLLM). Applying a breakthrough discrete diffusion approach, the model runs 5-10x faster than even speed optimized models like GPT-4.1 Nano and Claude 3.5 Haiku while matching their performance. Mercury's speed enables developers to provide responsive user experiences, including with voice agents, search interfaces, and chatbots. Read more in the blog post here. ",
+ "context_length": 32000,
+ "architecture": {
+ "modality": "text->text",
+ "input_modalities": [
+ "text"
+ ],
+ "output_modalities": [
+ "text"
+ ],
+ "tokenizer": "Other",
+ "instruct_type": null
+ },
+ "pricing": {
+ "prompt": "0.00001",
+ "completion": "0.00001",
+ "request": "0",
+ "image": "0",
+ "web_search": "0",
+ "internal_reasoning": "0"
+ },
+ "top_provider": {
+ "context_length": 32000,
+ "max_completion_tokens": 16000,
+ "is_moderated": false
+ },
+ "per_request_limits": null,
+ "supported_parameters": [
+ "max_tokens",
+ "frequency_penalty",
+ "presence_penalty",
+ "stop"
+ ]
+ },
+ {
+ "id": "morph/morph-v2",
+ "canonical_slug": "morph/morph-v2",
+ "hugging_face_id": "",
+ "name": "Morph: Morph V2",
+ "created": 1750946108,
+ "description": "Morph Apply is a specialized code-patching LLM that merges AI-suggested edits straight into your source files. It can apply updates from GPT-4o, Claude, and others into your files at 4000+ tokens per second.\n\nThe model requires the prompt to be in the following format: \n${originalCode}\\n${updateSnippet}\n\nLearn more about this model in their [documentation](https://docs.morphllm.com/)",
+ "context_length": 16000,
+ "architecture": {
+ "modality": "text->text",
+ "input_modalities": [
+ "text"
+ ],
+ "output_modalities": [
+ "text"
+ ],
+ "tokenizer": "Other",
+ "instruct_type": null
+ },
+ "pricing": {
+ "prompt": "0.0000012",
+ "completion": "0.0000027",
+ "request": "0",
+ "image": "0",
+ "web_search": "0",
+ "internal_reasoning": "0"
+ },
+ "top_provider": {
+ "context_length": 16000,
+ "max_completion_tokens": 16000,
+ "is_moderated": false
+ },
+ "per_request_limits": null,
+ "supported_parameters": []
+ },
+ {
+ "id": "mistralai/mistral-small-3.2-24b-instruct:free",
+ "canonical_slug": "mistralai/mistral-small-3.2-24b-instruct-2506",
+ "hugging_face_id": "mistralai/Mistral-Small-3.2-24B-Instruct-2506",
+ "name": "Mistral: Mistral Small 3.2 24B (free)",
+ "created": 1750443016,
+ "description": "Mistral-Small-3.2-24B-Instruct-2506 is an updated 24B parameter model from Mistral optimized for instruction following, repetition reduction, and improved function calling. Compared to the 3.1 release, version 3.2 significantly improves accuracy on WildBench and Arena Hard, reduces infinite generations, and delivers gains in tool use and structured output tasks.\n\nIt supports image and text inputs with structured outputs, function/tool calling, and strong performance across coding (HumanEval+, MBPP), STEM (MMLU, MATH, GPQA), and vision benchmarks (ChartQA, DocVQA).",
+ "context_length": 96000,
+ "architecture": {
+ "modality": "text+image->text",
+ "input_modalities": [
+ "image",
+ "text"
+ ],
+ "output_modalities": [
+ "text"
+ ],
+ "tokenizer": "Mistral",
+ "instruct_type": null
+ },
+ "pricing": {
+ "prompt": "0",
+ "completion": "0",
+ "request": "0",
+ "image": "0",
+ "web_search": "0",
+ "internal_reasoning": "0"
+ },
+ "top_provider": {
+ "context_length": 96000,
+ "max_completion_tokens": null,
+ "is_moderated": false
+ },
+ "per_request_limits": null,
+ "supported_parameters": [
+ "tools",
+ "tool_choice",
+ "max_tokens",
+ "temperature",
+ "top_p",
+ "structured_outputs",
+ "stop",
+ "frequency_penalty",
+ "presence_penalty",
+ "seed",
+ "top_k",
+ "min_p",
+ "repetition_penalty",
+ "logprobs",
+ "logit_bias",
+ "top_logprobs"
+ ]
+ },
+ {
+ "id": "mistralai/mistral-small-3.2-24b-instruct",
+ "canonical_slug": "mistralai/mistral-small-3.2-24b-instruct-2506",
+ "hugging_face_id": "mistralai/Mistral-Small-3.2-24B-Instruct-2506",
+ "name": "Mistral: Mistral Small 3.2 24B",
+ "created": 1750443016,
+ "description": "Mistral-Small-3.2-24B-Instruct-2506 is an updated 24B parameter model from Mistral optimized for instruction following, repetition reduction, and improved function calling. Compared to the 3.1 release, version 3.2 significantly improves accuracy on WildBench and Arena Hard, reduces infinite generations, and delivers gains in tool use and structured output tasks.\n\nIt supports image and text inputs with structured outputs, function/tool calling, and strong performance across coding (HumanEval+, MBPP), STEM (MMLU, MATH, GPQA), and vision benchmarks (ChartQA, DocVQA).",
+ "context_length": 128000,
+ "architecture": {
+ "modality": "text+image->text",
+ "input_modalities": [
+ "image",
+ "text"
+ ],
+ "output_modalities": [
+ "text"
+ ],
+ "tokenizer": "Mistral",
+ "instruct_type": null
+ },
+ "pricing": {
+ "prompt": "0.00000005",
+ "completion": "0.0000001",
+ "request": "0",
+ "image": "0",
+ "web_search": "0",
+ "internal_reasoning": "0"
+ },
+ "top_provider": {
+ "context_length": 128000,
+ "max_completion_tokens": null,
+ "is_moderated": false
+ },
+ "per_request_limits": null,
+ "supported_parameters": [
+ "tools",
+ "tool_choice",
+ "max_tokens",
+ "temperature",
+ "top_p",
+ "stop",
+ "frequency_penalty",
+ "presence_penalty",
+ "response_format",
+ "structured_outputs",
+ "seed",
+ "repetition_penalty",
+ "top_k",
+ "min_p"
+ ]
+ },
+ {
+ "id": "minimax/minimax-m1:extended",
+ "canonical_slug": "minimax/minimax-m1",
+ "hugging_face_id": "",
+ "name": "MiniMax: MiniMax M1 (extended)",
+ "created": 1750200414,
+ "description": "MiniMax-M1 is a large-scale, open-weight reasoning model designed for extended context and high-efficiency inference. It leverages a hybrid Mixture-of-Experts (MoE) architecture paired with a custom \"lightning attention\" mechanism, allowing it to process long sequences—up to 1 million tokens—while maintaining competitive FLOP efficiency. With 456 billion total parameters and 45.9B active per token, this variant is optimized for complex, multi-step reasoning tasks.\n\nTrained via a custom reinforcement learning pipeline (CISPO), M1 excels in long-context understanding, software engineering, agentic tool use, and mathematical reasoning. Benchmarks show strong performance across FullStackBench, SWE-bench, MATH, GPQA, and TAU-Bench, often outperforming other open models like DeepSeek R1 and Qwen3-235B.",
+ "context_length": 256000,
+ "architecture": {
+ "modality": "text->text",
+ "input_modalities": [
+ "text"
+ ],
+ "output_modalities": [
+ "text"
+ ],
+ "tokenizer": "Other",
+ "instruct_type": null
+ },
+ "pricing": {
+ "prompt": "0",
+ "completion": "0",
+ "request": "0",
+ "image": "0",
+ "web_search": "0",
+ "internal_reasoning": "0"
+ },
+ "top_provider": {
+ "context_length": 256000,
+ "max_completion_tokens": null,
+ "is_moderated": false
+ },
+ "per_request_limits": null,
+ "supported_parameters": [
+ "tools",
+ "tool_choice",
+ "max_tokens",
+ "temperature",
+ "top_p",
+ "reasoning",
+ "include_reasoning",
+ "stop",
+ "frequency_penalty",
+ "presence_penalty",
+ "seed",
+ "top_k",
+ "min_p",
+ "repetition_penalty",
+ "logprobs",
+ "logit_bias",
+ "top_logprobs"
+ ]
+ },
+ {
+ "id": "minimax/minimax-m1",
+ "canonical_slug": "minimax/minimax-m1",
+ "hugging_face_id": "",
+ "name": "MiniMax: MiniMax M1",
+ "created": 1750200414,
+ "description": "MiniMax-M1 is a large-scale, open-weight reasoning model designed for extended context and high-efficiency inference. It leverages a hybrid Mixture-of-Experts (MoE) architecture paired with a custom \"lightning attention\" mechanism, allowing it to process long sequences—up to 1 million tokens—while maintaining competitive FLOP efficiency. With 456 billion total parameters and 45.9B active per token, this variant is optimized for complex, multi-step reasoning tasks.\n\nTrained via a custom reinforcement learning pipeline (CISPO), M1 excels in long-context understanding, software engineering, agentic tool use, and mathematical reasoning. Benchmarks show strong performance across FullStackBench, SWE-bench, MATH, GPQA, and TAU-Bench, often outperforming other open models like DeepSeek R1 and Qwen3-235B.",
+ "context_length": 1000000,
+ "architecture": {
+ "modality": "text->text",
+ "input_modalities": [
+ "text"
+ ],
+ "output_modalities": [
+ "text"
+ ],
+ "tokenizer": "Other",
+ "instruct_type": null
+ },
+ "pricing": {
+ "prompt": "0.0000003",
+ "completion": "0.00000165",
+ "request": "0",
+ "image": "0",
+ "web_search": "0",
+ "internal_reasoning": "0"
+ },
+ "top_provider": {
+ "context_length": 1000000,
+ "max_completion_tokens": 40000,
+ "is_moderated": false
+ },
+ "per_request_limits": null,
+ "supported_parameters": [
+ "tools",
+ "tool_choice",
+ "max_tokens",
+ "temperature",
+ "top_p",
+ "reasoning",
+ "include_reasoning",
+ "structured_outputs",
+ "stop",
+ "frequency_penalty",
+ "presence_penalty",
+ "seed",
+ "top_k",
+ "min_p",
+ "repetition_penalty",
+ "logit_bias"
+ ]
+ },
+ {
+ "id": "google/gemini-2.5-flash-lite-preview-06-17",
+ "canonical_slug": "google/gemini-2.5-flash-lite-preview-06-17",
+ "hugging_face_id": "",
+ "name": "Google: Gemini 2.5 Flash Lite Preview 06-17",
+ "created": 1750173831,
+ "description": "Gemini 2.5 Flash-Lite is a lightweight reasoning model in the Gemini 2.5 family, optimized for ultra-low latency and cost efficiency. It offers improved throughput, faster token generation, and better performance across common benchmarks compared to earlier Flash models. By default, \"thinking\" (i.e. multi-pass reasoning) is disabled to prioritize speed, but developers can enable it via the [Reasoning API parameter](https://openrouter.ai/docs/use-cases/reasoning-tokens) to selectively trade off cost for intelligence. ",
+ "context_length": 1048576,
+ "architecture": {
+ "modality": "text+image->text",
+ "input_modalities": [
+ "file",
+ "image",
+ "text"
+ ],
+ "output_modalities": [
+ "text"
+ ],
+ "tokenizer": "Gemini",
+ "instruct_type": null
+ },
+ "pricing": {
+ "prompt": "0.0000001",
+ "completion": "0.0000004",
+ "request": "0",
+ "image": "0",
+ "web_search": "0",
+ "internal_reasoning": "0"
+ },
+ "top_provider": {
+ "context_length": 1048576,
+ "max_completion_tokens": 65535,
+ "is_moderated": false
+ },
+ "per_request_limits": null,
+ "supported_parameters": [
+ "max_tokens",
+ "temperature",
+ "top_p",
+ "tools",
+ "tool_choice",
+ "stop",
+ "response_format",
+ "structured_outputs"
+ ]
+ },
+ {
+ "id": "google/gemini-2.5-flash",
+ "canonical_slug": "google/gemini-2.5-flash",
+ "hugging_face_id": "",
+ "name": "Google: Gemini 2.5 Flash",
+ "created": 1750172488,
+ "description": "Gemini 2.5 Flash is Google's state-of-the-art workhorse model, specifically designed for advanced reasoning, coding, mathematics, and scientific tasks. It includes built-in \"thinking\" capabilities, enabling it to provide responses with greater accuracy and nuanced context handling. \n\nAdditionally, Gemini 2.5 Flash is configurable through the \"max tokens for reasoning\" parameter, as described in the documentation (https://openrouter.ai/docs/use-cases/reasoning-tokens#max-tokens-for-reasoning).",
+ "context_length": 1048576,
+ "architecture": {
+ "modality": "text+image->text",
+ "input_modalities": [
+ "file",
+ "image",
+ "text"
+ ],
+ "output_modalities": [
+ "text"
+ ],
+ "tokenizer": "Gemini",
+ "instruct_type": null
+ },
+ "pricing": {
+ "prompt": "0.0000003",
+ "completion": "0.0000025",
+ "request": "0",
+ "image": "0.001238",
+ "web_search": "0",
+ "internal_reasoning": "0",
+ "input_cache_read": "0.000000075",
+ "input_cache_write": "0.0000003833"
+ },
+ "top_provider": {
+ "context_length": 1048576,
+ "max_completion_tokens": 65535,
+ "is_moderated": false
+ },
+ "per_request_limits": null,
+ "supported_parameters": [
+ "max_tokens",
+ "temperature",
+ "top_p",
+ "tools",
+ "tool_choice",
+ "stop",
+ "response_format",
+ "structured_outputs"
+ ]
+ },
+ {
+ "id": "google/gemini-2.5-pro",
+ "canonical_slug": "google/gemini-2.5-pro",
+ "hugging_face_id": "",
+ "name": "Google: Gemini 2.5 Pro",
+ "created": 1750169544,
+ "description": "Gemini 2.5 Pro is Google’s state-of-the-art AI model designed for advanced reasoning, coding, mathematics, and scientific tasks. It employs “thinking” capabilities, enabling it to reason through responses with enhanced accuracy and nuanced context handling. Gemini 2.5 Pro achieves top-tier performance on multiple benchmarks, including first-place positioning on the LMArena leaderboard, reflecting superior human-preference alignment and complex problem-solving abilities.",
+ "context_length": 1048576,
+ "architecture": {
+ "modality": "text+image->text",
+ "input_modalities": [
+ "file",
+ "image",
+ "text"
+ ],
+ "output_modalities": [
+ "text"
+ ],
+ "tokenizer": "Gemini",
+ "instruct_type": null
+ },
+ "pricing": {
+ "prompt": "0.00000125",
+ "completion": "0.00001",
+ "request": "0",
+ "image": "0.00516",
+ "web_search": "0",
+ "internal_reasoning": "0",
+ "input_cache_read": "0.00000031",
+ "input_cache_write": "0.000001625"
+ },
+ "top_provider": {
+ "context_length": 1048576,
+ "max_completion_tokens": 65536,
+ "is_moderated": false
+ },
+ "per_request_limits": null,
+ "supported_parameters": [
+ "max_tokens",
+ "temperature",
+ "top_p",
+ "tools",
+ "tool_choice",
+ "stop",
+ "response_format",
+ "structured_outputs"
+ ]
+ },
+ {
+ "id": "moonshotai/kimi-dev-72b:free",
+ "canonical_slug": "moonshotai/kimi-dev-72b",
+ "hugging_face_id": "moonshotai/Kimi-Dev-72B",
+ "name": "Kimi Dev 72b (free)",
+ "created": 1750115909,
+ "description": "Kimi-Dev-72B is an open-source large language model fine-tuned for software engineering and issue resolution tasks. Based on Qwen2.5-72B, it is optimized using large-scale reinforcement learning that applies code patches in real repositories and validates them via full test suite execution—rewarding only correct, robust completions. The model achieves 60.4% on SWE-bench Verified, setting a new benchmark among open-source models for software bug fixing and code reasoning.",
+ "context_length": 131072,
+ "architecture": {
+ "modality": "text->text",
+ "input_modalities": [
+ "text"
+ ],
+ "output_modalities": [
+ "text"
+ ],
+ "tokenizer": "Other",
+ "instruct_type": null
+ },
+ "pricing": {
+ "prompt": "0",
+ "completion": "0",
+ "request": "0",
+ "image": "0",
+ "web_search": "0",
+ "internal_reasoning": "0"
+ },
+ "top_provider": {
+ "context_length": 131072,
+ "max_completion_tokens": null,
+ "is_moderated": false
+ },
+ "per_request_limits": null,
+ "supported_parameters": [
+ "max_tokens",
+ "temperature",
+ "top_p",
+ "reasoning",
+ "include_reasoning",
+ "stop",
+ "frequency_penalty",
+ "presence_penalty",
+ "seed",
+ "top_k",
+ "min_p",
+ "repetition_penalty",
+ "logprobs",
+ "logit_bias",
+ "top_logprobs"
+ ]
+ },
+ {
+ "id": "openai/o3-pro",
+ "canonical_slug": "openai/o3-pro-2025-06-10",
+ "hugging_face_id": "",
+ "name": "OpenAI: o3 Pro",
+ "created": 1749598352,
+ "description": "The o-series of models are trained with reinforcement learning to think before they answer and perform complex reasoning. The o3-pro model uses more compute to think harder and provide consistently better answers.\n\nNote that BYOK is required for this model. Set up here: https://openrouter.ai/settings/integrations",
+ "context_length": 200000,
+ "architecture": {
+ "modality": "text+image->text",
+ "input_modalities": [
+ "text",
+ "file",
+ "image"
+ ],
+ "output_modalities": [
+ "text"
+ ],
+ "tokenizer": "Other",
+ "instruct_type": null
+ },
+ "pricing": {
+ "prompt": "0.00002",
+ "completion": "0.00008",
+ "request": "0",
+ "image": "0.0153",
+ "web_search": "0",
+ "internal_reasoning": "0"
+ },
+ "top_provider": {
+ "context_length": 200000,
+ "max_completion_tokens": 100000,
+ "is_moderated": true
+ },
+ "per_request_limits": null,
+ "supported_parameters": [
+ "tools",
+ "tool_choice",
+ "seed",
+ "max_tokens",
+ "response_format",
+ "structured_outputs"
+ ]
+ },
+ {
+ "id": "x-ai/grok-3-mini",
+ "canonical_slug": "x-ai/grok-3-mini",
+ "hugging_face_id": "",
+ "name": "xAI: Grok 3 Mini",
+ "created": 1749583245,
+ "description": "A lightweight model that thinks before responding. Fast, smart, and great for logic-based tasks that do not require deep domain knowledge. The raw thinking traces are accessible.",
+ "context_length": 131072,
+ "architecture": {
+ "modality": "text->text",
+ "input_modalities": [
+ "text"
+ ],
+ "output_modalities": [
+ "text"
+ ],
+ "tokenizer": "Grok",
+ "instruct_type": null
+ },
+ "pricing": {
+ "prompt": "0.0000003",
+ "completion": "0.0000005",
+ "request": "0",
+ "image": "0",
+ "web_search": "0",
+ "internal_reasoning": "0",
+ "input_cache_read": "0.000000075"
+ },
+ "top_provider": {
+ "context_length": 131072,
+ "max_completion_tokens": null,
+ "is_moderated": false
+ },
+ "per_request_limits": null,
+ "supported_parameters": [
+ "tools",
+ "tool_choice",
+ "max_tokens",
+ "temperature",
+ "top_p",
+ "reasoning",
+ "include_reasoning",
+ "structured_outputs",
+ "stop",
+ "seed",
+ "logprobs",
+ "top_logprobs",
+ "response_format"
+ ]
+ },
+ {
+ "id": "x-ai/grok-3",
+ "canonical_slug": "x-ai/grok-3",
+ "hugging_face_id": "",
+ "name": "xAI: Grok 3",
+ "created": 1749582908,
+ "description": "Grok 3 is the latest model from xAI. It's their flagship model that excels at enterprise use cases like data extraction, coding, and text summarization. Possesses deep domain knowledge in finance, healthcare, law, and science.\n\n",
+ "context_length": 131072,
+ "architecture": {
+ "modality": "text->text",
+ "input_modalities": [
+ "text"
+ ],
+ "output_modalities": [
+ "text"
+ ],
+ "tokenizer": "Grok",
+ "instruct_type": null
+ },
+ "pricing": {
+ "prompt": "0.000003",
+ "completion": "0.000015",
+ "request": "0",
+ "image": "0",
+ "web_search": "0",
+ "internal_reasoning": "0",
+ "input_cache_read": "0.00000075"
+ },
+ "top_provider": {
+ "context_length": 131072,
+ "max_completion_tokens": null,
+ "is_moderated": false
+ },
+ "per_request_limits": null,
+ "supported_parameters": [
+ "tools",
+ "tool_choice",
+ "max_tokens",
+ "temperature",
+ "top_p",
+ "structured_outputs",
+ "stop",
+ "frequency_penalty",
+ "presence_penalty",
+ "seed",
+ "logprobs",
+ "top_logprobs",
+ "response_format"
+ ]
+ },
+ {
+ "id": "mistralai/magistral-small-2506",
+ "canonical_slug": "mistralai/magistral-small-2506",
+ "hugging_face_id": "mistralai/Magistral-Small-2506",
+ "name": "Mistral: Magistral Small 2506",
+ "created": 1749569561,
+ "description": "Magistral Small is a 24B parameter instruction-tuned model based on Mistral-Small-3.1 (2503), enhanced through supervised fine-tuning on traces from Magistral Medium and further refined via reinforcement learning. It is optimized for reasoning and supports a wide multilingual range, including over 20 languages.",
+ "context_length": 40000,
+ "architecture": {
+ "modality": "text->text",
+ "input_modalities": [
+ "text"
+ ],
+ "output_modalities": [
+ "text"
+ ],
+ "tokenizer": "Mistral",
+ "instruct_type": null
+ },
+ "pricing": {
+ "prompt": "0.0000005",
+ "completion": "0.0000015",
+ "request": "0",
+ "image": "0",
+ "web_search": "0",
+ "internal_reasoning": "0"
+ },
+ "top_provider": {
+ "context_length": 40000,
+ "max_completion_tokens": 40000,
+ "is_moderated": false
+ },
+ "per_request_limits": null,
+ "supported_parameters": [
+ "tools",
+ "tool_choice",
+ "max_tokens",
+ "temperature",
+ "top_p",
+ "reasoning",
+ "include_reasoning",
+ "structured_outputs",
+ "response_format",
+ "stop",
+ "frequency_penalty",
+ "presence_penalty",
+ "seed"
+ ]
+ },
+ {
+ "id": "mistralai/magistral-medium-2506",
+ "canonical_slug": "mistralai/magistral-medium-2506",
+ "hugging_face_id": "",
+ "name": "Mistral: Magistral Medium 2506",
+ "created": 1749354054,
+ "description": "Magistral is Mistral's first reasoning model. It is ideal for general purpose use requiring longer thought processing and better accuracy than with non-reasoning LLMs. From legal research and financial forecasting to software development and creative storytelling — this model solves multi-step challenges where transparency and precision are critical.",
+ "context_length": 40960,
+ "architecture": {
+ "modality": "text->text",
+ "input_modalities": [
+ "text"
+ ],
+ "output_modalities": [
+ "text"
+ ],
+ "tokenizer": "Mistral",
+ "instruct_type": null
+ },
+ "pricing": {
+ "prompt": "0.000002",
+ "completion": "0.000005",
+ "request": "0",
+ "image": "0",
+ "web_search": "0",
+ "internal_reasoning": "0"
+ },
+ "top_provider": {
+ "context_length": 40960,
+ "max_completion_tokens": 40000,
+ "is_moderated": false
+ },
+ "per_request_limits": null,
+ "supported_parameters": [
+ "tools",
+ "tool_choice",
+ "max_tokens",
+ "temperature",
+ "top_p",
+ "reasoning",
+ "include_reasoning",
+ "structured_outputs",
+ "response_format",
+ "stop",
+ "frequency_penalty",
+ "presence_penalty",
+ "seed"
+ ]
+ },
+ {
+ "id": "mistralai/magistral-medium-2506:thinking",
+ "canonical_slug": "mistralai/magistral-medium-2506",
+ "hugging_face_id": "",
+ "name": "Mistral: Magistral Medium 2506 (thinking)",
+ "created": 1749354054,
+ "description": "Magistral is Mistral's first reasoning model. It is ideal for general purpose use requiring longer thought processing and better accuracy than with non-reasoning LLMs. From legal research and financial forecasting to software development and creative storytelling — this model solves multi-step challenges where transparency and precision are critical.",
+ "context_length": 40960,
+ "architecture": {
+ "modality": "text->text",
+ "input_modalities": [
+ "text"
+ ],
+ "output_modalities": [
+ "text"
+ ],
+ "tokenizer": "Mistral",
+ "instruct_type": null
+ },
+ "pricing": {
+ "prompt": "0.000002",
+ "completion": "0.000005",
+ "request": "0",
+ "image": "0",
+ "web_search": "0",
+ "internal_reasoning": "0"
+ },
+ "top_provider": {
+ "context_length": 40960,
+ "max_completion_tokens": 40000,
+ "is_moderated": false
+ },
+ "per_request_limits": null,
+ "supported_parameters": [
+ "tools",
+ "tool_choice",
+ "max_tokens",
+ "temperature",
+ "top_p",
+ "reasoning",
+ "include_reasoning",
+ "structured_outputs",
+ "response_format",
+ "stop",
+ "frequency_penalty",
+ "presence_penalty",
+ "seed"
+ ]
+ },
{
"id": "google/gemini-2.5-pro-preview",
+ "canonical_slug": "google/gemini-2.5-pro-preview-06-05",
"hugging_face_id": "",
"name": "Google: Gemini 2.5 Pro Preview 06-05",
"created": 1749137257,
@@ -38,23 +807,19 @@
},
"per_request_limits": null,
"supported_parameters": [
- "tools",
- "tool_choice",
"max_tokens",
"temperature",
"top_p",
- "reasoning",
- "include_reasoning",
- "structured_outputs",
- "response_format",
+ "tools",
+ "tool_choice",
"stop",
- "frequency_penalty",
- "presence_penalty",
- "seed"
+ "response_format",
+ "structured_outputs"
]
},
{
"id": "sentientagi/dobby-mini-unhinged-plus-llama-3.1-8b",
+ "canonical_slug": "sentientagi/dobby-mini-unhinged-plus-llama-3.1-8b",
"hugging_face_id": "SentientAGI/Dobby-Mini-Unhinged-Plus-Llama-3.1-8B",
"name": "SentientAGI: Dobby Mini Plus Llama 3.1 8B",
"created": 1748885619,
@@ -103,6 +868,7 @@
},
{
"id": "deepseek/deepseek-r1-distill-qwen-7b",
+ "canonical_slug": "deepseek/deepseek-r1-distill-qwen-7b",
"hugging_face_id": "deepseek-ai/DeepSeek-R1-Distill-Qwen-7B",
"name": "DeepSeek: R1 Distill Qwen 7B",
"created": 1748628237,
@@ -144,6 +910,7 @@
},
{
"id": "deepseek/deepseek-r1-0528-qwen3-8b:free",
+ "canonical_slug": "deepseek/deepseek-r1-0528-qwen3-8b",
"hugging_face_id": "deepseek-ai/deepseek-r1-0528-qwen3-8b",
"name": "DeepSeek: Deepseek R1 0528 Qwen3 8B (free)",
"created": 1748538543,
@@ -194,11 +961,12 @@
},
{
"id": "deepseek/deepseek-r1-0528-qwen3-8b",
+ "canonical_slug": "deepseek/deepseek-r1-0528-qwen3-8b",
"hugging_face_id": "deepseek-ai/deepseek-r1-0528-qwen3-8b",
"name": "DeepSeek: Deepseek R1 0528 Qwen3 8B",
"created": 1748538543,
"description": "DeepSeek-R1-0528 is a lightly upgraded release of DeepSeek R1 that taps more compute and smarter post-training tricks, pushing its reasoning and inference to the brink of flagship models like O3 and Gemini 2.5 Pro.\nIt now tops math, programming, and logic leaderboards, showcasing a step-change in depth-of-thought.\nThe distilled variant, DeepSeek-R1-0528-Qwen3-8B, transfers this chain-of-thought into an 8 B-parameter form, beating standard Qwen3 8B by +10 pp and tying the 235 B “thinking” giant on AIME 2024.",
- "context_length": 128000,
+ "context_length": 32000,
"architecture": {
"modality": "text->text",
"input_modalities": [
@@ -211,16 +979,16 @@
"instruct_type": "deepseek-r1"
},
"pricing": {
- "prompt": "0.00000006",
- "completion": "0.00000009",
+ "prompt": "0.00000001",
+ "completion": "0.00000002",
"request": "0",
"image": "0",
"web_search": "0",
"internal_reasoning": "0"
},
"top_provider": {
- "context_length": 128000,
- "max_completion_tokens": 32000,
+ "context_length": 32000,
+ "max_completion_tokens": null,
"is_moderated": false
},
"per_request_limits": null,
@@ -230,64 +998,19 @@
"top_p",
"reasoning",
"include_reasoning",
- "stop",
- "frequency_penalty",
"presence_penalty",
- "seed",
- "top_k",
- "min_p",
+ "frequency_penalty",
"repetition_penalty",
+ "top_k",
+ "stop",
+ "seed",
+ "min_p",
"logit_bias"
]
},
- {
- "id": "google/gemma-2b-it",
- "hugging_face_id": "google/gemma-2b-it",
- "name": "Google: Gemma 1 2B",
- "created": 1748460815,
- "description": "Gemma 1 2B by Google is an open model built from the same research and technology used to create the [Gemini models](/models?q=gemini).\n\nGemma models are well-suited for a variety of text generation tasks, including question answering, summarization, and reasoning.\n\nUsage of Gemma is subject to Google's [Gemma Terms of Use](https://ai.google.dev/gemma/terms).",
- "context_length": 8192,
- "architecture": {
- "modality": "text->text",
- "input_modalities": [
- "text"
- ],
- "output_modalities": [
- "text"
- ],
- "tokenizer": "Gemini",
- "instruct_type": "gemma"
- },
- "pricing": {
- "prompt": "0.0000001",
- "completion": "0.0000001",
- "request": "0",
- "image": "0",
- "web_search": "0",
- "internal_reasoning": "0"
- },
- "top_provider": {
- "context_length": 8192,
- "max_completion_tokens": null,
- "is_moderated": false
- },
- "per_request_limits": null,
- "supported_parameters": [
- "max_tokens",
- "temperature",
- "top_p",
- "stop",
- "frequency_penalty",
- "presence_penalty",
- "top_k",
- "repetition_penalty",
- "logit_bias",
- "min_p",
- "response_format"
- ]
- },
{
"id": "deepseek/deepseek-r1-0528:free",
+ "canonical_slug": "deepseek/deepseek-r1-0528",
"hugging_face_id": "deepseek-ai/DeepSeek-R1-0528",
"name": "DeepSeek: R1 0528 (free)",
"created": 1748455170,
@@ -338,6 +1061,7 @@
},
{
"id": "deepseek/deepseek-r1-0528",
+ "canonical_slug": "deepseek/deepseek-r1-0528",
"hugging_face_id": "deepseek-ai/DeepSeek-R1-0528",
"name": "DeepSeek: R1 0528",
"created": 1748455170,
@@ -384,12 +1108,15 @@
"response_format",
"logprobs",
"top_logprobs",
+ "tools",
+ "tool_choice",
"seed",
"structured_outputs"
]
},
{
"id": "sarvamai/sarvam-m:free",
+ "canonical_slug": "sarvamai/sarvam-m",
"hugging_face_id": "sarvamai/sarvam-m",
"name": "Sarvam AI: Sarvam-M (free)",
"created": 1748188413,
@@ -436,50 +1163,9 @@
"top_logprobs"
]
},
- {
- "id": "sarvamai/sarvam-m",
- "hugging_face_id": "sarvamai/sarvam-m",
- "name": "Sarvam AI: Sarvam-M",
- "created": 1748188413,
- "description": "Sarvam-M is a 24 B-parameter, instruction-tuned derivative of Mistral-Small-3.1-24B-Base-2503, post-trained on English plus eleven major Indic languages (bn, hi, kn, gu, mr, ml, or, pa, ta, te). The model introduces a dual-mode interface: “non-think” for low-latency chat and a optional “think” phase that exposes chain-of-thought tokens for more demanding reasoning, math, and coding tasks. \n\nBenchmark reports show solid gains versus similarly sized open models on Indic-language QA, GSM-8K math, and SWE-Bench coding, making Sarvam-M a practical general-purpose choice for multilingual conversational agents as well as analytical workloads that mix English, native Indic scripts, or romanized text.",
- "context_length": 32768,
- "architecture": {
- "modality": "text->text",
- "input_modalities": [
- "text"
- ],
- "output_modalities": [
- "text"
- ],
- "tokenizer": "Other",
- "instruct_type": null
- },
- "pricing": {
- "prompt": "0.00000025",
- "completion": "0.00000075",
- "request": "0",
- "image": "0",
- "web_search": "0",
- "internal_reasoning": "0"
- },
- "top_provider": {
- "context_length": 32768,
- "max_completion_tokens": 32768,
- "is_moderated": false
- },
- "per_request_limits": null,
- "supported_parameters": [
- "max_tokens",
- "temperature",
- "top_p",
- "presence_penalty",
- "frequency_penalty",
- "repetition_penalty",
- "top_k"
- ]
- },
{
"id": "thedrummer/valkyrie-49b-v1",
+ "canonical_slug": "thedrummer/valkyrie-49b-v1",
"hugging_face_id": "TheDrummer/Valkyrie-49B-v1",
"name": "TheDrummer: Valkyrie 49B V1",
"created": 1748022670,
@@ -524,6 +1210,7 @@
},
{
"id": "anthropic/claude-opus-4",
+ "canonical_slug": "anthropic/claude-4-opus-20250522",
"hugging_face_id": "",
"name": "Anthropic: Claude Opus 4",
"created": 1747931245,
@@ -571,6 +1258,7 @@
},
{
"id": "anthropic/claude-sonnet-4",
+ "canonical_slug": "anthropic/claude-4-sonnet-20250522",
"hugging_face_id": "",
"name": "Anthropic: Claude Sonnet 4",
"created": 1747930371,
@@ -618,6 +1306,7 @@
},
{
"id": "mistralai/devstral-small:free",
+ "canonical_slug": "mistralai/devstral-small-2505",
"hugging_face_id": "mistralai/Devstral-Small-2505",
"name": "Mistral: Devstral Small (free)",
"created": 1747837379,
@@ -668,6 +1357,7 @@
},
{
"id": "mistralai/devstral-small",
+ "canonical_slug": "mistralai/devstral-small-2505",
"hugging_face_id": "mistralai/Devstral-Small-2505",
"name": "Mistral: Devstral Small",
"created": 1747837379,
@@ -712,12 +1402,16 @@
"min_p",
"tools",
"tool_choice",
+ "logit_bias",
+ "logprobs",
+ "top_logprobs",
"structured_outputs"
]
},
{
"id": "google/gemma-3n-e4b-it:free",
- "hugging_face_id": "",
+ "canonical_slug": "google/gemma-3n-e4b-it",
+ "hugging_face_id": "google/gemma-3n-E4B-it",
"name": "Google: Gemma 3n 4B (free)",
"created": 1747776824,
"description": "Gemma 3n E4B-it is optimized for efficient execution on mobile and low-resource devices, such as phones, laptops, and tablets. It supports multimodal inputs—including text, visual data, and audio—enabling diverse tasks such as text generation, speech recognition, translation, and image analysis. Leveraging innovations like Per-Layer Embedding (PLE) caching and the MatFormer architecture, Gemma 3n dynamically manages memory usage and computational load by selectively activating model parameters, significantly reducing runtime resource requirements.\n\nThis model supports a wide linguistic range (trained in over 140 languages) and features a flexible 32K token context window. Gemma 3n can selectively load parameters, optimizing memory and computational efficiency based on the task or device capabilities, making it well-suited for privacy-focused, offline-capable applications and on-device AI solutions. [Read more in the blog post](https://developers.googleblog.com/en/introducing-gemma-3n/)",
@@ -758,8 +1452,56 @@
"response_format"
]
},
+ {
+ "id": "google/gemma-3n-e4b-it",
+ "canonical_slug": "google/gemma-3n-e4b-it",
+ "hugging_face_id": "google/gemma-3n-E4B-it",
+ "name": "Google: Gemma 3n 4B",
+ "created": 1747776824,
+ "description": "Gemma 3n E4B-it is optimized for efficient execution on mobile and low-resource devices, such as phones, laptops, and tablets. It supports multimodal inputs—including text, visual data, and audio—enabling diverse tasks such as text generation, speech recognition, translation, and image analysis. Leveraging innovations like Per-Layer Embedding (PLE) caching and the MatFormer architecture, Gemma 3n dynamically manages memory usage and computational load by selectively activating model parameters, significantly reducing runtime resource requirements.\n\nThis model supports a wide linguistic range (trained in over 140 languages) and features a flexible 32K token context window. Gemma 3n can selectively load parameters, optimizing memory and computational efficiency based on the task or device capabilities, making it well-suited for privacy-focused, offline-capable applications and on-device AI solutions. [Read more in the blog post](https://developers.googleblog.com/en/introducing-gemma-3n/)",
+ "context_length": 32768,
+ "architecture": {
+ "modality": "text->text",
+ "input_modalities": [
+ "text"
+ ],
+ "output_modalities": [
+ "text"
+ ],
+ "tokenizer": "Other",
+ "instruct_type": null
+ },
+ "pricing": {
+ "prompt": "0.00000002",
+ "completion": "0.00000004",
+ "request": "0",
+ "image": "0",
+ "web_search": "0",
+ "internal_reasoning": "0"
+ },
+ "top_provider": {
+ "context_length": 32768,
+ "max_completion_tokens": null,
+ "is_moderated": false
+ },
+ "per_request_limits": null,
+ "supported_parameters": [
+ "max_tokens",
+ "temperature",
+ "top_p",
+ "stop",
+ "frequency_penalty",
+ "presence_penalty",
+ "top_k",
+ "repetition_penalty",
+ "logit_bias",
+ "min_p",
+ "response_format"
+ ]
+ },
{
"id": "google/gemini-2.5-flash-preview-05-20",
+ "canonical_slug": "google/gemini-2.5-flash-preview-05-20",
"hugging_face_id": "",
"name": "Google: Gemini 2.5 Flash Preview 05-20",
"created": 1747761924,
@@ -795,23 +1537,19 @@
},
"per_request_limits": null,
"supported_parameters": [
- "tools",
- "tool_choice",
"max_tokens",
"temperature",
"top_p",
- "reasoning",
- "include_reasoning",
- "structured_outputs",
- "response_format",
+ "tools",
+ "tool_choice",
"stop",
- "frequency_penalty",
- "presence_penalty",
- "seed"
+ "response_format",
+ "structured_outputs"
]
},
{
"id": "google/gemini-2.5-flash-preview-05-20:thinking",
+ "canonical_slug": "google/gemini-2.5-flash-preview-05-20",
"hugging_face_id": "",
"name": "Google: Gemini 2.5 Flash Preview 05-20 (thinking)",
"created": 1747761924,
@@ -847,23 +1585,19 @@
},
"per_request_limits": null,
"supported_parameters": [
- "tools",
- "tool_choice",
"max_tokens",
"temperature",
"top_p",
- "reasoning",
- "include_reasoning",
- "structured_outputs",
- "response_format",
+ "tools",
+ "tool_choice",
"stop",
- "frequency_penalty",
- "presence_penalty",
- "seed"
+ "response_format",
+ "structured_outputs"
]
},
{
"id": "openai/codex-mini",
+ "canonical_slug": "openai/codex-mini",
"hugging_face_id": "",
"name": "OpenAI: Codex Mini",
"created": 1747409761,
@@ -905,107 +1639,14 @@
"structured_outputs"
]
},
- {
- "id": "meta-llama/llama-3.3-8b-instruct:free",
- "hugging_face_id": "",
- "name": "Meta: Llama 3.3 8B Instruct (free)",
- "created": 1747230154,
- "description": "A lightweight and ultra-fast variant of Llama 3.3 70B, for use when quick response times are needed most.",
- "context_length": 128000,
- "architecture": {
- "modality": "text->text",
- "input_modalities": [
- "text"
- ],
- "output_modalities": [
- "text"
- ],
- "tokenizer": "Llama3",
- "instruct_type": null
- },
- "pricing": {
- "prompt": "0",
- "completion": "0",
- "request": "0",
- "image": "0",
- "web_search": "0",
- "internal_reasoning": "0"
- },
- "top_provider": {
- "context_length": 128000,
- "max_completion_tokens": 4028,
- "is_moderated": false
- },
- "per_request_limits": null,
- "supported_parameters": [
- "tools",
- "tool_choice",
- "max_tokens",
- "temperature",
- "top_p",
- "structured_outputs",
- "response_format",
- "repetition_penalty",
- "top_k"
- ]
- },
- {
- "id": "nousresearch/deephermes-3-mistral-24b-preview:free",
- "hugging_face_id": "NousResearch/DeepHermes-3-Mistral-24B-Preview",
- "name": "Nous: DeepHermes 3 Mistral 24B Preview (free)",
- "created": 1746830904,
- "description": "DeepHermes 3 (Mistral 24B Preview) is an instruction-tuned language model by Nous Research based on Mistral-Small-24B, designed for chat, function calling, and advanced multi-turn reasoning. It introduces a dual-mode system that toggles between intuitive chat responses and structured “deep reasoning” mode using special system prompts. Fine-tuned via distillation from R1, it supports structured output (JSON mode) and function call syntax for agent-based applications.\n\nDeepHermes 3 supports a **reasoning toggle via system prompt**, allowing users to switch between fast, intuitive responses and deliberate, multi-step reasoning. When activated with the following specific system instruction, the model enters a *\"deep thinking\"* mode—generating extended chains of thought wrapped in `` tags before delivering a final answer. \n\nSystem Prompt: You are a deep thinking AI, you may use extremely long chains of thought to deeply consider the problem and deliberate with yourself via systematic reasoning processes to help come to a correct solution prior to answering. You should enclose your thoughts and internal monologue inside tags, and then provide your solution or response to the problem.\n",
- "context_length": 32768,
- "architecture": {
- "modality": "text->text",
- "input_modalities": [
- "text"
- ],
- "output_modalities": [
- "text"
- ],
- "tokenizer": "Other",
- "instruct_type": null
- },
- "pricing": {
- "prompt": "0",
- "completion": "0",
- "request": "0",
- "image": "0",
- "web_search": "0",
- "internal_reasoning": "0"
- },
- "top_provider": {
- "context_length": 32768,
- "max_completion_tokens": null,
- "is_moderated": false
- },
- "per_request_limits": null,
- "supported_parameters": [
- "max_tokens",
- "temperature",
- "top_p",
- "reasoning",
- "include_reasoning",
- "stop",
- "frequency_penalty",
- "presence_penalty",
- "seed",
- "top_k",
- "min_p",
- "repetition_penalty",
- "logprobs",
- "logit_bias",
- "top_logprobs"
- ]
- },
{
"id": "mistralai/mistral-medium-3",
+ "canonical_slug": "mistralai/mistral-medium-3",
"hugging_face_id": "",
"name": "Mistral: Mistral Medium 3",
"created": 1746627341,
"description": "Mistral Medium 3 is a high-performance enterprise-grade language model designed to deliver frontier-level capabilities at significantly reduced operational cost. It balances state-of-the-art reasoning and multimodal performance with 8× lower cost compared to traditional large models, making it suitable for scalable deployments across professional and industrial use cases.\n\nThe model excels in domains such as coding, STEM reasoning, and enterprise adaptation. It supports hybrid, on-prem, and in-VPC deployments and is optimized for integration into custom workflows. Mistral Medium 3 offers competitive accuracy relative to larger models like Claude Sonnet 3.5/3.7, Llama 4 Maverick, and Command R+, while maintaining broad compatibility across cloud environments.",
- "context_length": 131072,
+ "context_length": 32768,
"architecture": {
"modality": "text+image->text",
"input_modalities": [
@@ -1027,7 +1668,7 @@
"internal_reasoning": "0"
},
"top_provider": {
- "context_length": 131072,
+ "context_length": 32768,
"max_completion_tokens": null,
"is_moderated": false
},
@@ -1048,6 +1689,7 @@
},
{
"id": "google/gemini-2.5-pro-preview-05-06",
+ "canonical_slug": "google/gemini-2.5-pro-preview-03-25",
"hugging_face_id": "",
"name": "Google: Gemini 2.5 Pro Preview 05-06",
"created": 1746578513,
@@ -1096,6 +1738,7 @@
},
{
"id": "arcee-ai/caller-large",
+ "canonical_slug": "arcee-ai/caller-large",
"hugging_face_id": "",
"name": "Arcee AI: Caller Large",
"created": 1746487869,
@@ -1144,6 +1787,7 @@
},
{
"id": "arcee-ai/spotlight",
+ "canonical_slug": "arcee-ai/spotlight",
"hugging_face_id": "",
"name": "Arcee AI: Spotlight",
"created": 1746481552,
@@ -1191,6 +1835,7 @@
},
{
"id": "arcee-ai/maestro-reasoning",
+ "canonical_slug": "arcee-ai/maestro-reasoning",
"hugging_face_id": "",
"name": "Arcee AI: Maestro Reasoning",
"created": 1746481269,
@@ -1237,6 +1882,7 @@
},
{
"id": "arcee-ai/virtuoso-large",
+ "canonical_slug": "arcee-ai/virtuoso-large",
"hugging_face_id": "",
"name": "Arcee AI: Virtuoso Large",
"created": 1746478885,
@@ -1268,6 +1914,8 @@
},
"per_request_limits": null,
"supported_parameters": [
+ "tools",
+ "tool_choice",
"max_tokens",
"temperature",
"top_p",
@@ -1283,6 +1931,7 @@
},
{
"id": "arcee-ai/coder-large",
+ "canonical_slug": "arcee-ai/coder-large",
"hugging_face_id": "",
"name": "Arcee AI: Coder Large",
"created": 1746478663,
@@ -1329,6 +1978,7 @@
},
{
"id": "arcee-ai/virtuoso-medium-v2",
+ "canonical_slug": "arcee-ai/virtuoso-medium-v2",
"hugging_face_id": "arcee-ai/Virtuoso-Medium-v2",
"name": "Arcee AI: Virtuoso Medium V2",
"created": 1746478434,
@@ -1360,6 +2010,8 @@
},
"per_request_limits": null,
"supported_parameters": [
+ "tools",
+ "tool_choice",
"max_tokens",
"temperature",
"top_p",
@@ -1375,6 +2027,7 @@
},
{
"id": "arcee-ai/arcee-blitz",
+ "canonical_slug": "arcee-ai/arcee-blitz",
"hugging_face_id": "arcee-ai/arcee-blitz",
"name": "Arcee AI: Arcee Blitz",
"created": 1746470100,
@@ -1419,58 +2072,9 @@
"response_format"
]
},
- {
- "id": "microsoft/phi-4-reasoning-plus:free",
- "hugging_face_id": "microsoft/Phi-4-reasoning-plus",
- "name": "Microsoft: Phi 4 Reasoning Plus (free)",
- "created": 1746130961,
- "description": "Phi-4-reasoning-plus is an enhanced 14B parameter model from Microsoft, fine-tuned from Phi-4 with additional reinforcement learning to boost accuracy on math, science, and code reasoning tasks. It uses the same dense decoder-only transformer architecture as Phi-4, but generates longer, more comprehensive outputs structured into a step-by-step reasoning trace and final answer.\n\nWhile it offers improved benchmark scores over Phi-4-reasoning across tasks like AIME, OmniMath, and HumanEvalPlus, its responses are typically ~50% longer, resulting in higher latency. Designed for English-only applications, it is well-suited for structured reasoning workflows where output quality takes priority over response speed.",
- "context_length": 32768,
- "architecture": {
- "modality": "text->text",
- "input_modalities": [
- "text"
- ],
- "output_modalities": [
- "text"
- ],
- "tokenizer": "Other",
- "instruct_type": null
- },
- "pricing": {
- "prompt": "0",
- "completion": "0",
- "request": "0",
- "image": "0",
- "web_search": "0",
- "internal_reasoning": "0"
- },
- "top_provider": {
- "context_length": 32768,
- "max_completion_tokens": null,
- "is_moderated": false
- },
- "per_request_limits": null,
- "supported_parameters": [
- "max_tokens",
- "temperature",
- "top_p",
- "reasoning",
- "include_reasoning",
- "stop",
- "frequency_penalty",
- "presence_penalty",
- "seed",
- "top_k",
- "min_p",
- "repetition_penalty",
- "logprobs",
- "logit_bias",
- "top_logprobs"
- ]
- },
{
"id": "microsoft/phi-4-reasoning-plus",
+ "canonical_slug": "microsoft/phi-4-reasoning-plus-04-30",
"hugging_face_id": "microsoft/Phi-4-reasoning-plus",
"name": "Microsoft: Phi 4 Reasoning Plus",
"created": 1746130961,
@@ -1517,58 +2121,9 @@
"min_p"
]
},
- {
- "id": "microsoft/phi-4-reasoning:free",
- "hugging_face_id": "microsoft/Phi-4-reasoning",
- "name": "Microsoft: Phi 4 Reasoning (free)",
- "created": 1746121275,
- "description": "Phi-4-reasoning is a 14B parameter dense decoder-only transformer developed by Microsoft, fine-tuned from Phi-4 to enhance complex reasoning capabilities. It uses a combination of supervised fine-tuning on chain-of-thought traces and reinforcement learning, targeting math, science, and code reasoning tasks. With a 32k context window and high inference efficiency, it is optimized for structured responses in a two-part format: reasoning trace followed by a final solution.\n\nThe model achieves strong results on specialized benchmarks such as AIME, OmniMath, and LiveCodeBench, outperforming many larger models in structured reasoning tasks. It is released under the MIT license and intended for use in latency-constrained, English-only environments requiring reliable step-by-step logic. Recommended usage includes ChatML prompts and structured reasoning format for best results.",
- "context_length": 32768,
- "architecture": {
- "modality": "text->text",
- "input_modalities": [
- "text"
- ],
- "output_modalities": [
- "text"
- ],
- "tokenizer": "Other",
- "instruct_type": null
- },
- "pricing": {
- "prompt": "0",
- "completion": "0",
- "request": "0",
- "image": "0",
- "web_search": "0",
- "internal_reasoning": "0"
- },
- "top_provider": {
- "context_length": 32768,
- "max_completion_tokens": null,
- "is_moderated": false
- },
- "per_request_limits": null,
- "supported_parameters": [
- "max_tokens",
- "temperature",
- "top_p",
- "reasoning",
- "include_reasoning",
- "stop",
- "frequency_penalty",
- "presence_penalty",
- "seed",
- "top_k",
- "min_p",
- "repetition_penalty",
- "logprobs",
- "logit_bias",
- "top_logprobs"
- ]
- },
{
"id": "inception/mercury-coder-small-beta",
+ "canonical_slug": "inception/mercury-coder-small-beta",
"hugging_face_id": "",
"name": "Inception: Mercury Coder Small Beta",
"created": 1746033880,
@@ -1607,12 +2162,13 @@
]
},
{
- "id": "opengvlab/internvl3-14b:free",
+ "id": "opengvlab/internvl3-14b",
+ "canonical_slug": "opengvlab/internvl3-14b",
"hugging_face_id": "OpenGVLab/InternVL3-14B",
- "name": "OpenGVLab: InternVL3 14B (free)",
+ "name": "OpenGVLab: InternVL3 14B",
"created": 1746021355,
"description": "The 14b version of the InternVL3 series. An advanced multimodal large language model (MLLM) series that demonstrates superior overall performance. Compared to InternVL 2.5, InternVL3 exhibits superior multimodal perception and reasoning capabilities, while further extending its multimodal capabilities to encompass tool usage, GUI agents, industrial image analysis, 3D vision perception, and more.",
- "context_length": 32000,
+ "context_length": 12288,
"architecture": {
"modality": "text+image->text",
"input_modalities": [
@@ -1626,15 +2182,15 @@
"instruct_type": null
},
"pricing": {
- "prompt": "0",
- "completion": "0",
+ "prompt": "0.0000002",
+ "completion": "0.0000004",
"request": "0",
"image": "0",
"web_search": "0",
"internal_reasoning": "0"
},
"top_provider": {
- "context_length": 32000,
+ "context_length": 12288,
"max_completion_tokens": null,
"is_moderated": false
},
@@ -1646,12 +2202,13 @@
]
},
{
- "id": "opengvlab/internvl3-2b:free",
+ "id": "opengvlab/internvl3-2b",
+ "canonical_slug": "opengvlab/internvl3-2b",
"hugging_face_id": "OpenGVLab/InternVL3-2B",
- "name": "OpenGVLab: InternVL3 2B (free)",
+ "name": "OpenGVLab: InternVL3 2B",
"created": 1746019807,
"description": "The 2b version of the InternVL3 series, for an even higher inference speed and very reasonable performance. An advanced multimodal large language model (MLLM) series that demonstrates superior overall performance. Compared to InternVL 2.5, InternVL3 exhibits superior multimodal perception and reasoning capabilities, while further extending its multimodal capabilities to encompass tool usage, GUI agents, industrial image analysis, 3D vision perception, and more.",
- "context_length": 32000,
+ "context_length": 12288,
"architecture": {
"modality": "text+image->text",
"input_modalities": [
@@ -1665,15 +2222,15 @@
"instruct_type": null
},
"pricing": {
- "prompt": "0",
- "completion": "0",
+ "prompt": "0.00000005",
+ "completion": "0.0000001",
"request": "0",
"image": "0",
"web_search": "0",
"internal_reasoning": "0"
},
"top_provider": {
- "context_length": 32000,
+ "context_length": 12288,
"max_completion_tokens": null,
"is_moderated": false
},
@@ -1684,56 +2241,9 @@
"top_p"
]
},
- {
- "id": "deepseek/deepseek-prover-v2:free",
- "hugging_face_id": "deepseek-ai/DeepSeek-Prover-V2-671B",
- "name": "DeepSeek: DeepSeek Prover V2 (free)",
- "created": 1746013094,
- "description": "DeepSeek Prover V2 is a 671B parameter model, speculated to be geared towards logic and mathematics. Likely an upgrade from [DeepSeek-Prover-V1.5](https://huggingface.co/deepseek-ai/DeepSeek-Prover-V1.5-RL) Not much is known about the model yet, as DeepSeek released it on Hugging Face without an announcement or description.",
- "context_length": 163840,
- "architecture": {
- "modality": "text->text",
- "input_modalities": [
- "text"
- ],
- "output_modalities": [
- "text"
- ],
- "tokenizer": "DeepSeek",
- "instruct_type": null
- },
- "pricing": {
- "prompt": "0",
- "completion": "0",
- "request": "0",
- "image": "0",
- "web_search": "0",
- "internal_reasoning": "0"
- },
- "top_provider": {
- "context_length": 163840,
- "max_completion_tokens": null,
- "is_moderated": false
- },
- "per_request_limits": null,
- "supported_parameters": [
- "max_tokens",
- "temperature",
- "top_p",
- "stop",
- "frequency_penalty",
- "presence_penalty",
- "seed",
- "top_k",
- "min_p",
- "repetition_penalty",
- "logprobs",
- "logit_bias",
- "top_logprobs"
- ]
- },
{
"id": "deepseek/deepseek-prover-v2",
+ "canonical_slug": "deepseek/deepseek-prover-v2",
"hugging_face_id": "deepseek-ai/DeepSeek-Prover-V2-671B",
"name": "DeepSeek: DeepSeek Prover V2",
"created": 1746013094,
@@ -1781,6 +2291,7 @@
},
{
"id": "meta-llama/llama-guard-4-12b",
+ "canonical_slug": "meta-llama/llama-guard-4-12b",
"hugging_face_id": "meta-llama/Llama-Guard-4-12B",
"name": "Meta: Llama Guard 4 12B",
"created": 1745975193,
@@ -1831,6 +2342,7 @@
},
{
"id": "qwen/qwen3-30b-a3b:free",
+ "canonical_slug": "qwen/qwen3-30b-a3b-04-28",
"hugging_face_id": "Qwen/Qwen3-30B-A3B",
"name": "Qwen: Qwen3 30B A3B (free)",
"created": 1745878604,
@@ -1881,6 +2393,7 @@
},
{
"id": "qwen/qwen3-30b-a3b",
+ "canonical_slug": "qwen/qwen3-30b-a3b-04-28",
"hugging_face_id": "Qwen/Qwen3-30B-A3B",
"name": "Qwen: Qwen3 30B A3B",
"created": 1745878604,
@@ -1919,22 +2432,23 @@
"top_p",
"reasoning",
"include_reasoning",
+ "response_format",
"stop",
"frequency_penalty",
"presence_penalty",
- "top_k",
"repetition_penalty",
- "response_format",
+ "top_k",
+ "seed",
+ "min_p",
"structured_outputs",
"logit_bias",
"logprobs",
- "top_logprobs",
- "seed",
- "min_p"
+ "top_logprobs"
]
},
{
"id": "qwen/qwen3-8b:free",
+ "canonical_slug": "qwen/qwen3-8b-04-28",
"hugging_face_id": "Qwen/Qwen3-8B",
"name": "Qwen: Qwen3 8B (free)",
"created": 1745876632,
@@ -1985,6 +2499,7 @@
},
{
"id": "qwen/qwen3-8b",
+ "canonical_slug": "qwen/qwen3-8b-04-28",
"hugging_face_id": "Qwen/Qwen3-8B",
"name": "Qwen: Qwen3 8B",
"created": 1745876632,
@@ -2033,6 +2548,7 @@
},
{
"id": "qwen/qwen3-14b:free",
+ "canonical_slug": "qwen/qwen3-14b-04-28",
"hugging_face_id": "Qwen/Qwen3-14B",
"name": "Qwen: Qwen3 14B (free)",
"created": 1745876478,
@@ -2083,6 +2599,7 @@
},
{
"id": "qwen/qwen3-14b",
+ "canonical_slug": "qwen/qwen3-14b-04-28",
"hugging_face_id": "Qwen/Qwen3-14B",
"name": "Qwen: Qwen3 14B",
"created": 1745876478,
@@ -2100,7 +2617,7 @@
"instruct_type": "qwen3"
},
"pricing": {
- "prompt": "0.00000007",
+ "prompt": "0.00000006",
"completion": "0.00000024",
"request": "0",
"image": "0",
@@ -2114,19 +2631,19 @@
},
"per_request_limits": null,
"supported_parameters": [
- "tools",
- "tool_choice",
"max_tokens",
"temperature",
"top_p",
"reasoning",
"include_reasoning",
- "stop",
- "frequency_penalty",
"presence_penalty",
+ "frequency_penalty",
"repetition_penalty",
- "response_format",
"top_k",
+ "tools",
+ "tool_choice",
+ "stop",
+ "response_format",
"seed",
"min_p",
"logit_bias",
@@ -2136,6 +2653,7 @@
},
{
"id": "qwen/qwen3-32b:free",
+ "canonical_slug": "qwen/qwen3-32b-04-28",
"hugging_face_id": "Qwen/Qwen3-32B",
"name": "Qwen: Qwen3 32B (free)",
"created": 1745875945,
@@ -2186,6 +2704,7 @@
},
{
"id": "qwen/qwen3-32b",
+ "canonical_slug": "qwen/qwen3-32b-04-28",
"hugging_face_id": "Qwen/Qwen3-32B",
"name": "Qwen: Qwen3 32B",
"created": 1745875945,
@@ -2222,24 +2741,25 @@
"top_p",
"reasoning",
"include_reasoning",
- "seed",
- "tools",
- "tool_choice",
"stop",
"frequency_penalty",
"presence_penalty",
- "repetition_penalty",
"response_format",
+ "top_logprobs",
+ "logprobs",
+ "logit_bias",
+ "seed",
+ "tools",
+ "tool_choice",
+ "repetition_penalty",
"top_k",
"min_p",
- "structured_outputs",
- "logprobs",
- "top_logprobs",
- "logit_bias"
+ "structured_outputs"
]
},
{
"id": "qwen/qwen3-235b-a22b:free",
+ "canonical_slug": "qwen/qwen3-235b-a22b-04-28",
"hugging_face_id": "Qwen/Qwen3-235B-A22B",
"name": "Qwen: Qwen3 235B A22B (free)",
"created": 1745875757,
@@ -2290,6 +2810,7 @@
},
{
"id": "qwen/qwen3-235b-a22b",
+ "canonical_slug": "qwen/qwen3-235b-a22b-04-28",
"hugging_face_id": "Qwen/Qwen3-235B-A22B",
"name": "Qwen: Qwen3 235B A22B",
"created": 1745875757,
@@ -2344,6 +2865,7 @@
},
{
"id": "tngtech/deepseek-r1t-chimera:free",
+ "canonical_slug": "tngtech/deepseek-r1t-chimera",
"hugging_face_id": "tngtech/DeepSeek-R1T-Chimera",
"name": "TNG: DeepSeek R1T Chimera (free)",
"created": 1745760875,
@@ -2394,6 +2916,7 @@
},
{
"id": "thudm/glm-z1-rumination-32b",
+ "canonical_slug": "thudm/glm-z1-rumination-32b-0414",
"hugging_face_id": "THUDM/GLM-Z1-Rumination-32B-0414",
"name": "THUDM: GLM Z1 Rumination 32B ",
"created": 1745601495,
@@ -2442,6 +2965,7 @@
},
{
"id": "microsoft/mai-ds-r1:free",
+ "canonical_slug": "microsoft/mai-ds-r1",
"hugging_face_id": "microsoft/MAI-DS-R1",
"name": "Microsoft: MAI DS R1 (free)",
"created": 1745194100,
@@ -2492,6 +3016,7 @@
},
{
"id": "thudm/glm-z1-32b:free",
+ "canonical_slug": "thudm/glm-z1-32b-0414",
"hugging_face_id": "THUDM/GLM-Z1-32B-0414",
"name": "THUDM: GLM Z1 32B (free)",
"created": 1744924148,
@@ -2542,6 +3067,7 @@
},
{
"id": "thudm/glm-z1-32b",
+ "canonical_slug": "thudm/glm-z1-32b-0414",
"hugging_face_id": "THUDM/GLM-Z1-32B-0414",
"name": "THUDM: GLM Z1 32B",
"created": 1744924148,
@@ -2590,6 +3116,7 @@
},
{
"id": "thudm/glm-4-32b:free",
+ "canonical_slug": "thudm/glm-4-32b-0414",
"hugging_face_id": "THUDM/GLM-4-32B-0414",
"name": "THUDM: GLM 4 32B (free)",
"created": 1744920915,
@@ -2638,6 +3165,7 @@
},
{
"id": "thudm/glm-4-32b",
+ "canonical_slug": "thudm/glm-4-32b-0414",
"hugging_face_id": "THUDM/GLM-4-32B-0414",
"name": "THUDM: GLM 4 32B",
"created": 1744920915,
@@ -2684,6 +3212,7 @@
},
{
"id": "google/gemini-2.5-flash-preview",
+ "canonical_slug": "google/gemini-2.5-flash-preview-04-17",
"hugging_face_id": "",
"name": "Google: Gemini 2.5 Flash Preview 04-17",
"created": 1744914667,
@@ -2731,6 +3260,7 @@
},
{
"id": "google/gemini-2.5-flash-preview:thinking",
+ "canonical_slug": "google/gemini-2.5-flash-preview-04-17",
"hugging_face_id": "",
"name": "Google: Gemini 2.5 Flash Preview 04-17 (thinking)",
"created": 1744914667,
@@ -2778,6 +3308,7 @@
},
{
"id": "openai/o4-mini-high",
+ "canonical_slug": "openai/o4-mini-high-2025-04-16",
"hugging_face_id": "",
"name": "OpenAI: o4 Mini High",
"created": 1744824212,
@@ -2822,6 +3353,7 @@
},
{
"id": "openai/o3",
+ "canonical_slug": "openai/o3-2025-04-16",
"hugging_face_id": "",
"name": "OpenAI: o3",
"created": 1744823457,
@@ -2841,13 +3373,13 @@
"instruct_type": null
},
"pricing": {
- "prompt": "0.00001",
- "completion": "0.00004",
+ "prompt": "0.000002",
+ "completion": "0.000008",
"request": "0",
- "image": "0.00765",
+ "image": "0.00153",
"web_search": "0",
"internal_reasoning": "0",
- "input_cache_read": "0.0000025"
+ "input_cache_read": "0.0000005"
},
"top_provider": {
"context_length": 200000,
@@ -2866,6 +3398,7 @@
},
{
"id": "openai/o4-mini",
+ "canonical_slug": "openai/o4-mini-2025-04-16",
"hugging_face_id": "",
"name": "OpenAI: o4 Mini",
"created": 1744820942,
@@ -2909,6 +3442,7 @@
},
{
"id": "shisa-ai/shisa-v2-llama3.3-70b:free",
+ "canonical_slug": "shisa-ai/shisa-v2-llama3.3-70b",
"hugging_face_id": "shisa-ai/shisa-v2-llama3.3-70b",
"name": "Shisa AI: Shisa V2 Llama 3.3 70B (free)",
"created": 1744754858,
@@ -2957,6 +3491,7 @@
},
{
"id": "openai/gpt-4.1",
+ "canonical_slug": "openai/gpt-4.1-2025-04-14",
"hugging_face_id": "",
"name": "OpenAI: GPT-4.1",
"created": 1744651385,
@@ -3010,6 +3545,7 @@
},
{
"id": "openai/gpt-4.1-mini",
+ "canonical_slug": "openai/gpt-4.1-mini-2025-04-14",
"hugging_face_id": "",
"name": "OpenAI: GPT-4.1 Mini",
"created": 1744651381,
@@ -3063,6 +3599,7 @@
},
{
"id": "openai/gpt-4.1-nano",
+ "canonical_slug": "openai/gpt-4.1-nano-2025-04-14",
"hugging_face_id": "",
"name": "OpenAI: GPT-4.1 Nano",
"created": 1744651369,
@@ -3115,6 +3652,7 @@
},
{
"id": "eleutherai/llemma_7b",
+ "canonical_slug": "eleutherai/llemma_7b",
"hugging_face_id": "EleutherAI/llemma_7b",
"name": "EleutherAI: Llemma 7b",
"created": 1744643225,
@@ -3160,6 +3698,7 @@
},
{
"id": "alfredpros/codellama-7b-instruct-solidity",
+ "canonical_slug": "alfredpros/codellama-7b-instruct-solidity",
"hugging_face_id": "AlfredPros/CodeLlama-7b-Instruct-Solidity",
"name": "AlfredPros: CodeLLaMa 7B Instruct Solidity",
"created": 1744641874,
@@ -3205,6 +3744,7 @@
},
{
"id": "arliai/qwq-32b-arliai-rpr-v1:free",
+ "canonical_slug": "arliai/qwq-32b-arliai-rpr-v1",
"hugging_face_id": "ArliAI/QwQ-32B-ArliAI-RpR-v1",
"name": "ArliAI: QwQ 32B RpR v1 (free)",
"created": 1744555982,
@@ -3255,6 +3795,7 @@
},
{
"id": "agentica-org/deepcoder-14b-preview:free",
+ "canonical_slug": "agentica-org/deepcoder-14b-preview",
"hugging_face_id": "agentica-org/DeepCoder-14B-Preview",
"name": "Agentica: Deepcoder 14B Preview (free)",
"created": 1744555395,
@@ -3305,6 +3846,7 @@
},
{
"id": "moonshotai/kimi-vl-a3b-thinking:free",
+ "canonical_slug": "moonshotai/kimi-vl-a3b-thinking",
"hugging_face_id": "moonshotai/Kimi-VL-A3B-Thinking",
"name": "Moonshot AI: Kimi VL A3B Thinking (free)",
"created": 1744304841,
@@ -3356,6 +3898,7 @@
},
{
"id": "x-ai/grok-3-mini-beta",
+ "canonical_slug": "x-ai/grok-3-mini-beta",
"hugging_face_id": "",
"name": "xAI: Grok 3 Mini Beta",
"created": 1744240195,
@@ -3378,7 +3921,8 @@
"request": "0",
"image": "0",
"web_search": "0",
- "internal_reasoning": "0"
+ "internal_reasoning": "0",
+ "input_cache_read": "0.000000075"
},
"top_provider": {
"context_length": 131072,
@@ -3403,6 +3947,7 @@
},
{
"id": "x-ai/grok-3-beta",
+ "canonical_slug": "x-ai/grok-3-beta",
"hugging_face_id": "",
"name": "xAI: Grok 3 Beta",
"created": 1744240068,
@@ -3425,7 +3970,8 @@
"request": "0",
"image": "0",
"web_search": "0",
- "internal_reasoning": "0"
+ "internal_reasoning": "0",
+ "input_cache_read": "0.00000075"
},
"top_provider": {
"context_length": 131072,
@@ -3450,6 +3996,7 @@
},
{
"id": "nvidia/llama-3.3-nemotron-super-49b-v1:free",
+ "canonical_slug": "nvidia/llama-3.3-nemotron-super-49b-v1",
"hugging_face_id": "nvidia/Llama-3_3-Nemotron-Super-49B-v1",
"name": "NVIDIA: Llama 3.3 Nemotron Super 49B v1 (free)",
"created": 1744119494,
@@ -3498,6 +4045,7 @@
},
{
"id": "nvidia/llama-3.3-nemotron-super-49b-v1",
+ "canonical_slug": "nvidia/llama-3.3-nemotron-super-49b-v1",
"hugging_face_id": "nvidia/Llama-3_3-Nemotron-Super-49B-v1",
"name": "NVIDIA: Llama 3.3 Nemotron Super 49B v1",
"created": 1744119494,
@@ -3544,6 +4092,7 @@
},
{
"id": "nvidia/llama-3.1-nemotron-ultra-253b-v1:free",
+ "canonical_slug": "nvidia/llama-3.1-nemotron-ultra-253b-v1",
"hugging_face_id": "nvidia/Llama-3_1-Nemotron-Ultra-253B-v1",
"name": "NVIDIA: Llama 3.1 Nemotron Ultra 253B v1 (free)",
"created": 1744115059,
@@ -3592,6 +4141,7 @@
},
{
"id": "nvidia/llama-3.1-nemotron-ultra-253b-v1",
+ "canonical_slug": "nvidia/llama-3.1-nemotron-ultra-253b-v1",
"hugging_face_id": "nvidia/Llama-3_1-Nemotron-Ultra-253B-v1",
"name": "NVIDIA: Llama 3.1 Nemotron Ultra 253B v1",
"created": 1744115059,
@@ -3640,6 +4190,7 @@
},
{
"id": "meta-llama/llama-4-maverick:free",
+ "canonical_slug": "meta-llama/llama-4-maverick-17b-128e-instruct",
"hugging_face_id": "meta-llama/Llama-4-Maverick-17B-128E-Instruct",
"name": "Meta: Llama 4 Maverick (free)",
"created": 1743881822,
@@ -3686,13 +4237,12 @@
"repetition_penalty",
"logprobs",
"logit_bias",
- "top_logprobs",
- "tools",
- "tool_choice"
+ "top_logprobs"
]
},
{
"id": "meta-llama/llama-4-maverick",
+ "canonical_slug": "meta-llama/llama-4-maverick-17b-128e-instruct",
"hugging_face_id": "meta-llama/Llama-4-Maverick-17B-128E-Instruct",
"name": "Meta: Llama 4 Maverick",
"created": 1743881822,
@@ -3737,20 +4287,21 @@
"seed",
"min_p",
"logit_bias",
- "logprobs",
- "top_logprobs",
"tools",
"tool_choice",
- "structured_outputs"
+ "structured_outputs",
+ "logprobs",
+ "top_logprobs"
]
},
{
"id": "meta-llama/llama-4-scout:free",
+ "canonical_slug": "meta-llama/llama-4-scout-17b-16e-instruct",
"hugging_face_id": "meta-llama/Llama-4-Scout-17B-16E-Instruct",
"name": "Meta: Llama 4 Scout (free)",
"created": 1743881519,
"description": "Llama 4 Scout 17B Instruct (16E) is a mixture-of-experts (MoE) language model developed by Meta, activating 17 billion parameters out of a total of 109B. It supports native multimodal input (text and image) and multilingual output (text and code) across 12 supported languages. Designed for assistant-style interaction and visual reasoning, Scout uses 16 experts per forward pass and features a context length of 10 million tokens, with a training corpus of ~40 trillion tokens.\n\nBuilt for high efficiency and local or commercial deployment, Llama 4 Scout incorporates early fusion for seamless modality integration. It is instruction-tuned for use in multilingual chat, captioning, and image understanding tasks. Released under the Llama 4 Community License, it was last trained on data up to August 2024 and launched publicly on April 5, 2025.",
- "context_length": 200000,
+ "context_length": 96000,
"architecture": {
"modality": "text+image->text",
"input_modalities": [
@@ -3772,7 +4323,7 @@
"internal_reasoning": "0"
},
"top_provider": {
- "context_length": 200000,
+ "context_length": 96000,
"max_completion_tokens": null,
"is_moderated": false
},
@@ -3792,13 +4343,12 @@
"repetition_penalty",
"logprobs",
"logit_bias",
- "top_logprobs",
- "tools",
- "tool_choice"
+ "top_logprobs"
]
},
{
"id": "meta-llama/llama-4-scout",
+ "canonical_slug": "meta-llama/llama-4-scout-17b-16e-instruct",
"hugging_face_id": "meta-llama/Llama-4-Scout-17B-16E-Instruct",
"name": "Meta: Llama 4 Scout",
"created": 1743881519,
@@ -3852,6 +4402,7 @@
},
{
"id": "all-hands/openhands-lm-32b-v0.1",
+ "canonical_slug": "all-hands/openhands-lm-32b-v0.1",
"hugging_face_id": "all-hands/openhands-lm-32b-v0.1",
"name": "OpenHands LM 32B V0.1",
"created": 1743613013,
@@ -3899,6 +4450,7 @@
},
{
"id": "deepseek/deepseek-v3-base:free",
+ "canonical_slug": "deepseek/deepseek-v3-base",
"hugging_face_id": "deepseek-ai/Deepseek-v3-base",
"name": "DeepSeek: DeepSeek V3 Base (free)",
"created": 1743272023,
@@ -3945,54 +4497,9 @@
"top_logprobs"
]
},
- {
- "id": "scb10x/llama3.1-typhoon2-8b-instruct",
- "hugging_face_id": "scb10x/llama3.1-typhoon2-8b-instruct",
- "name": "Typhoon2 8B Instruct",
- "created": 1743196511,
- "description": "Llama3.1-Typhoon2-8B-Instruct is a Thai-English instruction-tuned model with 8 billion parameters, built on Llama 3.1. It significantly improves over its base model in Thai reasoning, instruction-following, and function-calling tasks, while maintaining competitive English performance. The model is optimized for bilingual interaction and performs well on Thai-English code-switching, MT-Bench, IFEval, and tool-use benchmarks.\n\nDespite its smaller size, it demonstrates strong generalization across math, coding, and multilingual benchmarks, outperforming comparable 8B models across most Thai-specific tasks. Full benchmark results and methodology are available in the [technical report.](https://arxiv.org/abs/2412.13702)",
- "context_length": 8192,
- "architecture": {
- "modality": "text->text",
- "input_modalities": [
- "text"
- ],
- "output_modalities": [
- "text"
- ],
- "tokenizer": "Llama3",
- "instruct_type": "llama3"
- },
- "pricing": {
- "prompt": "0.00000018",
- "completion": "0.00000018",
- "request": "0",
- "image": "0",
- "web_search": "0",
- "internal_reasoning": "0"
- },
- "top_provider": {
- "context_length": 8192,
- "max_completion_tokens": null,
- "is_moderated": false
- },
- "per_request_limits": null,
- "supported_parameters": [
- "max_tokens",
- "temperature",
- "top_p",
- "stop",
- "frequency_penalty",
- "presence_penalty",
- "top_k",
- "repetition_penalty",
- "logit_bias",
- "min_p",
- "response_format"
- ]
- },
{
"id": "scb10x/llama3.1-typhoon2-70b-instruct",
+ "canonical_slug": "scb10x/llama3.1-typhoon2-70b-instruct",
"hugging_face_id": "scb10x/llama3.1-typhoon2-70b-instruct",
"name": "Typhoon2 70B Instruct",
"created": 1743196170,
@@ -4037,57 +4544,9 @@
"response_format"
]
},
- {
- "id": "qwen/qwen2.5-vl-3b-instruct:free",
- "hugging_face_id": "Qwen/Qwen2.5-VL-3B-Instruct",
- "name": "Qwen: Qwen2.5 VL 3B Instruct (free)",
- "created": 1743014573,
- "description": "Qwen2.5 VL 3B is a multimodal LLM from the Qwen Team with the following key enhancements:\n\n- SoTA understanding of images of various resolution & ratio: Qwen2.5-VL achieves state-of-the-art performance on visual understanding benchmarks, including MathVista, DocVQA, RealWorldQA, MTVQA, etc.\n\n- Agent that can operate your mobiles, robots, etc.: with the abilities of complex reasoning and decision making, Qwen2.5-VL can be integrated with devices like mobile phones, robots, etc., for automatic operation based on visual environment and text instructions.\n\n- Multilingual Support: to serve global users, besides English and Chinese, Qwen2.5-VL now supports the understanding of texts in different languages inside images, including most European languages, Japanese, Korean, Arabic, Vietnamese, etc.\n\nFor more details, see this [blog post](https://qwenlm.github.io/blog/qwen2-vl/) and [GitHub repo](https://github.com/QwenLM/Qwen2-VL).\n\nUsage of this model is subject to [Tongyi Qianwen LICENSE AGREEMENT](https://huggingface.co/Qwen/Qwen1.5-110B-Chat/blob/main/LICENSE).",
- "context_length": 64000,
- "architecture": {
- "modality": "text+image->text",
- "input_modalities": [
- "text",
- "image"
- ],
- "output_modalities": [
- "text"
- ],
- "tokenizer": "Qwen",
- "instruct_type": null
- },
- "pricing": {
- "prompt": "0",
- "completion": "0",
- "request": "0",
- "image": "0",
- "web_search": "0",
- "internal_reasoning": "0"
- },
- "top_provider": {
- "context_length": 64000,
- "max_completion_tokens": null,
- "is_moderated": false
- },
- "per_request_limits": null,
- "supported_parameters": [
- "max_tokens",
- "temperature",
- "top_p",
- "stop",
- "frequency_penalty",
- "presence_penalty",
- "seed",
- "top_k",
- "min_p",
- "repetition_penalty",
- "logprobs",
- "logit_bias",
- "top_logprobs"
- ]
- },
{
"id": "google/gemini-2.5-pro-exp-03-25",
+ "canonical_slug": "google/gemini-2.5-pro-exp-03-25",
"hugging_face_id": "",
"name": "Google: Gemini 2.5 Pro Experimental",
"created": 1742922099,
@@ -4134,6 +4593,7 @@
},
{
"id": "qwen/qwen2.5-vl-32b-instruct:free",
+ "canonical_slug": "qwen/qwen2.5-vl-32b-instruct",
"hugging_face_id": "Qwen/Qwen2.5-VL-32B-Instruct",
"name": "Qwen: Qwen2.5 VL 32B Instruct (free)",
"created": 1742839838,
@@ -4184,6 +4644,7 @@
},
{
"id": "qwen/qwen2.5-vl-32b-instruct",
+ "canonical_slug": "qwen/qwen2.5-vl-32b-instruct",
"hugging_face_id": "Qwen/Qwen2.5-VL-32B-Instruct",
"name": "Qwen: Qwen2.5 VL 32B Instruct",
"created": 1742839838,
@@ -4233,6 +4694,7 @@
},
{
"id": "deepseek/deepseek-chat-v3-0324:free",
+ "canonical_slug": "deepseek/deepseek-chat-v3-0324",
"hugging_face_id": "deepseek-ai/DeepSeek-V3-0324",
"name": "DeepSeek: DeepSeek V3 0324 (free)",
"created": 1742824755,
@@ -4284,6 +4746,7 @@
},
{
"id": "deepseek/deepseek-chat-v3-0324",
+ "canonical_slug": "deepseek/deepseek-chat-v3-0324",
"hugging_face_id": "deepseek-ai/DeepSeek-V3-0324",
"name": "DeepSeek: DeepSeek V3 0324",
"created": 1742824755,
@@ -4301,7 +4764,7 @@
"instruct_type": null
},
"pricing": {
- "prompt": "0.0000003",
+ "prompt": "0.00000028",
"completion": "0.00000088",
"request": "0",
"image": "0",
@@ -4315,27 +4778,28 @@
},
"per_request_limits": null,
"supported_parameters": [
- "tools",
- "tool_choice",
"max_tokens",
"temperature",
"top_p",
- "structured_outputs",
- "response_format",
"stop",
"frequency_penalty",
"presence_penalty",
- "top_k",
- "repetition_penalty",
+ "seed",
"logit_bias",
"logprobs",
"top_logprobs",
- "seed",
- "min_p"
+ "min_p",
+ "repetition_penalty",
+ "top_k",
+ "tools",
+ "tool_choice",
+ "structured_outputs",
+ "response_format"
]
},
{
"id": "featherless/qwerky-72b:free",
+ "canonical_slug": "featherless/qwerky-72b",
"hugging_face_id": "featherless-ai/Qwerky-72B",
"name": "Qwerky 72B (free)",
"created": 1742481597,
@@ -4381,6 +4845,7 @@
},
{
"id": "openai/o1-pro",
+ "canonical_slug": "openai/o1-pro",
"hugging_face_id": "",
"name": "OpenAI: o1-pro",
"created": 1742423211,
@@ -4421,10 +4886,11 @@
},
{
"id": "mistralai/mistral-small-3.1-24b-instruct:free",
+ "canonical_slug": "mistralai/mistral-small-3.1-24b-instruct-2503",
"hugging_face_id": "mistralai/Mistral-Small-3.1-24B-Instruct-2503",
"name": "Mistral: Mistral Small 3.1 24B (free)",
"created": 1742238937,
- "description": "Mistral Small 3.1 24B Instruct is an upgraded variant of Mistral Small 3 (2501), featuring 24 billion parameters with advanced multimodal capabilities. It provides state-of-the-art performance in text-based reasoning and vision tasks, including image analysis, programming, mathematical reasoning, and multilingual support across dozens of languages. Equipped with an extensive 128k token context window and optimized for efficient local inference, it supports use cases such as conversational agents, function calling, long-document comprehension, and privacy-sensitive deployments.",
+ "description": "Mistral Small 3.1 24B Instruct is an upgraded variant of Mistral Small 3 (2501), featuring 24 billion parameters with advanced multimodal capabilities. It provides state-of-the-art performance in text-based reasoning and vision tasks, including image analysis, programming, mathematical reasoning, and multilingual support across dozens of languages. Equipped with an extensive 128k token context window and optimized for efficient local inference, it supports use cases such as conversational agents, function calling, long-document comprehension, and privacy-sensitive deployments. The updated version is [Mistral Small 3.2](mistralai/mistral-small-3.2-24b-instruct)",
"context_length": 96000,
"architecture": {
"modality": "text+image->text",
@@ -4472,11 +4938,12 @@
},
{
"id": "mistralai/mistral-small-3.1-24b-instruct",
+ "canonical_slug": "mistralai/mistral-small-3.1-24b-instruct-2503",
"hugging_face_id": "mistralai/Mistral-Small-3.1-24B-Instruct-2503",
"name": "Mistral: Mistral Small 3.1 24B",
"created": 1742238937,
- "description": "Mistral Small 3.1 24B Instruct is an upgraded variant of Mistral Small 3 (2501), featuring 24 billion parameters with advanced multimodal capabilities. It provides state-of-the-art performance in text-based reasoning and vision tasks, including image analysis, programming, mathematical reasoning, and multilingual support across dozens of languages. Equipped with an extensive 128k token context window and optimized for efficient local inference, it supports use cases such as conversational agents, function calling, long-document comprehension, and privacy-sensitive deployments.",
- "context_length": 131072,
+ "description": "Mistral Small 3.1 24B Instruct is an upgraded variant of Mistral Small 3 (2501), featuring 24 billion parameters with advanced multimodal capabilities. It provides state-of-the-art performance in text-based reasoning and vision tasks, including image analysis, programming, mathematical reasoning, and multilingual support across dozens of languages. Equipped with an extensive 128k token context window and optimized for efficient local inference, it supports use cases such as conversational agents, function calling, long-document comprehension, and privacy-sensitive deployments. The updated version is [Mistral Small 3.2](mistralai/mistral-small-3.2-24b-instruct)",
+ "context_length": 128000,
"architecture": {
"modality": "text+image->text",
"input_modalities": [
@@ -4491,14 +4958,14 @@
},
"pricing": {
"prompt": "0.00000005",
- "completion": "0.00000015",
+ "completion": "0.0000001",
"request": "0",
"image": "0",
"web_search": "0",
"internal_reasoning": "0"
},
"top_provider": {
- "context_length": 131072,
+ "context_length": 128000,
"max_completion_tokens": null,
"is_moderated": false
},
@@ -4507,127 +4974,30 @@
"max_tokens",
"temperature",
"top_p",
- "presence_penalty",
+ "stop",
"frequency_penalty",
+ "presence_penalty",
"repetition_penalty",
+ "response_format",
"top_k",
+ "seed",
+ "min_p",
"tools",
"tool_choice",
- "stop",
- "response_format",
"structured_outputs",
- "seed",
"logit_bias",
"logprobs",
"top_logprobs"
]
},
- {
- "id": "open-r1/olympiccoder-32b:free",
- "hugging_face_id": "open-r1/OlympicCoder-32B",
- "name": "OlympicCoder 32B (free)",
- "created": 1742077228,
- "description": "OlympicCoder-32B is a high-performing open-source model fine-tuned using the CodeForces-CoTs dataset, containing approximately 100,000 chain-of-thought programming samples. It excels at complex competitive programming benchmarks, such as IOI 2024 and Codeforces-style challenges, frequently surpassing state-of-the-art closed-source models. OlympicCoder-32B provides advanced reasoning, coherent multi-step problem-solving, and robust code generation capabilities, demonstrating significant potential for olympiad-level competitive programming applications.",
- "context_length": 32768,
- "architecture": {
- "modality": "text->text",
- "input_modalities": [
- "text"
- ],
- "output_modalities": [
- "text"
- ],
- "tokenizer": "Other",
- "instruct_type": "deepseek-r1"
- },
- "pricing": {
- "prompt": "0",
- "completion": "0",
- "request": "0",
- "image": "0",
- "web_search": "0",
- "internal_reasoning": "0"
- },
- "top_provider": {
- "context_length": 32768,
- "max_completion_tokens": null,
- "is_moderated": false
- },
- "per_request_limits": null,
- "supported_parameters": [
- "max_tokens",
- "temperature",
- "top_p",
- "reasoning",
- "include_reasoning",
- "stop",
- "frequency_penalty",
- "presence_penalty",
- "seed",
- "top_k",
- "min_p",
- "repetition_penalty",
- "logprobs",
- "logit_bias",
- "top_logprobs"
- ]
- },
- {
- "id": "google/gemma-3-1b-it:free",
- "hugging_face_id": "google/gemma-3-1b-it",
- "name": "Google: Gemma 3 1B (free)",
- "created": 1741963556,
- "description": "Gemma 3 1B is the smallest of the new Gemma 3 family. It handles context windows up to 32k tokens, understands over 140 languages, and offers improved math, reasoning, and chat capabilities, including structured outputs and function calling. Note: Gemma 3 1B is not multimodal. For the smallest multimodal Gemma 3 model, please see [Gemma 3 4B](google/gemma-3-4b-it)",
- "context_length": 32768,
- "architecture": {
- "modality": "text+image->text",
- "input_modalities": [
- "text",
- "image"
- ],
- "output_modalities": [
- "text"
- ],
- "tokenizer": "Gemini",
- "instruct_type": "gemma"
- },
- "pricing": {
- "prompt": "0",
- "completion": "0",
- "request": "0",
- "image": "0",
- "web_search": "0",
- "internal_reasoning": "0"
- },
- "top_provider": {
- "context_length": 32768,
- "max_completion_tokens": 8192,
- "is_moderated": false
- },
- "per_request_limits": null,
- "supported_parameters": [
- "max_tokens",
- "temperature",
- "top_p",
- "stop",
- "frequency_penalty",
- "presence_penalty",
- "seed",
- "top_k",
- "min_p",
- "repetition_penalty",
- "logprobs",
- "logit_bias",
- "top_logprobs"
- ]
- },
{
"id": "google/gemma-3-4b-it:free",
+ "canonical_slug": "google/gemma-3-4b-it",
"hugging_face_id": "google/gemma-3-4b-it",
"name": "Google: Gemma 3 4B (free)",
"created": 1741905510,
"description": "Gemma 3 introduces multimodality, supporting vision-language input and text outputs. It handles context windows up to 128k tokens, understands over 140 languages, and offers improved math, reasoning, and chat capabilities, including structured outputs and function calling.",
- "context_length": 96000,
+ "context_length": 32768,
"architecture": {
"modality": "text+image->text",
"input_modalities": [
@@ -4649,7 +5019,7 @@
"internal_reasoning": "0"
},
"top_provider": {
- "context_length": 96000,
+ "context_length": 32768,
"max_completion_tokens": 8192,
"is_moderated": false
},
@@ -4658,22 +5028,14 @@
"max_tokens",
"temperature",
"top_p",
- "stop",
- "frequency_penalty",
- "presence_penalty",
"seed",
- "top_k",
- "min_p",
- "repetition_penalty",
- "logprobs",
- "logit_bias",
- "top_logprobs",
"response_format",
"structured_outputs"
]
},
{
"id": "google/gemma-3-4b-it",
+ "canonical_slug": "google/gemma-3-4b-it",
"hugging_face_id": "google/gemma-3-4b-it",
"name": "Google: Gemma 3 4B",
"created": 1741905510,
@@ -4721,6 +5083,7 @@
},
{
"id": "ai21/jamba-1.6-large",
+ "canonical_slug": "ai21/jamba-1.6-large",
"hugging_face_id": "ai21labs/AI21-Jamba-Large-1.6",
"name": "AI21: Jamba 1.6 Large",
"created": 1741905173,
@@ -4762,6 +5125,7 @@
},
{
"id": "ai21/jamba-1.6-mini",
+ "canonical_slug": "ai21/jamba-1.6-mini",
"hugging_face_id": "ai21labs/AI21-Jamba-Mini-1.6",
"name": "AI21: Jamba Mini 1.6",
"created": 1741905171,
@@ -4803,6 +5167,7 @@
},
{
"id": "google/gemma-3-12b-it:free",
+ "canonical_slug": "google/gemma-3-12b-it",
"hugging_face_id": "google/gemma-3-12b-it",
"name": "Google: Gemma 3 12B (free)",
"created": 1741902625,
@@ -4854,6 +5219,7 @@
},
{
"id": "google/gemma-3-12b-it",
+ "canonical_slug": "google/gemma-3-12b-it",
"hugging_face_id": "google/gemma-3-12b-it",
"name": "Google: Gemma 3 12B",
"created": 1741902625,
@@ -4901,6 +5267,7 @@
},
{
"id": "cohere/command-a",
+ "canonical_slug": "cohere/command-a-03-2025",
"hugging_face_id": "CohereForAI/c4ai-command-a-03-2025",
"name": "Cohere: Command A",
"created": 1741894342,
@@ -4946,6 +5313,7 @@
},
{
"id": "openai/gpt-4o-mini-search-preview",
+ "canonical_slug": "openai/gpt-4o-mini-search-preview-2025-03-11",
"hugging_face_id": "",
"name": "OpenAI: GPT-4o-mini Search Preview",
"created": 1741818122,
@@ -4985,6 +5353,7 @@
},
{
"id": "openai/gpt-4o-search-preview",
+ "canonical_slug": "openai/gpt-4o-search-preview-2025-03-11",
"hugging_face_id": "",
"name": "OpenAI: GPT-4o Search Preview",
"created": 1741817949,
@@ -5024,6 +5393,7 @@
},
{
"id": "rekaai/reka-flash-3:free",
+ "canonical_slug": "rekaai/reka-flash-3",
"hugging_face_id": "RekaAI/reka-flash-3",
"name": "Reka: Flash 3 (free)",
"created": 1741812813,
@@ -5074,6 +5444,7 @@
},
{
"id": "google/gemma-3-27b-it:free",
+ "canonical_slug": "google/gemma-3-27b-it",
"hugging_face_id": "",
"name": "Google: Gemma 3 27B (free)",
"created": 1741756359,
@@ -5125,11 +5496,12 @@
},
{
"id": "google/gemma-3-27b-it",
+ "canonical_slug": "google/gemma-3-27b-it",
"hugging_face_id": "",
"name": "Google: Gemma 3 27B",
"created": 1741756359,
"description": "Gemma 3 introduces multimodality, supporting vision-language input and text outputs. It handles context windows up to 128k tokens, understands over 140 languages, and offers improved math, reasoning, and chat capabilities, including structured outputs and function calling. Gemma 3 27B is Google's latest open source model, successor to [Gemma 2](google/gemma-2-27b-it)",
- "context_length": 131072,
+ "context_length": 64000,
"architecture": {
"modality": "text+image->text",
"input_modalities": [
@@ -5144,15 +5516,15 @@
},
"pricing": {
"prompt": "0.0000001",
- "completion": "0.0000002",
+ "completion": "0.00000018",
"request": "0",
- "image": "0.0000256",
+ "image": "0",
"web_search": "0",
"internal_reasoning": "0"
},
"top_provider": {
- "context_length": 131072,
- "max_completion_tokens": 16384,
+ "context_length": 64000,
+ "max_completion_tokens": 8000,
"is_moderated": false
},
"per_request_limits": null,
@@ -5175,6 +5547,7 @@
},
{
"id": "thedrummer/anubis-pro-105b-v1",
+ "canonical_slug": "thedrummer/anubis-pro-105b-v1",
"hugging_face_id": "TheDrummer/Anubis-Pro-105B-v1",
"name": "TheDrummer: Anubis Pro 105B V1",
"created": 1741642290,
@@ -5217,6 +5590,7 @@
},
{
"id": "thedrummer/skyfall-36b-v2",
+ "canonical_slug": "thedrummer/skyfall-36b-v2",
"hugging_face_id": "TheDrummer/Skyfall-36B-v2",
"name": "TheDrummer: Skyfall 36B V2",
"created": 1741636566,
@@ -5259,6 +5633,7 @@
},
{
"id": "microsoft/phi-4-multimodal-instruct",
+ "canonical_slug": "microsoft/phi-4-multimodal-instruct",
"hugging_face_id": "microsoft/Phi-4-multimodal-instruct",
"name": "Microsoft: Phi 4 Multimodal Instruct",
"created": 1741396284,
@@ -5306,6 +5681,7 @@
},
{
"id": "perplexity/sonar-reasoning-pro",
+ "canonical_slug": "perplexity/sonar-reasoning-pro",
"hugging_face_id": "",
"name": "Perplexity: Sonar Reasoning Pro",
"created": 1741313308,
@@ -5351,6 +5727,7 @@
},
{
"id": "perplexity/sonar-pro",
+ "canonical_slug": "perplexity/sonar-pro",
"hugging_face_id": "",
"name": "Perplexity: Sonar Pro",
"created": 1741312423,
@@ -5394,6 +5771,7 @@
},
{
"id": "perplexity/sonar-deep-research",
+ "canonical_slug": "perplexity/sonar-deep-research",
"hugging_face_id": "",
"name": "Perplexity: Sonar Deep Research",
"created": 1741311246,
@@ -5435,63 +5813,14 @@
"presence_penalty"
]
},
- {
- "id": "deepseek/deepseek-r1-zero:free",
- "hugging_face_id": "deepseek-ai/DeepSeek-R1-Zero",
- "name": "DeepSeek: DeepSeek R1 Zero (free)",
- "created": 1741297434,
- "description": "DeepSeek-R1-Zero is a model trained via large-scale reinforcement learning (RL) without supervised fine-tuning (SFT) as a preliminary step. It's 671B parameters in size, with 37B active in an inference pass.\n\nIt demonstrates remarkable performance on reasoning. With RL, DeepSeek-R1-Zero naturally emerged with numerous powerful and interesting reasoning behaviors.\n\nDeepSeek-R1-Zero encounters challenges such as endless repetition, poor readability, and language mixing. See [DeepSeek R1](/deepseek/deepseek-r1) for the SFT model.\n\n",
- "context_length": 163840,
- "architecture": {
- "modality": "text->text",
- "input_modalities": [
- "text"
- ],
- "output_modalities": [
- "text"
- ],
- "tokenizer": "Other",
- "instruct_type": "deepseek-r1"
- },
- "pricing": {
- "prompt": "0",
- "completion": "0",
- "request": "0",
- "image": "0",
- "web_search": "0",
- "internal_reasoning": "0"
- },
- "top_provider": {
- "context_length": 163840,
- "max_completion_tokens": null,
- "is_moderated": false
- },
- "per_request_limits": null,
- "supported_parameters": [
- "max_tokens",
- "temperature",
- "top_p",
- "reasoning",
- "include_reasoning",
- "stop",
- "frequency_penalty",
- "presence_penalty",
- "seed",
- "top_k",
- "min_p",
- "repetition_penalty",
- "logprobs",
- "logit_bias",
- "top_logprobs"
- ]
- },
{
"id": "qwen/qwq-32b:free",
+ "canonical_slug": "qwen/qwq-32b",
"hugging_face_id": "Qwen/QwQ-32B",
"name": "Qwen: QwQ 32B (free)",
"created": 1741208814,
"description": "QwQ is the reasoning model of the Qwen series. Compared with conventional instruction-tuned models, QwQ, which is capable of thinking and reasoning, can achieve significantly enhanced performance in downstream tasks, especially hard problems. QwQ-32B is the medium-sized reasoning model, which is capable of achieving competitive performance against state-of-the-art reasoning models, e.g., DeepSeek-R1, o1-mini.",
- "context_length": 40000,
+ "context_length": 131072,
"architecture": {
"modality": "text->text",
"input_modalities": [
@@ -5512,9 +5841,9 @@
"internal_reasoning": "0"
},
"top_provider": {
- "context_length": 40000,
- "max_completion_tokens": 40000,
- "is_moderated": false
+ "context_length": 131072,
+ "max_completion_tokens": null,
+ "is_moderated": true
},
"per_request_limits": null,
"supported_parameters": [
@@ -5537,6 +5866,7 @@
},
{
"id": "qwen/qwq-32b",
+ "canonical_slug": "qwen/qwq-32b",
"hugging_face_id": "Qwen/QwQ-32B",
"name": "Qwen: QwQ 32B",
"created": 1741208814,
@@ -5554,8 +5884,8 @@
"instruct_type": "qwq"
},
"pricing": {
- "prompt": "0.00000015",
- "completion": "0.0000002",
+ "prompt": "0.000000075",
+ "completion": "0.00000015",
"request": "0",
"image": "0",
"web_search": "0",
@@ -5587,56 +5917,9 @@
"structured_outputs"
]
},
- {
- "id": "moonshotai/moonlight-16b-a3b-instruct:free",
- "hugging_face_id": "moonshotai/Moonlight-16B-A3B-Instruct",
- "name": "Moonshot AI: Moonlight 16B A3B Instruct (free)",
- "created": 1740719801,
- "description": "Moonlight-16B-A3B-Instruct is a 16B-parameter Mixture-of-Experts (MoE) language model developed by Moonshot AI. It is optimized for instruction-following tasks with 3B activated parameters per inference. The model advances the Pareto frontier in performance per FLOP across English, coding, math, and Chinese benchmarks. It outperforms comparable models like Llama3-3B and Deepseek-v2-Lite while maintaining efficient deployment capabilities through Hugging Face integration and compatibility with popular inference engines like vLLM12.",
- "context_length": 8192,
- "architecture": {
- "modality": "text->text",
- "input_modalities": [
- "text"
- ],
- "output_modalities": [
- "text"
- ],
- "tokenizer": "Other",
- "instruct_type": null
- },
- "pricing": {
- "prompt": "0",
- "completion": "0",
- "request": "0",
- "image": "0",
- "web_search": "0",
- "internal_reasoning": "0"
- },
- "top_provider": {
- "context_length": 8192,
- "max_completion_tokens": null,
- "is_moderated": false
- },
- "per_request_limits": null,
- "supported_parameters": [
- "max_tokens",
- "temperature",
- "top_p",
- "stop",
- "frequency_penalty",
- "presence_penalty",
- "seed",
- "top_k",
- "min_p",
- "repetition_penalty",
- "logprobs",
- "logit_bias",
- "top_logprobs"
- ]
- },
{
"id": "nousresearch/deephermes-3-llama-3-8b-preview:free",
+ "canonical_slug": "nousresearch/deephermes-3-llama-3-8b-preview",
"hugging_face_id": "NousResearch/DeepHermes-3-Llama-3-8B-Preview",
"name": "Nous: DeepHermes 3 Llama 3 8B Preview (free)",
"created": 1740719372,
@@ -5685,6 +5968,7 @@
},
{
"id": "openai/gpt-4.5-preview",
+ "canonical_slug": "openai/gpt-4.5-preview-2025-02-27",
"hugging_face_id": "",
"name": "OpenAI: GPT-4.5 (Preview)",
"created": 1740687810,
@@ -5736,6 +6020,7 @@
},
{
"id": "google/gemini-2.0-flash-lite-001",
+ "canonical_slug": "google/gemini-2.0-flash-lite-001",
"hugging_face_id": "",
"name": "Google: Gemini 2.0 Flash Lite",
"created": 1740506212,
@@ -5769,21 +6054,19 @@
},
"per_request_limits": null,
"supported_parameters": [
- "tools",
- "tool_choice",
"max_tokens",
"temperature",
"top_p",
+ "tools",
+ "tool_choice",
"stop",
- "frequency_penalty",
- "presence_penalty",
- "seed",
"response_format",
"structured_outputs"
]
},
{
"id": "anthropic/claude-3.7-sonnet",
+ "canonical_slug": "anthropic/claude-3-7-sonnet-20250219",
"hugging_face_id": "",
"name": "Anthropic: Claude 3.7 Sonnet",
"created": 1740422110,
@@ -5831,6 +6114,7 @@
},
{
"id": "anthropic/claude-3.7-sonnet:beta",
+ "canonical_slug": "anthropic/claude-3-7-sonnet-20250219",
"hugging_face_id": "",
"name": "Anthropic: Claude 3.7 Sonnet (self-moderated)",
"created": 1740422110,
@@ -5876,6 +6160,7 @@
},
{
"id": "anthropic/claude-3.7-sonnet:thinking",
+ "canonical_slug": "anthropic/claude-3-7-sonnet-20250219",
"hugging_face_id": "",
"name": "Anthropic: Claude 3.7 Sonnet (thinking)",
"created": 1740422110,
@@ -5921,6 +6206,7 @@
},
{
"id": "perplexity/r1-1776",
+ "canonical_slug": "perplexity/r1-1776",
"hugging_face_id": "perplexity-ai/r1-1776",
"name": "Perplexity: R1 1776",
"created": 1740004929,
@@ -5969,6 +6255,7 @@
},
{
"id": "mistralai/mistral-saba",
+ "canonical_slug": "mistralai/mistral-saba-2502",
"hugging_face_id": "",
"name": "Mistral: Saba",
"created": 1739803239,
@@ -6018,6 +6305,7 @@
},
{
"id": "cognitivecomputations/dolphin3.0-r1-mistral-24b:free",
+ "canonical_slug": "cognitivecomputations/dolphin3.0-r1-mistral-24b",
"hugging_face_id": "cognitivecomputations/Dolphin3.0-R1-Mistral-24B",
"name": "Dolphin3.0 R1 Mistral 24B (free)",
"created": 1739462498,
@@ -6068,6 +6356,7 @@
},
{
"id": "cognitivecomputations/dolphin3.0-mistral-24b:free",
+ "canonical_slug": "cognitivecomputations/dolphin3.0-mistral-24b",
"hugging_face_id": "cognitivecomputations/Dolphin3.0-Mistral-24B",
"name": "Dolphin3.0 Mistral 24B (free)",
"created": 1739462019,
@@ -6116,6 +6405,7 @@
},
{
"id": "meta-llama/llama-guard-3-8b",
+ "canonical_slug": "meta-llama/llama-guard-3-8b",
"hugging_face_id": "meta-llama/Llama-Guard-3-8B",
"name": "Llama Guard 3 8B",
"created": 1739401318,
@@ -6166,6 +6456,7 @@
},
{
"id": "openai/o3-mini-high",
+ "canonical_slug": "openai/o3-mini-high-2025-01-31",
"hugging_face_id": "",
"name": "OpenAI: o3 Mini High",
"created": 1739372611,
@@ -6208,6 +6499,7 @@
},
{
"id": "deepseek/deepseek-r1-distill-llama-8b",
+ "canonical_slug": "deepseek/deepseek-r1-distill-llama-8b",
"hugging_face_id": "deepseek-ai/DeepSeek-R1-Distill-Llama-8B",
"name": "DeepSeek: R1 Distill Llama 8B",
"created": 1738937718,
@@ -6256,6 +6548,7 @@
},
{
"id": "google/gemini-2.0-flash-001",
+ "canonical_slug": "google/gemini-2.0-flash-001",
"hugging_face_id": "",
"name": "Google: Gemini 2.0 Flash",
"created": 1738769413,
@@ -6291,21 +6584,19 @@
},
"per_request_limits": null,
"supported_parameters": [
- "tools",
- "tool_choice",
"max_tokens",
"temperature",
"top_p",
+ "tools",
+ "tool_choice",
"stop",
- "frequency_penalty",
- "presence_penalty",
- "seed",
"response_format",
"structured_outputs"
]
},
{
"id": "qwen/qwen-vl-plus",
+ "canonical_slug": "qwen/qwen-vl-plus",
"hugging_face_id": "",
"name": "Qwen: Qwen VL Plus",
"created": 1738731255,
@@ -6348,6 +6639,7 @@
},
{
"id": "aion-labs/aion-1.0",
+ "canonical_slug": "aion-labs/aion-1.0",
"hugging_face_id": "",
"name": "AionLabs: Aion-1.0",
"created": 1738697557,
@@ -6388,6 +6680,7 @@
},
{
"id": "aion-labs/aion-1.0-mini",
+ "canonical_slug": "aion-labs/aion-1.0-mini",
"hugging_face_id": "FuseAI/FuseO1-DeepSeekR1-QwQ-SkyT1-32B-Preview",
"name": "AionLabs: Aion-1.0-Mini",
"created": 1738697107,
@@ -6428,6 +6721,7 @@
},
{
"id": "aion-labs/aion-rp-llama-3.1-8b",
+ "canonical_slug": "aion-labs/aion-rp-llama-3.1-8b",
"hugging_face_id": "",
"name": "AionLabs: Aion-RP 1.0 (8B)",
"created": 1738696718,
@@ -6466,6 +6760,7 @@
},
{
"id": "qwen/qwen-vl-max",
+ "canonical_slug": "qwen/qwen-vl-max-2025-01-25",
"hugging_face_id": "",
"name": "Qwen: Qwen VL Max",
"created": 1738434304,
@@ -6508,6 +6803,7 @@
},
{
"id": "qwen/qwen-turbo",
+ "canonical_slug": "qwen/qwen-turbo-2024-11-01",
"hugging_face_id": "",
"name": "Qwen: Qwen-Turbo",
"created": 1738410974,
@@ -6552,6 +6848,7 @@
},
{
"id": "qwen/qwen2.5-vl-72b-instruct:free",
+ "canonical_slug": "qwen/qwen2.5-vl-72b-instruct",
"hugging_face_id": "Qwen/Qwen2.5-VL-72B-Instruct",
"name": "Qwen: Qwen2.5 VL 72B Instruct (free)",
"created": 1738410311,
@@ -6602,6 +6899,7 @@
},
{
"id": "qwen/qwen2.5-vl-72b-instruct",
+ "canonical_slug": "qwen/qwen2.5-vl-72b-instruct",
"hugging_face_id": "Qwen/Qwen2.5-VL-72B-Instruct",
"name": "Qwen: Qwen2.5 VL 72B Instruct",
"created": 1738410311,
@@ -6652,6 +6950,7 @@
},
{
"id": "qwen/qwen-plus",
+ "canonical_slug": "qwen/qwen-plus-2025-01-25",
"hugging_face_id": "",
"name": "Qwen: Qwen-Plus",
"created": 1738409840,
@@ -6696,6 +6995,7 @@
},
{
"id": "qwen/qwen-max",
+ "canonical_slug": "qwen/qwen-max-2025-01-25",
"hugging_face_id": "",
"name": "Qwen: Qwen-Max ",
"created": 1738402289,
@@ -6740,6 +7040,7 @@
},
{
"id": "openai/o3-mini",
+ "canonical_slug": "openai/o3-mini-2025-01-31",
"hugging_face_id": "",
"name": "OpenAI: o3 Mini",
"created": 1738351721,
@@ -6782,6 +7083,7 @@
},
{
"id": "deepseek/deepseek-r1-distill-qwen-1.5b",
+ "canonical_slug": "deepseek/deepseek-r1-distill-qwen-1.5b",
"hugging_face_id": "deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B",
"name": "DeepSeek: R1 Distill Qwen 1.5B",
"created": 1738328067,
@@ -6830,6 +7132,7 @@
},
{
"id": "mistralai/mistral-small-24b-instruct-2501:free",
+ "canonical_slug": "mistralai/mistral-small-24b-instruct-2501",
"hugging_face_id": "mistralai/Mistral-Small-24B-Instruct-2501",
"name": "Mistral: Mistral Small 3 (free)",
"created": 1738255409,
@@ -6878,6 +7181,7 @@
},
{
"id": "mistralai/mistral-small-24b-instruct-2501",
+ "canonical_slug": "mistralai/mistral-small-24b-instruct-2501",
"hugging_face_id": "mistralai/Mistral-Small-24B-Instruct-2501",
"name": "Mistral: Mistral Small 3",
"created": 1738255409,
@@ -6896,7 +7200,7 @@
},
"pricing": {
"prompt": "0.00000005",
- "completion": "0.00000011",
+ "completion": "0.00000009",
"request": "0",
"image": "0",
"web_search": "0",
@@ -6928,48 +7232,9 @@
"structured_outputs"
]
},
- {
- "id": "deepseek/deepseek-r1-distill-qwen-32b:free",
- "hugging_face_id": "deepseek-ai/DeepSeek-R1-Distill-Qwen-32B",
- "name": "DeepSeek: R1 Distill Qwen 32B (free)",
- "created": 1738194830,
- "description": "DeepSeek R1 Distill Qwen 32B is a distilled large language model based on [Qwen 2.5 32B](https://huggingface.co/Qwen/Qwen2.5-32B), using outputs from [DeepSeek R1](/deepseek/deepseek-r1). It outperforms OpenAI's o1-mini across various benchmarks, achieving new state-of-the-art results for dense models.\\n\\nOther benchmark results include:\\n\\n- AIME 2024 pass@1: 72.6\\n- MATH-500 pass@1: 94.3\\n- CodeForces Rating: 1691\\n\\nThe model leverages fine-tuning from DeepSeek R1's outputs, enabling competitive performance comparable to larger frontier models.",
- "context_length": 16000,
- "architecture": {
- "modality": "text->text",
- "input_modalities": [
- "text"
- ],
- "output_modalities": [
- "text"
- ],
- "tokenizer": "Qwen",
- "instruct_type": "deepseek-r1"
- },
- "pricing": {
- "prompt": "0",
- "completion": "0",
- "request": "0",
- "image": "0",
- "web_search": "0",
- "internal_reasoning": "0"
- },
- "top_provider": {
- "context_length": 16000,
- "max_completion_tokens": 16000,
- "is_moderated": false
- },
- "per_request_limits": null,
- "supported_parameters": [
- "max_tokens",
- "temperature",
- "top_p",
- "reasoning",
- "include_reasoning"
- ]
- },
{
"id": "deepseek/deepseek-r1-distill-qwen-32b",
+ "canonical_slug": "deepseek/deepseek-r1-distill-qwen-32b",
"hugging_face_id": "deepseek-ai/DeepSeek-R1-Distill-Qwen-32B",
"name": "DeepSeek: R1 Distill Qwen 32B",
"created": 1738194830,
@@ -6987,8 +7252,8 @@
"instruct_type": "deepseek-r1"
},
"pricing": {
- "prompt": "0.00000012",
- "completion": "0.00000018",
+ "prompt": "0.000000075",
+ "completion": "0.00000015",
"request": "0",
"image": "0",
"web_search": "0",
@@ -7019,6 +7284,7 @@
},
{
"id": "deepseek/deepseek-r1-distill-qwen-14b:free",
+ "canonical_slug": "deepseek/deepseek-r1-distill-qwen-14b",
"hugging_face_id": "deepseek-ai/DeepSeek-R1-Distill-Qwen-14B",
"name": "DeepSeek: R1 Distill Qwen 14B (free)",
"created": 1738193940,
@@ -7069,6 +7335,7 @@
},
{
"id": "deepseek/deepseek-r1-distill-qwen-14b",
+ "canonical_slug": "deepseek/deepseek-r1-distill-qwen-14b",
"hugging_face_id": "deepseek-ai/DeepSeek-R1-Distill-Qwen-14B",
"name": "DeepSeek: R1 Distill Qwen 14B",
"created": 1738193940,
@@ -7118,6 +7385,7 @@
},
{
"id": "perplexity/sonar-reasoning",
+ "canonical_slug": "perplexity/sonar-reasoning",
"hugging_face_id": "",
"name": "Perplexity: Sonar Reasoning",
"created": 1738131107,
@@ -7162,6 +7430,7 @@
},
{
"id": "perplexity/sonar",
+ "canonical_slug": "perplexity/sonar",
"hugging_face_id": "",
"name": "Perplexity: Sonar",
"created": 1738013808,
@@ -7205,6 +7474,7 @@
},
{
"id": "liquid/lfm-7b",
+ "canonical_slug": "liquid/lfm-7b",
"hugging_face_id": "",
"name": "Liquid: LFM 7B",
"created": 1737806883,
@@ -7254,6 +7524,7 @@
},
{
"id": "liquid/lfm-3b",
+ "canonical_slug": "liquid/lfm-3b",
"hugging_face_id": "",
"name": "Liquid: LFM 3B",
"created": 1737806501,
@@ -7299,6 +7570,7 @@
},
{
"id": "deepseek/deepseek-r1-distill-llama-70b:free",
+ "canonical_slug": "deepseek/deepseek-r1-distill-llama-70b",
"hugging_face_id": "deepseek-ai/DeepSeek-R1-Distill-Llama-70B",
"name": "DeepSeek: R1 Distill Llama 70B (free)",
"created": 1737663169,
@@ -7350,6 +7622,7 @@
},
{
"id": "deepseek/deepseek-r1-distill-llama-70b",
+ "canonical_slug": "deepseek/deepseek-r1-distill-llama-70b",
"hugging_face_id": "deepseek-ai/DeepSeek-R1-Distill-Llama-70B",
"name": "DeepSeek: R1 Distill Llama 70B",
"created": 1737663169,
@@ -7404,6 +7677,7 @@
},
{
"id": "deepseek/deepseek-r1:free",
+ "canonical_slug": "deepseek/deepseek-r1",
"hugging_face_id": "deepseek-ai/DeepSeek-R1",
"name": "DeepSeek: R1 (free)",
"created": 1737381095,
@@ -7451,6 +7725,7 @@
},
{
"id": "deepseek/deepseek-r1",
+ "canonical_slug": "deepseek/deepseek-r1",
"hugging_face_id": "deepseek-ai/DeepSeek-R1",
"name": "DeepSeek: R1",
"created": 1737381095,
@@ -7505,6 +7780,7 @@
},
{
"id": "minimax/minimax-01",
+ "canonical_slug": "minimax/minimax-01",
"hugging_face_id": "MiniMaxAI/MiniMax-Text-01",
"name": "MiniMax: MiniMax-01",
"created": 1736915462,
@@ -7544,6 +7820,7 @@
},
{
"id": "mistralai/codestral-2501",
+ "canonical_slug": "mistralai/codestral-2501",
"hugging_face_id": "",
"name": "Mistral: Codestral 2501",
"created": 1736895522,
@@ -7590,6 +7867,7 @@
},
{
"id": "microsoft/phi-4",
+ "canonical_slug": "microsoft/phi-4",
"hugging_face_id": "microsoft/phi-4",
"name": "Microsoft: Phi 4",
"created": 1736489872,
@@ -7639,6 +7917,7 @@
},
{
"id": "deepseek/deepseek-chat:free",
+ "canonical_slug": "deepseek/deepseek-chat-v3",
"hugging_face_id": "deepseek-ai/DeepSeek-V3",
"name": "DeepSeek: DeepSeek V3 (free)",
"created": 1735241320,
@@ -7688,6 +7967,7 @@
},
{
"id": "deepseek/deepseek-chat",
+ "canonical_slug": "deepseek/deepseek-chat-v3",
"hugging_face_id": "deepseek-ai/DeepSeek-V3",
"name": "DeepSeek: DeepSeek V3",
"created": 1735241320,
@@ -7740,6 +8020,7 @@
},
{
"id": "sao10k/l3.3-euryale-70b",
+ "canonical_slug": "sao10k/l3.3-euryale-70b-v2.3",
"hugging_face_id": "Sao10K/L3.3-70B-Euryale-v2.3",
"name": "Sao10K: Llama 3.3 Euryale 70B",
"created": 1734535928,
@@ -7757,8 +8038,8 @@
"instruct_type": "llama3"
},
"pricing": {
- "prompt": "0.0000007",
- "completion": "0.0000008",
+ "prompt": "0.00000065",
+ "completion": "0.00000075",
"request": "0",
"image": "0",
"web_search": "0",
@@ -7766,7 +8047,7 @@
},
"top_provider": {
"context_length": 131072,
- "max_completion_tokens": 16384,
+ "max_completion_tokens": null,
"is_moderated": false
},
"per_request_limits": null,
@@ -7777,8 +8058,9 @@
"stop",
"frequency_penalty",
"presence_penalty",
- "repetition_penalty",
"response_format",
+ "structured_outputs",
+ "repetition_penalty",
"top_k",
"seed",
"min_p",
@@ -7787,6 +8069,7 @@
},
{
"id": "openai/o1",
+ "canonical_slug": "openai/o1-2024-12-17",
"hugging_face_id": "",
"name": "OpenAI: o1",
"created": 1734459999,
@@ -7830,6 +8113,7 @@
},
{
"id": "eva-unit-01/eva-llama-3.33-70b",
+ "canonical_slug": "eva-unit-01/eva-llama-3.33-70b",
"hugging_face_id": "EVA-UNIT-01/EVA-LLaMA-3.33-70B-v0.1",
"name": "EVA Llama 3.33 70B",
"created": 1734377303,
@@ -7875,6 +8159,7 @@
},
{
"id": "x-ai/grok-2-vision-1212",
+ "canonical_slug": "x-ai/grok-2-vision-1212",
"hugging_face_id": "",
"name": "xAI: Grok 2 Vision 1212",
"created": 1734237338,
@@ -7921,6 +8206,7 @@
},
{
"id": "x-ai/grok-2-1212",
+ "canonical_slug": "x-ai/grok-2-1212",
"hugging_face_id": "",
"name": "xAI: Grok 2 1212",
"created": 1734232814,
@@ -7968,6 +8254,7 @@
},
{
"id": "cohere/command-r7b-12-2024",
+ "canonical_slug": "cohere/command-r7b-12-2024",
"hugging_face_id": "",
"name": "Cohere: Command R7B (12-2024)",
"created": 1734158152,
@@ -8013,6 +8300,7 @@
},
{
"id": "google/gemini-2.0-flash-exp:free",
+ "canonical_slug": "google/gemini-2.0-flash-exp",
"hugging_face_id": "",
"name": "Google: Gemini 2.0 Flash Experimental (free)",
"created": 1733937523,
@@ -8053,6 +8341,7 @@
},
{
"id": "meta-llama/llama-3.3-70b-instruct:free",
+ "canonical_slug": "meta-llama/llama-3.3-70b-instruct",
"hugging_face_id": "meta-llama/Llama-3.3-70B-Instruct",
"name": "Meta: Llama 3.3 70B Instruct (free)",
"created": 1733506137,
@@ -8084,16 +8373,14 @@
},
"per_request_limits": null,
"supported_parameters": [
- "tools",
- "tool_choice",
"max_tokens",
"temperature",
"top_p",
- "repetition_penalty",
- "top_k",
"stop",
"frequency_penalty",
"presence_penalty",
+ "top_k",
+ "repetition_penalty",
"logit_bias",
"min_p",
"response_format"
@@ -8101,6 +8388,7 @@
},
{
"id": "meta-llama/llama-3.3-70b-instruct",
+ "canonical_slug": "meta-llama/llama-3.3-70b-instruct",
"hugging_face_id": "meta-llama/Llama-3.3-70B-Instruct",
"name": "Meta: Llama 3.3 70B Instruct",
"created": 1733506137,
@@ -8118,8 +8406,8 @@
"instruct_type": "llama3"
},
"pricing": {
- "prompt": "0.00000007",
- "completion": "0.00000025",
+ "prompt": "0.000000039",
+ "completion": "0.00000012",
"request": "0",
"image": "0",
"web_search": "0",
@@ -8127,7 +8415,7 @@
},
"top_provider": {
"context_length": 131072,
- "max_completion_tokens": 16384,
+ "max_completion_tokens": 8192,
"is_moderated": false
},
"per_request_limits": null,
@@ -8153,6 +8441,7 @@
},
{
"id": "amazon/nova-lite-v1",
+ "canonical_slug": "amazon/nova-lite-v1",
"hugging_face_id": "",
"name": "Amazon: Nova Lite 1.0",
"created": 1733437363,
@@ -8195,6 +8484,7 @@
},
{
"id": "amazon/nova-micro-v1",
+ "canonical_slug": "amazon/nova-micro-v1",
"hugging_face_id": "",
"name": "Amazon: Nova Micro 1.0",
"created": 1733437237,
@@ -8236,6 +8526,7 @@
},
{
"id": "amazon/nova-pro-v1",
+ "canonical_slug": "amazon/nova-pro-v1",
"hugging_face_id": "",
"name": "Amazon: Nova Pro 1.0",
"created": 1733436303,
@@ -8278,6 +8569,7 @@
},
{
"id": "qwen/qwq-32b-preview",
+ "canonical_slug": "qwen/qwq-32b-preview",
"hugging_face_id": "Qwen/QwQ-32B-Preview",
"name": "Qwen: QwQ 32B Preview",
"created": 1732754541,
@@ -8326,6 +8618,7 @@
},
{
"id": "eva-unit-01/eva-qwen-2.5-72b",
+ "canonical_slug": "eva-unit-01/eva-qwen-2.5-72b",
"hugging_face_id": "EVA-UNIT-01/EVA-Qwen2.5-72B-v0.1",
"name": "EVA Qwen2.5 72B",
"created": 1732210606,
@@ -8371,6 +8664,7 @@
},
{
"id": "openai/gpt-4o-2024-11-20",
+ "canonical_slug": "openai/gpt-4o-2024-11-20",
"hugging_face_id": "",
"name": "OpenAI: GPT-4o (2024-11-20)",
"created": 1732127594,
@@ -8424,6 +8718,7 @@
},
{
"id": "mistralai/mistral-large-2411",
+ "canonical_slug": "mistralai/mistral-large-2411",
"hugging_face_id": "",
"name": "Mistral Large 2411",
"created": 1731978685,
@@ -8470,6 +8765,7 @@
},
{
"id": "mistralai/mistral-large-2407",
+ "canonical_slug": "mistralai/mistral-large-2407",
"hugging_face_id": "",
"name": "Mistral Large 2407",
"created": 1731978415,
@@ -8516,11 +8812,12 @@
},
{
"id": "mistralai/pixtral-large-2411",
+ "canonical_slug": "mistralai/pixtral-large-2411",
"hugging_face_id": "",
"name": "Mistral: Pixtral Large 2411",
"created": 1731977388,
"description": "Pixtral Large is a 124B parameter, open-weight, multimodal model built on top of [Mistral Large 2](/mistralai/mistral-large-2411). The model is able to understand documents, charts and natural images.\n\nThe model is available under the Mistral Research License (MRL) for research and educational use, and the Mistral Commercial License for experimentation, testing, and production for commercial purposes.\n\n",
- "context_length": 131072,
+ "context_length": 32768,
"architecture": {
"modality": "text+image->text",
"input_modalities": [
@@ -8542,7 +8839,7 @@
"internal_reasoning": "0"
},
"top_provider": {
- "context_length": 131072,
+ "context_length": 32768,
"max_completion_tokens": null,
"is_moderated": false
},
@@ -8563,6 +8860,7 @@
},
{
"id": "x-ai/grok-vision-beta",
+ "canonical_slug": "x-ai/grok-vision-beta",
"hugging_face_id": "",
"name": "xAI: Grok Vision Beta",
"created": 1731976624,
@@ -8609,6 +8907,7 @@
},
{
"id": "infermatic/mn-inferor-12b",
+ "canonical_slug": "infermatic/mn-inferor-12b",
"hugging_face_id": "Infermatic/MN-12B-Inferor-v0.0",
"name": "Infermatic: Mistral Nemo Inferor 12B",
"created": 1731464428,
@@ -8654,6 +8953,7 @@
},
{
"id": "qwen/qwen-2.5-coder-32b-instruct:free",
+ "canonical_slug": "qwen/qwen-2.5-coder-32b-instruct",
"hugging_face_id": "Qwen/Qwen2.5-Coder-32B-Instruct",
"name": "Qwen2.5 Coder 32B Instruct (free)",
"created": 1731368400,
@@ -8702,6 +9002,7 @@
},
{
"id": "qwen/qwen-2.5-coder-32b-instruct",
+ "canonical_slug": "qwen/qwen-2.5-coder-32b-instruct",
"hugging_face_id": "Qwen/Qwen2.5-Coder-32B-Instruct",
"name": "Qwen2.5 Coder 32B Instruct",
"created": 1731368400,
@@ -8751,6 +9052,7 @@
},
{
"id": "raifle/sorcererlm-8x22b",
+ "canonical_slug": "raifle/sorcererlm-8x22b",
"hugging_face_id": "rAIfle/SorcererLM-8x22b-bf16",
"name": "SorcererLM 8x22B",
"created": 1731105083,
@@ -8797,6 +9099,7 @@
},
{
"id": "eva-unit-01/eva-qwen-2.5-32b",
+ "canonical_slug": "eva-unit-01/eva-qwen-2.5-32b",
"hugging_face_id": "EVA-UNIT-01/EVA-Qwen2.5-32B-v0.2",
"name": "EVA Qwen2.5 32B",
"created": 1731104847,
@@ -8842,11 +9145,12 @@
},
{
"id": "thedrummer/unslopnemo-12b",
+ "canonical_slug": "thedrummer/unslopnemo-12b",
"hugging_face_id": "TheDrummer/UnslopNemo-12B-v4.1",
"name": "TheDrummer: UnslopNemo 12B",
"created": 1731103448,
"description": "UnslopNemo v4.1 is the latest addition from the creator of Rocinante, designed for adventure writing and role-play scenarios.",
- "context_length": 32000,
+ "context_length": 32768,
"architecture": {
"modality": "text->text",
"input_modalities": [
@@ -8859,16 +9163,16 @@
"instruct_type": "mistral"
},
"pricing": {
- "prompt": "0.00000045",
- "completion": "0.00000045",
+ "prompt": "0.0000004",
+ "completion": "0.0000004",
"request": "0",
"image": "0",
"web_search": "0",
"internal_reasoning": "0"
},
"top_provider": {
- "context_length": 32000,
- "max_completion_tokens": 16000,
+ "context_length": 32768,
+ "max_completion_tokens": null,
"is_moderated": false
},
"per_request_limits": null,
@@ -8879,16 +9183,21 @@
"stop",
"frequency_penalty",
"presence_penalty",
+ "response_format",
+ "structured_outputs",
"repetition_penalty",
"logit_bias",
"top_k",
"min_p",
"seed",
+ "tools",
+ "tool_choice",
"logprobs"
]
},
{
"id": "anthropic/claude-3.5-haiku:beta",
+ "canonical_slug": "anthropic/claude-3-5-haiku",
"hugging_face_id": null,
"name": "Anthropic: Claude 3.5 Haiku (self-moderated)",
"created": 1730678400,
@@ -8934,6 +9243,7 @@
},
{
"id": "anthropic/claude-3.5-haiku",
+ "canonical_slug": "anthropic/claude-3-5-haiku",
"hugging_face_id": null,
"name": "Anthropic: Claude 3.5 Haiku",
"created": 1730678400,
@@ -8979,6 +9289,7 @@
},
{
"id": "anthropic/claude-3.5-haiku-20241022:beta",
+ "canonical_slug": "anthropic/claude-3-5-haiku-20241022",
"hugging_face_id": null,
"name": "Anthropic: Claude 3.5 Haiku (2024-10-22) (self-moderated)",
"created": 1730678400,
@@ -9024,6 +9335,7 @@
},
{
"id": "anthropic/claude-3.5-haiku-20241022",
+ "canonical_slug": "anthropic/claude-3-5-haiku-20241022",
"hugging_face_id": null,
"name": "Anthropic: Claude 3.5 Haiku (2024-10-22)",
"created": 1730678400,
@@ -9068,54 +9380,100 @@
]
},
{
- "id": "neversleep/llama-3.1-lumimaid-70b",
- "hugging_face_id": "NeverSleep/Lumimaid-v0.2-70B",
- "name": "NeverSleep: Lumimaid v0.2 70B",
+ "id": "anthropic/claude-3.5-sonnet:beta",
+ "canonical_slug": "anthropic/claude-3.5-sonnet",
+ "hugging_face_id": null,
+ "name": "Anthropic: Claude 3.5 Sonnet (self-moderated)",
"created": 1729555200,
- "description": "Lumimaid v0.2 70B is a finetune of [Llama 3.1 70B](/meta-llama/llama-3.1-70b-instruct) with a \"HUGE step up dataset wise\" compared to Lumimaid v0.1. Sloppy chats output were purged.\n\nUsage of this model is subject to [Meta's Acceptable Use Policy](https://llama.meta.com/llama3/use-policy/).",
- "context_length": 16384,
+ "description": "New Claude 3.5 Sonnet delivers better-than-Opus capabilities, faster-than-Sonnet speeds, at the same Sonnet prices. Sonnet is particularly good at:\n\n- Coding: Scores ~49% on SWE-Bench Verified, higher than the last best score, and without any fancy prompt scaffolding\n- Data science: Augments human data science expertise; navigates unstructured data while using multiple tools for insights\n- Visual processing: excelling at interpreting charts, graphs, and images, accurately transcribing text to derive insights beyond just the text alone\n- Agentic tasks: exceptional tool use, making it great at agentic tasks (i.e. complex, multi-step problem solving tasks that require engaging with other systems)\n\n#multimodal",
+ "context_length": 200000,
"architecture": {
- "modality": "text->text",
+ "modality": "text+image->text",
"input_modalities": [
- "text"
+ "text",
+ "image"
],
"output_modalities": [
"text"
],
- "tokenizer": "Llama3",
- "instruct_type": "llama3"
+ "tokenizer": "Claude",
+ "instruct_type": null
},
"pricing": {
- "prompt": "0.0000025",
- "completion": "0.000003",
+ "prompt": "0.000003",
+ "completion": "0.000015",
"request": "0",
- "image": "0",
+ "image": "0.0048",
"web_search": "0",
- "internal_reasoning": "0"
+ "internal_reasoning": "0",
+ "input_cache_read": "0.0000003",
+ "input_cache_write": "0.00000375"
},
"top_provider": {
- "context_length": 16384,
- "max_completion_tokens": 2048,
+ "context_length": 200000,
+ "max_completion_tokens": 8192,
"is_moderated": false
},
"per_request_limits": null,
"supported_parameters": [
+ "tools",
+ "tool_choice",
"max_tokens",
"temperature",
"top_p",
- "stop",
- "frequency_penalty",
- "presence_penalty",
- "repetition_penalty",
- "logit_bias",
"top_k",
- "min_p",
- "seed",
- "top_a"
+ "stop"
+ ]
+ },
+ {
+ "id": "anthropic/claude-3.5-sonnet",
+ "canonical_slug": "anthropic/claude-3.5-sonnet",
+ "hugging_face_id": null,
+ "name": "Anthropic: Claude 3.5 Sonnet",
+ "created": 1729555200,
+ "description": "New Claude 3.5 Sonnet delivers better-than-Opus capabilities, faster-than-Sonnet speeds, at the same Sonnet prices. Sonnet is particularly good at:\n\n- Coding: Scores ~49% on SWE-Bench Verified, higher than the last best score, and without any fancy prompt scaffolding\n- Data science: Augments human data science expertise; navigates unstructured data while using multiple tools for insights\n- Visual processing: excelling at interpreting charts, graphs, and images, accurately transcribing text to derive insights beyond just the text alone\n- Agentic tasks: exceptional tool use, making it great at agentic tasks (i.e. complex, multi-step problem solving tasks that require engaging with other systems)\n\n#multimodal",
+ "context_length": 200000,
+ "architecture": {
+ "modality": "text+image->text",
+ "input_modalities": [
+ "text",
+ "image"
+ ],
+ "output_modalities": [
+ "text"
+ ],
+ "tokenizer": "Claude",
+ "instruct_type": null
+ },
+ "pricing": {
+ "prompt": "0.000003",
+ "completion": "0.000015",
+ "request": "0",
+ "image": "0.0048",
+ "web_search": "0",
+ "internal_reasoning": "0",
+ "input_cache_read": "0.0000003",
+ "input_cache_write": "0.00000375"
+ },
+ "top_provider": {
+ "context_length": 200000,
+ "max_completion_tokens": 8192,
+ "is_moderated": true
+ },
+ "per_request_limits": null,
+ "supported_parameters": [
+ "tools",
+ "tool_choice",
+ "max_tokens",
+ "temperature",
+ "top_p",
+ "top_k",
+ "stop"
]
},
{
"id": "anthracite-org/magnum-v4-72b",
+ "canonical_slug": "anthracite-org/magnum-v4-72b",
"hugging_face_id": "anthracite-org/magnum-v4-72b",
"name": "Magnum v4 72B",
"created": 1729555200,
@@ -9162,97 +9520,56 @@
]
},
{
- "id": "anthropic/claude-3.5-sonnet:beta",
- "hugging_face_id": null,
- "name": "Anthropic: Claude 3.5 Sonnet (self-moderated)",
+ "id": "neversleep/llama-3.1-lumimaid-70b",
+ "canonical_slug": "neversleep/llama-3.1-lumimaid-70b",
+ "hugging_face_id": "NeverSleep/Lumimaid-v0.2-70B",
+ "name": "NeverSleep: Lumimaid v0.2 70B",
"created": 1729555200,
- "description": "New Claude 3.5 Sonnet delivers better-than-Opus capabilities, faster-than-Sonnet speeds, at the same Sonnet prices. Sonnet is particularly good at:\n\n- Coding: Scores ~49% on SWE-Bench Verified, higher than the last best score, and without any fancy prompt scaffolding\n- Data science: Augments human data science expertise; navigates unstructured data while using multiple tools for insights\n- Visual processing: excelling at interpreting charts, graphs, and images, accurately transcribing text to derive insights beyond just the text alone\n- Agentic tasks: exceptional tool use, making it great at agentic tasks (i.e. complex, multi-step problem solving tasks that require engaging with other systems)\n\n#multimodal",
- "context_length": 200000,
+ "description": "Lumimaid v0.2 70B is a finetune of [Llama 3.1 70B](/meta-llama/llama-3.1-70b-instruct) with a \"HUGE step up dataset wise\" compared to Lumimaid v0.1. Sloppy chats output were purged.\n\nUsage of this model is subject to [Meta's Acceptable Use Policy](https://llama.meta.com/llama3/use-policy/).",
+ "context_length": 16384,
"architecture": {
- "modality": "text+image->text",
+ "modality": "text->text",
"input_modalities": [
- "text",
- "image"
+ "text"
],
"output_modalities": [
"text"
],
- "tokenizer": "Claude",
- "instruct_type": null
+ "tokenizer": "Llama3",
+ "instruct_type": "llama3"
},
"pricing": {
- "prompt": "0.000003",
- "completion": "0.000015",
+ "prompt": "0.0000025",
+ "completion": "0.000003",
"request": "0",
- "image": "0.0048",
+ "image": "0",
"web_search": "0",
- "internal_reasoning": "0",
- "input_cache_read": "0.0000003",
- "input_cache_write": "0.00000375"
+ "internal_reasoning": "0"
},
"top_provider": {
- "context_length": 200000,
- "max_completion_tokens": 8192,
+ "context_length": 16384,
+ "max_completion_tokens": 2048,
"is_moderated": false
},
"per_request_limits": null,
"supported_parameters": [
- "tools",
- "tool_choice",
"max_tokens",
"temperature",
"top_p",
+ "stop",
+ "frequency_penalty",
+ "presence_penalty",
+ "repetition_penalty",
+ "logit_bias",
"top_k",
- "stop"
- ]
- },
- {
- "id": "anthropic/claude-3.5-sonnet",
- "hugging_face_id": null,
- "name": "Anthropic: Claude 3.5 Sonnet",
- "created": 1729555200,
- "description": "New Claude 3.5 Sonnet delivers better-than-Opus capabilities, faster-than-Sonnet speeds, at the same Sonnet prices. Sonnet is particularly good at:\n\n- Coding: Scores ~49% on SWE-Bench Verified, higher than the last best score, and without any fancy prompt scaffolding\n- Data science: Augments human data science expertise; navigates unstructured data while using multiple tools for insights\n- Visual processing: excelling at interpreting charts, graphs, and images, accurately transcribing text to derive insights beyond just the text alone\n- Agentic tasks: exceptional tool use, making it great at agentic tasks (i.e. complex, multi-step problem solving tasks that require engaging with other systems)\n\n#multimodal",
- "context_length": 200000,
- "architecture": {
- "modality": "text+image->text",
- "input_modalities": [
- "text",
- "image"
- ],
- "output_modalities": [
- "text"
- ],
- "tokenizer": "Claude",
- "instruct_type": null
- },
- "pricing": {
- "prompt": "0.000003",
- "completion": "0.000015",
- "request": "0",
- "image": "0.0048",
- "web_search": "0",
- "internal_reasoning": "0",
- "input_cache_read": "0.0000003",
- "input_cache_write": "0.00000375"
- },
- "top_provider": {
- "context_length": 200000,
- "max_completion_tokens": 8192,
- "is_moderated": true
- },
- "per_request_limits": null,
- "supported_parameters": [
- "tools",
- "tool_choice",
- "max_tokens",
- "temperature",
- "top_p",
- "top_k",
- "stop"
+ "min_p",
+ "seed",
+ "top_a"
]
},
{
"id": "x-ai/grok-beta",
+ "canonical_slug": "x-ai/grok-beta",
"hugging_face_id": null,
"name": "xAI: Grok Beta",
"created": 1729382400,
@@ -9300,6 +9617,7 @@
},
{
"id": "mistralai/ministral-8b",
+ "canonical_slug": "mistralai/ministral-8b",
"hugging_face_id": null,
"name": "Mistral: Ministral 8B",
"created": 1729123200,
@@ -9346,6 +9664,7 @@
},
{
"id": "mistralai/ministral-3b",
+ "canonical_slug": "mistralai/ministral-3b",
"hugging_face_id": null,
"name": "Mistral: Ministral 3B",
"created": 1729123200,
@@ -9390,46 +9709,9 @@
"seed"
]
},
- {
- "id": "qwen/qwen-2.5-7b-instruct:free",
- "hugging_face_id": "Qwen/Qwen2.5-7B-Instruct",
- "name": "Qwen2.5 7B Instruct (free)",
- "created": 1729036800,
- "description": "Qwen2.5 7B is the latest series of Qwen large language models. Qwen2.5 brings the following improvements upon Qwen2:\n\n- Significantly more knowledge and has greatly improved capabilities in coding and mathematics, thanks to our specialized expert models in these domains.\n\n- Significant improvements in instruction following, generating long texts (over 8K tokens), understanding structured data (e.g, tables), and generating structured outputs especially JSON. More resilient to the diversity of system prompts, enhancing role-play implementation and condition-setting for chatbots.\n\n- Long-context Support up to 128K tokens and can generate up to 8K tokens.\n\n- Multilingual support for over 29 languages, including Chinese, English, French, Spanish, Portuguese, German, Italian, Russian, Japanese, Korean, Vietnamese, Thai, Arabic, and more.\n\nUsage of this model is subject to [Tongyi Qianwen LICENSE AGREEMENT](https://huggingface.co/Qwen/Qwen1.5-110B-Chat/blob/main/LICENSE).",
- "context_length": 32768,
- "architecture": {
- "modality": "text->text",
- "input_modalities": [
- "text"
- ],
- "output_modalities": [
- "text"
- ],
- "tokenizer": "Qwen",
- "instruct_type": "chatml"
- },
- "pricing": {
- "prompt": "0",
- "completion": "0",
- "request": "0",
- "image": "0",
- "web_search": "0",
- "internal_reasoning": "0"
- },
- "top_provider": {
- "context_length": 32768,
- "max_completion_tokens": 32768,
- "is_moderated": false
- },
- "per_request_limits": null,
- "supported_parameters": [
- "max_tokens",
- "temperature",
- "top_p"
- ]
- },
{
"id": "qwen/qwen-2.5-7b-instruct",
+ "canonical_slug": "qwen/qwen-2.5-7b-instruct",
"hugging_face_id": "Qwen/Qwen2.5-7B-Instruct",
"name": "Qwen2.5 7B Instruct",
"created": 1729036800,
@@ -9456,7 +9738,7 @@
},
"top_provider": {
"context_length": 32768,
- "max_completion_tokens": null,
+ "max_completion_tokens": 16384,
"is_moderated": false
},
"per_request_limits": null,
@@ -9472,12 +9754,12 @@
"logit_bias",
"min_p",
"response_format",
- "structured_outputs",
"seed"
]
},
{
"id": "nvidia/llama-3.1-nemotron-70b-instruct",
+ "canonical_slug": "nvidia/llama-3.1-nemotron-70b-instruct",
"hugging_face_id": "nvidia/Llama-3.1-Nemotron-70B-Instruct-HF",
"name": "NVIDIA: Llama 3.1 Nemotron 70B Instruct",
"created": 1728950400,
@@ -9522,52 +9804,14 @@
"logprobs",
"top_logprobs",
"response_format",
- "top_k",
+ "min_p",
"repetition_penalty",
- "min_p"
- ]
- },
- {
- "id": "inflection/inflection-3-productivity",
- "hugging_face_id": null,
- "name": "Inflection: Inflection 3 Productivity",
- "created": 1728604800,
- "description": "Inflection 3 Productivity is optimized for following instructions. It is better for tasks requiring JSON output or precise adherence to provided guidelines. It has access to recent news.\n\nFor emotional intelligence similar to Pi, see [Inflect 3 Pi](/inflection/inflection-3-pi)\n\nSee [Inflection's announcement](https://inflection.ai/blog/enterprise) for more details.",
- "context_length": 8000,
- "architecture": {
- "modality": "text->text",
- "input_modalities": [
- "text"
- ],
- "output_modalities": [
- "text"
- ],
- "tokenizer": "Other",
- "instruct_type": null
- },
- "pricing": {
- "prompt": "0.0000025",
- "completion": "0.00001",
- "request": "0",
- "image": "0",
- "web_search": "0",
- "internal_reasoning": "0"
- },
- "top_provider": {
- "context_length": 8000,
- "max_completion_tokens": 1024,
- "is_moderated": false
- },
- "per_request_limits": null,
- "supported_parameters": [
- "max_tokens",
- "temperature",
- "top_p",
- "stop"
+ "top_k"
]
},
{
"id": "inflection/inflection-3-pi",
+ "canonical_slug": "inflection/inflection-3-pi",
"hugging_face_id": null,
"name": "Inflection: Inflection 3 Pi",
"created": 1728604800,
@@ -9605,8 +9849,49 @@
"stop"
]
},
+ {
+ "id": "inflection/inflection-3-productivity",
+ "canonical_slug": "inflection/inflection-3-productivity",
+ "hugging_face_id": null,
+ "name": "Inflection: Inflection 3 Productivity",
+ "created": 1728604800,
+ "description": "Inflection 3 Productivity is optimized for following instructions. It is better for tasks requiring JSON output or precise adherence to provided guidelines. It has access to recent news.\n\nFor emotional intelligence similar to Pi, see [Inflect 3 Pi](/inflection/inflection-3-pi)\n\nSee [Inflection's announcement](https://inflection.ai/blog/enterprise) for more details.",
+ "context_length": 8000,
+ "architecture": {
+ "modality": "text->text",
+ "input_modalities": [
+ "text"
+ ],
+ "output_modalities": [
+ "text"
+ ],
+ "tokenizer": "Other",
+ "instruct_type": null
+ },
+ "pricing": {
+ "prompt": "0.0000025",
+ "completion": "0.00001",
+ "request": "0",
+ "image": "0",
+ "web_search": "0",
+ "internal_reasoning": "0"
+ },
+ "top_provider": {
+ "context_length": 8000,
+ "max_completion_tokens": 1024,
+ "is_moderated": false
+ },
+ "per_request_limits": null,
+ "supported_parameters": [
+ "max_tokens",
+ "temperature",
+ "top_p",
+ "stop"
+ ]
+ },
{
"id": "google/gemini-flash-1.5-8b",
+ "canonical_slug": "google/gemini-flash-1.5-8b",
"hugging_face_id": null,
"name": "Google: Gemini 1.5 Flash 8B",
"created": 1727913600,
@@ -9654,100 +9939,9 @@
"structured_outputs"
]
},
- {
- "id": "thedrummer/rocinante-12b",
- "hugging_face_id": "TheDrummer/Rocinante-12B-v1.1",
- "name": "TheDrummer: Rocinante 12B",
- "created": 1727654400,
- "description": "Rocinante 12B is designed for engaging storytelling and rich prose.\n\nEarly testers have reported:\n- Expanded vocabulary with unique and expressive word choices\n- Enhanced creativity for vivid narratives\n- Adventure-filled and captivating stories",
- "context_length": 32768,
- "architecture": {
- "modality": "text->text",
- "input_modalities": [
- "text"
- ],
- "output_modalities": [
- "text"
- ],
- "tokenizer": "Qwen",
- "instruct_type": "chatml"
- },
- "pricing": {
- "prompt": "0.00000025",
- "completion": "0.0000005",
- "request": "0",
- "image": "0",
- "web_search": "0",
- "internal_reasoning": "0"
- },
- "top_provider": {
- "context_length": 32768,
- "max_completion_tokens": null,
- "is_moderated": false
- },
- "per_request_limits": null,
- "supported_parameters": [
- "max_tokens",
- "temperature",
- "top_p",
- "stop",
- "frequency_penalty",
- "presence_penalty",
- "repetition_penalty",
- "top_k",
- "min_p",
- "seed",
- "logit_bias"
- ]
- },
- {
- "id": "anthracite-org/magnum-v2-72b",
- "hugging_face_id": "anthracite-org/magnum-v2-72b",
- "name": "Magnum v2 72B",
- "created": 1727654400,
- "description": "From the maker of [Goliath](https://openrouter.ai/models/alpindale/goliath-120b), Magnum 72B is the seventh in a family of models designed to achieve the prose quality of the Claude 3 models, notably Opus & Sonnet.\n\nThe model is based on [Qwen2 72B](https://openrouter.ai/models/qwen/qwen-2-72b-instruct) and trained with 55 million tokens of highly curated roleplay (RP) data.",
- "context_length": 32768,
- "architecture": {
- "modality": "text->text",
- "input_modalities": [
- "text"
- ],
- "output_modalities": [
- "text"
- ],
- "tokenizer": "Qwen",
- "instruct_type": "chatml"
- },
- "pricing": {
- "prompt": "0.000003",
- "completion": "0.000003",
- "request": "0",
- "image": "0",
- "web_search": "0",
- "internal_reasoning": "0"
- },
- "top_provider": {
- "context_length": 32768,
- "max_completion_tokens": null,
- "is_moderated": false
- },
- "per_request_limits": null,
- "supported_parameters": [
- "max_tokens",
- "temperature",
- "top_p",
- "stop",
- "frequency_penalty",
- "presence_penalty",
- "repetition_penalty",
- "logit_bias",
- "top_k",
- "min_p",
- "seed"
- ]
- },
{
"id": "liquid/lfm-40b",
+ "canonical_slug": "liquid/lfm-40b",
"hugging_face_id": null,
"name": "Liquid: LFM 40B MoE",
"created": 1727654400,
@@ -9796,9 +9990,106 @@
]
},
{
- "id": "meta-llama/llama-3.2-3b-instruct:free",
+ "id": "anthracite-org/magnum-v2-72b",
+ "canonical_slug": "anthracite-org/magnum-v2-72b",
+ "hugging_face_id": "anthracite-org/magnum-v2-72b",
+ "name": "Magnum v2 72B",
+ "created": 1727654400,
+ "description": "From the maker of [Goliath](https://openrouter.ai/models/alpindale/goliath-120b), Magnum 72B is the seventh in a family of models designed to achieve the prose quality of the Claude 3 models, notably Opus & Sonnet.\n\nThe model is based on [Qwen2 72B](https://openrouter.ai/models/qwen/qwen-2-72b-instruct) and trained with 55 million tokens of highly curated roleplay (RP) data.",
+ "context_length": 32768,
+ "architecture": {
+ "modality": "text->text",
+ "input_modalities": [
+ "text"
+ ],
+ "output_modalities": [
+ "text"
+ ],
+ "tokenizer": "Qwen",
+ "instruct_type": "chatml"
+ },
+ "pricing": {
+ "prompt": "0.000003",
+ "completion": "0.000003",
+ "request": "0",
+ "image": "0",
+ "web_search": "0",
+ "internal_reasoning": "0"
+ },
+ "top_provider": {
+ "context_length": 32768,
+ "max_completion_tokens": null,
+ "is_moderated": false
+ },
+ "per_request_limits": null,
+ "supported_parameters": [
+ "max_tokens",
+ "temperature",
+ "top_p",
+ "stop",
+ "frequency_penalty",
+ "presence_penalty",
+ "repetition_penalty",
+ "logit_bias",
+ "top_k",
+ "min_p",
+ "seed"
+ ]
+ },
+ {
+ "id": "thedrummer/rocinante-12b",
+ "canonical_slug": "thedrummer/rocinante-12b",
+ "hugging_face_id": "TheDrummer/Rocinante-12B-v1.1",
+ "name": "TheDrummer: Rocinante 12B",
+ "created": 1727654400,
+ "description": "Rocinante 12B is designed for engaging storytelling and rich prose.\n\nEarly testers have reported:\n- Expanded vocabulary with unique and expressive word choices\n- Enhanced creativity for vivid narratives\n- Adventure-filled and captivating stories",
+ "context_length": 32768,
+ "architecture": {
+ "modality": "text->text",
+ "input_modalities": [
+ "text"
+ ],
+ "output_modalities": [
+ "text"
+ ],
+ "tokenizer": "Qwen",
+ "instruct_type": "chatml"
+ },
+ "pricing": {
+ "prompt": "0.0000002",
+ "completion": "0.0000005",
+ "request": "0",
+ "image": "0",
+ "web_search": "0",
+ "internal_reasoning": "0"
+ },
+ "top_provider": {
+ "context_length": 32768,
+ "max_completion_tokens": null,
+ "is_moderated": false
+ },
+ "per_request_limits": null,
+ "supported_parameters": [
+ "max_tokens",
+ "temperature",
+ "top_p",
+ "stop",
+ "frequency_penalty",
+ "presence_penalty",
+ "response_format",
+ "structured_outputs",
+ "repetition_penalty",
+ "top_k",
+ "min_p",
+ "seed",
+ "logit_bias"
+ ]
+ },
+ {
+ "id": "meta-llama/llama-3.2-3b-instruct",
+ "canonical_slug": "meta-llama/llama-3.2-3b-instruct",
"hugging_face_id": "meta-llama/Llama-3.2-3B-Instruct",
- "name": "Meta: Llama 3.2 3B Instruct (free)",
+ "name": "Meta: Llama 3.2 3B Instruct",
"created": 1727222400,
"description": "Llama 3.2 3B is a 3-billion-parameter multilingual large language model, optimized for advanced natural language processing tasks like dialogue generation, reasoning, and summarization. Designed with the latest transformer architecture, it supports eight languages, including English, Spanish, and Hindi, and is adaptable for additional languages.\n\nTrained on 9 trillion tokens, the Llama 3.2 3B model excels in instruction-following, complex reasoning, and tool use. Its balanced performance makes it ideal for applications needing accuracy and efficiency in text generation across multilingual settings.\n\nClick here for the [original model card](https://github.com/meta-llama/llama-models/blob/main/models/llama3_2/MODEL_CARD.md).\n\nUsage of this model is subject to [Meta's Acceptable Use Policy](https://www.llama.com/llama3/use-policy/).",
"context_length": 20000,
@@ -9814,8 +10105,8 @@
"instruct_type": "llama3"
},
"pricing": {
- "prompt": "0",
- "completion": "0",
+ "prompt": "0.000000003",
+ "completion": "0.000000006",
"request": "0",
"image": "0",
"web_search": "0",
@@ -9827,44 +10118,6 @@
"is_moderated": false
},
"per_request_limits": null,
- "supported_parameters": [
- "max_tokens",
- "temperature",
- "top_p"
- ]
- },
- {
- "id": "meta-llama/llama-3.2-3b-instruct",
- "hugging_face_id": "meta-llama/Llama-3.2-3B-Instruct",
- "name": "Meta: Llama 3.2 3B Instruct",
- "created": 1727222400,
- "description": "Llama 3.2 3B is a 3-billion-parameter multilingual large language model, optimized for advanced natural language processing tasks like dialogue generation, reasoning, and summarization. Designed with the latest transformer architecture, it supports eight languages, including English, Spanish, and Hindi, and is adaptable for additional languages.\n\nTrained on 9 trillion tokens, the Llama 3.2 3B model excels in instruction-following, complex reasoning, and tool use. Its balanced performance makes it ideal for applications needing accuracy and efficiency in text generation across multilingual settings.\n\nClick here for the [original model card](https://github.com/meta-llama/llama-models/blob/main/models/llama3_2/MODEL_CARD.md).\n\nUsage of this model is subject to [Meta's Acceptable Use Policy](https://www.llama.com/llama3/use-policy/).",
- "context_length": 131072,
- "architecture": {
- "modality": "text->text",
- "input_modalities": [
- "text"
- ],
- "output_modalities": [
- "text"
- ],
- "tokenizer": "Llama3",
- "instruct_type": "llama3"
- },
- "pricing": {
- "prompt": "0.00000001",
- "completion": "0.00000002",
- "request": "0",
- "image": "0",
- "web_search": "0",
- "internal_reasoning": "0"
- },
- "top_provider": {
- "context_length": 131072,
- "max_completion_tokens": 16384,
- "is_moderated": false
- },
- "per_request_limits": null,
"supported_parameters": [
"tools",
"tool_choice",
@@ -9879,103 +10132,14 @@
"logprobs",
"top_logprobs",
"response_format",
- "top_k",
"min_p",
- "repetition_penalty"
- ]
- },
- {
- "id": "meta-llama/llama-3.2-1b-instruct:free",
- "hugging_face_id": "meta-llama/Llama-3.2-1B-Instruct",
- "name": "Meta: Llama 3.2 1B Instruct (free)",
- "created": 1727222400,
- "description": "Llama 3.2 1B is a 1-billion-parameter language model focused on efficiently performing natural language tasks, such as summarization, dialogue, and multilingual text analysis. Its smaller size allows it to operate efficiently in low-resource environments while maintaining strong task performance.\n\nSupporting eight core languages and fine-tunable for more, Llama 1.3B is ideal for businesses or developers seeking lightweight yet powerful AI solutions that can operate in diverse multilingual settings without the high computational demand of larger models.\n\nClick here for the [original model card](https://github.com/meta-llama/llama-models/blob/main/models/llama3_2/MODEL_CARD.md).\n\nUsage of this model is subject to [Meta's Acceptable Use Policy](https://www.llama.com/llama3/use-policy/).",
- "context_length": 131072,
- "architecture": {
- "modality": "text->text",
- "input_modalities": [
- "text"
- ],
- "output_modalities": [
- "text"
- ],
- "tokenizer": "Llama3",
- "instruct_type": "llama3"
- },
- "pricing": {
- "prompt": "0",
- "completion": "0",
- "request": "0",
- "image": "0",
- "web_search": "0",
- "internal_reasoning": "0"
- },
- "top_provider": {
- "context_length": 131072,
- "max_completion_tokens": 8192,
- "is_moderated": false
- },
- "per_request_limits": null,
- "supported_parameters": [
- "max_tokens",
- "temperature",
- "top_p",
- "stop",
- "frequency_penalty",
- "presence_penalty",
- "seed"
- ]
- },
- {
- "id": "meta-llama/llama-3.2-1b-instruct",
- "hugging_face_id": "meta-llama/Llama-3.2-1B-Instruct",
- "name": "Meta: Llama 3.2 1B Instruct",
- "created": 1727222400,
- "description": "Llama 3.2 1B is a 1-billion-parameter language model focused on efficiently performing natural language tasks, such as summarization, dialogue, and multilingual text analysis. Its smaller size allows it to operate efficiently in low-resource environments while maintaining strong task performance.\n\nSupporting eight core languages and fine-tunable for more, Llama 1.3B is ideal for businesses or developers seeking lightweight yet powerful AI solutions that can operate in diverse multilingual settings without the high computational demand of larger models.\n\nClick here for the [original model card](https://github.com/meta-llama/llama-models/blob/main/models/llama3_2/MODEL_CARD.md).\n\nUsage of this model is subject to [Meta's Acceptable Use Policy](https://www.llama.com/llama3/use-policy/).",
- "context_length": 131072,
- "architecture": {
- "modality": "text->text",
- "input_modalities": [
- "text"
- ],
- "output_modalities": [
- "text"
- ],
- "tokenizer": "Llama3",
- "instruct_type": "llama3"
- },
- "pricing": {
- "prompt": "0.000000005",
- "completion": "0.00000001",
- "request": "0",
- "image": "0",
- "web_search": "0",
- "internal_reasoning": "0"
- },
- "top_provider": {
- "context_length": 131072,
- "max_completion_tokens": 16384,
- "is_moderated": false
- },
- "per_request_limits": null,
- "supported_parameters": [
- "max_tokens",
- "temperature",
- "top_p",
- "top_k",
- "stop",
- "frequency_penalty",
- "presence_penalty",
"repetition_penalty",
- "response_format",
- "seed",
- "min_p",
- "logit_bias",
- "top_logprobs"
+ "top_k"
]
},
{
"id": "meta-llama/llama-3.2-90b-vision-instruct",
+ "canonical_slug": "meta-llama/llama-3.2-90b-vision-instruct",
"hugging_face_id": "meta-llama/Llama-3.2-90B-Vision-Instruct",
"name": "Meta: Llama 3.2 90B Vision Instruct",
"created": 1727222400,
@@ -10024,6 +10188,7 @@
},
{
"id": "meta-llama/llama-3.2-11b-vision-instruct:free",
+ "canonical_slug": "meta-llama/llama-3.2-11b-vision-instruct",
"hugging_face_id": "meta-llama/Llama-3.2-11B-Vision-Instruct",
"name": "Meta: Llama 3.2 11B Vision Instruct (free)",
"created": 1727222400,
@@ -10071,6 +10236,7 @@
},
{
"id": "meta-llama/llama-3.2-11b-vision-instruct",
+ "canonical_slug": "meta-llama/llama-3.2-11b-vision-instruct",
"hugging_face_id": "meta-llama/Llama-3.2-11B-Vision-Instruct",
"name": "Meta: Llama 3.2 11B Vision Instruct",
"created": 1727222400,
@@ -10121,8 +10287,107 @@
"logprobs"
]
},
+ {
+ "id": "meta-llama/llama-3.2-1b-instruct:free",
+ "canonical_slug": "meta-llama/llama-3.2-1b-instruct",
+ "hugging_face_id": "meta-llama/Llama-3.2-1B-Instruct",
+ "name": "Meta: Llama 3.2 1B Instruct (free)",
+ "created": 1727222400,
+ "description": "Llama 3.2 1B is a 1-billion-parameter language model focused on efficiently performing natural language tasks, such as summarization, dialogue, and multilingual text analysis. Its smaller size allows it to operate efficiently in low-resource environments while maintaining strong task performance.\n\nSupporting eight core languages and fine-tunable for more, Llama 1.3B is ideal for businesses or developers seeking lightweight yet powerful AI solutions that can operate in diverse multilingual settings without the high computational demand of larger models.\n\nClick here for the [original model card](https://github.com/meta-llama/llama-models/blob/main/models/llama3_2/MODEL_CARD.md).\n\nUsage of this model is subject to [Meta's Acceptable Use Policy](https://www.llama.com/llama3/use-policy/).",
+ "context_length": 131072,
+ "architecture": {
+ "modality": "text->text",
+ "input_modalities": [
+ "text"
+ ],
+ "output_modalities": [
+ "text"
+ ],
+ "tokenizer": "Llama3",
+ "instruct_type": "llama3"
+ },
+ "pricing": {
+ "prompt": "0",
+ "completion": "0",
+ "request": "0",
+ "image": "0",
+ "web_search": "0",
+ "internal_reasoning": "0"
+ },
+ "top_provider": {
+ "context_length": 131072,
+ "max_completion_tokens": 8192,
+ "is_moderated": false
+ },
+ "per_request_limits": null,
+ "supported_parameters": [
+ "max_tokens",
+ "temperature",
+ "top_p",
+ "stop",
+ "frequency_penalty",
+ "presence_penalty",
+ "seed",
+ "logit_bias",
+ "logprobs",
+ "top_logprobs",
+ "min_p",
+ "repetition_penalty",
+ "top_k"
+ ]
+ },
+ {
+ "id": "meta-llama/llama-3.2-1b-instruct",
+ "canonical_slug": "meta-llama/llama-3.2-1b-instruct",
+ "hugging_face_id": "meta-llama/Llama-3.2-1B-Instruct",
+ "name": "Meta: Llama 3.2 1B Instruct",
+ "created": 1727222400,
+ "description": "Llama 3.2 1B is a 1-billion-parameter language model focused on efficiently performing natural language tasks, such as summarization, dialogue, and multilingual text analysis. Its smaller size allows it to operate efficiently in low-resource environments while maintaining strong task performance.\n\nSupporting eight core languages and fine-tunable for more, Llama 1.3B is ideal for businesses or developers seeking lightweight yet powerful AI solutions that can operate in diverse multilingual settings without the high computational demand of larger models.\n\nClick here for the [original model card](https://github.com/meta-llama/llama-models/blob/main/models/llama3_2/MODEL_CARD.md).\n\nUsage of this model is subject to [Meta's Acceptable Use Policy](https://www.llama.com/llama3/use-policy/).",
+ "context_length": 131072,
+ "architecture": {
+ "modality": "text->text",
+ "input_modalities": [
+ "text"
+ ],
+ "output_modalities": [
+ "text"
+ ],
+ "tokenizer": "Llama3",
+ "instruct_type": "llama3"
+ },
+ "pricing": {
+ "prompt": "0.000000005",
+ "completion": "0.00000001",
+ "request": "0",
+ "image": "0",
+ "web_search": "0",
+ "internal_reasoning": "0"
+ },
+ "top_provider": {
+ "context_length": 131072,
+ "max_completion_tokens": 16384,
+ "is_moderated": false
+ },
+ "per_request_limits": null,
+ "supported_parameters": [
+ "max_tokens",
+ "temperature",
+ "top_p",
+ "stop",
+ "frequency_penalty",
+ "presence_penalty",
+ "repetition_penalty",
+ "response_format",
+ "top_k",
+ "seed",
+ "min_p",
+ "logit_bias",
+ "top_logprobs"
+ ]
+ },
{
"id": "qwen/qwen-2.5-72b-instruct:free",
+ "canonical_slug": "qwen/qwen-2.5-72b-instruct",
"hugging_face_id": "Qwen/Qwen2.5-72B-Instruct",
"name": "Qwen2.5 72B Instruct (free)",
"created": 1726704000,
@@ -10171,6 +10436,7 @@
},
{
"id": "qwen/qwen-2.5-72b-instruct",
+ "canonical_slug": "qwen/qwen-2.5-72b-instruct",
"hugging_face_id": "Qwen/Qwen2.5-72B-Instruct",
"name": "Qwen2.5 72B Instruct",
"created": 1726704000,
@@ -10223,6 +10489,7 @@
},
{
"id": "neversleep/llama-3.1-lumimaid-8b",
+ "canonical_slug": "neversleep/llama-3.1-lumimaid-8b",
"hugging_face_id": "NeverSleep/Lumimaid-v0.2-8B",
"name": "NeverSleep: Lumimaid v0.2 8B",
"created": 1726358400,
@@ -10270,6 +10537,7 @@
},
{
"id": "openai/o1-preview",
+ "canonical_slug": "openai/o1-preview",
"hugging_face_id": null,
"name": "OpenAI: o1-preview",
"created": 1726099200,
@@ -10306,8 +10574,48 @@
"max_tokens"
]
},
+ {
+ "id": "openai/o1-mini-2024-09-12",
+ "canonical_slug": "openai/o1-mini-2024-09-12",
+ "hugging_face_id": null,
+ "name": "OpenAI: o1-mini (2024-09-12)",
+ "created": 1726099200,
+ "description": "The latest and strongest model family from OpenAI, o1 is designed to spend more time thinking before responding.\n\nThe o1 models are optimized for math, science, programming, and other STEM-related tasks. They consistently exhibit PhD-level accuracy on benchmarks in physics, chemistry, and biology. Learn more in the [launch announcement](https://openai.com/o1).\n\nNote: This model is currently experimental and not suitable for production use-cases, and may be heavily rate-limited.",
+ "context_length": 128000,
+ "architecture": {
+ "modality": "text->text",
+ "input_modalities": [
+ "text"
+ ],
+ "output_modalities": [
+ "text"
+ ],
+ "tokenizer": "GPT",
+ "instruct_type": null
+ },
+ "pricing": {
+ "prompt": "0.0000011",
+ "completion": "0.0000044",
+ "request": "0",
+ "image": "0",
+ "web_search": "0",
+ "internal_reasoning": "0",
+ "input_cache_read": "0.00000055"
+ },
+ "top_provider": {
+ "context_length": 128000,
+ "max_completion_tokens": 65536,
+ "is_moderated": true
+ },
+ "per_request_limits": null,
+ "supported_parameters": [
+ "seed",
+ "max_tokens"
+ ]
+ },
{
"id": "openai/o1-preview-2024-09-12",
+ "canonical_slug": "openai/o1-preview-2024-09-12",
"hugging_face_id": null,
"name": "OpenAI: o1-preview (2024-09-12)",
"created": 1726099200,
@@ -10346,6 +10654,7 @@
},
{
"id": "openai/o1-mini",
+ "canonical_slug": "openai/o1-mini",
"hugging_face_id": null,
"name": "OpenAI: o1-mini",
"created": 1726099200,
@@ -10382,46 +10691,9 @@
"max_tokens"
]
},
- {
- "id": "openai/o1-mini-2024-09-12",
- "hugging_face_id": null,
- "name": "OpenAI: o1-mini (2024-09-12)",
- "created": 1726099200,
- "description": "The latest and strongest model family from OpenAI, o1 is designed to spend more time thinking before responding.\n\nThe o1 models are optimized for math, science, programming, and other STEM-related tasks. They consistently exhibit PhD-level accuracy on benchmarks in physics, chemistry, and biology. Learn more in the [launch announcement](https://openai.com/o1).\n\nNote: This model is currently experimental and not suitable for production use-cases, and may be heavily rate-limited.",
- "context_length": 128000,
- "architecture": {
- "modality": "text->text",
- "input_modalities": [
- "text"
- ],
- "output_modalities": [
- "text"
- ],
- "tokenizer": "GPT",
- "instruct_type": null
- },
- "pricing": {
- "prompt": "0.0000011",
- "completion": "0.0000044",
- "request": "0",
- "image": "0",
- "web_search": "0",
- "internal_reasoning": "0",
- "input_cache_read": "0.00000055"
- },
- "top_provider": {
- "context_length": 128000,
- "max_completion_tokens": 65536,
- "is_moderated": true
- },
- "per_request_limits": null,
- "supported_parameters": [
- "seed",
- "max_tokens"
- ]
- },
{
"id": "mistralai/pixtral-12b",
+ "canonical_slug": "mistralai/pixtral-12b",
"hugging_face_id": "mistralai/Pixtral-12B-2409",
"name": "Mistral: Pixtral 12B",
"created": 1725926400,
@@ -10475,6 +10747,7 @@
},
{
"id": "cohere/command-r-plus-08-2024",
+ "canonical_slug": "cohere/command-r-plus-08-2024",
"hugging_face_id": null,
"name": "Cohere: Command R+ (08-2024)",
"created": 1724976000,
@@ -10521,6 +10794,7 @@
},
{
"id": "cohere/command-r-08-2024",
+ "canonical_slug": "cohere/command-r-08-2024",
"hugging_face_id": null,
"name": "Cohere: Command R (08-2024)",
"created": 1724976000,
@@ -10565,57 +10839,9 @@
"structured_outputs"
]
},
- {
- "id": "qwen/qwen-2.5-vl-7b-instruct:free",
- "hugging_face_id": "Qwen/Qwen2.5-VL-7B-Instruct",
- "name": "Qwen: Qwen2.5-VL 7B Instruct (free)",
- "created": 1724803200,
- "description": "Qwen2.5 VL 7B is a multimodal LLM from the Qwen Team with the following key enhancements:\n\n- SoTA understanding of images of various resolution & ratio: Qwen2.5-VL achieves state-of-the-art performance on visual understanding benchmarks, including MathVista, DocVQA, RealWorldQA, MTVQA, etc.\n\n- Understanding videos of 20min+: Qwen2.5-VL can understand videos over 20 minutes for high-quality video-based question answering, dialog, content creation, etc.\n\n- Agent that can operate your mobiles, robots, etc.: with the abilities of complex reasoning and decision making, Qwen2.5-VL can be integrated with devices like mobile phones, robots, etc., for automatic operation based on visual environment and text instructions.\n\n- Multilingual Support: to serve global users, besides English and Chinese, Qwen2.5-VL now supports the understanding of texts in different languages inside images, including most European languages, Japanese, Korean, Arabic, Vietnamese, etc.\n\nFor more details, see this [blog post](https://qwenlm.github.io/blog/qwen2-vl/) and [GitHub repo](https://github.com/QwenLM/Qwen2-VL).\n\nUsage of this model is subject to [Tongyi Qianwen LICENSE AGREEMENT](https://huggingface.co/Qwen/Qwen1.5-110B-Chat/blob/main/LICENSE).",
- "context_length": 32768,
- "architecture": {
- "modality": "text+image->text",
- "input_modalities": [
- "text",
- "image"
- ],
- "output_modalities": [
- "text"
- ],
- "tokenizer": "Qwen",
- "instruct_type": null
- },
- "pricing": {
- "prompt": "0",
- "completion": "0",
- "request": "0",
- "image": "0",
- "web_search": "0",
- "internal_reasoning": "0"
- },
- "top_provider": {
- "context_length": 32768,
- "max_completion_tokens": 32768,
- "is_moderated": false
- },
- "per_request_limits": null,
- "supported_parameters": [
- "max_tokens",
- "temperature",
- "top_p",
- "stop",
- "frequency_penalty",
- "presence_penalty",
- "seed",
- "top_k",
- "min_p",
- "repetition_penalty",
- "logprobs",
- "logit_bias",
- "top_logprobs"
- ]
- },
{
"id": "qwen/qwen-2.5-vl-7b-instruct",
+ "canonical_slug": "qwen/qwen-2-vl-7b-instruct",
"hugging_face_id": "Qwen/Qwen2.5-VL-7B-Instruct",
"name": "Qwen: Qwen2.5-VL 7B Instruct",
"created": 1724803200,
@@ -10665,11 +10891,12 @@
},
{
"id": "sao10k/l3.1-euryale-70b",
+ "canonical_slug": "sao10k/l3.1-euryale-70b",
"hugging_face_id": "Sao10K/L3.1-70B-Euryale-v2.2",
"name": "Sao10K: Llama 3.1 Euryale 70B v2.2",
"created": 1724803200,
"description": "Euryale L3.1 70B v2.2 is a model focused on creative roleplay from [Sao10k](https://ko-fi.com/sao10k). It is the successor of [Euryale L3 70B v2.1](/models/sao10k/l3-euryale-70b).",
- "context_length": 131072,
+ "context_length": 32768,
"architecture": {
"modality": "text->text",
"input_modalities": [
@@ -10682,16 +10909,16 @@
"instruct_type": "llama3"
},
"pricing": {
- "prompt": "0.0000007",
- "completion": "0.0000008",
+ "prompt": "0.00000065",
+ "completion": "0.00000075",
"request": "0",
"image": "0",
"web_search": "0",
"internal_reasoning": "0"
},
"top_provider": {
- "context_length": 131072,
- "max_completion_tokens": 16384,
+ "context_length": 32768,
+ "max_completion_tokens": null,
"is_moderated": false
},
"per_request_limits": null,
@@ -10702,16 +10929,18 @@
"stop",
"frequency_penalty",
"presence_penalty",
+ "response_format",
+ "structured_outputs",
"seed",
"top_k",
"min_p",
"repetition_penalty",
- "logit_bias",
- "response_format"
+ "logit_bias"
]
},
{
"id": "microsoft/phi-3.5-mini-128k-instruct",
+ "canonical_slug": "microsoft/phi-3.5-mini-128k-instruct",
"hugging_face_id": "microsoft/Phi-3.5-mini-instruct",
"name": "Microsoft: Phi-3.5 Mini 128K Instruct",
"created": 1724198400,
@@ -10752,6 +10981,7 @@
},
{
"id": "nousresearch/hermes-3-llama-3.1-70b",
+ "canonical_slug": "nousresearch/hermes-3-llama-3.1-70b",
"hugging_face_id": "NousResearch/Hermes-3-Llama-3.1-70B",
"name": "Nous: Hermes 3 70B Instruct",
"created": 1723939200,
@@ -10794,15 +11024,16 @@
"logprobs",
"top_logprobs",
"response_format",
- "tools",
- "tool_choice",
- "top_k",
"min_p",
- "repetition_penalty"
+ "repetition_penalty",
+ "top_k",
+ "tools",
+ "tool_choice"
]
},
{
"id": "nousresearch/hermes-3-llama-3.1-405b",
+ "canonical_slug": "nousresearch/hermes-3-llama-3.1-405b",
"hugging_face_id": "NousResearch/Hermes-3-Llama-3.1-405B",
"name": "Nous: Hermes 3 405B Instruct",
"created": 1723766400,
@@ -10845,13 +11076,14 @@
"logprobs",
"top_logprobs",
"response_format",
- "top_k",
+ "min_p",
"repetition_penalty",
- "min_p"
+ "top_k"
]
},
{
"id": "openai/chatgpt-4o-latest",
+ "canonical_slug": "openai/chatgpt-4o-latest",
"hugging_face_id": null,
"name": "OpenAI: ChatGPT-4o",
"created": 1723593600,
@@ -10898,8 +11130,55 @@
"structured_outputs"
]
},
+ {
+ "id": "aetherwiing/mn-starcannon-12b",
+ "canonical_slug": "aetherwiing/mn-starcannon-12b",
+ "hugging_face_id": "aetherwiing/MN-12B-Starcannon-v2",
+ "name": "Aetherwiing: Starcannon 12B",
+ "created": 1723507200,
+ "description": "Starcannon 12B v2 is a creative roleplay and story writing model, based on Mistral Nemo, using [nothingiisreal/mn-celeste-12b](/nothingiisreal/mn-celeste-12b) as a base, with [intervitens/mini-magnum-12b-v1.1](https://huggingface.co/intervitens/mini-magnum-12b-v1.1) merged in using the [TIES](https://arxiv.org/abs/2306.01708) method.\n\nAlthough more similar to Magnum overall, the model remains very creative, with a pleasant writing style. It is recommended for people wanting more variety than Magnum, and yet more verbose prose than Celeste.",
+ "context_length": 16384,
+ "architecture": {
+ "modality": "text->text",
+ "input_modalities": [
+ "text"
+ ],
+ "output_modalities": [
+ "text"
+ ],
+ "tokenizer": "Mistral",
+ "instruct_type": "chatml"
+ },
+ "pricing": {
+ "prompt": "0.0000008",
+ "completion": "0.0000012",
+ "request": "0",
+ "image": "0",
+ "web_search": "0",
+ "internal_reasoning": "0"
+ },
+ "top_provider": {
+ "context_length": 16384,
+ "max_completion_tokens": 4096,
+ "is_moderated": false
+ },
+ "per_request_limits": null,
+ "supported_parameters": [
+ "max_tokens",
+ "temperature",
+ "top_p",
+ "stop",
+ "frequency_penalty",
+ "presence_penalty",
+ "repetition_penalty",
+ "top_k",
+ "min_p",
+ "seed"
+ ]
+ },
{
"id": "sao10k/l3-lunaris-8b",
+ "canonical_slug": "sao10k/l3-lunaris-8b",
"hugging_face_id": "Sao10K/L3-8B-Lunaris-v1",
"name": "Sao10K: Llama 3 8B Lunaris",
"created": 1723507200,
@@ -10946,11 +11225,66 @@
]
},
{
- "id": "aetherwiing/mn-starcannon-12b",
- "hugging_face_id": "aetherwiing/MN-12B-Starcannon-v2",
- "name": "Aetherwiing: Starcannon 12B",
- "created": 1723507200,
- "description": "Starcannon 12B v2 is a creative roleplay and story writing model, based on Mistral Nemo, using [nothingiisreal/mn-celeste-12b](/nothingiisreal/mn-celeste-12b) as a base, with [intervitens/mini-magnum-12b-v1.1](https://huggingface.co/intervitens/mini-magnum-12b-v1.1) merged in using the [TIES](https://arxiv.org/abs/2306.01708) method.\n\nAlthough more similar to Magnum overall, the model remains very creative, with a pleasant writing style. It is recommended for people wanting more variety than Magnum, and yet more verbose prose than Celeste.",
+ "id": "openai/gpt-4o-2024-08-06",
+ "canonical_slug": "openai/gpt-4o-2024-08-06",
+ "hugging_face_id": null,
+ "name": "OpenAI: GPT-4o (2024-08-06)",
+ "created": 1722902400,
+ "description": "The 2024-08-06 version of GPT-4o offers improved performance in structured outputs, with the ability to supply a JSON schema in the respone_format. Read more [here](https://openai.com/index/introducing-structured-outputs-in-the-api/).\n\nGPT-4o (\"o\" for \"omni\") is OpenAI's latest AI model, supporting both text and image inputs with text outputs. It maintains the intelligence level of [GPT-4 Turbo](/models/openai/gpt-4-turbo) while being twice as fast and 50% more cost-effective. GPT-4o also offers improved performance in processing non-English languages and enhanced visual capabilities.\n\nFor benchmarking against other models, it was briefly called [\"im-also-a-good-gpt2-chatbot\"](https://twitter.com/LiamFedus/status/1790064963966370209)",
+ "context_length": 128000,
+ "architecture": {
+ "modality": "text+image->text",
+ "input_modalities": [
+ "text",
+ "image",
+ "file"
+ ],
+ "output_modalities": [
+ "text"
+ ],
+ "tokenizer": "GPT",
+ "instruct_type": null
+ },
+ "pricing": {
+ "prompt": "0.0000025",
+ "completion": "0.00001",
+ "request": "0",
+ "image": "0.003613",
+ "web_search": "0",
+ "internal_reasoning": "0",
+ "input_cache_read": "0.00000125"
+ },
+ "top_provider": {
+ "context_length": 128000,
+ "max_completion_tokens": 16384,
+ "is_moderated": false
+ },
+ "per_request_limits": null,
+ "supported_parameters": [
+ "tools",
+ "tool_choice",
+ "max_tokens",
+ "temperature",
+ "top_p",
+ "stop",
+ "frequency_penalty",
+ "presence_penalty",
+ "web_search_options",
+ "seed",
+ "logit_bias",
+ "logprobs",
+ "top_logprobs",
+ "response_format",
+ "structured_outputs"
+ ]
+ },
+ {
+ "id": "nothingiisreal/mn-celeste-12b",
+ "canonical_slug": "nothingiisreal/mn-celeste-12b",
+ "hugging_face_id": "nothingiisreal/MN-12B-Celeste-V1.9",
+ "name": "Mistral Nemo 12B Celeste",
+ "created": 1722556800,
+ "description": "A specialized story writing and roleplaying model based on Mistral's NeMo 12B Instruct. Fine-tuned on curated datasets including Reddit Writing Prompts and Opus Instruct 25K.\n\nThis model excels at creative writing, offering improved NSFW capabilities, with smarter and more active narration. It demonstrates remarkable versatility in both SFW and NSFW scenarios, with strong Out of Character (OOC) steering capabilities, allowing fine-tuned control over narrative direction and character behavior.\n\nCheck out the model's [HuggingFace page](https://huggingface.co/nothingiisreal/MN-12B-Celeste-V1.9) for details on what parameters and prompts work best!",
"context_length": 16384,
"architecture": {
"modality": "text->text",
@@ -10990,108 +11324,9 @@
"seed"
]
},
- {
- "id": "openai/gpt-4o-2024-08-06",
- "hugging_face_id": null,
- "name": "OpenAI: GPT-4o (2024-08-06)",
- "created": 1722902400,
- "description": "The 2024-08-06 version of GPT-4o offers improved performance in structured outputs, with the ability to supply a JSON schema in the respone_format. Read more [here](https://openai.com/index/introducing-structured-outputs-in-the-api/).\n\nGPT-4o (\"o\" for \"omni\") is OpenAI's latest AI model, supporting both text and image inputs with text outputs. It maintains the intelligence level of [GPT-4 Turbo](/models/openai/gpt-4-turbo) while being twice as fast and 50% more cost-effective. GPT-4o also offers improved performance in processing non-English languages and enhanced visual capabilities.\n\nFor benchmarking against other models, it was briefly called [\"im-also-a-good-gpt2-chatbot\"](https://twitter.com/LiamFedus/status/1790064963966370209)",
- "context_length": 128000,
- "architecture": {
- "modality": "text+image->text",
- "input_modalities": [
- "text",
- "image",
- "file"
- ],
- "output_modalities": [
- "text"
- ],
- "tokenizer": "GPT",
- "instruct_type": null
- },
- "pricing": {
- "prompt": "0.0000025",
- "completion": "0.00001",
- "request": "0",
- "image": "0.003613",
- "web_search": "0",
- "internal_reasoning": "0"
- },
- "top_provider": {
- "context_length": 128000,
- "max_completion_tokens": 16384,
- "is_moderated": false
- },
- "per_request_limits": null,
- "supported_parameters": [
- "tools",
- "tool_choice",
- "max_tokens",
- "temperature",
- "top_p",
- "stop",
- "frequency_penalty",
- "presence_penalty",
- "web_search_options",
- "seed",
- "logit_bias",
- "logprobs",
- "top_logprobs",
- "response_format",
- "structured_outputs"
- ]
- },
- {
- "id": "meta-llama/llama-3.1-405b:free",
- "hugging_face_id": "meta-llama/llama-3.1-405B",
- "name": "Meta: Llama 3.1 405B (base) (free)",
- "created": 1722556800,
- "description": "Meta's latest class of model (Llama 3.1) launched with a variety of sizes & flavors. This is the base 405B pre-trained version.\n\nIt has demonstrated strong performance compared to leading closed-source models in human evaluations.\n\nTo read more about the model release, [click here](https://ai.meta.com/blog/meta-llama-3/). Usage of this model is subject to [Meta's Acceptable Use Policy](https://llama.meta.com/llama3/use-policy/).",
- "context_length": 64000,
- "architecture": {
- "modality": "text->text",
- "input_modalities": [
- "text"
- ],
- "output_modalities": [
- "text"
- ],
- "tokenizer": "Llama3",
- "instruct_type": "none"
- },
- "pricing": {
- "prompt": "0",
- "completion": "0",
- "request": "0",
- "image": "0",
- "web_search": "0",
- "internal_reasoning": "0"
- },
- "top_provider": {
- "context_length": 64000,
- "max_completion_tokens": null,
- "is_moderated": false
- },
- "per_request_limits": null,
- "supported_parameters": [
- "max_tokens",
- "temperature",
- "top_p",
- "stop",
- "frequency_penalty",
- "presence_penalty",
- "seed",
- "top_k",
- "min_p",
- "repetition_penalty",
- "logprobs",
- "logit_bias",
- "top_logprobs"
- ]
- },
{
"id": "meta-llama/llama-3.1-405b",
+ "canonical_slug": "meta-llama/llama-3.1-405b",
"hugging_face_id": "meta-llama/llama-3.1-405B",
"name": "Meta: Llama 3.1 405B (base)",
"created": 1722556800,
@@ -11138,53 +11373,9 @@
"repetition_penalty"
]
},
- {
- "id": "nothingiisreal/mn-celeste-12b",
- "hugging_face_id": "nothingiisreal/MN-12B-Celeste-V1.9",
- "name": "Mistral Nemo 12B Celeste",
- "created": 1722556800,
- "description": "A specialized story writing and roleplaying model based on Mistral's NeMo 12B Instruct. Fine-tuned on curated datasets including Reddit Writing Prompts and Opus Instruct 25K.\n\nThis model excels at creative writing, offering improved NSFW capabilities, with smarter and more active narration. It demonstrates remarkable versatility in both SFW and NSFW scenarios, with strong Out of Character (OOC) steering capabilities, allowing fine-tuned control over narrative direction and character behavior.\n\nCheck out the model's [HuggingFace page](https://huggingface.co/nothingiisreal/MN-12B-Celeste-V1.9) for details on what parameters and prompts work best!",
- "context_length": 16384,
- "architecture": {
- "modality": "text->text",
- "input_modalities": [
- "text"
- ],
- "output_modalities": [
- "text"
- ],
- "tokenizer": "Mistral",
- "instruct_type": "chatml"
- },
- "pricing": {
- "prompt": "0.0000008",
- "completion": "0.0000012",
- "request": "0",
- "image": "0",
- "web_search": "0",
- "internal_reasoning": "0"
- },
- "top_provider": {
- "context_length": 16384,
- "max_completion_tokens": 4096,
- "is_moderated": false
- },
- "per_request_limits": null,
- "supported_parameters": [
- "max_tokens",
- "temperature",
- "top_p",
- "stop",
- "frequency_penalty",
- "presence_penalty",
- "repetition_penalty",
- "top_k",
- "min_p",
- "seed"
- ]
- },
{
"id": "perplexity/llama-3.1-sonar-small-128k-online",
+ "canonical_slug": "perplexity/llama-3.1-sonar-small-128k-online",
"hugging_face_id": null,
"name": "Perplexity: Llama 3.1 Sonar 8B Online",
"created": 1722470400,
@@ -11226,6 +11417,7 @@
},
{
"id": "perplexity/llama-3.1-sonar-large-128k-online",
+ "canonical_slug": "perplexity/llama-3.1-sonar-large-128k-online",
"hugging_face_id": null,
"name": "Perplexity: Llama 3.1 Sonar 70B Online",
"created": 1722470400,
@@ -11265,102 +11457,9 @@
"presence_penalty"
]
},
- {
- "id": "meta-llama/llama-3.1-8b-instruct:free",
- "hugging_face_id": "meta-llama/Meta-Llama-3.1-8B-Instruct",
- "name": "Meta: Llama 3.1 8B Instruct (free)",
- "created": 1721692800,
- "description": "Meta's latest class of model (Llama 3.1) launched with a variety of sizes & flavors. This 8B instruct-tuned version is fast and efficient.\n\nIt has demonstrated strong performance compared to leading closed-source models in human evaluations.\n\nTo read more about the model release, [click here](https://ai.meta.com/blog/meta-llama-3-1/). Usage of this model is subject to [Meta's Acceptable Use Policy](https://llama.meta.com/llama3/use-policy/).",
- "context_length": 131072,
- "architecture": {
- "modality": "text->text",
- "input_modalities": [
- "text"
- ],
- "output_modalities": [
- "text"
- ],
- "tokenizer": "Llama3",
- "instruct_type": "llama3"
- },
- "pricing": {
- "prompt": "0",
- "completion": "0",
- "request": "0",
- "image": "0",
- "web_search": "0",
- "internal_reasoning": "0"
- },
- "top_provider": {
- "context_length": 131072,
- "max_completion_tokens": 4096,
- "is_moderated": false
- },
- "per_request_limits": null,
- "supported_parameters": [
- "max_tokens",
- "temperature",
- "top_p",
- "stop",
- "frequency_penalty",
- "presence_penalty",
- "seed"
- ]
- },
- {
- "id": "meta-llama/llama-3.1-8b-instruct",
- "hugging_face_id": "meta-llama/Meta-Llama-3.1-8B-Instruct",
- "name": "Meta: Llama 3.1 8B Instruct",
- "created": 1721692800,
- "description": "Meta's latest class of model (Llama 3.1) launched with a variety of sizes & flavors. This 8B instruct-tuned version is fast and efficient.\n\nIt has demonstrated strong performance compared to leading closed-source models in human evaluations.\n\nTo read more about the model release, [click here](https://ai.meta.com/blog/meta-llama-3-1/). Usage of this model is subject to [Meta's Acceptable Use Policy](https://llama.meta.com/llama3/use-policy/).",
- "context_length": 16384,
- "architecture": {
- "modality": "text->text",
- "input_modalities": [
- "text"
- ],
- "output_modalities": [
- "text"
- ],
- "tokenizer": "Llama3",
- "instruct_type": "llama3"
- },
- "pricing": {
- "prompt": "0.00000002",
- "completion": "0.00000003",
- "request": "0",
- "image": "0",
- "web_search": "0",
- "internal_reasoning": "0"
- },
- "top_provider": {
- "context_length": 16384,
- "max_completion_tokens": 16384,
- "is_moderated": false
- },
- "per_request_limits": null,
- "supported_parameters": [
- "tools",
- "tool_choice",
- "max_tokens",
- "temperature",
- "top_p",
- "stop",
- "frequency_penalty",
- "presence_penalty",
- "seed",
- "logit_bias",
- "logprobs",
- "top_logprobs",
- "response_format",
- "top_k",
- "repetition_penalty",
- "structured_outputs",
- "min_p"
- ]
- },
{
"id": "meta-llama/llama-3.1-405b-instruct",
+ "canonical_slug": "meta-llama/llama-3.1-405b-instruct",
"hugging_face_id": "meta-llama/Meta-Llama-3.1-405B-Instruct",
"name": "Meta: Llama 3.1 405B Instruct",
"created": 1721692800,
@@ -11413,6 +11512,7 @@
},
{
"id": "meta-llama/llama-3.1-70b-instruct",
+ "canonical_slug": "meta-llama/llama-3.1-70b-instruct",
"hugging_face_id": "meta-llama/Meta-Llama-3.1-70B-Instruct",
"name": "Meta: Llama 3.1 70B Instruct",
"created": 1721692800,
@@ -11463,8 +11563,111 @@
"structured_outputs"
]
},
+ {
+ "id": "meta-llama/llama-3.1-8b-instruct:free",
+ "canonical_slug": "meta-llama/llama-3.1-8b-instruct",
+ "hugging_face_id": "meta-llama/Meta-Llama-3.1-8B-Instruct",
+ "name": "Meta: Llama 3.1 8B Instruct (free)",
+ "created": 1721692800,
+ "description": "Meta's latest class of model (Llama 3.1) launched with a variety of sizes & flavors. This 8B instruct-tuned version is fast and efficient.\n\nIt has demonstrated strong performance compared to leading closed-source models in human evaluations.\n\nTo read more about the model release, [click here](https://ai.meta.com/blog/meta-llama-3-1/). Usage of this model is subject to [Meta's Acceptable Use Policy](https://llama.meta.com/llama3/use-policy/).",
+ "context_length": 131072,
+ "architecture": {
+ "modality": "text->text",
+ "input_modalities": [
+ "text"
+ ],
+ "output_modalities": [
+ "text"
+ ],
+ "tokenizer": "Llama3",
+ "instruct_type": "llama3"
+ },
+ "pricing": {
+ "prompt": "0",
+ "completion": "0",
+ "request": "0",
+ "image": "0",
+ "web_search": "0",
+ "internal_reasoning": "0"
+ },
+ "top_provider": {
+ "context_length": 131072,
+ "max_completion_tokens": 4096,
+ "is_moderated": false
+ },
+ "per_request_limits": null,
+ "supported_parameters": [
+ "max_tokens",
+ "temperature",
+ "top_p",
+ "stop",
+ "frequency_penalty",
+ "presence_penalty",
+ "seed",
+ "logit_bias",
+ "logprobs",
+ "top_logprobs",
+ "min_p",
+ "repetition_penalty",
+ "top_k"
+ ]
+ },
+ {
+ "id": "meta-llama/llama-3.1-8b-instruct",
+ "canonical_slug": "meta-llama/llama-3.1-8b-instruct",
+ "hugging_face_id": "meta-llama/Meta-Llama-3.1-8B-Instruct",
+ "name": "Meta: Llama 3.1 8B Instruct",
+ "created": 1721692800,
+ "description": "Meta's latest class of model (Llama 3.1) launched with a variety of sizes & flavors. This 8B instruct-tuned version is fast and efficient.\n\nIt has demonstrated strong performance compared to leading closed-source models in human evaluations.\n\nTo read more about the model release, [click here](https://ai.meta.com/blog/meta-llama-3-1/). Usage of this model is subject to [Meta's Acceptable Use Policy](https://llama.meta.com/llama3/use-policy/).",
+ "context_length": 131000,
+ "architecture": {
+ "modality": "text->text",
+ "input_modalities": [
+ "text"
+ ],
+ "output_modalities": [
+ "text"
+ ],
+ "tokenizer": "Llama3",
+ "instruct_type": "llama3"
+ },
+ "pricing": {
+ "prompt": "0.000000016",
+ "completion": "0.000000021",
+ "request": "0",
+ "image": "0",
+ "web_search": "0",
+ "internal_reasoning": "0"
+ },
+ "top_provider": {
+ "context_length": 131000,
+ "max_completion_tokens": 131000,
+ "is_moderated": false
+ },
+ "per_request_limits": null,
+ "supported_parameters": [
+ "tools",
+ "tool_choice",
+ "max_tokens",
+ "temperature",
+ "top_p",
+ "stop",
+ "frequency_penalty",
+ "presence_penalty",
+ "seed",
+ "logit_bias",
+ "logprobs",
+ "top_logprobs",
+ "response_format",
+ "min_p",
+ "repetition_penalty",
+ "top_k",
+ "structured_outputs"
+ ]
+ },
{
"id": "mistralai/mistral-nemo:free",
+ "canonical_slug": "mistralai/mistral-nemo",
"hugging_face_id": "mistralai/Mistral-Nemo-Instruct-2407",
"name": "Mistral: Mistral Nemo (free)",
"created": 1721347200,
@@ -11513,6 +11716,7 @@
},
{
"id": "mistralai/mistral-nemo",
+ "canonical_slug": "mistralai/mistral-nemo",
"hugging_face_id": "mistralai/Mistral-Nemo-Instruct-2407",
"name": "Mistral: Mistral Nemo",
"created": 1721347200,
@@ -11531,7 +11735,7 @@
},
"pricing": {
"prompt": "0.00000001",
- "completion": "0.00000005",
+ "completion": "0.000000011",
"request": "0",
"image": "0",
"web_search": "0",
@@ -11550,14 +11754,14 @@
"stop",
"frequency_penalty",
"presence_penalty",
- "repetition_penalty",
- "response_format",
"top_k",
- "seed",
- "min_p",
+ "repetition_penalty",
"logit_bias",
"logprobs",
"top_logprobs",
+ "min_p",
+ "seed",
+ "response_format",
"tools",
"tool_choice",
"structured_outputs"
@@ -11565,6 +11769,7 @@
},
{
"id": "openai/gpt-4o-mini",
+ "canonical_slug": "openai/gpt-4o-mini",
"hugging_face_id": null,
"name": "OpenAI: GPT-4o-mini",
"created": 1721260800,
@@ -11618,6 +11823,7 @@
},
{
"id": "openai/gpt-4o-mini-2024-07-18",
+ "canonical_slug": "openai/gpt-4o-mini-2024-07-18",
"hugging_face_id": null,
"name": "OpenAI: GPT-4o-mini (2024-07-18)",
"created": 1721260800,
@@ -11671,6 +11877,7 @@
},
{
"id": "google/gemma-2-27b-it",
+ "canonical_slug": "google/gemma-2-27b-it",
"hugging_face_id": "google/gemma-2-27b-it",
"name": "Google: Gemma 2 27B",
"created": 1720828800,
@@ -11717,6 +11924,7 @@
},
{
"id": "alpindale/magnum-72b",
+ "canonical_slug": "alpindale/magnum-72b",
"hugging_face_id": "alpindale/magnum-72b-v1",
"name": "Magnum 72B",
"created": 1720656000,
@@ -11762,6 +11970,7 @@
},
{
"id": "google/gemma-2-9b-it:free",
+ "canonical_slug": "google/gemma-2-9b-it",
"hugging_face_id": "google/gemma-2-9b-it",
"name": "Google: Gemma 2 9B (free)",
"created": 1719532800,
@@ -11810,6 +12019,7 @@
},
{
"id": "google/gemma-2-9b-it",
+ "canonical_slug": "google/gemma-2-9b-it",
"hugging_face_id": "google/gemma-2-9b-it",
"name": "Google: Gemma 2 9B",
"created": 1719532800,
@@ -11856,6 +12066,7 @@
},
{
"id": "01-ai/yi-large",
+ "canonical_slug": "01-ai/yi-large",
"hugging_face_id": null,
"name": "01.AI: Yi Large",
"created": 1719273600,
@@ -11904,6 +12115,7 @@
},
{
"id": "anthropic/claude-3.5-sonnet-20240620:beta",
+ "canonical_slug": "anthropic/claude-3.5-sonnet-20240620",
"hugging_face_id": null,
"name": "Anthropic: Claude 3.5 Sonnet (2024-06-20) (self-moderated)",
"created": 1718841600,
@@ -11949,6 +12161,7 @@
},
{
"id": "anthropic/claude-3.5-sonnet-20240620",
+ "canonical_slug": "anthropic/claude-3.5-sonnet-20240620",
"hugging_face_id": null,
"name": "Anthropic: Claude 3.5 Sonnet (2024-06-20)",
"created": 1718841600,
@@ -11994,6 +12207,7 @@
},
{
"id": "sao10k/l3-euryale-70b",
+ "canonical_slug": "sao10k/l3-euryale-70b",
"hugging_face_id": "Sao10K/L3-70B-Euryale-v2.1",
"name": "Sao10k: Llama 3 Euryale 70B v2.1",
"created": 1718668800,
@@ -12040,6 +12254,7 @@
},
{
"id": "cognitivecomputations/dolphin-mixtral-8x22b",
+ "canonical_slug": "cognitivecomputations/dolphin-mixtral-8x22b",
"hugging_face_id": "cognitivecomputations/dolphin-2.9.2-mixtral-8x22b",
"name": "Dolphin 2.9.2 Mixtral 8x22B 🐬",
"created": 1717804800,
@@ -12086,6 +12301,7 @@
},
{
"id": "qwen/qwen-2-72b-instruct",
+ "canonical_slug": "qwen/qwen-2-72b-instruct",
"hugging_face_id": "Qwen/Qwen2-72B-Instruct",
"name": "Qwen 2 72B Instruct",
"created": 1717718400,
@@ -12130,155 +12346,9 @@
"response_format"
]
},
- {
- "id": "mistralai/mistral-7b-instruct:free",
- "hugging_face_id": "mistralai/Mistral-7B-Instruct-v0.3",
- "name": "Mistral: Mistral 7B Instruct (free)",
- "created": 1716768000,
- "description": "A high-performing, industry-standard 7.3B parameter model, with optimizations for speed and context length.\n\n*Mistral 7B Instruct has multiple version variants, and this is intended to be the latest version.*",
- "context_length": 32768,
- "architecture": {
- "modality": "text->text",
- "input_modalities": [
- "text"
- ],
- "output_modalities": [
- "text"
- ],
- "tokenizer": "Mistral",
- "instruct_type": "mistral"
- },
- "pricing": {
- "prompt": "0",
- "completion": "0",
- "request": "0",
- "image": "0",
- "web_search": "0",
- "internal_reasoning": "0"
- },
- "top_provider": {
- "context_length": 32768,
- "max_completion_tokens": 16384,
- "is_moderated": false
- },
- "per_request_limits": null,
- "supported_parameters": [
- "tools",
- "tool_choice",
- "max_tokens",
- "temperature",
- "top_p",
- "stop",
- "frequency_penalty",
- "presence_penalty",
- "repetition_penalty",
- "response_format",
- "top_k",
- "seed",
- "min_p"
- ]
- },
- {
- "id": "mistralai/mistral-7b-instruct",
- "hugging_face_id": "mistralai/Mistral-7B-Instruct-v0.3",
- "name": "Mistral: Mistral 7B Instruct",
- "created": 1716768000,
- "description": "A high-performing, industry-standard 7.3B parameter model, with optimizations for speed and context length.\n\n*Mistral 7B Instruct has multiple version variants, and this is intended to be the latest version.*",
- "context_length": 32768,
- "architecture": {
- "modality": "text->text",
- "input_modalities": [
- "text"
- ],
- "output_modalities": [
- "text"
- ],
- "tokenizer": "Mistral",
- "instruct_type": "mistral"
- },
- "pricing": {
- "prompt": "0.000000028",
- "completion": "0.000000054",
- "request": "0",
- "image": "0",
- "web_search": "0",
- "internal_reasoning": "0"
- },
- "top_provider": {
- "context_length": 32768,
- "max_completion_tokens": 16384,
- "is_moderated": false
- },
- "per_request_limits": null,
- "supported_parameters": [
- "max_tokens",
- "temperature",
- "top_p",
- "stop",
- "frequency_penalty",
- "presence_penalty",
- "top_k",
- "repetition_penalty",
- "logit_bias",
- "min_p",
- "response_format",
- "seed",
- "logprobs",
- "tools",
- "tool_choice"
- ]
- },
- {
- "id": "nousresearch/hermes-2-pro-llama-3-8b",
- "hugging_face_id": "NousResearch/Hermes-2-Pro-Llama-3-8B",
- "name": "NousResearch: Hermes 2 Pro - Llama-3 8B",
- "created": 1716768000,
- "description": "Hermes 2 Pro is an upgraded, retrained version of Nous Hermes 2, consisting of an updated and cleaned version of the OpenHermes 2.5 Dataset, as well as a newly introduced Function Calling and JSON Mode dataset developed in-house.",
- "context_length": 131072,
- "architecture": {
- "modality": "text->text",
- "input_modalities": [
- "text"
- ],
- "output_modalities": [
- "text"
- ],
- "tokenizer": "Llama3",
- "instruct_type": "chatml"
- },
- "pricing": {
- "prompt": "0.000000025",
- "completion": "0.00000004",
- "request": "0",
- "image": "0",
- "web_search": "0",
- "internal_reasoning": "0"
- },
- "top_provider": {
- "context_length": 131072,
- "max_completion_tokens": 131072,
- "is_moderated": false
- },
- "per_request_limits": null,
- "supported_parameters": [
- "max_tokens",
- "temperature",
- "top_p",
- "stop",
- "frequency_penalty",
- "presence_penalty",
- "seed",
- "logit_bias",
- "logprobs",
- "top_logprobs",
- "response_format",
- "top_k",
- "min_p",
- "repetition_penalty"
- ]
- },
{
"id": "mistralai/mistral-7b-instruct-v0.3",
+ "canonical_slug": "mistralai/mistral-7b-instruct-v0.3",
"hugging_face_id": "mistralai/Mistral-7B-Instruct-v0.3",
"name": "Mistral: Mistral 7B Instruct v0.3",
"created": 1716768000,
@@ -12327,8 +12397,159 @@
"logprobs"
]
},
+ {
+ "id": "nousresearch/hermes-2-pro-llama-3-8b",
+ "canonical_slug": "nousresearch/hermes-2-pro-llama-3-8b",
+ "hugging_face_id": "NousResearch/Hermes-2-Pro-Llama-3-8B",
+ "name": "NousResearch: Hermes 2 Pro - Llama-3 8B",
+ "created": 1716768000,
+ "description": "Hermes 2 Pro is an upgraded, retrained version of Nous Hermes 2, consisting of an updated and cleaned version of the OpenHermes 2.5 Dataset, as well as a newly introduced Function Calling and JSON Mode dataset developed in-house.",
+ "context_length": 131072,
+ "architecture": {
+ "modality": "text->text",
+ "input_modalities": [
+ "text"
+ ],
+ "output_modalities": [
+ "text"
+ ],
+ "tokenizer": "Llama3",
+ "instruct_type": "chatml"
+ },
+ "pricing": {
+ "prompt": "0.000000025",
+ "completion": "0.00000004",
+ "request": "0",
+ "image": "0",
+ "web_search": "0",
+ "internal_reasoning": "0"
+ },
+ "top_provider": {
+ "context_length": 131072,
+ "max_completion_tokens": 131072,
+ "is_moderated": false
+ },
+ "per_request_limits": null,
+ "supported_parameters": [
+ "max_tokens",
+ "temperature",
+ "top_p",
+ "stop",
+ "frequency_penalty",
+ "presence_penalty",
+ "seed",
+ "logit_bias",
+ "logprobs",
+ "top_logprobs",
+ "response_format",
+ "min_p",
+ "repetition_penalty",
+ "top_k"
+ ]
+ },
+ {
+ "id": "mistralai/mistral-7b-instruct:free",
+ "canonical_slug": "mistralai/mistral-7b-instruct",
+ "hugging_face_id": "mistralai/Mistral-7B-Instruct-v0.3",
+ "name": "Mistral: Mistral 7B Instruct (free)",
+ "created": 1716768000,
+ "description": "A high-performing, industry-standard 7.3B parameter model, with optimizations for speed and context length.\n\n*Mistral 7B Instruct has multiple version variants, and this is intended to be the latest version.*",
+ "context_length": 32768,
+ "architecture": {
+ "modality": "text->text",
+ "input_modalities": [
+ "text"
+ ],
+ "output_modalities": [
+ "text"
+ ],
+ "tokenizer": "Mistral",
+ "instruct_type": "mistral"
+ },
+ "pricing": {
+ "prompt": "0",
+ "completion": "0",
+ "request": "0",
+ "image": "0",
+ "web_search": "0",
+ "internal_reasoning": "0"
+ },
+ "top_provider": {
+ "context_length": 32768,
+ "max_completion_tokens": 16384,
+ "is_moderated": false
+ },
+ "per_request_limits": null,
+ "supported_parameters": [
+ "tools",
+ "tool_choice",
+ "max_tokens",
+ "temperature",
+ "top_p",
+ "stop",
+ "frequency_penalty",
+ "presence_penalty",
+ "repetition_penalty",
+ "response_format",
+ "top_k",
+ "seed",
+ "min_p"
+ ]
+ },
+ {
+ "id": "mistralai/mistral-7b-instruct",
+ "canonical_slug": "mistralai/mistral-7b-instruct",
+ "hugging_face_id": "mistralai/Mistral-7B-Instruct-v0.3",
+ "name": "Mistral: Mistral 7B Instruct",
+ "created": 1716768000,
+ "description": "A high-performing, industry-standard 7.3B parameter model, with optimizations for speed and context length.\n\n*Mistral 7B Instruct has multiple version variants, and this is intended to be the latest version.*",
+ "context_length": 32768,
+ "architecture": {
+ "modality": "text->text",
+ "input_modalities": [
+ "text"
+ ],
+ "output_modalities": [
+ "text"
+ ],
+ "tokenizer": "Mistral",
+ "instruct_type": "mistral"
+ },
+ "pricing": {
+ "prompt": "0.000000028",
+ "completion": "0.000000054",
+ "request": "0",
+ "image": "0",
+ "web_search": "0",
+ "internal_reasoning": "0"
+ },
+ "top_provider": {
+ "context_length": 32768,
+ "max_completion_tokens": 16384,
+ "is_moderated": false
+ },
+ "per_request_limits": null,
+ "supported_parameters": [
+ "max_tokens",
+ "temperature",
+ "top_p",
+ "stop",
+ "frequency_penalty",
+ "presence_penalty",
+ "top_k",
+ "repetition_penalty",
+ "logit_bias",
+ "min_p",
+ "response_format",
+ "seed",
+ "logprobs",
+ "tools",
+ "tool_choice"
+ ]
+ },
{
"id": "microsoft/phi-3-mini-128k-instruct",
+ "canonical_slug": "microsoft/phi-3-mini-128k-instruct",
"hugging_face_id": "microsoft/Phi-3-mini-128k-instruct",
"name": "Microsoft: Phi-3 Mini 128K Instruct",
"created": 1716681600,
@@ -12369,6 +12590,7 @@
},
{
"id": "microsoft/phi-3-medium-128k-instruct",
+ "canonical_slug": "microsoft/phi-3-medium-128k-instruct",
"hugging_face_id": "microsoft/Phi-3-medium-128k-instruct",
"name": "Microsoft: Phi-3 Medium 128K Instruct",
"created": 1716508800,
@@ -12409,6 +12631,7 @@
},
{
"id": "neversleep/llama-3-lumimaid-70b",
+ "canonical_slug": "neversleep/llama-3-lumimaid-70b",
"hugging_face_id": "NeverSleep/Llama-3-Lumimaid-70B-v0.1",
"name": "NeverSleep: Llama 3 Lumimaid 70B",
"created": 1715817600,
@@ -12454,6 +12677,7 @@
},
{
"id": "google/gemini-flash-1.5",
+ "canonical_slug": "google/gemini-flash-1.5",
"hugging_face_id": null,
"name": "Google: Gemini 1.5 Flash ",
"created": 1715644800,
@@ -12501,8 +12725,109 @@
"structured_outputs"
]
},
+ {
+ "id": "openai/gpt-4o-2024-05-13",
+ "canonical_slug": "openai/gpt-4o-2024-05-13",
+ "hugging_face_id": null,
+ "name": "OpenAI: GPT-4o (2024-05-13)",
+ "created": 1715558400,
+ "description": "GPT-4o (\"o\" for \"omni\") is OpenAI's latest AI model, supporting both text and image inputs with text outputs. It maintains the intelligence level of [GPT-4 Turbo](/models/openai/gpt-4-turbo) while being twice as fast and 50% more cost-effective. GPT-4o also offers improved performance in processing non-English languages and enhanced visual capabilities.\n\nFor benchmarking against other models, it was briefly called [\"im-also-a-good-gpt2-chatbot\"](https://twitter.com/LiamFedus/status/1790064963966370209)\n\n#multimodal",
+ "context_length": 128000,
+ "architecture": {
+ "modality": "text+image->text",
+ "input_modalities": [
+ "text",
+ "image",
+ "file"
+ ],
+ "output_modalities": [
+ "text"
+ ],
+ "tokenizer": "GPT",
+ "instruct_type": null
+ },
+ "pricing": {
+ "prompt": "0.000005",
+ "completion": "0.000015",
+ "request": "0",
+ "image": "0.007225",
+ "web_search": "0",
+ "internal_reasoning": "0"
+ },
+ "top_provider": {
+ "context_length": 128000,
+ "max_completion_tokens": 4096,
+ "is_moderated": true
+ },
+ "per_request_limits": null,
+ "supported_parameters": [
+ "tools",
+ "tool_choice",
+ "max_tokens",
+ "temperature",
+ "top_p",
+ "stop",
+ "frequency_penalty",
+ "presence_penalty",
+ "web_search_options",
+ "seed",
+ "logit_bias",
+ "logprobs",
+ "top_logprobs",
+ "response_format",
+ "structured_outputs"
+ ]
+ },
+ {
+ "id": "meta-llama/llama-guard-2-8b",
+ "canonical_slug": "meta-llama/llama-guard-2-8b",
+ "hugging_face_id": "meta-llama/Meta-Llama-Guard-2-8B",
+ "name": "Meta: LlamaGuard 2 8B",
+ "created": 1715558400,
+ "description": "This safeguard model has 8B parameters and is based on the Llama 3 family. Just like is predecessor, [LlamaGuard 1](https://huggingface.co/meta-llama/LlamaGuard-7b), it can do both prompt and response classification.\n\nLlamaGuard 2 acts as a normal LLM would, generating text that indicates whether the given input/output is safe/unsafe. If deemed unsafe, it will also share the content categories violated.\n\nFor best results, please use raw prompt input or the `/completions` endpoint, instead of the chat API.\n\nIt has demonstrated strong performance compared to leading closed-source models in human evaluations.\n\nTo read more about the model release, [click here](https://ai.meta.com/blog/meta-llama-3/). Usage of this model is subject to [Meta's Acceptable Use Policy](https://llama.meta.com/llama3/use-policy/).",
+ "context_length": 8192,
+ "architecture": {
+ "modality": "text->text",
+ "input_modalities": [
+ "text"
+ ],
+ "output_modalities": [
+ "text"
+ ],
+ "tokenizer": "Llama3",
+ "instruct_type": "none"
+ },
+ "pricing": {
+ "prompt": "0.0000002",
+ "completion": "0.0000002",
+ "request": "0",
+ "image": "0",
+ "web_search": "0",
+ "internal_reasoning": "0"
+ },
+ "top_provider": {
+ "context_length": 8192,
+ "max_completion_tokens": null,
+ "is_moderated": false
+ },
+ "per_request_limits": null,
+ "supported_parameters": [
+ "max_tokens",
+ "temperature",
+ "top_p",
+ "stop",
+ "frequency_penalty",
+ "presence_penalty",
+ "top_k",
+ "repetition_penalty",
+ "logit_bias",
+ "min_p",
+ "response_format"
+ ]
+ },
{
"id": "openai/gpt-4o",
+ "canonical_slug": "openai/gpt-4o",
"hugging_face_id": null,
"name": "OpenAI: GPT-4o",
"created": 1715558400,
@@ -12556,6 +12881,7 @@
},
{
"id": "openai/gpt-4o:extended",
+ "canonical_slug": "openai/gpt-4o",
"hugging_face_id": null,
"name": "OpenAI: GPT-4o (extended)",
"created": 1715558400,
@@ -12606,106 +12932,9 @@
"structured_outputs"
]
},
- {
- "id": "meta-llama/llama-guard-2-8b",
- "hugging_face_id": "meta-llama/Meta-Llama-Guard-2-8B",
- "name": "Meta: LlamaGuard 2 8B",
- "created": 1715558400,
- "description": "This safeguard model has 8B parameters and is based on the Llama 3 family. Just like is predecessor, [LlamaGuard 1](https://huggingface.co/meta-llama/LlamaGuard-7b), it can do both prompt and response classification.\n\nLlamaGuard 2 acts as a normal LLM would, generating text that indicates whether the given input/output is safe/unsafe. If deemed unsafe, it will also share the content categories violated.\n\nFor best results, please use raw prompt input or the `/completions` endpoint, instead of the chat API.\n\nIt has demonstrated strong performance compared to leading closed-source models in human evaluations.\n\nTo read more about the model release, [click here](https://ai.meta.com/blog/meta-llama-3/). Usage of this model is subject to [Meta's Acceptable Use Policy](https://llama.meta.com/llama3/use-policy/).",
- "context_length": 8192,
- "architecture": {
- "modality": "text->text",
- "input_modalities": [
- "text"
- ],
- "output_modalities": [
- "text"
- ],
- "tokenizer": "Llama3",
- "instruct_type": "none"
- },
- "pricing": {
- "prompt": "0.0000002",
- "completion": "0.0000002",
- "request": "0",
- "image": "0",
- "web_search": "0",
- "internal_reasoning": "0"
- },
- "top_provider": {
- "context_length": 8192,
- "max_completion_tokens": null,
- "is_moderated": false
- },
- "per_request_limits": null,
- "supported_parameters": [
- "max_tokens",
- "temperature",
- "top_p",
- "stop",
- "frequency_penalty",
- "presence_penalty",
- "top_k",
- "repetition_penalty",
- "logit_bias",
- "min_p",
- "response_format"
- ]
- },
- {
- "id": "openai/gpt-4o-2024-05-13",
- "hugging_face_id": null,
- "name": "OpenAI: GPT-4o (2024-05-13)",
- "created": 1715558400,
- "description": "GPT-4o (\"o\" for \"omni\") is OpenAI's latest AI model, supporting both text and image inputs with text outputs. It maintains the intelligence level of [GPT-4 Turbo](/models/openai/gpt-4-turbo) while being twice as fast and 50% more cost-effective. GPT-4o also offers improved performance in processing non-English languages and enhanced visual capabilities.\n\nFor benchmarking against other models, it was briefly called [\"im-also-a-good-gpt2-chatbot\"](https://twitter.com/LiamFedus/status/1790064963966370209)\n\n#multimodal",
- "context_length": 128000,
- "architecture": {
- "modality": "text+image->text",
- "input_modalities": [
- "text",
- "image",
- "file"
- ],
- "output_modalities": [
- "text"
- ],
- "tokenizer": "GPT",
- "instruct_type": null
- },
- "pricing": {
- "prompt": "0.000005",
- "completion": "0.000015",
- "request": "0",
- "image": "0.007225",
- "web_search": "0",
- "internal_reasoning": "0"
- },
- "top_provider": {
- "context_length": 128000,
- "max_completion_tokens": 4096,
- "is_moderated": true
- },
- "per_request_limits": null,
- "supported_parameters": [
- "tools",
- "tool_choice",
- "max_tokens",
- "temperature",
- "top_p",
- "stop",
- "frequency_penalty",
- "presence_penalty",
- "web_search_options",
- "seed",
- "logit_bias",
- "logprobs",
- "top_logprobs",
- "response_format",
- "structured_outputs"
- ]
- },
{
"id": "neversleep/llama-3-lumimaid-8b",
+ "canonical_slug": "neversleep/llama-3-lumimaid-8b",
"hugging_face_id": "NeverSleep/Llama-3-Lumimaid-8B-v0.1",
"name": "NeverSleep: Llama 3 Lumimaid 8B",
"created": 1714780800,
@@ -12753,6 +12982,7 @@
},
{
"id": "sao10k/fimbulvetr-11b-v2",
+ "canonical_slug": "sao10k/fimbulvetr-11b-v2",
"hugging_face_id": "Sao10K/Fimbulvetr-11B-v2",
"name": "Fimbulvetr 11B v2",
"created": 1713657600,
@@ -12798,6 +13028,7 @@
},
{
"id": "meta-llama/llama-3-8b-instruct",
+ "canonical_slug": "meta-llama/llama-3-8b-instruct",
"hugging_face_id": "meta-llama/Meta-Llama-3-8B-Instruct",
"name": "Meta: Llama 3 8B Instruct",
"created": 1713398400,
@@ -12850,6 +13081,7 @@
},
{
"id": "meta-llama/llama-3-70b-instruct",
+ "canonical_slug": "meta-llama/llama-3-70b-instruct",
"hugging_face_id": "meta-llama/Meta-Llama-3-70B-Instruct",
"name": "Meta: Llama 3 70B Instruct",
"created": 1713398400,
@@ -12901,6 +13133,7 @@
},
{
"id": "mistralai/mixtral-8x22b-instruct",
+ "canonical_slug": "mistralai/mixtral-8x22b-instruct",
"hugging_face_id": "mistralai/Mixtral-8x22B-Instruct-v0.1",
"name": "Mistral: Mixtral 8x22B Instruct",
"created": 1713312000,
@@ -12952,6 +13185,7 @@
},
{
"id": "microsoft/wizardlm-2-8x22b",
+ "canonical_slug": "microsoft/wizardlm-2-8x22b",
"hugging_face_id": "microsoft/WizardLM-2-8x22B",
"name": "WizardLM-2 8x22B",
"created": 1713225600,
@@ -12997,55 +13231,9 @@
"response_format"
]
},
- {
- "id": "google/gemini-pro-1.5",
- "hugging_face_id": null,
- "name": "Google: Gemini 1.5 Pro",
- "created": 1712620800,
- "description": "Google's latest multimodal model, supports image and video[0] in text or chat prompts.\n\nOptimized for language tasks including:\n\n- Code generation\n- Text generation\n- Text editing\n- Problem solving\n- Recommendations\n- Information extraction\n- Data extraction or generation\n- AI agents\n\nUsage of Gemini is subject to Google's [Gemini Terms of Use](https://ai.google.dev/terms).\n\n* [0]: Video input is not available through OpenRouter at this time.",
- "context_length": 2000000,
- "architecture": {
- "modality": "text+image->text",
- "input_modalities": [
- "text",
- "image"
- ],
- "output_modalities": [
- "text"
- ],
- "tokenizer": "Gemini",
- "instruct_type": null
- },
- "pricing": {
- "prompt": "0.00000125",
- "completion": "0.000005",
- "request": "0",
- "image": "0.0006575",
- "web_search": "0",
- "internal_reasoning": "0"
- },
- "top_provider": {
- "context_length": 2000000,
- "max_completion_tokens": 8192,
- "is_moderated": false
- },
- "per_request_limits": null,
- "supported_parameters": [
- "max_tokens",
- "temperature",
- "top_p",
- "stop",
- "frequency_penalty",
- "presence_penalty",
- "tools",
- "tool_choice",
- "seed",
- "response_format",
- "structured_outputs"
- ]
- },
{
"id": "openai/gpt-4-turbo",
+ "canonical_slug": "openai/gpt-4-turbo",
"hugging_face_id": null,
"name": "OpenAI: GPT-4 Turbo",
"created": 1712620800,
@@ -13093,8 +13281,57 @@
"response_format"
]
},
+ {
+ "id": "google/gemini-pro-1.5",
+ "canonical_slug": "google/gemini-pro-1.5",
+ "hugging_face_id": null,
+ "name": "Google: Gemini 1.5 Pro",
+ "created": 1712620800,
+ "description": "Google's latest multimodal model, supports image and video[0] in text or chat prompts.\n\nOptimized for language tasks including:\n\n- Code generation\n- Text generation\n- Text editing\n- Problem solving\n- Recommendations\n- Information extraction\n- Data extraction or generation\n- AI agents\n\nUsage of Gemini is subject to Google's [Gemini Terms of Use](https://ai.google.dev/terms).\n\n* [0]: Video input is not available through OpenRouter at this time.",
+ "context_length": 2000000,
+ "architecture": {
+ "modality": "text+image->text",
+ "input_modalities": [
+ "text",
+ "image"
+ ],
+ "output_modalities": [
+ "text"
+ ],
+ "tokenizer": "Gemini",
+ "instruct_type": null
+ },
+ "pricing": {
+ "prompt": "0.00000125",
+ "completion": "0.000005",
+ "request": "0",
+ "image": "0.0006575",
+ "web_search": "0",
+ "internal_reasoning": "0"
+ },
+ "top_provider": {
+ "context_length": 2000000,
+ "max_completion_tokens": 8192,
+ "is_moderated": false
+ },
+ "per_request_limits": null,
+ "supported_parameters": [
+ "max_tokens",
+ "temperature",
+ "top_p",
+ "stop",
+ "frequency_penalty",
+ "presence_penalty",
+ "tools",
+ "tool_choice",
+ "seed",
+ "response_format",
+ "structured_outputs"
+ ]
+ },
{
"id": "cohere/command-r-plus",
+ "canonical_slug": "cohere/command-r-plus",
"hugging_face_id": null,
"name": "Cohere: Command R+",
"created": 1712188800,
@@ -13141,6 +13378,7 @@
},
{
"id": "cohere/command-r-plus-04-2024",
+ "canonical_slug": "cohere/command-r-plus-04-2024",
"hugging_face_id": null,
"name": "Cohere: Command R+ (04-2024)",
"created": 1712016000,
@@ -13187,6 +13425,7 @@
},
{
"id": "sophosympatheia/midnight-rose-70b",
+ "canonical_slug": "sophosympatheia/midnight-rose-70b",
"hugging_face_id": "sophosympatheia/Midnight-Rose-70B-v2.0.3",
"name": "Midnight Rose 70B",
"created": 1711065600,
@@ -13231,53 +13470,9 @@
"logit_bias"
]
},
- {
- "id": "cohere/command",
- "hugging_face_id": null,
- "name": "Cohere: Command",
- "created": 1710374400,
- "description": "Command is an instruction-following conversational model that performs language tasks with high quality, more reliably and with a longer context than our base generative models.\n\nUse of this model is subject to Cohere's [Usage Policy](https://docs.cohere.com/docs/usage-policy) and [SaaS Agreement](https://cohere.com/saas-agreement).",
- "context_length": 4096,
- "architecture": {
- "modality": "text->text",
- "input_modalities": [
- "text"
- ],
- "output_modalities": [
- "text"
- ],
- "tokenizer": "Cohere",
- "instruct_type": null
- },
- "pricing": {
- "prompt": "0.000001",
- "completion": "0.000002",
- "request": "0",
- "image": "0",
- "web_search": "0",
- "internal_reasoning": "0"
- },
- "top_provider": {
- "context_length": 4096,
- "max_completion_tokens": 4000,
- "is_moderated": true
- },
- "per_request_limits": null,
- "supported_parameters": [
- "max_tokens",
- "temperature",
- "top_p",
- "stop",
- "frequency_penalty",
- "presence_penalty",
- "top_k",
- "seed",
- "response_format",
- "structured_outputs"
- ]
- },
{
"id": "cohere/command-r",
+ "canonical_slug": "cohere/command-r",
"hugging_face_id": null,
"name": "Cohere: Command R",
"created": 1710374400,
@@ -13322,8 +13517,55 @@
"structured_outputs"
]
},
+ {
+ "id": "cohere/command",
+ "canonical_slug": "cohere/command",
+ "hugging_face_id": null,
+ "name": "Cohere: Command",
+ "created": 1710374400,
+ "description": "Command is an instruction-following conversational model that performs language tasks with high quality, more reliably and with a longer context than our base generative models.\n\nUse of this model is subject to Cohere's [Usage Policy](https://docs.cohere.com/docs/usage-policy) and [SaaS Agreement](https://cohere.com/saas-agreement).",
+ "context_length": 4096,
+ "architecture": {
+ "modality": "text->text",
+ "input_modalities": [
+ "text"
+ ],
+ "output_modalities": [
+ "text"
+ ],
+ "tokenizer": "Cohere",
+ "instruct_type": null
+ },
+ "pricing": {
+ "prompt": "0.000001",
+ "completion": "0.000002",
+ "request": "0",
+ "image": "0",
+ "web_search": "0",
+ "internal_reasoning": "0"
+ },
+ "top_provider": {
+ "context_length": 4096,
+ "max_completion_tokens": 4000,
+ "is_moderated": true
+ },
+ "per_request_limits": null,
+ "supported_parameters": [
+ "max_tokens",
+ "temperature",
+ "top_p",
+ "stop",
+ "frequency_penalty",
+ "presence_penalty",
+ "top_k",
+ "seed",
+ "response_format",
+ "structured_outputs"
+ ]
+ },
{
"id": "anthropic/claude-3-haiku:beta",
+ "canonical_slug": "anthropic/claude-3-haiku",
"hugging_face_id": null,
"name": "Anthropic: Claude 3 Haiku (self-moderated)",
"created": 1710288000,
@@ -13369,6 +13611,7 @@
},
{
"id": "anthropic/claude-3-haiku",
+ "canonical_slug": "anthropic/claude-3-haiku",
"hugging_face_id": null,
"name": "Anthropic: Claude 3 Haiku",
"created": 1710288000,
@@ -13414,6 +13657,7 @@
},
{
"id": "anthropic/claude-3-opus:beta",
+ "canonical_slug": "anthropic/claude-3-opus",
"hugging_face_id": null,
"name": "Anthropic: Claude 3 Opus (self-moderated)",
"created": 1709596800,
@@ -13459,6 +13703,7 @@
},
{
"id": "anthropic/claude-3-opus",
+ "canonical_slug": "anthropic/claude-3-opus",
"hugging_face_id": null,
"name": "Anthropic: Claude 3 Opus",
"created": 1709596800,
@@ -13504,6 +13749,7 @@
},
{
"id": "anthropic/claude-3-sonnet:beta",
+ "canonical_slug": "anthropic/claude-3-sonnet",
"hugging_face_id": null,
"name": "Anthropic: Claude 3 Sonnet (self-moderated)",
"created": 1709596800,
@@ -13549,6 +13795,7 @@
},
{
"id": "anthropic/claude-3-sonnet",
+ "canonical_slug": "anthropic/claude-3-sonnet",
"hugging_face_id": null,
"name": "Anthropic: Claude 3 Sonnet",
"created": 1709596800,
@@ -13594,6 +13841,7 @@
},
{
"id": "cohere/command-r-03-2024",
+ "canonical_slug": "cohere/command-r-03-2024",
"hugging_face_id": null,
"name": "Cohere: Command R (03-2024)",
"created": 1709341200,
@@ -13640,6 +13888,7 @@
},
{
"id": "mistralai/mistral-large",
+ "canonical_slug": "mistralai/mistral-large",
"hugging_face_id": null,
"name": "Mistral Large",
"created": 1708905600,
@@ -13686,6 +13935,7 @@
},
{
"id": "openai/gpt-3.5-turbo-0613",
+ "canonical_slug": "openai/gpt-3.5-turbo-0613",
"hugging_face_id": null,
"name": "OpenAI: GPT-3.5 Turbo (older v0613)",
"created": 1706140800,
@@ -13735,6 +13985,7 @@
},
{
"id": "openai/gpt-4-turbo-preview",
+ "canonical_slug": "openai/gpt-4-turbo-preview",
"hugging_face_id": null,
"name": "OpenAI: GPT-4 Turbo Preview",
"created": 1706140800,
@@ -13784,6 +14035,7 @@
},
{
"id": "nousresearch/nous-hermes-2-mixtral-8x7b-dpo",
+ "canonical_slug": "nousresearch/nous-hermes-2-mixtral-8x7b-dpo",
"hugging_face_id": "NousResearch/Nous-Hermes-2-Mixtral-8x7B-DPO",
"name": "Nous: Hermes 2 Mixtral 8x7B DPO",
"created": 1705363200,
@@ -13828,54 +14080,9 @@
"response_format"
]
},
- {
- "id": "mistralai/mistral-medium",
- "hugging_face_id": null,
- "name": "Mistral Medium",
- "created": 1704844800,
- "description": "This is Mistral AI's closed-source, medium-sided model. It's powered by a closed-source prototype and excels at reasoning, code, JSON, chat, and more. In benchmarks, it compares with many of the flagship models of other companies.",
- "context_length": 32768,
- "architecture": {
- "modality": "text->text",
- "input_modalities": [
- "text"
- ],
- "output_modalities": [
- "text"
- ],
- "tokenizer": "Mistral",
- "instruct_type": null
- },
- "pricing": {
- "prompt": "0.00000275",
- "completion": "0.0000081",
- "request": "0",
- "image": "0",
- "web_search": "0",
- "internal_reasoning": "0"
- },
- "top_provider": {
- "context_length": 32768,
- "max_completion_tokens": null,
- "is_moderated": false
- },
- "per_request_limits": null,
- "supported_parameters": [
- "tools",
- "tool_choice",
- "max_tokens",
- "temperature",
- "top_p",
- "stop",
- "frequency_penalty",
- "presence_penalty",
- "response_format",
- "structured_outputs",
- "seed"
- ]
- },
{
"id": "mistralai/mistral-small",
+ "canonical_slug": "mistralai/mistral-small",
"hugging_face_id": null,
"name": "Mistral Small",
"created": 1704844800,
@@ -13922,6 +14129,7 @@
},
{
"id": "mistralai/mistral-tiny",
+ "canonical_slug": "mistralai/mistral-tiny",
"hugging_face_id": null,
"name": "Mistral Tiny",
"created": 1704844800,
@@ -13968,6 +14176,7 @@
},
{
"id": "mistralai/mistral-7b-instruct-v0.2",
+ "canonical_slug": "mistralai/mistral-7b-instruct-v0.2",
"hugging_face_id": "mistralai/Mistral-7B-Instruct-v0.2",
"name": "Mistral: Mistral 7B Instruct v0.2",
"created": 1703721600,
@@ -14014,6 +14223,7 @@
},
{
"id": "mistralai/mixtral-8x7b-instruct",
+ "canonical_slug": "mistralai/mixtral-8x7b-instruct",
"hugging_face_id": "mistralai/Mixtral-8x7B-Instruct-v0.1",
"name": "Mistral: Mixtral 8x7B Instruct",
"created": 1702166400,
@@ -14063,6 +14273,7 @@
},
{
"id": "neversleep/noromaid-20b",
+ "canonical_slug": "neversleep/noromaid-20b",
"hugging_face_id": "NeverSleep/Noromaid-20b-v0.1.1",
"name": "Noromaid 20B",
"created": 1700956800,
@@ -14108,88 +14319,9 @@
"top_a"
]
},
- {
- "id": "anthropic/claude-2.1:beta",
- "hugging_face_id": null,
- "name": "Anthropic: Claude v2.1 (self-moderated)",
- "created": 1700611200,
- "description": "Claude 2 delivers advancements in key capabilities for enterprises—including an industry-leading 200K token context window, significant reductions in rates of model hallucination, system prompts and a new beta feature: tool use.",
- "context_length": 200000,
- "architecture": {
- "modality": "text->text",
- "input_modalities": [
- "text"
- ],
- "output_modalities": [
- "text"
- ],
- "tokenizer": "Claude",
- "instruct_type": null
- },
- "pricing": {
- "prompt": "0.000008",
- "completion": "0.000024",
- "request": "0",
- "image": "0",
- "web_search": "0",
- "internal_reasoning": "0"
- },
- "top_provider": {
- "context_length": 200000,
- "max_completion_tokens": 4096,
- "is_moderated": false
- },
- "per_request_limits": null,
- "supported_parameters": [
- "max_tokens",
- "temperature",
- "top_p",
- "top_k",
- "stop"
- ]
- },
- {
- "id": "anthropic/claude-2.1",
- "hugging_face_id": null,
- "name": "Anthropic: Claude v2.1",
- "created": 1700611200,
- "description": "Claude 2 delivers advancements in key capabilities for enterprises—including an industry-leading 200K token context window, significant reductions in rates of model hallucination, system prompts and a new beta feature: tool use.",
- "context_length": 200000,
- "architecture": {
- "modality": "text->text",
- "input_modalities": [
- "text"
- ],
- "output_modalities": [
- "text"
- ],
- "tokenizer": "Claude",
- "instruct_type": null
- },
- "pricing": {
- "prompt": "0.000008",
- "completion": "0.000024",
- "request": "0",
- "image": "0",
- "web_search": "0",
- "internal_reasoning": "0"
- },
- "top_provider": {
- "context_length": 200000,
- "max_completion_tokens": 4096,
- "is_moderated": true
- },
- "per_request_limits": null,
- "supported_parameters": [
- "max_tokens",
- "temperature",
- "top_p",
- "top_k",
- "stop"
- ]
- },
{
"id": "anthropic/claude-2:beta",
+ "canonical_slug": "anthropic/claude-2",
"hugging_face_id": null,
"name": "Anthropic: Claude v2 (self-moderated)",
"created": 1700611200,
@@ -14230,6 +14362,7 @@
},
{
"id": "anthropic/claude-2",
+ "canonical_slug": "anthropic/claude-2",
"hugging_face_id": null,
"name": "Anthropic: Claude v2",
"created": 1700611200,
@@ -14268,8 +14401,91 @@
"stop"
]
},
+ {
+ "id": "anthropic/claude-2.1:beta",
+ "canonical_slug": "anthropic/claude-2.1",
+ "hugging_face_id": null,
+ "name": "Anthropic: Claude v2.1 (self-moderated)",
+ "created": 1700611200,
+ "description": "Claude 2 delivers advancements in key capabilities for enterprises—including an industry-leading 200K token context window, significant reductions in rates of model hallucination, system prompts and a new beta feature: tool use.",
+ "context_length": 200000,
+ "architecture": {
+ "modality": "text->text",
+ "input_modalities": [
+ "text"
+ ],
+ "output_modalities": [
+ "text"
+ ],
+ "tokenizer": "Claude",
+ "instruct_type": null
+ },
+ "pricing": {
+ "prompt": "0.000008",
+ "completion": "0.000024",
+ "request": "0",
+ "image": "0",
+ "web_search": "0",
+ "internal_reasoning": "0"
+ },
+ "top_provider": {
+ "context_length": 200000,
+ "max_completion_tokens": 4096,
+ "is_moderated": false
+ },
+ "per_request_limits": null,
+ "supported_parameters": [
+ "max_tokens",
+ "temperature",
+ "top_p",
+ "top_k",
+ "stop"
+ ]
+ },
+ {
+ "id": "anthropic/claude-2.1",
+ "canonical_slug": "anthropic/claude-2.1",
+ "hugging_face_id": null,
+ "name": "Anthropic: Claude v2.1",
+ "created": 1700611200,
+ "description": "Claude 2 delivers advancements in key capabilities for enterprises—including an industry-leading 200K token context window, significant reductions in rates of model hallucination, system prompts and a new beta feature: tool use.",
+ "context_length": 200000,
+ "architecture": {
+ "modality": "text->text",
+ "input_modalities": [
+ "text"
+ ],
+ "output_modalities": [
+ "text"
+ ],
+ "tokenizer": "Claude",
+ "instruct_type": null
+ },
+ "pricing": {
+ "prompt": "0.000008",
+ "completion": "0.000024",
+ "request": "0",
+ "image": "0",
+ "web_search": "0",
+ "internal_reasoning": "0"
+ },
+ "top_provider": {
+ "context_length": 200000,
+ "max_completion_tokens": 4096,
+ "is_moderated": true
+ },
+ "per_request_limits": null,
+ "supported_parameters": [
+ "max_tokens",
+ "temperature",
+ "top_p",
+ "top_k",
+ "stop"
+ ]
+ },
{
"id": "undi95/toppy-m-7b",
+ "canonical_slug": "undi95/toppy-m-7b",
"hugging_face_id": "Undi95/Toppy-M-7B",
"name": "Toppy M 7B",
"created": 1699574400,
@@ -14315,6 +14531,7 @@
},
{
"id": "alpindale/goliath-120b",
+ "canonical_slug": "alpindale/goliath-120b",
"hugging_face_id": "alpindale/goliath-120b",
"name": "Goliath 120B",
"created": 1699574400,
@@ -14332,8 +14549,8 @@
"instruct_type": "airoboros"
},
"pricing": {
- "prompt": "0.00001",
- "completion": "0.0000125",
+ "prompt": "0.000009",
+ "completion": "0.000011",
"request": "0",
"image": "0",
"web_search": "0",
@@ -14341,7 +14558,7 @@
},
"top_provider": {
"context_length": 6144,
- "max_completion_tokens": 512,
+ "max_completion_tokens": null,
"is_moderated": false
},
"per_request_limits": null,
@@ -14352,6 +14569,8 @@
"stop",
"frequency_penalty",
"presence_penalty",
+ "response_format",
+ "structured_outputs",
"repetition_penalty",
"logit_bias",
"top_k",
@@ -14362,6 +14581,7 @@
},
{
"id": "openrouter/auto",
+ "canonical_slug": "openrouter/auto",
"hugging_face_id": null,
"name": "Auto Router",
"created": 1699401600,
@@ -14390,57 +14610,9 @@
"per_request_limits": null,
"supported_parameters": []
},
- {
- "id": "openai/gpt-3.5-turbo-1106",
- "hugging_face_id": null,
- "name": "OpenAI: GPT-3.5 Turbo 16k (older v1106)",
- "created": 1699228800,
- "description": "An older GPT-3.5 Turbo model with improved instruction following, JSON mode, reproducible outputs, parallel function calling, and more. Training data: up to Sep 2021.",
- "context_length": 16385,
- "architecture": {
- "modality": "text->text",
- "input_modalities": [
- "text"
- ],
- "output_modalities": [
- "text"
- ],
- "tokenizer": "GPT",
- "instruct_type": null
- },
- "pricing": {
- "prompt": "0.000001",
- "completion": "0.000002",
- "request": "0",
- "image": "0",
- "web_search": "0",
- "internal_reasoning": "0"
- },
- "top_provider": {
- "context_length": 16385,
- "max_completion_tokens": 4096,
- "is_moderated": true
- },
- "per_request_limits": null,
- "supported_parameters": [
- "tools",
- "tool_choice",
- "max_tokens",
- "temperature",
- "top_p",
- "stop",
- "frequency_penalty",
- "presence_penalty",
- "seed",
- "logit_bias",
- "logprobs",
- "top_logprobs",
- "response_format",
- "structured_outputs"
- ]
- },
{
"id": "openai/gpt-4-1106-preview",
+ "canonical_slug": "openai/gpt-4-1106-preview",
"hugging_face_id": null,
"name": "OpenAI: GPT-4 Turbo (older v1106)",
"created": 1699228800,
@@ -14490,6 +14662,7 @@
},
{
"id": "openai/gpt-3.5-turbo-instruct",
+ "canonical_slug": "openai/gpt-3.5-turbo-instruct",
"hugging_face_id": null,
"name": "OpenAI: GPT-3.5 Turbo Instruct",
"created": 1695859200,
@@ -14536,6 +14709,7 @@
},
{
"id": "mistralai/mistral-7b-instruct-v0.1",
+ "canonical_slug": "mistralai/mistral-7b-instruct-v0.1",
"hugging_face_id": "mistralai/Mistral-7B-Instruct-v0.1",
"name": "Mistral: Mistral 7B Instruct v0.1",
"created": 1695859200,
@@ -14585,6 +14759,7 @@
},
{
"id": "pygmalionai/mythalion-13b",
+ "canonical_slug": "pygmalionai/mythalion-13b",
"hugging_face_id": "PygmalionAI/mythalion-13b",
"name": "Pygmalion: Mythalion 13B",
"created": 1693612800,
@@ -14630,6 +14805,7 @@
},
{
"id": "openai/gpt-3.5-turbo-16k",
+ "canonical_slug": "openai/gpt-3.5-turbo-16k",
"hugging_face_id": null,
"name": "OpenAI: GPT-3.5 Turbo 16k",
"created": 1693180800,
@@ -14678,6 +14854,7 @@
},
{
"id": "mancer/weaver",
+ "canonical_slug": "mancer/weaver",
"hugging_face_id": null,
"name": "Mancer: Weaver (alpha)",
"created": 1690934400,
@@ -14725,6 +14902,7 @@
},
{
"id": "anthropic/claude-2.0:beta",
+ "canonical_slug": "anthropic/claude-2.0",
"hugging_face_id": null,
"name": "Anthropic: Claude v2.0 (self-moderated)",
"created": 1690502400,
@@ -14765,6 +14943,7 @@
},
{
"id": "anthropic/claude-2.0",
+ "canonical_slug": "anthropic/claude-2.0",
"hugging_face_id": null,
"name": "Anthropic: Claude v2.0",
"created": 1690502400,
@@ -14805,6 +14984,7 @@
},
{
"id": "undi95/remm-slerp-l2-13b",
+ "canonical_slug": "undi95/remm-slerp-l2-13b",
"hugging_face_id": "Undi95/ReMM-SLERP-L2-13B",
"name": "ReMM SLERP 13B",
"created": 1689984000,
@@ -14852,6 +15032,7 @@
},
{
"id": "gryphe/mythomax-l2-13b",
+ "canonical_slug": "gryphe/mythomax-l2-13b",
"hugging_face_id": "Gryphe/MythoMax-L2-13b",
"name": "MythoMax 13B",
"created": 1688256000,
@@ -14886,162 +15067,21 @@
"max_tokens",
"temperature",
"top_p",
- "stop",
- "frequency_penalty",
"presence_penalty",
- "seed",
- "top_k",
- "min_p",
+ "frequency_penalty",
"repetition_penalty",
+ "top_k",
+ "stop",
+ "seed",
+ "min_p",
"logit_bias",
"response_format",
"top_a"
]
},
- {
- "id": "meta-llama/llama-2-70b-chat",
- "hugging_face_id": "meta-llama/Llama-2-70b-chat-hf",
- "name": "Meta: Llama 2 70B Chat",
- "created": 1687219200,
- "description": "The flagship, 70 billion parameter language model from Meta, fine tuned for chat completions. Llama 2 is an auto-regressive language model that uses an optimized transformer architecture. The tuned versions use supervised fine-tuning (SFT) and reinforcement learning with human feedback (RLHF) to align to human preferences for helpfulness and safety.",
- "context_length": 4096,
- "architecture": {
- "modality": "text->text",
- "input_modalities": [
- "text"
- ],
- "output_modalities": [
- "text"
- ],
- "tokenizer": "Llama2",
- "instruct_type": "llama2"
- },
- "pricing": {
- "prompt": "0.0000009",
- "completion": "0.0000009",
- "request": "0",
- "image": "0",
- "web_search": "0",
- "internal_reasoning": "0"
- },
- "top_provider": {
- "context_length": 4096,
- "max_completion_tokens": null,
- "is_moderated": false
- },
- "per_request_limits": null,
- "supported_parameters": [
- "max_tokens",
- "temperature",
- "top_p",
- "stop",
- "frequency_penalty",
- "presence_penalty",
- "top_k",
- "repetition_penalty",
- "logit_bias",
- "min_p",
- "response_format"
- ]
- },
- {
- "id": "openai/gpt-3.5-turbo",
- "hugging_face_id": null,
- "name": "OpenAI: GPT-3.5 Turbo",
- "created": 1685232000,
- "description": "GPT-3.5 Turbo is OpenAI's fastest model. It can understand and generate natural language or code, and is optimized for chat and traditional completion tasks.\n\nTraining data up to Sep 2021.",
- "context_length": 16385,
- "architecture": {
- "modality": "text->text",
- "input_modalities": [
- "text"
- ],
- "output_modalities": [
- "text"
- ],
- "tokenizer": "GPT",
- "instruct_type": null
- },
- "pricing": {
- "prompt": "0.0000005",
- "completion": "0.0000015",
- "request": "0",
- "image": "0",
- "web_search": "0",
- "internal_reasoning": "0"
- },
- "top_provider": {
- "context_length": 16385,
- "max_completion_tokens": 4096,
- "is_moderated": true
- },
- "per_request_limits": null,
- "supported_parameters": [
- "tools",
- "tool_choice",
- "max_tokens",
- "temperature",
- "top_p",
- "stop",
- "frequency_penalty",
- "presence_penalty",
- "seed",
- "logit_bias",
- "logprobs",
- "top_logprobs",
- "response_format"
- ]
- },
- {
- "id": "openai/gpt-3.5-turbo-0125",
- "hugging_face_id": null,
- "name": "OpenAI: GPT-3.5 Turbo 16k",
- "created": 1685232000,
- "description": "The latest GPT-3.5 Turbo model with improved instruction following, JSON mode, reproducible outputs, parallel function calling, and more. Training data: up to Sep 2021.\n\nThis version has a higher accuracy at responding in requested formats and a fix for a bug which caused a text encoding issue for non-English language function calls.",
- "context_length": 16385,
- "architecture": {
- "modality": "text->text",
- "input_modalities": [
- "text"
- ],
- "output_modalities": [
- "text"
- ],
- "tokenizer": "GPT",
- "instruct_type": null
- },
- "pricing": {
- "prompt": "0.0000005",
- "completion": "0.0000015",
- "request": "0",
- "image": "0",
- "web_search": "0",
- "internal_reasoning": "0"
- },
- "top_provider": {
- "context_length": 16385,
- "max_completion_tokens": 4096,
- "is_moderated": true
- },
- "per_request_limits": null,
- "supported_parameters": [
- "tools",
- "tool_choice",
- "max_tokens",
- "temperature",
- "top_p",
- "stop",
- "frequency_penalty",
- "presence_penalty",
- "seed",
- "logit_bias",
- "logprobs",
- "top_logprobs",
- "response_format"
- ]
- },
{
"id": "openai/gpt-4",
+ "canonical_slug": "openai/gpt-4",
"hugging_face_id": null,
"name": "OpenAI: GPT-4",
"created": 1685232000,
@@ -15090,6 +15130,7 @@
},
{
"id": "openai/gpt-4-0314",
+ "canonical_slug": "openai/gpt-4-0314",
"hugging_face_id": null,
"name": "OpenAI: GPT-4 (older v0314)",
"created": 1685232000,
diff --git a/packages/kbot/dist-in/src/models/cache/openai.ts b/packages/kbot/dist-in/src/models/cache/openai.ts
index e3ad829e..f45bb40e 100644
--- a/packages/kbot/dist-in/src/models/cache/openai.ts
+++ b/packages/kbot/dist-in/src/models/cache/openai.ts
@@ -1 +1 @@
-export const models = [{"id":"gpt-4-0613","object":"model","created":1686588896,"owned_by":"openai"},{"id":"gpt-4","object":"model","created":1687882411,"owned_by":"openai"},{"id":"gpt-3.5-turbo","object":"model","created":1677610602,"owned_by":"openai"},{"id":"gpt-4o-audio-preview-2025-06-03","object":"model","created":1748908498,"owned_by":"system"},{"id":"gpt-4.1-nano","object":"model","created":1744321707,"owned_by":"system"},{"id":"gpt-image-1","object":"model","created":1745517030,"owned_by":"system"},{"id":"codex-mini-latest","object":"model","created":1746673257,"owned_by":"system"},{"id":"gpt-4o-realtime-preview-2025-06-03","object":"model","created":1748907838,"owned_by":"system"},{"id":"davinci-002","object":"model","created":1692634301,"owned_by":"system"},{"id":"babbage-002","object":"model","created":1692634615,"owned_by":"system"},{"id":"gpt-3.5-turbo-instruct","object":"model","created":1692901427,"owned_by":"system"},{"id":"gpt-3.5-turbo-instruct-0914","object":"model","created":1694122472,"owned_by":"system"},{"id":"dall-e-3","object":"model","created":1698785189,"owned_by":"system"},{"id":"dall-e-2","object":"model","created":1698798177,"owned_by":"system"},{"id":"gpt-4-1106-preview","object":"model","created":1698957206,"owned_by":"system"},{"id":"gpt-3.5-turbo-1106","object":"model","created":1698959748,"owned_by":"system"},{"id":"tts-1-hd","object":"model","created":1699046015,"owned_by":"system"},{"id":"tts-1-1106","object":"model","created":1699053241,"owned_by":"system"},{"id":"tts-1-hd-1106","object":"model","created":1699053533,"owned_by":"system"},{"id":"text-embedding-3-small","object":"model","created":1705948997,"owned_by":"system"},{"id":"text-embedding-3-large","object":"model","created":1705953180,"owned_by":"system"},{"id":"gpt-4-0125-preview","object":"model","created":1706037612,"owned_by":"system"},{"id":"gpt-4-turbo-preview","object":"model","created":1706037777,"owned_by":"system"},{"id":"gpt-3.5-turbo-0125","object":"model","created":1706048358,"owned_by":"system"},{"id":"gpt-4-turbo","object":"model","created":1712361441,"owned_by":"system"},{"id":"gpt-4-turbo-2024-04-09","object":"model","created":1712601677,"owned_by":"system"},{"id":"gpt-4o","object":"model","created":1715367049,"owned_by":"system"},{"id":"gpt-4o-2024-05-13","object":"model","created":1715368132,"owned_by":"system"},{"id":"gpt-4o-mini-2024-07-18","object":"model","created":1721172717,"owned_by":"system"},{"id":"gpt-4o-mini","object":"model","created":1721172741,"owned_by":"system"},{"id":"gpt-4o-2024-08-06","object":"model","created":1722814719,"owned_by":"system"},{"id":"chatgpt-4o-latest","object":"model","created":1723515131,"owned_by":"system"},{"id":"o1-preview-2024-09-12","object":"model","created":1725648865,"owned_by":"system"},{"id":"o1-preview","object":"model","created":1725648897,"owned_by":"system"},{"id":"o1-mini-2024-09-12","object":"model","created":1725648979,"owned_by":"system"},{"id":"o1-mini","object":"model","created":1725649008,"owned_by":"system"},{"id":"gpt-4o-realtime-preview-2024-10-01","object":"model","created":1727131766,"owned_by":"system"},{"id":"gpt-4o-audio-preview-2024-10-01","object":"model","created":1727389042,"owned_by":"system"},{"id":"gpt-4o-audio-preview","object":"model","created":1727460443,"owned_by":"system"},{"id":"gpt-4o-realtime-preview","object":"model","created":1727659998,"owned_by":"system"},{"id":"omni-moderation-latest","object":"model","created":1731689265,"owned_by":"system"},{"id":"omni-moderation-2024-09-26","object":"model","created":1732734466,"owned_by":"system"},{"id":"gpt-4o-realtime-preview-2024-12-17","object":"model","created":1733945430,"owned_by":"system"},{"id":"gpt-4o-audio-preview-2024-12-17","object":"model","created":1734034239,"owned_by":"system"},{"id":"gpt-4o-mini-realtime-preview-2024-12-17","object":"model","created":1734112601,"owned_by":"system"},{"id":"gpt-4o-mini-audio-preview-2024-12-17","object":"model","created":1734115920,"owned_by":"system"},{"id":"o1-2024-12-17","object":"model","created":1734326976,"owned_by":"system"},{"id":"o1","object":"model","created":1734375816,"owned_by":"system"},{"id":"gpt-4o-mini-realtime-preview","object":"model","created":1734387380,"owned_by":"system"},{"id":"gpt-4o-mini-audio-preview","object":"model","created":1734387424,"owned_by":"system"},{"id":"o3-mini","object":"model","created":1737146383,"owned_by":"system"},{"id":"o3-mini-2025-01-31","object":"model","created":1738010200,"owned_by":"system"},{"id":"gpt-4o-2024-11-20","object":"model","created":1739331543,"owned_by":"system"},{"id":"gpt-4.5-preview","object":"model","created":1740623059,"owned_by":"system"},{"id":"gpt-4.5-preview-2025-02-27","object":"model","created":1740623304,"owned_by":"system"},{"id":"gpt-4o-search-preview-2025-03-11","object":"model","created":1741388170,"owned_by":"system"},{"id":"gpt-4o-search-preview","object":"model","created":1741388720,"owned_by":"system"},{"id":"gpt-4o-mini-search-preview-2025-03-11","object":"model","created":1741390858,"owned_by":"system"},{"id":"gpt-4o-mini-search-preview","object":"model","created":1741391161,"owned_by":"system"},{"id":"gpt-4o-transcribe","object":"model","created":1742068463,"owned_by":"system"},{"id":"gpt-4o-mini-transcribe","object":"model","created":1742068596,"owned_by":"system"},{"id":"o1-pro-2025-03-19","object":"model","created":1742251504,"owned_by":"system"},{"id":"o1-pro","object":"model","created":1742251791,"owned_by":"system"},{"id":"gpt-4o-mini-tts","object":"model","created":1742403959,"owned_by":"system"},{"id":"o4-mini-2025-04-16","object":"model","created":1744133506,"owned_by":"system"},{"id":"o4-mini","object":"model","created":1744225351,"owned_by":"system"},{"id":"gpt-4.1-2025-04-14","object":"model","created":1744315746,"owned_by":"system"},{"id":"gpt-4.1","object":"model","created":1744316542,"owned_by":"system"},{"id":"gpt-4.1-mini-2025-04-14","object":"model","created":1744317547,"owned_by":"system"},{"id":"gpt-4.1-mini","object":"model","created":1744318173,"owned_by":"system"},{"id":"gpt-4.1-nano-2025-04-14","object":"model","created":1744321025,"owned_by":"system"},{"id":"gpt-3.5-turbo-16k","object":"model","created":1683758102,"owned_by":"openai-internal"},{"id":"tts-1","object":"model","created":1681940951,"owned_by":"openai-internal"},{"id":"whisper-1","object":"model","created":1677532384,"owned_by":"openai-internal"},{"id":"text-embedding-ada-002","object":"model","created":1671217299,"owned_by":"openai-internal"}]
\ No newline at end of file
+export const models = [{"id":"gpt-4-0613","object":"model","created":1686588896,"owned_by":"openai"},{"id":"gpt-4","object":"model","created":1687882411,"owned_by":"openai"},{"id":"gpt-3.5-turbo","object":"model","created":1677610602,"owned_by":"openai"},{"id":"o4-mini-deep-research-2025-06-26","object":"model","created":1750866121,"owned_by":"system"},{"id":"codex-mini-latest","object":"model","created":1746673257,"owned_by":"system"},{"id":"gpt-4o-realtime-preview-2025-06-03","object":"model","created":1748907838,"owned_by":"system"},{"id":"gpt-4o-audio-preview-2025-06-03","object":"model","created":1748908498,"owned_by":"system"},{"id":"o4-mini-deep-research","object":"model","created":1749685485,"owned_by":"system"},{"id":"davinci-002","object":"model","created":1692634301,"owned_by":"system"},{"id":"babbage-002","object":"model","created":1692634615,"owned_by":"system"},{"id":"gpt-3.5-turbo-instruct","object":"model","created":1692901427,"owned_by":"system"},{"id":"gpt-3.5-turbo-instruct-0914","object":"model","created":1694122472,"owned_by":"system"},{"id":"dall-e-3","object":"model","created":1698785189,"owned_by":"system"},{"id":"dall-e-2","object":"model","created":1698798177,"owned_by":"system"},{"id":"gpt-4-1106-preview","object":"model","created":1698957206,"owned_by":"system"},{"id":"gpt-3.5-turbo-1106","object":"model","created":1698959748,"owned_by":"system"},{"id":"tts-1-hd","object":"model","created":1699046015,"owned_by":"system"},{"id":"tts-1-1106","object":"model","created":1699053241,"owned_by":"system"},{"id":"tts-1-hd-1106","object":"model","created":1699053533,"owned_by":"system"},{"id":"text-embedding-3-small","object":"model","created":1705948997,"owned_by":"system"},{"id":"text-embedding-3-large","object":"model","created":1705953180,"owned_by":"system"},{"id":"gpt-4-0125-preview","object":"model","created":1706037612,"owned_by":"system"},{"id":"gpt-4-turbo-preview","object":"model","created":1706037777,"owned_by":"system"},{"id":"gpt-3.5-turbo-0125","object":"model","created":1706048358,"owned_by":"system"},{"id":"gpt-4-turbo","object":"model","created":1712361441,"owned_by":"system"},{"id":"gpt-4-turbo-2024-04-09","object":"model","created":1712601677,"owned_by":"system"},{"id":"gpt-4o","object":"model","created":1715367049,"owned_by":"system"},{"id":"gpt-4o-2024-05-13","object":"model","created":1715368132,"owned_by":"system"},{"id":"gpt-4o-mini-2024-07-18","object":"model","created":1721172717,"owned_by":"system"},{"id":"gpt-4o-mini","object":"model","created":1721172741,"owned_by":"system"},{"id":"gpt-4o-2024-08-06","object":"model","created":1722814719,"owned_by":"system"},{"id":"chatgpt-4o-latest","object":"model","created":1723515131,"owned_by":"system"},{"id":"o1-preview-2024-09-12","object":"model","created":1725648865,"owned_by":"system"},{"id":"o1-preview","object":"model","created":1725648897,"owned_by":"system"},{"id":"o1-mini-2024-09-12","object":"model","created":1725648979,"owned_by":"system"},{"id":"o1-mini","object":"model","created":1725649008,"owned_by":"system"},{"id":"gpt-4o-realtime-preview-2024-10-01","object":"model","created":1727131766,"owned_by":"system"},{"id":"gpt-4o-audio-preview-2024-10-01","object":"model","created":1727389042,"owned_by":"system"},{"id":"gpt-4o-audio-preview","object":"model","created":1727460443,"owned_by":"system"},{"id":"gpt-4o-realtime-preview","object":"model","created":1727659998,"owned_by":"system"},{"id":"omni-moderation-latest","object":"model","created":1731689265,"owned_by":"system"},{"id":"omni-moderation-2024-09-26","object":"model","created":1732734466,"owned_by":"system"},{"id":"gpt-4o-realtime-preview-2024-12-17","object":"model","created":1733945430,"owned_by":"system"},{"id":"gpt-4o-audio-preview-2024-12-17","object":"model","created":1734034239,"owned_by":"system"},{"id":"gpt-4o-mini-realtime-preview-2024-12-17","object":"model","created":1734112601,"owned_by":"system"},{"id":"gpt-4o-mini-audio-preview-2024-12-17","object":"model","created":1734115920,"owned_by":"system"},{"id":"o1-2024-12-17","object":"model","created":1734326976,"owned_by":"system"},{"id":"o1","object":"model","created":1734375816,"owned_by":"system"},{"id":"gpt-4o-mini-realtime-preview","object":"model","created":1734387380,"owned_by":"system"},{"id":"gpt-4o-mini-audio-preview","object":"model","created":1734387424,"owned_by":"system"},{"id":"o3-mini","object":"model","created":1737146383,"owned_by":"system"},{"id":"o3-mini-2025-01-31","object":"model","created":1738010200,"owned_by":"system"},{"id":"gpt-4o-2024-11-20","object":"model","created":1739331543,"owned_by":"system"},{"id":"gpt-4.5-preview","object":"model","created":1740623059,"owned_by":"system"},{"id":"gpt-4.5-preview-2025-02-27","object":"model","created":1740623304,"owned_by":"system"},{"id":"gpt-4o-search-preview-2025-03-11","object":"model","created":1741388170,"owned_by":"system"},{"id":"gpt-4o-search-preview","object":"model","created":1741388720,"owned_by":"system"},{"id":"gpt-4o-mini-search-preview-2025-03-11","object":"model","created":1741390858,"owned_by":"system"},{"id":"gpt-4o-mini-search-preview","object":"model","created":1741391161,"owned_by":"system"},{"id":"gpt-4o-transcribe","object":"model","created":1742068463,"owned_by":"system"},{"id":"gpt-4o-mini-transcribe","object":"model","created":1742068596,"owned_by":"system"},{"id":"o1-pro-2025-03-19","object":"model","created":1742251504,"owned_by":"system"},{"id":"o1-pro","object":"model","created":1742251791,"owned_by":"system"},{"id":"gpt-4o-mini-tts","object":"model","created":1742403959,"owned_by":"system"},{"id":"o4-mini-2025-04-16","object":"model","created":1744133506,"owned_by":"system"},{"id":"o4-mini","object":"model","created":1744225351,"owned_by":"system"},{"id":"gpt-4.1-2025-04-14","object":"model","created":1744315746,"owned_by":"system"},{"id":"gpt-4.1","object":"model","created":1744316542,"owned_by":"system"},{"id":"gpt-4.1-mini-2025-04-14","object":"model","created":1744317547,"owned_by":"system"},{"id":"gpt-4.1-mini","object":"model","created":1744318173,"owned_by":"system"},{"id":"gpt-4.1-nano-2025-04-14","object":"model","created":1744321025,"owned_by":"system"},{"id":"gpt-4.1-nano","object":"model","created":1744321707,"owned_by":"system"},{"id":"gpt-image-1","object":"model","created":1745517030,"owned_by":"system"},{"id":"gpt-3.5-turbo-16k","object":"model","created":1683758102,"owned_by":"openai-internal"},{"id":"tts-1","object":"model","created":1681940951,"owned_by":"openai-internal"},{"id":"whisper-1","object":"model","created":1677532384,"owned_by":"openai-internal"},{"id":"text-embedding-ada-002","object":"model","created":1671217299,"owned_by":"openai-internal"}]
\ No newline at end of file
diff --git a/packages/kbot/dist-in/src/models/cache/openrouter.ts b/packages/kbot/dist-in/src/models/cache/openrouter.ts
index 351bf81a..9e3ef56b 100644
--- a/packages/kbot/dist-in/src/models/cache/openrouter.ts
+++ b/packages/kbot/dist-in/src/models/cache/openrouter.ts
@@ -1 +1 @@
-export const models = [{"id":"google/gemini-2.5-pro-preview","name":"Google: Gemini 2.5 Pro Preview 06-05","pricing":{"prompt":"0.00000125","completion":"0.00001","request":"0","image":"0.00516","web_search":"0","internal_reasoning":"0","input_cache_read":"0.00000031","input_cache_write":"0.000001625"},"created":1749137257,"top_provider":{"context_length":1048576,"max_completion_tokens":65536,"is_moderated":false}},{"id":"sentientagi/dobby-mini-unhinged-plus-llama-3.1-8b","name":"SentientAGI: Dobby Mini Plus Llama 3.1 8B","pricing":{"prompt":"0.0000002","completion":"0.0000002","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1748885619,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"deepseek/deepseek-r1-distill-qwen-7b","name":"DeepSeek: R1 Distill Qwen 7B","pricing":{"prompt":"0.0000001","completion":"0.0000002","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1748628237,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"deepseek/deepseek-r1-0528-qwen3-8b:free","name":"DeepSeek: Deepseek R1 0528 Qwen3 8B (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1748538543,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"deepseek/deepseek-r1-0528-qwen3-8b","name":"DeepSeek: Deepseek R1 0528 Qwen3 8B","pricing":{"prompt":"0.00000006","completion":"0.00000009","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1748538543,"top_provider":{"context_length":128000,"max_completion_tokens":32000,"is_moderated":false}},{"id":"google/gemma-2b-it","name":"Google: Gemma 1 2B","pricing":{"prompt":"0.0000001","completion":"0.0000001","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1748460815,"top_provider":{"context_length":8192,"max_completion_tokens":null,"is_moderated":false}},{"id":"deepseek/deepseek-r1-0528:free","name":"DeepSeek: R1 0528 (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1748455170,"top_provider":{"context_length":163840,"max_completion_tokens":null,"is_moderated":false}},{"id":"deepseek/deepseek-r1-0528","name":"DeepSeek: R1 0528","pricing":{"prompt":"0.0000005","completion":"0.00000215","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1748455170,"top_provider":{"context_length":128000,"max_completion_tokens":32768,"is_moderated":false}},{"id":"sarvamai/sarvam-m:free","name":"Sarvam AI: Sarvam-M (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1748188413,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"sarvamai/sarvam-m","name":"Sarvam AI: Sarvam-M","pricing":{"prompt":"0.00000025","completion":"0.00000075","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1748188413,"top_provider":{"context_length":32768,"max_completion_tokens":32768,"is_moderated":false}},{"id":"thedrummer/valkyrie-49b-v1","name":"TheDrummer: Valkyrie 49B V1","pricing":{"prompt":"0.0000005","completion":"0.0000008","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1748022670,"top_provider":{"context_length":131072,"max_completion_tokens":131072,"is_moderated":false}},{"id":"anthropic/claude-opus-4","name":"Anthropic: Claude Opus 4","pricing":{"prompt":"0.000015","completion":"0.000075","request":"0","image":"0.024","web_search":"0","internal_reasoning":"0","input_cache_read":"0.0000015","input_cache_write":"0.00001875"},"created":1747931245,"top_provider":{"context_length":200000,"max_completion_tokens":32000,"is_moderated":true}},{"id":"anthropic/claude-sonnet-4","name":"Anthropic: Claude Sonnet 4","pricing":{"prompt":"0.000003","completion":"0.000015","request":"0","image":"0.0048","web_search":"0","internal_reasoning":"0","input_cache_read":"0.0000003","input_cache_write":"0.00000375"},"created":1747930371,"top_provider":{"context_length":200000,"max_completion_tokens":64000,"is_moderated":false}},{"id":"mistralai/devstral-small:free","name":"Mistral: Devstral Small (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1747837379,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"mistralai/devstral-small","name":"Mistral: Devstral Small","pricing":{"prompt":"0.00000006","completion":"0.00000012","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1747837379,"top_provider":{"context_length":128000,"max_completion_tokens":null,"is_moderated":false}},{"id":"google/gemma-3n-e4b-it:free","name":"Google: Gemma 3n 4B (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1747776824,"top_provider":{"context_length":8192,"max_completion_tokens":2048,"is_moderated":false}},{"id":"google/gemini-2.5-flash-preview-05-20","name":"Google: Gemini 2.5 Flash Preview 05-20","pricing":{"prompt":"0.00000015","completion":"0.0000006","request":"0","image":"0.0006192","web_search":"0","internal_reasoning":"0","input_cache_read":"0.0000000375","input_cache_write":"0.0000002333"},"created":1747761924,"top_provider":{"context_length":1048576,"max_completion_tokens":65535,"is_moderated":false}},{"id":"google/gemini-2.5-flash-preview-05-20:thinking","name":"Google: Gemini 2.5 Flash Preview 05-20 (thinking)","pricing":{"prompt":"0.00000015","completion":"0.0000035","request":"0","image":"0.0006192","web_search":"0","internal_reasoning":"0","input_cache_read":"0.0000000375","input_cache_write":"0.0000002333"},"created":1747761924,"top_provider":{"context_length":1048576,"max_completion_tokens":65535,"is_moderated":false}},{"id":"openai/codex-mini","name":"OpenAI: Codex Mini","pricing":{"prompt":"0.0000015","completion":"0.000006","request":"0","image":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.000000375"},"created":1747409761,"top_provider":{"context_length":200000,"max_completion_tokens":100000,"is_moderated":true}},{"id":"meta-llama/llama-3.3-8b-instruct:free","name":"Meta: Llama 3.3 8B Instruct (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1747230154,"top_provider":{"context_length":128000,"max_completion_tokens":4028,"is_moderated":false}},{"id":"nousresearch/deephermes-3-mistral-24b-preview:free","name":"Nous: DeepHermes 3 Mistral 24B Preview (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1746830904,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"mistralai/mistral-medium-3","name":"Mistral: Mistral Medium 3","pricing":{"prompt":"0.0000004","completion":"0.000002","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1746627341,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"google/gemini-2.5-pro-preview-05-06","name":"Google: Gemini 2.5 Pro Preview 05-06","pricing":{"prompt":"0.00000125","completion":"0.00001","request":"0","image":"0.00516","web_search":"0","internal_reasoning":"0","input_cache_read":"0.00000031","input_cache_write":"0.000001625"},"created":1746578513,"top_provider":{"context_length":1048576,"max_completion_tokens":65535,"is_moderated":false}},{"id":"arcee-ai/caller-large","name":"Arcee AI: Caller Large","pricing":{"prompt":"0.00000055","completion":"0.00000085","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1746487869,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"arcee-ai/spotlight","name":"Arcee AI: Spotlight","pricing":{"prompt":"0.00000018","completion":"0.00000018","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1746481552,"top_provider":{"context_length":131072,"max_completion_tokens":65537,"is_moderated":false}},{"id":"arcee-ai/maestro-reasoning","name":"Arcee AI: Maestro Reasoning","pricing":{"prompt":"0.0000009","completion":"0.0000033","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1746481269,"top_provider":{"context_length":131072,"max_completion_tokens":32000,"is_moderated":false}},{"id":"arcee-ai/virtuoso-large","name":"Arcee AI: Virtuoso Large","pricing":{"prompt":"0.00000075","completion":"0.0000012","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1746478885,"top_provider":{"context_length":131072,"max_completion_tokens":64000,"is_moderated":false}},{"id":"arcee-ai/coder-large","name":"Arcee AI: Coder Large","pricing":{"prompt":"0.0000005","completion":"0.0000008","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1746478663,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"arcee-ai/virtuoso-medium-v2","name":"Arcee AI: Virtuoso Medium V2","pricing":{"prompt":"0.0000005","completion":"0.0000008","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1746478434,"top_provider":{"context_length":131072,"max_completion_tokens":32768,"is_moderated":false}},{"id":"arcee-ai/arcee-blitz","name":"Arcee AI: Arcee Blitz","pricing":{"prompt":"0.00000045","completion":"0.00000075","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1746470100,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"microsoft/phi-4-reasoning-plus:free","name":"Microsoft: Phi 4 Reasoning Plus (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1746130961,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"microsoft/phi-4-reasoning-plus","name":"Microsoft: Phi 4 Reasoning Plus","pricing":{"prompt":"0.00000007","completion":"0.00000035","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1746130961,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"microsoft/phi-4-reasoning:free","name":"Microsoft: Phi 4 Reasoning (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1746121275,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"inception/mercury-coder-small-beta","name":"Inception: Mercury Coder Small Beta","pricing":{"prompt":"0.00000025","completion":"0.000001","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1746033880,"top_provider":{"context_length":32000,"max_completion_tokens":null,"is_moderated":false}},{"id":"opengvlab/internvl3-14b:free","name":"OpenGVLab: InternVL3 14B (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1746021355,"top_provider":{"context_length":32000,"max_completion_tokens":null,"is_moderated":false}},{"id":"opengvlab/internvl3-2b:free","name":"OpenGVLab: InternVL3 2B (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1746019807,"top_provider":{"context_length":32000,"max_completion_tokens":null,"is_moderated":false}},{"id":"deepseek/deepseek-prover-v2:free","name":"DeepSeek: DeepSeek Prover V2 (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1746013094,"top_provider":{"context_length":163840,"max_completion_tokens":null,"is_moderated":false}},{"id":"deepseek/deepseek-prover-v2","name":"DeepSeek: DeepSeek Prover V2","pricing":{"prompt":"0.0000005","completion":"0.00000218","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1746013094,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"meta-llama/llama-guard-4-12b","name":"Meta: Llama Guard 4 12B","pricing":{"prompt":"0.00000005","completion":"0.00000005","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1745975193,"top_provider":{"context_length":163840,"max_completion_tokens":null,"is_moderated":false}},{"id":"qwen/qwen3-30b-a3b:free","name":"Qwen: Qwen3 30B A3B (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1745878604,"top_provider":{"context_length":40960,"max_completion_tokens":null,"is_moderated":false}},{"id":"qwen/qwen3-30b-a3b","name":"Qwen: Qwen3 30B A3B","pricing":{"prompt":"0.00000008","completion":"0.00000029","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1745878604,"top_provider":{"context_length":40960,"max_completion_tokens":40960,"is_moderated":false}},{"id":"qwen/qwen3-8b:free","name":"Qwen: Qwen3 8B (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1745876632,"top_provider":{"context_length":40960,"max_completion_tokens":40960,"is_moderated":false}},{"id":"qwen/qwen3-8b","name":"Qwen: Qwen3 8B","pricing":{"prompt":"0.000000035","completion":"0.000000138","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1745876632,"top_provider":{"context_length":128000,"max_completion_tokens":20000,"is_moderated":false}},{"id":"qwen/qwen3-14b:free","name":"Qwen: Qwen3 14B (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1745876478,"top_provider":{"context_length":40960,"max_completion_tokens":null,"is_moderated":false}},{"id":"qwen/qwen3-14b","name":"Qwen: Qwen3 14B","pricing":{"prompt":"0.00000007","completion":"0.00000024","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1745876478,"top_provider":{"context_length":40960,"max_completion_tokens":40960,"is_moderated":false}},{"id":"qwen/qwen3-32b:free","name":"Qwen: Qwen3 32B (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1745875945,"top_provider":{"context_length":40960,"max_completion_tokens":null,"is_moderated":false}},{"id":"qwen/qwen3-32b","name":"Qwen: Qwen3 32B","pricing":{"prompt":"0.0000001","completion":"0.0000003","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1745875945,"top_provider":{"context_length":40960,"max_completion_tokens":null,"is_moderated":false}},{"id":"qwen/qwen3-235b-a22b:free","name":"Qwen: Qwen3 235B A22B (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1745875757,"top_provider":{"context_length":40960,"max_completion_tokens":null,"is_moderated":false}},{"id":"qwen/qwen3-235b-a22b","name":"Qwen: Qwen3 235B A22B","pricing":{"prompt":"0.00000013","completion":"0.0000006","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1745875757,"top_provider":{"context_length":40960,"max_completion_tokens":40960,"is_moderated":false}},{"id":"tngtech/deepseek-r1t-chimera:free","name":"TNG: DeepSeek R1T Chimera (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1745760875,"top_provider":{"context_length":163840,"max_completion_tokens":null,"is_moderated":false}},{"id":"thudm/glm-z1-rumination-32b","name":"THUDM: GLM Z1 Rumination 32B ","pricing":{"prompt":"0.00000024","completion":"0.00000024","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1745601495,"top_provider":{"context_length":32000,"max_completion_tokens":32000,"is_moderated":false}},{"id":"microsoft/mai-ds-r1:free","name":"Microsoft: MAI DS R1 (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1745194100,"top_provider":{"context_length":163840,"max_completion_tokens":null,"is_moderated":false}},{"id":"thudm/glm-z1-32b:free","name":"THUDM: GLM Z1 32B (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1744924148,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"thudm/glm-z1-32b","name":"THUDM: GLM Z1 32B","pricing":{"prompt":"0.00000024","completion":"0.00000024","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1744924148,"top_provider":{"context_length":32000,"max_completion_tokens":32000,"is_moderated":false}},{"id":"thudm/glm-4-32b:free","name":"THUDM: GLM 4 32B (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1744920915,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"thudm/glm-4-32b","name":"THUDM: GLM 4 32B","pricing":{"prompt":"0.00000024","completion":"0.00000024","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1744920915,"top_provider":{"context_length":32000,"max_completion_tokens":32000,"is_moderated":false}},{"id":"google/gemini-2.5-flash-preview","name":"Google: Gemini 2.5 Flash Preview 04-17","pricing":{"prompt":"0.00000015","completion":"0.0000006","request":"0","image":"0.0006192","web_search":"0","internal_reasoning":"0","input_cache_read":"0.0000000375","input_cache_write":"0.0000002333"},"created":1744914667,"top_provider":{"context_length":1048576,"max_completion_tokens":65535,"is_moderated":false}},{"id":"google/gemini-2.5-flash-preview:thinking","name":"Google: Gemini 2.5 Flash Preview 04-17 (thinking)","pricing":{"prompt":"0.00000015","completion":"0.0000035","request":"0","image":"0.0006192","web_search":"0","internal_reasoning":"0","input_cache_read":"0.0000000375","input_cache_write":"0.0000002333"},"created":1744914667,"top_provider":{"context_length":1048576,"max_completion_tokens":65535,"is_moderated":false}},{"id":"openai/o4-mini-high","name":"OpenAI: o4 Mini High","pricing":{"prompt":"0.0000011","completion":"0.0000044","request":"0","image":"0.0008415","web_search":"0","internal_reasoning":"0","input_cache_read":"0.000000275"},"created":1744824212,"top_provider":{"context_length":200000,"max_completion_tokens":100000,"is_moderated":true}},{"id":"openai/o3","name":"OpenAI: o3","pricing":{"prompt":"0.00001","completion":"0.00004","request":"0","image":"0.00765","web_search":"0","internal_reasoning":"0","input_cache_read":"0.0000025"},"created":1744823457,"top_provider":{"context_length":200000,"max_completion_tokens":100000,"is_moderated":true}},{"id":"openai/o4-mini","name":"OpenAI: o4 Mini","pricing":{"prompt":"0.0000011","completion":"0.0000044","request":"0","image":"0.0008415","web_search":"0","internal_reasoning":"0","input_cache_read":"0.000000275"},"created":1744820942,"top_provider":{"context_length":200000,"max_completion_tokens":100000,"is_moderated":true}},{"id":"shisa-ai/shisa-v2-llama3.3-70b:free","name":"Shisa AI: Shisa V2 Llama 3.3 70B (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1744754858,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"openai/gpt-4.1","name":"OpenAI: GPT-4.1","pricing":{"prompt":"0.000002","completion":"0.000008","request":"0","image":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.0000005"},"created":1744651385,"top_provider":{"context_length":1047576,"max_completion_tokens":32768,"is_moderated":true}},{"id":"openai/gpt-4.1-mini","name":"OpenAI: GPT-4.1 Mini","pricing":{"prompt":"0.0000004","completion":"0.0000016","request":"0","image":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.0000001"},"created":1744651381,"top_provider":{"context_length":1047576,"max_completion_tokens":32768,"is_moderated":true}},{"id":"openai/gpt-4.1-nano","name":"OpenAI: GPT-4.1 Nano","pricing":{"prompt":"0.0000001","completion":"0.0000004","request":"0","image":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.000000025"},"created":1744651369,"top_provider":{"context_length":1047576,"max_completion_tokens":32768,"is_moderated":true}},{"id":"eleutherai/llemma_7b","name":"EleutherAI: Llemma 7b","pricing":{"prompt":"0.0000008","completion":"0.0000012","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1744643225,"top_provider":{"context_length":4096,"max_completion_tokens":4096,"is_moderated":false}},{"id":"alfredpros/codellama-7b-instruct-solidity","name":"AlfredPros: CodeLLaMa 7B Instruct Solidity","pricing":{"prompt":"0.0000008","completion":"0.0000012","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1744641874,"top_provider":{"context_length":4096,"max_completion_tokens":4096,"is_moderated":false}},{"id":"arliai/qwq-32b-arliai-rpr-v1:free","name":"ArliAI: QwQ 32B RpR v1 (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1744555982,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"agentica-org/deepcoder-14b-preview:free","name":"Agentica: Deepcoder 14B Preview (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1744555395,"top_provider":{"context_length":96000,"max_completion_tokens":null,"is_moderated":false}},{"id":"moonshotai/kimi-vl-a3b-thinking:free","name":"Moonshot AI: Kimi VL A3B Thinking (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1744304841,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"x-ai/grok-3-mini-beta","name":"xAI: Grok 3 Mini Beta","pricing":{"prompt":"0.0000003","completion":"0.0000005","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1744240195,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"x-ai/grok-3-beta","name":"xAI: Grok 3 Beta","pricing":{"prompt":"0.000003","completion":"0.000015","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1744240068,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"nvidia/llama-3.3-nemotron-super-49b-v1:free","name":"NVIDIA: Llama 3.3 Nemotron Super 49B v1 (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1744119494,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"nvidia/llama-3.3-nemotron-super-49b-v1","name":"NVIDIA: Llama 3.3 Nemotron Super 49B v1","pricing":{"prompt":"0.00000013","completion":"0.0000004","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1744119494,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"nvidia/llama-3.1-nemotron-ultra-253b-v1:free","name":"NVIDIA: Llama 3.1 Nemotron Ultra 253B v1 (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1744115059,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"nvidia/llama-3.1-nemotron-ultra-253b-v1","name":"NVIDIA: Llama 3.1 Nemotron Ultra 253B v1","pricing":{"prompt":"0.0000006","completion":"0.0000018","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1744115059,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"meta-llama/llama-4-maverick:free","name":"Meta: Llama 4 Maverick (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1743881822,"top_provider":{"context_length":128000,"max_completion_tokens":null,"is_moderated":false}},{"id":"meta-llama/llama-4-maverick","name":"Meta: Llama 4 Maverick","pricing":{"prompt":"0.00000015","completion":"0.0000006","request":"0","image":"0.0006684","web_search":"0","internal_reasoning":"0"},"created":1743881822,"top_provider":{"context_length":1048576,"max_completion_tokens":16384,"is_moderated":false}},{"id":"meta-llama/llama-4-scout:free","name":"Meta: Llama 4 Scout (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1743881519,"top_provider":{"context_length":200000,"max_completion_tokens":null,"is_moderated":false}},{"id":"meta-llama/llama-4-scout","name":"Meta: Llama 4 Scout","pricing":{"prompt":"0.00000008","completion":"0.0000003","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1743881519,"top_provider":{"context_length":1048576,"max_completion_tokens":1048576,"is_moderated":false}},{"id":"all-hands/openhands-lm-32b-v0.1","name":"OpenHands LM 32B V0.1","pricing":{"prompt":"0.0000026","completion":"0.0000034","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1743613013,"top_provider":{"context_length":16384,"max_completion_tokens":4096,"is_moderated":false}},{"id":"deepseek/deepseek-v3-base:free","name":"DeepSeek: DeepSeek V3 Base (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1743272023,"top_provider":{"context_length":163840,"max_completion_tokens":null,"is_moderated":false}},{"id":"scb10x/llama3.1-typhoon2-8b-instruct","name":"Typhoon2 8B Instruct","pricing":{"prompt":"0.00000018","completion":"0.00000018","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1743196511,"top_provider":{"context_length":8192,"max_completion_tokens":null,"is_moderated":false}},{"id":"scb10x/llama3.1-typhoon2-70b-instruct","name":"Typhoon2 70B Instruct","pricing":{"prompt":"0.00000088","completion":"0.00000088","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1743196170,"top_provider":{"context_length":8192,"max_completion_tokens":null,"is_moderated":false}},{"id":"qwen/qwen2.5-vl-3b-instruct:free","name":"Qwen: Qwen2.5 VL 3B Instruct (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1743014573,"top_provider":{"context_length":64000,"max_completion_tokens":null,"is_moderated":false}},{"id":"google/gemini-2.5-pro-exp-03-25","name":"Google: Gemini 2.5 Pro Experimental","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1742922099,"top_provider":{"context_length":1048576,"max_completion_tokens":65535,"is_moderated":false}},{"id":"qwen/qwen2.5-vl-32b-instruct:free","name":"Qwen: Qwen2.5 VL 32B Instruct (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1742839838,"top_provider":{"context_length":8192,"max_completion_tokens":null,"is_moderated":false}},{"id":"qwen/qwen2.5-vl-32b-instruct","name":"Qwen: Qwen2.5 VL 32B Instruct","pricing":{"prompt":"0.0000009","completion":"0.0000009","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1742839838,"top_provider":{"context_length":128000,"max_completion_tokens":null,"is_moderated":false}},{"id":"deepseek/deepseek-chat-v3-0324:free","name":"DeepSeek: DeepSeek V3 0324 (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1742824755,"top_provider":{"context_length":163840,"max_completion_tokens":null,"is_moderated":false}},{"id":"deepseek/deepseek-chat-v3-0324","name":"DeepSeek: DeepSeek V3 0324","pricing":{"prompt":"0.0000003","completion":"0.00000088","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1742824755,"top_provider":{"context_length":163840,"max_completion_tokens":null,"is_moderated":false}},{"id":"featherless/qwerky-72b:free","name":"Qwerky 72B (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1742481597,"top_provider":{"context_length":32768,"max_completion_tokens":4096,"is_moderated":false}},{"id":"openai/o1-pro","name":"OpenAI: o1-pro","pricing":{"prompt":"0.00015","completion":"0.0006","request":"0","image":"0.21675","web_search":"0","internal_reasoning":"0"},"created":1742423211,"top_provider":{"context_length":200000,"max_completion_tokens":100000,"is_moderated":true}},{"id":"mistralai/mistral-small-3.1-24b-instruct:free","name":"Mistral: Mistral Small 3.1 24B (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1742238937,"top_provider":{"context_length":96000,"max_completion_tokens":96000,"is_moderated":false}},{"id":"mistralai/mistral-small-3.1-24b-instruct","name":"Mistral: Mistral Small 3.1 24B","pricing":{"prompt":"0.00000005","completion":"0.00000015","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1742238937,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"open-r1/olympiccoder-32b:free","name":"OlympicCoder 32B (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1742077228,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"google/gemma-3-1b-it:free","name":"Google: Gemma 3 1B (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1741963556,"top_provider":{"context_length":32768,"max_completion_tokens":8192,"is_moderated":false}},{"id":"google/gemma-3-4b-it:free","name":"Google: Gemma 3 4B (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1741905510,"top_provider":{"context_length":96000,"max_completion_tokens":8192,"is_moderated":false}},{"id":"google/gemma-3-4b-it","name":"Google: Gemma 3 4B","pricing":{"prompt":"0.00000002","completion":"0.00000004","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1741905510,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"ai21/jamba-1.6-large","name":"AI21: Jamba 1.6 Large","pricing":{"prompt":"0.000002","completion":"0.000008","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1741905173,"top_provider":{"context_length":256000,"max_completion_tokens":4096,"is_moderated":false}},{"id":"ai21/jamba-1.6-mini","name":"AI21: Jamba Mini 1.6","pricing":{"prompt":"0.0000002","completion":"0.0000004","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1741905171,"top_provider":{"context_length":256000,"max_completion_tokens":4096,"is_moderated":false}},{"id":"google/gemma-3-12b-it:free","name":"Google: Gemma 3 12B (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1741902625,"top_provider":{"context_length":96000,"max_completion_tokens":8192,"is_moderated":false}},{"id":"google/gemma-3-12b-it","name":"Google: Gemma 3 12B","pricing":{"prompt":"0.00000005","completion":"0.0000001","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1741902625,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"cohere/command-a","name":"Cohere: Command A","pricing":{"prompt":"0.0000025","completion":"0.00001","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1741894342,"top_provider":{"context_length":256000,"max_completion_tokens":8192,"is_moderated":true}},{"id":"openai/gpt-4o-mini-search-preview","name":"OpenAI: GPT-4o-mini Search Preview","pricing":{"prompt":"0.00000015","completion":"0.0000006","request":"0.0275","image":"0.000217","web_search":"0","internal_reasoning":"0"},"created":1741818122,"top_provider":{"context_length":128000,"max_completion_tokens":16384,"is_moderated":true}},{"id":"openai/gpt-4o-search-preview","name":"OpenAI: GPT-4o Search Preview","pricing":{"prompt":"0.0000025","completion":"0.00001","request":"0.035","image":"0.003613","web_search":"0","internal_reasoning":"0"},"created":1741817949,"top_provider":{"context_length":128000,"max_completion_tokens":16384,"is_moderated":true}},{"id":"rekaai/reka-flash-3:free","name":"Reka: Flash 3 (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1741812813,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"google/gemma-3-27b-it:free","name":"Google: Gemma 3 27B (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1741756359,"top_provider":{"context_length":96000,"max_completion_tokens":8192,"is_moderated":false}},{"id":"google/gemma-3-27b-it","name":"Google: Gemma 3 27B","pricing":{"prompt":"0.0000001","completion":"0.0000002","request":"0","image":"0.0000256","web_search":"0","internal_reasoning":"0"},"created":1741756359,"top_provider":{"context_length":131072,"max_completion_tokens":16384,"is_moderated":false}},{"id":"thedrummer/anubis-pro-105b-v1","name":"TheDrummer: Anubis Pro 105B V1","pricing":{"prompt":"0.0000008","completion":"0.000001","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1741642290,"top_provider":{"context_length":131072,"max_completion_tokens":131072,"is_moderated":false}},{"id":"thedrummer/skyfall-36b-v2","name":"TheDrummer: Skyfall 36B V2","pricing":{"prompt":"0.0000005","completion":"0.0000008","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1741636566,"top_provider":{"context_length":32768,"max_completion_tokens":32768,"is_moderated":false}},{"id":"microsoft/phi-4-multimodal-instruct","name":"Microsoft: Phi 4 Multimodal Instruct","pricing":{"prompt":"0.00000005","completion":"0.0000001","request":"0","image":"0.00017685","web_search":"0","internal_reasoning":"0"},"created":1741396284,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"perplexity/sonar-reasoning-pro","name":"Perplexity: Sonar Reasoning Pro","pricing":{"prompt":"0.000002","completion":"0.000008","request":"0","image":"0","web_search":"0.005","internal_reasoning":"0"},"created":1741313308,"top_provider":{"context_length":128000,"max_completion_tokens":null,"is_moderated":false}},{"id":"perplexity/sonar-pro","name":"Perplexity: Sonar Pro","pricing":{"prompt":"0.000003","completion":"0.000015","request":"0","image":"0","web_search":"0.005","internal_reasoning":"0"},"created":1741312423,"top_provider":{"context_length":200000,"max_completion_tokens":8000,"is_moderated":false}},{"id":"perplexity/sonar-deep-research","name":"Perplexity: Sonar Deep Research","pricing":{"prompt":"0.000002","completion":"0.000008","request":"0","image":"0","web_search":"0.005","internal_reasoning":"0.000003"},"created":1741311246,"top_provider":{"context_length":128000,"max_completion_tokens":null,"is_moderated":false}},{"id":"deepseek/deepseek-r1-zero:free","name":"DeepSeek: DeepSeek R1 Zero (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1741297434,"top_provider":{"context_length":163840,"max_completion_tokens":null,"is_moderated":false}},{"id":"qwen/qwq-32b:free","name":"Qwen: QwQ 32B (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1741208814,"top_provider":{"context_length":40000,"max_completion_tokens":40000,"is_moderated":false}},{"id":"qwen/qwq-32b","name":"Qwen: QwQ 32B","pricing":{"prompt":"0.00000015","completion":"0.0000002","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1741208814,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"moonshotai/moonlight-16b-a3b-instruct:free","name":"Moonshot AI: Moonlight 16B A3B Instruct (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1740719801,"top_provider":{"context_length":8192,"max_completion_tokens":null,"is_moderated":false}},{"id":"nousresearch/deephermes-3-llama-3-8b-preview:free","name":"Nous: DeepHermes 3 Llama 3 8B Preview (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1740719372,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"openai/gpt-4.5-preview","name":"OpenAI: GPT-4.5 (Preview)","pricing":{"prompt":"0.000075","completion":"0.00015","request":"0","image":"0.108375","web_search":"0","internal_reasoning":"0","input_cache_read":"0.0000375"},"created":1740687810,"top_provider":{"context_length":128000,"max_completion_tokens":16384,"is_moderated":true}},{"id":"google/gemini-2.0-flash-lite-001","name":"Google: Gemini 2.0 Flash Lite","pricing":{"prompt":"0.000000075","completion":"0.0000003","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1740506212,"top_provider":{"context_length":1048576,"max_completion_tokens":8192,"is_moderated":false}},{"id":"anthropic/claude-3.7-sonnet","name":"Anthropic: Claude 3.7 Sonnet","pricing":{"prompt":"0.000003","completion":"0.000015","request":"0","image":"0.0048","web_search":"0","internal_reasoning":"0","input_cache_read":"0.0000003","input_cache_write":"0.00000375"},"created":1740422110,"top_provider":{"context_length":200000,"max_completion_tokens":64000,"is_moderated":false}},{"id":"anthropic/claude-3.7-sonnet:beta","name":"Anthropic: Claude 3.7 Sonnet (self-moderated)","pricing":{"prompt":"0.000003","completion":"0.000015","request":"0","image":"0.0048","web_search":"0","internal_reasoning":"0","input_cache_read":"0.0000003","input_cache_write":"0.00000375"},"created":1740422110,"top_provider":{"context_length":200000,"max_completion_tokens":128000,"is_moderated":false}},{"id":"anthropic/claude-3.7-sonnet:thinking","name":"Anthropic: Claude 3.7 Sonnet (thinking)","pricing":{"prompt":"0.000003","completion":"0.000015","request":"0","image":"0.0048","web_search":"0","internal_reasoning":"0","input_cache_read":"0.0000003","input_cache_write":"0.00000375"},"created":1740422110,"top_provider":{"context_length":200000,"max_completion_tokens":128000,"is_moderated":true}},{"id":"perplexity/r1-1776","name":"Perplexity: R1 1776","pricing":{"prompt":"0.000002","completion":"0.000008","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1740004929,"top_provider":{"context_length":128000,"max_completion_tokens":null,"is_moderated":false}},{"id":"mistralai/mistral-saba","name":"Mistral: Saba","pricing":{"prompt":"0.0000002","completion":"0.0000006","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1739803239,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"cognitivecomputations/dolphin3.0-r1-mistral-24b:free","name":"Dolphin3.0 R1 Mistral 24B (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1739462498,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"cognitivecomputations/dolphin3.0-mistral-24b:free","name":"Dolphin3.0 Mistral 24B (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1739462019,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"meta-llama/llama-guard-3-8b","name":"Llama Guard 3 8B","pricing":{"prompt":"0.00000002","completion":"0.00000006","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1739401318,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"openai/o3-mini-high","name":"OpenAI: o3 Mini High","pricing":{"prompt":"0.0000011","completion":"0.0000044","request":"0","image":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.00000055"},"created":1739372611,"top_provider":{"context_length":200000,"max_completion_tokens":100000,"is_moderated":true}},{"id":"deepseek/deepseek-r1-distill-llama-8b","name":"DeepSeek: R1 Distill Llama 8B","pricing":{"prompt":"0.00000004","completion":"0.00000004","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1738937718,"top_provider":{"context_length":32000,"max_completion_tokens":32000,"is_moderated":false}},{"id":"google/gemini-2.0-flash-001","name":"Google: Gemini 2.0 Flash","pricing":{"prompt":"0.0000001","completion":"0.0000004","request":"0","image":"0.0000258","web_search":"0","internal_reasoning":"0","input_cache_read":"0.000000025","input_cache_write":"0.0000001833"},"created":1738769413,"top_provider":{"context_length":1048576,"max_completion_tokens":8192,"is_moderated":false}},{"id":"qwen/qwen-vl-plus","name":"Qwen: Qwen VL Plus","pricing":{"prompt":"0.00000021","completion":"0.00000063","request":"0","image":"0.0002688","web_search":"0","internal_reasoning":"0"},"created":1738731255,"top_provider":{"context_length":7500,"max_completion_tokens":1500,"is_moderated":false}},{"id":"aion-labs/aion-1.0","name":"AionLabs: Aion-1.0","pricing":{"prompt":"0.000004","completion":"0.000008","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1738697557,"top_provider":{"context_length":131072,"max_completion_tokens":32768,"is_moderated":false}},{"id":"aion-labs/aion-1.0-mini","name":"AionLabs: Aion-1.0-Mini","pricing":{"prompt":"0.0000007","completion":"0.0000014","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1738697107,"top_provider":{"context_length":131072,"max_completion_tokens":32768,"is_moderated":false}},{"id":"aion-labs/aion-rp-llama-3.1-8b","name":"AionLabs: Aion-RP 1.0 (8B)","pricing":{"prompt":"0.0000002","completion":"0.0000002","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1738696718,"top_provider":{"context_length":32768,"max_completion_tokens":32768,"is_moderated":false}},{"id":"qwen/qwen-vl-max","name":"Qwen: Qwen VL Max","pricing":{"prompt":"0.0000008","completion":"0.0000032","request":"0","image":"0.001024","web_search":"0","internal_reasoning":"0"},"created":1738434304,"top_provider":{"context_length":7500,"max_completion_tokens":1500,"is_moderated":false}},{"id":"qwen/qwen-turbo","name":"Qwen: Qwen-Turbo","pricing":{"prompt":"0.00000005","completion":"0.0000002","request":"0","image":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.00000002"},"created":1738410974,"top_provider":{"context_length":1000000,"max_completion_tokens":8192,"is_moderated":false}},{"id":"qwen/qwen2.5-vl-72b-instruct:free","name":"Qwen: Qwen2.5 VL 72B Instruct (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1738410311,"top_provider":{"context_length":131072,"max_completion_tokens":2048,"is_moderated":false}},{"id":"qwen/qwen2.5-vl-72b-instruct","name":"Qwen: Qwen2.5 VL 72B Instruct","pricing":{"prompt":"0.00000025","completion":"0.00000075","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1738410311,"top_provider":{"context_length":32000,"max_completion_tokens":null,"is_moderated":false}},{"id":"qwen/qwen-plus","name":"Qwen: Qwen-Plus","pricing":{"prompt":"0.0000004","completion":"0.0000012","request":"0","image":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.00000016"},"created":1738409840,"top_provider":{"context_length":131072,"max_completion_tokens":8192,"is_moderated":false}},{"id":"qwen/qwen-max","name":"Qwen: Qwen-Max ","pricing":{"prompt":"0.0000016","completion":"0.0000064","request":"0","image":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.00000064"},"created":1738402289,"top_provider":{"context_length":32768,"max_completion_tokens":8192,"is_moderated":false}},{"id":"openai/o3-mini","name":"OpenAI: o3 Mini","pricing":{"prompt":"0.0000011","completion":"0.0000044","request":"0","image":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.00000055"},"created":1738351721,"top_provider":{"context_length":200000,"max_completion_tokens":100000,"is_moderated":true}},{"id":"deepseek/deepseek-r1-distill-qwen-1.5b","name":"DeepSeek: R1 Distill Qwen 1.5B","pricing":{"prompt":"0.00000018","completion":"0.00000018","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1738328067,"top_provider":{"context_length":131072,"max_completion_tokens":32768,"is_moderated":false}},{"id":"mistralai/mistral-small-24b-instruct-2501:free","name":"Mistral: Mistral Small 3 (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1738255409,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"mistralai/mistral-small-24b-instruct-2501","name":"Mistral: Mistral Small 3","pricing":{"prompt":"0.00000005","completion":"0.00000011","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1738255409,"top_provider":{"context_length":32768,"max_completion_tokens":32768,"is_moderated":false}},{"id":"deepseek/deepseek-r1-distill-qwen-32b:free","name":"DeepSeek: R1 Distill Qwen 32B (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1738194830,"top_provider":{"context_length":16000,"max_completion_tokens":16000,"is_moderated":false}},{"id":"deepseek/deepseek-r1-distill-qwen-32b","name":"DeepSeek: R1 Distill Qwen 32B","pricing":{"prompt":"0.00000012","completion":"0.00000018","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1738194830,"top_provider":{"context_length":131072,"max_completion_tokens":16384,"is_moderated":false}},{"id":"deepseek/deepseek-r1-distill-qwen-14b:free","name":"DeepSeek: R1 Distill Qwen 14B (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1738193940,"top_provider":{"context_length":64000,"max_completion_tokens":null,"is_moderated":false}},{"id":"deepseek/deepseek-r1-distill-qwen-14b","name":"DeepSeek: R1 Distill Qwen 14B","pricing":{"prompt":"0.00000015","completion":"0.00000015","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1738193940,"top_provider":{"context_length":64000,"max_completion_tokens":32000,"is_moderated":false}},{"id":"perplexity/sonar-reasoning","name":"Perplexity: Sonar Reasoning","pricing":{"prompt":"0.000001","completion":"0.000005","request":"0.005","image":"0","web_search":"0","internal_reasoning":"0"},"created":1738131107,"top_provider":{"context_length":127000,"max_completion_tokens":null,"is_moderated":false}},{"id":"perplexity/sonar","name":"Perplexity: Sonar","pricing":{"prompt":"0.000001","completion":"0.000001","request":"0.005","image":"0","web_search":"0","internal_reasoning":"0"},"created":1738013808,"top_provider":{"context_length":127072,"max_completion_tokens":null,"is_moderated":false}},{"id":"liquid/lfm-7b","name":"Liquid: LFM 7B","pricing":{"prompt":"0.00000001","completion":"0.00000001","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1737806883,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"liquid/lfm-3b","name":"Liquid: LFM 3B","pricing":{"prompt":"0.00000002","completion":"0.00000002","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1737806501,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"deepseek/deepseek-r1-distill-llama-70b:free","name":"DeepSeek: R1 Distill Llama 70B (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1737663169,"top_provider":{"context_length":8192,"max_completion_tokens":4096,"is_moderated":false}},{"id":"deepseek/deepseek-r1-distill-llama-70b","name":"DeepSeek: R1 Distill Llama 70B","pricing":{"prompt":"0.0000001","completion":"0.0000004","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1737663169,"top_provider":{"context_length":131072,"max_completion_tokens":16384,"is_moderated":false}},{"id":"deepseek/deepseek-r1:free","name":"DeepSeek: R1 (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1737381095,"top_provider":{"context_length":163840,"max_completion_tokens":null,"is_moderated":false}},{"id":"deepseek/deepseek-r1","name":"DeepSeek: R1","pricing":{"prompt":"0.00000045","completion":"0.00000215","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1737381095,"top_provider":{"context_length":128000,"max_completion_tokens":32768,"is_moderated":false}},{"id":"minimax/minimax-01","name":"MiniMax: MiniMax-01","pricing":{"prompt":"0.0000002","completion":"0.0000011","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1736915462,"top_provider":{"context_length":1000192,"max_completion_tokens":1000192,"is_moderated":false}},{"id":"mistralai/codestral-2501","name":"Mistral: Codestral 2501","pricing":{"prompt":"0.0000003","completion":"0.0000009","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1736895522,"top_provider":{"context_length":262144,"max_completion_tokens":null,"is_moderated":false}},{"id":"microsoft/phi-4","name":"Microsoft: Phi 4","pricing":{"prompt":"0.00000007","completion":"0.00000014","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1736489872,"top_provider":{"context_length":16384,"max_completion_tokens":16384,"is_moderated":false}},{"id":"deepseek/deepseek-chat:free","name":"DeepSeek: DeepSeek V3 (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1735241320,"top_provider":{"context_length":163840,"max_completion_tokens":null,"is_moderated":false}},{"id":"deepseek/deepseek-chat","name":"DeepSeek: DeepSeek V3","pricing":{"prompt":"0.00000038","completion":"0.00000089","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1735241320,"top_provider":{"context_length":163840,"max_completion_tokens":163840,"is_moderated":false}},{"id":"sao10k/l3.3-euryale-70b","name":"Sao10K: Llama 3.3 Euryale 70B","pricing":{"prompt":"0.0000007","completion":"0.0000008","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1734535928,"top_provider":{"context_length":131072,"max_completion_tokens":16384,"is_moderated":false}},{"id":"openai/o1","name":"OpenAI: o1","pricing":{"prompt":"0.000015","completion":"0.00006","request":"0","image":"0.021675","web_search":"0","internal_reasoning":"0","input_cache_read":"0.0000075"},"created":1734459999,"top_provider":{"context_length":200000,"max_completion_tokens":100000,"is_moderated":true}},{"id":"eva-unit-01/eva-llama-3.33-70b","name":"EVA Llama 3.33 70B","pricing":{"prompt":"0.000004","completion":"0.000006","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1734377303,"top_provider":{"context_length":16384,"max_completion_tokens":4096,"is_moderated":false}},{"id":"x-ai/grok-2-vision-1212","name":"xAI: Grok 2 Vision 1212","pricing":{"prompt":"0.000002","completion":"0.00001","request":"0","image":"0.0036","web_search":"0","internal_reasoning":"0"},"created":1734237338,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"x-ai/grok-2-1212","name":"xAI: Grok 2 1212","pricing":{"prompt":"0.000002","completion":"0.00001","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1734232814,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"cohere/command-r7b-12-2024","name":"Cohere: Command R7B (12-2024)","pricing":{"prompt":"0.0000000375","completion":"0.00000015","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1734158152,"top_provider":{"context_length":128000,"max_completion_tokens":4000,"is_moderated":true}},{"id":"google/gemini-2.0-flash-exp:free","name":"Google: Gemini 2.0 Flash Experimental (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1733937523,"top_provider":{"context_length":1048576,"max_completion_tokens":8192,"is_moderated":false}},{"id":"meta-llama/llama-3.3-70b-instruct:free","name":"Meta: Llama 3.3 70B Instruct (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1733506137,"top_provider":{"context_length":131072,"max_completion_tokens":2048,"is_moderated":false}},{"id":"meta-llama/llama-3.3-70b-instruct","name":"Meta: Llama 3.3 70B Instruct","pricing":{"prompt":"0.00000007","completion":"0.00000025","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1733506137,"top_provider":{"context_length":131072,"max_completion_tokens":16384,"is_moderated":false}},{"id":"amazon/nova-lite-v1","name":"Amazon: Nova Lite 1.0","pricing":{"prompt":"0.00000006","completion":"0.00000024","request":"0","image":"0.00009","web_search":"0","internal_reasoning":"0"},"created":1733437363,"top_provider":{"context_length":300000,"max_completion_tokens":5120,"is_moderated":true}},{"id":"amazon/nova-micro-v1","name":"Amazon: Nova Micro 1.0","pricing":{"prompt":"0.000000035","completion":"0.00000014","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1733437237,"top_provider":{"context_length":128000,"max_completion_tokens":5120,"is_moderated":true}},{"id":"amazon/nova-pro-v1","name":"Amazon: Nova Pro 1.0","pricing":{"prompt":"0.0000008","completion":"0.0000032","request":"0","image":"0.0012","web_search":"0","internal_reasoning":"0"},"created":1733436303,"top_provider":{"context_length":300000,"max_completion_tokens":5120,"is_moderated":true}},{"id":"qwen/qwq-32b-preview","name":"Qwen: QwQ 32B Preview","pricing":{"prompt":"0.0000002","completion":"0.0000002","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1732754541,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"eva-unit-01/eva-qwen-2.5-72b","name":"EVA Qwen2.5 72B","pricing":{"prompt":"0.000004","completion":"0.000006","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1732210606,"top_provider":{"context_length":16384,"max_completion_tokens":4096,"is_moderated":false}},{"id":"openai/gpt-4o-2024-11-20","name":"OpenAI: GPT-4o (2024-11-20)","pricing":{"prompt":"0.0000025","completion":"0.00001","request":"0","image":"0.003613","web_search":"0","internal_reasoning":"0","input_cache_read":"0.00000125"},"created":1732127594,"top_provider":{"context_length":128000,"max_completion_tokens":16384,"is_moderated":true}},{"id":"mistralai/mistral-large-2411","name":"Mistral Large 2411","pricing":{"prompt":"0.000002","completion":"0.000006","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1731978685,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"mistralai/mistral-large-2407","name":"Mistral Large 2407","pricing":{"prompt":"0.000002","completion":"0.000006","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1731978415,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"mistralai/pixtral-large-2411","name":"Mistral: Pixtral Large 2411","pricing":{"prompt":"0.000002","completion":"0.000006","request":"0","image":"0.002888","web_search":"0","internal_reasoning":"0"},"created":1731977388,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"x-ai/grok-vision-beta","name":"xAI: Grok Vision Beta","pricing":{"prompt":"0.000005","completion":"0.000015","request":"0","image":"0.009","web_search":"0","internal_reasoning":"0"},"created":1731976624,"top_provider":{"context_length":8192,"max_completion_tokens":null,"is_moderated":false}},{"id":"infermatic/mn-inferor-12b","name":"Infermatic: Mistral Nemo Inferor 12B","pricing":{"prompt":"0.0000008","completion":"0.0000012","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1731464428,"top_provider":{"context_length":16384,"max_completion_tokens":4096,"is_moderated":false}},{"id":"qwen/qwen-2.5-coder-32b-instruct:free","name":"Qwen2.5 Coder 32B Instruct (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1731368400,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"qwen/qwen-2.5-coder-32b-instruct","name":"Qwen2.5 Coder 32B Instruct","pricing":{"prompt":"0.00000006","completion":"0.00000015","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1731368400,"top_provider":{"context_length":32768,"max_completion_tokens":16384,"is_moderated":false}},{"id":"raifle/sorcererlm-8x22b","name":"SorcererLM 8x22B","pricing":{"prompt":"0.0000045","completion":"0.0000045","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1731105083,"top_provider":{"context_length":16000,"max_completion_tokens":null,"is_moderated":false}},{"id":"eva-unit-01/eva-qwen-2.5-32b","name":"EVA Qwen2.5 32B","pricing":{"prompt":"0.0000026","completion":"0.0000034","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1731104847,"top_provider":{"context_length":16384,"max_completion_tokens":4096,"is_moderated":false}},{"id":"thedrummer/unslopnemo-12b","name":"TheDrummer: UnslopNemo 12B","pricing":{"prompt":"0.00000045","completion":"0.00000045","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1731103448,"top_provider":{"context_length":32000,"max_completion_tokens":16000,"is_moderated":false}},{"id":"anthropic/claude-3.5-haiku:beta","name":"Anthropic: Claude 3.5 Haiku (self-moderated)","pricing":{"prompt":"0.0000008","completion":"0.000004","request":"0","image":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.00000008","input_cache_write":"0.000001"},"created":1730678400,"top_provider":{"context_length":200000,"max_completion_tokens":8192,"is_moderated":false}},{"id":"anthropic/claude-3.5-haiku","name":"Anthropic: Claude 3.5 Haiku","pricing":{"prompt":"0.0000008","completion":"0.000004","request":"0","image":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.00000008","input_cache_write":"0.000001"},"created":1730678400,"top_provider":{"context_length":200000,"max_completion_tokens":8192,"is_moderated":true}},{"id":"anthropic/claude-3.5-haiku-20241022:beta","name":"Anthropic: Claude 3.5 Haiku (2024-10-22) (self-moderated)","pricing":{"prompt":"0.0000008","completion":"0.000004","request":"0","image":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.00000008","input_cache_write":"0.000001"},"created":1730678400,"top_provider":{"context_length":200000,"max_completion_tokens":8192,"is_moderated":false}},{"id":"anthropic/claude-3.5-haiku-20241022","name":"Anthropic: Claude 3.5 Haiku (2024-10-22)","pricing":{"prompt":"0.0000008","completion":"0.000004","request":"0","image":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.00000008","input_cache_write":"0.000001"},"created":1730678400,"top_provider":{"context_length":200000,"max_completion_tokens":8192,"is_moderated":true}},{"id":"neversleep/llama-3.1-lumimaid-70b","name":"NeverSleep: Lumimaid v0.2 70B","pricing":{"prompt":"0.0000025","completion":"0.000003","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1729555200,"top_provider":{"context_length":16384,"max_completion_tokens":2048,"is_moderated":false}},{"id":"anthracite-org/magnum-v4-72b","name":"Magnum v4 72B","pricing":{"prompt":"0.0000025","completion":"0.000003","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1729555200,"top_provider":{"context_length":16384,"max_completion_tokens":1024,"is_moderated":false}},{"id":"anthropic/claude-3.5-sonnet:beta","name":"Anthropic: Claude 3.5 Sonnet (self-moderated)","pricing":{"prompt":"0.000003","completion":"0.000015","request":"0","image":"0.0048","web_search":"0","internal_reasoning":"0","input_cache_read":"0.0000003","input_cache_write":"0.00000375"},"created":1729555200,"top_provider":{"context_length":200000,"max_completion_tokens":8192,"is_moderated":false}},{"id":"anthropic/claude-3.5-sonnet","name":"Anthropic: Claude 3.5 Sonnet","pricing":{"prompt":"0.000003","completion":"0.000015","request":"0","image":"0.0048","web_search":"0","internal_reasoning":"0","input_cache_read":"0.0000003","input_cache_write":"0.00000375"},"created":1729555200,"top_provider":{"context_length":200000,"max_completion_tokens":8192,"is_moderated":true}},{"id":"x-ai/grok-beta","name":"xAI: Grok Beta","pricing":{"prompt":"0.000005","completion":"0.000015","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1729382400,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"mistralai/ministral-8b","name":"Mistral: Ministral 8B","pricing":{"prompt":"0.0000001","completion":"0.0000001","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1729123200,"top_provider":{"context_length":128000,"max_completion_tokens":null,"is_moderated":false}},{"id":"mistralai/ministral-3b","name":"Mistral: Ministral 3B","pricing":{"prompt":"0.00000004","completion":"0.00000004","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1729123200,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"qwen/qwen-2.5-7b-instruct:free","name":"Qwen2.5 7B Instruct (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1729036800,"top_provider":{"context_length":32768,"max_completion_tokens":32768,"is_moderated":false}},{"id":"qwen/qwen-2.5-7b-instruct","name":"Qwen2.5 7B Instruct","pricing":{"prompt":"0.00000004","completion":"0.0000001","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1729036800,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"nvidia/llama-3.1-nemotron-70b-instruct","name":"NVIDIA: Llama 3.1 Nemotron 70B Instruct","pricing":{"prompt":"0.00000012","completion":"0.0000003","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1728950400,"top_provider":{"context_length":131072,"max_completion_tokens":131072,"is_moderated":false}},{"id":"inflection/inflection-3-productivity","name":"Inflection: Inflection 3 Productivity","pricing":{"prompt":"0.0000025","completion":"0.00001","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1728604800,"top_provider":{"context_length":8000,"max_completion_tokens":1024,"is_moderated":false}},{"id":"inflection/inflection-3-pi","name":"Inflection: Inflection 3 Pi","pricing":{"prompt":"0.0000025","completion":"0.00001","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1728604800,"top_provider":{"context_length":8000,"max_completion_tokens":1024,"is_moderated":false}},{"id":"google/gemini-flash-1.5-8b","name":"Google: Gemini 1.5 Flash 8B","pricing":{"prompt":"0.0000000375","completion":"0.00000015","request":"0","image":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.00000001","input_cache_write":"0.0000000583"},"created":1727913600,"top_provider":{"context_length":1000000,"max_completion_tokens":8192,"is_moderated":false}},{"id":"thedrummer/rocinante-12b","name":"TheDrummer: Rocinante 12B","pricing":{"prompt":"0.00000025","completion":"0.0000005","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1727654400,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"anthracite-org/magnum-v2-72b","name":"Magnum v2 72B","pricing":{"prompt":"0.000003","completion":"0.000003","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1727654400,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"liquid/lfm-40b","name":"Liquid: LFM 40B MoE","pricing":{"prompt":"0.00000015","completion":"0.00000015","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1727654400,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"meta-llama/llama-3.2-3b-instruct:free","name":"Meta: Llama 3.2 3B Instruct (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1727222400,"top_provider":{"context_length":20000,"max_completion_tokens":20000,"is_moderated":false}},{"id":"meta-llama/llama-3.2-3b-instruct","name":"Meta: Llama 3.2 3B Instruct","pricing":{"prompt":"0.00000001","completion":"0.00000002","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1727222400,"top_provider":{"context_length":131072,"max_completion_tokens":16384,"is_moderated":false}},{"id":"meta-llama/llama-3.2-1b-instruct:free","name":"Meta: Llama 3.2 1B Instruct (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1727222400,"top_provider":{"context_length":131072,"max_completion_tokens":8192,"is_moderated":false}},{"id":"meta-llama/llama-3.2-1b-instruct","name":"Meta: Llama 3.2 1B Instruct","pricing":{"prompt":"0.000000005","completion":"0.00000001","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1727222400,"top_provider":{"context_length":131072,"max_completion_tokens":16384,"is_moderated":false}},{"id":"meta-llama/llama-3.2-90b-vision-instruct","name":"Meta: Llama 3.2 90B Vision Instruct","pricing":{"prompt":"0.0000012","completion":"0.0000012","request":"0","image":"0.001734","web_search":"0","internal_reasoning":"0"},"created":1727222400,"top_provider":{"context_length":131072,"max_completion_tokens":2048,"is_moderated":false}},{"id":"meta-llama/llama-3.2-11b-vision-instruct:free","name":"Meta: Llama 3.2 11B Vision Instruct (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1727222400,"top_provider":{"context_length":131072,"max_completion_tokens":2048,"is_moderated":false}},{"id":"meta-llama/llama-3.2-11b-vision-instruct","name":"Meta: Llama 3.2 11B Vision Instruct","pricing":{"prompt":"0.000000049","completion":"0.000000049","request":"0","image":"0.00007948","web_search":"0","internal_reasoning":"0"},"created":1727222400,"top_provider":{"context_length":131072,"max_completion_tokens":16384,"is_moderated":false}},{"id":"qwen/qwen-2.5-72b-instruct:free","name":"Qwen2.5 72B Instruct (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1726704000,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"qwen/qwen-2.5-72b-instruct","name":"Qwen2.5 72B Instruct","pricing":{"prompt":"0.00000012","completion":"0.00000039","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1726704000,"top_provider":{"context_length":32768,"max_completion_tokens":16384,"is_moderated":false}},{"id":"neversleep/llama-3.1-lumimaid-8b","name":"NeverSleep: Lumimaid v0.2 8B","pricing":{"prompt":"0.0000002","completion":"0.00000125","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1726358400,"top_provider":{"context_length":32768,"max_completion_tokens":2048,"is_moderated":false}},{"id":"openai/o1-preview","name":"OpenAI: o1-preview","pricing":{"prompt":"0.000015","completion":"0.00006","request":"0","image":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.0000075"},"created":1726099200,"top_provider":{"context_length":128000,"max_completion_tokens":32768,"is_moderated":true}},{"id":"openai/o1-preview-2024-09-12","name":"OpenAI: o1-preview (2024-09-12)","pricing":{"prompt":"0.000015","completion":"0.00006","request":"0","image":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.0000075"},"created":1726099200,"top_provider":{"context_length":128000,"max_completion_tokens":32768,"is_moderated":true}},{"id":"openai/o1-mini","name":"OpenAI: o1-mini","pricing":{"prompt":"0.0000011","completion":"0.0000044","request":"0","image":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.00000055"},"created":1726099200,"top_provider":{"context_length":128000,"max_completion_tokens":65536,"is_moderated":true}},{"id":"openai/o1-mini-2024-09-12","name":"OpenAI: o1-mini (2024-09-12)","pricing":{"prompt":"0.0000011","completion":"0.0000044","request":"0","image":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.00000055"},"created":1726099200,"top_provider":{"context_length":128000,"max_completion_tokens":65536,"is_moderated":true}},{"id":"mistralai/pixtral-12b","name":"Mistral: Pixtral 12B","pricing":{"prompt":"0.0000001","completion":"0.0000001","request":"0","image":"0.0001445","web_search":"0","internal_reasoning":"0"},"created":1725926400,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"cohere/command-r-plus-08-2024","name":"Cohere: Command R+ (08-2024)","pricing":{"prompt":"0.0000025","completion":"0.00001","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1724976000,"top_provider":{"context_length":128000,"max_completion_tokens":4000,"is_moderated":true}},{"id":"cohere/command-r-08-2024","name":"Cohere: Command R (08-2024)","pricing":{"prompt":"0.00000015","completion":"0.0000006","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1724976000,"top_provider":{"context_length":128000,"max_completion_tokens":4000,"is_moderated":true}},{"id":"qwen/qwen-2.5-vl-7b-instruct:free","name":"Qwen: Qwen2.5-VL 7B Instruct (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1724803200,"top_provider":{"context_length":32768,"max_completion_tokens":32768,"is_moderated":false}},{"id":"qwen/qwen-2.5-vl-7b-instruct","name":"Qwen: Qwen2.5-VL 7B Instruct","pricing":{"prompt":"0.0000002","completion":"0.0000002","request":"0","image":"0.0001445","web_search":"0","internal_reasoning":"0"},"created":1724803200,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"sao10k/l3.1-euryale-70b","name":"Sao10K: Llama 3.1 Euryale 70B v2.2","pricing":{"prompt":"0.0000007","completion":"0.0000008","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1724803200,"top_provider":{"context_length":131072,"max_completion_tokens":16384,"is_moderated":false}},{"id":"microsoft/phi-3.5-mini-128k-instruct","name":"Microsoft: Phi-3.5 Mini 128K Instruct","pricing":{"prompt":"0.0000001","completion":"0.0000001","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1724198400,"top_provider":{"context_length":128000,"max_completion_tokens":null,"is_moderated":false}},{"id":"nousresearch/hermes-3-llama-3.1-70b","name":"Nous: Hermes 3 70B Instruct","pricing":{"prompt":"0.00000012","completion":"0.0000003","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1723939200,"top_provider":{"context_length":131072,"max_completion_tokens":131072,"is_moderated":false}},{"id":"nousresearch/hermes-3-llama-3.1-405b","name":"Nous: Hermes 3 405B Instruct","pricing":{"prompt":"0.0000007","completion":"0.0000008","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1723766400,"top_provider":{"context_length":131072,"max_completion_tokens":16384,"is_moderated":false}},{"id":"openai/chatgpt-4o-latest","name":"OpenAI: ChatGPT-4o","pricing":{"prompt":"0.000005","completion":"0.000015","request":"0","image":"0.007225","web_search":"0","internal_reasoning":"0"},"created":1723593600,"top_provider":{"context_length":128000,"max_completion_tokens":16384,"is_moderated":true}},{"id":"sao10k/l3-lunaris-8b","name":"Sao10K: Llama 3 8B Lunaris","pricing":{"prompt":"0.00000002","completion":"0.00000005","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1723507200,"top_provider":{"context_length":8192,"max_completion_tokens":null,"is_moderated":false}},{"id":"aetherwiing/mn-starcannon-12b","name":"Aetherwiing: Starcannon 12B","pricing":{"prompt":"0.0000008","completion":"0.0000012","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1723507200,"top_provider":{"context_length":16384,"max_completion_tokens":4096,"is_moderated":false}},{"id":"openai/gpt-4o-2024-08-06","name":"OpenAI: GPT-4o (2024-08-06)","pricing":{"prompt":"0.0000025","completion":"0.00001","request":"0","image":"0.003613","web_search":"0","internal_reasoning":"0"},"created":1722902400,"top_provider":{"context_length":128000,"max_completion_tokens":16384,"is_moderated":false}},{"id":"meta-llama/llama-3.1-405b:free","name":"Meta: Llama 3.1 405B (base) (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1722556800,"top_provider":{"context_length":64000,"max_completion_tokens":null,"is_moderated":false}},{"id":"meta-llama/llama-3.1-405b","name":"Meta: Llama 3.1 405B (base)","pricing":{"prompt":"0.000002","completion":"0.000002","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1722556800,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"nothingiisreal/mn-celeste-12b","name":"Mistral Nemo 12B Celeste","pricing":{"prompt":"0.0000008","completion":"0.0000012","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1722556800,"top_provider":{"context_length":16384,"max_completion_tokens":4096,"is_moderated":false}},{"id":"perplexity/llama-3.1-sonar-small-128k-online","name":"Perplexity: Llama 3.1 Sonar 8B Online","pricing":{"prompt":"0.0000002","completion":"0.0000002","request":"0.005","image":"0","web_search":"0","internal_reasoning":"0"},"created":1722470400,"top_provider":{"context_length":127072,"max_completion_tokens":null,"is_moderated":false}},{"id":"perplexity/llama-3.1-sonar-large-128k-online","name":"Perplexity: Llama 3.1 Sonar 70B Online","pricing":{"prompt":"0.000001","completion":"0.000001","request":"0.005","image":"0","web_search":"0","internal_reasoning":"0"},"created":1722470400,"top_provider":{"context_length":127072,"max_completion_tokens":null,"is_moderated":false}},{"id":"meta-llama/llama-3.1-8b-instruct:free","name":"Meta: Llama 3.1 8B Instruct (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1721692800,"top_provider":{"context_length":131072,"max_completion_tokens":4096,"is_moderated":false}},{"id":"meta-llama/llama-3.1-8b-instruct","name":"Meta: Llama 3.1 8B Instruct","pricing":{"prompt":"0.00000002","completion":"0.00000003","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1721692800,"top_provider":{"context_length":16384,"max_completion_tokens":16384,"is_moderated":false}},{"id":"meta-llama/llama-3.1-405b-instruct","name":"Meta: Llama 3.1 405B Instruct","pricing":{"prompt":"0.0000008","completion":"0.0000008","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1721692800,"top_provider":{"context_length":32768,"max_completion_tokens":16384,"is_moderated":false}},{"id":"meta-llama/llama-3.1-70b-instruct","name":"Meta: Llama 3.1 70B Instruct","pricing":{"prompt":"0.0000001","completion":"0.00000028","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1721692800,"top_provider":{"context_length":131072,"max_completion_tokens":16384,"is_moderated":false}},{"id":"mistralai/mistral-nemo:free","name":"Mistral: Mistral Nemo (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1721347200,"top_provider":{"context_length":131072,"max_completion_tokens":128000,"is_moderated":false}},{"id":"mistralai/mistral-nemo","name":"Mistral: Mistral Nemo","pricing":{"prompt":"0.00000001","completion":"0.00000005","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1721347200,"top_provider":{"context_length":131072,"max_completion_tokens":131072,"is_moderated":false}},{"id":"openai/gpt-4o-mini","name":"OpenAI: GPT-4o-mini","pricing":{"prompt":"0.00000015","completion":"0.0000006","request":"0","image":"0.000217","web_search":"0","internal_reasoning":"0","input_cache_read":"0.000000075"},"created":1721260800,"top_provider":{"context_length":128000,"max_completion_tokens":16384,"is_moderated":true}},{"id":"openai/gpt-4o-mini-2024-07-18","name":"OpenAI: GPT-4o-mini (2024-07-18)","pricing":{"prompt":"0.00000015","completion":"0.0000006","request":"0","image":"0.007225","web_search":"0","internal_reasoning":"0","input_cache_read":"0.000000075"},"created":1721260800,"top_provider":{"context_length":128000,"max_completion_tokens":16384,"is_moderated":true}},{"id":"google/gemma-2-27b-it","name":"Google: Gemma 2 27B","pricing":{"prompt":"0.0000008","completion":"0.0000008","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1720828800,"top_provider":{"context_length":8192,"max_completion_tokens":2048,"is_moderated":false}},{"id":"alpindale/magnum-72b","name":"Magnum 72B","pricing":{"prompt":"0.000004","completion":"0.000006","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1720656000,"top_provider":{"context_length":16384,"max_completion_tokens":4096,"is_moderated":false}},{"id":"google/gemma-2-9b-it:free","name":"Google: Gemma 2 9B (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1719532800,"top_provider":{"context_length":8192,"max_completion_tokens":8192,"is_moderated":false}},{"id":"google/gemma-2-9b-it","name":"Google: Gemma 2 9B","pricing":{"prompt":"0.0000002","completion":"0.0000002","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1719532800,"top_provider":{"context_length":8192,"max_completion_tokens":8192,"is_moderated":false}},{"id":"01-ai/yi-large","name":"01.AI: Yi Large","pricing":{"prompt":"0.000003","completion":"0.000003","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1719273600,"top_provider":{"context_length":32768,"max_completion_tokens":4096,"is_moderated":false}},{"id":"anthropic/claude-3.5-sonnet-20240620:beta","name":"Anthropic: Claude 3.5 Sonnet (2024-06-20) (self-moderated)","pricing":{"prompt":"0.000003","completion":"0.000015","request":"0","image":"0.0048","web_search":"0","internal_reasoning":"0","input_cache_read":"0.0000003","input_cache_write":"0.00000375"},"created":1718841600,"top_provider":{"context_length":200000,"max_completion_tokens":8192,"is_moderated":false}},{"id":"anthropic/claude-3.5-sonnet-20240620","name":"Anthropic: Claude 3.5 Sonnet (2024-06-20)","pricing":{"prompt":"0.000003","completion":"0.000015","request":"0","image":"0.0048","web_search":"0","internal_reasoning":"0","input_cache_read":"0.0000003","input_cache_write":"0.00000375"},"created":1718841600,"top_provider":{"context_length":200000,"max_completion_tokens":8192,"is_moderated":true}},{"id":"sao10k/l3-euryale-70b","name":"Sao10k: Llama 3 Euryale 70B v2.1","pricing":{"prompt":"0.00000148","completion":"0.00000148","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1718668800,"top_provider":{"context_length":8192,"max_completion_tokens":8192,"is_moderated":false}},{"id":"cognitivecomputations/dolphin-mixtral-8x22b","name":"Dolphin 2.9.2 Mixtral 8x22B 🐬","pricing":{"prompt":"0.0000009","completion":"0.0000009","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1717804800,"top_provider":{"context_length":16000,"max_completion_tokens":8192,"is_moderated":false}},{"id":"qwen/qwen-2-72b-instruct","name":"Qwen 2 72B Instruct","pricing":{"prompt":"0.0000009","completion":"0.0000009","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1717718400,"top_provider":{"context_length":32768,"max_completion_tokens":4096,"is_moderated":false}},{"id":"mistralai/mistral-7b-instruct:free","name":"Mistral: Mistral 7B Instruct (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1716768000,"top_provider":{"context_length":32768,"max_completion_tokens":16384,"is_moderated":false}},{"id":"mistralai/mistral-7b-instruct","name":"Mistral: Mistral 7B Instruct","pricing":{"prompt":"0.000000028","completion":"0.000000054","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1716768000,"top_provider":{"context_length":32768,"max_completion_tokens":16384,"is_moderated":false}},{"id":"nousresearch/hermes-2-pro-llama-3-8b","name":"NousResearch: Hermes 2 Pro - Llama-3 8B","pricing":{"prompt":"0.000000025","completion":"0.00000004","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1716768000,"top_provider":{"context_length":131072,"max_completion_tokens":131072,"is_moderated":false}},{"id":"mistralai/mistral-7b-instruct-v0.3","name":"Mistral: Mistral 7B Instruct v0.3","pricing":{"prompt":"0.000000028","completion":"0.000000054","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1716768000,"top_provider":{"context_length":32768,"max_completion_tokens":16384,"is_moderated":false}},{"id":"microsoft/phi-3-mini-128k-instruct","name":"Microsoft: Phi-3 Mini 128K Instruct","pricing":{"prompt":"0.0000001","completion":"0.0000001","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1716681600,"top_provider":{"context_length":128000,"max_completion_tokens":null,"is_moderated":false}},{"id":"microsoft/phi-3-medium-128k-instruct","name":"Microsoft: Phi-3 Medium 128K Instruct","pricing":{"prompt":"0.000001","completion":"0.000001","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1716508800,"top_provider":{"context_length":128000,"max_completion_tokens":null,"is_moderated":false}},{"id":"neversleep/llama-3-lumimaid-70b","name":"NeverSleep: Llama 3 Lumimaid 70B","pricing":{"prompt":"0.000004","completion":"0.000006","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1715817600,"top_provider":{"context_length":8192,"max_completion_tokens":4096,"is_moderated":false}},{"id":"google/gemini-flash-1.5","name":"Google: Gemini 1.5 Flash ","pricing":{"prompt":"0.000000075","completion":"0.0000003","request":"0","image":"0.00004","web_search":"0","internal_reasoning":"0","input_cache_read":"0.00000001875","input_cache_write":"0.0000001583"},"created":1715644800,"top_provider":{"context_length":1000000,"max_completion_tokens":8192,"is_moderated":false}},{"id":"openai/gpt-4o","name":"OpenAI: GPT-4o","pricing":{"prompt":"0.0000025","completion":"0.00001","request":"0","image":"0.003613","web_search":"0","internal_reasoning":"0","input_cache_read":"0.00000125"},"created":1715558400,"top_provider":{"context_length":128000,"max_completion_tokens":16384,"is_moderated":true}},{"id":"openai/gpt-4o:extended","name":"OpenAI: GPT-4o (extended)","pricing":{"prompt":"0.000006","completion":"0.000018","request":"0","image":"0.007225","web_search":"0","internal_reasoning":"0"},"created":1715558400,"top_provider":{"context_length":128000,"max_completion_tokens":64000,"is_moderated":true}},{"id":"meta-llama/llama-guard-2-8b","name":"Meta: LlamaGuard 2 8B","pricing":{"prompt":"0.0000002","completion":"0.0000002","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1715558400,"top_provider":{"context_length":8192,"max_completion_tokens":null,"is_moderated":false}},{"id":"openai/gpt-4o-2024-05-13","name":"OpenAI: GPT-4o (2024-05-13)","pricing":{"prompt":"0.000005","completion":"0.000015","request":"0","image":"0.007225","web_search":"0","internal_reasoning":"0"},"created":1715558400,"top_provider":{"context_length":128000,"max_completion_tokens":4096,"is_moderated":true}},{"id":"neversleep/llama-3-lumimaid-8b","name":"NeverSleep: Llama 3 Lumimaid 8B","pricing":{"prompt":"0.0000002","completion":"0.00000125","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1714780800,"top_provider":{"context_length":24576,"max_completion_tokens":2048,"is_moderated":false}},{"id":"sao10k/fimbulvetr-11b-v2","name":"Fimbulvetr 11B v2","pricing":{"prompt":"0.0000008","completion":"0.0000012","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1713657600,"top_provider":{"context_length":4096,"max_completion_tokens":4096,"is_moderated":false}},{"id":"meta-llama/llama-3-8b-instruct","name":"Meta: Llama 3 8B Instruct","pricing":{"prompt":"0.00000003","completion":"0.00000006","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1713398400,"top_provider":{"context_length":8192,"max_completion_tokens":16384,"is_moderated":false}},{"id":"meta-llama/llama-3-70b-instruct","name":"Meta: Llama 3 70B Instruct","pricing":{"prompt":"0.0000003","completion":"0.0000004","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1713398400,"top_provider":{"context_length":8192,"max_completion_tokens":16384,"is_moderated":false}},{"id":"mistralai/mixtral-8x22b-instruct","name":"Mistral: Mixtral 8x22B Instruct","pricing":{"prompt":"0.0000009","completion":"0.0000009","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1713312000,"top_provider":{"context_length":65536,"max_completion_tokens":null,"is_moderated":false}},{"id":"microsoft/wizardlm-2-8x22b","name":"WizardLM-2 8x22B","pricing":{"prompt":"0.00000048","completion":"0.00000048","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1713225600,"top_provider":{"context_length":65536,"max_completion_tokens":65536,"is_moderated":false}},{"id":"google/gemini-pro-1.5","name":"Google: Gemini 1.5 Pro","pricing":{"prompt":"0.00000125","completion":"0.000005","request":"0","image":"0.0006575","web_search":"0","internal_reasoning":"0"},"created":1712620800,"top_provider":{"context_length":2000000,"max_completion_tokens":8192,"is_moderated":false}},{"id":"openai/gpt-4-turbo","name":"OpenAI: GPT-4 Turbo","pricing":{"prompt":"0.00001","completion":"0.00003","request":"0","image":"0.01445","web_search":"0","internal_reasoning":"0"},"created":1712620800,"top_provider":{"context_length":128000,"max_completion_tokens":4096,"is_moderated":true}},{"id":"cohere/command-r-plus","name":"Cohere: Command R+","pricing":{"prompt":"0.000003","completion":"0.000015","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1712188800,"top_provider":{"context_length":128000,"max_completion_tokens":4000,"is_moderated":true}},{"id":"cohere/command-r-plus-04-2024","name":"Cohere: Command R+ (04-2024)","pricing":{"prompt":"0.000003","completion":"0.000015","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1712016000,"top_provider":{"context_length":128000,"max_completion_tokens":4000,"is_moderated":true}},{"id":"sophosympatheia/midnight-rose-70b","name":"Midnight Rose 70B","pricing":{"prompt":"0.0000008","completion":"0.0000008","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1711065600,"top_provider":{"context_length":4096,"max_completion_tokens":2048,"is_moderated":false}},{"id":"cohere/command","name":"Cohere: Command","pricing":{"prompt":"0.000001","completion":"0.000002","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1710374400,"top_provider":{"context_length":4096,"max_completion_tokens":4000,"is_moderated":true}},{"id":"cohere/command-r","name":"Cohere: Command R","pricing":{"prompt":"0.0000005","completion":"0.0000015","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1710374400,"top_provider":{"context_length":128000,"max_completion_tokens":4000,"is_moderated":true}},{"id":"anthropic/claude-3-haiku:beta","name":"Anthropic: Claude 3 Haiku (self-moderated)","pricing":{"prompt":"0.00000025","completion":"0.00000125","request":"0","image":"0.0004","web_search":"0","internal_reasoning":"0","input_cache_read":"0.00000003","input_cache_write":"0.0000003"},"created":1710288000,"top_provider":{"context_length":200000,"max_completion_tokens":4096,"is_moderated":false}},{"id":"anthropic/claude-3-haiku","name":"Anthropic: Claude 3 Haiku","pricing":{"prompt":"0.00000025","completion":"0.00000125","request":"0","image":"0.0004","web_search":"0","internal_reasoning":"0","input_cache_read":"0.00000003","input_cache_write":"0.0000003"},"created":1710288000,"top_provider":{"context_length":200000,"max_completion_tokens":4096,"is_moderated":true}},{"id":"anthropic/claude-3-opus:beta","name":"Anthropic: Claude 3 Opus (self-moderated)","pricing":{"prompt":"0.000015","completion":"0.000075","request":"0","image":"0.024","web_search":"0","internal_reasoning":"0","input_cache_read":"0.0000015","input_cache_write":"0.00001875"},"created":1709596800,"top_provider":{"context_length":200000,"max_completion_tokens":4096,"is_moderated":false}},{"id":"anthropic/claude-3-opus","name":"Anthropic: Claude 3 Opus","pricing":{"prompt":"0.000015","completion":"0.000075","request":"0","image":"0.024","web_search":"0","internal_reasoning":"0","input_cache_read":"0.0000015","input_cache_write":"0.00001875"},"created":1709596800,"top_provider":{"context_length":200000,"max_completion_tokens":4096,"is_moderated":true}},{"id":"anthropic/claude-3-sonnet:beta","name":"Anthropic: Claude 3 Sonnet (self-moderated)","pricing":{"prompt":"0.000003","completion":"0.000015","request":"0","image":"0.0048","web_search":"0","internal_reasoning":"0","input_cache_read":"0.0000003","input_cache_write":"0.00000375"},"created":1709596800,"top_provider":{"context_length":200000,"max_completion_tokens":4096,"is_moderated":false}},{"id":"anthropic/claude-3-sonnet","name":"Anthropic: Claude 3 Sonnet","pricing":{"prompt":"0.000003","completion":"0.000015","request":"0","image":"0.0048","web_search":"0","internal_reasoning":"0","input_cache_read":"0.0000003","input_cache_write":"0.00000375"},"created":1709596800,"top_provider":{"context_length":200000,"max_completion_tokens":4096,"is_moderated":true}},{"id":"cohere/command-r-03-2024","name":"Cohere: Command R (03-2024)","pricing":{"prompt":"0.0000005","completion":"0.0000015","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1709341200,"top_provider":{"context_length":128000,"max_completion_tokens":4000,"is_moderated":true}},{"id":"mistralai/mistral-large","name":"Mistral Large","pricing":{"prompt":"0.000002","completion":"0.000006","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1708905600,"top_provider":{"context_length":128000,"max_completion_tokens":null,"is_moderated":false}},{"id":"openai/gpt-3.5-turbo-0613","name":"OpenAI: GPT-3.5 Turbo (older v0613)","pricing":{"prompt":"0.000001","completion":"0.000002","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1706140800,"top_provider":{"context_length":4095,"max_completion_tokens":4096,"is_moderated":false}},{"id":"openai/gpt-4-turbo-preview","name":"OpenAI: GPT-4 Turbo Preview","pricing":{"prompt":"0.00001","completion":"0.00003","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1706140800,"top_provider":{"context_length":128000,"max_completion_tokens":4096,"is_moderated":true}},{"id":"nousresearch/nous-hermes-2-mixtral-8x7b-dpo","name":"Nous: Hermes 2 Mixtral 8x7B DPO","pricing":{"prompt":"0.0000006","completion":"0.0000006","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1705363200,"top_provider":{"context_length":32768,"max_completion_tokens":2048,"is_moderated":false}},{"id":"mistralai/mistral-medium","name":"Mistral Medium","pricing":{"prompt":"0.00000275","completion":"0.0000081","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1704844800,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"mistralai/mistral-small","name":"Mistral Small","pricing":{"prompt":"0.0000002","completion":"0.0000006","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1704844800,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"mistralai/mistral-tiny","name":"Mistral Tiny","pricing":{"prompt":"0.00000025","completion":"0.00000025","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1704844800,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"mistralai/mistral-7b-instruct-v0.2","name":"Mistral: Mistral 7B Instruct v0.2","pricing":{"prompt":"0.0000002","completion":"0.0000002","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1703721600,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"mistralai/mixtral-8x7b-instruct","name":"Mistral: Mixtral 8x7B Instruct","pricing":{"prompt":"0.00000008","completion":"0.00000024","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1702166400,"top_provider":{"context_length":32768,"max_completion_tokens":16384,"is_moderated":false}},{"id":"neversleep/noromaid-20b","name":"Noromaid 20B","pricing":{"prompt":"0.00000125","completion":"0.000002","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1700956800,"top_provider":{"context_length":8192,"max_completion_tokens":2048,"is_moderated":false}},{"id":"anthropic/claude-2.1:beta","name":"Anthropic: Claude v2.1 (self-moderated)","pricing":{"prompt":"0.000008","completion":"0.000024","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1700611200,"top_provider":{"context_length":200000,"max_completion_tokens":4096,"is_moderated":false}},{"id":"anthropic/claude-2.1","name":"Anthropic: Claude v2.1","pricing":{"prompt":"0.000008","completion":"0.000024","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1700611200,"top_provider":{"context_length":200000,"max_completion_tokens":4096,"is_moderated":true}},{"id":"anthropic/claude-2:beta","name":"Anthropic: Claude v2 (self-moderated)","pricing":{"prompt":"0.000008","completion":"0.000024","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1700611200,"top_provider":{"context_length":200000,"max_completion_tokens":4096,"is_moderated":false}},{"id":"anthropic/claude-2","name":"Anthropic: Claude v2","pricing":{"prompt":"0.000008","completion":"0.000024","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1700611200,"top_provider":{"context_length":200000,"max_completion_tokens":4096,"is_moderated":true}},{"id":"undi95/toppy-m-7b","name":"Toppy M 7B","pricing":{"prompt":"0.0000008","completion":"0.0000012","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1699574400,"top_provider":{"context_length":4096,"max_completion_tokens":4096,"is_moderated":false}},{"id":"alpindale/goliath-120b","name":"Goliath 120B","pricing":{"prompt":"0.00001","completion":"0.0000125","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1699574400,"top_provider":{"context_length":6144,"max_completion_tokens":512,"is_moderated":false}},{"id":"openrouter/auto","name":"Auto Router","pricing":{"prompt":"-1","completion":"-1"},"created":1699401600,"top_provider":{"context_length":null,"max_completion_tokens":null,"is_moderated":false}},{"id":"openai/gpt-3.5-turbo-1106","name":"OpenAI: GPT-3.5 Turbo 16k (older v1106)","pricing":{"prompt":"0.000001","completion":"0.000002","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1699228800,"top_provider":{"context_length":16385,"max_completion_tokens":4096,"is_moderated":true}},{"id":"openai/gpt-4-1106-preview","name":"OpenAI: GPT-4 Turbo (older v1106)","pricing":{"prompt":"0.00001","completion":"0.00003","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1699228800,"top_provider":{"context_length":128000,"max_completion_tokens":4096,"is_moderated":true}},{"id":"openai/gpt-3.5-turbo-instruct","name":"OpenAI: GPT-3.5 Turbo Instruct","pricing":{"prompt":"0.0000015","completion":"0.000002","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1695859200,"top_provider":{"context_length":4095,"max_completion_tokens":4096,"is_moderated":true}},{"id":"mistralai/mistral-7b-instruct-v0.1","name":"Mistral: Mistral 7B Instruct v0.1","pricing":{"prompt":"0.00000011","completion":"0.00000019","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1695859200,"top_provider":{"context_length":2824,"max_completion_tokens":null,"is_moderated":false}},{"id":"pygmalionai/mythalion-13b","name":"Pygmalion: Mythalion 13B","pricing":{"prompt":"0.0000008","completion":"0.0000012","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1693612800,"top_provider":{"context_length":4096,"max_completion_tokens":4096,"is_moderated":false}},{"id":"openai/gpt-3.5-turbo-16k","name":"OpenAI: GPT-3.5 Turbo 16k","pricing":{"prompt":"0.000003","completion":"0.000004","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1693180800,"top_provider":{"context_length":16385,"max_completion_tokens":4096,"is_moderated":true}},{"id":"mancer/weaver","name":"Mancer: Weaver (alpha)","pricing":{"prompt":"0.0000015","completion":"0.0000015","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1690934400,"top_provider":{"context_length":8000,"max_completion_tokens":1000,"is_moderated":false}},{"id":"anthropic/claude-2.0:beta","name":"Anthropic: Claude v2.0 (self-moderated)","pricing":{"prompt":"0.000008","completion":"0.000024","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1690502400,"top_provider":{"context_length":100000,"max_completion_tokens":4096,"is_moderated":false}},{"id":"anthropic/claude-2.0","name":"Anthropic: Claude v2.0","pricing":{"prompt":"0.000008","completion":"0.000024","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1690502400,"top_provider":{"context_length":100000,"max_completion_tokens":4096,"is_moderated":true}},{"id":"undi95/remm-slerp-l2-13b","name":"ReMM SLERP 13B","pricing":{"prompt":"0.0000008","completion":"0.0000012","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1689984000,"top_provider":{"context_length":4096,"max_completion_tokens":4096,"is_moderated":false}},{"id":"gryphe/mythomax-l2-13b","name":"MythoMax 13B","pricing":{"prompt":"0.000000065","completion":"0.000000065","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1688256000,"top_provider":{"context_length":4096,"max_completion_tokens":4096,"is_moderated":false}},{"id":"meta-llama/llama-2-70b-chat","name":"Meta: Llama 2 70B Chat","pricing":{"prompt":"0.0000009","completion":"0.0000009","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1687219200,"top_provider":{"context_length":4096,"max_completion_tokens":null,"is_moderated":false}},{"id":"openai/gpt-3.5-turbo","name":"OpenAI: GPT-3.5 Turbo","pricing":{"prompt":"0.0000005","completion":"0.0000015","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1685232000,"top_provider":{"context_length":16385,"max_completion_tokens":4096,"is_moderated":true}},{"id":"openai/gpt-3.5-turbo-0125","name":"OpenAI: GPT-3.5 Turbo 16k","pricing":{"prompt":"0.0000005","completion":"0.0000015","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1685232000,"top_provider":{"context_length":16385,"max_completion_tokens":4096,"is_moderated":true}},{"id":"openai/gpt-4","name":"OpenAI: GPT-4","pricing":{"prompt":"0.00003","completion":"0.00006","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1685232000,"top_provider":{"context_length":8191,"max_completion_tokens":4096,"is_moderated":true}},{"id":"openai/gpt-4-0314","name":"OpenAI: GPT-4 (older v0314)","pricing":{"prompt":"0.00003","completion":"0.00006","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1685232000,"top_provider":{"context_length":8191,"max_completion_tokens":4096,"is_moderated":true}}]
\ No newline at end of file
+export const models = [{"id":"inception/mercury","name":"Inception: Mercury","pricing":{"prompt":"0.00001","completion":"0.00001","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1750973026,"top_provider":{"context_length":32000,"max_completion_tokens":16000,"is_moderated":false}},{"id":"morph/morph-v2","name":"Morph: Morph V2","pricing":{"prompt":"0.0000012","completion":"0.0000027","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1750946108,"top_provider":{"context_length":16000,"max_completion_tokens":16000,"is_moderated":false}},{"id":"mistralai/mistral-small-3.2-24b-instruct:free","name":"Mistral: Mistral Small 3.2 24B (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1750443016,"top_provider":{"context_length":96000,"max_completion_tokens":null,"is_moderated":false}},{"id":"mistralai/mistral-small-3.2-24b-instruct","name":"Mistral: Mistral Small 3.2 24B","pricing":{"prompt":"0.00000005","completion":"0.0000001","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1750443016,"top_provider":{"context_length":128000,"max_completion_tokens":null,"is_moderated":false}},{"id":"minimax/minimax-m1:extended","name":"MiniMax: MiniMax M1 (extended)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1750200414,"top_provider":{"context_length":256000,"max_completion_tokens":null,"is_moderated":false}},{"id":"minimax/minimax-m1","name":"MiniMax: MiniMax M1","pricing":{"prompt":"0.0000003","completion":"0.00000165","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1750200414,"top_provider":{"context_length":1000000,"max_completion_tokens":40000,"is_moderated":false}},{"id":"google/gemini-2.5-flash-lite-preview-06-17","name":"Google: Gemini 2.5 Flash Lite Preview 06-17","pricing":{"prompt":"0.0000001","completion":"0.0000004","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1750173831,"top_provider":{"context_length":1048576,"max_completion_tokens":65535,"is_moderated":false}},{"id":"google/gemini-2.5-flash","name":"Google: Gemini 2.5 Flash","pricing":{"prompt":"0.0000003","completion":"0.0000025","request":"0","image":"0.001238","web_search":"0","internal_reasoning":"0","input_cache_read":"0.000000075","input_cache_write":"0.0000003833"},"created":1750172488,"top_provider":{"context_length":1048576,"max_completion_tokens":65535,"is_moderated":false}},{"id":"google/gemini-2.5-pro","name":"Google: Gemini 2.5 Pro","pricing":{"prompt":"0.00000125","completion":"0.00001","request":"0","image":"0.00516","web_search":"0","internal_reasoning":"0","input_cache_read":"0.00000031","input_cache_write":"0.000001625"},"created":1750169544,"top_provider":{"context_length":1048576,"max_completion_tokens":65536,"is_moderated":false}},{"id":"moonshotai/kimi-dev-72b:free","name":"Kimi Dev 72b (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1750115909,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"openai/o3-pro","name":"OpenAI: o3 Pro","pricing":{"prompt":"0.00002","completion":"0.00008","request":"0","image":"0.0153","web_search":"0","internal_reasoning":"0"},"created":1749598352,"top_provider":{"context_length":200000,"max_completion_tokens":100000,"is_moderated":true}},{"id":"x-ai/grok-3-mini","name":"xAI: Grok 3 Mini","pricing":{"prompt":"0.0000003","completion":"0.0000005","request":"0","image":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.000000075"},"created":1749583245,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"x-ai/grok-3","name":"xAI: Grok 3","pricing":{"prompt":"0.000003","completion":"0.000015","request":"0","image":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.00000075"},"created":1749582908,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"mistralai/magistral-small-2506","name":"Mistral: Magistral Small 2506","pricing":{"prompt":"0.0000005","completion":"0.0000015","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1749569561,"top_provider":{"context_length":40000,"max_completion_tokens":40000,"is_moderated":false}},{"id":"mistralai/magistral-medium-2506","name":"Mistral: Magistral Medium 2506","pricing":{"prompt":"0.000002","completion":"0.000005","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1749354054,"top_provider":{"context_length":40960,"max_completion_tokens":40000,"is_moderated":false}},{"id":"mistralai/magistral-medium-2506:thinking","name":"Mistral: Magistral Medium 2506 (thinking)","pricing":{"prompt":"0.000002","completion":"0.000005","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1749354054,"top_provider":{"context_length":40960,"max_completion_tokens":40000,"is_moderated":false}},{"id":"google/gemini-2.5-pro-preview","name":"Google: Gemini 2.5 Pro Preview 06-05","pricing":{"prompt":"0.00000125","completion":"0.00001","request":"0","image":"0.00516","web_search":"0","internal_reasoning":"0","input_cache_read":"0.00000031","input_cache_write":"0.000001625"},"created":1749137257,"top_provider":{"context_length":1048576,"max_completion_tokens":65536,"is_moderated":false}},{"id":"sentientagi/dobby-mini-unhinged-plus-llama-3.1-8b","name":"SentientAGI: Dobby Mini Plus Llama 3.1 8B","pricing":{"prompt":"0.0000002","completion":"0.0000002","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1748885619,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"deepseek/deepseek-r1-distill-qwen-7b","name":"DeepSeek: R1 Distill Qwen 7B","pricing":{"prompt":"0.0000001","completion":"0.0000002","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1748628237,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"deepseek/deepseek-r1-0528-qwen3-8b:free","name":"DeepSeek: Deepseek R1 0528 Qwen3 8B (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1748538543,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"deepseek/deepseek-r1-0528-qwen3-8b","name":"DeepSeek: Deepseek R1 0528 Qwen3 8B","pricing":{"prompt":"0.00000001","completion":"0.00000002","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1748538543,"top_provider":{"context_length":32000,"max_completion_tokens":null,"is_moderated":false}},{"id":"deepseek/deepseek-r1-0528:free","name":"DeepSeek: R1 0528 (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1748455170,"top_provider":{"context_length":163840,"max_completion_tokens":null,"is_moderated":false}},{"id":"deepseek/deepseek-r1-0528","name":"DeepSeek: R1 0528","pricing":{"prompt":"0.0000005","completion":"0.00000215","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1748455170,"top_provider":{"context_length":128000,"max_completion_tokens":32768,"is_moderated":false}},{"id":"sarvamai/sarvam-m:free","name":"Sarvam AI: Sarvam-M (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1748188413,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"thedrummer/valkyrie-49b-v1","name":"TheDrummer: Valkyrie 49B V1","pricing":{"prompt":"0.0000005","completion":"0.0000008","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1748022670,"top_provider":{"context_length":131072,"max_completion_tokens":131072,"is_moderated":false}},{"id":"anthropic/claude-opus-4","name":"Anthropic: Claude Opus 4","pricing":{"prompt":"0.000015","completion":"0.000075","request":"0","image":"0.024","web_search":"0","internal_reasoning":"0","input_cache_read":"0.0000015","input_cache_write":"0.00001875"},"created":1747931245,"top_provider":{"context_length":200000,"max_completion_tokens":32000,"is_moderated":true}},{"id":"anthropic/claude-sonnet-4","name":"Anthropic: Claude Sonnet 4","pricing":{"prompt":"0.000003","completion":"0.000015","request":"0","image":"0.0048","web_search":"0","internal_reasoning":"0","input_cache_read":"0.0000003","input_cache_write":"0.00000375"},"created":1747930371,"top_provider":{"context_length":200000,"max_completion_tokens":64000,"is_moderated":false}},{"id":"mistralai/devstral-small:free","name":"Mistral: Devstral Small (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1747837379,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"mistralai/devstral-small","name":"Mistral: Devstral Small","pricing":{"prompt":"0.00000006","completion":"0.00000012","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1747837379,"top_provider":{"context_length":128000,"max_completion_tokens":null,"is_moderated":false}},{"id":"google/gemma-3n-e4b-it:free","name":"Google: Gemma 3n 4B (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1747776824,"top_provider":{"context_length":8192,"max_completion_tokens":2048,"is_moderated":false}},{"id":"google/gemma-3n-e4b-it","name":"Google: Gemma 3n 4B","pricing":{"prompt":"0.00000002","completion":"0.00000004","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1747776824,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"google/gemini-2.5-flash-preview-05-20","name":"Google: Gemini 2.5 Flash Preview 05-20","pricing":{"prompt":"0.00000015","completion":"0.0000006","request":"0","image":"0.0006192","web_search":"0","internal_reasoning":"0","input_cache_read":"0.0000000375","input_cache_write":"0.0000002333"},"created":1747761924,"top_provider":{"context_length":1048576,"max_completion_tokens":65535,"is_moderated":false}},{"id":"google/gemini-2.5-flash-preview-05-20:thinking","name":"Google: Gemini 2.5 Flash Preview 05-20 (thinking)","pricing":{"prompt":"0.00000015","completion":"0.0000035","request":"0","image":"0.0006192","web_search":"0","internal_reasoning":"0","input_cache_read":"0.0000000375","input_cache_write":"0.0000002333"},"created":1747761924,"top_provider":{"context_length":1048576,"max_completion_tokens":65535,"is_moderated":false}},{"id":"openai/codex-mini","name":"OpenAI: Codex Mini","pricing":{"prompt":"0.0000015","completion":"0.000006","request":"0","image":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.000000375"},"created":1747409761,"top_provider":{"context_length":200000,"max_completion_tokens":100000,"is_moderated":true}},{"id":"mistralai/mistral-medium-3","name":"Mistral: Mistral Medium 3","pricing":{"prompt":"0.0000004","completion":"0.000002","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1746627341,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"google/gemini-2.5-pro-preview-05-06","name":"Google: Gemini 2.5 Pro Preview 05-06","pricing":{"prompt":"0.00000125","completion":"0.00001","request":"0","image":"0.00516","web_search":"0","internal_reasoning":"0","input_cache_read":"0.00000031","input_cache_write":"0.000001625"},"created":1746578513,"top_provider":{"context_length":1048576,"max_completion_tokens":65535,"is_moderated":false}},{"id":"arcee-ai/caller-large","name":"Arcee AI: Caller Large","pricing":{"prompt":"0.00000055","completion":"0.00000085","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1746487869,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"arcee-ai/spotlight","name":"Arcee AI: Spotlight","pricing":{"prompt":"0.00000018","completion":"0.00000018","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1746481552,"top_provider":{"context_length":131072,"max_completion_tokens":65537,"is_moderated":false}},{"id":"arcee-ai/maestro-reasoning","name":"Arcee AI: Maestro Reasoning","pricing":{"prompt":"0.0000009","completion":"0.0000033","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1746481269,"top_provider":{"context_length":131072,"max_completion_tokens":32000,"is_moderated":false}},{"id":"arcee-ai/virtuoso-large","name":"Arcee AI: Virtuoso Large","pricing":{"prompt":"0.00000075","completion":"0.0000012","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1746478885,"top_provider":{"context_length":131072,"max_completion_tokens":64000,"is_moderated":false}},{"id":"arcee-ai/coder-large","name":"Arcee AI: Coder Large","pricing":{"prompt":"0.0000005","completion":"0.0000008","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1746478663,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"arcee-ai/virtuoso-medium-v2","name":"Arcee AI: Virtuoso Medium V2","pricing":{"prompt":"0.0000005","completion":"0.0000008","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1746478434,"top_provider":{"context_length":131072,"max_completion_tokens":32768,"is_moderated":false}},{"id":"arcee-ai/arcee-blitz","name":"Arcee AI: Arcee Blitz","pricing":{"prompt":"0.00000045","completion":"0.00000075","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1746470100,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"microsoft/phi-4-reasoning-plus","name":"Microsoft: Phi 4 Reasoning Plus","pricing":{"prompt":"0.00000007","completion":"0.00000035","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1746130961,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"inception/mercury-coder-small-beta","name":"Inception: Mercury Coder Small Beta","pricing":{"prompt":"0.00000025","completion":"0.000001","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1746033880,"top_provider":{"context_length":32000,"max_completion_tokens":null,"is_moderated":false}},{"id":"opengvlab/internvl3-14b","name":"OpenGVLab: InternVL3 14B","pricing":{"prompt":"0.0000002","completion":"0.0000004","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1746021355,"top_provider":{"context_length":12288,"max_completion_tokens":null,"is_moderated":false}},{"id":"opengvlab/internvl3-2b","name":"OpenGVLab: InternVL3 2B","pricing":{"prompt":"0.00000005","completion":"0.0000001","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1746019807,"top_provider":{"context_length":12288,"max_completion_tokens":null,"is_moderated":false}},{"id":"deepseek/deepseek-prover-v2","name":"DeepSeek: DeepSeek Prover V2","pricing":{"prompt":"0.0000005","completion":"0.00000218","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1746013094,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"meta-llama/llama-guard-4-12b","name":"Meta: Llama Guard 4 12B","pricing":{"prompt":"0.00000005","completion":"0.00000005","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1745975193,"top_provider":{"context_length":163840,"max_completion_tokens":null,"is_moderated":false}},{"id":"qwen/qwen3-30b-a3b:free","name":"Qwen: Qwen3 30B A3B (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1745878604,"top_provider":{"context_length":40960,"max_completion_tokens":null,"is_moderated":false}},{"id":"qwen/qwen3-30b-a3b","name":"Qwen: Qwen3 30B A3B","pricing":{"prompt":"0.00000008","completion":"0.00000029","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1745878604,"top_provider":{"context_length":40960,"max_completion_tokens":40960,"is_moderated":false}},{"id":"qwen/qwen3-8b:free","name":"Qwen: Qwen3 8B (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1745876632,"top_provider":{"context_length":40960,"max_completion_tokens":40960,"is_moderated":false}},{"id":"qwen/qwen3-8b","name":"Qwen: Qwen3 8B","pricing":{"prompt":"0.000000035","completion":"0.000000138","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1745876632,"top_provider":{"context_length":128000,"max_completion_tokens":20000,"is_moderated":false}},{"id":"qwen/qwen3-14b:free","name":"Qwen: Qwen3 14B (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1745876478,"top_provider":{"context_length":40960,"max_completion_tokens":null,"is_moderated":false}},{"id":"qwen/qwen3-14b","name":"Qwen: Qwen3 14B","pricing":{"prompt":"0.00000006","completion":"0.00000024","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1745876478,"top_provider":{"context_length":40960,"max_completion_tokens":40960,"is_moderated":false}},{"id":"qwen/qwen3-32b:free","name":"Qwen: Qwen3 32B (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1745875945,"top_provider":{"context_length":40960,"max_completion_tokens":null,"is_moderated":false}},{"id":"qwen/qwen3-32b","name":"Qwen: Qwen3 32B","pricing":{"prompt":"0.0000001","completion":"0.0000003","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1745875945,"top_provider":{"context_length":40960,"max_completion_tokens":null,"is_moderated":false}},{"id":"qwen/qwen3-235b-a22b:free","name":"Qwen: Qwen3 235B A22B (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1745875757,"top_provider":{"context_length":40960,"max_completion_tokens":null,"is_moderated":false}},{"id":"qwen/qwen3-235b-a22b","name":"Qwen: Qwen3 235B A22B","pricing":{"prompt":"0.00000013","completion":"0.0000006","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1745875757,"top_provider":{"context_length":40960,"max_completion_tokens":40960,"is_moderated":false}},{"id":"tngtech/deepseek-r1t-chimera:free","name":"TNG: DeepSeek R1T Chimera (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1745760875,"top_provider":{"context_length":163840,"max_completion_tokens":null,"is_moderated":false}},{"id":"thudm/glm-z1-rumination-32b","name":"THUDM: GLM Z1 Rumination 32B ","pricing":{"prompt":"0.00000024","completion":"0.00000024","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1745601495,"top_provider":{"context_length":32000,"max_completion_tokens":32000,"is_moderated":false}},{"id":"microsoft/mai-ds-r1:free","name":"Microsoft: MAI DS R1 (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1745194100,"top_provider":{"context_length":163840,"max_completion_tokens":null,"is_moderated":false}},{"id":"thudm/glm-z1-32b:free","name":"THUDM: GLM Z1 32B (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1744924148,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"thudm/glm-z1-32b","name":"THUDM: GLM Z1 32B","pricing":{"prompt":"0.00000024","completion":"0.00000024","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1744924148,"top_provider":{"context_length":32000,"max_completion_tokens":32000,"is_moderated":false}},{"id":"thudm/glm-4-32b:free","name":"THUDM: GLM 4 32B (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1744920915,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"thudm/glm-4-32b","name":"THUDM: GLM 4 32B","pricing":{"prompt":"0.00000024","completion":"0.00000024","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1744920915,"top_provider":{"context_length":32000,"max_completion_tokens":32000,"is_moderated":false}},{"id":"google/gemini-2.5-flash-preview","name":"Google: Gemini 2.5 Flash Preview 04-17","pricing":{"prompt":"0.00000015","completion":"0.0000006","request":"0","image":"0.0006192","web_search":"0","internal_reasoning":"0","input_cache_read":"0.0000000375","input_cache_write":"0.0000002333"},"created":1744914667,"top_provider":{"context_length":1048576,"max_completion_tokens":65535,"is_moderated":false}},{"id":"google/gemini-2.5-flash-preview:thinking","name":"Google: Gemini 2.5 Flash Preview 04-17 (thinking)","pricing":{"prompt":"0.00000015","completion":"0.0000035","request":"0","image":"0.0006192","web_search":"0","internal_reasoning":"0","input_cache_read":"0.0000000375","input_cache_write":"0.0000002333"},"created":1744914667,"top_provider":{"context_length":1048576,"max_completion_tokens":65535,"is_moderated":false}},{"id":"openai/o4-mini-high","name":"OpenAI: o4 Mini High","pricing":{"prompt":"0.0000011","completion":"0.0000044","request":"0","image":"0.0008415","web_search":"0","internal_reasoning":"0","input_cache_read":"0.000000275"},"created":1744824212,"top_provider":{"context_length":200000,"max_completion_tokens":100000,"is_moderated":true}},{"id":"openai/o3","name":"OpenAI: o3","pricing":{"prompt":"0.000002","completion":"0.000008","request":"0","image":"0.00153","web_search":"0","internal_reasoning":"0","input_cache_read":"0.0000005"},"created":1744823457,"top_provider":{"context_length":200000,"max_completion_tokens":100000,"is_moderated":true}},{"id":"openai/o4-mini","name":"OpenAI: o4 Mini","pricing":{"prompt":"0.0000011","completion":"0.0000044","request":"0","image":"0.0008415","web_search":"0","internal_reasoning":"0","input_cache_read":"0.000000275"},"created":1744820942,"top_provider":{"context_length":200000,"max_completion_tokens":100000,"is_moderated":true}},{"id":"shisa-ai/shisa-v2-llama3.3-70b:free","name":"Shisa AI: Shisa V2 Llama 3.3 70B (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1744754858,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"openai/gpt-4.1","name":"OpenAI: GPT-4.1","pricing":{"prompt":"0.000002","completion":"0.000008","request":"0","image":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.0000005"},"created":1744651385,"top_provider":{"context_length":1047576,"max_completion_tokens":32768,"is_moderated":true}},{"id":"openai/gpt-4.1-mini","name":"OpenAI: GPT-4.1 Mini","pricing":{"prompt":"0.0000004","completion":"0.0000016","request":"0","image":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.0000001"},"created":1744651381,"top_provider":{"context_length":1047576,"max_completion_tokens":32768,"is_moderated":true}},{"id":"openai/gpt-4.1-nano","name":"OpenAI: GPT-4.1 Nano","pricing":{"prompt":"0.0000001","completion":"0.0000004","request":"0","image":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.000000025"},"created":1744651369,"top_provider":{"context_length":1047576,"max_completion_tokens":32768,"is_moderated":true}},{"id":"eleutherai/llemma_7b","name":"EleutherAI: Llemma 7b","pricing":{"prompt":"0.0000008","completion":"0.0000012","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1744643225,"top_provider":{"context_length":4096,"max_completion_tokens":4096,"is_moderated":false}},{"id":"alfredpros/codellama-7b-instruct-solidity","name":"AlfredPros: CodeLLaMa 7B Instruct Solidity","pricing":{"prompt":"0.0000008","completion":"0.0000012","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1744641874,"top_provider":{"context_length":4096,"max_completion_tokens":4096,"is_moderated":false}},{"id":"arliai/qwq-32b-arliai-rpr-v1:free","name":"ArliAI: QwQ 32B RpR v1 (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1744555982,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"agentica-org/deepcoder-14b-preview:free","name":"Agentica: Deepcoder 14B Preview (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1744555395,"top_provider":{"context_length":96000,"max_completion_tokens":null,"is_moderated":false}},{"id":"moonshotai/kimi-vl-a3b-thinking:free","name":"Moonshot AI: Kimi VL A3B Thinking (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1744304841,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"x-ai/grok-3-mini-beta","name":"xAI: Grok 3 Mini Beta","pricing":{"prompt":"0.0000003","completion":"0.0000005","request":"0","image":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.000000075"},"created":1744240195,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"x-ai/grok-3-beta","name":"xAI: Grok 3 Beta","pricing":{"prompt":"0.000003","completion":"0.000015","request":"0","image":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.00000075"},"created":1744240068,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"nvidia/llama-3.3-nemotron-super-49b-v1:free","name":"NVIDIA: Llama 3.3 Nemotron Super 49B v1 (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1744119494,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"nvidia/llama-3.3-nemotron-super-49b-v1","name":"NVIDIA: Llama 3.3 Nemotron Super 49B v1","pricing":{"prompt":"0.00000013","completion":"0.0000004","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1744119494,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"nvidia/llama-3.1-nemotron-ultra-253b-v1:free","name":"NVIDIA: Llama 3.1 Nemotron Ultra 253B v1 (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1744115059,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"nvidia/llama-3.1-nemotron-ultra-253b-v1","name":"NVIDIA: Llama 3.1 Nemotron Ultra 253B v1","pricing":{"prompt":"0.0000006","completion":"0.0000018","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1744115059,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"meta-llama/llama-4-maverick:free","name":"Meta: Llama 4 Maverick (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1743881822,"top_provider":{"context_length":128000,"max_completion_tokens":null,"is_moderated":false}},{"id":"meta-llama/llama-4-maverick","name":"Meta: Llama 4 Maverick","pricing":{"prompt":"0.00000015","completion":"0.0000006","request":"0","image":"0.0006684","web_search":"0","internal_reasoning":"0"},"created":1743881822,"top_provider":{"context_length":1048576,"max_completion_tokens":16384,"is_moderated":false}},{"id":"meta-llama/llama-4-scout:free","name":"Meta: Llama 4 Scout (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1743881519,"top_provider":{"context_length":96000,"max_completion_tokens":null,"is_moderated":false}},{"id":"meta-llama/llama-4-scout","name":"Meta: Llama 4 Scout","pricing":{"prompt":"0.00000008","completion":"0.0000003","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1743881519,"top_provider":{"context_length":1048576,"max_completion_tokens":1048576,"is_moderated":false}},{"id":"all-hands/openhands-lm-32b-v0.1","name":"OpenHands LM 32B V0.1","pricing":{"prompt":"0.0000026","completion":"0.0000034","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1743613013,"top_provider":{"context_length":16384,"max_completion_tokens":4096,"is_moderated":false}},{"id":"deepseek/deepseek-v3-base:free","name":"DeepSeek: DeepSeek V3 Base (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1743272023,"top_provider":{"context_length":163840,"max_completion_tokens":null,"is_moderated":false}},{"id":"scb10x/llama3.1-typhoon2-70b-instruct","name":"Typhoon2 70B Instruct","pricing":{"prompt":"0.00000088","completion":"0.00000088","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1743196170,"top_provider":{"context_length":8192,"max_completion_tokens":null,"is_moderated":false}},{"id":"google/gemini-2.5-pro-exp-03-25","name":"Google: Gemini 2.5 Pro Experimental","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1742922099,"top_provider":{"context_length":1048576,"max_completion_tokens":65535,"is_moderated":false}},{"id":"qwen/qwen2.5-vl-32b-instruct:free","name":"Qwen: Qwen2.5 VL 32B Instruct (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1742839838,"top_provider":{"context_length":8192,"max_completion_tokens":null,"is_moderated":false}},{"id":"qwen/qwen2.5-vl-32b-instruct","name":"Qwen: Qwen2.5 VL 32B Instruct","pricing":{"prompt":"0.0000009","completion":"0.0000009","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1742839838,"top_provider":{"context_length":128000,"max_completion_tokens":null,"is_moderated":false}},{"id":"deepseek/deepseek-chat-v3-0324:free","name":"DeepSeek: DeepSeek V3 0324 (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1742824755,"top_provider":{"context_length":163840,"max_completion_tokens":null,"is_moderated":false}},{"id":"deepseek/deepseek-chat-v3-0324","name":"DeepSeek: DeepSeek V3 0324","pricing":{"prompt":"0.00000028","completion":"0.00000088","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1742824755,"top_provider":{"context_length":163840,"max_completion_tokens":null,"is_moderated":false}},{"id":"featherless/qwerky-72b:free","name":"Qwerky 72B (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1742481597,"top_provider":{"context_length":32768,"max_completion_tokens":4096,"is_moderated":false}},{"id":"openai/o1-pro","name":"OpenAI: o1-pro","pricing":{"prompt":"0.00015","completion":"0.0006","request":"0","image":"0.21675","web_search":"0","internal_reasoning":"0"},"created":1742423211,"top_provider":{"context_length":200000,"max_completion_tokens":100000,"is_moderated":true}},{"id":"mistralai/mistral-small-3.1-24b-instruct:free","name":"Mistral: Mistral Small 3.1 24B (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1742238937,"top_provider":{"context_length":96000,"max_completion_tokens":96000,"is_moderated":false}},{"id":"mistralai/mistral-small-3.1-24b-instruct","name":"Mistral: Mistral Small 3.1 24B","pricing":{"prompt":"0.00000005","completion":"0.0000001","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1742238937,"top_provider":{"context_length":128000,"max_completion_tokens":null,"is_moderated":false}},{"id":"google/gemma-3-4b-it:free","name":"Google: Gemma 3 4B (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1741905510,"top_provider":{"context_length":32768,"max_completion_tokens":8192,"is_moderated":false}},{"id":"google/gemma-3-4b-it","name":"Google: Gemma 3 4B","pricing":{"prompt":"0.00000002","completion":"0.00000004","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1741905510,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"ai21/jamba-1.6-large","name":"AI21: Jamba 1.6 Large","pricing":{"prompt":"0.000002","completion":"0.000008","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1741905173,"top_provider":{"context_length":256000,"max_completion_tokens":4096,"is_moderated":false}},{"id":"ai21/jamba-1.6-mini","name":"AI21: Jamba Mini 1.6","pricing":{"prompt":"0.0000002","completion":"0.0000004","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1741905171,"top_provider":{"context_length":256000,"max_completion_tokens":4096,"is_moderated":false}},{"id":"google/gemma-3-12b-it:free","name":"Google: Gemma 3 12B (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1741902625,"top_provider":{"context_length":96000,"max_completion_tokens":8192,"is_moderated":false}},{"id":"google/gemma-3-12b-it","name":"Google: Gemma 3 12B","pricing":{"prompt":"0.00000005","completion":"0.0000001","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1741902625,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"cohere/command-a","name":"Cohere: Command A","pricing":{"prompt":"0.0000025","completion":"0.00001","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1741894342,"top_provider":{"context_length":256000,"max_completion_tokens":8192,"is_moderated":true}},{"id":"openai/gpt-4o-mini-search-preview","name":"OpenAI: GPT-4o-mini Search Preview","pricing":{"prompt":"0.00000015","completion":"0.0000006","request":"0.0275","image":"0.000217","web_search":"0","internal_reasoning":"0"},"created":1741818122,"top_provider":{"context_length":128000,"max_completion_tokens":16384,"is_moderated":true}},{"id":"openai/gpt-4o-search-preview","name":"OpenAI: GPT-4o Search Preview","pricing":{"prompt":"0.0000025","completion":"0.00001","request":"0.035","image":"0.003613","web_search":"0","internal_reasoning":"0"},"created":1741817949,"top_provider":{"context_length":128000,"max_completion_tokens":16384,"is_moderated":true}},{"id":"rekaai/reka-flash-3:free","name":"Reka: Flash 3 (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1741812813,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"google/gemma-3-27b-it:free","name":"Google: Gemma 3 27B (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1741756359,"top_provider":{"context_length":96000,"max_completion_tokens":8192,"is_moderated":false}},{"id":"google/gemma-3-27b-it","name":"Google: Gemma 3 27B","pricing":{"prompt":"0.0000001","completion":"0.00000018","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1741756359,"top_provider":{"context_length":64000,"max_completion_tokens":8000,"is_moderated":false}},{"id":"thedrummer/anubis-pro-105b-v1","name":"TheDrummer: Anubis Pro 105B V1","pricing":{"prompt":"0.0000008","completion":"0.000001","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1741642290,"top_provider":{"context_length":131072,"max_completion_tokens":131072,"is_moderated":false}},{"id":"thedrummer/skyfall-36b-v2","name":"TheDrummer: Skyfall 36B V2","pricing":{"prompt":"0.0000005","completion":"0.0000008","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1741636566,"top_provider":{"context_length":32768,"max_completion_tokens":32768,"is_moderated":false}},{"id":"microsoft/phi-4-multimodal-instruct","name":"Microsoft: Phi 4 Multimodal Instruct","pricing":{"prompt":"0.00000005","completion":"0.0000001","request":"0","image":"0.00017685","web_search":"0","internal_reasoning":"0"},"created":1741396284,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"perplexity/sonar-reasoning-pro","name":"Perplexity: Sonar Reasoning Pro","pricing":{"prompt":"0.000002","completion":"0.000008","request":"0","image":"0","web_search":"0.005","internal_reasoning":"0"},"created":1741313308,"top_provider":{"context_length":128000,"max_completion_tokens":null,"is_moderated":false}},{"id":"perplexity/sonar-pro","name":"Perplexity: Sonar Pro","pricing":{"prompt":"0.000003","completion":"0.000015","request":"0","image":"0","web_search":"0.005","internal_reasoning":"0"},"created":1741312423,"top_provider":{"context_length":200000,"max_completion_tokens":8000,"is_moderated":false}},{"id":"perplexity/sonar-deep-research","name":"Perplexity: Sonar Deep Research","pricing":{"prompt":"0.000002","completion":"0.000008","request":"0","image":"0","web_search":"0.005","internal_reasoning":"0.000003"},"created":1741311246,"top_provider":{"context_length":128000,"max_completion_tokens":null,"is_moderated":false}},{"id":"qwen/qwq-32b:free","name":"Qwen: QwQ 32B (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1741208814,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":true}},{"id":"qwen/qwq-32b","name":"Qwen: QwQ 32B","pricing":{"prompt":"0.000000075","completion":"0.00000015","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1741208814,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"nousresearch/deephermes-3-llama-3-8b-preview:free","name":"Nous: DeepHermes 3 Llama 3 8B Preview (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1740719372,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"openai/gpt-4.5-preview","name":"OpenAI: GPT-4.5 (Preview)","pricing":{"prompt":"0.000075","completion":"0.00015","request":"0","image":"0.108375","web_search":"0","internal_reasoning":"0","input_cache_read":"0.0000375"},"created":1740687810,"top_provider":{"context_length":128000,"max_completion_tokens":16384,"is_moderated":true}},{"id":"google/gemini-2.0-flash-lite-001","name":"Google: Gemini 2.0 Flash Lite","pricing":{"prompt":"0.000000075","completion":"0.0000003","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1740506212,"top_provider":{"context_length":1048576,"max_completion_tokens":8192,"is_moderated":false}},{"id":"anthropic/claude-3.7-sonnet","name":"Anthropic: Claude 3.7 Sonnet","pricing":{"prompt":"0.000003","completion":"0.000015","request":"0","image":"0.0048","web_search":"0","internal_reasoning":"0","input_cache_read":"0.0000003","input_cache_write":"0.00000375"},"created":1740422110,"top_provider":{"context_length":200000,"max_completion_tokens":64000,"is_moderated":false}},{"id":"anthropic/claude-3.7-sonnet:beta","name":"Anthropic: Claude 3.7 Sonnet (self-moderated)","pricing":{"prompt":"0.000003","completion":"0.000015","request":"0","image":"0.0048","web_search":"0","internal_reasoning":"0","input_cache_read":"0.0000003","input_cache_write":"0.00000375"},"created":1740422110,"top_provider":{"context_length":200000,"max_completion_tokens":128000,"is_moderated":false}},{"id":"anthropic/claude-3.7-sonnet:thinking","name":"Anthropic: Claude 3.7 Sonnet (thinking)","pricing":{"prompt":"0.000003","completion":"0.000015","request":"0","image":"0.0048","web_search":"0","internal_reasoning":"0","input_cache_read":"0.0000003","input_cache_write":"0.00000375"},"created":1740422110,"top_provider":{"context_length":200000,"max_completion_tokens":128000,"is_moderated":true}},{"id":"perplexity/r1-1776","name":"Perplexity: R1 1776","pricing":{"prompt":"0.000002","completion":"0.000008","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1740004929,"top_provider":{"context_length":128000,"max_completion_tokens":null,"is_moderated":false}},{"id":"mistralai/mistral-saba","name":"Mistral: Saba","pricing":{"prompt":"0.0000002","completion":"0.0000006","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1739803239,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"cognitivecomputations/dolphin3.0-r1-mistral-24b:free","name":"Dolphin3.0 R1 Mistral 24B (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1739462498,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"cognitivecomputations/dolphin3.0-mistral-24b:free","name":"Dolphin3.0 Mistral 24B (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1739462019,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"meta-llama/llama-guard-3-8b","name":"Llama Guard 3 8B","pricing":{"prompt":"0.00000002","completion":"0.00000006","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1739401318,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"openai/o3-mini-high","name":"OpenAI: o3 Mini High","pricing":{"prompt":"0.0000011","completion":"0.0000044","request":"0","image":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.00000055"},"created":1739372611,"top_provider":{"context_length":200000,"max_completion_tokens":100000,"is_moderated":true}},{"id":"deepseek/deepseek-r1-distill-llama-8b","name":"DeepSeek: R1 Distill Llama 8B","pricing":{"prompt":"0.00000004","completion":"0.00000004","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1738937718,"top_provider":{"context_length":32000,"max_completion_tokens":32000,"is_moderated":false}},{"id":"google/gemini-2.0-flash-001","name":"Google: Gemini 2.0 Flash","pricing":{"prompt":"0.0000001","completion":"0.0000004","request":"0","image":"0.0000258","web_search":"0","internal_reasoning":"0","input_cache_read":"0.000000025","input_cache_write":"0.0000001833"},"created":1738769413,"top_provider":{"context_length":1048576,"max_completion_tokens":8192,"is_moderated":false}},{"id":"qwen/qwen-vl-plus","name":"Qwen: Qwen VL Plus","pricing":{"prompt":"0.00000021","completion":"0.00000063","request":"0","image":"0.0002688","web_search":"0","internal_reasoning":"0"},"created":1738731255,"top_provider":{"context_length":7500,"max_completion_tokens":1500,"is_moderated":false}},{"id":"aion-labs/aion-1.0","name":"AionLabs: Aion-1.0","pricing":{"prompt":"0.000004","completion":"0.000008","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1738697557,"top_provider":{"context_length":131072,"max_completion_tokens":32768,"is_moderated":false}},{"id":"aion-labs/aion-1.0-mini","name":"AionLabs: Aion-1.0-Mini","pricing":{"prompt":"0.0000007","completion":"0.0000014","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1738697107,"top_provider":{"context_length":131072,"max_completion_tokens":32768,"is_moderated":false}},{"id":"aion-labs/aion-rp-llama-3.1-8b","name":"AionLabs: Aion-RP 1.0 (8B)","pricing":{"prompt":"0.0000002","completion":"0.0000002","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1738696718,"top_provider":{"context_length":32768,"max_completion_tokens":32768,"is_moderated":false}},{"id":"qwen/qwen-vl-max","name":"Qwen: Qwen VL Max","pricing":{"prompt":"0.0000008","completion":"0.0000032","request":"0","image":"0.001024","web_search":"0","internal_reasoning":"0"},"created":1738434304,"top_provider":{"context_length":7500,"max_completion_tokens":1500,"is_moderated":false}},{"id":"qwen/qwen-turbo","name":"Qwen: Qwen-Turbo","pricing":{"prompt":"0.00000005","completion":"0.0000002","request":"0","image":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.00000002"},"created":1738410974,"top_provider":{"context_length":1000000,"max_completion_tokens":8192,"is_moderated":false}},{"id":"qwen/qwen2.5-vl-72b-instruct:free","name":"Qwen: Qwen2.5 VL 72B Instruct (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1738410311,"top_provider":{"context_length":131072,"max_completion_tokens":2048,"is_moderated":false}},{"id":"qwen/qwen2.5-vl-72b-instruct","name":"Qwen: Qwen2.5 VL 72B Instruct","pricing":{"prompt":"0.00000025","completion":"0.00000075","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1738410311,"top_provider":{"context_length":32000,"max_completion_tokens":null,"is_moderated":false}},{"id":"qwen/qwen-plus","name":"Qwen: Qwen-Plus","pricing":{"prompt":"0.0000004","completion":"0.0000012","request":"0","image":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.00000016"},"created":1738409840,"top_provider":{"context_length":131072,"max_completion_tokens":8192,"is_moderated":false}},{"id":"qwen/qwen-max","name":"Qwen: Qwen-Max ","pricing":{"prompt":"0.0000016","completion":"0.0000064","request":"0","image":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.00000064"},"created":1738402289,"top_provider":{"context_length":32768,"max_completion_tokens":8192,"is_moderated":false}},{"id":"openai/o3-mini","name":"OpenAI: o3 Mini","pricing":{"prompt":"0.0000011","completion":"0.0000044","request":"0","image":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.00000055"},"created":1738351721,"top_provider":{"context_length":200000,"max_completion_tokens":100000,"is_moderated":true}},{"id":"deepseek/deepseek-r1-distill-qwen-1.5b","name":"DeepSeek: R1 Distill Qwen 1.5B","pricing":{"prompt":"0.00000018","completion":"0.00000018","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1738328067,"top_provider":{"context_length":131072,"max_completion_tokens":32768,"is_moderated":false}},{"id":"mistralai/mistral-small-24b-instruct-2501:free","name":"Mistral: Mistral Small 3 (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1738255409,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"mistralai/mistral-small-24b-instruct-2501","name":"Mistral: Mistral Small 3","pricing":{"prompt":"0.00000005","completion":"0.00000009","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1738255409,"top_provider":{"context_length":32768,"max_completion_tokens":32768,"is_moderated":false}},{"id":"deepseek/deepseek-r1-distill-qwen-32b","name":"DeepSeek: R1 Distill Qwen 32B","pricing":{"prompt":"0.000000075","completion":"0.00000015","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1738194830,"top_provider":{"context_length":131072,"max_completion_tokens":16384,"is_moderated":false}},{"id":"deepseek/deepseek-r1-distill-qwen-14b:free","name":"DeepSeek: R1 Distill Qwen 14B (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1738193940,"top_provider":{"context_length":64000,"max_completion_tokens":null,"is_moderated":false}},{"id":"deepseek/deepseek-r1-distill-qwen-14b","name":"DeepSeek: R1 Distill Qwen 14B","pricing":{"prompt":"0.00000015","completion":"0.00000015","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1738193940,"top_provider":{"context_length":64000,"max_completion_tokens":32000,"is_moderated":false}},{"id":"perplexity/sonar-reasoning","name":"Perplexity: Sonar Reasoning","pricing":{"prompt":"0.000001","completion":"0.000005","request":"0.005","image":"0","web_search":"0","internal_reasoning":"0"},"created":1738131107,"top_provider":{"context_length":127000,"max_completion_tokens":null,"is_moderated":false}},{"id":"perplexity/sonar","name":"Perplexity: Sonar","pricing":{"prompt":"0.000001","completion":"0.000001","request":"0.005","image":"0","web_search":"0","internal_reasoning":"0"},"created":1738013808,"top_provider":{"context_length":127072,"max_completion_tokens":null,"is_moderated":false}},{"id":"liquid/lfm-7b","name":"Liquid: LFM 7B","pricing":{"prompt":"0.00000001","completion":"0.00000001","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1737806883,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"liquid/lfm-3b","name":"Liquid: LFM 3B","pricing":{"prompt":"0.00000002","completion":"0.00000002","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1737806501,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"deepseek/deepseek-r1-distill-llama-70b:free","name":"DeepSeek: R1 Distill Llama 70B (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1737663169,"top_provider":{"context_length":8192,"max_completion_tokens":4096,"is_moderated":false}},{"id":"deepseek/deepseek-r1-distill-llama-70b","name":"DeepSeek: R1 Distill Llama 70B","pricing":{"prompt":"0.0000001","completion":"0.0000004","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1737663169,"top_provider":{"context_length":131072,"max_completion_tokens":16384,"is_moderated":false}},{"id":"deepseek/deepseek-r1:free","name":"DeepSeek: R1 (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1737381095,"top_provider":{"context_length":163840,"max_completion_tokens":null,"is_moderated":false}},{"id":"deepseek/deepseek-r1","name":"DeepSeek: R1","pricing":{"prompt":"0.00000045","completion":"0.00000215","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1737381095,"top_provider":{"context_length":128000,"max_completion_tokens":32768,"is_moderated":false}},{"id":"minimax/minimax-01","name":"MiniMax: MiniMax-01","pricing":{"prompt":"0.0000002","completion":"0.0000011","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1736915462,"top_provider":{"context_length":1000192,"max_completion_tokens":1000192,"is_moderated":false}},{"id":"mistralai/codestral-2501","name":"Mistral: Codestral 2501","pricing":{"prompt":"0.0000003","completion":"0.0000009","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1736895522,"top_provider":{"context_length":262144,"max_completion_tokens":null,"is_moderated":false}},{"id":"microsoft/phi-4","name":"Microsoft: Phi 4","pricing":{"prompt":"0.00000007","completion":"0.00000014","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1736489872,"top_provider":{"context_length":16384,"max_completion_tokens":16384,"is_moderated":false}},{"id":"deepseek/deepseek-chat:free","name":"DeepSeek: DeepSeek V3 (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1735241320,"top_provider":{"context_length":163840,"max_completion_tokens":null,"is_moderated":false}},{"id":"deepseek/deepseek-chat","name":"DeepSeek: DeepSeek V3","pricing":{"prompt":"0.00000038","completion":"0.00000089","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1735241320,"top_provider":{"context_length":163840,"max_completion_tokens":163840,"is_moderated":false}},{"id":"sao10k/l3.3-euryale-70b","name":"Sao10K: Llama 3.3 Euryale 70B","pricing":{"prompt":"0.00000065","completion":"0.00000075","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1734535928,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"openai/o1","name":"OpenAI: o1","pricing":{"prompt":"0.000015","completion":"0.00006","request":"0","image":"0.021675","web_search":"0","internal_reasoning":"0","input_cache_read":"0.0000075"},"created":1734459999,"top_provider":{"context_length":200000,"max_completion_tokens":100000,"is_moderated":true}},{"id":"eva-unit-01/eva-llama-3.33-70b","name":"EVA Llama 3.33 70B","pricing":{"prompt":"0.000004","completion":"0.000006","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1734377303,"top_provider":{"context_length":16384,"max_completion_tokens":4096,"is_moderated":false}},{"id":"x-ai/grok-2-vision-1212","name":"xAI: Grok 2 Vision 1212","pricing":{"prompt":"0.000002","completion":"0.00001","request":"0","image":"0.0036","web_search":"0","internal_reasoning":"0"},"created":1734237338,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"x-ai/grok-2-1212","name":"xAI: Grok 2 1212","pricing":{"prompt":"0.000002","completion":"0.00001","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1734232814,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"cohere/command-r7b-12-2024","name":"Cohere: Command R7B (12-2024)","pricing":{"prompt":"0.0000000375","completion":"0.00000015","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1734158152,"top_provider":{"context_length":128000,"max_completion_tokens":4000,"is_moderated":true}},{"id":"google/gemini-2.0-flash-exp:free","name":"Google: Gemini 2.0 Flash Experimental (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1733937523,"top_provider":{"context_length":1048576,"max_completion_tokens":8192,"is_moderated":false}},{"id":"meta-llama/llama-3.3-70b-instruct:free","name":"Meta: Llama 3.3 70B Instruct (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1733506137,"top_provider":{"context_length":131072,"max_completion_tokens":2048,"is_moderated":false}},{"id":"meta-llama/llama-3.3-70b-instruct","name":"Meta: Llama 3.3 70B Instruct","pricing":{"prompt":"0.000000039","completion":"0.00000012","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1733506137,"top_provider":{"context_length":131072,"max_completion_tokens":8192,"is_moderated":false}},{"id":"amazon/nova-lite-v1","name":"Amazon: Nova Lite 1.0","pricing":{"prompt":"0.00000006","completion":"0.00000024","request":"0","image":"0.00009","web_search":"0","internal_reasoning":"0"},"created":1733437363,"top_provider":{"context_length":300000,"max_completion_tokens":5120,"is_moderated":true}},{"id":"amazon/nova-micro-v1","name":"Amazon: Nova Micro 1.0","pricing":{"prompt":"0.000000035","completion":"0.00000014","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1733437237,"top_provider":{"context_length":128000,"max_completion_tokens":5120,"is_moderated":true}},{"id":"amazon/nova-pro-v1","name":"Amazon: Nova Pro 1.0","pricing":{"prompt":"0.0000008","completion":"0.0000032","request":"0","image":"0.0012","web_search":"0","internal_reasoning":"0"},"created":1733436303,"top_provider":{"context_length":300000,"max_completion_tokens":5120,"is_moderated":true}},{"id":"qwen/qwq-32b-preview","name":"Qwen: QwQ 32B Preview","pricing":{"prompt":"0.0000002","completion":"0.0000002","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1732754541,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"eva-unit-01/eva-qwen-2.5-72b","name":"EVA Qwen2.5 72B","pricing":{"prompt":"0.000004","completion":"0.000006","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1732210606,"top_provider":{"context_length":16384,"max_completion_tokens":4096,"is_moderated":false}},{"id":"openai/gpt-4o-2024-11-20","name":"OpenAI: GPT-4o (2024-11-20)","pricing":{"prompt":"0.0000025","completion":"0.00001","request":"0","image":"0.003613","web_search":"0","internal_reasoning":"0","input_cache_read":"0.00000125"},"created":1732127594,"top_provider":{"context_length":128000,"max_completion_tokens":16384,"is_moderated":true}},{"id":"mistralai/mistral-large-2411","name":"Mistral Large 2411","pricing":{"prompt":"0.000002","completion":"0.000006","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1731978685,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"mistralai/mistral-large-2407","name":"Mistral Large 2407","pricing":{"prompt":"0.000002","completion":"0.000006","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1731978415,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"mistralai/pixtral-large-2411","name":"Mistral: Pixtral Large 2411","pricing":{"prompt":"0.000002","completion":"0.000006","request":"0","image":"0.002888","web_search":"0","internal_reasoning":"0"},"created":1731977388,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"x-ai/grok-vision-beta","name":"xAI: Grok Vision Beta","pricing":{"prompt":"0.000005","completion":"0.000015","request":"0","image":"0.009","web_search":"0","internal_reasoning":"0"},"created":1731976624,"top_provider":{"context_length":8192,"max_completion_tokens":null,"is_moderated":false}},{"id":"infermatic/mn-inferor-12b","name":"Infermatic: Mistral Nemo Inferor 12B","pricing":{"prompt":"0.0000008","completion":"0.0000012","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1731464428,"top_provider":{"context_length":16384,"max_completion_tokens":4096,"is_moderated":false}},{"id":"qwen/qwen-2.5-coder-32b-instruct:free","name":"Qwen2.5 Coder 32B Instruct (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1731368400,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"qwen/qwen-2.5-coder-32b-instruct","name":"Qwen2.5 Coder 32B Instruct","pricing":{"prompt":"0.00000006","completion":"0.00000015","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1731368400,"top_provider":{"context_length":32768,"max_completion_tokens":16384,"is_moderated":false}},{"id":"raifle/sorcererlm-8x22b","name":"SorcererLM 8x22B","pricing":{"prompt":"0.0000045","completion":"0.0000045","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1731105083,"top_provider":{"context_length":16000,"max_completion_tokens":null,"is_moderated":false}},{"id":"eva-unit-01/eva-qwen-2.5-32b","name":"EVA Qwen2.5 32B","pricing":{"prompt":"0.0000026","completion":"0.0000034","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1731104847,"top_provider":{"context_length":16384,"max_completion_tokens":4096,"is_moderated":false}},{"id":"thedrummer/unslopnemo-12b","name":"TheDrummer: UnslopNemo 12B","pricing":{"prompt":"0.0000004","completion":"0.0000004","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1731103448,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"anthropic/claude-3.5-haiku:beta","name":"Anthropic: Claude 3.5 Haiku (self-moderated)","pricing":{"prompt":"0.0000008","completion":"0.000004","request":"0","image":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.00000008","input_cache_write":"0.000001"},"created":1730678400,"top_provider":{"context_length":200000,"max_completion_tokens":8192,"is_moderated":false}},{"id":"anthropic/claude-3.5-haiku","name":"Anthropic: Claude 3.5 Haiku","pricing":{"prompt":"0.0000008","completion":"0.000004","request":"0","image":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.00000008","input_cache_write":"0.000001"},"created":1730678400,"top_provider":{"context_length":200000,"max_completion_tokens":8192,"is_moderated":true}},{"id":"anthropic/claude-3.5-haiku-20241022:beta","name":"Anthropic: Claude 3.5 Haiku (2024-10-22) (self-moderated)","pricing":{"prompt":"0.0000008","completion":"0.000004","request":"0","image":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.00000008","input_cache_write":"0.000001"},"created":1730678400,"top_provider":{"context_length":200000,"max_completion_tokens":8192,"is_moderated":false}},{"id":"anthropic/claude-3.5-haiku-20241022","name":"Anthropic: Claude 3.5 Haiku (2024-10-22)","pricing":{"prompt":"0.0000008","completion":"0.000004","request":"0","image":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.00000008","input_cache_write":"0.000001"},"created":1730678400,"top_provider":{"context_length":200000,"max_completion_tokens":8192,"is_moderated":true}},{"id":"anthropic/claude-3.5-sonnet:beta","name":"Anthropic: Claude 3.5 Sonnet (self-moderated)","pricing":{"prompt":"0.000003","completion":"0.000015","request":"0","image":"0.0048","web_search":"0","internal_reasoning":"0","input_cache_read":"0.0000003","input_cache_write":"0.00000375"},"created":1729555200,"top_provider":{"context_length":200000,"max_completion_tokens":8192,"is_moderated":false}},{"id":"anthropic/claude-3.5-sonnet","name":"Anthropic: Claude 3.5 Sonnet","pricing":{"prompt":"0.000003","completion":"0.000015","request":"0","image":"0.0048","web_search":"0","internal_reasoning":"0","input_cache_read":"0.0000003","input_cache_write":"0.00000375"},"created":1729555200,"top_provider":{"context_length":200000,"max_completion_tokens":8192,"is_moderated":true}},{"id":"anthracite-org/magnum-v4-72b","name":"Magnum v4 72B","pricing":{"prompt":"0.0000025","completion":"0.000003","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1729555200,"top_provider":{"context_length":16384,"max_completion_tokens":1024,"is_moderated":false}},{"id":"neversleep/llama-3.1-lumimaid-70b","name":"NeverSleep: Lumimaid v0.2 70B","pricing":{"prompt":"0.0000025","completion":"0.000003","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1729555200,"top_provider":{"context_length":16384,"max_completion_tokens":2048,"is_moderated":false}},{"id":"x-ai/grok-beta","name":"xAI: Grok Beta","pricing":{"prompt":"0.000005","completion":"0.000015","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1729382400,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"mistralai/ministral-8b","name":"Mistral: Ministral 8B","pricing":{"prompt":"0.0000001","completion":"0.0000001","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1729123200,"top_provider":{"context_length":128000,"max_completion_tokens":null,"is_moderated":false}},{"id":"mistralai/ministral-3b","name":"Mistral: Ministral 3B","pricing":{"prompt":"0.00000004","completion":"0.00000004","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1729123200,"top_provider":{"context_length":131072,"max_completion_tokens":null,"is_moderated":false}},{"id":"qwen/qwen-2.5-7b-instruct","name":"Qwen2.5 7B Instruct","pricing":{"prompt":"0.00000004","completion":"0.0000001","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1729036800,"top_provider":{"context_length":32768,"max_completion_tokens":16384,"is_moderated":false}},{"id":"nvidia/llama-3.1-nemotron-70b-instruct","name":"NVIDIA: Llama 3.1 Nemotron 70B Instruct","pricing":{"prompt":"0.00000012","completion":"0.0000003","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1728950400,"top_provider":{"context_length":131072,"max_completion_tokens":131072,"is_moderated":false}},{"id":"inflection/inflection-3-pi","name":"Inflection: Inflection 3 Pi","pricing":{"prompt":"0.0000025","completion":"0.00001","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1728604800,"top_provider":{"context_length":8000,"max_completion_tokens":1024,"is_moderated":false}},{"id":"inflection/inflection-3-productivity","name":"Inflection: Inflection 3 Productivity","pricing":{"prompt":"0.0000025","completion":"0.00001","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1728604800,"top_provider":{"context_length":8000,"max_completion_tokens":1024,"is_moderated":false}},{"id":"google/gemini-flash-1.5-8b","name":"Google: Gemini 1.5 Flash 8B","pricing":{"prompt":"0.0000000375","completion":"0.00000015","request":"0","image":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.00000001","input_cache_write":"0.0000000583"},"created":1727913600,"top_provider":{"context_length":1000000,"max_completion_tokens":8192,"is_moderated":false}},{"id":"liquid/lfm-40b","name":"Liquid: LFM 40B MoE","pricing":{"prompt":"0.00000015","completion":"0.00000015","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1727654400,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"anthracite-org/magnum-v2-72b","name":"Magnum v2 72B","pricing":{"prompt":"0.000003","completion":"0.000003","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1727654400,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"thedrummer/rocinante-12b","name":"TheDrummer: Rocinante 12B","pricing":{"prompt":"0.0000002","completion":"0.0000005","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1727654400,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"meta-llama/llama-3.2-3b-instruct","name":"Meta: Llama 3.2 3B Instruct","pricing":{"prompt":"0.000000003","completion":"0.000000006","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1727222400,"top_provider":{"context_length":20000,"max_completion_tokens":20000,"is_moderated":false}},{"id":"meta-llama/llama-3.2-90b-vision-instruct","name":"Meta: Llama 3.2 90B Vision Instruct","pricing":{"prompt":"0.0000012","completion":"0.0000012","request":"0","image":"0.001734","web_search":"0","internal_reasoning":"0"},"created":1727222400,"top_provider":{"context_length":131072,"max_completion_tokens":2048,"is_moderated":false}},{"id":"meta-llama/llama-3.2-11b-vision-instruct:free","name":"Meta: Llama 3.2 11B Vision Instruct (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1727222400,"top_provider":{"context_length":131072,"max_completion_tokens":2048,"is_moderated":false}},{"id":"meta-llama/llama-3.2-11b-vision-instruct","name":"Meta: Llama 3.2 11B Vision Instruct","pricing":{"prompt":"0.000000049","completion":"0.000000049","request":"0","image":"0.00007948","web_search":"0","internal_reasoning":"0"},"created":1727222400,"top_provider":{"context_length":131072,"max_completion_tokens":16384,"is_moderated":false}},{"id":"meta-llama/llama-3.2-1b-instruct:free","name":"Meta: Llama 3.2 1B Instruct (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1727222400,"top_provider":{"context_length":131072,"max_completion_tokens":8192,"is_moderated":false}},{"id":"meta-llama/llama-3.2-1b-instruct","name":"Meta: Llama 3.2 1B Instruct","pricing":{"prompt":"0.000000005","completion":"0.00000001","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1727222400,"top_provider":{"context_length":131072,"max_completion_tokens":16384,"is_moderated":false}},{"id":"qwen/qwen-2.5-72b-instruct:free","name":"Qwen2.5 72B Instruct (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1726704000,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"qwen/qwen-2.5-72b-instruct","name":"Qwen2.5 72B Instruct","pricing":{"prompt":"0.00000012","completion":"0.00000039","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1726704000,"top_provider":{"context_length":32768,"max_completion_tokens":16384,"is_moderated":false}},{"id":"neversleep/llama-3.1-lumimaid-8b","name":"NeverSleep: Lumimaid v0.2 8B","pricing":{"prompt":"0.0000002","completion":"0.00000125","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1726358400,"top_provider":{"context_length":32768,"max_completion_tokens":2048,"is_moderated":false}},{"id":"openai/o1-preview","name":"OpenAI: o1-preview","pricing":{"prompt":"0.000015","completion":"0.00006","request":"0","image":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.0000075"},"created":1726099200,"top_provider":{"context_length":128000,"max_completion_tokens":32768,"is_moderated":true}},{"id":"openai/o1-mini-2024-09-12","name":"OpenAI: o1-mini (2024-09-12)","pricing":{"prompt":"0.0000011","completion":"0.0000044","request":"0","image":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.00000055"},"created":1726099200,"top_provider":{"context_length":128000,"max_completion_tokens":65536,"is_moderated":true}},{"id":"openai/o1-preview-2024-09-12","name":"OpenAI: o1-preview (2024-09-12)","pricing":{"prompt":"0.000015","completion":"0.00006","request":"0","image":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.0000075"},"created":1726099200,"top_provider":{"context_length":128000,"max_completion_tokens":32768,"is_moderated":true}},{"id":"openai/o1-mini","name":"OpenAI: o1-mini","pricing":{"prompt":"0.0000011","completion":"0.0000044","request":"0","image":"0","web_search":"0","internal_reasoning":"0","input_cache_read":"0.00000055"},"created":1726099200,"top_provider":{"context_length":128000,"max_completion_tokens":65536,"is_moderated":true}},{"id":"mistralai/pixtral-12b","name":"Mistral: Pixtral 12B","pricing":{"prompt":"0.0000001","completion":"0.0000001","request":"0","image":"0.0001445","web_search":"0","internal_reasoning":"0"},"created":1725926400,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"cohere/command-r-plus-08-2024","name":"Cohere: Command R+ (08-2024)","pricing":{"prompt":"0.0000025","completion":"0.00001","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1724976000,"top_provider":{"context_length":128000,"max_completion_tokens":4000,"is_moderated":true}},{"id":"cohere/command-r-08-2024","name":"Cohere: Command R (08-2024)","pricing":{"prompt":"0.00000015","completion":"0.0000006","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1724976000,"top_provider":{"context_length":128000,"max_completion_tokens":4000,"is_moderated":true}},{"id":"qwen/qwen-2.5-vl-7b-instruct","name":"Qwen: Qwen2.5-VL 7B Instruct","pricing":{"prompt":"0.0000002","completion":"0.0000002","request":"0","image":"0.0001445","web_search":"0","internal_reasoning":"0"},"created":1724803200,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"sao10k/l3.1-euryale-70b","name":"Sao10K: Llama 3.1 Euryale 70B v2.2","pricing":{"prompt":"0.00000065","completion":"0.00000075","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1724803200,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"microsoft/phi-3.5-mini-128k-instruct","name":"Microsoft: Phi-3.5 Mini 128K Instruct","pricing":{"prompt":"0.0000001","completion":"0.0000001","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1724198400,"top_provider":{"context_length":128000,"max_completion_tokens":null,"is_moderated":false}},{"id":"nousresearch/hermes-3-llama-3.1-70b","name":"Nous: Hermes 3 70B Instruct","pricing":{"prompt":"0.00000012","completion":"0.0000003","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1723939200,"top_provider":{"context_length":131072,"max_completion_tokens":131072,"is_moderated":false}},{"id":"nousresearch/hermes-3-llama-3.1-405b","name":"Nous: Hermes 3 405B Instruct","pricing":{"prompt":"0.0000007","completion":"0.0000008","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1723766400,"top_provider":{"context_length":131072,"max_completion_tokens":16384,"is_moderated":false}},{"id":"openai/chatgpt-4o-latest","name":"OpenAI: ChatGPT-4o","pricing":{"prompt":"0.000005","completion":"0.000015","request":"0","image":"0.007225","web_search":"0","internal_reasoning":"0"},"created":1723593600,"top_provider":{"context_length":128000,"max_completion_tokens":16384,"is_moderated":true}},{"id":"aetherwiing/mn-starcannon-12b","name":"Aetherwiing: Starcannon 12B","pricing":{"prompt":"0.0000008","completion":"0.0000012","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1723507200,"top_provider":{"context_length":16384,"max_completion_tokens":4096,"is_moderated":false}},{"id":"sao10k/l3-lunaris-8b","name":"Sao10K: Llama 3 8B Lunaris","pricing":{"prompt":"0.00000002","completion":"0.00000005","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1723507200,"top_provider":{"context_length":8192,"max_completion_tokens":null,"is_moderated":false}},{"id":"openai/gpt-4o-2024-08-06","name":"OpenAI: GPT-4o (2024-08-06)","pricing":{"prompt":"0.0000025","completion":"0.00001","request":"0","image":"0.003613","web_search":"0","internal_reasoning":"0","input_cache_read":"0.00000125"},"created":1722902400,"top_provider":{"context_length":128000,"max_completion_tokens":16384,"is_moderated":false}},{"id":"nothingiisreal/mn-celeste-12b","name":"Mistral Nemo 12B Celeste","pricing":{"prompt":"0.0000008","completion":"0.0000012","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1722556800,"top_provider":{"context_length":16384,"max_completion_tokens":4096,"is_moderated":false}},{"id":"meta-llama/llama-3.1-405b","name":"Meta: Llama 3.1 405B (base)","pricing":{"prompt":"0.000002","completion":"0.000002","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1722556800,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"perplexity/llama-3.1-sonar-small-128k-online","name":"Perplexity: Llama 3.1 Sonar 8B Online","pricing":{"prompt":"0.0000002","completion":"0.0000002","request":"0.005","image":"0","web_search":"0","internal_reasoning":"0"},"created":1722470400,"top_provider":{"context_length":127072,"max_completion_tokens":null,"is_moderated":false}},{"id":"perplexity/llama-3.1-sonar-large-128k-online","name":"Perplexity: Llama 3.1 Sonar 70B Online","pricing":{"prompt":"0.000001","completion":"0.000001","request":"0.005","image":"0","web_search":"0","internal_reasoning":"0"},"created":1722470400,"top_provider":{"context_length":127072,"max_completion_tokens":null,"is_moderated":false}},{"id":"meta-llama/llama-3.1-405b-instruct","name":"Meta: Llama 3.1 405B Instruct","pricing":{"prompt":"0.0000008","completion":"0.0000008","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1721692800,"top_provider":{"context_length":32768,"max_completion_tokens":16384,"is_moderated":false}},{"id":"meta-llama/llama-3.1-70b-instruct","name":"Meta: Llama 3.1 70B Instruct","pricing":{"prompt":"0.0000001","completion":"0.00000028","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1721692800,"top_provider":{"context_length":131072,"max_completion_tokens":16384,"is_moderated":false}},{"id":"meta-llama/llama-3.1-8b-instruct:free","name":"Meta: Llama 3.1 8B Instruct (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1721692800,"top_provider":{"context_length":131072,"max_completion_tokens":4096,"is_moderated":false}},{"id":"meta-llama/llama-3.1-8b-instruct","name":"Meta: Llama 3.1 8B Instruct","pricing":{"prompt":"0.000000016","completion":"0.000000021","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1721692800,"top_provider":{"context_length":131000,"max_completion_tokens":131000,"is_moderated":false}},{"id":"mistralai/mistral-nemo:free","name":"Mistral: Mistral Nemo (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1721347200,"top_provider":{"context_length":131072,"max_completion_tokens":128000,"is_moderated":false}},{"id":"mistralai/mistral-nemo","name":"Mistral: Mistral Nemo","pricing":{"prompt":"0.00000001","completion":"0.000000011","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1721347200,"top_provider":{"context_length":131072,"max_completion_tokens":131072,"is_moderated":false}},{"id":"openai/gpt-4o-mini","name":"OpenAI: GPT-4o-mini","pricing":{"prompt":"0.00000015","completion":"0.0000006","request":"0","image":"0.000217","web_search":"0","internal_reasoning":"0","input_cache_read":"0.000000075"},"created":1721260800,"top_provider":{"context_length":128000,"max_completion_tokens":16384,"is_moderated":true}},{"id":"openai/gpt-4o-mini-2024-07-18","name":"OpenAI: GPT-4o-mini (2024-07-18)","pricing":{"prompt":"0.00000015","completion":"0.0000006","request":"0","image":"0.007225","web_search":"0","internal_reasoning":"0","input_cache_read":"0.000000075"},"created":1721260800,"top_provider":{"context_length":128000,"max_completion_tokens":16384,"is_moderated":true}},{"id":"google/gemma-2-27b-it","name":"Google: Gemma 2 27B","pricing":{"prompt":"0.0000008","completion":"0.0000008","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1720828800,"top_provider":{"context_length":8192,"max_completion_tokens":2048,"is_moderated":false}},{"id":"alpindale/magnum-72b","name":"Magnum 72B","pricing":{"prompt":"0.000004","completion":"0.000006","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1720656000,"top_provider":{"context_length":16384,"max_completion_tokens":4096,"is_moderated":false}},{"id":"google/gemma-2-9b-it:free","name":"Google: Gemma 2 9B (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1719532800,"top_provider":{"context_length":8192,"max_completion_tokens":8192,"is_moderated":false}},{"id":"google/gemma-2-9b-it","name":"Google: Gemma 2 9B","pricing":{"prompt":"0.0000002","completion":"0.0000002","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1719532800,"top_provider":{"context_length":8192,"max_completion_tokens":8192,"is_moderated":false}},{"id":"01-ai/yi-large","name":"01.AI: Yi Large","pricing":{"prompt":"0.000003","completion":"0.000003","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1719273600,"top_provider":{"context_length":32768,"max_completion_tokens":4096,"is_moderated":false}},{"id":"anthropic/claude-3.5-sonnet-20240620:beta","name":"Anthropic: Claude 3.5 Sonnet (2024-06-20) (self-moderated)","pricing":{"prompt":"0.000003","completion":"0.000015","request":"0","image":"0.0048","web_search":"0","internal_reasoning":"0","input_cache_read":"0.0000003","input_cache_write":"0.00000375"},"created":1718841600,"top_provider":{"context_length":200000,"max_completion_tokens":8192,"is_moderated":false}},{"id":"anthropic/claude-3.5-sonnet-20240620","name":"Anthropic: Claude 3.5 Sonnet (2024-06-20)","pricing":{"prompt":"0.000003","completion":"0.000015","request":"0","image":"0.0048","web_search":"0","internal_reasoning":"0","input_cache_read":"0.0000003","input_cache_write":"0.00000375"},"created":1718841600,"top_provider":{"context_length":200000,"max_completion_tokens":8192,"is_moderated":true}},{"id":"sao10k/l3-euryale-70b","name":"Sao10k: Llama 3 Euryale 70B v2.1","pricing":{"prompt":"0.00000148","completion":"0.00000148","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1718668800,"top_provider":{"context_length":8192,"max_completion_tokens":8192,"is_moderated":false}},{"id":"cognitivecomputations/dolphin-mixtral-8x22b","name":"Dolphin 2.9.2 Mixtral 8x22B 🐬","pricing":{"prompt":"0.0000009","completion":"0.0000009","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1717804800,"top_provider":{"context_length":16000,"max_completion_tokens":8192,"is_moderated":false}},{"id":"qwen/qwen-2-72b-instruct","name":"Qwen 2 72B Instruct","pricing":{"prompt":"0.0000009","completion":"0.0000009","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1717718400,"top_provider":{"context_length":32768,"max_completion_tokens":4096,"is_moderated":false}},{"id":"mistralai/mistral-7b-instruct-v0.3","name":"Mistral: Mistral 7B Instruct v0.3","pricing":{"prompt":"0.000000028","completion":"0.000000054","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1716768000,"top_provider":{"context_length":32768,"max_completion_tokens":16384,"is_moderated":false}},{"id":"nousresearch/hermes-2-pro-llama-3-8b","name":"NousResearch: Hermes 2 Pro - Llama-3 8B","pricing":{"prompt":"0.000000025","completion":"0.00000004","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1716768000,"top_provider":{"context_length":131072,"max_completion_tokens":131072,"is_moderated":false}},{"id":"mistralai/mistral-7b-instruct:free","name":"Mistral: Mistral 7B Instruct (free)","pricing":{"prompt":"0","completion":"0","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1716768000,"top_provider":{"context_length":32768,"max_completion_tokens":16384,"is_moderated":false}},{"id":"mistralai/mistral-7b-instruct","name":"Mistral: Mistral 7B Instruct","pricing":{"prompt":"0.000000028","completion":"0.000000054","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1716768000,"top_provider":{"context_length":32768,"max_completion_tokens":16384,"is_moderated":false}},{"id":"microsoft/phi-3-mini-128k-instruct","name":"Microsoft: Phi-3 Mini 128K Instruct","pricing":{"prompt":"0.0000001","completion":"0.0000001","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1716681600,"top_provider":{"context_length":128000,"max_completion_tokens":null,"is_moderated":false}},{"id":"microsoft/phi-3-medium-128k-instruct","name":"Microsoft: Phi-3 Medium 128K Instruct","pricing":{"prompt":"0.000001","completion":"0.000001","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1716508800,"top_provider":{"context_length":128000,"max_completion_tokens":null,"is_moderated":false}},{"id":"neversleep/llama-3-lumimaid-70b","name":"NeverSleep: Llama 3 Lumimaid 70B","pricing":{"prompt":"0.000004","completion":"0.000006","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1715817600,"top_provider":{"context_length":8192,"max_completion_tokens":4096,"is_moderated":false}},{"id":"google/gemini-flash-1.5","name":"Google: Gemini 1.5 Flash ","pricing":{"prompt":"0.000000075","completion":"0.0000003","request":"0","image":"0.00004","web_search":"0","internal_reasoning":"0","input_cache_read":"0.00000001875","input_cache_write":"0.0000001583"},"created":1715644800,"top_provider":{"context_length":1000000,"max_completion_tokens":8192,"is_moderated":false}},{"id":"openai/gpt-4o-2024-05-13","name":"OpenAI: GPT-4o (2024-05-13)","pricing":{"prompt":"0.000005","completion":"0.000015","request":"0","image":"0.007225","web_search":"0","internal_reasoning":"0"},"created":1715558400,"top_provider":{"context_length":128000,"max_completion_tokens":4096,"is_moderated":true}},{"id":"meta-llama/llama-guard-2-8b","name":"Meta: LlamaGuard 2 8B","pricing":{"prompt":"0.0000002","completion":"0.0000002","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1715558400,"top_provider":{"context_length":8192,"max_completion_tokens":null,"is_moderated":false}},{"id":"openai/gpt-4o","name":"OpenAI: GPT-4o","pricing":{"prompt":"0.0000025","completion":"0.00001","request":"0","image":"0.003613","web_search":"0","internal_reasoning":"0","input_cache_read":"0.00000125"},"created":1715558400,"top_provider":{"context_length":128000,"max_completion_tokens":16384,"is_moderated":true}},{"id":"openai/gpt-4o:extended","name":"OpenAI: GPT-4o (extended)","pricing":{"prompt":"0.000006","completion":"0.000018","request":"0","image":"0.007225","web_search":"0","internal_reasoning":"0"},"created":1715558400,"top_provider":{"context_length":128000,"max_completion_tokens":64000,"is_moderated":true}},{"id":"neversleep/llama-3-lumimaid-8b","name":"NeverSleep: Llama 3 Lumimaid 8B","pricing":{"prompt":"0.0000002","completion":"0.00000125","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1714780800,"top_provider":{"context_length":24576,"max_completion_tokens":2048,"is_moderated":false}},{"id":"sao10k/fimbulvetr-11b-v2","name":"Fimbulvetr 11B v2","pricing":{"prompt":"0.0000008","completion":"0.0000012","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1713657600,"top_provider":{"context_length":4096,"max_completion_tokens":4096,"is_moderated":false}},{"id":"meta-llama/llama-3-8b-instruct","name":"Meta: Llama 3 8B Instruct","pricing":{"prompt":"0.00000003","completion":"0.00000006","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1713398400,"top_provider":{"context_length":8192,"max_completion_tokens":16384,"is_moderated":false}},{"id":"meta-llama/llama-3-70b-instruct","name":"Meta: Llama 3 70B Instruct","pricing":{"prompt":"0.0000003","completion":"0.0000004","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1713398400,"top_provider":{"context_length":8192,"max_completion_tokens":16384,"is_moderated":false}},{"id":"mistralai/mixtral-8x22b-instruct","name":"Mistral: Mixtral 8x22B Instruct","pricing":{"prompt":"0.0000009","completion":"0.0000009","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1713312000,"top_provider":{"context_length":65536,"max_completion_tokens":null,"is_moderated":false}},{"id":"microsoft/wizardlm-2-8x22b","name":"WizardLM-2 8x22B","pricing":{"prompt":"0.00000048","completion":"0.00000048","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1713225600,"top_provider":{"context_length":65536,"max_completion_tokens":65536,"is_moderated":false}},{"id":"openai/gpt-4-turbo","name":"OpenAI: GPT-4 Turbo","pricing":{"prompt":"0.00001","completion":"0.00003","request":"0","image":"0.01445","web_search":"0","internal_reasoning":"0"},"created":1712620800,"top_provider":{"context_length":128000,"max_completion_tokens":4096,"is_moderated":true}},{"id":"google/gemini-pro-1.5","name":"Google: Gemini 1.5 Pro","pricing":{"prompt":"0.00000125","completion":"0.000005","request":"0","image":"0.0006575","web_search":"0","internal_reasoning":"0"},"created":1712620800,"top_provider":{"context_length":2000000,"max_completion_tokens":8192,"is_moderated":false}},{"id":"cohere/command-r-plus","name":"Cohere: Command R+","pricing":{"prompt":"0.000003","completion":"0.000015","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1712188800,"top_provider":{"context_length":128000,"max_completion_tokens":4000,"is_moderated":true}},{"id":"cohere/command-r-plus-04-2024","name":"Cohere: Command R+ (04-2024)","pricing":{"prompt":"0.000003","completion":"0.000015","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1712016000,"top_provider":{"context_length":128000,"max_completion_tokens":4000,"is_moderated":true}},{"id":"sophosympatheia/midnight-rose-70b","name":"Midnight Rose 70B","pricing":{"prompt":"0.0000008","completion":"0.0000008","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1711065600,"top_provider":{"context_length":4096,"max_completion_tokens":2048,"is_moderated":false}},{"id":"cohere/command-r","name":"Cohere: Command R","pricing":{"prompt":"0.0000005","completion":"0.0000015","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1710374400,"top_provider":{"context_length":128000,"max_completion_tokens":4000,"is_moderated":true}},{"id":"cohere/command","name":"Cohere: Command","pricing":{"prompt":"0.000001","completion":"0.000002","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1710374400,"top_provider":{"context_length":4096,"max_completion_tokens":4000,"is_moderated":true}},{"id":"anthropic/claude-3-haiku:beta","name":"Anthropic: Claude 3 Haiku (self-moderated)","pricing":{"prompt":"0.00000025","completion":"0.00000125","request":"0","image":"0.0004","web_search":"0","internal_reasoning":"0","input_cache_read":"0.00000003","input_cache_write":"0.0000003"},"created":1710288000,"top_provider":{"context_length":200000,"max_completion_tokens":4096,"is_moderated":false}},{"id":"anthropic/claude-3-haiku","name":"Anthropic: Claude 3 Haiku","pricing":{"prompt":"0.00000025","completion":"0.00000125","request":"0","image":"0.0004","web_search":"0","internal_reasoning":"0","input_cache_read":"0.00000003","input_cache_write":"0.0000003"},"created":1710288000,"top_provider":{"context_length":200000,"max_completion_tokens":4096,"is_moderated":true}},{"id":"anthropic/claude-3-opus:beta","name":"Anthropic: Claude 3 Opus (self-moderated)","pricing":{"prompt":"0.000015","completion":"0.000075","request":"0","image":"0.024","web_search":"0","internal_reasoning":"0","input_cache_read":"0.0000015","input_cache_write":"0.00001875"},"created":1709596800,"top_provider":{"context_length":200000,"max_completion_tokens":4096,"is_moderated":false}},{"id":"anthropic/claude-3-opus","name":"Anthropic: Claude 3 Opus","pricing":{"prompt":"0.000015","completion":"0.000075","request":"0","image":"0.024","web_search":"0","internal_reasoning":"0","input_cache_read":"0.0000015","input_cache_write":"0.00001875"},"created":1709596800,"top_provider":{"context_length":200000,"max_completion_tokens":4096,"is_moderated":true}},{"id":"anthropic/claude-3-sonnet:beta","name":"Anthropic: Claude 3 Sonnet (self-moderated)","pricing":{"prompt":"0.000003","completion":"0.000015","request":"0","image":"0.0048","web_search":"0","internal_reasoning":"0","input_cache_read":"0.0000003","input_cache_write":"0.00000375"},"created":1709596800,"top_provider":{"context_length":200000,"max_completion_tokens":4096,"is_moderated":false}},{"id":"anthropic/claude-3-sonnet","name":"Anthropic: Claude 3 Sonnet","pricing":{"prompt":"0.000003","completion":"0.000015","request":"0","image":"0.0048","web_search":"0","internal_reasoning":"0","input_cache_read":"0.0000003","input_cache_write":"0.00000375"},"created":1709596800,"top_provider":{"context_length":200000,"max_completion_tokens":4096,"is_moderated":true}},{"id":"cohere/command-r-03-2024","name":"Cohere: Command R (03-2024)","pricing":{"prompt":"0.0000005","completion":"0.0000015","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1709341200,"top_provider":{"context_length":128000,"max_completion_tokens":4000,"is_moderated":true}},{"id":"mistralai/mistral-large","name":"Mistral Large","pricing":{"prompt":"0.000002","completion":"0.000006","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1708905600,"top_provider":{"context_length":128000,"max_completion_tokens":null,"is_moderated":false}},{"id":"openai/gpt-3.5-turbo-0613","name":"OpenAI: GPT-3.5 Turbo (older v0613)","pricing":{"prompt":"0.000001","completion":"0.000002","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1706140800,"top_provider":{"context_length":4095,"max_completion_tokens":4096,"is_moderated":false}},{"id":"openai/gpt-4-turbo-preview","name":"OpenAI: GPT-4 Turbo Preview","pricing":{"prompt":"0.00001","completion":"0.00003","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1706140800,"top_provider":{"context_length":128000,"max_completion_tokens":4096,"is_moderated":true}},{"id":"nousresearch/nous-hermes-2-mixtral-8x7b-dpo","name":"Nous: Hermes 2 Mixtral 8x7B DPO","pricing":{"prompt":"0.0000006","completion":"0.0000006","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1705363200,"top_provider":{"context_length":32768,"max_completion_tokens":2048,"is_moderated":false}},{"id":"mistralai/mistral-small","name":"Mistral Small","pricing":{"prompt":"0.0000002","completion":"0.0000006","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1704844800,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"mistralai/mistral-tiny","name":"Mistral Tiny","pricing":{"prompt":"0.00000025","completion":"0.00000025","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1704844800,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"mistralai/mistral-7b-instruct-v0.2","name":"Mistral: Mistral 7B Instruct v0.2","pricing":{"prompt":"0.0000002","completion":"0.0000002","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1703721600,"top_provider":{"context_length":32768,"max_completion_tokens":null,"is_moderated":false}},{"id":"mistralai/mixtral-8x7b-instruct","name":"Mistral: Mixtral 8x7B Instruct","pricing":{"prompt":"0.00000008","completion":"0.00000024","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1702166400,"top_provider":{"context_length":32768,"max_completion_tokens":16384,"is_moderated":false}},{"id":"neversleep/noromaid-20b","name":"Noromaid 20B","pricing":{"prompt":"0.00000125","completion":"0.000002","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1700956800,"top_provider":{"context_length":8192,"max_completion_tokens":2048,"is_moderated":false}},{"id":"anthropic/claude-2:beta","name":"Anthropic: Claude v2 (self-moderated)","pricing":{"prompt":"0.000008","completion":"0.000024","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1700611200,"top_provider":{"context_length":200000,"max_completion_tokens":4096,"is_moderated":false}},{"id":"anthropic/claude-2","name":"Anthropic: Claude v2","pricing":{"prompt":"0.000008","completion":"0.000024","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1700611200,"top_provider":{"context_length":200000,"max_completion_tokens":4096,"is_moderated":true}},{"id":"anthropic/claude-2.1:beta","name":"Anthropic: Claude v2.1 (self-moderated)","pricing":{"prompt":"0.000008","completion":"0.000024","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1700611200,"top_provider":{"context_length":200000,"max_completion_tokens":4096,"is_moderated":false}},{"id":"anthropic/claude-2.1","name":"Anthropic: Claude v2.1","pricing":{"prompt":"0.000008","completion":"0.000024","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1700611200,"top_provider":{"context_length":200000,"max_completion_tokens":4096,"is_moderated":true}},{"id":"undi95/toppy-m-7b","name":"Toppy M 7B","pricing":{"prompt":"0.0000008","completion":"0.0000012","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1699574400,"top_provider":{"context_length":4096,"max_completion_tokens":4096,"is_moderated":false}},{"id":"alpindale/goliath-120b","name":"Goliath 120B","pricing":{"prompt":"0.000009","completion":"0.000011","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1699574400,"top_provider":{"context_length":6144,"max_completion_tokens":null,"is_moderated":false}},{"id":"openrouter/auto","name":"Auto Router","pricing":{"prompt":"-1","completion":"-1"},"created":1699401600,"top_provider":{"context_length":null,"max_completion_tokens":null,"is_moderated":false}},{"id":"openai/gpt-4-1106-preview","name":"OpenAI: GPT-4 Turbo (older v1106)","pricing":{"prompt":"0.00001","completion":"0.00003","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1699228800,"top_provider":{"context_length":128000,"max_completion_tokens":4096,"is_moderated":true}},{"id":"openai/gpt-3.5-turbo-instruct","name":"OpenAI: GPT-3.5 Turbo Instruct","pricing":{"prompt":"0.0000015","completion":"0.000002","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1695859200,"top_provider":{"context_length":4095,"max_completion_tokens":4096,"is_moderated":true}},{"id":"mistralai/mistral-7b-instruct-v0.1","name":"Mistral: Mistral 7B Instruct v0.1","pricing":{"prompt":"0.00000011","completion":"0.00000019","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1695859200,"top_provider":{"context_length":2824,"max_completion_tokens":null,"is_moderated":false}},{"id":"pygmalionai/mythalion-13b","name":"Pygmalion: Mythalion 13B","pricing":{"prompt":"0.0000008","completion":"0.0000012","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1693612800,"top_provider":{"context_length":4096,"max_completion_tokens":4096,"is_moderated":false}},{"id":"openai/gpt-3.5-turbo-16k","name":"OpenAI: GPT-3.5 Turbo 16k","pricing":{"prompt":"0.000003","completion":"0.000004","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1693180800,"top_provider":{"context_length":16385,"max_completion_tokens":4096,"is_moderated":true}},{"id":"mancer/weaver","name":"Mancer: Weaver (alpha)","pricing":{"prompt":"0.0000015","completion":"0.0000015","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1690934400,"top_provider":{"context_length":8000,"max_completion_tokens":1000,"is_moderated":false}},{"id":"anthropic/claude-2.0:beta","name":"Anthropic: Claude v2.0 (self-moderated)","pricing":{"prompt":"0.000008","completion":"0.000024","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1690502400,"top_provider":{"context_length":100000,"max_completion_tokens":4096,"is_moderated":false}},{"id":"anthropic/claude-2.0","name":"Anthropic: Claude v2.0","pricing":{"prompt":"0.000008","completion":"0.000024","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1690502400,"top_provider":{"context_length":100000,"max_completion_tokens":4096,"is_moderated":true}},{"id":"undi95/remm-slerp-l2-13b","name":"ReMM SLERP 13B","pricing":{"prompt":"0.0000008","completion":"0.0000012","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1689984000,"top_provider":{"context_length":4096,"max_completion_tokens":4096,"is_moderated":false}},{"id":"gryphe/mythomax-l2-13b","name":"MythoMax 13B","pricing":{"prompt":"0.000000065","completion":"0.000000065","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1688256000,"top_provider":{"context_length":4096,"max_completion_tokens":4096,"is_moderated":false}},{"id":"openai/gpt-4","name":"OpenAI: GPT-4","pricing":{"prompt":"0.00003","completion":"0.00006","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1685232000,"top_provider":{"context_length":8191,"max_completion_tokens":4096,"is_moderated":true}},{"id":"openai/gpt-4-0314","name":"OpenAI: GPT-4 (older v0314)","pricing":{"prompt":"0.00003","completion":"0.00006","request":"0","image":"0","web_search":"0","internal_reasoning":"0"},"created":1685232000,"top_provider":{"context_length":8191,"max_completion_tokens":4096,"is_moderated":true}}]
\ No newline at end of file
diff --git a/packages/kbot/dist/main_node.js b/packages/kbot/dist/main_node.js
index a229f4ca..b440a4c4 100644
--- a/packages/kbot/dist/main_node.js
+++ b/packages/kbot/dist/main_node.js
@@ -182560,6 +182560,7 @@ const complete_options = async (opts) => {
options.client = client;
options.variables = { ...options.variables, ...variables(options) };
options.collector = collector(options, client);
+ options.format = opts.format;
options.onRun = options.onRun || (async (options) => options);
return options;
}
@@ -182631,7 +182632,6 @@ const complete_params = async (options, messages) => {
if (options.mode === zod_schema/* E_Mode */.IA.TOOLS || options.mode === zod_schema/* E_Mode */.IA.ASSISTANT) {
params.tools = await (0,tools/* load */.Hh)(options);
params.tool_choice = 'auto';
- //params.parallel_tool_calls = false
}
return params;
};
@@ -182839,7 +182839,7 @@ const run = async (opts) => {
}
return ret;
};
-//# sourceMappingURL=data:application/json;base64,{"version":3,"file":"run.js","sourceRoot":"","sources":["../../src/commands/run.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,IAAI,MAAM,WAAW,CAAA;AACjC,OAAO,EAAE,MAAM,EAAE,MAAM,YAAY,CAAA;AACnC,OAAO,EAAE,GAAG,IAAI,UAAU,EAAE,MAAM,cAAc,CAAA;AAEhD,OAAO,EAAE,QAAQ,EAAE,MAAM,MAAM,CAAA;AAC/B,OAAO,EAAE,IAAI,IAAI,GAAG,EAAE,MAAM,kBAAkB,CAAA;AAC9C,OAAO,EAAE,IAAI,IAAI,MAAM,EAAE,MAAM,qBAAqB,CAAA;AACpD,OAAO,EAAE,IAAI,IAAI,KAAK,EAAE,MAAM,oBAAoB,CAAA;AAClD,OAAO,EAAE,IAAI,IAAI,IAAI,EAAE,MAAM,mBAAmB,CAAA;AAChD,OAAO,EAAE,aAAa,EAAE,OAAO,EAAE,MAAM,EAAE,UAAU,EAAE,MAAM,mBAAmB,CAAA;AAC9E,OAAO,EAAE,OAAO,EAAE,QAAQ,EAAE,MAAM,2BAA2B,CAAA;AAK7D,OAAO,EAAE,MAAM,EAAW,MAAM,OAAO,CAAA;AACvC,OAAO,EAAE,YAAY,EAAE,MAAM,cAAc,CAAA;AAC3C,OAAO,EAAE,aAAa,EAAE,MAAM,kBAAkB,CAAA;AAChD,OAAO,EAAE,GAAG,EAAE,MAAM,cAAc,CAAA;AAClC,OAAO,EAAE,QAAQ,EAAE,MAAM,YAAY,CAAA;AACrC,OAAO,EAAE,OAAO,EAAE,MAAM,mBAAmB,CAAA;AAC3C,OAAO,EAAE,SAAS,EAAE,MAAM,iBAAiB,CAAA;AAC3C,OAAO,EAAE,IAAI,IAAI,WAAW,EAAE,MAAM,eAAe,CAAA;AACnD,OAAO,EAAE,IAAI,IAAI,SAAS,EAAE,MAAM,aAAa,CAAA;AAC/C,OAAO,EAAE,WAAW,EAAE,MAAM,EAAE,MAAM,cAAc,CAAA;AAClD,OAAO,EAAE,SAAS,EAAE,MAAM,iBAAiB,CAAA;AAC3C,OAAO,EAAE,MAAM,EAAE,MAAM,kBAAkB,CAAA;AAEzC,OAAO,EAAE,aAAa,EAAE,MAAM,qBAAqB,CAAA;AACnD,OAAO,EAAE,QAAQ,EAAE,MAAM,gBAAgB,CAAA;AACzC,OAAO,EAAE,YAAY,EAAE,MAAM,oBAAoB,CAAA;AACjD,OAAO,EAAE,GAAG,EAAE,MAAM,oBAAoB,CAAA;AACxC,OAAO,EAAE,SAAS,EAAE,MAAM,aAAa,CAAA;AAIvC,MAAM,CAAC,MAAM,YAAY,GAAG,CAAC,KAAe,EAAY,EAAE;IACxD,MAAM,gBAAgB,GAAG,CAAC,QAAgB,EAAE,EAAE;QAC1C,MAAM,QAAQ,GAAG,IAAI,CAAC,KAAK,CAAC,QAAQ,CAAC,CAAC,IAAI,CAAC;QAC3C,MAAM,KAAK,GAAG,QAAQ,CAAC,KAAK,CAAC,eAAe,CAAC,CAAC,CAAC,wBAAwB;QACvE,MAAM,OAAO,GAAG,KAAK,CAAC,CAAC,CAAC,QAAQ,CAAC,KAAK,CAAC,CAAC,CAAC,EAAE,EAAE,CAAC,CAAC,CAAC,CAAC,GAAG,CAAC;QACrD,MAAM,QAAQ,GAAG,KAAK,CAAC,CAAC,CAAC,KAAK,CAAC,CAAC,CAAC,CAAC,CAAC,CAAC,QAAQ,CAAC,CAAC,oBAAoB;QAElE,OAAO,EAAE,OAAO,EAAE,QAAQ,EAAE,CAAC;IACjC,CAAC,CAAA;IACD,OAAO,KAAK,CAAC,IAAI,CAAC,CAAC,CAAC,EAAE,CAAC,EAAE,EAAE;QACvB,MAAM,EAAE,OAAO,EAAE,IAAI,EAAE,QAAQ,EAAE,KAAK,EAAE,GAAG,gBAAgB,CAAC,CAAC,CAAC,CAAA;QAC9D,MAAM,EAAE,OAAO,EAAE,IAAI,EAAE,QAAQ,EAAE,KAAK,EAAE,GAAG,gBAAgB,CAAC,CAAC,CAAC,CAAA;QAC9D,IAAI,CAAC,KAAK,CAAC,IAAI,CAAC,IAAI,CAAC,KAAK,CAAC,IAAI,CAAC,EAAE,CAAC;YAC/B,OAAO,IAAI,GAAG,IAAI,IAAI,KAAK,CAAC,aAAa,CAAC,KAAK,EAAE,SAAS,EAAE,EAAE,OAAO,EAAE,IAAI,EAAE,WAAW,EAAE,MAAM,EAAE,CAAC,CAAA;QACvG,CAAC;QACD,OAAO,KAAK,CAAC,aAAa,CAAC,KAAK,EAAE,SAAS,EAAE,EAAE,OAAO,EAAE,IAAI,EAAE,WAAW,EAAE,MAAM,EAAE,CAAC,CAAA;IACxF,CAAC,CAAC,CAAA;AACJ,CAAC,CAAA;AAID;;;;;GAKG;AACH,MAAM,CAAC,MAAM,gBAAgB,GAAG,KAAK,EAAE,IAAe,EAA6B,EAAE;IACnF,IAAI,CAAC,MAAM,GAAG,SAAS,CAAC,IAAI,CAAC,CAAA;IAE7B,MAAM,MAAM,GAAG,IAAI,CAAC,OAAO,CAAC,IAAI,CAAC,MAAM,IAAI,IAAI,CAAC,IAAI,IAAI,GAAG,CAAC,CAAA;IAC5D,IAAI,CAAC,MAAM,CAAC,MAAM,CAAC,EAAE,CAAC;QACpB,GAAG,CAAC,MAAM,CAAC,CAAA;IACb,CAAC;IAED,IAAI,CAAC,OAAO,GAAG,OAAO,CAAC,IAAI,CAAC,OAAO,CAAC,CAAA;IACpC,IAAI,CAAC,YAAY,GAAG,OAAO,CAAC,IAAI,CAAC,YAAY,CAAC,CAAA;IAC9C,IAAI,CAAC,OAAO,GAAG,OAAO,CAAC,IAAI,CAAC,OAAO,CAAC,CAAA;IACpC,IAAI,CAAC,OAAO,GAAG,OAAO,CAAC,IAAI,CAAC,OAAO,IAAI,EAAE,CAAC,CAAA;IAC1C,IAAI,CAAC,SAAS,GAAG,IAAI,CAAC,OAAO,CAAC,CAAC,CAAC,MAAM,WAAW,CAAC,IAAI,CAAC,CAAC,CAAC,CAAC,EAAE,CAAA;IAE5D,IAAI,CAAC;QACH,MAAM,MAAM,GAAG,aAAa,EAAE,CAAC,KAAK,CAAC,IAAI,CAAC,CAAC;QAC3C,MAAM,OAAO,GAAG,MAAmB,CAAC;QAEpC,MAAM,MAAM,GAAG,YAAY,CAAC,OAAO,CAAC,CAAA;QACpC,IAAI,CAAC,MAAM,EAAE,CAAC;YACZ,IAAI,CAAC,MAAM,CAAC,KAAK,CAAC,yBAAyB,CAAC,CAAA;YAC5C,OAAO,IAAI,CAAA;QACb,CAAC;QAED,OAAO,CAAC,MAAM,GAAG,MAAM,CAAA;QACvB,OAAO,CAAC,SAAS,GAAG,EAAE,GAAG,OAAO,CAAC,SAAS,EAAE,GAAG,SAAS,CAAC,OAAO,CAAC,EAAE,CAAA;QACnE,OAAO,CAAC,SAAS,GAAG,SAAS,CAAC,OAAO,EAAE,MAAM,CAAC,CAAA;QAC9C,OAAO,CAAC,KAAK,GAAG,OAAO,CAAC,KAAK,IAAI,CAAC,KAAK,EAAE,OAAO,EAAE,EAAE,CAAC,OAAO,CAAC,CAAA;QAE7D,OAAO,OAAO,CAAA;IAChB,CAAC;IAAC,OAAO,KAAK,EAAE,CAAC;QACf,IAAI,CAAC,MAAM,CAAC,KAAK,CAAC,0BAA0B,EAAE,KAAK,CAAC,OAAO,EAAE,KAAK,CAAC,MAAM,EAAE,KAAK,CAAC,KAAK,CAAC,CAAA;QACvF,OAAO,IAAI,CAAA;IACb,CAAC;AACH,CAAC,CAAA;AAED;;;;;;GAMG;AACH,MAAM,CAAC,MAAM,iBAAiB,GAAG,KAAK,EACpC,IAAe,EACf,OAAkB,EAIjB,EAAE;IACH,IAAI,YAAY,GAAsC,EAAE,CAAA;IAExD,MAAM,aAAa,GAAG,MAAM,MAAM,CAAC,IAAI,CAAC,CAAA;IACxC,IAAI,CAAC,aAAa,EAAE,OAAO,EAAE,CAAC;QAC5B,OAAO,EAAE,QAAQ,EAAE,EAAE,EAAE,KAAK,EAAE,EAAE,EAAE,CAAA;IACpC,CAAC;IACD,YAAY,CAAC,IAAI,CAAC,aAA2C,CAAC,CAAA;IAC9D,YAAY,CAAC,IAAI,CAAC,CAAC,MAAM,WAAW,CAAC,IAAI,CAAC,CAA+B,CAAC,CAAA;IAE1E,MAAM,WAAW,GAAG,IAAI,CAAC,OAAO,CAAC,OAAO,CAAC,IAAI,IAAI,GAAG,CAAC,CAAA;IACrD,IAAI,WAAW,GAAG,MAAM,GAAG,CAAC,WAAW,EAAE,OAAO,CAAC,OAAO,EAAE,OAAO,CAAC,IAAI,EAAE,CAAA;IAExE,MAAM,qBAAqB,GAAiC,WAAW,CAAC,GAAG,CAAC,CAAC,CAAC,EAAE;QAC9E,IAAI,CAAC,CAAC,IAAI,IAAI,CAAC,CAAC,OAAO,IAAI,OAAO,CAAC,CAAC,OAAO,KAAK,QAAQ,EAAE,CAAC;YACzD,MAAM,QAAQ,GAAG,MAAM,CAAC,CAAC,CAAC,IAAI,CAAC,CAAA;YAC/B,uGAAuG;YACvG,IAAI,QAAQ,IAAI,CAAC,QAAQ,CAAC,UAAU,CAAC,OAAO,CAAC,IAAI,CAAC,kBAAkB,EAAE,iBAAiB,EAAE,wBAAwB,EAAE,wBAAwB,CAAC,CAAC,QAAQ,CAAC,QAAQ,CAAC,CAAC,EAAE,CAAC;gBACjK,IAAI,OAAO,CAAC,IAAI,KAAK,MAAM,EAAE,CAAC;oBAC5B,MAAM,YAAY,GAAG,IAAI,CAAC,OAAO,CAAC,OAAO,CAAC,IAAI,IAAI,GAAG,EAAE,CAAC,CAAC,IAAI,CAAC,CAAA;oBAC9D,MAAM,YAAY,GAAG,CAAC,CAAC,IAAI,CAAA,CAAC,qDAAqD;oBACjF,MAAM,uBAAuB,GAAG,UAAU,EAAE,CAAA;oBAC5C,MAAM,UAAU,GAAG;;QAErB,YAAY;iBACH,YAAY;OACtB,uBAAuB;;;;CAI7B,CAAA;oBACS,OAAO,EAAE,GAAG,CAAC,EAAE,OAAO,EAAE,UAAU,GAAG,CAAC,CAAC,OAAO,EAAE,IAAI,EAAE,MAAM,EAAE,CAAA;gBAChE,CAAC;qBAAM,CAAC;oBACN,OAAO,EAAE,GAAG,CAAC,EAAE,IAAI,EAAE,MAAM,EAAE,CAAA;gBAC/B,CAAC;YACH,CAAC;QACH,CAAC;QACD,OAAO,EAAE,GAAG,CAAC,EAAE,IAAI,EAAE,MAAM,EAAE,CAAA;IAC/B,CAAC,CAAC,CAAA;IAEF,YAAY,GAAG,CAAC,GAAG,YAAY,EAAE,GAAG,qBAAqB,CAAC,CAAA;IAC1D,OAAO,EAAE,QAAQ,EAAE,YAAY,EAAE,KAAK,EAAE,WAAW,EAAE,CAAA;AACvD,CAAC,CAAA;AAED;;;;;;GAMG;AACH,MAAM,CAAC,MAAM,eAAe,GAAG,KAAK,EAClC,OAAkB,EAClB,QAA2C,EACG,EAAE;IAChD,MAAM,MAAM,GAAG;QACb,KAAK,EAAE,OAAO,CAAC,KAAK;QACpB,QAAQ;QACR,KAAK,EAAE,EAAE;KAC6B,CAAA;IAExC,IAAI,OAAO,CAAC,IAAI,KAAK,MAAM,CAAC,KAAK,IAAI,OAAO,CAAC,IAAI,KAAK,MAAM,CAAC,SAAS,EAAE,CAAC;QACvE,MAAM,CAAC,KAAK,GAAG,MAAM,SAAS,CAAC,OAAO,CAAC,CAAA;QACvC,MAAM,CAAC,WAAW,GAAG,MAAM,CAAA;QAC3B,oCAAoC;IACtC,CAAC;IAED,OAAO,MAAM,CAAA;AACf,CAAC,CAAA;AAED;;;;;;;GAOG;AACH,MAAM,CAAC,MAAM,eAAe,GAAG,KAAK,EAClC,OAAkB,EAClB,MAAW,EACX,MAA2C,EAChB,EAAE;IAC7B,IAAI,GAAG,GAAQ,IAAI,CAAA;IAEnB,IAAI,CAAC;QACH,QAAQ,OAAO,CAAC,IAAI,EAAE,CAAC;YACrB,KAAK,MAAM,CAAC,UAAU;gBACpB,GAAG,GAAG,MAAM,aAAa,CAAC,MAAM,EAAE,MAAM,EAAE,OAAO,CAAC,CAAA;gBAClD,MAAK;YAEP,KAAK,MAAM,CAAC,KAAK;gBACf,GAAG,GAAG,MAAM,QAAQ,CAAC,MAAM,EAAE,MAAM,EAAE,OAAO,CAAC,CAAA;gBAC7C,MAAK;YAEP,KAAK,MAAM,CAAC,SAAS;gBACnB,GAAG,GAAG,MAAM,YAAY,CAAC,MAAM,EAAE,MAAM,EAAE,OAAO,CAAC,CAAA;gBACjD,MAAK;YAEP;gBACE,MAAM,IAAI,KAAK,CAAC,qBAAqB,OAAO,CAAC,IAAI,EAAE,CAAC,CAAA;QACxD,CAAC;IACH,CAAC;IAAC,OAAO,CAAC,EAAE,CAAC;QACX,OAAO,CAAC,MAAM,EAAE,KAAK,CAAC,iBAAiB,OAAO,CAAC,IAAI,UAAU,CAAC,CAAC,OAAO,EAAE,CAAC,CAAA;IAC3E,CAAC;IAED,OAAO,GAAG,CAAA;AACZ,CAAC,CAAA;AAED;;;;;GAKG;AACH,MAAM,CAAC,MAAM,UAAU,GAAG,KAAK,EAAE,IAAe,EAA6B,EAAE;IAC7E,MAAM,cAAc,GAAG,CAAC,GAAG,IAAI,CAAC,OAAO,CAAC,CAAA;IACxC,IAAI,OAAO,GAAG,MAAM,gBAAgB,CAAC,IAAI,CAAC,CAAA;IAC1C,IAAI,CAAC,OAAO,EAAE,CAAC;QACb,OAAO,IAAI,CAAA;IACb,CAAC;IAED,MAAM,MAAM,GAAG,OAAO,CAAC,MAAM,CAAA;IAE7B,MAAM,EAAE,QAAQ,EAAE,KAAK,EAAE,GAAG,MAAM,iBAAiB,CAAC,IAAI,EAAE,OAAO,CAAC,CAAA;IAClE,IAAG,cAAc,CAAC,MAAM,GAAG,CAAC,IAAI,KAAK,CAAC,MAAM,KAAK,CAAC,EAAE,CAAC;QACnD,OAAO,EAAE,CAAA;IACX,CAAC;IACD,IAAI,QAAQ,CAAC,MAAM,KAAK,CAAC,EAAE,CAAC;QAC1B,OAAO,EAAE,CAAA;IACX,CAAC;IACD,MAAM,MAAM,GAAG,MAAM,eAAe,CAAC,OAAO,EAAE,QAAQ,CAAC,CAAA;IAEvD,MAAM,MAAM,GAAG,IAAI,CAAC,OAAO,CAAC,OAAO,CAAC,IAAI,CAAC,IAAI,IAAI,QAAQ,CAAC,CAAC,CAAA;IAC3D,MAAM,UAAU,GAAG,IAAI,CAAC,IAAI,CAAC,MAAM,EAAE,aAAa,CAAC,CAAA;IACnD,KAAK,CAAC,UAAU,EAAE,IAAI,CAAC,SAAS,CAAC,EAAE,GAAG,MAAM,EAAE,EAAE,IAAI,EAAE,CAAC,CAAC,CAAC,CAAA;IACzD,IAAI,IAAI,CAAC,IAAI,KAAK,MAAM,CAAC,KAAK,EAAE,CAAC;QAC/B,OAAO,CAAC,MAAM,CAAC,KAAK,CAAC,YAAY,OAAO,CAAC,IAAI,MAAM,OAAO,CAAC,KAAK,MAAM,OAAO,CAAC,MAAM,OAAO,KAAK,CAAC,MAAM,uBAAuB,IAAI,CAAC,OAAO,CAAC,OAAO,CAAC,IAAI,IAAI,GAAG,CAAC,SAAS,OAAO,CAAC,OAAO,EAAE,EAAE,KAAK,CAAC,GAAG,CAAC,CAAC,CAAC,EAAE,CAAC,CAAC,CAAC,IAAI,CAAC,EAAE,MAAM,CAAC,KAAK,CAAC,GAAG,CAAC,CAAC,CAAC,EAAE,CAAC,GAAG,CAAC,CAAC,QAAQ,CAAC,IAAI,MAAM,CAAC,CAAC,QAAQ,CAAC,WAAW,EAAE,CAAC,CAAC,CAAA;IAC7R,CAAC;SAAM,CAAC;QACN,OAAO,CAAC,MAAM,CAAC,KAAK,CAAC,YAAY,OAAO,CAAC,IAAI,MAAM,OAAO,CAAC,KAAK,MAAM,OAAO,CAAC,MAAM,OAAO,KAAK,CAAC,MAAM,uBAAuB,IAAI,CAAC,OAAO,CAAC,OAAO,CAAC,IAAI,IAAI,GAAG,CAAC,SAAS,OAAO,CAAC,OAAO,EAAE,EAAE,KAAK,CAAC,GAAG,CAAC,CAAC,CAAC,EAAE,CAAC,CAAC,CAAC,IAAI,CAAC,CAAC,CAAA;IACpN,CAAC;IACD,OAAO,GAAG,MAAM,OAAO,CAAC,KAAK,CAAC,OAAO,CAAC,IAAI,OAAO,CAAA;IACjD,MAAM,GAAG,GAAG,MAAM,eAAe,CAAC,OAAO,EAAE,MAAM,EAAE,MAAM,CAAC,CAAA;IAC1D,OAAO,GAAG,CAAA;AACZ,CAAC,CAAA;AAED;;;;;;;;;GASG;AACH,SAAS,YAAY,CAAC,KAAa;IACjC,+BAA+B;IAC/B,IAAI,QAAQ,CAAC,KAAK,CAAC,EAAE,CAAC;QACpB,OAAO,CAAC,KAAK,CAAC,CAAA;IAChB,CAAC;IAED,+CAA+C;IAC/C,IAAI,MAAM,CAAC,IAAI,CAAC,OAAO,CAAC,OAAO,CAAC,KAAK,CAAC,CAAC,CAAC,EAAE,CAAC;QACzC,OAAO,CAAC,KAAK,CAAC,CAAA;IAChB,CAAC;IAED,0CAA0C;IAC1C,IAAI,QAAQ,CAAC,KAAK,CAAC,EAAE,CAAC;QACpB,OAAO,CAAC,KAAK,CAAC,CAAA;IAChB,CAAC;IAED,6FAA6F;IAC7F,MAAM,cAAc,GAAG,iCAAiC,CAAC;IAEzD,MAAM,YAAY,GAAa,EAAE,CAAC;IAClC,IAAI,KAA6B,CAAC;IAElC,uCAAuC;IACvC,OAAO,CAAC,KAAK,GAAG,cAAc,CAAC,IAAI,CAAC,KAAK,CAAC,CAAC,KAAK,IAAI,EAAE,CAAC;QACrD,YAAY,CAAC,IAAI,CAAC,KAAK,CAAC,KAAK,CAAC,CAAC;IACjC,CAAC;IAED,8BAA8B;IAC9B,IAAI,CAAC,YAAY,CAAC,MAAM,EAAE,CAAC;QACzB,OAAO,EAAE,CAAC;IACZ,CAAC;IAED,gDAAgD;IAChD,MAAM,KAAK,GAAa,EAAE,CAAC;IAC3B,KAAK,IAAI,CAAC,GAAG,CAAC,EAAE,CAAC,GAAG,YAAY,CAAC,MAAM,EAAE,CAAC,EAAE,EAAE,CAAC;QAC7C,MAAM,KAAK,GAAG,YAAY,CAAC,CAAC,CAAC,CAAC;QAC9B,MAAM,GAAG,GAAG,CAAC,GAAG,YAAY,CAAC,MAAM,GAAG,CAAC,CAAC,CAAC,CAAC,YAAY,CAAC,CAAC,GAAG,CAAC,CAAC,CAAC,CAAC,CAAC,KAAK,CAAC,MAAM,CAAC;QAC7E,MAAM,IAAI,GAAG,KAAK,CAAC,SAAS,CAAC,KAAK,EAAE,GAAG,CAAC,CAAC,IAAI,EAAE,CAAC;QAChD,IAAI,IAAI,EAAE,CAAC;YACT,KAAK,CAAC,IAAI,CAAC,IAAI,CAAC,CAAC;QACnB,CAAC;IACH,CAAC;IAED,OAAO,KAAK,CAAC;AACf,CAAC;AAED,SAAS,aAAa,CAAI,MAAa;IACrC,OAAO,MAAM,CAAC,MAAM,CAAC,CAAC,WAAW,EAAE,OAAO,EAAE,EAAE;QAC5C,OAAO,WAAW,CAAC,MAAM,CAAC,OAAO,CAAC,CAAC;IACrC,CAAC,EAAE,EAAS,CAAC,CAAC;AAChB,CAAC;AAED,MAAM,CAAC,MAAM,GAAG,GAAG,KAAK,EAAE,IAAe,EAA+B,EAAE;IACxE,MAAM,GAAG,GAAuB,EAAE,CAAA;IAClC,MAAM,MAAM,GAAG,IAAI,MAAM,CAAU,EAAE,QAAQ,EAAE,IAAI,CAAC,QAAQ,IAAI,CAAC,EAAE,CAAC,CAAA;IACpE,IAAI,CAAC,MAAM,GAAG,MAAM,CAAA;IAEpB,IAAI,IAAI,CAAC,OAAO,EAAE,CAAC;QACjB,IAAI,QAAQ,CAAC,IAAI,CAAC,OAAO,CAAC,EAAE,CAAC;YAC3B,IAAI,CAAC,OAAO,GAAG,CAAC,IAAI,CAAC,OAAO,CAAC,CAAA;QAC/B,CAAC;QACD,IAAI,OAAO,CAAC,IAAI,CAAC,OAAO,CAAC,EAAE,CAAC;YAC1B,6CAA6C;YAC7C,MAAM,eAAe,GAAG,IAAI,CAAC,OAAO,CAAC,MAAM,CAAC,CAAC,IAAY,EAAE,EAAE,CAAC,QAAQ,CAAC,IAAI,CAAC,IAAI,QAAQ,CAAC,IAAI,CAAC,CAAC,CAAA;YAC/F,4BAA4B;YAC5B,MAAM,WAAW,GAAG,IAAI,CAAC,OAAO,CAAC,MAAM,CAAC,CAAC,IAAY,EAAE,EAAE,CAAC,CAAC,QAAQ,CAAC,IAAI,CAAC,IAAI,CAAC,QAAQ,CAAC,IAAI,CAAC,CAAC,CAAA;YAC7F,MAAM,cAAc,GAAG,aAAa,CAAC,WAAW,CAAC,GAAG,CAAC,YAAY,CAAC,CAAC,CAAA;YACnE,+CAA+C;YAC/C,IAAI,CAAC,OAAO,GAAG,CAAC,GAAG,eAAe,EAAE,GAAG,cAAc,CAAC,CAAA;QACxD,CAAC;IACH,CAAC;SAAM,CAAC;QACN,IAAI,CAAC,OAAO,GAAG,EAAE,CAAA;IACnB,CAAC;IAED,iDAAiD;IACjD,IAAI,IAAI,CAAC,OAAO,EAAE,CAAC;QACjB,IAAI,QAAQ,CAAC,IAAI,CAAC,OAAO,CAAC,EAAE,CAAC;YAC3B,IAAI,CAAC,OAAO,GAAG,CAAC,IAAI,CAAC,OAAO,CAAC,CAAA;QAC/B,CAAC;QACD,IAAI,OAAO,CAAC,IAAI,CAAC,OAAO,CAAC,EAAE,CAAC;YAC1B,MAAM,eAAe,GAAG,IAAI,CAAC,OAAO,CAAC,MAAM,CAAC,CAAC,IAAY,EAAE,EAAE,CAAC,QAAQ,CAAC,IAAI,CAAC,IAAI,QAAQ,CAAC,IAAI,CAAC,CAAC,CAAC;YAChG,MAAM,WAAW,GAAG,IAAI,CAAC,OAAO,CAAC,MAAM,CAAC,CAAC,IAAY,EAAE,EAAE,CAAC,CAAC,QAAQ,CAAC,IAAI,CAAC,IAAI,CAAC,QAAQ,CAAC,IAAI,CAAC,CAAC,CAAC;YAC9F,MAAM,cAAc,GAAG,aAAa,CAAC,WAAW,CAAC,GAAG,CAAC,YAAY,CAAC,CAAC,CAAC;YACpE,IAAI,CAAC,OAAO,GAAG,CAAC,GAAG,eAAe,EAAE,GAAG,cAAc,CAAC,CAAC;QACzD,CAAC;IACH,CAAC;SAAM,CAAC;QACN,IAAI,CAAC,OAAO,GAAG,EAAE,CAAA;IACnB,CAAC;IAED,IAAI,IAAI,CAAC,IAAI,EAAE,CAAC;QACd,IAAI,KAAK,GAAa,EAAE,CAAA;QACxB,IAAI,OAAO,CAAC,IAAI,CAAC,IAAI,CAAC,EAAE,CAAC;YACvB,KAAK,GAAG,IAAI,CAAC,IAAI,CAAA;QACnB,CAAC;aAAM,IAAI,MAAM,CAAC,IAAI,CAAC,IAAI,CAAC,IAAI,MAAM,CAAC,IAAI,CAAC,IAAI,CAAC,IAAI,IAAI,CAAC,KAAK,CAAC,IAAI,CAAC,IAAI,CAAC,CAAC,GAAG,KAAK,OAAO,EAAE,CAAC;YAC3F,KAAK,GAAG,IAAI,CAAC,IAAI,CAAC,IAAI,EAAE,MAAM,CAAO,IAAI,EAAE,CAAA;QAC7C,CAAC;aAAM,IAAI,QAAQ,CAAC,IAAI,CAAC,IAAI,CAAC,EAAE,CAAC;YAC/B,MAAM,IAAI,GAAG,UAAU,CAAC,aAAa,CAAC,IAAI,CAAC,OAAO,CAAC,OAAO,CAAC,IAAI,CAAC,IAAI,CAAC,CAAC,CAAC,EAAC,KAAK,EAAC;gBAC5E,QAAQ,EAAE,IAAI;aACf,CAAC,CAAA;YACF,KAAK,GAAG,YAAY,CAAC,IAAI,CAAC,KAAK,CAAC,CAAA;QAClC,CAAC;aAAM,IAAI,MAAM,CAAC,IAAI,CAAC,IAAI,CAAC,IAAI,MAAM,CAAC,IAAI,CAAC,IAAI,CAAC,EAAE,CAAC;YAClD,KAAK,GAAG,CAAC,IAAI,CAAC,IAAI,CAAC,CAAA;QACrB,CAAC;aAAM,IAAI,QAAQ,CAAC,IAAI,CAAC,IAAI,CAAC,EAAE,CAAC;YAC/B,KAAK,GAAG,IAAI,CAAC,IAAI,CAAC,KAAK,CAAC,GAAG,CAAC,CAAA;QAC9B,CAAC;QACD,IAAI,KAAK,CAAC,MAAM,KAAK,CAAC,EAAE,CAAC;YACvB,IAAI,CAAC,MAAM,CAAC,IAAI,CAAC,6BAA6B,IAAI,CAAC,IAAI,aAAa,IAAI,CAAC,IAAI,EAAE,CAAC,CAAA;YAChF,OAAO,GAAG,CAAA;QACZ,CAAC;QACD,IAAI,CAAC,MAAM,CAAC,IAAI,CAAC,cAAc,KAAK,CAAC,MAAM,2BAA2B,IAAI,CAAC,IAAI,KAAK,CAAC,CAAA;QACrF,MAAM,OAAO,GAAG,GAAG,EAAE,CAAA;QACrB,KAAK,MAAM,IAAI,IAAI,KAAK,EAAE,CAAC;YACzB,MAAM,QAAQ,GAAG;gBACf,GAAG,IAAI;gBACP,IAAI,EAAE,IAAI;gBACV,SAAS,EAAE,EAAE,IAAI,EAAE,IAAI,EAAE;aAC1B,CAAA;YACD,MAAM,KAAK,GAAG,OAAO,CAAC,IAAI,CAAC,CAAC,CAAC,EAAE,CAAC,CAAC,CAAC,EAAE,KAAK,IAAI,CAAC,CAAA;YAC9C,IAAI,KAAK,EAAE,CAAC;gBACV,QAAQ,CAAC,KAAK,GAAG,IAAI,CAAA;YACvB,CAAC;YAED,IAAI,2BAA2B,GAAG,CAAC,aAAa,CAAC,IAAI,CAAC,CAAC,CAAC;YAExD,QAAQ,CAAC,OAAO,GAAG,CAAC,GAAG,IAAI,CAAC,OAAO,EAAE,GAAG,2BAA2B,CAAC,CAAC;YAErE,MAAM,MAAM,GAAG,MAAM,UAAU,CAAC,QAAQ,CAAC,CAAA;YACzC,IAAI,MAAM,KAAK,SAAS,EAAE,CAAC;gBACzB,GAAG,CAAC,IAAI,CAAC,MAAM,CAAC,CAAA;YAClB,CAAC;QACH,CAAC;IACH,CAAC;SAAM,CAAC;QACN,MAAM,MAAM,GAAG,MAAM,UAAU,CAAC,IAAI,CAAC,CAAA;QACrC,IAAI,MAAM,KAAK,SAAS,EAAE,CAAC;YACzB,GAAG,CAAC,IAAI,CAAC,MAAM,CAAC,CAAA;QAClB,CAAC;IACH,CAAC;IACD,OAAO,GAAG,CAAA;AACZ,CAAC,CAAA"}
+//# sourceMappingURL=data:application/json;base64,{"version":3,"file":"run.js","sourceRoot":"","sources":["../../src/commands/run.ts"],"names":[],"mappings":"AAAA,OAAO,KAAK,IAAI,MAAM,WAAW,CAAA;AACjC,OAAO,EAAE,MAAM,EAAE,MAAM,YAAY,CAAA;AACnC,OAAO,EAAE,GAAG,IAAI,UAAU,EAAE,MAAM,cAAc,CAAA;AAEhD,OAAO,EAAE,QAAQ,EAAE,MAAM,MAAM,CAAA;AAC/B,OAAO,EAAE,IAAI,IAAI,GAAG,EAAE,MAAM,kBAAkB,CAAA;AAC9C,OAAO,EAAE,IAAI,IAAI,MAAM,EAAE,MAAM,qBAAqB,CAAA;AACpD,OAAO,EAAE,IAAI,IAAI,KAAK,EAAE,MAAM,oBAAoB,CAAA;AAClD,OAAO,EAAE,IAAI,IAAI,IAAI,EAAE,MAAM,mBAAmB,CAAA;AAChD,OAAO,EAAE,aAAa,EAAE,OAAO,EAAE,MAAM,EAAE,UAAU,EAAE,MAAM,mBAAmB,CAAA;AAC9E,OAAO,EAAE,OAAO,EAAE,QAAQ,EAAE,MAAM,2BAA2B,CAAA;AAK7D,OAAO,EAAE,MAAM,EAAW,MAAM,OAAO,CAAA;AACvC,OAAO,EAAE,YAAY,EAAE,MAAM,cAAc,CAAA;AAC3C,OAAO,EAAE,aAAa,EAAE,MAAM,kBAAkB,CAAA;AAChD,OAAO,EAAE,GAAG,EAAE,MAAM,cAAc,CAAA;AAClC,OAAO,EAAE,QAAQ,EAAE,MAAM,YAAY,CAAA;AACrC,OAAO,EAAE,OAAO,EAAE,MAAM,mBAAmB,CAAA;AAC3C,OAAO,EAAE,SAAS,EAAE,MAAM,iBAAiB,CAAA;AAC3C,OAAO,EAAE,IAAI,IAAI,WAAW,EAAE,MAAM,eAAe,CAAA;AACnD,OAAO,EAAE,IAAI,IAAI,SAAS,EAAE,MAAM,aAAa,CAAA;AAC/C,OAAO,EAAE,WAAW,EAAE,MAAM,EAAE,MAAM,cAAc,CAAA;AAClD,OAAO,EAAE,SAAS,EAAE,MAAM,iBAAiB,CAAA;AAC3C,OAAO,EAAE,MAAM,EAAE,MAAM,kBAAkB,CAAA;AAEzC,OAAO,EAAE,aAAa,EAAE,MAAM,qBAAqB,CAAA;AACnD,OAAO,EAAE,QAAQ,EAAE,MAAM,gBAAgB,CAAA;AACzC,OAAO,EAAE,YAAY,EAAE,MAAM,oBAAoB,CAAA;AACjD,OAAO,EAAE,GAAG,EAAE,MAAM,oBAAoB,CAAA;AACxC,OAAO,EAAE,SAAS,EAAE,MAAM,aAAa,CAAA;AAIvC,MAAM,CAAC,MAAM,YAAY,GAAG,CAAC,KAAe,EAAY,EAAE;IACxD,MAAM,gBAAgB,GAAG,CAAC,QAAgB,EAAE,EAAE;QAC1C,MAAM,QAAQ,GAAG,IAAI,CAAC,KAAK,CAAC,QAAQ,CAAC,CAAC,IAAI,CAAC;QAC3C,MAAM,KAAK,GAAG,QAAQ,CAAC,KAAK,CAAC,eAAe,CAAC,CAAC,CAAC,wBAAwB;QACvE,MAAM,OAAO,GAAG,KAAK,CAAC,CAAC,CAAC,QAAQ,CAAC,KAAK,CAAC,CAAC,CAAC,EAAE,EAAE,CAAC,CAAC,CAAC,CAAC,GAAG,CAAC;QACrD,MAAM,QAAQ,GAAG,KAAK,CAAC,CAAC,CAAC,KAAK,CAAC,CAAC,CAAC,CAAC,CAAC,CAAC,QAAQ,CAAC,CAAC,oBAAoB;QAElE,OAAO,EAAE,OAAO,EAAE,QAAQ,EAAE,CAAC;IACjC,CAAC,CAAA;IACD,OAAO,KAAK,CAAC,IAAI,CAAC,CAAC,CAAC,EAAE,CAAC,EAAE,EAAE;QACvB,MAAM,EAAE,OAAO,EAAE,IAAI,EAAE,QAAQ,EAAE,KAAK,EAAE,GAAG,gBAAgB,CAAC,CAAC,CAAC,CAAA;QAC9D,MAAM,EAAE,OAAO,EAAE,IAAI,EAAE,QAAQ,EAAE,KAAK,EAAE,GAAG,gBAAgB,CAAC,CAAC,CAAC,CAAA;QAC9D,IAAI,CAAC,KAAK,CAAC,IAAI,CAAC,IAAI,CAAC,KAAK,CAAC,IAAI,CAAC,EAAE,CAAC;YAC/B,OAAO,IAAI,GAAG,IAAI,IAAI,KAAK,CAAC,aAAa,CAAC,KAAK,EAAE,SAAS,EAAE,EAAE,OAAO,EAAE,IAAI,EAAE,WAAW,EAAE,MAAM,EAAE,CAAC,CAAA;QACvG,CAAC;QACD,OAAO,KAAK,CAAC,aAAa,CAAC,KAAK,EAAE,SAAS,EAAE,EAAE,OAAO,EAAE,IAAI,EAAE,WAAW,EAAE,MAAM,EAAE,CAAC,CAAA;IACxF,CAAC,CAAC,CAAA;AACJ,CAAC,CAAA;AAID;;;;;GAKG;AACH,MAAM,CAAC,MAAM,gBAAgB,GAAG,KAAK,EAAE,IAAe,EAA6B,EAAE;IACnF,IAAI,CAAC,MAAM,GAAG,SAAS,CAAC,IAAI,CAAC,CAAA;IAE7B,MAAM,MAAM,GAAG,IAAI,CAAC,OAAO,CAAC,IAAI,CAAC,MAAM,IAAI,IAAI,CAAC,IAAI,IAAI,GAAG,CAAC,CAAA;IAC5D,IAAI,CAAC,MAAM,CAAC,MAAM,CAAC,EAAE,CAAC;QACpB,GAAG,CAAC,MAAM,CAAC,CAAA;IACb,CAAC;IAED,IAAI,CAAC,OAAO,GAAG,OAAO,CAAC,IAAI,CAAC,OAAO,CAAC,CAAA;IACpC,IAAI,CAAC,YAAY,GAAG,OAAO,CAAC,IAAI,CAAC,YAAY,CAAC,CAAA;IAC9C,IAAI,CAAC,OAAO,GAAG,OAAO,CAAC,IAAI,CAAC,OAAO,CAAC,CAAA;IACpC,IAAI,CAAC,OAAO,GAAG,OAAO,CAAC,IAAI,CAAC,OAAO,IAAI,EAAE,CAAC,CAAA;IAC1C,IAAI,CAAC,SAAS,GAAG,IAAI,CAAC,OAAO,CAAC,CAAC,CAAC,MAAM,WAAW,CAAC,IAAI,CAAC,CAAC,CAAC,CAAC,EAAE,CAAA;IAE5D,IAAI,CAAC;QACH,MAAM,MAAM,GAAG,aAAa,EAAE,CAAC,KAAK,CAAC,IAAI,CAAC,CAAC;QAC3C,MAAM,OAAO,GAAG,MAAmB,CAAC;QAEpC,MAAM,MAAM,GAAG,YAAY,CAAC,OAAO,CAAC,CAAA;QACpC,IAAI,CAAC,MAAM,EAAE,CAAC;YACZ,IAAI,CAAC,MAAM,CAAC,KAAK,CAAC,yBAAyB,CAAC,CAAA;YAC5C,OAAO,IAAI,CAAA;QACb,CAAC;QAED,OAAO,CAAC,MAAM,GAAG,MAAM,CAAA;QACvB,OAAO,CAAC,SAAS,GAAG,EAAE,GAAG,OAAO,CAAC,SAAS,EAAE,GAAG,SAAS,CAAC,OAAO,CAAC,EAAE,CAAA;QACnE,OAAO,CAAC,SAAS,GAAG,SAAS,CAAC,OAAO,EAAE,MAAM,CAAC,CAAA;QAC9C,OAAO,CAAC,MAAM,GAAG,IAAI,CAAC,MAAM,CAAA;QAC5B,OAAO,CAAC,KAAK,GAAG,OAAO,CAAC,KAAK,IAAI,CAAC,KAAK,EAAE,OAAO,EAAE,EAAE,CAAC,OAAO,CAAC,CAAA;QAE7D,OAAO,OAAO,CAAA;IAChB,CAAC;IAAC,OAAO,KAAK,EAAE,CAAC;QACf,IAAI,CAAC,MAAM,CAAC,KAAK,CAAC,0BAA0B,EAAE,KAAK,CAAC,OAAO,EAAE,KAAK,CAAC,MAAM,EAAE,KAAK,CAAC,KAAK,CAAC,CAAA;QACvF,OAAO,IAAI,CAAA;IACb,CAAC;AACH,CAAC,CAAA;AAED;;;;;;GAMG;AACH,MAAM,CAAC,MAAM,iBAAiB,GAAG,KAAK,EACpC,IAAe,EACf,OAAkB,EAIjB,EAAE;IACH,IAAI,YAAY,GAAsC,EAAE,CAAA;IAExD,MAAM,aAAa,GAAG,MAAM,MAAM,CAAC,IAAI,CAAC,CAAA;IACxC,IAAI,CAAC,aAAa,EAAE,OAAO,EAAE,CAAC;QAC5B,OAAO,EAAE,QAAQ,EAAE,EAAE,EAAE,KAAK,EAAE,EAAE,EAAE,CAAA;IACpC,CAAC;IACD,YAAY,CAAC,IAAI,CAAC,aAA2C,CAAC,CAAA;IAC9D,YAAY,CAAC,IAAI,CAAC,CAAC,MAAM,WAAW,CAAC,IAAI,CAAC,CAA+B,CAAC,CAAA;IAE1E,MAAM,WAAW,GAAG,IAAI,CAAC,OAAO,CAAC,OAAO,CAAC,IAAI,IAAI,GAAG,CAAC,CAAA;IACrD,IAAI,WAAW,GAAG,MAAM,GAAG,CAAC,WAAW,EAAE,OAAO,CAAC,OAAO,EAAE,OAAO,CAAC,IAAI,EAAE,CAAA;IAExE,MAAM,qBAAqB,GAAiC,WAAW,CAAC,GAAG,CAAC,CAAC,CAAC,EAAE;QAC9E,IAAI,CAAC,CAAC,IAAI,IAAI,CAAC,CAAC,OAAO,IAAI,OAAO,CAAC,CAAC,OAAO,KAAK,QAAQ,EAAE,CAAC;YACzD,MAAM,QAAQ,GAAG,MAAM,CAAC,CAAC,CAAC,IAAI,CAAC,CAAA;YAC/B,uGAAuG;YACvG,IAAI,QAAQ,IAAI,CAAC,QAAQ,CAAC,UAAU,CAAC,OAAO,CAAC,IAAI,CAAC,kBAAkB,EAAE,iBAAiB,EAAE,wBAAwB,EAAE,wBAAwB,CAAC,CAAC,QAAQ,CAAC,QAAQ,CAAC,CAAC,EAAE,CAAC;gBACjK,IAAI,OAAO,CAAC,IAAI,KAAK,MAAM,EAAE,CAAC;oBAC5B,MAAM,YAAY,GAAG,IAAI,CAAC,OAAO,CAAC,OAAO,CAAC,IAAI,IAAI,GAAG,EAAE,CAAC,CAAC,IAAI,CAAC,CAAA;oBAC9D,MAAM,YAAY,GAAG,CAAC,CAAC,IAAI,CAAA,CAAC,qDAAqD;oBACjF,MAAM,uBAAuB,GAAG,UAAU,EAAE,CAAA;oBAC5C,MAAM,UAAU,GAAG;;QAErB,YAAY;iBACH,YAAY;OACtB,uBAAuB;;;;CAI7B,CAAA;oBACS,OAAO,EAAE,GAAG,CAAC,EAAE,OAAO,EAAE,UAAU,GAAG,CAAC,CAAC,OAAO,EAAE,IAAI,EAAE,MAAM,EAAE,CAAA;gBAChE,CAAC;qBAAM,CAAC;oBACN,OAAO,EAAE,GAAG,CAAC,EAAE,IAAI,EAAE,MAAM,EAAE,CAAA;gBAC/B,CAAC;YACH,CAAC;QACH,CAAC;QACD,OAAO,EAAE,GAAG,CAAC,EAAE,IAAI,EAAE,MAAM,EAAE,CAAA;IAC/B,CAAC,CAAC,CAAA;IAEF,YAAY,GAAG,CAAC,GAAG,YAAY,EAAE,GAAG,qBAAqB,CAAC,CAAA;IAC1D,OAAO,EAAE,QAAQ,EAAE,YAAY,EAAE,KAAK,EAAE,WAAW,EAAE,CAAA;AACvD,CAAC,CAAA;AAED;;;;;;GAMG;AACH,MAAM,CAAC,MAAM,eAAe,GAAG,KAAK,EAClC,OAAkB,EAClB,QAA2C,EACG,EAAE;IAChD,MAAM,MAAM,GAAG;QACb,KAAK,EAAE,OAAO,CAAC,KAAK;QACpB,QAAQ;QACR,KAAK,EAAE,EAAE;KAC6B,CAAA;IAExC,IAAI,OAAO,CAAC,IAAI,KAAK,MAAM,CAAC,KAAK,IAAI,OAAO,CAAC,IAAI,KAAK,MAAM,CAAC,SAAS,EAAE,CAAC;QACvE,MAAM,CAAC,KAAK,GAAG,MAAM,SAAS,CAAC,OAAO,CAAC,CAAA;QACvC,MAAM,CAAC,WAAW,GAAG,MAAM,CAAA;IAC7B,CAAC;IAED,OAAO,MAAM,CAAA;AACf,CAAC,CAAA;AAED;;;;;;;GAOG;AACH,MAAM,CAAC,MAAM,eAAe,GAAG,KAAK,EAClC,OAAkB,EAClB,MAAW,EACX,MAA2C,EAChB,EAAE;IAC7B,IAAI,GAAG,GAAQ,IAAI,CAAA;IAEnB,IAAI,CAAC;QACH,QAAQ,OAAO,CAAC,IAAI,EAAE,CAAC;YACrB,KAAK,MAAM,CAAC,UAAU;gBACpB,GAAG,GAAG,MAAM,aAAa,CAAC,MAAM,EAAE,MAAM,EAAE,OAAO,CAAC,CAAA;gBAClD,MAAK;YAEP,KAAK,MAAM,CAAC,KAAK;gBACf,GAAG,GAAG,MAAM,QAAQ,CAAC,MAAM,EAAE,MAAM,EAAE,OAAO,CAAC,CAAA;gBAC7C,MAAK;YAEP,KAAK,MAAM,CAAC,SAAS;gBACnB,GAAG,GAAG,MAAM,YAAY,CAAC,MAAM,EAAE,MAAM,EAAE,OAAO,CAAC,CAAA;gBACjD,MAAK;YAEP;gBACE,MAAM,IAAI,KAAK,CAAC,qBAAqB,OAAO,CAAC,IAAI,EAAE,CAAC,CAAA;QACxD,CAAC;IACH,CAAC;IAAC,OAAO,CAAC,EAAE,CAAC;QACX,OAAO,CAAC,MAAM,EAAE,KAAK,CAAC,iBAAiB,OAAO,CAAC,IAAI,UAAU,CAAC,CAAC,OAAO,EAAE,CAAC,CAAA;IAC3E,CAAC;IAED,OAAO,GAAG,CAAA;AACZ,CAAC,CAAA;AAED;;;;;GAKG;AACH,MAAM,CAAC,MAAM,UAAU,GAAG,KAAK,EAAE,IAAe,EAA6B,EAAE;IAC7E,MAAM,cAAc,GAAG,CAAC,GAAG,IAAI,CAAC,OAAO,CAAC,CAAA;IACxC,IAAI,OAAO,GAAG,MAAM,gBAAgB,CAAC,IAAI,CAAC,CAAA;IAC1C,IAAI,CAAC,OAAO,EAAE,CAAC;QACb,OAAO,IAAI,CAAA;IACb,CAAC;IAED,MAAM,MAAM,GAAG,OAAO,CAAC,MAAM,CAAA;IAE7B,MAAM,EAAE,QAAQ,EAAE,KAAK,EAAE,GAAG,MAAM,iBAAiB,CAAC,IAAI,EAAE,OAAO,CAAC,CAAA;IAClE,IAAG,cAAc,CAAC,MAAM,GAAG,CAAC,IAAI,KAAK,CAAC,MAAM,KAAK,CAAC,EAAE,CAAC;QACnD,OAAO,EAAE,CAAA;IACX,CAAC;IACD,IAAI,QAAQ,CAAC,MAAM,KAAK,CAAC,EAAE,CAAC;QAC1B,OAAO,EAAE,CAAA;IACX,CAAC;IACD,MAAM,MAAM,GAAG,MAAM,eAAe,CAAC,OAAO,EAAE,QAAQ,CAAC,CAAA;IAEvD,MAAM,MAAM,GAAG,IAAI,CAAC,OAAO,CAAC,OAAO,CAAC,IAAI,CAAC,IAAI,IAAI,QAAQ,CAAC,CAAC,CAAA;IAC3D,MAAM,UAAU,GAAG,IAAI,CAAC,IAAI,CAAC,MAAM,EAAE,aAAa,CAAC,CAAA;IACnD,KAAK,CAAC,UAAU,EAAE,IAAI,CAAC,SAAS,CAAC,EAAE,GAAG,MAAM,EAAE,EAAE,IAAI,EAAE,CAAC,CAAC,CAAC,CAAA;IACzD,IAAI,IAAI,CAAC,IAAI,KAAK,MAAM,CAAC,KAAK,EAAE,CAAC;QAC/B,OAAO,CAAC,MAAM,CAAC,KAAK,CAAC,YAAY,OAAO,CAAC,IAAI,MAAM,OAAO,CAAC,KAAK,MAAM,OAAO,CAAC,MAAM,OAAO,KAAK,CAAC,MAAM,uBAAuB,IAAI,CAAC,OAAO,CAAC,OAAO,CAAC,IAAI,IAAI,GAAG,CAAC,SAAS,OAAO,CAAC,OAAO,EAAE,EAAE,KAAK,CAAC,GAAG,CAAC,CAAC,CAAC,EAAE,CAAC,CAAC,CAAC,IAAI,CAAC,EAAE,MAAM,CAAC,KAAK,CAAC,GAAG,CAAC,CAAC,CAAC,EAAE,CAAC,GAAG,CAAC,CAAC,QAAQ,CAAC,IAAI,MAAM,CAAC,CAAC,QAAQ,CAAC,WAAW,EAAE,CAAC,CAAC,CAAA;IAC7R,CAAC;SAAM,CAAC;QACN,OAAO,CAAC,MAAM,CAAC,KAAK,CAAC,YAAY,OAAO,CAAC,IAAI,MAAM,OAAO,CAAC,KAAK,MAAM,OAAO,CAAC,MAAM,OAAO,KAAK,CAAC,MAAM,uBAAuB,IAAI,CAAC,OAAO,CAAC,OAAO,CAAC,IAAI,IAAI,GAAG,CAAC,SAAS,OAAO,CAAC,OAAO,EAAE,EAAE,KAAK,CAAC,GAAG,CAAC,CAAC,CAAC,EAAE,CAAC,CAAC,CAAC,IAAI,CAAC,CAAC,CAAA;IACpN,CAAC;IACD,OAAO,GAAG,MAAM,OAAO,CAAC,KAAK,CAAC,OAAO,CAAC,IAAI,OAAO,CAAA;IACjD,MAAM,GAAG,GAAG,MAAM,eAAe,CAAC,OAAO,EAAE,MAAM,EAAE,MAAM,CAAC,CAAA;IAC1D,OAAO,GAAG,CAAA;AACZ,CAAC,CAAA;AAED;;;;;;;;;GASG;AACH,SAAS,YAAY,CAAC,KAAa;IACjC,+BAA+B;IAC/B,IAAI,QAAQ,CAAC,KAAK,CAAC,EAAE,CAAC;QACpB,OAAO,CAAC,KAAK,CAAC,CAAA;IAChB,CAAC;IAED,+CAA+C;IAC/C,IAAI,MAAM,CAAC,IAAI,CAAC,OAAO,CAAC,OAAO,CAAC,KAAK,CAAC,CAAC,CAAC,EAAE,CAAC;QACzC,OAAO,CAAC,KAAK,CAAC,CAAA;IAChB,CAAC;IAED,0CAA0C;IAC1C,IAAI,QAAQ,CAAC,KAAK,CAAC,EAAE,CAAC;QACpB,OAAO,CAAC,KAAK,CAAC,CAAA;IAChB,CAAC;IAED,6FAA6F;IAC7F,MAAM,cAAc,GAAG,iCAAiC,CAAC;IAEzD,MAAM,YAAY,GAAa,EAAE,CAAC;IAClC,IAAI,KAA6B,CAAC;IAElC,uCAAuC;IACvC,OAAO,CAAC,KAAK,GAAG,cAAc,CAAC,IAAI,CAAC,KAAK,CAAC,CAAC,KAAK,IAAI,EAAE,CAAC;QACrD,YAAY,CAAC,IAAI,CAAC,KAAK,CAAC,KAAK,CAAC,CAAC;IACjC,CAAC;IAED,8BAA8B;IAC9B,IAAI,CAAC,YAAY,CAAC,MAAM,EAAE,CAAC;QACzB,OAAO,EAAE,CAAC;IACZ,CAAC;IAED,gDAAgD;IAChD,MAAM,KAAK,GAAa,EAAE,CAAC;IAC3B,KAAK,IAAI,CAAC,GAAG,CAAC,EAAE,CAAC,GAAG,YAAY,CAAC,MAAM,EAAE,CAAC,EAAE,EAAE,CAAC;QAC7C,MAAM,KAAK,GAAG,YAAY,CAAC,CAAC,CAAC,CAAC;QAC9B,MAAM,GAAG,GAAG,CAAC,GAAG,YAAY,CAAC,MAAM,GAAG,CAAC,CAAC,CAAC,CAAC,YAAY,CAAC,CAAC,GAAG,CAAC,CAAC,CAAC,CAAC,CAAC,KAAK,CAAC,MAAM,CAAC;QAC7E,MAAM,IAAI,GAAG,KAAK,CAAC,SAAS,CAAC,KAAK,EAAE,GAAG,CAAC,CAAC,IAAI,EAAE,CAAC;QAChD,IAAI,IAAI,EAAE,CAAC;YACT,KAAK,CAAC,IAAI,CAAC,IAAI,CAAC,CAAC;QACnB,CAAC;IACH,CAAC;IAED,OAAO,KAAK,CAAC;AACf,CAAC;AAED,SAAS,aAAa,CAAI,MAAa;IACrC,OAAO,MAAM,CAAC,MAAM,CAAC,CAAC,WAAW,EAAE,OAAO,EAAE,EAAE;QAC5C,OAAO,WAAW,CAAC,MAAM,CAAC,OAAO,CAAC,CAAC;IACrC,CAAC,EAAE,EAAS,CAAC,CAAC;AAChB,CAAC;AAED,MAAM,CAAC,MAAM,GAAG,GAAG,KAAK,EAAE,IAAe,EAA+B,EAAE;IACxE,MAAM,GAAG,GAAuB,EAAE,CAAA;IAClC,MAAM,MAAM,GAAG,IAAI,MAAM,CAAU,EAAE,QAAQ,EAAE,IAAI,CAAC,QAAQ,IAAI,CAAC,EAAE,CAAC,CAAA;IACpE,IAAI,CAAC,MAAM,GAAG,MAAM,CAAA;IAEpB,IAAI,IAAI,CAAC,OAAO,EAAE,CAAC;QACjB,IAAI,QAAQ,CAAC,IAAI,CAAC,OAAO,CAAC,EAAE,CAAC;YAC3B,IAAI,CAAC,OAAO,GAAG,CAAC,IAAI,CAAC,OAAO,CAAC,CAAA;QAC/B,CAAC;QACD,IAAI,OAAO,CAAC,IAAI,CAAC,OAAO,CAAC,EAAE,CAAC;YAC1B,6CAA6C;YAC7C,MAAM,eAAe,GAAG,IAAI,CAAC,OAAO,CAAC,MAAM,CAAC,CAAC,IAAY,EAAE,EAAE,CAAC,QAAQ,CAAC,IAAI,CAAC,IAAI,QAAQ,CAAC,IAAI,CAAC,CAAC,CAAA;YAC/F,4BAA4B;YAC5B,MAAM,WAAW,GAAG,IAAI,CAAC,OAAO,CAAC,MAAM,CAAC,CAAC,IAAY,EAAE,EAAE,CAAC,CAAC,QAAQ,CAAC,IAAI,CAAC,IAAI,CAAC,QAAQ,CAAC,IAAI,CAAC,CAAC,CAAA;YAC7F,MAAM,cAAc,GAAG,aAAa,CAAC,WAAW,CAAC,GAAG,CAAC,YAAY,CAAC,CAAC,CAAA;YACnE,+CAA+C;YAC/C,IAAI,CAAC,OAAO,GAAG,CAAC,GAAG,eAAe,EAAE,GAAG,cAAc,CAAC,CAAA;QACxD,CAAC;IACH,CAAC;SAAM,CAAC;QACN,IAAI,CAAC,OAAO,GAAG,EAAE,CAAA;IACnB,CAAC;IAED,iDAAiD;IACjD,IAAI,IAAI,CAAC,OAAO,EAAE,CAAC;QACjB,IAAI,QAAQ,CAAC,IAAI,CAAC,OAAO,CAAC,EAAE,CAAC;YAC3B,IAAI,CAAC,OAAO,GAAG,CAAC,IAAI,CAAC,OAAO,CAAC,CAAA;QAC/B,CAAC;QACD,IAAI,OAAO,CAAC,IAAI,CAAC,OAAO,CAAC,EAAE,CAAC;YAC1B,MAAM,eAAe,GAAG,IAAI,CAAC,OAAO,CAAC,MAAM,CAAC,CAAC,IAAY,EAAE,EAAE,CAAC,QAAQ,CAAC,IAAI,CAAC,IAAI,QAAQ,CAAC,IAAI,CAAC,CAAC,CAAC;YAChG,MAAM,WAAW,GAAG,IAAI,CAAC,OAAO,CAAC,MAAM,CAAC,CAAC,IAAY,EAAE,EAAE,CAAC,CAAC,QAAQ,CAAC,IAAI,CAAC,IAAI,CAAC,QAAQ,CAAC,IAAI,CAAC,CAAC,CAAC;YAC9F,MAAM,cAAc,GAAG,aAAa,CAAC,WAAW,CAAC,GAAG,CAAC,YAAY,CAAC,CAAC,CAAC;YACpE,IAAI,CAAC,OAAO,GAAG,CAAC,GAAG,eAAe,EAAE,GAAG,cAAc,CAAC,CAAC;QACzD,CAAC;IACH,CAAC;SAAM,CAAC;QACN,IAAI,CAAC,OAAO,GAAG,EAAE,CAAA;IACnB,CAAC;IAED,IAAI,IAAI,CAAC,IAAI,EAAE,CAAC;QACd,IAAI,KAAK,GAAa,EAAE,CAAA;QACxB,IAAI,OAAO,CAAC,IAAI,CAAC,IAAI,CAAC,EAAE,CAAC;YACvB,KAAK,GAAG,IAAI,CAAC,IAAI,CAAA;QACnB,CAAC;aAAM,IAAI,MAAM,CAAC,IAAI,CAAC,IAAI,CAAC,IAAI,MAAM,CAAC,IAAI,CAAC,IAAI,CAAC,IAAI,IAAI,CAAC,KAAK,CAAC,IAAI,CAAC,IAAI,CAAC,CAAC,GAAG,KAAK,OAAO,EAAE,CAAC;YAC3F,KAAK,GAAG,IAAI,CAAC,IAAI,CAAC,IAAI,EAAE,MAAM,CAAO,IAAI,EAAE,CAAA;QAC7C,CAAC;aAAM,IAAI,QAAQ,CAAC,IAAI,CAAC,IAAI,CAAC,EAAE,CAAC;YAC/B,MAAM,IAAI,GAAG,UAAU,CAAC,aAAa,CAAC,IAAI,CAAC,OAAO,CAAC,OAAO,CAAC,IAAI,CAAC,IAAI,CAAC,CAAC,CAAC,EAAC,KAAK,EAAC;gBAC5E,QAAQ,EAAE,IAAI;aACf,CAAC,CAAA;YACF,KAAK,GAAG,YAAY,CAAC,IAAI,CAAC,KAAK,CAAC,CAAA;QAClC,CAAC;aAAM,IAAI,MAAM,CAAC,IAAI,CAAC,IAAI,CAAC,IAAI,MAAM,CAAC,IAAI,CAAC,IAAI,CAAC,EAAE,CAAC;YAClD,KAAK,GAAG,CAAC,IAAI,CAAC,IAAI,CAAC,CAAA;QACrB,CAAC;aAAM,IAAI,QAAQ,CAAC,IAAI,CAAC,IAAI,CAAC,EAAE,CAAC;YAC/B,KAAK,GAAG,IAAI,CAAC,IAAI,CAAC,KAAK,CAAC,GAAG,CAAC,CAAA;QAC9B,CAAC;QACD,IAAI,KAAK,CAAC,MAAM,KAAK,CAAC,EAAE,CAAC;YACvB,IAAI,CAAC,MAAM,CAAC,IAAI,CAAC,6BAA6B,IAAI,CAAC,IAAI,aAAa,IAAI,CAAC,IAAI,EAAE,CAAC,CAAA;YAChF,OAAO,GAAG,CAAA;QACZ,CAAC;QACD,IAAI,CAAC,MAAM,CAAC,IAAI,CAAC,cAAc,KAAK,CAAC,MAAM,2BAA2B,IAAI,CAAC,IAAI,KAAK,CAAC,CAAA;QACrF,MAAM,OAAO,GAAG,GAAG,EAAE,CAAA;QACrB,KAAK,MAAM,IAAI,IAAI,KAAK,EAAE,CAAC;YACzB,MAAM,QAAQ,GAAG;gBACf,GAAG,IAAI;gBACP,IAAI,EAAE,IAAI;gBACV,SAAS,EAAE,EAAE,IAAI,EAAE,IAAI,EAAE;aAC1B,CAAA;YACD,MAAM,KAAK,GAAG,OAAO,CAAC,IAAI,CAAC,CAAC,CAAC,EAAE,CAAC,CAAC,CAAC,EAAE,KAAK,IAAI,CAAC,CAAA;YAC9C,IAAI,KAAK,EAAE,CAAC;gBACV,QAAQ,CAAC,KAAK,GAAG,IAAI,CAAA;YACvB,CAAC;YAED,IAAI,2BAA2B,GAAG,CAAC,aAAa,CAAC,IAAI,CAAC,CAAC,CAAC;YAExD,QAAQ,CAAC,OAAO,GAAG,CAAC,GAAG,IAAI,CAAC,OAAO,EAAE,GAAG,2BAA2B,CAAC,CAAC;YAErE,MAAM,MAAM,GAAG,MAAM,UAAU,CAAC,QAAQ,CAAC,CAAA;YACzC,IAAI,MAAM,KAAK,SAAS,EAAE,CAAC;gBACzB,GAAG,CAAC,IAAI,CAAC,MAAM,CAAC,CAAA;YAClB,CAAC;QACH,CAAC;IACH,CAAC;SAAM,CAAC;QACN,MAAM,MAAM,GAAG,MAAM,UAAU,CAAC,IAAI,CAAC,CAAA;QACrC,IAAI,MAAM,KAAK,SAAS,EAAE,CAAC;YACzB,GAAG,CAAC,IAAI,CAAC,MAAM,CAAC,CAAA;QAClB,CAAC;IACH,CAAC;IACD,OAAO,GAAG,CAAA;AACZ,CAAC,CAAA"}
/***/ }),
diff --git a/packages/kbot/dist/package-lock.json b/packages/kbot/dist/package-lock.json
index 68a0c48d..995bc506 100644
--- a/packages/kbot/dist/package-lock.json
+++ b/packages/kbot/dist/package-lock.json
@@ -1,12 +1,12 @@
{
"name": "@plastichub/kbot",
- "version": "1.1.40",
+ "version": "1.1.41",
"lockfileVersion": 3,
"requires": true,
"packages": {
"": {
"name": "@plastichub/kbot",
- "version": "1.1.40",
+ "version": "1.1.41",
"license": "ISC",
"dependencies": {
"node-emoji": "^2.2.0"
diff --git a/packages/kbot/dist/package.json b/packages/kbot/dist/package.json
index 1f6108a3..419b757c 100644
--- a/packages/kbot/dist/package.json
+++ b/packages/kbot/dist/package.json
@@ -1,6 +1,6 @@
{
"name": "@plastichub/kbot",
- "version": "1.1.40",
+ "version": "1.1.41",
"main": "main_node.js",
"author": "",
"license": "ISC",
diff --git a/packages/kbot/logs/params.json b/packages/kbot/logs/params.json
index 6633a7f0..91b5fcf8 100644
--- a/packages/kbot/logs/params.json
+++ b/packages/kbot/logs/params.json
@@ -1,5 +1,5 @@
{
- "model": "anthropic/claude-sonnet-4",
+ "model": "deepseek/deepseek-r1:free",
"messages": [
{
"role": "user",
diff --git a/packages/kbot/schema.json b/packages/kbot/schema.json
index 2bebd835..71f811a4 100644
--- a/packages/kbot/schema.json
+++ b/packages/kbot/schema.json
@@ -118,7 +118,7 @@
},
"model": {
"type": "string",
- "description": "AI model to use for processing. Available models:\n\u001b[35m\u001b[1m\u001b[22m\u001b[39m\n\u001b[35m\u001b[1m OpenRouter models:\u001b[22m\u001b[39m\n\u001b[35m\u001b[1m\u001b[22m\u001b[39m\n01-ai/yi-large | paid\naetherwiing/mn-starcannon-12b | paid\nagentica-org/deepcoder-14b-preview:free | free\nai21/jamba-1.6-large | paid\nai21/jamba-1.6-mini | paid\naion-labs/aion-1.0 | paid\naion-labs/aion-1.0-mini | paid\naion-labs/aion-rp-llama-3.1-8b | paid\nalfredpros/codellama-7b-instruct-solidity | paid\namazon/nova-lite-v1 | paid\namazon/nova-micro-v1 | paid\namazon/nova-pro-v1 | paid\nanthropic/claude-3-haiku | paid\nanthropic/claude-3-haiku:beta | paid\nanthropic/claude-3-opus | paid\nanthropic/claude-3-opus:beta | paid\nanthropic/claude-3-sonnet | paid\nanthropic/claude-3-sonnet:beta | paid\nanthropic/claude-3.5-haiku | paid\nanthropic/claude-3.5-haiku-20241022 | paid\nanthropic/claude-3.5-haiku-20241022:beta | paid\nanthropic/claude-3.5-haiku:beta | paid\nanthropic/claude-3.5-sonnet | paid\nanthropic/claude-3.5-sonnet-20240620 | paid\nanthropic/claude-3.5-sonnet-20240620:beta | paid\nanthropic/claude-3.5-sonnet:beta | paid\nanthropic/claude-3.7-sonnet | paid\nanthropic/claude-3.7-sonnet:beta | paid\nanthropic/claude-3.7-sonnet:thinking | paid\nanthropic/claude-opus-4 | paid\nanthropic/claude-sonnet-4 | paid\nanthropic/claude-2 | paid\nanthropic/claude-2:beta | paid\nanthropic/claude-2.0 | paid\nanthropic/claude-2.0:beta | paid\nanthropic/claude-2.1 | paid\nanthropic/claude-2.1:beta | paid\narcee-ai/arcee-blitz | paid\narcee-ai/caller-large | paid\narcee-ai/coder-large | paid\narcee-ai/maestro-reasoning | paid\narcee-ai/spotlight | paid\narcee-ai/virtuoso-large | paid\narcee-ai/virtuoso-medium-v2 | paid\narliai/qwq-32b-arliai-rpr-v1:free | free\nopenrouter/auto | paid\ncohere/command | paid\ncohere/command-a | paid\ncohere/command-r | paid\ncohere/command-r-03-2024 | paid\ncohere/command-r-08-2024 | paid\ncohere/command-r-plus | paid\ncohere/command-r-plus-04-2024 | paid\ncohere/command-r-plus-08-2024 | paid\ncohere/command-r7b-12-2024 | paid\ndeepseek/deepseek-prover-v2 | paid\ndeepseek/deepseek-prover-v2:free | free\ndeepseek/deepseek-r1-0528-qwen3-8b | paid\ndeepseek/deepseek-r1-0528-qwen3-8b:free | free\ndeepseek/deepseek-r1-zero:free | free\ndeepseek/deepseek-chat | paid\ndeepseek/deepseek-chat:free | free\ndeepseek/deepseek-chat-v3-0324 | paid\ndeepseek/deepseek-chat-v3-0324:free | free\ndeepseek/deepseek-v3-base:free | free\ndeepseek/deepseek-r1 | paid\ndeepseek/deepseek-r1:free | free\ndeepseek/deepseek-r1-0528 | paid\ndeepseek/deepseek-r1-0528:free | free\ndeepseek/deepseek-r1-distill-llama-70b | paid\ndeepseek/deepseek-r1-distill-llama-70b:free | free\ndeepseek/deepseek-r1-distill-llama-8b | paid\ndeepseek/deepseek-r1-distill-qwen-1.5b | paid\ndeepseek/deepseek-r1-distill-qwen-14b | paid\ndeepseek/deepseek-r1-distill-qwen-14b:free | free\ndeepseek/deepseek-r1-distill-qwen-32b | paid\ndeepseek/deepseek-r1-distill-qwen-32b:free | free\ndeepseek/deepseek-r1-distill-qwen-7b | paid\ncognitivecomputations/dolphin-mixtral-8x22b | paid\ncognitivecomputations/dolphin3.0-mistral-24b:free | free\ncognitivecomputations/dolphin3.0-r1-mistral-24b:free | free\neleutherai/llemma_7b | paid\neva-unit-01/eva-llama-3.33-70b | paid\neva-unit-01/eva-qwen-2.5-32b | paid\neva-unit-01/eva-qwen-2.5-72b | paid\nsao10k/fimbulvetr-11b-v2 | paid\nalpindale/goliath-120b | paid\ngoogle/gemini-flash-1.5 | paid\ngoogle/gemini-flash-1.5-8b | paid\ngoogle/gemini-pro-1.5 | paid\ngoogle/gemini-2.0-flash-001 | paid\ngoogle/gemini-2.0-flash-exp:free | free\ngoogle/gemini-2.0-flash-lite-001 | paid\ngoogle/gemini-2.5-flash-preview | paid\ngoogle/gemini-2.5-flash-preview:thinking | paid\ngoogle/gemini-2.5-flash-preview-05-20 | paid\ngoogle/gemini-2.5-flash-preview-05-20:thinking | paid\ngoogle/gemini-2.5-pro-exp-03-25 | paid\ngoogle/gemini-2.5-pro-preview | paid\ngoogle/gemma-2-27b-it | paid\ngoogle/gemma-2b-it | paid\ngoogle/gemma-2-9b-it | paid\ngoogle/gemma-2-9b-it:free | free\ngoogle/gemma-3-12b-it | paid\ngoogle/gemma-3-12b-it:free | free\ngoogle/gemma-3-1b-it:free | free\ngoogle/gemma-3-27b-it | paid\ngoogle/gemma-3-27b-it:free | free\ngoogle/gemma-3-4b-it | paid\ngoogle/gemma-3-4b-it:free | free\ngoogle/gemma-3n-e4b-it:free | free\ninception/mercury-coder-small-beta | paid\ninfermatic/mn-inferor-12b | paid\ninflection/inflection-3-pi | paid\ninflection/inflection-3-productivity | paid\nliquid/lfm-3b | paid\nliquid/lfm-40b | paid\nliquid/lfm-7b | paid\nmeta-llama/llama-guard-3-8b | paid\nalpindale/magnum-72b | paid\nanthracite-org/magnum-v2-72b | paid\nanthracite-org/magnum-v4-72b | paid\nmancer/weaver | paid\nmeta-llama/llama-2-70b-chat | paid\nmeta-llama/llama-3-70b-instruct | paid\nmeta-llama/llama-3-8b-instruct | paid\nmeta-llama/llama-3.1-405b | paid\nmeta-llama/llama-3.1-405b:free | free\nmeta-llama/llama-3.1-405b-instruct | paid\nmeta-llama/llama-3.1-70b-instruct | paid\nmeta-llama/llama-3.1-8b-instruct | paid\nmeta-llama/llama-3.1-8b-instruct:free | free\nmeta-llama/llama-3.2-11b-vision-instruct | paid\nmeta-llama/llama-3.2-11b-vision-instruct:free | free\nmeta-llama/llama-3.2-1b-instruct | paid\nmeta-llama/llama-3.2-1b-instruct:free | free\nmeta-llama/llama-3.2-3b-instruct | paid\nmeta-llama/llama-3.2-3b-instruct:free | free\nmeta-llama/llama-3.2-90b-vision-instruct | paid\nmeta-llama/llama-3.3-70b-instruct | paid\nmeta-llama/llama-3.3-70b-instruct:free | free\nmeta-llama/llama-3.3-8b-instruct:free | free\nmeta-llama/llama-4-maverick | paid\nmeta-llama/llama-4-maverick:free | free\nmeta-llama/llama-4-scout | paid\nmeta-llama/llama-4-scout:free | free\nmeta-llama/llama-guard-4-12b | paid\nmeta-llama/llama-guard-2-8b | paid\nmicrosoft/mai-ds-r1:free | free\nmicrosoft/phi-4 | paid\nmicrosoft/phi-4-multimodal-instruct | paid\nmicrosoft/phi-4-reasoning:free | free\nmicrosoft/phi-4-reasoning-plus | paid\nmicrosoft/phi-4-reasoning-plus:free | free\nmicrosoft/phi-3-medium-128k-instruct | paid\nmicrosoft/phi-3-mini-128k-instruct | paid\nmicrosoft/phi-3.5-mini-128k-instruct | paid\nsophosympatheia/midnight-rose-70b | paid\nminimax/minimax-01 | paid\nmistralai/mistral-large | paid\nmistralai/mistral-large-2407 | paid\nmistralai/mistral-large-2411 | paid\nmistralai/mistral-medium | paid\nnothingiisreal/mn-celeste-12b | paid\nmistralai/mistral-small | paid\nmistralai/mistral-tiny | paid\nmistralai/codestral-2501 | paid\nmistralai/devstral-small | paid\nmistralai/devstral-small:free | free\nmistralai/ministral-3b | paid\nmistralai/ministral-8b | paid\nmistralai/mistral-7b-instruct | paid\nmistralai/mistral-7b-instruct:free | free\nmistralai/mistral-7b-instruct-v0.1 | paid\nmistralai/mistral-7b-instruct-v0.2 | paid\nmistralai/mistral-7b-instruct-v0.3 | paid\nmistralai/mistral-medium-3 | paid\nmistralai/mistral-nemo | paid\nmistralai/mistral-nemo:free | free\nmistralai/mistral-small-24b-instruct-2501 | paid\nmistralai/mistral-small-24b-instruct-2501:free | free\nmistralai/mistral-small-3.1-24b-instruct | paid\nmistralai/mistral-small-3.1-24b-instruct:free | free\nmistralai/mixtral-8x22b-instruct | paid\nmistralai/mixtral-8x7b-instruct | paid\nmistralai/pixtral-12b | paid\nmistralai/pixtral-large-2411 | paid\nmistralai/mistral-saba | paid\nmoonshotai/kimi-vl-a3b-thinking:free | free\nmoonshotai/moonlight-16b-a3b-instruct:free | free\ngryphe/mythomax-l2-13b | paid\nneversleep/llama-3-lumimaid-70b | paid\nneversleep/llama-3-lumimaid-8b | paid\nneversleep/llama-3.1-lumimaid-70b | paid\nneversleep/llama-3.1-lumimaid-8b | paid\nneversleep/noromaid-20b | paid\nnousresearch/deephermes-3-llama-3-8b-preview:free | free\nnousresearch/deephermes-3-mistral-24b-preview:free | free\nnousresearch/nous-hermes-2-mixtral-8x7b-dpo | paid\nnousresearch/hermes-3-llama-3.1-405b | paid\nnousresearch/hermes-3-llama-3.1-70b | paid\nnousresearch/hermes-2-pro-llama-3-8b | paid\nnvidia/llama-3.1-nemotron-70b-instruct | paid\nnvidia/llama-3.1-nemotron-ultra-253b-v1 | paid\nnvidia/llama-3.1-nemotron-ultra-253b-v1:free | free\nnvidia/llama-3.3-nemotron-super-49b-v1 | paid\nnvidia/llama-3.3-nemotron-super-49b-v1:free | free\nopen-r1/olympiccoder-32b:free | free\nopenai/chatgpt-4o-latest | paid\nopenai/codex-mini | paid\nopenai/gpt-3.5-turbo | paid\nopenai/gpt-3.5-turbo-0613 | paid\nopenai/gpt-3.5-turbo-16k | paid\nopenai/gpt-3.5-turbo-0125 | paid\nopenai/gpt-3.5-turbo-1106 | paid\nopenai/gpt-3.5-turbo-instruct | paid\nopenai/gpt-4 | paid\nopenai/gpt-4-0314 | paid\nopenai/gpt-4-32k | paid\nopenai/gpt-4-32k-0314 | paid\nopenai/gpt-4-turbo | paid\nopenai/gpt-4-1106-preview | paid\nopenai/gpt-4-turbo-preview | paid\nopenai/gpt-4.1 | paid\nopenai/gpt-4.1-mini | paid\nopenai/gpt-4.1-nano | paid\nopenai/gpt-4.5-preview | paid\nopenai/gpt-4o | paid\nopenai/gpt-4o-2024-05-13 | paid\nopenai/gpt-4o-2024-08-06 | paid\nopenai/gpt-4o-2024-11-20 | paid\nopenai/gpt-4o:extended | paid\nopenai/gpt-4o-search-preview | paid\nopenai/gpt-4o-mini | paid\nopenai/gpt-4o-mini-2024-07-18 | paid\nopenai/gpt-4o-mini-search-preview | paid\nopenai/o1 | paid\nopenai/o1-mini | paid\nopenai/o1-mini-2024-09-12 | paid\nopenai/o1-preview | paid\nopenai/o1-preview-2024-09-12 | paid\nopenai/o1-pro | paid\nopenai/o3 | paid\nopenai/o3-mini | paid\nopenai/o3-mini-high | paid\nopenai/o4-mini | paid\nopenai/o4-mini-high | paid\nopengvlab/internvl3-14b:free | free\nopengvlab/internvl3-2b:free | free\nall-hands/openhands-lm-32b-v0.1 | paid\nperplexity/llama-3.1-sonar-large-128k-online | paid\nperplexity/llama-3.1-sonar-small-128k-online | paid\nperplexity/r1-1776 | paid\nperplexity/sonar | paid\nperplexity/sonar-deep-research | paid\nperplexity/sonar-pro | paid\nperplexity/sonar-reasoning | paid\nperplexity/sonar-reasoning-pro | paid\npygmalionai/mythalion-13b | paid\nqwen/qwen-2-72b-instruct | paid\nqwen/qwen-vl-max | paid\nqwen/qwen-vl-plus | paid\nqwen/qwen-max | paid\nqwen/qwen-plus | paid\nqwen/qwen-turbo | paid\nqwen/qwen2.5-vl-32b-instruct | paid\nqwen/qwen2.5-vl-32b-instruct:free | free\nqwen/qwen2.5-vl-3b-instruct:free | free\nqwen/qwen2.5-vl-72b-instruct | paid\nqwen/qwen2.5-vl-72b-instruct:free | free\nqwen/qwen-2.5-vl-7b-instruct | paid\nqwen/qwen-2.5-vl-7b-instruct:free | free\nqwen/qwen3-14b | paid\nqwen/qwen3-14b:free | free\nqwen/qwen3-235b-a22b | paid\nqwen/qwen3-235b-a22b:free | free\nqwen/qwen3-30b-a3b | paid\nqwen/qwen3-30b-a3b:free | free\nqwen/qwen3-32b | paid\nqwen/qwen3-32b:free | free\nqwen/qwen3-8b | paid\nqwen/qwen3-8b:free | free\nqwen/qwq-32b | paid\nqwen/qwq-32b:free | free\nqwen/qwq-32b-preview | paid\nqwen/qwen-2.5-72b-instruct | paid\nqwen/qwen-2.5-72b-instruct:free | free\nqwen/qwen-2.5-7b-instruct | paid\nqwen/qwen-2.5-7b-instruct:free | free\nqwen/qwen-2.5-coder-32b-instruct | paid\nqwen/qwen-2.5-coder-32b-instruct:free | free\nfeatherless/qwerky-72b:free | free\nrekaai/reka-flash-3:free | free\nundi95/remm-slerp-l2-13b | paid\nsao10k/l3-lunaris-8b | paid\nsao10k/l3-euryale-70b | paid\nsao10k/l3.1-euryale-70b | paid\nsao10k/l3.3-euryale-70b | paid\nsarvamai/sarvam-m | paid\nsarvamai/sarvam-m:free | free\nshisa-ai/shisa-v2-llama3.3-70b:free | free\nraifle/sorcererlm-8x22b | paid\nthedrummer/anubis-pro-105b-v1 | paid\nthedrummer/rocinante-12b | paid\nthedrummer/skyfall-36b-v2 | paid\nthedrummer/unslopnemo-12b | paid\nthedrummer/valkyrie-49b-v1 | paid\nthudm/glm-4-32b | paid\nthudm/glm-4-32b:free | free\nthudm/glm-z1-32b | paid\nthudm/glm-z1-32b:free | free\nthudm/glm-z1-rumination-32b | paid\ntngtech/deepseek-r1t-chimera:free | free\nundi95/toppy-m-7b | paid\nscb10x/llama3.1-typhoon2-70b-instruct | paid\nscb10x/llama3.1-typhoon2-8b-instruct | paid\nmicrosoft/wizardlm-2-8x22b | paid\nx-ai/grok-2-1212 | paid\nx-ai/grok-2-vision-1212 | paid\nx-ai/grok-3-beta | paid\nx-ai/grok-3-mini-beta | paid\nx-ai/grok-beta | paid\nx-ai/grok-vision-beta | paid\n\u001b[35m\u001b[1m\u001b[22m\u001b[39m\n\u001b[35m\u001b[1m OpenAI models:\u001b[22m\u001b[39m\n\u001b[35m\u001b[1m\u001b[22m\u001b[39m\nbabbage-002\nchatgpt-4o-latest\ncodex-mini-latest\ndall-e-2\ndall-e-3\ndavinci-002\ngpt-3.5-turbo\ngpt-3.5-turbo-0125\ngpt-3.5-turbo-1106\ngpt-3.5-turbo-16k\ngpt-3.5-turbo-instruct\ngpt-3.5-turbo-instruct-0914\ngpt-4\ngpt-4-0125-preview\ngpt-4-0613\ngpt-4-1106-preview\ngpt-4-turbo\ngpt-4-turbo-2024-04-09\ngpt-4-turbo-preview\ngpt-4.1\ngpt-4.1-2025-04-14\ngpt-4.1-mini\ngpt-4.1-mini-2025-04-14\ngpt-4.1-nano\ngpt-4.1-nano-2025-04-14\ngpt-4.5-preview\ngpt-4.5-preview-2025-02-27\ngpt-4o\ngpt-4o-2024-05-13\ngpt-4o-2024-08-06\ngpt-4o-2024-11-20\ngpt-4o-audio-preview\ngpt-4o-audio-preview-2024-10-01\ngpt-4o-audio-preview-2024-12-17\ngpt-4o-audio-preview-2025-06-03\ngpt-4o-mini\ngpt-4o-mini-2024-07-18\ngpt-4o-mini-audio-preview\ngpt-4o-mini-audio-preview-2024-12-17\ngpt-4o-mini-realtime-preview\ngpt-4o-mini-realtime-preview-2024-12-17\ngpt-4o-mini-search-preview\ngpt-4o-mini-search-preview-2025-03-11\ngpt-4o-mini-transcribe\ngpt-4o-mini-tts\ngpt-4o-realtime-preview\ngpt-4o-realtime-preview-2024-10-01\ngpt-4o-realtime-preview-2024-12-17\ngpt-4o-realtime-preview-2025-06-03\ngpt-4o-search-preview\ngpt-4o-search-preview-2025-03-11\ngpt-4o-transcribe\ngpt-image-1\no1\no1-2024-12-17\no1-mini\no1-mini-2024-09-12\no1-preview\no1-preview-2024-09-12\no1-pro\no1-pro-2025-03-19\no3-mini\no3-mini-2025-01-31\no4-mini\no4-mini-2025-04-16\nomni-moderation-2024-09-26\nomni-moderation-latest\ntext-embedding-3-large\ntext-embedding-3-small\ntext-embedding-ada-002\ntts-1\ntts-1-1106\ntts-1-hd\ntts-1-hd-1106\nwhisper-1\n-----\n\n\u001b[35m\u001b[1m\u001b[22m\u001b[39m\n\u001b[35m\u001b[1m Deepseek models:\u001b[22m\u001b[39m\n\u001b[35m\u001b[1m\u001b[22m\u001b[39m\ndeepseek-chat\ndeepseek-reasoner\n-----\n"
+ "description": "AI model to use for processing. Available models:\n\u001b[35m\u001b[1m\u001b[22m\u001b[39m\n\u001b[35m\u001b[1m OpenRouter models:\u001b[22m\u001b[39m\n\u001b[35m\u001b[1m\u001b[22m\u001b[39m\n01-ai/yi-large | paid\naetherwiing/mn-starcannon-12b | paid\nagentica-org/deepcoder-14b-preview:free | free\nai21/jamba-1.6-large | paid\nai21/jamba-1.6-mini | paid\naion-labs/aion-1.0 | paid\naion-labs/aion-1.0-mini | paid\naion-labs/aion-rp-llama-3.1-8b | paid\nalfredpros/codellama-7b-instruct-solidity | paid\namazon/nova-lite-v1 | paid\namazon/nova-micro-v1 | paid\namazon/nova-pro-v1 | paid\nanthropic/claude-3-haiku | paid\nanthropic/claude-3-haiku:beta | paid\nanthropic/claude-3-opus | paid\nanthropic/claude-3-opus:beta | paid\nanthropic/claude-3-sonnet | paid\nanthropic/claude-3-sonnet:beta | paid\nanthropic/claude-3.5-haiku | paid\nanthropic/claude-3.5-haiku-20241022 | paid\nanthropic/claude-3.5-haiku-20241022:beta | paid\nanthropic/claude-3.5-haiku:beta | paid\nanthropic/claude-3.5-sonnet | paid\nanthropic/claude-3.5-sonnet-20240620 | paid\nanthropic/claude-3.5-sonnet-20240620:beta | paid\nanthropic/claude-3.5-sonnet:beta | paid\nanthropic/claude-3.7-sonnet | paid\nanthropic/claude-3.7-sonnet:beta | paid\nanthropic/claude-3.7-sonnet:thinking | paid\nanthropic/claude-opus-4 | paid\nanthropic/claude-sonnet-4 | paid\nanthropic/claude-2 | paid\nanthropic/claude-2:beta | paid\nanthropic/claude-2.0 | paid\nanthropic/claude-2.0:beta | paid\nanthropic/claude-2.1 | paid\nanthropic/claude-2.1:beta | paid\narcee-ai/arcee-blitz | paid\narcee-ai/caller-large | paid\narcee-ai/coder-large | paid\narcee-ai/maestro-reasoning | paid\narcee-ai/spotlight | paid\narcee-ai/virtuoso-large | paid\narcee-ai/virtuoso-medium-v2 | paid\narliai/qwq-32b-arliai-rpr-v1:free | free\nopenrouter/auto | paid\ncohere/command | paid\ncohere/command-a | paid\ncohere/command-r | paid\ncohere/command-r-03-2024 | paid\ncohere/command-r-08-2024 | paid\ncohere/command-r-plus | paid\ncohere/command-r-plus-04-2024 | paid\ncohere/command-r-plus-08-2024 | paid\ncohere/command-r7b-12-2024 | paid\ndeepseek/deepseek-prover-v2 | paid\ndeepseek/deepseek-r1-0528-qwen3-8b | paid\ndeepseek/deepseek-r1-0528-qwen3-8b:free | free\ndeepseek/deepseek-chat | paid\ndeepseek/deepseek-chat:free | free\ndeepseek/deepseek-chat-v3-0324 | paid\ndeepseek/deepseek-chat-v3-0324:free | free\ndeepseek/deepseek-v3-base:free | free\ndeepseek/deepseek-r1 | paid\ndeepseek/deepseek-r1:free | free\ndeepseek/deepseek-r1-0528 | paid\ndeepseek/deepseek-r1-0528:free | free\ndeepseek/deepseek-r1-distill-llama-70b | paid\ndeepseek/deepseek-r1-distill-llama-70b:free | free\ndeepseek/deepseek-r1-distill-llama-8b | paid\ndeepseek/deepseek-r1-distill-qwen-1.5b | paid\ndeepseek/deepseek-r1-distill-qwen-14b | paid\ndeepseek/deepseek-r1-distill-qwen-14b:free | free\ndeepseek/deepseek-r1-distill-qwen-32b | paid\ndeepseek/deepseek-r1-distill-qwen-7b | paid\ncognitivecomputations/dolphin-mixtral-8x22b | paid\ncognitivecomputations/dolphin3.0-mistral-24b:free | free\ncognitivecomputations/dolphin3.0-r1-mistral-24b:free | free\neleutherai/llemma_7b | paid\neva-unit-01/eva-llama-3.33-70b | paid\neva-unit-01/eva-qwen-2.5-32b | paid\neva-unit-01/eva-qwen-2.5-72b | paid\nsao10k/fimbulvetr-11b-v2 | paid\nalpindale/goliath-120b | paid\ngoogle/gemini-flash-1.5 | paid\ngoogle/gemini-flash-1.5-8b | paid\ngoogle/gemini-pro-1.5 | paid\ngoogle/gemini-2.0-flash-001 | paid\ngoogle/gemini-2.0-flash-exp:free | free\ngoogle/gemini-2.0-flash-lite-001 | paid\ngoogle/gemini-2.5-flash | paid\ngoogle/gemini-2.5-flash-lite-preview-06-17 | paid\ngoogle/gemini-2.5-flash-preview | paid\ngoogle/gemini-2.5-flash-preview:thinking | paid\ngoogle/gemini-2.5-flash-preview-05-20 | paid\ngoogle/gemini-2.5-flash-preview-05-20:thinking | paid\ngoogle/gemini-2.5-pro | paid\ngoogle/gemini-2.5-pro-exp-03-25 | paid\ngoogle/gemini-2.5-pro-preview-05-06 | paid\ngoogle/gemini-2.5-pro-preview | paid\ngoogle/gemma-2-27b-it | paid\ngoogle/gemma-2-9b-it | paid\ngoogle/gemma-2-9b-it:free | free\ngoogle/gemma-3-12b-it | paid\ngoogle/gemma-3-12b-it:free | free\ngoogle/gemma-3-27b-it | paid\ngoogle/gemma-3-27b-it:free | free\ngoogle/gemma-3-4b-it | paid\ngoogle/gemma-3-4b-it:free | free\ngoogle/gemma-3n-e4b-it | paid\ngoogle/gemma-3n-e4b-it:free | free\ninception/mercury | paid\ninception/mercury-coder-small-beta | paid\ninfermatic/mn-inferor-12b | paid\ninflection/inflection-3-pi | paid\ninflection/inflection-3-productivity | paid\nmoonshotai/kimi-dev-72b:free | free\nliquid/lfm-3b | paid\nliquid/lfm-40b | paid\nliquid/lfm-7b | paid\nmeta-llama/llama-guard-3-8b | paid\nalpindale/magnum-72b | paid\nanthracite-org/magnum-v2-72b | paid\nanthracite-org/magnum-v4-72b | paid\nmancer/weaver | paid\nmeta-llama/llama-3-70b-instruct | paid\nmeta-llama/llama-3-8b-instruct | paid\nmeta-llama/llama-3.1-405b | paid\nmeta-llama/llama-3.1-405b-instruct | paid\nmeta-llama/llama-3.1-70b-instruct | paid\nmeta-llama/llama-3.1-8b-instruct | paid\nmeta-llama/llama-3.1-8b-instruct:free | free\nmeta-llama/llama-3.2-11b-vision-instruct | paid\nmeta-llama/llama-3.2-11b-vision-instruct:free | free\nmeta-llama/llama-3.2-1b-instruct | paid\nmeta-llama/llama-3.2-1b-instruct:free | free\nmeta-llama/llama-3.2-3b-instruct | paid\nmeta-llama/llama-3.2-90b-vision-instruct | paid\nmeta-llama/llama-3.3-70b-instruct | paid\nmeta-llama/llama-3.3-70b-instruct:free | free\nmeta-llama/llama-4-maverick | paid\nmeta-llama/llama-4-maverick:free | free\nmeta-llama/llama-4-scout | paid\nmeta-llama/llama-4-scout:free | free\nmeta-llama/llama-guard-4-12b | paid\nmeta-llama/llama-guard-2-8b | paid\nmicrosoft/mai-ds-r1:free | free\nmicrosoft/phi-4 | paid\nmicrosoft/phi-4-multimodal-instruct | paid\nmicrosoft/phi-4-reasoning-plus | paid\nmicrosoft/phi-3-medium-128k-instruct | paid\nmicrosoft/phi-3-mini-128k-instruct | paid\nmicrosoft/phi-3.5-mini-128k-instruct | paid\nsophosympatheia/midnight-rose-70b | paid\nminimax/minimax-m1 | paid\nminimax/minimax-m1:extended | paid\nminimax/minimax-01 | paid\nmistralai/mistral-large | paid\nmistralai/mistral-large-2407 | paid\nmistralai/mistral-large-2411 | paid\nnothingiisreal/mn-celeste-12b | paid\nmistralai/mistral-small | paid\nmistralai/mistral-tiny | paid\nmistralai/codestral-2501 | paid\nmistralai/devstral-small | paid\nmistralai/devstral-small:free | free\nmistralai/magistral-medium-2506 | paid\nmistralai/magistral-medium-2506:thinking | paid\nmistralai/magistral-small-2506 | paid\nmistralai/ministral-3b | paid\nmistralai/ministral-8b | paid\nmistralai/mistral-7b-instruct | paid\nmistralai/mistral-7b-instruct:free | free\nmistralai/mistral-7b-instruct-v0.1 | paid\nmistralai/mistral-7b-instruct-v0.2 | paid\nmistralai/mistral-7b-instruct-v0.3 | paid\nmistralai/mistral-medium-3 | paid\nmistralai/mistral-nemo | paid\nmistralai/mistral-nemo:free | free\nmistralai/mistral-small-24b-instruct-2501 | paid\nmistralai/mistral-small-24b-instruct-2501:free | free\nmistralai/mistral-small-3.1-24b-instruct | paid\nmistralai/mistral-small-3.1-24b-instruct:free | free\nmistralai/mistral-small-3.2-24b-instruct | paid\nmistralai/mistral-small-3.2-24b-instruct:free | free\nmistralai/mixtral-8x22b-instruct | paid\nmistralai/mixtral-8x7b-instruct | paid\nmistralai/pixtral-12b | paid\nmistralai/pixtral-large-2411 | paid\nmistralai/mistral-saba | paid\nmoonshotai/kimi-vl-a3b-thinking:free | free\nmorph/morph-v2 | paid\ngryphe/mythomax-l2-13b | paid\nneversleep/llama-3-lumimaid-70b | paid\nneversleep/llama-3-lumimaid-8b | paid\nneversleep/llama-3.1-lumimaid-70b | paid\nneversleep/llama-3.1-lumimaid-8b | paid\nneversleep/noromaid-20b | paid\nnousresearch/deephermes-3-llama-3-8b-preview:free | free\nnousresearch/nous-hermes-2-mixtral-8x7b-dpo | paid\nnousresearch/hermes-3-llama-3.1-405b | paid\nnousresearch/hermes-3-llama-3.1-70b | paid\nnousresearch/hermes-2-pro-llama-3-8b | paid\nnvidia/llama-3.1-nemotron-70b-instruct | paid\nnvidia/llama-3.1-nemotron-ultra-253b-v1 | paid\nnvidia/llama-3.1-nemotron-ultra-253b-v1:free | free\nnvidia/llama-3.3-nemotron-super-49b-v1 | paid\nnvidia/llama-3.3-nemotron-super-49b-v1:free | free\nopenai/chatgpt-4o-latest | paid\nopenai/codex-mini | paid\nopenai/gpt-3.5-turbo-0613 | paid\nopenai/gpt-3.5-turbo-16k | paid\nopenai/gpt-3.5-turbo-instruct | paid\nopenai/gpt-4 | paid\nopenai/gpt-4-0314 | paid\nopenai/gpt-4-turbo | paid\nopenai/gpt-4-1106-preview | paid\nopenai/gpt-4-turbo-preview | paid\nopenai/gpt-4.1 | paid\nopenai/gpt-4.1-mini | paid\nopenai/gpt-4.1-nano | paid\nopenai/gpt-4.5-preview | paid\nopenai/gpt-4o | paid\nopenai/gpt-4o-2024-05-13 | paid\nopenai/gpt-4o-2024-08-06 | paid\nopenai/gpt-4o-2024-11-20 | paid\nopenai/gpt-4o:extended | paid\nopenai/gpt-4o-search-preview | paid\nopenai/gpt-4o-mini | paid\nopenai/gpt-4o-mini-2024-07-18 | paid\nopenai/gpt-4o-mini-search-preview | paid\nopenai/o1 | paid\nopenai/o1-mini | paid\nopenai/o1-mini-2024-09-12 | paid\nopenai/o1-preview | paid\nopenai/o1-preview-2024-09-12 | paid\nopenai/o1-pro | paid\nopenai/o3 | paid\nopenai/o3-mini | paid\nopenai/o3-mini-high | paid\nopenai/o3-pro | paid\nopenai/o4-mini | paid\nopenai/o4-mini-high | paid\nopengvlab/internvl3-14b | paid\nopengvlab/internvl3-2b | paid\nall-hands/openhands-lm-32b-v0.1 | paid\nperplexity/llama-3.1-sonar-large-128k-online | paid\nperplexity/llama-3.1-sonar-small-128k-online | paid\nperplexity/r1-1776 | paid\nperplexity/sonar | paid\nperplexity/sonar-deep-research | paid\nperplexity/sonar-pro | paid\nperplexity/sonar-reasoning | paid\nperplexity/sonar-reasoning-pro | paid\npygmalionai/mythalion-13b | paid\nqwen/qwen-2-72b-instruct | paid\nqwen/qwen-vl-max | paid\nqwen/qwen-vl-plus | paid\nqwen/qwen-max | paid\nqwen/qwen-plus | paid\nqwen/qwen-turbo | paid\nqwen/qwen2.5-vl-32b-instruct | paid\nqwen/qwen2.5-vl-32b-instruct:free | free\nqwen/qwen2.5-vl-72b-instruct | paid\nqwen/qwen2.5-vl-72b-instruct:free | free\nqwen/qwen-2.5-vl-7b-instruct | paid\nqwen/qwen3-14b | paid\nqwen/qwen3-14b:free | free\nqwen/qwen3-235b-a22b | paid\nqwen/qwen3-235b-a22b:free | free\nqwen/qwen3-30b-a3b | paid\nqwen/qwen3-30b-a3b:free | free\nqwen/qwen3-32b | paid\nqwen/qwen3-32b:free | free\nqwen/qwen3-8b | paid\nqwen/qwen3-8b:free | free\nqwen/qwq-32b | paid\nqwen/qwq-32b:free | free\nqwen/qwq-32b-preview | paid\nqwen/qwen-2.5-72b-instruct | paid\nqwen/qwen-2.5-72b-instruct:free | free\nqwen/qwen-2.5-7b-instruct | paid\nqwen/qwen-2.5-coder-32b-instruct | paid\nqwen/qwen-2.5-coder-32b-instruct:free | free\nfeatherless/qwerky-72b:free | free\nrekaai/reka-flash-3:free | free\nundi95/remm-slerp-l2-13b | paid\nsao10k/l3-lunaris-8b | paid\nsao10k/l3-euryale-70b | paid\nsao10k/l3.1-euryale-70b | paid\nsao10k/l3.3-euryale-70b | paid\nsarvamai/sarvam-m:free | free\nsentientagi/dobby-mini-unhinged-plus-llama-3.1-8b | paid\nshisa-ai/shisa-v2-llama3.3-70b:free | free\nraifle/sorcererlm-8x22b | paid\nthedrummer/anubis-pro-105b-v1 | paid\nthedrummer/rocinante-12b | paid\nthedrummer/skyfall-36b-v2 | paid\nthedrummer/unslopnemo-12b | paid\nthedrummer/valkyrie-49b-v1 | paid\nthudm/glm-4-32b | paid\nthudm/glm-4-32b:free | free\nthudm/glm-z1-32b | paid\nthudm/glm-z1-32b:free | free\nthudm/glm-z1-rumination-32b | paid\ntngtech/deepseek-r1t-chimera:free | free\nundi95/toppy-m-7b | paid\nscb10x/llama3.1-typhoon2-70b-instruct | paid\nmicrosoft/wizardlm-2-8x22b | paid\nx-ai/grok-2-1212 | paid\nx-ai/grok-2-vision-1212 | paid\nx-ai/grok-3 | paid\nx-ai/grok-3-beta | paid\nx-ai/grok-3-mini | paid\nx-ai/grok-3-mini-beta | paid\nx-ai/grok-beta | paid\nx-ai/grok-vision-beta | paid\n\u001b[35m\u001b[1m\u001b[22m\u001b[39m\n\u001b[35m\u001b[1m OpenAI models:\u001b[22m\u001b[39m\n\u001b[35m\u001b[1m\u001b[22m\u001b[39m\nbabbage-002\nchatgpt-4o-latest\ncodex-mini-latest\ndall-e-2\ndall-e-3\ndavinci-002\ngpt-3.5-turbo\ngpt-3.5-turbo-0125\ngpt-3.5-turbo-1106\ngpt-3.5-turbo-16k\ngpt-3.5-turbo-instruct\ngpt-3.5-turbo-instruct-0914\ngpt-4\ngpt-4-0125-preview\ngpt-4-0613\ngpt-4-1106-preview\ngpt-4-turbo\ngpt-4-turbo-2024-04-09\ngpt-4-turbo-preview\ngpt-4.1\ngpt-4.1-2025-04-14\ngpt-4.1-mini\ngpt-4.1-mini-2025-04-14\ngpt-4.1-nano\ngpt-4.1-nano-2025-04-14\ngpt-4.5-preview\ngpt-4.5-preview-2025-02-27\ngpt-4o\ngpt-4o-2024-05-13\ngpt-4o-2024-08-06\ngpt-4o-2024-11-20\ngpt-4o-audio-preview\ngpt-4o-audio-preview-2024-10-01\ngpt-4o-audio-preview-2024-12-17\ngpt-4o-audio-preview-2025-06-03\ngpt-4o-mini\ngpt-4o-mini-2024-07-18\ngpt-4o-mini-audio-preview\ngpt-4o-mini-audio-preview-2024-12-17\ngpt-4o-mini-realtime-preview\ngpt-4o-mini-realtime-preview-2024-12-17\ngpt-4o-mini-search-preview\ngpt-4o-mini-search-preview-2025-03-11\ngpt-4o-mini-transcribe\ngpt-4o-mini-tts\ngpt-4o-realtime-preview\ngpt-4o-realtime-preview-2024-10-01\ngpt-4o-realtime-preview-2024-12-17\ngpt-4o-realtime-preview-2025-06-03\ngpt-4o-search-preview\ngpt-4o-search-preview-2025-03-11\ngpt-4o-transcribe\ngpt-image-1\no1\no1-2024-12-17\no1-mini\no1-mini-2024-09-12\no1-preview\no1-preview-2024-09-12\no1-pro\no1-pro-2025-03-19\no3-mini\no3-mini-2025-01-31\no4-mini\no4-mini-2025-04-16\no4-mini-deep-research\no4-mini-deep-research-2025-06-26\nomni-moderation-2024-09-26\nomni-moderation-latest\ntext-embedding-3-large\ntext-embedding-3-small\ntext-embedding-ada-002\ntts-1\ntts-1-1106\ntts-1-hd\ntts-1-hd-1106\nwhisper-1\n-----\n\n\u001b[35m\u001b[1m\u001b[22m\u001b[39m\n\u001b[35m\u001b[1m Deepseek models:\u001b[22m\u001b[39m\n\u001b[35m\u001b[1m\u001b[22m\u001b[39m\ndeepseek-chat\ndeepseek-reasoner\n-----\n"
},
"router": {
"type": "string",
diff --git a/packages/kbot/schema_ui.json b/packages/kbot/schema_ui.json
index cf2b1fed..022eed13 100644
--- a/packages/kbot/schema_ui.json
+++ b/packages/kbot/schema_ui.json
@@ -79,7 +79,7 @@
"ui:title": "Api_key"
},
"model": {
- "ui:description": "AI model to use for processing. Available models:\n\u001b[35m\u001b[1m\u001b[22m\u001b[39m\n\u001b[35m\u001b[1m OpenRouter models:\u001b[22m\u001b[39m\n\u001b[35m\u001b[1m\u001b[22m\u001b[39m\n01-ai/yi-large | paid\naetherwiing/mn-starcannon-12b | paid\nagentica-org/deepcoder-14b-preview:free | free\nai21/jamba-1.6-large | paid\nai21/jamba-1.6-mini | paid\naion-labs/aion-1.0 | paid\naion-labs/aion-1.0-mini | paid\naion-labs/aion-rp-llama-3.1-8b | paid\nalfredpros/codellama-7b-instruct-solidity | paid\namazon/nova-lite-v1 | paid\namazon/nova-micro-v1 | paid\namazon/nova-pro-v1 | paid\nanthropic/claude-3-haiku | paid\nanthropic/claude-3-haiku:beta | paid\nanthropic/claude-3-opus | paid\nanthropic/claude-3-opus:beta | paid\nanthropic/claude-3-sonnet | paid\nanthropic/claude-3-sonnet:beta | paid\nanthropic/claude-3.5-haiku | paid\nanthropic/claude-3.5-haiku-20241022 | paid\nanthropic/claude-3.5-haiku-20241022:beta | paid\nanthropic/claude-3.5-haiku:beta | paid\nanthropic/claude-3.5-sonnet | paid\nanthropic/claude-3.5-sonnet-20240620 | paid\nanthropic/claude-3.5-sonnet-20240620:beta | paid\nanthropic/claude-3.5-sonnet:beta | paid\nanthropic/claude-3.7-sonnet | paid\nanthropic/claude-3.7-sonnet:beta | paid\nanthropic/claude-3.7-sonnet:thinking | paid\nanthropic/claude-opus-4 | paid\nanthropic/claude-sonnet-4 | paid\nanthropic/claude-2 | paid\nanthropic/claude-2:beta | paid\nanthropic/claude-2.0 | paid\nanthropic/claude-2.0:beta | paid\nanthropic/claude-2.1 | paid\nanthropic/claude-2.1:beta | paid\narcee-ai/arcee-blitz | paid\narcee-ai/caller-large | paid\narcee-ai/coder-large | paid\narcee-ai/maestro-reasoning | paid\narcee-ai/spotlight | paid\narcee-ai/virtuoso-large | paid\narcee-ai/virtuoso-medium-v2 | paid\narliai/qwq-32b-arliai-rpr-v1:free | free\nopenrouter/auto | paid\ncohere/command | paid\ncohere/command-a | paid\ncohere/command-r | paid\ncohere/command-r-03-2024 | paid\ncohere/command-r-08-2024 | paid\ncohere/command-r-plus | paid\ncohere/command-r-plus-04-2024 | paid\ncohere/command-r-plus-08-2024 | paid\ncohere/command-r7b-12-2024 | paid\ndeepseek/deepseek-prover-v2 | paid\ndeepseek/deepseek-prover-v2:free | free\ndeepseek/deepseek-r1-0528-qwen3-8b | paid\ndeepseek/deepseek-r1-0528-qwen3-8b:free | free\ndeepseek/deepseek-r1-zero:free | free\ndeepseek/deepseek-chat | paid\ndeepseek/deepseek-chat:free | free\ndeepseek/deepseek-chat-v3-0324 | paid\ndeepseek/deepseek-chat-v3-0324:free | free\ndeepseek/deepseek-v3-base:free | free\ndeepseek/deepseek-r1 | paid\ndeepseek/deepseek-r1:free | free\ndeepseek/deepseek-r1-0528 | paid\ndeepseek/deepseek-r1-0528:free | free\ndeepseek/deepseek-r1-distill-llama-70b | paid\ndeepseek/deepseek-r1-distill-llama-70b:free | free\ndeepseek/deepseek-r1-distill-llama-8b | paid\ndeepseek/deepseek-r1-distill-qwen-1.5b | paid\ndeepseek/deepseek-r1-distill-qwen-14b | paid\ndeepseek/deepseek-r1-distill-qwen-14b:free | free\ndeepseek/deepseek-r1-distill-qwen-32b | paid\ndeepseek/deepseek-r1-distill-qwen-32b:free | free\ndeepseek/deepseek-r1-distill-qwen-7b | paid\ncognitivecomputations/dolphin-mixtral-8x22b | paid\ncognitivecomputations/dolphin3.0-mistral-24b:free | free\ncognitivecomputations/dolphin3.0-r1-mistral-24b:free | free\neleutherai/llemma_7b | paid\neva-unit-01/eva-llama-3.33-70b | paid\neva-unit-01/eva-qwen-2.5-32b | paid\neva-unit-01/eva-qwen-2.5-72b | paid\nsao10k/fimbulvetr-11b-v2 | paid\nalpindale/goliath-120b | paid\ngoogle/gemini-flash-1.5 | paid\ngoogle/gemini-flash-1.5-8b | paid\ngoogle/gemini-pro-1.5 | paid\ngoogle/gemini-2.0-flash-001 | paid\ngoogle/gemini-2.0-flash-exp:free | free\ngoogle/gemini-2.0-flash-lite-001 | paid\ngoogle/gemini-2.5-flash-preview | paid\ngoogle/gemini-2.5-flash-preview:thinking | paid\ngoogle/gemini-2.5-flash-preview-05-20 | paid\ngoogle/gemini-2.5-flash-preview-05-20:thinking | paid\ngoogle/gemini-2.5-pro-exp-03-25 | paid\ngoogle/gemini-2.5-pro-preview | paid\ngoogle/gemma-2-27b-it | paid\ngoogle/gemma-2b-it | paid\ngoogle/gemma-2-9b-it | paid\ngoogle/gemma-2-9b-it:free | free\ngoogle/gemma-3-12b-it | paid\ngoogle/gemma-3-12b-it:free | free\ngoogle/gemma-3-1b-it:free | free\ngoogle/gemma-3-27b-it | paid\ngoogle/gemma-3-27b-it:free | free\ngoogle/gemma-3-4b-it | paid\ngoogle/gemma-3-4b-it:free | free\ngoogle/gemma-3n-e4b-it:free | free\ninception/mercury-coder-small-beta | paid\ninfermatic/mn-inferor-12b | paid\ninflection/inflection-3-pi | paid\ninflection/inflection-3-productivity | paid\nliquid/lfm-3b | paid\nliquid/lfm-40b | paid\nliquid/lfm-7b | paid\nmeta-llama/llama-guard-3-8b | paid\nalpindale/magnum-72b | paid\nanthracite-org/magnum-v2-72b | paid\nanthracite-org/magnum-v4-72b | paid\nmancer/weaver | paid\nmeta-llama/llama-2-70b-chat | paid\nmeta-llama/llama-3-70b-instruct | paid\nmeta-llama/llama-3-8b-instruct | paid\nmeta-llama/llama-3.1-405b | paid\nmeta-llama/llama-3.1-405b:free | free\nmeta-llama/llama-3.1-405b-instruct | paid\nmeta-llama/llama-3.1-70b-instruct | paid\nmeta-llama/llama-3.1-8b-instruct | paid\nmeta-llama/llama-3.1-8b-instruct:free | free\nmeta-llama/llama-3.2-11b-vision-instruct | paid\nmeta-llama/llama-3.2-11b-vision-instruct:free | free\nmeta-llama/llama-3.2-1b-instruct | paid\nmeta-llama/llama-3.2-1b-instruct:free | free\nmeta-llama/llama-3.2-3b-instruct | paid\nmeta-llama/llama-3.2-3b-instruct:free | free\nmeta-llama/llama-3.2-90b-vision-instruct | paid\nmeta-llama/llama-3.3-70b-instruct | paid\nmeta-llama/llama-3.3-70b-instruct:free | free\nmeta-llama/llama-3.3-8b-instruct:free | free\nmeta-llama/llama-4-maverick | paid\nmeta-llama/llama-4-maverick:free | free\nmeta-llama/llama-4-scout | paid\nmeta-llama/llama-4-scout:free | free\nmeta-llama/llama-guard-4-12b | paid\nmeta-llama/llama-guard-2-8b | paid\nmicrosoft/mai-ds-r1:free | free\nmicrosoft/phi-4 | paid\nmicrosoft/phi-4-multimodal-instruct | paid\nmicrosoft/phi-4-reasoning:free | free\nmicrosoft/phi-4-reasoning-plus | paid\nmicrosoft/phi-4-reasoning-plus:free | free\nmicrosoft/phi-3-medium-128k-instruct | paid\nmicrosoft/phi-3-mini-128k-instruct | paid\nmicrosoft/phi-3.5-mini-128k-instruct | paid\nsophosympatheia/midnight-rose-70b | paid\nminimax/minimax-01 | paid\nmistralai/mistral-large | paid\nmistralai/mistral-large-2407 | paid\nmistralai/mistral-large-2411 | paid\nmistralai/mistral-medium | paid\nnothingiisreal/mn-celeste-12b | paid\nmistralai/mistral-small | paid\nmistralai/mistral-tiny | paid\nmistralai/codestral-2501 | paid\nmistralai/devstral-small | paid\nmistralai/devstral-small:free | free\nmistralai/ministral-3b | paid\nmistralai/ministral-8b | paid\nmistralai/mistral-7b-instruct | paid\nmistralai/mistral-7b-instruct:free | free\nmistralai/mistral-7b-instruct-v0.1 | paid\nmistralai/mistral-7b-instruct-v0.2 | paid\nmistralai/mistral-7b-instruct-v0.3 | paid\nmistralai/mistral-medium-3 | paid\nmistralai/mistral-nemo | paid\nmistralai/mistral-nemo:free | free\nmistralai/mistral-small-24b-instruct-2501 | paid\nmistralai/mistral-small-24b-instruct-2501:free | free\nmistralai/mistral-small-3.1-24b-instruct | paid\nmistralai/mistral-small-3.1-24b-instruct:free | free\nmistralai/mixtral-8x22b-instruct | paid\nmistralai/mixtral-8x7b-instruct | paid\nmistralai/pixtral-12b | paid\nmistralai/pixtral-large-2411 | paid\nmistralai/mistral-saba | paid\nmoonshotai/kimi-vl-a3b-thinking:free | free\nmoonshotai/moonlight-16b-a3b-instruct:free | free\ngryphe/mythomax-l2-13b | paid\nneversleep/llama-3-lumimaid-70b | paid\nneversleep/llama-3-lumimaid-8b | paid\nneversleep/llama-3.1-lumimaid-70b | paid\nneversleep/llama-3.1-lumimaid-8b | paid\nneversleep/noromaid-20b | paid\nnousresearch/deephermes-3-llama-3-8b-preview:free | free\nnousresearch/deephermes-3-mistral-24b-preview:free | free\nnousresearch/nous-hermes-2-mixtral-8x7b-dpo | paid\nnousresearch/hermes-3-llama-3.1-405b | paid\nnousresearch/hermes-3-llama-3.1-70b | paid\nnousresearch/hermes-2-pro-llama-3-8b | paid\nnvidia/llama-3.1-nemotron-70b-instruct | paid\nnvidia/llama-3.1-nemotron-ultra-253b-v1 | paid\nnvidia/llama-3.1-nemotron-ultra-253b-v1:free | free\nnvidia/llama-3.3-nemotron-super-49b-v1 | paid\nnvidia/llama-3.3-nemotron-super-49b-v1:free | free\nopen-r1/olympiccoder-32b:free | free\nopenai/chatgpt-4o-latest | paid\nopenai/codex-mini | paid\nopenai/gpt-3.5-turbo | paid\nopenai/gpt-3.5-turbo-0613 | paid\nopenai/gpt-3.5-turbo-16k | paid\nopenai/gpt-3.5-turbo-0125 | paid\nopenai/gpt-3.5-turbo-1106 | paid\nopenai/gpt-3.5-turbo-instruct | paid\nopenai/gpt-4 | paid\nopenai/gpt-4-0314 | paid\nopenai/gpt-4-32k | paid\nopenai/gpt-4-32k-0314 | paid\nopenai/gpt-4-turbo | paid\nopenai/gpt-4-1106-preview | paid\nopenai/gpt-4-turbo-preview | paid\nopenai/gpt-4.1 | paid\nopenai/gpt-4.1-mini | paid\nopenai/gpt-4.1-nano | paid\nopenai/gpt-4.5-preview | paid\nopenai/gpt-4o | paid\nopenai/gpt-4o-2024-05-13 | paid\nopenai/gpt-4o-2024-08-06 | paid\nopenai/gpt-4o-2024-11-20 | paid\nopenai/gpt-4o:extended | paid\nopenai/gpt-4o-search-preview | paid\nopenai/gpt-4o-mini | paid\nopenai/gpt-4o-mini-2024-07-18 | paid\nopenai/gpt-4o-mini-search-preview | paid\nopenai/o1 | paid\nopenai/o1-mini | paid\nopenai/o1-mini-2024-09-12 | paid\nopenai/o1-preview | paid\nopenai/o1-preview-2024-09-12 | paid\nopenai/o1-pro | paid\nopenai/o3 | paid\nopenai/o3-mini | paid\nopenai/o3-mini-high | paid\nopenai/o4-mini | paid\nopenai/o4-mini-high | paid\nopengvlab/internvl3-14b:free | free\nopengvlab/internvl3-2b:free | free\nall-hands/openhands-lm-32b-v0.1 | paid\nperplexity/llama-3.1-sonar-large-128k-online | paid\nperplexity/llama-3.1-sonar-small-128k-online | paid\nperplexity/r1-1776 | paid\nperplexity/sonar | paid\nperplexity/sonar-deep-research | paid\nperplexity/sonar-pro | paid\nperplexity/sonar-reasoning | paid\nperplexity/sonar-reasoning-pro | paid\npygmalionai/mythalion-13b | paid\nqwen/qwen-2-72b-instruct | paid\nqwen/qwen-vl-max | paid\nqwen/qwen-vl-plus | paid\nqwen/qwen-max | paid\nqwen/qwen-plus | paid\nqwen/qwen-turbo | paid\nqwen/qwen2.5-vl-32b-instruct | paid\nqwen/qwen2.5-vl-32b-instruct:free | free\nqwen/qwen2.5-vl-3b-instruct:free | free\nqwen/qwen2.5-vl-72b-instruct | paid\nqwen/qwen2.5-vl-72b-instruct:free | free\nqwen/qwen-2.5-vl-7b-instruct | paid\nqwen/qwen-2.5-vl-7b-instruct:free | free\nqwen/qwen3-14b | paid\nqwen/qwen3-14b:free | free\nqwen/qwen3-235b-a22b | paid\nqwen/qwen3-235b-a22b:free | free\nqwen/qwen3-30b-a3b | paid\nqwen/qwen3-30b-a3b:free | free\nqwen/qwen3-32b | paid\nqwen/qwen3-32b:free | free\nqwen/qwen3-8b | paid\nqwen/qwen3-8b:free | free\nqwen/qwq-32b | paid\nqwen/qwq-32b:free | free\nqwen/qwq-32b-preview | paid\nqwen/qwen-2.5-72b-instruct | paid\nqwen/qwen-2.5-72b-instruct:free | free\nqwen/qwen-2.5-7b-instruct | paid\nqwen/qwen-2.5-7b-instruct:free | free\nqwen/qwen-2.5-coder-32b-instruct | paid\nqwen/qwen-2.5-coder-32b-instruct:free | free\nfeatherless/qwerky-72b:free | free\nrekaai/reka-flash-3:free | free\nundi95/remm-slerp-l2-13b | paid\nsao10k/l3-lunaris-8b | paid\nsao10k/l3-euryale-70b | paid\nsao10k/l3.1-euryale-70b | paid\nsao10k/l3.3-euryale-70b | paid\nsarvamai/sarvam-m | paid\nsarvamai/sarvam-m:free | free\nshisa-ai/shisa-v2-llama3.3-70b:free | free\nraifle/sorcererlm-8x22b | paid\nthedrummer/anubis-pro-105b-v1 | paid\nthedrummer/rocinante-12b | paid\nthedrummer/skyfall-36b-v2 | paid\nthedrummer/unslopnemo-12b | paid\nthedrummer/valkyrie-49b-v1 | paid\nthudm/glm-4-32b | paid\nthudm/glm-4-32b:free | free\nthudm/glm-z1-32b | paid\nthudm/glm-z1-32b:free | free\nthudm/glm-z1-rumination-32b | paid\ntngtech/deepseek-r1t-chimera:free | free\nundi95/toppy-m-7b | paid\nscb10x/llama3.1-typhoon2-70b-instruct | paid\nscb10x/llama3.1-typhoon2-8b-instruct | paid\nmicrosoft/wizardlm-2-8x22b | paid\nx-ai/grok-2-1212 | paid\nx-ai/grok-2-vision-1212 | paid\nx-ai/grok-3-beta | paid\nx-ai/grok-3-mini-beta | paid\nx-ai/grok-beta | paid\nx-ai/grok-vision-beta | paid\n\u001b[35m\u001b[1m\u001b[22m\u001b[39m\n\u001b[35m\u001b[1m OpenAI models:\u001b[22m\u001b[39m\n\u001b[35m\u001b[1m\u001b[22m\u001b[39m\nbabbage-002\nchatgpt-4o-latest\ncodex-mini-latest\ndall-e-2\ndall-e-3\ndavinci-002\ngpt-3.5-turbo\ngpt-3.5-turbo-0125\ngpt-3.5-turbo-1106\ngpt-3.5-turbo-16k\ngpt-3.5-turbo-instruct\ngpt-3.5-turbo-instruct-0914\ngpt-4\ngpt-4-0125-preview\ngpt-4-0613\ngpt-4-1106-preview\ngpt-4-turbo\ngpt-4-turbo-2024-04-09\ngpt-4-turbo-preview\ngpt-4.1\ngpt-4.1-2025-04-14\ngpt-4.1-mini\ngpt-4.1-mini-2025-04-14\ngpt-4.1-nano\ngpt-4.1-nano-2025-04-14\ngpt-4.5-preview\ngpt-4.5-preview-2025-02-27\ngpt-4o\ngpt-4o-2024-05-13\ngpt-4o-2024-08-06\ngpt-4o-2024-11-20\ngpt-4o-audio-preview\ngpt-4o-audio-preview-2024-10-01\ngpt-4o-audio-preview-2024-12-17\ngpt-4o-audio-preview-2025-06-03\ngpt-4o-mini\ngpt-4o-mini-2024-07-18\ngpt-4o-mini-audio-preview\ngpt-4o-mini-audio-preview-2024-12-17\ngpt-4o-mini-realtime-preview\ngpt-4o-mini-realtime-preview-2024-12-17\ngpt-4o-mini-search-preview\ngpt-4o-mini-search-preview-2025-03-11\ngpt-4o-mini-transcribe\ngpt-4o-mini-tts\ngpt-4o-realtime-preview\ngpt-4o-realtime-preview-2024-10-01\ngpt-4o-realtime-preview-2024-12-17\ngpt-4o-realtime-preview-2025-06-03\ngpt-4o-search-preview\ngpt-4o-search-preview-2025-03-11\ngpt-4o-transcribe\ngpt-image-1\no1\no1-2024-12-17\no1-mini\no1-mini-2024-09-12\no1-preview\no1-preview-2024-09-12\no1-pro\no1-pro-2025-03-19\no3-mini\no3-mini-2025-01-31\no4-mini\no4-mini-2025-04-16\nomni-moderation-2024-09-26\nomni-moderation-latest\ntext-embedding-3-large\ntext-embedding-3-small\ntext-embedding-ada-002\ntts-1\ntts-1-1106\ntts-1-hd\ntts-1-hd-1106\nwhisper-1\n-----\n\n\u001b[35m\u001b[1m\u001b[22m\u001b[39m\n\u001b[35m\u001b[1m Deepseek models:\u001b[22m\u001b[39m\n\u001b[35m\u001b[1m\u001b[22m\u001b[39m\ndeepseek-chat\ndeepseek-reasoner\n-----\n",
+ "ui:description": "AI model to use for processing. Available models:\n\u001b[35m\u001b[1m\u001b[22m\u001b[39m\n\u001b[35m\u001b[1m OpenRouter models:\u001b[22m\u001b[39m\n\u001b[35m\u001b[1m\u001b[22m\u001b[39m\n01-ai/yi-large | paid\naetherwiing/mn-starcannon-12b | paid\nagentica-org/deepcoder-14b-preview:free | free\nai21/jamba-1.6-large | paid\nai21/jamba-1.6-mini | paid\naion-labs/aion-1.0 | paid\naion-labs/aion-1.0-mini | paid\naion-labs/aion-rp-llama-3.1-8b | paid\nalfredpros/codellama-7b-instruct-solidity | paid\namazon/nova-lite-v1 | paid\namazon/nova-micro-v1 | paid\namazon/nova-pro-v1 | paid\nanthropic/claude-3-haiku | paid\nanthropic/claude-3-haiku:beta | paid\nanthropic/claude-3-opus | paid\nanthropic/claude-3-opus:beta | paid\nanthropic/claude-3-sonnet | paid\nanthropic/claude-3-sonnet:beta | paid\nanthropic/claude-3.5-haiku | paid\nanthropic/claude-3.5-haiku-20241022 | paid\nanthropic/claude-3.5-haiku-20241022:beta | paid\nanthropic/claude-3.5-haiku:beta | paid\nanthropic/claude-3.5-sonnet | paid\nanthropic/claude-3.5-sonnet-20240620 | paid\nanthropic/claude-3.5-sonnet-20240620:beta | paid\nanthropic/claude-3.5-sonnet:beta | paid\nanthropic/claude-3.7-sonnet | paid\nanthropic/claude-3.7-sonnet:beta | paid\nanthropic/claude-3.7-sonnet:thinking | paid\nanthropic/claude-opus-4 | paid\nanthropic/claude-sonnet-4 | paid\nanthropic/claude-2 | paid\nanthropic/claude-2:beta | paid\nanthropic/claude-2.0 | paid\nanthropic/claude-2.0:beta | paid\nanthropic/claude-2.1 | paid\nanthropic/claude-2.1:beta | paid\narcee-ai/arcee-blitz | paid\narcee-ai/caller-large | paid\narcee-ai/coder-large | paid\narcee-ai/maestro-reasoning | paid\narcee-ai/spotlight | paid\narcee-ai/virtuoso-large | paid\narcee-ai/virtuoso-medium-v2 | paid\narliai/qwq-32b-arliai-rpr-v1:free | free\nopenrouter/auto | paid\ncohere/command | paid\ncohere/command-a | paid\ncohere/command-r | paid\ncohere/command-r-03-2024 | paid\ncohere/command-r-08-2024 | paid\ncohere/command-r-plus | paid\ncohere/command-r-plus-04-2024 | paid\ncohere/command-r-plus-08-2024 | paid\ncohere/command-r7b-12-2024 | paid\ndeepseek/deepseek-prover-v2 | paid\ndeepseek/deepseek-r1-0528-qwen3-8b | paid\ndeepseek/deepseek-r1-0528-qwen3-8b:free | free\ndeepseek/deepseek-chat | paid\ndeepseek/deepseek-chat:free | free\ndeepseek/deepseek-chat-v3-0324 | paid\ndeepseek/deepseek-chat-v3-0324:free | free\ndeepseek/deepseek-v3-base:free | free\ndeepseek/deepseek-r1 | paid\ndeepseek/deepseek-r1:free | free\ndeepseek/deepseek-r1-0528 | paid\ndeepseek/deepseek-r1-0528:free | free\ndeepseek/deepseek-r1-distill-llama-70b | paid\ndeepseek/deepseek-r1-distill-llama-70b:free | free\ndeepseek/deepseek-r1-distill-llama-8b | paid\ndeepseek/deepseek-r1-distill-qwen-1.5b | paid\ndeepseek/deepseek-r1-distill-qwen-14b | paid\ndeepseek/deepseek-r1-distill-qwen-14b:free | free\ndeepseek/deepseek-r1-distill-qwen-32b | paid\ndeepseek/deepseek-r1-distill-qwen-7b | paid\ncognitivecomputations/dolphin-mixtral-8x22b | paid\ncognitivecomputations/dolphin3.0-mistral-24b:free | free\ncognitivecomputations/dolphin3.0-r1-mistral-24b:free | free\neleutherai/llemma_7b | paid\neva-unit-01/eva-llama-3.33-70b | paid\neva-unit-01/eva-qwen-2.5-32b | paid\neva-unit-01/eva-qwen-2.5-72b | paid\nsao10k/fimbulvetr-11b-v2 | paid\nalpindale/goliath-120b | paid\ngoogle/gemini-flash-1.5 | paid\ngoogle/gemini-flash-1.5-8b | paid\ngoogle/gemini-pro-1.5 | paid\ngoogle/gemini-2.0-flash-001 | paid\ngoogle/gemini-2.0-flash-exp:free | free\ngoogle/gemini-2.0-flash-lite-001 | paid\ngoogle/gemini-2.5-flash | paid\ngoogle/gemini-2.5-flash-lite-preview-06-17 | paid\ngoogle/gemini-2.5-flash-preview | paid\ngoogle/gemini-2.5-flash-preview:thinking | paid\ngoogle/gemini-2.5-flash-preview-05-20 | paid\ngoogle/gemini-2.5-flash-preview-05-20:thinking | paid\ngoogle/gemini-2.5-pro | paid\ngoogle/gemini-2.5-pro-exp-03-25 | paid\ngoogle/gemini-2.5-pro-preview-05-06 | paid\ngoogle/gemini-2.5-pro-preview | paid\ngoogle/gemma-2-27b-it | paid\ngoogle/gemma-2-9b-it | paid\ngoogle/gemma-2-9b-it:free | free\ngoogle/gemma-3-12b-it | paid\ngoogle/gemma-3-12b-it:free | free\ngoogle/gemma-3-27b-it | paid\ngoogle/gemma-3-27b-it:free | free\ngoogle/gemma-3-4b-it | paid\ngoogle/gemma-3-4b-it:free | free\ngoogle/gemma-3n-e4b-it | paid\ngoogle/gemma-3n-e4b-it:free | free\ninception/mercury | paid\ninception/mercury-coder-small-beta | paid\ninfermatic/mn-inferor-12b | paid\ninflection/inflection-3-pi | paid\ninflection/inflection-3-productivity | paid\nmoonshotai/kimi-dev-72b:free | free\nliquid/lfm-3b | paid\nliquid/lfm-40b | paid\nliquid/lfm-7b | paid\nmeta-llama/llama-guard-3-8b | paid\nalpindale/magnum-72b | paid\nanthracite-org/magnum-v2-72b | paid\nanthracite-org/magnum-v4-72b | paid\nmancer/weaver | paid\nmeta-llama/llama-3-70b-instruct | paid\nmeta-llama/llama-3-8b-instruct | paid\nmeta-llama/llama-3.1-405b | paid\nmeta-llama/llama-3.1-405b-instruct | paid\nmeta-llama/llama-3.1-70b-instruct | paid\nmeta-llama/llama-3.1-8b-instruct | paid\nmeta-llama/llama-3.1-8b-instruct:free | free\nmeta-llama/llama-3.2-11b-vision-instruct | paid\nmeta-llama/llama-3.2-11b-vision-instruct:free | free\nmeta-llama/llama-3.2-1b-instruct | paid\nmeta-llama/llama-3.2-1b-instruct:free | free\nmeta-llama/llama-3.2-3b-instruct | paid\nmeta-llama/llama-3.2-90b-vision-instruct | paid\nmeta-llama/llama-3.3-70b-instruct | paid\nmeta-llama/llama-3.3-70b-instruct:free | free\nmeta-llama/llama-4-maverick | paid\nmeta-llama/llama-4-maverick:free | free\nmeta-llama/llama-4-scout | paid\nmeta-llama/llama-4-scout:free | free\nmeta-llama/llama-guard-4-12b | paid\nmeta-llama/llama-guard-2-8b | paid\nmicrosoft/mai-ds-r1:free | free\nmicrosoft/phi-4 | paid\nmicrosoft/phi-4-multimodal-instruct | paid\nmicrosoft/phi-4-reasoning-plus | paid\nmicrosoft/phi-3-medium-128k-instruct | paid\nmicrosoft/phi-3-mini-128k-instruct | paid\nmicrosoft/phi-3.5-mini-128k-instruct | paid\nsophosympatheia/midnight-rose-70b | paid\nminimax/minimax-m1 | paid\nminimax/minimax-m1:extended | paid\nminimax/minimax-01 | paid\nmistralai/mistral-large | paid\nmistralai/mistral-large-2407 | paid\nmistralai/mistral-large-2411 | paid\nnothingiisreal/mn-celeste-12b | paid\nmistralai/mistral-small | paid\nmistralai/mistral-tiny | paid\nmistralai/codestral-2501 | paid\nmistralai/devstral-small | paid\nmistralai/devstral-small:free | free\nmistralai/magistral-medium-2506 | paid\nmistralai/magistral-medium-2506:thinking | paid\nmistralai/magistral-small-2506 | paid\nmistralai/ministral-3b | paid\nmistralai/ministral-8b | paid\nmistralai/mistral-7b-instruct | paid\nmistralai/mistral-7b-instruct:free | free\nmistralai/mistral-7b-instruct-v0.1 | paid\nmistralai/mistral-7b-instruct-v0.2 | paid\nmistralai/mistral-7b-instruct-v0.3 | paid\nmistralai/mistral-medium-3 | paid\nmistralai/mistral-nemo | paid\nmistralai/mistral-nemo:free | free\nmistralai/mistral-small-24b-instruct-2501 | paid\nmistralai/mistral-small-24b-instruct-2501:free | free\nmistralai/mistral-small-3.1-24b-instruct | paid\nmistralai/mistral-small-3.1-24b-instruct:free | free\nmistralai/mistral-small-3.2-24b-instruct | paid\nmistralai/mistral-small-3.2-24b-instruct:free | free\nmistralai/mixtral-8x22b-instruct | paid\nmistralai/mixtral-8x7b-instruct | paid\nmistralai/pixtral-12b | paid\nmistralai/pixtral-large-2411 | paid\nmistralai/mistral-saba | paid\nmoonshotai/kimi-vl-a3b-thinking:free | free\nmorph/morph-v2 | paid\ngryphe/mythomax-l2-13b | paid\nneversleep/llama-3-lumimaid-70b | paid\nneversleep/llama-3-lumimaid-8b | paid\nneversleep/llama-3.1-lumimaid-70b | paid\nneversleep/llama-3.1-lumimaid-8b | paid\nneversleep/noromaid-20b | paid\nnousresearch/deephermes-3-llama-3-8b-preview:free | free\nnousresearch/nous-hermes-2-mixtral-8x7b-dpo | paid\nnousresearch/hermes-3-llama-3.1-405b | paid\nnousresearch/hermes-3-llama-3.1-70b | paid\nnousresearch/hermes-2-pro-llama-3-8b | paid\nnvidia/llama-3.1-nemotron-70b-instruct | paid\nnvidia/llama-3.1-nemotron-ultra-253b-v1 | paid\nnvidia/llama-3.1-nemotron-ultra-253b-v1:free | free\nnvidia/llama-3.3-nemotron-super-49b-v1 | paid\nnvidia/llama-3.3-nemotron-super-49b-v1:free | free\nopenai/chatgpt-4o-latest | paid\nopenai/codex-mini | paid\nopenai/gpt-3.5-turbo-0613 | paid\nopenai/gpt-3.5-turbo-16k | paid\nopenai/gpt-3.5-turbo-instruct | paid\nopenai/gpt-4 | paid\nopenai/gpt-4-0314 | paid\nopenai/gpt-4-turbo | paid\nopenai/gpt-4-1106-preview | paid\nopenai/gpt-4-turbo-preview | paid\nopenai/gpt-4.1 | paid\nopenai/gpt-4.1-mini | paid\nopenai/gpt-4.1-nano | paid\nopenai/gpt-4.5-preview | paid\nopenai/gpt-4o | paid\nopenai/gpt-4o-2024-05-13 | paid\nopenai/gpt-4o-2024-08-06 | paid\nopenai/gpt-4o-2024-11-20 | paid\nopenai/gpt-4o:extended | paid\nopenai/gpt-4o-search-preview | paid\nopenai/gpt-4o-mini | paid\nopenai/gpt-4o-mini-2024-07-18 | paid\nopenai/gpt-4o-mini-search-preview | paid\nopenai/o1 | paid\nopenai/o1-mini | paid\nopenai/o1-mini-2024-09-12 | paid\nopenai/o1-preview | paid\nopenai/o1-preview-2024-09-12 | paid\nopenai/o1-pro | paid\nopenai/o3 | paid\nopenai/o3-mini | paid\nopenai/o3-mini-high | paid\nopenai/o3-pro | paid\nopenai/o4-mini | paid\nopenai/o4-mini-high | paid\nopengvlab/internvl3-14b | paid\nopengvlab/internvl3-2b | paid\nall-hands/openhands-lm-32b-v0.1 | paid\nperplexity/llama-3.1-sonar-large-128k-online | paid\nperplexity/llama-3.1-sonar-small-128k-online | paid\nperplexity/r1-1776 | paid\nperplexity/sonar | paid\nperplexity/sonar-deep-research | paid\nperplexity/sonar-pro | paid\nperplexity/sonar-reasoning | paid\nperplexity/sonar-reasoning-pro | paid\npygmalionai/mythalion-13b | paid\nqwen/qwen-2-72b-instruct | paid\nqwen/qwen-vl-max | paid\nqwen/qwen-vl-plus | paid\nqwen/qwen-max | paid\nqwen/qwen-plus | paid\nqwen/qwen-turbo | paid\nqwen/qwen2.5-vl-32b-instruct | paid\nqwen/qwen2.5-vl-32b-instruct:free | free\nqwen/qwen2.5-vl-72b-instruct | paid\nqwen/qwen2.5-vl-72b-instruct:free | free\nqwen/qwen-2.5-vl-7b-instruct | paid\nqwen/qwen3-14b | paid\nqwen/qwen3-14b:free | free\nqwen/qwen3-235b-a22b | paid\nqwen/qwen3-235b-a22b:free | free\nqwen/qwen3-30b-a3b | paid\nqwen/qwen3-30b-a3b:free | free\nqwen/qwen3-32b | paid\nqwen/qwen3-32b:free | free\nqwen/qwen3-8b | paid\nqwen/qwen3-8b:free | free\nqwen/qwq-32b | paid\nqwen/qwq-32b:free | free\nqwen/qwq-32b-preview | paid\nqwen/qwen-2.5-72b-instruct | paid\nqwen/qwen-2.5-72b-instruct:free | free\nqwen/qwen-2.5-7b-instruct | paid\nqwen/qwen-2.5-coder-32b-instruct | paid\nqwen/qwen-2.5-coder-32b-instruct:free | free\nfeatherless/qwerky-72b:free | free\nrekaai/reka-flash-3:free | free\nundi95/remm-slerp-l2-13b | paid\nsao10k/l3-lunaris-8b | paid\nsao10k/l3-euryale-70b | paid\nsao10k/l3.1-euryale-70b | paid\nsao10k/l3.3-euryale-70b | paid\nsarvamai/sarvam-m:free | free\nsentientagi/dobby-mini-unhinged-plus-llama-3.1-8b | paid\nshisa-ai/shisa-v2-llama3.3-70b:free | free\nraifle/sorcererlm-8x22b | paid\nthedrummer/anubis-pro-105b-v1 | paid\nthedrummer/rocinante-12b | paid\nthedrummer/skyfall-36b-v2 | paid\nthedrummer/unslopnemo-12b | paid\nthedrummer/valkyrie-49b-v1 | paid\nthudm/glm-4-32b | paid\nthudm/glm-4-32b:free | free\nthudm/glm-z1-32b | paid\nthudm/glm-z1-32b:free | free\nthudm/glm-z1-rumination-32b | paid\ntngtech/deepseek-r1t-chimera:free | free\nundi95/toppy-m-7b | paid\nscb10x/llama3.1-typhoon2-70b-instruct | paid\nmicrosoft/wizardlm-2-8x22b | paid\nx-ai/grok-2-1212 | paid\nx-ai/grok-2-vision-1212 | paid\nx-ai/grok-3 | paid\nx-ai/grok-3-beta | paid\nx-ai/grok-3-mini | paid\nx-ai/grok-3-mini-beta | paid\nx-ai/grok-beta | paid\nx-ai/grok-vision-beta | paid\n\u001b[35m\u001b[1m\u001b[22m\u001b[39m\n\u001b[35m\u001b[1m OpenAI models:\u001b[22m\u001b[39m\n\u001b[35m\u001b[1m\u001b[22m\u001b[39m\nbabbage-002\nchatgpt-4o-latest\ncodex-mini-latest\ndall-e-2\ndall-e-3\ndavinci-002\ngpt-3.5-turbo\ngpt-3.5-turbo-0125\ngpt-3.5-turbo-1106\ngpt-3.5-turbo-16k\ngpt-3.5-turbo-instruct\ngpt-3.5-turbo-instruct-0914\ngpt-4\ngpt-4-0125-preview\ngpt-4-0613\ngpt-4-1106-preview\ngpt-4-turbo\ngpt-4-turbo-2024-04-09\ngpt-4-turbo-preview\ngpt-4.1\ngpt-4.1-2025-04-14\ngpt-4.1-mini\ngpt-4.1-mini-2025-04-14\ngpt-4.1-nano\ngpt-4.1-nano-2025-04-14\ngpt-4.5-preview\ngpt-4.5-preview-2025-02-27\ngpt-4o\ngpt-4o-2024-05-13\ngpt-4o-2024-08-06\ngpt-4o-2024-11-20\ngpt-4o-audio-preview\ngpt-4o-audio-preview-2024-10-01\ngpt-4o-audio-preview-2024-12-17\ngpt-4o-audio-preview-2025-06-03\ngpt-4o-mini\ngpt-4o-mini-2024-07-18\ngpt-4o-mini-audio-preview\ngpt-4o-mini-audio-preview-2024-12-17\ngpt-4o-mini-realtime-preview\ngpt-4o-mini-realtime-preview-2024-12-17\ngpt-4o-mini-search-preview\ngpt-4o-mini-search-preview-2025-03-11\ngpt-4o-mini-transcribe\ngpt-4o-mini-tts\ngpt-4o-realtime-preview\ngpt-4o-realtime-preview-2024-10-01\ngpt-4o-realtime-preview-2024-12-17\ngpt-4o-realtime-preview-2025-06-03\ngpt-4o-search-preview\ngpt-4o-search-preview-2025-03-11\ngpt-4o-transcribe\ngpt-image-1\no1\no1-2024-12-17\no1-mini\no1-mini-2024-09-12\no1-preview\no1-preview-2024-09-12\no1-pro\no1-pro-2025-03-19\no3-mini\no3-mini-2025-01-31\no4-mini\no4-mini-2025-04-16\no4-mini-deep-research\no4-mini-deep-research-2025-06-26\nomni-moderation-2024-09-26\nomni-moderation-latest\ntext-embedding-3-large\ntext-embedding-3-small\ntext-embedding-ada-002\ntts-1\ntts-1-1106\ntts-1-hd\ntts-1-hd-1106\nwhisper-1\n-----\n\n\u001b[35m\u001b[1m\u001b[22m\u001b[39m\n\u001b[35m\u001b[1m Deepseek models:\u001b[22m\u001b[39m\n\u001b[35m\u001b[1m\u001b[22m\u001b[39m\ndeepseek-chat\ndeepseek-reasoner\n-----\n",
"ui:title": "Model"
},
"router": {
diff --git a/packages/kbot/src/models/cache/openai-models.ts b/packages/kbot/src/models/cache/openai-models.ts
index 5046f529..6984f11d 100644
--- a/packages/kbot/src/models/cache/openai-models.ts
+++ b/packages/kbot/src/models/cache/openai-models.ts
@@ -2,11 +2,11 @@ export enum E_OPENAI_MODEL {
MODEL_GPT_4_0613 = "gpt-4-0613",
MODEL_GPT_4 = "gpt-4",
MODEL_GPT_3_5_TURBO = "gpt-3.5-turbo",
- MODEL_GPT_4O_AUDIO_PREVIEW_2025_06_03 = "gpt-4o-audio-preview-2025-06-03",
- MODEL_GPT_4_1_NANO = "gpt-4.1-nano",
- MODEL_GPT_IMAGE_1 = "gpt-image-1",
+ MODEL_O4_MINI_DEEP_RESEARCH_2025_06_26 = "o4-mini-deep-research-2025-06-26",
MODEL_CODEX_MINI_LATEST = "codex-mini-latest",
MODEL_GPT_4O_REALTIME_PREVIEW_2025_06_03 = "gpt-4o-realtime-preview-2025-06-03",
+ MODEL_GPT_4O_AUDIO_PREVIEW_2025_06_03 = "gpt-4o-audio-preview-2025-06-03",
+ MODEL_O4_MINI_DEEP_RESEARCH = "o4-mini-deep-research",
MODEL_DAVINCI_002 = "davinci-002",
MODEL_BABBAGE_002 = "babbage-002",
MODEL_GPT_3_5_TURBO_INSTRUCT = "gpt-3.5-turbo-instruct",
@@ -70,6 +70,8 @@ export enum E_OPENAI_MODEL {
MODEL_GPT_4_1_MINI_2025_04_14 = "gpt-4.1-mini-2025-04-14",
MODEL_GPT_4_1_MINI = "gpt-4.1-mini",
MODEL_GPT_4_1_NANO_2025_04_14 = "gpt-4.1-nano-2025-04-14",
+ MODEL_GPT_4_1_NANO = "gpt-4.1-nano",
+ MODEL_GPT_IMAGE_1 = "gpt-image-1",
MODEL_GPT_3_5_TURBO_16K = "gpt-3.5-turbo-16k",
MODEL_TTS_1 = "tts-1",
MODEL_WHISPER_1 = "whisper-1",
diff --git a/packages/kbot/src/models/cache/openrouter-models-free.ts b/packages/kbot/src/models/cache/openrouter-models-free.ts
index 38fc0ba3..63dbe27c 100644
--- a/packages/kbot/src/models/cache/openrouter-models-free.ts
+++ b/packages/kbot/src/models/cache/openrouter-models-free.ts
@@ -1,16 +1,12 @@
export enum E_OPENROUTER_MODEL_FREE {
+ MODEL_FREE_MISTRALAI_MISTRAL_SMALL_3_2_24B_INSTRUCT_FREE = "mistralai/mistral-small-3.2-24b-instruct:free",
+ MODEL_FREE_MINIMAX_MINIMAX_M1_EXTENDED = "minimax/minimax-m1:extended",
+ MODEL_FREE_MOONSHOTAI_KIMI_DEV_72B_FREE = "moonshotai/kimi-dev-72b:free",
MODEL_FREE_DEEPSEEK_DEEPSEEK_R1_0528_QWEN3_8B_FREE = "deepseek/deepseek-r1-0528-qwen3-8b:free",
MODEL_FREE_DEEPSEEK_DEEPSEEK_R1_0528_FREE = "deepseek/deepseek-r1-0528:free",
MODEL_FREE_SARVAMAI_SARVAM_M_FREE = "sarvamai/sarvam-m:free",
MODEL_FREE_MISTRALAI_DEVSTRAL_SMALL_FREE = "mistralai/devstral-small:free",
MODEL_FREE_GOOGLE_GEMMA_3N_E4B_IT_FREE = "google/gemma-3n-e4b-it:free",
- MODEL_FREE_META_LLAMA_LLAMA_3_3_8B_INSTRUCT_FREE = "meta-llama/llama-3.3-8b-instruct:free",
- MODEL_FREE_NOUSRESEARCH_DEEPHERMES_3_MISTRAL_24B_PREVIEW_FREE = "nousresearch/deephermes-3-mistral-24b-preview:free",
- MODEL_FREE_MICROSOFT_PHI_4_REASONING_PLUS_FREE = "microsoft/phi-4-reasoning-plus:free",
- MODEL_FREE_MICROSOFT_PHI_4_REASONING_FREE = "microsoft/phi-4-reasoning:free",
- MODEL_FREE_OPENGVLAB_INTERNVL3_14B_FREE = "opengvlab/internvl3-14b:free",
- MODEL_FREE_OPENGVLAB_INTERNVL3_2B_FREE = "opengvlab/internvl3-2b:free",
- MODEL_FREE_DEEPSEEK_DEEPSEEK_PROVER_V2_FREE = "deepseek/deepseek-prover-v2:free",
MODEL_FREE_QWEN_QWEN3_30B_A3B_FREE = "qwen/qwen3-30b-a3b:free",
MODEL_FREE_QWEN_QWEN3_8B_FREE = "qwen/qwen3-8b:free",
MODEL_FREE_QWEN_QWEN3_14B_FREE = "qwen/qwen3-14b:free",
@@ -29,27 +25,21 @@ export enum E_OPENROUTER_MODEL_FREE {
MODEL_FREE_META_LLAMA_LLAMA_4_MAVERICK_FREE = "meta-llama/llama-4-maverick:free",
MODEL_FREE_META_LLAMA_LLAMA_4_SCOUT_FREE = "meta-llama/llama-4-scout:free",
MODEL_FREE_DEEPSEEK_DEEPSEEK_V3_BASE_FREE = "deepseek/deepseek-v3-base:free",
- MODEL_FREE_QWEN_QWEN2_5_VL_3B_INSTRUCT_FREE = "qwen/qwen2.5-vl-3b-instruct:free",
MODEL_FREE_GOOGLE_GEMINI_2_5_PRO_EXP_03_25 = "google/gemini-2.5-pro-exp-03-25",
MODEL_FREE_QWEN_QWEN2_5_VL_32B_INSTRUCT_FREE = "qwen/qwen2.5-vl-32b-instruct:free",
MODEL_FREE_DEEPSEEK_DEEPSEEK_CHAT_V3_0324_FREE = "deepseek/deepseek-chat-v3-0324:free",
MODEL_FREE_FEATHERLESS_QWERKY_72B_FREE = "featherless/qwerky-72b:free",
MODEL_FREE_MISTRALAI_MISTRAL_SMALL_3_1_24B_INSTRUCT_FREE = "mistralai/mistral-small-3.1-24b-instruct:free",
- MODEL_FREE_OPEN_R1_OLYMPICCODER_32B_FREE = "open-r1/olympiccoder-32b:free",
- MODEL_FREE_GOOGLE_GEMMA_3_1B_IT_FREE = "google/gemma-3-1b-it:free",
MODEL_FREE_GOOGLE_GEMMA_3_4B_IT_FREE = "google/gemma-3-4b-it:free",
MODEL_FREE_GOOGLE_GEMMA_3_12B_IT_FREE = "google/gemma-3-12b-it:free",
MODEL_FREE_REKAAI_REKA_FLASH_3_FREE = "rekaai/reka-flash-3:free",
MODEL_FREE_GOOGLE_GEMMA_3_27B_IT_FREE = "google/gemma-3-27b-it:free",
- MODEL_FREE_DEEPSEEK_DEEPSEEK_R1_ZERO_FREE = "deepseek/deepseek-r1-zero:free",
MODEL_FREE_QWEN_QWQ_32B_FREE = "qwen/qwq-32b:free",
- MODEL_FREE_MOONSHOTAI_MOONLIGHT_16B_A3B_INSTRUCT_FREE = "moonshotai/moonlight-16b-a3b-instruct:free",
MODEL_FREE_NOUSRESEARCH_DEEPHERMES_3_LLAMA_3_8B_PREVIEW_FREE = "nousresearch/deephermes-3-llama-3-8b-preview:free",
MODEL_FREE_COGNITIVECOMPUTATIONS_DOLPHIN3_0_R1_MISTRAL_24B_FREE = "cognitivecomputations/dolphin3.0-r1-mistral-24b:free",
MODEL_FREE_COGNITIVECOMPUTATIONS_DOLPHIN3_0_MISTRAL_24B_FREE = "cognitivecomputations/dolphin3.0-mistral-24b:free",
MODEL_FREE_QWEN_QWEN2_5_VL_72B_INSTRUCT_FREE = "qwen/qwen2.5-vl-72b-instruct:free",
MODEL_FREE_MISTRALAI_MISTRAL_SMALL_24B_INSTRUCT_2501_FREE = "mistralai/mistral-small-24b-instruct-2501:free",
- MODEL_FREE_DEEPSEEK_DEEPSEEK_R1_DISTILL_QWEN_32B_FREE = "deepseek/deepseek-r1-distill-qwen-32b:free",
MODEL_FREE_DEEPSEEK_DEEPSEEK_R1_DISTILL_QWEN_14B_FREE = "deepseek/deepseek-r1-distill-qwen-14b:free",
MODEL_FREE_DEEPSEEK_DEEPSEEK_R1_DISTILL_LLAMA_70B_FREE = "deepseek/deepseek-r1-distill-llama-70b:free",
MODEL_FREE_DEEPSEEK_DEEPSEEK_R1_FREE = "deepseek/deepseek-r1:free",
@@ -57,13 +47,9 @@ export enum E_OPENROUTER_MODEL_FREE {
MODEL_FREE_GOOGLE_GEMINI_2_0_FLASH_EXP_FREE = "google/gemini-2.0-flash-exp:free",
MODEL_FREE_META_LLAMA_LLAMA_3_3_70B_INSTRUCT_FREE = "meta-llama/llama-3.3-70b-instruct:free",
MODEL_FREE_QWEN_QWEN_2_5_CODER_32B_INSTRUCT_FREE = "qwen/qwen-2.5-coder-32b-instruct:free",
- MODEL_FREE_QWEN_QWEN_2_5_7B_INSTRUCT_FREE = "qwen/qwen-2.5-7b-instruct:free",
- MODEL_FREE_META_LLAMA_LLAMA_3_2_3B_INSTRUCT_FREE = "meta-llama/llama-3.2-3b-instruct:free",
- MODEL_FREE_META_LLAMA_LLAMA_3_2_1B_INSTRUCT_FREE = "meta-llama/llama-3.2-1b-instruct:free",
MODEL_FREE_META_LLAMA_LLAMA_3_2_11B_VISION_INSTRUCT_FREE = "meta-llama/llama-3.2-11b-vision-instruct:free",
+ MODEL_FREE_META_LLAMA_LLAMA_3_2_1B_INSTRUCT_FREE = "meta-llama/llama-3.2-1b-instruct:free",
MODEL_FREE_QWEN_QWEN_2_5_72B_INSTRUCT_FREE = "qwen/qwen-2.5-72b-instruct:free",
- MODEL_FREE_QWEN_QWEN_2_5_VL_7B_INSTRUCT_FREE = "qwen/qwen-2.5-vl-7b-instruct:free",
- MODEL_FREE_META_LLAMA_LLAMA_3_1_405B_FREE = "meta-llama/llama-3.1-405b:free",
MODEL_FREE_META_LLAMA_LLAMA_3_1_8B_INSTRUCT_FREE = "meta-llama/llama-3.1-8b-instruct:free",
MODEL_FREE_MISTRALAI_MISTRAL_NEMO_FREE = "mistralai/mistral-nemo:free",
MODEL_FREE_GOOGLE_GEMMA_2_9B_IT_FREE = "google/gemma-2-9b-it:free",
diff --git a/packages/kbot/src/models/cache/openrouter-models.ts b/packages/kbot/src/models/cache/openrouter-models.ts
index 3e13917f..2428198d 100644
--- a/packages/kbot/src/models/cache/openrouter-models.ts
+++ b/packages/kbot/src/models/cache/openrouter-models.ts
@@ -1,25 +1,38 @@
export enum E_OPENROUTER_MODEL {
+ MODEL_INCEPTION_MERCURY = "inception/mercury",
+ MODEL_MORPH_MORPH_V2 = "morph/morph-v2",
+ MODEL_MISTRALAI_MISTRAL_SMALL_3_2_24B_INSTRUCT_FREE = "mistralai/mistral-small-3.2-24b-instruct:free",
+ MODEL_MISTRALAI_MISTRAL_SMALL_3_2_24B_INSTRUCT = "mistralai/mistral-small-3.2-24b-instruct",
+ MODEL_MINIMAX_MINIMAX_M1_EXTENDED = "minimax/minimax-m1:extended",
+ MODEL_MINIMAX_MINIMAX_M1 = "minimax/minimax-m1",
+ MODEL_GOOGLE_GEMINI_2_5_FLASH_LITE_PREVIEW_06_17 = "google/gemini-2.5-flash-lite-preview-06-17",
+ MODEL_GOOGLE_GEMINI_2_5_FLASH = "google/gemini-2.5-flash",
+ MODEL_GOOGLE_GEMINI_2_5_PRO = "google/gemini-2.5-pro",
+ MODEL_MOONSHOTAI_KIMI_DEV_72B_FREE = "moonshotai/kimi-dev-72b:free",
+ MODEL_OPENAI_O3_PRO = "openai/o3-pro",
+ MODEL_X_AI_GROK_3_MINI = "x-ai/grok-3-mini",
+ MODEL_X_AI_GROK_3 = "x-ai/grok-3",
+ MODEL_MISTRALAI_MAGISTRAL_SMALL_2506 = "mistralai/magistral-small-2506",
+ MODEL_MISTRALAI_MAGISTRAL_MEDIUM_2506 = "mistralai/magistral-medium-2506",
+ MODEL_MISTRALAI_MAGISTRAL_MEDIUM_2506_THINKING = "mistralai/magistral-medium-2506:thinking",
MODEL_GOOGLE_GEMINI_2_5_PRO_PREVIEW = "google/gemini-2.5-pro-preview",
MODEL_SENTIENTAGI_DOBBY_MINI_UNHINGED_PLUS_LLAMA_3_1_8B = "sentientagi/dobby-mini-unhinged-plus-llama-3.1-8b",
MODEL_DEEPSEEK_DEEPSEEK_R1_DISTILL_QWEN_7B = "deepseek/deepseek-r1-distill-qwen-7b",
MODEL_DEEPSEEK_DEEPSEEK_R1_0528_QWEN3_8B_FREE = "deepseek/deepseek-r1-0528-qwen3-8b:free",
MODEL_DEEPSEEK_DEEPSEEK_R1_0528_QWEN3_8B = "deepseek/deepseek-r1-0528-qwen3-8b",
- MODEL_GOOGLE_GEMMA_2B_IT = "google/gemma-2b-it",
MODEL_DEEPSEEK_DEEPSEEK_R1_0528_FREE = "deepseek/deepseek-r1-0528:free",
MODEL_DEEPSEEK_DEEPSEEK_R1_0528 = "deepseek/deepseek-r1-0528",
MODEL_SARVAMAI_SARVAM_M_FREE = "sarvamai/sarvam-m:free",
- MODEL_SARVAMAI_SARVAM_M = "sarvamai/sarvam-m",
MODEL_THEDRUMMER_VALKYRIE_49B_V1 = "thedrummer/valkyrie-49b-v1",
MODEL_ANTHROPIC_CLAUDE_OPUS_4 = "anthropic/claude-opus-4",
MODEL_ANTHROPIC_CLAUDE_SONNET_4 = "anthropic/claude-sonnet-4",
MODEL_MISTRALAI_DEVSTRAL_SMALL_FREE = "mistralai/devstral-small:free",
MODEL_MISTRALAI_DEVSTRAL_SMALL = "mistralai/devstral-small",
MODEL_GOOGLE_GEMMA_3N_E4B_IT_FREE = "google/gemma-3n-e4b-it:free",
+ MODEL_GOOGLE_GEMMA_3N_E4B_IT = "google/gemma-3n-e4b-it",
MODEL_GOOGLE_GEMINI_2_5_FLASH_PREVIEW_05_20 = "google/gemini-2.5-flash-preview-05-20",
MODEL_GOOGLE_GEMINI_2_5_FLASH_PREVIEW_05_20_THINKING = "google/gemini-2.5-flash-preview-05-20:thinking",
MODEL_OPENAI_CODEX_MINI = "openai/codex-mini",
- MODEL_META_LLAMA_LLAMA_3_3_8B_INSTRUCT_FREE = "meta-llama/llama-3.3-8b-instruct:free",
- MODEL_NOUSRESEARCH_DEEPHERMES_3_MISTRAL_24B_PREVIEW_FREE = "nousresearch/deephermes-3-mistral-24b-preview:free",
MODEL_MISTRALAI_MISTRAL_MEDIUM_3 = "mistralai/mistral-medium-3",
MODEL_GOOGLE_GEMINI_2_5_PRO_PREVIEW_05_06 = "google/gemini-2.5-pro-preview-05-06",
MODEL_ARCEE_AI_CALLER_LARGE = "arcee-ai/caller-large",
@@ -29,13 +42,10 @@ export enum E_OPENROUTER_MODEL {
MODEL_ARCEE_AI_CODER_LARGE = "arcee-ai/coder-large",
MODEL_ARCEE_AI_VIRTUOSO_MEDIUM_V2 = "arcee-ai/virtuoso-medium-v2",
MODEL_ARCEE_AI_ARCEE_BLITZ = "arcee-ai/arcee-blitz",
- MODEL_MICROSOFT_PHI_4_REASONING_PLUS_FREE = "microsoft/phi-4-reasoning-plus:free",
MODEL_MICROSOFT_PHI_4_REASONING_PLUS = "microsoft/phi-4-reasoning-plus",
- MODEL_MICROSOFT_PHI_4_REASONING_FREE = "microsoft/phi-4-reasoning:free",
MODEL_INCEPTION_MERCURY_CODER_SMALL_BETA = "inception/mercury-coder-small-beta",
- MODEL_OPENGVLAB_INTERNVL3_14B_FREE = "opengvlab/internvl3-14b:free",
- MODEL_OPENGVLAB_INTERNVL3_2B_FREE = "opengvlab/internvl3-2b:free",
- MODEL_DEEPSEEK_DEEPSEEK_PROVER_V2_FREE = "deepseek/deepseek-prover-v2:free",
+ MODEL_OPENGVLAB_INTERNVL3_14B = "opengvlab/internvl3-14b",
+ MODEL_OPENGVLAB_INTERNVL3_2B = "opengvlab/internvl3-2b",
MODEL_DEEPSEEK_DEEPSEEK_PROVER_V2 = "deepseek/deepseek-prover-v2",
MODEL_META_LLAMA_LLAMA_GUARD_4_12B = "meta-llama/llama-guard-4-12b",
MODEL_QWEN_QWEN3_30B_A3B_FREE = "qwen/qwen3-30b-a3b:free",
@@ -81,9 +91,7 @@ export enum E_OPENROUTER_MODEL {
MODEL_META_LLAMA_LLAMA_4_SCOUT = "meta-llama/llama-4-scout",
MODEL_ALL_HANDS_OPENHANDS_LM_32B_V0_1 = "all-hands/openhands-lm-32b-v0.1",
MODEL_DEEPSEEK_DEEPSEEK_V3_BASE_FREE = "deepseek/deepseek-v3-base:free",
- MODEL_SCB10X_LLAMA3_1_TYPHOON2_8B_INSTRUCT = "scb10x/llama3.1-typhoon2-8b-instruct",
MODEL_SCB10X_LLAMA3_1_TYPHOON2_70B_INSTRUCT = "scb10x/llama3.1-typhoon2-70b-instruct",
- MODEL_QWEN_QWEN2_5_VL_3B_INSTRUCT_FREE = "qwen/qwen2.5-vl-3b-instruct:free",
MODEL_GOOGLE_GEMINI_2_5_PRO_EXP_03_25 = "google/gemini-2.5-pro-exp-03-25",
MODEL_QWEN_QWEN2_5_VL_32B_INSTRUCT_FREE = "qwen/qwen2.5-vl-32b-instruct:free",
MODEL_QWEN_QWEN2_5_VL_32B_INSTRUCT = "qwen/qwen2.5-vl-32b-instruct",
@@ -93,8 +101,6 @@ export enum E_OPENROUTER_MODEL {
MODEL_OPENAI_O1_PRO = "openai/o1-pro",
MODEL_MISTRALAI_MISTRAL_SMALL_3_1_24B_INSTRUCT_FREE = "mistralai/mistral-small-3.1-24b-instruct:free",
MODEL_MISTRALAI_MISTRAL_SMALL_3_1_24B_INSTRUCT = "mistralai/mistral-small-3.1-24b-instruct",
- MODEL_OPEN_R1_OLYMPICCODER_32B_FREE = "open-r1/olympiccoder-32b:free",
- MODEL_GOOGLE_GEMMA_3_1B_IT_FREE = "google/gemma-3-1b-it:free",
MODEL_GOOGLE_GEMMA_3_4B_IT_FREE = "google/gemma-3-4b-it:free",
MODEL_GOOGLE_GEMMA_3_4B_IT = "google/gemma-3-4b-it",
MODEL_AI21_JAMBA_1_6_LARGE = "ai21/jamba-1.6-large",
@@ -113,10 +119,8 @@ export enum E_OPENROUTER_MODEL {
MODEL_PERPLEXITY_SONAR_REASONING_PRO = "perplexity/sonar-reasoning-pro",
MODEL_PERPLEXITY_SONAR_PRO = "perplexity/sonar-pro",
MODEL_PERPLEXITY_SONAR_DEEP_RESEARCH = "perplexity/sonar-deep-research",
- MODEL_DEEPSEEK_DEEPSEEK_R1_ZERO_FREE = "deepseek/deepseek-r1-zero:free",
MODEL_QWEN_QWQ_32B_FREE = "qwen/qwq-32b:free",
MODEL_QWEN_QWQ_32B = "qwen/qwq-32b",
- MODEL_MOONSHOTAI_MOONLIGHT_16B_A3B_INSTRUCT_FREE = "moonshotai/moonlight-16b-a3b-instruct:free",
MODEL_NOUSRESEARCH_DEEPHERMES_3_LLAMA_3_8B_PREVIEW_FREE = "nousresearch/deephermes-3-llama-3-8b-preview:free",
MODEL_OPENAI_GPT_4_5_PREVIEW = "openai/gpt-4.5-preview",
MODEL_GOOGLE_GEMINI_2_0_FLASH_LITE_001 = "google/gemini-2.0-flash-lite-001",
@@ -145,7 +149,6 @@ export enum E_OPENROUTER_MODEL {
MODEL_DEEPSEEK_DEEPSEEK_R1_DISTILL_QWEN_1_5B = "deepseek/deepseek-r1-distill-qwen-1.5b",
MODEL_MISTRALAI_MISTRAL_SMALL_24B_INSTRUCT_2501_FREE = "mistralai/mistral-small-24b-instruct-2501:free",
MODEL_MISTRALAI_MISTRAL_SMALL_24B_INSTRUCT_2501 = "mistralai/mistral-small-24b-instruct-2501",
- MODEL_DEEPSEEK_DEEPSEEK_R1_DISTILL_QWEN_32B_FREE = "deepseek/deepseek-r1-distill-qwen-32b:free",
MODEL_DEEPSEEK_DEEPSEEK_R1_DISTILL_QWEN_32B = "deepseek/deepseek-r1-distill-qwen-32b",
MODEL_DEEPSEEK_DEEPSEEK_R1_DISTILL_QWEN_14B_FREE = "deepseek/deepseek-r1-distill-qwen-14b:free",
MODEL_DEEPSEEK_DEEPSEEK_R1_DISTILL_QWEN_14B = "deepseek/deepseek-r1-distill-qwen-14b",
@@ -191,58 +194,54 @@ export enum E_OPENROUTER_MODEL {
MODEL_ANTHROPIC_CLAUDE_3_5_HAIKU = "anthropic/claude-3.5-haiku",
MODEL_ANTHROPIC_CLAUDE_3_5_HAIKU_20241022_BETA = "anthropic/claude-3.5-haiku-20241022:beta",
MODEL_ANTHROPIC_CLAUDE_3_5_HAIKU_20241022 = "anthropic/claude-3.5-haiku-20241022",
- MODEL_NEVERSLEEP_LLAMA_3_1_LUMIMAID_70B = "neversleep/llama-3.1-lumimaid-70b",
- MODEL_ANTHRACITE_ORG_MAGNUM_V4_72B = "anthracite-org/magnum-v4-72b",
MODEL_ANTHROPIC_CLAUDE_3_5_SONNET_BETA = "anthropic/claude-3.5-sonnet:beta",
MODEL_ANTHROPIC_CLAUDE_3_5_SONNET = "anthropic/claude-3.5-sonnet",
+ MODEL_ANTHRACITE_ORG_MAGNUM_V4_72B = "anthracite-org/magnum-v4-72b",
+ MODEL_NEVERSLEEP_LLAMA_3_1_LUMIMAID_70B = "neversleep/llama-3.1-lumimaid-70b",
MODEL_X_AI_GROK_BETA = "x-ai/grok-beta",
MODEL_MISTRALAI_MINISTRAL_8B = "mistralai/ministral-8b",
MODEL_MISTRALAI_MINISTRAL_3B = "mistralai/ministral-3b",
- MODEL_QWEN_QWEN_2_5_7B_INSTRUCT_FREE = "qwen/qwen-2.5-7b-instruct:free",
MODEL_QWEN_QWEN_2_5_7B_INSTRUCT = "qwen/qwen-2.5-7b-instruct",
MODEL_NVIDIA_LLAMA_3_1_NEMOTRON_70B_INSTRUCT = "nvidia/llama-3.1-nemotron-70b-instruct",
- MODEL_INFLECTION_INFLECTION_3_PRODUCTIVITY = "inflection/inflection-3-productivity",
MODEL_INFLECTION_INFLECTION_3_PI = "inflection/inflection-3-pi",
+ MODEL_INFLECTION_INFLECTION_3_PRODUCTIVITY = "inflection/inflection-3-productivity",
MODEL_GOOGLE_GEMINI_FLASH_1_5_8B = "google/gemini-flash-1.5-8b",
- MODEL_THEDRUMMER_ROCINANTE_12B = "thedrummer/rocinante-12b",
- MODEL_ANTHRACITE_ORG_MAGNUM_V2_72B = "anthracite-org/magnum-v2-72b",
MODEL_LIQUID_LFM_40B = "liquid/lfm-40b",
- MODEL_META_LLAMA_LLAMA_3_2_3B_INSTRUCT_FREE = "meta-llama/llama-3.2-3b-instruct:free",
+ MODEL_ANTHRACITE_ORG_MAGNUM_V2_72B = "anthracite-org/magnum-v2-72b",
+ MODEL_THEDRUMMER_ROCINANTE_12B = "thedrummer/rocinante-12b",
MODEL_META_LLAMA_LLAMA_3_2_3B_INSTRUCT = "meta-llama/llama-3.2-3b-instruct",
- MODEL_META_LLAMA_LLAMA_3_2_1B_INSTRUCT_FREE = "meta-llama/llama-3.2-1b-instruct:free",
- MODEL_META_LLAMA_LLAMA_3_2_1B_INSTRUCT = "meta-llama/llama-3.2-1b-instruct",
MODEL_META_LLAMA_LLAMA_3_2_90B_VISION_INSTRUCT = "meta-llama/llama-3.2-90b-vision-instruct",
MODEL_META_LLAMA_LLAMA_3_2_11B_VISION_INSTRUCT_FREE = "meta-llama/llama-3.2-11b-vision-instruct:free",
MODEL_META_LLAMA_LLAMA_3_2_11B_VISION_INSTRUCT = "meta-llama/llama-3.2-11b-vision-instruct",
+ MODEL_META_LLAMA_LLAMA_3_2_1B_INSTRUCT_FREE = "meta-llama/llama-3.2-1b-instruct:free",
+ MODEL_META_LLAMA_LLAMA_3_2_1B_INSTRUCT = "meta-llama/llama-3.2-1b-instruct",
MODEL_QWEN_QWEN_2_5_72B_INSTRUCT_FREE = "qwen/qwen-2.5-72b-instruct:free",
MODEL_QWEN_QWEN_2_5_72B_INSTRUCT = "qwen/qwen-2.5-72b-instruct",
MODEL_NEVERSLEEP_LLAMA_3_1_LUMIMAID_8B = "neversleep/llama-3.1-lumimaid-8b",
MODEL_OPENAI_O1_PREVIEW = "openai/o1-preview",
+ MODEL_OPENAI_O1_MINI_2024_09_12 = "openai/o1-mini-2024-09-12",
MODEL_OPENAI_O1_PREVIEW_2024_09_12 = "openai/o1-preview-2024-09-12",
MODEL_OPENAI_O1_MINI = "openai/o1-mini",
- MODEL_OPENAI_O1_MINI_2024_09_12 = "openai/o1-mini-2024-09-12",
MODEL_MISTRALAI_PIXTRAL_12B = "mistralai/pixtral-12b",
MODEL_COHERE_COMMAND_R_PLUS_08_2024 = "cohere/command-r-plus-08-2024",
MODEL_COHERE_COMMAND_R_08_2024 = "cohere/command-r-08-2024",
- MODEL_QWEN_QWEN_2_5_VL_7B_INSTRUCT_FREE = "qwen/qwen-2.5-vl-7b-instruct:free",
MODEL_QWEN_QWEN_2_5_VL_7B_INSTRUCT = "qwen/qwen-2.5-vl-7b-instruct",
MODEL_SAO10K_L3_1_EURYALE_70B = "sao10k/l3.1-euryale-70b",
MODEL_MICROSOFT_PHI_3_5_MINI_128K_INSTRUCT = "microsoft/phi-3.5-mini-128k-instruct",
MODEL_NOUSRESEARCH_HERMES_3_LLAMA_3_1_70B = "nousresearch/hermes-3-llama-3.1-70b",
MODEL_NOUSRESEARCH_HERMES_3_LLAMA_3_1_405B = "nousresearch/hermes-3-llama-3.1-405b",
MODEL_OPENAI_CHATGPT_4O_LATEST = "openai/chatgpt-4o-latest",
- MODEL_SAO10K_L3_LUNARIS_8B = "sao10k/l3-lunaris-8b",
MODEL_AETHERWIING_MN_STARCANNON_12B = "aetherwiing/mn-starcannon-12b",
+ MODEL_SAO10K_L3_LUNARIS_8B = "sao10k/l3-lunaris-8b",
MODEL_OPENAI_GPT_4O_2024_08_06 = "openai/gpt-4o-2024-08-06",
- MODEL_META_LLAMA_LLAMA_3_1_405B_FREE = "meta-llama/llama-3.1-405b:free",
- MODEL_META_LLAMA_LLAMA_3_1_405B = "meta-llama/llama-3.1-405b",
MODEL_NOTHINGIISREAL_MN_CELESTE_12B = "nothingiisreal/mn-celeste-12b",
+ MODEL_META_LLAMA_LLAMA_3_1_405B = "meta-llama/llama-3.1-405b",
MODEL_PERPLEXITY_LLAMA_3_1_SONAR_SMALL_128K_ONLINE = "perplexity/llama-3.1-sonar-small-128k-online",
MODEL_PERPLEXITY_LLAMA_3_1_SONAR_LARGE_128K_ONLINE = "perplexity/llama-3.1-sonar-large-128k-online",
- MODEL_META_LLAMA_LLAMA_3_1_8B_INSTRUCT_FREE = "meta-llama/llama-3.1-8b-instruct:free",
- MODEL_META_LLAMA_LLAMA_3_1_8B_INSTRUCT = "meta-llama/llama-3.1-8b-instruct",
MODEL_META_LLAMA_LLAMA_3_1_405B_INSTRUCT = "meta-llama/llama-3.1-405b-instruct",
MODEL_META_LLAMA_LLAMA_3_1_70B_INSTRUCT = "meta-llama/llama-3.1-70b-instruct",
+ MODEL_META_LLAMA_LLAMA_3_1_8B_INSTRUCT_FREE = "meta-llama/llama-3.1-8b-instruct:free",
+ MODEL_META_LLAMA_LLAMA_3_1_8B_INSTRUCT = "meta-llama/llama-3.1-8b-instruct",
MODEL_MISTRALAI_MISTRAL_NEMO_FREE = "mistralai/mistral-nemo:free",
MODEL_MISTRALAI_MISTRAL_NEMO = "mistralai/mistral-nemo",
MODEL_OPENAI_GPT_4O_MINI = "openai/gpt-4o-mini",
@@ -257,31 +256,31 @@ export enum E_OPENROUTER_MODEL {
MODEL_SAO10K_L3_EURYALE_70B = "sao10k/l3-euryale-70b",
MODEL_COGNITIVECOMPUTATIONS_DOLPHIN_MIXTRAL_8X22B = "cognitivecomputations/dolphin-mixtral-8x22b",
MODEL_QWEN_QWEN_2_72B_INSTRUCT = "qwen/qwen-2-72b-instruct",
+ MODEL_MISTRALAI_MISTRAL_7B_INSTRUCT_V0_3 = "mistralai/mistral-7b-instruct-v0.3",
+ MODEL_NOUSRESEARCH_HERMES_2_PRO_LLAMA_3_8B = "nousresearch/hermes-2-pro-llama-3-8b",
MODEL_MISTRALAI_MISTRAL_7B_INSTRUCT_FREE = "mistralai/mistral-7b-instruct:free",
MODEL_MISTRALAI_MISTRAL_7B_INSTRUCT = "mistralai/mistral-7b-instruct",
- MODEL_NOUSRESEARCH_HERMES_2_PRO_LLAMA_3_8B = "nousresearch/hermes-2-pro-llama-3-8b",
- MODEL_MISTRALAI_MISTRAL_7B_INSTRUCT_V0_3 = "mistralai/mistral-7b-instruct-v0.3",
MODEL_MICROSOFT_PHI_3_MINI_128K_INSTRUCT = "microsoft/phi-3-mini-128k-instruct",
MODEL_MICROSOFT_PHI_3_MEDIUM_128K_INSTRUCT = "microsoft/phi-3-medium-128k-instruct",
MODEL_NEVERSLEEP_LLAMA_3_LUMIMAID_70B = "neversleep/llama-3-lumimaid-70b",
MODEL_GOOGLE_GEMINI_FLASH_1_5 = "google/gemini-flash-1.5",
+ MODEL_OPENAI_GPT_4O_2024_05_13 = "openai/gpt-4o-2024-05-13",
+ MODEL_META_LLAMA_LLAMA_GUARD_2_8B = "meta-llama/llama-guard-2-8b",
MODEL_OPENAI_GPT_4O = "openai/gpt-4o",
MODEL_OPENAI_GPT_4O_EXTENDED = "openai/gpt-4o:extended",
- MODEL_META_LLAMA_LLAMA_GUARD_2_8B = "meta-llama/llama-guard-2-8b",
- MODEL_OPENAI_GPT_4O_2024_05_13 = "openai/gpt-4o-2024-05-13",
MODEL_NEVERSLEEP_LLAMA_3_LUMIMAID_8B = "neversleep/llama-3-lumimaid-8b",
MODEL_SAO10K_FIMBULVETR_11B_V2 = "sao10k/fimbulvetr-11b-v2",
MODEL_META_LLAMA_LLAMA_3_8B_INSTRUCT = "meta-llama/llama-3-8b-instruct",
MODEL_META_LLAMA_LLAMA_3_70B_INSTRUCT = "meta-llama/llama-3-70b-instruct",
MODEL_MISTRALAI_MIXTRAL_8X22B_INSTRUCT = "mistralai/mixtral-8x22b-instruct",
MODEL_MICROSOFT_WIZARDLM_2_8X22B = "microsoft/wizardlm-2-8x22b",
- MODEL_GOOGLE_GEMINI_PRO_1_5 = "google/gemini-pro-1.5",
MODEL_OPENAI_GPT_4_TURBO = "openai/gpt-4-turbo",
+ MODEL_GOOGLE_GEMINI_PRO_1_5 = "google/gemini-pro-1.5",
MODEL_COHERE_COMMAND_R_PLUS = "cohere/command-r-plus",
MODEL_COHERE_COMMAND_R_PLUS_04_2024 = "cohere/command-r-plus-04-2024",
MODEL_SOPHOSYMPATHEIA_MIDNIGHT_ROSE_70B = "sophosympatheia/midnight-rose-70b",
- MODEL_COHERE_COMMAND = "cohere/command",
MODEL_COHERE_COMMAND_R = "cohere/command-r",
+ MODEL_COHERE_COMMAND = "cohere/command",
MODEL_ANTHROPIC_CLAUDE_3_HAIKU_BETA = "anthropic/claude-3-haiku:beta",
MODEL_ANTHROPIC_CLAUDE_3_HAIKU = "anthropic/claude-3-haiku",
MODEL_ANTHROPIC_CLAUDE_3_OPUS_BETA = "anthropic/claude-3-opus:beta",
@@ -293,20 +292,18 @@ export enum E_OPENROUTER_MODEL {
MODEL_OPENAI_GPT_3_5_TURBO_0613 = "openai/gpt-3.5-turbo-0613",
MODEL_OPENAI_GPT_4_TURBO_PREVIEW = "openai/gpt-4-turbo-preview",
MODEL_NOUSRESEARCH_NOUS_HERMES_2_MIXTRAL_8X7B_DPO = "nousresearch/nous-hermes-2-mixtral-8x7b-dpo",
- MODEL_MISTRALAI_MISTRAL_MEDIUM = "mistralai/mistral-medium",
MODEL_MISTRALAI_MISTRAL_SMALL = "mistralai/mistral-small",
MODEL_MISTRALAI_MISTRAL_TINY = "mistralai/mistral-tiny",
MODEL_MISTRALAI_MISTRAL_7B_INSTRUCT_V0_2 = "mistralai/mistral-7b-instruct-v0.2",
MODEL_MISTRALAI_MIXTRAL_8X7B_INSTRUCT = "mistralai/mixtral-8x7b-instruct",
MODEL_NEVERSLEEP_NOROMAID_20B = "neversleep/noromaid-20b",
- MODEL_ANTHROPIC_CLAUDE_2_1_BETA = "anthropic/claude-2.1:beta",
- MODEL_ANTHROPIC_CLAUDE_2_1 = "anthropic/claude-2.1",
MODEL_ANTHROPIC_CLAUDE_2_BETA = "anthropic/claude-2:beta",
MODEL_ANTHROPIC_CLAUDE_2 = "anthropic/claude-2",
+ MODEL_ANTHROPIC_CLAUDE_2_1_BETA = "anthropic/claude-2.1:beta",
+ MODEL_ANTHROPIC_CLAUDE_2_1 = "anthropic/claude-2.1",
MODEL_UNDI95_TOPPY_M_7B = "undi95/toppy-m-7b",
MODEL_ALPINDALE_GOLIATH_120B = "alpindale/goliath-120b",
MODEL_OPENROUTER_AUTO = "openrouter/auto",
- MODEL_OPENAI_GPT_3_5_TURBO_1106 = "openai/gpt-3.5-turbo-1106",
MODEL_OPENAI_GPT_4_1106_PREVIEW = "openai/gpt-4-1106-preview",
MODEL_OPENAI_GPT_3_5_TURBO_INSTRUCT = "openai/gpt-3.5-turbo-instruct",
MODEL_MISTRALAI_MISTRAL_7B_INSTRUCT_V0_1 = "mistralai/mistral-7b-instruct-v0.1",
@@ -317,9 +314,6 @@ export enum E_OPENROUTER_MODEL {
MODEL_ANTHROPIC_CLAUDE_2_0 = "anthropic/claude-2.0",
MODEL_UNDI95_REMM_SLERP_L2_13B = "undi95/remm-slerp-l2-13b",
MODEL_GRYPHE_MYTHOMAX_L2_13B = "gryphe/mythomax-l2-13b",
- MODEL_META_LLAMA_LLAMA_2_70B_CHAT = "meta-llama/llama-2-70b-chat",
- MODEL_OPENAI_GPT_3_5_TURBO = "openai/gpt-3.5-turbo",
- MODEL_OPENAI_GPT_3_5_TURBO_0125 = "openai/gpt-3.5-turbo-0125",
MODEL_OPENAI_GPT_4 = "openai/gpt-4",
MODEL_OPENAI_GPT_4_0314 = "openai/gpt-4-0314"
}
\ No newline at end of file
diff --git a/packages/kbot/src/zod_types.ts b/packages/kbot/src/zod_types.ts
index ee1654e1..0459edef 100644
--- a/packages/kbot/src/zod_types.ts
+++ b/packages/kbot/src/zod_types.ts
@@ -87,10 +87,8 @@ export interface IKBotOptions {
cohere/command-r-plus-08-2024 | paid
cohere/command-r7b-12-2024 | paid
deepseek/deepseek-prover-v2 | paid
- deepseek/deepseek-prover-v2:free | free
deepseek/deepseek-r1-0528-qwen3-8b | paid
deepseek/deepseek-r1-0528-qwen3-8b:free | free
- deepseek/deepseek-r1-zero:free | free
deepseek/deepseek-chat | paid
deepseek/deepseek-chat:free | free
deepseek/deepseek-chat-v3-0324 | paid
@@ -107,7 +105,6 @@ export interface IKBotOptions {
deepseek/deepseek-r1-distill-qwen-14b | paid
deepseek/deepseek-r1-distill-qwen-14b:free | free
deepseek/deepseek-r1-distill-qwen-32b | paid
- deepseek/deepseek-r1-distill-qwen-32b:free | free
deepseek/deepseek-r1-distill-qwen-7b | paid
cognitivecomputations/dolphin-mixtral-8x22b | paid
cognitivecomputations/dolphin3.0-mistral-24b:free | free
@@ -124,28 +121,33 @@ export interface IKBotOptions {
google/gemini-2.0-flash-001 | paid
google/gemini-2.0-flash-exp:free | free
google/gemini-2.0-flash-lite-001 | paid
+ google/gemini-2.5-flash | paid
+ google/gemini-2.5-flash-lite-preview-06-17 | paid
google/gemini-2.5-flash-preview | paid
google/gemini-2.5-flash-preview:thinking | paid
google/gemini-2.5-flash-preview-05-20 | paid
google/gemini-2.5-flash-preview-05-20:thinking | paid
+ google/gemini-2.5-pro | paid
google/gemini-2.5-pro-exp-03-25 | paid
+ google/gemini-2.5-pro-preview-05-06 | paid
google/gemini-2.5-pro-preview | paid
google/gemma-2-27b-it | paid
- google/gemma-2b-it | paid
google/gemma-2-9b-it | paid
google/gemma-2-9b-it:free | free
google/gemma-3-12b-it | paid
google/gemma-3-12b-it:free | free
- google/gemma-3-1b-it:free | free
google/gemma-3-27b-it | paid
google/gemma-3-27b-it:free | free
google/gemma-3-4b-it | paid
google/gemma-3-4b-it:free | free
+ google/gemma-3n-e4b-it | paid
google/gemma-3n-e4b-it:free | free
+ inception/mercury | paid
inception/mercury-coder-small-beta | paid
infermatic/mn-inferor-12b | paid
inflection/inflection-3-pi | paid
inflection/inflection-3-productivity | paid
+ moonshotai/kimi-dev-72b:free | free
liquid/lfm-3b | paid
liquid/lfm-40b | paid
liquid/lfm-7b | paid
@@ -154,11 +156,9 @@ export interface IKBotOptions {
anthracite-org/magnum-v2-72b | paid
anthracite-org/magnum-v4-72b | paid
mancer/weaver | paid
- meta-llama/llama-2-70b-chat | paid
meta-llama/llama-3-70b-instruct | paid
meta-llama/llama-3-8b-instruct | paid
meta-llama/llama-3.1-405b | paid
- meta-llama/llama-3.1-405b:free | free
meta-llama/llama-3.1-405b-instruct | paid
meta-llama/llama-3.1-70b-instruct | paid
meta-llama/llama-3.1-8b-instruct | paid
@@ -168,11 +168,9 @@ export interface IKBotOptions {
meta-llama/llama-3.2-1b-instruct | paid
meta-llama/llama-3.2-1b-instruct:free | free
meta-llama/llama-3.2-3b-instruct | paid
- meta-llama/llama-3.2-3b-instruct:free | free
meta-llama/llama-3.2-90b-vision-instruct | paid
meta-llama/llama-3.3-70b-instruct | paid
meta-llama/llama-3.3-70b-instruct:free | free
- meta-llama/llama-3.3-8b-instruct:free | free
meta-llama/llama-4-maverick | paid
meta-llama/llama-4-maverick:free | free
meta-llama/llama-4-scout | paid
@@ -182,24 +180,26 @@ export interface IKBotOptions {
microsoft/mai-ds-r1:free | free
microsoft/phi-4 | paid
microsoft/phi-4-multimodal-instruct | paid
- microsoft/phi-4-reasoning:free | free
microsoft/phi-4-reasoning-plus | paid
- microsoft/phi-4-reasoning-plus:free | free
microsoft/phi-3-medium-128k-instruct | paid
microsoft/phi-3-mini-128k-instruct | paid
microsoft/phi-3.5-mini-128k-instruct | paid
sophosympatheia/midnight-rose-70b | paid
+ minimax/minimax-m1 | paid
+ minimax/minimax-m1:extended | paid
minimax/minimax-01 | paid
mistralai/mistral-large | paid
mistralai/mistral-large-2407 | paid
mistralai/mistral-large-2411 | paid
- mistralai/mistral-medium | paid
nothingiisreal/mn-celeste-12b | paid
mistralai/mistral-small | paid
mistralai/mistral-tiny | paid
mistralai/codestral-2501 | paid
mistralai/devstral-small | paid
mistralai/devstral-small:free | free
+ mistralai/magistral-medium-2506 | paid
+ mistralai/magistral-medium-2506:thinking | paid
+ mistralai/magistral-small-2506 | paid
mistralai/ministral-3b | paid
mistralai/ministral-8b | paid
mistralai/mistral-7b-instruct | paid
@@ -214,13 +214,15 @@ export interface IKBotOptions {
mistralai/mistral-small-24b-instruct-2501:free | free
mistralai/mistral-small-3.1-24b-instruct | paid
mistralai/mistral-small-3.1-24b-instruct:free | free
+ mistralai/mistral-small-3.2-24b-instruct | paid
+ mistralai/mistral-small-3.2-24b-instruct:free | free
mistralai/mixtral-8x22b-instruct | paid
mistralai/mixtral-8x7b-instruct | paid
mistralai/pixtral-12b | paid
mistralai/pixtral-large-2411 | paid
mistralai/mistral-saba | paid
moonshotai/kimi-vl-a3b-thinking:free | free
- moonshotai/moonlight-16b-a3b-instruct:free | free
+ morph/morph-v2 | paid
gryphe/mythomax-l2-13b | paid
neversleep/llama-3-lumimaid-70b | paid
neversleep/llama-3-lumimaid-8b | paid
@@ -228,7 +230,6 @@ export interface IKBotOptions {
neversleep/llama-3.1-lumimaid-8b | paid
neversleep/noromaid-20b | paid
nousresearch/deephermes-3-llama-3-8b-preview:free | free
- nousresearch/deephermes-3-mistral-24b-preview:free | free
nousresearch/nous-hermes-2-mixtral-8x7b-dpo | paid
nousresearch/hermes-3-llama-3.1-405b | paid
nousresearch/hermes-3-llama-3.1-70b | paid
@@ -238,19 +239,13 @@ export interface IKBotOptions {
nvidia/llama-3.1-nemotron-ultra-253b-v1:free | free
nvidia/llama-3.3-nemotron-super-49b-v1 | paid
nvidia/llama-3.3-nemotron-super-49b-v1:free | free
- open-r1/olympiccoder-32b:free | free
openai/chatgpt-4o-latest | paid
openai/codex-mini | paid
- openai/gpt-3.5-turbo | paid
openai/gpt-3.5-turbo-0613 | paid
openai/gpt-3.5-turbo-16k | paid
- openai/gpt-3.5-turbo-0125 | paid
- openai/gpt-3.5-turbo-1106 | paid
openai/gpt-3.5-turbo-instruct | paid
openai/gpt-4 | paid
openai/gpt-4-0314 | paid
- openai/gpt-4-32k | paid
- openai/gpt-4-32k-0314 | paid
openai/gpt-4-turbo | paid
openai/gpt-4-1106-preview | paid
openai/gpt-4-turbo-preview | paid
@@ -276,10 +271,11 @@ export interface IKBotOptions {
openai/o3 | paid
openai/o3-mini | paid
openai/o3-mini-high | paid
+ openai/o3-pro | paid
openai/o4-mini | paid
openai/o4-mini-high | paid
- opengvlab/internvl3-14b:free | free
- opengvlab/internvl3-2b:free | free
+ opengvlab/internvl3-14b | paid
+ opengvlab/internvl3-2b | paid
all-hands/openhands-lm-32b-v0.1 | paid
perplexity/llama-3.1-sonar-large-128k-online | paid
perplexity/llama-3.1-sonar-small-128k-online | paid
@@ -298,11 +294,9 @@ export interface IKBotOptions {
qwen/qwen-turbo | paid
qwen/qwen2.5-vl-32b-instruct | paid
qwen/qwen2.5-vl-32b-instruct:free | free
- qwen/qwen2.5-vl-3b-instruct:free | free
qwen/qwen2.5-vl-72b-instruct | paid
qwen/qwen2.5-vl-72b-instruct:free | free
qwen/qwen-2.5-vl-7b-instruct | paid
- qwen/qwen-2.5-vl-7b-instruct:free | free
qwen/qwen3-14b | paid
qwen/qwen3-14b:free | free
qwen/qwen3-235b-a22b | paid
@@ -319,7 +313,6 @@ export interface IKBotOptions {
qwen/qwen-2.5-72b-instruct | paid
qwen/qwen-2.5-72b-instruct:free | free
qwen/qwen-2.5-7b-instruct | paid
- qwen/qwen-2.5-7b-instruct:free | free
qwen/qwen-2.5-coder-32b-instruct | paid
qwen/qwen-2.5-coder-32b-instruct:free | free
featherless/qwerky-72b:free | free
@@ -329,8 +322,8 @@ export interface IKBotOptions {
sao10k/l3-euryale-70b | paid
sao10k/l3.1-euryale-70b | paid
sao10k/l3.3-euryale-70b | paid
- sarvamai/sarvam-m | paid
sarvamai/sarvam-m:free | free
+ sentientagi/dobby-mini-unhinged-plus-llama-3.1-8b | paid
shisa-ai/shisa-v2-llama3.3-70b:free | free
raifle/sorcererlm-8x22b | paid
thedrummer/anubis-pro-105b-v1 | paid
@@ -346,11 +339,12 @@ export interface IKBotOptions {
tngtech/deepseek-r1t-chimera:free | free
undi95/toppy-m-7b | paid
scb10x/llama3.1-typhoon2-70b-instruct | paid
- scb10x/llama3.1-typhoon2-8b-instruct | paid
microsoft/wizardlm-2-8x22b | paid
x-ai/grok-2-1212 | paid
x-ai/grok-2-vision-1212 | paid
+ x-ai/grok-3 | paid
x-ai/grok-3-beta | paid
+ x-ai/grok-3-mini | paid
x-ai/grok-3-mini-beta | paid
x-ai/grok-beta | paid
x-ai/grok-vision-beta | paid
@@ -422,6 +416,8 @@ export interface IKBotOptions {
o3-mini-2025-01-31
o4-mini
o4-mini-2025-04-16
+ o4-mini-deep-research
+ o4-mini-deep-research-2025-06-26
omni-moderation-2024-09-26
omni-moderation-latest
text-embedding-3-large
diff --git a/packages/kbot/tests/unit/reports/all.json b/packages/kbot/tests/unit/reports/all.json
index acde51dc..b1a1bbf4 100644
--- a/packages/kbot/tests/unit/reports/all.json
+++ b/packages/kbot/tests/unit/reports/all.json
@@ -5081,6 +5081,473 @@
"passed": true,
"duration": 1336,
"category": "language"
+ },
+ {
+ "test": "synonyms",
+ "prompt": "Provide a synonym for \"happy\". Return only the synonym, no explanation.",
+ "result": [
+ "Joyful"
+ ],
+ "expected": "joyful",
+ "model": "openai/gpt-4o-mini",
+ "router": "openai/gpt-4o-mini",
+ "timestamp": "2025-06-05T22:54:17.429Z",
+ "passed": true,
+ "duration": 561,
+ "category": "language"
+ },
+ {
+ "test": "synonyms",
+ "prompt": "Provide a synonym for \"happy\". Return only the synonym, no explanation.",
+ "result": [
+ "Joyful"
+ ],
+ "expected": "joyful",
+ "model": "deepseek/deepseek-r1:free",
+ "router": "deepseek/deepseek-r1:free",
+ "timestamp": "2025-06-05T22:54:21.005Z",
+ "passed": true,
+ "duration": 3571,
+ "category": "language"
+ },
+ {
+ "test": "addition",
+ "prompt": "add 5 and 3. Return only the number, no explanation.",
+ "result": [
+ "8"
+ ],
+ "expected": "8",
+ "model": "anthropic/claude-sonnet-4",
+ "router": "anthropic/claude-sonnet-4",
+ "timestamp": "2025-06-05T22:54:48.966Z",
+ "passed": true,
+ "duration": 1522,
+ "category": "basic"
+ },
+ {
+ "test": "addition",
+ "prompt": "add 5 and 3. Return only the number, no explanation.",
+ "result": [
+ "8"
+ ],
+ "expected": "8",
+ "model": "openai/gpt-4o-mini",
+ "router": "openai/gpt-4o-mini",
+ "timestamp": "2025-06-05T22:54:49.606Z",
+ "passed": true,
+ "duration": 634,
+ "category": "basic"
+ },
+ {
+ "test": "addition",
+ "prompt": "add 5 and 3. Return only the number, no explanation.",
+ "result": [
+ "8"
+ ],
+ "expected": "8",
+ "model": "deepseek/deepseek-r1:free",
+ "router": "deepseek/deepseek-r1:free",
+ "timestamp": "2025-06-05T22:54:53.004Z",
+ "passed": true,
+ "duration": 3394,
+ "category": "basic"
+ },
+ {
+ "test": "multiplication",
+ "prompt": "multiply 8 and 3. Return only the number, no explanation.",
+ "result": [
+ "24"
+ ],
+ "expected": "24",
+ "model": "anthropic/claude-sonnet-4",
+ "router": "anthropic/claude-sonnet-4",
+ "timestamp": "2025-06-05T22:54:53.710Z",
+ "passed": true,
+ "duration": 702,
+ "category": "basic"
+ },
+ {
+ "test": "multiplication",
+ "prompt": "multiply 8 and 3. Return only the number, no explanation.",
+ "result": [
+ "24"
+ ],
+ "expected": "24",
+ "model": "openai/gpt-4o-mini",
+ "router": "openai/gpt-4o-mini",
+ "timestamp": "2025-06-05T22:54:56.480Z",
+ "passed": true,
+ "duration": 2765,
+ "category": "basic"
+ },
+ {
+ "test": "multiplication",
+ "prompt": "multiply 8 and 3. Return only the number, no explanation.",
+ "result": [
+ "The result of multiplying 8 and 3 is \\boxed{24}."
+ ],
+ "expected": "24",
+ "model": "deepseek/deepseek-r1:free",
+ "router": "deepseek/deepseek-r1:free",
+ "timestamp": "2025-06-05T22:54:59.909Z",
+ "passed": false,
+ "duration": 3425,
+ "reason": "Expected 24, but got The result of multiplying 8 and 3 is \\boxed{24}.",
+ "category": "basic"
+ },
+ {
+ "test": "division",
+ "prompt": "divide 15 by 3. Return only the number, no explanation.",
+ "result": [
+ "5"
+ ],
+ "expected": "5",
+ "model": "anthropic/claude-sonnet-4",
+ "router": "anthropic/claude-sonnet-4",
+ "timestamp": "2025-06-05T22:55:01.169Z",
+ "passed": true,
+ "duration": 1252,
+ "category": "basic"
+ },
+ {
+ "test": "division",
+ "prompt": "divide 15 by 3. Return only the number, no explanation.",
+ "result": [
+ "5"
+ ],
+ "expected": "5",
+ "model": "openai/gpt-4o-mini",
+ "router": "openai/gpt-4o-mini",
+ "timestamp": "2025-06-05T22:55:01.737Z",
+ "passed": true,
+ "duration": 564,
+ "category": "basic"
+ },
+ {
+ "test": "division",
+ "prompt": "divide 15 by 3. Return only the number, no explanation.",
+ "result": [
+ "5"
+ ],
+ "expected": "5",
+ "model": "deepseek/deepseek-r1:free",
+ "router": "deepseek/deepseek-r1:free",
+ "timestamp": "2025-06-05T22:55:06.362Z",
+ "passed": true,
+ "duration": 4619,
+ "category": "basic"
+ },
+ {
+ "test": "web_content",
+ "prompt": "Check if the content contains a section about Human prehistory. Reply with \"yes\" if it does, \"no\" if it does not.",
+ "result": [
+ "Looking through the table of contents in the Wikipedia article on Kenya, I can see that there is indeed a section titled \"Prehistory\" under the History section.\n\nyes"
+ ],
+ "expected": "yes",
+ "model": "anthropic/claude-sonnet-4",
+ "router": "anthropic/claude-sonnet-4",
+ "timestamp": "2025-06-05T22:55:12.528Z",
+ "passed": false,
+ "duration": 6161,
+ "reason": "Expected yes, but got Looking through the table of contents in the Wikipedia article on Kenya, I can see that there is indeed a section titled \"Prehistory\" under the History section.\n\nyes",
+ "category": "basic"
+ },
+ {
+ "test": "web_content",
+ "prompt": "Check if the content contains a section about Human prehistory. Reply with \"yes\" if it does, \"no\" if it does not.",
+ "result": [
+ "yes"
+ ],
+ "expected": "yes",
+ "model": "openai/gpt-4o-mini",
+ "router": "openai/gpt-4o-mini",
+ "timestamp": "2025-06-05T22:55:18.757Z",
+ "passed": true,
+ "duration": 6225,
+ "category": "basic"
+ },
+ {
+ "test": "web_content",
+ "prompt": "Check if the content contains a section about Human prehistory. Reply with \"yes\" if it does, \"no\" if it does not.",
+ "result": [],
+ "expected": "yes",
+ "model": "deepseek/deepseek-r1:free",
+ "router": "deepseek/deepseek-r1:free",
+ "timestamp": "2025-06-05T22:55:25.642Z",
+ "passed": false,
+ "duration": 6879,
+ "reason": "Model returned empty response",
+ "category": "basic"
+ },
+ {
+ "test": "file-inclusion",
+ "prompt": "What animals are shown in these images?",
+ "result": [
+ "{\"animals\":[\"cat\",\"fox\"]}"
+ ],
+ "expected": "[\"cat\",\"fox\"]",
+ "model": "openai/gpt-4o",
+ "router": "openai/gpt-4o",
+ "timestamp": "2025-06-05T22:55:45.624Z",
+ "passed": true,
+ "duration": 5694,
+ "category": "files"
+ },
+ {
+ "test": "file-inclusion",
+ "prompt": "What is the name of the algorithm implemented in these files? Return only the name.",
+ "result": [
+ "bubbleSort"
+ ],
+ "expected": "bubble sort",
+ "model": "openai/gpt-4o",
+ "router": "openai/gpt-4o",
+ "timestamp": "2025-06-05T22:55:46.325Z",
+ "passed": false,
+ "duration": 694,
+ "reason": "Expected bubble sort, but got bubbleSort",
+ "category": "files"
+ },
+ {
+ "test": "file-inclusion",
+ "prompt": "List all algorithms implemented in these files, as JSON array.",
+ "result": [
+ "{\"algorithms\":[\"factorial\",\"bubbleSort\"]}"
+ ],
+ "expected": "[\"bubble sort\",\"factorial\"]",
+ "model": "openai/gpt-4o",
+ "router": "openai/gpt-4o",
+ "timestamp": "2025-06-05T22:55:47.088Z",
+ "passed": true,
+ "duration": 758,
+ "category": "files"
+ },
+ {
+ "test": "file-inclusion",
+ "prompt": "What is the title of the product in data.json? Return only the title.",
+ "result": [
+ "Injection Barrel"
+ ],
+ "expected": "Injection Barrel",
+ "model": "openai/gpt-4o",
+ "router": "openai/gpt-4o",
+ "timestamp": "2025-06-05T22:55:47.875Z",
+ "passed": true,
+ "duration": 782,
+ "category": "files"
+ },
+ {
+ "test": "summarization",
+ "prompt": "Summarize: \"The quick brown fox jumps over the dog\". Return only the summary, compact, no explanation.",
+ "result": [
+ "Fox jumps over dog."
+ ],
+ "expected": "A fox jumps over a dog",
+ "model": "anthropic/claude-sonnet-4",
+ "router": "anthropic/claude-sonnet-4",
+ "timestamp": "2025-06-05T22:55:58.691Z",
+ "passed": true,
+ "duration": 1621,
+ "category": "language"
+ },
+ {
+ "test": "summarization",
+ "prompt": "Summarize: \"The quick brown fox jumps over the dog\". Return only the summary, compact, no explanation.",
+ "result": [
+ "A fox jumps over a dog."
+ ],
+ "expected": "A fox jumps over a dog",
+ "model": "openai/gpt-4o-mini",
+ "router": "openai/gpt-4o-mini",
+ "timestamp": "2025-06-05T22:56:00.205Z",
+ "passed": true,
+ "duration": 1508,
+ "category": "language"
+ },
+ {
+ "test": "summarization",
+ "prompt": "Summarize: \"The quick brown fox jumps over the dog\". Return only the summary, compact, no explanation.",
+ "result": [
+ "A quick brown fox leaps over a dog."
+ ],
+ "expected": "A fox jumps over a dog",
+ "model": "deepseek/deepseek-r1:free",
+ "router": "deepseek/deepseek-r1:free",
+ "timestamp": "2025-06-05T22:56:06.210Z",
+ "passed": true,
+ "duration": 6000,
+ "category": "language"
+ },
+ {
+ "test": "translation",
+ "prompt": "Translate \"Hello, world!\" to Spanish. Return only the translation, no explanation.",
+ "result": [
+ "¡Hola, mundo!"
+ ],
+ "expected": "¡Hola, mundo!",
+ "model": "anthropic/claude-sonnet-4",
+ "router": "anthropic/claude-sonnet-4",
+ "timestamp": "2025-06-05T22:56:07.465Z",
+ "passed": true,
+ "duration": 1250,
+ "category": "language"
+ },
+ {
+ "test": "translation",
+ "prompt": "Translate \"Hello, world!\" to Spanish. Return only the translation, no explanation.",
+ "result": [
+ "¡Hola, mundo!"
+ ],
+ "expected": "¡Hola, mundo!",
+ "model": "openai/gpt-4o-mini",
+ "router": "openai/gpt-4o-mini",
+ "timestamp": "2025-06-05T22:56:07.992Z",
+ "passed": true,
+ "duration": 521,
+ "category": "language"
+ },
+ {
+ "test": "translation",
+ "prompt": "Translate \"Hello, world!\" to Spanish. Return only the translation, no explanation.",
+ "result": [
+ "¡Hola, mundo!"
+ ],
+ "expected": "¡Hola, mundo!",
+ "model": "deepseek/deepseek-r1:free",
+ "router": "deepseek/deepseek-r1:free",
+ "timestamp": "2025-06-05T22:56:10.966Z",
+ "passed": true,
+ "duration": 2969,
+ "category": "language"
+ },
+ {
+ "test": "grammar",
+ "prompt": "Correct the grammar in: \"I goes to the store yesterday\". Return only the corrected sentence, no explanation.",
+ "result": [
+ "I went to the store yesterday."
+ ],
+ "expected": "I went to the store yesterday",
+ "model": "anthropic/claude-sonnet-4",
+ "router": "anthropic/claude-sonnet-4",
+ "timestamp": "2025-06-05T22:56:12.216Z",
+ "passed": false,
+ "duration": 1246,
+ "reason": "Expected I went to the store yesterday, but got I went to the store yesterday.",
+ "category": "language"
+ },
+ {
+ "test": "grammar",
+ "prompt": "Correct the grammar in: \"I goes to the store yesterday\". Return only the corrected sentence, no explanation.",
+ "result": [
+ "\"I went to the store yesterday.\""
+ ],
+ "expected": "I went to the store yesterday",
+ "model": "openai/gpt-4o-mini",
+ "router": "openai/gpt-4o-mini",
+ "timestamp": "2025-06-05T22:56:12.844Z",
+ "passed": false,
+ "duration": 623,
+ "reason": "Expected I went to the store yesterday, but got \"I went to the store yesterday.\"",
+ "category": "language"
+ },
+ {
+ "test": "grammar",
+ "prompt": "Correct the grammar in: \"I goes to the store yesterday\". Return only the corrected sentence, no explanation.",
+ "result": [
+ "I went to the store yesterday."
+ ],
+ "expected": "I went to the store yesterday",
+ "model": "deepseek/deepseek-r1:free",
+ "router": "deepseek/deepseek-r1:free",
+ "timestamp": "2025-06-05T22:56:17.444Z",
+ "passed": false,
+ "duration": 4594,
+ "reason": "Expected I went to the store yesterday, but got I went to the store yesterday.",
+ "category": "language"
+ },
+ {
+ "test": "language_detection",
+ "prompt": "Identify the language of: \"Bonjour, comment allez-vous?\". Return only the language name, no explanation.",
+ "result": [
+ "French"
+ ],
+ "expected": "French",
+ "model": "anthropic/claude-sonnet-4",
+ "router": "anthropic/claude-sonnet-4",
+ "timestamp": "2025-06-05T22:56:18.557Z",
+ "passed": true,
+ "duration": 1107,
+ "category": "language"
+ },
+ {
+ "test": "language_detection",
+ "prompt": "Identify the language of: \"Bonjour, comment allez-vous?\". Return only the language name, no explanation.",
+ "result": [
+ "French"
+ ],
+ "expected": "French",
+ "model": "openai/gpt-4o-mini",
+ "router": "openai/gpt-4o-mini",
+ "timestamp": "2025-06-05T22:56:19.252Z",
+ "passed": true,
+ "duration": 689,
+ "category": "language"
+ },
+ {
+ "test": "language_detection",
+ "prompt": "Identify the language of: \"Bonjour, comment allez-vous?\". Return only the language name, no explanation.",
+ "result": [
+ "French"
+ ],
+ "expected": "French",
+ "model": "deepseek/deepseek-r1:free",
+ "router": "deepseek/deepseek-r1:free",
+ "timestamp": "2025-06-05T22:56:23.475Z",
+ "passed": true,
+ "duration": 4218,
+ "category": "language"
+ },
+ {
+ "test": "synonyms",
+ "prompt": "Provide a synonym for \"happy\". Return only the synonym, no explanation.",
+ "result": [
+ "Joyful"
+ ],
+ "expected": "joyful",
+ "model": "anthropic/claude-sonnet-4",
+ "router": "anthropic/claude-sonnet-4",
+ "timestamp": "2025-06-05T22:56:25.120Z",
+ "passed": true,
+ "duration": 1639,
+ "category": "language"
+ },
+ {
+ "test": "synonyms",
+ "prompt": "Provide a synonym for \"happy\". Return only the synonym, no explanation.",
+ "result": [
+ "Joyful"
+ ],
+ "expected": "joyful",
+ "model": "openai/gpt-4o-mini",
+ "router": "openai/gpt-4o-mini",
+ "timestamp": "2025-06-05T22:56:25.785Z",
+ "passed": true,
+ "duration": 661,
+ "category": "language"
+ },
+ {
+ "test": "synonyms",
+ "prompt": "Provide a synonym for \"happy\". Return only the synonym, no explanation.",
+ "result": [
+ "Joyful"
+ ],
+ "expected": "joyful",
+ "model": "deepseek/deepseek-r1:free",
+ "router": "deepseek/deepseek-r1:free",
+ "timestamp": "2025-06-05T22:56:29.261Z",
+ "passed": true,
+ "duration": 3471,
+ "category": "language"
}
],
"highscores": [
@@ -5224,8 +5691,8 @@
"rankings": [
{
"model": "openai/gpt-4o-mini",
- "duration": 639,
- "duration_secs": 0.639
+ "duration": 521,
+ "duration_secs": 0.521
},
{
"model": "openai/gpt-3.5-turbo",
@@ -5244,8 +5711,8 @@
},
{
"model": "openai/gpt-4o-mini",
- "duration": 641,
- "duration_secs": 0.641
+ "duration": 623,
+ "duration_secs": 0.623
}
]
},
@@ -5259,8 +5726,8 @@
},
{
"model": "openai/gpt-4o-mini",
- "duration": 8548,
- "duration_secs": 8.548
+ "duration": 1508,
+ "duration_secs": 1.508
}
]
},
@@ -5269,8 +5736,8 @@
"rankings": [
{
"model": "openai/gpt-4o-mini",
- "duration": 657,
- "duration_secs": 0.657
+ "duration": 689,
+ "duration_secs": 0.689
},
{
"model": "openai/gpt-3.5-turbo",
@@ -5289,8 +5756,8 @@
},
{
"model": "openai/gpt-4o-mini",
- "duration": 622,
- "duration_secs": 0.622
+ "duration": 661,
+ "duration_secs": 0.661
}
]
},
@@ -5299,8 +5766,8 @@
"rankings": [
{
"model": "openai/gpt-4o",
- "duration": 614,
- "duration_secs": 0.614
+ "duration": 782,
+ "duration_secs": 0.782
},
{
"model": "google/gemini-2.0-flash-exp:free",
@@ -5339,8 +5806,8 @@
"rankings": [
{
"model": "openai/gpt-4o-mini",
- "duration": 641,
- "duration_secs": 0.641
+ "duration": 634,
+ "duration_secs": 0.634
},
{
"model": "openai/gpt-3.5-turbo",
@@ -5362,15 +5829,15 @@
{
"test": "multiplication",
"rankings": [
- {
- "model": "openai/gpt-4o-mini",
- "duration": 585,
- "duration_secs": 0.585
- },
{
"model": "openai/gpt-3.5-turbo",
"duration": 624,
"duration_secs": 0.624
+ },
+ {
+ "model": "anthropic/claude-sonnet-4",
+ "duration": 702,
+ "duration_secs": 0.702
}
]
},
@@ -5402,15 +5869,15 @@
{
"test": "division",
"rankings": [
- {
- "model": "openai/gpt-4o-mini",
- "duration": 481,
- "duration_secs": 0.481
- },
{
"model": "openai/gpt-3.5-turbo",
"duration": 513,
"duration_secs": 0.513
+ },
+ {
+ "model": "openai/gpt-4o-mini",
+ "duration": 564,
+ "duration_secs": 0.564
}
]
},
@@ -5438,9 +5905,9 @@
"duration_secs": 0.22
},
{
- "model": "openai/gpt-4o-mini",
- "duration": 4917,
- "duration_secs": 4.917
+ "model": "anthropic/claude-sonnet-4",
+ "duration": 6161,
+ "duration_secs": 6.161
}
]
},
@@ -5495,5 +5962,5 @@
]
}
],
- "lastUpdated": "2025-06-05T22:53:48.765Z"
+ "lastUpdated": "2025-06-05T22:56:29.265Z"
}
\ No newline at end of file
diff --git a/packages/kbot/tests/unit/reports/basic.json b/packages/kbot/tests/unit/reports/basic.json
index 24c21875..b7f34fbc 100644
--- a/packages/kbot/tests/unit/reports/basic.json
+++ b/packages/kbot/tests/unit/reports/basic.json
@@ -347,6 +347,175 @@
"duration": 5147,
"reason": "Model returned empty response",
"category": "basic"
+ },
+ {
+ "test": "addition",
+ "prompt": "add 5 and 3. Return only the number, no explanation.",
+ "result": [
+ "8"
+ ],
+ "expected": "8",
+ "model": "anthropic/claude-sonnet-4",
+ "router": "anthropic/claude-sonnet-4",
+ "timestamp": "2025-06-05T22:54:48.966Z",
+ "passed": true,
+ "duration": 1522,
+ "category": "basic"
+ },
+ {
+ "test": "addition",
+ "prompt": "add 5 and 3. Return only the number, no explanation.",
+ "result": [
+ "8"
+ ],
+ "expected": "8",
+ "model": "openai/gpt-4o-mini",
+ "router": "openai/gpt-4o-mini",
+ "timestamp": "2025-06-05T22:54:49.606Z",
+ "passed": true,
+ "duration": 634,
+ "category": "basic"
+ },
+ {
+ "test": "addition",
+ "prompt": "add 5 and 3. Return only the number, no explanation.",
+ "result": [
+ "8"
+ ],
+ "expected": "8",
+ "model": "deepseek/deepseek-r1:free",
+ "router": "deepseek/deepseek-r1:free",
+ "timestamp": "2025-06-05T22:54:53.004Z",
+ "passed": true,
+ "duration": 3394,
+ "category": "basic"
+ },
+ {
+ "test": "multiplication",
+ "prompt": "multiply 8 and 3. Return only the number, no explanation.",
+ "result": [
+ "24"
+ ],
+ "expected": "24",
+ "model": "anthropic/claude-sonnet-4",
+ "router": "anthropic/claude-sonnet-4",
+ "timestamp": "2025-06-05T22:54:53.710Z",
+ "passed": true,
+ "duration": 702,
+ "category": "basic"
+ },
+ {
+ "test": "multiplication",
+ "prompt": "multiply 8 and 3. Return only the number, no explanation.",
+ "result": [
+ "24"
+ ],
+ "expected": "24",
+ "model": "openai/gpt-4o-mini",
+ "router": "openai/gpt-4o-mini",
+ "timestamp": "2025-06-05T22:54:56.480Z",
+ "passed": true,
+ "duration": 2765,
+ "category": "basic"
+ },
+ {
+ "test": "multiplication",
+ "prompt": "multiply 8 and 3. Return only the number, no explanation.",
+ "result": [
+ "The result of multiplying 8 and 3 is \\boxed{24}."
+ ],
+ "expected": "24",
+ "model": "deepseek/deepseek-r1:free",
+ "router": "deepseek/deepseek-r1:free",
+ "timestamp": "2025-06-05T22:54:59.909Z",
+ "passed": false,
+ "duration": 3425,
+ "reason": "Expected 24, but got The result of multiplying 8 and 3 is \\boxed{24}.",
+ "category": "basic"
+ },
+ {
+ "test": "division",
+ "prompt": "divide 15 by 3. Return only the number, no explanation.",
+ "result": [
+ "5"
+ ],
+ "expected": "5",
+ "model": "anthropic/claude-sonnet-4",
+ "router": "anthropic/claude-sonnet-4",
+ "timestamp": "2025-06-05T22:55:01.169Z",
+ "passed": true,
+ "duration": 1252,
+ "category": "basic"
+ },
+ {
+ "test": "division",
+ "prompt": "divide 15 by 3. Return only the number, no explanation.",
+ "result": [
+ "5"
+ ],
+ "expected": "5",
+ "model": "openai/gpt-4o-mini",
+ "router": "openai/gpt-4o-mini",
+ "timestamp": "2025-06-05T22:55:01.737Z",
+ "passed": true,
+ "duration": 564,
+ "category": "basic"
+ },
+ {
+ "test": "division",
+ "prompt": "divide 15 by 3. Return only the number, no explanation.",
+ "result": [
+ "5"
+ ],
+ "expected": "5",
+ "model": "deepseek/deepseek-r1:free",
+ "router": "deepseek/deepseek-r1:free",
+ "timestamp": "2025-06-05T22:55:06.362Z",
+ "passed": true,
+ "duration": 4619,
+ "category": "basic"
+ },
+ {
+ "test": "web_content",
+ "prompt": "Check if the content contains a section about Human prehistory. Reply with \"yes\" if it does, \"no\" if it does not.",
+ "result": [
+ "Looking through the table of contents in the Wikipedia article on Kenya, I can see that there is indeed a section titled \"Prehistory\" under the History section.\n\nyes"
+ ],
+ "expected": "yes",
+ "model": "anthropic/claude-sonnet-4",
+ "router": "anthropic/claude-sonnet-4",
+ "timestamp": "2025-06-05T22:55:12.528Z",
+ "passed": false,
+ "duration": 6161,
+ "reason": "Expected yes, but got Looking through the table of contents in the Wikipedia article on Kenya, I can see that there is indeed a section titled \"Prehistory\" under the History section.\n\nyes",
+ "category": "basic"
+ },
+ {
+ "test": "web_content",
+ "prompt": "Check if the content contains a section about Human prehistory. Reply with \"yes\" if it does, \"no\" if it does not.",
+ "result": [
+ "yes"
+ ],
+ "expected": "yes",
+ "model": "openai/gpt-4o-mini",
+ "router": "openai/gpt-4o-mini",
+ "timestamp": "2025-06-05T22:55:18.757Z",
+ "passed": true,
+ "duration": 6225,
+ "category": "basic"
+ },
+ {
+ "test": "web_content",
+ "prompt": "Check if the content contains a section about Human prehistory. Reply with \"yes\" if it does, \"no\" if it does not.",
+ "result": [],
+ "expected": "yes",
+ "model": "deepseek/deepseek-r1:free",
+ "router": "deepseek/deepseek-r1:free",
+ "timestamp": "2025-06-05T22:55:25.642Z",
+ "passed": false,
+ "duration": 6879,
+ "reason": "Model returned empty response",
+ "category": "basic"
}
],
"highscores": [
@@ -355,8 +524,8 @@
"rankings": [
{
"model": "openai/gpt-4o-mini",
- "duration": 641,
- "duration_secs": 0.641
+ "duration": 634,
+ "duration_secs": 0.634
},
{
"model": "openai/gpt-3.5-turbo",
@@ -368,30 +537,30 @@
{
"test": "multiplication",
"rankings": [
- {
- "model": "openai/gpt-4o-mini",
- "duration": 585,
- "duration_secs": 0.585
- },
{
"model": "openai/gpt-3.5-turbo",
"duration": 624,
"duration_secs": 0.624
+ },
+ {
+ "model": "anthropic/claude-sonnet-4",
+ "duration": 702,
+ "duration_secs": 0.702
}
]
},
{
"test": "division",
"rankings": [
- {
- "model": "openai/gpt-4o-mini",
- "duration": 481,
- "duration_secs": 0.481
- },
{
"model": "openai/gpt-3.5-turbo",
"duration": 513,
"duration_secs": 0.513
+ },
+ {
+ "model": "openai/gpt-4o-mini",
+ "duration": 564,
+ "duration_secs": 0.564
}
]
},
@@ -404,12 +573,12 @@
"duration_secs": 0.22
},
{
- "model": "openai/gpt-4o-mini",
- "duration": 4917,
- "duration_secs": 4.917
+ "model": "anthropic/claude-sonnet-4",
+ "duration": 6161,
+ "duration_secs": 6.161
}
]
}
],
- "lastUpdated": "2025-06-05T21:19:47.545Z"
+ "lastUpdated": "2025-06-05T22:55:25.642Z"
}
\ No newline at end of file
diff --git a/packages/kbot/tests/unit/reports/basic.md b/packages/kbot/tests/unit/reports/basic.md
index 25f4de44..3ded09e2 100644
--- a/packages/kbot/tests/unit/reports/basic.md
+++ b/packages/kbot/tests/unit/reports/basic.md
@@ -6,37 +6,59 @@
| Test | Model | Duration (ms) | Duration (s) |
|------|-------|--------------|--------------|
-| addition | openai/gpt-4o-mini | 641 | 0.64 |
-| addition | anthropic/claude-sonnet-4 | 2024 | 2.02 |
-| addition | deepseek/deepseek-r1:free | 3798 | 3.80 |
-| multiplication | openai/gpt-4o-mini | 585 | 0.58 |
-| multiplication | anthropic/claude-sonnet-4 | 1752 | 1.75 |
-| multiplication | deepseek/deepseek-r1:free | 3272 | 3.27 |
-| division | openai/gpt-4o-mini | 481 | 0.48 |
-| division | anthropic/claude-sonnet-4 | 1775 | 1.77 |
-| division | deepseek/deepseek-r1:free | 3406 | 3.41 |
-| web_content | openai/gpt-4o-mini | 4917 | 4.92 |
-| web_content | deepseek/deepseek-r1:free | 5147 | 5.15 |
-| web_content | anthropic/claude-sonnet-4 | 6020 | 6.02 |
+| addition | openai/gpt-4o-mini | 634 | 0.63 |
+| addition | anthropic/claude-sonnet-4 | 1522 | 1.52 |
+| addition | deepseek/deepseek-r1:free | 3394 | 3.39 |
+| multiplication | anthropic/claude-sonnet-4 | 702 | 0.70 |
+| multiplication | openai/gpt-4o-mini | 2765 | 2.77 |
+| multiplication | deepseek/deepseek-r1:free | 3425 | 3.42 |
+| division | openai/gpt-4o-mini | 564 | 0.56 |
+| division | anthropic/claude-sonnet-4 | 1252 | 1.25 |
+| division | deepseek/deepseek-r1:free | 4619 | 4.62 |
+| web_content | anthropic/claude-sonnet-4 | 6161 | 6.16 |
+| web_content | openai/gpt-4o-mini | 6225 | 6.22 |
+| web_content | deepseek/deepseek-r1:free | 6879 | 6.88 |
## Summary
- Total Tests: 12
-- Passed: 11
-- Failed: 1
-- Success Rate: 91.67%
-- Average Duration: 2818ms (2.82s)
+- Passed: 9
+- Failed: 3
+- Success Rate: 75.00%
+- Average Duration: 3179ms (3.18s)
## Failed Tests
+### multiplication - deepseek/deepseek-r1:free
+
+- Prompt: `multiply 8 and 3. Return only the number, no explanation.`
+- Expected: `24`
+- Actual: `The result of multiplying 8 and 3 is \boxed{24}.`
+- Duration: 3425ms (3.42s)
+- Reason: Expected 24, but got The result of multiplying 8 and 3 is \boxed{24}.
+- Timestamp: 6/6/2025, 12:54:59 AM
+
+### web_content - anthropic/claude-sonnet-4
+
+- Prompt: `Check if the content contains a section about Human prehistory. Reply with "yes" if it does, "no" if it does not.`
+- Expected: `yes`
+- Actual: `Looking through the table of contents in the Wikipedia article on Kenya, I can see that there is indeed a section titled "Prehistory" under the History section.
+
+yes`
+- Duration: 6161ms (6.16s)
+- Reason: Expected yes, but got Looking through the table of contents in the Wikipedia article on Kenya, I can see that there is indeed a section titled "Prehistory" under the History section.
+
+yes
+- Timestamp: 6/6/2025, 12:55:12 AM
+
### web_content - deepseek/deepseek-r1:free
- Prompt: `Check if the content contains a section about Human prehistory. Reply with "yes" if it does, "no" if it does not.`
- Expected: `yes`
- Actual: ``
-- Duration: 5147ms (5.15s)
+- Duration: 6879ms (6.88s)
- Reason: Model returned empty response
-- Timestamp: 6/5/2025, 11:19:47 PM
+- Timestamp: 6/6/2025, 12:55:25 AM
## Passed Tests
@@ -45,86 +67,70 @@
- Prompt: `add 5 and 3. Return only the number, no explanation.`
- Expected: `8`
- Actual: `8`
-- Duration: 2024ms (2.02s)
-- Timestamp: 6/5/2025, 11:19:15 PM
+- Duration: 1522ms (1.52s)
+- Timestamp: 6/6/2025, 12:54:48 AM
### addition - openai/gpt-4o-mini
- Prompt: `add 5 and 3. Return only the number, no explanation.`
- Expected: `8`
- Actual: `8`
-- Duration: 641ms (0.64s)
-- Timestamp: 6/5/2025, 11:19:16 PM
+- Duration: 634ms (0.63s)
+- Timestamp: 6/6/2025, 12:54:49 AM
### addition - deepseek/deepseek-r1:free
- Prompt: `add 5 and 3. Return only the number, no explanation.`
- Expected: `8`
- Actual: `8`
-- Duration: 3798ms (3.80s)
-- Timestamp: 6/5/2025, 11:19:20 PM
+- Duration: 3394ms (3.39s)
+- Timestamp: 6/6/2025, 12:54:53 AM
### multiplication - anthropic/claude-sonnet-4
- Prompt: `multiply 8 and 3. Return only the number, no explanation.`
- Expected: `24`
- Actual: `24`
-- Duration: 1752ms (1.75s)
-- Timestamp: 6/5/2025, 11:19:21 PM
+- Duration: 702ms (0.70s)
+- Timestamp: 6/6/2025, 12:54:53 AM
### multiplication - openai/gpt-4o-mini
- Prompt: `multiply 8 and 3. Return only the number, no explanation.`
- Expected: `24`
- Actual: `24`
-- Duration: 585ms (0.58s)
-- Timestamp: 6/5/2025, 11:19:22 PM
-
-### multiplication - deepseek/deepseek-r1:free
-
-- Prompt: `multiply 8 and 3. Return only the number, no explanation.`
-- Expected: `24`
-- Actual: `24`
-- Duration: 3272ms (3.27s)
-- Timestamp: 6/5/2025, 11:19:25 PM
+- Duration: 2765ms (2.77s)
+- Timestamp: 6/6/2025, 12:54:56 AM
### division - anthropic/claude-sonnet-4
- Prompt: `divide 15 by 3. Return only the number, no explanation.`
- Expected: `5`
- Actual: `5`
-- Duration: 1775ms (1.77s)
-- Timestamp: 6/5/2025, 11:19:27 PM
+- Duration: 1252ms (1.25s)
+- Timestamp: 6/6/2025, 12:55:01 AM
### division - openai/gpt-4o-mini
- Prompt: `divide 15 by 3. Return only the number, no explanation.`
- Expected: `5`
- Actual: `5`
-- Duration: 481ms (0.48s)
-- Timestamp: 6/5/2025, 11:19:28 PM
+- Duration: 564ms (0.56s)
+- Timestamp: 6/6/2025, 12:55:01 AM
### division - deepseek/deepseek-r1:free
- Prompt: `divide 15 by 3. Return only the number, no explanation.`
- Expected: `5`
- Actual: `5`
-- Duration: 3406ms (3.41s)
-- Timestamp: 6/5/2025, 11:19:31 PM
-
-### web_content - anthropic/claude-sonnet-4
-
-- Prompt: `Check if the content contains a section about Human prehistory. Reply with "yes" if it does, "no" if it does not.`
-- Expected: `yes`
-- Actual: `yes`
-- Duration: 6020ms (6.02s)
-- Timestamp: 6/5/2025, 11:19:37 PM
+- Duration: 4619ms (4.62s)
+- Timestamp: 6/6/2025, 12:55:06 AM
### web_content - openai/gpt-4o-mini
- Prompt: `Check if the content contains a section about Human prehistory. Reply with "yes" if it does, "no" if it does not.`
- Expected: `yes`
- Actual: `yes`
-- Duration: 4917ms (4.92s)
-- Timestamp: 6/5/2025, 11:19:42 PM
+- Duration: 6225ms (6.22s)
+- Timestamp: 6/6/2025, 12:55:18 AM
diff --git a/packages/kbot/tests/unit/reports/files.json b/packages/kbot/tests/unit/reports/files.json
index 8bbeeef2..54f91833 100644
--- a/packages/kbot/tests/unit/reports/files.json
+++ b/packages/kbot/tests/unit/reports/files.json
@@ -1996,6 +1996,63 @@
"passed": true,
"duration": 614,
"category": "files"
+ },
+ {
+ "test": "file-inclusion",
+ "prompt": "What animals are shown in these images?",
+ "result": [
+ "{\"animals\":[\"cat\",\"fox\"]}"
+ ],
+ "expected": "[\"cat\",\"fox\"]",
+ "model": "openai/gpt-4o",
+ "router": "openai/gpt-4o",
+ "timestamp": "2025-06-05T22:55:45.624Z",
+ "passed": true,
+ "duration": 5694,
+ "category": "files"
+ },
+ {
+ "test": "file-inclusion",
+ "prompt": "What is the name of the algorithm implemented in these files? Return only the name.",
+ "result": [
+ "bubbleSort"
+ ],
+ "expected": "bubble sort",
+ "model": "openai/gpt-4o",
+ "router": "openai/gpt-4o",
+ "timestamp": "2025-06-05T22:55:46.325Z",
+ "passed": false,
+ "duration": 694,
+ "reason": "Expected bubble sort, but got bubbleSort",
+ "category": "files"
+ },
+ {
+ "test": "file-inclusion",
+ "prompt": "List all algorithms implemented in these files, as JSON array.",
+ "result": [
+ "{\"algorithms\":[\"factorial\",\"bubbleSort\"]}"
+ ],
+ "expected": "[\"bubble sort\",\"factorial\"]",
+ "model": "openai/gpt-4o",
+ "router": "openai/gpt-4o",
+ "timestamp": "2025-06-05T22:55:47.088Z",
+ "passed": true,
+ "duration": 758,
+ "category": "files"
+ },
+ {
+ "test": "file-inclusion",
+ "prompt": "What is the title of the product in data.json? Return only the title.",
+ "result": [
+ "Injection Barrel"
+ ],
+ "expected": "Injection Barrel",
+ "model": "openai/gpt-4o",
+ "router": "openai/gpt-4o",
+ "timestamp": "2025-06-05T22:55:47.875Z",
+ "passed": true,
+ "duration": 782,
+ "category": "files"
}
],
"highscores": [
@@ -2004,8 +2061,8 @@
"rankings": [
{
"model": "openai/gpt-4o",
- "duration": 614,
- "duration_secs": 0.614
+ "duration": 782,
+ "duration_secs": 0.782
},
{
"model": "google/gemini-2.0-flash-exp:free",
@@ -2015,5 +2072,5 @@
]
}
],
- "lastUpdated": "2025-06-05T22:29:46.852Z"
+ "lastUpdated": "2025-06-05T22:55:47.876Z"
}
\ No newline at end of file
diff --git a/packages/kbot/tests/unit/reports/files.md b/packages/kbot/tests/unit/reports/files.md
index cee52d90..4632d850 100644
--- a/packages/kbot/tests/unit/reports/files.md
+++ b/packages/kbot/tests/unit/reports/files.md
@@ -6,7 +6,7 @@
| Test | Model | Duration (ms) | Duration (s) |
|------|-------|--------------|--------------|
-| file-inclusion | openai/gpt-4o | 614 | 0.61 |
+| file-inclusion | openai/gpt-4o | 782 | 0.78 |
## Summary
@@ -14,7 +14,7 @@
- Passed: 3
- Failed: 1
- Success Rate: 75.00%
-- Average Duration: 1380ms (1.38s)
+- Average Duration: 1982ms (1.98s)
## Failed Tests
@@ -27,6 +27,6 @@
- Prompt: `What is the title of the product in data.json? Return only the title.`
- Expected: `Injection Barrel`
- Actual: `Injection Barrel`
-- Duration: 614ms (0.61s)
-- Timestamp: 6/6/2025, 12:29:46 AM
+- Duration: 782ms (0.78s)
+- Timestamp: 6/6/2025, 12:55:47 AM
diff --git a/packages/kbot/tests/unit/reports/language.json b/packages/kbot/tests/unit/reports/language.json
index 4febcef2..b60f4f8c 100644
--- a/packages/kbot/tests/unit/reports/language.json
+++ b/packages/kbot/tests/unit/reports/language.json
@@ -1991,6 +1991,247 @@
"passed": true,
"duration": 1336,
"category": "language"
+ },
+ {
+ "test": "synonyms",
+ "prompt": "Provide a synonym for \"happy\". Return only the synonym, no explanation.",
+ "result": [
+ "Joyful"
+ ],
+ "expected": "joyful",
+ "model": "openai/gpt-4o-mini",
+ "router": "openai/gpt-4o-mini",
+ "timestamp": "2025-06-05T22:54:17.429Z",
+ "passed": true,
+ "duration": 561,
+ "category": "language"
+ },
+ {
+ "test": "synonyms",
+ "prompt": "Provide a synonym for \"happy\". Return only the synonym, no explanation.",
+ "result": [
+ "Joyful"
+ ],
+ "expected": "joyful",
+ "model": "deepseek/deepseek-r1:free",
+ "router": "deepseek/deepseek-r1:free",
+ "timestamp": "2025-06-05T22:54:21.005Z",
+ "passed": true,
+ "duration": 3571,
+ "category": "language"
+ },
+ {
+ "test": "summarization",
+ "prompt": "Summarize: \"The quick brown fox jumps over the dog\". Return only the summary, compact, no explanation.",
+ "result": [
+ "Fox jumps over dog."
+ ],
+ "expected": "A fox jumps over a dog",
+ "model": "anthropic/claude-sonnet-4",
+ "router": "anthropic/claude-sonnet-4",
+ "timestamp": "2025-06-05T22:55:58.691Z",
+ "passed": true,
+ "duration": 1621,
+ "category": "language"
+ },
+ {
+ "test": "summarization",
+ "prompt": "Summarize: \"The quick brown fox jumps over the dog\". Return only the summary, compact, no explanation.",
+ "result": [
+ "A fox jumps over a dog."
+ ],
+ "expected": "A fox jumps over a dog",
+ "model": "openai/gpt-4o-mini",
+ "router": "openai/gpt-4o-mini",
+ "timestamp": "2025-06-05T22:56:00.205Z",
+ "passed": true,
+ "duration": 1508,
+ "category": "language"
+ },
+ {
+ "test": "summarization",
+ "prompt": "Summarize: \"The quick brown fox jumps over the dog\". Return only the summary, compact, no explanation.",
+ "result": [
+ "A quick brown fox leaps over a dog."
+ ],
+ "expected": "A fox jumps over a dog",
+ "model": "deepseek/deepseek-r1:free",
+ "router": "deepseek/deepseek-r1:free",
+ "timestamp": "2025-06-05T22:56:06.210Z",
+ "passed": true,
+ "duration": 6000,
+ "category": "language"
+ },
+ {
+ "test": "translation",
+ "prompt": "Translate \"Hello, world!\" to Spanish. Return only the translation, no explanation.",
+ "result": [
+ "¡Hola, mundo!"
+ ],
+ "expected": "¡Hola, mundo!",
+ "model": "anthropic/claude-sonnet-4",
+ "router": "anthropic/claude-sonnet-4",
+ "timestamp": "2025-06-05T22:56:07.465Z",
+ "passed": true,
+ "duration": 1250,
+ "category": "language"
+ },
+ {
+ "test": "translation",
+ "prompt": "Translate \"Hello, world!\" to Spanish. Return only the translation, no explanation.",
+ "result": [
+ "¡Hola, mundo!"
+ ],
+ "expected": "¡Hola, mundo!",
+ "model": "openai/gpt-4o-mini",
+ "router": "openai/gpt-4o-mini",
+ "timestamp": "2025-06-05T22:56:07.992Z",
+ "passed": true,
+ "duration": 521,
+ "category": "language"
+ },
+ {
+ "test": "translation",
+ "prompt": "Translate \"Hello, world!\" to Spanish. Return only the translation, no explanation.",
+ "result": [
+ "¡Hola, mundo!"
+ ],
+ "expected": "¡Hola, mundo!",
+ "model": "deepseek/deepseek-r1:free",
+ "router": "deepseek/deepseek-r1:free",
+ "timestamp": "2025-06-05T22:56:10.966Z",
+ "passed": true,
+ "duration": 2969,
+ "category": "language"
+ },
+ {
+ "test": "grammar",
+ "prompt": "Correct the grammar in: \"I goes to the store yesterday\". Return only the corrected sentence, no explanation.",
+ "result": [
+ "I went to the store yesterday."
+ ],
+ "expected": "I went to the store yesterday",
+ "model": "anthropic/claude-sonnet-4",
+ "router": "anthropic/claude-sonnet-4",
+ "timestamp": "2025-06-05T22:56:12.216Z",
+ "passed": false,
+ "duration": 1246,
+ "reason": "Expected I went to the store yesterday, but got I went to the store yesterday.",
+ "category": "language"
+ },
+ {
+ "test": "grammar",
+ "prompt": "Correct the grammar in: \"I goes to the store yesterday\". Return only the corrected sentence, no explanation.",
+ "result": [
+ "\"I went to the store yesterday.\""
+ ],
+ "expected": "I went to the store yesterday",
+ "model": "openai/gpt-4o-mini",
+ "router": "openai/gpt-4o-mini",
+ "timestamp": "2025-06-05T22:56:12.844Z",
+ "passed": false,
+ "duration": 623,
+ "reason": "Expected I went to the store yesterday, but got \"I went to the store yesterday.\"",
+ "category": "language"
+ },
+ {
+ "test": "grammar",
+ "prompt": "Correct the grammar in: \"I goes to the store yesterday\". Return only the corrected sentence, no explanation.",
+ "result": [
+ "I went to the store yesterday."
+ ],
+ "expected": "I went to the store yesterday",
+ "model": "deepseek/deepseek-r1:free",
+ "router": "deepseek/deepseek-r1:free",
+ "timestamp": "2025-06-05T22:56:17.444Z",
+ "passed": false,
+ "duration": 4594,
+ "reason": "Expected I went to the store yesterday, but got I went to the store yesterday.",
+ "category": "language"
+ },
+ {
+ "test": "language_detection",
+ "prompt": "Identify the language of: \"Bonjour, comment allez-vous?\". Return only the language name, no explanation.",
+ "result": [
+ "French"
+ ],
+ "expected": "French",
+ "model": "anthropic/claude-sonnet-4",
+ "router": "anthropic/claude-sonnet-4",
+ "timestamp": "2025-06-05T22:56:18.557Z",
+ "passed": true,
+ "duration": 1107,
+ "category": "language"
+ },
+ {
+ "test": "language_detection",
+ "prompt": "Identify the language of: \"Bonjour, comment allez-vous?\". Return only the language name, no explanation.",
+ "result": [
+ "French"
+ ],
+ "expected": "French",
+ "model": "openai/gpt-4o-mini",
+ "router": "openai/gpt-4o-mini",
+ "timestamp": "2025-06-05T22:56:19.252Z",
+ "passed": true,
+ "duration": 689,
+ "category": "language"
+ },
+ {
+ "test": "language_detection",
+ "prompt": "Identify the language of: \"Bonjour, comment allez-vous?\". Return only the language name, no explanation.",
+ "result": [
+ "French"
+ ],
+ "expected": "French",
+ "model": "deepseek/deepseek-r1:free",
+ "router": "deepseek/deepseek-r1:free",
+ "timestamp": "2025-06-05T22:56:23.475Z",
+ "passed": true,
+ "duration": 4218,
+ "category": "language"
+ },
+ {
+ "test": "synonyms",
+ "prompt": "Provide a synonym for \"happy\". Return only the synonym, no explanation.",
+ "result": [
+ "Joyful"
+ ],
+ "expected": "joyful",
+ "model": "anthropic/claude-sonnet-4",
+ "router": "anthropic/claude-sonnet-4",
+ "timestamp": "2025-06-05T22:56:25.120Z",
+ "passed": true,
+ "duration": 1639,
+ "category": "language"
+ },
+ {
+ "test": "synonyms",
+ "prompt": "Provide a synonym for \"happy\". Return only the synonym, no explanation.",
+ "result": [
+ "Joyful"
+ ],
+ "expected": "joyful",
+ "model": "openai/gpt-4o-mini",
+ "router": "openai/gpt-4o-mini",
+ "timestamp": "2025-06-05T22:56:25.785Z",
+ "passed": true,
+ "duration": 661,
+ "category": "language"
+ },
+ {
+ "test": "synonyms",
+ "prompt": "Provide a synonym for \"happy\". Return only the synonym, no explanation.",
+ "result": [
+ "Joyful"
+ ],
+ "expected": "joyful",
+ "model": "deepseek/deepseek-r1:free",
+ "router": "deepseek/deepseek-r1:free",
+ "timestamp": "2025-06-05T22:56:29.261Z",
+ "passed": true,
+ "duration": 3471,
+ "category": "language"
}
],
"highscores": [
@@ -1999,8 +2240,8 @@
"rankings": [
{
"model": "openai/gpt-4o-mini",
- "duration": 639,
- "duration_secs": 0.639
+ "duration": 521,
+ "duration_secs": 0.521
},
{
"model": "openai/gpt-3.5-turbo",
@@ -2019,8 +2260,8 @@
},
{
"model": "openai/gpt-4o-mini",
- "duration": 641,
- "duration_secs": 0.641
+ "duration": 623,
+ "duration_secs": 0.623
}
]
},
@@ -2034,8 +2275,8 @@
},
{
"model": "openai/gpt-4o-mini",
- "duration": 8548,
- "duration_secs": 8.548
+ "duration": 1508,
+ "duration_secs": 1.508
}
]
},
@@ -2044,8 +2285,8 @@
"rankings": [
{
"model": "openai/gpt-4o-mini",
- "duration": 657,
- "duration_secs": 0.657
+ "duration": 689,
+ "duration_secs": 0.689
},
{
"model": "openai/gpt-3.5-turbo",
@@ -2064,11 +2305,11 @@
},
{
"model": "openai/gpt-4o-mini",
- "duration": 622,
- "duration_secs": 0.622
+ "duration": 661,
+ "duration_secs": 0.661
}
]
}
],
- "lastUpdated": "2025-06-05T22:53:48.763Z"
+ "lastUpdated": "2025-06-05T22:56:29.262Z"
}
\ No newline at end of file
diff --git a/packages/kbot/tests/unit/reports/language.md b/packages/kbot/tests/unit/reports/language.md
index 2102b9ea..3f2878b4 100644
--- a/packages/kbot/tests/unit/reports/language.md
+++ b/packages/kbot/tests/unit/reports/language.md
@@ -6,29 +6,29 @@
| Test | Model | Duration (ms) | Duration (s) |
|------|-------|--------------|--------------|
-| translation | openai/gpt-4o-mini | 1451 | 1.45 |
-| translation | anthropic/claude-sonnet-4 | 1560 | 1.56 |
-| translation | deepseek/deepseek-r1:free | 5434 | 5.43 |
-| grammar | openai/gpt-4o-mini | 695 | 0.69 |
-| grammar | anthropic/claude-sonnet-4 | 1201 | 1.20 |
-| grammar | deepseek/deepseek-r1:free | 3395 | 3.40 |
-| summarization | openai/gpt-4o-mini | 692 | 0.69 |
-| summarization | anthropic/claude-sonnet-4 | 1304 | 1.30 |
-| summarization | deepseek/deepseek-r1:free | 14038 | 14.04 |
-| language_detection | openai/gpt-4o-mini | 459 | 0.46 |
-| language_detection | anthropic/claude-sonnet-4 | 1137 | 1.14 |
-| language_detection | deepseek/deepseek-r1:free | 3924 | 3.92 |
-| synonyms | openai/gpt-4o-mini | 622 | 0.62 |
-| synonyms | anthropic/claude-sonnet-4 | 1251 | 1.25 |
-| synonyms | deepseek/deepseek-r1:free | 3836 | 3.84 |
+| summarization | openai/gpt-4o-mini | 1508 | 1.51 |
+| summarization | anthropic/claude-sonnet-4 | 1621 | 1.62 |
+| summarization | deepseek/deepseek-r1:free | 6000 | 6.00 |
+| translation | openai/gpt-4o-mini | 521 | 0.52 |
+| translation | anthropic/claude-sonnet-4 | 1250 | 1.25 |
+| translation | deepseek/deepseek-r1:free | 2969 | 2.97 |
+| grammar | openai/gpt-4o-mini | 623 | 0.62 |
+| grammar | anthropic/claude-sonnet-4 | 1246 | 1.25 |
+| grammar | deepseek/deepseek-r1:free | 4594 | 4.59 |
+| language_detection | openai/gpt-4o-mini | 689 | 0.69 |
+| language_detection | anthropic/claude-sonnet-4 | 1107 | 1.11 |
+| language_detection | deepseek/deepseek-r1:free | 4218 | 4.22 |
+| synonyms | openai/gpt-4o-mini | 661 | 0.66 |
+| synonyms | anthropic/claude-sonnet-4 | 1639 | 1.64 |
+| synonyms | deepseek/deepseek-r1:free | 3471 | 3.47 |
## Summary
- Total Tests: 15
-- Passed: 9
-- Failed: 6
-- Success Rate: 60.00%
-- Average Duration: 2733ms (2.73s)
+- Passed: 12
+- Failed: 3
+- Success Rate: 80.00%
+- Average Duration: 2141ms (2.14s)
## Failed Tests
@@ -37,126 +37,123 @@
- Prompt: `Correct the grammar in: "I goes to the store yesterday". Return only the corrected sentence, no explanation.`
- Expected: `I went to the store yesterday`
- Actual: `I went to the store yesterday.`
-- Duration: 1201ms (1.20s)
+- Duration: 1246ms (1.25s)
- Reason: Expected I went to the store yesterday, but got I went to the store yesterday.
-- Timestamp: 6/6/2025, 12:30:29 AM
+- Timestamp: 6/6/2025, 12:56:12 AM
### grammar - openai/gpt-4o-mini
- Prompt: `Correct the grammar in: "I goes to the store yesterday". Return only the corrected sentence, no explanation.`
- Expected: `I went to the store yesterday`
- Actual: `"I went to the store yesterday."`
-- Duration: 695ms (0.69s)
+- Duration: 623ms (0.62s)
- Reason: Expected I went to the store yesterday, but got "I went to the store yesterday."
-- Timestamp: 6/6/2025, 12:30:30 AM
+- Timestamp: 6/6/2025, 12:56:12 AM
### grammar - deepseek/deepseek-r1:free
- Prompt: `Correct the grammar in: "I goes to the store yesterday". Return only the corrected sentence, no explanation.`
- Expected: `I went to the store yesterday`
- Actual: `I went to the store yesterday.`
-- Duration: 3395ms (3.40s)
+- Duration: 4594ms (4.59s)
- Reason: Expected I went to the store yesterday, but got I went to the store yesterday.
-- Timestamp: 6/6/2025, 12:30:33 AM
+- Timestamp: 6/6/2025, 12:56:17 AM
+
+## Passed Tests
### summarization - anthropic/claude-sonnet-4
- Prompt: `Summarize: "The quick brown fox jumps over the dog". Return only the summary, compact, no explanation.`
- Expected: `A fox jumps over a dog`
-- Actual: `A brown fox leaps over a dog.`
-- Duration: 1304ms (1.30s)
-- Reason: Expected A fox jumps over a dog, but got A brown fox leaps over a dog.
-- Timestamp: 6/6/2025, 12:30:34 AM
+- Actual: `Fox jumps over dog.`
+- Duration: 1621ms (1.62s)
+- Timestamp: 6/6/2025, 12:55:58 AM
### summarization - openai/gpt-4o-mini
- Prompt: `Summarize: "The quick brown fox jumps over the dog". Return only the summary, compact, no explanation.`
- Expected: `A fox jumps over a dog`
- Actual: `A fox jumps over a dog.`
-- Duration: 692ms (0.69s)
-- Reason: Expected A fox jumps over a dog, but got A fox jumps over a dog.
-- Timestamp: 6/6/2025, 12:30:35 AM
+- Duration: 1508ms (1.51s)
+- Timestamp: 6/6/2025, 12:56:00 AM
### summarization - deepseek/deepseek-r1:free
- Prompt: `Summarize: "The quick brown fox jumps over the dog". Return only the summary, compact, no explanation.`
- Expected: `A fox jumps over a dog`
-- Actual: `"A quick brown fox leaps over a dog."`
-- Duration: 14038ms (14.04s)
-- Reason: Expected A fox jumps over a dog, but got "A quick brown fox leaps over a dog."
-- Timestamp: 6/6/2025, 12:30:49 AM
-
-## Passed Tests
+- Actual: `A quick brown fox leaps over a dog.`
+- Duration: 6000ms (6.00s)
+- Timestamp: 6/6/2025, 12:56:06 AM
### translation - anthropic/claude-sonnet-4
- Prompt: `Translate "Hello, world!" to Spanish. Return only the translation, no explanation.`
- Expected: `¡Hola, mundo!`
- Actual: `¡Hola, mundo!`
-- Duration: 1560ms (1.56s)
-- Timestamp: 6/6/2025, 12:30:21 AM
+- Duration: 1250ms (1.25s)
+- Timestamp: 6/6/2025, 12:56:07 AM
### translation - openai/gpt-4o-mini
- Prompt: `Translate "Hello, world!" to Spanish. Return only the translation, no explanation.`
- Expected: `¡Hola, mundo!`
- Actual: `¡Hola, mundo!`
-- Duration: 1451ms (1.45s)
-- Timestamp: 6/6/2025, 12:30:22 AM
+- Duration: 521ms (0.52s)
+- Timestamp: 6/6/2025, 12:56:07 AM
### translation - deepseek/deepseek-r1:free
- Prompt: `Translate "Hello, world!" to Spanish. Return only the translation, no explanation.`
- Expected: `¡Hola, mundo!`
- Actual: `¡Hola, mundo!`
-- Duration: 5434ms (5.43s)
-- Timestamp: 6/6/2025, 12:30:28 AM
+- Duration: 2969ms (2.97s)
+- Timestamp: 6/6/2025, 12:56:10 AM
### language_detection - anthropic/claude-sonnet-4
- Prompt: `Identify the language of: "Bonjour, comment allez-vous?". Return only the language name, no explanation.`
- Expected: `French`
- Actual: `French`
-- Duration: 1137ms (1.14s)
-- Timestamp: 6/6/2025, 12:30:50 AM
+- Duration: 1107ms (1.11s)
+- Timestamp: 6/6/2025, 12:56:18 AM
### language_detection - openai/gpt-4o-mini
- Prompt: `Identify the language of: "Bonjour, comment allez-vous?". Return only the language name, no explanation.`
- Expected: `French`
- Actual: `French`
-- Duration: 459ms (0.46s)
-- Timestamp: 6/6/2025, 12:30:51 AM
+- Duration: 689ms (0.69s)
+- Timestamp: 6/6/2025, 12:56:19 AM
### language_detection - deepseek/deepseek-r1:free
- Prompt: `Identify the language of: "Bonjour, comment allez-vous?". Return only the language name, no explanation.`
- Expected: `French`
- Actual: `French`
-- Duration: 3924ms (3.92s)
-- Timestamp: 6/6/2025, 12:30:55 AM
+- Duration: 4218ms (4.22s)
+- Timestamp: 6/6/2025, 12:56:23 AM
### synonyms - anthropic/claude-sonnet-4
- Prompt: `Provide a synonym for "happy". Return only the synonym, no explanation.`
- Expected: `joyful`
- Actual: `Joyful`
-- Duration: 1251ms (1.25s)
-- Timestamp: 6/6/2025, 12:30:56 AM
+- Duration: 1639ms (1.64s)
+- Timestamp: 6/6/2025, 12:56:25 AM
### synonyms - openai/gpt-4o-mini
- Prompt: `Provide a synonym for "happy". Return only the synonym, no explanation.`
- Expected: `joyful`
- Actual: `Joyful`
-- Duration: 622ms (0.62s)
-- Timestamp: 6/6/2025, 12:30:57 AM
+- Duration: 661ms (0.66s)
+- Timestamp: 6/6/2025, 12:56:25 AM
### synonyms - deepseek/deepseek-r1:free
- Prompt: `Provide a synonym for "happy". Return only the synonym, no explanation.`
- Expected: `joyful`
-- Actual: `joyful`
-- Duration: 3836ms (3.84s)
-- Timestamp: 6/6/2025, 12:31:00 AM
+- Actual: `Joyful`
+- Duration: 3471ms (3.47s)
+- Timestamp: 6/6/2025, 12:56:29 AM