maintainence love:)

This commit is contained in:
babayaga 2025-12-24 21:11:41 +01:00
parent 0bc9799dde
commit 8ecd654eda
9 changed files with 461 additions and 421 deletions

View File

@ -1,5 +1,5 @@
{ {
"timestamp": 1766607018996, "timestamp": 1766607089404,
"models": [ "models": [
{ {
"id": "gpt-4-0613", "id": "gpt-4-0613",

View File

@ -1,5 +1,5 @@
{ {
"timestamp": 1766607019100, "timestamp": 1766607089536,
"models": [ "models": [
{ {
"id": "bytedance-seed/seed-1.6-flash", "id": "bytedance-seed/seed-1.6-flash",
@ -14174,52 +14174,6 @@
"frequency_penalty": null "frequency_penalty": null
} }
}, },
{
"id": "anthropic/claude-3.5-sonnet",
"canonical_slug": "anthropic/claude-3.5-sonnet",
"hugging_face_id": null,
"name": "Anthropic: Claude 3.5 Sonnet",
"created": 1729555200,
"description": "New Claude 3.5 Sonnet delivers better-than-Opus capabilities, faster-than-Sonnet speeds, at the same Sonnet prices. Sonnet is particularly good at:\n\n- Coding: Scores ~49% on SWE-Bench Verified, higher than the last best score, and without any fancy prompt scaffolding\n- Data science: Augments human data science expertise; navigates unstructured data while using multiple tools for insights\n- Visual processing: excelling at interpreting charts, graphs, and images, accurately transcribing text to derive insights beyond just the text alone\n- Agentic tasks: exceptional tool use, making it great at agentic tasks (i.e. complex, multi-step problem solving tasks that require engaging with other systems)\n\n#multimodal",
"context_length": 200000,
"architecture": {
"modality": "text+image->text",
"input_modalities": [
"text",
"image",
"file"
],
"output_modalities": [
"text"
],
"tokenizer": "Claude",
"instruct_type": null
},
"pricing": {
"prompt": "0.000006",
"completion": "0.00003",
"request": "0",
"image": "0",
"web_search": "0",
"internal_reasoning": "0"
},
"top_provider": {
"context_length": 200000,
"max_completion_tokens": 8192,
"is_moderated": true
},
"per_request_limits": null,
"supported_parameters": [
"max_tokens",
"stop",
"temperature",
"tool_choice",
"tools",
"top_k",
"top_p"
],
"default_parameters": {}
},
{ {
"id": "anthracite-org/magnum-v4-72b", "id": "anthracite-org/magnum-v4-72b",
"canonical_slug": "anthracite-org/magnum-v4-72b", "canonical_slug": "anthracite-org/magnum-v4-72b",
@ -14272,6 +14226,52 @@
], ],
"default_parameters": {} "default_parameters": {}
}, },
{
"id": "anthropic/claude-3.5-sonnet",
"canonical_slug": "anthropic/claude-3.5-sonnet",
"hugging_face_id": null,
"name": "Anthropic: Claude 3.5 Sonnet",
"created": 1729555200,
"description": "New Claude 3.5 Sonnet delivers better-than-Opus capabilities, faster-than-Sonnet speeds, at the same Sonnet prices. Sonnet is particularly good at:\n\n- Coding: Scores ~49% on SWE-Bench Verified, higher than the last best score, and without any fancy prompt scaffolding\n- Data science: Augments human data science expertise; navigates unstructured data while using multiple tools for insights\n- Visual processing: excelling at interpreting charts, graphs, and images, accurately transcribing text to derive insights beyond just the text alone\n- Agentic tasks: exceptional tool use, making it great at agentic tasks (i.e. complex, multi-step problem solving tasks that require engaging with other systems)\n\n#multimodal",
"context_length": 200000,
"architecture": {
"modality": "text+image->text",
"input_modalities": [
"text",
"image",
"file"
],
"output_modalities": [
"text"
],
"tokenizer": "Claude",
"instruct_type": null
},
"pricing": {
"prompt": "0.000006",
"completion": "0.00003",
"request": "0",
"image": "0",
"web_search": "0",
"internal_reasoning": "0"
},
"top_provider": {
"context_length": 200000,
"max_completion_tokens": 8192,
"is_moderated": true
},
"per_request_limits": null,
"supported_parameters": [
"max_tokens",
"stop",
"temperature",
"tool_choice",
"tools",
"top_k",
"top_p"
],
"default_parameters": {}
},
{ {
"id": "mistralai/ministral-8b", "id": "mistralai/ministral-8b",
"canonical_slug": "mistralai/ministral-8b", "canonical_slug": "mistralai/ministral-8b",
@ -15097,6 +15097,57 @@
], ],
"default_parameters": {} "default_parameters": {}
}, },
{
"id": "sao10k/l3.1-euryale-70b",
"canonical_slug": "sao10k/l3.1-euryale-70b",
"hugging_face_id": "Sao10K/L3.1-70B-Euryale-v2.2",
"name": "Sao10K: Llama 3.1 Euryale 70B v2.2",
"created": 1724803200,
"description": "Euryale L3.1 70B v2.2 is a model focused on creative roleplay from [Sao10k](https://ko-fi.com/sao10k). It is the successor of [Euryale L3 70B v2.1](/models/sao10k/l3-euryale-70b).",
"context_length": 32768,
"architecture": {
"modality": "text->text",
"input_modalities": [
"text"
],
"output_modalities": [
"text"
],
"tokenizer": "Llama3",
"instruct_type": "llama3"
},
"pricing": {
"prompt": "0.00000065",
"completion": "0.00000075",
"request": "0",
"image": "0",
"web_search": "0",
"internal_reasoning": "0"
},
"top_provider": {
"context_length": 32768,
"max_completion_tokens": null,
"is_moderated": false
},
"per_request_limits": null,
"supported_parameters": [
"frequency_penalty",
"max_tokens",
"min_p",
"presence_penalty",
"repetition_penalty",
"response_format",
"seed",
"stop",
"structured_outputs",
"temperature",
"tool_choice",
"tools",
"top_k",
"top_p"
],
"default_parameters": {}
},
{ {
"id": "qwen/qwen-2.5-vl-7b-instruct:free", "id": "qwen/qwen-2.5-vl-7b-instruct:free",
"canonical_slug": "qwen/qwen-2-vl-7b-instruct", "canonical_slug": "qwen/qwen-2-vl-7b-instruct",
@ -15189,57 +15240,6 @@
], ],
"default_parameters": {} "default_parameters": {}
}, },
{
"id": "sao10k/l3.1-euryale-70b",
"canonical_slug": "sao10k/l3.1-euryale-70b",
"hugging_face_id": "Sao10K/L3.1-70B-Euryale-v2.2",
"name": "Sao10K: Llama 3.1 Euryale 70B v2.2",
"created": 1724803200,
"description": "Euryale L3.1 70B v2.2 is a model focused on creative roleplay from [Sao10k](https://ko-fi.com/sao10k). It is the successor of [Euryale L3 70B v2.1](/models/sao10k/l3-euryale-70b).",
"context_length": 32768,
"architecture": {
"modality": "text->text",
"input_modalities": [
"text"
],
"output_modalities": [
"text"
],
"tokenizer": "Llama3",
"instruct_type": "llama3"
},
"pricing": {
"prompt": "0.00000065",
"completion": "0.00000075",
"request": "0",
"image": "0",
"web_search": "0",
"internal_reasoning": "0"
},
"top_provider": {
"context_length": 32768,
"max_completion_tokens": null,
"is_moderated": false
},
"per_request_limits": null,
"supported_parameters": [
"frequency_penalty",
"max_tokens",
"min_p",
"presence_penalty",
"repetition_penalty",
"response_format",
"seed",
"stop",
"structured_outputs",
"temperature",
"tool_choice",
"tools",
"top_k",
"top_p"
],
"default_parameters": {}
},
{ {
"id": "microsoft/phi-3.5-mini-128k-instruct", "id": "microsoft/phi-3.5-mini-128k-instruct",
"canonical_slug": "microsoft/phi-3.5-mini-128k-instruct", "canonical_slug": "microsoft/phi-3.5-mini-128k-instruct",
@ -15625,57 +15625,6 @@
], ],
"default_parameters": {} "default_parameters": {}
}, },
{
"id": "meta-llama/llama-3.1-70b-instruct",
"canonical_slug": "meta-llama/llama-3.1-70b-instruct",
"hugging_face_id": "meta-llama/Meta-Llama-3.1-70B-Instruct",
"name": "Meta: Llama 3.1 70B Instruct",
"created": 1721692800,
"description": "Meta's latest class of model (Llama 3.1) launched with a variety of sizes & flavors. This 70B instruct-tuned version is optimized for high quality dialogue usecases.\n\nIt has demonstrated strong performance compared to leading closed-source models in human evaluations.\n\nTo read more about the model release, [click here](https://ai.meta.com/blog/meta-llama-3-1/). Usage of this model is subject to [Meta's Acceptable Use Policy](https://llama.meta.com/llama3/use-policy/).",
"context_length": 131072,
"architecture": {
"modality": "text->text",
"input_modalities": [
"text"
],
"output_modalities": [
"text"
],
"tokenizer": "Llama3",
"instruct_type": "llama3"
},
"pricing": {
"prompt": "0.0000004",
"completion": "0.0000004",
"request": "0",
"image": "0",
"web_search": "0",
"internal_reasoning": "0"
},
"top_provider": {
"context_length": 131072,
"max_completion_tokens": null,
"is_moderated": false
},
"per_request_limits": null,
"supported_parameters": [
"frequency_penalty",
"logit_bias",
"max_tokens",
"min_p",
"presence_penalty",
"repetition_penalty",
"response_format",
"seed",
"stop",
"temperature",
"tool_choice",
"tools",
"top_k",
"top_p"
],
"default_parameters": {}
},
{ {
"id": "meta-llama/llama-3.1-8b-instruct", "id": "meta-llama/llama-3.1-8b-instruct",
"canonical_slug": "meta-llama/llama-3.1-8b-instruct", "canonical_slug": "meta-llama/llama-3.1-8b-instruct",
@ -15824,6 +15773,57 @@
], ],
"default_parameters": {} "default_parameters": {}
}, },
{
"id": "meta-llama/llama-3.1-70b-instruct",
"canonical_slug": "meta-llama/llama-3.1-70b-instruct",
"hugging_face_id": "meta-llama/Meta-Llama-3.1-70B-Instruct",
"name": "Meta: Llama 3.1 70B Instruct",
"created": 1721692800,
"description": "Meta's latest class of model (Llama 3.1) launched with a variety of sizes & flavors. This 70B instruct-tuned version is optimized for high quality dialogue usecases.\n\nIt has demonstrated strong performance compared to leading closed-source models in human evaluations.\n\nTo read more about the model release, [click here](https://ai.meta.com/blog/meta-llama-3-1/). Usage of this model is subject to [Meta's Acceptable Use Policy](https://llama.meta.com/llama3/use-policy/).",
"context_length": 131072,
"architecture": {
"modality": "text->text",
"input_modalities": [
"text"
],
"output_modalities": [
"text"
],
"tokenizer": "Llama3",
"instruct_type": "llama3"
},
"pricing": {
"prompt": "0.0000004",
"completion": "0.0000004",
"request": "0",
"image": "0",
"web_search": "0",
"internal_reasoning": "0"
},
"top_provider": {
"context_length": 131072,
"max_completion_tokens": null,
"is_moderated": false
},
"per_request_limits": null,
"supported_parameters": [
"frequency_penalty",
"logit_bias",
"max_tokens",
"min_p",
"presence_penalty",
"repetition_penalty",
"response_format",
"seed",
"stop",
"temperature",
"tool_choice",
"tools",
"top_k",
"top_p"
],
"default_parameters": {}
},
{ {
"id": "mistralai/mistral-nemo", "id": "mistralai/mistral-nemo",
"canonical_slug": "mistralai/mistral-nemo", "canonical_slug": "mistralai/mistral-nemo",
@ -16124,6 +16124,54 @@
], ],
"default_parameters": {} "default_parameters": {}
}, },
{
"id": "nousresearch/hermes-2-pro-llama-3-8b",
"canonical_slug": "nousresearch/hermes-2-pro-llama-3-8b",
"hugging_face_id": "NousResearch/Hermes-2-Pro-Llama-3-8B",
"name": "NousResearch: Hermes 2 Pro - Llama-3 8B",
"created": 1716768000,
"description": "Hermes 2 Pro is an upgraded, retrained version of Nous Hermes 2, consisting of an updated and cleaned version of the OpenHermes 2.5 Dataset, as well as a newly introduced Function Calling and JSON Mode dataset developed in-house.",
"context_length": 8192,
"architecture": {
"modality": "text->text",
"input_modalities": [
"text"
],
"output_modalities": [
"text"
],
"tokenizer": "Llama3",
"instruct_type": "chatml"
},
"pricing": {
"prompt": "0.000000025",
"completion": "0.00000008",
"request": "0",
"image": "0",
"web_search": "0",
"internal_reasoning": "0"
},
"top_provider": {
"context_length": 8192,
"max_completion_tokens": 2048,
"is_moderated": false
},
"per_request_limits": null,
"supported_parameters": [
"frequency_penalty",
"max_tokens",
"presence_penalty",
"repetition_penalty",
"response_format",
"seed",
"stop",
"structured_outputs",
"temperature",
"top_k",
"top_p"
],
"default_parameters": {}
},
{ {
"id": "mistralai/mistral-7b-instruct:free", "id": "mistralai/mistral-7b-instruct:free",
"canonical_slug": "mistralai/mistral-7b-instruct", "canonical_slug": "mistralai/mistral-7b-instruct",
@ -16278,54 +16326,6 @@
"temperature": 0.3 "temperature": 0.3
} }
}, },
{
"id": "nousresearch/hermes-2-pro-llama-3-8b",
"canonical_slug": "nousresearch/hermes-2-pro-llama-3-8b",
"hugging_face_id": "NousResearch/Hermes-2-Pro-Llama-3-8B",
"name": "NousResearch: Hermes 2 Pro - Llama-3 8B",
"created": 1716768000,
"description": "Hermes 2 Pro is an upgraded, retrained version of Nous Hermes 2, consisting of an updated and cleaned version of the OpenHermes 2.5 Dataset, as well as a newly introduced Function Calling and JSON Mode dataset developed in-house.",
"context_length": 8192,
"architecture": {
"modality": "text->text",
"input_modalities": [
"text"
],
"output_modalities": [
"text"
],
"tokenizer": "Llama3",
"instruct_type": "chatml"
},
"pricing": {
"prompt": "0.000000025",
"completion": "0.00000008",
"request": "0",
"image": "0",
"web_search": "0",
"internal_reasoning": "0"
},
"top_provider": {
"context_length": 8192,
"max_completion_tokens": 2048,
"is_moderated": false
},
"per_request_limits": null,
"supported_parameters": [
"frequency_penalty",
"max_tokens",
"presence_penalty",
"repetition_penalty",
"response_format",
"seed",
"stop",
"structured_outputs",
"temperature",
"top_k",
"top_p"
],
"default_parameters": {}
},
{ {
"id": "microsoft/phi-3-mini-128k-instruct", "id": "microsoft/phi-3-mini-128k-instruct",
"canonical_slug": "microsoft/phi-3-mini-128k-instruct", "canonical_slug": "microsoft/phi-3-mini-128k-instruct",
@ -16410,6 +16410,107 @@
], ],
"default_parameters": {} "default_parameters": {}
}, },
{
"id": "meta-llama/llama-guard-2-8b",
"canonical_slug": "meta-llama/llama-guard-2-8b",
"hugging_face_id": "meta-llama/Meta-Llama-Guard-2-8B",
"name": "Meta: LlamaGuard 2 8B",
"created": 1715558400,
"description": "This safeguard model has 8B parameters and is based on the Llama 3 family. Just like is predecessor, [LlamaGuard 1](https://huggingface.co/meta-llama/LlamaGuard-7b), it can do both prompt and response classification.\n\nLlamaGuard 2 acts as a normal LLM would, generating text that indicates whether the given input/output is safe/unsafe. If deemed unsafe, it will also share the content categories violated.\n\nFor best results, please use raw prompt input or the `/completions` endpoint, instead of the chat API.\n\nIt has demonstrated strong performance compared to leading closed-source models in human evaluations.\n\nTo read more about the model release, [click here](https://ai.meta.com/blog/meta-llama-3/). Usage of this model is subject to [Meta's Acceptable Use Policy](https://llama.meta.com/llama3/use-policy/).",
"context_length": 8192,
"architecture": {
"modality": "text->text",
"input_modalities": [
"text"
],
"output_modalities": [
"text"
],
"tokenizer": "Llama3",
"instruct_type": "none"
},
"pricing": {
"prompt": "0.0000002",
"completion": "0.0000002",
"request": "0",
"image": "0",
"web_search": "0",
"internal_reasoning": "0"
},
"top_provider": {
"context_length": 8192,
"max_completion_tokens": null,
"is_moderated": false
},
"per_request_limits": null,
"supported_parameters": [
"frequency_penalty",
"logit_bias",
"max_tokens",
"min_p",
"presence_penalty",
"repetition_penalty",
"stop",
"temperature",
"top_k",
"top_p"
],
"default_parameters": {}
},
{
"id": "openai/gpt-4o-2024-05-13",
"canonical_slug": "openai/gpt-4o-2024-05-13",
"hugging_face_id": null,
"name": "OpenAI: GPT-4o (2024-05-13)",
"created": 1715558400,
"description": "GPT-4o (\"o\" for \"omni\") is OpenAI's latest AI model, supporting both text and image inputs with text outputs. It maintains the intelligence level of [GPT-4 Turbo](/models/openai/gpt-4-turbo) while being twice as fast and 50% more cost-effective. GPT-4o also offers improved performance in processing non-English languages and enhanced visual capabilities.\n\nFor benchmarking against other models, it was briefly called [\"im-also-a-good-gpt2-chatbot\"](https://twitter.com/LiamFedus/status/1790064963966370209)\n\n#multimodal",
"context_length": 128000,
"architecture": {
"modality": "text+image->text",
"input_modalities": [
"text",
"image",
"file"
],
"output_modalities": [
"text"
],
"tokenizer": "GPT",
"instruct_type": null
},
"pricing": {
"prompt": "0.000005",
"completion": "0.000015",
"request": "0",
"image": "0.007225",
"web_search": "0",
"internal_reasoning": "0"
},
"top_provider": {
"context_length": 128000,
"max_completion_tokens": 4096,
"is_moderated": true
},
"per_request_limits": null,
"supported_parameters": [
"frequency_penalty",
"logit_bias",
"logprobs",
"max_tokens",
"presence_penalty",
"response_format",
"seed",
"stop",
"structured_outputs",
"temperature",
"tool_choice",
"tools",
"top_logprobs",
"top_p",
"web_search_options"
],
"default_parameters": {}
},
{ {
"id": "openai/gpt-4o", "id": "openai/gpt-4o",
"canonical_slug": "openai/gpt-4o", "canonical_slug": "openai/gpt-4o",
@ -16519,107 +16620,6 @@
], ],
"default_parameters": {} "default_parameters": {}
}, },
{
"id": "openai/gpt-4o-2024-05-13",
"canonical_slug": "openai/gpt-4o-2024-05-13",
"hugging_face_id": null,
"name": "OpenAI: GPT-4o (2024-05-13)",
"created": 1715558400,
"description": "GPT-4o (\"o\" for \"omni\") is OpenAI's latest AI model, supporting both text and image inputs with text outputs. It maintains the intelligence level of [GPT-4 Turbo](/models/openai/gpt-4-turbo) while being twice as fast and 50% more cost-effective. GPT-4o also offers improved performance in processing non-English languages and enhanced visual capabilities.\n\nFor benchmarking against other models, it was briefly called [\"im-also-a-good-gpt2-chatbot\"](https://twitter.com/LiamFedus/status/1790064963966370209)\n\n#multimodal",
"context_length": 128000,
"architecture": {
"modality": "text+image->text",
"input_modalities": [
"text",
"image",
"file"
],
"output_modalities": [
"text"
],
"tokenizer": "GPT",
"instruct_type": null
},
"pricing": {
"prompt": "0.000005",
"completion": "0.000015",
"request": "0",
"image": "0.007225",
"web_search": "0",
"internal_reasoning": "0"
},
"top_provider": {
"context_length": 128000,
"max_completion_tokens": 4096,
"is_moderated": true
},
"per_request_limits": null,
"supported_parameters": [
"frequency_penalty",
"logit_bias",
"logprobs",
"max_tokens",
"presence_penalty",
"response_format",
"seed",
"stop",
"structured_outputs",
"temperature",
"tool_choice",
"tools",
"top_logprobs",
"top_p",
"web_search_options"
],
"default_parameters": {}
},
{
"id": "meta-llama/llama-guard-2-8b",
"canonical_slug": "meta-llama/llama-guard-2-8b",
"hugging_face_id": "meta-llama/Meta-Llama-Guard-2-8B",
"name": "Meta: LlamaGuard 2 8B",
"created": 1715558400,
"description": "This safeguard model has 8B parameters and is based on the Llama 3 family. Just like is predecessor, [LlamaGuard 1](https://huggingface.co/meta-llama/LlamaGuard-7b), it can do both prompt and response classification.\n\nLlamaGuard 2 acts as a normal LLM would, generating text that indicates whether the given input/output is safe/unsafe. If deemed unsafe, it will also share the content categories violated.\n\nFor best results, please use raw prompt input or the `/completions` endpoint, instead of the chat API.\n\nIt has demonstrated strong performance compared to leading closed-source models in human evaluations.\n\nTo read more about the model release, [click here](https://ai.meta.com/blog/meta-llama-3/). Usage of this model is subject to [Meta's Acceptable Use Policy](https://llama.meta.com/llama3/use-policy/).",
"context_length": 8192,
"architecture": {
"modality": "text->text",
"input_modalities": [
"text"
],
"output_modalities": [
"text"
],
"tokenizer": "Llama3",
"instruct_type": "none"
},
"pricing": {
"prompt": "0.0000002",
"completion": "0.0000002",
"request": "0",
"image": "0",
"web_search": "0",
"internal_reasoning": "0"
},
"top_provider": {
"context_length": 8192,
"max_completion_tokens": null,
"is_moderated": false
},
"per_request_limits": null,
"supported_parameters": [
"frequency_penalty",
"logit_bias",
"max_tokens",
"min_p",
"presence_penalty",
"repetition_penalty",
"stop",
"temperature",
"top_k",
"top_p"
],
"default_parameters": {}
},
{ {
"id": "meta-llama/llama-3-70b-instruct", "id": "meta-llama/llama-3-70b-instruct",
"canonical_slug": "meta-llama/llama-3-70b-instruct", "canonical_slug": "meta-llama/llama-3-70b-instruct",
@ -17017,57 +17017,6 @@
"temperature": 0.3 "temperature": 0.3
} }
}, },
{
"id": "openai/gpt-4-turbo-preview",
"canonical_slug": "openai/gpt-4-turbo-preview",
"hugging_face_id": null,
"name": "OpenAI: GPT-4 Turbo Preview",
"created": 1706140800,
"description": "The preview GPT-4 model with improved instruction following, JSON mode, reproducible outputs, parallel function calling, and more. Training data: up to Dec 2023.\n\n**Note:** heavily rate limited by OpenAI while in preview.",
"context_length": 128000,
"architecture": {
"modality": "text->text",
"input_modalities": [
"text"
],
"output_modalities": [
"text"
],
"tokenizer": "GPT",
"instruct_type": null
},
"pricing": {
"prompt": "0.00001",
"completion": "0.00003",
"request": "0",
"image": "0",
"web_search": "0",
"internal_reasoning": "0"
},
"top_provider": {
"context_length": 128000,
"max_completion_tokens": 4096,
"is_moderated": true
},
"per_request_limits": null,
"supported_parameters": [
"frequency_penalty",
"logit_bias",
"logprobs",
"max_tokens",
"presence_penalty",
"response_format",
"seed",
"stop",
"structured_outputs",
"temperature",
"tool_choice",
"tools",
"top_logprobs",
"top_p"
],
"default_parameters": {}
},
{ {
"id": "openai/gpt-3.5-turbo-0613", "id": "openai/gpt-3.5-turbo-0613",
"canonical_slug": "openai/gpt-3.5-turbo-0613", "canonical_slug": "openai/gpt-3.5-turbo-0613",
@ -17119,6 +17068,57 @@
], ],
"default_parameters": {} "default_parameters": {}
}, },
{
"id": "openai/gpt-4-turbo-preview",
"canonical_slug": "openai/gpt-4-turbo-preview",
"hugging_face_id": null,
"name": "OpenAI: GPT-4 Turbo Preview",
"created": 1706140800,
"description": "The preview GPT-4 model with improved instruction following, JSON mode, reproducible outputs, parallel function calling, and more. Training data: up to Dec 2023.\n\n**Note:** heavily rate limited by OpenAI while in preview.",
"context_length": 128000,
"architecture": {
"modality": "text->text",
"input_modalities": [
"text"
],
"output_modalities": [
"text"
],
"tokenizer": "GPT",
"instruct_type": null
},
"pricing": {
"prompt": "0.00001",
"completion": "0.00003",
"request": "0",
"image": "0",
"web_search": "0",
"internal_reasoning": "0"
},
"top_provider": {
"context_length": 128000,
"max_completion_tokens": 4096,
"is_moderated": true
},
"per_request_limits": null,
"supported_parameters": [
"frequency_penalty",
"logit_bias",
"logprobs",
"max_tokens",
"presence_penalty",
"response_format",
"seed",
"stop",
"structured_outputs",
"temperature",
"tool_choice",
"tools",
"top_logprobs",
"top_p"
],
"default_parameters": {}
},
{ {
"id": "mistralai/mistral-tiny", "id": "mistralai/mistral-tiny",
"canonical_slug": "mistralai/mistral-tiny", "canonical_slug": "mistralai/mistral-tiny",

File diff suppressed because one or more lines are too long

View File

@ -1,12 +1,12 @@
{ {
"name": "@plastichub/kbot", "name": "@plastichub/kbot",
"version": "1.1.56", "version": "1.1.57",
"lockfileVersion": 3, "lockfileVersion": 3,
"requires": true, "requires": true,
"packages": { "packages": {
"": { "": {
"name": "@plastichub/kbot", "name": "@plastichub/kbot",
"version": "1.1.56", "version": "1.1.57",
"license": "ISC", "license": "ISC",
"dependencies": { "dependencies": {
"node-emoji": "^2.2.0" "node-emoji": "^2.2.0"

View File

@ -1,6 +1,6 @@
{ {
"name": "@plastichub/kbot", "name": "@plastichub/kbot",
"version": "1.1.56", "version": "1.1.57",
"main": "main_node.js", "main": "main_node.js",
"author": "", "author": "",
"license": "ISC", "license": "ISC",

File diff suppressed because one or more lines are too long

File diff suppressed because one or more lines are too long

View File

@ -276,8 +276,8 @@ export enum E_OPENROUTER_MODEL {
MODEL_THEDRUMMER_UNSLOPNEMO_12B = "thedrummer/unslopnemo-12b", MODEL_THEDRUMMER_UNSLOPNEMO_12B = "thedrummer/unslopnemo-12b",
MODEL_ANTHROPIC_CLAUDE_3_5_HAIKU_20241022 = "anthropic/claude-3.5-haiku-20241022", MODEL_ANTHROPIC_CLAUDE_3_5_HAIKU_20241022 = "anthropic/claude-3.5-haiku-20241022",
MODEL_ANTHROPIC_CLAUDE_3_5_HAIKU = "anthropic/claude-3.5-haiku", MODEL_ANTHROPIC_CLAUDE_3_5_HAIKU = "anthropic/claude-3.5-haiku",
MODEL_ANTHROPIC_CLAUDE_3_5_SONNET = "anthropic/claude-3.5-sonnet",
MODEL_ANTHRACITE_ORG_MAGNUM_V4_72B = "anthracite-org/magnum-v4-72b", MODEL_ANTHRACITE_ORG_MAGNUM_V4_72B = "anthracite-org/magnum-v4-72b",
MODEL_ANTHROPIC_CLAUDE_3_5_SONNET = "anthropic/claude-3.5-sonnet",
MODEL_MISTRALAI_MINISTRAL_8B = "mistralai/ministral-8b", MODEL_MISTRALAI_MINISTRAL_8B = "mistralai/ministral-8b",
MODEL_MISTRALAI_MINISTRAL_3B = "mistralai/ministral-3b", MODEL_MISTRALAI_MINISTRAL_3B = "mistralai/ministral-3b",
MODEL_QWEN_QWEN_2_5_7B_INSTRUCT = "qwen/qwen-2.5-7b-instruct", MODEL_QWEN_QWEN_2_5_7B_INSTRUCT = "qwen/qwen-2.5-7b-instruct",
@ -295,9 +295,9 @@ export enum E_OPENROUTER_MODEL {
MODEL_MISTRALAI_PIXTRAL_12B = "mistralai/pixtral-12b", MODEL_MISTRALAI_PIXTRAL_12B = "mistralai/pixtral-12b",
MODEL_COHERE_COMMAND_R_08_2024 = "cohere/command-r-08-2024", MODEL_COHERE_COMMAND_R_08_2024 = "cohere/command-r-08-2024",
MODEL_COHERE_COMMAND_R_PLUS_08_2024 = "cohere/command-r-plus-08-2024", MODEL_COHERE_COMMAND_R_PLUS_08_2024 = "cohere/command-r-plus-08-2024",
MODEL_SAO10K_L3_1_EURYALE_70B = "sao10k/l3.1-euryale-70b",
MODEL_QWEN_QWEN_2_5_VL_7B_INSTRUCT_FREE = "qwen/qwen-2.5-vl-7b-instruct:free", MODEL_QWEN_QWEN_2_5_VL_7B_INSTRUCT_FREE = "qwen/qwen-2.5-vl-7b-instruct:free",
MODEL_QWEN_QWEN_2_5_VL_7B_INSTRUCT = "qwen/qwen-2.5-vl-7b-instruct", MODEL_QWEN_QWEN_2_5_VL_7B_INSTRUCT = "qwen/qwen-2.5-vl-7b-instruct",
MODEL_SAO10K_L3_1_EURYALE_70B = "sao10k/l3.1-euryale-70b",
MODEL_MICROSOFT_PHI_3_5_MINI_128K_INSTRUCT = "microsoft/phi-3.5-mini-128k-instruct", MODEL_MICROSOFT_PHI_3_5_MINI_128K_INSTRUCT = "microsoft/phi-3.5-mini-128k-instruct",
MODEL_NOUSRESEARCH_HERMES_3_LLAMA_3_1_70B = "nousresearch/hermes-3-llama-3.1-70b", MODEL_NOUSRESEARCH_HERMES_3_LLAMA_3_1_70B = "nousresearch/hermes-3-llama-3.1-70b",
MODEL_NOUSRESEARCH_HERMES_3_LLAMA_3_1_405B_FREE = "nousresearch/hermes-3-llama-3.1-405b:free", MODEL_NOUSRESEARCH_HERMES_3_LLAMA_3_1_405B_FREE = "nousresearch/hermes-3-llama-3.1-405b:free",
@ -306,26 +306,26 @@ export enum E_OPENROUTER_MODEL {
MODEL_SAO10K_L3_LUNARIS_8B = "sao10k/l3-lunaris-8b", MODEL_SAO10K_L3_LUNARIS_8B = "sao10k/l3-lunaris-8b",
MODEL_OPENAI_GPT_4O_2024_08_06 = "openai/gpt-4o-2024-08-06", MODEL_OPENAI_GPT_4O_2024_08_06 = "openai/gpt-4o-2024-08-06",
MODEL_META_LLAMA_LLAMA_3_1_405B = "meta-llama/llama-3.1-405b", MODEL_META_LLAMA_LLAMA_3_1_405B = "meta-llama/llama-3.1-405b",
MODEL_META_LLAMA_LLAMA_3_1_70B_INSTRUCT = "meta-llama/llama-3.1-70b-instruct",
MODEL_META_LLAMA_LLAMA_3_1_8B_INSTRUCT = "meta-llama/llama-3.1-8b-instruct", MODEL_META_LLAMA_LLAMA_3_1_8B_INSTRUCT = "meta-llama/llama-3.1-8b-instruct",
MODEL_META_LLAMA_LLAMA_3_1_405B_INSTRUCT_FREE = "meta-llama/llama-3.1-405b-instruct:free", MODEL_META_LLAMA_LLAMA_3_1_405B_INSTRUCT_FREE = "meta-llama/llama-3.1-405b-instruct:free",
MODEL_META_LLAMA_LLAMA_3_1_405B_INSTRUCT = "meta-llama/llama-3.1-405b-instruct", MODEL_META_LLAMA_LLAMA_3_1_405B_INSTRUCT = "meta-llama/llama-3.1-405b-instruct",
MODEL_META_LLAMA_LLAMA_3_1_70B_INSTRUCT = "meta-llama/llama-3.1-70b-instruct",
MODEL_MISTRALAI_MISTRAL_NEMO = "mistralai/mistral-nemo", MODEL_MISTRALAI_MISTRAL_NEMO = "mistralai/mistral-nemo",
MODEL_OPENAI_GPT_4O_MINI_2024_07_18 = "openai/gpt-4o-mini-2024-07-18", MODEL_OPENAI_GPT_4O_MINI_2024_07_18 = "openai/gpt-4o-mini-2024-07-18",
MODEL_OPENAI_GPT_4O_MINI = "openai/gpt-4o-mini", MODEL_OPENAI_GPT_4O_MINI = "openai/gpt-4o-mini",
MODEL_GOOGLE_GEMMA_2_27B_IT = "google/gemma-2-27b-it", MODEL_GOOGLE_GEMMA_2_27B_IT = "google/gemma-2-27b-it",
MODEL_GOOGLE_GEMMA_2_9B_IT = "google/gemma-2-9b-it", MODEL_GOOGLE_GEMMA_2_9B_IT = "google/gemma-2-9b-it",
MODEL_SAO10K_L3_EURYALE_70B = "sao10k/l3-euryale-70b", MODEL_SAO10K_L3_EURYALE_70B = "sao10k/l3-euryale-70b",
MODEL_NOUSRESEARCH_HERMES_2_PRO_LLAMA_3_8B = "nousresearch/hermes-2-pro-llama-3-8b",
MODEL_MISTRALAI_MISTRAL_7B_INSTRUCT_FREE = "mistralai/mistral-7b-instruct:free", MODEL_MISTRALAI_MISTRAL_7B_INSTRUCT_FREE = "mistralai/mistral-7b-instruct:free",
MODEL_MISTRALAI_MISTRAL_7B_INSTRUCT = "mistralai/mistral-7b-instruct", MODEL_MISTRALAI_MISTRAL_7B_INSTRUCT = "mistralai/mistral-7b-instruct",
MODEL_MISTRALAI_MISTRAL_7B_INSTRUCT_V0_3 = "mistralai/mistral-7b-instruct-v0.3", MODEL_MISTRALAI_MISTRAL_7B_INSTRUCT_V0_3 = "mistralai/mistral-7b-instruct-v0.3",
MODEL_NOUSRESEARCH_HERMES_2_PRO_LLAMA_3_8B = "nousresearch/hermes-2-pro-llama-3-8b",
MODEL_MICROSOFT_PHI_3_MINI_128K_INSTRUCT = "microsoft/phi-3-mini-128k-instruct", MODEL_MICROSOFT_PHI_3_MINI_128K_INSTRUCT = "microsoft/phi-3-mini-128k-instruct",
MODEL_MICROSOFT_PHI_3_MEDIUM_128K_INSTRUCT = "microsoft/phi-3-medium-128k-instruct", MODEL_MICROSOFT_PHI_3_MEDIUM_128K_INSTRUCT = "microsoft/phi-3-medium-128k-instruct",
MODEL_META_LLAMA_LLAMA_GUARD_2_8B = "meta-llama/llama-guard-2-8b",
MODEL_OPENAI_GPT_4O_2024_05_13 = "openai/gpt-4o-2024-05-13",
MODEL_OPENAI_GPT_4O = "openai/gpt-4o", MODEL_OPENAI_GPT_4O = "openai/gpt-4o",
MODEL_OPENAI_GPT_4O_EXTENDED = "openai/gpt-4o:extended", MODEL_OPENAI_GPT_4O_EXTENDED = "openai/gpt-4o:extended",
MODEL_OPENAI_GPT_4O_2024_05_13 = "openai/gpt-4o-2024-05-13",
MODEL_META_LLAMA_LLAMA_GUARD_2_8B = "meta-llama/llama-guard-2-8b",
MODEL_META_LLAMA_LLAMA_3_70B_INSTRUCT = "meta-llama/llama-3-70b-instruct", MODEL_META_LLAMA_LLAMA_3_70B_INSTRUCT = "meta-llama/llama-3-70b-instruct",
MODEL_META_LLAMA_LLAMA_3_8B_INSTRUCT = "meta-llama/llama-3-8b-instruct", MODEL_META_LLAMA_LLAMA_3_8B_INSTRUCT = "meta-llama/llama-3-8b-instruct",
MODEL_MISTRALAI_MIXTRAL_8X22B_INSTRUCT = "mistralai/mixtral-8x22b-instruct", MODEL_MISTRALAI_MIXTRAL_8X22B_INSTRUCT = "mistralai/mixtral-8x22b-instruct",
@ -334,8 +334,8 @@ export enum E_OPENROUTER_MODEL {
MODEL_ANTHROPIC_CLAUDE_3_HAIKU = "anthropic/claude-3-haiku", MODEL_ANTHROPIC_CLAUDE_3_HAIKU = "anthropic/claude-3-haiku",
MODEL_ANTHROPIC_CLAUDE_3_OPUS = "anthropic/claude-3-opus", MODEL_ANTHROPIC_CLAUDE_3_OPUS = "anthropic/claude-3-opus",
MODEL_MISTRALAI_MISTRAL_LARGE = "mistralai/mistral-large", MODEL_MISTRALAI_MISTRAL_LARGE = "mistralai/mistral-large",
MODEL_OPENAI_GPT_4_TURBO_PREVIEW = "openai/gpt-4-turbo-preview",
MODEL_OPENAI_GPT_3_5_TURBO_0613 = "openai/gpt-3.5-turbo-0613", MODEL_OPENAI_GPT_3_5_TURBO_0613 = "openai/gpt-3.5-turbo-0613",
MODEL_OPENAI_GPT_4_TURBO_PREVIEW = "openai/gpt-4-turbo-preview",
MODEL_MISTRALAI_MISTRAL_TINY = "mistralai/mistral-tiny", MODEL_MISTRALAI_MISTRAL_TINY = "mistralai/mistral-tiny",
MODEL_MISTRALAI_MISTRAL_7B_INSTRUCT_V0_2 = "mistralai/mistral-7b-instruct-v0.2", MODEL_MISTRALAI_MISTRAL_7B_INSTRUCT_V0_2 = "mistralai/mistral-7b-instruct-v0.2",
MODEL_MISTRALAI_MIXTRAL_8X7B_INSTRUCT = "mistralai/mixtral-8x7b-instruct", MODEL_MISTRALAI_MIXTRAL_8X7B_INSTRUCT = "mistralai/mixtral-8x7b-instruct",

View File

@ -31,72 +31,78 @@ export interface IKBotOptions {
 
 OpenRouter models:  OpenRouter models:
 
agentica-org/deepcoder-14b-preview | paid
agentica-org/deepcoder-14b-preview:free | free
ai21/jamba-large-1.7 | paid ai21/jamba-large-1.7 | paid
ai21/jamba-mini-1.7 | paid ai21/jamba-mini-1.7 | paid
aion-labs/aion-1.0 | paid aion-labs/aion-1.0 | paid
aion-labs/aion-1.0-mini | paid aion-labs/aion-1.0-mini | paid
aion-labs/aion-rp-llama-3.1-8b | paid aion-labs/aion-rp-llama-3.1-8b | paid
alfredpros/codellama-7b-instruct-solidity | paid alfredpros/codellama-7b-instruct-solidity | paid
allenai/molmo-7b-d | paid
allenai/olmo-2-0325-32b-instruct | paid allenai/olmo-2-0325-32b-instruct | paid
allenai/olmo-3-32b-think:free | free
allenai/olmo-3-7b-instruct | paid
allenai/olmo-3-7b-think | paid
allenai/olmo-3.1-32b-think:free | free
amazon/nova-2-lite-v1 | paid
amazon/nova-lite-v1 | paid amazon/nova-lite-v1 | paid
amazon/nova-micro-v1 | paid amazon/nova-micro-v1 | paid
amazon/nova-premier-v1 | paid
amazon/nova-pro-v1 | paid amazon/nova-pro-v1 | paid
anthropic/claude-3-haiku | paid anthropic/claude-3-haiku | paid
anthropic/claude-3-opus | paid anthropic/claude-3-opus | paid
anthropic/claude-3.5-haiku | paid anthropic/claude-3.5-haiku | paid
anthropic/claude-3.5-haiku-20241022 | paid anthropic/claude-3.5-haiku-20241022 | paid
anthropic/claude-3.5-sonnet | paid anthropic/claude-3.5-sonnet | paid
anthropic/claude-3.5-sonnet-20240620 | paid
anthropic/claude-3.7-sonnet | paid anthropic/claude-3.7-sonnet | paid
anthropic/claude-3.7-sonnet:thinking | paid anthropic/claude-3.7-sonnet:thinking | paid
anthropic/claude-haiku-4.5 | paid
anthropic/claude-opus-4 | paid anthropic/claude-opus-4 | paid
anthropic/claude-opus-4.1 | paid anthropic/claude-opus-4.1 | paid
anthropic/claude-opus-4.5 | paid
anthropic/claude-sonnet-4 | paid anthropic/claude-sonnet-4 | paid
anthropic/claude-sonnet-4.5 | paid anthropic/claude-sonnet-4.5 | paid
arcee-ai/afm-4.5b | paid
arcee-ai/coder-large | paid arcee-ai/coder-large | paid
arcee-ai/maestro-reasoning | paid arcee-ai/maestro-reasoning | paid
arcee-ai/spotlight | paid arcee-ai/spotlight | paid
arcee-ai/trinity-mini | paid
arcee-ai/trinity-mini:free | free
arcee-ai/virtuoso-large | paid arcee-ai/virtuoso-large | paid
arliai/qwq-32b-arliai-rpr-v1 | paid arliai/qwq-32b-arliai-rpr-v1 | paid
arliai/qwq-32b-arliai-rpr-v1:free | free
openrouter/auto | paid openrouter/auto | paid
baidu/ernie-4.5-21b-a3b | paid baidu/ernie-4.5-21b-a3b | paid
baidu/ernie-4.5-21b-a3b-thinking | paid baidu/ernie-4.5-21b-a3b-thinking | paid
baidu/ernie-4.5-300b-a47b | paid baidu/ernie-4.5-300b-a47b | paid
baidu/ernie-4.5-vl-28b-a3b | paid baidu/ernie-4.5-vl-28b-a3b | paid
baidu/ernie-4.5-vl-424b-a47b | paid baidu/ernie-4.5-vl-424b-a47b | paid
openrouter/bodybuilder | paid
bytedance-seed/seed-1.6 | paid
bytedance-seed/seed-1.6-flash | paid
bytedance/ui-tars-1.5-7b | paid bytedance/ui-tars-1.5-7b | paid
deepcogito/cogito-v2-preview-llama-109b-moe | paid deepcogito/cogito-v2-preview-llama-109b-moe | paid
cohere/command-a | paid cohere/command-a | paid
cohere/command-r-08-2024 | paid cohere/command-r-08-2024 | paid
cohere/command-r-plus-08-2024 | paid cohere/command-r-plus-08-2024 | paid
cohere/command-r7b-12-2024 | paid cohere/command-r7b-12-2024 | paid
deepcogito/cogito-v2-preview-deepseek-671b | paid deepcogito/cogito-v2-preview-llama-405b | paid
deepcogito/cogito-v2-preview-llama-70b | paid
deepcogito/cogito-v2.1-671b | paid
deepseek/deepseek-prover-v2 | paid deepseek/deepseek-prover-v2 | paid
deepseek/deepseek-r1-0528-qwen3-8b | paid deepseek/deepseek-r1-0528-qwen3-8b | paid
deepseek/deepseek-r1-0528-qwen3-8b:free | free
deepseek/deepseek-chat | paid deepseek/deepseek-chat | paid
deepseek/deepseek-chat-v3-0324 | paid deepseek/deepseek-chat-v3-0324 | paid
deepseek/deepseek-chat-v3-0324:free | free
deepseek/deepseek-chat-v3.1 | paid deepseek/deepseek-chat-v3.1 | paid
deepseek/deepseek-chat-v3.1:free | free
deepseek/deepseek-v3.1-terminus | paid deepseek/deepseek-v3.1-terminus | paid
deepseek/deepseek-v3.1-terminus:exacto | paid
deepseek/deepseek-v3.2 | paid
deepseek/deepseek-v3.2-exp | paid deepseek/deepseek-v3.2-exp | paid
deepseek/deepseek-v3.2-speciale | paid
deepseek/deepseek-r1 | paid deepseek/deepseek-r1 | paid
deepseek/deepseek-r1:free | free
deepseek/deepseek-r1-0528 | paid deepseek/deepseek-r1-0528 | paid
deepseek/deepseek-r1-0528:free | free deepseek/deepseek-r1-0528:free | free
deepseek/deepseek-r1-distill-llama-70b | paid deepseek/deepseek-r1-distill-llama-70b | paid
deepseek/deepseek-r1-distill-llama-70b:free | free
deepseek/deepseek-r1-distill-qwen-14b | paid deepseek/deepseek-r1-distill-qwen-14b | paid
deepseek/deepseek-r1-distill-qwen-32b | paid deepseek/deepseek-r1-distill-qwen-32b | paid
cognitivecomputations/dolphin3.0-mistral-24b | paid
cognitivecomputations/dolphin3.0-mistral-24b:free | free
eleutherai/llemma_7b | paid eleutherai/llemma_7b | paid
essentialai/rnj-1-instruct | paid
alpindale/goliath-120b | paid alpindale/goliath-120b | paid
google/gemini-2.0-flash-001 | paid google/gemini-2.0-flash-001 | paid
google/gemini-2.0-flash-exp:free | free google/gemini-2.0-flash-exp:free | free
@ -105,15 +111,15 @@ export interface IKBotOptions {
google/gemini-2.5-flash-image | paid google/gemini-2.5-flash-image | paid
google/gemini-2.5-flash-image-preview | paid google/gemini-2.5-flash-image-preview | paid
google/gemini-2.5-flash-lite | paid google/gemini-2.5-flash-lite | paid
google/gemini-2.5-flash-lite-preview-06-17 | paid
google/gemini-2.5-flash-lite-preview-09-2025 | paid google/gemini-2.5-flash-lite-preview-09-2025 | paid
google/gemini-2.5-flash-preview-09-2025 | paid google/gemini-2.5-flash-preview-09-2025 | paid
google/gemini-2.5-pro | paid google/gemini-2.5-pro | paid
google/gemini-2.5-pro-preview-05-06 | paid google/gemini-2.5-pro-preview-05-06 | paid
google/gemini-2.5-pro-preview | paid google/gemini-2.5-pro-preview | paid
google/gemini-3-flash-preview | paid
google/gemini-3-pro-preview | paid
google/gemma-2-27b-it | paid google/gemma-2-27b-it | paid
google/gemma-2-9b-it | paid google/gemma-2-9b-it | paid
google/gemma-2-9b-it:free | free
google/gemma-3-12b-it | paid google/gemma-3-12b-it | paid
google/gemma-3-12b-it:free | free google/gemma-3-12b-it:free | free
google/gemma-3-27b-it | paid google/gemma-3-27b-it | paid
@ -123,23 +129,24 @@ export interface IKBotOptions {
google/gemma-3n-e2b-it:free | free google/gemma-3n-e2b-it:free | free
google/gemma-3n-e4b-it | paid google/gemma-3n-e4b-it | paid
google/gemma-3n-e4b-it:free | free google/gemma-3n-e4b-it:free | free
google/gemini-3-pro-image-preview | paid
ibm-granite/granite-4.0-h-micro | paid
inception/mercury | paid inception/mercury | paid
inception/mercury-coder | paid inception/mercury-coder | paid
inclusionai/ling-1t | paid
inflection/inflection-3-pi | paid inflection/inflection-3-pi | paid
inflection/inflection-3-productivity | paid inflection/inflection-3-productivity | paid
liquid/lfm-3b | paid kwaipilot/kat-coder-pro:free | free
liquid/lfm-7b | paid liquid/lfm-2.2-6b | paid
liquid/lfm2-8b-a1b | paid
meta-llama/llama-guard-3-8b | paid meta-llama/llama-guard-3-8b | paid
anthracite-org/magnum-v2-72b | paid
anthracite-org/magnum-v4-72b | paid anthracite-org/magnum-v4-72b | paid
mancer/weaver | paid mancer/weaver | paid
meituan/longcat-flash-chat | paid meituan/longcat-flash-chat | paid
meituan/longcat-flash-chat:free | free
meta-llama/llama-3-70b-instruct | paid meta-llama/llama-3-70b-instruct | paid
meta-llama/llama-3-8b-instruct | paid meta-llama/llama-3-8b-instruct | paid
meta-llama/llama-3.1-405b | paid meta-llama/llama-3.1-405b | paid
meta-llama/llama-3.1-405b-instruct | paid meta-llama/llama-3.1-405b-instruct | paid
meta-llama/llama-3.1-405b-instruct:free | free
meta-llama/llama-3.1-70b-instruct | paid meta-llama/llama-3.1-70b-instruct | paid
meta-llama/llama-3.1-8b-instruct | paid meta-llama/llama-3.1-8b-instruct | paid
meta-llama/llama-3.2-11b-vision-instruct | paid meta-llama/llama-3.2-11b-vision-instruct | paid
@ -149,15 +156,10 @@ export interface IKBotOptions {
meta-llama/llama-3.2-90b-vision-instruct | paid meta-llama/llama-3.2-90b-vision-instruct | paid
meta-llama/llama-3.3-70b-instruct | paid meta-llama/llama-3.3-70b-instruct | paid
meta-llama/llama-3.3-70b-instruct:free | free meta-llama/llama-3.3-70b-instruct:free | free
meta-llama/llama-3.3-8b-instruct:free | free
meta-llama/llama-4-maverick | paid meta-llama/llama-4-maverick | paid
meta-llama/llama-4-maverick:free | free
meta-llama/llama-4-scout | paid meta-llama/llama-4-scout | paid
meta-llama/llama-4-scout:free | free
meta-llama/llama-guard-4-12b | paid meta-llama/llama-guard-4-12b | paid
meta-llama/llama-guard-2-8b | paid meta-llama/llama-guard-2-8b | paid
microsoft/mai-ds-r1 | paid
microsoft/mai-ds-r1:free | free
microsoft/phi-4 | paid microsoft/phi-4 | paid
microsoft/phi-4-multimodal-instruct | paid microsoft/phi-4-multimodal-instruct | paid
microsoft/phi-4-reasoning-plus | paid microsoft/phi-4-reasoning-plus | paid
@ -165,21 +167,22 @@ export interface IKBotOptions {
microsoft/phi-3-mini-128k-instruct | paid microsoft/phi-3-mini-128k-instruct | paid
microsoft/phi-3.5-mini-128k-instruct | paid microsoft/phi-3.5-mini-128k-instruct | paid
minimax/minimax-m1 | paid minimax/minimax-m1 | paid
minimax/minimax-m2 | paid
minimax/minimax-m2.1 | paid
minimax/minimax-01 | paid minimax/minimax-01 | paid
mistralai/mistral-large | paid mistralai/mistral-large | paid
mistralai/mistral-large-2407 | paid mistralai/mistral-large-2407 | paid
mistralai/mistral-large-2411 | paid mistralai/mistral-large-2411 | paid
mistralai/mistral-small | paid
mistralai/mistral-tiny | paid mistralai/mistral-tiny | paid
mistralai/codestral-2501 | paid
mistralai/codestral-2508 | paid mistralai/codestral-2508 | paid
mistralai/devstral-2512 | paid
mistralai/devstral-2512:free | free
mistralai/devstral-medium | paid mistralai/devstral-medium | paid
mistralai/devstral-small | paid mistralai/devstral-small | paid
mistralai/devstral-small-2505 | paid mistralai/devstral-small-2505 | paid
mistralai/devstral-small-2505:free | free mistralai/ministral-14b-2512 | paid
mistralai/magistral-medium-2506 | paid mistralai/ministral-3b-2512 | paid
mistralai/magistral-medium-2506:thinking | paid mistralai/ministral-8b-2512 | paid
mistralai/magistral-small-2506 | paid
mistralai/ministral-3b | paid mistralai/ministral-3b | paid
mistralai/ministral-8b | paid mistralai/ministral-8b | paid
mistralai/mistral-7b-instruct | paid mistralai/mistral-7b-instruct | paid
@ -187,35 +190,36 @@ export interface IKBotOptions {
mistralai/mistral-7b-instruct-v0.1 | paid mistralai/mistral-7b-instruct-v0.1 | paid
mistralai/mistral-7b-instruct-v0.2 | paid mistralai/mistral-7b-instruct-v0.2 | paid
mistralai/mistral-7b-instruct-v0.3 | paid mistralai/mistral-7b-instruct-v0.3 | paid
mistralai/mistral-large-2512 | paid
mistralai/mistral-medium-3 | paid mistralai/mistral-medium-3 | paid
mistralai/mistral-medium-3.1 | paid mistralai/mistral-medium-3.1 | paid
mistralai/mistral-nemo | paid mistralai/mistral-nemo | paid
mistralai/mistral-nemo:free | free
mistralai/mistral-small-24b-instruct-2501 | paid mistralai/mistral-small-24b-instruct-2501 | paid
mistralai/mistral-small-24b-instruct-2501:free | free
mistralai/mistral-small-3.1-24b-instruct | paid mistralai/mistral-small-3.1-24b-instruct | paid
mistralai/mistral-small-3.1-24b-instruct:free | free mistralai/mistral-small-3.1-24b-instruct:free | free
mistralai/mistral-small-3.2-24b-instruct | paid mistralai/mistral-small-3.2-24b-instruct | paid
mistralai/mistral-small-3.2-24b-instruct:free | free mistralai/mistral-small-creative | paid
mistralai/mixtral-8x22b-instruct | paid mistralai/mixtral-8x22b-instruct | paid
mistralai/mixtral-8x7b-instruct | paid mistralai/mixtral-8x7b-instruct | paid
mistralai/pixtral-12b | paid mistralai/pixtral-12b | paid
mistralai/pixtral-large-2411 | paid mistralai/pixtral-large-2411 | paid
mistralai/mistral-saba | paid mistralai/mistral-saba | paid
mistralai/voxtral-small-24b-2507 | paid
moonshotai/kimi-dev-72b | paid moonshotai/kimi-dev-72b | paid
moonshotai/kimi-dev-72b:free | free
moonshotai/kimi-k2 | paid moonshotai/kimi-k2 | paid
moonshotai/kimi-k2:free | free moonshotai/kimi-k2:free | free
moonshotai/kimi-k2-0905 | paid moonshotai/kimi-k2-0905 | paid
moonshotai/kimi-k2-0905:exacto | paid
moonshotai/kimi-k2-thinking | paid
morph/morph-v3-fast | paid morph/morph-v3-fast | paid
morph/morph-v3-large | paid morph/morph-v3-large | paid
gryphe/mythomax-l2-13b | paid gryphe/mythomax-l2-13b | paid
neversleep/llama-3.1-lumimaid-8b | paid neversleep/llama-3.1-lumimaid-8b | paid
nex-agi/deepseek-v3.1-nex-n1:free | free
neversleep/noromaid-20b | paid neversleep/noromaid-20b | paid
nousresearch/deephermes-3-llama-3-8b-preview | paid
nousresearch/deephermes-3-llama-3-8b-preview:free | free
nousresearch/deephermes-3-mistral-24b-preview | paid nousresearch/deephermes-3-mistral-24b-preview | paid
nousresearch/hermes-3-llama-3.1-405b | paid nousresearch/hermes-3-llama-3.1-405b | paid
nousresearch/hermes-3-llama-3.1-405b:free | free
nousresearch/hermes-3-llama-3.1-70b | paid nousresearch/hermes-3-llama-3.1-70b | paid
nousresearch/hermes-4-405b | paid nousresearch/hermes-4-405b | paid
nousresearch/hermes-4-70b | paid nousresearch/hermes-4-70b | paid
@ -223,6 +227,10 @@ export interface IKBotOptions {
nvidia/llama-3.1-nemotron-70b-instruct | paid nvidia/llama-3.1-nemotron-70b-instruct | paid
nvidia/llama-3.1-nemotron-ultra-253b-v1 | paid nvidia/llama-3.1-nemotron-ultra-253b-v1 | paid
nvidia/llama-3.3-nemotron-super-49b-v1.5 | paid nvidia/llama-3.3-nemotron-super-49b-v1.5 | paid
nvidia/nemotron-3-nano-30b-a3b | paid
nvidia/nemotron-3-nano-30b-a3b:free | free
nvidia/nemotron-nano-12b-v2-vl | paid
nvidia/nemotron-nano-12b-v2-vl:free | free
nvidia/nemotron-nano-9b-v2 | paid nvidia/nemotron-nano-9b-v2 | paid
nvidia/nemotron-nano-9b-v2:free | free nvidia/nemotron-nano-9b-v2:free | free
openai/chatgpt-4o-latest | paid openai/chatgpt-4o-latest | paid
@ -252,29 +260,43 @@ export interface IKBotOptions {
openai/gpt-5 | paid openai/gpt-5 | paid
openai/gpt-5-chat | paid openai/gpt-5-chat | paid
openai/gpt-5-codex | paid openai/gpt-5-codex | paid
openai/gpt-5-image | paid
openai/gpt-5-image-mini | paid
openai/gpt-5-mini | paid openai/gpt-5-mini | paid
openai/gpt-5-nano | paid openai/gpt-5-nano | paid
openai/gpt-5-pro | paid openai/gpt-5-pro | paid
openai/gpt-5.1 | paid
openai/gpt-5.1-chat | paid
openai/gpt-5.1-codex | paid
openai/gpt-5.1-codex-max | paid
openai/gpt-5.1-codex-mini | paid
openai/gpt-5.2 | paid
openai/gpt-5.2-chat | paid
openai/gpt-5.2-pro | paid
openai/gpt-oss-120b | paid openai/gpt-oss-120b | paid
openai/gpt-oss-120b:exacto | paid
openai/gpt-oss-120b:free | free
openai/gpt-oss-20b | paid openai/gpt-oss-20b | paid
openai/gpt-oss-20b:free | free openai/gpt-oss-20b:free | free
openai/gpt-oss-safeguard-20b | paid
openai/o1 | paid openai/o1 | paid
openai/o1-mini | paid
openai/o1-mini-2024-09-12 | paid
openai/o1-pro | paid openai/o1-pro | paid
openai/o3 | paid openai/o3 | paid
openai/o3-deep-research | paid
openai/o3-mini | paid openai/o3-mini | paid
openai/o3-mini-high | paid openai/o3-mini-high | paid
openai/o3-pro | paid openai/o3-pro | paid
openai/o4-mini | paid openai/o4-mini | paid
openai/o4-mini-deep-research | paid
openai/o4-mini-high | paid openai/o4-mini-high | paid
opengvlab/internvl3-78b | paid opengvlab/internvl3-78b | paid
perplexity/r1-1776 | paid
perplexity/sonar | paid perplexity/sonar | paid
perplexity/sonar-deep-research | paid perplexity/sonar-deep-research | paid
perplexity/sonar-pro | paid perplexity/sonar-pro | paid
perplexity/sonar-pro-search | paid
perplexity/sonar-reasoning | paid perplexity/sonar-reasoning | paid
perplexity/sonar-reasoning-pro | paid perplexity/sonar-reasoning-pro | paid
prime-intellect/intellect-3 | paid
qwen/qwen-plus-2025-07-28 | paid qwen/qwen-plus-2025-07-28 | paid
qwen/qwen-plus-2025-07-28:thinking | paid qwen/qwen-plus-2025-07-28:thinking | paid
qwen/qwen-vl-max | paid qwen/qwen-vl-max | paid
@ -285,26 +307,22 @@ export interface IKBotOptions {
qwen/qwen-2.5-7b-instruct | paid qwen/qwen-2.5-7b-instruct | paid
qwen/qwen2.5-coder-7b-instruct | paid qwen/qwen2.5-coder-7b-instruct | paid
qwen/qwen2.5-vl-32b-instruct | paid qwen/qwen2.5-vl-32b-instruct | paid
qwen/qwen2.5-vl-32b-instruct:free | free
qwen/qwen2.5-vl-72b-instruct | paid qwen/qwen2.5-vl-72b-instruct | paid
qwen/qwen2.5-vl-72b-instruct:free | free
qwen/qwen-2.5-vl-7b-instruct | paid qwen/qwen-2.5-vl-7b-instruct | paid
qwen/qwen-2.5-vl-7b-instruct:free | free
qwen/qwen3-14b | paid qwen/qwen3-14b | paid
qwen/qwen3-14b:free | free
qwen/qwen3-235b-a22b | paid qwen/qwen3-235b-a22b | paid
qwen/qwen3-235b-a22b:free | free
qwen/qwen3-235b-a22b-2507 | paid qwen/qwen3-235b-a22b-2507 | paid
qwen/qwen3-235b-a22b-thinking-2507 | paid qwen/qwen3-235b-a22b-thinking-2507 | paid
qwen/qwen3-30b-a3b | paid qwen/qwen3-30b-a3b | paid
qwen/qwen3-30b-a3b:free | free
qwen/qwen3-30b-a3b-instruct-2507 | paid qwen/qwen3-30b-a3b-instruct-2507 | paid
qwen/qwen3-30b-a3b-thinking-2507 | paid qwen/qwen3-30b-a3b-thinking-2507 | paid
qwen/qwen3-32b | paid qwen/qwen3-32b | paid
qwen/qwen3-4b:free | free qwen/qwen3-4b:free | free
qwen/qwen3-8b | paid qwen/qwen3-8b | paid
qwen/qwen3-8b:free | free
qwen/qwen3-coder-30b-a3b-instruct | paid qwen/qwen3-coder-30b-a3b-instruct | paid
qwen/qwen3-coder | paid qwen/qwen3-coder | paid
qwen/qwen3-coder:exacto | paid
qwen/qwen3-coder:free | free qwen/qwen3-coder:free | free
qwen/qwen3-coder-flash | paid qwen/qwen3-coder-flash | paid
qwen/qwen3-coder-plus | paid qwen/qwen3-coder-plus | paid
@ -315,36 +333,35 @@ export interface IKBotOptions {
qwen/qwen3-vl-235b-a22b-thinking | paid qwen/qwen3-vl-235b-a22b-thinking | paid
qwen/qwen3-vl-30b-a3b-instruct | paid qwen/qwen3-vl-30b-a3b-instruct | paid
qwen/qwen3-vl-30b-a3b-thinking | paid qwen/qwen3-vl-30b-a3b-thinking | paid
qwen/qwen3-vl-32b-instruct | paid
qwen/qwen3-vl-8b-instruct | paid
qwen/qwen3-vl-8b-thinking | paid
qwen/qwq-32b | paid qwen/qwq-32b | paid
qwen/qwen-2.5-72b-instruct | paid qwen/qwen-2.5-72b-instruct | paid
qwen/qwen-2.5-72b-instruct:free | free
qwen/qwen-2.5-coder-32b-instruct | paid qwen/qwen-2.5-coder-32b-instruct | paid
qwen/qwen-2.5-coder-32b-instruct:free | free
relace/relace-apply-3 | paid relace/relace-apply-3 | paid
relace/relace-search | paid
undi95/remm-slerp-l2-13b | paid undi95/remm-slerp-l2-13b | paid
sao10k/l3-lunaris-8b | paid sao10k/l3-lunaris-8b | paid
sao10k/l3-euryale-70b | paid sao10k/l3-euryale-70b | paid
sao10k/l3.1-70b-hanami-x1 | paid sao10k/l3.1-70b-hanami-x1 | paid
sao10k/l3.1-euryale-70b | paid sao10k/l3.1-euryale-70b | paid
sao10k/l3.3-euryale-70b | paid sao10k/l3.3-euryale-70b | paid
shisa-ai/shisa-v2-llama3.3-70b | paid
shisa-ai/shisa-v2-llama3.3-70b:free | free
raifle/sorcererlm-8x22b | paid raifle/sorcererlm-8x22b | paid
stepfun-ai/step3 | paid stepfun-ai/step3 | paid
switchpoint/router | paid switchpoint/router | paid
tencent/hunyuan-a13b-instruct | paid tencent/hunyuan-a13b-instruct | paid
tencent/hunyuan-a13b-instruct:free | free
thedrummer/anubis-70b-v1.1 | paid
thedrummer/cydonia-24b-v4.1 | paid thedrummer/cydonia-24b-v4.1 | paid
thedrummer/rocinante-12b | paid thedrummer/rocinante-12b | paid
thedrummer/skyfall-36b-v2 | paid thedrummer/skyfall-36b-v2 | paid
thedrummer/unslopnemo-12b | paid thedrummer/unslopnemo-12b | paid
thudm/glm-4.1v-9b-thinking | paid thudm/glm-4.1v-9b-thinking | paid
thudm/glm-z1-32b | paid
tngtech/deepseek-r1t-chimera | paid tngtech/deepseek-r1t-chimera | paid
tngtech/deepseek-r1t-chimera:free | free tngtech/deepseek-r1t-chimera:free | free
tngtech/deepseek-r1t2-chimera | paid tngtech/deepseek-r1t2-chimera | paid
tngtech/deepseek-r1t2-chimera:free | free tngtech/deepseek-r1t2-chimera:free | free
tngtech/tng-r1t-chimera | paid
tngtech/tng-r1t-chimera:free | free
alibaba/tongyi-deepresearch-30b-a3b | paid alibaba/tongyi-deepresearch-30b-a3b | paid
alibaba/tongyi-deepresearch-30b-a3b:free | free alibaba/tongyi-deepresearch-30b-a3b:free | free
cognitivecomputations/dolphin-mistral-24b-venice-edition:free | free cognitivecomputations/dolphin-mistral-24b-venice-edition:free | free
@ -355,18 +372,24 @@ export interface IKBotOptions {
x-ai/grok-3-mini-beta | paid x-ai/grok-3-mini-beta | paid
x-ai/grok-4 | paid x-ai/grok-4 | paid
x-ai/grok-4-fast | paid x-ai/grok-4-fast | paid
x-ai/grok-4.1-fast | paid
x-ai/grok-code-fast-1 | paid x-ai/grok-code-fast-1 | paid
xiaomi/mimo-v2-flash:free | free
z-ai/glm-4-32b | paid z-ai/glm-4-32b | paid
z-ai/glm-4.5 | paid z-ai/glm-4.5 | paid
z-ai/glm-4.5-air | paid z-ai/glm-4.5-air | paid
z-ai/glm-4.5-air:free | free z-ai/glm-4.5-air:free | free
z-ai/glm-4.5v | paid z-ai/glm-4.5v | paid
z-ai/glm-4.6 | paid z-ai/glm-4.6 | paid
z-ai/glm-4.6:exacto | paid
z-ai/glm-4.6v | paid
z-ai/glm-4.7 | paid
 
 OpenAI models:  OpenAI models:
 
babbage-002 babbage-002
chatgpt-4o-latest chatgpt-4o-latest
chatgpt-image-latest
codex-mini-latest codex-mini-latest
dall-e-2 dall-e-2
dall-e-3 dall-e-3
@ -395,7 +418,6 @@ export interface IKBotOptions {
gpt-4o-2024-08-06 gpt-4o-2024-08-06
gpt-4o-2024-11-20 gpt-4o-2024-11-20
gpt-4o-audio-preview gpt-4o-audio-preview
gpt-4o-audio-preview-2024-10-01
gpt-4o-audio-preview-2024-12-17 gpt-4o-audio-preview-2024-12-17
gpt-4o-audio-preview-2025-06-03 gpt-4o-audio-preview-2025-06-03
gpt-4o-mini gpt-4o-mini
@ -407,14 +429,18 @@ export interface IKBotOptions {
gpt-4o-mini-search-preview gpt-4o-mini-search-preview
gpt-4o-mini-search-preview-2025-03-11 gpt-4o-mini-search-preview-2025-03-11
gpt-4o-mini-transcribe gpt-4o-mini-transcribe
gpt-4o-mini-transcribe-2025-03-20
gpt-4o-mini-transcribe-2025-12-15
gpt-4o-mini-tts gpt-4o-mini-tts
gpt-4o-mini-tts-2025-03-20
gpt-4o-mini-tts-2025-12-15
gpt-4o-realtime-preview gpt-4o-realtime-preview
gpt-4o-realtime-preview-2024-10-01
gpt-4o-realtime-preview-2024-12-17 gpt-4o-realtime-preview-2024-12-17
gpt-4o-realtime-preview-2025-06-03 gpt-4o-realtime-preview-2025-06-03
gpt-4o-search-preview gpt-4o-search-preview
gpt-4o-search-preview-2025-03-11 gpt-4o-search-preview-2025-03-11
gpt-4o-transcribe gpt-4o-transcribe
gpt-4o-transcribe-diarize
gpt-5 gpt-5
gpt-5-2025-08-07 gpt-5-2025-08-07
gpt-5-chat-latest gpt-5-chat-latest
@ -425,20 +451,34 @@ export interface IKBotOptions {
gpt-5-nano-2025-08-07 gpt-5-nano-2025-08-07
gpt-5-pro gpt-5-pro
gpt-5-pro-2025-10-06 gpt-5-pro-2025-10-06
gpt-5-search-api
gpt-5-search-api-2025-10-14
gpt-5.1
gpt-5.1-2025-11-13
gpt-5.1-chat-latest
gpt-5.1-codex
gpt-5.1-codex-max
gpt-5.1-codex-mini
gpt-5.2
gpt-5.2-2025-12-11
gpt-5.2-chat-latest
gpt-5.2-pro
gpt-5.2-pro-2025-12-11
gpt-audio gpt-audio
gpt-audio-2025-08-28 gpt-audio-2025-08-28
gpt-audio-mini gpt-audio-mini
gpt-audio-mini-2025-10-06 gpt-audio-mini-2025-10-06
gpt-audio-mini-2025-12-15
gpt-image-1 gpt-image-1
gpt-image-1-mini gpt-image-1-mini
gpt-image-1.5
gpt-realtime gpt-realtime
gpt-realtime-2025-08-28 gpt-realtime-2025-08-28
gpt-realtime-mini gpt-realtime-mini
gpt-realtime-mini-2025-10-06 gpt-realtime-mini-2025-10-06
gpt-realtime-mini-2025-12-15
o1 o1
o1-2024-12-17 o1-2024-12-17
o1-mini
o1-mini-2024-09-12
o1-pro o1-pro
o1-pro-2025-03-19 o1-pro-2025-03-19
o3 o3