maintainence love:)
This commit is contained in:
parent
0bc9799dde
commit
8ecd654eda
@ -1,5 +1,5 @@
|
||||
{
|
||||
"timestamp": 1766607018996,
|
||||
"timestamp": 1766607089404,
|
||||
"models": [
|
||||
{
|
||||
"id": "gpt-4-0613",
|
||||
|
||||
@ -1,5 +1,5 @@
|
||||
{
|
||||
"timestamp": 1766607019100,
|
||||
"timestamp": 1766607089536,
|
||||
"models": [
|
||||
{
|
||||
"id": "bytedance-seed/seed-1.6-flash",
|
||||
@ -14174,52 +14174,6 @@
|
||||
"frequency_penalty": null
|
||||
}
|
||||
},
|
||||
{
|
||||
"id": "anthropic/claude-3.5-sonnet",
|
||||
"canonical_slug": "anthropic/claude-3.5-sonnet",
|
||||
"hugging_face_id": null,
|
||||
"name": "Anthropic: Claude 3.5 Sonnet",
|
||||
"created": 1729555200,
|
||||
"description": "New Claude 3.5 Sonnet delivers better-than-Opus capabilities, faster-than-Sonnet speeds, at the same Sonnet prices. Sonnet is particularly good at:\n\n- Coding: Scores ~49% on SWE-Bench Verified, higher than the last best score, and without any fancy prompt scaffolding\n- Data science: Augments human data science expertise; navigates unstructured data while using multiple tools for insights\n- Visual processing: excelling at interpreting charts, graphs, and images, accurately transcribing text to derive insights beyond just the text alone\n- Agentic tasks: exceptional tool use, making it great at agentic tasks (i.e. complex, multi-step problem solving tasks that require engaging with other systems)\n\n#multimodal",
|
||||
"context_length": 200000,
|
||||
"architecture": {
|
||||
"modality": "text+image->text",
|
||||
"input_modalities": [
|
||||
"text",
|
||||
"image",
|
||||
"file"
|
||||
],
|
||||
"output_modalities": [
|
||||
"text"
|
||||
],
|
||||
"tokenizer": "Claude",
|
||||
"instruct_type": null
|
||||
},
|
||||
"pricing": {
|
||||
"prompt": "0.000006",
|
||||
"completion": "0.00003",
|
||||
"request": "0",
|
||||
"image": "0",
|
||||
"web_search": "0",
|
||||
"internal_reasoning": "0"
|
||||
},
|
||||
"top_provider": {
|
||||
"context_length": 200000,
|
||||
"max_completion_tokens": 8192,
|
||||
"is_moderated": true
|
||||
},
|
||||
"per_request_limits": null,
|
||||
"supported_parameters": [
|
||||
"max_tokens",
|
||||
"stop",
|
||||
"temperature",
|
||||
"tool_choice",
|
||||
"tools",
|
||||
"top_k",
|
||||
"top_p"
|
||||
],
|
||||
"default_parameters": {}
|
||||
},
|
||||
{
|
||||
"id": "anthracite-org/magnum-v4-72b",
|
||||
"canonical_slug": "anthracite-org/magnum-v4-72b",
|
||||
@ -14272,6 +14226,52 @@
|
||||
],
|
||||
"default_parameters": {}
|
||||
},
|
||||
{
|
||||
"id": "anthropic/claude-3.5-sonnet",
|
||||
"canonical_slug": "anthropic/claude-3.5-sonnet",
|
||||
"hugging_face_id": null,
|
||||
"name": "Anthropic: Claude 3.5 Sonnet",
|
||||
"created": 1729555200,
|
||||
"description": "New Claude 3.5 Sonnet delivers better-than-Opus capabilities, faster-than-Sonnet speeds, at the same Sonnet prices. Sonnet is particularly good at:\n\n- Coding: Scores ~49% on SWE-Bench Verified, higher than the last best score, and without any fancy prompt scaffolding\n- Data science: Augments human data science expertise; navigates unstructured data while using multiple tools for insights\n- Visual processing: excelling at interpreting charts, graphs, and images, accurately transcribing text to derive insights beyond just the text alone\n- Agentic tasks: exceptional tool use, making it great at agentic tasks (i.e. complex, multi-step problem solving tasks that require engaging with other systems)\n\n#multimodal",
|
||||
"context_length": 200000,
|
||||
"architecture": {
|
||||
"modality": "text+image->text",
|
||||
"input_modalities": [
|
||||
"text",
|
||||
"image",
|
||||
"file"
|
||||
],
|
||||
"output_modalities": [
|
||||
"text"
|
||||
],
|
||||
"tokenizer": "Claude",
|
||||
"instruct_type": null
|
||||
},
|
||||
"pricing": {
|
||||
"prompt": "0.000006",
|
||||
"completion": "0.00003",
|
||||
"request": "0",
|
||||
"image": "0",
|
||||
"web_search": "0",
|
||||
"internal_reasoning": "0"
|
||||
},
|
||||
"top_provider": {
|
||||
"context_length": 200000,
|
||||
"max_completion_tokens": 8192,
|
||||
"is_moderated": true
|
||||
},
|
||||
"per_request_limits": null,
|
||||
"supported_parameters": [
|
||||
"max_tokens",
|
||||
"stop",
|
||||
"temperature",
|
||||
"tool_choice",
|
||||
"tools",
|
||||
"top_k",
|
||||
"top_p"
|
||||
],
|
||||
"default_parameters": {}
|
||||
},
|
||||
{
|
||||
"id": "mistralai/ministral-8b",
|
||||
"canonical_slug": "mistralai/ministral-8b",
|
||||
@ -15097,6 +15097,57 @@
|
||||
],
|
||||
"default_parameters": {}
|
||||
},
|
||||
{
|
||||
"id": "sao10k/l3.1-euryale-70b",
|
||||
"canonical_slug": "sao10k/l3.1-euryale-70b",
|
||||
"hugging_face_id": "Sao10K/L3.1-70B-Euryale-v2.2",
|
||||
"name": "Sao10K: Llama 3.1 Euryale 70B v2.2",
|
||||
"created": 1724803200,
|
||||
"description": "Euryale L3.1 70B v2.2 is a model focused on creative roleplay from [Sao10k](https://ko-fi.com/sao10k). It is the successor of [Euryale L3 70B v2.1](/models/sao10k/l3-euryale-70b).",
|
||||
"context_length": 32768,
|
||||
"architecture": {
|
||||
"modality": "text->text",
|
||||
"input_modalities": [
|
||||
"text"
|
||||
],
|
||||
"output_modalities": [
|
||||
"text"
|
||||
],
|
||||
"tokenizer": "Llama3",
|
||||
"instruct_type": "llama3"
|
||||
},
|
||||
"pricing": {
|
||||
"prompt": "0.00000065",
|
||||
"completion": "0.00000075",
|
||||
"request": "0",
|
||||
"image": "0",
|
||||
"web_search": "0",
|
||||
"internal_reasoning": "0"
|
||||
},
|
||||
"top_provider": {
|
||||
"context_length": 32768,
|
||||
"max_completion_tokens": null,
|
||||
"is_moderated": false
|
||||
},
|
||||
"per_request_limits": null,
|
||||
"supported_parameters": [
|
||||
"frequency_penalty",
|
||||
"max_tokens",
|
||||
"min_p",
|
||||
"presence_penalty",
|
||||
"repetition_penalty",
|
||||
"response_format",
|
||||
"seed",
|
||||
"stop",
|
||||
"structured_outputs",
|
||||
"temperature",
|
||||
"tool_choice",
|
||||
"tools",
|
||||
"top_k",
|
||||
"top_p"
|
||||
],
|
||||
"default_parameters": {}
|
||||
},
|
||||
{
|
||||
"id": "qwen/qwen-2.5-vl-7b-instruct:free",
|
||||
"canonical_slug": "qwen/qwen-2-vl-7b-instruct",
|
||||
@ -15189,57 +15240,6 @@
|
||||
],
|
||||
"default_parameters": {}
|
||||
},
|
||||
{
|
||||
"id": "sao10k/l3.1-euryale-70b",
|
||||
"canonical_slug": "sao10k/l3.1-euryale-70b",
|
||||
"hugging_face_id": "Sao10K/L3.1-70B-Euryale-v2.2",
|
||||
"name": "Sao10K: Llama 3.1 Euryale 70B v2.2",
|
||||
"created": 1724803200,
|
||||
"description": "Euryale L3.1 70B v2.2 is a model focused on creative roleplay from [Sao10k](https://ko-fi.com/sao10k). It is the successor of [Euryale L3 70B v2.1](/models/sao10k/l3-euryale-70b).",
|
||||
"context_length": 32768,
|
||||
"architecture": {
|
||||
"modality": "text->text",
|
||||
"input_modalities": [
|
||||
"text"
|
||||
],
|
||||
"output_modalities": [
|
||||
"text"
|
||||
],
|
||||
"tokenizer": "Llama3",
|
||||
"instruct_type": "llama3"
|
||||
},
|
||||
"pricing": {
|
||||
"prompt": "0.00000065",
|
||||
"completion": "0.00000075",
|
||||
"request": "0",
|
||||
"image": "0",
|
||||
"web_search": "0",
|
||||
"internal_reasoning": "0"
|
||||
},
|
||||
"top_provider": {
|
||||
"context_length": 32768,
|
||||
"max_completion_tokens": null,
|
||||
"is_moderated": false
|
||||
},
|
||||
"per_request_limits": null,
|
||||
"supported_parameters": [
|
||||
"frequency_penalty",
|
||||
"max_tokens",
|
||||
"min_p",
|
||||
"presence_penalty",
|
||||
"repetition_penalty",
|
||||
"response_format",
|
||||
"seed",
|
||||
"stop",
|
||||
"structured_outputs",
|
||||
"temperature",
|
||||
"tool_choice",
|
||||
"tools",
|
||||
"top_k",
|
||||
"top_p"
|
||||
],
|
||||
"default_parameters": {}
|
||||
},
|
||||
{
|
||||
"id": "microsoft/phi-3.5-mini-128k-instruct",
|
||||
"canonical_slug": "microsoft/phi-3.5-mini-128k-instruct",
|
||||
@ -15625,57 +15625,6 @@
|
||||
],
|
||||
"default_parameters": {}
|
||||
},
|
||||
{
|
||||
"id": "meta-llama/llama-3.1-70b-instruct",
|
||||
"canonical_slug": "meta-llama/llama-3.1-70b-instruct",
|
||||
"hugging_face_id": "meta-llama/Meta-Llama-3.1-70B-Instruct",
|
||||
"name": "Meta: Llama 3.1 70B Instruct",
|
||||
"created": 1721692800,
|
||||
"description": "Meta's latest class of model (Llama 3.1) launched with a variety of sizes & flavors. This 70B instruct-tuned version is optimized for high quality dialogue usecases.\n\nIt has demonstrated strong performance compared to leading closed-source models in human evaluations.\n\nTo read more about the model release, [click here](https://ai.meta.com/blog/meta-llama-3-1/). Usage of this model is subject to [Meta's Acceptable Use Policy](https://llama.meta.com/llama3/use-policy/).",
|
||||
"context_length": 131072,
|
||||
"architecture": {
|
||||
"modality": "text->text",
|
||||
"input_modalities": [
|
||||
"text"
|
||||
],
|
||||
"output_modalities": [
|
||||
"text"
|
||||
],
|
||||
"tokenizer": "Llama3",
|
||||
"instruct_type": "llama3"
|
||||
},
|
||||
"pricing": {
|
||||
"prompt": "0.0000004",
|
||||
"completion": "0.0000004",
|
||||
"request": "0",
|
||||
"image": "0",
|
||||
"web_search": "0",
|
||||
"internal_reasoning": "0"
|
||||
},
|
||||
"top_provider": {
|
||||
"context_length": 131072,
|
||||
"max_completion_tokens": null,
|
||||
"is_moderated": false
|
||||
},
|
||||
"per_request_limits": null,
|
||||
"supported_parameters": [
|
||||
"frequency_penalty",
|
||||
"logit_bias",
|
||||
"max_tokens",
|
||||
"min_p",
|
||||
"presence_penalty",
|
||||
"repetition_penalty",
|
||||
"response_format",
|
||||
"seed",
|
||||
"stop",
|
||||
"temperature",
|
||||
"tool_choice",
|
||||
"tools",
|
||||
"top_k",
|
||||
"top_p"
|
||||
],
|
||||
"default_parameters": {}
|
||||
},
|
||||
{
|
||||
"id": "meta-llama/llama-3.1-8b-instruct",
|
||||
"canonical_slug": "meta-llama/llama-3.1-8b-instruct",
|
||||
@ -15824,6 +15773,57 @@
|
||||
],
|
||||
"default_parameters": {}
|
||||
},
|
||||
{
|
||||
"id": "meta-llama/llama-3.1-70b-instruct",
|
||||
"canonical_slug": "meta-llama/llama-3.1-70b-instruct",
|
||||
"hugging_face_id": "meta-llama/Meta-Llama-3.1-70B-Instruct",
|
||||
"name": "Meta: Llama 3.1 70B Instruct",
|
||||
"created": 1721692800,
|
||||
"description": "Meta's latest class of model (Llama 3.1) launched with a variety of sizes & flavors. This 70B instruct-tuned version is optimized for high quality dialogue usecases.\n\nIt has demonstrated strong performance compared to leading closed-source models in human evaluations.\n\nTo read more about the model release, [click here](https://ai.meta.com/blog/meta-llama-3-1/). Usage of this model is subject to [Meta's Acceptable Use Policy](https://llama.meta.com/llama3/use-policy/).",
|
||||
"context_length": 131072,
|
||||
"architecture": {
|
||||
"modality": "text->text",
|
||||
"input_modalities": [
|
||||
"text"
|
||||
],
|
||||
"output_modalities": [
|
||||
"text"
|
||||
],
|
||||
"tokenizer": "Llama3",
|
||||
"instruct_type": "llama3"
|
||||
},
|
||||
"pricing": {
|
||||
"prompt": "0.0000004",
|
||||
"completion": "0.0000004",
|
||||
"request": "0",
|
||||
"image": "0",
|
||||
"web_search": "0",
|
||||
"internal_reasoning": "0"
|
||||
},
|
||||
"top_provider": {
|
||||
"context_length": 131072,
|
||||
"max_completion_tokens": null,
|
||||
"is_moderated": false
|
||||
},
|
||||
"per_request_limits": null,
|
||||
"supported_parameters": [
|
||||
"frequency_penalty",
|
||||
"logit_bias",
|
||||
"max_tokens",
|
||||
"min_p",
|
||||
"presence_penalty",
|
||||
"repetition_penalty",
|
||||
"response_format",
|
||||
"seed",
|
||||
"stop",
|
||||
"temperature",
|
||||
"tool_choice",
|
||||
"tools",
|
||||
"top_k",
|
||||
"top_p"
|
||||
],
|
||||
"default_parameters": {}
|
||||
},
|
||||
{
|
||||
"id": "mistralai/mistral-nemo",
|
||||
"canonical_slug": "mistralai/mistral-nemo",
|
||||
@ -16124,6 +16124,54 @@
|
||||
],
|
||||
"default_parameters": {}
|
||||
},
|
||||
{
|
||||
"id": "nousresearch/hermes-2-pro-llama-3-8b",
|
||||
"canonical_slug": "nousresearch/hermes-2-pro-llama-3-8b",
|
||||
"hugging_face_id": "NousResearch/Hermes-2-Pro-Llama-3-8B",
|
||||
"name": "NousResearch: Hermes 2 Pro - Llama-3 8B",
|
||||
"created": 1716768000,
|
||||
"description": "Hermes 2 Pro is an upgraded, retrained version of Nous Hermes 2, consisting of an updated and cleaned version of the OpenHermes 2.5 Dataset, as well as a newly introduced Function Calling and JSON Mode dataset developed in-house.",
|
||||
"context_length": 8192,
|
||||
"architecture": {
|
||||
"modality": "text->text",
|
||||
"input_modalities": [
|
||||
"text"
|
||||
],
|
||||
"output_modalities": [
|
||||
"text"
|
||||
],
|
||||
"tokenizer": "Llama3",
|
||||
"instruct_type": "chatml"
|
||||
},
|
||||
"pricing": {
|
||||
"prompt": "0.000000025",
|
||||
"completion": "0.00000008",
|
||||
"request": "0",
|
||||
"image": "0",
|
||||
"web_search": "0",
|
||||
"internal_reasoning": "0"
|
||||
},
|
||||
"top_provider": {
|
||||
"context_length": 8192,
|
||||
"max_completion_tokens": 2048,
|
||||
"is_moderated": false
|
||||
},
|
||||
"per_request_limits": null,
|
||||
"supported_parameters": [
|
||||
"frequency_penalty",
|
||||
"max_tokens",
|
||||
"presence_penalty",
|
||||
"repetition_penalty",
|
||||
"response_format",
|
||||
"seed",
|
||||
"stop",
|
||||
"structured_outputs",
|
||||
"temperature",
|
||||
"top_k",
|
||||
"top_p"
|
||||
],
|
||||
"default_parameters": {}
|
||||
},
|
||||
{
|
||||
"id": "mistralai/mistral-7b-instruct:free",
|
||||
"canonical_slug": "mistralai/mistral-7b-instruct",
|
||||
@ -16278,54 +16326,6 @@
|
||||
"temperature": 0.3
|
||||
}
|
||||
},
|
||||
{
|
||||
"id": "nousresearch/hermes-2-pro-llama-3-8b",
|
||||
"canonical_slug": "nousresearch/hermes-2-pro-llama-3-8b",
|
||||
"hugging_face_id": "NousResearch/Hermes-2-Pro-Llama-3-8B",
|
||||
"name": "NousResearch: Hermes 2 Pro - Llama-3 8B",
|
||||
"created": 1716768000,
|
||||
"description": "Hermes 2 Pro is an upgraded, retrained version of Nous Hermes 2, consisting of an updated and cleaned version of the OpenHermes 2.5 Dataset, as well as a newly introduced Function Calling and JSON Mode dataset developed in-house.",
|
||||
"context_length": 8192,
|
||||
"architecture": {
|
||||
"modality": "text->text",
|
||||
"input_modalities": [
|
||||
"text"
|
||||
],
|
||||
"output_modalities": [
|
||||
"text"
|
||||
],
|
||||
"tokenizer": "Llama3",
|
||||
"instruct_type": "chatml"
|
||||
},
|
||||
"pricing": {
|
||||
"prompt": "0.000000025",
|
||||
"completion": "0.00000008",
|
||||
"request": "0",
|
||||
"image": "0",
|
||||
"web_search": "0",
|
||||
"internal_reasoning": "0"
|
||||
},
|
||||
"top_provider": {
|
||||
"context_length": 8192,
|
||||
"max_completion_tokens": 2048,
|
||||
"is_moderated": false
|
||||
},
|
||||
"per_request_limits": null,
|
||||
"supported_parameters": [
|
||||
"frequency_penalty",
|
||||
"max_tokens",
|
||||
"presence_penalty",
|
||||
"repetition_penalty",
|
||||
"response_format",
|
||||
"seed",
|
||||
"stop",
|
||||
"structured_outputs",
|
||||
"temperature",
|
||||
"top_k",
|
||||
"top_p"
|
||||
],
|
||||
"default_parameters": {}
|
||||
},
|
||||
{
|
||||
"id": "microsoft/phi-3-mini-128k-instruct",
|
||||
"canonical_slug": "microsoft/phi-3-mini-128k-instruct",
|
||||
@ -16410,6 +16410,107 @@
|
||||
],
|
||||
"default_parameters": {}
|
||||
},
|
||||
{
|
||||
"id": "meta-llama/llama-guard-2-8b",
|
||||
"canonical_slug": "meta-llama/llama-guard-2-8b",
|
||||
"hugging_face_id": "meta-llama/Meta-Llama-Guard-2-8B",
|
||||
"name": "Meta: LlamaGuard 2 8B",
|
||||
"created": 1715558400,
|
||||
"description": "This safeguard model has 8B parameters and is based on the Llama 3 family. Just like is predecessor, [LlamaGuard 1](https://huggingface.co/meta-llama/LlamaGuard-7b), it can do both prompt and response classification.\n\nLlamaGuard 2 acts as a normal LLM would, generating text that indicates whether the given input/output is safe/unsafe. If deemed unsafe, it will also share the content categories violated.\n\nFor best results, please use raw prompt input or the `/completions` endpoint, instead of the chat API.\n\nIt has demonstrated strong performance compared to leading closed-source models in human evaluations.\n\nTo read more about the model release, [click here](https://ai.meta.com/blog/meta-llama-3/). Usage of this model is subject to [Meta's Acceptable Use Policy](https://llama.meta.com/llama3/use-policy/).",
|
||||
"context_length": 8192,
|
||||
"architecture": {
|
||||
"modality": "text->text",
|
||||
"input_modalities": [
|
||||
"text"
|
||||
],
|
||||
"output_modalities": [
|
||||
"text"
|
||||
],
|
||||
"tokenizer": "Llama3",
|
||||
"instruct_type": "none"
|
||||
},
|
||||
"pricing": {
|
||||
"prompt": "0.0000002",
|
||||
"completion": "0.0000002",
|
||||
"request": "0",
|
||||
"image": "0",
|
||||
"web_search": "0",
|
||||
"internal_reasoning": "0"
|
||||
},
|
||||
"top_provider": {
|
||||
"context_length": 8192,
|
||||
"max_completion_tokens": null,
|
||||
"is_moderated": false
|
||||
},
|
||||
"per_request_limits": null,
|
||||
"supported_parameters": [
|
||||
"frequency_penalty",
|
||||
"logit_bias",
|
||||
"max_tokens",
|
||||
"min_p",
|
||||
"presence_penalty",
|
||||
"repetition_penalty",
|
||||
"stop",
|
||||
"temperature",
|
||||
"top_k",
|
||||
"top_p"
|
||||
],
|
||||
"default_parameters": {}
|
||||
},
|
||||
{
|
||||
"id": "openai/gpt-4o-2024-05-13",
|
||||
"canonical_slug": "openai/gpt-4o-2024-05-13",
|
||||
"hugging_face_id": null,
|
||||
"name": "OpenAI: GPT-4o (2024-05-13)",
|
||||
"created": 1715558400,
|
||||
"description": "GPT-4o (\"o\" for \"omni\") is OpenAI's latest AI model, supporting both text and image inputs with text outputs. It maintains the intelligence level of [GPT-4 Turbo](/models/openai/gpt-4-turbo) while being twice as fast and 50% more cost-effective. GPT-4o also offers improved performance in processing non-English languages and enhanced visual capabilities.\n\nFor benchmarking against other models, it was briefly called [\"im-also-a-good-gpt2-chatbot\"](https://twitter.com/LiamFedus/status/1790064963966370209)\n\n#multimodal",
|
||||
"context_length": 128000,
|
||||
"architecture": {
|
||||
"modality": "text+image->text",
|
||||
"input_modalities": [
|
||||
"text",
|
||||
"image",
|
||||
"file"
|
||||
],
|
||||
"output_modalities": [
|
||||
"text"
|
||||
],
|
||||
"tokenizer": "GPT",
|
||||
"instruct_type": null
|
||||
},
|
||||
"pricing": {
|
||||
"prompt": "0.000005",
|
||||
"completion": "0.000015",
|
||||
"request": "0",
|
||||
"image": "0.007225",
|
||||
"web_search": "0",
|
||||
"internal_reasoning": "0"
|
||||
},
|
||||
"top_provider": {
|
||||
"context_length": 128000,
|
||||
"max_completion_tokens": 4096,
|
||||
"is_moderated": true
|
||||
},
|
||||
"per_request_limits": null,
|
||||
"supported_parameters": [
|
||||
"frequency_penalty",
|
||||
"logit_bias",
|
||||
"logprobs",
|
||||
"max_tokens",
|
||||
"presence_penalty",
|
||||
"response_format",
|
||||
"seed",
|
||||
"stop",
|
||||
"structured_outputs",
|
||||
"temperature",
|
||||
"tool_choice",
|
||||
"tools",
|
||||
"top_logprobs",
|
||||
"top_p",
|
||||
"web_search_options"
|
||||
],
|
||||
"default_parameters": {}
|
||||
},
|
||||
{
|
||||
"id": "openai/gpt-4o",
|
||||
"canonical_slug": "openai/gpt-4o",
|
||||
@ -16519,107 +16620,6 @@
|
||||
],
|
||||
"default_parameters": {}
|
||||
},
|
||||
{
|
||||
"id": "openai/gpt-4o-2024-05-13",
|
||||
"canonical_slug": "openai/gpt-4o-2024-05-13",
|
||||
"hugging_face_id": null,
|
||||
"name": "OpenAI: GPT-4o (2024-05-13)",
|
||||
"created": 1715558400,
|
||||
"description": "GPT-4o (\"o\" for \"omni\") is OpenAI's latest AI model, supporting both text and image inputs with text outputs. It maintains the intelligence level of [GPT-4 Turbo](/models/openai/gpt-4-turbo) while being twice as fast and 50% more cost-effective. GPT-4o also offers improved performance in processing non-English languages and enhanced visual capabilities.\n\nFor benchmarking against other models, it was briefly called [\"im-also-a-good-gpt2-chatbot\"](https://twitter.com/LiamFedus/status/1790064963966370209)\n\n#multimodal",
|
||||
"context_length": 128000,
|
||||
"architecture": {
|
||||
"modality": "text+image->text",
|
||||
"input_modalities": [
|
||||
"text",
|
||||
"image",
|
||||
"file"
|
||||
],
|
||||
"output_modalities": [
|
||||
"text"
|
||||
],
|
||||
"tokenizer": "GPT",
|
||||
"instruct_type": null
|
||||
},
|
||||
"pricing": {
|
||||
"prompt": "0.000005",
|
||||
"completion": "0.000015",
|
||||
"request": "0",
|
||||
"image": "0.007225",
|
||||
"web_search": "0",
|
||||
"internal_reasoning": "0"
|
||||
},
|
||||
"top_provider": {
|
||||
"context_length": 128000,
|
||||
"max_completion_tokens": 4096,
|
||||
"is_moderated": true
|
||||
},
|
||||
"per_request_limits": null,
|
||||
"supported_parameters": [
|
||||
"frequency_penalty",
|
||||
"logit_bias",
|
||||
"logprobs",
|
||||
"max_tokens",
|
||||
"presence_penalty",
|
||||
"response_format",
|
||||
"seed",
|
||||
"stop",
|
||||
"structured_outputs",
|
||||
"temperature",
|
||||
"tool_choice",
|
||||
"tools",
|
||||
"top_logprobs",
|
||||
"top_p",
|
||||
"web_search_options"
|
||||
],
|
||||
"default_parameters": {}
|
||||
},
|
||||
{
|
||||
"id": "meta-llama/llama-guard-2-8b",
|
||||
"canonical_slug": "meta-llama/llama-guard-2-8b",
|
||||
"hugging_face_id": "meta-llama/Meta-Llama-Guard-2-8B",
|
||||
"name": "Meta: LlamaGuard 2 8B",
|
||||
"created": 1715558400,
|
||||
"description": "This safeguard model has 8B parameters and is based on the Llama 3 family. Just like is predecessor, [LlamaGuard 1](https://huggingface.co/meta-llama/LlamaGuard-7b), it can do both prompt and response classification.\n\nLlamaGuard 2 acts as a normal LLM would, generating text that indicates whether the given input/output is safe/unsafe. If deemed unsafe, it will also share the content categories violated.\n\nFor best results, please use raw prompt input or the `/completions` endpoint, instead of the chat API.\n\nIt has demonstrated strong performance compared to leading closed-source models in human evaluations.\n\nTo read more about the model release, [click here](https://ai.meta.com/blog/meta-llama-3/). Usage of this model is subject to [Meta's Acceptable Use Policy](https://llama.meta.com/llama3/use-policy/).",
|
||||
"context_length": 8192,
|
||||
"architecture": {
|
||||
"modality": "text->text",
|
||||
"input_modalities": [
|
||||
"text"
|
||||
],
|
||||
"output_modalities": [
|
||||
"text"
|
||||
],
|
||||
"tokenizer": "Llama3",
|
||||
"instruct_type": "none"
|
||||
},
|
||||
"pricing": {
|
||||
"prompt": "0.0000002",
|
||||
"completion": "0.0000002",
|
||||
"request": "0",
|
||||
"image": "0",
|
||||
"web_search": "0",
|
||||
"internal_reasoning": "0"
|
||||
},
|
||||
"top_provider": {
|
||||
"context_length": 8192,
|
||||
"max_completion_tokens": null,
|
||||
"is_moderated": false
|
||||
},
|
||||
"per_request_limits": null,
|
||||
"supported_parameters": [
|
||||
"frequency_penalty",
|
||||
"logit_bias",
|
||||
"max_tokens",
|
||||
"min_p",
|
||||
"presence_penalty",
|
||||
"repetition_penalty",
|
||||
"stop",
|
||||
"temperature",
|
||||
"top_k",
|
||||
"top_p"
|
||||
],
|
||||
"default_parameters": {}
|
||||
},
|
||||
{
|
||||
"id": "meta-llama/llama-3-70b-instruct",
|
||||
"canonical_slug": "meta-llama/llama-3-70b-instruct",
|
||||
@ -17017,57 +17017,6 @@
|
||||
"temperature": 0.3
|
||||
}
|
||||
},
|
||||
{
|
||||
"id": "openai/gpt-4-turbo-preview",
|
||||
"canonical_slug": "openai/gpt-4-turbo-preview",
|
||||
"hugging_face_id": null,
|
||||
"name": "OpenAI: GPT-4 Turbo Preview",
|
||||
"created": 1706140800,
|
||||
"description": "The preview GPT-4 model with improved instruction following, JSON mode, reproducible outputs, parallel function calling, and more. Training data: up to Dec 2023.\n\n**Note:** heavily rate limited by OpenAI while in preview.",
|
||||
"context_length": 128000,
|
||||
"architecture": {
|
||||
"modality": "text->text",
|
||||
"input_modalities": [
|
||||
"text"
|
||||
],
|
||||
"output_modalities": [
|
||||
"text"
|
||||
],
|
||||
"tokenizer": "GPT",
|
||||
"instruct_type": null
|
||||
},
|
||||
"pricing": {
|
||||
"prompt": "0.00001",
|
||||
"completion": "0.00003",
|
||||
"request": "0",
|
||||
"image": "0",
|
||||
"web_search": "0",
|
||||
"internal_reasoning": "0"
|
||||
},
|
||||
"top_provider": {
|
||||
"context_length": 128000,
|
||||
"max_completion_tokens": 4096,
|
||||
"is_moderated": true
|
||||
},
|
||||
"per_request_limits": null,
|
||||
"supported_parameters": [
|
||||
"frequency_penalty",
|
||||
"logit_bias",
|
||||
"logprobs",
|
||||
"max_tokens",
|
||||
"presence_penalty",
|
||||
"response_format",
|
||||
"seed",
|
||||
"stop",
|
||||
"structured_outputs",
|
||||
"temperature",
|
||||
"tool_choice",
|
||||
"tools",
|
||||
"top_logprobs",
|
||||
"top_p"
|
||||
],
|
||||
"default_parameters": {}
|
||||
},
|
||||
{
|
||||
"id": "openai/gpt-3.5-turbo-0613",
|
||||
"canonical_slug": "openai/gpt-3.5-turbo-0613",
|
||||
@ -17119,6 +17068,57 @@
|
||||
],
|
||||
"default_parameters": {}
|
||||
},
|
||||
{
|
||||
"id": "openai/gpt-4-turbo-preview",
|
||||
"canonical_slug": "openai/gpt-4-turbo-preview",
|
||||
"hugging_face_id": null,
|
||||
"name": "OpenAI: GPT-4 Turbo Preview",
|
||||
"created": 1706140800,
|
||||
"description": "The preview GPT-4 model with improved instruction following, JSON mode, reproducible outputs, parallel function calling, and more. Training data: up to Dec 2023.\n\n**Note:** heavily rate limited by OpenAI while in preview.",
|
||||
"context_length": 128000,
|
||||
"architecture": {
|
||||
"modality": "text->text",
|
||||
"input_modalities": [
|
||||
"text"
|
||||
],
|
||||
"output_modalities": [
|
||||
"text"
|
||||
],
|
||||
"tokenizer": "GPT",
|
||||
"instruct_type": null
|
||||
},
|
||||
"pricing": {
|
||||
"prompt": "0.00001",
|
||||
"completion": "0.00003",
|
||||
"request": "0",
|
||||
"image": "0",
|
||||
"web_search": "0",
|
||||
"internal_reasoning": "0"
|
||||
},
|
||||
"top_provider": {
|
||||
"context_length": 128000,
|
||||
"max_completion_tokens": 4096,
|
||||
"is_moderated": true
|
||||
},
|
||||
"per_request_limits": null,
|
||||
"supported_parameters": [
|
||||
"frequency_penalty",
|
||||
"logit_bias",
|
||||
"logprobs",
|
||||
"max_tokens",
|
||||
"presence_penalty",
|
||||
"response_format",
|
||||
"seed",
|
||||
"stop",
|
||||
"structured_outputs",
|
||||
"temperature",
|
||||
"tool_choice",
|
||||
"tools",
|
||||
"top_logprobs",
|
||||
"top_p"
|
||||
],
|
||||
"default_parameters": {}
|
||||
},
|
||||
{
|
||||
"id": "mistralai/mistral-tiny",
|
||||
"canonical_slug": "mistralai/mistral-tiny",
|
||||
|
||||
File diff suppressed because one or more lines are too long
4
packages/kbot/dist/package-lock.json
generated
vendored
4
packages/kbot/dist/package-lock.json
generated
vendored
@ -1,12 +1,12 @@
|
||||
{
|
||||
"name": "@plastichub/kbot",
|
||||
"version": "1.1.56",
|
||||
"version": "1.1.57",
|
||||
"lockfileVersion": 3,
|
||||
"requires": true,
|
||||
"packages": {
|
||||
"": {
|
||||
"name": "@plastichub/kbot",
|
||||
"version": "1.1.56",
|
||||
"version": "1.1.57",
|
||||
"license": "ISC",
|
||||
"dependencies": {
|
||||
"node-emoji": "^2.2.0"
|
||||
|
||||
2
packages/kbot/dist/package.json
vendored
2
packages/kbot/dist/package.json
vendored
@ -1,6 +1,6 @@
|
||||
{
|
||||
"name": "@plastichub/kbot",
|
||||
"version": "1.1.56",
|
||||
"version": "1.1.57",
|
||||
"main": "main_node.js",
|
||||
"author": "",
|
||||
"license": "ISC",
|
||||
|
||||
File diff suppressed because one or more lines are too long
File diff suppressed because one or more lines are too long
@ -276,8 +276,8 @@ export enum E_OPENROUTER_MODEL {
|
||||
MODEL_THEDRUMMER_UNSLOPNEMO_12B = "thedrummer/unslopnemo-12b",
|
||||
MODEL_ANTHROPIC_CLAUDE_3_5_HAIKU_20241022 = "anthropic/claude-3.5-haiku-20241022",
|
||||
MODEL_ANTHROPIC_CLAUDE_3_5_HAIKU = "anthropic/claude-3.5-haiku",
|
||||
MODEL_ANTHROPIC_CLAUDE_3_5_SONNET = "anthropic/claude-3.5-sonnet",
|
||||
MODEL_ANTHRACITE_ORG_MAGNUM_V4_72B = "anthracite-org/magnum-v4-72b",
|
||||
MODEL_ANTHROPIC_CLAUDE_3_5_SONNET = "anthropic/claude-3.5-sonnet",
|
||||
MODEL_MISTRALAI_MINISTRAL_8B = "mistralai/ministral-8b",
|
||||
MODEL_MISTRALAI_MINISTRAL_3B = "mistralai/ministral-3b",
|
||||
MODEL_QWEN_QWEN_2_5_7B_INSTRUCT = "qwen/qwen-2.5-7b-instruct",
|
||||
@ -295,9 +295,9 @@ export enum E_OPENROUTER_MODEL {
|
||||
MODEL_MISTRALAI_PIXTRAL_12B = "mistralai/pixtral-12b",
|
||||
MODEL_COHERE_COMMAND_R_08_2024 = "cohere/command-r-08-2024",
|
||||
MODEL_COHERE_COMMAND_R_PLUS_08_2024 = "cohere/command-r-plus-08-2024",
|
||||
MODEL_SAO10K_L3_1_EURYALE_70B = "sao10k/l3.1-euryale-70b",
|
||||
MODEL_QWEN_QWEN_2_5_VL_7B_INSTRUCT_FREE = "qwen/qwen-2.5-vl-7b-instruct:free",
|
||||
MODEL_QWEN_QWEN_2_5_VL_7B_INSTRUCT = "qwen/qwen-2.5-vl-7b-instruct",
|
||||
MODEL_SAO10K_L3_1_EURYALE_70B = "sao10k/l3.1-euryale-70b",
|
||||
MODEL_MICROSOFT_PHI_3_5_MINI_128K_INSTRUCT = "microsoft/phi-3.5-mini-128k-instruct",
|
||||
MODEL_NOUSRESEARCH_HERMES_3_LLAMA_3_1_70B = "nousresearch/hermes-3-llama-3.1-70b",
|
||||
MODEL_NOUSRESEARCH_HERMES_3_LLAMA_3_1_405B_FREE = "nousresearch/hermes-3-llama-3.1-405b:free",
|
||||
@ -306,26 +306,26 @@ export enum E_OPENROUTER_MODEL {
|
||||
MODEL_SAO10K_L3_LUNARIS_8B = "sao10k/l3-lunaris-8b",
|
||||
MODEL_OPENAI_GPT_4O_2024_08_06 = "openai/gpt-4o-2024-08-06",
|
||||
MODEL_META_LLAMA_LLAMA_3_1_405B = "meta-llama/llama-3.1-405b",
|
||||
MODEL_META_LLAMA_LLAMA_3_1_70B_INSTRUCT = "meta-llama/llama-3.1-70b-instruct",
|
||||
MODEL_META_LLAMA_LLAMA_3_1_8B_INSTRUCT = "meta-llama/llama-3.1-8b-instruct",
|
||||
MODEL_META_LLAMA_LLAMA_3_1_405B_INSTRUCT_FREE = "meta-llama/llama-3.1-405b-instruct:free",
|
||||
MODEL_META_LLAMA_LLAMA_3_1_405B_INSTRUCT = "meta-llama/llama-3.1-405b-instruct",
|
||||
MODEL_META_LLAMA_LLAMA_3_1_70B_INSTRUCT = "meta-llama/llama-3.1-70b-instruct",
|
||||
MODEL_MISTRALAI_MISTRAL_NEMO = "mistralai/mistral-nemo",
|
||||
MODEL_OPENAI_GPT_4O_MINI_2024_07_18 = "openai/gpt-4o-mini-2024-07-18",
|
||||
MODEL_OPENAI_GPT_4O_MINI = "openai/gpt-4o-mini",
|
||||
MODEL_GOOGLE_GEMMA_2_27B_IT = "google/gemma-2-27b-it",
|
||||
MODEL_GOOGLE_GEMMA_2_9B_IT = "google/gemma-2-9b-it",
|
||||
MODEL_SAO10K_L3_EURYALE_70B = "sao10k/l3-euryale-70b",
|
||||
MODEL_NOUSRESEARCH_HERMES_2_PRO_LLAMA_3_8B = "nousresearch/hermes-2-pro-llama-3-8b",
|
||||
MODEL_MISTRALAI_MISTRAL_7B_INSTRUCT_FREE = "mistralai/mistral-7b-instruct:free",
|
||||
MODEL_MISTRALAI_MISTRAL_7B_INSTRUCT = "mistralai/mistral-7b-instruct",
|
||||
MODEL_MISTRALAI_MISTRAL_7B_INSTRUCT_V0_3 = "mistralai/mistral-7b-instruct-v0.3",
|
||||
MODEL_NOUSRESEARCH_HERMES_2_PRO_LLAMA_3_8B = "nousresearch/hermes-2-pro-llama-3-8b",
|
||||
MODEL_MICROSOFT_PHI_3_MINI_128K_INSTRUCT = "microsoft/phi-3-mini-128k-instruct",
|
||||
MODEL_MICROSOFT_PHI_3_MEDIUM_128K_INSTRUCT = "microsoft/phi-3-medium-128k-instruct",
|
||||
MODEL_META_LLAMA_LLAMA_GUARD_2_8B = "meta-llama/llama-guard-2-8b",
|
||||
MODEL_OPENAI_GPT_4O_2024_05_13 = "openai/gpt-4o-2024-05-13",
|
||||
MODEL_OPENAI_GPT_4O = "openai/gpt-4o",
|
||||
MODEL_OPENAI_GPT_4O_EXTENDED = "openai/gpt-4o:extended",
|
||||
MODEL_OPENAI_GPT_4O_2024_05_13 = "openai/gpt-4o-2024-05-13",
|
||||
MODEL_META_LLAMA_LLAMA_GUARD_2_8B = "meta-llama/llama-guard-2-8b",
|
||||
MODEL_META_LLAMA_LLAMA_3_70B_INSTRUCT = "meta-llama/llama-3-70b-instruct",
|
||||
MODEL_META_LLAMA_LLAMA_3_8B_INSTRUCT = "meta-llama/llama-3-8b-instruct",
|
||||
MODEL_MISTRALAI_MIXTRAL_8X22B_INSTRUCT = "mistralai/mixtral-8x22b-instruct",
|
||||
@ -334,8 +334,8 @@ export enum E_OPENROUTER_MODEL {
|
||||
MODEL_ANTHROPIC_CLAUDE_3_HAIKU = "anthropic/claude-3-haiku",
|
||||
MODEL_ANTHROPIC_CLAUDE_3_OPUS = "anthropic/claude-3-opus",
|
||||
MODEL_MISTRALAI_MISTRAL_LARGE = "mistralai/mistral-large",
|
||||
MODEL_OPENAI_GPT_4_TURBO_PREVIEW = "openai/gpt-4-turbo-preview",
|
||||
MODEL_OPENAI_GPT_3_5_TURBO_0613 = "openai/gpt-3.5-turbo-0613",
|
||||
MODEL_OPENAI_GPT_4_TURBO_PREVIEW = "openai/gpt-4-turbo-preview",
|
||||
MODEL_MISTRALAI_MISTRAL_TINY = "mistralai/mistral-tiny",
|
||||
MODEL_MISTRALAI_MISTRAL_7B_INSTRUCT_V0_2 = "mistralai/mistral-7b-instruct-v0.2",
|
||||
MODEL_MISTRALAI_MIXTRAL_8X7B_INSTRUCT = "mistralai/mixtral-8x7b-instruct",
|
||||
|
||||
@ -31,72 +31,78 @@ export interface IKBotOptions {
|
||||
[35m[1m[22m[39m
|
||||
[35m[1m OpenRouter models:[22m[39m
|
||||
[35m[1m[22m[39m
|
||||
agentica-org/deepcoder-14b-preview | paid
|
||||
agentica-org/deepcoder-14b-preview:free | free
|
||||
ai21/jamba-large-1.7 | paid
|
||||
ai21/jamba-mini-1.7 | paid
|
||||
aion-labs/aion-1.0 | paid
|
||||
aion-labs/aion-1.0-mini | paid
|
||||
aion-labs/aion-rp-llama-3.1-8b | paid
|
||||
alfredpros/codellama-7b-instruct-solidity | paid
|
||||
allenai/molmo-7b-d | paid
|
||||
allenai/olmo-2-0325-32b-instruct | paid
|
||||
allenai/olmo-3-32b-think:free | free
|
||||
allenai/olmo-3-7b-instruct | paid
|
||||
allenai/olmo-3-7b-think | paid
|
||||
allenai/olmo-3.1-32b-think:free | free
|
||||
amazon/nova-2-lite-v1 | paid
|
||||
amazon/nova-lite-v1 | paid
|
||||
amazon/nova-micro-v1 | paid
|
||||
amazon/nova-premier-v1 | paid
|
||||
amazon/nova-pro-v1 | paid
|
||||
anthropic/claude-3-haiku | paid
|
||||
anthropic/claude-3-opus | paid
|
||||
anthropic/claude-3.5-haiku | paid
|
||||
anthropic/claude-3.5-haiku-20241022 | paid
|
||||
anthropic/claude-3.5-sonnet | paid
|
||||
anthropic/claude-3.5-sonnet-20240620 | paid
|
||||
anthropic/claude-3.7-sonnet | paid
|
||||
anthropic/claude-3.7-sonnet:thinking | paid
|
||||
anthropic/claude-haiku-4.5 | paid
|
||||
anthropic/claude-opus-4 | paid
|
||||
anthropic/claude-opus-4.1 | paid
|
||||
anthropic/claude-opus-4.5 | paid
|
||||
anthropic/claude-sonnet-4 | paid
|
||||
anthropic/claude-sonnet-4.5 | paid
|
||||
arcee-ai/afm-4.5b | paid
|
||||
arcee-ai/coder-large | paid
|
||||
arcee-ai/maestro-reasoning | paid
|
||||
arcee-ai/spotlight | paid
|
||||
arcee-ai/trinity-mini | paid
|
||||
arcee-ai/trinity-mini:free | free
|
||||
arcee-ai/virtuoso-large | paid
|
||||
arliai/qwq-32b-arliai-rpr-v1 | paid
|
||||
arliai/qwq-32b-arliai-rpr-v1:free | free
|
||||
openrouter/auto | paid
|
||||
baidu/ernie-4.5-21b-a3b | paid
|
||||
baidu/ernie-4.5-21b-a3b-thinking | paid
|
||||
baidu/ernie-4.5-300b-a47b | paid
|
||||
baidu/ernie-4.5-vl-28b-a3b | paid
|
||||
baidu/ernie-4.5-vl-424b-a47b | paid
|
||||
openrouter/bodybuilder | paid
|
||||
bytedance-seed/seed-1.6 | paid
|
||||
bytedance-seed/seed-1.6-flash | paid
|
||||
bytedance/ui-tars-1.5-7b | paid
|
||||
deepcogito/cogito-v2-preview-llama-109b-moe | paid
|
||||
cohere/command-a | paid
|
||||
cohere/command-r-08-2024 | paid
|
||||
cohere/command-r-plus-08-2024 | paid
|
||||
cohere/command-r7b-12-2024 | paid
|
||||
deepcogito/cogito-v2-preview-deepseek-671b | paid
|
||||
deepcogito/cogito-v2-preview-llama-405b | paid
|
||||
deepcogito/cogito-v2-preview-llama-70b | paid
|
||||
deepcogito/cogito-v2.1-671b | paid
|
||||
deepseek/deepseek-prover-v2 | paid
|
||||
deepseek/deepseek-r1-0528-qwen3-8b | paid
|
||||
deepseek/deepseek-r1-0528-qwen3-8b:free | free
|
||||
deepseek/deepseek-chat | paid
|
||||
deepseek/deepseek-chat-v3-0324 | paid
|
||||
deepseek/deepseek-chat-v3-0324:free | free
|
||||
deepseek/deepseek-chat-v3.1 | paid
|
||||
deepseek/deepseek-chat-v3.1:free | free
|
||||
deepseek/deepseek-v3.1-terminus | paid
|
||||
deepseek/deepseek-v3.1-terminus:exacto | paid
|
||||
deepseek/deepseek-v3.2 | paid
|
||||
deepseek/deepseek-v3.2-exp | paid
|
||||
deepseek/deepseek-v3.2-speciale | paid
|
||||
deepseek/deepseek-r1 | paid
|
||||
deepseek/deepseek-r1:free | free
|
||||
deepseek/deepseek-r1-0528 | paid
|
||||
deepseek/deepseek-r1-0528:free | free
|
||||
deepseek/deepseek-r1-distill-llama-70b | paid
|
||||
deepseek/deepseek-r1-distill-llama-70b:free | free
|
||||
deepseek/deepseek-r1-distill-qwen-14b | paid
|
||||
deepseek/deepseek-r1-distill-qwen-32b | paid
|
||||
cognitivecomputations/dolphin3.0-mistral-24b | paid
|
||||
cognitivecomputations/dolphin3.0-mistral-24b:free | free
|
||||
eleutherai/llemma_7b | paid
|
||||
essentialai/rnj-1-instruct | paid
|
||||
alpindale/goliath-120b | paid
|
||||
google/gemini-2.0-flash-001 | paid
|
||||
google/gemini-2.0-flash-exp:free | free
|
||||
@ -105,15 +111,15 @@ export interface IKBotOptions {
|
||||
google/gemini-2.5-flash-image | paid
|
||||
google/gemini-2.5-flash-image-preview | paid
|
||||
google/gemini-2.5-flash-lite | paid
|
||||
google/gemini-2.5-flash-lite-preview-06-17 | paid
|
||||
google/gemini-2.5-flash-lite-preview-09-2025 | paid
|
||||
google/gemini-2.5-flash-preview-09-2025 | paid
|
||||
google/gemini-2.5-pro | paid
|
||||
google/gemini-2.5-pro-preview-05-06 | paid
|
||||
google/gemini-2.5-pro-preview | paid
|
||||
google/gemini-3-flash-preview | paid
|
||||
google/gemini-3-pro-preview | paid
|
||||
google/gemma-2-27b-it | paid
|
||||
google/gemma-2-9b-it | paid
|
||||
google/gemma-2-9b-it:free | free
|
||||
google/gemma-3-12b-it | paid
|
||||
google/gemma-3-12b-it:free | free
|
||||
google/gemma-3-27b-it | paid
|
||||
@ -123,23 +129,24 @@ export interface IKBotOptions {
|
||||
google/gemma-3n-e2b-it:free | free
|
||||
google/gemma-3n-e4b-it | paid
|
||||
google/gemma-3n-e4b-it:free | free
|
||||
google/gemini-3-pro-image-preview | paid
|
||||
ibm-granite/granite-4.0-h-micro | paid
|
||||
inception/mercury | paid
|
||||
inception/mercury-coder | paid
|
||||
inclusionai/ling-1t | paid
|
||||
inflection/inflection-3-pi | paid
|
||||
inflection/inflection-3-productivity | paid
|
||||
liquid/lfm-3b | paid
|
||||
liquid/lfm-7b | paid
|
||||
kwaipilot/kat-coder-pro:free | free
|
||||
liquid/lfm-2.2-6b | paid
|
||||
liquid/lfm2-8b-a1b | paid
|
||||
meta-llama/llama-guard-3-8b | paid
|
||||
anthracite-org/magnum-v2-72b | paid
|
||||
anthracite-org/magnum-v4-72b | paid
|
||||
mancer/weaver | paid
|
||||
meituan/longcat-flash-chat | paid
|
||||
meituan/longcat-flash-chat:free | free
|
||||
meta-llama/llama-3-70b-instruct | paid
|
||||
meta-llama/llama-3-8b-instruct | paid
|
||||
meta-llama/llama-3.1-405b | paid
|
||||
meta-llama/llama-3.1-405b-instruct | paid
|
||||
meta-llama/llama-3.1-405b-instruct:free | free
|
||||
meta-llama/llama-3.1-70b-instruct | paid
|
||||
meta-llama/llama-3.1-8b-instruct | paid
|
||||
meta-llama/llama-3.2-11b-vision-instruct | paid
|
||||
@ -149,15 +156,10 @@ export interface IKBotOptions {
|
||||
meta-llama/llama-3.2-90b-vision-instruct | paid
|
||||
meta-llama/llama-3.3-70b-instruct | paid
|
||||
meta-llama/llama-3.3-70b-instruct:free | free
|
||||
meta-llama/llama-3.3-8b-instruct:free | free
|
||||
meta-llama/llama-4-maverick | paid
|
||||
meta-llama/llama-4-maverick:free | free
|
||||
meta-llama/llama-4-scout | paid
|
||||
meta-llama/llama-4-scout:free | free
|
||||
meta-llama/llama-guard-4-12b | paid
|
||||
meta-llama/llama-guard-2-8b | paid
|
||||
microsoft/mai-ds-r1 | paid
|
||||
microsoft/mai-ds-r1:free | free
|
||||
microsoft/phi-4 | paid
|
||||
microsoft/phi-4-multimodal-instruct | paid
|
||||
microsoft/phi-4-reasoning-plus | paid
|
||||
@ -165,21 +167,22 @@ export interface IKBotOptions {
|
||||
microsoft/phi-3-mini-128k-instruct | paid
|
||||
microsoft/phi-3.5-mini-128k-instruct | paid
|
||||
minimax/minimax-m1 | paid
|
||||
minimax/minimax-m2 | paid
|
||||
minimax/minimax-m2.1 | paid
|
||||
minimax/minimax-01 | paid
|
||||
mistralai/mistral-large | paid
|
||||
mistralai/mistral-large-2407 | paid
|
||||
mistralai/mistral-large-2411 | paid
|
||||
mistralai/mistral-small | paid
|
||||
mistralai/mistral-tiny | paid
|
||||
mistralai/codestral-2501 | paid
|
||||
mistralai/codestral-2508 | paid
|
||||
mistralai/devstral-2512 | paid
|
||||
mistralai/devstral-2512:free | free
|
||||
mistralai/devstral-medium | paid
|
||||
mistralai/devstral-small | paid
|
||||
mistralai/devstral-small-2505 | paid
|
||||
mistralai/devstral-small-2505:free | free
|
||||
mistralai/magistral-medium-2506 | paid
|
||||
mistralai/magistral-medium-2506:thinking | paid
|
||||
mistralai/magistral-small-2506 | paid
|
||||
mistralai/ministral-14b-2512 | paid
|
||||
mistralai/ministral-3b-2512 | paid
|
||||
mistralai/ministral-8b-2512 | paid
|
||||
mistralai/ministral-3b | paid
|
||||
mistralai/ministral-8b | paid
|
||||
mistralai/mistral-7b-instruct | paid
|
||||
@ -187,35 +190,36 @@ export interface IKBotOptions {
|
||||
mistralai/mistral-7b-instruct-v0.1 | paid
|
||||
mistralai/mistral-7b-instruct-v0.2 | paid
|
||||
mistralai/mistral-7b-instruct-v0.3 | paid
|
||||
mistralai/mistral-large-2512 | paid
|
||||
mistralai/mistral-medium-3 | paid
|
||||
mistralai/mistral-medium-3.1 | paid
|
||||
mistralai/mistral-nemo | paid
|
||||
mistralai/mistral-nemo:free | free
|
||||
mistralai/mistral-small-24b-instruct-2501 | paid
|
||||
mistralai/mistral-small-24b-instruct-2501:free | free
|
||||
mistralai/mistral-small-3.1-24b-instruct | paid
|
||||
mistralai/mistral-small-3.1-24b-instruct:free | free
|
||||
mistralai/mistral-small-3.2-24b-instruct | paid
|
||||
mistralai/mistral-small-3.2-24b-instruct:free | free
|
||||
mistralai/mistral-small-creative | paid
|
||||
mistralai/mixtral-8x22b-instruct | paid
|
||||
mistralai/mixtral-8x7b-instruct | paid
|
||||
mistralai/pixtral-12b | paid
|
||||
mistralai/pixtral-large-2411 | paid
|
||||
mistralai/mistral-saba | paid
|
||||
mistralai/voxtral-small-24b-2507 | paid
|
||||
moonshotai/kimi-dev-72b | paid
|
||||
moonshotai/kimi-dev-72b:free | free
|
||||
moonshotai/kimi-k2 | paid
|
||||
moonshotai/kimi-k2:free | free
|
||||
moonshotai/kimi-k2-0905 | paid
|
||||
moonshotai/kimi-k2-0905:exacto | paid
|
||||
moonshotai/kimi-k2-thinking | paid
|
||||
morph/morph-v3-fast | paid
|
||||
morph/morph-v3-large | paid
|
||||
gryphe/mythomax-l2-13b | paid
|
||||
neversleep/llama-3.1-lumimaid-8b | paid
|
||||
nex-agi/deepseek-v3.1-nex-n1:free | free
|
||||
neversleep/noromaid-20b | paid
|
||||
nousresearch/deephermes-3-llama-3-8b-preview | paid
|
||||
nousresearch/deephermes-3-llama-3-8b-preview:free | free
|
||||
nousresearch/deephermes-3-mistral-24b-preview | paid
|
||||
nousresearch/hermes-3-llama-3.1-405b | paid
|
||||
nousresearch/hermes-3-llama-3.1-405b:free | free
|
||||
nousresearch/hermes-3-llama-3.1-70b | paid
|
||||
nousresearch/hermes-4-405b | paid
|
||||
nousresearch/hermes-4-70b | paid
|
||||
@ -223,6 +227,10 @@ export interface IKBotOptions {
|
||||
nvidia/llama-3.1-nemotron-70b-instruct | paid
|
||||
nvidia/llama-3.1-nemotron-ultra-253b-v1 | paid
|
||||
nvidia/llama-3.3-nemotron-super-49b-v1.5 | paid
|
||||
nvidia/nemotron-3-nano-30b-a3b | paid
|
||||
nvidia/nemotron-3-nano-30b-a3b:free | free
|
||||
nvidia/nemotron-nano-12b-v2-vl | paid
|
||||
nvidia/nemotron-nano-12b-v2-vl:free | free
|
||||
nvidia/nemotron-nano-9b-v2 | paid
|
||||
nvidia/nemotron-nano-9b-v2:free | free
|
||||
openai/chatgpt-4o-latest | paid
|
||||
@ -252,29 +260,43 @@ export interface IKBotOptions {
|
||||
openai/gpt-5 | paid
|
||||
openai/gpt-5-chat | paid
|
||||
openai/gpt-5-codex | paid
|
||||
openai/gpt-5-image | paid
|
||||
openai/gpt-5-image-mini | paid
|
||||
openai/gpt-5-mini | paid
|
||||
openai/gpt-5-nano | paid
|
||||
openai/gpt-5-pro | paid
|
||||
openai/gpt-5.1 | paid
|
||||
openai/gpt-5.1-chat | paid
|
||||
openai/gpt-5.1-codex | paid
|
||||
openai/gpt-5.1-codex-max | paid
|
||||
openai/gpt-5.1-codex-mini | paid
|
||||
openai/gpt-5.2 | paid
|
||||
openai/gpt-5.2-chat | paid
|
||||
openai/gpt-5.2-pro | paid
|
||||
openai/gpt-oss-120b | paid
|
||||
openai/gpt-oss-120b:exacto | paid
|
||||
openai/gpt-oss-120b:free | free
|
||||
openai/gpt-oss-20b | paid
|
||||
openai/gpt-oss-20b:free | free
|
||||
openai/gpt-oss-safeguard-20b | paid
|
||||
openai/o1 | paid
|
||||
openai/o1-mini | paid
|
||||
openai/o1-mini-2024-09-12 | paid
|
||||
openai/o1-pro | paid
|
||||
openai/o3 | paid
|
||||
openai/o3-deep-research | paid
|
||||
openai/o3-mini | paid
|
||||
openai/o3-mini-high | paid
|
||||
openai/o3-pro | paid
|
||||
openai/o4-mini | paid
|
||||
openai/o4-mini-deep-research | paid
|
||||
openai/o4-mini-high | paid
|
||||
opengvlab/internvl3-78b | paid
|
||||
perplexity/r1-1776 | paid
|
||||
perplexity/sonar | paid
|
||||
perplexity/sonar-deep-research | paid
|
||||
perplexity/sonar-pro | paid
|
||||
perplexity/sonar-pro-search | paid
|
||||
perplexity/sonar-reasoning | paid
|
||||
perplexity/sonar-reasoning-pro | paid
|
||||
prime-intellect/intellect-3 | paid
|
||||
qwen/qwen-plus-2025-07-28 | paid
|
||||
qwen/qwen-plus-2025-07-28:thinking | paid
|
||||
qwen/qwen-vl-max | paid
|
||||
@ -285,26 +307,22 @@ export interface IKBotOptions {
|
||||
qwen/qwen-2.5-7b-instruct | paid
|
||||
qwen/qwen2.5-coder-7b-instruct | paid
|
||||
qwen/qwen2.5-vl-32b-instruct | paid
|
||||
qwen/qwen2.5-vl-32b-instruct:free | free
|
||||
qwen/qwen2.5-vl-72b-instruct | paid
|
||||
qwen/qwen2.5-vl-72b-instruct:free | free
|
||||
qwen/qwen-2.5-vl-7b-instruct | paid
|
||||
qwen/qwen-2.5-vl-7b-instruct:free | free
|
||||
qwen/qwen3-14b | paid
|
||||
qwen/qwen3-14b:free | free
|
||||
qwen/qwen3-235b-a22b | paid
|
||||
qwen/qwen3-235b-a22b:free | free
|
||||
qwen/qwen3-235b-a22b-2507 | paid
|
||||
qwen/qwen3-235b-a22b-thinking-2507 | paid
|
||||
qwen/qwen3-30b-a3b | paid
|
||||
qwen/qwen3-30b-a3b:free | free
|
||||
qwen/qwen3-30b-a3b-instruct-2507 | paid
|
||||
qwen/qwen3-30b-a3b-thinking-2507 | paid
|
||||
qwen/qwen3-32b | paid
|
||||
qwen/qwen3-4b:free | free
|
||||
qwen/qwen3-8b | paid
|
||||
qwen/qwen3-8b:free | free
|
||||
qwen/qwen3-coder-30b-a3b-instruct | paid
|
||||
qwen/qwen3-coder | paid
|
||||
qwen/qwen3-coder:exacto | paid
|
||||
qwen/qwen3-coder:free | free
|
||||
qwen/qwen3-coder-flash | paid
|
||||
qwen/qwen3-coder-plus | paid
|
||||
@ -315,36 +333,35 @@ export interface IKBotOptions {
|
||||
qwen/qwen3-vl-235b-a22b-thinking | paid
|
||||
qwen/qwen3-vl-30b-a3b-instruct | paid
|
||||
qwen/qwen3-vl-30b-a3b-thinking | paid
|
||||
qwen/qwen3-vl-32b-instruct | paid
|
||||
qwen/qwen3-vl-8b-instruct | paid
|
||||
qwen/qwen3-vl-8b-thinking | paid
|
||||
qwen/qwq-32b | paid
|
||||
qwen/qwen-2.5-72b-instruct | paid
|
||||
qwen/qwen-2.5-72b-instruct:free | free
|
||||
qwen/qwen-2.5-coder-32b-instruct | paid
|
||||
qwen/qwen-2.5-coder-32b-instruct:free | free
|
||||
relace/relace-apply-3 | paid
|
||||
relace/relace-search | paid
|
||||
undi95/remm-slerp-l2-13b | paid
|
||||
sao10k/l3-lunaris-8b | paid
|
||||
sao10k/l3-euryale-70b | paid
|
||||
sao10k/l3.1-70b-hanami-x1 | paid
|
||||
sao10k/l3.1-euryale-70b | paid
|
||||
sao10k/l3.3-euryale-70b | paid
|
||||
shisa-ai/shisa-v2-llama3.3-70b | paid
|
||||
shisa-ai/shisa-v2-llama3.3-70b:free | free
|
||||
raifle/sorcererlm-8x22b | paid
|
||||
stepfun-ai/step3 | paid
|
||||
switchpoint/router | paid
|
||||
tencent/hunyuan-a13b-instruct | paid
|
||||
tencent/hunyuan-a13b-instruct:free | free
|
||||
thedrummer/anubis-70b-v1.1 | paid
|
||||
thedrummer/cydonia-24b-v4.1 | paid
|
||||
thedrummer/rocinante-12b | paid
|
||||
thedrummer/skyfall-36b-v2 | paid
|
||||
thedrummer/unslopnemo-12b | paid
|
||||
thudm/glm-4.1v-9b-thinking | paid
|
||||
thudm/glm-z1-32b | paid
|
||||
tngtech/deepseek-r1t-chimera | paid
|
||||
tngtech/deepseek-r1t-chimera:free | free
|
||||
tngtech/deepseek-r1t2-chimera | paid
|
||||
tngtech/deepseek-r1t2-chimera:free | free
|
||||
tngtech/tng-r1t-chimera | paid
|
||||
tngtech/tng-r1t-chimera:free | free
|
||||
alibaba/tongyi-deepresearch-30b-a3b | paid
|
||||
alibaba/tongyi-deepresearch-30b-a3b:free | free
|
||||
cognitivecomputations/dolphin-mistral-24b-venice-edition:free | free
|
||||
@ -355,18 +372,24 @@ export interface IKBotOptions {
|
||||
x-ai/grok-3-mini-beta | paid
|
||||
x-ai/grok-4 | paid
|
||||
x-ai/grok-4-fast | paid
|
||||
x-ai/grok-4.1-fast | paid
|
||||
x-ai/grok-code-fast-1 | paid
|
||||
xiaomi/mimo-v2-flash:free | free
|
||||
z-ai/glm-4-32b | paid
|
||||
z-ai/glm-4.5 | paid
|
||||
z-ai/glm-4.5-air | paid
|
||||
z-ai/glm-4.5-air:free | free
|
||||
z-ai/glm-4.5v | paid
|
||||
z-ai/glm-4.6 | paid
|
||||
z-ai/glm-4.6:exacto | paid
|
||||
z-ai/glm-4.6v | paid
|
||||
z-ai/glm-4.7 | paid
|
||||
[35m[1m[22m[39m
|
||||
[35m[1m OpenAI models:[22m[39m
|
||||
[35m[1m[22m[39m
|
||||
babbage-002
|
||||
chatgpt-4o-latest
|
||||
chatgpt-image-latest
|
||||
codex-mini-latest
|
||||
dall-e-2
|
||||
dall-e-3
|
||||
@ -395,7 +418,6 @@ export interface IKBotOptions {
|
||||
gpt-4o-2024-08-06
|
||||
gpt-4o-2024-11-20
|
||||
gpt-4o-audio-preview
|
||||
gpt-4o-audio-preview-2024-10-01
|
||||
gpt-4o-audio-preview-2024-12-17
|
||||
gpt-4o-audio-preview-2025-06-03
|
||||
gpt-4o-mini
|
||||
@ -407,14 +429,18 @@ export interface IKBotOptions {
|
||||
gpt-4o-mini-search-preview
|
||||
gpt-4o-mini-search-preview-2025-03-11
|
||||
gpt-4o-mini-transcribe
|
||||
gpt-4o-mini-transcribe-2025-03-20
|
||||
gpt-4o-mini-transcribe-2025-12-15
|
||||
gpt-4o-mini-tts
|
||||
gpt-4o-mini-tts-2025-03-20
|
||||
gpt-4o-mini-tts-2025-12-15
|
||||
gpt-4o-realtime-preview
|
||||
gpt-4o-realtime-preview-2024-10-01
|
||||
gpt-4o-realtime-preview-2024-12-17
|
||||
gpt-4o-realtime-preview-2025-06-03
|
||||
gpt-4o-search-preview
|
||||
gpt-4o-search-preview-2025-03-11
|
||||
gpt-4o-transcribe
|
||||
gpt-4o-transcribe-diarize
|
||||
gpt-5
|
||||
gpt-5-2025-08-07
|
||||
gpt-5-chat-latest
|
||||
@ -425,20 +451,34 @@ export interface IKBotOptions {
|
||||
gpt-5-nano-2025-08-07
|
||||
gpt-5-pro
|
||||
gpt-5-pro-2025-10-06
|
||||
gpt-5-search-api
|
||||
gpt-5-search-api-2025-10-14
|
||||
gpt-5.1
|
||||
gpt-5.1-2025-11-13
|
||||
gpt-5.1-chat-latest
|
||||
gpt-5.1-codex
|
||||
gpt-5.1-codex-max
|
||||
gpt-5.1-codex-mini
|
||||
gpt-5.2
|
||||
gpt-5.2-2025-12-11
|
||||
gpt-5.2-chat-latest
|
||||
gpt-5.2-pro
|
||||
gpt-5.2-pro-2025-12-11
|
||||
gpt-audio
|
||||
gpt-audio-2025-08-28
|
||||
gpt-audio-mini
|
||||
gpt-audio-mini-2025-10-06
|
||||
gpt-audio-mini-2025-12-15
|
||||
gpt-image-1
|
||||
gpt-image-1-mini
|
||||
gpt-image-1.5
|
||||
gpt-realtime
|
||||
gpt-realtime-2025-08-28
|
||||
gpt-realtime-mini
|
||||
gpt-realtime-mini-2025-10-06
|
||||
gpt-realtime-mini-2025-12-15
|
||||
o1
|
||||
o1-2024-12-17
|
||||
o1-mini
|
||||
o1-mini-2024-09-12
|
||||
o1-pro
|
||||
o1-pro-2025-03-19
|
||||
o3
|
||||
|
||||
Loading…
Reference in New Issue
Block a user