maintainence love:)
This commit is contained in:
parent
4726ebc0f0
commit
214eed04f5
File diff suppressed because one or more lines are too long
@ -1,5 +1,5 @@
|
|||||||
{
|
{
|
||||||
"timestamp": 1743511914173,
|
"timestamp": 1743769889560,
|
||||||
"models": [
|
"models": [
|
||||||
{
|
{
|
||||||
"id": "gpt-4o-realtime-preview-2024-12-17",
|
"id": "gpt-4o-realtime-preview-2024-12-17",
|
||||||
@ -32,15 +32,15 @@
|
|||||||
"owned_by": "system"
|
"owned_by": "system"
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
"id": "gpt-4o-mini-realtime-preview-2024-12-17",
|
"id": "o1-pro-2025-03-19",
|
||||||
"object": "model",
|
"object": "model",
|
||||||
"created": 1734112601,
|
"created": 1742251504,
|
||||||
"owned_by": "system"
|
"owned_by": "system"
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
"id": "gpt-4o-mini-realtime-preview",
|
"id": "o1-pro",
|
||||||
"object": "model",
|
"object": "model",
|
||||||
"created": 1734387380,
|
"created": 1742251791,
|
||||||
"owned_by": "system"
|
"owned_by": "system"
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
@ -49,12 +49,24 @@
|
|||||||
"created": 1727131766,
|
"created": 1727131766,
|
||||||
"owned_by": "system"
|
"owned_by": "system"
|
||||||
},
|
},
|
||||||
|
{
|
||||||
|
"id": "o1-2024-12-17",
|
||||||
|
"object": "model",
|
||||||
|
"created": 1734326976,
|
||||||
|
"owned_by": "system"
|
||||||
|
},
|
||||||
{
|
{
|
||||||
"id": "gpt-4o-transcribe",
|
"id": "gpt-4o-transcribe",
|
||||||
"object": "model",
|
"object": "model",
|
||||||
"created": 1742068463,
|
"created": 1742068463,
|
||||||
"owned_by": "system"
|
"owned_by": "system"
|
||||||
},
|
},
|
||||||
|
{
|
||||||
|
"id": "o1",
|
||||||
|
"object": "model",
|
||||||
|
"created": 1734375816,
|
||||||
|
"owned_by": "system"
|
||||||
|
},
|
||||||
{
|
{
|
||||||
"id": "gpt-4o-mini-transcribe",
|
"id": "gpt-4o-mini-transcribe",
|
||||||
"object": "model",
|
"object": "model",
|
||||||
@ -103,6 +115,12 @@
|
|||||||
"created": 1671217299,
|
"created": 1671217299,
|
||||||
"owned_by": "openai-internal"
|
"owned_by": "openai-internal"
|
||||||
},
|
},
|
||||||
|
{
|
||||||
|
"id": "omni-moderation-latest",
|
||||||
|
"object": "model",
|
||||||
|
"created": 1731689265,
|
||||||
|
"owned_by": "system"
|
||||||
|
},
|
||||||
{
|
{
|
||||||
"id": "tts-1-hd",
|
"id": "tts-1-hd",
|
||||||
"object": "model",
|
"object": "model",
|
||||||
@ -127,6 +145,18 @@
|
|||||||
"created": 1725648865,
|
"created": 1725648865,
|
||||||
"owned_by": "system"
|
"owned_by": "system"
|
||||||
},
|
},
|
||||||
|
{
|
||||||
|
"id": "gpt-4o-mini-realtime-preview",
|
||||||
|
"object": "model",
|
||||||
|
"created": 1734387380,
|
||||||
|
"owned_by": "system"
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"id": "gpt-4o-mini-realtime-preview-2024-12-17",
|
||||||
|
"object": "model",
|
||||||
|
"created": 1734112601,
|
||||||
|
"owned_by": "system"
|
||||||
|
},
|
||||||
{
|
{
|
||||||
"id": "gpt-3.5-turbo-instruct-0914",
|
"id": "gpt-3.5-turbo-instruct-0914",
|
||||||
"object": "model",
|
"object": "model",
|
||||||
@ -139,24 +169,12 @@
|
|||||||
"created": 1741391161,
|
"created": 1741391161,
|
||||||
"owned_by": "system"
|
"owned_by": "system"
|
||||||
},
|
},
|
||||||
{
|
|
||||||
"id": "o1",
|
|
||||||
"object": "model",
|
|
||||||
"created": 1734375816,
|
|
||||||
"owned_by": "system"
|
|
||||||
},
|
|
||||||
{
|
{
|
||||||
"id": "tts-1-1106",
|
"id": "tts-1-1106",
|
||||||
"object": "model",
|
"object": "model",
|
||||||
"created": 1699053241,
|
"created": 1699053241,
|
||||||
"owned_by": "system"
|
"owned_by": "system"
|
||||||
},
|
},
|
||||||
{
|
|
||||||
"id": "o1-2024-12-17",
|
|
||||||
"object": "model",
|
|
||||||
"created": 1734326976,
|
|
||||||
"owned_by": "system"
|
|
||||||
},
|
|
||||||
{
|
{
|
||||||
"id": "davinci-002",
|
"id": "davinci-002",
|
||||||
"object": "model",
|
"object": "model",
|
||||||
@ -169,18 +187,6 @@
|
|||||||
"created": 1698959748,
|
"created": 1698959748,
|
||||||
"owned_by": "system"
|
"owned_by": "system"
|
||||||
},
|
},
|
||||||
{
|
|
||||||
"id": "o1-pro",
|
|
||||||
"object": "model",
|
|
||||||
"created": 1742251791,
|
|
||||||
"owned_by": "system"
|
|
||||||
},
|
|
||||||
{
|
|
||||||
"id": "o1-pro-2025-03-19",
|
|
||||||
"object": "model",
|
|
||||||
"created": 1742251504,
|
|
||||||
"owned_by": "system"
|
|
||||||
},
|
|
||||||
{
|
{
|
||||||
"id": "gpt-4-turbo",
|
"id": "gpt-4-turbo",
|
||||||
"object": "model",
|
"object": "model",
|
||||||
@ -301,12 +307,6 @@
|
|||||||
"created": 1741388170,
|
"created": 1741388170,
|
||||||
"owned_by": "system"
|
"owned_by": "system"
|
||||||
},
|
},
|
||||||
{
|
|
||||||
"id": "omni-moderation-latest",
|
|
||||||
"object": "model",
|
|
||||||
"created": 1731689265,
|
|
||||||
"owned_by": "system"
|
|
||||||
},
|
|
||||||
{
|
{
|
||||||
"id": "tts-1",
|
"id": "tts-1",
|
||||||
"object": "model",
|
"object": "model",
|
||||||
@ -349,12 +349,6 @@
|
|||||||
"created": 1721172717,
|
"created": 1721172717,
|
||||||
"owned_by": "system"
|
"owned_by": "system"
|
||||||
},
|
},
|
||||||
{
|
|
||||||
"id": "gpt-4-turbo-preview",
|
|
||||||
"object": "model",
|
|
||||||
"created": 1706037777,
|
|
||||||
"owned_by": "system"
|
|
||||||
},
|
|
||||||
{
|
{
|
||||||
"id": "o1-mini",
|
"id": "o1-mini",
|
||||||
"object": "model",
|
"object": "model",
|
||||||
@ -378,6 +372,12 @@
|
|||||||
"object": "model",
|
"object": "model",
|
||||||
"created": 1706037612,
|
"created": 1706037612,
|
||||||
"owned_by": "system"
|
"owned_by": "system"
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"id": "gpt-4-turbo-preview",
|
||||||
|
"object": "model",
|
||||||
|
"created": 1706037777,
|
||||||
|
"owned_by": "system"
|
||||||
}
|
}
|
||||||
]
|
]
|
||||||
}
|
}
|
||||||
@ -1,6 +1,75 @@
|
|||||||
{
|
{
|
||||||
"timestamp": 1743511914343,
|
"timestamp": 1743769889892,
|
||||||
"models": [
|
"models": [
|
||||||
|
{
|
||||||
|
"id": "openrouter/quasar-alpha",
|
||||||
|
"name": "Quasar Alpha",
|
||||||
|
"created": 1743626809,
|
||||||
|
"description": "This is a cloaked model provided to the community to gather feedback. It’s a powerful, all-purpose model supporting long-context tasks, including code generation. All prompts and completions for this model are logged by the provider as well as OpenRouter.",
|
||||||
|
"context_length": 1000000,
|
||||||
|
"architecture": {
|
||||||
|
"modality": "text+image->text",
|
||||||
|
"input_modalities": [
|
||||||
|
"image",
|
||||||
|
"text"
|
||||||
|
],
|
||||||
|
"output_modalities": [
|
||||||
|
"text"
|
||||||
|
],
|
||||||
|
"tokenizer": "Other",
|
||||||
|
"instruct_type": null
|
||||||
|
},
|
||||||
|
"pricing": {
|
||||||
|
"prompt": "0",
|
||||||
|
"completion": "0",
|
||||||
|
"request": "0",
|
||||||
|
"image": "0",
|
||||||
|
"web_search": "0",
|
||||||
|
"internal_reasoning": "0",
|
||||||
|
"input_cache_read": "0",
|
||||||
|
"input_cache_write": "0"
|
||||||
|
},
|
||||||
|
"top_provider": {
|
||||||
|
"context_length": 1000000,
|
||||||
|
"max_completion_tokens": 32000,
|
||||||
|
"is_moderated": true
|
||||||
|
},
|
||||||
|
"per_request_limits": null
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"id": "all-hands/openhands-lm-32b-v0.1",
|
||||||
|
"name": "OpenHands LM 32B V0.1",
|
||||||
|
"created": 1743613013,
|
||||||
|
"description": "OpenHands LM v0.1 is a 32B open-source coding model fine-tuned from Qwen2.5-Coder-32B-Instruct using reinforcement learning techniques outlined in SWE-Gym. It is optimized for autonomous software development agents and achieves strong performance on SWE-Bench Verified, with a 37.2% resolve rate. The model supports a 128K token context window, making it well-suited for long-horizon code reasoning and large codebase tasks.\n\nOpenHands LM is designed for local deployment and runs on consumer-grade GPUs such as a single 3090. It enables fully offline agent workflows without dependency on proprietary APIs. This release is intended as a research preview, and future updates aim to improve generalizability, reduce repetition, and offer smaller variants.",
|
||||||
|
"context_length": 16384,
|
||||||
|
"architecture": {
|
||||||
|
"modality": "text->text",
|
||||||
|
"input_modalities": [
|
||||||
|
"text"
|
||||||
|
],
|
||||||
|
"output_modalities": [
|
||||||
|
"text"
|
||||||
|
],
|
||||||
|
"tokenizer": "Other",
|
||||||
|
"instruct_type": null
|
||||||
|
},
|
||||||
|
"pricing": {
|
||||||
|
"prompt": "0.0000026",
|
||||||
|
"completion": "0.0000034",
|
||||||
|
"request": "0",
|
||||||
|
"image": "0",
|
||||||
|
"web_search": "0",
|
||||||
|
"internal_reasoning": "0",
|
||||||
|
"input_cache_read": "0",
|
||||||
|
"input_cache_write": "0"
|
||||||
|
},
|
||||||
|
"top_provider": {
|
||||||
|
"context_length": 16384,
|
||||||
|
"max_completion_tokens": 4095,
|
||||||
|
"is_moderated": false
|
||||||
|
},
|
||||||
|
"per_request_limits": null
|
||||||
|
},
|
||||||
{
|
{
|
||||||
"id": "mistral/ministral-8b",
|
"id": "mistral/ministral-8b",
|
||||||
"name": "Mistral: Ministral 8b",
|
"name": "Mistral: Ministral 8b",
|
||||||
@ -1465,8 +1534,8 @@
|
|||||||
"instruct_type": "qwq"
|
"instruct_type": "qwq"
|
||||||
},
|
},
|
||||||
"pricing": {
|
"pricing": {
|
||||||
"prompt": "0.00000012",
|
"prompt": "0.00000015",
|
||||||
"completion": "0.00000018",
|
"completion": "0.0000002",
|
||||||
"request": "0",
|
"request": "0",
|
||||||
"image": "0",
|
"image": "0",
|
||||||
"web_search": "0",
|
"web_search": "0",
|
||||||
@ -1653,41 +1722,6 @@
|
|||||||
},
|
},
|
||||||
"per_request_limits": null
|
"per_request_limits": null
|
||||||
},
|
},
|
||||||
{
|
|
||||||
"id": "anthropic/claude-3.7-sonnet:beta",
|
|
||||||
"name": "Anthropic: Claude 3.7 Sonnet (self-moderated)",
|
|
||||||
"created": 1740422110,
|
|
||||||
"description": "Claude 3.7 Sonnet is an advanced large language model with improved reasoning, coding, and problem-solving capabilities. It introduces a hybrid reasoning approach, allowing users to choose between rapid responses and extended, step-by-step processing for complex tasks. The model demonstrates notable improvements in coding, particularly in front-end development and full-stack updates, and excels in agentic workflows, where it can autonomously navigate multi-step processes. \n\nClaude 3.7 Sonnet maintains performance parity with its predecessor in standard mode while offering an extended reasoning mode for enhanced accuracy in math, coding, and instruction-following tasks.\n\nRead more at the [blog post here](https://www.anthropic.com/news/claude-3-7-sonnet)",
|
|
||||||
"context_length": 200000,
|
|
||||||
"architecture": {
|
|
||||||
"modality": "text+image->text",
|
|
||||||
"input_modalities": [
|
|
||||||
"text",
|
|
||||||
"image"
|
|
||||||
],
|
|
||||||
"output_modalities": [
|
|
||||||
"text"
|
|
||||||
],
|
|
||||||
"tokenizer": "Claude",
|
|
||||||
"instruct_type": null
|
|
||||||
},
|
|
||||||
"pricing": {
|
|
||||||
"prompt": "0.000003",
|
|
||||||
"completion": "0.000015",
|
|
||||||
"request": "0",
|
|
||||||
"image": "0.0048",
|
|
||||||
"web_search": "0",
|
|
||||||
"internal_reasoning": "0",
|
|
||||||
"input_cache_read": "0",
|
|
||||||
"input_cache_write": "0"
|
|
||||||
},
|
|
||||||
"top_provider": {
|
|
||||||
"context_length": 200000,
|
|
||||||
"max_completion_tokens": 128000,
|
|
||||||
"is_moderated": false
|
|
||||||
},
|
|
||||||
"per_request_limits": null
|
|
||||||
},
|
|
||||||
{
|
{
|
||||||
"id": "anthropic/claude-3.7-sonnet",
|
"id": "anthropic/claude-3.7-sonnet",
|
||||||
"name": "Anthropic: Claude 3.7 Sonnet",
|
"name": "Anthropic: Claude 3.7 Sonnet",
|
||||||
@ -1718,8 +1752,8 @@
|
|||||||
},
|
},
|
||||||
"top_provider": {
|
"top_provider": {
|
||||||
"context_length": 200000,
|
"context_length": 200000,
|
||||||
"max_completion_tokens": 128000,
|
"max_completion_tokens": 64000,
|
||||||
"is_moderated": true
|
"is_moderated": false
|
||||||
},
|
},
|
||||||
"per_request_limits": null
|
"per_request_limits": null
|
||||||
},
|
},
|
||||||
@ -1751,10 +1785,45 @@
|
|||||||
"input_cache_read": "0",
|
"input_cache_read": "0",
|
||||||
"input_cache_write": "0"
|
"input_cache_write": "0"
|
||||||
},
|
},
|
||||||
|
"top_provider": {
|
||||||
|
"context_length": 200000,
|
||||||
|
"max_completion_tokens": 64000,
|
||||||
|
"is_moderated": false
|
||||||
|
},
|
||||||
|
"per_request_limits": null
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"id": "anthropic/claude-3.7-sonnet:beta",
|
||||||
|
"name": "Anthropic: Claude 3.7 Sonnet (self-moderated)",
|
||||||
|
"created": 1740422110,
|
||||||
|
"description": "Claude 3.7 Sonnet is an advanced large language model with improved reasoning, coding, and problem-solving capabilities. It introduces a hybrid reasoning approach, allowing users to choose between rapid responses and extended, step-by-step processing for complex tasks. The model demonstrates notable improvements in coding, particularly in front-end development and full-stack updates, and excels in agentic workflows, where it can autonomously navigate multi-step processes. \n\nClaude 3.7 Sonnet maintains performance parity with its predecessor in standard mode while offering an extended reasoning mode for enhanced accuracy in math, coding, and instruction-following tasks.\n\nRead more at the [blog post here](https://www.anthropic.com/news/claude-3-7-sonnet)",
|
||||||
|
"context_length": 200000,
|
||||||
|
"architecture": {
|
||||||
|
"modality": "text+image->text",
|
||||||
|
"input_modalities": [
|
||||||
|
"text",
|
||||||
|
"image"
|
||||||
|
],
|
||||||
|
"output_modalities": [
|
||||||
|
"text"
|
||||||
|
],
|
||||||
|
"tokenizer": "Claude",
|
||||||
|
"instruct_type": null
|
||||||
|
},
|
||||||
|
"pricing": {
|
||||||
|
"prompt": "0.000003",
|
||||||
|
"completion": "0.000015",
|
||||||
|
"request": "0",
|
||||||
|
"image": "0.0048",
|
||||||
|
"web_search": "0",
|
||||||
|
"internal_reasoning": "0",
|
||||||
|
"input_cache_read": "0",
|
||||||
|
"input_cache_write": "0"
|
||||||
|
},
|
||||||
"top_provider": {
|
"top_provider": {
|
||||||
"context_length": 200000,
|
"context_length": 200000,
|
||||||
"max_completion_tokens": 128000,
|
"max_completion_tokens": 128000,
|
||||||
"is_moderated": true
|
"is_moderated": false
|
||||||
},
|
},
|
||||||
"per_request_limits": null
|
"per_request_limits": null
|
||||||
},
|
},
|
||||||
@ -2312,7 +2381,7 @@
|
|||||||
"name": "Qwen: Qwen2.5 VL 72B Instruct",
|
"name": "Qwen: Qwen2.5 VL 72B Instruct",
|
||||||
"created": 1738410311,
|
"created": 1738410311,
|
||||||
"description": "Qwen2.5-VL is proficient in recognizing common objects such as flowers, birds, fish, and insects. It is also highly capable of analyzing texts, charts, icons, graphics, and layouts within images.",
|
"description": "Qwen2.5-VL is proficient in recognizing common objects such as flowers, birds, fish, and insects. It is also highly capable of analyzing texts, charts, icons, graphics, and layouts within images.",
|
||||||
"context_length": 128000,
|
"context_length": 32000,
|
||||||
"architecture": {
|
"architecture": {
|
||||||
"modality": "text+image->text",
|
"modality": "text+image->text",
|
||||||
"input_modalities": [
|
"input_modalities": [
|
||||||
@ -2336,8 +2405,8 @@
|
|||||||
"input_cache_write": "0"
|
"input_cache_write": "0"
|
||||||
},
|
},
|
||||||
"top_provider": {
|
"top_provider": {
|
||||||
"context_length": 128000,
|
"context_length": 32000,
|
||||||
"max_completion_tokens": 128000,
|
"max_completion_tokens": 32000,
|
||||||
"is_moderated": false
|
"is_moderated": false
|
||||||
},
|
},
|
||||||
"per_request_limits": null
|
"per_request_limits": null
|
||||||
@ -2823,7 +2892,7 @@
|
|||||||
"name": "DeepSeek: R1 Distill Llama 70B (free)",
|
"name": "DeepSeek: R1 Distill Llama 70B (free)",
|
||||||
"created": 1737663169,
|
"created": 1737663169,
|
||||||
"description": "DeepSeek R1 Distill Llama 70B is a distilled large language model based on [Llama-3.3-70B-Instruct](/meta-llama/llama-3.3-70b-instruct), using outputs from [DeepSeek R1](/deepseek/deepseek-r1). The model combines advanced distillation techniques to achieve high performance across multiple benchmarks, including:\n\n- AIME 2024 pass@1: 70.0\n- MATH-500 pass@1: 94.5\n- CodeForces Rating: 1633\n\nThe model leverages fine-tuning from DeepSeek R1's outputs, enabling competitive performance comparable to larger frontier models.",
|
"description": "DeepSeek R1 Distill Llama 70B is a distilled large language model based on [Llama-3.3-70B-Instruct](/meta-llama/llama-3.3-70b-instruct), using outputs from [DeepSeek R1](/deepseek/deepseek-r1). The model combines advanced distillation techniques to achieve high performance across multiple benchmarks, including:\n\n- AIME 2024 pass@1: 70.0\n- MATH-500 pass@1: 94.5\n- CodeForces Rating: 1633\n\nThe model leverages fine-tuning from DeepSeek R1's outputs, enabling competitive performance comparable to larger frontier models.",
|
||||||
"context_length": 8192,
|
"context_length": 128000,
|
||||||
"architecture": {
|
"architecture": {
|
||||||
"modality": "text->text",
|
"modality": "text->text",
|
||||||
"input_modalities": [
|
"input_modalities": [
|
||||||
@ -2846,8 +2915,8 @@
|
|||||||
"input_cache_write": "0"
|
"input_cache_write": "0"
|
||||||
},
|
},
|
||||||
"top_provider": {
|
"top_provider": {
|
||||||
"context_length": 8192,
|
"context_length": 128000,
|
||||||
"max_completion_tokens": 4096,
|
"max_completion_tokens": null,
|
||||||
"is_moderated": false
|
"is_moderated": false
|
||||||
},
|
},
|
||||||
"per_request_limits": null
|
"per_request_limits": null
|
||||||
@ -5088,7 +5157,7 @@
|
|||||||
"name": "Meta: Llama 3.2 11B Vision Instruct",
|
"name": "Meta: Llama 3.2 11B Vision Instruct",
|
||||||
"created": 1727222400,
|
"created": 1727222400,
|
||||||
"description": "Llama 3.2 11B Vision is a multimodal model with 11 billion parameters, designed to handle tasks combining visual and textual data. It excels in tasks such as image captioning and visual question answering, bridging the gap between language generation and visual reasoning. Pre-trained on a massive dataset of image-text pairs, it performs well in complex, high-accuracy image analysis.\n\nIts ability to integrate visual understanding with language processing makes it an ideal solution for industries requiring comprehensive visual-linguistic AI applications, such as content creation, AI-driven customer service, and research.\n\nClick here for the [original model card](https://github.com/meta-llama/llama-models/blob/main/models/llama3_2/MODEL_CARD_VISION.md).\n\nUsage of this model is subject to [Meta's Acceptable Use Policy](https://www.llama.com/llama3/use-policy/).",
|
"description": "Llama 3.2 11B Vision is a multimodal model with 11 billion parameters, designed to handle tasks combining visual and textual data. It excels in tasks such as image captioning and visual question answering, bridging the gap between language generation and visual reasoning. Pre-trained on a massive dataset of image-text pairs, it performs well in complex, high-accuracy image analysis.\n\nIts ability to integrate visual understanding with language processing makes it an ideal solution for industries requiring comprehensive visual-linguistic AI applications, such as content creation, AI-driven customer service, and research.\n\nClick here for the [original model card](https://github.com/meta-llama/llama-models/blob/main/models/llama3_2/MODEL_CARD_VISION.md).\n\nUsage of this model is subject to [Meta's Acceptable Use Policy](https://www.llama.com/llama3/use-policy/).",
|
||||||
"context_length": 16384,
|
"context_length": 131072,
|
||||||
"architecture": {
|
"architecture": {
|
||||||
"modality": "text+image->text",
|
"modality": "text+image->text",
|
||||||
"input_modalities": [
|
"input_modalities": [
|
||||||
@ -5105,15 +5174,15 @@
|
|||||||
"prompt": "0.000000055",
|
"prompt": "0.000000055",
|
||||||
"completion": "0.000000055",
|
"completion": "0.000000055",
|
||||||
"request": "0",
|
"request": "0",
|
||||||
"image": "0",
|
"image": "0.00007948",
|
||||||
"web_search": "0",
|
"web_search": "0",
|
||||||
"internal_reasoning": "0",
|
"internal_reasoning": "0",
|
||||||
"input_cache_read": "0",
|
"input_cache_read": "0",
|
||||||
"input_cache_write": "0"
|
"input_cache_write": "0"
|
||||||
},
|
},
|
||||||
"top_provider": {
|
"top_provider": {
|
||||||
"context_length": 16384,
|
"context_length": 131072,
|
||||||
"max_completion_tokens": 16384,
|
"max_completion_tokens": 8192,
|
||||||
"is_moderated": false
|
"is_moderated": false
|
||||||
},
|
},
|
||||||
"per_request_limits": null
|
"per_request_limits": null
|
||||||
@ -5444,8 +5513,8 @@
|
|||||||
"instruct_type": null
|
"instruct_type": null
|
||||||
},
|
},
|
||||||
"pricing": {
|
"pricing": {
|
||||||
"prompt": "0.000002375",
|
"prompt": "0.0000025",
|
||||||
"completion": "0.0000095",
|
"completion": "0.00001",
|
||||||
"request": "0",
|
"request": "0",
|
||||||
"image": "0",
|
"image": "0",
|
||||||
"web_search": "0",
|
"web_search": "0",
|
||||||
@ -5478,8 +5547,8 @@
|
|||||||
"instruct_type": null
|
"instruct_type": null
|
||||||
},
|
},
|
||||||
"pricing": {
|
"pricing": {
|
||||||
"prompt": "0.0000001425",
|
"prompt": "0.00000015",
|
||||||
"completion": "0.00000057",
|
"completion": "0.0000006",
|
||||||
"request": "0",
|
"request": "0",
|
||||||
"image": "0",
|
"image": "0",
|
||||||
"web_search": "0",
|
"web_search": "0",
|
||||||
@ -5856,8 +5925,8 @@
|
|||||||
"instruct_type": "llama3"
|
"instruct_type": "llama3"
|
||||||
},
|
},
|
||||||
"pricing": {
|
"pricing": {
|
||||||
"prompt": "0.00000003",
|
"prompt": "0.00000005",
|
||||||
"completion": "0.00000006",
|
"completion": "0.00000005",
|
||||||
"request": "0",
|
"request": "0",
|
||||||
"image": "0",
|
"image": "0",
|
||||||
"web_search": "0",
|
"web_search": "0",
|
||||||
@ -5867,7 +5936,7 @@
|
|||||||
},
|
},
|
||||||
"top_provider": {
|
"top_provider": {
|
||||||
"context_length": 8192,
|
"context_length": 8192,
|
||||||
"max_completion_tokens": 8192,
|
"max_completion_tokens": null,
|
||||||
"is_moderated": false
|
"is_moderated": false
|
||||||
},
|
},
|
||||||
"per_request_limits": null
|
"per_request_limits": null
|
||||||
@ -6385,74 +6454,6 @@
|
|||||||
},
|
},
|
||||||
"per_request_limits": null
|
"per_request_limits": null
|
||||||
},
|
},
|
||||||
{
|
|
||||||
"id": "qwen/qwen-2-7b-instruct:free",
|
|
||||||
"name": "Qwen 2 7B Instruct (free)",
|
|
||||||
"created": 1721088000,
|
|
||||||
"description": "Qwen2 7B is a transformer-based model that excels in language understanding, multilingual capabilities, coding, mathematics, and reasoning.\n\nIt features SwiGLU activation, attention QKV bias, and group query attention. It is pretrained on extensive data with supervised finetuning and direct preference optimization.\n\nFor more details, see this [blog post](https://qwenlm.github.io/blog/qwen2/) and [GitHub repo](https://github.com/QwenLM/Qwen2).\n\nUsage of this model is subject to [Tongyi Qianwen LICENSE AGREEMENT](https://huggingface.co/Qwen/Qwen1.5-110B-Chat/blob/main/LICENSE).",
|
|
||||||
"context_length": 8192,
|
|
||||||
"architecture": {
|
|
||||||
"modality": "text->text",
|
|
||||||
"input_modalities": [
|
|
||||||
"text"
|
|
||||||
],
|
|
||||||
"output_modalities": [
|
|
||||||
"text"
|
|
||||||
],
|
|
||||||
"tokenizer": "Qwen",
|
|
||||||
"instruct_type": "chatml"
|
|
||||||
},
|
|
||||||
"pricing": {
|
|
||||||
"prompt": "0",
|
|
||||||
"completion": "0",
|
|
||||||
"request": "0",
|
|
||||||
"image": "0",
|
|
||||||
"web_search": "0",
|
|
||||||
"internal_reasoning": "0",
|
|
||||||
"input_cache_read": "0",
|
|
||||||
"input_cache_write": "0"
|
|
||||||
},
|
|
||||||
"top_provider": {
|
|
||||||
"context_length": 8192,
|
|
||||||
"max_completion_tokens": 4096,
|
|
||||||
"is_moderated": false
|
|
||||||
},
|
|
||||||
"per_request_limits": null
|
|
||||||
},
|
|
||||||
{
|
|
||||||
"id": "qwen/qwen-2-7b-instruct",
|
|
||||||
"name": "Qwen 2 7B Instruct",
|
|
||||||
"created": 1721088000,
|
|
||||||
"description": "Qwen2 7B is a transformer-based model that excels in language understanding, multilingual capabilities, coding, mathematics, and reasoning.\n\nIt features SwiGLU activation, attention QKV bias, and group query attention. It is pretrained on extensive data with supervised finetuning and direct preference optimization.\n\nFor more details, see this [blog post](https://qwenlm.github.io/blog/qwen2/) and [GitHub repo](https://github.com/QwenLM/Qwen2).\n\nUsage of this model is subject to [Tongyi Qianwen LICENSE AGREEMENT](https://huggingface.co/Qwen/Qwen1.5-110B-Chat/blob/main/LICENSE).",
|
|
||||||
"context_length": 32768,
|
|
||||||
"architecture": {
|
|
||||||
"modality": "text->text",
|
|
||||||
"input_modalities": [
|
|
||||||
"text"
|
|
||||||
],
|
|
||||||
"output_modalities": [
|
|
||||||
"text"
|
|
||||||
],
|
|
||||||
"tokenizer": "Qwen",
|
|
||||||
"instruct_type": "chatml"
|
|
||||||
},
|
|
||||||
"pricing": {
|
|
||||||
"prompt": "0.000000054",
|
|
||||||
"completion": "0.000000054",
|
|
||||||
"request": "0",
|
|
||||||
"image": "0",
|
|
||||||
"web_search": "0",
|
|
||||||
"internal_reasoning": "0",
|
|
||||||
"input_cache_read": "0",
|
|
||||||
"input_cache_write": "0"
|
|
||||||
},
|
|
||||||
"top_provider": {
|
|
||||||
"context_length": 32768,
|
|
||||||
"max_completion_tokens": null,
|
|
||||||
"is_moderated": false
|
|
||||||
},
|
|
||||||
"per_request_limits": null
|
|
||||||
},
|
|
||||||
{
|
{
|
||||||
"id": "google/gemma-2-27b-it",
|
"id": "google/gemma-2-27b-it",
|
||||||
"name": "Google: Gemma 2 27B",
|
"name": "Google: Gemma 2 27B",
|
||||||
@ -6471,8 +6472,8 @@
|
|||||||
"instruct_type": "gemma"
|
"instruct_type": "gemma"
|
||||||
},
|
},
|
||||||
"pricing": {
|
"pricing": {
|
||||||
"prompt": "0.00000027",
|
"prompt": "0.0000008",
|
||||||
"completion": "0.00000027",
|
"completion": "0.0000008",
|
||||||
"request": "0",
|
"request": "0",
|
||||||
"image": "0",
|
"image": "0",
|
||||||
"web_search": "0",
|
"web_search": "0",
|
||||||
@ -6482,7 +6483,7 @@
|
|||||||
},
|
},
|
||||||
"top_provider": {
|
"top_provider": {
|
||||||
"context_length": 8192,
|
"context_length": 8192,
|
||||||
"max_completion_tokens": 8192,
|
"max_completion_tokens": 2048,
|
||||||
"is_moderated": false
|
"is_moderated": false
|
||||||
},
|
},
|
||||||
"per_request_limits": null
|
"per_request_limits": null
|
||||||
@ -6573,8 +6574,8 @@
|
|||||||
"instruct_type": "gemma"
|
"instruct_type": "gemma"
|
||||||
},
|
},
|
||||||
"pricing": {
|
"pricing": {
|
||||||
"prompt": "0.00000003",
|
"prompt": "0.00000007",
|
||||||
"completion": "0.00000006",
|
"completion": "0.00000007",
|
||||||
"request": "0",
|
"request": "0",
|
||||||
"image": "0",
|
"image": "0",
|
||||||
"web_search": "0",
|
"web_search": "0",
|
||||||
@ -6584,7 +6585,7 @@
|
|||||||
},
|
},
|
||||||
"top_provider": {
|
"top_provider": {
|
||||||
"context_length": 8192,
|
"context_length": 8192,
|
||||||
"max_completion_tokens": 8192,
|
"max_completion_tokens": null,
|
||||||
"is_moderated": false
|
"is_moderated": false
|
||||||
},
|
},
|
||||||
"per_request_limits": null
|
"per_request_limits": null
|
||||||
@ -6745,8 +6746,8 @@
|
|||||||
"instruct_type": "llama3"
|
"instruct_type": "llama3"
|
||||||
},
|
},
|
||||||
"pricing": {
|
"pricing": {
|
||||||
"prompt": "0.0000007",
|
"prompt": "0.00000148",
|
||||||
"completion": "0.0000008",
|
"completion": "0.00000148",
|
||||||
"request": "0",
|
"request": "0",
|
||||||
"image": "0",
|
"image": "0",
|
||||||
"web_search": "0",
|
"web_search": "0",
|
||||||
@ -7411,40 +7412,6 @@
|
|||||||
},
|
},
|
||||||
"per_request_limits": null
|
"per_request_limits": null
|
||||||
},
|
},
|
||||||
{
|
|
||||||
"id": "meta-llama/llama-3-8b-instruct:free",
|
|
||||||
"name": "Meta: Llama 3 8B Instruct (free)",
|
|
||||||
"created": 1713398400,
|
|
||||||
"description": "Meta's latest class of model (Llama 3) launched with a variety of sizes & flavors. This 8B instruct-tuned version was optimized for high quality dialogue usecases.\n\nIt has demonstrated strong performance compared to leading closed-source models in human evaluations.\n\nTo read more about the model release, [click here](https://ai.meta.com/blog/meta-llama-3/). Usage of this model is subject to [Meta's Acceptable Use Policy](https://llama.meta.com/llama3/use-policy/).",
|
|
||||||
"context_length": 8192,
|
|
||||||
"architecture": {
|
|
||||||
"modality": "text->text",
|
|
||||||
"input_modalities": [
|
|
||||||
"text"
|
|
||||||
],
|
|
||||||
"output_modalities": [
|
|
||||||
"text"
|
|
||||||
],
|
|
||||||
"tokenizer": "Llama3",
|
|
||||||
"instruct_type": "llama3"
|
|
||||||
},
|
|
||||||
"pricing": {
|
|
||||||
"prompt": "0",
|
|
||||||
"completion": "0",
|
|
||||||
"request": "0",
|
|
||||||
"image": "0",
|
|
||||||
"web_search": "0",
|
|
||||||
"internal_reasoning": "0",
|
|
||||||
"input_cache_read": "0",
|
|
||||||
"input_cache_write": "0"
|
|
||||||
},
|
|
||||||
"top_provider": {
|
|
||||||
"context_length": 8192,
|
|
||||||
"max_completion_tokens": 4096,
|
|
||||||
"is_moderated": false
|
|
||||||
},
|
|
||||||
"per_request_limits": null
|
|
||||||
},
|
|
||||||
{
|
{
|
||||||
"id": "meta-llama/llama-3-8b-instruct",
|
"id": "meta-llama/llama-3-8b-instruct",
|
||||||
"name": "Meta: Llama 3 8B Instruct",
|
"name": "Meta: Llama 3 8B Instruct",
|
||||||
@ -7703,8 +7670,8 @@
|
|||||||
"instruct_type": null
|
"instruct_type": null
|
||||||
},
|
},
|
||||||
"pricing": {
|
"pricing": {
|
||||||
"prompt": "0.00000285",
|
"prompt": "0.000003",
|
||||||
"completion": "0.00001425",
|
"completion": "0.000015",
|
||||||
"request": "0",
|
"request": "0",
|
||||||
"image": "0",
|
"image": "0",
|
||||||
"web_search": "0",
|
"web_search": "0",
|
||||||
@ -7737,8 +7704,8 @@
|
|||||||
"instruct_type": null
|
"instruct_type": null
|
||||||
},
|
},
|
||||||
"pricing": {
|
"pricing": {
|
||||||
"prompt": "0.00000285",
|
"prompt": "0.000003",
|
||||||
"completion": "0.00001425",
|
"completion": "0.000015",
|
||||||
"request": "0",
|
"request": "0",
|
||||||
"image": "0",
|
"image": "0",
|
||||||
"web_search": "0",
|
"web_search": "0",
|
||||||
@ -7805,8 +7772,8 @@
|
|||||||
"instruct_type": null
|
"instruct_type": null
|
||||||
},
|
},
|
||||||
"pricing": {
|
"pricing": {
|
||||||
"prompt": "0.00000095",
|
"prompt": "0.000001",
|
||||||
"completion": "0.0000019",
|
"completion": "0.000002",
|
||||||
"request": "0",
|
"request": "0",
|
||||||
"image": "0",
|
"image": "0",
|
||||||
"web_search": "0",
|
"web_search": "0",
|
||||||
@ -7839,8 +7806,8 @@
|
|||||||
"instruct_type": null
|
"instruct_type": null
|
||||||
},
|
},
|
||||||
"pricing": {
|
"pricing": {
|
||||||
"prompt": "0.000000475",
|
"prompt": "0.0000005",
|
||||||
"completion": "0.000001425",
|
"completion": "0.0000015",
|
||||||
"request": "0",
|
"request": "0",
|
||||||
"image": "0",
|
"image": "0",
|
||||||
"web_search": "0",
|
"web_search": "0",
|
||||||
@ -8083,8 +8050,8 @@
|
|||||||
"instruct_type": null
|
"instruct_type": null
|
||||||
},
|
},
|
||||||
"pricing": {
|
"pricing": {
|
||||||
"prompt": "0.000000475",
|
"prompt": "0.0000005",
|
||||||
"completion": "0.000001425",
|
"completion": "0.0000015",
|
||||||
"request": "0",
|
"request": "0",
|
||||||
"image": "0",
|
"image": "0",
|
||||||
"web_search": "0",
|
"web_search": "0",
|
||||||
@ -8133,40 +8100,6 @@
|
|||||||
},
|
},
|
||||||
"per_request_limits": null
|
"per_request_limits": null
|
||||||
},
|
},
|
||||||
{
|
|
||||||
"id": "google/gemma-7b-it",
|
|
||||||
"name": "Google: Gemma 7B",
|
|
||||||
"created": 1708560000,
|
|
||||||
"description": "Gemma by Google is an advanced, open-source language model family, leveraging the latest in decoder-only, text-to-text technology. It offers English language capabilities across text generation tasks like question answering, summarization, and reasoning. The Gemma 7B variant is comparable in performance to leading open source models.\n\nUsage of Gemma is subject to Google's [Gemma Terms of Use](https://ai.google.dev/gemma/terms).",
|
|
||||||
"context_length": 8192,
|
|
||||||
"architecture": {
|
|
||||||
"modality": "text->text",
|
|
||||||
"input_modalities": [
|
|
||||||
"text"
|
|
||||||
],
|
|
||||||
"output_modalities": [
|
|
||||||
"text"
|
|
||||||
],
|
|
||||||
"tokenizer": "Gemini",
|
|
||||||
"instruct_type": "gemma"
|
|
||||||
},
|
|
||||||
"pricing": {
|
|
||||||
"prompt": "0.00000015",
|
|
||||||
"completion": "0.00000015",
|
|
||||||
"request": "0",
|
|
||||||
"image": "0",
|
|
||||||
"web_search": "0",
|
|
||||||
"internal_reasoning": "0",
|
|
||||||
"input_cache_read": "0",
|
|
||||||
"input_cache_write": "0"
|
|
||||||
},
|
|
||||||
"top_provider": {
|
|
||||||
"context_length": 8192,
|
|
||||||
"max_completion_tokens": null,
|
|
||||||
"is_moderated": false
|
|
||||||
},
|
|
||||||
"per_request_limits": null
|
|
||||||
},
|
|
||||||
{
|
{
|
||||||
"id": "openai/gpt-3.5-turbo-0613",
|
"id": "openai/gpt-3.5-turbo-0613",
|
||||||
"name": "OpenAI: GPT-3.5 Turbo (older v0613)",
|
"name": "OpenAI: GPT-3.5 Turbo (older v0613)",
|
||||||
@ -8628,8 +8561,8 @@
|
|||||||
"instruct_type": "openchat"
|
"instruct_type": "openchat"
|
||||||
},
|
},
|
||||||
"pricing": {
|
"pricing": {
|
||||||
"prompt": "0.000000055",
|
"prompt": "0.00000007",
|
||||||
"completion": "0.000000055",
|
"completion": "0.00000007",
|
||||||
"request": "0",
|
"request": "0",
|
||||||
"image": "0",
|
"image": "0",
|
||||||
"web_search": "0",
|
"web_search": "0",
|
||||||
@ -8639,7 +8572,7 @@
|
|||||||
},
|
},
|
||||||
"top_provider": {
|
"top_provider": {
|
||||||
"context_length": 8192,
|
"context_length": 8192,
|
||||||
"max_completion_tokens": 8192,
|
"max_completion_tokens": null,
|
||||||
"is_moderated": false
|
"is_moderated": false
|
||||||
},
|
},
|
||||||
"per_request_limits": null
|
"per_request_limits": null
|
||||||
@ -8814,40 +8747,6 @@
|
|||||||
},
|
},
|
||||||
"per_request_limits": null
|
"per_request_limits": null
|
||||||
},
|
},
|
||||||
{
|
|
||||||
"id": "teknium/openhermes-2.5-mistral-7b",
|
|
||||||
"name": "OpenHermes 2.5 Mistral 7B",
|
|
||||||
"created": 1700438400,
|
|
||||||
"description": "A continuation of [OpenHermes 2 model](/models/teknium/openhermes-2-mistral-7b), trained on additional code datasets.\nPotentially the most interesting finding from training on a good ratio (est. of around 7-14% of the total dataset) of code instruction was that it has boosted several non-code benchmarks, including TruthfulQA, AGIEval, and GPT4All suite. It did however reduce BigBench benchmark score, but the net gain overall is significant.",
|
|
||||||
"context_length": 4096,
|
|
||||||
"architecture": {
|
|
||||||
"modality": "text->text",
|
|
||||||
"input_modalities": [
|
|
||||||
"text"
|
|
||||||
],
|
|
||||||
"output_modalities": [
|
|
||||||
"text"
|
|
||||||
],
|
|
||||||
"tokenizer": "Mistral",
|
|
||||||
"instruct_type": "chatml"
|
|
||||||
},
|
|
||||||
"pricing": {
|
|
||||||
"prompt": "0.00000017",
|
|
||||||
"completion": "0.00000017",
|
|
||||||
"request": "0",
|
|
||||||
"image": "0",
|
|
||||||
"web_search": "0",
|
|
||||||
"internal_reasoning": "0",
|
|
||||||
"input_cache_read": "0",
|
|
||||||
"input_cache_write": "0"
|
|
||||||
},
|
|
||||||
"top_provider": {
|
|
||||||
"context_length": 4096,
|
|
||||||
"max_completion_tokens": 4096,
|
|
||||||
"is_moderated": false
|
|
||||||
},
|
|
||||||
"per_request_limits": null
|
|
||||||
},
|
|
||||||
{
|
{
|
||||||
"id": "undi95/toppy-m-7b:free",
|
"id": "undi95/toppy-m-7b:free",
|
||||||
"name": "Toppy M 7B (free)",
|
"name": "Toppy M 7B (free)",
|
||||||
@ -9664,40 +9563,6 @@
|
|||||||
},
|
},
|
||||||
"per_request_limits": null
|
"per_request_limits": null
|
||||||
},
|
},
|
||||||
{
|
|
||||||
"id": "gryphe/mythomax-l2-13b:free",
|
|
||||||
"name": "MythoMax 13B (free)",
|
|
||||||
"created": 1688256000,
|
|
||||||
"description": "One of the highest performing and most popular fine-tunes of Llama 2 13B, with rich descriptions and roleplay. #merge",
|
|
||||||
"context_length": 4096,
|
|
||||||
"architecture": {
|
|
||||||
"modality": "text->text",
|
|
||||||
"input_modalities": [
|
|
||||||
"text"
|
|
||||||
],
|
|
||||||
"output_modalities": [
|
|
||||||
"text"
|
|
||||||
],
|
|
||||||
"tokenizer": "Llama2",
|
|
||||||
"instruct_type": "alpaca"
|
|
||||||
},
|
|
||||||
"pricing": {
|
|
||||||
"prompt": "0",
|
|
||||||
"completion": "0",
|
|
||||||
"request": "0",
|
|
||||||
"image": "0",
|
|
||||||
"web_search": "0",
|
|
||||||
"internal_reasoning": "0",
|
|
||||||
"input_cache_read": "0",
|
|
||||||
"input_cache_write": "0"
|
|
||||||
},
|
|
||||||
"top_provider": {
|
|
||||||
"context_length": 4096,
|
|
||||||
"max_completion_tokens": 2048,
|
|
||||||
"is_moderated": false
|
|
||||||
},
|
|
||||||
"per_request_limits": null
|
|
||||||
},
|
|
||||||
{
|
{
|
||||||
"id": "gryphe/mythomax-l2-13b",
|
"id": "gryphe/mythomax-l2-13b",
|
||||||
"name": "MythoMax 13B",
|
"name": "MythoMax 13B",
|
||||||
|
|||||||
@ -4,10 +4,12 @@ export declare enum E_OPENAI_MODEL {
|
|||||||
MODEL_DALL_E_3 = "dall-e-3",
|
MODEL_DALL_E_3 = "dall-e-3",
|
||||||
MODEL_DALL_E_2 = "dall-e-2",
|
MODEL_DALL_E_2 = "dall-e-2",
|
||||||
MODEL_GPT_4O_AUDIO_PREVIEW_2024_10_01 = "gpt-4o-audio-preview-2024-10-01",
|
MODEL_GPT_4O_AUDIO_PREVIEW_2024_10_01 = "gpt-4o-audio-preview-2024-10-01",
|
||||||
MODEL_GPT_4O_MINI_REALTIME_PREVIEW_2024_12_17 = "gpt-4o-mini-realtime-preview-2024-12-17",
|
MODEL_O1_PRO_2025_03_19 = "o1-pro-2025-03-19",
|
||||||
MODEL_GPT_4O_MINI_REALTIME_PREVIEW = "gpt-4o-mini-realtime-preview",
|
MODEL_O1_PRO = "o1-pro",
|
||||||
MODEL_GPT_4O_REALTIME_PREVIEW_2024_10_01 = "gpt-4o-realtime-preview-2024-10-01",
|
MODEL_GPT_4O_REALTIME_PREVIEW_2024_10_01 = "gpt-4o-realtime-preview-2024-10-01",
|
||||||
|
MODEL_O1_2024_12_17 = "o1-2024-12-17",
|
||||||
MODEL_GPT_4O_TRANSCRIBE = "gpt-4o-transcribe",
|
MODEL_GPT_4O_TRANSCRIBE = "gpt-4o-transcribe",
|
||||||
|
MODEL_O1 = "o1",
|
||||||
MODEL_GPT_4O_MINI_TRANSCRIBE = "gpt-4o-mini-transcribe",
|
MODEL_GPT_4O_MINI_TRANSCRIBE = "gpt-4o-mini-transcribe",
|
||||||
MODEL_GPT_4O_REALTIME_PREVIEW = "gpt-4o-realtime-preview",
|
MODEL_GPT_4O_REALTIME_PREVIEW = "gpt-4o-realtime-preview",
|
||||||
MODEL_BABBAGE_002 = "babbage-002",
|
MODEL_BABBAGE_002 = "babbage-002",
|
||||||
@ -16,19 +18,18 @@ export declare enum E_OPENAI_MODEL {
|
|||||||
MODEL_TEXT_EMBEDDING_3_LARGE = "text-embedding-3-large",
|
MODEL_TEXT_EMBEDDING_3_LARGE = "text-embedding-3-large",
|
||||||
MODEL_GPT_4 = "gpt-4",
|
MODEL_GPT_4 = "gpt-4",
|
||||||
MODEL_TEXT_EMBEDDING_ADA_002 = "text-embedding-ada-002",
|
MODEL_TEXT_EMBEDDING_ADA_002 = "text-embedding-ada-002",
|
||||||
|
MODEL_OMNI_MODERATION_LATEST = "omni-moderation-latest",
|
||||||
MODEL_TTS_1_HD = "tts-1-hd",
|
MODEL_TTS_1_HD = "tts-1-hd",
|
||||||
MODEL_GPT_4O_MINI_AUDIO_PREVIEW = "gpt-4o-mini-audio-preview",
|
MODEL_GPT_4O_MINI_AUDIO_PREVIEW = "gpt-4o-mini-audio-preview",
|
||||||
MODEL_GPT_4O_AUDIO_PREVIEW = "gpt-4o-audio-preview",
|
MODEL_GPT_4O_AUDIO_PREVIEW = "gpt-4o-audio-preview",
|
||||||
MODEL_O1_PREVIEW_2024_09_12 = "o1-preview-2024-09-12",
|
MODEL_O1_PREVIEW_2024_09_12 = "o1-preview-2024-09-12",
|
||||||
|
MODEL_GPT_4O_MINI_REALTIME_PREVIEW = "gpt-4o-mini-realtime-preview",
|
||||||
|
MODEL_GPT_4O_MINI_REALTIME_PREVIEW_2024_12_17 = "gpt-4o-mini-realtime-preview-2024-12-17",
|
||||||
MODEL_GPT_3_5_TURBO_INSTRUCT_0914 = "gpt-3.5-turbo-instruct-0914",
|
MODEL_GPT_3_5_TURBO_INSTRUCT_0914 = "gpt-3.5-turbo-instruct-0914",
|
||||||
MODEL_GPT_4O_MINI_SEARCH_PREVIEW = "gpt-4o-mini-search-preview",
|
MODEL_GPT_4O_MINI_SEARCH_PREVIEW = "gpt-4o-mini-search-preview",
|
||||||
MODEL_O1 = "o1",
|
|
||||||
MODEL_TTS_1_1106 = "tts-1-1106",
|
MODEL_TTS_1_1106 = "tts-1-1106",
|
||||||
MODEL_O1_2024_12_17 = "o1-2024-12-17",
|
|
||||||
MODEL_DAVINCI_002 = "davinci-002",
|
MODEL_DAVINCI_002 = "davinci-002",
|
||||||
MODEL_GPT_3_5_TURBO_1106 = "gpt-3.5-turbo-1106",
|
MODEL_GPT_3_5_TURBO_1106 = "gpt-3.5-turbo-1106",
|
||||||
MODEL_O1_PRO = "o1-pro",
|
|
||||||
MODEL_O1_PRO_2025_03_19 = "o1-pro-2025-03-19",
|
|
||||||
MODEL_GPT_4_TURBO = "gpt-4-turbo",
|
MODEL_GPT_4_TURBO = "gpt-4-turbo",
|
||||||
MODEL_GPT_3_5_TURBO_INSTRUCT = "gpt-3.5-turbo-instruct",
|
MODEL_GPT_3_5_TURBO_INSTRUCT = "gpt-3.5-turbo-instruct",
|
||||||
MODEL_GPT_3_5_TURBO = "gpt-3.5-turbo",
|
MODEL_GPT_3_5_TURBO = "gpt-3.5-turbo",
|
||||||
@ -49,7 +50,6 @@ export declare enum E_OPENAI_MODEL {
|
|||||||
MODEL_GPT_4_5_PREVIEW = "gpt-4.5-preview",
|
MODEL_GPT_4_5_PREVIEW = "gpt-4.5-preview",
|
||||||
MODEL_GPT_4_5_PREVIEW_2025_02_27 = "gpt-4.5-preview-2025-02-27",
|
MODEL_GPT_4_5_PREVIEW_2025_02_27 = "gpt-4.5-preview-2025-02-27",
|
||||||
MODEL_GPT_4O_SEARCH_PREVIEW_2025_03_11 = "gpt-4o-search-preview-2025-03-11",
|
MODEL_GPT_4O_SEARCH_PREVIEW_2025_03_11 = "gpt-4o-search-preview-2025-03-11",
|
||||||
MODEL_OMNI_MODERATION_LATEST = "omni-moderation-latest",
|
|
||||||
MODEL_TTS_1 = "tts-1",
|
MODEL_TTS_1 = "tts-1",
|
||||||
MODEL_OMNI_MODERATION_2024_09_26 = "omni-moderation-2024-09-26",
|
MODEL_OMNI_MODERATION_2024_09_26 = "omni-moderation-2024-09-26",
|
||||||
MODEL_TEXT_EMBEDDING_3_SMALL = "text-embedding-3-small",
|
MODEL_TEXT_EMBEDDING_3_SMALL = "text-embedding-3-small",
|
||||||
@ -57,9 +57,9 @@ export declare enum E_OPENAI_MODEL {
|
|||||||
MODEL_GPT_4O_MINI = "gpt-4o-mini",
|
MODEL_GPT_4O_MINI = "gpt-4o-mini",
|
||||||
MODEL_GPT_4O_2024_08_06 = "gpt-4o-2024-08-06",
|
MODEL_GPT_4O_2024_08_06 = "gpt-4o-2024-08-06",
|
||||||
MODEL_GPT_4O_MINI_2024_07_18 = "gpt-4o-mini-2024-07-18",
|
MODEL_GPT_4O_MINI_2024_07_18 = "gpt-4o-mini-2024-07-18",
|
||||||
MODEL_GPT_4_TURBO_PREVIEW = "gpt-4-turbo-preview",
|
|
||||||
MODEL_O1_MINI = "o1-mini",
|
MODEL_O1_MINI = "o1-mini",
|
||||||
MODEL_GPT_4O_MINI_AUDIO_PREVIEW_2024_12_17 = "gpt-4o-mini-audio-preview-2024-12-17",
|
MODEL_GPT_4O_MINI_AUDIO_PREVIEW_2024_12_17 = "gpt-4o-mini-audio-preview-2024-12-17",
|
||||||
MODEL_O1_MINI_2024_09_12 = "o1-mini-2024-09-12",
|
MODEL_O1_MINI_2024_09_12 = "o1-mini-2024-09-12",
|
||||||
MODEL_GPT_4_0125_PREVIEW = "gpt-4-0125-preview"
|
MODEL_GPT_4_0125_PREVIEW = "gpt-4-0125-preview",
|
||||||
|
MODEL_GPT_4_TURBO_PREVIEW = "gpt-4-turbo-preview"
|
||||||
}
|
}
|
||||||
|
|||||||
@ -5,10 +5,12 @@ export var E_OPENAI_MODEL;
|
|||||||
E_OPENAI_MODEL["MODEL_DALL_E_3"] = "dall-e-3";
|
E_OPENAI_MODEL["MODEL_DALL_E_3"] = "dall-e-3";
|
||||||
E_OPENAI_MODEL["MODEL_DALL_E_2"] = "dall-e-2";
|
E_OPENAI_MODEL["MODEL_DALL_E_2"] = "dall-e-2";
|
||||||
E_OPENAI_MODEL["MODEL_GPT_4O_AUDIO_PREVIEW_2024_10_01"] = "gpt-4o-audio-preview-2024-10-01";
|
E_OPENAI_MODEL["MODEL_GPT_4O_AUDIO_PREVIEW_2024_10_01"] = "gpt-4o-audio-preview-2024-10-01";
|
||||||
E_OPENAI_MODEL["MODEL_GPT_4O_MINI_REALTIME_PREVIEW_2024_12_17"] = "gpt-4o-mini-realtime-preview-2024-12-17";
|
E_OPENAI_MODEL["MODEL_O1_PRO_2025_03_19"] = "o1-pro-2025-03-19";
|
||||||
E_OPENAI_MODEL["MODEL_GPT_4O_MINI_REALTIME_PREVIEW"] = "gpt-4o-mini-realtime-preview";
|
E_OPENAI_MODEL["MODEL_O1_PRO"] = "o1-pro";
|
||||||
E_OPENAI_MODEL["MODEL_GPT_4O_REALTIME_PREVIEW_2024_10_01"] = "gpt-4o-realtime-preview-2024-10-01";
|
E_OPENAI_MODEL["MODEL_GPT_4O_REALTIME_PREVIEW_2024_10_01"] = "gpt-4o-realtime-preview-2024-10-01";
|
||||||
|
E_OPENAI_MODEL["MODEL_O1_2024_12_17"] = "o1-2024-12-17";
|
||||||
E_OPENAI_MODEL["MODEL_GPT_4O_TRANSCRIBE"] = "gpt-4o-transcribe";
|
E_OPENAI_MODEL["MODEL_GPT_4O_TRANSCRIBE"] = "gpt-4o-transcribe";
|
||||||
|
E_OPENAI_MODEL["MODEL_O1"] = "o1";
|
||||||
E_OPENAI_MODEL["MODEL_GPT_4O_MINI_TRANSCRIBE"] = "gpt-4o-mini-transcribe";
|
E_OPENAI_MODEL["MODEL_GPT_4O_MINI_TRANSCRIBE"] = "gpt-4o-mini-transcribe";
|
||||||
E_OPENAI_MODEL["MODEL_GPT_4O_REALTIME_PREVIEW"] = "gpt-4o-realtime-preview";
|
E_OPENAI_MODEL["MODEL_GPT_4O_REALTIME_PREVIEW"] = "gpt-4o-realtime-preview";
|
||||||
E_OPENAI_MODEL["MODEL_BABBAGE_002"] = "babbage-002";
|
E_OPENAI_MODEL["MODEL_BABBAGE_002"] = "babbage-002";
|
||||||
@ -17,19 +19,18 @@ export var E_OPENAI_MODEL;
|
|||||||
E_OPENAI_MODEL["MODEL_TEXT_EMBEDDING_3_LARGE"] = "text-embedding-3-large";
|
E_OPENAI_MODEL["MODEL_TEXT_EMBEDDING_3_LARGE"] = "text-embedding-3-large";
|
||||||
E_OPENAI_MODEL["MODEL_GPT_4"] = "gpt-4";
|
E_OPENAI_MODEL["MODEL_GPT_4"] = "gpt-4";
|
||||||
E_OPENAI_MODEL["MODEL_TEXT_EMBEDDING_ADA_002"] = "text-embedding-ada-002";
|
E_OPENAI_MODEL["MODEL_TEXT_EMBEDDING_ADA_002"] = "text-embedding-ada-002";
|
||||||
|
E_OPENAI_MODEL["MODEL_OMNI_MODERATION_LATEST"] = "omni-moderation-latest";
|
||||||
E_OPENAI_MODEL["MODEL_TTS_1_HD"] = "tts-1-hd";
|
E_OPENAI_MODEL["MODEL_TTS_1_HD"] = "tts-1-hd";
|
||||||
E_OPENAI_MODEL["MODEL_GPT_4O_MINI_AUDIO_PREVIEW"] = "gpt-4o-mini-audio-preview";
|
E_OPENAI_MODEL["MODEL_GPT_4O_MINI_AUDIO_PREVIEW"] = "gpt-4o-mini-audio-preview";
|
||||||
E_OPENAI_MODEL["MODEL_GPT_4O_AUDIO_PREVIEW"] = "gpt-4o-audio-preview";
|
E_OPENAI_MODEL["MODEL_GPT_4O_AUDIO_PREVIEW"] = "gpt-4o-audio-preview";
|
||||||
E_OPENAI_MODEL["MODEL_O1_PREVIEW_2024_09_12"] = "o1-preview-2024-09-12";
|
E_OPENAI_MODEL["MODEL_O1_PREVIEW_2024_09_12"] = "o1-preview-2024-09-12";
|
||||||
|
E_OPENAI_MODEL["MODEL_GPT_4O_MINI_REALTIME_PREVIEW"] = "gpt-4o-mini-realtime-preview";
|
||||||
|
E_OPENAI_MODEL["MODEL_GPT_4O_MINI_REALTIME_PREVIEW_2024_12_17"] = "gpt-4o-mini-realtime-preview-2024-12-17";
|
||||||
E_OPENAI_MODEL["MODEL_GPT_3_5_TURBO_INSTRUCT_0914"] = "gpt-3.5-turbo-instruct-0914";
|
E_OPENAI_MODEL["MODEL_GPT_3_5_TURBO_INSTRUCT_0914"] = "gpt-3.5-turbo-instruct-0914";
|
||||||
E_OPENAI_MODEL["MODEL_GPT_4O_MINI_SEARCH_PREVIEW"] = "gpt-4o-mini-search-preview";
|
E_OPENAI_MODEL["MODEL_GPT_4O_MINI_SEARCH_PREVIEW"] = "gpt-4o-mini-search-preview";
|
||||||
E_OPENAI_MODEL["MODEL_O1"] = "o1";
|
|
||||||
E_OPENAI_MODEL["MODEL_TTS_1_1106"] = "tts-1-1106";
|
E_OPENAI_MODEL["MODEL_TTS_1_1106"] = "tts-1-1106";
|
||||||
E_OPENAI_MODEL["MODEL_O1_2024_12_17"] = "o1-2024-12-17";
|
|
||||||
E_OPENAI_MODEL["MODEL_DAVINCI_002"] = "davinci-002";
|
E_OPENAI_MODEL["MODEL_DAVINCI_002"] = "davinci-002";
|
||||||
E_OPENAI_MODEL["MODEL_GPT_3_5_TURBO_1106"] = "gpt-3.5-turbo-1106";
|
E_OPENAI_MODEL["MODEL_GPT_3_5_TURBO_1106"] = "gpt-3.5-turbo-1106";
|
||||||
E_OPENAI_MODEL["MODEL_O1_PRO"] = "o1-pro";
|
|
||||||
E_OPENAI_MODEL["MODEL_O1_PRO_2025_03_19"] = "o1-pro-2025-03-19";
|
|
||||||
E_OPENAI_MODEL["MODEL_GPT_4_TURBO"] = "gpt-4-turbo";
|
E_OPENAI_MODEL["MODEL_GPT_4_TURBO"] = "gpt-4-turbo";
|
||||||
E_OPENAI_MODEL["MODEL_GPT_3_5_TURBO_INSTRUCT"] = "gpt-3.5-turbo-instruct";
|
E_OPENAI_MODEL["MODEL_GPT_3_5_TURBO_INSTRUCT"] = "gpt-3.5-turbo-instruct";
|
||||||
E_OPENAI_MODEL["MODEL_GPT_3_5_TURBO"] = "gpt-3.5-turbo";
|
E_OPENAI_MODEL["MODEL_GPT_3_5_TURBO"] = "gpt-3.5-turbo";
|
||||||
@ -50,7 +51,6 @@ export var E_OPENAI_MODEL;
|
|||||||
E_OPENAI_MODEL["MODEL_GPT_4_5_PREVIEW"] = "gpt-4.5-preview";
|
E_OPENAI_MODEL["MODEL_GPT_4_5_PREVIEW"] = "gpt-4.5-preview";
|
||||||
E_OPENAI_MODEL["MODEL_GPT_4_5_PREVIEW_2025_02_27"] = "gpt-4.5-preview-2025-02-27";
|
E_OPENAI_MODEL["MODEL_GPT_4_5_PREVIEW_2025_02_27"] = "gpt-4.5-preview-2025-02-27";
|
||||||
E_OPENAI_MODEL["MODEL_GPT_4O_SEARCH_PREVIEW_2025_03_11"] = "gpt-4o-search-preview-2025-03-11";
|
E_OPENAI_MODEL["MODEL_GPT_4O_SEARCH_PREVIEW_2025_03_11"] = "gpt-4o-search-preview-2025-03-11";
|
||||||
E_OPENAI_MODEL["MODEL_OMNI_MODERATION_LATEST"] = "omni-moderation-latest";
|
|
||||||
E_OPENAI_MODEL["MODEL_TTS_1"] = "tts-1";
|
E_OPENAI_MODEL["MODEL_TTS_1"] = "tts-1";
|
||||||
E_OPENAI_MODEL["MODEL_OMNI_MODERATION_2024_09_26"] = "omni-moderation-2024-09-26";
|
E_OPENAI_MODEL["MODEL_OMNI_MODERATION_2024_09_26"] = "omni-moderation-2024-09-26";
|
||||||
E_OPENAI_MODEL["MODEL_TEXT_EMBEDDING_3_SMALL"] = "text-embedding-3-small";
|
E_OPENAI_MODEL["MODEL_TEXT_EMBEDDING_3_SMALL"] = "text-embedding-3-small";
|
||||||
@ -58,10 +58,10 @@ export var E_OPENAI_MODEL;
|
|||||||
E_OPENAI_MODEL["MODEL_GPT_4O_MINI"] = "gpt-4o-mini";
|
E_OPENAI_MODEL["MODEL_GPT_4O_MINI"] = "gpt-4o-mini";
|
||||||
E_OPENAI_MODEL["MODEL_GPT_4O_2024_08_06"] = "gpt-4o-2024-08-06";
|
E_OPENAI_MODEL["MODEL_GPT_4O_2024_08_06"] = "gpt-4o-2024-08-06";
|
||||||
E_OPENAI_MODEL["MODEL_GPT_4O_MINI_2024_07_18"] = "gpt-4o-mini-2024-07-18";
|
E_OPENAI_MODEL["MODEL_GPT_4O_MINI_2024_07_18"] = "gpt-4o-mini-2024-07-18";
|
||||||
E_OPENAI_MODEL["MODEL_GPT_4_TURBO_PREVIEW"] = "gpt-4-turbo-preview";
|
|
||||||
E_OPENAI_MODEL["MODEL_O1_MINI"] = "o1-mini";
|
E_OPENAI_MODEL["MODEL_O1_MINI"] = "o1-mini";
|
||||||
E_OPENAI_MODEL["MODEL_GPT_4O_MINI_AUDIO_PREVIEW_2024_12_17"] = "gpt-4o-mini-audio-preview-2024-12-17";
|
E_OPENAI_MODEL["MODEL_GPT_4O_MINI_AUDIO_PREVIEW_2024_12_17"] = "gpt-4o-mini-audio-preview-2024-12-17";
|
||||||
E_OPENAI_MODEL["MODEL_O1_MINI_2024_09_12"] = "o1-mini-2024-09-12";
|
E_OPENAI_MODEL["MODEL_O1_MINI_2024_09_12"] = "o1-mini-2024-09-12";
|
||||||
E_OPENAI_MODEL["MODEL_GPT_4_0125_PREVIEW"] = "gpt-4-0125-preview";
|
E_OPENAI_MODEL["MODEL_GPT_4_0125_PREVIEW"] = "gpt-4-0125-preview";
|
||||||
|
E_OPENAI_MODEL["MODEL_GPT_4_TURBO_PREVIEW"] = "gpt-4-turbo-preview";
|
||||||
})(E_OPENAI_MODEL || (E_OPENAI_MODEL = {}));
|
})(E_OPENAI_MODEL || (E_OPENAI_MODEL = {}));
|
||||||
//# sourceMappingURL=data:application/json;base64,eyJ2ZXJzaW9uIjozLCJmaWxlIjoib3BlbmFpLW1vZGVscy5qcyIsInNvdXJjZVJvb3QiOiIiLCJzb3VyY2VzIjpbIi4uLy4uLy4uL3NyYy9tb2RlbHMvY2FjaGUvb3BlbmFpLW1vZGVscy50cyJdLCJuYW1lcyI6W10sIm1hcHBpbmdzIjoiQUFBQSxNQUFNLENBQU4sSUFBWSxjQWdFWDtBQWhFRCxXQUFZLGNBQWM7SUFDeEIsaUdBQStFLENBQUE7SUFDL0UsMkZBQXlFLENBQUE7SUFDekUsNkNBQTJCLENBQUE7SUFDM0IsNkNBQTJCLENBQUE7SUFDM0IsMkZBQXlFLENBQUE7SUFDekUsMkdBQXlGLENBQUE7SUFDekYscUZBQW1FLENBQUE7SUFDbkUsaUdBQStFLENBQUE7SUFDL0UsK0RBQTZDLENBQUE7SUFDN0MseUVBQXVELENBQUE7SUFDdkQsMkVBQXlELENBQUE7SUFDekQsbURBQWlDLENBQUE7SUFDakMsMkRBQXlDLENBQUE7SUFDekMsdURBQXFDLENBQUE7SUFDckMseUVBQXVELENBQUE7SUFDdkQsdUNBQXFCLENBQUE7SUFDckIseUVBQXVELENBQUE7SUFDdkQsNkNBQTJCLENBQUE7SUFDM0IsK0VBQTZELENBQUE7SUFDN0QscUVBQW1ELENBQUE7SUFDbkQsdUVBQXFELENBQUE7SUFDckQsbUZBQWlFLENBQUE7SUFDakUsaUZBQStELENBQUE7SUFDL0QsaUNBQWUsQ0FBQTtJQUNmLGlEQUErQixDQUFBO0lBQy9CLHVEQUFxQyxDQUFBO0lBQ3JDLG1EQUFpQyxDQUFBO0lBQ2pDLGlFQUErQyxDQUFBO0lBQy9DLHlDQUF1QixDQUFBO0lBQ3ZCLCtEQUE2QyxDQUFBO0lBQzdDLG1EQUFpQyxDQUFBO0lBQ2pDLHlFQUF1RCxDQUFBO0lBQ3ZELHVEQUFxQyxDQUFBO0lBQ3JDLCtEQUE2QyxDQUFBO0lBQzdDLHVHQUFxRixDQUFBO0lBQ3JGLCtEQUE2QyxDQUFBO0lBQzdDLCtDQUE2QixDQUFBO0lBQzdCLGlFQUErQyxDQUFBO0lBQy9DLCtEQUE2QyxDQUFBO0lBQzdDLCtEQUE2QyxDQUFBO0lBQzdDLHlFQUF1RCxDQUFBO0lBQ3ZELGlFQUErQyxDQUFBO0lBQy9DLGlEQUErQixDQUFBO0lBQy9CLGlEQUErQixDQUFBO0lBQy9CLHVFQUFxRCxDQUFBO0lBQ3JELGlFQUErQyxDQUFBO0lBQy9DLDJDQUF5QixDQUFBO0lBQ3pCLDJEQUF5QyxDQUFBO0lBQ3pDLGlGQUErRCxDQUFBO0lBQy9ELDZGQUEyRSxDQUFBO0lBQzNFLHlFQUF1RCxDQUFBO0lBQ3ZELHVDQUFxQixDQUFBO0lBQ3JCLGlGQUErRCxDQUFBO0lBQy9ELHlFQUF1RCxDQUFBO0lBQ3ZELHlDQUF1QixDQUFBO0lBQ3ZCLG1EQUFpQyxDQUFBO0lBQ2pDLCtEQUE2QyxDQUFBO0lBQzdDLHlFQUF1RCxDQUFBO0lBQ3ZELG1FQUFpRCxDQUFBO0lBQ2pELDJDQUF5QixDQUFBO0lBQ3pCLHFHQUFtRixDQUFBO0lBQ25GLGlFQUErQyxDQUFBO0lBQy9DLGlFQUErQyxDQUFBO0FBQ2pELENBQUMsRUFoRVcsY0FBYyxLQUFkLGNBQWMsUUFnRXpCIn0=
|
//# sourceMappingURL=data:application/json;base64,eyJ2ZXJzaW9uIjozLCJmaWxlIjoib3BlbmFpLW1vZGVscy5qcyIsInNvdXJjZVJvb3QiOiIiLCJzb3VyY2VzIjpbIi4uLy4uLy4uL3NyYy9tb2RlbHMvY2FjaGUvb3BlbmFpLW1vZGVscy50cyJdLCJuYW1lcyI6W10sIm1hcHBpbmdzIjoiQUFBQSxNQUFNLENBQU4sSUFBWSxjQWdFWDtBQWhFRCxXQUFZLGNBQWM7SUFDeEIsaUdBQStFLENBQUE7SUFDL0UsMkZBQXlFLENBQUE7SUFDekUsNkNBQTJCLENBQUE7SUFDM0IsNkNBQTJCLENBQUE7SUFDM0IsMkZBQXlFLENBQUE7SUFDekUsK0RBQTZDLENBQUE7SUFDN0MseUNBQXVCLENBQUE7SUFDdkIsaUdBQStFLENBQUE7SUFDL0UsdURBQXFDLENBQUE7SUFDckMsK0RBQTZDLENBQUE7SUFDN0MsaUNBQWUsQ0FBQTtJQUNmLHlFQUF1RCxDQUFBO0lBQ3ZELDJFQUF5RCxDQUFBO0lBQ3pELG1EQUFpQyxDQUFBO0lBQ2pDLDJEQUF5QyxDQUFBO0lBQ3pDLHVEQUFxQyxDQUFBO0lBQ3JDLHlFQUF1RCxDQUFBO0lBQ3ZELHVDQUFxQixDQUFBO0lBQ3JCLHlFQUF1RCxDQUFBO0lBQ3ZELHlFQUF1RCxDQUFBO0lBQ3ZELDZDQUEyQixDQUFBO0lBQzNCLCtFQUE2RCxDQUFBO0lBQzdELHFFQUFtRCxDQUFBO0lBQ25ELHVFQUFxRCxDQUFBO0lBQ3JELHFGQUFtRSxDQUFBO0lBQ25FLDJHQUF5RixDQUFBO0lBQ3pGLG1GQUFpRSxDQUFBO0lBQ2pFLGlGQUErRCxDQUFBO0lBQy9ELGlEQUErQixDQUFBO0lBQy9CLG1EQUFpQyxDQUFBO0lBQ2pDLGlFQUErQyxDQUFBO0lBQy9DLG1EQUFpQyxDQUFBO0lBQ2pDLHlFQUF1RCxDQUFBO0lBQ3ZELHVEQUFxQyxDQUFBO0lBQ3JDLCtEQUE2QyxDQUFBO0lBQzdDLHVHQUFxRixDQUFBO0lBQ3JGLCtEQUE2QyxDQUFBO0lBQzdDLCtDQUE2QixDQUFBO0lBQzdCLGlFQUErQyxDQUFBO0lBQy9DLCtEQUE2QyxDQUFBO0lBQzdDLCtEQUE2QyxDQUFBO0lBQzdDLHlFQUF1RCxDQUFBO0lBQ3ZELGlFQUErQyxDQUFBO0lBQy9DLGlEQUErQixDQUFBO0lBQy9CLGlEQUErQixDQUFBO0lBQy9CLHVFQUFxRCxDQUFBO0lBQ3JELGlFQUErQyxDQUFBO0lBQy9DLDJDQUF5QixDQUFBO0lBQ3pCLDJEQUF5QyxDQUFBO0lBQ3pDLGlGQUErRCxDQUFBO0lBQy9ELDZGQUEyRSxDQUFBO0lBQzNFLHVDQUFxQixDQUFBO0lBQ3JCLGlGQUErRCxDQUFBO0lBQy9ELHlFQUF1RCxDQUFBO0lBQ3ZELHlDQUF1QixDQUFBO0lBQ3ZCLG1EQUFpQyxDQUFBO0lBQ2pDLCtEQUE2QyxDQUFBO0lBQzdDLHlFQUF1RCxDQUFBO0lBQ3ZELDJDQUF5QixDQUFBO0lBQ3pCLHFHQUFtRixDQUFBO0lBQ25GLGlFQUErQyxDQUFBO0lBQy9DLGlFQUErQyxDQUFBO0lBQy9DLG1FQUFpRCxDQUFBO0FBQ25ELENBQUMsRUFoRVcsY0FBYyxLQUFkLGNBQWMsUUFnRXpCIn0=
|
||||||
@ -1,4 +1,5 @@
|
|||||||
export declare enum E_OPENROUTER_MODEL_FREE {
|
export declare enum E_OPENROUTER_MODEL_FREE {
|
||||||
|
MODEL_FREE_OPENROUTER_QUASAR_ALPHA = "openrouter/quasar-alpha",
|
||||||
MODEL_FREE_DEEPSEEK_DEEPSEEK_V3_BASE_FREE = "deepseek/deepseek-v3-base:free",
|
MODEL_FREE_DEEPSEEK_DEEPSEEK_V3_BASE_FREE = "deepseek/deepseek-v3-base:free",
|
||||||
MODEL_FREE_ALLENAI_MOLMO_7B_D_FREE = "allenai/molmo-7b-d:free",
|
MODEL_FREE_ALLENAI_MOLMO_7B_D_FREE = "allenai/molmo-7b-d:free",
|
||||||
MODEL_FREE_BYTEDANCE_RESEARCH_UI_TARS_72B_FREE = "bytedance-research/ui-tars-72b:free",
|
MODEL_FREE_BYTEDANCE_RESEARCH_UI_TARS_72B_FREE = "bytedance-research/ui-tars-72b:free",
|
||||||
@ -46,14 +47,11 @@ export declare enum E_OPENROUTER_MODEL_FREE {
|
|||||||
MODEL_FREE_GOOGLE_GEMINI_FLASH_1_5_8B_EXP = "google/gemini-flash-1.5-8b-exp",
|
MODEL_FREE_GOOGLE_GEMINI_FLASH_1_5_8B_EXP = "google/gemini-flash-1.5-8b-exp",
|
||||||
MODEL_FREE_META_LLAMA_LLAMA_3_1_8B_INSTRUCT_FREE = "meta-llama/llama-3.1-8b-instruct:free",
|
MODEL_FREE_META_LLAMA_LLAMA_3_1_8B_INSTRUCT_FREE = "meta-llama/llama-3.1-8b-instruct:free",
|
||||||
MODEL_FREE_MISTRALAI_MISTRAL_NEMO_FREE = "mistralai/mistral-nemo:free",
|
MODEL_FREE_MISTRALAI_MISTRAL_NEMO_FREE = "mistralai/mistral-nemo:free",
|
||||||
MODEL_FREE_QWEN_QWEN_2_7B_INSTRUCT_FREE = "qwen/qwen-2-7b-instruct:free",
|
|
||||||
MODEL_FREE_GOOGLE_GEMMA_2_9B_IT_FREE = "google/gemma-2-9b-it:free",
|
MODEL_FREE_GOOGLE_GEMMA_2_9B_IT_FREE = "google/gemma-2-9b-it:free",
|
||||||
MODEL_FREE_MISTRALAI_MISTRAL_7B_INSTRUCT_FREE = "mistralai/mistral-7b-instruct:free",
|
MODEL_FREE_MISTRALAI_MISTRAL_7B_INSTRUCT_FREE = "mistralai/mistral-7b-instruct:free",
|
||||||
MODEL_FREE_MICROSOFT_PHI_3_MINI_128K_INSTRUCT_FREE = "microsoft/phi-3-mini-128k-instruct:free",
|
MODEL_FREE_MICROSOFT_PHI_3_MINI_128K_INSTRUCT_FREE = "microsoft/phi-3-mini-128k-instruct:free",
|
||||||
MODEL_FREE_MICROSOFT_PHI_3_MEDIUM_128K_INSTRUCT_FREE = "microsoft/phi-3-medium-128k-instruct:free",
|
MODEL_FREE_MICROSOFT_PHI_3_MEDIUM_128K_INSTRUCT_FREE = "microsoft/phi-3-medium-128k-instruct:free",
|
||||||
MODEL_FREE_META_LLAMA_LLAMA_3_8B_INSTRUCT_FREE = "meta-llama/llama-3-8b-instruct:free",
|
|
||||||
MODEL_FREE_OPENCHAT_OPENCHAT_7B_FREE = "openchat/openchat-7b:free",
|
MODEL_FREE_OPENCHAT_OPENCHAT_7B_FREE = "openchat/openchat-7b:free",
|
||||||
MODEL_FREE_UNDI95_TOPPY_M_7B_FREE = "undi95/toppy-m-7b:free",
|
MODEL_FREE_UNDI95_TOPPY_M_7B_FREE = "undi95/toppy-m-7b:free",
|
||||||
MODEL_FREE_HUGGINGFACEH4_ZEPHYR_7B_BETA_FREE = "huggingfaceh4/zephyr-7b-beta:free",
|
MODEL_FREE_HUGGINGFACEH4_ZEPHYR_7B_BETA_FREE = "huggingfaceh4/zephyr-7b-beta:free"
|
||||||
MODEL_FREE_GRYPHE_MYTHOMAX_L2_13B_FREE = "gryphe/mythomax-l2-13b:free"
|
|
||||||
}
|
}
|
||||||
|
|||||||
@ -1,5 +1,6 @@
|
|||||||
export var E_OPENROUTER_MODEL_FREE;
|
export var E_OPENROUTER_MODEL_FREE;
|
||||||
(function (E_OPENROUTER_MODEL_FREE) {
|
(function (E_OPENROUTER_MODEL_FREE) {
|
||||||
|
E_OPENROUTER_MODEL_FREE["MODEL_FREE_OPENROUTER_QUASAR_ALPHA"] = "openrouter/quasar-alpha";
|
||||||
E_OPENROUTER_MODEL_FREE["MODEL_FREE_DEEPSEEK_DEEPSEEK_V3_BASE_FREE"] = "deepseek/deepseek-v3-base:free";
|
E_OPENROUTER_MODEL_FREE["MODEL_FREE_DEEPSEEK_DEEPSEEK_V3_BASE_FREE"] = "deepseek/deepseek-v3-base:free";
|
||||||
E_OPENROUTER_MODEL_FREE["MODEL_FREE_ALLENAI_MOLMO_7B_D_FREE"] = "allenai/molmo-7b-d:free";
|
E_OPENROUTER_MODEL_FREE["MODEL_FREE_ALLENAI_MOLMO_7B_D_FREE"] = "allenai/molmo-7b-d:free";
|
||||||
E_OPENROUTER_MODEL_FREE["MODEL_FREE_BYTEDANCE_RESEARCH_UI_TARS_72B_FREE"] = "bytedance-research/ui-tars-72b:free";
|
E_OPENROUTER_MODEL_FREE["MODEL_FREE_BYTEDANCE_RESEARCH_UI_TARS_72B_FREE"] = "bytedance-research/ui-tars-72b:free";
|
||||||
@ -47,15 +48,12 @@ export var E_OPENROUTER_MODEL_FREE;
|
|||||||
E_OPENROUTER_MODEL_FREE["MODEL_FREE_GOOGLE_GEMINI_FLASH_1_5_8B_EXP"] = "google/gemini-flash-1.5-8b-exp";
|
E_OPENROUTER_MODEL_FREE["MODEL_FREE_GOOGLE_GEMINI_FLASH_1_5_8B_EXP"] = "google/gemini-flash-1.5-8b-exp";
|
||||||
E_OPENROUTER_MODEL_FREE["MODEL_FREE_META_LLAMA_LLAMA_3_1_8B_INSTRUCT_FREE"] = "meta-llama/llama-3.1-8b-instruct:free";
|
E_OPENROUTER_MODEL_FREE["MODEL_FREE_META_LLAMA_LLAMA_3_1_8B_INSTRUCT_FREE"] = "meta-llama/llama-3.1-8b-instruct:free";
|
||||||
E_OPENROUTER_MODEL_FREE["MODEL_FREE_MISTRALAI_MISTRAL_NEMO_FREE"] = "mistralai/mistral-nemo:free";
|
E_OPENROUTER_MODEL_FREE["MODEL_FREE_MISTRALAI_MISTRAL_NEMO_FREE"] = "mistralai/mistral-nemo:free";
|
||||||
E_OPENROUTER_MODEL_FREE["MODEL_FREE_QWEN_QWEN_2_7B_INSTRUCT_FREE"] = "qwen/qwen-2-7b-instruct:free";
|
|
||||||
E_OPENROUTER_MODEL_FREE["MODEL_FREE_GOOGLE_GEMMA_2_9B_IT_FREE"] = "google/gemma-2-9b-it:free";
|
E_OPENROUTER_MODEL_FREE["MODEL_FREE_GOOGLE_GEMMA_2_9B_IT_FREE"] = "google/gemma-2-9b-it:free";
|
||||||
E_OPENROUTER_MODEL_FREE["MODEL_FREE_MISTRALAI_MISTRAL_7B_INSTRUCT_FREE"] = "mistralai/mistral-7b-instruct:free";
|
E_OPENROUTER_MODEL_FREE["MODEL_FREE_MISTRALAI_MISTRAL_7B_INSTRUCT_FREE"] = "mistralai/mistral-7b-instruct:free";
|
||||||
E_OPENROUTER_MODEL_FREE["MODEL_FREE_MICROSOFT_PHI_3_MINI_128K_INSTRUCT_FREE"] = "microsoft/phi-3-mini-128k-instruct:free";
|
E_OPENROUTER_MODEL_FREE["MODEL_FREE_MICROSOFT_PHI_3_MINI_128K_INSTRUCT_FREE"] = "microsoft/phi-3-mini-128k-instruct:free";
|
||||||
E_OPENROUTER_MODEL_FREE["MODEL_FREE_MICROSOFT_PHI_3_MEDIUM_128K_INSTRUCT_FREE"] = "microsoft/phi-3-medium-128k-instruct:free";
|
E_OPENROUTER_MODEL_FREE["MODEL_FREE_MICROSOFT_PHI_3_MEDIUM_128K_INSTRUCT_FREE"] = "microsoft/phi-3-medium-128k-instruct:free";
|
||||||
E_OPENROUTER_MODEL_FREE["MODEL_FREE_META_LLAMA_LLAMA_3_8B_INSTRUCT_FREE"] = "meta-llama/llama-3-8b-instruct:free";
|
|
||||||
E_OPENROUTER_MODEL_FREE["MODEL_FREE_OPENCHAT_OPENCHAT_7B_FREE"] = "openchat/openchat-7b:free";
|
E_OPENROUTER_MODEL_FREE["MODEL_FREE_OPENCHAT_OPENCHAT_7B_FREE"] = "openchat/openchat-7b:free";
|
||||||
E_OPENROUTER_MODEL_FREE["MODEL_FREE_UNDI95_TOPPY_M_7B_FREE"] = "undi95/toppy-m-7b:free";
|
E_OPENROUTER_MODEL_FREE["MODEL_FREE_UNDI95_TOPPY_M_7B_FREE"] = "undi95/toppy-m-7b:free";
|
||||||
E_OPENROUTER_MODEL_FREE["MODEL_FREE_HUGGINGFACEH4_ZEPHYR_7B_BETA_FREE"] = "huggingfaceh4/zephyr-7b-beta:free";
|
E_OPENROUTER_MODEL_FREE["MODEL_FREE_HUGGINGFACEH4_ZEPHYR_7B_BETA_FREE"] = "huggingfaceh4/zephyr-7b-beta:free";
|
||||||
E_OPENROUTER_MODEL_FREE["MODEL_FREE_GRYPHE_MYTHOMAX_L2_13B_FREE"] = "gryphe/mythomax-l2-13b:free";
|
|
||||||
})(E_OPENROUTER_MODEL_FREE || (E_OPENROUTER_MODEL_FREE = {}));
|
})(E_OPENROUTER_MODEL_FREE || (E_OPENROUTER_MODEL_FREE = {}));
|
||||||
//# sourceMappingURL=data:application/json;base64,eyJ2ZXJzaW9uIjozLCJmaWxlIjoib3BlbnJvdXRlci1tb2RlbHMtZnJlZS5qcyIsInNvdXJjZVJvb3QiOiIiLCJzb3VyY2VzIjpbIi4uLy4uLy4uL3NyYy9tb2RlbHMvY2FjaGUvb3BlbnJvdXRlci1tb2RlbHMtZnJlZS50cyJdLCJuYW1lcyI6W10sIm1hcHBpbmdzIjoiQUFBQSxNQUFNLENBQU4sSUFBWSx1QkEwRFg7QUExREQsV0FBWSx1QkFBdUI7SUFDakMsdUdBQTRFLENBQUE7SUFDNUUseUZBQThELENBQUE7SUFDOUQsaUhBQXNGLENBQUE7SUFDdEYsMkdBQWdGLENBQUE7SUFDaEYsbUhBQXdGLENBQUE7SUFDeEYsNkdBQWtGLENBQUE7SUFDbEYsaUhBQXNGLENBQUE7SUFDdEYsaUdBQXNFLENBQUE7SUFDdEUscUlBQTBHLENBQUE7SUFDMUcsbUdBQXdFLENBQUE7SUFDeEUscUdBQTBFLENBQUE7SUFDMUUsNkZBQWtFLENBQUE7SUFDbEUsNkZBQWtFLENBQUE7SUFDbEUsK0ZBQW9FLENBQUE7SUFDcEUsMkZBQWdFLENBQUE7SUFDaEUsK0ZBQW9FLENBQUE7SUFDcEUsdUdBQTRFLENBQUE7SUFDNUUsNkVBQWtELENBQUE7SUFDbEQsK0hBQW9HLENBQUE7SUFDcEcsNklBQWtILENBQUE7SUFDbEgsbUpBQXdILENBQUE7SUFDeEgsNklBQWtILENBQUE7SUFDbEgsbUhBQXdGLENBQUE7SUFDeEYsNkdBQWtGLENBQUE7SUFDbEYsdUlBQTRHLENBQUE7SUFDNUcsK0hBQW9HLENBQUE7SUFDcEcsK0hBQW9HLENBQUE7SUFDcEcsaUlBQXNHLENBQUE7SUFDdEcsNkhBQWtHLENBQUE7SUFDbEcsNkZBQWtFLENBQUE7SUFDbEUsNkhBQWtHLENBQUE7SUFDbEcsaUdBQXNFLENBQUE7SUFDdEUsdUlBQTRHLENBQUE7SUFDNUcsMkdBQWdGLENBQUE7SUFDaEYsdUhBQTRGLENBQUE7SUFDNUYsNkZBQWtFLENBQUE7SUFDbEUsMkhBQWdHLENBQUE7SUFDaEcscUhBQTBGLENBQUE7SUFDMUYsaUlBQXNHLENBQUE7SUFDdEcscUhBQTBGLENBQUE7SUFDMUYscUhBQTBGLENBQUE7SUFDMUYscUlBQTBHLENBQUE7SUFDMUcseUdBQThFLENBQUE7SUFDOUUsNkdBQWtGLENBQUE7SUFDbEYsdUdBQTRFLENBQUE7SUFDNUUscUhBQTBGLENBQUE7SUFDMUYsaUdBQXNFLENBQUE7SUFDdEUsbUdBQXdFLENBQUE7SUFDeEUsNkZBQWtFLENBQUE7SUFDbEUsK0dBQW9GLENBQUE7SUFDcEYseUhBQThGLENBQUE7SUFDOUYsNkhBQWtHLENBQUE7SUFDbEcsaUhBQXNGLENBQUE7SUFDdEYsNkZBQWtFLENBQUE7SUFDbEUsdUZBQTRELENBQUE7SUFDNUQsNkdBQWtGLENBQUE7SUFDbEYsaUdBQXNFLENBQUE7QUFDeEUsQ0FBQyxFQTFEVyx1QkFBdUIsS0FBdkIsdUJBQXVCLFFBMERsQyJ9
|
//# sourceMappingURL=data:application/json;base64,eyJ2ZXJzaW9uIjozLCJmaWxlIjoib3BlbnJvdXRlci1tb2RlbHMtZnJlZS5qcyIsInNvdXJjZVJvb3QiOiIiLCJzb3VyY2VzIjpbIi4uLy4uLy4uL3NyYy9tb2RlbHMvY2FjaGUvb3BlbnJvdXRlci1tb2RlbHMtZnJlZS50cyJdLCJuYW1lcyI6W10sIm1hcHBpbmdzIjoiQUFBQSxNQUFNLENBQU4sSUFBWSx1QkF3RFg7QUF4REQsV0FBWSx1QkFBdUI7SUFDakMseUZBQThELENBQUE7SUFDOUQsdUdBQTRFLENBQUE7SUFDNUUseUZBQThELENBQUE7SUFDOUQsaUhBQXNGLENBQUE7SUFDdEYsMkdBQWdGLENBQUE7SUFDaEYsbUhBQXdGLENBQUE7SUFDeEYsNkdBQWtGLENBQUE7SUFDbEYsaUhBQXNGLENBQUE7SUFDdEYsaUdBQXNFLENBQUE7SUFDdEUscUlBQTBHLENBQUE7SUFDMUcsbUdBQXdFLENBQUE7SUFDeEUscUdBQTBFLENBQUE7SUFDMUUsNkZBQWtFLENBQUE7SUFDbEUsNkZBQWtFLENBQUE7SUFDbEUsK0ZBQW9FLENBQUE7SUFDcEUsMkZBQWdFLENBQUE7SUFDaEUsK0ZBQW9FLENBQUE7SUFDcEUsdUdBQTRFLENBQUE7SUFDNUUsNkVBQWtELENBQUE7SUFDbEQsK0hBQW9HLENBQUE7SUFDcEcsNklBQWtILENBQUE7SUFDbEgsbUpBQXdILENBQUE7SUFDeEgsNklBQWtILENBQUE7SUFDbEgsbUhBQXdGLENBQUE7SUFDeEYsNkdBQWtGLENBQUE7SUFDbEYsdUlBQTRHLENBQUE7SUFDNUcsK0hBQW9HLENBQUE7SUFDcEcsK0hBQW9HLENBQUE7SUFDcEcsaUlBQXNHLENBQUE7SUFDdEcsNkhBQWtHLENBQUE7SUFDbEcsNkZBQWtFLENBQUE7SUFDbEUsNkhBQWtHLENBQUE7SUFDbEcsaUdBQXNFLENBQUE7SUFDdEUsdUlBQTRHLENBQUE7SUFDNUcsMkdBQWdGLENBQUE7SUFDaEYsdUhBQTRGLENBQUE7SUFDNUYsNkZBQWtFLENBQUE7SUFDbEUsMkhBQWdHLENBQUE7SUFDaEcscUhBQTBGLENBQUE7SUFDMUYsaUlBQXNHLENBQUE7SUFDdEcscUhBQTBGLENBQUE7SUFDMUYscUhBQTBGLENBQUE7SUFDMUYscUlBQTBHLENBQUE7SUFDMUcseUdBQThFLENBQUE7SUFDOUUsNkdBQWtGLENBQUE7SUFDbEYsdUdBQTRFLENBQUE7SUFDNUUscUhBQTBGLENBQUE7SUFDMUYsaUdBQXNFLENBQUE7SUFDdEUsNkZBQWtFLENBQUE7SUFDbEUsK0dBQW9GLENBQUE7SUFDcEYseUhBQThGLENBQUE7SUFDOUYsNkhBQWtHLENBQUE7SUFDbEcsNkZBQWtFLENBQUE7SUFDbEUsdUZBQTRELENBQUE7SUFDNUQsNkdBQWtGLENBQUE7QUFDcEYsQ0FBQyxFQXhEVyx1QkFBdUIsS0FBdkIsdUJBQXVCLFFBd0RsQyJ9
|
||||||
@ -1,4 +1,6 @@
|
|||||||
export declare enum E_OPENROUTER_MODEL {
|
export declare enum E_OPENROUTER_MODEL {
|
||||||
|
MODEL_OPENROUTER_QUASAR_ALPHA = "openrouter/quasar-alpha",
|
||||||
|
MODEL_ALL_HANDS_OPENHANDS_LM_32B_V0_1 = "all-hands/openhands-lm-32b-v0.1",
|
||||||
MODEL_MISTRAL_MINISTRAL_8B = "mistral/ministral-8b",
|
MODEL_MISTRAL_MINISTRAL_8B = "mistral/ministral-8b",
|
||||||
MODEL_DEEPSEEK_DEEPSEEK_V3_BASE_FREE = "deepseek/deepseek-v3-base:free",
|
MODEL_DEEPSEEK_DEEPSEEK_V3_BASE_FREE = "deepseek/deepseek-v3-base:free",
|
||||||
MODEL_SCB10X_LLAMA3_1_TYPHOON2_8B_INSTRUCT = "scb10x/llama3.1-typhoon2-8b-instruct",
|
MODEL_SCB10X_LLAMA3_1_TYPHOON2_8B_INSTRUCT = "scb10x/llama3.1-typhoon2-8b-instruct",
|
||||||
@ -47,9 +49,9 @@ export declare enum E_OPENROUTER_MODEL {
|
|||||||
MODEL_NOUSRESEARCH_DEEPHERMES_3_LLAMA_3_8B_PREVIEW_FREE = "nousresearch/deephermes-3-llama-3-8b-preview:free",
|
MODEL_NOUSRESEARCH_DEEPHERMES_3_LLAMA_3_8B_PREVIEW_FREE = "nousresearch/deephermes-3-llama-3-8b-preview:free",
|
||||||
MODEL_OPENAI_GPT_4_5_PREVIEW = "openai/gpt-4.5-preview",
|
MODEL_OPENAI_GPT_4_5_PREVIEW = "openai/gpt-4.5-preview",
|
||||||
MODEL_GOOGLE_GEMINI_2_0_FLASH_LITE_001 = "google/gemini-2.0-flash-lite-001",
|
MODEL_GOOGLE_GEMINI_2_0_FLASH_LITE_001 = "google/gemini-2.0-flash-lite-001",
|
||||||
MODEL_ANTHROPIC_CLAUDE_3_7_SONNET_BETA = "anthropic/claude-3.7-sonnet:beta",
|
|
||||||
MODEL_ANTHROPIC_CLAUDE_3_7_SONNET = "anthropic/claude-3.7-sonnet",
|
MODEL_ANTHROPIC_CLAUDE_3_7_SONNET = "anthropic/claude-3.7-sonnet",
|
||||||
MODEL_ANTHROPIC_CLAUDE_3_7_SONNET_THINKING = "anthropic/claude-3.7-sonnet:thinking",
|
MODEL_ANTHROPIC_CLAUDE_3_7_SONNET_THINKING = "anthropic/claude-3.7-sonnet:thinking",
|
||||||
|
MODEL_ANTHROPIC_CLAUDE_3_7_SONNET_BETA = "anthropic/claude-3.7-sonnet:beta",
|
||||||
MODEL_PERPLEXITY_R1_1776 = "perplexity/r1-1776",
|
MODEL_PERPLEXITY_R1_1776 = "perplexity/r1-1776",
|
||||||
MODEL_MISTRALAI_MISTRAL_SABA = "mistralai/mistral-saba",
|
MODEL_MISTRALAI_MISTRAL_SABA = "mistralai/mistral-saba",
|
||||||
MODEL_COGNITIVECOMPUTATIONS_DOLPHIN3_0_R1_MISTRAL_24B_FREE = "cognitivecomputations/dolphin3.0-r1-mistral-24b:free",
|
MODEL_COGNITIVECOMPUTATIONS_DOLPHIN3_0_R1_MISTRAL_24B_FREE = "cognitivecomputations/dolphin3.0-r1-mistral-24b:free",
|
||||||
@ -185,8 +187,6 @@ export declare enum E_OPENROUTER_MODEL {
|
|||||||
MODEL_MISTRALAI_MISTRAL_NEMO = "mistralai/mistral-nemo",
|
MODEL_MISTRALAI_MISTRAL_NEMO = "mistralai/mistral-nemo",
|
||||||
MODEL_OPENAI_GPT_4O_MINI = "openai/gpt-4o-mini",
|
MODEL_OPENAI_GPT_4O_MINI = "openai/gpt-4o-mini",
|
||||||
MODEL_OPENAI_GPT_4O_MINI_2024_07_18 = "openai/gpt-4o-mini-2024-07-18",
|
MODEL_OPENAI_GPT_4O_MINI_2024_07_18 = "openai/gpt-4o-mini-2024-07-18",
|
||||||
MODEL_QWEN_QWEN_2_7B_INSTRUCT_FREE = "qwen/qwen-2-7b-instruct:free",
|
|
||||||
MODEL_QWEN_QWEN_2_7B_INSTRUCT = "qwen/qwen-2-7b-instruct",
|
|
||||||
MODEL_GOOGLE_GEMMA_2_27B_IT = "google/gemma-2-27b-it",
|
MODEL_GOOGLE_GEMMA_2_27B_IT = "google/gemma-2-27b-it",
|
||||||
MODEL_ALPINDALE_MAGNUM_72B = "alpindale/magnum-72b",
|
MODEL_ALPINDALE_MAGNUM_72B = "alpindale/magnum-72b",
|
||||||
MODEL_GOOGLE_GEMMA_2_9B_IT_FREE = "google/gemma-2-9b-it:free",
|
MODEL_GOOGLE_GEMMA_2_9B_IT_FREE = "google/gemma-2-9b-it:free",
|
||||||
@ -215,7 +215,6 @@ export declare enum E_OPENROUTER_MODEL {
|
|||||||
MODEL_NEVERSLEEP_LLAMA_3_LUMIMAID_8B_EXTENDED = "neversleep/llama-3-lumimaid-8b:extended",
|
MODEL_NEVERSLEEP_LLAMA_3_LUMIMAID_8B_EXTENDED = "neversleep/llama-3-lumimaid-8b:extended",
|
||||||
MODEL_NEVERSLEEP_LLAMA_3_LUMIMAID_8B = "neversleep/llama-3-lumimaid-8b",
|
MODEL_NEVERSLEEP_LLAMA_3_LUMIMAID_8B = "neversleep/llama-3-lumimaid-8b",
|
||||||
MODEL_SAO10K_FIMBULVETR_11B_V2 = "sao10k/fimbulvetr-11b-v2",
|
MODEL_SAO10K_FIMBULVETR_11B_V2 = "sao10k/fimbulvetr-11b-v2",
|
||||||
MODEL_META_LLAMA_LLAMA_3_8B_INSTRUCT_FREE = "meta-llama/llama-3-8b-instruct:free",
|
|
||||||
MODEL_META_LLAMA_LLAMA_3_8B_INSTRUCT = "meta-llama/llama-3-8b-instruct",
|
MODEL_META_LLAMA_LLAMA_3_8B_INSTRUCT = "meta-llama/llama-3-8b-instruct",
|
||||||
MODEL_META_LLAMA_LLAMA_3_70B_INSTRUCT = "meta-llama/llama-3-70b-instruct",
|
MODEL_META_LLAMA_LLAMA_3_70B_INSTRUCT = "meta-llama/llama-3-70b-instruct",
|
||||||
MODEL_MISTRALAI_MIXTRAL_8X22B_INSTRUCT = "mistralai/mixtral-8x22b-instruct",
|
MODEL_MISTRALAI_MIXTRAL_8X22B_INSTRUCT = "mistralai/mixtral-8x22b-instruct",
|
||||||
@ -236,7 +235,6 @@ export declare enum E_OPENROUTER_MODEL {
|
|||||||
MODEL_ANTHROPIC_CLAUDE_3_SONNET = "anthropic/claude-3-sonnet",
|
MODEL_ANTHROPIC_CLAUDE_3_SONNET = "anthropic/claude-3-sonnet",
|
||||||
MODEL_COHERE_COMMAND_R_03_2024 = "cohere/command-r-03-2024",
|
MODEL_COHERE_COMMAND_R_03_2024 = "cohere/command-r-03-2024",
|
||||||
MODEL_MISTRALAI_MISTRAL_LARGE = "mistralai/mistral-large",
|
MODEL_MISTRALAI_MISTRAL_LARGE = "mistralai/mistral-large",
|
||||||
MODEL_GOOGLE_GEMMA_7B_IT = "google/gemma-7b-it",
|
|
||||||
MODEL_OPENAI_GPT_3_5_TURBO_0613 = "openai/gpt-3.5-turbo-0613",
|
MODEL_OPENAI_GPT_3_5_TURBO_0613 = "openai/gpt-3.5-turbo-0613",
|
||||||
MODEL_OPENAI_GPT_4_TURBO_PREVIEW = "openai/gpt-4-turbo-preview",
|
MODEL_OPENAI_GPT_4_TURBO_PREVIEW = "openai/gpt-4-turbo-preview",
|
||||||
MODEL_NOUSRESEARCH_NOUS_HERMES_2_MIXTRAL_8X7B_DPO = "nousresearch/nous-hermes-2-mixtral-8x7b-dpo",
|
MODEL_NOUSRESEARCH_NOUS_HERMES_2_MIXTRAL_8X7B_DPO = "nousresearch/nous-hermes-2-mixtral-8x7b-dpo",
|
||||||
@ -256,7 +254,6 @@ export declare enum E_OPENROUTER_MODEL {
|
|||||||
MODEL_ANTHROPIC_CLAUDE_2_1 = "anthropic/claude-2.1",
|
MODEL_ANTHROPIC_CLAUDE_2_1 = "anthropic/claude-2.1",
|
||||||
MODEL_ANTHROPIC_CLAUDE_2_BETA = "anthropic/claude-2:beta",
|
MODEL_ANTHROPIC_CLAUDE_2_BETA = "anthropic/claude-2:beta",
|
||||||
MODEL_ANTHROPIC_CLAUDE_2 = "anthropic/claude-2",
|
MODEL_ANTHROPIC_CLAUDE_2 = "anthropic/claude-2",
|
||||||
MODEL_TEKNIUM_OPENHERMES_2_5_MISTRAL_7B = "teknium/openhermes-2.5-mistral-7b",
|
|
||||||
MODEL_UNDI95_TOPPY_M_7B_FREE = "undi95/toppy-m-7b:free",
|
MODEL_UNDI95_TOPPY_M_7B_FREE = "undi95/toppy-m-7b:free",
|
||||||
MODEL_UNDI95_TOPPY_M_7B = "undi95/toppy-m-7b",
|
MODEL_UNDI95_TOPPY_M_7B = "undi95/toppy-m-7b",
|
||||||
MODEL_ALPINDALE_GOLIATH_120B = "alpindale/goliath-120b",
|
MODEL_ALPINDALE_GOLIATH_120B = "alpindale/goliath-120b",
|
||||||
@ -281,7 +278,6 @@ export declare enum E_OPENROUTER_MODEL {
|
|||||||
MODEL_UNDI95_REMM_SLERP_L2_13B = "undi95/remm-slerp-l2-13b",
|
MODEL_UNDI95_REMM_SLERP_L2_13B = "undi95/remm-slerp-l2-13b",
|
||||||
MODEL_GOOGLE_PALM_2_CHAT_BISON = "google/palm-2-chat-bison",
|
MODEL_GOOGLE_PALM_2_CHAT_BISON = "google/palm-2-chat-bison",
|
||||||
MODEL_GOOGLE_PALM_2_CODECHAT_BISON = "google/palm-2-codechat-bison",
|
MODEL_GOOGLE_PALM_2_CODECHAT_BISON = "google/palm-2-codechat-bison",
|
||||||
MODEL_GRYPHE_MYTHOMAX_L2_13B_FREE = "gryphe/mythomax-l2-13b:free",
|
|
||||||
MODEL_GRYPHE_MYTHOMAX_L2_13B = "gryphe/mythomax-l2-13b",
|
MODEL_GRYPHE_MYTHOMAX_L2_13B = "gryphe/mythomax-l2-13b",
|
||||||
MODEL_META_LLAMA_LLAMA_2_13B_CHAT = "meta-llama/llama-2-13b-chat",
|
MODEL_META_LLAMA_LLAMA_2_13B_CHAT = "meta-llama/llama-2-13b-chat",
|
||||||
MODEL_META_LLAMA_LLAMA_2_70B_CHAT = "meta-llama/llama-2-70b-chat",
|
MODEL_META_LLAMA_LLAMA_2_70B_CHAT = "meta-llama/llama-2-70b-chat",
|
||||||
|
|||||||
File diff suppressed because one or more lines are too long
File diff suppressed because one or more lines are too long
File diff suppressed because one or more lines are too long
8
packages/kbot/dist-in/zod_types.d.ts
vendored
8
packages/kbot/dist-in/zod_types.d.ts
vendored
@ -121,7 +121,6 @@ export interface IKBotOptions {
|
|||||||
google/gemma-3-27b-it:free | free
|
google/gemma-3-27b-it:free | free
|
||||||
google/gemma-3-4b-it | paid
|
google/gemma-3-4b-it | paid
|
||||||
google/gemma-3-4b-it:free | free
|
google/gemma-3-4b-it:free | free
|
||||||
google/gemma-7b-it | paid
|
|
||||||
google/learnlm-1.5-pro-experimental:free | free
|
google/learnlm-1.5-pro-experimental:free | free
|
||||||
google/palm-2-chat-bison | paid
|
google/palm-2-chat-bison | paid
|
||||||
google/palm-2-chat-bison-32k | paid
|
google/palm-2-chat-bison-32k | paid
|
||||||
@ -144,7 +143,6 @@ export interface IKBotOptions {
|
|||||||
meta-llama/llama-2-70b-chat | paid
|
meta-llama/llama-2-70b-chat | paid
|
||||||
meta-llama/llama-3-70b-instruct | paid
|
meta-llama/llama-3-70b-instruct | paid
|
||||||
meta-llama/llama-3-8b-instruct | paid
|
meta-llama/llama-3-8b-instruct | paid
|
||||||
meta-llama/llama-3-8b-instruct:free | free
|
|
||||||
meta-llama/llama-3.1-405b | paid
|
meta-llama/llama-3.1-405b | paid
|
||||||
meta-llama/llama-3.1-405b-instruct | paid
|
meta-llama/llama-3.1-405b-instruct | paid
|
||||||
meta-llama/llama-3.1-70b-instruct | paid
|
meta-llama/llama-3.1-70b-instruct | paid
|
||||||
@ -200,7 +198,6 @@ export interface IKBotOptions {
|
|||||||
mistralai/mistral-saba | paid
|
mistralai/mistral-saba | paid
|
||||||
moonshotai/moonlight-16b-a3b-instruct:free | free
|
moonshotai/moonlight-16b-a3b-instruct:free | free
|
||||||
gryphe/mythomax-l2-13b | paid
|
gryphe/mythomax-l2-13b | paid
|
||||||
gryphe/mythomax-l2-13b:free | free
|
|
||||||
neversleep/llama-3-lumimaid-70b | paid
|
neversleep/llama-3-lumimaid-70b | paid
|
||||||
neversleep/llama-3-lumimaid-8b | paid
|
neversleep/llama-3-lumimaid-8b | paid
|
||||||
neversleep/llama-3-lumimaid-8b:extended | paid
|
neversleep/llama-3-lumimaid-8b:extended | paid
|
||||||
@ -251,7 +248,7 @@ export interface IKBotOptions {
|
|||||||
openai/o3-mini-high | paid
|
openai/o3-mini-high | paid
|
||||||
openchat/openchat-7b | paid
|
openchat/openchat-7b | paid
|
||||||
openchat/openchat-7b:free | free
|
openchat/openchat-7b:free | free
|
||||||
teknium/openhermes-2.5-mistral-7b | paid
|
all-hands/openhands-lm-32b-v0.1 | paid
|
||||||
perplexity/llama-3.1-sonar-large-128k-online | paid
|
perplexity/llama-3.1-sonar-large-128k-online | paid
|
||||||
perplexity/llama-3.1-sonar-small-128k-online | paid
|
perplexity/llama-3.1-sonar-small-128k-online | paid
|
||||||
perplexity/r1-1776 | paid
|
perplexity/r1-1776 | paid
|
||||||
@ -261,9 +258,8 @@ export interface IKBotOptions {
|
|||||||
perplexity/sonar-reasoning | paid
|
perplexity/sonar-reasoning | paid
|
||||||
perplexity/sonar-reasoning-pro | paid
|
perplexity/sonar-reasoning-pro | paid
|
||||||
pygmalionai/mythalion-13b | paid
|
pygmalionai/mythalion-13b | paid
|
||||||
|
openrouter/quasar-alpha | paid
|
||||||
qwen/qwen-2-72b-instruct | paid
|
qwen/qwen-2-72b-instruct | paid
|
||||||
qwen/qwen-2-7b-instruct | paid
|
|
||||||
qwen/qwen-2-7b-instruct:free | free
|
|
||||||
qwen/qwen-vl-max | paid
|
qwen/qwen-vl-max | paid
|
||||||
qwen/qwen-vl-plus | paid
|
qwen/qwen-vl-plus | paid
|
||||||
qwen/qwen-max | paid
|
qwen/qwen-max | paid
|
||||||
|
|||||||
File diff suppressed because one or more lines are too long
File diff suppressed because one or more lines are too long
@ -52,7 +52,7 @@ ${toolModels.map(model => ` MODEL_TOOLS_${model.id.replace(/[^a-zA-Z0-9]/g, '_'
|
|||||||
}
|
}
|
||||||
|
|
||||||
export const build = async () => {
|
export const build = async () => {
|
||||||
const examplesSrc = path.resolve(__dirname, '../docs_/examples.md')
|
const examplesSrc = path.resolve(__dirname, '../../docs_/examples.md')
|
||||||
if(exists(examplesSrc)) {
|
if(exists(examplesSrc)) {
|
||||||
const examples = read(examplesSrc,'string') || ''
|
const examples = read(examplesSrc,'string') || ''
|
||||||
const examplesPath = path.resolve(__dirname, '../src/docs-internal/examples.ts')
|
const examplesPath = path.resolve(__dirname, '../src/docs-internal/examples.ts')
|
||||||
|
|||||||
18
packages/kbot/src/models/cache/openai-models.ts
vendored
18
packages/kbot/src/models/cache/openai-models.ts
vendored
@ -4,10 +4,12 @@ export enum E_OPENAI_MODEL {
|
|||||||
MODEL_DALL_E_3 = "dall-e-3",
|
MODEL_DALL_E_3 = "dall-e-3",
|
||||||
MODEL_DALL_E_2 = "dall-e-2",
|
MODEL_DALL_E_2 = "dall-e-2",
|
||||||
MODEL_GPT_4O_AUDIO_PREVIEW_2024_10_01 = "gpt-4o-audio-preview-2024-10-01",
|
MODEL_GPT_4O_AUDIO_PREVIEW_2024_10_01 = "gpt-4o-audio-preview-2024-10-01",
|
||||||
MODEL_GPT_4O_MINI_REALTIME_PREVIEW_2024_12_17 = "gpt-4o-mini-realtime-preview-2024-12-17",
|
MODEL_O1_PRO_2025_03_19 = "o1-pro-2025-03-19",
|
||||||
MODEL_GPT_4O_MINI_REALTIME_PREVIEW = "gpt-4o-mini-realtime-preview",
|
MODEL_O1_PRO = "o1-pro",
|
||||||
MODEL_GPT_4O_REALTIME_PREVIEW_2024_10_01 = "gpt-4o-realtime-preview-2024-10-01",
|
MODEL_GPT_4O_REALTIME_PREVIEW_2024_10_01 = "gpt-4o-realtime-preview-2024-10-01",
|
||||||
|
MODEL_O1_2024_12_17 = "o1-2024-12-17",
|
||||||
MODEL_GPT_4O_TRANSCRIBE = "gpt-4o-transcribe",
|
MODEL_GPT_4O_TRANSCRIBE = "gpt-4o-transcribe",
|
||||||
|
MODEL_O1 = "o1",
|
||||||
MODEL_GPT_4O_MINI_TRANSCRIBE = "gpt-4o-mini-transcribe",
|
MODEL_GPT_4O_MINI_TRANSCRIBE = "gpt-4o-mini-transcribe",
|
||||||
MODEL_GPT_4O_REALTIME_PREVIEW = "gpt-4o-realtime-preview",
|
MODEL_GPT_4O_REALTIME_PREVIEW = "gpt-4o-realtime-preview",
|
||||||
MODEL_BABBAGE_002 = "babbage-002",
|
MODEL_BABBAGE_002 = "babbage-002",
|
||||||
@ -16,19 +18,18 @@ export enum E_OPENAI_MODEL {
|
|||||||
MODEL_TEXT_EMBEDDING_3_LARGE = "text-embedding-3-large",
|
MODEL_TEXT_EMBEDDING_3_LARGE = "text-embedding-3-large",
|
||||||
MODEL_GPT_4 = "gpt-4",
|
MODEL_GPT_4 = "gpt-4",
|
||||||
MODEL_TEXT_EMBEDDING_ADA_002 = "text-embedding-ada-002",
|
MODEL_TEXT_EMBEDDING_ADA_002 = "text-embedding-ada-002",
|
||||||
|
MODEL_OMNI_MODERATION_LATEST = "omni-moderation-latest",
|
||||||
MODEL_TTS_1_HD = "tts-1-hd",
|
MODEL_TTS_1_HD = "tts-1-hd",
|
||||||
MODEL_GPT_4O_MINI_AUDIO_PREVIEW = "gpt-4o-mini-audio-preview",
|
MODEL_GPT_4O_MINI_AUDIO_PREVIEW = "gpt-4o-mini-audio-preview",
|
||||||
MODEL_GPT_4O_AUDIO_PREVIEW = "gpt-4o-audio-preview",
|
MODEL_GPT_4O_AUDIO_PREVIEW = "gpt-4o-audio-preview",
|
||||||
MODEL_O1_PREVIEW_2024_09_12 = "o1-preview-2024-09-12",
|
MODEL_O1_PREVIEW_2024_09_12 = "o1-preview-2024-09-12",
|
||||||
|
MODEL_GPT_4O_MINI_REALTIME_PREVIEW = "gpt-4o-mini-realtime-preview",
|
||||||
|
MODEL_GPT_4O_MINI_REALTIME_PREVIEW_2024_12_17 = "gpt-4o-mini-realtime-preview-2024-12-17",
|
||||||
MODEL_GPT_3_5_TURBO_INSTRUCT_0914 = "gpt-3.5-turbo-instruct-0914",
|
MODEL_GPT_3_5_TURBO_INSTRUCT_0914 = "gpt-3.5-turbo-instruct-0914",
|
||||||
MODEL_GPT_4O_MINI_SEARCH_PREVIEW = "gpt-4o-mini-search-preview",
|
MODEL_GPT_4O_MINI_SEARCH_PREVIEW = "gpt-4o-mini-search-preview",
|
||||||
MODEL_O1 = "o1",
|
|
||||||
MODEL_TTS_1_1106 = "tts-1-1106",
|
MODEL_TTS_1_1106 = "tts-1-1106",
|
||||||
MODEL_O1_2024_12_17 = "o1-2024-12-17",
|
|
||||||
MODEL_DAVINCI_002 = "davinci-002",
|
MODEL_DAVINCI_002 = "davinci-002",
|
||||||
MODEL_GPT_3_5_TURBO_1106 = "gpt-3.5-turbo-1106",
|
MODEL_GPT_3_5_TURBO_1106 = "gpt-3.5-turbo-1106",
|
||||||
MODEL_O1_PRO = "o1-pro",
|
|
||||||
MODEL_O1_PRO_2025_03_19 = "o1-pro-2025-03-19",
|
|
||||||
MODEL_GPT_4_TURBO = "gpt-4-turbo",
|
MODEL_GPT_4_TURBO = "gpt-4-turbo",
|
||||||
MODEL_GPT_3_5_TURBO_INSTRUCT = "gpt-3.5-turbo-instruct",
|
MODEL_GPT_3_5_TURBO_INSTRUCT = "gpt-3.5-turbo-instruct",
|
||||||
MODEL_GPT_3_5_TURBO = "gpt-3.5-turbo",
|
MODEL_GPT_3_5_TURBO = "gpt-3.5-turbo",
|
||||||
@ -49,7 +50,6 @@ export enum E_OPENAI_MODEL {
|
|||||||
MODEL_GPT_4_5_PREVIEW = "gpt-4.5-preview",
|
MODEL_GPT_4_5_PREVIEW = "gpt-4.5-preview",
|
||||||
MODEL_GPT_4_5_PREVIEW_2025_02_27 = "gpt-4.5-preview-2025-02-27",
|
MODEL_GPT_4_5_PREVIEW_2025_02_27 = "gpt-4.5-preview-2025-02-27",
|
||||||
MODEL_GPT_4O_SEARCH_PREVIEW_2025_03_11 = "gpt-4o-search-preview-2025-03-11",
|
MODEL_GPT_4O_SEARCH_PREVIEW_2025_03_11 = "gpt-4o-search-preview-2025-03-11",
|
||||||
MODEL_OMNI_MODERATION_LATEST = "omni-moderation-latest",
|
|
||||||
MODEL_TTS_1 = "tts-1",
|
MODEL_TTS_1 = "tts-1",
|
||||||
MODEL_OMNI_MODERATION_2024_09_26 = "omni-moderation-2024-09-26",
|
MODEL_OMNI_MODERATION_2024_09_26 = "omni-moderation-2024-09-26",
|
||||||
MODEL_TEXT_EMBEDDING_3_SMALL = "text-embedding-3-small",
|
MODEL_TEXT_EMBEDDING_3_SMALL = "text-embedding-3-small",
|
||||||
@ -57,9 +57,9 @@ export enum E_OPENAI_MODEL {
|
|||||||
MODEL_GPT_4O_MINI = "gpt-4o-mini",
|
MODEL_GPT_4O_MINI = "gpt-4o-mini",
|
||||||
MODEL_GPT_4O_2024_08_06 = "gpt-4o-2024-08-06",
|
MODEL_GPT_4O_2024_08_06 = "gpt-4o-2024-08-06",
|
||||||
MODEL_GPT_4O_MINI_2024_07_18 = "gpt-4o-mini-2024-07-18",
|
MODEL_GPT_4O_MINI_2024_07_18 = "gpt-4o-mini-2024-07-18",
|
||||||
MODEL_GPT_4_TURBO_PREVIEW = "gpt-4-turbo-preview",
|
|
||||||
MODEL_O1_MINI = "o1-mini",
|
MODEL_O1_MINI = "o1-mini",
|
||||||
MODEL_GPT_4O_MINI_AUDIO_PREVIEW_2024_12_17 = "gpt-4o-mini-audio-preview-2024-12-17",
|
MODEL_GPT_4O_MINI_AUDIO_PREVIEW_2024_12_17 = "gpt-4o-mini-audio-preview-2024-12-17",
|
||||||
MODEL_O1_MINI_2024_09_12 = "o1-mini-2024-09-12",
|
MODEL_O1_MINI_2024_09_12 = "o1-mini-2024-09-12",
|
||||||
MODEL_GPT_4_0125_PREVIEW = "gpt-4-0125-preview"
|
MODEL_GPT_4_0125_PREVIEW = "gpt-4-0125-preview",
|
||||||
|
MODEL_GPT_4_TURBO_PREVIEW = "gpt-4-turbo-preview"
|
||||||
}
|
}
|
||||||
@ -1,4 +1,5 @@
|
|||||||
export enum E_OPENROUTER_MODEL_FREE {
|
export enum E_OPENROUTER_MODEL_FREE {
|
||||||
|
MODEL_FREE_OPENROUTER_QUASAR_ALPHA = "openrouter/quasar-alpha",
|
||||||
MODEL_FREE_DEEPSEEK_DEEPSEEK_V3_BASE_FREE = "deepseek/deepseek-v3-base:free",
|
MODEL_FREE_DEEPSEEK_DEEPSEEK_V3_BASE_FREE = "deepseek/deepseek-v3-base:free",
|
||||||
MODEL_FREE_ALLENAI_MOLMO_7B_D_FREE = "allenai/molmo-7b-d:free",
|
MODEL_FREE_ALLENAI_MOLMO_7B_D_FREE = "allenai/molmo-7b-d:free",
|
||||||
MODEL_FREE_BYTEDANCE_RESEARCH_UI_TARS_72B_FREE = "bytedance-research/ui-tars-72b:free",
|
MODEL_FREE_BYTEDANCE_RESEARCH_UI_TARS_72B_FREE = "bytedance-research/ui-tars-72b:free",
|
||||||
@ -46,14 +47,11 @@ export enum E_OPENROUTER_MODEL_FREE {
|
|||||||
MODEL_FREE_GOOGLE_GEMINI_FLASH_1_5_8B_EXP = "google/gemini-flash-1.5-8b-exp",
|
MODEL_FREE_GOOGLE_GEMINI_FLASH_1_5_8B_EXP = "google/gemini-flash-1.5-8b-exp",
|
||||||
MODEL_FREE_META_LLAMA_LLAMA_3_1_8B_INSTRUCT_FREE = "meta-llama/llama-3.1-8b-instruct:free",
|
MODEL_FREE_META_LLAMA_LLAMA_3_1_8B_INSTRUCT_FREE = "meta-llama/llama-3.1-8b-instruct:free",
|
||||||
MODEL_FREE_MISTRALAI_MISTRAL_NEMO_FREE = "mistralai/mistral-nemo:free",
|
MODEL_FREE_MISTRALAI_MISTRAL_NEMO_FREE = "mistralai/mistral-nemo:free",
|
||||||
MODEL_FREE_QWEN_QWEN_2_7B_INSTRUCT_FREE = "qwen/qwen-2-7b-instruct:free",
|
|
||||||
MODEL_FREE_GOOGLE_GEMMA_2_9B_IT_FREE = "google/gemma-2-9b-it:free",
|
MODEL_FREE_GOOGLE_GEMMA_2_9B_IT_FREE = "google/gemma-2-9b-it:free",
|
||||||
MODEL_FREE_MISTRALAI_MISTRAL_7B_INSTRUCT_FREE = "mistralai/mistral-7b-instruct:free",
|
MODEL_FREE_MISTRALAI_MISTRAL_7B_INSTRUCT_FREE = "mistralai/mistral-7b-instruct:free",
|
||||||
MODEL_FREE_MICROSOFT_PHI_3_MINI_128K_INSTRUCT_FREE = "microsoft/phi-3-mini-128k-instruct:free",
|
MODEL_FREE_MICROSOFT_PHI_3_MINI_128K_INSTRUCT_FREE = "microsoft/phi-3-mini-128k-instruct:free",
|
||||||
MODEL_FREE_MICROSOFT_PHI_3_MEDIUM_128K_INSTRUCT_FREE = "microsoft/phi-3-medium-128k-instruct:free",
|
MODEL_FREE_MICROSOFT_PHI_3_MEDIUM_128K_INSTRUCT_FREE = "microsoft/phi-3-medium-128k-instruct:free",
|
||||||
MODEL_FREE_META_LLAMA_LLAMA_3_8B_INSTRUCT_FREE = "meta-llama/llama-3-8b-instruct:free",
|
|
||||||
MODEL_FREE_OPENCHAT_OPENCHAT_7B_FREE = "openchat/openchat-7b:free",
|
MODEL_FREE_OPENCHAT_OPENCHAT_7B_FREE = "openchat/openchat-7b:free",
|
||||||
MODEL_FREE_UNDI95_TOPPY_M_7B_FREE = "undi95/toppy-m-7b:free",
|
MODEL_FREE_UNDI95_TOPPY_M_7B_FREE = "undi95/toppy-m-7b:free",
|
||||||
MODEL_FREE_HUGGINGFACEH4_ZEPHYR_7B_BETA_FREE = "huggingfaceh4/zephyr-7b-beta:free",
|
MODEL_FREE_HUGGINGFACEH4_ZEPHYR_7B_BETA_FREE = "huggingfaceh4/zephyr-7b-beta:free"
|
||||||
MODEL_FREE_GRYPHE_MYTHOMAX_L2_13B_FREE = "gryphe/mythomax-l2-13b:free"
|
|
||||||
}
|
}
|
||||||
@ -1,4 +1,6 @@
|
|||||||
export enum E_OPENROUTER_MODEL {
|
export enum E_OPENROUTER_MODEL {
|
||||||
|
MODEL_OPENROUTER_QUASAR_ALPHA = "openrouter/quasar-alpha",
|
||||||
|
MODEL_ALL_HANDS_OPENHANDS_LM_32B_V0_1 = "all-hands/openhands-lm-32b-v0.1",
|
||||||
MODEL_MISTRAL_MINISTRAL_8B = "mistral/ministral-8b",
|
MODEL_MISTRAL_MINISTRAL_8B = "mistral/ministral-8b",
|
||||||
MODEL_DEEPSEEK_DEEPSEEK_V3_BASE_FREE = "deepseek/deepseek-v3-base:free",
|
MODEL_DEEPSEEK_DEEPSEEK_V3_BASE_FREE = "deepseek/deepseek-v3-base:free",
|
||||||
MODEL_SCB10X_LLAMA3_1_TYPHOON2_8B_INSTRUCT = "scb10x/llama3.1-typhoon2-8b-instruct",
|
MODEL_SCB10X_LLAMA3_1_TYPHOON2_8B_INSTRUCT = "scb10x/llama3.1-typhoon2-8b-instruct",
|
||||||
@ -47,9 +49,9 @@ export enum E_OPENROUTER_MODEL {
|
|||||||
MODEL_NOUSRESEARCH_DEEPHERMES_3_LLAMA_3_8B_PREVIEW_FREE = "nousresearch/deephermes-3-llama-3-8b-preview:free",
|
MODEL_NOUSRESEARCH_DEEPHERMES_3_LLAMA_3_8B_PREVIEW_FREE = "nousresearch/deephermes-3-llama-3-8b-preview:free",
|
||||||
MODEL_OPENAI_GPT_4_5_PREVIEW = "openai/gpt-4.5-preview",
|
MODEL_OPENAI_GPT_4_5_PREVIEW = "openai/gpt-4.5-preview",
|
||||||
MODEL_GOOGLE_GEMINI_2_0_FLASH_LITE_001 = "google/gemini-2.0-flash-lite-001",
|
MODEL_GOOGLE_GEMINI_2_0_FLASH_LITE_001 = "google/gemini-2.0-flash-lite-001",
|
||||||
MODEL_ANTHROPIC_CLAUDE_3_7_SONNET_BETA = "anthropic/claude-3.7-sonnet:beta",
|
|
||||||
MODEL_ANTHROPIC_CLAUDE_3_7_SONNET = "anthropic/claude-3.7-sonnet",
|
MODEL_ANTHROPIC_CLAUDE_3_7_SONNET = "anthropic/claude-3.7-sonnet",
|
||||||
MODEL_ANTHROPIC_CLAUDE_3_7_SONNET_THINKING = "anthropic/claude-3.7-sonnet:thinking",
|
MODEL_ANTHROPIC_CLAUDE_3_7_SONNET_THINKING = "anthropic/claude-3.7-sonnet:thinking",
|
||||||
|
MODEL_ANTHROPIC_CLAUDE_3_7_SONNET_BETA = "anthropic/claude-3.7-sonnet:beta",
|
||||||
MODEL_PERPLEXITY_R1_1776 = "perplexity/r1-1776",
|
MODEL_PERPLEXITY_R1_1776 = "perplexity/r1-1776",
|
||||||
MODEL_MISTRALAI_MISTRAL_SABA = "mistralai/mistral-saba",
|
MODEL_MISTRALAI_MISTRAL_SABA = "mistralai/mistral-saba",
|
||||||
MODEL_COGNITIVECOMPUTATIONS_DOLPHIN3_0_R1_MISTRAL_24B_FREE = "cognitivecomputations/dolphin3.0-r1-mistral-24b:free",
|
MODEL_COGNITIVECOMPUTATIONS_DOLPHIN3_0_R1_MISTRAL_24B_FREE = "cognitivecomputations/dolphin3.0-r1-mistral-24b:free",
|
||||||
@ -185,8 +187,6 @@ export enum E_OPENROUTER_MODEL {
|
|||||||
MODEL_MISTRALAI_MISTRAL_NEMO = "mistralai/mistral-nemo",
|
MODEL_MISTRALAI_MISTRAL_NEMO = "mistralai/mistral-nemo",
|
||||||
MODEL_OPENAI_GPT_4O_MINI = "openai/gpt-4o-mini",
|
MODEL_OPENAI_GPT_4O_MINI = "openai/gpt-4o-mini",
|
||||||
MODEL_OPENAI_GPT_4O_MINI_2024_07_18 = "openai/gpt-4o-mini-2024-07-18",
|
MODEL_OPENAI_GPT_4O_MINI_2024_07_18 = "openai/gpt-4o-mini-2024-07-18",
|
||||||
MODEL_QWEN_QWEN_2_7B_INSTRUCT_FREE = "qwen/qwen-2-7b-instruct:free",
|
|
||||||
MODEL_QWEN_QWEN_2_7B_INSTRUCT = "qwen/qwen-2-7b-instruct",
|
|
||||||
MODEL_GOOGLE_GEMMA_2_27B_IT = "google/gemma-2-27b-it",
|
MODEL_GOOGLE_GEMMA_2_27B_IT = "google/gemma-2-27b-it",
|
||||||
MODEL_ALPINDALE_MAGNUM_72B = "alpindale/magnum-72b",
|
MODEL_ALPINDALE_MAGNUM_72B = "alpindale/magnum-72b",
|
||||||
MODEL_GOOGLE_GEMMA_2_9B_IT_FREE = "google/gemma-2-9b-it:free",
|
MODEL_GOOGLE_GEMMA_2_9B_IT_FREE = "google/gemma-2-9b-it:free",
|
||||||
@ -215,7 +215,6 @@ export enum E_OPENROUTER_MODEL {
|
|||||||
MODEL_NEVERSLEEP_LLAMA_3_LUMIMAID_8B_EXTENDED = "neversleep/llama-3-lumimaid-8b:extended",
|
MODEL_NEVERSLEEP_LLAMA_3_LUMIMAID_8B_EXTENDED = "neversleep/llama-3-lumimaid-8b:extended",
|
||||||
MODEL_NEVERSLEEP_LLAMA_3_LUMIMAID_8B = "neversleep/llama-3-lumimaid-8b",
|
MODEL_NEVERSLEEP_LLAMA_3_LUMIMAID_8B = "neversleep/llama-3-lumimaid-8b",
|
||||||
MODEL_SAO10K_FIMBULVETR_11B_V2 = "sao10k/fimbulvetr-11b-v2",
|
MODEL_SAO10K_FIMBULVETR_11B_V2 = "sao10k/fimbulvetr-11b-v2",
|
||||||
MODEL_META_LLAMA_LLAMA_3_8B_INSTRUCT_FREE = "meta-llama/llama-3-8b-instruct:free",
|
|
||||||
MODEL_META_LLAMA_LLAMA_3_8B_INSTRUCT = "meta-llama/llama-3-8b-instruct",
|
MODEL_META_LLAMA_LLAMA_3_8B_INSTRUCT = "meta-llama/llama-3-8b-instruct",
|
||||||
MODEL_META_LLAMA_LLAMA_3_70B_INSTRUCT = "meta-llama/llama-3-70b-instruct",
|
MODEL_META_LLAMA_LLAMA_3_70B_INSTRUCT = "meta-llama/llama-3-70b-instruct",
|
||||||
MODEL_MISTRALAI_MIXTRAL_8X22B_INSTRUCT = "mistralai/mixtral-8x22b-instruct",
|
MODEL_MISTRALAI_MIXTRAL_8X22B_INSTRUCT = "mistralai/mixtral-8x22b-instruct",
|
||||||
@ -236,7 +235,6 @@ export enum E_OPENROUTER_MODEL {
|
|||||||
MODEL_ANTHROPIC_CLAUDE_3_SONNET = "anthropic/claude-3-sonnet",
|
MODEL_ANTHROPIC_CLAUDE_3_SONNET = "anthropic/claude-3-sonnet",
|
||||||
MODEL_COHERE_COMMAND_R_03_2024 = "cohere/command-r-03-2024",
|
MODEL_COHERE_COMMAND_R_03_2024 = "cohere/command-r-03-2024",
|
||||||
MODEL_MISTRALAI_MISTRAL_LARGE = "mistralai/mistral-large",
|
MODEL_MISTRALAI_MISTRAL_LARGE = "mistralai/mistral-large",
|
||||||
MODEL_GOOGLE_GEMMA_7B_IT = "google/gemma-7b-it",
|
|
||||||
MODEL_OPENAI_GPT_3_5_TURBO_0613 = "openai/gpt-3.5-turbo-0613",
|
MODEL_OPENAI_GPT_3_5_TURBO_0613 = "openai/gpt-3.5-turbo-0613",
|
||||||
MODEL_OPENAI_GPT_4_TURBO_PREVIEW = "openai/gpt-4-turbo-preview",
|
MODEL_OPENAI_GPT_4_TURBO_PREVIEW = "openai/gpt-4-turbo-preview",
|
||||||
MODEL_NOUSRESEARCH_NOUS_HERMES_2_MIXTRAL_8X7B_DPO = "nousresearch/nous-hermes-2-mixtral-8x7b-dpo",
|
MODEL_NOUSRESEARCH_NOUS_HERMES_2_MIXTRAL_8X7B_DPO = "nousresearch/nous-hermes-2-mixtral-8x7b-dpo",
|
||||||
@ -256,7 +254,6 @@ export enum E_OPENROUTER_MODEL {
|
|||||||
MODEL_ANTHROPIC_CLAUDE_2_1 = "anthropic/claude-2.1",
|
MODEL_ANTHROPIC_CLAUDE_2_1 = "anthropic/claude-2.1",
|
||||||
MODEL_ANTHROPIC_CLAUDE_2_BETA = "anthropic/claude-2:beta",
|
MODEL_ANTHROPIC_CLAUDE_2_BETA = "anthropic/claude-2:beta",
|
||||||
MODEL_ANTHROPIC_CLAUDE_2 = "anthropic/claude-2",
|
MODEL_ANTHROPIC_CLAUDE_2 = "anthropic/claude-2",
|
||||||
MODEL_TEKNIUM_OPENHERMES_2_5_MISTRAL_7B = "teknium/openhermes-2.5-mistral-7b",
|
|
||||||
MODEL_UNDI95_TOPPY_M_7B_FREE = "undi95/toppy-m-7b:free",
|
MODEL_UNDI95_TOPPY_M_7B_FREE = "undi95/toppy-m-7b:free",
|
||||||
MODEL_UNDI95_TOPPY_M_7B = "undi95/toppy-m-7b",
|
MODEL_UNDI95_TOPPY_M_7B = "undi95/toppy-m-7b",
|
||||||
MODEL_ALPINDALE_GOLIATH_120B = "alpindale/goliath-120b",
|
MODEL_ALPINDALE_GOLIATH_120B = "alpindale/goliath-120b",
|
||||||
@ -281,7 +278,6 @@ export enum E_OPENROUTER_MODEL {
|
|||||||
MODEL_UNDI95_REMM_SLERP_L2_13B = "undi95/remm-slerp-l2-13b",
|
MODEL_UNDI95_REMM_SLERP_L2_13B = "undi95/remm-slerp-l2-13b",
|
||||||
MODEL_GOOGLE_PALM_2_CHAT_BISON = "google/palm-2-chat-bison",
|
MODEL_GOOGLE_PALM_2_CHAT_BISON = "google/palm-2-chat-bison",
|
||||||
MODEL_GOOGLE_PALM_2_CODECHAT_BISON = "google/palm-2-codechat-bison",
|
MODEL_GOOGLE_PALM_2_CODECHAT_BISON = "google/palm-2-codechat-bison",
|
||||||
MODEL_GRYPHE_MYTHOMAX_L2_13B_FREE = "gryphe/mythomax-l2-13b:free",
|
|
||||||
MODEL_GRYPHE_MYTHOMAX_L2_13B = "gryphe/mythomax-l2-13b",
|
MODEL_GRYPHE_MYTHOMAX_L2_13B = "gryphe/mythomax-l2-13b",
|
||||||
MODEL_META_LLAMA_LLAMA_2_13B_CHAT = "meta-llama/llama-2-13b-chat",
|
MODEL_META_LLAMA_LLAMA_2_13B_CHAT = "meta-llama/llama-2-13b-chat",
|
||||||
MODEL_META_LLAMA_LLAMA_2_70B_CHAT = "meta-llama/llama-2-70b-chat",
|
MODEL_META_LLAMA_LLAMA_2_70B_CHAT = "meta-llama/llama-2-70b-chat",
|
||||||
|
|||||||
@ -121,7 +121,6 @@ export interface IKBotOptions {
|
|||||||
google/gemma-3-27b-it:free | free
|
google/gemma-3-27b-it:free | free
|
||||||
google/gemma-3-4b-it | paid
|
google/gemma-3-4b-it | paid
|
||||||
google/gemma-3-4b-it:free | free
|
google/gemma-3-4b-it:free | free
|
||||||
google/gemma-7b-it | paid
|
|
||||||
google/learnlm-1.5-pro-experimental:free | free
|
google/learnlm-1.5-pro-experimental:free | free
|
||||||
google/palm-2-chat-bison | paid
|
google/palm-2-chat-bison | paid
|
||||||
google/palm-2-chat-bison-32k | paid
|
google/palm-2-chat-bison-32k | paid
|
||||||
@ -144,7 +143,6 @@ export interface IKBotOptions {
|
|||||||
meta-llama/llama-2-70b-chat | paid
|
meta-llama/llama-2-70b-chat | paid
|
||||||
meta-llama/llama-3-70b-instruct | paid
|
meta-llama/llama-3-70b-instruct | paid
|
||||||
meta-llama/llama-3-8b-instruct | paid
|
meta-llama/llama-3-8b-instruct | paid
|
||||||
meta-llama/llama-3-8b-instruct:free | free
|
|
||||||
meta-llama/llama-3.1-405b | paid
|
meta-llama/llama-3.1-405b | paid
|
||||||
meta-llama/llama-3.1-405b-instruct | paid
|
meta-llama/llama-3.1-405b-instruct | paid
|
||||||
meta-llama/llama-3.1-70b-instruct | paid
|
meta-llama/llama-3.1-70b-instruct | paid
|
||||||
@ -200,7 +198,6 @@ export interface IKBotOptions {
|
|||||||
mistralai/mistral-saba | paid
|
mistralai/mistral-saba | paid
|
||||||
moonshotai/moonlight-16b-a3b-instruct:free | free
|
moonshotai/moonlight-16b-a3b-instruct:free | free
|
||||||
gryphe/mythomax-l2-13b | paid
|
gryphe/mythomax-l2-13b | paid
|
||||||
gryphe/mythomax-l2-13b:free | free
|
|
||||||
neversleep/llama-3-lumimaid-70b | paid
|
neversleep/llama-3-lumimaid-70b | paid
|
||||||
neversleep/llama-3-lumimaid-8b | paid
|
neversleep/llama-3-lumimaid-8b | paid
|
||||||
neversleep/llama-3-lumimaid-8b:extended | paid
|
neversleep/llama-3-lumimaid-8b:extended | paid
|
||||||
@ -251,7 +248,7 @@ export interface IKBotOptions {
|
|||||||
openai/o3-mini-high | paid
|
openai/o3-mini-high | paid
|
||||||
openchat/openchat-7b | paid
|
openchat/openchat-7b | paid
|
||||||
openchat/openchat-7b:free | free
|
openchat/openchat-7b:free | free
|
||||||
teknium/openhermes-2.5-mistral-7b | paid
|
all-hands/openhands-lm-32b-v0.1 | paid
|
||||||
perplexity/llama-3.1-sonar-large-128k-online | paid
|
perplexity/llama-3.1-sonar-large-128k-online | paid
|
||||||
perplexity/llama-3.1-sonar-small-128k-online | paid
|
perplexity/llama-3.1-sonar-small-128k-online | paid
|
||||||
perplexity/r1-1776 | paid
|
perplexity/r1-1776 | paid
|
||||||
@ -261,9 +258,8 @@ export interface IKBotOptions {
|
|||||||
perplexity/sonar-reasoning | paid
|
perplexity/sonar-reasoning | paid
|
||||||
perplexity/sonar-reasoning-pro | paid
|
perplexity/sonar-reasoning-pro | paid
|
||||||
pygmalionai/mythalion-13b | paid
|
pygmalionai/mythalion-13b | paid
|
||||||
|
openrouter/quasar-alpha | paid
|
||||||
qwen/qwen-2-72b-instruct | paid
|
qwen/qwen-2-72b-instruct | paid
|
||||||
qwen/qwen-2-7b-instruct | paid
|
|
||||||
qwen/qwen-2-7b-instruct:free | free
|
|
||||||
qwen/qwen-vl-max | paid
|
qwen/qwen-vl-max | paid
|
||||||
qwen/qwen-vl-plus | paid
|
qwen/qwen-vl-plus | paid
|
||||||
qwen/qwen-max | paid
|
qwen/qwen-max | paid
|
||||||
|
|||||||
Loading…
Reference in New Issue
Block a user