maintainence love:)
This commit is contained in:
parent
4726ebc0f0
commit
214eed04f5
File diff suppressed because one or more lines are too long
@ -1,5 +1,5 @@
|
||||
{
|
||||
"timestamp": 1743511914173,
|
||||
"timestamp": 1743769889560,
|
||||
"models": [
|
||||
{
|
||||
"id": "gpt-4o-realtime-preview-2024-12-17",
|
||||
@ -32,15 +32,15 @@
|
||||
"owned_by": "system"
|
||||
},
|
||||
{
|
||||
"id": "gpt-4o-mini-realtime-preview-2024-12-17",
|
||||
"id": "o1-pro-2025-03-19",
|
||||
"object": "model",
|
||||
"created": 1734112601,
|
||||
"created": 1742251504,
|
||||
"owned_by": "system"
|
||||
},
|
||||
{
|
||||
"id": "gpt-4o-mini-realtime-preview",
|
||||
"id": "o1-pro",
|
||||
"object": "model",
|
||||
"created": 1734387380,
|
||||
"created": 1742251791,
|
||||
"owned_by": "system"
|
||||
},
|
||||
{
|
||||
@ -49,12 +49,24 @@
|
||||
"created": 1727131766,
|
||||
"owned_by": "system"
|
||||
},
|
||||
{
|
||||
"id": "o1-2024-12-17",
|
||||
"object": "model",
|
||||
"created": 1734326976,
|
||||
"owned_by": "system"
|
||||
},
|
||||
{
|
||||
"id": "gpt-4o-transcribe",
|
||||
"object": "model",
|
||||
"created": 1742068463,
|
||||
"owned_by": "system"
|
||||
},
|
||||
{
|
||||
"id": "o1",
|
||||
"object": "model",
|
||||
"created": 1734375816,
|
||||
"owned_by": "system"
|
||||
},
|
||||
{
|
||||
"id": "gpt-4o-mini-transcribe",
|
||||
"object": "model",
|
||||
@ -103,6 +115,12 @@
|
||||
"created": 1671217299,
|
||||
"owned_by": "openai-internal"
|
||||
},
|
||||
{
|
||||
"id": "omni-moderation-latest",
|
||||
"object": "model",
|
||||
"created": 1731689265,
|
||||
"owned_by": "system"
|
||||
},
|
||||
{
|
||||
"id": "tts-1-hd",
|
||||
"object": "model",
|
||||
@ -127,6 +145,18 @@
|
||||
"created": 1725648865,
|
||||
"owned_by": "system"
|
||||
},
|
||||
{
|
||||
"id": "gpt-4o-mini-realtime-preview",
|
||||
"object": "model",
|
||||
"created": 1734387380,
|
||||
"owned_by": "system"
|
||||
},
|
||||
{
|
||||
"id": "gpt-4o-mini-realtime-preview-2024-12-17",
|
||||
"object": "model",
|
||||
"created": 1734112601,
|
||||
"owned_by": "system"
|
||||
},
|
||||
{
|
||||
"id": "gpt-3.5-turbo-instruct-0914",
|
||||
"object": "model",
|
||||
@ -139,24 +169,12 @@
|
||||
"created": 1741391161,
|
||||
"owned_by": "system"
|
||||
},
|
||||
{
|
||||
"id": "o1",
|
||||
"object": "model",
|
||||
"created": 1734375816,
|
||||
"owned_by": "system"
|
||||
},
|
||||
{
|
||||
"id": "tts-1-1106",
|
||||
"object": "model",
|
||||
"created": 1699053241,
|
||||
"owned_by": "system"
|
||||
},
|
||||
{
|
||||
"id": "o1-2024-12-17",
|
||||
"object": "model",
|
||||
"created": 1734326976,
|
||||
"owned_by": "system"
|
||||
},
|
||||
{
|
||||
"id": "davinci-002",
|
||||
"object": "model",
|
||||
@ -169,18 +187,6 @@
|
||||
"created": 1698959748,
|
||||
"owned_by": "system"
|
||||
},
|
||||
{
|
||||
"id": "o1-pro",
|
||||
"object": "model",
|
||||
"created": 1742251791,
|
||||
"owned_by": "system"
|
||||
},
|
||||
{
|
||||
"id": "o1-pro-2025-03-19",
|
||||
"object": "model",
|
||||
"created": 1742251504,
|
||||
"owned_by": "system"
|
||||
},
|
||||
{
|
||||
"id": "gpt-4-turbo",
|
||||
"object": "model",
|
||||
@ -301,12 +307,6 @@
|
||||
"created": 1741388170,
|
||||
"owned_by": "system"
|
||||
},
|
||||
{
|
||||
"id": "omni-moderation-latest",
|
||||
"object": "model",
|
||||
"created": 1731689265,
|
||||
"owned_by": "system"
|
||||
},
|
||||
{
|
||||
"id": "tts-1",
|
||||
"object": "model",
|
||||
@ -349,12 +349,6 @@
|
||||
"created": 1721172717,
|
||||
"owned_by": "system"
|
||||
},
|
||||
{
|
||||
"id": "gpt-4-turbo-preview",
|
||||
"object": "model",
|
||||
"created": 1706037777,
|
||||
"owned_by": "system"
|
||||
},
|
||||
{
|
||||
"id": "o1-mini",
|
||||
"object": "model",
|
||||
@ -378,6 +372,12 @@
|
||||
"object": "model",
|
||||
"created": 1706037612,
|
||||
"owned_by": "system"
|
||||
},
|
||||
{
|
||||
"id": "gpt-4-turbo-preview",
|
||||
"object": "model",
|
||||
"created": 1706037777,
|
||||
"owned_by": "system"
|
||||
}
|
||||
]
|
||||
}
|
||||
@ -1,6 +1,75 @@
|
||||
{
|
||||
"timestamp": 1743511914343,
|
||||
"timestamp": 1743769889892,
|
||||
"models": [
|
||||
{
|
||||
"id": "openrouter/quasar-alpha",
|
||||
"name": "Quasar Alpha",
|
||||
"created": 1743626809,
|
||||
"description": "This is a cloaked model provided to the community to gather feedback. It’s a powerful, all-purpose model supporting long-context tasks, including code generation. All prompts and completions for this model are logged by the provider as well as OpenRouter.",
|
||||
"context_length": 1000000,
|
||||
"architecture": {
|
||||
"modality": "text+image->text",
|
||||
"input_modalities": [
|
||||
"image",
|
||||
"text"
|
||||
],
|
||||
"output_modalities": [
|
||||
"text"
|
||||
],
|
||||
"tokenizer": "Other",
|
||||
"instruct_type": null
|
||||
},
|
||||
"pricing": {
|
||||
"prompt": "0",
|
||||
"completion": "0",
|
||||
"request": "0",
|
||||
"image": "0",
|
||||
"web_search": "0",
|
||||
"internal_reasoning": "0",
|
||||
"input_cache_read": "0",
|
||||
"input_cache_write": "0"
|
||||
},
|
||||
"top_provider": {
|
||||
"context_length": 1000000,
|
||||
"max_completion_tokens": 32000,
|
||||
"is_moderated": true
|
||||
},
|
||||
"per_request_limits": null
|
||||
},
|
||||
{
|
||||
"id": "all-hands/openhands-lm-32b-v0.1",
|
||||
"name": "OpenHands LM 32B V0.1",
|
||||
"created": 1743613013,
|
||||
"description": "OpenHands LM v0.1 is a 32B open-source coding model fine-tuned from Qwen2.5-Coder-32B-Instruct using reinforcement learning techniques outlined in SWE-Gym. It is optimized for autonomous software development agents and achieves strong performance on SWE-Bench Verified, with a 37.2% resolve rate. The model supports a 128K token context window, making it well-suited for long-horizon code reasoning and large codebase tasks.\n\nOpenHands LM is designed for local deployment and runs on consumer-grade GPUs such as a single 3090. It enables fully offline agent workflows without dependency on proprietary APIs. This release is intended as a research preview, and future updates aim to improve generalizability, reduce repetition, and offer smaller variants.",
|
||||
"context_length": 16384,
|
||||
"architecture": {
|
||||
"modality": "text->text",
|
||||
"input_modalities": [
|
||||
"text"
|
||||
],
|
||||
"output_modalities": [
|
||||
"text"
|
||||
],
|
||||
"tokenizer": "Other",
|
||||
"instruct_type": null
|
||||
},
|
||||
"pricing": {
|
||||
"prompt": "0.0000026",
|
||||
"completion": "0.0000034",
|
||||
"request": "0",
|
||||
"image": "0",
|
||||
"web_search": "0",
|
||||
"internal_reasoning": "0",
|
||||
"input_cache_read": "0",
|
||||
"input_cache_write": "0"
|
||||
},
|
||||
"top_provider": {
|
||||
"context_length": 16384,
|
||||
"max_completion_tokens": 4095,
|
||||
"is_moderated": false
|
||||
},
|
||||
"per_request_limits": null
|
||||
},
|
||||
{
|
||||
"id": "mistral/ministral-8b",
|
||||
"name": "Mistral: Ministral 8b",
|
||||
@ -1465,8 +1534,8 @@
|
||||
"instruct_type": "qwq"
|
||||
},
|
||||
"pricing": {
|
||||
"prompt": "0.00000012",
|
||||
"completion": "0.00000018",
|
||||
"prompt": "0.00000015",
|
||||
"completion": "0.0000002",
|
||||
"request": "0",
|
||||
"image": "0",
|
||||
"web_search": "0",
|
||||
@ -1653,41 +1722,6 @@
|
||||
},
|
||||
"per_request_limits": null
|
||||
},
|
||||
{
|
||||
"id": "anthropic/claude-3.7-sonnet:beta",
|
||||
"name": "Anthropic: Claude 3.7 Sonnet (self-moderated)",
|
||||
"created": 1740422110,
|
||||
"description": "Claude 3.7 Sonnet is an advanced large language model with improved reasoning, coding, and problem-solving capabilities. It introduces a hybrid reasoning approach, allowing users to choose between rapid responses and extended, step-by-step processing for complex tasks. The model demonstrates notable improvements in coding, particularly in front-end development and full-stack updates, and excels in agentic workflows, where it can autonomously navigate multi-step processes. \n\nClaude 3.7 Sonnet maintains performance parity with its predecessor in standard mode while offering an extended reasoning mode for enhanced accuracy in math, coding, and instruction-following tasks.\n\nRead more at the [blog post here](https://www.anthropic.com/news/claude-3-7-sonnet)",
|
||||
"context_length": 200000,
|
||||
"architecture": {
|
||||
"modality": "text+image->text",
|
||||
"input_modalities": [
|
||||
"text",
|
||||
"image"
|
||||
],
|
||||
"output_modalities": [
|
||||
"text"
|
||||
],
|
||||
"tokenizer": "Claude",
|
||||
"instruct_type": null
|
||||
},
|
||||
"pricing": {
|
||||
"prompt": "0.000003",
|
||||
"completion": "0.000015",
|
||||
"request": "0",
|
||||
"image": "0.0048",
|
||||
"web_search": "0",
|
||||
"internal_reasoning": "0",
|
||||
"input_cache_read": "0",
|
||||
"input_cache_write": "0"
|
||||
},
|
||||
"top_provider": {
|
||||
"context_length": 200000,
|
||||
"max_completion_tokens": 128000,
|
||||
"is_moderated": false
|
||||
},
|
||||
"per_request_limits": null
|
||||
},
|
||||
{
|
||||
"id": "anthropic/claude-3.7-sonnet",
|
||||
"name": "Anthropic: Claude 3.7 Sonnet",
|
||||
@ -1718,8 +1752,8 @@
|
||||
},
|
||||
"top_provider": {
|
||||
"context_length": 200000,
|
||||
"max_completion_tokens": 128000,
|
||||
"is_moderated": true
|
||||
"max_completion_tokens": 64000,
|
||||
"is_moderated": false
|
||||
},
|
||||
"per_request_limits": null
|
||||
},
|
||||
@ -1751,10 +1785,45 @@
|
||||
"input_cache_read": "0",
|
||||
"input_cache_write": "0"
|
||||
},
|
||||
"top_provider": {
|
||||
"context_length": 200000,
|
||||
"max_completion_tokens": 64000,
|
||||
"is_moderated": false
|
||||
},
|
||||
"per_request_limits": null
|
||||
},
|
||||
{
|
||||
"id": "anthropic/claude-3.7-sonnet:beta",
|
||||
"name": "Anthropic: Claude 3.7 Sonnet (self-moderated)",
|
||||
"created": 1740422110,
|
||||
"description": "Claude 3.7 Sonnet is an advanced large language model with improved reasoning, coding, and problem-solving capabilities. It introduces a hybrid reasoning approach, allowing users to choose between rapid responses and extended, step-by-step processing for complex tasks. The model demonstrates notable improvements in coding, particularly in front-end development and full-stack updates, and excels in agentic workflows, where it can autonomously navigate multi-step processes. \n\nClaude 3.7 Sonnet maintains performance parity with its predecessor in standard mode while offering an extended reasoning mode for enhanced accuracy in math, coding, and instruction-following tasks.\n\nRead more at the [blog post here](https://www.anthropic.com/news/claude-3-7-sonnet)",
|
||||
"context_length": 200000,
|
||||
"architecture": {
|
||||
"modality": "text+image->text",
|
||||
"input_modalities": [
|
||||
"text",
|
||||
"image"
|
||||
],
|
||||
"output_modalities": [
|
||||
"text"
|
||||
],
|
||||
"tokenizer": "Claude",
|
||||
"instruct_type": null
|
||||
},
|
||||
"pricing": {
|
||||
"prompt": "0.000003",
|
||||
"completion": "0.000015",
|
||||
"request": "0",
|
||||
"image": "0.0048",
|
||||
"web_search": "0",
|
||||
"internal_reasoning": "0",
|
||||
"input_cache_read": "0",
|
||||
"input_cache_write": "0"
|
||||
},
|
||||
"top_provider": {
|
||||
"context_length": 200000,
|
||||
"max_completion_tokens": 128000,
|
||||
"is_moderated": true
|
||||
"is_moderated": false
|
||||
},
|
||||
"per_request_limits": null
|
||||
},
|
||||
@ -2312,7 +2381,7 @@
|
||||
"name": "Qwen: Qwen2.5 VL 72B Instruct",
|
||||
"created": 1738410311,
|
||||
"description": "Qwen2.5-VL is proficient in recognizing common objects such as flowers, birds, fish, and insects. It is also highly capable of analyzing texts, charts, icons, graphics, and layouts within images.",
|
||||
"context_length": 128000,
|
||||
"context_length": 32000,
|
||||
"architecture": {
|
||||
"modality": "text+image->text",
|
||||
"input_modalities": [
|
||||
@ -2336,8 +2405,8 @@
|
||||
"input_cache_write": "0"
|
||||
},
|
||||
"top_provider": {
|
||||
"context_length": 128000,
|
||||
"max_completion_tokens": 128000,
|
||||
"context_length": 32000,
|
||||
"max_completion_tokens": 32000,
|
||||
"is_moderated": false
|
||||
},
|
||||
"per_request_limits": null
|
||||
@ -2823,7 +2892,7 @@
|
||||
"name": "DeepSeek: R1 Distill Llama 70B (free)",
|
||||
"created": 1737663169,
|
||||
"description": "DeepSeek R1 Distill Llama 70B is a distilled large language model based on [Llama-3.3-70B-Instruct](/meta-llama/llama-3.3-70b-instruct), using outputs from [DeepSeek R1](/deepseek/deepseek-r1). The model combines advanced distillation techniques to achieve high performance across multiple benchmarks, including:\n\n- AIME 2024 pass@1: 70.0\n- MATH-500 pass@1: 94.5\n- CodeForces Rating: 1633\n\nThe model leverages fine-tuning from DeepSeek R1's outputs, enabling competitive performance comparable to larger frontier models.",
|
||||
"context_length": 8192,
|
||||
"context_length": 128000,
|
||||
"architecture": {
|
||||
"modality": "text->text",
|
||||
"input_modalities": [
|
||||
@ -2846,8 +2915,8 @@
|
||||
"input_cache_write": "0"
|
||||
},
|
||||
"top_provider": {
|
||||
"context_length": 8192,
|
||||
"max_completion_tokens": 4096,
|
||||
"context_length": 128000,
|
||||
"max_completion_tokens": null,
|
||||
"is_moderated": false
|
||||
},
|
||||
"per_request_limits": null
|
||||
@ -5088,7 +5157,7 @@
|
||||
"name": "Meta: Llama 3.2 11B Vision Instruct",
|
||||
"created": 1727222400,
|
||||
"description": "Llama 3.2 11B Vision is a multimodal model with 11 billion parameters, designed to handle tasks combining visual and textual data. It excels in tasks such as image captioning and visual question answering, bridging the gap between language generation and visual reasoning. Pre-trained on a massive dataset of image-text pairs, it performs well in complex, high-accuracy image analysis.\n\nIts ability to integrate visual understanding with language processing makes it an ideal solution for industries requiring comprehensive visual-linguistic AI applications, such as content creation, AI-driven customer service, and research.\n\nClick here for the [original model card](https://github.com/meta-llama/llama-models/blob/main/models/llama3_2/MODEL_CARD_VISION.md).\n\nUsage of this model is subject to [Meta's Acceptable Use Policy](https://www.llama.com/llama3/use-policy/).",
|
||||
"context_length": 16384,
|
||||
"context_length": 131072,
|
||||
"architecture": {
|
||||
"modality": "text+image->text",
|
||||
"input_modalities": [
|
||||
@ -5105,15 +5174,15 @@
|
||||
"prompt": "0.000000055",
|
||||
"completion": "0.000000055",
|
||||
"request": "0",
|
||||
"image": "0",
|
||||
"image": "0.00007948",
|
||||
"web_search": "0",
|
||||
"internal_reasoning": "0",
|
||||
"input_cache_read": "0",
|
||||
"input_cache_write": "0"
|
||||
},
|
||||
"top_provider": {
|
||||
"context_length": 16384,
|
||||
"max_completion_tokens": 16384,
|
||||
"context_length": 131072,
|
||||
"max_completion_tokens": 8192,
|
||||
"is_moderated": false
|
||||
},
|
||||
"per_request_limits": null
|
||||
@ -5444,8 +5513,8 @@
|
||||
"instruct_type": null
|
||||
},
|
||||
"pricing": {
|
||||
"prompt": "0.000002375",
|
||||
"completion": "0.0000095",
|
||||
"prompt": "0.0000025",
|
||||
"completion": "0.00001",
|
||||
"request": "0",
|
||||
"image": "0",
|
||||
"web_search": "0",
|
||||
@ -5478,8 +5547,8 @@
|
||||
"instruct_type": null
|
||||
},
|
||||
"pricing": {
|
||||
"prompt": "0.0000001425",
|
||||
"completion": "0.00000057",
|
||||
"prompt": "0.00000015",
|
||||
"completion": "0.0000006",
|
||||
"request": "0",
|
||||
"image": "0",
|
||||
"web_search": "0",
|
||||
@ -5856,8 +5925,8 @@
|
||||
"instruct_type": "llama3"
|
||||
},
|
||||
"pricing": {
|
||||
"prompt": "0.00000003",
|
||||
"completion": "0.00000006",
|
||||
"prompt": "0.00000005",
|
||||
"completion": "0.00000005",
|
||||
"request": "0",
|
||||
"image": "0",
|
||||
"web_search": "0",
|
||||
@ -5867,7 +5936,7 @@
|
||||
},
|
||||
"top_provider": {
|
||||
"context_length": 8192,
|
||||
"max_completion_tokens": 8192,
|
||||
"max_completion_tokens": null,
|
||||
"is_moderated": false
|
||||
},
|
||||
"per_request_limits": null
|
||||
@ -6385,74 +6454,6 @@
|
||||
},
|
||||
"per_request_limits": null
|
||||
},
|
||||
{
|
||||
"id": "qwen/qwen-2-7b-instruct:free",
|
||||
"name": "Qwen 2 7B Instruct (free)",
|
||||
"created": 1721088000,
|
||||
"description": "Qwen2 7B is a transformer-based model that excels in language understanding, multilingual capabilities, coding, mathematics, and reasoning.\n\nIt features SwiGLU activation, attention QKV bias, and group query attention. It is pretrained on extensive data with supervised finetuning and direct preference optimization.\n\nFor more details, see this [blog post](https://qwenlm.github.io/blog/qwen2/) and [GitHub repo](https://github.com/QwenLM/Qwen2).\n\nUsage of this model is subject to [Tongyi Qianwen LICENSE AGREEMENT](https://huggingface.co/Qwen/Qwen1.5-110B-Chat/blob/main/LICENSE).",
|
||||
"context_length": 8192,
|
||||
"architecture": {
|
||||
"modality": "text->text",
|
||||
"input_modalities": [
|
||||
"text"
|
||||
],
|
||||
"output_modalities": [
|
||||
"text"
|
||||
],
|
||||
"tokenizer": "Qwen",
|
||||
"instruct_type": "chatml"
|
||||
},
|
||||
"pricing": {
|
||||
"prompt": "0",
|
||||
"completion": "0",
|
||||
"request": "0",
|
||||
"image": "0",
|
||||
"web_search": "0",
|
||||
"internal_reasoning": "0",
|
||||
"input_cache_read": "0",
|
||||
"input_cache_write": "0"
|
||||
},
|
||||
"top_provider": {
|
||||
"context_length": 8192,
|
||||
"max_completion_tokens": 4096,
|
||||
"is_moderated": false
|
||||
},
|
||||
"per_request_limits": null
|
||||
},
|
||||
{
|
||||
"id": "qwen/qwen-2-7b-instruct",
|
||||
"name": "Qwen 2 7B Instruct",
|
||||
"created": 1721088000,
|
||||
"description": "Qwen2 7B is a transformer-based model that excels in language understanding, multilingual capabilities, coding, mathematics, and reasoning.\n\nIt features SwiGLU activation, attention QKV bias, and group query attention. It is pretrained on extensive data with supervised finetuning and direct preference optimization.\n\nFor more details, see this [blog post](https://qwenlm.github.io/blog/qwen2/) and [GitHub repo](https://github.com/QwenLM/Qwen2).\n\nUsage of this model is subject to [Tongyi Qianwen LICENSE AGREEMENT](https://huggingface.co/Qwen/Qwen1.5-110B-Chat/blob/main/LICENSE).",
|
||||
"context_length": 32768,
|
||||
"architecture": {
|
||||
"modality": "text->text",
|
||||
"input_modalities": [
|
||||
"text"
|
||||
],
|
||||
"output_modalities": [
|
||||
"text"
|
||||
],
|
||||
"tokenizer": "Qwen",
|
||||
"instruct_type": "chatml"
|
||||
},
|
||||
"pricing": {
|
||||
"prompt": "0.000000054",
|
||||
"completion": "0.000000054",
|
||||
"request": "0",
|
||||
"image": "0",
|
||||
"web_search": "0",
|
||||
"internal_reasoning": "0",
|
||||
"input_cache_read": "0",
|
||||
"input_cache_write": "0"
|
||||
},
|
||||
"top_provider": {
|
||||
"context_length": 32768,
|
||||
"max_completion_tokens": null,
|
||||
"is_moderated": false
|
||||
},
|
||||
"per_request_limits": null
|
||||
},
|
||||
{
|
||||
"id": "google/gemma-2-27b-it",
|
||||
"name": "Google: Gemma 2 27B",
|
||||
@ -6471,8 +6472,8 @@
|
||||
"instruct_type": "gemma"
|
||||
},
|
||||
"pricing": {
|
||||
"prompt": "0.00000027",
|
||||
"completion": "0.00000027",
|
||||
"prompt": "0.0000008",
|
||||
"completion": "0.0000008",
|
||||
"request": "0",
|
||||
"image": "0",
|
||||
"web_search": "0",
|
||||
@ -6482,7 +6483,7 @@
|
||||
},
|
||||
"top_provider": {
|
||||
"context_length": 8192,
|
||||
"max_completion_tokens": 8192,
|
||||
"max_completion_tokens": 2048,
|
||||
"is_moderated": false
|
||||
},
|
||||
"per_request_limits": null
|
||||
@ -6573,8 +6574,8 @@
|
||||
"instruct_type": "gemma"
|
||||
},
|
||||
"pricing": {
|
||||
"prompt": "0.00000003",
|
||||
"completion": "0.00000006",
|
||||
"prompt": "0.00000007",
|
||||
"completion": "0.00000007",
|
||||
"request": "0",
|
||||
"image": "0",
|
||||
"web_search": "0",
|
||||
@ -6584,7 +6585,7 @@
|
||||
},
|
||||
"top_provider": {
|
||||
"context_length": 8192,
|
||||
"max_completion_tokens": 8192,
|
||||
"max_completion_tokens": null,
|
||||
"is_moderated": false
|
||||
},
|
||||
"per_request_limits": null
|
||||
@ -6745,8 +6746,8 @@
|
||||
"instruct_type": "llama3"
|
||||
},
|
||||
"pricing": {
|
||||
"prompt": "0.0000007",
|
||||
"completion": "0.0000008",
|
||||
"prompt": "0.00000148",
|
||||
"completion": "0.00000148",
|
||||
"request": "0",
|
||||
"image": "0",
|
||||
"web_search": "0",
|
||||
@ -7411,40 +7412,6 @@
|
||||
},
|
||||
"per_request_limits": null
|
||||
},
|
||||
{
|
||||
"id": "meta-llama/llama-3-8b-instruct:free",
|
||||
"name": "Meta: Llama 3 8B Instruct (free)",
|
||||
"created": 1713398400,
|
||||
"description": "Meta's latest class of model (Llama 3) launched with a variety of sizes & flavors. This 8B instruct-tuned version was optimized for high quality dialogue usecases.\n\nIt has demonstrated strong performance compared to leading closed-source models in human evaluations.\n\nTo read more about the model release, [click here](https://ai.meta.com/blog/meta-llama-3/). Usage of this model is subject to [Meta's Acceptable Use Policy](https://llama.meta.com/llama3/use-policy/).",
|
||||
"context_length": 8192,
|
||||
"architecture": {
|
||||
"modality": "text->text",
|
||||
"input_modalities": [
|
||||
"text"
|
||||
],
|
||||
"output_modalities": [
|
||||
"text"
|
||||
],
|
||||
"tokenizer": "Llama3",
|
||||
"instruct_type": "llama3"
|
||||
},
|
||||
"pricing": {
|
||||
"prompt": "0",
|
||||
"completion": "0",
|
||||
"request": "0",
|
||||
"image": "0",
|
||||
"web_search": "0",
|
||||
"internal_reasoning": "0",
|
||||
"input_cache_read": "0",
|
||||
"input_cache_write": "0"
|
||||
},
|
||||
"top_provider": {
|
||||
"context_length": 8192,
|
||||
"max_completion_tokens": 4096,
|
||||
"is_moderated": false
|
||||
},
|
||||
"per_request_limits": null
|
||||
},
|
||||
{
|
||||
"id": "meta-llama/llama-3-8b-instruct",
|
||||
"name": "Meta: Llama 3 8B Instruct",
|
||||
@ -7703,8 +7670,8 @@
|
||||
"instruct_type": null
|
||||
},
|
||||
"pricing": {
|
||||
"prompt": "0.00000285",
|
||||
"completion": "0.00001425",
|
||||
"prompt": "0.000003",
|
||||
"completion": "0.000015",
|
||||
"request": "0",
|
||||
"image": "0",
|
||||
"web_search": "0",
|
||||
@ -7737,8 +7704,8 @@
|
||||
"instruct_type": null
|
||||
},
|
||||
"pricing": {
|
||||
"prompt": "0.00000285",
|
||||
"completion": "0.00001425",
|
||||
"prompt": "0.000003",
|
||||
"completion": "0.000015",
|
||||
"request": "0",
|
||||
"image": "0",
|
||||
"web_search": "0",
|
||||
@ -7805,8 +7772,8 @@
|
||||
"instruct_type": null
|
||||
},
|
||||
"pricing": {
|
||||
"prompt": "0.00000095",
|
||||
"completion": "0.0000019",
|
||||
"prompt": "0.000001",
|
||||
"completion": "0.000002",
|
||||
"request": "0",
|
||||
"image": "0",
|
||||
"web_search": "0",
|
||||
@ -7839,8 +7806,8 @@
|
||||
"instruct_type": null
|
||||
},
|
||||
"pricing": {
|
||||
"prompt": "0.000000475",
|
||||
"completion": "0.000001425",
|
||||
"prompt": "0.0000005",
|
||||
"completion": "0.0000015",
|
||||
"request": "0",
|
||||
"image": "0",
|
||||
"web_search": "0",
|
||||
@ -8083,8 +8050,8 @@
|
||||
"instruct_type": null
|
||||
},
|
||||
"pricing": {
|
||||
"prompt": "0.000000475",
|
||||
"completion": "0.000001425",
|
||||
"prompt": "0.0000005",
|
||||
"completion": "0.0000015",
|
||||
"request": "0",
|
||||
"image": "0",
|
||||
"web_search": "0",
|
||||
@ -8133,40 +8100,6 @@
|
||||
},
|
||||
"per_request_limits": null
|
||||
},
|
||||
{
|
||||
"id": "google/gemma-7b-it",
|
||||
"name": "Google: Gemma 7B",
|
||||
"created": 1708560000,
|
||||
"description": "Gemma by Google is an advanced, open-source language model family, leveraging the latest in decoder-only, text-to-text technology. It offers English language capabilities across text generation tasks like question answering, summarization, and reasoning. The Gemma 7B variant is comparable in performance to leading open source models.\n\nUsage of Gemma is subject to Google's [Gemma Terms of Use](https://ai.google.dev/gemma/terms).",
|
||||
"context_length": 8192,
|
||||
"architecture": {
|
||||
"modality": "text->text",
|
||||
"input_modalities": [
|
||||
"text"
|
||||
],
|
||||
"output_modalities": [
|
||||
"text"
|
||||
],
|
||||
"tokenizer": "Gemini",
|
||||
"instruct_type": "gemma"
|
||||
},
|
||||
"pricing": {
|
||||
"prompt": "0.00000015",
|
||||
"completion": "0.00000015",
|
||||
"request": "0",
|
||||
"image": "0",
|
||||
"web_search": "0",
|
||||
"internal_reasoning": "0",
|
||||
"input_cache_read": "0",
|
||||
"input_cache_write": "0"
|
||||
},
|
||||
"top_provider": {
|
||||
"context_length": 8192,
|
||||
"max_completion_tokens": null,
|
||||
"is_moderated": false
|
||||
},
|
||||
"per_request_limits": null
|
||||
},
|
||||
{
|
||||
"id": "openai/gpt-3.5-turbo-0613",
|
||||
"name": "OpenAI: GPT-3.5 Turbo (older v0613)",
|
||||
@ -8628,8 +8561,8 @@
|
||||
"instruct_type": "openchat"
|
||||
},
|
||||
"pricing": {
|
||||
"prompt": "0.000000055",
|
||||
"completion": "0.000000055",
|
||||
"prompt": "0.00000007",
|
||||
"completion": "0.00000007",
|
||||
"request": "0",
|
||||
"image": "0",
|
||||
"web_search": "0",
|
||||
@ -8639,7 +8572,7 @@
|
||||
},
|
||||
"top_provider": {
|
||||
"context_length": 8192,
|
||||
"max_completion_tokens": 8192,
|
||||
"max_completion_tokens": null,
|
||||
"is_moderated": false
|
||||
},
|
||||
"per_request_limits": null
|
||||
@ -8814,40 +8747,6 @@
|
||||
},
|
||||
"per_request_limits": null
|
||||
},
|
||||
{
|
||||
"id": "teknium/openhermes-2.5-mistral-7b",
|
||||
"name": "OpenHermes 2.5 Mistral 7B",
|
||||
"created": 1700438400,
|
||||
"description": "A continuation of [OpenHermes 2 model](/models/teknium/openhermes-2-mistral-7b), trained on additional code datasets.\nPotentially the most interesting finding from training on a good ratio (est. of around 7-14% of the total dataset) of code instruction was that it has boosted several non-code benchmarks, including TruthfulQA, AGIEval, and GPT4All suite. It did however reduce BigBench benchmark score, but the net gain overall is significant.",
|
||||
"context_length": 4096,
|
||||
"architecture": {
|
||||
"modality": "text->text",
|
||||
"input_modalities": [
|
||||
"text"
|
||||
],
|
||||
"output_modalities": [
|
||||
"text"
|
||||
],
|
||||
"tokenizer": "Mistral",
|
||||
"instruct_type": "chatml"
|
||||
},
|
||||
"pricing": {
|
||||
"prompt": "0.00000017",
|
||||
"completion": "0.00000017",
|
||||
"request": "0",
|
||||
"image": "0",
|
||||
"web_search": "0",
|
||||
"internal_reasoning": "0",
|
||||
"input_cache_read": "0",
|
||||
"input_cache_write": "0"
|
||||
},
|
||||
"top_provider": {
|
||||
"context_length": 4096,
|
||||
"max_completion_tokens": 4096,
|
||||
"is_moderated": false
|
||||
},
|
||||
"per_request_limits": null
|
||||
},
|
||||
{
|
||||
"id": "undi95/toppy-m-7b:free",
|
||||
"name": "Toppy M 7B (free)",
|
||||
@ -9664,40 +9563,6 @@
|
||||
},
|
||||
"per_request_limits": null
|
||||
},
|
||||
{
|
||||
"id": "gryphe/mythomax-l2-13b:free",
|
||||
"name": "MythoMax 13B (free)",
|
||||
"created": 1688256000,
|
||||
"description": "One of the highest performing and most popular fine-tunes of Llama 2 13B, with rich descriptions and roleplay. #merge",
|
||||
"context_length": 4096,
|
||||
"architecture": {
|
||||
"modality": "text->text",
|
||||
"input_modalities": [
|
||||
"text"
|
||||
],
|
||||
"output_modalities": [
|
||||
"text"
|
||||
],
|
||||
"tokenizer": "Llama2",
|
||||
"instruct_type": "alpaca"
|
||||
},
|
||||
"pricing": {
|
||||
"prompt": "0",
|
||||
"completion": "0",
|
||||
"request": "0",
|
||||
"image": "0",
|
||||
"web_search": "0",
|
||||
"internal_reasoning": "0",
|
||||
"input_cache_read": "0",
|
||||
"input_cache_write": "0"
|
||||
},
|
||||
"top_provider": {
|
||||
"context_length": 4096,
|
||||
"max_completion_tokens": 2048,
|
||||
"is_moderated": false
|
||||
},
|
||||
"per_request_limits": null
|
||||
},
|
||||
{
|
||||
"id": "gryphe/mythomax-l2-13b",
|
||||
"name": "MythoMax 13B",
|
||||
|
||||
@ -4,10 +4,12 @@ export declare enum E_OPENAI_MODEL {
|
||||
MODEL_DALL_E_3 = "dall-e-3",
|
||||
MODEL_DALL_E_2 = "dall-e-2",
|
||||
MODEL_GPT_4O_AUDIO_PREVIEW_2024_10_01 = "gpt-4o-audio-preview-2024-10-01",
|
||||
MODEL_GPT_4O_MINI_REALTIME_PREVIEW_2024_12_17 = "gpt-4o-mini-realtime-preview-2024-12-17",
|
||||
MODEL_GPT_4O_MINI_REALTIME_PREVIEW = "gpt-4o-mini-realtime-preview",
|
||||
MODEL_O1_PRO_2025_03_19 = "o1-pro-2025-03-19",
|
||||
MODEL_O1_PRO = "o1-pro",
|
||||
MODEL_GPT_4O_REALTIME_PREVIEW_2024_10_01 = "gpt-4o-realtime-preview-2024-10-01",
|
||||
MODEL_O1_2024_12_17 = "o1-2024-12-17",
|
||||
MODEL_GPT_4O_TRANSCRIBE = "gpt-4o-transcribe",
|
||||
MODEL_O1 = "o1",
|
||||
MODEL_GPT_4O_MINI_TRANSCRIBE = "gpt-4o-mini-transcribe",
|
||||
MODEL_GPT_4O_REALTIME_PREVIEW = "gpt-4o-realtime-preview",
|
||||
MODEL_BABBAGE_002 = "babbage-002",
|
||||
@ -16,19 +18,18 @@ export declare enum E_OPENAI_MODEL {
|
||||
MODEL_TEXT_EMBEDDING_3_LARGE = "text-embedding-3-large",
|
||||
MODEL_GPT_4 = "gpt-4",
|
||||
MODEL_TEXT_EMBEDDING_ADA_002 = "text-embedding-ada-002",
|
||||
MODEL_OMNI_MODERATION_LATEST = "omni-moderation-latest",
|
||||
MODEL_TTS_1_HD = "tts-1-hd",
|
||||
MODEL_GPT_4O_MINI_AUDIO_PREVIEW = "gpt-4o-mini-audio-preview",
|
||||
MODEL_GPT_4O_AUDIO_PREVIEW = "gpt-4o-audio-preview",
|
||||
MODEL_O1_PREVIEW_2024_09_12 = "o1-preview-2024-09-12",
|
||||
MODEL_GPT_4O_MINI_REALTIME_PREVIEW = "gpt-4o-mini-realtime-preview",
|
||||
MODEL_GPT_4O_MINI_REALTIME_PREVIEW_2024_12_17 = "gpt-4o-mini-realtime-preview-2024-12-17",
|
||||
MODEL_GPT_3_5_TURBO_INSTRUCT_0914 = "gpt-3.5-turbo-instruct-0914",
|
||||
MODEL_GPT_4O_MINI_SEARCH_PREVIEW = "gpt-4o-mini-search-preview",
|
||||
MODEL_O1 = "o1",
|
||||
MODEL_TTS_1_1106 = "tts-1-1106",
|
||||
MODEL_O1_2024_12_17 = "o1-2024-12-17",
|
||||
MODEL_DAVINCI_002 = "davinci-002",
|
||||
MODEL_GPT_3_5_TURBO_1106 = "gpt-3.5-turbo-1106",
|
||||
MODEL_O1_PRO = "o1-pro",
|
||||
MODEL_O1_PRO_2025_03_19 = "o1-pro-2025-03-19",
|
||||
MODEL_GPT_4_TURBO = "gpt-4-turbo",
|
||||
MODEL_GPT_3_5_TURBO_INSTRUCT = "gpt-3.5-turbo-instruct",
|
||||
MODEL_GPT_3_5_TURBO = "gpt-3.5-turbo",
|
||||
@ -49,7 +50,6 @@ export declare enum E_OPENAI_MODEL {
|
||||
MODEL_GPT_4_5_PREVIEW = "gpt-4.5-preview",
|
||||
MODEL_GPT_4_5_PREVIEW_2025_02_27 = "gpt-4.5-preview-2025-02-27",
|
||||
MODEL_GPT_4O_SEARCH_PREVIEW_2025_03_11 = "gpt-4o-search-preview-2025-03-11",
|
||||
MODEL_OMNI_MODERATION_LATEST = "omni-moderation-latest",
|
||||
MODEL_TTS_1 = "tts-1",
|
||||
MODEL_OMNI_MODERATION_2024_09_26 = "omni-moderation-2024-09-26",
|
||||
MODEL_TEXT_EMBEDDING_3_SMALL = "text-embedding-3-small",
|
||||
@ -57,9 +57,9 @@ export declare enum E_OPENAI_MODEL {
|
||||
MODEL_GPT_4O_MINI = "gpt-4o-mini",
|
||||
MODEL_GPT_4O_2024_08_06 = "gpt-4o-2024-08-06",
|
||||
MODEL_GPT_4O_MINI_2024_07_18 = "gpt-4o-mini-2024-07-18",
|
||||
MODEL_GPT_4_TURBO_PREVIEW = "gpt-4-turbo-preview",
|
||||
MODEL_O1_MINI = "o1-mini",
|
||||
MODEL_GPT_4O_MINI_AUDIO_PREVIEW_2024_12_17 = "gpt-4o-mini-audio-preview-2024-12-17",
|
||||
MODEL_O1_MINI_2024_09_12 = "o1-mini-2024-09-12",
|
||||
MODEL_GPT_4_0125_PREVIEW = "gpt-4-0125-preview"
|
||||
MODEL_GPT_4_0125_PREVIEW = "gpt-4-0125-preview",
|
||||
MODEL_GPT_4_TURBO_PREVIEW = "gpt-4-turbo-preview"
|
||||
}
|
||||
|
||||
@ -5,10 +5,12 @@ export var E_OPENAI_MODEL;
|
||||
E_OPENAI_MODEL["MODEL_DALL_E_3"] = "dall-e-3";
|
||||
E_OPENAI_MODEL["MODEL_DALL_E_2"] = "dall-e-2";
|
||||
E_OPENAI_MODEL["MODEL_GPT_4O_AUDIO_PREVIEW_2024_10_01"] = "gpt-4o-audio-preview-2024-10-01";
|
||||
E_OPENAI_MODEL["MODEL_GPT_4O_MINI_REALTIME_PREVIEW_2024_12_17"] = "gpt-4o-mini-realtime-preview-2024-12-17";
|
||||
E_OPENAI_MODEL["MODEL_GPT_4O_MINI_REALTIME_PREVIEW"] = "gpt-4o-mini-realtime-preview";
|
||||
E_OPENAI_MODEL["MODEL_O1_PRO_2025_03_19"] = "o1-pro-2025-03-19";
|
||||
E_OPENAI_MODEL["MODEL_O1_PRO"] = "o1-pro";
|
||||
E_OPENAI_MODEL["MODEL_GPT_4O_REALTIME_PREVIEW_2024_10_01"] = "gpt-4o-realtime-preview-2024-10-01";
|
||||
E_OPENAI_MODEL["MODEL_O1_2024_12_17"] = "o1-2024-12-17";
|
||||
E_OPENAI_MODEL["MODEL_GPT_4O_TRANSCRIBE"] = "gpt-4o-transcribe";
|
||||
E_OPENAI_MODEL["MODEL_O1"] = "o1";
|
||||
E_OPENAI_MODEL["MODEL_GPT_4O_MINI_TRANSCRIBE"] = "gpt-4o-mini-transcribe";
|
||||
E_OPENAI_MODEL["MODEL_GPT_4O_REALTIME_PREVIEW"] = "gpt-4o-realtime-preview";
|
||||
E_OPENAI_MODEL["MODEL_BABBAGE_002"] = "babbage-002";
|
||||
@ -17,19 +19,18 @@ export var E_OPENAI_MODEL;
|
||||
E_OPENAI_MODEL["MODEL_TEXT_EMBEDDING_3_LARGE"] = "text-embedding-3-large";
|
||||
E_OPENAI_MODEL["MODEL_GPT_4"] = "gpt-4";
|
||||
E_OPENAI_MODEL["MODEL_TEXT_EMBEDDING_ADA_002"] = "text-embedding-ada-002";
|
||||
E_OPENAI_MODEL["MODEL_OMNI_MODERATION_LATEST"] = "omni-moderation-latest";
|
||||
E_OPENAI_MODEL["MODEL_TTS_1_HD"] = "tts-1-hd";
|
||||
E_OPENAI_MODEL["MODEL_GPT_4O_MINI_AUDIO_PREVIEW"] = "gpt-4o-mini-audio-preview";
|
||||
E_OPENAI_MODEL["MODEL_GPT_4O_AUDIO_PREVIEW"] = "gpt-4o-audio-preview";
|
||||
E_OPENAI_MODEL["MODEL_O1_PREVIEW_2024_09_12"] = "o1-preview-2024-09-12";
|
||||
E_OPENAI_MODEL["MODEL_GPT_4O_MINI_REALTIME_PREVIEW"] = "gpt-4o-mini-realtime-preview";
|
||||
E_OPENAI_MODEL["MODEL_GPT_4O_MINI_REALTIME_PREVIEW_2024_12_17"] = "gpt-4o-mini-realtime-preview-2024-12-17";
|
||||
E_OPENAI_MODEL["MODEL_GPT_3_5_TURBO_INSTRUCT_0914"] = "gpt-3.5-turbo-instruct-0914";
|
||||
E_OPENAI_MODEL["MODEL_GPT_4O_MINI_SEARCH_PREVIEW"] = "gpt-4o-mini-search-preview";
|
||||
E_OPENAI_MODEL["MODEL_O1"] = "o1";
|
||||
E_OPENAI_MODEL["MODEL_TTS_1_1106"] = "tts-1-1106";
|
||||
E_OPENAI_MODEL["MODEL_O1_2024_12_17"] = "o1-2024-12-17";
|
||||
E_OPENAI_MODEL["MODEL_DAVINCI_002"] = "davinci-002";
|
||||
E_OPENAI_MODEL["MODEL_GPT_3_5_TURBO_1106"] = "gpt-3.5-turbo-1106";
|
||||
E_OPENAI_MODEL["MODEL_O1_PRO"] = "o1-pro";
|
||||
E_OPENAI_MODEL["MODEL_O1_PRO_2025_03_19"] = "o1-pro-2025-03-19";
|
||||
E_OPENAI_MODEL["MODEL_GPT_4_TURBO"] = "gpt-4-turbo";
|
||||
E_OPENAI_MODEL["MODEL_GPT_3_5_TURBO_INSTRUCT"] = "gpt-3.5-turbo-instruct";
|
||||
E_OPENAI_MODEL["MODEL_GPT_3_5_TURBO"] = "gpt-3.5-turbo";
|
||||
@ -50,7 +51,6 @@ export var E_OPENAI_MODEL;
|
||||
E_OPENAI_MODEL["MODEL_GPT_4_5_PREVIEW"] = "gpt-4.5-preview";
|
||||
E_OPENAI_MODEL["MODEL_GPT_4_5_PREVIEW_2025_02_27"] = "gpt-4.5-preview-2025-02-27";
|
||||
E_OPENAI_MODEL["MODEL_GPT_4O_SEARCH_PREVIEW_2025_03_11"] = "gpt-4o-search-preview-2025-03-11";
|
||||
E_OPENAI_MODEL["MODEL_OMNI_MODERATION_LATEST"] = "omni-moderation-latest";
|
||||
E_OPENAI_MODEL["MODEL_TTS_1"] = "tts-1";
|
||||
E_OPENAI_MODEL["MODEL_OMNI_MODERATION_2024_09_26"] = "omni-moderation-2024-09-26";
|
||||
E_OPENAI_MODEL["MODEL_TEXT_EMBEDDING_3_SMALL"] = "text-embedding-3-small";
|
||||
@ -58,10 +58,10 @@ export var E_OPENAI_MODEL;
|
||||
E_OPENAI_MODEL["MODEL_GPT_4O_MINI"] = "gpt-4o-mini";
|
||||
E_OPENAI_MODEL["MODEL_GPT_4O_2024_08_06"] = "gpt-4o-2024-08-06";
|
||||
E_OPENAI_MODEL["MODEL_GPT_4O_MINI_2024_07_18"] = "gpt-4o-mini-2024-07-18";
|
||||
E_OPENAI_MODEL["MODEL_GPT_4_TURBO_PREVIEW"] = "gpt-4-turbo-preview";
|
||||
E_OPENAI_MODEL["MODEL_O1_MINI"] = "o1-mini";
|
||||
E_OPENAI_MODEL["MODEL_GPT_4O_MINI_AUDIO_PREVIEW_2024_12_17"] = "gpt-4o-mini-audio-preview-2024-12-17";
|
||||
E_OPENAI_MODEL["MODEL_O1_MINI_2024_09_12"] = "o1-mini-2024-09-12";
|
||||
E_OPENAI_MODEL["MODEL_GPT_4_0125_PREVIEW"] = "gpt-4-0125-preview";
|
||||
E_OPENAI_MODEL["MODEL_GPT_4_TURBO_PREVIEW"] = "gpt-4-turbo-preview";
|
||||
})(E_OPENAI_MODEL || (E_OPENAI_MODEL = {}));
|
||||
//# sourceMappingURL=data:application/json;base64,eyJ2ZXJzaW9uIjozLCJmaWxlIjoib3BlbmFpLW1vZGVscy5qcyIsInNvdXJjZVJvb3QiOiIiLCJzb3VyY2VzIjpbIi4uLy4uLy4uL3NyYy9tb2RlbHMvY2FjaGUvb3BlbmFpLW1vZGVscy50cyJdLCJuYW1lcyI6W10sIm1hcHBpbmdzIjoiQUFBQSxNQUFNLENBQU4sSUFBWSxjQWdFWDtBQWhFRCxXQUFZLGNBQWM7SUFDeEIsaUdBQStFLENBQUE7SUFDL0UsMkZBQXlFLENBQUE7SUFDekUsNkNBQTJCLENBQUE7SUFDM0IsNkNBQTJCLENBQUE7SUFDM0IsMkZBQXlFLENBQUE7SUFDekUsMkdBQXlGLENBQUE7SUFDekYscUZBQW1FLENBQUE7SUFDbkUsaUdBQStFLENBQUE7SUFDL0UsK0RBQTZDLENBQUE7SUFDN0MseUVBQXVELENBQUE7SUFDdkQsMkVBQXlELENBQUE7SUFDekQsbURBQWlDLENBQUE7SUFDakMsMkRBQXlDLENBQUE7SUFDekMsdURBQXFDLENBQUE7SUFDckMseUVBQXVELENBQUE7SUFDdkQsdUNBQXFCLENBQUE7SUFDckIseUVBQXVELENBQUE7SUFDdkQsNkNBQTJCLENBQUE7SUFDM0IsK0VBQTZELENBQUE7SUFDN0QscUVBQW1ELENBQUE7SUFDbkQsdUVBQXFELENBQUE7SUFDckQsbUZBQWlFLENBQUE7SUFDakUsaUZBQStELENBQUE7SUFDL0QsaUNBQWUsQ0FBQTtJQUNmLGlEQUErQixDQUFBO0lBQy9CLHVEQUFxQyxDQUFBO0lBQ3JDLG1EQUFpQyxDQUFBO0lBQ2pDLGlFQUErQyxDQUFBO0lBQy9DLHlDQUF1QixDQUFBO0lBQ3ZCLCtEQUE2QyxDQUFBO0lBQzdDLG1EQUFpQyxDQUFBO0lBQ2pDLHlFQUF1RCxDQUFBO0lBQ3ZELHVEQUFxQyxDQUFBO0lBQ3JDLCtEQUE2QyxDQUFBO0lBQzdDLHVHQUFxRixDQUFBO0lBQ3JGLCtEQUE2QyxDQUFBO0lBQzdDLCtDQUE2QixDQUFBO0lBQzdCLGlFQUErQyxDQUFBO0lBQy9DLCtEQUE2QyxDQUFBO0lBQzdDLCtEQUE2QyxDQUFBO0lBQzdDLHlFQUF1RCxDQUFBO0lBQ3ZELGlFQUErQyxDQUFBO0lBQy9DLGlEQUErQixDQUFBO0lBQy9CLGlEQUErQixDQUFBO0lBQy9CLHVFQUFxRCxDQUFBO0lBQ3JELGlFQUErQyxDQUFBO0lBQy9DLDJDQUF5QixDQUFBO0lBQ3pCLDJEQUF5QyxDQUFBO0lBQ3pDLGlGQUErRCxDQUFBO0lBQy9ELDZGQUEyRSxDQUFBO0lBQzNFLHlFQUF1RCxDQUFBO0lBQ3ZELHVDQUFxQixDQUFBO0lBQ3JCLGlGQUErRCxDQUFBO0lBQy9ELHlFQUF1RCxDQUFBO0lBQ3ZELHlDQUF1QixDQUFBO0lBQ3ZCLG1EQUFpQyxDQUFBO0lBQ2pDLCtEQUE2QyxDQUFBO0lBQzdDLHlFQUF1RCxDQUFBO0lBQ3ZELG1FQUFpRCxDQUFBO0lBQ2pELDJDQUF5QixDQUFBO0lBQ3pCLHFHQUFtRixDQUFBO0lBQ25GLGlFQUErQyxDQUFBO0lBQy9DLGlFQUErQyxDQUFBO0FBQ2pELENBQUMsRUFoRVcsY0FBYyxLQUFkLGNBQWMsUUFnRXpCIn0=
|
||||
//# sourceMappingURL=data:application/json;base64,eyJ2ZXJzaW9uIjozLCJmaWxlIjoib3BlbmFpLW1vZGVscy5qcyIsInNvdXJjZVJvb3QiOiIiLCJzb3VyY2VzIjpbIi4uLy4uLy4uL3NyYy9tb2RlbHMvY2FjaGUvb3BlbmFpLW1vZGVscy50cyJdLCJuYW1lcyI6W10sIm1hcHBpbmdzIjoiQUFBQSxNQUFNLENBQU4sSUFBWSxjQWdFWDtBQWhFRCxXQUFZLGNBQWM7SUFDeEIsaUdBQStFLENBQUE7SUFDL0UsMkZBQXlFLENBQUE7SUFDekUsNkNBQTJCLENBQUE7SUFDM0IsNkNBQTJCLENBQUE7SUFDM0IsMkZBQXlFLENBQUE7SUFDekUsK0RBQTZDLENBQUE7SUFDN0MseUNBQXVCLENBQUE7SUFDdkIsaUdBQStFLENBQUE7SUFDL0UsdURBQXFDLENBQUE7SUFDckMsK0RBQTZDLENBQUE7SUFDN0MsaUNBQWUsQ0FBQTtJQUNmLHlFQUF1RCxDQUFBO0lBQ3ZELDJFQUF5RCxDQUFBO0lBQ3pELG1EQUFpQyxDQUFBO0lBQ2pDLDJEQUF5QyxDQUFBO0lBQ3pDLHVEQUFxQyxDQUFBO0lBQ3JDLHlFQUF1RCxDQUFBO0lBQ3ZELHVDQUFxQixDQUFBO0lBQ3JCLHlFQUF1RCxDQUFBO0lBQ3ZELHlFQUF1RCxDQUFBO0lBQ3ZELDZDQUEyQixDQUFBO0lBQzNCLCtFQUE2RCxDQUFBO0lBQzdELHFFQUFtRCxDQUFBO0lBQ25ELHVFQUFxRCxDQUFBO0lBQ3JELHFGQUFtRSxDQUFBO0lBQ25FLDJHQUF5RixDQUFBO0lBQ3pGLG1GQUFpRSxDQUFBO0lBQ2pFLGlGQUErRCxDQUFBO0lBQy9ELGlEQUErQixDQUFBO0lBQy9CLG1EQUFpQyxDQUFBO0lBQ2pDLGlFQUErQyxDQUFBO0lBQy9DLG1EQUFpQyxDQUFBO0lBQ2pDLHlFQUF1RCxDQUFBO0lBQ3ZELHVEQUFxQyxDQUFBO0lBQ3JDLCtEQUE2QyxDQUFBO0lBQzdDLHVHQUFxRixDQUFBO0lBQ3JGLCtEQUE2QyxDQUFBO0lBQzdDLCtDQUE2QixDQUFBO0lBQzdCLGlFQUErQyxDQUFBO0lBQy9DLCtEQUE2QyxDQUFBO0lBQzdDLCtEQUE2QyxDQUFBO0lBQzdDLHlFQUF1RCxDQUFBO0lBQ3ZELGlFQUErQyxDQUFBO0lBQy9DLGlEQUErQixDQUFBO0lBQy9CLGlEQUErQixDQUFBO0lBQy9CLHVFQUFxRCxDQUFBO0lBQ3JELGlFQUErQyxDQUFBO0lBQy9DLDJDQUF5QixDQUFBO0lBQ3pCLDJEQUF5QyxDQUFBO0lBQ3pDLGlGQUErRCxDQUFBO0lBQy9ELDZGQUEyRSxDQUFBO0lBQzNFLHVDQUFxQixDQUFBO0lBQ3JCLGlGQUErRCxDQUFBO0lBQy9ELHlFQUF1RCxDQUFBO0lBQ3ZELHlDQUF1QixDQUFBO0lBQ3ZCLG1EQUFpQyxDQUFBO0lBQ2pDLCtEQUE2QyxDQUFBO0lBQzdDLHlFQUF1RCxDQUFBO0lBQ3ZELDJDQUF5QixDQUFBO0lBQ3pCLHFHQUFtRixDQUFBO0lBQ25GLGlFQUErQyxDQUFBO0lBQy9DLGlFQUErQyxDQUFBO0lBQy9DLG1FQUFpRCxDQUFBO0FBQ25ELENBQUMsRUFoRVcsY0FBYyxLQUFkLGNBQWMsUUFnRXpCIn0=
|
||||
@ -1,4 +1,5 @@
|
||||
export declare enum E_OPENROUTER_MODEL_FREE {
|
||||
MODEL_FREE_OPENROUTER_QUASAR_ALPHA = "openrouter/quasar-alpha",
|
||||
MODEL_FREE_DEEPSEEK_DEEPSEEK_V3_BASE_FREE = "deepseek/deepseek-v3-base:free",
|
||||
MODEL_FREE_ALLENAI_MOLMO_7B_D_FREE = "allenai/molmo-7b-d:free",
|
||||
MODEL_FREE_BYTEDANCE_RESEARCH_UI_TARS_72B_FREE = "bytedance-research/ui-tars-72b:free",
|
||||
@ -46,14 +47,11 @@ export declare enum E_OPENROUTER_MODEL_FREE {
|
||||
MODEL_FREE_GOOGLE_GEMINI_FLASH_1_5_8B_EXP = "google/gemini-flash-1.5-8b-exp",
|
||||
MODEL_FREE_META_LLAMA_LLAMA_3_1_8B_INSTRUCT_FREE = "meta-llama/llama-3.1-8b-instruct:free",
|
||||
MODEL_FREE_MISTRALAI_MISTRAL_NEMO_FREE = "mistralai/mistral-nemo:free",
|
||||
MODEL_FREE_QWEN_QWEN_2_7B_INSTRUCT_FREE = "qwen/qwen-2-7b-instruct:free",
|
||||
MODEL_FREE_GOOGLE_GEMMA_2_9B_IT_FREE = "google/gemma-2-9b-it:free",
|
||||
MODEL_FREE_MISTRALAI_MISTRAL_7B_INSTRUCT_FREE = "mistralai/mistral-7b-instruct:free",
|
||||
MODEL_FREE_MICROSOFT_PHI_3_MINI_128K_INSTRUCT_FREE = "microsoft/phi-3-mini-128k-instruct:free",
|
||||
MODEL_FREE_MICROSOFT_PHI_3_MEDIUM_128K_INSTRUCT_FREE = "microsoft/phi-3-medium-128k-instruct:free",
|
||||
MODEL_FREE_META_LLAMA_LLAMA_3_8B_INSTRUCT_FREE = "meta-llama/llama-3-8b-instruct:free",
|
||||
MODEL_FREE_OPENCHAT_OPENCHAT_7B_FREE = "openchat/openchat-7b:free",
|
||||
MODEL_FREE_UNDI95_TOPPY_M_7B_FREE = "undi95/toppy-m-7b:free",
|
||||
MODEL_FREE_HUGGINGFACEH4_ZEPHYR_7B_BETA_FREE = "huggingfaceh4/zephyr-7b-beta:free",
|
||||
MODEL_FREE_GRYPHE_MYTHOMAX_L2_13B_FREE = "gryphe/mythomax-l2-13b:free"
|
||||
MODEL_FREE_HUGGINGFACEH4_ZEPHYR_7B_BETA_FREE = "huggingfaceh4/zephyr-7b-beta:free"
|
||||
}
|
||||
|
||||
@ -1,5 +1,6 @@
|
||||
export var E_OPENROUTER_MODEL_FREE;
|
||||
(function (E_OPENROUTER_MODEL_FREE) {
|
||||
E_OPENROUTER_MODEL_FREE["MODEL_FREE_OPENROUTER_QUASAR_ALPHA"] = "openrouter/quasar-alpha";
|
||||
E_OPENROUTER_MODEL_FREE["MODEL_FREE_DEEPSEEK_DEEPSEEK_V3_BASE_FREE"] = "deepseek/deepseek-v3-base:free";
|
||||
E_OPENROUTER_MODEL_FREE["MODEL_FREE_ALLENAI_MOLMO_7B_D_FREE"] = "allenai/molmo-7b-d:free";
|
||||
E_OPENROUTER_MODEL_FREE["MODEL_FREE_BYTEDANCE_RESEARCH_UI_TARS_72B_FREE"] = "bytedance-research/ui-tars-72b:free";
|
||||
@ -47,15 +48,12 @@ export var E_OPENROUTER_MODEL_FREE;
|
||||
E_OPENROUTER_MODEL_FREE["MODEL_FREE_GOOGLE_GEMINI_FLASH_1_5_8B_EXP"] = "google/gemini-flash-1.5-8b-exp";
|
||||
E_OPENROUTER_MODEL_FREE["MODEL_FREE_META_LLAMA_LLAMA_3_1_8B_INSTRUCT_FREE"] = "meta-llama/llama-3.1-8b-instruct:free";
|
||||
E_OPENROUTER_MODEL_FREE["MODEL_FREE_MISTRALAI_MISTRAL_NEMO_FREE"] = "mistralai/mistral-nemo:free";
|
||||
E_OPENROUTER_MODEL_FREE["MODEL_FREE_QWEN_QWEN_2_7B_INSTRUCT_FREE"] = "qwen/qwen-2-7b-instruct:free";
|
||||
E_OPENROUTER_MODEL_FREE["MODEL_FREE_GOOGLE_GEMMA_2_9B_IT_FREE"] = "google/gemma-2-9b-it:free";
|
||||
E_OPENROUTER_MODEL_FREE["MODEL_FREE_MISTRALAI_MISTRAL_7B_INSTRUCT_FREE"] = "mistralai/mistral-7b-instruct:free";
|
||||
E_OPENROUTER_MODEL_FREE["MODEL_FREE_MICROSOFT_PHI_3_MINI_128K_INSTRUCT_FREE"] = "microsoft/phi-3-mini-128k-instruct:free";
|
||||
E_OPENROUTER_MODEL_FREE["MODEL_FREE_MICROSOFT_PHI_3_MEDIUM_128K_INSTRUCT_FREE"] = "microsoft/phi-3-medium-128k-instruct:free";
|
||||
E_OPENROUTER_MODEL_FREE["MODEL_FREE_META_LLAMA_LLAMA_3_8B_INSTRUCT_FREE"] = "meta-llama/llama-3-8b-instruct:free";
|
||||
E_OPENROUTER_MODEL_FREE["MODEL_FREE_OPENCHAT_OPENCHAT_7B_FREE"] = "openchat/openchat-7b:free";
|
||||
E_OPENROUTER_MODEL_FREE["MODEL_FREE_UNDI95_TOPPY_M_7B_FREE"] = "undi95/toppy-m-7b:free";
|
||||
E_OPENROUTER_MODEL_FREE["MODEL_FREE_HUGGINGFACEH4_ZEPHYR_7B_BETA_FREE"] = "huggingfaceh4/zephyr-7b-beta:free";
|
||||
E_OPENROUTER_MODEL_FREE["MODEL_FREE_GRYPHE_MYTHOMAX_L2_13B_FREE"] = "gryphe/mythomax-l2-13b:free";
|
||||
})(E_OPENROUTER_MODEL_FREE || (E_OPENROUTER_MODEL_FREE = {}));
|
||||
//# sourceMappingURL=data:application/json;base64,eyJ2ZXJzaW9uIjozLCJmaWxlIjoib3BlbnJvdXRlci1tb2RlbHMtZnJlZS5qcyIsInNvdXJjZVJvb3QiOiIiLCJzb3VyY2VzIjpbIi4uLy4uLy4uL3NyYy9tb2RlbHMvY2FjaGUvb3BlbnJvdXRlci1tb2RlbHMtZnJlZS50cyJdLCJuYW1lcyI6W10sIm1hcHBpbmdzIjoiQUFBQSxNQUFNLENBQU4sSUFBWSx1QkEwRFg7QUExREQsV0FBWSx1QkFBdUI7SUFDakMsdUdBQTRFLENBQUE7SUFDNUUseUZBQThELENBQUE7SUFDOUQsaUhBQXNGLENBQUE7SUFDdEYsMkdBQWdGLENBQUE7SUFDaEYsbUhBQXdGLENBQUE7SUFDeEYsNkdBQWtGLENBQUE7SUFDbEYsaUhBQXNGLENBQUE7SUFDdEYsaUdBQXNFLENBQUE7SUFDdEUscUlBQTBHLENBQUE7SUFDMUcsbUdBQXdFLENBQUE7SUFDeEUscUdBQTBFLENBQUE7SUFDMUUsNkZBQWtFLENBQUE7SUFDbEUsNkZBQWtFLENBQUE7SUFDbEUsK0ZBQW9FLENBQUE7SUFDcEUsMkZBQWdFLENBQUE7SUFDaEUsK0ZBQW9FLENBQUE7SUFDcEUsdUdBQTRFLENBQUE7SUFDNUUsNkVBQWtELENBQUE7SUFDbEQsK0hBQW9HLENBQUE7SUFDcEcsNklBQWtILENBQUE7SUFDbEgsbUpBQXdILENBQUE7SUFDeEgsNklBQWtILENBQUE7SUFDbEgsbUhBQXdGLENBQUE7SUFDeEYsNkdBQWtGLENBQUE7SUFDbEYsdUlBQTRHLENBQUE7SUFDNUcsK0hBQW9HLENBQUE7SUFDcEcsK0hBQW9HLENBQUE7SUFDcEcsaUlBQXNHLENBQUE7SUFDdEcsNkhBQWtHLENBQUE7SUFDbEcsNkZBQWtFLENBQUE7SUFDbEUsNkhBQWtHLENBQUE7SUFDbEcsaUdBQXNFLENBQUE7SUFDdEUsdUlBQTRHLENBQUE7SUFDNUcsMkdBQWdGLENBQUE7SUFDaEYsdUhBQTRGLENBQUE7SUFDNUYsNkZBQWtFLENBQUE7SUFDbEUsMkhBQWdHLENBQUE7SUFDaEcscUhBQTBGLENBQUE7SUFDMUYsaUlBQXNHLENBQUE7SUFDdEcscUhBQTBGLENBQUE7SUFDMUYscUhBQTBGLENBQUE7SUFDMUYscUlBQTBHLENBQUE7SUFDMUcseUdBQThFLENBQUE7SUFDOUUsNkdBQWtGLENBQUE7SUFDbEYsdUdBQTRFLENBQUE7SUFDNUUscUhBQTBGLENBQUE7SUFDMUYsaUdBQXNFLENBQUE7SUFDdEUsbUdBQXdFLENBQUE7SUFDeEUsNkZBQWtFLENBQUE7SUFDbEUsK0dBQW9GLENBQUE7SUFDcEYseUhBQThGLENBQUE7SUFDOUYsNkhBQWtHLENBQUE7SUFDbEcsaUhBQXNGLENBQUE7SUFDdEYsNkZBQWtFLENBQUE7SUFDbEUsdUZBQTRELENBQUE7SUFDNUQsNkdBQWtGLENBQUE7SUFDbEYsaUdBQXNFLENBQUE7QUFDeEUsQ0FBQyxFQTFEVyx1QkFBdUIsS0FBdkIsdUJBQXVCLFFBMERsQyJ9
|
||||
//# sourceMappingURL=data:application/json;base64,eyJ2ZXJzaW9uIjozLCJmaWxlIjoib3BlbnJvdXRlci1tb2RlbHMtZnJlZS5qcyIsInNvdXJjZVJvb3QiOiIiLCJzb3VyY2VzIjpbIi4uLy4uLy4uL3NyYy9tb2RlbHMvY2FjaGUvb3BlbnJvdXRlci1tb2RlbHMtZnJlZS50cyJdLCJuYW1lcyI6W10sIm1hcHBpbmdzIjoiQUFBQSxNQUFNLENBQU4sSUFBWSx1QkF3RFg7QUF4REQsV0FBWSx1QkFBdUI7SUFDakMseUZBQThELENBQUE7SUFDOUQsdUdBQTRFLENBQUE7SUFDNUUseUZBQThELENBQUE7SUFDOUQsaUhBQXNGLENBQUE7SUFDdEYsMkdBQWdGLENBQUE7SUFDaEYsbUhBQXdGLENBQUE7SUFDeEYsNkdBQWtGLENBQUE7SUFDbEYsaUhBQXNGLENBQUE7SUFDdEYsaUdBQXNFLENBQUE7SUFDdEUscUlBQTBHLENBQUE7SUFDMUcsbUdBQXdFLENBQUE7SUFDeEUscUdBQTBFLENBQUE7SUFDMUUsNkZBQWtFLENBQUE7SUFDbEUsNkZBQWtFLENBQUE7SUFDbEUsK0ZBQW9FLENBQUE7SUFDcEUsMkZBQWdFLENBQUE7SUFDaEUsK0ZBQW9FLENBQUE7SUFDcEUsdUdBQTRFLENBQUE7SUFDNUUsNkVBQWtELENBQUE7SUFDbEQsK0hBQW9HLENBQUE7SUFDcEcsNklBQWtILENBQUE7SUFDbEgsbUpBQXdILENBQUE7SUFDeEgsNklBQWtILENBQUE7SUFDbEgsbUhBQXdGLENBQUE7SUFDeEYsNkdBQWtGLENBQUE7SUFDbEYsdUlBQTRHLENBQUE7SUFDNUcsK0hBQW9HLENBQUE7SUFDcEcsK0hBQW9HLENBQUE7SUFDcEcsaUlBQXNHLENBQUE7SUFDdEcsNkhBQWtHLENBQUE7SUFDbEcsNkZBQWtFLENBQUE7SUFDbEUsNkhBQWtHLENBQUE7SUFDbEcsaUdBQXNFLENBQUE7SUFDdEUsdUlBQTRHLENBQUE7SUFDNUcsMkdBQWdGLENBQUE7SUFDaEYsdUhBQTRGLENBQUE7SUFDNUYsNkZBQWtFLENBQUE7SUFDbEUsMkhBQWdHLENBQUE7SUFDaEcscUhBQTBGLENBQUE7SUFDMUYsaUlBQXNHLENBQUE7SUFDdEcscUhBQTBGLENBQUE7SUFDMUYscUhBQTBGLENBQUE7SUFDMUYscUlBQTBHLENBQUE7SUFDMUcseUdBQThFLENBQUE7SUFDOUUsNkdBQWtGLENBQUE7SUFDbEYsdUdBQTRFLENBQUE7SUFDNUUscUhBQTBGLENBQUE7SUFDMUYsaUdBQXNFLENBQUE7SUFDdEUsNkZBQWtFLENBQUE7SUFDbEUsK0dBQW9GLENBQUE7SUFDcEYseUhBQThGLENBQUE7SUFDOUYsNkhBQWtHLENBQUE7SUFDbEcsNkZBQWtFLENBQUE7SUFDbEUsdUZBQTRELENBQUE7SUFDNUQsNkdBQWtGLENBQUE7QUFDcEYsQ0FBQyxFQXhEVyx1QkFBdUIsS0FBdkIsdUJBQXVCLFFBd0RsQyJ9
|
||||
@ -1,4 +1,6 @@
|
||||
export declare enum E_OPENROUTER_MODEL {
|
||||
MODEL_OPENROUTER_QUASAR_ALPHA = "openrouter/quasar-alpha",
|
||||
MODEL_ALL_HANDS_OPENHANDS_LM_32B_V0_1 = "all-hands/openhands-lm-32b-v0.1",
|
||||
MODEL_MISTRAL_MINISTRAL_8B = "mistral/ministral-8b",
|
||||
MODEL_DEEPSEEK_DEEPSEEK_V3_BASE_FREE = "deepseek/deepseek-v3-base:free",
|
||||
MODEL_SCB10X_LLAMA3_1_TYPHOON2_8B_INSTRUCT = "scb10x/llama3.1-typhoon2-8b-instruct",
|
||||
@ -47,9 +49,9 @@ export declare enum E_OPENROUTER_MODEL {
|
||||
MODEL_NOUSRESEARCH_DEEPHERMES_3_LLAMA_3_8B_PREVIEW_FREE = "nousresearch/deephermes-3-llama-3-8b-preview:free",
|
||||
MODEL_OPENAI_GPT_4_5_PREVIEW = "openai/gpt-4.5-preview",
|
||||
MODEL_GOOGLE_GEMINI_2_0_FLASH_LITE_001 = "google/gemini-2.0-flash-lite-001",
|
||||
MODEL_ANTHROPIC_CLAUDE_3_7_SONNET_BETA = "anthropic/claude-3.7-sonnet:beta",
|
||||
MODEL_ANTHROPIC_CLAUDE_3_7_SONNET = "anthropic/claude-3.7-sonnet",
|
||||
MODEL_ANTHROPIC_CLAUDE_3_7_SONNET_THINKING = "anthropic/claude-3.7-sonnet:thinking",
|
||||
MODEL_ANTHROPIC_CLAUDE_3_7_SONNET_BETA = "anthropic/claude-3.7-sonnet:beta",
|
||||
MODEL_PERPLEXITY_R1_1776 = "perplexity/r1-1776",
|
||||
MODEL_MISTRALAI_MISTRAL_SABA = "mistralai/mistral-saba",
|
||||
MODEL_COGNITIVECOMPUTATIONS_DOLPHIN3_0_R1_MISTRAL_24B_FREE = "cognitivecomputations/dolphin3.0-r1-mistral-24b:free",
|
||||
@ -185,8 +187,6 @@ export declare enum E_OPENROUTER_MODEL {
|
||||
MODEL_MISTRALAI_MISTRAL_NEMO = "mistralai/mistral-nemo",
|
||||
MODEL_OPENAI_GPT_4O_MINI = "openai/gpt-4o-mini",
|
||||
MODEL_OPENAI_GPT_4O_MINI_2024_07_18 = "openai/gpt-4o-mini-2024-07-18",
|
||||
MODEL_QWEN_QWEN_2_7B_INSTRUCT_FREE = "qwen/qwen-2-7b-instruct:free",
|
||||
MODEL_QWEN_QWEN_2_7B_INSTRUCT = "qwen/qwen-2-7b-instruct",
|
||||
MODEL_GOOGLE_GEMMA_2_27B_IT = "google/gemma-2-27b-it",
|
||||
MODEL_ALPINDALE_MAGNUM_72B = "alpindale/magnum-72b",
|
||||
MODEL_GOOGLE_GEMMA_2_9B_IT_FREE = "google/gemma-2-9b-it:free",
|
||||
@ -215,7 +215,6 @@ export declare enum E_OPENROUTER_MODEL {
|
||||
MODEL_NEVERSLEEP_LLAMA_3_LUMIMAID_8B_EXTENDED = "neversleep/llama-3-lumimaid-8b:extended",
|
||||
MODEL_NEVERSLEEP_LLAMA_3_LUMIMAID_8B = "neversleep/llama-3-lumimaid-8b",
|
||||
MODEL_SAO10K_FIMBULVETR_11B_V2 = "sao10k/fimbulvetr-11b-v2",
|
||||
MODEL_META_LLAMA_LLAMA_3_8B_INSTRUCT_FREE = "meta-llama/llama-3-8b-instruct:free",
|
||||
MODEL_META_LLAMA_LLAMA_3_8B_INSTRUCT = "meta-llama/llama-3-8b-instruct",
|
||||
MODEL_META_LLAMA_LLAMA_3_70B_INSTRUCT = "meta-llama/llama-3-70b-instruct",
|
||||
MODEL_MISTRALAI_MIXTRAL_8X22B_INSTRUCT = "mistralai/mixtral-8x22b-instruct",
|
||||
@ -236,7 +235,6 @@ export declare enum E_OPENROUTER_MODEL {
|
||||
MODEL_ANTHROPIC_CLAUDE_3_SONNET = "anthropic/claude-3-sonnet",
|
||||
MODEL_COHERE_COMMAND_R_03_2024 = "cohere/command-r-03-2024",
|
||||
MODEL_MISTRALAI_MISTRAL_LARGE = "mistralai/mistral-large",
|
||||
MODEL_GOOGLE_GEMMA_7B_IT = "google/gemma-7b-it",
|
||||
MODEL_OPENAI_GPT_3_5_TURBO_0613 = "openai/gpt-3.5-turbo-0613",
|
||||
MODEL_OPENAI_GPT_4_TURBO_PREVIEW = "openai/gpt-4-turbo-preview",
|
||||
MODEL_NOUSRESEARCH_NOUS_HERMES_2_MIXTRAL_8X7B_DPO = "nousresearch/nous-hermes-2-mixtral-8x7b-dpo",
|
||||
@ -256,7 +254,6 @@ export declare enum E_OPENROUTER_MODEL {
|
||||
MODEL_ANTHROPIC_CLAUDE_2_1 = "anthropic/claude-2.1",
|
||||
MODEL_ANTHROPIC_CLAUDE_2_BETA = "anthropic/claude-2:beta",
|
||||
MODEL_ANTHROPIC_CLAUDE_2 = "anthropic/claude-2",
|
||||
MODEL_TEKNIUM_OPENHERMES_2_5_MISTRAL_7B = "teknium/openhermes-2.5-mistral-7b",
|
||||
MODEL_UNDI95_TOPPY_M_7B_FREE = "undi95/toppy-m-7b:free",
|
||||
MODEL_UNDI95_TOPPY_M_7B = "undi95/toppy-m-7b",
|
||||
MODEL_ALPINDALE_GOLIATH_120B = "alpindale/goliath-120b",
|
||||
@ -281,7 +278,6 @@ export declare enum E_OPENROUTER_MODEL {
|
||||
MODEL_UNDI95_REMM_SLERP_L2_13B = "undi95/remm-slerp-l2-13b",
|
||||
MODEL_GOOGLE_PALM_2_CHAT_BISON = "google/palm-2-chat-bison",
|
||||
MODEL_GOOGLE_PALM_2_CODECHAT_BISON = "google/palm-2-codechat-bison",
|
||||
MODEL_GRYPHE_MYTHOMAX_L2_13B_FREE = "gryphe/mythomax-l2-13b:free",
|
||||
MODEL_GRYPHE_MYTHOMAX_L2_13B = "gryphe/mythomax-l2-13b",
|
||||
MODEL_META_LLAMA_LLAMA_2_13B_CHAT = "meta-llama/llama-2-13b-chat",
|
||||
MODEL_META_LLAMA_LLAMA_2_70B_CHAT = "meta-llama/llama-2-70b-chat",
|
||||
|
||||
File diff suppressed because one or more lines are too long
File diff suppressed because one or more lines are too long
File diff suppressed because one or more lines are too long
8
packages/kbot/dist-in/zod_types.d.ts
vendored
8
packages/kbot/dist-in/zod_types.d.ts
vendored
@ -121,7 +121,6 @@ export interface IKBotOptions {
|
||||
google/gemma-3-27b-it:free | free
|
||||
google/gemma-3-4b-it | paid
|
||||
google/gemma-3-4b-it:free | free
|
||||
google/gemma-7b-it | paid
|
||||
google/learnlm-1.5-pro-experimental:free | free
|
||||
google/palm-2-chat-bison | paid
|
||||
google/palm-2-chat-bison-32k | paid
|
||||
@ -144,7 +143,6 @@ export interface IKBotOptions {
|
||||
meta-llama/llama-2-70b-chat | paid
|
||||
meta-llama/llama-3-70b-instruct | paid
|
||||
meta-llama/llama-3-8b-instruct | paid
|
||||
meta-llama/llama-3-8b-instruct:free | free
|
||||
meta-llama/llama-3.1-405b | paid
|
||||
meta-llama/llama-3.1-405b-instruct | paid
|
||||
meta-llama/llama-3.1-70b-instruct | paid
|
||||
@ -200,7 +198,6 @@ export interface IKBotOptions {
|
||||
mistralai/mistral-saba | paid
|
||||
moonshotai/moonlight-16b-a3b-instruct:free | free
|
||||
gryphe/mythomax-l2-13b | paid
|
||||
gryphe/mythomax-l2-13b:free | free
|
||||
neversleep/llama-3-lumimaid-70b | paid
|
||||
neversleep/llama-3-lumimaid-8b | paid
|
||||
neversleep/llama-3-lumimaid-8b:extended | paid
|
||||
@ -251,7 +248,7 @@ export interface IKBotOptions {
|
||||
openai/o3-mini-high | paid
|
||||
openchat/openchat-7b | paid
|
||||
openchat/openchat-7b:free | free
|
||||
teknium/openhermes-2.5-mistral-7b | paid
|
||||
all-hands/openhands-lm-32b-v0.1 | paid
|
||||
perplexity/llama-3.1-sonar-large-128k-online | paid
|
||||
perplexity/llama-3.1-sonar-small-128k-online | paid
|
||||
perplexity/r1-1776 | paid
|
||||
@ -261,9 +258,8 @@ export interface IKBotOptions {
|
||||
perplexity/sonar-reasoning | paid
|
||||
perplexity/sonar-reasoning-pro | paid
|
||||
pygmalionai/mythalion-13b | paid
|
||||
openrouter/quasar-alpha | paid
|
||||
qwen/qwen-2-72b-instruct | paid
|
||||
qwen/qwen-2-7b-instruct | paid
|
||||
qwen/qwen-2-7b-instruct:free | free
|
||||
qwen/qwen-vl-max | paid
|
||||
qwen/qwen-vl-plus | paid
|
||||
qwen/qwen-max | paid
|
||||
|
||||
File diff suppressed because one or more lines are too long
File diff suppressed because one or more lines are too long
@ -52,7 +52,7 @@ ${toolModels.map(model => ` MODEL_TOOLS_${model.id.replace(/[^a-zA-Z0-9]/g, '_'
|
||||
}
|
||||
|
||||
export const build = async () => {
|
||||
const examplesSrc = path.resolve(__dirname, '../docs_/examples.md')
|
||||
const examplesSrc = path.resolve(__dirname, '../../docs_/examples.md')
|
||||
if(exists(examplesSrc)) {
|
||||
const examples = read(examplesSrc,'string') || ''
|
||||
const examplesPath = path.resolve(__dirname, '../src/docs-internal/examples.ts')
|
||||
|
||||
18
packages/kbot/src/models/cache/openai-models.ts
vendored
18
packages/kbot/src/models/cache/openai-models.ts
vendored
@ -4,10 +4,12 @@ export enum E_OPENAI_MODEL {
|
||||
MODEL_DALL_E_3 = "dall-e-3",
|
||||
MODEL_DALL_E_2 = "dall-e-2",
|
||||
MODEL_GPT_4O_AUDIO_PREVIEW_2024_10_01 = "gpt-4o-audio-preview-2024-10-01",
|
||||
MODEL_GPT_4O_MINI_REALTIME_PREVIEW_2024_12_17 = "gpt-4o-mini-realtime-preview-2024-12-17",
|
||||
MODEL_GPT_4O_MINI_REALTIME_PREVIEW = "gpt-4o-mini-realtime-preview",
|
||||
MODEL_O1_PRO_2025_03_19 = "o1-pro-2025-03-19",
|
||||
MODEL_O1_PRO = "o1-pro",
|
||||
MODEL_GPT_4O_REALTIME_PREVIEW_2024_10_01 = "gpt-4o-realtime-preview-2024-10-01",
|
||||
MODEL_O1_2024_12_17 = "o1-2024-12-17",
|
||||
MODEL_GPT_4O_TRANSCRIBE = "gpt-4o-transcribe",
|
||||
MODEL_O1 = "o1",
|
||||
MODEL_GPT_4O_MINI_TRANSCRIBE = "gpt-4o-mini-transcribe",
|
||||
MODEL_GPT_4O_REALTIME_PREVIEW = "gpt-4o-realtime-preview",
|
||||
MODEL_BABBAGE_002 = "babbage-002",
|
||||
@ -16,19 +18,18 @@ export enum E_OPENAI_MODEL {
|
||||
MODEL_TEXT_EMBEDDING_3_LARGE = "text-embedding-3-large",
|
||||
MODEL_GPT_4 = "gpt-4",
|
||||
MODEL_TEXT_EMBEDDING_ADA_002 = "text-embedding-ada-002",
|
||||
MODEL_OMNI_MODERATION_LATEST = "omni-moderation-latest",
|
||||
MODEL_TTS_1_HD = "tts-1-hd",
|
||||
MODEL_GPT_4O_MINI_AUDIO_PREVIEW = "gpt-4o-mini-audio-preview",
|
||||
MODEL_GPT_4O_AUDIO_PREVIEW = "gpt-4o-audio-preview",
|
||||
MODEL_O1_PREVIEW_2024_09_12 = "o1-preview-2024-09-12",
|
||||
MODEL_GPT_4O_MINI_REALTIME_PREVIEW = "gpt-4o-mini-realtime-preview",
|
||||
MODEL_GPT_4O_MINI_REALTIME_PREVIEW_2024_12_17 = "gpt-4o-mini-realtime-preview-2024-12-17",
|
||||
MODEL_GPT_3_5_TURBO_INSTRUCT_0914 = "gpt-3.5-turbo-instruct-0914",
|
||||
MODEL_GPT_4O_MINI_SEARCH_PREVIEW = "gpt-4o-mini-search-preview",
|
||||
MODEL_O1 = "o1",
|
||||
MODEL_TTS_1_1106 = "tts-1-1106",
|
||||
MODEL_O1_2024_12_17 = "o1-2024-12-17",
|
||||
MODEL_DAVINCI_002 = "davinci-002",
|
||||
MODEL_GPT_3_5_TURBO_1106 = "gpt-3.5-turbo-1106",
|
||||
MODEL_O1_PRO = "o1-pro",
|
||||
MODEL_O1_PRO_2025_03_19 = "o1-pro-2025-03-19",
|
||||
MODEL_GPT_4_TURBO = "gpt-4-turbo",
|
||||
MODEL_GPT_3_5_TURBO_INSTRUCT = "gpt-3.5-turbo-instruct",
|
||||
MODEL_GPT_3_5_TURBO = "gpt-3.5-turbo",
|
||||
@ -49,7 +50,6 @@ export enum E_OPENAI_MODEL {
|
||||
MODEL_GPT_4_5_PREVIEW = "gpt-4.5-preview",
|
||||
MODEL_GPT_4_5_PREVIEW_2025_02_27 = "gpt-4.5-preview-2025-02-27",
|
||||
MODEL_GPT_4O_SEARCH_PREVIEW_2025_03_11 = "gpt-4o-search-preview-2025-03-11",
|
||||
MODEL_OMNI_MODERATION_LATEST = "omni-moderation-latest",
|
||||
MODEL_TTS_1 = "tts-1",
|
||||
MODEL_OMNI_MODERATION_2024_09_26 = "omni-moderation-2024-09-26",
|
||||
MODEL_TEXT_EMBEDDING_3_SMALL = "text-embedding-3-small",
|
||||
@ -57,9 +57,9 @@ export enum E_OPENAI_MODEL {
|
||||
MODEL_GPT_4O_MINI = "gpt-4o-mini",
|
||||
MODEL_GPT_4O_2024_08_06 = "gpt-4o-2024-08-06",
|
||||
MODEL_GPT_4O_MINI_2024_07_18 = "gpt-4o-mini-2024-07-18",
|
||||
MODEL_GPT_4_TURBO_PREVIEW = "gpt-4-turbo-preview",
|
||||
MODEL_O1_MINI = "o1-mini",
|
||||
MODEL_GPT_4O_MINI_AUDIO_PREVIEW_2024_12_17 = "gpt-4o-mini-audio-preview-2024-12-17",
|
||||
MODEL_O1_MINI_2024_09_12 = "o1-mini-2024-09-12",
|
||||
MODEL_GPT_4_0125_PREVIEW = "gpt-4-0125-preview"
|
||||
MODEL_GPT_4_0125_PREVIEW = "gpt-4-0125-preview",
|
||||
MODEL_GPT_4_TURBO_PREVIEW = "gpt-4-turbo-preview"
|
||||
}
|
||||
@ -1,4 +1,5 @@
|
||||
export enum E_OPENROUTER_MODEL_FREE {
|
||||
MODEL_FREE_OPENROUTER_QUASAR_ALPHA = "openrouter/quasar-alpha",
|
||||
MODEL_FREE_DEEPSEEK_DEEPSEEK_V3_BASE_FREE = "deepseek/deepseek-v3-base:free",
|
||||
MODEL_FREE_ALLENAI_MOLMO_7B_D_FREE = "allenai/molmo-7b-d:free",
|
||||
MODEL_FREE_BYTEDANCE_RESEARCH_UI_TARS_72B_FREE = "bytedance-research/ui-tars-72b:free",
|
||||
@ -46,14 +47,11 @@ export enum E_OPENROUTER_MODEL_FREE {
|
||||
MODEL_FREE_GOOGLE_GEMINI_FLASH_1_5_8B_EXP = "google/gemini-flash-1.5-8b-exp",
|
||||
MODEL_FREE_META_LLAMA_LLAMA_3_1_8B_INSTRUCT_FREE = "meta-llama/llama-3.1-8b-instruct:free",
|
||||
MODEL_FREE_MISTRALAI_MISTRAL_NEMO_FREE = "mistralai/mistral-nemo:free",
|
||||
MODEL_FREE_QWEN_QWEN_2_7B_INSTRUCT_FREE = "qwen/qwen-2-7b-instruct:free",
|
||||
MODEL_FREE_GOOGLE_GEMMA_2_9B_IT_FREE = "google/gemma-2-9b-it:free",
|
||||
MODEL_FREE_MISTRALAI_MISTRAL_7B_INSTRUCT_FREE = "mistralai/mistral-7b-instruct:free",
|
||||
MODEL_FREE_MICROSOFT_PHI_3_MINI_128K_INSTRUCT_FREE = "microsoft/phi-3-mini-128k-instruct:free",
|
||||
MODEL_FREE_MICROSOFT_PHI_3_MEDIUM_128K_INSTRUCT_FREE = "microsoft/phi-3-medium-128k-instruct:free",
|
||||
MODEL_FREE_META_LLAMA_LLAMA_3_8B_INSTRUCT_FREE = "meta-llama/llama-3-8b-instruct:free",
|
||||
MODEL_FREE_OPENCHAT_OPENCHAT_7B_FREE = "openchat/openchat-7b:free",
|
||||
MODEL_FREE_UNDI95_TOPPY_M_7B_FREE = "undi95/toppy-m-7b:free",
|
||||
MODEL_FREE_HUGGINGFACEH4_ZEPHYR_7B_BETA_FREE = "huggingfaceh4/zephyr-7b-beta:free",
|
||||
MODEL_FREE_GRYPHE_MYTHOMAX_L2_13B_FREE = "gryphe/mythomax-l2-13b:free"
|
||||
MODEL_FREE_HUGGINGFACEH4_ZEPHYR_7B_BETA_FREE = "huggingfaceh4/zephyr-7b-beta:free"
|
||||
}
|
||||
@ -1,4 +1,6 @@
|
||||
export enum E_OPENROUTER_MODEL {
|
||||
MODEL_OPENROUTER_QUASAR_ALPHA = "openrouter/quasar-alpha",
|
||||
MODEL_ALL_HANDS_OPENHANDS_LM_32B_V0_1 = "all-hands/openhands-lm-32b-v0.1",
|
||||
MODEL_MISTRAL_MINISTRAL_8B = "mistral/ministral-8b",
|
||||
MODEL_DEEPSEEK_DEEPSEEK_V3_BASE_FREE = "deepseek/deepseek-v3-base:free",
|
||||
MODEL_SCB10X_LLAMA3_1_TYPHOON2_8B_INSTRUCT = "scb10x/llama3.1-typhoon2-8b-instruct",
|
||||
@ -47,9 +49,9 @@ export enum E_OPENROUTER_MODEL {
|
||||
MODEL_NOUSRESEARCH_DEEPHERMES_3_LLAMA_3_8B_PREVIEW_FREE = "nousresearch/deephermes-3-llama-3-8b-preview:free",
|
||||
MODEL_OPENAI_GPT_4_5_PREVIEW = "openai/gpt-4.5-preview",
|
||||
MODEL_GOOGLE_GEMINI_2_0_FLASH_LITE_001 = "google/gemini-2.0-flash-lite-001",
|
||||
MODEL_ANTHROPIC_CLAUDE_3_7_SONNET_BETA = "anthropic/claude-3.7-sonnet:beta",
|
||||
MODEL_ANTHROPIC_CLAUDE_3_7_SONNET = "anthropic/claude-3.7-sonnet",
|
||||
MODEL_ANTHROPIC_CLAUDE_3_7_SONNET_THINKING = "anthropic/claude-3.7-sonnet:thinking",
|
||||
MODEL_ANTHROPIC_CLAUDE_3_7_SONNET_BETA = "anthropic/claude-3.7-sonnet:beta",
|
||||
MODEL_PERPLEXITY_R1_1776 = "perplexity/r1-1776",
|
||||
MODEL_MISTRALAI_MISTRAL_SABA = "mistralai/mistral-saba",
|
||||
MODEL_COGNITIVECOMPUTATIONS_DOLPHIN3_0_R1_MISTRAL_24B_FREE = "cognitivecomputations/dolphin3.0-r1-mistral-24b:free",
|
||||
@ -185,8 +187,6 @@ export enum E_OPENROUTER_MODEL {
|
||||
MODEL_MISTRALAI_MISTRAL_NEMO = "mistralai/mistral-nemo",
|
||||
MODEL_OPENAI_GPT_4O_MINI = "openai/gpt-4o-mini",
|
||||
MODEL_OPENAI_GPT_4O_MINI_2024_07_18 = "openai/gpt-4o-mini-2024-07-18",
|
||||
MODEL_QWEN_QWEN_2_7B_INSTRUCT_FREE = "qwen/qwen-2-7b-instruct:free",
|
||||
MODEL_QWEN_QWEN_2_7B_INSTRUCT = "qwen/qwen-2-7b-instruct",
|
||||
MODEL_GOOGLE_GEMMA_2_27B_IT = "google/gemma-2-27b-it",
|
||||
MODEL_ALPINDALE_MAGNUM_72B = "alpindale/magnum-72b",
|
||||
MODEL_GOOGLE_GEMMA_2_9B_IT_FREE = "google/gemma-2-9b-it:free",
|
||||
@ -215,7 +215,6 @@ export enum E_OPENROUTER_MODEL {
|
||||
MODEL_NEVERSLEEP_LLAMA_3_LUMIMAID_8B_EXTENDED = "neversleep/llama-3-lumimaid-8b:extended",
|
||||
MODEL_NEVERSLEEP_LLAMA_3_LUMIMAID_8B = "neversleep/llama-3-lumimaid-8b",
|
||||
MODEL_SAO10K_FIMBULVETR_11B_V2 = "sao10k/fimbulvetr-11b-v2",
|
||||
MODEL_META_LLAMA_LLAMA_3_8B_INSTRUCT_FREE = "meta-llama/llama-3-8b-instruct:free",
|
||||
MODEL_META_LLAMA_LLAMA_3_8B_INSTRUCT = "meta-llama/llama-3-8b-instruct",
|
||||
MODEL_META_LLAMA_LLAMA_3_70B_INSTRUCT = "meta-llama/llama-3-70b-instruct",
|
||||
MODEL_MISTRALAI_MIXTRAL_8X22B_INSTRUCT = "mistralai/mixtral-8x22b-instruct",
|
||||
@ -236,7 +235,6 @@ export enum E_OPENROUTER_MODEL {
|
||||
MODEL_ANTHROPIC_CLAUDE_3_SONNET = "anthropic/claude-3-sonnet",
|
||||
MODEL_COHERE_COMMAND_R_03_2024 = "cohere/command-r-03-2024",
|
||||
MODEL_MISTRALAI_MISTRAL_LARGE = "mistralai/mistral-large",
|
||||
MODEL_GOOGLE_GEMMA_7B_IT = "google/gemma-7b-it",
|
||||
MODEL_OPENAI_GPT_3_5_TURBO_0613 = "openai/gpt-3.5-turbo-0613",
|
||||
MODEL_OPENAI_GPT_4_TURBO_PREVIEW = "openai/gpt-4-turbo-preview",
|
||||
MODEL_NOUSRESEARCH_NOUS_HERMES_2_MIXTRAL_8X7B_DPO = "nousresearch/nous-hermes-2-mixtral-8x7b-dpo",
|
||||
@ -256,7 +254,6 @@ export enum E_OPENROUTER_MODEL {
|
||||
MODEL_ANTHROPIC_CLAUDE_2_1 = "anthropic/claude-2.1",
|
||||
MODEL_ANTHROPIC_CLAUDE_2_BETA = "anthropic/claude-2:beta",
|
||||
MODEL_ANTHROPIC_CLAUDE_2 = "anthropic/claude-2",
|
||||
MODEL_TEKNIUM_OPENHERMES_2_5_MISTRAL_7B = "teknium/openhermes-2.5-mistral-7b",
|
||||
MODEL_UNDI95_TOPPY_M_7B_FREE = "undi95/toppy-m-7b:free",
|
||||
MODEL_UNDI95_TOPPY_M_7B = "undi95/toppy-m-7b",
|
||||
MODEL_ALPINDALE_GOLIATH_120B = "alpindale/goliath-120b",
|
||||
@ -281,7 +278,6 @@ export enum E_OPENROUTER_MODEL {
|
||||
MODEL_UNDI95_REMM_SLERP_L2_13B = "undi95/remm-slerp-l2-13b",
|
||||
MODEL_GOOGLE_PALM_2_CHAT_BISON = "google/palm-2-chat-bison",
|
||||
MODEL_GOOGLE_PALM_2_CODECHAT_BISON = "google/palm-2-codechat-bison",
|
||||
MODEL_GRYPHE_MYTHOMAX_L2_13B_FREE = "gryphe/mythomax-l2-13b:free",
|
||||
MODEL_GRYPHE_MYTHOMAX_L2_13B = "gryphe/mythomax-l2-13b",
|
||||
MODEL_META_LLAMA_LLAMA_2_13B_CHAT = "meta-llama/llama-2-13b-chat",
|
||||
MODEL_META_LLAMA_LLAMA_2_70B_CHAT = "meta-llama/llama-2-70b-chat",
|
||||
|
||||
@ -121,7 +121,6 @@ export interface IKBotOptions {
|
||||
google/gemma-3-27b-it:free | free
|
||||
google/gemma-3-4b-it | paid
|
||||
google/gemma-3-4b-it:free | free
|
||||
google/gemma-7b-it | paid
|
||||
google/learnlm-1.5-pro-experimental:free | free
|
||||
google/palm-2-chat-bison | paid
|
||||
google/palm-2-chat-bison-32k | paid
|
||||
@ -144,7 +143,6 @@ export interface IKBotOptions {
|
||||
meta-llama/llama-2-70b-chat | paid
|
||||
meta-llama/llama-3-70b-instruct | paid
|
||||
meta-llama/llama-3-8b-instruct | paid
|
||||
meta-llama/llama-3-8b-instruct:free | free
|
||||
meta-llama/llama-3.1-405b | paid
|
||||
meta-llama/llama-3.1-405b-instruct | paid
|
||||
meta-llama/llama-3.1-70b-instruct | paid
|
||||
@ -200,7 +198,6 @@ export interface IKBotOptions {
|
||||
mistralai/mistral-saba | paid
|
||||
moonshotai/moonlight-16b-a3b-instruct:free | free
|
||||
gryphe/mythomax-l2-13b | paid
|
||||
gryphe/mythomax-l2-13b:free | free
|
||||
neversleep/llama-3-lumimaid-70b | paid
|
||||
neversleep/llama-3-lumimaid-8b | paid
|
||||
neversleep/llama-3-lumimaid-8b:extended | paid
|
||||
@ -251,7 +248,7 @@ export interface IKBotOptions {
|
||||
openai/o3-mini-high | paid
|
||||
openchat/openchat-7b | paid
|
||||
openchat/openchat-7b:free | free
|
||||
teknium/openhermes-2.5-mistral-7b | paid
|
||||
all-hands/openhands-lm-32b-v0.1 | paid
|
||||
perplexity/llama-3.1-sonar-large-128k-online | paid
|
||||
perplexity/llama-3.1-sonar-small-128k-online | paid
|
||||
perplexity/r1-1776 | paid
|
||||
@ -261,9 +258,8 @@ export interface IKBotOptions {
|
||||
perplexity/sonar-reasoning | paid
|
||||
perplexity/sonar-reasoning-pro | paid
|
||||
pygmalionai/mythalion-13b | paid
|
||||
openrouter/quasar-alpha | paid
|
||||
qwen/qwen-2-72b-instruct | paid
|
||||
qwen/qwen-2-7b-instruct | paid
|
||||
qwen/qwen-2-7b-instruct:free | free
|
||||
qwen/qwen-vl-max | paid
|
||||
qwen/qwen-vl-plus | paid
|
||||
qwen/qwen-max | paid
|
||||
|
||||
Loading…
Reference in New Issue
Block a user