maintainence love:)

This commit is contained in:
lovebird 2025-04-04 14:33:56 +02:00
parent 4726ebc0f0
commit 214eed04f5
19 changed files with 243 additions and 404 deletions

File diff suppressed because one or more lines are too long

View File

@ -1,5 +1,5 @@
{
"timestamp": 1743511914173,
"timestamp": 1743769889560,
"models": [
{
"id": "gpt-4o-realtime-preview-2024-12-17",
@ -32,15 +32,15 @@
"owned_by": "system"
},
{
"id": "gpt-4o-mini-realtime-preview-2024-12-17",
"id": "o1-pro-2025-03-19",
"object": "model",
"created": 1734112601,
"created": 1742251504,
"owned_by": "system"
},
{
"id": "gpt-4o-mini-realtime-preview",
"id": "o1-pro",
"object": "model",
"created": 1734387380,
"created": 1742251791,
"owned_by": "system"
},
{
@ -49,12 +49,24 @@
"created": 1727131766,
"owned_by": "system"
},
{
"id": "o1-2024-12-17",
"object": "model",
"created": 1734326976,
"owned_by": "system"
},
{
"id": "gpt-4o-transcribe",
"object": "model",
"created": 1742068463,
"owned_by": "system"
},
{
"id": "o1",
"object": "model",
"created": 1734375816,
"owned_by": "system"
},
{
"id": "gpt-4o-mini-transcribe",
"object": "model",
@ -103,6 +115,12 @@
"created": 1671217299,
"owned_by": "openai-internal"
},
{
"id": "omni-moderation-latest",
"object": "model",
"created": 1731689265,
"owned_by": "system"
},
{
"id": "tts-1-hd",
"object": "model",
@ -127,6 +145,18 @@
"created": 1725648865,
"owned_by": "system"
},
{
"id": "gpt-4o-mini-realtime-preview",
"object": "model",
"created": 1734387380,
"owned_by": "system"
},
{
"id": "gpt-4o-mini-realtime-preview-2024-12-17",
"object": "model",
"created": 1734112601,
"owned_by": "system"
},
{
"id": "gpt-3.5-turbo-instruct-0914",
"object": "model",
@ -139,24 +169,12 @@
"created": 1741391161,
"owned_by": "system"
},
{
"id": "o1",
"object": "model",
"created": 1734375816,
"owned_by": "system"
},
{
"id": "tts-1-1106",
"object": "model",
"created": 1699053241,
"owned_by": "system"
},
{
"id": "o1-2024-12-17",
"object": "model",
"created": 1734326976,
"owned_by": "system"
},
{
"id": "davinci-002",
"object": "model",
@ -169,18 +187,6 @@
"created": 1698959748,
"owned_by": "system"
},
{
"id": "o1-pro",
"object": "model",
"created": 1742251791,
"owned_by": "system"
},
{
"id": "o1-pro-2025-03-19",
"object": "model",
"created": 1742251504,
"owned_by": "system"
},
{
"id": "gpt-4-turbo",
"object": "model",
@ -301,12 +307,6 @@
"created": 1741388170,
"owned_by": "system"
},
{
"id": "omni-moderation-latest",
"object": "model",
"created": 1731689265,
"owned_by": "system"
},
{
"id": "tts-1",
"object": "model",
@ -349,12 +349,6 @@
"created": 1721172717,
"owned_by": "system"
},
{
"id": "gpt-4-turbo-preview",
"object": "model",
"created": 1706037777,
"owned_by": "system"
},
{
"id": "o1-mini",
"object": "model",
@ -378,6 +372,12 @@
"object": "model",
"created": 1706037612,
"owned_by": "system"
},
{
"id": "gpt-4-turbo-preview",
"object": "model",
"created": 1706037777,
"owned_by": "system"
}
]
}

View File

@ -1,6 +1,75 @@
{
"timestamp": 1743511914343,
"timestamp": 1743769889892,
"models": [
{
"id": "openrouter/quasar-alpha",
"name": "Quasar Alpha",
"created": 1743626809,
"description": "This is a cloaked model provided to the community to gather feedback. Its a powerful, all-purpose model supporting long-context tasks, including code generation. All prompts and completions for this model are logged by the provider as well as OpenRouter.",
"context_length": 1000000,
"architecture": {
"modality": "text+image->text",
"input_modalities": [
"image",
"text"
],
"output_modalities": [
"text"
],
"tokenizer": "Other",
"instruct_type": null
},
"pricing": {
"prompt": "0",
"completion": "0",
"request": "0",
"image": "0",
"web_search": "0",
"internal_reasoning": "0",
"input_cache_read": "0",
"input_cache_write": "0"
},
"top_provider": {
"context_length": 1000000,
"max_completion_tokens": 32000,
"is_moderated": true
},
"per_request_limits": null
},
{
"id": "all-hands/openhands-lm-32b-v0.1",
"name": "OpenHands LM 32B V0.1",
"created": 1743613013,
"description": "OpenHands LM v0.1 is a 32B open-source coding model fine-tuned from Qwen2.5-Coder-32B-Instruct using reinforcement learning techniques outlined in SWE-Gym. It is optimized for autonomous software development agents and achieves strong performance on SWE-Bench Verified, with a 37.2% resolve rate. The model supports a 128K token context window, making it well-suited for long-horizon code reasoning and large codebase tasks.\n\nOpenHands LM is designed for local deployment and runs on consumer-grade GPUs such as a single 3090. It enables fully offline agent workflows without dependency on proprietary APIs. This release is intended as a research preview, and future updates aim to improve generalizability, reduce repetition, and offer smaller variants.",
"context_length": 16384,
"architecture": {
"modality": "text->text",
"input_modalities": [
"text"
],
"output_modalities": [
"text"
],
"tokenizer": "Other",
"instruct_type": null
},
"pricing": {
"prompt": "0.0000026",
"completion": "0.0000034",
"request": "0",
"image": "0",
"web_search": "0",
"internal_reasoning": "0",
"input_cache_read": "0",
"input_cache_write": "0"
},
"top_provider": {
"context_length": 16384,
"max_completion_tokens": 4095,
"is_moderated": false
},
"per_request_limits": null
},
{
"id": "mistral/ministral-8b",
"name": "Mistral: Ministral 8b",
@ -1465,8 +1534,8 @@
"instruct_type": "qwq"
},
"pricing": {
"prompt": "0.00000012",
"completion": "0.00000018",
"prompt": "0.00000015",
"completion": "0.0000002",
"request": "0",
"image": "0",
"web_search": "0",
@ -1653,41 +1722,6 @@
},
"per_request_limits": null
},
{
"id": "anthropic/claude-3.7-sonnet:beta",
"name": "Anthropic: Claude 3.7 Sonnet (self-moderated)",
"created": 1740422110,
"description": "Claude 3.7 Sonnet is an advanced large language model with improved reasoning, coding, and problem-solving capabilities. It introduces a hybrid reasoning approach, allowing users to choose between rapid responses and extended, step-by-step processing for complex tasks. The model demonstrates notable improvements in coding, particularly in front-end development and full-stack updates, and excels in agentic workflows, where it can autonomously navigate multi-step processes. \n\nClaude 3.7 Sonnet maintains performance parity with its predecessor in standard mode while offering an extended reasoning mode for enhanced accuracy in math, coding, and instruction-following tasks.\n\nRead more at the [blog post here](https://www.anthropic.com/news/claude-3-7-sonnet)",
"context_length": 200000,
"architecture": {
"modality": "text+image->text",
"input_modalities": [
"text",
"image"
],
"output_modalities": [
"text"
],
"tokenizer": "Claude",
"instruct_type": null
},
"pricing": {
"prompt": "0.000003",
"completion": "0.000015",
"request": "0",
"image": "0.0048",
"web_search": "0",
"internal_reasoning": "0",
"input_cache_read": "0",
"input_cache_write": "0"
},
"top_provider": {
"context_length": 200000,
"max_completion_tokens": 128000,
"is_moderated": false
},
"per_request_limits": null
},
{
"id": "anthropic/claude-3.7-sonnet",
"name": "Anthropic: Claude 3.7 Sonnet",
@ -1718,8 +1752,8 @@
},
"top_provider": {
"context_length": 200000,
"max_completion_tokens": 128000,
"is_moderated": true
"max_completion_tokens": 64000,
"is_moderated": false
},
"per_request_limits": null
},
@ -1751,10 +1785,45 @@
"input_cache_read": "0",
"input_cache_write": "0"
},
"top_provider": {
"context_length": 200000,
"max_completion_tokens": 64000,
"is_moderated": false
},
"per_request_limits": null
},
{
"id": "anthropic/claude-3.7-sonnet:beta",
"name": "Anthropic: Claude 3.7 Sonnet (self-moderated)",
"created": 1740422110,
"description": "Claude 3.7 Sonnet is an advanced large language model with improved reasoning, coding, and problem-solving capabilities. It introduces a hybrid reasoning approach, allowing users to choose between rapid responses and extended, step-by-step processing for complex tasks. The model demonstrates notable improvements in coding, particularly in front-end development and full-stack updates, and excels in agentic workflows, where it can autonomously navigate multi-step processes. \n\nClaude 3.7 Sonnet maintains performance parity with its predecessor in standard mode while offering an extended reasoning mode for enhanced accuracy in math, coding, and instruction-following tasks.\n\nRead more at the [blog post here](https://www.anthropic.com/news/claude-3-7-sonnet)",
"context_length": 200000,
"architecture": {
"modality": "text+image->text",
"input_modalities": [
"text",
"image"
],
"output_modalities": [
"text"
],
"tokenizer": "Claude",
"instruct_type": null
},
"pricing": {
"prompt": "0.000003",
"completion": "0.000015",
"request": "0",
"image": "0.0048",
"web_search": "0",
"internal_reasoning": "0",
"input_cache_read": "0",
"input_cache_write": "0"
},
"top_provider": {
"context_length": 200000,
"max_completion_tokens": 128000,
"is_moderated": true
"is_moderated": false
},
"per_request_limits": null
},
@ -2312,7 +2381,7 @@
"name": "Qwen: Qwen2.5 VL 72B Instruct",
"created": 1738410311,
"description": "Qwen2.5-VL is proficient in recognizing common objects such as flowers, birds, fish, and insects. It is also highly capable of analyzing texts, charts, icons, graphics, and layouts within images.",
"context_length": 128000,
"context_length": 32000,
"architecture": {
"modality": "text+image->text",
"input_modalities": [
@ -2336,8 +2405,8 @@
"input_cache_write": "0"
},
"top_provider": {
"context_length": 128000,
"max_completion_tokens": 128000,
"context_length": 32000,
"max_completion_tokens": 32000,
"is_moderated": false
},
"per_request_limits": null
@ -2823,7 +2892,7 @@
"name": "DeepSeek: R1 Distill Llama 70B (free)",
"created": 1737663169,
"description": "DeepSeek R1 Distill Llama 70B is a distilled large language model based on [Llama-3.3-70B-Instruct](/meta-llama/llama-3.3-70b-instruct), using outputs from [DeepSeek R1](/deepseek/deepseek-r1). The model combines advanced distillation techniques to achieve high performance across multiple benchmarks, including:\n\n- AIME 2024 pass@1: 70.0\n- MATH-500 pass@1: 94.5\n- CodeForces Rating: 1633\n\nThe model leverages fine-tuning from DeepSeek R1's outputs, enabling competitive performance comparable to larger frontier models.",
"context_length": 8192,
"context_length": 128000,
"architecture": {
"modality": "text->text",
"input_modalities": [
@ -2846,8 +2915,8 @@
"input_cache_write": "0"
},
"top_provider": {
"context_length": 8192,
"max_completion_tokens": 4096,
"context_length": 128000,
"max_completion_tokens": null,
"is_moderated": false
},
"per_request_limits": null
@ -5088,7 +5157,7 @@
"name": "Meta: Llama 3.2 11B Vision Instruct",
"created": 1727222400,
"description": "Llama 3.2 11B Vision is a multimodal model with 11 billion parameters, designed to handle tasks combining visual and textual data. It excels in tasks such as image captioning and visual question answering, bridging the gap between language generation and visual reasoning. Pre-trained on a massive dataset of image-text pairs, it performs well in complex, high-accuracy image analysis.\n\nIts ability to integrate visual understanding with language processing makes it an ideal solution for industries requiring comprehensive visual-linguistic AI applications, such as content creation, AI-driven customer service, and research.\n\nClick here for the [original model card](https://github.com/meta-llama/llama-models/blob/main/models/llama3_2/MODEL_CARD_VISION.md).\n\nUsage of this model is subject to [Meta's Acceptable Use Policy](https://www.llama.com/llama3/use-policy/).",
"context_length": 16384,
"context_length": 131072,
"architecture": {
"modality": "text+image->text",
"input_modalities": [
@ -5105,15 +5174,15 @@
"prompt": "0.000000055",
"completion": "0.000000055",
"request": "0",
"image": "0",
"image": "0.00007948",
"web_search": "0",
"internal_reasoning": "0",
"input_cache_read": "0",
"input_cache_write": "0"
},
"top_provider": {
"context_length": 16384,
"max_completion_tokens": 16384,
"context_length": 131072,
"max_completion_tokens": 8192,
"is_moderated": false
},
"per_request_limits": null
@ -5444,8 +5513,8 @@
"instruct_type": null
},
"pricing": {
"prompt": "0.000002375",
"completion": "0.0000095",
"prompt": "0.0000025",
"completion": "0.00001",
"request": "0",
"image": "0",
"web_search": "0",
@ -5478,8 +5547,8 @@
"instruct_type": null
},
"pricing": {
"prompt": "0.0000001425",
"completion": "0.00000057",
"prompt": "0.00000015",
"completion": "0.0000006",
"request": "0",
"image": "0",
"web_search": "0",
@ -5856,8 +5925,8 @@
"instruct_type": "llama3"
},
"pricing": {
"prompt": "0.00000003",
"completion": "0.00000006",
"prompt": "0.00000005",
"completion": "0.00000005",
"request": "0",
"image": "0",
"web_search": "0",
@ -5867,7 +5936,7 @@
},
"top_provider": {
"context_length": 8192,
"max_completion_tokens": 8192,
"max_completion_tokens": null,
"is_moderated": false
},
"per_request_limits": null
@ -6385,74 +6454,6 @@
},
"per_request_limits": null
},
{
"id": "qwen/qwen-2-7b-instruct:free",
"name": "Qwen 2 7B Instruct (free)",
"created": 1721088000,
"description": "Qwen2 7B is a transformer-based model that excels in language understanding, multilingual capabilities, coding, mathematics, and reasoning.\n\nIt features SwiGLU activation, attention QKV bias, and group query attention. It is pretrained on extensive data with supervised finetuning and direct preference optimization.\n\nFor more details, see this [blog post](https://qwenlm.github.io/blog/qwen2/) and [GitHub repo](https://github.com/QwenLM/Qwen2).\n\nUsage of this model is subject to [Tongyi Qianwen LICENSE AGREEMENT](https://huggingface.co/Qwen/Qwen1.5-110B-Chat/blob/main/LICENSE).",
"context_length": 8192,
"architecture": {
"modality": "text->text",
"input_modalities": [
"text"
],
"output_modalities": [
"text"
],
"tokenizer": "Qwen",
"instruct_type": "chatml"
},
"pricing": {
"prompt": "0",
"completion": "0",
"request": "0",
"image": "0",
"web_search": "0",
"internal_reasoning": "0",
"input_cache_read": "0",
"input_cache_write": "0"
},
"top_provider": {
"context_length": 8192,
"max_completion_tokens": 4096,
"is_moderated": false
},
"per_request_limits": null
},
{
"id": "qwen/qwen-2-7b-instruct",
"name": "Qwen 2 7B Instruct",
"created": 1721088000,
"description": "Qwen2 7B is a transformer-based model that excels in language understanding, multilingual capabilities, coding, mathematics, and reasoning.\n\nIt features SwiGLU activation, attention QKV bias, and group query attention. It is pretrained on extensive data with supervised finetuning and direct preference optimization.\n\nFor more details, see this [blog post](https://qwenlm.github.io/blog/qwen2/) and [GitHub repo](https://github.com/QwenLM/Qwen2).\n\nUsage of this model is subject to [Tongyi Qianwen LICENSE AGREEMENT](https://huggingface.co/Qwen/Qwen1.5-110B-Chat/blob/main/LICENSE).",
"context_length": 32768,
"architecture": {
"modality": "text->text",
"input_modalities": [
"text"
],
"output_modalities": [
"text"
],
"tokenizer": "Qwen",
"instruct_type": "chatml"
},
"pricing": {
"prompt": "0.000000054",
"completion": "0.000000054",
"request": "0",
"image": "0",
"web_search": "0",
"internal_reasoning": "0",
"input_cache_read": "0",
"input_cache_write": "0"
},
"top_provider": {
"context_length": 32768,
"max_completion_tokens": null,
"is_moderated": false
},
"per_request_limits": null
},
{
"id": "google/gemma-2-27b-it",
"name": "Google: Gemma 2 27B",
@ -6471,8 +6472,8 @@
"instruct_type": "gemma"
},
"pricing": {
"prompt": "0.00000027",
"completion": "0.00000027",
"prompt": "0.0000008",
"completion": "0.0000008",
"request": "0",
"image": "0",
"web_search": "0",
@ -6482,7 +6483,7 @@
},
"top_provider": {
"context_length": 8192,
"max_completion_tokens": 8192,
"max_completion_tokens": 2048,
"is_moderated": false
},
"per_request_limits": null
@ -6573,8 +6574,8 @@
"instruct_type": "gemma"
},
"pricing": {
"prompt": "0.00000003",
"completion": "0.00000006",
"prompt": "0.00000007",
"completion": "0.00000007",
"request": "0",
"image": "0",
"web_search": "0",
@ -6584,7 +6585,7 @@
},
"top_provider": {
"context_length": 8192,
"max_completion_tokens": 8192,
"max_completion_tokens": null,
"is_moderated": false
},
"per_request_limits": null
@ -6745,8 +6746,8 @@
"instruct_type": "llama3"
},
"pricing": {
"prompt": "0.0000007",
"completion": "0.0000008",
"prompt": "0.00000148",
"completion": "0.00000148",
"request": "0",
"image": "0",
"web_search": "0",
@ -7411,40 +7412,6 @@
},
"per_request_limits": null
},
{
"id": "meta-llama/llama-3-8b-instruct:free",
"name": "Meta: Llama 3 8B Instruct (free)",
"created": 1713398400,
"description": "Meta's latest class of model (Llama 3) launched with a variety of sizes & flavors. This 8B instruct-tuned version was optimized for high quality dialogue usecases.\n\nIt has demonstrated strong performance compared to leading closed-source models in human evaluations.\n\nTo read more about the model release, [click here](https://ai.meta.com/blog/meta-llama-3/). Usage of this model is subject to [Meta's Acceptable Use Policy](https://llama.meta.com/llama3/use-policy/).",
"context_length": 8192,
"architecture": {
"modality": "text->text",
"input_modalities": [
"text"
],
"output_modalities": [
"text"
],
"tokenizer": "Llama3",
"instruct_type": "llama3"
},
"pricing": {
"prompt": "0",
"completion": "0",
"request": "0",
"image": "0",
"web_search": "0",
"internal_reasoning": "0",
"input_cache_read": "0",
"input_cache_write": "0"
},
"top_provider": {
"context_length": 8192,
"max_completion_tokens": 4096,
"is_moderated": false
},
"per_request_limits": null
},
{
"id": "meta-llama/llama-3-8b-instruct",
"name": "Meta: Llama 3 8B Instruct",
@ -7703,8 +7670,8 @@
"instruct_type": null
},
"pricing": {
"prompt": "0.00000285",
"completion": "0.00001425",
"prompt": "0.000003",
"completion": "0.000015",
"request": "0",
"image": "0",
"web_search": "0",
@ -7737,8 +7704,8 @@
"instruct_type": null
},
"pricing": {
"prompt": "0.00000285",
"completion": "0.00001425",
"prompt": "0.000003",
"completion": "0.000015",
"request": "0",
"image": "0",
"web_search": "0",
@ -7805,8 +7772,8 @@
"instruct_type": null
},
"pricing": {
"prompt": "0.00000095",
"completion": "0.0000019",
"prompt": "0.000001",
"completion": "0.000002",
"request": "0",
"image": "0",
"web_search": "0",
@ -7839,8 +7806,8 @@
"instruct_type": null
},
"pricing": {
"prompt": "0.000000475",
"completion": "0.000001425",
"prompt": "0.0000005",
"completion": "0.0000015",
"request": "0",
"image": "0",
"web_search": "0",
@ -8083,8 +8050,8 @@
"instruct_type": null
},
"pricing": {
"prompt": "0.000000475",
"completion": "0.000001425",
"prompt": "0.0000005",
"completion": "0.0000015",
"request": "0",
"image": "0",
"web_search": "0",
@ -8133,40 +8100,6 @@
},
"per_request_limits": null
},
{
"id": "google/gemma-7b-it",
"name": "Google: Gemma 7B",
"created": 1708560000,
"description": "Gemma by Google is an advanced, open-source language model family, leveraging the latest in decoder-only, text-to-text technology. It offers English language capabilities across text generation tasks like question answering, summarization, and reasoning. The Gemma 7B variant is comparable in performance to leading open source models.\n\nUsage of Gemma is subject to Google's [Gemma Terms of Use](https://ai.google.dev/gemma/terms).",
"context_length": 8192,
"architecture": {
"modality": "text->text",
"input_modalities": [
"text"
],
"output_modalities": [
"text"
],
"tokenizer": "Gemini",
"instruct_type": "gemma"
},
"pricing": {
"prompt": "0.00000015",
"completion": "0.00000015",
"request": "0",
"image": "0",
"web_search": "0",
"internal_reasoning": "0",
"input_cache_read": "0",
"input_cache_write": "0"
},
"top_provider": {
"context_length": 8192,
"max_completion_tokens": null,
"is_moderated": false
},
"per_request_limits": null
},
{
"id": "openai/gpt-3.5-turbo-0613",
"name": "OpenAI: GPT-3.5 Turbo (older v0613)",
@ -8628,8 +8561,8 @@
"instruct_type": "openchat"
},
"pricing": {
"prompt": "0.000000055",
"completion": "0.000000055",
"prompt": "0.00000007",
"completion": "0.00000007",
"request": "0",
"image": "0",
"web_search": "0",
@ -8639,7 +8572,7 @@
},
"top_provider": {
"context_length": 8192,
"max_completion_tokens": 8192,
"max_completion_tokens": null,
"is_moderated": false
},
"per_request_limits": null
@ -8814,40 +8747,6 @@
},
"per_request_limits": null
},
{
"id": "teknium/openhermes-2.5-mistral-7b",
"name": "OpenHermes 2.5 Mistral 7B",
"created": 1700438400,
"description": "A continuation of [OpenHermes 2 model](/models/teknium/openhermes-2-mistral-7b), trained on additional code datasets.\nPotentially the most interesting finding from training on a good ratio (est. of around 7-14% of the total dataset) of code instruction was that it has boosted several non-code benchmarks, including TruthfulQA, AGIEval, and GPT4All suite. It did however reduce BigBench benchmark score, but the net gain overall is significant.",
"context_length": 4096,
"architecture": {
"modality": "text->text",
"input_modalities": [
"text"
],
"output_modalities": [
"text"
],
"tokenizer": "Mistral",
"instruct_type": "chatml"
},
"pricing": {
"prompt": "0.00000017",
"completion": "0.00000017",
"request": "0",
"image": "0",
"web_search": "0",
"internal_reasoning": "0",
"input_cache_read": "0",
"input_cache_write": "0"
},
"top_provider": {
"context_length": 4096,
"max_completion_tokens": 4096,
"is_moderated": false
},
"per_request_limits": null
},
{
"id": "undi95/toppy-m-7b:free",
"name": "Toppy M 7B (free)",
@ -9664,40 +9563,6 @@
},
"per_request_limits": null
},
{
"id": "gryphe/mythomax-l2-13b:free",
"name": "MythoMax 13B (free)",
"created": 1688256000,
"description": "One of the highest performing and most popular fine-tunes of Llama 2 13B, with rich descriptions and roleplay. #merge",
"context_length": 4096,
"architecture": {
"modality": "text->text",
"input_modalities": [
"text"
],
"output_modalities": [
"text"
],
"tokenizer": "Llama2",
"instruct_type": "alpaca"
},
"pricing": {
"prompt": "0",
"completion": "0",
"request": "0",
"image": "0",
"web_search": "0",
"internal_reasoning": "0",
"input_cache_read": "0",
"input_cache_write": "0"
},
"top_provider": {
"context_length": 4096,
"max_completion_tokens": 2048,
"is_moderated": false
},
"per_request_limits": null
},
{
"id": "gryphe/mythomax-l2-13b",
"name": "MythoMax 13B",

View File

@ -4,10 +4,12 @@ export declare enum E_OPENAI_MODEL {
MODEL_DALL_E_3 = "dall-e-3",
MODEL_DALL_E_2 = "dall-e-2",
MODEL_GPT_4O_AUDIO_PREVIEW_2024_10_01 = "gpt-4o-audio-preview-2024-10-01",
MODEL_GPT_4O_MINI_REALTIME_PREVIEW_2024_12_17 = "gpt-4o-mini-realtime-preview-2024-12-17",
MODEL_GPT_4O_MINI_REALTIME_PREVIEW = "gpt-4o-mini-realtime-preview",
MODEL_O1_PRO_2025_03_19 = "o1-pro-2025-03-19",
MODEL_O1_PRO = "o1-pro",
MODEL_GPT_4O_REALTIME_PREVIEW_2024_10_01 = "gpt-4o-realtime-preview-2024-10-01",
MODEL_O1_2024_12_17 = "o1-2024-12-17",
MODEL_GPT_4O_TRANSCRIBE = "gpt-4o-transcribe",
MODEL_O1 = "o1",
MODEL_GPT_4O_MINI_TRANSCRIBE = "gpt-4o-mini-transcribe",
MODEL_GPT_4O_REALTIME_PREVIEW = "gpt-4o-realtime-preview",
MODEL_BABBAGE_002 = "babbage-002",
@ -16,19 +18,18 @@ export declare enum E_OPENAI_MODEL {
MODEL_TEXT_EMBEDDING_3_LARGE = "text-embedding-3-large",
MODEL_GPT_4 = "gpt-4",
MODEL_TEXT_EMBEDDING_ADA_002 = "text-embedding-ada-002",
MODEL_OMNI_MODERATION_LATEST = "omni-moderation-latest",
MODEL_TTS_1_HD = "tts-1-hd",
MODEL_GPT_4O_MINI_AUDIO_PREVIEW = "gpt-4o-mini-audio-preview",
MODEL_GPT_4O_AUDIO_PREVIEW = "gpt-4o-audio-preview",
MODEL_O1_PREVIEW_2024_09_12 = "o1-preview-2024-09-12",
MODEL_GPT_4O_MINI_REALTIME_PREVIEW = "gpt-4o-mini-realtime-preview",
MODEL_GPT_4O_MINI_REALTIME_PREVIEW_2024_12_17 = "gpt-4o-mini-realtime-preview-2024-12-17",
MODEL_GPT_3_5_TURBO_INSTRUCT_0914 = "gpt-3.5-turbo-instruct-0914",
MODEL_GPT_4O_MINI_SEARCH_PREVIEW = "gpt-4o-mini-search-preview",
MODEL_O1 = "o1",
MODEL_TTS_1_1106 = "tts-1-1106",
MODEL_O1_2024_12_17 = "o1-2024-12-17",
MODEL_DAVINCI_002 = "davinci-002",
MODEL_GPT_3_5_TURBO_1106 = "gpt-3.5-turbo-1106",
MODEL_O1_PRO = "o1-pro",
MODEL_O1_PRO_2025_03_19 = "o1-pro-2025-03-19",
MODEL_GPT_4_TURBO = "gpt-4-turbo",
MODEL_GPT_3_5_TURBO_INSTRUCT = "gpt-3.5-turbo-instruct",
MODEL_GPT_3_5_TURBO = "gpt-3.5-turbo",
@ -49,7 +50,6 @@ export declare enum E_OPENAI_MODEL {
MODEL_GPT_4_5_PREVIEW = "gpt-4.5-preview",
MODEL_GPT_4_5_PREVIEW_2025_02_27 = "gpt-4.5-preview-2025-02-27",
MODEL_GPT_4O_SEARCH_PREVIEW_2025_03_11 = "gpt-4o-search-preview-2025-03-11",
MODEL_OMNI_MODERATION_LATEST = "omni-moderation-latest",
MODEL_TTS_1 = "tts-1",
MODEL_OMNI_MODERATION_2024_09_26 = "omni-moderation-2024-09-26",
MODEL_TEXT_EMBEDDING_3_SMALL = "text-embedding-3-small",
@ -57,9 +57,9 @@ export declare enum E_OPENAI_MODEL {
MODEL_GPT_4O_MINI = "gpt-4o-mini",
MODEL_GPT_4O_2024_08_06 = "gpt-4o-2024-08-06",
MODEL_GPT_4O_MINI_2024_07_18 = "gpt-4o-mini-2024-07-18",
MODEL_GPT_4_TURBO_PREVIEW = "gpt-4-turbo-preview",
MODEL_O1_MINI = "o1-mini",
MODEL_GPT_4O_MINI_AUDIO_PREVIEW_2024_12_17 = "gpt-4o-mini-audio-preview-2024-12-17",
MODEL_O1_MINI_2024_09_12 = "o1-mini-2024-09-12",
MODEL_GPT_4_0125_PREVIEW = "gpt-4-0125-preview"
MODEL_GPT_4_0125_PREVIEW = "gpt-4-0125-preview",
MODEL_GPT_4_TURBO_PREVIEW = "gpt-4-turbo-preview"
}

View File

@ -5,10 +5,12 @@ export var E_OPENAI_MODEL;
E_OPENAI_MODEL["MODEL_DALL_E_3"] = "dall-e-3";
E_OPENAI_MODEL["MODEL_DALL_E_2"] = "dall-e-2";
E_OPENAI_MODEL["MODEL_GPT_4O_AUDIO_PREVIEW_2024_10_01"] = "gpt-4o-audio-preview-2024-10-01";
E_OPENAI_MODEL["MODEL_GPT_4O_MINI_REALTIME_PREVIEW_2024_12_17"] = "gpt-4o-mini-realtime-preview-2024-12-17";
E_OPENAI_MODEL["MODEL_GPT_4O_MINI_REALTIME_PREVIEW"] = "gpt-4o-mini-realtime-preview";
E_OPENAI_MODEL["MODEL_O1_PRO_2025_03_19"] = "o1-pro-2025-03-19";
E_OPENAI_MODEL["MODEL_O1_PRO"] = "o1-pro";
E_OPENAI_MODEL["MODEL_GPT_4O_REALTIME_PREVIEW_2024_10_01"] = "gpt-4o-realtime-preview-2024-10-01";
E_OPENAI_MODEL["MODEL_O1_2024_12_17"] = "o1-2024-12-17";
E_OPENAI_MODEL["MODEL_GPT_4O_TRANSCRIBE"] = "gpt-4o-transcribe";
E_OPENAI_MODEL["MODEL_O1"] = "o1";
E_OPENAI_MODEL["MODEL_GPT_4O_MINI_TRANSCRIBE"] = "gpt-4o-mini-transcribe";
E_OPENAI_MODEL["MODEL_GPT_4O_REALTIME_PREVIEW"] = "gpt-4o-realtime-preview";
E_OPENAI_MODEL["MODEL_BABBAGE_002"] = "babbage-002";
@ -17,19 +19,18 @@ export var E_OPENAI_MODEL;
E_OPENAI_MODEL["MODEL_TEXT_EMBEDDING_3_LARGE"] = "text-embedding-3-large";
E_OPENAI_MODEL["MODEL_GPT_4"] = "gpt-4";
E_OPENAI_MODEL["MODEL_TEXT_EMBEDDING_ADA_002"] = "text-embedding-ada-002";
E_OPENAI_MODEL["MODEL_OMNI_MODERATION_LATEST"] = "omni-moderation-latest";
E_OPENAI_MODEL["MODEL_TTS_1_HD"] = "tts-1-hd";
E_OPENAI_MODEL["MODEL_GPT_4O_MINI_AUDIO_PREVIEW"] = "gpt-4o-mini-audio-preview";
E_OPENAI_MODEL["MODEL_GPT_4O_AUDIO_PREVIEW"] = "gpt-4o-audio-preview";
E_OPENAI_MODEL["MODEL_O1_PREVIEW_2024_09_12"] = "o1-preview-2024-09-12";
E_OPENAI_MODEL["MODEL_GPT_4O_MINI_REALTIME_PREVIEW"] = "gpt-4o-mini-realtime-preview";
E_OPENAI_MODEL["MODEL_GPT_4O_MINI_REALTIME_PREVIEW_2024_12_17"] = "gpt-4o-mini-realtime-preview-2024-12-17";
E_OPENAI_MODEL["MODEL_GPT_3_5_TURBO_INSTRUCT_0914"] = "gpt-3.5-turbo-instruct-0914";
E_OPENAI_MODEL["MODEL_GPT_4O_MINI_SEARCH_PREVIEW"] = "gpt-4o-mini-search-preview";
E_OPENAI_MODEL["MODEL_O1"] = "o1";
E_OPENAI_MODEL["MODEL_TTS_1_1106"] = "tts-1-1106";
E_OPENAI_MODEL["MODEL_O1_2024_12_17"] = "o1-2024-12-17";
E_OPENAI_MODEL["MODEL_DAVINCI_002"] = "davinci-002";
E_OPENAI_MODEL["MODEL_GPT_3_5_TURBO_1106"] = "gpt-3.5-turbo-1106";
E_OPENAI_MODEL["MODEL_O1_PRO"] = "o1-pro";
E_OPENAI_MODEL["MODEL_O1_PRO_2025_03_19"] = "o1-pro-2025-03-19";
E_OPENAI_MODEL["MODEL_GPT_4_TURBO"] = "gpt-4-turbo";
E_OPENAI_MODEL["MODEL_GPT_3_5_TURBO_INSTRUCT"] = "gpt-3.5-turbo-instruct";
E_OPENAI_MODEL["MODEL_GPT_3_5_TURBO"] = "gpt-3.5-turbo";
@ -50,7 +51,6 @@ export var E_OPENAI_MODEL;
E_OPENAI_MODEL["MODEL_GPT_4_5_PREVIEW"] = "gpt-4.5-preview";
E_OPENAI_MODEL["MODEL_GPT_4_5_PREVIEW_2025_02_27"] = "gpt-4.5-preview-2025-02-27";
E_OPENAI_MODEL["MODEL_GPT_4O_SEARCH_PREVIEW_2025_03_11"] = "gpt-4o-search-preview-2025-03-11";
E_OPENAI_MODEL["MODEL_OMNI_MODERATION_LATEST"] = "omni-moderation-latest";
E_OPENAI_MODEL["MODEL_TTS_1"] = "tts-1";
E_OPENAI_MODEL["MODEL_OMNI_MODERATION_2024_09_26"] = "omni-moderation-2024-09-26";
E_OPENAI_MODEL["MODEL_TEXT_EMBEDDING_3_SMALL"] = "text-embedding-3-small";
@ -58,10 +58,10 @@ export var E_OPENAI_MODEL;
E_OPENAI_MODEL["MODEL_GPT_4O_MINI"] = "gpt-4o-mini";
E_OPENAI_MODEL["MODEL_GPT_4O_2024_08_06"] = "gpt-4o-2024-08-06";
E_OPENAI_MODEL["MODEL_GPT_4O_MINI_2024_07_18"] = "gpt-4o-mini-2024-07-18";
E_OPENAI_MODEL["MODEL_GPT_4_TURBO_PREVIEW"] = "gpt-4-turbo-preview";
E_OPENAI_MODEL["MODEL_O1_MINI"] = "o1-mini";
E_OPENAI_MODEL["MODEL_GPT_4O_MINI_AUDIO_PREVIEW_2024_12_17"] = "gpt-4o-mini-audio-preview-2024-12-17";
E_OPENAI_MODEL["MODEL_O1_MINI_2024_09_12"] = "o1-mini-2024-09-12";
E_OPENAI_MODEL["MODEL_GPT_4_0125_PREVIEW"] = "gpt-4-0125-preview";
E_OPENAI_MODEL["MODEL_GPT_4_TURBO_PREVIEW"] = "gpt-4-turbo-preview";
})(E_OPENAI_MODEL || (E_OPENAI_MODEL = {}));
//# sourceMappingURL=data:application/json;base64,eyJ2ZXJzaW9uIjozLCJmaWxlIjoib3BlbmFpLW1vZGVscy5qcyIsInNvdXJjZVJvb3QiOiIiLCJzb3VyY2VzIjpbIi4uLy4uLy4uL3NyYy9tb2RlbHMvY2FjaGUvb3BlbmFpLW1vZGVscy50cyJdLCJuYW1lcyI6W10sIm1hcHBpbmdzIjoiQUFBQSxNQUFNLENBQU4sSUFBWSxjQWdFWDtBQWhFRCxXQUFZLGNBQWM7SUFDeEIsaUdBQStFLENBQUE7SUFDL0UsMkZBQXlFLENBQUE7SUFDekUsNkNBQTJCLENBQUE7SUFDM0IsNkNBQTJCLENBQUE7SUFDM0IsMkZBQXlFLENBQUE7SUFDekUsMkdBQXlGLENBQUE7SUFDekYscUZBQW1FLENBQUE7SUFDbkUsaUdBQStFLENBQUE7SUFDL0UsK0RBQTZDLENBQUE7SUFDN0MseUVBQXVELENBQUE7SUFDdkQsMkVBQXlELENBQUE7SUFDekQsbURBQWlDLENBQUE7SUFDakMsMkRBQXlDLENBQUE7SUFDekMsdURBQXFDLENBQUE7SUFDckMseUVBQXVELENBQUE7SUFDdkQsdUNBQXFCLENBQUE7SUFDckIseUVBQXVELENBQUE7SUFDdkQsNkNBQTJCLENBQUE7SUFDM0IsK0VBQTZELENBQUE7SUFDN0QscUVBQW1ELENBQUE7SUFDbkQsdUVBQXFELENBQUE7SUFDckQsbUZBQWlFLENBQUE7SUFDakUsaUZBQStELENBQUE7SUFDL0QsaUNBQWUsQ0FBQTtJQUNmLGlEQUErQixDQUFBO0lBQy9CLHVEQUFxQyxDQUFBO0lBQ3JDLG1EQUFpQyxDQUFBO0lBQ2pDLGlFQUErQyxDQUFBO0lBQy9DLHlDQUF1QixDQUFBO0lBQ3ZCLCtEQUE2QyxDQUFBO0lBQzdDLG1EQUFpQyxDQUFBO0lBQ2pDLHlFQUF1RCxDQUFBO0lBQ3ZELHVEQUFxQyxDQUFBO0lBQ3JDLCtEQUE2QyxDQUFBO0lBQzdDLHVHQUFxRixDQUFBO0lBQ3JGLCtEQUE2QyxDQUFBO0lBQzdDLCtDQUE2QixDQUFBO0lBQzdCLGlFQUErQyxDQUFBO0lBQy9DLCtEQUE2QyxDQUFBO0lBQzdDLCtEQUE2QyxDQUFBO0lBQzdDLHlFQUF1RCxDQUFBO0lBQ3ZELGlFQUErQyxDQUFBO0lBQy9DLGlEQUErQixDQUFBO0lBQy9CLGlEQUErQixDQUFBO0lBQy9CLHVFQUFxRCxDQUFBO0lBQ3JELGlFQUErQyxDQUFBO0lBQy9DLDJDQUF5QixDQUFBO0lBQ3pCLDJEQUF5QyxDQUFBO0lBQ3pDLGlGQUErRCxDQUFBO0lBQy9ELDZGQUEyRSxDQUFBO0lBQzNFLHlFQUF1RCxDQUFBO0lBQ3ZELHVDQUFxQixDQUFBO0lBQ3JCLGlGQUErRCxDQUFBO0lBQy9ELHlFQUF1RCxDQUFBO0lBQ3ZELHlDQUF1QixDQUFBO0lBQ3ZCLG1EQUFpQyxDQUFBO0lBQ2pDLCtEQUE2QyxDQUFBO0lBQzdDLHlFQUF1RCxDQUFBO0lBQ3ZELG1FQUFpRCxDQUFBO0lBQ2pELDJDQUF5QixDQUFBO0lBQ3pCLHFHQUFtRixDQUFBO0lBQ25GLGlFQUErQyxDQUFBO0lBQy9DLGlFQUErQyxDQUFBO0FBQ2pELENBQUMsRUFoRVcsY0FBYyxLQUFkLGNBQWMsUUFnRXpCIn0=
//# sourceMappingURL=data:application/json;base64,eyJ2ZXJzaW9uIjozLCJmaWxlIjoib3BlbmFpLW1vZGVscy5qcyIsInNvdXJjZVJvb3QiOiIiLCJzb3VyY2VzIjpbIi4uLy4uLy4uL3NyYy9tb2RlbHMvY2FjaGUvb3BlbmFpLW1vZGVscy50cyJdLCJuYW1lcyI6W10sIm1hcHBpbmdzIjoiQUFBQSxNQUFNLENBQU4sSUFBWSxjQWdFWDtBQWhFRCxXQUFZLGNBQWM7SUFDeEIsaUdBQStFLENBQUE7SUFDL0UsMkZBQXlFLENBQUE7SUFDekUsNkNBQTJCLENBQUE7SUFDM0IsNkNBQTJCLENBQUE7SUFDM0IsMkZBQXlFLENBQUE7SUFDekUsK0RBQTZDLENBQUE7SUFDN0MseUNBQXVCLENBQUE7SUFDdkIsaUdBQStFLENBQUE7SUFDL0UsdURBQXFDLENBQUE7SUFDckMsK0RBQTZDLENBQUE7SUFDN0MsaUNBQWUsQ0FBQTtJQUNmLHlFQUF1RCxDQUFBO0lBQ3ZELDJFQUF5RCxDQUFBO0lBQ3pELG1EQUFpQyxDQUFBO0lBQ2pDLDJEQUF5QyxDQUFBO0lBQ3pDLHVEQUFxQyxDQUFBO0lBQ3JDLHlFQUF1RCxDQUFBO0lBQ3ZELHVDQUFxQixDQUFBO0lBQ3JCLHlFQUF1RCxDQUFBO0lBQ3ZELHlFQUF1RCxDQUFBO0lBQ3ZELDZDQUEyQixDQUFBO0lBQzNCLCtFQUE2RCxDQUFBO0lBQzdELHFFQUFtRCxDQUFBO0lBQ25ELHVFQUFxRCxDQUFBO0lBQ3JELHFGQUFtRSxDQUFBO0lBQ25FLDJHQUF5RixDQUFBO0lBQ3pGLG1GQUFpRSxDQUFBO0lBQ2pFLGlGQUErRCxDQUFBO0lBQy9ELGlEQUErQixDQUFBO0lBQy9CLG1EQUFpQyxDQUFBO0lBQ2pDLGlFQUErQyxDQUFBO0lBQy9DLG1EQUFpQyxDQUFBO0lBQ2pDLHlFQUF1RCxDQUFBO0lBQ3ZELHVEQUFxQyxDQUFBO0lBQ3JDLCtEQUE2QyxDQUFBO0lBQzdDLHVHQUFxRixDQUFBO0lBQ3JGLCtEQUE2QyxDQUFBO0lBQzdDLCtDQUE2QixDQUFBO0lBQzdCLGlFQUErQyxDQUFBO0lBQy9DLCtEQUE2QyxDQUFBO0lBQzdDLCtEQUE2QyxDQUFBO0lBQzdDLHlFQUF1RCxDQUFBO0lBQ3ZELGlFQUErQyxDQUFBO0lBQy9DLGlEQUErQixDQUFBO0lBQy9CLGlEQUErQixDQUFBO0lBQy9CLHVFQUFxRCxDQUFBO0lBQ3JELGlFQUErQyxDQUFBO0lBQy9DLDJDQUF5QixDQUFBO0lBQ3pCLDJEQUF5QyxDQUFBO0lBQ3pDLGlGQUErRCxDQUFBO0lBQy9ELDZGQUEyRSxDQUFBO0lBQzNFLHVDQUFxQixDQUFBO0lBQ3JCLGlGQUErRCxDQUFBO0lBQy9ELHlFQUF1RCxDQUFBO0lBQ3ZELHlDQUF1QixDQUFBO0lBQ3ZCLG1EQUFpQyxDQUFBO0lBQ2pDLCtEQUE2QyxDQUFBO0lBQzdDLHlFQUF1RCxDQUFBO0lBQ3ZELDJDQUF5QixDQUFBO0lBQ3pCLHFHQUFtRixDQUFBO0lBQ25GLGlFQUErQyxDQUFBO0lBQy9DLGlFQUErQyxDQUFBO0lBQy9DLG1FQUFpRCxDQUFBO0FBQ25ELENBQUMsRUFoRVcsY0FBYyxLQUFkLGNBQWMsUUFnRXpCIn0=

View File

@ -1,4 +1,5 @@
export declare enum E_OPENROUTER_MODEL_FREE {
MODEL_FREE_OPENROUTER_QUASAR_ALPHA = "openrouter/quasar-alpha",
MODEL_FREE_DEEPSEEK_DEEPSEEK_V3_BASE_FREE = "deepseek/deepseek-v3-base:free",
MODEL_FREE_ALLENAI_MOLMO_7B_D_FREE = "allenai/molmo-7b-d:free",
MODEL_FREE_BYTEDANCE_RESEARCH_UI_TARS_72B_FREE = "bytedance-research/ui-tars-72b:free",
@ -46,14 +47,11 @@ export declare enum E_OPENROUTER_MODEL_FREE {
MODEL_FREE_GOOGLE_GEMINI_FLASH_1_5_8B_EXP = "google/gemini-flash-1.5-8b-exp",
MODEL_FREE_META_LLAMA_LLAMA_3_1_8B_INSTRUCT_FREE = "meta-llama/llama-3.1-8b-instruct:free",
MODEL_FREE_MISTRALAI_MISTRAL_NEMO_FREE = "mistralai/mistral-nemo:free",
MODEL_FREE_QWEN_QWEN_2_7B_INSTRUCT_FREE = "qwen/qwen-2-7b-instruct:free",
MODEL_FREE_GOOGLE_GEMMA_2_9B_IT_FREE = "google/gemma-2-9b-it:free",
MODEL_FREE_MISTRALAI_MISTRAL_7B_INSTRUCT_FREE = "mistralai/mistral-7b-instruct:free",
MODEL_FREE_MICROSOFT_PHI_3_MINI_128K_INSTRUCT_FREE = "microsoft/phi-3-mini-128k-instruct:free",
MODEL_FREE_MICROSOFT_PHI_3_MEDIUM_128K_INSTRUCT_FREE = "microsoft/phi-3-medium-128k-instruct:free",
MODEL_FREE_META_LLAMA_LLAMA_3_8B_INSTRUCT_FREE = "meta-llama/llama-3-8b-instruct:free",
MODEL_FREE_OPENCHAT_OPENCHAT_7B_FREE = "openchat/openchat-7b:free",
MODEL_FREE_UNDI95_TOPPY_M_7B_FREE = "undi95/toppy-m-7b:free",
MODEL_FREE_HUGGINGFACEH4_ZEPHYR_7B_BETA_FREE = "huggingfaceh4/zephyr-7b-beta:free",
MODEL_FREE_GRYPHE_MYTHOMAX_L2_13B_FREE = "gryphe/mythomax-l2-13b:free"
MODEL_FREE_HUGGINGFACEH4_ZEPHYR_7B_BETA_FREE = "huggingfaceh4/zephyr-7b-beta:free"
}

View File

@ -1,5 +1,6 @@
export var E_OPENROUTER_MODEL_FREE;
(function (E_OPENROUTER_MODEL_FREE) {
E_OPENROUTER_MODEL_FREE["MODEL_FREE_OPENROUTER_QUASAR_ALPHA"] = "openrouter/quasar-alpha";
E_OPENROUTER_MODEL_FREE["MODEL_FREE_DEEPSEEK_DEEPSEEK_V3_BASE_FREE"] = "deepseek/deepseek-v3-base:free";
E_OPENROUTER_MODEL_FREE["MODEL_FREE_ALLENAI_MOLMO_7B_D_FREE"] = "allenai/molmo-7b-d:free";
E_OPENROUTER_MODEL_FREE["MODEL_FREE_BYTEDANCE_RESEARCH_UI_TARS_72B_FREE"] = "bytedance-research/ui-tars-72b:free";
@ -47,15 +48,12 @@ export var E_OPENROUTER_MODEL_FREE;
E_OPENROUTER_MODEL_FREE["MODEL_FREE_GOOGLE_GEMINI_FLASH_1_5_8B_EXP"] = "google/gemini-flash-1.5-8b-exp";
E_OPENROUTER_MODEL_FREE["MODEL_FREE_META_LLAMA_LLAMA_3_1_8B_INSTRUCT_FREE"] = "meta-llama/llama-3.1-8b-instruct:free";
E_OPENROUTER_MODEL_FREE["MODEL_FREE_MISTRALAI_MISTRAL_NEMO_FREE"] = "mistralai/mistral-nemo:free";
E_OPENROUTER_MODEL_FREE["MODEL_FREE_QWEN_QWEN_2_7B_INSTRUCT_FREE"] = "qwen/qwen-2-7b-instruct:free";
E_OPENROUTER_MODEL_FREE["MODEL_FREE_GOOGLE_GEMMA_2_9B_IT_FREE"] = "google/gemma-2-9b-it:free";
E_OPENROUTER_MODEL_FREE["MODEL_FREE_MISTRALAI_MISTRAL_7B_INSTRUCT_FREE"] = "mistralai/mistral-7b-instruct:free";
E_OPENROUTER_MODEL_FREE["MODEL_FREE_MICROSOFT_PHI_3_MINI_128K_INSTRUCT_FREE"] = "microsoft/phi-3-mini-128k-instruct:free";
E_OPENROUTER_MODEL_FREE["MODEL_FREE_MICROSOFT_PHI_3_MEDIUM_128K_INSTRUCT_FREE"] = "microsoft/phi-3-medium-128k-instruct:free";
E_OPENROUTER_MODEL_FREE["MODEL_FREE_META_LLAMA_LLAMA_3_8B_INSTRUCT_FREE"] = "meta-llama/llama-3-8b-instruct:free";
E_OPENROUTER_MODEL_FREE["MODEL_FREE_OPENCHAT_OPENCHAT_7B_FREE"] = "openchat/openchat-7b:free";
E_OPENROUTER_MODEL_FREE["MODEL_FREE_UNDI95_TOPPY_M_7B_FREE"] = "undi95/toppy-m-7b:free";
E_OPENROUTER_MODEL_FREE["MODEL_FREE_HUGGINGFACEH4_ZEPHYR_7B_BETA_FREE"] = "huggingfaceh4/zephyr-7b-beta:free";
E_OPENROUTER_MODEL_FREE["MODEL_FREE_GRYPHE_MYTHOMAX_L2_13B_FREE"] = "gryphe/mythomax-l2-13b:free";
})(E_OPENROUTER_MODEL_FREE || (E_OPENROUTER_MODEL_FREE = {}));
//# sourceMappingURL=data:application/json;base64,eyJ2ZXJzaW9uIjozLCJmaWxlIjoib3BlbnJvdXRlci1tb2RlbHMtZnJlZS5qcyIsInNvdXJjZVJvb3QiOiIiLCJzb3VyY2VzIjpbIi4uLy4uLy4uL3NyYy9tb2RlbHMvY2FjaGUvb3BlbnJvdXRlci1tb2RlbHMtZnJlZS50cyJdLCJuYW1lcyI6W10sIm1hcHBpbmdzIjoiQUFBQSxNQUFNLENBQU4sSUFBWSx1QkEwRFg7QUExREQsV0FBWSx1QkFBdUI7SUFDakMsdUdBQTRFLENBQUE7SUFDNUUseUZBQThELENBQUE7SUFDOUQsaUhBQXNGLENBQUE7SUFDdEYsMkdBQWdGLENBQUE7SUFDaEYsbUhBQXdGLENBQUE7SUFDeEYsNkdBQWtGLENBQUE7SUFDbEYsaUhBQXNGLENBQUE7SUFDdEYsaUdBQXNFLENBQUE7SUFDdEUscUlBQTBHLENBQUE7SUFDMUcsbUdBQXdFLENBQUE7SUFDeEUscUdBQTBFLENBQUE7SUFDMUUsNkZBQWtFLENBQUE7SUFDbEUsNkZBQWtFLENBQUE7SUFDbEUsK0ZBQW9FLENBQUE7SUFDcEUsMkZBQWdFLENBQUE7SUFDaEUsK0ZBQW9FLENBQUE7SUFDcEUsdUdBQTRFLENBQUE7SUFDNUUsNkVBQWtELENBQUE7SUFDbEQsK0hBQW9HLENBQUE7SUFDcEcsNklBQWtILENBQUE7SUFDbEgsbUpBQXdILENBQUE7SUFDeEgsNklBQWtILENBQUE7SUFDbEgsbUhBQXdGLENBQUE7SUFDeEYsNkdBQWtGLENBQUE7SUFDbEYsdUlBQTRHLENBQUE7SUFDNUcsK0hBQW9HLENBQUE7SUFDcEcsK0hBQW9HLENBQUE7SUFDcEcsaUlBQXNHLENBQUE7SUFDdEcsNkhBQWtHLENBQUE7SUFDbEcsNkZBQWtFLENBQUE7SUFDbEUsNkhBQWtHLENBQUE7SUFDbEcsaUdBQXNFLENBQUE7SUFDdEUsdUlBQTRHLENBQUE7SUFDNUcsMkdBQWdGLENBQUE7SUFDaEYsdUhBQTRGLENBQUE7SUFDNUYsNkZBQWtFLENBQUE7SUFDbEUsMkhBQWdHLENBQUE7SUFDaEcscUhBQTBGLENBQUE7SUFDMUYsaUlBQXNHLENBQUE7SUFDdEcscUhBQTBGLENBQUE7SUFDMUYscUhBQTBGLENBQUE7SUFDMUYscUlBQTBHLENBQUE7SUFDMUcseUdBQThFLENBQUE7SUFDOUUsNkdBQWtGLENBQUE7SUFDbEYsdUdBQTRFLENBQUE7SUFDNUUscUhBQTBGLENBQUE7SUFDMUYsaUdBQXNFLENBQUE7SUFDdEUsbUdBQXdFLENBQUE7SUFDeEUsNkZBQWtFLENBQUE7SUFDbEUsK0dBQW9GLENBQUE7SUFDcEYseUhBQThGLENBQUE7SUFDOUYsNkhBQWtHLENBQUE7SUFDbEcsaUhBQXNGLENBQUE7SUFDdEYsNkZBQWtFLENBQUE7SUFDbEUsdUZBQTRELENBQUE7SUFDNUQsNkdBQWtGLENBQUE7SUFDbEYsaUdBQXNFLENBQUE7QUFDeEUsQ0FBQyxFQTFEVyx1QkFBdUIsS0FBdkIsdUJBQXVCLFFBMERsQyJ9
//# sourceMappingURL=data:application/json;base64,eyJ2ZXJzaW9uIjozLCJmaWxlIjoib3BlbnJvdXRlci1tb2RlbHMtZnJlZS5qcyIsInNvdXJjZVJvb3QiOiIiLCJzb3VyY2VzIjpbIi4uLy4uLy4uL3NyYy9tb2RlbHMvY2FjaGUvb3BlbnJvdXRlci1tb2RlbHMtZnJlZS50cyJdLCJuYW1lcyI6W10sIm1hcHBpbmdzIjoiQUFBQSxNQUFNLENBQU4sSUFBWSx1QkF3RFg7QUF4REQsV0FBWSx1QkFBdUI7SUFDakMseUZBQThELENBQUE7SUFDOUQsdUdBQTRFLENBQUE7SUFDNUUseUZBQThELENBQUE7SUFDOUQsaUhBQXNGLENBQUE7SUFDdEYsMkdBQWdGLENBQUE7SUFDaEYsbUhBQXdGLENBQUE7SUFDeEYsNkdBQWtGLENBQUE7SUFDbEYsaUhBQXNGLENBQUE7SUFDdEYsaUdBQXNFLENBQUE7SUFDdEUscUlBQTBHLENBQUE7SUFDMUcsbUdBQXdFLENBQUE7SUFDeEUscUdBQTBFLENBQUE7SUFDMUUsNkZBQWtFLENBQUE7SUFDbEUsNkZBQWtFLENBQUE7SUFDbEUsK0ZBQW9FLENBQUE7SUFDcEUsMkZBQWdFLENBQUE7SUFDaEUsK0ZBQW9FLENBQUE7SUFDcEUsdUdBQTRFLENBQUE7SUFDNUUsNkVBQWtELENBQUE7SUFDbEQsK0hBQW9HLENBQUE7SUFDcEcsNklBQWtILENBQUE7SUFDbEgsbUpBQXdILENBQUE7SUFDeEgsNklBQWtILENBQUE7SUFDbEgsbUhBQXdGLENBQUE7SUFDeEYsNkdBQWtGLENBQUE7SUFDbEYsdUlBQTRHLENBQUE7SUFDNUcsK0hBQW9HLENBQUE7SUFDcEcsK0hBQW9HLENBQUE7SUFDcEcsaUlBQXNHLENBQUE7SUFDdEcsNkhBQWtHLENBQUE7SUFDbEcsNkZBQWtFLENBQUE7SUFDbEUsNkhBQWtHLENBQUE7SUFDbEcsaUdBQXNFLENBQUE7SUFDdEUsdUlBQTRHLENBQUE7SUFDNUcsMkdBQWdGLENBQUE7SUFDaEYsdUhBQTRGLENBQUE7SUFDNUYsNkZBQWtFLENBQUE7SUFDbEUsMkhBQWdHLENBQUE7SUFDaEcscUhBQTBGLENBQUE7SUFDMUYsaUlBQXNHLENBQUE7SUFDdEcscUhBQTBGLENBQUE7SUFDMUYscUhBQTBGLENBQUE7SUFDMUYscUlBQTBHLENBQUE7SUFDMUcseUdBQThFLENBQUE7SUFDOUUsNkdBQWtGLENBQUE7SUFDbEYsdUdBQTRFLENBQUE7SUFDNUUscUhBQTBGLENBQUE7SUFDMUYsaUdBQXNFLENBQUE7SUFDdEUsNkZBQWtFLENBQUE7SUFDbEUsK0dBQW9GLENBQUE7SUFDcEYseUhBQThGLENBQUE7SUFDOUYsNkhBQWtHLENBQUE7SUFDbEcsNkZBQWtFLENBQUE7SUFDbEUsdUZBQTRELENBQUE7SUFDNUQsNkdBQWtGLENBQUE7QUFDcEYsQ0FBQyxFQXhEVyx1QkFBdUIsS0FBdkIsdUJBQXVCLFFBd0RsQyJ9

View File

@ -1,4 +1,6 @@
export declare enum E_OPENROUTER_MODEL {
MODEL_OPENROUTER_QUASAR_ALPHA = "openrouter/quasar-alpha",
MODEL_ALL_HANDS_OPENHANDS_LM_32B_V0_1 = "all-hands/openhands-lm-32b-v0.1",
MODEL_MISTRAL_MINISTRAL_8B = "mistral/ministral-8b",
MODEL_DEEPSEEK_DEEPSEEK_V3_BASE_FREE = "deepseek/deepseek-v3-base:free",
MODEL_SCB10X_LLAMA3_1_TYPHOON2_8B_INSTRUCT = "scb10x/llama3.1-typhoon2-8b-instruct",
@ -47,9 +49,9 @@ export declare enum E_OPENROUTER_MODEL {
MODEL_NOUSRESEARCH_DEEPHERMES_3_LLAMA_3_8B_PREVIEW_FREE = "nousresearch/deephermes-3-llama-3-8b-preview:free",
MODEL_OPENAI_GPT_4_5_PREVIEW = "openai/gpt-4.5-preview",
MODEL_GOOGLE_GEMINI_2_0_FLASH_LITE_001 = "google/gemini-2.0-flash-lite-001",
MODEL_ANTHROPIC_CLAUDE_3_7_SONNET_BETA = "anthropic/claude-3.7-sonnet:beta",
MODEL_ANTHROPIC_CLAUDE_3_7_SONNET = "anthropic/claude-3.7-sonnet",
MODEL_ANTHROPIC_CLAUDE_3_7_SONNET_THINKING = "anthropic/claude-3.7-sonnet:thinking",
MODEL_ANTHROPIC_CLAUDE_3_7_SONNET_BETA = "anthropic/claude-3.7-sonnet:beta",
MODEL_PERPLEXITY_R1_1776 = "perplexity/r1-1776",
MODEL_MISTRALAI_MISTRAL_SABA = "mistralai/mistral-saba",
MODEL_COGNITIVECOMPUTATIONS_DOLPHIN3_0_R1_MISTRAL_24B_FREE = "cognitivecomputations/dolphin3.0-r1-mistral-24b:free",
@ -185,8 +187,6 @@ export declare enum E_OPENROUTER_MODEL {
MODEL_MISTRALAI_MISTRAL_NEMO = "mistralai/mistral-nemo",
MODEL_OPENAI_GPT_4O_MINI = "openai/gpt-4o-mini",
MODEL_OPENAI_GPT_4O_MINI_2024_07_18 = "openai/gpt-4o-mini-2024-07-18",
MODEL_QWEN_QWEN_2_7B_INSTRUCT_FREE = "qwen/qwen-2-7b-instruct:free",
MODEL_QWEN_QWEN_2_7B_INSTRUCT = "qwen/qwen-2-7b-instruct",
MODEL_GOOGLE_GEMMA_2_27B_IT = "google/gemma-2-27b-it",
MODEL_ALPINDALE_MAGNUM_72B = "alpindale/magnum-72b",
MODEL_GOOGLE_GEMMA_2_9B_IT_FREE = "google/gemma-2-9b-it:free",
@ -215,7 +215,6 @@ export declare enum E_OPENROUTER_MODEL {
MODEL_NEVERSLEEP_LLAMA_3_LUMIMAID_8B_EXTENDED = "neversleep/llama-3-lumimaid-8b:extended",
MODEL_NEVERSLEEP_LLAMA_3_LUMIMAID_8B = "neversleep/llama-3-lumimaid-8b",
MODEL_SAO10K_FIMBULVETR_11B_V2 = "sao10k/fimbulvetr-11b-v2",
MODEL_META_LLAMA_LLAMA_3_8B_INSTRUCT_FREE = "meta-llama/llama-3-8b-instruct:free",
MODEL_META_LLAMA_LLAMA_3_8B_INSTRUCT = "meta-llama/llama-3-8b-instruct",
MODEL_META_LLAMA_LLAMA_3_70B_INSTRUCT = "meta-llama/llama-3-70b-instruct",
MODEL_MISTRALAI_MIXTRAL_8X22B_INSTRUCT = "mistralai/mixtral-8x22b-instruct",
@ -236,7 +235,6 @@ export declare enum E_OPENROUTER_MODEL {
MODEL_ANTHROPIC_CLAUDE_3_SONNET = "anthropic/claude-3-sonnet",
MODEL_COHERE_COMMAND_R_03_2024 = "cohere/command-r-03-2024",
MODEL_MISTRALAI_MISTRAL_LARGE = "mistralai/mistral-large",
MODEL_GOOGLE_GEMMA_7B_IT = "google/gemma-7b-it",
MODEL_OPENAI_GPT_3_5_TURBO_0613 = "openai/gpt-3.5-turbo-0613",
MODEL_OPENAI_GPT_4_TURBO_PREVIEW = "openai/gpt-4-turbo-preview",
MODEL_NOUSRESEARCH_NOUS_HERMES_2_MIXTRAL_8X7B_DPO = "nousresearch/nous-hermes-2-mixtral-8x7b-dpo",
@ -256,7 +254,6 @@ export declare enum E_OPENROUTER_MODEL {
MODEL_ANTHROPIC_CLAUDE_2_1 = "anthropic/claude-2.1",
MODEL_ANTHROPIC_CLAUDE_2_BETA = "anthropic/claude-2:beta",
MODEL_ANTHROPIC_CLAUDE_2 = "anthropic/claude-2",
MODEL_TEKNIUM_OPENHERMES_2_5_MISTRAL_7B = "teknium/openhermes-2.5-mistral-7b",
MODEL_UNDI95_TOPPY_M_7B_FREE = "undi95/toppy-m-7b:free",
MODEL_UNDI95_TOPPY_M_7B = "undi95/toppy-m-7b",
MODEL_ALPINDALE_GOLIATH_120B = "alpindale/goliath-120b",
@ -281,7 +278,6 @@ export declare enum E_OPENROUTER_MODEL {
MODEL_UNDI95_REMM_SLERP_L2_13B = "undi95/remm-slerp-l2-13b",
MODEL_GOOGLE_PALM_2_CHAT_BISON = "google/palm-2-chat-bison",
MODEL_GOOGLE_PALM_2_CODECHAT_BISON = "google/palm-2-codechat-bison",
MODEL_GRYPHE_MYTHOMAX_L2_13B_FREE = "gryphe/mythomax-l2-13b:free",
MODEL_GRYPHE_MYTHOMAX_L2_13B = "gryphe/mythomax-l2-13b",
MODEL_META_LLAMA_LLAMA_2_13B_CHAT = "meta-llama/llama-2-13b-chat",
MODEL_META_LLAMA_LLAMA_2_70B_CHAT = "meta-llama/llama-2-70b-chat",

File diff suppressed because one or more lines are too long

File diff suppressed because one or more lines are too long

File diff suppressed because one or more lines are too long

View File

@ -121,7 +121,6 @@ export interface IKBotOptions {
google/gemma-3-27b-it:free | free
google/gemma-3-4b-it | paid
google/gemma-3-4b-it:free | free
google/gemma-7b-it | paid
google/learnlm-1.5-pro-experimental:free | free
google/palm-2-chat-bison | paid
google/palm-2-chat-bison-32k | paid
@ -144,7 +143,6 @@ export interface IKBotOptions {
meta-llama/llama-2-70b-chat | paid
meta-llama/llama-3-70b-instruct | paid
meta-llama/llama-3-8b-instruct | paid
meta-llama/llama-3-8b-instruct:free | free
meta-llama/llama-3.1-405b | paid
meta-llama/llama-3.1-405b-instruct | paid
meta-llama/llama-3.1-70b-instruct | paid
@ -200,7 +198,6 @@ export interface IKBotOptions {
mistralai/mistral-saba | paid
moonshotai/moonlight-16b-a3b-instruct:free | free
gryphe/mythomax-l2-13b | paid
gryphe/mythomax-l2-13b:free | free
neversleep/llama-3-lumimaid-70b | paid
neversleep/llama-3-lumimaid-8b | paid
neversleep/llama-3-lumimaid-8b:extended | paid
@ -251,7 +248,7 @@ export interface IKBotOptions {
openai/o3-mini-high | paid
openchat/openchat-7b | paid
openchat/openchat-7b:free | free
teknium/openhermes-2.5-mistral-7b | paid
all-hands/openhands-lm-32b-v0.1 | paid
perplexity/llama-3.1-sonar-large-128k-online | paid
perplexity/llama-3.1-sonar-small-128k-online | paid
perplexity/r1-1776 | paid
@ -261,9 +258,8 @@ export interface IKBotOptions {
perplexity/sonar-reasoning | paid
perplexity/sonar-reasoning-pro | paid
pygmalionai/mythalion-13b | paid
openrouter/quasar-alpha | paid
qwen/qwen-2-72b-instruct | paid
qwen/qwen-2-7b-instruct | paid
qwen/qwen-2-7b-instruct:free | free
qwen/qwen-vl-max | paid
qwen/qwen-vl-plus | paid
qwen/qwen-max | paid

File diff suppressed because one or more lines are too long

File diff suppressed because one or more lines are too long

View File

@ -52,7 +52,7 @@ ${toolModels.map(model => ` MODEL_TOOLS_${model.id.replace(/[^a-zA-Z0-9]/g, '_'
}
export const build = async () => {
const examplesSrc = path.resolve(__dirname, '../docs_/examples.md')
const examplesSrc = path.resolve(__dirname, '../../docs_/examples.md')
if(exists(examplesSrc)) {
const examples = read(examplesSrc,'string') || ''
const examplesPath = path.resolve(__dirname, '../src/docs-internal/examples.ts')

View File

@ -4,10 +4,12 @@ export enum E_OPENAI_MODEL {
MODEL_DALL_E_3 = "dall-e-3",
MODEL_DALL_E_2 = "dall-e-2",
MODEL_GPT_4O_AUDIO_PREVIEW_2024_10_01 = "gpt-4o-audio-preview-2024-10-01",
MODEL_GPT_4O_MINI_REALTIME_PREVIEW_2024_12_17 = "gpt-4o-mini-realtime-preview-2024-12-17",
MODEL_GPT_4O_MINI_REALTIME_PREVIEW = "gpt-4o-mini-realtime-preview",
MODEL_O1_PRO_2025_03_19 = "o1-pro-2025-03-19",
MODEL_O1_PRO = "o1-pro",
MODEL_GPT_4O_REALTIME_PREVIEW_2024_10_01 = "gpt-4o-realtime-preview-2024-10-01",
MODEL_O1_2024_12_17 = "o1-2024-12-17",
MODEL_GPT_4O_TRANSCRIBE = "gpt-4o-transcribe",
MODEL_O1 = "o1",
MODEL_GPT_4O_MINI_TRANSCRIBE = "gpt-4o-mini-transcribe",
MODEL_GPT_4O_REALTIME_PREVIEW = "gpt-4o-realtime-preview",
MODEL_BABBAGE_002 = "babbage-002",
@ -16,19 +18,18 @@ export enum E_OPENAI_MODEL {
MODEL_TEXT_EMBEDDING_3_LARGE = "text-embedding-3-large",
MODEL_GPT_4 = "gpt-4",
MODEL_TEXT_EMBEDDING_ADA_002 = "text-embedding-ada-002",
MODEL_OMNI_MODERATION_LATEST = "omni-moderation-latest",
MODEL_TTS_1_HD = "tts-1-hd",
MODEL_GPT_4O_MINI_AUDIO_PREVIEW = "gpt-4o-mini-audio-preview",
MODEL_GPT_4O_AUDIO_PREVIEW = "gpt-4o-audio-preview",
MODEL_O1_PREVIEW_2024_09_12 = "o1-preview-2024-09-12",
MODEL_GPT_4O_MINI_REALTIME_PREVIEW = "gpt-4o-mini-realtime-preview",
MODEL_GPT_4O_MINI_REALTIME_PREVIEW_2024_12_17 = "gpt-4o-mini-realtime-preview-2024-12-17",
MODEL_GPT_3_5_TURBO_INSTRUCT_0914 = "gpt-3.5-turbo-instruct-0914",
MODEL_GPT_4O_MINI_SEARCH_PREVIEW = "gpt-4o-mini-search-preview",
MODEL_O1 = "o1",
MODEL_TTS_1_1106 = "tts-1-1106",
MODEL_O1_2024_12_17 = "o1-2024-12-17",
MODEL_DAVINCI_002 = "davinci-002",
MODEL_GPT_3_5_TURBO_1106 = "gpt-3.5-turbo-1106",
MODEL_O1_PRO = "o1-pro",
MODEL_O1_PRO_2025_03_19 = "o1-pro-2025-03-19",
MODEL_GPT_4_TURBO = "gpt-4-turbo",
MODEL_GPT_3_5_TURBO_INSTRUCT = "gpt-3.5-turbo-instruct",
MODEL_GPT_3_5_TURBO = "gpt-3.5-turbo",
@ -49,7 +50,6 @@ export enum E_OPENAI_MODEL {
MODEL_GPT_4_5_PREVIEW = "gpt-4.5-preview",
MODEL_GPT_4_5_PREVIEW_2025_02_27 = "gpt-4.5-preview-2025-02-27",
MODEL_GPT_4O_SEARCH_PREVIEW_2025_03_11 = "gpt-4o-search-preview-2025-03-11",
MODEL_OMNI_MODERATION_LATEST = "omni-moderation-latest",
MODEL_TTS_1 = "tts-1",
MODEL_OMNI_MODERATION_2024_09_26 = "omni-moderation-2024-09-26",
MODEL_TEXT_EMBEDDING_3_SMALL = "text-embedding-3-small",
@ -57,9 +57,9 @@ export enum E_OPENAI_MODEL {
MODEL_GPT_4O_MINI = "gpt-4o-mini",
MODEL_GPT_4O_2024_08_06 = "gpt-4o-2024-08-06",
MODEL_GPT_4O_MINI_2024_07_18 = "gpt-4o-mini-2024-07-18",
MODEL_GPT_4_TURBO_PREVIEW = "gpt-4-turbo-preview",
MODEL_O1_MINI = "o1-mini",
MODEL_GPT_4O_MINI_AUDIO_PREVIEW_2024_12_17 = "gpt-4o-mini-audio-preview-2024-12-17",
MODEL_O1_MINI_2024_09_12 = "o1-mini-2024-09-12",
MODEL_GPT_4_0125_PREVIEW = "gpt-4-0125-preview"
MODEL_GPT_4_0125_PREVIEW = "gpt-4-0125-preview",
MODEL_GPT_4_TURBO_PREVIEW = "gpt-4-turbo-preview"
}

View File

@ -1,4 +1,5 @@
export enum E_OPENROUTER_MODEL_FREE {
MODEL_FREE_OPENROUTER_QUASAR_ALPHA = "openrouter/quasar-alpha",
MODEL_FREE_DEEPSEEK_DEEPSEEK_V3_BASE_FREE = "deepseek/deepseek-v3-base:free",
MODEL_FREE_ALLENAI_MOLMO_7B_D_FREE = "allenai/molmo-7b-d:free",
MODEL_FREE_BYTEDANCE_RESEARCH_UI_TARS_72B_FREE = "bytedance-research/ui-tars-72b:free",
@ -46,14 +47,11 @@ export enum E_OPENROUTER_MODEL_FREE {
MODEL_FREE_GOOGLE_GEMINI_FLASH_1_5_8B_EXP = "google/gemini-flash-1.5-8b-exp",
MODEL_FREE_META_LLAMA_LLAMA_3_1_8B_INSTRUCT_FREE = "meta-llama/llama-3.1-8b-instruct:free",
MODEL_FREE_MISTRALAI_MISTRAL_NEMO_FREE = "mistralai/mistral-nemo:free",
MODEL_FREE_QWEN_QWEN_2_7B_INSTRUCT_FREE = "qwen/qwen-2-7b-instruct:free",
MODEL_FREE_GOOGLE_GEMMA_2_9B_IT_FREE = "google/gemma-2-9b-it:free",
MODEL_FREE_MISTRALAI_MISTRAL_7B_INSTRUCT_FREE = "mistralai/mistral-7b-instruct:free",
MODEL_FREE_MICROSOFT_PHI_3_MINI_128K_INSTRUCT_FREE = "microsoft/phi-3-mini-128k-instruct:free",
MODEL_FREE_MICROSOFT_PHI_3_MEDIUM_128K_INSTRUCT_FREE = "microsoft/phi-3-medium-128k-instruct:free",
MODEL_FREE_META_LLAMA_LLAMA_3_8B_INSTRUCT_FREE = "meta-llama/llama-3-8b-instruct:free",
MODEL_FREE_OPENCHAT_OPENCHAT_7B_FREE = "openchat/openchat-7b:free",
MODEL_FREE_UNDI95_TOPPY_M_7B_FREE = "undi95/toppy-m-7b:free",
MODEL_FREE_HUGGINGFACEH4_ZEPHYR_7B_BETA_FREE = "huggingfaceh4/zephyr-7b-beta:free",
MODEL_FREE_GRYPHE_MYTHOMAX_L2_13B_FREE = "gryphe/mythomax-l2-13b:free"
MODEL_FREE_HUGGINGFACEH4_ZEPHYR_7B_BETA_FREE = "huggingfaceh4/zephyr-7b-beta:free"
}

View File

@ -1,4 +1,6 @@
export enum E_OPENROUTER_MODEL {
MODEL_OPENROUTER_QUASAR_ALPHA = "openrouter/quasar-alpha",
MODEL_ALL_HANDS_OPENHANDS_LM_32B_V0_1 = "all-hands/openhands-lm-32b-v0.1",
MODEL_MISTRAL_MINISTRAL_8B = "mistral/ministral-8b",
MODEL_DEEPSEEK_DEEPSEEK_V3_BASE_FREE = "deepseek/deepseek-v3-base:free",
MODEL_SCB10X_LLAMA3_1_TYPHOON2_8B_INSTRUCT = "scb10x/llama3.1-typhoon2-8b-instruct",
@ -47,9 +49,9 @@ export enum E_OPENROUTER_MODEL {
MODEL_NOUSRESEARCH_DEEPHERMES_3_LLAMA_3_8B_PREVIEW_FREE = "nousresearch/deephermes-3-llama-3-8b-preview:free",
MODEL_OPENAI_GPT_4_5_PREVIEW = "openai/gpt-4.5-preview",
MODEL_GOOGLE_GEMINI_2_0_FLASH_LITE_001 = "google/gemini-2.0-flash-lite-001",
MODEL_ANTHROPIC_CLAUDE_3_7_SONNET_BETA = "anthropic/claude-3.7-sonnet:beta",
MODEL_ANTHROPIC_CLAUDE_3_7_SONNET = "anthropic/claude-3.7-sonnet",
MODEL_ANTHROPIC_CLAUDE_3_7_SONNET_THINKING = "anthropic/claude-3.7-sonnet:thinking",
MODEL_ANTHROPIC_CLAUDE_3_7_SONNET_BETA = "anthropic/claude-3.7-sonnet:beta",
MODEL_PERPLEXITY_R1_1776 = "perplexity/r1-1776",
MODEL_MISTRALAI_MISTRAL_SABA = "mistralai/mistral-saba",
MODEL_COGNITIVECOMPUTATIONS_DOLPHIN3_0_R1_MISTRAL_24B_FREE = "cognitivecomputations/dolphin3.0-r1-mistral-24b:free",
@ -185,8 +187,6 @@ export enum E_OPENROUTER_MODEL {
MODEL_MISTRALAI_MISTRAL_NEMO = "mistralai/mistral-nemo",
MODEL_OPENAI_GPT_4O_MINI = "openai/gpt-4o-mini",
MODEL_OPENAI_GPT_4O_MINI_2024_07_18 = "openai/gpt-4o-mini-2024-07-18",
MODEL_QWEN_QWEN_2_7B_INSTRUCT_FREE = "qwen/qwen-2-7b-instruct:free",
MODEL_QWEN_QWEN_2_7B_INSTRUCT = "qwen/qwen-2-7b-instruct",
MODEL_GOOGLE_GEMMA_2_27B_IT = "google/gemma-2-27b-it",
MODEL_ALPINDALE_MAGNUM_72B = "alpindale/magnum-72b",
MODEL_GOOGLE_GEMMA_2_9B_IT_FREE = "google/gemma-2-9b-it:free",
@ -215,7 +215,6 @@ export enum E_OPENROUTER_MODEL {
MODEL_NEVERSLEEP_LLAMA_3_LUMIMAID_8B_EXTENDED = "neversleep/llama-3-lumimaid-8b:extended",
MODEL_NEVERSLEEP_LLAMA_3_LUMIMAID_8B = "neversleep/llama-3-lumimaid-8b",
MODEL_SAO10K_FIMBULVETR_11B_V2 = "sao10k/fimbulvetr-11b-v2",
MODEL_META_LLAMA_LLAMA_3_8B_INSTRUCT_FREE = "meta-llama/llama-3-8b-instruct:free",
MODEL_META_LLAMA_LLAMA_3_8B_INSTRUCT = "meta-llama/llama-3-8b-instruct",
MODEL_META_LLAMA_LLAMA_3_70B_INSTRUCT = "meta-llama/llama-3-70b-instruct",
MODEL_MISTRALAI_MIXTRAL_8X22B_INSTRUCT = "mistralai/mixtral-8x22b-instruct",
@ -236,7 +235,6 @@ export enum E_OPENROUTER_MODEL {
MODEL_ANTHROPIC_CLAUDE_3_SONNET = "anthropic/claude-3-sonnet",
MODEL_COHERE_COMMAND_R_03_2024 = "cohere/command-r-03-2024",
MODEL_MISTRALAI_MISTRAL_LARGE = "mistralai/mistral-large",
MODEL_GOOGLE_GEMMA_7B_IT = "google/gemma-7b-it",
MODEL_OPENAI_GPT_3_5_TURBO_0613 = "openai/gpt-3.5-turbo-0613",
MODEL_OPENAI_GPT_4_TURBO_PREVIEW = "openai/gpt-4-turbo-preview",
MODEL_NOUSRESEARCH_NOUS_HERMES_2_MIXTRAL_8X7B_DPO = "nousresearch/nous-hermes-2-mixtral-8x7b-dpo",
@ -256,7 +254,6 @@ export enum E_OPENROUTER_MODEL {
MODEL_ANTHROPIC_CLAUDE_2_1 = "anthropic/claude-2.1",
MODEL_ANTHROPIC_CLAUDE_2_BETA = "anthropic/claude-2:beta",
MODEL_ANTHROPIC_CLAUDE_2 = "anthropic/claude-2",
MODEL_TEKNIUM_OPENHERMES_2_5_MISTRAL_7B = "teknium/openhermes-2.5-mistral-7b",
MODEL_UNDI95_TOPPY_M_7B_FREE = "undi95/toppy-m-7b:free",
MODEL_UNDI95_TOPPY_M_7B = "undi95/toppy-m-7b",
MODEL_ALPINDALE_GOLIATH_120B = "alpindale/goliath-120b",
@ -281,7 +278,6 @@ export enum E_OPENROUTER_MODEL {
MODEL_UNDI95_REMM_SLERP_L2_13B = "undi95/remm-slerp-l2-13b",
MODEL_GOOGLE_PALM_2_CHAT_BISON = "google/palm-2-chat-bison",
MODEL_GOOGLE_PALM_2_CODECHAT_BISON = "google/palm-2-codechat-bison",
MODEL_GRYPHE_MYTHOMAX_L2_13B_FREE = "gryphe/mythomax-l2-13b:free",
MODEL_GRYPHE_MYTHOMAX_L2_13B = "gryphe/mythomax-l2-13b",
MODEL_META_LLAMA_LLAMA_2_13B_CHAT = "meta-llama/llama-2-13b-chat",
MODEL_META_LLAMA_LLAMA_2_70B_CHAT = "meta-llama/llama-2-70b-chat",

View File

@ -121,7 +121,6 @@ export interface IKBotOptions {
google/gemma-3-27b-it:free | free
google/gemma-3-4b-it | paid
google/gemma-3-4b-it:free | free
google/gemma-7b-it | paid
google/learnlm-1.5-pro-experimental:free | free
google/palm-2-chat-bison | paid
google/palm-2-chat-bison-32k | paid
@ -144,7 +143,6 @@ export interface IKBotOptions {
meta-llama/llama-2-70b-chat | paid
meta-llama/llama-3-70b-instruct | paid
meta-llama/llama-3-8b-instruct | paid
meta-llama/llama-3-8b-instruct:free | free
meta-llama/llama-3.1-405b | paid
meta-llama/llama-3.1-405b-instruct | paid
meta-llama/llama-3.1-70b-instruct | paid
@ -200,7 +198,6 @@ export interface IKBotOptions {
mistralai/mistral-saba | paid
moonshotai/moonlight-16b-a3b-instruct:free | free
gryphe/mythomax-l2-13b | paid
gryphe/mythomax-l2-13b:free | free
neversleep/llama-3-lumimaid-70b | paid
neversleep/llama-3-lumimaid-8b | paid
neversleep/llama-3-lumimaid-8b:extended | paid
@ -251,7 +248,7 @@ export interface IKBotOptions {
openai/o3-mini-high | paid
openchat/openchat-7b | paid
openchat/openchat-7b:free | free
teknium/openhermes-2.5-mistral-7b | paid
all-hands/openhands-lm-32b-v0.1 | paid
perplexity/llama-3.1-sonar-large-128k-online | paid
perplexity/llama-3.1-sonar-small-128k-online | paid
perplexity/r1-1776 | paid
@ -261,9 +258,8 @@ export interface IKBotOptions {
perplexity/sonar-reasoning | paid
perplexity/sonar-reasoning-pro | paid
pygmalionai/mythalion-13b | paid
openrouter/quasar-alpha | paid
qwen/qwen-2-72b-instruct | paid
qwen/qwen-2-7b-instruct | paid
qwen/qwen-2-7b-instruct:free | free
qwen/qwen-vl-max | paid
qwen/qwen-vl-plus | paid
qwen/qwen-max | paid