maintainence love:)
This commit is contained in:
parent
d2e960d231
commit
b2498d406c
@ -1,5 +1,5 @@
|
||||
{
|
||||
"timestamp": 1746201109298,
|
||||
"timestamp": 1746389240702,
|
||||
"models": [
|
||||
{
|
||||
"id": "gpt-4o-audio-preview-2024-12-17",
|
||||
@ -25,12 +25,30 @@
|
||||
"created": 1727389042,
|
||||
"owned_by": "system"
|
||||
},
|
||||
{
|
||||
"id": "gpt-4-turbo-preview",
|
||||
"object": "model",
|
||||
"created": 1706037777,
|
||||
"owned_by": "system"
|
||||
},
|
||||
{
|
||||
"id": "text-embedding-3-small",
|
||||
"object": "model",
|
||||
"created": 1705948997,
|
||||
"owned_by": "system"
|
||||
},
|
||||
{
|
||||
"id": "gpt-4-turbo",
|
||||
"object": "model",
|
||||
"created": 1712361441,
|
||||
"owned_by": "system"
|
||||
},
|
||||
{
|
||||
"id": "gpt-4-turbo-2024-04-09",
|
||||
"object": "model",
|
||||
"created": 1712601677,
|
||||
"owned_by": "system"
|
||||
},
|
||||
{
|
||||
"id": "gpt-4.1-nano",
|
||||
"object": "model",
|
||||
@ -49,48 +67,24 @@
|
||||
"created": 1727131766,
|
||||
"owned_by": "system"
|
||||
},
|
||||
{
|
||||
"id": "o1-pro-2025-03-19",
|
||||
"object": "model",
|
||||
"created": 1742251504,
|
||||
"owned_by": "system"
|
||||
},
|
||||
{
|
||||
"id": "gpt-4o-realtime-preview",
|
||||
"object": "model",
|
||||
"created": 1727659998,
|
||||
"owned_by": "system"
|
||||
},
|
||||
{
|
||||
"id": "o1-pro",
|
||||
"object": "model",
|
||||
"created": 1742251791,
|
||||
"owned_by": "system"
|
||||
},
|
||||
{
|
||||
"id": "babbage-002",
|
||||
"object": "model",
|
||||
"created": 1692634615,
|
||||
"owned_by": "system"
|
||||
},
|
||||
{
|
||||
"id": "o1",
|
||||
"object": "model",
|
||||
"created": 1734375816,
|
||||
"owned_by": "system"
|
||||
},
|
||||
{
|
||||
"id": "gpt-4",
|
||||
"object": "model",
|
||||
"created": 1687882411,
|
||||
"owned_by": "openai"
|
||||
},
|
||||
{
|
||||
"id": "o1-2024-12-17",
|
||||
"object": "model",
|
||||
"created": 1734326976,
|
||||
"owned_by": "system"
|
||||
},
|
||||
{
|
||||
"id": "text-embedding-ada-002",
|
||||
"object": "model",
|
||||
@ -163,6 +157,18 @@
|
||||
"created": 1744317547,
|
||||
"owned_by": "system"
|
||||
},
|
||||
{
|
||||
"id": "o1",
|
||||
"object": "model",
|
||||
"created": 1734375816,
|
||||
"owned_by": "system"
|
||||
},
|
||||
{
|
||||
"id": "o1-2024-12-17",
|
||||
"object": "model",
|
||||
"created": 1734326976,
|
||||
"owned_by": "system"
|
||||
},
|
||||
{
|
||||
"id": "davinci-002",
|
||||
"object": "model",
|
||||
@ -181,12 +187,6 @@
|
||||
"created": 1741388720,
|
||||
"owned_by": "system"
|
||||
},
|
||||
{
|
||||
"id": "gpt-4-turbo",
|
||||
"object": "model",
|
||||
"created": 1712361441,
|
||||
"owned_by": "system"
|
||||
},
|
||||
{
|
||||
"id": "gpt-3.5-turbo-instruct",
|
||||
"object": "model",
|
||||
@ -199,12 +199,6 @@
|
||||
"created": 1677610602,
|
||||
"owned_by": "openai"
|
||||
},
|
||||
{
|
||||
"id": "gpt-4-turbo-preview",
|
||||
"object": "model",
|
||||
"created": 1706037777,
|
||||
"owned_by": "system"
|
||||
},
|
||||
{
|
||||
"id": "gpt-4o-mini-search-preview-2025-03-11",
|
||||
"object": "model",
|
||||
@ -236,9 +230,9 @@
|
||||
"owned_by": "system"
|
||||
},
|
||||
{
|
||||
"id": "gpt-4-turbo-2024-04-09",
|
||||
"id": "o1-pro",
|
||||
"object": "model",
|
||||
"created": 1712601677,
|
||||
"created": 1742251791,
|
||||
"owned_by": "system"
|
||||
},
|
||||
{
|
||||
@ -253,6 +247,12 @@
|
||||
"created": 1745517030,
|
||||
"owned_by": "system"
|
||||
},
|
||||
{
|
||||
"id": "o1-pro-2025-03-19",
|
||||
"object": "model",
|
||||
"created": 1742251504,
|
||||
"owned_by": "system"
|
||||
},
|
||||
{
|
||||
"id": "o1-preview",
|
||||
"object": "model",
|
||||
|
||||
@ -1,5 +1,5 @@
|
||||
{
|
||||
"timestamp": 1746201109462,
|
||||
"timestamp": 1746389240875,
|
||||
"models": [
|
||||
{
|
||||
"id": "microsoft/phi-4-reasoning-plus:free",
|
||||
@ -533,11 +533,12 @@
|
||||
"stop",
|
||||
"frequency_penalty",
|
||||
"presence_penalty",
|
||||
"repetition_penalty",
|
||||
"response_format",
|
||||
"top_k",
|
||||
"seed",
|
||||
"min_p"
|
||||
"repetition_penalty",
|
||||
"logit_bias",
|
||||
"min_p",
|
||||
"response_format",
|
||||
"seed"
|
||||
]
|
||||
},
|
||||
{
|
||||
@ -992,7 +993,7 @@
|
||||
"name": "Qwen: Qwen3 235B A22B",
|
||||
"created": 1745875757,
|
||||
"description": "Qwen3-235B-A22B is a 235B parameter mixture-of-experts (MoE) model developed by Qwen, activating 22B parameters per forward pass. It supports seamless switching between a \"thinking\" mode for complex reasoning, math, and code tasks, and a \"non-thinking\" mode for general conversational efficiency. The model demonstrates strong reasoning ability, multilingual support (100+ languages and dialects), advanced instruction-following, and agent tool-calling capabilities. It natively handles a 32K token context window and extends up to 131K tokens using YaRN-based scaling.",
|
||||
"context_length": 131072,
|
||||
"context_length": 40960,
|
||||
"architecture": {
|
||||
"modality": "text->text",
|
||||
"input_modalities": [
|
||||
@ -1013,8 +1014,8 @@
|
||||
"internal_reasoning": "0"
|
||||
},
|
||||
"top_provider": {
|
||||
"context_length": 131072,
|
||||
"max_completion_tokens": null,
|
||||
"context_length": 40960,
|
||||
"max_completion_tokens": 40960,
|
||||
"is_moderated": false
|
||||
},
|
||||
"per_request_limits": null,
|
||||
@ -1028,13 +1029,13 @@
|
||||
"frequency_penalty",
|
||||
"repetition_penalty",
|
||||
"top_k",
|
||||
"seed",
|
||||
"stop",
|
||||
"response_format",
|
||||
"min_p",
|
||||
"tools",
|
||||
"tool_choice",
|
||||
"stop",
|
||||
"seed",
|
||||
"min_p",
|
||||
"logit_bias",
|
||||
"response_format",
|
||||
"structured_outputs",
|
||||
"logprobs",
|
||||
"top_logprobs"
|
||||
@ -1302,7 +1303,9 @@
|
||||
"request": "0",
|
||||
"image": "0.00516",
|
||||
"web_search": "0",
|
||||
"internal_reasoning": "0"
|
||||
"internal_reasoning": "0",
|
||||
"input_cache_read": "0.000000625",
|
||||
"input_cache_write": "0.00000285"
|
||||
},
|
||||
"top_provider": {
|
||||
"context_length": 1048576,
|
||||
@ -1544,19 +1547,14 @@
|
||||
},
|
||||
"per_request_limits": null,
|
||||
"supported_parameters": [
|
||||
"tools",
|
||||
"tool_choice",
|
||||
"max_tokens",
|
||||
"temperature",
|
||||
"top_p",
|
||||
"reasoning",
|
||||
"include_reasoning",
|
||||
"structured_outputs",
|
||||
"response_format",
|
||||
"tools",
|
||||
"tool_choice",
|
||||
"stop",
|
||||
"frequency_penalty",
|
||||
"presence_penalty",
|
||||
"seed"
|
||||
"response_format",
|
||||
"structured_outputs"
|
||||
]
|
||||
},
|
||||
{
|
||||
@ -1593,19 +1591,14 @@
|
||||
},
|
||||
"per_request_limits": null,
|
||||
"supported_parameters": [
|
||||
"tools",
|
||||
"tool_choice",
|
||||
"max_tokens",
|
||||
"temperature",
|
||||
"top_p",
|
||||
"reasoning",
|
||||
"include_reasoning",
|
||||
"structured_outputs",
|
||||
"response_format",
|
||||
"tools",
|
||||
"tool_choice",
|
||||
"stop",
|
||||
"frequency_penalty",
|
||||
"presence_penalty",
|
||||
"seed"
|
||||
"response_format",
|
||||
"structured_outputs"
|
||||
]
|
||||
},
|
||||
{
|
||||
@ -3506,47 +3499,6 @@
|
||||
"top_logprobs"
|
||||
]
|
||||
},
|
||||
{
|
||||
"id": "steelskull/l3.3-electra-r1-70b",
|
||||
"name": "SteelSkull: L3.3 Electra R1 70B",
|
||||
"created": 1742067611,
|
||||
"description": "L3.3-Electra-R1-70 is the newest release of the Unnamed series. Built on a DeepSeek R1 Distill base, Electra-R1 integrates various models together to provide an intelligent and coherent model capable of providing deep character insights. Through proper prompting, the model demonstrates advanced reasoning capabilities and unprompted exploration of character inner thoughts and motivations. Read more about the model and [prompting here](https://huggingface.co/Steelskull/L3.3-Electra-R1-70b)",
|
||||
"context_length": 131072,
|
||||
"architecture": {
|
||||
"modality": "text->text",
|
||||
"input_modalities": [
|
||||
"text"
|
||||
],
|
||||
"output_modalities": [
|
||||
"text"
|
||||
],
|
||||
"tokenizer": "Other",
|
||||
"instruct_type": "deepseek-r1"
|
||||
},
|
||||
"pricing": {
|
||||
"prompt": "0.0000007",
|
||||
"completion": "0.00000095",
|
||||
"request": "0",
|
||||
"image": "0",
|
||||
"web_search": "0",
|
||||
"internal_reasoning": "0"
|
||||
},
|
||||
"top_provider": {
|
||||
"context_length": 131072,
|
||||
"max_completion_tokens": 131072,
|
||||
"is_moderated": false
|
||||
},
|
||||
"per_request_limits": null,
|
||||
"supported_parameters": [
|
||||
"max_tokens",
|
||||
"temperature",
|
||||
"top_p",
|
||||
"presence_penalty",
|
||||
"frequency_penalty",
|
||||
"repetition_penalty",
|
||||
"top_k"
|
||||
]
|
||||
},
|
||||
{
|
||||
"id": "google/gemma-3-1b-it:free",
|
||||
"name": "Google: Gemma 3 1B (free)",
|
||||
@ -4172,47 +4124,6 @@
|
||||
"top_k"
|
||||
]
|
||||
},
|
||||
{
|
||||
"id": "latitudegames/wayfarer-large-70b-llama-3.3",
|
||||
"name": "LatitudeGames: Wayfarer Large 70B Llama 3.3",
|
||||
"created": 1741636885,
|
||||
"description": "Wayfarer Large 70B is a roleplay and text-adventure model fine-tuned from Meta’s Llama-3.3-70B-Instruct. Specifically optimized for narrative-driven, challenging scenarios, it introduces realistic stakes, conflicts, and consequences often avoided by standard RLHF-aligned models. Trained using a curated blend of adventure, roleplay, and instructive fiction datasets, Wayfarer emphasizes tense storytelling, authentic player failure scenarios, and robust narrative immersion, making it uniquely suited for interactive fiction and gaming experiences.",
|
||||
"context_length": 131072,
|
||||
"architecture": {
|
||||
"modality": "text->text",
|
||||
"input_modalities": [
|
||||
"text"
|
||||
],
|
||||
"output_modalities": [
|
||||
"text"
|
||||
],
|
||||
"tokenizer": "Llama3",
|
||||
"instruct_type": null
|
||||
},
|
||||
"pricing": {
|
||||
"prompt": "0.0000008",
|
||||
"completion": "0.0000009",
|
||||
"request": "0",
|
||||
"image": "0",
|
||||
"web_search": "0",
|
||||
"internal_reasoning": "0"
|
||||
},
|
||||
"top_provider": {
|
||||
"context_length": 131072,
|
||||
"max_completion_tokens": 131072,
|
||||
"is_moderated": false
|
||||
},
|
||||
"per_request_limits": null,
|
||||
"supported_parameters": [
|
||||
"max_tokens",
|
||||
"temperature",
|
||||
"top_p",
|
||||
"presence_penalty",
|
||||
"frequency_penalty",
|
||||
"repetition_penalty",
|
||||
"top_k"
|
||||
]
|
||||
},
|
||||
{
|
||||
"id": "thedrummer/skyfall-36b-v2",
|
||||
"name": "TheDrummer: Skyfall 36B V2",
|
||||
@ -4290,12 +4201,12 @@
|
||||
"max_tokens",
|
||||
"temperature",
|
||||
"top_p",
|
||||
"presence_penalty",
|
||||
"frequency_penalty",
|
||||
"repetition_penalty",
|
||||
"top_k",
|
||||
"stop",
|
||||
"frequency_penalty",
|
||||
"presence_penalty",
|
||||
"repetition_penalty",
|
||||
"response_format",
|
||||
"top_k",
|
||||
"seed",
|
||||
"min_p"
|
||||
]
|
||||
@ -6969,17 +6880,10 @@
|
||||
},
|
||||
"per_request_limits": null,
|
||||
"supported_parameters": [
|
||||
"tools",
|
||||
"tool_choice",
|
||||
"max_tokens",
|
||||
"temperature",
|
||||
"top_p",
|
||||
"stop",
|
||||
"frequency_penalty",
|
||||
"presence_penalty",
|
||||
"seed",
|
||||
"response_format",
|
||||
"structured_outputs"
|
||||
"stop"
|
||||
]
|
||||
},
|
||||
{
|
||||
@ -7344,7 +7248,7 @@
|
||||
"name": "EVA Qwen2.5 72B",
|
||||
"created": 1732210606,
|
||||
"description": "EVA Qwen2.5 72B is a roleplay and storywriting specialist model. It's a full-parameter finetune of Qwen2.5-72B on mixture of synthetic and natural data.\n\nIt uses Celeste 70B 0.1 data mixture, greatly expanding it to improve versatility, creativity and \"flavor\" of the resulting model.",
|
||||
"context_length": 131072,
|
||||
"context_length": 16384,
|
||||
"architecture": {
|
||||
"modality": "text->text",
|
||||
"input_modalities": [
|
||||
@ -7357,16 +7261,16 @@
|
||||
"instruct_type": "chatml"
|
||||
},
|
||||
"pricing": {
|
||||
"prompt": "0.0000009",
|
||||
"completion": "0.0000012",
|
||||
"prompt": "0.000004",
|
||||
"completion": "0.000006",
|
||||
"request": "0",
|
||||
"image": "0",
|
||||
"web_search": "0",
|
||||
"internal_reasoning": "0"
|
||||
},
|
||||
"top_provider": {
|
||||
"context_length": 131072,
|
||||
"max_completion_tokens": 131072,
|
||||
"context_length": 16384,
|
||||
"max_completion_tokens": 4096,
|
||||
"is_moderated": false
|
||||
},
|
||||
"per_request_limits": null,
|
||||
@ -7374,11 +7278,11 @@
|
||||
"max_tokens",
|
||||
"temperature",
|
||||
"top_p",
|
||||
"presence_penalty",
|
||||
"stop",
|
||||
"frequency_penalty",
|
||||
"presence_penalty",
|
||||
"repetition_penalty",
|
||||
"top_k",
|
||||
"stop",
|
||||
"min_p",
|
||||
"seed"
|
||||
]
|
||||
|
||||
File diff suppressed because one or more lines are too long
File diff suppressed because one or more lines are too long
4
packages/kbot/dist/package-lock.json
generated
vendored
4
packages/kbot/dist/package-lock.json
generated
vendored
@ -1,12 +1,12 @@
|
||||
{
|
||||
"name": "@plastichub/kbot",
|
||||
"version": "1.1.30",
|
||||
"version": "1.1.31",
|
||||
"lockfileVersion": 3,
|
||||
"requires": true,
|
||||
"packages": {
|
||||
"": {
|
||||
"name": "@plastichub/kbot",
|
||||
"version": "1.1.30",
|
||||
"version": "1.1.31",
|
||||
"license": "ISC",
|
||||
"dependencies": {
|
||||
"node-emoji": "^2.2.0"
|
||||
|
||||
2
packages/kbot/dist/package.json
vendored
2
packages/kbot/dist/package.json
vendored
@ -1,6 +1,6 @@
|
||||
{
|
||||
"name": "@plastichub/kbot",
|
||||
"version": "1.1.30",
|
||||
"version": "1.1.31",
|
||||
"main": "main_node.js",
|
||||
"author": "",
|
||||
"license": "ISC",
|
||||
|
||||
File diff suppressed because one or more lines are too long
File diff suppressed because one or more lines are too long
14
packages/kbot/src/models/cache/openai-models.ts
vendored
14
packages/kbot/src/models/cache/openai-models.ts
vendored
@ -3,17 +3,16 @@ export enum E_OPENAI_MODEL {
|
||||
MODEL_DALL_E_3 = "dall-e-3",
|
||||
MODEL_DALL_E_2 = "dall-e-2",
|
||||
MODEL_GPT_4O_AUDIO_PREVIEW_2024_10_01 = "gpt-4o-audio-preview-2024-10-01",
|
||||
MODEL_GPT_4_TURBO_PREVIEW = "gpt-4-turbo-preview",
|
||||
MODEL_TEXT_EMBEDDING_3_SMALL = "text-embedding-3-small",
|
||||
MODEL_GPT_4_TURBO = "gpt-4-turbo",
|
||||
MODEL_GPT_4_TURBO_2024_04_09 = "gpt-4-turbo-2024-04-09",
|
||||
MODEL_GPT_4_1_NANO = "gpt-4.1-nano",
|
||||
MODEL_GPT_4_1_NANO_2025_04_14 = "gpt-4.1-nano-2025-04-14",
|
||||
MODEL_GPT_4O_REALTIME_PREVIEW_2024_10_01 = "gpt-4o-realtime-preview-2024-10-01",
|
||||
MODEL_O1_PRO_2025_03_19 = "o1-pro-2025-03-19",
|
||||
MODEL_GPT_4O_REALTIME_PREVIEW = "gpt-4o-realtime-preview",
|
||||
MODEL_O1_PRO = "o1-pro",
|
||||
MODEL_BABBAGE_002 = "babbage-002",
|
||||
MODEL_O1 = "o1",
|
||||
MODEL_GPT_4 = "gpt-4",
|
||||
MODEL_O1_2024_12_17 = "o1-2024-12-17",
|
||||
MODEL_TEXT_EMBEDDING_ADA_002 = "text-embedding-ada-002",
|
||||
MODEL_CHATGPT_4O_LATEST = "chatgpt-4o-latest",
|
||||
MODEL_GPT_4O_REALTIME_PREVIEW_2024_12_17 = "gpt-4o-realtime-preview-2024-12-17",
|
||||
@ -26,21 +25,22 @@ export enum E_OPENAI_MODEL {
|
||||
MODEL_GPT_3_5_TURBO_INSTRUCT_0914 = "gpt-3.5-turbo-instruct-0914",
|
||||
MODEL_GPT_4O_MINI_SEARCH_PREVIEW = "gpt-4o-mini-search-preview",
|
||||
MODEL_GPT_4_1_MINI_2025_04_14 = "gpt-4.1-mini-2025-04-14",
|
||||
MODEL_O1 = "o1",
|
||||
MODEL_O1_2024_12_17 = "o1-2024-12-17",
|
||||
MODEL_DAVINCI_002 = "davinci-002",
|
||||
MODEL_GPT_3_5_TURBO_1106 = "gpt-3.5-turbo-1106",
|
||||
MODEL_GPT_4O_SEARCH_PREVIEW = "gpt-4o-search-preview",
|
||||
MODEL_GPT_4_TURBO = "gpt-4-turbo",
|
||||
MODEL_GPT_3_5_TURBO_INSTRUCT = "gpt-3.5-turbo-instruct",
|
||||
MODEL_GPT_3_5_TURBO = "gpt-3.5-turbo",
|
||||
MODEL_GPT_4_TURBO_PREVIEW = "gpt-4-turbo-preview",
|
||||
MODEL_GPT_4O_MINI_SEARCH_PREVIEW_2025_03_11 = "gpt-4o-mini-search-preview-2025-03-11",
|
||||
MODEL_GPT_4_0125_PREVIEW = "gpt-4-0125-preview",
|
||||
MODEL_GPT_4O_2024_11_20 = "gpt-4o-2024-11-20",
|
||||
MODEL_WHISPER_1 = "whisper-1",
|
||||
MODEL_GPT_4O_2024_05_13 = "gpt-4o-2024-05-13",
|
||||
MODEL_GPT_4_TURBO_2024_04_09 = "gpt-4-turbo-2024-04-09",
|
||||
MODEL_O1_PRO = "o1-pro",
|
||||
MODEL_GPT_3_5_TURBO_16K = "gpt-3.5-turbo-16k",
|
||||
MODEL_GPT_IMAGE_1 = "gpt-image-1",
|
||||
MODEL_O1_PRO_2025_03_19 = "o1-pro-2025-03-19",
|
||||
MODEL_O1_PREVIEW = "o1-preview",
|
||||
MODEL_GPT_4_0613 = "gpt-4-0613",
|
||||
MODEL_TEXT_EMBEDDING_3_LARGE = "text-embedding-3-large",
|
||||
|
||||
@ -73,7 +73,6 @@ export enum E_OPENROUTER_MODEL {
|
||||
MODEL_MISTRALAI_MISTRAL_SMALL_3_1_24B_INSTRUCT_FREE = "mistralai/mistral-small-3.1-24b-instruct:free",
|
||||
MODEL_MISTRALAI_MISTRAL_SMALL_3_1_24B_INSTRUCT = "mistralai/mistral-small-3.1-24b-instruct",
|
||||
MODEL_OPEN_R1_OLYMPICCODER_32B_FREE = "open-r1/olympiccoder-32b:free",
|
||||
MODEL_STEELSKULL_L3_3_ELECTRA_R1_70B = "steelskull/l3.3-electra-r1-70b",
|
||||
MODEL_GOOGLE_GEMMA_3_1B_IT_FREE = "google/gemma-3-1b-it:free",
|
||||
MODEL_GOOGLE_GEMMA_3_4B_IT_FREE = "google/gemma-3-4b-it:free",
|
||||
MODEL_GOOGLE_GEMMA_3_4B_IT = "google/gemma-3-4b-it",
|
||||
@ -88,7 +87,6 @@ export enum E_OPENROUTER_MODEL {
|
||||
MODEL_GOOGLE_GEMMA_3_27B_IT_FREE = "google/gemma-3-27b-it:free",
|
||||
MODEL_GOOGLE_GEMMA_3_27B_IT = "google/gemma-3-27b-it",
|
||||
MODEL_THEDRUMMER_ANUBIS_PRO_105B_V1 = "thedrummer/anubis-pro-105b-v1",
|
||||
MODEL_LATITUDEGAMES_WAYFARER_LARGE_70B_LLAMA_3_3 = "latitudegames/wayfarer-large-70b-llama-3.3",
|
||||
MODEL_THEDRUMMER_SKYFALL_36B_V2 = "thedrummer/skyfall-36b-v2",
|
||||
MODEL_MICROSOFT_PHI_4_MULTIMODAL_INSTRUCT = "microsoft/phi-4-multimodal-instruct",
|
||||
MODEL_PERPLEXITY_SONAR_REASONING_PRO = "perplexity/sonar-reasoning-pro",
|
||||
|
||||
@ -77,6 +77,9 @@ export interface IKBotOptions {
|
||||
cohere/command-r-plus-04-2024 | paid
|
||||
cohere/command-r-plus-08-2024 | paid
|
||||
cohere/command-r7b-12-2024 | paid
|
||||
deepseek/deepseek-coder | paid
|
||||
deepseek/deepseek-prover-v2 | paid
|
||||
deepseek/deepseek-prover-v2:free | free
|
||||
deepseek/deepseek-r1-zero:free | free
|
||||
deepseek/deepseek-chat | paid
|
||||
deepseek/deepseek-chat:free | free
|
||||
@ -130,6 +133,7 @@ export interface IKBotOptions {
|
||||
google/palm-2-codechat-bison | paid
|
||||
google/palm-2-codechat-bison-32k | paid
|
||||
huggingfaceh4/zephyr-7b-beta:free | free
|
||||
inception/mercury-coder-small-beta | paid
|
||||
infermatic/mn-inferor-12b | paid
|
||||
inflection/inflection-3-pi | paid
|
||||
inflection/inflection-3-productivity | paid
|
||||
@ -164,10 +168,14 @@ export interface IKBotOptions {
|
||||
meta-llama/llama-4-maverick:free | free
|
||||
meta-llama/llama-4-scout | paid
|
||||
meta-llama/llama-4-scout:free | free
|
||||
meta-llama/llama-guard-4-12b | paid
|
||||
meta-llama/llama-guard-2-8b | paid
|
||||
microsoft/mai-ds-r1:free | free
|
||||
microsoft/phi-4 | paid
|
||||
microsoft/phi-4-multimodal-instruct | paid
|
||||
microsoft/phi-4-reasoning:free | free
|
||||
microsoft/phi-4-reasoning-plus | paid
|
||||
microsoft/phi-4-reasoning-plus:free | free
|
||||
microsoft/phi-3-medium-128k-instruct | paid
|
||||
microsoft/phi-3-mini-128k-instruct | paid
|
||||
microsoft/phi-3.5-mini-128k-instruct | paid
|
||||
@ -217,7 +225,9 @@ export interface IKBotOptions {
|
||||
nousresearch/hermes-2-pro-llama-3-8b | paid
|
||||
nvidia/llama-3.1-nemotron-70b-instruct | paid
|
||||
nvidia/llama-3.1-nemotron-ultra-253b-v1:free | free
|
||||
nvidia/llama-3.3-nemotron-super-49b-v1 | paid
|
||||
nvidia/llama-3.3-nemotron-super-49b-v1:free | free
|
||||
allenai/olmo-7b-instruct | paid
|
||||
open-r1/olympiccoder-32b:free | free
|
||||
openai/chatgpt-4o-latest | paid
|
||||
openai/gpt-3.5-turbo | paid
|
||||
@ -257,6 +267,8 @@ export interface IKBotOptions {
|
||||
openai/o3-mini-high | paid
|
||||
openai/o4-mini | paid
|
||||
openai/o4-mini-high | paid
|
||||
opengvlab/internvl3-14b:free | free
|
||||
opengvlab/internvl3-2b:free | free
|
||||
all-hands/openhands-lm-32b-v0.1 | paid
|
||||
perplexity/llama-3.1-sonar-large-128k-online | paid
|
||||
perplexity/llama-3.1-sonar-small-128k-online | paid
|
||||
@ -273,6 +285,7 @@ export interface IKBotOptions {
|
||||
qwen/qwen-max | paid
|
||||
qwen/qwen-plus | paid
|
||||
qwen/qwen-turbo | paid
|
||||
qwen/qwen2.5-coder-7b-instruct | paid
|
||||
qwen/qwen2.5-vl-32b-instruct | paid
|
||||
qwen/qwen2.5-vl-32b-instruct:free | free
|
||||
qwen/qwen2.5-vl-3b-instruct:free | free
|
||||
@ -281,6 +294,8 @@ export interface IKBotOptions {
|
||||
qwen/qwen-2.5-vl-72b-instruct | paid
|
||||
qwen/qwen-2.5-vl-7b-instruct | paid
|
||||
qwen/qwen-2.5-vl-7b-instruct:free | free
|
||||
qwen/qwen3-0.6b-04-28:free | free
|
||||
qwen/qwen3-1.7b:free | free
|
||||
qwen/qwen3-14b | paid
|
||||
qwen/qwen3-14b:free | free
|
||||
qwen/qwen3-235b-a22b | paid
|
||||
@ -289,6 +304,8 @@ export interface IKBotOptions {
|
||||
qwen/qwen3-30b-a3b:free | free
|
||||
qwen/qwen3-32b | paid
|
||||
qwen/qwen3-32b:free | free
|
||||
qwen/qwen3-4b:free | free
|
||||
qwen/qwen3-8b | paid
|
||||
qwen/qwen3-8b:free | free
|
||||
qwen/qwq-32b | paid
|
||||
qwen/qwq-32b:free | free
|
||||
@ -304,7 +321,6 @@ export interface IKBotOptions {
|
||||
rekaai/reka-flash-3:free | free
|
||||
undi95/remm-slerp-l2-13b | paid
|
||||
thedrummer/rocinante-12b | paid
|
||||
sophosympatheia/rogue-rose-103b-v0.2:free | free
|
||||
sao10k/l3-lunaris-8b | paid
|
||||
sao10k/l3-euryale-70b | paid
|
||||
sao10k/l3.1-euryale-70b | paid
|
||||
@ -322,6 +338,7 @@ export interface IKBotOptions {
|
||||
thudm/glm-z1-9b:free | free
|
||||
thudm/glm-z1-rumination-32b | paid
|
||||
tngtech/deepseek-r1t-chimera:free | free
|
||||
undi95/toppy-m-7b | paid
|
||||
scb10x/llama3.1-typhoon2-70b-instruct | paid
|
||||
scb10x/llama3.1-typhoon2-8b-instruct | paid
|
||||
thedrummer/unslopnemo-12b | paid
|
||||
|
||||
@ -1,179 +0,0 @@
|
||||
|
||||
### Modbus Registers
|
||||
|
||||
| type | description | address (Hex) | address (dec) | parameter group |
|
||||
| :--------------------- | :------------------------- | :------------ | :------------ | :-------------- |
|
||||
| write holding register | P0-16 functional parameters | F010H | 61456 | P0 |
|
||||
| write holding register | AC-08 functional parameters | AC08H | 44040 | AC |
|
||||
|
||||
|
||||
|
||||
|
||||
### Modbus Registers (Page 1)
|
||||
|
||||
| type | description | address in Hex | address in decimal | parameter group |
|
||||
| :------------------------- | :------------------------------------------ | :--------------- | :------------------- | :---------------- |
|
||||
| Read Holding Register (03) | U-Group monitoring parameter U0-11 (Example) | 700BH | 28683 | U-Group (U0-11) |
|
||||
| Read Holding Register (03) | Frequency converter fault code | 8000H | 32768 | F9-14 (context) |
|
||||
| Read Holding Register (03) | Frequency converter operating status | 3000H | 12288 | |
|
||||
|
||||
|
||||
### Control Parameters Modbus Registers
|
||||
|
||||
| type | description | address (Hex) | address (dec) | parameter group |
|
||||
| :-------------------- | :------------------------------------------------ | :------------ | :------------ | :---------------- |
|
||||
| write holding register | Communication Command (1: fwd, 2: rev, 3: stop) | 3000H | 12288 | communication |
|
||||
| write holding register | Control Command (1: fwd, 2: rev, 3: fwd jog) | 2000H | 8192 | Control command |
|
||||
|
||||
|
||||
### Modbus Registers
|
||||
|
||||
| type | description | address Hex | address decimal | parameter group |
|
||||
| :-------------------- | :----------------------------------------------------------------- | :---------- | :-------------- | :------------------------------ |
|
||||
| write holding register | Communication setting value (Frequency/torque/PID sources, etc.) | 1000H | 4096 | Communication setting value |
|
||||
| write holding register | Digital output terminal control (BIT0: DO1 output control) | 2001H | 8193 | Digital output terminal control |
|
||||
|
||||
|
||||
### Output Control Registers
|
||||
|
||||
| type | description | address (Hex) | address (decimal) | parameter group |
|
||||
| :------------------------- | :---------------------------------- | :------------ | :---------------- | :---------------------------------------------------------- |
|
||||
| Write Holding Register (FC06/16) | Analog output AO1 | 2002H | 8194 | Analog output AO1, AO2, high-speed pulse output FMP control |
|
||||
| Write Holding Register (FC06/16) | Analog output AO2 | 2003H | 8195 | Analog output AO1, AO2, high-speed pulse output FMP control |
|
||||
| Write Holding Register (FC06/16) | High-speed pulse output FMP control | 2004H | 8196 | Analog output AO1, AO2, high-speed pulse output FMP control |
|
||||
|
||||
|
||||
### Parameter Initialization Registers (Page 1)
|
||||
|
||||
|
||||
| type | description | address (Hex) | address (decimal) | parameter group |
|
||||
| :------------------- | :------------------------------------------------------------------------------------------------------------------------------------------------ | :------------ | :---------------- | :---------------- |
|
||||
| Write Holding Reg. | User password verification | 1F00H | 7936 | Parameter Control |
|
||||
| Write Holding Reg. | Parameter initialization command (1: Restore factory, 2: Clear record info, 4: Restore user backup, 5: Backup user current parameters) | 1F01H | 7937 | Parameter Control |
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
### Modbus Register Mapping (Page 1)
|
||||
|
||||
|
||||
| type | description | address in Hex | address in decimal | parameter group |
|
||||
| :--------------------- | :--------------------------------- | :-------------- | :----------------- | :-------------- |
|
||||
| Read Holding Register | Read P0-PE Group Parameters | 0xF000 - 0xFEFF | 61440 - 65279 | P0 ~ PE Group |
|
||||
| Write Holding Register | Write P0-PE Group Parameters (RAM) | 0x0000 - 0x0EFF | 0 - 2303 | P0 ~ PE Group |
|
||||
| Read Holding Register | Read A0-AC Group Parameters | 0xA000 - 0xACFF | 40960 - 44287 | A0 ~ AC Group |
|
||||
| Write Holding Register | Write A0-AC Group Parameters (RAM) | 0x4000 - 0x4CFF | 16384 - 19711 | A0 ~ AC Group |
|
||||
| Read Holding Register | Read U0 Group Parameters | 0x7000 - 0x70FF | 28672 - 28927 | U0 Group |
|
||||
|
||||
|
||||
### Modbus Registers
|
||||
|
||||
| type | description | address in Hex | address in decimal | parameter group |
|
||||
| :------------------ | :---------------------------------------------- | :------------- | :----------------- | :---------------- |
|
||||
| Write Holding (06/16) | Functional Code P3-12 (Write-only RAM) | 030C | 780 | P group |
|
||||
| Write Holding (06/16) | Functional Code A0-05 (Write-only RAM) | 4005 | 16389 | A group |
|
||||
| Write Holding (06/16) | Data input from host computer (2 decimal points) | 1000 | 4096 | Not specified |
|
||||
|
||||
|
||||
### Modbus Registers
|
||||
|
||||
| type | description | address in Hex | address in decimal |
|
||||
| :---------------------- | :------------------------------------------------------------ | :--------------- | :------------------- |
|
||||
| Write Holding Reg (06/16) | Communication setting value (decimal system) -10000 - 10000 | 1000H | 4096 |
|
||||
| Read Holding Reg (03) | Running frequency | 1001H | 4097 |
|
||||
| Read Holding Reg (03) | Busbar voltage | 1002H | 4098 |
|
||||
| Read Holding Reg (03) | Output voltage | 1003H | 4099 |
|
||||
| Read Holding Reg (03) | Output current | 1004H | 4100 |
|
||||
| Read Holding Reg (03) | Output power | 1005H | 4101 |
|
||||
| Read Holding Reg (03) | Output torque | 1006H | 4102 |
|
||||
| Read Holding Reg (03) | Running speed | 1007H | 4103 |
|
||||
| Write Holding Reg (06/16) | PID set | 1010H | 4112 |
|
||||
| Read Holding Reg (03) | PID feedback | 1011H | 4113 |
|
||||
| Read Holding Reg (03) | PLC process | 1012H | 4114 |
|
||||
| Read Holding Reg (03) | Pulse input frequency, unit: 0.01kHz | 1013H | 4115 |
|
||||
| Read Holding Reg (03) | Feedback speed, unit 0.1Hz | 1014H | 4116 |
|
||||
| Read Holding Reg (03) | Remaining run time | 1015H | 4117 |
|
||||
| Read Holding Reg (03) | AI1 Pre-correction voltage | 1016H | 4118 |
|
||||
| Read Holding Reg (03) | AI2 Pre-correction voltage | 1017H | 4119 |
|
||||
|
||||
|
||||
|
||||
### Modbus Registers
|
||||
|
||||
| type | description | address in Hex | address in decimal | parameter group |
|
||||
| :-------------------------- | :------------------------------ | :------------- | :----------------- | :---------------- |
|
||||
| read holding register (FC03)| DI input sign | 1008H | 4104 | |
|
||||
| read holding register (FC03)| DO output sign | 1009H | 4105 | |
|
||||
| read holding register (FC03)| AI1 voltage | 100AH | 4106 | |
|
||||
| read holding register (FC03)| AI2 voltage | 100BH | 4107 | |
|
||||
| read holding register (FC03)| AI3 voltage | 100CH | 4108 | |
|
||||
| read holding register (FC03)| Count value input | 100DH | 4109 | |
|
||||
| read holding register (FC03)| Length value input | 100EH | 4110 | |
|
||||
| read holding register (FC03)| Loading speed | 100FH | 4111 | |
|
||||
| read holding register (FC03)| AI3 Pre-correction voltage | 1018H | 4120 | |
|
||||
| read holding register (FC03)| Linear speed | 1019H | 4121 | |
|
||||
| read holding register (FC03)| Current power on time | 101AH | 4122 | |
|
||||
| read holding register (FC03)| Current running time | 101BH | 4123 | |
|
||||
| read holding register (FC03)| Pulse input frequency, unit: 1Hz| 101CH | 4124 | |
|
||||
| read holding register (FC03)| Communication set value | 101DH | 4125 | |
|
||||
| read holding register (FC03)| Actual feedback speed | 101EH | 4126 | |
|
||||
| read holding register (FC03)| Main frequency X display | 101FH | 4127 | |
|
||||
| read holding register (FC03)| Auxiliary frequency Y display | 1020H | 4128 | |
|
||||
|
||||
|
||||
|
||||
### Modbus Registers
|
||||
|
||||
|
||||
| type | description | address in Hex | address in decimal | parameter group |
|
||||
| :--- | :---------------------- | :------------- | :----------------- | :-------------- |
|
||||
| FC16 | Input Control Command | 2000H | 8192 | |
|
||||
| FC03 | Frequency Conv. Status | 3000H | 12288 | |
|
||||
|
||||
|
||||
|
||||
### Modbus Registers
|
||||
|
||||
|
||||
| type | description | address in Hex | address in decimal | parameter group |
|
||||
| :------------ | :------------------------------ | :------------- | :----------------- | :------------------------------ |
|
||||
| FC03 | Parameter lock password check | 1F00 | 7936 | Parameter lock password check |
|
||||
| FC06 / FC16 | Digital output terminal control | 2001 | 8193 | Digital output terminal control |
|
||||
|
||||
|
||||
### Modbus Registers
|
||||
|
||||
| type | description | address in Hex | address in decimal | parameter group |
|
||||
|:-------------|:----------------------------------------|:---------------|:-------------------|:------------------|
|
||||
| FC 06/16 | Analog output AO1 control (write only) | 2002H | 8194 | |
|
||||
| FC 06/16 | Analog output AO2 control (write only) | 2003H | 8195 | |
|
||||
| FC 06/16 | Pulse output control (write only) | 2004H | 8196 | |
|
||||
| FC 03/04 (?) | Fault description of frequency converter | 8000H | 32768 | Fault information |
|
||||
|
||||
|
||||
|
||||
### Modbus Registers
|
||||
|
||||
| type | description | address in Hex | address in decimal | parameter group |
|
||||
| :-------------------- | :---------- | :------------- | :----------------- | :-------------- |
|
||||
| Write Holding Register | Baud rate | Fd00 | 64768 | FD |
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
### Modbus Register Examples
|
||||
|
||||
| type (FC) | description | address (Hex) | address (decimal) | parameter group |
|
||||
| :-------- | :------------------------------------------------------- | :------------ | :---------------- | :---------------- |
|
||||
| 06 | Communication command channel selection (Value 2) | F002 | 61442 | P002 |
|
||||
| 06 | Main frequency source selection (Value 9) | F003 | 61443 | P003 |
|
||||
| 06 | Run Command (Start = Value 1) | 2000 | 8192 | Run Control |
|
||||
| 06 | Set Running Frequency (e.g., 10Hz = Value 2000h?) | 1000 | 4096 | Frequency Setting |
|
||||
| 06 | Run Command (Stop = Value 6) | 2000 | 8192 | Run Control |
|
||||
File diff suppressed because it is too large
Load Diff
Loading…
Reference in New Issue
Block a user