maintainence love:)
This commit is contained in:
parent
7b07f1a55a
commit
f0aee9ce27
@ -1,5 +1,5 @@
|
||||
{
|
||||
"timestamp": 1773934586082,
|
||||
"timestamp": 1774802354444,
|
||||
"models": [
|
||||
{
|
||||
"id": "gpt-4-0613",
|
||||
@ -85,12 +85,6 @@
|
||||
"created": 1698798177,
|
||||
"owned_by": "system"
|
||||
},
|
||||
{
|
||||
"id": "gpt-4-1106-preview",
|
||||
"object": "model",
|
||||
"created": 1698957206,
|
||||
"owned_by": "system"
|
||||
},
|
||||
{
|
||||
"id": "gpt-3.5-turbo-1106",
|
||||
"object": "model",
|
||||
@ -127,18 +121,6 @@
|
||||
"created": 1705953180,
|
||||
"owned_by": "system"
|
||||
},
|
||||
{
|
||||
"id": "gpt-4-0125-preview",
|
||||
"object": "model",
|
||||
"created": 1706037612,
|
||||
"owned_by": "system"
|
||||
},
|
||||
{
|
||||
"id": "gpt-4-turbo-preview",
|
||||
"object": "model",
|
||||
"created": 1706037777,
|
||||
"owned_by": "system"
|
||||
},
|
||||
{
|
||||
"id": "gpt-3.5-turbo-0125",
|
||||
"object": "model",
|
||||
|
||||
File diff suppressed because it is too large
Load Diff
@ -13,15 +13,12 @@ export declare enum E_OPENAI_MODEL {
|
||||
MODEL_GPT_3_5_TURBO_INSTRUCT_0914 = "gpt-3.5-turbo-instruct-0914",
|
||||
MODEL_DALL_E_3 = "dall-e-3",
|
||||
MODEL_DALL_E_2 = "dall-e-2",
|
||||
MODEL_GPT_4_1106_PREVIEW = "gpt-4-1106-preview",
|
||||
MODEL_GPT_3_5_TURBO_1106 = "gpt-3.5-turbo-1106",
|
||||
MODEL_TTS_1_HD = "tts-1-hd",
|
||||
MODEL_TTS_1_1106 = "tts-1-1106",
|
||||
MODEL_TTS_1_HD_1106 = "tts-1-hd-1106",
|
||||
MODEL_TEXT_EMBEDDING_3_SMALL = "text-embedding-3-small",
|
||||
MODEL_TEXT_EMBEDDING_3_LARGE = "text-embedding-3-large",
|
||||
MODEL_GPT_4_0125_PREVIEW = "gpt-4-0125-preview",
|
||||
MODEL_GPT_4_TURBO_PREVIEW = "gpt-4-turbo-preview",
|
||||
MODEL_GPT_3_5_TURBO_0125 = "gpt-3.5-turbo-0125",
|
||||
MODEL_GPT_4_TURBO = "gpt-4-turbo",
|
||||
MODEL_GPT_4_TURBO_2024_04_09 = "gpt-4-turbo-2024-04-09",
|
||||
|
||||
@ -14,15 +14,12 @@ export var E_OPENAI_MODEL;
|
||||
E_OPENAI_MODEL["MODEL_GPT_3_5_TURBO_INSTRUCT_0914"] = "gpt-3.5-turbo-instruct-0914";
|
||||
E_OPENAI_MODEL["MODEL_DALL_E_3"] = "dall-e-3";
|
||||
E_OPENAI_MODEL["MODEL_DALL_E_2"] = "dall-e-2";
|
||||
E_OPENAI_MODEL["MODEL_GPT_4_1106_PREVIEW"] = "gpt-4-1106-preview";
|
||||
E_OPENAI_MODEL["MODEL_GPT_3_5_TURBO_1106"] = "gpt-3.5-turbo-1106";
|
||||
E_OPENAI_MODEL["MODEL_TTS_1_HD"] = "tts-1-hd";
|
||||
E_OPENAI_MODEL["MODEL_TTS_1_1106"] = "tts-1-1106";
|
||||
E_OPENAI_MODEL["MODEL_TTS_1_HD_1106"] = "tts-1-hd-1106";
|
||||
E_OPENAI_MODEL["MODEL_TEXT_EMBEDDING_3_SMALL"] = "text-embedding-3-small";
|
||||
E_OPENAI_MODEL["MODEL_TEXT_EMBEDDING_3_LARGE"] = "text-embedding-3-large";
|
||||
E_OPENAI_MODEL["MODEL_GPT_4_0125_PREVIEW"] = "gpt-4-0125-preview";
|
||||
E_OPENAI_MODEL["MODEL_GPT_4_TURBO_PREVIEW"] = "gpt-4-turbo-preview";
|
||||
E_OPENAI_MODEL["MODEL_GPT_3_5_TURBO_0125"] = "gpt-3.5-turbo-0125";
|
||||
E_OPENAI_MODEL["MODEL_GPT_4_TURBO"] = "gpt-4-turbo";
|
||||
E_OPENAI_MODEL["MODEL_GPT_4_TURBO_2024_04_09"] = "gpt-4-turbo-2024-04-09";
|
||||
@ -126,4 +123,4 @@ export var E_OPENAI_MODEL;
|
||||
E_OPENAI_MODEL["MODEL_WHISPER_1"] = "whisper-1";
|
||||
E_OPENAI_MODEL["MODEL_TEXT_EMBEDDING_ADA_002"] = "text-embedding-ada-002";
|
||||
})(E_OPENAI_MODEL || (E_OPENAI_MODEL = {}));
|
||||
//# sourceMappingURL=data:application/json;base64,eyJ2ZXJzaW9uIjozLCJmaWxlIjoib3BlbmFpLW1vZGVscy5qcyIsInNvdXJjZVJvb3QiOiIiLCJzb3VyY2VzIjpbIi4uLy4uLy4uL3NyYy9tb2RlbHMvY2FjaGUvb3BlbmFpLW1vZGVscy50cyJdLCJuYW1lcyI6W10sIm1hcHBpbmdzIjoiQUFBQSxNQUFNLENBQU4sSUFBWSxjQThIWDtBQTlIRCxXQUFZLGNBQWM7SUFDeEIsaURBQStCLENBQUE7SUFDL0IsdUNBQXFCLENBQUE7SUFDckIsdURBQXFDLENBQUE7SUFDckMscURBQW1DLENBQUE7SUFDbkMsMkNBQXlCLENBQUE7SUFDekIsMkVBQXlELENBQUE7SUFDekQscURBQW1DLENBQUE7SUFDbkMsMkVBQXlELENBQUE7SUFDekQsbURBQWlDLENBQUE7SUFDakMsbURBQWlDLENBQUE7SUFDakMseUVBQXVELENBQUE7SUFDdkQsbUZBQWlFLENBQUE7SUFDakUsNkNBQTJCLENBQUE7SUFDM0IsNkNBQTJCLENBQUE7SUFDM0IsaUVBQStDLENBQUE7SUFDL0MsaUVBQStDLENBQUE7SUFDL0MsNkNBQTJCLENBQUE7SUFDM0IsaURBQStCLENBQUE7SUFDL0IsdURBQXFDLENBQUE7SUFDckMseUVBQXVELENBQUE7SUFDdkQseUVBQXVELENBQUE7SUFDdkQsaUVBQStDLENBQUE7SUFDL0MsbUVBQWlELENBQUE7SUFDakQsaUVBQStDLENBQUE7SUFDL0MsbURBQWlDLENBQUE7SUFDakMseUVBQXVELENBQUE7SUFDdkQseUNBQXVCLENBQUE7SUFDdkIsK0RBQTZDLENBQUE7SUFDN0MseUVBQXVELENBQUE7SUFDdkQsbURBQWlDLENBQUE7SUFDakMsK0RBQTZDLENBQUE7SUFDN0MscUVBQW1ELENBQUE7SUFDbkQsMkVBQXlELENBQUE7SUFDekQseUVBQXVELENBQUE7SUFDdkQsaUZBQStELENBQUE7SUFDL0QsaUdBQStFLENBQUE7SUFDL0UsMkZBQXlFLENBQUE7SUFDekUsMkdBQXlGLENBQUE7SUFDekYscUdBQW1GLENBQUE7SUFDbkYsdURBQXFDLENBQUE7SUFDckMsaUNBQWUsQ0FBQTtJQUNmLHFGQUFtRSxDQUFBO0lBQ25FLCtFQUE2RCxDQUFBO0lBQzdELDJDQUF5QixDQUFBO0lBQ3pCLGlFQUErQyxDQUFBO0lBQy9DLCtEQUE2QyxDQUFBO0lBQzdDLHVHQUFxRixDQUFBO0lBQ3JGLGlGQUErRCxDQUFBO0lBQy9ELCtEQUE2QyxDQUFBO0lBQzdDLHlFQUF1RCxDQUFBO0lBQ3ZELCtEQUE2QyxDQUFBO0lBQzdDLHlDQUF1QixDQUFBO0lBQ3ZCLDJEQUF5QyxDQUFBO0lBQ3pDLHVEQUFxQyxDQUFBO0lBQ3JDLGlFQUErQyxDQUFBO0lBQy9DLGlDQUFlLENBQUE7SUFDZiwyQ0FBeUIsQ0FBQTtJQUN6QixpRUFBK0MsQ0FBQTtJQUMvQywyQ0FBeUIsQ0FBQTtJQUN6QiwyRUFBeUQsQ0FBQTtJQUN6RCxxREFBbUMsQ0FBQTtJQUNuQywyRUFBeUQsQ0FBQTtJQUN6RCxxREFBbUMsQ0FBQTtJQUNuQyxtREFBaUMsQ0FBQTtJQUNqQyxpR0FBK0UsQ0FBQTtJQUMvRSwyRkFBeUUsQ0FBQTtJQUN6RSx1RUFBcUQsQ0FBQTtJQUNyRCwrRUFBNkQsQ0FBQTtJQUM3RCw2RkFBMkUsQ0FBQTtJQUMzRSwrREFBNkMsQ0FBQTtJQUM3Qyw2REFBMkMsQ0FBQTtJQUMzQyx1Q0FBcUIsQ0FBQTtJQUNyQix1RUFBcUQsQ0FBQTtJQUNyRCxpREFBK0IsQ0FBQTtJQUMvQix1RUFBcUQsQ0FBQTtJQUNyRCxpREFBK0IsQ0FBQTtJQUMvQixxRUFBbUQsQ0FBQTtJQUNuRCxxREFBbUMsQ0FBQTtJQUNuQywyRUFBeUQsQ0FBQTtJQUN6RCwrQ0FBNkIsQ0FBQTtJQUM3QixtREFBaUMsQ0FBQTtJQUNqQyw2REFBMkMsQ0FBQTtJQUMzQyxxRUFBbUQsQ0FBQTtJQUNuRCwrQ0FBNkIsQ0FBQTtJQUM3Qix5REFBdUMsQ0FBQTtJQUN2QywrRUFBNkQsQ0FBQTtJQUM3RCw2REFBMkMsQ0FBQTtJQUMzQywrREFBNkMsQ0FBQTtJQUM3QyxxRkFBbUUsQ0FBQTtJQUNuRSx5Q0FBdUIsQ0FBQTtJQUN2QixpREFBK0IsQ0FBQTtJQUMvQixtRkFBaUUsQ0FBQTtJQUNqRSxtRUFBaUQsQ0FBQTtJQUNqRCxpRUFBK0MsQ0FBQTtJQUMvQywyQ0FBeUIsQ0FBQTtJQUN6Qix1REFBcUMsQ0FBQTtJQUNyQyxpRUFBK0MsQ0FBQTtJQUMvQywrREFBNkMsQ0FBQTtJQUM3Qyx1REFBcUMsQ0FBQTtJQUNyQyxpRUFBK0MsQ0FBQTtJQUMvQywyQ0FBeUIsQ0FBQTtJQUN6Qix5RUFBdUQsQ0FBQTtJQUN2RCxtREFBaUMsQ0FBQTtJQUNqQyxtRUFBaUQsQ0FBQTtJQUNqRCwrRkFBNkUsQ0FBQTtJQUM3RSwrRkFBNkUsQ0FBQTtJQUM3RSxpRkFBK0QsQ0FBQTtJQUMvRCxpRkFBK0QsQ0FBQTtJQUMvRCxxRkFBbUUsQ0FBQTtJQUNuRSwrRUFBNkQsQ0FBQTtJQUM3RCxxRUFBbUQsQ0FBQTtJQUNuRCx1REFBcUMsQ0FBQTtJQUNyQyx1REFBcUMsQ0FBQTtJQUNyQyw2REFBMkMsQ0FBQTtJQUMzQyx1REFBcUMsQ0FBQTtJQUNyQyx1RUFBcUQsQ0FBQTtJQUNyRCw2RkFBMkUsQ0FBQTtJQUMzRSxtRUFBaUQsQ0FBQTtJQUNqRCxpRUFBK0MsQ0FBQTtJQUMvQyxtREFBaUMsQ0FBQTtJQUNqQyx5RUFBdUQsQ0FBQTtJQUN2RCwrREFBNkMsQ0FBQTtJQUM3Qyx1Q0FBcUIsQ0FBQTtJQUNyQiwrQ0FBNkIsQ0FBQTtJQUM3Qix5RUFBdUQsQ0FBQTtBQUN6RCxDQUFDLEVBOUhXLGNBQWMsS0FBZCxjQUFjLFFBOEh6QiJ9
|
||||
//# sourceMappingURL=data:application/json;base64,eyJ2ZXJzaW9uIjozLCJmaWxlIjoib3BlbmFpLW1vZGVscy5qcyIsInNvdXJjZVJvb3QiOiIiLCJzb3VyY2VzIjpbIi4uLy4uLy4uL3NyYy9tb2RlbHMvY2FjaGUvb3BlbmFpLW1vZGVscy50cyJdLCJuYW1lcyI6W10sIm1hcHBpbmdzIjoiQUFBQSxNQUFNLENBQU4sSUFBWSxjQTJIWDtBQTNIRCxXQUFZLGNBQWM7SUFDeEIsaURBQStCLENBQUE7SUFDL0IsdUNBQXFCLENBQUE7SUFDckIsdURBQXFDLENBQUE7SUFDckMscURBQW1DLENBQUE7SUFDbkMsMkNBQXlCLENBQUE7SUFDekIsMkVBQXlELENBQUE7SUFDekQscURBQW1DLENBQUE7SUFDbkMsMkVBQXlELENBQUE7SUFDekQsbURBQWlDLENBQUE7SUFDakMsbURBQWlDLENBQUE7SUFDakMseUVBQXVELENBQUE7SUFDdkQsbUZBQWlFLENBQUE7SUFDakUsNkNBQTJCLENBQUE7SUFDM0IsNkNBQTJCLENBQUE7SUFDM0IsaUVBQStDLENBQUE7SUFDL0MsNkNBQTJCLENBQUE7SUFDM0IsaURBQStCLENBQUE7SUFDL0IsdURBQXFDLENBQUE7SUFDckMseUVBQXVELENBQUE7SUFDdkQseUVBQXVELENBQUE7SUFDdkQsaUVBQStDLENBQUE7SUFDL0MsbURBQWlDLENBQUE7SUFDakMseUVBQXVELENBQUE7SUFDdkQseUNBQXVCLENBQUE7SUFDdkIsK0RBQTZDLENBQUE7SUFDN0MseUVBQXVELENBQUE7SUFDdkQsbURBQWlDLENBQUE7SUFDakMsK0RBQTZDLENBQUE7SUFDN0MscUVBQW1ELENBQUE7SUFDbkQsMkVBQXlELENBQUE7SUFDekQseUVBQXVELENBQUE7SUFDdkQsaUZBQStELENBQUE7SUFDL0QsaUdBQStFLENBQUE7SUFDL0UsMkZBQXlFLENBQUE7SUFDekUsMkdBQXlGLENBQUE7SUFDekYscUdBQW1GLENBQUE7SUFDbkYsdURBQXFDLENBQUE7SUFDckMsaUNBQWUsQ0FBQTtJQUNmLHFGQUFtRSxDQUFBO0lBQ25FLCtFQUE2RCxDQUFBO0lBQzdELDJDQUF5QixDQUFBO0lBQ3pCLGlFQUErQyxDQUFBO0lBQy9DLCtEQUE2QyxDQUFBO0lBQzdDLHVHQUFxRixDQUFBO0lBQ3JGLGlGQUErRCxDQUFBO0lBQy9ELCtEQUE2QyxDQUFBO0lBQzdDLHlFQUF1RCxDQUFBO0lBQ3ZELCtEQUE2QyxDQUFBO0lBQzdDLHlDQUF1QixDQUFBO0lBQ3ZCLDJEQUF5QyxDQUFBO0lBQ3pDLHVEQUFxQyxDQUFBO0lBQ3JDLGlFQUErQyxDQUFBO0lBQy9DLGlDQUFlLENBQUE7SUFDZiwyQ0FBeUIsQ0FBQTtJQUN6QixpRUFBK0MsQ0FBQTtJQUMvQywyQ0FBeUIsQ0FBQTtJQUN6QiwyRUFBeUQsQ0FBQTtJQUN6RCxxREFBbUMsQ0FBQTtJQUNuQywyRUFBeUQsQ0FBQTtJQUN6RCxxREFBbUMsQ0FBQTtJQUNuQyxtREFBaUMsQ0FBQTtJQUNqQyxpR0FBK0UsQ0FBQTtJQUMvRSwyRkFBeUUsQ0FBQTtJQUN6RSx1RUFBcUQsQ0FBQTtJQUNyRCwrRUFBNkQsQ0FBQTtJQUM3RCw2RkFBMkUsQ0FBQTtJQUMzRSwrREFBNkMsQ0FBQTtJQUM3Qyw2REFBMkMsQ0FBQTtJQUMzQyx1Q0FBcUIsQ0FBQTtJQUNyQix1RUFBcUQsQ0FBQTtJQUNyRCxpREFBK0IsQ0FBQTtJQUMvQix1RUFBcUQsQ0FBQTtJQUNyRCxpREFBK0IsQ0FBQTtJQUMvQixxRUFBbUQsQ0FBQTtJQUNuRCxxREFBbUMsQ0FBQTtJQUNuQywyRUFBeUQsQ0FBQTtJQUN6RCwrQ0FBNkIsQ0FBQTtJQUM3QixtREFBaUMsQ0FBQTtJQUNqQyw2REFBMkMsQ0FBQTtJQUMzQyxxRUFBbUQsQ0FBQTtJQUNuRCwrQ0FBNkIsQ0FBQTtJQUM3Qix5REFBdUMsQ0FBQTtJQUN2QywrRUFBNkQsQ0FBQTtJQUM3RCw2REFBMkMsQ0FBQTtJQUMzQywrREFBNkMsQ0FBQTtJQUM3QyxxRkFBbUUsQ0FBQTtJQUNuRSx5Q0FBdUIsQ0FBQTtJQUN2QixpREFBK0IsQ0FBQTtJQUMvQixtRkFBaUUsQ0FBQTtJQUNqRSxtRUFBaUQsQ0FBQTtJQUNqRCxpRUFBK0MsQ0FBQTtJQUMvQywyQ0FBeUIsQ0FBQTtJQUN6Qix1REFBcUMsQ0FBQTtJQUNyQyxpRUFBK0MsQ0FBQTtJQUMvQywrREFBNkMsQ0FBQTtJQUM3Qyx1REFBcUMsQ0FBQTtJQUNyQyxpRUFBK0MsQ0FBQTtJQUMvQywyQ0FBeUIsQ0FBQTtJQUN6Qix5RUFBdUQsQ0FBQTtJQUN2RCxtREFBaUMsQ0FBQTtJQUNqQyxtRUFBaUQsQ0FBQTtJQUNqRCwrRkFBNkUsQ0FBQTtJQUM3RSwrRkFBNkUsQ0FBQTtJQUM3RSxpRkFBK0QsQ0FBQTtJQUMvRCxpRkFBK0QsQ0FBQTtJQUMvRCxxRkFBbUUsQ0FBQTtJQUNuRSwrRUFBNkQsQ0FBQTtJQUM3RCxxRUFBbUQsQ0FBQTtJQUNuRCx1REFBcUMsQ0FBQTtJQUNyQyx1REFBcUMsQ0FBQTtJQUNyQyw2REFBMkMsQ0FBQTtJQUMzQyx1REFBcUMsQ0FBQTtJQUNyQyx1RUFBcUQsQ0FBQTtJQUNyRCw2RkFBMkUsQ0FBQTtJQUMzRSxtRUFBaUQsQ0FBQTtJQUNqRCxpRUFBK0MsQ0FBQTtJQUMvQyxtREFBaUMsQ0FBQTtJQUNqQyx5RUFBdUQsQ0FBQTtJQUN2RCwrREFBNkMsQ0FBQTtJQUM3Qyx1Q0FBcUIsQ0FBQTtJQUNyQiwrQ0FBNkIsQ0FBQTtJQUM3Qix5RUFBdUQsQ0FBQTtBQUN6RCxDQUFDLEVBM0hXLGNBQWMsS0FBZCxjQUFjLFFBMkh6QiJ9
|
||||
@ -1,4 +1,6 @@
|
||||
export declare enum E_OPENROUTER_MODEL {
|
||||
MODEL_KWAIPILOT_KAT_CODER_PRO_V2 = "kwaipilot/kat-coder-pro-v2",
|
||||
MODEL_REKA_REKA_EDGE = "reka/reka-edge",
|
||||
MODEL_XIAOMI_MIMO_V2_OMNI = "xiaomi/mimo-v2-omni",
|
||||
MODEL_XIAOMI_MIMO_V2_PRO = "xiaomi/mimo-v2-pro",
|
||||
MODEL_MINIMAX_MINIMAX_M2_7 = "minimax/minimax-m2.7",
|
||||
@ -51,7 +53,6 @@ export declare enum E_OPENROUTER_MODEL {
|
||||
MODEL_OPENAI_GPT_AUDIO_MINI = "openai/gpt-audio-mini",
|
||||
MODEL_Z_AI_GLM_4_7_FLASH = "z-ai/glm-4.7-flash",
|
||||
MODEL_OPENAI_GPT_5_2_CODEX = "openai/gpt-5.2-codex",
|
||||
MODEL_ALLENAI_MOLMO_2_8B = "allenai/molmo-2-8b",
|
||||
MODEL_ALLENAI_OLMO_3_1_32B_INSTRUCT = "allenai/olmo-3.1-32b-instruct",
|
||||
MODEL_BYTEDANCE_SEED_SEED_1_6_FLASH = "bytedance-seed/seed-1.6-flash",
|
||||
MODEL_BYTEDANCE_SEED_SEED_1_6 = "bytedance-seed/seed-1.6",
|
||||
@ -85,11 +86,8 @@ export declare enum E_OPENROUTER_MODEL {
|
||||
MODEL_PRIME_INTELLECT_INTELLECT_3 = "prime-intellect/intellect-3",
|
||||
MODEL_ANTHROPIC_CLAUDE_OPUS_4_5 = "anthropic/claude-opus-4.5",
|
||||
MODEL_ALLENAI_OLMO_3_32B_THINK = "allenai/olmo-3-32b-think",
|
||||
MODEL_ALLENAI_OLMO_3_7B_INSTRUCT = "allenai/olmo-3-7b-instruct",
|
||||
MODEL_ALLENAI_OLMO_3_7B_THINK = "allenai/olmo-3-7b-think",
|
||||
MODEL_GOOGLE_GEMINI_3_PRO_IMAGE_PREVIEW = "google/gemini-3-pro-image-preview",
|
||||
MODEL_X_AI_GROK_4_1_FAST = "x-ai/grok-4.1-fast",
|
||||
MODEL_GOOGLE_GEMINI_3_PRO_PREVIEW = "google/gemini-3-pro-preview",
|
||||
MODEL_DEEPCOGITO_COGITO_V2_1_671B = "deepcogito/cogito-v2.1-671b",
|
||||
MODEL_OPENAI_GPT_5_1 = "openai/gpt-5.1",
|
||||
MODEL_OPENAI_GPT_5_1_CHAT = "openai/gpt-5.1-chat",
|
||||
@ -212,7 +210,6 @@ export declare enum E_OPENROUTER_MODEL {
|
||||
MODEL_ARCEE_AI_VIRTUOSO_LARGE = "arcee-ai/virtuoso-large",
|
||||
MODEL_ARCEE_AI_CODER_LARGE = "arcee-ai/coder-large",
|
||||
MODEL_INCEPTION_MERCURY_CODER = "inception/mercury-coder",
|
||||
MODEL_QWEN_QWEN3_4B_FREE = "qwen/qwen3-4b:free",
|
||||
MODEL_META_LLAMA_LLAMA_GUARD_4_12B = "meta-llama/llama-guard-4-12b",
|
||||
MODEL_QWEN_QWEN3_30B_A3B = "qwen/qwen3-30b-a3b",
|
||||
MODEL_QWEN_QWEN3_8B = "qwen/qwen3-8b",
|
||||
@ -230,12 +227,12 @@ export declare enum E_OPENROUTER_MODEL {
|
||||
MODEL_ALFREDPROS_CODELLAMA_7B_INSTRUCT_SOLIDITY = "alfredpros/codellama-7b-instruct-solidity",
|
||||
MODEL_X_AI_GROK_3_MINI_BETA = "x-ai/grok-3-mini-beta",
|
||||
MODEL_X_AI_GROK_3_BETA = "x-ai/grok-3-beta",
|
||||
MODEL_NVIDIA_LLAMA_3_1_NEMOTRON_ULTRA_253B_V1 = "nvidia/llama-3.1-nemotron-ultra-253b-v1",
|
||||
MODEL_META_LLAMA_LLAMA_4_MAVERICK = "meta-llama/llama-4-maverick",
|
||||
MODEL_META_LLAMA_LLAMA_4_SCOUT = "meta-llama/llama-4-scout",
|
||||
MODEL_QWEN_QWEN2_5_VL_32B_INSTRUCT = "qwen/qwen2.5-vl-32b-instruct",
|
||||
MODEL_DEEPSEEK_DEEPSEEK_CHAT_V3_0324 = "deepseek/deepseek-chat-v3-0324",
|
||||
MODEL_OPENAI_O1_PRO = "openai/o1-pro",
|
||||
MODEL_MISTRALAI_MISTRAL_SMALL_3_1_24B_INSTRUCT_FREE = "mistralai/mistral-small-3.1-24b-instruct:free",
|
||||
MODEL_MISTRALAI_MISTRAL_SMALL_3_1_24B_INSTRUCT = "mistralai/mistral-small-3.1-24b-instruct",
|
||||
MODEL_ALLENAI_OLMO_2_0325_32B_INSTRUCT = "allenai/olmo-2-0325-32b-instruct",
|
||||
MODEL_GOOGLE_GEMMA_3_4B_IT_FREE = "google/gemma-3-4b-it:free",
|
||||
@ -297,45 +294,43 @@ export declare enum E_OPENROUTER_MODEL {
|
||||
MODEL_ANTHROPIC_CLAUDE_3_5_SONNET = "anthropic/claude-3.5-sonnet",
|
||||
MODEL_QWEN_QWEN_2_5_7B_INSTRUCT = "qwen/qwen-2.5-7b-instruct",
|
||||
MODEL_NVIDIA_LLAMA_3_1_NEMOTRON_70B_INSTRUCT = "nvidia/llama-3.1-nemotron-70b-instruct",
|
||||
MODEL_INFLECTION_INFLECTION_3_PI = "inflection/inflection-3-pi",
|
||||
MODEL_INFLECTION_INFLECTION_3_PRODUCTIVITY = "inflection/inflection-3-productivity",
|
||||
MODEL_INFLECTION_INFLECTION_3_PI = "inflection/inflection-3-pi",
|
||||
MODEL_THEDRUMMER_ROCINANTE_12B = "thedrummer/rocinante-12b",
|
||||
MODEL_META_LLAMA_LLAMA_3_2_3B_INSTRUCT_FREE = "meta-llama/llama-3.2-3b-instruct:free",
|
||||
MODEL_META_LLAMA_LLAMA_3_2_3B_INSTRUCT = "meta-llama/llama-3.2-3b-instruct",
|
||||
MODEL_META_LLAMA_LLAMA_3_2_1B_INSTRUCT = "meta-llama/llama-3.2-1b-instruct",
|
||||
MODEL_META_LLAMA_LLAMA_3_2_11B_VISION_INSTRUCT = "meta-llama/llama-3.2-11b-vision-instruct",
|
||||
MODEL_QWEN_QWEN_2_5_72B_INSTRUCT = "qwen/qwen-2.5-72b-instruct",
|
||||
MODEL_COHERE_COMMAND_R_08_2024 = "cohere/command-r-08-2024",
|
||||
MODEL_COHERE_COMMAND_R_PLUS_08_2024 = "cohere/command-r-plus-08-2024",
|
||||
MODEL_COHERE_COMMAND_R_08_2024 = "cohere/command-r-08-2024",
|
||||
MODEL_SAO10K_L3_1_EURYALE_70B = "sao10k/l3.1-euryale-70b",
|
||||
MODEL_QWEN_QWEN_2_5_VL_7B_INSTRUCT = "qwen/qwen-2.5-vl-7b-instruct",
|
||||
MODEL_NOUSRESEARCH_HERMES_3_LLAMA_3_1_70B = "nousresearch/hermes-3-llama-3.1-70b",
|
||||
MODEL_NOUSRESEARCH_HERMES_3_LLAMA_3_1_405B_FREE = "nousresearch/hermes-3-llama-3.1-405b:free",
|
||||
MODEL_NOUSRESEARCH_HERMES_3_LLAMA_3_1_405B = "nousresearch/hermes-3-llama-3.1-405b",
|
||||
MODEL_SAO10K_L3_LUNARIS_8B = "sao10k/l3-lunaris-8b",
|
||||
MODEL_OPENAI_GPT_4O_2024_08_06 = "openai/gpt-4o-2024-08-06",
|
||||
MODEL_META_LLAMA_LLAMA_3_1_405B = "meta-llama/llama-3.1-405b",
|
||||
MODEL_META_LLAMA_LLAMA_3_1_8B_INSTRUCT = "meta-llama/llama-3.1-8b-instruct",
|
||||
MODEL_META_LLAMA_LLAMA_3_1_70B_INSTRUCT = "meta-llama/llama-3.1-70b-instruct",
|
||||
MODEL_MISTRALAI_MISTRAL_NEMO = "mistralai/mistral-nemo",
|
||||
MODEL_OPENAI_GPT_4O_MINI_2024_07_18 = "openai/gpt-4o-mini-2024-07-18",
|
||||
MODEL_OPENAI_GPT_4O_MINI = "openai/gpt-4o-mini",
|
||||
MODEL_OPENAI_GPT_4O_MINI_2024_07_18 = "openai/gpt-4o-mini-2024-07-18",
|
||||
MODEL_GOOGLE_GEMMA_2_27B_IT = "google/gemma-2-27b-it",
|
||||
MODEL_GOOGLE_GEMMA_2_9B_IT = "google/gemma-2-9b-it",
|
||||
MODEL_SAO10K_L3_EURYALE_70B = "sao10k/l3-euryale-70b",
|
||||
MODEL_NOUSRESEARCH_HERMES_2_PRO_LLAMA_3_8B = "nousresearch/hermes-2-pro-llama-3-8b",
|
||||
MODEL_OPENAI_GPT_4O_2024_05_13 = "openai/gpt-4o-2024-05-13",
|
||||
MODEL_OPENAI_GPT_4O = "openai/gpt-4o",
|
||||
MODEL_OPENAI_GPT_4O_EXTENDED = "openai/gpt-4o:extended",
|
||||
MODEL_META_LLAMA_LLAMA_3_70B_INSTRUCT = "meta-llama/llama-3-70b-instruct",
|
||||
MODEL_OPENAI_GPT_4O_2024_05_13 = "openai/gpt-4o-2024-05-13",
|
||||
MODEL_META_LLAMA_LLAMA_3_8B_INSTRUCT = "meta-llama/llama-3-8b-instruct",
|
||||
MODEL_META_LLAMA_LLAMA_3_70B_INSTRUCT = "meta-llama/llama-3-70b-instruct",
|
||||
MODEL_MISTRALAI_MIXTRAL_8X22B_INSTRUCT = "mistralai/mixtral-8x22b-instruct",
|
||||
MODEL_MICROSOFT_WIZARDLM_2_8X22B = "microsoft/wizardlm-2-8x22b",
|
||||
MODEL_OPENAI_GPT_4_TURBO = "openai/gpt-4-turbo",
|
||||
MODEL_ANTHROPIC_CLAUDE_3_HAIKU = "anthropic/claude-3-haiku",
|
||||
MODEL_MISTRALAI_MISTRAL_LARGE = "mistralai/mistral-large",
|
||||
MODEL_OPENAI_GPT_3_5_TURBO_0613 = "openai/gpt-3.5-turbo-0613",
|
||||
MODEL_OPENAI_GPT_4_TURBO_PREVIEW = "openai/gpt-4-turbo-preview",
|
||||
MODEL_OPENAI_GPT_3_5_TURBO_0613 = "openai/gpt-3.5-turbo-0613",
|
||||
MODEL_MISTRALAI_MIXTRAL_8X7B_INSTRUCT = "mistralai/mixtral-8x7b-instruct",
|
||||
MODEL_ALPINDALE_GOLIATH_120B = "alpindale/goliath-120b",
|
||||
MODEL_OPENROUTER_AUTO = "openrouter/auto",
|
||||
@ -346,7 +341,7 @@ export declare enum E_OPENROUTER_MODEL {
|
||||
MODEL_MANCER_WEAVER = "mancer/weaver",
|
||||
MODEL_UNDI95_REMM_SLERP_L2_13B = "undi95/remm-slerp-l2-13b",
|
||||
MODEL_GRYPHE_MYTHOMAX_L2_13B = "gryphe/mythomax-l2-13b",
|
||||
MODEL_OPENAI_GPT_4_0314 = "openai/gpt-4-0314",
|
||||
MODEL_OPENAI_GPT_4 = "openai/gpt-4",
|
||||
MODEL_OPENAI_GPT_4_0314 = "openai/gpt-4-0314",
|
||||
MODEL_OPENAI_GPT_3_5_TURBO = "openai/gpt-3.5-turbo"
|
||||
}
|
||||
|
||||
File diff suppressed because one or more lines are too long
File diff suppressed because one or more lines are too long
File diff suppressed because one or more lines are too long
175029
packages/kbot/dist/main_node.js
vendored
175029
packages/kbot/dist/main_node.js
vendored
File diff suppressed because one or more lines are too long
@ -13,15 +13,12 @@ export enum E_OPENAI_MODEL {
|
||||
MODEL_GPT_3_5_TURBO_INSTRUCT_0914 = "gpt-3.5-turbo-instruct-0914",
|
||||
MODEL_DALL_E_3 = "dall-e-3",
|
||||
MODEL_DALL_E_2 = "dall-e-2",
|
||||
MODEL_GPT_4_1106_PREVIEW = "gpt-4-1106-preview",
|
||||
MODEL_GPT_3_5_TURBO_1106 = "gpt-3.5-turbo-1106",
|
||||
MODEL_TTS_1_HD = "tts-1-hd",
|
||||
MODEL_TTS_1_1106 = "tts-1-1106",
|
||||
MODEL_TTS_1_HD_1106 = "tts-1-hd-1106",
|
||||
MODEL_TEXT_EMBEDDING_3_SMALL = "text-embedding-3-small",
|
||||
MODEL_TEXT_EMBEDDING_3_LARGE = "text-embedding-3-large",
|
||||
MODEL_GPT_4_0125_PREVIEW = "gpt-4-0125-preview",
|
||||
MODEL_GPT_4_TURBO_PREVIEW = "gpt-4-turbo-preview",
|
||||
MODEL_GPT_3_5_TURBO_0125 = "gpt-3.5-turbo-0125",
|
||||
MODEL_GPT_4_TURBO = "gpt-4-turbo",
|
||||
MODEL_GPT_4_TURBO_2024_04_09 = "gpt-4-turbo-2024-04-09",
|
||||
|
||||
@ -1,4 +1,6 @@
|
||||
export enum E_OPENROUTER_MODEL {
|
||||
MODEL_KWAIPILOT_KAT_CODER_PRO_V2 = "kwaipilot/kat-coder-pro-v2",
|
||||
MODEL_REKA_REKA_EDGE = "reka/reka-edge",
|
||||
MODEL_XIAOMI_MIMO_V2_OMNI = "xiaomi/mimo-v2-omni",
|
||||
MODEL_XIAOMI_MIMO_V2_PRO = "xiaomi/mimo-v2-pro",
|
||||
MODEL_MINIMAX_MINIMAX_M2_7 = "minimax/minimax-m2.7",
|
||||
@ -51,7 +53,6 @@ export enum E_OPENROUTER_MODEL {
|
||||
MODEL_OPENAI_GPT_AUDIO_MINI = "openai/gpt-audio-mini",
|
||||
MODEL_Z_AI_GLM_4_7_FLASH = "z-ai/glm-4.7-flash",
|
||||
MODEL_OPENAI_GPT_5_2_CODEX = "openai/gpt-5.2-codex",
|
||||
MODEL_ALLENAI_MOLMO_2_8B = "allenai/molmo-2-8b",
|
||||
MODEL_ALLENAI_OLMO_3_1_32B_INSTRUCT = "allenai/olmo-3.1-32b-instruct",
|
||||
MODEL_BYTEDANCE_SEED_SEED_1_6_FLASH = "bytedance-seed/seed-1.6-flash",
|
||||
MODEL_BYTEDANCE_SEED_SEED_1_6 = "bytedance-seed/seed-1.6",
|
||||
@ -85,11 +86,8 @@ export enum E_OPENROUTER_MODEL {
|
||||
MODEL_PRIME_INTELLECT_INTELLECT_3 = "prime-intellect/intellect-3",
|
||||
MODEL_ANTHROPIC_CLAUDE_OPUS_4_5 = "anthropic/claude-opus-4.5",
|
||||
MODEL_ALLENAI_OLMO_3_32B_THINK = "allenai/olmo-3-32b-think",
|
||||
MODEL_ALLENAI_OLMO_3_7B_INSTRUCT = "allenai/olmo-3-7b-instruct",
|
||||
MODEL_ALLENAI_OLMO_3_7B_THINK = "allenai/olmo-3-7b-think",
|
||||
MODEL_GOOGLE_GEMINI_3_PRO_IMAGE_PREVIEW = "google/gemini-3-pro-image-preview",
|
||||
MODEL_X_AI_GROK_4_1_FAST = "x-ai/grok-4.1-fast",
|
||||
MODEL_GOOGLE_GEMINI_3_PRO_PREVIEW = "google/gemini-3-pro-preview",
|
||||
MODEL_DEEPCOGITO_COGITO_V2_1_671B = "deepcogito/cogito-v2.1-671b",
|
||||
MODEL_OPENAI_GPT_5_1 = "openai/gpt-5.1",
|
||||
MODEL_OPENAI_GPT_5_1_CHAT = "openai/gpt-5.1-chat",
|
||||
@ -212,7 +210,6 @@ export enum E_OPENROUTER_MODEL {
|
||||
MODEL_ARCEE_AI_VIRTUOSO_LARGE = "arcee-ai/virtuoso-large",
|
||||
MODEL_ARCEE_AI_CODER_LARGE = "arcee-ai/coder-large",
|
||||
MODEL_INCEPTION_MERCURY_CODER = "inception/mercury-coder",
|
||||
MODEL_QWEN_QWEN3_4B_FREE = "qwen/qwen3-4b:free",
|
||||
MODEL_META_LLAMA_LLAMA_GUARD_4_12B = "meta-llama/llama-guard-4-12b",
|
||||
MODEL_QWEN_QWEN3_30B_A3B = "qwen/qwen3-30b-a3b",
|
||||
MODEL_QWEN_QWEN3_8B = "qwen/qwen3-8b",
|
||||
@ -230,12 +227,12 @@ export enum E_OPENROUTER_MODEL {
|
||||
MODEL_ALFREDPROS_CODELLAMA_7B_INSTRUCT_SOLIDITY = "alfredpros/codellama-7b-instruct-solidity",
|
||||
MODEL_X_AI_GROK_3_MINI_BETA = "x-ai/grok-3-mini-beta",
|
||||
MODEL_X_AI_GROK_3_BETA = "x-ai/grok-3-beta",
|
||||
MODEL_NVIDIA_LLAMA_3_1_NEMOTRON_ULTRA_253B_V1 = "nvidia/llama-3.1-nemotron-ultra-253b-v1",
|
||||
MODEL_META_LLAMA_LLAMA_4_MAVERICK = "meta-llama/llama-4-maverick",
|
||||
MODEL_META_LLAMA_LLAMA_4_SCOUT = "meta-llama/llama-4-scout",
|
||||
MODEL_QWEN_QWEN2_5_VL_32B_INSTRUCT = "qwen/qwen2.5-vl-32b-instruct",
|
||||
MODEL_DEEPSEEK_DEEPSEEK_CHAT_V3_0324 = "deepseek/deepseek-chat-v3-0324",
|
||||
MODEL_OPENAI_O1_PRO = "openai/o1-pro",
|
||||
MODEL_MISTRALAI_MISTRAL_SMALL_3_1_24B_INSTRUCT_FREE = "mistralai/mistral-small-3.1-24b-instruct:free",
|
||||
MODEL_MISTRALAI_MISTRAL_SMALL_3_1_24B_INSTRUCT = "mistralai/mistral-small-3.1-24b-instruct",
|
||||
MODEL_ALLENAI_OLMO_2_0325_32B_INSTRUCT = "allenai/olmo-2-0325-32b-instruct",
|
||||
MODEL_GOOGLE_GEMMA_3_4B_IT_FREE = "google/gemma-3-4b-it:free",
|
||||
@ -297,45 +294,43 @@ export enum E_OPENROUTER_MODEL {
|
||||
MODEL_ANTHROPIC_CLAUDE_3_5_SONNET = "anthropic/claude-3.5-sonnet",
|
||||
MODEL_QWEN_QWEN_2_5_7B_INSTRUCT = "qwen/qwen-2.5-7b-instruct",
|
||||
MODEL_NVIDIA_LLAMA_3_1_NEMOTRON_70B_INSTRUCT = "nvidia/llama-3.1-nemotron-70b-instruct",
|
||||
MODEL_INFLECTION_INFLECTION_3_PI = "inflection/inflection-3-pi",
|
||||
MODEL_INFLECTION_INFLECTION_3_PRODUCTIVITY = "inflection/inflection-3-productivity",
|
||||
MODEL_INFLECTION_INFLECTION_3_PI = "inflection/inflection-3-pi",
|
||||
MODEL_THEDRUMMER_ROCINANTE_12B = "thedrummer/rocinante-12b",
|
||||
MODEL_META_LLAMA_LLAMA_3_2_3B_INSTRUCT_FREE = "meta-llama/llama-3.2-3b-instruct:free",
|
||||
MODEL_META_LLAMA_LLAMA_3_2_3B_INSTRUCT = "meta-llama/llama-3.2-3b-instruct",
|
||||
MODEL_META_LLAMA_LLAMA_3_2_1B_INSTRUCT = "meta-llama/llama-3.2-1b-instruct",
|
||||
MODEL_META_LLAMA_LLAMA_3_2_11B_VISION_INSTRUCT = "meta-llama/llama-3.2-11b-vision-instruct",
|
||||
MODEL_QWEN_QWEN_2_5_72B_INSTRUCT = "qwen/qwen-2.5-72b-instruct",
|
||||
MODEL_COHERE_COMMAND_R_08_2024 = "cohere/command-r-08-2024",
|
||||
MODEL_COHERE_COMMAND_R_PLUS_08_2024 = "cohere/command-r-plus-08-2024",
|
||||
MODEL_COHERE_COMMAND_R_08_2024 = "cohere/command-r-08-2024",
|
||||
MODEL_SAO10K_L3_1_EURYALE_70B = "sao10k/l3.1-euryale-70b",
|
||||
MODEL_QWEN_QWEN_2_5_VL_7B_INSTRUCT = "qwen/qwen-2.5-vl-7b-instruct",
|
||||
MODEL_NOUSRESEARCH_HERMES_3_LLAMA_3_1_70B = "nousresearch/hermes-3-llama-3.1-70b",
|
||||
MODEL_NOUSRESEARCH_HERMES_3_LLAMA_3_1_405B_FREE = "nousresearch/hermes-3-llama-3.1-405b:free",
|
||||
MODEL_NOUSRESEARCH_HERMES_3_LLAMA_3_1_405B = "nousresearch/hermes-3-llama-3.1-405b",
|
||||
MODEL_SAO10K_L3_LUNARIS_8B = "sao10k/l3-lunaris-8b",
|
||||
MODEL_OPENAI_GPT_4O_2024_08_06 = "openai/gpt-4o-2024-08-06",
|
||||
MODEL_META_LLAMA_LLAMA_3_1_405B = "meta-llama/llama-3.1-405b",
|
||||
MODEL_META_LLAMA_LLAMA_3_1_8B_INSTRUCT = "meta-llama/llama-3.1-8b-instruct",
|
||||
MODEL_META_LLAMA_LLAMA_3_1_70B_INSTRUCT = "meta-llama/llama-3.1-70b-instruct",
|
||||
MODEL_MISTRALAI_MISTRAL_NEMO = "mistralai/mistral-nemo",
|
||||
MODEL_OPENAI_GPT_4O_MINI_2024_07_18 = "openai/gpt-4o-mini-2024-07-18",
|
||||
MODEL_OPENAI_GPT_4O_MINI = "openai/gpt-4o-mini",
|
||||
MODEL_OPENAI_GPT_4O_MINI_2024_07_18 = "openai/gpt-4o-mini-2024-07-18",
|
||||
MODEL_GOOGLE_GEMMA_2_27B_IT = "google/gemma-2-27b-it",
|
||||
MODEL_GOOGLE_GEMMA_2_9B_IT = "google/gemma-2-9b-it",
|
||||
MODEL_SAO10K_L3_EURYALE_70B = "sao10k/l3-euryale-70b",
|
||||
MODEL_NOUSRESEARCH_HERMES_2_PRO_LLAMA_3_8B = "nousresearch/hermes-2-pro-llama-3-8b",
|
||||
MODEL_OPENAI_GPT_4O_2024_05_13 = "openai/gpt-4o-2024-05-13",
|
||||
MODEL_OPENAI_GPT_4O = "openai/gpt-4o",
|
||||
MODEL_OPENAI_GPT_4O_EXTENDED = "openai/gpt-4o:extended",
|
||||
MODEL_META_LLAMA_LLAMA_3_70B_INSTRUCT = "meta-llama/llama-3-70b-instruct",
|
||||
MODEL_OPENAI_GPT_4O_2024_05_13 = "openai/gpt-4o-2024-05-13",
|
||||
MODEL_META_LLAMA_LLAMA_3_8B_INSTRUCT = "meta-llama/llama-3-8b-instruct",
|
||||
MODEL_META_LLAMA_LLAMA_3_70B_INSTRUCT = "meta-llama/llama-3-70b-instruct",
|
||||
MODEL_MISTRALAI_MIXTRAL_8X22B_INSTRUCT = "mistralai/mixtral-8x22b-instruct",
|
||||
MODEL_MICROSOFT_WIZARDLM_2_8X22B = "microsoft/wizardlm-2-8x22b",
|
||||
MODEL_OPENAI_GPT_4_TURBO = "openai/gpt-4-turbo",
|
||||
MODEL_ANTHROPIC_CLAUDE_3_HAIKU = "anthropic/claude-3-haiku",
|
||||
MODEL_MISTRALAI_MISTRAL_LARGE = "mistralai/mistral-large",
|
||||
MODEL_OPENAI_GPT_3_5_TURBO_0613 = "openai/gpt-3.5-turbo-0613",
|
||||
MODEL_OPENAI_GPT_4_TURBO_PREVIEW = "openai/gpt-4-turbo-preview",
|
||||
MODEL_OPENAI_GPT_3_5_TURBO_0613 = "openai/gpt-3.5-turbo-0613",
|
||||
MODEL_MISTRALAI_MIXTRAL_8X7B_INSTRUCT = "mistralai/mixtral-8x7b-instruct",
|
||||
MODEL_ALPINDALE_GOLIATH_120B = "alpindale/goliath-120b",
|
||||
MODEL_OPENROUTER_AUTO = "openrouter/auto",
|
||||
@ -346,7 +341,7 @@ export enum E_OPENROUTER_MODEL {
|
||||
MODEL_MANCER_WEAVER = "mancer/weaver",
|
||||
MODEL_UNDI95_REMM_SLERP_L2_13B = "undi95/remm-slerp-l2-13b",
|
||||
MODEL_GRYPHE_MYTHOMAX_L2_13B = "gryphe/mythomax-l2-13b",
|
||||
MODEL_OPENAI_GPT_4_0314 = "openai/gpt-4-0314",
|
||||
MODEL_OPENAI_GPT_4 = "openai/gpt-4",
|
||||
MODEL_OPENAI_GPT_4_0314 = "openai/gpt-4-0314",
|
||||
MODEL_OPENAI_GPT_3_5_TURBO = "openai/gpt-3.5-turbo"
|
||||
}
|
||||
Loading…
Reference in New Issue
Block a user