maintainence love:)
This commit is contained in:
parent
934f0e4fe6
commit
55c5c9c4a5
@ -1,23 +1,24 @@
|
||||
import { marked } from 'marked';
|
||||
import { markedTerminal } from 'marked-terminal';
|
||||
import * as path from 'node:path';
|
||||
import { sync as write } from '@polymech/fs/write';
|
||||
import { resolve } from '@polymech/commons';
|
||||
import { dumpAsScript } from '../utils/script.js';
|
||||
import { applyFilters } from '../filters.js';
|
||||
import { variables } from '../variables.js';
|
||||
import { writeOrAppend } from '../utils/merge.js';
|
||||
export const onCompletion = async (result = "", options) => {
|
||||
result = applyFilters(result, options.filters || []);
|
||||
let vars = variables(options);
|
||||
if (options.dst) {
|
||||
vars = {
|
||||
...vars,
|
||||
MODEL: path.parse(options.model).name,
|
||||
ROUTER: options.router,
|
||||
MODEL: options.model ? path.parse(options.model).name : 'unknown_model',
|
||||
ROUTER: options.router || 'unknown_router',
|
||||
};
|
||||
const dstPath = path.resolve(resolve(options.dst, false, vars));
|
||||
write(dstPath, result);
|
||||
options.logger.debug(`Wrote completion result to ${dstPath} : ${options.dst}`);
|
||||
writeOrAppend(dstPath, result, options.append);
|
||||
const action = options.append ? `${options.append} content to` : 'Wrote completion result to';
|
||||
options.logger?.debug(`${action} ${dstPath} : ${options.dst}`);
|
||||
}
|
||||
else {
|
||||
marked.use(markedTerminal({
|
||||
@ -35,7 +36,6 @@ export const runCompletion = async (client, params, options) => {
|
||||
options.logger.info('Dry run - skipping API call');
|
||||
return false;
|
||||
}
|
||||
// await client.beta.chat.completions.parse
|
||||
const completion = await client.chat.completions.create({
|
||||
model: options.model,
|
||||
messages: params.messages,
|
||||
@ -48,4 +48,4 @@ export const runCompletion = async (client, params, options) => {
|
||||
result = await onCompletion(result, options);
|
||||
return result;
|
||||
};
|
||||
//# sourceMappingURL=data:application/json;base64,eyJ2ZXJzaW9uIjozLCJmaWxlIjoicnVuLWNvbXBsZXRpb24uanMiLCJzb3VyY2VSb290IjoiIiwic291cmNlcyI6WyIuLi8uLi9zcmMvY29tbWFuZHMvcnVuLWNvbXBsZXRpb24udHMiXSwibmFtZXMiOltdLCJtYXBwaW5ncyI6IkFBQ0EsT0FBTyxFQUFFLE1BQU0sRUFBRSxNQUFNLFFBQVEsQ0FBQTtBQUMvQixPQUFPLEVBQUUsY0FBYyxFQUFFLE1BQU0saUJBQWlCLENBQUE7QUFDaEQsT0FBTyxLQUFLLElBQUksTUFBTSxXQUFXLENBQUE7QUFDakMsT0FBTyxFQUFFLElBQUksSUFBSSxLQUFLLEVBQUUsTUFBTSxvQkFBb0IsQ0FBQTtBQUNsRCxPQUFPLEVBQUUsT0FBTyxFQUFFLE1BQU0sbUJBQW1CLENBQUE7QUFHM0MsT0FBTyxFQUFFLFlBQVksRUFBRSxNQUFNLG9CQUFvQixDQUFBO0FBQ2pELE9BQU8sRUFBRSxZQUFZLEVBQVUsTUFBTSxlQUFlLENBQUE7QUFDcEQsT0FBTyxFQUFFLFNBQVMsRUFBRSxNQUFNLGlCQUFpQixDQUFBO0FBRTNDLE1BQU0sQ0FBQyxNQUFNLFlBQVksR0FBRyxLQUFLLEVBQUUsU0FBYyxFQUFFLEVBQUUsT0FBa0IsRUFBRSxFQUFFO0lBQ3pFLE1BQU0sR0FBRyxZQUFZLENBQUMsTUFBTSxFQUFFLE9BQU8sQ0FBQyxPQUFtQixJQUFJLEVBQUUsQ0FBQyxDQUFBO0lBQ2hFLElBQUksSUFBSSxHQUFHLFNBQVMsQ0FBQyxPQUFPLENBQUMsQ0FBQTtJQUM3QixJQUFJLE9BQU8sQ0FBQyxHQUFHLEVBQUUsQ0FBQztRQUNoQixJQUFJLEdBQUc7WUFDTCxHQUFHLElBQUk7WUFDUCxLQUFLLEVBQUUsSUFBSSxDQUFDLEtBQUssQ0FBQyxPQUFPLENBQUMsS0FBSyxDQUFDLENBQUMsSUFBSTtZQUNyQyxNQUFNLEVBQUUsT0FBTyxDQUFDLE1BQU07U0FDdkIsQ0FBQTtRQUNELE1BQU0sT0FBTyxHQUFHLElBQUksQ0FBQyxPQUFPLENBQUMsT0FBTyxDQUFDLE9BQU8sQ0FBQyxHQUFHLEVBQUUsS0FBSyxFQUFFLElBQUksQ0FBQyxDQUFDLENBQUE7UUFDL0QsS0FBSyxDQUFDLE9BQU8sRUFBRSxNQUFNLENBQUMsQ0FBQTtRQUN0QixPQUFPLENBQUMsTUFBTSxDQUFDLEtBQUssQ0FBQyw4QkFBOEIsT0FBTyxNQUFNLE9BQU8sQ0FBQyxHQUFHLEVBQUUsQ0FBQyxDQUFBO0lBQ2hGLENBQUM7U0FBTSxDQUFDO1FBQ04sTUFBTSxDQUFDLEdBQUcsQ0FBQyxjQUFjLENBQUM7WUFDeEIsS0FBSyxFQUFFLEtBQUs7U0FDYixDQUFDLENBQUMsQ0FBQTtRQUNILE1BQU0sT0FBTyxHQUFXLE1BQU0sQ0FBQyxNQUFNLENBQVcsQ0FBQztRQUNqRCxPQUFPLENBQUMsTUFBTSxDQUFDLEtBQUssQ0FBQyxPQUFPLENBQUMsQ0FBQTtJQUMvQixDQUFDO0lBQ0QsWUFBWSxDQUFDLE9BQU8sQ0FBQyxDQUFBO0lBQ3JCLGtCQUFrQjtJQUNsQixPQUFPLE1BQU0sQ0FBQTtBQUNmLENBQUMsQ0FBQTtBQUVELE1BQU0sQ0FBQyxNQUFNLGFBQWEsR0FBRyxLQUFLLEVBQUUsTUFBYyxFQUFFLE1BQVcsRUFBRSxPQUFrQixFQUFFLEVBQUU7SUFDckYsSUFBSSxPQUFPLENBQUMsR0FBRyxFQUFFLENBQUM7UUFDaEIsT0FBTyxDQUFDLE1BQU0sQ0FBQyxJQUFJLENBQUMsNkJBQTZCLENBQUMsQ0FBQTtRQUNsRCxPQUFPLEtBQUssQ0FBQTtJQUNkLENBQUM7SUFDRCwyQ0FBMkM7SUFDM0MsTUFBTSxVQUFVLEdBQUcsTUFBTSxNQUFNLENBQUMsSUFBSSxDQUFDLFdBQVcsQ0FBQyxNQUFNLENBQUM7UUFDdEQsS0FBSyxFQUFFLE9BQU8sQ0FBQyxLQUFLO1FBQ3BCLFFBQVEsRUFBRSxNQUFNLENBQUMsUUFBUTtRQUN6QixlQUFlLEVBQUUsT0FBTyxDQUFDLE1BQWE7S0FDdkMsQ0FBQyxDQUFBO0lBQ0YsSUFBSSxDQUFDLFVBQVUsSUFBSSxDQUFDLFVBQVUsQ0FBQyxPQUFPLElBQUksQ0FBQyxVQUFVLENBQUMsT0FBTyxDQUFDLENBQUMsQ0FBQyxJQUFJLENBQUMsVUFBVSxDQUFDLE9BQU8sQ0FBQyxDQUFDLENBQUMsQ0FBQyxPQUFPLElBQUksQ0FBQyxVQUFVLENBQUMsT0FBTyxDQUFDLENBQUMsQ0FBQyxDQUFDLE9BQU8sQ0FBQyxPQUFPLEVBQUUsQ0FBQztRQUM3SSxPQUFPLEVBQUUsQ0FBQTtJQUNYLENBQUM7SUFDRCxJQUFJLE1BQU0sR0FBRyxVQUFVLENBQUMsT0FBTyxDQUFDLENBQUMsQ0FBQyxDQUFDLE9BQU8sQ0FBQyxPQUFPLENBQUE7SUFDbEQsTUFBTSxHQUFHLE1BQU0sWUFBWSxDQUFDLE1BQU0sRUFBRSxPQUFPLENBQUMsQ0FBQTtJQUM1QyxPQUFPLE1BQU0sQ0FBQTtBQUNmLENBQUMsQ0FBQSJ9
|
||||
//# sourceMappingURL=data:application/json;base64,eyJ2ZXJzaW9uIjozLCJmaWxlIjoicnVuLWNvbXBsZXRpb24uanMiLCJzb3VyY2VSb290IjoiIiwic291cmNlcyI6WyIuLi8uLi9zcmMvY29tbWFuZHMvcnVuLWNvbXBsZXRpb24udHMiXSwibmFtZXMiOltdLCJtYXBwaW5ncyI6IkFBQ0EsT0FBTyxFQUFFLE1BQU0sRUFBRSxNQUFNLFFBQVEsQ0FBQTtBQUMvQixPQUFPLEVBQUUsY0FBYyxFQUFFLE1BQU0saUJBQWlCLENBQUE7QUFDaEQsT0FBTyxLQUFLLElBQUksTUFBTSxXQUFXLENBQUE7QUFDakMsT0FBTyxFQUFFLE9BQU8sRUFBRSxNQUFNLG1CQUFtQixDQUFBO0FBRzNDLE9BQU8sRUFBRSxZQUFZLEVBQUUsTUFBTSxvQkFBb0IsQ0FBQTtBQUNqRCxPQUFPLEVBQUUsWUFBWSxFQUFVLE1BQU0sZUFBZSxDQUFBO0FBQ3BELE9BQU8sRUFBRSxTQUFTLEVBQUUsTUFBTSxpQkFBaUIsQ0FBQTtBQUMzQyxPQUFPLEVBQUUsYUFBYSxFQUFFLE1BQU0sbUJBQW1CLENBQUE7QUFHakQsTUFBTSxDQUFDLE1BQU0sWUFBWSxHQUFHLEtBQUssRUFBRSxTQUFjLEVBQUUsRUFBRSxPQUFrQixFQUFFLEVBQUU7SUFDekUsTUFBTSxHQUFHLFlBQVksQ0FBQyxNQUFNLEVBQUUsT0FBTyxDQUFDLE9BQW1CLElBQUksRUFBRSxDQUFDLENBQUE7SUFDaEUsSUFBSSxJQUFJLEdBQUcsU0FBUyxDQUFDLE9BQU8sQ0FBQyxDQUFBO0lBQzdCLElBQUksT0FBTyxDQUFDLEdBQUcsRUFBRSxDQUFDO1FBQ2hCLElBQUksR0FBRztZQUNMLEdBQUcsSUFBSTtZQUNQLEtBQUssRUFBRSxPQUFPLENBQUMsS0FBSyxDQUFDLENBQUMsQ0FBQyxJQUFJLENBQUMsS0FBSyxDQUFDLE9BQU8sQ0FBQyxLQUFLLENBQUMsQ0FBQyxJQUFJLENBQUMsQ0FBQyxDQUFDLGVBQWU7WUFDdkUsTUFBTSxFQUFFLE9BQU8sQ0FBQyxNQUFNLElBQUksZ0JBQWdCO1NBQzNDLENBQUE7UUFDRCxNQUFNLE9BQU8sR0FBRyxJQUFJLENBQUMsT0FBTyxDQUFDLE9BQU8sQ0FBQyxPQUFPLENBQUMsR0FBRyxFQUFFLEtBQUssRUFBRSxJQUFJLENBQUMsQ0FBQyxDQUFBO1FBQy9ELGFBQWEsQ0FBQyxPQUFPLEVBQUUsTUFBTSxFQUFFLE9BQU8sQ0FBQyxNQUEwQixDQUFDLENBQUE7UUFDbEUsTUFBTSxNQUFNLEdBQUcsT0FBTyxDQUFDLE1BQU0sQ0FBQyxDQUFDLENBQUMsR0FBRyxPQUFPLENBQUMsTUFBTSxhQUFhLENBQUMsQ0FBQyxDQUFDLDRCQUE0QixDQUFBO1FBQzdGLE9BQU8sQ0FBQyxNQUFNLEVBQUUsS0FBSyxDQUFDLEdBQUcsTUFBTSxJQUFJLE9BQU8sTUFBTSxPQUFPLENBQUMsR0FBRyxFQUFFLENBQUMsQ0FBQTtJQUNoRSxDQUFDO1NBQU0sQ0FBQztRQUNOLE1BQU0sQ0FBQyxHQUFHLENBQUMsY0FBYyxDQUFDO1lBQ3hCLEtBQUssRUFBRSxLQUFLO1NBQ2IsQ0FBQyxDQUFDLENBQUE7UUFDSCxNQUFNLE9BQU8sR0FBVyxNQUFNLENBQUMsTUFBTSxDQUFXLENBQUM7UUFDakQsT0FBTyxDQUFDLE1BQU0sQ0FBQyxLQUFLLENBQUMsT0FBTyxDQUFDLENBQUE7SUFDL0IsQ0FBQztJQUNELFlBQVksQ0FBQyxPQUFPLENBQUMsQ0FBQTtJQUNyQixrQkFBa0I7SUFDbEIsT0FBTyxNQUFNLENBQUE7QUFDZixDQUFDLENBQUE7QUFFRCxNQUFNLENBQUMsTUFBTSxhQUFhLEdBQUcsS0FBSyxFQUFFLE1BQWMsRUFBRSxNQUFXLEVBQUUsT0FBa0IsRUFBRSxFQUFFO0lBQ3JGLElBQUksT0FBTyxDQUFDLEdBQUcsRUFBRSxDQUFDO1FBQ2hCLE9BQU8sQ0FBQyxNQUFNLENBQUMsSUFBSSxDQUFDLDZCQUE2QixDQUFDLENBQUE7UUFDbEQsT0FBTyxLQUFLLENBQUE7SUFDZCxDQUFDO0lBQ0QsTUFBTSxVQUFVLEdBQUcsTUFBTSxNQUFNLENBQUMsSUFBSSxDQUFDLFdBQVcsQ0FBQyxNQUFNLENBQUM7UUFDdEQsS0FBSyxFQUFFLE9BQU8sQ0FBQyxLQUFLO1FBQ3BCLFFBQVEsRUFBRSxNQUFNLENBQUMsUUFBUTtRQUN6QixlQUFlLEVBQUUsT0FBTyxDQUFDLE1BQWE7S0FDdkMsQ0FBQyxDQUFBO0lBQ0YsSUFBSSxDQUFDLFVBQVUsSUFBSSxDQUFDLFVBQVUsQ0FBQyxPQUFPLElBQUksQ0FBQyxVQUFVLENBQUMsT0FBTyxDQUFDLENBQUMsQ0FBQyxJQUFJLENBQUMsVUFBVSxDQUFDLE9BQU8sQ0FBQyxDQUFDLENBQUMsQ0FBQyxPQUFPLElBQUksQ0FBQyxVQUFVLENBQUMsT0FBTyxDQUFDLENBQUMsQ0FBQyxDQUFDLE9BQU8sQ0FBQyxPQUFPLEVBQUUsQ0FBQztRQUM3SSxPQUFPLEVBQUUsQ0FBQTtJQUNYLENBQUM7SUFDRCxJQUFJLE1BQU0sR0FBRyxVQUFVLENBQUMsT0FBTyxDQUFDLENBQUMsQ0FBQyxDQUFDLE9BQU8sQ0FBQyxPQUFPLENBQUE7SUFDbEQsTUFBTSxHQUFHLE1BQU0sWUFBWSxDQUFDLE1BQU0sRUFBRSxPQUFPLENBQUMsQ0FBQTtJQUM1QyxPQUFPLE1BQU0sQ0FBQTtBQUNmLENBQUMsQ0FBQSJ9
|
||||
@ -1,5 +1,5 @@
|
||||
{
|
||||
"timestamp": 1745239682440,
|
||||
"timestamp": 1745500189322,
|
||||
"models": [
|
||||
{
|
||||
"id": "gpt-4o-audio-preview-2024-12-17",
|
||||
@ -13,30 +13,24 @@
|
||||
"created": 1698785189,
|
||||
"owned_by": "system"
|
||||
},
|
||||
{
|
||||
"id": "text-embedding-3-large",
|
||||
"object": "model",
|
||||
"created": 1705953180,
|
||||
"owned_by": "system"
|
||||
},
|
||||
{
|
||||
"id": "dall-e-2",
|
||||
"object": "model",
|
||||
"created": 1698798177,
|
||||
"owned_by": "system"
|
||||
},
|
||||
{
|
||||
"id": "o4-mini-2025-04-16",
|
||||
"object": "model",
|
||||
"created": 1744133506,
|
||||
"owned_by": "system"
|
||||
},
|
||||
{
|
||||
"id": "gpt-4o-audio-preview-2024-10-01",
|
||||
"object": "model",
|
||||
"created": 1727389042,
|
||||
"owned_by": "system"
|
||||
},
|
||||
{
|
||||
"id": "text-embedding-3-small",
|
||||
"object": "model",
|
||||
"created": 1705948997,
|
||||
"owned_by": "system"
|
||||
},
|
||||
{
|
||||
"id": "o4-mini",
|
||||
"object": "model",
|
||||
@ -61,6 +55,12 @@
|
||||
"created": 1727131766,
|
||||
"owned_by": "system"
|
||||
},
|
||||
{
|
||||
"id": "o4-mini-2025-04-16",
|
||||
"object": "model",
|
||||
"created": 1744133506,
|
||||
"owned_by": "system"
|
||||
},
|
||||
{
|
||||
"id": "gpt-4o-realtime-preview",
|
||||
"object": "model",
|
||||
@ -85,6 +85,12 @@
|
||||
"created": 1671217299,
|
||||
"owned_by": "openai-internal"
|
||||
},
|
||||
{
|
||||
"id": "text-embedding-3-large",
|
||||
"object": "model",
|
||||
"created": 1705953180,
|
||||
"owned_by": "system"
|
||||
},
|
||||
{
|
||||
"id": "gpt-4o-mini-audio-preview",
|
||||
"object": "model",
|
||||
@ -103,18 +109,6 @@
|
||||
"created": 1725648865,
|
||||
"owned_by": "system"
|
||||
},
|
||||
{
|
||||
"id": "o1-pro",
|
||||
"object": "model",
|
||||
"created": 1742251791,
|
||||
"owned_by": "system"
|
||||
},
|
||||
{
|
||||
"id": "o1-2024-12-17",
|
||||
"object": "model",
|
||||
"created": 1734326976,
|
||||
"owned_by": "system"
|
||||
},
|
||||
{
|
||||
"id": "gpt-4o-mini-realtime-preview",
|
||||
"object": "model",
|
||||
@ -145,24 +139,12 @@
|
||||
"created": 1741391161,
|
||||
"owned_by": "system"
|
||||
},
|
||||
{
|
||||
"id": "o1",
|
||||
"object": "model",
|
||||
"created": 1734375816,
|
||||
"owned_by": "system"
|
||||
},
|
||||
{
|
||||
"id": "gpt-4.1-mini-2025-04-14",
|
||||
"object": "model",
|
||||
"created": 1744317547,
|
||||
"owned_by": "system"
|
||||
},
|
||||
{
|
||||
"id": "o1-pro-2025-03-19",
|
||||
"object": "model",
|
||||
"created": 1742251504,
|
||||
"owned_by": "system"
|
||||
},
|
||||
{
|
||||
"id": "chatgpt-4o-latest",
|
||||
"object": "model",
|
||||
@ -277,18 +259,6 @@
|
||||
"created": 1740623059,
|
||||
"owned_by": "system"
|
||||
},
|
||||
{
|
||||
"id": "o3-mini",
|
||||
"object": "model",
|
||||
"created": 1737146383,
|
||||
"owned_by": "system"
|
||||
},
|
||||
{
|
||||
"id": "o3-mini-2025-01-31",
|
||||
"object": "model",
|
||||
"created": 1738010200,
|
||||
"owned_by": "system"
|
||||
},
|
||||
{
|
||||
"id": "gpt-4.5-preview-2025-02-27",
|
||||
"object": "model",
|
||||
@ -308,15 +278,15 @@
|
||||
"owned_by": "system"
|
||||
},
|
||||
{
|
||||
"id": "text-embedding-3-small",
|
||||
"id": "o3-mini-2025-01-31",
|
||||
"object": "model",
|
||||
"created": 1705948997,
|
||||
"created": 1738010200,
|
||||
"owned_by": "system"
|
||||
},
|
||||
{
|
||||
"id": "gpt-4o-mini-tts",
|
||||
"id": "o3-mini",
|
||||
"object": "model",
|
||||
"created": 1742403959,
|
||||
"created": 1737146383,
|
||||
"owned_by": "system"
|
||||
},
|
||||
{
|
||||
@ -367,6 +337,12 @@
|
||||
"created": 1744315746,
|
||||
"owned_by": "system"
|
||||
},
|
||||
{
|
||||
"id": "o1-2024-12-17",
|
||||
"object": "model",
|
||||
"created": 1734326976,
|
||||
"owned_by": "system"
|
||||
},
|
||||
{
|
||||
"id": "gpt-4o-mini-2024-07-18",
|
||||
"object": "model",
|
||||
@ -415,12 +391,36 @@
|
||||
"created": 1698957206,
|
||||
"owned_by": "system"
|
||||
},
|
||||
{
|
||||
"id": "gpt-4o-mini-tts",
|
||||
"object": "model",
|
||||
"created": 1742403959,
|
||||
"owned_by": "system"
|
||||
},
|
||||
{
|
||||
"id": "tts-1-1106",
|
||||
"object": "model",
|
||||
"created": 1699053241,
|
||||
"owned_by": "system"
|
||||
},
|
||||
{
|
||||
"id": "o1",
|
||||
"object": "model",
|
||||
"created": 1734375816,
|
||||
"owned_by": "system"
|
||||
},
|
||||
{
|
||||
"id": "o1-pro",
|
||||
"object": "model",
|
||||
"created": 1742251791,
|
||||
"owned_by": "system"
|
||||
},
|
||||
{
|
||||
"id": "o1-pro-2025-03-19",
|
||||
"object": "model",
|
||||
"created": 1742251504,
|
||||
"owned_by": "system"
|
||||
},
|
||||
{
|
||||
"id": "omni-moderation-latest",
|
||||
"object": "model",
|
||||
|
||||
@ -1,6 +1,38 @@
|
||||
{
|
||||
"timestamp": 1745239682608,
|
||||
"timestamp": 1745500189490,
|
||||
"models": [
|
||||
{
|
||||
"id": "microsoft/mai-ds-r1:free",
|
||||
"name": "Microsoft: MAI DS R1 (free)",
|
||||
"created": 1745194100,
|
||||
"description": "MAI-DS-R1 is a post-trained variant of DeepSeek-R1 developed by the Microsoft AI team to improve the model’s responsiveness on previously blocked topics while enhancing its safety profile. Built on top of DeepSeek-R1’s reasoning foundation, it integrates 110k examples from the Tulu-3 SFT dataset and 350k internally curated multilingual safety-alignment samples. The model retains strong reasoning, coding, and problem-solving capabilities, while unblocking a wide range of prompts previously restricted in R1.\n\nMAI-DS-R1 demonstrates improved performance on harm mitigation benchmarks and maintains competitive results across general reasoning tasks. It surpasses R1-1776 in satisfaction metrics for blocked queries and reduces leakage in harmful content categories. The model is based on a transformer MoE architecture and is suitable for general-purpose use cases, excluding high-stakes domains such as legal, medical, or autonomous systems.",
|
||||
"context_length": 163840,
|
||||
"architecture": {
|
||||
"modality": "text->text",
|
||||
"input_modalities": [
|
||||
"text"
|
||||
],
|
||||
"output_modalities": [
|
||||
"text"
|
||||
],
|
||||
"tokenizer": "DeepSeek",
|
||||
"instruct_type": "deepseek-r1"
|
||||
},
|
||||
"pricing": {
|
||||
"prompt": "0",
|
||||
"completion": "0",
|
||||
"request": "0",
|
||||
"image": "0",
|
||||
"web_search": "0",
|
||||
"internal_reasoning": "0"
|
||||
},
|
||||
"top_provider": {
|
||||
"context_length": 163840,
|
||||
"max_completion_tokens": null,
|
||||
"is_moderated": false
|
||||
},
|
||||
"per_request_limits": null
|
||||
},
|
||||
{
|
||||
"id": "google/gemini-2.5-pro-preview-03-25",
|
||||
"name": "Google: Gemini 2.5 Pro Preview",
|
||||
@ -11,7 +43,8 @@
|
||||
"modality": "text+image->text",
|
||||
"input_modalities": [
|
||||
"text",
|
||||
"image"
|
||||
"image",
|
||||
"file"
|
||||
],
|
||||
"output_modalities": [
|
||||
"text"
|
||||
@ -26,7 +59,7 @@
|
||||
"image": "0.00516",
|
||||
"web_search": "0",
|
||||
"internal_reasoning": "0",
|
||||
"input_cache_read": "0.0000003125",
|
||||
"input_cache_read": "0.000000625",
|
||||
"input_cache_write": "0"
|
||||
},
|
||||
"top_provider": {
|
||||
@ -110,7 +143,8 @@
|
||||
"modality": "text+image->text",
|
||||
"input_modalities": [
|
||||
"image",
|
||||
"text"
|
||||
"text",
|
||||
"file"
|
||||
],
|
||||
"output_modalities": [
|
||||
"text"
|
||||
@ -143,7 +177,8 @@
|
||||
"modality": "text+image->text",
|
||||
"input_modalities": [
|
||||
"image",
|
||||
"text"
|
||||
"text",
|
||||
"file"
|
||||
],
|
||||
"output_modalities": [
|
||||
"text"
|
||||
@ -176,7 +211,8 @@
|
||||
"modality": "text+image->text",
|
||||
"input_modalities": [
|
||||
"image",
|
||||
"text"
|
||||
"text",
|
||||
"file"
|
||||
],
|
||||
"output_modalities": [
|
||||
"text"
|
||||
@ -210,7 +246,8 @@
|
||||
"modality": "text+image->text",
|
||||
"input_modalities": [
|
||||
"image",
|
||||
"text"
|
||||
"text",
|
||||
"file"
|
||||
],
|
||||
"output_modalities": [
|
||||
"text"
|
||||
@ -808,7 +845,7 @@
|
||||
},
|
||||
"pricing": {
|
||||
"prompt": "0.00000017",
|
||||
"completion": "0.00000085",
|
||||
"completion": "0.0000006",
|
||||
"request": "0",
|
||||
"image": "0.0006684",
|
||||
"web_search": "0",
|
||||
@ -816,7 +853,7 @@
|
||||
},
|
||||
"top_provider": {
|
||||
"context_length": 1048576,
|
||||
"max_completion_tokens": 1048576,
|
||||
"max_completion_tokens": 16384,
|
||||
"is_moderated": false
|
||||
},
|
||||
"per_request_limits": null
|
||||
@ -859,7 +896,7 @@
|
||||
"name": "Meta: Llama 4 Scout",
|
||||
"created": 1743881519,
|
||||
"description": "Llama 4 Scout 17B Instruct (16E) is a mixture-of-experts (MoE) language model developed by Meta, activating 17 billion parameters out of a total of 109B. It supports native multimodal input (text and image) and multilingual output (text and code) across 12 supported languages. Designed for assistant-style interaction and visual reasoning, Scout uses 16 experts per forward pass and features a context length of 10 million tokens, with a training corpus of ~40 trillion tokens.\n\nBuilt for high efficiency and local or commercial deployment, Llama 4 Scout incorporates early fusion for seamless modality integration. It is instruction-tuned for use in multilingual chat, captioning, and image understanding tasks. Released under the Llama 4 Community License, it was last trained on data up to August 2024 and launched publicly on April 5, 2025.",
|
||||
"context_length": 327680,
|
||||
"context_length": 1048576,
|
||||
"architecture": {
|
||||
"modality": "text+image->text",
|
||||
"input_modalities": [
|
||||
@ -876,13 +913,13 @@
|
||||
"prompt": "0.00000008",
|
||||
"completion": "0.0000003",
|
||||
"request": "0",
|
||||
"image": "0.0003342",
|
||||
"image": "0",
|
||||
"web_search": "0",
|
||||
"internal_reasoning": "0"
|
||||
},
|
||||
"top_provider": {
|
||||
"context_length": 327680,
|
||||
"max_completion_tokens": 8192,
|
||||
"context_length": 1048576,
|
||||
"max_completion_tokens": 1048576,
|
||||
"is_moderated": false
|
||||
},
|
||||
"per_request_limits": null
|
||||
@ -1413,7 +1450,7 @@
|
||||
"name": "Mistral: Mistral Small 3.1 24B",
|
||||
"created": 1742238937,
|
||||
"description": "Mistral Small 3.1 24B Instruct is an upgraded variant of Mistral Small 3 (2501), featuring 24 billion parameters with advanced multimodal capabilities. It provides state-of-the-art performance in text-based reasoning and vision tasks, including image analysis, programming, mathematical reasoning, and multilingual support across dozens of languages. Equipped with an extensive 128k token context window and optimized for efficient local inference, it supports use cases such as conversational agents, function calling, long-document comprehension, and privacy-sensitive deployments.",
|
||||
"context_length": 32768,
|
||||
"context_length": 128000,
|
||||
"architecture": {
|
||||
"modality": "text+image->text",
|
||||
"input_modalities": [
|
||||
@ -1430,45 +1467,13 @@
|
||||
"prompt": "0.0000001",
|
||||
"completion": "0.0000003",
|
||||
"request": "0",
|
||||
"image": "0.0009264",
|
||||
"image": "0.000926",
|
||||
"web_search": "0",
|
||||
"internal_reasoning": "0"
|
||||
},
|
||||
"top_provider": {
|
||||
"context_length": 32768,
|
||||
"max_completion_tokens": null,
|
||||
"is_moderated": false
|
||||
},
|
||||
"per_request_limits": null
|
||||
},
|
||||
{
|
||||
"id": "open-r1/olympiccoder-7b:free",
|
||||
"name": "OlympicCoder 7B (free)",
|
||||
"created": 1742078265,
|
||||
"description": "OlympicCoder-7B is an open-source language model fine-tuned on the CodeForces-CoTs dataset, consisting of nearly 100,000 high-quality chain-of-thought examples from competitive programming contexts. Optimized specifically for olympiad-level coding problems, this model demonstrates strong chain-of-thought reasoning and competitive code generation capabilities, achieving performance competitive with frontier closed-source models on tasks from the IOI 2024 and similar coding contests.",
|
||||
"context_length": 32768,
|
||||
"architecture": {
|
||||
"modality": "text->text",
|
||||
"input_modalities": [
|
||||
"text"
|
||||
],
|
||||
"output_modalities": [
|
||||
"text"
|
||||
],
|
||||
"tokenizer": "Other",
|
||||
"instruct_type": "deepseek-r1"
|
||||
},
|
||||
"pricing": {
|
||||
"prompt": "0",
|
||||
"completion": "0",
|
||||
"request": "0",
|
||||
"image": "0",
|
||||
"web_search": "0",
|
||||
"internal_reasoning": "0"
|
||||
},
|
||||
"top_provider": {
|
||||
"context_length": 32768,
|
||||
"max_completion_tokens": null,
|
||||
"context_length": 128000,
|
||||
"max_completion_tokens": 128000,
|
||||
"is_moderated": false
|
||||
},
|
||||
"per_request_limits": null
|
||||
@ -1805,10 +1810,9 @@
|
||||
"description": "GPT-4o mini Search Preview is a specialized model for web search in Chat Completions. It is trained to understand and execute web search queries.",
|
||||
"context_length": 128000,
|
||||
"architecture": {
|
||||
"modality": "text+image->text",
|
||||
"modality": "text->text",
|
||||
"input_modalities": [
|
||||
"text",
|
||||
"image"
|
||||
"text"
|
||||
],
|
||||
"output_modalities": [
|
||||
"text"
|
||||
@ -1838,10 +1842,9 @@
|
||||
"description": "GPT-4o Search Previewis a specialized model for web search in Chat Completions. It is trained to understand and execute web search queries.",
|
||||
"context_length": 128000,
|
||||
"architecture": {
|
||||
"modality": "text+image->text",
|
||||
"modality": "text->text",
|
||||
"input_modalities": [
|
||||
"text",
|
||||
"image"
|
||||
"text"
|
||||
],
|
||||
"output_modalities": [
|
||||
"text"
|
||||
@ -1957,7 +1960,7 @@
|
||||
},
|
||||
"top_provider": {
|
||||
"context_length": 131072,
|
||||
"max_completion_tokens": 8192,
|
||||
"max_completion_tokens": 16384,
|
||||
"is_moderated": false
|
||||
},
|
||||
"per_request_limits": null
|
||||
@ -2391,7 +2394,8 @@
|
||||
"modality": "text+image->text",
|
||||
"input_modalities": [
|
||||
"text",
|
||||
"image"
|
||||
"image",
|
||||
"file"
|
||||
],
|
||||
"output_modalities": [
|
||||
"text"
|
||||
@ -2665,8 +2669,8 @@
|
||||
"instruct_type": "none"
|
||||
},
|
||||
"pricing": {
|
||||
"prompt": "0.0000001",
|
||||
"completion": "0.0000001",
|
||||
"prompt": "0.000000055",
|
||||
"completion": "0.000000055",
|
||||
"request": "0",
|
||||
"image": "0",
|
||||
"web_search": "0",
|
||||
@ -2754,7 +2758,8 @@
|
||||
"modality": "text+image->text",
|
||||
"input_modalities": [
|
||||
"text",
|
||||
"image"
|
||||
"image",
|
||||
"file"
|
||||
],
|
||||
"output_modalities": [
|
||||
"text"
|
||||
@ -2768,7 +2773,9 @@
|
||||
"request": "0",
|
||||
"image": "0.0000258",
|
||||
"web_search": "0",
|
||||
"internal_reasoning": "0"
|
||||
"internal_reasoning": "0",
|
||||
"input_cache_read": "0.000000025",
|
||||
"input_cache_write": "0.0000001833"
|
||||
},
|
||||
"top_provider": {
|
||||
"context_length": 1000000,
|
||||
@ -3225,7 +3232,7 @@
|
||||
},
|
||||
"top_provider": {
|
||||
"context_length": 32768,
|
||||
"max_completion_tokens": 8192,
|
||||
"max_completion_tokens": 16384,
|
||||
"is_moderated": false
|
||||
},
|
||||
"per_request_limits": null
|
||||
@ -3289,7 +3296,7 @@
|
||||
},
|
||||
"top_provider": {
|
||||
"context_length": 131072,
|
||||
"max_completion_tokens": 8192,
|
||||
"max_completion_tokens": 16384,
|
||||
"is_moderated": false
|
||||
},
|
||||
"per_request_limits": null
|
||||
@ -3523,7 +3530,7 @@
|
||||
"name": "DeepSeek: R1 Distill Llama 70B",
|
||||
"created": 1737663169,
|
||||
"description": "DeepSeek R1 Distill Llama 70B is a distilled large language model based on [Llama-3.3-70B-Instruct](/meta-llama/llama-3.3-70b-instruct), using outputs from [DeepSeek R1](/deepseek/deepseek-r1). The model combines advanced distillation techniques to achieve high performance across multiple benchmarks, including:\n\n- AIME 2024 pass@1: 70.0\n- MATH-500 pass@1: 94.5\n- CodeForces Rating: 1633\n\nThe model leverages fine-tuning from DeepSeek R1's outputs, enabling competitive performance comparable to larger frontier models.",
|
||||
"context_length": 128000,
|
||||
"context_length": 131072,
|
||||
"architecture": {
|
||||
"modality": "text->text",
|
||||
"input_modalities": [
|
||||
@ -3544,45 +3551,12 @@
|
||||
"internal_reasoning": "0"
|
||||
},
|
||||
"top_provider": {
|
||||
"context_length": 128000,
|
||||
"context_length": 131072,
|
||||
"max_completion_tokens": 16384,
|
||||
"is_moderated": false
|
||||
},
|
||||
"per_request_limits": null
|
||||
},
|
||||
{
|
||||
"id": "google/gemini-2.0-flash-thinking-exp:free",
|
||||
"name": "Google: Gemini 2.0 Flash Thinking Experimental 01-21 (free)",
|
||||
"created": 1737547899,
|
||||
"description": "Gemini 2.0 Flash Thinking Experimental (01-21) is a snapshot of Gemini 2.0 Flash Thinking Experimental.\n\nGemini 2.0 Flash Thinking Mode is an experimental model that's trained to generate the \"thinking process\" the model goes through as part of its response. As a result, Thinking Mode is capable of stronger reasoning capabilities in its responses than the [base Gemini 2.0 Flash model](/google/gemini-2.0-flash-exp).",
|
||||
"context_length": 1048576,
|
||||
"architecture": {
|
||||
"modality": "text+image->text",
|
||||
"input_modalities": [
|
||||
"text",
|
||||
"image"
|
||||
],
|
||||
"output_modalities": [
|
||||
"text"
|
||||
],
|
||||
"tokenizer": "Gemini",
|
||||
"instruct_type": null
|
||||
},
|
||||
"pricing": {
|
||||
"prompt": "0",
|
||||
"completion": "0",
|
||||
"request": "0",
|
||||
"image": "0",
|
||||
"web_search": "0",
|
||||
"internal_reasoning": "0"
|
||||
},
|
||||
"top_provider": {
|
||||
"context_length": 1048576,
|
||||
"max_completion_tokens": 65536,
|
||||
"is_moderated": false
|
||||
},
|
||||
"per_request_limits": null
|
||||
},
|
||||
{
|
||||
"id": "deepseek/deepseek-r1:free",
|
||||
"name": "DeepSeek: R1 (free)",
|
||||
@ -3620,7 +3594,7 @@
|
||||
"name": "DeepSeek: R1",
|
||||
"created": 1737381095,
|
||||
"description": "DeepSeek R1 is here: Performance on par with [OpenAI o1](/openai/o1), but open-sourced and with fully open reasoning tokens. It's 671B parameters in size, with 37B active in an inference pass.\n\nFully open-source model & [technical report](https://api-docs.deepseek.com/news/news250120).\n\nMIT licensed: Distill & commercialize freely!",
|
||||
"context_length": 128000,
|
||||
"context_length": 163840,
|
||||
"architecture": {
|
||||
"modality": "text->text",
|
||||
"input_modalities": [
|
||||
@ -3634,15 +3608,15 @@
|
||||
},
|
||||
"pricing": {
|
||||
"prompt": "0.0000005",
|
||||
"completion": "0.000003",
|
||||
"completion": "0.00000218",
|
||||
"request": "0",
|
||||
"image": "0",
|
||||
"web_search": "0",
|
||||
"internal_reasoning": "0"
|
||||
},
|
||||
"top_provider": {
|
||||
"context_length": 128000,
|
||||
"max_completion_tokens": 32768,
|
||||
"context_length": 163840,
|
||||
"max_completion_tokens": 163840,
|
||||
"is_moderated": false
|
||||
},
|
||||
"per_request_limits": null
|
||||
@ -3771,7 +3745,7 @@
|
||||
},
|
||||
"top_provider": {
|
||||
"context_length": 16384,
|
||||
"max_completion_tokens": 8192,
|
||||
"max_completion_tokens": 16384,
|
||||
"is_moderated": false
|
||||
},
|
||||
"per_request_limits": null
|
||||
@ -3840,39 +3814,6 @@
|
||||
},
|
||||
"per_request_limits": null
|
||||
},
|
||||
{
|
||||
"id": "google/gemini-2.0-flash-thinking-exp-1219:free",
|
||||
"name": "Google: Gemini 2.0 Flash Thinking Experimental (free)",
|
||||
"created": 1734650026,
|
||||
"description": "Gemini 2.0 Flash Thinking Mode is an experimental model that's trained to generate the \"thinking process\" the model goes through as part of its response. As a result, Thinking Mode is capable of stronger reasoning capabilities in its responses than the [base Gemini 2.0 Flash model](/google/gemini-2.0-flash-exp).",
|
||||
"context_length": 40000,
|
||||
"architecture": {
|
||||
"modality": "text+image->text",
|
||||
"input_modalities": [
|
||||
"text",
|
||||
"image"
|
||||
],
|
||||
"output_modalities": [
|
||||
"text"
|
||||
],
|
||||
"tokenizer": "Gemini",
|
||||
"instruct_type": null
|
||||
},
|
||||
"pricing": {
|
||||
"prompt": "0",
|
||||
"completion": "0",
|
||||
"request": "0",
|
||||
"image": "0",
|
||||
"web_search": "0",
|
||||
"internal_reasoning": "0"
|
||||
},
|
||||
"top_provider": {
|
||||
"context_length": 40000,
|
||||
"max_completion_tokens": 8000,
|
||||
"is_moderated": false
|
||||
},
|
||||
"per_request_limits": null
|
||||
},
|
||||
{
|
||||
"id": "sao10k/l3.3-euryale-70b",
|
||||
"name": "Sao10K: Llama 3.3 Euryale 70B",
|
||||
@ -3900,7 +3841,7 @@
|
||||
},
|
||||
"top_provider": {
|
||||
"context_length": 131072,
|
||||
"max_completion_tokens": 8192,
|
||||
"max_completion_tokens": 16384,
|
||||
"is_moderated": false
|
||||
},
|
||||
"per_request_limits": null
|
||||
@ -4402,7 +4343,8 @@
|
||||
"modality": "text+image->text",
|
||||
"input_modalities": [
|
||||
"text",
|
||||
"image"
|
||||
"image",
|
||||
"file"
|
||||
],
|
||||
"output_modalities": [
|
||||
"text"
|
||||
@ -4647,7 +4589,7 @@
|
||||
},
|
||||
"top_provider": {
|
||||
"context_length": 32768,
|
||||
"max_completion_tokens": 8192,
|
||||
"max_completion_tokens": 16384,
|
||||
"is_moderated": false
|
||||
},
|
||||
"per_request_limits": null
|
||||
@ -5177,7 +5119,7 @@
|
||||
},
|
||||
"top_provider": {
|
||||
"context_length": 32768,
|
||||
"max_completion_tokens": 8192,
|
||||
"max_completion_tokens": 16384,
|
||||
"is_moderated": false
|
||||
},
|
||||
"per_request_limits": null
|
||||
@ -5663,7 +5605,7 @@
|
||||
},
|
||||
"top_provider": {
|
||||
"context_length": 131072,
|
||||
"max_completion_tokens": 8192,
|
||||
"max_completion_tokens": 16384,
|
||||
"is_moderated": false
|
||||
},
|
||||
"per_request_limits": null
|
||||
@ -5727,7 +5669,7 @@
|
||||
},
|
||||
"top_provider": {
|
||||
"context_length": 32768,
|
||||
"max_completion_tokens": 8192,
|
||||
"max_completion_tokens": 16384,
|
||||
"is_moderated": false
|
||||
},
|
||||
"per_request_limits": null
|
||||
@ -6119,7 +6061,7 @@
|
||||
},
|
||||
"top_provider": {
|
||||
"context_length": 131072,
|
||||
"max_completion_tokens": 8192,
|
||||
"max_completion_tokens": 16384,
|
||||
"is_moderated": false
|
||||
},
|
||||
"per_request_limits": null
|
||||
@ -6424,7 +6366,8 @@
|
||||
"modality": "text+image->text",
|
||||
"input_modalities": [
|
||||
"text",
|
||||
"image"
|
||||
"image",
|
||||
"file"
|
||||
],
|
||||
"output_modalities": [
|
||||
"text"
|
||||
@ -6699,7 +6642,7 @@
|
||||
},
|
||||
"top_provider": {
|
||||
"context_length": 32768,
|
||||
"max_completion_tokens": 8192,
|
||||
"max_completion_tokens": 16384,
|
||||
"is_moderated": false
|
||||
},
|
||||
"per_request_limits": null
|
||||
@ -6709,7 +6652,7 @@
|
||||
"name": "Meta: Llama 3.1 70B Instruct",
|
||||
"created": 1721692800,
|
||||
"description": "Meta's latest class of model (Llama 3.1) launched with a variety of sizes & flavors. This 70B instruct-tuned version is optimized for high quality dialogue usecases.\n\nIt has demonstrated strong performance compared to leading closed-source models in human evaluations.\n\nTo read more about the model release, [click here](https://ai.meta.com/blog/meta-llama-3-1/). Usage of this model is subject to [Meta's Acceptable Use Policy](https://llama.meta.com/llama3/use-policy/).",
|
||||
"context_length": 32768,
|
||||
"context_length": 131072,
|
||||
"architecture": {
|
||||
"modality": "text->text",
|
||||
"input_modalities": [
|
||||
@ -6722,16 +6665,16 @@
|
||||
"instruct_type": "llama3"
|
||||
},
|
||||
"pricing": {
|
||||
"prompt": "0.000000119",
|
||||
"completion": "0.00000039",
|
||||
"prompt": "0.0000001",
|
||||
"completion": "0.00000028",
|
||||
"request": "0",
|
||||
"image": "0",
|
||||
"web_search": "0",
|
||||
"internal_reasoning": "0"
|
||||
},
|
||||
"top_provider": {
|
||||
"context_length": 32768,
|
||||
"max_completion_tokens": 32768,
|
||||
"context_length": 131072,
|
||||
"max_completion_tokens": 16384,
|
||||
"is_moderated": false
|
||||
},
|
||||
"per_request_limits": null
|
||||
@ -6805,7 +6748,7 @@
|
||||
"name": "Mistral: Mistral Nemo",
|
||||
"created": 1721347200,
|
||||
"description": "A 12B parameter model with a 128k token context length built by Mistral in collaboration with NVIDIA.\n\nThe model is multilingual, supporting English, French, German, Spanish, Italian, Portuguese, Chinese, Japanese, Korean, Arabic, and Hindi.\n\nIt supports function calling and is released under the Apache 2.0 license.",
|
||||
"context_length": 131072,
|
||||
"context_length": 98304,
|
||||
"architecture": {
|
||||
"modality": "text->text",
|
||||
"input_modalities": [
|
||||
@ -6818,16 +6761,16 @@
|
||||
"instruct_type": "mistral"
|
||||
},
|
||||
"pricing": {
|
||||
"prompt": "0.000000035",
|
||||
"completion": "0.00000008",
|
||||
"prompt": "0.00000003",
|
||||
"completion": "0.00000007",
|
||||
"request": "0",
|
||||
"image": "0",
|
||||
"web_search": "0",
|
||||
"internal_reasoning": "0"
|
||||
},
|
||||
"top_provider": {
|
||||
"context_length": 131072,
|
||||
"max_completion_tokens": 8192,
|
||||
"context_length": 98304,
|
||||
"max_completion_tokens": 49152,
|
||||
"is_moderated": false
|
||||
},
|
||||
"per_request_limits": null
|
||||
@ -6842,7 +6785,8 @@
|
||||
"modality": "text+image->text",
|
||||
"input_modalities": [
|
||||
"text",
|
||||
"image"
|
||||
"image",
|
||||
"file"
|
||||
],
|
||||
"output_modalities": [
|
||||
"text"
|
||||
@ -6876,7 +6820,8 @@
|
||||
"modality": "text+image->text",
|
||||
"input_modalities": [
|
||||
"text",
|
||||
"image"
|
||||
"image",
|
||||
"file"
|
||||
],
|
||||
"output_modalities": [
|
||||
"text"
|
||||
@ -7285,7 +7230,7 @@
|
||||
},
|
||||
"top_provider": {
|
||||
"context_length": 32768,
|
||||
"max_completion_tokens": 8192,
|
||||
"max_completion_tokens": 16384,
|
||||
"is_moderated": false
|
||||
},
|
||||
"per_request_limits": null
|
||||
@ -7309,7 +7254,7 @@
|
||||
},
|
||||
"pricing": {
|
||||
"prompt": "0.000000029",
|
||||
"completion": "0.000000059",
|
||||
"completion": "0.000000055",
|
||||
"request": "0",
|
||||
"image": "0",
|
||||
"web_search": "0",
|
||||
@ -7317,7 +7262,7 @@
|
||||
},
|
||||
"top_provider": {
|
||||
"context_length": 32768,
|
||||
"max_completion_tokens": null,
|
||||
"max_completion_tokens": 16384,
|
||||
"is_moderated": false
|
||||
},
|
||||
"per_request_limits": null
|
||||
@ -7341,7 +7286,7 @@
|
||||
},
|
||||
"pricing": {
|
||||
"prompt": "0.000000029",
|
||||
"completion": "0.000000059",
|
||||
"completion": "0.000000055",
|
||||
"request": "0",
|
||||
"image": "0",
|
||||
"web_search": "0",
|
||||
@ -7349,7 +7294,7 @@
|
||||
},
|
||||
"top_provider": {
|
||||
"context_length": 32768,
|
||||
"max_completion_tokens": null,
|
||||
"max_completion_tokens": 16384,
|
||||
"is_moderated": false
|
||||
},
|
||||
"per_request_limits": null
|
||||
@ -7527,7 +7472,8 @@
|
||||
"modality": "text+image->text",
|
||||
"input_modalities": [
|
||||
"text",
|
||||
"image"
|
||||
"image",
|
||||
"file"
|
||||
],
|
||||
"output_modalities": [
|
||||
"text"
|
||||
@ -7561,7 +7507,8 @@
|
||||
"modality": "text+image->text",
|
||||
"input_modalities": [
|
||||
"text",
|
||||
"image"
|
||||
"image",
|
||||
"file"
|
||||
],
|
||||
"output_modalities": [
|
||||
"text"
|
||||
@ -7584,39 +7531,6 @@
|
||||
},
|
||||
"per_request_limits": null
|
||||
},
|
||||
{
|
||||
"id": "openai/gpt-4o-2024-05-13",
|
||||
"name": "OpenAI: GPT-4o (2024-05-13)",
|
||||
"created": 1715558400,
|
||||
"description": "GPT-4o (\"o\" for \"omni\") is OpenAI's latest AI model, supporting both text and image inputs with text outputs. It maintains the intelligence level of [GPT-4 Turbo](/models/openai/gpt-4-turbo) while being twice as fast and 50% more cost-effective. GPT-4o also offers improved performance in processing non-English languages and enhanced visual capabilities.\n\nFor benchmarking against other models, it was briefly called [\"im-also-a-good-gpt2-chatbot\"](https://twitter.com/LiamFedus/status/1790064963966370209)\n\n#multimodal",
|
||||
"context_length": 128000,
|
||||
"architecture": {
|
||||
"modality": "text+image->text",
|
||||
"input_modalities": [
|
||||
"text",
|
||||
"image"
|
||||
],
|
||||
"output_modalities": [
|
||||
"text"
|
||||
],
|
||||
"tokenizer": "GPT",
|
||||
"instruct_type": null
|
||||
},
|
||||
"pricing": {
|
||||
"prompt": "0.000005",
|
||||
"completion": "0.000015",
|
||||
"request": "0",
|
||||
"image": "0.007225",
|
||||
"web_search": "0",
|
||||
"internal_reasoning": "0"
|
||||
},
|
||||
"top_provider": {
|
||||
"context_length": 128000,
|
||||
"max_completion_tokens": 4096,
|
||||
"is_moderated": true
|
||||
},
|
||||
"per_request_limits": null
|
||||
},
|
||||
{
|
||||
"id": "meta-llama/llama-guard-2-8b",
|
||||
"name": "Meta: LlamaGuard 2 8B",
|
||||
@ -7649,6 +7563,40 @@
|
||||
},
|
||||
"per_request_limits": null
|
||||
},
|
||||
{
|
||||
"id": "openai/gpt-4o-2024-05-13",
|
||||
"name": "OpenAI: GPT-4o (2024-05-13)",
|
||||
"created": 1715558400,
|
||||
"description": "GPT-4o (\"o\" for \"omni\") is OpenAI's latest AI model, supporting both text and image inputs with text outputs. It maintains the intelligence level of [GPT-4 Turbo](/models/openai/gpt-4-turbo) while being twice as fast and 50% more cost-effective. GPT-4o also offers improved performance in processing non-English languages and enhanced visual capabilities.\n\nFor benchmarking against other models, it was briefly called [\"im-also-a-good-gpt2-chatbot\"](https://twitter.com/LiamFedus/status/1790064963966370209)\n\n#multimodal",
|
||||
"context_length": 128000,
|
||||
"architecture": {
|
||||
"modality": "text+image->text",
|
||||
"input_modalities": [
|
||||
"text",
|
||||
"image",
|
||||
"file"
|
||||
],
|
||||
"output_modalities": [
|
||||
"text"
|
||||
],
|
||||
"tokenizer": "GPT",
|
||||
"instruct_type": null
|
||||
},
|
||||
"pricing": {
|
||||
"prompt": "0.000005",
|
||||
"completion": "0.000015",
|
||||
"request": "0",
|
||||
"image": "0.007225",
|
||||
"web_search": "0",
|
||||
"internal_reasoning": "0"
|
||||
},
|
||||
"top_provider": {
|
||||
"context_length": 128000,
|
||||
"max_completion_tokens": 4096,
|
||||
"is_moderated": true
|
||||
},
|
||||
"per_request_limits": null
|
||||
},
|
||||
{
|
||||
"id": "neversleep/llama-3-lumimaid-8b:extended",
|
||||
"name": "NeverSleep: Llama 3 Lumimaid 8B (extended)",
|
||||
@ -7772,7 +7720,7 @@
|
||||
},
|
||||
"top_provider": {
|
||||
"context_length": 8192,
|
||||
"max_completion_tokens": 8192,
|
||||
"max_completion_tokens": 16384,
|
||||
"is_moderated": false
|
||||
},
|
||||
"per_request_limits": null
|
||||
@ -7804,7 +7752,7 @@
|
||||
},
|
||||
"top_provider": {
|
||||
"context_length": 8192,
|
||||
"max_completion_tokens": 8192,
|
||||
"max_completion_tokens": 16384,
|
||||
"is_moderated": false
|
||||
},
|
||||
"per_request_limits": null
|
||||
@ -7868,7 +7816,7 @@
|
||||
},
|
||||
"top_provider": {
|
||||
"context_length": 65536,
|
||||
"max_completion_tokens": 8192,
|
||||
"max_completion_tokens": 16384,
|
||||
"is_moderated": false
|
||||
},
|
||||
"per_request_limits": null
|
||||
@ -8694,38 +8642,6 @@
|
||||
},
|
||||
"per_request_limits": null
|
||||
},
|
||||
{
|
||||
"id": "google/gemini-pro",
|
||||
"name": "Google: Gemini Pro 1.0",
|
||||
"created": 1702425600,
|
||||
"description": "Google's flagship text generation model. Designed to handle natural language tasks, multiturn text and code chat, and code generation.\n\nSee the benchmarks and prompting guidelines from [Deepmind](https://deepmind.google/technologies/gemini/).\n\nUsage of Gemini is subject to Google's [Gemini Terms of Use](https://ai.google.dev/terms).",
|
||||
"context_length": 32760,
|
||||
"architecture": {
|
||||
"modality": "text->text",
|
||||
"input_modalities": [
|
||||
"text"
|
||||
],
|
||||
"output_modalities": [
|
||||
"text"
|
||||
],
|
||||
"tokenizer": "Gemini",
|
||||
"instruct_type": null
|
||||
},
|
||||
"pricing": {
|
||||
"prompt": "0.0000005",
|
||||
"completion": "0.0000015",
|
||||
"request": "0",
|
||||
"image": "0.0025",
|
||||
"web_search": "0",
|
||||
"internal_reasoning": "0"
|
||||
},
|
||||
"top_provider": {
|
||||
"context_length": 32760,
|
||||
"max_completion_tokens": 8192,
|
||||
"is_moderated": false
|
||||
},
|
||||
"per_request_limits": null
|
||||
},
|
||||
{
|
||||
"id": "mistralai/mixtral-8x7b-instruct",
|
||||
"name": "Mistral: Mixtral 8x7B Instruct",
|
||||
@ -8753,7 +8669,7 @@
|
||||
},
|
||||
"top_provider": {
|
||||
"context_length": 32768,
|
||||
"max_completion_tokens": 8192,
|
||||
"max_completion_tokens": 16384,
|
||||
"is_moderated": false
|
||||
},
|
||||
"per_request_limits": null
|
||||
|
||||
@ -1,31 +1,28 @@
|
||||
export declare enum E_OPENAI_MODEL {
|
||||
MODEL_GPT_4O_AUDIO_PREVIEW_2024_12_17 = "gpt-4o-audio-preview-2024-12-17",
|
||||
MODEL_DALL_E_3 = "dall-e-3",
|
||||
MODEL_TEXT_EMBEDDING_3_LARGE = "text-embedding-3-large",
|
||||
MODEL_DALL_E_2 = "dall-e-2",
|
||||
MODEL_O4_MINI_2025_04_16 = "o4-mini-2025-04-16",
|
||||
MODEL_GPT_4O_AUDIO_PREVIEW_2024_10_01 = "gpt-4o-audio-preview-2024-10-01",
|
||||
MODEL_TEXT_EMBEDDING_3_SMALL = "text-embedding-3-small",
|
||||
MODEL_O4_MINI = "o4-mini",
|
||||
MODEL_GPT_4_1_NANO = "gpt-4.1-nano",
|
||||
MODEL_GPT_4_1_NANO_2025_04_14 = "gpt-4.1-nano-2025-04-14",
|
||||
MODEL_GPT_4O_REALTIME_PREVIEW_2024_10_01 = "gpt-4o-realtime-preview-2024-10-01",
|
||||
MODEL_O4_MINI_2025_04_16 = "o4-mini-2025-04-16",
|
||||
MODEL_GPT_4O_REALTIME_PREVIEW = "gpt-4o-realtime-preview",
|
||||
MODEL_BABBAGE_002 = "babbage-002",
|
||||
MODEL_GPT_4 = "gpt-4",
|
||||
MODEL_TEXT_EMBEDDING_ADA_002 = "text-embedding-ada-002",
|
||||
MODEL_TEXT_EMBEDDING_3_LARGE = "text-embedding-3-large",
|
||||
MODEL_GPT_4O_MINI_AUDIO_PREVIEW = "gpt-4o-mini-audio-preview",
|
||||
MODEL_GPT_4O_AUDIO_PREVIEW = "gpt-4o-audio-preview",
|
||||
MODEL_O1_PREVIEW_2024_09_12 = "o1-preview-2024-09-12",
|
||||
MODEL_O1_PRO = "o1-pro",
|
||||
MODEL_O1_2024_12_17 = "o1-2024-12-17",
|
||||
MODEL_GPT_4O_MINI_REALTIME_PREVIEW = "gpt-4o-mini-realtime-preview",
|
||||
MODEL_GPT_4_1_MINI = "gpt-4.1-mini",
|
||||
MODEL_GPT_4O_MINI_REALTIME_PREVIEW_2024_12_17 = "gpt-4o-mini-realtime-preview-2024-12-17",
|
||||
MODEL_GPT_3_5_TURBO_INSTRUCT_0914 = "gpt-3.5-turbo-instruct-0914",
|
||||
MODEL_GPT_4O_MINI_SEARCH_PREVIEW = "gpt-4o-mini-search-preview",
|
||||
MODEL_O1 = "o1",
|
||||
MODEL_GPT_4_1_MINI_2025_04_14 = "gpt-4.1-mini-2025-04-14",
|
||||
MODEL_O1_PRO_2025_03_19 = "o1-pro-2025-03-19",
|
||||
MODEL_CHATGPT_4O_LATEST = "chatgpt-4o-latest",
|
||||
MODEL_DAVINCI_002 = "davinci-002",
|
||||
MODEL_GPT_3_5_TURBO_1106 = "gpt-3.5-turbo-1106",
|
||||
@ -45,13 +42,11 @@ export declare enum E_OPENAI_MODEL {
|
||||
MODEL_O1_PREVIEW = "o1-preview",
|
||||
MODEL_GPT_4_0613 = "gpt-4-0613",
|
||||
MODEL_GPT_4_5_PREVIEW = "gpt-4.5-preview",
|
||||
MODEL_O3_MINI = "o3-mini",
|
||||
MODEL_O3_MINI_2025_01_31 = "o3-mini-2025-01-31",
|
||||
MODEL_GPT_4_5_PREVIEW_2025_02_27 = "gpt-4.5-preview-2025-02-27",
|
||||
MODEL_GPT_4O_SEARCH_PREVIEW_2025_03_11 = "gpt-4o-search-preview-2025-03-11",
|
||||
MODEL_OMNI_MODERATION_2024_09_26 = "omni-moderation-2024-09-26",
|
||||
MODEL_TEXT_EMBEDDING_3_SMALL = "text-embedding-3-small",
|
||||
MODEL_GPT_4O_MINI_TTS = "gpt-4o-mini-tts",
|
||||
MODEL_O3_MINI_2025_01_31 = "o3-mini-2025-01-31",
|
||||
MODEL_O3_MINI = "o3-mini",
|
||||
MODEL_TTS_1_HD = "tts-1-hd",
|
||||
MODEL_GPT_4O = "gpt-4o",
|
||||
MODEL_TTS_1_HD_1106 = "tts-1-hd-1106",
|
||||
@ -60,6 +55,7 @@ export declare enum E_OPENAI_MODEL {
|
||||
MODEL_GPT_4_1 = "gpt-4.1",
|
||||
MODEL_GPT_4O_TRANSCRIBE = "gpt-4o-transcribe",
|
||||
MODEL_GPT_4_1_2025_04_14 = "gpt-4.1-2025-04-14",
|
||||
MODEL_O1_2024_12_17 = "o1-2024-12-17",
|
||||
MODEL_GPT_4O_MINI_2024_07_18 = "gpt-4o-mini-2024-07-18",
|
||||
MODEL_GPT_4O_MINI_TRANSCRIBE = "gpt-4o-mini-transcribe",
|
||||
MODEL_O1_MINI = "o1-mini",
|
||||
@ -68,6 +64,10 @@ export declare enum E_OPENAI_MODEL {
|
||||
MODEL_O1_MINI_2024_09_12 = "o1-mini-2024-09-12",
|
||||
MODEL_TTS_1 = "tts-1",
|
||||
MODEL_GPT_4_1106_PREVIEW = "gpt-4-1106-preview",
|
||||
MODEL_GPT_4O_MINI_TTS = "gpt-4o-mini-tts",
|
||||
MODEL_TTS_1_1106 = "tts-1-1106",
|
||||
MODEL_O1 = "o1",
|
||||
MODEL_O1_PRO = "o1-pro",
|
||||
MODEL_O1_PRO_2025_03_19 = "o1-pro-2025-03-19",
|
||||
MODEL_OMNI_MODERATION_LATEST = "omni-moderation-latest"
|
||||
}
|
||||
|
||||
@ -2,31 +2,28 @@ export var E_OPENAI_MODEL;
|
||||
(function (E_OPENAI_MODEL) {
|
||||
E_OPENAI_MODEL["MODEL_GPT_4O_AUDIO_PREVIEW_2024_12_17"] = "gpt-4o-audio-preview-2024-12-17";
|
||||
E_OPENAI_MODEL["MODEL_DALL_E_3"] = "dall-e-3";
|
||||
E_OPENAI_MODEL["MODEL_TEXT_EMBEDDING_3_LARGE"] = "text-embedding-3-large";
|
||||
E_OPENAI_MODEL["MODEL_DALL_E_2"] = "dall-e-2";
|
||||
E_OPENAI_MODEL["MODEL_O4_MINI_2025_04_16"] = "o4-mini-2025-04-16";
|
||||
E_OPENAI_MODEL["MODEL_GPT_4O_AUDIO_PREVIEW_2024_10_01"] = "gpt-4o-audio-preview-2024-10-01";
|
||||
E_OPENAI_MODEL["MODEL_TEXT_EMBEDDING_3_SMALL"] = "text-embedding-3-small";
|
||||
E_OPENAI_MODEL["MODEL_O4_MINI"] = "o4-mini";
|
||||
E_OPENAI_MODEL["MODEL_GPT_4_1_NANO"] = "gpt-4.1-nano";
|
||||
E_OPENAI_MODEL["MODEL_GPT_4_1_NANO_2025_04_14"] = "gpt-4.1-nano-2025-04-14";
|
||||
E_OPENAI_MODEL["MODEL_GPT_4O_REALTIME_PREVIEW_2024_10_01"] = "gpt-4o-realtime-preview-2024-10-01";
|
||||
E_OPENAI_MODEL["MODEL_O4_MINI_2025_04_16"] = "o4-mini-2025-04-16";
|
||||
E_OPENAI_MODEL["MODEL_GPT_4O_REALTIME_PREVIEW"] = "gpt-4o-realtime-preview";
|
||||
E_OPENAI_MODEL["MODEL_BABBAGE_002"] = "babbage-002";
|
||||
E_OPENAI_MODEL["MODEL_GPT_4"] = "gpt-4";
|
||||
E_OPENAI_MODEL["MODEL_TEXT_EMBEDDING_ADA_002"] = "text-embedding-ada-002";
|
||||
E_OPENAI_MODEL["MODEL_TEXT_EMBEDDING_3_LARGE"] = "text-embedding-3-large";
|
||||
E_OPENAI_MODEL["MODEL_GPT_4O_MINI_AUDIO_PREVIEW"] = "gpt-4o-mini-audio-preview";
|
||||
E_OPENAI_MODEL["MODEL_GPT_4O_AUDIO_PREVIEW"] = "gpt-4o-audio-preview";
|
||||
E_OPENAI_MODEL["MODEL_O1_PREVIEW_2024_09_12"] = "o1-preview-2024-09-12";
|
||||
E_OPENAI_MODEL["MODEL_O1_PRO"] = "o1-pro";
|
||||
E_OPENAI_MODEL["MODEL_O1_2024_12_17"] = "o1-2024-12-17";
|
||||
E_OPENAI_MODEL["MODEL_GPT_4O_MINI_REALTIME_PREVIEW"] = "gpt-4o-mini-realtime-preview";
|
||||
E_OPENAI_MODEL["MODEL_GPT_4_1_MINI"] = "gpt-4.1-mini";
|
||||
E_OPENAI_MODEL["MODEL_GPT_4O_MINI_REALTIME_PREVIEW_2024_12_17"] = "gpt-4o-mini-realtime-preview-2024-12-17";
|
||||
E_OPENAI_MODEL["MODEL_GPT_3_5_TURBO_INSTRUCT_0914"] = "gpt-3.5-turbo-instruct-0914";
|
||||
E_OPENAI_MODEL["MODEL_GPT_4O_MINI_SEARCH_PREVIEW"] = "gpt-4o-mini-search-preview";
|
||||
E_OPENAI_MODEL["MODEL_O1"] = "o1";
|
||||
E_OPENAI_MODEL["MODEL_GPT_4_1_MINI_2025_04_14"] = "gpt-4.1-mini-2025-04-14";
|
||||
E_OPENAI_MODEL["MODEL_O1_PRO_2025_03_19"] = "o1-pro-2025-03-19";
|
||||
E_OPENAI_MODEL["MODEL_CHATGPT_4O_LATEST"] = "chatgpt-4o-latest";
|
||||
E_OPENAI_MODEL["MODEL_DAVINCI_002"] = "davinci-002";
|
||||
E_OPENAI_MODEL["MODEL_GPT_3_5_TURBO_1106"] = "gpt-3.5-turbo-1106";
|
||||
@ -46,13 +43,11 @@ export var E_OPENAI_MODEL;
|
||||
E_OPENAI_MODEL["MODEL_O1_PREVIEW"] = "o1-preview";
|
||||
E_OPENAI_MODEL["MODEL_GPT_4_0613"] = "gpt-4-0613";
|
||||
E_OPENAI_MODEL["MODEL_GPT_4_5_PREVIEW"] = "gpt-4.5-preview";
|
||||
E_OPENAI_MODEL["MODEL_O3_MINI"] = "o3-mini";
|
||||
E_OPENAI_MODEL["MODEL_O3_MINI_2025_01_31"] = "o3-mini-2025-01-31";
|
||||
E_OPENAI_MODEL["MODEL_GPT_4_5_PREVIEW_2025_02_27"] = "gpt-4.5-preview-2025-02-27";
|
||||
E_OPENAI_MODEL["MODEL_GPT_4O_SEARCH_PREVIEW_2025_03_11"] = "gpt-4o-search-preview-2025-03-11";
|
||||
E_OPENAI_MODEL["MODEL_OMNI_MODERATION_2024_09_26"] = "omni-moderation-2024-09-26";
|
||||
E_OPENAI_MODEL["MODEL_TEXT_EMBEDDING_3_SMALL"] = "text-embedding-3-small";
|
||||
E_OPENAI_MODEL["MODEL_GPT_4O_MINI_TTS"] = "gpt-4o-mini-tts";
|
||||
E_OPENAI_MODEL["MODEL_O3_MINI_2025_01_31"] = "o3-mini-2025-01-31";
|
||||
E_OPENAI_MODEL["MODEL_O3_MINI"] = "o3-mini";
|
||||
E_OPENAI_MODEL["MODEL_TTS_1_HD"] = "tts-1-hd";
|
||||
E_OPENAI_MODEL["MODEL_GPT_4O"] = "gpt-4o";
|
||||
E_OPENAI_MODEL["MODEL_TTS_1_HD_1106"] = "tts-1-hd-1106";
|
||||
@ -61,6 +56,7 @@ export var E_OPENAI_MODEL;
|
||||
E_OPENAI_MODEL["MODEL_GPT_4_1"] = "gpt-4.1";
|
||||
E_OPENAI_MODEL["MODEL_GPT_4O_TRANSCRIBE"] = "gpt-4o-transcribe";
|
||||
E_OPENAI_MODEL["MODEL_GPT_4_1_2025_04_14"] = "gpt-4.1-2025-04-14";
|
||||
E_OPENAI_MODEL["MODEL_O1_2024_12_17"] = "o1-2024-12-17";
|
||||
E_OPENAI_MODEL["MODEL_GPT_4O_MINI_2024_07_18"] = "gpt-4o-mini-2024-07-18";
|
||||
E_OPENAI_MODEL["MODEL_GPT_4O_MINI_TRANSCRIBE"] = "gpt-4o-mini-transcribe";
|
||||
E_OPENAI_MODEL["MODEL_O1_MINI"] = "o1-mini";
|
||||
@ -69,7 +65,11 @@ export var E_OPENAI_MODEL;
|
||||
E_OPENAI_MODEL["MODEL_O1_MINI_2024_09_12"] = "o1-mini-2024-09-12";
|
||||
E_OPENAI_MODEL["MODEL_TTS_1"] = "tts-1";
|
||||
E_OPENAI_MODEL["MODEL_GPT_4_1106_PREVIEW"] = "gpt-4-1106-preview";
|
||||
E_OPENAI_MODEL["MODEL_GPT_4O_MINI_TTS"] = "gpt-4o-mini-tts";
|
||||
E_OPENAI_MODEL["MODEL_TTS_1_1106"] = "tts-1-1106";
|
||||
E_OPENAI_MODEL["MODEL_O1"] = "o1";
|
||||
E_OPENAI_MODEL["MODEL_O1_PRO"] = "o1-pro";
|
||||
E_OPENAI_MODEL["MODEL_O1_PRO_2025_03_19"] = "o1-pro-2025-03-19";
|
||||
E_OPENAI_MODEL["MODEL_OMNI_MODERATION_LATEST"] = "omni-moderation-latest";
|
||||
})(E_OPENAI_MODEL || (E_OPENAI_MODEL = {}));
|
||||
//# sourceMappingURL=data:application/json;base64,eyJ2ZXJzaW9uIjozLCJmaWxlIjoib3BlbmFpLW1vZGVscy5qcyIsInNvdXJjZVJvb3QiOiIiLCJzb3VyY2VzIjpbIi4uLy4uLy4uL3NyYy9tb2RlbHMvY2FjaGUvb3BlbmFpLW1vZGVscy50cyJdLCJuYW1lcyI6W10sIm1hcHBpbmdzIjoiQUFBQSxNQUFNLENBQU4sSUFBWSxjQXdFWDtBQXhFRCxXQUFZLGNBQWM7SUFDeEIsMkZBQXlFLENBQUE7SUFDekUsNkNBQTJCLENBQUE7SUFDM0IseUVBQXVELENBQUE7SUFDdkQsNkNBQTJCLENBQUE7SUFDM0IsaUVBQStDLENBQUE7SUFDL0MsMkZBQXlFLENBQUE7SUFDekUsMkNBQXlCLENBQUE7SUFDekIscURBQW1DLENBQUE7SUFDbkMsMkVBQXlELENBQUE7SUFDekQsaUdBQStFLENBQUE7SUFDL0UsMkVBQXlELENBQUE7SUFDekQsbURBQWlDLENBQUE7SUFDakMsdUNBQXFCLENBQUE7SUFDckIseUVBQXVELENBQUE7SUFDdkQsK0VBQTZELENBQUE7SUFDN0QscUVBQW1ELENBQUE7SUFDbkQsdUVBQXFELENBQUE7SUFDckQseUNBQXVCLENBQUE7SUFDdkIsdURBQXFDLENBQUE7SUFDckMscUZBQW1FLENBQUE7SUFDbkUscURBQW1DLENBQUE7SUFDbkMsMkdBQXlGLENBQUE7SUFDekYsbUZBQWlFLENBQUE7SUFDakUsaUZBQStELENBQUE7SUFDL0QsaUNBQWUsQ0FBQTtJQUNmLDJFQUF5RCxDQUFBO0lBQ3pELCtEQUE2QyxDQUFBO0lBQzdDLCtEQUE2QyxDQUFBO0lBQzdDLG1EQUFpQyxDQUFBO0lBQ2pDLGlFQUErQyxDQUFBO0lBQy9DLHVFQUFxRCxDQUFBO0lBQ3JELG1EQUFpQyxDQUFBO0lBQ2pDLGlHQUErRSxDQUFBO0lBQy9FLHlFQUF1RCxDQUFBO0lBQ3ZELHVEQUFxQyxDQUFBO0lBQ3JDLG1FQUFpRCxDQUFBO0lBQ2pELHVHQUFxRixDQUFBO0lBQ3JGLGlFQUErQyxDQUFBO0lBQy9DLCtEQUE2QyxDQUFBO0lBQzdDLCtDQUE2QixDQUFBO0lBQzdCLCtEQUE2QyxDQUFBO0lBQzdDLHlFQUF1RCxDQUFBO0lBQ3ZELCtEQUE2QyxDQUFBO0lBQzdDLGlEQUErQixDQUFBO0lBQy9CLGlEQUErQixDQUFBO0lBQy9CLDJEQUF5QyxDQUFBO0lBQ3pDLDJDQUF5QixDQUFBO0lBQ3pCLGlFQUErQyxDQUFBO0lBQy9DLGlGQUErRCxDQUFBO0lBQy9ELDZGQUEyRSxDQUFBO0lBQzNFLGlGQUErRCxDQUFBO0lBQy9ELHlFQUF1RCxDQUFBO0lBQ3ZELDJEQUF5QyxDQUFBO0lBQ3pDLDZDQUEyQixDQUFBO0lBQzNCLHlDQUF1QixDQUFBO0lBQ3ZCLHVEQUFxQyxDQUFBO0lBQ3JDLG1EQUFpQyxDQUFBO0lBQ2pDLCtEQUE2QyxDQUFBO0lBQzdDLDJDQUF5QixDQUFBO0lBQ3pCLCtEQUE2QyxDQUFBO0lBQzdDLGlFQUErQyxDQUFBO0lBQy9DLHlFQUF1RCxDQUFBO0lBQ3ZELHlFQUF1RCxDQUFBO0lBQ3ZELDJDQUF5QixDQUFBO0lBQ3pCLHFHQUFtRixDQUFBO0lBQ25GLGlFQUErQyxDQUFBO0lBQy9DLGlFQUErQyxDQUFBO0lBQy9DLHVDQUFxQixDQUFBO0lBQ3JCLGlFQUErQyxDQUFBO0lBQy9DLGlEQUErQixDQUFBO0lBQy9CLHlFQUF1RCxDQUFBO0FBQ3pELENBQUMsRUF4RVcsY0FBYyxLQUFkLGNBQWMsUUF3RXpCIn0=
|
||||
//# sourceMappingURL=data:application/json;base64,eyJ2ZXJzaW9uIjozLCJmaWxlIjoib3BlbmFpLW1vZGVscy5qcyIsInNvdXJjZVJvb3QiOiIiLCJzb3VyY2VzIjpbIi4uLy4uLy4uL3NyYy9tb2RlbHMvY2FjaGUvb3BlbmFpLW1vZGVscy50cyJdLCJuYW1lcyI6W10sIm1hcHBpbmdzIjoiQUFBQSxNQUFNLENBQU4sSUFBWSxjQXdFWDtBQXhFRCxXQUFZLGNBQWM7SUFDeEIsMkZBQXlFLENBQUE7SUFDekUsNkNBQTJCLENBQUE7SUFDM0IsNkNBQTJCLENBQUE7SUFDM0IsMkZBQXlFLENBQUE7SUFDekUseUVBQXVELENBQUE7SUFDdkQsMkNBQXlCLENBQUE7SUFDekIscURBQW1DLENBQUE7SUFDbkMsMkVBQXlELENBQUE7SUFDekQsaUdBQStFLENBQUE7SUFDL0UsaUVBQStDLENBQUE7SUFDL0MsMkVBQXlELENBQUE7SUFDekQsbURBQWlDLENBQUE7SUFDakMsdUNBQXFCLENBQUE7SUFDckIseUVBQXVELENBQUE7SUFDdkQseUVBQXVELENBQUE7SUFDdkQsK0VBQTZELENBQUE7SUFDN0QscUVBQW1ELENBQUE7SUFDbkQsdUVBQXFELENBQUE7SUFDckQscUZBQW1FLENBQUE7SUFDbkUscURBQW1DLENBQUE7SUFDbkMsMkdBQXlGLENBQUE7SUFDekYsbUZBQWlFLENBQUE7SUFDakUsaUZBQStELENBQUE7SUFDL0QsMkVBQXlELENBQUE7SUFDekQsK0RBQTZDLENBQUE7SUFDN0MsbURBQWlDLENBQUE7SUFDakMsaUVBQStDLENBQUE7SUFDL0MsdUVBQXFELENBQUE7SUFDckQsbURBQWlDLENBQUE7SUFDakMsaUdBQStFLENBQUE7SUFDL0UseUVBQXVELENBQUE7SUFDdkQsdURBQXFDLENBQUE7SUFDckMsbUVBQWlELENBQUE7SUFDakQsdUdBQXFGLENBQUE7SUFDckYsaUVBQStDLENBQUE7SUFDL0MsK0RBQTZDLENBQUE7SUFDN0MsK0NBQTZCLENBQUE7SUFDN0IsK0RBQTZDLENBQUE7SUFDN0MseUVBQXVELENBQUE7SUFDdkQsK0RBQTZDLENBQUE7SUFDN0MsaURBQStCLENBQUE7SUFDL0IsaURBQStCLENBQUE7SUFDL0IsMkRBQXlDLENBQUE7SUFDekMsaUZBQStELENBQUE7SUFDL0QsNkZBQTJFLENBQUE7SUFDM0UsaUZBQStELENBQUE7SUFDL0QsaUVBQStDLENBQUE7SUFDL0MsMkNBQXlCLENBQUE7SUFDekIsNkNBQTJCLENBQUE7SUFDM0IseUNBQXVCLENBQUE7SUFDdkIsdURBQXFDLENBQUE7SUFDckMsbURBQWlDLENBQUE7SUFDakMsK0RBQTZDLENBQUE7SUFDN0MsMkNBQXlCLENBQUE7SUFDekIsK0RBQTZDLENBQUE7SUFDN0MsaUVBQStDLENBQUE7SUFDL0MsdURBQXFDLENBQUE7SUFDckMseUVBQXVELENBQUE7SUFDdkQseUVBQXVELENBQUE7SUFDdkQsMkNBQXlCLENBQUE7SUFDekIscUdBQW1GLENBQUE7SUFDbkYsaUVBQStDLENBQUE7SUFDL0MsaUVBQStDLENBQUE7SUFDL0MsdUNBQXFCLENBQUE7SUFDckIsaUVBQStDLENBQUE7SUFDL0MsMkRBQXlDLENBQUE7SUFDekMsaURBQStCLENBQUE7SUFDL0IsaUNBQWUsQ0FBQTtJQUNmLHlDQUF1QixDQUFBO0lBQ3ZCLCtEQUE2QyxDQUFBO0lBQzdDLHlFQUF1RCxDQUFBO0FBQ3pELENBQUMsRUF4RVcsY0FBYyxLQUFkLGNBQWMsUUF3RXpCIn0=
|
||||
@ -1,4 +1,5 @@
|
||||
export declare enum E_OPENROUTER_MODEL_FREE {
|
||||
MODEL_FREE_MICROSOFT_MAI_DS_R1_FREE = "microsoft/mai-ds-r1:free",
|
||||
MODEL_FREE_THUDM_GLM_Z1_32B_FREE = "thudm/glm-z1-32b:free",
|
||||
MODEL_FREE_THUDM_GLM_4_32B_FREE = "thudm/glm-4-32b:free",
|
||||
MODEL_FREE_SHISA_AI_SHISA_V2_LLAMA3_3_70B_FREE = "shisa-ai/shisa-v2-llama3.3-70b:free",
|
||||
@ -19,7 +20,6 @@ export declare enum E_OPENROUTER_MODEL_FREE {
|
||||
MODEL_FREE_DEEPSEEK_DEEPSEEK_CHAT_V3_0324_FREE = "deepseek/deepseek-chat-v3-0324:free",
|
||||
MODEL_FREE_FEATHERLESS_QWERKY_72B_FREE = "featherless/qwerky-72b:free",
|
||||
MODEL_FREE_MISTRALAI_MISTRAL_SMALL_3_1_24B_INSTRUCT_FREE = "mistralai/mistral-small-3.1-24b-instruct:free",
|
||||
MODEL_FREE_OPEN_R1_OLYMPICCODER_7B_FREE = "open-r1/olympiccoder-7b:free",
|
||||
MODEL_FREE_OPEN_R1_OLYMPICCODER_32B_FREE = "open-r1/olympiccoder-32b:free",
|
||||
MODEL_FREE_GOOGLE_GEMMA_3_1B_IT_FREE = "google/gemma-3-1b-it:free",
|
||||
MODEL_FREE_GOOGLE_GEMMA_3_4B_IT_FREE = "google/gemma-3-4b-it:free",
|
||||
@ -37,11 +37,9 @@ export declare enum E_OPENROUTER_MODEL_FREE {
|
||||
MODEL_FREE_DEEPSEEK_DEEPSEEK_R1_DISTILL_QWEN_32B_FREE = "deepseek/deepseek-r1-distill-qwen-32b:free",
|
||||
MODEL_FREE_DEEPSEEK_DEEPSEEK_R1_DISTILL_QWEN_14B_FREE = "deepseek/deepseek-r1-distill-qwen-14b:free",
|
||||
MODEL_FREE_DEEPSEEK_DEEPSEEK_R1_DISTILL_LLAMA_70B_FREE = "deepseek/deepseek-r1-distill-llama-70b:free",
|
||||
MODEL_FREE_GOOGLE_GEMINI_2_0_FLASH_THINKING_EXP_FREE = "google/gemini-2.0-flash-thinking-exp:free",
|
||||
MODEL_FREE_DEEPSEEK_DEEPSEEK_R1_FREE = "deepseek/deepseek-r1:free",
|
||||
MODEL_FREE_SOPHOSYMPATHEIA_ROGUE_ROSE_103B_V0_2_FREE = "sophosympatheia/rogue-rose-103b-v0.2:free",
|
||||
MODEL_FREE_DEEPSEEK_DEEPSEEK_CHAT_FREE = "deepseek/deepseek-chat:free",
|
||||
MODEL_FREE_GOOGLE_GEMINI_2_0_FLASH_THINKING_EXP_1219_FREE = "google/gemini-2.0-flash-thinking-exp-1219:free",
|
||||
MODEL_FREE_GOOGLE_GEMINI_2_0_FLASH_EXP_FREE = "google/gemini-2.0-flash-exp:free",
|
||||
MODEL_FREE_META_LLAMA_LLAMA_3_3_70B_INSTRUCT_FREE = "meta-llama/llama-3.3-70b-instruct:free",
|
||||
MODEL_FREE_QWEN_QWQ_32B_PREVIEW_FREE = "qwen/qwq-32b-preview:free",
|
||||
|
||||
@ -1,5 +1,6 @@
|
||||
export var E_OPENROUTER_MODEL_FREE;
|
||||
(function (E_OPENROUTER_MODEL_FREE) {
|
||||
E_OPENROUTER_MODEL_FREE["MODEL_FREE_MICROSOFT_MAI_DS_R1_FREE"] = "microsoft/mai-ds-r1:free";
|
||||
E_OPENROUTER_MODEL_FREE["MODEL_FREE_THUDM_GLM_Z1_32B_FREE"] = "thudm/glm-z1-32b:free";
|
||||
E_OPENROUTER_MODEL_FREE["MODEL_FREE_THUDM_GLM_4_32B_FREE"] = "thudm/glm-4-32b:free";
|
||||
E_OPENROUTER_MODEL_FREE["MODEL_FREE_SHISA_AI_SHISA_V2_LLAMA3_3_70B_FREE"] = "shisa-ai/shisa-v2-llama3.3-70b:free";
|
||||
@ -20,7 +21,6 @@ export var E_OPENROUTER_MODEL_FREE;
|
||||
E_OPENROUTER_MODEL_FREE["MODEL_FREE_DEEPSEEK_DEEPSEEK_CHAT_V3_0324_FREE"] = "deepseek/deepseek-chat-v3-0324:free";
|
||||
E_OPENROUTER_MODEL_FREE["MODEL_FREE_FEATHERLESS_QWERKY_72B_FREE"] = "featherless/qwerky-72b:free";
|
||||
E_OPENROUTER_MODEL_FREE["MODEL_FREE_MISTRALAI_MISTRAL_SMALL_3_1_24B_INSTRUCT_FREE"] = "mistralai/mistral-small-3.1-24b-instruct:free";
|
||||
E_OPENROUTER_MODEL_FREE["MODEL_FREE_OPEN_R1_OLYMPICCODER_7B_FREE"] = "open-r1/olympiccoder-7b:free";
|
||||
E_OPENROUTER_MODEL_FREE["MODEL_FREE_OPEN_R1_OLYMPICCODER_32B_FREE"] = "open-r1/olympiccoder-32b:free";
|
||||
E_OPENROUTER_MODEL_FREE["MODEL_FREE_GOOGLE_GEMMA_3_1B_IT_FREE"] = "google/gemma-3-1b-it:free";
|
||||
E_OPENROUTER_MODEL_FREE["MODEL_FREE_GOOGLE_GEMMA_3_4B_IT_FREE"] = "google/gemma-3-4b-it:free";
|
||||
@ -38,11 +38,9 @@ export var E_OPENROUTER_MODEL_FREE;
|
||||
E_OPENROUTER_MODEL_FREE["MODEL_FREE_DEEPSEEK_DEEPSEEK_R1_DISTILL_QWEN_32B_FREE"] = "deepseek/deepseek-r1-distill-qwen-32b:free";
|
||||
E_OPENROUTER_MODEL_FREE["MODEL_FREE_DEEPSEEK_DEEPSEEK_R1_DISTILL_QWEN_14B_FREE"] = "deepseek/deepseek-r1-distill-qwen-14b:free";
|
||||
E_OPENROUTER_MODEL_FREE["MODEL_FREE_DEEPSEEK_DEEPSEEK_R1_DISTILL_LLAMA_70B_FREE"] = "deepseek/deepseek-r1-distill-llama-70b:free";
|
||||
E_OPENROUTER_MODEL_FREE["MODEL_FREE_GOOGLE_GEMINI_2_0_FLASH_THINKING_EXP_FREE"] = "google/gemini-2.0-flash-thinking-exp:free";
|
||||
E_OPENROUTER_MODEL_FREE["MODEL_FREE_DEEPSEEK_DEEPSEEK_R1_FREE"] = "deepseek/deepseek-r1:free";
|
||||
E_OPENROUTER_MODEL_FREE["MODEL_FREE_SOPHOSYMPATHEIA_ROGUE_ROSE_103B_V0_2_FREE"] = "sophosympatheia/rogue-rose-103b-v0.2:free";
|
||||
E_OPENROUTER_MODEL_FREE["MODEL_FREE_DEEPSEEK_DEEPSEEK_CHAT_FREE"] = "deepseek/deepseek-chat:free";
|
||||
E_OPENROUTER_MODEL_FREE["MODEL_FREE_GOOGLE_GEMINI_2_0_FLASH_THINKING_EXP_1219_FREE"] = "google/gemini-2.0-flash-thinking-exp-1219:free";
|
||||
E_OPENROUTER_MODEL_FREE["MODEL_FREE_GOOGLE_GEMINI_2_0_FLASH_EXP_FREE"] = "google/gemini-2.0-flash-exp:free";
|
||||
E_OPENROUTER_MODEL_FREE["MODEL_FREE_META_LLAMA_LLAMA_3_3_70B_INSTRUCT_FREE"] = "meta-llama/llama-3.3-70b-instruct:free";
|
||||
E_OPENROUTER_MODEL_FREE["MODEL_FREE_QWEN_QWQ_32B_PREVIEW_FREE"] = "qwen/qwq-32b-preview:free";
|
||||
@ -63,4 +61,4 @@ export var E_OPENROUTER_MODEL_FREE;
|
||||
E_OPENROUTER_MODEL_FREE["MODEL_FREE_MISTRALAI_MISTRAL_7B_INSTRUCT_FREE"] = "mistralai/mistral-7b-instruct:free";
|
||||
E_OPENROUTER_MODEL_FREE["MODEL_FREE_HUGGINGFACEH4_ZEPHYR_7B_BETA_FREE"] = "huggingfaceh4/zephyr-7b-beta:free";
|
||||
})(E_OPENROUTER_MODEL_FREE || (E_OPENROUTER_MODEL_FREE = {}));
|
||||
//# sourceMappingURL=data:application/json;base64,eyJ2ZXJzaW9uIjozLCJmaWxlIjoib3BlbnJvdXRlci1tb2RlbHMtZnJlZS5qcyIsInNvdXJjZVJvb3QiOiIiLCJzb3VyY2VzIjpbIi4uLy4uLy4uL3NyYy9tb2RlbHMvY2FjaGUvb3BlbnJvdXRlci1tb2RlbHMtZnJlZS50cyJdLCJuYW1lcyI6W10sIm1hcHBpbmdzIjoiQUFBQSxNQUFNLENBQU4sSUFBWSx1QkErRFg7QUEvREQsV0FBWSx1QkFBdUI7SUFDakMscUZBQTBELENBQUE7SUFDMUQsbUZBQXdELENBQUE7SUFDeEQsaUhBQXNGLENBQUE7SUFDdEYsNkdBQWtGLENBQUE7SUFDbEYseUhBQThGLENBQUE7SUFDOUYsbUhBQXdGLENBQUE7SUFDeEYsNkhBQWtHLENBQUE7SUFDbEcsaUlBQXNHLENBQUE7SUFDdEcsbUlBQXdHLENBQUE7SUFDeEcsMkdBQWdGLENBQUE7SUFDaEYscUdBQTBFLENBQUE7SUFDMUUsdUdBQTRFLENBQUE7SUFDNUUseUZBQThELENBQUE7SUFDOUQsaUhBQXNGLENBQUE7SUFDdEYsMkdBQWdGLENBQUE7SUFDaEYsbUhBQXdGLENBQUE7SUFDeEYsNkdBQWtGLENBQUE7SUFDbEYsaUhBQXNGLENBQUE7SUFDdEYsaUdBQXNFLENBQUE7SUFDdEUscUlBQTBHLENBQUE7SUFDMUcsbUdBQXdFLENBQUE7SUFDeEUscUdBQTBFLENBQUE7SUFDMUUsNkZBQWtFLENBQUE7SUFDbEUsNkZBQWtFLENBQUE7SUFDbEUsK0ZBQW9FLENBQUE7SUFDcEUsMkZBQWdFLENBQUE7SUFDaEUsK0ZBQW9FLENBQUE7SUFDcEUsdUdBQTRFLENBQUE7SUFDNUUsNkVBQWtELENBQUE7SUFDbEQsK0hBQW9HLENBQUE7SUFDcEcsNklBQWtILENBQUE7SUFDbEgsbUpBQXdILENBQUE7SUFDeEgsNklBQWtILENBQUE7SUFDbEgsNkdBQWtGLENBQUE7SUFDbEYsdUlBQTRHLENBQUE7SUFDNUcsK0hBQW9HLENBQUE7SUFDcEcsK0hBQW9HLENBQUE7SUFDcEcsaUlBQXNHLENBQUE7SUFDdEcsNkhBQWtHLENBQUE7SUFDbEcsNkZBQWtFLENBQUE7SUFDbEUsNkhBQWtHLENBQUE7SUFDbEcsaUdBQXNFLENBQUE7SUFDdEUsdUlBQTRHLENBQUE7SUFDNUcsMkdBQWdGLENBQUE7SUFDaEYsdUhBQTRGLENBQUE7SUFDNUYsNkZBQWtFLENBQUE7SUFDbEUsMkhBQWdHLENBQUE7SUFDaEcscUhBQTBGLENBQUE7SUFDMUYsdUdBQTRFLENBQUE7SUFDNUUsaUlBQXNHLENBQUE7SUFDdEcscUhBQTBGLENBQUE7SUFDMUYscUhBQTBGLENBQUE7SUFDMUYscUlBQTBHLENBQUE7SUFDMUcseUdBQThFLENBQUE7SUFDOUUsNkdBQWtGLENBQUE7SUFDbEYsdUdBQTRFLENBQUE7SUFDNUUsdUdBQTRFLENBQUE7SUFDNUUscUhBQTBGLENBQUE7SUFDMUYsaUdBQXNFLENBQUE7SUFDdEUsNkZBQWtFLENBQUE7SUFDbEUsK0dBQW9GLENBQUE7SUFDcEYsNkdBQWtGLENBQUE7QUFDcEYsQ0FBQyxFQS9EVyx1QkFBdUIsS0FBdkIsdUJBQXVCLFFBK0RsQyJ9
|
||||
//# sourceMappingURL=data:application/json;base64,eyJ2ZXJzaW9uIjozLCJmaWxlIjoib3BlbnJvdXRlci1tb2RlbHMtZnJlZS5qcyIsInNvdXJjZVJvb3QiOiIiLCJzb3VyY2VzIjpbIi4uLy4uLy4uL3NyYy9tb2RlbHMvY2FjaGUvb3BlbnJvdXRlci1tb2RlbHMtZnJlZS50cyJdLCJuYW1lcyI6W10sIm1hcHBpbmdzIjoiQUFBQSxNQUFNLENBQU4sSUFBWSx1QkE2RFg7QUE3REQsV0FBWSx1QkFBdUI7SUFDakMsMkZBQWdFLENBQUE7SUFDaEUscUZBQTBELENBQUE7SUFDMUQsbUZBQXdELENBQUE7SUFDeEQsaUhBQXNGLENBQUE7SUFDdEYsNkdBQWtGLENBQUE7SUFDbEYseUhBQThGLENBQUE7SUFDOUYsbUhBQXdGLENBQUE7SUFDeEYsNkhBQWtHLENBQUE7SUFDbEcsaUlBQXNHLENBQUE7SUFDdEcsbUlBQXdHLENBQUE7SUFDeEcsMkdBQWdGLENBQUE7SUFDaEYscUdBQTBFLENBQUE7SUFDMUUsdUdBQTRFLENBQUE7SUFDNUUseUZBQThELENBQUE7SUFDOUQsaUhBQXNGLENBQUE7SUFDdEYsMkdBQWdGLENBQUE7SUFDaEYsbUhBQXdGLENBQUE7SUFDeEYsNkdBQWtGLENBQUE7SUFDbEYsaUhBQXNGLENBQUE7SUFDdEYsaUdBQXNFLENBQUE7SUFDdEUscUlBQTBHLENBQUE7SUFDMUcscUdBQTBFLENBQUE7SUFDMUUsNkZBQWtFLENBQUE7SUFDbEUsNkZBQWtFLENBQUE7SUFDbEUsK0ZBQW9FLENBQUE7SUFDcEUsMkZBQWdFLENBQUE7SUFDaEUsK0ZBQW9FLENBQUE7SUFDcEUsdUdBQTRFLENBQUE7SUFDNUUsNkVBQWtELENBQUE7SUFDbEQsK0hBQW9HLENBQUE7SUFDcEcsNklBQWtILENBQUE7SUFDbEgsbUpBQXdILENBQUE7SUFDeEgsNklBQWtILENBQUE7SUFDbEgsNkdBQWtGLENBQUE7SUFDbEYsdUlBQTRHLENBQUE7SUFDNUcsK0hBQW9HLENBQUE7SUFDcEcsK0hBQW9HLENBQUE7SUFDcEcsaUlBQXNHLENBQUE7SUFDdEcsNkZBQWtFLENBQUE7SUFDbEUsNkhBQWtHLENBQUE7SUFDbEcsaUdBQXNFLENBQUE7SUFDdEUsMkdBQWdGLENBQUE7SUFDaEYsdUhBQTRGLENBQUE7SUFDNUYsNkZBQWtFLENBQUE7SUFDbEUsMkhBQWdHLENBQUE7SUFDaEcscUhBQTBGLENBQUE7SUFDMUYsdUdBQTRFLENBQUE7SUFDNUUsaUlBQXNHLENBQUE7SUFDdEcscUhBQTBGLENBQUE7SUFDMUYscUhBQTBGLENBQUE7SUFDMUYscUlBQTBHLENBQUE7SUFDMUcseUdBQThFLENBQUE7SUFDOUUsNkdBQWtGLENBQUE7SUFDbEYsdUdBQTRFLENBQUE7SUFDNUUsdUdBQTRFLENBQUE7SUFDNUUscUhBQTBGLENBQUE7SUFDMUYsaUdBQXNFLENBQUE7SUFDdEUsNkZBQWtFLENBQUE7SUFDbEUsK0dBQW9GLENBQUE7SUFDcEYsNkdBQWtGLENBQUE7QUFDcEYsQ0FBQyxFQTdEVyx1QkFBdUIsS0FBdkIsdUJBQXVCLFFBNkRsQyJ9
|
||||
@ -1,4 +1,5 @@
|
||||
export declare enum E_OPENROUTER_MODEL {
|
||||
MODEL_MICROSOFT_MAI_DS_R1_FREE = "microsoft/mai-ds-r1:free",
|
||||
MODEL_GOOGLE_GEMINI_2_5_PRO_PREVIEW_03_25 = "google/gemini-2.5-pro-preview-03-25",
|
||||
MODEL_THUDM_GLM_Z1_32B_FREE = "thudm/glm-z1-32b:free",
|
||||
MODEL_THUDM_GLM_4_32B_FREE = "thudm/glm-4-32b:free",
|
||||
@ -43,7 +44,6 @@ export declare enum E_OPENROUTER_MODEL {
|
||||
MODEL_OPENAI_O1_PRO = "openai/o1-pro",
|
||||
MODEL_MISTRALAI_MISTRAL_SMALL_3_1_24B_INSTRUCT_FREE = "mistralai/mistral-small-3.1-24b-instruct:free",
|
||||
MODEL_MISTRALAI_MISTRAL_SMALL_3_1_24B_INSTRUCT = "mistralai/mistral-small-3.1-24b-instruct",
|
||||
MODEL_OPEN_R1_OLYMPICCODER_7B_FREE = "open-r1/olympiccoder-7b:free",
|
||||
MODEL_OPEN_R1_OLYMPICCODER_32B_FREE = "open-r1/olympiccoder-32b:free",
|
||||
MODEL_STEELSKULL_L3_3_ELECTRA_R1_70B = "steelskull/l3.3-electra-r1-70b",
|
||||
MODEL_GOOGLE_GEMMA_3_1B_IT_FREE = "google/gemma-3-1b-it:free",
|
||||
@ -108,7 +108,6 @@ export declare enum E_OPENROUTER_MODEL {
|
||||
MODEL_LIQUID_LFM_3B = "liquid/lfm-3b",
|
||||
MODEL_DEEPSEEK_DEEPSEEK_R1_DISTILL_LLAMA_70B_FREE = "deepseek/deepseek-r1-distill-llama-70b:free",
|
||||
MODEL_DEEPSEEK_DEEPSEEK_R1_DISTILL_LLAMA_70B = "deepseek/deepseek-r1-distill-llama-70b",
|
||||
MODEL_GOOGLE_GEMINI_2_0_FLASH_THINKING_EXP_FREE = "google/gemini-2.0-flash-thinking-exp:free",
|
||||
MODEL_DEEPSEEK_DEEPSEEK_R1_FREE = "deepseek/deepseek-r1:free",
|
||||
MODEL_DEEPSEEK_DEEPSEEK_R1 = "deepseek/deepseek-r1",
|
||||
MODEL_SOPHOSYMPATHEIA_ROGUE_ROSE_103B_V0_2_FREE = "sophosympatheia/rogue-rose-103b-v0.2:free",
|
||||
@ -117,7 +116,6 @@ export declare enum E_OPENROUTER_MODEL {
|
||||
MODEL_MICROSOFT_PHI_4 = "microsoft/phi-4",
|
||||
MODEL_DEEPSEEK_DEEPSEEK_CHAT_FREE = "deepseek/deepseek-chat:free",
|
||||
MODEL_DEEPSEEK_DEEPSEEK_CHAT = "deepseek/deepseek-chat",
|
||||
MODEL_GOOGLE_GEMINI_2_0_FLASH_THINKING_EXP_1219_FREE = "google/gemini-2.0-flash-thinking-exp-1219:free",
|
||||
MODEL_SAO10K_L3_3_EURYALE_70B = "sao10k/l3.3-euryale-70b",
|
||||
MODEL_OPENAI_O1 = "openai/o1",
|
||||
MODEL_EVA_UNIT_01_EVA_LLAMA_3_33_70B = "eva-unit-01/eva-llama-3.33-70b",
|
||||
@ -232,8 +230,8 @@ export declare enum E_OPENROUTER_MODEL {
|
||||
MODEL_GOOGLE_GEMINI_FLASH_1_5 = "google/gemini-flash-1.5",
|
||||
MODEL_OPENAI_GPT_4O = "openai/gpt-4o",
|
||||
MODEL_OPENAI_GPT_4O_EXTENDED = "openai/gpt-4o:extended",
|
||||
MODEL_OPENAI_GPT_4O_2024_05_13 = "openai/gpt-4o-2024-05-13",
|
||||
MODEL_META_LLAMA_LLAMA_GUARD_2_8B = "meta-llama/llama-guard-2-8b",
|
||||
MODEL_OPENAI_GPT_4O_2024_05_13 = "openai/gpt-4o-2024-05-13",
|
||||
MODEL_NEVERSLEEP_LLAMA_3_LUMIMAID_8B_EXTENDED = "neversleep/llama-3-lumimaid-8b:extended",
|
||||
MODEL_NEVERSLEEP_LLAMA_3_LUMIMAID_8B = "neversleep/llama-3-lumimaid-8b",
|
||||
MODEL_SAO10K_FIMBULVETR_11B_V2 = "sao10k/fimbulvetr-11b-v2",
|
||||
@ -266,7 +264,6 @@ export declare enum E_OPENROUTER_MODEL {
|
||||
MODEL_MISTRALAI_MISTRAL_7B_INSTRUCT_V0_2 = "mistralai/mistral-7b-instruct-v0.2",
|
||||
MODEL_COGNITIVECOMPUTATIONS_DOLPHIN_MIXTRAL_8X7B = "cognitivecomputations/dolphin-mixtral-8x7b",
|
||||
MODEL_GOOGLE_GEMINI_PRO_VISION = "google/gemini-pro-vision",
|
||||
MODEL_GOOGLE_GEMINI_PRO = "google/gemini-pro",
|
||||
MODEL_MISTRALAI_MIXTRAL_8X7B_INSTRUCT = "mistralai/mixtral-8x7b-instruct",
|
||||
MODEL_OPENCHAT_OPENCHAT_7B = "openchat/openchat-7b",
|
||||
MODEL_NEVERSLEEP_NOROMAID_20B = "neversleep/noromaid-20b",
|
||||
|
||||
File diff suppressed because one or more lines are too long
File diff suppressed because one or more lines are too long
File diff suppressed because one or more lines are too long
3
packages/kbot/dist-in/zod_schema.d.ts
vendored
3
packages/kbot/dist-in/zod_schema.d.ts
vendored
@ -12,7 +12,8 @@ export declare const E_Mode: {
|
||||
readonly CUSTOM: "custom";
|
||||
};
|
||||
export declare const EType: z.ZodEnum<["completion", "tools", "assistant", "custom"]>;
|
||||
export declare const SplitType: z.ZodEnum<["none", "langchain/CharacterTextSplitter", "langchain/LatexTextSplitter", "langchain/MarkdownTextSplitter", "langchain/RecursiveCharacterTextSplitter", "langchain/TokenTextSplitter"]>;
|
||||
export declare const E_AppendMode: z.ZodEnum<["concat", "merge"]>;
|
||||
export type E_AppendModeType = z.infer<typeof E_AppendMode>;
|
||||
export type OptionsSchemaMeta = Record<string, unknown>;
|
||||
export { fetchOpenRouterModels, listModelsAsStrings as listOpenRouterModelsAsStrings } from './models/openrouter.js';
|
||||
export { fetchOpenAIModels, listModelsAsStrings as listOpenAIModelsAsStrings } from './models/openai.js';
|
||||
|
||||
File diff suppressed because one or more lines are too long
5
packages/kbot/dist-in/zod_types.d.ts
vendored
5
packages/kbot/dist-in/zod_types.d.ts
vendored
@ -7,6 +7,8 @@ export interface IKBotOptions {
|
||||
output?: string | undefined;
|
||||
/** Optional destination path for the result, will substitute ${MODEL_NAME} and ${ROUTER} in the path. Optional, used for "completion" mode */
|
||||
dst?: string | undefined;
|
||||
/** How to handle output if --dst file already exists: "concat" (append) or "merge" (try to merge structures if possible, otherwise append). Only used if --dst is specified. */
|
||||
append?: ("concat" | "merge") | undefined;
|
||||
/** Iterate over items, supported: GLOB | Path to JSON File | array of strings (comma separated). To test different models, use --each="gpt-3.5-turbo,gpt-4o", the actual string will exposed as variable `ITEM`, eg: --dst="${ITEM}-output.md" */
|
||||
each?: string | undefined;
|
||||
/** Disable tools categories, eg: --disable=fs,git,interact,terminal,search,web,email,user */
|
||||
@ -149,6 +151,7 @@ export interface IKBotOptions {
|
||||
meta-llama/llama-3-70b-instruct | paid
|
||||
meta-llama/llama-3-8b-instruct | paid
|
||||
meta-llama/llama-3.1-405b | paid
|
||||
meta-llama/llama-3.1-405b:free | free
|
||||
meta-llama/llama-3.1-405b-instruct | paid
|
||||
meta-llama/llama-3.1-70b-instruct | paid
|
||||
meta-llama/llama-3.1-8b-instruct | paid
|
||||
@ -305,7 +308,6 @@ export interface IKBotOptions {
|
||||
sophosympatheia/rogue-rose-103b-v0.2:free | free
|
||||
sao10k/l3-lunaris-8b | paid
|
||||
sao10k/l3-euryale-70b | paid
|
||||
sao10k/l3.1-70b-hanami-x1 | paid
|
||||
sao10k/l3.1-euryale-70b | paid
|
||||
sao10k/l3.3-euryale-70b | paid
|
||||
shisa-ai/shisa-v2-llama3.3-70b:free | free
|
||||
@ -327,7 +329,6 @@ export interface IKBotOptions {
|
||||
x-ai/grok-3-mini-beta | paid
|
||||
x-ai/grok-beta | paid
|
||||
x-ai/grok-vision-beta | paid
|
||||
xwin-lm/xwin-lm-70b | paid
|
||||
[35m[1m[22m[39m
|
||||
[35m[1m OpenAI models:[22m[39m
|
||||
[35m[1m[22m[39m
|
||||
|
||||
523
packages/kbot/dist/main_node.js
vendored
523
packages/kbot/dist/main_node.js
vendored
File diff suppressed because one or more lines are too long
4
packages/kbot/dist/package-lock.json
generated
vendored
4
packages/kbot/dist/package-lock.json
generated
vendored
@ -1,12 +1,12 @@
|
||||
{
|
||||
"name": "@plastichub/kbot",
|
||||
"version": "1.1.27",
|
||||
"version": "1.1.28",
|
||||
"lockfileVersion": 3,
|
||||
"requires": true,
|
||||
"packages": {
|
||||
"": {
|
||||
"name": "@plastichub/kbot",
|
||||
"version": "1.1.27",
|
||||
"version": "1.1.28",
|
||||
"license": "ISC",
|
||||
"dependencies": {
|
||||
"node-emoji": "^2.2.0"
|
||||
|
||||
2
packages/kbot/dist/package.json
vendored
2
packages/kbot/dist/package.json
vendored
@ -1,6 +1,6 @@
|
||||
{
|
||||
"name": "@plastichub/kbot",
|
||||
"version": "1.1.27",
|
||||
"version": "1.1.28",
|
||||
"main": "main_node.js",
|
||||
"author": "",
|
||||
"license": "ISC",
|
||||
|
||||
24
packages/kbot/package-lock.json
generated
24
packages/kbot/package-lock.json
generated
@ -29,7 +29,7 @@
|
||||
"marked": "14.1.4",
|
||||
"marked-terminal": "7.2.1",
|
||||
"mime-types": "2.1.35",
|
||||
"openai": "4.91.0",
|
||||
"openai": "4.96.0",
|
||||
"p-map": "7.0.3",
|
||||
"p-throttle": "7.0.0",
|
||||
"remark-parse": "11.0.0",
|
||||
@ -40,7 +40,7 @@
|
||||
"unified": "11.0.5",
|
||||
"unist-util-visit": "5.0.0",
|
||||
"yargs": "17.7.2",
|
||||
"zod": "3.24.2"
|
||||
"zod": "3.24.3"
|
||||
},
|
||||
"bin": {
|
||||
"kbot-d": "dist-in/main.js"
|
||||
@ -84,7 +84,7 @@
|
||||
"marked": "^15.0.4",
|
||||
"mime-types": "^2.1.35",
|
||||
"nodemailer": "^6.9.16",
|
||||
"openai": "^4.87.4",
|
||||
"openai": "^4.96.0",
|
||||
"p-map": "^7.0.3",
|
||||
"rotating-file-stream": "^3.2.6",
|
||||
"screenshot-desktop": "^1.15.0",
|
||||
@ -96,7 +96,7 @@
|
||||
"type-fest": "^4.30.2",
|
||||
"winston": "^3.17.0",
|
||||
"yargs": "^17.7.2",
|
||||
"zod": "^3.24.1",
|
||||
"zod": "^3.24.3",
|
||||
"zod-to-json-schema": "^3.24.1"
|
||||
},
|
||||
"bin": {
|
||||
@ -155,8 +155,8 @@
|
||||
"tslog": "^3.3.3",
|
||||
"tsup": "^2.0.3",
|
||||
"yargs": "^17.7.2",
|
||||
"zod": "^3.24.2",
|
||||
"zod-to-json-schema": "^3.24.1",
|
||||
"zod": "^3.24.3",
|
||||
"zod-to-json-schema": "^3.24.5",
|
||||
"zod-to-ts": "^1.2.0"
|
||||
},
|
||||
"devDependencies": {
|
||||
@ -5868,9 +5868,9 @@
|
||||
}
|
||||
},
|
||||
"node_modules/openai": {
|
||||
"version": "4.91.0",
|
||||
"resolved": "https://registry.npmjs.org/openai/-/openai-4.91.0.tgz",
|
||||
"integrity": "sha512-zdDg6eyvUmCP58QAW7/aPb+XdeavJ51pK6AcwZOWG5QNSLIovVz0XonRL9vARGJRmw8iImmvf2A31Q7hoh544w==",
|
||||
"version": "4.96.0",
|
||||
"resolved": "https://registry.npmjs.org/openai/-/openai-4.96.0.tgz",
|
||||
"integrity": "sha512-dKoW56i02Prv2XQolJ9Rl9Svqubqkzg3QpwEOBuSVZLk05Shelu7s+ErRTwFc1Bs3JZ2qBqBfVpXQiJhwOGG8A==",
|
||||
"license": "Apache-2.0",
|
||||
"dependencies": {
|
||||
"@types/node": "^18.11.18",
|
||||
@ -8843,9 +8843,9 @@
|
||||
}
|
||||
},
|
||||
"node_modules/zod": {
|
||||
"version": "3.24.2",
|
||||
"resolved": "https://registry.npmjs.org/zod/-/zod-3.24.2.tgz",
|
||||
"integrity": "sha512-lY7CDW43ECgW9u1TcT3IoXHflywfVqDYze4waEz812jR/bZ8FHDsl7pFQoSZTz5N+2NqRXs8GBwnAwo3ZNxqhQ==",
|
||||
"version": "3.24.3",
|
||||
"resolved": "https://registry.npmjs.org/zod/-/zod-3.24.3.tgz",
|
||||
"integrity": "sha512-HhY1oqzWCQWuUqvBFnsyrtZRhyPeR7SUGv+C4+MsisMuVfSPx8HpwWqH8tRahSlt6M3PiFAcoeFhZAqIXTxoSg==",
|
||||
"license": "MIT",
|
||||
"funding": {
|
||||
"url": "https://github.com/sponsors/colinhacks"
|
||||
|
||||
@ -70,7 +70,7 @@
|
||||
"marked": "14.1.4",
|
||||
"marked-terminal": "7.2.1",
|
||||
"mime-types": "2.1.35",
|
||||
"openai": "4.91.0",
|
||||
"openai": "4.96.0",
|
||||
"p-map": "7.0.3",
|
||||
"p-throttle": "7.0.0",
|
||||
"remark-parse": "11.0.0",
|
||||
@ -81,7 +81,7 @@
|
||||
"unified": "11.0.5",
|
||||
"unist-util-visit": "5.0.0",
|
||||
"yargs": "17.7.2",
|
||||
"zod": "3.24.2"
|
||||
"zod": "3.24.3"
|
||||
},
|
||||
"keywords": [
|
||||
"ai",
|
||||
|
||||
File diff suppressed because one or more lines are too long
File diff suppressed because one or more lines are too long
@ -2,13 +2,14 @@ import OpenAI from 'openai'
|
||||
import { marked } from 'marked'
|
||||
import { markedTerminal } from 'marked-terminal'
|
||||
import * as path from 'node:path'
|
||||
import { sync as write } from '@polymech/fs/write'
|
||||
import { resolve } from '@polymech/commons'
|
||||
import { IKBotTask } from '@polymech/ai-tools'
|
||||
|
||||
import { dumpAsScript } from '../utils/script.js'
|
||||
import { applyFilters, Filter } from '../filters.js'
|
||||
import { variables } from '../variables.js'
|
||||
import { writeOrAppend } from '../utils/merge.js'
|
||||
import { E_AppendModeType } from '../zod_schema.js'
|
||||
|
||||
export const onCompletion = async (result: any = "", options: IKBotTask) => {
|
||||
result = applyFilters(result, options.filters as Filter[] || [])
|
||||
@ -16,12 +17,13 @@ export const onCompletion = async (result: any = "", options: IKBotTask) => {
|
||||
if (options.dst) {
|
||||
vars = {
|
||||
...vars,
|
||||
MODEL: path.parse(options.model).name,
|
||||
ROUTER: options.router,
|
||||
MODEL: options.model ? path.parse(options.model).name : 'unknown_model',
|
||||
ROUTER: options.router || 'unknown_router',
|
||||
}
|
||||
const dstPath = path.resolve(resolve(options.dst, false, vars))
|
||||
write(dstPath, result)
|
||||
options.logger.debug(`Wrote completion result to ${dstPath} : ${options.dst}`)
|
||||
writeOrAppend(dstPath, result, options.append as E_AppendModeType)
|
||||
const action = options.append ? `${options.append} content to` : 'Wrote completion result to'
|
||||
options.logger?.debug(`${action} ${dstPath} : ${options.dst}`)
|
||||
} else {
|
||||
marked.use(markedTerminal({
|
||||
emoji: false,
|
||||
@ -39,7 +41,6 @@ export const runCompletion = async (client: OpenAI, params: any, options: IKBotT
|
||||
options.logger.info('Dry run - skipping API call')
|
||||
return false
|
||||
}
|
||||
// await client.beta.chat.completions.parse
|
||||
const completion = await client.chat.completions.create({
|
||||
model: options.model,
|
||||
messages: params.messages,
|
||||
|
||||
20
packages/kbot/src/models/cache/openai-models.ts
vendored
20
packages/kbot/src/models/cache/openai-models.ts
vendored
@ -1,31 +1,28 @@
|
||||
export enum E_OPENAI_MODEL {
|
||||
MODEL_GPT_4O_AUDIO_PREVIEW_2024_12_17 = "gpt-4o-audio-preview-2024-12-17",
|
||||
MODEL_DALL_E_3 = "dall-e-3",
|
||||
MODEL_TEXT_EMBEDDING_3_LARGE = "text-embedding-3-large",
|
||||
MODEL_DALL_E_2 = "dall-e-2",
|
||||
MODEL_O4_MINI_2025_04_16 = "o4-mini-2025-04-16",
|
||||
MODEL_GPT_4O_AUDIO_PREVIEW_2024_10_01 = "gpt-4o-audio-preview-2024-10-01",
|
||||
MODEL_TEXT_EMBEDDING_3_SMALL = "text-embedding-3-small",
|
||||
MODEL_O4_MINI = "o4-mini",
|
||||
MODEL_GPT_4_1_NANO = "gpt-4.1-nano",
|
||||
MODEL_GPT_4_1_NANO_2025_04_14 = "gpt-4.1-nano-2025-04-14",
|
||||
MODEL_GPT_4O_REALTIME_PREVIEW_2024_10_01 = "gpt-4o-realtime-preview-2024-10-01",
|
||||
MODEL_O4_MINI_2025_04_16 = "o4-mini-2025-04-16",
|
||||
MODEL_GPT_4O_REALTIME_PREVIEW = "gpt-4o-realtime-preview",
|
||||
MODEL_BABBAGE_002 = "babbage-002",
|
||||
MODEL_GPT_4 = "gpt-4",
|
||||
MODEL_TEXT_EMBEDDING_ADA_002 = "text-embedding-ada-002",
|
||||
MODEL_TEXT_EMBEDDING_3_LARGE = "text-embedding-3-large",
|
||||
MODEL_GPT_4O_MINI_AUDIO_PREVIEW = "gpt-4o-mini-audio-preview",
|
||||
MODEL_GPT_4O_AUDIO_PREVIEW = "gpt-4o-audio-preview",
|
||||
MODEL_O1_PREVIEW_2024_09_12 = "o1-preview-2024-09-12",
|
||||
MODEL_O1_PRO = "o1-pro",
|
||||
MODEL_O1_2024_12_17 = "o1-2024-12-17",
|
||||
MODEL_GPT_4O_MINI_REALTIME_PREVIEW = "gpt-4o-mini-realtime-preview",
|
||||
MODEL_GPT_4_1_MINI = "gpt-4.1-mini",
|
||||
MODEL_GPT_4O_MINI_REALTIME_PREVIEW_2024_12_17 = "gpt-4o-mini-realtime-preview-2024-12-17",
|
||||
MODEL_GPT_3_5_TURBO_INSTRUCT_0914 = "gpt-3.5-turbo-instruct-0914",
|
||||
MODEL_GPT_4O_MINI_SEARCH_PREVIEW = "gpt-4o-mini-search-preview",
|
||||
MODEL_O1 = "o1",
|
||||
MODEL_GPT_4_1_MINI_2025_04_14 = "gpt-4.1-mini-2025-04-14",
|
||||
MODEL_O1_PRO_2025_03_19 = "o1-pro-2025-03-19",
|
||||
MODEL_CHATGPT_4O_LATEST = "chatgpt-4o-latest",
|
||||
MODEL_DAVINCI_002 = "davinci-002",
|
||||
MODEL_GPT_3_5_TURBO_1106 = "gpt-3.5-turbo-1106",
|
||||
@ -45,13 +42,11 @@ export enum E_OPENAI_MODEL {
|
||||
MODEL_O1_PREVIEW = "o1-preview",
|
||||
MODEL_GPT_4_0613 = "gpt-4-0613",
|
||||
MODEL_GPT_4_5_PREVIEW = "gpt-4.5-preview",
|
||||
MODEL_O3_MINI = "o3-mini",
|
||||
MODEL_O3_MINI_2025_01_31 = "o3-mini-2025-01-31",
|
||||
MODEL_GPT_4_5_PREVIEW_2025_02_27 = "gpt-4.5-preview-2025-02-27",
|
||||
MODEL_GPT_4O_SEARCH_PREVIEW_2025_03_11 = "gpt-4o-search-preview-2025-03-11",
|
||||
MODEL_OMNI_MODERATION_2024_09_26 = "omni-moderation-2024-09-26",
|
||||
MODEL_TEXT_EMBEDDING_3_SMALL = "text-embedding-3-small",
|
||||
MODEL_GPT_4O_MINI_TTS = "gpt-4o-mini-tts",
|
||||
MODEL_O3_MINI_2025_01_31 = "o3-mini-2025-01-31",
|
||||
MODEL_O3_MINI = "o3-mini",
|
||||
MODEL_TTS_1_HD = "tts-1-hd",
|
||||
MODEL_GPT_4O = "gpt-4o",
|
||||
MODEL_TTS_1_HD_1106 = "tts-1-hd-1106",
|
||||
@ -60,6 +55,7 @@ export enum E_OPENAI_MODEL {
|
||||
MODEL_GPT_4_1 = "gpt-4.1",
|
||||
MODEL_GPT_4O_TRANSCRIBE = "gpt-4o-transcribe",
|
||||
MODEL_GPT_4_1_2025_04_14 = "gpt-4.1-2025-04-14",
|
||||
MODEL_O1_2024_12_17 = "o1-2024-12-17",
|
||||
MODEL_GPT_4O_MINI_2024_07_18 = "gpt-4o-mini-2024-07-18",
|
||||
MODEL_GPT_4O_MINI_TRANSCRIBE = "gpt-4o-mini-transcribe",
|
||||
MODEL_O1_MINI = "o1-mini",
|
||||
@ -68,6 +64,10 @@ export enum E_OPENAI_MODEL {
|
||||
MODEL_O1_MINI_2024_09_12 = "o1-mini-2024-09-12",
|
||||
MODEL_TTS_1 = "tts-1",
|
||||
MODEL_GPT_4_1106_PREVIEW = "gpt-4-1106-preview",
|
||||
MODEL_GPT_4O_MINI_TTS = "gpt-4o-mini-tts",
|
||||
MODEL_TTS_1_1106 = "tts-1-1106",
|
||||
MODEL_O1 = "o1",
|
||||
MODEL_O1_PRO = "o1-pro",
|
||||
MODEL_O1_PRO_2025_03_19 = "o1-pro-2025-03-19",
|
||||
MODEL_OMNI_MODERATION_LATEST = "omni-moderation-latest"
|
||||
}
|
||||
@ -1,4 +1,5 @@
|
||||
export enum E_OPENROUTER_MODEL_FREE {
|
||||
MODEL_FREE_MICROSOFT_MAI_DS_R1_FREE = "microsoft/mai-ds-r1:free",
|
||||
MODEL_FREE_THUDM_GLM_Z1_32B_FREE = "thudm/glm-z1-32b:free",
|
||||
MODEL_FREE_THUDM_GLM_4_32B_FREE = "thudm/glm-4-32b:free",
|
||||
MODEL_FREE_SHISA_AI_SHISA_V2_LLAMA3_3_70B_FREE = "shisa-ai/shisa-v2-llama3.3-70b:free",
|
||||
@ -19,7 +20,6 @@ export enum E_OPENROUTER_MODEL_FREE {
|
||||
MODEL_FREE_DEEPSEEK_DEEPSEEK_CHAT_V3_0324_FREE = "deepseek/deepseek-chat-v3-0324:free",
|
||||
MODEL_FREE_FEATHERLESS_QWERKY_72B_FREE = "featherless/qwerky-72b:free",
|
||||
MODEL_FREE_MISTRALAI_MISTRAL_SMALL_3_1_24B_INSTRUCT_FREE = "mistralai/mistral-small-3.1-24b-instruct:free",
|
||||
MODEL_FREE_OPEN_R1_OLYMPICCODER_7B_FREE = "open-r1/olympiccoder-7b:free",
|
||||
MODEL_FREE_OPEN_R1_OLYMPICCODER_32B_FREE = "open-r1/olympiccoder-32b:free",
|
||||
MODEL_FREE_GOOGLE_GEMMA_3_1B_IT_FREE = "google/gemma-3-1b-it:free",
|
||||
MODEL_FREE_GOOGLE_GEMMA_3_4B_IT_FREE = "google/gemma-3-4b-it:free",
|
||||
@ -37,11 +37,9 @@ export enum E_OPENROUTER_MODEL_FREE {
|
||||
MODEL_FREE_DEEPSEEK_DEEPSEEK_R1_DISTILL_QWEN_32B_FREE = "deepseek/deepseek-r1-distill-qwen-32b:free",
|
||||
MODEL_FREE_DEEPSEEK_DEEPSEEK_R1_DISTILL_QWEN_14B_FREE = "deepseek/deepseek-r1-distill-qwen-14b:free",
|
||||
MODEL_FREE_DEEPSEEK_DEEPSEEK_R1_DISTILL_LLAMA_70B_FREE = "deepseek/deepseek-r1-distill-llama-70b:free",
|
||||
MODEL_FREE_GOOGLE_GEMINI_2_0_FLASH_THINKING_EXP_FREE = "google/gemini-2.0-flash-thinking-exp:free",
|
||||
MODEL_FREE_DEEPSEEK_DEEPSEEK_R1_FREE = "deepseek/deepseek-r1:free",
|
||||
MODEL_FREE_SOPHOSYMPATHEIA_ROGUE_ROSE_103B_V0_2_FREE = "sophosympatheia/rogue-rose-103b-v0.2:free",
|
||||
MODEL_FREE_DEEPSEEK_DEEPSEEK_CHAT_FREE = "deepseek/deepseek-chat:free",
|
||||
MODEL_FREE_GOOGLE_GEMINI_2_0_FLASH_THINKING_EXP_1219_FREE = "google/gemini-2.0-flash-thinking-exp-1219:free",
|
||||
MODEL_FREE_GOOGLE_GEMINI_2_0_FLASH_EXP_FREE = "google/gemini-2.0-flash-exp:free",
|
||||
MODEL_FREE_META_LLAMA_LLAMA_3_3_70B_INSTRUCT_FREE = "meta-llama/llama-3.3-70b-instruct:free",
|
||||
MODEL_FREE_QWEN_QWQ_32B_PREVIEW_FREE = "qwen/qwq-32b-preview:free",
|
||||
|
||||
@ -1,4 +1,5 @@
|
||||
export enum E_OPENROUTER_MODEL {
|
||||
MODEL_MICROSOFT_MAI_DS_R1_FREE = "microsoft/mai-ds-r1:free",
|
||||
MODEL_GOOGLE_GEMINI_2_5_PRO_PREVIEW_03_25 = "google/gemini-2.5-pro-preview-03-25",
|
||||
MODEL_THUDM_GLM_Z1_32B_FREE = "thudm/glm-z1-32b:free",
|
||||
MODEL_THUDM_GLM_4_32B_FREE = "thudm/glm-4-32b:free",
|
||||
@ -43,7 +44,6 @@ export enum E_OPENROUTER_MODEL {
|
||||
MODEL_OPENAI_O1_PRO = "openai/o1-pro",
|
||||
MODEL_MISTRALAI_MISTRAL_SMALL_3_1_24B_INSTRUCT_FREE = "mistralai/mistral-small-3.1-24b-instruct:free",
|
||||
MODEL_MISTRALAI_MISTRAL_SMALL_3_1_24B_INSTRUCT = "mistralai/mistral-small-3.1-24b-instruct",
|
||||
MODEL_OPEN_R1_OLYMPICCODER_7B_FREE = "open-r1/olympiccoder-7b:free",
|
||||
MODEL_OPEN_R1_OLYMPICCODER_32B_FREE = "open-r1/olympiccoder-32b:free",
|
||||
MODEL_STEELSKULL_L3_3_ELECTRA_R1_70B = "steelskull/l3.3-electra-r1-70b",
|
||||
MODEL_GOOGLE_GEMMA_3_1B_IT_FREE = "google/gemma-3-1b-it:free",
|
||||
@ -108,7 +108,6 @@ export enum E_OPENROUTER_MODEL {
|
||||
MODEL_LIQUID_LFM_3B = "liquid/lfm-3b",
|
||||
MODEL_DEEPSEEK_DEEPSEEK_R1_DISTILL_LLAMA_70B_FREE = "deepseek/deepseek-r1-distill-llama-70b:free",
|
||||
MODEL_DEEPSEEK_DEEPSEEK_R1_DISTILL_LLAMA_70B = "deepseek/deepseek-r1-distill-llama-70b",
|
||||
MODEL_GOOGLE_GEMINI_2_0_FLASH_THINKING_EXP_FREE = "google/gemini-2.0-flash-thinking-exp:free",
|
||||
MODEL_DEEPSEEK_DEEPSEEK_R1_FREE = "deepseek/deepseek-r1:free",
|
||||
MODEL_DEEPSEEK_DEEPSEEK_R1 = "deepseek/deepseek-r1",
|
||||
MODEL_SOPHOSYMPATHEIA_ROGUE_ROSE_103B_V0_2_FREE = "sophosympatheia/rogue-rose-103b-v0.2:free",
|
||||
@ -117,7 +116,6 @@ export enum E_OPENROUTER_MODEL {
|
||||
MODEL_MICROSOFT_PHI_4 = "microsoft/phi-4",
|
||||
MODEL_DEEPSEEK_DEEPSEEK_CHAT_FREE = "deepseek/deepseek-chat:free",
|
||||
MODEL_DEEPSEEK_DEEPSEEK_CHAT = "deepseek/deepseek-chat",
|
||||
MODEL_GOOGLE_GEMINI_2_0_FLASH_THINKING_EXP_1219_FREE = "google/gemini-2.0-flash-thinking-exp-1219:free",
|
||||
MODEL_SAO10K_L3_3_EURYALE_70B = "sao10k/l3.3-euryale-70b",
|
||||
MODEL_OPENAI_O1 = "openai/o1",
|
||||
MODEL_EVA_UNIT_01_EVA_LLAMA_3_33_70B = "eva-unit-01/eva-llama-3.33-70b",
|
||||
@ -232,8 +230,8 @@ export enum E_OPENROUTER_MODEL {
|
||||
MODEL_GOOGLE_GEMINI_FLASH_1_5 = "google/gemini-flash-1.5",
|
||||
MODEL_OPENAI_GPT_4O = "openai/gpt-4o",
|
||||
MODEL_OPENAI_GPT_4O_EXTENDED = "openai/gpt-4o:extended",
|
||||
MODEL_OPENAI_GPT_4O_2024_05_13 = "openai/gpt-4o-2024-05-13",
|
||||
MODEL_META_LLAMA_LLAMA_GUARD_2_8B = "meta-llama/llama-guard-2-8b",
|
||||
MODEL_OPENAI_GPT_4O_2024_05_13 = "openai/gpt-4o-2024-05-13",
|
||||
MODEL_NEVERSLEEP_LLAMA_3_LUMIMAID_8B_EXTENDED = "neversleep/llama-3-lumimaid-8b:extended",
|
||||
MODEL_NEVERSLEEP_LLAMA_3_LUMIMAID_8B = "neversleep/llama-3-lumimaid-8b",
|
||||
MODEL_SAO10K_FIMBULVETR_11B_V2 = "sao10k/fimbulvetr-11b-v2",
|
||||
@ -266,7 +264,6 @@ export enum E_OPENROUTER_MODEL {
|
||||
MODEL_MISTRALAI_MISTRAL_7B_INSTRUCT_V0_2 = "mistralai/mistral-7b-instruct-v0.2",
|
||||
MODEL_COGNITIVECOMPUTATIONS_DOLPHIN_MIXTRAL_8X7B = "cognitivecomputations/dolphin-mixtral-8x7b",
|
||||
MODEL_GOOGLE_GEMINI_PRO_VISION = "google/gemini-pro-vision",
|
||||
MODEL_GOOGLE_GEMINI_PRO = "google/gemini-pro",
|
||||
MODEL_MISTRALAI_MIXTRAL_8X7B_INSTRUCT = "mistralai/mixtral-8x7b-instruct",
|
||||
MODEL_OPENCHAT_OPENCHAT_7B = "openchat/openchat-7b",
|
||||
MODEL_NEVERSLEEP_NOROMAID_20B = "neversleep/noromaid-20b",
|
||||
|
||||
@ -37,23 +37,10 @@ export const EType = z.enum([
|
||||
E_Mode.ASSISTANT,
|
||||
E_Mode.CUSTOM
|
||||
])
|
||||
// https://rjsf-team.github.io/react-jsonschema-form/docs/usage/widgets#file-widgets
|
||||
/*
|
||||
langchain/CharacterTextSplitter,
|
||||
langchain/LatexTextSplitter,
|
||||
langchain/MarkdownTextSplitter,
|
||||
langchain/RecursiveCharacterTextSplitter,
|
||||
langchain/TokenTextSplitter
|
||||
*/
|
||||
export const SplitType = z.enum([
|
||||
'none',
|
||||
'langchain/CharacterTextSplitter',
|
||||
'langchain/LatexTextSplitter',
|
||||
'langchain/MarkdownTextSplitter',
|
||||
'langchain/RecursiveCharacterTextSplitter',
|
||||
'langchain/TokenTextSplitter'
|
||||
]).describe('Type of text splitter to use')
|
||||
|
||||
// Define the new enum for append modes
|
||||
export const E_AppendMode = z.enum(['concat', 'merge'])
|
||||
export type E_AppendModeType = z.infer<typeof E_AppendMode>
|
||||
|
||||
export type OptionsSchemaMeta = Record<string, unknown>
|
||||
|
||||
@ -90,6 +77,12 @@ export const OptionsSchema = (opts?: any): any => {
|
||||
.optional()
|
||||
.describe('Optional destination path for the result, will substitute ${MODEL_NAME} and ${ROUTER} in the path. Optional, used for "completion" mode')
|
||||
)
|
||||
.add(
|
||||
'append',
|
||||
E_AppendMode
|
||||
.optional()
|
||||
.describe('How to handle output if --dst file already exists: "concat" (append) or "merge" (try to merge structures if possible, otherwise append). Only used if --dst is specified.')
|
||||
)
|
||||
.add(
|
||||
'each',
|
||||
z.string()
|
||||
@ -203,8 +196,6 @@ export const OptionsSchema = (opts?: any): any => {
|
||||
.default('default')
|
||||
.describe('Environment (in profile)')
|
||||
)
|
||||
// If "split" was relevant, re-add here:
|
||||
// .add('split', SplitType.default('none').describe('Type of text splitter to use'))
|
||||
.add(
|
||||
'variables',
|
||||
z.record(z.string(), z.string())
|
||||
|
||||
@ -7,6 +7,8 @@ export interface IKBotOptions {
|
||||
output?: string | undefined;
|
||||
/** Optional destination path for the result, will substitute ${MODEL_NAME} and ${ROUTER} in the path. Optional, used for "completion" mode */
|
||||
dst?: string | undefined;
|
||||
/** How to handle output if --dst file already exists: "concat" (append) or "merge" (try to merge structures if possible, otherwise append). Only used if --dst is specified. */
|
||||
append?: ("concat" | "merge") | undefined;
|
||||
/** Iterate over items, supported: GLOB | Path to JSON File | array of strings (comma separated). To test different models, use --each="gpt-3.5-turbo,gpt-4o", the actual string will exposed as variable `ITEM`, eg: --dst="${ITEM}-output.md" */
|
||||
each?: string | undefined;
|
||||
/** Disable tools categories, eg: --disable=fs,git,interact,terminal,search,web,email,user */
|
||||
@ -149,6 +151,7 @@ export interface IKBotOptions {
|
||||
meta-llama/llama-3-70b-instruct | paid
|
||||
meta-llama/llama-3-8b-instruct | paid
|
||||
meta-llama/llama-3.1-405b | paid
|
||||
meta-llama/llama-3.1-405b:free | free
|
||||
meta-llama/llama-3.1-405b-instruct | paid
|
||||
meta-llama/llama-3.1-70b-instruct | paid
|
||||
meta-llama/llama-3.1-8b-instruct | paid
|
||||
@ -305,7 +308,6 @@ export interface IKBotOptions {
|
||||
sophosympatheia/rogue-rose-103b-v0.2:free | free
|
||||
sao10k/l3-lunaris-8b | paid
|
||||
sao10k/l3-euryale-70b | paid
|
||||
sao10k/l3.1-70b-hanami-x1 | paid
|
||||
sao10k/l3.1-euryale-70b | paid
|
||||
sao10k/l3.3-euryale-70b | paid
|
||||
shisa-ai/shisa-v2-llama3.3-70b:free | free
|
||||
@ -327,7 +329,6 @@ export interface IKBotOptions {
|
||||
x-ai/grok-3-mini-beta | paid
|
||||
x-ai/grok-beta | paid
|
||||
x-ai/grok-vision-beta | paid
|
||||
xwin-lm/xwin-lm-70b | paid
|
||||
[35m[1m[22m[39m
|
||||
[35m[1m OpenAI models:[22m[39m
|
||||
[35m[1m[22m[39m
|
||||
|
||||
Loading…
Reference in New Issue
Block a user