maintainence love:)
This commit is contained in:
parent
9d1eeb31f8
commit
a335e388a8
@ -1,10 +1,6 @@
|
||||
import { marked } from 'marked';
|
||||
import { markedTerminal } from 'marked-terminal';
|
||||
import { examples as content } from '../docs-internal/examples.js';
|
||||
export const examples = () => {
|
||||
marked.use(markedTerminal({
|
||||
emoji: false,
|
||||
}));
|
||||
process.stdout.write(marked(content));
|
||||
};
|
||||
//# sourceMappingURL=data:application/json;base64,eyJ2ZXJzaW9uIjozLCJmaWxlIjoiZXhhbXBsZXMuanMiLCJzb3VyY2VSb290IjoiIiwic291cmNlcyI6WyIuLi8uLi9zcmMvY29tbWFuZHMvZXhhbXBsZXMudHMiXSwibmFtZXMiOltdLCJtYXBwaW5ncyI6IkFBQUEsT0FBTyxFQUFFLE1BQU0sRUFBRSxNQUFNLFFBQVEsQ0FBQTtBQUMvQixPQUFPLEVBQUUsY0FBYyxFQUFFLE1BQU0saUJBQWlCLENBQUE7QUFDaEQsT0FBTyxFQUFFLFFBQVEsSUFBSSxPQUFPLEVBQUUsTUFBTSw4QkFBOEIsQ0FBQTtBQUVsRSxNQUFNLENBQUMsTUFBTSxRQUFRLEdBQUcsR0FBRyxFQUFFO0lBQzNCLE1BQU0sQ0FBQyxHQUFHLENBQUMsY0FBYyxDQUN2QjtRQUNFLEtBQUssRUFBRSxLQUFLO0tBQ2IsQ0FDRixDQUFDLENBQUE7SUFDRixPQUFPLENBQUMsTUFBTSxDQUFDLEtBQUssQ0FBQyxNQUFNLENBQUMsT0FBTyxDQUFXLENBQUMsQ0FBQTtBQUNqRCxDQUFDLENBQUEifQ==
|
||||
//# sourceMappingURL=data:application/json;base64,eyJ2ZXJzaW9uIjozLCJmaWxlIjoiZXhhbXBsZXMuanMiLCJzb3VyY2VSb290IjoiIiwic291cmNlcyI6WyIuLi8uLi9zcmMvY29tbWFuZHMvZXhhbXBsZXMudHMiXSwibmFtZXMiOltdLCJtYXBwaW5ncyI6IkFBQUEsT0FBTyxFQUFFLE1BQU0sRUFBRSxNQUFNLFFBQVEsQ0FBQTtBQUMvQixPQUFPLEVBQUUsUUFBUSxJQUFJLE9BQU8sRUFBRSxNQUFNLDhCQUE4QixDQUFBO0FBRWxFLE1BQU0sQ0FBQyxNQUFNLFFBQVEsR0FBRyxHQUFHLEVBQUU7SUFDM0IsT0FBTyxDQUFDLE1BQU0sQ0FBQyxLQUFLLENBQUMsTUFBTSxDQUFDLE9BQU8sQ0FBVyxDQUFDLENBQUE7QUFDakQsQ0FBQyxDQUFBIn0=
|
||||
@ -1,6 +1,5 @@
|
||||
import { z } from 'zod';
|
||||
import { marked } from 'marked';
|
||||
import { markedTerminal } from 'marked-terminal';
|
||||
import { getDefaultValue, getDescription } from '@polymech/commons';
|
||||
import { OptionsSchema } from '../zod_schema.js';
|
||||
export const help = () => {
|
||||
@ -35,9 +34,6 @@ export const help = () => {
|
||||
md.push(`Required: ${(!isOptional).toString()}`);
|
||||
md.push('');
|
||||
}
|
||||
marked.use(markedTerminal({
|
||||
emoji: false
|
||||
}));
|
||||
const content = marked(md.join('\n'));
|
||||
process.stdout.write(content);
|
||||
return content;
|
||||
@ -45,4 +41,4 @@ export const help = () => {
|
||||
export default async (argv) => {
|
||||
return help();
|
||||
};
|
||||
//# sourceMappingURL=data:application/json;base64,eyJ2ZXJzaW9uIjozLCJmaWxlIjoiaGVscC5qcyIsInNvdXJjZVJvb3QiOiIiLCJzb3VyY2VzIjpbIi4uLy4uL3NyYy9jb21tYW5kcy9oZWxwLnRzIl0sIm5hbWVzIjpbXSwibWFwcGluZ3MiOiJBQUFBLE9BQU8sRUFBRSxDQUFDLEVBQUUsTUFBTSxLQUFLLENBQUE7QUFDdkIsT0FBTyxFQUFFLE1BQU0sRUFBRSxNQUFNLFFBQVEsQ0FBQTtBQUMvQixPQUFPLEVBQUUsY0FBYyxFQUFFLE1BQU0saUJBQWlCLENBQUE7QUFDaEQsT0FBTyxFQUFFLGVBQWUsRUFBRSxjQUFjLEVBQUUsTUFBTSxtQkFBbUIsQ0FBQTtBQUVuRSxPQUFPLEVBQUUsYUFBYSxFQUFFLE1BQU0sa0JBQWtCLENBQUE7QUFFaEQsTUFBTSxDQUFDLE1BQU0sSUFBSSxHQUFHLEdBQUcsRUFBRTtJQUN2QixNQUFNLE1BQU0sR0FBRyxhQUFhLEVBQUUsQ0FBQTtJQUM5QixNQUFNLElBQUksR0FBRyxNQUFNLENBQUMsSUFBSSxDQUFDLFdBQVcsQ0FBQTtJQUNwQyxNQUFNLEtBQUssR0FBRyxNQUFNLENBQUMsS0FBSyxDQUFBO0lBRTFCLE1BQU0sRUFBRSxHQUFHO1FBQ1QsK0JBQStCO1FBQy9CLEVBQUU7UUFDRixTQUFTO1FBQ1QsaUNBQWlDO1FBQ2pDLEtBQUs7UUFDTCxFQUFFO1FBQ0YsZ0JBQWdCO1FBQ2hCLEVBQUU7UUFDRixJQUFJLElBQUksZUFBZTtRQUN2QixFQUFFO1FBQ0YsWUFBWTtRQUNaLEVBQUU7S0FDSCxDQUFBO0lBRUQseUJBQXlCO0lBQ3pCLEtBQUssTUFBTSxDQUFDLEdBQUcsRUFBRSxHQUFHLENBQUMsSUFBSSxNQUFNLENBQUMsT0FBTyxDQUFDLEtBQUssQ0FBQyxFQUFFLENBQUM7UUFDL0MsTUFBTSxVQUFVLEdBQUcsR0FBRyxZQUFZLENBQUMsQ0FBQyxXQUFXLENBQUE7UUFDL0MsTUFBTSxZQUFZLEdBQUcsZUFBZSxDQUFDLEdBQVUsQ0FBQyxDQUFBO1FBQ2hELE1BQU0sV0FBVyxHQUFHLGNBQWMsQ0FBQyxHQUFVLENBQUMsQ0FBQTtRQUM5QyxFQUFFLENBQUMsSUFBSSxDQUFDLEtBQUssR0FBRyxFQUFFLENBQUMsQ0FBQTtRQUNuQixFQUFFLENBQUMsSUFBSSxDQUFDLEVBQUUsQ0FBQyxDQUFBO1FBQ1gsRUFBRSxDQUFDLElBQUksQ0FBQyxnQkFBZ0IsV0FBVyxFQUFFLENBQUMsQ0FBQTtRQUN0QyxJQUFJLFlBQVksS0FBSyxTQUFTLEVBQUUsQ0FBQztZQUMvQixFQUFFLENBQUMsSUFBSSxDQUFDLGNBQWMsSUFBSSxDQUFDLFNBQVMsQ0FBQyxZQUFZLENBQUMsSUFBSSxDQUFDLENBQUE7UUFDekQsQ0FBQztRQUNELEVBQUUsQ0FBQyxJQUFJLENBQUMsYUFBYSxDQUFDLENBQUMsVUFBVSxDQUFDLENBQUMsUUFBUSxFQUFFLEVBQUUsQ0FBQyxDQUFBO1FBQ2hELEVBQUUsQ0FBQyxJQUFJLENBQUMsRUFBRSxDQUFDLENBQUE7SUFDYixDQUFDO0lBRUQsTUFBTSxDQUFDLEdBQUcsQ0FBQyxjQUFjLENBQUM7UUFDeEIsS0FBSyxFQUFFLEtBQUs7S0FDYixDQUFDLENBQUMsQ0FBQTtJQUNILE1BQU0sT0FBTyxHQUFXLE1BQU0sQ0FBQyxFQUFFLENBQUMsSUFBSSxDQUFDLElBQUksQ0FBQyxDQUFXLENBQUM7SUFDeEQsT0FBTyxDQUFDLE1BQU0sQ0FBQyxLQUFLLENBQUMsT0FBTyxDQUFDLENBQUE7SUFDN0IsT0FBTyxPQUFPLENBQUE7QUFDaEIsQ0FBQyxDQUFBO0FBRUQsZUFBZSxLQUFLLEVBQUUsSUFBUyxFQUFFLEVBQUU7SUFDakMsT0FBTyxJQUFJLEVBQUUsQ0FBQTtBQUNmLENBQUMsQ0FBQSJ9
|
||||
//# sourceMappingURL=data:application/json;base64,eyJ2ZXJzaW9uIjozLCJmaWxlIjoiaGVscC5qcyIsInNvdXJjZVJvb3QiOiIiLCJzb3VyY2VzIjpbIi4uLy4uL3NyYy9jb21tYW5kcy9oZWxwLnRzIl0sIm5hbWVzIjpbXSwibWFwcGluZ3MiOiJBQUFBLE9BQU8sRUFBRSxDQUFDLEVBQUUsTUFBTSxLQUFLLENBQUE7QUFDdkIsT0FBTyxFQUFFLE1BQU0sRUFBRSxNQUFNLFFBQVEsQ0FBQTtBQUMvQixPQUFPLEVBQUUsZUFBZSxFQUFFLGNBQWMsRUFBRSxNQUFNLG1CQUFtQixDQUFBO0FBRW5FLE9BQU8sRUFBRSxhQUFhLEVBQUUsTUFBTSxrQkFBa0IsQ0FBQTtBQUVoRCxNQUFNLENBQUMsTUFBTSxJQUFJLEdBQUcsR0FBRyxFQUFFO0lBQ3ZCLE1BQU0sTUFBTSxHQUFHLGFBQWEsRUFBRSxDQUFBO0lBQzlCLE1BQU0sSUFBSSxHQUFHLE1BQU0sQ0FBQyxJQUFJLENBQUMsV0FBVyxDQUFBO0lBQ3BDLE1BQU0sS0FBSyxHQUFHLE1BQU0sQ0FBQyxLQUFLLENBQUE7SUFFMUIsTUFBTSxFQUFFLEdBQUc7UUFDVCwrQkFBK0I7UUFDL0IsRUFBRTtRQUNGLFNBQVM7UUFDVCxpQ0FBaUM7UUFDakMsS0FBSztRQUNMLEVBQUU7UUFDRixnQkFBZ0I7UUFDaEIsRUFBRTtRQUNGLElBQUksSUFBSSxlQUFlO1FBQ3ZCLEVBQUU7UUFDRixZQUFZO1FBQ1osRUFBRTtLQUNILENBQUE7SUFFRCx5QkFBeUI7SUFDekIsS0FBSyxNQUFNLENBQUMsR0FBRyxFQUFFLEdBQUcsQ0FBQyxJQUFJLE1BQU0sQ0FBQyxPQUFPLENBQUMsS0FBSyxDQUFDLEVBQUUsQ0FBQztRQUMvQyxNQUFNLFVBQVUsR0FBRyxHQUFHLFlBQVksQ0FBQyxDQUFDLFdBQVcsQ0FBQTtRQUMvQyxNQUFNLFlBQVksR0FBRyxlQUFlLENBQUMsR0FBVSxDQUFDLENBQUE7UUFDaEQsTUFBTSxXQUFXLEdBQUcsY0FBYyxDQUFDLEdBQVUsQ0FBQyxDQUFBO1FBQzlDLEVBQUUsQ0FBQyxJQUFJLENBQUMsS0FBSyxHQUFHLEVBQUUsQ0FBQyxDQUFBO1FBQ25CLEVBQUUsQ0FBQyxJQUFJLENBQUMsRUFBRSxDQUFDLENBQUE7UUFDWCxFQUFFLENBQUMsSUFBSSxDQUFDLGdCQUFnQixXQUFXLEVBQUUsQ0FBQyxDQUFBO1FBQ3RDLElBQUksWUFBWSxLQUFLLFNBQVMsRUFBRSxDQUFDO1lBQy9CLEVBQUUsQ0FBQyxJQUFJLENBQUMsY0FBYyxJQUFJLENBQUMsU0FBUyxDQUFDLFlBQVksQ0FBQyxJQUFJLENBQUMsQ0FBQTtRQUN6RCxDQUFDO1FBQ0QsRUFBRSxDQUFDLElBQUksQ0FBQyxhQUFhLENBQUMsQ0FBQyxVQUFVLENBQUMsQ0FBQyxRQUFRLEVBQUUsRUFBRSxDQUFDLENBQUE7UUFDaEQsRUFBRSxDQUFDLElBQUksQ0FBQyxFQUFFLENBQUMsQ0FBQTtJQUNiLENBQUM7SUFFRCxNQUFNLE9BQU8sR0FBVyxNQUFNLENBQUMsRUFBRSxDQUFDLElBQUksQ0FBQyxJQUFJLENBQUMsQ0FBVyxDQUFDO0lBQ3hELE9BQU8sQ0FBQyxNQUFNLENBQUMsS0FBSyxDQUFDLE9BQU8sQ0FBQyxDQUFBO0lBQzdCLE9BQU8sT0FBTyxDQUFBO0FBQ2hCLENBQUMsQ0FBQTtBQUVELGVBQWUsS0FBSyxFQUFFLElBQVMsRUFBRSxFQUFFO0lBQ2pDLE9BQU8sSUFBSSxFQUFFLENBQUE7QUFDZixDQUFDLENBQUEifQ==
|
||||
@ -1,5 +1,4 @@
|
||||
import { marked } from 'marked';
|
||||
import { markedTerminal } from 'marked-terminal';
|
||||
import * as path from 'node:path';
|
||||
import { resolve } from '@polymech/commons';
|
||||
import { dumpAsScript } from '../utils/script.js';
|
||||
@ -21,9 +20,6 @@ export const onCompletion = async (result = "", options) => {
|
||||
options.logger?.debug(`${action} ${dstPath} : ${options.dst}`);
|
||||
}
|
||||
else {
|
||||
marked.use(markedTerminal({
|
||||
emoji: false,
|
||||
}));
|
||||
const content = marked(result);
|
||||
process.stdout.write(content);
|
||||
}
|
||||
@ -48,4 +44,4 @@ export const runCompletion = async (client, params, options) => {
|
||||
result = await onCompletion(result, options);
|
||||
return result;
|
||||
};
|
||||
//# sourceMappingURL=data:application/json;base64,eyJ2ZXJzaW9uIjozLCJmaWxlIjoicnVuLWNvbXBsZXRpb24uanMiLCJzb3VyY2VSb290IjoiIiwic291cmNlcyI6WyIuLi8uLi9zcmMvY29tbWFuZHMvcnVuLWNvbXBsZXRpb24udHMiXSwibmFtZXMiOltdLCJtYXBwaW5ncyI6IkFBQ0EsT0FBTyxFQUFFLE1BQU0sRUFBRSxNQUFNLFFBQVEsQ0FBQTtBQUMvQixPQUFPLEVBQUUsY0FBYyxFQUFFLE1BQU0saUJBQWlCLENBQUE7QUFDaEQsT0FBTyxLQUFLLElBQUksTUFBTSxXQUFXLENBQUE7QUFDakMsT0FBTyxFQUFFLE9BQU8sRUFBRSxNQUFNLG1CQUFtQixDQUFBO0FBRzNDLE9BQU8sRUFBRSxZQUFZLEVBQUUsTUFBTSxvQkFBb0IsQ0FBQTtBQUNqRCxPQUFPLEVBQUUsWUFBWSxFQUFVLE1BQU0sZUFBZSxDQUFBO0FBQ3BELE9BQU8sRUFBRSxTQUFTLEVBQUUsTUFBTSxpQkFBaUIsQ0FBQTtBQUMzQyxPQUFPLEVBQUUsYUFBYSxFQUFFLE1BQU0sbUJBQW1CLENBQUE7QUFHakQsTUFBTSxDQUFDLE1BQU0sWUFBWSxHQUFHLEtBQUssRUFBRSxTQUFjLEVBQUUsRUFBRSxPQUFrQixFQUFFLEVBQUU7SUFDekUsTUFBTSxHQUFHLFlBQVksQ0FBQyxNQUFNLEVBQUUsT0FBTyxDQUFDLE9BQW1CLElBQUksRUFBRSxDQUFDLENBQUE7SUFDaEUsSUFBSSxJQUFJLEdBQUcsU0FBUyxDQUFDLE9BQU8sQ0FBQyxDQUFBO0lBQzdCLElBQUksT0FBTyxDQUFDLEdBQUcsRUFBRSxDQUFDO1FBQ2hCLElBQUksR0FBRztZQUNMLEdBQUcsSUFBSTtZQUNQLEtBQUssRUFBRSxPQUFPLENBQUMsS0FBSyxDQUFDLENBQUMsQ0FBQyxJQUFJLENBQUMsS0FBSyxDQUFDLE9BQU8sQ0FBQyxLQUFLLENBQUMsQ0FBQyxJQUFJLENBQUMsQ0FBQyxDQUFDLGVBQWU7WUFDdkUsTUFBTSxFQUFFLE9BQU8sQ0FBQyxNQUFNLElBQUksZ0JBQWdCO1NBQzNDLENBQUE7UUFDRCxNQUFNLE9BQU8sR0FBRyxJQUFJLENBQUMsT0FBTyxDQUFDLE9BQU8sQ0FBQyxPQUFPLENBQUMsR0FBRyxFQUFFLE9BQU8sQ0FBQyxHQUFHLEVBQUUsSUFBSSxDQUFDLENBQUMsQ0FBQTtRQUNyRSxhQUFhLENBQUMsT0FBTyxFQUFFLE1BQU0sRUFBRSxPQUFPLENBQUMsTUFBMEIsQ0FBQyxDQUFBO1FBQ2xFLE1BQU0sTUFBTSxHQUFHLE9BQU8sQ0FBQyxNQUFNLENBQUMsQ0FBQyxDQUFDLEdBQUcsT0FBTyxDQUFDLE1BQU0sYUFBYSxDQUFDLENBQUMsQ0FBQyw0QkFBNEIsQ0FBQTtRQUM3RixPQUFPLENBQUMsTUFBTSxFQUFFLEtBQUssQ0FBQyxHQUFHLE1BQU0sSUFBSSxPQUFPLE1BQU0sT0FBTyxDQUFDLEdBQUcsRUFBRSxDQUFDLENBQUE7SUFDaEUsQ0FBQztTQUFNLENBQUM7UUFDTixNQUFNLENBQUMsR0FBRyxDQUFDLGNBQWMsQ0FBQztZQUN4QixLQUFLLEVBQUUsS0FBSztTQUNiLENBQUMsQ0FBQyxDQUFBO1FBQ0gsTUFBTSxPQUFPLEdBQVcsTUFBTSxDQUFDLE1BQU0sQ0FBVyxDQUFDO1FBQ2pELE9BQU8sQ0FBQyxNQUFNLENBQUMsS0FBSyxDQUFDLE9BQU8sQ0FBQyxDQUFBO0lBQy9CLENBQUM7SUFDRCxZQUFZLENBQUMsT0FBTyxDQUFDLENBQUE7SUFDckIsa0JBQWtCO0lBQ2xCLE9BQU8sTUFBTSxDQUFBO0FBQ2YsQ0FBQyxDQUFBO0FBRUQsTUFBTSxDQUFDLE1BQU0sYUFBYSxHQUFHLEtBQUssRUFBRSxNQUFjLEVBQUUsTUFBVyxFQUFFLE9BQWtCLEVBQUUsRUFBRTtJQUNyRixJQUFJLE9BQU8sQ0FBQyxHQUFHLEVBQUUsQ0FBQztRQUNoQixPQUFPLENBQUMsTUFBTSxDQUFDLElBQUksQ0FBQyw2QkFBNkIsQ0FBQyxDQUFBO1FBQ2xELE9BQU8sS0FBSyxDQUFBO0lBQ2QsQ0FBQztJQUNELE1BQU0sVUFBVSxHQUFHLE1BQU0sTUFBTSxDQUFDLElBQUksQ0FBQyxXQUFXLENBQUMsTUFBTSxDQUFDO1FBQ3RELEtBQUssRUFBRSxPQUFPLENBQUMsS0FBSztRQUNwQixRQUFRLEVBQUUsTUFBTSxDQUFDLFFBQVE7UUFDekIsZUFBZSxFQUFFLE9BQU8sQ0FBQyxNQUFhO0tBQ3ZDLENBQUMsQ0FBQTtJQUNGLElBQUksQ0FBQyxVQUFVLElBQUksQ0FBQyxVQUFVLENBQUMsT0FBTyxJQUFJLENBQUMsVUFBVSxDQUFDLE9BQU8sQ0FBQyxDQUFDLENBQUMsSUFBSSxDQUFDLFVBQVUsQ0FBQyxPQUFPLENBQUMsQ0FBQyxDQUFDLENBQUMsT0FBTyxJQUFJLENBQUMsVUFBVSxDQUFDLE9BQU8sQ0FBQyxDQUFDLENBQUMsQ0FBQyxPQUFPLENBQUMsT0FBTyxFQUFFLENBQUM7UUFDN0ksT0FBTyxFQUFFLENBQUE7SUFDWCxDQUFDO0lBQ0QsSUFBSSxNQUFNLEdBQUcsVUFBVSxDQUFDLE9BQU8sQ0FBQyxDQUFDLENBQUMsQ0FBQyxPQUFPLENBQUMsT0FBTyxDQUFBO0lBQ2xELE1BQU0sR0FBRyxNQUFNLFlBQVksQ0FBQyxNQUFNLEVBQUUsT0FBTyxDQUFDLENBQUE7SUFDNUMsT0FBTyxNQUFNLENBQUE7QUFDZixDQUFDLENBQUEifQ==
|
||||
//# sourceMappingURL=data:application/json;base64,eyJ2ZXJzaW9uIjozLCJmaWxlIjoicnVuLWNvbXBsZXRpb24uanMiLCJzb3VyY2VSb290IjoiIiwic291cmNlcyI6WyIuLi8uLi9zcmMvY29tbWFuZHMvcnVuLWNvbXBsZXRpb24udHMiXSwibmFtZXMiOltdLCJtYXBwaW5ncyI6IkFBQ0EsT0FBTyxFQUFFLE1BQU0sRUFBRSxNQUFNLFFBQVEsQ0FBQTtBQUMvQixPQUFPLEtBQUssSUFBSSxNQUFNLFdBQVcsQ0FBQTtBQUNqQyxPQUFPLEVBQUUsT0FBTyxFQUFFLE1BQU0sbUJBQW1CLENBQUE7QUFHM0MsT0FBTyxFQUFFLFlBQVksRUFBRSxNQUFNLG9CQUFvQixDQUFBO0FBQ2pELE9BQU8sRUFBRSxZQUFZLEVBQVUsTUFBTSxlQUFlLENBQUE7QUFDcEQsT0FBTyxFQUFFLFNBQVMsRUFBRSxNQUFNLGlCQUFpQixDQUFBO0FBQzNDLE9BQU8sRUFBRSxhQUFhLEVBQUUsTUFBTSxtQkFBbUIsQ0FBQTtBQUdqRCxNQUFNLENBQUMsTUFBTSxZQUFZLEdBQUcsS0FBSyxFQUFFLFNBQWMsRUFBRSxFQUFFLE9BQWtCLEVBQUUsRUFBRTtJQUN6RSxNQUFNLEdBQUcsWUFBWSxDQUFDLE1BQU0sRUFBRSxPQUFPLENBQUMsT0FBbUIsSUFBSSxFQUFFLENBQUMsQ0FBQTtJQUNoRSxJQUFJLElBQUksR0FBRyxTQUFTLENBQUMsT0FBTyxDQUFDLENBQUE7SUFDN0IsSUFBSSxPQUFPLENBQUMsR0FBRyxFQUFFLENBQUM7UUFDaEIsSUFBSSxHQUFHO1lBQ0wsR0FBRyxJQUFJO1lBQ1AsS0FBSyxFQUFFLE9BQU8sQ0FBQyxLQUFLLENBQUMsQ0FBQyxDQUFDLElBQUksQ0FBQyxLQUFLLENBQUMsT0FBTyxDQUFDLEtBQUssQ0FBQyxDQUFDLElBQUksQ0FBQyxDQUFDLENBQUMsZUFBZTtZQUN2RSxNQUFNLEVBQUUsT0FBTyxDQUFDLE1BQU0sSUFBSSxnQkFBZ0I7U0FDM0MsQ0FBQTtRQUNELE1BQU0sT0FBTyxHQUFHLElBQUksQ0FBQyxPQUFPLENBQUMsT0FBTyxDQUFDLE9BQU8sQ0FBQyxHQUFHLEVBQUUsT0FBTyxDQUFDLEdBQUcsRUFBRSxJQUFJLENBQUMsQ0FBQyxDQUFBO1FBQ3JFLGFBQWEsQ0FBQyxPQUFPLEVBQUUsTUFBTSxFQUFFLE9BQU8sQ0FBQyxNQUEwQixDQUFDLENBQUE7UUFDbEUsTUFBTSxNQUFNLEdBQUcsT0FBTyxDQUFDLE1BQU0sQ0FBQyxDQUFDLENBQUMsR0FBRyxPQUFPLENBQUMsTUFBTSxhQUFhLENBQUMsQ0FBQyxDQUFDLDRCQUE0QixDQUFBO1FBQzdGLE9BQU8sQ0FBQyxNQUFNLEVBQUUsS0FBSyxDQUFDLEdBQUcsTUFBTSxJQUFJLE9BQU8sTUFBTSxPQUFPLENBQUMsR0FBRyxFQUFFLENBQUMsQ0FBQTtJQUNoRSxDQUFDO1NBQU0sQ0FBQztRQUNOLE1BQU0sT0FBTyxHQUFXLE1BQU0sQ0FBQyxNQUFNLENBQVcsQ0FBQztRQUNqRCxPQUFPLENBQUMsTUFBTSxDQUFDLEtBQUssQ0FBQyxPQUFPLENBQUMsQ0FBQTtJQUMvQixDQUFDO0lBQ0QsWUFBWSxDQUFDLE9BQU8sQ0FBQyxDQUFBO0lBQ3JCLGtCQUFrQjtJQUNsQixPQUFPLE1BQU0sQ0FBQTtBQUNmLENBQUMsQ0FBQTtBQUVELE1BQU0sQ0FBQyxNQUFNLGFBQWEsR0FBRyxLQUFLLEVBQUUsTUFBYyxFQUFFLE1BQVcsRUFBRSxPQUFrQixFQUFFLEVBQUU7SUFDckYsSUFBSSxPQUFPLENBQUMsR0FBRyxFQUFFLENBQUM7UUFDaEIsT0FBTyxDQUFDLE1BQU0sQ0FBQyxJQUFJLENBQUMsNkJBQTZCLENBQUMsQ0FBQTtRQUNsRCxPQUFPLEtBQUssQ0FBQTtJQUNkLENBQUM7SUFDRCxNQUFNLFVBQVUsR0FBRyxNQUFNLE1BQU0sQ0FBQyxJQUFJLENBQUMsV0FBVyxDQUFDLE1BQU0sQ0FBQztRQUN0RCxLQUFLLEVBQUUsT0FBTyxDQUFDLEtBQUs7UUFDcEIsUUFBUSxFQUFFLE1BQU0sQ0FBQyxRQUFRO1FBQ3pCLGVBQWUsRUFBRSxPQUFPLENBQUMsTUFBYTtLQUN2QyxDQUFDLENBQUE7SUFDRixJQUFJLENBQUMsVUFBVSxJQUFJLENBQUMsVUFBVSxDQUFDLE9BQU8sSUFBSSxDQUFDLFVBQVUsQ0FBQyxPQUFPLENBQUMsQ0FBQyxDQUFDLElBQUksQ0FBQyxVQUFVLENBQUMsT0FBTyxDQUFDLENBQUMsQ0FBQyxDQUFDLE9BQU8sSUFBSSxDQUFDLFVBQVUsQ0FBQyxPQUFPLENBQUMsQ0FBQyxDQUFDLENBQUMsT0FBTyxDQUFDLE9BQU8sRUFBRSxDQUFDO1FBQzdJLE9BQU8sRUFBRSxDQUFBO0lBQ1gsQ0FBQztJQUNELElBQUksTUFBTSxHQUFHLFVBQVUsQ0FBQyxPQUFPLENBQUMsQ0FBQyxDQUFDLENBQUMsT0FBTyxDQUFDLE9BQU8sQ0FBQTtJQUNsRCxNQUFNLEdBQUcsTUFBTSxZQUFZLENBQUMsTUFBTSxFQUFFLE9BQU8sQ0FBQyxDQUFBO0lBQzVDLE9BQU8sTUFBTSxDQUFBO0FBQ2YsQ0FBQyxDQUFBIn0=
|
||||
@ -1,5 +1,5 @@
|
||||
{
|
||||
"timestamp": 1776026492123,
|
||||
"timestamp": 1776333984881,
|
||||
"models": [
|
||||
{
|
||||
"id": "gpt-4-0613",
|
||||
|
||||
@ -1,6 +1,58 @@
|
||||
{
|
||||
"timestamp": 1776026492258,
|
||||
"timestamp": 1776333985020,
|
||||
"models": [
|
||||
{
|
||||
"id": "openrouter/elephant-alpha",
|
||||
"canonical_slug": "openrouter/elephant-alpha",
|
||||
"hugging_face_id": null,
|
||||
"name": "Elephant",
|
||||
"created": 1776052598,
|
||||
"description": "Elephant Alpha is a 100B-parameter text model focused on intelligence efficiency, delivering strong performance while minimizing token usage. It supports a 256K context window with up to 32K output tokens,...",
|
||||
"context_length": 262144,
|
||||
"architecture": {
|
||||
"modality": "text->text",
|
||||
"input_modalities": [
|
||||
"text"
|
||||
],
|
||||
"output_modalities": [
|
||||
"text"
|
||||
],
|
||||
"tokenizer": "Other",
|
||||
"instruct_type": null
|
||||
},
|
||||
"pricing": {
|
||||
"prompt": "0",
|
||||
"completion": "0"
|
||||
},
|
||||
"top_provider": {
|
||||
"context_length": 262144,
|
||||
"max_completion_tokens": 32768,
|
||||
"is_moderated": false
|
||||
},
|
||||
"per_request_limits": null,
|
||||
"supported_parameters": [
|
||||
"max_tokens",
|
||||
"response_format",
|
||||
"structured_outputs",
|
||||
"temperature",
|
||||
"tool_choice",
|
||||
"tools",
|
||||
"top_p"
|
||||
],
|
||||
"default_parameters": {
|
||||
"temperature": null,
|
||||
"top_p": null,
|
||||
"top_k": null,
|
||||
"frequency_penalty": null,
|
||||
"presence_penalty": null,
|
||||
"repetition_penalty": null
|
||||
},
|
||||
"knowledge_cutoff": null,
|
||||
"expiration_date": null,
|
||||
"links": {
|
||||
"details": "/api/v1/models/openrouter/elephant-alpha/endpoints"
|
||||
}
|
||||
},
|
||||
{
|
||||
"id": "anthropic/claude-opus-4.6-fast",
|
||||
"canonical_slug": "anthropic/claude-4.6-opus-fast-20260407",
|
||||
@ -98,8 +150,10 @@
|
||||
"logprobs",
|
||||
"max_tokens",
|
||||
"min_p",
|
||||
"parallel_tool_calls",
|
||||
"presence_penalty",
|
||||
"reasoning",
|
||||
"reasoning_effort",
|
||||
"repetition_penalty",
|
||||
"response_format",
|
||||
"seed",
|
||||
@ -201,12 +255,13 @@
|
||||
"instruct_type": null
|
||||
},
|
||||
"pricing": {
|
||||
"prompt": "0.00000012",
|
||||
"completion": "0.0000004"
|
||||
"prompt": "0.00000008",
|
||||
"completion": "0.00000035",
|
||||
"input_cache_read": "0.00000001"
|
||||
},
|
||||
"top_provider": {
|
||||
"context_length": 262144,
|
||||
"max_completion_tokens": 262144,
|
||||
"max_completion_tokens": null,
|
||||
"is_moderated": false
|
||||
},
|
||||
"per_request_limits": null,
|
||||
@ -214,7 +269,9 @@
|
||||
"frequency_penalty",
|
||||
"include_reasoning",
|
||||
"logit_bias",
|
||||
"logprobs",
|
||||
"max_tokens",
|
||||
"min_p",
|
||||
"presence_penalty",
|
||||
"reasoning",
|
||||
"repetition_penalty",
|
||||
@ -226,6 +283,7 @@
|
||||
"tool_choice",
|
||||
"tools",
|
||||
"top_k",
|
||||
"top_logprobs",
|
||||
"top_p"
|
||||
],
|
||||
"default_parameters": {
|
||||
@ -317,12 +375,13 @@
|
||||
"instruct_type": null
|
||||
},
|
||||
"pricing": {
|
||||
"prompt": "0.00000014",
|
||||
"completion": "0.0000004"
|
||||
"prompt": "0.00000013",
|
||||
"completion": "0.00000038",
|
||||
"input_cache_read": "0.0000000199999995"
|
||||
},
|
||||
"top_provider": {
|
||||
"context_length": 262144,
|
||||
"max_completion_tokens": 131072,
|
||||
"max_completion_tokens": null,
|
||||
"is_moderated": false
|
||||
},
|
||||
"per_request_limits": null,
|
||||
@ -331,6 +390,7 @@
|
||||
"include_reasoning",
|
||||
"logit_bias",
|
||||
"max_tokens",
|
||||
"min_p",
|
||||
"presence_penalty",
|
||||
"reasoning",
|
||||
"repetition_penalty",
|
||||
@ -493,7 +553,8 @@
|
||||
},
|
||||
"pricing": {
|
||||
"prompt": "0.00000022",
|
||||
"completion": "0.00000085"
|
||||
"completion": "0.00000085",
|
||||
"input_cache_read": "0.00000006"
|
||||
},
|
||||
"top_provider": {
|
||||
"context_length": 262144,
|
||||
@ -1000,7 +1061,7 @@
|
||||
"name": "MiniMax: MiniMax M2.7",
|
||||
"created": 1773836697,
|
||||
"description": "MiniMax-M2.7 is a next-generation large language model designed for autonomous, real-world productivity and continuous improvement. Built to actively participate in its own evolution, M2.7 integrates advanced agentic capabilities through multi-agent...",
|
||||
"context_length": 204800,
|
||||
"context_length": 196608,
|
||||
"architecture": {
|
||||
"modality": "text->text",
|
||||
"input_modalities": [
|
||||
@ -1015,22 +1076,32 @@
|
||||
"pricing": {
|
||||
"prompt": "0.0000003",
|
||||
"completion": "0.0000012",
|
||||
"input_cache_read": "0.00000006"
|
||||
"input_cache_read": "0.000000059"
|
||||
},
|
||||
"top_provider": {
|
||||
"context_length": 204800,
|
||||
"max_completion_tokens": 131072,
|
||||
"context_length": 196608,
|
||||
"max_completion_tokens": null,
|
||||
"is_moderated": false
|
||||
},
|
||||
"per_request_limits": null,
|
||||
"supported_parameters": [
|
||||
"frequency_penalty",
|
||||
"include_reasoning",
|
||||
"logit_bias",
|
||||
"logprobs",
|
||||
"max_tokens",
|
||||
"min_p",
|
||||
"presence_penalty",
|
||||
"reasoning",
|
||||
"repetition_penalty",
|
||||
"response_format",
|
||||
"stop",
|
||||
"structured_outputs",
|
||||
"temperature",
|
||||
"tool_choice",
|
||||
"tools",
|
||||
"top_k",
|
||||
"top_logprobs",
|
||||
"top_p"
|
||||
],
|
||||
"default_parameters": {
|
||||
@ -3095,6 +3166,7 @@
|
||||
"min_p",
|
||||
"presence_penalty",
|
||||
"reasoning",
|
||||
"reasoning_effort",
|
||||
"repetition_penalty",
|
||||
"response_format",
|
||||
"seed",
|
||||
@ -3915,7 +3987,6 @@
|
||||
"frequency_penalty",
|
||||
"include_reasoning",
|
||||
"logit_bias",
|
||||
"logprobs",
|
||||
"max_tokens",
|
||||
"min_p",
|
||||
"presence_penalty",
|
||||
@ -3929,7 +4000,6 @@
|
||||
"tool_choice",
|
||||
"tools",
|
||||
"top_k",
|
||||
"top_logprobs",
|
||||
"top_p"
|
||||
],
|
||||
"default_parameters": {
|
||||
@ -5371,7 +5441,10 @@
|
||||
"default_parameters": {
|
||||
"temperature": null,
|
||||
"top_p": null,
|
||||
"frequency_penalty": null
|
||||
"top_k": null,
|
||||
"frequency_penalty": null,
|
||||
"presence_penalty": null,
|
||||
"repetition_penalty": null
|
||||
},
|
||||
"knowledge_cutoff": null,
|
||||
"expiration_date": null,
|
||||
@ -5853,11 +5926,12 @@
|
||||
},
|
||||
"pricing": {
|
||||
"prompt": "0.0000006",
|
||||
"completion": "0.0000025"
|
||||
"completion": "0.0000025",
|
||||
"input_cache_read": "0.00000015"
|
||||
},
|
||||
"top_provider": {
|
||||
"context_length": 262144,
|
||||
"max_completion_tokens": null,
|
||||
"max_completion_tokens": 262144,
|
||||
"is_moderated": false
|
||||
},
|
||||
"per_request_limits": null,
|
||||
@ -8142,60 +8216,6 @@
|
||||
"details": "/api/v1/models/qwen/qwen3-next-80b-a3b-instruct-2509/endpoints"
|
||||
}
|
||||
},
|
||||
{
|
||||
"id": "meituan/longcat-flash-chat",
|
||||
"canonical_slug": "meituan/longcat-flash-chat",
|
||||
"hugging_face_id": "meituan-longcat/LongCat-Flash-Chat",
|
||||
"name": "Meituan: LongCat Flash Chat",
|
||||
"created": 1757427658,
|
||||
"description": "LongCat-Flash-Chat is a large-scale Mixture-of-Experts (MoE) model with 560B total parameters, of which 18.6B–31.3B (≈27B on average) are dynamically activated per input. It introduces a shortcut-connected MoE design to reduce...",
|
||||
"context_length": 131072,
|
||||
"architecture": {
|
||||
"modality": "text->text",
|
||||
"input_modalities": [
|
||||
"text"
|
||||
],
|
||||
"output_modalities": [
|
||||
"text"
|
||||
],
|
||||
"tokenizer": "Other",
|
||||
"instruct_type": null
|
||||
},
|
||||
"pricing": {
|
||||
"prompt": "0.0000002",
|
||||
"completion": "0.0000008",
|
||||
"input_cache_read": "0.0000002"
|
||||
},
|
||||
"top_provider": {
|
||||
"context_length": 131072,
|
||||
"max_completion_tokens": 131072,
|
||||
"is_moderated": false
|
||||
},
|
||||
"per_request_limits": null,
|
||||
"supported_parameters": [
|
||||
"frequency_penalty",
|
||||
"logit_bias",
|
||||
"max_tokens",
|
||||
"min_p",
|
||||
"presence_penalty",
|
||||
"repetition_penalty",
|
||||
"response_format",
|
||||
"seed",
|
||||
"stop",
|
||||
"structured_outputs",
|
||||
"temperature",
|
||||
"tool_choice",
|
||||
"tools",
|
||||
"top_k",
|
||||
"top_p"
|
||||
],
|
||||
"default_parameters": {},
|
||||
"knowledge_cutoff": "2025-03-31",
|
||||
"expiration_date": null,
|
||||
"links": {
|
||||
"details": "/api/v1/models/meituan/longcat-flash-chat/endpoints"
|
||||
}
|
||||
},
|
||||
{
|
||||
"id": "qwen/qwen-plus-2025-07-28:thinking",
|
||||
"canonical_slug": "qwen/qwen-plus-2025-07-28",
|
||||
@ -8440,8 +8460,6 @@
|
||||
"per_request_limits": null,
|
||||
"supported_parameters": [
|
||||
"frequency_penalty",
|
||||
"logit_bias",
|
||||
"logprobs",
|
||||
"max_tokens",
|
||||
"presence_penalty",
|
||||
"repetition_penalty",
|
||||
@ -8453,7 +8471,6 @@
|
||||
"tool_choice",
|
||||
"tools",
|
||||
"top_k",
|
||||
"top_logprobs",
|
||||
"top_p"
|
||||
],
|
||||
"default_parameters": {},
|
||||
@ -9743,7 +9760,6 @@
|
||||
"response_format",
|
||||
"seed",
|
||||
"stop",
|
||||
"structured_outputs",
|
||||
"temperature",
|
||||
"tool_choice",
|
||||
"tools",
|
||||
@ -9847,10 +9863,8 @@
|
||||
"presence_penalty",
|
||||
"reasoning",
|
||||
"repetition_penalty",
|
||||
"response_format",
|
||||
"seed",
|
||||
"stop",
|
||||
"structured_outputs",
|
||||
"temperature",
|
||||
"tool_choice",
|
||||
"tools",
|
||||
@ -10097,7 +10111,8 @@
|
||||
},
|
||||
"pricing": {
|
||||
"prompt": "0.0000001",
|
||||
"completion": "0.0000002"
|
||||
"completion": "0.0000002",
|
||||
"input_cache_read": "0.0000001"
|
||||
},
|
||||
"top_provider": {
|
||||
"context_length": 128000,
|
||||
@ -10315,7 +10330,7 @@
|
||||
},
|
||||
"top_provider": {
|
||||
"context_length": 131072,
|
||||
"max_completion_tokens": 131072,
|
||||
"max_completion_tokens": 32768,
|
||||
"is_moderated": false
|
||||
},
|
||||
"per_request_limits": null,
|
||||
@ -10883,56 +10898,6 @@
|
||||
"details": "/api/v1/models/baidu/ernie-4.5-300b-a47b/endpoints"
|
||||
}
|
||||
},
|
||||
{
|
||||
"id": "inception/mercury",
|
||||
"canonical_slug": "inception/mercury",
|
||||
"hugging_face_id": "",
|
||||
"name": "Inception: Mercury",
|
||||
"created": 1750973026,
|
||||
"description": "Mercury is the first diffusion large language model (dLLM). Applying a breakthrough discrete diffusion approach, the model runs 5-10x faster than even speed optimized models like GPT-4.1 Nano and Claude...",
|
||||
"context_length": 128000,
|
||||
"architecture": {
|
||||
"modality": "text->text",
|
||||
"input_modalities": [
|
||||
"text"
|
||||
],
|
||||
"output_modalities": [
|
||||
"text"
|
||||
],
|
||||
"tokenizer": "Other",
|
||||
"instruct_type": null
|
||||
},
|
||||
"pricing": {
|
||||
"prompt": "0.00000025",
|
||||
"completion": "0.00000075",
|
||||
"input_cache_read": "0.000000025"
|
||||
},
|
||||
"top_provider": {
|
||||
"context_length": 128000,
|
||||
"max_completion_tokens": 32000,
|
||||
"is_moderated": false
|
||||
},
|
||||
"per_request_limits": null,
|
||||
"supported_parameters": [
|
||||
"max_tokens",
|
||||
"response_format",
|
||||
"stop",
|
||||
"structured_outputs",
|
||||
"temperature",
|
||||
"tool_choice",
|
||||
"tools"
|
||||
],
|
||||
"default_parameters": {
|
||||
"temperature": 0,
|
||||
"top_p": null,
|
||||
"frequency_penalty": null
|
||||
},
|
||||
"knowledge_cutoff": "2025-01-31",
|
||||
"expiration_date": "2026-04-15",
|
||||
"links": {
|
||||
"details": "/api/v1/models/inception/mercury/endpoints"
|
||||
}
|
||||
},
|
||||
{
|
||||
"id": "mistralai/mistral-small-3.2-24b-instruct",
|
||||
"canonical_slug": "mistralai/mistral-small-3.2-24b-instruct-2506",
|
||||
@ -11629,8 +11594,8 @@
|
||||
"instruct_type": null
|
||||
},
|
||||
"pricing": {
|
||||
"prompt": "0.00000002",
|
||||
"completion": "0.00000004"
|
||||
"prompt": "0.00000006",
|
||||
"completion": "0.00000012"
|
||||
},
|
||||
"top_provider": {
|
||||
"context_length": 32768,
|
||||
@ -11968,16 +11933,17 @@
|
||||
}
|
||||
},
|
||||
{
|
||||
"id": "inception/mercury-coder",
|
||||
"canonical_slug": "inception/mercury-coder-small-beta",
|
||||
"hugging_face_id": "",
|
||||
"name": "Inception: Mercury Coder",
|
||||
"created": 1746033880,
|
||||
"description": "Mercury Coder is the first diffusion large language model (dLLM). Applying a breakthrough discrete diffusion approach, the model runs 5-10x faster than even speed optimized models like Claude 3.5 Haiku...",
|
||||
"context_length": 128000,
|
||||
"id": "meta-llama/llama-guard-4-12b:free",
|
||||
"canonical_slug": "meta-llama/llama-guard-4-12b",
|
||||
"hugging_face_id": "meta-llama/Llama-Guard-4-12B",
|
||||
"name": "Meta: Llama Guard 4 12B (free)",
|
||||
"created": 1745975193,
|
||||
"description": "Llama Guard 4 is a Llama 4 Scout-derived multimodal pretrained model, fine-tuned for content safety classification. Similar to previous versions, it can be used to classify content in both LLM...",
|
||||
"context_length": 163840,
|
||||
"architecture": {
|
||||
"modality": "text->text",
|
||||
"modality": "text+image->text",
|
||||
"input_modalities": [
|
||||
"image",
|
||||
"text"
|
||||
],
|
||||
"output_modalities": [
|
||||
@ -11987,34 +11953,26 @@
|
||||
"instruct_type": null
|
||||
},
|
||||
"pricing": {
|
||||
"prompt": "0.00000025",
|
||||
"completion": "0.00000075",
|
||||
"input_cache_read": "0.000000025"
|
||||
"prompt": "0",
|
||||
"completion": "0"
|
||||
},
|
||||
"top_provider": {
|
||||
"context_length": 128000,
|
||||
"max_completion_tokens": 32000,
|
||||
"context_length": 163840,
|
||||
"max_completion_tokens": 65000,
|
||||
"is_moderated": false
|
||||
},
|
||||
"per_request_limits": null,
|
||||
"supported_parameters": [
|
||||
"max_tokens",
|
||||
"response_format",
|
||||
"stop",
|
||||
"structured_outputs",
|
||||
"seed",
|
||||
"temperature",
|
||||
"tool_choice",
|
||||
"tools"
|
||||
"top_p"
|
||||
],
|
||||
"default_parameters": {
|
||||
"temperature": 0,
|
||||
"top_p": null,
|
||||
"frequency_penalty": null
|
||||
},
|
||||
"knowledge_cutoff": "2025-01-31",
|
||||
"expiration_date": "2026-04-15",
|
||||
"default_parameters": {},
|
||||
"knowledge_cutoff": "2024-08-31",
|
||||
"expiration_date": null,
|
||||
"links": {
|
||||
"details": "/api/v1/models/inception/mercury-coder-small-beta/endpoints"
|
||||
"details": "/api/v1/models/meta-llama/llama-guard-4-12b/endpoints"
|
||||
}
|
||||
},
|
||||
{
|
||||
@ -12505,53 +12463,6 @@
|
||||
"details": "/api/v1/models/openai/o4-mini-2025-04-16/endpoints"
|
||||
}
|
||||
},
|
||||
{
|
||||
"id": "qwen/qwen2.5-coder-7b-instruct",
|
||||
"canonical_slug": "qwen/qwen2.5-coder-7b-instruct",
|
||||
"hugging_face_id": "Qwen/Qwen2.5-Coder-7B-Instruct",
|
||||
"name": "Qwen: Qwen2.5 Coder 7B Instruct",
|
||||
"created": 1744734887,
|
||||
"description": "Qwen2.5-Coder-7B-Instruct is a 7B parameter instruction-tuned language model optimized for code-related tasks such as code generation, reasoning, and bug fixing. Based on the Qwen2.5 architecture, it incorporates enhancements like RoPE,...",
|
||||
"context_length": 32768,
|
||||
"architecture": {
|
||||
"modality": "text->text",
|
||||
"input_modalities": [
|
||||
"text"
|
||||
],
|
||||
"output_modalities": [
|
||||
"text"
|
||||
],
|
||||
"tokenizer": "Qwen",
|
||||
"instruct_type": null
|
||||
},
|
||||
"pricing": {
|
||||
"prompt": "0.00000003",
|
||||
"completion": "0.00000009"
|
||||
},
|
||||
"top_provider": {
|
||||
"context_length": 32768,
|
||||
"max_completion_tokens": null,
|
||||
"is_moderated": false
|
||||
},
|
||||
"per_request_limits": null,
|
||||
"supported_parameters": [
|
||||
"frequency_penalty",
|
||||
"max_tokens",
|
||||
"presence_penalty",
|
||||
"repetition_penalty",
|
||||
"response_format",
|
||||
"structured_outputs",
|
||||
"temperature",
|
||||
"top_k",
|
||||
"top_p"
|
||||
],
|
||||
"default_parameters": {},
|
||||
"knowledge_cutoff": "2024-06-30",
|
||||
"expiration_date": null,
|
||||
"links": {
|
||||
"details": "/api/v1/models/qwen/qwen2.5-coder-7b-instruct/endpoints"
|
||||
}
|
||||
},
|
||||
{
|
||||
"id": "openai/gpt-4.1",
|
||||
"canonical_slug": "openai/gpt-4.1-2025-04-14",
|
||||
@ -12704,54 +12615,6 @@
|
||||
"details": "/api/v1/models/openai/gpt-4.1-nano-2025-04-14/endpoints"
|
||||
}
|
||||
},
|
||||
{
|
||||
"id": "eleutherai/llemma_7b",
|
||||
"canonical_slug": "eleutherai/llemma_7b",
|
||||
"hugging_face_id": "EleutherAI/llemma_7b",
|
||||
"name": "EleutherAI: Llemma 7b",
|
||||
"created": 1744643225,
|
||||
"description": "Llemma 7B is a language model for mathematics. It was initialized with Code Llama 7B weights, and trained on the Proof-Pile-2 for 200B tokens. Llemma models are particularly strong at...",
|
||||
"context_length": 4096,
|
||||
"architecture": {
|
||||
"modality": "text->text",
|
||||
"input_modalities": [
|
||||
"text"
|
||||
],
|
||||
"output_modalities": [
|
||||
"text"
|
||||
],
|
||||
"tokenizer": "Other",
|
||||
"instruct_type": "code-llama"
|
||||
},
|
||||
"pricing": {
|
||||
"prompt": "0.0000008",
|
||||
"completion": "0.0000012"
|
||||
},
|
||||
"top_provider": {
|
||||
"context_length": 4096,
|
||||
"max_completion_tokens": 4096,
|
||||
"is_moderated": false
|
||||
},
|
||||
"per_request_limits": null,
|
||||
"supported_parameters": [
|
||||
"frequency_penalty",
|
||||
"max_tokens",
|
||||
"min_p",
|
||||
"presence_penalty",
|
||||
"repetition_penalty",
|
||||
"seed",
|
||||
"stop",
|
||||
"temperature",
|
||||
"top_k",
|
||||
"top_p"
|
||||
],
|
||||
"default_parameters": {},
|
||||
"knowledge_cutoff": "2023-04-30",
|
||||
"expiration_date": null,
|
||||
"links": {
|
||||
"details": "/api/v1/models/eleutherai/llemma_7b/endpoints"
|
||||
}
|
||||
},
|
||||
{
|
||||
"id": "alfredpros/codellama-7b-instruct-solidity",
|
||||
"canonical_slug": "alfredpros/codellama-7b-instruct-solidity",
|
||||
@ -12908,62 +12771,6 @@
|
||||
"details": "/api/v1/models/x-ai/grok-3-beta/endpoints"
|
||||
}
|
||||
},
|
||||
{
|
||||
"id": "nvidia/llama-3.1-nemotron-ultra-253b-v1",
|
||||
"canonical_slug": "nvidia/llama-3.1-nemotron-ultra-253b-v1",
|
||||
"hugging_face_id": "nvidia/Llama-3_1-Nemotron-Ultra-253B-v1",
|
||||
"name": "NVIDIA: Llama 3.1 Nemotron Ultra 253B v1",
|
||||
"created": 1744115059,
|
||||
"description": "Llama-3.1-Nemotron-Ultra-253B-v1 is a large language model (LLM) optimized for advanced reasoning, human-interactive chat, retrieval-augmented generation (RAG), and tool-calling tasks. Derived from Meta’s Llama-3.1-405B-Instruct, it has been significantly customized using Neural...",
|
||||
"context_length": 131072,
|
||||
"architecture": {
|
||||
"modality": "text->text",
|
||||
"input_modalities": [
|
||||
"text"
|
||||
],
|
||||
"output_modalities": [
|
||||
"text"
|
||||
],
|
||||
"tokenizer": "Llama3",
|
||||
"instruct_type": null
|
||||
},
|
||||
"pricing": {
|
||||
"prompt": "0.0000006",
|
||||
"completion": "0.0000018"
|
||||
},
|
||||
"top_provider": {
|
||||
"context_length": 131072,
|
||||
"max_completion_tokens": null,
|
||||
"is_moderated": false
|
||||
},
|
||||
"per_request_limits": null,
|
||||
"supported_parameters": [
|
||||
"frequency_penalty",
|
||||
"include_reasoning",
|
||||
"max_tokens",
|
||||
"presence_penalty",
|
||||
"reasoning",
|
||||
"repetition_penalty",
|
||||
"response_format",
|
||||
"structured_outputs",
|
||||
"temperature",
|
||||
"top_k",
|
||||
"top_p"
|
||||
],
|
||||
"default_parameters": {
|
||||
"temperature": null,
|
||||
"top_p": null,
|
||||
"top_k": null,
|
||||
"frequency_penalty": null,
|
||||
"presence_penalty": null,
|
||||
"repetition_penalty": null
|
||||
},
|
||||
"knowledge_cutoff": "2024-03-31",
|
||||
"expiration_date": null,
|
||||
"links": {
|
||||
"details": "/api/v1/models/nvidia/llama-3.1-nemotron-ultra-253b-v1/endpoints"
|
||||
}
|
||||
},
|
||||
{
|
||||
"id": "meta-llama/llama-4-maverick",
|
||||
"canonical_slug": "meta-llama/llama-4-maverick-17b-128e-instruct",
|
||||
@ -13816,7 +13623,8 @@
|
||||
},
|
||||
"pricing": {
|
||||
"prompt": "0.00000055",
|
||||
"completion": "0.0000008"
|
||||
"completion": "0.0000008",
|
||||
"input_cache_read": "0.00000025"
|
||||
},
|
||||
"top_provider": {
|
||||
"context_length": 32768,
|
||||
@ -14281,8 +14089,8 @@
|
||||
"instruct_type": "none"
|
||||
},
|
||||
"pricing": {
|
||||
"prompt": "0.00000002",
|
||||
"completion": "0.00000006"
|
||||
"prompt": "0.00000048",
|
||||
"completion": "0.00000003"
|
||||
},
|
||||
"top_provider": {
|
||||
"context_length": 131072,
|
||||
@ -17360,51 +17168,6 @@
|
||||
"details": "/api/v1/models/google/gemma-2-27b-it/endpoints"
|
||||
}
|
||||
},
|
||||
{
|
||||
"id": "google/gemma-2-9b-it",
|
||||
"canonical_slug": "google/gemma-2-9b-it",
|
||||
"hugging_face_id": "google/gemma-2-9b-it",
|
||||
"name": "Google: Gemma 2 9B",
|
||||
"created": 1719532800,
|
||||
"description": "Gemma 2 9B by Google is an advanced, open-source language model that sets a new standard for efficiency and performance in its size class. Designed for a wide variety of...",
|
||||
"context_length": 8192,
|
||||
"architecture": {
|
||||
"modality": "text->text",
|
||||
"input_modalities": [
|
||||
"text"
|
||||
],
|
||||
"output_modalities": [
|
||||
"text"
|
||||
],
|
||||
"tokenizer": "Gemini",
|
||||
"instruct_type": "gemma"
|
||||
},
|
||||
"pricing": {
|
||||
"prompt": "0.00000003",
|
||||
"completion": "0.00000009"
|
||||
},
|
||||
"top_provider": {
|
||||
"context_length": 8192,
|
||||
"max_completion_tokens": null,
|
||||
"is_moderated": false
|
||||
},
|
||||
"per_request_limits": null,
|
||||
"supported_parameters": [
|
||||
"frequency_penalty",
|
||||
"max_tokens",
|
||||
"presence_penalty",
|
||||
"repetition_penalty",
|
||||
"temperature",
|
||||
"top_k",
|
||||
"top_p"
|
||||
],
|
||||
"default_parameters": {},
|
||||
"knowledge_cutoff": "2024-06-30",
|
||||
"expiration_date": null,
|
||||
"links": {
|
||||
"details": "/api/v1/models/google/gemma-2-9b-it/endpoints"
|
||||
}
|
||||
},
|
||||
{
|
||||
"id": "sao10k/l3-euryale-70b",
|
||||
"canonical_slug": "sao10k/l3-euryale-70b",
|
||||
@ -18156,7 +17919,6 @@
|
||||
"per_request_limits": null,
|
||||
"supported_parameters": [
|
||||
"frequency_penalty",
|
||||
"logit_bias",
|
||||
"max_tokens",
|
||||
"min_p",
|
||||
"presence_penalty",
|
||||
|
||||
@ -1,6 +1,19 @@
|
||||
export declare enum E_OPENROUTER_MODEL {
|
||||
MODEL_ANTHROPIC_CLAUDE_OPUS_4_6_FAST = "anthropic/claude-opus-4.6-fast",
|
||||
MODEL_Z_AI_GLM_5_1 = "z-ai/glm-5.1",
|
||||
MODEL_GOOGLE_GEMMA_4_26B_A4B_IT_FREE = "google/gemma-4-26b-a4b-it:free",
|
||||
MODEL_GOOGLE_GEMMA_4_26B_A4B_IT = "google/gemma-4-26b-a4b-it",
|
||||
MODEL_GOOGLE_GEMMA_4_31B_IT_FREE = "google/gemma-4-31b-it:free",
|
||||
MODEL_GOOGLE_GEMMA_4_31B_IT = "google/gemma-4-31b-it",
|
||||
MODEL_QWEN_QWEN3_6_PLUS = "qwen/qwen3.6-plus",
|
||||
MODEL_Z_AI_GLM_5V_TURBO = "z-ai/glm-5v-turbo",
|
||||
MODEL_ARCEE_AI_TRINITY_LARGE_THINKING = "arcee-ai/trinity-large-thinking",
|
||||
MODEL_X_AI_GROK_4_20_MULTI_AGENT = "x-ai/grok-4.20-multi-agent",
|
||||
MODEL_X_AI_GROK_4_20 = "x-ai/grok-4.20",
|
||||
MODEL_GOOGLE_LYRIA_3_PRO_PREVIEW = "google/lyria-3-pro-preview",
|
||||
MODEL_GOOGLE_LYRIA_3_CLIP_PREVIEW = "google/lyria-3-clip-preview",
|
||||
MODEL_KWAIPILOT_KAT_CODER_PRO_V2 = "kwaipilot/kat-coder-pro-v2",
|
||||
MODEL_REKA_REKA_EDGE = "reka/reka-edge",
|
||||
MODEL_REKAAI_REKA_EDGE = "rekaai/reka-edge",
|
||||
MODEL_XIAOMI_MIMO_V2_OMNI = "xiaomi/mimo-v2-omni",
|
||||
MODEL_XIAOMI_MIMO_V2_PRO = "xiaomi/mimo-v2-pro",
|
||||
MODEL_MINIMAX_MINIMAX_M2_7 = "minimax/minimax-m2.7",
|
||||
@ -8,8 +21,6 @@ export declare enum E_OPENROUTER_MODEL {
|
||||
MODEL_OPENAI_GPT_5_4_MINI = "openai/gpt-5.4-mini",
|
||||
MODEL_MISTRALAI_MISTRAL_SMALL_2603 = "mistralai/mistral-small-2603",
|
||||
MODEL_Z_AI_GLM_5_TURBO = "z-ai/glm-5-turbo",
|
||||
MODEL_X_AI_GROK_4_20_MULTI_AGENT_BETA = "x-ai/grok-4.20-multi-agent-beta",
|
||||
MODEL_X_AI_GROK_4_20_BETA = "x-ai/grok-4.20-beta",
|
||||
MODEL_NVIDIA_NEMOTRON_3_SUPER_120B_A12B_FREE = "nvidia/nemotron-3-super-120b-a12b:free",
|
||||
MODEL_NVIDIA_NEMOTRON_3_SUPER_120B_A12B = "nvidia/nemotron-3-super-120b-a12b",
|
||||
MODEL_BYTEDANCE_SEED_SEED_2_0_LITE = "bytedance-seed/seed-2.0-lite",
|
||||
@ -40,7 +51,6 @@ export declare enum E_OPENROUTER_MODEL {
|
||||
MODEL_ANTHROPIC_CLAUDE_OPUS_4_6 = "anthropic/claude-opus-4.6",
|
||||
MODEL_QWEN_QWEN3_CODER_NEXT = "qwen/qwen3-coder-next",
|
||||
MODEL_OPENROUTER_FREE = "openrouter/free",
|
||||
MODEL_STEPFUN_STEP_3_5_FLASH_FREE = "stepfun/step-3.5-flash:free",
|
||||
MODEL_STEPFUN_STEP_3_5_FLASH = "stepfun/step-3.5-flash",
|
||||
MODEL_ARCEE_AI_TRINITY_LARGE_PREVIEW_FREE = "arcee-ai/trinity-large-preview:free",
|
||||
MODEL_MOONSHOTAI_KIMI_K2_5 = "moonshotai/kimi-k2.5",
|
||||
@ -60,7 +70,6 @@ export declare enum E_OPENROUTER_MODEL {
|
||||
MODEL_Z_AI_GLM_4_7 = "z-ai/glm-4.7",
|
||||
MODEL_GOOGLE_GEMINI_3_FLASH_PREVIEW = "google/gemini-3-flash-preview",
|
||||
MODEL_MISTRALAI_MISTRAL_SMALL_CREATIVE = "mistralai/mistral-small-creative",
|
||||
MODEL_ALLENAI_OLMO_3_1_32B_THINK = "allenai/olmo-3.1-32b-think",
|
||||
MODEL_XIAOMI_MIMO_V2_FLASH = "xiaomi/mimo-v2-flash",
|
||||
MODEL_NVIDIA_NEMOTRON_3_NANO_30B_A3B_FREE = "nvidia/nemotron-3-nano-30b-a3b:free",
|
||||
MODEL_NVIDIA_NEMOTRON_3_NANO_30B_A3B = "nvidia/nemotron-3-nano-30b-a3b",
|
||||
@ -79,7 +88,6 @@ export declare enum E_OPENROUTER_MODEL {
|
||||
MODEL_MISTRALAI_MINISTRAL_8B_2512 = "mistralai/ministral-8b-2512",
|
||||
MODEL_MISTRALAI_MINISTRAL_3B_2512 = "mistralai/ministral-3b-2512",
|
||||
MODEL_MISTRALAI_MISTRAL_LARGE_2512 = "mistralai/mistral-large-2512",
|
||||
MODEL_ARCEE_AI_TRINITY_MINI_FREE = "arcee-ai/trinity-mini:free",
|
||||
MODEL_ARCEE_AI_TRINITY_MINI = "arcee-ai/trinity-mini",
|
||||
MODEL_DEEPSEEK_DEEPSEEK_V3_2_SPECIALE = "deepseek/deepseek-v3.2-speciale",
|
||||
MODEL_DEEPSEEK_DEEPSEEK_V3_2 = "deepseek/deepseek-v3.2",
|
||||
@ -93,7 +101,6 @@ export declare enum E_OPENROUTER_MODEL {
|
||||
MODEL_OPENAI_GPT_5_1_CHAT = "openai/gpt-5.1-chat",
|
||||
MODEL_OPENAI_GPT_5_1_CODEX = "openai/gpt-5.1-codex",
|
||||
MODEL_OPENAI_GPT_5_1_CODEX_MINI = "openai/gpt-5.1-codex-mini",
|
||||
MODEL_KWAIPILOT_KAT_CODER_PRO = "kwaipilot/kat-coder-pro",
|
||||
MODEL_MOONSHOTAI_KIMI_K2_THINKING = "moonshotai/kimi-k2-thinking",
|
||||
MODEL_AMAZON_NOVA_PREMIER_V1 = "amazon/nova-premier-v1",
|
||||
MODEL_PERPLEXITY_SONAR_PRO_SEARCH = "perplexity/sonar-pro-search",
|
||||
@ -103,8 +110,6 @@ export declare enum E_OPENROUTER_MODEL {
|
||||
MODEL_NVIDIA_NEMOTRON_NANO_12B_V2_VL = "nvidia/nemotron-nano-12b-v2-vl",
|
||||
MODEL_MINIMAX_MINIMAX_M2 = "minimax/minimax-m2",
|
||||
MODEL_QWEN_QWEN3_VL_32B_INSTRUCT = "qwen/qwen3-vl-32b-instruct",
|
||||
MODEL_LIQUID_LFM2_8B_A1B = "liquid/lfm2-8b-a1b",
|
||||
MODEL_LIQUID_LFM_2_2_6B = "liquid/lfm-2.2-6b",
|
||||
MODEL_IBM_GRANITE_GRANITE_4_0_H_MICRO = "ibm-granite/granite-4.0-h-micro",
|
||||
MODEL_OPENAI_GPT_5_IMAGE_MINI = "openai/gpt-5-image-mini",
|
||||
MODEL_ANTHROPIC_CLAUDE_HAIKU_4_5 = "anthropic/claude-haiku-4.5",
|
||||
@ -242,6 +247,7 @@ export declare enum E_OPENROUTER_MODEL {
|
||||
MODEL_COHERE_COMMAND_A = "cohere/command-a",
|
||||
MODEL_OPENAI_GPT_4O_MINI_SEARCH_PREVIEW = "openai/gpt-4o-mini-search-preview",
|
||||
MODEL_OPENAI_GPT_4O_SEARCH_PREVIEW = "openai/gpt-4o-search-preview",
|
||||
MODEL_REKAAI_REKA_FLASH_3 = "rekaai/reka-flash-3",
|
||||
MODEL_GOOGLE_GEMMA_3_27B_IT_FREE = "google/gemma-3-27b-it:free",
|
||||
MODEL_GOOGLE_GEMMA_3_27B_IT = "google/gemma-3-27b-it",
|
||||
MODEL_THEDRUMMER_SKYFALL_36B_V2 = "thedrummer/skyfall-36b-v2",
|
||||
@ -291,7 +297,6 @@ export declare enum E_OPENROUTER_MODEL {
|
||||
MODEL_THEDRUMMER_UNSLOPNEMO_12B = "thedrummer/unslopnemo-12b",
|
||||
MODEL_ANTHROPIC_CLAUDE_3_5_HAIKU = "anthropic/claude-3.5-haiku",
|
||||
MODEL_ANTHRACITE_ORG_MAGNUM_V4_72B = "anthracite-org/magnum-v4-72b",
|
||||
MODEL_ANTHROPIC_CLAUDE_3_5_SONNET = "anthropic/claude-3.5-sonnet",
|
||||
MODEL_QWEN_QWEN_2_5_7B_INSTRUCT = "qwen/qwen-2.5-7b-instruct",
|
||||
MODEL_NVIDIA_LLAMA_3_1_NEMOTRON_70B_INSTRUCT = "nvidia/llama-3.1-nemotron-70b-instruct",
|
||||
MODEL_INFLECTION_INFLECTION_3_PRODUCTIVITY = "inflection/inflection-3-productivity",
|
||||
@ -313,15 +318,15 @@ export declare enum E_OPENROUTER_MODEL {
|
||||
MODEL_META_LLAMA_LLAMA_3_1_8B_INSTRUCT = "meta-llama/llama-3.1-8b-instruct",
|
||||
MODEL_META_LLAMA_LLAMA_3_1_70B_INSTRUCT = "meta-llama/llama-3.1-70b-instruct",
|
||||
MODEL_MISTRALAI_MISTRAL_NEMO = "mistralai/mistral-nemo",
|
||||
MODEL_OPENAI_GPT_4O_MINI = "openai/gpt-4o-mini",
|
||||
MODEL_OPENAI_GPT_4O_MINI_2024_07_18 = "openai/gpt-4o-mini-2024-07-18",
|
||||
MODEL_OPENAI_GPT_4O_MINI = "openai/gpt-4o-mini",
|
||||
MODEL_GOOGLE_GEMMA_2_27B_IT = "google/gemma-2-27b-it",
|
||||
MODEL_GOOGLE_GEMMA_2_9B_IT = "google/gemma-2-9b-it",
|
||||
MODEL_SAO10K_L3_EURYALE_70B = "sao10k/l3-euryale-70b",
|
||||
MODEL_NOUSRESEARCH_HERMES_2_PRO_LLAMA_3_8B = "nousresearch/hermes-2-pro-llama-3-8b",
|
||||
MODEL_OPENAI_GPT_4O_2024_05_13 = "openai/gpt-4o-2024-05-13",
|
||||
MODEL_OPENAI_GPT_4O = "openai/gpt-4o",
|
||||
MODEL_OPENAI_GPT_4O_EXTENDED = "openai/gpt-4o:extended",
|
||||
MODEL_OPENAI_GPT_4O_2024_05_13 = "openai/gpt-4o-2024-05-13",
|
||||
MODEL_META_LLAMA_LLAMA_3_8B_INSTRUCT = "meta-llama/llama-3-8b-instruct",
|
||||
MODEL_META_LLAMA_LLAMA_3_70B_INSTRUCT = "meta-llama/llama-3-70b-instruct",
|
||||
MODEL_MISTRALAI_MIXTRAL_8X22B_INSTRUCT = "mistralai/mixtral-8x22b-instruct",
|
||||
@ -341,7 +346,7 @@ export declare enum E_OPENROUTER_MODEL {
|
||||
MODEL_MANCER_WEAVER = "mancer/weaver",
|
||||
MODEL_UNDI95_REMM_SLERP_L2_13B = "undi95/remm-slerp-l2-13b",
|
||||
MODEL_GRYPHE_MYTHOMAX_L2_13B = "gryphe/mythomax-l2-13b",
|
||||
MODEL_OPENAI_GPT_4 = "openai/gpt-4",
|
||||
MODEL_OPENAI_GPT_4_0314 = "openai/gpt-4-0314",
|
||||
MODEL_OPENAI_GPT_4 = "openai/gpt-4",
|
||||
MODEL_OPENAI_GPT_3_5_TURBO = "openai/gpt-3.5-turbo"
|
||||
}
|
||||
|
||||
File diff suppressed because one or more lines are too long
File diff suppressed because one or more lines are too long
60201
packages/kbot/dist/main_node.js
vendored
60201
packages/kbot/dist/main_node.js
vendored
File diff suppressed because one or more lines are too long
317
packages/kbot/package-lock.json
generated
317
packages/kbot/package-lock.json
generated
@ -40,7 +40,6 @@
|
||||
"keyv": "^5.5.0",
|
||||
"keyv-file": "^5.1.3",
|
||||
"marked": "14.1.4",
|
||||
"marked-terminal": "7.2.1",
|
||||
"mime-types": "2.1.35",
|
||||
"nodemailer": "^7.0.13",
|
||||
"openai": "6.32.0",
|
||||
@ -350,16 +349,6 @@
|
||||
"node": ">=18"
|
||||
}
|
||||
},
|
||||
"node_modules/@colors/colors": {
|
||||
"version": "1.5.0",
|
||||
"resolved": "https://registry.npmjs.org/@colors/colors/-/colors-1.5.0.tgz",
|
||||
"integrity": "sha512-ooWCrlZP11i8GImSjTHYHLkvFDP48nS4+204nGb1RiX/WXYHmJA2III9/e2DWVabCESdW7hBAEzHRqUn9OUVvQ==",
|
||||
"license": "MIT",
|
||||
"optional": true,
|
||||
"engines": {
|
||||
"node": ">=0.1.90"
|
||||
}
|
||||
},
|
||||
"node_modules/@cspotcode/source-map-support": {
|
||||
"version": "0.8.1",
|
||||
"resolved": "https://registry.npmjs.org/@cspotcode/source-map-support/-/source-map-support-0.8.1.tgz",
|
||||
@ -2065,18 +2054,6 @@
|
||||
"dev": true,
|
||||
"license": "MIT"
|
||||
},
|
||||
"node_modules/@sindresorhus/is": {
|
||||
"version": "4.6.0",
|
||||
"resolved": "https://registry.npmjs.org/@sindresorhus/is/-/is-4.6.0.tgz",
|
||||
"integrity": "sha512-t09vSN3MdfsyCHoFcTRCH/iUtG7OJ0CsjzB8cjAmKc/va/kIgeDI/TxsigdncE/4be734m0cvIYwNaV4i2XqAw==",
|
||||
"license": "MIT",
|
||||
"engines": {
|
||||
"node": ">=10"
|
||||
},
|
||||
"funding": {
|
||||
"url": "https://github.com/sindresorhus/is?sponsor=1"
|
||||
}
|
||||
},
|
||||
"node_modules/@so-ric/colorspace": {
|
||||
"version": "1.1.6",
|
||||
"resolved": "https://registry.npmjs.org/@so-ric/colorspace/-/colorspace-1.1.6.tgz",
|
||||
@ -3023,33 +3000,6 @@
|
||||
"dev": true,
|
||||
"license": "MIT"
|
||||
},
|
||||
"node_modules/ansi-escapes": {
|
||||
"version": "7.3.0",
|
||||
"resolved": "https://registry.npmjs.org/ansi-escapes/-/ansi-escapes-7.3.0.tgz",
|
||||
"integrity": "sha512-BvU8nYgGQBxcmMuEeUEmNTvrMVjJNSH7RgW24vXexN4Ven6qCvy4TntnvlnwnMLTVlcRQQdbRY8NKnaIoeWDNg==",
|
||||
"license": "MIT",
|
||||
"dependencies": {
|
||||
"environment": "^1.0.0"
|
||||
},
|
||||
"engines": {
|
||||
"node": ">=18"
|
||||
},
|
||||
"funding": {
|
||||
"url": "https://github.com/sponsors/sindresorhus"
|
||||
}
|
||||
},
|
||||
"node_modules/ansi-regex": {
|
||||
"version": "6.2.2",
|
||||
"resolved": "https://registry.npmjs.org/ansi-regex/-/ansi-regex-6.2.2.tgz",
|
||||
"integrity": "sha512-Bq3SmSpyFHaWjPk8If9yc6svM8c56dB5BAtW4Qbw5jHTwwXXcTLoRMkpDJp6VL0XzlWaCHTXrkFURMYmD0sLqg==",
|
||||
"license": "MIT",
|
||||
"engines": {
|
||||
"node": ">=12"
|
||||
},
|
||||
"funding": {
|
||||
"url": "https://github.com/chalk/ansi-regex?sponsor=1"
|
||||
}
|
||||
},
|
||||
"node_modules/ansi-styles": {
|
||||
"version": "4.3.0",
|
||||
"resolved": "https://registry.npmjs.org/ansi-styles/-/ansi-styles-4.3.0.tgz",
|
||||
@ -3065,12 +3015,6 @@
|
||||
"url": "https://github.com/chalk/ansi-styles?sponsor=1"
|
||||
}
|
||||
},
|
||||
"node_modules/any-promise": {
|
||||
"version": "1.3.0",
|
||||
"resolved": "https://registry.npmjs.org/any-promise/-/any-promise-1.3.0.tgz",
|
||||
"integrity": "sha512-7UvmKalWRt1wgjL1RrGxoSJW/0QZFIegpeGvZG9kjp8vrRu55XTHbwnqq2GpXm9uLbcuhxm3IqX9OB4MZR1b2A==",
|
||||
"license": "MIT"
|
||||
},
|
||||
"node_modules/arg": {
|
||||
"version": "4.1.3",
|
||||
"resolved": "https://registry.npmjs.org/arg/-/arg-4.1.3.tgz",
|
||||
@ -3815,15 +3759,6 @@
|
||||
"node": ">=8"
|
||||
}
|
||||
},
|
||||
"node_modules/char-regex": {
|
||||
"version": "1.0.2",
|
||||
"resolved": "https://registry.npmjs.org/char-regex/-/char-regex-1.0.2.tgz",
|
||||
"integrity": "sha512-kWWXztvZ5SBQV+eRgKFeh8q5sLuZY2+8WUIzlxWVTg+oGwY14qylx1KbKzHd8P6ZYkAg0xyIDU9JMHhyJMZ1jw==",
|
||||
"license": "MIT",
|
||||
"engines": {
|
||||
"node": ">=10"
|
||||
}
|
||||
},
|
||||
"node_modules/character-entities": {
|
||||
"version": "2.0.2",
|
||||
"resolved": "https://registry.npmjs.org/character-entities/-/character-entities-2.0.2.tgz",
|
||||
@ -3944,80 +3879,6 @@
|
||||
"node": ">= 0.10"
|
||||
}
|
||||
},
|
||||
"node_modules/cli-highlight": {
|
||||
"version": "2.1.11",
|
||||
"resolved": "https://registry.npmjs.org/cli-highlight/-/cli-highlight-2.1.11.tgz",
|
||||
"integrity": "sha512-9KDcoEVwyUXrjcJNvHD0NFc/hiwe/WPVYIleQh2O1N2Zro5gWJZ/K+3DGn8w8P/F6FxOgzyC5bxDyHIgCSPhGg==",
|
||||
"license": "ISC",
|
||||
"dependencies": {
|
||||
"chalk": "^4.0.0",
|
||||
"highlight.js": "^10.7.1",
|
||||
"mz": "^2.4.0",
|
||||
"parse5": "^5.1.1",
|
||||
"parse5-htmlparser2-tree-adapter": "^6.0.0",
|
||||
"yargs": "^16.0.0"
|
||||
},
|
||||
"bin": {
|
||||
"highlight": "bin/highlight"
|
||||
},
|
||||
"engines": {
|
||||
"node": ">=8.0.0",
|
||||
"npm": ">=5.0.0"
|
||||
}
|
||||
},
|
||||
"node_modules/cli-highlight/node_modules/cliui": {
|
||||
"version": "7.0.4",
|
||||
"resolved": "https://registry.npmjs.org/cliui/-/cliui-7.0.4.tgz",
|
||||
"integrity": "sha512-OcRE68cOsVMXp1Yvonl/fzkQOyjLSu/8bhPDfQt0e0/Eb283TKP20Fs2MqoPsr9SwA595rRCA+QMzYc9nBP+JQ==",
|
||||
"license": "ISC",
|
||||
"dependencies": {
|
||||
"string-width": "^4.2.0",
|
||||
"strip-ansi": "^6.0.0",
|
||||
"wrap-ansi": "^7.0.0"
|
||||
}
|
||||
},
|
||||
"node_modules/cli-highlight/node_modules/yargs": {
|
||||
"version": "16.2.0",
|
||||
"resolved": "https://registry.npmjs.org/yargs/-/yargs-16.2.0.tgz",
|
||||
"integrity": "sha512-D1mvvtDG0L5ft/jGWkLpG1+m0eQxOfaBvTNELraWj22wSVUMWxZUvYgJYcKh6jGGIkJFhH4IZPQhR4TKpc8mBw==",
|
||||
"license": "MIT",
|
||||
"dependencies": {
|
||||
"cliui": "^7.0.2",
|
||||
"escalade": "^3.1.1",
|
||||
"get-caller-file": "^2.0.5",
|
||||
"require-directory": "^2.1.1",
|
||||
"string-width": "^4.2.0",
|
||||
"y18n": "^5.0.5",
|
||||
"yargs-parser": "^20.2.2"
|
||||
},
|
||||
"engines": {
|
||||
"node": ">=10"
|
||||
}
|
||||
},
|
||||
"node_modules/cli-highlight/node_modules/yargs-parser": {
|
||||
"version": "20.2.9",
|
||||
"resolved": "https://registry.npmjs.org/yargs-parser/-/yargs-parser-20.2.9.tgz",
|
||||
"integrity": "sha512-y11nGElTIV+CT3Zv9t7VKl+Q3hTQoT9a1Qzezhhl6Rp21gJ/IVTW7Z3y9EWXhuUBC2Shnf+DX0antecpAwSP8w==",
|
||||
"license": "ISC",
|
||||
"engines": {
|
||||
"node": ">=10"
|
||||
}
|
||||
},
|
||||
"node_modules/cli-table3": {
|
||||
"version": "0.6.5",
|
||||
"resolved": "https://registry.npmjs.org/cli-table3/-/cli-table3-0.6.5.tgz",
|
||||
"integrity": "sha512-+W/5efTR7y5HRD7gACw9yQjqMVvEMLBHmboM/kPWam+H+Hmyrgjh6YncVKK122YZkXrLudzTuAukUw9FnMf7IQ==",
|
||||
"license": "MIT",
|
||||
"dependencies": {
|
||||
"string-width": "^4.2.0"
|
||||
},
|
||||
"engines": {
|
||||
"node": "10.* || >= 12.*"
|
||||
},
|
||||
"optionalDependencies": {
|
||||
"@colors/colors": "1.5.0"
|
||||
}
|
||||
},
|
||||
"node_modules/cli-width": {
|
||||
"version": "4.1.0",
|
||||
"resolved": "https://registry.npmjs.org/cli-width/-/cli-width-4.1.0.tgz",
|
||||
@ -5023,18 +4884,6 @@
|
||||
"node": ">=4"
|
||||
}
|
||||
},
|
||||
"node_modules/environment": {
|
||||
"version": "1.1.0",
|
||||
"resolved": "https://registry.npmjs.org/environment/-/environment-1.1.0.tgz",
|
||||
"integrity": "sha512-xUtoPkMggbz0MPyPiIWr1Kp4aeWJjDZ6SMvURhimjdZgsRuDplF5/s9hcgGhyXMhs+6vpnuoiZ2kFiu3FMnS8Q==",
|
||||
"license": "MIT",
|
||||
"engines": {
|
||||
"node": ">=18"
|
||||
},
|
||||
"funding": {
|
||||
"url": "https://github.com/sponsors/sindresorhus"
|
||||
}
|
||||
},
|
||||
"node_modules/error-ex": {
|
||||
"version": "1.3.4",
|
||||
"resolved": "https://registry.npmjs.org/error-ex/-/error-ex-1.3.4.tgz",
|
||||
@ -6262,15 +6111,6 @@
|
||||
"node": ">= 0.4"
|
||||
}
|
||||
},
|
||||
"node_modules/highlight.js": {
|
||||
"version": "10.7.3",
|
||||
"resolved": "https://registry.npmjs.org/highlight.js/-/highlight.js-10.7.3.tgz",
|
||||
"integrity": "sha512-tzcUFauisWKNHaRkN4Wjl/ZA07gENAjFl3J/c480dprkGTg5EQstgaNFqBfUqCq54kZRIEcreTsAgF/m2quD7A==",
|
||||
"license": "BSD-3-Clause",
|
||||
"engines": {
|
||||
"node": "*"
|
||||
}
|
||||
},
|
||||
"node_modules/hmac-drbg": {
|
||||
"version": "1.0.1",
|
||||
"resolved": "https://registry.npmjs.org/hmac-drbg/-/hmac-drbg-1.0.1.tgz",
|
||||
@ -7837,39 +7677,6 @@
|
||||
"node": ">= 18"
|
||||
}
|
||||
},
|
||||
"node_modules/marked-terminal": {
|
||||
"version": "7.2.1",
|
||||
"resolved": "https://registry.npmjs.org/marked-terminal/-/marked-terminal-7.2.1.tgz",
|
||||
"integrity": "sha512-rQ1MoMFXZICWNsKMiiHwP/Z+92PLKskTPXj+e7uwXmuMPkNn7iTqC+IvDekVm1MPeC9wYQeLxeFaOvudRR/XbQ==",
|
||||
"license": "MIT",
|
||||
"dependencies": {
|
||||
"ansi-escapes": "^7.0.0",
|
||||
"ansi-regex": "^6.1.0",
|
||||
"chalk": "^5.3.0",
|
||||
"cli-highlight": "^2.1.11",
|
||||
"cli-table3": "^0.6.5",
|
||||
"node-emoji": "^2.1.3",
|
||||
"supports-hyperlinks": "^3.1.0"
|
||||
},
|
||||
"engines": {
|
||||
"node": ">=16.0.0"
|
||||
},
|
||||
"peerDependencies": {
|
||||
"marked": ">=1 <15"
|
||||
}
|
||||
},
|
||||
"node_modules/marked-terminal/node_modules/chalk": {
|
||||
"version": "5.6.2",
|
||||
"resolved": "https://registry.npmjs.org/chalk/-/chalk-5.6.2.tgz",
|
||||
"integrity": "sha512-7NzBL0rN6fMUW+f7A6Io4h40qQlG+xGmtMxfbnH/K7TAtt8JQWVQK+6g0UXKMeVJoyV5EkkNsErQ8pVD3bLHbA==",
|
||||
"license": "MIT",
|
||||
"engines": {
|
||||
"node": "^12.17.0 || ^14.13 || >=16.0.0"
|
||||
},
|
||||
"funding": {
|
||||
"url": "https://github.com/chalk/chalk?sponsor=1"
|
||||
}
|
||||
},
|
||||
"node_modules/math-intrinsics": {
|
||||
"version": "1.1.0",
|
||||
"resolved": "https://registry.npmjs.org/math-intrinsics/-/math-intrinsics-1.1.0.tgz",
|
||||
@ -8595,17 +8402,6 @@
|
||||
"node": "^18.17.0 || >=20.5.0"
|
||||
}
|
||||
},
|
||||
"node_modules/mz": {
|
||||
"version": "2.7.0",
|
||||
"resolved": "https://registry.npmjs.org/mz/-/mz-2.7.0.tgz",
|
||||
"integrity": "sha512-z81GNO7nnYMEhrGh9LeymoE4+Yr0Wn5McHIZMK5cfQCl+NDX08sCZgUc9/6MHni9IWuFLm1Z3HTCXu2z9fN62Q==",
|
||||
"license": "MIT",
|
||||
"dependencies": {
|
||||
"any-promise": "^1.0.0",
|
||||
"object-assign": "^4.0.1",
|
||||
"thenify-all": "^1.0.0"
|
||||
}
|
||||
},
|
||||
"node_modules/nanoid": {
|
||||
"version": "3.3.11",
|
||||
"resolved": "https://registry.npmjs.org/nanoid/-/nanoid-3.3.11.tgz",
|
||||
@ -8852,27 +8648,6 @@
|
||||
"lodash.assign": "^4.0.6"
|
||||
}
|
||||
},
|
||||
"node_modules/node-emoji": {
|
||||
"version": "2.2.0",
|
||||
"resolved": "https://registry.npmjs.org/node-emoji/-/node-emoji-2.2.0.tgz",
|
||||
"integrity": "sha512-Z3lTE9pLaJF47NyMhd4ww1yFTAP8YhYI8SleJiHzM46Fgpm5cnNzSl9XfzFNqbaz+VlJrIj3fXQ4DeN1Rjm6cw==",
|
||||
"license": "MIT",
|
||||
"dependencies": {
|
||||
"@sindresorhus/is": "^4.6.0",
|
||||
"char-regex": "^1.0.2",
|
||||
"emojilib": "^2.4.0",
|
||||
"skin-tone": "^2.0.0"
|
||||
},
|
||||
"engines": {
|
||||
"node": ">=18"
|
||||
}
|
||||
},
|
||||
"node_modules/node-emoji/node_modules/emojilib": {
|
||||
"version": "2.4.0",
|
||||
"resolved": "https://registry.npmjs.org/emojilib/-/emojilib-2.4.0.tgz",
|
||||
"integrity": "sha512-5U0rVMU5Y2n2+ykNLQqMoqklN9ICBT/KsvC1Gz6vqHbz2AXXGkG+Pm5rMWk/8Vjrr/mY9985Hi8DYzn1F09Nyw==",
|
||||
"license": "MIT"
|
||||
},
|
||||
"node_modules/node-fetch": {
|
||||
"version": "2.7.0",
|
||||
"resolved": "https://registry.npmjs.org/node-fetch/-/node-fetch-2.7.0.tgz",
|
||||
@ -8984,6 +8759,7 @@
|
||||
"version": "4.1.1",
|
||||
"resolved": "https://registry.npmjs.org/object-assign/-/object-assign-4.1.1.tgz",
|
||||
"integrity": "sha512-rJgTQnkUnH1sFw8yT6VSU3zD3sWmu6sZhIseY8VX+GRu3P6F7Fu+JNDoXfklElbLJSnc3FUQHVe4cU5hj+BcUg==",
|
||||
"dev": true,
|
||||
"license": "MIT",
|
||||
"engines": {
|
||||
"node": ">=0.10.0"
|
||||
@ -9251,27 +9027,6 @@
|
||||
"node": ">=0.10.0"
|
||||
}
|
||||
},
|
||||
"node_modules/parse5": {
|
||||
"version": "5.1.1",
|
||||
"resolved": "https://registry.npmjs.org/parse5/-/parse5-5.1.1.tgz",
|
||||
"integrity": "sha512-ugq4DFI0Ptb+WWjAdOK16+u/nHfiIrcE+sh8kZMaM0WllQKLI9rOUq6c2b7cwPkXdzfQESqvoqK6ug7U/Yyzug==",
|
||||
"license": "MIT"
|
||||
},
|
||||
"node_modules/parse5-htmlparser2-tree-adapter": {
|
||||
"version": "6.0.1",
|
||||
"resolved": "https://registry.npmjs.org/parse5-htmlparser2-tree-adapter/-/parse5-htmlparser2-tree-adapter-6.0.1.tgz",
|
||||
"integrity": "sha512-qPuWvbLgvDGilKc5BoicRovlT4MtYT6JfJyBOMDsKoiT+GiuP5qyrPCnR9HcPECIJJmZh5jRndyNThnhhb/vlA==",
|
||||
"license": "MIT",
|
||||
"dependencies": {
|
||||
"parse5": "^6.0.1"
|
||||
}
|
||||
},
|
||||
"node_modules/parse5-htmlparser2-tree-adapter/node_modules/parse5": {
|
||||
"version": "6.0.1",
|
||||
"resolved": "https://registry.npmjs.org/parse5/-/parse5-6.0.1.tgz",
|
||||
"integrity": "sha512-Ofn/CTFzRGTTxwpNEs9PP93gXShHcTq255nzRYSKe8AkVpZY7e1fpmTfOyoIvjP5HG7Z2ZM7VS9PPhQGW2pOpw==",
|
||||
"license": "MIT"
|
||||
},
|
||||
"node_modules/parse5-parser-stream": {
|
||||
"version": "7.1.2",
|
||||
"resolved": "https://registry.npmjs.org/parse5-parser-stream/-/parse5-parser-stream-7.1.2.tgz",
|
||||
@ -10737,18 +10492,6 @@
|
||||
"node": ">=18"
|
||||
}
|
||||
},
|
||||
"node_modules/skin-tone": {
|
||||
"version": "2.0.0",
|
||||
"resolved": "https://registry.npmjs.org/skin-tone/-/skin-tone-2.0.0.tgz",
|
||||
"integrity": "sha512-kUMbT1oBJCpgrnKoSr0o6wPtvRWT9W9UKvGLwfJYO2WuahZRHOpEyL1ckyMGgMWh0UdpmaoFqKKD29WTomNEGA==",
|
||||
"license": "MIT",
|
||||
"dependencies": {
|
||||
"unicode-emoji-modifier-base": "^1.0.0"
|
||||
},
|
||||
"engines": {
|
||||
"node": ">=8"
|
||||
}
|
||||
},
|
||||
"node_modules/source-map": {
|
||||
"version": "0.6.1",
|
||||
"resolved": "https://registry.npmjs.org/source-map/-/source-map-0.6.1.tgz",
|
||||
@ -11015,34 +10758,6 @@
|
||||
"url": "https://github.com/chalk/supports-color?sponsor=1"
|
||||
}
|
||||
},
|
||||
"node_modules/supports-hyperlinks": {
|
||||
"version": "3.2.0",
|
||||
"resolved": "https://registry.npmjs.org/supports-hyperlinks/-/supports-hyperlinks-3.2.0.tgz",
|
||||
"integrity": "sha512-zFObLMyZeEwzAoKCyu1B91U79K2t7ApXuQfo8OuxwXLDgcKxuwM+YvcbIhm6QWqz7mHUH1TVytR1PwVVjEuMig==",
|
||||
"license": "MIT",
|
||||
"dependencies": {
|
||||
"has-flag": "^4.0.0",
|
||||
"supports-color": "^7.0.0"
|
||||
},
|
||||
"engines": {
|
||||
"node": ">=14.18"
|
||||
},
|
||||
"funding": {
|
||||
"url": "https://github.com/chalk/supports-hyperlinks?sponsor=1"
|
||||
}
|
||||
},
|
||||
"node_modules/supports-hyperlinks/node_modules/supports-color": {
|
||||
"version": "7.2.0",
|
||||
"resolved": "https://registry.npmjs.org/supports-color/-/supports-color-7.2.0.tgz",
|
||||
"integrity": "sha512-qpCAvRl9stuOHveKsn7HncJRvv501qIacKzQlO/+Lwxc9+0q2wLyv4Dfvt80/DPn2pqOBsJdDiogXGR9+OvwRw==",
|
||||
"license": "MIT",
|
||||
"dependencies": {
|
||||
"has-flag": "^4.0.0"
|
||||
},
|
||||
"engines": {
|
||||
"node": ">=8"
|
||||
}
|
||||
},
|
||||
"node_modules/supports-preserve-symlinks-flag": {
|
||||
"version": "1.0.0",
|
||||
"resolved": "https://registry.npmjs.org/supports-preserve-symlinks-flag/-/supports-preserve-symlinks-flag-1.0.0.tgz",
|
||||
@ -11254,27 +10969,6 @@
|
||||
"dev": true,
|
||||
"license": "MIT"
|
||||
},
|
||||
"node_modules/thenify": {
|
||||
"version": "3.3.1",
|
||||
"resolved": "https://registry.npmjs.org/thenify/-/thenify-3.3.1.tgz",
|
||||
"integrity": "sha512-RVZSIV5IG10Hk3enotrhvz0T9em6cyHBLkH/YAZuKqd8hRkKhSfCGIcP2KUY0EPxndzANBmNllzWPwak+bheSw==",
|
||||
"license": "MIT",
|
||||
"dependencies": {
|
||||
"any-promise": "^1.0.0"
|
||||
}
|
||||
},
|
||||
"node_modules/thenify-all": {
|
||||
"version": "1.6.0",
|
||||
"resolved": "https://registry.npmjs.org/thenify-all/-/thenify-all-1.6.0.tgz",
|
||||
"integrity": "sha512-RNxQH/qI8/t3thXJDwcstUO4zeqo64+Uy/+sNVRBx4Xn2OX+OZ9oP+iJnNFqplFra2ZUVeKCSa2oVWi3T4uVmA==",
|
||||
"license": "MIT",
|
||||
"dependencies": {
|
||||
"thenify": ">= 3.1.0 < 4"
|
||||
},
|
||||
"engines": {
|
||||
"node": ">=0.8"
|
||||
}
|
||||
},
|
||||
"node_modules/through": {
|
||||
"version": "2.3.8",
|
||||
"resolved": "https://registry.npmjs.org/through/-/through-2.3.8.tgz",
|
||||
@ -11829,15 +11523,6 @@
|
||||
"devOptional": true,
|
||||
"license": "MIT"
|
||||
},
|
||||
"node_modules/unicode-emoji-modifier-base": {
|
||||
"version": "1.0.0",
|
||||
"resolved": "https://registry.npmjs.org/unicode-emoji-modifier-base/-/unicode-emoji-modifier-base-1.0.0.tgz",
|
||||
"integrity": "sha512-yLSH4py7oFH3oG/9K+XWrz1pSi3dfUrWEnInbxMfArOfc1+33BlGPQtLsOYwvdMy11AwUBetYuaRxSPqgkq+8g==",
|
||||
"license": "MIT",
|
||||
"engines": {
|
||||
"node": ">=4"
|
||||
}
|
||||
},
|
||||
"node_modules/unicorn-magic": {
|
||||
"version": "0.1.0",
|
||||
"resolved": "https://registry.npmjs.org/unicorn-magic/-/unicorn-magic-0.1.0.tgz",
|
||||
|
||||
@ -96,7 +96,6 @@
|
||||
"keyv": "^5.5.0",
|
||||
"keyv-file": "^5.1.3",
|
||||
"marked": "14.1.4",
|
||||
"marked-terminal": "7.2.1",
|
||||
"mime-types": "2.1.35",
|
||||
"nodemailer": "^7.0.13",
|
||||
"openai": "6.32.0",
|
||||
|
||||
@ -1,12 +1,6 @@
|
||||
import { marked } from 'marked'
|
||||
import { markedTerminal } from 'marked-terminal'
|
||||
import { examples as content } from '../docs-internal/examples.js'
|
||||
|
||||
export const examples = () => {
|
||||
marked.use(markedTerminal(
|
||||
{
|
||||
emoji: false,
|
||||
}
|
||||
))
|
||||
process.stdout.write(marked(content) as string)
|
||||
}
|
||||
|
||||
@ -1,6 +1,5 @@
|
||||
import { z } from 'zod'
|
||||
import { marked } from 'marked'
|
||||
import { markedTerminal } from 'marked-terminal'
|
||||
import { getDefaultValue, getDescription } from '@polymech/commons'
|
||||
|
||||
import { OptionsSchema } from '../zod_schema.js'
|
||||
@ -40,9 +39,6 @@ export const help = () => {
|
||||
md.push('')
|
||||
}
|
||||
|
||||
marked.use(markedTerminal({
|
||||
emoji: false
|
||||
}))
|
||||
const content: string = marked(md.join('\n')) as string;
|
||||
process.stdout.write(content)
|
||||
return content
|
||||
|
||||
@ -1,6 +1,5 @@
|
||||
import OpenAI from 'openai'
|
||||
import { marked } from 'marked'
|
||||
import { markedTerminal } from 'marked-terminal'
|
||||
import * as path from 'node:path'
|
||||
import { resolve } from '@polymech/commons'
|
||||
import { IKBotTask } from '../ai-tools/index.js'
|
||||
@ -25,9 +24,6 @@ export const onCompletion = async (result: any = "", options: IKBotTask) => {
|
||||
const action = options.append ? `${options.append} content to` : 'Wrote completion result to'
|
||||
options.logger?.debug(`${action} ${dstPath} : ${options.dst}`)
|
||||
} else {
|
||||
marked.use(markedTerminal({
|
||||
emoji: false,
|
||||
}))
|
||||
const content: string = marked(result) as string;
|
||||
process.stdout.write(content)
|
||||
}
|
||||
|
||||
@ -1,4 +1,5 @@
|
||||
export enum E_OPENROUTER_MODEL {
|
||||
MODEL_OPENROUTER_ELEPHANT_ALPHA = "openrouter/elephant-alpha",
|
||||
MODEL_ANTHROPIC_CLAUDE_OPUS_4_6_FAST = "anthropic/claude-opus-4.6-fast",
|
||||
MODEL_Z_AI_GLM_5_1 = "z-ai/glm-5.1",
|
||||
MODEL_GOOGLE_GEMMA_4_26B_A4B_IT_FREE = "google/gemma-4-26b-a4b-it:free",
|
||||
@ -142,7 +143,6 @@ export enum E_OPENROUTER_MODEL {
|
||||
MODEL_QWEN_QWEN3_NEXT_80B_A3B_THINKING = "qwen/qwen3-next-80b-a3b-thinking",
|
||||
MODEL_QWEN_QWEN3_NEXT_80B_A3B_INSTRUCT_FREE = "qwen/qwen3-next-80b-a3b-instruct:free",
|
||||
MODEL_QWEN_QWEN3_NEXT_80B_A3B_INSTRUCT = "qwen/qwen3-next-80b-a3b-instruct",
|
||||
MODEL_MEITUAN_LONGCAT_FLASH_CHAT = "meituan/longcat-flash-chat",
|
||||
MODEL_QWEN_QWEN_PLUS_2025_07_28_THINKING = "qwen/qwen-plus-2025-07-28:thinking",
|
||||
MODEL_QWEN_QWEN_PLUS_2025_07_28 = "qwen/qwen-plus-2025-07-28",
|
||||
MODEL_NVIDIA_NEMOTRON_NANO_9B_V2_FREE = "nvidia/nemotron-nano-9b-v2:free",
|
||||
@ -194,7 +194,6 @@ export enum E_OPENROUTER_MODEL {
|
||||
MODEL_MORPH_MORPH_V3_FAST = "morph/morph-v3-fast",
|
||||
MODEL_BAIDU_ERNIE_4_5_VL_424B_A47B = "baidu/ernie-4.5-vl-424b-a47b",
|
||||
MODEL_BAIDU_ERNIE_4_5_300B_A47B = "baidu/ernie-4.5-300b-a47b",
|
||||
MODEL_INCEPTION_MERCURY = "inception/mercury",
|
||||
MODEL_MISTRALAI_MISTRAL_SMALL_3_2_24B_INSTRUCT = "mistralai/mistral-small-3.2-24b-instruct",
|
||||
MODEL_MINIMAX_MINIMAX_M1 = "minimax/minimax-m1",
|
||||
MODEL_GOOGLE_GEMINI_2_5_FLASH = "google/gemini-2.5-flash",
|
||||
@ -214,7 +213,7 @@ export enum E_OPENROUTER_MODEL {
|
||||
MODEL_ARCEE_AI_MAESTRO_REASONING = "arcee-ai/maestro-reasoning",
|
||||
MODEL_ARCEE_AI_VIRTUOSO_LARGE = "arcee-ai/virtuoso-large",
|
||||
MODEL_ARCEE_AI_CODER_LARGE = "arcee-ai/coder-large",
|
||||
MODEL_INCEPTION_MERCURY_CODER = "inception/mercury-coder",
|
||||
MODEL_META_LLAMA_LLAMA_GUARD_4_12B_FREE = "meta-llama/llama-guard-4-12b:free",
|
||||
MODEL_META_LLAMA_LLAMA_GUARD_4_12B = "meta-llama/llama-guard-4-12b",
|
||||
MODEL_QWEN_QWEN3_30B_A3B = "qwen/qwen3-30b-a3b",
|
||||
MODEL_QWEN_QWEN3_8B = "qwen/qwen3-8b",
|
||||
@ -224,15 +223,12 @@ export enum E_OPENROUTER_MODEL {
|
||||
MODEL_OPENAI_O4_MINI_HIGH = "openai/o4-mini-high",
|
||||
MODEL_OPENAI_O3 = "openai/o3",
|
||||
MODEL_OPENAI_O4_MINI = "openai/o4-mini",
|
||||
MODEL_QWEN_QWEN2_5_CODER_7B_INSTRUCT = "qwen/qwen2.5-coder-7b-instruct",
|
||||
MODEL_OPENAI_GPT_4_1 = "openai/gpt-4.1",
|
||||
MODEL_OPENAI_GPT_4_1_MINI = "openai/gpt-4.1-mini",
|
||||
MODEL_OPENAI_GPT_4_1_NANO = "openai/gpt-4.1-nano",
|
||||
MODEL_ELEUTHERAI_LLEMMA_7B = "eleutherai/llemma_7b",
|
||||
MODEL_ALFREDPROS_CODELLAMA_7B_INSTRUCT_SOLIDITY = "alfredpros/codellama-7b-instruct-solidity",
|
||||
MODEL_X_AI_GROK_3_MINI_BETA = "x-ai/grok-3-mini-beta",
|
||||
MODEL_X_AI_GROK_3_BETA = "x-ai/grok-3-beta",
|
||||
MODEL_NVIDIA_LLAMA_3_1_NEMOTRON_ULTRA_253B_V1 = "nvidia/llama-3.1-nemotron-ultra-253b-v1",
|
||||
MODEL_META_LLAMA_LLAMA_4_MAVERICK = "meta-llama/llama-4-maverick",
|
||||
MODEL_META_LLAMA_LLAMA_4_SCOUT = "meta-llama/llama-4-scout",
|
||||
MODEL_QWEN_QWEN2_5_VL_32B_INSTRUCT = "qwen/qwen2.5-vl-32b-instruct",
|
||||
@ -321,7 +317,6 @@ export enum E_OPENROUTER_MODEL {
|
||||
MODEL_OPENAI_GPT_4O_MINI_2024_07_18 = "openai/gpt-4o-mini-2024-07-18",
|
||||
MODEL_OPENAI_GPT_4O_MINI = "openai/gpt-4o-mini",
|
||||
MODEL_GOOGLE_GEMMA_2_27B_IT = "google/gemma-2-27b-it",
|
||||
MODEL_GOOGLE_GEMMA_2_9B_IT = "google/gemma-2-9b-it",
|
||||
MODEL_SAO10K_L3_EURYALE_70B = "sao10k/l3-euryale-70b",
|
||||
MODEL_NOUSRESEARCH_HERMES_2_PRO_LLAMA_3_8B = "nousresearch/hermes-2-pro-llama-3-8b",
|
||||
MODEL_OPENAI_GPT_4O_2024_05_13 = "openai/gpt-4o-2024-05-13",
|
||||
|
||||
Loading…
Reference in New Issue
Block a user