forked from XiaoMo/ChatGPT-Next-Web
57 lines
1.2 KiB
TypeScript
57 lines
1.2 KiB
TypeScript
import { LLMModel } from "../client/api";
|
|
|
|
export function collectModelTable(
|
|
models: readonly LLMModel[],
|
|
customModels: string,
|
|
) {
|
|
const modelTable: Record<
|
|
string,
|
|
{ available: boolean; name: string; displayName: string }
|
|
> = {};
|
|
|
|
// default models
|
|
models.forEach(
|
|
(m) =>
|
|
(modelTable[m.name] = {
|
|
...m,
|
|
displayName: m.name,
|
|
}),
|
|
);
|
|
|
|
// server custom models
|
|
customModels
|
|
.split(",")
|
|
.filter((v) => !!v && v.length > 0)
|
|
.map((m) => {
|
|
const available = !m.startsWith("-");
|
|
const nameConfig =
|
|
m.startsWith("+") || m.startsWith("-") ? m.slice(1) : m;
|
|
const [name, displayName] = nameConfig.split("=");
|
|
|
|
// enable or disable all models
|
|
if (name === "all") {
|
|
Object.values(modelTable).forEach((m) => (m.available = available));
|
|
}
|
|
|
|
modelTable[name] = {
|
|
name,
|
|
displayName: displayName || name,
|
|
available,
|
|
};
|
|
});
|
|
return modelTable;
|
|
}
|
|
|
|
/**
|
|
* Generate full model table.
|
|
*/
|
|
export function collectModels(
|
|
models: readonly LLMModel[],
|
|
customModels: string,
|
|
) {
|
|
const modelTable = collectModelTable(models, customModels);
|
|
const allModels = Object.values(modelTable);
|
|
|
|
return allModels;
|
|
}
|