59 lines
1.4 KiB
TypeScript
59 lines
1.4 KiB
TypeScript
import { ChatOpenAI, OpenAIEmbeddings } from '@langchain/openai';
|
|
import { getOpenaiApiKey } from '../../config';
|
|
import logger from '../../utils/logger';
|
|
|
|
export const loadOpenAIChatModels = async () => {
|
|
const openAIApiKey = getOpenaiApiKey();
|
|
|
|
try {
|
|
const chatModels = {
|
|
'GPT-3.5 turbo': new ChatOpenAI({
|
|
openAIApiKey,
|
|
modelName: 'gpt-3.5-turbo',
|
|
temperature: 0.7,
|
|
}),
|
|
'GPT-4': new ChatOpenAI({
|
|
openAIApiKey,
|
|
modelName: 'gpt-4',
|
|
temperature: 0.7,
|
|
}),
|
|
'GPT-4 turbo': new ChatOpenAI({
|
|
openAIApiKey,
|
|
modelName: 'gpt-4-turbo',
|
|
temperature: 0.7,
|
|
}),
|
|
'GPT-4 omni': new ChatOpenAI({
|
|
openAIApiKey,
|
|
modelName: 'gpt-4o',
|
|
temperature: 0.7,
|
|
}),
|
|
};
|
|
|
|
return chatModels;
|
|
} catch (err) {
|
|
logger.error(`Error loading OpenAI models: ${err}`);
|
|
return {};
|
|
}
|
|
};
|
|
|
|
export const loadOpenAIEmbeddingsModel = async () => {
|
|
const openAIApiKey = getOpenaiApiKey();
|
|
|
|
try {
|
|
const embeddingModels = {
|
|
'Text embedding 3 small': new OpenAIEmbeddings({
|
|
openAIApiKey,
|
|
modelName: 'text-embedding-3-small',
|
|
}),
|
|
'Text embedding 3 large': new OpenAIEmbeddings({
|
|
openAIApiKey,
|
|
modelName: 'text-embedding-3-large',
|
|
}),
|
|
};
|
|
|
|
return embeddingModels;
|
|
} catch (err) {
|
|
logger.error(`Error loading OpenAI embeddings model: ${err}`);
|
|
return {};
|
|
}
|
|
};
|