From 5924690df297de67ee5c114b338e29bfd9a68bd4 Mon Sep 17 00:00:00 2001 From: ItzCrazyKns Date: Sat, 20 Apr 2024 22:12:07 +0530 Subject: [PATCH] feat(image-search): Use LLM from config --- src/routes/images.ts | 23 +++++++++++++++++------ 1 file changed, 17 insertions(+), 6 deletions(-) diff --git a/src/routes/images.ts b/src/routes/images.ts index dd3925f..97f0b31 100644 --- a/src/routes/images.ts +++ b/src/routes/images.ts @@ -1,7 +1,8 @@ import express from 'express'; import handleImageSearch from '../agents/imageSearchAgent'; -import { ChatOpenAI } from '@langchain/openai'; -import { getOpenaiApiKey } from '../config'; +import { BaseChatModel } from '@langchain/core/language_models/chat_models'; +import { getAvailableProviders } from '../lib/providers'; +import { getChatModel, getChatModelProvider } from '../config'; const router = express.Router(); @@ -9,10 +10,20 @@ router.post('/', async (req, res) => { try { const { query, chat_history } = req.body; - const llm = new ChatOpenAI({ - temperature: 0.7, - openAIApiKey: getOpenaiApiKey(), - }); + const models = await getAvailableProviders(); + const provider = getChatModelProvider(); + const chatModel = getChatModel(); + + let llm: BaseChatModel | undefined; + + if (models[provider] && models[provider][chatModel]) { + llm = models[provider][chatModel] as BaseChatModel | undefined; + } + + if (!llm) { + res.status(500).json({ message: 'Invalid LLM model selected' }); + return; + } const images = await handleImageSearch({ query, chat_history }, llm);