diff --git a/src/agents/academicSearchAgent.ts b/src/agents/academicSearchAgent.ts index a72e3a2..4a10c98 100644 --- a/src/agents/academicSearchAgent.ts +++ b/src/agents/academicSearchAgent.ts @@ -118,7 +118,6 @@ const createBasicAcademicSearchRetrieverChain = (llm: BaseChatModel) => { engines: [ 'arxiv', 'google scholar', - 'internetarchivescholar', 'pubmed', ], }); @@ -143,6 +142,7 @@ const createBasicAcademicSearchRetrieverChain = (llm: BaseChatModel) => { const createBasicAcademicSearchAnsweringChain = ( llm: BaseChatModel, embeddings: Embeddings, + optimizationMode: 'speed' | 'balanced' | 'quality', ) => { const basicAcademicSearchRetrieverChain = createBasicAcademicSearchRetrieverChain(llm); @@ -168,26 +168,33 @@ const createBasicAcademicSearchAnsweringChain = ( (doc) => doc.pageContent && doc.pageContent.length > 0, ); - const [docEmbeddings, queryEmbedding] = await Promise.all([ - embeddings.embedDocuments(docsWithContent.map((doc) => doc.pageContent)), - embeddings.embedQuery(query), - ]); + if (optimizationMode === 'speed') { + return docsWithContent.slice(0, 15); + } else if (optimizationMode === 'balanced') { + console.log('Balanced mode'); + const [docEmbeddings, queryEmbedding] = await Promise.all([ + embeddings.embedDocuments( + docsWithContent.map((doc) => doc.pageContent), + ), + embeddings.embedQuery(query), + ]); - const similarity = docEmbeddings.map((docEmbedding, i) => { - const sim = computeSimilarity(queryEmbedding, docEmbedding); + const similarity = docEmbeddings.map((docEmbedding, i) => { + const sim = computeSimilarity(queryEmbedding, docEmbedding); - return { - index: i, - similarity: sim, - }; - }); + return { + index: i, + similarity: sim, + }; + }); - const sortedDocs = similarity - .sort((a, b) => b.similarity - a.similarity) - .slice(0, 15) - .map((sim) => docsWithContent[sim.index]); + const sortedDocs = similarity + .sort((a, b) => b.similarity - a.similarity) + .slice(0, 15) + .map((sim) => docsWithContent[sim.index]); - return sortedDocs; + return sortedDocs; + } }; return RunnableSequence.from([ @@ -224,12 +231,17 @@ const basicAcademicSearch = ( history: BaseMessage[], llm: BaseChatModel, embeddings: Embeddings, + optimizationMode: 'speed' | 'balanced' | 'quality', ) => { const emitter = new eventEmitter(); try { const basicAcademicSearchAnsweringChain = - createBasicAcademicSearchAnsweringChain(llm, embeddings); + createBasicAcademicSearchAnsweringChain( + llm, + embeddings, + optimizationMode, + ); const stream = basicAcademicSearchAnsweringChain.streamEvents( { @@ -258,8 +270,15 @@ const handleAcademicSearch = ( history: BaseMessage[], llm: BaseChatModel, embeddings: Embeddings, + optimizationMode: 'speed' | 'balanced' | 'quality', ) => { - const emitter = basicAcademicSearch(message, history, llm, embeddings); + const emitter = basicAcademicSearch( + message, + history, + llm, + embeddings, + optimizationMode, + ); return emitter; }; diff --git a/src/agents/redditSearchAgent.ts b/src/agents/redditSearchAgent.ts index 9c2c443..2c44c13 100644 --- a/src/agents/redditSearchAgent.ts +++ b/src/agents/redditSearchAgent.ts @@ -138,6 +138,7 @@ const createBasicRedditSearchRetrieverChain = (llm: BaseChatModel) => { const createBasicRedditSearchAnsweringChain = ( llm: BaseChatModel, embeddings: Embeddings, + optimizationMode: 'speed' | 'balanced' | 'quality', ) => { const basicRedditSearchRetrieverChain = createBasicRedditSearchRetrieverChain(llm); @@ -163,27 +164,33 @@ const createBasicRedditSearchAnsweringChain = ( (doc) => doc.pageContent && doc.pageContent.length > 0, ); - const [docEmbeddings, queryEmbedding] = await Promise.all([ - embeddings.embedDocuments(docsWithContent.map((doc) => doc.pageContent)), - embeddings.embedQuery(query), - ]); + if (optimizationMode === 'speed') { + return docsWithContent.slice(0, 15); + } else if (optimizationMode === 'balanced') { + const [docEmbeddings, queryEmbedding] = await Promise.all([ + embeddings.embedDocuments( + docsWithContent.map((doc) => doc.pageContent), + ), + embeddings.embedQuery(query), + ]); - const similarity = docEmbeddings.map((docEmbedding, i) => { - const sim = computeSimilarity(queryEmbedding, docEmbedding); + const similarity = docEmbeddings.map((docEmbedding, i) => { + const sim = computeSimilarity(queryEmbedding, docEmbedding); - return { - index: i, - similarity: sim, - }; - }); + return { + index: i, + similarity: sim, + }; + }); - const sortedDocs = similarity - .filter((sim) => sim.similarity > 0.3) - .sort((a, b) => b.similarity - a.similarity) - .slice(0, 15) - .map((sim) => docsWithContent[sim.index]); + const sortedDocs = similarity + .filter((sim) => sim.similarity > 0.3) + .sort((a, b) => b.similarity - a.similarity) + .slice(0, 15) + .map((sim) => docsWithContent[sim.index]); - return sortedDocs; + return sortedDocs; + } }; return RunnableSequence.from([ @@ -220,12 +227,13 @@ const basicRedditSearch = ( history: BaseMessage[], llm: BaseChatModel, embeddings: Embeddings, + optimizationMode: 'speed' | 'balanced' | 'quality', ) => { const emitter = new eventEmitter(); try { const basicRedditSearchAnsweringChain = - createBasicRedditSearchAnsweringChain(llm, embeddings); + createBasicRedditSearchAnsweringChain(llm, embeddings, optimizationMode); const stream = basicRedditSearchAnsweringChain.streamEvents( { chat_history: history, @@ -253,8 +261,15 @@ const handleRedditSearch = ( history: BaseMessage[], llm: BaseChatModel, embeddings: Embeddings, + optimizationMode: 'speed' | 'balanced' | 'quality', ) => { - const emitter = basicRedditSearch(message, history, llm, embeddings); + const emitter = basicRedditSearch( + message, + history, + llm, + embeddings, + optimizationMode, + ); return emitter; }; diff --git a/src/agents/webSearchAgent.ts b/src/agents/webSearchAgent.ts index 77ec181..51653a0 100644 --- a/src/agents/webSearchAgent.ts +++ b/src/agents/webSearchAgent.ts @@ -216,12 +216,34 @@ const createBasicWebSearchRetrieverChain = (llm: BaseChatModel) => { await Promise.all( docGroups.map(async (doc) => { const res = await llm.invoke(` - You are a text summarizer. You need to summarize the text provided inside the \`text\` XML block. - You need to summarize the text into 1 or 2 sentences capturing the main idea of the text. - You need to make sure that you don't miss any point while summarizing the text. - You will also be given a \`query\` XML block which will contain the query of the user. Try to answer the query in the summary from the text provided. - If the query says Summarize then you just need to summarize the text without answering the query. - Only return the summarized text without any other messages, text or XML block. + You are a web search summarizer, tasked with summarizing a piece of text retrieved from a web search. Your job is to summarize the + text into a detailed, 2-4 paragraph explanation that captures the main ideas and provides a comprehensive answer to the query. + If the query is \"summarize\", you should provide a detailed summary of the text. If the query is a specific question, you should answer it in the summary. + + - **Journalistic tone**: The summary should sound professional and journalistic, not too casual or vague. + - **Thorough and detailed**: Ensure that every key point from the text is captured and that the summary directly answers the query. + - **Not too lengthy, but detailed**: The summary should be informative but not excessively long. Focus on providing detailed information in a concise format. + + The text will be shared inside the \`text\` XML tag, and the query inside the \`query\` XML tag. + + + + Docker is a set of platform-as-a-service products that use OS-level virtualization to deliver software in packages called containers. + It was first released in 2013 and is developed by Docker, Inc. Docker is designed to make it easier to create, deploy, and run applications + by using containers. + + + + What is Docker and how does it work? + + + Response: + Docker is a revolutionary platform-as-a-service product developed by Docker, Inc., that uses container technology to make application + deployment more efficient. It allows developers to package their software with all necessary dependencies, making it easier to run in + any environment. Released in 2013, Docker has transformed the way applications are built, deployed, and managed. + + + Everything below is the actual data you will be working with. Good luck! ${question} @@ -273,6 +295,7 @@ const createBasicWebSearchRetrieverChain = (llm: BaseChatModel) => { const createBasicWebSearchAnsweringChain = ( llm: BaseChatModel, embeddings: Embeddings, + optimizationMode: 'speed' | 'balanced' | 'quality', ) => { const basicWebSearchRetrieverChain = createBasicWebSearchRetrieverChain(llm); @@ -301,27 +324,33 @@ const createBasicWebSearchAnsweringChain = ( (doc) => doc.pageContent && doc.pageContent.length > 0, ); - const [docEmbeddings, queryEmbedding] = await Promise.all([ - embeddings.embedDocuments(docsWithContent.map((doc) => doc.pageContent)), - embeddings.embedQuery(query), - ]); + if (optimizationMode === 'speed') { + return docsWithContent.slice(0, 15); + } else if (optimizationMode === 'balanced') { + const [docEmbeddings, queryEmbedding] = await Promise.all([ + embeddings.embedDocuments( + docsWithContent.map((doc) => doc.pageContent), + ), + embeddings.embedQuery(query), + ]); - const similarity = docEmbeddings.map((docEmbedding, i) => { - const sim = computeSimilarity(queryEmbedding, docEmbedding); + const similarity = docEmbeddings.map((docEmbedding, i) => { + const sim = computeSimilarity(queryEmbedding, docEmbedding); - return { - index: i, - similarity: sim, - }; - }); + return { + index: i, + similarity: sim, + }; + }); - const sortedDocs = similarity - .filter((sim) => sim.similarity > 0.3) - .sort((a, b) => b.similarity - a.similarity) - .slice(0, 15) - .map((sim) => docsWithContent[sim.index]); + const sortedDocs = similarity + .filter((sim) => sim.similarity > 0.3) + .sort((a, b) => b.similarity - a.similarity) + .slice(0, 15) + .map((sim) => docsWithContent[sim.index]); - return sortedDocs; + return sortedDocs; + } }; return RunnableSequence.from([ @@ -358,6 +387,7 @@ const basicWebSearch = ( history: BaseMessage[], llm: BaseChatModel, embeddings: Embeddings, + optimizationMode: 'speed' | 'balanced' | 'quality', ) => { const emitter = new eventEmitter(); @@ -365,6 +395,7 @@ const basicWebSearch = ( const basicWebSearchAnsweringChain = createBasicWebSearchAnsweringChain( llm, embeddings, + optimizationMode, ); const stream = basicWebSearchAnsweringChain.streamEvents( @@ -394,8 +425,15 @@ const handleWebSearch = ( history: BaseMessage[], llm: BaseChatModel, embeddings: Embeddings, + optimizationMode: 'speed' | 'balanced' | 'quality', ) => { - const emitter = basicWebSearch(message, history, llm, embeddings); + const emitter = basicWebSearch( + message, + history, + llm, + embeddings, + optimizationMode, + ); return emitter; }; diff --git a/src/agents/youtubeSearchAgent.ts b/src/agents/youtubeSearchAgent.ts index e9b6553..2f53bc9 100644 --- a/src/agents/youtubeSearchAgent.ts +++ b/src/agents/youtubeSearchAgent.ts @@ -138,6 +138,7 @@ const createBasicYoutubeSearchRetrieverChain = (llm: BaseChatModel) => { const createBasicYoutubeSearchAnsweringChain = ( llm: BaseChatModel, embeddings: Embeddings, + optimizationMode: 'speed' | 'balanced' | 'quality', ) => { const basicYoutubeSearchRetrieverChain = createBasicYoutubeSearchRetrieverChain(llm); @@ -163,27 +164,33 @@ const createBasicYoutubeSearchAnsweringChain = ( (doc) => doc.pageContent && doc.pageContent.length > 0, ); - const [docEmbeddings, queryEmbedding] = await Promise.all([ - embeddings.embedDocuments(docsWithContent.map((doc) => doc.pageContent)), - embeddings.embedQuery(query), - ]); + if (optimizationMode === 'speed') { + return docsWithContent.slice(0, 15); + } else { + const [docEmbeddings, queryEmbedding] = await Promise.all([ + embeddings.embedDocuments( + docsWithContent.map((doc) => doc.pageContent), + ), + embeddings.embedQuery(query), + ]); - const similarity = docEmbeddings.map((docEmbedding, i) => { - const sim = computeSimilarity(queryEmbedding, docEmbedding); + const similarity = docEmbeddings.map((docEmbedding, i) => { + const sim = computeSimilarity(queryEmbedding, docEmbedding); - return { - index: i, - similarity: sim, - }; - }); + return { + index: i, + similarity: sim, + }; + }); - const sortedDocs = similarity - .filter((sim) => sim.similarity > 0.3) - .sort((a, b) => b.similarity - a.similarity) - .slice(0, 15) - .map((sim) => docsWithContent[sim.index]); + const sortedDocs = similarity + .filter((sim) => sim.similarity > 0.3) + .sort((a, b) => b.similarity - a.similarity) + .slice(0, 15) + .map((sim) => docsWithContent[sim.index]); - return sortedDocs; + return sortedDocs; + } }; return RunnableSequence.from([ @@ -220,12 +227,13 @@ const basicYoutubeSearch = ( history: BaseMessage[], llm: BaseChatModel, embeddings: Embeddings, + optimizationMode: 'speed' | 'balanced' | 'quality', ) => { const emitter = new eventEmitter(); try { const basicYoutubeSearchAnsweringChain = - createBasicYoutubeSearchAnsweringChain(llm, embeddings); + createBasicYoutubeSearchAnsweringChain(llm, embeddings, optimizationMode); const stream = basicYoutubeSearchAnsweringChain.streamEvents( { @@ -254,8 +262,15 @@ const handleYoutubeSearch = ( history: BaseMessage[], llm: BaseChatModel, embeddings: Embeddings, + optimizationMode: 'speed' | 'balanced' | 'quality', ) => { - const emitter = basicYoutubeSearch(message, history, llm, embeddings); + const emitter = basicYoutubeSearch( + message, + history, + llm, + embeddings, + optimizationMode, + ); return emitter; }; diff --git a/src/websocket/messageHandler.ts b/src/websocket/messageHandler.ts index 332910c..d230386 100644 --- a/src/websocket/messageHandler.ts +++ b/src/websocket/messageHandler.ts @@ -22,7 +22,7 @@ type Message = { type WSMessage = { message: Message; - copilot: boolean; + optimizationMode: string; type: string; focusMode: string; history: Array<[string, string]>; @@ -138,6 +138,7 @@ export const handleMessage = async ( history, llm, embeddings, + parsedWSMessage.optimizationMode, ); handleEmitterEvents(emitter, ws, id, parsedMessage.chatId); diff --git a/ui/components/ChatWindow.tsx b/ui/components/ChatWindow.tsx index b3d0089..b67ca3a 100644 --- a/ui/components/ChatWindow.tsx +++ b/ui/components/ChatWindow.tsx @@ -278,6 +278,7 @@ const ChatWindow = ({ id }: { id?: string }) => { const [messages, setMessages] = useState([]); const [focusMode, setFocusMode] = useState('webSearch'); + const [optimizationMode, setOptimizationMode] = useState('speed'); const [isMessagesLoaded, setIsMessagesLoaded] = useState(false); @@ -346,6 +347,7 @@ const ChatWindow = ({ id }: { id?: string }) => { content: message, }, focusMode: focusMode, + optimizationMode: optimizationMode, history: [...chatHistory, ['human', message]], }), ); @@ -508,6 +510,8 @@ const ChatWindow = ({ id }: { id?: string }) => { sendMessage={sendMessage} focusMode={focusMode} setFocusMode={setFocusMode} + optimizationMode={optimizationMode} + setOptimizationMode={setOptimizationMode} /> )} diff --git a/ui/components/EmptyChat.tsx b/ui/components/EmptyChat.tsx index ea3642b..63f186c 100644 --- a/ui/components/EmptyChat.tsx +++ b/ui/components/EmptyChat.tsx @@ -4,10 +4,14 @@ const EmptyChat = ({ sendMessage, focusMode, setFocusMode, + optimizationMode, + setOptimizationMode, }: { sendMessage: (message: string) => void; focusMode: string; setFocusMode: (mode: string) => void; + optimizationMode: string; + setOptimizationMode: (mode: string) => void; }) => { return (
@@ -19,6 +23,8 @@ const EmptyChat = ({ sendMessage={sendMessage} focusMode={focusMode} setFocusMode={setFocusMode} + optimizationMode={optimizationMode} + setOptimizationMode={setOptimizationMode} />
diff --git a/ui/components/EmptyChatMessageInput.tsx b/ui/components/EmptyChatMessageInput.tsx index 39d3f16..845bed9 100644 --- a/ui/components/EmptyChatMessageInput.tsx +++ b/ui/components/EmptyChatMessageInput.tsx @@ -3,15 +3,20 @@ import { useEffect, useRef, useState } from 'react'; import TextareaAutosize from 'react-textarea-autosize'; import CopilotToggle from './MessageInputActions/Copilot'; import Focus from './MessageInputActions/Focus'; +import Optimization from './MessageInputActions/Optimization'; const EmptyChatMessageInput = ({ sendMessage, focusMode, setFocusMode, + optimizationMode, + setOptimizationMode, }: { sendMessage: (message: string) => void; focusMode: string; setFocusMode: (mode: string) => void; + optimizationMode: string; + setOptimizationMode: (mode: string) => void; }) => { const [copilotEnabled, setCopilotEnabled] = useState(false); const [message, setMessage] = useState(''); @@ -66,14 +71,13 @@ const EmptyChatMessageInput = ({ placeholder="Ask anything..." />
-
+
- {/* */}
-
- +