Merge remote-tracking branch 'origin/master' into ollama-auth
This commit is contained in:
commit
4447cf7b5f
|
@ -1,6 +1,6 @@
|
||||||
{
|
{
|
||||||
"name": "perplexica-backend",
|
"name": "perplexica-backend",
|
||||||
"version": "1.9.0-rc1",
|
"version": "1.9.0-rc2",
|
||||||
"license": "MIT",
|
"license": "MIT",
|
||||||
"author": "ItzCrazyKns",
|
"author": "ItzCrazyKns",
|
||||||
"scripts": {
|
"scripts": {
|
||||||
|
|
2345
searxng/settings.yml
2345
searxng/settings.yml
File diff suppressed because it is too large
Load Diff
|
@ -19,6 +19,7 @@ import formatChatHistoryAsString from '../utils/formatHistory';
|
||||||
import eventEmitter from 'events';
|
import eventEmitter from 'events';
|
||||||
import computeSimilarity from '../utils/computeSimilarity';
|
import computeSimilarity from '../utils/computeSimilarity';
|
||||||
import logger from '../utils/logger';
|
import logger from '../utils/logger';
|
||||||
|
import { IterableReadableStream } from '@langchain/core/utils/stream';
|
||||||
|
|
||||||
const basicAcademicSearchRetrieverPrompt = `
|
const basicAcademicSearchRetrieverPrompt = `
|
||||||
You will be given a conversation below and a follow up question. You need to rephrase the follow-up question if needed so it is a standalone question that can be used by the LLM to search the web for information.
|
You will be given a conversation below and a follow up question. You need to rephrase the follow-up question if needed so it is a standalone question that can be used by the LLM to search the web for information.
|
||||||
|
@ -66,7 +67,7 @@ const basicAcademicSearchResponsePrompt = `
|
||||||
const strParser = new StringOutputParser();
|
const strParser = new StringOutputParser();
|
||||||
|
|
||||||
const handleStream = async (
|
const handleStream = async (
|
||||||
stream: AsyncGenerator<StreamEvent, any, unknown>,
|
stream: IterableReadableStream<StreamEvent>,
|
||||||
emitter: eventEmitter,
|
emitter: eventEmitter,
|
||||||
) => {
|
) => {
|
||||||
for await (const event of stream) {
|
for await (const event of stream) {
|
||||||
|
|
|
@ -19,6 +19,7 @@ import formatChatHistoryAsString from '../utils/formatHistory';
|
||||||
import eventEmitter from 'events';
|
import eventEmitter from 'events';
|
||||||
import computeSimilarity from '../utils/computeSimilarity';
|
import computeSimilarity from '../utils/computeSimilarity';
|
||||||
import logger from '../utils/logger';
|
import logger from '../utils/logger';
|
||||||
|
import { IterableReadableStream } from '@langchain/core/utils/stream';
|
||||||
|
|
||||||
const basicRedditSearchRetrieverPrompt = `
|
const basicRedditSearchRetrieverPrompt = `
|
||||||
You will be given a conversation below and a follow up question. You need to rephrase the follow-up question if needed so it is a standalone question that can be used by the LLM to search the web for information.
|
You will be given a conversation below and a follow up question. You need to rephrase the follow-up question if needed so it is a standalone question that can be used by the LLM to search the web for information.
|
||||||
|
@ -66,7 +67,7 @@ const basicRedditSearchResponsePrompt = `
|
||||||
const strParser = new StringOutputParser();
|
const strParser = new StringOutputParser();
|
||||||
|
|
||||||
const handleStream = async (
|
const handleStream = async (
|
||||||
stream: AsyncGenerator<StreamEvent, any, unknown>,
|
stream: IterableReadableStream<StreamEvent>,
|
||||||
emitter: eventEmitter,
|
emitter: eventEmitter,
|
||||||
) => {
|
) => {
|
||||||
for await (const event of stream) {
|
for await (const event of stream) {
|
||||||
|
|
|
@ -22,22 +22,38 @@ import logger from '../utils/logger';
|
||||||
import LineListOutputParser from '../lib/outputParsers/listLineOutputParser';
|
import LineListOutputParser from '../lib/outputParsers/listLineOutputParser';
|
||||||
import { getDocumentsFromLinks } from '../lib/linkDocument';
|
import { getDocumentsFromLinks } from '../lib/linkDocument';
|
||||||
import LineOutputParser from '../lib/outputParsers/lineOutputParser';
|
import LineOutputParser from '../lib/outputParsers/lineOutputParser';
|
||||||
|
import { IterableReadableStream } from '@langchain/core/utils/stream';
|
||||||
|
import { ChatOpenAI } from '@langchain/openai';
|
||||||
|
|
||||||
const basicSearchRetrieverPrompt = `
|
const basicSearchRetrieverPrompt = `
|
||||||
You will be given a conversation below and a follow up question. You need to rephrase the follow-up question if needed so it is a standalone question that can be used by the LLM to search the web for information.
|
You are an AI question rephraser. You will be given a conversation and a follow-up question, you will have to rephrase the follow up question so it is a standalone question and can be used by another LLM to search the web for information to answer it.
|
||||||
If it is a writing task or a simple hi, hello rather than a question, you need to return \`not_needed\` as the response.
|
If it is a smple writing task or a greeting (unless the greeting contains a question after it) like Hi, Hello, How are you, etc. than a question then you need to return \`not_needed\` as the response (This is because the LLM won't need to search the web for finding information on this topic).
|
||||||
If the question contains some links and asks to answer from those links or even if they don't you need to return the links inside 'links' XML block and the question inside 'question' XML block. If there are no links then you need to return the question without any XML block.
|
If the user asks some question from some URL or wants you to summarize a PDF or a webpage (via URL) you need to return the links inside the \`links\` XML block and the question inside the \`question\` XML block. If the user wants to you to summarize the webpage or the PDF you need to return \`summarize\` inside the \`question\` XML block in place of a question and the link to summarize in the \`links\` XML block.
|
||||||
If the user asks to summarrize the content from some links you need to return \`Summarize\` as the question inside the 'question' XML block and the links inside the 'links' XML block.
|
You must always return the rephrased question inside the \`question\` XML block, if there are no links in the follow-up question then don't insert a \`links\` XML block in your response.
|
||||||
|
|
||||||
Example:
|
There are several examples attached for your reference inside the below \`examples\` XML block
|
||||||
1. Follow up question: What is the capital of France?
|
|
||||||
Rephrased question: \`Capital of france\`
|
|
||||||
|
|
||||||
2. Follow up question: What is the population of New York City?
|
<examples>
|
||||||
Rephrased question: \`Population of New York City\`
|
1. Follow up question: What is the capital of France
|
||||||
|
Rephrased question:\`
|
||||||
|
<question>
|
||||||
|
Capital of france
|
||||||
|
</question>
|
||||||
|
\`
|
||||||
|
|
||||||
|
2. Hi, how are you?
|
||||||
|
Rephrased question\`
|
||||||
|
<question>
|
||||||
|
not_needed
|
||||||
|
</question>
|
||||||
|
\`
|
||||||
|
|
||||||
3. Follow up question: What is Docker?
|
3. Follow up question: What is Docker?
|
||||||
Rephrased question: \`What is Docker\`
|
Rephrased question: \`
|
||||||
|
<question>
|
||||||
|
What is Docker
|
||||||
|
</question>
|
||||||
|
\`
|
||||||
|
|
||||||
4. Follow up question: Can you tell me what is X from https://example.com
|
4. Follow up question: Can you tell me what is X from https://example.com
|
||||||
Rephrased question: \`
|
Rephrased question: \`
|
||||||
|
@ -53,16 +69,20 @@ https://example.com
|
||||||
5. Follow up question: Summarize the content from https://example.com
|
5. Follow up question: Summarize the content from https://example.com
|
||||||
Rephrased question: \`
|
Rephrased question: \`
|
||||||
<question>
|
<question>
|
||||||
Summarize
|
summarize
|
||||||
</question>
|
</question>
|
||||||
|
|
||||||
<links>
|
<links>
|
||||||
https://example.com
|
https://example.com
|
||||||
</links>
|
</links>
|
||||||
\`
|
\`
|
||||||
|
</examples>
|
||||||
|
|
||||||
Conversation:
|
Anything below is the part of the actual conversation and you need to use conversation and the follow-up question to rephrase the follow-up question as a standalone question based on the guidelines shared above.
|
||||||
|
|
||||||
|
<conversation>
|
||||||
{chat_history}
|
{chat_history}
|
||||||
|
</conversation>
|
||||||
|
|
||||||
Follow up question: {query}
|
Follow up question: {query}
|
||||||
Rephrased question:
|
Rephrased question:
|
||||||
|
@ -95,7 +115,7 @@ const basicWebSearchResponsePrompt = `
|
||||||
const strParser = new StringOutputParser();
|
const strParser = new StringOutputParser();
|
||||||
|
|
||||||
const handleStream = async (
|
const handleStream = async (
|
||||||
stream: AsyncGenerator<StreamEvent, any, unknown>,
|
stream: IterableReadableStream<StreamEvent>,
|
||||||
emitter: eventEmitter,
|
emitter: eventEmitter,
|
||||||
) => {
|
) => {
|
||||||
for await (const event of stream) {
|
for await (const event of stream) {
|
||||||
|
@ -132,15 +152,13 @@ type BasicChainInput = {
|
||||||
};
|
};
|
||||||
|
|
||||||
const createBasicWebSearchRetrieverChain = (llm: BaseChatModel) => {
|
const createBasicWebSearchRetrieverChain = (llm: BaseChatModel) => {
|
||||||
|
(llm as unknown as ChatOpenAI).temperature = 0;
|
||||||
|
|
||||||
return RunnableSequence.from([
|
return RunnableSequence.from([
|
||||||
PromptTemplate.fromTemplate(basicSearchRetrieverPrompt),
|
PromptTemplate.fromTemplate(basicSearchRetrieverPrompt),
|
||||||
llm,
|
llm,
|
||||||
strParser,
|
strParser,
|
||||||
RunnableLambda.from(async (input: string) => {
|
RunnableLambda.from(async (input: string) => {
|
||||||
if (input === 'not_needed') {
|
|
||||||
return { query: '', docs: [] };
|
|
||||||
}
|
|
||||||
|
|
||||||
const linksOutputParser = new LineListOutputParser({
|
const linksOutputParser = new LineListOutputParser({
|
||||||
key: 'links',
|
key: 'links',
|
||||||
});
|
});
|
||||||
|
@ -152,9 +170,13 @@ const createBasicWebSearchRetrieverChain = (llm: BaseChatModel) => {
|
||||||
const links = await linksOutputParser.parse(input);
|
const links = await linksOutputParser.parse(input);
|
||||||
let question = await questionOutputParser.parse(input);
|
let question = await questionOutputParser.parse(input);
|
||||||
|
|
||||||
|
if (question === 'not_needed') {
|
||||||
|
return { query: '', docs: [] };
|
||||||
|
}
|
||||||
|
|
||||||
if (links.length > 0) {
|
if (links.length > 0) {
|
||||||
if (question.length === 0) {
|
if (question.length === 0) {
|
||||||
question = 'Summarize';
|
question = 'summarize';
|
||||||
}
|
}
|
||||||
|
|
||||||
let docs = [];
|
let docs = [];
|
||||||
|
@ -226,7 +248,7 @@ const createBasicWebSearchRetrieverChain = (llm: BaseChatModel) => {
|
||||||
|
|
||||||
return { query: question, docs: docs };
|
return { query: question, docs: docs };
|
||||||
} else {
|
} else {
|
||||||
const res = await searchSearxng(input, {
|
const res = await searchSearxng(question, {
|
||||||
language: 'en',
|
language: 'en',
|
||||||
});
|
});
|
||||||
|
|
||||||
|
@ -242,7 +264,7 @@ const createBasicWebSearchRetrieverChain = (llm: BaseChatModel) => {
|
||||||
}),
|
}),
|
||||||
);
|
);
|
||||||
|
|
||||||
return { query: input, docs: documents };
|
return { query: question, docs: documents };
|
||||||
}
|
}
|
||||||
}),
|
}),
|
||||||
]);
|
]);
|
||||||
|
@ -271,7 +293,7 @@ const createBasicWebSearchAnsweringChain = (
|
||||||
return docs;
|
return docs;
|
||||||
}
|
}
|
||||||
|
|
||||||
if (query === 'Summarize') {
|
if (query.toLocaleLowerCase() === 'summarize') {
|
||||||
return docs;
|
return docs;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -294,7 +316,7 @@ const createBasicWebSearchAnsweringChain = (
|
||||||
});
|
});
|
||||||
|
|
||||||
const sortedDocs = similarity
|
const sortedDocs = similarity
|
||||||
.filter((sim) => sim.similarity > 0.5)
|
.filter((sim) => sim.similarity > 0.3)
|
||||||
.sort((a, b) => b.similarity - a.similarity)
|
.sort((a, b) => b.similarity - a.similarity)
|
||||||
.slice(0, 15)
|
.slice(0, 15)
|
||||||
.map((sim) => docsWithContent[sim.index]);
|
.map((sim) => docsWithContent[sim.index]);
|
||||||
|
|
|
@ -18,6 +18,7 @@ import type { Embeddings } from '@langchain/core/embeddings';
|
||||||
import formatChatHistoryAsString from '../utils/formatHistory';
|
import formatChatHistoryAsString from '../utils/formatHistory';
|
||||||
import eventEmitter from 'events';
|
import eventEmitter from 'events';
|
||||||
import logger from '../utils/logger';
|
import logger from '../utils/logger';
|
||||||
|
import { IterableReadableStream } from '@langchain/core/utils/stream';
|
||||||
|
|
||||||
const basicWolframAlphaSearchRetrieverPrompt = `
|
const basicWolframAlphaSearchRetrieverPrompt = `
|
||||||
You will be given a conversation below and a follow up question. You need to rephrase the follow-up question if needed so it is a standalone question that can be used by the LLM to search the web for information.
|
You will be given a conversation below and a follow up question. You need to rephrase the follow-up question if needed so it is a standalone question that can be used by the LLM to search the web for information.
|
||||||
|
@ -65,7 +66,7 @@ const basicWolframAlphaSearchResponsePrompt = `
|
||||||
const strParser = new StringOutputParser();
|
const strParser = new StringOutputParser();
|
||||||
|
|
||||||
const handleStream = async (
|
const handleStream = async (
|
||||||
stream: AsyncGenerator<StreamEvent, any, unknown>,
|
stream: IterableReadableStream<StreamEvent>,
|
||||||
emitter: eventEmitter,
|
emitter: eventEmitter,
|
||||||
) => {
|
) => {
|
||||||
for await (const event of stream) {
|
for await (const event of stream) {
|
||||||
|
|
|
@ -10,6 +10,7 @@ import eventEmitter from 'events';
|
||||||
import type { BaseChatModel } from '@langchain/core/language_models/chat_models';
|
import type { BaseChatModel } from '@langchain/core/language_models/chat_models';
|
||||||
import type { Embeddings } from '@langchain/core/embeddings';
|
import type { Embeddings } from '@langchain/core/embeddings';
|
||||||
import logger from '../utils/logger';
|
import logger from '../utils/logger';
|
||||||
|
import { IterableReadableStream } from '@langchain/core/utils/stream';
|
||||||
|
|
||||||
const writingAssistantPrompt = `
|
const writingAssistantPrompt = `
|
||||||
You are Perplexica, an AI model who is expert at searching the web and answering user's queries. You are currently set on focus mode 'Writing Assistant', this means you will be helping the user write a response to a given query.
|
You are Perplexica, an AI model who is expert at searching the web and answering user's queries. You are currently set on focus mode 'Writing Assistant', this means you will be helping the user write a response to a given query.
|
||||||
|
@ -19,7 +20,7 @@ Since you are a writing assistant, you would not perform web searches. If you th
|
||||||
const strParser = new StringOutputParser();
|
const strParser = new StringOutputParser();
|
||||||
|
|
||||||
const handleStream = async (
|
const handleStream = async (
|
||||||
stream: AsyncGenerator<StreamEvent, any, unknown>,
|
stream: IterableReadableStream<StreamEvent>,
|
||||||
emitter: eventEmitter,
|
emitter: eventEmitter,
|
||||||
) => {
|
) => {
|
||||||
for await (const event of stream) {
|
for await (const event of stream) {
|
||||||
|
|
|
@ -19,6 +19,7 @@ import formatChatHistoryAsString from '../utils/formatHistory';
|
||||||
import eventEmitter from 'events';
|
import eventEmitter from 'events';
|
||||||
import computeSimilarity from '../utils/computeSimilarity';
|
import computeSimilarity from '../utils/computeSimilarity';
|
||||||
import logger from '../utils/logger';
|
import logger from '../utils/logger';
|
||||||
|
import { IterableReadableStream } from '@langchain/core/utils/stream';
|
||||||
|
|
||||||
const basicYoutubeSearchRetrieverPrompt = `
|
const basicYoutubeSearchRetrieverPrompt = `
|
||||||
You will be given a conversation below and a follow up question. You need to rephrase the follow-up question if needed so it is a standalone question that can be used by the LLM to search the web for information.
|
You will be given a conversation below and a follow up question. You need to rephrase the follow-up question if needed so it is a standalone question that can be used by the LLM to search the web for information.
|
||||||
|
@ -66,7 +67,7 @@ const basicYoutubeSearchResponsePrompt = `
|
||||||
const strParser = new StringOutputParser();
|
const strParser = new StringOutputParser();
|
||||||
|
|
||||||
const handleStream = async (
|
const handleStream = async (
|
||||||
stream: AsyncGenerator<StreamEvent, any, unknown>,
|
stream: IterableReadableStream<StreamEvent>,
|
||||||
emitter: eventEmitter,
|
emitter: eventEmitter,
|
||||||
) => {
|
) => {
|
||||||
for await (const event of stream) {
|
for await (const event of stream) {
|
||||||
|
|
|
@ -3,6 +3,7 @@ import { htmlToText } from 'html-to-text';
|
||||||
import { RecursiveCharacterTextSplitter } from 'langchain/text_splitter';
|
import { RecursiveCharacterTextSplitter } from 'langchain/text_splitter';
|
||||||
import { Document } from '@langchain/core/documents';
|
import { Document } from '@langchain/core/documents';
|
||||||
import pdfParse from 'pdf-parse';
|
import pdfParse from 'pdf-parse';
|
||||||
|
import logger from '../utils/logger';
|
||||||
|
|
||||||
export const getDocumentsFromLinks = async ({ links }: { links: string[] }) => {
|
export const getDocumentsFromLinks = async ({ links }: { links: string[] }) => {
|
||||||
const splitter = new RecursiveCharacterTextSplitter();
|
const splitter = new RecursiveCharacterTextSplitter();
|
||||||
|
@ -16,6 +17,7 @@ export const getDocumentsFromLinks = async ({ links }: { links: string[] }) => {
|
||||||
? link
|
? link
|
||||||
: `https://${link}`;
|
: `https://${link}`;
|
||||||
|
|
||||||
|
try {
|
||||||
const res = await axios.get(link, {
|
const res = await axios.get(link, {
|
||||||
responseType: 'arraybuffer',
|
responseType: 'arraybuffer',
|
||||||
});
|
});
|
||||||
|
@ -76,6 +78,20 @@ export const getDocumentsFromLinks = async ({ links }: { links: string[] }) => {
|
||||||
});
|
});
|
||||||
|
|
||||||
docs.push(...linkDocs);
|
docs.push(...linkDocs);
|
||||||
|
} catch (err) {
|
||||||
|
logger.error(
|
||||||
|
`Error at generating documents from links: ${err.message}`,
|
||||||
|
);
|
||||||
|
docs.push(
|
||||||
|
new Document({
|
||||||
|
pageContent: `Failed to retrieve content from the link: ${err.message}`,
|
||||||
|
metadata: {
|
||||||
|
title: 'Failed to retrieve content',
|
||||||
|
url: link,
|
||||||
|
},
|
||||||
|
}),
|
||||||
|
);
|
||||||
|
}
|
||||||
}),
|
}),
|
||||||
);
|
);
|
||||||
|
|
||||||
|
|
|
@ -83,7 +83,9 @@ const useSocket = (
|
||||||
chatModelProvider = Object.keys(chatModelProviders)[0];
|
chatModelProvider = Object.keys(chatModelProviders)[0];
|
||||||
|
|
||||||
if (chatModelProvider === 'custom_openai') {
|
if (chatModelProvider === 'custom_openai') {
|
||||||
toast.error('Seems like you are using the custom OpenAI provider, please open the settings and configure the API key and base URL');
|
toast.error(
|
||||||
|
'Seems like you are using the custom OpenAI provider, please open the settings and configure the API key and base URL',
|
||||||
|
);
|
||||||
setError(true);
|
setError(true);
|
||||||
return;
|
return;
|
||||||
} else {
|
} else {
|
||||||
|
@ -220,7 +222,7 @@ const useSocket = (
|
||||||
if (data.type === 'error') {
|
if (data.type === 'error') {
|
||||||
toast.error(data.data);
|
toast.error(data.data);
|
||||||
}
|
}
|
||||||
})
|
});
|
||||||
|
|
||||||
setWs(ws);
|
setWs(ws);
|
||||||
};
|
};
|
||||||
|
@ -235,13 +237,6 @@ const useSocket = (
|
||||||
setError(true);
|
setError(true);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
return () => {
|
|
||||||
if (ws?.readyState === 1) {
|
|
||||||
ws?.close();
|
|
||||||
console.log('[DEBUG] closed');
|
|
||||||
}
|
|
||||||
};
|
|
||||||
}, [ws, url, setIsWSReady, setError]);
|
}, [ws, url, setIsWSReady, setError]);
|
||||||
|
|
||||||
return ws;
|
return ws;
|
||||||
|
@ -348,6 +343,15 @@ const ChatWindow = ({ id }: { id?: string }) => {
|
||||||
// eslint-disable-next-line react-hooks/exhaustive-deps
|
// eslint-disable-next-line react-hooks/exhaustive-deps
|
||||||
}, []);
|
}, []);
|
||||||
|
|
||||||
|
useEffect(() => {
|
||||||
|
return () => {
|
||||||
|
if (ws?.readyState === 1) {
|
||||||
|
ws.close();
|
||||||
|
console.log('[DEBUG] closed');
|
||||||
|
}
|
||||||
|
};
|
||||||
|
}, []);
|
||||||
|
|
||||||
const messagesRef = useRef<Message[]>([]);
|
const messagesRef = useRef<Message[]>([]);
|
||||||
|
|
||||||
useEffect(() => {
|
useEffect(() => {
|
||||||
|
|
|
@ -1,6 +1,6 @@
|
||||||
{
|
{
|
||||||
"name": "perplexica-frontend",
|
"name": "perplexica-frontend",
|
||||||
"version": "1.9.0-rc1",
|
"version": "1.9.0-rc2",
|
||||||
"license": "MIT",
|
"license": "MIT",
|
||||||
"author": "ItzCrazyKns",
|
"author": "ItzCrazyKns",
|
||||||
"scripts": {
|
"scripts": {
|
||||||
|
|
Loading…
Reference in New Issue