Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Support historical message image inputs/attachments for n+1 queries #2919

Merged
merged 7 commits into from
Jan 16, 2025
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
3 changes: 2 additions & 1 deletion server/utils/AiProviders/anthropic/index.js
Original file line number Diff line number Diff line change
Expand Up @@ -2,6 +2,7 @@ const { v4 } = require("uuid");
const {
writeResponseChunk,
clientAbortedHandler,
formatChatHistory,
} = require("../../helpers/chat/responses");
const { NativeEmbedder } = require("../../EmbeddingEngines/native");
const { MODEL_MAP } = require("../modelMap");
Expand Down Expand Up @@ -99,7 +100,7 @@ class AnthropicLLM {

return [
prompt,
...chatHistory,
...formatChatHistory(chatHistory, this.#generateContent),
{
role: "user",
content: this.#generateContent({ userPrompt, attachments }),
Expand Down
3 changes: 2 additions & 1 deletion server/utils/AiProviders/apipie/index.js
Original file line number Diff line number Diff line change
Expand Up @@ -3,6 +3,7 @@ const { v4: uuidv4 } = require("uuid");
const {
writeResponseChunk,
clientAbortedHandler,
formatChatHistory,
} = require("../../helpers/chat/responses");
const fs = require("fs");
const path = require("path");
Expand Down Expand Up @@ -177,7 +178,7 @@ class ApiPieLLM {
};
return [
prompt,
...chatHistory,
...formatChatHistory(chatHistory, this.#generateContent),
{
role: "user",
content: this.#generateContent({ userPrompt, attachments }),
Expand Down
3 changes: 2 additions & 1 deletion server/utils/AiProviders/azureOpenAi/index.js
Original file line number Diff line number Diff line change
Expand Up @@ -5,6 +5,7 @@ const {
const {
writeResponseChunk,
clientAbortedHandler,
formatChatHistory,
} = require("../../helpers/chat/responses");

class AzureOpenAiLLM {
Expand Down Expand Up @@ -103,7 +104,7 @@ class AzureOpenAiLLM {
};
return [
prompt,
...chatHistory,
...formatChatHistory(chatHistory, this.#generateContent),
{
role: "user",
content: this.#generateContent({ userPrompt, attachments }),
Expand Down
5 changes: 3 additions & 2 deletions server/utils/AiProviders/bedrock/index.js
Original file line number Diff line number Diff line change
Expand Up @@ -2,6 +2,7 @@ const { StringOutputParser } = require("@langchain/core/output_parsers");
const {
writeResponseChunk,
clientAbortedHandler,
formatChatHistory,
} = require("../../helpers/chat/responses");
const { NativeEmbedder } = require("../../EmbeddingEngines/native");
const {
Expand Down Expand Up @@ -199,7 +200,7 @@ class AWSBedrockLLM {
// AWS Mistral models do not support system prompts
if (this.model.startsWith("mistral"))
return [
...chatHistory,
...formatChatHistory(chatHistory, this.#generateContent, "spread"),
{
role: "user",
...this.#generateContent({ userPrompt, attachments }),
Expand All @@ -212,7 +213,7 @@ class AWSBedrockLLM {
};
return [
prompt,
...chatHistory,
...formatChatHistory(chatHistory, this.#generateContent, "spread"),
{
role: "user",
...this.#generateContent({ userPrompt, attachments }),
Expand Down
15 changes: 13 additions & 2 deletions server/utils/AiProviders/gemini/index.js
Original file line number Diff line number Diff line change
Expand Up @@ -7,6 +7,7 @@ const {
const {
writeResponseChunk,
clientAbortedHandler,
formatChatHistory,
} = require("../../helpers/chat/responses");
const { MODEL_MAP } = require("../modelMap");
const { defaultGeminiModels, v1BetaModels } = require("./defaultModels");
Expand Down Expand Up @@ -254,6 +255,7 @@ class GeminiLLM {
const models = await this.fetchModels(process.env.GEMINI_API_KEY);
return models.some((model) => model.id === modelName);
}

/**
* Generates appropriate content array for a message + attachments.
* @param {{userPrompt:string, attachments: import("../../helpers").Attachment[]}}
Expand Down Expand Up @@ -290,7 +292,7 @@ class GeminiLLM {
return [
prompt,
{ role: "assistant", content: "Okay." },
...chatHistory,
...formatChatHistory(chatHistory, this.#generateContent),
{
role: "USER_PROMPT",
content: this.#generateContent({ userPrompt, attachments }),
Expand All @@ -306,8 +308,17 @@ class GeminiLLM {
.map((message) => {
if (message.role === "system")
return { role: "user", parts: [{ text: message.content }] };
if (message.role === "user")

if (message.role === "user") {
// If the content is an array - then we have already formatted the context so return it directly.
if (Array.isArray(message.content))
return { role: "user", parts: message.content };

// Otherwise, this was a regular user message with no attachments
// so we need to format it for Gemini
return { role: "user", parts: [{ text: message.content }] };
}

if (message.role === "assistant")
return { role: "model", parts: [{ text: message.content }] };
return null;
Expand Down
3 changes: 2 additions & 1 deletion server/utils/AiProviders/genericOpenAi/index.js
Original file line number Diff line number Diff line change
Expand Up @@ -4,6 +4,7 @@ const {
} = require("../../helpers/chat/LLMPerformanceMonitor");
const {
handleDefaultStreamResponseV2,
formatChatHistory,
} = require("../../helpers/chat/responses");
const { toValidNumber } = require("../../http");

Expand Down Expand Up @@ -133,7 +134,7 @@ class GenericOpenAiLLM {
};
return [
prompt,
...chatHistory,
...formatChatHistory(chatHistory, this.#generateContent),
{
role: "user",
content: this.#generateContent({ userPrompt, attachments }),
Expand Down
2 changes: 2 additions & 0 deletions server/utils/AiProviders/groq/index.js
Original file line number Diff line number Diff line change
Expand Up @@ -89,6 +89,8 @@ class GroqLLM {
* Since we can only explicitly support the current models, this is a temporary solution.
* If the attachments are empty or the model is not a vision model, we will return the default prompt structure which will work for all models.
* If the attachments are present and the model is a vision model - we only return the user prompt with attachments - see comment at end of function for more.
*
* Historical attachments are also omitted from prompt chat history for the reasons above. (TDC: Dec 30, 2024)
*/
#conditionalPromptStruct({
systemPrompt = "",
Expand Down
3 changes: 2 additions & 1 deletion server/utils/AiProviders/koboldCPP/index.js
Original file line number Diff line number Diff line change
Expand Up @@ -2,6 +2,7 @@ const { NativeEmbedder } = require("../../EmbeddingEngines/native");
const {
clientAbortedHandler,
writeResponseChunk,
formatChatHistory,
} = require("../../helpers/chat/responses");
const {
LLMPerformanceMonitor,
Expand Down Expand Up @@ -116,7 +117,7 @@ class KoboldCPPLLM {
};
return [
prompt,
...chatHistory,
...formatChatHistory(chatHistory, this.#generateContent),
{
role: "user",
content: this.#generateContent({ userPrompt, attachments }),
Expand Down
3 changes: 2 additions & 1 deletion server/utils/AiProviders/liteLLM/index.js
Original file line number Diff line number Diff line change
Expand Up @@ -4,6 +4,7 @@ const {
} = require("../../helpers/chat/LLMPerformanceMonitor");
const {
handleDefaultStreamResponseV2,
formatChatHistory,
} = require("../../helpers/chat/responses");

class LiteLLM {
Expand Down Expand Up @@ -115,7 +116,7 @@ class LiteLLM {
};
return [
prompt,
...chatHistory,
...formatChatHistory(chatHistory, this.#generateContent),
{
role: "user",
content: this.#generateContent({ userPrompt, attachments }),
Expand Down
3 changes: 2 additions & 1 deletion server/utils/AiProviders/lmStudio/index.js
Original file line number Diff line number Diff line change
@@ -1,6 +1,7 @@
const { NativeEmbedder } = require("../../EmbeddingEngines/native");
const {
handleDefaultStreamResponseV2,
formatChatHistory,
} = require("../../helpers/chat/responses");
const {
LLMPerformanceMonitor,
Expand Down Expand Up @@ -117,7 +118,7 @@ class LMStudioLLM {
};
return [
prompt,
...chatHistory,
...formatChatHistory(chatHistory, this.#generateContent),
{
role: "user",
content: this.#generateContent({ userPrompt, attachments }),
Expand Down
3 changes: 2 additions & 1 deletion server/utils/AiProviders/localAi/index.js
Original file line number Diff line number Diff line change
Expand Up @@ -4,6 +4,7 @@ const {
} = require("../../helpers/chat/LLMPerformanceMonitor");
const {
handleDefaultStreamResponseV2,
formatChatHistory,
} = require("../../helpers/chat/responses");

class LocalAiLLM {
Expand Down Expand Up @@ -103,7 +104,7 @@ class LocalAiLLM {
};
return [
prompt,
...chatHistory,
...formatChatHistory(chatHistory, this.#generateContent),
{
role: "user",
content: this.#generateContent({ userPrompt, attachments }),
Expand Down
8 changes: 7 additions & 1 deletion server/utils/AiProviders/mistral/index.js
Original file line number Diff line number Diff line change
Expand Up @@ -4,6 +4,7 @@ const {
} = require("../../helpers/chat/LLMPerformanceMonitor");
const {
handleDefaultStreamResponseV2,
formatChatHistory,
} = require("../../helpers/chat/responses");

class MistralLLM {
Expand All @@ -26,6 +27,11 @@ class MistralLLM {

this.embedder = embedder ?? new NativeEmbedder();
this.defaultTemp = 0.0;
this.log("Initialized with model:", this.model);
}

log(text, ...args) {
console.log(`\x1b[36m[${this.constructor.name}]\x1b[0m ${text}`, ...args);
}

#appendContext(contextTexts = []) {
Expand Down Expand Up @@ -92,7 +98,7 @@ class MistralLLM {
};
return [
prompt,
...chatHistory,
...formatChatHistory(chatHistory, this.#generateContent),
{
role: "user",
content: this.#generateContent({ userPrompt, attachments }),
Expand Down
3 changes: 2 additions & 1 deletion server/utils/AiProviders/novita/index.js
Original file line number Diff line number Diff line change
Expand Up @@ -3,6 +3,7 @@ const { v4: uuidv4 } = require("uuid");
const {
writeResponseChunk,
clientAbortedHandler,
formatChatHistory,
} = require("../../helpers/chat/responses");
const fs = require("fs");
const path = require("path");
Expand Down Expand Up @@ -177,7 +178,7 @@ class NovitaLLM {
};
return [
prompt,
...chatHistory,
...formatChatHistory(chatHistory, this.#generateContent),
{
role: "user",
content: this.#generateContent({ userPrompt, attachments }),
Expand Down
3 changes: 2 additions & 1 deletion server/utils/AiProviders/nvidiaNim/index.js
Original file line number Diff line number Diff line change
Expand Up @@ -4,6 +4,7 @@ const {
} = require("../../helpers/chat/LLMPerformanceMonitor");
const {
handleDefaultStreamResponseV2,
formatChatHistory,
} = require("../../helpers/chat/responses");

class NvidiaNimLLM {
Expand Down Expand Up @@ -142,7 +143,7 @@ class NvidiaNimLLM {
};
return [
prompt,
...chatHistory,
...formatChatHistory(chatHistory, this.#generateContent),
{
role: "user",
content: this.#generateContent({ userPrompt, attachments }),
Expand Down
3 changes: 2 additions & 1 deletion server/utils/AiProviders/ollama/index.js
Original file line number Diff line number Diff line change
@@ -1,6 +1,7 @@
const {
writeResponseChunk,
clientAbortedHandler,
formatChatHistory,
} = require("../../helpers/chat/responses");
const { NativeEmbedder } = require("../../EmbeddingEngines/native");
const {
Expand Down Expand Up @@ -120,7 +121,7 @@ class OllamaAILLM {
};
return [
prompt,
...chatHistory,
...formatChatHistory(chatHistory, this.#generateContent, "spread"),
{
role: "user",
...this.#generateContent({ userPrompt, attachments }),
Expand Down
3 changes: 2 additions & 1 deletion server/utils/AiProviders/openAi/index.js
Original file line number Diff line number Diff line change
@@ -1,6 +1,7 @@
const { NativeEmbedder } = require("../../EmbeddingEngines/native");
const {
handleDefaultStreamResponseV2,
formatChatHistory,
} = require("../../helpers/chat/responses");
const { MODEL_MAP } = require("../modelMap");
const {
Expand Down Expand Up @@ -121,7 +122,7 @@ class OpenAiLLM {
};
return [
prompt,
...chatHistory,
...formatChatHistory(chatHistory, this.#generateContent),
{
role: "user",
content: this.#generateContent({ userPrompt, attachments }),
Expand Down
5 changes: 3 additions & 2 deletions server/utils/AiProviders/openRouter/index.js
Original file line number Diff line number Diff line change
Expand Up @@ -3,6 +3,7 @@ const { v4: uuidv4 } = require("uuid");
const {
writeResponseChunk,
clientAbortedHandler,
formatChatHistory,
} = require("../../helpers/chat/responses");
const fs = require("fs");
const path = require("path");
Expand Down Expand Up @@ -47,6 +48,7 @@ class OpenRouterLLM {
fs.mkdirSync(cacheFolder, { recursive: true });
this.cacheModelPath = path.resolve(cacheFolder, "models.json");
this.cacheAtPath = path.resolve(cacheFolder, ".cached_at");
this.log("Initialized with model:", this.model);
}

log(text, ...args) {
Expand Down Expand Up @@ -162,7 +164,6 @@ class OpenRouterLLM {
},
});
}
console.log(content.flat());
return content.flat();
}

Expand All @@ -179,7 +180,7 @@ class OpenRouterLLM {
};
return [
prompt,
...chatHistory,
...formatChatHistory(chatHistory, this.#generateContent),
{
role: "user",
content: this.#generateContent({ userPrompt, attachments }),
Expand Down
3 changes: 2 additions & 1 deletion server/utils/AiProviders/textGenWebUI/index.js
Original file line number Diff line number Diff line change
@@ -1,6 +1,7 @@
const { NativeEmbedder } = require("../../EmbeddingEngines/native");
const {
handleDefaultStreamResponseV2,
formatChatHistory,
} = require("../../helpers/chat/responses");
const {
LLMPerformanceMonitor,
Expand Down Expand Up @@ -113,7 +114,7 @@ class TextGenWebUILLM {
};
return [
prompt,
...chatHistory,
...formatChatHistory(chatHistory, this.#generateContent),
{
role: "user",
content: this.#generateContent({ userPrompt, attachments }),
Expand Down
Loading