diff --git a/server/utils/AiProviders/anthropic/index.js b/server/utils/AiProviders/anthropic/index.js index 450b376b3e0..2170ba83f71 100644 --- a/server/utils/AiProviders/anthropic/index.js +++ b/server/utils/AiProviders/anthropic/index.js @@ -15,6 +15,7 @@ class AnthropicLLM { if (!process.env.ANTHROPIC_API_KEY) throw new Error("No Anthropic API key was set."); + this.className = "AnthropicLLM"; // Docs: https://www.npmjs.com/package/@anthropic-ai/sdk const AnthropicAI = require("@anthropic-ai/sdk"); const anthropic = new AnthropicAI({ @@ -37,7 +38,7 @@ class AnthropicLLM { } log(text, ...args) { - console.log(`\x1b[36m[${this.constructor.name}]\x1b[0m ${text}`, ...args); + console.log(`\x1b[36m[${this.className}]\x1b[0m ${text}`, ...args); } streamingEnabled() { diff --git a/server/utils/AiProviders/apipie/index.js b/server/utils/AiProviders/apipie/index.js index bd794d38ea6..534e7c27ca5 100644 --- a/server/utils/AiProviders/apipie/index.js +++ b/server/utils/AiProviders/apipie/index.js @@ -23,6 +23,7 @@ class ApiPieLLM { if (!process.env.APIPIE_LLM_API_KEY) throw new Error("No ApiPie LLM API key was set."); + this.className = "ApiPieLLM"; const { OpenAI: OpenAIApi } = require("openai"); this.basePath = "https://apipie.ai/v1"; this.openai = new OpenAIApi({ @@ -49,7 +50,7 @@ class ApiPieLLM { } log(text, ...args) { - console.log(`\x1b[36m[${this.constructor.name}]\x1b[0m ${text}`, ...args); + console.log(`\x1b[36m[${this.className}]\x1b[0m ${text}`, ...args); } // This checks if the .cached_at file has a timestamp that is more than 1Week (in millis) diff --git a/server/utils/AiProviders/cometapi/index.js b/server/utils/AiProviders/cometapi/index.js index fca0b0cc50a..23e069d5b38 100644 --- a/server/utils/AiProviders/cometapi/index.js +++ b/server/utils/AiProviders/cometapi/index.js @@ -24,6 +24,7 @@ class CometApiLLM { if (!process.env.COMETAPI_LLM_API_KEY) throw new Error("No CometAPI API key was set."); + this.className = "CometApiLLM"; const { OpenAI: OpenAIApi } = require("openai"); this.basePath = "https://api.cometapi.com/v1"; this.openai = new OpenAIApi({ @@ -55,7 +56,7 @@ class CometApiLLM { } log(text, ...args) { - console.log(`\x1b[36m[${this.constructor.name}]\x1b[0m ${text}`, ...args); + console.log(`\x1b[36m[${this.className}]\x1b[0m ${text}`, ...args); } /** diff --git a/server/utils/AiProviders/deepseek/index.js b/server/utils/AiProviders/deepseek/index.js index 4b5b334bfe0..379012d008d 100644 --- a/server/utils/AiProviders/deepseek/index.js +++ b/server/utils/AiProviders/deepseek/index.js @@ -13,6 +13,7 @@ class DeepSeekLLM { constructor(embedder = null, modelPreference = null) { if (!process.env.DEEPSEEK_API_KEY) throw new Error("No DeepSeek API key was set."); + this.className = "DeepSeekLLM"; const { OpenAI: OpenAIApi } = require("openai"); this.openai = new OpenAIApi({ @@ -35,7 +36,7 @@ class DeepSeekLLM { } log(text, ...args) { - console.log(`\x1b[36m[${this.constructor.name}]\x1b[0m ${text}`, ...args); + console.log(`\x1b[36m[${this.className}]\x1b[0m ${text}`, ...args); } #appendContext(contextTexts = []) { diff --git a/server/utils/AiProviders/dellProAiStudio/index.js b/server/utils/AiProviders/dellProAiStudio/index.js index 2cca4c8bf66..7c233ec26b0 100644 --- a/server/utils/AiProviders/dellProAiStudio/index.js +++ b/server/utils/AiProviders/dellProAiStudio/index.js @@ -13,6 +13,7 @@ class DellProAiStudioLLM { if (!process.env.DPAIS_LLM_BASE_PATH) throw new Error("No Dell Pro AI Studio Base Path was set."); + this.className = "DellProAiStudioLLM"; const { OpenAI: OpenAIApi } = require("openai"); this.dpais = new OpenAIApi({ baseURL: DellProAiStudioLLM.parseBasePath(), @@ -50,7 +51,7 @@ class DellProAiStudioLLM { } log(text, ...args) { - console.log(`\x1b[36m[${this.constructor.name}]\x1b[0m ${text}`, ...args); + console.log(`\x1b[36m[${this.className}]\x1b[0m ${text}`, ...args); } #appendContext(contextTexts = []) { diff --git a/server/utils/AiProviders/gemini/index.js b/server/utils/AiProviders/gemini/index.js index c7d39bc782a..de76ec79c82 100644 --- a/server/utils/AiProviders/gemini/index.js +++ b/server/utils/AiProviders/gemini/index.js @@ -29,6 +29,7 @@ class GeminiLLM { if (!process.env.GEMINI_API_KEY) throw new Error("No Gemini API key was set."); + this.className = "GeminiLLM"; const { OpenAI: OpenAIApi } = require("openai"); this.model = modelPreference || @@ -71,7 +72,7 @@ class GeminiLLM { } #log(text, ...args) { - console.log(`\x1b[32m[GeminiLLM]\x1b[0m ${text}`, ...args); + console.log(`\x1b[32m[${this.className}]\x1b[0m ${text}`, ...args); } // This checks if the .cached_at file has a timestamp that is more than 1Week (in millis) diff --git a/server/utils/AiProviders/genericOpenAi/index.js b/server/utils/AiProviders/genericOpenAi/index.js index 3b4c179d1ed..8e32ad5921d 100644 --- a/server/utils/AiProviders/genericOpenAi/index.js +++ b/server/utils/AiProviders/genericOpenAi/index.js @@ -18,6 +18,7 @@ class GenericOpenAiLLM { "GenericOpenAI must have a valid base path to use for the api." ); + this.className = "GenericOpenAiLLM"; this.basePath = process.env.GENERIC_OPEN_AI_BASE_PATH; this.openai = new OpenAIApi({ baseURL: this.basePath, @@ -45,7 +46,7 @@ class GenericOpenAiLLM { } log(text, ...args) { - console.log(`\x1b[36m[${this.constructor.name}]\x1b[0m ${text}`, ...args); + console.log(`\x1b[36m[${this.className}]\x1b[0m ${text}`, ...args); } #appendContext(contextTexts = []) { diff --git a/server/utils/AiProviders/koboldCPP/index.js b/server/utils/AiProviders/koboldCPP/index.js index 64fa8c456ca..b795b152227 100644 --- a/server/utils/AiProviders/koboldCPP/index.js +++ b/server/utils/AiProviders/koboldCPP/index.js @@ -17,6 +17,7 @@ class KoboldCPPLLM { "KoboldCPP must have a valid base path to use for the api." ); + this.className = "KoboldCPPLLM"; this.basePath = process.env.KOBOLD_CPP_BASE_PATH; this.openai = new OpenAIApi({ baseURL: this.basePath, @@ -37,7 +38,7 @@ class KoboldCPPLLM { } log(text, ...args) { - console.log(`\x1b[36m[${this.constructor.name}]\x1b[0m ${text}`, ...args); + console.log(`\x1b[36m[${this.className}]\x1b[0m ${text}`, ...args); } #appendContext(contextTexts = []) { diff --git a/server/utils/AiProviders/liteLLM/index.js b/server/utils/AiProviders/liteLLM/index.js index 2017d7774f8..91aa8517749 100644 --- a/server/utils/AiProviders/liteLLM/index.js +++ b/server/utils/AiProviders/liteLLM/index.js @@ -15,6 +15,7 @@ class LiteLLM { "LiteLLM must have a valid base path to use for the api." ); + this.className = "LiteLLM"; this.basePath = process.env.LITE_LLM_BASE_PATH; this.openai = new OpenAIApi({ baseURL: this.basePath, @@ -35,7 +36,7 @@ class LiteLLM { } log(text, ...args) { - console.log(`\x1b[36m[${this.constructor.name}]\x1b[0m ${text}`, ...args); + console.log(`\x1b[36m[${this.className}]\x1b[0m ${text}`, ...args); } #appendContext(contextTexts = []) { diff --git a/server/utils/AiProviders/mistral/index.js b/server/utils/AiProviders/mistral/index.js index 6c637857b35..4cf547ccc7c 100644 --- a/server/utils/AiProviders/mistral/index.js +++ b/server/utils/AiProviders/mistral/index.js @@ -12,6 +12,7 @@ class MistralLLM { if (!process.env.MISTRAL_API_KEY) throw new Error("No Mistral API key was set."); + this.className = "MistralLLM"; const { OpenAI: OpenAIApi } = require("openai"); this.openai = new OpenAIApi({ baseURL: "https://api.mistral.ai/v1", @@ -31,7 +32,7 @@ class MistralLLM { } log(text, ...args) { - console.log(`\x1b[36m[${this.constructor.name}]\x1b[0m ${text}`, ...args); + console.log(`\x1b[36m[${this.className}]\x1b[0m ${text}`, ...args); } #appendContext(contextTexts = []) { diff --git a/server/utils/AiProviders/moonshotAi/index.js b/server/utils/AiProviders/moonshotAi/index.js index c4bc7b65bbd..b00f72134ec 100644 --- a/server/utils/AiProviders/moonshotAi/index.js +++ b/server/utils/AiProviders/moonshotAi/index.js @@ -12,6 +12,7 @@ class MoonshotAiLLM { constructor(embedder = null, modelPreference = null) { if (!process.env.MOONSHOT_AI_API_KEY) throw new Error("No Moonshot AI API key was set."); + this.className = "MoonshotAiLLM"; const { OpenAI: OpenAIApi } = require("openai"); this.openai = new OpenAIApi({ @@ -36,7 +37,7 @@ class MoonshotAiLLM { } log(text, ...args) { - console.log(`\x1b[36m[${this.constructor.name}]\x1b[0m ${text}`, ...args); + console.log(`\x1b[36m[${this.className}]\x1b[0m ${text}`, ...args); } #appendContext(contextTexts = []) { diff --git a/server/utils/AiProviders/novita/index.js b/server/utils/AiProviders/novita/index.js index 08be9cf83cb..69c06753752 100644 --- a/server/utils/AiProviders/novita/index.js +++ b/server/utils/AiProviders/novita/index.js @@ -24,6 +24,7 @@ class NovitaLLM { if (!process.env.NOVITA_LLM_API_KEY) throw new Error("No Novita API key was set."); + this.className = "NovitaLLM"; const { OpenAI: OpenAIApi } = require("openai"); this.basePath = "https://api.novita.ai/v3/openai"; this.openai = new OpenAIApi({ @@ -57,7 +58,7 @@ class NovitaLLM { } log(text, ...args) { - console.log(`\x1b[36m[${this.constructor.name}]\x1b[0m ${text}`, ...args); + console.log(`\x1b[36m[${this.className}]\x1b[0m ${text}`, ...args); } /** diff --git a/server/utils/AiProviders/nvidiaNim/index.js b/server/utils/AiProviders/nvidiaNim/index.js index b421fdc15ba..f932625cb04 100644 --- a/server/utils/AiProviders/nvidiaNim/index.js +++ b/server/utils/AiProviders/nvidiaNim/index.js @@ -12,6 +12,7 @@ class NvidiaNimLLM { if (!process.env.NVIDIA_NIM_LLM_BASE_PATH) throw new Error("No NVIDIA NIM API Base Path was set."); + this.className = "NvidiaNimLLM"; const { OpenAI: OpenAIApi } = require("openai"); this.nvidiaNim = new OpenAIApi({ baseURL: parseNvidiaNimBasePath(process.env.NVIDIA_NIM_LLM_BASE_PATH), @@ -33,7 +34,7 @@ class NvidiaNimLLM { } #log(text, ...args) { - console.log(`\x1b[36m[${this.constructor.name}]\x1b[0m ${text}`, ...args); + console.log(`\x1b[36m[${this.className}]\x1b[0m ${text}`, ...args); } #appendContext(contextTexts = []) { diff --git a/server/utils/AiProviders/openAi/index.js b/server/utils/AiProviders/openAi/index.js index bc159d76f29..001c27ea65f 100644 --- a/server/utils/AiProviders/openAi/index.js +++ b/server/utils/AiProviders/openAi/index.js @@ -13,6 +13,7 @@ const { class OpenAiLLM { constructor(embedder = null, modelPreference = null) { if (!process.env.OPEN_AI_KEY) throw new Error("No OpenAI API key was set."); + this.className = "OpenAiLLM"; const { OpenAI: OpenAIApi } = require("openai"); this.openai = new OpenAIApi({ @@ -33,7 +34,7 @@ class OpenAiLLM { } log(text, ...args) { - console.log(`\x1b[36m[${this.constructor.name}]\x1b[0m ${text}`, ...args); + console.log(`\x1b[36m[${this.className}]\x1b[0m ${text}`, ...args); } #appendContext(contextTexts = []) { diff --git a/server/utils/AiProviders/openRouter/index.js b/server/utils/AiProviders/openRouter/index.js index ea1665c04db..7a0fc1c355f 100644 --- a/server/utils/AiProviders/openRouter/index.js +++ b/server/utils/AiProviders/openRouter/index.js @@ -32,6 +32,7 @@ class OpenRouterLLM { if (!process.env.OPENROUTER_API_KEY) throw new Error("No OpenRouter API key was set."); + this.className = "OpenRouterLLM"; const { OpenAI: OpenAIApi } = require("openai"); this.basePath = "https://openrouter.ai/api/v1"; this.openai = new OpenAIApi({ @@ -88,7 +89,7 @@ class OpenRouterLLM { } log(text, ...args) { - console.log(`\x1b[36m[${this.constructor.name}]\x1b[0m ${text}`, ...args); + console.log(`\x1b[36m[${this.className}]\x1b[0m ${text}`, ...args); } /** diff --git a/server/utils/AiProviders/ppio/index.js b/server/utils/AiProviders/ppio/index.js index 677cd4cd04f..bef7fada68a 100644 --- a/server/utils/AiProviders/ppio/index.js +++ b/server/utils/AiProviders/ppio/index.js @@ -18,6 +18,7 @@ class PPIOLLM { constructor(embedder = null, modelPreference = null) { if (!process.env.PPIO_API_KEY) throw new Error("No PPIO API key was set."); + this.className = "PPIOLLM"; const { OpenAI: OpenAIApi } = require("openai"); this.basePath = "https://api.ppinfra.com/v3/openai/"; this.openai = new OpenAIApi({ @@ -50,7 +51,7 @@ class PPIOLLM { } log(text, ...args) { - console.log(`\x1b[36m[${this.constructor.name}]\x1b[0m ${text}`, ...args); + console.log(`\x1b[36m[${this.className}]\x1b[0m ${text}`, ...args); } async #syncModels() { diff --git a/server/utils/AiProviders/textGenWebUI/index.js b/server/utils/AiProviders/textGenWebUI/index.js index f3647c06d45..2de1d8cd567 100644 --- a/server/utils/AiProviders/textGenWebUI/index.js +++ b/server/utils/AiProviders/textGenWebUI/index.js @@ -15,6 +15,7 @@ class TextGenWebUILLM { "TextGenWebUI must have a valid base path to use for the api." ); + this.className = "TextGenWebUILLM"; this.basePath = process.env.TEXT_GEN_WEB_UI_BASE_PATH; this.openai = new OpenAIApi({ baseURL: this.basePath, @@ -33,7 +34,7 @@ class TextGenWebUILLM { } log(text, ...args) { - console.log(`\x1b[36m[${this.constructor.name}]\x1b[0m ${text}`, ...args); + console.log(`\x1b[36m[${this.className}]\x1b[0m ${text}`, ...args); } #appendContext(contextTexts = []) { diff --git a/server/utils/AiProviders/xai/index.js b/server/utils/AiProviders/xai/index.js index f9116270946..df925715736 100644 --- a/server/utils/AiProviders/xai/index.js +++ b/server/utils/AiProviders/xai/index.js @@ -12,6 +12,7 @@ class XAiLLM { constructor(embedder = null, modelPreference = null) { if (!process.env.XAI_LLM_API_KEY) throw new Error("No xAI API key was set."); + this.className = "XAiLLM"; const { OpenAI: OpenAIApi } = require("openai"); this.openai = new OpenAIApi({ @@ -34,7 +35,7 @@ class XAiLLM { } log(text, ...args) { - console.log(`\x1b[36m[${this.constructor.name}]\x1b[0m ${text}`, ...args); + console.log(`\x1b[36m[${this.className}]\x1b[0m ${text}`, ...args); } #appendContext(contextTexts = []) { diff --git a/server/utils/EmbeddingEngines/azureOpenAi/index.js b/server/utils/EmbeddingEngines/azureOpenAi/index.js index 57907f45fd0..79fd000d1c5 100644 --- a/server/utils/EmbeddingEngines/azureOpenAi/index.js +++ b/server/utils/EmbeddingEngines/azureOpenAi/index.js @@ -8,6 +8,7 @@ class AzureOpenAiEmbedder { if (!process.env.AZURE_OPENAI_KEY) throw new Error("No Azure API key was set."); + this.className = "AzureOpenAiEmbedder"; this.apiVersion = "2024-12-01-preview"; const openai = new AzureOpenAI({ apiKey: process.env.AZURE_OPENAI_KEY, @@ -29,7 +30,7 @@ class AzureOpenAiEmbedder { } log(text, ...args) { - console.log(`\x1b[36m[AzureOpenAiEmbedder]\x1b[0m ${text}`, ...args); + console.log(`\x1b[36m[${this.className}]\x1b[0m ${text}`, ...args); } async embedTextInput(textInput) { diff --git a/server/utils/EmbeddingEngines/gemini/index.js b/server/utils/EmbeddingEngines/gemini/index.js index 8c505466728..59eb22cb2d6 100644 --- a/server/utils/EmbeddingEngines/gemini/index.js +++ b/server/utils/EmbeddingEngines/gemini/index.js @@ -11,6 +11,7 @@ class GeminiEmbedder { if (!process.env.GEMINI_EMBEDDING_API_KEY) throw new Error("No Gemini API key was set."); + this.className = "GeminiEmbedder"; const { OpenAI: OpenAIApi } = require("openai"); this.model = process.env.EMBEDDING_MODEL_PREF || "text-embedding-004"; this.openai = new OpenAIApi({ @@ -29,7 +30,7 @@ class GeminiEmbedder { } log(text, ...args) { - console.log(`\x1b[36m[GeminiEmbedder]\x1b[0m ${text}`, ...args); + console.log(`\x1b[36m[${this.className}]\x1b[0m ${text}`, ...args); } /** diff --git a/server/utils/EmbeddingEngines/genericOpenAi/index.js b/server/utils/EmbeddingEngines/genericOpenAi/index.js index a8a3ac1a584..f8885f811c6 100644 --- a/server/utils/EmbeddingEngines/genericOpenAi/index.js +++ b/server/utils/EmbeddingEngines/genericOpenAi/index.js @@ -6,6 +6,7 @@ class GenericOpenAiEmbedder { throw new Error( "GenericOpenAI must have a valid base path to use for the api." ); + this.className = "GenericOpenAiEmbedder"; const { OpenAI: OpenAIApi } = require("openai"); this.basePath = process.env.EMBEDDING_BASE_PATH; this.openai = new OpenAIApi({ @@ -25,7 +26,7 @@ class GenericOpenAiEmbedder { } log(text, ...args) { - console.log(`\x1b[36m[GenericOpenAiEmbedder]\x1b[0m ${text}`, ...args); + console.log(`\x1b[36m[${this.className}]\x1b[0m ${text}`, ...args); } /** diff --git a/server/utils/EmbeddingEngines/lmstudio/index.js b/server/utils/EmbeddingEngines/lmstudio/index.js index e94f45aaba0..0e9a5e0ebb6 100644 --- a/server/utils/EmbeddingEngines/lmstudio/index.js +++ b/server/utils/EmbeddingEngines/lmstudio/index.js @@ -8,6 +8,7 @@ class LMStudioEmbedder { if (!process.env.EMBEDDING_MODEL_PREF) throw new Error("No embedding model was set."); + this.className = "LMStudioEmbedder"; const { OpenAI: OpenAIApi } = require("openai"); this.lmstudio = new OpenAIApi({ baseURL: parseLMStudioBasePath(process.env.EMBEDDING_BASE_PATH), @@ -21,7 +22,7 @@ class LMStudioEmbedder { } log(text, ...args) { - console.log(`\x1b[36m[${this.constructor.name}]\x1b[0m ${text}`, ...args); + console.log(`\x1b[36m[${this.className}]\x1b[0m ${text}`, ...args); } async #isAlive() { diff --git a/server/utils/EmbeddingEngines/native/index.js b/server/utils/EmbeddingEngines/native/index.js index 9142d3b3a9c..21773fcb564 100644 --- a/server/utils/EmbeddingEngines/native/index.js +++ b/server/utils/EmbeddingEngines/native/index.js @@ -30,6 +30,7 @@ class NativeEmbedder { #fallbackHost = "https://cdn.anythingllm.com/support/models/"; constructor() { + this.className = "NativeEmbedder"; this.model = this.getEmbeddingModel(); this.modelInfo = this.getEmbedderInfo(); this.cacheDir = path.resolve( @@ -50,7 +51,7 @@ class NativeEmbedder { } log(text, ...args) { - console.log(`\x1b[36m[NativeEmbedder]\x1b[0m ${text}`, ...args); + console.log(`\x1b[36m[${this.className}]\x1b[0m ${text}`, ...args); } /** diff --git a/server/utils/EmbeddingEngines/ollama/index.js b/server/utils/EmbeddingEngines/ollama/index.js index 7e2f636e259..2f7241e6a90 100644 --- a/server/utils/EmbeddingEngines/ollama/index.js +++ b/server/utils/EmbeddingEngines/ollama/index.js @@ -8,6 +8,7 @@ class OllamaEmbedder { if (!process.env.EMBEDDING_MODEL_PREF) throw new Error("No embedding model was set."); + this.className = "OllamaEmbedder"; this.basePath = process.env.EMBEDDING_BASE_PATH; this.model = process.env.EMBEDDING_MODEL_PREF; // Limit of how many strings we can process in a single pass to stay with resource or network limits @@ -20,7 +21,7 @@ class OllamaEmbedder { } log(text, ...args) { - console.log(`\x1b[36m[${this.constructor.name}]\x1b[0m ${text}`, ...args); + console.log(`\x1b[36m[${this.className}]\x1b[0m ${text}`, ...args); } /** diff --git a/server/utils/EmbeddingEngines/openAi/index.js b/server/utils/EmbeddingEngines/openAi/index.js index 9976ef54d14..dd517ae3030 100644 --- a/server/utils/EmbeddingEngines/openAi/index.js +++ b/server/utils/EmbeddingEngines/openAi/index.js @@ -3,6 +3,7 @@ const { toChunks } = require("../../helpers"); class OpenAiEmbedder { constructor() { if (!process.env.OPEN_AI_KEY) throw new Error("No OpenAI API key was set."); + this.className = "OpenAiEmbedder"; const { OpenAI: OpenAIApi } = require("openai"); this.openai = new OpenAIApi({ apiKey: process.env.OPEN_AI_KEY, @@ -17,7 +18,7 @@ class OpenAiEmbedder { } log(text, ...args) { - console.log(`\x1b[36m[OpenAiEmbedder]\x1b[0m ${text}`, ...args); + console.log(`\x1b[36m[${this.className}]\x1b[0m ${text}`, ...args); } async embedTextInput(textInput) { diff --git a/server/utils/MCP/hypervisor/index.js b/server/utils/MCP/hypervisor/index.js index 0cd1f6cecbb..861274f20b7 100644 --- a/server/utils/MCP/hypervisor/index.js +++ b/server/utils/MCP/hypervisor/index.js @@ -53,6 +53,7 @@ class MCPHypervisor { constructor() { if (MCPHypervisor._instance) return MCPHypervisor._instance; MCPHypervisor._instance = this; + this.className = "MCPHypervisor"; this.log("Initializing MCP Hypervisor - subsequent calls will boot faster"); this.#setupConfigFile(); return this; @@ -88,7 +89,7 @@ class MCPHypervisor { } log(text, ...args) { - console.log(`\x1b[36m[${this.constructor.name}]\x1b[0m ${text}`, ...args); + console.log(`\x1b[36m[${this.className}]\x1b[0m ${text}`, ...args); } /** diff --git a/server/utils/agents/aibitat/plugins/sql-agent/SQLConnectors/MSSQL.js b/server/utils/agents/aibitat/plugins/sql-agent/SQLConnectors/MSSQL.js index d11314f6f26..584e2f638c4 100644 --- a/server/utils/agents/aibitat/plugins/sql-agent/SQLConnectors/MSSQL.js +++ b/server/utils/agents/aibitat/plugins/sql-agent/SQLConnectors/MSSQL.js @@ -28,6 +28,7 @@ class MSSQLConnector { connectionString: null, // we will force into RFC-3986 } ) { + this.className = "MSSQLConnector"; this.connectionString = config.connectionString; this._client = null; this.#parseDatabase(); @@ -72,7 +73,7 @@ class MSSQLConnector { result.rows = query.recordset; result.count = query.rowsAffected.reduce((sum, a) => sum + a, 0); } catch (err) { - console.log(this.constructor.name, err); + console.log(this.className, err); result.error = err.message; } finally { // Check client is connected before closing since we use this for validation diff --git a/server/utils/agents/aibitat/plugins/sql-agent/SQLConnectors/MySQL.js b/server/utils/agents/aibitat/plugins/sql-agent/SQLConnectors/MySQL.js index 7a566cb725a..7fa4c6a512b 100644 --- a/server/utils/agents/aibitat/plugins/sql-agent/SQLConnectors/MySQL.js +++ b/server/utils/agents/aibitat/plugins/sql-agent/SQLConnectors/MySQL.js @@ -9,6 +9,7 @@ class MySQLConnector { connectionString: null, } ) { + this.className = "MySQLConnector"; this.connectionString = config.connectionString; this._client = null; this.database_id = this.#parseDatabase(); @@ -39,7 +40,7 @@ class MySQLConnector { result.rows = query; result.count = query?.length; } catch (err) { - console.log(this.constructor.name, err); + console.log(this.className, err); result.error = err.message; } finally { // Check client is connected before closing since we use this for validation diff --git a/server/utils/agents/aibitat/plugins/sql-agent/SQLConnectors/Postgresql.js b/server/utils/agents/aibitat/plugins/sql-agent/SQLConnectors/Postgresql.js index 15ca6dde428..d77c1bf524e 100644 --- a/server/utils/agents/aibitat/plugins/sql-agent/SQLConnectors/Postgresql.js +++ b/server/utils/agents/aibitat/plugins/sql-agent/SQLConnectors/Postgresql.js @@ -8,6 +8,7 @@ class PostgresSQLConnector { schema: null, } ) { + this.className = "PostgresSQLConnector"; this.connectionString = config.connectionString; this.schema = config.schema || "public"; this._client = new pgSql.Client({ @@ -34,7 +35,7 @@ class PostgresSQLConnector { result.rows = query.rows; result.count = query.rowCount; } catch (err) { - console.log(this.constructor.name, err); + console.log(this.className, err); result.error = err.message; } finally { // Check client is connected before closing since we use this for validation diff --git a/server/utils/agents/aibitat/providers/gemini.js b/server/utils/agents/aibitat/providers/gemini.js index c62357bb6d3..d624bd64f21 100644 --- a/server/utils/agents/aibitat/providers/gemini.js +++ b/server/utils/agents/aibitat/providers/gemini.js @@ -17,6 +17,7 @@ class GeminiProvider extends InheritMultiple([Provider, UnTooled]) { constructor(config = {}) { const { model = "gemini-2.0-flash-lite" } = config; super(); + this.className = "GeminiProvider"; const client = new OpenAI({ baseURL: "https://generativelanguage.googleapis.com/v1beta/openai/", apiKey: process.env.GEMINI_API_KEY, @@ -134,7 +135,7 @@ class GeminiProvider extends InheritMultiple([Provider, UnTooled]) { } catch (error) { throw new APIError( error?.message - ? `${this.constructor.name} encountered an error while executing the request: ${error.message}` + ? `${this.className} encountered an error while executing the request: ${error.message}` : "There was an error with the Gemini provider executing the request" ); }