From 7899b71111771bd5654ffa89d0bf3432cd90ac62 Mon Sep 17 00:00:00 2001 From: angelplusultra Date: Wed, 1 Oct 2025 10:39:17 -0700 Subject: [PATCH 1/2] Add stream options to Gemini LLM for usage tracking --- server/utils/AiProviders/gemini/index.js | 3 +++ 1 file changed, 3 insertions(+) diff --git a/server/utils/AiProviders/gemini/index.js b/server/utils/AiProviders/gemini/index.js index de76ec79c82..66158302eea 100644 --- a/server/utils/AiProviders/gemini/index.js +++ b/server/utils/AiProviders/gemini/index.js @@ -416,6 +416,9 @@ class GeminiLLM { stream: true, messages, temperature: temperature, + stream_options: { + include_usage: true, + }, }), messages, true From 2775657311eaf878c8d0f18f1b391eee2db80625 Mon Sep 17 00:00:00 2001 From: angelplusultra Date: Wed, 1 Oct 2025 13:15:18 -0700 Subject: [PATCH 2/2] Update Gemini LLM to disable prompt token calculation --- server/utils/AiProviders/gemini/index.js | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/server/utils/AiProviders/gemini/index.js b/server/utils/AiProviders/gemini/index.js index 66158302eea..4c352557b54 100644 --- a/server/utils/AiProviders/gemini/index.js +++ b/server/utils/AiProviders/gemini/index.js @@ -421,7 +421,7 @@ class GeminiLLM { }, }), messages, - true + false ); return measuredStreamRequest;