-
-
Notifications
You must be signed in to change notification settings - Fork 5.4k
Description
How are you running AnythingLLM?
Docker (local)
What happened?
When selecting the first "Gemini Experimental 1206" from the Chat Model Selection list (there are six overall on the list), It's apparently using gemini-1.5-pro-exp-0801, not 1206, since I'm getting this error:
Could not respond to message.
[GoogleGenerativeAI Error]: Error fetching from https://generativelanguage.googleapis.com/v1beta/models/gemini-1.5-pro-exp-0801:streamGenerateContent?alt=sse: [500 Internal Server Error] An internal error has occurred. Please retry or report in https://developers.generativeai.google/guide/troubleshooting
When I select the correct "Gemini Experimental 1206" (the forth one top to bottom on the list) and try to send a prompt message, I get the same error just with the correct model.
Could not respond to message.
[GoogleGenerativeAI Error]: Error fetching from https://generativelanguage.googleapis.com/v1beta/models/gemini-exp-1206:streamGenerateContent?alt=sse: [500 Internal Server Error] An internal error has occurred. Please retry or report in https://developers.generativeai.google/guide/troubleshooting
Sucks that I've apparently been using 0801all this time before this error started, at least now I know.
Edit: The 500 Internal Server Error is gone now, the model names bug is still there though. I also have a few more comments:
-
The metrics say the LLM's reply takes a few seconds but it actually takes a few minutes to stop loading and show up sometimes, why is that?
-
Gemini 2.0 Flash Thinking Experimental is showing the thinking process in the reply and gets broken easily, using wrong context for its replies.