-
-
Notifications
You must be signed in to change notification settings - Fork 5.4k
Closed
Labels
investigatingCore team or maintainer will or is currently looking into this issueCore team or maintainer will or is currently looking into this issuepossible bugBug was reported but is not confirmed or is unable to be replicated.Bug was reported but is not confirmed or is unable to be replicated.
Description
How are you running AnythingLLM?
AnythingLLM desktop app
What happened?
I'm attempting to use AnythingLLM with KoboldCPP as the provider. No matter what I tried, the max response length is always 512 tokens. No matter the context size.
This, combined with the lack of a "Continue" button makes certain tasks hard, annoying or borderline impossible.
Are there known steps to reproduce?
Connect KoboldCPP. Do any kind of chat. Look at the KoboldCPP console to confirm that the requested response has a maximum of 512 tokens.
Metadata
Metadata
Assignees
Labels
investigatingCore team or maintainer will or is currently looking into this issueCore team or maintainer will or is currently looking into this issuepossible bugBug was reported but is not confirmed or is unable to be replicated.Bug was reported but is not confirmed or is unable to be replicated.