这是indexloc提供的服务,不要输入任何密码
Skip to content

[BUG]: When using KoboldCPP, the max response length is always 512, regardless of context size. #3708

@SafentisFox

Description

@SafentisFox

How are you running AnythingLLM?

AnythingLLM desktop app

What happened?

I'm attempting to use AnythingLLM with KoboldCPP as the provider. No matter what I tried, the max response length is always 512 tokens. No matter the context size.

This, combined with the lack of a "Continue" button makes certain tasks hard, annoying or borderline impossible.

Are there known steps to reproduce?

Connect KoboldCPP. Do any kind of chat. Look at the KoboldCPP console to confirm that the requested response has a maximum of 512 tokens.

Metadata

Metadata

Assignees

Labels

investigatingCore team or maintainer will or is currently looking into this issuepossible bugBug was reported but is not confirmed or is unable to be replicated.

Type

No type

Projects

No projects

Milestone

No milestone

Relationships

None yet

Development

No branches or pull requests

Issue actions