这是indexloc提供的服务,不要输入任何密码
Skip to content

[BUG]: llama3.1 8B Context Size Max Tokens Ignored in Both Performance Modes #2442

@rurhrlaub

Description

@rurhrlaub

How are you running AnythingLLM?

AnythingLLM desktop app

What happened?

anythingfllm_context

When using "Base" as the "Performance Mode", the Max Tokens setting is ignored and Llama 3.1 is invoked with 8K context size. When setting Performance Mode to "Maximum", the Max Tokens settings is ignored and Llama 3.1 is invoked with 128K context size. Created a modelfile to enforce 32K context size but the result was 128K. Workspace was set to use the system defined LLM settings.

Are there known steps to reproduce?

See above

Metadata

Metadata

Labels

core-team-onlyinvestigatingCore team or maintainer will or is currently looking into this issueneeds info / can't replicateIssues that require additional information and/or cannot currently be replicated, but possible bugpossible bugBug was reported but is not confirmed or is unable to be replicated.

Type

No type

Projects

No projects

Milestone

No milestone

Relationships

None yet

Development

No branches or pull requests

Issue actions