-
-
Notifications
You must be signed in to change notification settings - Fork 5.4k
[Feature] AnythingLLM use locally hosted Llama.cpp and GGUF files for inferencing #413
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
stop showing prisma queries during dev
…-llm into built-in-llm
hide localLlama when on hosted
|
Is GPU Acceleration possible for this feature? |
node-llama-cpp using CUDA is supported And once the node-llama-cpp binaries are built it will use GPU acceleration for the llama client, however for docker instances this is CPU only so some additional work will need to be done to bind GPUs to the instance so they can be used. |
It is possible to bind the GPU: As I use WSL, I would need to install the toolkit too, I will see how this plays out later on. Thanks. |
update dev comment on token model size
…ain, just use the app!
… inferencing (Mintplex-Labs#413) * Implement use of native embedder (all-Mini-L6-v2) stop showing prisma queries during dev * Add native embedder as an available embedder selection * wrap model loader in try/catch * print progress on download * add built-in LLM support (expiermental) * Update to progress output for embedder * move embedder selection options to component * saftey checks for modelfile * update ref * Hide selection when on hosted subdomain * update documentation hide localLlama when on hosted * saftey checks for storage of models * update dockerfile to pre-build Llama.cpp bindings * update lockfile * add langchain doc comment * remove extraneous --no-metal option * Show data handling for private LLM * persist model in memory for N+1 chats * update import update dev comment on token model size * update primary README * chore: more readme updates and remove screenshots - too much to maintain, just use the app! * remove screeshot link
Uh oh!
There was an error while loading. Please reload this page.