这是indexloc提供的服务,不要输入任何密码
Skip to content

Conversation

@krrishdholakia
Copy link

Hi @varunvummadi,

Notice you forked chat-ui. if you're trying to test other LLMs - besides GigaML's finetuned ones - (codellama, wizardcoder, etc.) with it, I just wrote a 1-click proxy to translate openai calls to huggingface, anthropic, togetherai, etc. api calls.

code

$ pip install litellm

$ litellm --model huggingface/bigcode/starcoder

#INFO:     Uvicorn running on http://0.0.0.0:8000

>> openai.api_base = "http://0.0.0.0:8000"

Here's the PR on adding openai to chat-ui: huggingface#452

I'd love to know if this solves a problem for you

nsarrazin and others added 21 commits July 12, 2023 07:51
* add optional timestamp field to messages

* Add a `hashConv` function that only uses a subset of the message for hashing
* Add ability to define custom model/dataset URLs

* lint

---------

Co-authored-by: Nathan Sarrazin <sarrazin.nathan@gmail.com>
* Update README.md

* Update README.md

Co-authored-by: Julien Chaumond <julien@huggingface.co>

* Align with header

* lint

* fixed markdown table of content

---------

Co-authored-by: Julien Chaumond <julien@huggingface.co>
Co-authored-by: Nathan Sarrazin <sarrazin.nathan@gmail.com>
* disable login on first message

* update banner here too

* modal wording tweaks

* prevent NaN

---------

Co-authored-by: Victor Mustar <victor.mustar@gmail.com>
* disable login on first message

* update banner here too

* modal wording tweaks

* prevent NaN

* fix login wall

* fix flicker

* lint

* put modal text behind login check

* fix bug with sending messages without login

* fix misalignment between ui and api

* fix data update on disable login

---------

Co-authored-by: Nathan Sarrazin <sarrazin.nathan@gmail.com>
* Update README.md

* Update README.md

Co-authored-by: Julien Chaumond <julien@huggingface.co>

* Update README.md

---------

Co-authored-by: Julien Chaumond <julien@huggingface.co>
The userMessageToken, assistantMessageToken, messageEndToken, and
parameters.stop settings in `MODELS` do not have to be a token. They can
be any string.
* rm open assistant branding

* Update SettingsModal.svelte

* make settings  work with a dynamic list of models

* fixed types

---------

Co-authored-by: Nathan Sarrazin <sarrazin.nathan@gmail.com>
The chat generation removes parameters.stop and <|endoftext|>
from the generated text. And additionally trims trailing whitespace.

This PR copies that behavior to the summarize functionality, when the
summary is produced by a the chat model.
* allow different user and assistant end-token

For models like Llama2, the EndToken is not the same for a userMessage
and an assistantMessage. This implements `userMessageEndToken` and
`assistantMessageEndToken` which overwrites the messageEndToken
behavior.

This PR also allows empty strings as userMessageToken and
assistantMessageToken and makes this the default. This adds additional
flexibility, which is required in the case of Llama2 where the first
userMessage is effectively different because of the system message.

Note that because `userMessageEndToken` and `assistantMessageToken` are
nearly always concatenated, it is almost redundant to have both. The
exception is `generateQuery` for websearch which have several
consecutive user messages.

* Make model branding customizable based on env var (huggingface#345)

* rm open assistant branding

* Update SettingsModal.svelte

* make settings  work with a dynamic list of models

* fixed types

---------

Co-authored-by: Nathan Sarrazin <sarrazin.nathan@gmail.com>

* trim and remove stop-suffixes from summary (huggingface#369)

The chat generation removes parameters.stop and <|endoftext|>
from the generated text. And additionally trims trailing whitespace.

This PR copies that behavior to the summarize functionality, when the
summary is produced by a the chat model.

* add a login button when users are logged out (huggingface#381)

* add fallback to message end token if there's no specified tokens for user & assistant

---------

Co-authored-by: Florian Zimmermeister <flozi00.fz@gmail.com>
Co-authored-by: Nathan Sarrazin <sarrazin.nathan@gmail.com>
* Use modelUrl instead of building it from model name

* Preserve compatibility with optional modelUrl config

Use modelUrl if informed, else use the previous behavior.
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

10 participants