Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Web GUI for longchat #12

Open
VVNMA opened this issue Jul 4, 2023 · 3 comments
Open

Web GUI for longchat #12

VVNMA opened this issue Jul 4, 2023 · 3 comments

Comments

@VVNMA
Copy link

VVNMA commented Jul 4, 2023

Hi, thanks for the great job!
But I'd like to know is there a web GUI for this longchat model, it seems that the model worker in Vicuna does not support this model directly, there should be a condense ratio process.

@DachengLi1
Copy link
Owner

@VVNMA Thanks for trying out! LongChat is compatible in FastChat that we have implemented the monkey patch in the load_model function. The model worker is using load_model so it should work. Let me know if it does not work and we will take time to fix it!

@VVNMA
Copy link
Author

VVNMA commented Jul 5, 2023

@DachengLi1 Thanks. I saw the replace in the model adapter. But for the gradio server, I have to modify the input and max output token to fit the long chat.

@jalajthanaki
Copy link

How can I see an option on UI to increase the context length from 1024 to 16K?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants