Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Usage]: How to make model response information appear in the vllm backend logs #10602

Open
1 task done
nora647 opened this issue Nov 24, 2024 · 1 comment
Open
1 task done
Labels
usage How to use vllm

Comments

@nora647
Copy link

nora647 commented Nov 24, 2024

Your current environment

I am deploying a qwen2.5-7b model via vllm serve Qwen/Qwen2.5-7B-Instruct with api calls using the langchain framework.

vllm serve Qwen/Qwen2.5-7B-Instruct

But in the process of using it, I found that the vllm background log only records the log of requests to llm, and does not record the log of llm replies, which is very inconvenient in program debugging.
1db554ad959899f82124ab9cd80d923

I would like to know how to make vllm background log can record model reply information in real time.For example, as shown in the following figure
af26d8187b66d9c6eac2abba5cccf81

How would you like to use vllm

I would like to know how to make vllm background log can record model reply information in real time.For example, as shown in the following figure.Thank you!

af26d8187b66d9c6eac2abba5cccf81

Before submitting a new issue...

  • Make sure you already searched for relevant issues, and asked the chatbot living at the bottom right corner of the documentation page, which can answer lots of frequently asked questions.
@nora647 nora647 added the usage How to use vllm label Nov 24, 2024
@DarkLight1337
Copy link
Member

DarkLight1337 commented Nov 25, 2024

The output isn't logged in our code, you'll have to edit chat_completion_full_generator and/or chat_completion_stream_generator inside vllm/entrypoints/openai/serving_chat.py to log this information.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
usage How to use vllm
Projects
None yet
Development

No branches or pull requests

2 participants