[Usage]: How to make model response information appear in the vllm backend logs #10602
Open
1 task done
Labels
usage
How to use vllm
Your current environment
I am deploying a qwen2.5-7b model via vllm serve Qwen/Qwen2.5-7B-Instruct with api calls using the langchain framework.
But in the process of using it, I found that the vllm background log only records the log of requests to llm, and does not record the log of llm replies, which is very inconvenient in program debugging.
I would like to know how to make vllm background log can record model reply information in real time.For example, as shown in the following figure
How would you like to use vllm
I would like to know how to make vllm background log can record model reply information in real time.For example, as shown in the following figure.Thank you!
Before submitting a new issue...
The text was updated successfully, but these errors were encountered: