[Usage]: A problem when use llm.generate() for several times in one LLM case #12651
Open
1 task done
Labels
usage
How to use vllm
Your current environment
How would you like to use vllm
I' m trying to test the performance of vllm so I need to test the time need for vllm when the input and output tokens numbers are fixed. Here I tried to run each situation for 20 times and get the average. But I got a problems that the answer returned after the first round become very strange(they are not nature language any more). It seems that the returns after the first round were interfered. What' s the problem here, what should I do to solve the problem? Thank!
### Here's my code:
### And the first and second returns
Before submitting a new issue...
The text was updated successfully, but these errors were encountered: