-
Notifications
You must be signed in to change notification settings - Fork 110
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Error when running triton server with whisper model #522
Comments
I got same issue. but it work properly.
|
Could you have a look at this issue? |
@jwkyeongzz You mean using V100 is good. The issue only happened with RTX3090 GPU ? |
@jackNhat May I ask what's your GPU's name? Also, would you mind attaching more details? e.g. how to reproduce the error. |
I thought the test environment might be the problem. At first, since the error environment was in Windows' virtual Ubuntu 20.04, it was assumed that there was a problem with cuda memory allocation. In addition, it seems that it may have occurred due to insufficient memory of the RTX 3090. Therefore, it seems that the rtx3090 is not necessarily the problem. |
When i ran client.py, i got errror message:
tritonclient.utils.InferenceServerException: [StatusCode.INTERNAL] in ensemble 'whisper', Failed to process the request(s) for model instance 'scorer_0', message: AssertionError: <EMPTY MESSAGE>
How to fix?
I ran triton server with whisper model verson large-v2
The text was updated successfully, but these errors were encountered: