-
Notifications
You must be signed in to change notification settings - Fork 12
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Error encountered when using metricx with batch_size > 1 #2
Comments
Hi, It seems there is a padding issue occurring due to variations in the lengths of input sequences within the same batch size. To resolve this, all input sequences within a batch should have the same length. In the get_dataset() function in predict.py, I've added a new function called _pad. This function is responsible for filling input sequences with padding tokens to make them all the same length, equal to max_input_length `def get_dataset(
` |
Hi @nicolasdahan, |
@nicolasdahan @samirsalman
If you need to pad to
|
Issue Description:
Hi there,
I'm currently encountering an error while utilizing metricx, specifically when the batch_size parameter is set to a value greater than 1. Below, I've outlined the command I'm using along with the error message received:
Command:
Error Message:
It seems that the error occurs due to tensor size discrepancies within the specified batch when batch_size is set higher than 1.
Request for Assistance:
Could you please provide guidance on resolving this issue? Any suggestions or insights would be greatly appreciated.
Thank you!
The text was updated successfully, but these errors were encountered: