Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

val_data_size and test_data_size have no effect. #252

Open
hyeok9855 opened this issue Feb 14, 2025 · 3 comments
Open

val_data_size and test_data_size have no effect. #252

hyeok9855 opened this issue Feb 14, 2025 · 3 comments

Comments

@hyeok9855
Copy link
Contributor

The configs val_data_size and test_data_size (see here) have no effect. It seems that the default env.val_file and env.test_file are used.

You can simply check this by

  1. change those configs to some value
  2. python run.py with the changed config file.
@fedebotu
Copy link
Member

Right, that is working as intended though.

There are two cases:

  1. User provides val and/or test files. Then, we should use them
  2. User does not provide them OR they cannot be loaded for some reason. Then, we should generate a validation and test dataset

I agree though that it can be confusing 🤔

How about having a log.warning that says something , e.g. during val: "env.val_file has been loaded. val_data_size will have no effect"? Or do you have better ideas?

@hyeok9855
Copy link
Contributor Author

How about slicing the dataset according to the size when both val/test_file and val/test_data_size are provided?

@fedebotu
Copy link
Member

That sounds like a good idea!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants