Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[FEA]: Remove/Deprecate the model_max_batch_size config option #421

Open
2 tasks done
mdemoret-nv opened this issue Oct 27, 2022 · 0 comments
Open
2 tasks done

[FEA]: Remove/Deprecate the model_max_batch_size config option #421

mdemoret-nv opened this issue Oct 27, 2022 · 0 comments
Labels
feature request New feature or request

Comments

@mdemoret-nv
Copy link
Contributor

Is this a new feature, an improvement, or a change to existing functionality?

Improvement

How would you describe the priority of this feature request

Medium

Please provide a clear description of problem this feature solves

As mentioned in issue #420, the different options for batch size, model_max_batch_size and pipeline_max_batch_size, can be confusing to users and it's not clear how they interact or impact performance. This config option is a legacy value from one of the first iterations of Morpheus where multiple stages would need this option to coordinate the size of messages. Since it's only used by one stage (Inference), it does not make sense to be a config option any more.

Describe your ideal solution

The model_max_batch_size option should be removed/deprecated. Where it is still needed, (on the InferenceStage implementations), we can automatically determine the max batch size either from the model or the service. For example, the TritonInferenceStage can determine the model_max_batch_size during the initialization step.

To allow for backward compatibility, we could add a model_max_batch_size property to the InferenceStage itself to override any automatically determined value.

Describe any alternatives you have considered

No response

Additional context

No response

Code of Conduct

  • I agree to follow this project's Code of Conduct
  • I have searched the open feature requests and have found no duplicates for this feature request
@mdemoret-nv mdemoret-nv added the feature request New feature or request label Oct 27, 2022
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
feature request New feature or request
Projects
Status: Todo
Development

No branches or pull requests

1 participant