Trainning Standard ONNX models #5969
Replies: 2 comments 1 reply
-
Hello, thank you for trying out the ONNX Runtime Training feature. Currently ONNX Runtime Training is designed to accelerate distributed training for PyTorch models in a multi-GPU training environment. There is also a limitation in the ONNX spec with no operators available for training. However, we would love to learn more about your use case of training standalone ONNX models. Please share any additional context. We also encourage you to try ORT Training for Transformer based models in PyTorch. |
Beta Was this translation helpful? Give feedback.
-
Hello, thank you for the additional context, this was helpful in understanding the scenario better. Regarding the limitation of the ONNX spec, please raise an issue in the ONNX repo for operator support for training. Regarding your question around the API to run the ONNX model and get its outputs, currently we do not support such an option and the faster route would be the ONNX spec update for training. |
Beta Was this translation helpful? Give feedback.
-
Hello, I couldn't find a way to run a standalone training ONNX graph using its embedded training information, is there a way to do it? I could only find examples focused on running Pytorch models.
The hacky workaround I'm using here is configuring an ORTTrainer and accessing the private member
trainer._training_session.get_state()
to get the updated parameters.I'm wondering if there is a better way to run a trainable ONNX model and gets it's outputs (updated weights + loss), preferably using the embedded training and Gradient operator so have more control over the training loop.
Thank you for the attention.
Beta Was this translation helpful? Give feedback.
All reactions