Introduction to the inference server project using the ONNX Runtime #17496
kibae
started this conversation in
Show & Tell
Replies: 0 comments
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Hi, all 🤗
Thank you for creating and maintaining the ONNX Runtime.
I would like to introduce my personal project using ONNX Runtime C++ API.
https://github.com/kibae/onnxruntime-server
The onnxruntime-server can easily provide an inference API for ONNX files. It provides TCP, HTTP/HTTPS REST API backend and has built-in swagger API documentation. It also supports hardware acceleration using CUDA.
Just copy the ONNX files to the model directory of onnxruntime-server and it provides onnx session management and execution without any code.
I hope you're interested and if you have any questions, feel free to let me know!
Beta Was this translation helpful? Give feedback.
All reactions