Skip to content

ONNX Runtime v0.5.0

Compare
Choose a tag to compare
@hariharans29 hariharans29 released this 01 Aug 19:48
1f8019b
  • Execution Provider updates
    • MKL-DNN provider (subgraph based execution) for improved performance
    • Intel OpenVINO EP now available for Public Preview - build instructions
    • Update to CUDA 10 for inferencing with NVIDIA GPUs
    • Base CPU EP has faster convolution performance using the NCHWc blocked layout. This layout optimization can be enabled by setting graph optimization level to 3 in the session options.
  • C++ API for inferencing (wrapper on C API)
  • ONNX Runtime Server (Beta) for inferencing with HTTP and GRPC endpoints
  • Python Operator (Beta) to support custom Python code in a single node of an ONNX graph to make it easier for experimentation of custom operators
  • Support of Keras-based Mask R-CNN model. The model relies on some custom operators pending to be added in ONNX; in the meantime, it can be converted using this script for inferencing using ONNX Runtime 0.5. Other object detection models can be found from the ONNX Model Zoo.
  • Minor updates to the C API
    • For consistency, all C APIs now return an ORT status code
  • Code coverage for this release is 83%