Is there any documentation available on how to perform inference with an ONNX model in C++?
Yes, TVM does have interfaces for different languages. Most of the components of TVM are implemented in C++. I think a better solution is
- Import, convert, and compile the ONNX model in advance with Python. Now you will get a runtime executable.
- Then you can deploy the executable on the target device with TVM runtime (without Python).
You can refer to https://github.com/apache/tvm/tree/main/apps/bundle_deploy
1 Like
right, python is common used for compilation and deployment happens at the language of your choice