The TensorRT execution we use in TVM is not asynchronous, so there is no need to sync. module.run() won’t return until inference is completed. Actually I think run() is never asynchronous in TVM?
5ms is not an unreasonable inference time for mobilenet v2 with TensorRT on xavier, although I am getting around 10ms. But your model may be different.