I’m having some problems when deploying this model on x86 architecture using relay.
When setting opt_level=3
performance get worse by almost 3 times compared to using opt_level=1
.
Any clue why does this happen?
I’ve also tuned convolution on x86 using autotvm
, but inference is still quite slow. Is there any other way I can further speed up the inference?