Hello everyone,
I am wondering about TVM’s ability to fold batch norm layers. As a refenrece, I will use the equations as shown here.
I know that TVM will constant fold the values in (1) which are to be multiplied/added with the output of the bias addition. In other words, there will be no square root nor division during inference. What I am wondering is can TVM do (4), (5) and (6). In other words, fold the batch norm onto the weights and biases.
If so, what is the pass that does this?
Thanks a lot