Hello,
I am trying to reproduce benchmarking results, as explained in https://github.com/dmlc/tvm/tree/master/apps/benchmark. I have the latest TVM source code and am deploying to a Raspberry Pi 3B.
When I run the benchmark using
python3 arm_cpu_imagenet_bench.py --device rasp3b --rpc-key rasp3b
I observe the following runtimes:
--------------------------------------------------
Network Name Mean Inference Time (std dev)
--------------------------------------------------
squeezenet v1.1 204.01 ms (1.97 ms)
mobilenet 412.53 ms (79.38 ms)
resnet-18 775.99 ms (46.59 ms)
These appear noticeably slower than the ones reported on the repo page (shown below for reference):
--------------------------------------------------
Network Name Mean Inference Time (std dev)
--------------------------------------------------
squeezenet v1.1 92.34 ms (0.07 ms)
mobilenet 145.22 ms (0.11 ms)
resnet-18 325.06 ms (0.23 ms)
I’ve tried two Raspberry Pi 3Bs and two different host CPUs but cannot reproduce the reported results. Has anyone encountered this?