A general framework for inferring CNNs efficiently. Reduce the inference latency of MobileNet-V3 by 1.3x on an iPhone XS Max without sacrificing accuracy.
Thanks for your great work which inspires me a lot.
I understand the process of budgeted batch classification, you set a threshold to every classifier, so you can calcuate the FLOPs via different stages with a rate. But in the baseline in figure4 just like MobileNet-V3,EfficientNet, how do you get the result of multi flops with different accuracy. You esemble them with early-exit? or other settings? Thank you.
I am interested in your project and your paper.
I find flops = checkpoint['flops'] in inference.py. But when I train a new model by myself, I couldn't find 'flops' during the process of saving checkpoint.
I think I can set this parameter manually. But what should I based on, if I want to set this parameter manually?
Looking forward to your reply.
I find you have dynamic_threshold = checkpoint['dynamic_threshold '] in inference.py and this code is necessary when eval_mode=1. But where does 'dynamic_threshold' come from? You did not save this value when you saved you checkpoint.
very great work,great paper ,l have a question just is for example,in first layer Classifier fc,l know fc is RNN,but l dont know what is the fc to next fc pass through h1C,thanks,vevy nice!
Hello,
Thanks for your great work, it's a really fascinating
Can you release the code for visualization ? I mean how can we see the sequence of patches in network ?