Beyond Network Architecture Search (NAS): Once-For-All network for train once and deploy on FPGAs

Beyond Network Architecture Search (NAS): Once-For-All network for train once and deploy on FPGAs

Thursday, September 23, 2021 10:20 AM to 10:45 AM · 25 min. (Europe/London)
Presentation
POC to Production – Headline Stage

Information

We address the challenging problem of efficient inference across many devices and resource constraints. Conventional approaches either manually design or use neural architecture search (NAS) to find a specialized neural network and train it from scratch for each case, which is computationally prohibitive (causing CO2 emission as much as 5 cars' lifetime) thus unscalable. In this work, we propose to train a once-for-all (OFA) network that supports diverse architectural settings by decoupling training and search, to reduce the cost. We can quickly get a specialized sub-network by selecting from the OFA network without additional training. OFA consistently outperforms state-of-the-art (SOTA) NAS methods (up to 4.0% ImageNet top1 accuracy improvement over MobileNetV3, or same accuracy but 1.5x faster than MobileNetV3, 2.6x faster than EfficientNet w.r.t measured latency) while reducing many orders of magnitude GPU hours and CO2 emission.  

Day
Day 2
Virtual session?
Yes

Log in