Towards Energy-Efficient Convolutional Neural Network Inference von Lukas Arno Jakob Cavigelli | ISBN 9783866286511

Towards Energy-Efficient Convolutional Neural Network Inference

von Lukas Arno Jakob Cavigelli, herausgegeben von Qiuting Huang, Andreas Schenk, Mathieu Maurice Luisier und Bernd Witzigmann
Mitwirkende
Autor / AutorinLukas Arno Jakob Cavigelli
Herausgegeben vonQiuting Huang
Herausgegeben vonAndreas Schenk
Herausgegeben vonMathieu Maurice Luisier
Herausgegeben vonBernd Witzigmann
Buchcover Towards Energy-Efficient Convolutional Neural Network Inference | Lukas Arno Jakob Cavigelli | EAN 9783866286511 | ISBN 3-86628-651-1 | ISBN 978-3-86628-651-1

Towards Energy-Efficient Convolutional Neural Network Inference

von Lukas Arno Jakob Cavigelli, herausgegeben von Qiuting Huang, Andreas Schenk, Mathieu Maurice Luisier und Bernd Witzigmann
Mitwirkende
Autor / AutorinLukas Arno Jakob Cavigelli
Herausgegeben vonQiuting Huang
Herausgegeben vonAndreas Schenk
Herausgegeben vonMathieu Maurice Luisier
Herausgegeben vonBernd Witzigmann
Deep learning and particularly convolutional neural networks (CNNs) have become the method of choice for most computer vision tasks. The achieved leap in accuracy has dramatically increased the range of possibilities and created a demand for running these compute and memory intensive algorithms on embedded and mobile devices. In this thesis, we evaluate the capabilities of software-programmable hardware, dive into specialized accelerators, and explore the potential of extremely quantized CNNs—all with special consideration to external memory bandwidth, which dominates the overall energy cost. We establish that—including I/O—software-programmable platforms can achieve 10–40 GOp/s/W, our specialized accelerator for fixedpoint CNNs achieves 630 GOp/s/W, binary-weight CNNs can be implemented with up to 5.9 TOp/s/W and very small binarized neural networks implementable with purely combinational logic could be run directly on the sensor with 670 TOp/s/W.