Algorithm-Hardware Co-Design for Energy-Efficient Neural Network Inference
Paul Whatmough ( Arm ML Research Lab Boston / Harvard University )
- 13:00 7th March 2019 ( week 8, Hilary Term 2019 )Lecture Theatre B, Wolfson Building
Deep neural networks (DNNs) have quickly become an essential workload across computing form factors, including IoT, mobile, automotive and datacenter. However, DNN inference demands an enormous number of arithmetic operations and a large memory footprint. In this talk, we will explore the co-design of DNN models and hardware to achieve state-of-the-art performance for real-time, energy-constrained inference applications.