Redundancy in Deep Neural Networks and Its Impacts to Hardware Accelerator Design
- π€ Speaker: Jiang Su, Imperial College London
- π Date & Time: Monday 24 July 2017, 11:00 - 12:00
- π Venue: SW01, Computer Laboratory
Abstract
Hardware systems for neural networks are limited in their applicability to power-constrained hardware environments as they are both highly compute and memory intensive. As a result, model-level redundancy approaches such as dropout, pruning and parameter compression have been proposed to increase classification accuracy and/or lower hardware complexity. Additionally, significant data-level redundancy of the weight parameters has been consistently shown to produce comparable classification accuracy to their floating point equivalent models. As a result, thereβs recently been a growing interest in networks with low-precision weight representations, especially ones with only 1 or 2 bits. Such computational structures significantly reduce the compute, spatial complexity, and memory footprint which ultimately improves their applicability to power-constrained application scenarios.
In this talk, these two levels of redundancy are introduced as well as their impacts to hardware system design. Some personal opinions about efficient deep neural network acceleration system design are finally proposed for more open discussion.
Series This talk is part of the Computer Laboratory Computer Architecture Group Meeting series.
Included in Lists
- All Talks (aka the CURE list)
- bld31
- Cambridge talks
- Computer Laboratory Computer Architecture Group Meeting
- Department of Computer Science and Technology talks and seminars
- Interested Talks
- School of Technology
- SW01, Computer Laboratory
- Trust & Technology Initiative - interesting events
- yk449
Note: Ex-directory lists are not shown.
![[Talks.cam]](/static/images/talkslogosmall.gif)


Monday 24 July 2017, 11:00-12:00