Deep Neural Networks (DNNs) although promising for image processing, have significant computational complexity. This impedes implementation in resource-constrained systems. This paper presents effective heuristic approaches for porting DNNs onto mobile devices. Four sets of heuristics are studied: (1) heuristics based on the reuse of transferred weight matrices and weight pruning; (2) heuristics based on parameter reduction, network acceleration and non-tensor layer improvements; (3) a suite of heuristics for low power acceleration of DNNs based on dataflow, near memory and in-memory processing, transform schemes and analog based approaches; and (4) heuristics based on feature and feature map pruning utilizing cosine distances. These sets of heuristics achieve significant complexity, memory and power reductions with minimal reduction of accuracy across an assortment of state-of-the-art DNNs and applications.
|