2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2019
DOI: 10.1109/cvpr.2019.01146
|View full text |Cite
|
Sign up to set email alerts
|

ECC: Platform-Independent Energy-Constrained Deep Neural Network Compression via a Bilinear Regression Model

Abstract: Many DNN-enabled vision applications constantly operate under severe energy constraints such as unmanned aerial vehicles, Augmented Reality headsets, and smartphones. Designing DNNs that can meet a stringent energy budget is becoming increasingly important. This paper proposes ECC, a framework that compresses DNNs to meet a given energy constraint while minimizing accuracy loss.The key idea of ECC is to model the DNN energy consumption via a novel bilinear regression function. The energy estimate model allows … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
24
0

Year Published

2019
2019
2022
2022

Publication Types

Select...
5
5

Relationship

2
8

Authors

Journals

citations
Cited by 37 publications
(24 citation statements)
references
References 36 publications
(70 reference statements)
0
24
0
Order By: Relevance
“…Pruning from later layers that process smaller input resolution might not achieve as much speedup as pruning from early layers. Constraint aware optimization using Alternating Direction Method of Multipliers (ADMM) [2] such as proposed in [44] can be further integrated with our method to optimize over latency instead of FLOPs.…”
Section: Theoretical Vs Practical Speedupmentioning
confidence: 99%
“…Pruning from later layers that process smaller input resolution might not achieve as much speedup as pruning from early layers. Constraint aware optimization using Alternating Direction Method of Multipliers (ADMM) [2] such as proposed in [44] can be further integrated with our method to optimize over latency instead of FLOPs.…”
Section: Theoretical Vs Practical Speedupmentioning
confidence: 99%
“…In contrast, pruning-based methods construct smaller networks from a pretrained over-parameterized neural network by gradually removing the least important neurons. Various pruning strategies have been developed based on different heuristics (e.g., Han et al, 2016;Luo et al, 2017;He et al, 2017b;Peng et al, 2019), including energy-aware pruning methods that use energy consumption related metrics to guide the pruning process (e.g., Gordon et al, 2018;He et al, 2018;Yang et al, 2019). However, a common issue of these methods is to alter the standard training objective with sparsity-induced regularization which necessities sensitive hyperparameters tuning.…”
Section: Related Workmentioning
confidence: 99%
“…Liu et al [29] used a hyper-network in the ES algorithm to find the layer-wise sparsity for channel pruning. Instead of regarding the layer-wise sparsity as hyper-parameters, recently proposed energy-constrained compression methods [43,44] used optimization-based approaches to prune the DNNs under a given energy budget. Besides the above, there are some methods on searching efficient neural architectures [2,36], while our work mainly concentrates on compressing a given architecture.…”
Section: Automated Model Compressionmentioning
confidence: 99%