On energy complexity of fully-connected layers
Language English Country United States Media print-electronic
Document type Journal Article
PubMed
38861836
DOI
10.1016/j.neunet.2024.106419
PII: S0893-6080(24)00343-5
Knihovny.cz E-resources
- Keywords
- Convolutional neural networks, Dataflow, Deep neural networks, Energy complexity, Energy consumption, Fully-connected layer,
- MeSH
- Algorithms MeSH
- Deep Learning MeSH
- Neural Networks, Computer * MeSH
- Computers MeSH
- Programming, Linear MeSH
- Publication type
- Journal Article MeSH
The massive increase in the size of deep neural networks (DNNs) is accompanied by a significant increase in energy consumption of their hardware implementations which is critical for their widespread deployment in low-power mobile devices. In our previous work, an abstract hardware-independent model of energy complexity for convolutional neural networks (CNNs) has been proposed and experimentally validated. Based on this model, we provide a theoretical analysis of energy complexity related to the computation of a fully-connected layer when its inputs, outputs, and weights are transferred between two kinds of memories (DRAM and Buffer). First, we establish a general lower bound on this energy complexity. Then, we present two dataflows and calculate their energy costs to achieve the corresponding upper bounds. In the case of a partitioned Buffer, we prove by the weak duality theorem from linear programming that the lower and upper bounds coincide up to an additive constant, and therefore establish the optimal energy complexity. Finally, the asymptotically optimal quadratic energy complexity of fully-connected layers is experimentally validated by estimating their energy consumption on the Simba and Eyeriss hardware.
References provided by Crossref.org