On energy complexity of fully-connected layers
The result's identifiers
Result code in IS VaVaI
<a href="https://www.isvavai.cz/riv?ss=detail&h=RIV%2F67985807%3A_____%2F24%3A00586300" target="_blank" >RIV/67985807:_____/24:00586300 - isvavai.cz</a>
Result on the web
<a href="https://doi.org/10.1016/j.neunet.2024.106419" target="_blank" >https://doi.org/10.1016/j.neunet.2024.106419</a>
DOI - Digital Object Identifier
<a href="http://dx.doi.org/10.1016/j.neunet.2024.106419" target="_blank" >10.1016/j.neunet.2024.106419</a>
Alternative languages
Result language
angličtina
Original language name
On energy complexity of fully-connected layers
Original language description
The massive increase in the size of deep neural networks (DNNs) is accompanied by a significant increase in energy consumption of their hardware implementations which is critical for their widespread deployment in low-power mobile devices. In our previous work, an abstract hardware-independent model of energy complexity for convolutional neural networks (CNNs) has been proposed and experimentally validated. Based on this model, we provide a theoretical analysis of energy complexity related to the computation of a fully-connected layer when its inputs, outputs, and weights are transferred between two kinds of memories (DRAM and Buffer). First, we establish a general lower bound on this energy complexity. Then, we present two dataflows and calculate their energy costs to achieve the corresponding upper bounds. In the case of a partitioned Buffer, we prove by the weak duality theorem from linear programming that the lower and upper bounds coincide up to an additive constant, and therefore establish the optimal energy complexity. Finally, the asymptotically optimal quadratic energy complexity of fully-connected layers is experimentally validated by estimating their energy consumption on the Simba and Eyeriss hardware.
Czech name
—
Czech description
—
Classification
Type
J<sub>imp</sub> - Article in a specialist periodical, which is included in the Web of Science database
CEP classification
—
OECD FORD branch
10201 - Computer sciences, information science, bioinformathics (hardware development to be 2.2, social aspect to be 5.8)
Result continuities
Project
<a href="/en/project/GA22-02067S" target="_blank" >GA22-02067S: AppNeCo: Approximate Neurocomputing</a><br>
Continuities
I - Institucionalni podpora na dlouhodoby koncepcni rozvoj vyzkumne organizace
Others
Publication year
2024
Confidentiality
S - Úplné a pravdivé údaje o projektu nepodléhají ochraně podle zvláštních právních předpisů
Data specific for result type
Name of the periodical
Neural Networks
ISSN
0893-6080
e-ISSN
1879-2782
Volume of the periodical
178
Issue of the periodical within the volume
October 2024
Country of publishing house
GB - UNITED KINGDOM
Number of pages
11
Pages from-to
106419
UT code for WoS article
001281632200001
EID of the result in the Scopus database
2-s2.0-85195428301