Detail publikace

Reducing Memory Requirements of Convolutional Neural Networks for Inference at the Edge

BRAVENEC, T. FRÝZA, T.

Originální název

Reducing Memory Requirements of Convolutional Neural Networks for Inference at the Edge

Typ

článek ve sborníku ve WoS nebo Scopus

Jazyk

angličtina

Originální abstrakt

The main focus of this paper is to use post training quantization to analyse the influence of using lower precision data types in neural networks, while avoiding the process of retraining the networks in question. The main idea is to enable usage of high accuracy neural networks in devices other than high performance servers or super computers and bring the neural network compute closer to the device collecting the data. There are two main issues with using neural networks on edge devices, the memory constraint and the computational performance. Both of these issues could be diminished if the usage of lower precision data types does not considerably reduce the accuracy of the networks in question.

Klíčová slova

deep learning; neural networks; computer vision; machine learning; parallel computing; inference optimization; inference at the edge; reduced precision computing

Autoři

BRAVENEC, T.; FRÝZA, T.

Vydáno

20. 4. 2021

Nakladatel

Vysoké učení technické v Brně

Místo

Brno

ISBN

978-0-7381-4436-8

Kniha

International Conference Radioelektronika 2021

Strany od

1

Strany do

6

Strany počet

6

URL

BibTex

@inproceedings{BUT171248,
  author="Tomáš {Bravenec} and Tomáš {Frýza}",
  title="Reducing Memory Requirements of Convolutional Neural Networks for Inference at the Edge",
  booktitle="International Conference Radioelektronika 2021",
  year="2021",
  pages="1--6",
  publisher="Vysoké učení technické v Brně",
  address="Brno",
  doi="10.1109/RADIOELEKTRONIKA52220.2021.9420214",
  isbn="978-0-7381-4436-8",
  url="https://ieeexplore.ieee.org/document/9420214"
}