TY - GEN
T1 - Accuracy Evaluation of Transposed Convolution-Based Quantized Neural Networks
AU - Sestito, Cristian
AU - Perri, Stefania
AU - Stewart, Robert James
N1 - Funding Information:
ACKNOWLEDGMENTS This work was supported by: “POR Calabria FSE/FESR 2014-2020 – International Mobility of PhD students and research grants/type A Researchers” – Actions 10.5.6 and 10.5.12 actuated by Regione Calabria, Italy; The Engineering and Physical Research Council: HAFLANG (EP/W009447/1); Border Patrol (EP/N028201/1); Serious Coding (EP/T017511/1).
Publisher Copyright:
© 2022 IEEE.
PY - 2022/9/30
Y1 - 2022/9/30
N2 - Several modern applications in the field of Artificial Intelligence exploit deep learning to make accurate decisions. Recent work on compression techniques allows for deep learning applications, such as computer vision, to run on Edge Computing devices. For instance, quantizing the precision of deep learning architectures allows Edge Computing devices to achieve high throughput at low power. Quantization has been mainly focused on multilayer perceptrons and convolution-based models for classification problems. However, its impact over more complex scenarios, such as image up-sampling, is still underexplored. This paper presents a systematic evaluation of the accuracy achieved by quantized neural networks when performing image up-sampling in three different applications: image compression/decompression, synthetic image generation and semantic segmentation. Taking into account the promising attitude of learnable filters to predict pixels, transposed convolutional layers are used for up-sampling. Experimental results based on analytical metrics show that acceptable accuracies are reached with quantization spanning between 3 and 7 bits. Based on the visual inspection, the range 2–6 bits guarantees appropriate accuracy.
AB - Several modern applications in the field of Artificial Intelligence exploit deep learning to make accurate decisions. Recent work on compression techniques allows for deep learning applications, such as computer vision, to run on Edge Computing devices. For instance, quantizing the precision of deep learning architectures allows Edge Computing devices to achieve high throughput at low power. Quantization has been mainly focused on multilayer perceptrons and convolution-based models for classification problems. However, its impact over more complex scenarios, such as image up-sampling, is still underexplored. This paper presents a systematic evaluation of the accuracy achieved by quantized neural networks when performing image up-sampling in three different applications: image compression/decompression, synthetic image generation and semantic segmentation. Taking into account the promising attitude of learnable filters to predict pixels, transposed convolutional layers are used for up-sampling. Experimental results based on analytical metrics show that acceptable accuracies are reached with quantization spanning between 3 and 7 bits. Based on the visual inspection, the range 2–6 bits guarantees appropriate accuracy.
KW - accuracy
KW - neural networks
KW - quantization
KW - transposed convolution
UR - http://www.scopus.com/inward/record.url?scp=85140734134&partnerID=8YFLogxK
U2 - 10.1109/IJCNN55064.2022.9892671
DO - 10.1109/IJCNN55064.2022.9892671
M3 - Conference contribution
BT - 2022 International Joint Conference on Neural Networks (IJCNN)
PB - IEEE
ER -