Constraints on Hyper-parameters in Deep Learning Convolutional Neural Networks

Ubaid M. Al-Saggaf, Abdelaziz Botalb, Muhammad Faisal, Muhammad Moinuddin, Abdulrahman U. Alsaggaf, Sulhi Ali Alfakeh

Research output: Contribution to journalArticlepeer-review

1 Scopus citations

Abstract

Convolutional Neural Network (CNN), a type of Deep Learning, has a very large number of hyper-meters in contrast to the Artificial Neural Network (ANN) which makes the task of CNN training more demanding. The reason why the task of tuning parameters optimization is difficult in the CNN is the existence of a huge optimization space comprising a large number of hyper-parameters such as the number of layers, number of neurons, number of kernels, stride, padding, rows or columns truncation, parameters of the backpropagation algorithm, etc. Moreover, most of the existing techniques in the literature for the selection of these parameters are based on random practice which is developed for some specific datasets. In this work, we empirically investigated and proved that CNN performance is linked not only to choosing the right hyper-parameters but also to its implementation. More specifically, it is found that the performance is also depending on how it deals when the CNN operations require setting of hyper-parameters that do not symmetrically fit the input volume. We demonstrated two different implementations, crop or pad the input volume to make it fit. Our analysis shows that padding performs better than cropping in terms of prediction accuracy (85.58% in contrast to 82.62%) while takes lesser training time (8 minutes lesser).

Original languageEnglish
Pages (from-to)439-449
Number of pages11
JournalInternational Journal of Advanced Computer Science and Applications
Volume13
Issue number11
DOIs
StatePublished - 2022

Bibliographical note

Funding Information:
FUNDING STATEMENT This research work is funded by Institutional Fund Projects by the Ministry of Education, Saudi Arabia, under grant no. (IFPRC-118-135-2020).

Funding Information:
ACKNOWLEDGMENT The authors extend their appreciation to the Deputyship for Research and Innovation, Ministry of Education in Saudi Arabia, for funding this research work through the project number (IFPRC-118-135-2020) and King Abdulaziz University, DSR, Jeddah, Saudi Arabia.

Publisher Copyright:
© This is an open-access article distributed under the terms of the Creative Commons Attribution Non-Commercial License (http://creativecommons.org/licenses/by-nc/4.0) which permits unrestricted non-commercial use, distribution, and reproduction in any medium, provided the original work is properly cited.

Keywords

  • Back-propagation
  • Cnn
  • Convolution
  • Deep learning
  • Hyper-parameters
  • Neural networks
  • Pooling
  • Stride
  • Zero-padding

ASJC Scopus subject areas

  • General Computer Science

Fingerprint

Dive into the research topics of 'Constraints on Hyper-parameters in Deep Learning Convolutional Neural Networks'. Together they form a unique fingerprint.

Cite this