Abstract
Convolutional Neural Network (CNN) is the most popular method of deep learning in the machine learning field. Training a CNN has always been a demanding task compared to other machine learning paradigms, and this is due to its big space of hyper-parameters such as convolutional kernel size, number of strides, number of layers, pooling window size, etc. What makes the CNN's huge hyper-parameters space optimization harder is that there is no universal robust theory supporting it, and any work flow proposed so far in literature is based on heuristics that are just rules of thumb and only depend on the dataset and problem at hand. In this work, it is empirically illustrated that the performance of a CNN is not linked only with the choice of the right hyper-parameters, but also linked to how some of the CNN operations are implemented. More specifically, the CNN performance is contrasted for two different implementations: cropping and padding the input volume. The results state that padding the input volume achieves higher accuracy and takes less time in training compared with cropping method.
Original language | English |
---|---|
Title of host publication | International Conference on Intelligent and Advanced Systems |
Subtitle of host publication | Enhance the Present for a Sustainable Future, ICIAS 2021 |
Publisher | Institute of Electrical and Electronics Engineers Inc. |
ISBN (Electronic) | 9781728176666 |
DOIs | |
State | Published - 2021 |
Externally published | Yes |
Publication series
Name | International Conference on Intelligent and Advanced Systems: Enhance the Present for a Sustainable Future, ICIAS 2021 |
---|
Bibliographical note
Publisher Copyright:© 2021 IEEE.
Keywords
- CNN
- Convolution
- Cropping
- Hyper-parameters
- Padding
- Pooling
ASJC Scopus subject areas
- Artificial Intelligence
- Computer Science Applications
- Electrical and Electronic Engineering
- Electronic, Optical and Magnetic Materials
- Instrumentation