Evolving deep neural networks: A new prospect

Sreenivas Sremath Tirumala, S. Ali, C. Phani Ramesh

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

29 Scopus citations

Abstract

The success of Deep Neural Networks (DNNs) for various applications like language processing (NLP), image processing, character recognition inspired to use machine learning (ML) and Evolutionary Computation (EC) techniques for improving learning process. Using evolutionary algorithms to improve the efficiency of deep learning attained some success. However, these techniques are unable to reduce the learning time which is the key concern for deep learning. The main problem with DNN is that, it uses a random topology to start with (similar to artificial neural networks). If the topology is not suitable, training procedure will restart with a new topology and this process continues till expected results are obtained. In this paper, we propose, for the first time, a new prospect for evolving optimized deep neural networks which can provide a warm start to the training process compared to heuristic random initial architecture. We discuss the theoretical approach towards possibility of optimizing the learning process inspired from the existing un-conventional approaches. The training process of DNN with EC approach is faster than regular approach by a considerable difference of over 6 hours for MNIST data set. Further, we observed a considerable improvement in the classification accuracies. Our approaches resulted in an improved classification accuracy of 2% and 4.4%for MNIST data set and 1.2% and 1.4% for IRIS data set compared to heuristic random weights approach. Our initial experimental results prove that evolutionary approaches provides a warm start to the deep learning, thus, reducing the training time.

Original languageEnglish
Title of host publication2016 12th International Conference on Natural Computation, Fuzzy Systems and Knowledge Discovery, ICNC-FSKD 2016
EditorsJiayi Du, Chubo Liu, Kenli Li, Lipo Wang, Zhao Tong, Maozhen Li, Ning Xiong
PublisherInstitute of Electrical and Electronics Engineers Inc.
Pages69-74
Number of pages6
ISBN (Electronic)9781509040933
DOIs
StatePublished - 19 Oct 2016
Externally publishedYes
Event12th International Conference on Natural Computation, Fuzzy Systems and Knowledge Discovery, ICNC-FSKD 2016 - Changsha, China
Duration: 13 Aug 201615 Aug 2016

Publication series

Name2016 12th International Conference on Natural Computation, Fuzzy Systems and Knowledge Discovery, ICNC-FSKD 2016

Conference

Conference12th International Conference on Natural Computation, Fuzzy Systems and Knowledge Discovery, ICNC-FSKD 2016
Country/TerritoryChina
CityChangsha
Period13/08/1615/08/16

Bibliographical note

Publisher Copyright:
© 2016 IEEE.

Keywords

  • Co-evolution
  • Deep Neural Networks
  • Evolving Neural Networks
  • Multi-population evolution

ASJC Scopus subject areas

  • Computational Theory and Mathematics
  • Control and Optimization
  • Computer Science Applications
  • Signal Processing

Fingerprint

Dive into the research topics of 'Evolving deep neural networks: A new prospect'. Together they form a unique fingerprint.

Cite this