Evolutionary Stochastic Gradient Descent For Optimization Of Deep Neural Networks

Authors:
Xiaodong Cui IBM T. J. Watson Research Center
Wei Zhang IBM T.J.Watson Research Center
Zoltán Tüske IBM Research AI
Michael Picheny IBM T. J. Watson Research Center

Introduction:

The authors propose a population-based Evolutionary Stochastic Gradient Descent (ESGD) framework for optimizing deep neural networks.

Abstract:

We propose a population-based Evolutionary Stochastic Gradient Descent (ESGD) framework for optimizing deep neural networks. ESGD combines SGD and gradient-free evolutionary algorithms as complementary algorithms in one framework in which the optimization alternates between the SGD step and evolution step to improve the average fitness of the population. With a back-off strategy in the SGD step and an elitist strategy in the evolution step, it guarantees that the best fitness in the population will never degrade. In addition, individuals in the population optimized with various SGD-based optimizers using distinct hyper-parameters in the SGD step are considered as competing species in a coevolution setting such that the complementarity of the optimizers is also taken into account. The effectiveness of ESGD is demonstrated across multiple applications including speech recognition, image recognition and language modeling, using networks with a variety of deep architectures.

You may want to know: