skip to main content
OSTI.GOV title logo U.S. Department of Energy
Office of Scientific and Technical Information

Title: Evolving Deep Networks Using HPC

Conference ·

While a large number of deep learning networks have been studiedand published that produce outstanding results on natural imagedatasets, these datasets only make up a fraction of those to whichdeep learning can be applied. These datasets include text data, audiodata, and arrays of sensors that have very different characteristicsthan natural images. As these “best” networks for natural imageshave been largely discovered through experimentation and cannotbe proven optimal on some theoretical basis, there is no reasonto believe that they are the optimal network for these drasticallydifferent datasets. Hyperparameter search is thus often a very im-portant process when applying deep learning to a new problem. Inthis work we present an evolutionary approach to searching thepossible space of network hyperparameters and construction thatcan scale to 18, 000 nodes. This approach is applied to datasets ofvarying types and characteristics where we demonstrate the abilityto rapidly find best hyperparameters in order to enable practitionersto quickly iterate between idea and result.

Research Organization:
Oak Ridge National Lab. (ORNL), Oak Ridge, TN (United States). Oak Ridge Leadership Computing Facility (OLCF)
Sponsoring Organization:
USDOE Office of Science (SC), Advanced Scientific Computing Research (ASCR)
DOE Contract Number:
AC05-00OR22725
OSTI ID:
1410912
Resource Relation:
Conference: 3rd Workshop on Machine Learning in High Performance Computing Environments (in conjuction with SC17) - Denver, Colorado, United States of America - 11/13/2017 10:00:00 AM-11/13/2017 10:00:00 AM
Country of Publication:
United States
Language:
English

References (10)

The <mml:math altimg="si0011.gif" overflow="scroll" xmlns:xocs="http://www.elsevier.com/xml/xocs/dtd" xmlns:xs="http://www.w3.org/2001/XMLSchema" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xmlns="http://www.elsevier.com/xml/ja/dtd" xmlns:ja="http://www.elsevier.com/xml/ja/dtd" xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:tb="http://www.elsevier.com/xml/common/table/dtd" xmlns:sb="http://www.elsevier.com/xml/common/struct-bib/dtd" xmlns:ce="http://www.elsevier.com/xml/common/dtd" xmlns:xlink="http://www.w3.org/1999/xlink" xmlns:cals="http://www.elsevier.com/xml/common/cals/dtd"><mml:mi>MINER</mml:mi><mml:mi mathvariant="normal">ν</mml:mi><mml:mi mathvariant="normal">A</mml:mi></mml:math> data acquisition system and infrastructure
  • Perdue, G. N.; Bagby, L.; Baldin, B.
  • Nuclear Instruments and Methods in Physics Research Section A: Accelerators, Spectrometers, Detectors and Associated Equipment, Vol. 694 https://doi.org/10.1016/j.nima.2012.08.024
journal December 2012
Deep Residual Learning for Image Recognition conference June 2016
Going deeper with convolutions conference June 2015
Caffe: Convolutional Architecture for Fast Feature Embedding conference January 2014
Recent advances in deep learning for speech research at Microsoft conference May 2013
An empirical evaluation of deep architectures on problems with many factors of variation conference January 2007
A convolutional neural network neutrino event classifier journal September 2016
ImageNet Large Scale Visual Recognition Challenge journal April 2015
Evolving Neural Networks through Augmenting Topologies journal June 2002
A survey on deep learning in medical image analysis journal December 2017