Skip to main content
U.S. Department of Energy
Office of Scientific and Technical Information

Scalable deep text comprehension for Cancer surveillance on high-performance computing

Conference ·
Background: Deep Learning (DL) has advanced the state-of-the-art capabilities in bioinformatics applications which has resulted in trends of increasingly sophisticated and computationally demanding models trained by larger and larger data sets. This vastly increased computational demand challenges the feasibility of conducting cutting-edge research. One solution is to distribute the vast computational workload across multiple computing cluster nodes with data parallelism algorithms. In this study, we used a High-Performance Computing environment and implemented the Downpour Stochastic Gradient Descent algorithm for data parallelism to train a Convolutional Neural Network (CNN) for the natural language processing task of information extraction from a massive dataset of cancer pathology reports. We evaluated the scalability improvements using data parallelism training and the Titan supercomputer at Oak Ridge Leadership Computing Facility. To evaluate scalability, we used different numbers of worker nodes and performed a set of experiments comparing the effects of different training batch sizes and optimizer functions.Results: We found that Adadelta would consistently converge at a lower validation loss, though requiring over twice as many training epochs as the fastest converging optimizer, RMSProp. The Adam optimizer consistently achieved a close 2nd place minimum validation loss significantly faster; using a batch size of 16 and 32 allowed the network to converge in only 4.5 training epochs.Conclusions: We demonstrated that the networked training process is scalable across multiple compute nodes communicating with message passing interface while achieving higher classification accuracy compared to a traditional machine learning algorithm.
Research Organization:
Oak Ridge National Laboratory (ORNL), Oak Ridge, TN (United States)
Sponsoring Organization:
USDOE
DOE Contract Number:
AC05-00OR22725
OSTI ID:
1491345
Country of Publication:
United States
Language:
English

References (9)

Large-Scale Machine Learning with Stochastic Gradient Descent book January 2010
American Society of Clinical Oncology 1998 Update of Recommended Breast Cancer Surveillance Guidelines journal March 1999
Deep Learning for Automated Extraction of Primary Sites From Cancer Pathology Reports journal January 2018
Deep learning journal May 2015
Using Natural Language Processing to Improve Efficiency of Manual Chart Abstraction in Research: The Case of Breast Cancer Recurrence journal January 2014
Distributed asynchronous deterministic and stochastic gradient optimization algorithms journal September 1986
Accelerator: using data parallelism to program GPUs for general-purpose uses journal October 2006
Information Extraction: Past, Present and Future book July 2012
Cancer statistics, 2016: Cancer Statistics, 2016 journal January 2016

Similar Records

The Case for Strong Scaling in Deep Learning: Training Large 3D CNNs with Hybrid Parallelism
Journal Article · Tue Dec 29 19:00:00 EST 2020 · IEEE Transactions on Parallel and Distributed Systems · OSTI ID:1959404

Convergence of Hyperbolic Neural Networks Under Riemannian Stochastic Gradient Descent
Journal Article · Wed Oct 04 20:00:00 EDT 2023 · Communications on Applied Mathematics and Computation · OSTI ID:2007675

Related Subjects