Skip to main content
U.S. Department of Energy
Office of Scientific and Technical Information

Towards Compact Neural Networks via End-to-End Training: A Bayesian Tensor Approach with Automatic Rank Determination

Journal Article · · SIAM Journal on Mathematics of Data Science
DOI:https://doi.org/10.1137/21M1391444· OSTI ID:2889955
 [1];  [2];  [1]
  1. University of California, Santa Barbara, CA (United States)
  2. Facebook AI Systems Hardware/Software Co-design, Menlo Park, CA (United States)
Post-training model compression can reduce the inference costs of deep neural networks, but uncompressed training still consumes enormous hardware resources and energy. To enable low-energy training on edge devices, it is highly desirable to directly train a compact neural network from scratch with a low memory cost. Low-rank tensor decomposition is an effective approach to reduce the memory and computing costs of large neural networks. However, directly training low-rank tensorized neural networks is a very challenging task because it is hard to determine a proper tensor rank a priori, and the tensor rank controls both model complexity and accuracy. Here, this paper presents a novel end-to-end framework for low-rank tensorized training. We first develop a Bayesian model that supports various low-rank tensor formats (e.g., CANDECOMP/PARAFAC, Tucker, tensor-train, and tensor-train matrix) and reduces neural network parameters with automatic rank determination during training. Then we develop a customized Bayesian solver to train large-scale tensorized neural networks. Our training methods shows orders-of-magnitude parameter reduction and little accuracy loss (or even better accuracy) in the experiments. On a very large deep learning recommendation system with over 4.2 ×109 model parameters, our method can reduce the parameter number to 1.6 ×105 automatically in the training process (i.e., by 2.6 ×104 times) while achieving almost the same accuracy. Code is available at https://github.com/colehawkins/bayesian-tensor-rank-determination.
Research Organization:
University of California, Santa Barbara, CA (United States)
Sponsoring Organization:
National Science Foundation (NSF); USDOE
Grant/Contract Number:
SC0021323
OSTI ID:
2889955
Journal Information:
SIAM Journal on Mathematics of Data Science, Journal Name: SIAM Journal on Mathematics of Data Science Journal Issue: 1 Vol. 4; ISSN 2577-0187
Publisher:
Society for Industrial & Applied Mathematics (SIAM)Copyright Statement
Country of Publication:
United States
Language:
English

References (15)

Some mathematical notes on three-mode factor analysis journal September 1966
Analysis of individual differences in multidimensional scaling via an n-way generalization of “Eckart-Young” decomposition journal September 1970
PARAFAC: Parallel factor analysis journal August 1994
Bayesian tensorized neural networks with automatic rank selection journal September 2021
Tensor Regression with Applications in Neuroimaging Data Analysis journal June 2013
Tensor completion and low-n-rank tensor recovery via convex optimization journal January 2011
The horseshoe estimator for sparse signals journal April 2010
Tensor Dropout for Robust Learning journal April 2021
Bayesian Robust Tensor Factorization for Incomplete Multiway Data journal April 2016
Bayesian CP Factorization of Incomplete Tensors with Automatic Rank Determination journal September 2015
Tensor Decompositions and Applications journal August 2009
Tensor-Train Decomposition journal January 2011
Robust Low-Rank Tensor Recovery: Models and Algorithms journal January 2014
Tie conference June 2019
Energy and Policy Considerations for Deep Learning in NLP conference January 2019

Similar Records

Bayesian tensorized neural networks with automatic rank selection
Journal Article · Mon May 03 20:00:00 EDT 2021 · Neurocomputing · OSTI ID:2875828

General-Purpose Bayesian Tensor Learning With Automatic Rank Determination and Uncertainty Quantification
Journal Article · Wed Jan 05 19:00:00 EST 2022 · Frontiers in Artificial Intelligence · OSTI ID:2953825

pnnl/HiParTI
Software · Tue Jan 12 19:00:00 EST 2021 · OSTI ID:code-49911