Skip to main content
U.S. Department of Energy
Office of Scientific and Technical Information

How Low Can You Go? Using Synthetic 3D Imagery to Drastically Reduce Real-World Training Data for Object Detection

Technical Report ·
DOI:https://doi.org/10.2172/1670874· OSTI ID:1670874
 [1];  [1]
  1. Sandia National Laboratories (SNL), Albuquerque, NM, and Livermore, CA (United States)

Deep convolutional neural networks (DCNNs) currently provide state-of-the-art performance on image classification and object detection tasks, and there are many global security mission areas where such models could be extremely useful. Crucially, the success of these models is driven in large part by the widespread availability of high-quality open source data sets such as Image Net, Common Objects in Context (COCO), and KITTI, which contain millions of images with thousands of unique labels. However, global security relevant objects-of-interest can be difficult to obtain: relevant events are low frequency and high consequence; the content of relevant images is sensitive; and adversaries and proliferators seek to obscure their activities. For these cases where exemplar data is hard to come-by, even fine-tuning an existing model with available data can be effectively impossible. Recent work demonstrated that models can be trained using a combination of real-world and synthetic images generated from 3D representations; that such models can exceed the performance of models trained using real-world data alone; and that the generated images need not be perfectly realistic (Tremblay, et al., 2018). However, this approach still required hundreds to thousands of real-world images for training and fine tuning, which for sparse, global security-relevant datasets can be an unrealistic hurdle. In this research, we validate the performance and behavior of DCNN models as we drive the number of real-world images used for training object detection tasks down to a minimal set. We perform multiple experiments to identify the best approach to train DCNNs from an extremely small set of real-world images. In doing so, we: Develop state-of-the-art, parameterized 3D models based on real-world images and sample from their parameters to increase the variance in synthetic image training data; Use machine learning explainability techniques to highlight and correct through targeted training the biases that result from training using completely synthetic images; and Validate our results by comparing the performance of the models trained on synthetic data to one another, and to a control model created by fine-tuning an existing ImageNet-trained model with a limited number (hundreds) of real-world images.

Research Organization:
Sandia National Laboratories (SNL-NM), Albuquerque, NM (United States)
Sponsoring Organization:
USDOE National Nuclear Security Administration (NNSA); USDOE Laboratory Directed Research and Development (LDRD) Program
DOE Contract Number:
AC04-94AL85000; NA0003525
OSTI ID:
1670874
Report Number(s):
SAND--2020-10244; 691109
Country of Publication:
United States
Language:
English

Similar Records

Safeguards-Informed Hybrid Imagery Dataset [Poster]
Conference · Thu Apr 01 00:00:00 EDT 2021 · OSTI ID:1884463

Data Validation Experiments with a Computer-Generated Imagery Dataset for International Nuclear Safeguards
Journal Article · Tue Dec 12 23:00:00 EST 2023 · ESARDA Bulletin · OSTI ID:2311269

Deep Learning for Full Waveform Inversion of Elastic Active-Source Seismic Data to Estimate P-Wave Velocity Models
Technical Report · Sun Sep 01 00:00:00 EDT 2024 · OSTI ID:2462924