Skip to main content
U.S. Department of Energy
Office of Scientific and Technical Information

SAMBA: S parsity A ware In- M emory Computing B ased Machine Learning A ccelerator

Journal Article · · IEEE Transactions on Computers
Not provided.
Research Organization:
Argonne National Laboratory (ANL), Argonne, IL (United States)
Sponsoring Organization:
USDOE
DOE Contract Number:
AC02-06CH11357
OSTI ID:
2423209
Journal Information:
IEEE Transactions on Computers, Journal Name: IEEE Transactions on Computers Journal Issue: 9 Vol. 72; ISSN 0018-9340
Publisher:
IEEE
Country of Publication:
United States
Language:
English

References (28)

Eyeriss: An Energy-Efficient Reconfigurable Accelerator for Deep Convolutional Neural Networks journal January 2017
A Massively Parallel Coprocessor for Convolutional Neural Networks
  • Sankaradas, Murugan; Jakkula, Venkata; Cadambi, Srihari
  • 2009 20th IEEE International Conference on Application-specific Systems, Architectures and Processors https://doi.org/10.1109/ASAP.2009.25
conference July 2009
Resistive Crossbars as Approximate Hardware Building Blocks for Machine Learning: Opportunities and Challenges journal December 2020
PRIME: a novel processing-in-memory architecture for neural network computation in ReRAM-based main memory journal October 2016
TraNNsformer: Neural network transformation for memristive crossbar based neuromorphic system design conference November 2017
15.3 A 65nm 3T Dynamic Analog RAM-Based Computing-in-Memory Macro and CNN Accelerator with Retention Enhancement, Adaptive Analog Sparsity and 44TOPS/W System Energy Efficiency conference February 2021
ShiDianNao: shifting vision processing closer to the sensor
  • Du, Zidong; Fasthuber, Robert; Chen, Tianshi
  • ISCA '15: The 42nd Annual International Symposium on Computer Architecture, Proceedings of the 42nd Annual International Symposium on Computer Architecture https://doi.org/10.1145/2749469.2750389
conference June 2015
NVIDIA Tensor Core Programmability, Performance & Precision conference May 2018
Origami: A 803-GOp/s/W Convolutional Network Accelerator journal November 2017
PipeLayer: A Pipelined ReRAM-Based Accelerator for Deep Learning
  • Song, Linghao; Qian, Xuehai; Li, Hai
  • 2017 IEEE International Symposium on High-Performance Computer Architecture (HPCA), 2017 IEEE International Symposium on High Performance Computer Architecture (HPCA) https://doi.org/10.1109/HPCA.2017.55
conference February 2017
ISAAC: a convolutional neural network accelerator with in-situ analog arithmetic in crossbars journal June 2016
Enabling Robust SOT-MTJ Crossbars for Machine Learning using Sparsity-Aware Device-Circuit Co-design conference July 2021
ReCom: An efficient resistive accelerator for compressed deep neural networks conference March 2018
DaDianNao: A Machine-Learning Supercomputer conference December 2014
Towards an embedded biologically-inspired machine vision processor conference December 2010
Learning Phrase Representations using RNN Encoder–Decoder for Statistical Machine Translation
  • Cho, Kyunghyun; van Merrienboer, Bart; Gulcehre, Caglar
  • Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP) https://doi.org/10.3115/v1/D14-1179
conference January 2014
DianNao: a small-footprint high-throughput accelerator for ubiquitous machine-learning
  • Chen, Tianshi; Du, Zidong; Sun, Ninghui
  • Proceedings of the 19th international conference on Architectural support for programming languages and operating systems - ASPLOS '14 https://doi.org/10.1145/2541940.2541967
conference January 2014
Memory-centric accelerator design for Convolutional Neural Networks conference October 2013
Sparse ReRAM engine conference June 2019
PUMA: A Programmable Ultra-efficient Memristor-based Accelerator for Machine Learning Inference
  • Ankit, Aayush; Hajj, Izzat El; Chalamalasetti, Sai Rahul
  • ASPLOS '19: Architectural Support for Programming Languages and Operating Systems, Proceedings of the Twenty-Fourth International Conference on Architectural Support for Programming Languages and Operating Systems https://doi.org/10.1145/3297858.3304049
conference April 2019
Resparc conference June 2017
In-Datacenter Performance Analysis of a Tensor Processing Unit conference January 2017
SNrram: An Efficient Sparse Neural Network Computation Architecture Based on Resistive Random-Access Memory conference June 2018
Efficient Processing of Deep Neural Networks: A Tutorial and Survey journal December 2017
A dynamically configurable coprocessor for convolutional neural networks conference January 2010
Optimizing FPGA-based Accelerator Design for Deep Convolutional Neural Networks conference January 2015
Learning the sparsity for ReRAM: mapping and pruning sparse neural network for ReRAM based accelerator
  • Lin, Jilan; Zhu, Zhenhua; Wang, Yu
  • ASPDAC '19: 24th Asia and South Pacific Design Automation Conference, Proceedings of the 24th Asia and South Pacific Design Automation Conference https://doi.org/10.1145/3287624.3287715
conference January 2019
A 35.5-127.2 TOPS/W Dynamic Sparsity-Aware Reconfigurable-Precision Compute-in-Memory SRAM Macro for Machine Learning journal January 2021

Similar Records

Boosting $ H\to b\overline{b} $ with machine learning
Journal Article · 2018 · Journal of High Energy Physics (Online) · OSTI ID:1544187

Rebooting Computers as Learning Machines
Journal Article · 2016 · Computer · OSTI ID:1411876

Machine Learning in Computational Science.
Conference · 2021 · OSTI ID:1899485

Related Subjects