Skip to main content
U.S. Department of Energy
Office of Scientific and Technical Information

ML-CGRA: An Integrated Compilation Framework to Enable Efficient Machine Learning Acceleration on CGRAs

Conference ·
Coarse-Grained Reconfigurable Arrays (CGRAs) can achieve higher energy-efficiency than general-purpose processors and accelerators or fine-grained reconfigurable devices, while maintaining adaptability to different computational patterns. CGRAs have shown some success as a platform to accelerate machine learning (ML) thanks to their flexibility, which allows them to support new models not considered by fixed accelerators. However, current solutions for CGRAs employ low level instruction-based compiler approaches and lack specialized compilation infrastructures from high-level ML frameworks that could leverage semantic information from the models, limiting the ability to efficiently map them on the recon- figurable substrate. This paper proposes ML-CGRA, an integrated compilation framework based on the MLIR infrastructure that en- ables efficient ML acceleration on CGRAs. ML-CGRA provides an end-to-end solution for mapping ML models on CGRAs that out- performs conventional approaches by 3.15× and 6.02 × on 4×4 and 8×8 CGRAs, respectively.
Research Organization:
Pacific Northwest National Laboratory (PNNL), Richland, WA (United States)
Sponsoring Organization:
USDOE
DOE Contract Number:
AC05-76RL01830
OSTI ID:
2280647
Report Number(s):
PNNL-SA-180015
Country of Publication:
United States
Language:
English

Similar Records

ICED: An Integrated CGRA Framework Enabling DFVS-Aware Acceleration
Conference · Sun Nov 03 23:00:00 EST 2024 · OSTI ID:2564121

CQNN: a CGRA-based QNN Framework
Conference · Mon Dec 21 23:00:00 EST 2020 · OSTI ID:1763313

DRIPS: Dynamic Rebalancing of Pipelined Streaming Applications on CGRAs
Conference · Sat Apr 02 00:00:00 EDT 2022 · OSTI ID:1877109

Related Subjects