Abstract
The Coarse-Grained Reconfigurable Array (CGRA) architecture for Efficient Processingin-Memory (PIM) systems is presented in this article. PIM architectures that incorporate computational capabilities directly into memory present a promising solution to mitigate the memory wall issue. There are several difficulties in CGRA architecture optimization for PIM systems especially when it comes to striking a balance between area efficiency, power consumption and performance. The proposed framework tackles these problems by examining crucial design components like processing element (PE) architecture memory hierarchy integration and interconnect design. Using a design space exploration (DSE) methodology we assess various CGRA configurations to find the optimal trade-offs between computation throughput, power consumption and silicon area utilization. To assist in selecting effective architectures that meet different application workloads the framework combines performance analysis and advanced modeling techniques. Based on test results, the optimized CGRA architecture for PIM achieves significant improvements in processing performance (20 percent increase in throughput), area reduction and energy efficiency (up to 40 percent reduction in power consumption) when compared to conventional PIM designs. Our architecture is well-suited for data-intensive applications such as machine learning and graph analytics since these enhancements are achieved without compromising computational accuracy or scalability.
Original language | English |
---|---|
Pages (from-to) | 11-18 |
Number of pages | 8 |
Journal | Journal of VLSI Circuits and Systems |
Volume | 7 |
Issue number | 1 |
DOIs | |
State | Published - 20 Jan 2025 |
Keywords
- Coarse-Grained Reconfigurable Array (CGRA)
- Design Space Exploration (DSE)
- Energy efficiency.
- Processing-in-Memory (PIM)
- memory wall