Design and Implement of Sharable Multi-Channel On-Chip Memory for Embedded CMP System

2011 ◽  
Vol 217-218 ◽  
pp. 1147-1152 ◽  
Author(s):  
Cai Xia Liu ◽  
Zhi Bin Zhang ◽  
Feng Qi Wei ◽  
Xiao Dong Xu

A kind of shared multi-channel on-chip memory architecture (SMC-OCM) for embedded CMPs is proposed in this article. To implement SMC-OCM architecture, the sharable multi-channel on-chip memory (MC-OCM) is designed and implemented based on FPGA. The characteristic of multiple data channel of MC-OCM assures good parallel responsiveness of SMC-OCM system. Experiments showed that the access latency of SMC-OCM is lower than that of the-state-of arts. SMC-OCM architecture satisfies the performance requirements for memory system by embedded applications

2020 ◽  
Vol 12 (2) ◽  
pp. 116-121
Author(s):  
Rastislav Struharik ◽  
Vuk Vranjković

Data movement between the Convolutional Neural Network (CNN) accelerators and off-chip memory is critical concerning the overall power consumption. Minimizing power consumption is particularly important for low power embedded applications. Specific CNN computes patterns offer a possibility of significant data reuse, leading to the idea of using specialized on-chip cache memories which enable a significant improvement in power consumption. However, due to the unique caching pattern present within CNNs, standard cache memories would not be efficient. In this paper, a novel on-chip cache memory architecture, based on the idea of input feature map striping, is proposed, which requires significantly less on-chip memory resources compared to previously proposed solutions. Experiment results show that the proposed cache architecture can reduce on-chip memory size by a factor of 16 or more, while increasing power consumption no more than 15%, compared to some of the previously proposed solutions.


2012 ◽  
Vol 629 ◽  
pp. 542-547
Author(s):  
Cai Xia Liu ◽  
Xiao Qing Tian ◽  
Zhi Bin Zhang

A kind of shared multi-channel on-chip memory CMP architecture is proposed in this article to efficiently support embedded applications. For the multi-channel on-chip memory being scarce resource, optimal space management mechanism of multi-channel on-chip memory is proposed including automatic space allocation strategy based on application parallelization mapping pattern and optimal space utilization scheme. ILP-model-based analysis of system performance verifies that the proposed optimal space management mechanism can deeply exploit the efficiency of multi-channel on-chip memory to improve system performance.


Author(s):  
Junfeng Song ◽  
Xianshu Luo ◽  
Yanzhe Tang ◽  
Chao Li ◽  
Lianxi Jia ◽  
...  
Keyword(s):  

Author(s):  
J. Santhi ◽  
L. Srinivas

Multi-pattern matching is known to require intensive memory accesses and is often a performance bottleneck. Hence specialized hardware-accelerated algorithms are being developed for line-speed packet processing. While several pattern matching algorithms have already been developed for such applications, we find that most of them suffer from scalability issues. We present a hardware-implementable pattern matching algorithm for content filtering applications, which is scalable in terms of speed, the number of patterns and the pattern length. We modify the classic Aho-Corasick algorithm to consider multiple characters at a time for higher throughput. Furthermore, we suppress a large fraction of memory accesses by using Bloom filters implemented with a small amount of on-chip memory. The resulting algorithm can support matching of several thousands of patterns at more than 10 Gbps with the help of a less than 50 KBytes of embedded memory and a few megabytes of external SRAM.


2013 ◽  
pp. 140-159
Author(s):  
Giorgio C. Buttazzo

The number of computer-controlled systems has increased dramatically in our daily life. Processors and microcontrollers are embedded in most of the devices we use every day, such as mobile phones, cameras, media players, navigators, washing machines, biomedical devices, and cars. The complexity of such systems is increasing exponentially, pushed by the demand of new products with extra functionality, higher performance requirements, and low energy consumption. To cope with such a complex scenario, many embedded systems are adopting more powerful and highly integrated hardware components, such as multi-core systems, network-on-chip architectures, inertial subsystems, and special purpose co-processors. However, developing, analyzing, and testing the application software on these architectures is not easy, and new methodologies are being investigated in the research community to guarantee high predictability and efficiency in next generation embedded devices. This chapter presents some recent approaches proposed within the real-time research community aimed at achieving predictability, high modularity, efficiency, and adaptability in modern embedded computing systems.


Sign in / Sign up

Export Citation Format

Share Document