instruction cache
Recently Published Documents


TOTAL DOCUMENTS

235
(FIVE YEARS 11)

H-INDEX

21
(FIVE YEARS 0)

Author(s):  
Tanvir Ahmed Khan ◽  
Dexin Zhang ◽  
Akshitha Sriraman ◽  
Joseph Devietti ◽  
Gilles Pokam ◽  
...  

Author(s):  
Qing-Qing Li ◽  
Zhi-Guo Yu ◽  
Yi Sun ◽  
Jing-He Wei ◽  
Xiao-Feng Gu

2021 ◽  
Author(s):  
Utku Sirin ◽  
Pınar Tözün ◽  
Danica Porobic ◽  
Ahmad Yasin ◽  
Anastasia Ailamaki

AbstractMicro-architectural behavior of traditional disk-based online transaction processing (OLTP) systems has been investigated extensively over the past couple of decades. Results show that traditional OLTP systems mostly under-utilize the available micro-architectural resources. In-memory OLTP systems, on the other hand, process all the data in main-memory and, therefore, can omit the buffer pool. Furthermore, they usually adopt more lightweight concurrency control mechanisms, cache-conscious data structures, and cleaner codebases since they are usually designed from scratch. Hence, we expect significant differences in micro-architectural behavior when running OLTP on platforms optimized for in-memory processing as opposed to disk-based database systems. In particular, we expect that in-memory systems exploit micro-architectural features such as instruction and data caches significantly better than disk-based systems. This paper sheds light on the micro-architectural behavior of in-memory database systems by analyzing and contrasting it to the behavior of disk-based systems when running OLTP workloads. The results show that, despite all the design changes, in-memory OLTP exhibits very similar micro-architectural behavior to disk-based OLTP: more than half of the execution time goes to memory stalls where instruction cache misses or the long-latency data misses from the last-level cache (LLC) are the dominant factors in the overall execution time. Even though ground-up designed in-memory systems can eliminate the instruction cache misses, the reduction in instruction stalls amplifies the impact of LLC data misses. As a result, only 30% of the CPU cycles are used to retire instructions, and 70% of the CPU cycles are wasted to stalls for both traditional disk-based and new generation in-memory OLTP.


2021 ◽  
Author(s):  
Chun-Chang Yu ◽  
Yu Hen Hu ◽  
Yi-Chang Lu ◽  
Charlie Chung-Ping Chen

IEEE Access ◽  
2020 ◽  
Vol 8 ◽  
pp. 208003-208015
Author(s):  
Tingxu Zhang ◽  
Wenguang Zheng ◽  
Yingyuan Xiao ◽  
Guangping Xu

2020 ◽  
pp. 1-1
Author(s):  
Jinkwon Kim ◽  
Seokin Hong ◽  
Jeongkyu Hong ◽  
Soontae Kim

2019 ◽  
Vol 28 (12) ◽  
pp. 1950203
Author(s):  
Sajjad Rostami-Sani ◽  
Mojtaba Valinataj ◽  
Saeideh Alinezhad Chamazcoti

The cache system dissipates a significant amount of energy compared to the other memory components. This will be intensified if a cache is designed with a set-associative structure to improve the system performance because the parallel accesses to the entries of a set for tag comparisons lead to even more energy consumption. In this paper, a novel method is proposed as a combination of a counting Bloom filter and partial tags to mitigate the energy consumption of set-associative caches. This new hybrid method noticeably decreases the cache energy consumption especially in highly-associative instruction caches. In fact, it uses an enhanced counting Bloom filter to predict cache misses with a high accuracy as well as partial tags to decrease the overall cache size. This way, unnecessary tag comparisons can be prevented and therefore, the cache energy consumption is considerably reduced. Based on the simulation results, the proposed method provides the energy reduction from 22% to 31% for 4-way–32-way set-associative L1 caches bigger than 16[Formula: see text]kB running the MiBench programs. The improvements are attained with a negligible system performance degradation compared to the traditional cache system.


Sign in / Sign up

Export Citation Format

Share Document