Underclocked Software Prefetching: More Cores, Less Energy

IEEE Micro ◽  
2012 ◽  
Vol 32 (4) ◽  
pp. 32-41 ◽  
Author(s):  
Md Kamruzzaman ◽  
Steven Swanson ◽  
Dean M. Tullsen
Keyword(s):  

Author(s):  
Marina Shimchenko ◽  
Rubén Titos-Gil ◽  
Ricardo Fernández-Pascual ◽  
Manuel E. Acacio ◽  
Stefanos Kaxiras ◽  
...  


2021 ◽  
Vol 16 (1) ◽  
Author(s):  
Jens Zentgraf ◽  
Sven Rahmann

Abstract Motivation With an increasing number of patient-derived xenograft (PDX) models being created and subsequently sequenced to study tumor heterogeneity and to guide therapy decisions, there is a similarly increasing need for methods to separate reads originating from the graft (human) tumor and reads originating from the host species’ (mouse) surrounding tissue. Two kinds of methods are in use: On the one hand, alignment-based tools require that reads are mapped and aligned (by an external mapper/aligner) to the host and graft genomes separately first; the tool itself then processes the resulting alignments and quality metrics (typically BAM files) to assign each read or read pair. On the other hand, alignment-free tools work directly on the raw read data (typically FASTQ files). Recent studies compare different approaches and tools, with varying results. Results We show that alignment-free methods for xenograft sorting are superior concerning CPU time usage and equivalent in accuracy. We improve upon the state of the art sorting by presenting a fast lightweight approach based on three-way bucketed quotiented Cuckoo hashing. Our hash table requires memory comparable to an FM index typically used for read alignment and less than other alignment-free approaches. It allows extremely fast lookups and uses less CPU time than other alignment-free methods and alignment-based methods at similar accuracy. Several engineering steps (e.g., shortcuts for unsuccessful lookups, software prefetching) improve the performance even further. Availability Our software xengsort is available under the MIT license at http://gitlab.com/genomeinformatics/xengsort. It is written in numba-compiled Python and comes with sample Snakemake workflows for hash table construction and dataset processing.



2001 ◽  
Vol 27 (9) ◽  
pp. 1173-1195
Author(s):  
Daeyeon Park ◽  
Byeong Hag Seong ◽  
Rafael H Saavedra




2007 ◽  
Vol 16 (05) ◽  
pp. 745-767
Author(s):  
SUMITKUMAR N. PAMNANI ◽  
DEEPAK N. AGARWAL ◽  
GANG QU ◽  
DONALD YEUNG

Performance-enhancement techniques improve CPU speed at the cost of other valuable system resources such as power and energy. Software prefetching is one such technique, tolerating memory latency for high performance. In this article, we quantitatively study this technique's impact on system performance and power/energy consumption. First, we demonstrate that software prefetching achieves an average of 36% performance improvement with 8% additional energy consumption and 69% higher power consumption on six memory-intensive benchmarks. Then we combine software prefetching with a (unrealistic) static voltage scaling technique to show that this performance gain can be converted to an average of 48% energy saving. This suggests that it is promising to build low power systems with techniques traditionally known for performance enhancement. We thus propose a practical online profiling based dynamic voltage scaling (DVS) algorithm. The algorithm monitors system's performance and adapts the voltage level accordingly to save energy while maintaining the observed system performance. Our proposed online profiling DVS algorithm achieves 38% energy saving without any significant performance loss.







Sign in / Sign up

Export Citation Format

Share Document