QSM: A general purpose shared-memory model for parallel computation

Author(s):  
Vijaya Ramachandran

It is shown that any program written for the idealized shared-memory model of parallel computation can be simulated on a hypercube architecture with only constant factor inefficiency, provided that the original program has a certain amount of parallel slackness.


1999 ◽  
Vol 32 (3) ◽  
pp. 327-359 ◽  
Author(s):  
P. B. Gibbons ◽  
Y. Matias

2018 ◽  
Vol 3 (7) ◽  
pp. 12
Author(s):  
DaeHwan Kim

Nowadays, GPU processors are widely used for general-purpose parallel computation applications. In the GPU programming, thread and block configuration is one of the most important decisions to be made, which increases parallelism and hides instruction latency. However, in many cases, it is often difficult to have sufficient parallelism to hide all the latencies, where the high latencies are often caused by the global memory accesses. In order to reduce the number of  those accesses, the shared memory is instead used which is  much faster than the global memory being located on a chip. The performance of the proposed thread configuration is evaluated on the GPU 960 processor. The experimental result shows that the best configuration improves the performance by 7.3 times compared to the worst configuration in the experiment. The experiences are also discussed for the shared memory performance when compared to that of the global memory.


Author(s):  
Atanu Barai ◽  
Gopinath Chennupati ◽  
Nandakishore Santhi ◽  
Abdel-Hameed Badawy ◽  
Yehia Arafa ◽  
...  
Keyword(s):  

Author(s):  
Kazuhiko Ohno ◽  
Dai Michiura ◽  
Masaki Matsumoto ◽  
Takahiro Sasaki ◽  
Toshio Kondo

Author(s):  
Kazuhiko Ohno ◽  
Dai Michiura ◽  
Masaki Matsumoto ◽  
Takahiro Sasaki ◽  
Toshio Kondo

2020 ◽  
Vol 60 (1) ◽  
pp. 25-37
Author(s):  
Michal Bošanský ◽  
Bořek Patzák

The efficient codes can take an advantage of multiple threads and/or processing nodes to partition a work that can be processed concurrently. This can reduce the overall run-time or make the solution of a large problem feasible. This paper deals with evaluation of different parallelization strategies of assembly operations for global vectors and matrices, which are one of the critical operations in any finite element software. Different assembly strategies for systems with a shared memory model are proposed and evaluated, using Open Multi-Processing (OpenMP), Portable Operating System Interface (POSIX), and C++11 Threads. The considered strategies are based on simple synchronization directives, various block locking algorithms and, finally, on smart locking free processing based on a colouring algorithm. The different strategies were implemented in a free finite element code with object-oriented architecture OOFEM [1].


Sign in / Sign up

Export Citation Format

Share Document