A Dense, Massively Parallel Architecture

Author(s):  
Thilo Reski ◽  
Willy B. Strothmann
1993 ◽  
Vol 04 (01) ◽  
pp. 5-16 ◽  
Author(s):  
ALBERTO BROGGI ◽  
VINCENZO D'ANDREA ◽  
GIULIO DESTRI

In this paper we discuss the use of the Cellular Automata (CA) computational model in computer vision applications on massively parallel architectures. Motivations and guidelines of this approach to low-level vision in the frame of the PROMETHEUS project are discussed. The hard real-time requirement of actual application can be only satisfied using an ad hoc VLSI massively parallel architecture (PAPRICA). The hardware solutions and the specific algorithms can be efficiently verified and tested only using, as a simulator, a general purpose machine with a parent architecture (CM-2). An example of application related to feature extraction is discussed.


Author(s):  
Takashi Yoza ◽  
Retsu Moriwaki ◽  
Yuki Torigai ◽  
Yuki Kamikubo ◽  
Takayuki Kubota ◽  
...  

2018 ◽  
Vol 62 (4) ◽  
pp. 134-143 ◽  
Author(s):  
László Szirmay-Kalos ◽  
Ágota Kacsó ◽  
Milán Magdics ◽  
Balázs Tóth

Dynamic Positron Emission Tomography (PET) reconstructs the space-time concentration function of a radiotracer by observing the detector hits of gamma-photon pairs born during the radiotracer decay. The computation is based on the maximum likelihood principle, i.e. we look for the space-time function that maximizes the probability of the actual measurements. The number of finite elements representing the spatio-temporal concentration and the number of events detected by the tomograph may be higher than a billion, thus the reconstruction requires supercomputer performance. The enormous computational burden can be handled by graphics processors (GPU) if the algorithm is decomposed to parallel, independent threads, and the storage requirements are kept under control. This paper proposes a scalable dynamic reconstruction system where the algorithm is decomposed to phases where each phase is efficiently mapped onto the massively parallel architecture of the GPU.


Author(s):  
A. D. Romig ◽  
J. R. Michael ◽  
S. J. Plimpton

Monte Carlo electron trajectory simulations have been adapted to run on massively parallel supercomputers. An nCUBE2 parallel supercomputer with 1024 processors has been used in these studies. The advantage of the parallel architecture is the great increase in computational speed and the fact that few changes in the standard serial Monte Carlo algorithms are required. The temporal performance of the massively parallel Monte Carlo electron trajectory simulation run on 1024 nodes has been compared with Monte Carlo codes run on other types of supercomputers (CRAY-YMP). It was found to be as much as 100 times faster than the CRAY-YMP and over 2000 times faster than a VAX 785. This increase in computational speed allows the exploration of problems, in particular those involving small probability events, which are not normally amenable to solution by traditional serial Monte Carlo simulations due tothe time intensive nature of the calculations. For example, the calculation of 1,000,000 electrons at 100 kV through a thin foil takes about 6 seconds on the nCUBE.


Sign in / Sign up

Export Citation Format

Share Document