Reduction of computation time using Graphics Processing Unit for the detection of a crack in a large scale concrete structure

2012 ◽  
Vol 131 (4) ◽  
pp. 3443-3443 ◽  
Author(s):  
Yuhei Katsurakawa ◽  
Toyota Fujioka ◽  
Yoshifumi Nagata ◽  
Masato Abe
Author(s):  
Alan Gray ◽  
Kevin Stratford

Leading high performance computing systems achieve their status through use of highly parallel devices such as NVIDIA graphics processing units or Intel Xeon Phi many-core CPUs. The concept of performance portability across such architectures, as well as traditional CPUs, is vital for the application programmer. In this paper we describe targetDP, a lightweight abstraction layer which allows grid-based applications to target data parallel hardware in a platform agnostic manner. We demonstrate the effectiveness of our pragmatic approach by presenting performance results for a complex fluid application (with which the model was co-designed), plus separate lattice quantum chromodynamics particle physics code. For each application, a single source code base is seen to achieve portable performance, as assessed within the context of the Roofline model. TargetDP can be combined with Message Passing Interface (MPI) to allow use on systems containing multiple nodes: we demonstrate this through provision of scaling results on traditional and graphics processing unit-accelerated large scale supercomputers.


2020 ◽  
Vol 10 (7) ◽  
pp. 2359
Author(s):  
Sajad Mohammadi ◽  
Hamidreza Karami ◽  
Mohammad Azadifar ◽  
Farhad Rachidi

An open accelerator (OpenACC)-aided graphics processing unit (GPU)-based finite difference time domain (FDTD) method is presented for the first time for the 3D evaluation of lightning radiated electromagnetic fields along a complex terrain with arbitrary topography. The OpenACC directive-based programming model is used to enhance the computational performance, and the results are compared with those obtained by using a CPU-based model. It is shown that OpenACC GPUs can provide very accurate results, and they are more than 20 times faster than CPUs. The presented results support the use of OpenACC not only in relation to lightning electromagnetics problems, but also to large-scale realistic electromagnetic compatibility (EMC) applications in which computation time efficiency is a critical factor.


Author(s):  
Timothy Dykes ◽  
Claudio Gheller ◽  
Marzia Rivi ◽  
Mel Krokos

With the increasing size and complexity of data produced by large-scale numerical simulations, it is of primary importance for scientists to be able to exploit all available hardware in heterogenous high-performance computing environments for increased throughput and efficiency. We focus on the porting and optimization of Splotch, a scalable visualization algorithm, to utilize the Xeon Phi, Intel’s coprocessor based upon the new many integrated core architecture. We discuss steps taken to offload data to the coprocessor and algorithmic modifications to aid faster processing on the many-core architecture and make use of the uniquely wide vector capabilities of the device, with accompanying performance results using multiple Xeon Phi. Finally we compare performance against results achieved with the Graphics Processing Unit (GPU) based implementation of Splotch.


Author(s):  
Masatomo Inui ◽  
Kouhei Nishimiya ◽  
Nobuyuki Umezu

Abstract Clearance is a basic parameter in the design of mechanical products, generally specified as the distance between two shape elements, for example, the width of a slot. This definition is unsuitable for evaluating the clearance during assembly or manufacturing tasks, where the depth information is also critical. In this paper, we propose a novel definition of clearance for the surface of three-dimensional objects. Unlike the typical methods used to define clearance, the proposed method can simultaneously handle the relationship between the width and depth in the clearance, and thus, obtain an intuitive understanding regarding the assembly and manufacturing capability of a product. Our definition is based on the accessibility cone of a point on the object’s surface; further, the peak angle of the accessibility cone corresponds to the clearance at this point. A computation method of the clearance is presented and the results of its application are demonstrated. Our method uses the rendering function of a graphics processing unit to compute the clearance. A large computation time necessary for the analysis is considered as a problem regarding the practical use of this clearance definition.


Author(s):  
Shen Lu ◽  
Richard S. Segall

Big data is large-scale data and can be either discrete or continuous. This article entails research that discusses the continuous case of big data often called “data streaming.” More and more businesses will depend on being able to process and make decisions on streams of data. This article utilizes the algorithmic side of data stream processing often called “stream analytics” or “stream mining.” Data streaming Windows Join can be improved by using graphics processing unit (GPU) for higher performance computing. Data streams are generated by two independent threads: one thread can be used to generate Data Stream A, and the other thread can be used to generate Data Stream B. One would use a Windows Join thread to merge the two data streams, which is also the process of “Data Stream Window Join.” The Window Join process can be implemented in parallel that can efficiently improve the computing speed. Experiments are provided for Data Stream Window Joins using both static and dynamic data.


2013 ◽  
Vol 2013 ◽  
pp. 1-7 ◽  
Author(s):  
Congying Han ◽  
Tingting Feng ◽  
Guoping He ◽  
Tiande Guo

A modified parallel variable distribution (PVD) algorithm for solving large-scale constrained optimization problems is developed, which modifies quadratic subproblemQPlat each iteration instead of theQPl0of the SQP-type PVD algorithm proposed by C. A. Sagastizábal and M. V. Solodov in 2002. The algorithm can circumvent the difficulties associated with the possible inconsistency ofQPl0subproblem of the original SQP method. Moreover, we introduce a nonmonotone technique instead of the penalty function to carry out the line search procedure with more flexibly. Under appropriate conditions, the global convergence of the method is established. In the final part, parallel numerical experiments are implemented on CUDA based on GPU (Graphics Processing unit).


Sign in / Sign up

Export Citation Format

Share Document