A High-Performance Parallel Computing Framework for Uncertainty Quantification Analysis of RF Devices*

Author(s):  
George Stantchev ◽  
Simon Cooke ◽  
Kyle Elliott ◽  
John Petillo

Usage of high-performance computing (HPC) infrastructure adopting cloud-computing environment offers an efficient solution for executing data intensive application. MapReduce (MR) is the favored high performance parallel computing framework used in BigData study, scientific, and data intensive applications. Hadoop is one of the significantly used MR based parallel computing framework by various organization as it is freely available open source framework from Apache foundation. The existing Hadoop MapReduce (HMR) based makespan model incurs memory and I/O overhead. Thus, affecting makespan performance. For overcoming research issues and challenges, this manuscript presented an efficient parallel HMR (PHMR) makespan model. The PHMR includes a parallel execution scheme in virtual computing worker to reduce makespan times using cloud computing framework. The PHMR model provides efficient memory management design within the virtual computing workers to minimize memory allocation and transmission overheads. For evaluating performance of PHMR of over existing model experiment are conducted on public cloud environment using Azure HDInsight cloud platform. Different application such as bioinformatics, tex mining, stream, and nonstream application is considered. The overall result obtained shows superior performance is attained by PHMR over existing model in term of makespan time reduction and correlation among practical and theoretical makespan values.


2019 ◽  
Author(s):  
Rohitash Chandra ◽  
Danial Azam ◽  
Arpit Kapoor ◽  
R. Dietmar Mulller

Abstract. The complex and computationally expensive features of the forward landscape and sedimentary basin evolution models pose a major challenge in the development of efficient inference and optimization methods. Bayesian inference provides a methodology for estimation and uncertainty quantification of free model parameters. In our previous work, parallel tempering Bayeslands was developed as a framework for parameter estimation and uncertainty quantification for the landscape and basin evolution modelling software Badlands. Parallel tempering Bayeslands features high-performance computing with dozens of processing cores running in parallel to enhance computational efficiency. Although parallel computing is used, the procedure remains computationally challenging since thousands of samples need to be drawn and evaluated. In large-scale landscape and basin evolution problems, a single model evaluation can take from several minutes to hours, and in certain cases, even days. Surrogate-assisted optimization has been with successfully applied to a number of engineering problems This motivates its use in optimisation and inference methods suited for complex models in geology and geophysics. Surrogates can speed up parallel tempering Bayeslands by developing computationally inexpensive surrogates to mimic expensive models. In this paper, we present an application of surrogate-assisted parallel tempering where that surrogate mimics a landscape evolution model including erosion, sediment transport and deposition, by estimating the likelihood function that is given by the model. We employ a machine learning model as a surrogate that learns from the samples generated by the parallel tempering algorithm and the likelihood from the model. The entire framework is developed in a parallel computing infrastructure to take advantage of parallelization. The results show that the proposed methodology is effective in lowering the overall computational cost significantly while retaining the quality of solutions.


2012 ◽  
Vol 17 (4) ◽  
pp. 207-216 ◽  
Author(s):  
Magdalena Szymczyk ◽  
Piotr Szymczyk

Abstract The MATLAB is a technical computing language used in a variety of fields, such as control systems, image and signal processing, visualization, financial process simulations in an easy-to-use environment. MATLAB offers "toolboxes" which are specialized libraries for variety scientific domains, and a simplified interface to high-performance libraries (LAPACK, BLAS, FFTW too). Now MATLAB is enriched by the possibility of parallel computing with the Parallel Computing ToolboxTM and MATLAB Distributed Computing ServerTM. In this article we present some of the key features of MATLAB parallel applications focused on using GPU processors for image processing.


Sign in / Sign up

Export Citation Format

Share Document