multiple gpus
Recently Published Documents


TOTAL DOCUMENTS

185
(FIVE YEARS 55)

H-INDEX

16
(FIVE YEARS 2)

2022 ◽  
Vol 27 (1) ◽  
pp. 114-126
Author(s):  
Yu Tang ◽  
Zhigang Kan ◽  
Lujia Yin ◽  
Zhiquan Lai ◽  
Zhaoning Zhang ◽  
...  
Keyword(s):  

Sensors ◽  
2021 ◽  
Vol 21 (24) ◽  
pp. 8258
Author(s):  
Seokwon Lee ◽  
Inmo Ban ◽  
Myeongjin Lee ◽  
Yunho Jung ◽  
Wookyung Lee

This paper explores novel architectures for fast backprojection based video synthetic aperture radar (BP-VISAR) with multiple GPUs. The video SAR frame rate is analyzed for non-overlapped and overlapped aperture modes. For the parallelization of the backprojection process, a processing data unit is defined as the phase history data or range profile data from partial synthetic-apertures divided from the full resolution target data. Considering whether full-aperture processing is performed and range compression or backprojection are parallelized on a GPU basis, we propose six distinct architectures, each having a single-stream pipeline with a single GPU. The performance of these architectures is evaluated in both non-overlapped and overlapped modes. The efficiency of the BP-VISAR architecture with sub-aperture processing in the overlapped mode is accelerated further by filling the processing gap from the idling GPU resources with multi-stream based backprojection on multiple GPUs. The frame rate of the proposed BP-VISAR architecture with sub-aperture processing is scalable with the number of GPU devices for large pixel resolution. It can generate 4096 × 4096 video SAR frames of 0.5 m cross-range resolution in 23.0 Hz on a single GPU and 73.5 Hz on quad GPUs.


2021 ◽  
pp. 108263
Author(s):  
Joseph O'Connor ◽  
José M. Domínguez ◽  
Benedict D. Rogers ◽  
Steven J. Lind ◽  
Peter K. Stansby

2021 ◽  
Vol E104.D (12) ◽  
pp. 2057-2067
Author(s):  
Tomoya ITSUBO ◽  
Michihiro KOIBUCHI ◽  
Hideharu AMANO ◽  
Hiroki MATSUTANI
Keyword(s):  

2021 ◽  
Author(s):  
Tomohiro Imanaga ◽  
Koji Nakano ◽  
Ryota Yasudo ◽  
Yasuaki Ito ◽  
Yuya Kawamata ◽  
...  
Keyword(s):  

Author(s):  
Bahzad Taha Chicho ◽  
◽  
Amira Bibo Sallow ◽  

Python is one of the most widely adopted programming languages, having replaced a number of those in the field. Python is popular with developers for a variety of reasons, one of which is because it has an incredibly diverse collection of libraries that users can run. The most compelling reasons for adopting Keras come from its guiding principles, particularly those related to usability. Aside from the simplicity of learning and model construction, Keras has a wide variety of production deployment options and robust support for multiple GPUs and distributed training. A strong and easy-to-use free, open-source Python library is the most important tool for developing and evaluating deep learning models. The aim of this paper is to provide the most current survey of Keras in different aspects, which is a Python-based deep learning Application Programming Interface (API) that runs on top of the machine learning framework, TensorFlow. The mentioned library is used in conjunction with TensorFlow, PyTorch, CODEEPNEATM, and Pygame to allow integration of deep learning models such as cardiovascular disease diagnostics, graph neural networks, identifying health issues, COVID-19 recognition, skin tumors, image detection, and so on, in the applied area. Furthermore, the author used Keras's details, goals, challenges, significant outcomes, and the findings obtained using this method.


Algorithms ◽  
2021 ◽  
Vol 14 (7) ◽  
pp. 204
Author(s):  
Wenpeng Ma ◽  
Wu Yuan ◽  
Xiazhen Liu

Incomplete Sparse Approximate Inverses (ISAI) has shown some advantages over sparse triangular solves on GPUs when it is used for the incomplete LU based preconditioner. In this paper, we extend the single GPU method for Block–ISAI to multiple GPUs algorithm by coupling Block–Jacobi preconditioner, and introduce the detailed implementation in the open source numerical package PETSc. In the experiments, two representative cases are performed and a comparative study of Block–ISAI on up to four GPUs are conducted on two major generations of NVIDIA’s GPUs (Tesla K20 and Tesla V100). Block–Jacobi preconditioning with Block–ISAI (BJPB-ISAI) shows an advantage over the level-scheduling based triangular solves from the cuSPARSE library for the cases, and the overhead of setting up Block–ISAI and the total wall clock times of GMRES is greatly reduced using Tesla V100 GPUs compared to Tesla K20 GPUs.


Sign in / Sign up

Export Citation Format

Share Document