scholarly journals GPU-ACCELERATED SPH MODEL FOR WATER WAVES AND FREE SURFACE FLOWS

2011 ◽  
Vol 1 (32) ◽  
pp. 9 ◽  
Author(s):  
Robert Anthony Dalrymple ◽  
Alexis Herault ◽  
Giuseppe Bilotta ◽  
Rozita Jalali Farahani

This paper discusses the meshless numerical method Smoothed Particle Hydrodynamics and its application to water waves and nearshore circulation. In particularly we focus on an implementation of the model on the graphics processing unit (GPU) of computers, which permits low-cost supercomputing capabilities for certain types of computational problems. The implementation here runs on Nvidia graphics cards, from off-the-shelf laptops to the top-of-line Tesla cards for workstations with their current 480 massively parallel streaming processors. Here we apply the model to breaking waves and nearshore circulation, demonstrating that SPH can model changes in wave properties due to shoaling, refraction, and diffraction and wave-current interaction; as well as nonlinear phenomena such as harmonic generation, and, by using wave-period averaged quantities, such aspects of nearshore circulation as wave set-up, longshore currents, rip currents, and nearshore circulation gyres.

2021 ◽  
Author(s):  
Airidas Korolkovas ◽  
Alexander Katsevich ◽  
Michael Frenkel ◽  
William Thompson ◽  
Edward Morton

X-ray computed tomography (CT) can provide 3D images of density, and possibly the atomic number, for large objects like passenger luggage. This information, while generally very useful, is often insufficient to identify threats like explosives and narcotics, which can have a similar average composition as benign everyday materials such as plastics, glass, light metals, etc. A much more specific material signature can be measured with X-ray diffraction (XRD). Unfortunately, XRD signal is very faint compared to the transmitted one, and also challenging to reconstruct for objects larger than a small laboratory sample. In this article we analyze a novel low-cost scanner design which captures CT and XRD signals simultaneously, and uses the least possible collimation to maximize the flux. To simulate a realistic instrument, we derive a formula for the resolution of any diffraction pathway, taking into account the polychromatic spectrum, and the finite size of the source, detector, and each voxel. We then show how to reconstruct XRD patterns from a large phantom with multiple diffracting objects. Our approach includes a reasonable amount of photon counting noise (Poisson statistics), as well as measurement bias, in particular incoherent Compton scattering. The resolution of our reconstruction is sufficient to provide significantly more information than standard CT, thus increasing the accuracy of threat detection. Our theoretical model is implemented in GPU (Graphics Processing Unit) accelerated software which can be used to assess and further optimize scanner designs for specific applications in security, healthcare, and manufacturing quality control.


2011 ◽  
Vol 21 (01) ◽  
pp. 31-47 ◽  
Author(s):  
NOEL LOPES ◽  
BERNARDETE RIBEIRO

The Graphics Processing Unit (GPU) originally designed for rendering graphics and which is difficult to program for other tasks, has since evolved into a device suitable for general-purpose computations. As a result graphics hardware has become progressively more attractive yielding unprecedented performance at a relatively low cost. Thus, it is the ideal candidate to accelerate a wide variety of data parallel tasks in many fields such as in Machine Learning (ML). As problems become more and more demanding, parallel implementations of learning algorithms are crucial for a useful application. In particular, the implementation of Neural Networks (NNs) in GPUs can significantly reduce the long training times during the learning process. In this paper we present a GPU parallel implementation of the Back-Propagation (BP) and Multiple Back-Propagation (MBP) algorithms, and describe the GPU kernels needed for this task. The results obtained on well-known benchmarks show faster training times and improved performances as compared to the implementation in traditional hardware, due to maximized floating-point throughput and memory bandwidth. Moreover, a preliminary GPU based Autonomous Training System (ATS) is developed which aims at automatically finding high-quality NNs-based solutions for a given problem.


2009 ◽  
Vol 17 (25) ◽  
pp. 22320 ◽  
Author(s):  
Claudio Vinegoni ◽  
Lyuba Fexon ◽  
Paolo Fumene Feruglio ◽  
Misha Pivovarov ◽  
Jose-Luiz Figueiredo ◽  
...  

2012 ◽  
Vol 463-464 ◽  
pp. 1073-1076
Author(s):  
Helmar Alvares ◽  
Eliana Prado Lopes Aude ◽  
Ernesto Prado Lopes

This work proposes a Web-Based laboratory where researchers share the facilities of a simulation environment for parallel algorithms which solves scheduling problems known as Job Shop Problem (JSP). The environment supports multi-language platforms and uses a low cost, high performance Graphics Processing Unit (GPU) connected to a Java application server to help design more efficient solutions for JSP. Within a single web environment one can analyze and compare different methods and meta-heuristics. Each newly developed method is stored in an environment library and made available to all other users of the environment. This amassment of openly accessible solution methods will allow for the rapid convergence towards optimal solutions for JSP. The algorithm uses the parallel architecture of the system to handle threads. Each thread represents a job operation and the number of threads scales with the problem’s size. The threads exchange information in order to find the best solution. This cooperation decreases response times by one or two orders of magnitude.


2018 ◽  
Vol 65 ◽  
pp. 249-265
Author(s):  
Shahrdad G. Sajjadi ◽  
Julian C. R. Hunt

Michael Longuet-Higgins was a geometer and applied mathematician who made notable contributions to geophysics and physical oceanography, particularly to the theory of oceanic microseism and to the dynamics of finite amplitude, sharp-crested wind-generated surface waves. The latter led to his pioneering studies on breaking waves. On a much larger scale, he showed how ocean waves produce currents around islands in the ocean. He considered wider aspects of the physics of waves, including wave-driven transport of sand along beaches, and the electrical effects of tidal streams. He also contributed to subjects of a geometrical character such as the growth of quasi-crystals, the assembly of protein sheaths in viruses, to chains of circle themes and to a wide variety of other topics. He was an extraordinary applied mathematician, using the simplest forms of mathematics to demonstrate and discover highly complex nonlinear phenomena. In particular, he often thought of problems involving water waves using his unique knowledge of geometry and then tested his theories by experiment. Along with Brooke Benjamin FRS, Sir James Lighthill FRS, Walter Munk FRS, John Miles and Andrei Monin, Michael Longuet-Higgins stands out as one of the towering figures of theoretical fluid dynamics in the twentieth century. His contributions will have a continuing influence on our attempts to understand better the processes that influence the oceans.


2021 ◽  
Vol 13 (4) ◽  
pp. 690
Author(s):  
Dylan Anderson ◽  
A. Spicer Bak ◽  
Katherine L. Brodie ◽  
Nicholas Cohn ◽  
Rob A. Holman ◽  
...  

Complex two-dimensional nearshore current patterns are generated by feedbacks between sub-aqueous morphology and momentum imparted on the water column by breaking waves, winds, and tides. These non-stationary features, such as rip currents and circulation cells, respond to changing environmental conditions and underlying morphology. However, using fixed instruments to observe nearshore currents is limiting due to the high costs and logistics necessary to achieve adequate spatial sampling resolution. A new technique for processing surf-zone imagery, WAMFlow, quantifies fluid velocities to reveal complex, multi-scale (10 s–1000 s meters) nearshore surface circulation patterns. We apply the concept of a wave-averaged movie (WAM) to measure surf-zone circulation patterns on spatial scales of kilometers in the alongshore and 100 s of meters in the cross-shore. The approach uses a rolling average of 2 Hz optical imagery, removing the dominant optical clutter of incident waves, to leave the residual foam or water turbidity features carried by the flow. These residual features are tracked as quasi-passive tracers in space and time using optical flow, which solves for u and v as a function of image intensity gradients in x, y, and t. Surf zone drifters were deployed over multiple days with varying nearshore circulations to validate the optically derived flow patterns. Root mean square error are reduced to 0.1 m per second after filtering based on image attributes. The optically derived patterns captured longshore currents, rip currents, and gyres within the surf zone. Quantifying nearshore circulation patterns using low-cost image platforms and open-source computer vision algorithms presents the potential to further our understanding of fundamental surf zone dynamics.


2012 ◽  
Vol 10 (H16) ◽  
pp. 679-680
Author(s):  
Christopher J. Fluke

AbstractAs we move ever closer to the Square Kilometre Array era, support for real-time, interactive visualisation and analysis of tera-scale (and beyond) data cubes will be crucial for on-going knowledge discovery. However, the data-on-the-desktop approach to analysis and visualisation that most astronomers are comfortable with will no longer be feasible: tera-scale data volumes exceed the memory and processing capabilities of standard desktop computing environments. Instead, there will be an increasing need for astronomers to utilise remote high performance computing (HPC) resources. In recent years, the graphics processing unit (GPU) has emerged as a credible, low cost option for HPC. A growing number of supercomputing centres are now investing heavily in GPU technologies to provide O(100) Teraflop/s processing. I describe how a GPU-powered computing cluster allows us to overcome the analysis and visualisation challenges of tera-scale data. With a GPU-based architecture, we have moved the bottleneck from processing-limited to bandwidth-limited, achieving exceptional real-time performance for common visualisation and data analysis tasks.


2021 ◽  
Vol 12 ◽  
Author(s):  
Yosuke Toda ◽  
Toshiaki Tameshige ◽  
Masakazu Tomiyama ◽  
Toshinori Kinoshita ◽  
Kentaro K. Shimizu

Recent technical advances in the computer-vision domain have facilitated the development of various methods for achieving image-based quantification of stomata-related traits. However, the installation cost of such a system and the difficulties of operating it on-site have been hurdles for experimental biologists. Here, we present a platform that allows real-time stomata detection during microscopic observation. The proposed system consists of a deep neural network model-based stomata detector and an upright microscope connected to a USB camera and a graphics processing unit (GPU)-supported single-board computer. All the hardware components are commercially available at common electronic commerce stores at a reasonable price. Moreover, the machine-learning model is prepared based on freely available cloud services. This approach allows users to set up a phenotyping platform at low cost. As a proof of concept, we trained our model to detect dumbbell-shaped stomata from wheat leaf imprints. Using this platform, we collected a comprehensive range of stomatal phenotypes from wheat leaves. We confirmed notable differences in stomatal density (SD) between adaxial and abaxial surfaces and in stomatal size (SS) between wheat-related species of different ploidy. Utilizing such a platform is expected to accelerate research that involves all aspects of stomata phenotyping.


2021 ◽  
Author(s):  
Airidas Korolkovas ◽  
Alexander Katsevich ◽  
Michael Frenkel ◽  
William Thompson ◽  
Edward Morton

X-ray computed tomography (CT) can provide 3D images of density, and possibly the atomic number, for large objects like passenger luggage. This information, while generally very useful, is often insufficient to identify threats like explosives and narcotics, which can have a similar average composition as benign everyday materials such as plastics, glass, light metals, etc. A much more specific material signature can be measured with X-ray diffraction (XRD). Unfortunately, XRD signal is very faint compared to the transmitted one, and also challenging to reconstruct for objects larger than a small laboratory sample. In this article we analyze a novel low-cost scanner design which captures CT and XRD signals simultaneously, and uses the least possible collimation to maximize the flux. To simulate a realistic instrument, we derive a formula for the resolution of any diffraction pathway, taking into account the polychromatic spectrum, and the finite size of the source, detector, and each voxel. We then show how to reconstruct XRD patterns from a large phantom with multiple diffracting objects. Our approach includes a reasonable amount of photon counting noise (Poisson statistics), as well as measurement bias, in particular incoherent Compton scattering. The resolution of our reconstruction is sufficient to provide significantly more information than standard CT, thus increasing the accuracy of threat detection. Our theoretical model is implemented in GPU (Graphics Processing Unit) accelerated software which can be used to assess and further optimize scanner designs for specific applications in security, healthcare, and manufacturing quality control.


Sign in / Sign up

Export Citation Format

Share Document