Increased space-parallelism via time-simultaneous Newton-multigrid methods for nonstationary nonlinear PDE problems

Author(s):  
Jonas Dünnebacke ◽  
Stefan Turek ◽  
Christoph Lohmann ◽  
Andriy Sokolov ◽  
Peter Zajac

We discuss how “parallel-in-space & simultaneous-in-time” Newton-multigrid approaches can be designed which improve the scaling behavior of the spatial parallelism by reducing the latency costs. The idea is to solve many time steps at once and therefore solving fewer but larger systems. These large systems are reordered and interpreted as a space-only problem leading to multigrid algorithm with semi-coarsening in space and line smoothing in time direction. The smoother is further improved by embedding it as a preconditioner in a Krylov subspace method. As a prototypical application, we concentrate on scalar partial differential equations (PDEs) with up to many thousands of time steps which are discretized in time, resp., space by finite difference, resp., finite element methods. For linear PDEs, the resulting method is closely related to multigrid waveform relaxation and its theoretical framework. In our parabolic test problems the numerical behavior of this multigrid approach is robust w.r.t. the spatial and temporal grid size and the number of simultaneously treated time steps. Moreover, we illustrate how corresponding time-simultaneous fixed-point and Newton-type solvers can be derived for nonlinear nonstationary problems that require the described solution of linearized problems in each outer nonlinear step. As the main result, we are able to generate much larger problem sizes to be treated by a large number of cores so that the combination of the robustly scaling multigrid solvers together with a larger degree of parallelism allows a faster solution procedure for nonstationary problems.

Author(s):  
Yuka Hashimoto ◽  
Takashi Nodera

AbstractThe Krylov subspace method has been investigated and refined for approximating the behaviors of finite or infinite dimensional linear operators. It has been used for approximating eigenvalues, solutions of linear equations, and operator functions acting on vectors. Recently, for time-series data analysis, much attention is being paid to the Krylov subspace method as a viable method for estimating the multiplications of a vector by an unknown linear operator referred to as a transfer operator. In this paper, we investigate a convergence analysis for Krylov subspace methods for estimating operator-vector multiplications.


Author(s):  
Shin-ichi Ito ◽  
Takeru Matsuda ◽  
Yuto Miyatake

AbstractWe consider a scalar function depending on a numerical solution of an initial value problem, and its second-derivative (Hessian) matrix for the initial value. The need to extract the information of the Hessian or to solve a linear system having the Hessian as a coefficient matrix arises in many research fields such as optimization, Bayesian estimation, and uncertainty quantification. From the perspective of memory efficiency, these tasks often employ a Krylov subspace method that does not need to hold the Hessian matrix explicitly and only requires computing the multiplication of the Hessian and a given vector. One of the ways to obtain an approximation of such Hessian-vector multiplication is to integrate the so-called second-order adjoint system numerically. However, the error in the approximation could be significant even if the numerical integration to the second-order adjoint system is sufficiently accurate. This paper presents a novel algorithm that computes the intended Hessian-vector multiplication exactly and efficiently. For this aim, we give a new concise derivation of the second-order adjoint system and show that the intended multiplication can be computed exactly by applying a particular numerical method to the second-order adjoint system. In the discussion, symplectic partitioned Runge–Kutta methods play an essential role.


Author(s):  
Sheldon Tan ◽  
Mehdi Tahoori ◽  
Taeyoung Kim ◽  
Shengcheng Wang ◽  
Zeyu Sun ◽  
...  

Author(s):  
Gaoxin Zhou ◽  
Zhi Gang

In recent years, high order harmonic (or eigenvector) of neutron diffusion equation has been widely used in on-line monitoring system of reactor power. There are two kinds of calculation method to solve the equation: corrected power iteration method and Krylov subspace methods. Fu Li used the corrected power iteration method. When solving for the ith harmonic, it tries to eliminate the influence of the front harmonics using the orthogonality of the harmonic function. But its convergence speed depends on the occupation ratio. When the dominant ratios equal to 1 or close to 1, convergence speed of fixed source iteration method is slow or convergence can’t be achieved. Another method is the Krylov subspace method, the main idea of this method is to project the eigenvalue and eigenvector of large-scale matrix to a small one. Then we can solve the small matrix eigenvalue and eigenvector to get the large ones. In recent years, the restart Arnoldi method emerged as a development of Krylov subspace method. The method uses continuous reboot Arnoldi decomposition, limiting expanding subspace, and the orthogonality of the subspace is guaranteed using orthogonalization method. This paper studied the refined algorithms, a method based on the Krylov subspace method of solving eigenvalue problem for large sparse matrix of neutron diffusion equation. Two improvements have been made for a restarted Arnoldi method. One is that using an ingenious linear combination of the refined Ritz vector forms an initial vector and then generates a new Krylov subspace. Another is that retaining the refined Ritz vector in the new subspace, called, augmented Krylov subspace. This way retains useful information and makes the resulting algorithm converge faster. Several numerical examples are the new algorithm with the implicitly restart Arnoldi algorithm (IRA) and the implicitly restarted refined Arnoldi algorithm (IRRA). Numerical results confirm efficiency of the new algorithm.


2014 ◽  
Vol 2014 ◽  
pp. 1-15 ◽  
Author(s):  
Yury Gryazin

We consider an efficient iterative approach to the solution of the discrete Helmholtz equation with Dirichlet, Neumann, and Sommerfeld-like boundary conditions based on a compact sixth order approximation scheme and lower order preconditioned Krylov subspace methodology. The resulting systems of finite-difference equations are solved by different preconditioned Krylov subspace-based methods. In the analysis of the lower order preconditioning developed here, we introduce the term “kth order preconditioned matrix” in addition to the commonly used “an optimal preconditioner.” The necessity of the new criterion is justified by the fact that the condition number of the preconditioned matrix in some of our test problems improves with the decrease of the grid step size. In a simple 1D case, we are able to prove this analytically. This new parameter could serve as a guide in the construction of new preconditioners. The lower order direct preconditioner used in our algorithms is based on a combination of the separation of variables technique and fast Fourier transform (FFT) type methods. The resulting numerical methods allow efficient implementation on parallel computers. Numerical results confirm the high efficiency of the proposed iterative approach.


Sign in / Sign up

Export Citation Format

Share Document