En poursuivant votre navigation sur ce site, vous acceptez l'utilisation d'un simple cookie d'identification. Aucune autre exploitation n'est faite de ce cookie. OK

Documents 65N22 7 résultats

Filtrer
Sélectionner : Tous / Aucun
Q
Déposez votre fichier ici pour le déplacer vers cet enregistrement.
y

Linear solvers for reservoir simulation - Hénon, Pascal (Auteur de la Conférence) | CIRM H

Multi angle

In this presentation, we will first present the main goals and principles of reservoir simulation. Then we will focus on linear systems that arise in such simulation. The main HPC challenge is to solve those systems efficiently on massively parallel computers. The specificity of those systems is that their convergence is mostly governed by the elliptic part of the equations and the linear solver needs to take advantage of it to be efficient. The reference method in reservoir simulation is CPR-AMG which usually relies on AMG to solve the quasi elliptic part of the system. We will present some works on improving AMG scalability for the reservoir linear systems (work done in collaboration with CERFACS). We will then introduce an on-going work with INRIA to take advantage of their enlarged Krylov method (EGMRES) in the CPR method.[-]
In this presentation, we will first present the main goals and principles of reservoir simulation. Then we will focus on linear systems that arise in such simulation. The main HPC challenge is to solve those systems efficiently on massively parallel computers. The specificity of those systems is that their convergence is mostly governed by the elliptic part of the equations and the linear solver needs to take advantage of it to be efficient. The ...[+]

65F10 ; 65N22 ; 65Y05

Sélection Signaler une erreur
Déposez votre fichier ici pour le déplacer vers cet enregistrement.
y
We review how to bound the error between the unknown weak solution of a PDE and its numerical approximation via a fully computable a posteriori estimate. We focus on approximations obtained at an arbitrary step of a linearization (Newton-Raphson, fixed point, ...) and algebraic solver (conjugate gradients, multigrid, domain decomposition, ...). Identifying the discretization, linearization, and algebraic error components, we design local stopping criteria which keep them in balance. This gives rise to a fully adaptive inexact Newton method. Numerical experiments are presented in confirmation of the theory.[-]
We review how to bound the error between the unknown weak solution of a PDE and its numerical approximation via a fully computable a posteriori estimate. We focus on approximations obtained at an arbitrary step of a linearization (Newton-Raphson, fixed point, ...) and algebraic solver (conjugate gradients, multigrid, domain decomposition, ...). Identifying the discretization, linearization, and algebraic error components, we design local ...[+]

65N15 ; 65N22 ; 65Y05

Sélection Signaler une erreur
Déposez votre fichier ici pour le déplacer vers cet enregistrement.
y
I will present an efficient implementation of the highly robust and scalable GenEO preconditioner in the high-performance PDE framework DUNE. The GenEO coarse space is constructed by combining low energy solutions of local generalised eigenproblems using a partition of unity. In this talk, both weak and strong scaling for the GenEO solver on over 15,000 cores will be demonstrated by solving an industrially motivated problem with over 200 million degrees of freedom in aerospace composites modelling. Further, it will be shown that for highly complex parameter distributions in certain real-world applications, established methods can become intractable while GenEO remains fully effective. In the context of multilevel Markov chain Monte Carlo (MLMCMC), the GenEO coarse space also plays an important role as an effective surrogate model in PDE-constrained Bayesian inference. The second part will therefore focus on the approximation properties of the GenEO coarse space and on a high-performance parallel implementation of MLMCMC.
This is joint work with Tim Dodwell (Exeter), Anne Reinarz (TU Munich) and Linus Seelinger (Heidelberg).[-]
I will present an efficient implementation of the highly robust and scalable GenEO preconditioner in the high-performance PDE framework DUNE. The GenEO coarse space is constructed by combining low energy solutions of local generalised eigenproblems using a partition of unity. In this talk, both weak and strong scaling for the GenEO solver on over 15,000 cores will be demonstrated by solving an industrially motivated problem with over 200 million ...[+]

65F08 ; 65N22 ; 65N30 ; 65N55

Sélection Signaler une erreur
Déposez votre fichier ici pour le déplacer vers cet enregistrement.
y
Both multigrid and domain decomposition methods are so called optimal solvers for Laplace type problems, but how do they compare? I will start by showing in what sense these methods are optimal for the Laplace equation, which will reveal that while both multigrid and domain decomposition are iterative solvers, there are fundamental differences between them. Multigrid for Laplace's equation is a standalone solver, while classical domain decomposition methods like the additive Schwarz method or Neumann-Neumann and FETI methods need Krylov acceleration to work. I will explain in detail for each case why this is so, and then also present modifications so that Krylov acceleration is not necessary any more. For overlapping methods, this leads to the use of partitions of unity, while for non-overlapping methods, the coarse space can be a remedy. Good coarse spaces in domain decomposition methods are very different from coarse spaces in multigrid, due to the very aggressive coarsening in domain decomposition. I will introduce the concept of optimal coarse spaces for domain decomposition in a sense very different from the optimal above, and then present approximations of this coarse space. Together with optimized transmission conditions, this leads to a two level domain decomposition method of Schwarz type which is competitive with multigrid for Laplace's equation in wallclock time.[-]
Both multigrid and domain decomposition methods are so called optimal solvers for Laplace type problems, but how do they compare? I will start by showing in what sense these methods are optimal for the Laplace equation, which will reveal that while both multigrid and domain decomposition are iterative solvers, there are fundamental differences between them. Multigrid for Laplace's equation is a standalone solver, while classical domain ...[+]

65N55 ; 65N22 ; 65F10

Sélection Signaler une erreur
Déposez votre fichier ici pour le déplacer vers cet enregistrement.
y
Iterative methods for linear systems were invented for the same reasons as they are used today,namely to reduce computational cost. Gauss states in a letter to his friend Gerling in 1823: 'you will in the future hardly eliminate directly, at least not when you have more than two unknowns'.
Richardson's paper from 1910 was then very influential, and is a model of a modern numerical analysis paper: modeling, discretization, approximate solution of the discrete problem,and a real application. Richardson's method is much more sophisticated that how it is usually presented today, and his dream became reality in the PhD thesis of Gene Golub.
The work of Stiefel, Hestenes and Lanczos in the early 1950 sparked the success story of Krylov methods, and these methods can also be understood in the context of extrapolation, pioneered by Brezinski and Sidi, based on seminal work by Wynn.
This brings us to the modern iterative methods for solving partial differential equations,which come in two main classes: domain decomposition methods and multigrid methods. Domain decomposition methods go back to the alternating Schwarz method invented by Herman Amandus Schwarz in 1869 to close a gap in the proof of Riemann's famous Mapping Theorem. Multigrid goes back to the seminal work by Fedorenko in 1961, with main contributions by Brandt and Hackbusch in the Seventies.
I will show in my presentation how these methods function on the same model problem ofthe temperature distribution in a simple room. All these methods are today used as preconditioners for Krylov methods, which leads to the most powerful iterative solvers currently knownfor linear systems.[-]
Iterative methods for linear systems were invented for the same reasons as they are used today,namely to reduce computational cost. Gauss states in a letter to his friend Gerling in 1823: 'you will in the future hardly eliminate directly, at least not when you have more than two unknowns'.
Richardson's paper from 1910 was then very influential, and is a model of a modern numerical analysis paper: modeling, discretization, approximate solution of ...[+]

65N22 ; 65F10 ; 65B05 ; 65-02 ; 65-03

Sélection Signaler une erreur
Déposez votre fichier ici pour le déplacer vers cet enregistrement.
y
We construct a hierarchy of hybrid numerical methods for multi-scale kinetic equations based on moment realizability matrices, a concept introduced by Levermore, Morokoff and Nadiga. Following such a criterion, one can consider hybrid scheme where the hydrodynamic part is given either by the compressible Euler or Navier-Stokes equations, or even with more general models, such as the Burnett or super-Burnett systems.
PDE - numerical methods - Boltzmann equation - fluid models - hybrid methods[-]
We construct a hierarchy of hybrid numerical methods for multi-scale kinetic equations based on moment realizability matrices, a concept introduced by Levermore, Morokoff and Nadiga. Following such a criterion, one can consider hybrid scheme where the hydrodynamic part is given either by the compressible Euler or Navier-Stokes equations, or even with more general models, such as the Burnett or super-Burnett systems.
PDE - numerical methods - ...[+]

35Q35 ; 65N08 ; 65N22

Sélection Signaler une erreur
Déposez votre fichier ici pour le déplacer vers cet enregistrement.
y
In domain decomposition methods, most of the computational cost lies in the successive solutions of the local problems in subdomains via forward-backward substitutions and in the orthogonalization of interface search directions. All these operations are performed, in the best case, via BLAS-1 or BLAS-2 routines which are inefficient on multicore systems with hierarchical memory. A way to improve the parallel efficiency of the method consists in working with several search directions, since multiple forward-backward substitutions and reorthogonalizations involve BLAS-3 routines. In the case of a problem with several right-hand-sides, using a block Krylov method is a straightforward way to work with multiple search directions. This will be illustrated with an application in electromagnetism using FETI-2LM method. For problems with a single right-hand-side, deriving several search directions that make sense from the optimal one constructed by the Krylov method is not so easy. The recently developed S-FETI method gives a very good approach that does not only improve parallel efficiency but can also reduce the global computational cost in the case of very heterogeneous problems.[-]
In domain decomposition methods, most of the computational cost lies in the successive solutions of the local problems in subdomains via forward-backward substitutions and in the orthogonalization of interface search directions. All these operations are performed, in the best case, via BLAS-1 or BLAS-2 routines which are inefficient on multicore systems with hierarchical memory. A way to improve the parallel efficiency of the method consists in ...[+]

65N22 ; 65N30 ; 65N55 ; 65Y05 ; 65F10

Sélection Signaler une erreur