Jump to content

Parareal: Difference between revisions

From Wikipedia, the free encyclopedia
Content deleted Content added
Danrup (talk | contribs)
First version of page describing the Parareal parallel-in-time integration method
(No difference)

Revision as of 10:47, 29 August 2015

Parareal is an algorithm from numerical analysis and used for the solution of initial value problems.[1] In contrast to e.g. Runge-Kutta or multi-step methods, some of the computations in Parareal can be performed in parallel and Parareal is therefore one example of a parallel-in-time integration method. While historically most efforts to parallelize the numerical solution of partial differential equations focussed on the spatial discretization, in view of the challenges from exascale computing, parallel methods for temporal discretization have been identified as a possible way to increase concurrency in numerical software.[2] Because Parareal computes the numerical solution for multiple time steps in parallel, it is categorized as a parallel across the steps method.[3] This is in contrast to approaches using parallelism across the method like parallel Runge-Kutta method where independent stages can be computed in parallel or parallel across the system methods like waveform relaxation.

History

Parareal can be derived as both a multigrid method in time method or as multiple shooting along the time axis.[4] Both ideas, multigrid in time as well as adopting multiple shooting for time integration, go back to the 1980s and 1990s.[5][6] Parareal is a widely studied method and has been used and modified for a range of different applications.[7] Ideas to parallelize the solution of initial value problems go back even further: the first paper proposing a parallel-in-time integration method appeared in 1964.[8]

Algorithm

Parareal solves an initial value problem of the form

Here, the right hand side can correspond to the spatial discretization of a partial differential equation in a method of lines approach.

Parareal now requires a decomposition of the time interval into so-called time slices such that

Each time slice is assigned to one processing unit when parallelizing the algorithm, so that is equal to the number of processing units used for Parareal: in an MPI based code for example, this would be the number of processes, while in an OpenMP based code, would be equal to the number of threads.

Parareal is based on the iterative application of two methods for integration of ordinary differential equations. One, commonly labeled , should be of high accuracy and computational cost while the other, typically labelled , must be computationally cheap but can be much less accurate. Typically, some form of Runge-Kutta method is chosen for both coarse and fine integrator, where might be of lower order and use a larger time step than . If the initial value problem stems from the discretization of a PDE, can also use a coarser spatial discretization. The result of numerical integration with one of these methods over a time slice for some starting value given at is then written as

or .

Serial time integration with the fine method would then correspond to a step by step computation of

Parareal instead uses the following iteration

where is the iteration counter. As the iteration converges and , the terms from the coarse method cancel out and Parareal reproduces the solution that is obtained by the serial execution of the fine method only. It can be shown that Parareal converges after a maximum of iterations.[4]. For Parareal to provide speedup, however, it has to convergence in a number of iterations significantly smaller than the number of time slices, that is .

In the Parareal iteration, the computationally expensive evaluation of can be performed in parallel on processing units. By contrast, the dependency of on means that the coarse correction has to be computed in serial order.

Speedup

Under some assumptions, a simple theoretical model for the speedup of Parareal can be derived.[9] Although in applications these assumptions can be too restrictive, the model still is useful to illustrate the trade offs that are involved in obtaining speedup with Parareal.

First, assume that every time slice consists of exactly steps of the fine integrator and of steps of the coarse integrator. This includes in particular the assumption that all time slices are of identical length and that both coarse and fine integrator use a constant step size over the full simulation. Second, denote by and the computing time required for a single step of the coarse or fine method and assume that both are constant. This is typically not exactly true when an implicit method is used, because then runtimes vary depending on the number of iterations required by the iterative solver.

Under these two assumptions, the runtime for the fine method integrating over time slices can be modelled as

The runtime of Parareal using processing units and performing iterations is

Speedup of Parareal then is

These two bounds illustrate the trade off that has to be made in choosing the coarse method: one the one hand, it has to be cheap and/or use a much larger time step to make the first bound as large as possible, on the other hand the number of iterations has to be kept low to keep the second bound large. In particular, Parareal's parallel efficiency is bounded by

that is by the inverse of the number of required iterations.

References

  1. ^ Lions, Jacques-Louis; Maday, Yvon; Turinici, Gabriel (2015). "A "parareal" in time discretization of PDE's". Comptes Rendus de l’Académie des Sciences - Series I - Mathematics. 332 (7). Elsevier: 661–668. doi:10.1016/S0764-4442(00)01793-6. Retrieved August 2015. {{cite journal}}: Check date values in: |access-date= (help)
  2. ^ Jack Dongarra (March 2014). Applied Mathematics Research for Exascale Computing (PDF) (Report). US Department of Energy. Retrieved August 2015. {{cite report}}: Check date values in: |accessdate= (help); Unknown parameter |coauthors= ignored (|author= suggested) (help)
  3. ^ Burrage, Kevin (1997). "Parallel methods for ODEs". Advances in Computational Mathematics. 7 (1–2). Kluwer Academic Publishers: 1–31. doi:10.1023/A:1018997130884. {{cite journal}}: |access-date= requires |url= (help); Check date values in: |access-date= (help)
  4. ^ a b Gander, Martin J.; Vandewalle, Stefan (2007). "Analysis of the Parareal Time‐Parallel Time‐Integration Method". SIAM Journal on Scientific Computing. 29 (2). SIAM: 556–578. doi:10.1137/05064607X. Retrieved August 2015. {{cite journal}}: Check date values in: |access-date= (help)
  5. ^ Hackbusch, Wolfgang (1985). "Parabolic multi-grid methods". Computing Methods in Applied Sciences and Engineering, VI. North-Holland Publishing Co: 189–197. Retrieved August 2015. {{cite journal}}: Check date values in: |access-date= (help)
  6. ^ Kiehl, Martin (1994). "Parallel multiple shooting for the solution of initial value problems". Parallel Computing. 20 (3). Elsevier: 275–295. doi:10.1016/S0167-8191(06)80013-X. Retrieved August 2015. {{cite journal}}: Check date values in: |access-date= (help)
  7. ^ Gander, Martin J. (2015). 50 years of Time Parallel Time Integration. Contributions in Mathematical and Computational Sciences. Vol. 9 (1 ed.). Springer International Publishing. doi:10.1007/978-3-319-23321-5. ISBN 978-3-319-23321-5.
  8. ^ Nievergelt, Jürg (1964). "Parallel methods for integrating ordinary differential equations". Communications of the ACM. 7 (12). ACM New York: 731–733. doi:10.1145/355588.365137. Retrieved August 2015. {{cite journal}}: Check date values in: |access-date= (help)
  9. ^ Minion, Michael L. (2010). "A Hybrid Parareal Spectral Deferred Corrections Method". Communications in Applied Mathematics and Computational Science. 5 (2): 265–301. doi:10.2140/camcos.2010.5.265. Retrieved August 2015. {{cite journal}}: Check date values in: |access-date= (help)