1#ifndef AMREX_RUNGE_KUTTA_H_
2#define AMREX_RUNGE_KUTTA_H_
3#include <AMReX_Config.H>
52 template <
typename MF>
53 std::enable_if_t<IsFabArray<MF>::value>
operator() (
int, MF&)
const {}
59void rk_update (MF& Unew, MF
const& Uold, MF
const& dUdt, Real dt)
61 auto const& snew = Unew.arrays();
62 auto const& sold = Uold.const_arrays();
63 auto const& sdot = dUdt.const_arrays();
65 (
int bi,
int i,
int j,
int k,
int n)
noexcept
67 snew[bi](i,j,k,n) = sold[bi](i,j,k,n) + dt*sdot[bi](i,j,k,n);
74void rk_update (MF& Unew, MF
const& Uold, MF
const& dUdt1, MF
const& dUdt2, Real dt)
76 auto const& snew = Unew.arrays();
77 auto const& sold = Uold.const_arrays();
78 auto const& sdot1 = dUdt1.const_arrays();
79 auto const& sdot2 = dUdt2.const_arrays();
81 (
int bi,
int i,
int j,
int k,
int n)
noexcept
83 snew[bi](i,j,k,n) = sold[bi](i,j,k,n) + dt*(sdot1[bi](i,j,k,n) +
91void rk2_update_2 (MF& Unew, MF
const& Uold, MF
const& dUdt, Real dt)
93 auto const& snew = Unew.arrays();
94 auto const& sold = Uold.const_arrays();
95 auto const& sdot = dUdt.const_arrays();
97 (
int bi,
int i,
int j,
int k,
int n)
noexcept
99 snew[bi](i,j,k,n) = Real(0.5)*(snew[bi](i,j,k,n) +
101 sdot[bi](i,j,k,n) * dt);
107template <
typename MF>
110 auto const& snew = Unew.arrays();
111 auto const& sold = Uold.const_arrays();
112 auto const& k1 = rkk[0].const_arrays();
113 auto const& k2 = rkk[1].const_arrays();
114 auto const& k3 = rkk[2].const_arrays();
116 (
int bi,
int i,
int j,
int k,
int n)
noexcept
118 snew[bi](i,j,k,n) = sold[bi](i,j,k,n)
119 + dt6 * (k1[bi](i,j,k,n) + k2[bi](i,j,k,n)
120 + Real(4.) * k3[bi](i,j,k,n));
126template <
typename MF>
129 auto const& snew = Unew.arrays();
130 auto const& sold = Uold.const_arrays();
131 auto const& k1 = rkk[0].const_arrays();
132 auto const& k2 = rkk[1].const_arrays();
133 auto const& k3 = rkk[2].const_arrays();
134 auto const& k4 = rkk[3].const_arrays();
136 (
int bi,
int i,
int j,
int k,
int n)
noexcept
138 snew[bi](i,j,k,n) = sold[bi](i,j,k,n)
139 + dt6 * ( k1[bi](i,j,k,n) + k4[bi](i,j,k,n)
140 + Real(2.)*(k2[bi](i,j,k,n) + k3[bi](i,j,k,n)));
157template <
typename MF,
typename F,
typename FB,
typename P = PostStageNoOp>
158void RK2 (MF& Uold, MF& Unew, Real time, Real dt,
F const& frhs, FB
const& fillbndry,
163 MF dUdt(Unew.boxArray(), Unew.DistributionMap(), Unew.nComp(), 0,
164 MFInfo(), Unew.Factory());
167 fillbndry(1, Uold, time);
168 frhs(1, dUdt, Uold, time, Real(0.5)*dt);
174 fillbndry(2, Unew, time+dt);
175 frhs(2, dUdt, Unew, time, Real(0.5)*dt);
194template <
typename MF,
typename F,
typename FB,
typename R,
195 typename P = PostStageNoOp>
196void RK3 (MF& Uold, MF& Unew, Real time, Real dt,
F const& frhs, FB
const& fillbndry,
197 R
const& store_crse_data, P
const& post_stage =
PostStageNoOp())
202 for (
auto& mf : rkk) {
203 mf.define(Unew.boxArray(), Unew.DistributionMap(), Unew.nComp(), 0,
204 MFInfo(), Unew.Factory());
208 fillbndry(1, Uold, time);
209 frhs(1, rkk[0], Uold, time, dt/Real(6.));
215 fillbndry(2, Unew, time+dt);
216 frhs(2, rkk[1], Unew, time+dt, dt/Real(6.));
222 Real t_half = time + Real(0.5)*dt;
223 fillbndry(3, Unew, t_half);
224 frhs(3, rkk[2], Unew, t_half, dt*Real(2./3.));
229 store_crse_data(rkk);
244template <
typename MF,
typename F,
typename FB,
typename R,
245 typename P = PostStageNoOp>
246void RK4 (MF& Uold, MF& Unew, Real time, Real dt,
F const& frhs, FB
const& fillbndry,
247 R
const& store_crse_data, P
const& post_stage =
PostStageNoOp())
252 for (
auto& mf : rkk) {
253 mf.define(Unew.boxArray(), Unew.DistributionMap(), Unew.nComp(), 0,
254 MFInfo(), Unew.Factory());
258 fillbndry(1, Uold, time);
259 frhs(1, rkk[0], Uold, time, dt/Real(6.));
265 Real t_half = time + Real(0.5)*dt;
266 fillbndry(2, Unew, t_half);
267 frhs(2, rkk[1], Unew, t_half, dt/Real(3.));
273 fillbndry(3, Unew, t_half);
274 frhs(3, rkk[2], Unew, t_half, dt/Real(3.));
280 fillbndry(4, Unew, time+dt);
281 frhs(4, rkk[3], Unew, time+dt, dt/Real(6.));
286 store_crse_data(rkk);
#define BL_PROFILE(a)
Definition AMReX_BLProfiler.H:551
#define AMREX_GPU_DEVICE
Definition AMReX_GpuQualifiers.H:18
void streamSynchronize() noexcept
Definition AMReX_GpuDevice.H:237
void rk3_update_3(MF &Unew, MF const &Uold, Array< MF, 3 > const &rkk, Real dt6)
Unew = Uold + (k1 + k2 + 4*k3) * dt6, where dt6 = dt/6.
Definition AMReX_RungeKutta.H:108
void rk2_update_2(MF &Unew, MF const &Uold, MF const &dUdt, Real dt)
Unew = (Uold+Unew)/2 + dUdt * dt/2.
Definition AMReX_RungeKutta.H:91
void rk4_update_4(MF &Unew, MF const &Uold, Array< MF, 4 > const &rkk, Real dt6)
Unew = Uold + (k1+k4+2*(k2+k3))*dt6, where dt6 = dt/6.
Definition AMReX_RungeKutta.H:127
void rk_update(MF &Unew, MF const &Uold, MF const &dUdt, Real dt)
Unew = Uold + dUdt * dt.
Definition AMReX_RungeKutta.H:59
Functions for Runge-Kutta methods.
Definition AMReX_RungeKutta.H:49
void RK2(MF &Uold, MF &Unew, Real time, Real dt, F const &frhs, FB const &fillbndry, P const &post_stage=PostStageNoOp())
Time stepping with RK2.
Definition AMReX_RungeKutta.H:158
void RK3(MF &Uold, MF &Unew, Real time, Real dt, F const &frhs, FB const &fillbndry, R const &store_crse_data, P const &post_stage=PostStageNoOp())
Time stepping with RK3.
Definition AMReX_RungeKutta.H:196
std::enable_if_t< std::is_integral_v< T > > ParallelFor(TypeList< CTOs... > ctos, std::array< int, sizeof...(CTOs)> const &runtime_options, T N, F &&f)
Definition AMReX_CTOParallelForImpl.H:191
IntVectND< AMREX_SPACEDIM > IntVect
Definition AMReX_BaseFwd.H:30
std::array< T, N > Array
Definition AMReX_Array.H:24
Definition AMReX_FabArrayCommI.H:896
FabArray memory allocation information.
Definition AMReX_FabArray.H:66
Definition AMReX_RungeKutta.H:51
std::enable_if_t< IsFabArray< MF >::value > operator()(int, MF &) const
Definition AMReX_RungeKutta.H:53