1#ifndef AMREX_FFT_HELPER_H_
2#define AMREX_FFT_HELPER_H_
3#include <AMReX_Config.H>
22#if defined(AMREX_USE_CUDA)
24# include <cuComplex.h>
25#elif defined(AMREX_USE_HIP)
26# if __has_include(<rocfft/rocfft.h>)
27# include <rocfft/rocfft.h>
31# include <hip/hip_complex.h>
32#elif defined(AMREX_USE_SYCL)
33# if __has_include(<oneapi/mkl/dft.hpp>)
34# include <oneapi/mkl/dft.hpp>
36# define AMREX_USE_MKL_DFTI_2024 1
37# include <oneapi/mkl/dfti.hpp>
84 int nprocs = std::numeric_limits<int>::max();
136namespace detail {
void hip_execute (rocfft_plan plan,
void **in,
void **out); }
144template <
typename T, Direction direction,
typename P,
typename TI,
typename TO>
145void sycl_execute (P* plan, TI* in, TO* out)
147#ifndef AMREX_USE_MKL_DFTI_2024
148 std::int64_t workspaceSize = 0;
150 std::size_t workspaceSize = 0;
152 plan->get_value(oneapi::mkl::dft::config_param::WORKSPACE_BYTES,
155 plan->set_workspace(buffer);
157 if (std::is_same_v<TI,TO>) {
159 if constexpr (direction == Direction::forward) {
160 r = oneapi::mkl::dft::compute_forward(*plan, out);
162 r = oneapi::mkl::dft::compute_backward(*plan, out);
165 if constexpr (direction == Direction::forward) {
166 r = oneapi::mkl::dft::compute_forward(*plan, in, out);
168 r = oneapi::mkl::dft::compute_backward(*plan, in, out);
181#if defined(AMREX_USE_CUDA)
184 cuComplex, cuDoubleComplex>;
185#elif defined(AMREX_USE_HIP)
187 using VendorComplex = std::conditional_t<std::is_same_v<float,T>,
189#elif defined(AMREX_USE_SYCL)
190 using mkl_desc_r = oneapi::mkl::dft::descriptor<std::is_same_v<float,T>
191 ? oneapi::mkl::dft::precision::SINGLE
192 : oneapi::mkl::dft::precision::DOUBLE,
193 oneapi::mkl::dft::domain::REAL>;
194 using mkl_desc_c = oneapi::mkl::dft::descriptor<std::is_same_v<float,T>
195 ? oneapi::mkl::dft::precision::SINGLE
196 : oneapi::mkl::dft::precision::DOUBLE,
197 oneapi::mkl::dft::domain::COMPLEX>;
198 using VendorPlan = std::variant<mkl_desc_r*,mkl_desc_c*>;
201 using VendorPlan = std::conditional_t<std::is_same_v<float,T>,
202 fftwf_plan, fftw_plan>;
203 using VendorComplex = std::conditional_t<std::is_same_v<float,T>,
204 fftwf_complex, fftw_complex>;
240#if !defined(AMREX_USE_GPU)
257 template <Direction D>
262 int rank = is_2d_transform ? 2 : 1;
277 int nr = (rank == 1) ? len[0] : len[0]*len[1];
279 int nc = (rank == 1) ? (len[0]/2+1) : (len[1]/2+1)*len[0];
280#if (AMREX_SPACEDIM == 1)
290#if defined(AMREX_USE_CUDA)
294 std::size_t work_size;
296 cufftType fwd_type = std::is_same_v<float,T> ? CUFFT_R2C : CUFFT_D2Z;
298 (cufftMakePlanMany(
plan, rank, len,
nullptr, 1, nr,
nullptr, 1, nc, fwd_type,
howmany, &work_size));
300 cufftType bwd_type = std::is_same_v<float,T> ? CUFFT_C2R : CUFFT_Z2D;
302 (cufftMakePlanMany(
plan, rank, len,
nullptr, 1, nc,
nullptr, 1, nr, bwd_type,
howmany, &work_size));
305#elif defined(AMREX_USE_HIP)
307 auto prec = std::is_same_v<float,T> ? rocfft_precision_single : rocfft_precision_double;
309 std::size_t
length[2] = {std::size_t(len[1]), std::size_t(len[0])};
311 AMREX_ROCFFT_SAFE_CALL
312 (rocfft_plan_create(&
plan, rocfft_placement_notinplace,
313 rocfft_transform_type_real_forward, prec, rank,
316 AMREX_ROCFFT_SAFE_CALL
317 (rocfft_plan_create(&
plan, rocfft_placement_notinplace,
318 rocfft_transform_type_real_inverse, prec, rank,
322#elif defined(AMREX_USE_SYCL)
326 pp =
new mkl_desc_r(len[0]);
328 pp =
new mkl_desc_r({std::int64_t(len[0]), std::int64_t(len[1])});
330#ifndef AMREX_USE_MKL_DFTI_2024
331 pp->set_value(oneapi::mkl::dft::config_param::PLACEMENT,
332 oneapi::mkl::dft::config_value::NOT_INPLACE);
334 pp->set_value(oneapi::mkl::dft::config_param::PLACEMENT, DFTI_NOT_INPLACE);
336 pp->set_value(oneapi::mkl::dft::config_param::NUMBER_OF_TRANSFORMS,
howmany);
337 pp->set_value(oneapi::mkl::dft::config_param::FWD_DISTANCE, nr);
338 pp->set_value(oneapi::mkl::dft::config_param::BWD_DISTANCE, nc);
339 std::vector<std::int64_t> strides;
340 strides.push_back(0);
341 if (rank == 2) { strides.push_back(len[1]); }
342 strides.push_back(1);
343#ifndef AMREX_USE_MKL_DFTI_2024
344 pp->set_value(oneapi::mkl::dft::config_param::FWD_STRIDES, strides);
347 pp->set_value(oneapi::mkl::dft::config_param::FWD_STRIDES, strides.data());
350 pp->set_value(oneapi::mkl::dft::config_param::WORKSPACE,
351 oneapi::mkl::dft::config_value::WORKSPACE_EXTERNAL);
352 pp->commit(amrex::Gpu::Device::streamQueue());
357 if constexpr (std::is_same_v<float,T>) {
359 plan = fftwf_plan_many_dft_r2c
360 (rank, len,
howmany, pr,
nullptr, 1, nr, pc,
nullptr, 1, nc,
361 FFTW_ESTIMATE | FFTW_DESTROY_INPUT);
363 plan = fftwf_plan_many_dft_c2r
364 (rank, len,
howmany, pc,
nullptr, 1, nc, pr,
nullptr, 1, nr,
365 FFTW_ESTIMATE | FFTW_DESTROY_INPUT);
369 plan = fftw_plan_many_dft_r2c
370 (rank, len,
howmany, pr,
nullptr, 1, nr, pc,
nullptr, 1, nc,
371 FFTW_ESTIMATE | FFTW_DESTROY_INPUT);
373 plan = fftw_plan_many_dft_c2r
374 (rank, len,
howmany, pc,
nullptr, 1, nc, pr,
nullptr, 1, nr,
375 FFTW_ESTIMATE | FFTW_DESTROY_INPUT);
390 template <Direction D,
int M>
408 template <Direction D>
426#if (AMREX_SPACEDIM >= 2)
427 else if (ndims == 2) {
429#if (AMREX_SPACEDIM == 2)
437#if (AMREX_SPACEDIM == 3)
438 else if (ndims == 3) {
448#if defined(AMREX_USE_CUDA)
452 cufftType t = std::is_same_v<float,T> ? CUFFT_C2C : CUFFT_Z2Z;
453 std::size_t work_size;
455 (cufftMakePlanMany(
plan, ndims, len,
nullptr, 1,
n,
nullptr, 1,
n, t,
howmany, &work_size));
457#elif defined(AMREX_USE_HIP)
459 auto prec = std::is_same_v<float,T> ? rocfft_precision_single
460 : rocfft_precision_double;
462 : rocfft_transform_type_complex_inverse;
466 }
else if (ndims == 2) {
474 AMREX_ROCFFT_SAFE_CALL
475 (rocfft_plan_create(&
plan, rocfft_placement_inplace, dir, prec, ndims,
478#elif defined(AMREX_USE_SYCL)
482 pp =
new mkl_desc_c(
n);
483 }
else if (ndims == 2) {
484 pp =
new mkl_desc_c({std::int64_t(len[0]), std::int64_t(len[1])});
486 pp =
new mkl_desc_c({std::int64_t(len[0]), std::int64_t(len[1]), std::int64_t(len[2])});
488#ifndef AMREX_USE_MKL_DFTI_2024
489 pp->set_value(oneapi::mkl::dft::config_param::PLACEMENT,
490 oneapi::mkl::dft::config_value::INPLACE);
492 pp->set_value(oneapi::mkl::dft::config_param::PLACEMENT, DFTI_INPLACE);
494 pp->set_value(oneapi::mkl::dft::config_param::NUMBER_OF_TRANSFORMS,
howmany);
495 pp->set_value(oneapi::mkl::dft::config_param::FWD_DISTANCE,
n);
496 pp->set_value(oneapi::mkl::dft::config_param::BWD_DISTANCE,
n);
497 std::vector<std::int64_t> strides(ndims+1);
500 for (
int i = ndims-1; i >= 1; --i) {
501 strides[i] = strides[i+1] * len[i];
503#ifndef AMREX_USE_MKL_DFTI_2024
504 pp->set_value(oneapi::mkl::dft::config_param::FWD_STRIDES, strides);
505 pp->set_value(oneapi::mkl::dft::config_param::BWD_STRIDES, strides);
507 pp->set_value(oneapi::mkl::dft::config_param::FWD_STRIDES, strides.data());
508 pp->set_value(oneapi::mkl::dft::config_param::BWD_STRIDES, strides.data());
510 pp->set_value(oneapi::mkl::dft::config_param::WORKSPACE,
511 oneapi::mkl::dft::config_value::WORKSPACE_EXTERNAL);
512 pp->commit(amrex::Gpu::Device::streamQueue());
517 if constexpr (std::is_same_v<float,T>) {
519 plan = fftwf_plan_many_dft
520 (ndims, len,
howmany, p,
nullptr, 1,
n, p,
nullptr, 1,
n, -1,
523 plan = fftwf_plan_many_dft
524 (ndims, len,
howmany, p,
nullptr, 1,
n, p,
nullptr, 1,
n, +1,
529 plan = fftw_plan_many_dft
530 (ndims, len,
howmany, p,
nullptr, 1,
n, p,
nullptr, 1,
n, -1,
533 plan = fftw_plan_many_dft
534 (ndims, len,
howmany, p,
nullptr, 1,
n, p,
nullptr, 1,
n, +1,
542 template <Direction D>
543 fftw_r2r_kind get_fftw_kind (std::pair<Boundary,Boundary>
const& bc)
563 return fftw_r2r_kind{};
575 template <Direction D>
609 template <Direction D>
610 void init_r2r (
Box const& box, T* p, std::pair<Boundary,Boundary>
const& bc,
611 int howmany_initval = 1)
615 kind = get_r2r_kind<D>(bc);
623#if defined(AMREX_USE_GPU)
643 int nc = (nex/2) + 1;
645#if defined (AMREX_USE_CUDA)
649 cufftType fwd_type = std::is_same_v<float,T> ? CUFFT_R2C : CUFFT_D2Z;
650 std::size_t work_size;
652 (cufftMakePlanMany(
plan, 1, &nex,
nullptr, 1, nc*2,
nullptr, 1, nc, fwd_type,
howmany, &work_size));
654#elif defined(AMREX_USE_HIP)
657 auto prec = std::is_same_v<float,T> ? rocfft_precision_single : rocfft_precision_double;
658 const std::size_t
length = nex;
659 AMREX_ROCFFT_SAFE_CALL
660 (rocfft_plan_create(&
plan, rocfft_placement_inplace,
661 rocfft_transform_type_real_forward, prec, 1,
664#elif defined(AMREX_USE_SYCL)
666 auto*
pp =
new mkl_desc_r(nex);
667#ifndef AMREX_USE_MKL_DFTI_2024
668 pp->set_value(oneapi::mkl::dft::config_param::PLACEMENT,
669 oneapi::mkl::dft::config_value::INPLACE);
671 pp->set_value(oneapi::mkl::dft::config_param::PLACEMENT, DFTI_INPLACE);
673 pp->set_value(oneapi::mkl::dft::config_param::NUMBER_OF_TRANSFORMS,
howmany);
674 pp->set_value(oneapi::mkl::dft::config_param::FWD_DISTANCE, nc*2);
675 pp->set_value(oneapi::mkl::dft::config_param::BWD_DISTANCE, nc);
676 std::vector<std::int64_t> strides = {0,1};
677#ifndef AMREX_USE_MKL_DFTI_2024
678 pp->set_value(oneapi::mkl::dft::config_param::FWD_STRIDES, strides);
679 pp->set_value(oneapi::mkl::dft::config_param::BWD_STRIDES, strides);
681 pp->set_value(oneapi::mkl::dft::config_param::FWD_STRIDES, strides.data());
682 pp->set_value(oneapi::mkl::dft::config_param::BWD_STRIDES, strides.data());
684 pp->set_value(oneapi::mkl::dft::config_param::WORKSPACE,
685 oneapi::mkl::dft::config_value::WORKSPACE_EXTERNAL);
686 pp->commit(amrex::Gpu::Device::streamQueue());
692 auto fftw_kind = get_fftw_kind<D>(bc);
693 if constexpr (std::is_same_v<float,T>) {
694 plan = fftwf_plan_many_r2r
695 (1, &
n,
howmany, p,
nullptr, 1,
n, p,
nullptr, 1,
n, &fftw_kind,
698 plan = fftw_plan_many_r2r
699 (1, &
n,
howmany, p,
nullptr, 1,
n, p,
nullptr, 1,
n, &fftw_kind,
712 template <Direction D>
714 std::pair<Boundary,Boundary>
const& bc)
720#if defined(AMREX_USE_GPU)
722 init_r2r<D>(box, p, bc, 2);
727 kind = get_r2r_kind<D>(bc);
736 auto fftw_kind = get_fftw_kind<D>(bc);
737 if constexpr (std::is_same_v<float,T>) {
738 plan = fftwf_plan_many_r2r
739 (1, &
n,
howmany, p,
nullptr, 2,
n*2, p,
nullptr, 2,
n*2, &fftw_kind,
741 plan2 = fftwf_plan_many_r2r
742 (1, &
n,
howmany, p+1,
nullptr, 2,
n*2, p+1,
nullptr, 2,
n*2, &fftw_kind,
745 plan = fftw_plan_many_r2r
746 (1, &
n,
howmany, p,
nullptr, 2,
n*2, p,
nullptr, 2,
n*2, &fftw_kind,
748 plan2 = fftw_plan_many_r2r
749 (1, &
n,
howmany, p+1,
nullptr, 2,
n*2, p+1,
nullptr, 2,
n*2, &fftw_kind,
760 template <Direction D>
771#if defined(AMREX_USE_CUDA)
774 std::size_t work_size = 0;
781 if constexpr (std::is_same_v<float,T>) {
787 if constexpr (std::is_same_v<float,T>) {
795#elif defined(AMREX_USE_HIP)
796 detail::hip_execute(
plan, (
void**)&pi, (
void**)&po);
797#elif defined(AMREX_USE_SYCL)
798 detail::sycl_execute<T,D>(std::get<0>(
plan), pi, po);
801 if constexpr (std::is_same_v<float,T>) {
812 template <Direction D>
820#if defined(AMREX_USE_CUDA)
823 std::size_t work_size = 0;
830 if constexpr (std::is_same_v<float,T>) {
837#elif defined(AMREX_USE_HIP)
838 detail::hip_execute(
plan, (
void**)&p, (
void**)&p);
839#elif defined(AMREX_USE_SYCL)
840 detail::sycl_execute<T,D>(std::get<1>(
plan), p, p);
843 if constexpr (std::is_same_v<float,T>) {
866 amrex::Abort(
"FFT: alloc_scratch_space: unsupported kind");
884 auto*
pdst = (T*) pbuf;
887 int ostride = (
n+1)*2;
895 auto batch = ielem /
Long(nex);
896 auto i =
int(ielem - batch*nex);
897 for (
int ir = 0; ir < 2; ++ir) {
898 auto* po =
pdst + (2*batch+ir)*ostride + i;
899 auto const* pi = psrc + 2*batch*istride + ir;
903 *po = sign * pi[(2*norig-1-i)*2];
910 auto batch = ielem /
Long(nex);
911 auto i =
int(ielem - batch*nex);
912 auto* po =
pdst + batch*ostride + i;
913 auto const* pi = psrc + batch*istride;
917 *po = sign * pi[2*norig-1-i];
922 int ostride = (2*
n+1)*2;
930 auto batch = ielem /
Long(nex);
931 auto i =
int(ielem - batch*nex);
932 for (
int ir = 0; ir < 2; ++ir) {
933 auto* po =
pdst + (2*batch+ir)*ostride + i;
934 auto const* pi = psrc + 2*batch*istride + ir;
937 }
else if (i < (2*norig-1)) {
938 *po = pi[(2*norig-2-i)*2];
939 }
else if (i == (2*norig-1)) {
941 }
else if (i < (3*norig)) {
942 *po = -pi[(i-2*norig)*2];
943 }
else if (i < (4*norig-1)) {
944 *po = -pi[(4*norig-2-i)*2];
953 auto batch = ielem /
Long(nex);
954 auto i =
int(ielem - batch*nex);
955 auto* po =
pdst + batch*ostride + i;
956 auto const* pi = psrc + batch*istride;
959 }
else if (i < (2*norig-1)) {
960 *po = pi[2*norig-2-i];
961 }
else if (i == (2*norig-1)) {
963 }
else if (i < (3*norig)) {
964 *po = -pi[i-2*norig];
965 }
else if (i < (4*norig-1)) {
966 *po = -pi[4*norig-2-i];
973 int ostride = (2*
n+1)*2;
981 auto batch = ielem /
Long(nex);
982 auto i =
int(ielem - batch*nex);
983 for (
int ir = 0; ir < 2; ++ir) {
984 auto* po =
pdst + (2*batch+ir)*ostride + i;
985 auto const* pi = psrc + 2*batch*istride + ir;
988 }
else if (i == norig) {
990 }
else if (i < (2*norig+1)) {
991 *po = -pi[(2*norig-i)*2];
992 }
else if (i < (3*norig)) {
993 *po = -pi[(i-2*norig)*2];
994 }
else if (i == 3*norig) {
997 *po = pi[(4*norig-i)*2];
1004 auto batch = ielem /
Long(nex);
1005 auto i =
int(ielem - batch*nex);
1006 auto* po =
pdst + batch*ostride + i;
1007 auto const* pi = psrc + batch*istride;
1010 }
else if (i == norig) {
1012 }
else if (i < (2*norig+1)) {
1013 *po = -pi[2*norig-i];
1014 }
else if (i < (3*norig)) {
1015 *po = -pi[i-2*norig];
1016 }
else if (i == 3*norig) {
1019 *po = pi[4*norig-i];
1024 int ostride = (2*
n+1)*2;
1032 auto batch = ielem /
Long(nex);
1033 auto i =
int(ielem - batch*nex);
1034 for (
int ir = 0; ir < 2; ++ir) {
1035 auto* po =
pdst + (2*batch+ir)*ostride + i;
1036 auto const* pi = psrc + 2*batch*istride + ir;
1039 }
else if (i < (2*norig)) {
1040 *po = -pi[(2*norig-1-i)*2];
1041 }
else if (i < (3*norig)) {
1042 *po = -pi[(i-2*norig)*2];
1044 *po = pi[(4*norig-1-i)*2];
1051 auto batch = ielem /
Long(nex);
1052 auto i =
int(ielem - batch*nex);
1053 auto* po =
pdst + batch*ostride + i;
1054 auto const* pi = psrc + batch*istride;
1057 }
else if (i < (2*norig)) {
1058 *po = -pi[2*norig-1-i];
1059 }
else if (i < (3*norig)) {
1060 *po = -pi[i-2*norig];
1062 *po = pi[4*norig-1-i];
1067 int ostride = (2*
n+1)*2;
1075 auto batch = ielem /
Long(nex);
1076 auto i =
int(ielem - batch*nex);
1077 for (
int ir = 0; ir < 2; ++ir) {
1078 auto* po =
pdst + (2*batch+ir)*ostride + i;
1079 auto const* pi = psrc + 2*batch*istride + ir;
1082 }
else if (i < (2*norig)) {
1083 *po = pi[(2*norig-1-i)*2];
1084 }
else if (i < (3*norig)) {
1085 *po = -pi[(i-2*norig)*2];
1087 *po = -pi[(4*norig-1-i)*2];
1094 auto batch = ielem /
Long(nex);
1095 auto i =
int(ielem - batch*nex);
1096 auto* po =
pdst + batch*ostride + i;
1097 auto const* pi = psrc + batch*istride;
1100 }
else if (i < (2*norig)) {
1101 *po = pi[2*norig-1-i];
1102 }
else if (i < (3*norig)) {
1103 *po = -pi[i-2*norig];
1105 *po = -pi[4*norig-1-i];
1110 amrex::Abort(
"FFT: pack_r2r_buffer: unsupported kind");
1132 auto batch = ielem /
Long(norig);
1133 auto k =
int(ielem - batch*norig);
1135 for (
int ir = 0; ir < 2; ++ir) {
1136 auto const& yk = psrc[(2*batch+ir)*istride+k+1];
1137 pdst[2*batch*ostride+ir+k*2] = s * yk.real() - c * yk.imag();
1143 auto batch = ielem /
Long(norig);
1144 auto k =
int(ielem - batch*norig);
1146 auto const& yk = psrc[batch*istride+k+1];
1147 pdst[batch*ostride+k] = s * yk.real() - c * yk.imag();
1151 int istride = 2*
n+1;
1155 auto batch = ielem /
Long(norig);
1156 auto k =
int(ielem - batch*norig);
1158 for (
int ir = 0; ir < 2; ++ir) {
1159 auto const& yk = psrc[(2*batch+ir)*istride+2*k+1];
1160 pdst[2*batch*ostride+ir+k*2] = T(0.5)*(s * yk.real() - c * yk.imag());
1166 auto batch = ielem /
Long(norig);
1167 auto k =
int(ielem - batch*norig);
1169 auto const& yk = psrc[batch*istride+2*k+1];
1170 pdst[batch*ostride+k] = T(0.5)*(s * yk.real() - c * yk.imag());
1178 auto batch = ielem /
Long(norig);
1179 auto k =
int(ielem - batch*norig);
1181 for (
int ir = 0; ir < 2; ++ir) {
1182 auto const& yk = psrc[(2*batch+ir)*istride+k];
1183 pdst[2*batch*ostride+ir+k*2] = c * yk.real() + s * yk.imag();
1189 auto batch = ielem /
Long(norig);
1190 auto k =
int(ielem - batch*norig);
1192 auto const& yk = psrc[batch*istride+k];
1193 pdst[batch*ostride+k] = c * yk.real() + s * yk.imag();
1197 int istride = 2*
n+1;
1201 auto batch = ielem /
Long(norig);
1202 auto k =
int(ielem - batch*norig);
1203 for (
int ir = 0; ir < 2; ++ir) {
1204 auto const& yk = psrc[(2*batch+ir)*istride+2*k+1];
1205 pdst[2*batch*ostride+ir+k*2] = T(0.5) * yk.real();
1211 auto batch = ielem /
Long(norig);
1212 auto k =
int(ielem - batch*norig);
1213 auto const& yk = psrc[batch*istride+2*k+1];
1214 pdst[batch*ostride+k] = T(0.5) * yk.real();
1218 int istride = 2*
n+1;
1222 auto batch = ielem /
Long(norig);
1223 auto k =
int(ielem - batch*norig);
1225 for (
int ir = 0; ir < 2; ++ir) {
1226 auto const& yk = psrc[(2*batch+ir)*istride+2*k+1];
1227 pdst[2*batch*ostride+ir+k*2] = T(0.5) * (c * yk.real() + s * yk.imag());
1233 auto batch = ielem /
Long(norig);
1234 auto k =
int(ielem - batch*norig);
1236 auto const& yk = psrc[batch*istride+2*k+1];
1237 pdst[batch*ostride+k] = T(0.5) * (c * yk.real() + s * yk.imag());
1241 int istride = 2*
n+1;
1245 auto batch = ielem /
Long(norig);
1246 auto k =
int(ielem - batch*norig);
1248 for (
int ir = 0; ir < 2; ++ir) {
1249 auto const& yk = psrc[(2*batch+ir)*istride+2*k+1];
1250 pdst[2*batch*ostride+ir+k*2] = T(0.5) * (s * yk.real() - c * yk.imag());
1256 auto batch = ielem /
Long(norig);
1257 auto k =
int(ielem - batch*norig);
1259 auto const& yk = psrc[batch*istride+2*k+1];
1260 pdst[batch*ostride+k] = T(0.5) * (s * yk.real() - c * yk.imag());
1264 amrex::Abort(
"FFT: unpack_r2r_buffer: unsupported kind");
1272 template <Direction D>
1278#if defined(AMREX_USE_GPU)
1284#if defined(AMREX_USE_CUDA)
1288 std::size_t work_size = 0;
1294 if constexpr (std::is_same_v<float,T>) {
1300#elif defined(AMREX_USE_HIP)
1301 detail::hip_execute(
plan, (
void**)&pscratch, (
void**)&pscratch);
1302#elif defined(AMREX_USE_SYCL)
1303 detail::sycl_execute<T,Direction::forward>(std::get<0>(
plan), (T*)pscratch, (
VendorComplex*)pscratch);
1310#if defined(AMREX_USE_CUDA)
1316 if constexpr (std::is_same_v<float,T>) {
1317 fftwf_execute(
plan);
1332#if defined(AMREX_USE_CUDA)
1334#elif defined(AMREX_USE_HIP)
1335 AMREX_ROCFFT_SAFE_CALL(rocfft_plan_destroy(
plan));
1336#elif defined(AMREX_USE_SYCL)
1337 std::visit([](
auto&& p) {
delete p; },
plan);
1339 if constexpr (std::is_same_v<float,T>) {
1340 fftwf_destroy_plan(
plan);
1342 fftw_destroy_plan(
plan);
1355 PlanD* get_vendor_plan_d (Key
const& key);
1356 PlanF* get_vendor_plan_f (Key
const& key);
1358 void add_vendor_plan_d (Key
const& key, PlanD plan);
1359 void add_vendor_plan_f (Key
const& key, PlanF plan);
1363template <
typename T>
1364template <Direction D,
int M>
1375 for (
auto s : fft_size) { n *= s; }
1378#if defined(AMREX_USE_GPU)
1379 Key key = {fft_size.template expand<3>(), ncomp, D, kind};
1382 if constexpr (std::is_same_v<float,T>) {
1383 cached_plan = detail::get_vendor_plan_f(key);
1385 cached_plan = detail::get_vendor_plan_d(key);
1388 plan = *cached_plan;
1397 for (
int i = 0; i < M; ++i) {
1398 len[i] = fft_size[M-1-i];
1401 int nc = fft_size[0]/2+1;
1402 for (
int i = 1; i < M; ++i) {
1406#if defined(AMREX_USE_CUDA)
1413 type = std::is_same_v<float,T> ? CUFFT_R2C : CUFFT_D2Z;
1417 type = std::is_same_v<float,T> ? CUFFT_C2R : CUFFT_Z2D;
1421 std::size_t work_size;
1423 (cufftMakePlanMany(plan, M, len,
nullptr, 1, n_in,
nullptr, 1, n_out, type, howmany, &work_size));
1425#elif defined(AMREX_USE_HIP)
1427 auto prec = std::is_same_v<float,T> ? rocfft_precision_single : rocfft_precision_double;
1429 for (
int idim = 0; idim < M; ++idim) {
length[idim] = fft_size[idim]; }
1431 AMREX_ROCFFT_SAFE_CALL
1432 (rocfft_plan_create(&plan, rocfft_placement_notinplace,
1433 rocfft_transform_type_real_forward, prec, M,
1434 length, howmany,
nullptr));
1436 AMREX_ROCFFT_SAFE_CALL
1437 (rocfft_plan_create(&plan, rocfft_placement_notinplace,
1438 rocfft_transform_type_real_inverse, prec, M,
1439 length, howmany,
nullptr));
1442#elif defined(AMREX_USE_SYCL)
1446 pp =
new mkl_desc_r(fft_size[0]);
1448 std::vector<std::int64_t> len64(M);
1449 for (
int idim = 0; idim < M; ++idim) {
1450 len64[idim] = len[idim];
1452 pp =
new mkl_desc_r(len64);
1454#ifndef AMREX_USE_MKL_DFTI_2024
1455 pp->set_value(oneapi::mkl::dft::config_param::PLACEMENT,
1456 oneapi::mkl::dft::config_value::NOT_INPLACE);
1458 pp->set_value(oneapi::mkl::dft::config_param::PLACEMENT, DFTI_NOT_INPLACE);
1460 pp->set_value(oneapi::mkl::dft::config_param::NUMBER_OF_TRANSFORMS, howmany);
1461 pp->set_value(oneapi::mkl::dft::config_param::FWD_DISTANCE, n);
1462 pp->set_value(oneapi::mkl::dft::config_param::BWD_DISTANCE, nc);
1463 std::vector<std::int64_t> strides(M+1);
1466 for (
int i = M-1; i >= 1; --i) {
1467 strides[i] = strides[i+1] * fft_size[M-1-i];
1470#ifndef AMREX_USE_MKL_DFTI_2024
1471 pp->set_value(oneapi::mkl::dft::config_param::FWD_STRIDES, strides);
1474 pp->set_value(oneapi::mkl::dft::config_param::FWD_STRIDES, strides.data());
1477 pp->set_value(oneapi::mkl::dft::config_param::WORKSPACE,
1478 oneapi::mkl::dft::config_value::WORKSPACE_EXTERNAL);
1479 pp->commit(amrex::Gpu::Device::streamQueue());
1484 if (pf ==
nullptr || pb ==
nullptr) {
1489 if constexpr (std::is_same_v<float,T>) {
1491 plan = fftwf_plan_many_dft_r2c
1492 (M, len, howmany, (
float*)pf,
nullptr, 1, n, (fftwf_complex*)pb,
nullptr, 1, nc,
1495 plan = fftwf_plan_many_dft_c2r
1496 (M, len, howmany, (fftwf_complex*)pb,
nullptr, 1, nc, (
float*)pf,
nullptr, 1, n,
1501 plan = fftw_plan_many_dft_r2c
1502 (M, len, howmany, (
double*)pf,
nullptr, 1, n, (fftw_complex*)pb,
nullptr, 1, nc,
1505 plan = fftw_plan_many_dft_c2r
1506 (M, len, howmany, (fftw_complex*)pb,
nullptr, 1, nc, (
double*)pf,
nullptr, 1, n,
1512#if defined(AMREX_USE_GPU)
1514 if constexpr (std::is_same_v<float,T>) {
1515 detail::add_vendor_plan_f(key, plan);
1517 detail::add_vendor_plan_d(key, plan);
1528 template <
typename FA>
1529 typename FA::FABType::value_type * get_fab (FA& fa)
1531 auto myproc = ParallelContext::MyProcSub();
1532 if (myproc < fa.size()) {
1533 return fa.fabPtr(myproc);
1539 template <
typename FA1,
typename FA2>
1540 std::unique_ptr<char,DataDeleter> make_mfs_share (FA1& fa1, FA2& fa2)
1542 bool not_same_fa =
true;
1543 if constexpr (std::is_same_v<FA1,FA2>) {
1544 not_same_fa = (&fa1 != &fa2);
1546 using FAB1 =
typename FA1::FABType::value_type;
1547 using FAB2 =
typename FA2::FABType::value_type;
1548 using T1 =
typename FAB1::value_type;
1549 using T2 =
typename FAB2::value_type;
1550 auto myproc = ParallelContext::MyProcSub();
1551 bool alloc_1 = (myproc < fa1.size());
1552 bool alloc_2 = (myproc < fa2.size()) && not_same_fa;
1554 if (alloc_1 && alloc_2) {
1555 Box const& box1 = fa1.fabbox(myproc);
1556 Box const& box2 = fa2.fabbox(myproc);
1557 int ncomp1 = fa1.nComp();
1558 int ncomp2 = fa2.nComp();
1560 sizeof(T2)*box2.numPts()*ncomp2));
1561 fa1.setFab(myproc, FAB1(box1, ncomp1, (T1*)p));
1562 fa2.setFab(myproc, FAB2(box2, ncomp2, (T2*)p));
1563 }
else if (alloc_1) {
1564 Box const& box1 = fa1.fabbox(myproc);
1565 int ncomp1 = fa1.nComp();
1567 fa1.setFab(myproc, FAB1(box1, ncomp1, (T1*)p));
1568 }
else if (alloc_2) {
1569 Box const& box2 = fa2.fabbox(myproc);
1570 int ncomp2 = fa2.nComp();
1572 fa2.setFab(myproc, FAB2(box2, ncomp2, (T2*)p));
1576 return std::unique_ptr<char,DataDeleter>((
char*)p, DataDeleter{
The_Arena()});
1585 [[nodiscard]]
constexpr Dim3 operator() (Dim3 i)
const noexcept
1587 return {i.y, i.x, i.z};
1590 static constexpr Dim3 Inverse (Dim3 i)
1592 return {i.y, i.x, i.z};
1595 [[nodiscard]]
constexpr IndexType operator() (IndexType it)
const noexcept
1600 static constexpr IndexType Inverse (IndexType it)
1608 [[nodiscard]]
constexpr Dim3 operator() (Dim3 i)
const noexcept
1610 return {i.z, i.y, i.x};
1613 static constexpr Dim3 Inverse (Dim3 i)
1615 return {i.z, i.y, i.x};
1618 [[nodiscard]]
constexpr IndexType operator() (IndexType it)
const noexcept
1623 static constexpr IndexType Inverse (IndexType it)
1632 [[nodiscard]]
constexpr Dim3 operator() (Dim3 i)
const noexcept
1634 return {i.y, i.z, i.x};
1638 static constexpr Dim3 Inverse (Dim3 i)
1640 return {i.z, i.x, i.y};
1643 [[nodiscard]]
constexpr IndexType operator() (IndexType it)
const noexcept
1648 static constexpr IndexType Inverse (IndexType it)
1657 [[nodiscard]]
constexpr Dim3 operator() (Dim3 i)
const noexcept
1659 return {i.z, i.x, i.y};
1663 static constexpr Dim3 Inverse (Dim3 i)
1665 return {i.y, i.z, i.x};
1668 [[nodiscard]]
constexpr IndexType operator() (IndexType it)
const noexcept
1673 static constexpr IndexType Inverse (IndexType it)
1686 explicit SubHelper (Box
const& domain);
1688 [[nodiscard]]
Box make_box (Box
const& box)
const;
1690 [[nodiscard]] Periodicity make_periodicity (Periodicity
const& period)
const;
1692 [[nodiscard]]
bool ghost_safe (IntVect
const& ng)
const;
1695 [[nodiscard]]
IntVect make_iv (IntVect
const& iv)
const;
1698 [[nodiscard]]
IntVect make_safe_ghost (IntVect
const& ng)
const;
1700 [[nodiscard]] BoxArray inverse_boxarray (BoxArray
const& ba)
const;
1702 [[nodiscard]]
IntVect inverse_order (IntVect
const& order)
const;
1704 template <
typename T>
1705 [[nodiscard]] T make_array (T
const& a)
const
1707#if (AMREX_SPACEDIM == 1)
1710#elif (AMREX_SPACEDIM == 2)
1711 if (m_case == case_1n) {
1712 return T{a[1],a[0]};
1717 if (m_case == case_11n) {
1718 return T{a[2],a[0],a[1]};
1719 }
else if (m_case == case_1n1) {
1720 return T{a[1],a[0],a[2]};
1721 }
else if (m_case == case_1nn) {
1722 return T{a[1],a[2],a[0]};
1723 }
else if (m_case == case_n1n) {
1724 return T{a[0],a[2],a[1]};
1731 [[nodiscard]] GpuArray<int,3> xyz_order ()
const;
1733 template <
typename FA>
1734 FA make_alias_mf (FA
const& mf)
1736 BoxList bl = mf.boxArray().boxList();
1737 for (
auto& b : bl) {
1740 auto const& ng = make_iv(mf.nGrowVect());
1741 FA submf(BoxArray(std::move(bl)), mf.DistributionMap(), mf.nComp(), ng, MFInfo{}.SetAlloc(
false));
1742 using FAB =
typename FA::fab_type;
1743 for (MFIter mfi(submf, MFItInfo().DisableDeviceSync()); mfi.isValid(); ++mfi) {
1744 submf.setFab(mfi, FAB(mfi.fabbox(), mf.nComp(), mf[mfi].dataPtr()));
1749#if (AMREX_SPACEDIM == 2)
1750 enum Case { case_1n, case_other };
1751 int m_case = case_other;
1752#elif (AMREX_SPACEDIM == 3)
1753 enum Case { case_11n, case_1n1, case_1nn, case_n1n, case_other };
1754 int m_case = case_other;
#define AMREX_ENUM(CLASS,...)
Definition AMReX_Enum.H:208
#define AMREX_CUFFT_SAFE_CALL(call)
Definition AMReX_GpuError.H:92
#define AMREX_GPU_DEVICE
Definition AMReX_GpuQualifiers.H:18
amrex::ParmParse pp
Input file parser instance for the given namespace.
Definition AMReX_HypreIJIface.cpp:15
Real * pdst
Definition AMReX_HypreMLABecLap.cpp:1090
#define AMREX_D_TERM(a, b, c)
Definition AMReX_SPACE.H:172
virtual void free(void *pt)=0
A pure virtual function for deleting the arena pointed to by pt.
virtual void * alloc(std::size_t sz)=0
__host__ __device__ IntVectND< dim > length() const noexcept
Return the length of the BoxND.
Definition AMReX_Box.H:154
Calculates the distribution of FABs to MPI processes.
Definition AMReX_DistributionMapping.H:43
An Integer Vector in dim-Dimensional Space.
Definition AMReX_IntVect.H:57
amrex_long Long
Definition AMReX_INT.H:30
void ParallelForOMP(T n, L const &f) noexcept
Performance-portable kernel launch function with optional OpenMP threading.
Definition AMReX_GpuLaunch.H:319
__host__ __device__ Dim3 length(Array4< T > const &a) noexcept
Return the spatial extents of an Array4 in Dim3 form.
Definition AMReX_Array4.H:1345
Arena * The_Arena()
Definition AMReX_Arena.cpp:783
Definition AMReX_FFT_Helper.H:52
Direction
Definition AMReX_FFT_Helper.H:54
Boundary
Definition AMReX_FFT_Helper.H:58
DomainStrategy
Definition AMReX_FFT_Helper.H:56
Kind
Definition AMReX_FFT_Helper.H:60
void streamSynchronize() noexcept
Definition AMReX_GpuDevice.H:263
gpuStream_t gpuStream() noexcept
Definition AMReX_GpuDevice.H:244
__host__ __device__ std::pair< double, double > sincospi(double x)
Return sin(pi*x) and cos(pi*x) given x.
Definition AMReX_Math.H:204
__host__ __device__ void ignore_unused(const Ts &...)
This shuts up the compiler about unused variables.
Definition AMReX.H:139
BoxND< 3 > Box
Box is an alias for amrex::BoxND instantiated with AMREX_SPACEDIM.
Definition AMReX_BaseFwd.H:30
IndexTypeND< 3 > IndexType
IndexType is an alias for amrex::IndexTypeND instantiated with AMREX_SPACEDIM.
Definition AMReX_BaseFwd.H:36
IntVectND< 3 > IntVect
IntVect is an alias for amrex::IntVectND instantiated with AMREX_SPACEDIM.
Definition AMReX_BaseFwd.H:33
void Abort(const std::string &msg)
Print out message to cerr and exit via abort().
Definition AMReX.cpp:230
const int[]
Definition AMReX_BLProfiler.cpp:1664
Definition AMReX_FFT_Helper.H:64
bool twod_mode
Definition AMReX_FFT_Helper.H:75
Info & setNumProcs(int n)
Cap the number of MPI ranks used by FFT.
Definition AMReX_FFT_Helper.H:131
bool oned_mode
We might have a special twod_mode: nx or ny == 1 && nz > 1.
Definition AMReX_FFT_Helper.H:78
int batch_size
Batched FFT size. Only support in R2C, not R2X.
Definition AMReX_FFT_Helper.H:81
Info & setDomainStrategy(DomainStrategy s)
Select how the domain is decomposed across MPI ranks.
Definition AMReX_FFT_Helper.H:92
DomainStrategy domain_strategy
Domain composition strategy.
Definition AMReX_FFT_Helper.H:66
int nprocs
Max number of processes to use.
Definition AMReX_FFT_Helper.H:84
int pencil_threshold
Definition AMReX_FFT_Helper.H:70
Info & setOneDMode(bool x)
Flag the degenerate 2-D mode (nx==1 or ny==1) that still batches along z.
Definition AMReX_FFT_Helper.H:117
Info & setBatchSize(int bsize)
Specify the batch size for FFT.
Definition AMReX_FFT_Helper.H:124
Info & setPencilThreshold(int t)
Override the slab→pencil break-even threshold for the automatic strategy.
Definition AMReX_FFT_Helper.H:99
Info & setTwoDMode(bool x)
Restrict transforms to the first two dimensions (3-D problems only).
Definition AMReX_FFT_Helper.H:106
Definition AMReX_FFT_Helper.H:180
void * pf
Definition AMReX_FFT_Helper.H:215
void unpack_r2r_buffer(T *pdst, void const *pbuf) const
Collapse the spectral R2R result back into the original real layout.
Definition AMReX_FFT_Helper.H:1120
std::conditional_t< std::is_same_v< float, T >, cuComplex, cuDoubleComplex > VendorComplex
Definition AMReX_FFT_Helper.H:184
VendorPlan plan2
Definition AMReX_FFT_Helper.H:214
void init_r2c(IntVectND< M > const &fft_size, void *pbf, void *pbb, bool cache, int ncomp=1)
Initialize an M-dimensional batched real-to-complex plan.
Definition AMReX_FFT_Helper.H:1365
int n
Definition AMReX_FFT_Helper.H:207
void destroy()
Release any vendor FFT plan objects owned by this Plan.
Definition AMReX_FFT_Helper.H:234
bool defined2
Definition AMReX_FFT_Helper.H:212
void init_r2r(Box const &box, VendorComplex *pc, std::pair< Boundary, Boundary > const &bc)
Initialize a real-to-real plan that reads/writes complex storage.
Definition AMReX_FFT_Helper.H:713
void pack_r2r_buffer(void *pbuf, T const *psrc) const
Expand the real R2R input into the symmetry-extended buffer expected by CUFFT/rocFFT.
Definition AMReX_FFT_Helper.H:882
static void free_scratch_space(void *p)
Release GPU scratch allocated via alloc_scratch_space().
Definition AMReX_FFT_Helper.H:874
static void destroy_vendor_plan(VendorPlan plan)
Helper that destroys a vendor plan of the appropriate backend type.
Definition AMReX_FFT_Helper.H:1330
Kind get_r2r_kind(std::pair< Boundary, Boundary > const &bc)
Map boundary conditions to the Plan Kind for real-to-real transforms.
Definition AMReX_FFT_Helper.H:576
cufftHandle VendorPlan
Definition AMReX_FFT_Helper.H:182
Kind kind
Definition AMReX_FFT_Helper.H:209
void init_c2c(Box const &box, VendorComplex *p, int ncomp=1, int ndims=1)
Initialize a complex-to-complex plan across 1/2/3 dimensions.
Definition AMReX_FFT_Helper.H:409
int howmany
Definition AMReX_FFT_Helper.H:208
void * pb
Definition AMReX_FFT_Helper.H:216
void init_r2c(Box const &box, T *pr, VendorComplex *pc, bool is_2d_transform=false, int ncomp=1)
Initialize a 1-D or 2-D real-to-complex plan over the supplied Box.
Definition AMReX_FFT_Helper.H:258
void compute_r2r()
Execute the real-to-real plan, including GPU packing/unpacking.
Definition AMReX_FFT_Helper.H:1273
void compute_c2c()
Execute the complex-to-complex plan in place.
Definition AMReX_FFT_Helper.H:813
bool r2r_data_is_complex
Definition AMReX_FFT_Helper.H:210
void * alloc_scratch_space() const
Allocate GPU scratch space large enough to hold packed R2R data.
Definition AMReX_FFT_Helper.H:857
VendorPlan plan
Definition AMReX_FFT_Helper.H:213
void compute_r2c()
Execute the previously initialized real-to-complex plan.
Definition AMReX_FFT_Helper.H:761
bool defined
Definition AMReX_FFT_Helper.H:211
void set_ptrs(void *p0, void *p1)
Register device pointers used by the forward/backward executions.
Definition AMReX_FFT_Helper.H:225
void init_r2r(Box const &box, T *p, std::pair< Boundary, Boundary > const &bc, int howmany_initval=1)
Initialize a real-to-real (cosine/sine) plan that operates on real buffers.
Definition AMReX_FFT_Helper.H:610
A host / device complex number type, because std::complex doesn't work in device code with Cuda yet.
Definition AMReX_GpuComplex.H:30