1#ifndef AMREX_FFT_HELPER_H_
2#define AMREX_FFT_HELPER_H_
3#include <AMReX_Config.H>
22#if defined(AMREX_USE_CUDA)
24# include <cuComplex.h>
25#elif defined(AMREX_USE_HIP)
26# if __has_include(<rocfft/rocfft.h>)
27# include <rocfft/rocfft.h>
31# include <hip/hip_complex.h>
32#elif defined(AMREX_USE_SYCL)
33# if __has_include(<oneapi/mkl/dft.hpp>)
34# include <oneapi/mkl/dft.hpp>
36# define AMREX_USE_MKL_DFTI_2024 1
37# include <oneapi/mkl/dfti.hpp>
90 int nprocs = std::numeric_limits<int>::max();
142namespace detail {
void hip_execute (rocfft_plan plan,
void **in,
void **out); }
150inline void assert_no_external_stream ()
154 "SYCL FFT does not support external GPU streams.");
157template <
typename T, Direction direction,
typename P,
typename TI,
typename TO>
158void sycl_execute (P* plan, TI* in, TO* out)
160 assert_no_external_stream();
161#ifndef AMREX_USE_MKL_DFTI_2024
162 std::int64_t workspaceSize = 0;
164 std::size_t workspaceSize = 0;
166 plan->get_value(oneapi::mkl::dft::config_param::WORKSPACE_BYTES,
169 plan->set_workspace(buffer);
171 if (std::is_same_v<TI,TO>) {
173 if constexpr (direction == Direction::forward) {
174 r = oneapi::mkl::dft::compute_forward(*plan, out);
176 r = oneapi::mkl::dft::compute_backward(*plan, out);
179 if constexpr (direction == Direction::forward) {
180 r = oneapi::mkl::dft::compute_forward(*plan, in, out);
182 r = oneapi::mkl::dft::compute_backward(*plan, in, out);
195#if defined(AMREX_USE_CUDA)
198 cuComplex, cuDoubleComplex>;
199#elif defined(AMREX_USE_HIP)
201 using VendorComplex = std::conditional_t<std::is_same_v<float,T>,
203#elif defined(AMREX_USE_SYCL)
204 using mkl_desc_r = oneapi::mkl::dft::descriptor<std::is_same_v<float,T>
205 ? oneapi::mkl::dft::precision::SINGLE
206 : oneapi::mkl::dft::precision::DOUBLE,
207 oneapi::mkl::dft::domain::REAL>;
208 using mkl_desc_c = oneapi::mkl::dft::descriptor<std::is_same_v<float,T>
209 ? oneapi::mkl::dft::precision::SINGLE
210 : oneapi::mkl::dft::precision::DOUBLE,
211 oneapi::mkl::dft::domain::COMPLEX>;
212 using VendorPlan = std::variant<mkl_desc_r*,mkl_desc_c*>;
215 using VendorPlan = std::conditional_t<std::is_same_v<float,T>,
216 fftwf_plan, fftw_plan>;
217 using VendorComplex = std::conditional_t<std::is_same_v<float,T>,
218 fftwf_complex, fftw_complex>;
254#if !defined(AMREX_USE_GPU)
271 template <Direction D>
276 int rank = is_2d_transform ? 2 : 1;
291 int nr = (rank == 1) ? len[0] : len[0]*len[1];
293 int nc = (rank == 1) ? (len[0]/2+1) : (len[1]/2+1)*len[0];
294#if (AMREX_SPACEDIM == 1)
304#if defined(AMREX_USE_CUDA)
308 std::size_t work_size;
310 cufftType fwd_type = std::is_same_v<float,T> ? CUFFT_R2C : CUFFT_D2Z;
312 (cufftMakePlanMany(
plan, rank, len,
nullptr, 1, nr,
nullptr, 1, nc, fwd_type,
howmany, &work_size));
314 cufftType bwd_type = std::is_same_v<float,T> ? CUFFT_C2R : CUFFT_Z2D;
316 (cufftMakePlanMany(
plan, rank, len,
nullptr, 1, nc,
nullptr, 1, nr, bwd_type,
howmany, &work_size));
319#elif defined(AMREX_USE_HIP)
321 auto prec = std::is_same_v<float,T> ? rocfft_precision_single : rocfft_precision_double;
323 std::size_t
length[2] = {std::size_t(len[1]), std::size_t(len[0])};
325 AMREX_ROCFFT_SAFE_CALL
326 (rocfft_plan_create(&
plan, rocfft_placement_notinplace,
327 rocfft_transform_type_real_forward, prec, rank,
330 AMREX_ROCFFT_SAFE_CALL
331 (rocfft_plan_create(&
plan, rocfft_placement_notinplace,
332 rocfft_transform_type_real_inverse, prec, rank,
336#elif defined(AMREX_USE_SYCL)
340 pp =
new mkl_desc_r(len[0]);
342 pp =
new mkl_desc_r({std::int64_t(len[0]), std::int64_t(len[1])});
344#ifndef AMREX_USE_MKL_DFTI_2024
345 pp->set_value(oneapi::mkl::dft::config_param::PLACEMENT,
346 oneapi::mkl::dft::config_value::NOT_INPLACE);
348 pp->set_value(oneapi::mkl::dft::config_param::PLACEMENT, DFTI_NOT_INPLACE);
350 pp->set_value(oneapi::mkl::dft::config_param::NUMBER_OF_TRANSFORMS,
howmany);
351 pp->set_value(oneapi::mkl::dft::config_param::FWD_DISTANCE, nr);
352 pp->set_value(oneapi::mkl::dft::config_param::BWD_DISTANCE, nc);
353 std::vector<std::int64_t> strides;
354 strides.push_back(0);
355 if (rank == 2) { strides.push_back(len[1]); }
356 strides.push_back(1);
357#ifndef AMREX_USE_MKL_DFTI_2024
358 pp->set_value(oneapi::mkl::dft::config_param::FWD_STRIDES, strides);
361 pp->set_value(oneapi::mkl::dft::config_param::FWD_STRIDES, strides.data());
364 pp->set_value(oneapi::mkl::dft::config_param::WORKSPACE,
365 oneapi::mkl::dft::config_value::WORKSPACE_EXTERNAL);
366 detail::assert_no_external_stream();
367 pp->commit(amrex::Gpu::Device::streamQueue());
372 if constexpr (std::is_same_v<float,T>) {
374 plan = fftwf_plan_many_dft_r2c
375 (rank, len,
howmany, pr,
nullptr, 1, nr, pc,
nullptr, 1, nc,
376 FFTW_ESTIMATE | FFTW_DESTROY_INPUT);
378 plan = fftwf_plan_many_dft_c2r
379 (rank, len,
howmany, pc,
nullptr, 1, nc, pr,
nullptr, 1, nr,
380 FFTW_ESTIMATE | FFTW_DESTROY_INPUT);
384 plan = fftw_plan_many_dft_r2c
385 (rank, len,
howmany, pr,
nullptr, 1, nr, pc,
nullptr, 1, nc,
386 FFTW_ESTIMATE | FFTW_DESTROY_INPUT);
388 plan = fftw_plan_many_dft_c2r
389 (rank, len,
howmany, pc,
nullptr, 1, nc, pr,
nullptr, 1, nr,
390 FFTW_ESTIMATE | FFTW_DESTROY_INPUT);
405 template <Direction D,
int M>
423 template <Direction D>
441#if (AMREX_SPACEDIM >= 2)
442 else if (ndims == 2) {
444#if (AMREX_SPACEDIM == 2)
452#if (AMREX_SPACEDIM == 3)
453 else if (ndims == 3) {
463#if defined(AMREX_USE_CUDA)
467 cufftType t = std::is_same_v<float,T> ? CUFFT_C2C : CUFFT_Z2Z;
468 std::size_t work_size;
470 (cufftMakePlanMany(
plan, ndims, len,
nullptr, 1,
n,
nullptr, 1,
n, t,
howmany, &work_size));
472#elif defined(AMREX_USE_HIP)
474 auto prec = std::is_same_v<float,T> ? rocfft_precision_single
475 : rocfft_precision_double;
477 : rocfft_transform_type_complex_inverse;
481 }
else if (ndims == 2) {
489 AMREX_ROCFFT_SAFE_CALL
490 (rocfft_plan_create(&
plan, rocfft_placement_inplace, dir, prec, ndims,
493#elif defined(AMREX_USE_SYCL)
497 pp =
new mkl_desc_c(
n);
498 }
else if (ndims == 2) {
499 pp =
new mkl_desc_c({std::int64_t(len[0]), std::int64_t(len[1])});
501 pp =
new mkl_desc_c({std::int64_t(len[0]), std::int64_t(len[1]), std::int64_t(len[2])});
503#ifndef AMREX_USE_MKL_DFTI_2024
504 pp->set_value(oneapi::mkl::dft::config_param::PLACEMENT,
505 oneapi::mkl::dft::config_value::INPLACE);
507 pp->set_value(oneapi::mkl::dft::config_param::PLACEMENT, DFTI_INPLACE);
509 pp->set_value(oneapi::mkl::dft::config_param::NUMBER_OF_TRANSFORMS,
howmany);
510 pp->set_value(oneapi::mkl::dft::config_param::FWD_DISTANCE,
n);
511 pp->set_value(oneapi::mkl::dft::config_param::BWD_DISTANCE,
n);
512 std::vector<std::int64_t> strides(ndims+1);
515 for (
int i = ndims-1; i >= 1; --i) {
516 strides[i] = strides[i+1] * len[i];
518#ifndef AMREX_USE_MKL_DFTI_2024
519 pp->set_value(oneapi::mkl::dft::config_param::FWD_STRIDES, strides);
520 pp->set_value(oneapi::mkl::dft::config_param::BWD_STRIDES, strides);
522 pp->set_value(oneapi::mkl::dft::config_param::FWD_STRIDES, strides.data());
523 pp->set_value(oneapi::mkl::dft::config_param::BWD_STRIDES, strides.data());
525 pp->set_value(oneapi::mkl::dft::config_param::WORKSPACE,
526 oneapi::mkl::dft::config_value::WORKSPACE_EXTERNAL);
527 detail::assert_no_external_stream();
528 pp->commit(amrex::Gpu::Device::streamQueue());
533 if constexpr (std::is_same_v<float,T>) {
535 plan = fftwf_plan_many_dft
536 (ndims, len,
howmany, p,
nullptr, 1,
n, p,
nullptr, 1,
n, -1,
539 plan = fftwf_plan_many_dft
540 (ndims, len,
howmany, p,
nullptr, 1,
n, p,
nullptr, 1,
n, +1,
545 plan = fftw_plan_many_dft
546 (ndims, len,
howmany, p,
nullptr, 1,
n, p,
nullptr, 1,
n, -1,
549 plan = fftw_plan_many_dft
550 (ndims, len,
howmany, p,
nullptr, 1,
n, p,
nullptr, 1,
n, +1,
558 template <Direction D>
559 fftw_r2r_kind get_fftw_kind (std::pair<Boundary,Boundary>
const& bc)
579 return fftw_r2r_kind{};
591 template <Direction D>
625 template <Direction D>
626 void init_r2r (
Box const& box, T* p, std::pair<Boundary,Boundary>
const& bc,
627 int howmany_initval = 1)
631 kind = get_r2r_kind<D>(bc);
639#if defined(AMREX_USE_GPU)
659 int nc = (nex/2) + 1;
661#if defined (AMREX_USE_CUDA)
665 cufftType fwd_type = std::is_same_v<float,T> ? CUFFT_R2C : CUFFT_D2Z;
666 std::size_t work_size;
668 (cufftMakePlanMany(
plan, 1, &nex,
nullptr, 1, nc*2,
nullptr, 1, nc, fwd_type,
howmany, &work_size));
670#elif defined(AMREX_USE_HIP)
673 auto prec = std::is_same_v<float,T> ? rocfft_precision_single : rocfft_precision_double;
674 const std::size_t
length = nex;
675 AMREX_ROCFFT_SAFE_CALL
676 (rocfft_plan_create(&
plan, rocfft_placement_inplace,
677 rocfft_transform_type_real_forward, prec, 1,
680#elif defined(AMREX_USE_SYCL)
682 auto*
pp =
new mkl_desc_r(nex);
683#ifndef AMREX_USE_MKL_DFTI_2024
684 pp->set_value(oneapi::mkl::dft::config_param::PLACEMENT,
685 oneapi::mkl::dft::config_value::INPLACE);
687 pp->set_value(oneapi::mkl::dft::config_param::PLACEMENT, DFTI_INPLACE);
689 pp->set_value(oneapi::mkl::dft::config_param::NUMBER_OF_TRANSFORMS,
howmany);
690 pp->set_value(oneapi::mkl::dft::config_param::FWD_DISTANCE, nc*2);
691 pp->set_value(oneapi::mkl::dft::config_param::BWD_DISTANCE, nc);
692 std::vector<std::int64_t> strides = {0,1};
693#ifndef AMREX_USE_MKL_DFTI_2024
694 pp->set_value(oneapi::mkl::dft::config_param::FWD_STRIDES, strides);
695 pp->set_value(oneapi::mkl::dft::config_param::BWD_STRIDES, strides);
697 pp->set_value(oneapi::mkl::dft::config_param::FWD_STRIDES, strides.data());
698 pp->set_value(oneapi::mkl::dft::config_param::BWD_STRIDES, strides.data());
700 pp->set_value(oneapi::mkl::dft::config_param::WORKSPACE,
701 oneapi::mkl::dft::config_value::WORKSPACE_EXTERNAL);
702 detail::assert_no_external_stream();
703 pp->commit(amrex::Gpu::Device::streamQueue());
709 auto fftw_kind = get_fftw_kind<D>(bc);
710 if constexpr (std::is_same_v<float,T>) {
711 plan = fftwf_plan_many_r2r
712 (1, &
n,
howmany, p,
nullptr, 1,
n, p,
nullptr, 1,
n, &fftw_kind,
715 plan = fftw_plan_many_r2r
716 (1, &
n,
howmany, p,
nullptr, 1,
n, p,
nullptr, 1,
n, &fftw_kind,
729 template <Direction D>
731 std::pair<Boundary,Boundary>
const& bc)
737#if defined(AMREX_USE_GPU)
739 init_r2r<D>(box, p, bc, 2);
744 kind = get_r2r_kind<D>(bc);
753 auto fftw_kind = get_fftw_kind<D>(bc);
754 if constexpr (std::is_same_v<float,T>) {
755 plan = fftwf_plan_many_r2r
756 (1, &
n,
howmany, p,
nullptr, 2,
n*2, p,
nullptr, 2,
n*2, &fftw_kind,
758 plan2 = fftwf_plan_many_r2r
759 (1, &
n,
howmany, p+1,
nullptr, 2,
n*2, p+1,
nullptr, 2,
n*2, &fftw_kind,
762 plan = fftw_plan_many_r2r
763 (1, &
n,
howmany, p,
nullptr, 2,
n*2, p,
nullptr, 2,
n*2, &fftw_kind,
765 plan2 = fftw_plan_many_r2r
766 (1, &
n,
howmany, p+1,
nullptr, 2,
n*2, p+1,
nullptr, 2,
n*2, &fftw_kind,
777 template <Direction D>
788#if defined(AMREX_USE_CUDA)
791 std::size_t work_size = 0;
798 if constexpr (std::is_same_v<float,T>) {
804 if constexpr (std::is_same_v<float,T>) {
812#elif defined(AMREX_USE_HIP)
813 detail::hip_execute(
plan, (
void**)&pi, (
void**)&po);
814#elif defined(AMREX_USE_SYCL)
815 detail::sycl_execute<T,D>(std::get<0>(
plan), pi, po);
818 if constexpr (std::is_same_v<float,T>) {
829 template <Direction D>
837#if defined(AMREX_USE_CUDA)
840 std::size_t work_size = 0;
847 if constexpr (std::is_same_v<float,T>) {
854#elif defined(AMREX_USE_HIP)
855 detail::hip_execute(
plan, (
void**)&p, (
void**)&p);
856#elif defined(AMREX_USE_SYCL)
857 detail::sycl_execute<T,D>(std::get<1>(
plan), p, p);
860 if constexpr (std::is_same_v<float,T>) {
883 amrex::Abort(
"FFT: alloc_scratch_space: unsupported kind");
901 auto*
pdst = (T*) pbuf;
904 int ostride = (
n+1)*2;
912 auto batch = ielem /
Long(nex);
913 auto i =
int(ielem - batch*nex);
914 for (
int ir = 0; ir < 2; ++ir) {
915 auto* po =
pdst + (2*batch+ir)*ostride + i;
916 auto const* pi = psrc + 2*batch*istride + ir;
920 *po = sign * pi[(2*norig-1-i)*2];
927 auto batch = ielem /
Long(nex);
928 auto i =
int(ielem - batch*nex);
929 auto* po =
pdst + batch*ostride + i;
930 auto const* pi = psrc + batch*istride;
934 *po = sign * pi[2*norig-1-i];
939 int ostride = (2*
n+1)*2;
947 auto batch = ielem /
Long(nex);
948 auto i =
int(ielem - batch*nex);
949 for (
int ir = 0; ir < 2; ++ir) {
950 auto* po =
pdst + (2*batch+ir)*ostride + i;
951 auto const* pi = psrc + 2*batch*istride + ir;
954 }
else if (i < (2*norig-1)) {
955 *po = pi[(2*norig-2-i)*2];
956 }
else if (i == (2*norig-1)) {
958 }
else if (i < (3*norig)) {
959 *po = -pi[(i-2*norig)*2];
960 }
else if (i < (4*norig-1)) {
961 *po = -pi[(4*norig-2-i)*2];
970 auto batch = ielem /
Long(nex);
971 auto i =
int(ielem - batch*nex);
972 auto* po =
pdst + batch*ostride + i;
973 auto const* pi = psrc + batch*istride;
976 }
else if (i < (2*norig-1)) {
977 *po = pi[2*norig-2-i];
978 }
else if (i == (2*norig-1)) {
980 }
else if (i < (3*norig)) {
981 *po = -pi[i-2*norig];
982 }
else if (i < (4*norig-1)) {
983 *po = -pi[4*norig-2-i];
990 int ostride = (2*
n+1)*2;
998 auto batch = ielem /
Long(nex);
999 auto i =
int(ielem - batch*nex);
1000 for (
int ir = 0; ir < 2; ++ir) {
1001 auto* po =
pdst + (2*batch+ir)*ostride + i;
1002 auto const* pi = psrc + 2*batch*istride + ir;
1005 }
else if (i == norig) {
1007 }
else if (i < (2*norig+1)) {
1008 *po = -pi[(2*norig-i)*2];
1009 }
else if (i < (3*norig)) {
1010 *po = -pi[(i-2*norig)*2];
1011 }
else if (i == 3*norig) {
1014 *po = pi[(4*norig-i)*2];
1021 auto batch = ielem /
Long(nex);
1022 auto i =
int(ielem - batch*nex);
1023 auto* po =
pdst + batch*ostride + i;
1024 auto const* pi = psrc + batch*istride;
1027 }
else if (i == norig) {
1029 }
else if (i < (2*norig+1)) {
1030 *po = -pi[2*norig-i];
1031 }
else if (i < (3*norig)) {
1032 *po = -pi[i-2*norig];
1033 }
else if (i == 3*norig) {
1036 *po = pi[4*norig-i];
1041 int ostride = (2*
n+1)*2;
1049 auto batch = ielem /
Long(nex);
1050 auto i =
int(ielem - batch*nex);
1051 for (
int ir = 0; ir < 2; ++ir) {
1052 auto* po =
pdst + (2*batch+ir)*ostride + i;
1053 auto const* pi = psrc + 2*batch*istride + ir;
1056 }
else if (i < (2*norig)) {
1057 *po = -pi[(2*norig-1-i)*2];
1058 }
else if (i < (3*norig)) {
1059 *po = -pi[(i-2*norig)*2];
1061 *po = pi[(4*norig-1-i)*2];
1068 auto batch = ielem /
Long(nex);
1069 auto i =
int(ielem - batch*nex);
1070 auto* po =
pdst + batch*ostride + i;
1071 auto const* pi = psrc + batch*istride;
1074 }
else if (i < (2*norig)) {
1075 *po = -pi[2*norig-1-i];
1076 }
else if (i < (3*norig)) {
1077 *po = -pi[i-2*norig];
1079 *po = pi[4*norig-1-i];
1084 int ostride = (2*
n+1)*2;
1092 auto batch = ielem /
Long(nex);
1093 auto i =
int(ielem - batch*nex);
1094 for (
int ir = 0; ir < 2; ++ir) {
1095 auto* po =
pdst + (2*batch+ir)*ostride + i;
1096 auto const* pi = psrc + 2*batch*istride + ir;
1099 }
else if (i < (2*norig)) {
1100 *po = pi[(2*norig-1-i)*2];
1101 }
else if (i < (3*norig)) {
1102 *po = -pi[(i-2*norig)*2];
1104 *po = -pi[(4*norig-1-i)*2];
1111 auto batch = ielem /
Long(nex);
1112 auto i =
int(ielem - batch*nex);
1113 auto* po =
pdst + batch*ostride + i;
1114 auto const* pi = psrc + batch*istride;
1117 }
else if (i < (2*norig)) {
1118 *po = pi[2*norig-1-i];
1119 }
else if (i < (3*norig)) {
1120 *po = -pi[i-2*norig];
1122 *po = -pi[4*norig-1-i];
1127 amrex::Abort(
"FFT: pack_r2r_buffer: unsupported kind");
1149 auto batch = ielem /
Long(norig);
1150 auto k =
int(ielem - batch*norig);
1152 for (
int ir = 0; ir < 2; ++ir) {
1153 auto const& yk = psrc[(2*batch+ir)*istride+k+1];
1154 pdst[2*batch*ostride+ir+k*2] = s * yk.real() - c * yk.imag();
1160 auto batch = ielem /
Long(norig);
1161 auto k =
int(ielem - batch*norig);
1163 auto const& yk = psrc[batch*istride+k+1];
1164 pdst[batch*ostride+k] = s * yk.real() - c * yk.imag();
1168 int istride = 2*
n+1;
1172 auto batch = ielem /
Long(norig);
1173 auto k =
int(ielem - batch*norig);
1175 for (
int ir = 0; ir < 2; ++ir) {
1176 auto const& yk = psrc[(2*batch+ir)*istride+2*k+1];
1177 pdst[2*batch*ostride+ir+k*2] = T(0.5)*(s * yk.real() - c * yk.imag());
1183 auto batch = ielem /
Long(norig);
1184 auto k =
int(ielem - batch*norig);
1186 auto const& yk = psrc[batch*istride+2*k+1];
1187 pdst[batch*ostride+k] = T(0.5)*(s * yk.real() - c * yk.imag());
1195 auto batch = ielem /
Long(norig);
1196 auto k =
int(ielem - batch*norig);
1198 for (
int ir = 0; ir < 2; ++ir) {
1199 auto const& yk = psrc[(2*batch+ir)*istride+k];
1200 pdst[2*batch*ostride+ir+k*2] = c * yk.real() + s * yk.imag();
1206 auto batch = ielem /
Long(norig);
1207 auto k =
int(ielem - batch*norig);
1209 auto const& yk = psrc[batch*istride+k];
1210 pdst[batch*ostride+k] = c * yk.real() + s * yk.imag();
1214 int istride = 2*
n+1;
1218 auto batch = ielem /
Long(norig);
1219 auto k =
int(ielem - batch*norig);
1220 for (
int ir = 0; ir < 2; ++ir) {
1221 auto const& yk = psrc[(2*batch+ir)*istride+2*k+1];
1222 pdst[2*batch*ostride+ir+k*2] = T(0.5) * yk.real();
1228 auto batch = ielem /
Long(norig);
1229 auto k =
int(ielem - batch*norig);
1230 auto const& yk = psrc[batch*istride+2*k+1];
1231 pdst[batch*ostride+k] = T(0.5) * yk.real();
1235 int istride = 2*
n+1;
1239 auto batch = ielem /
Long(norig);
1240 auto k =
int(ielem - batch*norig);
1242 for (
int ir = 0; ir < 2; ++ir) {
1243 auto const& yk = psrc[(2*batch+ir)*istride+2*k+1];
1244 pdst[2*batch*ostride+ir+k*2] = T(0.5) * (c * yk.real() + s * yk.imag());
1250 auto batch = ielem /
Long(norig);
1251 auto k =
int(ielem - batch*norig);
1253 auto const& yk = psrc[batch*istride+2*k+1];
1254 pdst[batch*ostride+k] = T(0.5) * (c * yk.real() + s * yk.imag());
1258 int istride = 2*
n+1;
1262 auto batch = ielem /
Long(norig);
1263 auto k =
int(ielem - batch*norig);
1265 for (
int ir = 0; ir < 2; ++ir) {
1266 auto const& yk = psrc[(2*batch+ir)*istride+2*k+1];
1267 pdst[2*batch*ostride+ir+k*2] = T(0.5) * (s * yk.real() - c * yk.imag());
1273 auto batch = ielem /
Long(norig);
1274 auto k =
int(ielem - batch*norig);
1276 auto const& yk = psrc[batch*istride+2*k+1];
1277 pdst[batch*ostride+k] = T(0.5) * (s * yk.real() - c * yk.imag());
1281 amrex::Abort(
"FFT: unpack_r2r_buffer: unsupported kind");
1289 template <Direction D>
1295#if defined(AMREX_USE_GPU)
1301#if defined(AMREX_USE_CUDA)
1305 std::size_t work_size = 0;
1311 if constexpr (std::is_same_v<float,T>) {
1317#elif defined(AMREX_USE_HIP)
1318 detail::hip_execute(
plan, (
void**)&pscratch, (
void**)&pscratch);
1319#elif defined(AMREX_USE_SYCL)
1320 detail::sycl_execute<T,Direction::forward>(std::get<0>(
plan), (T*)pscratch, (
VendorComplex*)pscratch);
1327#if defined(AMREX_USE_CUDA)
1333 if constexpr (std::is_same_v<float,T>) {
1334 fftwf_execute(
plan);
1349#if defined(AMREX_USE_CUDA)
1351#elif defined(AMREX_USE_HIP)
1352 AMREX_ROCFFT_SAFE_CALL(rocfft_plan_destroy(
plan));
1353#elif defined(AMREX_USE_SYCL)
1354 std::visit([](
auto&& p) {
delete p; },
plan);
1356 if constexpr (std::is_same_v<float,T>) {
1357 fftwf_destroy_plan(
plan);
1359 fftw_destroy_plan(
plan);
1372 PlanD* get_vendor_plan_d (Key
const& key);
1373 PlanF* get_vendor_plan_f (Key
const& key);
1375 void add_vendor_plan_d (Key
const& key, PlanD plan);
1376 void add_vendor_plan_f (Key
const& key, PlanF plan);
1380template <
typename T>
1381template <Direction D,
int M>
1392 for (
auto s : fft_size) { n *= s; }
1395#if defined(AMREX_USE_GPU)
1396 Key key = {fft_size.template expand<3>(), ncomp, D, kind};
1399 if constexpr (std::is_same_v<float,T>) {
1400 cached_plan = detail::get_vendor_plan_f(key);
1402 cached_plan = detail::get_vendor_plan_d(key);
1405 plan = *cached_plan;
1414 for (
int i = 0; i < M; ++i) {
1415 len[i] = fft_size[M-1-i];
1418 int nc = fft_size[0]/2+1;
1419 for (
int i = 1; i < M; ++i) {
1423#if defined(AMREX_USE_CUDA)
1430 type = std::is_same_v<float,T> ? CUFFT_R2C : CUFFT_D2Z;
1434 type = std::is_same_v<float,T> ? CUFFT_C2R : CUFFT_Z2D;
1438 std::size_t work_size;
1440 (cufftMakePlanMany(plan, M, len,
nullptr, 1, n_in,
nullptr, 1, n_out, type, howmany, &work_size));
1442#elif defined(AMREX_USE_HIP)
1444 auto prec = std::is_same_v<float,T> ? rocfft_precision_single : rocfft_precision_double;
1446 for (
int idim = 0; idim < M; ++idim) {
length[idim] = fft_size[idim]; }
1448 AMREX_ROCFFT_SAFE_CALL
1449 (rocfft_plan_create(&plan, rocfft_placement_notinplace,
1450 rocfft_transform_type_real_forward, prec, M,
1451 length, howmany,
nullptr));
1453 AMREX_ROCFFT_SAFE_CALL
1454 (rocfft_plan_create(&plan, rocfft_placement_notinplace,
1455 rocfft_transform_type_real_inverse, prec, M,
1456 length, howmany,
nullptr));
1459#elif defined(AMREX_USE_SYCL)
1463 pp =
new mkl_desc_r(fft_size[0]);
1465 std::vector<std::int64_t> len64(M);
1466 for (
int idim = 0; idim < M; ++idim) {
1467 len64[idim] = len[idim];
1469 pp =
new mkl_desc_r(len64);
1471#ifndef AMREX_USE_MKL_DFTI_2024
1472 pp->set_value(oneapi::mkl::dft::config_param::PLACEMENT,
1473 oneapi::mkl::dft::config_value::NOT_INPLACE);
1475 pp->set_value(oneapi::mkl::dft::config_param::PLACEMENT, DFTI_NOT_INPLACE);
1477 pp->set_value(oneapi::mkl::dft::config_param::NUMBER_OF_TRANSFORMS, howmany);
1478 pp->set_value(oneapi::mkl::dft::config_param::FWD_DISTANCE, n);
1479 pp->set_value(oneapi::mkl::dft::config_param::BWD_DISTANCE, nc);
1480 std::vector<std::int64_t> strides(M+1);
1483 for (
int i = M-1; i >= 1; --i) {
1484 strides[i] = strides[i+1] * fft_size[M-1-i];
1487#ifndef AMREX_USE_MKL_DFTI_2024
1488 pp->set_value(oneapi::mkl::dft::config_param::FWD_STRIDES, strides);
1491 pp->set_value(oneapi::mkl::dft::config_param::FWD_STRIDES, strides.data());
1494 pp->set_value(oneapi::mkl::dft::config_param::WORKSPACE,
1495 oneapi::mkl::dft::config_value::WORKSPACE_EXTERNAL);
1496 detail::assert_no_external_stream();
1497 pp->commit(amrex::Gpu::Device::streamQueue());
1502 if (pf ==
nullptr || pb ==
nullptr) {
1507 if constexpr (std::is_same_v<float,T>) {
1509 plan = fftwf_plan_many_dft_r2c
1510 (M, len, howmany, (
float*)pf,
nullptr, 1, n, (fftwf_complex*)pb,
nullptr, 1, nc,
1513 plan = fftwf_plan_many_dft_c2r
1514 (M, len, howmany, (fftwf_complex*)pb,
nullptr, 1, nc, (
float*)pf,
nullptr, 1, n,
1519 plan = fftw_plan_many_dft_r2c
1520 (M, len, howmany, (
double*)pf,
nullptr, 1, n, (fftw_complex*)pb,
nullptr, 1, nc,
1523 plan = fftw_plan_many_dft_c2r
1524 (M, len, howmany, (fftw_complex*)pb,
nullptr, 1, nc, (
double*)pf,
nullptr, 1, n,
1530#if defined(AMREX_USE_GPU)
1532 if constexpr (std::is_same_v<float,T>) {
1533 detail::add_vendor_plan_f(key, plan);
1535 detail::add_vendor_plan_d(key, plan);
1546 template <
typename FA>
1547 typename FA::FABType::value_type * get_fab (FA& fa)
1549 auto myproc = ParallelContext::MyProcSub();
1550 if (myproc < fa.size()) {
1551 return fa.fabPtr(myproc);
1557 template <
typename FA1,
typename FA2>
1558 std::unique_ptr<char,DataDeleter> make_mfs_share (FA1& fa1, FA2& fa2)
1560 bool not_same_fa =
true;
1561 if constexpr (std::is_same_v<FA1,FA2>) {
1562 not_same_fa = (&fa1 != &fa2);
1564 using FAB1 =
typename FA1::FABType::value_type;
1565 using FAB2 =
typename FA2::FABType::value_type;
1566 using T1 =
typename FAB1::value_type;
1567 using T2 =
typename FAB2::value_type;
1568 auto myproc = ParallelContext::MyProcSub();
1569 bool alloc_1 = (myproc < fa1.size());
1570 bool alloc_2 = (myproc < fa2.size()) && not_same_fa;
1572 if (alloc_1 && alloc_2) {
1573 Box const& box1 = fa1.fabbox(myproc);
1574 Box const& box2 = fa2.fabbox(myproc);
1575 int ncomp1 = fa1.nComp();
1576 int ncomp2 = fa2.nComp();
1578 sizeof(T2)*box2.numPts()*ncomp2));
1579 fa1.setFab(myproc, FAB1(box1, ncomp1, (T1*)p));
1580 fa2.setFab(myproc, FAB2(box2, ncomp2, (T2*)p));
1581 }
else if (alloc_1) {
1582 Box const& box1 = fa1.fabbox(myproc);
1583 int ncomp1 = fa1.nComp();
1585 fa1.setFab(myproc, FAB1(box1, ncomp1, (T1*)p));
1586 }
else if (alloc_2) {
1587 Box const& box2 = fa2.fabbox(myproc);
1588 int ncomp2 = fa2.nComp();
1590 fa2.setFab(myproc, FAB2(box2, ncomp2, (T2*)p));
1594 return std::unique_ptr<char,DataDeleter>((
char*)p, DataDeleter{
The_Arena()});
1603 [[nodiscard]]
constexpr Dim3 operator() (Dim3 i)
const noexcept
1605 return {i.y, i.x, i.z};
1608 static constexpr Dim3 Inverse (Dim3 i)
1610 return {i.y, i.x, i.z};
1613 [[nodiscard]]
constexpr IndexType operator() (IndexType it)
const noexcept
1618 static constexpr IndexType Inverse (IndexType it)
1626 [[nodiscard]]
constexpr Dim3 operator() (Dim3 i)
const noexcept
1628 return {i.z, i.y, i.x};
1631 static constexpr Dim3 Inverse (Dim3 i)
1633 return {i.z, i.y, i.x};
1636 [[nodiscard]]
constexpr IndexType operator() (IndexType it)
const noexcept
1641 static constexpr IndexType Inverse (IndexType it)
1650 [[nodiscard]]
constexpr Dim3 operator() (Dim3 i)
const noexcept
1652 return {i.y, i.z, i.x};
1656 static constexpr Dim3 Inverse (Dim3 i)
1658 return {i.z, i.x, i.y};
1661 [[nodiscard]]
constexpr IndexType operator() (IndexType it)
const noexcept
1666 static constexpr IndexType Inverse (IndexType it)
1675 [[nodiscard]]
constexpr Dim3 operator() (Dim3 i)
const noexcept
1677 return {i.z, i.x, i.y};
1681 static constexpr Dim3 Inverse (Dim3 i)
1683 return {i.y, i.z, i.x};
1686 [[nodiscard]]
constexpr IndexType operator() (IndexType it)
const noexcept
1691 static constexpr IndexType Inverse (IndexType it)
1704 explicit SubHelper (Box
const& domain);
1706 [[nodiscard]]
Box make_box (Box
const& box)
const;
1708 [[nodiscard]] Periodicity make_periodicity (Periodicity
const& period)
const;
1710 [[nodiscard]]
bool ghost_safe (IntVect
const& ng)
const;
1713 [[nodiscard]]
IntVect make_iv (IntVect
const& iv)
const;
1716 [[nodiscard]]
IntVect make_safe_ghost (IntVect
const& ng)
const;
1718 [[nodiscard]] BoxArray inverse_boxarray (BoxArray
const& ba)
const;
1720 [[nodiscard]]
IntVect inverse_order (IntVect
const& order)
const;
1722 template <
typename T>
1723 [[nodiscard]] T make_array (T
const& a)
const
1725#if (AMREX_SPACEDIM == 1)
1728#elif (AMREX_SPACEDIM == 2)
1729 if (m_case == case_1n) {
1730 return T{a[1],a[0]};
1735 if (m_case == case_11n) {
1736 return T{a[2],a[0],a[1]};
1737 }
else if (m_case == case_1n1) {
1738 return T{a[1],a[0],a[2]};
1739 }
else if (m_case == case_1nn) {
1740 return T{a[1],a[2],a[0]};
1741 }
else if (m_case == case_n1n) {
1742 return T{a[0],a[2],a[1]};
1749 [[nodiscard]] GpuArray<int,3> xyz_order ()
const;
1751 template <
typename FA>
1752 FA make_alias_mf (FA
const& mf)
1754 BoxList bl = mf.boxArray().boxList();
1755 for (
auto& b : bl) {
1758 auto const& ng = make_iv(mf.nGrowVect());
1759 FA submf(BoxArray(std::move(bl)), mf.DistributionMap(), mf.nComp(), ng, MFInfo{}.SetAlloc(
false));
1760 using FAB =
typename FA::fab_type;
1761 for (MFIter mfi(submf, MFItInfo().DisableDeviceSync()); mfi.isValid(); ++mfi) {
1762 submf.setFab(mfi, FAB(mfi.fabbox(), mf.nComp(), mf[mfi].dataPtr()));
1767#if (AMREX_SPACEDIM == 2)
1768 enum Case { case_1n, case_other };
1769 int m_case = case_other;
1770#elif (AMREX_SPACEDIM == 3)
1771 enum Case { case_11n, case_1n1, case_1nn, case_n1n, case_other };
1772 int m_case = case_other;
#define AMREX_ALWAYS_ASSERT_WITH_MESSAGE(EX, MSG)
Definition AMReX_BLassert.H:49
#define AMREX_ENUM(CLASS,...)
Definition AMReX_Enum.H:209
#define AMREX_CUFFT_SAFE_CALL(call)
Definition AMReX_GpuError.H:92
#define AMREX_GPU_DEVICE
Definition AMReX_GpuQualifiers.H:18
amrex::ParmParse pp
Input file parser instance for the given namespace.
Definition AMReX_HypreIJIface.cpp:15
Real * pdst
Definition AMReX_HypreMLABecLap.cpp:1140
#define AMREX_D_TERM(a, b, c)
Definition AMReX_SPACE.H:172
virtual void free(void *pt)=0
A pure virtual function for deleting the arena pointed to by pt.
virtual void * alloc(std::size_t sz)=0
__host__ __device__ IntVectND< dim > length() const noexcept
Return the length of the BoxND.
Definition AMReX_Box.H:154
Calculates the distribution of FABs to MPI processes.
Definition AMReX_DistributionMapping.H:43
static bool usingExternalStream() noexcept
Definition AMReX_GpuDevice.cpp:836
An Integer Vector in dim-Dimensional Space.
Definition AMReX_IntVect.H:57
amrex_long Long
Definition AMReX_INT.H:30
void ParallelForOMP(T n, L const &f) noexcept
Performance-portable kernel launch function with optional OpenMP threading.
Definition AMReX_GpuLaunch.H:319
__host__ __device__ Dim3 length(Array4< T > const &a) noexcept
Return the spatial extents of an Array4 in Dim3 form.
Definition AMReX_Array4.H:1345
Arena * The_Arena()
Definition AMReX_Arena.cpp:805
Definition AMReX_FFT_Helper.H:52
Direction
Definition AMReX_FFT_Helper.H:54
Boundary
Definition AMReX_FFT_Helper.H:58
DomainStrategy
Definition AMReX_FFT_Helper.H:56
Kind
Definition AMReX_FFT_Helper.H:60
void streamSynchronize() noexcept
Definition AMReX_GpuDevice.H:310
gpuStream_t gpuStream() noexcept
Definition AMReX_GpuDevice.H:291
__host__ __device__ std::pair< double, double > sincospi(double x)
Return sin(pi*x) and cos(pi*x) given x.
Definition AMReX_Math.H:204
__host__ __device__ void ignore_unused(const Ts &...)
This shuts up the compiler about unused variables.
Definition AMReX.H:139
BoxND< 3 > Box
Box is an alias for amrex::BoxND instantiated with AMREX_SPACEDIM.
Definition AMReX_BaseFwd.H:30
IndexTypeND< 3 > IndexType
IndexType is an alias for amrex::IndexTypeND instantiated with AMREX_SPACEDIM.
Definition AMReX_BaseFwd.H:36
IntVectND< 3 > IntVect
IntVect is an alias for amrex::IntVectND instantiated with AMREX_SPACEDIM.
Definition AMReX_BaseFwd.H:33
void Abort(const std::string &msg)
Print out message to cerr and exit via abort().
Definition AMReX.cpp:240
const int[]
Definition AMReX_BLProfiler.cpp:1664
Definition AMReX_FFT_Helper.H:64
bool twod_mode
Definition AMReX_FFT_Helper.H:75
Info & setNumProcs(int n)
Cap the number of MPI ranks used by FFT.
Definition AMReX_FFT_Helper.H:137
bool oned_mode
Definition AMReX_FFT_Helper.H:84
int batch_size
Batched FFT size. Only support in R2C, not R2X.
Definition AMReX_FFT_Helper.H:87
Info & setDomainStrategy(DomainStrategy s)
Select how the domain is decomposed across MPI ranks.
Definition AMReX_FFT_Helper.H:98
DomainStrategy domain_strategy
Domain composition strategy.
Definition AMReX_FFT_Helper.H:66
int nprocs
Max number of processes to use.
Definition AMReX_FFT_Helper.H:90
int pencil_threshold
Definition AMReX_FFT_Helper.H:70
Info & setOneDMode(bool x)
Flag the degenerate 2-D mode (nx==1 or ny==1) that still batches along z.
Definition AMReX_FFT_Helper.H:123
Info & setBatchSize(int bsize)
Specify the batch size for FFT.
Definition AMReX_FFT_Helper.H:130
Info & setPencilThreshold(int t)
Override the slab→pencil break-even threshold for the automatic strategy.
Definition AMReX_FFT_Helper.H:105
Info & setTwoDMode(bool x)
Restrict transforms to the first two dimensions (3-D problems only).
Definition AMReX_FFT_Helper.H:112
Definition AMReX_FFT_Helper.H:194
void * pf
Definition AMReX_FFT_Helper.H:229
void unpack_r2r_buffer(T *pdst, void const *pbuf) const
Collapse the spectral R2R result back into the original real layout.
Definition AMReX_FFT_Helper.H:1137
std::conditional_t< std::is_same_v< float, T >, cuComplex, cuDoubleComplex > VendorComplex
Definition AMReX_FFT_Helper.H:198
VendorPlan plan2
Definition AMReX_FFT_Helper.H:228
void init_r2c(IntVectND< M > const &fft_size, void *pbf, void *pbb, bool cache, int ncomp=1)
Initialize an M-dimensional batched real-to-complex plan.
Definition AMReX_FFT_Helper.H:1382
int n
Definition AMReX_FFT_Helper.H:221
void destroy()
Release any vendor FFT plan objects owned by this Plan.
Definition AMReX_FFT_Helper.H:248
bool defined2
Definition AMReX_FFT_Helper.H:226
void init_r2r(Box const &box, VendorComplex *pc, std::pair< Boundary, Boundary > const &bc)
Initialize a real-to-real plan that reads/writes complex storage.
Definition AMReX_FFT_Helper.H:730
void pack_r2r_buffer(void *pbuf, T const *psrc) const
Expand the real R2R input into the symmetry-extended buffer expected by CUFFT/rocFFT.
Definition AMReX_FFT_Helper.H:899
static void free_scratch_space(void *p)
Release GPU scratch allocated via alloc_scratch_space().
Definition AMReX_FFT_Helper.H:891
static void destroy_vendor_plan(VendorPlan plan)
Helper that destroys a vendor plan of the appropriate backend type.
Definition AMReX_FFT_Helper.H:1347
Kind get_r2r_kind(std::pair< Boundary, Boundary > const &bc)
Map boundary conditions to the Plan Kind for real-to-real transforms.
Definition AMReX_FFT_Helper.H:592
cufftHandle VendorPlan
Definition AMReX_FFT_Helper.H:196
Kind kind
Definition AMReX_FFT_Helper.H:223
void init_c2c(Box const &box, VendorComplex *p, int ncomp=1, int ndims=1)
Initialize a complex-to-complex plan across 1/2/3 dimensions.
Definition AMReX_FFT_Helper.H:424
int howmany
Definition AMReX_FFT_Helper.H:222
void * pb
Definition AMReX_FFT_Helper.H:230
void init_r2c(Box const &box, T *pr, VendorComplex *pc, bool is_2d_transform=false, int ncomp=1)
Initialize a 1-D or 2-D real-to-complex plan over the supplied Box.
Definition AMReX_FFT_Helper.H:272
void compute_r2r()
Execute the real-to-real plan, including GPU packing/unpacking.
Definition AMReX_FFT_Helper.H:1290
void compute_c2c()
Execute the complex-to-complex plan in place.
Definition AMReX_FFT_Helper.H:830
bool r2r_data_is_complex
Definition AMReX_FFT_Helper.H:224
void * alloc_scratch_space() const
Allocate GPU scratch space large enough to hold packed R2R data.
Definition AMReX_FFT_Helper.H:874
VendorPlan plan
Definition AMReX_FFT_Helper.H:227
void compute_r2c()
Execute the previously initialized real-to-complex plan.
Definition AMReX_FFT_Helper.H:778
bool defined
Definition AMReX_FFT_Helper.H:225
void set_ptrs(void *p0, void *p1)
Register device pointers used by the forward/backward executions.
Definition AMReX_FFT_Helper.H:239
void init_r2r(Box const &box, T *p, std::pair< Boundary, Boundary > const &bc, int howmany_initval=1)
Initialize a real-to-real (cosine/sine) plan that operates on real buffers.
Definition AMReX_FFT_Helper.H:626
A host / device complex number type, because std::complex doesn't work in device code with Cuda yet.
Definition AMReX_GpuComplex.H:30