Add switchable C RK4 kernel and build toggle
This commit is contained in:
@@ -72,13 +72,21 @@ TwoPunctureABE.o: TwoPunctureABE.C
|
||||
# Input files
|
||||
|
||||
## Kernel implementation switch (set USE_CXX_KERNELS=0 to fall back to Fortran)
|
||||
ifeq ($(USE_CXX_KERNELS),0)
|
||||
# Fortran mode: no C rewrite files; bssn_rhs.o is included via F90FILES below
|
||||
CFILES =
|
||||
else
|
||||
ifeq ($(USE_CXX_KERNELS),0)
|
||||
# Fortran mode: no C rewrite files; bssn_rhs.o is included via F90FILES below
|
||||
CFILES =
|
||||
else
|
||||
# C++ mode (default): C rewrite of bssn_rhs and helper kernels
|
||||
CFILES = bssn_rhs_c.o fderivs_c.o fdderivs_c.o kodiss_c.o lopsided_c.o lopsided_kodis_c.o
|
||||
endif
|
||||
|
||||
## RK4 kernel switch (independent from USE_CXX_KERNELS)
|
||||
ifeq ($(USE_CXX_RK4),1)
|
||||
CFILES += rungekutta4_rout_c.o
|
||||
RK4_F90_OBJ =
|
||||
else
|
||||
RK4_F90_OBJ = rungekutta4_rout.o
|
||||
endif
|
||||
|
||||
C++FILES = ABE.o Ansorg.o Block.o misc.o monitor.o Parallel.o MPatch.o var.o\
|
||||
cgh.o bssn_class.o surface_integral.o ShellPatch.o\
|
||||
@@ -95,12 +103,12 @@ C++FILES_GPU = ABE.o Ansorg.o Block.o misc.o monitor.o Parallel.o MPatch.o var.o
|
||||
NullShellPatch2_Evo.o \
|
||||
bssn_gpu_class.o bssn_step_gpu.o bssn_macro.o writefile_f.o
|
||||
|
||||
F90FILES_BASE = enforce_algebra.o fmisc.o initial_puncture.o prolongrestrict.o\
|
||||
prolongrestrict_cell.o prolongrestrict_vertex.o\
|
||||
rungekutta4_rout.o diff_new.o kodiss.o kodiss_sh.o\
|
||||
lopsidediff.o sommerfeld_rout.o getnp4.o diff_new_sh.o\
|
||||
shellfunctions.o bssn_rhs_ss.o Set_Rho_ADM.o\
|
||||
getnp4EScalar.o bssnEScalar_rhs.o bssn_constraint.o ricci_gamma.o\
|
||||
F90FILES_BASE = enforce_algebra.o fmisc.o initial_puncture.o prolongrestrict.o\
|
||||
prolongrestrict_cell.o prolongrestrict_vertex.o\
|
||||
$(RK4_F90_OBJ) diff_new.o kodiss.o kodiss_sh.o\
|
||||
lopsidediff.o sommerfeld_rout.o getnp4.o diff_new_sh.o\
|
||||
shellfunctions.o bssn_rhs_ss.o Set_Rho_ADM.o\
|
||||
getnp4EScalar.o bssnEScalar_rhs.o bssn_constraint.o ricci_gamma.o\
|
||||
fadmquantites_bssn.o Z4c_rhs.o Z4c_rhs_ss.o point_diff_new_sh.o\
|
||||
cpbc.o getnp4old.o NullEvol.o initial_null.o initial_maxwell.o\
|
||||
getnpem2.o empart.o NullNews.o fourdcurvature.o\
|
||||
|
||||
@@ -47,6 +47,12 @@ endif
|
||||
## 1 (default) : use C++ rewrite of bssn_rhs and helper kernels (faster)
|
||||
## 0 : fall back to original Fortran kernels
|
||||
USE_CXX_KERNELS ?= 1
|
||||
|
||||
## RK4 kernel implementation switch
|
||||
## 1 (default) : use C/C++ rewrite of rungekutta4_rout (for optimization experiments)
|
||||
## 0 : use original Fortran rungekutta4_rout.o
|
||||
USE_CXX_RK4 ?= 1
|
||||
|
||||
f90 = ifx
|
||||
f77 = ifx
|
||||
CXX = icpx
|
||||
|
||||
155
AMSS_NCKU_source/rungekutta4_rout_c.C
Normal file
155
AMSS_NCKU_source/rungekutta4_rout_c.C
Normal file
@@ -0,0 +1,155 @@
|
||||
#include "rungekutta4_rout.h"
|
||||
#include <cstdio>
|
||||
#include <cstdlib>
|
||||
#include <cstddef>
|
||||
#include <immintrin.h>
|
||||
|
||||
namespace {
|
||||
|
||||
inline void rk4_stage0(std::size_t n,
|
||||
const double *__restrict f0,
|
||||
const double *__restrict frhs,
|
||||
double *__restrict f1,
|
||||
double c) {
|
||||
std::size_t i = 0;
|
||||
#if defined(__AVX512F__)
|
||||
const __m512d vc = _mm512_set1_pd(c);
|
||||
for (; i + 7 < n; i += 8) {
|
||||
const __m512d v0 = _mm512_loadu_pd(f0 + i);
|
||||
const __m512d vr = _mm512_loadu_pd(frhs + i);
|
||||
_mm512_storeu_pd(f1 + i, _mm512_fmadd_pd(vc, vr, v0));
|
||||
}
|
||||
#elif defined(__AVX2__)
|
||||
const __m256d vc = _mm256_set1_pd(c);
|
||||
for (; i + 3 < n; i += 4) {
|
||||
const __m256d v0 = _mm256_loadu_pd(f0 + i);
|
||||
const __m256d vr = _mm256_loadu_pd(frhs + i);
|
||||
_mm256_storeu_pd(f1 + i, _mm256_fmadd_pd(vc, vr, v0));
|
||||
}
|
||||
#endif
|
||||
#pragma ivdep
|
||||
for (; i < n; ++i) {
|
||||
f1[i] = f0[i] + c * frhs[i];
|
||||
}
|
||||
}
|
||||
|
||||
inline void rk4_rhs_accum(std::size_t n,
|
||||
const double *__restrict f1,
|
||||
double *__restrict frhs) {
|
||||
std::size_t i = 0;
|
||||
#if defined(__AVX512F__)
|
||||
const __m512d v2 = _mm512_set1_pd(2.0);
|
||||
for (; i + 7 < n; i += 8) {
|
||||
const __m512d v1 = _mm512_loadu_pd(f1 + i);
|
||||
const __m512d vrhs = _mm512_loadu_pd(frhs + i);
|
||||
_mm512_storeu_pd(frhs + i, _mm512_fmadd_pd(v2, v1, vrhs));
|
||||
}
|
||||
#elif defined(__AVX2__)
|
||||
const __m256d v2 = _mm256_set1_pd(2.0);
|
||||
for (; i + 3 < n; i += 4) {
|
||||
const __m256d v1 = _mm256_loadu_pd(f1 + i);
|
||||
const __m256d vrhs = _mm256_loadu_pd(frhs + i);
|
||||
_mm256_storeu_pd(frhs + i, _mm256_fmadd_pd(v2, v1, vrhs));
|
||||
}
|
||||
#endif
|
||||
#pragma ivdep
|
||||
for (; i < n; ++i) {
|
||||
frhs[i] = frhs[i] + 2.0 * f1[i];
|
||||
}
|
||||
}
|
||||
|
||||
inline void rk4_f1_from_f0_f1(std::size_t n,
|
||||
const double *__restrict f0,
|
||||
double *__restrict f1,
|
||||
double c) {
|
||||
std::size_t i = 0;
|
||||
#if defined(__AVX512F__)
|
||||
const __m512d vc = _mm512_set1_pd(c);
|
||||
for (; i + 7 < n; i += 8) {
|
||||
const __m512d v0 = _mm512_loadu_pd(f0 + i);
|
||||
const __m512d v1 = _mm512_loadu_pd(f1 + i);
|
||||
_mm512_storeu_pd(f1 + i, _mm512_fmadd_pd(vc, v1, v0));
|
||||
}
|
||||
#elif defined(__AVX2__)
|
||||
const __m256d vc = _mm256_set1_pd(c);
|
||||
for (; i + 3 < n; i += 4) {
|
||||
const __m256d v0 = _mm256_loadu_pd(f0 + i);
|
||||
const __m256d v1 = _mm256_loadu_pd(f1 + i);
|
||||
_mm256_storeu_pd(f1 + i, _mm256_fmadd_pd(vc, v1, v0));
|
||||
}
|
||||
#endif
|
||||
#pragma ivdep
|
||||
for (; i < n; ++i) {
|
||||
f1[i] = f0[i] + c * f1[i];
|
||||
}
|
||||
}
|
||||
|
||||
inline void rk4_stage3(std::size_t n,
|
||||
const double *__restrict f0,
|
||||
double *__restrict f1,
|
||||
const double *__restrict frhs,
|
||||
double c) {
|
||||
std::size_t i = 0;
|
||||
#if defined(__AVX512F__)
|
||||
const __m512d vc = _mm512_set1_pd(c);
|
||||
for (; i + 7 < n; i += 8) {
|
||||
const __m512d v0 = _mm512_loadu_pd(f0 + i);
|
||||
const __m512d v1 = _mm512_loadu_pd(f1 + i);
|
||||
const __m512d vr = _mm512_loadu_pd(frhs + i);
|
||||
_mm512_storeu_pd(f1 + i, _mm512_fmadd_pd(vc, _mm512_add_pd(v1, vr), v0));
|
||||
}
|
||||
#elif defined(__AVX2__)
|
||||
const __m256d vc = _mm256_set1_pd(c);
|
||||
for (; i + 3 < n; i += 4) {
|
||||
const __m256d v0 = _mm256_loadu_pd(f0 + i);
|
||||
const __m256d v1 = _mm256_loadu_pd(f1 + i);
|
||||
const __m256d vr = _mm256_loadu_pd(frhs + i);
|
||||
_mm256_storeu_pd(f1 + i, _mm256_fmadd_pd(vc, _mm256_add_pd(v1, vr), v0));
|
||||
}
|
||||
#endif
|
||||
#pragma ivdep
|
||||
for (; i < n; ++i) {
|
||||
f1[i] = f0[i] + c * (f1[i] + frhs[i]);
|
||||
}
|
||||
}
|
||||
|
||||
} // namespace
|
||||
|
||||
extern "C" {
|
||||
|
||||
int f_rungekutta4_rout(int *ex, double &dT,
|
||||
double *f0, double *f1, double *f_rhs,
|
||||
int &RK4) {
|
||||
const std::size_t n = static_cast<std::size_t>(ex[0]) *
|
||||
static_cast<std::size_t>(ex[1]) *
|
||||
static_cast<std::size_t>(ex[2]);
|
||||
const double *const __restrict f0r = f0;
|
||||
double *const __restrict f1r = f1;
|
||||
double *const __restrict frhs = f_rhs;
|
||||
|
||||
if (__builtin_expect(static_cast<unsigned>(RK4) > 3u, 0)) {
|
||||
std::fprintf(stderr, "rungekutta4_rout_c: invalid RK4 stage %d\n", RK4);
|
||||
std::abort();
|
||||
}
|
||||
|
||||
switch (RK4) {
|
||||
case 0:
|
||||
rk4_stage0(n, f0r, frhs, f1r, 0.5 * dT);
|
||||
break;
|
||||
case 1:
|
||||
rk4_rhs_accum(n, f1r, frhs);
|
||||
rk4_f1_from_f0_f1(n, f0r, f1r, 0.5 * dT);
|
||||
break;
|
||||
case 2:
|
||||
rk4_rhs_accum(n, f1r, frhs);
|
||||
rk4_f1_from_f0_f1(n, f0r, f1r, dT);
|
||||
break;
|
||||
default:
|
||||
rk4_stage3(n, f0r, f1r, frhs, (1.0 / 6.0) * dT);
|
||||
break;
|
||||
}
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
} // extern "C"
|
||||
Reference in New Issue
Block a user