Compare commits
7 Commits
chb-replac
...
chb-copilo
| Author | SHA1 | Date | |
|---|---|---|---|
|
8b68b5d782
|
|||
|
dd2443c926
|
|||
|
2d7ba5c60c
|
|||
|
4777cad4ed
|
|||
|
afd4006da2
|
|||
|
|
a918dc103e
|
||
|
|
38c2c30186
|
@@ -341,9 +341,8 @@ void Patch::Interp_Points(MyList<var> *VarList,
|
|||||||
double *Shellf, int Symmetry)
|
double *Shellf, int Symmetry)
|
||||||
{
|
{
|
||||||
// NOTE: we do not Synchnize variables here, make sure of that before calling this routine
|
// NOTE: we do not Synchnize variables here, make sure of that before calling this routine
|
||||||
int myrank, nprocs;
|
int myrank;
|
||||||
MPI_Comm_rank(MPI_COMM_WORLD, &myrank);
|
MPI_Comm_rank(MPI_COMM_WORLD, &myrank);
|
||||||
MPI_Comm_size(MPI_COMM_WORLD, &nprocs);
|
|
||||||
|
|
||||||
int ordn = 2 * ghost_width;
|
int ordn = 2 * ghost_width;
|
||||||
MyList<var> *varl;
|
MyList<var> *varl;
|
||||||
@@ -355,18 +354,24 @@ void Patch::Interp_Points(MyList<var> *VarList,
|
|||||||
varl = varl->next;
|
varl = varl->next;
|
||||||
}
|
}
|
||||||
|
|
||||||
memset(Shellf, 0, sizeof(double) * NN * num_var);
|
double *shellf;
|
||||||
|
shellf = new double[NN * num_var];
|
||||||
|
memset(shellf, 0, sizeof(double) * NN * num_var);
|
||||||
|
|
||||||
// owner_rank[j] records which MPI rank owns point j
|
// we use weight to monitor code, later some day we can move it for optimization
|
||||||
// All ranks traverse the same block list so they all agree on ownership
|
int *weight;
|
||||||
int *owner_rank;
|
weight = new int[NN];
|
||||||
owner_rank = new int[NN];
|
memset(weight, 0, sizeof(int) * NN);
|
||||||
for (int j = 0; j < NN; j++)
|
|
||||||
owner_rank[j] = -1;
|
double *DH, *llb, *uub;
|
||||||
|
DH = new double[dim];
|
||||||
|
|
||||||
double DH[dim], llb[dim], uub[dim];
|
|
||||||
for (int i = 0; i < dim; i++)
|
for (int i = 0; i < dim; i++)
|
||||||
|
{
|
||||||
DH[i] = getdX(i);
|
DH[i] = getdX(i);
|
||||||
|
}
|
||||||
|
llb = new double[dim];
|
||||||
|
uub = new double[dim];
|
||||||
|
|
||||||
for (int j = 0; j < NN; j++) // run along points
|
for (int j = 0; j < NN; j++) // run along points
|
||||||
{
|
{
|
||||||
@@ -398,6 +403,12 @@ void Patch::Interp_Points(MyList<var> *VarList,
|
|||||||
bool flag = true;
|
bool flag = true;
|
||||||
for (int i = 0; i < dim; i++)
|
for (int i = 0; i < dim; i++)
|
||||||
{
|
{
|
||||||
|
// NOTE: our dividing structure is (exclude ghost)
|
||||||
|
// -1 0
|
||||||
|
// 1 2
|
||||||
|
// so (0,1) does not belong to any part for vertex structure
|
||||||
|
// here we put (0,0.5) to left part and (0.5,1) to right part
|
||||||
|
// BUT for cell structure the bbox is (-1.5,0.5) and (0.5,2.5), there is no missing region at all
|
||||||
#ifdef Vertex
|
#ifdef Vertex
|
||||||
#ifdef Cell
|
#ifdef Cell
|
||||||
#error Both Cell and Vertex are defined
|
#error Both Cell and Vertex are defined
|
||||||
@@ -422,7 +433,6 @@ void Patch::Interp_Points(MyList<var> *VarList,
|
|||||||
if (flag)
|
if (flag)
|
||||||
{
|
{
|
||||||
notfind = false;
|
notfind = false;
|
||||||
owner_rank[j] = BP->rank;
|
|
||||||
if (myrank == BP->rank)
|
if (myrank == BP->rank)
|
||||||
{
|
{
|
||||||
//---> interpolation
|
//---> interpolation
|
||||||
@@ -430,11 +440,14 @@ void Patch::Interp_Points(MyList<var> *VarList,
|
|||||||
int k = 0;
|
int k = 0;
|
||||||
while (varl) // run along variables
|
while (varl) // run along variables
|
||||||
{
|
{
|
||||||
f_global_interp(BP->shape, BP->X[0], BP->X[1], BP->X[2], BP->fgfs[varl->data->sgfn], Shellf[j * num_var + k],
|
// shellf[j*num_var+k] = Parallel::global_interp(dim,BP->shape,BP->X,BP->fgfs[varl->data->sgfn],
|
||||||
|
// pox,ordn,varl->data->SoA,Symmetry);
|
||||||
|
f_global_interp(BP->shape, BP->X[0], BP->X[1], BP->X[2], BP->fgfs[varl->data->sgfn], shellf[j * num_var + k],
|
||||||
pox[0], pox[1], pox[2], ordn, varl->data->SoA, Symmetry);
|
pox[0], pox[1], pox[2], ordn, varl->data->SoA, Symmetry);
|
||||||
varl = varl->next;
|
varl = varl->next;
|
||||||
k++;
|
k++;
|
||||||
}
|
}
|
||||||
|
weight[j] = 1;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
if (Bp == ble)
|
if (Bp == ble)
|
||||||
@@ -443,125 +456,61 @@ void Patch::Interp_Points(MyList<var> *VarList,
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// Replace MPI_Allreduce with per-owner MPI_Bcast:
|
MPI_Allreduce(shellf, Shellf, NN * num_var, MPI_DOUBLE, MPI_SUM, MPI_COMM_WORLD);
|
||||||
// Group consecutive points by owner rank and broadcast each group.
|
int *Weight;
|
||||||
// Since each point's data is non-zero only on the owner rank,
|
Weight = new int[NN];
|
||||||
// Bcast from owner is equivalent to Allreduce(MPI_SUM) but much cheaper.
|
MPI_Allreduce(weight, Weight, NN, MPI_INT, MPI_SUM, MPI_COMM_WORLD);
|
||||||
|
|
||||||
|
// misc::tillherecheck("print me");
|
||||||
|
|
||||||
|
for (int i = 0; i < NN; i++)
|
||||||
{
|
{
|
||||||
int j = 0;
|
if (Weight[i] > 1)
|
||||||
while (j < NN)
|
|
||||||
{
|
|
||||||
int cur_owner = owner_rank[j];
|
|
||||||
if (cur_owner < 0)
|
|
||||||
{
|
{
|
||||||
if (myrank == 0)
|
if (myrank == 0)
|
||||||
|
cout << "WARNING: Patch::Interp_Points meets multiple weight" << endl;
|
||||||
|
for (int j = 0; j < num_var; j++)
|
||||||
|
Shellf[j + i * num_var] = Shellf[j + i * num_var] / Weight[i];
|
||||||
|
}
|
||||||
|
else if (Weight[i] == 0 && myrank == 0)
|
||||||
{
|
{
|
||||||
cout << "ERROR: Patch::Interp_Points fails to find point (";
|
cout << "ERROR: Patch::Interp_Points fails to find point (";
|
||||||
for (int d = 0; d < dim; d++)
|
for (int j = 0; j < dim; j++)
|
||||||
{
|
{
|
||||||
cout << XX[d][j];
|
cout << XX[j][i];
|
||||||
if (d < dim - 1)
|
if (j < dim - 1)
|
||||||
cout << ",";
|
cout << ",";
|
||||||
else
|
else
|
||||||
cout << ")";
|
cout << ")";
|
||||||
}
|
}
|
||||||
cout << " on Patch (";
|
cout << " on Patch (";
|
||||||
for (int d = 0; d < dim; d++)
|
for (int j = 0; j < dim; j++)
|
||||||
{
|
{
|
||||||
cout << bbox[d] << "+" << lli[d] * DH[d];
|
cout << bbox[j] << "+" << lli[j] * getdX(j);
|
||||||
if (d < dim - 1)
|
if (j < dim - 1)
|
||||||
cout << ",";
|
cout << ",";
|
||||||
else
|
else
|
||||||
cout << ")--";
|
cout << ")--";
|
||||||
}
|
}
|
||||||
cout << "(";
|
cout << "(";
|
||||||
for (int d = 0; d < dim; d++)
|
for (int j = 0; j < dim; j++)
|
||||||
{
|
{
|
||||||
cout << bbox[dim + d] << "-" << uui[d] * DH[d];
|
cout << bbox[dim + j] << "-" << uui[j] * getdX(j);
|
||||||
if (d < dim - 1)
|
if (j < dim - 1)
|
||||||
cout << ",";
|
cout << ",";
|
||||||
else
|
else
|
||||||
cout << ")" << endl;
|
cout << ")" << endl;
|
||||||
}
|
}
|
||||||
MPI_Abort(MPI_COMM_WORLD, 1);
|
#if 0
|
||||||
}
|
checkBlock();
|
||||||
j++;
|
#else
|
||||||
continue;
|
cout << "splited domains:" << endl;
|
||||||
}
|
|
||||||
// Find contiguous run of points with the same owner
|
|
||||||
int jstart = j;
|
|
||||||
while (j < NN && owner_rank[j] == cur_owner)
|
|
||||||
j++;
|
|
||||||
int count = (j - jstart) * num_var;
|
|
||||||
MPI_Bcast(Shellf + jstart * num_var, count, MPI_DOUBLE, cur_owner, MPI_COMM_WORLD);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
delete[] owner_rank;
|
|
||||||
}
|
|
||||||
void Patch::Interp_Points(MyList<var> *VarList,
|
|
||||||
int NN, double **XX,
|
|
||||||
double *Shellf, int Symmetry,
|
|
||||||
int Nmin_consumer, int Nmax_consumer)
|
|
||||||
{
|
{
|
||||||
// Targeted point-to-point overload: each owner sends each point only to
|
|
||||||
// the one rank that needs it for integration (consumer), reducing
|
|
||||||
// communication volume by ~nprocs times compared to the Bcast version.
|
|
||||||
int myrank, nprocs;
|
|
||||||
MPI_Comm_rank(MPI_COMM_WORLD, &myrank);
|
|
||||||
MPI_Comm_size(MPI_COMM_WORLD, &nprocs);
|
|
||||||
|
|
||||||
int ordn = 2 * ghost_width;
|
|
||||||
MyList<var> *varl;
|
|
||||||
int num_var = 0;
|
|
||||||
varl = VarList;
|
|
||||||
while (varl)
|
|
||||||
{
|
|
||||||
num_var++;
|
|
||||||
varl = varl->next;
|
|
||||||
}
|
|
||||||
|
|
||||||
memset(Shellf, 0, sizeof(double) * NN * num_var);
|
|
||||||
|
|
||||||
// owner_rank[j] records which MPI rank owns point j
|
|
||||||
int *owner_rank;
|
|
||||||
owner_rank = new int[NN];
|
|
||||||
for (int j = 0; j < NN; j++)
|
|
||||||
owner_rank[j] = -1;
|
|
||||||
|
|
||||||
double DH[dim], llb[dim], uub[dim];
|
|
||||||
for (int i = 0; i < dim; i++)
|
|
||||||
DH[i] = getdX(i);
|
|
||||||
|
|
||||||
// --- Interpolation phase (identical to original) ---
|
|
||||||
for (int j = 0; j < NN; j++)
|
|
||||||
{
|
|
||||||
double pox[dim];
|
|
||||||
for (int i = 0; i < dim; i++)
|
|
||||||
{
|
|
||||||
pox[i] = XX[i][j];
|
|
||||||
if (myrank == 0 && (XX[i][j] < bbox[i] + lli[i] * DH[i] || XX[i][j] > bbox[dim + i] - uui[i] * DH[i]))
|
|
||||||
{
|
|
||||||
cout << "Patch::Interp_Points: point (";
|
|
||||||
for (int k = 0; k < dim; k++)
|
|
||||||
{
|
|
||||||
cout << XX[k][j];
|
|
||||||
if (k < dim - 1)
|
|
||||||
cout << ",";
|
|
||||||
else
|
|
||||||
cout << ") is out of current Patch." << endl;
|
|
||||||
}
|
|
||||||
MPI_Abort(MPI_COMM_WORLD, 1);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
MyList<Block> *Bp = blb;
|
MyList<Block> *Bp = blb;
|
||||||
bool notfind = true;
|
while (Bp)
|
||||||
while (notfind && Bp)
|
|
||||||
{
|
{
|
||||||
Block *BP = Bp->data;
|
Block *BP = Bp->data;
|
||||||
|
|
||||||
bool flag = true;
|
|
||||||
for (int i = 0; i < dim; i++)
|
for (int i = 0; i < dim; i++)
|
||||||
{
|
{
|
||||||
#ifdef Vertex
|
#ifdef Vertex
|
||||||
@@ -578,192 +527,32 @@ void Patch::Interp_Points(MyList<var> *VarList,
|
|||||||
#error Not define Vertex nor Cell
|
#error Not define Vertex nor Cell
|
||||||
#endif
|
#endif
|
||||||
#endif
|
#endif
|
||||||
if (XX[i][j] - llb[i] < -DH[i] / 2 || XX[i][j] - uub[i] > DH[i] / 2)
|
|
||||||
{
|
|
||||||
flag = false;
|
|
||||||
break;
|
|
||||||
}
|
}
|
||||||
}
|
cout << "(";
|
||||||
|
for (int j = 0; j < dim; j++)
|
||||||
if (flag)
|
|
||||||
{
|
{
|
||||||
notfind = false;
|
cout << llb[j] << ":" << uub[j];
|
||||||
owner_rank[j] = BP->rank;
|
if (j < dim - 1)
|
||||||
if (myrank == BP->rank)
|
cout << ",";
|
||||||
{
|
else
|
||||||
varl = VarList;
|
cout << ")" << endl;
|
||||||
int k = 0;
|
|
||||||
while (varl)
|
|
||||||
{
|
|
||||||
f_global_interp(BP->shape, BP->X[0], BP->X[1], BP->X[2], BP->fgfs[varl->data->sgfn], Shellf[j * num_var + k],
|
|
||||||
pox[0], pox[1], pox[2], ordn, varl->data->SoA, Symmetry);
|
|
||||||
varl = varl->next;
|
|
||||||
k++;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
if (Bp == ble)
|
if (Bp == ble)
|
||||||
break;
|
break;
|
||||||
Bp = Bp->next;
|
Bp = Bp->next;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
#endif
|
||||||
// --- Error check for unfound points ---
|
|
||||||
for (int j = 0; j < NN; j++)
|
|
||||||
{
|
|
||||||
if (owner_rank[j] < 0 && myrank == 0)
|
|
||||||
{
|
|
||||||
cout << "ERROR: Patch::Interp_Points fails to find point (";
|
|
||||||
for (int d = 0; d < dim; d++)
|
|
||||||
{
|
|
||||||
cout << XX[d][j];
|
|
||||||
if (d < dim - 1)
|
|
||||||
cout << ",";
|
|
||||||
else
|
|
||||||
cout << ")";
|
|
||||||
}
|
|
||||||
cout << " on Patch (";
|
|
||||||
for (int d = 0; d < dim; d++)
|
|
||||||
{
|
|
||||||
cout << bbox[d] << "+" << lli[d] * DH[d];
|
|
||||||
if (d < dim - 1)
|
|
||||||
cout << ",";
|
|
||||||
else
|
|
||||||
cout << ")--";
|
|
||||||
}
|
|
||||||
cout << "(";
|
|
||||||
for (int d = 0; d < dim; d++)
|
|
||||||
{
|
|
||||||
cout << bbox[dim + d] << "-" << uui[d] * DH[d];
|
|
||||||
if (d < dim - 1)
|
|
||||||
cout << ",";
|
|
||||||
else
|
|
||||||
cout << ")" << endl;
|
|
||||||
}
|
|
||||||
MPI_Abort(MPI_COMM_WORLD, 1);
|
MPI_Abort(MPI_COMM_WORLD, 1);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// --- Targeted point-to-point communication phase ---
|
delete[] shellf;
|
||||||
// Compute consumer_rank[j] using the same deterministic formula as surface_integral
|
delete[] weight;
|
||||||
int *consumer_rank = new int[NN];
|
delete[] Weight;
|
||||||
{
|
delete[] DH;
|
||||||
int mp = NN / nprocs;
|
delete[] llb;
|
||||||
int Lp = NN - nprocs * mp;
|
delete[] uub;
|
||||||
for (int j = 0; j < NN; j++)
|
|
||||||
{
|
|
||||||
if (j < Lp * (mp + 1))
|
|
||||||
consumer_rank[j] = j / (mp + 1);
|
|
||||||
else
|
|
||||||
consumer_rank[j] = Lp + (j - Lp * (mp + 1)) / mp;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// Count sends and recvs per rank
|
|
||||||
int *send_count = new int[nprocs];
|
|
||||||
int *recv_count = new int[nprocs];
|
|
||||||
memset(send_count, 0, sizeof(int) * nprocs);
|
|
||||||
memset(recv_count, 0, sizeof(int) * nprocs);
|
|
||||||
|
|
||||||
for (int j = 0; j < NN; j++)
|
|
||||||
{
|
|
||||||
int own = owner_rank[j];
|
|
||||||
int con = consumer_rank[j];
|
|
||||||
if (own == con)
|
|
||||||
continue; // local — no communication needed
|
|
||||||
if (own == myrank)
|
|
||||||
send_count[con]++;
|
|
||||||
if (con == myrank)
|
|
||||||
recv_count[own]++;
|
|
||||||
}
|
|
||||||
|
|
||||||
// Build send buffers: for each destination rank, pack (index, data) pairs
|
|
||||||
// Each entry: 1 int (point index j) + num_var doubles
|
|
||||||
int total_send = 0, total_recv = 0;
|
|
||||||
int *send_offset = new int[nprocs];
|
|
||||||
int *recv_offset = new int[nprocs];
|
|
||||||
for (int r = 0; r < nprocs; r++)
|
|
||||||
{
|
|
||||||
send_offset[r] = total_send;
|
|
||||||
total_send += send_count[r];
|
|
||||||
recv_offset[r] = total_recv;
|
|
||||||
total_recv += recv_count[r];
|
|
||||||
}
|
|
||||||
|
|
||||||
// Pack send buffers: each message contains (j, data[0..num_var-1]) per point
|
|
||||||
int stride = 1 + num_var; // 1 double for index + num_var doubles for data
|
|
||||||
double *sendbuf = new double[total_send * stride];
|
|
||||||
double *recvbuf = new double[total_recv * stride];
|
|
||||||
|
|
||||||
// Temporary counters for packing
|
|
||||||
int *pack_pos = new int[nprocs];
|
|
||||||
memset(pack_pos, 0, sizeof(int) * nprocs);
|
|
||||||
|
|
||||||
for (int j = 0; j < NN; j++)
|
|
||||||
{
|
|
||||||
int own = owner_rank[j];
|
|
||||||
int con = consumer_rank[j];
|
|
||||||
if (own != myrank || con == myrank)
|
|
||||||
continue;
|
|
||||||
int pos = (send_offset[con] + pack_pos[con]) * stride;
|
|
||||||
sendbuf[pos] = (double)j; // point index
|
|
||||||
for (int v = 0; v < num_var; v++)
|
|
||||||
sendbuf[pos + 1 + v] = Shellf[j * num_var + v];
|
|
||||||
pack_pos[con]++;
|
|
||||||
}
|
|
||||||
|
|
||||||
// Post non-blocking recvs and sends
|
|
||||||
int n_req = 0;
|
|
||||||
for (int r = 0; r < nprocs; r++)
|
|
||||||
{
|
|
||||||
if (recv_count[r] > 0) n_req++;
|
|
||||||
if (send_count[r] > 0) n_req++;
|
|
||||||
}
|
|
||||||
|
|
||||||
MPI_Request *reqs = new MPI_Request[n_req];
|
|
||||||
int req_idx = 0;
|
|
||||||
|
|
||||||
for (int r = 0; r < nprocs; r++)
|
|
||||||
{
|
|
||||||
if (recv_count[r] > 0)
|
|
||||||
{
|
|
||||||
MPI_Irecv(recvbuf + recv_offset[r] * stride,
|
|
||||||
recv_count[r] * stride, MPI_DOUBLE,
|
|
||||||
r, 0, MPI_COMM_WORLD, &reqs[req_idx++]);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
for (int r = 0; r < nprocs; r++)
|
|
||||||
{
|
|
||||||
if (send_count[r] > 0)
|
|
||||||
{
|
|
||||||
MPI_Isend(sendbuf + send_offset[r] * stride,
|
|
||||||
send_count[r] * stride, MPI_DOUBLE,
|
|
||||||
r, 0, MPI_COMM_WORLD, &reqs[req_idx++]);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
if (n_req > 0)
|
|
||||||
MPI_Waitall(n_req, reqs, MPI_STATUSES_IGNORE);
|
|
||||||
|
|
||||||
// Unpack recv buffers into Shellf
|
|
||||||
for (int i = 0; i < total_recv; i++)
|
|
||||||
{
|
|
||||||
int pos = i * stride;
|
|
||||||
int j = (int)recvbuf[pos];
|
|
||||||
for (int v = 0; v < num_var; v++)
|
|
||||||
Shellf[j * num_var + v] = recvbuf[pos + 1 + v];
|
|
||||||
}
|
|
||||||
|
|
||||||
delete[] reqs;
|
|
||||||
delete[] sendbuf;
|
|
||||||
delete[] recvbuf;
|
|
||||||
delete[] pack_pos;
|
|
||||||
delete[] send_offset;
|
|
||||||
delete[] recv_offset;
|
|
||||||
delete[] send_count;
|
|
||||||
delete[] recv_count;
|
|
||||||
delete[] consumer_rank;
|
|
||||||
delete[] owner_rank;
|
|
||||||
}
|
}
|
||||||
void Patch::Interp_Points(MyList<var> *VarList,
|
void Patch::Interp_Points(MyList<var> *VarList,
|
||||||
int NN, double **XX,
|
int NN, double **XX,
|
||||||
@@ -784,22 +573,24 @@ void Patch::Interp_Points(MyList<var> *VarList,
|
|||||||
varl = varl->next;
|
varl = varl->next;
|
||||||
}
|
}
|
||||||
|
|
||||||
memset(Shellf, 0, sizeof(double) * NN * num_var);
|
double *shellf;
|
||||||
|
shellf = new double[NN * num_var];
|
||||||
|
memset(shellf, 0, sizeof(double) * NN * num_var);
|
||||||
|
|
||||||
// owner_rank[j] stores the global rank that owns point j
|
// we use weight to monitor code, later some day we can move it for optimization
|
||||||
int *owner_rank;
|
int *weight;
|
||||||
owner_rank = new int[NN];
|
weight = new int[NN];
|
||||||
for (int j = 0; j < NN; j++)
|
memset(weight, 0, sizeof(int) * NN);
|
||||||
owner_rank[j] = -1;
|
|
||||||
|
|
||||||
// Build global-to-local rank translation for Comm_here
|
double *DH, *llb, *uub;
|
||||||
MPI_Group world_group, local_group;
|
DH = new double[dim];
|
||||||
MPI_Comm_group(MPI_COMM_WORLD, &world_group);
|
|
||||||
MPI_Comm_group(Comm_here, &local_group);
|
|
||||||
|
|
||||||
double DH[dim], llb[dim], uub[dim];
|
|
||||||
for (int i = 0; i < dim; i++)
|
for (int i = 0; i < dim; i++)
|
||||||
|
{
|
||||||
DH[i] = getdX(i);
|
DH[i] = getdX(i);
|
||||||
|
}
|
||||||
|
llb = new double[dim];
|
||||||
|
uub = new double[dim];
|
||||||
|
|
||||||
for (int j = 0; j < NN; j++) // run along points
|
for (int j = 0; j < NN; j++) // run along points
|
||||||
{
|
{
|
||||||
@@ -831,6 +622,12 @@ void Patch::Interp_Points(MyList<var> *VarList,
|
|||||||
bool flag = true;
|
bool flag = true;
|
||||||
for (int i = 0; i < dim; i++)
|
for (int i = 0; i < dim; i++)
|
||||||
{
|
{
|
||||||
|
// NOTE: our dividing structure is (exclude ghost)
|
||||||
|
// -1 0
|
||||||
|
// 1 2
|
||||||
|
// so (0,1) does not belong to any part for vertex structure
|
||||||
|
// here we put (0,0.5) to left part and (0.5,1) to right part
|
||||||
|
// BUT for cell structure the bbox is (-1.5,0.5) and (0.5,2.5), there is no missing region at all
|
||||||
#ifdef Vertex
|
#ifdef Vertex
|
||||||
#ifdef Cell
|
#ifdef Cell
|
||||||
#error Both Cell and Vertex are defined
|
#error Both Cell and Vertex are defined
|
||||||
@@ -855,7 +652,6 @@ void Patch::Interp_Points(MyList<var> *VarList,
|
|||||||
if (flag)
|
if (flag)
|
||||||
{
|
{
|
||||||
notfind = false;
|
notfind = false;
|
||||||
owner_rank[j] = BP->rank;
|
|
||||||
if (myrank == BP->rank)
|
if (myrank == BP->rank)
|
||||||
{
|
{
|
||||||
//---> interpolation
|
//---> interpolation
|
||||||
@@ -863,11 +659,14 @@ void Patch::Interp_Points(MyList<var> *VarList,
|
|||||||
int k = 0;
|
int k = 0;
|
||||||
while (varl) // run along variables
|
while (varl) // run along variables
|
||||||
{
|
{
|
||||||
f_global_interp(BP->shape, BP->X[0], BP->X[1], BP->X[2], BP->fgfs[varl->data->sgfn], Shellf[j * num_var + k],
|
// shellf[j*num_var+k] = Parallel::global_interp(dim,BP->shape,BP->X,BP->fgfs[varl->data->sgfn],
|
||||||
|
// pox,ordn,varl->data->SoA,Symmetry);
|
||||||
|
f_global_interp(BP->shape, BP->X[0], BP->X[1], BP->X[2], BP->fgfs[varl->data->sgfn], shellf[j * num_var + k],
|
||||||
pox[0], pox[1], pox[2], ordn, varl->data->SoA, Symmetry);
|
pox[0], pox[1], pox[2], ordn, varl->data->SoA, Symmetry);
|
||||||
varl = varl->next;
|
varl = varl->next;
|
||||||
k++;
|
k++;
|
||||||
}
|
}
|
||||||
|
weight[j] = 1;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
if (Bp == ble)
|
if (Bp == ble)
|
||||||
@@ -876,35 +675,97 @@ void Patch::Interp_Points(MyList<var> *VarList,
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// Collect unique global owner ranks and translate to local ranks in Comm_here
|
MPI_Allreduce(shellf, Shellf, NN * num_var, MPI_DOUBLE, MPI_SUM, Comm_here);
|
||||||
// Then broadcast each owner's points via MPI_Bcast on Comm_here
|
int *Weight;
|
||||||
{
|
Weight = new int[NN];
|
||||||
int j = 0;
|
MPI_Allreduce(weight, Weight, NN, MPI_INT, MPI_SUM, Comm_here);
|
||||||
while (j < NN)
|
|
||||||
{
|
|
||||||
int cur_owner_global = owner_rank[j];
|
|
||||||
if (cur_owner_global < 0)
|
|
||||||
{
|
|
||||||
// Point not found — skip (error check disabled for sub-communicator levels)
|
|
||||||
j++;
|
|
||||||
continue;
|
|
||||||
}
|
|
||||||
// Translate global rank to local rank in Comm_here
|
|
||||||
int cur_owner_local;
|
|
||||||
MPI_Group_translate_ranks(world_group, 1, &cur_owner_global, local_group, &cur_owner_local);
|
|
||||||
|
|
||||||
// Find contiguous run of points with the same owner
|
// misc::tillherecheck("print me");
|
||||||
int jstart = j;
|
// if(lmyrank == 0) cout<<"myrank = "<<myrank<<"print me"<<endl;
|
||||||
while (j < NN && owner_rank[j] == cur_owner_global)
|
|
||||||
j++;
|
for (int i = 0; i < NN; i++)
|
||||||
int count = (j - jstart) * num_var;
|
{
|
||||||
MPI_Bcast(Shellf + jstart * num_var, count, MPI_DOUBLE, cur_owner_local, Comm_here);
|
if (Weight[i] > 1)
|
||||||
|
{
|
||||||
|
if (lmyrank == 0)
|
||||||
|
cout << "WARNING: Patch::Interp_Points meets multiple weight" << endl;
|
||||||
|
for (int j = 0; j < num_var; j++)
|
||||||
|
Shellf[j + i * num_var] = Shellf[j + i * num_var] / Weight[i];
|
||||||
}
|
}
|
||||||
|
#if 0 // for not involved levels, this may fail
|
||||||
|
else if(Weight[i] == 0 && lmyrank == 0)
|
||||||
|
{
|
||||||
|
cout<<"ERROR: Patch::Interp_Points fails to find point (";
|
||||||
|
for(int j=0;j<dim;j++)
|
||||||
|
{
|
||||||
|
cout<<XX[j][i];
|
||||||
|
if(j<dim-1) cout<<",";
|
||||||
|
else cout<<")";
|
||||||
|
}
|
||||||
|
cout<<" on Patch (";
|
||||||
|
for(int j=0;j<dim;j++)
|
||||||
|
{
|
||||||
|
cout<<bbox[j]<<"+"<<lli[j]*getdX(j);
|
||||||
|
if(j<dim-1) cout<<",";
|
||||||
|
else cout<<")--";
|
||||||
|
}
|
||||||
|
cout<<"(";
|
||||||
|
for(int j=0;j<dim;j++)
|
||||||
|
{
|
||||||
|
cout<<bbox[dim+j]<<"-"<<uui[j]*getdX(j);
|
||||||
|
if(j<dim-1) cout<<",";
|
||||||
|
else cout<<")"<<endl;
|
||||||
|
}
|
||||||
|
#if 0
|
||||||
|
checkBlock();
|
||||||
|
#else
|
||||||
|
cout<<"splited domains:"<<endl;
|
||||||
|
{
|
||||||
|
MyList<Block> *Bp=blb;
|
||||||
|
while(Bp)
|
||||||
|
{
|
||||||
|
Block *BP=Bp->data;
|
||||||
|
|
||||||
|
for(int i=0;i<dim;i++)
|
||||||
|
{
|
||||||
|
#ifdef Vertex
|
||||||
|
#ifdef Cell
|
||||||
|
#error Both Cell and Vertex are defined
|
||||||
|
#endif
|
||||||
|
llb[i] = (feq(BP->bbox[i] ,bbox[i] ,DH[i]/2)) ? BP->bbox[i]+lli[i]*DH[i] : BP->bbox[i] +(ghost_width-0.5)*DH[i];
|
||||||
|
uub[i] = (feq(BP->bbox[dim+i],bbox[dim+i],DH[i]/2)) ? BP->bbox[dim+i]-uui[i]*DH[i] : BP->bbox[dim+i]-(ghost_width-0.5)*DH[i];
|
||||||
|
#else
|
||||||
|
#ifdef Cell
|
||||||
|
llb[i] = (feq(BP->bbox[i] ,bbox[i] ,DH[i]/2)) ? BP->bbox[i]+lli[i]*DH[i] : BP->bbox[i] +ghost_width*DH[i];
|
||||||
|
uub[i] = (feq(BP->bbox[dim+i],bbox[dim+i],DH[i]/2)) ? BP->bbox[dim+i]-uui[i]*DH[i] : BP->bbox[dim+i]-ghost_width*DH[i];
|
||||||
|
#else
|
||||||
|
#error Not define Vertex nor Cell
|
||||||
|
#endif
|
||||||
|
#endif
|
||||||
|
}
|
||||||
|
cout<<"(";
|
||||||
|
for(int j=0;j<dim;j++)
|
||||||
|
{
|
||||||
|
cout<<llb[j]<<":"<<uub[j];
|
||||||
|
if(j<dim-1) cout<<",";
|
||||||
|
else cout<<")"<<endl;
|
||||||
|
}
|
||||||
|
if(Bp == ble) break;
|
||||||
|
Bp=Bp->next;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
#endif
|
||||||
|
MPI_Abort(MPI_COMM_WORLD,1);
|
||||||
|
}
|
||||||
|
#endif
|
||||||
}
|
}
|
||||||
|
|
||||||
MPI_Group_free(&world_group);
|
delete[] shellf;
|
||||||
MPI_Group_free(&local_group);
|
delete[] weight;
|
||||||
delete[] owner_rank;
|
delete[] Weight;
|
||||||
|
delete[] DH;
|
||||||
|
delete[] llb;
|
||||||
|
delete[] uub;
|
||||||
}
|
}
|
||||||
void Patch::checkBlock()
|
void Patch::checkBlock()
|
||||||
{
|
{
|
||||||
|
|||||||
@@ -39,10 +39,6 @@ public:
|
|||||||
|
|
||||||
bool Find_Point(double *XX);
|
bool Find_Point(double *XX);
|
||||||
|
|
||||||
void Interp_Points(MyList<var> *VarList,
|
|
||||||
int NN, double **XX,
|
|
||||||
double *Shellf, int Symmetry,
|
|
||||||
int Nmin_consumer, int Nmax_consumer);
|
|
||||||
void Interp_Points(MyList<var> *VarList,
|
void Interp_Points(MyList<var> *VarList,
|
||||||
int NN, double **XX,
|
int NN, double **XX,
|
||||||
double *Shellf, int Symmetry, MPI_Comm Comm_here);
|
double *Shellf, int Symmetry, MPI_Comm Comm_here);
|
||||||
|
|||||||
File diff suppressed because it is too large
Load Diff
@@ -81,43 +81,53 @@ namespace Parallel
|
|||||||
int Symmetry);
|
int Symmetry);
|
||||||
void Sync(Patch *Pat, MyList<var> *VarList, int Symmetry);
|
void Sync(Patch *Pat, MyList<var> *VarList, int Symmetry);
|
||||||
void Sync(MyList<Patch> *PatL, MyList<var> *VarList, int Symmetry);
|
void Sync(MyList<Patch> *PatL, MyList<var> *VarList, int Symmetry);
|
||||||
void Sync_merged(MyList<Patch> *PatL, MyList<var> *VarList, int Symmetry);
|
|
||||||
|
|
||||||
struct SyncCache {
|
// Async Sync: overlap MPI communication with computation
|
||||||
bool valid;
|
struct TransferState
|
||||||
int cpusize;
|
{
|
||||||
MyList<gridseg> **combined_src;
|
|
||||||
MyList<gridseg> **combined_dst;
|
|
||||||
int *send_lengths;
|
|
||||||
int *recv_lengths;
|
|
||||||
double **send_bufs;
|
|
||||||
double **recv_bufs;
|
|
||||||
int *send_buf_caps;
|
|
||||||
int *recv_buf_caps;
|
|
||||||
MPI_Request *reqs;
|
MPI_Request *reqs;
|
||||||
MPI_Status *stats;
|
MPI_Status *stats;
|
||||||
int max_reqs;
|
|
||||||
bool lengths_valid;
|
|
||||||
SyncCache();
|
|
||||||
void invalidate();
|
|
||||||
void destroy();
|
|
||||||
};
|
|
||||||
|
|
||||||
void Sync_cached(MyList<Patch> *PatL, MyList<var> *VarList, int Symmetry, SyncCache &cache);
|
|
||||||
void transfer_cached(MyList<gridseg> **src, MyList<gridseg> **dst,
|
|
||||||
MyList<var> *VarList1, MyList<var> *VarList2,
|
|
||||||
int Symmetry, SyncCache &cache);
|
|
||||||
|
|
||||||
struct AsyncSyncState {
|
|
||||||
int req_no;
|
int req_no;
|
||||||
bool active;
|
double **send_data;
|
||||||
AsyncSyncState() : req_no(0), active(false) {}
|
double **rec_data;
|
||||||
|
int cpusize;
|
||||||
|
MyList<gridseg> **transfer_src;
|
||||||
|
MyList<gridseg> **transfer_dst;
|
||||||
|
MyList<gridseg> **src;
|
||||||
|
MyList<gridseg> *dst;
|
||||||
|
MyList<var> *VarList1;
|
||||||
|
MyList<var> *VarList2;
|
||||||
|
int Symmetry;
|
||||||
|
bool owns_gsl; // true if this state owns and should free the GSLs
|
||||||
};
|
};
|
||||||
|
struct SyncHandle
|
||||||
|
{
|
||||||
|
TransferState *states;
|
||||||
|
int num_states;
|
||||||
|
};
|
||||||
|
SyncHandle *SyncBegin(Patch *Pat, MyList<var> *VarList, int Symmetry);
|
||||||
|
SyncHandle *SyncBegin(MyList<Patch> *PatL, MyList<var> *VarList, int Symmetry);
|
||||||
|
void SyncEnd(SyncHandle *handle);
|
||||||
|
|
||||||
void Sync_start(MyList<Patch> *PatL, MyList<var> *VarList, int Symmetry,
|
// Cached GSL plan: pre-build grid segment lists once, reuse across multiple Sync calls
|
||||||
SyncCache &cache, AsyncSyncState &state);
|
struct SyncPlanEntry
|
||||||
void Sync_finish(SyncCache &cache, AsyncSyncState &state,
|
{
|
||||||
MyList<var> *VarList, int Symmetry);
|
int cpusize;
|
||||||
|
MyList<gridseg> **transfer_src;
|
||||||
|
MyList<gridseg> **transfer_dst;
|
||||||
|
MyList<gridseg> **src;
|
||||||
|
MyList<gridseg> *dst;
|
||||||
|
};
|
||||||
|
struct SyncPlan
|
||||||
|
{
|
||||||
|
SyncPlanEntry *entries;
|
||||||
|
int num_entries;
|
||||||
|
int Symmetry;
|
||||||
|
};
|
||||||
|
SyncPlan *SyncPreparePlan(MyList<Patch> *PatL, int Symmetry);
|
||||||
|
void SyncFreePlan(SyncPlan *plan);
|
||||||
|
SyncHandle *SyncBeginWithPlan(SyncPlan *plan, MyList<var> *VarList);
|
||||||
|
SyncHandle *SyncBeginWithPlan(SyncPlan *plan, MyList<var> *VarList1, MyList<var> *VarList2);
|
||||||
void OutBdLow2Hi(Patch *Patc, Patch *Patf,
|
void OutBdLow2Hi(Patch *Patc, Patch *Patf,
|
||||||
MyList<var> *VarList1 /* source */, MyList<var> *VarList2 /* target */,
|
MyList<var> *VarList1 /* source */, MyList<var> *VarList2 /* target */,
|
||||||
int Symmetry);
|
int Symmetry);
|
||||||
@@ -130,15 +140,6 @@ namespace Parallel
|
|||||||
void OutBdLow2Himix(MyList<Patch> *PatcL, MyList<Patch> *PatfL,
|
void OutBdLow2Himix(MyList<Patch> *PatcL, MyList<Patch> *PatfL,
|
||||||
MyList<var> *VarList1 /* source */, MyList<var> *VarList2 /* target */,
|
MyList<var> *VarList1 /* source */, MyList<var> *VarList2 /* target */,
|
||||||
int Symmetry);
|
int Symmetry);
|
||||||
void Restrict_cached(MyList<Patch> *PatcL, MyList<Patch> *PatfL,
|
|
||||||
MyList<var> *VarList1, MyList<var> *VarList2,
|
|
||||||
int Symmetry, SyncCache &cache);
|
|
||||||
void OutBdLow2Hi_cached(MyList<Patch> *PatcL, MyList<Patch> *PatfL,
|
|
||||||
MyList<var> *VarList1, MyList<var> *VarList2,
|
|
||||||
int Symmetry, SyncCache &cache);
|
|
||||||
void OutBdLow2Himix_cached(MyList<Patch> *PatcL, MyList<Patch> *PatfL,
|
|
||||||
MyList<var> *VarList1, MyList<var> *VarList2,
|
|
||||||
int Symmetry, SyncCache &cache);
|
|
||||||
void Prolong(Patch *Patc, Patch *Patf,
|
void Prolong(Patch *Patc, Patch *Patf,
|
||||||
MyList<var> *VarList1 /* source */, MyList<var> *VarList2 /* target */,
|
MyList<var> *VarList1 /* source */, MyList<var> *VarList2 /* target */,
|
||||||
int Symmetry);
|
int Symmetry);
|
||||||
|
|||||||
@@ -186,6 +186,12 @@ void Z4c_class::Step(int lev, int YN)
|
|||||||
int ERROR = 0;
|
int ERROR = 0;
|
||||||
|
|
||||||
MyList<ss_patch> *sPp;
|
MyList<ss_patch> *sPp;
|
||||||
|
|
||||||
|
// Pre-build grid segment lists once for this level's patches.
|
||||||
|
// These are reused across predictor + 3 corrector SyncBegin calls,
|
||||||
|
// avoiding O(cpusize * blocks^2) rebuild each time.
|
||||||
|
Parallel::SyncPlan *sync_plan = Parallel::SyncPreparePlan(GH->PatL[lev], Symmetry);
|
||||||
|
|
||||||
// Predictor
|
// Predictor
|
||||||
MyList<Patch> *Pp = GH->PatL[lev];
|
MyList<Patch> *Pp = GH->PatL[lev];
|
||||||
while (Pp)
|
while (Pp)
|
||||||
@@ -321,13 +327,17 @@ void Z4c_class::Step(int lev, int YN)
|
|||||||
}
|
}
|
||||||
Pp = Pp->next;
|
Pp = Pp->next;
|
||||||
}
|
}
|
||||||
// check error information
|
// Start async ghost zone exchange - overlaps with error check and Shell computation
|
||||||
|
Parallel::SyncHandle *sync_pre = Parallel::SyncBeginWithPlan(sync_plan, SynchList_pre);
|
||||||
|
|
||||||
|
// check error information (overlaps with MPI transfer)
|
||||||
{
|
{
|
||||||
int erh = ERROR;
|
int erh = ERROR;
|
||||||
MPI_Allreduce(&erh, &ERROR, 1, MPI_INT, MPI_SUM, MPI_COMM_WORLD);
|
MPI_Allreduce(&erh, &ERROR, 1, MPI_INT, MPI_SUM, MPI_COMM_WORLD);
|
||||||
}
|
}
|
||||||
if (ERROR)
|
if (ERROR)
|
||||||
{
|
{
|
||||||
|
Parallel::SyncEnd(sync_pre); sync_pre = 0;
|
||||||
Parallel::Dump_Data(GH->PatL[lev], StateList, 0, PhysTime, dT_lev);
|
Parallel::Dump_Data(GH->PatL[lev], StateList, 0, PhysTime, dT_lev);
|
||||||
if (myrank == 0)
|
if (myrank == 0)
|
||||||
{
|
{
|
||||||
@@ -475,6 +485,7 @@ void Z4c_class::Step(int lev, int YN)
|
|||||||
}
|
}
|
||||||
if (ERROR)
|
if (ERROR)
|
||||||
{
|
{
|
||||||
|
Parallel::SyncEnd(sync_pre); sync_pre = 0;
|
||||||
SH->Dump_Data(StateList, 0, PhysTime, dT_lev);
|
SH->Dump_Data(StateList, 0, PhysTime, dT_lev);
|
||||||
if (myrank == 0)
|
if (myrank == 0)
|
||||||
{
|
{
|
||||||
@@ -485,7 +496,26 @@ void Z4c_class::Step(int lev, int YN)
|
|||||||
}
|
}
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
// CA-RK4: skip post-prediction sync (redundant; ghost cells computable locally)
|
// Complete async ghost zone exchange
|
||||||
|
if (sync_pre) Parallel::SyncEnd(sync_pre);
|
||||||
|
|
||||||
|
#ifdef WithShell
|
||||||
|
if (lev == 0)
|
||||||
|
{
|
||||||
|
clock_t prev_clock, curr_clock;
|
||||||
|
if (myrank == 0)
|
||||||
|
curr_clock = clock();
|
||||||
|
SH->Synch(SynchList_pre, Symmetry);
|
||||||
|
if (myrank == 0)
|
||||||
|
{
|
||||||
|
prev_clock = curr_clock;
|
||||||
|
curr_clock = clock();
|
||||||
|
cout << " Shell stuff synchronization used "
|
||||||
|
<< (double)(curr_clock - prev_clock) / ((double)CLOCKS_PER_SEC)
|
||||||
|
<< " seconds! " << endl;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
#endif
|
||||||
|
|
||||||
// for black hole position
|
// for black hole position
|
||||||
if (BH_num > 0 && lev == GH->levels - 1)
|
if (BH_num > 0 && lev == GH->levels - 1)
|
||||||
@@ -675,13 +705,17 @@ void Z4c_class::Step(int lev, int YN)
|
|||||||
Pp = Pp->next;
|
Pp = Pp->next;
|
||||||
}
|
}
|
||||||
|
|
||||||
// check error information
|
// Start async ghost zone exchange - overlaps with error check and Shell computation
|
||||||
|
Parallel::SyncHandle *sync_cor = Parallel::SyncBeginWithPlan(sync_plan, SynchList_cor);
|
||||||
|
|
||||||
|
// check error information (overlaps with MPI transfer)
|
||||||
{
|
{
|
||||||
int erh = ERROR;
|
int erh = ERROR;
|
||||||
MPI_Allreduce(&erh, &ERROR, 1, MPI_INT, MPI_SUM, MPI_COMM_WORLD);
|
MPI_Allreduce(&erh, &ERROR, 1, MPI_INT, MPI_SUM, MPI_COMM_WORLD);
|
||||||
}
|
}
|
||||||
if (ERROR)
|
if (ERROR)
|
||||||
{
|
{
|
||||||
|
Parallel::SyncEnd(sync_cor); sync_cor = 0;
|
||||||
Parallel::Dump_Data(GH->PatL[lev], SynchList_pre, 0, PhysTime, dT_lev);
|
Parallel::Dump_Data(GH->PatL[lev], SynchList_pre, 0, PhysTime, dT_lev);
|
||||||
if (myrank == 0)
|
if (myrank == 0)
|
||||||
{
|
{
|
||||||
@@ -839,6 +873,7 @@ void Z4c_class::Step(int lev, int YN)
|
|||||||
}
|
}
|
||||||
if (ERROR)
|
if (ERROR)
|
||||||
{
|
{
|
||||||
|
Parallel::SyncEnd(sync_cor); sync_cor = 0;
|
||||||
SH->Dump_Data(SynchList_pre, 0, PhysTime, dT_lev);
|
SH->Dump_Data(SynchList_pre, 0, PhysTime, dT_lev);
|
||||||
if (myrank == 0)
|
if (myrank == 0)
|
||||||
{
|
{
|
||||||
@@ -850,9 +885,8 @@ void Z4c_class::Step(int lev, int YN)
|
|||||||
}
|
}
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
// CA-RK4: only sync after last corrector (iter_count == 3); stages 1 & 2 are redundant
|
// Complete async ghost zone exchange
|
||||||
if (iter_count == 3) {
|
if (sync_cor) Parallel::SyncEnd(sync_cor);
|
||||||
Parallel::Sync(GH->PatL[lev], SynchList_cor, Symmetry);
|
|
||||||
|
|
||||||
#ifdef WithShell
|
#ifdef WithShell
|
||||||
if (lev == 0)
|
if (lev == 0)
|
||||||
@@ -871,7 +905,6 @@ void Z4c_class::Step(int lev, int YN)
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
#endif
|
#endif
|
||||||
} // end CA-RK4 guard
|
|
||||||
// for black hole position
|
// for black hole position
|
||||||
if (BH_num > 0 && lev == GH->levels - 1)
|
if (BH_num > 0 && lev == GH->levels - 1)
|
||||||
{
|
{
|
||||||
@@ -1027,6 +1060,8 @@ void Z4c_class::Step(int lev, int YN)
|
|||||||
Porg0[ithBH][2] = Porg1[ithBH][2];
|
Porg0[ithBH][2] = Porg1[ithBH][2];
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
Parallel::SyncFreePlan(sync_plan);
|
||||||
}
|
}
|
||||||
#else
|
#else
|
||||||
// for constraint preserving boundary (CPBC)
|
// for constraint preserving boundary (CPBC)
|
||||||
@@ -1060,6 +1095,10 @@ void Z4c_class::Step(int lev, int YN)
|
|||||||
int ERROR = 0;
|
int ERROR = 0;
|
||||||
|
|
||||||
MyList<ss_patch> *sPp;
|
MyList<ss_patch> *sPp;
|
||||||
|
|
||||||
|
// Pre-build grid segment lists once for this level's patches.
|
||||||
|
Parallel::SyncPlan *sync_plan = Parallel::SyncPreparePlan(GH->PatL[lev], Symmetry);
|
||||||
|
|
||||||
// Predictor
|
// Predictor
|
||||||
MyList<Patch> *Pp = GH->PatL[lev];
|
MyList<Patch> *Pp = GH->PatL[lev];
|
||||||
while (Pp)
|
while (Pp)
|
||||||
@@ -1527,13 +1566,17 @@ void Z4c_class::Step(int lev, int YN)
|
|||||||
}
|
}
|
||||||
#endif
|
#endif
|
||||||
}
|
}
|
||||||
// check error information
|
// Start async ghost zone exchange - overlaps with error check
|
||||||
|
Parallel::SyncHandle *sync_pre = Parallel::SyncBeginWithPlan(sync_plan, SynchList_pre);
|
||||||
|
|
||||||
|
// check error information (overlaps with MPI transfer)
|
||||||
{
|
{
|
||||||
int erh = ERROR;
|
int erh = ERROR;
|
||||||
MPI_Allreduce(&erh, &ERROR, 1, MPI_INT, MPI_SUM, MPI_COMM_WORLD);
|
MPI_Allreduce(&erh, &ERROR, 1, MPI_INT, MPI_SUM, MPI_COMM_WORLD);
|
||||||
}
|
}
|
||||||
if (ERROR)
|
if (ERROR)
|
||||||
{
|
{
|
||||||
|
Parallel::SyncEnd(sync_pre); sync_pre = 0;
|
||||||
SH->Dump_Data(StateList, 0, PhysTime, dT_lev);
|
SH->Dump_Data(StateList, 0, PhysTime, dT_lev);
|
||||||
if (myrank == 0)
|
if (myrank == 0)
|
||||||
{
|
{
|
||||||
@@ -1543,7 +1586,8 @@ void Z4c_class::Step(int lev, int YN)
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// CA-RK4: skip post-prediction MPI ghost sync (redundant; ghost cells computable locally)
|
// Complete async ghost zone exchange
|
||||||
|
if (sync_pre) Parallel::SyncEnd(sync_pre);
|
||||||
|
|
||||||
if (lev == 0)
|
if (lev == 0)
|
||||||
{
|
{
|
||||||
@@ -2088,13 +2132,17 @@ void Z4c_class::Step(int lev, int YN)
|
|||||||
sPp = sPp->next;
|
sPp = sPp->next;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
// check error information
|
// Start async ghost zone exchange - overlaps with error check
|
||||||
|
Parallel::SyncHandle *sync_cor = Parallel::SyncBeginWithPlan(sync_plan, SynchList_cor);
|
||||||
|
|
||||||
|
// check error information (overlaps with MPI transfer)
|
||||||
{
|
{
|
||||||
int erh = ERROR;
|
int erh = ERROR;
|
||||||
MPI_Allreduce(&erh, &ERROR, 1, MPI_INT, MPI_SUM, MPI_COMM_WORLD);
|
MPI_Allreduce(&erh, &ERROR, 1, MPI_INT, MPI_SUM, MPI_COMM_WORLD);
|
||||||
}
|
}
|
||||||
if (ERROR)
|
if (ERROR)
|
||||||
{
|
{
|
||||||
|
Parallel::SyncEnd(sync_cor); sync_cor = 0;
|
||||||
SH->Dump_Data(SynchList_pre, 0, PhysTime, dT_lev);
|
SH->Dump_Data(SynchList_pre, 0, PhysTime, dT_lev);
|
||||||
if (myrank == 0)
|
if (myrank == 0)
|
||||||
{
|
{
|
||||||
@@ -2105,9 +2153,8 @@ void Z4c_class::Step(int lev, int YN)
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// CA-RK4: only MPI sync after last corrector (iter_count == 3); stages 1 & 2 are redundant
|
// Complete async ghost zone exchange
|
||||||
if (iter_count == 3)
|
if (sync_cor) Parallel::SyncEnd(sync_cor);
|
||||||
Parallel::Sync(GH->PatL[lev], SynchList_cor, Symmetry);
|
|
||||||
|
|
||||||
if (lev == 0)
|
if (lev == 0)
|
||||||
{
|
{
|
||||||
@@ -2333,6 +2380,8 @@ void Z4c_class::Step(int lev, int YN)
|
|||||||
DG_List->clearList();
|
DG_List->clearList();
|
||||||
}
|
}
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
|
Parallel::SyncFreePlan(sync_plan);
|
||||||
}
|
}
|
||||||
#endif
|
#endif
|
||||||
#undef MRBD
|
#undef MRBD
|
||||||
|
|||||||
@@ -1221,7 +1221,25 @@ void bssnEM_class::Step(int lev, int YN)
|
|||||||
}
|
}
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
// CA-RK4: skip post-prediction sync (redundant; ghost cells computable locally)
|
Parallel::Sync(GH->PatL[lev], SynchList_pre, Symmetry);
|
||||||
|
|
||||||
|
#ifdef WithShell
|
||||||
|
if (lev == 0)
|
||||||
|
{
|
||||||
|
clock_t prev_clock, curr_clock;
|
||||||
|
if (myrank == 0)
|
||||||
|
curr_clock = clock();
|
||||||
|
SH->Synch(SynchList_pre, Symmetry);
|
||||||
|
if (myrank == 0)
|
||||||
|
{
|
||||||
|
prev_clock = curr_clock;
|
||||||
|
curr_clock = clock();
|
||||||
|
cout << " Shell stuff synchronization used "
|
||||||
|
<< (double)(curr_clock - prev_clock) / ((double)CLOCKS_PER_SEC)
|
||||||
|
<< " seconds! " << endl;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
#endif
|
||||||
|
|
||||||
// for black hole position
|
// for black hole position
|
||||||
if (BH_num > 0 && lev == GH->levels - 1)
|
if (BH_num > 0 && lev == GH->levels - 1)
|
||||||
@@ -1665,8 +1683,6 @@ void bssnEM_class::Step(int lev, int YN)
|
|||||||
}
|
}
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
// CA-RK4: only sync after last corrector (iter_count == 3); stages 1 & 2 are redundant
|
|
||||||
if (iter_count == 3) {
|
|
||||||
Parallel::Sync(GH->PatL[lev], SynchList_cor, Symmetry);
|
Parallel::Sync(GH->PatL[lev], SynchList_cor, Symmetry);
|
||||||
|
|
||||||
#ifdef WithShell
|
#ifdef WithShell
|
||||||
@@ -1686,7 +1702,6 @@ void bssnEM_class::Step(int lev, int YN)
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
#endif
|
#endif
|
||||||
} // end CA-RK4 guard
|
|
||||||
// for black hole position
|
// for black hole position
|
||||||
if (BH_num > 0 && lev == GH->levels - 1)
|
if (BH_num > 0 && lev == GH->levels - 1)
|
||||||
{
|
{
|
||||||
|
|||||||
@@ -730,12 +730,6 @@ void bssn_class::Initialize()
|
|||||||
PhysTime = StartTime;
|
PhysTime = StartTime;
|
||||||
Setup_Black_Hole_position();
|
Setup_Black_Hole_position();
|
||||||
}
|
}
|
||||||
|
|
||||||
// Initialize sync caches (per-level, for predictor and corrector)
|
|
||||||
sync_cache_pre = new Parallel::SyncCache[GH->levels];
|
|
||||||
sync_cache_cor = new Parallel::SyncCache[GH->levels];
|
|
||||||
sync_cache_rp_coarse = new Parallel::SyncCache[GH->levels];
|
|
||||||
sync_cache_rp_fine = new Parallel::SyncCache[GH->levels];
|
|
||||||
}
|
}
|
||||||
|
|
||||||
//================================================================================================
|
//================================================================================================
|
||||||
@@ -987,32 +981,6 @@ bssn_class::~bssn_class()
|
|||||||
delete Azzz;
|
delete Azzz;
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
// Destroy sync caches before GH
|
|
||||||
if (sync_cache_pre)
|
|
||||||
{
|
|
||||||
for (int i = 0; i < GH->levels; i++)
|
|
||||||
sync_cache_pre[i].destroy();
|
|
||||||
delete[] sync_cache_pre;
|
|
||||||
}
|
|
||||||
if (sync_cache_cor)
|
|
||||||
{
|
|
||||||
for (int i = 0; i < GH->levels; i++)
|
|
||||||
sync_cache_cor[i].destroy();
|
|
||||||
delete[] sync_cache_cor;
|
|
||||||
}
|
|
||||||
if (sync_cache_rp_coarse)
|
|
||||||
{
|
|
||||||
for (int i = 0; i < GH->levels; i++)
|
|
||||||
sync_cache_rp_coarse[i].destroy();
|
|
||||||
delete[] sync_cache_rp_coarse;
|
|
||||||
}
|
|
||||||
if (sync_cache_rp_fine)
|
|
||||||
{
|
|
||||||
for (int i = 0; i < GH->levels; i++)
|
|
||||||
sync_cache_rp_fine[i].destroy();
|
|
||||||
delete[] sync_cache_rp_fine;
|
|
||||||
}
|
|
||||||
|
|
||||||
delete GH;
|
delete GH;
|
||||||
#ifdef WithShell
|
#ifdef WithShell
|
||||||
delete SH;
|
delete SH;
|
||||||
@@ -2213,7 +2181,6 @@ void bssn_class::Evolve(int Steps)
|
|||||||
GH->Regrid(Symmetry, BH_num, Porgbr, Porg0,
|
GH->Regrid(Symmetry, BH_num, Porgbr, Porg0,
|
||||||
SynchList_cor, OldStateList, StateList, SynchList_pre,
|
SynchList_cor, OldStateList, StateList, SynchList_pre,
|
||||||
fgt(PhysTime - dT_mon, StartTime, dT_mon / 2), ErrorMonitor);
|
fgt(PhysTime - dT_mon, StartTime, dT_mon / 2), ErrorMonitor);
|
||||||
for (int il = 0; il < GH->levels; il++) { sync_cache_pre[il].invalidate(); sync_cache_cor[il].invalidate(); sync_cache_rp_coarse[il].invalidate(); sync_cache_rp_fine[il].invalidate(); }
|
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
#if (REGLEV == 0 && (PSTR == 1 || PSTR == 2))
|
#if (REGLEV == 0 && (PSTR == 1 || PSTR == 2))
|
||||||
@@ -2426,10 +2393,9 @@ void bssn_class::RecursiveStep(int lev)
|
|||||||
#endif
|
#endif
|
||||||
|
|
||||||
#if (REGLEV == 0)
|
#if (REGLEV == 0)
|
||||||
if (GH->Regrid_Onelevel(lev, Symmetry, BH_num, Porgbr, Porg0,
|
GH->Regrid_Onelevel(lev, Symmetry, BH_num, Porgbr, Porg0,
|
||||||
SynchList_cor, OldStateList, StateList, SynchList_pre,
|
SynchList_cor, OldStateList, StateList, SynchList_pre,
|
||||||
fgt(PhysTime - dT_lev, StartTime, dT_lev / 2), ErrorMonitor))
|
fgt(PhysTime - dT_lev, StartTime, dT_lev / 2), ErrorMonitor);
|
||||||
for (int il = 0; il < GH->levels; il++) { sync_cache_pre[il].invalidate(); sync_cache_cor[il].invalidate(); sync_cache_rp_coarse[il].invalidate(); sync_cache_rp_fine[il].invalidate(); }
|
|
||||||
#endif
|
#endif
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -2605,10 +2571,9 @@ void bssn_class::ParallelStep()
|
|||||||
delete[] tporg;
|
delete[] tporg;
|
||||||
delete[] tporgo;
|
delete[] tporgo;
|
||||||
#if (REGLEV == 0)
|
#if (REGLEV == 0)
|
||||||
if (GH->Regrid_Onelevel(GH->mylev, Symmetry, BH_num, Porgbr, Porg0,
|
GH->Regrid_Onelevel(GH->mylev, Symmetry, BH_num, Porgbr, Porg0,
|
||||||
SynchList_cor, OldStateList, StateList, SynchList_pre,
|
SynchList_cor, OldStateList, StateList, SynchList_pre,
|
||||||
fgt(PhysTime - dT_lev, StartTime, dT_lev / 2), ErrorMonitor))
|
fgt(PhysTime - dT_lev, StartTime, dT_lev / 2), ErrorMonitor);
|
||||||
for (int il = 0; il < GH->levels; il++) { sync_cache_pre[il].invalidate(); sync_cache_cor[il].invalidate(); sync_cache_rp_coarse[il].invalidate(); sync_cache_rp_fine[il].invalidate(); }
|
|
||||||
#endif
|
#endif
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -2772,10 +2737,9 @@ void bssn_class::ParallelStep()
|
|||||||
if (lev + 1 >= GH->movls)
|
if (lev + 1 >= GH->movls)
|
||||||
{
|
{
|
||||||
// GH->Regrid_Onelevel_aux(lev,Symmetry,BH_num,Porgbr,Porg0,
|
// GH->Regrid_Onelevel_aux(lev,Symmetry,BH_num,Porgbr,Porg0,
|
||||||
if (GH->Regrid_Onelevel(lev + 1, Symmetry, BH_num, Porgbr, Porg0,
|
GH->Regrid_Onelevel(lev + 1, Symmetry, BH_num, Porgbr, Porg0,
|
||||||
SynchList_cor, OldStateList, StateList, SynchList_pre,
|
SynchList_cor, OldStateList, StateList, SynchList_pre,
|
||||||
fgt(PhysTime - dT_levp1, StartTime, dT_levp1 / 2), ErrorMonitor))
|
fgt(PhysTime - dT_levp1, StartTime, dT_levp1 / 2), ErrorMonitor);
|
||||||
for (int il = 0; il < GH->levels; il++) { sync_cache_pre[il].invalidate(); sync_cache_cor[il].invalidate(); sync_cache_rp_coarse[il].invalidate(); sync_cache_rp_fine[il].invalidate(); }
|
|
||||||
|
|
||||||
// a_stream.clear();
|
// a_stream.clear();
|
||||||
// a_stream.str("");
|
// a_stream.str("");
|
||||||
@@ -2787,10 +2751,9 @@ void bssn_class::ParallelStep()
|
|||||||
// for this level
|
// for this level
|
||||||
if (YN == 1)
|
if (YN == 1)
|
||||||
{
|
{
|
||||||
if (GH->Regrid_Onelevel(lev, Symmetry, BH_num, Porgbr, Porg0,
|
GH->Regrid_Onelevel(lev, Symmetry, BH_num, Porgbr, Porg0,
|
||||||
SynchList_cor, OldStateList, StateList, SynchList_pre,
|
SynchList_cor, OldStateList, StateList, SynchList_pre,
|
||||||
fgt(PhysTime - dT_lev, StartTime, dT_lev / 2), ErrorMonitor))
|
fgt(PhysTime - dT_lev, StartTime, dT_lev / 2), ErrorMonitor);
|
||||||
for (int il = 0; il < GH->levels; il++) { sync_cache_pre[il].invalidate(); sync_cache_cor[il].invalidate(); sync_cache_rp_coarse[il].invalidate(); sync_cache_rp_fine[il].invalidate(); }
|
|
||||||
|
|
||||||
// a_stream.clear();
|
// a_stream.clear();
|
||||||
// a_stream.str("");
|
// a_stream.str("");
|
||||||
@@ -2806,10 +2769,9 @@ void bssn_class::ParallelStep()
|
|||||||
if (YN == 1)
|
if (YN == 1)
|
||||||
{
|
{
|
||||||
// GH->Regrid_Onelevel_aux(lev-2,Symmetry,BH_num,Porgbr,Porg0,
|
// GH->Regrid_Onelevel_aux(lev-2,Symmetry,BH_num,Porgbr,Porg0,
|
||||||
if (GH->Regrid_Onelevel(lev - 1, Symmetry, BH_num, Porgbr, Porg0,
|
GH->Regrid_Onelevel(lev - 1, Symmetry, BH_num, Porgbr, Porg0,
|
||||||
SynchList_cor, OldStateList, StateList, SynchList_pre,
|
SynchList_cor, OldStateList, StateList, SynchList_pre,
|
||||||
fgt(PhysTime - dT_lev, StartTime, dT_levm1 / 2), ErrorMonitor))
|
fgt(PhysTime - dT_lev, StartTime, dT_levm1 / 2), ErrorMonitor);
|
||||||
for (int il = 0; il < GH->levels; il++) { sync_cache_pre[il].invalidate(); sync_cache_cor[il].invalidate(); sync_cache_rp_coarse[il].invalidate(); sync_cache_rp_fine[il].invalidate(); }
|
|
||||||
|
|
||||||
// a_stream.clear();
|
// a_stream.clear();
|
||||||
// a_stream.str("");
|
// a_stream.str("");
|
||||||
@@ -2822,10 +2784,9 @@ void bssn_class::ParallelStep()
|
|||||||
if (i % 4 == 3)
|
if (i % 4 == 3)
|
||||||
{
|
{
|
||||||
// GH->Regrid_Onelevel_aux(lev-2,Symmetry,BH_num,Porgbr,Porg0,
|
// GH->Regrid_Onelevel_aux(lev-2,Symmetry,BH_num,Porgbr,Porg0,
|
||||||
if (GH->Regrid_Onelevel(lev - 1, Symmetry, BH_num, Porgbr, Porg0,
|
GH->Regrid_Onelevel(lev - 1, Symmetry, BH_num, Porgbr, Porg0,
|
||||||
SynchList_cor, OldStateList, StateList, SynchList_pre,
|
SynchList_cor, OldStateList, StateList, SynchList_pre,
|
||||||
fgt(PhysTime - dT_lev, StartTime, dT_levm1 / 2), ErrorMonitor))
|
fgt(PhysTime - dT_lev, StartTime, dT_levm1 / 2), ErrorMonitor);
|
||||||
for (int il = 0; il < GH->levels; il++) { sync_cache_pre[il].invalidate(); sync_cache_cor[il].invalidate(); sync_cache_rp_coarse[il].invalidate(); sync_cache_rp_fine[il].invalidate(); }
|
|
||||||
|
|
||||||
// a_stream.clear();
|
// a_stream.clear();
|
||||||
// a_stream.str("");
|
// a_stream.str("");
|
||||||
@@ -3074,6 +3035,12 @@ void bssn_class::Step(int lev, int YN)
|
|||||||
int ERROR = 0;
|
int ERROR = 0;
|
||||||
|
|
||||||
MyList<ss_patch> *sPp;
|
MyList<ss_patch> *sPp;
|
||||||
|
|
||||||
|
// Pre-build grid segment lists once for this level's patches.
|
||||||
|
// These are reused across predictor + 3 corrector SyncBegin calls,
|
||||||
|
// avoiding O(cpusize * blocks^2) rebuild each time.
|
||||||
|
Parallel::SyncPlan *sync_plan = Parallel::SyncPreparePlan(GH->PatL[lev], Symmetry);
|
||||||
|
|
||||||
// Predictor
|
// Predictor
|
||||||
MyList<Patch> *Pp = GH->PatL[lev];
|
MyList<Patch> *Pp = GH->PatL[lev];
|
||||||
while (Pp)
|
while (Pp)
|
||||||
@@ -3197,7 +3164,26 @@ void bssn_class::Step(int lev, int YN)
|
|||||||
}
|
}
|
||||||
Pp = Pp->next;
|
Pp = Pp->next;
|
||||||
}
|
}
|
||||||
// NOTE: error check deferred to after Shell Patch computation to reduce MPI_Allreduce calls
|
|
||||||
|
// Start async ghost zone exchange - overlaps with error check and Shell computation
|
||||||
|
Parallel::SyncHandle *sync_pre = Parallel::SyncBeginWithPlan(sync_plan, SynchList_pre);
|
||||||
|
|
||||||
|
// check error information (overlaps with MPI transfer)
|
||||||
|
{
|
||||||
|
int erh = ERROR;
|
||||||
|
MPI_Allreduce(&erh, &ERROR, 1, MPI_INT, MPI_SUM, MPI_COMM_WORLD);
|
||||||
|
}
|
||||||
|
if (ERROR)
|
||||||
|
{
|
||||||
|
Parallel::SyncEnd(sync_pre); sync_pre = 0;
|
||||||
|
Parallel::Dump_Data(GH->PatL[lev], StateList, 0, PhysTime, dT_lev);
|
||||||
|
if (myrank == 0)
|
||||||
|
{
|
||||||
|
if (ErrorMonitor->outfile)
|
||||||
|
ErrorMonitor->outfile << "find NaN in state variables at t = " << PhysTime << ", lev = " << lev << endl;
|
||||||
|
MPI_Abort(MPI_COMM_WORLD, 1);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
#ifdef WithShell
|
#ifdef WithShell
|
||||||
// evolve Shell Patches
|
// evolve Shell Patches
|
||||||
@@ -3341,32 +3327,46 @@ void bssn_class::Step(int lev, int YN)
|
|||||||
#endif
|
#endif
|
||||||
}
|
}
|
||||||
|
|
||||||
// Non-blocking error reduction overlapped with Sync to hide Allreduce latency
|
// check error information
|
||||||
MPI_Request err_req;
|
|
||||||
{
|
{
|
||||||
int erh = ERROR;
|
int erh = ERROR;
|
||||||
MPI_Iallreduce(&erh, &ERROR, 1, MPI_INT, MPI_SUM, MPI_COMM_WORLD, &err_req);
|
MPI_Allreduce(&erh, &ERROR, 1, MPI_INT, MPI_SUM, MPI_COMM_WORLD);
|
||||||
}
|
}
|
||||||
#endif
|
|
||||||
|
|
||||||
// CA-RK4: skip post-prediction sync (redundant; ghost cells computable locally)
|
|
||||||
|
|
||||||
#ifdef WithShell
|
|
||||||
// Complete non-blocking error reduction and check
|
|
||||||
MPI_Wait(&err_req, MPI_STATUS_IGNORE);
|
|
||||||
if (ERROR)
|
if (ERROR)
|
||||||
{
|
{
|
||||||
Parallel::Dump_Data(GH->PatL[lev], StateList, 0, PhysTime, dT_lev);
|
Parallel::SyncEnd(sync_pre); sync_pre = 0;
|
||||||
SH->Dump_Data(StateList, 0, PhysTime, dT_lev);
|
SH->Dump_Data(StateList, 0, PhysTime, dT_lev);
|
||||||
if (myrank == 0)
|
if (myrank == 0)
|
||||||
{
|
{
|
||||||
if (ErrorMonitor->outfile)
|
if (ErrorMonitor->outfile)
|
||||||
ErrorMonitor->outfile << "find NaN in state variables at t = " << PhysTime << ", lev = " << lev << endl;
|
ErrorMonitor->outfile << "find NaN in state variables on Shell Patches at t = " << PhysTime << endl;
|
||||||
MPI_Abort(MPI_COMM_WORLD, 1);
|
MPI_Abort(MPI_COMM_WORLD, 1);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
|
// Complete async ghost zone exchange
|
||||||
|
if (sync_pre) Parallel::SyncEnd(sync_pre);
|
||||||
|
|
||||||
|
#ifdef WithShell
|
||||||
|
if (lev == 0)
|
||||||
|
{
|
||||||
|
clock_t prev_clock, curr_clock;
|
||||||
|
if (myrank == 0)
|
||||||
|
curr_clock = clock();
|
||||||
|
SH->Synch(SynchList_pre, Symmetry);
|
||||||
|
if (myrank == 0)
|
||||||
|
{
|
||||||
|
prev_clock = curr_clock;
|
||||||
|
curr_clock = clock();
|
||||||
|
cout << " Shell stuff synchronization used "
|
||||||
|
<< (double)(curr_clock - prev_clock) / ((double)CLOCKS_PER_SEC)
|
||||||
|
<< " seconds! " << endl;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
#endif
|
||||||
|
|
||||||
#if (MAPBH == 0)
|
#if (MAPBH == 0)
|
||||||
// for black hole position
|
// for black hole position
|
||||||
if (BH_num > 0 && lev == GH->levels - 1)
|
if (BH_num > 0 && lev == GH->levels - 1)
|
||||||
@@ -3541,7 +3541,28 @@ void bssn_class::Step(int lev, int YN)
|
|||||||
Pp = Pp->next;
|
Pp = Pp->next;
|
||||||
}
|
}
|
||||||
|
|
||||||
// NOTE: error check deferred to after Shell Patch computation to reduce MPI_Allreduce calls
|
// Start async ghost zone exchange - overlaps with error check and Shell computation
|
||||||
|
Parallel::SyncHandle *sync_cor = Parallel::SyncBeginWithPlan(sync_plan, SynchList_cor);
|
||||||
|
|
||||||
|
// check error information (overlaps with MPI transfer)
|
||||||
|
{
|
||||||
|
int erh = ERROR;
|
||||||
|
MPI_Allreduce(&erh, &ERROR, 1, MPI_INT, MPI_SUM, MPI_COMM_WORLD);
|
||||||
|
}
|
||||||
|
|
||||||
|
if (ERROR)
|
||||||
|
{
|
||||||
|
Parallel::SyncEnd(sync_cor); sync_cor = 0;
|
||||||
|
Parallel::Dump_Data(GH->PatL[lev], SynchList_pre, 0, PhysTime, dT_lev);
|
||||||
|
if (myrank == 0)
|
||||||
|
{
|
||||||
|
if (ErrorMonitor->outfile)
|
||||||
|
ErrorMonitor->outfile << "find NaN in RK4 substep#" << iter_count
|
||||||
|
<< " variables at t = " << PhysTime
|
||||||
|
<< ", lev = " << lev << endl;
|
||||||
|
MPI_Abort(MPI_COMM_WORLD, 1);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
#ifdef WithShell
|
#ifdef WithShell
|
||||||
// evolve Shell Patches
|
// evolve Shell Patches
|
||||||
@@ -3681,18 +3702,28 @@ void bssn_class::Step(int lev, int YN)
|
|||||||
sPp = sPp->next;
|
sPp = sPp->next;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
// Non-blocking error reduction overlapped with Sync to hide Allreduce latency
|
// check error information
|
||||||
MPI_Request err_req_cor;
|
|
||||||
{
|
{
|
||||||
int erh = ERROR;
|
int erh = ERROR;
|
||||||
MPI_Iallreduce(&erh, &ERROR, 1, MPI_INT, MPI_SUM, MPI_COMM_WORLD, &err_req_cor);
|
MPI_Allreduce(&erh, &ERROR, 1, MPI_INT, MPI_SUM, MPI_COMM_WORLD);
|
||||||
|
}
|
||||||
|
if (ERROR)
|
||||||
|
{
|
||||||
|
Parallel::SyncEnd(sync_cor); sync_cor = 0;
|
||||||
|
SH->Dump_Data(SynchList_pre, 0, PhysTime, dT_lev);
|
||||||
|
if (myrank == 0)
|
||||||
|
{
|
||||||
|
if (ErrorMonitor->outfile)
|
||||||
|
ErrorMonitor->outfile << "find NaN on Shell Patches in RK4 substep#"
|
||||||
|
<< iter_count << " variables at t = "
|
||||||
|
<< PhysTime << endl;
|
||||||
|
MPI_Abort(MPI_COMM_WORLD, 1);
|
||||||
|
}
|
||||||
}
|
}
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
// CA-RK4: only sync after last corrector (iter_count == 3); stages 1 & 2 are redundant
|
// Complete async ghost zone exchange
|
||||||
if (iter_count == 3) {
|
if (sync_cor) Parallel::SyncEnd(sync_cor);
|
||||||
Parallel::AsyncSyncState async_cor;
|
|
||||||
Parallel::Sync_start(GH->PatL[lev], SynchList_cor, Symmetry, sync_cache_cor[lev], async_cor);
|
|
||||||
|
|
||||||
#ifdef WithShell
|
#ifdef WithShell
|
||||||
if (lev == 0)
|
if (lev == 0)
|
||||||
@@ -3711,26 +3742,6 @@ void bssn_class::Step(int lev, int YN)
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
#endif
|
#endif
|
||||||
Parallel::Sync_finish(sync_cache_cor[lev], async_cor, SynchList_cor, Symmetry);
|
|
||||||
} // end CA-RK4 guard
|
|
||||||
|
|
||||||
#ifdef WithShell
|
|
||||||
// Complete non-blocking error reduction and check
|
|
||||||
MPI_Wait(&err_req_cor, MPI_STATUS_IGNORE);
|
|
||||||
if (ERROR)
|
|
||||||
{
|
|
||||||
Parallel::Dump_Data(GH->PatL[lev], SynchList_pre, 0, PhysTime, dT_lev);
|
|
||||||
SH->Dump_Data(SynchList_pre, 0, PhysTime, dT_lev);
|
|
||||||
if (myrank == 0)
|
|
||||||
{
|
|
||||||
if (ErrorMonitor->outfile)
|
|
||||||
ErrorMonitor->outfile << "find NaN in RK4 substep#" << iter_count
|
|
||||||
<< " variables at t = " << PhysTime
|
|
||||||
<< ", lev = " << lev << endl;
|
|
||||||
MPI_Abort(MPI_COMM_WORLD, 1);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
#endif
|
|
||||||
|
|
||||||
#if (MAPBH == 0)
|
#if (MAPBH == 0)
|
||||||
// for black hole position
|
// for black hole position
|
||||||
@@ -3903,6 +3914,8 @@ void bssn_class::Step(int lev, int YN)
|
|||||||
Porg0[ithBH][2] = Porg1[ithBH][2];
|
Porg0[ithBH][2] = Porg1[ithBH][2];
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
Parallel::SyncFreePlan(sync_plan);
|
||||||
}
|
}
|
||||||
|
|
||||||
//================================================================================================
|
//================================================================================================
|
||||||
@@ -4042,7 +4055,22 @@ void bssn_class::Step(int lev, int YN)
|
|||||||
}
|
}
|
||||||
Pp = Pp->next;
|
Pp = Pp->next;
|
||||||
}
|
}
|
||||||
// NOTE: error check deferred to after Shell Patch computation to reduce MPI_Allreduce calls
|
// check error information
|
||||||
|
{
|
||||||
|
int erh = ERROR;
|
||||||
|
MPI_Allreduce(&erh, &ERROR, 1, MPI_INT, MPI_SUM, MPI_COMM_WORLD);
|
||||||
|
}
|
||||||
|
if (ERROR)
|
||||||
|
{
|
||||||
|
Parallel::Dump_Data(GH->PatL[lev], StateList, 0, PhysTime, dT_lev);
|
||||||
|
if (myrank == 0)
|
||||||
|
{
|
||||||
|
if (ErrorMonitor->outfile)
|
||||||
|
ErrorMonitor->outfile << "find NaN in state variables at t = " << PhysTime
|
||||||
|
<< ", lev = " << lev << endl;
|
||||||
|
MPI_Abort(MPI_COMM_WORLD, 1);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
#ifdef WithShell
|
#ifdef WithShell
|
||||||
// evolve Shell Patches
|
// evolve Shell Patches
|
||||||
@@ -4183,16 +4211,25 @@ void bssn_class::Step(int lev, int YN)
|
|||||||
}
|
}
|
||||||
#endif
|
#endif
|
||||||
}
|
}
|
||||||
// Non-blocking error reduction overlapped with Sync to hide Allreduce latency
|
// check error information
|
||||||
MPI_Request err_req;
|
|
||||||
{
|
{
|
||||||
int erh = ERROR;
|
int erh = ERROR;
|
||||||
MPI_Iallreduce(&erh, &ERROR, 1, MPI_INT, MPI_SUM, MPI_COMM_WORLD, &err_req);
|
MPI_Allreduce(&erh, &ERROR, 1, MPI_INT, MPI_SUM, MPI_COMM_WORLD);
|
||||||
|
}
|
||||||
|
if (ERROR)
|
||||||
|
{
|
||||||
|
SH->Dump_Data(StateList, 0, PhysTime, dT_lev);
|
||||||
|
if (myrank == 0)
|
||||||
|
{
|
||||||
|
if (ErrorMonitor->outfile)
|
||||||
|
ErrorMonitor->outfile << "find NaN in state variables on Shell Patches at t = "
|
||||||
|
<< PhysTime << endl;
|
||||||
|
MPI_Abort(MPI_COMM_WORLD, 1);
|
||||||
|
}
|
||||||
}
|
}
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
Parallel::AsyncSyncState async_pre;
|
Parallel::Sync(GH->PatL[lev], SynchList_pre, Symmetry);
|
||||||
Parallel::Sync_start(GH->PatL[lev], SynchList_pre, Symmetry, sync_cache_pre[lev], async_pre);
|
|
||||||
|
|
||||||
#ifdef WithShell
|
#ifdef WithShell
|
||||||
if (lev == 0)
|
if (lev == 0)
|
||||||
@@ -4211,24 +4248,6 @@ void bssn_class::Step(int lev, int YN)
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
#endif
|
#endif
|
||||||
Parallel::Sync_finish(sync_cache_pre[lev], async_pre, SynchList_pre, Symmetry);
|
|
||||||
|
|
||||||
#ifdef WithShell
|
|
||||||
// Complete non-blocking error reduction and check
|
|
||||||
MPI_Wait(&err_req, MPI_STATUS_IGNORE);
|
|
||||||
if (ERROR)
|
|
||||||
{
|
|
||||||
Parallel::Dump_Data(GH->PatL[lev], StateList, 0, PhysTime, dT_lev);
|
|
||||||
SH->Dump_Data(StateList, 0, PhysTime, dT_lev);
|
|
||||||
if (myrank == 0)
|
|
||||||
{
|
|
||||||
if (ErrorMonitor->outfile)
|
|
||||||
ErrorMonitor->outfile << "find NaN in state variables at t = " << PhysTime
|
|
||||||
<< ", lev = " << lev << endl;
|
|
||||||
MPI_Abort(MPI_COMM_WORLD, 1);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
#endif
|
|
||||||
|
|
||||||
// for black hole position
|
// for black hole position
|
||||||
if (BH_num > 0 && lev == GH->levels - 1)
|
if (BH_num > 0 && lev == GH->levels - 1)
|
||||||
@@ -4388,7 +4407,23 @@ void bssn_class::Step(int lev, int YN)
|
|||||||
Pp = Pp->next;
|
Pp = Pp->next;
|
||||||
}
|
}
|
||||||
|
|
||||||
// NOTE: error check deferred to after Shell Patch computation to reduce MPI_Allreduce calls
|
// check error information
|
||||||
|
{
|
||||||
|
int erh = ERROR;
|
||||||
|
MPI_Allreduce(&erh, &ERROR, 1, MPI_INT, MPI_SUM, MPI_COMM_WORLD);
|
||||||
|
}
|
||||||
|
if (ERROR)
|
||||||
|
{
|
||||||
|
Parallel::Dump_Data(GH->PatL[lev], SynchList_pre, 0, PhysTime, dT_lev);
|
||||||
|
if (myrank == 0)
|
||||||
|
{
|
||||||
|
if (ErrorMonitor->outfile)
|
||||||
|
ErrorMonitor->outfile << "find NaN in RK4 substep#" << iter_count
|
||||||
|
<< " variables at t = " << PhysTime
|
||||||
|
<< ", lev = " << lev << endl;
|
||||||
|
MPI_Abort(MPI_COMM_WORLD, 1);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
#ifdef WithShell
|
#ifdef WithShell
|
||||||
// evolve Shell Patches
|
// evolve Shell Patches
|
||||||
@@ -4528,16 +4563,25 @@ void bssn_class::Step(int lev, int YN)
|
|||||||
sPp = sPp->next;
|
sPp = sPp->next;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
// Non-blocking error reduction overlapped with Sync to hide Allreduce latency
|
// check error information
|
||||||
MPI_Request err_req_cor;
|
|
||||||
{
|
{
|
||||||
int erh = ERROR;
|
int erh = ERROR;
|
||||||
MPI_Iallreduce(&erh, &ERROR, 1, MPI_INT, MPI_SUM, MPI_COMM_WORLD, &err_req_cor);
|
MPI_Allreduce(&erh, &ERROR, 1, MPI_INT, MPI_SUM, MPI_COMM_WORLD);
|
||||||
|
}
|
||||||
|
if (ERROR)
|
||||||
|
{
|
||||||
|
SH->Dump_Data(SynchList_pre, 0, PhysTime, dT_lev);
|
||||||
|
if (myrank == 0)
|
||||||
|
{
|
||||||
|
if (ErrorMonitor->outfile)
|
||||||
|
ErrorMonitor->outfile << "find NaN on Shell Patches in RK4 substep#" << iter_count
|
||||||
|
<< " variables at t = " << PhysTime << endl;
|
||||||
|
MPI_Abort(MPI_COMM_WORLD, 1);
|
||||||
|
}
|
||||||
}
|
}
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
Parallel::AsyncSyncState async_cor;
|
Parallel::Sync(GH->PatL[lev], SynchList_cor, Symmetry);
|
||||||
Parallel::Sync_start(GH->PatL[lev], SynchList_cor, Symmetry, sync_cache_cor[lev], async_cor);
|
|
||||||
|
|
||||||
#ifdef WithShell
|
#ifdef WithShell
|
||||||
if (lev == 0)
|
if (lev == 0)
|
||||||
@@ -4555,25 +4599,6 @@ void bssn_class::Step(int lev, int YN)
|
|||||||
<< " seconds! " << endl;
|
<< " seconds! " << endl;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
#endif
|
|
||||||
Parallel::Sync_finish(sync_cache_cor[lev], async_cor, SynchList_cor, Symmetry);
|
|
||||||
|
|
||||||
#ifdef WithShell
|
|
||||||
// Complete non-blocking error reduction and check
|
|
||||||
MPI_Wait(&err_req_cor, MPI_STATUS_IGNORE);
|
|
||||||
if (ERROR)
|
|
||||||
{
|
|
||||||
Parallel::Dump_Data(GH->PatL[lev], SynchList_pre, 0, PhysTime, dT_lev);
|
|
||||||
SH->Dump_Data(SynchList_pre, 0, PhysTime, dT_lev);
|
|
||||||
if (myrank == 0)
|
|
||||||
{
|
|
||||||
if (ErrorMonitor->outfile)
|
|
||||||
ErrorMonitor->outfile << "find NaN in RK4 substep#" << iter_count
|
|
||||||
<< " variables at t = " << PhysTime
|
|
||||||
<< ", lev = " << lev << endl;
|
|
||||||
MPI_Abort(MPI_COMM_WORLD, 1);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
#endif
|
#endif
|
||||||
// for black hole position
|
// for black hole position
|
||||||
if (BH_num > 0 && lev == GH->levels - 1)
|
if (BH_num > 0 && lev == GH->levels - 1)
|
||||||
@@ -4813,6 +4838,12 @@ void bssn_class::Step(int lev, int YN)
|
|||||||
int ERROR = 0;
|
int ERROR = 0;
|
||||||
|
|
||||||
MyList<ss_patch> *sPp;
|
MyList<ss_patch> *sPp;
|
||||||
|
|
||||||
|
// Pre-build grid segment lists once for this level's patches.
|
||||||
|
// These are reused across predictor + 3 corrector SyncBegin calls,
|
||||||
|
// avoiding O(cpusize * blocks^2) rebuild each time.
|
||||||
|
Parallel::SyncPlan *sync_plan = Parallel::SyncPreparePlan(GH->PatL[lev], Symmetry);
|
||||||
|
|
||||||
// Predictor
|
// Predictor
|
||||||
MyList<Patch> *Pp = GH->PatL[lev];
|
MyList<Patch> *Pp = GH->PatL[lev];
|
||||||
while (Pp)
|
while (Pp)
|
||||||
@@ -4939,21 +4970,17 @@ void bssn_class::Step(int lev, int YN)
|
|||||||
|
|
||||||
// misc::tillherecheck(GH->Commlev[lev],GH->start_rank[lev],"after Predictor rhs calculation");
|
// misc::tillherecheck(GH->Commlev[lev],GH->start_rank[lev],"after Predictor rhs calculation");
|
||||||
|
|
||||||
// Non-blocking error reduction overlapped with Sync to hide Allreduce latency
|
// Start async ghost zone exchange - overlaps with error check and BH position
|
||||||
MPI_Request err_req;
|
Parallel::SyncHandle *sync_pre = Parallel::SyncBeginWithPlan(sync_plan, SynchList_pre);
|
||||||
|
|
||||||
|
// check error information (overlaps with MPI transfer)
|
||||||
{
|
{
|
||||||
int erh = ERROR;
|
int erh = ERROR;
|
||||||
MPI_Iallreduce(&erh, &ERROR, 1, MPI_INT, MPI_SUM, GH->Commlev[lev], &err_req);
|
MPI_Allreduce(&erh, &ERROR, 1, MPI_INT, MPI_SUM, GH->Commlev[lev]);
|
||||||
}
|
}
|
||||||
|
|
||||||
// misc::tillherecheck(GH->Commlev[lev],GH->start_rank[lev],"before Predictor sync");
|
|
||||||
|
|
||||||
Parallel::Sync_cached(GH->PatL[lev], SynchList_pre, Symmetry, sync_cache_pre[lev]);
|
|
||||||
|
|
||||||
// Complete non-blocking error reduction and check
|
|
||||||
MPI_Wait(&err_req, MPI_STATUS_IGNORE);
|
|
||||||
if (ERROR)
|
if (ERROR)
|
||||||
{
|
{
|
||||||
|
Parallel::SyncEnd(sync_pre); sync_pre = 0;
|
||||||
Parallel::Dump_Data(GH->PatL[lev], StateList, 0, PhysTime, dT_lev);
|
Parallel::Dump_Data(GH->PatL[lev], StateList, 0, PhysTime, dT_lev);
|
||||||
if (myrank == 0)
|
if (myrank == 0)
|
||||||
{
|
{
|
||||||
@@ -4963,6 +4990,11 @@ void bssn_class::Step(int lev, int YN)
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// misc::tillherecheck(GH->Commlev[lev],GH->start_rank[lev],"before Predictor sync");
|
||||||
|
|
||||||
|
// Complete async ghost zone exchange
|
||||||
|
if (sync_pre) Parallel::SyncEnd(sync_pre);
|
||||||
|
|
||||||
#if (MAPBH == 0)
|
#if (MAPBH == 0)
|
||||||
// for black hole position
|
// for black hole position
|
||||||
if (BH_num > 0 && lev == GH->levels - 1)
|
if (BH_num > 0 && lev == GH->levels - 1)
|
||||||
@@ -5140,23 +5172,17 @@ void bssn_class::Step(int lev, int YN)
|
|||||||
|
|
||||||
// misc::tillherecheck(GH->Commlev[lev],GH->start_rank[lev],"before Corrector error check");
|
// misc::tillherecheck(GH->Commlev[lev],GH->start_rank[lev],"before Corrector error check");
|
||||||
|
|
||||||
// Non-blocking error reduction overlapped with Sync to hide Allreduce latency
|
// Start async ghost zone exchange - overlaps with error check and BH position
|
||||||
MPI_Request err_req_cor;
|
Parallel::SyncHandle *sync_cor = Parallel::SyncBeginWithPlan(sync_plan, SynchList_cor);
|
||||||
|
|
||||||
|
// check error information (overlaps with MPI transfer)
|
||||||
{
|
{
|
||||||
int erh = ERROR;
|
int erh = ERROR;
|
||||||
MPI_Iallreduce(&erh, &ERROR, 1, MPI_INT, MPI_SUM, GH->Commlev[lev], &err_req_cor);
|
MPI_Allreduce(&erh, &ERROR, 1, MPI_INT, MPI_SUM, GH->Commlev[lev]);
|
||||||
}
|
}
|
||||||
|
|
||||||
// misc::tillherecheck(GH->Commlev[lev],GH->start_rank[lev],"before Corrector sync");
|
|
||||||
|
|
||||||
Parallel::Sync_cached(GH->PatL[lev], SynchList_cor, Symmetry, sync_cache_cor[lev]);
|
|
||||||
|
|
||||||
// misc::tillherecheck(GH->Commlev[lev],GH->start_rank[lev],"after Corrector sync");
|
|
||||||
|
|
||||||
// Complete non-blocking error reduction and check
|
|
||||||
MPI_Wait(&err_req_cor, MPI_STATUS_IGNORE);
|
|
||||||
if (ERROR)
|
if (ERROR)
|
||||||
{
|
{
|
||||||
|
Parallel::SyncEnd(sync_cor); sync_cor = 0;
|
||||||
Parallel::Dump_Data(GH->PatL[lev], SynchList_pre, 0, PhysTime, dT_lev);
|
Parallel::Dump_Data(GH->PatL[lev], SynchList_pre, 0, PhysTime, dT_lev);
|
||||||
if (myrank == 0)
|
if (myrank == 0)
|
||||||
{
|
{
|
||||||
@@ -5168,6 +5194,13 @@ void bssn_class::Step(int lev, int YN)
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// misc::tillherecheck(GH->Commlev[lev],GH->start_rank[lev],"before Corrector sync");
|
||||||
|
|
||||||
|
// Complete async ghost zone exchange
|
||||||
|
if (sync_cor) Parallel::SyncEnd(sync_cor);
|
||||||
|
|
||||||
|
// misc::tillherecheck(GH->Commlev[lev],GH->start_rank[lev],"after Corrector sync");
|
||||||
|
|
||||||
#if (MAPBH == 0)
|
#if (MAPBH == 0)
|
||||||
// for black hole position
|
// for black hole position
|
||||||
if (BH_num > 0 && lev == GH->levels - 1)
|
if (BH_num > 0 && lev == GH->levels - 1)
|
||||||
@@ -5280,6 +5313,8 @@ void bssn_class::Step(int lev, int YN)
|
|||||||
|
|
||||||
// if(myrank==GH->start_rank[lev]) cout<<GH->mylev<<endl;
|
// if(myrank==GH->start_rank[lev]) cout<<GH->mylev<<endl;
|
||||||
// misc::tillherecheck(GH->Commlev[lev],GH->start_rank[lev],"complet GH Step");
|
// misc::tillherecheck(GH->Commlev[lev],GH->start_rank[lev],"complet GH Step");
|
||||||
|
|
||||||
|
Parallel::SyncFreePlan(sync_plan);
|
||||||
}
|
}
|
||||||
|
|
||||||
//================================================================================================
|
//================================================================================================
|
||||||
@@ -5451,11 +5486,21 @@ void bssn_class::SHStep()
|
|||||||
#if (PSTR == 1 || PSTR == 2)
|
#if (PSTR == 1 || PSTR == 2)
|
||||||
// misc::tillherecheck(GH->Commlev[lev],GH->start_rank[lev],"before Predictor's error check");
|
// misc::tillherecheck(GH->Commlev[lev],GH->start_rank[lev],"before Predictor's error check");
|
||||||
#endif
|
#endif
|
||||||
// Non-blocking error reduction overlapped with Synch to hide Allreduce latency
|
// check error information
|
||||||
MPI_Request err_req;
|
|
||||||
{
|
{
|
||||||
int erh = ERROR;
|
int erh = ERROR;
|
||||||
MPI_Iallreduce(&erh, &ERROR, 1, MPI_INT, MPI_SUM, MPI_COMM_WORLD, &err_req);
|
MPI_Allreduce(&erh, &ERROR, 1, MPI_INT, MPI_SUM, MPI_COMM_WORLD);
|
||||||
|
}
|
||||||
|
|
||||||
|
if (ERROR)
|
||||||
|
{
|
||||||
|
SH->Dump_Data(StateList, 0, PhysTime, dT_lev);
|
||||||
|
if (myrank == 0)
|
||||||
|
{
|
||||||
|
if (ErrorMonitor->outfile)
|
||||||
|
ErrorMonitor->outfile << "find NaN in state variables on Shell Patches at t = " << PhysTime << endl;
|
||||||
|
MPI_Abort(MPI_COMM_WORLD, 1);
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
{
|
{
|
||||||
@@ -5473,19 +5518,6 @@ void bssn_class::SHStep()
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// Complete non-blocking error reduction and check
|
|
||||||
MPI_Wait(&err_req, MPI_STATUS_IGNORE);
|
|
||||||
if (ERROR)
|
|
||||||
{
|
|
||||||
SH->Dump_Data(StateList, 0, PhysTime, dT_lev);
|
|
||||||
if (myrank == 0)
|
|
||||||
{
|
|
||||||
if (ErrorMonitor->outfile)
|
|
||||||
ErrorMonitor->outfile << "find NaN in state variables on Shell Patches at t = " << PhysTime << endl;
|
|
||||||
MPI_Abort(MPI_COMM_WORLD, 1);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// corrector
|
// corrector
|
||||||
for (iter_count = 1; iter_count < 4; iter_count++)
|
for (iter_count = 1; iter_count < 4; iter_count++)
|
||||||
{
|
{
|
||||||
@@ -5628,11 +5660,21 @@ void bssn_class::SHStep()
|
|||||||
sPp = sPp->next;
|
sPp = sPp->next;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
// Non-blocking error reduction overlapped with Synch to hide Allreduce latency
|
// check error information
|
||||||
MPI_Request err_req_cor;
|
|
||||||
{
|
{
|
||||||
int erh = ERROR;
|
int erh = ERROR;
|
||||||
MPI_Iallreduce(&erh, &ERROR, 1, MPI_INT, MPI_SUM, MPI_COMM_WORLD, &err_req_cor);
|
MPI_Allreduce(&erh, &ERROR, 1, MPI_INT, MPI_SUM, MPI_COMM_WORLD);
|
||||||
|
}
|
||||||
|
if (ERROR)
|
||||||
|
{
|
||||||
|
SH->Dump_Data(SynchList_pre, 0, PhysTime, dT_lev);
|
||||||
|
if (myrank == 0)
|
||||||
|
{
|
||||||
|
if (ErrorMonitor->outfile)
|
||||||
|
ErrorMonitor->outfile << "find NaN on Shell Patches in RK4 substep#" << iter_count
|
||||||
|
<< " variables at t = " << PhysTime << endl;
|
||||||
|
MPI_Abort(MPI_COMM_WORLD, 1);
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
{
|
{
|
||||||
@@ -5650,20 +5692,6 @@ void bssn_class::SHStep()
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// Complete non-blocking error reduction and check
|
|
||||||
MPI_Wait(&err_req_cor, MPI_STATUS_IGNORE);
|
|
||||||
if (ERROR)
|
|
||||||
{
|
|
||||||
SH->Dump_Data(SynchList_pre, 0, PhysTime, dT_lev);
|
|
||||||
if (myrank == 0)
|
|
||||||
{
|
|
||||||
if (ErrorMonitor->outfile)
|
|
||||||
ErrorMonitor->outfile << "find NaN on Shell Patches in RK4 substep#" << iter_count
|
|
||||||
<< " variables at t = " << PhysTime << endl;
|
|
||||||
MPI_Abort(MPI_COMM_WORLD, 1);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
sPp = SH->PatL;
|
sPp = SH->PatL;
|
||||||
while (sPp)
|
while (sPp)
|
||||||
{
|
{
|
||||||
@@ -5792,7 +5820,7 @@ void bssn_class::RestrictProlong(int lev, int YN, bool BB,
|
|||||||
// misc::tillherecheck(GH->Commlev[GH->mylev],GH->start_rank[GH->mylev],a_stream.str());
|
// misc::tillherecheck(GH->Commlev[GH->mylev],GH->start_rank[GH->mylev],a_stream.str());
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
Parallel::Sync_cached(GH->PatL[lev - 1], SynchList_pre, Symmetry, sync_cache_rp_coarse[lev]);
|
Parallel::Sync(GH->PatL[lev - 1], SynchList_pre, Symmetry);
|
||||||
|
|
||||||
#if (PSTR == 1 || PSTR == 2)
|
#if (PSTR == 1 || PSTR == 2)
|
||||||
// a_stream.clear();
|
// a_stream.clear();
|
||||||
@@ -5802,11 +5830,21 @@ void bssn_class::RestrictProlong(int lev, int YN, bool BB,
|
|||||||
#endif
|
#endif
|
||||||
|
|
||||||
#if (RPB == 0)
|
#if (RPB == 0)
|
||||||
|
Ppc = GH->PatL[lev - 1];
|
||||||
|
while (Ppc)
|
||||||
|
{
|
||||||
|
Pp = GH->PatL[lev];
|
||||||
|
while (Pp)
|
||||||
|
{
|
||||||
#if (MIXOUTB == 0)
|
#if (MIXOUTB == 0)
|
||||||
Parallel::OutBdLow2Hi(GH->PatL[lev - 1], GH->PatL[lev], SynchList_pre, SL, Symmetry);
|
Parallel::OutBdLow2Hi(Ppc->data, Pp->data, SynchList_pre, SL, Symmetry);
|
||||||
#elif (MIXOUTB == 1)
|
#elif (MIXOUTB == 1)
|
||||||
Parallel::OutBdLow2Himix(GH->PatL[lev - 1], GH->PatL[lev], SynchList_pre, SL, Symmetry);
|
Parallel::OutBdLow2Himix(Ppc->data, Pp->data, SynchList_pre, SL, Symmetry);
|
||||||
#endif
|
#endif
|
||||||
|
Pp = Pp->next;
|
||||||
|
}
|
||||||
|
Ppc = Ppc->next;
|
||||||
|
}
|
||||||
#elif (RPB == 1)
|
#elif (RPB == 1)
|
||||||
// Parallel::OutBdLow2Hi_bam(GH->PatL[lev-1],GH->PatL[lev],SynchList_pre,SL,Symmetry);
|
// Parallel::OutBdLow2Hi_bam(GH->PatL[lev-1],GH->PatL[lev],SynchList_pre,SL,Symmetry);
|
||||||
Parallel::OutBdLow2Hi_bam(GH->PatL[lev - 1], GH->PatL[lev], SynchList_pre, SL, GH->bdsul[lev], Symmetry);
|
Parallel::OutBdLow2Hi_bam(GH->PatL[lev - 1], GH->PatL[lev], SynchList_pre, SL, GH->bdsul[lev], Symmetry);
|
||||||
@@ -5843,7 +5881,7 @@ void bssn_class::RestrictProlong(int lev, int YN, bool BB,
|
|||||||
// misc::tillherecheck(GH->Commlev[GH->mylev],GH->start_rank[GH->mylev],a_stream.str());
|
// misc::tillherecheck(GH->Commlev[GH->mylev],GH->start_rank[GH->mylev],a_stream.str());
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
Parallel::Sync_cached(GH->PatL[lev - 1], SL, Symmetry, sync_cache_rp_coarse[lev]);
|
Parallel::Sync(GH->PatL[lev - 1], SL, Symmetry);
|
||||||
|
|
||||||
#if (PSTR == 1 || PSTR == 2)
|
#if (PSTR == 1 || PSTR == 2)
|
||||||
// a_stream.clear();
|
// a_stream.clear();
|
||||||
@@ -5853,11 +5891,21 @@ void bssn_class::RestrictProlong(int lev, int YN, bool BB,
|
|||||||
#endif
|
#endif
|
||||||
|
|
||||||
#if (RPB == 0)
|
#if (RPB == 0)
|
||||||
|
Ppc = GH->PatL[lev - 1];
|
||||||
|
while (Ppc)
|
||||||
|
{
|
||||||
|
Pp = GH->PatL[lev];
|
||||||
|
while (Pp)
|
||||||
|
{
|
||||||
#if (MIXOUTB == 0)
|
#if (MIXOUTB == 0)
|
||||||
Parallel::OutBdLow2Hi(GH->PatL[lev - 1], GH->PatL[lev], SL, SL, Symmetry);
|
Parallel::OutBdLow2Hi(Ppc->data, Pp->data, SL, SL, Symmetry);
|
||||||
#elif (MIXOUTB == 1)
|
#elif (MIXOUTB == 1)
|
||||||
Parallel::OutBdLow2Himix(GH->PatL[lev - 1], GH->PatL[lev], SL, SL, Symmetry);
|
Parallel::OutBdLow2Himix(Ppc->data, Pp->data, SL, SL, Symmetry);
|
||||||
#endif
|
#endif
|
||||||
|
Pp = Pp->next;
|
||||||
|
}
|
||||||
|
Ppc = Ppc->next;
|
||||||
|
}
|
||||||
#elif (RPB == 1)
|
#elif (RPB == 1)
|
||||||
// Parallel::OutBdLow2Hi_bam(GH->PatL[lev-1],GH->PatL[lev],SL,SL,Symmetry);
|
// Parallel::OutBdLow2Hi_bam(GH->PatL[lev-1],GH->PatL[lev],SL,SL,Symmetry);
|
||||||
Parallel::OutBdLow2Hi_bam(GH->PatL[lev - 1], GH->PatL[lev], SL, SL, GH->bdsul[lev], Symmetry);
|
Parallel::OutBdLow2Hi_bam(GH->PatL[lev - 1], GH->PatL[lev], SL, SL, GH->bdsul[lev], Symmetry);
|
||||||
@@ -5871,7 +5919,7 @@ void bssn_class::RestrictProlong(int lev, int YN, bool BB,
|
|||||||
#endif
|
#endif
|
||||||
}
|
}
|
||||||
|
|
||||||
Parallel::Sync_cached(GH->PatL[lev], SL, Symmetry, sync_cache_rp_fine[lev]);
|
Parallel::Sync(GH->PatL[lev], SL, Symmetry);
|
||||||
|
|
||||||
#if (PSTR == 1 || PSTR == 2)
|
#if (PSTR == 1 || PSTR == 2)
|
||||||
// a_stream.clear();
|
// a_stream.clear();
|
||||||
@@ -5929,14 +5977,24 @@ void bssn_class::RestrictProlong_aux(int lev, int YN, bool BB,
|
|||||||
Parallel::Restrict_bam(GH->PatL[lev - 1], GH->PatL[lev], SL, SynchList_pre, GH->rsul[lev], Symmetry);
|
Parallel::Restrict_bam(GH->PatL[lev - 1], GH->PatL[lev], SL, SynchList_pre, GH->rsul[lev], Symmetry);
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
Parallel::Sync_cached(GH->PatL[lev - 1], SynchList_pre, Symmetry, sync_cache_rp_coarse[lev]);
|
Parallel::Sync(GH->PatL[lev - 1], SynchList_pre, Symmetry);
|
||||||
|
|
||||||
#if (RPB == 0)
|
#if (RPB == 0)
|
||||||
|
Ppc = GH->PatL[lev - 1];
|
||||||
|
while (Ppc)
|
||||||
|
{
|
||||||
|
Pp = GH->PatL[lev];
|
||||||
|
while (Pp)
|
||||||
|
{
|
||||||
#if (MIXOUTB == 0)
|
#if (MIXOUTB == 0)
|
||||||
Parallel::OutBdLow2Hi(GH->PatL[lev - 1], GH->PatL[lev], SynchList_pre, SL, Symmetry);
|
Parallel::OutBdLow2Hi(Ppc->data, Pp->data, SynchList_pre, SL, Symmetry);
|
||||||
#elif (MIXOUTB == 1)
|
#elif (MIXOUTB == 1)
|
||||||
Parallel::OutBdLow2Himix(GH->PatL[lev - 1], GH->PatL[lev], SynchList_pre, SL, Symmetry);
|
Parallel::OutBdLow2Himix(Ppc->data, Pp->data, SynchList_pre, SL, Symmetry);
|
||||||
#endif
|
#endif
|
||||||
|
Pp = Pp->next;
|
||||||
|
}
|
||||||
|
Ppc = Ppc->next;
|
||||||
|
}
|
||||||
#elif (RPB == 1)
|
#elif (RPB == 1)
|
||||||
// Parallel::OutBdLow2Hi_bam(GH->PatL[lev-1],GH->PatL[lev],SynchList_pre,SL,Symmetry);
|
// Parallel::OutBdLow2Hi_bam(GH->PatL[lev-1],GH->PatL[lev],SynchList_pre,SL,Symmetry);
|
||||||
Parallel::OutBdLow2Hi_bam(GH->PatL[lev - 1], GH->PatL[lev], SynchList_pre, SL, GH->bdsul[lev], Symmetry);
|
Parallel::OutBdLow2Hi_bam(GH->PatL[lev - 1], GH->PatL[lev], SynchList_pre, SL, GH->bdsul[lev], Symmetry);
|
||||||
@@ -5951,21 +6009,31 @@ void bssn_class::RestrictProlong_aux(int lev, int YN, bool BB,
|
|||||||
Parallel::Restrict_bam(GH->PatL[lev - 1], GH->PatL[lev], SL, SL, GH->rsul[lev], Symmetry);
|
Parallel::Restrict_bam(GH->PatL[lev - 1], GH->PatL[lev], SL, SL, GH->rsul[lev], Symmetry);
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
Parallel::Sync_cached(GH->PatL[lev - 1], SL, Symmetry, sync_cache_rp_coarse[lev]);
|
Parallel::Sync(GH->PatL[lev - 1], SL, Symmetry);
|
||||||
|
|
||||||
#if (RPB == 0)
|
#if (RPB == 0)
|
||||||
|
Ppc = GH->PatL[lev - 1];
|
||||||
|
while (Ppc)
|
||||||
|
{
|
||||||
|
Pp = GH->PatL[lev];
|
||||||
|
while (Pp)
|
||||||
|
{
|
||||||
#if (MIXOUTB == 0)
|
#if (MIXOUTB == 0)
|
||||||
Parallel::OutBdLow2Hi(GH->PatL[lev - 1], GH->PatL[lev], SL, SL, Symmetry);
|
Parallel::OutBdLow2Hi(Ppc->data, Pp->data, SL, SL, Symmetry);
|
||||||
#elif (MIXOUTB == 1)
|
#elif (MIXOUTB == 1)
|
||||||
Parallel::OutBdLow2Himix(GH->PatL[lev - 1], GH->PatL[lev], SL, SL, Symmetry);
|
Parallel::OutBdLow2Himix(Ppc->data, Pp->data, SL, SL, Symmetry);
|
||||||
#endif
|
#endif
|
||||||
|
Pp = Pp->next;
|
||||||
|
}
|
||||||
|
Ppc = Ppc->next;
|
||||||
|
}
|
||||||
#elif (RPB == 1)
|
#elif (RPB == 1)
|
||||||
// Parallel::OutBdLow2Hi_bam(GH->PatL[lev-1],GH->PatL[lev],SL,SL,Symmetry);
|
// Parallel::OutBdLow2Hi_bam(GH->PatL[lev-1],GH->PatL[lev],SL,SL,Symmetry);
|
||||||
Parallel::OutBdLow2Hi_bam(GH->PatL[lev - 1], GH->PatL[lev], SL, SL, GH->bdsul[lev], Symmetry);
|
Parallel::OutBdLow2Hi_bam(GH->PatL[lev - 1], GH->PatL[lev], SL, SL, GH->bdsul[lev], Symmetry);
|
||||||
#endif
|
#endif
|
||||||
}
|
}
|
||||||
|
|
||||||
Parallel::Sync_cached(GH->PatL[lev], SL, Symmetry, sync_cache_rp_fine[lev]);
|
Parallel::Sync(GH->PatL[lev], SL, Symmetry);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -6016,14 +6084,24 @@ void bssn_class::RestrictProlong(int lev, int YN, bool BB)
|
|||||||
Parallel::Restrict_bam(GH->PatL[lev - 1], GH->PatL[lev], SynchList_cor, SynchList_pre, GH->rsul[lev], Symmetry);
|
Parallel::Restrict_bam(GH->PatL[lev - 1], GH->PatL[lev], SynchList_cor, SynchList_pre, GH->rsul[lev], Symmetry);
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
Parallel::Sync_cached(GH->PatL[lev - 1], SynchList_pre, Symmetry, sync_cache_rp_coarse[lev]);
|
Parallel::Sync(GH->PatL[lev - 1], SynchList_pre, Symmetry);
|
||||||
|
|
||||||
#if (RPB == 0)
|
#if (RPB == 0)
|
||||||
|
Ppc = GH->PatL[lev - 1];
|
||||||
|
while (Ppc)
|
||||||
|
{
|
||||||
|
Pp = GH->PatL[lev];
|
||||||
|
while (Pp)
|
||||||
|
{
|
||||||
#if (MIXOUTB == 0)
|
#if (MIXOUTB == 0)
|
||||||
Parallel::OutBdLow2Hi(GH->PatL[lev - 1], GH->PatL[lev], SynchList_pre, SynchList_cor, Symmetry);
|
Parallel::OutBdLow2Hi(Ppc->data, Pp->data, SynchList_pre, SynchList_cor, Symmetry);
|
||||||
#elif (MIXOUTB == 1)
|
#elif (MIXOUTB == 1)
|
||||||
Parallel::OutBdLow2Himix(GH->PatL[lev - 1], GH->PatL[lev], SynchList_pre, SynchList_cor, Symmetry);
|
Parallel::OutBdLow2Himix(Ppc->data, Pp->data, SynchList_pre, SynchList_cor, Symmetry);
|
||||||
#endif
|
#endif
|
||||||
|
Pp = Pp->next;
|
||||||
|
}
|
||||||
|
Ppc = Ppc->next;
|
||||||
|
}
|
||||||
#elif (RPB == 1)
|
#elif (RPB == 1)
|
||||||
// Parallel::OutBdLow2Hi_bam(GH->PatL[lev-1],GH->PatL[lev],SynchList_pre,SynchList_cor,Symmetry);
|
// Parallel::OutBdLow2Hi_bam(GH->PatL[lev-1],GH->PatL[lev],SynchList_pre,SynchList_cor,Symmetry);
|
||||||
Parallel::OutBdLow2Hi_bam(GH->PatL[lev - 1], GH->PatL[lev], SynchList_pre, SynchList_cor, GH->bdsul[lev], Symmetry);
|
Parallel::OutBdLow2Hi_bam(GH->PatL[lev - 1], GH->PatL[lev], SynchList_pre, SynchList_cor, GH->bdsul[lev], Symmetry);
|
||||||
@@ -6040,21 +6118,31 @@ void bssn_class::RestrictProlong(int lev, int YN, bool BB)
|
|||||||
Parallel::Restrict_bam(GH->PatL[lev - 1], GH->PatL[lev], SynchList_cor, StateList, GH->rsul[lev], Symmetry);
|
Parallel::Restrict_bam(GH->PatL[lev - 1], GH->PatL[lev], SynchList_cor, StateList, GH->rsul[lev], Symmetry);
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
Parallel::Sync_cached(GH->PatL[lev - 1], StateList, Symmetry, sync_cache_rp_coarse[lev]);
|
Parallel::Sync(GH->PatL[lev - 1], StateList, Symmetry);
|
||||||
|
|
||||||
#if (RPB == 0)
|
#if (RPB == 0)
|
||||||
|
Ppc = GH->PatL[lev - 1];
|
||||||
|
while (Ppc)
|
||||||
|
{
|
||||||
|
Pp = GH->PatL[lev];
|
||||||
|
while (Pp)
|
||||||
|
{
|
||||||
#if (MIXOUTB == 0)
|
#if (MIXOUTB == 0)
|
||||||
Parallel::OutBdLow2Hi(GH->PatL[lev - 1], GH->PatL[lev], StateList, SynchList_cor, Symmetry);
|
Parallel::OutBdLow2Hi(Ppc->data, Pp->data, StateList, SynchList_cor, Symmetry);
|
||||||
#elif (MIXOUTB == 1)
|
#elif (MIXOUTB == 1)
|
||||||
Parallel::OutBdLow2Himix(GH->PatL[lev - 1], GH->PatL[lev], StateList, SynchList_cor, Symmetry);
|
Parallel::OutBdLow2Himix(Ppc->data, Pp->data, StateList, SynchList_cor, Symmetry);
|
||||||
#endif
|
#endif
|
||||||
|
Pp = Pp->next;
|
||||||
|
}
|
||||||
|
Ppc = Ppc->next;
|
||||||
|
}
|
||||||
#elif (RPB == 1)
|
#elif (RPB == 1)
|
||||||
// Parallel::OutBdLow2Hi_bam(GH->PatL[lev-1],GH->PatL[lev],StateList,SynchList_cor,Symmetry);
|
// Parallel::OutBdLow2Hi_bam(GH->PatL[lev-1],GH->PatL[lev],StateList,SynchList_cor,Symmetry);
|
||||||
Parallel::OutBdLow2Hi_bam(GH->PatL[lev - 1], GH->PatL[lev], StateList, SynchList_cor, GH->bdsul[lev], Symmetry);
|
Parallel::OutBdLow2Hi_bam(GH->PatL[lev - 1], GH->PatL[lev], StateList, SynchList_cor, GH->bdsul[lev], Symmetry);
|
||||||
#endif
|
#endif
|
||||||
}
|
}
|
||||||
|
|
||||||
Parallel::Sync_cached(GH->PatL[lev], SynchList_cor, Symmetry, sync_cache_rp_fine[lev]);
|
Parallel::Sync(GH->PatL[lev], SynchList_cor, Symmetry);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -6084,11 +6172,21 @@ void bssn_class::ProlongRestrict(int lev, int YN, bool BB)
|
|||||||
}
|
}
|
||||||
|
|
||||||
#if (RPB == 0)
|
#if (RPB == 0)
|
||||||
|
Ppc = GH->PatL[lev - 1];
|
||||||
|
while (Ppc)
|
||||||
|
{
|
||||||
|
Pp = GH->PatL[lev];
|
||||||
|
while (Pp)
|
||||||
|
{
|
||||||
#if (MIXOUTB == 0)
|
#if (MIXOUTB == 0)
|
||||||
Parallel::OutBdLow2Hi(GH->PatL[lev - 1], GH->PatL[lev], SynchList_pre, SynchList_cor, Symmetry);
|
Parallel::OutBdLow2Hi(Ppc->data, Pp->data, SynchList_pre, SynchList_cor, Symmetry);
|
||||||
#elif (MIXOUTB == 1)
|
#elif (MIXOUTB == 1)
|
||||||
Parallel::OutBdLow2Himix(GH->PatL[lev - 1], GH->PatL[lev], SynchList_pre, SynchList_cor, Symmetry);
|
Parallel::OutBdLow2Himix(Ppc->data, Pp->data, SynchList_pre, SynchList_cor, Symmetry);
|
||||||
#endif
|
#endif
|
||||||
|
Pp = Pp->next;
|
||||||
|
}
|
||||||
|
Ppc = Ppc->next;
|
||||||
|
}
|
||||||
#elif (RPB == 1)
|
#elif (RPB == 1)
|
||||||
// Parallel::OutBdLow2Hi_bam(GH->PatL[lev-1],GH->PatL[lev],SynchList_pre,SynchList_cor,Symmetry);
|
// Parallel::OutBdLow2Hi_bam(GH->PatL[lev-1],GH->PatL[lev],SynchList_pre,SynchList_cor,Symmetry);
|
||||||
Parallel::OutBdLow2Hi_bam(GH->PatL[lev - 1], GH->PatL[lev], SynchList_pre, SynchList_cor, GH->bdsul[lev], Symmetry);
|
Parallel::OutBdLow2Hi_bam(GH->PatL[lev - 1], GH->PatL[lev], SynchList_pre, SynchList_cor, GH->bdsul[lev], Symmetry);
|
||||||
@@ -6097,11 +6195,21 @@ void bssn_class::ProlongRestrict(int lev, int YN, bool BB)
|
|||||||
else // no time refinement levels and for all same time levels
|
else // no time refinement levels and for all same time levels
|
||||||
{
|
{
|
||||||
#if (RPB == 0)
|
#if (RPB == 0)
|
||||||
|
Ppc = GH->PatL[lev - 1];
|
||||||
|
while (Ppc)
|
||||||
|
{
|
||||||
|
Pp = GH->PatL[lev];
|
||||||
|
while (Pp)
|
||||||
|
{
|
||||||
#if (MIXOUTB == 0)
|
#if (MIXOUTB == 0)
|
||||||
Parallel::OutBdLow2Hi(GH->PatL[lev - 1], GH->PatL[lev], StateList, SynchList_cor, Symmetry);
|
Parallel::OutBdLow2Hi(Ppc->data, Pp->data, StateList, SynchList_cor, Symmetry);
|
||||||
#elif (MIXOUTB == 1)
|
#elif (MIXOUTB == 1)
|
||||||
Parallel::OutBdLow2Himix(GH->PatL[lev - 1], GH->PatL[lev], StateList, SynchList_cor, Symmetry);
|
Parallel::OutBdLow2Himix(Ppc->data, Pp->data, StateList, SynchList_cor, Symmetry);
|
||||||
#endif
|
#endif
|
||||||
|
Pp = Pp->next;
|
||||||
|
}
|
||||||
|
Ppc = Ppc->next;
|
||||||
|
}
|
||||||
#elif (RPB == 1)
|
#elif (RPB == 1)
|
||||||
// Parallel::OutBdLow2Hi_bam(GH->PatL[lev-1],GH->PatL[lev],StateList,SynchList_cor,Symmetry);
|
// Parallel::OutBdLow2Hi_bam(GH->PatL[lev-1],GH->PatL[lev],StateList,SynchList_cor,Symmetry);
|
||||||
Parallel::OutBdLow2Hi_bam(GH->PatL[lev - 1], GH->PatL[lev], StateList, SynchList_cor, GH->bdsul[lev], Symmetry);
|
Parallel::OutBdLow2Hi_bam(GH->PatL[lev - 1], GH->PatL[lev], StateList, SynchList_cor, GH->bdsul[lev], Symmetry);
|
||||||
@@ -6117,10 +6225,10 @@ void bssn_class::ProlongRestrict(int lev, int YN, bool BB)
|
|||||||
#else
|
#else
|
||||||
Parallel::Restrict_after(GH->PatL[lev - 1], GH->PatL[lev], SynchList_cor, StateList, Symmetry);
|
Parallel::Restrict_after(GH->PatL[lev - 1], GH->PatL[lev], SynchList_cor, StateList, Symmetry);
|
||||||
#endif
|
#endif
|
||||||
Parallel::Sync_cached(GH->PatL[lev - 1], StateList, Symmetry, sync_cache_rp_coarse[lev]);
|
Parallel::Sync(GH->PatL[lev - 1], StateList, Symmetry);
|
||||||
}
|
}
|
||||||
|
|
||||||
Parallel::Sync_cached(GH->PatL[lev], SynchList_cor, Symmetry, sync_cache_rp_fine[lev]);
|
Parallel::Sync(GH->PatL[lev], SynchList_cor, Symmetry);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
#undef MIXOUTB
|
#undef MIXOUTB
|
||||||
|
|||||||
@@ -126,11 +126,6 @@ public:
|
|||||||
MyList<var> *OldStateList, *DumpList;
|
MyList<var> *OldStateList, *DumpList;
|
||||||
MyList<var> *ConstraintList;
|
MyList<var> *ConstraintList;
|
||||||
|
|
||||||
Parallel::SyncCache *sync_cache_pre; // per-level cache for predictor sync
|
|
||||||
Parallel::SyncCache *sync_cache_cor; // per-level cache for corrector sync
|
|
||||||
Parallel::SyncCache *sync_cache_rp_coarse; // RestrictProlong sync on PatL[lev-1]
|
|
||||||
Parallel::SyncCache *sync_cache_rp_fine; // RestrictProlong sync on PatL[lev]
|
|
||||||
|
|
||||||
monitor *ErrorMonitor, *Psi4Monitor, *BHMonitor, *MAPMonitor;
|
monitor *ErrorMonitor, *Psi4Monitor, *BHMonitor, *MAPMonitor;
|
||||||
monitor *ConVMonitor;
|
monitor *ConVMonitor;
|
||||||
surface_integral *Waveshell;
|
surface_integral *Waveshell;
|
||||||
|
|||||||
File diff suppressed because it is too large
Load Diff
@@ -1301,13 +1301,13 @@ bool cgh::Interp_One_Point(MyList<var> *VarList,
|
|||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
bool cgh::Regrid_Onelevel(int lev, int Symmetry, int BH_num, double **Porgbr, double **Porg0,
|
void cgh::Regrid_Onelevel(int lev, int Symmetry, int BH_num, double **Porgbr, double **Porg0,
|
||||||
MyList<var> *OldList, MyList<var> *StateList,
|
MyList<var> *OldList, MyList<var> *StateList,
|
||||||
MyList<var> *FutureList, MyList<var> *tmList, bool BB,
|
MyList<var> *FutureList, MyList<var> *tmList, bool BB,
|
||||||
monitor *ErrorMonitor)
|
monitor *ErrorMonitor)
|
||||||
{
|
{
|
||||||
if (lev < movls)
|
if (lev < movls)
|
||||||
return false;
|
return;
|
||||||
|
|
||||||
#if (0)
|
#if (0)
|
||||||
// #if (PSTR == 1 || PSTR == 2)
|
// #if (PSTR == 1 || PSTR == 2)
|
||||||
@@ -1396,7 +1396,7 @@ bool cgh::Regrid_Onelevel(int lev, int Symmetry, int BH_num, double **Porgbr, do
|
|||||||
for (bhi = 0; bhi < BH_num; bhi++)
|
for (bhi = 0; bhi < BH_num; bhi++)
|
||||||
delete[] tmpPorg[bhi];
|
delete[] tmpPorg[bhi];
|
||||||
delete[] tmpPorg;
|
delete[] tmpPorg;
|
||||||
return false;
|
return;
|
||||||
}
|
}
|
||||||
// x direction
|
// x direction
|
||||||
rr = (Porg0[bhi][0] - handle[lev][grd][0]) / dX;
|
rr = (Porg0[bhi][0] - handle[lev][grd][0]) / dX;
|
||||||
@@ -1500,7 +1500,6 @@ bool cgh::Regrid_Onelevel(int lev, int Symmetry, int BH_num, double **Porgbr, do
|
|||||||
for (int bhi = 0; bhi < BH_num; bhi++)
|
for (int bhi = 0; bhi < BH_num; bhi++)
|
||||||
delete[] tmpPorg[bhi];
|
delete[] tmpPorg[bhi];
|
||||||
delete[] tmpPorg;
|
delete[] tmpPorg;
|
||||||
return tot_flag;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
|
|||||||
@@ -74,7 +74,7 @@ public:
|
|||||||
MyList<var> *OldList, MyList<var> *StateList,
|
MyList<var> *OldList, MyList<var> *StateList,
|
||||||
MyList<var> *FutureList, MyList<var> *tmList,
|
MyList<var> *FutureList, MyList<var> *tmList,
|
||||||
int Symmetry, bool BB);
|
int Symmetry, bool BB);
|
||||||
bool Regrid_Onelevel(int lev, int Symmetry, int BH_num, double **Porgbr, double **Porg0,
|
void Regrid_Onelevel(int lev, int Symmetry, int BH_num, double **Porgbr, double **Porg0,
|
||||||
MyList<var> *OldList, MyList<var> *StateList,
|
MyList<var> *OldList, MyList<var> *StateList,
|
||||||
MyList<var> *FutureList, MyList<var> *tmList, bool BB,
|
MyList<var> *FutureList, MyList<var> *tmList, bool BB,
|
||||||
monitor *ErrorMonitor);
|
monitor *ErrorMonitor);
|
||||||
|
|||||||
@@ -69,8 +69,6 @@
|
|||||||
fy = ZEO
|
fy = ZEO
|
||||||
fz = ZEO
|
fz = ZEO
|
||||||
|
|
||||||
!DIR$ SIMD VECTORLENGTHFOR(KNOWN_INTEGER=8)
|
|
||||||
!DIR$ UNROLL PARTIAL(4)
|
|
||||||
do k=1,ex(3)-1
|
do k=1,ex(3)-1
|
||||||
do j=1,ex(2)-1
|
do j=1,ex(2)-1
|
||||||
do i=1,ex(1)-1
|
do i=1,ex(1)-1
|
||||||
@@ -373,8 +371,6 @@
|
|||||||
fxz = ZEO
|
fxz = ZEO
|
||||||
fyz = ZEO
|
fyz = ZEO
|
||||||
|
|
||||||
!DIR$ SIMD VECTORLENGTHFOR(KNOWN_INTEGER=8)
|
|
||||||
!DIR$ UNROLL PARTIAL(4)
|
|
||||||
do k=1,ex(3)-1
|
do k=1,ex(3)-1
|
||||||
do j=1,ex(2)-1
|
do j=1,ex(2)-1
|
||||||
do i=1,ex(1)-1
|
do i=1,ex(1)-1
|
||||||
|
|||||||
@@ -1,268 +0,0 @@
|
|||||||
#include "tool.h"
|
|
||||||
void fdderivs(const int ex[3],
|
|
||||||
const double *f,
|
|
||||||
double *fxx, double *fxy, double *fxz,
|
|
||||||
double *fyy, double *fyz, double *fzz,
|
|
||||||
const double *X, const double *Y, const double *Z,
|
|
||||||
double SYM1, double SYM2, double SYM3,
|
|
||||||
int Symmetry, int onoff)
|
|
||||||
{
|
|
||||||
(void)onoff;
|
|
||||||
|
|
||||||
const int NO_SYMM = 0, EQ_SYMM = 1;
|
|
||||||
const double ZEO = 0.0, ONE = 1.0, TWO = 2.0;
|
|
||||||
const double F1o4 = 2.5e-1; // 1/4
|
|
||||||
const double F8 = 8.0;
|
|
||||||
const double F16 = 16.0;
|
|
||||||
const double F30 = 30.0;
|
|
||||||
const double F1o12 = ONE / 12.0;
|
|
||||||
const double F1o144 = ONE / 144.0;
|
|
||||||
|
|
||||||
const int ex1 = ex[0], ex2 = ex[1], ex3 = ex[2];
|
|
||||||
|
|
||||||
const double dX = X[1] - X[0];
|
|
||||||
const double dY = Y[1] - Y[0];
|
|
||||||
const double dZ = Z[1] - Z[0];
|
|
||||||
|
|
||||||
const int imaxF = ex1;
|
|
||||||
const int jmaxF = ex2;
|
|
||||||
const int kmaxF = ex3;
|
|
||||||
|
|
||||||
int iminF = 1, jminF = 1, kminF = 1;
|
|
||||||
if (Symmetry > NO_SYMM && fabs(Z[0]) < dZ) kminF = -1;
|
|
||||||
if (Symmetry > EQ_SYMM && fabs(X[0]) < dX) iminF = -1;
|
|
||||||
if (Symmetry > EQ_SYMM && fabs(Y[0]) < dY) jminF = -1;
|
|
||||||
|
|
||||||
const double SoA[3] = { SYM1, SYM2, SYM3 };
|
|
||||||
|
|
||||||
/* fh: (ex1+2)*(ex2+2)*(ex3+2) because ord=2 */
|
|
||||||
const size_t nx = (size_t)ex1 + 2;
|
|
||||||
const size_t ny = (size_t)ex2 + 2;
|
|
||||||
const size_t nz = (size_t)ex3 + 2;
|
|
||||||
const size_t fh_size = nx * ny * nz;
|
|
||||||
|
|
||||||
static double *fh = NULL;
|
|
||||||
static size_t cap = 0;
|
|
||||||
|
|
||||||
if (fh_size > cap) {
|
|
||||||
free(fh);
|
|
||||||
fh = (double*)aligned_alloc(64, fh_size * sizeof(double));
|
|
||||||
cap = fh_size;
|
|
||||||
}
|
|
||||||
// double *fh = (double*)malloc(fh_size * sizeof(double));
|
|
||||||
if (!fh) return;
|
|
||||||
|
|
||||||
symmetry_bd(2, ex, f, fh, SoA);
|
|
||||||
|
|
||||||
/* 系数:按 Fortran 原式 */
|
|
||||||
const double Sdxdx = ONE / (dX * dX);
|
|
||||||
const double Sdydy = ONE / (dY * dY);
|
|
||||||
const double Sdzdz = ONE / (dZ * dZ);
|
|
||||||
|
|
||||||
const double Fdxdx = F1o12 / (dX * dX);
|
|
||||||
const double Fdydy = F1o12 / (dY * dY);
|
|
||||||
const double Fdzdz = F1o12 / (dZ * dZ);
|
|
||||||
|
|
||||||
const double Sdxdy = F1o4 / (dX * dY);
|
|
||||||
const double Sdxdz = F1o4 / (dX * dZ);
|
|
||||||
const double Sdydz = F1o4 / (dY * dZ);
|
|
||||||
|
|
||||||
const double Fdxdy = F1o144 / (dX * dY);
|
|
||||||
const double Fdxdz = F1o144 / (dX * dZ);
|
|
||||||
const double Fdydz = F1o144 / (dY * dZ);
|
|
||||||
|
|
||||||
/* 输出清零:fxx,fyy,fzz,fxy,fxz,fyz = 0 */
|
|
||||||
const size_t all = (size_t)ex1 * (size_t)ex2 * (size_t)ex3;
|
|
||||||
for (size_t p = 0; p < all; ++p) {
|
|
||||||
fxx[p] = ZEO; fyy[p] = ZEO; fzz[p] = ZEO;
|
|
||||||
fxy[p] = ZEO; fxz[p] = ZEO; fyz[p] = ZEO;
|
|
||||||
}
|
|
||||||
|
|
||||||
/*
|
|
||||||
* Fortran:
|
|
||||||
* do k=1,ex3-1
|
|
||||||
* do j=1,ex2-1
|
|
||||||
* do i=1,ex1-1
|
|
||||||
*/
|
|
||||||
|
|
||||||
for (int k0 = 0; k0 <= ex3 - 2; ++k0) {
|
|
||||||
const int kF = k0 + 1;
|
|
||||||
for (int j0 = 0; j0 <= ex2 - 2; ++j0) {
|
|
||||||
const int jF = j0 + 1;
|
|
||||||
for (int i0 = 0; i0 <= ex1 - 2; ++i0) {
|
|
||||||
const int iF = i0 + 1;
|
|
||||||
const size_t p = idx_ex(i0, j0, k0, ex);
|
|
||||||
|
|
||||||
/* 高阶分支:i±2,j±2,k±2 都在范围内 */
|
|
||||||
if ((iF + 2) <= imaxF && (iF - 2) >= iminF &&
|
|
||||||
(jF + 2) <= jmaxF && (jF - 2) >= jminF &&
|
|
||||||
(kF + 2) <= kmaxF && (kF - 2) >= kminF)
|
|
||||||
{
|
|
||||||
fxx[p] = Fdxdx * (
|
|
||||||
-fh[idx_fh_F_ord2(iF - 2, jF, kF, ex)] +
|
|
||||||
F16 * fh[idx_fh_F_ord2(iF - 1, jF, kF, ex)] -
|
|
||||||
F30 * fh[idx_fh_F_ord2(iF, jF, kF, ex)] -
|
|
||||||
fh[idx_fh_F_ord2(iF + 2, jF, kF, ex)] +
|
|
||||||
F16 * fh[idx_fh_F_ord2(iF + 1, jF, kF, ex)]
|
|
||||||
);
|
|
||||||
|
|
||||||
fyy[p] = Fdydy * (
|
|
||||||
-fh[idx_fh_F_ord2(iF, jF - 2, kF, ex)] +
|
|
||||||
F16 * fh[idx_fh_F_ord2(iF, jF - 1, kF, ex)] -
|
|
||||||
F30 * fh[idx_fh_F_ord2(iF, jF, kF, ex)] -
|
|
||||||
fh[idx_fh_F_ord2(iF, jF + 2, kF, ex)] +
|
|
||||||
F16 * fh[idx_fh_F_ord2(iF, jF + 1, kF, ex)]
|
|
||||||
);
|
|
||||||
|
|
||||||
fzz[p] = Fdzdz * (
|
|
||||||
-fh[idx_fh_F_ord2(iF, jF, kF - 2, ex)] +
|
|
||||||
F16 * fh[idx_fh_F_ord2(iF, jF, kF - 1, ex)] -
|
|
||||||
F30 * fh[idx_fh_F_ord2(iF, jF, kF, ex)] -
|
|
||||||
fh[idx_fh_F_ord2(iF, jF, kF + 2, ex)] +
|
|
||||||
F16 * fh[idx_fh_F_ord2(iF, jF, kF + 1, ex)]
|
|
||||||
);
|
|
||||||
|
|
||||||
/* fxy 高阶:完全照搬 Fortran 的括号结构 */
|
|
||||||
{
|
|
||||||
const double t_jm2 =
|
|
||||||
( fh[idx_fh_F_ord2(iF - 2, jF - 2, kF, ex)]
|
|
||||||
-F8*fh[idx_fh_F_ord2(iF - 1, jF - 2, kF, ex)]
|
|
||||||
+F8*fh[idx_fh_F_ord2(iF + 1, jF - 2, kF, ex)]
|
|
||||||
- fh[idx_fh_F_ord2(iF + 2, jF - 2, kF, ex)] );
|
|
||||||
|
|
||||||
const double t_jm1 =
|
|
||||||
( fh[idx_fh_F_ord2(iF - 2, jF - 1, kF, ex)]
|
|
||||||
-F8*fh[idx_fh_F_ord2(iF - 1, jF - 1, kF, ex)]
|
|
||||||
+F8*fh[idx_fh_F_ord2(iF + 1, jF - 1, kF, ex)]
|
|
||||||
- fh[idx_fh_F_ord2(iF + 2, jF - 1, kF, ex)] );
|
|
||||||
|
|
||||||
const double t_jp1 =
|
|
||||||
( fh[idx_fh_F_ord2(iF - 2, jF + 1, kF, ex)]
|
|
||||||
-F8*fh[idx_fh_F_ord2(iF - 1, jF + 1, kF, ex)]
|
|
||||||
+F8*fh[idx_fh_F_ord2(iF + 1, jF + 1, kF, ex)]
|
|
||||||
- fh[idx_fh_F_ord2(iF + 2, jF + 1, kF, ex)] );
|
|
||||||
|
|
||||||
const double t_jp2 =
|
|
||||||
( fh[idx_fh_F_ord2(iF - 2, jF + 2, kF, ex)]
|
|
||||||
-F8*fh[idx_fh_F_ord2(iF - 1, jF + 2, kF, ex)]
|
|
||||||
+F8*fh[idx_fh_F_ord2(iF + 1, jF + 2, kF, ex)]
|
|
||||||
- fh[idx_fh_F_ord2(iF + 2, jF + 2, kF, ex)] );
|
|
||||||
|
|
||||||
fxy[p] = Fdxdy * ( t_jm2 - F8 * t_jm1 + F8 * t_jp1 - t_jp2 );
|
|
||||||
}
|
|
||||||
|
|
||||||
/* fxz 高阶 */
|
|
||||||
{
|
|
||||||
const double t_km2 =
|
|
||||||
( fh[idx_fh_F_ord2(iF - 2, jF, kF - 2, ex)]
|
|
||||||
-F8*fh[idx_fh_F_ord2(iF - 1, jF, kF - 2, ex)]
|
|
||||||
+F8*fh[idx_fh_F_ord2(iF + 1, jF, kF - 2, ex)]
|
|
||||||
- fh[idx_fh_F_ord2(iF + 2, jF, kF - 2, ex)] );
|
|
||||||
|
|
||||||
const double t_km1 =
|
|
||||||
( fh[idx_fh_F_ord2(iF - 2, jF, kF - 1, ex)]
|
|
||||||
-F8*fh[idx_fh_F_ord2(iF - 1, jF, kF - 1, ex)]
|
|
||||||
+F8*fh[idx_fh_F_ord2(iF + 1, jF, kF - 1, ex)]
|
|
||||||
- fh[idx_fh_F_ord2(iF + 2, jF, kF - 1, ex)] );
|
|
||||||
|
|
||||||
const double t_kp1 =
|
|
||||||
( fh[idx_fh_F_ord2(iF - 2, jF, kF + 1, ex)]
|
|
||||||
-F8*fh[idx_fh_F_ord2(iF - 1, jF, kF + 1, ex)]
|
|
||||||
+F8*fh[idx_fh_F_ord2(iF + 1, jF, kF + 1, ex)]
|
|
||||||
- fh[idx_fh_F_ord2(iF + 2, jF, kF + 1, ex)] );
|
|
||||||
|
|
||||||
const double t_kp2 =
|
|
||||||
( fh[idx_fh_F_ord2(iF - 2, jF, kF + 2, ex)]
|
|
||||||
-F8*fh[idx_fh_F_ord2(iF - 1, jF, kF + 2, ex)]
|
|
||||||
+F8*fh[idx_fh_F_ord2(iF + 1, jF, kF + 2, ex)]
|
|
||||||
- fh[idx_fh_F_ord2(iF + 2, jF, kF + 2, ex)] );
|
|
||||||
|
|
||||||
fxz[p] = Fdxdz * ( t_km2 - F8 * t_km1 + F8 * t_kp1 - t_kp2 );
|
|
||||||
}
|
|
||||||
|
|
||||||
/* fyz 高阶 */
|
|
||||||
{
|
|
||||||
const double t_km2 =
|
|
||||||
( fh[idx_fh_F_ord2(iF, jF - 2, kF - 2, ex)]
|
|
||||||
-F8*fh[idx_fh_F_ord2(iF, jF - 1, kF - 2, ex)]
|
|
||||||
+F8*fh[idx_fh_F_ord2(iF, jF + 1, kF - 2, ex)]
|
|
||||||
- fh[idx_fh_F_ord2(iF, jF + 2, kF - 2, ex)] );
|
|
||||||
|
|
||||||
const double t_km1 =
|
|
||||||
( fh[idx_fh_F_ord2(iF, jF - 2, kF - 1, ex)]
|
|
||||||
-F8*fh[idx_fh_F_ord2(iF, jF - 1, kF - 1, ex)]
|
|
||||||
+F8*fh[idx_fh_F_ord2(iF, jF + 1, kF - 1, ex)]
|
|
||||||
- fh[idx_fh_F_ord2(iF, jF + 2, kF - 1, ex)] );
|
|
||||||
|
|
||||||
const double t_kp1 =
|
|
||||||
( fh[idx_fh_F_ord2(iF, jF - 2, kF + 1, ex)]
|
|
||||||
-F8*fh[idx_fh_F_ord2(iF, jF - 1, kF + 1, ex)]
|
|
||||||
+F8*fh[idx_fh_F_ord2(iF, jF + 1, kF + 1, ex)]
|
|
||||||
- fh[idx_fh_F_ord2(iF, jF + 2, kF + 1, ex)] );
|
|
||||||
|
|
||||||
const double t_kp2 =
|
|
||||||
( fh[idx_fh_F_ord2(iF, jF - 2, kF + 2, ex)]
|
|
||||||
-F8*fh[idx_fh_F_ord2(iF, jF - 1, kF + 2, ex)]
|
|
||||||
+F8*fh[idx_fh_F_ord2(iF, jF + 1, kF + 2, ex)]
|
|
||||||
- fh[idx_fh_F_ord2(iF, jF + 2, kF + 2, ex)] );
|
|
||||||
|
|
||||||
fyz[p] = Fdydz * ( t_km2 - F8 * t_km1 + F8 * t_kp1 - t_kp2 );
|
|
||||||
}
|
|
||||||
}
|
|
||||||
/* 二阶分支:i±1,j±1,k±1 在范围内 */
|
|
||||||
else if ((iF + 1) <= imaxF && (iF - 1) >= iminF &&
|
|
||||||
(jF + 1) <= jmaxF && (jF - 1) >= jminF &&
|
|
||||||
(kF + 1) <= kmaxF && (kF - 1) >= kminF)
|
|
||||||
{
|
|
||||||
fxx[p] = Sdxdx * (
|
|
||||||
fh[idx_fh_F_ord2(iF - 1, jF, kF, ex)] -
|
|
||||||
TWO * fh[idx_fh_F_ord2(iF, jF, kF, ex)] +
|
|
||||||
fh[idx_fh_F_ord2(iF + 1, jF, kF, ex)]
|
|
||||||
);
|
|
||||||
|
|
||||||
fyy[p] = Sdydy * (
|
|
||||||
fh[idx_fh_F_ord2(iF, jF - 1, kF, ex)] -
|
|
||||||
TWO * fh[idx_fh_F_ord2(iF, jF, kF, ex)] +
|
|
||||||
fh[idx_fh_F_ord2(iF, jF + 1, kF, ex)]
|
|
||||||
);
|
|
||||||
|
|
||||||
fzz[p] = Sdzdz * (
|
|
||||||
fh[idx_fh_F_ord2(iF, jF, kF - 1, ex)] -
|
|
||||||
TWO * fh[idx_fh_F_ord2(iF, jF, kF, ex)] +
|
|
||||||
fh[idx_fh_F_ord2(iF, jF, kF + 1, ex)]
|
|
||||||
);
|
|
||||||
|
|
||||||
fxy[p] = Sdxdy * (
|
|
||||||
fh[idx_fh_F_ord2(iF - 1, jF - 1, kF, ex)] -
|
|
||||||
fh[idx_fh_F_ord2(iF + 1, jF - 1, kF, ex)] -
|
|
||||||
fh[idx_fh_F_ord2(iF - 1, jF + 1, kF, ex)] +
|
|
||||||
fh[idx_fh_F_ord2(iF + 1, jF + 1, kF, ex)]
|
|
||||||
);
|
|
||||||
|
|
||||||
fxz[p] = Sdxdz * (
|
|
||||||
fh[idx_fh_F_ord2(iF - 1, jF, kF - 1, ex)] -
|
|
||||||
fh[idx_fh_F_ord2(iF + 1, jF, kF - 1, ex)] -
|
|
||||||
fh[idx_fh_F_ord2(iF - 1, jF, kF + 1, ex)] +
|
|
||||||
fh[idx_fh_F_ord2(iF + 1, jF, kF + 1, ex)]
|
|
||||||
);
|
|
||||||
|
|
||||||
fyz[p] = Sdydz * (
|
|
||||||
fh[idx_fh_F_ord2(iF, jF - 1, kF - 1, ex)] -
|
|
||||||
fh[idx_fh_F_ord2(iF, jF + 1, kF - 1, ex)] -
|
|
||||||
fh[idx_fh_F_ord2(iF, jF - 1, kF + 1, ex)] +
|
|
||||||
fh[idx_fh_F_ord2(iF, jF + 1, kF + 1, ex)]
|
|
||||||
);
|
|
||||||
}else{
|
|
||||||
fxx[p] = 0.0;
|
|
||||||
fyy[p] = 0.0;
|
|
||||||
fzz[p] = 0.0;
|
|
||||||
fxy[p] = 0.0;
|
|
||||||
fxz[p] = 0.0;
|
|
||||||
fyz[p] = 0.0;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// free(fh);
|
|
||||||
}
|
|
||||||
@@ -1,150 +0,0 @@
|
|||||||
#include "tool.h"
|
|
||||||
|
|
||||||
/*
|
|
||||||
* C 版 fderivs
|
|
||||||
*
|
|
||||||
* Fortran:
|
|
||||||
* subroutine fderivs(ex,f,fx,fy,fz,X,Y,Z,SYM1,SYM2,SYM3,symmetry,onoff)
|
|
||||||
*
|
|
||||||
* 约定:
|
|
||||||
* f, fx, fy, fz: ex1*ex2*ex3,按 idx_ex 布局
|
|
||||||
* X: ex1, Y: ex2, Z: ex3
|
|
||||||
*/
|
|
||||||
void fderivs(const int ex[3],
|
|
||||||
const double *f,
|
|
||||||
double *fx, double *fy, double *fz,
|
|
||||||
const double *X, const double *Y, const double *Z,
|
|
||||||
double SYM1, double SYM2, double SYM3,
|
|
||||||
int Symmetry, int onoff)
|
|
||||||
{
|
|
||||||
(void)onoff; // Fortran 里没用到
|
|
||||||
|
|
||||||
const double ZEO = 0.0, ONE = 1.0;
|
|
||||||
const double TWO = 2.0, EIT = 8.0;
|
|
||||||
const double F12 = 12.0;
|
|
||||||
|
|
||||||
const int NO_SYMM = 0, EQ_SYMM = 1; // OCTANT=2 在本子程序里不直接用
|
|
||||||
|
|
||||||
const int ex1 = ex[0], ex2 = ex[1], ex3 = ex[2];
|
|
||||||
|
|
||||||
// dX = X(2)-X(1) -> C: X[1]-X[0]
|
|
||||||
const double dX = X[1] - X[0];
|
|
||||||
const double dY = Y[1] - Y[0];
|
|
||||||
const double dZ = Z[1] - Z[0];
|
|
||||||
|
|
||||||
// Fortran 1-based bounds
|
|
||||||
const int imaxF = ex1;
|
|
||||||
const int jmaxF = ex2;
|
|
||||||
const int kmaxF = ex3;
|
|
||||||
|
|
||||||
int iminF = 1, jminF = 1, kminF = 1;
|
|
||||||
if (Symmetry > NO_SYMM && fabs(Z[0]) < dZ) kminF = -1;
|
|
||||||
if (Symmetry > EQ_SYMM && fabs(X[0]) < dX) iminF = -1;
|
|
||||||
if (Symmetry > EQ_SYMM && fabs(Y[0]) < dY) jminF = -1;
|
|
||||||
|
|
||||||
// SoA(1:3) = SYM1,SYM2,SYM3
|
|
||||||
const double SoA[3] = { SYM1, SYM2, SYM3 };
|
|
||||||
|
|
||||||
// fh: (ex1+2)*(ex2+2)*(ex3+2) because ord=2
|
|
||||||
const size_t nx = (size_t)ex1 + 2;
|
|
||||||
const size_t ny = (size_t)ex2 + 2;
|
|
||||||
const size_t nz = (size_t)ex3 + 2;
|
|
||||||
const size_t fh_size = nx * ny * nz;
|
|
||||||
static double *fh = NULL;
|
|
||||||
static size_t cap = 0;
|
|
||||||
|
|
||||||
if (fh_size > cap) {
|
|
||||||
free(fh);
|
|
||||||
fh = (double*)aligned_alloc(64, fh_size * sizeof(double));
|
|
||||||
cap = fh_size;
|
|
||||||
}
|
|
||||||
// double *fh = (double*)malloc(fh_size * sizeof(double));
|
|
||||||
if (!fh) return;
|
|
||||||
|
|
||||||
// call symmetry_bd(2,ex,f,fh,SoA)
|
|
||||||
symmetry_bd(2, ex, f, fh, SoA);
|
|
||||||
|
|
||||||
const double d12dx = ONE / F12 / dX;
|
|
||||||
const double d12dy = ONE / F12 / dY;
|
|
||||||
const double d12dz = ONE / F12 / dZ;
|
|
||||||
|
|
||||||
const double d2dx = ONE / TWO / dX;
|
|
||||||
const double d2dy = ONE / TWO / dY;
|
|
||||||
const double d2dz = ONE / TWO / dZ;
|
|
||||||
|
|
||||||
// fx = fy = fz = 0
|
|
||||||
const size_t all = (size_t)ex1 * (size_t)ex2 * (size_t)ex3;
|
|
||||||
for (size_t p = 0; p < all; ++p) {
|
|
||||||
fx[p] = ZEO;
|
|
||||||
fy[p] = ZEO;
|
|
||||||
fz[p] = ZEO;
|
|
||||||
}
|
|
||||||
|
|
||||||
/*
|
|
||||||
* Fortran loops:
|
|
||||||
* do k=1,ex3-1
|
|
||||||
* do j=1,ex2-1
|
|
||||||
* do i=1,ex1-1
|
|
||||||
*
|
|
||||||
* C: k0=0..ex3-2, j0=0..ex2-2, i0=0..ex1-2
|
|
||||||
*/
|
|
||||||
for (int k0 = 0; k0 <= ex3 - 2; ++k0) {
|
|
||||||
const int kF = k0 + 1;
|
|
||||||
for (int j0 = 0; j0 <= ex2 - 2; ++j0) {
|
|
||||||
const int jF = j0 + 1;
|
|
||||||
for (int i0 = 0; i0 <= ex1 - 2; ++i0) {
|
|
||||||
const int iF = i0 + 1;
|
|
||||||
const size_t p = idx_ex(i0, j0, k0, ex);
|
|
||||||
|
|
||||||
// if(i+2 <= imax .and. i-2 >= imin ... ) (全是 Fortran 索引)
|
|
||||||
if ((iF + 2) <= imaxF && (iF - 2) >= iminF &&
|
|
||||||
(jF + 2) <= jmaxF && (jF - 2) >= jminF &&
|
|
||||||
(kF + 2) <= kmaxF && (kF - 2) >= kminF)
|
|
||||||
{
|
|
||||||
fx[p] = d12dx * (
|
|
||||||
fh[idx_fh_F_ord2(iF - 2, jF, kF, ex)] -
|
|
||||||
EIT * fh[idx_fh_F_ord2(iF - 1, jF, kF, ex)] +
|
|
||||||
EIT * fh[idx_fh_F_ord2(iF + 1, jF, kF, ex)] -
|
|
||||||
fh[idx_fh_F_ord2(iF + 2, jF, kF, ex)]
|
|
||||||
);
|
|
||||||
|
|
||||||
fy[p] = d12dy * (
|
|
||||||
fh[idx_fh_F_ord2(iF, jF - 2, kF, ex)] -
|
|
||||||
EIT * fh[idx_fh_F_ord2(iF, jF - 1, kF, ex)] +
|
|
||||||
EIT * fh[idx_fh_F_ord2(iF, jF + 1, kF, ex)] -
|
|
||||||
fh[idx_fh_F_ord2(iF, jF + 2, kF, ex)]
|
|
||||||
);
|
|
||||||
|
|
||||||
fz[p] = d12dz * (
|
|
||||||
fh[idx_fh_F_ord2(iF, jF, kF - 2, ex)] -
|
|
||||||
EIT * fh[idx_fh_F_ord2(iF, jF, kF - 1, ex)] +
|
|
||||||
EIT * fh[idx_fh_F_ord2(iF, jF, kF + 1, ex)] -
|
|
||||||
fh[idx_fh_F_ord2(iF, jF, kF + 2, ex)]
|
|
||||||
);
|
|
||||||
}
|
|
||||||
// elseif(i+1 <= imax .and. i-1 >= imin ...)
|
|
||||||
else if ((iF + 1) <= imaxF && (iF - 1) >= iminF &&
|
|
||||||
(jF + 1) <= jmaxF && (jF - 1) >= jminF &&
|
|
||||||
(kF + 1) <= kmaxF && (kF - 1) >= kminF)
|
|
||||||
{
|
|
||||||
fx[p] = d2dx * (
|
|
||||||
-fh[idx_fh_F_ord2(iF - 1, jF, kF, ex)] +
|
|
||||||
fh[idx_fh_F_ord2(iF + 1, jF, kF, ex)]
|
|
||||||
);
|
|
||||||
|
|
||||||
fy[p] = d2dy * (
|
|
||||||
-fh[idx_fh_F_ord2(iF, jF - 1, kF, ex)] +
|
|
||||||
fh[idx_fh_F_ord2(iF, jF + 1, kF, ex)]
|
|
||||||
);
|
|
||||||
|
|
||||||
fz[p] = d2dz * (
|
|
||||||
-fh[idx_fh_F_ord2(iF, jF, kF - 1, ex)] +
|
|
||||||
fh[idx_fh_F_ord2(iF, jF, kF + 1, ex)]
|
|
||||||
);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// free(fh);
|
|
||||||
}
|
|
||||||
@@ -883,17 +883,13 @@ subroutine symmetry_bd(ord,extc,func,funcc,SoA)
|
|||||||
|
|
||||||
integer::i
|
integer::i
|
||||||
|
|
||||||
!DIR$ SIMD VECTORLENGTHFOR(KNOWN_INTEGER=8)
|
|
||||||
funcc(1:extc(1),1:extc(2),1:extc(3)) = func
|
funcc(1:extc(1),1:extc(2),1:extc(3)) = func
|
||||||
!DIR$ SIMD VECTORLENGTHFOR(KNOWN_INTEGER=8)
|
|
||||||
do i=0,ord-1
|
do i=0,ord-1
|
||||||
funcc(-i,1:extc(2),1:extc(3)) = funcc(i+1,1:extc(2),1:extc(3))*SoA(1)
|
funcc(-i,1:extc(2),1:extc(3)) = funcc(i+1,1:extc(2),1:extc(3))*SoA(1)
|
||||||
enddo
|
enddo
|
||||||
!DIR$ SIMD VECTORLENGTHFOR(KNOWN_INTEGER=8)
|
|
||||||
do i=0,ord-1
|
do i=0,ord-1
|
||||||
funcc(:,-i,1:extc(3)) = funcc(:,i+1,1:extc(3))*SoA(2)
|
funcc(:,-i,1:extc(3)) = funcc(:,i+1,1:extc(3))*SoA(2)
|
||||||
enddo
|
enddo
|
||||||
!DIR$ SIMD VECTORLENGTHFOR(KNOWN_INTEGER=8)
|
|
||||||
do i=0,ord-1
|
do i=0,ord-1
|
||||||
funcc(:,:,-i) = funcc(:,:,i+1)*SoA(3)
|
funcc(:,:,-i) = funcc(:,:,i+1)*SoA(3)
|
||||||
enddo
|
enddo
|
||||||
@@ -1116,7 +1112,6 @@ end subroutine d2dump
|
|||||||
! Lagrangian polynomial interpolation
|
! Lagrangian polynomial interpolation
|
||||||
!------------------------------------------------------------------------------
|
!------------------------------------------------------------------------------
|
||||||
|
|
||||||
!DIR$ ATTRIBUTES FORCEINLINE :: polint
|
|
||||||
subroutine polint(xa, ya, x, y, dy, ordn)
|
subroutine polint(xa, ya, x, y, dy, ordn)
|
||||||
implicit none
|
implicit none
|
||||||
|
|
||||||
|
|||||||
@@ -65,8 +65,6 @@ real*8,intent(in) :: eps
|
|||||||
! dx^4
|
! dx^4
|
||||||
|
|
||||||
! note the sign (-1)^r-1, now r=2
|
! note the sign (-1)^r-1, now r=2
|
||||||
!DIR$ SIMD VECTORLENGTHFOR(KNOWN_INTEGER=8)
|
|
||||||
!DIR$ UNROLL PARTIAL(4)
|
|
||||||
do k=1,ex(3)
|
do k=1,ex(3)
|
||||||
do j=1,ex(2)
|
do j=1,ex(2)
|
||||||
do i=1,ex(1)
|
do i=1,ex(1)
|
||||||
|
|||||||
@@ -1,109 +0,0 @@
|
|||||||
#include "tool.h"
|
|
||||||
|
|
||||||
/*
|
|
||||||
* C 版 kodis
|
|
||||||
*
|
|
||||||
* Fortran signature:
|
|
||||||
* subroutine kodis(ex,X,Y,Z,f,f_rhs,SoA,Symmetry,eps)
|
|
||||||
*
|
|
||||||
* 约定:
|
|
||||||
* X: ex1, Y: ex2, Z: ex3
|
|
||||||
* f, f_rhs: ex1*ex2*ex3 按 idx_ex 布局
|
|
||||||
* SoA[3]
|
|
||||||
* eps: double
|
|
||||||
*/
|
|
||||||
void kodis(const int ex[3],
|
|
||||||
const double *X, const double *Y, const double *Z,
|
|
||||||
const double *f, double *f_rhs,
|
|
||||||
const double SoA[3],
|
|
||||||
int Symmetry, double eps)
|
|
||||||
{
|
|
||||||
const double ONE = 1.0, SIX = 6.0, FIT = 15.0, TWT = 20.0;
|
|
||||||
const double cof = 64.0; // 2^6
|
|
||||||
const int NO_SYMM = 0, OCTANT = 2;
|
|
||||||
|
|
||||||
const int ex1 = ex[0], ex2 = ex[1], ex3 = ex[2];
|
|
||||||
|
|
||||||
// Fortran: dX = X(2)-X(1) -> C: X[1]-X[0]
|
|
||||||
const double dX = X[1] - X[0];
|
|
||||||
const double dY = Y[1] - Y[0];
|
|
||||||
const double dZ = Z[1] - Z[0];
|
|
||||||
(void)ONE; // ONE 在原 Fortran 里只是参数,这里不一定用得上
|
|
||||||
|
|
||||||
// Fortran: imax=ex(1) 等是 1-based 上界
|
|
||||||
const int imaxF = ex1;
|
|
||||||
const int jmaxF = ex2;
|
|
||||||
const int kmaxF = ex3;
|
|
||||||
|
|
||||||
// Fortran: imin=jmin=kmin=1,某些对称情况变 -2
|
|
||||||
int iminF = 1, jminF = 1, kminF = 1;
|
|
||||||
|
|
||||||
if (Symmetry > NO_SYMM && fabs(Z[0]) < dZ) kminF = -2;
|
|
||||||
if (Symmetry == OCTANT && fabs(X[0]) < dX) iminF = -2;
|
|
||||||
if (Symmetry == OCTANT && fabs(Y[0]) < dY) jminF = -2;
|
|
||||||
|
|
||||||
// 分配 fh:大小 (ex1+3)*(ex2+3)*(ex3+3),对应 ord=3
|
|
||||||
const size_t nx = (size_t)ex1 + 3;
|
|
||||||
const size_t ny = (size_t)ex2 + 3;
|
|
||||||
const size_t nz = (size_t)ex3 + 3;
|
|
||||||
const size_t fh_size = nx * ny * nz;
|
|
||||||
|
|
||||||
double *fh = (double*)malloc(fh_size * sizeof(double));
|
|
||||||
if (!fh) return;
|
|
||||||
|
|
||||||
// Fortran: call symmetry_bd(3,ex,f,fh,SoA)
|
|
||||||
symmetry_bd(3, ex, f, fh, SoA);
|
|
||||||
|
|
||||||
/*
|
|
||||||
* Fortran loops:
|
|
||||||
* do k=1,ex3
|
|
||||||
* do j=1,ex2
|
|
||||||
* do i=1,ex1
|
|
||||||
*
|
|
||||||
* C: k0=0..ex3-1, j0=0..ex2-1, i0=0..ex1-1
|
|
||||||
* 并定义 Fortran index: iF=i0+1, ...
|
|
||||||
*/
|
|
||||||
for (int k0 = 0; k0 < ex3; ++k0) {
|
|
||||||
const int kF = k0 + 1;
|
|
||||||
for (int j0 = 0; j0 < ex2; ++j0) {
|
|
||||||
const int jF = j0 + 1;
|
|
||||||
for (int i0 = 0; i0 < ex1; ++i0) {
|
|
||||||
const int iF = i0 + 1;
|
|
||||||
|
|
||||||
// Fortran if 条件:
|
|
||||||
// i-3 >= imin .and. i+3 <= imax 等(都是 Fortran 索引)
|
|
||||||
if ((iF - 3) >= iminF && (iF + 3) <= imaxF &&
|
|
||||||
(jF - 3) >= jminF && (jF + 3) <= jmaxF &&
|
|
||||||
(kF - 3) >= kminF && (kF + 3) <= kmaxF)
|
|
||||||
{
|
|
||||||
const size_t p = idx_ex(i0, j0, k0, ex);
|
|
||||||
|
|
||||||
// 三个方向各一份同型的 7 点组合(实际上是对称的 6th-order dissipation/filter 核)
|
|
||||||
const double Dx_term =
|
|
||||||
( (fh[idx_fh_F(iF - 3, jF, kF, ex)] + fh[idx_fh_F(iF + 3, jF, kF, ex)]) -
|
|
||||||
SIX * (fh[idx_fh_F(iF - 2, jF, kF, ex)] + fh[idx_fh_F(iF + 2, jF, kF, ex)]) +
|
|
||||||
FIT * (fh[idx_fh_F(iF - 1, jF, kF, ex)] + fh[idx_fh_F(iF + 1, jF, kF, ex)]) -
|
|
||||||
TWT * fh[idx_fh_F(iF , jF, kF, ex)] ) / dX;
|
|
||||||
|
|
||||||
const double Dy_term =
|
|
||||||
( (fh[idx_fh_F(iF, jF - 3, kF, ex)] + fh[idx_fh_F(iF, jF + 3, kF, ex)]) -
|
|
||||||
SIX * (fh[idx_fh_F(iF, jF - 2, kF, ex)] + fh[idx_fh_F(iF, jF + 2, kF, ex)]) +
|
|
||||||
FIT * (fh[idx_fh_F(iF, jF - 1, kF, ex)] + fh[idx_fh_F(iF, jF + 1, kF, ex)]) -
|
|
||||||
TWT * fh[idx_fh_F(iF, jF , kF, ex)] ) / dY;
|
|
||||||
|
|
||||||
const double Dz_term =
|
|
||||||
( (fh[idx_fh_F(iF, jF, kF - 3, ex)] + fh[idx_fh_F(iF, jF, kF + 3, ex)]) -
|
|
||||||
SIX * (fh[idx_fh_F(iF, jF, kF - 2, ex)] + fh[idx_fh_F(iF, jF, kF + 2, ex)]) +
|
|
||||||
FIT * (fh[idx_fh_F(iF, jF, kF - 1, ex)] + fh[idx_fh_F(iF, jF, kF + 1, ex)]) -
|
|
||||||
TWT * fh[idx_fh_F(iF, jF, kF , ex)] ) / dZ;
|
|
||||||
|
|
||||||
// Fortran:
|
|
||||||
// f_rhs(i,j,k) = f_rhs(i,j,k) + eps/cof*(Dx_term + Dy_term + Dz_term)
|
|
||||||
f_rhs[p] += (eps / cof) * (Dx_term + Dy_term + Dz_term);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
free(fh);
|
|
||||||
}
|
|
||||||
@@ -1,255 +0,0 @@
|
|||||||
#include "tool.h"
|
|
||||||
/*
|
|
||||||
* 你需要提供 symmetry_bd 的 C 版本(或 Fortran 绑到 C 的接口)。
|
|
||||||
* Fortran: call symmetry_bd(3,ex,f,fh,SoA)
|
|
||||||
*
|
|
||||||
* 约定:
|
|
||||||
* nghost = 3
|
|
||||||
* ex[3] = {ex1,ex2,ex3}
|
|
||||||
* f = 原始网格 (ex1*ex2*ex3)
|
|
||||||
* fh = 扩展网格 ((ex1+3)*(ex2+3)*(ex3+3)),对应 Fortran 的 (-2:ex1, ...)
|
|
||||||
* SoA[3] = 输入参数
|
|
||||||
*/
|
|
||||||
void lopsided(const int ex[3],
|
|
||||||
const double *X, const double *Y, const double *Z,
|
|
||||||
const double *f, double *f_rhs,
|
|
||||||
const double *Sfx, const double *Sfy, const double *Sfz,
|
|
||||||
int Symmetry, const double SoA[3])
|
|
||||||
{
|
|
||||||
const double ZEO = 0.0, ONE = 1.0, F3 = 3.0;
|
|
||||||
const double TWO = 2.0, F6 = 6.0, F18 = 18.0;
|
|
||||||
const double F12 = 12.0, F10 = 10.0, EIT = 8.0;
|
|
||||||
|
|
||||||
const int NO_SYMM = 0, EQ_SYMM = 1, OCTANT = 2;
|
|
||||||
(void)OCTANT; // 这里和 Fortran 一样只是定义了不用也没关系
|
|
||||||
|
|
||||||
const int ex1 = ex[0], ex2 = ex[1], ex3 = ex[2];
|
|
||||||
|
|
||||||
// 对应 Fortran: dX = X(2)-X(1) (Fortran 1-based)
|
|
||||||
// C: X[1]-X[0]
|
|
||||||
const double dX = X[1] - X[0];
|
|
||||||
const double dY = Y[1] - Y[0];
|
|
||||||
const double dZ = Z[1] - Z[0];
|
|
||||||
|
|
||||||
const double d12dx = ONE / F12 / dX;
|
|
||||||
const double d12dy = ONE / F12 / dY;
|
|
||||||
const double d12dz = ONE / F12 / dZ;
|
|
||||||
|
|
||||||
// Fortran 里算了 d2dx/d2dy/d2dz 但本 subroutine 里没用到(保持一致也算出来)
|
|
||||||
const double d2dx = ONE / TWO / dX;
|
|
||||||
const double d2dy = ONE / TWO / dY;
|
|
||||||
const double d2dz = ONE / TWO / dZ;
|
|
||||||
(void)d2dx; (void)d2dy; (void)d2dz;
|
|
||||||
|
|
||||||
// Fortran:
|
|
||||||
// imax = ex(1); jmax = ex(2); kmax = ex(3)
|
|
||||||
const int imaxF = ex1;
|
|
||||||
const int jmaxF = ex2;
|
|
||||||
const int kmaxF = ex3;
|
|
||||||
|
|
||||||
// Fortran:
|
|
||||||
// imin=jmin=kmin=1; 若满足对称条件则设为 -2
|
|
||||||
int iminF = 1, jminF = 1, kminF = 1;
|
|
||||||
if (Symmetry > NO_SYMM && fabs(Z[0]) < dZ) kminF = -2;
|
|
||||||
if (Symmetry > EQ_SYMM && fabs(X[0]) < dX) iminF = -2;
|
|
||||||
if (Symmetry > EQ_SYMM && fabs(Y[0]) < dY) jminF = -2;
|
|
||||||
|
|
||||||
// 分配 fh:大小 (ex1+3)*(ex2+3)*(ex3+3)
|
|
||||||
const size_t nx = (size_t)ex1 + 3;
|
|
||||||
const size_t ny = (size_t)ex2 + 3;
|
|
||||||
const size_t nz = (size_t)ex3 + 3;
|
|
||||||
const size_t fh_size = nx * ny * nz;
|
|
||||||
|
|
||||||
double *fh = (double*)malloc(fh_size * sizeof(double));
|
|
||||||
if (!fh) return; // 内存不足:直接返回(你也可以改成 abort/报错)
|
|
||||||
|
|
||||||
// Fortran: call symmetry_bd(3,ex,f,fh,SoA)
|
|
||||||
symmetry_bd(3, ex, f, fh, SoA);
|
|
||||||
|
|
||||||
/*
|
|
||||||
* Fortran 主循环:
|
|
||||||
* do k=1,ex(3)-1
|
|
||||||
* do j=1,ex(2)-1
|
|
||||||
* do i=1,ex(1)-1
|
|
||||||
*
|
|
||||||
* 转成 C 0-based:
|
|
||||||
* k0 = 0..ex3-2, j0 = 0..ex2-2, i0 = 0..ex1-2
|
|
||||||
*
|
|
||||||
* 并且 Fortran 里的 i/j/k 在 fh 访问时,仍然是 Fortran 索引值:
|
|
||||||
* iF=i0+1, jF=j0+1, kF=k0+1
|
|
||||||
*/
|
|
||||||
for (int k0 = 0; k0 <= ex3 - 2; ++k0) {
|
|
||||||
const int kF = k0 + 1;
|
|
||||||
for (int j0 = 0; j0 <= ex2 - 2; ++j0) {
|
|
||||||
const int jF = j0 + 1;
|
|
||||||
for (int i0 = 0; i0 <= ex1 - 2; ++i0) {
|
|
||||||
const int iF = i0 + 1;
|
|
||||||
|
|
||||||
const size_t p = idx_ex(i0, j0, k0, ex);
|
|
||||||
|
|
||||||
// ---------------- x direction ----------------
|
|
||||||
const double sfx = Sfx[p];
|
|
||||||
if (sfx > ZEO) {
|
|
||||||
// Fortran: if(i+3 <= imax)
|
|
||||||
// iF+3 <= ex1 <=> i0+4 <= ex1 <=> i0 <= ex1-4
|
|
||||||
if (i0 <= ex1 - 4) {
|
|
||||||
f_rhs[p] += sfx * d12dx *
|
|
||||||
(-F3 * fh[idx_fh_F(iF - 1, jF, kF, ex)]
|
|
||||||
-F10 * fh[idx_fh_F(iF , jF, kF, ex)]
|
|
||||||
+F18 * fh[idx_fh_F(iF + 1, jF, kF, ex)]
|
|
||||||
-F6 * fh[idx_fh_F(iF + 2, jF, kF, ex)]
|
|
||||||
+ fh[idx_fh_F(iF + 3, jF, kF, ex)]);
|
|
||||||
}
|
|
||||||
// elseif(i+2 <= imax) <=> i0 <= ex1-3
|
|
||||||
else if (i0 <= ex1 - 3) {
|
|
||||||
f_rhs[p] += sfx * d12dx *
|
|
||||||
( fh[idx_fh_F(iF - 2, jF, kF, ex)]
|
|
||||||
-EIT * fh[idx_fh_F(iF - 1, jF, kF, ex)]
|
|
||||||
+EIT * fh[idx_fh_F(iF + 1, jF, kF, ex)]
|
|
||||||
- fh[idx_fh_F(iF + 2, jF, kF, ex)]);
|
|
||||||
}
|
|
||||||
// elseif(i+1 <= imax) <=> i0 <= ex1-2(循环里总成立)
|
|
||||||
else if (i0 <= ex1 - 2) {
|
|
||||||
f_rhs[p] -= sfx * d12dx *
|
|
||||||
(-F3 * fh[idx_fh_F(iF + 1, jF, kF, ex)]
|
|
||||||
-F10 * fh[idx_fh_F(iF , jF, kF, ex)]
|
|
||||||
+F18 * fh[idx_fh_F(iF - 1, jF, kF, ex)]
|
|
||||||
-F6 * fh[idx_fh_F(iF - 2, jF, kF, ex)]
|
|
||||||
+ fh[idx_fh_F(iF - 3, jF, kF, ex)]);
|
|
||||||
}
|
|
||||||
} else if (sfx < ZEO) {
|
|
||||||
// Fortran: if(i-3 >= imin)
|
|
||||||
// (iF-3) >= iminF <=> (i0-2) >= iminF
|
|
||||||
if ((i0 - 2) >= iminF) {
|
|
||||||
f_rhs[p] -= sfx * d12dx *
|
|
||||||
(-F3 * fh[idx_fh_F(iF + 1, jF, kF, ex)]
|
|
||||||
-F10 * fh[idx_fh_F(iF , jF, kF, ex)]
|
|
||||||
+F18 * fh[idx_fh_F(iF - 1, jF, kF, ex)]
|
|
||||||
-F6 * fh[idx_fh_F(iF - 2, jF, kF, ex)]
|
|
||||||
+ fh[idx_fh_F(iF - 3, jF, kF, ex)]);
|
|
||||||
}
|
|
||||||
// elseif(i-2 >= imin) <=> (i0-1) >= iminF
|
|
||||||
else if ((i0 - 1) >= iminF) {
|
|
||||||
f_rhs[p] += sfx * d12dx *
|
|
||||||
( fh[idx_fh_F(iF - 2, jF, kF, ex)]
|
|
||||||
-EIT * fh[idx_fh_F(iF - 1, jF, kF, ex)]
|
|
||||||
+EIT * fh[idx_fh_F(iF + 1, jF, kF, ex)]
|
|
||||||
- fh[idx_fh_F(iF + 2, jF, kF, ex)]);
|
|
||||||
}
|
|
||||||
// elseif(i-1 >= imin) <=> i0 >= iminF
|
|
||||||
else if (i0 >= iminF) {
|
|
||||||
f_rhs[p] += sfx * d12dx *
|
|
||||||
(-F3 * fh[idx_fh_F(iF - 1, jF, kF, ex)]
|
|
||||||
-F10 * fh[idx_fh_F(iF , jF, kF, ex)]
|
|
||||||
+F18 * fh[idx_fh_F(iF + 1, jF, kF, ex)]
|
|
||||||
-F6 * fh[idx_fh_F(iF + 2, jF, kF, ex)]
|
|
||||||
+ fh[idx_fh_F(iF + 3, jF, kF, ex)]);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// ---------------- y direction ----------------
|
|
||||||
const double sfy = Sfy[p];
|
|
||||||
if (sfy > ZEO) {
|
|
||||||
// jF+3 <= ex2 <=> j0+4 <= ex2 <=> j0 <= ex2-4
|
|
||||||
if (j0 <= ex2 - 4) {
|
|
||||||
f_rhs[p] += sfy * d12dy *
|
|
||||||
(-F3 * fh[idx_fh_F(iF, jF - 1, kF, ex)]
|
|
||||||
-F10 * fh[idx_fh_F(iF, jF , kF, ex)]
|
|
||||||
+F18 * fh[idx_fh_F(iF, jF + 1, kF, ex)]
|
|
||||||
-F6 * fh[idx_fh_F(iF, jF + 2, kF, ex)]
|
|
||||||
+ fh[idx_fh_F(iF, jF + 3, kF, ex)]);
|
|
||||||
} else if (j0 <= ex2 - 3) {
|
|
||||||
f_rhs[p] += sfy * d12dy *
|
|
||||||
( fh[idx_fh_F(iF, jF - 2, kF, ex)]
|
|
||||||
-EIT * fh[idx_fh_F(iF, jF - 1, kF, ex)]
|
|
||||||
+EIT * fh[idx_fh_F(iF, jF + 1, kF, ex)]
|
|
||||||
- fh[idx_fh_F(iF, jF + 2, kF, ex)]);
|
|
||||||
} else if (j0 <= ex2 - 2) {
|
|
||||||
f_rhs[p] -= sfy * d12dy *
|
|
||||||
(-F3 * fh[idx_fh_F(iF, jF + 1, kF, ex)]
|
|
||||||
-F10 * fh[idx_fh_F(iF, jF , kF, ex)]
|
|
||||||
+F18 * fh[idx_fh_F(iF, jF - 1, kF, ex)]
|
|
||||||
-F6 * fh[idx_fh_F(iF, jF - 2, kF, ex)]
|
|
||||||
+ fh[idx_fh_F(iF, jF - 3, kF, ex)]);
|
|
||||||
}
|
|
||||||
} else if (sfy < ZEO) {
|
|
||||||
if ((j0 - 2) >= jminF) {
|
|
||||||
f_rhs[p] -= sfy * d12dy *
|
|
||||||
(-F3 * fh[idx_fh_F(iF, jF + 1, kF, ex)]
|
|
||||||
-F10 * fh[idx_fh_F(iF, jF , kF, ex)]
|
|
||||||
+F18 * fh[idx_fh_F(iF, jF - 1, kF, ex)]
|
|
||||||
-F6 * fh[idx_fh_F(iF, jF - 2, kF, ex)]
|
|
||||||
+ fh[idx_fh_F(iF, jF - 3, kF, ex)]);
|
|
||||||
} else if ((j0 - 1) >= jminF) {
|
|
||||||
f_rhs[p] += sfy * d12dy *
|
|
||||||
( fh[idx_fh_F(iF, jF - 2, kF, ex)]
|
|
||||||
-EIT * fh[idx_fh_F(iF, jF - 1, kF, ex)]
|
|
||||||
+EIT * fh[idx_fh_F(iF, jF + 1, kF, ex)]
|
|
||||||
- fh[idx_fh_F(iF, jF + 2, kF, ex)]);
|
|
||||||
} else if (j0 >= jminF) {
|
|
||||||
f_rhs[p] += sfy * d12dy *
|
|
||||||
(-F3 * fh[idx_fh_F(iF, jF - 1, kF, ex)]
|
|
||||||
-F10 * fh[idx_fh_F(iF, jF , kF, ex)]
|
|
||||||
+F18 * fh[idx_fh_F(iF, jF + 1, kF, ex)]
|
|
||||||
-F6 * fh[idx_fh_F(iF, jF + 2, kF, ex)]
|
|
||||||
+ fh[idx_fh_F(iF, jF + 3, kF, ex)]);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// ---------------- z direction ----------------
|
|
||||||
const double sfz = Sfz[p];
|
|
||||||
if (sfz > ZEO) {
|
|
||||||
if (k0 <= ex3 - 4) {
|
|
||||||
f_rhs[p] += sfz * d12dz *
|
|
||||||
(-F3 * fh[idx_fh_F(iF, jF, kF - 1, ex)]
|
|
||||||
-F10 * fh[idx_fh_F(iF, jF, kF , ex)]
|
|
||||||
+F18 * fh[idx_fh_F(iF, jF, kF + 1, ex)]
|
|
||||||
-F6 * fh[idx_fh_F(iF, jF, kF + 2, ex)]
|
|
||||||
+ fh[idx_fh_F(iF, jF, kF + 3, ex)]);
|
|
||||||
} else if (k0 <= ex3 - 3) {
|
|
||||||
f_rhs[p] += sfz * d12dz *
|
|
||||||
( fh[idx_fh_F(iF, jF, kF - 2, ex)]
|
|
||||||
-EIT * fh[idx_fh_F(iF, jF, kF - 1, ex)]
|
|
||||||
+EIT * fh[idx_fh_F(iF, jF, kF + 1, ex)]
|
|
||||||
- fh[idx_fh_F(iF, jF, kF + 2, ex)]);
|
|
||||||
} else if (k0 <= ex3 - 2) {
|
|
||||||
f_rhs[p] -= sfz * d12dz *
|
|
||||||
(-F3 * fh[idx_fh_F(iF, jF, kF + 1, ex)]
|
|
||||||
-F10 * fh[idx_fh_F(iF, jF, kF , ex)]
|
|
||||||
+F18 * fh[idx_fh_F(iF, jF, kF - 1, ex)]
|
|
||||||
-F6 * fh[idx_fh_F(iF, jF, kF - 2, ex)]
|
|
||||||
+ fh[idx_fh_F(iF, jF, kF - 3, ex)]);
|
|
||||||
}
|
|
||||||
} else if (sfz < ZEO) {
|
|
||||||
if ((k0 - 2) >= kminF) {
|
|
||||||
f_rhs[p] -= sfz * d12dz *
|
|
||||||
(-F3 * fh[idx_fh_F(iF, jF, kF + 1, ex)]
|
|
||||||
-F10 * fh[idx_fh_F(iF, jF, kF , ex)]
|
|
||||||
+F18 * fh[idx_fh_F(iF, jF, kF - 1, ex)]
|
|
||||||
-F6 * fh[idx_fh_F(iF, jF, kF - 2, ex)]
|
|
||||||
+ fh[idx_fh_F(iF, jF, kF - 3, ex)]);
|
|
||||||
} else if ((k0 - 1) >= kminF) {
|
|
||||||
f_rhs[p] += sfz * d12dz *
|
|
||||||
( fh[idx_fh_F(iF, jF, kF - 2, ex)]
|
|
||||||
-EIT * fh[idx_fh_F(iF, jF, kF - 1, ex)]
|
|
||||||
+EIT * fh[idx_fh_F(iF, jF, kF + 1, ex)]
|
|
||||||
- fh[idx_fh_F(iF, jF, kF + 2, ex)]);
|
|
||||||
} else if (k0 >= kminF) {
|
|
||||||
f_rhs[p] += sfz * d12dz *
|
|
||||||
(-F3 * fh[idx_fh_F(iF, jF, kF - 1, ex)]
|
|
||||||
-F10 * fh[idx_fh_F(iF, jF, kF , ex)]
|
|
||||||
+F18 * fh[idx_fh_F(iF, jF, kF + 1, ex)]
|
|
||||||
-F6 * fh[idx_fh_F(iF, jF, kF + 2, ex)]
|
|
||||||
+ fh[idx_fh_F(iF, jF, kF + 3, ex)]);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
free(fh);
|
|
||||||
}
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
@@ -1,23 +1,7 @@
|
|||||||
|
|
||||||
#define tetradtype 2
|
|
||||||
|
|
||||||
#define Cell
|
|
||||||
|
|
||||||
#define ghost_width 3
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
#define GAUGE 0
|
|
||||||
|
|
||||||
#define CPBC_ghost_width (ghost_width)
|
|
||||||
|
|
||||||
#define ABV 0
|
|
||||||
|
|
||||||
#define EScalar_CC 2
|
|
||||||
|
|
||||||
#if 0
|
#if 0
|
||||||
|
note here
|
||||||
define tetradtype
|
|
||||||
v:r; u: phi; w: theta
|
v:r; u: phi; w: theta
|
||||||
tetradtype 0
|
tetradtype 0
|
||||||
v^a = (x,y,z)
|
v^a = (x,y,z)
|
||||||
@@ -30,48 +14,70 @@ define tetradtype
|
|||||||
v_a = (x,y,z)
|
v_a = (x,y,z)
|
||||||
orthonormal order: v,u,w
|
orthonormal order: v,u,w
|
||||||
m = (phi - i theta)/sqrt(2) following Frans, Eq.(8) of PRD 75, 124018(2007)
|
m = (phi - i theta)/sqrt(2) following Frans, Eq.(8) of PRD 75, 124018(2007)
|
||||||
|
#endif
|
||||||
|
#define tetradtype 2
|
||||||
|
|
||||||
define Cell or Vertex
|
#if 0
|
||||||
|
note here
|
||||||
Cell center or Vertex center
|
Cell center or Vertex center
|
||||||
|
#endif
|
||||||
|
#define Cell
|
||||||
|
|
||||||
define ghost_width
|
#if 0
|
||||||
|
note here
|
||||||
2nd order: 2
|
2nd order: 2
|
||||||
4th order: 3
|
4th order: 3
|
||||||
6th order: 4
|
6th order: 4
|
||||||
8th order: 5
|
8th order: 5
|
||||||
|
#endif
|
||||||
|
#define ghost_width 3
|
||||||
|
|
||||||
define WithShell
|
#if 0
|
||||||
|
note here
|
||||||
use shell or not
|
use shell or not
|
||||||
|
#endif
|
||||||
|
#define WithShell
|
||||||
|
|
||||||
define CPBC
|
#if 0
|
||||||
|
note here
|
||||||
use constraint preserving boundary condition or not
|
use constraint preserving boundary condition or not
|
||||||
only affect Z4c
|
only affect Z4c
|
||||||
CPBC only supports WithShell
|
#endif
|
||||||
|
#define CPBC
|
||||||
|
|
||||||
define GAUGE
|
#if 0
|
||||||
|
note here
|
||||||
|
Gauge condition type
|
||||||
0: B^i gauge
|
0: B^i gauge
|
||||||
1: David puncture gauge
|
1: David's puncture gauge
|
||||||
2: MB B^i gauge
|
2: MB B^i gauge
|
||||||
3: RIT B^i gauge
|
3: RIT B^i gauge
|
||||||
4: MB beta gauge (beta gauge not means Eq.(3) of PRD 84, 124006)
|
4: MB beta gauge (beta gauge not means Eq.(3) of PRD 84, 124006)
|
||||||
5: RIT beta gauge (beta gauge not means Eq.(3) of PRD 84, 124006)
|
5: RIT beta gauge (beta gauge not means Eq.(3) of PRD 84, 124006)
|
||||||
6: MGB1 B^i gauge
|
6: MGB1 B^i gauge
|
||||||
7: MGB2 B^i gauge
|
7: MGB2 B^i gauge
|
||||||
|
#endif
|
||||||
|
#define GAUGE 2
|
||||||
|
|
||||||
define CPBC_ghost_width (ghost_width)
|
#if 0
|
||||||
buffer points for CPBC boundary
|
buffer points for CPBC boundary
|
||||||
|
#endif
|
||||||
|
#define CPBC_ghost_width (ghost_width)
|
||||||
|
|
||||||
define ABV
|
#if 0
|
||||||
0: using BSSN variable for constraint violation and psi4 calculation
|
using BSSN variable for constraint violation and psi4 calculation: 0
|
||||||
1: using ADM variable for constraint violation and psi4 calculation
|
using ADM variable for constraint violation and psi4 calculation: 1
|
||||||
|
#endif
|
||||||
|
#define ABV 0
|
||||||
|
|
||||||
define EScalar_CC
|
#if 0
|
||||||
Type of Potential and Scalar Distribution in F(R) Scalar-Tensor Theory
|
Type of Potential and Scalar Distribution in F(R) Scalar-Tensor Theory
|
||||||
1: Case C of 1112.3928, V=0
|
1: Case C of 1112.3928, V=0
|
||||||
2: shell with phi(r) = phi0 * a2^2/(1+a2^2), f(R) = R+a2*R^2 induced V
|
2: shell with a2^2*phi0/(1+a2^2), f(R) = R+a2*R^2 induced V
|
||||||
3: ground state of Schrodinger-Newton system, f(R) = R+a2*R^2 induced V
|
3: ground state of Schrodinger-Newton system, f(R) = R+a2*R^2 induced V
|
||||||
4: a2 = +oo and phi(r) = phi0 * 0.5 * ( tanh((r+r0)/sigma) - tanh((r-r0)/sigma) )
|
4: a2 = oo and phi(r) = phi0 * 0.5 * ( tanh((r+r0)/sigma) - tanh((r-r0)/sigma) )
|
||||||
5: shell with phi(r) = phi0*Exp(-(r-r0)**2/sigma), V = 0
|
5: shell with phi(r) = phi0*Exp(-(r-r0)**2/sigma), V = 0
|
||||||
|
|
||||||
#endif
|
#endif
|
||||||
|
#define EScalar_CC 2
|
||||||
|
|
||||||
|
|
||||||
|
|||||||
@@ -6,124 +6,92 @@
|
|||||||
|
|
||||||
// application parameters
|
// application parameters
|
||||||
|
|
||||||
|
/// ****
|
||||||
|
// sommerfeld boundary type
|
||||||
|
// 0: bam, 1: shibata
|
||||||
#define SommerType 0
|
#define SommerType 0
|
||||||
|
|
||||||
|
/// ****
|
||||||
|
// for Using Gauss-Legendre quadrature in theta direction
|
||||||
#define GaussInt
|
#define GaussInt
|
||||||
|
|
||||||
#define ABEtype 0
|
/// ****
|
||||||
|
|
||||||
//#define With_AHF
|
|
||||||
#define Psi4type 0
|
|
||||||
|
|
||||||
//#define Point_Psi4
|
|
||||||
|
|
||||||
#define RPS 1
|
|
||||||
|
|
||||||
#define AGM 0
|
|
||||||
|
|
||||||
#define RPB 0
|
|
||||||
|
|
||||||
#define MAPBH 1
|
|
||||||
|
|
||||||
#define PSTR 0
|
|
||||||
|
|
||||||
#define REGLEV 0
|
|
||||||
|
|
||||||
//#define USE_GPU
|
|
||||||
|
|
||||||
//#define CHECKDETAIL
|
|
||||||
|
|
||||||
//#define FAKECHECK
|
|
||||||
|
|
||||||
//
|
|
||||||
// define SommerType
|
|
||||||
// sommerfeld boundary type
|
|
||||||
// 0: bam
|
|
||||||
// 1: shibata
|
|
||||||
//
|
|
||||||
// define GaussInt
|
|
||||||
// for Using Gauss-Legendre quadrature in theta direction
|
|
||||||
//
|
|
||||||
// define ABEtype
|
|
||||||
// 0: BSSN vacuum
|
// 0: BSSN vacuum
|
||||||
// 1: coupled to scalar field
|
// 1: coupled to scalar field
|
||||||
// 2: Z4c vacuum
|
// 2: Z4c vacuum
|
||||||
// 3: coupled to Maxwell field
|
// 3: coupled to Maxwell field
|
||||||
//
|
//
|
||||||
// define With_AHF
|
#define ABEtype 2
|
||||||
|
|
||||||
|
/// ****
|
||||||
// using Apparent Horizon Finder
|
// using Apparent Horizon Finder
|
||||||
//
|
//#define With_AHF
|
||||||
// define Psi4type
|
|
||||||
|
/// ****
|
||||||
// Psi4 calculation method
|
// Psi4 calculation method
|
||||||
// 0: EB method
|
// 0: EB method
|
||||||
// 1: 4-D method
|
// 1: 4-D method
|
||||||
//
|
//
|
||||||
// define Point_Psi4
|
#define Psi4type 0
|
||||||
|
|
||||||
|
/// ****
|
||||||
// for Using point psi4 or not
|
// for Using point psi4 or not
|
||||||
//
|
//#define Point_Psi4
|
||||||
// define RPS
|
|
||||||
|
/// ****
|
||||||
// RestrictProlong in Step (0) or after Step (1)
|
// RestrictProlong in Step (0) or after Step (1)
|
||||||
//
|
#define RPS 1
|
||||||
// define AGM
|
|
||||||
|
/// ****
|
||||||
// Enforce algebra constraint
|
// Enforce algebra constraint
|
||||||
// for every RK4 sub step: 0
|
// for every RK4 sub step: 0
|
||||||
// only when iter_count == 3: 1
|
// only when iter_count == 3: 1
|
||||||
// after routine Step: 2
|
// after routine Step: 2
|
||||||
//
|
#define AGM 0
|
||||||
// define RPB
|
|
||||||
// Restrict Prolong using BAM style 1 or old style 0
|
|
||||||
//
|
|
||||||
// define MAPBH
|
|
||||||
// 1: move Analysis out ot 4 sub steps and treat PBH with Euler method
|
|
||||||
//
|
|
||||||
// define PSTR
|
|
||||||
// parallel structure
|
|
||||||
// 0: level by level
|
|
||||||
// 1: considering all levels
|
|
||||||
// 2: as 1 but reverse the CPU order
|
|
||||||
// 3: Frank's scheme
|
|
||||||
//
|
|
||||||
// define REGLEV
|
|
||||||
// regrid for every level or for all levels at a time
|
|
||||||
// 0: for every level;
|
|
||||||
// 1: for all
|
|
||||||
//
|
|
||||||
// define USE_GPU
|
|
||||||
// use gpu or not
|
|
||||||
//
|
|
||||||
// define CHECKDETAIL
|
|
||||||
// use checkpoint for every process
|
|
||||||
//
|
|
||||||
// define FAKECHECK
|
|
||||||
// use FakeCheckPrepare to write CheckPoint
|
|
||||||
//
|
|
||||||
|
|
||||||
|
/// ****
|
||||||
|
// Restrict Prolong using BAM style 1 or old style 0
|
||||||
|
#define RPB 0
|
||||||
|
|
||||||
|
/// ****
|
||||||
|
// 1: move Analysis out ot 4 sub steps and treat PBH with Euler method
|
||||||
|
#define MAPBH 1
|
||||||
|
|
||||||
|
/// ****
|
||||||
|
// parallel structure, 0: level by level, 1: considering all levels, 2: as 1 but reverse the CPU order, 3: Frank's scheme
|
||||||
|
#define PSTR 0
|
||||||
|
|
||||||
|
/// ****
|
||||||
|
// regrid for every level or for all levels at a time
|
||||||
|
// 0: for every level; 1: for all
|
||||||
|
#define REGLEV 0
|
||||||
|
|
||||||
|
/// ****
|
||||||
|
// use gpu or not
|
||||||
|
//#define USE_GPU
|
||||||
|
|
||||||
|
/// ****
|
||||||
|
// use checkpoint for every process
|
||||||
|
//#define CHECKDETAIL
|
||||||
|
|
||||||
|
/// ****
|
||||||
|
// use FakeCheckPrepare to write CheckPoint
|
||||||
|
//#define FAKECHECK
|
||||||
////================================================================
|
////================================================================
|
||||||
// some basic parameters for numerical calculation
|
// some basic parameters for numerical calculation
|
||||||
////================================================================
|
|
||||||
|
|
||||||
#define dim 3
|
#define dim 3
|
||||||
|
|
||||||
//#define Cell or Vertex in "macrodef.fh"
|
//#define Cell or Vertex in "microdef.fh"
|
||||||
|
|
||||||
|
// ******
|
||||||
|
// buffer point number for mesh refinement interface
|
||||||
#define buffer_width 6
|
#define buffer_width 6
|
||||||
|
|
||||||
#define SC_width buffer_width
|
// ******
|
||||||
|
|
||||||
#define CS_width (2*buffer_width)
|
|
||||||
|
|
||||||
//
|
|
||||||
// define Cell or Vertex in "macrodef.fh"
|
|
||||||
//
|
|
||||||
// define buffer_width
|
|
||||||
// buffer point number for mesh refinement interface
|
|
||||||
//
|
|
||||||
// define SC_width buffer_width
|
|
||||||
// buffer point number shell-box interface, on shell
|
// buffer point number shell-box interface, on shell
|
||||||
//
|
#define SC_width buffer_width
|
||||||
// define CS_width
|
|
||||||
// buffer point number shell-box interface, on box
|
// buffer point number shell-box interface, on box
|
||||||
//
|
#define CS_width (2*buffer_width)
|
||||||
|
|
||||||
#if(buffer_width < ghost_width)
|
#if(buffer_width < ghost_width)
|
||||||
#error we always assume buffer_width>ghost_width
|
#error we always assume buffer_width>ghost_width
|
||||||
@@ -142,4 +110,3 @@
|
|||||||
#define TINY 1e-10
|
#define TINY 1e-10
|
||||||
|
|
||||||
#endif /* MICRODEF_H */
|
#endif /* MICRODEF_H */
|
||||||
|
|
||||||
|
|||||||
@@ -2,27 +2,6 @@
|
|||||||
|
|
||||||
include makefile.inc
|
include makefile.inc
|
||||||
|
|
||||||
## ABE build flags selected by PGO_MODE (set in makefile.inc, default: opt)
|
|
||||||
## make -> opt (PGO-guided, maximum performance)
|
|
||||||
## make PGO_MODE=instrument -> instrument (Phase 1: collect fresh profile data)
|
|
||||||
PROFDATA = /home/$(shell whoami)/AMSS-NCKU/pgo_profile/default.profdata
|
|
||||||
|
|
||||||
ifeq ($(PGO_MODE),instrument)
|
|
||||||
## Phase 1: instrumentation — omit -ipo/-fp-model fast=2 for faster build and numerical stability
|
|
||||||
CXXAPPFLAGS = -O3 -xHost -fma -fprofile-instr-generate -ipo \
|
|
||||||
-Dfortran3 -Dnewc -I${MKLROOT}/include
|
|
||||||
f90appflags = -O3 -xHost -fma -fprofile-instr-generate -ipo \
|
|
||||||
-align array64byte -fpp -I${MKLROOT}/include
|
|
||||||
else
|
|
||||||
## opt (default): maximum performance with PGO profile data
|
|
||||||
CXXAPPFLAGS = -O3 -xHost -fp-model fast=2 -fma -ipo \
|
|
||||||
-fprofile-instr-use=$(PROFDATA) \
|
|
||||||
-Dfortran3 -Dnewc -I${MKLROOT}/include
|
|
||||||
f90appflags = -O3 -xHost -fp-model fast=2 -fma -ipo \
|
|
||||||
-fprofile-instr-use=$(PROFDATA) \
|
|
||||||
-align array64byte -fpp -I${MKLROOT}/include
|
|
||||||
endif
|
|
||||||
|
|
||||||
.SUFFIXES: .o .f90 .C .for .cu
|
.SUFFIXES: .o .f90 .C .for .cu
|
||||||
|
|
||||||
.f90.o:
|
.f90.o:
|
||||||
@@ -37,36 +16,13 @@ endif
|
|||||||
.cu.o:
|
.cu.o:
|
||||||
$(Cu) $(CUDA_APP_FLAGS) -c $< -o $@ $(CUDA_LIB_PATH)
|
$(Cu) $(CUDA_APP_FLAGS) -c $< -o $@ $(CUDA_LIB_PATH)
|
||||||
|
|
||||||
# C rewrite of BSSN RHS kernel and helpers
|
|
||||||
bssn_rhs_c.o: bssn_rhs_c.C
|
|
||||||
${CXX} $(CXXAPPFLAGS) -c $< $(filein) -o $@
|
|
||||||
|
|
||||||
fderivs_c.o: fderivs_c.C
|
|
||||||
${CXX} $(CXXAPPFLAGS) -c $< $(filein) -o $@
|
|
||||||
|
|
||||||
fdderivs_c.o: fdderivs_c.C
|
|
||||||
${CXX} $(CXXAPPFLAGS) -c $< $(filein) -o $@
|
|
||||||
|
|
||||||
kodiss_c.o: kodiss_c.C
|
|
||||||
${CXX} $(CXXAPPFLAGS) -c $< $(filein) -o $@
|
|
||||||
|
|
||||||
lopsided_c.o: lopsided_c.C
|
|
||||||
${CXX} $(CXXAPPFLAGS) -c $< $(filein) -o $@
|
|
||||||
|
|
||||||
## TwoPunctureABE uses fixed optimal flags, independent of CXXAPPFLAGS (which may be PGO-instrumented)
|
|
||||||
TP_OPTFLAGS = -O3 -xHost -fp-model fast=2 -fma -ipo -Dfortran3 -Dnewc -I${MKLROOT}/include
|
|
||||||
|
|
||||||
TwoPunctures.o: TwoPunctures.C
|
TwoPunctures.o: TwoPunctures.C
|
||||||
${CXX} $(TP_OPTFLAGS) -qopenmp -c $< -o $@
|
${CXX} $(CXXAPPFLAGS) -qopenmp -c $< -o $@
|
||||||
|
|
||||||
TwoPunctureABE.o: TwoPunctureABE.C
|
TwoPunctureABE.o: TwoPunctureABE.C
|
||||||
${CXX} $(TP_OPTFLAGS) -qopenmp -c $< -o $@
|
${CXX} $(CXXAPPFLAGS) -qopenmp -c $< -o $@
|
||||||
|
|
||||||
# Input files
|
# Input files
|
||||||
|
|
||||||
# C rewrite files
|
|
||||||
CFILES = bssn_rhs_c.o fderivs_c.o fdderivs_c.o kodiss_c.o lopsided_c.o
|
|
||||||
|
|
||||||
C++FILES = ABE.o Ansorg.o Block.o misc.o monitor.o Parallel.o MPatch.o var.o\
|
C++FILES = ABE.o Ansorg.o Block.o misc.o monitor.o Parallel.o MPatch.o var.o\
|
||||||
cgh.o bssn_class.o surface_integral.o ShellPatch.o\
|
cgh.o bssn_class.o surface_integral.o ShellPatch.o\
|
||||||
bssnEScalar_class.o perf.o Z4c_class.o NullShellPatch.o\
|
bssnEScalar_class.o perf.o Z4c_class.o NullShellPatch.o\
|
||||||
@@ -84,7 +40,7 @@ C++FILES_GPU = ABE.o Ansorg.o Block.o misc.o monitor.o Parallel.o MPatch.o var.o
|
|||||||
|
|
||||||
F90FILES = enforce_algebra.o fmisc.o initial_puncture.o prolongrestrict.o\
|
F90FILES = enforce_algebra.o fmisc.o initial_puncture.o prolongrestrict.o\
|
||||||
prolongrestrict_cell.o prolongrestrict_vertex.o\
|
prolongrestrict_cell.o prolongrestrict_vertex.o\
|
||||||
rungekutta4_rout.o diff_new.o kodiss.o kodiss_sh.o\
|
rungekutta4_rout.o bssn_rhs.o diff_new.o kodiss.o kodiss_sh.o\
|
||||||
lopsidediff.o sommerfeld_rout.o getnp4.o diff_new_sh.o\
|
lopsidediff.o sommerfeld_rout.o getnp4.o diff_new_sh.o\
|
||||||
shellfunctions.o bssn_rhs_ss.o Set_Rho_ADM.o\
|
shellfunctions.o bssn_rhs_ss.o Set_Rho_ADM.o\
|
||||||
getnp4EScalar.o bssnEScalar_rhs.o bssn_constraint.o ricci_gamma.o\
|
getnp4EScalar.o bssnEScalar_rhs.o bssn_constraint.o ricci_gamma.o\
|
||||||
@@ -107,7 +63,7 @@ TwoPunctureFILES = TwoPunctureABE.o TwoPunctures.o
|
|||||||
CUDAFILES = bssn_gpu.o bssn_gpu_rhs_ss.o
|
CUDAFILES = bssn_gpu.o bssn_gpu_rhs_ss.o
|
||||||
|
|
||||||
# file dependences
|
# file dependences
|
||||||
$(C++FILES) $(C++FILES_GPU) $(F90FILES) $(CFILES) $(AHFDOBJS) $(CUDAFILES): macrodef.fh
|
$(C++FILES) $(C++FILESGPU) $(F90FILES) $(AHFDOBJS) $(CUDAFILES): macrodef.fh
|
||||||
|
|
||||||
$(C++FILES): Block.h enforce_algebra.h fmisc.h initial_puncture.h macrodef.h\
|
$(C++FILES): Block.h enforce_algebra.h fmisc.h initial_puncture.h macrodef.h\
|
||||||
misc.h monitor.h MyList.h Parallel.h MPatch.h prolongrestrict.h\
|
misc.h monitor.h MyList.h Parallel.h MPatch.h prolongrestrict.h\
|
||||||
@@ -130,7 +86,7 @@ $(C++FILES_GPU): Block.h enforce_algebra.h fmisc.h initial_puncture.h macrodef.h
|
|||||||
|
|
||||||
$(AHFDOBJS): cctk.h cctk_Config.h cctk_Types.h cctk_Constants.h myglobal.h
|
$(AHFDOBJS): cctk.h cctk_Config.h cctk_Types.h cctk_Constants.h myglobal.h
|
||||||
|
|
||||||
$(C++FILES) $(C++FILES_GPU) $(CFILES) $(AHFDOBJS) $(CUDAFILES): macrodef.h
|
$(C++FILES) $(C++FILES_GPU) $(AHFDOBJS) $(CUDAFILES): macrodef.h
|
||||||
|
|
||||||
TwoPunctureFILES: TwoPunctures.h
|
TwoPunctureFILES: TwoPunctures.h
|
||||||
|
|
||||||
@@ -139,14 +95,14 @@ $(CUDAFILES): bssn_gpu.h gpu_mem.h gpu_rhsSS_mem.h
|
|||||||
misc.o : zbesh.o
|
misc.o : zbesh.o
|
||||||
|
|
||||||
# projects
|
# projects
|
||||||
ABE: $(C++FILES) $(CFILES) $(F90FILES) $(F77FILES) $(AHFDOBJS)
|
ABE: $(C++FILES) $(F90FILES) $(F77FILES) $(AHFDOBJS)
|
||||||
$(CLINKER) $(CXXAPPFLAGS) -o $@ $(C++FILES) $(CFILES) $(F90FILES) $(F77FILES) $(AHFDOBJS) $(LDLIBS)
|
$(CLINKER) $(CXXAPPFLAGS) -o $@ $(C++FILES) $(F90FILES) $(F77FILES) $(AHFDOBJS) $(LDLIBS)
|
||||||
|
|
||||||
ABEGPU: $(C++FILES_GPU) $(CFILES) $(F90FILES) $(F77FILES) $(AHFDOBJS) $(CUDAFILES)
|
ABEGPU: $(C++FILES_GPU) $(F90FILES) $(F77FILES) $(AHFDOBJS) $(CUDAFILES)
|
||||||
$(CLINKER) $(CXXAPPFLAGS) -o $@ $(C++FILES_GPU) $(CFILES) $(F90FILES) $(F77FILES) $(AHFDOBJS) $(CUDAFILES) $(LDLIBS)
|
$(CLINKER) $(CXXAPPFLAGS) -o $@ $(C++FILES_GPU) $(F90FILES) $(F77FILES) $(AHFDOBJS) $(CUDAFILES) $(LDLIBS)
|
||||||
|
|
||||||
TwoPunctureABE: $(TwoPunctureFILES)
|
TwoPunctureABE: $(TwoPunctureFILES)
|
||||||
$(CLINKER) $(TP_OPTFLAGS) -qopenmp -o $@ $(TwoPunctureFILES) $(LDLIBS)
|
$(CLINKER) $(CXXAPPFLAGS) -qopenmp -o $@ $(TwoPunctureFILES) $(LDLIBS)
|
||||||
|
|
||||||
clean:
|
clean:
|
||||||
rm *.o ABE ABEGPU TwoPunctureABE make.log -f
|
rm *.o ABE ABEGPU TwoPunctureABE make.log -f
|
||||||
|
|||||||
@@ -8,12 +8,17 @@ filein = -I/usr/include/ -I${MKLROOT}/include
|
|||||||
|
|
||||||
## Using sequential MKL (OpenMP disabled for better single-threaded performance)
|
## Using sequential MKL (OpenMP disabled for better single-threaded performance)
|
||||||
## Added -lifcore for Intel Fortran runtime and -limf for Intel math library
|
## Added -lifcore for Intel Fortran runtime and -limf for Intel math library
|
||||||
LDLIBS = -L${MKLROOT}/lib -lmkl_intel_lp64 -lmkl_sequential -lmkl_core -lifcore -limf -lpthread -lm -ldl -liomp5
|
LDLIBS = -L${MKLROOT}/lib -lmkl_intel_lp64 -lmkl_sequential -lmkl_core -lifcore -limf -lpthread -lm -ldl
|
||||||
|
|
||||||
## PGO build mode switch (ABE only; TwoPunctureABE always uses opt flags)
|
## Aggressive optimization flags:
|
||||||
## opt : (default) maximum performance with PGO profile-guided optimization
|
## -O3: Maximum optimization
|
||||||
## instrument : PGO Phase 1 instrumentation to collect fresh profile data
|
## -xHost: Optimize for the host CPU architecture (Intel/AMD compatible)
|
||||||
PGO_MODE ?= opt
|
## -fp-model fast=2: Aggressive floating-point optimizations
|
||||||
|
## -fma: Enable fused multiply-add instructions
|
||||||
|
CXXAPPFLAGS = -O3 -xHost -fp-model fast=2 -fma -ipo \
|
||||||
|
-Dfortran3 -Dnewc -I${MKLROOT}/include
|
||||||
|
f90appflags = -O3 -xHost -fp-model fast=2 -fma -ipo \
|
||||||
|
-align array64byte -fpp -I${MKLROOT}/include
|
||||||
f90 = ifx
|
f90 = ifx
|
||||||
f77 = ifx
|
f77 = ifx
|
||||||
CXX = icpx
|
CXX = icpx
|
||||||
|
|||||||
@@ -1,146 +0,0 @@
|
|||||||
#ifndef SHARE_FUNC_H
|
|
||||||
#define SHARE_FUNC_H
|
|
||||||
|
|
||||||
#include <stdlib.h>
|
|
||||||
#include <stddef.h>
|
|
||||||
#include <math.h>
|
|
||||||
#include <stdio.h>
|
|
||||||
/* 主网格:0-based -> 1D */
|
|
||||||
static inline size_t idx_ex(int i0, int j0, int k0, const int ex[3]) {
|
|
||||||
const int ex1 = ex[0], ex2 = ex[1];
|
|
||||||
return (size_t)i0 + (size_t)j0 * (size_t)ex1 + (size_t)k0 * (size_t)ex1 * (size_t)ex2;
|
|
||||||
}
|
|
||||||
|
|
||||||
/*
|
|
||||||
* fh 对应 Fortran: fh(-1:ex1, -1:ex2, -1:ex3)
|
|
||||||
* ord=2 => shift=1
|
|
||||||
* iF/jF/kF 为 Fortran 索引(可为 -1,0,1..ex)
|
|
||||||
*/
|
|
||||||
static inline size_t idx_fh_F_ord2(int iF, int jF, int kF, const int ex[3]) {
|
|
||||||
const int shift = 1;
|
|
||||||
const int nx = ex[0] + 2; // ex1 + ord
|
|
||||||
const int ny = ex[1] + 2;
|
|
||||||
|
|
||||||
const int ii = iF + shift; // 0..ex1+1
|
|
||||||
const int jj = jF + shift; // 0..ex2+1
|
|
||||||
const int kk = kF + shift; // 0..ex3+1
|
|
||||||
|
|
||||||
return (size_t)ii + (size_t)jj * (size_t)nx + (size_t)kk * (size_t)nx * (size_t)ny;
|
|
||||||
}
|
|
||||||
|
|
||||||
/*
|
|
||||||
* fh 对应 Fortran: fh(-2:ex1, -2:ex2, -2:ex3)
|
|
||||||
* ord=3 => shift=2
|
|
||||||
* iF/jF/kF 是 Fortran 索引(可为负)
|
|
||||||
*/
|
|
||||||
static inline size_t idx_fh_F(int iF, int jF, int kF, const int ex[3]) {
|
|
||||||
const int shift = 2; // ord=3 -> -2..ex
|
|
||||||
const int nx = ex[0] + 3; // ex1 + ord
|
|
||||||
const int ny = ex[1] + 3;
|
|
||||||
|
|
||||||
const int ii = iF + shift; // 0..ex1+2
|
|
||||||
const int jj = jF + shift; // 0..ex2+2
|
|
||||||
const int kk = kF + shift; // 0..ex3+2
|
|
||||||
|
|
||||||
return (size_t)ii + (size_t)jj * (size_t)nx + (size_t)kk * (size_t)nx * (size_t)ny;
|
|
||||||
}
|
|
||||||
|
|
||||||
/*
|
|
||||||
* func: (1..extc1, 1..extc2, 1..extc3) 1-based in Fortran
|
|
||||||
* funcc: (-ord+1..extc1, -ord+1..extc2, -ord+1..extc3) in Fortran
|
|
||||||
*
|
|
||||||
* C 里我们把:
|
|
||||||
* func 视为 0-based: i0=0..extc1-1, j0=0..extc2-1, k0=0..extc3-1
|
|
||||||
* funcc 用“平移下标”存为一维数组:
|
|
||||||
* iF in [-ord+1..extc1] -> ii = iF + (ord-1) in [0..extc1+ord-1]
|
|
||||||
* 总长度 nx = extc1 + ord
|
|
||||||
* 同理 ny = extc2 + ord, nz = extc3 + ord
|
|
||||||
*/
|
|
||||||
|
|
||||||
static inline size_t idx_func0(int i0, int j0, int k0, const int extc[3]) {
|
|
||||||
const int nx = extc[0], ny = extc[1];
|
|
||||||
return (size_t)i0 + (size_t)j0 * (size_t)nx + (size_t)k0 * (size_t)nx * (size_t)ny;
|
|
||||||
}
|
|
||||||
|
|
||||||
static inline size_t idx_funcc_F(int iF, int jF, int kF, int ord, const int extc[3]) {
|
|
||||||
const int shift = ord - 1; // iF = -shift .. extc1
|
|
||||||
const int nx = extc[0] + ord; // [-shift..extc1] 共 extc1+ord 个
|
|
||||||
const int ny = extc[1] + ord;
|
|
||||||
|
|
||||||
const int ii = iF + shift; // 0..extc1+shift
|
|
||||||
const int jj = jF + shift; // 0..extc2+shift
|
|
||||||
const int kk = kF + shift; // 0..extc3+shift
|
|
||||||
|
|
||||||
return (size_t)ii + (size_t)jj * (size_t)nx + (size_t)kk * (size_t)nx * (size_t)ny;
|
|
||||||
}
|
|
||||||
|
|
||||||
/*
|
|
||||||
* 等价于 Fortran:
|
|
||||||
* funcc(1:extc1,1:extc2,1:extc3)=func
|
|
||||||
* do i=0,ord-1
|
|
||||||
* funcc(-i,1:extc2,1:extc3) = funcc(i+1,1:extc2,1:extc3)*SoA(1)
|
|
||||||
* enddo
|
|
||||||
* do i=0,ord-1
|
|
||||||
* funcc(:,-i,1:extc3) = funcc(:,i+1,1:extc3)*SoA(2)
|
|
||||||
* enddo
|
|
||||||
* do i=0,ord-1
|
|
||||||
* funcc(:,:,-i) = funcc(:,:,i+1)*SoA(3)
|
|
||||||
* enddo
|
|
||||||
*/
|
|
||||||
static inline void symmetry_bd(int ord,
|
|
||||||
const int extc[3],
|
|
||||||
const double *func,
|
|
||||||
double *funcc,
|
|
||||||
const double SoA[3])
|
|
||||||
{
|
|
||||||
const int extc1 = extc[0], extc2 = extc[1], extc3 = extc[2];
|
|
||||||
|
|
||||||
// 1) funcc(1:extc1,1:extc2,1:extc3) = func
|
|
||||||
// Fortran 的 (iF=1..extc1) 对应 C 的 func(i0=0..extc1-1)
|
|
||||||
for (int k0 = 0; k0 < extc3; ++k0) {
|
|
||||||
for (int j0 = 0; j0 < extc2; ++j0) {
|
|
||||||
for (int i0 = 0; i0 < extc1; ++i0) {
|
|
||||||
const int iF = i0 + 1, jF = j0 + 1, kF = k0 + 1;
|
|
||||||
funcc[idx_funcc_F(iF, jF, kF, ord, extc)] = func[idx_func0(i0, j0, k0, extc)];
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// 2) do i=0..ord-1: funcc(-i, 1:extc2, 1:extc3) = funcc(i+1, ...)*SoA(1)
|
|
||||||
for (int ii = 0; ii <= ord - 1; ++ii) {
|
|
||||||
const int iF_dst = -ii; // 0, -1, -2, ...
|
|
||||||
const int iF_src = ii + 1; // 1, 2, 3, ...
|
|
||||||
for (int kF = 1; kF <= extc3; ++kF) {
|
|
||||||
for (int jF = 1; jF <= extc2; ++jF) {
|
|
||||||
funcc[idx_funcc_F(iF_dst, jF, kF, ord, extc)] =
|
|
||||||
funcc[idx_funcc_F(iF_src, jF, kF, ord, extc)] * SoA[0];
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// 3) do i=0..ord-1: funcc(:,-i, 1:extc3) = funcc(:, i+1, 1:extc3)*SoA(2)
|
|
||||||
// 注意 Fortran 这里的 ":" 表示 iF 从 (-ord+1..extc1) 全覆盖
|
|
||||||
for (int jj = 0; jj <= ord - 1; ++jj) {
|
|
||||||
const int jF_dst = -jj;
|
|
||||||
const int jF_src = jj + 1;
|
|
||||||
for (int kF = 1; kF <= extc3; ++kF) {
|
|
||||||
for (int iF = -ord + 1; iF <= extc1; ++iF) {
|
|
||||||
funcc[idx_funcc_F(iF, jF_dst, kF, ord, extc)] =
|
|
||||||
funcc[idx_funcc_F(iF, jF_src, kF, ord, extc)] * SoA[1];
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// 4) do i=0..ord-1: funcc(:,:,-i) = funcc(:,:, i+1)*SoA(3)
|
|
||||||
for (int kk = 0; kk <= ord - 1; ++kk) {
|
|
||||||
const int kF_dst = -kk;
|
|
||||||
const int kF_src = kk + 1;
|
|
||||||
for (int jF = -ord + 1; jF <= extc2; ++jF) {
|
|
||||||
for (int iF = -ord + 1; iF <= extc1; ++iF) {
|
|
||||||
funcc[idx_funcc_F(iF, jF, kF_dst, ord, extc)] =
|
|
||||||
funcc[idx_funcc_F(iF, jF, kF_src, ord, extc)] * SoA[2];
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
#endif
|
|
||||||
@@ -220,9 +220,16 @@ void surface_integral::surf_Wave(double rex, int lev, cgh *GH, var *Rpsi4, var *
|
|||||||
pox[2][n] = rex * nz_g[n];
|
pox[2][n] = rex * nz_g[n];
|
||||||
}
|
}
|
||||||
|
|
||||||
|
double *shellf;
|
||||||
|
shellf = new double[n_tot * InList];
|
||||||
|
|
||||||
|
GH->PatL[lev]->data->Interp_Points(DG_List, n_tot, pox, shellf, Symmetry);
|
||||||
|
|
||||||
int mp, Lp, Nmin, Nmax;
|
int mp, Lp, Nmin, Nmax;
|
||||||
|
|
||||||
mp = n_tot / cpusize;
|
mp = n_tot / cpusize;
|
||||||
Lp = n_tot - cpusize * mp;
|
Lp = n_tot - cpusize * mp;
|
||||||
|
|
||||||
if (Lp > myrank)
|
if (Lp > myrank)
|
||||||
{
|
{
|
||||||
Nmin = myrank * mp + myrank;
|
Nmin = myrank * mp + myrank;
|
||||||
@@ -234,11 +241,6 @@ void surface_integral::surf_Wave(double rex, int lev, cgh *GH, var *Rpsi4, var *
|
|||||||
Nmax = Nmin + mp - 1;
|
Nmax = Nmin + mp - 1;
|
||||||
}
|
}
|
||||||
|
|
||||||
double *shellf;
|
|
||||||
shellf = new double[n_tot * InList];
|
|
||||||
|
|
||||||
GH->PatL[lev]->data->Interp_Points(DG_List, n_tot, pox, shellf, Symmetry, Nmin, Nmax);
|
|
||||||
|
|
||||||
//|~~~~~> Integrate the dot product of Dphi with the surface normal.
|
//|~~~~~> Integrate the dot product of Dphi with the surface normal.
|
||||||
|
|
||||||
double *RP_out, *IP_out;
|
double *RP_out, *IP_out;
|
||||||
@@ -361,17 +363,8 @@ void surface_integral::surf_Wave(double rex, int lev, cgh *GH, var *Rpsi4, var *
|
|||||||
}
|
}
|
||||||
//|------+ Communicate and sum the results from each processor.
|
//|------+ Communicate and sum the results from each processor.
|
||||||
|
|
||||||
{
|
MPI_Allreduce(RP_out, RP, NN, MPI_DOUBLE, MPI_SUM, MPI_COMM_WORLD);
|
||||||
double *RPIP_out = new double[2 * NN];
|
MPI_Allreduce(IP_out, IP, NN, MPI_DOUBLE, MPI_SUM, MPI_COMM_WORLD);
|
||||||
double *RPIP = new double[2 * NN];
|
|
||||||
memcpy(RPIP_out, RP_out, NN * sizeof(double));
|
|
||||||
memcpy(RPIP_out + NN, IP_out, NN * sizeof(double));
|
|
||||||
MPI_Allreduce(RPIP_out, RPIP, 2 * NN, MPI_DOUBLE, MPI_SUM, MPI_COMM_WORLD);
|
|
||||||
memcpy(RP, RPIP, NN * sizeof(double));
|
|
||||||
memcpy(IP, RPIP + NN, NN * sizeof(double));
|
|
||||||
delete[] RPIP_out;
|
|
||||||
delete[] RPIP;
|
|
||||||
}
|
|
||||||
|
|
||||||
//|------= Free memory.
|
//|------= Free memory.
|
||||||
|
|
||||||
@@ -563,17 +556,8 @@ void surface_integral::surf_Wave(double rex, int lev, cgh *GH, var *Rpsi4, var *
|
|||||||
}
|
}
|
||||||
//|------+ Communicate and sum the results from each processor.
|
//|------+ Communicate and sum the results from each processor.
|
||||||
|
|
||||||
{
|
MPI_Allreduce(RP_out, RP, NN, MPI_DOUBLE, MPI_SUM, Comm_here);
|
||||||
double *RPIP_out = new double[2 * NN];
|
MPI_Allreduce(IP_out, IP, NN, MPI_DOUBLE, MPI_SUM, Comm_here);
|
||||||
double *RPIP = new double[2 * NN];
|
|
||||||
memcpy(RPIP_out, RP_out, NN * sizeof(double));
|
|
||||||
memcpy(RPIP_out + NN, IP_out, NN * sizeof(double));
|
|
||||||
MPI_Allreduce(RPIP_out, RPIP, 2 * NN, MPI_DOUBLE, MPI_SUM, Comm_here);
|
|
||||||
memcpy(RP, RPIP, NN * sizeof(double));
|
|
||||||
memcpy(IP, RPIP + NN, NN * sizeof(double));
|
|
||||||
delete[] RPIP_out;
|
|
||||||
delete[] RPIP;
|
|
||||||
}
|
|
||||||
|
|
||||||
//|------= Free memory.
|
//|------= Free memory.
|
||||||
|
|
||||||
@@ -751,17 +735,8 @@ void surface_integral::surf_Wave(double rex, int lev, ShellPatch *GH, var *Rpsi4
|
|||||||
}
|
}
|
||||||
//|------+ Communicate and sum the results from each processor.
|
//|------+ Communicate and sum the results from each processor.
|
||||||
|
|
||||||
{
|
MPI_Allreduce(RP_out, RP, NN, MPI_DOUBLE, MPI_SUM, MPI_COMM_WORLD);
|
||||||
double *RPIP_out = new double[2 * NN];
|
MPI_Allreduce(IP_out, IP, NN, MPI_DOUBLE, MPI_SUM, MPI_COMM_WORLD);
|
||||||
double *RPIP = new double[2 * NN];
|
|
||||||
memcpy(RPIP_out, RP_out, NN * sizeof(double));
|
|
||||||
memcpy(RPIP_out + NN, IP_out, NN * sizeof(double));
|
|
||||||
MPI_Allreduce(RPIP_out, RPIP, 2 * NN, MPI_DOUBLE, MPI_SUM, MPI_COMM_WORLD);
|
|
||||||
memcpy(RP, RPIP, NN * sizeof(double));
|
|
||||||
memcpy(IP, RPIP + NN, NN * sizeof(double));
|
|
||||||
delete[] RPIP_out;
|
|
||||||
delete[] RPIP;
|
|
||||||
}
|
|
||||||
|
|
||||||
//|------= Free memory.
|
//|------= Free memory.
|
||||||
|
|
||||||
@@ -1009,17 +984,8 @@ void surface_integral::surf_Wave(double rex, int lev, ShellPatch *GH,
|
|||||||
}
|
}
|
||||||
//|------+ Communicate and sum the results from each processor.
|
//|------+ Communicate and sum the results from each processor.
|
||||||
|
|
||||||
{
|
MPI_Allreduce(RP_out, RP, NN, MPI_DOUBLE, MPI_SUM, MPI_COMM_WORLD);
|
||||||
double *RPIP_out = new double[2 * NN];
|
MPI_Allreduce(IP_out, IP, NN, MPI_DOUBLE, MPI_SUM, MPI_COMM_WORLD);
|
||||||
double *RPIP = new double[2 * NN];
|
|
||||||
memcpy(RPIP_out, RP_out, NN * sizeof(double));
|
|
||||||
memcpy(RPIP_out + NN, IP_out, NN * sizeof(double));
|
|
||||||
MPI_Allreduce(RPIP_out, RPIP, 2 * NN, MPI_DOUBLE, MPI_SUM, MPI_COMM_WORLD);
|
|
||||||
memcpy(RP, RPIP, NN * sizeof(double));
|
|
||||||
memcpy(IP, RPIP + NN, NN * sizeof(double));
|
|
||||||
delete[] RPIP_out;
|
|
||||||
delete[] RPIP;
|
|
||||||
}
|
|
||||||
|
|
||||||
//|------= Free memory.
|
//|------= Free memory.
|
||||||
|
|
||||||
@@ -1453,17 +1419,8 @@ void surface_integral::surf_Wave(double rex, int lev, ShellPatch *GH,
|
|||||||
}
|
}
|
||||||
//|------+ Communicate and sum the results from each processor.
|
//|------+ Communicate and sum the results from each processor.
|
||||||
|
|
||||||
{
|
MPI_Allreduce(RP_out, RP, NN, MPI_DOUBLE, MPI_SUM, MPI_COMM_WORLD);
|
||||||
double *RPIP_out = new double[2 * NN];
|
MPI_Allreduce(IP_out, IP, NN, MPI_DOUBLE, MPI_SUM, MPI_COMM_WORLD);
|
||||||
double *RPIP = new double[2 * NN];
|
|
||||||
memcpy(RPIP_out, RP_out, NN * sizeof(double));
|
|
||||||
memcpy(RPIP_out + NN, IP_out, NN * sizeof(double));
|
|
||||||
MPI_Allreduce(RPIP_out, RPIP, 2 * NN, MPI_DOUBLE, MPI_SUM, MPI_COMM_WORLD);
|
|
||||||
memcpy(RP, RPIP, NN * sizeof(double));
|
|
||||||
memcpy(IP, RPIP + NN, NN * sizeof(double));
|
|
||||||
delete[] RPIP_out;
|
|
||||||
delete[] RPIP;
|
|
||||||
}
|
|
||||||
|
|
||||||
//|------= Free memory.
|
//|------= Free memory.
|
||||||
|
|
||||||
@@ -1897,17 +1854,8 @@ void surface_integral::surf_Wave(double rex, int lev, cgh *GH,
|
|||||||
}
|
}
|
||||||
//|------+ Communicate and sum the results from each processor.
|
//|------+ Communicate and sum the results from each processor.
|
||||||
|
|
||||||
{
|
MPI_Allreduce(RP_out, RP, NN, MPI_DOUBLE, MPI_SUM, MPI_COMM_WORLD);
|
||||||
double *RPIP_out = new double[2 * NN];
|
MPI_Allreduce(IP_out, IP, NN, MPI_DOUBLE, MPI_SUM, MPI_COMM_WORLD);
|
||||||
double *RPIP = new double[2 * NN];
|
|
||||||
memcpy(RPIP_out, RP_out, NN * sizeof(double));
|
|
||||||
memcpy(RPIP_out + NN, IP_out, NN * sizeof(double));
|
|
||||||
MPI_Allreduce(RPIP_out, RPIP, 2 * NN, MPI_DOUBLE, MPI_SUM, MPI_COMM_WORLD);
|
|
||||||
memcpy(RP, RPIP, NN * sizeof(double));
|
|
||||||
memcpy(IP, RPIP + NN, NN * sizeof(double));
|
|
||||||
delete[] RPIP_out;
|
|
||||||
delete[] RPIP;
|
|
||||||
}
|
|
||||||
|
|
||||||
//|------= Free memory.
|
//|------= Free memory.
|
||||||
|
|
||||||
@@ -2092,17 +2040,8 @@ void surface_integral::surf_Wave(double rex, int lev, NullShellPatch2 *GH, var *
|
|||||||
}
|
}
|
||||||
//|------+ Communicate and sum the results from each processor.
|
//|------+ Communicate and sum the results from each processor.
|
||||||
|
|
||||||
{
|
MPI_Allreduce(RP_out, RP, NN, MPI_DOUBLE, MPI_SUM, MPI_COMM_WORLD);
|
||||||
double *RPIP_out = new double[2 * NN];
|
MPI_Allreduce(IP_out, IP, NN, MPI_DOUBLE, MPI_SUM, MPI_COMM_WORLD);
|
||||||
double *RPIP = new double[2 * NN];
|
|
||||||
memcpy(RPIP_out, RP_out, NN * sizeof(double));
|
|
||||||
memcpy(RPIP_out + NN, IP_out, NN * sizeof(double));
|
|
||||||
MPI_Allreduce(RPIP_out, RPIP, 2 * NN, MPI_DOUBLE, MPI_SUM, MPI_COMM_WORLD);
|
|
||||||
memcpy(RP, RPIP, NN * sizeof(double));
|
|
||||||
memcpy(IP, RPIP + NN, NN * sizeof(double));
|
|
||||||
delete[] RPIP_out;
|
|
||||||
delete[] RPIP;
|
|
||||||
}
|
|
||||||
|
|
||||||
//|------= Free memory.
|
//|------= Free memory.
|
||||||
|
|
||||||
@@ -2287,17 +2226,8 @@ void surface_integral::surf_Wave(double rex, int lev, NullShellPatch *GH, var *R
|
|||||||
}
|
}
|
||||||
//|------+ Communicate and sum the results from each processor.
|
//|------+ Communicate and sum the results from each processor.
|
||||||
|
|
||||||
{
|
MPI_Allreduce(RP_out, RP, NN, MPI_DOUBLE, MPI_SUM, MPI_COMM_WORLD);
|
||||||
double *RPIP_out = new double[2 * NN];
|
MPI_Allreduce(IP_out, IP, NN, MPI_DOUBLE, MPI_SUM, MPI_COMM_WORLD);
|
||||||
double *RPIP = new double[2 * NN];
|
|
||||||
memcpy(RPIP_out, RP_out, NN * sizeof(double));
|
|
||||||
memcpy(RPIP_out + NN, IP_out, NN * sizeof(double));
|
|
||||||
MPI_Allreduce(RPIP_out, RPIP, 2 * NN, MPI_DOUBLE, MPI_SUM, MPI_COMM_WORLD);
|
|
||||||
memcpy(RP, RPIP, NN * sizeof(double));
|
|
||||||
memcpy(IP, RPIP + NN, NN * sizeof(double));
|
|
||||||
delete[] RPIP_out;
|
|
||||||
delete[] RPIP;
|
|
||||||
}
|
|
||||||
|
|
||||||
//|------= Free memory.
|
//|------= Free memory.
|
||||||
|
|
||||||
@@ -2384,9 +2314,25 @@ void surface_integral::surf_MassPAng(double rex, int lev, cgh *GH, var *chi, var
|
|||||||
pox[2][n] = rex * nz_g[n];
|
pox[2][n] = rex * nz_g[n];
|
||||||
}
|
}
|
||||||
|
|
||||||
|
double *shellf;
|
||||||
|
shellf = new double[n_tot * InList];
|
||||||
|
|
||||||
|
// we have assumed there is only one box on this level,
|
||||||
|
// so we do not need loop boxes
|
||||||
|
GH->PatL[lev]->data->Interp_Points(DG_List, n_tot, pox, shellf, Symmetry);
|
||||||
|
|
||||||
|
double Mass_out = 0;
|
||||||
|
double ang_outx, ang_outy, ang_outz;
|
||||||
|
double p_outx, p_outy, p_outz;
|
||||||
|
ang_outx = ang_outy = ang_outz = 0.0;
|
||||||
|
p_outx = p_outy = p_outz = 0.0;
|
||||||
|
const double f1o8 = 0.125;
|
||||||
|
|
||||||
int mp, Lp, Nmin, Nmax;
|
int mp, Lp, Nmin, Nmax;
|
||||||
|
|
||||||
mp = n_tot / cpusize;
|
mp = n_tot / cpusize;
|
||||||
Lp = n_tot - cpusize * mp;
|
Lp = n_tot - cpusize * mp;
|
||||||
|
|
||||||
if (Lp > myrank)
|
if (Lp > myrank)
|
||||||
{
|
{
|
||||||
Nmin = myrank * mp + myrank;
|
Nmin = myrank * mp + myrank;
|
||||||
@@ -2398,20 +2344,6 @@ void surface_integral::surf_MassPAng(double rex, int lev, cgh *GH, var *chi, var
|
|||||||
Nmax = Nmin + mp - 1;
|
Nmax = Nmin + mp - 1;
|
||||||
}
|
}
|
||||||
|
|
||||||
double *shellf;
|
|
||||||
shellf = new double[n_tot * InList];
|
|
||||||
|
|
||||||
// we have assumed there is only one box on this level,
|
|
||||||
// so we do not need loop boxes
|
|
||||||
GH->PatL[lev]->data->Interp_Points(DG_List, n_tot, pox, shellf, Symmetry, Nmin, Nmax);
|
|
||||||
|
|
||||||
double Mass_out = 0;
|
|
||||||
double ang_outx, ang_outy, ang_outz;
|
|
||||||
double p_outx, p_outy, p_outz;
|
|
||||||
ang_outx = ang_outy = ang_outz = 0.0;
|
|
||||||
p_outx = p_outy = p_outz = 0.0;
|
|
||||||
const double f1o8 = 0.125;
|
|
||||||
|
|
||||||
double Chi, Psi;
|
double Chi, Psi;
|
||||||
double Gxx, Gxy, Gxz, Gyy, Gyz, Gzz;
|
double Gxx, Gxy, Gxz, Gyy, Gyz, Gzz;
|
||||||
double gupxx, gupxy, gupxz, gupyy, gupyz, gupzz;
|
double gupxx, gupxy, gupxz, gupyy, gupyz, gupzz;
|
||||||
@@ -2532,13 +2464,15 @@ void surface_integral::surf_MassPAng(double rex, int lev, cgh *GH, var *chi, var
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
{
|
MPI_Allreduce(&Mass_out, &mass, 1, MPI_DOUBLE, MPI_SUM, MPI_COMM_WORLD);
|
||||||
double scalar_out[7] = {Mass_out, ang_outx, ang_outy, ang_outz, p_outx, p_outy, p_outz};
|
|
||||||
double scalar_in[7];
|
MPI_Allreduce(&ang_outx, &sx, 1, MPI_DOUBLE, MPI_SUM, MPI_COMM_WORLD);
|
||||||
MPI_Allreduce(scalar_out, scalar_in, 7, MPI_DOUBLE, MPI_SUM, MPI_COMM_WORLD);
|
MPI_Allreduce(&ang_outy, &sy, 1, MPI_DOUBLE, MPI_SUM, MPI_COMM_WORLD);
|
||||||
mass = scalar_in[0]; sx = scalar_in[1]; sy = scalar_in[2]; sz = scalar_in[3];
|
MPI_Allreduce(&ang_outz, &sz, 1, MPI_DOUBLE, MPI_SUM, MPI_COMM_WORLD);
|
||||||
px = scalar_in[4]; py = scalar_in[5]; pz = scalar_in[6];
|
|
||||||
}
|
MPI_Allreduce(&p_outx, &px, 1, MPI_DOUBLE, MPI_SUM, MPI_COMM_WORLD);
|
||||||
|
MPI_Allreduce(&p_outy, &py, 1, MPI_DOUBLE, MPI_SUM, MPI_COMM_WORLD);
|
||||||
|
MPI_Allreduce(&p_outz, &pz, 1, MPI_DOUBLE, MPI_SUM, MPI_COMM_WORLD);
|
||||||
|
|
||||||
#ifdef GaussInt
|
#ifdef GaussInt
|
||||||
mass = mass * rex * rex * dphi * factor;
|
mass = mass * rex * rex * dphi * factor;
|
||||||
@@ -2801,13 +2735,15 @@ void surface_integral::surf_MassPAng(double rex, int lev, cgh *GH, var *chi, var
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
{
|
MPI_Allreduce(&Mass_out, &mass, 1, MPI_DOUBLE, MPI_SUM, Comm_here);
|
||||||
double scalar_out[7] = {Mass_out, ang_outx, ang_outy, ang_outz, p_outx, p_outy, p_outz};
|
|
||||||
double scalar_in[7];
|
MPI_Allreduce(&ang_outx, &sx, 1, MPI_DOUBLE, MPI_SUM, Comm_here);
|
||||||
MPI_Allreduce(scalar_out, scalar_in, 7, MPI_DOUBLE, MPI_SUM, Comm_here);
|
MPI_Allreduce(&ang_outy, &sy, 1, MPI_DOUBLE, MPI_SUM, Comm_here);
|
||||||
mass = scalar_in[0]; sx = scalar_in[1]; sy = scalar_in[2]; sz = scalar_in[3];
|
MPI_Allreduce(&ang_outz, &sz, 1, MPI_DOUBLE, MPI_SUM, Comm_here);
|
||||||
px = scalar_in[4]; py = scalar_in[5]; pz = scalar_in[6];
|
|
||||||
}
|
MPI_Allreduce(&p_outx, &px, 1, MPI_DOUBLE, MPI_SUM, Comm_here);
|
||||||
|
MPI_Allreduce(&p_outy, &py, 1, MPI_DOUBLE, MPI_SUM, Comm_here);
|
||||||
|
MPI_Allreduce(&p_outz, &pz, 1, MPI_DOUBLE, MPI_SUM, Comm_here);
|
||||||
|
|
||||||
#ifdef GaussInt
|
#ifdef GaussInt
|
||||||
mass = mass * rex * rex * dphi * factor;
|
mass = mass * rex * rex * dphi * factor;
|
||||||
@@ -3084,13 +3020,15 @@ void surface_integral::surf_MassPAng(double rex, int lev, ShellPatch *GH, var *c
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
{
|
MPI_Allreduce(&Mass_out, &mass, 1, MPI_DOUBLE, MPI_SUM, MPI_COMM_WORLD);
|
||||||
double scalar_out[7] = {Mass_out, ang_outx, ang_outy, ang_outz, p_outx, p_outy, p_outz};
|
|
||||||
double scalar_in[7];
|
MPI_Allreduce(&ang_outx, &sx, 1, MPI_DOUBLE, MPI_SUM, MPI_COMM_WORLD);
|
||||||
MPI_Allreduce(scalar_out, scalar_in, 7, MPI_DOUBLE, MPI_SUM, MPI_COMM_WORLD);
|
MPI_Allreduce(&ang_outy, &sy, 1, MPI_DOUBLE, MPI_SUM, MPI_COMM_WORLD);
|
||||||
mass = scalar_in[0]; sx = scalar_in[1]; sy = scalar_in[2]; sz = scalar_in[3];
|
MPI_Allreduce(&ang_outz, &sz, 1, MPI_DOUBLE, MPI_SUM, MPI_COMM_WORLD);
|
||||||
px = scalar_in[4]; py = scalar_in[5]; pz = scalar_in[6];
|
|
||||||
}
|
MPI_Allreduce(&p_outx, &px, 1, MPI_DOUBLE, MPI_SUM, MPI_COMM_WORLD);
|
||||||
|
MPI_Allreduce(&p_outy, &py, 1, MPI_DOUBLE, MPI_SUM, MPI_COMM_WORLD);
|
||||||
|
MPI_Allreduce(&p_outz, &pz, 1, MPI_DOUBLE, MPI_SUM, MPI_COMM_WORLD);
|
||||||
|
|
||||||
#ifdef GaussInt
|
#ifdef GaussInt
|
||||||
mass = mass * rex * rex * dphi * factor;
|
mass = mass * rex * rex * dphi * factor;
|
||||||
@@ -3669,17 +3607,8 @@ void surface_integral::surf_Wave(double rex, cgh *GH, ShellPatch *SH,
|
|||||||
}
|
}
|
||||||
//|------+ Communicate and sum the results from each processor.
|
//|------+ Communicate and sum the results from each processor.
|
||||||
|
|
||||||
{
|
MPI_Allreduce(RP_out, RP, NN, MPI_DOUBLE, MPI_SUM, MPI_COMM_WORLD);
|
||||||
double *RPIP_out = new double[2 * NN];
|
MPI_Allreduce(IP_out, IP, NN, MPI_DOUBLE, MPI_SUM, MPI_COMM_WORLD);
|
||||||
double *RPIP = new double[2 * NN];
|
|
||||||
memcpy(RPIP_out, RP_out, NN * sizeof(double));
|
|
||||||
memcpy(RPIP_out + NN, IP_out, NN * sizeof(double));
|
|
||||||
MPI_Allreduce(RPIP_out, RPIP, 2 * NN, MPI_DOUBLE, MPI_SUM, MPI_COMM_WORLD);
|
|
||||||
memcpy(RP, RPIP, NN * sizeof(double));
|
|
||||||
memcpy(IP, RPIP + NN, NN * sizeof(double));
|
|
||||||
delete[] RPIP_out;
|
|
||||||
delete[] RPIP;
|
|
||||||
}
|
|
||||||
|
|
||||||
//|------= Free memory.
|
//|------= Free memory.
|
||||||
|
|
||||||
|
|||||||
@@ -1,27 +0,0 @@
|
|||||||
#include "share_func.h"
|
|
||||||
void fdderivs(const int ex[3],
|
|
||||||
const double *f,
|
|
||||||
double *fxx, double *fxy, double *fxz,
|
|
||||||
double *fyy, double *fyz, double *fzz,
|
|
||||||
const double *X, const double *Y, const double *Z,
|
|
||||||
double SYM1, double SYM2, double SYM3,
|
|
||||||
int Symmetry, int onoff);
|
|
||||||
|
|
||||||
void fderivs(const int ex[3],
|
|
||||||
const double *f,
|
|
||||||
double *fx, double *fy, double *fz,
|
|
||||||
const double *X, const double *Y, const double *Z,
|
|
||||||
double SYM1, double SYM2, double SYM3,
|
|
||||||
int Symmetry, int onoff);
|
|
||||||
|
|
||||||
void kodis(const int ex[3],
|
|
||||||
const double *X, const double *Y, const double *Z,
|
|
||||||
const double *f, double *f_rhs,
|
|
||||||
const double SoA[3],
|
|
||||||
int Symmetry, double eps);
|
|
||||||
|
|
||||||
void lopsided(const int ex[3],
|
|
||||||
const double *X, const double *Y, const double *Z,
|
|
||||||
const double *f, double *f_rhs,
|
|
||||||
const double *Sfx, const double *Sfy, const double *Sfz,
|
|
||||||
int Symmetry, const double SoA[3]);
|
|
||||||
@@ -11,46 +11,16 @@
|
|||||||
import AMSS_NCKU_Input as input_data
|
import AMSS_NCKU_Input as input_data
|
||||||
import subprocess
|
import subprocess
|
||||||
import time
|
import time
|
||||||
|
## CPU core binding configuration using taskset
|
||||||
|
## taskset ensures all child processes inherit the CPU affinity mask
|
||||||
|
## This forces make and all compiler processes to use only nohz_full cores (4-55, 60-111)
|
||||||
|
## Format: taskset -c 4-55,60-111 ensures processes only run on these cores
|
||||||
|
NUMACTL_CPU_BIND = "taskset -c 0-111"
|
||||||
|
|
||||||
|
## Build parallelism configuration
|
||||||
def get_last_n_cores_per_socket(n=32):
|
## Use nohz_full cores (4-55, 60-111) for compilation: 52 + 52 = 104 cores
|
||||||
"""
|
## Set make -j to utilize available cores for faster builds
|
||||||
Read CPU topology via lscpu and return a taskset -c string
|
BUILD_JOBS = 104
|
||||||
selecting the last `n` cores of each NUMA node (socket).
|
|
||||||
|
|
||||||
Example: 2 sockets x 56 cores each, n=32 -> node0: 24-55, node1: 80-111
|
|
||||||
-> "taskset -c 24-55,80-111"
|
|
||||||
"""
|
|
||||||
result = subprocess.run(["lscpu", "--parse=NODE,CPU"], capture_output=True, text=True)
|
|
||||||
|
|
||||||
# Build a dict: node_id -> sorted list of CPU ids
|
|
||||||
node_cpus = {}
|
|
||||||
for line in result.stdout.splitlines():
|
|
||||||
if line.startswith("#") or not line.strip():
|
|
||||||
continue
|
|
||||||
parts = line.split(",")
|
|
||||||
if len(parts) < 2:
|
|
||||||
continue
|
|
||||||
node_id, cpu_id = int(parts[0]), int(parts[1])
|
|
||||||
node_cpus.setdefault(node_id, []).append(cpu_id)
|
|
||||||
|
|
||||||
segments = []
|
|
||||||
for node_id in sorted(node_cpus):
|
|
||||||
cpus = sorted(node_cpus[node_id])
|
|
||||||
selected = cpus[-n:] # last n cores of this socket
|
|
||||||
segments.append(f"{selected[0]}-{selected[-1]}")
|
|
||||||
|
|
||||||
cpu_str = ",".join(segments)
|
|
||||||
total = len(segments) * n
|
|
||||||
print(f" CPU binding: taskset -c {cpu_str} ({total} cores, last {n} per socket)")
|
|
||||||
return f"taskset -c {cpu_str}"
|
|
||||||
|
|
||||||
|
|
||||||
## CPU core binding: dynamically select the last 32 cores of each socket (64 cores total)
|
|
||||||
NUMACTL_CPU_BIND = get_last_n_cores_per_socket(n=32)
|
|
||||||
|
|
||||||
## Build parallelism: match the number of bound cores
|
|
||||||
BUILD_JOBS = 64
|
|
||||||
|
|
||||||
|
|
||||||
##################################################################
|
##################################################################
|
||||||
@@ -147,7 +117,6 @@ def run_ABE():
|
|||||||
|
|
||||||
if (input_data.GPU_Calculation == "no"):
|
if (input_data.GPU_Calculation == "no"):
|
||||||
mpi_command = NUMACTL_CPU_BIND + " mpirun -np " + str(input_data.MPI_processes) + " ./ABE"
|
mpi_command = NUMACTL_CPU_BIND + " mpirun -np " + str(input_data.MPI_processes) + " ./ABE"
|
||||||
#mpi_command = " mpirun -np " + str(input_data.MPI_processes) + " ./ABE"
|
|
||||||
mpi_command_outfile = "ABE_out.log"
|
mpi_command_outfile = "ABE_out.log"
|
||||||
elif (input_data.GPU_Calculation == "yes"):
|
elif (input_data.GPU_Calculation == "yes"):
|
||||||
mpi_command = NUMACTL_CPU_BIND + " mpirun -np " + str(input_data.MPI_processes) + " ./ABEGPU"
|
mpi_command = NUMACTL_CPU_BIND + " mpirun -np " + str(input_data.MPI_processes) + " ./ABEGPU"
|
||||||
@@ -189,8 +158,7 @@ def run_TwoPunctureABE():
|
|||||||
print( )
|
print( )
|
||||||
|
|
||||||
## Define the command to run
|
## Define the command to run
|
||||||
#TwoPuncture_command = NUMACTL_CPU_BIND + " ./TwoPunctureABE"
|
TwoPuncture_command = NUMACTL_CPU_BIND + " ./TwoPunctureABE"
|
||||||
TwoPuncture_command = " ./TwoPunctureABE"
|
|
||||||
TwoPuncture_command_outfile = "TwoPunctureABE_out.log"
|
TwoPuncture_command_outfile = "TwoPunctureABE_out.log"
|
||||||
|
|
||||||
## Execute the command with subprocess.Popen and stream output
|
## Execute the command with subprocess.Popen and stream output
|
||||||
|
|||||||
@@ -1,97 +0,0 @@
|
|||||||
# AMSS-NCKU PGO Profile Analysis Report
|
|
||||||
|
|
||||||
## 1. Profiling Environment
|
|
||||||
|
|
||||||
| Item | Value |
|
|
||||||
|------|-------|
|
|
||||||
| Compiler | Intel oneAPI DPC++/C++ 2025.3.0 (icpx/ifx) |
|
|
||||||
| Instrumentation Flag | `-fprofile-instr-generate` |
|
|
||||||
| Optimization Level (instrumented) | `-O2 -xHost -fma` |
|
|
||||||
| MPI Processes | 1 (single process to avoid MPI+instrumentation deadlock) |
|
|
||||||
| Profile File | `default_9725750769337483397_0.profraw` (327 KB) |
|
|
||||||
| Merged Profile | `default.profdata` (394 KB) |
|
|
||||||
| llvm-profdata | `/home/intel/oneapi/compiler/2025.3/bin/compiler/llvm-profdata` |
|
|
||||||
|
|
||||||
## 2. Reduced Simulation Parameters (for profiling run)
|
|
||||||
|
|
||||||
| Parameter | Production Value | Profiling Value |
|
|
||||||
|-----------|-----------------|-----------------|
|
|
||||||
| MPI_processes | 64 | 1 |
|
|
||||||
| grid_level | 9 | 4 |
|
|
||||||
| static_grid_level | 5 | 3 |
|
|
||||||
| static_grid_number | 96 | 24 |
|
|
||||||
| moving_grid_number | 48 | 16 |
|
|
||||||
| largest_box_xyz_max | 320^3 | 160^3 |
|
|
||||||
| Final_Evolution_Time | 1000.0 | 10.0 |
|
|
||||||
| Evolution_Step_Number | 10,000,000 | 1,000 |
|
|
||||||
| Detector_Number | 12 | 2 |
|
|
||||||
|
|
||||||
## 3. Profile Summary
|
|
||||||
|
|
||||||
| Metric | Value |
|
|
||||||
|--------|-------|
|
|
||||||
| Total instrumented functions | 1,392 |
|
|
||||||
| Functions with non-zero counts | 117 (8.4%) |
|
|
||||||
| Functions with zero counts | 1,275 (91.6%) |
|
|
||||||
| Maximum function entry count | 386,459,248 |
|
|
||||||
| Maximum internal block count | 370,477,680 |
|
|
||||||
| Total block count | 4,198,023,118 |
|
|
||||||
|
|
||||||
## 4. Top 20 Hotspot Functions
|
|
||||||
|
|
||||||
| Rank | Total Count | Max Block Count | Function | Category |
|
|
||||||
|------|------------|-----------------|----------|----------|
|
|
||||||
| 1 | 1,241,601,732 | 370,477,680 | `polint_` | Interpolation |
|
|
||||||
| 2 | 755,994,435 | 230,156,640 | `prolong3_` | Grid prolongation |
|
|
||||||
| 3 | 667,964,095 | 3,697,792 | `compute_rhs_bssn_` | BSSN RHS evolution |
|
|
||||||
| 4 | 539,736,051 | 386,459,248 | `symmetry_bd_` | Symmetry boundary |
|
|
||||||
| 5 | 277,310,808 | 53,170,728 | `lopsided_` | Lopsided FD stencil |
|
|
||||||
| 6 | 155,534,488 | 94,535,040 | `decide3d_` | 3D grid decision |
|
|
||||||
| 7 | 119,267,712 | 19,266,048 | `rungekutta4_rout_` | RK4 time integrator |
|
|
||||||
| 8 | 91,574,616 | 48,824,160 | `kodis_` | Kreiss-Oliger dissipation |
|
|
||||||
| 9 | 67,555,389 | 43,243,680 | `fderivs_` | Finite differences |
|
|
||||||
| 10 | 55,296,000 | 42,246,144 | `misc::fact(int)` | Factorial utility |
|
|
||||||
| 11 | 43,191,071 | 27,663,328 | `fdderivs_` | 2nd-order FD derivatives |
|
|
||||||
| 12 | 36,233,965 | 22,429,440 | `restrict3_` | Grid restriction |
|
|
||||||
| 13 | 24,698,512 | 17,231,520 | `polin3_` | Polynomial interpolation |
|
|
||||||
| 14 | 22,962,942 | 20,968,768 | `copy_` | Data copy |
|
|
||||||
| 15 | 20,135,696 | 17,259,168 | `Ansorg::barycentric(...)` | Spectral interpolation |
|
|
||||||
| 16 | 14,650,224 | 7,224,768 | `Ansorg::barycentric_omega(...)` | Spectral weights |
|
|
||||||
| 17 | 13,242,296 | 2,871,920 | `global_interp_` | Global interpolation |
|
|
||||||
| 18 | 12,672,000 | 7,734,528 | `sommerfeld_rout_` | Sommerfeld boundary |
|
|
||||||
| 19 | 6,872,832 | 1,880,064 | `sommerfeld_routbam_` | Sommerfeld boundary (BAM) |
|
|
||||||
| 20 | 5,709,900 | 2,809,632 | `l2normhelper_` | L2 norm computation |
|
|
||||||
|
|
||||||
## 5. Hotspot Category Breakdown
|
|
||||||
|
|
||||||
Top 20 functions account for ~98% of total execution counts:
|
|
||||||
|
|
||||||
| Category | Functions | Combined Count | Share |
|
|
||||||
|----------|-----------|---------------|-------|
|
|
||||||
| Interpolation / Prolongation / Restriction | polint_, prolong3_, restrict3_, polin3_, global_interp_, Ansorg::* | ~2,093M | ~50% |
|
|
||||||
| BSSN RHS + FD stencils | compute_rhs_bssn_, lopsided_, fderivs_, fdderivs_ | ~1,056M | ~25% |
|
|
||||||
| Boundary conditions | symmetry_bd_, sommerfeld_rout_, sommerfeld_routbam_ | ~559M | ~13% |
|
|
||||||
| Time integration | rungekutta4_rout_ | ~119M | ~3% |
|
|
||||||
| Dissipation | kodis_ | ~92M | ~2% |
|
|
||||||
| Utilities | misc::fact, decide3d_, copy_, l2normhelper_ | ~256M | ~6% |
|
|
||||||
|
|
||||||
## 6. Conclusions
|
|
||||||
|
|
||||||
1. **Profile data is valid**: 1,392 functions instrumented, 117 exercised with ~4.2 billion total counts.
|
|
||||||
2. **Hotspot concentration is high**: Top 5 functions alone account for ~76% of all counts, which is ideal for PGO — the compiler has strong branch/layout optimization targets.
|
|
||||||
3. **Fortran numerical kernels dominate**: `polint_`, `prolong3_`, `compute_rhs_bssn_`, `symmetry_bd_`, `lopsided_` are all Fortran routines in the inner evolution loop. PGO will optimize their branch prediction and basic block layout.
|
|
||||||
4. **91.6% of functions have zero counts**: These are code paths for unused features (GPU, BSSN-EScalar, BSSN-EM, Z4C, etc.). PGO will deprioritize them, improving instruction cache utilization.
|
|
||||||
5. **Profile is representative**: Despite the reduced grid size, the code path coverage matches production — the same kernels (RHS, prolongation, restriction, boundary) are exercised. PGO branch probabilities from this profile will transfer well to full-scale runs.
|
|
||||||
|
|
||||||
## 7. PGO Phase 2 Usage
|
|
||||||
|
|
||||||
To apply the profile, use the following flags in `makefile.inc`:
|
|
||||||
|
|
||||||
```makefile
|
|
||||||
CXXAPPFLAGS = -O3 -xHost -fp-model fast=2 -fma -ipo \
|
|
||||||
-fprofile-instr-use=/home/amss/AMSS-NCKU/pgo_profile/default.profdata \
|
|
||||||
-Dfortran3 -Dnewc -I${MKLROOT}/include
|
|
||||||
f90appflags = -O3 -xHost -fp-model fast=2 -fma -ipo \
|
|
||||||
-fprofile-instr-use=/home/amss/AMSS-NCKU/pgo_profile/default.profdata \
|
|
||||||
-align array64byte -fpp -I${MKLROOT}/include
|
|
||||||
```
|
|
||||||
Binary file not shown.
Binary file not shown.
Binary file not shown.
Binary file not shown.
Binary file not shown.
Binary file not shown.
Binary file not shown.
Reference in New Issue
Block a user