Compare commits
2 Commits
cjy-oneapi
...
yx-fmisc
| Author | SHA1 | Date | |
|---|---|---|---|
| 3f7e20f702 | |||
| 673dd20722 |
3
.gitignore
vendored
3
.gitignore
vendored
@@ -1,6 +1,3 @@
|
|||||||
__pycache__
|
__pycache__
|
||||||
GW150914
|
GW150914
|
||||||
GW150914-origin
|
GW150914-origin
|
||||||
docs
|
|
||||||
*.tmp
|
|
||||||
|
|
||||||
|
|||||||
445
AMSS_NCKU_ABEtest.py
Normal file
445
AMSS_NCKU_ABEtest.py
Normal file
@@ -0,0 +1,445 @@
|
|||||||
|
|
||||||
|
##################################################################
|
||||||
|
##
|
||||||
|
## AMSS-NCKU ABE Test Program (Skip TwoPuncture if data exists)
|
||||||
|
## Modified from AMSS_NCKU_Program.py
|
||||||
|
## Author: Xiaoqu
|
||||||
|
## Modified: 2026/02/01
|
||||||
|
##
|
||||||
|
##################################################################
|
||||||
|
|
||||||
|
|
||||||
|
##################################################################
|
||||||
|
|
||||||
|
## Print program introduction
|
||||||
|
|
||||||
|
import print_information
|
||||||
|
|
||||||
|
print_information.print_program_introduction()
|
||||||
|
|
||||||
|
##################################################################
|
||||||
|
|
||||||
|
import AMSS_NCKU_Input as input_data
|
||||||
|
|
||||||
|
##################################################################
|
||||||
|
|
||||||
|
## Create directories to store program run data
|
||||||
|
|
||||||
|
import os
|
||||||
|
import shutil
|
||||||
|
import sys
|
||||||
|
import time
|
||||||
|
|
||||||
|
## Set the output directory according to the input file
|
||||||
|
File_directory = os.path.join(input_data.File_directory)
|
||||||
|
|
||||||
|
## Check if output directory exists and if TwoPuncture data is available
|
||||||
|
skip_twopuncture = False
|
||||||
|
output_directory = os.path.join(File_directory, "AMSS_NCKU_output")
|
||||||
|
binary_results_directory = os.path.join(output_directory, input_data.Output_directory)
|
||||||
|
|
||||||
|
if os.path.exists(File_directory):
|
||||||
|
print( " Output directory already exists." )
|
||||||
|
print()
|
||||||
|
|
||||||
|
# Check if TwoPuncture initial data files exist
|
||||||
|
if (input_data.Initial_Data_Method == "Ansorg-TwoPuncture"):
|
||||||
|
twopuncture_output = os.path.join(output_directory, "TwoPunctureABE")
|
||||||
|
input_par = os.path.join(output_directory, "input.par")
|
||||||
|
|
||||||
|
if os.path.exists(twopuncture_output) and os.path.exists(input_par):
|
||||||
|
print( " Found existing TwoPuncture initial data." )
|
||||||
|
print( " Do you want to skip TwoPuncture phase and reuse existing data?" )
|
||||||
|
print( " Input 'skip' to skip TwoPuncture and start ABE directly" )
|
||||||
|
print( " Input 'regenerate' to regenerate everything from scratch" )
|
||||||
|
print()
|
||||||
|
|
||||||
|
while True:
|
||||||
|
try:
|
||||||
|
inputvalue = input()
|
||||||
|
if ( inputvalue == "skip" ):
|
||||||
|
print( " Skipping TwoPuncture phase, will reuse existing initial data." )
|
||||||
|
print()
|
||||||
|
skip_twopuncture = True
|
||||||
|
break
|
||||||
|
elif ( inputvalue == "regenerate" ):
|
||||||
|
print( " Regenerating everything from scratch." )
|
||||||
|
print()
|
||||||
|
skip_twopuncture = False
|
||||||
|
break
|
||||||
|
else:
|
||||||
|
print( " Please input 'skip' or 'regenerate'." )
|
||||||
|
except ValueError:
|
||||||
|
print( " Please input 'skip' or 'regenerate'." )
|
||||||
|
else:
|
||||||
|
print( " TwoPuncture initial data not found, will regenerate everything." )
|
||||||
|
print()
|
||||||
|
|
||||||
|
# If not skipping, remove and recreate directory
|
||||||
|
if not skip_twopuncture:
|
||||||
|
shutil.rmtree(File_directory, ignore_errors=True)
|
||||||
|
os.mkdir(File_directory)
|
||||||
|
os.mkdir(output_directory)
|
||||||
|
os.mkdir(binary_results_directory)
|
||||||
|
figure_directory = os.path.join(File_directory, "figure")
|
||||||
|
os.mkdir(figure_directory)
|
||||||
|
shutil.copy("AMSS_NCKU_Input.py", File_directory)
|
||||||
|
print( " Output directory has been regenerated." )
|
||||||
|
print()
|
||||||
|
else:
|
||||||
|
# Create fresh directory structure
|
||||||
|
os.mkdir(File_directory)
|
||||||
|
shutil.copy("AMSS_NCKU_Input.py", File_directory)
|
||||||
|
os.mkdir(output_directory)
|
||||||
|
os.mkdir(binary_results_directory)
|
||||||
|
figure_directory = os.path.join(File_directory, "figure")
|
||||||
|
os.mkdir(figure_directory)
|
||||||
|
print( " Output directory has been generated." )
|
||||||
|
print()
|
||||||
|
|
||||||
|
# Ensure figure directory exists
|
||||||
|
figure_directory = os.path.join(File_directory, "figure")
|
||||||
|
if not os.path.exists(figure_directory):
|
||||||
|
os.mkdir(figure_directory)
|
||||||
|
|
||||||
|
##################################################################
|
||||||
|
|
||||||
|
## Output related parameter information
|
||||||
|
|
||||||
|
import setup
|
||||||
|
|
||||||
|
## Print and save input parameter information
|
||||||
|
setup.print_input_data( File_directory )
|
||||||
|
|
||||||
|
if not skip_twopuncture:
|
||||||
|
setup.generate_AMSSNCKU_input()
|
||||||
|
|
||||||
|
setup.print_puncture_information()
|
||||||
|
|
||||||
|
|
||||||
|
##################################################################
|
||||||
|
|
||||||
|
## Generate AMSS-NCKU program input files based on the configured parameters
|
||||||
|
|
||||||
|
if not skip_twopuncture:
|
||||||
|
print()
|
||||||
|
print( " Generating the AMSS-NCKU input parfile for the ABE executable." )
|
||||||
|
print()
|
||||||
|
|
||||||
|
## Generate cgh-related input files from the grid information
|
||||||
|
|
||||||
|
import numerical_grid
|
||||||
|
|
||||||
|
numerical_grid.append_AMSSNCKU_cgh_input()
|
||||||
|
|
||||||
|
print()
|
||||||
|
print( " The input parfile for AMSS-NCKU C++ executable file ABE has been generated." )
|
||||||
|
print( " However, the input relevant to TwoPuncture need to be appended later." )
|
||||||
|
print()
|
||||||
|
|
||||||
|
|
||||||
|
##################################################################
|
||||||
|
|
||||||
|
## Plot the initial grid configuration
|
||||||
|
|
||||||
|
if not skip_twopuncture:
|
||||||
|
print()
|
||||||
|
print( " Schematically plot the numerical grid structure." )
|
||||||
|
print()
|
||||||
|
|
||||||
|
import numerical_grid
|
||||||
|
numerical_grid.plot_initial_grid()
|
||||||
|
|
||||||
|
|
||||||
|
##################################################################
|
||||||
|
|
||||||
|
## Generate AMSS-NCKU macro files according to the numerical scheme and parameters
|
||||||
|
|
||||||
|
if not skip_twopuncture:
|
||||||
|
print()
|
||||||
|
print( " Automatically generating the macro file for AMSS-NCKU C++ executable file ABE " )
|
||||||
|
print( " (Based on the finite-difference numerical scheme) " )
|
||||||
|
print()
|
||||||
|
|
||||||
|
import generate_macrodef
|
||||||
|
|
||||||
|
generate_macrodef.generate_macrodef_h()
|
||||||
|
print( " AMSS-NCKU macro file macrodef.h has been generated. " )
|
||||||
|
|
||||||
|
generate_macrodef.generate_macrodef_fh()
|
||||||
|
print( " AMSS-NCKU macro file macrodef.fh has been generated. " )
|
||||||
|
|
||||||
|
|
||||||
|
##################################################################
|
||||||
|
|
||||||
|
# Compile the AMSS-NCKU program according to user requirements
|
||||||
|
# NOTE: ABE compilation is always performed, even when skipping TwoPuncture
|
||||||
|
|
||||||
|
print()
|
||||||
|
print( " Preparing to compile and run the AMSS-NCKU code as requested " )
|
||||||
|
print( " Compiling the AMSS-NCKU code based on the generated macro files " )
|
||||||
|
print()
|
||||||
|
|
||||||
|
AMSS_NCKU_source_path = "AMSS_NCKU_source"
|
||||||
|
AMSS_NCKU_source_copy = os.path.join(File_directory, "AMSS_NCKU_source_copy")
|
||||||
|
|
||||||
|
## If AMSS_NCKU source folder is missing, create it and prompt the user
|
||||||
|
if not os.path.exists(AMSS_NCKU_source_path):
|
||||||
|
os.makedirs(AMSS_NCKU_source_path)
|
||||||
|
print( " The AMSS-NCKU source files are incomplete; copy all source files into ./AMSS_NCKU_source. " )
|
||||||
|
print( " Press Enter to continue. " )
|
||||||
|
inputvalue = input()
|
||||||
|
|
||||||
|
# Copy AMSS-NCKU source files to prepare for compilation
|
||||||
|
# If skipping TwoPuncture and source_copy already exists, remove it first
|
||||||
|
if skip_twopuncture and os.path.exists(AMSS_NCKU_source_copy):
|
||||||
|
shutil.rmtree(AMSS_NCKU_source_copy)
|
||||||
|
|
||||||
|
shutil.copytree(AMSS_NCKU_source_path, AMSS_NCKU_source_copy)
|
||||||
|
|
||||||
|
# Copy the generated macro files into the AMSS_NCKU source folder
|
||||||
|
if not skip_twopuncture:
|
||||||
|
macrodef_h_path = os.path.join(File_directory, "macrodef.h")
|
||||||
|
macrodef_fh_path = os.path.join(File_directory, "macrodef.fh")
|
||||||
|
else:
|
||||||
|
# When skipping TwoPuncture, use existing macro files from previous run
|
||||||
|
macrodef_h_path = os.path.join(File_directory, "macrodef.h")
|
||||||
|
macrodef_fh_path = os.path.join(File_directory, "macrodef.fh")
|
||||||
|
|
||||||
|
shutil.copy2(macrodef_h_path, AMSS_NCKU_source_copy)
|
||||||
|
shutil.copy2(macrodef_fh_path, AMSS_NCKU_source_copy)
|
||||||
|
|
||||||
|
# Compile related programs
|
||||||
|
import makefile_and_run
|
||||||
|
|
||||||
|
## Change working directory to the target source copy
|
||||||
|
os.chdir(AMSS_NCKU_source_copy)
|
||||||
|
|
||||||
|
## Build the main AMSS-NCKU executable (ABE or ABEGPU)
|
||||||
|
makefile_and_run.makefile_ABE()
|
||||||
|
|
||||||
|
## If the initial-data method is Ansorg-TwoPuncture, build the TwoPunctureABE executable
|
||||||
|
## Only build TwoPunctureABE if not skipping TwoPuncture phase
|
||||||
|
if (input_data.Initial_Data_Method == "Ansorg-TwoPuncture" ) and not skip_twopuncture:
|
||||||
|
makefile_and_run.makefile_TwoPunctureABE()
|
||||||
|
|
||||||
|
## Change current working directory back up two levels
|
||||||
|
os.chdir('..')
|
||||||
|
os.chdir('..')
|
||||||
|
|
||||||
|
print()
|
||||||
|
|
||||||
|
##################################################################
|
||||||
|
|
||||||
|
## Copy the AMSS-NCKU executable (ABE/ABEGPU) to the run directory
|
||||||
|
|
||||||
|
if (input_data.GPU_Calculation == "no"):
|
||||||
|
ABE_file = os.path.join(AMSS_NCKU_source_copy, "ABE")
|
||||||
|
elif (input_data.GPU_Calculation == "yes"):
|
||||||
|
ABE_file = os.path.join(AMSS_NCKU_source_copy, "ABEGPU")
|
||||||
|
|
||||||
|
if not os.path.exists( ABE_file ):
|
||||||
|
print()
|
||||||
|
print( " Lack of AMSS-NCKU executable file ABE/ABEGPU; recompile AMSS_NCKU_source manually. " )
|
||||||
|
print( " When recompilation is finished, press Enter to continue. " )
|
||||||
|
inputvalue = input()
|
||||||
|
|
||||||
|
## Copy the executable ABE (or ABEGPU) into the run directory
|
||||||
|
shutil.copy2(ABE_file, output_directory)
|
||||||
|
|
||||||
|
## If the initial-data method is TwoPuncture, copy the TwoPunctureABE executable to the run directory
|
||||||
|
## Only copy TwoPunctureABE if not skipping TwoPuncture phase
|
||||||
|
if (input_data.Initial_Data_Method == "Ansorg-TwoPuncture" ) and not skip_twopuncture:
|
||||||
|
TwoPuncture_file = os.path.join(AMSS_NCKU_source_copy, "TwoPunctureABE")
|
||||||
|
|
||||||
|
if not os.path.exists( TwoPuncture_file ):
|
||||||
|
print()
|
||||||
|
print( " Lack of AMSS-NCKU executable file TwoPunctureABE; recompile TwoPunctureABE in AMSS_NCKU_source. " )
|
||||||
|
print( " When recompilation is finished, press Enter to continue. " )
|
||||||
|
inputvalue = input()
|
||||||
|
|
||||||
|
## Copy the TwoPunctureABE executable into the run directory
|
||||||
|
shutil.copy2(TwoPuncture_file, output_directory)
|
||||||
|
|
||||||
|
##################################################################
|
||||||
|
|
||||||
|
## If the initial-data method is TwoPuncture, generate the TwoPuncture input files
|
||||||
|
|
||||||
|
if (input_data.Initial_Data_Method == "Ansorg-TwoPuncture" ) and not skip_twopuncture:
|
||||||
|
|
||||||
|
print()
|
||||||
|
print( " Initial data is chosen as Ansorg-TwoPuncture" )
|
||||||
|
print()
|
||||||
|
|
||||||
|
print()
|
||||||
|
print( " Automatically generating the input parfile for the TwoPunctureABE executable " )
|
||||||
|
print()
|
||||||
|
|
||||||
|
import generate_TwoPuncture_input
|
||||||
|
|
||||||
|
generate_TwoPuncture_input.generate_AMSSNCKU_TwoPuncture_input()
|
||||||
|
|
||||||
|
print()
|
||||||
|
print( " The input parfile for the TwoPunctureABE executable has been generated. " )
|
||||||
|
print()
|
||||||
|
|
||||||
|
## Generated AMSS-NCKU TwoPuncture input filename
|
||||||
|
AMSS_NCKU_TwoPuncture_inputfile = 'AMSS-NCKU-TwoPuncture.input'
|
||||||
|
AMSS_NCKU_TwoPuncture_inputfile_path = os.path.join( File_directory, AMSS_NCKU_TwoPuncture_inputfile )
|
||||||
|
|
||||||
|
## Copy and rename the file
|
||||||
|
shutil.copy2( AMSS_NCKU_TwoPuncture_inputfile_path, os.path.join(output_directory, 'TwoPunctureinput.par') )
|
||||||
|
|
||||||
|
## Run TwoPuncture to generate initial-data files
|
||||||
|
|
||||||
|
start_time = time.time() # Record start time
|
||||||
|
|
||||||
|
print()
|
||||||
|
print()
|
||||||
|
|
||||||
|
## Change to the output (run) directory
|
||||||
|
os.chdir(output_directory)
|
||||||
|
|
||||||
|
## Run the TwoPuncture executable
|
||||||
|
import makefile_and_run
|
||||||
|
makefile_and_run.run_TwoPunctureABE()
|
||||||
|
|
||||||
|
## Change current working directory back up two levels
|
||||||
|
os.chdir('..')
|
||||||
|
os.chdir('..')
|
||||||
|
|
||||||
|
elif (input_data.Initial_Data_Method == "Ansorg-TwoPuncture" ) and skip_twopuncture:
|
||||||
|
print()
|
||||||
|
print( " Skipping TwoPuncture execution, using existing initial data." )
|
||||||
|
print()
|
||||||
|
start_time = time.time() # Record start time for ABE only
|
||||||
|
else:
|
||||||
|
start_time = time.time() # Record start time
|
||||||
|
|
||||||
|
##################################################################
|
||||||
|
|
||||||
|
## Update puncture data based on TwoPuncture run results
|
||||||
|
|
||||||
|
if not skip_twopuncture:
|
||||||
|
import renew_puncture_parameter
|
||||||
|
renew_puncture_parameter.append_AMSSNCKU_BSSN_input(File_directory, output_directory)
|
||||||
|
|
||||||
|
## Generated AMSS-NCKU input filename
|
||||||
|
AMSS_NCKU_inputfile = 'AMSS-NCKU.input'
|
||||||
|
AMSS_NCKU_inputfile_path = os.path.join(File_directory, AMSS_NCKU_inputfile)
|
||||||
|
|
||||||
|
## Copy and rename the file
|
||||||
|
shutil.copy2( AMSS_NCKU_inputfile_path, os.path.join(output_directory, 'input.par') )
|
||||||
|
|
||||||
|
print()
|
||||||
|
print( " Successfully copy all AMSS-NCKU input parfile to target dictionary. " )
|
||||||
|
print()
|
||||||
|
else:
|
||||||
|
print()
|
||||||
|
print( " Using existing input.par file from previous run." )
|
||||||
|
print()
|
||||||
|
|
||||||
|
##################################################################
|
||||||
|
|
||||||
|
## Launch the AMSS-NCKU program
|
||||||
|
|
||||||
|
print()
|
||||||
|
print()
|
||||||
|
|
||||||
|
## Change to the run directory
|
||||||
|
os.chdir( output_directory )
|
||||||
|
|
||||||
|
import makefile_and_run
|
||||||
|
makefile_and_run.run_ABE()
|
||||||
|
|
||||||
|
## Change current working directory back up two levels
|
||||||
|
os.chdir('..')
|
||||||
|
os.chdir('..')
|
||||||
|
|
||||||
|
end_time = time.time()
|
||||||
|
elapsed_time = end_time - start_time
|
||||||
|
|
||||||
|
##################################################################
|
||||||
|
|
||||||
|
## Copy some basic input and log files out to facilitate debugging
|
||||||
|
|
||||||
|
## Path to the file that stores calculation settings
|
||||||
|
AMSS_NCKU_error_file_path = os.path.join(binary_results_directory, "setting.par")
|
||||||
|
## Copy and rename the file for easier inspection
|
||||||
|
shutil.copy( AMSS_NCKU_error_file_path, os.path.join(output_directory, "AMSSNCKU_setting_parameter") )
|
||||||
|
|
||||||
|
## Path to the error log file
|
||||||
|
AMSS_NCKU_error_file_path = os.path.join(binary_results_directory, "Error.log")
|
||||||
|
## Copy and rename the error log
|
||||||
|
shutil.copy( AMSS_NCKU_error_file_path, os.path.join(output_directory, "Error.log") )
|
||||||
|
|
||||||
|
## Primary program outputs
|
||||||
|
AMSS_NCKU_BH_data = os.path.join(binary_results_directory, "bssn_BH.dat" )
|
||||||
|
AMSS_NCKU_ADM_data = os.path.join(binary_results_directory, "bssn_ADMQs.dat" )
|
||||||
|
AMSS_NCKU_psi4_data = os.path.join(binary_results_directory, "bssn_psi4.dat" )
|
||||||
|
AMSS_NCKU_constraint_data = os.path.join(binary_results_directory, "bssn_constraint.dat")
|
||||||
|
## copy and rename the file
|
||||||
|
shutil.copy( AMSS_NCKU_BH_data, os.path.join(output_directory, "bssn_BH.dat" ) )
|
||||||
|
shutil.copy( AMSS_NCKU_ADM_data, os.path.join(output_directory, "bssn_ADMQs.dat" ) )
|
||||||
|
shutil.copy( AMSS_NCKU_psi4_data, os.path.join(output_directory, "bssn_psi4.dat" ) )
|
||||||
|
shutil.copy( AMSS_NCKU_constraint_data, os.path.join(output_directory, "bssn_constraint.dat") )
|
||||||
|
|
||||||
|
## Additional program outputs
|
||||||
|
if (input_data.Equation_Class == "BSSN-EM"):
|
||||||
|
AMSS_NCKU_phi1_data = os.path.join(binary_results_directory, "bssn_phi1.dat" )
|
||||||
|
AMSS_NCKU_phi2_data = os.path.join(binary_results_directory, "bssn_phi2.dat" )
|
||||||
|
shutil.copy( AMSS_NCKU_phi1_data, os.path.join(output_directory, "bssn_phi1.dat" ) )
|
||||||
|
shutil.copy( AMSS_NCKU_phi2_data, os.path.join(output_directory, "bssn_phi2.dat" ) )
|
||||||
|
elif (input_data.Equation_Class == "BSSN-EScalar"):
|
||||||
|
AMSS_NCKU_maxs_data = os.path.join(binary_results_directory, "bssn_maxs.dat" )
|
||||||
|
shutil.copy( AMSS_NCKU_maxs_data, os.path.join(output_directory, "bssn_maxs.dat" ) )
|
||||||
|
|
||||||
|
##################################################################
|
||||||
|
|
||||||
|
## Plot the AMSS-NCKU program results
|
||||||
|
|
||||||
|
print()
|
||||||
|
print( " Plotting the txt and binary results data from the AMSS-NCKU simulation " )
|
||||||
|
print()
|
||||||
|
|
||||||
|
|
||||||
|
import plot_xiaoqu
|
||||||
|
import plot_GW_strain_amplitude_xiaoqu
|
||||||
|
|
||||||
|
## Plot black hole trajectory
|
||||||
|
plot_xiaoqu.generate_puncture_orbit_plot( binary_results_directory, figure_directory )
|
||||||
|
plot_xiaoqu.generate_puncture_orbit_plot3D( binary_results_directory, figure_directory )
|
||||||
|
|
||||||
|
## Plot black hole separation vs. time
|
||||||
|
plot_xiaoqu.generate_puncture_distence_plot( binary_results_directory, figure_directory )
|
||||||
|
|
||||||
|
## Plot gravitational waveforms (psi4 and strain amplitude)
|
||||||
|
for i in range(input_data.Detector_Number):
|
||||||
|
plot_xiaoqu.generate_gravitational_wave_psi4_plot( binary_results_directory, figure_directory, i )
|
||||||
|
plot_GW_strain_amplitude_xiaoqu.generate_gravitational_wave_amplitude_plot( binary_results_directory, figure_directory, i )
|
||||||
|
|
||||||
|
## Plot ADM mass evolution
|
||||||
|
for i in range(input_data.Detector_Number):
|
||||||
|
plot_xiaoqu.generate_ADMmass_plot( binary_results_directory, figure_directory, i )
|
||||||
|
|
||||||
|
## Plot Hamiltonian constraint violation over time
|
||||||
|
for i in range(input_data.grid_level):
|
||||||
|
plot_xiaoqu.generate_constraint_check_plot( binary_results_directory, figure_directory, i )
|
||||||
|
|
||||||
|
## Plot stored binary data
|
||||||
|
plot_xiaoqu.generate_binary_data_plot( binary_results_directory, figure_directory )
|
||||||
|
|
||||||
|
print()
|
||||||
|
print( f" This Program Cost = {elapsed_time} Seconds " )
|
||||||
|
print()
|
||||||
|
|
||||||
|
|
||||||
|
##################################################################
|
||||||
|
|
||||||
|
print()
|
||||||
|
print( " The AMSS-NCKU-Python simulation is successfully finished, thanks for using !!! " )
|
||||||
|
print()
|
||||||
|
|
||||||
|
##################################################################
|
||||||
|
|
||||||
|
|
||||||
@@ -16,7 +16,7 @@ import numpy
|
|||||||
File_directory = "GW150914" ## output file directory
|
File_directory = "GW150914" ## output file directory
|
||||||
Output_directory = "binary_output" ## binary data file directory
|
Output_directory = "binary_output" ## binary data file directory
|
||||||
## The file directory name should not be too long
|
## The file directory name should not be too long
|
||||||
MPI_processes = 1 ## number of processes (MPI removed, single-process mode)
|
MPI_processes = 64 ## number of mpi processes used in the simulation
|
||||||
|
|
||||||
GPU_Calculation = "no" ## Use GPU or not
|
GPU_Calculation = "no" ## Use GPU or not
|
||||||
## (prefer "no" in the current version, because the GPU part may have bugs when integrated in this Python interface)
|
## (prefer "no" in the current version, because the GPU part may have bugs when integrated in this Python interface)
|
||||||
|
|||||||
@@ -277,3 +277,4 @@ def main():
|
|||||||
|
|
||||||
if __name__ == "__main__":
|
if __name__ == "__main__":
|
||||||
main()
|
main()
|
||||||
|
|
||||||
|
|||||||
@@ -20,11 +20,7 @@ using namespace std;
|
|||||||
#include <map.h>
|
#include <map.h>
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
#ifdef MPI_STUB
|
|
||||||
#include "mpi_stub.h"
|
|
||||||
#else
|
|
||||||
#include <mpi.h>
|
#include <mpi.h>
|
||||||
#endif
|
|
||||||
|
|
||||||
#include "misc.h"
|
#include "misc.h"
|
||||||
#include "macrodef.h"
|
#include "macrodef.h"
|
||||||
|
|||||||
@@ -19,11 +19,7 @@ using namespace std;
|
|||||||
#include <math.h>
|
#include <math.h>
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
#ifdef MPI_STUB
|
|
||||||
#include "mpi_stub.h"
|
|
||||||
#else
|
|
||||||
#include <mpi.h>
|
#include <mpi.h>
|
||||||
#endif
|
|
||||||
|
|
||||||
#define PI M_PI
|
#define PI M_PI
|
||||||
|
|
||||||
|
|||||||
@@ -2,11 +2,7 @@
|
|||||||
#ifndef BLOCK_H
|
#ifndef BLOCK_H
|
||||||
#define BLOCK_H
|
#define BLOCK_H
|
||||||
|
|
||||||
#ifdef MPI_STUB
|
|
||||||
#include "mpi_stub.h"
|
|
||||||
#else
|
|
||||||
#include <mpi.h>
|
#include <mpi.h>
|
||||||
#endif
|
|
||||||
#include "macrodef.h" //need dim here; Vertex or Cell
|
#include "macrodef.h" //need dim here; Vertex or Cell
|
||||||
#include "var.h"
|
#include "var.h"
|
||||||
#include "MyList.h"
|
#include "MyList.h"
|
||||||
|
|||||||
@@ -37,51 +37,57 @@ close(77)
|
|||||||
end program checkFFT
|
end program checkFFT
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
!-------------
|
|
||||||
! Optimized FFT using Intel oneMKL DFTI
|
|
||||||
! Mathematical equivalence: Standard DFT definition
|
|
||||||
! Forward (isign=1): X[k] = sum_{n=0}^{N-1} x[n] * exp(-2*pi*i*k*n/N)
|
|
||||||
! Backward (isign=-1): X[k] = sum_{n=0}^{N-1} x[n] * exp(+2*pi*i*k*n/N)
|
|
||||||
! Input/Output: dataa is interleaved complex array [Re(0),Im(0),Re(1),Im(1),...]
|
|
||||||
!-------------
|
!-------------
|
||||||
SUBROUTINE four1(dataa,nn,isign)
|
SUBROUTINE four1(dataa,nn,isign)
|
||||||
use MKL_DFTI
|
|
||||||
implicit none
|
implicit none
|
||||||
INTEGER, intent(in) :: isign, nn
|
INTEGER::isign,nn
|
||||||
DOUBLE PRECISION, dimension(2*nn), intent(inout) :: dataa
|
double precision,dimension(2*nn)::dataa
|
||||||
|
INTEGER::i,istep,j,m,mmax,n
|
||||||
type(DFTI_DESCRIPTOR), pointer :: desc
|
double precision::tempi,tempr
|
||||||
integer :: status
|
DOUBLE PRECISION::theta,wi,wpi,wpr,wr,wtemp
|
||||||
|
n=2*nn
|
||||||
! Create DFTI descriptor for 1D complex-to-complex transform
|
j=1
|
||||||
status = DftiCreateDescriptor(desc, DFTI_DOUBLE, DFTI_COMPLEX, 1, nn)
|
do i=1,n,2
|
||||||
if (status /= 0) return
|
if(j.gt.i)then
|
||||||
|
tempr=dataa(j)
|
||||||
! Set input/output storage as interleaved complex (default)
|
tempi=dataa(j+1)
|
||||||
status = DftiSetValue(desc, DFTI_PLACEMENT, DFTI_INPLACE)
|
dataa(j)=dataa(i)
|
||||||
if (status /= 0) then
|
dataa(j+1)=dataa(i+1)
|
||||||
status = DftiFreeDescriptor(desc)
|
dataa(i)=tempr
|
||||||
return
|
dataa(i+1)=tempi
|
||||||
endif
|
endif
|
||||||
|
m=nn
|
||||||
! Commit the descriptor
|
1 if ((m.ge.2).and.(j.gt.m)) then
|
||||||
status = DftiCommitDescriptor(desc)
|
j=j-m
|
||||||
if (status /= 0) then
|
m=m/2
|
||||||
status = DftiFreeDescriptor(desc)
|
goto 1
|
||||||
return
|
|
||||||
endif
|
endif
|
||||||
|
j=j+m
|
||||||
! Execute FFT based on direction
|
enddo
|
||||||
if (isign == 1) then
|
mmax=2
|
||||||
! Forward FFT: exp(-2*pi*i*k*n/N)
|
2 if (n.gt.mmax) then
|
||||||
status = DftiComputeForward(desc, dataa)
|
istep=2*mmax
|
||||||
else
|
theta=6.28318530717959d0/(isign*mmax)
|
||||||
! Backward FFT: exp(+2*pi*i*k*n/N)
|
wpr=-2.d0*sin(0.5d0*theta)**2
|
||||||
status = DftiComputeBackward(desc, dataa)
|
wpi=sin(theta)
|
||||||
|
wr=1.d0
|
||||||
|
wi=0.d0
|
||||||
|
do m=1,mmax,2
|
||||||
|
do i=m,n,istep
|
||||||
|
j=i+mmax
|
||||||
|
tempr=sngl(wr)*dataa(j)-sngl(wi)*dataa(j+1)
|
||||||
|
tempi=sngl(wr)*dataa(j+1)+sngl(wi)*dataa(j)
|
||||||
|
dataa(j)=dataa(i)-tempr
|
||||||
|
dataa(j+1)=dataa(i+1)-tempi
|
||||||
|
dataa(i)=dataa(i)+tempr
|
||||||
|
dataa(i+1)=dataa(i+1)+tempi
|
||||||
|
enddo
|
||||||
|
wtemp=wr
|
||||||
|
wr=wr*wpr-wi*wpi+wr
|
||||||
|
wi=wi*wpr+wtemp*wpi+wi
|
||||||
|
enddo
|
||||||
|
mmax=istep
|
||||||
|
goto 2
|
||||||
endif
|
endif
|
||||||
|
|
||||||
! Free descriptor
|
|
||||||
status = DftiFreeDescriptor(desc)
|
|
||||||
|
|
||||||
return
|
return
|
||||||
END SUBROUTINE four1
|
END SUBROUTINE four1
|
||||||
|
|||||||
@@ -4,11 +4,7 @@
|
|||||||
#include <stdarg.h>
|
#include <stdarg.h>
|
||||||
#include <string.h>
|
#include <string.h>
|
||||||
|
|
||||||
#ifdef MPI_STUB
|
|
||||||
#include "mpi_stub.h"
|
|
||||||
#else
|
|
||||||
#include <mpi.h>
|
#include <mpi.h>
|
||||||
#endif
|
|
||||||
|
|
||||||
#include "myglobal.h"
|
#include "myglobal.h"
|
||||||
|
|
||||||
|
|||||||
@@ -341,9 +341,8 @@ void Patch::Interp_Points(MyList<var> *VarList,
|
|||||||
double *Shellf, int Symmetry)
|
double *Shellf, int Symmetry)
|
||||||
{
|
{
|
||||||
// NOTE: we do not Synchnize variables here, make sure of that before calling this routine
|
// NOTE: we do not Synchnize variables here, make sure of that before calling this routine
|
||||||
int myrank, nprocs;
|
int myrank;
|
||||||
MPI_Comm_rank(MPI_COMM_WORLD, &myrank);
|
MPI_Comm_rank(MPI_COMM_WORLD, &myrank);
|
||||||
MPI_Comm_size(MPI_COMM_WORLD, &nprocs);
|
|
||||||
|
|
||||||
int ordn = 2 * ghost_width;
|
int ordn = 2 * ghost_width;
|
||||||
MyList<var> *varl;
|
MyList<var> *varl;
|
||||||
@@ -355,18 +354,24 @@ void Patch::Interp_Points(MyList<var> *VarList,
|
|||||||
varl = varl->next;
|
varl = varl->next;
|
||||||
}
|
}
|
||||||
|
|
||||||
memset(Shellf, 0, sizeof(double) * NN * num_var);
|
double *shellf;
|
||||||
|
shellf = new double[NN * num_var];
|
||||||
|
memset(shellf, 0, sizeof(double) * NN * num_var);
|
||||||
|
|
||||||
// owner_rank[j] records which MPI rank owns point j
|
// we use weight to monitor code, later some day we can move it for optimization
|
||||||
// All ranks traverse the same block list so they all agree on ownership
|
int *weight;
|
||||||
int *owner_rank;
|
weight = new int[NN];
|
||||||
owner_rank = new int[NN];
|
memset(weight, 0, sizeof(int) * NN);
|
||||||
for (int j = 0; j < NN; j++)
|
|
||||||
owner_rank[j] = -1;
|
double *DH, *llb, *uub;
|
||||||
|
DH = new double[dim];
|
||||||
|
|
||||||
double DH[dim], llb[dim], uub[dim];
|
|
||||||
for (int i = 0; i < dim; i++)
|
for (int i = 0; i < dim; i++)
|
||||||
|
{
|
||||||
DH[i] = getdX(i);
|
DH[i] = getdX(i);
|
||||||
|
}
|
||||||
|
llb = new double[dim];
|
||||||
|
uub = new double[dim];
|
||||||
|
|
||||||
for (int j = 0; j < NN; j++) // run along points
|
for (int j = 0; j < NN; j++) // run along points
|
||||||
{
|
{
|
||||||
@@ -398,6 +403,12 @@ void Patch::Interp_Points(MyList<var> *VarList,
|
|||||||
bool flag = true;
|
bool flag = true;
|
||||||
for (int i = 0; i < dim; i++)
|
for (int i = 0; i < dim; i++)
|
||||||
{
|
{
|
||||||
|
// NOTE: our dividing structure is (exclude ghost)
|
||||||
|
// -1 0
|
||||||
|
// 1 2
|
||||||
|
// so (0,1) does not belong to any part for vertex structure
|
||||||
|
// here we put (0,0.5) to left part and (0.5,1) to right part
|
||||||
|
// BUT for cell structure the bbox is (-1.5,0.5) and (0.5,2.5), there is no missing region at all
|
||||||
#ifdef Vertex
|
#ifdef Vertex
|
||||||
#ifdef Cell
|
#ifdef Cell
|
||||||
#error Both Cell and Vertex are defined
|
#error Both Cell and Vertex are defined
|
||||||
@@ -422,7 +433,6 @@ void Patch::Interp_Points(MyList<var> *VarList,
|
|||||||
if (flag)
|
if (flag)
|
||||||
{
|
{
|
||||||
notfind = false;
|
notfind = false;
|
||||||
owner_rank[j] = BP->rank;
|
|
||||||
if (myrank == BP->rank)
|
if (myrank == BP->rank)
|
||||||
{
|
{
|
||||||
//---> interpolation
|
//---> interpolation
|
||||||
@@ -430,11 +440,14 @@ void Patch::Interp_Points(MyList<var> *VarList,
|
|||||||
int k = 0;
|
int k = 0;
|
||||||
while (varl) // run along variables
|
while (varl) // run along variables
|
||||||
{
|
{
|
||||||
f_global_interp(BP->shape, BP->X[0], BP->X[1], BP->X[2], BP->fgfs[varl->data->sgfn], Shellf[j * num_var + k],
|
// shellf[j*num_var+k] = Parallel::global_interp(dim,BP->shape,BP->X,BP->fgfs[varl->data->sgfn],
|
||||||
|
// pox,ordn,varl->data->SoA,Symmetry);
|
||||||
|
f_global_interp(BP->shape, BP->X[0], BP->X[1], BP->X[2], BP->fgfs[varl->data->sgfn], shellf[j * num_var + k],
|
||||||
pox[0], pox[1], pox[2], ordn, varl->data->SoA, Symmetry);
|
pox[0], pox[1], pox[2], ordn, varl->data->SoA, Symmetry);
|
||||||
varl = varl->next;
|
varl = varl->next;
|
||||||
k++;
|
k++;
|
||||||
}
|
}
|
||||||
|
weight[j] = 1;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
if (Bp == ble)
|
if (Bp == ble)
|
||||||
@@ -443,125 +456,61 @@ void Patch::Interp_Points(MyList<var> *VarList,
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// Replace MPI_Allreduce with per-owner MPI_Bcast:
|
MPI_Allreduce(shellf, Shellf, NN * num_var, MPI_DOUBLE, MPI_SUM, MPI_COMM_WORLD);
|
||||||
// Group consecutive points by owner rank and broadcast each group.
|
int *Weight;
|
||||||
// Since each point's data is non-zero only on the owner rank,
|
Weight = new int[NN];
|
||||||
// Bcast from owner is equivalent to Allreduce(MPI_SUM) but much cheaper.
|
MPI_Allreduce(weight, Weight, NN, MPI_INT, MPI_SUM, MPI_COMM_WORLD);
|
||||||
|
|
||||||
|
// misc::tillherecheck("print me");
|
||||||
|
|
||||||
|
for (int i = 0; i < NN; i++)
|
||||||
{
|
{
|
||||||
int j = 0;
|
if (Weight[i] > 1)
|
||||||
while (j < NN)
|
|
||||||
{
|
|
||||||
int cur_owner = owner_rank[j];
|
|
||||||
if (cur_owner < 0)
|
|
||||||
{
|
{
|
||||||
if (myrank == 0)
|
if (myrank == 0)
|
||||||
|
cout << "WARNING: Patch::Interp_Points meets multiple weight" << endl;
|
||||||
|
for (int j = 0; j < num_var; j++)
|
||||||
|
Shellf[j + i * num_var] = Shellf[j + i * num_var] / Weight[i];
|
||||||
|
}
|
||||||
|
else if (Weight[i] == 0 && myrank == 0)
|
||||||
{
|
{
|
||||||
cout << "ERROR: Patch::Interp_Points fails to find point (";
|
cout << "ERROR: Patch::Interp_Points fails to find point (";
|
||||||
for (int d = 0; d < dim; d++)
|
for (int j = 0; j < dim; j++)
|
||||||
{
|
{
|
||||||
cout << XX[d][j];
|
cout << XX[j][i];
|
||||||
if (d < dim - 1)
|
if (j < dim - 1)
|
||||||
cout << ",";
|
cout << ",";
|
||||||
else
|
else
|
||||||
cout << ")";
|
cout << ")";
|
||||||
}
|
}
|
||||||
cout << " on Patch (";
|
cout << " on Patch (";
|
||||||
for (int d = 0; d < dim; d++)
|
for (int j = 0; j < dim; j++)
|
||||||
{
|
{
|
||||||
cout << bbox[d] << "+" << lli[d] * DH[d];
|
cout << bbox[j] << "+" << lli[j] * getdX(j);
|
||||||
if (d < dim - 1)
|
if (j < dim - 1)
|
||||||
cout << ",";
|
cout << ",";
|
||||||
else
|
else
|
||||||
cout << ")--";
|
cout << ")--";
|
||||||
}
|
}
|
||||||
cout << "(";
|
cout << "(";
|
||||||
for (int d = 0; d < dim; d++)
|
for (int j = 0; j < dim; j++)
|
||||||
{
|
{
|
||||||
cout << bbox[dim + d] << "-" << uui[d] * DH[d];
|
cout << bbox[dim + j] << "-" << uui[j] * getdX(j);
|
||||||
if (d < dim - 1)
|
if (j < dim - 1)
|
||||||
cout << ",";
|
cout << ",";
|
||||||
else
|
else
|
||||||
cout << ")" << endl;
|
cout << ")" << endl;
|
||||||
}
|
}
|
||||||
MPI_Abort(MPI_COMM_WORLD, 1);
|
#if 0
|
||||||
}
|
checkBlock();
|
||||||
j++;
|
#else
|
||||||
continue;
|
cout << "splited domains:" << endl;
|
||||||
}
|
|
||||||
// Find contiguous run of points with the same owner
|
|
||||||
int jstart = j;
|
|
||||||
while (j < NN && owner_rank[j] == cur_owner)
|
|
||||||
j++;
|
|
||||||
int count = (j - jstart) * num_var;
|
|
||||||
MPI_Bcast(Shellf + jstart * num_var, count, MPI_DOUBLE, cur_owner, MPI_COMM_WORLD);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
delete[] owner_rank;
|
|
||||||
}
|
|
||||||
void Patch::Interp_Points(MyList<var> *VarList,
|
|
||||||
int NN, double **XX,
|
|
||||||
double *Shellf, int Symmetry,
|
|
||||||
int Nmin_consumer, int Nmax_consumer)
|
|
||||||
{
|
{
|
||||||
// Targeted point-to-point overload: each owner sends each point only to
|
|
||||||
// the one rank that needs it for integration (consumer), reducing
|
|
||||||
// communication volume by ~nprocs times compared to the Bcast version.
|
|
||||||
int myrank, nprocs;
|
|
||||||
MPI_Comm_rank(MPI_COMM_WORLD, &myrank);
|
|
||||||
MPI_Comm_size(MPI_COMM_WORLD, &nprocs);
|
|
||||||
|
|
||||||
int ordn = 2 * ghost_width;
|
|
||||||
MyList<var> *varl;
|
|
||||||
int num_var = 0;
|
|
||||||
varl = VarList;
|
|
||||||
while (varl)
|
|
||||||
{
|
|
||||||
num_var++;
|
|
||||||
varl = varl->next;
|
|
||||||
}
|
|
||||||
|
|
||||||
memset(Shellf, 0, sizeof(double) * NN * num_var);
|
|
||||||
|
|
||||||
// owner_rank[j] records which MPI rank owns point j
|
|
||||||
int *owner_rank;
|
|
||||||
owner_rank = new int[NN];
|
|
||||||
for (int j = 0; j < NN; j++)
|
|
||||||
owner_rank[j] = -1;
|
|
||||||
|
|
||||||
double DH[dim], llb[dim], uub[dim];
|
|
||||||
for (int i = 0; i < dim; i++)
|
|
||||||
DH[i] = getdX(i);
|
|
||||||
|
|
||||||
// --- Interpolation phase (identical to original) ---
|
|
||||||
for (int j = 0; j < NN; j++)
|
|
||||||
{
|
|
||||||
double pox[dim];
|
|
||||||
for (int i = 0; i < dim; i++)
|
|
||||||
{
|
|
||||||
pox[i] = XX[i][j];
|
|
||||||
if (myrank == 0 && (XX[i][j] < bbox[i] + lli[i] * DH[i] || XX[i][j] > bbox[dim + i] - uui[i] * DH[i]))
|
|
||||||
{
|
|
||||||
cout << "Patch::Interp_Points: point (";
|
|
||||||
for (int k = 0; k < dim; k++)
|
|
||||||
{
|
|
||||||
cout << XX[k][j];
|
|
||||||
if (k < dim - 1)
|
|
||||||
cout << ",";
|
|
||||||
else
|
|
||||||
cout << ") is out of current Patch." << endl;
|
|
||||||
}
|
|
||||||
MPI_Abort(MPI_COMM_WORLD, 1);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
MyList<Block> *Bp = blb;
|
MyList<Block> *Bp = blb;
|
||||||
bool notfind = true;
|
while (Bp)
|
||||||
while (notfind && Bp)
|
|
||||||
{
|
{
|
||||||
Block *BP = Bp->data;
|
Block *BP = Bp->data;
|
||||||
|
|
||||||
bool flag = true;
|
|
||||||
for (int i = 0; i < dim; i++)
|
for (int i = 0; i < dim; i++)
|
||||||
{
|
{
|
||||||
#ifdef Vertex
|
#ifdef Vertex
|
||||||
@@ -578,192 +527,32 @@ void Patch::Interp_Points(MyList<var> *VarList,
|
|||||||
#error Not define Vertex nor Cell
|
#error Not define Vertex nor Cell
|
||||||
#endif
|
#endif
|
||||||
#endif
|
#endif
|
||||||
if (XX[i][j] - llb[i] < -DH[i] / 2 || XX[i][j] - uub[i] > DH[i] / 2)
|
|
||||||
{
|
|
||||||
flag = false;
|
|
||||||
break;
|
|
||||||
}
|
}
|
||||||
}
|
cout << "(";
|
||||||
|
for (int j = 0; j < dim; j++)
|
||||||
if (flag)
|
|
||||||
{
|
{
|
||||||
notfind = false;
|
cout << llb[j] << ":" << uub[j];
|
||||||
owner_rank[j] = BP->rank;
|
if (j < dim - 1)
|
||||||
if (myrank == BP->rank)
|
cout << ",";
|
||||||
{
|
else
|
||||||
varl = VarList;
|
cout << ")" << endl;
|
||||||
int k = 0;
|
|
||||||
while (varl)
|
|
||||||
{
|
|
||||||
f_global_interp(BP->shape, BP->X[0], BP->X[1], BP->X[2], BP->fgfs[varl->data->sgfn], Shellf[j * num_var + k],
|
|
||||||
pox[0], pox[1], pox[2], ordn, varl->data->SoA, Symmetry);
|
|
||||||
varl = varl->next;
|
|
||||||
k++;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
if (Bp == ble)
|
if (Bp == ble)
|
||||||
break;
|
break;
|
||||||
Bp = Bp->next;
|
Bp = Bp->next;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
#endif
|
||||||
// --- Error check for unfound points ---
|
|
||||||
for (int j = 0; j < NN; j++)
|
|
||||||
{
|
|
||||||
if (owner_rank[j] < 0 && myrank == 0)
|
|
||||||
{
|
|
||||||
cout << "ERROR: Patch::Interp_Points fails to find point (";
|
|
||||||
for (int d = 0; d < dim; d++)
|
|
||||||
{
|
|
||||||
cout << XX[d][j];
|
|
||||||
if (d < dim - 1)
|
|
||||||
cout << ",";
|
|
||||||
else
|
|
||||||
cout << ")";
|
|
||||||
}
|
|
||||||
cout << " on Patch (";
|
|
||||||
for (int d = 0; d < dim; d++)
|
|
||||||
{
|
|
||||||
cout << bbox[d] << "+" << lli[d] * DH[d];
|
|
||||||
if (d < dim - 1)
|
|
||||||
cout << ",";
|
|
||||||
else
|
|
||||||
cout << ")--";
|
|
||||||
}
|
|
||||||
cout << "(";
|
|
||||||
for (int d = 0; d < dim; d++)
|
|
||||||
{
|
|
||||||
cout << bbox[dim + d] << "-" << uui[d] * DH[d];
|
|
||||||
if (d < dim - 1)
|
|
||||||
cout << ",";
|
|
||||||
else
|
|
||||||
cout << ")" << endl;
|
|
||||||
}
|
|
||||||
MPI_Abort(MPI_COMM_WORLD, 1);
|
MPI_Abort(MPI_COMM_WORLD, 1);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// --- Targeted point-to-point communication phase ---
|
delete[] shellf;
|
||||||
// Compute consumer_rank[j] using the same deterministic formula as surface_integral
|
delete[] weight;
|
||||||
int *consumer_rank = new int[NN];
|
delete[] Weight;
|
||||||
{
|
delete[] DH;
|
||||||
int mp = NN / nprocs;
|
delete[] llb;
|
||||||
int Lp = NN - nprocs * mp;
|
delete[] uub;
|
||||||
for (int j = 0; j < NN; j++)
|
|
||||||
{
|
|
||||||
if (j < Lp * (mp + 1))
|
|
||||||
consumer_rank[j] = j / (mp + 1);
|
|
||||||
else
|
|
||||||
consumer_rank[j] = Lp + (j - Lp * (mp + 1)) / mp;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// Count sends and recvs per rank
|
|
||||||
int *send_count = new int[nprocs];
|
|
||||||
int *recv_count = new int[nprocs];
|
|
||||||
memset(send_count, 0, sizeof(int) * nprocs);
|
|
||||||
memset(recv_count, 0, sizeof(int) * nprocs);
|
|
||||||
|
|
||||||
for (int j = 0; j < NN; j++)
|
|
||||||
{
|
|
||||||
int own = owner_rank[j];
|
|
||||||
int con = consumer_rank[j];
|
|
||||||
if (own == con)
|
|
||||||
continue; // local — no communication needed
|
|
||||||
if (own == myrank)
|
|
||||||
send_count[con]++;
|
|
||||||
if (con == myrank)
|
|
||||||
recv_count[own]++;
|
|
||||||
}
|
|
||||||
|
|
||||||
// Build send buffers: for each destination rank, pack (index, data) pairs
|
|
||||||
// Each entry: 1 int (point index j) + num_var doubles
|
|
||||||
int total_send = 0, total_recv = 0;
|
|
||||||
int *send_offset = new int[nprocs];
|
|
||||||
int *recv_offset = new int[nprocs];
|
|
||||||
for (int r = 0; r < nprocs; r++)
|
|
||||||
{
|
|
||||||
send_offset[r] = total_send;
|
|
||||||
total_send += send_count[r];
|
|
||||||
recv_offset[r] = total_recv;
|
|
||||||
total_recv += recv_count[r];
|
|
||||||
}
|
|
||||||
|
|
||||||
// Pack send buffers: each message contains (j, data[0..num_var-1]) per point
|
|
||||||
int stride = 1 + num_var; // 1 double for index + num_var doubles for data
|
|
||||||
double *sendbuf = new double[total_send * stride];
|
|
||||||
double *recvbuf = new double[total_recv * stride];
|
|
||||||
|
|
||||||
// Temporary counters for packing
|
|
||||||
int *pack_pos = new int[nprocs];
|
|
||||||
memset(pack_pos, 0, sizeof(int) * nprocs);
|
|
||||||
|
|
||||||
for (int j = 0; j < NN; j++)
|
|
||||||
{
|
|
||||||
int own = owner_rank[j];
|
|
||||||
int con = consumer_rank[j];
|
|
||||||
if (own != myrank || con == myrank)
|
|
||||||
continue;
|
|
||||||
int pos = (send_offset[con] + pack_pos[con]) * stride;
|
|
||||||
sendbuf[pos] = (double)j; // point index
|
|
||||||
for (int v = 0; v < num_var; v++)
|
|
||||||
sendbuf[pos + 1 + v] = Shellf[j * num_var + v];
|
|
||||||
pack_pos[con]++;
|
|
||||||
}
|
|
||||||
|
|
||||||
// Post non-blocking recvs and sends
|
|
||||||
int n_req = 0;
|
|
||||||
for (int r = 0; r < nprocs; r++)
|
|
||||||
{
|
|
||||||
if (recv_count[r] > 0) n_req++;
|
|
||||||
if (send_count[r] > 0) n_req++;
|
|
||||||
}
|
|
||||||
|
|
||||||
MPI_Request *reqs = new MPI_Request[n_req];
|
|
||||||
int req_idx = 0;
|
|
||||||
|
|
||||||
for (int r = 0; r < nprocs; r++)
|
|
||||||
{
|
|
||||||
if (recv_count[r] > 0)
|
|
||||||
{
|
|
||||||
MPI_Irecv(recvbuf + recv_offset[r] * stride,
|
|
||||||
recv_count[r] * stride, MPI_DOUBLE,
|
|
||||||
r, 0, MPI_COMM_WORLD, &reqs[req_idx++]);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
for (int r = 0; r < nprocs; r++)
|
|
||||||
{
|
|
||||||
if (send_count[r] > 0)
|
|
||||||
{
|
|
||||||
MPI_Isend(sendbuf + send_offset[r] * stride,
|
|
||||||
send_count[r] * stride, MPI_DOUBLE,
|
|
||||||
r, 0, MPI_COMM_WORLD, &reqs[req_idx++]);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
if (n_req > 0)
|
|
||||||
MPI_Waitall(n_req, reqs, MPI_STATUSES_IGNORE);
|
|
||||||
|
|
||||||
// Unpack recv buffers into Shellf
|
|
||||||
for (int i = 0; i < total_recv; i++)
|
|
||||||
{
|
|
||||||
int pos = i * stride;
|
|
||||||
int j = (int)recvbuf[pos];
|
|
||||||
for (int v = 0; v < num_var; v++)
|
|
||||||
Shellf[j * num_var + v] = recvbuf[pos + 1 + v];
|
|
||||||
}
|
|
||||||
|
|
||||||
delete[] reqs;
|
|
||||||
delete[] sendbuf;
|
|
||||||
delete[] recvbuf;
|
|
||||||
delete[] pack_pos;
|
|
||||||
delete[] send_offset;
|
|
||||||
delete[] recv_offset;
|
|
||||||
delete[] send_count;
|
|
||||||
delete[] recv_count;
|
|
||||||
delete[] consumer_rank;
|
|
||||||
delete[] owner_rank;
|
|
||||||
}
|
}
|
||||||
void Patch::Interp_Points(MyList<var> *VarList,
|
void Patch::Interp_Points(MyList<var> *VarList,
|
||||||
int NN, double **XX,
|
int NN, double **XX,
|
||||||
@@ -784,22 +573,24 @@ void Patch::Interp_Points(MyList<var> *VarList,
|
|||||||
varl = varl->next;
|
varl = varl->next;
|
||||||
}
|
}
|
||||||
|
|
||||||
memset(Shellf, 0, sizeof(double) * NN * num_var);
|
double *shellf;
|
||||||
|
shellf = new double[NN * num_var];
|
||||||
|
memset(shellf, 0, sizeof(double) * NN * num_var);
|
||||||
|
|
||||||
// owner_rank[j] stores the global rank that owns point j
|
// we use weight to monitor code, later some day we can move it for optimization
|
||||||
int *owner_rank;
|
int *weight;
|
||||||
owner_rank = new int[NN];
|
weight = new int[NN];
|
||||||
for (int j = 0; j < NN; j++)
|
memset(weight, 0, sizeof(int) * NN);
|
||||||
owner_rank[j] = -1;
|
|
||||||
|
|
||||||
// Build global-to-local rank translation for Comm_here
|
double *DH, *llb, *uub;
|
||||||
MPI_Group world_group, local_group;
|
DH = new double[dim];
|
||||||
MPI_Comm_group(MPI_COMM_WORLD, &world_group);
|
|
||||||
MPI_Comm_group(Comm_here, &local_group);
|
|
||||||
|
|
||||||
double DH[dim], llb[dim], uub[dim];
|
|
||||||
for (int i = 0; i < dim; i++)
|
for (int i = 0; i < dim; i++)
|
||||||
|
{
|
||||||
DH[i] = getdX(i);
|
DH[i] = getdX(i);
|
||||||
|
}
|
||||||
|
llb = new double[dim];
|
||||||
|
uub = new double[dim];
|
||||||
|
|
||||||
for (int j = 0; j < NN; j++) // run along points
|
for (int j = 0; j < NN; j++) // run along points
|
||||||
{
|
{
|
||||||
@@ -831,6 +622,12 @@ void Patch::Interp_Points(MyList<var> *VarList,
|
|||||||
bool flag = true;
|
bool flag = true;
|
||||||
for (int i = 0; i < dim; i++)
|
for (int i = 0; i < dim; i++)
|
||||||
{
|
{
|
||||||
|
// NOTE: our dividing structure is (exclude ghost)
|
||||||
|
// -1 0
|
||||||
|
// 1 2
|
||||||
|
// so (0,1) does not belong to any part for vertex structure
|
||||||
|
// here we put (0,0.5) to left part and (0.5,1) to right part
|
||||||
|
// BUT for cell structure the bbox is (-1.5,0.5) and (0.5,2.5), there is no missing region at all
|
||||||
#ifdef Vertex
|
#ifdef Vertex
|
||||||
#ifdef Cell
|
#ifdef Cell
|
||||||
#error Both Cell and Vertex are defined
|
#error Both Cell and Vertex are defined
|
||||||
@@ -855,7 +652,6 @@ void Patch::Interp_Points(MyList<var> *VarList,
|
|||||||
if (flag)
|
if (flag)
|
||||||
{
|
{
|
||||||
notfind = false;
|
notfind = false;
|
||||||
owner_rank[j] = BP->rank;
|
|
||||||
if (myrank == BP->rank)
|
if (myrank == BP->rank)
|
||||||
{
|
{
|
||||||
//---> interpolation
|
//---> interpolation
|
||||||
@@ -863,11 +659,14 @@ void Patch::Interp_Points(MyList<var> *VarList,
|
|||||||
int k = 0;
|
int k = 0;
|
||||||
while (varl) // run along variables
|
while (varl) // run along variables
|
||||||
{
|
{
|
||||||
f_global_interp(BP->shape, BP->X[0], BP->X[1], BP->X[2], BP->fgfs[varl->data->sgfn], Shellf[j * num_var + k],
|
// shellf[j*num_var+k] = Parallel::global_interp(dim,BP->shape,BP->X,BP->fgfs[varl->data->sgfn],
|
||||||
|
// pox,ordn,varl->data->SoA,Symmetry);
|
||||||
|
f_global_interp(BP->shape, BP->X[0], BP->X[1], BP->X[2], BP->fgfs[varl->data->sgfn], shellf[j * num_var + k],
|
||||||
pox[0], pox[1], pox[2], ordn, varl->data->SoA, Symmetry);
|
pox[0], pox[1], pox[2], ordn, varl->data->SoA, Symmetry);
|
||||||
varl = varl->next;
|
varl = varl->next;
|
||||||
k++;
|
k++;
|
||||||
}
|
}
|
||||||
|
weight[j] = 1;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
if (Bp == ble)
|
if (Bp == ble)
|
||||||
@@ -876,35 +675,97 @@ void Patch::Interp_Points(MyList<var> *VarList,
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// Collect unique global owner ranks and translate to local ranks in Comm_here
|
MPI_Allreduce(shellf, Shellf, NN * num_var, MPI_DOUBLE, MPI_SUM, Comm_here);
|
||||||
// Then broadcast each owner's points via MPI_Bcast on Comm_here
|
int *Weight;
|
||||||
{
|
Weight = new int[NN];
|
||||||
int j = 0;
|
MPI_Allreduce(weight, Weight, NN, MPI_INT, MPI_SUM, Comm_here);
|
||||||
while (j < NN)
|
|
||||||
{
|
|
||||||
int cur_owner_global = owner_rank[j];
|
|
||||||
if (cur_owner_global < 0)
|
|
||||||
{
|
|
||||||
// Point not found — skip (error check disabled for sub-communicator levels)
|
|
||||||
j++;
|
|
||||||
continue;
|
|
||||||
}
|
|
||||||
// Translate global rank to local rank in Comm_here
|
|
||||||
int cur_owner_local;
|
|
||||||
MPI_Group_translate_ranks(world_group, 1, &cur_owner_global, local_group, &cur_owner_local);
|
|
||||||
|
|
||||||
// Find contiguous run of points with the same owner
|
// misc::tillherecheck("print me");
|
||||||
int jstart = j;
|
// if(lmyrank == 0) cout<<"myrank = "<<myrank<<"print me"<<endl;
|
||||||
while (j < NN && owner_rank[j] == cur_owner_global)
|
|
||||||
j++;
|
for (int i = 0; i < NN; i++)
|
||||||
int count = (j - jstart) * num_var;
|
{
|
||||||
MPI_Bcast(Shellf + jstart * num_var, count, MPI_DOUBLE, cur_owner_local, Comm_here);
|
if (Weight[i] > 1)
|
||||||
|
{
|
||||||
|
if (lmyrank == 0)
|
||||||
|
cout << "WARNING: Patch::Interp_Points meets multiple weight" << endl;
|
||||||
|
for (int j = 0; j < num_var; j++)
|
||||||
|
Shellf[j + i * num_var] = Shellf[j + i * num_var] / Weight[i];
|
||||||
}
|
}
|
||||||
|
#if 0 // for not involved levels, this may fail
|
||||||
|
else if(Weight[i] == 0 && lmyrank == 0)
|
||||||
|
{
|
||||||
|
cout<<"ERROR: Patch::Interp_Points fails to find point (";
|
||||||
|
for(int j=0;j<dim;j++)
|
||||||
|
{
|
||||||
|
cout<<XX[j][i];
|
||||||
|
if(j<dim-1) cout<<",";
|
||||||
|
else cout<<")";
|
||||||
|
}
|
||||||
|
cout<<" on Patch (";
|
||||||
|
for(int j=0;j<dim;j++)
|
||||||
|
{
|
||||||
|
cout<<bbox[j]<<"+"<<lli[j]*getdX(j);
|
||||||
|
if(j<dim-1) cout<<",";
|
||||||
|
else cout<<")--";
|
||||||
|
}
|
||||||
|
cout<<"(";
|
||||||
|
for(int j=0;j<dim;j++)
|
||||||
|
{
|
||||||
|
cout<<bbox[dim+j]<<"-"<<uui[j]*getdX(j);
|
||||||
|
if(j<dim-1) cout<<",";
|
||||||
|
else cout<<")"<<endl;
|
||||||
|
}
|
||||||
|
#if 0
|
||||||
|
checkBlock();
|
||||||
|
#else
|
||||||
|
cout<<"splited domains:"<<endl;
|
||||||
|
{
|
||||||
|
MyList<Block> *Bp=blb;
|
||||||
|
while(Bp)
|
||||||
|
{
|
||||||
|
Block *BP=Bp->data;
|
||||||
|
|
||||||
|
for(int i=0;i<dim;i++)
|
||||||
|
{
|
||||||
|
#ifdef Vertex
|
||||||
|
#ifdef Cell
|
||||||
|
#error Both Cell and Vertex are defined
|
||||||
|
#endif
|
||||||
|
llb[i] = (feq(BP->bbox[i] ,bbox[i] ,DH[i]/2)) ? BP->bbox[i]+lli[i]*DH[i] : BP->bbox[i] +(ghost_width-0.5)*DH[i];
|
||||||
|
uub[i] = (feq(BP->bbox[dim+i],bbox[dim+i],DH[i]/2)) ? BP->bbox[dim+i]-uui[i]*DH[i] : BP->bbox[dim+i]-(ghost_width-0.5)*DH[i];
|
||||||
|
#else
|
||||||
|
#ifdef Cell
|
||||||
|
llb[i] = (feq(BP->bbox[i] ,bbox[i] ,DH[i]/2)) ? BP->bbox[i]+lli[i]*DH[i] : BP->bbox[i] +ghost_width*DH[i];
|
||||||
|
uub[i] = (feq(BP->bbox[dim+i],bbox[dim+i],DH[i]/2)) ? BP->bbox[dim+i]-uui[i]*DH[i] : BP->bbox[dim+i]-ghost_width*DH[i];
|
||||||
|
#else
|
||||||
|
#error Not define Vertex nor Cell
|
||||||
|
#endif
|
||||||
|
#endif
|
||||||
|
}
|
||||||
|
cout<<"(";
|
||||||
|
for(int j=0;j<dim;j++)
|
||||||
|
{
|
||||||
|
cout<<llb[j]<<":"<<uub[j];
|
||||||
|
if(j<dim-1) cout<<",";
|
||||||
|
else cout<<")"<<endl;
|
||||||
|
}
|
||||||
|
if(Bp == ble) break;
|
||||||
|
Bp=Bp->next;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
#endif
|
||||||
|
MPI_Abort(MPI_COMM_WORLD,1);
|
||||||
|
}
|
||||||
|
#endif
|
||||||
}
|
}
|
||||||
|
|
||||||
MPI_Group_free(&world_group);
|
delete[] shellf;
|
||||||
MPI_Group_free(&local_group);
|
delete[] weight;
|
||||||
delete[] owner_rank;
|
delete[] Weight;
|
||||||
|
delete[] DH;
|
||||||
|
delete[] llb;
|
||||||
|
delete[] uub;
|
||||||
}
|
}
|
||||||
void Patch::checkBlock()
|
void Patch::checkBlock()
|
||||||
{
|
{
|
||||||
|
|||||||
@@ -2,11 +2,7 @@
|
|||||||
#ifndef PATCH_H
|
#ifndef PATCH_H
|
||||||
#define PATCH_H
|
#define PATCH_H
|
||||||
|
|
||||||
#ifdef MPI_STUB
|
|
||||||
#include "mpi_stub.h"
|
|
||||||
#else
|
|
||||||
#include <mpi.h>
|
#include <mpi.h>
|
||||||
#endif
|
|
||||||
#include "MyList.h"
|
#include "MyList.h"
|
||||||
#include "Block.h"
|
#include "Block.h"
|
||||||
#include "var.h"
|
#include "var.h"
|
||||||
@@ -43,10 +39,6 @@ public:
|
|||||||
|
|
||||||
bool Find_Point(double *XX);
|
bool Find_Point(double *XX);
|
||||||
|
|
||||||
void Interp_Points(MyList<var> *VarList,
|
|
||||||
int NN, double **XX,
|
|
||||||
double *Shellf, int Symmetry,
|
|
||||||
int Nmin_consumer, int Nmax_consumer);
|
|
||||||
void Interp_Points(MyList<var> *VarList,
|
void Interp_Points(MyList<var> *VarList,
|
||||||
int NN, double **XX,
|
int NN, double **XX,
|
||||||
double *Shellf, int Symmetry, MPI_Comm Comm_here);
|
double *Shellf, int Symmetry, MPI_Comm Comm_here);
|
||||||
|
|||||||
@@ -8,11 +8,7 @@
|
|||||||
#include <limits.h>
|
#include <limits.h>
|
||||||
#include <float.h>
|
#include <float.h>
|
||||||
#include <math.h>
|
#include <math.h>
|
||||||
#ifdef MPI_STUB
|
|
||||||
#include "mpi_stub.h"
|
|
||||||
#else
|
|
||||||
#include <mpi.h>
|
#include <mpi.h>
|
||||||
#endif
|
|
||||||
|
|
||||||
#include "util_Table.h"
|
#include "util_Table.h"
|
||||||
#include "cctk.h"
|
#include "cctk.h"
|
||||||
|
|||||||
@@ -23,11 +23,7 @@ using namespace std;
|
|||||||
#include <complex.h>
|
#include <complex.h>
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
#ifdef MPI_STUB
|
|
||||||
#include "mpi_stub.h"
|
|
||||||
#else
|
|
||||||
#include <mpi.h>
|
#include <mpi.h>
|
||||||
#endif
|
|
||||||
#include "MyList.h"
|
#include "MyList.h"
|
||||||
#include "Block.h"
|
#include "Block.h"
|
||||||
#include "Parallel.h"
|
#include "Parallel.h"
|
||||||
|
|||||||
@@ -23,11 +23,7 @@ using namespace std;
|
|||||||
#include <complex.h>
|
#include <complex.h>
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
#ifdef MPI_STUB
|
|
||||||
#include "mpi_stub.h"
|
|
||||||
#else
|
|
||||||
#include <mpi.h>
|
#include <mpi.h>
|
||||||
#endif
|
|
||||||
#include "MyList.h"
|
#include "MyList.h"
|
||||||
#include "Block.h"
|
#include "Block.h"
|
||||||
#include "Parallel.h"
|
#include "Parallel.h"
|
||||||
|
|||||||
@@ -3756,484 +3756,6 @@ void Parallel::Sync(MyList<Patch> *PatL, MyList<var> *VarList, int Symmetry)
|
|||||||
delete[] transfer_src;
|
delete[] transfer_src;
|
||||||
delete[] transfer_dst;
|
delete[] transfer_dst;
|
||||||
}
|
}
|
||||||
// Merged Sync: collect all intra-patch and inter-patch grid segment lists,
|
|
||||||
// then issue a single transfer() call instead of N+1 separate ones.
|
|
||||||
void Parallel::Sync_merged(MyList<Patch> *PatL, MyList<var> *VarList, int Symmetry)
|
|
||||||
{
|
|
||||||
int cpusize;
|
|
||||||
MPI_Comm_size(MPI_COMM_WORLD, &cpusize);
|
|
||||||
|
|
||||||
MyList<Parallel::gridseg> **combined_src = new MyList<Parallel::gridseg> *[cpusize];
|
|
||||||
MyList<Parallel::gridseg> **combined_dst = new MyList<Parallel::gridseg> *[cpusize];
|
|
||||||
for (int node = 0; node < cpusize; node++)
|
|
||||||
combined_src[node] = combined_dst[node] = 0;
|
|
||||||
|
|
||||||
// Phase A: Intra-patch ghost exchange segments
|
|
||||||
MyList<Patch> *Pp = PatL;
|
|
||||||
while (Pp)
|
|
||||||
{
|
|
||||||
Patch *Pat = Pp->data;
|
|
||||||
MyList<Parallel::gridseg> *dst_ghost = build_ghost_gsl(Pat);
|
|
||||||
|
|
||||||
for (int node = 0; node < cpusize; node++)
|
|
||||||
{
|
|
||||||
MyList<Parallel::gridseg> *src_owned = build_owned_gsl0(Pat, node);
|
|
||||||
MyList<Parallel::gridseg> *tsrc = 0, *tdst = 0;
|
|
||||||
build_gstl(src_owned, dst_ghost, &tsrc, &tdst);
|
|
||||||
|
|
||||||
if (tsrc)
|
|
||||||
{
|
|
||||||
if (combined_src[node])
|
|
||||||
combined_src[node]->catList(tsrc);
|
|
||||||
else
|
|
||||||
combined_src[node] = tsrc;
|
|
||||||
}
|
|
||||||
if (tdst)
|
|
||||||
{
|
|
||||||
if (combined_dst[node])
|
|
||||||
combined_dst[node]->catList(tdst);
|
|
||||||
else
|
|
||||||
combined_dst[node] = tdst;
|
|
||||||
}
|
|
||||||
|
|
||||||
if (src_owned)
|
|
||||||
src_owned->destroyList();
|
|
||||||
}
|
|
||||||
|
|
||||||
if (dst_ghost)
|
|
||||||
dst_ghost->destroyList();
|
|
||||||
|
|
||||||
Pp = Pp->next;
|
|
||||||
}
|
|
||||||
|
|
||||||
// Phase B: Inter-patch buffer exchange segments
|
|
||||||
MyList<Parallel::gridseg> *dst_buffer = build_buffer_gsl(PatL);
|
|
||||||
for (int node = 0; node < cpusize; node++)
|
|
||||||
{
|
|
||||||
MyList<Parallel::gridseg> *src_owned = build_owned_gsl(PatL, node, 5, Symmetry);
|
|
||||||
MyList<Parallel::gridseg> *tsrc = 0, *tdst = 0;
|
|
||||||
build_gstl(src_owned, dst_buffer, &tsrc, &tdst);
|
|
||||||
|
|
||||||
if (tsrc)
|
|
||||||
{
|
|
||||||
if (combined_src[node])
|
|
||||||
combined_src[node]->catList(tsrc);
|
|
||||||
else
|
|
||||||
combined_src[node] = tsrc;
|
|
||||||
}
|
|
||||||
if (tdst)
|
|
||||||
{
|
|
||||||
if (combined_dst[node])
|
|
||||||
combined_dst[node]->catList(tdst);
|
|
||||||
else
|
|
||||||
combined_dst[node] = tdst;
|
|
||||||
}
|
|
||||||
|
|
||||||
if (src_owned)
|
|
||||||
src_owned->destroyList();
|
|
||||||
}
|
|
||||||
if (dst_buffer)
|
|
||||||
dst_buffer->destroyList();
|
|
||||||
|
|
||||||
// Phase C: Single transfer
|
|
||||||
transfer(combined_src, combined_dst, VarList, VarList, Symmetry);
|
|
||||||
|
|
||||||
// Phase D: Cleanup
|
|
||||||
for (int node = 0; node < cpusize; node++)
|
|
||||||
{
|
|
||||||
if (combined_src[node])
|
|
||||||
combined_src[node]->destroyList();
|
|
||||||
if (combined_dst[node])
|
|
||||||
combined_dst[node]->destroyList();
|
|
||||||
}
|
|
||||||
delete[] combined_src;
|
|
||||||
delete[] combined_dst;
|
|
||||||
}
|
|
||||||
// SyncCache constructor
|
|
||||||
Parallel::SyncCache::SyncCache()
|
|
||||||
: valid(false), cpusize(0), combined_src(0), combined_dst(0),
|
|
||||||
send_lengths(0), recv_lengths(0), send_bufs(0), recv_bufs(0),
|
|
||||||
send_buf_caps(0), recv_buf_caps(0), reqs(0), stats(0), max_reqs(0)
|
|
||||||
{
|
|
||||||
}
|
|
||||||
// SyncCache invalidate: free grid segment lists but keep buffers
|
|
||||||
void Parallel::SyncCache::invalidate()
|
|
||||||
{
|
|
||||||
if (!valid)
|
|
||||||
return;
|
|
||||||
for (int i = 0; i < cpusize; i++)
|
|
||||||
{
|
|
||||||
if (combined_src[i])
|
|
||||||
combined_src[i]->destroyList();
|
|
||||||
if (combined_dst[i])
|
|
||||||
combined_dst[i]->destroyList();
|
|
||||||
combined_src[i] = combined_dst[i] = 0;
|
|
||||||
send_lengths[i] = recv_lengths[i] = 0;
|
|
||||||
}
|
|
||||||
valid = false;
|
|
||||||
}
|
|
||||||
// SyncCache destroy: free everything
|
|
||||||
void Parallel::SyncCache::destroy()
|
|
||||||
{
|
|
||||||
invalidate();
|
|
||||||
if (combined_src) delete[] combined_src;
|
|
||||||
if (combined_dst) delete[] combined_dst;
|
|
||||||
if (send_lengths) delete[] send_lengths;
|
|
||||||
if (recv_lengths) delete[] recv_lengths;
|
|
||||||
if (send_buf_caps) delete[] send_buf_caps;
|
|
||||||
if (recv_buf_caps) delete[] recv_buf_caps;
|
|
||||||
for (int i = 0; i < cpusize; i++)
|
|
||||||
{
|
|
||||||
if (send_bufs && send_bufs[i]) delete[] send_bufs[i];
|
|
||||||
if (recv_bufs && recv_bufs[i]) delete[] recv_bufs[i];
|
|
||||||
}
|
|
||||||
if (send_bufs) delete[] send_bufs;
|
|
||||||
if (recv_bufs) delete[] recv_bufs;
|
|
||||||
if (reqs) delete[] reqs;
|
|
||||||
if (stats) delete[] stats;
|
|
||||||
combined_src = combined_dst = 0;
|
|
||||||
send_lengths = recv_lengths = 0;
|
|
||||||
send_buf_caps = recv_buf_caps = 0;
|
|
||||||
send_bufs = recv_bufs = 0;
|
|
||||||
reqs = 0; stats = 0;
|
|
||||||
cpusize = 0; max_reqs = 0;
|
|
||||||
}
|
|
||||||
// transfer_cached: reuse pre-allocated buffers from SyncCache
|
|
||||||
void Parallel::transfer_cached(MyList<Parallel::gridseg> **src, MyList<Parallel::gridseg> **dst,
|
|
||||||
MyList<var> *VarList1, MyList<var> *VarList2,
|
|
||||||
int Symmetry, SyncCache &cache)
|
|
||||||
{
|
|
||||||
int myrank;
|
|
||||||
MPI_Comm_size(MPI_COMM_WORLD, &cache.cpusize);
|
|
||||||
MPI_Comm_rank(MPI_COMM_WORLD, &myrank);
|
|
||||||
int cpusize = cache.cpusize;
|
|
||||||
|
|
||||||
int req_no = 0;
|
|
||||||
int node;
|
|
||||||
|
|
||||||
for (node = 0; node < cpusize; node++)
|
|
||||||
{
|
|
||||||
if (node == myrank)
|
|
||||||
{
|
|
||||||
int length = data_packer(0, src[myrank], dst[myrank], node, PACK, VarList1, VarList2, Symmetry);
|
|
||||||
cache.recv_lengths[node] = length;
|
|
||||||
if (length > 0)
|
|
||||||
{
|
|
||||||
if (length > cache.recv_buf_caps[node])
|
|
||||||
{
|
|
||||||
if (cache.recv_bufs[node]) delete[] cache.recv_bufs[node];
|
|
||||||
cache.recv_bufs[node] = new double[length];
|
|
||||||
cache.recv_buf_caps[node] = length;
|
|
||||||
}
|
|
||||||
data_packer(cache.recv_bufs[node], src[myrank], dst[myrank], node, PACK, VarList1, VarList2, Symmetry);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
else
|
|
||||||
{
|
|
||||||
// send
|
|
||||||
int slength = data_packer(0, src[myrank], dst[myrank], node, PACK, VarList1, VarList2, Symmetry);
|
|
||||||
cache.send_lengths[node] = slength;
|
|
||||||
if (slength > 0)
|
|
||||||
{
|
|
||||||
if (slength > cache.send_buf_caps[node])
|
|
||||||
{
|
|
||||||
if (cache.send_bufs[node]) delete[] cache.send_bufs[node];
|
|
||||||
cache.send_bufs[node] = new double[slength];
|
|
||||||
cache.send_buf_caps[node] = slength;
|
|
||||||
}
|
|
||||||
data_packer(cache.send_bufs[node], src[myrank], dst[myrank], node, PACK, VarList1, VarList2, Symmetry);
|
|
||||||
MPI_Isend((void *)cache.send_bufs[node], slength, MPI_DOUBLE, node, 1, MPI_COMM_WORLD, cache.reqs + req_no++);
|
|
||||||
}
|
|
||||||
// recv
|
|
||||||
int rlength = data_packer(0, src[node], dst[node], node, UNPACK, VarList1, VarList2, Symmetry);
|
|
||||||
cache.recv_lengths[node] = rlength;
|
|
||||||
if (rlength > 0)
|
|
||||||
{
|
|
||||||
if (rlength > cache.recv_buf_caps[node])
|
|
||||||
{
|
|
||||||
if (cache.recv_bufs[node]) delete[] cache.recv_bufs[node];
|
|
||||||
cache.recv_bufs[node] = new double[rlength];
|
|
||||||
cache.recv_buf_caps[node] = rlength;
|
|
||||||
}
|
|
||||||
MPI_Irecv((void *)cache.recv_bufs[node], rlength, MPI_DOUBLE, node, 1, MPI_COMM_WORLD, cache.reqs + req_no++);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
MPI_Waitall(req_no, cache.reqs, cache.stats);
|
|
||||||
|
|
||||||
for (node = 0; node < cpusize; node++)
|
|
||||||
if (cache.recv_bufs[node] && cache.recv_lengths[node] > 0)
|
|
||||||
data_packer(cache.recv_bufs[node], src[node], dst[node], node, UNPACK, VarList1, VarList2, Symmetry);
|
|
||||||
}
|
|
||||||
// Sync_cached: build grid segment lists on first call, reuse on subsequent calls
|
|
||||||
void Parallel::Sync_cached(MyList<Patch> *PatL, MyList<var> *VarList, int Symmetry, SyncCache &cache)
|
|
||||||
{
|
|
||||||
if (!cache.valid)
|
|
||||||
{
|
|
||||||
int cpusize;
|
|
||||||
MPI_Comm_size(MPI_COMM_WORLD, &cpusize);
|
|
||||||
cache.cpusize = cpusize;
|
|
||||||
|
|
||||||
// Allocate cache arrays if needed
|
|
||||||
if (!cache.combined_src)
|
|
||||||
{
|
|
||||||
cache.combined_src = new MyList<Parallel::gridseg> *[cpusize];
|
|
||||||
cache.combined_dst = new MyList<Parallel::gridseg> *[cpusize];
|
|
||||||
cache.send_lengths = new int[cpusize];
|
|
||||||
cache.recv_lengths = new int[cpusize];
|
|
||||||
cache.send_bufs = new double *[cpusize];
|
|
||||||
cache.recv_bufs = new double *[cpusize];
|
|
||||||
cache.send_buf_caps = new int[cpusize];
|
|
||||||
cache.recv_buf_caps = new int[cpusize];
|
|
||||||
for (int i = 0; i < cpusize; i++)
|
|
||||||
{
|
|
||||||
cache.send_bufs[i] = cache.recv_bufs[i] = 0;
|
|
||||||
cache.send_buf_caps[i] = cache.recv_buf_caps[i] = 0;
|
|
||||||
}
|
|
||||||
cache.max_reqs = 2 * cpusize;
|
|
||||||
cache.reqs = new MPI_Request[cache.max_reqs];
|
|
||||||
cache.stats = new MPI_Status[cache.max_reqs];
|
|
||||||
}
|
|
||||||
|
|
||||||
for (int node = 0; node < cpusize; node++)
|
|
||||||
{
|
|
||||||
cache.combined_src[node] = cache.combined_dst[node] = 0;
|
|
||||||
cache.send_lengths[node] = cache.recv_lengths[node] = 0;
|
|
||||||
}
|
|
||||||
|
|
||||||
// Build intra-patch segments (same as Sync_merged Phase A)
|
|
||||||
MyList<Patch> *Pp = PatL;
|
|
||||||
while (Pp)
|
|
||||||
{
|
|
||||||
Patch *Pat = Pp->data;
|
|
||||||
MyList<Parallel::gridseg> *dst_ghost = build_ghost_gsl(Pat);
|
|
||||||
for (int node = 0; node < cpusize; node++)
|
|
||||||
{
|
|
||||||
MyList<Parallel::gridseg> *src_owned = build_owned_gsl0(Pat, node);
|
|
||||||
MyList<Parallel::gridseg> *tsrc = 0, *tdst = 0;
|
|
||||||
build_gstl(src_owned, dst_ghost, &tsrc, &tdst);
|
|
||||||
if (tsrc)
|
|
||||||
{
|
|
||||||
if (cache.combined_src[node])
|
|
||||||
cache.combined_src[node]->catList(tsrc);
|
|
||||||
else
|
|
||||||
cache.combined_src[node] = tsrc;
|
|
||||||
}
|
|
||||||
if (tdst)
|
|
||||||
{
|
|
||||||
if (cache.combined_dst[node])
|
|
||||||
cache.combined_dst[node]->catList(tdst);
|
|
||||||
else
|
|
||||||
cache.combined_dst[node] = tdst;
|
|
||||||
}
|
|
||||||
if (src_owned) src_owned->destroyList();
|
|
||||||
}
|
|
||||||
if (dst_ghost) dst_ghost->destroyList();
|
|
||||||
Pp = Pp->next;
|
|
||||||
}
|
|
||||||
|
|
||||||
// Build inter-patch segments (same as Sync_merged Phase B)
|
|
||||||
MyList<Parallel::gridseg> *dst_buffer = build_buffer_gsl(PatL);
|
|
||||||
for (int node = 0; node < cpusize; node++)
|
|
||||||
{
|
|
||||||
MyList<Parallel::gridseg> *src_owned = build_owned_gsl(PatL, node, 5, Symmetry);
|
|
||||||
MyList<Parallel::gridseg> *tsrc = 0, *tdst = 0;
|
|
||||||
build_gstl(src_owned, dst_buffer, &tsrc, &tdst);
|
|
||||||
if (tsrc)
|
|
||||||
{
|
|
||||||
if (cache.combined_src[node])
|
|
||||||
cache.combined_src[node]->catList(tsrc);
|
|
||||||
else
|
|
||||||
cache.combined_src[node] = tsrc;
|
|
||||||
}
|
|
||||||
if (tdst)
|
|
||||||
{
|
|
||||||
if (cache.combined_dst[node])
|
|
||||||
cache.combined_dst[node]->catList(tdst);
|
|
||||||
else
|
|
||||||
cache.combined_dst[node] = tdst;
|
|
||||||
}
|
|
||||||
if (src_owned) src_owned->destroyList();
|
|
||||||
}
|
|
||||||
if (dst_buffer) dst_buffer->destroyList();
|
|
||||||
|
|
||||||
cache.valid = true;
|
|
||||||
}
|
|
||||||
|
|
||||||
// Use cached lists with buffer-reusing transfer
|
|
||||||
transfer_cached(cache.combined_src, cache.combined_dst, VarList, VarList, Symmetry, cache);
|
|
||||||
}
|
|
||||||
// Sync_start: pack and post MPI_Isend/Irecv, return immediately
|
|
||||||
void Parallel::Sync_start(MyList<Patch> *PatL, MyList<var> *VarList, int Symmetry,
|
|
||||||
SyncCache &cache, AsyncSyncState &state)
|
|
||||||
{
|
|
||||||
// Ensure cache is built
|
|
||||||
if (!cache.valid)
|
|
||||||
{
|
|
||||||
// Build cache (same logic as Sync_cached)
|
|
||||||
int cpusize;
|
|
||||||
MPI_Comm_size(MPI_COMM_WORLD, &cpusize);
|
|
||||||
cache.cpusize = cpusize;
|
|
||||||
|
|
||||||
if (!cache.combined_src)
|
|
||||||
{
|
|
||||||
cache.combined_src = new MyList<Parallel::gridseg> *[cpusize];
|
|
||||||
cache.combined_dst = new MyList<Parallel::gridseg> *[cpusize];
|
|
||||||
cache.send_lengths = new int[cpusize];
|
|
||||||
cache.recv_lengths = new int[cpusize];
|
|
||||||
cache.send_bufs = new double *[cpusize];
|
|
||||||
cache.recv_bufs = new double *[cpusize];
|
|
||||||
cache.send_buf_caps = new int[cpusize];
|
|
||||||
cache.recv_buf_caps = new int[cpusize];
|
|
||||||
for (int i = 0; i < cpusize; i++)
|
|
||||||
{
|
|
||||||
cache.send_bufs[i] = cache.recv_bufs[i] = 0;
|
|
||||||
cache.send_buf_caps[i] = cache.recv_buf_caps[i] = 0;
|
|
||||||
}
|
|
||||||
cache.max_reqs = 2 * cpusize;
|
|
||||||
cache.reqs = new MPI_Request[cache.max_reqs];
|
|
||||||
cache.stats = new MPI_Status[cache.max_reqs];
|
|
||||||
}
|
|
||||||
|
|
||||||
for (int node = 0; node < cpusize; node++)
|
|
||||||
{
|
|
||||||
cache.combined_src[node] = cache.combined_dst[node] = 0;
|
|
||||||
cache.send_lengths[node] = cache.recv_lengths[node] = 0;
|
|
||||||
}
|
|
||||||
|
|
||||||
MyList<Patch> *Pp = PatL;
|
|
||||||
while (Pp)
|
|
||||||
{
|
|
||||||
Patch *Pat = Pp->data;
|
|
||||||
MyList<Parallel::gridseg> *dst_ghost = build_ghost_gsl(Pat);
|
|
||||||
for (int node = 0; node < cpusize; node++)
|
|
||||||
{
|
|
||||||
MyList<Parallel::gridseg> *src_owned = build_owned_gsl0(Pat, node);
|
|
||||||
MyList<Parallel::gridseg> *tsrc = 0, *tdst = 0;
|
|
||||||
build_gstl(src_owned, dst_ghost, &tsrc, &tdst);
|
|
||||||
if (tsrc)
|
|
||||||
{
|
|
||||||
if (cache.combined_src[node])
|
|
||||||
cache.combined_src[node]->catList(tsrc);
|
|
||||||
else
|
|
||||||
cache.combined_src[node] = tsrc;
|
|
||||||
}
|
|
||||||
if (tdst)
|
|
||||||
{
|
|
||||||
if (cache.combined_dst[node])
|
|
||||||
cache.combined_dst[node]->catList(tdst);
|
|
||||||
else
|
|
||||||
cache.combined_dst[node] = tdst;
|
|
||||||
}
|
|
||||||
if (src_owned) src_owned->destroyList();
|
|
||||||
}
|
|
||||||
if (dst_ghost) dst_ghost->destroyList();
|
|
||||||
Pp = Pp->next;
|
|
||||||
}
|
|
||||||
|
|
||||||
MyList<Parallel::gridseg> *dst_buffer = build_buffer_gsl(PatL);
|
|
||||||
for (int node = 0; node < cpusize; node++)
|
|
||||||
{
|
|
||||||
MyList<Parallel::gridseg> *src_owned = build_owned_gsl(PatL, node, 5, Symmetry);
|
|
||||||
MyList<Parallel::gridseg> *tsrc = 0, *tdst = 0;
|
|
||||||
build_gstl(src_owned, dst_buffer, &tsrc, &tdst);
|
|
||||||
if (tsrc)
|
|
||||||
{
|
|
||||||
if (cache.combined_src[node])
|
|
||||||
cache.combined_src[node]->catList(tsrc);
|
|
||||||
else
|
|
||||||
cache.combined_src[node] = tsrc;
|
|
||||||
}
|
|
||||||
if (tdst)
|
|
||||||
{
|
|
||||||
if (cache.combined_dst[node])
|
|
||||||
cache.combined_dst[node]->catList(tdst);
|
|
||||||
else
|
|
||||||
cache.combined_dst[node] = tdst;
|
|
||||||
}
|
|
||||||
if (src_owned) src_owned->destroyList();
|
|
||||||
}
|
|
||||||
if (dst_buffer) dst_buffer->destroyList();
|
|
||||||
cache.valid = true;
|
|
||||||
}
|
|
||||||
|
|
||||||
// Now pack and post async MPI operations
|
|
||||||
int myrank;
|
|
||||||
MPI_Comm_rank(MPI_COMM_WORLD, &myrank);
|
|
||||||
int cpusize = cache.cpusize;
|
|
||||||
state.req_no = 0;
|
|
||||||
state.active = true;
|
|
||||||
|
|
||||||
MyList<Parallel::gridseg> **src = cache.combined_src;
|
|
||||||
MyList<Parallel::gridseg> **dst = cache.combined_dst;
|
|
||||||
|
|
||||||
for (int node = 0; node < cpusize; node++)
|
|
||||||
{
|
|
||||||
if (node == myrank)
|
|
||||||
{
|
|
||||||
int length = data_packer(0, src[myrank], dst[myrank], node, PACK, VarList, VarList, Symmetry);
|
|
||||||
cache.recv_lengths[node] = length;
|
|
||||||
if (length > 0)
|
|
||||||
{
|
|
||||||
if (length > cache.recv_buf_caps[node])
|
|
||||||
{
|
|
||||||
if (cache.recv_bufs[node]) delete[] cache.recv_bufs[node];
|
|
||||||
cache.recv_bufs[node] = new double[length];
|
|
||||||
cache.recv_buf_caps[node] = length;
|
|
||||||
}
|
|
||||||
data_packer(cache.recv_bufs[node], src[myrank], dst[myrank], node, PACK, VarList, VarList, Symmetry);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
else
|
|
||||||
{
|
|
||||||
int slength = data_packer(0, src[myrank], dst[myrank], node, PACK, VarList, VarList, Symmetry);
|
|
||||||
cache.send_lengths[node] = slength;
|
|
||||||
if (slength > 0)
|
|
||||||
{
|
|
||||||
if (slength > cache.send_buf_caps[node])
|
|
||||||
{
|
|
||||||
if (cache.send_bufs[node]) delete[] cache.send_bufs[node];
|
|
||||||
cache.send_bufs[node] = new double[slength];
|
|
||||||
cache.send_buf_caps[node] = slength;
|
|
||||||
}
|
|
||||||
data_packer(cache.send_bufs[node], src[myrank], dst[myrank], node, PACK, VarList, VarList, Symmetry);
|
|
||||||
MPI_Isend((void *)cache.send_bufs[node], slength, MPI_DOUBLE, node, 2, MPI_COMM_WORLD, cache.reqs + state.req_no++);
|
|
||||||
}
|
|
||||||
int rlength = data_packer(0, src[node], dst[node], node, UNPACK, VarList, VarList, Symmetry);
|
|
||||||
cache.recv_lengths[node] = rlength;
|
|
||||||
if (rlength > 0)
|
|
||||||
{
|
|
||||||
if (rlength > cache.recv_buf_caps[node])
|
|
||||||
{
|
|
||||||
if (cache.recv_bufs[node]) delete[] cache.recv_bufs[node];
|
|
||||||
cache.recv_bufs[node] = new double[rlength];
|
|
||||||
cache.recv_buf_caps[node] = rlength;
|
|
||||||
}
|
|
||||||
MPI_Irecv((void *)cache.recv_bufs[node], rlength, MPI_DOUBLE, node, 2, MPI_COMM_WORLD, cache.reqs + state.req_no++);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
// Sync_finish: wait for async MPI operations and unpack
|
|
||||||
void Parallel::Sync_finish(SyncCache &cache, AsyncSyncState &state,
|
|
||||||
MyList<var> *VarList, int Symmetry)
|
|
||||||
{
|
|
||||||
if (!state.active)
|
|
||||||
return;
|
|
||||||
|
|
||||||
MPI_Waitall(state.req_no, cache.reqs, cache.stats);
|
|
||||||
|
|
||||||
int cpusize = cache.cpusize;
|
|
||||||
MyList<Parallel::gridseg> **src = cache.combined_src;
|
|
||||||
MyList<Parallel::gridseg> **dst = cache.combined_dst;
|
|
||||||
|
|
||||||
for (int node = 0; node < cpusize; node++)
|
|
||||||
if (cache.recv_bufs[node] && cache.recv_lengths[node] > 0)
|
|
||||||
data_packer(cache.recv_bufs[node], src[node], dst[node], node, UNPACK, VarList, VarList, Symmetry);
|
|
||||||
|
|
||||||
state.active = false;
|
|
||||||
}
|
|
||||||
// collect buffer grid segments or blocks for the periodic boundary condition of given patch
|
// collect buffer grid segments or blocks for the periodic boundary condition of given patch
|
||||||
// ---------------------------------------------------
|
// ---------------------------------------------------
|
||||||
// |con | |con |
|
// |con | |con |
|
||||||
|
|||||||
@@ -81,42 +81,6 @@ namespace Parallel
|
|||||||
int Symmetry);
|
int Symmetry);
|
||||||
void Sync(Patch *Pat, MyList<var> *VarList, int Symmetry);
|
void Sync(Patch *Pat, MyList<var> *VarList, int Symmetry);
|
||||||
void Sync(MyList<Patch> *PatL, MyList<var> *VarList, int Symmetry);
|
void Sync(MyList<Patch> *PatL, MyList<var> *VarList, int Symmetry);
|
||||||
void Sync_merged(MyList<Patch> *PatL, MyList<var> *VarList, int Symmetry);
|
|
||||||
|
|
||||||
struct SyncCache {
|
|
||||||
bool valid;
|
|
||||||
int cpusize;
|
|
||||||
MyList<gridseg> **combined_src;
|
|
||||||
MyList<gridseg> **combined_dst;
|
|
||||||
int *send_lengths;
|
|
||||||
int *recv_lengths;
|
|
||||||
double **send_bufs;
|
|
||||||
double **recv_bufs;
|
|
||||||
int *send_buf_caps;
|
|
||||||
int *recv_buf_caps;
|
|
||||||
MPI_Request *reqs;
|
|
||||||
MPI_Status *stats;
|
|
||||||
int max_reqs;
|
|
||||||
SyncCache();
|
|
||||||
void invalidate();
|
|
||||||
void destroy();
|
|
||||||
};
|
|
||||||
|
|
||||||
void Sync_cached(MyList<Patch> *PatL, MyList<var> *VarList, int Symmetry, SyncCache &cache);
|
|
||||||
void transfer_cached(MyList<gridseg> **src, MyList<gridseg> **dst,
|
|
||||||
MyList<var> *VarList1, MyList<var> *VarList2,
|
|
||||||
int Symmetry, SyncCache &cache);
|
|
||||||
|
|
||||||
struct AsyncSyncState {
|
|
||||||
int req_no;
|
|
||||||
bool active;
|
|
||||||
AsyncSyncState() : req_no(0), active(false) {}
|
|
||||||
};
|
|
||||||
|
|
||||||
void Sync_start(MyList<Patch> *PatL, MyList<var> *VarList, int Symmetry,
|
|
||||||
SyncCache &cache, AsyncSyncState &state);
|
|
||||||
void Sync_finish(SyncCache &cache, AsyncSyncState &state,
|
|
||||||
MyList<var> *VarList, int Symmetry);
|
|
||||||
void OutBdLow2Hi(Patch *Patc, Patch *Patf,
|
void OutBdLow2Hi(Patch *Patc, Patch *Patf,
|
||||||
MyList<var> *VarList1 /* source */, MyList<var> *VarList2 /* target */,
|
MyList<var> *VarList1 /* source */, MyList<var> *VarList2 /* target */,
|
||||||
int Symmetry);
|
int Symmetry);
|
||||||
|
|||||||
@@ -2,11 +2,7 @@
|
|||||||
#ifndef SHELLPATCH_H
|
#ifndef SHELLPATCH_H
|
||||||
#define SHELLPATCH_H
|
#define SHELLPATCH_H
|
||||||
|
|
||||||
#ifdef MPI_STUB
|
|
||||||
#include "mpi_stub.h"
|
|
||||||
#else
|
|
||||||
#include <mpi.h>
|
#include <mpi.h>
|
||||||
#endif
|
|
||||||
#include "MyList.h"
|
#include "MyList.h"
|
||||||
#include "Block.h"
|
#include "Block.h"
|
||||||
#include "Parallel.h"
|
#include "Parallel.h"
|
||||||
|
|||||||
File diff suppressed because it is too large
Load Diff
@@ -1,8 +1,7 @@
|
|||||||
|
|
||||||
#ifndef TWO_PUNCTURES_H
|
#ifndef TWO_PUNCTURES_H
|
||||||
#define TWO_PUNCTURES_H
|
#define TWO_PUNCTURES_H
|
||||||
|
|
||||||
#include <omp.h>
|
|
||||||
|
|
||||||
#define StencilSize 19
|
#define StencilSize 19
|
||||||
#define N_PlaneRelax 1
|
#define N_PlaneRelax 1
|
||||||
#define NRELAX 200
|
#define NRELAX 200
|
||||||
@@ -43,18 +42,6 @@ private:
|
|||||||
|
|
||||||
int ntotal;
|
int ntotal;
|
||||||
|
|
||||||
// ===== Precomputed spectral derivative matrices =====
|
|
||||||
double *D1_A, *D2_A;
|
|
||||||
double *D1_B, *D2_B;
|
|
||||||
double *DF1_phi, *DF2_phi;
|
|
||||||
|
|
||||||
// ===== Pre-allocated workspace for LineRelax (per-thread) =====
|
|
||||||
int max_threads;
|
|
||||||
double **ws_diag_be, **ws_e_be, **ws_f_be, **ws_b_be, **ws_x_be;
|
|
||||||
double **ws_l_be, **ws_u_be, **ws_d_be, **ws_y_be;
|
|
||||||
double **ws_diag_al, **ws_e_al, **ws_f_al, **ws_b_al, **ws_x_al;
|
|
||||||
double **ws_l_al, **ws_u_al, **ws_d_al, **ws_y_al;
|
|
||||||
|
|
||||||
struct parameters
|
struct parameters
|
||||||
{
|
{
|
||||||
int nvar, n1, n2, n3;
|
int nvar, n1, n2, n3;
|
||||||
@@ -71,28 +58,6 @@ public:
|
|||||||
int Newtonmaxit);
|
int Newtonmaxit);
|
||||||
~TwoPunctures();
|
~TwoPunctures();
|
||||||
|
|
||||||
// 02/07: New/modified methods
|
|
||||||
void allocate_workspace();
|
|
||||||
void free_workspace();
|
|
||||||
void precompute_derivative_matrices();
|
|
||||||
void build_cheb_deriv_matrices(int n, double *D1, double *D2);
|
|
||||||
void build_fourier_deriv_matrices(int N, double *DF1, double *DF2);
|
|
||||||
void Derivatives_AB3_MatMul(int nvar, int n1, int n2, int n3, derivs v);
|
|
||||||
void ThomasAlgorithm_ws(int N, double *b, double *a, double *c, double *x, double *q,
|
|
||||||
double *l, double *u_ws, double *d, double *y);
|
|
||||||
void LineRelax_be_omp(double *dv,
|
|
||||||
int const i, int const k, int const nvar,
|
|
||||||
int const n1, int const n2, int const n3,
|
|
||||||
double const *rhs, int const *ncols, int **cols,
|
|
||||||
double **JFD, int tid);
|
|
||||||
void LineRelax_al_omp(double *dv,
|
|
||||||
int const j, int const k, int const nvar,
|
|
||||||
int const n1, int const n2, int const n3,
|
|
||||||
double const *rhs, int const *ncols,
|
|
||||||
int **cols, double **JFD, int tid);
|
|
||||||
void relax_omp(double *dv, int const nvar, int const n1, int const n2, int const n3,
|
|
||||||
double const *rhs, int const *ncols, int **cols, double **JFD);
|
|
||||||
|
|
||||||
void Solve();
|
void Solve();
|
||||||
void set_initial_guess(derivs v);
|
void set_initial_guess(derivs v);
|
||||||
int index(int i, int j, int k, int l, int a, int b, int c, int d);
|
int index(int i, int j, int k, int l, int a, int b, int c, int d);
|
||||||
@@ -151,11 +116,23 @@ public:
|
|||||||
double BY_KKofxyz(double x, double y, double z);
|
double BY_KKofxyz(double x, double y, double z);
|
||||||
void SetMatrix_JFD(int nvar, int n1, int n2, int n3, derivs u, int *ncols, int **cols, double **Matrix);
|
void SetMatrix_JFD(int nvar, int n1, int n2, int n3, derivs u, int *ncols, int **cols, double **Matrix);
|
||||||
void J_times_dv(int nvar, int n1, int n2, int n3, derivs dv, double *Jdv, derivs u);
|
void J_times_dv(int nvar, int n1, int n2, int n3, derivs dv, double *Jdv, derivs u);
|
||||||
|
void relax(double *dv, int const nvar, int const n1, int const n2, int const n3,
|
||||||
|
double const *rhs, int const *ncols, int **cols, double **JFD);
|
||||||
|
void LineRelax_be(double *dv,
|
||||||
|
int const i, int const k, int const nvar,
|
||||||
|
int const n1, int const n2, int const n3,
|
||||||
|
double const *rhs, int const *ncols, int **cols,
|
||||||
|
double **JFD);
|
||||||
void JFD_times_dv(int i, int j, int k, int nvar, int n1, int n2,
|
void JFD_times_dv(int i, int j, int k, int nvar, int n1, int n2,
|
||||||
int n3, derivs dv, derivs u, double *values);
|
int n3, derivs dv, derivs u, double *values);
|
||||||
void LinEquations(double A, double B, double X, double R,
|
void LinEquations(double A, double B, double X, double R,
|
||||||
double x, double r, double phi,
|
double x, double r, double phi,
|
||||||
double y, double z, derivs dU, derivs U, double *values);
|
double y, double z, derivs dU, derivs U, double *values);
|
||||||
|
void LineRelax_al(double *dv,
|
||||||
|
int const j, int const k, int const nvar,
|
||||||
|
int const n1, int const n2, int const n3,
|
||||||
|
double const *rhs, int const *ncols,
|
||||||
|
int **cols, double **JFD);
|
||||||
void ThomasAlgorithm(int N, double *b, double *a, double *c, double *x, double *q);
|
void ThomasAlgorithm(int N, double *b, double *a, double *c, double *x, double *q);
|
||||||
void Save(char *fname);
|
void Save(char *fname);
|
||||||
// provided by Vasileios Paschalidis (vpaschal@illinois.edu)
|
// provided by Vasileios Paschalidis (vpaschal@illinois.edu)
|
||||||
|
|||||||
@@ -19,11 +19,7 @@ using namespace std;
|
|||||||
#include <math.h>
|
#include <math.h>
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
#ifdef MPI_STUB
|
|
||||||
#include "mpi_stub.h"
|
|
||||||
#else
|
|
||||||
#include <mpi.h>
|
#include <mpi.h>
|
||||||
#endif
|
|
||||||
|
|
||||||
#include "cgh.h"
|
#include "cgh.h"
|
||||||
#include "ShellPatch.h"
|
#include "ShellPatch.h"
|
||||||
|
|||||||
@@ -19,11 +19,7 @@ using namespace std;
|
|||||||
#include <math.h>
|
#include <math.h>
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
#ifdef MPI_STUB
|
|
||||||
#include "mpi_stub.h"
|
|
||||||
#else
|
|
||||||
#include <mpi.h>
|
#include <mpi.h>
|
||||||
#endif
|
|
||||||
|
|
||||||
#include "cgh.h"
|
#include "cgh.h"
|
||||||
#include "ShellPatch.h"
|
#include "ShellPatch.h"
|
||||||
|
|||||||
@@ -19,11 +19,7 @@ using namespace std;
|
|||||||
#include <math.h>
|
#include <math.h>
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
#ifdef MPI_STUB
|
|
||||||
#include "mpi_stub.h"
|
|
||||||
#else
|
|
||||||
#include <mpi.h>
|
#include <mpi.h>
|
||||||
#endif
|
|
||||||
|
|
||||||
#include "cgh.h"
|
#include "cgh.h"
|
||||||
#include "ShellPatch.h"
|
#include "ShellPatch.h"
|
||||||
|
|||||||
@@ -730,12 +730,6 @@ void bssn_class::Initialize()
|
|||||||
PhysTime = StartTime;
|
PhysTime = StartTime;
|
||||||
Setup_Black_Hole_position();
|
Setup_Black_Hole_position();
|
||||||
}
|
}
|
||||||
|
|
||||||
// Initialize sync caches (per-level, for predictor and corrector)
|
|
||||||
sync_cache_pre = new Parallel::SyncCache[GH->levels];
|
|
||||||
sync_cache_cor = new Parallel::SyncCache[GH->levels];
|
|
||||||
sync_cache_rp_coarse = new Parallel::SyncCache[GH->levels];
|
|
||||||
sync_cache_rp_fine = new Parallel::SyncCache[GH->levels];
|
|
||||||
}
|
}
|
||||||
|
|
||||||
//================================================================================================
|
//================================================================================================
|
||||||
@@ -987,32 +981,6 @@ bssn_class::~bssn_class()
|
|||||||
delete Azzz;
|
delete Azzz;
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
// Destroy sync caches before GH
|
|
||||||
if (sync_cache_pre)
|
|
||||||
{
|
|
||||||
for (int i = 0; i < GH->levels; i++)
|
|
||||||
sync_cache_pre[i].destroy();
|
|
||||||
delete[] sync_cache_pre;
|
|
||||||
}
|
|
||||||
if (sync_cache_cor)
|
|
||||||
{
|
|
||||||
for (int i = 0; i < GH->levels; i++)
|
|
||||||
sync_cache_cor[i].destroy();
|
|
||||||
delete[] sync_cache_cor;
|
|
||||||
}
|
|
||||||
if (sync_cache_rp_coarse)
|
|
||||||
{
|
|
||||||
for (int i = 0; i < GH->levels; i++)
|
|
||||||
sync_cache_rp_coarse[i].destroy();
|
|
||||||
delete[] sync_cache_rp_coarse;
|
|
||||||
}
|
|
||||||
if (sync_cache_rp_fine)
|
|
||||||
{
|
|
||||||
for (int i = 0; i < GH->levels; i++)
|
|
||||||
sync_cache_rp_fine[i].destroy();
|
|
||||||
delete[] sync_cache_rp_fine;
|
|
||||||
}
|
|
||||||
|
|
||||||
delete GH;
|
delete GH;
|
||||||
#ifdef WithShell
|
#ifdef WithShell
|
||||||
delete SH;
|
delete SH;
|
||||||
@@ -2213,7 +2181,6 @@ void bssn_class::Evolve(int Steps)
|
|||||||
GH->Regrid(Symmetry, BH_num, Porgbr, Porg0,
|
GH->Regrid(Symmetry, BH_num, Porgbr, Porg0,
|
||||||
SynchList_cor, OldStateList, StateList, SynchList_pre,
|
SynchList_cor, OldStateList, StateList, SynchList_pre,
|
||||||
fgt(PhysTime - dT_mon, StartTime, dT_mon / 2), ErrorMonitor);
|
fgt(PhysTime - dT_mon, StartTime, dT_mon / 2), ErrorMonitor);
|
||||||
for (int il = 0; il < GH->levels; il++) { sync_cache_pre[il].invalidate(); sync_cache_cor[il].invalidate(); sync_cache_rp_coarse[il].invalidate(); sync_cache_rp_fine[il].invalidate(); }
|
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
#if (REGLEV == 0 && (PSTR == 1 || PSTR == 2))
|
#if (REGLEV == 0 && (PSTR == 1 || PSTR == 2))
|
||||||
@@ -2429,7 +2396,6 @@ void bssn_class::RecursiveStep(int lev)
|
|||||||
GH->Regrid_Onelevel(lev, Symmetry, BH_num, Porgbr, Porg0,
|
GH->Regrid_Onelevel(lev, Symmetry, BH_num, Porgbr, Porg0,
|
||||||
SynchList_cor, OldStateList, StateList, SynchList_pre,
|
SynchList_cor, OldStateList, StateList, SynchList_pre,
|
||||||
fgt(PhysTime - dT_lev, StartTime, dT_lev / 2), ErrorMonitor);
|
fgt(PhysTime - dT_lev, StartTime, dT_lev / 2), ErrorMonitor);
|
||||||
for (int il = 0; il < GH->levels; il++) { sync_cache_pre[il].invalidate(); sync_cache_cor[il].invalidate(); sync_cache_rp_coarse[il].invalidate(); sync_cache_rp_fine[il].invalidate(); }
|
|
||||||
#endif
|
#endif
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -2608,7 +2574,6 @@ void bssn_class::ParallelStep()
|
|||||||
GH->Regrid_Onelevel(GH->mylev, Symmetry, BH_num, Porgbr, Porg0,
|
GH->Regrid_Onelevel(GH->mylev, Symmetry, BH_num, Porgbr, Porg0,
|
||||||
SynchList_cor, OldStateList, StateList, SynchList_pre,
|
SynchList_cor, OldStateList, StateList, SynchList_pre,
|
||||||
fgt(PhysTime - dT_lev, StartTime, dT_lev / 2), ErrorMonitor);
|
fgt(PhysTime - dT_lev, StartTime, dT_lev / 2), ErrorMonitor);
|
||||||
for (int il = 0; il < GH->levels; il++) { sync_cache_pre[il].invalidate(); sync_cache_cor[il].invalidate(); sync_cache_rp_coarse[il].invalidate(); sync_cache_rp_fine[il].invalidate(); }
|
|
||||||
#endif
|
#endif
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -2775,7 +2740,6 @@ void bssn_class::ParallelStep()
|
|||||||
GH->Regrid_Onelevel(lev + 1, Symmetry, BH_num, Porgbr, Porg0,
|
GH->Regrid_Onelevel(lev + 1, Symmetry, BH_num, Porgbr, Porg0,
|
||||||
SynchList_cor, OldStateList, StateList, SynchList_pre,
|
SynchList_cor, OldStateList, StateList, SynchList_pre,
|
||||||
fgt(PhysTime - dT_levp1, StartTime, dT_levp1 / 2), ErrorMonitor);
|
fgt(PhysTime - dT_levp1, StartTime, dT_levp1 / 2), ErrorMonitor);
|
||||||
for (int il = 0; il < GH->levels; il++) { sync_cache_pre[il].invalidate(); sync_cache_cor[il].invalidate(); sync_cache_rp_coarse[il].invalidate(); sync_cache_rp_fine[il].invalidate(); }
|
|
||||||
|
|
||||||
// a_stream.clear();
|
// a_stream.clear();
|
||||||
// a_stream.str("");
|
// a_stream.str("");
|
||||||
@@ -2790,7 +2754,6 @@ void bssn_class::ParallelStep()
|
|||||||
GH->Regrid_Onelevel(lev, Symmetry, BH_num, Porgbr, Porg0,
|
GH->Regrid_Onelevel(lev, Symmetry, BH_num, Porgbr, Porg0,
|
||||||
SynchList_cor, OldStateList, StateList, SynchList_pre,
|
SynchList_cor, OldStateList, StateList, SynchList_pre,
|
||||||
fgt(PhysTime - dT_lev, StartTime, dT_lev / 2), ErrorMonitor);
|
fgt(PhysTime - dT_lev, StartTime, dT_lev / 2), ErrorMonitor);
|
||||||
for (int il = 0; il < GH->levels; il++) { sync_cache_pre[il].invalidate(); sync_cache_cor[il].invalidate(); sync_cache_rp_coarse[il].invalidate(); sync_cache_rp_fine[il].invalidate(); }
|
|
||||||
|
|
||||||
// a_stream.clear();
|
// a_stream.clear();
|
||||||
// a_stream.str("");
|
// a_stream.str("");
|
||||||
@@ -2809,7 +2772,6 @@ void bssn_class::ParallelStep()
|
|||||||
GH->Regrid_Onelevel(lev - 1, Symmetry, BH_num, Porgbr, Porg0,
|
GH->Regrid_Onelevel(lev - 1, Symmetry, BH_num, Porgbr, Porg0,
|
||||||
SynchList_cor, OldStateList, StateList, SynchList_pre,
|
SynchList_cor, OldStateList, StateList, SynchList_pre,
|
||||||
fgt(PhysTime - dT_lev, StartTime, dT_levm1 / 2), ErrorMonitor);
|
fgt(PhysTime - dT_lev, StartTime, dT_levm1 / 2), ErrorMonitor);
|
||||||
for (int il = 0; il < GH->levels; il++) { sync_cache_pre[il].invalidate(); sync_cache_cor[il].invalidate(); sync_cache_rp_coarse[il].invalidate(); sync_cache_rp_fine[il].invalidate(); }
|
|
||||||
|
|
||||||
// a_stream.clear();
|
// a_stream.clear();
|
||||||
// a_stream.str("");
|
// a_stream.str("");
|
||||||
@@ -2825,7 +2787,6 @@ void bssn_class::ParallelStep()
|
|||||||
GH->Regrid_Onelevel(lev - 1, Symmetry, BH_num, Porgbr, Porg0,
|
GH->Regrid_Onelevel(lev - 1, Symmetry, BH_num, Porgbr, Porg0,
|
||||||
SynchList_cor, OldStateList, StateList, SynchList_pre,
|
SynchList_cor, OldStateList, StateList, SynchList_pre,
|
||||||
fgt(PhysTime - dT_lev, StartTime, dT_levm1 / 2), ErrorMonitor);
|
fgt(PhysTime - dT_lev, StartTime, dT_levm1 / 2), ErrorMonitor);
|
||||||
for (int il = 0; il < GH->levels; il++) { sync_cache_pre[il].invalidate(); sync_cache_cor[il].invalidate(); sync_cache_rp_coarse[il].invalidate(); sync_cache_rp_fine[il].invalidate(); }
|
|
||||||
|
|
||||||
// a_stream.clear();
|
// a_stream.clear();
|
||||||
// a_stream.str("");
|
// a_stream.str("");
|
||||||
@@ -3197,7 +3158,21 @@ void bssn_class::Step(int lev, int YN)
|
|||||||
}
|
}
|
||||||
Pp = Pp->next;
|
Pp = Pp->next;
|
||||||
}
|
}
|
||||||
// NOTE: error check deferred to after Shell Patch computation to reduce MPI_Allreduce calls
|
// check error information
|
||||||
|
{
|
||||||
|
int erh = ERROR;
|
||||||
|
MPI_Allreduce(&erh, &ERROR, 1, MPI_INT, MPI_SUM, MPI_COMM_WORLD);
|
||||||
|
}
|
||||||
|
if (ERROR)
|
||||||
|
{
|
||||||
|
Parallel::Dump_Data(GH->PatL[lev], StateList, 0, PhysTime, dT_lev);
|
||||||
|
if (myrank == 0)
|
||||||
|
{
|
||||||
|
if (ErrorMonitor->outfile)
|
||||||
|
ErrorMonitor->outfile << "find NaN in state variables at t = " << PhysTime << ", lev = " << lev << endl;
|
||||||
|
MPI_Abort(MPI_COMM_WORLD, 1);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
#ifdef WithShell
|
#ifdef WithShell
|
||||||
// evolve Shell Patches
|
// evolve Shell Patches
|
||||||
@@ -3341,16 +3316,25 @@ void bssn_class::Step(int lev, int YN)
|
|||||||
#endif
|
#endif
|
||||||
}
|
}
|
||||||
|
|
||||||
// Non-blocking error reduction overlapped with Sync to hide Allreduce latency
|
// check error information
|
||||||
MPI_Request err_req;
|
|
||||||
{
|
{
|
||||||
int erh = ERROR;
|
int erh = ERROR;
|
||||||
MPI_Iallreduce(&erh, &ERROR, 1, MPI_INT, MPI_SUM, MPI_COMM_WORLD, &err_req);
|
MPI_Allreduce(&erh, &ERROR, 1, MPI_INT, MPI_SUM, MPI_COMM_WORLD);
|
||||||
|
}
|
||||||
|
|
||||||
|
if (ERROR)
|
||||||
|
{
|
||||||
|
SH->Dump_Data(StateList, 0, PhysTime, dT_lev);
|
||||||
|
if (myrank == 0)
|
||||||
|
{
|
||||||
|
if (ErrorMonitor->outfile)
|
||||||
|
ErrorMonitor->outfile << "find NaN in state variables on Shell Patches at t = " << PhysTime << endl;
|
||||||
|
MPI_Abort(MPI_COMM_WORLD, 1);
|
||||||
|
}
|
||||||
}
|
}
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
Parallel::AsyncSyncState async_pre;
|
Parallel::Sync(GH->PatL[lev], SynchList_pre, Symmetry);
|
||||||
Parallel::Sync_start(GH->PatL[lev], SynchList_pre, Symmetry, sync_cache_pre[lev], async_pre);
|
|
||||||
|
|
||||||
#ifdef WithShell
|
#ifdef WithShell
|
||||||
if (lev == 0)
|
if (lev == 0)
|
||||||
@@ -3369,23 +3353,6 @@ void bssn_class::Step(int lev, int YN)
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
#endif
|
#endif
|
||||||
Parallel::Sync_finish(sync_cache_pre[lev], async_pre, SynchList_pre, Symmetry);
|
|
||||||
|
|
||||||
#ifdef WithShell
|
|
||||||
// Complete non-blocking error reduction and check
|
|
||||||
MPI_Wait(&err_req, MPI_STATUS_IGNORE);
|
|
||||||
if (ERROR)
|
|
||||||
{
|
|
||||||
Parallel::Dump_Data(GH->PatL[lev], StateList, 0, PhysTime, dT_lev);
|
|
||||||
SH->Dump_Data(StateList, 0, PhysTime, dT_lev);
|
|
||||||
if (myrank == 0)
|
|
||||||
{
|
|
||||||
if (ErrorMonitor->outfile)
|
|
||||||
ErrorMonitor->outfile << "find NaN in state variables at t = " << PhysTime << ", lev = " << lev << endl;
|
|
||||||
MPI_Abort(MPI_COMM_WORLD, 1);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
#endif
|
|
||||||
|
|
||||||
#if (MAPBH == 0)
|
#if (MAPBH == 0)
|
||||||
// for black hole position
|
// for black hole position
|
||||||
@@ -3561,7 +3528,24 @@ void bssn_class::Step(int lev, int YN)
|
|||||||
Pp = Pp->next;
|
Pp = Pp->next;
|
||||||
}
|
}
|
||||||
|
|
||||||
// NOTE: error check deferred to after Shell Patch computation to reduce MPI_Allreduce calls
|
// check error information
|
||||||
|
{
|
||||||
|
int erh = ERROR;
|
||||||
|
MPI_Allreduce(&erh, &ERROR, 1, MPI_INT, MPI_SUM, MPI_COMM_WORLD);
|
||||||
|
}
|
||||||
|
|
||||||
|
if (ERROR)
|
||||||
|
{
|
||||||
|
Parallel::Dump_Data(GH->PatL[lev], SynchList_pre, 0, PhysTime, dT_lev);
|
||||||
|
if (myrank == 0)
|
||||||
|
{
|
||||||
|
if (ErrorMonitor->outfile)
|
||||||
|
ErrorMonitor->outfile << "find NaN in RK4 substep#" << iter_count
|
||||||
|
<< " variables at t = " << PhysTime
|
||||||
|
<< ", lev = " << lev << endl;
|
||||||
|
MPI_Abort(MPI_COMM_WORLD, 1);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
#ifdef WithShell
|
#ifdef WithShell
|
||||||
// evolve Shell Patches
|
// evolve Shell Patches
|
||||||
@@ -3701,16 +3685,26 @@ void bssn_class::Step(int lev, int YN)
|
|||||||
sPp = sPp->next;
|
sPp = sPp->next;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
// Non-blocking error reduction overlapped with Sync to hide Allreduce latency
|
// check error information
|
||||||
MPI_Request err_req_cor;
|
|
||||||
{
|
{
|
||||||
int erh = ERROR;
|
int erh = ERROR;
|
||||||
MPI_Iallreduce(&erh, &ERROR, 1, MPI_INT, MPI_SUM, MPI_COMM_WORLD, &err_req_cor);
|
MPI_Allreduce(&erh, &ERROR, 1, MPI_INT, MPI_SUM, MPI_COMM_WORLD);
|
||||||
|
}
|
||||||
|
if (ERROR)
|
||||||
|
{
|
||||||
|
SH->Dump_Data(SynchList_pre, 0, PhysTime, dT_lev);
|
||||||
|
if (myrank == 0)
|
||||||
|
{
|
||||||
|
if (ErrorMonitor->outfile)
|
||||||
|
ErrorMonitor->outfile << "find NaN on Shell Patches in RK4 substep#"
|
||||||
|
<< iter_count << " variables at t = "
|
||||||
|
<< PhysTime << endl;
|
||||||
|
MPI_Abort(MPI_COMM_WORLD, 1);
|
||||||
|
}
|
||||||
}
|
}
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
Parallel::AsyncSyncState async_cor;
|
Parallel::Sync(GH->PatL[lev], SynchList_cor, Symmetry);
|
||||||
Parallel::Sync_start(GH->PatL[lev], SynchList_cor, Symmetry, sync_cache_cor[lev], async_cor);
|
|
||||||
|
|
||||||
#ifdef WithShell
|
#ifdef WithShell
|
||||||
if (lev == 0)
|
if (lev == 0)
|
||||||
@@ -3729,25 +3723,6 @@ void bssn_class::Step(int lev, int YN)
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
#endif
|
#endif
|
||||||
Parallel::Sync_finish(sync_cache_cor[lev], async_cor, SynchList_cor, Symmetry);
|
|
||||||
|
|
||||||
#ifdef WithShell
|
|
||||||
// Complete non-blocking error reduction and check
|
|
||||||
MPI_Wait(&err_req_cor, MPI_STATUS_IGNORE);
|
|
||||||
if (ERROR)
|
|
||||||
{
|
|
||||||
Parallel::Dump_Data(GH->PatL[lev], SynchList_pre, 0, PhysTime, dT_lev);
|
|
||||||
SH->Dump_Data(SynchList_pre, 0, PhysTime, dT_lev);
|
|
||||||
if (myrank == 0)
|
|
||||||
{
|
|
||||||
if (ErrorMonitor->outfile)
|
|
||||||
ErrorMonitor->outfile << "find NaN in RK4 substep#" << iter_count
|
|
||||||
<< " variables at t = " << PhysTime
|
|
||||||
<< ", lev = " << lev << endl;
|
|
||||||
MPI_Abort(MPI_COMM_WORLD, 1);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
#endif
|
|
||||||
|
|
||||||
#if (MAPBH == 0)
|
#if (MAPBH == 0)
|
||||||
// for black hole position
|
// for black hole position
|
||||||
@@ -4059,7 +4034,22 @@ void bssn_class::Step(int lev, int YN)
|
|||||||
}
|
}
|
||||||
Pp = Pp->next;
|
Pp = Pp->next;
|
||||||
}
|
}
|
||||||
// NOTE: error check deferred to after Shell Patch computation to reduce MPI_Allreduce calls
|
// check error information
|
||||||
|
{
|
||||||
|
int erh = ERROR;
|
||||||
|
MPI_Allreduce(&erh, &ERROR, 1, MPI_INT, MPI_SUM, MPI_COMM_WORLD);
|
||||||
|
}
|
||||||
|
if (ERROR)
|
||||||
|
{
|
||||||
|
Parallel::Dump_Data(GH->PatL[lev], StateList, 0, PhysTime, dT_lev);
|
||||||
|
if (myrank == 0)
|
||||||
|
{
|
||||||
|
if (ErrorMonitor->outfile)
|
||||||
|
ErrorMonitor->outfile << "find NaN in state variables at t = " << PhysTime
|
||||||
|
<< ", lev = " << lev << endl;
|
||||||
|
MPI_Abort(MPI_COMM_WORLD, 1);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
#ifdef WithShell
|
#ifdef WithShell
|
||||||
// evolve Shell Patches
|
// evolve Shell Patches
|
||||||
@@ -4200,16 +4190,25 @@ void bssn_class::Step(int lev, int YN)
|
|||||||
}
|
}
|
||||||
#endif
|
#endif
|
||||||
}
|
}
|
||||||
// Non-blocking error reduction overlapped with Sync to hide Allreduce latency
|
// check error information
|
||||||
MPI_Request err_req;
|
|
||||||
{
|
{
|
||||||
int erh = ERROR;
|
int erh = ERROR;
|
||||||
MPI_Iallreduce(&erh, &ERROR, 1, MPI_INT, MPI_SUM, MPI_COMM_WORLD, &err_req);
|
MPI_Allreduce(&erh, &ERROR, 1, MPI_INT, MPI_SUM, MPI_COMM_WORLD);
|
||||||
|
}
|
||||||
|
if (ERROR)
|
||||||
|
{
|
||||||
|
SH->Dump_Data(StateList, 0, PhysTime, dT_lev);
|
||||||
|
if (myrank == 0)
|
||||||
|
{
|
||||||
|
if (ErrorMonitor->outfile)
|
||||||
|
ErrorMonitor->outfile << "find NaN in state variables on Shell Patches at t = "
|
||||||
|
<< PhysTime << endl;
|
||||||
|
MPI_Abort(MPI_COMM_WORLD, 1);
|
||||||
|
}
|
||||||
}
|
}
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
Parallel::AsyncSyncState async_pre;
|
Parallel::Sync(GH->PatL[lev], SynchList_pre, Symmetry);
|
||||||
Parallel::Sync_start(GH->PatL[lev], SynchList_pre, Symmetry, sync_cache_pre[lev], async_pre);
|
|
||||||
|
|
||||||
#ifdef WithShell
|
#ifdef WithShell
|
||||||
if (lev == 0)
|
if (lev == 0)
|
||||||
@@ -4228,24 +4227,6 @@ void bssn_class::Step(int lev, int YN)
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
#endif
|
#endif
|
||||||
Parallel::Sync_finish(sync_cache_pre[lev], async_pre, SynchList_pre, Symmetry);
|
|
||||||
|
|
||||||
#ifdef WithShell
|
|
||||||
// Complete non-blocking error reduction and check
|
|
||||||
MPI_Wait(&err_req, MPI_STATUS_IGNORE);
|
|
||||||
if (ERROR)
|
|
||||||
{
|
|
||||||
Parallel::Dump_Data(GH->PatL[lev], StateList, 0, PhysTime, dT_lev);
|
|
||||||
SH->Dump_Data(StateList, 0, PhysTime, dT_lev);
|
|
||||||
if (myrank == 0)
|
|
||||||
{
|
|
||||||
if (ErrorMonitor->outfile)
|
|
||||||
ErrorMonitor->outfile << "find NaN in state variables at t = " << PhysTime
|
|
||||||
<< ", lev = " << lev << endl;
|
|
||||||
MPI_Abort(MPI_COMM_WORLD, 1);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
#endif
|
|
||||||
|
|
||||||
// for black hole position
|
// for black hole position
|
||||||
if (BH_num > 0 && lev == GH->levels - 1)
|
if (BH_num > 0 && lev == GH->levels - 1)
|
||||||
@@ -4405,7 +4386,23 @@ void bssn_class::Step(int lev, int YN)
|
|||||||
Pp = Pp->next;
|
Pp = Pp->next;
|
||||||
}
|
}
|
||||||
|
|
||||||
// NOTE: error check deferred to after Shell Patch computation to reduce MPI_Allreduce calls
|
// check error information
|
||||||
|
{
|
||||||
|
int erh = ERROR;
|
||||||
|
MPI_Allreduce(&erh, &ERROR, 1, MPI_INT, MPI_SUM, MPI_COMM_WORLD);
|
||||||
|
}
|
||||||
|
if (ERROR)
|
||||||
|
{
|
||||||
|
Parallel::Dump_Data(GH->PatL[lev], SynchList_pre, 0, PhysTime, dT_lev);
|
||||||
|
if (myrank == 0)
|
||||||
|
{
|
||||||
|
if (ErrorMonitor->outfile)
|
||||||
|
ErrorMonitor->outfile << "find NaN in RK4 substep#" << iter_count
|
||||||
|
<< " variables at t = " << PhysTime
|
||||||
|
<< ", lev = " << lev << endl;
|
||||||
|
MPI_Abort(MPI_COMM_WORLD, 1);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
#ifdef WithShell
|
#ifdef WithShell
|
||||||
// evolve Shell Patches
|
// evolve Shell Patches
|
||||||
@@ -4545,16 +4542,25 @@ void bssn_class::Step(int lev, int YN)
|
|||||||
sPp = sPp->next;
|
sPp = sPp->next;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
// Non-blocking error reduction overlapped with Sync to hide Allreduce latency
|
// check error information
|
||||||
MPI_Request err_req_cor;
|
|
||||||
{
|
{
|
||||||
int erh = ERROR;
|
int erh = ERROR;
|
||||||
MPI_Iallreduce(&erh, &ERROR, 1, MPI_INT, MPI_SUM, MPI_COMM_WORLD, &err_req_cor);
|
MPI_Allreduce(&erh, &ERROR, 1, MPI_INT, MPI_SUM, MPI_COMM_WORLD);
|
||||||
|
}
|
||||||
|
if (ERROR)
|
||||||
|
{
|
||||||
|
SH->Dump_Data(SynchList_pre, 0, PhysTime, dT_lev);
|
||||||
|
if (myrank == 0)
|
||||||
|
{
|
||||||
|
if (ErrorMonitor->outfile)
|
||||||
|
ErrorMonitor->outfile << "find NaN on Shell Patches in RK4 substep#" << iter_count
|
||||||
|
<< " variables at t = " << PhysTime << endl;
|
||||||
|
MPI_Abort(MPI_COMM_WORLD, 1);
|
||||||
|
}
|
||||||
}
|
}
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
Parallel::AsyncSyncState async_cor;
|
Parallel::Sync(GH->PatL[lev], SynchList_cor, Symmetry);
|
||||||
Parallel::Sync_start(GH->PatL[lev], SynchList_cor, Symmetry, sync_cache_cor[lev], async_cor);
|
|
||||||
|
|
||||||
#ifdef WithShell
|
#ifdef WithShell
|
||||||
if (lev == 0)
|
if (lev == 0)
|
||||||
@@ -4572,25 +4578,6 @@ void bssn_class::Step(int lev, int YN)
|
|||||||
<< " seconds! " << endl;
|
<< " seconds! " << endl;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
#endif
|
|
||||||
Parallel::Sync_finish(sync_cache_cor[lev], async_cor, SynchList_cor, Symmetry);
|
|
||||||
|
|
||||||
#ifdef WithShell
|
|
||||||
// Complete non-blocking error reduction and check
|
|
||||||
MPI_Wait(&err_req_cor, MPI_STATUS_IGNORE);
|
|
||||||
if (ERROR)
|
|
||||||
{
|
|
||||||
Parallel::Dump_Data(GH->PatL[lev], SynchList_pre, 0, PhysTime, dT_lev);
|
|
||||||
SH->Dump_Data(SynchList_pre, 0, PhysTime, dT_lev);
|
|
||||||
if (myrank == 0)
|
|
||||||
{
|
|
||||||
if (ErrorMonitor->outfile)
|
|
||||||
ErrorMonitor->outfile << "find NaN in RK4 substep#" << iter_count
|
|
||||||
<< " variables at t = " << PhysTime
|
|
||||||
<< ", lev = " << lev << endl;
|
|
||||||
MPI_Abort(MPI_COMM_WORLD, 1);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
#endif
|
#endif
|
||||||
// for black hole position
|
// for black hole position
|
||||||
if (BH_num > 0 && lev == GH->levels - 1)
|
if (BH_num > 0 && lev == GH->levels - 1)
|
||||||
@@ -4956,19 +4943,11 @@ void bssn_class::Step(int lev, int YN)
|
|||||||
|
|
||||||
// misc::tillherecheck(GH->Commlev[lev],GH->start_rank[lev],"after Predictor rhs calculation");
|
// misc::tillherecheck(GH->Commlev[lev],GH->start_rank[lev],"after Predictor rhs calculation");
|
||||||
|
|
||||||
// Non-blocking error reduction overlapped with Sync to hide Allreduce latency
|
// check error information
|
||||||
MPI_Request err_req;
|
|
||||||
{
|
{
|
||||||
int erh = ERROR;
|
int erh = ERROR;
|
||||||
MPI_Iallreduce(&erh, &ERROR, 1, MPI_INT, MPI_SUM, GH->Commlev[lev], &err_req);
|
MPI_Allreduce(&erh, &ERROR, 1, MPI_INT, MPI_SUM, GH->Commlev[lev]);
|
||||||
}
|
}
|
||||||
|
|
||||||
// misc::tillherecheck(GH->Commlev[lev],GH->start_rank[lev],"before Predictor sync");
|
|
||||||
|
|
||||||
Parallel::Sync_cached(GH->PatL[lev], SynchList_pre, Symmetry, sync_cache_pre[lev]);
|
|
||||||
|
|
||||||
// Complete non-blocking error reduction and check
|
|
||||||
MPI_Wait(&err_req, MPI_STATUS_IGNORE);
|
|
||||||
if (ERROR)
|
if (ERROR)
|
||||||
{
|
{
|
||||||
Parallel::Dump_Data(GH->PatL[lev], StateList, 0, PhysTime, dT_lev);
|
Parallel::Dump_Data(GH->PatL[lev], StateList, 0, PhysTime, dT_lev);
|
||||||
@@ -4980,6 +4959,10 @@ void bssn_class::Step(int lev, int YN)
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// misc::tillherecheck(GH->Commlev[lev],GH->start_rank[lev],"before Predictor sync");
|
||||||
|
|
||||||
|
Parallel::Sync(GH->PatL[lev], SynchList_pre, Symmetry);
|
||||||
|
|
||||||
#if (MAPBH == 0)
|
#if (MAPBH == 0)
|
||||||
// for black hole position
|
// for black hole position
|
||||||
if (BH_num > 0 && lev == GH->levels - 1)
|
if (BH_num > 0 && lev == GH->levels - 1)
|
||||||
@@ -5157,21 +5140,11 @@ void bssn_class::Step(int lev, int YN)
|
|||||||
|
|
||||||
// misc::tillherecheck(GH->Commlev[lev],GH->start_rank[lev],"before Corrector error check");
|
// misc::tillherecheck(GH->Commlev[lev],GH->start_rank[lev],"before Corrector error check");
|
||||||
|
|
||||||
// Non-blocking error reduction overlapped with Sync to hide Allreduce latency
|
// check error information
|
||||||
MPI_Request err_req_cor;
|
|
||||||
{
|
{
|
||||||
int erh = ERROR;
|
int erh = ERROR;
|
||||||
MPI_Iallreduce(&erh, &ERROR, 1, MPI_INT, MPI_SUM, GH->Commlev[lev], &err_req_cor);
|
MPI_Allreduce(&erh, &ERROR, 1, MPI_INT, MPI_SUM, GH->Commlev[lev]);
|
||||||
}
|
}
|
||||||
|
|
||||||
// misc::tillherecheck(GH->Commlev[lev],GH->start_rank[lev],"before Corrector sync");
|
|
||||||
|
|
||||||
Parallel::Sync_cached(GH->PatL[lev], SynchList_cor, Symmetry, sync_cache_cor[lev]);
|
|
||||||
|
|
||||||
// misc::tillherecheck(GH->Commlev[lev],GH->start_rank[lev],"after Corrector sync");
|
|
||||||
|
|
||||||
// Complete non-blocking error reduction and check
|
|
||||||
MPI_Wait(&err_req_cor, MPI_STATUS_IGNORE);
|
|
||||||
if (ERROR)
|
if (ERROR)
|
||||||
{
|
{
|
||||||
Parallel::Dump_Data(GH->PatL[lev], SynchList_pre, 0, PhysTime, dT_lev);
|
Parallel::Dump_Data(GH->PatL[lev], SynchList_pre, 0, PhysTime, dT_lev);
|
||||||
@@ -5185,6 +5158,12 @@ void bssn_class::Step(int lev, int YN)
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// misc::tillherecheck(GH->Commlev[lev],GH->start_rank[lev],"before Corrector sync");
|
||||||
|
|
||||||
|
Parallel::Sync(GH->PatL[lev], SynchList_cor, Symmetry);
|
||||||
|
|
||||||
|
// misc::tillherecheck(GH->Commlev[lev],GH->start_rank[lev],"after Corrector sync");
|
||||||
|
|
||||||
#if (MAPBH == 0)
|
#if (MAPBH == 0)
|
||||||
// for black hole position
|
// for black hole position
|
||||||
if (BH_num > 0 && lev == GH->levels - 1)
|
if (BH_num > 0 && lev == GH->levels - 1)
|
||||||
@@ -5468,11 +5447,21 @@ void bssn_class::SHStep()
|
|||||||
#if (PSTR == 1 || PSTR == 2)
|
#if (PSTR == 1 || PSTR == 2)
|
||||||
// misc::tillherecheck(GH->Commlev[lev],GH->start_rank[lev],"before Predictor's error check");
|
// misc::tillherecheck(GH->Commlev[lev],GH->start_rank[lev],"before Predictor's error check");
|
||||||
#endif
|
#endif
|
||||||
// Non-blocking error reduction overlapped with Synch to hide Allreduce latency
|
// check error information
|
||||||
MPI_Request err_req;
|
|
||||||
{
|
{
|
||||||
int erh = ERROR;
|
int erh = ERROR;
|
||||||
MPI_Iallreduce(&erh, &ERROR, 1, MPI_INT, MPI_SUM, MPI_COMM_WORLD, &err_req);
|
MPI_Allreduce(&erh, &ERROR, 1, MPI_INT, MPI_SUM, MPI_COMM_WORLD);
|
||||||
|
}
|
||||||
|
|
||||||
|
if (ERROR)
|
||||||
|
{
|
||||||
|
SH->Dump_Data(StateList, 0, PhysTime, dT_lev);
|
||||||
|
if (myrank == 0)
|
||||||
|
{
|
||||||
|
if (ErrorMonitor->outfile)
|
||||||
|
ErrorMonitor->outfile << "find NaN in state variables on Shell Patches at t = " << PhysTime << endl;
|
||||||
|
MPI_Abort(MPI_COMM_WORLD, 1);
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
{
|
{
|
||||||
@@ -5490,19 +5479,6 @@ void bssn_class::SHStep()
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// Complete non-blocking error reduction and check
|
|
||||||
MPI_Wait(&err_req, MPI_STATUS_IGNORE);
|
|
||||||
if (ERROR)
|
|
||||||
{
|
|
||||||
SH->Dump_Data(StateList, 0, PhysTime, dT_lev);
|
|
||||||
if (myrank == 0)
|
|
||||||
{
|
|
||||||
if (ErrorMonitor->outfile)
|
|
||||||
ErrorMonitor->outfile << "find NaN in state variables on Shell Patches at t = " << PhysTime << endl;
|
|
||||||
MPI_Abort(MPI_COMM_WORLD, 1);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// corrector
|
// corrector
|
||||||
for (iter_count = 1; iter_count < 4; iter_count++)
|
for (iter_count = 1; iter_count < 4; iter_count++)
|
||||||
{
|
{
|
||||||
@@ -5645,11 +5621,21 @@ void bssn_class::SHStep()
|
|||||||
sPp = sPp->next;
|
sPp = sPp->next;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
// Non-blocking error reduction overlapped with Synch to hide Allreduce latency
|
// check error information
|
||||||
MPI_Request err_req_cor;
|
|
||||||
{
|
{
|
||||||
int erh = ERROR;
|
int erh = ERROR;
|
||||||
MPI_Iallreduce(&erh, &ERROR, 1, MPI_INT, MPI_SUM, MPI_COMM_WORLD, &err_req_cor);
|
MPI_Allreduce(&erh, &ERROR, 1, MPI_INT, MPI_SUM, MPI_COMM_WORLD);
|
||||||
|
}
|
||||||
|
if (ERROR)
|
||||||
|
{
|
||||||
|
SH->Dump_Data(SynchList_pre, 0, PhysTime, dT_lev);
|
||||||
|
if (myrank == 0)
|
||||||
|
{
|
||||||
|
if (ErrorMonitor->outfile)
|
||||||
|
ErrorMonitor->outfile << "find NaN on Shell Patches in RK4 substep#" << iter_count
|
||||||
|
<< " variables at t = " << PhysTime << endl;
|
||||||
|
MPI_Abort(MPI_COMM_WORLD, 1);
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
{
|
{
|
||||||
@@ -5667,20 +5653,6 @@ void bssn_class::SHStep()
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// Complete non-blocking error reduction and check
|
|
||||||
MPI_Wait(&err_req_cor, MPI_STATUS_IGNORE);
|
|
||||||
if (ERROR)
|
|
||||||
{
|
|
||||||
SH->Dump_Data(SynchList_pre, 0, PhysTime, dT_lev);
|
|
||||||
if (myrank == 0)
|
|
||||||
{
|
|
||||||
if (ErrorMonitor->outfile)
|
|
||||||
ErrorMonitor->outfile << "find NaN on Shell Patches in RK4 substep#" << iter_count
|
|
||||||
<< " variables at t = " << PhysTime << endl;
|
|
||||||
MPI_Abort(MPI_COMM_WORLD, 1);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
sPp = SH->PatL;
|
sPp = SH->PatL;
|
||||||
while (sPp)
|
while (sPp)
|
||||||
{
|
{
|
||||||
@@ -5809,7 +5781,7 @@ void bssn_class::RestrictProlong(int lev, int YN, bool BB,
|
|||||||
// misc::tillherecheck(GH->Commlev[GH->mylev],GH->start_rank[GH->mylev],a_stream.str());
|
// misc::tillherecheck(GH->Commlev[GH->mylev],GH->start_rank[GH->mylev],a_stream.str());
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
Parallel::Sync_cached(GH->PatL[lev - 1], SynchList_pre, Symmetry, sync_cache_rp_coarse[lev]);
|
Parallel::Sync(GH->PatL[lev - 1], SynchList_pre, Symmetry);
|
||||||
|
|
||||||
#if (PSTR == 1 || PSTR == 2)
|
#if (PSTR == 1 || PSTR == 2)
|
||||||
// a_stream.clear();
|
// a_stream.clear();
|
||||||
@@ -5870,7 +5842,7 @@ void bssn_class::RestrictProlong(int lev, int YN, bool BB,
|
|||||||
// misc::tillherecheck(GH->Commlev[GH->mylev],GH->start_rank[GH->mylev],a_stream.str());
|
// misc::tillherecheck(GH->Commlev[GH->mylev],GH->start_rank[GH->mylev],a_stream.str());
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
Parallel::Sync_cached(GH->PatL[lev - 1], SL, Symmetry, sync_cache_rp_coarse[lev]);
|
Parallel::Sync(GH->PatL[lev - 1], SL, Symmetry);
|
||||||
|
|
||||||
#if (PSTR == 1 || PSTR == 2)
|
#if (PSTR == 1 || PSTR == 2)
|
||||||
// a_stream.clear();
|
// a_stream.clear();
|
||||||
@@ -5908,7 +5880,7 @@ void bssn_class::RestrictProlong(int lev, int YN, bool BB,
|
|||||||
#endif
|
#endif
|
||||||
}
|
}
|
||||||
|
|
||||||
Parallel::Sync_cached(GH->PatL[lev], SL, Symmetry, sync_cache_rp_fine[lev]);
|
Parallel::Sync(GH->PatL[lev], SL, Symmetry);
|
||||||
|
|
||||||
#if (PSTR == 1 || PSTR == 2)
|
#if (PSTR == 1 || PSTR == 2)
|
||||||
// a_stream.clear();
|
// a_stream.clear();
|
||||||
@@ -5966,7 +5938,7 @@ void bssn_class::RestrictProlong_aux(int lev, int YN, bool BB,
|
|||||||
Parallel::Restrict_bam(GH->PatL[lev - 1], GH->PatL[lev], SL, SynchList_pre, GH->rsul[lev], Symmetry);
|
Parallel::Restrict_bam(GH->PatL[lev - 1], GH->PatL[lev], SL, SynchList_pre, GH->rsul[lev], Symmetry);
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
Parallel::Sync_cached(GH->PatL[lev - 1], SynchList_pre, Symmetry, sync_cache_rp_coarse[lev]);
|
Parallel::Sync(GH->PatL[lev - 1], SynchList_pre, Symmetry);
|
||||||
|
|
||||||
#if (RPB == 0)
|
#if (RPB == 0)
|
||||||
Ppc = GH->PatL[lev - 1];
|
Ppc = GH->PatL[lev - 1];
|
||||||
@@ -5998,7 +5970,7 @@ void bssn_class::RestrictProlong_aux(int lev, int YN, bool BB,
|
|||||||
Parallel::Restrict_bam(GH->PatL[lev - 1], GH->PatL[lev], SL, SL, GH->rsul[lev], Symmetry);
|
Parallel::Restrict_bam(GH->PatL[lev - 1], GH->PatL[lev], SL, SL, GH->rsul[lev], Symmetry);
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
Parallel::Sync_cached(GH->PatL[lev - 1], SL, Symmetry, sync_cache_rp_coarse[lev]);
|
Parallel::Sync(GH->PatL[lev - 1], SL, Symmetry);
|
||||||
|
|
||||||
#if (RPB == 0)
|
#if (RPB == 0)
|
||||||
Ppc = GH->PatL[lev - 1];
|
Ppc = GH->PatL[lev - 1];
|
||||||
@@ -6022,7 +5994,7 @@ void bssn_class::RestrictProlong_aux(int lev, int YN, bool BB,
|
|||||||
#endif
|
#endif
|
||||||
}
|
}
|
||||||
|
|
||||||
Parallel::Sync_cached(GH->PatL[lev], SL, Symmetry, sync_cache_rp_fine[lev]);
|
Parallel::Sync(GH->PatL[lev], SL, Symmetry);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -6073,7 +6045,7 @@ void bssn_class::RestrictProlong(int lev, int YN, bool BB)
|
|||||||
Parallel::Restrict_bam(GH->PatL[lev - 1], GH->PatL[lev], SynchList_cor, SynchList_pre, GH->rsul[lev], Symmetry);
|
Parallel::Restrict_bam(GH->PatL[lev - 1], GH->PatL[lev], SynchList_cor, SynchList_pre, GH->rsul[lev], Symmetry);
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
Parallel::Sync_cached(GH->PatL[lev - 1], SynchList_pre, Symmetry, sync_cache_rp_coarse[lev]);
|
Parallel::Sync(GH->PatL[lev - 1], SynchList_pre, Symmetry);
|
||||||
|
|
||||||
#if (RPB == 0)
|
#if (RPB == 0)
|
||||||
Ppc = GH->PatL[lev - 1];
|
Ppc = GH->PatL[lev - 1];
|
||||||
@@ -6107,7 +6079,7 @@ void bssn_class::RestrictProlong(int lev, int YN, bool BB)
|
|||||||
Parallel::Restrict_bam(GH->PatL[lev - 1], GH->PatL[lev], SynchList_cor, StateList, GH->rsul[lev], Symmetry);
|
Parallel::Restrict_bam(GH->PatL[lev - 1], GH->PatL[lev], SynchList_cor, StateList, GH->rsul[lev], Symmetry);
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
Parallel::Sync_cached(GH->PatL[lev - 1], StateList, Symmetry, sync_cache_rp_coarse[lev]);
|
Parallel::Sync(GH->PatL[lev - 1], StateList, Symmetry);
|
||||||
|
|
||||||
#if (RPB == 0)
|
#if (RPB == 0)
|
||||||
Ppc = GH->PatL[lev - 1];
|
Ppc = GH->PatL[lev - 1];
|
||||||
@@ -6131,7 +6103,7 @@ void bssn_class::RestrictProlong(int lev, int YN, bool BB)
|
|||||||
#endif
|
#endif
|
||||||
}
|
}
|
||||||
|
|
||||||
Parallel::Sync_cached(GH->PatL[lev], SynchList_cor, Symmetry, sync_cache_rp_fine[lev]);
|
Parallel::Sync(GH->PatL[lev], SynchList_cor, Symmetry);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -6214,10 +6186,10 @@ void bssn_class::ProlongRestrict(int lev, int YN, bool BB)
|
|||||||
#else
|
#else
|
||||||
Parallel::Restrict_after(GH->PatL[lev - 1], GH->PatL[lev], SynchList_cor, StateList, Symmetry);
|
Parallel::Restrict_after(GH->PatL[lev - 1], GH->PatL[lev], SynchList_cor, StateList, Symmetry);
|
||||||
#endif
|
#endif
|
||||||
Parallel::Sync_cached(GH->PatL[lev - 1], StateList, Symmetry, sync_cache_rp_coarse[lev]);
|
Parallel::Sync(GH->PatL[lev - 1], StateList, Symmetry);
|
||||||
}
|
}
|
||||||
|
|
||||||
Parallel::Sync_cached(GH->PatL[lev], SynchList_cor, Symmetry, sync_cache_rp_fine[lev]);
|
Parallel::Sync(GH->PatL[lev], SynchList_cor, Symmetry);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
#undef MIXOUTB
|
#undef MIXOUTB
|
||||||
|
|||||||
@@ -19,11 +19,7 @@ using namespace std;
|
|||||||
#include <math.h>
|
#include <math.h>
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
#ifdef MPI_STUB
|
|
||||||
#include "mpi_stub.h"
|
|
||||||
#else
|
|
||||||
#include <mpi.h>
|
#include <mpi.h>
|
||||||
#endif
|
|
||||||
|
|
||||||
#include "macrodef.h"
|
#include "macrodef.h"
|
||||||
#include "cgh.h"
|
#include "cgh.h"
|
||||||
@@ -130,11 +126,6 @@ public:
|
|||||||
MyList<var> *OldStateList, *DumpList;
|
MyList<var> *OldStateList, *DumpList;
|
||||||
MyList<var> *ConstraintList;
|
MyList<var> *ConstraintList;
|
||||||
|
|
||||||
Parallel::SyncCache *sync_cache_pre; // per-level cache for predictor sync
|
|
||||||
Parallel::SyncCache *sync_cache_cor; // per-level cache for corrector sync
|
|
||||||
Parallel::SyncCache *sync_cache_rp_coarse; // RestrictProlong sync on PatL[lev-1]
|
|
||||||
Parallel::SyncCache *sync_cache_rp_fine; // RestrictProlong sync on PatL[lev]
|
|
||||||
|
|
||||||
monitor *ErrorMonitor, *Psi4Monitor, *BHMonitor, *MAPMonitor;
|
monitor *ErrorMonitor, *Psi4Monitor, *BHMonitor, *MAPMonitor;
|
||||||
monitor *ConVMonitor;
|
monitor *ConVMonitor;
|
||||||
surface_integral *Waveshell;
|
surface_integral *Waveshell;
|
||||||
|
|||||||
@@ -19,11 +19,7 @@ using namespace std;
|
|||||||
#include <math.h>
|
#include <math.h>
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
#ifdef MPI_STUB
|
|
||||||
#include "mpi_stub.h"
|
|
||||||
#else
|
|
||||||
#include <mpi.h>
|
#include <mpi.h>
|
||||||
#endif
|
|
||||||
|
|
||||||
#include "macrodef.h"
|
#include "macrodef.h"
|
||||||
#include "cgh.h"
|
#include "cgh.h"
|
||||||
|
|||||||
@@ -106,8 +106,7 @@
|
|||||||
call getpbh(BHN,Porg,Mass)
|
call getpbh(BHN,Porg,Mass)
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
!!! sanity check (disabled in production builds for performance)
|
!!! sanity check
|
||||||
#ifdef DEBUG
|
|
||||||
dX = sum(chi)+sum(trK)+sum(dxx)+sum(gxy)+sum(gxz)+sum(dyy)+sum(gyz)+sum(dzz) &
|
dX = sum(chi)+sum(trK)+sum(dxx)+sum(gxy)+sum(gxz)+sum(dyy)+sum(gyz)+sum(dzz) &
|
||||||
+sum(Axx)+sum(Axy)+sum(Axz)+sum(Ayy)+sum(Ayz)+sum(Azz) &
|
+sum(Axx)+sum(Axy)+sum(Axz)+sum(Ayy)+sum(Ayz)+sum(Azz) &
|
||||||
+sum(Gamx)+sum(Gamy)+sum(Gamz) &
|
+sum(Gamx)+sum(Gamy)+sum(Gamz) &
|
||||||
@@ -137,7 +136,6 @@
|
|||||||
gont = 1
|
gont = 1
|
||||||
return
|
return
|
||||||
endif
|
endif
|
||||||
#endif
|
|
||||||
|
|
||||||
PI = dacos(-ONE)
|
PI = dacos(-ONE)
|
||||||
|
|
||||||
|
|||||||
@@ -20,11 +20,7 @@ using namespace std;
|
|||||||
#include <map.h>
|
#include <map.h>
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
#ifdef MPI_STUB
|
|
||||||
#include "mpi_stub.h"
|
|
||||||
#else
|
|
||||||
#include <mpi.h>
|
#include <mpi.h>
|
||||||
#endif
|
|
||||||
|
|
||||||
#include "macrodef.h"
|
#include "macrodef.h"
|
||||||
#include "misc.h"
|
#include "misc.h"
|
||||||
|
|||||||
@@ -2,11 +2,7 @@
|
|||||||
#ifndef CGH_H
|
#ifndef CGH_H
|
||||||
#define CGH_H
|
#define CGH_H
|
||||||
|
|
||||||
#ifdef MPI_STUB
|
|
||||||
#include "mpi_stub.h"
|
|
||||||
#else
|
|
||||||
#include <mpi.h>
|
#include <mpi.h>
|
||||||
#endif
|
|
||||||
#include "MyList.h"
|
#include "MyList.h"
|
||||||
#include "MPatch.h"
|
#include "MPatch.h"
|
||||||
#include "macrodef.h"
|
#include "macrodef.h"
|
||||||
|
|||||||
@@ -19,11 +19,7 @@ using namespace std;
|
|||||||
#include <time.h>
|
#include <time.h>
|
||||||
#include <stdlib.h>
|
#include <stdlib.h>
|
||||||
|
|
||||||
#ifdef MPI_STUB
|
|
||||||
#include "mpi_stub.h"
|
|
||||||
#else
|
|
||||||
#include <mpi.h>
|
#include <mpi.h>
|
||||||
#endif
|
|
||||||
|
|
||||||
#include "var.h"
|
#include "var.h"
|
||||||
#include "MyList.h"
|
#include "MyList.h"
|
||||||
|
|||||||
File diff suppressed because it is too large
Load Diff
@@ -81,7 +81,6 @@
|
|||||||
fy = ZEO
|
fy = ZEO
|
||||||
fz = ZEO
|
fz = ZEO
|
||||||
|
|
||||||
!$omp parallel do collapse(2) private(i,j,k)
|
|
||||||
do k=1,ex(3)
|
do k=1,ex(3)
|
||||||
do j=1,ex(2)
|
do j=1,ex(2)
|
||||||
do i=1,ex(1)
|
do i=1,ex(1)
|
||||||
@@ -180,7 +179,6 @@
|
|||||||
|
|
||||||
fx = ZEO
|
fx = ZEO
|
||||||
|
|
||||||
!$omp parallel do collapse(2) private(i,j,k)
|
|
||||||
do k=1,ex(3)
|
do k=1,ex(3)
|
||||||
do j=1,ex(2)
|
do j=1,ex(2)
|
||||||
do i=1,ex(1)
|
do i=1,ex(1)
|
||||||
@@ -264,7 +262,6 @@
|
|||||||
|
|
||||||
fy = ZEO
|
fy = ZEO
|
||||||
|
|
||||||
!$omp parallel do collapse(2) private(i,j,k)
|
|
||||||
do k=1,ex(3)
|
do k=1,ex(3)
|
||||||
do j=1,ex(2)
|
do j=1,ex(2)
|
||||||
do i=1,ex(1)
|
do i=1,ex(1)
|
||||||
@@ -345,7 +342,6 @@
|
|||||||
|
|
||||||
fz = ZEO
|
fz = ZEO
|
||||||
|
|
||||||
!$omp parallel do collapse(2) private(i,j,k)
|
|
||||||
do k=1,ex(3)
|
do k=1,ex(3)
|
||||||
do j=1,ex(2)
|
do j=1,ex(2)
|
||||||
do i=1,ex(1)
|
do i=1,ex(1)
|
||||||
@@ -447,7 +443,6 @@
|
|||||||
fxz = ZEO
|
fxz = ZEO
|
||||||
fyz = ZEO
|
fyz = ZEO
|
||||||
|
|
||||||
!$omp parallel do collapse(2) private(i,j,k)
|
|
||||||
do k=1,ex(3)
|
do k=1,ex(3)
|
||||||
do j=1,ex(2)
|
do j=1,ex(2)
|
||||||
do i=1,ex(1)
|
do i=1,ex(1)
|
||||||
@@ -558,7 +553,6 @@
|
|||||||
|
|
||||||
fxx = ZEO
|
fxx = ZEO
|
||||||
|
|
||||||
!$omp parallel do collapse(2) private(i,j,k)
|
|
||||||
do k=1,ex(3)
|
do k=1,ex(3)
|
||||||
do j=1,ex(2)
|
do j=1,ex(2)
|
||||||
do i=1,ex(1)
|
do i=1,ex(1)
|
||||||
@@ -633,7 +627,6 @@
|
|||||||
|
|
||||||
fyy = ZEO
|
fyy = ZEO
|
||||||
|
|
||||||
!$omp parallel do collapse(2) private(i,j,k)
|
|
||||||
do k=1,ex(3)
|
do k=1,ex(3)
|
||||||
do j=1,ex(2)
|
do j=1,ex(2)
|
||||||
do i=1,ex(1)
|
do i=1,ex(1)
|
||||||
@@ -709,7 +702,6 @@
|
|||||||
|
|
||||||
fzz = ZEO
|
fzz = ZEO
|
||||||
|
|
||||||
!$omp parallel do collapse(2) private(i,j,k)
|
|
||||||
do k=1,ex(3)
|
do k=1,ex(3)
|
||||||
do j=1,ex(2)
|
do j=1,ex(2)
|
||||||
do i=1,ex(1)
|
do i=1,ex(1)
|
||||||
@@ -785,7 +777,6 @@
|
|||||||
|
|
||||||
fxy = ZEO
|
fxy = ZEO
|
||||||
|
|
||||||
!$omp parallel do collapse(2) private(i,j,k)
|
|
||||||
do k=1,ex(3)
|
do k=1,ex(3)
|
||||||
do j=1,ex(2)
|
do j=1,ex(2)
|
||||||
do i=1,ex(1)
|
do i=1,ex(1)
|
||||||
@@ -860,7 +851,6 @@
|
|||||||
|
|
||||||
fxz = ZEO
|
fxz = ZEO
|
||||||
|
|
||||||
!$omp parallel do collapse(2) private(i,j,k)
|
|
||||||
do k=1,ex(3)
|
do k=1,ex(3)
|
||||||
do j=1,ex(2)
|
do j=1,ex(2)
|
||||||
do i=1,ex(1)
|
do i=1,ex(1)
|
||||||
@@ -934,7 +924,6 @@
|
|||||||
|
|
||||||
fyz = ZEO
|
fyz = ZEO
|
||||||
|
|
||||||
!$omp parallel do collapse(2) private(i,j,k)
|
|
||||||
do k=1,ex(3)
|
do k=1,ex(3)
|
||||||
do j=1,ex(2)
|
do j=1,ex(2)
|
||||||
do i=1,ex(1)
|
do i=1,ex(1)
|
||||||
@@ -1030,7 +1019,6 @@
|
|||||||
fy = ZEO
|
fy = ZEO
|
||||||
fz = ZEO
|
fz = ZEO
|
||||||
|
|
||||||
!$omp parallel do collapse(2) private(i,j,k)
|
|
||||||
do k=1,ex(3)
|
do k=1,ex(3)
|
||||||
do j=1,ex(2)
|
do j=1,ex(2)
|
||||||
do i=1,ex(1)
|
do i=1,ex(1)
|
||||||
@@ -1146,7 +1134,6 @@
|
|||||||
|
|
||||||
fx = ZEO
|
fx = ZEO
|
||||||
|
|
||||||
!$omp parallel do collapse(2) private(i,j,k)
|
|
||||||
do k=1,ex(3)
|
do k=1,ex(3)
|
||||||
do j=1,ex(2)
|
do j=1,ex(2)
|
||||||
do i=1,ex(1)
|
do i=1,ex(1)
|
||||||
@@ -1240,7 +1227,6 @@
|
|||||||
|
|
||||||
fy = ZEO
|
fy = ZEO
|
||||||
|
|
||||||
!$omp parallel do collapse(2) private(i,j,k)
|
|
||||||
do k=1,ex(3)
|
do k=1,ex(3)
|
||||||
do j=1,ex(2)
|
do j=1,ex(2)
|
||||||
do i=1,ex(1)
|
do i=1,ex(1)
|
||||||
@@ -1328,7 +1314,6 @@
|
|||||||
|
|
||||||
fz = ZEO
|
fz = ZEO
|
||||||
|
|
||||||
!$omp parallel do collapse(2) private(i,j,k)
|
|
||||||
do k=1,ex(3)
|
do k=1,ex(3)
|
||||||
do j=1,ex(2)
|
do j=1,ex(2)
|
||||||
do i=1,ex(1)
|
do i=1,ex(1)
|
||||||
@@ -1445,7 +1430,6 @@
|
|||||||
fxz = ZEO
|
fxz = ZEO
|
||||||
fyz = ZEO
|
fyz = ZEO
|
||||||
|
|
||||||
!$omp parallel do collapse(2) private(i,j,k)
|
|
||||||
do k=1,ex(3)
|
do k=1,ex(3)
|
||||||
do j=1,ex(2)
|
do j=1,ex(2)
|
||||||
do i=1,ex(1)
|
do i=1,ex(1)
|
||||||
@@ -1596,7 +1580,6 @@
|
|||||||
|
|
||||||
fxx = ZEO
|
fxx = ZEO
|
||||||
|
|
||||||
!$omp parallel do collapse(2) private(i,j,k)
|
|
||||||
do k=1,ex(3)
|
do k=1,ex(3)
|
||||||
do j=1,ex(2)
|
do j=1,ex(2)
|
||||||
do i=1,ex(1)
|
do i=1,ex(1)
|
||||||
@@ -1676,7 +1659,6 @@
|
|||||||
|
|
||||||
fyy = ZEO
|
fyy = ZEO
|
||||||
|
|
||||||
!$omp parallel do collapse(2) private(i,j,k)
|
|
||||||
do k=1,ex(3)
|
do k=1,ex(3)
|
||||||
do j=1,ex(2)
|
do j=1,ex(2)
|
||||||
do i=1,ex(1)
|
do i=1,ex(1)
|
||||||
@@ -1758,7 +1740,6 @@
|
|||||||
|
|
||||||
fzz = ZEO
|
fzz = ZEO
|
||||||
|
|
||||||
!$omp parallel do collapse(2) private(i,j,k)
|
|
||||||
do k=1,ex(3)
|
do k=1,ex(3)
|
||||||
do j=1,ex(2)
|
do j=1,ex(2)
|
||||||
do i=1,ex(1)
|
do i=1,ex(1)
|
||||||
@@ -1840,7 +1821,6 @@
|
|||||||
|
|
||||||
fxy = ZEO
|
fxy = ZEO
|
||||||
|
|
||||||
!$omp parallel do collapse(2) private(i,j,k)
|
|
||||||
do k=1,ex(3)
|
do k=1,ex(3)
|
||||||
do j=1,ex(2)
|
do j=1,ex(2)
|
||||||
do i=1,ex(1)
|
do i=1,ex(1)
|
||||||
@@ -1923,7 +1903,6 @@
|
|||||||
|
|
||||||
fxz = ZEO
|
fxz = ZEO
|
||||||
|
|
||||||
!$omp parallel do collapse(2) private(i,j,k)
|
|
||||||
do k=1,ex(3)
|
do k=1,ex(3)
|
||||||
do j=1,ex(2)
|
do j=1,ex(2)
|
||||||
do i=1,ex(1)
|
do i=1,ex(1)
|
||||||
@@ -2004,7 +1983,6 @@
|
|||||||
|
|
||||||
fyz = ZEO
|
fyz = ZEO
|
||||||
|
|
||||||
!$omp parallel do collapse(2) private(i,j,k)
|
|
||||||
do k=1,ex(3)
|
do k=1,ex(3)
|
||||||
do j=1,ex(2)
|
do j=1,ex(2)
|
||||||
do i=1,ex(1)
|
do i=1,ex(1)
|
||||||
@@ -2109,7 +2087,6 @@
|
|||||||
fy = ZEO
|
fy = ZEO
|
||||||
fz = ZEO
|
fz = ZEO
|
||||||
|
|
||||||
!$omp parallel do collapse(2) private(i,j,k)
|
|
||||||
do k=1,ex(3)
|
do k=1,ex(3)
|
||||||
do j=1,ex(2)
|
do j=1,ex(2)
|
||||||
do i=1,ex(1)
|
do i=1,ex(1)
|
||||||
@@ -2242,7 +2219,6 @@
|
|||||||
|
|
||||||
fx = ZEO
|
fx = ZEO
|
||||||
|
|
||||||
!$omp parallel do collapse(2) private(i,j,k)
|
|
||||||
do k=1,ex(3)
|
do k=1,ex(3)
|
||||||
do j=1,ex(2)
|
do j=1,ex(2)
|
||||||
do i=1,ex(1)
|
do i=1,ex(1)
|
||||||
@@ -2345,7 +2321,6 @@
|
|||||||
|
|
||||||
fy = ZEO
|
fy = ZEO
|
||||||
|
|
||||||
!$omp parallel do collapse(2) private(i,j,k)
|
|
||||||
do k=1,ex(3)
|
do k=1,ex(3)
|
||||||
do j=1,ex(2)
|
do j=1,ex(2)
|
||||||
do i=1,ex(1)
|
do i=1,ex(1)
|
||||||
@@ -2439,7 +2414,6 @@
|
|||||||
|
|
||||||
fz = ZEO
|
fz = ZEO
|
||||||
|
|
||||||
!$omp parallel do collapse(2) private(i,j,k)
|
|
||||||
do k=1,ex(3)
|
do k=1,ex(3)
|
||||||
do j=1,ex(2)
|
do j=1,ex(2)
|
||||||
do i=1,ex(1)
|
do i=1,ex(1)
|
||||||
@@ -2570,7 +2544,6 @@
|
|||||||
fxz = ZEO
|
fxz = ZEO
|
||||||
fyz = ZEO
|
fyz = ZEO
|
||||||
|
|
||||||
!$omp parallel do collapse(2) private(i,j,k)
|
|
||||||
do k=1,ex(3)
|
do k=1,ex(3)
|
||||||
do j=1,ex(2)
|
do j=1,ex(2)
|
||||||
do i=1,ex(1)
|
do i=1,ex(1)
|
||||||
@@ -2770,7 +2743,6 @@
|
|||||||
|
|
||||||
fxx = ZEO
|
fxx = ZEO
|
||||||
|
|
||||||
!$omp parallel do collapse(2) private(i,j,k)
|
|
||||||
do k=1,ex(3)
|
do k=1,ex(3)
|
||||||
do j=1,ex(2)
|
do j=1,ex(2)
|
||||||
do i=1,ex(1)
|
do i=1,ex(1)
|
||||||
@@ -2855,7 +2827,6 @@
|
|||||||
|
|
||||||
fyy = ZEO
|
fyy = ZEO
|
||||||
|
|
||||||
!$omp parallel do collapse(2) private(i,j,k)
|
|
||||||
do k=1,ex(3)
|
do k=1,ex(3)
|
||||||
do j=1,ex(2)
|
do j=1,ex(2)
|
||||||
do i=1,ex(1)
|
do i=1,ex(1)
|
||||||
@@ -2943,7 +2914,6 @@
|
|||||||
|
|
||||||
fzz = ZEO
|
fzz = ZEO
|
||||||
|
|
||||||
!$omp parallel do collapse(2) private(i,j,k)
|
|
||||||
do k=1,ex(3)
|
do k=1,ex(3)
|
||||||
do j=1,ex(2)
|
do j=1,ex(2)
|
||||||
do i=1,ex(1)
|
do i=1,ex(1)
|
||||||
@@ -3031,7 +3001,6 @@
|
|||||||
|
|
||||||
fxy = ZEO
|
fxy = ZEO
|
||||||
|
|
||||||
!$omp parallel do collapse(2) private(i,j,k)
|
|
||||||
do k=1,ex(3)
|
do k=1,ex(3)
|
||||||
do j=1,ex(2)
|
do j=1,ex(2)
|
||||||
do i=1,ex(1)
|
do i=1,ex(1)
|
||||||
@@ -3124,7 +3093,6 @@
|
|||||||
|
|
||||||
fxz = ZEO
|
fxz = ZEO
|
||||||
|
|
||||||
!$omp parallel do collapse(2) private(i,j,k)
|
|
||||||
do k=1,ex(3)
|
do k=1,ex(3)
|
||||||
do j=1,ex(2)
|
do j=1,ex(2)
|
||||||
do i=1,ex(1)
|
do i=1,ex(1)
|
||||||
@@ -3215,7 +3183,6 @@
|
|||||||
|
|
||||||
fyz = ZEO
|
fyz = ZEO
|
||||||
|
|
||||||
!$omp parallel do collapse(2) private(i,j,k)
|
|
||||||
do k=1,ex(3)
|
do k=1,ex(3)
|
||||||
do j=1,ex(2)
|
do j=1,ex(2)
|
||||||
do i=1,ex(1)
|
do i=1,ex(1)
|
||||||
@@ -3335,7 +3302,6 @@
|
|||||||
fy = ZEO
|
fy = ZEO
|
||||||
fz = ZEO
|
fz = ZEO
|
||||||
|
|
||||||
!$omp parallel do collapse(2) private(i,j,k)
|
|
||||||
do k=1,ex(3)
|
do k=1,ex(3)
|
||||||
do j=1,ex(2)
|
do j=1,ex(2)
|
||||||
do i=1,ex(1)
|
do i=1,ex(1)
|
||||||
@@ -3488,7 +3454,6 @@
|
|||||||
|
|
||||||
fx = ZEO
|
fx = ZEO
|
||||||
|
|
||||||
!$omp parallel do collapse(2) private(i,j,k)
|
|
||||||
do k=1,ex(3)
|
do k=1,ex(3)
|
||||||
do j=1,ex(2)
|
do j=1,ex(2)
|
||||||
do i=1,ex(1)
|
do i=1,ex(1)
|
||||||
@@ -3601,7 +3566,6 @@
|
|||||||
|
|
||||||
fy = ZEO
|
fy = ZEO
|
||||||
|
|
||||||
!$omp parallel do collapse(2) private(i,j,k)
|
|
||||||
do k=1,ex(3)
|
do k=1,ex(3)
|
||||||
do j=1,ex(2)
|
do j=1,ex(2)
|
||||||
do i=1,ex(1)
|
do i=1,ex(1)
|
||||||
@@ -3703,7 +3667,6 @@
|
|||||||
|
|
||||||
fz = ZEO
|
fz = ZEO
|
||||||
|
|
||||||
!$omp parallel do collapse(2) private(i,j,k)
|
|
||||||
do k=1,ex(3)
|
do k=1,ex(3)
|
||||||
do j=1,ex(2)
|
do j=1,ex(2)
|
||||||
do i=1,ex(1)
|
do i=1,ex(1)
|
||||||
@@ -3851,7 +3814,6 @@
|
|||||||
fxz = ZEO
|
fxz = ZEO
|
||||||
fyz = ZEO
|
fyz = ZEO
|
||||||
|
|
||||||
!$omp parallel do collapse(2) private(i,j,k)
|
|
||||||
do k=1,ex(3)
|
do k=1,ex(3)
|
||||||
do j=1,ex(2)
|
do j=1,ex(2)
|
||||||
do i=1,ex(1)
|
do i=1,ex(1)
|
||||||
@@ -4136,7 +4098,6 @@
|
|||||||
|
|
||||||
fxx = ZEO
|
fxx = ZEO
|
||||||
|
|
||||||
!$omp parallel do collapse(2) private(i,j,k)
|
|
||||||
do k=1,ex(3)
|
do k=1,ex(3)
|
||||||
do j=1,ex(2)
|
do j=1,ex(2)
|
||||||
do i=1,ex(1)
|
do i=1,ex(1)
|
||||||
@@ -4230,7 +4191,6 @@
|
|||||||
|
|
||||||
fyy = ZEO
|
fyy = ZEO
|
||||||
|
|
||||||
!$omp parallel do collapse(2) private(i,j,k)
|
|
||||||
do k=1,ex(3)
|
do k=1,ex(3)
|
||||||
do j=1,ex(2)
|
do j=1,ex(2)
|
||||||
do i=1,ex(1)
|
do i=1,ex(1)
|
||||||
@@ -4327,7 +4287,6 @@
|
|||||||
|
|
||||||
fzz = ZEO
|
fzz = ZEO
|
||||||
|
|
||||||
!$omp parallel do collapse(2) private(i,j,k)
|
|
||||||
do k=1,ex(3)
|
do k=1,ex(3)
|
||||||
do j=1,ex(2)
|
do j=1,ex(2)
|
||||||
do i=1,ex(1)
|
do i=1,ex(1)
|
||||||
@@ -4424,7 +4383,6 @@
|
|||||||
|
|
||||||
fxy = ZEO
|
fxy = ZEO
|
||||||
|
|
||||||
!$omp parallel do collapse(2) private(i,j,k)
|
|
||||||
do k=1,ex(3)
|
do k=1,ex(3)
|
||||||
do j=1,ex(2)
|
do j=1,ex(2)
|
||||||
do i=1,ex(1)
|
do i=1,ex(1)
|
||||||
@@ -4539,7 +4497,6 @@
|
|||||||
|
|
||||||
fxz = ZEO
|
fxz = ZEO
|
||||||
|
|
||||||
!$omp parallel do collapse(2) private(i,j,k)
|
|
||||||
do k=1,ex(3)
|
do k=1,ex(3)
|
||||||
do j=1,ex(2)
|
do j=1,ex(2)
|
||||||
do i=1,ex(1)
|
do i=1,ex(1)
|
||||||
@@ -4652,7 +4609,6 @@
|
|||||||
|
|
||||||
fyz = ZEO
|
fyz = ZEO
|
||||||
|
|
||||||
!$omp parallel do collapse(2) private(i,j,k)
|
|
||||||
do k=1,ex(3)
|
do k=1,ex(3)
|
||||||
do j=1,ex(2)
|
do j=1,ex(2)
|
||||||
do i=1,ex(1)
|
do i=1,ex(1)
|
||||||
@@ -4723,7 +4679,6 @@ subroutine fderivs_shc(ex,f,fx,fy,fz,crho,sigma,R,SYM1,SYM2,SYM3,Symmetry,Lev,ss
|
|||||||
#if 0
|
#if 0
|
||||||
integer :: i,j,k
|
integer :: i,j,k
|
||||||
|
|
||||||
!$omp parallel do collapse(2) private(i,j,k)
|
|
||||||
do k=1,ex(3)
|
do k=1,ex(3)
|
||||||
do j=1,ex(2)
|
do j=1,ex(2)
|
||||||
do i=1,ex(1)
|
do i=1,ex(1)
|
||||||
@@ -4774,7 +4729,6 @@ subroutine fdderivs_shc(ex,f,fxx,fxy,fxz,fyy,fyz,fzz,crho,sigma,R,SYM1,SYM2,SYM3
|
|||||||
#if 0
|
#if 0
|
||||||
integer :: i,j,k
|
integer :: i,j,k
|
||||||
|
|
||||||
!$omp parallel do collapse(2) private(i,j,k)
|
|
||||||
do k=1,ex(3)
|
do k=1,ex(3)
|
||||||
do j=1,ex(2)
|
do j=1,ex(2)
|
||||||
do i=1,ex(1)
|
do i=1,ex(1)
|
||||||
|
|||||||
@@ -19,61 +19,48 @@
|
|||||||
|
|
||||||
!~~~~~~~> Local variable:
|
!~~~~~~~> Local variable:
|
||||||
|
|
||||||
integer :: i,j,k
|
real*8, dimension(ex(1),ex(2),ex(3)) :: trA,detg
|
||||||
real*8 :: lgxx,lgyy,lgzz,ldetg
|
real*8, dimension(ex(1),ex(2),ex(3)) :: gxx,gyy,gzz
|
||||||
real*8 :: lgupxx,lgupxy,lgupxz,lgupyy,lgupyz,lgupzz
|
real*8, dimension(ex(1),ex(2),ex(3)) :: gupxx,gupxy,gupxz,gupyy,gupyz,gupzz
|
||||||
real*8 :: ltrA,lscale
|
|
||||||
real*8, parameter :: F1o3 = 1.D0 / 3.D0, ONE = 1.D0, TWO = 2.D0
|
real*8, parameter :: F1o3 = 1.D0 / 3.D0, ONE = 1.D0, TWO = 2.D0
|
||||||
|
|
||||||
!~~~~~~>
|
!~~~~~~>
|
||||||
|
|
||||||
!$omp parallel do collapse(2) private(i,j,k,lgxx,lgyy,lgzz,ldetg,lgupxx,lgupxy,lgupxz,lgupyy,lgupyz,lgupzz,ltrA,lscale)
|
gxx = dxx + ONE
|
||||||
do k=1,ex(3)
|
gyy = dyy + ONE
|
||||||
do j=1,ex(2)
|
gzz = dzz + ONE
|
||||||
do i=1,ex(1)
|
|
||||||
|
|
||||||
lgxx = dxx(i,j,k) + ONE
|
detg = gxx * gyy * gzz + gxy * gyz * gxz + gxz * gxy * gyz - &
|
||||||
lgyy = dyy(i,j,k) + ONE
|
gxz * gyy * gxz - gxy * gxy * gzz - gxx * gyz * gyz
|
||||||
lgzz = dzz(i,j,k) + ONE
|
gupxx = ( gyy * gzz - gyz * gyz ) / detg
|
||||||
|
gupxy = - ( gxy * gzz - gyz * gxz ) / detg
|
||||||
|
gupxz = ( gxy * gyz - gyy * gxz ) / detg
|
||||||
|
gupyy = ( gxx * gzz - gxz * gxz ) / detg
|
||||||
|
gupyz = - ( gxx * gyz - gxy * gxz ) / detg
|
||||||
|
gupzz = ( gxx * gyy - gxy * gxy ) / detg
|
||||||
|
|
||||||
ldetg = lgxx * lgyy * lgzz &
|
trA = gupxx * Axx + gupyy * Ayy + gupzz * Azz &
|
||||||
+ gxy(i,j,k) * gyz(i,j,k) * gxz(i,j,k) &
|
+ TWO * (gupxy * Axy + gupxz * Axz + gupyz * Ayz)
|
||||||
+ gxz(i,j,k) * gxy(i,j,k) * gyz(i,j,k) &
|
|
||||||
- gxz(i,j,k) * lgyy * gxz(i,j,k) &
|
|
||||||
- gxy(i,j,k) * gxy(i,j,k) * lgzz &
|
|
||||||
- lgxx * gyz(i,j,k) * gyz(i,j,k)
|
|
||||||
|
|
||||||
lgupxx = ( lgyy * lgzz - gyz(i,j,k) * gyz(i,j,k) ) / ldetg
|
Axx = Axx - F1o3 * gxx * trA
|
||||||
lgupxy = - ( gxy(i,j,k) * lgzz - gyz(i,j,k) * gxz(i,j,k) ) / ldetg
|
Axy = Axy - F1o3 * gxy * trA
|
||||||
lgupxz = ( gxy(i,j,k) * gyz(i,j,k) - lgyy * gxz(i,j,k) ) / ldetg
|
Axz = Axz - F1o3 * gxz * trA
|
||||||
lgupyy = ( lgxx * lgzz - gxz(i,j,k) * gxz(i,j,k) ) / ldetg
|
Ayy = Ayy - F1o3 * gyy * trA
|
||||||
lgupyz = - ( lgxx * gyz(i,j,k) - gxy(i,j,k) * gxz(i,j,k) ) / ldetg
|
Ayz = Ayz - F1o3 * gyz * trA
|
||||||
lgupzz = ( lgxx * lgyy - gxy(i,j,k) * gxy(i,j,k) ) / ldetg
|
Azz = Azz - F1o3 * gzz * trA
|
||||||
|
|
||||||
ltrA = lgupxx * Axx(i,j,k) + lgupyy * Ayy(i,j,k) &
|
detg = ONE / ( detg ** F1o3 )
|
||||||
+ lgupzz * Azz(i,j,k) &
|
|
||||||
+ TWO * (lgupxy * Axy(i,j,k) + lgupxz * Axz(i,j,k) &
|
|
||||||
+ lgupyz * Ayz(i,j,k))
|
|
||||||
|
|
||||||
Axx(i,j,k) = Axx(i,j,k) - F1o3 * lgxx * ltrA
|
gxx = gxx * detg
|
||||||
Axy(i,j,k) = Axy(i,j,k) - F1o3 * gxy(i,j,k) * ltrA
|
gxy = gxy * detg
|
||||||
Axz(i,j,k) = Axz(i,j,k) - F1o3 * gxz(i,j,k) * ltrA
|
gxz = gxz * detg
|
||||||
Ayy(i,j,k) = Ayy(i,j,k) - F1o3 * lgyy * ltrA
|
gyy = gyy * detg
|
||||||
Ayz(i,j,k) = Ayz(i,j,k) - F1o3 * gyz(i,j,k) * ltrA
|
gyz = gyz * detg
|
||||||
Azz(i,j,k) = Azz(i,j,k) - F1o3 * lgzz * ltrA
|
gzz = gzz * detg
|
||||||
|
|
||||||
lscale = ONE / ( ldetg ** F1o3 )
|
dxx = gxx - ONE
|
||||||
|
dyy = gyy - ONE
|
||||||
dxx(i,j,k) = lgxx * lscale - ONE
|
dzz = gzz - ONE
|
||||||
gxy(i,j,k) = gxy(i,j,k) * lscale
|
|
||||||
gxz(i,j,k) = gxz(i,j,k) * lscale
|
|
||||||
dyy(i,j,k) = lgyy * lscale - ONE
|
|
||||||
gyz(i,j,k) = gyz(i,j,k) * lscale
|
|
||||||
dzz(i,j,k) = lgzz * lscale - ONE
|
|
||||||
|
|
||||||
enddo
|
|
||||||
enddo
|
|
||||||
enddo
|
|
||||||
|
|
||||||
return
|
return
|
||||||
|
|
||||||
@@ -96,71 +83,50 @@
|
|||||||
|
|
||||||
!~~~~~~~> Local variable:
|
!~~~~~~~> Local variable:
|
||||||
|
|
||||||
integer :: i,j,k
|
real*8, dimension(ex(1),ex(2),ex(3)) :: trA
|
||||||
real*8 :: lgxx,lgyy,lgzz,lscale
|
real*8, dimension(ex(1),ex(2),ex(3)) :: gxx,gyy,gzz
|
||||||
real*8 :: lgxy,lgxz,lgyz
|
real*8, dimension(ex(1),ex(2),ex(3)) :: gupxx,gupxy,gupxz,gupyy,gupyz,gupzz
|
||||||
real*8 :: lgupxx,lgupxy,lgupxz,lgupyy,lgupyz,lgupzz
|
|
||||||
real*8 :: ltrA
|
|
||||||
real*8, parameter :: F1o3 = 1.D0 / 3.D0, ONE = 1.D0, TWO = 2.D0
|
real*8, parameter :: F1o3 = 1.D0 / 3.D0, ONE = 1.D0, TWO = 2.D0
|
||||||
|
|
||||||
!~~~~~~>
|
!~~~~~~>
|
||||||
|
|
||||||
!$omp parallel do collapse(2) private(i,j,k,lgxx,lgyy,lgzz,lscale,lgxy,lgxz,lgyz,lgupxx,lgupxy,lgupxz,lgupyy,lgupyz,lgupzz,ltrA)
|
gxx = dxx + ONE
|
||||||
do k=1,ex(3)
|
gyy = dyy + ONE
|
||||||
do j=1,ex(2)
|
gzz = dzz + ONE
|
||||||
do i=1,ex(1)
|
! for g
|
||||||
|
gupzz = gxx * gyy * gzz + gxy * gyz * gxz + gxz * gxy * gyz - &
|
||||||
|
gxz * gyy * gxz - gxy * gxy * gzz - gxx * gyz * gyz
|
||||||
|
|
||||||
! for g: normalize determinant first
|
gupzz = ONE / ( gupzz ** F1o3 )
|
||||||
lgxx = dxx(i,j,k) + ONE
|
|
||||||
lgyy = dyy(i,j,k) + ONE
|
|
||||||
lgzz = dzz(i,j,k) + ONE
|
|
||||||
lgxy = gxy(i,j,k)
|
|
||||||
lgxz = gxz(i,j,k)
|
|
||||||
lgyz = gyz(i,j,k)
|
|
||||||
|
|
||||||
lscale = lgxx * lgyy * lgzz + lgxy * lgyz * lgxz &
|
gxx = gxx * gupzz
|
||||||
+ lgxz * lgxy * lgyz - lgxz * lgyy * lgxz &
|
gxy = gxy * gupzz
|
||||||
- lgxy * lgxy * lgzz - lgxx * lgyz * lgyz
|
gxz = gxz * gupzz
|
||||||
|
gyy = gyy * gupzz
|
||||||
|
gyz = gyz * gupzz
|
||||||
|
gzz = gzz * gupzz
|
||||||
|
|
||||||
lscale = ONE / ( lscale ** F1o3 )
|
dxx = gxx - ONE
|
||||||
|
dyy = gyy - ONE
|
||||||
|
dzz = gzz - ONE
|
||||||
|
! for A
|
||||||
|
|
||||||
lgxx = lgxx * lscale
|
gupxx = ( gyy * gzz - gyz * gyz )
|
||||||
lgxy = lgxy * lscale
|
gupxy = - ( gxy * gzz - gyz * gxz )
|
||||||
lgxz = lgxz * lscale
|
gupxz = ( gxy * gyz - gyy * gxz )
|
||||||
lgyy = lgyy * lscale
|
gupyy = ( gxx * gzz - gxz * gxz )
|
||||||
lgyz = lgyz * lscale
|
gupyz = - ( gxx * gyz - gxy * gxz )
|
||||||
lgzz = lgzz * lscale
|
gupzz = ( gxx * gyy - gxy * gxy )
|
||||||
|
|
||||||
dxx(i,j,k) = lgxx - ONE
|
trA = gupxx * Axx + gupyy * Ayy + gupzz * Azz &
|
||||||
gxy(i,j,k) = lgxy
|
+ TWO * (gupxy * Axy + gupxz * Axz + gupyz * Ayz)
|
||||||
gxz(i,j,k) = lgxz
|
|
||||||
dyy(i,j,k) = lgyy - ONE
|
|
||||||
gyz(i,j,k) = lgyz
|
|
||||||
dzz(i,j,k) = lgzz - ONE
|
|
||||||
|
|
||||||
! for A: trace-free using normalized metric (det=1, no division needed)
|
Axx = Axx - F1o3 * gxx * trA
|
||||||
lgupxx = ( lgyy * lgzz - lgyz * lgyz )
|
Axy = Axy - F1o3 * gxy * trA
|
||||||
lgupxy = - ( lgxy * lgzz - lgyz * lgxz )
|
Axz = Axz - F1o3 * gxz * trA
|
||||||
lgupxz = ( lgxy * lgyz - lgyy * lgxz )
|
Ayy = Ayy - F1o3 * gyy * trA
|
||||||
lgupyy = ( lgxx * lgzz - lgxz * lgxz )
|
Ayz = Ayz - F1o3 * gyz * trA
|
||||||
lgupyz = - ( lgxx * lgyz - lgxy * lgxz )
|
Azz = Azz - F1o3 * gzz * trA
|
||||||
lgupzz = ( lgxx * lgyy - lgxy * lgxy )
|
|
||||||
|
|
||||||
ltrA = lgupxx * Axx(i,j,k) + lgupyy * Ayy(i,j,k) &
|
|
||||||
+ lgupzz * Azz(i,j,k) &
|
|
||||||
+ TWO * (lgupxy * Axy(i,j,k) + lgupxz * Axz(i,j,k) &
|
|
||||||
+ lgupyz * Ayz(i,j,k))
|
|
||||||
|
|
||||||
Axx(i,j,k) = Axx(i,j,k) - F1o3 * lgxx * ltrA
|
|
||||||
Axy(i,j,k) = Axy(i,j,k) - F1o3 * lgxy * ltrA
|
|
||||||
Axz(i,j,k) = Axz(i,j,k) - F1o3 * lgxz * ltrA
|
|
||||||
Ayy(i,j,k) = Ayy(i,j,k) - F1o3 * lgyy * ltrA
|
|
||||||
Ayz(i,j,k) = Ayz(i,j,k) - F1o3 * lgyz * ltrA
|
|
||||||
Azz(i,j,k) = Azz(i,j,k) - F1o3 * lgzz * ltrA
|
|
||||||
|
|
||||||
enddo
|
|
||||||
enddo
|
|
||||||
enddo
|
|
||||||
|
|
||||||
return
|
return
|
||||||
|
|
||||||
|
|||||||
@@ -6,11 +6,7 @@
|
|||||||
#include <stdio.h>
|
#include <stdio.h>
|
||||||
#include <assert.h>
|
#include <assert.h>
|
||||||
#include <math.h>
|
#include <math.h>
|
||||||
#ifdef MPI_STUB
|
|
||||||
#include "mpi_stub.h"
|
|
||||||
#else
|
|
||||||
#include <mpi.h>
|
#include <mpi.h>
|
||||||
#endif
|
|
||||||
|
|
||||||
#include "util_Table.h"
|
#include "util_Table.h"
|
||||||
#include "cctk.h"
|
#include "cctk.h"
|
||||||
|
|||||||
@@ -6,11 +6,7 @@
|
|||||||
#include <stdio.h>
|
#include <stdio.h>
|
||||||
#include <assert.h>
|
#include <assert.h>
|
||||||
#include <math.h>
|
#include <math.h>
|
||||||
#ifdef MPI_STUB
|
|
||||||
#include "mpi_stub.h"
|
|
||||||
#else
|
|
||||||
#include <mpi.h>
|
#include <mpi.h>
|
||||||
#endif
|
|
||||||
|
|
||||||
#include "cctk.h"
|
#include "cctk.h"
|
||||||
|
|
||||||
|
|||||||
@@ -324,6 +324,7 @@ subroutine symmetry_bd(ord,extc,func,funcc,SoA)
|
|||||||
|
|
||||||
integer::i
|
integer::i
|
||||||
|
|
||||||
|
funcc = 0.d0
|
||||||
funcc(1:extc(1),1:extc(2),1:extc(3)) = func
|
funcc(1:extc(1),1:extc(2),1:extc(3)) = func
|
||||||
do i=0,ord-1
|
do i=0,ord-1
|
||||||
funcc(-i,1:extc(2),1:extc(3)) = funcc(i+2,1:extc(2),1:extc(3))*SoA(1)
|
funcc(-i,1:extc(2),1:extc(3)) = funcc(i+2,1:extc(2),1:extc(3))*SoA(1)
|
||||||
@@ -349,6 +350,7 @@ subroutine symmetry_tbd(ord,extc,func,funcc,SoA)
|
|||||||
|
|
||||||
integer::i
|
integer::i
|
||||||
|
|
||||||
|
funcc = 0.d0
|
||||||
funcc(1:extc(1),1:extc(2),1:extc(3)) = func
|
funcc(1:extc(1),1:extc(2),1:extc(3)) = func
|
||||||
do i=0,ord-1
|
do i=0,ord-1
|
||||||
funcc(-i,1:extc(2),1:extc(3)) = funcc(i+2,1:extc(2),1:extc(3))*SoA(1)
|
funcc(-i,1:extc(2),1:extc(3)) = funcc(i+2,1:extc(2),1:extc(3))*SoA(1)
|
||||||
@@ -377,6 +379,7 @@ subroutine symmetry_stbd(ord,extc,func,funcc,SoA)
|
|||||||
|
|
||||||
integer::i
|
integer::i
|
||||||
|
|
||||||
|
funcc = 0.d0
|
||||||
funcc(1:extc(1),1:extc(2),1:extc(3)) = func
|
funcc(1:extc(1),1:extc(2),1:extc(3)) = func
|
||||||
do i=0,ord-1
|
do i=0,ord-1
|
||||||
funcc(-i,1:extc(2),1:extc(3)) = funcc(i+2,1:extc(2),1:extc(3))*SoA(1)
|
funcc(-i,1:extc(2),1:extc(3)) = funcc(i+2,1:extc(2),1:extc(3))*SoA(1)
|
||||||
@@ -883,6 +886,7 @@ subroutine symmetry_bd(ord,extc,func,funcc,SoA)
|
|||||||
|
|
||||||
integer::i
|
integer::i
|
||||||
|
|
||||||
|
funcc = 0.d0
|
||||||
funcc(1:extc(1),1:extc(2),1:extc(3)) = func
|
funcc(1:extc(1),1:extc(2),1:extc(3)) = func
|
||||||
do i=0,ord-1
|
do i=0,ord-1
|
||||||
funcc(-i,1:extc(2),1:extc(3)) = funcc(i+1,1:extc(2),1:extc(3))*SoA(1)
|
funcc(-i,1:extc(2),1:extc(3)) = funcc(i+1,1:extc(2),1:extc(3))*SoA(1)
|
||||||
@@ -908,6 +912,7 @@ subroutine symmetry_tbd(ord,extc,func,funcc,SoA)
|
|||||||
|
|
||||||
integer::i
|
integer::i
|
||||||
|
|
||||||
|
funcc = 0.d0
|
||||||
funcc(1:extc(1),1:extc(2),1:extc(3)) = func
|
funcc(1:extc(1),1:extc(2),1:extc(3)) = func
|
||||||
do i=0,ord-1
|
do i=0,ord-1
|
||||||
funcc(-i,1:extc(2),1:extc(3)) = funcc(i+1,1:extc(2),1:extc(3))*SoA(1)
|
funcc(-i,1:extc(2),1:extc(3)) = funcc(i+1,1:extc(2),1:extc(3))*SoA(1)
|
||||||
@@ -936,6 +941,7 @@ subroutine symmetry_stbd(ord,extc,func,funcc,SoA)
|
|||||||
|
|
||||||
integer::i
|
integer::i
|
||||||
|
|
||||||
|
funcc = 0.d0
|
||||||
funcc(1:extc(1),1:extc(2),1:extc(3)) = func
|
funcc(1:extc(1),1:extc(2),1:extc(3)) = func
|
||||||
do i=0,ord-1
|
do i=0,ord-1
|
||||||
funcc(-i,1:extc(2),1:extc(3)) = funcc(i+1,1:extc(2),1:extc(3))*SoA(1)
|
funcc(-i,1:extc(2),1:extc(3)) = funcc(i+1,1:extc(2),1:extc(3))*SoA(1)
|
||||||
@@ -1111,7 +1117,6 @@ end subroutine d2dump
|
|||||||
!------------------------------------------------------------------------------
|
!------------------------------------------------------------------------------
|
||||||
! Lagrangian polynomial interpolation
|
! Lagrangian polynomial interpolation
|
||||||
!------------------------------------------------------------------------------
|
!------------------------------------------------------------------------------
|
||||||
|
|
||||||
subroutine polint(xa, ya, x, y, dy, ordn)
|
subroutine polint(xa, ya, x, y, dy, ordn)
|
||||||
implicit none
|
implicit none
|
||||||
|
|
||||||
@@ -1124,6 +1129,7 @@ end subroutine d2dump
|
|||||||
real*8, dimension(ordn) :: c, d, ho
|
real*8, dimension(ordn) :: c, d, ho
|
||||||
real*8 :: dif, dift, hp, h, den_val
|
real*8 :: dif, dift, hp, h, den_val
|
||||||
|
|
||||||
|
! Initialization
|
||||||
c = ya
|
c = ya
|
||||||
d = ya
|
d = ya
|
||||||
ho = xa - x
|
ho = xa - x
|
||||||
@@ -1131,6 +1137,7 @@ end subroutine d2dump
|
|||||||
ns = 1
|
ns = 1
|
||||||
dif = abs(x - xa(1))
|
dif = abs(x - xa(1))
|
||||||
|
|
||||||
|
! Find the index of the closest table entry
|
||||||
do i = 2, ordn
|
do i = 2, ordn
|
||||||
dift = abs(x - xa(i))
|
dift = abs(x - xa(i))
|
||||||
if (dift < dif) then
|
if (dift < dif) then
|
||||||
@@ -1142,6 +1149,7 @@ end subroutine d2dump
|
|||||||
y = ya(ns)
|
y = ya(ns)
|
||||||
ns = ns - 1
|
ns = ns - 1
|
||||||
|
|
||||||
|
! Main Neville's algorithm loop
|
||||||
do m = 1, ordn - 1
|
do m = 1, ordn - 1
|
||||||
n_m = ordn - m
|
n_m = ordn - m
|
||||||
do i = 1, n_m
|
do i = 1, n_m
|
||||||
@@ -1149,18 +1157,22 @@ end subroutine d2dump
|
|||||||
h = ho(i+m)
|
h = ho(i+m)
|
||||||
den_val = hp - h
|
den_val = hp - h
|
||||||
|
|
||||||
|
! Check for division by zero locally
|
||||||
if (den_val == 0.0d0) then
|
if (den_val == 0.0d0) then
|
||||||
write(*,*) 'failure in polint for point',x
|
write(*,*) 'failure in polint for point',x
|
||||||
write(*,*) 'with input points: ',xa
|
write(*,*) 'with input points: ',xa
|
||||||
stop
|
stop
|
||||||
end if
|
end if
|
||||||
|
|
||||||
|
! Reuse den_val to avoid redundant divisions
|
||||||
den_val = (c(i+1) - d(i)) / den_val
|
den_val = (c(i+1) - d(i)) / den_val
|
||||||
|
|
||||||
|
! Update c and d in place
|
||||||
d(i) = h * den_val
|
d(i) = h * den_val
|
||||||
c(i) = hp * den_val
|
c(i) = hp * den_val
|
||||||
end do
|
end do
|
||||||
|
|
||||||
|
! Decide which path (up or down the tableau) to take
|
||||||
if (2 * ns < n_m) then
|
if (2 * ns < n_m) then
|
||||||
dy = c(ns + 1)
|
dy = c(ns + 1)
|
||||||
else
|
else
|
||||||
@@ -1179,34 +1191,24 @@ end subroutine d2dump
|
|||||||
!------------------------------------------------------------------------------
|
!------------------------------------------------------------------------------
|
||||||
subroutine polin2(x1a,x2a,ya,x1,x2,y,dy,ordn)
|
subroutine polin2(x1a,x2a,ya,x1,x2,y,dy,ordn)
|
||||||
implicit none
|
implicit none
|
||||||
|
|
||||||
integer,intent(in) :: ordn
|
integer,intent(in) :: ordn
|
||||||
real*8, dimension(1:ordn), intent(in) :: x1a,x2a
|
real*8, dimension(ordn), intent(in) :: x1a,x2a
|
||||||
real*8, dimension(1:ordn,1:ordn), intent(in) :: ya
|
real*8, dimension(ordn,ordn), intent(in) :: ya
|
||||||
real*8, intent(in) :: x1,x2
|
real*8, intent(in) :: x1,x2
|
||||||
real*8, intent(out) :: y,dy
|
real*8, intent(out) :: y,dy
|
||||||
|
|
||||||
#ifdef POLINT_LEGACY_ORDER
|
|
||||||
integer :: i,m
|
|
||||||
real*8, dimension(ordn) :: ymtmp
|
|
||||||
real*8, dimension(ordn) :: yntmp
|
|
||||||
|
|
||||||
m=size(x1a)
|
|
||||||
do i=1,m
|
|
||||||
yntmp=ya(i,:)
|
|
||||||
call polint(x2a,yntmp,x2,ymtmp(i),dy,ordn)
|
|
||||||
end do
|
|
||||||
call polint(x1a,ymtmp,x1,y,dy,ordn)
|
|
||||||
#else
|
|
||||||
integer :: j
|
integer :: j
|
||||||
real*8, dimension(ordn) :: ymtmp
|
real*8, dimension(ordn) :: ymtmp
|
||||||
real*8 :: dy_temp
|
real*8 :: dy_temp ! Local variable to prevent overwriting result
|
||||||
|
|
||||||
|
! Optimized sequence: Loop over columns (j)
|
||||||
|
! ya(:,j) is a contiguous memory block in Fortran
|
||||||
do j=1,ordn
|
do j=1,ordn
|
||||||
call polint(x1a, ya(:,j), x1, ymtmp(j), dy_temp, ordn)
|
call polint(x1a, ya(:,j), x1, ymtmp(j), dy_temp, ordn)
|
||||||
end do
|
end do
|
||||||
|
|
||||||
|
! Final interpolation on the results
|
||||||
call polint(x2a, ymtmp, x2, y, dy, ordn)
|
call polint(x2a, ymtmp, x2, y, dy, ordn)
|
||||||
#endif
|
|
||||||
|
|
||||||
return
|
return
|
||||||
end subroutine polin2
|
end subroutine polin2
|
||||||
@@ -1217,47 +1219,33 @@ end subroutine d2dump
|
|||||||
!------------------------------------------------------------------------------
|
!------------------------------------------------------------------------------
|
||||||
subroutine polin3(x1a,x2a,x3a,ya,x1,x2,x3,y,dy,ordn)
|
subroutine polin3(x1a,x2a,x3a,ya,x1,x2,x3,y,dy,ordn)
|
||||||
implicit none
|
implicit none
|
||||||
|
|
||||||
integer,intent(in) :: ordn
|
integer,intent(in) :: ordn
|
||||||
real*8, dimension(1:ordn), intent(in) :: x1a,x2a,x3a
|
real*8, dimension(ordn), intent(in) :: x1a,x2a,x3a
|
||||||
real*8, dimension(1:ordn,1:ordn,1:ordn), intent(in) :: ya
|
real*8, dimension(ordn,ordn,ordn), intent(in) :: ya
|
||||||
real*8, intent(in) :: x1,x2,x3
|
real*8, intent(in) :: x1,x2,x3
|
||||||
real*8, intent(out) :: y,dy
|
real*8, intent(out) :: y,dy
|
||||||
|
|
||||||
#ifdef POLINT_LEGACY_ORDER
|
|
||||||
integer :: i,j,m,n
|
|
||||||
real*8, dimension(ordn,ordn) :: yatmp
|
|
||||||
real*8, dimension(ordn) :: ymtmp
|
|
||||||
real*8, dimension(ordn) :: yntmp
|
|
||||||
real*8, dimension(ordn) :: yqtmp
|
|
||||||
|
|
||||||
m=size(x1a)
|
|
||||||
n=size(x2a)
|
|
||||||
do i=1,m
|
|
||||||
do j=1,n
|
|
||||||
yqtmp=ya(i,j,:)
|
|
||||||
call polint(x3a,yqtmp,x3,yatmp(i,j),dy,ordn)
|
|
||||||
end do
|
|
||||||
yntmp=yatmp(i,:)
|
|
||||||
call polint(x2a,yntmp,x2,ymtmp(i),dy,ordn)
|
|
||||||
end do
|
|
||||||
call polint(x1a,ymtmp,x1,y,dy,ordn)
|
|
||||||
#else
|
|
||||||
integer :: j, k
|
integer :: j, k
|
||||||
real*8, dimension(ordn,ordn) :: yatmp
|
real*8, dimension(ordn,ordn) :: yatmp
|
||||||
real*8, dimension(ordn) :: ymtmp
|
real*8, dimension(ordn) :: ymtmp
|
||||||
real*8 :: dy_temp
|
real*8 :: dy_temp
|
||||||
|
|
||||||
|
! Sequence change: Process the contiguous first dimension (x1) first.
|
||||||
|
! We loop through the 'slow' planes (j, k) to extract 'fast' columns.
|
||||||
do k=1,ordn
|
do k=1,ordn
|
||||||
do j=1,ordn
|
do j=1,ordn
|
||||||
|
! ya(:,j,k) is contiguous; much faster than ya(i,j,:)
|
||||||
call polint(x1a, ya(:,j,k), x1, yatmp(j,k), dy_temp, ordn)
|
call polint(x1a, ya(:,j,k), x1, yatmp(j,k), dy_temp, ordn)
|
||||||
end do
|
end do
|
||||||
end do
|
end do
|
||||||
|
|
||||||
|
! Now process the second dimension
|
||||||
do k=1,ordn
|
do k=1,ordn
|
||||||
call polint(x2a, yatmp(:,k), x2, ymtmp(k), dy_temp, ordn)
|
call polint(x2a, yatmp(:,k), x2, ymtmp(k), dy_temp, ordn)
|
||||||
end do
|
end do
|
||||||
|
|
||||||
|
! Final dimension
|
||||||
call polint(x3a, ymtmp, x3, y, dy, ordn)
|
call polint(x3a, ymtmp, x3, y, dy, ordn)
|
||||||
#endif
|
|
||||||
|
|
||||||
return
|
return
|
||||||
end subroutine polin3
|
end subroutine polin3
|
||||||
@@ -1279,9 +1267,7 @@ end subroutine d2dump
|
|||||||
real*8 :: dX, dY, dZ
|
real*8 :: dX, dY, dZ
|
||||||
integer::imin,jmin,kmin
|
integer::imin,jmin,kmin
|
||||||
integer::imax,jmax,kmax
|
integer::imax,jmax,kmax
|
||||||
integer::i,j,k,n_elements
|
integer::i,j,k
|
||||||
real*8, dimension(:), allocatable :: f_flat
|
|
||||||
real*8, external :: DDOT
|
|
||||||
|
|
||||||
dX = X(2) - X(1)
|
dX = X(2) - X(1)
|
||||||
dY = Y(2) - Y(1)
|
dY = Y(2) - Y(1)
|
||||||
@@ -1305,12 +1291,7 @@ if(dabs(X(1)-xmin) < dX) imin = 1
|
|||||||
if(dabs(Y(1)-ymin) < dY) jmin = 1
|
if(dabs(Y(1)-ymin) < dY) jmin = 1
|
||||||
if(dabs(Z(1)-zmin) < dZ) kmin = 1
|
if(dabs(Z(1)-zmin) < dZ) kmin = 1
|
||||||
|
|
||||||
! Optimized with oneMKL BLAS DDOT for dot product
|
f_out = sum(f(imin:imax,jmin:jmax,kmin:kmax)*f(imin:imax,jmin:jmax,kmin:kmax))
|
||||||
n_elements = (imax-imin+1)*(jmax-jmin+1)*(kmax-kmin+1)
|
|
||||||
allocate(f_flat(n_elements))
|
|
||||||
f_flat = reshape(f(imin:imax,jmin:jmax,kmin:kmax), [n_elements])
|
|
||||||
f_out = DDOT(n_elements, f_flat, 1, f_flat, 1)
|
|
||||||
deallocate(f_flat)
|
|
||||||
|
|
||||||
f_out = f_out*dX*dY*dZ
|
f_out = f_out*dX*dY*dZ
|
||||||
|
|
||||||
@@ -1335,9 +1316,7 @@ f_out = f_out*dX*dY*dZ
|
|||||||
real*8 :: dX, dY, dZ
|
real*8 :: dX, dY, dZ
|
||||||
integer::imin,jmin,kmin
|
integer::imin,jmin,kmin
|
||||||
integer::imax,jmax,kmax
|
integer::imax,jmax,kmax
|
||||||
integer::i,j,k,n_elements
|
integer::i,j,k
|
||||||
real*8, dimension(:), allocatable :: f_flat
|
|
||||||
real*8, external :: DDOT
|
|
||||||
|
|
||||||
real*8 :: PIo4
|
real*8 :: PIo4
|
||||||
|
|
||||||
@@ -1400,12 +1379,7 @@ if(Symmetry==2)then
|
|||||||
if(dabs(ymin+gw*dY)<dY.and.Y(1)<0.d0) jmin = gw+1
|
if(dabs(ymin+gw*dY)<dY.and.Y(1)<0.d0) jmin = gw+1
|
||||||
endif
|
endif
|
||||||
|
|
||||||
! Optimized with oneMKL BLAS DDOT for dot product
|
f_out = sum(f(imin:imax,jmin:jmax,kmin:kmax)*f(imin:imax,jmin:jmax,kmin:kmax))
|
||||||
n_elements = (imax-imin+1)*(jmax-jmin+1)*(kmax-kmin+1)
|
|
||||||
allocate(f_flat(n_elements))
|
|
||||||
f_flat = reshape(f(imin:imax,jmin:jmax,kmin:kmax), [n_elements])
|
|
||||||
f_out = DDOT(n_elements, f_flat, 1, f_flat, 1)
|
|
||||||
deallocate(f_flat)
|
|
||||||
|
|
||||||
f_out = f_out*dX*dY*dZ
|
f_out = f_out*dX*dY*dZ
|
||||||
|
|
||||||
@@ -1433,8 +1407,6 @@ f_out = f_out*dX*dY*dZ
|
|||||||
integer::imin,jmin,kmin
|
integer::imin,jmin,kmin
|
||||||
integer::imax,jmax,kmax
|
integer::imax,jmax,kmax
|
||||||
integer::i,j,k
|
integer::i,j,k
|
||||||
real*8, dimension(:), allocatable :: f_flat
|
|
||||||
real*8, external :: DDOT
|
|
||||||
|
|
||||||
real*8 :: PIo4
|
real*8 :: PIo4
|
||||||
|
|
||||||
@@ -1497,12 +1469,11 @@ if(Symmetry==2)then
|
|||||||
if(dabs(ymin+gw*dY)<dY.and.Y(1)<0.d0) jmin = gw+1
|
if(dabs(ymin+gw*dY)<dY.and.Y(1)<0.d0) jmin = gw+1
|
||||||
endif
|
endif
|
||||||
|
|
||||||
! Optimized with oneMKL BLAS DDOT for dot product
|
f_out = sum(f(imin:imax,jmin:jmax,kmin:kmax)*f(imin:imax,jmin:jmax,kmin:kmax))
|
||||||
|
|
||||||
|
f_out = f_out
|
||||||
|
|
||||||
Nout = (imax-imin+1)*(jmax-jmin+1)*(kmax-kmin+1)
|
Nout = (imax-imin+1)*(jmax-jmin+1)*(kmax-kmin+1)
|
||||||
allocate(f_flat(Nout))
|
|
||||||
f_flat = reshape(f(imin:imax,jmin:jmax,kmin:kmax), [Nout])
|
|
||||||
f_out = DDOT(Nout, f_flat, 1, f_flat, 1)
|
|
||||||
deallocate(f_flat)
|
|
||||||
|
|
||||||
return
|
return
|
||||||
|
|
||||||
@@ -1700,7 +1671,6 @@ deallocate(f_flat)
|
|||||||
real*8, dimension(ORDN,ORDN) :: tmp2
|
real*8, dimension(ORDN,ORDN) :: tmp2
|
||||||
real*8, dimension(ORDN) :: tmp1
|
real*8, dimension(ORDN) :: tmp1
|
||||||
real*8, dimension(3) :: SoAh
|
real*8, dimension(3) :: SoAh
|
||||||
real*8, external :: DDOT
|
|
||||||
|
|
||||||
! +1 because c++ gives 0 for first point
|
! +1 because c++ gives 0 for first point
|
||||||
cxB = inds+1
|
cxB = inds+1
|
||||||
@@ -1736,21 +1706,20 @@ deallocate(f_flat)
|
|||||||
ya=fh(cxB(1):cxT(1),cxB(2):cxT(2),cxB(3):cxT(3))
|
ya=fh(cxB(1):cxT(1),cxB(2):cxT(2),cxB(3):cxT(3))
|
||||||
endif
|
endif
|
||||||
|
|
||||||
! Optimized with BLAS operations for better performance
|
|
||||||
! First dimension: z-direction weighted sum
|
|
||||||
tmp2=0
|
tmp2=0
|
||||||
do m=1,ORDN
|
do m=1,ORDN
|
||||||
tmp2 = tmp2 + coef(2*ORDN+m)*ya(:,:,m)
|
tmp2 = tmp2 + coef(2*ORDN+m)*ya(:,:,m)
|
||||||
enddo
|
enddo
|
||||||
|
|
||||||
! Second dimension: y-direction weighted sum
|
|
||||||
tmp1=0
|
tmp1=0
|
||||||
do m=1,ORDN
|
do m=1,ORDN
|
||||||
tmp1 = tmp1 + coef(ORDN+m)*tmp2(:,m)
|
tmp1 = tmp1 + coef(ORDN+m)*tmp2(:,m)
|
||||||
enddo
|
enddo
|
||||||
|
|
||||||
! Third dimension: x-direction weighted sum using BLAS DDOT
|
f_int=0
|
||||||
f_int = DDOT(ORDN, coef(1:ORDN), 1, tmp1, 1)
|
do m=1,ORDN
|
||||||
|
f_int = f_int + coef(m)*tmp1(m)
|
||||||
|
enddo
|
||||||
|
|
||||||
return
|
return
|
||||||
|
|
||||||
@@ -1780,7 +1749,6 @@ deallocate(f_flat)
|
|||||||
real*8, dimension(ORDN,ORDN) :: ya
|
real*8, dimension(ORDN,ORDN) :: ya
|
||||||
real*8, dimension(ORDN) :: tmp1
|
real*8, dimension(ORDN) :: tmp1
|
||||||
real*8, dimension(2) :: SoAh
|
real*8, dimension(2) :: SoAh
|
||||||
real*8, external :: DDOT
|
|
||||||
|
|
||||||
! +1 because c++ gives 0 for first point
|
! +1 because c++ gives 0 for first point
|
||||||
cxB = inds(1:2)+1
|
cxB = inds(1:2)+1
|
||||||
@@ -1810,14 +1778,15 @@ deallocate(f_flat)
|
|||||||
ya=fh(cxB(1):cxT(1),cxB(2):cxT(2),inds(3))
|
ya=fh(cxB(1):cxT(1),cxB(2):cxT(2),inds(3))
|
||||||
endif
|
endif
|
||||||
|
|
||||||
! Optimized with BLAS operations
|
|
||||||
tmp1=0
|
tmp1=0
|
||||||
do m=1,ORDN
|
do m=1,ORDN
|
||||||
tmp1 = tmp1 + coef(ORDN+m)*ya(:,m)
|
tmp1 = tmp1 + coef(ORDN+m)*ya(:,m)
|
||||||
enddo
|
enddo
|
||||||
|
|
||||||
! Use BLAS DDOT for final weighted sum
|
f_int=0
|
||||||
f_int = DDOT(ORDN, coef(1:ORDN), 1, tmp1, 1)
|
do m=1,ORDN
|
||||||
|
f_int = f_int + coef(m)*tmp1(m)
|
||||||
|
enddo
|
||||||
|
|
||||||
return
|
return
|
||||||
|
|
||||||
@@ -1848,7 +1817,6 @@ deallocate(f_flat)
|
|||||||
real*8, dimension(ORDN) :: ya
|
real*8, dimension(ORDN) :: ya
|
||||||
real*8 :: SoAh
|
real*8 :: SoAh
|
||||||
integer,dimension(3) :: inds
|
integer,dimension(3) :: inds
|
||||||
real*8, external :: DDOT
|
|
||||||
|
|
||||||
! +1 because c++ gives 0 for first point
|
! +1 because c++ gives 0 for first point
|
||||||
inds = indsi + 1
|
inds = indsi + 1
|
||||||
@@ -1909,8 +1877,10 @@ deallocate(f_flat)
|
|||||||
write(*,*)"error in global_interpind1d, not recognized dumyd = ",dumyd
|
write(*,*)"error in global_interpind1d, not recognized dumyd = ",dumyd
|
||||||
endif
|
endif
|
||||||
|
|
||||||
! Optimized with BLAS DDOT for weighted sum
|
f_int=0
|
||||||
f_int = DDOT(ORDN, coef, 1, ya, 1)
|
do m=1,ORDN
|
||||||
|
f_int = f_int + coef(m)*ya(m)
|
||||||
|
enddo
|
||||||
|
|
||||||
return
|
return
|
||||||
|
|
||||||
@@ -2142,38 +2112,24 @@ deallocate(f_flat)
|
|||||||
|
|
||||||
end function fWigner_d_function
|
end function fWigner_d_function
|
||||||
!----------------------------------
|
!----------------------------------
|
||||||
! Optimized factorial function using lookup table for small N
|
|
||||||
! and log-gamma for large N to avoid overflow
|
|
||||||
function ffact(N) result(gont)
|
function ffact(N) result(gont)
|
||||||
implicit none
|
implicit none
|
||||||
integer,intent(in) :: N
|
integer,intent(in) :: N
|
||||||
|
|
||||||
real*8 :: gont
|
real*8 :: gont
|
||||||
integer :: i
|
|
||||||
|
|
||||||
! Lookup table for factorials 0! to 20! (precomputed)
|
integer :: i
|
||||||
real*8, parameter, dimension(0:20) :: fact_table = [ &
|
|
||||||
1.d0, 1.d0, 2.d0, 6.d0, 24.d0, 120.d0, 720.d0, 5040.d0, 40320.d0, &
|
|
||||||
362880.d0, 3628800.d0, 39916800.d0, 479001600.d0, 6227020800.d0, &
|
|
||||||
87178291200.d0, 1307674368000.d0, 20922789888000.d0, &
|
|
||||||
355687428096000.d0, 6402373705728000.d0, 121645100408832000.d0, &
|
|
||||||
2432902008176640000.d0 ]
|
|
||||||
|
|
||||||
! sanity check
|
! sanity check
|
||||||
if(N < 0)then
|
if(N < 0)then
|
||||||
write(*,*) "ffact: error input for factorial"
|
write(*,*) "ffact: error input for factorial"
|
||||||
gont = 1.d0
|
|
||||||
return
|
return
|
||||||
endif
|
endif
|
||||||
|
|
||||||
! Use lookup table for small N (fast path)
|
gont = 1.d0
|
||||||
if(N <= 20)then
|
do i=1,N
|
||||||
gont = fact_table(N)
|
gont = gont*i
|
||||||
else
|
enddo
|
||||||
! Use log-gamma function for large N: N! = exp(log_gamma(N+1))
|
|
||||||
! This avoids overflow and is computed efficiently
|
|
||||||
gont = exp(log_gamma(dble(N+1)))
|
|
||||||
endif
|
|
||||||
|
|
||||||
return
|
return
|
||||||
|
|
||||||
@@ -2307,3 +2263,4 @@ subroutine find_maximum(ext,X,Y,Z,fun,val,pos,llb,uub)
|
|||||||
return
|
return
|
||||||
|
|
||||||
end subroutine
|
end subroutine
|
||||||
|
|
||||||
|
|||||||
@@ -16,66 +16,115 @@ using namespace std;
|
|||||||
#include <string.h>
|
#include <string.h>
|
||||||
#include <math.h>
|
#include <math.h>
|
||||||
#endif
|
#endif
|
||||||
|
/* Linear equation solution by Gauss-Jordan elimination.
|
||||||
// Intel oneMKL LAPACK interface
|
|
||||||
#include <mkl_lapacke.h>
|
|
||||||
/* Linear equation solution using Intel oneMKL LAPACK.
|
|
||||||
a[0..n-1][0..n-1] is the input matrix. b[0..n-1] is input
|
a[0..n-1][0..n-1] is the input matrix. b[0..n-1] is input
|
||||||
containing the right-hand side vectors. On output a is
|
containing the right-hand side vectors. On output a is
|
||||||
replaced by its matrix inverse, and b is replaced by the
|
replaced by its matrix inverse, and b is replaced by the
|
||||||
corresponding set of solution vectors.
|
corresponding set of solution vectors */
|
||||||
|
|
||||||
Mathematical equivalence:
|
|
||||||
Solves: A * x = b => x = A^(-1) * b
|
|
||||||
Original Gauss-Jordan and LAPACK dgesv/dgetri produce identical results
|
|
||||||
within numerical precision. */
|
|
||||||
|
|
||||||
int gaussj(double *a, double *b, int n)
|
int gaussj(double *a, double *b, int n)
|
||||||
{
|
{
|
||||||
// Allocate pivot array and workspace
|
double swap;
|
||||||
lapack_int *ipiv = new lapack_int[n];
|
|
||||||
lapack_int info;
|
|
||||||
|
|
||||||
// Make a copy of matrix a for solving (dgesv modifies it to LU form)
|
int *indxc, *indxr, *ipiv;
|
||||||
double *a_copy = new double[n * n];
|
indxc = new int[n];
|
||||||
for (int i = 0; i < n * n; i++) {
|
indxr = new int[n];
|
||||||
a_copy[i] = a[i];
|
ipiv = new int[n];
|
||||||
|
|
||||||
|
int i, icol, irow, j, k, l, ll;
|
||||||
|
double big, dum, pivinv, temp;
|
||||||
|
|
||||||
|
for (j = 0; j < n; j++)
|
||||||
|
ipiv[j] = 0;
|
||||||
|
for (i = 0; i < n; i++)
|
||||||
|
{
|
||||||
|
big = 0.0;
|
||||||
|
for (j = 0; j < n; j++)
|
||||||
|
if (ipiv[j] != 1)
|
||||||
|
for (k = 0; k < n; k++)
|
||||||
|
{
|
||||||
|
if (ipiv[k] == 0)
|
||||||
|
{
|
||||||
|
if (fabs(a[j * n + k]) >= big)
|
||||||
|
{
|
||||||
|
big = fabs(a[j * n + k]);
|
||||||
|
irow = j;
|
||||||
|
icol = k;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
else if (ipiv[k] > 1)
|
||||||
|
{
|
||||||
|
cout << "gaussj: Singular Matrix-1" << endl;
|
||||||
|
for (int ii = 0; ii < n; ii++)
|
||||||
|
{
|
||||||
|
for (int jj = 0; jj < n; jj++)
|
||||||
|
cout << a[ii * n + jj] << " ";
|
||||||
|
cout << endl;
|
||||||
|
}
|
||||||
|
return 1; // error return
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// Step 1: Solve linear system A*x = b using LU decomposition
|
ipiv[icol] = ipiv[icol] + 1;
|
||||||
// LAPACKE_dgesv uses column-major by default, but we use row-major
|
if (irow != icol)
|
||||||
info = LAPACKE_dgesv(LAPACK_ROW_MAJOR, n, 1, a_copy, n, ipiv, b, 1);
|
{
|
||||||
|
for (l = 0; l < n; l++)
|
||||||
if (info != 0) {
|
{
|
||||||
cout << "gaussj: Singular Matrix (dgesv info=" << info << ")" << endl;
|
swap = a[irow * n + l];
|
||||||
delete[] ipiv;
|
a[irow * n + l] = a[icol * n + l];
|
||||||
delete[] a_copy;
|
a[icol * n + l] = swap;
|
||||||
return 1;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// Step 2: Compute matrix inverse A^(-1) using LU factorization
|
swap = b[irow];
|
||||||
// First do LU factorization of original matrix a
|
b[irow] = b[icol];
|
||||||
info = LAPACKE_dgetrf(LAPACK_ROW_MAJOR, n, n, a, n, ipiv);
|
b[icol] = swap;
|
||||||
|
|
||||||
if (info != 0) {
|
|
||||||
cout << "gaussj: Singular Matrix (dgetrf info=" << info << ")" << endl;
|
|
||||||
delete[] ipiv;
|
|
||||||
delete[] a_copy;
|
|
||||||
return 1;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// Then compute inverse from LU factorization
|
indxr[i] = irow;
|
||||||
info = LAPACKE_dgetri(LAPACK_ROW_MAJOR, n, a, n, ipiv);
|
indxc[i] = icol;
|
||||||
|
|
||||||
if (info != 0) {
|
if (a[icol * n + icol] == 0.0)
|
||||||
cout << "gaussj: Singular Matrix (dgetri info=" << info << ")" << endl;
|
{
|
||||||
delete[] ipiv;
|
cout << "gaussj: Singular Matrix-2" << endl;
|
||||||
delete[] a_copy;
|
for (int ii = 0; ii < n; ii++)
|
||||||
return 1;
|
{
|
||||||
|
for (int jj = 0; jj < n; jj++)
|
||||||
|
cout << a[ii * n + jj] << " ";
|
||||||
|
cout << endl;
|
||||||
|
}
|
||||||
|
return 1; // error return
|
||||||
}
|
}
|
||||||
|
|
||||||
|
pivinv = 1.0 / a[icol * n + icol];
|
||||||
|
a[icol * n + icol] = 1.0;
|
||||||
|
for (l = 0; l < n; l++)
|
||||||
|
a[icol * n + l] *= pivinv;
|
||||||
|
b[icol] *= pivinv;
|
||||||
|
for (ll = 0; ll < n; ll++)
|
||||||
|
if (ll != icol)
|
||||||
|
{
|
||||||
|
dum = a[ll * n + icol];
|
||||||
|
a[ll * n + icol] = 0.0;
|
||||||
|
for (l = 0; l < n; l++)
|
||||||
|
a[ll * n + l] -= a[icol * n + l] * dum;
|
||||||
|
b[ll] -= b[icol] * dum;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
for (l = n - 1; l >= 0; l--)
|
||||||
|
{
|
||||||
|
if (indxr[l] != indxc[l])
|
||||||
|
for (k = 0; k < n; k++)
|
||||||
|
{
|
||||||
|
swap = a[k * n + indxr[l]];
|
||||||
|
a[k * n + indxr[l]] = a[k * n + indxc[l]];
|
||||||
|
a[k * n + indxc[l]] = swap;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
delete[] indxc;
|
||||||
|
delete[] indxr;
|
||||||
delete[] ipiv;
|
delete[] ipiv;
|
||||||
delete[] a_copy;
|
|
||||||
|
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -512,10 +512,11 @@
|
|||||||
IMPLICIT DOUBLE PRECISION (A-H,O-Z)
|
IMPLICIT DOUBLE PRECISION (A-H,O-Z)
|
||||||
DIMENSION V(N),W(N)
|
DIMENSION V(N),W(N)
|
||||||
! SUBROUTINE TO COMPUTE DOUBLE PRECISION VECTOR DOT PRODUCT.
|
! SUBROUTINE TO COMPUTE DOUBLE PRECISION VECTOR DOT PRODUCT.
|
||||||
! Optimized using Intel oneMKL BLAS ddot
|
|
||||||
! Mathematical equivalence: DGVV = sum_{i=1}^{N} V(i)*W(i)
|
|
||||||
|
|
||||||
DOUBLE PRECISION, EXTERNAL :: DDOT
|
SUM = 0.0D0
|
||||||
DGVV = DDOT(N, V, 1, W, 1)
|
DO 10 I = 1,N
|
||||||
|
SUM = SUM + V(I)*W(I)
|
||||||
|
10 CONTINUE
|
||||||
|
DGVV = SUM
|
||||||
RETURN
|
RETURN
|
||||||
END
|
END
|
||||||
|
|||||||
@@ -65,7 +65,6 @@ real*8,intent(in) :: eps
|
|||||||
! dx^4
|
! dx^4
|
||||||
|
|
||||||
! note the sign (-1)^r-1, now r=2
|
! note the sign (-1)^r-1, now r=2
|
||||||
!$omp parallel do collapse(2) private(i,j,k)
|
|
||||||
do k=1,ex(3)
|
do k=1,ex(3)
|
||||||
do j=1,ex(2)
|
do j=1,ex(2)
|
||||||
do i=1,ex(1)
|
do i=1,ex(1)
|
||||||
@@ -160,7 +159,6 @@ integer, parameter :: NO_SYMM=0, OCTANT=2
|
|||||||
|
|
||||||
call symmetry_bd(3,ex,f,fh,SoA)
|
call symmetry_bd(3,ex,f,fh,SoA)
|
||||||
|
|
||||||
!$omp parallel do collapse(2) private(i,j,k)
|
|
||||||
do k=1,ex(3)
|
do k=1,ex(3)
|
||||||
do j=1,ex(2)
|
do j=1,ex(2)
|
||||||
do i=1,ex(1)
|
do i=1,ex(1)
|
||||||
@@ -275,7 +273,6 @@ real*8,intent(in) :: eps
|
|||||||
! dx^8
|
! dx^8
|
||||||
|
|
||||||
! note the sign (-1)^r-1, now r=4
|
! note the sign (-1)^r-1, now r=4
|
||||||
!$omp parallel do collapse(2) private(i,j,k)
|
|
||||||
do k=1,ex(3)
|
do k=1,ex(3)
|
||||||
do j=1,ex(2)
|
do j=1,ex(2)
|
||||||
do i=1,ex(1)
|
do i=1,ex(1)
|
||||||
@@ -388,7 +385,6 @@ real*8,intent(in) :: eps
|
|||||||
! dx^10
|
! dx^10
|
||||||
|
|
||||||
! note the sign (-1)^r-1, now r=5
|
! note the sign (-1)^r-1, now r=5
|
||||||
!$omp parallel do collapse(2) private(i,j,k)
|
|
||||||
do k=1,ex(3)
|
do k=1,ex(3)
|
||||||
do j=1,ex(2)
|
do j=1,ex(2)
|
||||||
do i=1,ex(1)
|
do i=1,ex(1)
|
||||||
|
|||||||
@@ -80,7 +80,6 @@ real*8,intent(in) :: eps
|
|||||||
! dx^4
|
! dx^4
|
||||||
|
|
||||||
! note the sign (-1)^r-1, now r=2
|
! note the sign (-1)^r-1, now r=2
|
||||||
!$omp parallel do collapse(2) private(i,j,k)
|
|
||||||
do k=1,ex(3)
|
do k=1,ex(3)
|
||||||
do j=1,ex(2)
|
do j=1,ex(2)
|
||||||
do i=1,ex(1)
|
do i=1,ex(1)
|
||||||
@@ -179,7 +178,6 @@ real*8,intent(in) :: eps
|
|||||||
! dx^4
|
! dx^4
|
||||||
|
|
||||||
! note the sign (-1)^r-1, now r=2
|
! note the sign (-1)^r-1, now r=2
|
||||||
!$omp parallel do collapse(2) private(i,j,k)
|
|
||||||
do k=1,ex(3)
|
do k=1,ex(3)
|
||||||
do j=1,ex(2)
|
do j=1,ex(2)
|
||||||
do i=1,ex(1)
|
do i=1,ex(1)
|
||||||
@@ -275,7 +273,6 @@ integer, parameter :: NO_SYMM=0, EQ_SYMM=1, OCTANT=2
|
|||||||
|
|
||||||
call symmetry_stbd(2,ex,f,fh,SoA)
|
call symmetry_stbd(2,ex,f,fh,SoA)
|
||||||
|
|
||||||
!$omp parallel do collapse(2) private(i,j,k)
|
|
||||||
do k=1,ex(3)
|
do k=1,ex(3)
|
||||||
do j=1,ex(2)
|
do j=1,ex(2)
|
||||||
do i=1,ex(1)
|
do i=1,ex(1)
|
||||||
@@ -372,7 +369,6 @@ integer, parameter :: NO_SYMM=0, EQ_SYMM=1, OCTANT=2
|
|||||||
|
|
||||||
call symmetry_stbd(3,ex,f,fh,SoA)
|
call symmetry_stbd(3,ex,f,fh,SoA)
|
||||||
|
|
||||||
!$omp parallel do collapse(2) private(i,j,k)
|
|
||||||
do k=1,ex(3)
|
do k=1,ex(3)
|
||||||
do j=1,ex(2)
|
do j=1,ex(2)
|
||||||
do i=1,ex(1)
|
do i=1,ex(1)
|
||||||
@@ -514,7 +510,6 @@ integer, parameter :: NO_SYMM=0, EQ_SYMM=1, OCTANT=2
|
|||||||
|
|
||||||
call symmetry_stbd(3,ex,f,fh,SoA)
|
call symmetry_stbd(3,ex,f,fh,SoA)
|
||||||
|
|
||||||
!$omp parallel do collapse(2) private(i,j,k)
|
|
||||||
do k=1,ex(3)
|
do k=1,ex(3)
|
||||||
do j=1,ex(2)
|
do j=1,ex(2)
|
||||||
do i=1,ex(1)
|
do i=1,ex(1)
|
||||||
@@ -603,7 +598,6 @@ integer, parameter :: NO_SYMM=0, EQ_SYMM=1, OCTANT=2
|
|||||||
|
|
||||||
call symmetry_stbd(3,ex,f,fh,SoA)
|
call symmetry_stbd(3,ex,f,fh,SoA)
|
||||||
|
|
||||||
!$omp parallel do collapse(2) private(i,j,k)
|
|
||||||
do k=1,ex(3)
|
do k=1,ex(3)
|
||||||
do j=1,ex(2)
|
do j=1,ex(2)
|
||||||
do i=1,ex(1)
|
do i=1,ex(1)
|
||||||
@@ -700,7 +694,6 @@ real*8,intent(in) :: eps
|
|||||||
! dx^8
|
! dx^8
|
||||||
|
|
||||||
! note the sign (-1)^r-1, now r=4
|
! note the sign (-1)^r-1, now r=4
|
||||||
!$omp parallel do collapse(2) private(i,j,k)
|
|
||||||
do k=1,ex(3)
|
do k=1,ex(3)
|
||||||
do j=1,ex(2)
|
do j=1,ex(2)
|
||||||
do i=1,ex(1)
|
do i=1,ex(1)
|
||||||
@@ -801,7 +794,6 @@ integer, parameter :: NO_SYMM=0, EQ_SYMM=1, OCTANT=2
|
|||||||
|
|
||||||
call symmetry_stbd(4,ex,f,fh,SoA)
|
call symmetry_stbd(4,ex,f,fh,SoA)
|
||||||
|
|
||||||
!$omp parallel do collapse(2) private(i,j,k)
|
|
||||||
do k=1,ex(3)
|
do k=1,ex(3)
|
||||||
do j=1,ex(2)
|
do j=1,ex(2)
|
||||||
do i=1,ex(1)
|
do i=1,ex(1)
|
||||||
@@ -911,7 +903,6 @@ real*8,intent(in) :: eps
|
|||||||
! dx^10
|
! dx^10
|
||||||
|
|
||||||
! note the sign (-1)^r-1, now r=5
|
! note the sign (-1)^r-1, now r=5
|
||||||
!$omp parallel do collapse(2) private(i,j,k)
|
|
||||||
do k=1,ex(3)
|
do k=1,ex(3)
|
||||||
do j=1,ex(2)
|
do j=1,ex(2)
|
||||||
do i=1,ex(1)
|
do i=1,ex(1)
|
||||||
@@ -1015,7 +1006,6 @@ integer, parameter :: NO_SYMM=0, EQ_SYMM=1, OCTANT=2
|
|||||||
|
|
||||||
call symmetry_stbd(5,ex,f,fh,SoA)
|
call symmetry_stbd(5,ex,f,fh,SoA)
|
||||||
|
|
||||||
!$omp parallel do collapse(2) private(i,j,k)
|
|
||||||
do k=1,ex(3)
|
do k=1,ex(3)
|
||||||
do j=1,ex(2)
|
do j=1,ex(2)
|
||||||
do i=1,ex(1)
|
do i=1,ex(1)
|
||||||
|
|||||||
@@ -68,7 +68,6 @@ subroutine lopsided(ex,X,Y,Z,f,f_rhs,Sfx,Sfy,Sfz,Symmetry,SoA)
|
|||||||
|
|
||||||
! upper bound set ex-1 only for efficiency,
|
! upper bound set ex-1 only for efficiency,
|
||||||
! the loop body will set ex 0 also
|
! the loop body will set ex 0 also
|
||||||
!$omp parallel do collapse(2) private(i,j,k)
|
|
||||||
do k=1,ex(3)-1
|
do k=1,ex(3)-1
|
||||||
do j=1,ex(2)-1
|
do j=1,ex(2)-1
|
||||||
do i=1,ex(1)-1
|
do i=1,ex(1)-1
|
||||||
@@ -234,7 +233,6 @@ subroutine lopsided(ex,X,Y,Z,f,f_rhs,Sfx,Sfy,Sfz,Symmetry,SoA)
|
|||||||
|
|
||||||
! upper bound set ex-1 only for efficiency,
|
! upper bound set ex-1 only for efficiency,
|
||||||
! the loop body will set ex 0 also
|
! the loop body will set ex 0 also
|
||||||
!$omp parallel do collapse(2) private(i,j,k)
|
|
||||||
do k=1,ex(3)-1
|
do k=1,ex(3)-1
|
||||||
do j=1,ex(2)-1
|
do j=1,ex(2)-1
|
||||||
do i=1,ex(1)-1
|
do i=1,ex(1)-1
|
||||||
@@ -560,7 +558,6 @@ subroutine lopsided(ex,X,Y,Z,f,f_rhs,Sfx,Sfy,Sfz,Symmetry,SoA)
|
|||||||
|
|
||||||
! upper bound set ex-1 only for efficiency,
|
! upper bound set ex-1 only for efficiency,
|
||||||
! the loop body will set ex 0 also
|
! the loop body will set ex 0 also
|
||||||
!$omp parallel do collapse(2) private(i,j,k)
|
|
||||||
do k=1,ex(3)-1
|
do k=1,ex(3)-1
|
||||||
do j=1,ex(2)-1
|
do j=1,ex(2)-1
|
||||||
do i=1,ex(1)-1
|
do i=1,ex(1)-1
|
||||||
@@ -777,7 +774,6 @@ subroutine lopsided(ex,X,Y,Z,f,f_rhs,Sfx,Sfy,Sfz,Symmetry,SoA)
|
|||||||
|
|
||||||
! upper bound set ex-1 only for efficiency,
|
! upper bound set ex-1 only for efficiency,
|
||||||
! the loop body will set ex 0 also
|
! the loop body will set ex 0 also
|
||||||
!$omp parallel do collapse(2) private(i,j,k)
|
|
||||||
do k=1,ex(3)-1
|
do k=1,ex(3)-1
|
||||||
do j=1,ex(2)-1
|
do j=1,ex(2)-1
|
||||||
do i=1,ex(1)-1
|
do i=1,ex(1)-1
|
||||||
|
|||||||
@@ -2,7 +2,7 @@
|
|||||||
#ifndef MICRODEF_H
|
#ifndef MICRODEF_H
|
||||||
#define MICRODEF_H
|
#define MICRODEF_H
|
||||||
|
|
||||||
#include "macrodef.fh"
|
#include "microdef.fh"
|
||||||
|
|
||||||
// application parameters
|
// application parameters
|
||||||
|
|
||||||
|
|||||||
@@ -16,12 +16,6 @@ include makefile.inc
|
|||||||
.cu.o:
|
.cu.o:
|
||||||
$(Cu) $(CUDA_APP_FLAGS) -c $< -o $@ $(CUDA_LIB_PATH)
|
$(Cu) $(CUDA_APP_FLAGS) -c $< -o $@ $(CUDA_LIB_PATH)
|
||||||
|
|
||||||
TwoPunctures.o: TwoPunctures.C
|
|
||||||
${CXX} $(CXXAPPFLAGS) -qopenmp -c $< -o $@
|
|
||||||
|
|
||||||
TwoPunctureABE.o: TwoPunctureABE.C
|
|
||||||
${CXX} $(CXXAPPFLAGS) -qopenmp -c $< -o $@
|
|
||||||
|
|
||||||
# Input files
|
# Input files
|
||||||
C++FILES = ABE.o Ansorg.o Block.o misc.o monitor.o Parallel.o MPatch.o var.o\
|
C++FILES = ABE.o Ansorg.o Block.o misc.o monitor.o Parallel.o MPatch.o var.o\
|
||||||
cgh.o bssn_class.o surface_integral.o ShellPatch.o\
|
cgh.o bssn_class.o surface_integral.o ShellPatch.o\
|
||||||
@@ -96,13 +90,13 @@ misc.o : zbesh.o
|
|||||||
|
|
||||||
# projects
|
# projects
|
||||||
ABE: $(C++FILES) $(F90FILES) $(F77FILES) $(AHFDOBJS)
|
ABE: $(C++FILES) $(F90FILES) $(F77FILES) $(AHFDOBJS)
|
||||||
$(CLINKER) $(CXXAPPFLAGS) -qopenmp -o $@ $(C++FILES) $(F90FILES) $(F77FILES) $(AHFDOBJS) $(LDLIBS)
|
$(CLINKER) $(CXXAPPFLAGS) -o $@ $(C++FILES) $(F90FILES) $(F77FILES) $(AHFDOBJS) $(LDLIBS)
|
||||||
|
|
||||||
ABEGPU: $(C++FILES_GPU) $(F90FILES) $(F77FILES) $(AHFDOBJS) $(CUDAFILES)
|
ABEGPU: $(C++FILES_GPU) $(F90FILES) $(F77FILES) $(AHFDOBJS) $(CUDAFILES)
|
||||||
$(CLINKER) $(CXXAPPFLAGS) -o $@ $(C++FILES_GPU) $(F90FILES) $(F77FILES) $(AHFDOBJS) $(CUDAFILES) $(LDLIBS)
|
$(CLINKER) $(CXXAPPFLAGS) -o $@ $(C++FILES_GPU) $(F90FILES) $(F77FILES) $(AHFDOBJS) $(CUDAFILES) $(LDLIBS)
|
||||||
|
|
||||||
TwoPunctureABE: $(TwoPunctureFILES)
|
TwoPunctureABE: $(TwoPunctureFILES)
|
||||||
$(CLINKER) $(CXXAPPFLAGS) -qopenmp -o $@ $(TwoPunctureFILES) $(LDLIBS)
|
$(CLINKER) $(CXXAPPFLAGS) -o $@ $(TwoPunctureFILES) $(LDLIBS)
|
||||||
|
|
||||||
clean:
|
clean:
|
||||||
rm *.o ABE ABEGPU TwoPunctureABE make.log -f
|
rm *.o ABE ABEGPU TwoPunctureABE make.log -f
|
||||||
|
|||||||
@@ -6,25 +6,28 @@
|
|||||||
## Intel oneAPI version with oneMKL (Optimized for performance)
|
## Intel oneAPI version with oneMKL (Optimized for performance)
|
||||||
filein = -I/usr/include/ -I${MKLROOT}/include
|
filein = -I/usr/include/ -I${MKLROOT}/include
|
||||||
|
|
||||||
## Using OpenMP-threaded MKL for parallel performance
|
## Using sequential MKL (OpenMP disabled for better single-threaded performance)
|
||||||
## Added -lifcore for Intel Fortran runtime and -limf for Intel math library
|
## Added -lifcore for Intel Fortran runtime and -limf for Intel math library
|
||||||
LDLIBS = -L${MKLROOT}/lib -lmkl_intel_lp64 -lmkl_intel_thread -lmkl_core -liomp5 -lifcore -limf -lpthread -lm -ldl
|
LDLIBS = -L${MKLROOT}/lib -lmkl_intel_lp64 -lmkl_sequential -lmkl_core -lifcore -limf -lpthread -lm -ldl
|
||||||
|
|
||||||
## Aggressive optimization flags + PGO Phase 2 (profile-guided optimization)
|
## Aggressive optimization flags:
|
||||||
## -fprofile-instr-use: use collected profile data to guide optimization decisions
|
## -O3: Maximum optimization
|
||||||
## (branch prediction, basic block layout, inlining, loop unrolling)
|
## -xHost: Optimize for the host CPU architecture (Intel/AMD compatible)
|
||||||
PROFDATA = /home/amss/AMSS-NCKU/pgo_profile/default.profdata
|
## -fp-model fast=2: Aggressive floating-point optimizations
|
||||||
CXXAPPFLAGS = -O3 -march=native -fp-model fast=2 -fma -ipo -qopenmp \
|
## -fma: Enable fused multiply-add instructions
|
||||||
-DMPI_STUB -Dfortran3 -Dnewc -I${MKLROOT}/include
|
## Note: OpenMP has been disabled (-qopenmp removed) due to performance issues
|
||||||
f90appflags = -O3 -march=native -fp-model fast=2 -fma -ipo -qopenmp \
|
CXXAPPFLAGS = -O3 -xHost -fp-model fast=2 -fma \
|
||||||
-align array64byte -fpp -I${MKLROOT}/include
|
-Dfortran3 -Dnewc -I${MKLROOT}/include
|
||||||
|
f90appflags = -O3 -xHost -fp-model fast=2 -fma \
|
||||||
|
-fpp -I${MKLROOT}/include
|
||||||
f90 = ifx
|
f90 = ifx
|
||||||
f77 = ifx
|
f77 = ifx
|
||||||
CXX = icpx
|
CXX = icpx
|
||||||
CC = icx
|
CC = icx
|
||||||
CLINKER = icpx
|
CLINKER = mpiicpx
|
||||||
|
|
||||||
Cu = nvcc
|
Cu = nvcc
|
||||||
CUDA_LIB_PATH = -L/usr/lib/cuda/lib64 -I/usr/include -I/usr/lib/cuda/include
|
CUDA_LIB_PATH = -L/usr/lib/cuda/lib64 -I/usr/include -I/usr/lib/cuda/include
|
||||||
#CUDA_APP_FLAGS = -c -g -O3 --ptxas-options=-v -arch compute_13 -code compute_13,sm_13 -Dfortran3 -Dnewc
|
#CUDA_APP_FLAGS = -c -g -O3 --ptxas-options=-v -arch compute_13 -code compute_13,sm_13 -Dfortran3 -Dnewc
|
||||||
CUDA_APP_FLAGS = -c -g -O3 --ptxas-options=-v -Dfortran3 -Dnewc
|
CUDA_APP_FLAGS = -c -g -O3 --ptxas-options=-v -Dfortran3 -Dnewc
|
||||||
|
|
||||||
|
|||||||
@@ -14,11 +14,7 @@ using namespace std;
|
|||||||
#include <string.h>
|
#include <string.h>
|
||||||
#include <math.h>
|
#include <math.h>
|
||||||
#endif
|
#endif
|
||||||
#ifdef MPI_STUB
|
|
||||||
#include "mpi_stub.h"
|
|
||||||
#else
|
|
||||||
#include <mpi.h>
|
#include <mpi.h>
|
||||||
#endif
|
|
||||||
|
|
||||||
#include "misc.h"
|
#include "misc.h"
|
||||||
#include "macrodef.h"
|
#include "macrodef.h"
|
||||||
|
|||||||
@@ -24,11 +24,7 @@ using namespace std;
|
|||||||
#include <complex.h>
|
#include <complex.h>
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
#ifdef MPI_STUB
|
|
||||||
#include "mpi_stub.h"
|
|
||||||
#else
|
|
||||||
#include <mpi.h>
|
#include <mpi.h>
|
||||||
#endif
|
|
||||||
|
|
||||||
namespace misc
|
namespace misc
|
||||||
{
|
{
|
||||||
|
|||||||
@@ -20,11 +20,7 @@ using namespace std;
|
|||||||
#endif
|
#endif
|
||||||
#include <time.h>
|
#include <time.h>
|
||||||
|
|
||||||
#ifdef MPI_STUB
|
|
||||||
#include "mpi_stub.h"
|
|
||||||
#else
|
|
||||||
#include <mpi.h>
|
#include <mpi.h>
|
||||||
#endif
|
|
||||||
|
|
||||||
class monitor
|
class monitor
|
||||||
{
|
{
|
||||||
|
|||||||
@@ -1,153 +0,0 @@
|
|||||||
#ifndef MPI_STUB_H
|
|
||||||
#define MPI_STUB_H
|
|
||||||
|
|
||||||
/*
|
|
||||||
* MPI Stub Header — single-process shim for AMSS-NCKU ABE solver.
|
|
||||||
* Provides all MPI types, constants, and functions used in the codebase
|
|
||||||
* as no-ops or trivial implementations for nprocs=1, myrank=0.
|
|
||||||
*/
|
|
||||||
|
|
||||||
#include <cstring>
|
|
||||||
#include <cstdlib>
|
|
||||||
#include <cstdio>
|
|
||||||
#include <time.h>
|
|
||||||
|
|
||||||
/* ── Types ─────────────────────────────────────────────────────────── */
|
|
||||||
|
|
||||||
typedef int MPI_Comm;
|
|
||||||
typedef int MPI_Datatype;
|
|
||||||
typedef int MPI_Op;
|
|
||||||
typedef int MPI_Request;
|
|
||||||
typedef int MPI_Group;
|
|
||||||
|
|
||||||
typedef struct MPI_Status {
|
|
||||||
int MPI_SOURCE;
|
|
||||||
int MPI_TAG;
|
|
||||||
int MPI_ERROR;
|
|
||||||
} MPI_Status;
|
|
||||||
|
|
||||||
/* ── Constants ─────────────────────────────────────────────────────── */
|
|
||||||
|
|
||||||
#define MPI_COMM_WORLD 0
|
|
||||||
|
|
||||||
#define MPI_INT 1
|
|
||||||
#define MPI_DOUBLE 2
|
|
||||||
#define MPI_DOUBLE_PRECISION 2
|
|
||||||
#define MPI_DOUBLE_INT 3
|
|
||||||
|
|
||||||
#define MPI_SUM 1
|
|
||||||
#define MPI_MAX 2
|
|
||||||
#define MPI_MAXLOC 3
|
|
||||||
|
|
||||||
#define MPI_STATUS_IGNORE ((MPI_Status *)0)
|
|
||||||
#define MPI_STATUSES_IGNORE ((MPI_Status *)0)
|
|
||||||
|
|
||||||
#define MPI_MAX_PROCESSOR_NAME 256
|
|
||||||
|
|
||||||
/* ── Helper: sizeof for MPI_Datatype ──────────────────────────────── */
|
|
||||||
|
|
||||||
static inline size_t mpi_stub_sizeof(MPI_Datatype type) {
|
|
||||||
switch (type) {
|
|
||||||
case MPI_INT: return sizeof(int);
|
|
||||||
case MPI_DOUBLE: return sizeof(double);
|
|
||||||
case MPI_DOUBLE_INT: return sizeof(double) + sizeof(int);
|
|
||||||
default: return 0;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
/* ── Init / Finalize ──────────────────────────────────────────────── */
|
|
||||||
|
|
||||||
static inline int MPI_Init(int *, char ***) { return 0; }
|
|
||||||
static inline int MPI_Finalize() { return 0; }
|
|
||||||
|
|
||||||
/* ── Communicator queries ─────────────────────────────────────────── */
|
|
||||||
|
|
||||||
static inline int MPI_Comm_rank(MPI_Comm, int *rank) { *rank = 0; return 0; }
|
|
||||||
static inline int MPI_Comm_size(MPI_Comm, int *size) { *size = 1; return 0; }
|
|
||||||
static inline int MPI_Comm_split(MPI_Comm comm, int, int, MPI_Comm *newcomm) {
|
|
||||||
*newcomm = comm;
|
|
||||||
return 0;
|
|
||||||
}
|
|
||||||
static inline int MPI_Comm_free(MPI_Comm *) { return 0; }
|
|
||||||
|
|
||||||
/* ── Group operations ─────────────────────────────────────────────── */
|
|
||||||
|
|
||||||
static inline int MPI_Comm_group(MPI_Comm, MPI_Group *group) {
|
|
||||||
*group = 0;
|
|
||||||
return 0;
|
|
||||||
}
|
|
||||||
static inline int MPI_Group_translate_ranks(MPI_Group, int n,
|
|
||||||
const int *ranks1, MPI_Group, int *ranks2) {
|
|
||||||
for (int i = 0; i < n; ++i) ranks2[i] = ranks1[i];
|
|
||||||
return 0;
|
|
||||||
}
|
|
||||||
static inline int MPI_Group_free(MPI_Group *) { return 0; }
|
|
||||||
|
|
||||||
/* ── Collective operations ────────────────────────────────────────── */
|
|
||||||
|
|
||||||
static inline int MPI_Allreduce(const void *sendbuf, void *recvbuf,
|
|
||||||
int count, MPI_Datatype datatype, MPI_Op, MPI_Comm) {
|
|
||||||
std::memcpy(recvbuf, sendbuf, count * mpi_stub_sizeof(datatype));
|
|
||||||
return 0;
|
|
||||||
}
|
|
||||||
|
|
||||||
static inline int MPI_Iallreduce(const void *sendbuf, void *recvbuf,
|
|
||||||
int count, MPI_Datatype datatype, MPI_Op, MPI_Comm,
|
|
||||||
MPI_Request *request) {
|
|
||||||
std::memcpy(recvbuf, sendbuf, count * mpi_stub_sizeof(datatype));
|
|
||||||
*request = 0;
|
|
||||||
return 0;
|
|
||||||
}
|
|
||||||
|
|
||||||
static inline int MPI_Bcast(void *, int, MPI_Datatype, int, MPI_Comm) {
|
|
||||||
return 0;
|
|
||||||
}
|
|
||||||
|
|
||||||
static inline int MPI_Barrier(MPI_Comm) { return 0; }
|
|
||||||
|
|
||||||
/* ── Point-to-point (never reached with nprocs=1) ─────────────────── */
|
|
||||||
|
|
||||||
static inline int MPI_Send(const void *, int, MPI_Datatype, int, int, MPI_Comm) {
|
|
||||||
return 0;
|
|
||||||
}
|
|
||||||
static inline int MPI_Recv(void *, int, MPI_Datatype, int, int, MPI_Comm, MPI_Status *) {
|
|
||||||
return 0;
|
|
||||||
}
|
|
||||||
static inline int MPI_Isend(const void *, int, MPI_Datatype, int, int, MPI_Comm,
|
|
||||||
MPI_Request *req) {
|
|
||||||
*req = 0;
|
|
||||||
return 0;
|
|
||||||
}
|
|
||||||
static inline int MPI_Irecv(void *, int, MPI_Datatype, int, int, MPI_Comm,
|
|
||||||
MPI_Request *req) {
|
|
||||||
*req = 0;
|
|
||||||
return 0;
|
|
||||||
}
|
|
||||||
|
|
||||||
/* ── Completion ───────────────────────────────────────────────────── */
|
|
||||||
|
|
||||||
static inline int MPI_Wait(MPI_Request *, MPI_Status *) { return 0; }
|
|
||||||
static inline int MPI_Waitall(int, MPI_Request *, MPI_Status *) { return 0; }
|
|
||||||
|
|
||||||
/* ── Utility ──────────────────────────────────────────────────────── */
|
|
||||||
|
|
||||||
static inline int MPI_Abort(MPI_Comm, int error_code) {
|
|
||||||
std::fprintf(stderr, "MPI_Abort called with error code %d\n", error_code);
|
|
||||||
std::exit(error_code);
|
|
||||||
return 0;
|
|
||||||
}
|
|
||||||
|
|
||||||
static inline double MPI_Wtime() {
|
|
||||||
struct timespec ts;
|
|
||||||
clock_gettime(CLOCK_MONOTONIC, &ts);
|
|
||||||
return (double)ts.tv_sec + (double)ts.tv_nsec * 1.0e-9;
|
|
||||||
}
|
|
||||||
|
|
||||||
static inline int MPI_Get_processor_name(char *name, int *resultlen) {
|
|
||||||
const char *stub_name = "localhost";
|
|
||||||
std::strcpy(name, stub_name);
|
|
||||||
*resultlen = (int)std::strlen(stub_name);
|
|
||||||
return 0;
|
|
||||||
}
|
|
||||||
|
|
||||||
#endif /* MPI_STUB_H */
|
|
||||||
@@ -24,11 +24,7 @@ using namespace std;
|
|||||||
#include <map.h>
|
#include <map.h>
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
#ifdef MPI_STUB
|
|
||||||
#include "mpi_stub.h"
|
|
||||||
#else
|
|
||||||
#include <mpi.h>
|
#include <mpi.h>
|
||||||
#endif
|
|
||||||
|
|
||||||
namespace parameters
|
namespace parameters
|
||||||
{
|
{
|
||||||
|
|||||||
@@ -30,11 +30,7 @@ using namespace std;
|
|||||||
#include <sys/time.h>
|
#include <sys/time.h>
|
||||||
#include <sys/resource.h>
|
#include <sys/resource.h>
|
||||||
|
|
||||||
#ifdef MPI_STUB
|
|
||||||
#include "mpi_stub.h"
|
|
||||||
#else
|
|
||||||
#include <mpi.h>
|
#include <mpi.h>
|
||||||
#endif
|
|
||||||
|
|
||||||
/* Real time */
|
/* Real time */
|
||||||
#define TimerSignal SIGALRM
|
#define TimerSignal SIGALRM
|
||||||
|
|||||||
@@ -109,33 +109,23 @@
|
|||||||
|
|
||||||
if( RK4 == 0 ) then
|
if( RK4 == 0 ) then
|
||||||
|
|
||||||
!$omp parallel workshare
|
|
||||||
f1 = f0 + HLF * dT * f_rhs
|
f1 = f0 + HLF * dT * f_rhs
|
||||||
!$omp end parallel workshare
|
|
||||||
|
|
||||||
elseif(RK4 == 1 ) then
|
elseif(RK4 == 1 ) then
|
||||||
|
|
||||||
!$omp parallel workshare
|
|
||||||
f_rhs = f_rhs + TWO * f1
|
f_rhs = f_rhs + TWO * f1
|
||||||
!$omp end parallel workshare
|
|
||||||
!$omp parallel workshare
|
|
||||||
f1 = f0 + HLF * dT * f1
|
f1 = f0 + HLF * dT * f1
|
||||||
!$omp end parallel workshare
|
|
||||||
|
|
||||||
elseif(RK4 == 2 ) then
|
elseif(RK4 == 2 ) then
|
||||||
|
|
||||||
!$omp parallel workshare
|
|
||||||
f_rhs = f_rhs + TWO * f1
|
f_rhs = f_rhs + TWO * f1
|
||||||
!$omp end parallel workshare
|
|
||||||
!$omp parallel workshare
|
|
||||||
f1 = f0 + dT * f1
|
f1 = f0 + dT * f1
|
||||||
!$omp end parallel workshare
|
|
||||||
|
|
||||||
elseif( RK4 == 3 ) then
|
elseif( RK4 == 3 ) then
|
||||||
|
|
||||||
!$omp parallel workshare
|
|
||||||
f1 = f0 +F1o6 * dT *(f1 + f_rhs)
|
f1 = f0 +F1o6 * dT *(f1 + f_rhs)
|
||||||
!$omp end parallel workshare
|
|
||||||
|
|
||||||
else
|
else
|
||||||
|
|
||||||
@@ -225,15 +215,11 @@
|
|||||||
|
|
||||||
if( RK4 == 0 ) then
|
if( RK4 == 0 ) then
|
||||||
|
|
||||||
!$omp parallel workshare
|
|
||||||
f1 = f0 + dT * f_rhs
|
f1 = f0 + dT * f_rhs
|
||||||
!$omp end parallel workshare
|
|
||||||
|
|
||||||
else
|
else
|
||||||
|
|
||||||
!$omp parallel workshare
|
|
||||||
f1 = f0 + HLF * dT * (f1+f_rhs)
|
f1 = f0 + HLF * dT * (f1+f_rhs)
|
||||||
!$omp end parallel workshare
|
|
||||||
|
|
||||||
endif
|
endif
|
||||||
|
|
||||||
@@ -253,9 +239,7 @@
|
|||||||
real*8, dimension(ex(1),ex(2),ex(3)),intent(in) ::f_rhs
|
real*8, dimension(ex(1),ex(2),ex(3)),intent(in) ::f_rhs
|
||||||
real*8, dimension(ex(1),ex(2),ex(3)),intent(out) ::f1
|
real*8, dimension(ex(1),ex(2),ex(3)),intent(out) ::f1
|
||||||
|
|
||||||
!$omp parallel workshare
|
|
||||||
f1 = f0 + dT * f_rhs
|
f1 = f0 + dT * f_rhs
|
||||||
!$omp end parallel workshare
|
|
||||||
|
|
||||||
return
|
return
|
||||||
|
|
||||||
|
|||||||
@@ -19,11 +19,7 @@ using namespace std;
|
|||||||
#include <math.h>
|
#include <math.h>
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
#ifdef MPI_STUB
|
|
||||||
#include "mpi_stub.h"
|
|
||||||
#else
|
|
||||||
#include <mpi.h>
|
#include <mpi.h>
|
||||||
#endif
|
|
||||||
|
|
||||||
#include "cgh.h"
|
#include "cgh.h"
|
||||||
#include "ShellPatch.h"
|
#include "ShellPatch.h"
|
||||||
|
|||||||
@@ -18,11 +18,7 @@ using namespace std;
|
|||||||
#include <math.h>
|
#include <math.h>
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
#ifdef MPI_STUB
|
|
||||||
#include "mpi_stub.h"
|
|
||||||
#else
|
|
||||||
#include <mpi.h>
|
#include <mpi.h>
|
||||||
#endif
|
|
||||||
|
|
||||||
#include "misc.h"
|
#include "misc.h"
|
||||||
#include "microdef.h"
|
#include "microdef.h"
|
||||||
|
|||||||
@@ -3,11 +3,7 @@
|
|||||||
#include <math.h>
|
#include <math.h>
|
||||||
#include <string.h>
|
#include <string.h>
|
||||||
|
|
||||||
#ifdef MPI_STUB
|
|
||||||
#include "mpi_stub.h"
|
|
||||||
#else
|
|
||||||
#include <mpi.h>
|
#include <mpi.h>
|
||||||
#endif
|
|
||||||
|
|
||||||
#include "util_Table.h"
|
#include "util_Table.h"
|
||||||
#include "cctk.h"
|
#include "cctk.h"
|
||||||
|
|||||||
@@ -20,11 +20,7 @@ using namespace std;
|
|||||||
#include <math.h>
|
#include <math.h>
|
||||||
#include <map.h>
|
#include <map.h>
|
||||||
#endif
|
#endif
|
||||||
#ifdef MPI_STUB
|
|
||||||
#include "mpi_stub.h"
|
|
||||||
#else
|
|
||||||
#include <mpi.h>
|
#include <mpi.h>
|
||||||
#endif
|
|
||||||
|
|
||||||
#include "misc.h"
|
#include "misc.h"
|
||||||
#include "cgh.h"
|
#include "cgh.h"
|
||||||
@@ -224,9 +220,16 @@ void surface_integral::surf_Wave(double rex, int lev, cgh *GH, var *Rpsi4, var *
|
|||||||
pox[2][n] = rex * nz_g[n];
|
pox[2][n] = rex * nz_g[n];
|
||||||
}
|
}
|
||||||
|
|
||||||
|
double *shellf;
|
||||||
|
shellf = new double[n_tot * InList];
|
||||||
|
|
||||||
|
GH->PatL[lev]->data->Interp_Points(DG_List, n_tot, pox, shellf, Symmetry);
|
||||||
|
|
||||||
int mp, Lp, Nmin, Nmax;
|
int mp, Lp, Nmin, Nmax;
|
||||||
|
|
||||||
mp = n_tot / cpusize;
|
mp = n_tot / cpusize;
|
||||||
Lp = n_tot - cpusize * mp;
|
Lp = n_tot - cpusize * mp;
|
||||||
|
|
||||||
if (Lp > myrank)
|
if (Lp > myrank)
|
||||||
{
|
{
|
||||||
Nmin = myrank * mp + myrank;
|
Nmin = myrank * mp + myrank;
|
||||||
@@ -238,11 +241,6 @@ void surface_integral::surf_Wave(double rex, int lev, cgh *GH, var *Rpsi4, var *
|
|||||||
Nmax = Nmin + mp - 1;
|
Nmax = Nmin + mp - 1;
|
||||||
}
|
}
|
||||||
|
|
||||||
double *shellf;
|
|
||||||
shellf = new double[n_tot * InList];
|
|
||||||
|
|
||||||
GH->PatL[lev]->data->Interp_Points(DG_List, n_tot, pox, shellf, Symmetry, Nmin, Nmax);
|
|
||||||
|
|
||||||
//|~~~~~> Integrate the dot product of Dphi with the surface normal.
|
//|~~~~~> Integrate the dot product of Dphi with the surface normal.
|
||||||
|
|
||||||
double *RP_out, *IP_out;
|
double *RP_out, *IP_out;
|
||||||
@@ -365,17 +363,8 @@ void surface_integral::surf_Wave(double rex, int lev, cgh *GH, var *Rpsi4, var *
|
|||||||
}
|
}
|
||||||
//|------+ Communicate and sum the results from each processor.
|
//|------+ Communicate and sum the results from each processor.
|
||||||
|
|
||||||
{
|
MPI_Allreduce(RP_out, RP, NN, MPI_DOUBLE, MPI_SUM, MPI_COMM_WORLD);
|
||||||
double *RPIP_out = new double[2 * NN];
|
MPI_Allreduce(IP_out, IP, NN, MPI_DOUBLE, MPI_SUM, MPI_COMM_WORLD);
|
||||||
double *RPIP = new double[2 * NN];
|
|
||||||
memcpy(RPIP_out, RP_out, NN * sizeof(double));
|
|
||||||
memcpy(RPIP_out + NN, IP_out, NN * sizeof(double));
|
|
||||||
MPI_Allreduce(RPIP_out, RPIP, 2 * NN, MPI_DOUBLE, MPI_SUM, MPI_COMM_WORLD);
|
|
||||||
memcpy(RP, RPIP, NN * sizeof(double));
|
|
||||||
memcpy(IP, RPIP + NN, NN * sizeof(double));
|
|
||||||
delete[] RPIP_out;
|
|
||||||
delete[] RPIP;
|
|
||||||
}
|
|
||||||
|
|
||||||
//|------= Free memory.
|
//|------= Free memory.
|
||||||
|
|
||||||
@@ -567,17 +556,8 @@ void surface_integral::surf_Wave(double rex, int lev, cgh *GH, var *Rpsi4, var *
|
|||||||
}
|
}
|
||||||
//|------+ Communicate and sum the results from each processor.
|
//|------+ Communicate and sum the results from each processor.
|
||||||
|
|
||||||
{
|
MPI_Allreduce(RP_out, RP, NN, MPI_DOUBLE, MPI_SUM, Comm_here);
|
||||||
double *RPIP_out = new double[2 * NN];
|
MPI_Allreduce(IP_out, IP, NN, MPI_DOUBLE, MPI_SUM, Comm_here);
|
||||||
double *RPIP = new double[2 * NN];
|
|
||||||
memcpy(RPIP_out, RP_out, NN * sizeof(double));
|
|
||||||
memcpy(RPIP_out + NN, IP_out, NN * sizeof(double));
|
|
||||||
MPI_Allreduce(RPIP_out, RPIP, 2 * NN, MPI_DOUBLE, MPI_SUM, Comm_here);
|
|
||||||
memcpy(RP, RPIP, NN * sizeof(double));
|
|
||||||
memcpy(IP, RPIP + NN, NN * sizeof(double));
|
|
||||||
delete[] RPIP_out;
|
|
||||||
delete[] RPIP;
|
|
||||||
}
|
|
||||||
|
|
||||||
//|------= Free memory.
|
//|------= Free memory.
|
||||||
|
|
||||||
@@ -755,17 +735,8 @@ void surface_integral::surf_Wave(double rex, int lev, ShellPatch *GH, var *Rpsi4
|
|||||||
}
|
}
|
||||||
//|------+ Communicate and sum the results from each processor.
|
//|------+ Communicate and sum the results from each processor.
|
||||||
|
|
||||||
{
|
MPI_Allreduce(RP_out, RP, NN, MPI_DOUBLE, MPI_SUM, MPI_COMM_WORLD);
|
||||||
double *RPIP_out = new double[2 * NN];
|
MPI_Allreduce(IP_out, IP, NN, MPI_DOUBLE, MPI_SUM, MPI_COMM_WORLD);
|
||||||
double *RPIP = new double[2 * NN];
|
|
||||||
memcpy(RPIP_out, RP_out, NN * sizeof(double));
|
|
||||||
memcpy(RPIP_out + NN, IP_out, NN * sizeof(double));
|
|
||||||
MPI_Allreduce(RPIP_out, RPIP, 2 * NN, MPI_DOUBLE, MPI_SUM, MPI_COMM_WORLD);
|
|
||||||
memcpy(RP, RPIP, NN * sizeof(double));
|
|
||||||
memcpy(IP, RPIP + NN, NN * sizeof(double));
|
|
||||||
delete[] RPIP_out;
|
|
||||||
delete[] RPIP;
|
|
||||||
}
|
|
||||||
|
|
||||||
//|------= Free memory.
|
//|------= Free memory.
|
||||||
|
|
||||||
@@ -1013,17 +984,8 @@ void surface_integral::surf_Wave(double rex, int lev, ShellPatch *GH,
|
|||||||
}
|
}
|
||||||
//|------+ Communicate and sum the results from each processor.
|
//|------+ Communicate and sum the results from each processor.
|
||||||
|
|
||||||
{
|
MPI_Allreduce(RP_out, RP, NN, MPI_DOUBLE, MPI_SUM, MPI_COMM_WORLD);
|
||||||
double *RPIP_out = new double[2 * NN];
|
MPI_Allreduce(IP_out, IP, NN, MPI_DOUBLE, MPI_SUM, MPI_COMM_WORLD);
|
||||||
double *RPIP = new double[2 * NN];
|
|
||||||
memcpy(RPIP_out, RP_out, NN * sizeof(double));
|
|
||||||
memcpy(RPIP_out + NN, IP_out, NN * sizeof(double));
|
|
||||||
MPI_Allreduce(RPIP_out, RPIP, 2 * NN, MPI_DOUBLE, MPI_SUM, MPI_COMM_WORLD);
|
|
||||||
memcpy(RP, RPIP, NN * sizeof(double));
|
|
||||||
memcpy(IP, RPIP + NN, NN * sizeof(double));
|
|
||||||
delete[] RPIP_out;
|
|
||||||
delete[] RPIP;
|
|
||||||
}
|
|
||||||
|
|
||||||
//|------= Free memory.
|
//|------= Free memory.
|
||||||
|
|
||||||
@@ -1457,17 +1419,8 @@ void surface_integral::surf_Wave(double rex, int lev, ShellPatch *GH,
|
|||||||
}
|
}
|
||||||
//|------+ Communicate and sum the results from each processor.
|
//|------+ Communicate and sum the results from each processor.
|
||||||
|
|
||||||
{
|
MPI_Allreduce(RP_out, RP, NN, MPI_DOUBLE, MPI_SUM, MPI_COMM_WORLD);
|
||||||
double *RPIP_out = new double[2 * NN];
|
MPI_Allreduce(IP_out, IP, NN, MPI_DOUBLE, MPI_SUM, MPI_COMM_WORLD);
|
||||||
double *RPIP = new double[2 * NN];
|
|
||||||
memcpy(RPIP_out, RP_out, NN * sizeof(double));
|
|
||||||
memcpy(RPIP_out + NN, IP_out, NN * sizeof(double));
|
|
||||||
MPI_Allreduce(RPIP_out, RPIP, 2 * NN, MPI_DOUBLE, MPI_SUM, MPI_COMM_WORLD);
|
|
||||||
memcpy(RP, RPIP, NN * sizeof(double));
|
|
||||||
memcpy(IP, RPIP + NN, NN * sizeof(double));
|
|
||||||
delete[] RPIP_out;
|
|
||||||
delete[] RPIP;
|
|
||||||
}
|
|
||||||
|
|
||||||
//|------= Free memory.
|
//|------= Free memory.
|
||||||
|
|
||||||
@@ -1901,17 +1854,8 @@ void surface_integral::surf_Wave(double rex, int lev, cgh *GH,
|
|||||||
}
|
}
|
||||||
//|------+ Communicate and sum the results from each processor.
|
//|------+ Communicate and sum the results from each processor.
|
||||||
|
|
||||||
{
|
MPI_Allreduce(RP_out, RP, NN, MPI_DOUBLE, MPI_SUM, MPI_COMM_WORLD);
|
||||||
double *RPIP_out = new double[2 * NN];
|
MPI_Allreduce(IP_out, IP, NN, MPI_DOUBLE, MPI_SUM, MPI_COMM_WORLD);
|
||||||
double *RPIP = new double[2 * NN];
|
|
||||||
memcpy(RPIP_out, RP_out, NN * sizeof(double));
|
|
||||||
memcpy(RPIP_out + NN, IP_out, NN * sizeof(double));
|
|
||||||
MPI_Allreduce(RPIP_out, RPIP, 2 * NN, MPI_DOUBLE, MPI_SUM, MPI_COMM_WORLD);
|
|
||||||
memcpy(RP, RPIP, NN * sizeof(double));
|
|
||||||
memcpy(IP, RPIP + NN, NN * sizeof(double));
|
|
||||||
delete[] RPIP_out;
|
|
||||||
delete[] RPIP;
|
|
||||||
}
|
|
||||||
|
|
||||||
//|------= Free memory.
|
//|------= Free memory.
|
||||||
|
|
||||||
@@ -2096,17 +2040,8 @@ void surface_integral::surf_Wave(double rex, int lev, NullShellPatch2 *GH, var *
|
|||||||
}
|
}
|
||||||
//|------+ Communicate and sum the results from each processor.
|
//|------+ Communicate and sum the results from each processor.
|
||||||
|
|
||||||
{
|
MPI_Allreduce(RP_out, RP, NN, MPI_DOUBLE, MPI_SUM, MPI_COMM_WORLD);
|
||||||
double *RPIP_out = new double[2 * NN];
|
MPI_Allreduce(IP_out, IP, NN, MPI_DOUBLE, MPI_SUM, MPI_COMM_WORLD);
|
||||||
double *RPIP = new double[2 * NN];
|
|
||||||
memcpy(RPIP_out, RP_out, NN * sizeof(double));
|
|
||||||
memcpy(RPIP_out + NN, IP_out, NN * sizeof(double));
|
|
||||||
MPI_Allreduce(RPIP_out, RPIP, 2 * NN, MPI_DOUBLE, MPI_SUM, MPI_COMM_WORLD);
|
|
||||||
memcpy(RP, RPIP, NN * sizeof(double));
|
|
||||||
memcpy(IP, RPIP + NN, NN * sizeof(double));
|
|
||||||
delete[] RPIP_out;
|
|
||||||
delete[] RPIP;
|
|
||||||
}
|
|
||||||
|
|
||||||
//|------= Free memory.
|
//|------= Free memory.
|
||||||
|
|
||||||
@@ -2291,17 +2226,8 @@ void surface_integral::surf_Wave(double rex, int lev, NullShellPatch *GH, var *R
|
|||||||
}
|
}
|
||||||
//|------+ Communicate and sum the results from each processor.
|
//|------+ Communicate and sum the results from each processor.
|
||||||
|
|
||||||
{
|
MPI_Allreduce(RP_out, RP, NN, MPI_DOUBLE, MPI_SUM, MPI_COMM_WORLD);
|
||||||
double *RPIP_out = new double[2 * NN];
|
MPI_Allreduce(IP_out, IP, NN, MPI_DOUBLE, MPI_SUM, MPI_COMM_WORLD);
|
||||||
double *RPIP = new double[2 * NN];
|
|
||||||
memcpy(RPIP_out, RP_out, NN * sizeof(double));
|
|
||||||
memcpy(RPIP_out + NN, IP_out, NN * sizeof(double));
|
|
||||||
MPI_Allreduce(RPIP_out, RPIP, 2 * NN, MPI_DOUBLE, MPI_SUM, MPI_COMM_WORLD);
|
|
||||||
memcpy(RP, RPIP, NN * sizeof(double));
|
|
||||||
memcpy(IP, RPIP + NN, NN * sizeof(double));
|
|
||||||
delete[] RPIP_out;
|
|
||||||
delete[] RPIP;
|
|
||||||
}
|
|
||||||
|
|
||||||
//|------= Free memory.
|
//|------= Free memory.
|
||||||
|
|
||||||
@@ -2388,9 +2314,25 @@ void surface_integral::surf_MassPAng(double rex, int lev, cgh *GH, var *chi, var
|
|||||||
pox[2][n] = rex * nz_g[n];
|
pox[2][n] = rex * nz_g[n];
|
||||||
}
|
}
|
||||||
|
|
||||||
|
double *shellf;
|
||||||
|
shellf = new double[n_tot * InList];
|
||||||
|
|
||||||
|
// we have assumed there is only one box on this level,
|
||||||
|
// so we do not need loop boxes
|
||||||
|
GH->PatL[lev]->data->Interp_Points(DG_List, n_tot, pox, shellf, Symmetry);
|
||||||
|
|
||||||
|
double Mass_out = 0;
|
||||||
|
double ang_outx, ang_outy, ang_outz;
|
||||||
|
double p_outx, p_outy, p_outz;
|
||||||
|
ang_outx = ang_outy = ang_outz = 0.0;
|
||||||
|
p_outx = p_outy = p_outz = 0.0;
|
||||||
|
const double f1o8 = 0.125;
|
||||||
|
|
||||||
int mp, Lp, Nmin, Nmax;
|
int mp, Lp, Nmin, Nmax;
|
||||||
|
|
||||||
mp = n_tot / cpusize;
|
mp = n_tot / cpusize;
|
||||||
Lp = n_tot - cpusize * mp;
|
Lp = n_tot - cpusize * mp;
|
||||||
|
|
||||||
if (Lp > myrank)
|
if (Lp > myrank)
|
||||||
{
|
{
|
||||||
Nmin = myrank * mp + myrank;
|
Nmin = myrank * mp + myrank;
|
||||||
@@ -2402,20 +2344,6 @@ void surface_integral::surf_MassPAng(double rex, int lev, cgh *GH, var *chi, var
|
|||||||
Nmax = Nmin + mp - 1;
|
Nmax = Nmin + mp - 1;
|
||||||
}
|
}
|
||||||
|
|
||||||
double *shellf;
|
|
||||||
shellf = new double[n_tot * InList];
|
|
||||||
|
|
||||||
// we have assumed there is only one box on this level,
|
|
||||||
// so we do not need loop boxes
|
|
||||||
GH->PatL[lev]->data->Interp_Points(DG_List, n_tot, pox, shellf, Symmetry, Nmin, Nmax);
|
|
||||||
|
|
||||||
double Mass_out = 0;
|
|
||||||
double ang_outx, ang_outy, ang_outz;
|
|
||||||
double p_outx, p_outy, p_outz;
|
|
||||||
ang_outx = ang_outy = ang_outz = 0.0;
|
|
||||||
p_outx = p_outy = p_outz = 0.0;
|
|
||||||
const double f1o8 = 0.125;
|
|
||||||
|
|
||||||
double Chi, Psi;
|
double Chi, Psi;
|
||||||
double Gxx, Gxy, Gxz, Gyy, Gyz, Gzz;
|
double Gxx, Gxy, Gxz, Gyy, Gyz, Gzz;
|
||||||
double gupxx, gupxy, gupxz, gupyy, gupyz, gupzz;
|
double gupxx, gupxy, gupxz, gupyy, gupyz, gupzz;
|
||||||
@@ -2536,13 +2464,15 @@ void surface_integral::surf_MassPAng(double rex, int lev, cgh *GH, var *chi, var
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
{
|
MPI_Allreduce(&Mass_out, &mass, 1, MPI_DOUBLE, MPI_SUM, MPI_COMM_WORLD);
|
||||||
double scalar_out[7] = {Mass_out, ang_outx, ang_outy, ang_outz, p_outx, p_outy, p_outz};
|
|
||||||
double scalar_in[7];
|
MPI_Allreduce(&ang_outx, &sx, 1, MPI_DOUBLE, MPI_SUM, MPI_COMM_WORLD);
|
||||||
MPI_Allreduce(scalar_out, scalar_in, 7, MPI_DOUBLE, MPI_SUM, MPI_COMM_WORLD);
|
MPI_Allreduce(&ang_outy, &sy, 1, MPI_DOUBLE, MPI_SUM, MPI_COMM_WORLD);
|
||||||
mass = scalar_in[0]; sx = scalar_in[1]; sy = scalar_in[2]; sz = scalar_in[3];
|
MPI_Allreduce(&ang_outz, &sz, 1, MPI_DOUBLE, MPI_SUM, MPI_COMM_WORLD);
|
||||||
px = scalar_in[4]; py = scalar_in[5]; pz = scalar_in[6];
|
|
||||||
}
|
MPI_Allreduce(&p_outx, &px, 1, MPI_DOUBLE, MPI_SUM, MPI_COMM_WORLD);
|
||||||
|
MPI_Allreduce(&p_outy, &py, 1, MPI_DOUBLE, MPI_SUM, MPI_COMM_WORLD);
|
||||||
|
MPI_Allreduce(&p_outz, &pz, 1, MPI_DOUBLE, MPI_SUM, MPI_COMM_WORLD);
|
||||||
|
|
||||||
#ifdef GaussInt
|
#ifdef GaussInt
|
||||||
mass = mass * rex * rex * dphi * factor;
|
mass = mass * rex * rex * dphi * factor;
|
||||||
@@ -2805,13 +2735,15 @@ void surface_integral::surf_MassPAng(double rex, int lev, cgh *GH, var *chi, var
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
{
|
MPI_Allreduce(&Mass_out, &mass, 1, MPI_DOUBLE, MPI_SUM, Comm_here);
|
||||||
double scalar_out[7] = {Mass_out, ang_outx, ang_outy, ang_outz, p_outx, p_outy, p_outz};
|
|
||||||
double scalar_in[7];
|
MPI_Allreduce(&ang_outx, &sx, 1, MPI_DOUBLE, MPI_SUM, Comm_here);
|
||||||
MPI_Allreduce(scalar_out, scalar_in, 7, MPI_DOUBLE, MPI_SUM, Comm_here);
|
MPI_Allreduce(&ang_outy, &sy, 1, MPI_DOUBLE, MPI_SUM, Comm_here);
|
||||||
mass = scalar_in[0]; sx = scalar_in[1]; sy = scalar_in[2]; sz = scalar_in[3];
|
MPI_Allreduce(&ang_outz, &sz, 1, MPI_DOUBLE, MPI_SUM, Comm_here);
|
||||||
px = scalar_in[4]; py = scalar_in[5]; pz = scalar_in[6];
|
|
||||||
}
|
MPI_Allreduce(&p_outx, &px, 1, MPI_DOUBLE, MPI_SUM, Comm_here);
|
||||||
|
MPI_Allreduce(&p_outy, &py, 1, MPI_DOUBLE, MPI_SUM, Comm_here);
|
||||||
|
MPI_Allreduce(&p_outz, &pz, 1, MPI_DOUBLE, MPI_SUM, Comm_here);
|
||||||
|
|
||||||
#ifdef GaussInt
|
#ifdef GaussInt
|
||||||
mass = mass * rex * rex * dphi * factor;
|
mass = mass * rex * rex * dphi * factor;
|
||||||
@@ -3088,13 +3020,15 @@ void surface_integral::surf_MassPAng(double rex, int lev, ShellPatch *GH, var *c
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
{
|
MPI_Allreduce(&Mass_out, &mass, 1, MPI_DOUBLE, MPI_SUM, MPI_COMM_WORLD);
|
||||||
double scalar_out[7] = {Mass_out, ang_outx, ang_outy, ang_outz, p_outx, p_outy, p_outz};
|
|
||||||
double scalar_in[7];
|
MPI_Allreduce(&ang_outx, &sx, 1, MPI_DOUBLE, MPI_SUM, MPI_COMM_WORLD);
|
||||||
MPI_Allreduce(scalar_out, scalar_in, 7, MPI_DOUBLE, MPI_SUM, MPI_COMM_WORLD);
|
MPI_Allreduce(&ang_outy, &sy, 1, MPI_DOUBLE, MPI_SUM, MPI_COMM_WORLD);
|
||||||
mass = scalar_in[0]; sx = scalar_in[1]; sy = scalar_in[2]; sz = scalar_in[3];
|
MPI_Allreduce(&ang_outz, &sz, 1, MPI_DOUBLE, MPI_SUM, MPI_COMM_WORLD);
|
||||||
px = scalar_in[4]; py = scalar_in[5]; pz = scalar_in[6];
|
|
||||||
}
|
MPI_Allreduce(&p_outx, &px, 1, MPI_DOUBLE, MPI_SUM, MPI_COMM_WORLD);
|
||||||
|
MPI_Allreduce(&p_outy, &py, 1, MPI_DOUBLE, MPI_SUM, MPI_COMM_WORLD);
|
||||||
|
MPI_Allreduce(&p_outz, &pz, 1, MPI_DOUBLE, MPI_SUM, MPI_COMM_WORLD);
|
||||||
|
|
||||||
#ifdef GaussInt
|
#ifdef GaussInt
|
||||||
mass = mass * rex * rex * dphi * factor;
|
mass = mass * rex * rex * dphi * factor;
|
||||||
@@ -3673,17 +3607,8 @@ void surface_integral::surf_Wave(double rex, cgh *GH, ShellPatch *SH,
|
|||||||
}
|
}
|
||||||
//|------+ Communicate and sum the results from each processor.
|
//|------+ Communicate and sum the results from each processor.
|
||||||
|
|
||||||
{
|
MPI_Allreduce(RP_out, RP, NN, MPI_DOUBLE, MPI_SUM, MPI_COMM_WORLD);
|
||||||
double *RPIP_out = new double[2 * NN];
|
MPI_Allreduce(IP_out, IP, NN, MPI_DOUBLE, MPI_SUM, MPI_COMM_WORLD);
|
||||||
double *RPIP = new double[2 * NN];
|
|
||||||
memcpy(RPIP_out, RP_out, NN * sizeof(double));
|
|
||||||
memcpy(RPIP_out + NN, IP_out, NN * sizeof(double));
|
|
||||||
MPI_Allreduce(RPIP_out, RPIP, 2 * NN, MPI_DOUBLE, MPI_SUM, MPI_COMM_WORLD);
|
|
||||||
memcpy(RP, RPIP, NN * sizeof(double));
|
|
||||||
memcpy(IP, RPIP + NN, NN * sizeof(double));
|
|
||||||
delete[] RPIP_out;
|
|
||||||
delete[] RPIP;
|
|
||||||
}
|
|
||||||
|
|
||||||
//|------= Free memory.
|
//|------= Free memory.
|
||||||
|
|
||||||
|
|||||||
@@ -18,11 +18,7 @@ using namespace std;
|
|||||||
#include <math.h>
|
#include <math.h>
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
#ifdef MPI_STUB
|
|
||||||
#include "mpi_stub.h"
|
|
||||||
#else
|
|
||||||
#include <mpi.h>
|
#include <mpi.h>
|
||||||
#endif
|
|
||||||
|
|
||||||
#include "misc.h"
|
#include "misc.h"
|
||||||
#include "macrodef.h"
|
#include "macrodef.h"
|
||||||
|
|||||||
@@ -20,11 +20,7 @@ using namespace std;
|
|||||||
#include <map.h>
|
#include <map.h>
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
#ifdef MPI_STUB
|
|
||||||
#include "mpi_stub.h"
|
|
||||||
#else
|
|
||||||
#include <mpi.h>
|
#include <mpi.h>
|
||||||
#endif
|
|
||||||
|
|
||||||
#include "misc.h"
|
#include "misc.h"
|
||||||
#include "macrodef.h"
|
#include "macrodef.h"
|
||||||
|
|||||||
@@ -9,11 +9,7 @@
|
|||||||
using namespace std;
|
using namespace std;
|
||||||
|
|
||||||
#include <time.h>
|
#include <time.h>
|
||||||
#ifdef MPI_STUB
|
|
||||||
#include "mpi_stub.h"
|
|
||||||
#else
|
|
||||||
#include <mpi.h>
|
#include <mpi.h>
|
||||||
#endif
|
|
||||||
|
|
||||||
#include "var.h"
|
#include "var.h"
|
||||||
|
|
||||||
|
|||||||
@@ -10,31 +10,6 @@
|
|||||||
|
|
||||||
import AMSS_NCKU_Input as input_data
|
import AMSS_NCKU_Input as input_data
|
||||||
import subprocess
|
import subprocess
|
||||||
import time
|
|
||||||
import os
|
|
||||||
|
|
||||||
## OpenMP configuration for threaded Fortran kernels
|
|
||||||
## OMP_NUM_THREADS: set to number of physical cores (not hyperthreads)
|
|
||||||
## OMP_PROC_BIND: bind threads to cores to avoid migration overhead
|
|
||||||
## OMP_STACKSIZE: each thread needs stack space for fh arrays (~3.6MB)
|
|
||||||
if "OMP_NUM_THREADS" not in os.environ:
|
|
||||||
os.environ["OMP_NUM_THREADS"] = "96"
|
|
||||||
os.environ["OMP_STACKSIZE"] = "16M"
|
|
||||||
os.environ["OMP_PROC_BIND"] = "close"
|
|
||||||
os.environ["OMP_PLACES"] = "cores"
|
|
||||||
## CPU core binding configuration using taskset
|
|
||||||
## taskset ensures all child processes inherit the CPU affinity mask
|
|
||||||
## This forces make and all compiler processes to use only nohz_full cores (4-55, 60-111)
|
|
||||||
## Format: taskset -c 4-55,60-111 ensures processes only run on these cores
|
|
||||||
#NUMACTL_CPU_BIND = "taskset -c 0-111"
|
|
||||||
#NUMACTL_CPU_BIND = "taskset -c 16-47,64-95"
|
|
||||||
#NUMACTL_CPU_BIND = "taskset -c 8-15"
|
|
||||||
NUMACTL_CPU_BIND = ""
|
|
||||||
|
|
||||||
## Build parallelism configuration
|
|
||||||
## Use nohz_full cores (4-55, 60-111) for compilation: 52 + 52 = 104 cores
|
|
||||||
## Set make -j to utilize available cores for faster builds
|
|
||||||
BUILD_JOBS = 16
|
|
||||||
|
|
||||||
|
|
||||||
##################################################################
|
##################################################################
|
||||||
@@ -51,11 +26,11 @@ def makefile_ABE():
|
|||||||
print( " Compiling the AMSS-NCKU executable file ABE/ABEGPU " )
|
print( " Compiling the AMSS-NCKU executable file ABE/ABEGPU " )
|
||||||
print( )
|
print( )
|
||||||
|
|
||||||
## Build command with CPU binding to nohz_full cores
|
## Build command
|
||||||
if (input_data.GPU_Calculation == "no"):
|
if (input_data.GPU_Calculation == "no"):
|
||||||
makefile_command = f"{NUMACTL_CPU_BIND} make -j{BUILD_JOBS} ABE"
|
makefile_command = "make -j4" + " ABE"
|
||||||
elif (input_data.GPU_Calculation == "yes"):
|
elif (input_data.GPU_Calculation == "yes"):
|
||||||
makefile_command = f"{NUMACTL_CPU_BIND} make -j{BUILD_JOBS} ABEGPU"
|
makefile_command = "make -j4" + " ABEGPU"
|
||||||
else:
|
else:
|
||||||
print( " CPU/GPU numerical calculation setting is wrong " )
|
print( " CPU/GPU numerical calculation setting is wrong " )
|
||||||
print( )
|
print( )
|
||||||
@@ -92,8 +67,8 @@ def makefile_TwoPunctureABE():
|
|||||||
print( " Compiling the AMSS-NCKU executable file TwoPunctureABE " )
|
print( " Compiling the AMSS-NCKU executable file TwoPunctureABE " )
|
||||||
print( )
|
print( )
|
||||||
|
|
||||||
## Build command with CPU binding to nohz_full cores
|
## Build command
|
||||||
makefile_command = f"{NUMACTL_CPU_BIND} make -j{BUILD_JOBS} TwoPunctureABE"
|
makefile_command = "make" + " TwoPunctureABE"
|
||||||
|
|
||||||
## Execute the command with subprocess.Popen and stream output
|
## Execute the command with subprocess.Popen and stream output
|
||||||
makefile_process = subprocess.Popen(makefile_command, shell=True, stdout=subprocess.PIPE, stderr=subprocess.STDOUT, text=True)
|
makefile_process = subprocess.Popen(makefile_command, shell=True, stdout=subprocess.PIPE, stderr=subprocess.STDOUT, text=True)
|
||||||
@@ -130,26 +105,26 @@ def run_ABE():
|
|||||||
## Define the command to run; cast other values to strings as needed
|
## Define the command to run; cast other values to strings as needed
|
||||||
|
|
||||||
if (input_data.GPU_Calculation == "no"):
|
if (input_data.GPU_Calculation == "no"):
|
||||||
run_command = NUMACTL_CPU_BIND + " ./ABE"
|
mpi_command = "mpirun -np " + str(input_data.MPI_processes) + " ./ABE"
|
||||||
run_command_outfile = "ABE_out.log"
|
mpi_command_outfile = "ABE_out.log"
|
||||||
elif (input_data.GPU_Calculation == "yes"):
|
elif (input_data.GPU_Calculation == "yes"):
|
||||||
run_command = NUMACTL_CPU_BIND + " ./ABEGPU"
|
mpi_command = "mpirun -np " + str(input_data.MPI_processes) + " ./ABEGPU"
|
||||||
run_command_outfile = "ABEGPU_out.log"
|
mpi_command_outfile = "ABEGPU_out.log"
|
||||||
|
|
||||||
## Execute the command and stream output
|
## Execute the MPI command and stream output
|
||||||
run_process = subprocess.Popen(run_command, shell=True, stdout=subprocess.PIPE, stderr=subprocess.STDOUT, text=True)
|
mpi_process = subprocess.Popen(mpi_command, shell=True, stdout=subprocess.PIPE, stderr=subprocess.STDOUT, text=True)
|
||||||
|
|
||||||
## Write ABE run output to file while printing to stdout
|
## Write ABE run output to file while printing to stdout
|
||||||
with open(run_command_outfile, 'w') as file0:
|
with open(mpi_command_outfile, 'w') as file0:
|
||||||
## Read and print output lines; also write each line to file
|
## Read and print output lines; also write each line to file
|
||||||
for line in run_process.stdout:
|
for line in mpi_process.stdout:
|
||||||
print(line, end='') # stream output in real time
|
print(line, end='') # stream output in real time
|
||||||
file0.write(line) # write the line to file
|
file0.write(line) # write the line to file
|
||||||
file0.flush() # flush to ensure each line is written immediately (optional)
|
file0.flush() # flush to ensure each line is written immediately (optional)
|
||||||
file0.close()
|
file0.close()
|
||||||
|
|
||||||
## Wait for the process to finish
|
## Wait for the process to finish
|
||||||
run_return_code = run_process.wait()
|
mpi_return_code = mpi_process.wait()
|
||||||
|
|
||||||
print( )
|
print( )
|
||||||
print( " The ABE/ABEGPU simulation is finished " )
|
print( " The ABE/ABEGPU simulation is finished " )
|
||||||
@@ -166,13 +141,12 @@ def run_ABE():
|
|||||||
## Run the AMSS-NCKU TwoPuncture program TwoPunctureABE
|
## Run the AMSS-NCKU TwoPuncture program TwoPunctureABE
|
||||||
|
|
||||||
def run_TwoPunctureABE():
|
def run_TwoPunctureABE():
|
||||||
tp_time1=time.time()
|
|
||||||
print( )
|
print( )
|
||||||
print( " Running the AMSS-NCKU executable file TwoPunctureABE " )
|
print( " Running the AMSS-NCKU executable file TwoPunctureABE " )
|
||||||
print( )
|
print( )
|
||||||
|
|
||||||
## Define the command to run
|
## Define the command to run
|
||||||
#TwoPuncture_command = NUMACTL_CPU_BIND + " ./TwoPunctureABE"
|
|
||||||
TwoPuncture_command = "./TwoPunctureABE"
|
TwoPuncture_command = "./TwoPunctureABE"
|
||||||
TwoPuncture_command_outfile = "TwoPunctureABE_out.log"
|
TwoPuncture_command_outfile = "TwoPunctureABE_out.log"
|
||||||
|
|
||||||
@@ -194,9 +168,7 @@ def run_TwoPunctureABE():
|
|||||||
print( )
|
print( )
|
||||||
print( " The TwoPunctureABE simulation is finished " )
|
print( " The TwoPunctureABE simulation is finished " )
|
||||||
print( )
|
print( )
|
||||||
tp_time2=time.time()
|
|
||||||
et=tp_time2-tp_time1
|
|
||||||
print(f"Used time: {et}")
|
|
||||||
return
|
return
|
||||||
|
|
||||||
##################################################################
|
##################################################################
|
||||||
|
|||||||
@@ -1,97 +0,0 @@
|
|||||||
# AMSS-NCKU PGO Profile Analysis Report
|
|
||||||
|
|
||||||
## 1. Profiling Environment
|
|
||||||
|
|
||||||
| Item | Value |
|
|
||||||
|------|-------|
|
|
||||||
| Compiler | Intel oneAPI DPC++/C++ 2025.3.0 (icpx/ifx) |
|
|
||||||
| Instrumentation Flag | `-fprofile-instr-generate` |
|
|
||||||
| Optimization Level (instrumented) | `-O2 -xHost -fma` |
|
|
||||||
| MPI Processes | 1 (single process to avoid MPI+instrumentation deadlock) |
|
|
||||||
| Profile File | `default_9725750769337483397_0.profraw` (327 KB) |
|
|
||||||
| Merged Profile | `default.profdata` (394 KB) |
|
|
||||||
| llvm-profdata | `/home/intel/oneapi/compiler/2025.3/bin/compiler/llvm-profdata` |
|
|
||||||
|
|
||||||
## 2. Reduced Simulation Parameters (for profiling run)
|
|
||||||
|
|
||||||
| Parameter | Production Value | Profiling Value |
|
|
||||||
|-----------|-----------------|-----------------|
|
|
||||||
| MPI_processes | 64 | 1 |
|
|
||||||
| grid_level | 9 | 4 |
|
|
||||||
| static_grid_level | 5 | 3 |
|
|
||||||
| static_grid_number | 96 | 24 |
|
|
||||||
| moving_grid_number | 48 | 16 |
|
|
||||||
| largest_box_xyz_max | 320^3 | 160^3 |
|
|
||||||
| Final_Evolution_Time | 1000.0 | 10.0 |
|
|
||||||
| Evolution_Step_Number | 10,000,000 | 1,000 |
|
|
||||||
| Detector_Number | 12 | 2 |
|
|
||||||
|
|
||||||
## 3. Profile Summary
|
|
||||||
|
|
||||||
| Metric | Value |
|
|
||||||
|--------|-------|
|
|
||||||
| Total instrumented functions | 1,392 |
|
|
||||||
| Functions with non-zero counts | 117 (8.4%) |
|
|
||||||
| Functions with zero counts | 1,275 (91.6%) |
|
|
||||||
| Maximum function entry count | 386,459,248 |
|
|
||||||
| Maximum internal block count | 370,477,680 |
|
|
||||||
| Total block count | 4,198,023,118 |
|
|
||||||
|
|
||||||
## 4. Top 20 Hotspot Functions
|
|
||||||
|
|
||||||
| Rank | Total Count | Max Block Count | Function | Category |
|
|
||||||
|------|------------|-----------------|----------|----------|
|
|
||||||
| 1 | 1,241,601,732 | 370,477,680 | `polint_` | Interpolation |
|
|
||||||
| 2 | 755,994,435 | 230,156,640 | `prolong3_` | Grid prolongation |
|
|
||||||
| 3 | 667,964,095 | 3,697,792 | `compute_rhs_bssn_` | BSSN RHS evolution |
|
|
||||||
| 4 | 539,736,051 | 386,459,248 | `symmetry_bd_` | Symmetry boundary |
|
|
||||||
| 5 | 277,310,808 | 53,170,728 | `lopsided_` | Lopsided FD stencil |
|
|
||||||
| 6 | 155,534,488 | 94,535,040 | `decide3d_` | 3D grid decision |
|
|
||||||
| 7 | 119,267,712 | 19,266,048 | `rungekutta4_rout_` | RK4 time integrator |
|
|
||||||
| 8 | 91,574,616 | 48,824,160 | `kodis_` | Kreiss-Oliger dissipation |
|
|
||||||
| 9 | 67,555,389 | 43,243,680 | `fderivs_` | Finite differences |
|
|
||||||
| 10 | 55,296,000 | 42,246,144 | `misc::fact(int)` | Factorial utility |
|
|
||||||
| 11 | 43,191,071 | 27,663,328 | `fdderivs_` | 2nd-order FD derivatives |
|
|
||||||
| 12 | 36,233,965 | 22,429,440 | `restrict3_` | Grid restriction |
|
|
||||||
| 13 | 24,698,512 | 17,231,520 | `polin3_` | Polynomial interpolation |
|
|
||||||
| 14 | 22,962,942 | 20,968,768 | `copy_` | Data copy |
|
|
||||||
| 15 | 20,135,696 | 17,259,168 | `Ansorg::barycentric(...)` | Spectral interpolation |
|
|
||||||
| 16 | 14,650,224 | 7,224,768 | `Ansorg::barycentric_omega(...)` | Spectral weights |
|
|
||||||
| 17 | 13,242,296 | 2,871,920 | `global_interp_` | Global interpolation |
|
|
||||||
| 18 | 12,672,000 | 7,734,528 | `sommerfeld_rout_` | Sommerfeld boundary |
|
|
||||||
| 19 | 6,872,832 | 1,880,064 | `sommerfeld_routbam_` | Sommerfeld boundary (BAM) |
|
|
||||||
| 20 | 5,709,900 | 2,809,632 | `l2normhelper_` | L2 norm computation |
|
|
||||||
|
|
||||||
## 5. Hotspot Category Breakdown
|
|
||||||
|
|
||||||
Top 20 functions account for ~98% of total execution counts:
|
|
||||||
|
|
||||||
| Category | Functions | Combined Count | Share |
|
|
||||||
|----------|-----------|---------------|-------|
|
|
||||||
| Interpolation / Prolongation / Restriction | polint_, prolong3_, restrict3_, polin3_, global_interp_, Ansorg::* | ~2,093M | ~50% |
|
|
||||||
| BSSN RHS + FD stencils | compute_rhs_bssn_, lopsided_, fderivs_, fdderivs_ | ~1,056M | ~25% |
|
|
||||||
| Boundary conditions | symmetry_bd_, sommerfeld_rout_, sommerfeld_routbam_ | ~559M | ~13% |
|
|
||||||
| Time integration | rungekutta4_rout_ | ~119M | ~3% |
|
|
||||||
| Dissipation | kodis_ | ~92M | ~2% |
|
|
||||||
| Utilities | misc::fact, decide3d_, copy_, l2normhelper_ | ~256M | ~6% |
|
|
||||||
|
|
||||||
## 6. Conclusions
|
|
||||||
|
|
||||||
1. **Profile data is valid**: 1,392 functions instrumented, 117 exercised with ~4.2 billion total counts.
|
|
||||||
2. **Hotspot concentration is high**: Top 5 functions alone account for ~76% of all counts, which is ideal for PGO — the compiler has strong branch/layout optimization targets.
|
|
||||||
3. **Fortran numerical kernels dominate**: `polint_`, `prolong3_`, `compute_rhs_bssn_`, `symmetry_bd_`, `lopsided_` are all Fortran routines in the inner evolution loop. PGO will optimize their branch prediction and basic block layout.
|
|
||||||
4. **91.6% of functions have zero counts**: These are code paths for unused features (GPU, BSSN-EScalar, BSSN-EM, Z4C, etc.). PGO will deprioritize them, improving instruction cache utilization.
|
|
||||||
5. **Profile is representative**: Despite the reduced grid size, the code path coverage matches production — the same kernels (RHS, prolongation, restriction, boundary) are exercised. PGO branch probabilities from this profile will transfer well to full-scale runs.
|
|
||||||
|
|
||||||
## 7. PGO Phase 2 Usage
|
|
||||||
|
|
||||||
To apply the profile, use the following flags in `makefile.inc`:
|
|
||||||
|
|
||||||
```makefile
|
|
||||||
CXXAPPFLAGS = -O3 -xHost -fp-model fast=2 -fma -ipo \
|
|
||||||
-fprofile-instr-use=/home/amss/AMSS-NCKU/pgo_profile/default.profdata \
|
|
||||||
-Dfortran3 -Dnewc -I${MKLROOT}/include
|
|
||||||
f90appflags = -O3 -xHost -fp-model fast=2 -fma -ipo \
|
|
||||||
-fprofile-instr-use=/home/amss/AMSS-NCKU/pgo_profile/default.profdata \
|
|
||||||
-align array64byte -fpp -I${MKLROOT}/include
|
|
||||||
```
|
|
||||||
Binary file not shown.
Binary file not shown.
Reference in New Issue
Block a user