[FLASH-USERS] Fatal error in PMPI_Waitall: See the MPI_ERROR field in MPI_Status for the error code

guido g.granda at irya.unam.mx
Tue Dec 10 13:34:11 EST 2019


Hello,

Thank you, but I'm confused.  You said that the more processors I use 
the more memory each processor will require, however from maxblocks ~ 
1.02*num_blocks_requesed/num_procs , max blocks is inverse proportional

to the number of processors and since the memory each processor needs is 
NUNK_VARSxNXBxNYBxNZBxMAXBLOCKS , then the memory is inverse 
proportional to the num_procs , right?

Another question is if NUNK_VARS is the number of plot_var written into 
the plot files.

Cheers,

On 06/12/19 22:53, Ryan Farber wrote:
> Hi Guido,
>
> I've seen FLASH crash with similar stdout due to insufficient memory 
> many times. Keep in mind that the more processors you use, the more 
> memory the run requires (since each processor has its own solution 
> data array of NUNK_VARSxNXBxNYBxNZBxMAXBLOCKS).
>
> I was going to suggest reducing maxblocks since 3000 sounds to me 
> rather extreme. My simulations are happiest with MAXBLOCKS=200 (the 
> default value).
>
> Assuming static mesh refinement, you only need maxblocks to be ~1.02 * 
> num_blks_requested / num_procs (in your case, 1.02*97801/procs).
>
> So, if you want to use more nodes, decrease maxblocks accordingly.
>
> Best,
> --------
> Ryan
>
>
> On Fri, Dec 6, 2019 at 8:42 PM guido <g.granda at irya.unam.mx 
> <mailto:g.granda at irya.unam.mx>> wrote:
>
>     Hello,
>     Thank you for replaying. I tried your suggestion and I still got
>     the same errors. However, I tried increasing the number of nodes
>     from 8 to 16 and the simulations runs. So, I guess the crash was
>     due to lack of resources such as ram memory or sth similar. How
>     can I check that it was the cause of the crash ?
>
>     Cheers,
>
>     On 12/6/19 6:13 PM, Eddie Hansen wrote:
>>     I haven't seen this error before, so I don't know if this is the
>>     solution, but here's something I noticed...
>>
>>     You mentioned that you have loaded a version of hdf5 with
>>     parallel mode enabled, yet the IO unit that is loaded during
>>     setup is IO/IOMain/hdf5/serial. Try using the setup
>>     shortcut +hdf5typeio. This will load, by default, the parallel IO
>>     unit. Resetup, recompile, rerun, and see if you get the same error.
>>
>>
>>     On Fri, Dec 6, 2019 at 5:46 PM guido <g.granda at irya.unam.mx
>>     <mailto:g.granda at irya.unam.mx>> wrote:
>>
>>         Hello Flash users,
>>
>>         I’m having trouble running a simulation with FLASH 4.5. It
>>         compiled FLASH 4.5 using:
>>
>>         - hdf5-1.8.20 with parallel mode enabled
>>
>>         - mpich-3.2.1
>>
>>         - gcc-7.3.0
>>
>>         The simulations run fine till I increase the maximum level of
>>         refinement and the number of nodes used. For example, for
>>         simulation with lmax=9, it runs slow but safe with 8 nodes.
>>         However, when I increase the number of nodes to 10, the
>>         simulation crash trying to write the Coefficients in the
>>         Taylor expansion into file.
>>
>>         The log file is the following:
>>
>>
>>         FLASH log file: 12-05-2019 12:01:48.640 Run number: 1
>>
>>         ==============================================================================
>>
>>         Number of MPI tasks: 1
>>
>>         MPI version: 3
>>
>>         MPI subversion: 1
>>
>>         Dimensionality: 3
>>
>>         Max Number of Blocks/Proc: 3000
>>
>>         Number x zones: 8
>>
>>         Number y zones: 8
>>
>>         Number z zones: 8
>>
>>         Setup stamp: Wed Sep 18 16:44:51 2019
>>
>>         Build stamp: Wed Sep 18 16:45:20 2019
>>
>>         System info: Linux mouruka.crya.privado
>>         2.6.32-504.16.2.el6.x86_64 #1 SMP Wed Apr 22 06:48:29
>>
>>         Version: FLASH 4.5_release
>>
>>         Build directory: /home/guido/FLASH4.5/obj_mc
>>
>>         Setup syntax: /home/guido/FLASH4.5/bin/setup.py
>>         magnetoHd/MC_evolution -auto -3d -maxblocks=3000
>>         -objdir=obj_mc -site=irya.guido
>>         --without-unit=physics/Hydro/HydroMain/split/MHD_8Wave
>>
>>         f compiler flags:
>>         /home/guido/libraries/compiled_with_gcc-7.3.0/mpich-3.2.1/bin/mpif90
>>         -g -c -O0 -fdefault-real-8 -fdefault-double-8
>>         -ffree-line-length-none -Wuninitialized -g -c
>>         -fdefault-real-8 -fdefault-double-8 -ffree-line-length-none
>>         -Wuninitialized -DMAXBLOCKS=3000 -DNXB=8 -DNYB=8 -DNZB=8
>>         -DN_DIM=3
>>
>>         c compiler flags:
>>         /home/guido/libraries/compiled_with_gcc-7.3.0/mpich-3.2.1/bin/mpicc
>>         -I/home/guido/libraries/compiled_with_gcc-7.3.0/hdf5-1.8.20/include
>>         -DH5_USE_16_API -O0 -c -DMAXBLOCKS=3000 -DNXB=8 -DNYB=8
>>         -DNZB=8 -DN_DIM=3
>>
>>         ==============================================================================
>>
>>         Comment: MC evolution
>>
>>         ==============================================================================
>>
>>         FLASH Units used:
>>
>>         Driver/DriverMain/Split
>>
>>         Driver/localAPI
>>
>>         Grid/GridBoundaryConditions
>>
>>         Grid/GridMain/paramesh/interpolation/Paramesh4/prolong
>>
>>         Grid/GridMain/paramesh/interpolation/prolong
>>
>>         Grid/GridMain/paramesh/paramesh4/Paramesh4dev/PM4_package/headers
>>
>>         Grid/GridMain/paramesh/paramesh4/Paramesh4dev/PM4_package/mpi_source
>>
>>         Grid/GridMain/paramesh/paramesh4/Paramesh4dev/PM4_package/source
>>
>>         Grid/GridMain/paramesh/paramesh4/Paramesh4dev/PM4_package/source
>>
>>         Grid/GridMain/paramesh/paramesh4/Paramesh4dev/PM4_package/utilities/multigrid
>>
>>         Grid/GridMain/paramesh/paramesh4/Paramesh4dev/flash_avoid_orrery
>>
>>         Grid/GridSolvers/BHTree/Wunsch
>>
>>         Grid/localAPI
>>
>>         IO/IOMain/hdf5/serial/PM
>>
>>         IO/localAPI
>>
>>         PhysicalConstants/PhysicalConstantsMain
>>
>>         RuntimeParameters/RuntimeParametersMain
>>
>>         Simulation/SimulationMain/magnetoHD/MC_evolution
>>
>>         flashUtilities/contiguousConversion
>>
>>         flashUtilities/general
>>
>>         flashUtilities/interpolation/oneDim
>>
>>         flashUtilities/nameValueLL
>>
>>         flashUtilities/sorting/quicksort
>>
>>         flashUtilities/system/memoryUsage/legacy
>>
>>         monitors/Logfile/LogfileMain
>>
>>         monitors/Timers/TimersMain/MPINative
>>
>>         physics/Eos/EosMain/Gamma
>>
>>         physics/Eos/localAPI
>>
>>         physics/Gravity/GravityMain/Poisson/BHTree
>>
>>         physics/Hydro/HydroMain/split/Bouchut/Bouchut3
>>
>>         physics/Hydro/localAPI
>>
>>         physics/sourceTerms/Polytrope/PolytropeMain
>>
>>         ==============================================================================
>>
>>         RuntimeParameters:
>>
>>         ==============================================================================
>>
>>         bndpriorityone = 1
>>
>>         bndprioritythree = 3
>>
>>         bndprioritytwo = 2
>>
>>         checkpointfileintervalstep = 1000000 [CHANGED]
>>
>>         checkpointfilenumber = 0
>>
>>         dr_abortpause = 2
>>
>>         dr_dtminbelowaction = 1
>>
>>         dr_numposdefvars = 4
>>
>>         drift_break_inst = 0
>>
>>         drift_trunc_mantissa = 2
>>
>>         drift_verbose_inst = 0
>>
>>         eos_entrelescalechoice = 6
>>
>>         eos_loglevel = 700
>>
>>         fileformatversion = 9
>>
>>         forcedplotfilenumber = 0
>>
>>         gr_bhtwmaxqueuesize = 10000
>>
>>         gr_lrefmaxtimevalue_1 = -1
>>
>>         gr_lrefmaxtimevalue_10 = -1
>>
>>         gr_lrefmaxtimevalue_11 = -1
>>
>>         gr_lrefmaxtimevalue_12 = -1
>>
>>         gr_lrefmaxtimevalue_13 = -1
>>
>>         gr_lrefmaxtimevalue_14 = -1
>>
>>         gr_lrefmaxtimevalue_15 = -1
>>
>>         gr_lrefmaxtimevalue_16 = -1
>>
>>         gr_lrefmaxtimevalue_17 = -1
>>
>>         gr_lrefmaxtimevalue_18 = -1
>>
>>         gr_lrefmaxtimevalue_19 = -1
>>
>>         gr_lrefmaxtimevalue_2 = -1
>>
>>         gr_lrefmaxtimevalue_20 = -1
>>
>>         gr_lrefmaxtimevalue_3 = -1
>>
>>         gr_lrefmaxtimevalue_4 = -1
>>
>>         gr_lrefmaxtimevalue_5 = -1
>>
>>         gr_lrefmaxtimevalue_6 = -1
>>
>>         gr_lrefmaxtimevalue_7 = -1
>>
>>         gr_lrefmaxtimevalue_8 = -1
>>
>>         gr_lrefmaxtimevalue_9 = -1
>>
>>         gr_pmrpdivergencefree = 1
>>
>>         gr_pmrpifaceoff = 0
>>
>>         gr_pmrpl2p5d = 0
>>
>>         gr_pmrpmaxblocks = -1
>>
>>         gr_pmrpmflags = 1
>>
>>         gr_pmrpnboundaries = 6
>>
>>         gr_pmrpndim = 3
>>
>>         gr_pmrpnedgevar1 = -1
>>
>>         gr_pmrpnfacevar = -1
>>
>>         gr_pmrpnfielddivf = -1
>>
>>         gr_pmrpnfluxvar = -1
>>
>>         gr_pmrpnguard = -1
>>
>>         gr_pmrpnguardwork = -1
>>
>>         gr_pmrpnvar = -1
>>
>>         gr_pmrpnvarwork = 1
>>
>>         gr_pmrpnvarcorn = 0
>>
>>         gr_pmrpnvaredge = 0
>>
>>         gr_pmrpnxb = -1
>>
>>         gr_pmrpnyb = -1
>>
>>         gr_pmrpnzb = -1
>>
>>         gr_restrictallmethod = 3
>>
>>         gr_sanitizedatamode = 3 [CHANGED]
>>
>>         gr_sanitizeverbosity = 5
>>
>>         grv_bhewaldfieldnxv42 = 32
>>
>>         grv_bhewaldfieldnyv42 = 32
>>
>>         grv_bhewaldfieldnzv42 = 32
>>
>>         grv_bhewaldnper = 32
>>
>>         grv_bhewaldnrefv42 = -1
>>
>>         grv_bhewaldseriesn = 10
>>
>>         grv_bhmpdegree = 2
>>
>>         iprocs = 1
>>
>>         interpol_order = 2
>>
>>         irenorm = 0
>>
>>         jprocs = 1
>>
>>         kprocs = 1
>>
>>         lrefine_del = 0
>>
>>         lrefine_max = 9 [CHANGED]
>>
>>         lrefine_min = 3 [CHANGED]
>>
>>         lrefine_min_init = 1
>>
>>         max_particles_per_blk = 100
>>
>>         memory_stat_freq = 100000
>>
>>         meshcopycount = 1
>>
>>         min_particles_per_blk = 1
>>
>>         nbegin = 1
>>
>>         nblockx = 1
>>
>>         nblocky = 1
>>
>>         nblockz = 1
>>
>>         nend = 1000000 [CHANGED]
>>
>>         nrefs = 2
>>
>>         nsteptotalsts = 5
>>
>>         outputsplitnum = 1
>>
>>         plotfileintervalstep = 1000000 [CHANGED]
>>
>>         plotfilenumber = 0
>>
>>         refine_var_count = 4
>>
>>         rolling_checkpoint = 10000
>>
>>         sim_refine = 9 [CHANGED]
>>
>>         sweeporder = 123
>>
>>         wr_integrals_freq = 1
>>
>>         cfl = 0.800E+00
>>
>>         checkpointfileintervaltime = 0.316E+14 [CHANGED]
>>
>>         checkpointfileintervalz = 0.180+309
>>
>>         derefine_cutoff_1 = 0.300E+00 [CHANGED]
>>
>>         derefine_cutoff_2 = 0.200E+00
>>
>>         derefine_cutoff_3 = 0.200E+00
>>
>>         derefine_cutoff_4 = 0.200E+00
>>
>>         dr_dtmincontinue = 0.000E+00
>>
>>         dr_posdefdtfactor = 0.100E+01
>>
>>         dr_tstepslowstartfactor = 0.100E+00
>>
>>         dtinit = 0.316E+11 [CHANGED]
>>
>>         dtmax = 0.316E+13 [CHANGED]
>>
>>         dtmin = 0.316E+10 [CHANGED]
>>
>>         eintswitch = 0.100E-03 [CHANGED]
>>
>>         eos_singlespeciesa = 0.100E+01
>>
>>         eos_singlespeciesz = 0.100E+01
>>
>>         gamma = 0.167E+01
>>
>>         gr_bhtreelimangle = 0.500E+00
>>
>>         gr_bhtreemaxcellmass = 0.100+100
>>
>>         gr_bhtreemincellmass = 0.100E-98
>>
>>         gr_bhtreesafebox = 0.120E+01
>>
>>         gr_lrefinemaxredlogbase = 0.100E+02
>>
>>         gr_lrefinemaxredradiusfact = 0.000E+00
>>
>>         gr_lrefinemaxredtref = 0.000E+00
>>
>>         gr_lrefinemaxredtimescale = 0.100E+01
>>
>>         gr_lrefmaxtime_1 = -0.100E+01
>>
>>         gr_lrefmaxtime_10 = -0.100E+01
>>
>>         gr_lrefmaxtime_11 = -0.100E+01
>>
>>         gr_lrefmaxtime_12 = -0.100E+01
>>
>>         gr_lrefmaxtime_13 = -0.100E+01
>>
>>         gr_lrefmaxtime_14 = -0.100E+01
>>
>>         gr_lrefmaxtime_15 = -0.100E+01
>>
>>         gr_lrefmaxtime_16 = -0.100E+01
>>
>>         gr_lrefmaxtime_17 = -0.100E+01
>>
>>         gr_lrefmaxtime_18 = -0.100E+01
>>
>>         gr_lrefmaxtime_19 = -0.100E+01
>>
>>         gr_lrefmaxtime_2 = -0.100E+01
>>
>>         gr_lrefmaxtime_20 = -0.100E+01
>>
>>         gr_lrefmaxtime_3 = -0.100E+01
>>
>>         gr_lrefmaxtime_4 = -0.100E+01
>>
>>         gr_lrefmaxtime_5 = -0.100E+01
>>
>>         gr_lrefmaxtime_6 = -0.100E+01
>>
>>         gr_lrefmaxtime_7 = -0.100E+01
>>
>>         gr_lrefmaxtime_8 = -0.100E+01
>>
>>         gr_lrefmaxtime_9 = -0.100E+01
>>
>>         grv_bhaccerr = 0.100E+00
>>
>>         grv_bhextrnpotcenterx = 0.000E+00
>>
>>         grv_bhextrnpotcentery = 0.000E+00
>>
>>         grv_bhextrnpotcenterz = 0.000E+00
>>
>>         grv_bhnewton = -0.100E+01
>>
>>         hall_parameter = 0.000E+00
>>
>>         hyperresistivity = 0.000E+00
>>
>>         nusts = 0.100E+00
>>
>>         plotfileintervaltime = 0.316E+13 [CHANGED]
>>
>>         plotfileintervalz = 0.180+309
>>
>>         point_mass = 0.000E+00
>>
>>         point_mass_rsoft = 0.000E+00
>>
>>         polytropedens1 = 0.212E-24 [CHANGED]
>>
>>         polytropedens2 = 0.100+100
>>
>>         polytropedens3 = 0.100+100
>>
>>         polytropedens4 = 0.100+100
>>
>>         polytropedens5 = 0.100+100
>>
>>         polytropegamma1 = 0.100E+01
>>
>>         polytropegamma2 = 0.100E+01
>>
>>         polytropegamma3 = 0.100E+01
>>
>>         polytropegamma4 = 0.100E+01
>>
>>         polytropegamma5 = 0.100E+01
>>
>>         polytropekonst = 0.400E+09 [CHANGED]
>>
>>         refine_cutoff_1 = 0.700E+00 [CHANGED]
>>
>>         refine_cutoff_2 = 0.800E+00
>>
>>         refine_cutoff_3 = 0.800E+00
>>
>>         refine_cutoff_4 = 0.800E+00
>>
>>         refine_filter_1 = 0.100E-01
>>
>>         refine_filter_2 = 0.100E-01
>>
>>         refine_filter_3 = 0.100E-01
>>
>>         refine_filter_4 = 0.100E-01
>>
>>         rss_limit = -0.100E+01
>>
>>         sim_bx = 0.710E-05 [CHANGED]
>>
>>         sim_by = 0.000E+00
>>
>>         sim_bz = 0.000E+00
>>
>>         sim_cs = 0.200E+05 [CHANGED]
>>
>>         sim_delta = 0.463E+19 [CHANGED]
>>
>>         sim_mu_mol = 0.127E+01
>>
>>         sim_pert = 0.150E+01 [CHANGED]
>>
>>         sim_rho = 0.212E-21 [CHANGED]
>>
>>         sim_sigma = 0.250E+00 [CHANGED]
>>
>>         sim_vx = 0.000E+00
>>
>>         sim_vy = 0.000E+00
>>
>>         sim_vz = 0.000E+00
>>
>>         sim_x_cent = 0.000E+00
>>
>>         sim_y_cent = 0.000E+00
>>
>>         sim_z_cent = 0.000E+00
>>
>>         small = 0.100E-39 [CHANGED]
>>
>>         smalle = 0.100E-09
>>
>>         smallp = 0.100E-21 [CHANGED]
>>
>>         smallt = 0.100E+01 [CHANGED]
>>
>>         smallu = 0.100E-39 [CHANGED]
>>
>>         smallx = 0.100E-09
>>
>>         smlrho = 0.100E-39 [CHANGED]
>>
>>         tinitial = 0.000E+00
>>
>>         tmax = 0.252E+15 [CHANGED]
>>
>>         tstep_change_factor = 0.200E+01
>>
>>         wall_clock_checkpoint = 0.432E+05
>>
>>         wall_clock_time_limit = 0.605E+06
>>
>>         x_refine_center = 0.000E+00
>>
>>         xmax = 0.154E+20 [CHANGED]
>>
>>         xmin = -0.154E+20 [CHANGED]
>>
>>         y_refine_center = 0.000E+00
>>
>>         ymax = 0.154E+20 [CHANGED]
>>
>>         ymin = -0.154E+20 [CHANGED]
>>
>>         zfinal = 0.000E+00
>>
>>         zinitial = -0.100E+01
>>
>>         z_refine_center = 0.000E+00
>>
>>         zmax = 0.154E+20 [CHANGED]
>>
>>         zmin = -0.154E+20 [CHANGED]
>>
>>         unitsystem = CGS [CHANGED]
>>
>>         basenm = mc_evolution_ [CHANGED]
>>
>>         dr_posdefvar_1 = none
>>
>>         dr_posdefvar_2 = none
>>
>>         dr_posdefvar_3 = none
>>
>>         dr_posdefvar_4 = none
>>
>>         eosmode = dens_ie
>>
>>         eosmodeinit = dens_ie
>>
>>         geometry = cartesian
>>
>>         gr_pmrpoutputdir = ./
>>
>>         grav_boundary_type = periodic [CHANGED]
>>
>>         grav_boundary_type_x = periodic [CHANGED]
>>
>>         grav_boundary_type_y = periodic [CHANGED]
>>
>>         grav_boundary_type_z = periodic [CHANGED]
>>
>>         grv_bhewaldfname = ewald_coeffs
>>
>>         grv_bhewaldfnameaccv42 = ewald_field_acc
>>
>>         grv_bhewaldfnamepotv42 = ewald_field_pot
>>
>>         grv_bhextrnpotfile = external_potential.dat
>>
>>         grv_bhextrnpottype = planez
>>
>>         grv_bhmac = ApproxPartialErr
>>
>>         log_file = mc_evolution.log [CHANGED]
>>
>>         output_directory =
>>
>>         pc_unitsbase = CGS
>>
>>         plot_grid_var_1 = none
>>
>>         plot_grid_var_10 = none
>>
>>         plot_grid_var_11 = none
>>
>>         plot_grid_var_12 = none
>>
>>         plot_grid_var_2 = none
>>
>>         plot_grid_var_3 = none
>>
>>         plot_grid_var_4 = none
>>
>>         plot_grid_var_5 = none
>>
>>         plot_grid_var_6 = none
>>
>>         plot_grid_var_7 = none
>>
>>         plot_grid_var_8 = none
>>
>>         plot_grid_var_9 = none
>>
>>         plot_var_1 = dens [CHANGED]
>>
>>         plot_var_10 = magz [CHANGED]
>>
>>         plot_var_11 = none
>>
>>         plot_var_12 = none
>>
>>         plot_var_13 = none
>>
>>         plot_var_14 = none
>>
>>         plot_var_15 = none
>>
>>         plot_var_16 = none
>>
>>         plot_var_17 = none
>>
>>         plot_var_18 = none
>>
>>         plot_var_2 = pres [CHANGED]
>>
>>         plot_var_3 = none
>>
>>         plot_var_4 = eint [CHANGED]
>>
>>         plot_var_5 = velx [CHANGED]
>>
>>         plot_var_6 = vely [CHANGED]
>>
>>         plot_var_7 = velz [CHANGED]
>>
>>         plot_var_8 = magx [CHANGED]
>>
>>         plot_var_9 = magy [CHANGED]
>>
>>         prof_file = profile.dat
>>
>>         refine_var_1 = dens [CHANGED]
>>
>>         refine_var_2 = none
>>
>>         refine_var_3 = none
>>
>>         refine_var_4 = none
>>
>>         refine_var_thresh = dens [CHANGED]
>>
>>         run_comment = MC evolution [CHANGED]
>>
>>         run_number = 1
>>
>>         stats_file = mc.dat [CHANGED]
>>
>>         xl_boundary_type = periodic
>>
>>         xr_boundary_type = periodic
>>
>>         yl_boundary_type = periodic
>>
>>         yr_boundary_type = periodic
>>
>>         zl_boundary_type = periodic
>>
>>         zr_boundary_type = periodic
>>
>>         allowdtstsdominate = F
>>
>>         alwayscomputeuservars = T
>>
>>         alwaysrestrictcheckpoint = T
>>
>>         chkguardcellsinput = F
>>
>>         chkguardcellsoutput = F
>>
>>         converttoconsvdformeshcalls = F
>>
>>         converttoconsvdinmeshinterp = T
>>
>>         corners = F
>>
>>         dr_printtsteploc = T
>>
>>         dr_shortenlaststepbeforetmax = F
>>
>>         dr_useposdefcomputedt = F
>>
>>         drift_tuples = F
>>
>>         eachprocwritesownabortlog = F
>>
>>         eachprocwritessummary = F
>>
>>         earlyblockdistadjustment = T
>>
>>         enablemaskedgcfill = F
>>
>>         flux_correct = T
>>
>>         geometryoverride = F
>>
>>         gr_bcenableapplymixedgds = T
>>
>>         gr_bhphysmaccomm = F
>>
>>         gr_bhphysmactw = F
>>
>>         gr_bhuseunifiedtw = F [CHANGED]
>>
>>         gr_lrefinemaxbytime = F
>>
>>         gr_lrefinemaxreddobylogr = F
>>
>>         gr_lrefinemaxreddobytime = F
>>
>>         gr_pmrpadvancealllevels = F
>>
>>         gr_pmrpamrerrorchecking = F
>>
>>         gr_pmrpcartesianpm = F
>>
>>         gr_pmrpconserve = F
>>
>>         gr_pmrpconsvfluxdensities = T
>>
>>         gr_pmrpconsvfluxes = F
>>
>>         gr_pmrpcurvilinear = F
>>
>>         gr_pmrpcurvilinearconserve = F
>>
>>         gr_pmrpcylindricalpm = F
>>
>>         gr_pmrpdiagonals = T
>>
>>         gr_pmrpedgevalue = T
>>
>>         gr_pmrpedgevalueinteg = F
>>
>>         gr_pmrpemptycells = F
>>
>>         gr_pmrpforceconsistency = T
>>
>>         gr_pmrplsingularline = F
>>
>>         gr_pmrpnopermanentguardcells = F
>>
>>         gr_pmrppolarpm = F
>>
>>         gr_pmrppredcorr = F
>>
>>         gr_pmrpsphericalpm = F
>>
>>         gr_pmrptimingmpi = F
>>
>>         gr_pmrptimingmpix = F
>>
>>         gr_pmrpvardt = F
>>
>>         grav_temporal_extrp = F
>>
>>         grav_unjunkpden = T
>>
>>         grv_bhewaldalwaysgenerate = T
>>
>>         grv_bhlinearinterpolonlyv42 = T
>>
>>         grv_bhuserelaccerr = F
>>
>>         grv_useexternalpotential = F
>>
>>         grv_usepoissonpotential = T
>>
>>         ignoreforcedplot = F
>>
>>         io_writemscalarintegrals = F
>>
>>         killdivb = T [CHANGED]
>>
>>         plotfilegridquantitydp = F
>>
>>         plotfilemetadatadp = F
>>
>>         reducegcellfills = F
>>
>>         refine_on_particle_count = F
>>
>>         restart = F
>>
>>         summaryoutputonly = F
>>
>>         threadblocklistbuild = F
>>
>>         threaddriverblocklist = F
>>
>>         threaddriverwithinblock = F
>>
>>         threadeoswithinblock = F
>>
>>         threadhydroblocklist = F
>>
>>         threadhydrowithinblock = F
>>
>>         threadraytracebuild = F
>>
>>         threadwithinblockbuild = F
>>
>>         typematchedxfer = T
>>
>>         unbiased_geometry = F
>>
>>         updategravity = T
>>
>>         updatehydrofluxes = T
>>
>>         useburn = F
>>
>>         usecollectivehdf5 = T
>>
>>         useconductivity = F
>>
>>         usecool = F
>>
>>         usecosmology = F
>>
>>         usedeleptonize = F
>>
>>         usediffuse = F
>>
>>         usediffusecomputedtspecies = F
>>
>>         usediffusecomputedttherm = F
>>
>>         usediffusecomputedtvisc = F
>>
>>         usediffusecomputedtmagnetic = F
>>
>>         useenergydeposition = F
>>
>>         useflame = F
>>
>>         usegravity = T
>>
>>         useheat = F
>>
>>         useheatexchange = F
>>
>>         usehydro = T
>>
>>         useincompns = F
>>
>>         useionize = F
>>
>>         uselegacylabels = T
>>
>>         usemagneticresistivity = F
>>
>>         usemassdiffusivity = F
>>
>>         useopacity = F
>>
>>         useparticles = F
>>
>>         useplasmastate = F
>>
>>         usepolytrope = T [CHANGED]
>>
>>         useprimordialchemistry = F
>>
>>         useprotonemission = F
>>
>>         useprotonimaging = F
>>
>>         useradtrans = F
>>
>>         useraytrace = F
>>
>>         usests = F
>>
>>         usestsfordiffusion = F
>>
>>         usestir = F
>>
>>         usethomsonscattering = F
>>
>>         usetreeray = F
>>
>>         useturb = T
>>
>>         useviscosity = F
>>
>>         usexrayimaging = F
>>
>>         use_cma_advection = F
>>
>>         use_cma_flattening = F
>>
>>         use_flash_surr_blks_fill = T
>>
>>         use_reduced_orrery = T
>>
>>         use_steepening = T
>>
>>         writestatsummary = T
>>
>>
>>         ==============================================================================
>>
>>
>>         WARNING: Ignored Parameters :
>>
>>         These parameters were found in the flash.par file, but they were
>>
>>         not declared in any Config file for the simulation!
>>
>>
>>         energyFix
>>
>>
>>         ==============================================================================
>>
>>
>>         Known units of measurement:
>>
>>
>>         Unit CGS Value Base Unit
>>
>>         1 cm 1.0000 cm
>>
>>         2 s 1.0000 s
>>
>>         3 g 1.0000 g
>>
>>         4 K 1.0000 K
>>
>>         5 esu 1.0000 esu
>>
>>         6 mol 1.0000 mol
>>
>>         7 m 100.00 cm
>>
>>         8 km 1.00000E+05 cm
>>
>>         9 pc 3.08568E+18 cm
>>
>>         10 kpc 3.08568E+21 cm
>>
>>         11 Mpc 3.08568E+24 cm
>>
>>         12 Gpc 3.08568E+27 cm
>>
>>         13 Rsun 6.96000E+10 cm
>>
>>         14 AU 1.49598E+13 cm
>>
>>         15 yr 3.15569E+07 s
>>
>>         16 Myr 3.15569E+13 s
>>
>>         17 Gyr 3.15569E+16 s
>>
>>         18 kg 1000.0 g
>>
>>         19 Msun 1.98892E+33 g
>>
>>         20 amu 1.66054E-24 g
>>
>>         21 eV 11605. K
>>
>>         22 C 2.99792E+09 esu
>>
>>         23 LFLY 3.08568E+24 cm
>>
>>         24 TFLY 2.05759E+17 s
>>
>>         25 MFLY 9.88470E+45 g
>>
>>         26 clLength 3.08568E+24 cm
>>
>>         27 clTime 3.15569E+16 s
>>
>>         28 clMass 1.98892E+48 g
>>
>>         29 clTemp 1.16045E+07 K
>>
>>         -----------End of Units--------------------
>>
>>
>>         Known physical constants:
>>
>>
>>         Constant Name Constant Value cm s g K esu mol
>>
>>         1 Newton 6.67408E-08 3.0 -2.0 -1.0 0.0 0.0 0.0
>>
>>         2 speed of light 2.99792E+10 1.0 -1.0 0.0 0.0 0.0 0.0
>>
>>         3 Planck 6.62607E-27 2.0 -1.0 1.0 0.0 0.0 0.0
>>
>>         4 electron charge 4.80320E-10 0.0 0.0 0.0 0.0 1.0 0.0
>>
>>         5 electron mass 9.10938E-28 0.0 0.0 1.0 0.0 0.0 0.0
>>
>>         6 proton mass 1.67262E-24 0.0 0.0 1.0 0.0 0.0 0.0
>>
>>         7 fine-structure 7.29735E-03 0.0 0.0 0.0 0.0 0.0 0.0
>>
>>         8 Avogadro 6.02214E+23 0.0 0.0 0.0 0.0 0.0 -1.0
>>
>>         9 Boltzmann 1.38065E-16 2.0 -2.0 1.0 -1.0 0.0 0.0
>>
>>         10 ideal gas constant 8.31446E+07 2.0 -2.0 1.0 -1.0 0.0 -1.0
>>
>>         11 Wien 0.28978 1.0 0.0 0.0 1.0 0.0 0.0
>>
>>         12 Stefan-Boltzmann 5.67037E-05 0.0 -3.0 1.0 -4.0 0.0 0.0
>>
>>         13 Radiation Constant 7.56572E-15 -1.0 -2.0 1.0 -4.0 0.0 0.0
>>
>>         14 pi 3.1416 0.0 0.0 0.0 0.0 0.0 0.0
>>
>>         15 e 2.7183 0.0 0.0 0.0 0.0 0.0 0.0
>>
>>         16 Euler 0.57722 0.0 0.0 0.0 0.0 0.0 0.0
>>
>>         ==============================================================================
>>
>>
>>         Multifluid database: not configured in
>>
>>
>>         ==============================================================================
>>
>>         [ 12-05-2019 12:02:26.975 ] [gr_initGeometry] checking BCs
>>         for idir: 1
>>
>>         [ 12-05-2019 12:02:26.976 ] [gr_initGeometry] checking BCs
>>         for idir: 2
>>
>>         [ 12-05-2019 12:02:26.976 ] [gr_initGeometry] checking BCs
>>         for idir: 3
>>
>>         [ 12-05-2019 12:02:27.074 ] [GRID amr_refine_derefine]:
>>         initiating refinement
>>
>>         [ 12-05-2019 12:02:27.302 ] [GRID amr_refine_derefine]:
>>         redist. phase. tot blks requested: 9
>>
>>         [GRID amr_refine_derefine] min blks 0 max blks 1 tot blks 9
>>
>>         [GRID amr_refine_derefine] min leaf blks 0 max leaf blks 1
>>         tot leaf blks 8
>>
>>         [ 12-05-2019 12:02:27.543 ] [GRID amr_refine_derefine]:
>>         refinement complete
>>
>>         [ 12-05-2019 12:02:27.968 ] [GRID gr_expandDomain]:
>>         iteration=1, create level=3
>>
>>         INFO: Grid_fillGuardCells is ignoring masking.
>>
>>         [ 12-05-2019 12:02:28.078 ] [mpi_amr_comm_setup]:
>>         buffer_dim_send=74785, buffer_dim_recv=1
>>
>>         [ 12-05-2019 12:02:28.157 ] [GRID amr_refine_derefine]:
>>         initiating refinement
>>
>>         [ 12-05-2019 12:02:29.011 ] [GRID amr_refine_derefine]:
>>         redist. phase. tot blks requested: 73
>>
>>         [GRID amr_refine_derefine] min blks 0 max blks 1 tot blks 73
>>
>>         [GRID amr_refine_derefine] min leaf blks 0 max leaf blks 1
>>         tot leaf blks 64
>>
>>         [ 12-05-2019 12:02:29.440 ] [GRID amr_refine_derefine]:
>>         refinement complete
>>
>>         [ 12-05-2019 12:02:29.833 ] [GRID gr_expandDomain]:
>>         iteration=2, create level=4
>>
>>         [ 12-05-2019 12:02:30.253 ] [GRID amr_refine_derefine]:
>>         initiating refinement
>>
>>         [ 12-05-2019 12:02:30.322 ] [GRID amr_refine_derefine]:
>>         redist. phase. tot blks requested: 585
>>
>>         [GRID amr_refine_derefine] min blks 2 max blks 4 tot blks 585
>>
>>         [GRID amr_refine_derefine] min leaf blks 1 max leaf blks 3
>>         tot leaf blks 512
>>
>>         [ 12-05-2019 12:02:32.046 ] [GRID amr_refine_derefine]:
>>         refinement complete
>>
>>         [ 12-05-2019 12:02:32.564 ] [GRID gr_expandDomain]:
>>         iteration=3, create level=5
>>
>>         [ 12-05-2019 12:02:32.684 ] [mpi_amr_comm_setup]:
>>         buffer_dim_send=233185, buffer_dim_recv=135889
>>
>>         [ 12-05-2019 12:02:33.443 ] [GRID amr_refine_derefine]:
>>         initiating refinement
>>
>>         [ 12-05-2019 12:02:34.807 ] [GRID amr_refine_derefine]:
>>         redist. phase. tot blks requested: 3033
>>
>>         [GRID amr_refine_derefine] min blks 11 max blks 14 tot blks 3033
>>
>>         [GRID amr_refine_derefine] min leaf blks 9 max leaf blks 11
>>         tot leaf blks 2654
>>
>>         [ 12-05-2019 12:02:35.561 ] [GRID amr_refine_derefine]:
>>         refinement complete
>>
>>         [ 12-05-2019 12:02:35.977 ] [GRID gr_expandDomain]:
>>         iteration=4, create level=6
>>
>>         [ 12-05-2019 12:02:36.173 ] [mpi_amr_comm_setup]:
>>         buffer_dim_send=575965, buffer_dim_recv=351529
>>
>>         [ 12-05-2019 12:02:37.611 ] [GRID amr_refine_derefine]:
>>         initiating refinement
>>
>>         [ 12-05-2019 12:02:40.401 ] [GRID amr_refine_derefine]:
>>         redist. phase. tot blks requested: 6297
>>
>>         [GRID amr_refine_derefine] min blks 24 max blks 26 tot blks 6297
>>
>>         [GRID amr_refine_derefine] min leaf blks 20 max leaf blks 23
>>         tot leaf blks 5510
>>
>>         [ 12-05-2019 12:02:41.154 ] [GRID amr_refine_derefine]:
>>         refinement complete
>>
>>         [ 12-05-2019 12:02:41.544 ] [GRID gr_expandDomain]:
>>         iteration=5, create level=7
>>
>>         [ 12-05-2019 12:02:41.815 ] [mpi_amr_comm_setup]:
>>         buffer_dim_send=540997, buffer_dim_recv=411109
>>
>>         [ 12-05-2019 12:02:44.513 ] [GRID amr_refine_derefine]:
>>         initiating refinement
>>
>>         [ 12-05-2019 12:02:46.394 ] [GRID amr_refine_derefine]:
>>         redist. phase. tot blks requested: 16185
>>
>>         [GRID amr_refine_derefine] min blks 62 max blks 65 tot blks 16185
>>
>>         [GRID amr_refine_derefine] min leaf blks 54 max leaf blks 57
>>         tot leaf blks 14162
>>
>>         [ 12-05-2019 12:02:47.277 ] [GRID amr_refine_derefine]:
>>         refinement complete
>>
>>         [ 12-05-2019 12:02:47.696 ] [GRID gr_expandDomain]:
>>         iteration=6, create level=8
>>
>>         [ 12-05-2019 12:02:48.106 ] [mpi_amr_comm_setup]:
>>         buffer_dim_send=881329, buffer_dim_recv=699133
>>
>>         [ 12-05-2019 12:02:50.281 ] [GRID amr_refine_derefine]:
>>         initiating refinement
>>
>>         [ 12-05-2019 12:02:51.436 ] [GRID amr_refine_derefine]:
>>         redist. phase. tot blks requested: 89497
>>
>>         [GRID amr_refine_derefine] min blks 348 max blks 351 tot blks
>>         89497
>>
>>         [GRID amr_refine_derefine] min leaf blks 304 max leaf blks
>>         308 tot leaf blks 78310
>>
>>         [ 12-05-2019 12:02:53.153 ] [GRID amr_refine_derefine]:
>>         refinement complete
>>
>>         [ 12-05-2019 12:02:53.443 ] [GRID gr_expandDomain]:
>>         iteration=7, create level=9
>>
>>         [ 12-05-2019 12:02:56.118 ] [mpi_amr_comm_setup]:
>>         buffer_dim_send=1952545, buffer_dim_recv=1773109
>>
>>         [ 12-05-2019 12:03:03.333 ] [GRID amr_refine_derefine]:
>>         initiating refinement
>>
>>         [ 12-05-2019 12:03:07.980 ] [GRID amr_refine_derefine]:
>>         redist. phase. tot blks requested: 626345
>>
>>         [GRID amr_refine_derefine] min blks 2445 max blks 2449 tot
>>         blks 626345
>>
>>         [GRID amr_refine_derefine] min leaf blks 2139 max leaf blks
>>         2142 tot leaf blks 548052
>>
>>         [ 12-05-2019 12:03:13.985 ] [GRID amr_refine_derefine]:
>>         refinement complete
>>
>>         [ 12-05-2019 12:03:14.055 ] [GRID gr_expandDomain]:
>>         iteration=8, create level=9
>>
>>         [ 12-05-2019 12:03:25.399 ] [mpi_amr_comm_setup]:
>>         buffer_dim_send=6183577, buffer_dim_recv=5623945
>>
>>         [ 12-05-2019 12:03:55.420 ] [GRID amr_refine_derefine]:
>>         initiating refinement
>>
>>         [ 12-05-2019 12:03:58.377 ] [GRID amr_refine_derefine]:
>>         redist. phase. tot blks requested: 97801
>>
>>         [GRID amr_refine_derefine] min blks 381 max blks 384 tot blks
>>         97801
>>
>>         [GRID amr_refine_derefine] min leaf blks 333 max leaf blks
>>         336 tot leaf blks 85576
>>
>>         [ 12-05-2019 12:04:01.822 ] [GRID amr_refine_derefine]:
>>         refinement complete
>>
>>         [ 12-05-2019 12:04:01.849 ] [GRID gr_expandDomain]:
>>         iteration=9, create level=9
>>
>>         [ 12-05-2019 12:04:03.709 ] [mpi_amr_comm_setup]:
>>         buffer_dim_send=2129845, buffer_dim_recv=1958629
>>
>>         [ 12-05-2019 12:04:11.340 ] [GRID amr_refine_derefine]:
>>         initiating refinement
>>
>>         [ 12-05-2019 12:04:15.464 ] [GRID amr_refine_derefine]:
>>         redist. phase. tot blks requested: 626345
>>
>>         [GRID amr_refine_derefine] min blks 2445 max blks 2449 tot
>>         blks 626345
>>
>>         [GRID amr_refine_derefine] min leaf blks 2139 max leaf blks
>>         2142 tot leaf blks 548052
>>
>>         [ 12-05-2019 12:04:24.519 ] [GRID amr_refine_derefine]:
>>         refinement complete
>>
>>         [ 12-05-2019 12:04:24.588 ] [GRID gr_expandDomain]:
>>         iteration=10, create level=9
>>
>>         [ 12-05-2019 12:04:35.959 ] [mpi_amr_comm_setup]:
>>         buffer_dim_send=6183577, buffer_dim_recv=5623945
>>
>>         [ 12-05-2019 12:05:06.370 ] [GRID amr_refine_derefine]:
>>         initiating refinement
>>
>>         [ 12-05-2019 12:05:09.335 ] [GRID amr_refine_derefine]:
>>         redist. phase. tot blks requested: 97801
>>
>>         [GRID amr_refine_derefine] min blks 381 max blks 384 tot blks
>>         97801
>>
>>         [GRID amr_refine_derefine] min leaf blks 333 max leaf blks
>>         336 tot leaf blks 85576
>>
>>         [ 12-05-2019 12:05:12.763 ] [GRID amr_refine_derefine]:
>>         refinement complete
>>
>>         [ 12-05-2019 12:05:12.790 ] [GRID gr_expandDomain]:
>>         iteration=11, create level=9
>>
>>         [ 12-05-2019 12:05:14.289 ] [gr_ensureValidNeighborInfo]
>>         found mpi_pattern_id: -10
>>
>>         [ 12-05-2019 12:05:14.317 ] [mpi_amr_comm_setup]:
>>         buffer_dim_send=73525, buffer_dim_recv=62437
>>
>>         [ 12-05-2019 12:05:14.806 ] memory: /proc vsize (MiB):
>>         2918.54 (min) 3030.76 (max) 2946.03 (avg)
>>
>>         [ 12-05-2019 12:05:14.807 ] memory: /proc rss (MiB): 1798.14
>>         (min) 1883.50 (max) 1830.61 (avg)
>>
>>         [ 12-05-2019 12:05:14.808 ] memory: /proc vsize (MiB):
>>         2918.54 (min) 3030.76 (max) 2946.03 (avg)
>>
>>         [ 12-05-2019 12:05:14.809 ] memory: /proc rss (MiB): 1798.18
>>         (min) 1883.53 (max) 1830.67 (avg)
>>
>>         [ 12-05-2019 12:05:17.553 ] [BHTree]: Coefficients in the
>>         Taylor expansion written into file
>>
>>
>>
>>         The error file says:
>>
>>
>>         Fatal error in PMPI_Waitall: See the MPI_ERROR field in
>>         MPI_Status for the error code
>>
>>
>>         The out file says:
>>
>>
>>         RuntimeParameters_read: ignoring unknown parameter "energyFix"...
>>
>>         Grid_init: resolution based on runtime params:
>>
>>         lrefine dx dy dz
>>
>>         3 9.643E+17 9.643E+17 9.643E+17
>>
>>         4 4.821E+17 4.821E+17 4.821E+17
>>
>>         5 2.411E+17 2.411E+17 2.411E+17
>>
>>         6 1.205E+17 1.205E+17 1.205E+17
>>
>>         7 6.027E+16 6.027E+16 6.027E+16
>>
>>         8 3.013E+16 3.013E+16 3.013E+16
>>
>>         9 1.507E+16 1.507E+16 1.507E+16
>>
>>         MaterialProperties initialized
>>
>>         Cosmology initialized
>>
>>         Initializing Polytropic Equation of State
>>
>>         Source terms initialized
>>
>>         iteration, no. not moved = 0 0
>>
>>         refined: total leaf blocks = 8
>>
>>         refined: total blocks = 9
>>
>>         [amr_morton_process]: Initializing surr_blks using standard
>>         orrery implementation
>>
>>         INFO: Grid_fillGuardCells is ignoring masking.
>>
>>         iteration, no. not moved = 0 6
>>
>>         iteration, no. not moved = 1 0
>>
>>         refined: total leaf blocks = 64
>>
>>         refined: total blocks = 73
>>
>>         iteration, no. not moved = 0 46
>>
>>         iteration, no. not moved = 1 2
>>
>>         iteration, no. not moved = 2 0
>>
>>         refined: total leaf blocks = 512
>>
>>         refined: total blocks = 585
>>
>>         iteration, no. not moved = 0 459
>>
>>         iteration, no. not moved = 1 58
>>
>>         iteration, no. not moved = 2 0
>>
>>         refined: total leaf blocks = 2654
>>
>>         refined: total blocks = 3033
>>
>>         iteration, no. not moved = 0 2675
>>
>>         iteration, no. not moved = 1 765
>>
>>         iteration, no. not moved = 2 0
>>
>>         refined: total leaf blocks = 5510
>>
>>         refined: total blocks = 6297
>>
>>         iteration, no. not moved = 0 5642
>>
>>         iteration, no. not moved = 1 1146
>>
>>         iteration, no. not moved = 2 0
>>
>>         refined: total leaf blocks = 14162
>>
>>         refined: total blocks = 16185
>>
>>         iteration, no. not moved = 0 15209
>>
>>         iteration, no. not moved = 1 1503
>>
>>         iteration, no. not moved = 2 0
>>
>>         refined: total leaf blocks = 78310
>>
>>         refined: total blocks = 89497
>>
>>         iteration, no. not moved = 0 79298
>>
>>         iteration, no. not moved = 1 4927
>>
>>         iteration, no. not moved = 2 0
>>
>>         refined: total leaf blocks = 548052
>>
>>         refined: total blocks = 626345
>>
>>         iteration, no. not moved = 0 91484
>>
>>         iteration, no. not moved = 1 36351
>>
>>         iteration, no. not moved = 2 0
>>
>>         refined: total leaf blocks = 85576
>>
>>         refined: total blocks = 97801
>>
>>         iteration, no. not moved = 0 89804
>>
>>         iteration, no. not moved = 1 6879
>>
>>         iteration, no. not moved = 2 0
>>
>>         refined: total leaf blocks = 548052
>>
>>         refined: total blocks = 626345
>>
>>         iteration, no. not moved = 0 91447
>>
>>         iteration, no. not moved = 1 36080
>>
>>         iteration, no. not moved = 2 0
>>
>>         refined: total leaf blocks = 85576
>>
>>         refined: total blocks = 97801
>>
>>         NSIZE = 4 1
>>
>>         Tree solver initialized
>>
>>         Finished with Grid_initDomain, no restart
>>
>>         Ready to call Hydro_init
>>
>>         Hydro initialized
>>
>>         Gravity initialized
>>
>>         Initial dt verified
>>
>>
>>         I’m not sure what is the reason for this kind of crash, have
>>         you ever experienced something similar?
>>
>>         My guess is that it is a communication problem due to the use
>>         of mpich-3.2.1 but i’m not sure.
>>
>>
>>         Cheers,
>>
>>
>>
>>     -- 
>>     Eddie Hansen, PhD
>>     Postdoctoral Scholar
>>     University of Chicago
>>     607-341-6126 | Flash Center
>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://flash.rochester.edu/pipermail/flash-users/attachments/20191210/d54fa968/attachment.htm>


More information about the flash-users mailing list