[FLASH-USERS] Issue with Cray Compiler
Alan Calder
alan.calder at stonybrook.edu
Wed Nov 11 21:20:25 EST 2020
Hello,
I am writing to describe an issue with running Flash4 on a Cray A64FX
machine with the Cray compiler, version 10.0.1. We are running Flash
4.6.2. I thought I would describe what I am seeing so as to ask for any
advice before digging in further. I am not completely sure if this is a
bug, hence my sending it here.
First I should note that things work well with the Gnu compilers. We do not
see this problem when compiling with the gnu compilers.
The shortest possible summary is that the code seems to enter an unphysical
state and crash, and with the split hydro solver I see a warning
WARNING after gc filling: min. unk(EINT_VAR)=9.9999999735241242E-11
PE=4 block=6
type=1
We compile with -c -g -G 2 -s real64 -s integer32 and the code generates a
few warnings, none of which seem relevant. We've tried a few problems, too,
and are seeing the same issues. For this report I ran the Sedov problem.
With the default units, and set up with
./setup Sedov -auto -site=ookami.stonybrook.edu
the code runs a while and then suddenly the solution seems to hit some bad
state causing the time step to suddenly drop and then the code crashes.
[acalder at fj-debug1 object]$ !mpi
mpiexec -n 8 ./flash4
Grid_init: resolution based on runtime params:
lrefine dx dy
1 0.125 0.125
2 0.063 0.063
3 0.031 0.031
4 0.016 0.016
5 0.008 0.008
6 0.004 0.004
MaterialProperties initialized
Cosmology initialized
Source terms initialized
sim_rhoAmbient is 1.
iteration, no. not moved = 2*0
refined: total leaf blocks = 1
refined: total blocks = 1
[amr_morton_process]: Initializing surr_blks using standard orrery
implementation
INFO: Grid_fillGuardCells is ignoring masking.
iteration, no. not moved = 2*0
refined: total leaf blocks = 4
refined: total blocks = 5
iteration, no. not moved = 0, 3
iteration, no. not moved = 2*1
iteration, no. not moved = 2, 0
refined: total leaf blocks = 16
refined: total blocks = 21
iteration, no. not moved = 0, 6
iteration, no. not moved = 1, 0
refined: total leaf blocks = 28
refined: total blocks = 37
iteration, no. not moved = 0, 9
iteration, no. not moved = 1, 2
iteration, no. not moved = 2, 0
refined: total leaf blocks = 40
refined: total blocks = 53
iteration, no. not moved = 0, 16
iteration, no. not moved = 1, 2
iteration, no. not moved = 2, 0
refined: total leaf blocks = 52
refined: total blocks = 69
Finished with Grid_initDomain, no restart
Ready to call Hydro_init
Hydro initialized
Gravity initialized
Initial dt verified
*** Wrote checkpoint file to sedov_hdf5_chk_0000 ****
*** Wrote plotfile to sedov_hdf5_plt_cnt_0000 ****
Initial plotfile written
Driver init all done
n t dt ( x, y, z) |
dt_hydro
1 1.0000E-10 2.0000E-10 ( 0.498 , 0.498 , 0.00 ) |
1.012E-04
2 3.0000E-10 4.0000E-10 ( 0.498 , 0.494 , 0.00 ) |
1.012E-04
3 7.0000E-10 8.0000E-10 ( 0.498 , 0.494 , 0.00 ) |
1.012E-04
4 1.5000E-09 1.6000E-09 ( 0.498 , 0.494 , 0.00 ) |
1.012E-04
snip
26 4.4575E-04 4.7891E-05 ( 0.502 , 0.510 , 0.00 ) |
4.789E-05
27 4.9364E-04 4.5144E-05 ( 0.498 , 0.510 , 0.00 ) |
4.514E-05
28 5.3879E-04 4.1959E-05 ( 0.498 , 0.510 , 0.00 ) |
4.196E-05
29 5.8075E-04 2.3721E-05 ( 0.506 , 0.498 , 0.00 ) |
2.372E-05
iteration, no. not moved = 0, 17
iteration, no. not moved = 1, 2
iteration, no. not moved = 2, 0
refined: total leaf blocks = 88
refined: total blocks = 117
30 6.0447E-04 1.0000E-10 ( 0.506 , 0.498 , 0.00 ) |
4.269E-18
31 6.0447E-04 1.0000E-10 ( 0.498 , 0.498 , 0.00 ) |
2.897E-56
32 6.0447E-04 1.0000E-10 ( 0.510 , 0.494 , 0.00 ) |
1.206-156
dtCheck= 0.
DRIVER_ABORT: [Hydro]: Computed dt is not positive! Aborting!
--------------------------------------------------------------------------
MPI_ABORT was invoked on rank 1 in communicator MPI_COMM_WORLD
with errorcode 1.
NOTE: invoking MPI_ABORT causes Open MPI to kill all MPI processes.
You may or may not see output from other processes, depending on
exactly when Open MPI kills them.
--------------------------------------------------------------------------
In the above case, the glitch happened after a refinement, but it does not
always happen after a refinement. Here I changed the max level of
refinement. After step 69 the glitch happens.
[acalder at fj-debug1 object]$ vi flash.par
[acalder at fj-debug1 object]$ !mpi
mpiexec -n 8 ./flash4
Grid_init: resolution based on runtime params:
lrefine dx dy
1 0.125 0.125
2 0.063 0.063
3 0.031 0.031
4 0.016 0.016
5 0.008 0.008
6 0.004 0.004
7 0.002 0.002
MaterialProperties initialized
Cosmology initialized
Source terms initialized
sim_rhoAmbient is 1.
iteration, no. not moved = 2*0
refined: total leaf blocks = 1
refined: total blocks = 1
[amr_morton_process]: Initializing surr_blks using standard orrery
implementation
INFO: Grid_fillGuardCells is ignoring masking.
iteration, no. not moved = 2*0
refined: total leaf blocks = 4
refined: total blocks = 5
iteration, no. not moved = 0, 3
iteration, no. not moved = 2*1
iteration, no. not moved = 2, 0
refined: total leaf blocks = 16
refined: total blocks = 21
iteration, no. not moved = 0, 6
iteration, no. not moved = 1, 0
refined: total leaf blocks = 28
refined: total blocks = 37
iteration, no. not moved = 0, 9
iteration, no. not moved = 1, 2
iteration, no. not moved = 2, 0
refined: total leaf blocks = 40
refined: total blocks = 53
iteration, no. not moved = 0, 16
iteration, no. not moved = 1, 2
iteration, no. not moved = 2, 0
refined: total leaf blocks = 52
refined: total blocks = 69
iteration, no. not moved = 0, 20
iteration, no. not moved = 2*1
iteration, no. not moved = 2, 0
refined: total leaf blocks = 64
refined: total blocks = 85
Finished with Grid_initDomain, no restart
Ready to call Hydro_init
Hydro initialized
Gravity initialized
Initial dt verified
*** Wrote checkpoint file to sedov_hdf5_chk_0000 ****
*** Wrote plotfile to sedov_hdf5_plt_cnt_0000 ****
Initial plotfile written
Driver init all done
n t dt ( x, y, z) |
dt_hydro
1 1.0000E-10 2.0000E-10 ( 0.499 , 0.491 , 0.00 ) |
5.060E-05
2 3.0000E-10 4.0000E-10 ( 0.499 , 0.493 , 0.00 ) |
5.060E-05
3 7.0000E-10 8.0000E-10 ( 0.499 , 0.493 , 0.00 ) |
5.060E-05
4 1.5000E-09 1.6000E-09 ( 0.499 , 0.493 , 0.00 ) |
5.060E-05
5 3.1000E-09 3.2000E-09 ( 0.499 , 0.493 , 0.00 ) |
5.060E-05
6 6.3000E-09 6.4000E-09 ( 0.499 , 0.493 , 0.00 ) |
5.059E-05
7 1.2700E-08 1.2800E-08 ( 0.499 , 0.493 , 0.00 ) |
5.059E-05
8 2.5500E-08 2.5600E-08 ( 0.499 , 0.493 , 0.00 ) |
5.058E-05
9 5.1100E-08 5.1200E-08 ( 0.499 , 0.495 , 0.00 ) |
5.057E-05
10 1.0230E-07 1.0240E-07 ( 0.499 , 0.493 , 0.00 ) |
5.054E-05
11 2.0470E-07 2.0480E-07 ( 0.499 , 0.493 , 0.00 ) |
5.048E-05
12 4.0950E-07 4.0960E-07 ( 0.499 , 0.493 , 0.00 ) |
5.036E-05
13 8.1910E-07 8.1920E-07 ( 0.499 , 0.493 , 0.00 ) |
5.014E-05
iteration, no. not moved = 0, 19
iteration, no. not moved = 2*1
iteration, no. not moved = 2, 0
refined: total leaf blocks = 112
refined: total blocks = 149
14 1.6383E-06 1.6384E-06 ( 0.499 , 0.493 , 0.00 ) |
4.968E-05
15 3.2767E-06 3.2768E-06 ( 0.499 , 0.493 , 0.00 ) |
4.881E-05
16 6.5535E-06 6.5536E-06 ( 0.499 , 0.493 , 0.00 ) |
4.719E-05
17 1.3107E-05 1.3107E-05 ( 0.499 , 0.487 , 0.00 ) |
4.400E-05
18 2.6214E-05 2.6214E-05 ( 0.499 , 0.487 , 0.00 ) |
3.861E-05
19 5.2429E-05 3.1822E-05 ( 0.499 , 0.487 , 0.00 ) |
3.182E-05
iteration, no. not moved = 0, 56
iteration, no. not moved = 1, 2
iteration, no. not moved = 2, 0
refined: total leaf blocks = 160
refined: total blocks = 213
20 8.4250E-05 2.8211E-05 ( 0.499 , 0.487 , 0.00 ) |
2.821E-05
21 1.1246E-04 2.6569E-05 ( 0.499 , 0.487 , 0.00 ) |
2.657E-05
22 1.3903E-04 2.5651E-05 ( 0.499 , 0.487 , 0.00 ) |
2.565E-05
snip
64 5.0247E-04 1.7003E-06 ( 0.501 , 0.501 , 0.00 ) |
1.700E-06
65 5.0417E-04 1.6235E-06 ( 0.501 , 0.501 , 0.00 ) |
1.624E-06
66 5.0580E-04 1.5568E-06 ( 0.501 , 0.501 , 0.00 ) |
1.557E-06
67 5.0735E-04 1.4925E-06 ( 0.501 , 0.501 , 0.00 ) |
1.493E-06
68 5.0885E-04 1.4258E-06 ( 0.501 , 0.501 , 0.00 ) |
1.426E-06
69 5.1027E-04 1.3543E-06 ( 0.501 , 0.501 , 0.00 ) |
1.354E-06
70 5.1163E-04 1.0000E-10 ( 0.513 , 0.489 , 0.00 ) |
1.888E-14
71 5.1163E-04 1.0000E-10 ( 0.487 , 0.489 , 0.00 ) |
3.189E-44
dtCheck= 0.
DRIVER_ABORT: [Hydro]: Computed dt is not positive! Aborting!
--------------------------------------------------------------------------
MPI_ABORT was invoked on rank 1 in communicator MPI_COMM_WORLD
with errorcode 1.
NOTE: invoking MPI_ABORT causes Open MPI to kill all MPI processes.
You may or may not see output from other processes, depending on
exactly when Open MPI kills them.
--------------------------------------------------------------------------
[acalder at fj-debug1 object]$
Just to play with it, I tried the split hydro solver. In Units, I see among
others
INCLUDE Driver/DriverMain/Split
INCLUDE physics/Hydro/HydroMain/split/PPM/PPMKernel
And when I run it, I see an intriguing warning
WARNING after gc filling: min. unk(EINT_VAR)=9.9999999735241242E-11
PE=4 block=6
type=1
Here's the output of a run:
[acalder at fj-debug1 object]$ !mpi
mpiexec -n 8 ./flash4
RuntimeParameters_read: ignoring unknown parameter "order"...
RuntimeParameters_read: ignoring unknown parameter "slopeLimiter"...
RuntimeParameters_read: ignoring unknown parameter "LimitedSlopeBeta"...
RuntimeParameters_read: ignoring unknown parameter "use_avisc"...
RuntimeParameters_read: ignoring unknown parameter "use_flattening"...
RuntimeParameters_read: ignoring unknown parameter "use_upwindTVD"...
RuntimeParameters_read: ignoring unknown parameter "RiemannSolver"...
RuntimeParameters_read: ignoring unknown parameter "entropy"...
RuntimeParameters_read: ignoring unknown parameter "shockDetect"...
Grid_init: resolution based on runtime params:
lrefine dx dy
1 0.125 0.125
2 0.063 0.063
3 0.031 0.031
4 0.016 0.016
5 0.008 0.008
6 0.004 0.004
MaterialProperties initialized
Cosmology initialized
Source terms initialized
sim_rhoAmbient is 1.
iteration, no. not moved = 2*0
refined: total leaf blocks = 1
refined: total blocks = 1
[amr_morton_process]: Initializing surr_blks using standard orrery
implementation
INFO: Grid_fillGuardCells is ignoring masking.
iteration, no. not moved = 2*0
refined: total leaf blocks = 4
refined: total blocks = 5
iteration, no. not moved = 0, 3
iteration, no. not moved = 2*1
iteration, no. not moved = 2, 0
refined: total leaf blocks = 16
refined: total blocks = 21
iteration, no. not moved = 0, 6
iteration, no. not moved = 1, 0
refined: total leaf blocks = 28
refined: total blocks = 37
iteration, no. not moved = 0, 9
iteration, no. not moved = 1, 2
iteration, no. not moved = 2, 0
refined: total leaf blocks = 40
refined: total blocks = 53
iteration, no. not moved = 0, 16
iteration, no. not moved = 1, 2
iteration, no. not moved = 2, 0
refined: total leaf blocks = 52
refined: total blocks = 69
Finished with Grid_initDomain, no restart
Ready to call Hydro_init
Info: Hydro_init has set hy_specialFluxVars to 3
Hydro initialized
Gravity initialized
Initial dt verified
*** Wrote checkpoint file to sedov_hdf5_chk_0000 ****
*** Wrote plotfile to sedov_hdf5_plt_cnt_0000 ****
Initial plotfile written
Driver init all done
n t dt ( x, y, z) |
dt_hydro
1 2.0000E-10 2.0000E-10 ( 0.498 , 0.502 , 0.00 ) |
7.969E-05
WARNING after gc filling: min. unk(EINT_VAR)=9.9999999735241242E-11
PE=4 block=6
type=1
16 0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04
0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.
25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04
15 0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04
0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.
25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04
14 0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04
0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.
25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04
13 0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04
0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.
25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04
12 0.25E-04 0.25E-04 0.10E-09 0.10E-09 0.25E-04 0.25E-04 0.25E-04
0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.
25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04
11 0.25E-04 0.25E-04 0.10E-09 0.10E-09 0.25E-04 0.25E-04 0.25E-04
0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.
25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04
10 0.25E-04 0.25E-04 0.10E-09 0.10E-09 0.25E-04 0.25E-04 0.25E-04
0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.
25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04
9 0.25E-04 0.25E-04 0.10E-09 0.10E-09 0.25E-04 0.25E-04 0.25E-04
0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.
25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04
8 0.25E-04 0.25E-04 0.10E-09 0.10E-09 0.25E-04 0.25E-04 0.25E-04
0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.
10E-09 0.25E-04 0.25E-04 0.25E-04 0.25E-04
7 0.25E-04 0.25E-04 0.10E-09 0.10E-09 0.25E-04 0.25E-04 0.25E-04
0.25E-04 0.25E-04 0.25E-04 0.26E-04 0.
10E-09 0.25E-04 0.25E-04 0.25E-04 0.25E-04
6 0.25E-04 0.25E-04 0.10E-09 0.10E-09 0.25E-04 0.25E-04 0.25E-04
0.25E-04 0.25E-04 0.25E-04 0.26E-04 0.
10E-09 0.25E-04 0.25E-04 0.25E-04 0.18E-04
5 0.25E-04 0.25E-04 0.10E-09 0.10E-09 0.25E-04 0.25E-04 0.25E-04
0.25E-04 0.25E-04 0.25E-04 0.26E-04 0.
10E-09 0.25E-04 0.25E-04 0.25E-04 0.22E-04
4 0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04
0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.
15E-04 0.25E-04 0.25E-04 0.25E-04 0.13E-04
3 0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04
0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.
15E-04 0.25E-04 0.25E-04 0.25E-04 0.17E-04
2 0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04
0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.
15E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04
1 0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04
0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.
23E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04
WARNING after gc filling: min. unk(EINT_VAR)=9.9999999735241242E-11
PE=5 block=3
type=1
16 0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04
0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.
25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04
15 0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04
0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.
25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04
14 0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04
0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.
25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04
13 0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04
0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.
25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04
12 0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04
0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.
25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04
11 0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04
0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.
25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04
10 0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04
0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.
25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04
9 0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04
0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.
25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04
8 0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04
0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.
25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04
7 0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04
0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.
25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04
6 0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04
0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.
25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04
5 0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04
0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.
25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04
4 0.25E-04 0.25E-04 0.10E-09 0.10E-09 0.25E-04 0.25E-04 0.25E-04
0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.
25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04
3 0.25E-04 0.25E-04 0.10E-09 0.10E-09 0.25E-04 0.25E-04 0.25E-04
0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.
25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04
2 0.25E-04 0.25E-04 0.10E-09 0.10E-09 0.25E-04 0.25E-04 0.25E-04
0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.
25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04
1 0.25E-04 0.25E-04 0.10E-09 0.10E-09 0.25E-04 0.25E-04 0.25E-04
0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.
25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04
WARNING after gc filling: min. unk(EINT_VAR)=9.9999999384891323E-11
PE=4 block=6
type=1
16 0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04
0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.
25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04
15 0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04
0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.
25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04
14 0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04
0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.
25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04
13 0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04
0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.
25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04
12 0.25E-04 0.25E-04 0.10E-09 0.10E-09 0.25E-04 0.25E-04 0.25E-04
0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.
25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04
11 0.25E-04 0.25E-04 0.10E-09 0.10E-09 0.25E-04 0.25E-04 0.25E-04
0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.
25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04
10 0.25E-04 0.25E-04 0.10E-09 0.10E-09 0.25E-04 0.25E-04 0.25E-04
0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.
25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04
9 0.25E-04 0.25E-04 0.10E-09 0.10E-09 0.25E-04 0.25E-04 0.25E-04
0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.
25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04
8 0.25E-04 0.25E-04 0.10E-09 0.10E-09 0.25E-04 0.25E-04 0.25E-04
0.25E-04 0.25E-04 0.25E-04 0.26E-04 0.
10E-09 0.25E-04 0.25E-04 0.25E-04 0.25E-04
7 0.25E-04 0.25E-04 0.10E-09 0.10E-09 0.25E-04 0.25E-04 0.25E-04
0.25E-04 0.25E-04 0.25E-04 0.34E-04 0.
10E-09 0.25E-04 0.25E-04 0.25E-04 0.25E-04
6 0.25E-04 0.25E-04 0.10E-09 0.10E-09 0.25E-04 0.25E-04 0.25E-04
0.25E-04 0.25E-04 0.25E-04 0.34E-04 0.
10E-09 0.25E-04 0.25E-04 0.25E-04 0.50E-04
5 0.25E-04 0.25E-04 0.10E-09 0.10E-09 0.25E-04 0.25E-04 0.25E-04
0.25E-04 0.25E-04 0.25E-04 0.34E-04 0.
10E-09 0.25E-04 0.25E-04 0.25E-04 0.81E-04
4 0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04
0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.
28E-05 0.25E-04 0.25E-04 0.25E-04 0.13E-04
3 0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04
0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.
28E-05 0.25E-04 0.25E-04 0.25E-04 0.15E-04
2 0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04
0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.
28E-05 0.24E-04 0.25E-04 0.25E-04 0.25E-04
1 0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.25E-04
0.25E-04 0.25E-04 0.25E-04 0.25E-04 0.
21E-04 0.24E-04 0.25E-04 0.25E-04 0.25E-04
WARNING after gc filling: min. unk(EINT_VAR)=9.9999999384891323E-11
PE=5 block=3
type=1
Finally, here are warnings.
Several things like this:
USE_16_API -O0 -c -DMAXBLOCKS=1000 -DNXB=8 -DNYB=8 -DNZB=1 -DN_DIM=2
io_attribute.c
io_attribute.c: In function ‘io_attribute_create_’:
io_attribute.c:66:5: warning: implicit declaration of function
‘Driver_abortFlashC’ [-Wimplicit-function-d
eclaration]
Driver_abortFlashC("[io_attribute_create]: Unknown I/O");
^~~~~~~~~~~~~~~~~~
And this one:
/lustre/projects/global/software/a64fx/openmpi/4.0.5/bin/mpicc
-I/cm/shared/apps/hdf5/1.10.1/include -DH5_
USE_16_API -O0 -c -DMAXBLOCKS=1000 -DNXB=8 -DNYB=8 -DNZB=1 -DN_DIM=2
io_h5write_generic_real_arr.c
io_h5write_generic_int_arr.c: In function ‘io_h5write_generic_int_arr_’:
io_h5write_generic_int_arr.c:82:31: warning: implicit declaration of
function ‘malloc’ [-Wimplicit-functio
n-declaration]
dataset_name_new = (char *) malloc((*name_len) + 1 * sizeof(char));
^~~~~~
io_h5write_generic_int_arr.c:82:31: warning: incompatible implicit
declaration of built-in function ‘mallo
c’
io_h5write_generic_int_arr.c:82:31: note: include ‘<stdlib.h>’ or provide a
declaration of ‘malloc’
io_h5write_generic_int_arr.c:6:1:
+#include <stdlib.h>
io_h5write_generic_int_arr.c:82:31:
dataset_name_new = (char *) malloc((*name_len) + 1 * sizeof(char));
^~~~~~
io_h5write_generic_int_arr.c:87:3: warning: implicit declaration of
function ‘strncpy’ [-Wimplicit-functio
n-declaration]
strncpy(dataset_name_new, dataset_name, *name_len);
^~~~~~~
io_h5write_generic_int_arr.c:87:3: warning: incompatible implicit
declaration of built-in function ‘strncp
y’
io_h5write_generic_int_arr.c:87:3: note: include ‘<string.h>’ or provide a
declaration of ‘strncpy’
io_h5write_generic_int_arr.c:6:1:
+#include <string.h>
io_h5write_generic_int_arr.c:87:3:
strncpy(dataset_name_new, dataset_name, *name_len);
^~~~~~~
io_h5write_generic_int_arr.c:187:3: warning: implicit declaration of
function ‘free’ [-Wimplicit-function-
declaration]
free(dataset_name_new);
^~~~
io_h5write_generic_int_arr.c:187:3: warning: incompatible implicit
declaration of built-in function ‘free’
io_h5write_generic_int_arr.c:187:3: note: include ‘<stdlib.h>’ or provide a
declaration of ‘free’
/lustre/projects/global/software/a64fx/openmpi/4.0.5/bin/mpicc
-I/cm/shared/apps/hdf5/1.10.1/include -DH5_
USE_16_API -O0 -c -DMAXBLOCKS=1000 -DNXB=8 -DNYB=8 -DNZB=1 -DN_DIM=2
io_h5write_generic_real_arr.c
io_h5write_generic_real_arr.c: In function ‘io_h5write_generic_real_arr_’:
io_h5write_generic_real_arr.c:33:31: warning: implicit declaration of
function ‘malloc’ [-Wimplicit-functi
on-declaration]
dataset_name_new = (char *) malloc((*name_len) + 1 * sizeof(char));
^~~~~~
io_h5write_generic_real_arr.c:33:31: warning: incompatible implicit
declaration of built-in function ‘mall
oc’
io_h5write_generic_real_arr.c:33:31: note: include ‘<stdlib.h>’ or provide
a declaration of ‘malloc’
io_h5write_generic_real_arr.c:6:1:
+#include <stdlib.h>
io_h5write_generic_real_arr.c:33:31:
dataset_name_new = (char *) malloc((*name_len) + 1 * sizeof(char));
^~~~~~
io_h5write_generic_real_arr.c:38:3: warning: implicit declaration of
function ‘strncpy’ [-Wimplicit-functi
on-declaration]
strncpy(dataset_name_new, dataset_name, *name_len);
^~~~~~~
io_h5write_generic_real_arr.c:38:3: warning: incompatible implicit
declaration of built-in function ‘strnc
py’
io_h5write_generic_real_arr.c:38:3: note: include ‘<string.h>’ or provide a
declaration of ‘strncpy’
io_h5write_generic_real_arr.c:6:1:
+#include <string.h>
io_h5write_generic_real_arr.c:38:3:
strncpy(dataset_name_new, dataset_name, *name_len);
^~~~~~~
io_h5write_generic_real_arr.c:139:3: warning: implicit declaration of
function ‘free’ [-Wimplicit-function
-declaration]
free(dataset_name_new);
^~~~
io_h5write_generic_real_arr.c:139:3: warning: incompatible implicit
declaration of built-in function ‘free
’
io_h5write_generic_real_arr.c:139:3: note: include ‘<stdlib.h>’ or provide
a declaration of ‘free’
Thanks for any advice or suggestions!
Alan Calder
--
Alan C. Calder
Department of Physics and Astronomy
State University of New York at Stony Brook
Stony Brook, NY 11794-3800
office: ESS 438
phone: (631) 632-1176
fax: (631) 632-1745
web: http://www.astro.sunysb.edu/acalder
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://flash.rochester.edu/pipermail/flash-users/attachments/20201111/2beef963/attachment.htm>
More information about the flash-users
mailing list