[FLASH-USERS] HDF5 IO problem
Zylstra, Alex
zylstra at lanl.gov
Fri Jun 22 16:15:37 EDT 2018
Hi everybody,
I am having a (perhaps) similar issue on Mac with the following versions:
FLASH4.5
mpich-3.2.1
hdf5-1.8.20
hypre-2.11.2
parallel-netcdf-1.9.0
Sedov with serial IO works fine for me, but when I try the laser slab problem I get a segmentation fault as it starts to write the initial checkpoint. From the log file:
[ 06-22-2018 13:05:22.498 ] [IO_writeCheckpoint] open: type=checkpoint name=lasslab_hdf5_chk_0000
[ 06-22-2018 13:05:22.533 ] [io_writeData]: wrote 58 blocks
And the end of the console output is
Hydro initialized
Gravity initialized
Initial dt verified
Program received signal SIGSEGV: Segmentation fault - invalid memory reference.
Backtrace for this error:
#0 0x106649f3c
#1 0x1066492e3
#2 0x7fff79d38f59
Segmentation fault: 11
Any help appreciated!
-Alex
From: <flash-users-bounces at flash.uchicago.edu> on behalf of "t_schlegel at t-online.de" <t_schlegel at t-online.de>
Reply-To: "t_schlegel at t-online.de" <t_schlegel at t-online.de>
Date: Thursday, June 21, 2018 at 3:14 AM
To: "flash-users at flash.uchicago.edu" <flash-users at flash.uchicago.edu>
Subject: [FLASH-USERS] HDF5 IO problem
Hello,
I have installed the modules:
GCC/7.2.0 5)
OpenMPI/3.0.0-GCC-7.2.0
FLASH/4.5-OpenMPI-3.0.0-GCC-7.2.0
hypre/2.11.2-OpenMPI-3.0.0-GCC-7.2.0
HDF5/1.10.1-OpenMPI-3.0.0-GCC-7.2.0 for parallel output
either HDF5/1.10.2-GCC-7.2.0 for serial one
parallel_netcdf/1.9.0-OpenMPI-3.0.0-GCC-7.2.0 on our cluster.
The Sedov example with serial HDF5 works as expected,
but using the HDF5/1.10.1-OpenMPI-3.0.0-GCC-7.2.0 module instead of serial IO
(./setup Sedov -auto +parallelio, putting the right path for HDF5/1.10.1-OpenMPI-3.0.0-GCC-7.2.0
in Makefile.h)
the job will be aborted at first write of a ckeckpoint file.
Part of the error output:
HDF5-DIAG: Error detected in HDF5 (1.10.1) MPI-process 3:
#000: H5Ddeprec.c line 187 in H5Dopen1(): not a location
major: Invalid arguments to routine
minor: Inappropriate type
#001: H5Gloc.c line 251 in H5G_loc(): invalid object ID
major: Invalid arguments to routine
minor: Bad value
flash4: io_h5_attribute.c:49: io_h5_attribute_create: Assertion `dsetID >= 0' failed.
Program received signal SIGABRT: Process abort signal.
A similar situation I observed with the more complex example LaserSlab:
HDF5-DIAG: Error detected in HDF5 (1.10.1) MPI-process 4:
#000: H5Dio.c line 269 in H5Dwrite(): can't prepare for writing data
major: Dataset
minor: Write failed
#001: H5Dio.c line 345 in H5D__pre_write(): can't write data
major: Dataset
minor: Write failed
#002: H5Dio.c line 791 in H5D__write(): can't write data
major: Dataset
minor: Write failed
#003: H5Dmpio.c line 527 in H5D__contig_collective_write(): couldn't finish shared collective MPI-IO
major: Low-level I/O
minor: Write failed
#004: H5Dmpio.c line 1398 in H5D__inter_collective_io(): couldn't finish collective MPI-IO
major: Low-level I/O
minor: Can't get value
#005: H5Dmpio.c line 1442 in H5D__final_collective_io(): optimized write failed
major: Dataset
minor: Write failed
#006: H5Dmpio.c line 295 in H5D__mpio_select_write(): can't finish collective parallel write
major: Low-level I/O
minor: Write failed
#007: H5Fio.c line 195 in H5F_block_write(): write through page buffer failed
major: Low-level I/O
minor: Write failed
#008: H5PB.c line 1041 in H5PB_write(): write through metadata accumulator failed
major: Page Buffering
minor: Write failed
#009: H5Faccum.c line 834 in H5F__accum_write(): file write failed
major: Low-level I/O
minor: Write failed
#010: H5FDint.c line 308 in H5FD_write(): driver write request failed
major: Virtual File Layer
minor: Write failed
#011: H5FDmpio.c line 1864 in H5FD_mpio_write(): file write failed
major: Low-level I/O
minor: Write failed
flash4: io_h5_xfer_mesh_dataset.c:210: io_h5_xfer_mesh_dataset: Assertion `err >= 0' failed.
Program received signal SIGABRT: Process abort signal.
How can I overcome these problems?
Thanks in advancce,
Theo
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://flash.rochester.edu/pipermail/flash-users/attachments/20180622/a1bc75d2/attachment-0001.htm>
More information about the flash-users
mailing list