| Commit message (Collapse) | Author | Age | Files | Lines |
| |
|
| |
|
|\
| |
| |
| | |
https://bitbucket.hdfgroup.org/scm/~jake.smith/hdf5 into feature/vfd_splitter_mirror_a
|
| |
| |
| |
| |
| | |
occur. Next commit will demote the warnings promoted to errors back to
warnings again.
|
|/
|
|
|
|
|
|
|
|
|
|
|
|
|
| |
* "Simultaneous and equivalent" Read-Write and Write-Only channels for
file I/O.
* Only supports drivers with the H5FD_FEAT_DEFAULT_VFD_COMPATIBLE flag for
now, preventing issues with multi-file drivers.
Add Mirror VFD to library.
* Write-only operations over a network.
* Uses TCP/IP sockets.
* Server and auxiliary server-shutdown programs provided in a new directory,
`utils/mirror_vfd`.
* Automated testing via loopback ("remote" of localhost).
|
|
|
|
| |
level and fixing some code.' (commit 5c911d8baf3)"
|
|\
| |
| |
| | |
to pick up the NetBSD-compatibility changes that I recently merged.
|
| |
| |
| |
| |
| |
| |
| |
| |
| | |
This is handy for NetBSD where HDF5 examples are installed
by convention in $prefix/share/examples/hdf5/ rather than in
${prefix}/share/hdf5_examples/, which is the HDF5 default.
Place hdf5_examples/ under ${datarootdir} which on most systems will be
${prefix}/share/, anyway.
|
| | |
|
|/ |
|
| |
|
| |
|
|
|
|
| |
branch.
|
|
|
|
| |
This reverts commit 431123503d387bce30b1f6db7dc4ad3a33bc51bb.
|
| |
|
| |
|
| |
|
|
|
|
|
|
|
|
|
|
|
| |
Merge changes from update_merged_S3_HDFS branch into develop.
commit d5034315aea88629929ac0c9c59ebfafd5f21a31
Merge: 9c48823 d3fdcd8
Author: Larry Knox <lrknox@hdfgroup.org>
Date: Thu Jul 25 08:24:53 2019 -0500
Merge branch 'develop' into update_merged_S3_HDFS
|
| |
|
|
|
|
|
|
|
| |
should always be built and installed whether tools are enabled or
disabled. Also added Makefile.am to bin to build h5redeploy and
to install and uninstall them. h5cc is created from h5cc.in by
configure.
|
|
|
|
|
|
|
| |
disabling tests.
Moved h5cc.in from tools/src/misc to src directory to always create h5cc
whether or not tools are enabled.
Added configuration status of tools and tests to libhdf5.settings.
|
|
|
|
| |
is enabled for each.
|
|
|
|
| |
Update version to 1.11.6.
|
|\
| |
| |
| | |
https://bitbucket.hdfgroup.org/scm/~songyulu/hdf5_ray into HDFFV-10658-performance-drop-from-1-8
|
| |
| |
| |
| | |
Update version to 1.11.5.
|
|/
|
|
|
|
| |
1. restored the datatype, dataspace, and LCPL of the dataset for VOL connector back to the properties.
2. splitted external.c and vds.c because they called HDsetenv in the program, instead using shell scripts to set the environment variables.
3. changed H5CX_get_vds_prefix and H5CX_get_ext_file_prefix to use H5P_peek instead of H5P_get.
|
|
|
|
| |
available (can be controlled via a configure/CMake option)
|
|\
| |
| |
| |
| |
| |
| |
| |
| | |
~SONGYULU/hdf5_ray:TRILAB-98-dt_arith-and-cpp_testhdf5-tests to develop
* commit 'b02de315b93ac29d2483a91d526b110a25073505':
NNSA Tri-LabsTRILAB-98: Another two test cases out.
NNSA Tri-Labs TRILAB-98: Taking out a few more test cases.
NNSA Tri-Labs TRILAB-98 dt_arith and cpp_testhdf5 tests fail on sierra.llnl.gov: According to the group decision, simply provide a macro to disable some failing test cases on sierra (IBM power9 cpu). All failing cases involve long double data type.
|
| | |
|
| |
| |
| |
| | |
sierra.llnl.gov: According to the group decision, simply provide a macro to disable some failing test cases on sierra (IBM power9 cpu). All failing cases involve long double data type.
|
|/ |
|
|\
| |
| |
| |
| |
| |
| |
| | |
develop
Resolve conflicts:
Conflicts:
test/Makefile.am
|
| |\
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | | |
~DEROBINS/hdf5_der:openmpi_tweaks to develop
* commit 'd6c2a96ac2f103d90b96d5b39814810e6a31ef99':
Updated the parallel install docs.
Eliminated the need for a separate script variable.
Added a helpful message to the flush script.
Added a shell script so we can run the parallel flush test on OpenMPI.
|
| | | |
|
| | |\ |
|
| | | |
| | | |
| | | |
| | | | |
on OpenMPI.
|
| |\ \ \
| | | |/
| | |/| |
|
| | |/
| |/|
| | |
| | |
| | | |
Autotools only for right now, but this will be fleshed out in future
work.
|
| | | |
|
| |/
|/|
| |
| |
| |
| |
| |
| | |
after deleting attributes in densed storage.
The fix: When deleting attribute nodes from the name index v2 B-tree,
if an attribute is found in the intermediate B-tree nodes, which may be
merged/redistributed in the process, we need to free the dynamically
allocated spaces for the intermediate decoded attribute.
|
|/ |
|
| |
|
|\
| |
| |
| |
| |
| | |
* commit 'f03758613a6ec83e5ddd664590d162ccc888f13e':
Snapshot version 1.11 release 2.
Normalization with vol_integration (test code and H5Xtest.c)
|
| | |
|
|/ |
|
| |
|
| |
|
|\
| |
| |
| |
| |
| |
| |
| |
| |
| | |
~JHENDERSON/hdf5:bugfix/MPI2_only_fixes to develop
* commit '9aa2eaeb9ed9c94f63973cb55f13de256558497c':
Add Autotools and CMake checks for big I/O MPI-3 functions
Add hdf5settings section for parallel compression status in CMake builds
Switch to CheckSymbolExists in CMake
Add configure check for MPI_Mprobe and MPI_Imrecv functions
|
| | |
|
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| | |
Add line to libhdf5settings file for status of Parallel writes to
filtered datasets status
Surround Parallel Compression code in MPI_VERSION >= 3 checks
Add disabled message for Parallel Compression built w/ MPI-2
Modify Parallel Compression tests to only run the parallel filtered read
tests when parallel filtered writes are disabled
Update big I/O code to handle being built with MPI-2
Add checks to CMakeLists.txt for MPI_Mprobe and MPI_Imrecv
|