| Commit message (Collapse) | Author | Age | Files | Lines |
|
|
|
|
|
| |
a) multi/split drivers and
b) persisting free-space or using paged aggregation strategy
because the library will fail file creation (temporary) for the above conditions.
|
|
|
|
| |
Test the changes in a branch via daily testing.
|
|\
| |
| |
| |
| |
| |
| | |
hdf5_1_10
* commit 'fc76c7785251482a26f31efabb10d38725facb6a':
Fix for HDFFV-10217 infinite loop in H5VM_power2up(). The function H5VM_power2up() returns the next power of 2 for n. When n exceeds 2^63, it overflows and becomes 0 causing the infinite looping. The fix ensures that the function checks for n >= 2^63 and returns 0.
|
| |
| |
| |
| |
| |
| | |
The function H5VM_power2up() returns the next power of 2 for n.
When n exceeds 2^63, it overflows and becomes 0 causing the infinite looping.
The fix ensures that the function checks for n >= 2^63 and returns 0.
|
|/ |
|
|
|
|
| |
Changes made so that raw data for dataset objects are copied from cached info when possible instead of flushing objects to file and read them back in again.
|
|
|
|
|
|
|
|
|
|
| |
Fixes for:
* HDFFV-10247 (autotools Fortran long double bug)
* HDFFV-10248 (autotools deprecated debug/production tweak)
* HDFFV-10250 (-fno-omit-frame-pointer w/ symbols)
* Disabled missing parameter warnings in Java constant code
* Added H5I_VFL ID checks to tmisc.c
* Added 'get a dummy VFD' call to h5test library
|
|
|
|
|
|
|
|
|
| |
* H5FD_FEAT_DEFAULT_VFD_COMPATIBLE VFD feature flag (HDFFV-10214)
* H5FDdriver_query() API call (HDFFV-10215)
* Fix to skip SWMR tests that when the current VFD is not compatible
with the default VFD (HDFFV-10197)
* Removed obsolete compiler versions from config/gnu-flags (HDFFV-9937)
* Many warning fixes
|
|
|
|
|
| |
changed to H5Dopen2 to fix failure with --with-default-api-version=v16
configure option.
|
|
|
|
|
| |
Modifications made based on the review comments from pull request #494
Tested on moohan, mayll, kituo, platypus, ostrich, osx1010test, quail, emu.
|
|
|
|
|
|
|
|
|
|
|
| |
(1) Repeated open/close of a compact dataset fails due to the
increment of ndims in the dataset structure for every open.
--This is done only for chunked dataset via H5D__chunk_set_sizes().
(2) layout "dirty" flag for a compact dataset is not reset
properly after flushing the data at dataset close.
--Reset the "dirty" flag before flushing the message to the object header
via H5O_msg_write().
Tested on moohan, kituo, platypus, ostrich, osx1010test, quail, emu.
|
| |
|
| |
|
| |
|
|
|
|
| |
the parallel library.
|
| |
|
| |
|
|\
| |
| |
| |
| |
| |
| |
| | |
* commit 'b67e626ba9095f35dd9a8f9e8ca906320f85914c':
HDFFV-10173 remove hook for noinst libs
HDFFV-10170 fix noinst for test libs
HDFFV-10173 fix compile on VS2008
Due to HDFVIEW-19 add read VL as string function
|
| | |
|
| | |
|
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| | |
Cimmit changes to Makefile.in(s) and H5PL.c that resulted from running
autogen.sh.
Conflicts:
README.txt
c++/src/Makefile.in
fortran/src/Makefile.in
hl/c++/src/Makefile.in
hl/fortran/src/Makefile.in
hl/src/Makefile.in
java/src/jni/Makefile.in
release_docs/RELEASE.txt
src/Makefile.in
|
| |
| |
| |
| |
| |
| |
| |
| |
| | |
hdf5_1_10_1
* commit '0d05cb607e5951f3a28a101be56ca194022a7fa1':
Reverted addition of The HDF Group copyright headers to 3rd party CMake files for Java.
Add Copyright headers to 90 files (mostly .cmake files) without Copyright headers.
Add "Copyright by The HDF Group."" to copyright header.
|
|/
|
|
|
|
|
|
| |
and replace it with new url for COPYING file.
Fix 2 lines in java error test expected output file where messages
include line numbers changed by reducing the copyright header by 2
lines.
|
|
|
|
|
| |
Skip the two tests for page buffering in test/fheap.c because this feature is disabled in parallel.
Activate full testing when this feature is re-enabled in the future for parallel build.
|
|\
| |
| |
| |
| |
| | |
* commit '2821b8bd768138db6b5f917a5b7325f878f8f7d0':
Fix URL name
DAILYTEST-250 change test props to reduce timeout
|
| | |
|
|/
|
|
|
| |
1) Calculate chunk index for extensive array index based on swizzled max chunks when unlim_dim > 0
2) Minor fixes to test/fheap.c that somehow were missed from last check in. See pull request #396 review comments.
|
|\
| |
| |
| |
| |
| |
| | |
* commit '4a05d27c6fcef81612e005b6f6429ab362247492':
HDFFV-10143 add missing javadoc param
Add missing test status
Fix CMake regex commands
|
| | |
|
|/
|
|
|
|
|
|
| |
buffering
Modify test/fheap.c to run with different combinations of file space strategies and page buffering
only when ExpressMode is 0 (HDF5TestExpress is 0).
Tested on ostrich, platypus, mayll, emu, osx1010test, quail, kite, kituo.
|
|\
| |
| |
| |
| | |
* commit '60bb3fd1108e2006a79b008d031a0f68a1e6b393':
Re-enabled fixed array index testing in the test_random_rank4_vl() test in test/set_extent. This was fixed some time ago, but the test was never re-enabled for that index type.
|
| |
| |
| |
| |
| | |
in test/set_extent. This was fixed some time ago, but the test was
never re-enabled for that index type.
|
| | |
|
| | |
|
| | |
|
|\ \
| |/
| |
| |
| |
| |
| |
| |
| |
| | |
* commit '2412158ed8326a3f3d62fbd947e470667d0b5951':
Add new file COPYING_LBNL_HDF5.
Revert "Clear hdf5 1.10 entries from RELEASE.txt in the develop branch. Entries"
Add LBNL license file and modify COPYING file accordingly.
Omnibus checkin for several relatively minor modifications:
Clear hdf5 1.10 entries from RELEASE.txt in the develop branch. Entries in this branch version of RELEASE.txt should be intended for the future 1.12.0 release.
Fix HDFFV-8089 Description: Some code within an "ifdef H5D_CHUNK_DEBUG" block was using outdated data structure but not caught because the case of H5D_CHUNK_DEBUG being defined was never tested. It was commented out. I defined H5D_CHUNK_DEBUG, tested, and commented out again. Platforms tested: Linux/32 2.6 (jam) Linux/64 (platypus) Darwin (osx1010test)
|
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| | |
1) Added code test/page_buffer.c to verify that page buffering is
disabled in parallel builds.
2) Added code to test/cache_image.c to verify correct interaction
between evict on close and cache image -- in particular management
of a file containing a cache image containing dirty metadata that
has been opened R/O. Also fix for the bug exposed.
3) Added code to testpar/t_cache_image.c to verify expected procedure
for reading cache images, and also supporting stats collection code
needed for the test.
4) Repair of an overactive sanity check in H5C__reconstruct_cache_contents().
5) Other minor tidies in passing.
Tested serial and parallel, debug and production on Jelly.
|
| | |
|
| | |
|
| | |
|
| | |
|
| | |
|
| | |
|
| | |
|
| | |
|
| | |
|
|/ |
|
|
|
|
| |
directories to avoid accidental deletion of the signal files.
|
|
|
|
| |
file deletion by other tests.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| |
Comment added to address HDFFV-8270. The sample ontput in the
file's comments are not up-to-date with the scripts in the file
that remove output unique to certain systems when running test
scripts. This output doesn't match expected output files for
the tests, causing them to fail. Ther output_filter.sh file
removes such output. Currently we don't have access to these
systems to update the comments.
testflushrefresh.sh.in:
flushrefresh and other SWMR tests have been failing
occasionally but regularly since bringing them to the develop
branch. The logs and debugging point to unexpected deletion of
signal files as the point of failure, and John Mainzer suggests
that two such tests running simultaneously in the same directory
may lead to the operating system deleting the other test's
signal file. Running the flushrefresh test in its own directory
seems to solve the problem.
|