| Commit message (Collapse) | Author | Age | Files | Lines |
| |
|
|
|
|
|
|
| |
HDFFV-5897.
Tested: jam, durandal (too minor for full h5committest)
|
|
|
|
|
|
|
|
|
|
|
| |
update for HDFFV-7837 - h5diff: incorrect behavior with exclude-path option when unique object exist only in one file
Description:
Add the new test output list to test shell script.
This is based on the previous commit r22114.
Tested:
jam (linux32-LE), koala (linux64-LE), ostrich (linuxppc64-BE)
|
| |
|
|
|
|
|
|
| |
structure in H5D_read as I did in H5D_write.
Tested on jam - simple change.
|
|
|
|
| |
Tested: koala (gnu 4.1)
|
|
|
|
| |
Added check that HLDS fortran function do remove NULL character.
|
|
|
|
|
|
| |
in h5lt_get_attribute_string_f
Fixed size +1 error in HD5f2cstring
|
|
|
|
|
|
|
|
|
|
| |
Description:
Closed various HDF5 objects in DataSet::getInMemDataSize and
Attribute::getInMemDataSize to remove some memory leaks.
Platforms tested:
Linux/32 2.6 (jam)
Linux/64 2.6 (amani)
SunOS 5.10 (linew)
|
| |
|
|
|
|
|
|
|
|
| |
memory and caused seg fault. I added checks in two places to make sure the library returns error stack
when it fails to allocate memory. I didn't add any test to the test suite since there is no good way to test it. But I tested and verified the error stack by hand.
Tested on jam, koala, ostrich.
|
|
|
|
|
|
|
|
|
|
|
|
|
| |
core VFD) and HDFFV-7603 (core VFD has trouble with 2GB+ files on Windows).
Propagates the SEC2 driver fixes from HDF5 1.8.8 to the core VFD (mainly concerning the backing store). These fixes also conveniently fixed 7603 as well.
Tested on:
64-bit Windows 7
jam
koala
ostrich
|
|
|
|
|
|
| |
private functions and a macro according to Quincey's suggestion.
Tested on jam - simple change.
|
|
|
|
|
|
|
| |
opened with the MPI-IO VFD
Add test cases for these two routines
Jira issue HDFFV-7961
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| |
Fix for HDFFV-7837 - h5diff: incorrect behavior with exclude-path option when unique object exist only in one file
Description:
If unique objects exists only in one file and try to exclude
the unique objects with --exclude-path option, h5diff missed
excluding some objects.
Fixed to exclude objects correctly in such case.
Test cases were added and tagged with jira#.
Tested:
jam (linux32-LE), koala (linux64-LE), ostrich (linuxppc64-BE), tejeda (mac32-LE), linew (solaris-BE), Windows (32-LE cmake), Cmake (jam)
|
|
|
|
|
|
|
|
| |
Minor code safety issue in test/fheap.c and whitespace in other files.
Tested on:
Mac OSX/64 10.7.3 (amazon) w/debug
(Too minor to require h5committest)
|
| |
|
| |
|
| |
|
|
|
|
|
|
|
|
|
|
|
| |
HDFFV-7523 JPSS: h5dump fails to open an attribute when using the -a option to specify an attribute with a slash in the name
Description:
Update command line help page and test generating file.
Previous commits: r22052, r22063
Tested:
jam (linux32-LE), koala (linux64-LE), cmake (jam)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| |
VFD is set).
This occurs due to the istore test creating very large files on systems which
do not have POSIX-like sparse file semantics. The large amount of I/O causes
the test to run for a very long period of time.
The fix was to copy the "big" test's sparse file check and only run the
largest sparse file test when POSIX-like sparse file semantics are found.
Tested on:
jam (nfs)
ostrich (nfs)
loyalty (ufs)
64-bit linux VM (ext4)
64-bit OS-X Lion (hfs, detected as not POSIX sparse)
64-bit Windows 7 (NTFS, detected as not POSIX sparse)
The OS-X failure to pass the sparse check is likely an error and will be
entered as a new bug.
|
|
|
|
| |
Remove "STAR" test as some platforms vary in interpretation.
|
| |
|
|
|
|
| |
Tested: jam (gnu and intel)
|
|
|
|
|
|
| |
(11.8-0) on jam and koala has trouble with the command "*p++ = *p OP tree_val" in the macro definition of H5Z_XFORM_DO_OP1 of H5Ztrans.c. It increments P first before doing the operation. So I break down the command into two lines: "*p = *p OP tree_val; p++;" I also reported the problem to PGI.
Tested on jam, koala, and ostrich.
|
| |
|
| |
|
|
|
|
| |
No test is needed.
|
| |
|
|
|
|
|
|
| |
Changed H5G*() to H5G*2() APIs for compatability.
Tested: (jam intel, --with-default-api-version=v16)
|
|
|
|
|
|
| |
above
tested: none
|
| |
|
|
|
|
|
|
|
|
|
| |
REF: HDFFV-1233
A high-level "H5LTpath_valid" routine for checking if a path is correct
Both Fortran and C version.
Tested: jam (gnu and intel)
|
|
|
|
| |
Tested: local linux
|
|
|
|
|
|
|
|
|
| |
operations like x*-100.
The parser mistaked "-" as substraction. I fixed it and also fixed another problem
with some special cases like 100-x and 2/x.
Tested on jam, koala, and ostrich.
|
|
|
|
| |
Tested: local linux
|
| |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| |
Description:
When using the new object header format, it was possible for corruption to occur
if the first object header chunk changed size such that the lenght of the "chunk
0 size" field changed. This only occurred if there were messages that had not
been decoded. The original algorithm that changed the object header chunk size
marked all messages as dirty, causing those that had not been decoded to have
both the raw and native form invalidated. Changed the algorithm to avoid
marking messages dirty and added assertions to catch the case where messages
are dirtied without being decoded (or recently created) first.
Tested: jam, koala, ostrich (h5committest), durandal
|
|
|
|
|
|
|
|
|
|
| |
Description:
When using the new object header format and adding an attribute with a size near
64K, it was possible for file corruption to occur. This happened only if the
first object header chunk was smaller than 256 bytes and then grew to larger
than 64K after the attribute was added.
Tested: ostrich, jam, koala (h5committest), durandal
|
|
|
|
|
|
| |
The test scripts for cmake, on windows, and the h5ls shell script do not work properly.
Tested: h5committest
|
| |
|
|
|
|
| |
cmake parsing
|
|
|
|
|
|
|
|
| |
Change to use versioned API call, to avoid problems when defaulting to
older AI versions.
Tested on:
Mac OSX/64 10.7.3 (amazon) w/debug
|
|
|
|
|
|
|
| |
Better fix for zero-sized dataset error (r22053).
Tested on:
Mac OSX/64 10.7.3 (amazon) w/debug & parallel
|
| |
|
|
|
|
|
|
|
|
|
|
| |
Correct corner case for creating a contiguous dataset with a zero-sized
dataspace, when the allocation time is set to early.
Also clean up a few compiler warnings in the dataspace code.
Tested on:
Mac OSX/64 10.7.3 (amazon) w/debug & parallel
|
|
|
|
|
|
|
|
| |
attributes containing "/"
by using an escape character. For example, for a dataset "/dset"
containing attribute "speed(m/h)", use "h5dump -a "/dset/speed(\/h)"
to show the content of the attribute. See details at HDFFV-7523
|
|
|
|
| |
runTest.cmake script
|
| |
|
|
|
|
|
| |
Add tools error stack and check return from H5Dreference call.
Tested: CMake and h5committest
|