summaryrefslogtreecommitdiffstats
path: root/src/H5Dio.c
Commit message (Collapse)AuthorAgeFilesLines
* [svn-r7218] Purpose:Quincey Koziol2003-07-141-16/+16
| | | | | | | | | | | | | | Bug fix Description: H5T_BKG_TEMP was accidentally removed from library code, but is used by application's datatype conversion routines. Solution: Revert removal of H5T_BKG_TEMP. Platforms tested: h5committest
* [svn-r7116] Purpose:Quincey Koziol2003-06-271-3/+4
| | | | | | | | | | | | Update code Description: Move "PABLO_MASK" definition above header file inclusion to avoid problems with inline functions in header files. Platforms tested: FreeBSD 4.8 (sleipnir) h5committest
* [svn-r7018] Purpose:Quincey Koziol2003-06-101-0/+8
| | | | | | | | | | | | | | | | | | | | | Bug fix Description: The dataset's modification time was getting set whenever raw data was written with H5Dwrite. Unfortunately, this is a metadata change (which are required to be performed collectively) and H5Dwrite may be called independently from a parallel program, resulting in metadata cache corruption and/or program hangs. Solution: Don't update the modification time when raw data it written. :-( Platforms tested: FreeBSD 4.8 (sleipnir) w/parallel h5committest Misc. update: Noted in release notes and also sent to Frank for updating the docs.
* [svn-r6975] Purpose:Quincey Koziol2003-06-051-8/+8
| | | | | | | | | | | Bug fix Description: Don't attempt to perform collective I/O on chunked datasets. Platforms tested: FreeBSD 4.8 (sleipnir) w/parallel h5committested
* [svn-r6951] Purpose:Quincey Koziol2003-06-041-6/+4
| | | | | | | | | | | | Code cleanup Description: Clean up various warnings & comment out unused code. Platforms tested: FreeBSD 4.8 (sleipnir) w/C++ FreeBSD 4.8 (sleipnir) w/parallel h5committested
* [svn-r6942] Purpose:Quincey Koziol2003-05-311-139/+511
| | | | | | | | | | | | | | | | | Performance improment Description: Speed up chunked dataset I/O. This breaks down into several areas: - Compute chunk selections in the file by using hyperslab operations instead of iterating over each element in the selection. - If the file and memory selections are the same shape, use the file chunk selections to compute the memory chunk selections. This required several additional dataspace, dataspace selection and hyperslab routines. Platforms tested: h5committestted (although Fortran tests failed for some reason)
* [svn-r6909] Purpose:Quincey Koziol2003-05-211-10/+14
| | | | | | | | | | | | Bug fix Description: Correct typo which didn't show up during my previous testing in production mode. Platforms tested: FreeBSD 4.8 (sleipnir) h5committest not necessary.
* [svn-r6908] Purpose:Quincey Koziol2003-05-211-6/+22
| | | | | | | | | | | Code cleanup Description: Fix some unused parameter warnings. Platforms tested: FreeBSD 4.8 (sleipnir) h5committest not necessary.
* [svn-r6907] Purpose:Quincey Koziol2003-05-201-0/+8
| | | | | | | | | | | | | | Bug fix Description: The chunk dataspace selection information for raw data I/O are being leaked. Solution: Free the chunk information during the cleanup code. Platforms tested: Solaris 2.6 (baldric) w/purify h5committest not needed.
* [svn-r6891] Purpose:Quincey Koziol2003-05-171-299/+639
| | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | Bug fix Description: Raw data I/O on chunked datasets would attempt to allocate data structures proportional to the number of chunks in the dataset on disk, instead of just the number of chunks that the I/O operation would interact with, causing operations on datasets with large #'s of chunks to fail (or become very slow), even though the actual I/O operation was very modest. Solution: This is the "scalability fix" for chunked datasets that I've mentioned we need to do, althought it's not the complete fix for the issue. Read on for the details... Only create data structures for the chunks that the I/O operation will actually act on, reducing the amount of information allocated in memory, normally. I say "normally", because this algorithm has the same problems as the original algorithm (worse actually, since the data structure for each chunk is larger now) if _all_ the chunks in a dataset with a lot of chunks are actually involved in the I/O operation. If that is the case, this code will fail in a similar way. To truly fix the problem, we would need to only create data structures for a particular number of chunks, perform the I/O on just those chunks, then release the data structures for those chunks and create data structures for the next set of chunks to access, etc. However, I think this case is pretty rare right now and we should worry about it after the 1.6.0 release. Platforms tested: h5committested
* [svn-r6860] Purpose:Quincey Koziol2003-05-131-8/+20
| | | | | | | | | | | | | Small code cleanup Description: Remove redundant sanity check. Added a note about a potential bug. Platforms tested: FreeBSD 4.8 (sleipnir) h5committest not necessary.
* [svn-r6857] Purpose: bug fixRaymond Lu2003-05-131-2/+18
| | | | | | | | | Description: when a datatype element size is bigger than I/O datatype conversion buffer(1MB), library reports failure. Solution: adjust conversion buffer to be at least equal to one element size Platforms tested: h5committest
* [svn-r6843] Purpose:Quincey Koziol2003-05-091-9/+3
| | | | | | | | | | | | Code cleanup Description: Clean up warnings exposed by compiling on O2K. Also, revert some of Bill and my changes to the H5S_mpi_opt_types_g, etc. and settle them back into their original location. Platforms tested: h5committested.
* [svn-r6841] Purpose:Bill Wendling2003-05-081-0/+5
| | | | | | | | | | | | | | | | Bug Fix Description: A variable wasn't declared in this file. Solution: Put the extern declaration of the variable in this file. It's defined in the H5S.c module. Platforms tested: Modi4 (This affects only parallel platforms) Misc. update:
* [svn-r6836] Purpose:Quincey Koziol2003-05-081-68/+72
| | | | | | | | | | Bug fix Description: Clean up small memory leak. Platforms tested: h5committested.
* [svn-r6829] Purpose:Quincey Koziol2003-05-081-8/+0
| | | | | | | | | | | Code cleanup. Description: Cleaned up an unusued variable and some commented out code. Platforms tested: FreeBSD 4.8 (sleipnir) Triple check not needed.
* [svn-r6825] Purpose:Quincey Koziol2003-05-071-0/+2293
New feature/enhancement Description: Chunked datasets are handled poorly in several circumstances involving certain selections and chunks that are too large for the chunk cache and/or chunks with filters, causing the chunk to be read from disk multiple times. Solution: Rearrange raw data I/O infrastructure to handle chunked datasets in a much more friendly way by creating a selection in memory and on disk for each chunk in a chunked dataset and performing all of the I/O on that chunk at one time. There are still some scalability (the current code attempts to create a selection for all the chunks in the dataset, instead of just the chunks that are accessed, requiring portions of the istore.c and fillval.c tests to be commented out) and performance issues, but checking this in will allow the changes to be tested by a much wider audience while I address the remaining issues. Platforms tested: h5committested, FreeBSD 4.8 (sleipnir) serial & parallel, Linux 2.4 (eirene)