summaryrefslogtreecommitdiffstats
path: root/src/H5P.c
diff options
context:
space:
mode:
authorQuincey Koziol <koziol@hdfgroup.org>2002-05-17 12:54:42 (GMT)
committerQuincey Koziol <koziol@hdfgroup.org>2002-05-17 12:54:42 (GMT)
commite02ae065f6a2728d33af673e97405a3b409c9202 (patch)
tree5dd3fc9364b57b8e597adecbfa51988a28d9e177 /src/H5P.c
parenta6b4cba798a494dea1d29474cc5658f7003615d9 (diff)
downloadhdf5-e02ae065f6a2728d33af673e97405a3b409c9202.zip
hdf5-e02ae065f6a2728d33af673e97405a3b409c9202.tar.gz
hdf5-e02ae065f6a2728d33af673e97405a3b409c9202.tar.bz2
[svn-r5430] Purpose:
Document Bug fix/Code improvement below: Description: Currently, the chunk data allocation routine invoked to allocate space for the entire dataset is inefficient. It writes out each chunk in the dataset, whether it is already allocated or not. Additionally, this happens not only when it is created, but also anytime it is opened for writing, or the dataset is extended. Worse, there's too much parallel I/O syncronization, which slows things down even more. Solution: Only attempt to write out chunks that don't already exist. Additionally, share the I/O writing between all the nodes, instead of writing everything with process 0. Then, only block with MPI_Barrier if chunks were actually created.
Diffstat (limited to 'src/H5P.c')
0 files changed, 0 insertions, 0 deletions