summaryrefslogtreecommitdiffstats
diff options
context:
space:
mode:
authorAlbert Cheng <acheng@hdfgroup.org>2002-05-22 15:29:08 (GMT)
committerAlbert Cheng <acheng@hdfgroup.org>2002-05-22 15:29:08 (GMT)
commit5a7304a219af9d3f8a8a231fc399d7a77e2e3b35 (patch)
tree1af678faacf83021cda86a2f386e945cd5c96982
parent3808fa2002f36b7b4e43d6ee1a643a4b36967cdc (diff)
downloadhdf5-5a7304a219af9d3f8a8a231fc399d7a77e2e3b35.zip
hdf5-5a7304a219af9d3f8a8a231fc399d7a77e2e3b35.tar.gz
hdf5-5a7304a219af9d3f8a8a231fc399d7a77e2e3b35.tar.bz2
[svn-r5451] Purpose:
Merged in a bunch of changes Quincey made. Added some of mine here and there. Description: Below is Quincey's summary. The -r and -t changes were changed to "-D r" and "-D t" as they are of the debug purposes. Need to save option letters for hundreds of future features. The nofill for chunk storage is not supported since v1.4 does not have the feature. - Added option (-r) to display the throughput when doing I/O on just the dataset data. This is somewhat redundant for the posix I/O functions and MPI-I/O, but it's useful for HDF5. - Added option (-t) to display the times along with the throughputs. - Added options (-a <size> and -T <size>) to align objects in the HDF5 file. These are ignored for posix and MPI-I/O benchmarks. - Added option (-c) to create the HDF5 datasets using chunked storage. The chunks are created with the same size as the buffer size used for I/O, so all I/O to the dataset is chunk-aligned. - Changed code to allow 0-sized datasets to be created. This is useful when measuring the overhead for creating lots of datasets in HDF5. - Cleaned up code to eliminate warnings and modularize things better. Platforms tested: modi4 (pp)
-rw-r--r--perform/pio_engine.c229
-rw-r--r--perform/pio_perf.c349
-rw-r--r--perform/pio_perf.h6
-rw-r--r--perform/pio_timer.c93
-rw-r--r--perform/pio_timer.h18
5 files changed, 436 insertions, 259 deletions
diff --git a/perform/pio_engine.c b/perform/pio_engine.c
index 9475d79..54478a8 100644
--- a/perform/pio_engine.c
+++ b/perform/pio_engine.c
@@ -105,11 +105,11 @@ typedef union _file_descr {
/* local functions */
static char *pio_create_filename(iotype iot, const char *base_name,
char *fullname, size_t size);
-static herr_t do_write(file_descr *fd, iotype iot, long ndsets,
- long nelmts, size_t buf_size, void *buffer);
-static herr_t do_read(file_descr *fd, iotype iot, long ndsets,
- long nelmts, size_t buf_size, void *buffer /*out*/);
-static herr_t do_fopen(iotype iot, char *fname, file_descr *fd /*out*/,
+static herr_t do_write(results *res, file_descr *fd, parameters *parms,
+ long ndsets, off_t nelmts, size_t buf_size, void *buffer);
+static herr_t do_read(results *res, file_descr *fd, parameters *parms,
+ long ndsets, off_t nelmts, size_t buf_size, void *buffer /*out*/);
+static herr_t do_fopen(parameters *param, char *fname, file_descr *fd /*out*/,
int flags);
static herr_t do_fclose(iotype iot, file_descr *fd);
static void do_cleanupfile(iotype iot, char *fname);
@@ -135,7 +135,7 @@ do_pio(parameters param)
int maxprocs;
int nfiles, nf;
long ndsets;
- long nelmts;
+ off_t nelmts;
char *buffer = NULL; /*data buffer pointer */
size_t buf_size; /*data buffer size in bytes */
@@ -191,13 +191,6 @@ do_pio(parameters param)
GOTOERROR(FAIL);
}
- if (nelmts <= 0 ) {
- fprintf(stderr,
- "number of elements per dataset must be > 0 (%ld)\n",
- nelmts);
- GOTOERROR(FAIL);
- }
-
if (maxprocs <= 0 ) {
fprintf(stderr,
"maximum number of process to use must be > 0 (%d)\n",
@@ -205,11 +198,6 @@ do_pio(parameters param)
GOTOERROR(FAIL);
}
- if (buf_size <= 0 ){
- fprintf(stderr,
- "buffer size must be > 0 (%ld)\n", buf_size);
- GOTOERROR(FAIL);
- }
#if akcdebug
/* debug*/
@@ -227,12 +215,14 @@ buf_size=MIN(1024*1024, buf_size);
#endif
/* allocate data buffer */
- buffer = malloc(buf_size);
+ if(buf_size>0) {
+ buffer = malloc(buf_size);
- if (buffer == NULL){
- fprintf(stderr, "malloc for data buffer size (%ld) failed\n",
- buf_size);
- GOTOERROR(FAIL);
+ if (buffer == NULL){
+ fprintf(stderr, "malloc for data buffer size (%ld) failed\n",
+ buf_size);
+ GOTOERROR(FAIL);
+ }
}
if (pio_debug_level >= 4) {
@@ -261,12 +251,12 @@ fprintf(stderr, "filename=%s\n", fname);
#endif
set_time(res.timers, HDF5_GROSS_WRITE_FIXED_DIMS, START);
- hrc = do_fopen(iot, fname, &fd, PIO_CREATE | PIO_WRITE);
+ hrc = do_fopen(&param, fname, &fd, PIO_CREATE | PIO_WRITE);
VRFY((hrc == SUCCESS), "do_fopen failed");
set_time(res.timers, HDF5_FINE_WRITE_FIXED_DIMS, START);
- hrc = do_write(&fd, iot, ndsets, nelmts, buf_size, buffer);
+ hrc = do_write(&res, &fd, &param, ndsets, nelmts, buf_size, buffer);
set_time(res.timers, HDF5_FINE_WRITE_FIXED_DIMS, STOP);
VRFY((hrc == SUCCESS), "do_write failed");
@@ -284,12 +274,12 @@ fprintf(stderr, "filename=%s\n", fname);
*/
/* Open file for read */
set_time(res.timers, HDF5_GROSS_READ_FIXED_DIMS, START);
- hrc = do_fopen(iot, fname, &fd, PIO_READ);
+ hrc = do_fopen(&param, fname, &fd, PIO_READ);
VRFY((hrc == SUCCESS), "do_fopen failed");
set_time(res.timers, HDF5_FINE_READ_FIXED_DIMS, START);
- hrc = do_read(&fd, iot, ndsets, nelmts, buf_size, buffer);
+ hrc = do_read(&res, &fd, &param, ndsets, nelmts, buf_size, buffer);
set_time(res.timers, HDF5_FINE_READ_FIXED_DIMS, STOP);
VRFY((hrc == SUCCESS), "do_read failed");
@@ -326,7 +316,8 @@ done:
}
/* release generic resources */
- free(buffer);
+ if(buffer)
+ free(buffer);
res.ret_code = ret_code;
return res;
}
@@ -450,23 +441,25 @@ pio_create_filename(iotype iot, const char *base_name, char *fullname, size_t si
* Modifications:
*/
static herr_t
-do_write(file_descr *fd, iotype iot, long ndsets,
- long nelmts, size_t buf_size, void *buffer)
+do_write(results *res, file_descr *fd, parameters *parms, long ndsets,
+ off_t nelmts, size_t buf_size, void *buffer)
{
int ret_code = SUCCESS;
- int rc; /*routine return code */
+ long rc; /*routine return code */
int mrc; /*MPI return code */
MPI_Offset mpi_offset;
MPI_Status mpi_status;
long ndset;
- long nelmts_towrite, nelmts_written;
+ off_t nelmts_written;
+ size_t nelmts_towrite;
char dname[64];
off_t dset_offset; /*dataset offset in a file */
off_t file_offset; /*file offset of the next transfer */
off_t dset_size; /*one dataset size in bytes */
- long nelmts_in_buf;
- long elmts_begin; /*first elmt this process transfer */
- long elmts_count; /*number of elmts this process transfer */
+ size_t nelmts_in_buf;
+ off_t elmts_begin; /*first elmt this process transfer */
+ off_t elmts_count; /*number of elmts this process transfer */
+ hid_t dcpl = -1; /* Dataset creation property list */
/* HDF5 variables */
herr_t hrc; /*HDF5 return code */
@@ -487,16 +480,28 @@ fprintf(stderr, "buffer size=%ld\n", buf_size);
nelmts_in_buf = buf_size/ELMT_SIZE;
/* hdf5 data space setup */
- if (iot == PHDF5){
- /* define a contiquous dataset of nelmts native ints */
- h5dims[0] = nelmts;
- h5dset_space_id = H5Screate_simple(1, h5dims, NULL);
- VRFY((h5dset_space_id >= 0), "H5Screate_simple");
+ if (parms->io_type == PHDF5){
+ if(nelmts>0) {
+ /* define a contiquous dataset of nelmts native ints */
+ h5dims[0] = nelmts;
+ h5dset_space_id = H5Screate_simple(1, h5dims, NULL);
+ VRFY((h5dset_space_id >= 0), "H5Screate_simple");
+ } /* end if */
+ else {
+ h5dset_space_id = H5Screate(H5S_SCALAR);
+ VRFY((h5dset_space_id >= 0), "H5Screate");
+ } /* end else */
/* create the memory dataspace that corresponds to the xfer buffer */
- h5dims[0] = nelmts_in_buf;
- h5mem_space_id = H5Screate_simple(1, h5dims, NULL);
- VRFY((h5mem_space_id >= 0), "H5Screate_simple");
+ if(nelmts_in_buf>0) {
+ h5dims[0] = nelmts_in_buf;
+ h5mem_space_id = H5Screate_simple(1, h5dims, NULL);
+ VRFY((h5mem_space_id >= 0), "H5Screate_simple");
+ } /* end if */
+ else {
+ h5mem_space_id = H5Screate(H5S_SCALAR);
+ VRFY((h5mem_space_id >= 0), "H5Screate");
+ } /* end else */
}
for (ndset = 1; ndset <= ndsets; ++ndset) {
@@ -504,7 +509,7 @@ fprintf(stderr, "buffer size=%ld\n", buf_size);
/* Calculate dataset offset within a file */
/* create dataset */
- switch (iot) {
+ switch (parms->io_type) {
case POSIXIO:
case MPIO:
/* both posix and mpi io just need dataset offset in file*/
@@ -512,15 +517,50 @@ fprintf(stderr, "buffer size=%ld\n", buf_size);
break;
case PHDF5:
+ dcpl = H5Pcreate(H5P_DATASET_CREATE);
+ if (dcpl < 0) {
+ fprintf(stderr, "HDF5 Property List Create failed\n");
+ GOTOERROR(FAIL);
+ }
+
+ /* Make the dataset chunked if asked */
+ if(parms->h5_use_chunks) {
+ /* Set the chunk size to be the same as the buffer size */
+ h5dims[0] = nelmts_in_buf;
+ hrc = H5Pset_chunk(dcpl, 1, h5dims);
+ if (hrc < 0) {
+ fprintf(stderr, "HDF5 Property List Set failed\n");
+ GOTOERROR(FAIL);
+ } /* end if */
+ } /* end if */
+
+#if H5_VERS_MAJOR > 1 || H5_VERS_MINOR > 4
+ /* Disable writing fill values if asked */
+ if(parms->h5_no_fill) {
+ hrc = H5Pset_fill_time(dcpl, H5D_FILL_TIME_NEVER);
+ if (hrc < 0) {
+ fprintf(stderr, "HDF5 Property List Set failed\n");
+ GOTOERROR(FAIL);
+ } /* end if */
+ } /* end if */
+#endif
+
sprintf(dname, "Dataset_%ld", ndset);
h5ds_id = H5Dcreate(fd->h5fd, dname, H5T_NATIVE_INT,
- h5dset_space_id, H5P_DEFAULT);
+ h5dset_space_id, dcpl);
if (h5ds_id < 0) {
fprintf(stderr, "HDF5 Dataset Create failed\n");
GOTOERROR(FAIL);
}
+ hrc = H5Pclose(dcpl);
+ /* verifying the close of the dcpl */
+ if (hrc < 0) {
+ fprintf(stderr, "HDF5 Property List Close failed\n");
+ GOTOERROR(FAIL);
+ }
+
break;
}
@@ -529,11 +569,11 @@ fprintf(stderr, "buffer size=%ld\n", buf_size);
* and the next process. Count of elements is the difference between
* these two beginnings. This way, it avoids any rounding errors.
*/
- elmts_begin = (nelmts*1.0)/pio_mpi_nprocs_g*pio_mpi_rank_g;
+ elmts_begin = (long)(((double)nelmts)/pio_mpi_nprocs_g*pio_mpi_rank_g);
if (pio_mpi_rank_g < (pio_mpi_nprocs_g - 1))
- elmts_count = ((nelmts * 1.0) / pio_mpi_nprocs_g * (pio_mpi_rank_g + 1))
- - elmts_begin;
+ elmts_count = (long)((((double)nelmts) / pio_mpi_nprocs_g * (pio_mpi_rank_g + 1))
+ - (double)elmts_begin);
else
/* last process. Take whatever are left */
elmts_count = nelmts - elmts_begin;
@@ -545,10 +585,13 @@ fprintf(stderr, "proc %d: elmts_begin=%ld, elmts_count=%ld\n",
nelmts_written = 0 ;
+ /* Start "raw data" write timer */
+ set_time(res->timers, HDF5_RAW_WRITE_FIXED_DIMS, START);
+
while (nelmts_written < elmts_count){
nelmts_towrite = elmts_count - nelmts_written;
- if (elmts_count - nelmts_written >= nelmts_in_buf) {
+ if ((elmts_count - nelmts_written) >= nelmts_in_buf) {
nelmts_towrite = nelmts_in_buf;
} else {
/* last write of a partial buffer */
@@ -568,7 +611,7 @@ fprintf(stderr, "proc %d: elmts_begin=%ld, elmts_count=%ld\n",
/* Write */
/* Calculate offset of write within a dataset/file */
- switch (iot) {
+ switch (parms->io_type) {
case POSIXIO:
/* need to (off_t) the elmnts_begin expression because they */
/* may be of smaller sized integer types */
@@ -594,7 +637,7 @@ fprintf(stderr, "proc %d: writes %ld bytes at mpi-offset %ld\n",
#endif
mrc = MPI_File_write_at(fd->mpifd, mpi_offset, buffer,
- nelmts_towrite * ELMT_SIZE, MPI_CHAR,
+ (int)(nelmts_towrite*ELMT_SIZE), MPI_CHAR,
&mpi_status);
VRFY((mrc==MPI_SUCCESS), "MPIO_WRITE");
break;
@@ -628,10 +671,13 @@ fprintf(stderr, "proc %d: writes %ld bytes at mpi-offset %ld\n",
nelmts_written += nelmts_towrite;
}
+ /* Stop "raw data" write timer */
+ set_time(res->timers, HDF5_RAW_WRITE_FIXED_DIMS, STOP);
+
/* Calculate write time */
/* Close dataset. Only HDF5 needs to do an explicit close. */
- if (iot == PHDF5){
+ if (parms->io_type == PHDF5){
hrc = H5Dclose(h5ds_id);
if (hrc < 0) {
@@ -676,23 +722,24 @@ done:
* Modifications:
*/
static herr_t
-do_read(file_descr *fd, iotype iot, long ndsets,
- long nelmts, size_t buf_size, void *buffer /*out*/)
+do_read(results *res, file_descr *fd, parameters *parms, long ndsets,
+ off_t nelmts, size_t buf_size, void *buffer /*out*/)
{
int ret_code = SUCCESS;
- int rc; /*routine return code */
+ long rc; /*routine return code */
int mrc; /*MPI return code */
MPI_Offset mpi_offset;
MPI_Status mpi_status;
long ndset;
- long nelmts_toread, nelmts_read;
+ size_t nelmts_toread;
+ off_t nelmts_read;
char dname[64];
off_t dset_offset; /*dataset offset in a file */
off_t file_offset; /*file offset of the next transfer */
off_t dset_size; /*one dataset size in bytes */
- long nelmts_in_buf;
- long elmts_begin; /*first elmt this process transfer */
- long elmts_count; /*number of elmts this process transfer */
+ size_t nelmts_in_buf;
+ off_t elmts_begin; /*first elmt this process transfer */
+ off_t elmts_count; /*number of elmts this process transfer */
/* HDF5 variables */
herr_t hrc; /*HDF5 return code */
@@ -713,23 +760,35 @@ fprintf(stderr, "buffer size=%ld\n", buf_size);
nelmts_in_buf = buf_size/ELMT_SIZE;
/* hdf5 data space setup */
- if (iot == PHDF5){
- /* define a contiquous dataset of nelmts native ints */
- h5dims[0] = nelmts;
- h5dset_space_id = H5Screate_simple(1, h5dims, NULL);
- VRFY((h5dset_space_id >= 0), "H5Screate_simple");
+ if (parms->io_type == PHDF5){
+ if(nelmts>0) {
+ /* define a contiquous dataset of nelmts native ints */
+ h5dims[0] = nelmts;
+ h5dset_space_id = H5Screate_simple(1, h5dims, NULL);
+ VRFY((h5dset_space_id >= 0), "H5Screate_simple");
+ } /* end if */
+ else {
+ h5dset_space_id = H5Screate(H5S_SCALAR);
+ VRFY((h5dset_space_id >= 0), "H5Screate");
+ } /* end else */
/* create the memory dataspace that corresponds to the xfer buffer */
- h5dims[0] = nelmts_in_buf;
- h5mem_space_id = H5Screate_simple(1, h5dims, NULL);
- VRFY((h5mem_space_id >= 0), "H5Screate_simple");
+ if(nelmts_in_buf>0) {
+ h5dims[0] = nelmts_in_buf;
+ h5mem_space_id = H5Screate_simple(1, h5dims, NULL);
+ VRFY((h5mem_space_id >= 0), "H5Screate_simple");
+ } /* end if */
+ else {
+ h5mem_space_id = H5Screate(H5S_SCALAR);
+ VRFY((h5mem_space_id >= 0), "H5Screate");
+ } /* end else */
}
for (ndset = 1; ndset <= ndsets; ++ndset) {
/* Calculate dataset offset within a file */
/* create dataset */
- switch (iot) {
+ switch (parms->io_type) {
case POSIXIO:
case MPIO:
/* both posix and mpi io just need dataset offset in file*/
@@ -753,11 +812,11 @@ fprintf(stderr, "buffer size=%ld\n", buf_size);
* and the next process. Count of elements is the difference between
* these two beginnings. This way, it avoids any rounding errors.
*/
- elmts_begin = (nelmts*1.0)/pio_mpi_nprocs_g*pio_mpi_rank_g;
+ elmts_begin = (long)(((double)nelmts)/pio_mpi_nprocs_g*pio_mpi_rank_g);
if (pio_mpi_rank_g < (pio_mpi_nprocs_g - 1))
- elmts_count = ((nelmts * 1.0) / pio_mpi_nprocs_g * (pio_mpi_rank_g + 1)) -
- elmts_begin;
+ elmts_count = (long)((((double)nelmts) / pio_mpi_nprocs_g * (pio_mpi_rank_g + 1)) -
+ (double)elmts_begin);
else
/* last process. Take whatever are left */
elmts_count = nelmts - elmts_begin;
@@ -769,10 +828,13 @@ fprintf(stderr, "proc %d: elmts_begin=%ld, elmts_count=%ld\n",
nelmts_read = 0 ;
+ /* Start "raw data" read timer */
+ set_time(res->timers, HDF5_RAW_READ_FIXED_DIMS, START);
+
while (nelmts_read < elmts_count){
nelmts_toread = elmts_count - nelmts_read;
- if (elmts_count - nelmts_read >= nelmts_in_buf)
+ if ((elmts_count - nelmts_read) >= nelmts_in_buf)
nelmts_toread = nelmts_in_buf;
else
/* last read of a partial buffer */
@@ -780,7 +842,7 @@ fprintf(stderr, "proc %d: elmts_begin=%ld, elmts_count=%ld\n",
/* read */
/* Calculate offset of read within a dataset/file */
- switch (iot){
+ switch (parms->io_type){
case POSIXIO:
/* need to (off_t) the elmnts_begin expression because they */
/* may be of smaller sized integer types */
@@ -806,7 +868,7 @@ fprintf(stderr, "proc %d: read %ld bytes at mpi-offset %ld\n",
#endif
mrc = MPI_File_read_at(fd->mpifd, mpi_offset, buffer,
- nelmts_toread*ELMT_SIZE, MPI_CHAR,
+ (int)(nelmts_toread*ELMT_SIZE), MPI_CHAR,
&mpi_status);
VRFY((mrc==MPI_SUCCESS), "MPIO_read");
break;
@@ -853,10 +915,13 @@ fprintf(stderr, "proc %d: read %ld bytes at mpi-offset %ld\n",
nelmts_read += nelmts_toread;
}
+ /* Stop "raw data" read timer */
+ set_time(res->timers, HDF5_RAW_READ_FIXED_DIMS, STOP);
+
/* Calculate read time */
/* Close dataset. Only HDF5 needs to do an explicit close. */
- if (iot == PHDF5){
+ if (parms->io_type == PHDF5){
hrc = H5Dclose(h5ds_id);
if (hrc < 0) {
@@ -901,13 +966,13 @@ done:
* Modifications:
*/
static herr_t
-do_fopen(iotype iot, char *fname, file_descr *fd /*out*/, int flags)
+do_fopen(parameters *param, char *fname, file_descr *fd /*out*/, int flags)
{
int ret_code = SUCCESS, mrc;
herr_t hrc;
hid_t acc_tpl = -1; /* file access templates */
- switch (iot) {
+ switch (param->io_type) {
case POSIXIO:
if (flags & (PIO_CREATE | PIO_WRITE))
fd->posixfd = POSIXCREATE(fname);
@@ -920,7 +985,7 @@ do_fopen(iotype iot, char *fname, file_descr *fd /*out*/, int flags)
}
- /* The perils of POSIX I/O API in a parallel environment. The problem is:
+ /* The perils of POSIX I/O in a parallel environment. The problem is:
*
* - Process n opens a file with truncation and then starts
* writing to the file.
@@ -965,14 +1030,20 @@ do_fopen(iotype iot, char *fname, file_descr *fd /*out*/, int flags)
case PHDF5:
acc_tpl = H5Pcreate(H5P_FILE_ACCESS);
-
if (acc_tpl < 0) {
fprintf(stderr, "HDF5 Property List Create failed\n");
GOTOERROR(FAIL);
}
+ /* Set the file driver to the MPI-I/O driver */
hrc = H5Pset_fapl_mpio(acc_tpl, pio_comm_g, pio_info_g);
+ if (hrc < 0) {
+ fprintf(stderr, "HDF5 Property List Set failed\n");
+ GOTOERROR(FAIL);
+ }
+ /* Set the alignment of objects in HDF5 file */
+ hrc = H5Pset_alignment(acc_tpl, param->h5_thresh, param->h5_align);
if (hrc < 0) {
fprintf(stderr, "HDF5 Property List Set failed\n");
GOTOERROR(FAIL);
diff --git a/perform/pio_perf.c b/perform/pio_perf.c
index 03bc915..c3c43e8 100644
--- a/perform/pio_perf.c
+++ b/perform/pio_perf.c
@@ -76,7 +76,8 @@
#define PIO_MPI 0x20
#define PIO_HDF5 0x40
-#define MB_PER_SEC(bytes,t) (((bytes) / ONE_MB) / t)
+/* report 0.0 in case t is zero too */
+#define MB_PER_SEC(bytes,t) (((t)==0.0) ? 0.0 : ((((double)bytes) / ONE_MB) / (t)))
/* global variables */
FILE *output; /* output file */
@@ -102,16 +103,16 @@ static const char *progname = "pio_perf";
* adding more, make sure that they don't clash with each other.
*/
#if 1
-static const char *s_opts = "hD:f:HP:p:X:x:md:F:i:o:r";
+static const char *s_opts = "hD:f:HP:p:X:x:md:F:i:o:rsta:T:cn";
#else
-static const char *s_opts = "hbD:f:HP:p:X:x:md:F:i:o:r";
+static const char *s_opts = "hbD:f:HP:p:X:x:md:F:i:o:rsta:T:cn";
#endif /* 1 */
static struct long_options l_opts[] = {
{ "help", no_arg, 'h' },
{ "hel", no_arg, 'h' },
{ "he", no_arg, 'h' },
#if 0
- /* a siting of the elusive binary option */
+ /* a sighting of the elusive binary option */
{ "binary", no_arg, 'b' },
{ "binar", no_arg, 'b' },
{ "bina", no_arg, 'b' },
@@ -198,6 +199,28 @@ static struct long_options l_opts[] = {
{ "ou", require_arg, 'o' },
{ "raw", no_arg, 'r' },
{ "ra", no_arg, 'r' },
+ { "align", require_arg, 'a' },
+ { "alig", require_arg, 'a' },
+ { "ali", require_arg, 'a' },
+ { "al", require_arg, 'a' },
+ { "threshold", require_arg, 'T' },
+ { "threshol", require_arg, 'T' },
+ { "thresho", require_arg, 'T' },
+ { "thresh", require_arg, 'T' },
+ { "thres", require_arg, 'T' },
+ { "thre", require_arg, 'T' },
+ { "thr", require_arg, 'T' },
+ { "th", require_arg, 'T' },
+ { "chunk", no_arg, 'c' },
+ { "chun", no_arg, 'c' },
+ { "chu", no_arg, 'c' },
+ { "ch", no_arg, 'c' },
+ { "no-fill", no_arg, 'n' },
+ { "no-fil", no_arg, 'n' },
+ { "no-fi", no_arg, 'n' },
+ { "no-f", no_arg, 'n' },
+ { "no-", no_arg, 'n' },
+ { "no", no_arg, 'n' },
{ NULL, 0, '\0' }
};
@@ -208,10 +231,16 @@ struct options {
long num_dsets; /* number of datasets */
long num_files; /* number of files */
long num_iters; /* number of iterations */
- long max_num_procs; /* maximum number of processes to use */
- long min_num_procs; /* minimum number of processes to use */
+ int max_num_procs; /* maximum number of processes to use */
+ int min_num_procs; /* minimum number of processes to use */
size_t max_xfer_size; /* maximum transfer buffer size */
size_t min_xfer_size; /* minimum transfer buffer size */
+ int print_times; /* print times as well as throughputs */
+ int print_raw; /* print raw data throughput info */
+ off_t h5_alignment; /* alignment in HDF5 file */
+ off_t h5_threshold; /* threshold for alignment in HDF5 file */
+ int h5_use_chunks; /* Make HDF5 dataset chunked */
+ int h5_no_fill; /* Disable HDF5 writing fill values */
};
typedef struct _minmax {
@@ -225,12 +254,13 @@ typedef struct _minmax {
static off_t parse_size_directive(const char *size);
static struct options *parse_command_line(int argc, char *argv[]);
static void run_test_loop(struct options *options);
-static int run_test(iotype iot, parameters parms);
+static int run_test(iotype iot, parameters parms, struct options *opts);
static void output_all_info(minmax *mm, int count, int indent_level);
static void get_minmax(minmax *mm, double val);
-static minmax accumulate_minmax_stuff(minmax *mm, off_t raw_size, int count);
+static minmax accumulate_minmax_stuff(minmax *mm, int count);
static int create_comm_world(int num_procs, int *doing_pio);
static int destroy_comm_world(void);
+static void output_results(const struct options *options, const char *name, minmax *table, int table_size,off_t data_size);
static void output_report(const char *fmt, ...);
static void print_indent(register int indent);
static void usage(const char *prog);
@@ -331,7 +361,7 @@ static void
run_test_loop(struct options *opts)
{
parameters parms;
- long num_procs;
+ int num_procs;
int doing_pio; /* if this process is doing PIO */
int io_runs = PIO_HDF5 | PIO_MPI | PIO_POSIX; /* default to run all tests */
@@ -352,6 +382,10 @@ run_test_loop(struct options *opts)
parms.num_files = opts->num_files;
parms.num_dsets = opts->num_dsets;
parms.num_iters = opts->num_iters;
+ parms.h5_align = opts->h5_alignment;
+ parms.h5_thresh = opts->h5_threshold;
+ parms.h5_use_chunks = opts->h5_use_chunks;
+ parms.h5_no_fill = opts->h5_no_fill;
/* start with max_num_procs and decrement it by half for each loop. */
/* if performance needs restart, fewer processes may be needed. */
@@ -373,24 +407,28 @@ run_test_loop(struct options *opts)
for (buf_size = opts->min_xfer_size;
buf_size <= opts->max_xfer_size; buf_size <<= 1) {
parms.buf_size = buf_size;
- parms.num_elmts = opts->file_size / (parms.num_dsets * sizeof(int));
+ parms.num_elmts = opts->file_size / (off_t)(parms.num_dsets * sizeof(int));
print_indent(1);
output_report("Transfer Buffer Size: %ld bytes, File size: %.2f MBs\n",
buf_size,
- ((double)parms.num_dsets * parms.num_elmts * sizeof(int)) / ONE_MB);
+ ((double)parms.num_dsets * (double)parms.num_elmts * (double)sizeof(int)) / ONE_MB);
print_indent(1);
output_report(" # of files: %ld, # of dsets: %ld, # of elmts per dset: %ld\n",
parms.num_files, parms.num_dsets, parms.num_elmts);
if (io_runs & PIO_POSIX)
- run_test(POSIXIO, parms);
+ run_test(POSIXIO, parms, opts);
if (io_runs & PIO_MPI)
- run_test(MPIO, parms);
+ run_test(MPIO, parms, opts);
if (io_runs & PIO_HDF5)
- run_test(PHDF5, parms);
+ run_test(PHDF5, parms, opts);
+
+ /* Run the tests once if buf_size==0, but then break out */
+ if(buf_size==0)
+ break;
}
if (destroy_comm_world() != SUCCESS) {
@@ -408,27 +446,30 @@ run_test_loop(struct options *opts)
* Modifications:
*/
static int
-run_test(iotype iot, parameters parms)
+run_test(iotype iot, parameters parms, struct options *opts)
{
results res;
register int i, ret_value = SUCCESS;
int comm_size;
off_t raw_size;
- minmax total_mm;
- minmax *write_mpi_mm_table;
- minmax *write_mm_table;
- minmax *write_gross_mm_table;
- minmax *read_mpi_mm_table;
- minmax *read_mm_table;
- minmax *read_gross_mm_table;
+ minmax *write_mpi_mm_table=NULL;
+ minmax *write_mm_table=NULL;
+ minmax *write_gross_mm_table=NULL;
+ minmax *write_raw_mm_table=NULL;
+ minmax *read_mpi_mm_table=NULL;
+ minmax *read_mm_table=NULL;
+ minmax *read_gross_mm_table=NULL;
+ minmax *read_raw_mm_table=NULL;
minmax write_mpi_mm = {0.0, 0.0, 0.0, 0};
minmax write_mm = {0.0, 0.0, 0.0, 0};
minmax write_gross_mm = {0.0, 0.0, 0.0, 0};
+ minmax write_raw_mm = {0.0, 0.0, 0.0, 0};
minmax read_mpi_mm = {0.0, 0.0, 0.0, 0};
minmax read_mm = {0.0, 0.0, 0.0, 0};
minmax read_gross_mm = {0.0, 0.0, 0.0, 0};
+ minmax read_raw_mm = {0.0, 0.0, 0.0, 0};
- raw_size = parms.num_dsets * parms.num_elmts * sizeof(int);
+ raw_size = (off_t)parms.num_dsets * (off_t)parms.num_elmts * (off_t)sizeof(int);
parms.io_type = iot;
print_indent(2);
output_report("IO API = ");
@@ -447,46 +488,18 @@ run_test(iotype iot, parameters parms)
MPI_Comm_size(pio_comm_g, &comm_size);
- write_mpi_mm_table = malloc(parms.num_iters * sizeof(minmax));
- write_mm_table = malloc(parms.num_iters * sizeof(minmax));
- write_gross_mm_table = malloc(parms.num_iters * sizeof(minmax));
- read_mpi_mm_table = malloc(parms.num_iters * sizeof(minmax));
- read_mm_table = malloc(parms.num_iters * sizeof(minmax));
- read_gross_mm_table = malloc(parms.num_iters * sizeof(minmax));
-
- for (i = 0; i < parms.num_iters; ++i) {
- write_mpi_mm_table[i].min = 0.0;
- write_mpi_mm_table[i].max = 0.0;
- write_mpi_mm_table[i].sum = 0.0;
- write_mpi_mm_table[i].num = 0;
-
- write_mm_table[i].min = 0.0;
- write_mm_table[i].max = 0.0;
- write_mm_table[i].sum = 0.0;
- write_mm_table[i].num = 0;
-
- write_gross_mm_table[i].min = 0.0;
- write_gross_mm_table[i].max = 0.0;
- write_gross_mm_table[i].sum = 0.0;
- write_gross_mm_table[i].num = 0;
-
- read_mpi_mm_table[i].min = 0.0;
- read_mpi_mm_table[i].max = 0.0;
- read_mpi_mm_table[i].sum = 0.0;
- read_mpi_mm_table[i].num = 0;
-
- read_mm_table[i].min = 0.0;
- read_mm_table[i].max = 0.0;
- read_mm_table[i].sum = 0.0;
- read_mm_table[i].num = 0;
-
- read_gross_mm_table[i].min = 0.0;
- read_gross_mm_table[i].max = 0.0;
- read_gross_mm_table[i].sum = 0.0;
- read_gross_mm_table[i].num = 0;
- }
-
- /* Do IO iteration times, collecting statics each time */
+ /* allocate space for tables minmax and that it is sufficient */
+ /* to initialize all elements to zeros by calloc. */
+ write_mpi_mm_table = calloc(parms.num_iters , sizeof(minmax));
+ write_mm_table = calloc(parms.num_iters , sizeof(minmax));
+ write_gross_mm_table = calloc(parms.num_iters , sizeof(minmax));
+ write_raw_mm_table = calloc(parms.num_iters , sizeof(minmax));
+ read_mpi_mm_table = calloc(parms.num_iters , sizeof(minmax));
+ read_mm_table = calloc(parms.num_iters , sizeof(minmax));
+ read_gross_mm_table = calloc(parms.num_iters , sizeof(minmax));
+ read_raw_mm_table = calloc(parms.num_iters , sizeof(minmax));
+
+ /* Do IO iteration times, collecting statistics each time */
for (i = 0; i < parms.num_iters; ++i) {
double t;
@@ -511,6 +524,12 @@ run_test(iotype iot, parameters parms)
write_gross_mm_table[i] = write_gross_mm;
+ /* gather all of the raw "write" times */
+ t = get_time(res.timers, HDF5_RAW_WRITE_FIXED_DIMS);
+ get_minmax(&write_raw_mm, t);
+
+ write_raw_mm_table[i] = write_raw_mm;
+
/* gather all of the "mpi read" times */
t = get_time(res.timers, HDF5_MPI_READ);
get_minmax(&read_mpi_mm, t);
@@ -528,12 +547,32 @@ run_test(iotype iot, parameters parms)
get_minmax(&read_gross_mm, t);
read_gross_mm_table[i] = read_gross_mm;
+
+ /* gather all of the raw "read" times */
+ t = get_time(res.timers, HDF5_RAW_READ_FIXED_DIMS);
+ get_minmax(&read_raw_mm, t);
+
+ read_raw_mm_table[i] = read_raw_mm;
pio_time_destroy(res.timers);
}
/*
- * Show various statics
+ * Show various statistics
*/
+ /* Write statistics */
+ /* Print the raw data throughput if desired */
+ if(opts->print_raw) {
+ /* accumulate and output the max, min, and average "raw write" times */
+ if (pio_debug_level >= 3) {
+ /* output all of the times for all iterations */
+ print_indent(3);
+ output_report("Raw Data Write details:\n");
+ output_all_info(write_raw_mm_table, parms.num_iters, 4);
+ }
+
+ output_results(opts,"Raw Data Write",write_raw_mm_table,parms.num_iters,raw_size);
+ } /* end if */
+
/* show mpi write statics */
if (pio_debug_level >= 3) {
/* output all of the times for all iterations */
@@ -542,6 +581,8 @@ run_test(iotype iot, parameters parms)
output_all_info(write_mpi_mm_table, parms.num_iters, 4);
}
+ /* We don't currently output the MPI write results */
+
/* accumulate and output the max, min, and average "write" times */
if (pio_debug_level >= 3) {
/* output all of the times for all iterations */
@@ -550,18 +591,7 @@ run_test(iotype iot, parameters parms)
output_all_info(write_mm_table, parms.num_iters, 4);
}
- total_mm = accumulate_minmax_stuff(write_mm_table, raw_size, parms.num_iters);
-
- print_indent(3);
- output_report("Write (%d iteration(s)):\n", parms.num_iters);
-
- print_indent(4);
- output_report("Minimum Throughput: %.2f MB/s\n", total_mm.min);
- print_indent(4);
- output_report("Maximum Throughput: %.2f MB/s\n", total_mm.max);
- print_indent(4);
- output_report("Average Throughput: %.2f MB/s\n",
- total_mm.sum / total_mm.num);
+ output_results(opts,"Write",write_mm_table,parms.num_iters,raw_size);
/* accumulate and output the max, min, and average "gross write" times */
if (pio_debug_level >= 3) {
@@ -571,18 +601,21 @@ run_test(iotype iot, parameters parms)
output_all_info(write_gross_mm_table, parms.num_iters, 4);
}
- total_mm = accumulate_minmax_stuff(write_gross_mm_table, raw_size, parms.num_iters);
-
- print_indent(3);
- output_report("Write Open-Close (%d iteration(s)):\n", parms.num_iters);
+ output_results(opts,"Write Open-Close",write_gross_mm_table,parms.num_iters,raw_size);
+
+ /* Read statistics */
+ /* Print the raw data throughput if desired */
+ if(opts->print_raw) {
+ /* accumulate and output the max, min, and average "raw read" times */
+ if (pio_debug_level >= 3) {
+ /* output all of the times for all iterations */
+ print_indent(3);
+ output_report("Raw Data Read details:\n");
+ output_all_info(read_raw_mm_table, parms.num_iters, 4);
+ }
- print_indent(4);
- output_report("Minimum Throughput: %.2f MB/s\n", total_mm.min);
- print_indent(4);
- output_report("Maximum Throughput: %.2f MB/s\n", total_mm.max);
- print_indent(4);
- output_report("Average Throughput: %.2f MB/s\n",
- total_mm.sum / total_mm.num);
+ output_results(opts,"Raw Data Read",read_raw_mm_table,parms.num_iters,raw_size);
+ } /* end if */
/* show mpi read statics */
if (pio_debug_level >= 3) {
@@ -592,6 +625,8 @@ run_test(iotype iot, parameters parms)
output_all_info(read_mpi_mm_table, parms.num_iters, 4);
}
+ /* We don't currently output the MPI read results */
+
/* accumulate and output the max, min, and average "read" times */
if (pio_debug_level >= 3) {
/* output all of the times for all iterations */
@@ -600,18 +635,7 @@ run_test(iotype iot, parameters parms)
output_all_info(read_mm_table, parms.num_iters, 4);
}
- total_mm = accumulate_minmax_stuff(read_mm_table, raw_size, parms.num_iters);
-
- print_indent(3);
- output_report("Read (%d iteration(s)):\n", parms.num_iters);
-
- print_indent(4);
- output_report("Minimum Throughput: %.2f MB/s\n", total_mm.min);
- print_indent(4);
- output_report("Maximum Throughput: %.2f MB/s\n", total_mm.max);
- print_indent(4);
- output_report("Average Throughput: %.2f MB/s\n",
- total_mm.sum / total_mm.num);
+ output_results(opts,"Read",read_mm_table,parms.num_iters,raw_size);
/* accumulate and output the max, min, and average "gross read" times */
if (pio_debug_level >= 3) {
@@ -621,18 +645,7 @@ run_test(iotype iot, parameters parms)
output_all_info(read_gross_mm_table, parms.num_iters, 4);
}
- total_mm = accumulate_minmax_stuff(read_gross_mm_table, raw_size, parms.num_iters);
-
- print_indent(3);
- output_report("Read Open-Close (%d iteration(s)):\n", parms.num_iters);
-
- print_indent(4);
- output_report("Minimum Throughput: %.2f MB/s\n", total_mm.min);
- print_indent(4);
- output_report("Maximum Throughput: %.2f MB/s\n", total_mm.max);
- print_indent(4);
- output_report("Average Throughput: %.2f MB/s\n",
- total_mm.sum / total_mm.num);
+ output_results(opts,"Read Open-Close",read_gross_mm_table,parms.num_iters,raw_size);
/* clean up our mess */
free(write_mpi_mm_table);
@@ -641,6 +654,8 @@ run_test(iotype iot, parameters parms)
free(read_mm_table);
free(write_gross_mm_table);
free(read_gross_mm_table);
+ free(write_raw_mm_table);
+ free(read_raw_mm_table);
return ret_value;
}
@@ -667,7 +682,7 @@ output_all_info(minmax *mm, int count, int indent_level)
}
/*
- * Function: get_minmax_stuff
+ * Function: get_minmax
* Purpose: Gather all the min, max and total of val.
* Return: Nothing
* Programmer: Bill Wendling, 21. December 2001
@@ -694,18 +709,21 @@ get_minmax(minmax *mm, double val)
* Return: TOTAL_MM - the total of all of these.
* Programmer: Bill Wendling, 21. December 2001
* Modifications:
+ * Changed to use seconds instead of MB/s - QAK, 5/9/02
*/
static minmax
-accumulate_minmax_stuff(minmax *mm, off_t raw_size, int count)
+accumulate_minmax_stuff(minmax *mm, int count)
{
register int i;
minmax total_mm;
- total_mm.sum = total_mm.max = total_mm.min = MB_PER_SEC(raw_size, mm[0].max);
+ total_mm.sum = 0.0;
+ total_mm.max = -DBL_MAX;
+ total_mm.min = DBL_MAX;
total_mm.num = count;
- for (i = 1; i < count; ++i) {
- double m = MB_PER_SEC(raw_size, mm[i].max);
+ for (i = 0; i < count; ++i) {
+ double m = mm[i].max;
total_mm.sum += m;
@@ -801,6 +819,50 @@ destroy_comm_world(void)
}
/*
+ * Function: output_results
+ * Purpose: Print information about the time & bandwidth for a given
+ * minmax & # of iterations.
+ * Return: Nothing
+ * Programmer: Quincey Koziol, 9. May 2002
+ * Modifications:
+ */
+static void
+output_results(const struct options *opts, const char *name, minmax *table,
+ int table_size,off_t data_size)
+{
+ minmax total_mm;
+
+ total_mm = accumulate_minmax_stuff(table, table_size);
+
+ print_indent(3);
+ output_report("%s (%d iteration(s)):\n", name,(int)table_size);
+
+ /* Note: The maximum throughput uses the minimum amount of time & vice versa */
+
+ print_indent(4);
+ output_report("Maximum Throughput: %6.2f MB/s", MB_PER_SEC(data_size,total_mm.min));
+ if(opts->print_times)
+ output_report(" (%7.3f s)\n", total_mm.min);
+ else
+ output_report("\n");
+
+ print_indent(4);
+ output_report("Average Throughput: %6.2f MB/s",
+ MB_PER_SEC(data_size,total_mm.sum / total_mm.num));
+ if(opts->print_times)
+ output_report(" (%7.3f s)\n", (total_mm.sum / total_mm.num));
+ else
+ output_report("\n");
+
+ print_indent(4);
+ output_report("Minimum Throughput: %6.2f MB/s", MB_PER_SEC(data_size,total_mm.max));
+ if(opts->print_times)
+ output_report(" (%7.3f s)\n", total_mm.max);
+ else
+ output_report("\n");
+}
+
+/*
* Function: output_report
* Purpose: Print a line of the report. Only do so if I'm the 0 process.
* Return: Nothing
@@ -872,56 +934,83 @@ parse_command_line(int argc, char *argv[])
cl_opts->min_num_procs = 1;
cl_opts->max_xfer_size = 1 * ONE_MB;
cl_opts->min_xfer_size = 128 * ONE_KB;
+ cl_opts->print_times = 0; /* Printing times is off by default */
+ cl_opts->print_raw = 0; /* Printing raw data throughput is off by default */
+ cl_opts->h5_alignment = 1; /* No alignment for HDF5 objects by default */
+ cl_opts->h5_threshold = 1; /* No threshold for aligning HDF5 objects by default */
+ cl_opts->h5_use_chunks = 0; /* Don't chunk the HDF5 dataset by default */
+ cl_opts->h5_no_fill = 0; /* Write fill values by default */
while ((opt = get_option(argc, (const char **)argv, s_opts, l_opts)) != EOF) {
switch ((char)opt) {
+ case 'a':
+ cl_opts->h5_alignment = parse_size_directive(opt_arg);
+ break;
#if 0
case 'b':
/* the future "binary" option */
break;
#endif /* 0 */
+ case 'c': /* Turn on chunked HDF5 dataset creation */
+ cl_opts->h5_use_chunks = 1;
+ break;
case 'd':
- cl_opts->num_dsets = strtol(opt_arg, NULL, 10);
+ cl_opts->num_dsets = atoi(opt_arg);
break;
case 'D':
- pio_debug_level = strtol(opt_arg, NULL, 10);
-
- if (pio_debug_level > 4)
- pio_debug_level = 4;
- else if (pio_debug_level < 0)
- pio_debug_level = 0;
+ switch (*opt_arg){ /* Debug sub-options */
+ case 'r': /* Turn on raw data throughput info */
+ cl_opts->print_raw = 1;
+ break;
+ case 't': /* Turn on time printing */
+ cl_opts->print_times = 1;
+ break;
+ default: /* assume debug level */
+ pio_debug_level = atoi(opt_arg);
+
+ if (pio_debug_level > 4)
+ pio_debug_level = 4;
+ else if (pio_debug_level < 0)
+ pio_debug_level = 0;
+ }
break;
case 'f':
cl_opts->file_size = parse_size_directive(opt_arg);
break;
case 'F':
- cl_opts->num_files = strtol(opt_arg, NULL, 10);
+ cl_opts->num_files = atoi(opt_arg);
break;
case 'H':
cl_opts->io_types &= ~0x7;
cl_opts->io_types |= PIO_HDF5;
break;
case 'i':
- cl_opts->num_iters = strtol(opt_arg, NULL, 10);
+ cl_opts->num_iters = atoi(opt_arg);
break;
case 'm':
cl_opts->io_types &= ~0x7;
cl_opts->io_types |= PIO_MPI;
break;
+ case 'n': /* Turn off writing fill values */
+ cl_opts->h5_no_fill = 1;
+ break;
case 'o':
cl_opts->output_file = opt_arg;
break;
case 'p':
- cl_opts->min_num_procs = strtol(opt_arg, NULL, 10);
+ cl_opts->min_num_procs = atoi(opt_arg);
break;
case 'P':
- cl_opts->max_num_procs = strtol(opt_arg, NULL, 10);
+ cl_opts->max_num_procs = atoi(opt_arg);
break;
case 'r':
cl_opts->io_types &= ~0x7;
cl_opts->io_types |= PIO_POSIX;
break;
+ case 'T':
+ cl_opts->h5_threshold = parse_size_directive(opt_arg);
+ break;
case 'x':
cl_opts->min_xfer_size = parse_size_directive(opt_arg);
break;
@@ -1008,17 +1097,23 @@ usage(const char *prog)
fprintf(stdout, "usage: %s [OPTIONS]\n", prog);
fprintf(stdout, " OPTIONS\n");
fprintf(stdout, " -h, --help Print a usage message and exit\n");
+ fprintf(stdout, " -a S, --align=S Alignment of objects in HDF5 file [default: 1]\n");
+ fprintf(stdout, " -c, --chunk Create HDF5 datasets chunked [default: off]\n");
fprintf(stdout, " -d N, --num-dsets=N Number of datasets per file [default:1]\n");
fprintf(stdout, " -D N, --debug=N Indicate the debugging level [default:0]\n");
+ fprintf(stdout, " -D r Print raw data I/O throughput information [default: off]\n");
+ fprintf(stdout, " -D t Print times as well as throughputs [default: off]\n");
fprintf(stdout, " -f S, --file-size=S Size of a single file [default: 64M]\n");
fprintf(stdout, " -F N, --num-files=N Number of files [default: 1]\n");
fprintf(stdout, " -H, --hdf5 Run HDF5 performance test\n");
fprintf(stdout, " -i, --num-iterations Number of iterations to perform [default: 1]\n");
fprintf(stdout, " -m, --mpiio Run MPI/IO performance test\n");
+ fprintf(stdout, " -n, --no-fill Don't write fill values to HDF5 dataset [default: off (i.e. write fill values)]\n");
fprintf(stdout, " -o F, --output=F Output raw data into file F [default: none]\n");
fprintf(stdout, " -P N, --max-num-processes=N Maximum number of processes to use [default: all MPI_COMM_WORLD processes ]\n");
fprintf(stdout, " -p N, --min-num-processes=N Minimum number of processes to use [default: 1]\n");
fprintf(stdout, " -r, --raw Run raw (POSIX) performance test\n");
+ fprintf(stdout, " -T S, --threshold=S Threshold for alignment of objects in HDF5 file [default: 1]\n");
fprintf(stdout, " -X S, --max-xfer-size=S Maximum transfer buffer size [default: 1M]\n");
fprintf(stdout, " -x S, --min-xfer-size=S Minimum transfer buffer size [default: 128K]\n");
fprintf(stdout, "\n");
diff --git a/perform/pio_perf.h b/perform/pio_perf.h
index 5f75ad1..89aa879 100644
--- a/perform/pio_perf.h
+++ b/perform/pio_perf.h
@@ -22,9 +22,13 @@ typedef struct parameters_ {
int num_procs; /* Maximum number of processes to use */
int num_files; /* Number of files to create */
long num_dsets; /* Number of datasets to create */
- long num_elmts; /* Number of native ints in each dset */
+ off_t num_elmts; /* Number of native ints in each dset */
int num_iters; /* Number of times to loop doing the IO */
size_t buf_size; /* Buffer size */
+ hsize_t h5_align; /* HDF5 object alignment */
+ hsize_t h5_thresh; /* HDF5 object alignment threshold */
+ unsigned h5_use_chunks; /* Make HDF5 dataset chunked */
+ unsigned h5_no_fill; /* Disable HDF5 writing fill values */
} parameters;
typedef struct results_ {
diff --git a/perform/pio_timer.c b/perform/pio_timer.c
index d053e10..588ea1d 100644
--- a/perform/pio_timer.c
+++ b/perform/pio_timer.c
@@ -41,15 +41,12 @@ pio_time *timer_g; /* timer: global for stub functions */
* Modifications:
*/
pio_time *
-pio_time_new(unsigned int type)
+pio_time_new(clock_type type)
{
pio_time *pt = (pio_time *)calloc(1, sizeof(struct pio_time_));
- register int i;
/* set global timer variable */
timer_g = pt;
- for (i = 0; i < NUM_TIMERS; ++i)
- pt->total_time[i] = 0.0;
pt->type = type;
return pt;
@@ -82,7 +79,7 @@ pio_time_destroy(pio_time *pt)
* Modifications:
*/
void
-set_timer_type(pio_time *pt, timer_type type)
+set_timer_type(pio_time *pt, clock_type type)
{
pt->type = type;
}
@@ -94,7 +91,7 @@ set_timer_type(pio_time *pt, timer_type type)
* Programmer: Bill Wendling, 04. October 2001
* Modifications:
*/
-timer_type
+clock_type
get_timer_type(pio_time *pt)
{
return pt->type;
@@ -131,47 +128,53 @@ set_time(pio_time *pt, timer_type t, int start_stop)
((double)pt->sys_timer[t].tv_usec) / MICROSECOND);
}
}
- }
- if (pio_debug_level >= 4) {
- char *msg;
- int myrank;
-
- MPI_Comm_rank(pio_comm_g, &myrank);
-
- switch (t) {
- case HDF5_FILE_OPENCLOSE:
- msg = "File Open/Close";
- break;
- case HDF5_DATASET_CREATE:
- msg = "Dataset Create";
- break;
- case HDF5_MPI_WRITE:
- msg = "MPI Write";
- break;
- case HDF5_MPI_READ:
- msg = "MPI Read";
- break;
- case HDF5_FINE_WRITE_FIXED_DIMS:
- msg = "Fine Write";
- break;
- case HDF5_FINE_READ_FIXED_DIMS:
- msg = "Fine Read";
- break;
- case HDF5_GROSS_WRITE_FIXED_DIMS:
- msg = "Gross Write";
- break;
- case HDF5_GROSS_READ_FIXED_DIMS:
- msg = "Gross Read";
- break;
- default:
- msg = "Unknown Timer";
- break;
- }
+ if (pio_debug_level >= 4) {
+ char *msg;
+ int myrank;
+
+ MPI_Comm_rank(pio_comm_g, &myrank);
+
+ switch (t) {
+ case HDF5_FILE_OPENCLOSE:
+ msg = "File Open/Close";
+ break;
+ case HDF5_DATASET_CREATE:
+ msg = "Dataset Create";
+ break;
+ case HDF5_MPI_WRITE:
+ msg = "MPI Write";
+ break;
+ case HDF5_MPI_READ:
+ msg = "MPI Read";
+ break;
+ case HDF5_FINE_WRITE_FIXED_DIMS:
+ msg = "Fine Write";
+ break;
+ case HDF5_FINE_READ_FIXED_DIMS:
+ msg = "Fine Read";
+ break;
+ case HDF5_GROSS_WRITE_FIXED_DIMS:
+ msg = "Gross Write";
+ break;
+ case HDF5_GROSS_READ_FIXED_DIMS:
+ msg = "Gross Read";
+ break;
+ case HDF5_RAW_WRITE_FIXED_DIMS:
+ msg = "Raw Write";
+ break;
+ case HDF5_RAW_READ_FIXED_DIMS:
+ msg = "Raw Read";
+ break;
+ default:
+ msg = "Unknown Timer";
+ break;
+ }
- fprintf(output, " Proc %d: %s %s: %.2f\n", myrank, msg,
- (start_stop == START ? "Start" : "Stop"),
- pt->total_time[t]);
+ fprintf(output, " Proc %d: %s %s: %.2f\n", myrank, msg,
+ (start_stop == START ? "Start" : "Stop"),
+ pt->total_time[t]);
+ }
}
return pt;
diff --git a/perform/pio_timer.h b/perform/pio_timer.h
index 4144825..81916dd 100644
--- a/perform/pio_timer.h
+++ b/perform/pio_timer.h
@@ -28,21 +28,25 @@ typedef enum timer_type_ {
HDF5_FINE_READ_FIXED_DIMS,
HDF5_GROSS_WRITE_FIXED_DIMS,
HDF5_GROSS_READ_FIXED_DIMS,
+ HDF5_RAW_WRITE_FIXED_DIMS,
+ HDF5_RAW_READ_FIXED_DIMS,
NUM_TIMERS
} timer_type;
-/* Miscellaneous identifiers */
-enum {
+typedef enum clock_type_ {
MPI_TIMER = 0, /* Use MPI timer to measure time */
- SYS_TIMER = 1, /* Use system clock to measure time */
+ SYS_TIMER = 1 /* Use system clock to measure time */
+} clock_type;
+/* Miscellaneous identifiers */
+enum {
START, /* Start a specified timer */
STOP /* Stop a specified timer */
};
/* The performance time structure */
typedef struct pio_time_ {
- unsigned int type : 1;
+ clock_type type;
double total_time[NUM_TIMERS];
double mpi_timer[NUM_TIMERS];
struct timeval sys_timer[NUM_TIMERS];
@@ -52,10 +56,10 @@ typedef struct pio_time_ {
#ifdef __cplusplus
extern "C" {
#endif /* __cplusplus */
-extern pio_time *pio_time_new(unsigned int);
+extern pio_time *pio_time_new(clock_type t);
extern void pio_time_destroy(pio_time *pt);
-extern void set_timer_type(pio_time *pt, timer_type type);
-extern timer_type get_timer_type(pio_time *pt);
+extern void set_timer_type(pio_time *pt, clock_type type);
+extern clock_type get_timer_type(pio_time *pt);
extern pio_time *set_time(pio_time *pt, timer_type t, int start_stop);
extern double get_time(pio_time *pt, timer_type t);
#ifdef __cplusplus