summaryrefslogtreecommitdiffstats
path: root/perform
diff options
context:
space:
mode:
authorcvs2svn <no_author@cvs2svn>2002-01-31 20:27:18 (GMT)
committercvs2svn <no_author@cvs2svn>2002-01-31 20:27:18 (GMT)
commitffab5ce759c52ff0b61803b259bd3d786195b724 (patch)
tree5b16dd60a4949fd81041d58fe08295016283c367 /perform
parent7821aaa0631c7f07f45419a633ee1e89f2e691de (diff)
downloadhdf5-ffab5ce759c52ff0b61803b259bd3d786195b724.zip
hdf5-ffab5ce759c52ff0b61803b259bd3d786195b724.tar.gz
hdf5-ffab5ce759c52ff0b61803b259bd3d786195b724.tar.bz2
[svn-r4894] This commit was manufactured by cvs2svn to create branch 'hdf5_1_4'.
Diffstat (limited to 'perform')
-rw-r--r--perform/pio_perf.c985
-rw-r--r--perform/pio_perf.h60
2 files changed, 1045 insertions, 0 deletions
diff --git a/perform/pio_perf.c b/perform/pio_perf.c
new file mode 100644
index 0000000..6da8aae
--- /dev/null
+++ b/perform/pio_perf.c
@@ -0,0 +1,985 @@
+/*
+ * Copyright (C) 2001, 2002
+ * National Center for Supercomputing Applications
+ * All rights reserved.
+ *
+ */
+
+/*
+ * Parallel HDF5 Performance Testing Code
+ * --------------------------------------
+ *
+ * Portable code to test performance on the different platforms we support.
+ * This is what the report should look like:
+ *
+ * nprocs = Max#Procs
+ * IO Type = Raw
+ * # Files = 1, # of dsets = 1000, Elements per dset = 37000
+ * Write Results = x MB/s
+ * Read Results = x MB/s
+ * # Files = 1, # of dsets = 3000, Elements per dset = 37000
+ * Write Results = x MB/s
+ * Read Results = x MB/s
+ *
+ * . . .
+ *
+ * IO Type = MPIO
+ * # Files = 1, # of dsets = 1000, Elements per dset = 37000
+ * Write Results = x MB/s
+ * Read Results = x MB/s
+ * # Files = 1, # of dsets = 3000, Elements per dset = 37000
+ * Write Results = x MB/s
+ * Read Results = x MB/s
+ *
+ * . . .
+ *
+ * IO Type = PHDF5
+ * # Files = 1, # of dsets = 1000, Elements per dset = 37000
+ * Write Results = x MB/s
+ * Read Results = x MB/s
+ * # Files = 1, # of dsets = 3000, Elements per dset = 37000
+ * Write Results = x MB/s
+ * Read Results = x MB/s
+ *
+ * . . .
+ *
+ * nprocs = Max#Procs / 2
+ *
+ * . . .
+ *
+ */
+
+/* system header files */
+#include <stdarg.h>
+#include <stdio.h>
+#include <stdlib.h>
+
+#include "hdf5.h"
+
+#ifdef H5_HAVE_PARALLEL
+
+/* library header files */
+#include <mpi.h>
+
+/* our header files */
+#include "h5tools_utils.h"
+#include "pio_perf.h"
+
+/* useful macros */
+#define TAB_SPACE 4
+
+#define ONE_KB 1024
+#define ONE_MB (ONE_KB * ONE_KB)
+#define ONE_GB (ONE_MB * ONE_KB)
+
+#define PIO_RAW 0x10
+#define PIO_MPI 0x20
+#define PIO_HDF5 0x40
+
+#define MB_PER_SEC(bytes,t) (((bytes) / ONE_MB) / t)
+
+/* global variables */
+MPI_Comm pio_comm_g; /* Communicator to run the PIO */
+int pio_mpi_rank_g; /* MPI rank of pio_comm_g */
+int pio_mpi_nprocs_g; /* Number of processes of pio_comm_g */
+int pio_debug_level = 0;/* The debug level:
+ * 0 - Off
+ * 1 - Minimal
+ * 2 - Some more
+ * 3 - Maximal
+ */
+
+/* local variables */
+static const char *progname = "pio_perf";
+
+/*
+ * Command-line options: The user can specify short or long-named
+ * parameters. The long-named ones can be partially spelled. When
+ * adding more, make sure that they don't clash with each other.
+ */
+#if 1
+static const char *s_opts = "hD:f:HP:p:X:x:md:F:i:o:r";
+#else
+static const char *s_opts = "hbD:f:HP:p:X:x:md:F:i:o:r";
+#endif /* 1 */
+static struct long_options l_opts[] = {
+ { "help", no_arg, 'h' },
+ { "hel", no_arg, 'h' },
+ { "he", no_arg, 'h' },
+#if 0
+ /* a siting of the elusive binary option */
+ { "binary", no_arg, 'b' },
+ { "binar", no_arg, 'b' },
+ { "bina", no_arg, 'b' },
+ { "bin", no_arg, 'b' },
+ { "bi", no_arg, 'b' },
+#endif /* 0 */
+ { "debug", require_arg, 'D' },
+ { "debu", require_arg, 'D' },
+ { "deb", require_arg, 'D' },
+ { "de", require_arg, 'D' },
+ { "d", require_arg, 'D' },
+ { "file-size", require_arg, 'f' },
+ { "file-siz", require_arg, 'f' },
+ { "file-si", require_arg, 'f' },
+ { "file-s", require_arg, 'f' },
+ { "file", require_arg, 'f' },
+ { "fil", require_arg, 'f' },
+ { "fi", require_arg, 'f' },
+ { "hdf5", no_arg, 'H' },
+ { "hdf", no_arg, 'H' },
+ { "hd", no_arg, 'H' },
+ { "max-num-processes", require_arg, 'P' },
+ { "max-num-processe", require_arg, 'P' },
+ { "max-num-process", require_arg, 'P' },
+ { "max-num-proces", require_arg, 'P' },
+ { "max-num-proce", require_arg, 'P' },
+ { "max-num-proc", require_arg, 'P' },
+ { "max-num-pro", require_arg, 'P' },
+ { "max-num-pr", require_arg, 'P' },
+ { "max-num-p", require_arg, 'P' },
+ { "min-num-processes", require_arg, 'p' },
+ { "min-num-processe", require_arg, 'p' },
+ { "min-num-process", require_arg, 'p' },
+ { "min-num-proces", require_arg, 'p' },
+ { "min-num-proce", require_arg, 'p' },
+ { "min-num-proc", require_arg, 'p' },
+ { "min-num-pro", require_arg, 'p' },
+ { "min-num-pr", require_arg, 'p' },
+ { "min-num-p", require_arg, 'p' },
+ { "max-xfer-size", require_arg, 'X' },
+ { "max-xfer-siz", require_arg, 'X' },
+ { "max-xfer-si", require_arg, 'X' },
+ { "max-xfer-s", require_arg, 'X' },
+ { "max-xfer", require_arg, 'X' },
+ { "max-xfe", require_arg, 'X' },
+ { "max-xf", require_arg, 'X' },
+ { "max-x", require_arg, 'X' },
+ { "min-xfer-size", require_arg, 'x' },
+ { "min-xfer-siz", require_arg, 'x' },
+ { "min-xfer-si", require_arg, 'x' },
+ { "min-xfer-s", require_arg, 'x' },
+ { "min-xfer", require_arg, 'x' },
+ { "min-xfe", require_arg, 'x' },
+ { "min-xf", require_arg, 'x' },
+ { "min-x", require_arg, 'x' },
+ { "mpiio", no_arg, 'm' },
+ { "mpii", no_arg, 'm' },
+ { "mpi", no_arg, 'm' },
+ { "mp", no_arg, 'm' },
+ { "num-dsets", require_arg, 'd' },
+ { "num-dset", require_arg, 'd' },
+ { "num-dse", require_arg, 'd' },
+ { "num-ds", require_arg, 'd' },
+ { "num-d", require_arg, 'd' },
+ { "num-files", require_arg, 'F' },
+ { "num-file", require_arg, 'F' },
+ { "num-fil", require_arg, 'F' },
+ { "num-fi", require_arg, 'F' },
+ { "num-f", require_arg, 'F' },
+ { "num-iterations", require_arg, 'i' },
+ { "num-iteration", require_arg, 'i' },
+ { "num-iteratio", require_arg, 'i' },
+ { "num-iterati", require_arg, 'i' },
+ { "num-iterat", require_arg, 'i' },
+ { "num-itera", require_arg, 'i' },
+ { "num-iter", require_arg, 'i' },
+ { "num-ite", require_arg, 'i' },
+ { "num-it", require_arg, 'i' },
+ { "num-i", require_arg, 'i' },
+ { "output", require_arg, 'o' },
+ { "outpu", require_arg, 'o' },
+ { "outp", require_arg, 'o' },
+ { "out", require_arg, 'o' },
+ { "ou", require_arg, 'o' },
+ { "raw", no_arg, 'r' },
+ { "ra", no_arg, 'r' },
+ { NULL, 0, '\0' }
+};
+
+struct options {
+ long io_types; /* bitmask of which I/O types to test */
+ const char *output_file; /* file to print report to */
+ long file_size; /* size of file */
+ long num_dsets; /* number of datasets */
+ long num_files; /* number of files */
+ long num_iters; /* number of iterations */
+ long max_num_procs; /* maximum number of processes to use */
+ long min_num_procs; /* minimum number of processes to use */
+ long max_xfer_size; /* maximum transfer buffer size */
+ long min_xfer_size; /* minimum transfer buffer size */
+};
+
+typedef struct _minmax {
+ double min;
+ double max;
+ double sum;
+ int num;
+} minmax;
+
+/* local functions */
+static long parse_size_directive(const char *size);
+static struct options *parse_command_line(int argc, char *argv[]);
+static void run_test_loop(FILE *output, struct options *options);
+static int run_test(FILE *output, iotype iot, parameters parms);
+static void output_all_info(FILE *output, minmax *mm, int count, int indent_level);
+static void get_minmax(minmax *mm, double val);
+static minmax accumulate_minmax_stuff(minmax *mm, long raw_size, int count);
+static int create_comm_world(int num_procs, int *doing_pio);
+static int destroy_comm_world(void);
+static void output_report(FILE *output, const char *fmt, ...);
+static void print_indent(register FILE *output, register int indent);
+static void usage(const char *prog);
+
+/*
+ * Function: main
+ * Purpose: Start things up. Initialize MPI and then call the test looping
+ * function.
+ * Return: EXIT_SUCCESS or EXIT_FAILURE
+ * Programmer: Bill Wendling, 30. October 2001
+ * Modifications:
+ */
+int
+main(int argc, char **argv)
+{
+ int world_size, ret;
+ int exit_value = EXIT_SUCCESS;
+ FILE *output = stdout;
+ struct options *opts = NULL;
+
+ /* initialize MPI and get the maximum num of processors we started with */
+ MPI_Init(&argc, &argv);
+ ret = MPI_Comm_size(MPI_COMM_WORLD, &world_size);
+
+ if (ret != MPI_SUCCESS) {
+ fprintf(stderr, "%s: MPI_Comm_size call failed\n", progname);
+
+ if (ret == MPI_ERR_COMM)
+ fprintf(stderr, "invalid MPI communicator\n");
+ else
+ fprintf(stderr, "invalid argument\n");
+
+ exit_value = EXIT_FAILURE;
+ goto finish;
+ }
+
+ pio_comm_g = MPI_COMM_WORLD;
+
+ opts = parse_command_line(argc, argv);
+
+ if (!opts) {
+ exit_value = EXIT_FAILURE;
+ goto finish;
+ }
+
+ if (opts->output_file) {
+ if ((output = fopen(opts->output_file, "w")) == NULL) {
+ fprintf(stderr, "%s: cannot open output file\n", progname);
+ perror(opts->output_file);
+ goto finish;
+ }
+ }
+
+ run_test_loop(output, opts);
+
+finish:
+ MPI_Finalize();
+ free(opts);
+ return exit_value;
+}
+
+/*
+ * Function: run_test_loop
+ * Purpose: Run the I/O tests. Write the results to OUTPUT.
+ *
+ * - The slowest changing part of the test is the number of
+ * processors to use. For each loop iteration, we divide that
+ * number by 2 and rerun the test.
+ *
+ * - The second slowest is what type of IO to perform. We have
+ * three choices: RAW, MPI-IO, and PHDF5.
+ *
+ * - Then we change the size of the buffer. This information is
+ * inferred from the number of datasets to create and the number
+ * of integers to put into each dataset. The backend code figures
+ * this out.
+ *
+ * Return: Nothing
+ * Programmer: Bill Wendling, 30. October 2001
+ * Modifications:
+ */
+static void
+run_test_loop(FILE *output, struct options *opts)
+{
+ parameters parms;
+ long num_procs;
+ int doing_pio; /* if this process is doing PIO */
+ int io_runs = PIO_HDF5 | PIO_MPI | PIO_RAW; /* default to run all tests */
+
+ if (opts->io_types & ~0x7) {
+ /* we want to run only a select subset of these tests */
+ io_runs = 0;
+
+ if (opts->io_types & PIO_HDF5)
+ io_runs |= PIO_HDF5;
+
+ if (opts->io_types & PIO_MPI)
+ io_runs |= PIO_MPI;
+
+ if (opts->io_types & PIO_RAW)
+ io_runs |= PIO_RAW;
+ }
+
+ parms.num_files = opts->num_files;
+ parms.num_dsets = opts->num_dsets;
+ parms.num_iters = opts->num_iters;
+
+ /* multiply the maximum number of processors by 2 for each loop iter */
+ for (num_procs = opts->min_num_procs;
+ num_procs <= opts->max_num_procs; num_procs <<= 1) {
+ register long buf_size;
+
+ parms.num_procs = num_procs;
+
+ if (create_comm_world(parms.num_procs, &doing_pio) != SUCCESS) {
+ /* do something harsh */
+ }
+
+ /* only processes doing PIO will run the tests */
+ if (doing_pio){
+ output_report(output, "Number of processors = %ld\n", parms.num_procs);
+
+ /* multiply the xfer buffer size by 2 for each loop iteration */
+ for (buf_size = opts->min_xfer_size;
+ buf_size <= opts->max_xfer_size; buf_size <<= 1) {
+ parms.buf_size = buf_size;
+ parms.num_elmts = opts->file_size / (parms.num_dsets * sizeof(int));
+
+ print_indent(output, 1);
+ output_report(output, "Transfer Buffer Size: %ld bytes, File size: %.2f MBs\n",
+ buf_size,
+ ((double)parms.num_dsets * parms.num_elmts * sizeof(int)) / ONE_MB);
+ print_indent(output, 1);
+ output_report(output,
+ " # of files: %ld, # of dsets: %ld, # of elmts per dset: %ld\n",
+ parms.num_files, parms.num_dsets, parms.num_elmts);
+
+ if (io_runs & PIO_RAW)
+ run_test(output, RAW, parms);
+
+ if (io_runs & PIO_MPI)
+ run_test(output, MPIO, parms);
+
+ if (io_runs & PIO_HDF5)
+ run_test(output, PHDF5, parms);
+ }
+
+ if (destroy_comm_world() != SUCCESS) {
+ /* do something harsh */
+ }
+ }
+ }
+}
+
+/*
+ * Function: run_test
+ * Purpose: Inner loop call to actually run the I/O test.
+ * Return: Nothing
+ * Programmer: Bill Wendling, 18. December 2001
+ * Modifications:
+ */
+static int
+run_test(FILE *output, iotype iot, parameters parms)
+{
+ results res;
+ register int i, ret_value = SUCCESS;
+ int comm_size;
+ long raw_size;
+ minmax total_mm;
+ minmax *write_mm_table;
+ minmax *write_gross_mm_table;
+ minmax *read_mm_table;
+ minmax *read_gross_mm_table;
+ minmax write_mm = {0.0, 0.0, 0.0, 0};
+ minmax write_gross_mm = {0.0, 0.0, 0.0, 0};
+ minmax read_mm = {0.0, 0.0, 0.0, 0};
+ minmax read_gross_mm = {0.0, 0.0, 0.0, 0};
+
+ raw_size = parms.num_dsets * parms.num_elmts * sizeof(int);
+ parms.io_type = iot;
+ print_indent(output, 2);
+ output_report(output, "Type of IO = ");
+
+ switch (iot) {
+ case RAW:
+ output_report(output, "Raw\n");
+ break;
+ case MPIO:
+ output_report(output, "MPIO\n");
+ break;
+ case PHDF5:
+ output_report(output, "PHDF5\n");
+ break;
+ }
+
+ MPI_Comm_size(pio_comm_g, &comm_size);
+
+ write_mm_table = malloc(parms.num_iters * sizeof(minmax));
+ write_gross_mm_table = malloc(parms.num_iters * sizeof(minmax));
+ read_mm_table = malloc(parms.num_iters * sizeof(minmax));
+ read_gross_mm_table = malloc(parms.num_iters * sizeof(minmax));
+
+ for (i = 0; i < parms.num_iters; ++i) {
+ write_mm_table[i].min = 0.0;
+ write_mm_table[i].max = 0.0;
+ write_mm_table[i].sum = 0.0;
+ write_mm_table[i].num = 0;
+
+ write_gross_mm_table[i].min = 0.0;
+ write_gross_mm_table[i].max = 0.0;
+ write_gross_mm_table[i].sum = 0.0;
+ write_gross_mm_table[i].num = 0;
+
+ read_mm_table[i].min = 0.0;
+ read_mm_table[i].max = 0.0;
+ read_mm_table[i].sum = 0.0;
+ read_mm_table[i].num = 0;
+
+ read_gross_mm_table[i].min = 0.0;
+ read_gross_mm_table[i].max = 0.0;
+ read_gross_mm_table[i].sum = 0.0;
+ read_gross_mm_table[i].num = 0;
+ }
+
+ /* call Albert's testing here */
+ for (i = 0; i < parms.num_iters; ++i) {
+ double t;
+
+ MPI_Barrier(pio_comm_g);
+ res = do_pio(parms);
+
+ /* gather all of the "write" times */
+ t = get_time(res.timers, HDF5_FINE_WRITE_FIXED_DIMS);
+ get_minmax(&write_mm, t);
+
+ write_mm_table[i] = write_mm;
+
+ /* gather all of the "write" times from open to close */
+ t = get_time(res.timers, HDF5_GROSS_WRITE_FIXED_DIMS);
+ get_minmax(&write_gross_mm, t);
+
+ write_gross_mm_table[i] = write_gross_mm;
+
+ /* gather all of the "read" times */
+ t = get_time(res.timers, HDF5_FINE_READ_FIXED_DIMS);
+ get_minmax(&read_mm, t);
+
+ read_mm_table[i] = read_mm;
+
+ /* gather all of the "read" times from open to close */
+ t = get_time(res.timers, HDF5_GROSS_READ_FIXED_DIMS);
+ get_minmax(&read_gross_mm, t);
+
+ read_gross_mm_table[i] = read_gross_mm;
+ }
+
+ /* accumulate and output the max, min, and average "write" times */
+ if (pio_debug_level == 3) {
+ /* output all of the times for all iterations */
+ print_indent(output, 3);
+ output_report(output, "Write details:\n");
+ output_all_info(output, write_mm_table, parms.num_iters, 4);
+ }
+
+ total_mm = accumulate_minmax_stuff(write_mm_table, raw_size, parms.num_iters);
+
+ print_indent(output, 3);
+ output_report(output, "Write (%d iteration(s)):\n", parms.num_iters);
+
+ print_indent(output, 4);
+ output_report(output, "Minimum Throughput: %.2f MB/s\n", total_mm.min);
+ print_indent(output, 4);
+ output_report(output, "Maximum Throughput: %.2f MB/s\n", total_mm.max);
+ print_indent(output, 4);
+ output_report(output, "Average Throughput: %.2f MB/s\n",
+ total_mm.sum / total_mm.num);
+
+ /* accumulate and output the max, min, and average "gross write" times */
+ if (pio_debug_level == 3) {
+ /* output all of the times for all iterations */
+ print_indent(output, 3);
+ output_report(output, "Write Open-Close details:\n");
+ output_all_info(output, write_gross_mm_table, parms.num_iters, 4);
+ }
+
+ total_mm = accumulate_minmax_stuff(write_gross_mm_table, raw_size, parms.num_iters);
+
+ print_indent(output, 3);
+ output_report(output, "Write Open-Close (%d iteration(s)):\n", parms.num_iters);
+
+ print_indent(output, 4);
+ output_report(output, "Minimum Throughput: %.2f MB/s\n", total_mm.min);
+ print_indent(output, 4);
+ output_report(output, "Maximum Throughput: %.2f MB/s\n", total_mm.max);
+ print_indent(output, 4);
+ output_report(output, "Average Throughput: %.2f MB/s\n",
+ total_mm.sum / total_mm.num);
+
+ /* accumulate and output the max, min, and average "read" times */
+ if (pio_debug_level == 3) {
+ /* output all of the times for all iterations */
+ print_indent(output, 3);
+ output_report(output, "Read details:\n");
+ output_all_info(output, read_mm_table, parms.num_iters, 4);
+ }
+
+ total_mm = accumulate_minmax_stuff(read_mm_table, raw_size, parms.num_iters);
+
+ print_indent(output, 3);
+ output_report(output, "Read (%d iteration(s)):\n", parms.num_iters);
+
+ print_indent(output, 4);
+ output_report(output, "Minimum Throughput: %.2f MB/s\n", total_mm.min);
+ print_indent(output, 4);
+ output_report(output, "Maximum Throughput: %.2f MB/s\n", total_mm.max);
+ print_indent(output, 4);
+ output_report(output, "Average Throughput: %.2f MB/s\n",
+ total_mm.sum / total_mm.num);
+
+ /* accumulate and output the max, min, and average "gross read" times */
+ if (pio_debug_level == 3) {
+ /* output all of the times for all iterations */
+ print_indent(output, 3);
+ output_report(output, "Read Open-Close details:\n");
+ output_all_info(output, read_gross_mm_table, parms.num_iters, 4);
+ }
+
+ total_mm = accumulate_minmax_stuff(read_gross_mm_table, raw_size, parms.num_iters);
+
+ print_indent(output, 3);
+ output_report(output, "Read Open-Close (%d iteration(s)):\n", parms.num_iters);
+
+ print_indent(output, 4);
+ output_report(output, "Minimum Throughput: %.2f MB/s\n", total_mm.min);
+ print_indent(output, 4);
+ output_report(output, "Maximum Throughput: %.2f MB/s\n", total_mm.max);
+ print_indent(output, 4);
+ output_report(output, "Average Throughput: %.2f MB/s\n",
+ total_mm.sum / total_mm.num);
+
+ free(write_mm_table);
+ free(read_mm_table);
+ pio_time_destroy(res.timers);
+ return ret_value;
+}
+
+/*
+ * Function: output_all_info
+ * Purpose:
+ * Return: Nothing
+ * Programmer: Bill Wendling, 29. January 2002
+ * Modifications:
+ */
+static void
+output_all_info(FILE *output, minmax *mm, int count, int indent_level)
+{
+ register int i;
+
+ for (i = 0; i < count; ++i) {
+ print_indent(output, indent_level);
+ output_report(output, "Iteration %d:\n", i + 1);
+ print_indent(output, indent_level + 1);
+ output_report(output, "Minimum Time: %.2fs\n", mm[i].min);
+ print_indent(output, indent_level + 1);
+ output_report(output, "Maximum Time: %.2fs\n", mm[i].max);
+ }
+}
+
+/*
+ * Function: get_minmax_stuff
+ * Purpose: Gather all the min, max and total of val.
+ * Return: Nothing
+ * Programmer: Bill Wendling, 21. December 2001
+ * Modifications:
+ * Use MPI_Allreduce to do it. -akc, 2002/01/11
+ */
+static void
+get_minmax(minmax *mm, double val)
+{
+ int myrank;
+
+ MPI_Comm_rank(pio_comm_g, &myrank);
+ MPI_Comm_size(pio_comm_g, &mm->num);
+
+ MPI_Allreduce(&val, &mm->max, 1, MPI_DOUBLE, MPI_MAX, pio_comm_g);
+ MPI_Allreduce(&val, &mm->min, 1, MPI_DOUBLE, MPI_MIN, pio_comm_g);
+ MPI_Allreduce(&val, &mm->sum, 1, MPI_DOUBLE, MPI_SUM, pio_comm_g);
+}
+
+/*
+ * Function: accumulate_minmax_stuff
+ * Purpose: Accumulate the minimum, maximum, and average of the times
+ * across all processes.
+ * Return: TOTAL_MM - the total of all of these.
+ * Programmer: Bill Wendling, 21. December 2001
+ * Modifications:
+ */
+static minmax
+accumulate_minmax_stuff(minmax *mm, long raw_size, int count)
+{
+ register int i;
+ minmax total_mm;
+
+ total_mm.sum = total_mm.max = total_mm.min = MB_PER_SEC(raw_size, mm[0].max);
+ total_mm.num = count;
+
+ for (i = 1; i < count; ++i) {
+ double m = MB_PER_SEC(raw_size, mm[i].max);
+
+ total_mm.sum += m;
+
+ if (m < total_mm.min)
+ total_mm.min = m;
+
+ if (m > total_mm.max)
+ total_mm.max = m;
+ }
+
+ return total_mm;
+}
+
+/*
+ * Function: create_comm_world
+ * Purpose: Create an MPI Comm world and store it in pio_comm_g, which
+ * is a global variable.
+ * Return: SUCCESS on success.
+ * FAIL otherwise.
+ * Programmer: Bill Wendling, 19. December 2001
+ * Modifications:
+ */
+static int
+create_comm_world(int num_procs, int *doing_pio)
+{
+ /* MPI variables */
+ int mrc, ret_value; /* return values */
+ int color; /* for communicator creation */
+ int myrank, nprocs;
+
+ pio_comm_g = MPI_COMM_NULL;
+
+ /*
+ * Create a sub communicator for this PIO run. Easier to use the first N
+ * processes.
+ */
+ MPI_Comm_size(MPI_COMM_WORLD, &nprocs);
+
+ if (num_procs > nprocs) {
+ fprintf(stderr,
+ "number of process(%d) must be <= number of processes in MPI_COMM_WORLD(%d)\n",
+ num_procs, nprocs);
+ goto error_done;
+ }
+
+ MPI_Comm_rank(MPI_COMM_WORLD, &myrank);
+ color = (myrank < num_procs);
+ mrc = MPI_Comm_split(MPI_COMM_WORLD, color, myrank, &pio_comm_g);
+
+ if (mrc != MPI_SUCCESS) {
+ fprintf(stderr, "MPI_Comm_split failed\n");
+ goto error_done;
+ }
+
+ if (!color) {
+ /* not involved in this run */
+ mrc = destroy_comm_world();
+ goto done;
+ }
+
+ /* determine the MPI rank in the PIO communicator */
+ MPI_Comm_size(pio_comm_g, &pio_mpi_nprocs_g);
+ MPI_Comm_rank(pio_comm_g, &pio_mpi_rank_g);
+
+done:
+ *doing_pio = color;
+ return ret_value;
+
+error_done:
+ destroy_comm_world();
+ return FAIL;
+}
+
+/*
+ * Function: destroy_comm_world
+ * Purpose: Destroy the created MPI Comm world which is stored in the
+ * pio_comm_g global variable.
+ * Return: SUCCESS on success.
+ * FAIL otherwise.
+ * Programmer: Bill Wendling, 19. December 2001
+ * Modifications:
+ */
+static int
+destroy_comm_world(void)
+{
+ int mrc = SUCCESS; /* return code */
+
+ /* release MPI resources */
+ if (pio_comm_g != MPI_COMM_NULL)
+ mrc = (MPI_Comm_free(&pio_comm_g) == MPI_SUCCESS ? SUCCESS : FAIL);
+
+ return mrc;
+}
+
+/*
+ * Function: output_report
+ * Purpose: Print a line of the report. Only do so if I'm the 0 process.
+ * Return: Nothing
+ * Programmer: Bill Wendling, 19. December 2001
+ * Modifications:
+ */
+static void
+output_report(FILE *output, const char *fmt, ...)
+{
+ int myrank;
+
+ MPI_Comm_rank(pio_comm_g, &myrank);
+
+ if (myrank == 0) {
+ va_list ap;
+
+ va_start(ap, fmt);
+ vfprintf(output, fmt, ap);
+ va_end(ap);
+ }
+}
+
+/*
+ * Function: print_indent
+ * Purpose: Print spaces to indent a new line of text for pretty printing
+ * things.
+ * Return: Nothing
+ * Programmer: Bill Wendling, 29. October 2001
+ * Modifications:
+ */
+static void
+print_indent(register FILE *output, register int indent)
+{
+ indent *= TAB_SPACE;
+
+ for (; indent > 0; --indent)
+ fputc(' ', output);
+}
+
+/*
+ * Function: parse_command_line
+ * Purpose: Parse the command line options and return a STRUCT OPTIONS
+ * structure which will need to be freed by the calling function.
+ * Return: Pointer to an OPTIONS structure
+ * Programmer: Bill Wendling, 31. October 2001
+ * Modifications:
+ */
+static struct options *
+parse_command_line(int argc, char *argv[])
+{
+ register int opt;
+ struct options *cl_opts;
+
+ cl_opts = (struct options *)malloc(sizeof(struct options));
+
+ cl_opts->output_file = NULL;
+ cl_opts->file_size = 64 * ONE_MB;
+ cl_opts->io_types = 0x7; /* bottom bits indicate default type to run */
+ cl_opts->num_dsets = 1;
+ cl_opts->num_files = 1;
+ cl_opts->num_iters = 1;
+ cl_opts->max_num_procs = 1;
+ cl_opts->min_num_procs = 1;
+ cl_opts->max_xfer_size = 1 * ONE_MB;
+ cl_opts->min_xfer_size = 128 * ONE_KB;
+
+ while ((opt = get_option(argc, (const char **)argv, s_opts, l_opts)) != EOF) {
+ switch ((char)opt) {
+#if 0
+ case 'b':
+ /* the future "binary" option */
+ break;
+#endif /* 0 */
+ case 'd':
+ cl_opts->num_dsets = strtol(opt_arg, NULL, 10);
+ break;
+ case 'D':
+ pio_debug_level = strtol(opt_arg, NULL, 10);
+
+ if (pio_debug_level > 3)
+ pio_debug_level = 3;
+ else if (pio_debug_level < 0)
+ pio_debug_level = 0;
+
+ break;
+ case 'f':
+ cl_opts->file_size = parse_size_directive(opt_arg);
+ break;
+ case 'F':
+ cl_opts->num_files = strtol(opt_arg, NULL, 10);
+ break;
+ case 'H':
+ cl_opts->io_types &= ~0x7;
+ cl_opts->io_types |= PIO_HDF5;
+ break;
+ case 'i':
+ cl_opts->num_iters = strtol(opt_arg, NULL, 10);
+ break;
+ case 'm':
+ cl_opts->io_types &= ~0x7;
+ cl_opts->io_types |= PIO_MPI;
+ break;
+ case 'o':
+ cl_opts->output_file = opt_arg;
+ break;
+ case 'p':
+ cl_opts->min_num_procs = strtol(opt_arg, NULL, 10);
+ break;
+ case 'P':
+ cl_opts->max_num_procs = strtol(opt_arg, NULL, 10);
+ break;
+ case 'r':
+ cl_opts->io_types &= ~0x7;
+ cl_opts->io_types |= PIO_RAW;
+ break;
+ case 'x':
+ cl_opts->min_xfer_size = parse_size_directive(opt_arg);
+ break;
+ case 'X':
+ cl_opts->max_xfer_size = parse_size_directive(opt_arg);
+ break;
+ case 'h':
+ case '?':
+ default:
+ usage(progname);
+ free(cl_opts);
+ return NULL;
+ }
+ }
+
+ return cl_opts;
+}
+
+/*
+ * Function: parse_size_directive
+ * Purpose: Parse the size directive passed on the commandline. The size
+ * directive is an integer followed by a size indicator:
+ *
+ * K, k - Kilobyte
+ * M, m - Megabyte
+ * G, g - Gigabyte
+ *
+ * Return: The size as a LONG. If an unknown size indicator is used, then
+ * the program will exit with EXIT_FAILURE as the return value.
+ * Programmer: Bill Wendling, 18. December 2001
+ * Modifications:
+ */
+static long
+parse_size_directive(const char *size)
+{
+ long s;
+ char *endptr;
+
+ s = strtol(size, &endptr, 10);
+
+ if (endptr && *endptr) {
+ while (*endptr != '\0' && (*endptr == ' ' || *endptr == '\t'))
+ ++endptr;
+
+ switch (*endptr) {
+ case 'K':
+ case 'k':
+ s *= ONE_KB;
+ break;
+ case 'M':
+ case 'm':
+ s *= ONE_MB;
+ break;
+ case 'G':
+ case 'g':
+ s *= ONE_GB;
+ break;
+ default:
+ fprintf(stderr, "Illegal size specifier '%c'\n", *endptr);
+ exit(EXIT_FAILURE);
+ }
+ }
+
+ return s;
+}
+
+/*
+ * Function: usage
+ * Purpose: Print a usage message and then exit.
+ * Return: Nothing
+ * Programmer: Bill Wendling, 31. October 2001
+ * Modifications:
+ */
+static void
+usage(const char *prog)
+{
+ int myrank;
+
+ MPI_Comm_rank(pio_comm_g, &myrank);
+
+ if (myrank == 0) {
+ fflush(stdout);
+ fprintf(stdout, "usage: %s [OPTIONS]\n", prog);
+ fprintf(stdout, " OPTIONS\n");
+ fprintf(stdout, " -h, --help Print a usage message and exit\n");
+ fprintf(stdout, " -d N, --num-dsets=N Number of datasets per file [default:1]\n");
+ fprintf(stdout, " -D N, --debug=N Indicate the debugging level [default:0]\n");
+ fprintf(stdout, " -f S, --file-size=S Size of a single file [default: 64M]\n");
+ fprintf(stdout, " -F N, --num-files=N Number of files [default: 1]\n");
+ fprintf(stdout, " -H, --hdf5 Run HDF5 performance test\n");
+ fprintf(stdout, " -i, --num-iterations Number of iterations to perform [default: 1]\n");
+ fprintf(stdout, " -m, --mpiio Run MPI/IO performance test\n");
+ fprintf(stdout, " -o F, --output=F Output raw data into file F [default: none]\n");
+ fprintf(stdout, " -P N, --max-num-processes=N Maximum number of processes to use [default: 1]\n");
+ fprintf(stdout, " -p N, --min-num-processes=N Minimum number of processes to use [default: 1]\n");
+ fprintf(stdout, " -r, --raw Run raw (UNIX) performance test\n");
+ fprintf(stdout, " -X S, --max-xfer-size=S Maximum transfer buffer size [default: 1M]\n");
+ fprintf(stdout, " -x S, --min-xfer-size=S Minimum transfer buffer size [default: 128K]\n");
+ fprintf(stdout, "\n");
+ fprintf(stdout, " F - is a filename.\n");
+ fprintf(stdout, " N - is an integer >=0.\n");
+ fprintf(stdout, " S - is a size specifier, an integer >=0 followed by a size indicator:\n");
+ fprintf(stdout, "\n");
+ fprintf(stdout, " K - Kilobyte\n");
+ fprintf(stdout, " M - Megabyte\n");
+ fprintf(stdout, " G - Gigabyte\n");
+ fprintf(stdout, "\n");
+ fprintf(stdout, " Example: 37M = 37 Megabytes\n");
+ fprintf(stdout, "\n");
+ fprintf(stdout, " Debugging levels are:\n");
+ fprintf(stdout, "\n");
+ fprintf(stdout, " 0 - None\n");
+ fprintf(stdout, " 1 - Minimal\n");
+ fprintf(stdout, " 2 - Not quite everything\n");
+ fprintf(stdout, " 3 - Everything\n");
+ fprintf(stdout, "\n");
+ fflush(stdout);
+ }
+}
+
+#else /* H5_HAVE_PARALLEL */
+
+/*
+ * Function: main
+ * Purpose: Dummy main() function for if HDF5 was configured without
+ * parallel stuff.
+ * Return: EXIT_SUCCESS
+ * Programmer: Bill Wendling, 14. November 2001
+ * Modifications:
+ */
+int
+main(void)
+{
+ printf("No parallel IO performance because parallel is not configured\n");
+ return EXIT_SUCCESS;
+}
+
+#endif /* !H5_HAVE_PARALLEL */
diff --git a/perform/pio_perf.h b/perform/pio_perf.h
new file mode 100644
index 0000000..cf84f55
--- /dev/null
+++ b/perform/pio_perf.h
@@ -0,0 +1,60 @@
+/*
+ * Copyright (C) 2001
+ * National Center for Supercomputing Applications
+ * All rights reserved.
+ *
+ */
+#ifndef PIO_PERF_H__
+#define PIO_PERF_H__
+
+#include "pio_timer.h"
+
+typedef enum iotype_ {
+ RAW,
+ MPIO,
+ PHDF5
+ /*NUM_TYPES*/
+} iotype;
+
+typedef struct parameters_ {
+ iotype io_type; /* The type of IO test to perform */
+ int num_procs; /* Maximum number of processes to use */
+ int num_files; /* Number of files to create */
+ long num_dsets; /* Number of datasets to create */
+ long num_elmts; /* Number of native ints in each dset */
+ int num_iters; /* Number of times to loop doing the IO */
+ long buf_size; /* Buffer size */
+} parameters;
+
+typedef struct results_ {
+ herr_t ret_code;
+ pio_time *timers;
+} results;
+
+#ifndef SUCCESS
+#define SUCCESS 0
+#endif /* !SUCCESS */
+
+#ifndef FAIL
+#define FAIL -1
+#endif /* !FAIL */
+
+extern MPI_Comm pio_comm_g; /* Communicator to run the PIO */
+extern int pio_mpi_rank_g; /* MPI rank of pio_comm_g */
+extern int pio_mpi_nprocs_g; /* number of processes of pio_comm_g */
+extern int pio_debug_level; /* The debug level:
+ * 0 - Off
+ * 1 - Minimal
+ * 2 - Some more
+ * 3 - Maximal
+ */
+
+#ifdef __cplusplus
+extern "C" {
+#endif /* __cplusplus */
+extern results do_pio(parameters param);
+#ifdef __cplusplus
+}
+#endif /* __cplusplus */
+
+#endif /* PIO_PERF_H__ */