summaryrefslogtreecommitdiffstats
path: root/release_docs
diff options
context:
space:
mode:
Diffstat (limited to 'release_docs')
-rw-r--r--release_docs/INSTALL20
-rw-r--r--release_docs/INSTALL_TFLOPS175
-rw-r--r--release_docs/INSTALL_VFL131
3 files changed, 9 insertions, 317 deletions
diff --git a/release_docs/INSTALL b/release_docs/INSTALL
index e5ae047..924c29b 100644
--- a/release_docs/INSTALL
+++ b/release_docs/INSTALL
@@ -13,9 +13,8 @@
2.4. Windows/NT
3. Quick installation
- 3.1. TFLOPS
- 3.2. Windows
- 3.3. Certain Virtual File Layer(VFL)
+ 3.1. Windows
+ 3.2. RedStorm (Cray XT3)
4. HDF5 dependencies
4.1. Zlib
@@ -121,17 +120,16 @@
$ make check
$ make install
-3.1. TFLOPS
- Users of the Intel TFLOPS machine, after reading this file,
- should see the INSTALL_TFLOPS for more instructions.
-
-3.2. Windows
+3.1. Windows
Users of Microsoft Windows should see the INSTALL_Windows for
detailed instructions.
-3.3. Certain Virtual File Layer(VFL)
- If users want to install with special Virtual File Layer(VFL),
- please go to read INSTALL_VFL file.
+3.2. RedStorm (Cray Xt3)
+ Users of the Red Storm machine, after reading this file, should read
+ the Red Storm section in the INSTALL_parallel file for specific
+ instructions for the Red Storm machine. The same instructions would
+ probably work for other Cray XT3 systems but they have not been
+ verified.
4. HDF5 dependencies
diff --git a/release_docs/INSTALL_TFLOPS b/release_docs/INSTALL_TFLOPS
deleted file mode 100644
index 66dafc2..0000000
--- a/release_docs/INSTALL_TFLOPS
+++ /dev/null
@@ -1,175 +0,0 @@
-FOR THE INTEL TFLOPS MACHINE:
-
-Below are the step-by-step procedures for building, testing, and
-installing the parallel and sequential versions of the HDF5 library.
-
------------------
-Software locations
-------------------
-The zlib compression library is installed in /usr/community/hdf5/ZLIB.
-The latest version is zlib v1.1.4.
-
-The mpich library, including mpi-io support, is now supported by the
-TFLOPS system staff. Check Sasn100:/usr/local/FAQ/R4.4.0_Release_Notes
-for details. A modified version of mpicc is created to simply the
-command. It is installed as /usr/community/hdf5/tflop-mpich/bin/mpicc.
-
----------------
-gmake recommended
----------------
-Both Sasn100 and Janus have multiple versions of the make command.
-We recommand the use of the Gnu gmake which has some features
-(e.g., -j and --srcdir support) that makes (sic) life easier.
-
----------------
-Parallel HDF5:
----------------
-
-The setup process for building the parallel version of the HDF5 library for the
-ASCI Red machine is very similar to the sequential version. Since TFLOPS
-does not support MPIO, we have prepared a shell-script file that configures
-with the appropriate MPI library.
-
-Assuming you have already unpacked the HDF5 tar-file into the
-directory <hdf5>, follow the steps below:
-FROM SASN100,
-
-1) cd <hdf5>
-
-2) CC=/usr/community/hdf5/tflop-mpich/bin/mpicc \
- ./configure --host=tflops --with-zlib=/usr/community/hdf5/ZLIB
-
- Alternately, you may specify the host explicitly:
-
- CC=/usr/community/hdf5/tflop-mpich/bin/mpicc \
- ./configure --host=i386-intel-osf1 --with-zlib=/usr/community/hdf5/ZLIB
-
- Skip the "--with-zlib=..." option if you do not wish to include the zlib
- compression feature. Without the zlib compression feature, the library
- will not be able to access zlib compressed datasets.
-
- You may safely ignore the WARNING message,
- =========
- configure: WARNING: If you wanted to set the --build type, don't use --host.
- If a cross compiler is detected then cross compile mode will be used.
- =========
- You may add the option "--build=i386-intel-osf1" to get rid of the WARNING.
-
- (The previous bugs in src/Makefile and test/Makefile have been resolved.
- You don't need to edit them any more.)
-
-3) gmake H5detect
-
-
-FROM JANUS,
-
-4) cd <hdf5>
-
-5) gmake H5Tinit.c
-
-
-FROM SASN100,
-
-6) gmake
-
-
-When everything is finished compiling and linking,
-FROM JANUS,
-
-7) gmake check
- (We have not encountered the following problem for a year.)
- Sometimes the "gmake check" fails in the sub-directories of test
- or tools with a message as "print not found". This is due to the
- "gmake" of Janus thinking some binary code needs to be recompiled.
- The easiest way to fix it is
- FROM SASN100
- cd <hdf5>/test # or cd <hdf5>/tools
- gmake clean; gmake # re-make all binary
-
-
-Once satisfied with the parallel test results, as long as you
-have the correct permission,
-FROM SASN100,
-
-8) gmake install
-
-
----------------
-Sequential HDF5:
----------------
-(**NOTE** We have stopped testing sequential HDF5 for the Tflops machine
-since it has little practical value to build sequential applications for
-the Tflops machine. The instruction below are kept more for historical
-purpose.)
-
-The setup process for building the sequential HDF5 library for the
-ASCI Red machine is done by a coordination of events from sasn100 and
-janus. Though janus can do compiling, it is better to build it
-from sasn100 which has more complete building tools and runs faster.
-It is also anti-social to tie up janus with compiling. The HDF5 building
-requires the use of janus because one of steps is to execute a program
-to find out the run-time characteristics of the TFLOPS machine.
-
-Assuming you have already unpacked the HDF5 tar-file into the
-directory <hdf5>, follow the steps below:
-
-FROM SASN100,
-
-1) cd <hdf5>
-
-2) ./configure --host=tflops --with-zlib=/usr/community/hdf5/ZLIB
-
- Alternately, you may specify the host explicitly:
-
- ./configure --host=i386-intel-osf1 --with-zlib=/usr/community/hdf5/ZLIB
-
- Skip the "--with-zlib=..." option if you do not wish to include the zlib
- compression feature. Without the zlib compression feature, the library
- will not be able to access zlib compressed datasets.
-
- You may safely ignore the WARNING message,
- =========
- configure: WARNING: If you wanted to set the --build type, don't use --host.
- If a cross compiler is detected then cross compile mode will be used.
- =========
- You may add the option "--build=i386-intel-osf1" to get rid of the WARNING.
-
- (The previous bugs in src/Makefile and test/Makefile have been resolved.
- You don't need to edit them any more.)
-
-3) gmake H5detect
-
-
-FROM JANUS,
-
-4) cd <hdf5>
-
-5) gmake H5Tinit.c
-
-
-FROM SASN100,
-
-6) gmake
-
-
-When everything is finished compiling and linking,
-you can run the tests by
-FROM JANUS,
-
-7) gmake check
- Sometimes the "gmake check" fails in the sub-directories of test
- or tools with a message as "print not found". This is due to the
- "gmake" of Janus thinking some binary code needs to be recompiled.
- The easiest way to fix it is
- FROM SASN100
- cd <hdf5>/test # or cd <hdf5>/tools
- gmake clean; gmake # re-make all binary
-
-
-Once satisfied with the test results, you can install
-the software by
-FROM SASN100,
-
-8) gmake install
-
-
diff --git a/release_docs/INSTALL_VFL b/release_docs/INSTALL_VFL
deleted file mode 100644
index 4385c26..0000000
--- a/release_docs/INSTALL_VFL
+++ /dev/null
@@ -1,131 +0,0 @@
- Installation Instructions for HDF5
- with Different Virtual File Layer
-
-Support for SRB and GASS drivers was removed. See RELEASE.txt file
- * * *
-
-This file contains installation instructions for HDF5 with certain Virtual File
-Layer to handle file I/O. We currently have documented SRB and Globus-GASS.
-
-
-
- --- Part I. SRB ---
-I. Overview
------------
-This part contains instructions for remote-accessing HDF5 through SRB. The
-SRB version 1.1.7 on Sun Solaris 2.7 platform has been tested. If you have
-difficulties installing the software in your system, please send mails to me
-(Raymond Lu) at
-
- slu@ncsa.uiuc.edu
-
-First, you must obtain and unpack the HDF5 source as described in the file
-INSTALL. You need the SRB library(client part) installed. You should also
-have access to SRB server.
-
-
-The Storage Resource Broker(SRB) from San Diego Supercomputer Center is client-
-server middleware that provides a uniform interface for connecting to
-heterogeneous data resources over a network and accessig replicated data sets.
-SRB, in conjunction with the Metadata Catalog(MCAT), provides a way to access
-data sets and resources based on their attributes rather than their names or
-physical locations. Their webpage is at http://www.npaci.edu/Research/DI/srb.
-
-HDF5 is built on the top of SRB as a client to remotely access files on SRB
-server through SRB. Right now, HDF-SRB only support low-level file transfer of
-SRB. The MCAT part is not supported yet. Low-level file transfer means files
-are treated just like Unix type files. Files can be read, written and
-appended. Partial access(read and write to a chunk of file without transferrig
-the whole) is also supported.
-
-
-II. Installation Steps
-----------------------
-The installation steps are similar to the ones in INSTALL file:
-
-1. Run 'configure' file with SRB options:
- configure --with-srb=$SRB/include,$SRB/lib
- where $SRB is your SRB installed library directory.
-
- For example, below is a script file to run 'configure':
- #! /bin/sh
- # how to configure to use the SRB
-
- SRB_DIR=/afs/ncsa.uiuc.edu/projects/hdf/users/slu/srb_install
- configure --with-srb=$SRB_DIR/include,$SRB_DIR/lib
-
-2. Run 'make'
-
-3. Run 'make check'
-
-4. Run 'make install'
-
-5. Run testing program(Optional):
- Go to the testing directory(cd test) and run srb_write, srb_read,
- srb_append. Actually, these tests have been run in step 3.
-
- srb_write: Connect to SRB server, write an HDF5 file with an integer
- dataset to SRB server.
- srb_read: Connect to SRB server, read part of HDF5 file on the SRB server.
- srb_append: Connect to SRB server, append an integer dataset to an existent
- HDF5 file on the SRB server.
-
-6. For using HDF-SRB, please read comments in srb_write.c, srb_read.c,
- srb_append.c in the hdf5/test directory.
-
-
-
- --- Part II. Globus-GASS ---
-
-I. Overview
------------
-This part contains instructions for remote-accessing HDF5 through Globus-GASS.
-The SGI IRIX64(and IRIX) 6.5 platform have been tested. If you have
-difficulties installing the software in your system, please send mails to me
-(Raymond Lu) at
- slu@ncsa.uiuc.edu
-
-First, you must obtain and unpack the HDF5 source as described in the file
-INSTALL. You need the Globus 1.1.x and SSL(should have come with Globus)
-packages.
-
-HDF5 is built on the top of Globus-GASS(1.1.x) to handle remote access.
-Globus-GASS(1.1.x) only supports HTTP and HTTPS protocals for 'whole file
-read and write'. More features may be added in the future.
-
-II. Installation Steps
-----------------------
-The installation steps are similar to the ones in INSTALL file:
-
-1. Run 'configure' file with SSL and GASS options:
- configure --with-ssl=$SSL/lib --with-gass=$GASS/include,$GASS/lib
- where $SSL is your SSL directory, and $GASS is your Globus directory.
-
- For example, below is a script file to run 'configure':
- #! /bin/sh
- # how to configure to use the Globus-GASS(1.1.x)
-
- GASS_DIR=/usr/local/globus-install-1.1.1/development/mips-sgi-irix6.5-64_nothreads_standard_debug
- SSL_LIB=/usr/local/ssl/lib
-
- configure --with-ssl=$SSL_LIB --with-gass=$GASS_DIR/include,$GASS_DIR/lib
-
-2. Run 'make'
-
-3. Run 'make check'
-
-4. Run 'make install'
-
-5. Run testing program:
- There is one read testing program called 'gass_read' in the 'test'
- directory. It does whole file read through HTTP protocal. The URL is
- hard coded as
- http://hdf.ncsa.uiuc.edu/GLOBUS/a.h5
-
- The writing really depends on your web server. You have to set up your
- server in right way to accept writing in files. We have tested it using
- Apache Server(1.3.12) without authentication. If you need more details
- about our testing, please contact us. Globus suggests using their GASS
- server.
-
- There is another program called 'gass_append' used for experiments.