summaryrefslogtreecommitdiffstats
path: root/test/dsets.c
diff options
context:
space:
mode:
authorQuincey Koziol <koziol@hdfgroup.org>2010-01-30 04:29:13 (GMT)
committerQuincey Koziol <koziol@hdfgroup.org>2010-01-30 04:29:13 (GMT)
commitfd70b2afa883f94718ffb7f4f33d104d76e3fe0a (patch)
treec1add8db2a4848202d86a9b274bfaf8c7b80e961 /test/dsets.c
parent35b0159a0a5f1f4b80e305204ea51a742b052403 (diff)
downloadhdf5-fd70b2afa883f94718ffb7f4f33d104d76e3fe0a.zip
hdf5-fd70b2afa883f94718ffb7f4f33d104d76e3fe0a.tar.gz
hdf5-fd70b2afa883f94718ffb7f4f33d104d76e3fe0a.tar.bz2
[svn-r18197] Description:
Trim trailing whitespace from source code files with this command: find . \( -name "*.[ch]" -or -name "*.cpp" -or -name "*.f90" \) -print |xargs -n 1 sed -i "" 's/[[:blank:]]*$//' Tested on: None - eyeballed only
Diffstat (limited to 'test/dsets.c')
-rw-r--r--test/dsets.c56
1 files changed, 28 insertions, 28 deletions
diff --git a/test/dsets.c b/test/dsets.c
index 1041385..9c70875 100644
--- a/test/dsets.c
+++ b/test/dsets.c
@@ -183,7 +183,7 @@ const char *FILENAME[] = {
#define BYPASS_DIM 1000
#define BYPASS_CHUNK_DIM 500
#define BYPASS_FILL_VALUE 7
-
+
/* Shared global arrays */
#define DSET_DIM1 100
#define DSET_DIM2 200
@@ -813,7 +813,7 @@ error:
HDfree(wbuf);
if(rbuf)
HDfree(rbuf);
-
+
H5E_BEGIN_TRY {
/* Close file */
H5Sclose(space);
@@ -1051,16 +1051,16 @@ test_tconv(hid_t file)
return 0;
error:
- if(out)
+ if(out)
HDfree(out);
- if(in)
+ if(in)
HDfree(in);
-
+
H5E_BEGIN_TRY {
H5Dclose(dataset);
H5Sclose(space);
} H5E_END_TRY;
-
+
return -1;
}
@@ -1730,7 +1730,7 @@ error:
*
* Modifications:
* Make copy of data file since the test writes to the file.
- * Larry Knox, October 14, 2009
+ * Larry Knox, October 14, 2009
*
*-------------------------------------------------------------------------
*/
@@ -1748,12 +1748,12 @@ test_filter_noencoder(const char *dset_name)
int test_ints[10] = { 12 };
int read_buf[10];
int i;
-
+
/* Make a local copy of the file since this test writes to the data file
- from svn. */
- if (h5_make_local_copy(NOENCODER_FILENAME, NOENCODER_COPY_FILENAME) < 0)
+ from svn. */
+ if (h5_make_local_copy(NOENCODER_FILENAME, NOENCODER_COPY_FILENAME) < 0)
goto error;
-
+
/* Open file */
file_id = H5Fopen(NOENCODER_COPY_FILENAME, H5F_ACC_RDWR, H5P_DEFAULT);
if(file_id < 0) goto error;
@@ -5561,9 +5561,9 @@ error:
/*-------------------------------------------------------------------------
* Function: test_copy_dcpl
*
- * Purpose: Verifies whether the copy of dataset creation property
+ * Purpose: Verifies whether the copy of dataset creation property
* list works. It tests the DCPL for chunked layout with
- * filter and for contiguous layout with external storage.
+ * filter and for contiguous layout with external storage.
* (Please see #1608 in Bugzilla)
*
* Return: Success: 0
@@ -5599,7 +5599,7 @@ test_copy_dcpl(hid_t file, hid_t fapl)
if(H5Pset_fletcher32(dcpl) < 0) TEST_ERROR
/* Create first dataset of chunking with filter */
- if((dsid1 = H5Dcreate2(file, DSET_COPY_DCPL_NAME_1, H5T_NATIVE_INT, sid, H5P_DEFAULT, dcpl,
+ if((dsid1 = H5Dcreate2(file, DSET_COPY_DCPL_NAME_1, H5T_NATIVE_INT, sid, H5P_DEFAULT, dcpl,
H5P_DEFAULT)) < 0) TEST_ERROR
/* Close dataset */
@@ -5624,7 +5624,7 @@ test_copy_dcpl(hid_t file, hid_t fapl)
if(H5Pset_external(dcpl, COPY_DCPL_EXTFILE_NAME, 0, 500*4096*sizeof(int)) < 0) TEST_ERROR
/* Create second dataset of contiguous layout with external storage */
- if((dsid2 = H5Dcreate2(file, DSET_COPY_DCPL_NAME_2, H5T_NATIVE_INT, sid, H5P_DEFAULT, dcpl,
+ if((dsid2 = H5Dcreate2(file, DSET_COPY_DCPL_NAME_2, H5T_NATIVE_INT, sid, H5P_DEFAULT, dcpl,
H5P_DEFAULT)) < 0) TEST_ERROR
/* Close dataset */
@@ -5641,7 +5641,7 @@ test_copy_dcpl(hid_t file, hid_t fapl)
if(H5Dclose (dsid2) < 0) TEST_ERROR
/* Create a second file and create 2 datasets with the copies of the DCPLs in the first
- * file. Test whether the copies of DCPLs work. */
+ * file. Test whether the copies of DCPLs work. */
h5_fixname(FILENAME[11], fapl, filename, sizeof filename);
if((new_file = H5Fcreate(filename, H5F_ACC_TRUNC, H5P_DEFAULT, H5P_DEFAULT)) < 0)
TEST_ERROR
@@ -5657,10 +5657,10 @@ test_copy_dcpl(hid_t file, hid_t fapl)
/* Close datasets */
if(H5Dclose (new_dsid1) < 0) TEST_ERROR
- if(H5Dclose (new_dsid2) < 0) TEST_ERROR
+ if(H5Dclose (new_dsid2) < 0) TEST_ERROR
/* Close the second file */
- if(H5Fclose (new_file) < 0) TEST_ERROR
+ if(H5Fclose (new_file) < 0) TEST_ERROR
/* Close dataset creation property lists */
if(H5Pclose(dcpl) < 0) TEST_ERROR
@@ -6906,7 +6906,7 @@ error:
*
* Purpose: When the chunk size is bigger than the cache size and the
* chunk isn't on disk, this test verifies that the library
- * bypasses the cache.
+ * bypasses the cache.
*
* Note: This test is not very conclusive - it doesn't actually check
* if the chunks bypass the cache... :-( -QAK
@@ -6932,7 +6932,7 @@ test_big_chunks_bypass_cache(hid_t fapl)
size_t rdcc_nelmts, rdcc_nbytes;
int fvalue = BYPASS_FILL_VALUE;
hsize_t count, stride, offset, block;
- static int wdata[BYPASS_CHUNK_DIM/2], rdata1[BYPASS_DIM],
+ static int wdata[BYPASS_CHUNK_DIM/2], rdata1[BYPASS_DIM],
rdata2[BYPASS_CHUNK_DIM/2];
int i, j;
@@ -6976,7 +6976,7 @@ test_big_chunks_bypass_cache(hid_t fapl)
count = 1;
stride = 1;
block = BYPASS_CHUNK_DIM / 2;
- if(H5Sselect_hyperslab(sid, H5S_SELECT_SET, &offset, &stride, &count, &block) < 0)
+ if(H5Sselect_hyperslab(sid, H5S_SELECT_SET, &offset, &stride, &count, &block) < 0)
FAIL_STACK_ERROR
/* Initialize data to write */
@@ -6992,9 +6992,9 @@ test_big_chunks_bypass_cache(hid_t fapl)
/* Reopen the dataset */
if((dsid = H5Dopen2(fid, BYPASS_DATASET1, H5P_DEFAULT)) < 0) FAIL_STACK_ERROR
- /* Reads both 2 chunks. Reading the second chunk should bypass the cache because the
+ /* Reads both 2 chunks. Reading the second chunk should bypass the cache because the
* chunk is bigger than the cache size and it isn't allocated on disk. */
- if(H5Dread(dsid, H5T_NATIVE_INT, H5S_ALL, H5S_ALL, H5P_DEFAULT, rdata1) < 0)
+ if(H5Dread(dsid, H5T_NATIVE_INT, H5S_ALL, H5S_ALL, H5P_DEFAULT, rdata1) < 0)
FAIL_STACK_ERROR
for(i = 0; i < BYPASS_CHUNK_DIM / 2; i++)
@@ -7013,9 +7013,9 @@ test_big_chunks_bypass_cache(hid_t fapl)
/* Close the first dataset */
if(H5Dclose(dsid) < 0) FAIL_STACK_ERROR
-
+
/* Create a second dataset without fill value. This time, both write
- * and read should bypass the cache because the chunk is bigger than the
+ * and read should bypass the cache because the chunk is bigger than the
* cache size and it's not allocated on disk. */
if(H5Pset_fill_time(dcpl, H5D_FILL_TIME_NEVER) < 0) FAIL_STACK_ERROR
@@ -7030,10 +7030,10 @@ test_big_chunks_bypass_cache(hid_t fapl)
/* Reopen the dataset */
if((dsid = H5Dopen2(fid, BYPASS_DATASET2, H5P_DEFAULT)) < 0) FAIL_STACK_ERROR
- /* Read back only the part that was written to the file. Reading the
+ /* Read back only the part that was written to the file. Reading the
* half chunk should bypass the cache because the chunk is bigger than
* the cache size. */
- if(H5Dread(dsid, H5T_NATIVE_INT, H5S_ALL, sid, H5P_DEFAULT, rdata2) < 0)
+ if(H5Dread(dsid, H5T_NATIVE_INT, H5S_ALL, sid, H5P_DEFAULT, rdata2) < 0)
for(i = 0; i < BYPASS_CHUNK_DIM / 2; i++)
if(rdata2[i] != i) {
@@ -7547,7 +7547,7 @@ main(void)
printf("All dataset tests passed.\n");
#ifdef H5_HAVE_FILTER_SZIP
if (GetTestCleanup())
- HDremove(NOENCODER_COPY_FILENAME);
+ HDremove(NOENCODER_COPY_FILENAME);
#endif /* H5_HAVE_FILTER_SZIP */
h5_cleanup(FILENAME, fapl);