1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
|
! * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * *
! Copyright by The HDF Group. *
! All rights reserved. *
! *
! This file is part of HDF5. The full HDF5 copyright notice, including *
! terms governing use, modification, and redistribution, is contained in *
! the COPYING file, which can be found at the root of the source code *
! distribution tree, or in https://www.hdfgroup.org/licenses. *
! If you do not have access to either file, you may request a copy from *
! help@hdfgroup.org. *
! * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * *
!
! writes/reads dataset by hyperslabs
!
SUBROUTINE hyper(length,do_collective,do_chunk, mpi_size, mpi_rank, nerrors)
USE HDF5
USE MPI
USE TH5_MISC
IMPLICIT NONE
INTEGER, INTENT(in) :: length ! array length
LOGICAL, INTENT(in) :: do_collective ! use collective I/O
LOGICAL, INTENT(in) :: do_chunk ! use chunking
INTEGER, INTENT(in) :: mpi_size ! number of processes in the group of communicator
INTEGER, INTENT(in) :: mpi_rank ! rank of the calling process in the communicator
INTEGER, INTENT(inout) :: nerrors ! number of errors
INTEGER :: hdferror ! HDF hdferror flag
INTEGER(hsize_t), DIMENSION(1) :: dims ! dataset dimensions
INTEGER(hsize_t), DIMENSION(1) :: cdims ! chunk dimensions
INTEGER, ALLOCATABLE :: wbuf(:) ! write buffer
INTEGER, ALLOCATABLE :: rbuf(:) ! read buffer
INTEGER(hsize_t), DIMENSION(1) :: counti ! hyperslab selection
INTEGER(hsize_t), DIMENSION(1) :: start ! hyperslab selection
INTEGER(hid_t) :: fapl_id ! file access identifier
INTEGER(hid_t) :: dxpl_id ! dataset transfer property list
INTEGER(hid_t) :: dcpl_id ! dataset creation property list
INTEGER(hid_t) :: file_id ! file identifier
INTEGER(hid_t) :: dset_id ! dataset identifier
INTEGER(hid_t) :: fspace_id ! file space identifier
INTEGER(hid_t) :: mspace_id ! memory space identifier
INTEGER(hid_t) :: driver_id ! low-level file driver identifier
INTEGER :: istart ! start position in array
INTEGER :: iend ! end position in array
INTEGER :: icount ! number of elements in array
CHARACTER(len=80) :: filename ! filename
INTEGER :: i
INTEGER :: actual_io_mode ! The type of I/O performed by this process
LOGICAL :: is_coll
LOGICAL :: is_coll_true = .TRUE.
!
! initialize the array data between the processes (3)
! for the 12 size array we get
! p0 = 1,2,3,4
! p1 = 5,6,7,8
! p2 = 9,10,11,12
!
ALLOCATE(wbuf(0:length-1),stat=hdferror)
IF (hdferror /= 0) THEN
WRITE(*,*) 'allocate error'
RETURN
ENDIF
ALLOCATE(rbuf(0:length-1),stat=hdferror)
IF (hdferror /= 0) THEN
WRITE(*,*) 'allocate error'
RETURN
ENDIF
icount = length/mpi_size ! divide the array by the number of processes
istart = mpi_rank*icount ! start position
iend = istart + icount ! end position
DO i = istart, iend-1
wbuf(i) = i
ENDDO
!
! HDF5 I/O
!
dims(1) = length
cdims(1) = length/mpi_size ! define chunks as the number of processes
!
! setup file access property list with parallel I/O access
!
CALL h5pcreate_f(H5P_FILE_ACCESS_F, fapl_id, hdferror)
CALL check("h5pcreate_f", hdferror, nerrors)
CALL h5pset_fapl_mpio_f(fapl_id, MPI_COMM_WORLD, MPI_INFO_NULL, hdferror)
CALL check("h5pset_fapl_mpio_f", hdferror, nerrors)
CALL h5pget_driver_f(fapl_id, driver_id, hdferror)
CALL check("h5pget_driver_f", hdferror, nerrors)
IF( driver_id /= H5FD_MPIO_F) THEN
WRITE(*,*) "Wrong driver information returned"
nerrors = nerrors + 1
ENDIF
!
! create the file collectively
!
CALL h5_fixname_f("parf1", filename, fapl_id, hdferror)
IF(do_collective)THEN
! verify settings for file access properties
! Collective metadata writes
CALL h5pget_coll_metadata_write_f(fapl_id, is_coll, hdferror)
CALL check("h5pget_coll_metadata_write_f", hdferror, nerrors)
IF(is_coll .NEQV. .FALSE.)THEN
PRINT*, "Incorrect property setting for coll metadata writes"
nerrors = nerrors + 1
ENDIF
! Collective metadata read API calling requirement
CALL h5pget_all_coll_metadata_ops_f(fapl_id, is_coll, hdferror)
CALL check("h5pget_all_coll_metadata_ops_f", hdferror, nerrors)
IF(is_coll .NEQV. .FALSE.)THEN
PRINT*, "Incorrect property setting for coll metadata API calls requirement"
nerrors = nerrors + 1
ENDIF
! Collective metadata writes
CALL h5pset_coll_metadata_write_f(fapl_id, .TRUE., hdferror)
CALL check("h5pset_coll_metadata_write_f", hdferror, nerrors)
! Collective metadata READ API calling requirement
CALL h5pset_all_coll_metadata_ops_f(fapl_id, is_coll_true, hdferror)
CALL check("h5pset_all_coll_metadata_ops_f", hdferror, nerrors)
CALL h5fcreate_f(filename, H5F_ACC_TRUNC_F, file_id, hdferror, access_prp = fapl_id)
CALL check("h5fcreate_f", hdferror, nerrors)
! close fapl and retrieve it from file
CALL h5pclose_f(fapl_id, hdferror)
CALL check("h5pclose_f", hdferror, nerrors)
CALL h5fget_access_plist_f(file_id, fapl_id, hdferror)
CALL check("h5fget_access_plist_f", hdferror, nerrors)
! verify settings for file access properties
! Collective metadata writes
CALL h5pget_coll_metadata_write_f(fapl_id, is_coll, hdferror)
CALL check("h5pget_coll_metadata_write_f", hdferror, nerrors)
IF(is_coll .NEQV. .TRUE.)THEN
PRINT*, "Incorrect property setting for coll metadata writes"
nerrors = nerrors + 1
ENDIF
! Collective metadata read API calling requirement
CALL h5pget_all_coll_metadata_ops_f(fapl_id, is_coll, hdferror)
CALL check("h5pget_all_coll_metadata_ops_f", hdferror, nerrors)
IF(is_coll .NEQV. .TRUE.)THEN
PRINT*, "Incorrect property setting for coll metadata API calls requirement"
nerrors = nerrors + 1
ENDIF
ELSE
CALL h5fcreate_f(filename, H5F_ACC_TRUNC_F, file_id, hdferror, access_prp = fapl_id)
CALL check("h5fcreate_f", hdferror, nerrors)
ENDIF
CALL h5screate_simple_f(1, dims, fspace_id, hdferror)
CALL check("h5screate_simple_f", hdferror, nerrors)
CALL h5screate_simple_f(1, dims, mspace_id, hdferror)
CALL check("h5screate_simple_f", hdferror, nerrors)
!
! modify dataset creation properties to enable chunking
!
CALL h5pcreate_f(H5P_DATASET_CREATE_F, dcpl_id, hdferror)
CALL check("h5pcreate_f", hdferror, nerrors)
IF (do_chunk) THEN
CALL h5pset_chunk_f(dcpl_id, 1, cdims, hdferror)
CALL check("h5pset_chunk_f", hdferror, nerrors)
ENDIF
!
! create the dataset
!
CALL h5dcreate_f(file_id, "dset", H5T_NATIVE_INTEGER, fspace_id, dset_id, hdferror, dcpl_id)
CALL check("h5dcreate_f", hdferror, nerrors)
!
! define hyperslab
!
counti(1) = icount
start(1) = istart
!
! select hyperslab in memory
!
CALL h5sselect_hyperslab_f(mspace_id, H5S_SELECT_SET_F, start, counti, hdferror)
CALL check("h5sselect_hyperslab_f", hdferror, nerrors)
!
! select hyperslab in the file
!
CALL h5sselect_hyperslab_f(fspace_id, H5S_SELECT_SET_F, start, counti, hdferror)
CALL check("h5sselect_hyperslab_f", hdferror, nerrors)
!
! create a property list for collective dataset write
!
CALL h5pcreate_f(H5P_DATASET_XFER_F, dxpl_id, hdferror)
CALL check("h5pcreate_f", hdferror, nerrors)
IF (do_collective) THEN
CALL h5pset_dxpl_mpio_f(dxpl_id, H5FD_MPIO_COLLECTIVE_F, hdferror)
CALL check("h5pset_dxpl_mpio_f", hdferror, nerrors)
ENDIF
!
! write dataset
!
CALL h5dwrite_f(dset_id,H5T_NATIVE_INTEGER,wbuf,dims,hdferror,file_space_id=fspace_id,mem_space_id=mspace_id,xfer_prp=dxpl_id)
CALL check("h5dwrite_f", hdferror, nerrors)
! Check h5pget_mpio_actual_io_mode_f function
CALL h5pget_mpio_actual_io_mode_f(dxpl_id, actual_io_mode, hdferror)
CALL check("h5pget_mpio_actual_io_mode_f", hdferror, nerrors)
! MSB -- TODO FIX: skipping for now since multi-dataset
! has no specific path for contiguous collective
!
! IF(do_collective.AND.do_chunk)THEN
! IF(actual_io_mode.NE.H5D_MPIO_CHUNK_COLLECTIVE_F)THEN
! CALL check("h5pget_mpio_actual_io_mode_f", -1, nerrors)
! ENDIF
! ELSEIF(.NOT.do_collective)THEN
! IF(actual_io_mode.NE.H5D_MPIO_NO_COLLECTIVE_F)THEN
! CALL check("h5pget_mpio_actual_io_mode_f", -1, nerrors)
! ENDIF
! ELSEIF( do_collective.AND.(.NOT.do_chunk))THEN
! IF(actual_io_mode.NE.H5D_MPIO_CONTIG_COLLECTIVE_F)THEN
! CALL check("h5pget_mpio_actual_io_mode_f", -1, nerrors)
! ENDIF
! ENDIF
! MSB
!
! close HDF5 I/O
!
CALL h5pclose_f(fapl_id, hdferror)
CALL check("h5pclose_f", hdferror, nerrors)
CALL h5pclose_f(dcpl_id, hdferror)
CALL check("h5pclose_f", hdferror, nerrors)
CALL h5pclose_f(dxpl_id, hdferror)
CALL check("h5pclose_f", hdferror, nerrors)
CALL h5sclose_f(mspace_id, hdferror)
CALL check("h5sclose_f", hdferror, nerrors)
CALL h5sclose_f(fspace_id, hdferror)
CALL check("h5sclose_f", hdferror, nerrors)
CALL h5dclose_f(dset_id, hdferror)
CALL check("h5dclose_f", hdferror, nerrors)
CALL h5fclose_f(file_id, hdferror)
CALL check("h5fclose_f", hdferror, nerrors)
!
! reopen file with read access
!
CALL h5pcreate_f(H5P_FILE_ACCESS_F, fapl_id, hdferror)
CALL check("h5pcreate_f", hdferror, nerrors)
CALL h5pset_fapl_mpio_f(fapl_id, MPI_COMM_WORLD, MPI_INFO_NULL, hdferror)
CALL check("h5pset_fapl_mpio_f", hdferror, nerrors)
CALL h5fopen_f(filename, H5F_ACC_RDWR_F, file_id, hdferror, access_prp = fapl_id)
CALL check("h5fopen_f", hdferror, nerrors)
CALL h5screate_simple_f(1, dims, fspace_id, hdferror)
CALL check("h5screate_simple_f", hdferror, nerrors)
CALL h5screate_simple_f(1, dims, mspace_id, hdferror)
CALL check(" h5screate_simple_f", hdferror, nerrors)
CALL h5dopen_f(file_id, "dset", dset_id, hdferror)
CALL check("h5dopen_f", hdferror, nerrors)
!
! select hyperslab in memory
!
CALL h5sselect_hyperslab_f(mspace_id, H5S_SELECT_SET_F, start, counti, hdferror)
CALL check("h5sselect_hyperslab_f", hdferror, nerrors)
!
! select hyperslab in the file
!
CALL h5sselect_hyperslab_f(fspace_id, H5S_SELECT_SET_F, start, counti, hdferror)
CALL check("h5sselect_hyperslab_f", hdferror, nerrors)
!
! create a property list for collective dataset read
!
CALL h5pcreate_f(H5P_DATASET_XFER_F, dxpl_id, hdferror)
CALL check("h5pcreate_f", hdferror, nerrors)
IF (do_collective) THEN
CALL h5pset_dxpl_mpio_f(dxpl_id, H5FD_MPIO_COLLECTIVE_F, hdferror)
CALL check("h5pset_dxpl_mpio_f", hdferror, nerrors)
ENDIF
!
! read dataset
!
CALL h5dread_f(dset_id,H5T_NATIVE_INTEGER,rbuf,dims,hdferror,file_space_id=fspace_id,mem_space_id=mspace_id,xfer_prp=dxpl_id)
CALL check("h5dread_f", hdferror, nerrors)
!
! close HDF5 I/O
!
CALL h5pclose_f(fapl_id, hdferror)
CALL check("h5pclose_f", hdferror, nerrors)
CALL h5pclose_f(dxpl_id, hdferror)
CALL check("h5pclose_f", hdferror, nerrors)
CALL h5sclose_f(fspace_id, hdferror)
CALL check("h5sclose_f", hdferror, nerrors)
CALL h5sclose_f(mspace_id, hdferror)
CALL check("h5sclose_f", hdferror, nerrors)
CALL h5dclose_f(dset_id, hdferror)
CALL check("h5dclose_f", hdferror, nerrors)
CALL h5fclose_f(file_id, hdferror)
CALL check("h5fclose_f", hdferror, nerrors)
!
! compare read and write data. each process compares a subset of the array
!
DO i = istart, iend-1
IF( wbuf(i) /= rbuf(i)) THEN
WRITE(*,*) 'buffers differs at ', i, rbuf(i), wbuf(i)
nerrors = nerrors + 1
ENDIF
ENDDO
DEALLOCATE(wbuf)
DEALLOCATE(rbuf)
END SUBROUTINE hyper
|