summaryrefslogtreecommitdiffstats
path: root/Tools
diff options
context:
space:
mode:
Diffstat (limited to 'Tools')
-rw-r--r--Tools/webchecker/README23
-rw-r--r--Tools/webchecker/tktools.py366
-rwxr-xr-xTools/webchecker/wcgui.py456
-rw-r--r--Tools/webchecker/wcmac.py9
-rwxr-xr-xTools/webchecker/webchecker.py890
-rwxr-xr-xTools/webchecker/websucker.py123
-rwxr-xr-xTools/webchecker/wsgui.py240
7 files changed, 0 insertions, 2107 deletions
diff --git a/Tools/webchecker/README b/Tools/webchecker/README
deleted file mode 100644
index a51bb3d..0000000
--- a/Tools/webchecker/README
+++ /dev/null
@@ -1,23 +0,0 @@
-Webchecker
-----------
-
-This is a simple web tree checker, useful to find bad links in a web
-tree. It currently checks links pointing within the same subweb for
-validity. The main program is "webchecker.py". See its doc string
-(or invoke it with the option "-?") for more defails.
-
-History:
-
-- Jan 1997. First release. The module robotparser.py was written by
-Skip Montanaro; the rest is original work by Guido van Rossum.
-
-- May 1999. Sam Bayer contributed a new version, wcnew.py, which
-supports checking internal links (#spam fragments in URLs) and some
-other options.
-
-- Nov 1999. Sam Bayer contributed patches to reintegrate wcnew.py
-into webchecker.py, and corresponding mods to wcgui.py and
-websucker.py.
-
-- Mar 2004. Chris Herborth contributed a patch to let webchecker.py
-handle XHTML's 'id' attribute.
diff --git a/Tools/webchecker/tktools.py b/Tools/webchecker/tktools.py
deleted file mode 100644
index 3a68f9a..0000000
--- a/Tools/webchecker/tktools.py
+++ /dev/null
@@ -1,366 +0,0 @@
-"""Assorted Tk-related subroutines used in Grail."""
-
-
-from types import *
-from Tkinter import *
-
-def _clear_entry_widget(event):
- try:
- widget = event.widget
- widget.delete(0, INSERT)
- except: pass
-def install_keybindings(root):
- root.bind_class('Entry', '<Control-u>', _clear_entry_widget)
-
-
-def make_toplevel(master, title=None, class_=None):
- """Create a Toplevel widget.
-
- This is a shortcut for a Toplevel() instantiation plus calls to
- set the title and icon name of the widget.
-
- """
-
- if class_:
- widget = Toplevel(master, class_=class_)
- else:
- widget = Toplevel(master)
- if title:
- widget.title(title)
- widget.iconname(title)
- return widget
-
-def set_transient(widget, master, relx=0.5, rely=0.3, expose=1):
- """Make an existing toplevel widget transient for a master.
-
- The widget must exist but should not yet have been placed; in
- other words, this should be called after creating all the
- subwidget but before letting the user interact.
- """
-
- widget.withdraw() # Remain invisible while we figure out the geometry
- widget.transient(master)
- widget.update_idletasks() # Actualize geometry information
- if master.winfo_ismapped():
- m_width = master.winfo_width()
- m_height = master.winfo_height()
- m_x = master.winfo_rootx()
- m_y = master.winfo_rooty()
- else:
- m_width = master.winfo_screenwidth()
- m_height = master.winfo_screenheight()
- m_x = m_y = 0
- w_width = widget.winfo_reqwidth()
- w_height = widget.winfo_reqheight()
- x = m_x + (m_width - w_width) * relx
- y = m_y + (m_height - w_height) * rely
- widget.geometry("+%d+%d" % (x, y))
- if expose:
- widget.deiconify() # Become visible at the desired location
- return widget
-
-
-def make_scrollbars(parent, hbar, vbar, pack=1, class_=None, name=None,
- takefocus=0):
-
- """Subroutine to create a frame with scrollbars.
-
- This is used by make_text_box and similar routines.
-
- Note: the caller is responsible for setting the x/y scroll command
- properties (e.g. by calling set_scroll_commands()).
-
- Return a tuple containing the hbar, the vbar, and the frame, where
- hbar and vbar are None if not requested.
-
- """
- if class_:
- if name: frame = Frame(parent, class_=class_, name=name)
- else: frame = Frame(parent, class_=class_)
- else:
- if name: frame = Frame(parent, name=name)
- else: frame = Frame(parent)
-
- if pack:
- frame.pack(fill=BOTH, expand=1)
-
- corner = None
- if vbar:
- if not hbar:
- vbar = Scrollbar(frame, takefocus=takefocus)
- vbar.pack(fill=Y, side=RIGHT)
- else:
- vbarframe = Frame(frame, borderwidth=0)
- vbarframe.pack(fill=Y, side=RIGHT)
- vbar = Scrollbar(frame, name="vbar", takefocus=takefocus)
- vbar.pack(in_=vbarframe, expand=1, fill=Y, side=TOP)
- sbwidth = vbar.winfo_reqwidth()
- corner = Frame(vbarframe, width=sbwidth, height=sbwidth)
- corner.propagate(0)
- corner.pack(side=BOTTOM)
- else:
- vbar = None
-
- if hbar:
- hbar = Scrollbar(frame, orient=HORIZONTAL, name="hbar",
- takefocus=takefocus)
- hbar.pack(fill=X, side=BOTTOM)
- else:
- hbar = None
-
- return hbar, vbar, frame
-
-
-def set_scroll_commands(widget, hbar, vbar):
-
- """Link a scrollable widget to its scroll bars.
-
- The scroll bars may be empty.
-
- """
-
- if vbar:
- widget['yscrollcommand'] = (vbar, 'set')
- vbar['command'] = (widget, 'yview')
-
- if hbar:
- widget['xscrollcommand'] = (hbar, 'set')
- hbar['command'] = (widget, 'xview')
-
- widget.vbar = vbar
- widget.hbar = hbar
-
-
-def make_text_box(parent, width=0, height=0, hbar=0, vbar=1,
- fill=BOTH, expand=1, wrap=WORD, pack=1,
- class_=None, name=None, takefocus=None):
-
- """Subroutine to create a text box.
-
- Create:
- - a both-ways filling and expanding frame, containing:
- - a text widget on the left, and
- - possibly a vertical scroll bar on the right.
- - possibly a horizonta; scroll bar at the bottom.
-
- Return the text widget and the frame widget.
-
- """
- hbar, vbar, frame = make_scrollbars(parent, hbar, vbar, pack,
- class_=class_, name=name,
- takefocus=takefocus)
-
- widget = Text(frame, wrap=wrap, name="text")
- if width: widget.config(width=width)
- if height: widget.config(height=height)
- widget.pack(expand=expand, fill=fill, side=LEFT)
-
- set_scroll_commands(widget, hbar, vbar)
-
- return widget, frame
-
-
-def make_list_box(parent, width=0, height=0, hbar=0, vbar=1,
- fill=BOTH, expand=1, pack=1, class_=None, name=None,
- takefocus=None):
-
- """Subroutine to create a list box.
-
- Like make_text_box().
- """
- hbar, vbar, frame = make_scrollbars(parent, hbar, vbar, pack,
- class_=class_, name=name,
- takefocus=takefocus)
-
- widget = Listbox(frame, name="listbox")
- if width: widget.config(width=width)
- if height: widget.config(height=height)
- widget.pack(expand=expand, fill=fill, side=LEFT)
-
- set_scroll_commands(widget, hbar, vbar)
-
- return widget, frame
-
-
-def make_canvas(parent, width=0, height=0, hbar=1, vbar=1,
- fill=BOTH, expand=1, pack=1, class_=None, name=None,
- takefocus=None):
-
- """Subroutine to create a canvas.
-
- Like make_text_box().
-
- """
-
- hbar, vbar, frame = make_scrollbars(parent, hbar, vbar, pack,
- class_=class_, name=name,
- takefocus=takefocus)
-
- widget = Canvas(frame, scrollregion=(0, 0, width, height), name="canvas")
- if width: widget.config(width=width)
- if height: widget.config(height=height)
- widget.pack(expand=expand, fill=fill, side=LEFT)
-
- set_scroll_commands(widget, hbar, vbar)
-
- return widget, frame
-
-
-
-def make_form_entry(parent, label, borderwidth=None):
-
- """Subroutine to create a form entry.
-
- Create:
- - a horizontally filling and expanding frame, containing:
- - a label on the left, and
- - a text entry on the right.
-
- Return the entry widget and the frame widget.
-
- """
-
- frame = Frame(parent)
- frame.pack(fill=X)
-
- label = Label(frame, text=label)
- label.pack(side=LEFT)
-
- if borderwidth is None:
- entry = Entry(frame, relief=SUNKEN)
- else:
- entry = Entry(frame, relief=SUNKEN, borderwidth=borderwidth)
- entry.pack(side=LEFT, fill=X, expand=1)
-
- return entry, frame
-
-# This is a slightly modified version of the function above. This
-# version does the proper alighnment of labels with their fields. It
-# should probably eventually replace make_form_entry altogether.
-#
-# The one annoying bug is that the text entry field should be
-# expandable while still aligning the colons. This doesn't work yet.
-#
-def make_labeled_form_entry(parent, label, entrywidth=20, entryheight=1,
- labelwidth=0, borderwidth=None,
- takefocus=None):
- """Subroutine to create a form entry.
-
- Create:
- - a horizontally filling and expanding frame, containing:
- - a label on the left, and
- - a text entry on the right.
-
- Return the entry widget and the frame widget.
- """
- if label and label[-1] != ':': label = label + ':'
-
- frame = Frame(parent)
-
- label = Label(frame, text=label, width=labelwidth, anchor=E)
- label.pack(side=LEFT)
- if entryheight == 1:
- if borderwidth is None:
- entry = Entry(frame, relief=SUNKEN, width=entrywidth)
- else:
- entry = Entry(frame, relief=SUNKEN, width=entrywidth,
- borderwidth=borderwidth)
- entry.pack(side=RIGHT, expand=1, fill=X)
- frame.pack(fill=X)
- else:
- entry = make_text_box(frame, entrywidth, entryheight, 1, 1,
- takefocus=takefocus)
- frame.pack(fill=BOTH, expand=1)
-
- return entry, frame, label
-
-
-def make_double_frame(master=None, class_=None, name=None, relief=RAISED,
- borderwidth=1):
- """Create a pair of frames suitable for 'hosting' a dialog."""
- if name:
- if class_: frame = Frame(master, class_=class_, name=name)
- else: frame = Frame(master, name=name)
- else:
- if class_: frame = Frame(master, class_=class_)
- else: frame = Frame(master)
- top = Frame(frame, name="topframe", relief=relief,
- borderwidth=borderwidth)
- bottom = Frame(frame, name="bottomframe")
- bottom.pack(fill=X, padx='1m', pady='1m', side=BOTTOM)
- top.pack(expand=1, fill=BOTH, padx='1m', pady='1m')
- frame.pack(expand=1, fill=BOTH)
- top = Frame(top)
- top.pack(expand=1, fill=BOTH, padx='2m', pady='2m')
-
- return frame, top, bottom
-
-
-def make_group_frame(master, name=None, label=None, fill=Y,
- side=None, expand=None, font=None):
- """Create nested frames with a border and optional label.
-
- The outer frame is only used to provide the decorative border, to
- control packing, and to host the label. The inner frame is packed
- to fill the outer frame and should be used as the parent of all
- sub-widgets. Only the inner frame is returned.
-
- """
- font = font or "-*-helvetica-medium-r-normal-*-*-100-*-*-*-*-*-*"
- outer = Frame(master, borderwidth=2, relief=GROOVE)
- outer.pack(expand=expand, fill=fill, side=side)
- if label:
- Label(outer, text=label, font=font, anchor=W).pack(fill=X)
- inner = Frame(master, borderwidth='1m', name=name)
- inner.pack(expand=1, fill=BOTH, in_=outer)
- inner.forget = outer.forget
- return inner
-
-
-def unify_button_widths(*buttons):
- """Make buttons passed in all have the same width.
-
- Works for labels and other widgets with the 'text' option.
-
- """
- wid = 0
- for btn in buttons:
- wid = max(wid, len(btn["text"]))
- for btn in buttons:
- btn["width"] = wid
-
-
-def flatten(msg):
- """Turn a list or tuple into a single string -- recursively."""
- t = type(msg)
- if t in (ListType, TupleType):
- msg = ' '.join(map(flatten, msg))
- elif t is ClassType:
- msg = msg.__name__
- else:
- msg = str(msg)
- return msg
-
-
-def boolean(s):
- """Test whether a string is a Tk boolean, without error checking."""
- if s.lower() in ('', '0', 'no', 'off', 'false'): return 0
- else: return 1
-
-
-def test():
- """Test make_text_box(), make_form_entry(), flatten(), boolean()."""
- import sys
- root = Tk()
- entry, eframe = make_form_entry(root, 'Boolean:')
- text, tframe = make_text_box(root)
- def enter(event, entry=entry, text=text):
- s = boolean(entry.get()) and '\nyes' or '\nno'
- text.insert('end', s)
- entry.bind('<Return>', enter)
- entry.insert(END, flatten(sys.argv))
- root.mainloop()
-
-
-if __name__ == '__main__':
- test()
diff --git a/Tools/webchecker/wcgui.py b/Tools/webchecker/wcgui.py
deleted file mode 100755
index b0c69d3..0000000
--- a/Tools/webchecker/wcgui.py
+++ /dev/null
@@ -1,456 +0,0 @@
-#! /usr/bin/env python3
-
-"""GUI interface to webchecker.
-
-This works as a Grail applet too! E.g.
-
- <APPLET CODE=wcgui.py NAME=CheckerWindow></APPLET>
-
-Checkpoints are not (yet??? ever???) supported.
-
-User interface:
-
-Enter a root to check in the text entry box. To enter more than one root,
-enter them one at a time and press <Return> for each one.
-
-Command buttons Start, Stop and "Check one" govern the checking process in
-the obvious way. Start and "Check one" also enter the root from the text
-entry box if one is present. There's also a check box (enabled by default)
-to decide whether actually to follow external links (since this can slow
-the checking down considerably). Finally there's a Quit button.
-
-A series of checkbuttons determines whether the corresponding output panel
-is shown. List panels are also automatically shown or hidden when their
-status changes between empty to non-empty. There are six panels:
-
-Log -- raw output from the checker (-v, -q affect this)
-To check -- links discovered but not yet checked
-Checked -- links that have been checked
-Bad links -- links that failed upon checking
-Errors -- pages containing at least one bad link
-Details -- details about one URL; double click on a URL in any of
- the above list panels (not in Log) will show details
- for that URL
-
-Use your window manager's Close command to quit.
-
-Command line options:
-
--m bytes -- skip HTML pages larger than this size (default %(MAXPAGE)d)
--q -- quiet operation (also suppresses external links report)
--v -- verbose operation; repeating -v will increase verbosity
--t root -- specify root dir which should be treated as internal (can repeat)
--a -- don't check name anchors
-
-Command line arguments:
-
-rooturl -- URL to start checking
- (default %(DEFROOT)s)
-
-XXX The command line options (-m, -q, -v) should be GUI accessible.
-
-XXX The roots should be visible as a list (?).
-
-XXX The multipanel user interface is clumsy.
-
-"""
-
-# ' Emacs bait
-
-
-import sys
-import getopt
-from Tkinter import *
-import tktools
-import webchecker
-
-def main():
- try:
- opts, args = getopt.getopt(sys.argv[1:], 't:m:qva')
- except getopt.error as msg:
- sys.stdout = sys.stderr
- print(msg)
- print(__doc__%vars(webchecker))
- sys.exit(2)
- webchecker.verbose = webchecker.VERBOSE
- webchecker.nonames = webchecker.NONAMES
- webchecker.maxpage = webchecker.MAXPAGE
- extra_roots = []
- for o, a in opts:
- if o == '-m':
- webchecker.maxpage = int(a)
- if o == '-q':
- webchecker.verbose = 0
- if o == '-v':
- webchecker.verbose = webchecker.verbose + 1
- if o == '-t':
- extra_roots.append(a)
- if o == '-a':
- webchecker.nonames = not webchecker.nonames
- root = Tk(className='Webchecker')
- root.protocol("WM_DELETE_WINDOW", root.quit)
- c = CheckerWindow(root)
- c.setflags(verbose=webchecker.verbose, maxpage=webchecker.maxpage,
- nonames=webchecker.nonames)
- if args:
- for arg in args[:-1]:
- c.addroot(arg)
- c.suggestroot(args[-1])
- # Usually conditioned on whether external links
- # will be checked, but since that's not a command
- # line option, just toss them in.
- for url_root in extra_roots:
- # Make sure it's terminated by a slash,
- # so that addroot doesn't discard the last
- # directory component.
- if url_root[-1] != "/":
- url_root = url_root + "/"
- c.addroot(url_root, add_to_do = 0)
- root.mainloop()
-
-
-class CheckerWindow(webchecker.Checker):
-
- def __init__(self, parent, root=webchecker.DEFROOT):
- self.__parent = parent
-
- self.__topcontrols = Frame(parent)
- self.__topcontrols.pack(side=TOP, fill=X)
- self.__label = Label(self.__topcontrols, text="Root URL:")
- self.__label.pack(side=LEFT)
- self.__rootentry = Entry(self.__topcontrols, width=60)
- self.__rootentry.pack(side=LEFT)
- self.__rootentry.bind('<Return>', self.enterroot)
- self.__rootentry.focus_set()
-
- self.__controls = Frame(parent)
- self.__controls.pack(side=TOP, fill=X)
- self.__running = 0
- self.__start = Button(self.__controls, text="Run", command=self.start)
- self.__start.pack(side=LEFT)
- self.__stop = Button(self.__controls, text="Stop", command=self.stop,
- state=DISABLED)
- self.__stop.pack(side=LEFT)
- self.__step = Button(self.__controls, text="Check one",
- command=self.step)
- self.__step.pack(side=LEFT)
- self.__cv = BooleanVar(parent)
- self.__cv.set(self.checkext)
- self.__checkext = Checkbutton(self.__controls, variable=self.__cv,
- command=self.update_checkext,
- text="Check nonlocal links",)
- self.__checkext.pack(side=LEFT)
- self.__reset = Button(self.__controls, text="Start over", command=self.reset)
- self.__reset.pack(side=LEFT)
- if __name__ == '__main__': # No Quit button under Grail!
- self.__quit = Button(self.__controls, text="Quit",
- command=self.__parent.quit)
- self.__quit.pack(side=RIGHT)
-
- self.__status = Label(parent, text="Status: initial", anchor=W)
- self.__status.pack(side=TOP, fill=X)
- self.__checking = Label(parent, text="Idle", anchor=W)
- self.__checking.pack(side=TOP, fill=X)
- self.__mp = mp = MultiPanel(parent)
- sys.stdout = self.__log = LogPanel(mp, "Log")
- self.__todo = ListPanel(mp, "To check", self, self.showinfo)
- self.__done = ListPanel(mp, "Checked", self, self.showinfo)
- self.__bad = ListPanel(mp, "Bad links", self, self.showinfo)
- self.__errors = ListPanel(mp, "Pages w/ bad links", self, self.showinfo)
- self.__details = LogPanel(mp, "Details")
- self.root_seed = None
- webchecker.Checker.__init__(self)
- if root:
- root = str(root).strip()
- if root:
- self.suggestroot(root)
- self.newstatus()
-
- def reset(self):
- webchecker.Checker.reset(self)
- for p in self.__todo, self.__done, self.__bad, self.__errors:
- p.clear()
- if self.root_seed:
- self.suggestroot(self.root_seed)
-
- def suggestroot(self, root):
- self.__rootentry.delete(0, END)
- self.__rootentry.insert(END, root)
- self.__rootentry.select_range(0, END)
- self.root_seed = root
-
- def enterroot(self, event=None):
- root = self.__rootentry.get()
- root = root.strip()
- if root:
- self.__checking.config(text="Adding root "+root)
- self.__checking.update_idletasks()
- self.addroot(root)
- self.__checking.config(text="Idle")
- try:
- i = self.__todo.items.index(root)
- except (ValueError, IndexError):
- pass
- else:
- self.__todo.list.select_clear(0, END)
- self.__todo.list.select_set(i)
- self.__todo.list.yview(i)
- self.__rootentry.delete(0, END)
-
- def start(self):
- self.__start.config(state=DISABLED, relief=SUNKEN)
- self.__stop.config(state=NORMAL)
- self.__step.config(state=DISABLED)
- self.enterroot()
- self.__running = 1
- self.go()
-
- def stop(self):
- self.__stop.config(state=DISABLED, relief=SUNKEN)
- self.__running = 0
-
- def step(self):
- self.__start.config(state=DISABLED)
- self.__step.config(state=DISABLED, relief=SUNKEN)
- self.enterroot()
- self.__running = 0
- self.dosomething()
-
- def go(self):
- if self.__running:
- self.__parent.after_idle(self.dosomething)
- else:
- self.__checking.config(text="Idle")
- self.__start.config(state=NORMAL, relief=RAISED)
- self.__stop.config(state=DISABLED, relief=RAISED)
- self.__step.config(state=NORMAL, relief=RAISED)
-
- __busy = 0
-
- def dosomething(self):
- if self.__busy: return
- self.__busy = 1
- if self.todo:
- l = self.__todo.selectedindices()
- if l:
- i = l[0]
- else:
- i = 0
- self.__todo.list.select_set(i)
- self.__todo.list.yview(i)
- url = self.__todo.items[i]
- self.__checking.config(text="Checking "+self.format_url(url))
- self.__parent.update()
- self.dopage(url)
- else:
- self.stop()
- self.__busy = 0
- self.go()
-
- def showinfo(self, url):
- d = self.__details
- d.clear()
- d.put("URL: %s\n" % self.format_url(url))
- if url in self.bad:
- d.put("Error: %s\n" % str(self.bad[url]))
- if url in self.roots:
- d.put("Note: This is a root URL\n")
- if url in self.done:
- d.put("Status: checked\n")
- o = self.done[url]
- elif url in self.todo:
- d.put("Status: to check\n")
- o = self.todo[url]
- else:
- d.put("Status: unknown (!)\n")
- o = []
- if (not url[1]) and url[0] in self.errors:
- d.put("Bad links from this page:\n")
- for triple in self.errors[url[0]]:
- link, rawlink, msg = triple
- d.put(" HREF %s" % self.format_url(link))
- if self.format_url(link) != rawlink: d.put(" (%s)" %rawlink)
- d.put("\n")
- d.put(" error %s\n" % str(msg))
- self.__mp.showpanel("Details")
- for source, rawlink in o:
- d.put("Origin: %s" % source)
- if rawlink != self.format_url(url):
- d.put(" (%s)" % rawlink)
- d.put("\n")
- d.text.yview("1.0")
-
- def setbad(self, url, msg):
- webchecker.Checker.setbad(self, url, msg)
- self.__bad.insert(url)
- self.newstatus()
-
- def setgood(self, url):
- webchecker.Checker.setgood(self, url)
- self.__bad.remove(url)
- self.newstatus()
-
- def newlink(self, url, origin):
- webchecker.Checker.newlink(self, url, origin)
- if url in self.done:
- self.__done.insert(url)
- elif url in self.todo:
- self.__todo.insert(url)
- self.newstatus()
-
- def markdone(self, url):
- webchecker.Checker.markdone(self, url)
- self.__done.insert(url)
- self.__todo.remove(url)
- self.newstatus()
-
- def seterror(self, url, triple):
- webchecker.Checker.seterror(self, url, triple)
- self.__errors.insert((url, ''))
- self.newstatus()
-
- def newstatus(self):
- self.__status.config(text="Status: "+self.status())
- self.__parent.update()
-
- def update_checkext(self):
- self.checkext = self.__cv.get()
-
-
-class ListPanel:
-
- def __init__(self, mp, name, checker, showinfo=None):
- self.mp = mp
- self.name = name
- self.showinfo = showinfo
- self.checker = checker
- self.panel = mp.addpanel(name)
- self.list, self.frame = tktools.make_list_box(
- self.panel, width=60, height=5)
- self.list.config(exportselection=0)
- if showinfo:
- self.list.bind('<Double-Button-1>', self.doubleclick)
- self.items = []
-
- def clear(self):
- self.items = []
- self.list.delete(0, END)
- self.mp.hidepanel(self.name)
-
- def doubleclick(self, event):
- l = self.selectedindices()
- if l:
- self.showinfo(self.items[l[0]])
-
- def selectedindices(self):
- l = self.list.curselection()
- if not l: return []
- return list(map(int, l))
-
- def insert(self, url):
- if url not in self.items:
- if not self.items:
- self.mp.showpanel(self.name)
- # (I tried sorting alphabetically, but the display is too jumpy)
- i = len(self.items)
- self.list.insert(i, self.checker.format_url(url))
- self.list.yview(i)
- self.items.insert(i, url)
-
- def remove(self, url):
- try:
- i = self.items.index(url)
- except (ValueError, IndexError):
- pass
- else:
- was_selected = i in self.selectedindices()
- self.list.delete(i)
- del self.items[i]
- if not self.items:
- self.mp.hidepanel(self.name)
- elif was_selected:
- if i >= len(self.items):
- i = len(self.items) - 1
- self.list.select_set(i)
-
-
-class LogPanel:
-
- def __init__(self, mp, name):
- self.mp = mp
- self.name = name
- self.panel = mp.addpanel(name)
- self.text, self.frame = tktools.make_text_box(self.panel, height=10)
- self.text.config(wrap=NONE)
-
- def clear(self):
- self.text.delete("1.0", END)
- self.text.yview("1.0")
-
- def put(self, s):
- self.text.insert(END, s)
- if '\n' in s:
- self.text.yview(END)
-
- def write(self, s):
- self.text.insert(END, s)
- if '\n' in s:
- self.text.yview(END)
- self.panel.update()
-
-
-class MultiPanel:
-
- def __init__(self, parent):
- self.parent = parent
- self.frame = Frame(self.parent)
- self.frame.pack(expand=1, fill=BOTH)
- self.topframe = Frame(self.frame, borderwidth=2, relief=RAISED)
- self.topframe.pack(fill=X)
- self.botframe = Frame(self.frame)
- self.botframe.pack(expand=1, fill=BOTH)
- self.panelnames = []
- self.panels = {}
-
- def addpanel(self, name, on=0):
- v = StringVar(self.parent)
- if on:
- v.set(name)
- else:
- v.set("")
- check = Checkbutton(self.topframe, text=name,
- offvalue="", onvalue=name, variable=v,
- command=self.checkpanel)
- check.pack(side=LEFT)
- panel = Frame(self.botframe)
- label = Label(panel, text=name, borderwidth=2, relief=RAISED, anchor=W)
- label.pack(side=TOP, fill=X)
- t = v, check, panel
- self.panelnames.append(name)
- self.panels[name] = t
- if on:
- panel.pack(expand=1, fill=BOTH)
- return panel
-
- def showpanel(self, name):
- v, check, panel = self.panels[name]
- v.set(name)
- panel.pack(expand=1, fill=BOTH)
-
- def hidepanel(self, name):
- v, check, panel = self.panels[name]
- v.set("")
- panel.pack_forget()
-
- def checkpanel(self):
- for name in self.panelnames:
- v, check, panel = self.panels[name]
- panel.pack_forget()
- for name in self.panelnames:
- v, check, panel = self.panels[name]
- if v.get():
- panel.pack(expand=1, fill=BOTH)
-
-
-if __name__ == '__main__':
- main()
diff --git a/Tools/webchecker/wcmac.py b/Tools/webchecker/wcmac.py
deleted file mode 100644
index 9edcd5d..0000000
--- a/Tools/webchecker/wcmac.py
+++ /dev/null
@@ -1,9 +0,0 @@
-import webchecker, sys
-webchecker.DEFROOT = "http://www.python.org/python/"
-webchecker.MAXPAGE = 50000
-webchecker.verbose = 2
-sys.argv.append('-x')
-webchecker.main()
-sys.stdout.write("\nCR to exit: ")
-sys.stdout.flush()
-sys.stdin.readline()
diff --git a/Tools/webchecker/webchecker.py b/Tools/webchecker/webchecker.py
deleted file mode 100755
index 7c3124c..0000000
--- a/Tools/webchecker/webchecker.py
+++ /dev/null
@@ -1,890 +0,0 @@
-#! /usr/bin/env python3
-
-# Original code by Guido van Rossum; extensive changes by Sam Bayer,
-# including code to check URL fragments.
-
-"""Web tree checker.
-
-This utility is handy to check a subweb of the world-wide web for
-errors. A subweb is specified by giving one or more ``root URLs''; a
-page belongs to the subweb if one of the root URLs is an initial
-prefix of it.
-
-File URL extension:
-
-In order to easy the checking of subwebs via the local file system,
-the interpretation of ``file:'' URLs is extended to mimic the behavior
-of your average HTTP daemon: if a directory pathname is given, the
-file index.html in that directory is returned if it exists, otherwise
-a directory listing is returned. Now, you can point webchecker to the
-document tree in the local file system of your HTTP daemon, and have
-most of it checked. In fact the default works this way if your local
-web tree is located at /usr/local/etc/httpd/htdpcs (the default for
-the NCSA HTTP daemon and probably others).
-
-Report printed:
-
-When done, it reports pages with bad links within the subweb. When
-interrupted, it reports for the pages that it has checked already.
-
-In verbose mode, additional messages are printed during the
-information gathering phase. By default, it prints a summary of its
-work status every 50 URLs (adjustable with the -r option), and it
-reports errors as they are encountered. Use the -q option to disable
-this output.
-
-Checkpoint feature:
-
-Whether interrupted or not, it dumps its state (a Python pickle) to a
-checkpoint file and the -R option allows it to restart from the
-checkpoint (assuming that the pages on the subweb that were already
-processed haven't changed). Even when it has run till completion, -R
-can still be useful -- it will print the reports again, and -Rq prints
-the errors only. In this case, the checkpoint file is not written
-again. The checkpoint file can be set with the -d option.
-
-The checkpoint file is written as a Python pickle. Remember that
-Python's pickle module is currently quite slow. Give it the time it
-needs to load and save the checkpoint file. When interrupted while
-writing the checkpoint file, the old checkpoint file is not
-overwritten, but all work done in the current run is lost.
-
-Miscellaneous:
-
-- You may find the (Tk-based) GUI version easier to use. See wcgui.py.
-
-- Webchecker honors the "robots.txt" convention. Thanks to Skip
-Montanaro for his robotparser.py module (included in this directory)!
-The agent name is hardwired to "webchecker". URLs that are disallowed
-by the robots.txt file are reported as external URLs.
-
-- Because the SGML parser is a bit slow, very large SGML files are
-skipped. The size limit can be set with the -m option.
-
-- When the server or protocol does not tell us a file's type, we guess
-it based on the URL's suffix. The mimetypes.py module (also in this
-directory) has a built-in table mapping most currently known suffixes,
-and in addition attempts to read the mime.types configuration files in
-the default locations of Netscape and the NCSA HTTP daemon.
-
-- We follow links indicated by <A>, <FRAME> and <IMG> tags. We also
-honor the <BASE> tag.
-
-- We now check internal NAME anchor links, as well as toplevel links.
-
-- Checking external links is now done by default; use -x to *disable*
-this feature. External links are now checked during normal
-processing. (XXX The status of a checked link could be categorized
-better. Later...)
-
-- If external links are not checked, you can use the -t flag to
-provide specific overrides to -x.
-
-Usage: webchecker.py [option] ... [rooturl] ...
-
-Options:
-
--R -- restart from checkpoint file
--d file -- checkpoint filename (default %(DUMPFILE)s)
--m bytes -- skip HTML pages larger than this size (default %(MAXPAGE)d)
--n -- reports only, no checking (use with -R)
--q -- quiet operation (also suppresses external links report)
--r number -- number of links processed per round (default %(ROUNDSIZE)d)
--t root -- specify root dir which should be treated as internal (can repeat)
--v -- verbose operation; repeating -v will increase verbosity
--x -- don't check external links (these are often slow to check)
--a -- don't check name anchors
-
-Arguments:
-
-rooturl -- URL to start checking
- (default %(DEFROOT)s)
-
-"""
-
-
-__version__ = "$Revision$"
-
-
-import sys
-import os
-from types import *
-import io
-import getopt
-import pickle
-
-import urllib.request
-import urllib.parse as urlparse
-import sgmllib
-import cgi
-
-import mimetypes
-from urllib import robotparser
-
-# Extract real version number if necessary
-if __version__[0] == '$':
- _v = __version__.split()
- if len(_v) == 3:
- __version__ = _v[1]
-
-
-# Tunable parameters
-DEFROOT = "file:/usr/local/etc/httpd/htdocs/" # Default root URL
-CHECKEXT = 1 # Check external references (1 deep)
-VERBOSE = 1 # Verbosity level (0-3)
-MAXPAGE = 150000 # Ignore files bigger than this
-ROUNDSIZE = 50 # Number of links processed per round
-DUMPFILE = "@webchecker.pickle" # Pickled checkpoint
-AGENTNAME = "webchecker" # Agent name for robots.txt parser
-NONAMES = 0 # Force name anchor checking
-
-
-# Global variables
-
-
-def main():
- checkext = CHECKEXT
- verbose = VERBOSE
- maxpage = MAXPAGE
- roundsize = ROUNDSIZE
- dumpfile = DUMPFILE
- restart = 0
- norun = 0
-
- try:
- opts, args = getopt.getopt(sys.argv[1:], 'Rd:m:nqr:t:vxa')
- except getopt.error as msg:
- sys.stdout = sys.stderr
- print(msg)
- print(__doc__%globals())
- sys.exit(2)
-
- # The extra_roots variable collects extra roots.
- extra_roots = []
- nonames = NONAMES
-
- for o, a in opts:
- if o == '-R':
- restart = 1
- if o == '-d':
- dumpfile = a
- if o == '-m':
- maxpage = int(a)
- if o == '-n':
- norun = 1
- if o == '-q':
- verbose = 0
- if o == '-r':
- roundsize = int(a)
- if o == '-t':
- extra_roots.append(a)
- if o == '-a':
- nonames = not nonames
- if o == '-v':
- verbose = verbose + 1
- if o == '-x':
- checkext = not checkext
-
- if verbose > 0:
- print(AGENTNAME, "version", __version__)
-
- if restart:
- c = load_pickle(dumpfile=dumpfile, verbose=verbose)
- else:
- c = Checker()
-
- c.setflags(checkext=checkext, verbose=verbose,
- maxpage=maxpage, roundsize=roundsize,
- nonames=nonames
- )
-
- if not restart and not args:
- args.append(DEFROOT)
-
- for arg in args:
- c.addroot(arg)
-
- # The -t flag is only needed if external links are not to be
- # checked. So -t values are ignored unless -x was specified.
- if not checkext:
- for root in extra_roots:
- # Make sure it's terminated by a slash,
- # so that addroot doesn't discard the last
- # directory component.
- if root[-1] != "/":
- root = root + "/"
- c.addroot(root, add_to_do = 0)
-
- try:
-
- if not norun:
- try:
- c.run()
- except KeyboardInterrupt:
- if verbose > 0:
- print("[run interrupted]")
-
- try:
- c.report()
- except KeyboardInterrupt:
- if verbose > 0:
- print("[report interrupted]")
-
- finally:
- if c.save_pickle(dumpfile):
- if dumpfile == DUMPFILE:
- print("Use ``%s -R'' to restart." % sys.argv[0])
- else:
- print("Use ``%s -R -d %s'' to restart." % (sys.argv[0],
- dumpfile))
-
-
-def load_pickle(dumpfile=DUMPFILE, verbose=VERBOSE):
- if verbose > 0:
- print("Loading checkpoint from %s ..." % dumpfile)
- f = open(dumpfile, "rb")
- c = pickle.load(f)
- f.close()
- if verbose > 0:
- print("Done.")
- print("Root:", "\n ".join(c.roots))
- return c
-
-
-class Checker:
-
- checkext = CHECKEXT
- verbose = VERBOSE
- maxpage = MAXPAGE
- roundsize = ROUNDSIZE
- nonames = NONAMES
-
- validflags = tuple(dir())
-
- def __init__(self):
- self.reset()
-
- def setflags(self, **kw):
- for key in kw:
- if key not in self.validflags:
- raise NameError("invalid keyword argument: %s" % str(key))
- for key, value in kw.items():
- setattr(self, key, value)
-
- def reset(self):
- self.roots = []
- self.todo = {}
- self.done = {}
- self.bad = {}
-
- # Add a name table, so that the name URLs can be checked. Also
- # serves as an implicit cache for which URLs are done.
- self.name_table = {}
-
- self.round = 0
- # The following are not pickled:
- self.robots = {}
- self.errors = {}
- self.urlopener = MyURLopener()
- self.changed = 0
-
- def note(self, level, format, *args):
- if self.verbose > level:
- if args:
- format = format%args
- self.message(format)
-
- def message(self, format, *args):
- if args:
- format = format%args
- print(format)
-
- def __getstate__(self):
- return (self.roots, self.todo, self.done, self.bad, self.round)
-
- def __setstate__(self, state):
- self.reset()
- (self.roots, self.todo, self.done, self.bad, self.round) = state
- for root in self.roots:
- self.addrobot(root)
- for url in self.bad:
- self.markerror(url)
-
- def addroot(self, root, add_to_do = 1):
- if root not in self.roots:
- troot = root
- scheme, netloc, path, params, query, fragment = \
- urlparse.urlparse(root)
- i = path.rfind("/") + 1
- if 0 < i < len(path):
- path = path[:i]
- troot = urlparse.urlunparse((scheme, netloc, path,
- params, query, fragment))
- self.roots.append(troot)
- self.addrobot(root)
- if add_to_do:
- self.newlink((root, ""), ("<root>", root))
-
- def addrobot(self, root):
- root = urlparse.urljoin(root, "/")
- if root in self.robots: return
- url = urlparse.urljoin(root, "/robots.txt")
- self.robots[root] = rp = robotparser.RobotFileParser()
- self.note(2, "Parsing %s", url)
- rp.debug = self.verbose > 3
- rp.set_url(url)
- try:
- rp.read()
- except (OSError, IOError) as msg:
- self.note(1, "I/O error parsing %s: %s", url, msg)
-
- def run(self):
- while self.todo:
- self.round = self.round + 1
- self.note(0, "\nRound %d (%s)\n", self.round, self.status())
- urls = sorted(self.todo.keys())
- del urls[self.roundsize:]
- for url in urls:
- self.dopage(url)
-
- def status(self):
- return "%d total, %d to do, %d done, %d bad" % (
- len(self.todo)+len(self.done),
- len(self.todo), len(self.done),
- len(self.bad))
-
- def report(self):
- self.message("")
- if not self.todo: s = "Final"
- else: s = "Interim"
- self.message("%s Report (%s)", s, self.status())
- self.report_errors()
-
- def report_errors(self):
- if not self.bad:
- self.message("\nNo errors")
- return
- self.message("\nError Report:")
- sources = sorted(self.errors.keys())
- for source in sources:
- triples = self.errors[source]
- self.message("")
- if len(triples) > 1:
- self.message("%d Errors in %s", len(triples), source)
- else:
- self.message("Error in %s", source)
- # Call self.format_url() instead of referring
- # to the URL directly, since the URLs in these
- # triples is now a (URL, fragment) pair. The value
- # of the "source" variable comes from the list of
- # origins, and is a URL, not a pair.
- for url, rawlink, msg in triples:
- if rawlink != self.format_url(url): s = " (%s)" % rawlink
- else: s = ""
- self.message(" HREF %s%s\n msg %s",
- self.format_url(url), s, msg)
-
- def dopage(self, url_pair):
-
- # All printing of URLs uses format_url(); argument changed to
- # url_pair for clarity.
- if self.verbose > 1:
- if self.verbose > 2:
- self.show("Check ", self.format_url(url_pair),
- " from", self.todo[url_pair])
- else:
- self.message("Check %s", self.format_url(url_pair))
- url, local_fragment = url_pair
- if local_fragment and self.nonames:
- self.markdone(url_pair)
- return
- try:
- page = self.getpage(url_pair)
- except sgmllib.SGMLParseError as msg:
- msg = self.sanitize(msg)
- self.note(0, "Error parsing %s: %s",
- self.format_url(url_pair), msg)
- # Dont actually mark the URL as bad - it exists, just
- # we can't parse it!
- page = None
- if page:
- # Store the page which corresponds to this URL.
- self.name_table[url] = page
- # If there is a fragment in this url_pair, and it's not
- # in the list of names for the page, call setbad(), since
- # it's a missing anchor.
- if local_fragment and local_fragment not in page.getnames():
- self.setbad(url_pair, ("Missing name anchor `%s'" % local_fragment))
- for info in page.getlinkinfos():
- # getlinkinfos() now returns the fragment as well,
- # and we store that fragment here in the "todo" dictionary.
- link, rawlink, fragment = info
- # However, we don't want the fragment as the origin, since
- # the origin is logically a page.
- origin = url, rawlink
- self.newlink((link, fragment), origin)
- else:
- # If no page has been created yet, we want to
- # record that fact.
- self.name_table[url_pair[0]] = None
- self.markdone(url_pair)
-
- def newlink(self, url, origin):
- if url in self.done:
- self.newdonelink(url, origin)
- else:
- self.newtodolink(url, origin)
-
- def newdonelink(self, url, origin):
- if origin not in self.done[url]:
- self.done[url].append(origin)
-
- # Call self.format_url(), since the URL here
- # is now a (URL, fragment) pair.
- self.note(3, " Done link %s", self.format_url(url))
-
- # Make sure that if it's bad, that the origin gets added.
- if url in self.bad:
- source, rawlink = origin
- triple = url, rawlink, self.bad[url]
- self.seterror(source, triple)
-
- def newtodolink(self, url, origin):
- # Call self.format_url(), since the URL here
- # is now a (URL, fragment) pair.
- if url in self.todo:
- if origin not in self.todo[url]:
- self.todo[url].append(origin)
- self.note(3, " Seen todo link %s", self.format_url(url))
- else:
- self.todo[url] = [origin]
- self.note(3, " New todo link %s", self.format_url(url))
-
- def format_url(self, url):
- link, fragment = url
- if fragment: return link + "#" + fragment
- else: return link
-
- def markdone(self, url):
- self.done[url] = self.todo[url]
- del self.todo[url]
- self.changed = 1
-
- def inroots(self, url):
- for root in self.roots:
- if url[:len(root)] == root:
- return self.isallowed(root, url)
- return 0
-
- def isallowed(self, root, url):
- root = urlparse.urljoin(root, "/")
- return self.robots[root].can_fetch(AGENTNAME, url)
-
- def getpage(self, url_pair):
- # Incoming argument name is a (URL, fragment) pair.
- # The page may have been cached in the name_table variable.
- url, fragment = url_pair
- if url in self.name_table:
- return self.name_table[url]
-
- scheme, path = urllib.request.splittype(url)
- if scheme in ('mailto', 'news', 'javascript', 'telnet'):
- self.note(1, " Not checking %s URL" % scheme)
- return None
- isint = self.inroots(url)
-
- # Ensure that openpage gets the URL pair to
- # print out its error message and record the error pair
- # correctly.
- if not isint:
- if not self.checkext:
- self.note(1, " Not checking ext link")
- return None
- f = self.openpage(url_pair)
- if f:
- self.safeclose(f)
- return None
- text, nurl = self.readhtml(url_pair)
-
- if nurl != url:
- self.note(1, " Redirected to %s", nurl)
- url = nurl
- if text:
- return Page(text, url, maxpage=self.maxpage, checker=self)
-
- # These next three functions take (URL, fragment) pairs as
- # arguments, so that openpage() receives the appropriate tuple to
- # record error messages.
- def readhtml(self, url_pair):
- url, fragment = url_pair
- text = None
- f, url = self.openhtml(url_pair)
- if f:
- text = f.read()
- f.close()
- return text, url
-
- def openhtml(self, url_pair):
- url, fragment = url_pair
- f = self.openpage(url_pair)
- if f:
- url = f.geturl()
- info = f.info()
- if not self.checkforhtml(info, url):
- self.safeclose(f)
- f = None
- return f, url
-
- def openpage(self, url_pair):
- url, fragment = url_pair
- try:
- return self.urlopener.open(url)
- except (OSError, IOError) as msg:
- msg = self.sanitize(msg)
- self.note(0, "Error %s", msg)
- if self.verbose > 0:
- self.show(" HREF ", url, " from", self.todo[url_pair])
- self.setbad(url_pair, msg)
- return None
-
- def checkforhtml(self, info, url):
- if 'content-type' in info:
- ctype = cgi.parse_header(info['content-type'])[0].lower()
- if ';' in ctype:
- # handle content-type: text/html; charset=iso8859-1 :
- ctype = ctype.split(';', 1)[0].strip()
- else:
- if url[-1:] == "/":
- return 1
- ctype, encoding = mimetypes.guess_type(url)
- if ctype == 'text/html':
- return 1
- else:
- self.note(1, " Not HTML, mime type %s", ctype)
- return 0
-
- def setgood(self, url):
- if url in self.bad:
- del self.bad[url]
- self.changed = 1
- self.note(0, "(Clear previously seen error)")
-
- def setbad(self, url, msg):
- if url in self.bad and self.bad[url] == msg:
- self.note(0, "(Seen this error before)")
- return
- self.bad[url] = msg
- self.changed = 1
- self.markerror(url)
-
- def markerror(self, url):
- try:
- origins = self.todo[url]
- except KeyError:
- origins = self.done[url]
- for source, rawlink in origins:
- triple = url, rawlink, self.bad[url]
- self.seterror(source, triple)
-
- def seterror(self, url, triple):
- try:
- # Because of the way the URLs are now processed, I need to
- # check to make sure the URL hasn't been entered in the
- # error list. The first element of the triple here is a
- # (URL, fragment) pair, but the URL key is not, since it's
- # from the list of origins.
- if triple not in self.errors[url]:
- self.errors[url].append(triple)
- except KeyError:
- self.errors[url] = [triple]
-
- # The following used to be toplevel functions; they have been
- # changed into methods so they can be overridden in subclasses.
-
- def show(self, p1, link, p2, origins):
- self.message("%s %s", p1, link)
- i = 0
- for source, rawlink in origins:
- i = i+1
- if i == 2:
- p2 = ' '*len(p2)
- if rawlink != link: s = " (%s)" % rawlink
- else: s = ""
- self.message("%s %s%s", p2, source, s)
-
- def sanitize(self, msg):
- if isinstance(IOError, ClassType) and isinstance(msg, IOError):
- # Do the other branch recursively
- msg.args = self.sanitize(msg.args)
- elif isinstance(msg, TupleType):
- if len(msg) >= 4 and msg[0] == 'http error' and \
- isinstance(msg[3], InstanceType):
- # Remove the Message instance -- it may contain
- # a file object which prevents pickling.
- msg = msg[:3] + msg[4:]
- return msg
-
- def safeclose(self, f):
- try:
- url = f.geturl()
- except AttributeError:
- pass
- else:
- if url[:4] == 'ftp:' or url[:7] == 'file://':
- # Apparently ftp connections don't like to be closed
- # prematurely...
- text = f.read()
- f.close()
-
- def save_pickle(self, dumpfile=DUMPFILE):
- if not self.changed:
- self.note(0, "\nNo need to save checkpoint")
- elif not dumpfile:
- self.note(0, "No dumpfile, won't save checkpoint")
- else:
- self.note(0, "\nSaving checkpoint to %s ...", dumpfile)
- newfile = dumpfile + ".new"
- f = open(newfile, "wb")
- pickle.dump(self, f)
- f.close()
- try:
- os.unlink(dumpfile)
- except os.error:
- pass
- os.rename(newfile, dumpfile)
- self.note(0, "Done.")
- return 1
-
-
-class Page:
-
- def __init__(self, text, url, verbose=VERBOSE, maxpage=MAXPAGE, checker=None):
- self.text = text
- self.url = url
- self.verbose = verbose
- self.maxpage = maxpage
- self.checker = checker
-
- # The parsing of the page is done in the __init__() routine in
- # order to initialize the list of names the file
- # contains. Stored the parser in an instance variable. Passed
- # the URL to MyHTMLParser().
- size = len(self.text)
- if size > self.maxpage:
- self.note(0, "Skip huge file %s (%.0f Kbytes)", self.url, (size*0.001))
- self.parser = None
- return
- self.checker.note(2, " Parsing %s (%d bytes)", self.url, size)
- self.parser = MyHTMLParser(url, verbose=self.verbose,
- checker=self.checker)
- self.parser.feed(self.text)
- self.parser.close()
-
- def note(self, level, msg, *args):
- if self.checker:
- self.checker.note(level, msg, *args)
- else:
- if self.verbose >= level:
- if args:
- msg = msg%args
- print(msg)
-
- # Method to retrieve names.
- def getnames(self):
- if self.parser:
- return self.parser.names
- else:
- return []
-
- def getlinkinfos(self):
- # File reading is done in __init__() routine. Store parser in
- # local variable to indicate success of parsing.
-
- # If no parser was stored, fail.
- if not self.parser: return []
-
- rawlinks = self.parser.getlinks()
- base = urlparse.urljoin(self.url, self.parser.getbase() or "")
- infos = []
- for rawlink in rawlinks:
- t = urlparse.urlparse(rawlink)
- # DON'T DISCARD THE FRAGMENT! Instead, include
- # it in the tuples which are returned. See Checker.dopage().
- fragment = t[-1]
- t = t[:-1] + ('',)
- rawlink = urlparse.urlunparse(t)
- link = urlparse.urljoin(base, rawlink)
- infos.append((link, rawlink, fragment))
-
- return infos
-
-
-class MyStringIO(io.StringIO):
-
- def __init__(self, url, info):
- self.__url = url
- self.__info = info
- super(MyStringIO, self).__init__(self)
-
- def info(self):
- return self.__info
-
- def geturl(self):
- return self.__url
-
-
-class MyURLopener(urllib.request.FancyURLopener):
-
- http_error_default = urllib.request.URLopener.http_error_default
-
- def __init__(*args):
- self = args[0]
- urllib.request.FancyURLopener.__init__(*args)
- self.addheaders = [
- ('User-agent', 'Python-webchecker/%s' % __version__),
- ]
-
- def http_error_401(self, url, fp, errcode, errmsg, headers):
- return None
-
- def open_file(self, url):
- path = urllib.url2pathname(urllib.unquote(url))
- if os.path.isdir(path):
- if path[-1] != os.sep:
- url = url + '/'
- indexpath = os.path.join(path, "index.html")
- if os.path.exists(indexpath):
- return self.open_file(url + "index.html")
- try:
- names = os.listdir(path)
- except os.error as msg:
- exc_type, exc_value, exc_tb = sys.exc_info()
- raise IOError(msg).with_traceback(exc_tb)
- names.sort()
- s = MyStringIO("file:"+url, {'content-type': 'text/html'})
- s.write('<BASE HREF="file:%s">\n' %
- urllib.quote(os.path.join(path, "")))
- for name in names:
- q = urllib.quote(name)
- s.write('<A HREF="%s">%s</A>\n' % (q, q))
- s.seek(0)
- return s
- return urllib.request.FancyURLopener.open_file(self, url)
-
-
-class MyHTMLParser(sgmllib.SGMLParser):
-
- def __init__(self, url, verbose=VERBOSE, checker=None):
- self.myverbose = verbose # now unused
- self.checker = checker
- self.base = None
- self.links = {}
- self.names = []
- self.url = url
- sgmllib.SGMLParser.__init__(self)
-
- def check_name_id(self, attributes):
- """ Check the name or id attributes on an element.
- """
- # We must rescue the NAME or id (name is deprecated in XHTML)
- # attributes from the anchor, in order to
- # cache the internal anchors which are made
- # available in the page.
- for name, value in attributes:
- if name == "name" or name == "id":
- if value in self.names:
- self.checker.message("WARNING: duplicate ID name %s in %s",
- value, self.url)
- else: self.names.append(value)
- break
-
- def unknown_starttag(self, tag, attributes):
- """ In XHTML, you can have id attributes on any element.
- """
- self.check_name_id(attributes)
-
- def start_a(self, attributes):
- self.link_attr(attributes, 'href')
- self.check_name_id(attributes)
-
- def end_a(self): pass
-
- def do_area(self, attributes):
- self.link_attr(attributes, 'href')
- self.check_name_id(attributes)
-
- def do_body(self, attributes):
- self.link_attr(attributes, 'background', 'bgsound')
- self.check_name_id(attributes)
-
- def do_img(self, attributes):
- self.link_attr(attributes, 'src', 'lowsrc')
- self.check_name_id(attributes)
-
- def do_frame(self, attributes):
- self.link_attr(attributes, 'src', 'longdesc')
- self.check_name_id(attributes)
-
- def do_iframe(self, attributes):
- self.link_attr(attributes, 'src', 'longdesc')
- self.check_name_id(attributes)
-
- def do_link(self, attributes):
- for name, value in attributes:
- if name == "rel":
- parts = value.lower().split()
- if ( parts == ["stylesheet"]
- or parts == ["alternate", "stylesheet"]):
- self.link_attr(attributes, "href")
- break
- self.check_name_id(attributes)
-
- def do_object(self, attributes):
- self.link_attr(attributes, 'data', 'usemap')
- self.check_name_id(attributes)
-
- def do_script(self, attributes):
- self.link_attr(attributes, 'src')
- self.check_name_id(attributes)
-
- def do_table(self, attributes):
- self.link_attr(attributes, 'background')
- self.check_name_id(attributes)
-
- def do_td(self, attributes):
- self.link_attr(attributes, 'background')
- self.check_name_id(attributes)
-
- def do_th(self, attributes):
- self.link_attr(attributes, 'background')
- self.check_name_id(attributes)
-
- def do_tr(self, attributes):
- self.link_attr(attributes, 'background')
- self.check_name_id(attributes)
-
- def link_attr(self, attributes, *args):
- for name, value in attributes:
- if name in args:
- if value: value = value.strip()
- if value: self.links[value] = None
-
- def do_base(self, attributes):
- for name, value in attributes:
- if name == 'href':
- if value: value = value.strip()
- if value:
- if self.checker:
- self.checker.note(1, " Base %s", value)
- self.base = value
- self.check_name_id(attributes)
-
- def getlinks(self):
- return list(self.links.keys())
-
- def getbase(self):
- return self.base
-
-
-if __name__ == '__main__':
- main()
diff --git a/Tools/webchecker/websucker.py b/Tools/webchecker/websucker.py
deleted file mode 100755
index 4657b52..0000000
--- a/Tools/webchecker/websucker.py
+++ /dev/null
@@ -1,123 +0,0 @@
-#! /usr/bin/env python3
-
-"""A variant on webchecker that creates a mirror copy of a remote site."""
-
-__version__ = "$Revision$"
-
-import os
-import sys
-import getopt
-import urllib.parse
-
-import webchecker
-
-# Extract real version number if necessary
-if __version__[0] == '$':
- _v = __version__.split()
- if len(_v) == 3:
- __version__ = _v[1]
-
-def main():
- verbose = webchecker.VERBOSE
- try:
- opts, args = getopt.getopt(sys.argv[1:], "qv")
- except getopt.error as msg:
- print(msg)
- print("usage:", sys.argv[0], "[-qv] ... [rooturl] ...")
- return 2
- for o, a in opts:
- if o == "-q":
- verbose = 0
- if o == "-v":
- verbose = verbose + 1
- c = Sucker()
- c.setflags(verbose=verbose)
- c.urlopener.addheaders = [
- ('User-agent', 'websucker/%s' % __version__),
- ]
- for arg in args:
- print("Adding root", arg)
- c.addroot(arg)
- print("Run...")
- c.run()
-
-class Sucker(webchecker.Checker):
-
- checkext = 0
- nonames = 1
-
- # SAM 11/13/99: in general, URLs are now URL pairs.
- # Since we've suppressed name anchor checking,
- # we can ignore the second dimension.
-
- def readhtml(self, url_pair):
- url = url_pair[0]
- text = None
- path = self.savefilename(url)
- try:
- f = open(path, "rb")
- except IOError:
- f = self.openpage(url_pair)
- if f:
- info = f.info()
- nurl = f.geturl()
- if nurl != url:
- url = nurl
- path = self.savefilename(url)
- text = f.read()
- f.close()
- self.savefile(text, path)
- if not self.checkforhtml(info, url):
- text = None
- else:
- if self.checkforhtml({}, url):
- text = f.read()
- f.close()
- return text, url
-
- def savefile(self, text, path):
- dir, base = os.path.split(path)
- makedirs(dir)
- try:
- f = open(path, "wb")
- f.write(text)
- f.close()
- self.message("saved %s", path)
- except IOError as msg:
- self.message("didn't save %s: %s", path, str(msg))
-
- def savefilename(self, url):
- type, rest = urllib.parse.splittype(url)
- host, path = urllib.parse.splithost(rest)
- path = path.lstrip("/")
- user, host = urllib.parse.splituser(host)
- host, port = urllib.parse.splitnport(host)
- host = host.lower()
- if not path or path[-1] == "/":
- path = path + "index.html"
- if os.sep != "/":
- path = os.sep.join(path.split("/"))
- path = os.path.join(host, path)
- return path
-
-def makedirs(dir):
- if not dir:
- return
- if os.path.exists(dir):
- if not os.path.isdir(dir):
- try:
- os.rename(dir, dir + ".bak")
- os.mkdir(dir)
- os.rename(dir + ".bak", os.path.join(dir, "index.html"))
- except os.error:
- pass
- return
- head, tail = os.path.split(dir)
- if not tail:
- print("Huh? Don't know how to make dir", dir)
- return
- makedirs(head)
- os.mkdir(dir, 0o777)
-
-if __name__ == '__main__':
- sys.exit(main() or 0)
diff --git a/Tools/webchecker/wsgui.py b/Tools/webchecker/wsgui.py
deleted file mode 100755
index 032523b..0000000
--- a/Tools/webchecker/wsgui.py
+++ /dev/null
@@ -1,240 +0,0 @@
-#! /usr/bin/env python3
-
-"""Tkinter-based GUI for websucker.
-
-Easy use: type or paste source URL and destination directory in
-their respective text boxes, click GO or hit return, and presto.
-"""
-
-from Tkinter import *
-import websucker
-import os
-import threading
-import queue
-import time
-
-VERBOSE = 2
-
-
-try:
- class Canceled(Exception):
- "Exception used to cancel run()."
-except (NameError, TypeError):
- Canceled = __name__ + ".Canceled"
-
-
-class SuckerThread(websucker.Sucker):
-
- stopit = 0
- savedir = None
- rootdir = None
-
- def __init__(self, msgq):
- self.msgq = msgq
- websucker.Sucker.__init__(self)
- self.setflags(verbose=VERBOSE)
- self.urlopener.addheaders = [
- ('User-agent', 'websucker/%s' % websucker.__version__),
- ]
-
- def message(self, format, *args):
- if args:
- format = format%args
- ##print format
- self.msgq.put(format)
-
- def run1(self, url):
- try:
- try:
- self.reset()
- self.addroot(url)
- self.run()
- except Canceled:
- self.message("[canceled]")
- else:
- self.message("[done]")
- finally:
- self.msgq.put(None)
-
- def savefile(self, text, path):
- if self.stopit:
- raise Canceled
- websucker.Sucker.savefile(self, text, path)
-
- def getpage(self, url):
- if self.stopit:
- raise Canceled
- return websucker.Sucker.getpage(self, url)
-
- def savefilename(self, url):
- path = websucker.Sucker.savefilename(self, url)
- if self.savedir:
- n = len(self.rootdir)
- if path[:n] == self.rootdir:
- path = path[n:]
- while path[:1] == os.sep:
- path = path[1:]
- path = os.path.join(self.savedir, path)
- return path
-
- def XXXaddrobot(self, *args):
- pass
-
- def XXXisallowed(self, *args):
- return 1
-
-
-class App:
-
- sucker = None
- msgq = None
-
- def __init__(self, top):
- self.top = top
- top.columnconfigure(99, weight=1)
- self.url_label = Label(top, text="URL:")
- self.url_label.grid(row=0, column=0, sticky='e')
- self.url_entry = Entry(top, width=60, exportselection=0)
- self.url_entry.grid(row=0, column=1, sticky='we',
- columnspan=99)
- self.url_entry.focus_set()
- self.url_entry.bind("<Key-Return>", self.go)
- self.dir_label = Label(top, text="Directory:")
- self.dir_label.grid(row=1, column=0, sticky='e')
- self.dir_entry = Entry(top)
- self.dir_entry.grid(row=1, column=1, sticky='we',
- columnspan=99)
- self.go_button = Button(top, text="Go", command=self.go)
- self.go_button.grid(row=2, column=1, sticky='w')
- self.cancel_button = Button(top, text="Cancel",
- command=self.cancel,
- state=DISABLED)
- self.cancel_button.grid(row=2, column=2, sticky='w')
- self.auto_button = Button(top, text="Paste+Go",
- command=self.auto)
- self.auto_button.grid(row=2, column=3, sticky='w')
- self.status_label = Label(top, text="[idle]")
- self.status_label.grid(row=2, column=4, sticky='w')
- self.top.update_idletasks()
- self.top.grid_propagate(0)
-
- def message(self, text, *args):
- if args:
- text = text % args
- self.status_label.config(text=text)
-
- def check_msgq(self):
- while not self.msgq.empty():
- msg = self.msgq.get()
- if msg is None:
- self.go_button.configure(state=NORMAL)
- self.auto_button.configure(state=NORMAL)
- self.cancel_button.configure(state=DISABLED)
- if self.sucker:
- self.sucker.stopit = 0
- self.top.bell()
- else:
- self.message(msg)
- self.top.after(100, self.check_msgq)
-
- def go(self, event=None):
- if not self.msgq:
- self.msgq = queue.Queue(0)
- self.check_msgq()
- if not self.sucker:
- self.sucker = SuckerThread(self.msgq)
- if self.sucker.stopit:
- return
- self.url_entry.selection_range(0, END)
- url = self.url_entry.get()
- url = url.strip()
- if not url:
- self.top.bell()
- self.message("[Error: No URL entered]")
- return
- self.rooturl = url
- dir = self.dir_entry.get().strip()
- if not dir:
- self.sucker.savedir = None
- else:
- self.sucker.savedir = dir
- self.sucker.rootdir = os.path.dirname(
- websucker.Sucker.savefilename(self.sucker, url))
- self.go_button.configure(state=DISABLED)
- self.auto_button.configure(state=DISABLED)
- self.cancel_button.configure(state=NORMAL)
- self.message( '[running...]')
- self.sucker.stopit = 0
- t = threading.Thread(target=self.sucker.run1, args=(url,))
- t.start()
-
- def cancel(self):
- if self.sucker:
- self.sucker.stopit = 1
- self.message("[canceling...]")
-
- def auto(self):
- tries = ['PRIMARY', 'CLIPBOARD']
- text = ""
- for t in tries:
- try:
- text = self.top.selection_get(selection=t)
- except TclError:
- continue
- text = text.strip()
- if text:
- break
- if not text:
- self.top.bell()
- self.message("[Error: clipboard is empty]")
- return
- self.url_entry.delete(0, END)
- self.url_entry.insert(0, text)
- self.go()
-
-
-class AppArray:
-
- def __init__(self, top=None):
- if not top:
- top = Tk()
- top.title("websucker GUI")
- top.iconname("wsgui")
- top.wm_protocol('WM_DELETE_WINDOW', self.exit)
- self.top = top
- self.appframe = Frame(self.top)
- self.appframe.pack(fill='both')
- self.applist = []
- self.exit_button = Button(top, text="Exit", command=self.exit)
- self.exit_button.pack(side=RIGHT)
- self.new_button = Button(top, text="New", command=self.addsucker)
- self.new_button.pack(side=LEFT)
- self.addsucker()
- ##self.applist[0].url_entry.insert(END, "http://www.python.org/doc/essays/")
-
- def addsucker(self):
- self.top.geometry("")
- frame = Frame(self.appframe, borderwidth=2, relief=GROOVE)
- frame.pack(fill='x')
- app = App(frame)
- self.applist.append(app)
-
- done = 0
-
- def mainloop(self):
- while not self.done:
- time.sleep(0.1)
- self.top.update()
-
- def exit(self):
- for app in self.applist:
- app.cancel()
- app.message("[exiting...]")
- self.done = 1
-
-
-def main():
- AppArray().mainloop()
-
-if __name__ == '__main__':
- main()
style='width: 0.8%;'/> -rw-r--r--src/H5Tpkg.h2
-rw-r--r--src/H5Tpublic.h8
-rw-r--r--src/H5VLcallback.c8
-rw-r--r--src/H5VLint.c4
-rw-r--r--src/H5VLnative.c6
-rw-r--r--src/H5VLnative_introspect.c2
-rw-r--r--src/H5VLpassthru.c2
-rw-r--r--src/H5VM.c2
-rw-r--r--src/H5WB.c2
-rw-r--r--src/H5Z.c2
-rw-r--r--src/H5Znbit.c2
-rw-r--r--src/H5Zscaleoffset.c4
-rw-r--r--src/H5Ztrans.c6
-rw-r--r--src/H5detect.c4
-rw-r--r--src/H5mpi.c2
-rw-r--r--src/H5private.h2
-rw-r--r--src/H5public.h2
-rw-r--r--test/SWMR_POSIX_Order_UG.txt4
-rw-r--r--test/SWMR_UseCase_UG.txt6
-rw-r--r--test/accum.c4
-rw-r--r--test/atomic_writer.c2
-rw-r--r--test/big.c10
-rw-r--r--test/btree2.c130
-rw-r--r--test/cache.c66
-rw-r--r--test/cache_api.c8
-rw-r--r--test/cache_common.c10
-rw-r--r--test/cache_common.h4
-rw-r--r--test/cache_image.c12
-rw-r--r--test/cache_tagging.c46
-rw-r--r--test/chunk_info.c4
-rw-r--r--test/cmpd_dset.c2
-rw-r--r--test/cork.c6
-rw-r--r--test/direct_chunk.c14
-rw-r--r--test/dsets.c40
-rw-r--r--test/dt_arith.c24
-rw-r--r--test/dtransform.c2
-rw-r--r--test/dtypes.c12
-rw-r--r--test/earray.c18
-rw-r--r--test/enc_dec_plist_cross_platform.c2
-rw-r--r--test/enum.c2
-rw-r--r--test/external.c6
-rw-r--r--test/farray.c12
-rw-r--r--test/fheap.c8
-rw-r--r--test/file_image.c8
-rw-r--r--test/fillval.c4
-rw-r--r--test/filter_fail.c4
-rw-r--r--test/filter_plugin1_dsets.c2
-rw-r--r--test/filter_plugin3_dsets.c2
-rw-r--r--test/flushrefresh.c12
-rw-r--r--test/gen_bounds.c12
-rw-r--r--test/gen_new_array.c4
-rw-r--r--test/gen_old_group.c2
-rw-r--r--test/gen_plist.c2
-rw-r--r--test/genall5.c2
-rw-r--r--test/getname.c2
-rw-r--r--test/gheap.c2
-rw-r--r--test/h5test.c2
-rw-r--r--test/links.c8
-rw-r--r--test/mf.c34
-rw-r--r--test/mirror_vfd.c2
-rw-r--r--test/mount.c4
-rw-r--r--test/mtime.c2
-rw-r--r--test/objcopy.c12
-rw-r--r--test/ohdr.c6
-rw-r--r--test/page_buffer.c24
-rw-r--r--test/ros3.c2
-rw-r--r--test/s3comms.c26
-rw-r--r--test/set_extent.c14
-rw-r--r--test/stab.c6
-rw-r--r--test/swmr.c10
-rw-r--r--test/swmr_generator.c2
-rw-r--r--test/swmr_start_write.c2
-rw-r--r--test/tattr.c12
-rw-r--r--test/tcoords.c2
-rw-r--r--test/testcheck_version.sh.in4
-rw-r--r--test/testlibinfo.sh.in2
-rw-r--r--test/testmeta.c2
-rw-r--r--test/testswmr.sh.in12
-rw-r--r--test/testvdsswmr.sh.in4
-rw-r--r--test/tfile.c50
-rw-r--r--test/tgenprop.c18
-rw-r--r--test/th5o.c2
-rw-r--r--test/th5s.c2
-rw-r--r--test/tid.c4
-rw-r--r--test/tmisc.c4
-rw-r--r--test/trefer.c2
-rw-r--r--test/tselect.c40
-rw-r--r--test/tsohm.c10
-rw-r--r--test/ttsafe_attr_vlen.c2
-rw-r--r--test/tvltypes.c2
-rw-r--r--test/twriteorder.c2
-rw-r--r--test/unlink.c6
-rw-r--r--test/use_common.c4
-rw-r--r--test/use_disable_mdc_flushes.c2
-rw-r--r--test/vds.c16
-rw-r--r--test/vds_env.c2
-rw-r--r--test/vfd.c20
-rw-r--r--testpar/t_2Gio.c50
-rw-r--r--testpar/t_bigio.c4
-rw-r--r--testpar/t_cache.c32
-rw-r--r--testpar/t_cache_image.c28
-rw-r--r--testpar/t_coll_chunk.c16
-rw-r--r--testpar/t_coll_md_read.c2
-rw-r--r--testpar/t_dset.c48
-rw-r--r--testpar/t_file.c2
-rw-r--r--testpar/t_mdset.c6
-rw-r--r--testpar/t_mpi.c4
-rw-r--r--testpar/t_ph5basic.c2
-rw-r--r--testpar/t_pread.c10
-rw-r--r--testpar/t_shapesame.c64
-rw-r--r--testpar/t_span_tree.c30
-rw-r--r--testpar/testphdf5.h4
-rw-r--r--tools/lib/h5diff.c2
-rw-r--r--tools/lib/h5diff.h2
-rw-r--r--tools/lib/h5tools.h10
-rw-r--r--tools/lib/h5tools_dump.c10
-rw-r--r--tools/lib/h5tools_ref.c6
-rw-r--r--tools/lib/h5tools_str.c8
-rw-r--r--tools/lib/h5tools_utils.c6
-rw-r--r--tools/lib/h5tools_utils.h2
-rw-r--r--tools/lib/h5trav.c6
-rw-r--r--tools/lib/h5trav.h2
-rw-r--r--tools/lib/io_timer.c2
-rw-r--r--tools/libtest/h5tools_test_utils.c16
-rw-r--r--tools/src/h5diff/h5diff_common.c4
-rw-r--r--tools/src/h5diff/h5diff_main.c2
-rw-r--r--tools/src/h5dump/h5dump.c2
-rw-r--r--tools/src/h5dump/h5dump_ddl.c8
-rw-r--r--tools/src/h5dump/h5dump_xml.c8
-rw-r--r--tools/src/h5import/h5import.c14
-rw-r--r--tools/src/h5jam/h5jam.c10
-rw-r--r--tools/src/h5ls/h5ls.c2
-rw-r--r--tools/src/h5perf/pio_engine.c4
-rw-r--r--tools/src/h5perf/pio_perf.c4
-rw-r--r--tools/src/h5perf/sio_engine.c2
-rw-r--r--tools/src/h5perf/sio_perf.c6
-rw-r--r--tools/src/h5repack/h5repack.h2
-rw-r--r--tools/src/h5repack/h5repack_copy.c46
-rw-r--r--tools/src/h5repack/h5repack_filters.c4
-rw-r--r--tools/src/h5repack/h5repack_main.c6
-rw-r--r--tools/src/h5repack/h5repack_refs.c2
-rw-r--r--tools/src/h5repack/h5repack_verify.c2
-rw-r--r--tools/src/h5stat/h5stat.c2
-rw-r--r--tools/src/misc/h5debug.c8
-rw-r--r--tools/test/h5copy/CMakeTests.cmake2
-rw-r--r--tools/test/h5copy/h5copygentest.c2
-rw-r--r--tools/test/h5copy/testh5copy.sh.in2
-rw-r--r--tools/test/h5diff/CMakeTests.cmake4
-rw-r--r--tools/test/h5diff/h5diff_plugin.sh.in6
-rw-r--r--tools/test/h5diff/h5diffgentest.c12
-rw-r--r--tools/test/h5diff/testfiles/h5diff_10.txt2
-rw-r--r--tools/test/h5diff/testfiles/h5diff_600.txt2
-rw-r--r--tools/test/h5diff/testfiles/h5diff_603.txt2
-rw-r--r--tools/test/h5diff/testfiles/h5diff_606.txt2
-rw-r--r--tools/test/h5diff/testfiles/h5diff_612.txt2
-rw-r--r--tools/test/h5diff/testfiles/h5diff_615.txt2
-rw-r--r--tools/test/h5diff/testfiles/h5diff_621.txt2
-rw-r--r--tools/test/h5diff/testfiles/h5diff_622.txt2
-rw-r--r--tools/test/h5diff/testfiles/h5diff_623.txt2
-rw-r--r--tools/test/h5diff/testfiles/h5diff_624.txt2
-rw-r--r--tools/test/h5diff/testh5diff.sh.in8
-rw-r--r--tools/test/h5dump/CMakeTests.cmake2
-rw-r--r--tools/test/h5dump/CMakeTestsPBITS.cmake2
-rw-r--r--tools/test/h5dump/h5dumpgentest.c30
-rw-r--r--tools/test/h5dump/testh5dump.sh.in2
-rw-r--r--tools/test/h5dump/testh5dumppbits.sh.in2
-rw-r--r--tools/test/h5format_convert/h5fc_gentest.c2
-rw-r--r--tools/test/h5format_convert/testh5fc.sh.in2
-rw-r--r--tools/test/h5jam/testh5jam.sh.in2
-rw-r--r--tools/test/h5ls/h5ls_plugin.sh.in2
-rw-r--r--tools/test/h5ls/testh5ls.sh.in4
-rw-r--r--tools/test/h5ls/testh5lsvds.sh.in4
-rw-r--r--tools/test/h5repack/CMakeTests.cmake2
-rw-r--r--tools/test/h5repack/dynlib_rpk.c2
-rw-r--r--tools/test/h5repack/dynlib_vrpk.c2
-rw-r--r--tools/test/h5repack/h5repack.sh.in6
-rw-r--r--tools/test/h5repack/h5repackgentest.c4
-rw-r--r--tools/test/h5repack/h5repacktst.c16
-rw-r--r--tools/test/h5repack/testfiles/README2
-rw-r--r--tools/test/h5stat/testh5stat.sh.in2
-rw-r--r--tools/test/misc/h5clear_gentest.c2
-rw-r--r--tools/test/misc/h5perf_gentest.c6
-rw-r--r--tools/test/perform/direct_write_perf.c2
-rw-r--r--tools/test/perform/zip_perf.c4
-rw-r--r--utils/mirror_vfd/mirror_server.c2
-rw-r--r--utils/mirror_vfd/mirror_server_stop.c4
-rw-r--r--utils/tools/h5dwalk/h5dwalk.12
-rw-r--r--utils/tools/h5dwalk/h5dwalk.c6
452 files changed, 1617 insertions, 1604 deletions
diff --git a/.github/workflows/codespell.yml b/.github/workflows/codespell.yml
new file mode 100644
index 0000000..163353e
--- /dev/null
+++ b/.github/workflows/codespell.yml
@@ -0,0 +1,15 @@
+# GitHub Action to automate the identification of common misspellings in text files
+# https://github.com/codespell-project/codespell
+# https://github.com/codespell-project/actions-codespell
+name: codespell
+on: [push, pull_request]
+jobs:
+ codespell:
+ name: Check for spelling errors
+ runs-on: ubuntu-latest
+ steps:
+ - uses: actions/checkout@v2
+ - uses: codespell-project/actions-codespell@master
+ with:
+ skip: ./bin/trace,./hl/tools/h5watch/h5watch.c,./tools/test/h5jam/tellub.c
+ ignore_words_list: isnt,inout,nd,parms,parm,ba,offsetP,ser,ois,had,fiter,fo,clude,refere,minnum,offsetp,creat,ans:,eiter,lastr,ans,isn't,ifset,sur,trun,dne,tthe,hda,filname,te,htmp,minnum
diff --git a/MANIFEST b/MANIFEST
index 1550a94..796e4b3 100644
--- a/MANIFEST
+++ b/MANIFEST
@@ -39,6 +39,7 @@
./.github/workflows/clang-format-check.yml _DO_NOT_DISTRIBUTE_
./.github/workflows/main.yml _DO_NOT_DISTRIBUTE_
./.github/workflows/pr-check.yml _DO_NOT_DISTRIBUTE_
+./.github/workflows/codespell.yml _DO_NOT_DISTRIBUTE_
./m4/aclocal_fc.m4
./m4/aclocal_fc.f90
diff --git a/bin/bbrelease b/bin/bbrelease
index 388b9e4..cdb8678 100755
--- a/bin/bbrelease
+++ b/bin/bbrelease
@@ -48,7 +48,7 @@ USAGE()
{
cat << EOF
Usage: $0 -d <dir> [--docver BRANCHNAME] [-h] [--nocheck] [--private] <methods> ...
- -d DIR The name of the directory where the releas(es) should be
+ -d DIR The name of the directory where the release(s) should be
placed.
--docver BRANCHNAME This is added for 1.8 and beyond to get the correct
version of documentation files from the hdf5docs
@@ -103,7 +103,7 @@ EOF
# Modifications
#
# Steps:
-# 1. untar the tarball in a temporay directory;
+# 1. untar the tarball in a temporary directory;
# Note: do this in a temporary directory to avoid changing
# the original source directory which maybe around.
# 2. convert all its text files to DOS (LF-CR) style;
@@ -246,7 +246,7 @@ if [ "X$methods" = "X" ]; then
methods="tar"
fi
-# Create the temporay work directory.
+# Create the temporary work directory.
if mkdir $tmpdir; then
echo "temporary work directory for release. "\
"Can be deleted after release completes." > $tmpdir/README
diff --git a/bin/buildhdf5 b/bin/buildhdf5
index 113a278..786c35d 100755
--- a/bin/buildhdf5
+++ b/bin/buildhdf5
@@ -178,7 +178,7 @@ LOCATE_SZLIB()
esac
;; # end of case ncsa
unknown)
- # Unknow domain. Give a shot at the some standard places.
+ # Unknown domain. Give a shot at the some standard places.
szlibpaths="/usr/local"
;;
esac # end of case $mydomain
diff --git a/bin/cmakehdf5 b/bin/cmakehdf5
index da1aef8..e59c772 100755
--- a/bin/cmakehdf5
+++ b/bin/cmakehdf5
@@ -199,7 +199,7 @@ DUMP_LOGFILE()
# Show a start time stamp
TIMESTAMP
-# Initialize njobs if $AMKE is defined
+# Initialize njobs if $MAKE is defined
if [ -n "$MAKE" ]; then
# assume all arguments are for --jobs
njobs=`echo $MAKE | cut -s -d' ' -f2-`
@@ -365,7 +365,7 @@ STEP "Test the library and tools..." "ctest . -C Release $njobs" $testlog
# 7. Create an install image with this command:
STEP "Create an install image..." "cpack -C Release CPackConfig.cmake" $packlog
-# The implementation of installation is imcomplete (only works for linux).
+# The implementation of installation is incomplete (only works for linux).
# Screen it out for now till it is completed.
if false; then
# 8. Install with this command:
diff --git a/bin/genparser b/bin/genparser
index 462c2c2..9ea3152 100755
--- a/bin/genparser
+++ b/bin/genparser
@@ -27,7 +27,7 @@
# IMPORTANT OS X NOTE
#
# If you are using OS X, you will probably not have flex or bison
-# installed. In addtion, even if you do have bison installed, the bison
+# installed. In addition, even if you do have bison installed, the bison
# version you have installed may also have a bug that makes it unable to
# process our input files.
#
diff --git a/bin/h5cc.in b/bin/h5cc.in
index 6024b4d..6835b79 100644
--- a/bin/h5cc.in
+++ b/bin/h5cc.in
@@ -40,7 +40,7 @@ HL="@HL@"
## $CLINKER $H5BLD_CPPFLAGS $CPPFLAGS $H5BLD_CFLAGS $CFLAGS $LDFLAGS ##
## $LIBS $clibpath $link_objs $link_args $shared_link ##
## ##
-## These settings can be overriden by setting HDF5_CFLAGS, ##
+## These settings can be overridden by setting HDF5_CFLAGS, ##
## HDF5_CPPFLAGS, HDF5_LDFLAGS, or HDF5_LIBS in the environment. ##
## ##
############################################################################
@@ -144,7 +144,7 @@ usage() {
echo " shared libraries]"
echo " "
echo " You can also add or change paths and flags to the compile line using"
- echo " the following environment varibles or by assigning them to their counterparts"
+ echo " the following environment variables or by assigning them to their counterparts"
echo " in the 'Things You Can Modify to Override...'" section of $prog_name
echo " "
echo " Variable Current value to be replaced"
@@ -318,7 +318,7 @@ fi
if test "x$do_link" = "xyes"; then
shared_link=""
-# conditionnaly link with the hl library
+# conditionally link with the hl library
if test "X$HL" = "Xhl"; then
libraries=" $libraries -lhdf5_hl -lhdf5 "
else
diff --git a/bin/newer b/bin/newer
index e60ec45..5f0fdca 100755
--- a/bin/newer
+++ b/bin/newer
@@ -19,7 +19,7 @@
# Created Date: 2005/07/06
# Modification:
# Albert Cheng 2005/8/30
-# Changed from two arguments to mulitple arguments.
+# Changed from two arguments to multiple arguments.
if test $# -lt 2; then
exit 1
diff --git a/bin/release b/bin/release
index 8774851..e40c3d3 100755
--- a/bin/release
+++ b/bin/release
@@ -39,7 +39,7 @@ USAGE()
{
cat << EOF
Usage: $0 -d <dir> [--docver BRANCHNAME] [-h] [--nocheck] [--private] <methods> ...
- -d DIR The name of the directory where the releas(es) should be
+ -d DIR The name of the directory where the release(es) should be
placed.
--docver BRANCHNAME This is added for 1.8 and beyond to get the correct
version of documentation files from the hdf5docs
@@ -60,7 +60,7 @@ for compressing the resulting tar archive (if none are given then
cmake-tgz -- create a tar file using the gzip default level with a build-unix.sh
command file and all other CMake files needed to build HDF5 source
using CMake on unix machines.
- cmake-zip -- convert all text files to DOS style and create a zip file inluding cmake
+ cmake-zip -- convert all text files to DOS style and create a zip file including cmake
scripts and .bat files to build HDF5 source using CMake on Windows.
hpc-cmake-tgz
-- create a tar file using the gzip default level with a build-unix.sh
@@ -512,7 +512,7 @@ if [ "X$methods" = "X" ]; then
methods="tar"
fi
-# Create the temporay work directory.
+# Create the temporary work directory.
if mkdir $tmpdir; then
echo "temporary work directory for release. "\
"Can be deleted after release completes." > $tmpdir/README
diff --git a/bin/runtest b/bin/runtest
index 50acb76..5e05abb 100755
--- a/bin/runtest
+++ b/bin/runtest
@@ -630,7 +630,7 @@ while [ $# -gt 0 ]; do
# setup the directory structure for snapshot test.
CMD=setup
;;
- -*) # Unknow option
+ -*) # Unknown option
PRINT "Unknown option ($1)"
USAGE
exit 1
@@ -830,7 +830,7 @@ if [ -z "$NOCVS" ]; then
else
# make sure the cvs update, if done by another host, has completed.
# First wait for the presence of $CVSLOG which signals some host
- # has started the cvs update. Then wait for the absense of $CVSLOG_LOCK
+ # has started the cvs update. Then wait for the absence of $CVSLOG_LOCK
# which signals the host has completed the cvs update.
WAITFOR $CVSLOG 90
if [ $WAIT_STATUS -ne 0 ]; then
@@ -911,7 +911,7 @@ if [ -n "$TESTHOST" -a $HOSTNAME != "$TESTHOST" ]; then
PRINT $h is not reachable "(`date`)"
;;
*)
- PRINT "CHECK_RSH for $h returned unknow result ($RSH)"
+ PRINT "CHECK_RSH for $h returned unknown result ($RSH)"
;;
esac >> $TMP_OUTPUT 2>&1
done
diff --git a/bin/snapshot b/bin/snapshot
index 20c73cd..7fcf3ab 100755
--- a/bin/snapshot
+++ b/bin/snapshot
@@ -156,7 +156,7 @@ Usage: $PROGNAME [all] [checkout] [ftp <URL> [diff] [test] [srcdir] [release] [h
"snapshot srcdir" is equivalent to "snapshot srcdir all"
"snapshot srcdir checkout" is equivalent to "snapshot checkout"
srcdirname <dir>:
- Use <dir> as the srcdir testing directory if srcdir is choosen.
+ Use <dir> as the srcdir testing directory if srcdir is chosen.
If <dir> starts with '-', it is append to the default name
E.g., "snapshot srcdir srcdirname -xx" uses hostname-xx
[Default is hostname]
@@ -422,7 +422,7 @@ while [ $# -gt 0 ] ; do
fi
cmd=""
EXEC_CMD_ARG="$@"
- # exit the parsing while loop since all arguments have been consummed.
+ # exit the parsing while loop since all arguments have been consumed.
break
;;
check-vfd)
@@ -455,7 +455,7 @@ while [ $# -gt 0 ] ; do
OP_CONFIGURE="$OP_CONFIGURE $1"
;;
*)
- echo "Unkown option $1"
+ echo "Unknown option $1"
errcode=1
cmd="help"
break
@@ -623,7 +623,7 @@ if [ "$cmd" = "all" -o -n "$cmdtest" -o -n "$cmddiff" ]; then
fi
fi
- # if diff is choosen, exit 0 if no significant differences are found.
+ # if diff is chosen, exit 0 if no significant differences are found.
# otherwise, exit 1. This includes cases of other failures.
if [ -n "$cmddiff" ]; then
if [ $snapshot = no ]; then
@@ -799,7 +799,7 @@ if [ "$cmd" = "all" -o -n "$cmdrel" ]; then
# Replace the previous version with the current version.
# Should check if the errcode of the release process but there
# are other failures after release was done (e.g. h5vers or git failures)
- # that should allow the replacement to occure.
+ # that should allow the replacement to occur.
rm -rf ${PREVIOUS}
mv ${CURRENT} ${PREVIOUS}
fi #Release snapshot
diff --git a/bin/trace b/bin/trace
index 7afc3e7..169719c 100755
--- a/bin/trace
+++ b/bin/trace
@@ -381,9 +381,9 @@ sub rewrite_func ($$$$$) {
# Wrap lines that will be longer than the limit
if ($len_if_added > $max_trace_macro_line_len) {
- # Wrap line, with indention
+ # Wrap line, with indentation
$trace .= ",\n ";
- $len = 13; # Set to 13, for indention
+ $len = 13; # Set to 13, for indentation
# Indent an extra space to account for extra digit in 'H5TRACE' macro
if (scalar(@arg_str) >= 10) {
diff --git a/bin/warnhist b/bin/warnhist
index 742a6b5..0150138 100755
--- a/bin/warnhist
+++ b/bin/warnhist
@@ -78,7 +78,7 @@ sub do_help {
print "\t-S <file string list>\tDisplay warnings for files which contain a string, <file string list>\n";
print "\t\t<file string list> is a comma separated list, with no spaces\n";
print "\t\tFor example: 'H5Fint' or 'H5Fint,H5Gnode'\n";
- print "\t-l\tDisplay line nunbers for file/warning\n";
+ print "\t-l\tDisplay line numbers for file/warning\n";
print "\t-u\tLeave 'unique' types in warnings, instead of genericizing them\n";
print "\t-i <name list>\tIgnore named files, <name list>\n";
print "\t\t<name list> is a comma separated list, with no spaces\n";
diff --git a/c++/src/H5Attribute.cpp b/c++/src/H5Attribute.cpp
index 954222c..520a4f6 100644
--- a/c++/src/H5Attribute.cpp
+++ b/c++/src/H5Attribute.cpp
@@ -562,7 +562,7 @@ Attribute::p_read_variable_len(const DataType &mem_type, H5std_string &strg) con
///\exception H5::IdComponentException when the attempt to close the HDF5
/// object fails
// Description:
-// The underlaying reference counting in the C library ensures
+// The underlying reference counting in the C library ensures
// that the current valid id of this object is properly closed.
// Then the object's id is reset to the new id.
// Programmer Binh-Minh Ribler - 2000
diff --git a/c++/src/H5DataSet.cpp b/c++/src/H5DataSet.cpp
index ce56e30..627d81b 100644
--- a/c++/src/H5DataSet.cpp
+++ b/c++/src/H5DataSet.cpp
@@ -769,7 +769,7 @@ DataSet::p_read_variable_len(const hid_t mem_type_id, const hid_t mem_space_id,
///\exception H5::IdComponentException when the attempt to close the HDF5
/// object fails
// Description:
-// The underlaying reference counting in the C library ensures
+// The underlying reference counting in the C library ensures
// that the current valid id of this object is properly closed.
// Then the object's id is reset to the new id.
//--------------------------------------------------------------------------
diff --git a/c++/src/H5DataSpace.cpp b/c++/src/H5DataSpace.cpp
index 342e9fa..e4aef83 100644
--- a/c++/src/H5DataSpace.cpp
+++ b/c++/src/H5DataSpace.cpp
@@ -629,7 +629,7 @@ DataSpace::getId() const
///\exception H5::IdComponentException when the attempt to close the HDF5
/// object fails
// Description:
-// The underlaying reference counting in the C library ensures
+// The underlying reference counting in the C library ensures
// that the current valid id of this object is properly closed.
// Then the object's id is reset to the new id.
// Programmer Binh-Minh Ribler - 2000
diff --git a/c++/src/H5DataType.cpp b/c++/src/H5DataType.cpp
index af58a90..ff8f6dc 100644
--- a/c++/src/H5DataType.cpp
+++ b/c++/src/H5DataType.cpp
@@ -927,7 +927,7 @@ DataType::p_opentype(const H5Location &loc, const char *dtype_name) const
///\exception H5::IdComponentException when the attempt to close the HDF5
/// object fails
// Description:
-// The underlaying reference counting in the C library ensures
+// The underlying reference counting in the C library ensures
// that the current valid id of this object is properly closed.
// Then the object's id is reset to the new id.
// Programmer Binh-Minh Ribler - 2000
@@ -983,7 +983,7 @@ DataType::close()
// - Replaced decRefCount with close() to let the C library
// handle the reference counting - BMR, Jun 1, 2006
// - Added the use of H5CPP_EXITED to terminate the HDF5 library
-// and elimiate previous memory leaks. See comments in the
+// and eliminate previous memory leaks. See comments in the
// header file "H5PredType.h" for details. - BMR, Mar 30, 2012
// - Major re-implementation of the global constants was done
// to avoid relying on the order of the creation and deletion
diff --git a/c++/src/H5DcreatProp.cpp b/c++/src/H5DcreatProp.cpp
index a9270fd..da19d8d 100644
--- a/c++/src/H5DcreatProp.cpp
+++ b/c++/src/H5DcreatProp.cpp
@@ -552,7 +552,7 @@ DSetCreatPropList::setShuffle() const
///\exception H5::PropListIException
///\par Description
/// The values of space allocation time can be one of the
-/// followings:
+/// following:
/// \li \c H5D_ALLOC_TIME_DEFAULT
/// \li \c H5D_ALLOC_TIME_EARLY
/// \li \c H5D_ALLOC_TIME_LATE
diff --git a/c++/src/H5File.cpp b/c++/src/H5File.cpp
index f92171b..92a8d59 100644
--- a/c++/src/H5File.cpp
+++ b/c++/src/H5File.cpp
@@ -665,7 +665,7 @@ H5File::getLocId() const
///\exception H5::IdComponentException when the attempt to close the HDF5
/// object fails
// Description
-// The underlaying reference counting in the C library ensures
+// The underlying reference counting in the C library ensures
// that the current valid id of this object is properly closed.
// Then the object's id is reset to the new id.
// December 2000
diff --git a/c++/src/H5Group.cpp b/c++/src/H5Group.cpp
index fe791da..b017f4f 100644
--- a/c++/src/H5Group.cpp
+++ b/c++/src/H5Group.cpp
@@ -208,7 +208,7 @@ Group::getId() const
///\exception H5::IdComponentException when the attempt to close the HDF5
/// object fails
// Description:
-// The underlaying reference counting in the C library ensures
+// The underlying reference counting in the C library ensures
// that the current valid id of this object is properly closed.
// Then the object's id is reset to the new id.
// Programmer Binh-Minh Ribler - 2000
diff --git a/c++/src/H5Location.cpp b/c++/src/H5Location.cpp
index eb8dd4c..13a89aa 100644
--- a/c++/src/H5Location.cpp
+++ b/c++/src/H5Location.cpp
@@ -1745,7 +1745,7 @@ H5Location::getObjinfo(const H5std_string &name, hbool_t follow_link, H5G_stat_t
// Function: H5Location::getObjinfo
///\brief This is an overloaded member function, provided for convenience.
/// It differs from the above functions in that it doesn't have
-/// the paramemter \a follow_link.
+/// the parameter \a follow_link.
// Nov, 2005
//--------------------------------------------------------------------------
void
diff --git a/c++/src/H5Location.h b/c++/src/H5Location.h
index 740d0ce..e990ec8 100644
--- a/c++/src/H5Location.h
+++ b/c++/src/H5Location.h
@@ -15,7 +15,7 @@
#ifndef H5Location_H
#define H5Location_H
-#include "H5Classes.h" // constains forward class declarations
+#include "H5Classes.h" // contains forward class declarations
namespace H5 {
diff --git a/c++/src/H5PredType.cpp b/c++/src/H5PredType.cpp
index c407c33..9458fb7 100644
--- a/c++/src/H5PredType.cpp
+++ b/c++/src/H5PredType.cpp
@@ -790,7 +790,7 @@ September 2015:
classes, such as PropList, PredType, DataSpace, etc... Previously,
these global constants were declared statically and the C++ library used
a constant, called PredType::AtExit, to detect when all the global
- contants are destroyed then close the C library (H5close). This method
+ constants are destroyed then close the C library (H5close). This method
relied on the order of the constants being created and destroyed and
that PredType constants be the last to be destroyed. In September
2015, it was recognized that the order in which the global constants were
diff --git a/c++/src/H5PropList.cpp b/c++/src/H5PropList.cpp
index 7ee8395..e7a83af 100644
--- a/c++/src/H5PropList.cpp
+++ b/c++/src/H5PropList.cpp
@@ -301,7 +301,7 @@ PropList::getId() const
///\exception H5::IdComponentException when the attempt to close the HDF5
/// object fails
// Description:
-// The underlaying reference counting in the C library ensures
+// The underlying reference counting in the C library ensures
// that the current valid id of this object is properly closed.
// Then the object's id is reset to the new id.
// Programmer Binh-Minh Ribler - 2000
diff --git a/c++/test/tobject.cpp b/c++/test/tobject.cpp
index 2b694da..f040007 100644
--- a/c++/test/tobject.cpp
+++ b/c++/test/tobject.cpp
@@ -740,7 +740,7 @@ test_object()
MESSAGE(5, ("Testing Object Functions\n"));
test_get_objname(); // Test get object name from groups/datasets
- test_existance(); // Test check for object existance
+ test_existance(); // Test check for object existence
test_get_objname_ontypes(); // Test get object name from types
test_get_objtype(); // Test get object type
test_open_object_header(); // Test object header functions (H5O)
diff --git a/c++/test/ttypes.cpp b/c++/test/ttypes.cpp
index 69c93d1..8afb4bb 100644
--- a/c++/test/ttypes.cpp
+++ b/c++/test/ttypes.cpp
@@ -28,7 +28,7 @@ using namespace H5;
#include "h5cpputil.h" // C++ utilility header file
/*
- * Offset from alinged memory returned by malloc(). This can be used to test
+ * Offset from aligned memory returned by malloc(). This can be used to test
* that type conversions handle non-aligned buffers correctly.
*/
#define ALIGNMENT 1
diff --git a/config/BlankForm b/config/BlankForm
index ce4bf53..14131b2 100644
--- a/config/BlankForm
+++ b/config/BlankForm
@@ -61,7 +61,7 @@ fi
#
# PROD_CFLAGS Flags to pass to the compiler to create a
# PROD_CPPFLAGS production version of the library. These
-# usualy exclude symbolic debugging switches
+# usually exclude symbolic debugging switches
# (like `-g') and include optimization switches
# (like `-O').
#
@@ -84,7 +84,7 @@ fi
# compiler or a descendent of gcc such as EGCS or PGCC.
#
# AM_CFLAGS Flags added directly into this variable will
-# be propogated to the compiler wrapper scripts (h5cc,
+# be propagated to the compiler wrapper scripts (h5cc,
# h5c++, et cetera) in addition to being used to compile
# the library.
#
diff --git a/config/cmake/ConfigureChecks.cmake b/config/cmake/ConfigureChecks.cmake
index 483e551..43aa243 100644
--- a/config/cmake/ConfigureChecks.cmake
+++ b/config/cmake/ConfigureChecks.cmake
@@ -30,7 +30,7 @@ MARK_AS_ADVANCED (HDF5_STRICT_FORMAT_CHECKS)
# ----------------------------------------------------------------------
# Decide whether the data accuracy has higher priority during data
-# conversions. If not, some hard conversions will still be prefered even
+# conversions. If not, some hard conversions will still be preferred even
# though the data may be wrong (for example, some compilers don't
# support denormalized floating values) to maximize speed.
#-----------------------------------------------------------------------------
diff --git a/config/cmake/H5pubconf.h.in b/config/cmake/H5pubconf.h.in
index a6b238c..1f7f4b1 100644
--- a/config/cmake/H5pubconf.h.in
+++ b/config/cmake/H5pubconf.h.in
@@ -663,7 +663,7 @@
/* Version number of package */
#define H5_VERSION "@HDF5_PACKAGE_VERSION_STRING@"
-/* Data accuracy is prefered to speed during data conversions */
+/* Data accuracy is preferred to speed during data conversions */
#cmakedefine H5_WANT_DATA_ACCURACY @H5_WANT_DATA_ACCURACY@
/* Check exception handling functions during data conversions */
diff --git a/config/cmake/HDF5UseFortran.cmake b/config/cmake/HDF5UseFortran.cmake
index 0d05e20..43e006c 100644
--- a/config/cmake/HDF5UseFortran.cmake
+++ b/config/cmake/HDF5UseFortran.cmake
@@ -134,8 +134,8 @@ FORTRAN_RUN ("REAL and INTEGER KINDs"
PROG_OUTPUT
)
# dnl The output from the above program will be:
-# dnl -- LINE 1 -- valid integer kinds (comma seperated list)
-# dnl -- LINE 2 -- valid real kinds (comma seperated list)
+# dnl -- LINE 1 -- valid integer kinds (comma separated list)
+# dnl -- LINE 2 -- valid real kinds (comma separated list)
# dnl -- LINE 3 -- max decimal precision for reals
# dnl -- LINE 4 -- number of valid integer kinds
# dnl -- LINE 5 -- number of valid real kinds
diff --git a/config/cmake/HDFCXXCompilerFlags.cmake b/config/cmake/HDFCXXCompilerFlags.cmake
index a5af43c..e20ed6b 100644
--- a/config/cmake/HDFCXXCompilerFlags.cmake
+++ b/config/cmake/HDFCXXCompilerFlags.cmake
@@ -240,7 +240,7 @@ if (CMAKE_CXX_COMPILER_ID STREQUAL "GNU")
# Append more extra warning flags that only gcc 9.3+ knows about
if (NOT CMAKE_CXX_COMPILER_VERSION VERSION_LESS 9.3)
- # do not use C warnings, gnu-warnings 9.3, no cxx warniings
+ # do not use C warnings, gnu-warnings 9.3, no cxx warnings
# ADD_H5_FLAGS (H5_CXXFLAGS "${HDF5_SOURCE_DIR}/config/gnu-warnings/9.3")
endif ()
diff --git a/config/cmake_ext_mod/ConfigureChecks.cmake b/config/cmake_ext_mod/ConfigureChecks.cmake
index 1dd1440..6a9c274 100644
--- a/config/cmake_ext_mod/ConfigureChecks.cmake
+++ b/config/cmake_ext_mod/ConfigureChecks.cmake
@@ -129,7 +129,7 @@ CHECK_INCLUDE_FILE_CONCAT ("netinet/in.h" ${HDF_PREFIX}_HAVE_NETINET_IN_H)
CHECK_INCLUDE_FILE_CONCAT ("netdb.h" ${HDF_PREFIX}_HAVE_NETDB_H)
CHECK_INCLUDE_FILE_CONCAT ("arpa/inet.h" ${HDF_PREFIX}_HAVE_ARPA_INET_H)
-## Check for non-standard extenstion quadmath.h
+## Check for non-standard extension quadmath.h
CHECK_INCLUDE_FILES(quadmath.h C_HAVE_QUADMATH)
if (${C_HAVE_QUADMATH})
@@ -547,7 +547,7 @@ if (WINDOWS)
message (VERBOSE "Performing Test InitOnceExecuteOnce - Success")
endif ()
file (APPEND ${CMAKE_BINARY_DIR}${CMAKE_FILES_DIRECTORY}/CMakeOutput.log
- "Performing C SOURCE FILE Test InitOnceExecuteOnce succeded with the following output:\n"
+ "Performing C SOURCE FILE Test InitOnceExecuteOnce succeeded with the following output:\n"
"${OUTPUT}\n"
"Return value: ${HAVE_IOEO}\n")
else ()
diff --git a/config/cmake_ext_mod/HDFUseFortran.cmake b/config/cmake_ext_mod/HDFUseFortran.cmake
index 9cbcd76..c96298f 100644
--- a/config/cmake_ext_mod/HDFUseFortran.cmake
+++ b/config/cmake_ext_mod/HDFUseFortran.cmake
@@ -22,7 +22,7 @@ if (NOT CMAKE_VERSION VERSION_LESS "3.14.0")
endif ()
#-------------------------------------------------------------------------------
-# Fix Fortran flags if we are compiling staticly on Windows using
+# Fix Fortran flags if we are compiling statically on Windows using
# Windows_MT.cmake from config/cmake/UserMacros
#-------------------------------------------------------------------------------
if (BUILD_STATIC_CRT_LIBS)
diff --git a/config/cmake_ext_mod/NSIS.template.in b/config/cmake_ext_mod/NSIS.template.in
index ecced05..db79466 100644
--- a/config/cmake_ext_mod/NSIS.template.in
+++ b/config/cmake_ext_mod/NSIS.template.in
@@ -849,7 +849,7 @@ Section "Uninstall"
@CPACK_NSIS_DELETE_ICONS@
@CPACK_NSIS_DELETE_ICONS_EXTRA@
- ;Delete empty start menu parent diretories
+ ;Delete empty start menu parent directories
StrCpy $MUI_TEMP "$SMPROGRAMS\$MUI_TEMP"
startMenuDeleteLoop:
@@ -868,7 +868,7 @@ Section "Uninstall"
Delete "$SMPROGRAMS\$MUI_TEMP\Uninstall.lnk"
@CPACK_NSIS_DELETE_ICONS_EXTRA@
- ;Delete empty start menu parent diretories
+ ;Delete empty start menu parent directories
StrCpy $MUI_TEMP "$SMPROGRAMS\$MUI_TEMP"
secondStartMenuDeleteLoop:
diff --git a/config/gnu-cxxflags b/config/gnu-cxxflags
index 56e89bc..b544813 100644
--- a/config/gnu-cxxflags
+++ b/config/gnu-cxxflags
@@ -58,7 +58,7 @@ load_gnu_arguments()
if test "X-" = "X-$cxx_flags_set"; then
# PathScale compiler spits out gcc version string too. Need to
# filter it out.
- # icc beginning with version 12 includes a "gcc version compatiblilty"
+ # icc beginning with version 12 includes a "gcc version compatibility"
# string, causing the gcc H5_CXXFLAGS to be erroneously added. The line
# "grep -v 'icc version'" causes the discarding of any output
# containing 'icc version'. The cxx_version for icc is correctly determined
diff --git a/config/gnu-flags b/config/gnu-flags
index 89239df..fe6626d 100644
--- a/config/gnu-flags
+++ b/config/gnu-flags
@@ -58,7 +58,7 @@ load_gnu_arguments()
if test "X-" = "X-$cc_flags_set"; then
# PathScale compiler spits out gcc version string too. Need to
# filter it out.
- # icc beginning with version 12 includes a "gcc version compatiblilty"
+ # icc beginning with version 12 includes a "gcc version compatibility"
# string, causing the gcc H5_CFLAGS to be erroneously added. The line
# "grep -v 'icc version'" causes the discarding of any output
# containing 'icc version'. The cc_version for icc is correctly determined
diff --git a/config/solaris b/config/solaris
index b7c7963..6b6b74d 100644
--- a/config/solaris
+++ b/config/solaris
@@ -39,7 +39,7 @@ if test "X-" = "X-$cc_flags_set"; then
# Debug
# NDEBUG is handled explicitly by the configure script
- # -g is hanled by the symbols flags
+ # -g is handled by the symbols flags
DEBUG_CFLAGS=
# Symbols
@@ -101,7 +101,7 @@ if test "X-" = "X-$f9x_flags_set"; then
# Debug
# NDEBUG is handled explicitly by the configure script
- # -g is hanled by the symbols flags
+ # -g is handled by the symbols flags
DEBUG_FCFLAGS=
# Symbols
@@ -157,7 +157,7 @@ if test -z "$cxx_flags_set"; then
# Debug
# NDEBUG is handled explicitly by the configure script
- # -g is hanled by the symbols flags
+ # -g is handled by the symbols flags
DEBUG_CXXFLAGS=
# Symbols
diff --git a/configure.ac b/configure.ac
index 44ffd47..31578a6 100644
--- a/configure.ac
+++ b/configure.ac
@@ -390,7 +390,7 @@ if test "X$CC_BASENAME" = "Xclang"; then
# 'address;undefined'. Which and which combinations of these are
# supported varies by compiler version, but unsupported options
# or combinations will result in configure errors reported in config.log.
- # Comma separated lists of sanitize options wil be entered intact in
+ # Comma separated lists of sanitize options will be entered intact in
# one -fsanitize=<list> flag. Space separated lists will be entered in
# separate -fsanitize=<item> flags.
# NOTE: No sanity checking done here!
@@ -599,7 +599,7 @@ if test "X$HDF_FORTRAN" = "Xyes"; then
AC_SUBST([OBJECT_NAMELEN_DEFAULT_F])
## --------------------------------------------------------------------
- ## Fortran source extention
+ ## Fortran source extension
##
AC_FC_SRCEXT([f90])
@@ -3479,7 +3479,7 @@ fi
## ----------------------------------------------------------------------
## Decide whether the data accuracy has higher priority during data
-## conversions. If not, some hard conversions will still be prefered even
+## conversions. If not, some hard conversions will still be preferred even
## though the data may be wrong (for example, some compilers don't
## support denormalized floating values) to maximize speed.
##
@@ -3493,7 +3493,7 @@ AC_ARG_ENABLE([dconv-accuracy],
if test "$DATA_ACCURACY" = "yes"; then
AC_MSG_RESULT([yes])
AC_DEFINE([WANT_DATA_ACCURACY], [1],
- [Data accuracy is prefered to speed during data conversions])
+ [Data accuracy is preferred to speed during data conversions])
else
AC_MSG_RESULT([no])
fi
diff --git a/doxygen/dox/H5AC_cache_config_t.dox b/doxygen/dox/H5AC_cache_config_t.dox
index 9b9862b..3faecd5 100644
--- a/doxygen/dox/H5AC_cache_config_t.dox
+++ b/doxygen/dox/H5AC_cache_config_t.dox
@@ -26,7 +26,7 @@
*
* \Emph{*** DEPRECATED ***} Use \Code{H5Fstart/stop} logging functions instead
*
- * The trace file is a debuging feature that allow the capture of
+ * The trace file is a debugging feature that allow the capture of
* top level metadata cache requests for purposes of debugging and/or
* optimization. This field should normally be set to \c FALSE, as
* trace file collection imposes considerable overhead.
@@ -82,7 +82,7 @@
* H5C_incr__off ) && ( decr_mode == H5C_decr__off )}). There
* is no logical reason why this should be so, but it simplifies
* implementation and testing, and I can't think of any reason
- * why it would be desireable. If you can think of one, I'll
+ * why it would be desirable. If you can think of one, I'll
* revisit the issue. (JM)
* \endparblock
*
@@ -383,7 +383,7 @@
* to disk.\n
* When the sync point is reached (or when there is a user generated
* flush), process zero flushes sufficient entries to bring it into
- * complience with its min clean size (or flushes all dirty entries in
+ * compliance with its min clean size (or flushes all dirty entries in
* the case of a user generated flush), broad casts the list of
* entries just cleaned to all the other processes, and then exits
* the sync point.\n
diff --git a/doxygen/dox/Overview.dox b/doxygen/dox/Overview.dox
index 754722e..3b190d6 100644
--- a/doxygen/dox/Overview.dox
+++ b/doxygen/dox/Overview.dox
@@ -4,7 +4,7 @@
This is the documentation set for HDF5. You can
<a href="hdf5-doc.tgz">download</a> it as a tgz archive for offline reading.
-This is the documention set for HDF5 in terms of specifications and software
+This is the documentation set for HDF5 in terms of specifications and software
developed and maintained by <a href="https://www.hdfgroup.org/">The HDF
Group</a>. It is impractical to document the entire HDF5 ecosystem in one place,
and you should also consult the documentation sets of the many outstanding
diff --git a/doxygen/examples/DebuggingHDF5Applications.html b/doxygen/examples/DebuggingHDF5Applications.html
index c6aaf74..3390887 100644
--- a/doxygen/examples/DebuggingHDF5Applications.html
+++ b/doxygen/examples/DebuggingHDF5Applications.html
@@ -173,7 +173,7 @@ IOT Trap, core dumped.
<tr>
<td align="center">mm</td>
<td align="center">Yes</td>
- <td>Library memory managment</td>
+ <td>Library memory management</td>
</tr>
<tr>
<td align="center">o</td>
diff --git a/doxygen/examples/Filters.html b/doxygen/examples/Filters.html
index 2d5bc5e..7054a3b 100644
--- a/doxygen/examples/Filters.html
+++ b/doxygen/examples/Filters.html
@@ -54,7 +54,7 @@
<p>Two types of filters can be applied to raw data I/O: permanent
filters and transient filters. The permanent filter pipeline is
- defned when the dataset is created while the transient pipeline
+ defined when the dataset is created while the transient pipeline
is defined for each I/O operation. During an
<code>H5Dwrite()</code> the transient filters are applied first
in the order defined and then the permanent filters are applied
@@ -211,7 +211,7 @@
should be turned on so such cases can be handled gracefully by
storing the original data instead of the compressed data. The
<em>cd_nvalues</em> should be one with <em>cd_value[0]</em>
- being a compression agression level between zero and nine,
+ being a compression aggression level between zero and nine,
inclusive (zero is the fastest compression while nine results in
the best compression ratio).
@@ -252,7 +252,7 @@
</p><p>The function that acts as the filter always returns zero
(failure) if the <code>md5()</code> function was not detected at
- configuration time (left as an excercise for the reader).
+ configuration time (left as an exercise for the reader).
Otherwise the function is broken down to an input and output
half. The output half calculates a checksum, increases the size
of the output buffer if necessary, and appends the checksum to
@@ -374,7 +374,7 @@
<tr valign="top">
<td>Method</td>
<td>This is the name of the method as defined with
- <code>H5Zregister()</code> with the charaters
+ <code>H5Zregister()</code> with the characters
"&lt; or "&gt;" prepended to indicate
input or output.</td>
</tr>
diff --git a/doxygen/examples/H5.format.1.0.html b/doxygen/examples/H5.format.1.0.html
index 2d3ffbe..cdc19ec7 100644
--- a/doxygen/examples/H5.format.1.0.html
+++ b/doxygen/examples/H5.format.1.0.html
@@ -495,7 +495,7 @@ Elena> "Free-space object"
<td>End of File Address</td>
<td>This is the relative file address of the first byte past
the end of all HDF5 data. It is used to determine whether a
- file has been accidently truncated and as an address where
+ file has been accidentally truncated and as an address where
file data allocation can occur if the free list is not
used.</td>
</tr>
@@ -2414,7 +2414,7 @@ Elena> "Free-space object"
<td><b>Normalization.</b> The value can be 0 if there is no
normalization, 1 if the most significant bit of the
mantissa is always set (except for 0.0), and 2 if the most
- signficant bit of the mantissa is not stored but is
+ significant bit of the mantissa is not stored but is
implied to be set. The value 3 is reserved and will not
appear in this field.</td>
</tr>
@@ -2916,7 +2916,7 @@ Elena> "Free-space object"
<p>The fill value message stores a single data point value which
is returned to the application when an uninitialized data point
- is read from the dataset. The fill value is interpretted with
+ is read from the dataset. The fill value is interpreted with
the same datatype as the dataset. If no fill value message is
present then a fill value of all zero is assumed.
@@ -3327,7 +3327,7 @@ Elena> "Free-space object"
<p>
<center>
- <table border align=center cellpadding=4 witdh="80%">
+ <table border align=center cellpadding=4 width="80%">
<caption align=top>
<b>Filter Pipeline Message</b>
</caption>
@@ -3386,7 +3386,7 @@ Elena> "Free-space object"
<p>
<center>
- <table border align=center cellpadding=4 witdh="80%">
+ <table border align=center cellpadding=4 width="80%">
<caption align=top>
<b>Filter Pipeline Message</b>
</caption>
diff --git a/doxygen/examples/H5.format.1.1.html b/doxygen/examples/H5.format.1.1.html
index ebbbe8e..9894fad 100644
--- a/doxygen/examples/H5.format.1.1.html
+++ b/doxygen/examples/H5.format.1.1.html
@@ -616,7 +616,7 @@ TABLE.list TD { border:none; }
<td>
<P>This is the absolute file address of the first byte past
the end of all HDF5 data. It is used to determine whether a
- file has been accidently truncated and as an address where
+ file has been accidentally truncated and as an address where
file data allocation can occur if space from the free list is
not used.
</P>
@@ -3184,7 +3184,7 @@ TABLE.list TD { border:none; }
<td><b>Normalization.</b> The value can be 0 if there is no
normalization, 1 if the most significant bit of the
mantissa is always set (except for 0.0), and 2 if the most
- signficant bit of the mantissa is not stored but is
+ significant bit of the mantissa is not stored but is
implied to be set. The value 3 is reserved and will not
appear in this field.</td>
</tr>
diff --git a/doxygen/examples/H5.format.2.0.html b/doxygen/examples/H5.format.2.0.html
index 3653489..4a5fe37 100644
--- a/doxygen/examples/H5.format.2.0.html
+++ b/doxygen/examples/H5.format.2.0.html
@@ -821,7 +821,7 @@ II.A. Disk Format: Level 0A - Format Signature and Superblock</a></h3>
<td>
<p>This is the absolute file address of the first byte past
the end of all HDF5 data. It is used to determine whether a
- file has been accidently truncated and as an address where
+ file has been accidentally truncated and as an address where
file data allocation can occur if space from the free list is
not used.
</p>
@@ -4910,7 +4910,7 @@ III.F. Disk Format: Level 1F - Fractal Heap</a></h3>
enough to store objects greater than 16 bytes or not. If the
heap ID length is 18 bytes or smaller, the &lsquo;normal&rsquo; tiny heap ID
form is used. If the heap ID length is greater than 18 bytes in
- length, the &ldquo;extented&rdquo; form is used. See format description below
+ length, the &ldquo;extended&rdquo; form is used. See format description below
for both sub-types.
</p>
</td>
@@ -7884,7 +7884,7 @@ IV.A.2. Disk Format: Level 2A2 - Data Object Header Messages</a></h3>
<tr>
<td align="center"><code>3</code></td>
- <td>Message stored is not shared, but is sharable.
+ <td>Message stored is not shared, but is shareable.
</td>
</tr>
diff --git a/doxygen/examples/H5.format.html b/doxygen/examples/H5.format.html
index e16805f..cbcb387 100644
--- a/doxygen/examples/H5.format.html
+++ b/doxygen/examples/H5.format.html
@@ -936,7 +936,7 @@
<td>
<p>This is the absolute file address of the first byte past
the end of all HDF5 data. It is used to determine whether a
- file has been accidently truncated and as an address where
+ file has been accidentally truncated and as an address where
file data allocation can occur if space from the free list is
not used.
</p>
@@ -8691,7 +8691,7 @@ three rows are needed.
<tr>
<td align="center"><code>3</code></td>
- <td>Message stored is not shared, but is sharable.
+ <td>Message stored is not shared, but is shareable.
</td>
</tr>
@@ -20202,7 +20202,7 @@ disk address for the chunk.</p>
</tr>
<tr>
- <td><p>Length fo External File Name</p></td>
+ <td><p>Length of External File Name</p></td>
<td><p>This is the length for the external file name.
<p>This field exists if bit 0 of <em>flags</em> is set.</p>
</p>
diff --git a/doxygen/examples/ThreadSafeLibrary.html b/doxygen/examples/ThreadSafeLibrary.html
index 2ea42b4..97f7742 100644
--- a/doxygen/examples/ThreadSafeLibrary.html
+++ b/doxygen/examples/ThreadSafeLibrary.html
@@ -165,7 +165,7 @@ The structure is defined in <code>H5private.h</code> as:
<blockquote>
<pre>
- /* cancelability structure */
+ /* cancellability structure */
typedef struct H5_cancel_struct {
int previous_state;
unsigned int cancel_count;
diff --git a/doxygen/examples/VFL.html b/doxygen/examples/VFL.html
index 9776f96..624d942 100644
--- a/doxygen/examples/VFL.html
+++ b/doxygen/examples/VFL.html
@@ -306,7 +306,7 @@ H5Dread(dataset, type, mspace, fspace, buffer, dxpl);
</PRE>
<P>
-The transfer propery list can be queried in a manner similar to the file
+The transfer property list can be queried in a manner similar to the file
access property list: the driver provides a function (or functions) to return
various information about the transfer property list:
@@ -1210,7 +1210,7 @@ Flush all data for file <VAR>file</VAR> to storage.
</P>
<P>
<STRONG>Example:</STRONG> The sec2 driver doesn't cache any data but it also doesn't
-extend the Unix file as agressively as it should. Therefore, when finalizing a
+extend the Unix file as aggressively as it should. Therefore, when finalizing a
file it should write a zero to the last byte of the allocated region so that
when reopening the file later the EOF marker will be at least as large as the
EOA marker saved in the superblock (otherwise HDF5 will refuse to open the
diff --git a/doxygen/hdf5_navtree_hacks.js b/doxygen/hdf5_navtree_hacks.js
index 942970c..dda8984 100644
--- a/doxygen/hdf5_navtree_hacks.js
+++ b/doxygen/hdf5_navtree_hacks.js
@@ -223,7 +223,7 @@ $(document).ready(function() {
(function (){ // wait until the first "selected" element has been created
try {
- // this line will triger an exception if there is no #selected element, i.e., before the tree structure is complete.
+ // this line will trigger an exception if there is no #selected element, i.e., before the tree structure is complete.
document.getElementById("selected").className = "item selected";
// ok, the default tree has been created, we can keep going...
diff --git a/examples/h5_mount.c b/examples/h5_mount.c
index e984d5b..3f087f7 100644
--- a/examples/h5_mount.c
+++ b/examples/h5_mount.c
@@ -109,7 +109,7 @@ main(void)
}
/*
- * Close all identifers
+ * Close all identifiers
*/
H5Tclose(tid);
H5Dclose(did);
diff --git a/examples/h5_shared_mesg.c b/examples/h5_shared_mesg.c
index 0f623b9..f26b360 100644
--- a/examples/h5_shared_mesg.c
+++ b/examples/h5_shared_mesg.c
@@ -109,7 +109,7 @@ main(void)
* This keeps the indexes simple when only a few messages are shared,
* but allows them to scale for many messages. If many messages are
* deleted from the B-tree, it scales back down into a list.
- * A "resonable" setting for maximum list size and minimum btree size
+ * A "reasonable" setting for maximum list size and minimum btree size
* depends on what kinds of messages will be stored in the file.
* These numbers are the same for all indexes in a file.
* We'll guess at some numbers, though we could just as easily have kept
@@ -227,7 +227,7 @@ error:
/*-------------------------------------------------------------------------
* Function: create_standard_file
*
- * Purpose: A helper functon for the example. Creates an HDF5 file
+ * Purpose: A helper function for the example. Creates an HDF5 file
* with many repeated messages using the file creation
* property list FCPL.
*
diff --git a/examples/h5_vds-percival-unlim-maxmin.c b/examples/h5_vds-percival-unlim-maxmin.c
index 7e029a3..f459c20 100644
--- a/examples/h5_vds-percival-unlim-maxmin.c
+++ b/examples/h5_vds-percival-unlim-maxmin.c
@@ -152,7 +152,7 @@ main(void)
wdata[j] = 10 * (i + 1);
/*
- * Open the source files and datasets. Appen data to each dataset and
+ * Open the source files and datasets. Append data to each dataset and
* close all resources.
*/
diff --git a/examples/ph5example.c b/examples/ph5example.c
index a5f3c75..87632e5 100644
--- a/examples/ph5example.c
+++ b/examples/ph5example.c
@@ -30,7 +30,7 @@
* for parallel I/O and there is no standard pathname for parallel file
* systems. In some cases, the parallel file name may even needs some
* parallel file type prefix such as: "pfs:/GF/...". Therefore, this
- * example requires an explicite parallel file prefix. See the usage
+ * example requires an explicit parallel file prefix. See the usage
* for more detail.
*/
@@ -898,7 +898,7 @@ usage(void)
{
printf("Usage: testphdf5 [-f <prefix>] [-r] [-w] [-v]\n");
printf("\t-f\tfile prefix for parallel test files.\n");
- printf("\t \te.g. pfs:/PFS/myname\n");
+ printf("\t \t e.g. pfs:/PFS/myname\n");
printf("\t \tcan be set via $" PARAPREFIX ".\n");
printf("\t \tDefault is current directory.\n");
printf("\t-c\tno cleanup\n");
diff --git a/examples/testh5cc.sh.in b/examples/testh5cc.sh.in
index ba80f2d..a8974bf 100644
--- a/examples/testh5cc.sh.in
+++ b/examples/testh5cc.sh.in
@@ -80,7 +80,7 @@ applib=libapp${H5TOOL}.a
# short hands
# Caution: if some *.h5 files must be cleaned here, list them by names.
# Don't use the wildcard form of *.h5 as it will wipe out even *.h5 generated
-# by otehr test programs. This will cause a racing condition error when
+# by other test programs. This will cause a racing condition error when
# parallel make (e.g., gmake -j 4) is used.
temp_SRC="$hdf5main $v16main $v18main $v110main $v112main $appmain $prog1 $prog2"
temp_OBJ=`echo $temp_SRC | sed -e "s/\.${suffix}/.o/g"`
diff --git a/fortran/examples/compound.f90 b/fortran/examples/compound.f90
index d9f8235..57624c6 100644
--- a/fortran/examples/compound.f90
+++ b/fortran/examples/compound.f90
@@ -41,7 +41,7 @@
INTEGER(HID_T) :: dt3_id ! Memory datatype identifier (for double precision field)
INTEGER(HID_T) :: dt4_id ! Memory datatype identifier (for real field)
INTEGER(HID_T) :: dt5_id ! Memory datatype identifier
- INTEGER(HID_T) :: plist_id ! Dataset trasfer property
+ INTEGER(HID_T) :: plist_id ! Dataset transfer property
INTEGER(SIZE_T) :: typesize
@@ -111,21 +111,21 @@
type_size = type_sizec + type_sizei + type_sized + type_sizer
CALL h5tcreate_f(H5T_COMPOUND_F, type_size, dtype_id, error)
!
- ! Insert memebers
+ ! Insert members
!
- ! CHARACTER*2 memeber
+ ! CHARACTER*2 member
!
offset = 0
CALL h5tinsert_f(dtype_id, "char_field", offset, dt5_id, error)
!
! INTEGER member
!
- offset = offset + type_sizec ! Offset of the second memeber is 2
+ offset = offset + type_sizec ! Offset of the second member is 2
CALL h5tinsert_f(dtype_id, "integer_field", offset, H5T_NATIVE_INTEGER, error)
!
! DOUBLE PRECISION member
!
- offset = offset + type_sizei ! Offset of the third memeber is 6
+ offset = offset + type_sizei ! Offset of the third member is 6
CALL h5tinsert_f(dtype_id, "double_field", offset, H5T_NATIVE_DOUBLE, error)
!
! REAL member
@@ -198,7 +198,7 @@
!
CALL h5dopen_f(file_id, dsetname, dset_id, error)
!
- ! Create memeory datatyoe to read character member of the compound datatype.
+ ! Create memory datatype to read character member of the compound datatype.
!
CALL h5tcopy_f(H5T_NATIVE_CHARACTER, dt2_id, error)
typesize = 2
@@ -208,7 +208,7 @@
offset = 0
CALL h5tinsert_f(dt1_id, "char_field", offset, dt2_id, error)
!
- ! Read part of the datatset and display it.
+ ! Read part of the dataset and display it.
!
CALL h5dread_f(dset_id, dt1_id, char_member_out, data_dims, error)
write(*,*) (char_member_out(i), i=1, dimsize)
diff --git a/fortran/examples/h5_cmprss.f90 b/fortran/examples/h5_cmprss.f90
index 61efc30..86d9cfa 100644
--- a/fortran/examples/h5_cmprss.f90
+++ b/fortran/examples/h5_cmprss.f90
@@ -51,7 +51,7 @@ PROGRAM h5_cmprss
!INTEGER :: szip_pixels_per_block
!
- !Initialize FORTRAN predifined datatypes
+ !Initialize FORTRAN predefined datatypes
!
CALL h5open_f(error)
!
diff --git a/fortran/examples/h5_crtatt.f90 b/fortran/examples/h5_crtatt.f90
index 996a506..64d2bde 100644
--- a/fortran/examples/h5_crtatt.f90
+++ b/fortran/examples/h5_crtatt.f90
@@ -35,7 +35,7 @@ PROGRAM H5_CRTATT
INTEGER(HID_T) :: aspace_id ! Attribute Dataspace identifier
INTEGER(HID_T) :: atype_id ! Attribute Dataspace identifier
INTEGER(HSIZE_T), DIMENSION(1) :: adims = (/2/) ! Attribute dimension
- INTEGER :: arank = 1 ! Attribure rank
+ INTEGER :: arank = 1 ! Attribute rank
INTEGER(SIZE_T) :: attrlen ! Length of the attribute string
CHARACTER(LEN=80), DIMENSION(2) :: attr_data ! Attribute data
diff --git a/fortran/examples/h5_extend.f90 b/fortran/examples/h5_extend.f90
index 0ad6ef9..193f33e 100644
--- a/fortran/examples/h5_extend.f90
+++ b/fortran/examples/h5_extend.f90
@@ -86,7 +86,7 @@ PROGRAM H5_EXTEND
INTEGER, DIMENSION(1:3,1:10) :: rdata
!
- !Initialize FORTRAN predifined datatypes
+ !Initialize FORTRAN predefined datatypes
!
CALL h5open_f(error)
@@ -110,7 +110,7 @@ PROGRAM H5_EXTEND
CALL h5pset_chunk_f(crp_list, RANK, dimsc, error)
!
- !Create a dataset with 3X3 dimensions using cparms creation propertie .
+ !Create a dataset with 3X3 dimensions using cparms creation properties .
!
CALL h5dcreate_f(file_id, dsetname, H5T_NATIVE_INTEGER, dataspace, &
dset_id, error, crp_list )
diff --git a/fortran/examples/refobjexample.f90 b/fortran/examples/refobjexample.f90
index d420192..873b420 100644
--- a/fortran/examples/refobjexample.f90
+++ b/fortran/examples/refobjexample.f90
@@ -26,7 +26,7 @@
IMPLICIT NONE
CHARACTER(LEN=10), PARAMETER :: filename = "FORTRAN.h5" ! File
CHARACTER(LEN=8), PARAMETER :: dsetnamei = "INTEGERS" ! Dataset with the integer data
- CHARACTER(LEN=17), PARAMETER :: dsetnamer = "OBJECT_REFERENCES" ! Dataset wtih object
+ CHARACTER(LEN=17), PARAMETER :: dsetnamer = "OBJECT_REFERENCES" ! Dataset with object
! references
CHARACTER(LEN=6), PARAMETER :: groupname1 = "GROUP1" ! Groups in the file
CHARACTER(LEN=6), PARAMETER :: groupname2 = "GROUP2" !
diff --git a/fortran/examples/testh5fc.sh.in b/fortran/examples/testh5fc.sh.in
index db687d8..71f41f8 100644
--- a/fortran/examples/testh5fc.sh.in
+++ b/fortran/examples/testh5fc.sh.in
@@ -61,7 +61,7 @@ applib=libapp${H5TOOL}.a
# short hands
# Caution: if some *.h5 files must be cleaned here, list them by names.
# Don't use the wildcard form of *.h5 as it will wipe out even *.h5 generated
-# by otehr test programs. This will cause a racing condition error when
+# by other test programs. This will cause a racing condition error when
# parallel make (e.g., gmake -j 4) is used.
temp_SRC="$hdf5main $appmain $prog1 $prog2"
temp_OBJ=`echo $temp_SRC | sed -e "s/\.${suffix}/.o/g"`
diff --git a/fortran/src/H5Af.c b/fortran/src/H5Af.c
index f78ade2..db1c040 100644
--- a/fortran/src/H5Af.c
+++ b/fortran/src/H5Af.c
@@ -272,7 +272,7 @@ done:
* PURPOSE
* Call H5Aopen to open an attribute
* INPUTS
- * obj_id - Identifer for object to which attribute is attached
+ * obj_id - Identifier for object to which attribute is attached
* attr_name - Attribute access property list
* attr_namelen - size of attr_name
* aapl_id - Link access property list
@@ -317,7 +317,7 @@ done:
* PURPOSE
* Call h5adelete_by_name to remove an attribute from a specified location
* INPUTS
- * loc_id - identifer for object to which attribute is attached
+ * loc_id - identifier for object to which attribute is attached
* obj_name - object identifier
* obj_namelen - name length
* attr_name - name of the attribute
@@ -424,7 +424,7 @@ done:
* INPUTS
*
*
- * loc_id - Identifer for object to which attribute is attached
+ * loc_id - Identifier for object to which attribute is attached
* obj_name - Name of object, relative to location,
* from which attribute is to be removed *TEST* check NULL
* idx_type - Type of index; Possible values are:
@@ -532,7 +532,7 @@ done:
* aapl_id - Attribute access property list
* lapl_id - Link access property list
* OUTPUTS
- * attr_id - attribute identifer
+ * attr_id - attribute identifier
* RETURNS
* 0 on success, -1 on failure
* AUTHOR
diff --git a/fortran/src/H5Aff.F90 b/fortran/src/H5Aff.F90
index 8f56993..53f0a39 100644
--- a/fortran/src/H5Aff.F90
+++ b/fortran/src/H5Aff.F90
@@ -211,7 +211,7 @@ CONTAINS
!
! INPUTS
! obj_id - identifier of a group, dataset, or named
-! datatype atttribute to be attached to
+! datatype attribute to be attached to
! name - attribute name
! OUTPUTS
! attr_id - attribute identifier
@@ -487,7 +487,7 @@ CONTAINS
SUBROUTINE h5aget_name_by_idx_f(loc_id, obj_name, idx_type, order, &
n, name, hdferr, size, lapl_id)
IMPLICIT NONE
- INTEGER(HID_T), INTENT(IN) :: loc_id ! Identifer for object to which attribute is attached
+ INTEGER(HID_T), INTENT(IN) :: loc_id ! Identifier for object to which attribute is attached
CHARACTER(LEN=*), INTENT(IN) :: obj_name ! Name of object, relative to location,
! from which attribute is to be removed *TEST* check NULL
INTEGER, INTENT(IN) :: idx_type ! Type of index; Possible values are:
@@ -846,7 +846,7 @@ CONTAINS
! identifier and attribute name
!
! INPUTS
-! obj_id - Identifer for object to which attribute is attached
+! obj_id - Identifier for object to which attribute is attached
! attr_name - Name of attribute to open
! OUTPUTS
! attr_id - attribute identifier
@@ -934,7 +934,7 @@ CONTAINS
! SOURCE
SUBROUTINE H5Adelete_by_idx_f(loc_id, obj_name, idx_type, order, n, hdferr, lapl_id)
IMPLICIT NONE
- INTEGER(HID_T), INTENT(IN) :: loc_id ! Identifer for object to which attribute is attached
+ INTEGER(HID_T), INTENT(IN) :: loc_id ! Identifier for object to which attribute is attached
CHARACTER(LEN=*), INTENT(IN) :: obj_name ! Name of object, relative to location,
! from which attribute is to be removed
INTEGER, INTENT(IN) :: idx_type ! Type of index; Possible values are:
@@ -991,7 +991,7 @@ CONTAINS
! Removes an attribute from a specified location
!
! INPUTS
-! loc_id - Identifer for object to which attribute is attached
+! loc_id - Identifier for object to which attribute is attached
! obj_name - Name of attribute to open
! attr_name - Attribute access property list
! lapl_id - Link access property list
@@ -1005,7 +1005,7 @@ CONTAINS
! SOURCE
SUBROUTINE H5Adelete_by_name_f(loc_id, obj_name, attr_name, hdferr, lapl_id)
IMPLICIT NONE
- INTEGER(HID_T), INTENT(IN) :: loc_id ! Identifer for object to which attribute is attached
+ INTEGER(HID_T), INTENT(IN) :: loc_id ! Identifier for object to which attribute is attached
CHARACTER(LEN=*), INTENT(IN) :: obj_name ! Name of object, relative to location,
! from which attribute is to be removed
CHARACTER(LEN=*), INTENT(IN) :: attr_name ! Name of attribute to delete
@@ -1841,7 +1841,7 @@ CONTAINS
! dims parameter was added to make code portable;
! Aprile 4, 2001
!
-! Changed buf intent to INOUT to be consistant
+! Changed buf intent to INOUT to be consistent
! with how the C functions handles it. The pg
! compiler will return 0 if a buf value is not set.
! February, 2008
diff --git a/fortran/src/H5Df.c b/fortran/src/H5Df.c
index bcf792c..5c46cb3 100644
--- a/fortran/src/H5Df.c
+++ b/fortran/src/H5Df.c
@@ -145,7 +145,7 @@ DONE:
* Elena Pourmal
* Tuesday, May 14, 2002
* HISTORY
- * This function was added to accomodate h5dwrite_f with the
+ * This function was added to accommodate h5dwrite_f with the
* dims argument being of INTEGER(HSIZE_T) type
* SOURCE
*/
@@ -217,7 +217,7 @@ h5dwrite_ref_reg_c(hid_t_f *dset_id, hid_t_f *mem_type_id, hid_t_f *mem_space_id
* Elena Pourmal
* Wednesday, May 15, 2002
* HISTORY
- * This function was added to accomodate h5dread_f subroutine
+ * This function was added to accommodate h5dread_f subroutine
* with the dims parameter being of INTEGER(HSIZE_T_F) size.
* SOURCE
*/
@@ -783,7 +783,7 @@ h5dwrite_vl_string_c(hid_t_f *dset_id, hid_t_f *mem_type_id, hid_t_f *mem_space_
return ret_value;
}
/*
- * Move data from temorary buffer
+ * Move data from temporary buffer
*/
tmp_p = tmp;
for (i = 0; i < num_elem; i++) {
diff --git a/fortran/src/H5Dff.F90 b/fortran/src/H5Dff.F90
index 75af58e..655cf1c 100644
--- a/fortran/src/H5Dff.F90
+++ b/fortran/src/H5Dff.F90
@@ -476,7 +476,7 @@ CONTAINS
!
! Changed name from the now obsolete h5dextend_f
! to h5dset_extent_f. Provided interface to old name
-! for backward compatability. -MSB- March 14, 2008
+! for backward compatibility. -MSB- March 14, 2008
!
! SOURCE
SUBROUTINE h5dset_extent_f(dataset_id, size, hdferr)
@@ -1195,7 +1195,7 @@ CONTAINS
IMPLICIT NONE
INTEGER(HID_T), INTENT(IN) :: dset_id ! Dataset identifier
INTEGER(HID_T), INTENT(IN) :: mem_type_id ! Memory datatype identifier
- INTEGER(HSIZE_T), DIMENSION(*), INTENT(IN) :: dims ! size of the bufffer buf
+ INTEGER(HSIZE_T), DIMENSION(*), INTENT(IN) :: dims ! size of the buffer buf
TYPE(hobj_ref_t_f), DIMENSION(dims(1)), INTENT(IN), TARGET :: buf ! Data buffer
INTEGER, INTENT(OUT) :: hdferr ! Error code
INTEGER(HID_T), OPTIONAL, INTENT(IN) :: mem_space_id ! Memory dataspace identfier
@@ -1227,7 +1227,7 @@ CONTAINS
IMPLICIT NONE
INTEGER(HID_T), INTENT(IN) :: dset_id ! Dataset identifier
INTEGER(HID_T), INTENT(IN) :: mem_type_id ! Memory datatype identifier
- INTEGER(HSIZE_T), DIMENSION(*), INTENT(IN) :: dims ! size of the bufffer buf
+ INTEGER(HSIZE_T), DIMENSION(*), INTENT(IN) :: dims ! size of the buffer buf
TYPE(hdset_reg_ref_t_f), DIMENSION(dims(1)), INTENT(IN), TARGET :: buf ! Data buffer
INTEGER, INTENT(OUT) :: hdferr ! Error code
INTEGER(HID_T), OPTIONAL, INTENT(IN) :: mem_space_id ! Memory dataspace identfier
diff --git a/fortran/src/H5Fff.F90 b/fortran/src/H5Fff.F90
index a273431..ecb40b7 100644
--- a/fortran/src/H5Fff.F90
+++ b/fortran/src/H5Fff.F90
@@ -97,7 +97,7 @@ CONTAINS
INTEGER(HID_T), INTENT(OUT) :: file_id ! File identifier
INTEGER, INTENT(OUT) :: hdferr ! Error code
INTEGER(HID_T), OPTIONAL, INTENT(IN) :: creation_prp
- ! File creation propertly
+ ! File creation property
! list identifier
INTEGER(HID_T), OPTIONAL, INTENT(IN) :: access_prp
! File access property list
diff --git a/fortran/src/H5Gf.c b/fortran/src/H5Gf.c
index f445fee..0201761 100644
--- a/fortran/src/H5Gf.c
+++ b/fortran/src/H5Gf.c
@@ -390,13 +390,13 @@ DONE:
* Call H5Glink2 to link the specified type
* INPUTS
* cur_loc_id - identifier of file or group
- * cur_name - name of the existing object for hard link releative
+ * cur_name - name of the existing object for hard link relative
* to cur_loc_id location,
* anything for the soft link
* current_namelen - current name length
* link_type - link type
* new_loc_id - location identifier
- * new_name - new name for the object releative to the new_loc_id
+ * new_name - new name for the object relative to the new_loc_id
* location
* new_namelen - new_name length
* RETURNS
@@ -879,7 +879,7 @@ done:
* February 15, 2008
* HISTORY
*
- * - Added 'mounted' paramater
+ * - Added 'mounted' parameter
* M. Scot Breitenfeld
* July 16, 2008
* SOURCE
@@ -1014,7 +1014,7 @@ done:
* February 18, 2008
* HISTORY
*
- * - Added 'mounted' paramater
+ * - Added 'mounted' parameter
* M. Scot Breitenfeld
* July 16, 2008
* SOURCE
diff --git a/fortran/src/H5Gff.F90 b/fortran/src/H5Gff.F90
index 9e7665f..bfca595 100644
--- a/fortran/src/H5Gff.F90
+++ b/fortran/src/H5Gff.F90
@@ -73,7 +73,7 @@ CONTAINS
! called C functions (it is needed for Windows
! port). March 5, 2001
!
-! Added additional optional paramaters in 1.8
+! Added additional optional parameters in 1.8
! MSB - February 27, 2008
!
! SOURCE
@@ -491,8 +491,8 @@ CONTAINS
INTEGER, INTENT(OUT) :: hdferr ! Error code
!*****
- INTEGER :: current_namelen ! Lenghth of the current_name string
- INTEGER :: new_namelen ! Lenghth of the new_name string
+ INTEGER :: current_namelen ! Length of the current_name string
+ INTEGER :: new_namelen ! Length of the new_name string
INTERFACE
INTEGER FUNCTION h5glink_c(loc_id, link_type, current_name, &
@@ -524,7 +524,7 @@ CONTAINS
! PURPOSE
! Creates a link of the specified type from new_name
! to current_name. current_name and new_name are interpreted
-! releative to current and new location identifiers.
+! relative to current and new location identifiers.
!
! INPUTS
! cur_loc_id - location identifier
@@ -559,8 +559,8 @@ CONTAINS
INTEGER, INTENT(OUT) :: hdferr ! Error code
!*****
- INTEGER :: cur_namelen ! Lenghth of the current_name string
- INTEGER :: new_namelen ! Lenghth of the new_name string
+ INTEGER :: cur_namelen ! Length of the current_name string
+ INTEGER :: new_namelen ! Length of the new_name string
INTERFACE
INTEGER FUNCTION h5glink2_c(cur_loc_id, cur_name, cur_namelen, &
@@ -617,7 +617,7 @@ CONTAINS
CHARACTER(LEN=*), INTENT(IN) :: name ! Name of an object
INTEGER, INTENT(OUT) :: hdferr ! Error code
!*****
- INTEGER :: namelen ! Lenghth of the name character string
+ INTEGER :: namelen ! Length of the name character string
INTERFACE
INTEGER FUNCTION h5gunlink_c(loc_id, name, namelen) BIND(C,NAME='h5gunlink_c')
@@ -666,8 +666,8 @@ CONTAINS
CHARACTER(LEN=*), INTENT(IN) :: new_name ! New name of an object
INTEGER, INTENT(OUT) :: hdferr ! Error code
!*****
- INTEGER :: namelen ! Lenghth of the current_name string
- INTEGER :: new_namelen ! Lenghth of the new_name string
+ INTEGER :: namelen ! Length of the current_name string
+ INTEGER :: new_namelen ! Length of the new_name string
INTERFACE
INTEGER FUNCTION h5gmove_c(loc_id, name, namelen, new_name, new_namelen) BIND(C,NAME='h5gmove_c')
@@ -715,7 +715,7 @@ CONTAINS
INTEGER, INTENT(OUT) :: hdferr ! Error code
!*****
INTEGER :: src_namelen ! Length of the current_name string
- INTEGER :: dst_namelen ! Lenghth of the new_name string
+ INTEGER :: dst_namelen ! Length of the new_name string
INTERFACE
INTEGER FUNCTION h5gmove2_c(src_loc_id, src_name, src_namelen, &
@@ -776,7 +776,7 @@ CONTAINS
! points to
INTEGER, INTENT(OUT) :: hdferr ! Error code
!*****
- INTEGER :: namelen ! Lenghth of the current_name string
+ INTEGER :: namelen ! Length of the current_name string
INTERFACE
INTEGER FUNCTION h5gget_linkval_c(loc_id, name, namelen, size, buffer) BIND(C,NAME='h5gget_linkval_c')
@@ -827,8 +827,8 @@ CONTAINS
CHARACTER(LEN=*), INTENT(IN) :: comment ! New name of an object
INTEGER, INTENT(OUT) :: hdferr ! Error code
!*****
- INTEGER :: namelen ! Lenghth of the current_name string
- INTEGER :: commentlen ! Lenghth of the comment string
+ INTEGER :: namelen ! Length of the current_name string
+ INTEGER :: commentlen ! Length of the comment string
INTERFACE
INTEGER FUNCTION h5gset_comment_c(loc_id, name, namelen, &
@@ -1024,7 +1024,7 @@ CONTAINS
!
! HISTORY
!
-! - Added 'mounted' paramater
+! - Added 'mounted' parameter
! M. Scot Breitenfeld
! July 16, 2008
!
@@ -1104,7 +1104,7 @@ CONTAINS
! February 18, 2008
!
! HISTORY
-! Added 'mounted' paramater
+! Added 'mounted' parameter
! M. Scot Breitenfeld
! July 16, 2008
!
@@ -1206,7 +1206,7 @@ CONTAINS
! February 18, 2008
!
! HISTORY
-! Added 'mounted' paramater
+! Added 'mounted' parameter
! M. Scot Breitenfeld
! July 16, 2008
! SOURCE
diff --git a/fortran/src/H5Lff.F90 b/fortran/src/H5Lff.F90
index 3bfcf65..a568824 100644
--- a/fortran/src/H5Lff.F90
+++ b/fortran/src/H5Lff.F90
@@ -469,7 +469,7 @@ CONTAINS
! SOURCE
SUBROUTINE h5ldelete_by_idx_f(loc_id, group_name, index_field, order, n, hdferr, lapl_id)
IMPLICIT NONE
- INTEGER(HID_T), INTENT(IN) :: loc_id ! Identifer for object to which attribute is attached
+ INTEGER(HID_T), INTENT(IN) :: loc_id ! Identifier for object to which attribute is attached
CHARACTER(LEN=*), INTENT(IN) :: group_name ! Name of object, relative to location,
! from which attribute is to be removed
INTEGER, INTENT(IN) :: index_field ! Type of index; Possible values are:
@@ -1315,7 +1315,7 @@ CONTAINS
! Inputs:
! group_id - Identifier specifying subject group
! index_type - Type of index which determines the order:
-! H5_INDEX_NAME_F - Alpha-numeric index on name
+! H5_INDEX_NAME_F - Alphanumeric index on name
! H5_INDEX_CRT_ORDER_F - Index on creation order
! order - Order within index:
! H5_ITER_INC_F - Increasing order
@@ -1391,7 +1391,7 @@ CONTAINS
! loc_id - File or group identifier specifying location of subject group
! group_name - Name of subject group
! index_type - Type of index which determines the order:
-! H5_INDEX_NAME_F - Alpha-numeric index on name
+! H5_INDEX_NAME_F - Alphanumeric index on name
! H5_INDEX_CRT_ORDER_F - Index on creation order
! order - Order within index:
! H5_ITER_INC_F - Increasing order
diff --git a/fortran/src/H5Pf.c b/fortran/src/H5Pf.c
index 1f74955..ab09ab9 100644
--- a/fortran/src/H5Pf.c
+++ b/fortran/src/H5Pf.c
@@ -145,7 +145,7 @@ done:
* plist1_id - property list identifier
* plist2_id - property list identifier
* OUTPUTS
- * c_flag - flag to indicate that lists are eqaul
+ * c_flag - flag to indicate that lists are equal
* RETURNS
* 0 on success, -1 on failure
* AUTHOR
@@ -936,7 +936,7 @@ h5pset_fapl_stdio_c(hid_t_f *prp_id)
* NAME
* h5pget_fapl_stdio_c
* PURPOSE
- * Call H5Pget_fapl_stdio to determine whther the low level file driver
+ * Call H5Pget_fapl_stdio to determine whether the low level file driver
* uses the functions declared in the stdio.h
* INPUTS
* prp_id - property list identifier
@@ -1013,7 +1013,7 @@ h5pset_fapl_sec2_c(hid_t_f *prp_id)
* NAME
* h5pget_fapl_sec2_c
* PURPOSE
- * Call H5Pget_fapl_stdio to determine whther the low level file driver
+ * Call H5Pget_fapl_stdio to determine whether the low level file driver
* uses the functions declared in the unistd.h
* INPUTS
* prp_id - property list identifier
@@ -2186,7 +2186,7 @@ h5pset_fclose_degree_c(hid_t_f *fapl_id, int_f *degree)
* PURPOSE
* Call H5Pset_buffer to set size of conversion buffer
* INPUTS
- * prp_id - t`dataset trasfer property list identifier
+ * prp_id - t`dataset transfer property list identifier
* size - size of the buffer
* OUTPUTS
* NONE
@@ -2221,7 +2221,7 @@ h5pset_buffer_c(hid_t_f *prp_id, hsize_t_f *size)
* PURPOSE
* Call H5Pget_buffer to get size of conversion buffer
* INPUTS
- * prp_id - t`dataset trasfer property list identifier
+ * prp_id - t`dataset transfer property list identifier
* OUTPUTS
* size - size of conversion buffer
* RETURNS
@@ -3005,11 +3005,11 @@ h5pget_nprops_c(hid_t_f *plist, size_t_f *nprops)
* h5pget_class_parent_c
* PURPOSE
* Call H5Pget_class_parent to get the parent class of
- * a genereic property class
+ * a generic property class
* INPUTS
* prp_id - property list to query
* OUTPUTS
- * parent_id - parent classs identifier
+ * parent_id - parent class identifier
* RETURNS
* 0 on success, -1 on failure
* AUTHOR
@@ -3505,7 +3505,7 @@ h5pset_family_offset_c(hid_t_f *prp_id, hsize_t_f *offset)
* NAME
* h5pset_fapl_multi_c
* PURPOSE
- * Call H5Pset_fapl_multi to set multi file dirver
+ * Call H5Pset_fapl_multi to set multi file driver
* INPUTS
* prp_id - file_creation property list identifier
* mem_map - memory mapping array
@@ -3570,7 +3570,7 @@ h5pset_fapl_multi_c(hid_t_f *prp_id, int_f *memb_map, hid_t_f *memb_fapl, _fcd m
tmp_p = tmp_p + c_lenmax;
}
/*
- * Take care of othe arguments
+ * Take care of other arguments
*/
tmp_max_addr = (long double)(HADDR_MAX);
c_prp_id = (hid_t)*prp_id;
@@ -3603,7 +3603,7 @@ DONE:
* NAME
* h5pset_fapl_multi_sc
* PURPOSE
- * Call H5Pset_fapl_multi to set multi file dirver
+ * Call H5Pset_fapl_multi to set multi file driver
* INPUTS
* prp_id - file_creation property list identifier
* RETURNS
@@ -3641,7 +3641,7 @@ h5pset_fapl_multi_sc(hid_t_f *prp_id, int_f *flag)
* NAME
* h5pget_fapl_multi_c
* PURPOSE
- * Call H5Pget_fapl_multi to set multi file dirver
+ * Call H5Pget_fapl_multi to set multi file driver
* INPUTS
* prp_id - file_creation property list identifier
* lenmax - length of the name a sdeclared in Fortran
@@ -4607,7 +4607,7 @@ h5pget_copy_object_c(hid_t_f *ocp_plist_id, int_f *copy_options)
* INPUTS
*
* prp_id - property list identifier to query
- * expression_len - buffer size transorm expression
+ * expression_len - buffer size transform expression
*
* Output:
* expression - buffer to hold transform expression
@@ -4670,7 +4670,7 @@ done:
*
* prp_id - property list identifier to query
* expression - buffer to hold transform expression
- * expression_len - buffer size transorm expression
+ * expression_len - buffer size transform expression
*
* Output:
*
@@ -5478,7 +5478,7 @@ h5pget_fapl_mpio_c(hid_t_f *prp_id, int_f *comm, int_f *info)
* h5pset_dxpl_mpio_c
* PURPOSE
* Call H5Pset_dxpl_mpio to set transfer mode of the dataset
- * trasfer property list
+ * transfer property list
* INPUTS
* prp_id - property list identifier
* data_xfer_mode - transfer mode
@@ -5530,7 +5530,7 @@ h5pset_dxpl_mpio_c(hid_t_f *prp_id, int_f *data_xfer_mode)
* h5pget_dxpl_mpio_c
* PURPOSE
* Call H5Pget_dxpl_mpio to get transfer mode of the dataset
- * trasfer property list
+ * transfer property list
* INPUTS
* prp_id - property list identifier
* data_xfer_mode - buffer to retrieve transfer mode
diff --git a/fortran/src/H5Pff.F90 b/fortran/src/H5Pff.F90
index a7f2366..c55b7d9 100644
--- a/fortran/src/H5Pff.F90
+++ b/fortran/src/H5Pff.F90
@@ -500,7 +500,7 @@ CONTAINS
! a chunked layout dataset.
!
! INPUTS
-! prp_id - datatset creation property list identifier
+! prp_id - dataset creation property list identifier
! ndims - number of dimensions for each chunk
! dims - array with dimension sizes for each chunk
! OUTPUTS
@@ -1880,7 +1880,7 @@ CONTAINS
! INPUTS
!
! prp_id - file access property list identifier
-! gc_reference - flag for stting garbage collection on
+! gc_reference - flag for setting garbage collection on
! and off (1 or 0)
! OUTPUTS
!
@@ -1931,7 +1931,7 @@ CONTAINS
! prp_id - file access property list identifier
! OUTPUTS
!
-! gc_reference - flag for stting garbage collection on
+! gc_reference - flag for setting garbage collection on
! and off (1 or 0)
! hdferr - error code
! Success: 0
@@ -2659,7 +2659,7 @@ CONTAINS
! h5pequal_f
!
! PURPOSE
-! Checks if two property lists are eqaul
+! Checks if two property lists are equal
!
! INPUTS
!
@@ -3059,7 +3059,7 @@ CONTAINS
! INPUTS
!
! plist_id - file access property list identifier
-! size - metatdata block size
+! size - metadata block size
! OUTPUTS
!
! hdferr - error code
@@ -3103,7 +3103,7 @@ CONTAINS
! plist_id - file access property list identifier
! OUTPUTS
!
-! size - metatdata block size
+! size - metadata block size
! hdferr - error code
! Success: 0
! Failure: -1
@@ -3596,7 +3596,7 @@ CONTAINS
! h5pget_class_parent_f
!
! PURPOSE
-! Retrieves the parent class of a genric property class.
+! Retrieves the parent class of a generic property class.
!
! INPUTS
!
@@ -5391,7 +5391,7 @@ CONTAINS
!
! HISTORY
!
-! Should hdferr return just 0 or 1 and add another arguement for the size?
+! Should hdferr return just 0 or 1 and add another argument for the size?
! Fortran90 Interface:
SUBROUTINE h5pget_data_transform_f(plist_id, expression, hdferr, size)
IMPLICIT NONE
diff --git a/fortran/src/H5Sf.c b/fortran/src/H5Sf.c
index 341acfd..ea42a31 100644
--- a/fortran/src/H5Sf.c
+++ b/fortran/src/H5Sf.c
@@ -388,7 +388,7 @@ done:
* PURPOSE
* Call H5Sget_select_elem_pointlist
* get a list of element points in the
- * current dataspace selectin.
+ * current dataspace selection.
* Starting with the startpoint-th point in the
* list of points, numpoints points are put into the user's
* buffer. If the user's buffer fills up before numpoints
@@ -759,7 +759,7 @@ done:
* space_id - identifier of the dataspace
* rank - dataspace rank
* current_size - array with the new dimension sizes
- * maximum_size - aray with maximum sizes of dimensions
+ * maximum_size - array with maximum sizes of dimensions
* RETURNS
* 0 on success, -1 on failure
* AUTHOR
@@ -804,7 +804,7 @@ done:
* space_id - identifier of the dataspace
* OUTPUTS
* dims - array with the dimension sizes
- * maxdims - aray with maximum sizes of dimensions
+ * maxdims - array with maximum sizes of dimensions
* RETURNS
* number of dataspace dimensions (rank) on success, -1 on failure
* AUTHOR
@@ -852,7 +852,7 @@ done:
* NAME
* h5sis_simple_c
* PURPOSE
- * Call H5Sis_simple to detrmine if the dataspace
+ * Call H5Sis_simple to determine if the dataspace
* is simple.
* INPUTS
* space_id - identifier of the dataspace
diff --git a/fortran/src/H5Sff.F90 b/fortran/src/H5Sff.F90
index 62d665a..76b0dea 100644
--- a/fortran/src/H5Sff.F90
+++ b/fortran/src/H5Sff.F90
@@ -447,7 +447,7 @@ CONTAINS
! INPUTS
! space_id - dataspace identifier
! startpoint - element point to start with
-! num_points - number of elemnt points to get
+! num_points - number of element points to get
! OUTPUTS
! buf - buffer with element points selected
! hdferr - Returns 0 if successful and -1 if fails
diff --git a/fortran/src/H5Tf.c b/fortran/src/H5Tf.c
index 1516cc7..f3016f1 100644
--- a/fortran/src/H5Tf.c
+++ b/fortran/src/H5Tf.c
@@ -1560,7 +1560,7 @@ h5tget_member_type_c(hid_t_f *type_id, int_f *field_idx, hid_t_f *datatype)
* Call H5Tcreate to create a datatype
* INPUTS
* cls - class type
- * size - size of the class memeber
+ * size - size of the class member
* RETURNS
* 0 on success, -1 on failure
* AUTHOR
@@ -2033,7 +2033,7 @@ h5tvlen_create_c(hid_t_f *type_id, hid_t_f *vltype_id)
* NAME
* h5tis_variable_str_c
* PURPOSE
- * Call H5Tis_variable_str to detrmine if the datatype
+ * Call H5Tis_variable_str to determine if the datatype
* is a variable string.
* INPUTS
* type_id - identifier of the dataspace
@@ -2069,7 +2069,7 @@ h5tis_variable_str_c(hid_t_f *type_id, int_f *flag)
* NAME
* h5tget_member_class_c
* PURPOSE
- * Call H5Tget_member_class to detrmine ithe class of the compound
+ * Call H5Tget_member_class to determine ithe class of the compound
* datatype member
* INPUTS
* type_id - identifier of the dataspace
diff --git a/fortran/src/H5Tff.F90 b/fortran/src/H5Tff.F90
index 0eab39e..7582dab 100644
--- a/fortran/src/H5Tff.F90
+++ b/fortran/src/H5Tff.F90
@@ -86,7 +86,7 @@ CONTAINS
! called C functions (it is needed for Windows
! port). March 7, 2001
!
-! Added optional parameter 'tapl_id' for compatability
+! Added optional parameter 'tapl_id' for compatibility
! with H5Topen2. April 9, 2009.
!
! SOURCE
@@ -212,7 +212,7 @@ CONTAINS
! h5tcopy_f
!
! PURPOSE
-! Creates a copy of exisiting datatype.
+! Creates a copy of existing datatype.
!
! INPUTS
! type_id - datatype identifier
diff --git a/fortran/src/H5_buildiface.F90 b/fortran/src/H5_buildiface.F90
index fb3a002..090b6db 100644
--- a/fortran/src/H5_buildiface.F90
+++ b/fortran/src/H5_buildiface.F90
@@ -14,7 +14,7 @@
!
! NOTES
! This program uses the Fortran 2008 intrinsic function STORAGE_SIZE or SIZEOF
-! depending on availablity.It generates code that makes use of
+! depending on availability.It generates code that makes use of
! STORAGE_SIZE/SIZEOF in H5_gen.F90. STORAGE_SIZE is standard
! compliant and should always be chosen over SIZEOF.
!
@@ -437,7 +437,7 @@ PROGRAM H5_buildiface
! dims parameter was added to make code portable;
! Aprile 4, 2001
!
-! Changed buf intent to INOUT to be consistant
+! Changed buf intent to INOUT to be consistent
! with how the C functions handles it. The pg
! compiler will return 0 if a buf value is not set.
! February, 2008
@@ -564,7 +564,7 @@ PROGRAM H5_buildiface
! dims parameter was added to make code portable;
! Aprile 4, 2001
!
-! Changed buf intent to INOUT to be consistant
+! Changed buf intent to INOUT to be consistent
! with how the C functions handles it. The pg
! compiler will return 0 if a buf value is not set.
! February, 2008
@@ -677,7 +677,7 @@ PROGRAM H5_buildiface
! Optional parameters:
! mem_space_id - memory dataspace identifier
! file_space_id - file dataspace identifier
-! xfer_prp - trasfer property list identifier
+! xfer_prp - transfer property list identifier
!
! AUTHOR
! Elena Pourmal
diff --git a/fortran/src/H5_ff.F90 b/fortran/src/H5_ff.F90
index f56cdd9..350d978 100644
--- a/fortran/src/H5_ff.F90
+++ b/fortran/src/H5_ff.F90
@@ -701,7 +701,7 @@ CONTAINS
!
! Outputs:
! majnum - major version of the library
-! minum - minor version of the library
+! minnum - minor version of the library
! relnum - release version of the library
! error - Returns 0 if successful and -1 if fails
!
@@ -736,7 +736,7 @@ CONTAINS
!
! Inputs:
! majnum - major version of the library
-! minum - minor version of the library
+! minnum - minor version of the library
! relnum - release version of the library
!
! Outputs:
diff --git a/fortran/src/H5f90global.F90 b/fortran/src/H5f90global.F90
index 8f2b5ae..eb3b87e 100644
--- a/fortran/src/H5f90global.F90
+++ b/fortran/src/H5f90global.F90
@@ -855,7 +855,7 @@ CONTAINS
END SUBROUTINE H5_Fortran_string_f2c
-! Copy Fortran string to C charater array, assuming the C array is one-char
+! Copy Fortran string to C character array, assuming the C array is one-char
! longer for the terminating null char.
! fstring : the Fortran input string
! cstring : the C output string (with memory already allocated)
@@ -883,7 +883,7 @@ CONTAINS
!!$ cstring(j) = C_NULL_CHAR
!!$end subroutine MPIR_Fortran_string_f2c
!!$
-!!$! Copy C charater array to Fortran string
+!!$! Copy C character array to Fortran string
!!$subroutine MPIR_Fortran_string_c2f(cstring, fstring)
!!$ implicit none
!!$ character(kind=c_char), intent(in) :: cstring(:)
diff --git a/fortran/src/H5f90proto.h b/fortran/src/H5f90proto.h
index fbbf7c2..4aa33f6 100644
--- a/fortran/src/H5f90proto.h
+++ b/fortran/src/H5f90proto.h
@@ -540,7 +540,7 @@ H5_FCDLL int_f h5pget_dxpl_mpio_c(hid_t_f *prp_id, int_f *data_xfer_mode);
H5_FCDLL int_f h5pset_dxpl_mpio_c(hid_t_f *prp_id, int_f *data_xfer_mode);
#endif
/*
- * Functions frome H5Rf.c
+ * Functions from H5Rf.c
*/
H5_FCDLL int_f h5rcreate_region_c(int_f *ref, hid_t_f *loc_id, _fcd name, int_f *namelen, hid_t_f *space_id);
H5_FCDLL int_f h5rcreate_ptr_c(void *ref, hid_t_f *loc_id, _fcd name, int_f *namelen, int_f *ref_type,
diff --git a/fortran/src/h5fc.in b/fortran/src/h5fc.in
index a56d38d..661fde5 100644
--- a/fortran/src/h5fc.in
+++ b/fortran/src/h5fc.in
@@ -38,7 +38,7 @@ HL="@HL@"
## $FLINKER $FCFLAGS $H5BLD_FCFLAGS $F9XSUFFIXFLAG $LDFLAGS $LIBS ##
## $fmodules $link_objs $link_args $shared_link ##
## ##
-## These settings can be overriden by setting HDF5_FCFLAGS, ##
+## These settings can be overridden by setting HDF5_FCFLAGS, ##
## HDF5_LDFLAGS, or HDF5_LIBS in the environment. ##
## ##
############################################################################
@@ -136,7 +136,7 @@ usage() {
echo " shared libraries]"
echo " "
echo " You can also add or change paths and flags to the compile line using"
- echo " the following environment varibles or by assigning them to their counterparts"
+ echo " the following environment variables or by assigning them to their counterparts"
echo " in the 'Things You Can Modify to Override...'" section of $prog_name
echo " "
echo " Variable Current value to be replaced"
@@ -298,7 +298,7 @@ fi
if test "x$do_link" = "xyes"; then
shared_link=""
-# conditionnaly link with the hl library
+# conditionally link with the hl library
if test "X$HL" = "Xhl"; then
libraries=" $libraries -lhdf5hl_fortran -lhdf5_hl -lhdf5_fortran -lhdf5 "
else
diff --git a/fortran/test/H5_test_buildiface.F90 b/fortran/test/H5_test_buildiface.F90
index 94beab1..302fa38 100644
--- a/fortran/test/H5_test_buildiface.F90
+++ b/fortran/test/H5_test_buildiface.F90
@@ -14,7 +14,7 @@
!
! NOTES
! This program uses the Fortran 2008 intrinsic function STORAGE_SIZE or SIZEOF
-! depending on availablity.It generates code that makes use of
+! depending on availability.It generates code that makes use of
! STORAGE_SIZE/SIZEOF in H5fortran_detect.f90. STORAGE_SIZE is standard
! compliant and should always be chosen over SIZEOF.
!
diff --git a/fortran/test/fflush1.F90 b/fortran/test/fflush1.F90
index c192d93..5684f81 100644
--- a/fortran/test/fflush1.F90
+++ b/fortran/test/fflush1.F90
@@ -88,7 +88,7 @@
data_dims(2) = NY
!
- !Initialize FORTRAN predifined datatypes
+ !Initialize FORTRAN predefined datatypes
!
CALL h5open_f(error)
CALL check("h5open_f",error,total_error)
diff --git a/fortran/test/fflush2.F90 b/fortran/test/fflush2.F90
index dc3bc85..a6e16c1 100644
--- a/fortran/test/fflush2.F90
+++ b/fortran/test/fflush2.F90
@@ -81,7 +81,7 @@ PROGRAM FFLUSH2EXAMPLE
data_dims(2) = NY
!
- !Initialize FORTRAN predifined datatypes
+ !Initialize FORTRAN predefined datatypes
!
CALL h5open_f(error)
CALL check("h5open_f",error,total_error)
@@ -163,7 +163,7 @@ PROGRAM FFLUSH2EXAMPLE
CALL check("h5fclose_f",error,total_error)
!
- !Close FORTRAN predifined datatypes
+ !Close FORTRAN predefined datatypes
!
CALL h5_cleanup_f(filename, H5P_DEFAULT_F, error)
CALL h5close_f(error)
diff --git a/fortran/test/tH5A.F90 b/fortran/test/tH5A.F90
index d5ce9a2..2e76dad 100644
--- a/fortran/test/tH5A.F90
+++ b/fortran/test/tH5A.F90
@@ -81,7 +81,7 @@ CONTAINS
INTEGER(HID_T) :: atype5_id !Integer Attribute Datatype identifier
INTEGER(HSIZE_T), DIMENSION(1) :: adims = (/2/) ! Attribute dimension
INTEGER(HSIZE_T), DIMENSION(1) :: adims2 = (/1/) ! Attribute dimension
- INTEGER :: arank = 1 ! Attribure rank
+ INTEGER :: arank = 1 ! Attribute rank
INTEGER(SIZE_T) :: attrlen ! Length of the attribute string
INTEGER(HID_T) :: attr_space !Returned String Attribute Space identifier
diff --git a/fortran/test/tH5A_1_8.F90 b/fortran/test/tH5A_1_8.F90
index 4e02c58..b245b1c 100644
--- a/fortran/test/tH5A_1_8.F90
+++ b/fortran/test/tH5A_1_8.F90
@@ -64,7 +64,7 @@ SUBROUTINE attribute_test_1_8(cleanup, total_error)
INTEGER :: ret_total_error
! ********************
-! test_attr equivelent
+! test_attr equivalent
! ********************
! WRITE(*,*) "TESTING ATTRIBUTES"
@@ -777,7 +777,6 @@ SUBROUTINE test_attr_info_by_idx(new_format, fcpl, fapl, total_error)
INTEGER :: Input1
INTEGER(HSIZE_T) :: hzero = 0_HSIZE_T
INTEGER :: minusone = -1
- INTEGER(HSIZE_T) :: htmp
data_dims = 0
@@ -840,10 +839,6 @@ SUBROUTINE test_attr_info_by_idx(new_format, fcpl, fapl, total_error)
my_dataset = dset3
END SELECT
- ! Check for query on non-existant attribute
-
- n = 0
-
! -- CHECK PASSING AN INTEGER CONSTANT IN DIFFERENT FORMS --
! 1) call by passing an integer with the _hsize_t declaration
@@ -896,8 +891,8 @@ SUBROUTINE test_attr_info_by_idx(new_format, fcpl, fapl, total_error)
! Verify information for new attribute
!EP CALL attr_info_by_idx_check(my_dataset, attrname, INT(j,HSIZE_T), use_index(i), total_error )
- htmp = j
- CALL attr_info_by_idx_check(my_dataset, attrname, htmp, use_index(i), total_error )
+ n = INT(j, HSIZE_T)
+ CALL attr_info_by_idx_check(my_dataset, attrname, n, use_index(i), total_error )
!CHECK(ret, FAIL, "attr_info_by_idx_check");
ENDDO
@@ -1111,7 +1106,7 @@ SUBROUTINE test_attr_shared_rename( fcpl, fapl, total_error)
INTEGER :: test_shared
INTEGER(HSIZE_T), DIMENSION(1) :: adims2 = (/1/) ! Attribute dimension
- INTEGER :: arank = 1 ! Attribure rank
+ INTEGER :: arank = 1 ! Attribute rank
! Initialize "big" attribute data
@@ -1497,7 +1492,7 @@ SUBROUTINE test_attr_delete_by_idx(new_format, fcpl, fapl, total_error)
END SELECT
- ! Check for deleting non-existant attribute
+ ! Check for deleting non-existent attribute
!EP CALL H5Adelete_by_idx_f(my_dataset, '.', idx_type, order, 0_HSIZE_T,error, lapl_id=H5P_DEFAULT_F)
CALL H5Adelete_by_idx_f(my_dataset, '.', idx_type, order, hzero,error, lapl_id=H5P_DEFAULT_F)
CALL verify("H5Adelete_by_idx_f",error,minusone,total_error)
@@ -1773,7 +1768,7 @@ SUBROUTINE test_attr_shared_delete(fcpl, fapl, total_error)
INTEGER :: test_shared
INTEGER(HSIZE_T), DIMENSION(1) :: adims2 = (/1/) ! Attribute dimension
- INTEGER :: arank = 1 ! Attribure rank
+ INTEGER :: arank = 1 ! Attribute rank
! Output message about test being performed
diff --git a/fortran/test/tH5D.F90 b/fortran/test/tH5D.F90
index 006d62a..eb9e831 100644
--- a/fortran/test/tH5D.F90
+++ b/fortran/test/tH5D.F90
@@ -187,7 +187,7 @@ CONTAINS
DO i = 1, 4
DO j = 1, 6
IF (data_out(i,j) .NE. dset_data(i, j)) THEN
- WRITE(*, *) "dataset test error occured"
+ WRITE(*, *) "dataset test error occurred"
WRITE(*,*) "data read is not the same as the data written"
END IF
END DO
@@ -196,7 +196,7 @@ CONTAINS
! Check if no change to null_dset_data
!
IF (null_dset_data .NE. 1) THEN
- WRITE(*, *) "null dataset test error occured"
+ WRITE(*, *) "null dataset test error occurred"
END IF
!
! End access to the dataset and release resources used by it.
@@ -258,7 +258,7 @@ CONTAINS
INTEGER(HID_T) :: dset_id ! Dataset identifier
INTEGER(HID_T) :: dataspace ! Dataspace identifier
INTEGER(HID_T) :: memspace ! memory Dataspace identifier
- INTEGER(HID_T) :: crp_list ! dataset creatation property identifier
+ INTEGER(HID_T) :: crp_list ! dataset creation property identifier
!
!dataset dimensions at creation time
@@ -313,7 +313,7 @@ CONTAINS
END DO
!
- !Initialize FORTRAN predifined datatypes
+ !Initialize FORTRAN predefined datatypes
!
! CALL h5init_types_f(error)
! CALL check("h5init_types_f",error,total_error)
@@ -347,7 +347,7 @@ CONTAINS
CALL check("h5pset_chunk_f",error,total_error)
!
- !Create a dataset with 3X3 dimensions using cparms creation propertie .
+ !Create a dataset with 3X3 dimensions using cparms creation properties .
!
CALL h5dcreate_f(file_id, dsetname, H5T_NATIVE_INTEGER, dataspace, dset_id, error, crp_list )
CALL check("h5dcreate_f",error,total_error)
@@ -426,7 +426,7 @@ CONTAINS
CALL h5sget_simple_extent_ndims_f(dataspace, rankr, error)
CALL check("h5sget_simple_extent_ndims_f",error,total_error)
IF (rankr .NE. RANK) THEN
- WRITE(*,*) "dataset rank error occured"
+ WRITE(*,*) "dataset rank error occurred"
STOP
END IF
@@ -436,7 +436,7 @@ CONTAINS
CALL h5sget_simple_extent_dims_f(dataspace, dimsr, maxdimsr, error)
CALL check("h5sget_simple_extent_dims_f",error,total_error)
IF ((dimsr(1) .NE. dims1(1)) .OR. (dimsr(2) .NE. dims1(2))) THEN
- WRITE(*,*) "dataset dimensions error occured"
+ WRITE(*,*) "dataset dimensions error occurred"
STOP
END IF
@@ -466,7 +466,7 @@ CONTAINS
DO ih = 1, dims1(1)
DO jh = 1, dims1(2)
IF (data_out(ih,jh) .NE. data_in(ih, jh)) THEN
- WRITE(*, *) "extend dataset test error occured"
+ WRITE(*, *) "extend dataset test error occurred"
WRITE(*, *) "read value is not the same as the written values"
END IF
END DO
diff --git a/fortran/test/tH5F.F90 b/fortran/test/tH5F.F90
index 06dc6de..3affed0 100644
--- a/fortran/test/tH5F.F90
+++ b/fortran/test/tH5F.F90
@@ -554,7 +554,7 @@ CONTAINS
do i = 1, NX
do j = 1, NY
IF (data_out(i,j) .NE. dset_data(i, j)) THEN
- write(*, *) "reopen test error occured"
+ write(*, *) "reopen test error occurred"
END IF
end do
end do
diff --git a/fortran/test/tH5G_1_8.F90 b/fortran/test/tH5G_1_8.F90
index 534a2d4..222ba9d 100644
--- a/fortran/test/tH5G_1_8.F90
+++ b/fortran/test/tH5G_1_8.F90
@@ -654,7 +654,7 @@ SUBROUTINE group_info(cleanup, fapl, total_error)
INTEGER(HID_T) :: file, scalar, grp, d1
CHARACTER(LEN=12), PARAMETER :: filename ='TestLinks.h5'
INTEGER(HSIZE_T), DIMENSION(1) :: adims2 = (/1/) ! Attribute dimension
- INTEGER :: arank = 1 ! Attribure rank
+ INTEGER :: arank = 1 ! Attribute rank
INTEGER :: error
INTEGER :: cset ! Indicates the character set used for the link’s name.
diff --git a/fortran/test/tH5MISC_1_8.F90 b/fortran/test/tH5MISC_1_8.F90
index c294f21..e35ef1f 100644
--- a/fortran/test/tH5MISC_1_8.F90
+++ b/fortran/test/tH5MISC_1_8.F90
@@ -160,7 +160,7 @@ SUBROUTINE test_genprop_basic_class(total_error)
CALL verify("H5Pequal_f", flag, .TRUE., total_error)
- ! Make certain false postives aren't being returned
+ ! Make certain false positives aren't being returned
CALL H5Pequal_f(cid2, H5P_FILE_CREATE_F, flag, error)
CALL check("H5Pequal_f", error, total_error)
CALL verify("H5Pequal_f", flag, .FALSE., total_error)
diff --git a/fortran/test/tH5P.F90 b/fortran/test/tH5P.F90
index 1d1208c..75e4e72 100644
--- a/fortran/test/tH5P.F90
+++ b/fortran/test/tH5P.F90
@@ -349,8 +349,8 @@ SUBROUTINE multi_file_test(cleanup, total_error)
DO i = 1, 4
DO j = 1, 6
IF (data_out(i,j) .NE. dset_data(i, j)) THEN
- WRITE(*, *) "dataset test error occured"
- WRITE(*,*) "data read is not the same as the data writen"
+ WRITE(*, *) "dataset test error occurred"
+ WRITE(*,*) "data read is not the same as the data written"
END IF
END DO
END DO
@@ -575,7 +575,7 @@ SUBROUTINE test_chunk_cache(cleanup, total_error)
CALL H5Pclose_f(dapl2,error)
CALL check("H5Pclose_f", error, total_error)
- ! Similary, test use of H5Dcreate2 with H5P_DEFAULT
+ ! Similarly, test use of H5Dcreate2 with H5P_DEFAULT
CALL H5Dclose_f(dsid, error)
CALL check("H5Dclose_f", error, total_error)
diff --git a/fortran/test/tH5P_F03.F90 b/fortran/test/tH5P_F03.F90
index 245a588..43dd1c0 100644
--- a/fortran/test/tH5P_F03.F90
+++ b/fortran/test/tH5P_F03.F90
@@ -1013,28 +1013,28 @@ SUBROUTINE test_vds(total_error)
CALL check("H5Pget_virtual_filename_f", error, total_error)
IF(nsize.NE.LEN(SRC_FILE_LEN_EXACT))THEN
- PRINT*,"virtual filenname size is incorrect"
+ PRINT*,"virtual filename size is incorrect"
total_error = total_error + 1
ENDIF
! check passing a buffer that is very small
CALL H5Pget_virtual_filename_f(dcpl, INT(i-1, size_t), SRC_FILE_LEN_TINY, error)
CALL check("H5Pget_virtual_filename_f", error, total_error)
IF(SRC_FILE_LEN_TINY.NE.SRC_FILE(i)(1:LEN(SRC_FILE_LEN_TINY)))THEN
- PRINT*,"virtual filenname returned is incorrect"
+ PRINT*,"virtual filename returned is incorrect"
total_error = total_error + 1
ENDIF
! check passing a buffer that small by one
CALL H5Pget_virtual_filename_f(dcpl, INT(i-1, size_t), SRC_FILE_LEN_SMALL, error)
CALL check("H5Pget_virtual_filename_f", error, total_error)
IF(SRC_FILE_LEN_SMALL.NE.SRC_FILE(i)(1:LEN(SRC_FILE_LEN_SMALL)))THEN
- PRINT*,"virtual filenname returned is incorrect"
+ PRINT*,"virtual filename returned is incorrect"
total_error = total_error + 1
ENDIF
! check passing a buffer that is exact
CALL H5Pget_virtual_filename_f(dcpl, INT(i-1, size_t), SRC_FILE_LEN_EXACT, error)
CALL check("H5Pget_virtual_filename_f", error, total_error)
IF(SRC_FILE_LEN_EXACT.NE.SRC_FILE(i)(1:LEN(SRC_FILE_LEN_EXACT)))THEN
- PRINT*,"virtual filenname returned is incorrect"
+ PRINT*,"virtual filename returned is incorrect"
total_error = total_error + 1
ENDIF
! check passing a buffer that bigger by one
@@ -1042,7 +1042,7 @@ SUBROUTINE test_vds(total_error)
CALL check("H5Pget_virtual_filename_f", error, total_error)
IF(SRC_FILE_LEN_LARGE(1:LEN(SRC_FILE_LEN_EXACT)).NE.SRC_FILE(i)(1:LEN(SRC_FILE_LEN_EXACT)).AND. &
SRC_FILE_LEN_LARGE(LEN(SRC_FILE_LEN_EXACT):).NE.'')THEN
- PRINT*,"virtual filenname returned is incorrect"
+ PRINT*,"virtual filename returned is incorrect"
total_error = total_error + 1
ENDIF
! check passing a buffer that is very big
@@ -1050,7 +1050,7 @@ SUBROUTINE test_vds(total_error)
CALL check("H5Pget_virtual_filename_f", error, total_error)
IF(SRC_FILE_LEN_HUGE(1:LEN(SRC_FILE_LEN_EXACT)).NE.SRC_FILE(i)(1:LEN(SRC_FILE_LEN_EXACT)).AND. &
SRC_FILE_LEN_HUGE(LEN(SRC_FILE_LEN_EXACT):).NE.'')THEN
- PRINT*,"virtual filenname returned is incorrect"
+ PRINT*,"virtual filename returned is incorrect"
total_error = total_error + 1
ENDIF
! Get source dataset name
diff --git a/fortran/test/tH5R.F90 b/fortran/test/tH5R.F90
index 5144d83..4e62b20 100644
--- a/fortran/test/tH5R.F90
+++ b/fortran/test/tH5R.F90
@@ -142,7 +142,7 @@ SUBROUTINE refobjtest(cleanup, total_error)
CALL check("h5gclose_f",error,total_error)
!
- ! Craete references to two groups, integer dataset and shared datatype
+ ! Create references to two groups, integer dataset and shared datatype
! and write it to the dataset in the file
!
CALL h5rcreate_f(file_id, groupname1, ref(1), error)
diff --git a/fortran/test/tH5S.F90 b/fortran/test/tH5S.F90
index 08f0b59..302db51 100644
--- a/fortran/test/tH5S.F90
+++ b/fortran/test/tH5S.F90
@@ -151,7 +151,7 @@ CONTAINS
CALL h5sget_simple_extent_dims_f(space2_id, dimsout, maxdimsout, error)
CALL check("h5sget_simple_extent_dims_f", error, total_error)
IF ((dimsout(1) .NE. dims1(1)) .OR. (dimsout(2) .NE. dims1(2)) ) THEN
- write(*,*)"error occured, copied dims not same"
+ write(*,*)"error occurred, copied dims not same"
END IF
!
@@ -159,14 +159,14 @@ CONTAINS
!
CALL h5sget_simple_extent_ndims_f(space2_id, rank2, error)
CALL check("h5sget_simple_extent_ndims_f", error, total_error)
- IF (rank2 .NE. rank1) write(*,*)"error occured, copied ranks not same"
+ IF (rank2 .NE. rank1) write(*,*)"error occurred, copied ranks not same"
!
!get the copied space's number of elements.
!
CALL h5sget_simple_extent_npoints_f(space2_id, npoints, error)
CALL check("h5sget_simple_extent_npoints_f", error, total_error)
- IF (npoints .NE. 24) write(*,*)"error occured, number of elements not correct"
+ IF (npoints .NE. 24) write(*,*)"error occurred, number of elements not correct"
!
@@ -194,7 +194,7 @@ CONTAINS
CALL h5sget_simple_extent_dims_f(space2_id, dimsout, maxdimsout, error)
CALL check("h5sget_simple_extent_dims_f", error, total_error)
IF ((dimsout(1) .NE. dims2(1)) .OR. (dimsout(2) .NE. dims2(2)) ) THEN
- write(*,*)"error occured, copied dims not same"
+ write(*,*)"error occurred, copied dims not same"
END IF
!
@@ -235,8 +235,8 @@ CONTAINS
do i = 1, 4
do j = 1, 6
IF (data1_out(i,j) .NE. data1_in(i, j)) THEN
- write(*, *) "dataset test error occured"
- write(*,*) "data read is not the same as the data writen"
+ write(*, *) "dataset test error occurred"
+ write(*,*) "data read is not the same as the data written"
END IF
end do
end do
@@ -256,8 +256,8 @@ CONTAINS
do i = 1, 6
do j = 1, 6
IF (data2_out(i,j) .NE. data2_in(i, j)) THEN
- write(*, *) "dataset test error occured"
- write(*,*) "data read is not the same as the data writen"
+ write(*, *) "dataset test error occurred"
+ write(*,*) "data read is not the same as the data written"
END IF
end do
end do
diff --git a/fortran/test/tH5Sselect.F90 b/fortran/test/tH5Sselect.F90
index 178ac32..41cd16f 100644
--- a/fortran/test/tH5Sselect.F90
+++ b/fortran/test/tH5Sselect.F90
@@ -149,7 +149,7 @@ CONTAINS
!
!
- !Initialize FORTRAN predifined datatypes
+ !Initialize FORTRAN predefined datatypes
!
! CALL h5init_types_f(error)
! CALL check("h5init_types_f", error, total_error)
@@ -430,7 +430,7 @@ CONTAINS
end do
!
- !Initialize FORTRAN predifined datatypes
+ !Initialize FORTRAN predefined datatypes
!
! CALL h5init_types_f(error)
! CALL check("h5init_types_f", error, total_error)
@@ -912,7 +912,7 @@ CONTAINS
!
CALL h5sget_select_hyper_nblocks_f(dataspace, num_blocks, error)
CALL check("h5sget_select_hyper_nblocks_f", error, total_error)
- IF (num_blocks .NE. 4) write (*,*) "error occured with num_blocks"
+ IF (num_blocks .NE. 4) write (*,*) "error occurred with num_blocks"
!write(*,*) num_blocks
!result of num_blocks is 4
@@ -945,11 +945,11 @@ CONTAINS
CALL h5sget_select_bounds_f(dataspace, startout, endout, error)
CALL check("h5sget_select_bounds_f", error, total_error)
IF ( (startout(1) .ne. 1) .or. (startout(2) .ne. 1) ) THEN
- write(*,*) "error occured to select_bounds's start position"
+ write(*,*) "error occurred to select_bounds's start position"
END IF
IF ( (endout(1) .ne. 5) .or. (endout(2) .ne. 5) ) THEN
- write(*,*) "error occured to select_bounds's end position"
+ write(*,*) "error occurred to select_bounds's end position"
END IF
!write(*,*) (startout(i), i = 1, RANK)
!result of startout is 0, 0
@@ -978,7 +978,7 @@ CONTAINS
!
CALL h5sget_select_elem_npoints_f(dataspace, num_points, error)
CALL check("h5sget_select_elem_npoints_f", error, total_error)
- IF (num_points .NE. 10) write(*,*) "error occured with num_points"
+ IF (num_points .NE. 10) write(*,*) "error occurred with num_points"
!write(*,*) num_points
! result of num_points is 10
diff --git a/fortran/test/tH5T.F90 b/fortran/test/tH5T.F90
index 8b10136..d24ac89 100644
--- a/fortran/test/tH5T.F90
+++ b/fortran/test/tH5T.F90
@@ -72,7 +72,7 @@ CONTAINS
INTEGER(HID_T) :: dt4_id ! Memory datatype identifier (for real field)
INTEGER(HID_T) :: dt5_id ! Memory datatype identifier
INTEGER(HID_T) :: membtype_id ! Datatype identifier
- INTEGER(HID_T) :: plist_id ! Dataset trasfer property
+ INTEGER(HID_T) :: plist_id ! Dataset transfer property
INTEGER(HSIZE_T), DIMENSION(1) :: dims = (/dimsize/) ! Dataset dimensions
@@ -98,7 +98,7 @@ CONTAINS
INTEGER :: class ! Datatype class
INTEGER :: num_members ! Number of members in the compound datatype
CHARACTER(LEN=256) :: member_name
- INTEGER :: len ! Lenght of the name of the compound datatype member
+ INTEGER :: len ! Length of the name of the compound datatype member
INTEGER :: member_index ! index of the field
INTEGER(HSIZE_T), DIMENSION(3) :: array_dims=(/2,3,4/)
INTEGER :: array_dims_range = 3
@@ -219,9 +219,9 @@ CONTAINS
CALL h5tcreate_f(H5T_COMPOUND_F, type_size, dtype_id, error)
CALL check("h5tcreate_f", error, total_error)
!
- ! Insert memebers
+ ! Insert members
!
- ! CHARACTER*2 memeber
+ ! CHARACTER*2 member
!
offset = 0
CALL h5tinsert_f(dtype_id, "char_field", offset, dt5_id, error)
@@ -229,13 +229,13 @@ CONTAINS
!
! INTEGER member
!
- offset = offset + type_sizec ! Offset of the second memeber is 2
+ offset = offset + type_sizec ! Offset of the second member is 2
CALL h5tinsert_f(dtype_id, "integer_field", offset, H5T_NATIVE_INTEGER, error)
CALL check("h5tinsert_f", error, total_error)
!
! DOUBLE PRECISION member
!
- offset = offset + type_sizei ! Offset of the third memeber is 6
+ offset = offset + type_sizei ! Offset of the third member is 6
CALL h5tinsert_f(dtype_id, "double_field", offset, H5T_NATIVE_DOUBLE, error)
CALL check("h5tinsert_f", error, total_error)
!
@@ -416,7 +416,7 @@ CONTAINS
CALL h5tequal_f(membtype_id, H5T_NATIVE_INTEGER, flag, error)
CALL check("h5tequal_f", error, total_error)
if(.not. flag) then
- write(*,*) "Wrong member type returned for integer memebr"
+ write(*,*) "Wrong member type returned for integer member"
total_error = total_error + 1
endif
CALL h5tget_member_class_f(dtype_id, i-1, class, error)
@@ -435,7 +435,7 @@ CONTAINS
CALL h5tequal_f(membtype_id, H5T_NATIVE_DOUBLE, flag, error)
CALL check("h5tequal_f", error, total_error)
if(.not. flag) then
- write(*,*) "Wrong member type returned for double precision memebr"
+ write(*,*) "Wrong member type returned for double precision member"
total_error = total_error + 1
endif
CALL h5tget_member_class_f(dtype_id, i-1, class, error)
@@ -454,7 +454,7 @@ CONTAINS
CALL h5tequal_f(membtype_id, H5T_NATIVE_REAL, flag, error)
CALL check("h5tequal_f", error, total_error)
if(.not. flag) then
- write(*,*) "Wrong member type returned for real memebr"
+ write(*,*) "Wrong member type returned for real member"
total_error = total_error + 1
endif
CALL h5tget_member_class_f(dtype_id, i-1, class, error)
diff --git a/fortran/test/tH5VL.F90 b/fortran/test/tH5VL.F90
index 2269fec..eb0470c 100644
--- a/fortran/test/tH5VL.F90
+++ b/fortran/test/tH5VL.F90
@@ -147,7 +147,7 @@ CONTAINS
CALL check("h5dvlen_get_max_len_f", error, total_error)
if(max_len .ne. data_dims(1)) then
total_error = total_error + 1
- write(*,*) "Wrong number of elemets returned by h5dvlen_get_max_len_f"
+ write(*,*) "Wrong number of elements returned by h5dvlen_get_max_len_f"
endif
!
! Read the dataset.
@@ -317,7 +317,7 @@ CONTAINS
CALL check("h5dvlen_get_max_len_f", error, total_error)
if(max_len .ne. data_dims(1)) then
total_error = total_error + 1
- write(*,*) "Wrong number of elemets returned by h5dvlen_get_max_len_f"
+ write(*,*) "Wrong number of elements returned by h5dvlen_get_max_len_f"
endif
!
! Read the dataset.
diff --git a/fortran/test/tH5Z.F90 b/fortran/test/tH5Z.F90
index 8c39fea..799067a 100644
--- a/fortran/test/tH5Z.F90
+++ b/fortran/test/tH5Z.F90
@@ -374,7 +374,7 @@ CONTAINS
do i = 1, N
do j = 1, M
IF (data_out(i,j) .NE. dset_data(i, j)) THEN
- write(*, *) "dataset test error occured"
+ write(*, *) "dataset test error occurred"
write(*,*) "data read is not the same as the data written"
num_errors = num_errors + 1
IF (num_errors .GE. 512) THEN
diff --git a/hl/c++/test/ptableTest.cpp b/hl/c++/test/ptableTest.cpp
index ab49303..9db56e0 100644
--- a/hl/c++/test/ptableTest.cpp
+++ b/hl/c++/test/ptableTest.cpp
@@ -223,7 +223,7 @@ TestGetNext()
for (record = 1; record < 6; record++)
wrapper.AppendPacket(&record);
- /* Ensure that we can interate through the records and get the right ones */
+ /* Ensure that we can iterate through the records and get the right ones */
for (i = 1; i < 6; i++) {
wrapper.GetNextPacket(&record);
if (record != i)
@@ -237,7 +237,7 @@ TestGetNext()
if (error < 0)
goto error;
- /* Ensure that we can interate through the records and get the right ones */
+ /* Ensure that we can iterate through the records and get the right ones */
for (i = 1; i < 6; i++) {
error = wrapper.GetNextPacket(&record);
if (record != i || error < 0)
diff --git a/hl/examples/ex_image1.c b/hl/examples/ex_image1.c
index ead1715..1294ab3 100644
--- a/hl/examples/ex_image1.c
+++ b/hl/examples/ex_image1.c
@@ -56,10 +56,10 @@ main(void)
H5IMmake_image_8bit(file_id, "image1", (hsize_t)WIDTH, (hsize_t)HEIGHT, buf);
/* make a palette */
- H5IMmake_palette(file_id, "pallete", pal_dims, pal);
+ H5IMmake_palette(file_id, "palette", pal_dims, pal);
/* attach the palette to the image */
- H5IMlink_palette(file_id, "image1", "pallete");
+ H5IMlink_palette(file_id, "image1", "palette");
/* close the file. */
H5Fclose(file_id);
diff --git a/hl/fortran/src/H5DSff.F90 b/hl/fortran/src/H5DSff.F90
index b241e9a..bbd9918 100644
--- a/hl/fortran/src/H5DSff.F90
+++ b/hl/fortran/src/H5DSff.F90
@@ -216,7 +216,7 @@ CONTAINS
END SUBROUTINE H5DSis_attached_f
!
-! H5DSiterate_scales: Impliment in F2003
+! H5DSiterate_scales: Implement in F2003
!
!-------------------------------------------------------------------------
diff --git a/hl/fortran/src/H5HL_buildiface.F90 b/hl/fortran/src/H5HL_buildiface.F90
index 4f34575..ca66ebf 100644
--- a/hl/fortran/src/H5HL_buildiface.F90
+++ b/hl/fortran/src/H5HL_buildiface.F90
@@ -14,7 +14,7 @@
!
! NOTES
! This program uses the Fortran 2008 intrinsic function STORAGE_SIZE or SIZEOF
-! depending on availablity.It generates code that makes use of
+! depending on availability.It generates code that makes use of
! STORAGE_SIZE/SIZEOF in H5fortran_detect.f90. STORAGE_SIZE is standard
! compliant and should always be chosen over SIZEOF.
!
diff --git a/hl/fortran/src/H5LTff.F90 b/hl/fortran/src/H5LTff.F90
index ba4b770..a4ab247 100644
--- a/hl/fortran/src/H5LTff.F90
+++ b/hl/fortran/src/H5LTff.F90
@@ -88,7 +88,7 @@ MODULE H5LT_CONST
TYPE(C_PTR), VALUE :: buf ! data buffer
CHARACTER(KIND=C_CHAR), DIMENSION(*), INTENT(in) :: dtype ! flag indicating the datatype of the
! the buffer:
- ! R=Real, D=DOUBLE, I=Interger, C=Character
+ ! R=Real, D=DOUBLE, I=Integer, C=Character
INTEGER(size_t) :: SizeOf_buf ! Sizeof the buf datatype
END FUNCTION h5ltset_attribute_c
END INTERFACE
@@ -107,7 +107,7 @@ MODULE H5LT_CONST
TYPE(C_PTR), VALUE :: buf ! data buffer
CHARACTER(KIND=C_CHAR), DIMENSION(*), INTENT(in) :: dtype ! flag indicating the datatype of the
! the buffer:
- ! R=Real, D=DOUBLE, I=Interger
+ ! R=Real, D=DOUBLE, I=Integer
INTEGER(size_t), INTENT(in) :: SizeOf_buf ! Sizeof the buf data type
END FUNCTION h5ltget_attribute_c
END INTERFACE
diff --git a/hl/fortran/src/H5TBfc.c b/hl/fortran/src/H5TBfc.c
index 23e3377..91c420b 100644
--- a/hl/fortran/src/H5TBfc.c
+++ b/hl/fortran/src/H5TBfc.c
@@ -83,7 +83,7 @@ h5tbmake_table_c(size_t_f *namelen1, _fcd name1, hid_t_f *loc_id, size_t_f *name
(tmp = (char *)HD5f2cstring(field_names, (size_t) * (max_char_size_field_names) * (size_t)num_elem)))
HGOTO_DONE(FAIL)
/*
- * move data from temorary buffer
+ * move data from temporary buffer
*/
tmp_p = tmp;
for (i = 0; i < num_elem; i++) {
@@ -188,7 +188,7 @@ h5tbmake_table_ptr_c(size_t_f *namelen1, _fcd name1, hid_t_f *loc_id, size_t_f *
(tmp = (char *)HD5f2cstring(field_names, (size_t) * (max_char_size_field_names) * (size_t)num_elem)))
HGOTO_DONE(FAIL)
/*
- * move data from temorary buffer
+ * move data from temporary buffer
*/
tmp_p = tmp;
for (i = 0; i < num_elem; i++) {
diff --git a/hl/fortran/src/H5TBff.F90 b/hl/fortran/src/H5TBff.F90
index 35a88b8..82f34a8 100644
--- a/hl/fortran/src/H5TBff.F90
+++ b/hl/fortran/src/H5TBff.F90
@@ -961,7 +961,7 @@ CONTAINS
INTEGER(size_t), DIMENSION(1:nfields), INTENT(inout) :: field_offsets ! field offsets
INTEGER(size_t), INTENT(inout):: type_size ! type size
INTEGER(size_t) :: namelen ! name length
- INTEGER(size_t) :: maxlen ! maxiumum length of input field names
+ INTEGER(size_t) :: maxlen ! maximum length of input field names
INTEGER(size_t), DIMENSION(1:nfields) :: namelen2 ! name lengths
INTEGER(size_t) :: c_maxlen_out ! maximum character length of a field array element
END FUNCTION h5tbget_field_info_c
diff --git a/hl/fortran/test/tsttable.F90 b/hl/fortran/test/tsttable.F90
index a77068c..e7bd88e 100644
--- a/hl/fortran/test/tsttable.F90
+++ b/hl/fortran/test/tsttable.F90
@@ -172,11 +172,11 @@ SUBROUTINE test_table1()
!
offset = 0
field_offset(1) = offset
- offset = offset + type_sizec ! Offset of the second memeber is 2
+ offset = offset + type_sizec ! Offset of the second member is 2
field_offset(2) = offset
- offset = offset + type_sizei ! Offset of the second memeber is 6
+ offset = offset + type_sizei ! Offset of the second member is 6
field_offset(3) = offset
- offset = offset + type_sized ! Offset of the second memeber is 14
+ offset = offset + type_sized ! Offset of the second member is 14
field_offset(4) = offset
!-------------------------------------------------------------------------
@@ -571,7 +571,7 @@ SUBROUTINE test_table2()
INTEGER(HSIZE_T), PARAMETER :: nrecords = 8 ! nrecords
CHARACTER(len=8), PARAMETER :: filename = "f2tab.h5" ! File name
- CHARACTER(LEN=5), PARAMETER :: table_name = "tabel" ! table name
+ CHARACTER(LEN=5), PARAMETER :: table_name = "table" ! table name
CHARACTER(LEN=10), PARAMETER :: table_name_fill = "tabel_fill" ! table name
! Define field information
diff --git a/hl/src/H5DO.c b/hl/src/H5DO.c
index 3bdb4e2..1703669 100644
--- a/hl/src/H5DO.c
+++ b/hl/src/H5DO.c
@@ -182,7 +182,7 @@ H5DOappend(hid_t dset_id, hid_t dxpl_id, unsigned axis, size_t extension, hid_t
if (FAIL == H5Sselect_hyperslab(new_space_id, H5S_SELECT_SET, start, stride, count, block))
goto done;
- /* The # of elemnts in the new extended dataspace */
+ /* The # of elements in the new extended dataspace */
if ((snelmts = H5Sget_select_npoints(new_space_id)) < 0)
goto done;
nelmts = (hsize_t)snelmts;
diff --git a/hl/src/H5DS.c b/hl/src/H5DS.c
index c947d16..387419b 100644
--- a/hl/src/H5DS.c
+++ b/hl/src/H5DS.c
@@ -88,7 +88,7 @@ H5DSset_scale(hid_t dsid, const char *dimname)
return FAIL;
/*-------------------------------------------------------------------------
- * check if the dataset is a dataset wich has references to dimension scales
+ * check if the dataset is a dataset which has references to dimension scales
*-------------------------------------------------------------------------
*/
@@ -1122,7 +1122,7 @@ H5DSdetach_scale(hid_t did, hid_t dsid, unsigned int idx)
goto out;
} /* nelmts */
- /* Free refrences */
+ /* Free references */
if (is_new_ref) {
if (H5Treclaim(tid, sid, H5P_DEFAULT, ndsbuf) < 0)
goto out;
@@ -1189,7 +1189,7 @@ out:
dsbuf = NULL;
}
if (buf) {
- /* Failure occured before H5Treclaim was called;
+ /* Failure occurred before H5Treclaim was called;
free the pointers allocated when we read data in */
for (i = 0; i < rank; i++) {
if (buf[i].p)
@@ -1633,7 +1633,7 @@ H5DSiterate_scales(hid_t did, unsigned int dim, int *ds_idx, H5DS_iterate_t visi
if (H5DSwith_new_ref(did, &is_new_ref) < 0)
return FAIL;
- /* get the number of scales assotiated with this DIM */
+ /* get the number of scales associated with this DIM */
if ((nscales = H5DSget_num_scales(did, dim)) < 0)
return FAIL;
diff --git a/hl/src/H5LTpublic.h b/hl/src/H5LTpublic.h
index f19d353..a6c7c84 100644
--- a/hl/src/H5LTpublic.h
+++ b/hl/src/H5LTpublic.h
@@ -19,7 +19,7 @@
#define H5LT_FILE_IMAGE_DONT_COPY 0x0002 /* The HDF5 lib won't copy */
/* user supplied image buffer. The same image is open with the core driver. */
#define H5LT_FILE_IMAGE_DONT_RELEASE 0x0004 /* The HDF5 lib won't */
-/* deallocate user supplied image buffer. The user application is reponsible */
+/* deallocate user supplied image buffer. The user application is responsible */
/* for doing so. */
#define H5LT_FILE_IMAGE_ALL 0x0007
diff --git a/hl/test/test_ds.c b/hl/test/test_ds.c
index c66ef49..12b6840 100644
--- a/hl/test/test_ds.c
+++ b/hl/test/test_ds.c
@@ -3021,7 +3021,7 @@ test_simple(void)
}
/*-------------------------------------------------------------------------
- * dettach for DIM0
+ * detach for DIM0
*-------------------------------------------------------------------------
*/
@@ -3319,11 +3319,11 @@ test_simple(void)
goto out;
dim = 0;
- /* iterate trough the 1st dimension of "dset_a" and verify that its DS is valid */
+ /* iterate through the 1st dimension of "dset_a" and verify that its DS is valid */
if (H5DSiterate_scales(did, dim, NULL, verify_scale, NULL) < 0)
goto out;
- /* iterate trough the 2nd dimension of "dset_a" and verify that its DS is valid
+ /* iterate through the 2nd dimension of "dset_a" and verify that its DS is valid
start at DS index 2 */
dim = 1;
scale_idx = 2;
@@ -3344,11 +3344,11 @@ test_simple(void)
goto out;
dim = 0;
- /* iterate trough the 1st dimension of "dset_a" and read the DS */
+ /* iterate through the 1st dimension of "dset_a" and read the DS */
if (H5DSiterate_scales(did, dim, NULL, read_scale, s1_wbuf) < 0)
goto out;
- /* iterate trough the 2nd dimension of "dset_a" and read the DS
+ /* iterate through the 2nd dimension of "dset_a" and read the DS
start at DS index 2 */
dim = 1;
scale_idx = 2;
@@ -3381,9 +3381,9 @@ test_simple(void)
goto out;
{
int match_size; /* does this scale size matches the dataset DIM size */
- int idx = 0; /* scale index to start iterating, on return, index where iterator stoped */
+ int idx = 0; /* scale index to start iterating, on return, index where iterator stopped */
- /* iterate trough all the dimensions */
+ /* iterate through all the dimensions */
for (dim = 0; dim < (unsigned)rank; dim++) {
if ((match_size = H5DSiterate_scales(did, dim, &idx, match_dim_scale, NULL)) < 0)
goto out;
@@ -3475,9 +3475,9 @@ test_simple(void)
goto out;
{
int match_size; /* does this scale size matches the dataset DIM size */
- int idx; /* scale index to start iterating, on return, index where iterator stoped */
+ int idx; /* scale index to start iterating, on return, index where iterator stopped */
- /* iterate trough all the dimensions */
+ /* iterate through all the dimensions */
for (dim = 0; dim < (unsigned)rank; dim++) {
/* always start at 1st scale */
idx = 0;
@@ -4025,7 +4025,7 @@ test_errors(void)
PASSED();
/*-------------------------------------------------------------------------
- * try to attach a scale to an image, pallete or table
+ * try to attach a scale to an image, palette or table
*-------------------------------------------------------------------------
*/
@@ -4036,7 +4036,7 @@ test_errors(void)
goto out;
/* make a palette */
- if (H5IMmake_palette(fid, "pallete", pal_dims, NULL) < 0)
+ if (H5IMmake_palette(fid, "palette", pal_dims, NULL) < 0)
goto out;
/* open the previous written "ds_b" */
@@ -4248,7 +4248,7 @@ test_iterators(void)
if ((did = H5Dopen2(fid, "dset_a", H5P_DEFAULT)) < 0)
goto out;
- /* try to iterate trough the 1st dimension of "dset_a", return error */
+ /* try to iterate through the 1st dimension of "dset_a", return error */
if (H5DSiterate_scales(did, 0, NULL, verify_scale, NULL) < 0)
goto out;
@@ -4269,7 +4269,7 @@ test_iterators(void)
if ((did = H5Dopen2(fid, "dset_a", H5P_DEFAULT)) < 0)
goto out;
- /* try to iterate trough the 3rd dimension of "dset_a", return error */
+ /* try to iterate through the 3rd dimension of "dset_a", return error */
if (H5DSiterate_scales(did, 3, NULL, verify_scale, NULL) == SUCCEED)
goto out;
@@ -4306,7 +4306,7 @@ test_iterators(void)
goto out;
}
- /* iterate trough the 1st dimension of "dset_a" */
+ /* iterate through the 1st dimension of "dset_a" */
if (H5DSiterate_scales(did, 0, NULL, op_continue, NULL) < 0)
goto out;
diff --git a/hl/test/test_file_image.c b/hl/test/test_file_image.c
index bde8adc..6bb36e4 100644
--- a/hl/test/test_file_image.c
+++ b/hl/test/test_file_image.c
@@ -243,7 +243,7 @@ test_file_image(size_t open_images, size_t nflags, const unsigned *flags)
*/
HDmemset((uint8_t *)tmp_ptr + SUPER_STATUS_FLAGS_OFF_V0_V1, (int)0,
(size_t)SUPER_STATUS_FLAGS_SIZE_V0_V1);
- /* Does the comparision */
+ /* Does the comparison */
if (HDmemcmp(tmp_ptr, buf_ptr[i], (size_t)buf_size[i]) != 0)
FAIL_PUTS_ERROR("comparison of TMP vfd and user buffer failed");
/* Free the temporary buffer */
diff --git a/hl/test/test_h5do_compat.c b/hl/test/test_h5do_compat.c
index c554ae7..6f19d4b 100644
--- a/hl/test/test_h5do_compat.c
+++ b/hl/test/test_h5do_compat.c
@@ -31,7 +31,7 @@
*
* Purpose: Test the basic functionality of H5DOwrite_chunk
*
- * Return: Success: An identifer for the dataset used in the tests
+ * Return: Success: An identifier for the dataset used in the tests
* Failure: H5I_INVALID_HID
*
*-------------------------------------------------------------------------
diff --git a/hl/test/test_image.c b/hl/test/test_image.c
index 81340c1..2251a04 100644
--- a/hl/test/test_image.c
+++ b/hl/test/test_image.c
@@ -219,7 +219,7 @@ test_simple(void)
*-------------------------------------------------------------------------
*/
- HL_TESTING2("pallete functions");
+ HL_TESTING2("palette functions");
if (H5IMget_npalettes(fid, IMAGE1_NAME, &npals) < 0)
goto out;
@@ -422,7 +422,7 @@ test_data(void)
/*-------------------------------------------------------------------------
* palette #4. blue-red
- * make a palette whith blue to red colors
+ * make a palette with blue to red colors
*-------------------------------------------------------------------------
*/
for (i = 0, n = 0; i < 256 * 3; i += 3, n++) {
diff --git a/hl/test/test_ld.c b/hl/test/test_ld.c
index dc5755c..150d04c 100644
--- a/hl/test/test_ld.c
+++ b/hl/test/test_ld.c
@@ -189,7 +189,7 @@ typedef struct set_t {
* This will fail on heiwa and amani when VALID_FIELDS1 is "field1,field3,field4"
* because of alignment problems:
* amani and heiwa - 8 byte alignment
- * jam - 4 byte alignemnt
+ * jam - 4 byte alignment
* This will need to be fixed in the library for H5Tget_native_type().
*/
/* VALID_FIELDS1 "field1,field2.a,field3,field4" */
@@ -213,8 +213,8 @@ typedef struct test_valid_fields2 {
#define TEST_BUF_SIZE 100
int * iibuf; /* buffer for storing retrieved elements */
int * ibuf; /* buffer for storing retrieved elements (integer) */
-set_t * cbuf; /* buffer for storing retrieved elemnets (compound) */
-set_t * ccbuf; /* buffer for storing retrieved elemnets (compound) */
+set_t * cbuf; /* buffer for storing retrieved elements (compound) */
+set_t * ccbuf; /* buffer for storing retrieved elements (compound) */
test_valid_fields1 *vbuf1; /* buffer for storing retrieved elements (FIELDS1) */
test_valid_fields2 *vbuf2; /* buffer for storing retrieved elements (FIELDS2) */
diff --git a/hl/test/test_packet_vlen.c b/hl/test/test_packet_vlen.c
index 6d6bf34..e704944 100644
--- a/hl/test/test_packet_vlen.c
+++ b/hl/test/test_packet_vlen.c
@@ -677,7 +677,7 @@ error: /* An error has occurred. Clean up and exit. */
*
* Description:
* - Added a fixed-length packet table to the file for variety
- * - Use the helper funtion verify_ptlengthtype to test H5PTis_varlen
+ * - Use the helper function verify_ptlengthtype to test H5PTis_varlen
* on each packet table.
*
* 2016/01/27 -BMR
diff --git a/hl/test/test_table.c b/hl/test/test_table.c
index 6199e27..90a930c 100644
--- a/hl/test/test_table.c
+++ b/hl/test/test_table.c
@@ -1464,7 +1464,7 @@ test_table(hid_t fid, int do_write)
HL_TESTING2("getting field info");
- /* alocate */
+ /* allocate */
names_out = (char **)HDmalloc(sizeof(char *) * (size_t)NFIELDS);
for (i = 0; i < NFIELDS; i++) {
names_out[i] = (char *)HDmalloc(sizeof(char) * 255);
diff --git a/hl/tools/gif2h5/decompress.c b/hl/tools/gif2h5/decompress.c
index 64d9643..ca33f86 100644
--- a/hl/tools/gif2h5/decompress.c
+++ b/hl/tools/gif2h5/decompress.c
@@ -178,7 +178,7 @@ Decompress(GIFIMAGEDESC *GifImageDesc, GIFHEAD *GifHead)
DataMask = (1 << ((GifHead->PackedField & 0x07) + 1)) - 1;
Raster = GifImageDesc->GIFImage;
- /* Check for image seperator */
+ /* Check for image separator */
/* Now read in values from the image descriptor */
IWidth = GifImageDesc->ImageWidth;
@@ -193,7 +193,7 @@ Decompress(GIFIMAGEDESC *GifImageDesc, GIFHEAD *GifHead)
*/
/*
- * Start reading the raster data. First we get the intial code size and
+ * Start reading the raster data. First we get the initial code size and
* compute decompressor constant values, based on this code size.
*/
diff --git a/hl/tools/gif2h5/gif.h b/hl/tools/gif2h5/gif.h
index 1a8cfe4..5ea8633 100644
--- a/hl/tools/gif2h5/gif.h
+++ b/hl/tools/gif2h5/gif.h
@@ -129,7 +129,7 @@ typedef struct _GifCommentExtension {
** in the HDF file.
** I have assumed that the ImageDescriptors and GraphicControls follow
** one another, ie. I have not associated them with each other. The driver
-** must assume a 1-1 correspondance. The same discussion with plain text
+** must assume a 1-1 correspondence. The same discussion with plain text
** extension.
*/
typedef struct _GifToMem {
diff --git a/hl/tools/gif2h5/gif2mem.c b/hl/tools/gif2h5/gif2mem.c
index c66250b..2c2225e 100644
--- a/hl/tools/gif2h5/gif2mem.c
+++ b/hl/tools/gif2h5/gif2mem.c
@@ -40,7 +40,7 @@ Gif2Mem(GIFBYTE *MemGif, GIFTOMEM *GifMemoryStruct)
{
/*
* The gif structure outline for passing data to memory is given in gif.h.
- * These pointers are redunant, should take them out in ver. 2
+ * These pointers are redundant, should take them out in ver. 2
*/
GIFHEAD * gifHead; /* GIF Header structure */
GIFIMAGEDESC ** gifImageDesc; /* Logical Image Descriptor struct */
diff --git a/hl/tools/gif2h5/gifread.c b/hl/tools/gif2h5/gifread.c
index dd9e5be..3a6bcae 100644
--- a/hl/tools/gif2h5/gifread.c
+++ b/hl/tools/gif2h5/gifread.c
@@ -50,15 +50,15 @@ GetByte(const GIFBYTE *MemGif)
* Global Color Table (if any) from a GIF image file. The information
* is stored in a GIFHEAD structure.
*
- * Returns: -1 if a FILE stream error occured during the read,
- * otherwise 0 if no error occured.
+ * Returns: -1 if a FILE stream error occurred during the read,
+ * otherwise 0 if no error occurred.
*/
int
ReadGifHeader(GIFHEAD * GifHead, /* Pointer to GIF header structure */
GIFBYTE **MemGif2) /* GIF image file input FILE stream */
{
GIFWORD i; /* Loop counter */
- GIFWORD tableSize; /* Number of entires in the Global Color Table */
+ GIFWORD tableSize; /* Number of entries in the Global Color Table */
GifHead->TableSize = 0;
for (i = 0; i < 6; i++) {
@@ -103,7 +103,7 @@ ReadGifHeader(GIFHEAD * GifHead, /* Pointer to GIF header structure */
return -1;
#endif /* 0 */
- return 0; /* No FILE stream error occured */
+ return 0; /* No FILE stream error occurred */
}
/*
@@ -116,8 +116,8 @@ ReadGifHeader(GIFHEAD * GifHead, /* Pointer to GIF header structure */
** Note that the ImageSeparator field value in the GIFIMAGEDESC
** structure is assigned by the function calling ReadGifImageDesc().
**
-** Returns: -1 if a FILE stream error occured during the read,
-** otherwise 0 if no error occured.
+** Returns: -1 if a FILE stream error occurred during the read,
+** otherwise 0 if no error occurred.
*/
int
ReadGifImageDesc(GIFIMAGEDESC *GifImageDesc, /* Pointer to GIF image descriptor structure */
@@ -196,7 +196,7 @@ ReadGifImageDesc(GIFIMAGEDESC *GifImageDesc, /* Pointer to GIF image descriptor
*TempPtr++ = *(*MemGif2)++;
} while (ch1);
- return (0); /* No FILE stream error occured */
+ return (0); /* No FILE stream error occurred */
}
/*
@@ -205,8 +205,8 @@ ReadGifImageDesc(GIFIMAGEDESC *GifImageDesc, /* Pointer to GIF image descriptor
** Note that the Introducer and Label field values in the GIFGRAPHICCONTROL
** structure are assigned by the function calling ReadGifGraphicControl().
**
-** Returns: -1 if a FILE stream error occured during the read,
-** otherwise 0 if no error occured.
+** Returns: -1 if a FILE stream error occurred during the read,
+** otherwise 0 if no error occurred.
*/
int
ReadGifGraphicControl(GIFGRAPHICCONTROL *GifGraphicControl, /* Pointer to GC Extension structure */
@@ -219,7 +219,7 @@ ReadGifGraphicControl(GIFGRAPHICCONTROL *GifGraphicControl, /* Pointer to GC Ext
GifGraphicControl->GCEDump[i] = *(*MemGif2)++;
}
- return (0); /* No FILE stream error occured */
+ return (0); /* No FILE stream error occurred */
}
/*
@@ -228,8 +228,8 @@ ReadGifGraphicControl(GIFGRAPHICCONTROL *GifGraphicControl, /* Pointer to GC Ext
** Note that the Introducer and Label field values in the GIFLPLAINTEXT
** structure are assigned by the function calling ReadGifPlainText().
**
-** Returns: -1 if a FILE stream error occured during the read,
-** otherwise 0 if no error occured.
+** Returns: -1 if a FILE stream error occurred during the read,
+** otherwise 0 if no error occurred.
*/
int
ReadGifPlainText(GIFPLAINTEXT *GifPlainText, /* Pointer to Plain Text Extension structure */
@@ -256,7 +256,7 @@ ReadGifPlainText(GIFPLAINTEXT *GifPlainText, /* Pointer to Plain Text Extension
return(-1);
*/
- return (0); /* No FILE stream error occured */
+ return (0); /* No FILE stream error occurred */
}
/*
@@ -265,8 +265,8 @@ ReadGifPlainText(GIFPLAINTEXT *GifPlainText, /* Pointer to Plain Text Extension
** Note that the Introducer and Label field values in the GIFAPPLICATION
** structure are assigned by the function calling ReadGifApplication().
**
-** Returns: -1 if a FILE stream error occured during the read,
-** otherwise 0 if no error occured.
+** Returns: -1 if a FILE stream error occurred during the read,
+** otherwise 0 if no error occurred.
*/
int
ReadGifApplication(GIFAPPLICATION *GifApplication, /* Pointer to Application Extension structure */
@@ -292,7 +292,7 @@ ReadGifApplication(GIFAPPLICATION *GifApplication, /* Pointer to Application Ext
return(-1);
*/
- return (0); /* No FILE stream error occured */
+ return (0); /* No FILE stream error occurred */
}
/*
@@ -301,8 +301,8 @@ ReadGifApplication(GIFAPPLICATION *GifApplication, /* Pointer to Application Ext
** Note that the Introducer and Label field values in the GIFCOMMENT
** structure are assigned by the function calling ReadGifComment().
**
-** Returns: -1 if a FILE stream error occured during the read,
-** otherwise 0 if no error occured.
+** Returns: -1 if a FILE stream error occurred during the read,
+** otherwise 0 if no error occurred.
*/
int
ReadGifComment(GIFCOMMENT *GifComment, /* Pointer to GIF Comment Extension structure */
@@ -316,7 +316,7 @@ ReadGifComment(GIFCOMMENT *GifComment, /* Pointer to GIF Comment Extension struc
GifComment->Terminator = 0;
- return (0); /* No FILE stream error occured */
+ return (0); /* No FILE stream error occurred */
}
/*
@@ -326,8 +326,8 @@ ReadGifComment(GIFCOMMENT *GifComment, /* Pointer to GIF Comment Extension struc
** A GIF "sub-block" is a single count byte followed by 1 to 255
** additional data bytes.
**
-** Returns: A NULL pointer if a memory allocation error occured,
-** otherwise a valid pointer if no error occured.
+** Returns: A NULL pointer if a memory allocation error occurred,
+** otherwise a valid pointer if no error occurred.
*/
static GIFBYTE *
ReadDataSubBlocks(GIFBYTE **MemGif2, /* GIF image file input FILE stream */
@@ -364,7 +364,7 @@ ReadDataSubBlocks(GIFBYTE **MemGif2, /* GIF image file input FILE stream
if ((dataSize = *(*MemGif2)++) == 0)
break; /* Block Terminator encountered */
- /* Increase the buffer size to accomodate the next sub-block */
+ /* Increase the buffer size to accommodate the next sub-block */
if (!(ptr1 = ptr2 = (GIFBYTE *)realloc(ptr2, bufSize + dataSize + 1)))
return ((GIFBYTE *)NULL);
diff --git a/hl/tools/gif2h5/h52giftest.sh.in b/hl/tools/gif2h5/h52giftest.sh.in
index 2f84ef6..4f4057f 100644
--- a/hl/tools/gif2h5/h52giftest.sh.in
+++ b/hl/tools/gif2h5/h52giftest.sh.in
@@ -46,7 +46,7 @@ TOOLTEST()
# Verify the test runs with failure (return code is not 0)
# Use for testing if tool can handle error conditions like
-# illegal input, bad arguments, exeeding limits, ...
+# illegal input, bad arguments, exceeding limits, ...
TOOLTESTFAIL()
{
# for now, discard any error messages generated.
diff --git a/hl/tools/gif2h5/writehdf.c b/hl/tools/gif2h5/writehdf.c
index 6ec6af4..c321519 100644
--- a/hl/tools/gif2h5/writehdf.c
+++ b/hl/tools/gif2h5/writehdf.c
@@ -27,7 +27,7 @@
* Programmer: Unknown
*
* Modifications: pvn
- * Use the HDF5 IMAGE API to write the HDF5 image and pallete
+ * Use the HDF5 IMAGE API to write the HDF5 image and palette
*
* Date: January, 31, 2006
*
@@ -68,7 +68,7 @@ WriteHDF(GIFTOMEM GifMemoryStruct, char *HDFName)
#endif /* UNUSED */
if ((file_id = H5Fcreate(HDFName, H5F_ACC_TRUNC, H5P_DEFAULT, H5P_DEFAULT)) < 0) {
- /* error occured opening the HDF File for write */
+ /* error occurred opening the HDF File for write */
fprintf(stderr, "HDF file could not be opened for writing\n");
fprintf(stderr,
"NOTE: GIF file must be present in the same directory as the binary on UNIX systems.\n");
diff --git a/hl/tools/h5watch/extend_dset.c b/hl/tools/h5watch/extend_dset.c
index 282f5c2..624eebc 100644
--- a/hl/tools/h5watch/extend_dset.c
+++ b/hl/tools/h5watch/extend_dset.c
@@ -83,7 +83,7 @@ extend_dset_two(const char *file, char *dname, int action1, int action2)
size_t dtype_size; /* size of the dataset's datatype */
unsigned num_elmts; /* number of elements in the dataset */
int * ibuf = NULL; /* buffer for storing retrieved elements (integer) */
- set_t * cbuf = NULL; /* buffer for storing retrieved elemnets (compound) */
+ set_t * cbuf = NULL; /* buffer for storing retrieved elements (compound) */
/* Allocate memory */
if (NULL == (ibuf = (int *)HDcalloc(TEST_BUF_SIZE, sizeof(int))))
@@ -237,7 +237,7 @@ extend_dset_one(const char *file, char *dname, int action)
hsize_t count[1]; /* dimension sizes of appended data */
size_t dtype_size; /* size of the dataset's datatype */
int * ibuf = NULL; /* buffer for storing retrieved elements (integer) */
- set_t * cbuf = NULL; /* buffer for storing retrieved elemnets (compound) */
+ set_t * cbuf = NULL; /* buffer for storing retrieved elements (compound) */
/* Allocate memory */
if (NULL == (ibuf = (int *)HDcalloc(TEST_BUF_SIZE, sizeof(int))))
diff --git a/hl/tools/h5watch/h5watch.c b/hl/tools/h5watch/h5watch.c
index afbc31d..c897c91 100644
--- a/hl/tools/h5watch/h5watch.c
+++ b/hl/tools/h5watch/h5watch.c
@@ -921,7 +921,7 @@ main(int argc, const char *argv[])
}
}
- /* If everything is fine, start monitoring the datset */
+ /* If everything is fine, start monitoring the dataset */
if (h5tools_getstatus() != EXIT_FAILURE)
if (monitor_dataset(fid, dname) < 0)
h5tools_setstatus(EXIT_FAILURE);
diff --git a/java/examples/datasets/H5Ex_D_UnlimitedMod.java b/java/examples/datasets/H5Ex_D_UnlimitedMod.java
index 273ac3e..0708898 100644
--- a/java/examples/datasets/H5Ex_D_UnlimitedMod.java
+++ b/java/examples/datasets/H5Ex_D_UnlimitedMod.java
@@ -17,7 +17,7 @@
file. Next, it reopens the file, reads back the data,
outputs it to the screen, extends the dataset, and writes
new data to the entire extended dataset. Finally it
- reopens the file again, reads back the data, and utputs it
+ reopens the file again, reads back the data, and outputs it
to the screen.
************************************************************/
package examples.datasets;
diff --git a/java/examples/groups/H5Ex_G_Corder.java b/java/examples/groups/H5Ex_G_Corder.java
index 5df850e..4fa5aa2 100644
--- a/java/examples/groups/H5Ex_G_Corder.java
+++ b/java/examples/groups/H5Ex_G_Corder.java
@@ -11,7 +11,7 @@
* * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * */
/************************************************************
Creating a file with creation properties and traverse the
- groups in alpabetical and creation order.
+ groups in alphabetical and creation order.
************************************************************/
package examples.groups;
diff --git a/java/src/hdf/hdf5lib/H5.java b/java/src/hdf/hdf5lib/H5.java
index 4129f50..49a539b 100644
--- a/java/src/hdf/hdf5lib/H5.java
+++ b/java/src/hdf/hdf5lib/H5.java
@@ -161,7 +161,7 @@ import hdf.hdf5lib.structs.H5O_token_t;
* disk (source) and in memory (destination).
* <p>
* For Java, this ``ANY'' is a problem, as the type of data must always be declared. Furthermore, multidimensional
- * arrays are definitely <i>not</i> layed out contiguously in memory. It would be infeasible to declare a separate
+ * arrays are definitely <i>not</i> laid out contiguously in memory. It would be infeasible to declare a separate
* routine for every combination of number type and dimensionality. For that reason, the
* <a href="./hdf.hdf5lib.HDFArray.html"><b>HDFArray</b></a> class is used to discover the type, shape, and size of the
* data array at run time, and to convert to and from a contiguous array of bytes in synchronized static native C order.
@@ -442,7 +442,7 @@ public class H5 implements java.io.Serializable {
/**
* Turn on error handling. By default, the C library prints the error stack of the HDF-5 C library on stdout. This
- * behavior may be reenabled by calling H5error_on().
+ * behavior may be re-enabled by calling H5error_on().
*/
public synchronized static native void H5error_on();
@@ -2097,7 +2097,7 @@ public class H5 implements java.io.Serializable {
* IN: Identifier for object to which attributes are attached; may be group, dataset, or named datatype.
* @param idx_type
* IN: The type of index specified by idx_type can be one of the following:
- * H5_INDEX_NAME An alpha-numeric index by attribute name.
+ * H5_INDEX_NAME An alphanumeric index by attribute name.
* H5_INDEX_CRT_ORDER An index by creation order.
* @param order
* IN: The order in which the index is to be traversed, as specified by order, can be one of the following:
@@ -2137,7 +2137,7 @@ public class H5 implements java.io.Serializable {
* IN: Name of object, relative to location.
* @param idx_type
* IN: The type of index specified by idx_type can be one of the following:
- * H5_INDEX_NAME An alpha-numeric index by attribute name.
+ * H5_INDEX_NAME An alphanumeric index by attribute name.
* H5_INDEX_CRT_ORDER An index by creation order.
* @param order
* IN: The order in which the index is to be traversed, as specified by order, can be one of the following:
@@ -4577,7 +4577,7 @@ public class H5 implements java.io.Serializable {
throws HDF5LibraryException, NullPointerException;
/**
- * Given a mount point, H5Funmount dissassociates the mount point's file from the file mounted there.
+ * Given a mount point, H5Funmount disassociates the mount point's file from the file mounted there.
*
* @param loc_id
* The identifier for the location at which the specified file is to be unmounted.
@@ -5205,7 +5205,7 @@ public class H5 implements java.io.Serializable {
*/
/**
* retrieves information of all objects (recurvisely) under the group (name) located in the file or group specified
- * by loc_id upto maximum specified by objMax.
+ * by loc_id up to maximum specified by objMax.
*
* @param loc_id
* IN: File or group identifier
@@ -7182,7 +7182,7 @@ public class H5 implements java.io.Serializable {
* @param size
* IN: Size the property value.
* @param def_value
- * IN: Defaul value of the property
+ * IN: Default value of the property
*
* @exception HDF5LibraryException
* - Error from the HDF-5 Library.
@@ -7202,7 +7202,7 @@ public class H5 implements java.io.Serializable {
* @param size
* IN: Size the property value.
* @param value
- * IN: Defaul value of the property
+ * IN: Default value of the property
*
* @exception HDF5LibraryException
* - Error from the HDF-5 Library.
@@ -7245,7 +7245,7 @@ public class H5 implements java.io.Serializable {
* @param ocpl_id
* IN: : Object (dataset or group) creation property list identifier
* @param attributes
- * The maximun and minimum no. of attributes to be stored.
+ * The maximum and minimum no. of attributes to be stored.
*
* <pre>
* attributes[0] = The maximum number of attributes to be stored in compact storage
@@ -7353,7 +7353,7 @@ public class H5 implements java.io.Serializable {
* transfer property list. The FLAGS argument specifies certain
* general properties of the filter and is documented below.
* The CD_VALUES is an array of CD_NELMTS integers which are
- * auxiliary data for the filter. The integer vlues will be
+ * auxiliary data for the filter. The integer values will be
* stored in the dataset object header as part of the filter
* information.
*<p>
@@ -8054,9 +8054,9 @@ public class H5 implements java.io.Serializable {
* @exception HDF5LibraryException
* - Error from the HDF-5 Library.
* @exception NullPointerException
- * - aligment array is null.
+ * - alignment array is null.
* @exception IllegalArgumentException
- * - aligment array is invalid.
+ * - alignment array is invalid.
**/
public synchronized static native int H5Pget_alignment(long plist, long[] alignment) throws HDF5LibraryException,
NullPointerException, IllegalArgumentException;
@@ -8202,7 +8202,7 @@ public class H5 implements java.io.Serializable {
throws HDF5LibraryException;
/**
- * H5Pget_gc_references Returns the current setting for the garbage collection refernces property from a file access
+ * H5Pget_gc_references Returns the current setting for the garbage collection references property from a file access
* property list.
*
* @param fapl_id
@@ -9254,7 +9254,7 @@ public class H5 implements java.io.Serializable {
* the raw data chunk cache on a per-datset basis.
*
* @param dapl_id
- * IN: Identifier of the datset access property list.
+ * IN: Identifier of the dataset access property list.
* @param rdcc_nslots
* IN: Number of elements (objects) in the raw data chunk cache.
* @param rdcc_nbytes
diff --git a/java/src/hdf/hdf5lib/exceptions/HDF5LibraryException.java b/java/src/hdf/hdf5lib/exceptions/HDF5LibraryException.java
index 532355e..9f94882 100644
--- a/java/src/hdf/hdf5lib/exceptions/HDF5LibraryException.java
+++ b/java/src/hdf/hdf5lib/exceptions/HDF5LibraryException.java
@@ -114,7 +114,7 @@ public class HDF5LibraryException extends HDF5Exception {
return "special zero no error";
}
else if (err_code == HDF5Constants.H5E_UNINITIALIZED) {
- return "information is unitialized";
+ return "information is uninitialized";
}
else if (err_code == HDF5Constants.H5E_UNSUPPORTED) {
return "feature is unsupported";
diff --git a/java/src/hdf/hdf5lib/package-info.java b/java/src/hdf/hdf5lib/package-info.java
index 838a266..7edfcb2 100644
--- a/java/src/hdf/hdf5lib/package-info.java
+++ b/java/src/hdf/hdf5lib/package-info.java
@@ -114,7 +114,7 @@
* disk (source) and in memory (destination).
* <p>
* For Java, this ``ANY'' is a problem, as the type of data must always be declared. Furthermore, multidimensional
- * arrays are definitely <i>not</i> layed out contiguously in memory. It would be infeasible to declare a separate
+ * arrays are definitely <i>not</i> laid out contiguously in memory. It would be infeasible to declare a separate
* routine for every combination of number type and dimensionality. For that reason, the
* <a href="./hdf.hdf5lib.HDFArray.html"><b>HDFArray</b></a> class is used to discover the type, shape, and size of the
* data array at run time, and to convert to and from a contiguous array of bytes in synchronized static native C order.
diff --git a/java/src/hdf/hdf5lib/structs/H5F_info2_t.java b/java/src/hdf/hdf5lib/structs/H5F_info2_t.java
index f951bb4..78e6d6c 100644
--- a/java/src/hdf/hdf5lib/structs/H5F_info2_t.java
+++ b/java/src/hdf/hdf5lib/structs/H5F_info2_t.java
@@ -40,7 +40,7 @@ public class H5F_info2_t implements Serializable{
public H5_ih_info_t sohm_msgs_info;
/**
- * Constructor fot current "global" information about file
+ * Constructor for current "global" information about file
* @param super_version: Superblock version number
* @param super_size: Superblock size
* @param super_ext_size: Superblock extension size
diff --git a/java/src/jni/h5util.c b/java/src/jni/h5util.c
index 7313c28..02e99af 100644
--- a/java/src/jni/h5util.c
+++ b/java/src/jni/h5util.c
@@ -1348,7 +1348,7 @@ done:
* This is a special case subfunction to print the data in a region reference of type blocks.
*
* Return:
- * The function returns FAIL if there was an error, otherwise SUCEED
+ * The function returns FAIL if there was an error, otherwise SUCCEED
*-------------------------------------------------------------------------
*/
static int
@@ -1771,7 +1771,7 @@ done:
* Purpose: Recursive check for variable length string of a datatype.
*
* Return:
- * TRUE : type conatains any variable length string
+ * TRUE : type contains any variable length string
* FALSE : type doesn't contain any variable length string
* Negative value: error occur
*
@@ -2300,7 +2300,7 @@ done:
* This is a special case subfunction to print the data in a region reference of type blocks.
*
* Return:
- * The function returns FAIL if there was an error, otherwise SUCEED
+ * The function returns FAIL if there was an error, otherwise SUCCEED
*
*-------------------------------------------------------------------------
*/
diff --git a/java/test/TestH5P.java b/java/test/TestH5P.java
index 3879128..6ae93d9 100644
--- a/java/test/TestH5P.java
+++ b/java/test/TestH5P.java
@@ -1173,12 +1173,12 @@ public class TestH5P {
strategy = H5.H5Pget_file_space_strategy(fcpl_id, persist, threshold);
assertTrue("strategy(default): "+strategy, strategy == HDF5Constants.H5F_FSPACE_STRATEGY_FSM_AGGR);
assertTrue("persist(default): "+persist[0], persist[0] == false);
- assertTrue("theshold(default): "+threshold[0], threshold[0] == 1);
+ assertTrue("threshold(default): "+threshold[0], threshold[0] == 1);
H5.H5Pset_file_space_strategy(fcpl_id, HDF5Constants.H5F_FSPACE_STRATEGY_PAGE, true, 1);
strategy = H5.H5Pget_file_space_strategy(fcpl_id, persist, threshold);
assertTrue("strategy: "+strategy, strategy == HDF5Constants.H5F_FSPACE_STRATEGY_PAGE);
assertTrue("persist: "+persist[0], persist[0] == true);
- assertTrue("theshold: "+threshold[0], threshold[0] == 1);
+ assertTrue("threshold: "+threshold[0], threshold[0] == 1);
}
catch (Throwable err) {
err.printStackTrace();
diff --git a/java/test/TestH5Pfapl.java b/java/test/TestH5Pfapl.java
index a65b52e..f6be24f 100644
--- a/java/test/TestH5Pfapl.java
+++ b/java/test/TestH5Pfapl.java
@@ -892,7 +892,7 @@ public class TestH5Pfapl {
try {
H5.H5Pset_fapl_multi(fapl_id, member_map, member_fapl, member_name, member_addr, true);
long driver_type = H5.H5Pget_driver(fapl_id);
- assertTrue("H5Pget_driver: muti = "+ driver_type, HDF5Constants.H5FD_MULTI==driver_type);
+ assertTrue("H5Pget_driver: multi = "+ driver_type, HDF5Constants.H5FD_MULTI==driver_type);
}
catch (Throwable err) {
err.printStackTrace();
@@ -932,7 +932,7 @@ public class TestH5Pfapl {
try {
H5.H5Pset_fapl_multi(fapl_id, member_map, member_fapl, member_name, member_addr, true);
long driver_type = H5.H5Pget_driver(fapl_id);
- assertTrue("H5Pget_driver: muti = "+ driver_type, HDF5Constants.H5FD_MULTI==driver_type);
+ assertTrue("H5Pget_driver: multi = "+ driver_type, HDF5Constants.H5FD_MULTI==driver_type);
}
catch (Throwable err) {
err.printStackTrace();
@@ -1001,7 +1001,7 @@ public class TestH5Pfapl {
try {
H5.H5Pset_fapl_multi(fapl_id, member_map, member_fapl, member_name, member_addr, true);
long driver_type = H5.H5Pget_driver(fapl_id);
- assertTrue("H5Pget_driver: muti = "+ driver_type, HDF5Constants.H5FD_MULTI==driver_type);
+ assertTrue("H5Pget_driver: multi = "+ driver_type, HDF5Constants.H5FD_MULTI==driver_type);
}
catch (Throwable err) {
err.printStackTrace();
@@ -1223,7 +1223,7 @@ public class TestH5Pfapl {
try {
H5.H5Pset_fapl_multi(fapl_id, member_map, member_fapl, member_name, member_addr, true);
long driver_type = H5.H5Pget_driver(fapl_id);
- assertTrue("H5Pget_driver: muti = "+ driver_type, HDF5Constants.H5FD_MULTI==driver_type);
+ assertTrue("H5Pget_driver: multi = "+ driver_type, HDF5Constants.H5FD_MULTI==driver_type);
}
catch (Throwable err) {
err.printStackTrace();
diff --git a/java/test/TestH5Plist.java b/java/test/TestH5Plist.java
index 0d53071..03e1c8f 100644
--- a/java/test/TestH5Plist.java
+++ b/java/test/TestH5Plist.java
@@ -148,7 +148,7 @@ public class TestH5Plist {
}
assertTrue("H5Pequal cid2", status >= 0);
- // Make certain false postives aren't being returned
+ // Make certain false positives aren't being returned
try {
status = H5.H5Pequal(cid2, HDF5Constants.H5P_FILE_CREATE);
}
@@ -295,7 +295,7 @@ public class TestH5Plist {
}
assertTrue("H5Pget_nprops: "+nprops, nprops==0);
- // Check the existance of the first property (should fail)
+ // Check the existence of the first property (should fail)
try {
status = H5.H5Pexist(plist_class_id, PROP1_NAME);
}
@@ -326,7 +326,7 @@ public class TestH5Plist {
catch (Throwable err) {
}
- // Check the existance of the first property
+ // Check the existence of the first property
try {
status = H5.H5Pexist(plist_class_id, PROP1_NAME);
}
@@ -377,7 +377,7 @@ public class TestH5Plist {
catch (Throwable err) {
}
- // Check the existance of the second property
+ // Check the existence of the second property
try {
status = H5.H5Pexist(plist_class_id, PROP2_NAME);
}
@@ -418,7 +418,7 @@ public class TestH5Plist {
fail("H5Pregister2 plist_class_id: "+PROP3_NAME + err);
}
- // Check the existance of the third property
+ // Check the existence of the third property
try {
status = H5.H5Pexist(plist_class_id, PROP3_NAME);
}
diff --git a/m4/aclocal_fc.f90 b/m4/aclocal_fc.f90
index 240a768..bcefab5 100644
--- a/m4/aclocal_fc.f90
+++ b/m4/aclocal_fc.f90
@@ -120,7 +120,7 @@ PROGRAM FC_AVAIL_KINDS
k = SELECTED_REAL_KIND(ik,jk)
IF(k.LT.0) EXIT exp
IF(k.NE.prev_rkind)THEN
- ! Check if we aleady have that kind
+ ! Check if we already have that kind
new_kind = .TRUE.
DO kk = 1, num_rkinds
IF(k.EQ.list_rkinds(kk))THEN
diff --git a/m4/aclocal_fc.m4 b/m4/aclocal_fc.m4
index 49ff485..9d01f47 100644
--- a/m4/aclocal_fc.m4
+++ b/m4/aclocal_fc.m4
@@ -20,7 +20,7 @@ dnl Contains Macros for HDF5 Fortran
dnl *********************************
dnl
dnl Special characteristics that have no autoconf counterpart but that
-dnl we need as part of the C++ support. To distinquish these, they
+dnl we need as part of the C++ support. To distinguish these, they
dnl have a [PAC] prefix.
dnl
dnl -------------------------------------------------------------------------
@@ -260,8 +260,8 @@ TEST_SRC="`sed -n '/PROGRAM FC_AVAIL_KINDS/,/END PROGRAM FC_AVAIL_KINDS/p' $srcd
AC_RUN_IFELSE([$TEST_SRC],
[
dnl The output from the above program will be:
- dnl -- LINE 1 -- valid integer kinds (comma seperated list)
- dnl -- LINE 2 -- valid real kinds (comma seperated list)
+ dnl -- LINE 1 -- valid integer kinds (comma separated list)
+ dnl -- LINE 2 -- valid real kinds (comma separated list)
dnl -- LINE 3 -- max decimal precision for reals
dnl -- LINE 4 -- number of valid integer kinds
dnl -- LINE 5 -- number of valid real kinds
diff --git a/release_docs/HISTORY-1_0-1_8_0_rc3.txt b/release_docs/HISTORY-1_0-1_8_0_rc3.txt
index f23a2c5..f54ba64 100644
--- a/release_docs/HISTORY-1_0-1_8_0_rc3.txt
+++ b/release_docs/HISTORY-1_0-1_8_0_rc3.txt
@@ -103,9 +103,9 @@ New Features
-JML 2006/6/21
- If both shared and static libraries are installed, now both will be
tested during 'make install'. -MAM 2006/06/21
- - Added support to explicity enable stream_vfd or shared libraries
+ - Added support to explicitly enable stream_vfd or shared libraries
when using parallel via the '--enable-stream_vfd' and
- '--enable-shared' options, respectively. If not explicity defined,
+ '--enable-shared' options, respectively. If not explicitly defined,
These settings default to enabled when parallel is not used,
and disabled when parallel is used. -MAM 2006/06/17
- Remove the flexible parallel code and the --enable-fphdf5
@@ -169,7 +169,7 @@ New Features
variable $FC, not $F9X. F9X still works, but is depreciated.
The output of make may be different. This should be only a
cosmetic effect.
- make depened (or make dep) is no longer recognized, since automake
+ make depend (or make dep) is no longer recognized, since automake
handles dependency tracking.
Some new configure options exist. --enable-dependency-tracking
and --disable-dependency-tracking are used to control automake's
@@ -306,11 +306,11 @@ New Features
- Added managements of collective IO supports for chunking storage
inside parallel HDF5
1) Implemented One IO with collective mode for all chunks in the
- application by building one MPI derived datatype accross all
+ application by building one MPI derived datatype across all
chunks.
2) Implemented the decision-making support to do collective IO inside
MPI-IO per chunk.
- 3) Added the decision-making support to do one IO accross all chunks
+ 3) Added the decision-making support to do one IO across all chunks
or to do multiple IOs with each IO per chunk.
4) Added the support to handle the case some processes won't do any IOs in
collectively.
@@ -319,7 +319,7 @@ New Features
contributions to IOs, a special macro is added to change
collective IO mode to independent IO mode inside HDF5 library.
- Currently we find that MPICH at Linux and vender MPI-IO package at NCSA
+ Currently we find that MPICH at Linux and vendor MPI-IO package at NCSA
Altix cannot handle this case.
"hdf5_mpi_special_collective_io_works=${hdf5_mpi_special_collective_io_works='no'}"
@@ -469,7 +469,7 @@ New Features
object's object header. QAK 2003/10/06
- Added new H5Fget_freespace() routine to query the free space in a
given file. QAK 2003/10/06
- - Added backward compatability with v1.6 for new Error API. SLU -
+ - Added backward compatibility with v1.6 for new Error API. SLU -
2003/09/24
- Changed 'objno' field in H5G_stat_t structure from 'unsigned long[2]'
to 'haddr_t'. QAK - 2003/08/08
@@ -496,7 +496,7 @@ New Features
- Add options of using atomicity and file-sync to test_mpio_1wMr.
AKC - 2003/11/13
- Added parallel test, test_mpio_1wMr, which tests if the
- underlaying parallel I/O system is conforming to the POSIX
+ underlying parallel I/O system is conforming to the POSIX
write/read requirement. AKC - 2003/11/12
Fortran Library:
@@ -606,7 +606,7 @@ New Features
and H5LTdtype_to_text. H5LTtext_to_dtype creates a HDF5 data type
given a text description; H5LTdtype_to_text converts a data type
to text description. Only DDL definition is supported as text
- desciption now. SLU - 2006/05/17
+ description now. SLU - 2006/05/17
- Added Packet Table API for creating tables with less overhead than
H5TB API. Added C++ wrapper for Packet Tables. See documentation.
JML - 2004/03/28
@@ -639,7 +639,7 @@ Support for new platforms, languages and compilers.
- Added support for HPUX11.23 for both 32 and 64-bit; HDF5 C++
- Added support for 64-bit Windows with Visual Studio .NET and 2005.
SJW - 2007/06/25
- - Added suport for HPUX11.23 for both 32 and 64-bit; HDF5 C++
+ - Added support for HPUX11.23 for both 32 and 64-bit; HDF5 C++
shared library is not supported with +DD64 flag.
EIP - 2006/06/22
- Added support for VAX floating numbers for Alpha Open VMS 7.3.2
@@ -655,7 +655,7 @@ Support for new platforms, languages and compilers.
- HDF5 Fortran APIs are supported on Mac OSX with IBM XL Fortran
compiler version 8.1. This is a default compiler.
- HDF5 Fortran APIs are supported on MAC OSX with Absoft F95 compiler
- version 8.2; set F9X environment varibale to f95, for example
+ version 8.2; set F9X environment variable to f95, for example
setenv F9X f95
Use --disable-shared --enable-static configure flags when Absoft
compiler is used.
@@ -700,7 +700,7 @@ Bug Fixes since HDF5-1.6.0 release
- Re-open the file
- Delete one of the attributes on the object
- Add a smaller attribute to the object
- - Delete the smaller atttribute on the object
+ - Delete the smaller attribute on the object
- Add a larger attribute on the object
After this, the number of header messages stored for the object
@@ -855,7 +855,7 @@ Bug Fixes since HDF5-1.6.0 release
- Fixed dtypes "sw long double -> double" failure in QSC class
machines. AKC - 2004/4/16
- Fixed problem with fletcher32 filter when converting data of different
- endianess. PVN - 2004/03/10
+ endianness. PVN - 2004/03/10
- Fixed problem with H5Tget_native_type() not handling opaque fields
correctly. QAK - 2004/01/31
- Fixed several errors in B-tree deletion code which could cause a
@@ -1271,7 +1271,7 @@ Known Problems
sometimes. This can be verified with t_mpi test under testpar.
* On IRIX6.5, when C compiler version >7.4, the complicate MPI derived data type
code will work. However, the user should be aware to enlarge MPI_TYPE_MAX environment
- variable to some certian value in order to use collective irregular selection code.
+ variable to some certain value in order to use collective irregular selection code.
For example, the current parallel HDF5 test needs to enlarge MPI_TYPE_MAX to
200,000 to make the test pass.
* The dataset created or rewritten with the v1.6.3 library or after can't
@@ -1428,7 +1428,7 @@ Known Problems
exit $RETURN_VALUE
- You get the HDF make files and test scipts to execute your filter script
+ You get the HDF make files and test scripts to execute your filter script
by setting the environment variable "RUNSERIAL" to the full path of the
script prior to running configure for parallel builds. Remember to
"unsetenv RUNSERIAL" before running configure for a serial build.
@@ -2098,7 +2098,7 @@ Bug Fixes since HDF5-1.6.5 Release
- Re-open the file.
- Delete one of the attributes on the object.
- Add a smaller attribute to the object.
- - Delete the smaller atttribute on the object.
+ - Delete the smaller attribute on the object.
- Add a larger attribute on the object.
After this, the number of header messages stored for the object
@@ -2399,11 +2399,11 @@ Known Problems
For 32-bit:
You may do this, under csh:
- setenv LD_LIBRARYN32_PATH .......(existing pathes):[full path of HDF5
+ setenv LD_LIBRARYN32_PATH .......(existing paths):[full path of HDF5
directory/test/.libs]:[full path of HDF5 directory/src/.libs]
For 64-bit:
- setenv LD_LIBRARY64_PATH ......(existing pathes):[full path of HDF5
+ setenv LD_LIBRARY64_PATH ......(existing paths):[full path of HDF5
directory/test/.libs]:[full path of HDF5 directory/src/.libs]
NOTE: This problem ONLY affects the testing of the HDF5 library when you
@@ -2723,7 +2723,7 @@ Bug Fixes since HDF5-1.6.4 Release
-------
- Fixed collective IO in chunking-storage. HDF5 may have called the
wrong routine when the shape of the dataspace in the file and in
- the bufferred chunk were different. This bug was fixed to make sure
+ the buffered chunk were different. This bug was fixed to make sure
the correct routine is called. KY - 2005/10/19
- Fixed core dump when closing root groups opened through two different
file handles that operate on the same actual file. QAK - 2005/10/02
@@ -2744,7 +2744,7 @@ Bug Fixes since HDF5-1.6.4 Release
together all have the same "file close degree". QAK - 2005/07/19
- More bug fixes on holding open files that are mounted and have
IDs open. QAK - 2005/07/14
- - Dataset sieve buffer cache was inadvertantly disabled; it has been
+ - Dataset sieve buffer cache was inadvertently disabled; it has been
re-enabled. QAK - 2005/07/08
- Don't unmount child files until the parent file actually closes.
(Previously, if an object was holding open a file, the child files
@@ -3024,7 +3024,7 @@ Known Problems
to intel forum.
* When testing parallel HDF5 with the C compiler version MIPSpro 7.4.3 at IRIX
- 6.5, set enviroment variable MPI_TYPE_MAX to be a bigger number, for example
+ 6.5, set environment variable MPI_TYPE_MAX to be a bigger number, for example
120000, in order to pass the complicated collective IO tests inside parallel
HDF5 library. This is not a problem inside parallel HDF5 library. You can
always set a bigger number in your system.
@@ -4448,7 +4448,7 @@ New Features
verbosity via the '-v' option. The default is less verbose
than before. AKC - 2004/01/23
- Added parallel test, test_mpio_1wMr, which tests if the
- underlaying parallel I/O system is conforming to the POSIX
+ underlying parallel I/O system is conforming to the POSIX
write/read requirement. This version includes options of using
atomicity and file-sync. AKC - 2003/11/27
@@ -4821,7 +4821,7 @@ Known Problems
Testing random sw long double -> double conversions
If -ieee is used, the converted doubles spread over the range 0.0 to 1**-308.
If -ieee is not used, the converted double values are mostly 0.0 but
- occassionaly as 1**-308. This has been reported to the system staff.
+ occasionally as 1**-308. This has been reported to the system staff.
All other tests have passed.
* On AIX 5.1 when 64-bit parallel C Library is built with zlib configured in,
@@ -5187,7 +5187,7 @@ Linux 2.4 IA64 Intel y n y n y y y
Known Problems
==============
* Fortran subroutine h5pget_driver_f doesn't return a correct driver information.
- The fix willl be available in the 1.6.2 release.
+ The fix will be available in the 1.6.2 release.
* There are two h5dump xml tests(h5dump --xml thlink.h5 and h5dump --xml tmany.h5)
failed on windows xp with .NET for debug and debug dll. Release and Release
dll work fine.
@@ -5368,7 +5368,7 @@ Library:
* The first version of szip compression support were implemented.
User should have static szlib library installed. Using function
H5Pset_szip to pass the szip parameters to the HDF5 library.
- More detailed decription of the process will be followed.
+ More detailed description of the process will be followed.
KY-2003/04/01
* Added Fletcher32 checksum as a filter in pipeline. It only works in
chunked dataset. SLU - 2003/2/11
@@ -5416,15 +5416,15 @@ Library:
be in different locations in the same file. The old functions H5Glink
and H5Gmove remain valid. SLU - 2002/04/26
* Fill-value's behaviors for contiguous dataset have been redefined.
- Basicly, dataset won't allocate space until it's necessary. Full details
+ Basically, dataset won't allocate space until it's necessary. Full details
are available at http://hdf.ncsa.uiuc.edu/RFC/Fill_Value, at this moment.
SLU - 2002/04/11
* Added new routine "H5Dfill" to fill a selection with a particular value
in memory. QAK - 2002/04/09
* Improved performance of "regular" hyperslab I/O when using MPI-IO and the
- datatype conversion is unneccessary. QAK - 2002/04/02
+ datatype conversion is unnecessary. QAK - 2002/04/02
* Improved performance of single hyperslab I/O when datatype conversion is
- unneccessary. QAK - 2002/04/02
+ unnecessary. QAK - 2002/04/02
* Added new "H5Sget_select_type" API function to determine which type of
selection is defined for a dataspace ("all", "none", "hyperslab" or
"point"). QAK - 2002/02/07
@@ -5438,7 +5438,7 @@ Library:
compatibility is enabled.
* Added a new file access property, file close degree, to control file
close behavior. It has four values, H5F_CLOSE_WEAK, H5F_CLOSE_SEMI,
- H5F_CLOSE_STRONG, and H5F_CLOSE_DEFAULT. Two correspont functions
+ H5F_CLOSE_STRONG, and H5F_CLOSE_DEFAULT. Two correspond functions
H5Pset_fclose_degree and H5Pget_fclose_degree are also provided. Two
new functions H5Fget_obj_count and H5Fget_obj_ids are offerted to assist
this new feature. For full details, please refer to the reference
@@ -5458,7 +5458,7 @@ Library:
* Improved regular hyperslab I/O by about a factor of 6 or so.
* Modified the Pablo build procedure to permit building of the instrumented
library to link either with the Trace libraries as before or with the
- Pablo Performance Caputure Facility.
+ Pablo Performance Capture Facility.
* Added new F90 APIs for generic properties, new filters, and
time/space allocation properties.
* C++ API:
@@ -5536,7 +5536,7 @@ Library
* Fixed data corruption problem which could occur when fill values were
written to a contiguously stored dataset in parallel. QAK - 2002/08/27
* Fixed VL memory leak when data is overwritten. The heap objects holding
- old data are freed. If the fill value writting time is set to
+ old data are freed. If the fill value writing time is set to
H5D_FILL_TIME_NEVER, the library prohibits user to create VL type dataset.
The library free all the heap objects storing VL type if there is nested
VL type(a VL type contains another VL type). SLU - 2002/07/10
@@ -6157,7 +6157,7 @@ Known Problems
* IRIX 6.5 fails to compile if configured with --enable-static-exec.
- * The executables are always dynamic on Solaris 2.7 ans 2.8(64 and n32)
+ * The executables are always dynamic on Solaris 2.7 and 2.8(64 and n32)
even if they are configured with --enable-static-exec.
* The HDF5_MPI_OPT_TYPES optimization code in the parallel HDF5 will cause
@@ -6175,7 +6175,7 @@ Known Problems
compiler optimization causes errors in many HDF5 Library tests. Use -g -h zero flags
to build HDF5 Library.
- * On Cray SV1 10.0.1. 0 datatype convertion test fails. Please check HDF FTP site
+ * On Cray SV1 10.0.1. 0 datatype conversion test fails. Please check HDF FTP site
if patch is available. We will try to provide one in the nearest future.
* For configuration, building and testing with Intel and PGI compilers see
@@ -6932,7 +6932,7 @@ Known Problems
HDF5_MPI_OPT_TYPES to a non-zero value such as 1.
* On IA64 systems one has to use -DIA64 compilation flag to compile
- h4toh5 and h5toh4 utilites. After configuration step manually modify
+ h4toh5 and h5toh4 utilities. After configuration step manually modify
Makefile in the tools/h4toh4 and tools/h5toh4 directories to add
-DIA64 to the compilation flags.
@@ -6998,7 +6998,7 @@ New Features
and other version information in the source code.
* Parallel HDF5 now runs on the HP V2500 and HP N4000 machines.
* F90 API:
- - Added aditional parameter "dims" to the h5dread_f/h5dwrite_f and
+ - Added additional parameter "dims" to the h5dread_f/h5dwrite_f and
h5aread_f/h5awrite_f subroutines. This parameter is a 1-D array
of size 7 and contains the sizes of the data buffer dimensions.
This change enables portability between Windows and UNIX platforms.
@@ -7012,8 +7012,8 @@ New Features
* F90 static library is available on Windows platforms.
See INSTALL_Windows_withF90.txt for details.
* F90 APIs are available on HPUX 11.00 and 10.20 and IBM SP platforms.
- * H5 <-> GIF convertor has been added. This is available under
- tools/gifconv. The convertor supports the ability to create animated
+ * H5 <-> GIF converter has been added. This is available under
+ tools/gifconv. The converter supports the ability to create animated
gifs as well.
* Verified correct operation of library on Solaris 2.8 in both 64-bit and
32-bit compilation modes. See INSTALL document for instructions on
@@ -7064,7 +7064,7 @@ Bug Fixes since HDF5-1.4.1 Release
when --enable-hdf5v1_2 configure flag is enabled.
* Added --enable-linux-lfs flag to allow more control over whether to
enable or disable large file support on Linux.
- * Fixed various bugs releated to SDS dimensional scale conversions in H4->H5
+ * Fixed various bugs related to SDS dimensional scale conversions in H4->H5
converter.
* Fixed a bug to correctly convert HDF4 objects with fill value into HDF5.
* Fixed a bug of H5pubconf.h causing repeated definitions if it is included
@@ -7345,7 +7345,7 @@ Bug Fixes since HDF5-1.4.0 Release
causing errors during I/O
* The RCSID string in H5public.h was causing the C++ compiling problem
because when it was included multiple times, C++ did not like
- multiple definitions of the same static variable. All occurance of
+ multiple definitions of the same static variable. All occurrence of
RCSID definition are removed since we have not used it consistently
before.
@@ -8111,7 +8111,7 @@ Known Problems
o SunOS 5.6 with C WorkShop Compilers 4.2: hyperslab selections will
fail if library is compiled using optimization of any level.
o TFLOPS: dsets test fails if compiled with optimization turned on.
-o J90: tools fail to dispay data for the datasets with a compound datatype.
+o J90: tools fail to display data for the datasets with a compound datatype.
Platforms Tested
================
@@ -8379,7 +8379,7 @@ C. Changes Since Version 1.0.1
* New query functions for selections.
* I/O operations bypass the stripmining loop and go directly to
- storage for certain contiguous selections in the absense of type
+ storage for certain contiguous selections in the absence of type
conversions. In other cases the stripmining buffers are used more
effectively.
diff --git a/release_docs/HISTORY-1_10.txt b/release_docs/HISTORY-1_10.txt
index 127b821..4796ebb 100644
--- a/release_docs/HISTORY-1_10.txt
+++ b/release_docs/HISTORY-1_10.txt
@@ -161,7 +161,7 @@ New Features
YES NO Default to STATIC
NO YES Default to SHARED
NO NO Default to SHARED
- The defaults can be overriden by setting the config option
+ The defaults can be overridden by setting the config option
HDF5_INSTALL_MOD_FORTRAN to one of NO, SHARED, or STATIC
(ADB - 2020/07/09, HDFFV-11116)
@@ -610,7 +610,7 @@ Bug Fixes since HDF5-1.10.6 release
- h5diff fixed a command line parsing error.
h5diff would ignore the argument to -d (delta) if it is smaller than DBL_EPSILON.
- The macro H5_DBL_ABS_EQUAL was removed and a direct value comparision was used.
+ The macro H5_DBL_ABS_EQUAL was removed and a direct value comparison was used.
(ADB - 2020/07/20, HDFFV-10897)
@@ -618,7 +618,7 @@ Bug Fixes since HDF5-1.10.6 release
h5diff would ignore all objects with a supplied path if the exclude-path argument is used.
Adding the exclude-attribute argument will only eclude attributes, with the supplied path,
- from comparision.
+ from comparison.
(ADB - 2020/07/20, HDFFV-5935)
@@ -1434,7 +1434,7 @@ Known Problems
and thus does not suffer from the 32 bit size limitation described here.
OpenMPI releases prior to v3.1.3 appear to have other datatype issues however,
e.g. within a single parallel test (testphdf5) the subtests (cdsetr, eidsetr)
- report data verfication errors before eventually aborting.
+ report data verification errors before eventually aborting.
The most recent versions of OpenMPI (v3.1.3 or newer) have evidently
resolved these isses and parallel HDF5 testing does not currently report
errors though occasional hangs have been observed.
@@ -1636,7 +1636,7 @@ New Features
(DER - 2019/02/03, HDFFV-10696)
- - Add ability to minimze dataset object headers.
+ - Add ability to minimize dataset object headers.
Creation of many, very small datasets resulted in extensive file bloat
due to extra space in the dataset object headers -- this space is
@@ -2137,7 +2137,7 @@ Known Problems
and thus does not suffer from the 32 bit size limitation described here.
OpenMPI releases prior to v3.1.3 appear to have other datatype issues however,
e.g. within a single parallel test (testphdf5) the subtests (cdsetr, eidsetr)
- report data verfication errors before eventually aborting.
+ report data verification errors before eventually aborting.
The most recent versions of OpenMPI (v3.1.3 or newer) have evidently
resolved these isses and parallel HDF5 testing does not currently report
errors though occasional hangs have been observed.
@@ -2865,7 +2865,7 @@ Bug Fixes since HDF5-1.10.2 release
When the function is called with the 'allgather' parameter set
to a non-true value, the function will receive data from all MPI
- ranks and gather it to the single rank specied by the 'root'
+ ranks and gather it to the single rank specified by the 'root'
parameter. However, the bug in the function caused memory for
the received data to be allocated on all MPI ranks, not just the
singular rank specified as the receiver. In some circumstances,
@@ -3431,11 +3431,11 @@ New Features
file driver in HDF5 by previously having called H5Pset_fapl_mpio().
HDF5 parallel file operations which do not employ multiple ranks
- e.g. specifiying MPI_COMM_SELF (whose MPI_Comm_size == 1)
+ e.g. specifying MPI_COMM_SELF (whose MPI_Comm_size == 1)
as opposed to MPI_COMM_WORLD, will not be affected by this
optimization. Conversely, parallel file operations on subgroups
of MPI_COMM_WORLD are allowed to be run in parallel with each
- subgroup operating as an independant collection of processes.
+ subgroup operating as an independent collection of processes.
(RAW - 2017/10/10, HDFFV-10294)
@@ -3574,7 +3574,7 @@ New Features
(ADB - 2017/12/20)
- The H5I_REFERENCE value in the H5I_type_t enum (defined in H5Ipublic.h)
- has been marked as deprectated.
+ has been marked as deprecated.
JNI code which refers to this value will be removed in a future
major version of the library. The code will remain unchanged in the
@@ -3895,7 +3895,7 @@ Bug Fixes since HDF5-1.10.1 release
Use of this VFD feature flag was not in line with the documentation in
the public H5FDpublic.h file. In particular, it was being used as a
- proxy for determining if SWMR I/O is allowed. This is unecessary as we
+ proxy for determining if SWMR I/O is allowed. This is unnecessary as we
already have a feature flag for this (H5FD_SUPPORTS_SWMR_IO).
(DER - 2017/05/31, HDFFV-10214)
@@ -4796,7 +4796,7 @@ Bug Fixes since HDF5-1.10.0-patch1 release
files to ensure that the library fails gracefully in these cases:
* Writing to a corrupted file that has an object message which is
- incorrectly marked as sharable on disk results in a buffer overflow /
+ incorrectly marked as shareable on disk results in a buffer overflow /
invalid write instead of a clean error message.
* Decoding data from a corrupted file with a dataset encoded with the
@@ -5258,7 +5258,7 @@ This release supports the following features:
identically to the 1.8 library with the --with-default-api-version=v18
configure flag. This allows existing code to be compiled with the 1.10
library without requiring immediate changes to the application source
- code. For addtional configuration options and other details, see
+ code. For additional configuration options and other details, see
"API Compatibility Macros in HDF5" at
https://www.hdfgroup.org/HDF5/doc/RM/APICompatMacros.html.
@@ -5342,7 +5342,7 @@ This release supports the following features:
Library
--------
- - Concurrent Access to HDF5 Files - Single Writer/ Multple Reader (SWMR)
+ - Concurrent Access to HDF5 Files - Single Writer/ Multiple Reader (SWMR)
The Single Writer/ Multiple Reader or SWMR feature enables users to
read data concurrently while writing it. Communications between the
@@ -5970,7 +5970,7 @@ This release supports the following features:
identically to the 1.8 library with the --with-default-api-version=v18
configure flag. This allows existing code to be compiled with the 1.10
library without requiring immediate changes to the application source
- code. For addtional configuration options and other details, see
+ code. For additional configuration options and other details, see
"API Compatibility Macros in HDF5" at
https://www.hdfgroup.org/HDF5/doc/RM/APICompatMacros.html.
@@ -6054,7 +6054,7 @@ This release supports the following features:
Library
--------
- - Concurrent Access to HDF5 Files - Single Writer/ Multple Reader (SWMR)
+ - Concurrent Access to HDF5 Files - Single Writer/ Multiple Reader (SWMR)
The Single Writer/ Multiple Reader or SWMR feature enables users to
read data concurrently while writing it. Communications between the
diff --git a/release_docs/HISTORY-1_10_0-1_12_0.txt b/release_docs/HISTORY-1_10_0-1_12_0.txt
index 60516fc..a83e58d 100644
--- a/release_docs/HISTORY-1_10_0-1_12_0.txt
+++ b/release_docs/HISTORY-1_10_0-1_12_0.txt
@@ -810,7 +810,7 @@ Bug Fixes since HDF5-1.10.2 release
When the function is called with the 'allgather' parameter set
to a non-true value, the function will receive data from all MPI
- ranks and gather it to the single rank specied by the 'root'
+ ranks and gather it to the single rank specified by the 'root'
parameter. However, the bug in the function caused memory for
the received data to be allocated on all MPI ranks, not just the
singular rank specified as the receiver. In some circumstances,
diff --git a/release_docs/HISTORY-1_8.txt b/release_docs/HISTORY-1_8.txt
index da3d849..1e31010 100644
--- a/release_docs/HISTORY-1_8.txt
+++ b/release_docs/HISTORY-1_8.txt
@@ -1213,7 +1213,7 @@ New Features
H5PLreplace - Replace the path at the specified index.
H5PLinsert - Insert a plugin path at the specified index, moving
other paths after the index.
- H5PLremove - Remove the plugin path at the specifed index and
+ H5PLremove - Remove the plugin path at the specified index and
compacting the list.
H5PLget - Query the plugin path at the specified index.
H5PLsize - Query the size of the current list of plugin paths.
@@ -1251,7 +1251,7 @@ New Features
// Returns a copy of the creation property list of a datatype.
DataType::getCreatePlist() - C API H5Tget_create_plist()
- // Opens an object within a location, regarless its type.
+ // Opens an object within a location, regardless its type.
H5Location::openObjId() - C API H5Oopen()
H5Location::openObjId() - C API H5Oopen()
@@ -1730,7 +1730,7 @@ Bug Fixes since HDF5-1.8.17
files to ensure that the library fails gracefully in these cases:
* Writing to a corrupted file that has an object message which is
- incorrectly marked as sharable on disk results in a buffer overflow /
+ incorrectly marked as shareable on disk results in a buffer overflow /
invalid write instead of a clean error message.
* Decoding data from a corrupted file with a dataset encoded with the
@@ -4622,7 +4622,7 @@ Bug Fixes since HDF5-1.8.13
Removed the module name from the package list of Fortran modules because
that module was never generated. This was only an issue for Fortran
- compliers that are not F2003 compatible.
+ compilers that are not F2003 compatible.
(ADB - 2014/10/16, HDFFV-8932)
@@ -4632,7 +4632,7 @@ Bug Fixes since HDF5-1.8.13
change was made which required the renaming of several public symbols in
H5Ppublic.h.
- The shared libary interface version number ("soname") has been increased
+ The shared library interface version number ("soname") has been increased
on account of these changes. For a full list of the changed symbols, see
the interface compatibility report, which is available as a link off of
the 'HDF5 Software Changes from Release to Release' document:
@@ -5891,7 +5891,7 @@ Known Problems
* The data conversion test dt_arith.c fails in "long double" to integer
conversion on Ubuntu 11.10 (3.0.0.13 kernel) with GCC 4.6.1 if the library
- is built with optimization -O3 or -O2. The older GCC (4.5) or newer kernal
+ is built with optimization -O3 or -O2. The older GCC (4.5) or newer kernel
(3.2.2 on Fedora) doesn't have the problem. Users should lower the
optimization level (-O1 or -O0) by defining CFLAGS in the command line of
"configure" like:
@@ -6746,7 +6746,7 @@ Bug Fixes since HDF5-1.8.10
- Fixed a bug in the core VFD that causes failures when opening files
> 2 GB.
Fixes HDFFV-8124 and HDFFV-8158 (DER 2013/01/07)
- - Fixed a bug where unintialized memory was read during variable-length
+ - Fixed a bug where uninitialized memory was read during variable-length
type conversion. This caused segfaults in netCDF. Fixes HDFFV-8159.
(DER 2013/03/30)
- Removed the H5Pset/get_dxpl_multi functions from the library. The
@@ -7094,7 +7094,7 @@ Known Problems
* The data conversion test dt_arith.c fails in "long double" to integer
conversion on Ubuntu 11.10 (3.0.0.13 kernel) with GCC 4.6.1 if the library
- is built with optimization -O3 or -O2. The older GCC (4.5) or newer kernal
+ is built with optimization -O3 or -O2. The older GCC (4.5) or newer kernel
(3.2.2 on Fedora) doesn't have the problem. Users should lower the
optimization level (-O1 or -O0) by defining CFLAGS in the command line of
"configure" like:
@@ -7718,8 +7718,8 @@ Known Problems
(MSC - 2012/04/18)
* The data conversion test dt_arith.c fails in "long double" to integer
- conversion on Ubuntu 11.10 (3.0.0.13 kernal) with GCC 4.6.1 if the library
- is built with optimization -O3 or -O2. The older GCC (4.5) or newer kernal
+ conversion on Ubuntu 11.10 (3.0.0.13 kernel) with GCC 4.6.1 if the library
+ is built with optimization -O3 or -O2. The older GCC (4.5) or newer kernel
(3.2.2 on Fedora) doesn't have the problem. Users should lower the
optimization level (-O1 or -O0) by defining CFLAGS in the command line of
"configure" like:
@@ -8317,8 +8317,8 @@ Known Problems
(MSC - 2012/04/18)
* The data conversion test dt_arith.c fails in "long double" to integer
- conversion on Ubuntu 11.10 (3.0.0.13 kernal) with GCC 4.6.1 if the library
- is built with optimization -O3 or -O2. The older GCC (4.5) or newer kernal
+ conversion on Ubuntu 11.10 (3.0.0.13 kernel) with GCC 4.6.1 if the library
+ is built with optimization -O3 or -O2. The older GCC (4.5) or newer kernel
(3.2.2 on Fedora) doesn't have the problem. Users should lower the
optimization level (-O1 or -O0) by defining CFLAGS in the command line of
"configure" like:
@@ -8696,7 +8696,7 @@ Bug Fixes since HDF5-1.8.8
------
- h5ltget_attribute_string_f: The h5ltget_attribute_string_f used to return
the C NULL character in the returned character buffer. The returned
- charactor buffer now does not return the C NULL character; the buffer
+ character buffer now does not return the C NULL character; the buffer
is blank-padded if needed. (MSB - 2012/03/23)
@@ -8799,7 +8799,7 @@ The following platforms and compilers have been tested for this release.
Mac OS X Snow Leopard 10.6.8 i686-apple-darwin10-gcc-4.2.1 (GCC) 4.2.1 (gcc)
Darwin Kernel Version 10.8.0 i686-apple-darwin10-g++-4.2.1 (GCC) 4.2.1 (g++)
Intel 64-bit (Apple Inc. build 5666) (dot 3)
- (fred) GNU Fortan (GCC) 4.6.1 (gfortran)
+ (fred) GNU Fortran (GCC) 4.6.1 (gfortran)
Intel C (icc), Fortran (ifort), C++ (icpc)
12.1.0.038 Build 20110811
@@ -8976,8 +8976,8 @@ Known Problems
(MSC - 2012/04/18)
* The data conversion test dt_arith.c fails in "long double" to integer
- conversion on Ubuntu 11.10 (3.0.0.13 kernal) with GCC 4.6.1 if the library
- is built with optimization -O3 or -O2. The older GCC (4.5) or newer kernal
+ conversion on Ubuntu 11.10 (3.0.0.13 kernel) with GCC 4.6.1 if the library
+ is built with optimization -O3 or -O2. The older GCC (4.5) or newer kernel
(3.2.2 on Fedora) doesn't have the problem. Users should lower the
optimization level (-O1 or -O0) by defining CFLAGS in the command line of
"configure" like:
@@ -9394,7 +9394,7 @@ Bug Fixes since HDF5-1.8.7
Tools
-----
- h5diff: fixed segfault over non-comparable attribute with different
- dimention or rank, along with '-c' option to display details.
+ dimension or rank, along with '-c' option to display details.
(JKM - 2011/10/24 - HDFFV-7770)
- Fixed h5diff to display all the comparable objects and attributes
regardless of detecting non-comparables. (JKM - 2011/09/16 - HDFFV-7693)
@@ -9433,7 +9433,7 @@ Bug Fixes since HDF5-1.8.7
(-d, -p and --use-system-epsilon) are used together.
(JKM - 2011/07/07 - HDFFV-7600)
- Fixed h5dump so that it displays the first line of each element in correct
- position for multiple dimention array types. Before this fix,
+ position for multiple dimension array types. Before this fix,
the first line of each element in an array was
displayed after the last line of previous element without
moving to the next line (+indentation).
@@ -10024,7 +10024,7 @@ Bug Fixes since HDF5-1.8.6
ibm-aix. (AKC - 2011/4/14)
- Shared C libraries are no longer disabled on Mac when Fortran
is enabled. Shared Fortran libraries are still not supported on Mac,
- so configure will disable them by default, but this is overrideable
+ so configure will disable them by default, but this is overridable
with the new --enable-unsupported configure option. The configure
summary has been updated to reflect the fact that the shared-ness of
the C++/Fortran wrapper libraries may not align with the C library.
@@ -10666,7 +10666,7 @@ Bug Fixes since HDF5-1.8.5
- Added a new set of unit tests that are run during 'make check' to verify
the behavior of the metadata accumulator. (MAM - 2010/10/15)
- Modified library to always cache symbol table information. Libraries
- from version 1.6.3 and earler have a bug which causes them to require
+ from version 1.6.3 and earlier have a bug which causes them to require
this information for some operations. (NAF - 2010/09/21 - Bug 1864)
- Fixed a bug where the library could generate an assertion/core dump when
a file that had been created with H5Pset_libver_bounds(fapl,
@@ -11826,7 +11826,7 @@ Bug Fixes since HDF5-1.8.3
created files exhibiting this problem should now be readable.
NAF - 2009/10/27 - 1483
- Fixed error in library's internal caching mechanisms which could cause
- an assertion failure (and attendent core dump) when encountering an
+ an assertion failure (and attendant core dump) when encountering an
unusually formatted file. (QAK - 2009/10/13)
- Fixed incorrect return value for H5Pget_preserve. AKC - 2009/10/08 - 1628
- Fixed an assertion failure that occurred when H5Ocopy was called on a
@@ -12350,7 +12350,7 @@ Bug Fixes since HDF5-1.8.2
can only be successfully opened with write access. NAF - 2009/03/23 - 1189
- Removed the long_long #define and replaced all instances with
"long long". This caused problems with third party products. All
- currently supported compliers support the type. ADB - 2009/03/05
+ currently supported compilers support the type. ADB - 2009/03/05
- Fixed various bugs that could prevent the fill value from being written
in certain rare cases. NAF - 2009/02/26 - 1469
- Fixed a bug that prevented more than one dataset chunk from being cached
@@ -13488,7 +13488,7 @@ Bug Fixes since HDF5-1.8.0
- The h5tset(get)_fields subroutines were missing the parameter to
specify a sign position; fixed. (EIP - 2008/05/23)
- Many APIs were fixed to work with the 8-byte integers in Fortran vs.
- 4-byte integers in C. This change is trasparent to user applications.
+ 4-byte integers in C. This change is transparent to user applications.
C++ API
------
@@ -13882,7 +13882,7 @@ Known Problems
exit $RETURN_VALUE
- You get the HDF5 make files and test scipts to execute your filter script
+ You get the HDF5 make files and test scripts to execute your filter script
by setting the environment variable "RUNSERIAL" to the full path of the
script prior to running configure for parallel builds. Remember to
"unsetenv RUNSERIAL" before running configure for a serial build.
@@ -14419,7 +14419,7 @@ Known Problems
exit $RETURN_VALUE
- You get the HDF5 make files and test scipts to execute your filter script
+ You get the HDF5 make files and test scripts to execute your filter script
by setting the environment variable "RUNSERIAL" to the full path of the
script prior to running configure for parallel builds. Remember to
"unsetenv RUNSERIAL" before running configure for a serial build.
diff --git a/release_docs/HISTORY-1_8_0-1_10_0.txt b/release_docs/HISTORY-1_8_0-1_10_0.txt
index 616aa24..575d070 100644
--- a/release_docs/HISTORY-1_8_0-1_10_0.txt
+++ b/release_docs/HISTORY-1_8_0-1_10_0.txt
@@ -211,7 +211,7 @@ New Features
Parallel Library:
-----------------
- - Add H5Pget_mpio_no_collective_cause() function that retrive reasons
+ - Add H5Pget_mpio_no_collective_cause() function that retrieve reasons
why the collective I/O was broken during read/write IO access.
(JKM - 2012/08/30 HDFFV-8143)
- Special Collective IO (IO when some processes do not contribute to the
@@ -331,10 +331,10 @@ New Features
- h5repack: Improved performance for big chunked datasets (size > 128MB)
when used with layout (-l) or compression (-f) option.
It would perform much better prior to the improvement,
- especially for cases that chunk dimentions looks like
+ especially for cases that chunk dimensions looks like
"1024x5x1" (compare to "1x5x1024"). When bigger numbers
are toward front and smaller number is toward back in chunk
- dimentions. HDFFV-7862 (JKM - 2012/03/01)
+ dimensions. HDFFV-7862 (JKM - 2012/03/01)
- h5dump: Added new option --no-compact-subset. This option will not
interpret the '[' character as starting the compact form of
subsetting. This is useful when the "h5dump error: unable to
@@ -564,7 +564,7 @@ Bug Fixes since HDF5-1.8.0 release
- Fixed rare corruption bugs that could occur when using the new object
header format. (NAF - 2012/3/15 - HDFFV-7879)
- Creating a dataset in a read-only file caused seg fault when the file
- is closed. It's fixed. The attemp to create a dataset will fail
+ is closed. It's fixed. The attempt to create a dataset will fail
with the error stack indicating the file is read-only. (SLU -
2012/1/25. Issue 7756)
- Fixed a seg fault that could occur when shrinking a dataset with chunks
@@ -700,7 +700,7 @@ Bug Fixes since HDF5-1.8.0 release
(NAF - 2009/03/23 - 1189)
- Removed the long_long #define and replaced all instances with
"long long". This caused problems with third party products. All
- currently supported compliers support the type. (ADB - 2009/03/05)
+ currently supported compilers support the type. (ADB - 2009/03/05)
- Fixed various bugs that could prevent the fill value from being written
in certain rare cases. (NAF - 2009/02/26 - 1469)
- Fixed a bug that prevented more than one dataset chunk from being cached
@@ -995,7 +995,7 @@ Bug Fixes since HDF5-1.8.0 release
- h5ls: fixed segfault when access region reference data in an
attribute. HDFFV-7838 (JKM 2011/12/29)
- h5diff: fixed segfault over non-comparable attribute with different
- dimention or rank, along with '-c' option to display details.
+ dimension or rank, along with '-c' option to display details.
HDFFV-7770 (JKM 2011/10/24)
- Fixed h5diff to display all the comparable object and attribute
regardless of non-comparables. HDFFV-7693 (JKM 2011/09/16)
@@ -1025,7 +1025,7 @@ Bug Fixes since HDF5-1.8.0 release
when mutually exclusive options (-d, -p and --use-system-epsilon)
are used together. HDFFV-7600 (JKM 2011/07/07)
- Fixed h5dump to display the first line of each element into correct
- position for multiple dimention array type.
+ position for multiple dimension array type.
Before this fix, the first line of each element in array were
displayed after the last line of previous element without
moving to the next line (+indentation).
@@ -1077,7 +1077,7 @@ Bug Fixes since HDF5-1.8.0 release
as source file. (XCAO 2011/3/8)
- Fixed h5dump for skipping some values for long array type dataset on
Windows. This issue only occurred on Windows due to the different
- return behavior from _vsnprintf() funtion. Bug#2161 (JKM 2011/3/3)
+ return behavior from _vsnprintf() function. Bug#2161 (JKM 2011/3/3)
- Fixed h5dump for skipping array indices every certain number
when the array type dataset is relatively big. The certain number
varies according to the size of array. Bug#2092 (JKM 2011/2/15).
@@ -1202,7 +1202,7 @@ Bug Fixes since HDF5-1.8.0 release
------
- Lite: The h5ltget_attribute_string_f used to return the C NULL character in the
- returned character buffer. The returned Fortran charactor buffer now does
+ returned character buffer. The returned Fortran character buffer now does
not return the C NULL character. (MSB - 2012/3/23)
- Lite: The h5ltget_dataset_info_f function (gets information about a dataset)
was not correctly returning the dimension array. (PVN - 2009/3/23)
@@ -1262,7 +1262,7 @@ Bug Fixes since HDF5-1.8.0 release
The strings command of Mac inspects library files and older
versions of strings may not know newer library format, resulting
in errors. Fixed by sending the library file as stdin to the strings
- coommand to avoid this problem. (AKC - 2013/03/08 HDFFV-8305)
+ command to avoid this problem. (AKC - 2013/03/08 HDFFV-8305)
- Fixed a typo in the ERROR macro in test/testhdf5.h. It segmentation
faulted when used before. (AKC - 2013/02/12 HDFFV-8267)
@@ -1535,8 +1535,8 @@ Known Problems
of GCC. (Issue 8017. SLU - 2012/6/12)
* The data conversion test dt_arith.c fails in "long double" to integer
- conversion on Ubuntu 11.10 (3.0.0.13 kernal) with GCC 4.6.1 if the library
- is built with optimization -O3 or -O2. The older GCC (4.5) or newer kernal
+ conversion on Ubuntu 11.10 (3.0.0.13 kernel) with GCC 4.6.1 if the library
+ is built with optimization -O3 or -O2. The older GCC (4.5) or newer kernel
(3.2.2 on Fedora) doesn't have the problem. Users should lower down the
optimization level (-O1 or -O0) by defining CFLAGS in the command line of
"configure" like:
@@ -1719,7 +1719,7 @@ Known Problems
exit $RETURN_VALUE
- You get the HDF make files and test scipts to execute your filter script
+ You get the HDF make files and test scripts to execute your filter script
by setting the environment variable "RUNSERIAL" to the full path of the
script prior to running configure for parallel builds. Remember to
"unsetenv RUNSERIAL" before running configure for a serial build.
diff --git a/release_docs/INSTALL_Cygwin.txt b/release_docs/INSTALL_Cygwin.txt
index eebffba..34d3e5a 100644
--- a/release_docs/INSTALL_Cygwin.txt
+++ b/release_docs/INSTALL_Cygwin.txt
@@ -2,7 +2,7 @@
HDF5 Build and Install Instructions for Cygwin
************************************************************************
-This document is a instruction on how to build, test and install HDF5 libary on
+This document is a instruction on how to build, test and install HDF5 library on
Cygwin. See detailed information in hdf5/INSTALL.
NOTE: hdf5 can be built with CMake, see the INSTALL_CMake.txt file for more guidance.
@@ -66,7 +66,7 @@ Preconditions:
2.2.2 Szip
The HDF5 library has a predefined compression filter that uses
the extended-Rice lossless compression algorithm for chunked
- datatsets. For more information on Szip compression, license terms,
+ datasets. For more information on Szip compression, license terms,
and obtaining the Szip source code, see:
https://portal.hdfgroup.org/display/HDF5/Szip+Compression+in+HDF+Products
diff --git a/release_docs/RELEASE.txt b/release_docs/RELEASE.txt
index 017d750..a828798 100644
--- a/release_docs/RELEASE.txt
+++ b/release_docs/RELEASE.txt
@@ -556,7 +556,7 @@ New Features
void **output);
This newly-added "ctl" callback allows Virtual File
- Drivers to intercept and handle arbitary operations
+ Drivers to intercept and handle arbitrary operations
identified by an operation code. Its parameters are
as follows:
@@ -1118,7 +1118,7 @@ Bug Fixes since HDF5-1.12.0 release
dimension sizes.
The problem was fixed by ensuring that decoding the dimension sizes
- and max values will not go beyong the end of the buffer.
+ and max values will not go beyond the end of the buffer.
(BMR - 2021/05/12, HDFFV-11223)
@@ -1324,7 +1324,7 @@ Bug Fixes since HDF5-1.12.0 release
- Fixed CMake C++ compiler flags
- A recent refactoring of the C++ configure files accidently removed the
+ A recent refactoring of the C++ configure files accidentally removed the
file that executed the enable_language command for C++ needed by the
HDFCXXCompilerFlags.cmake file. Also updated the intel warnings files,
including adding support for windows platforms.
@@ -1452,7 +1452,7 @@ Bug Fixes since HDF5-1.12.0 release
- h5diff fixed a command line parsing error.
h5diff would ignore the argument to -d (delta) if it is smaller than DBL_EPSILON.
- The macro H5_DBL_ABS_EQUAL was removed and a direct value comparision was used.
+ The macro H5_DBL_ABS_EQUAL was removed and a direct value comparison was used.
(ADB - 2020/07/20, HDFFV-10897)
diff --git a/src/H5A.c b/src/H5A.c
index c986d1f..b5cc77a 100644
--- a/src/H5A.c
+++ b/src/H5A.c
@@ -1868,7 +1868,7 @@ done:
if all attributes were processed.
DESCRIPTION
- This function interates over the attributes of dataset or group
+ This function iterates over the attributes of dataset or group
specified with 'loc_id' & 'obj_name'. For each attribute of the object,
the 'op_data' and some additional information (specified below) are passed
to the 'op' function. The iteration begins with the '*idx'
@@ -1956,7 +1956,7 @@ done:
if all attributes were processed.
DESCRIPTION
- This function interates over the attributes of dataset or group
+ This function iterates over the attributes of dataset or group
specified with 'loc_id' & 'obj_name'. For each attribute of the object,
the 'op_data' and some additional information (specified below) are passed
to the 'op' function. The iteration begins with the '*idx'
@@ -2410,7 +2410,7 @@ done:
htri_t
H5Aexists(hid_t obj_id, const char *attr_name)
{
- hbool_t exists; /* Flag for attribute existance */
+ hbool_t exists; /* Flag for attribute existence */
htri_t ret_value = FAIL; /* Return value */
FUNC_ENTER_API(FAIL)
@@ -2526,7 +2526,7 @@ done:
htri_t
H5Aexists_by_name(hid_t loc_id, const char *obj_name, const char *attr_name, hid_t lapl_id)
{
- hbool_t exists; /* Flag for attribute existance */
+ hbool_t exists; /* Flag for attribute existence */
htri_t ret_value = FAIL; /* Return value */
FUNC_ENTER_API(FAIL)
diff --git a/src/H5AC.c b/src/H5AC.c
index e727a9e..6fbc63e 100644
--- a/src/H5AC.c
+++ b/src/H5AC.c
@@ -1199,7 +1199,7 @@ done:
* metadata cache flush.
*
* Initially, this means setting up the slist prior to the
- * flush. We do this in a seperate call because
+ * flush. We do this in a separate call because
* H5F__flush_phase2() make repeated calls to H5AC_flush().
* Handling this detail in separate calls allows us to avoid
* the overhead of setting up and taking down the skip list
@@ -1251,7 +1251,7 @@ done:
* flush.
*
* Initially, this means taking down the slist after the
- * flush. We do this in a seperate call because
+ * flush. We do this in a separate call because
* H5F__flush_phase2() make repeated calls to H5AC_flush().
* Handling this detail in separate calls allows us to avoid
* the overhead of setting up and taking down the skip list
diff --git a/src/H5ACdbg.c b/src/H5ACdbg.c
index 2d24adb..5e24517 100644
--- a/src/H5ACdbg.c
+++ b/src/H5ACdbg.c
@@ -197,9 +197,9 @@ done:
* If either the parent or the child is not in the metadata
* cache, the function sets *fd_exists_ptr to FALSE.
*
- * If both are in the cache, the childs list of parents is
+ * If both are in the cache, the child's list of parents is
* searched for the proposed parent. If the proposed parent
- * is found in the childs parent list, the function sets
+ * is found in the child's parent list, the function sets
* *fd_exists_ptr to TRUE. In all other non-error cases,
* the function sets *fd_exists_ptr FALSE.
*
diff --git a/src/H5ACmpio.c b/src/H5ACmpio.c
index 7d2ba25..500a05a 100644
--- a/src/H5ACmpio.c
+++ b/src/H5ACmpio.c
@@ -1132,7 +1132,7 @@ done:
*
* While the list of candidate cache entries is prepared
* elsewhere, this function is the main routine for distributing
- * and applying the list. It must be run simultaniously on
+ * and applying the list. It must be run simultaneously on
* all processes that have the relevant file open. To ensure
* proper synchronization, there is a barrier at the beginning
* of this function.
@@ -1311,7 +1311,7 @@ done:
*
* This function is the main routine for handling this
* notification procedure. It must be called
- * simultaniously on all processes that have the relevant
+ * simultaneously on all processes that have the relevant
* file open. To this end, it is called only during a
* sync point, with a barrier prior to the call.
*
diff --git a/src/H5ACpkg.h b/src/H5ACpkg.h
index 521f85a..9b75034 100644
--- a/src/H5ACpkg.h
+++ b/src/H5ACpkg.h
@@ -346,7 +346,7 @@ H5FL_EXTERN(H5AC_aux_t);
*
* The following field supports the metadata cache image feature.
*
- * p0_image_len: unsiged integer containing the length of the metadata cache
+ * p0_image_len: unsigned integer containing the length of the metadata cache
* image constructed by MPI process 0. This field should be 0
* if the value is unknown, or if cache image is not enabled.
*
diff --git a/src/H5ACpublic.h b/src/H5ACpublic.h
index f8f4f28..c853794 100644
--- a/src/H5ACpublic.h
+++ b/src/H5ACpublic.h
@@ -76,7 +76,7 @@ extern "C" {
*
* *** DEPRECATED *** Use H5Fstart/stop logging functions instead
*
- * The trace file is a debuging feature that allow the capture of
+ * The trace file is a debugging feature that allow the capture of
* top level metadata cache requests for purposes of debugging and/or
* optimization. This field should normally be set to FALSE, as
* trace file collection imposes considerable overhead.
@@ -123,7 +123,7 @@ extern "C" {
* H5C_incr__off ) && ( decr_mode == H5C_decr__off )). There
* is no logical reason why this should be so, but it simplifies
* implementation and testing, and I can't think of any reason
- * why it would be desireable. If you can think of one, I'll
+ * why it would be desirable. If you can think of one, I'll
* revisit the issue.
*
* set_initial_size: Boolean flag indicating whether the size of the
@@ -396,7 +396,7 @@ extern "C" {
*
* When the sync point is reached (or when there is a user generated
* flush), process zero flushes sufficient entries to bring it into
- * complience with its min clean size (or flushes all dirty entries in
+ * compliance with its min clean size (or flushes all dirty entries in
* the case of a user generated flush), broad casts the list of
* entries just cleaned to all the other processes, and then exits
* the sync point.
@@ -576,7 +576,7 @@ typedef struct H5AC_cache_config_t {
size_t min_size;
/**< Lower bound (in bytes) on the range of values that the
- * adaptive cache resize code can select as the mininum cache * size. */
+ * adaptive cache resize code can select as the minimum cache * size. */
long int epoch_length;
/**< Number of cache accesses between runs of the adaptive cache resize
@@ -708,13 +708,13 @@ typedef struct H5AC_cache_config_t {
* of bytes of dirty metadata created since the last synchronization exceeds
* this limit.\n This field only applies to the parallel case. While it is
* ignored elsewhere, it can still draw a value out of bounds error.\n It
- * must be consistant across all caches on any given file.\n By default,
+ * must be consistent across all caches on any given file.\n By default,
* this field is set to 256 KB. It shouldn't be more than half the current
* max cache size times the min clean fraction. */
int metadata_write_strategy;
/**< Desired metadata write strategy. The valid values for this field
- * are:\n #H5AC_METADATA_WRITE_STRATEGY__PROCESS_0_ONLY: Specifies tha only
+ * are:\n #H5AC_METADATA_WRITE_STRATEGY__PROCESS_0_ONLY: Specifies the only
* process zero is allowed to write dirty metadata to disk.\n
* #H5AC_METADATA_WRITE_STRATEGY__DISTRIBUTED: Specifies that process zero
* still makes the decisions as to what entries should be flushed, but the
diff --git a/src/H5Adense.c b/src/H5Adense.c
index b0d793c..fa51f9d 100644
--- a/src/H5Adense.c
+++ b/src/H5Adense.c
@@ -337,7 +337,7 @@ H5A__dense_open(H5F_t *f, const H5O_ainfo_t *ainfo, const char *name)
H5HF_t * fheap = NULL; /* Fractal heap handle */
H5HF_t * shared_fheap = NULL; /* Fractal heap handle for shared header messages */
H5B2_t * bt2_name = NULL; /* v2 B-tree handle for name index */
- htri_t attr_sharable; /* Flag indicating attributes are sharable */
+ htri_t attr_sharable; /* Flag indicating attributes are shareable */
hbool_t attr_exists; /* Attribute exists in v2 B-tree */
H5A_t * ret_value = NULL; /* Return value */
@@ -356,7 +356,7 @@ H5A__dense_open(H5F_t *f, const H5O_ainfo_t *ainfo, const char *name)
if ((attr_sharable = H5SM_type_shared(f, H5O_ATTR_ID)) < 0)
HGOTO_ERROR(H5E_ATTR, H5E_CANTGET, NULL, "can't determine if attributes are shared")
- /* Get handle for shared message heap, if attributes are sharable */
+ /* Get handle for shared message heap, if attributes are shareable */
if (attr_sharable) {
haddr_t shared_fheap_addr; /* Address of fractal heap to use */
@@ -429,7 +429,7 @@ H5A__dense_insert(H5F_t *f, const H5O_ainfo_t *ainfo, H5A_t *attr)
H5WB_t * wb = NULL; /* Wrapped buffer for attribute data */
uint8_t attr_buf[H5A_ATTR_BUF_SIZE]; /* Buffer for serializing message */
unsigned mesg_flags = 0; /* Flags for storing message */
- htri_t attr_sharable; /* Flag indicating attributes are sharable */
+ htri_t attr_sharable; /* Flag indicating attributes are shareable */
herr_t ret_value = SUCCEED; /* Return value */
FUNC_ENTER_PACKAGE
@@ -443,7 +443,7 @@ H5A__dense_insert(H5F_t *f, const H5O_ainfo_t *ainfo, H5A_t *attr)
if ((attr_sharable = H5SM_type_shared(f, H5O_ATTR_ID)) < 0)
HGOTO_ERROR(H5E_ATTR, H5E_CANTGET, FAIL, "can't determine if attributes are shared")
- /* Get handle for shared message heap, if attributes are sharable */
+ /* Get handle for shared message heap, if attributes are shareable */
if (attr_sharable) {
haddr_t shared_fheap_addr; /* Address of fractal heap to use */
htri_t shared_mesg; /* Should this message be stored in the Shared Message table? */
@@ -727,7 +727,7 @@ H5A__dense_write(H5F_t *f, const H5O_ainfo_t *ainfo, H5A_t *attr)
H5HF_t * fheap = NULL; /* Fractal heap handle */
H5HF_t * shared_fheap = NULL; /* Fractal heap handle for shared header messages */
H5B2_t * bt2_name = NULL; /* v2 B-tree handle for name index */
- htri_t attr_sharable; /* Flag indicating attributes are sharable */
+ htri_t attr_sharable; /* Flag indicating attributes are shareable */
herr_t ret_value = SUCCEED; /* Return value */
FUNC_ENTER_PACKAGE
@@ -743,7 +743,7 @@ H5A__dense_write(H5F_t *f, const H5O_ainfo_t *ainfo, H5A_t *attr)
if ((attr_sharable = H5SM_type_shared(f, H5O_ATTR_ID)) < 0)
HGOTO_ERROR(H5E_ATTR, H5E_CANTGET, FAIL, "can't determine if attributes are shared")
- /* Get handle for shared message heap, if attributes are sharable */
+ /* Get handle for shared message heap, if attributes are shareable */
if (attr_sharable) {
haddr_t shared_fheap_addr; /* Address of fractal heap to use */
@@ -865,7 +865,7 @@ H5A__dense_rename(H5F_t *f, const H5O_ainfo_t *ainfo, const char *old_name, cons
H5B2_t * bt2_name = NULL; /* v2 B-tree handle for name index */
H5B2_t * bt2_corder = NULL; /* v2 B-tree handle for creation order ndex */
H5A_t * attr_copy = NULL; /* Copy of attribute to rename */
- htri_t attr_sharable; /* Flag indicating attributes are sharable */
+ htri_t attr_sharable; /* Flag indicating attributes are shareable */
htri_t shared_mesg; /* Should this message be stored in the Shared Message table? */
hbool_t attr_exists; /* Attribute exists in v2 B-tree */
herr_t ret_value = SUCCEED; /* Return value */
@@ -882,7 +882,7 @@ H5A__dense_rename(H5F_t *f, const H5O_ainfo_t *ainfo, const char *old_name, cons
if ((attr_sharable = H5SM_type_shared(f, H5O_ATTR_ID)) < 0)
HGOTO_ERROR(H5E_ATTR, H5E_CANTGET, FAIL, "can't determine if attributes are shared")
- /* Get handle for shared message heap, if attributes are sharable */
+ /* Get handle for shared message heap, if attributes are shareable */
if (attr_sharable) {
haddr_t shared_fheap_addr; /* Address of fractal heap to use */
@@ -1181,7 +1181,7 @@ H5A__dense_iterate(H5F_t *f, hid_t loc_id, const H5O_ainfo_t *ainfo, H5_index_t
/* Check on iteration order */
if (order == H5_ITER_NATIVE && H5F_addr_defined(bt2_addr)) {
H5A_bt2_ud_it_t udata; /* User data for iterator callback */
- htri_t attr_sharable; /* Flag indicating attributes are sharable */
+ htri_t attr_sharable; /* Flag indicating attributes are shareable */
/* Open the fractal heap */
if (NULL == (fheap = H5HF_open(f, ainfo->fheap_addr)))
@@ -1191,7 +1191,7 @@ H5A__dense_iterate(H5F_t *f, hid_t loc_id, const H5O_ainfo_t *ainfo, H5_index_t
if ((attr_sharable = H5SM_type_shared(f, H5O_ATTR_ID)) < 0)
HGOTO_ERROR(H5E_ATTR, H5E_CANTGET, FAIL, "can't determine if attributes are shared")
- /* Get handle for shared message heap, if attributes are sharable */
+ /* Get handle for shared message heap, if attributes are shareable */
if (attr_sharable) {
haddr_t shared_fheap_addr; /* Address of fractal heap to use */
@@ -1338,7 +1338,7 @@ H5A__dense_remove(H5F_t *f, const H5O_ainfo_t *ainfo, const char *name)
H5HF_t * shared_fheap = NULL; /* Fractal heap handle for shared header messages */
H5B2_t * bt2_name = NULL; /* v2 B-tree handle for name index */
H5A_t * attr_copy = NULL; /* Copy of attribute to remove */
- htri_t attr_sharable; /* Flag indicating attributes are sharable */
+ htri_t attr_sharable; /* Flag indicating attributes are shareable */
herr_t ret_value = SUCCEED; /* Return value */
FUNC_ENTER_PACKAGE
@@ -1356,7 +1356,7 @@ H5A__dense_remove(H5F_t *f, const H5O_ainfo_t *ainfo, const char *name)
if ((attr_sharable = H5SM_type_shared(f, H5O_ATTR_ID)) < 0)
HGOTO_ERROR(H5E_ATTR, H5E_CANTGET, FAIL, "can't determine if attributes are shared")
- /* Get handle for shared message heap, if attributes are sharable */
+ /* Get handle for shared message heap, if attributes are shareable */
if (attr_sharable) {
haddr_t shared_fheap_addr; /* Address of fractal heap to use */
@@ -1585,7 +1585,7 @@ H5A__dense_remove_by_idx(H5F_t *f, const H5O_ainfo_t *ainfo, H5_index_t idx_type
/* If there is an index defined for the field, use it */
if (H5F_addr_defined(bt2_addr)) {
H5A_bt2_ud_rmbi_t udata; /* User data for v2 B-tree record removal */
- htri_t attr_sharable; /* Flag indicating attributes are sharable */
+ htri_t attr_sharable; /* Flag indicating attributes are shareable */
/* Open the fractal heap */
if (NULL == (fheap = H5HF_open(f, ainfo->fheap_addr)))
@@ -1595,7 +1595,7 @@ H5A__dense_remove_by_idx(H5F_t *f, const H5O_ainfo_t *ainfo, H5_index_t idx_type
if ((attr_sharable = H5SM_type_shared(f, H5O_ATTR_ID)) < 0)
HGOTO_ERROR(H5E_ATTR, H5E_CANTGET, FAIL, "can't determine if attributes are shared")
- /* Get handle for shared message heap, if attributes are sharable */
+ /* Get handle for shared message heap, if attributes are shareable */
if (attr_sharable) {
haddr_t shared_fheap_addr; /* Address of fractal heap to use */
@@ -1675,7 +1675,7 @@ H5A__dense_exists(H5F_t *f, const H5O_ainfo_t *ainfo, const char *name, hbool_t
H5HF_t * fheap = NULL; /* Fractal heap handle */
H5HF_t * shared_fheap = NULL; /* Fractal heap handle for shared header messages */
H5B2_t * bt2_name = NULL; /* v2 B-tree handle for name index */
- htri_t attr_sharable; /* Flag indicating attributes are sharable */
+ htri_t attr_sharable; /* Flag indicating attributes are shareable */
herr_t ret_value = SUCCEED; /* Return value */
FUNC_ENTER_PACKAGE
@@ -1694,7 +1694,7 @@ H5A__dense_exists(H5F_t *f, const H5O_ainfo_t *ainfo, const char *name, hbool_t
if ((attr_sharable = H5SM_type_shared(f, H5O_ATTR_ID)) < 0)
HGOTO_ERROR(H5E_ATTR, H5E_CANTGET, FAIL, "can't determine if attributes are shared")
- /* Get handle for shared message heap, if attributes are sharable */
+ /* Get handle for shared message heap, if attributes are shareable */
if (attr_sharable) {
haddr_t shared_fheap_addr; /* Address of fractal heap to use */
diff --git a/src/H5Adeprec.c b/src/H5Adeprec.c
index 8ae4e41..05e4eed 100644
--- a/src/H5Adeprec.c
+++ b/src/H5Adeprec.c
@@ -356,7 +356,7 @@ done:
last operator if it was non-zero, or zero if all attributes were processed.
DESCRIPTION
- This function interates over the attributes of dataset or group
+ This function iterates over the attributes of dataset or group
specified with 'loc_id'. For each attribute of the object, the
'op_data' and some additional information (specified below) are passed
to the 'op' function. The iteration begins with the '*attr_number'
diff --git a/src/H5Apublic.h b/src/H5Apublic.h
index b78ae05..4ac6a53 100644
--- a/src/H5Apublic.h
+++ b/src/H5Apublic.h
@@ -622,7 +622,7 @@ H5_DLL hid_t H5Aget_type(hid_t attr_id);
* For example, if \p idx_type, \p order, and \p idx are set to
* #H5_INDEX_NAME, #H5_ITER_INC, and 5, respectively, the attribute
* in question is the fifth attribute from the beginning of the
- * alpha-numeric index of attribute names. If \p order were set to
+ * alphanumeric index of attribute names. If \p order were set to
* #H5_ITER_DEC, it would be the fifth attribute from the end of
* the index.
*
@@ -677,7 +677,7 @@ H5_DLL herr_t H5Aiterate2(hid_t loc_id, H5_index_t idx_type, H5_iter_order_t ord
* For example, if \p idx_type, \p order, and \p idx are set to
* #H5_INDEX_NAME, #H5_ITER_INC, and 5, respectively, the attribute
* in question is the fifth attribute from the beginning of the
- * alpha-numeric index of attribute names. If \p order were set to
+ * alphanumeric index of attribute names. If \p order were set to
* #H5_ITER_DEC, it would be the fifth attribute from the end of
* the index.
*
diff --git a/src/H5C.c b/src/H5C.c
index 275afea..09c19ce 100644
--- a/src/H5C.c
+++ b/src/H5C.c
@@ -778,7 +778,7 @@ H5C_prep_for_file_close(H5F_t *f)
*
* 2) Since the FSM settle routines are only invoked once during
* file close, invoking them now will prevent their invocation
- * during a flush, and thus avoid any resulting entrie dirties,
+ * during a flush, and thus avoid any resulting entry dirties,
* deletions, insertion, or moves during the flush.
*/
if (H5C__serialize_cache(f) < 0)
@@ -1484,7 +1484,7 @@ H5C_insert_entry(H5F_t *f, const H5C_class_t *type, haddr_t addr, void *thing, u
* Finally, we usually don't check to see if the cache is
* oversized at the end of an unprotect. As a result, it is
* possible to have a vastly oversized cache with no protected
- * entries as long as all the protects preceed the unprotects.
+ * entries as long as all the protects proceed the unprotects.
*
* Since items 1 and 2 are not changing any time soon, I see
* no point in worrying about the third.
@@ -2444,7 +2444,7 @@ H5C_protect(H5F_t *f, const H5C_class_t *type, haddr_t addr, void *udata, unsign
* Finally, we usually don't check to see if the cache is
* oversized at the end of an unprotect. As a result, it is
* possible to have a vastly oversized cache with no protected
- * entries as long as all the protects preceed the unprotects.
+ * entries as long as all the protects proceed the unprotects.
*
* Since items 1, 2, and 3 are not changing any time soon, I
* see no point in worrying about the fourth.
@@ -2559,7 +2559,7 @@ H5C_protect(H5F_t *f, const H5C_class_t *type, haddr_t addr, void *udata, unsign
*
* Also, if the min_clean_size requirement is not met, we
* should also call H5C__make_space_in_cache() to bring us
- * into complience.
+ * into compliance.
*/
if (cache_ptr->index_size >= cache_ptr->max_cache_size)
@@ -4848,7 +4848,7 @@ H5C__autoadjust__ageout__evict_aged_out_entries(H5F_t *f, hbool_t write_permitte
/* for now at least, don't bother to maintain the minimum clean size,
* as the cache should now be less than its maximum size. Due to
- * the vaguries of the cache size reduction algorthim, we may not
+ * the vaguries of the cache size reduction algorithm, we may not
* reduce the size of the cache.
*
* If we do, we will calculate a new minimum clean size, which will
@@ -5860,7 +5860,7 @@ H5C__flush_invalidate_ring(H5F_t *f, H5C_ring_t ring, unsigned flags)
* If either of these happen, and one of the target
* or proxy entries happens to be the next entry in
* the hash bucket, we could either find ourselves
- * either scanning a non-existant entry, scanning
+ * either scanning a non-existent entry, scanning
* through a different bucket, or skipping an entry.
*
* Neither of these are good, so restart the
@@ -8059,7 +8059,7 @@ done:
* To cork the object
* Return error if the object is already corked
* H5C__UNCORK:
- * To uncork the obejct
+ * To uncork the object
* Return error if the object is not corked
* H5C__GET_CORKED:
* To retrieve the cork status of an object in
@@ -8490,7 +8490,7 @@ H5C__serialize_cache(H5F_t *f)
* each entry in the cache to zero before we start the serialization.
* This allows us to detect the case in which any entry is serialized
* more than once (a performance issues), and more importantly, the
- * case is which any flush depencency parent is serializes more than
+ * case is which any flush dependency parent is serializes more than
* once (a correctness issue).
*/
{
@@ -8647,8 +8647,8 @@ H5C__serialize_ring(H5F_t *f, H5C_ring_t ring)
* are serialized correctly, it doesn't matter if we have to go back
* and serialize an entry a second time.
*
- * These possible actions result in the following modfications to
- * tha basic algorithm:
+ * These possible actions result in the following modifications to
+ * the basic algorithm:
*
* 1) In the event of an entry expunge, eviction or removal, we must
* restart the scan as it is possible that the next entry in our
@@ -8698,7 +8698,7 @@ H5C__serialize_ring(H5F_t *f, H5C_ring_t ring)
* condition appears.
*
* Observe that either eviction or removal of entries as a result of
- * a serialization is not a problem as long as the flush depencency
+ * a serialization is not a problem as long as the flush dependency
* tree does not change beyond the removal of a leaf.
*/
while (!done) {
diff --git a/src/H5CS.c b/src/H5CS.c
index d7cb6f1..0fabc0d 100644
--- a/src/H5CS.c
+++ b/src/H5CS.c
@@ -47,7 +47,7 @@ typedef struct H5CS_t {
* each thread individually. The association of stacks to threads will
* be handled by the pthread library.
*
- * In order for this macro to work, H5CS_get_my_stack() must be preceeded
+ * In order for this macro to work, H5CS_get_my_stack() must be preceded
* by "H5CS_t *fstack =".
*/
static H5CS_t *H5CS__get_stack(void);
@@ -127,7 +127,7 @@ H5CS__get_stack(void)
herr_t
H5CS_print_stack(const H5CS_t *fstack, FILE *stream)
{
- const int indent = 2; /* Indention level */
+ const int indent = 2; /* Indentation level */
int i; /* Local index ariable */
/* Don't push this function on the function stack... :-) */
diff --git a/src/H5CX.c b/src/H5CX.c
index 7c99055..b9ce682 100644
--- a/src/H5CX.c
+++ b/src/H5CX.c
@@ -50,7 +50,7 @@
* each thread individually. The association of contexts to threads will
* be handled by the pthread library.
*
- * In order for this macro to work, H5CX_get_my_context() must be preceeded
+ * In order for this macro to work, H5CX_get_my_context() must be preceded
* by "H5CX_node_t *ctx =".
*/
#define H5CX_get_my_context() H5CX__get_context()
diff --git a/src/H5Cdbg.c b/src/H5Cdbg.c
index 104f1af..de9019c 100644
--- a/src/H5Cdbg.c
+++ b/src/H5Cdbg.c
@@ -844,9 +844,9 @@ H5C_stats__reset(H5C_t H5_ATTR_UNUSED *cache_ptr)
* If either the parent or the child is not in the metadata
* cache, the function sets *fd_exists_ptr to FALSE.
*
- * If both are in the cache, the childs list of parents is
+ * If both are in the cache, the child's list of parents is
* searched for the proposed parent. If the proposed parent
- * is found in the childs parent list, the function sets
+ * is found in the child's parent list, the function sets
* *fd_exists_ptr to TRUE. In all other non-error cases,
* the function sets *fd_exists_ptr FALSE.
*
diff --git a/src/H5Cimage.c b/src/H5Cimage.c
index f17ff48..7421c90 100644
--- a/src/H5Cimage.c
+++ b/src/H5Cimage.c
@@ -591,7 +591,7 @@ H5C__deserialize_prefetched_entry(H5F_t *f, H5C_t *cache_ptr, H5C_cache_entry_t
*
* Note that at present, dirty can't be set to true with prefetched
* entries. However this may change, so include this functionality
- * against that posibility.
+ * against that possibility.
*
* Also, note that it is possible for a prefetched entry to be dirty --
* hence the value assigned to ds_entry_ptr->is_dirty below.
@@ -1129,7 +1129,7 @@ done:
* image superblock extension message must be deleted from
* the superblock extension and the image block freed
*
- * Contrawise, if the file is openened R/O, the metadata
+ * Contrawise, if the file is opened R/O, the metadata
* cache image superblock extension message and image block
* must be left as is. Further, any dirty entries in the
* cache image block must be marked as clean to avoid
@@ -1821,7 +1821,7 @@ done:
* Purpose: Decode the metadata cache image entry from the supplied
* buffer into the supplied instance of H5C_image_entry_t.
* This includes allocating a buffer for the entry image,
- * loading it, and seting ie_ptr->image_ptr to point to
+ * loading it, and setting ie_ptr->image_ptr to point to
* the buffer.
*
* Advances the buffer pointer to the first byte
@@ -2338,7 +2338,7 @@ done:
* also be a flush dependency child.
*
* Finally, note that for purposes of the cache image, flush
- * dependency height ends when a flush dependecy relation
+ * dependency height ends when a flush dependency relation
* passes off the cache image.
*
* On exit, the flush dependency height of each entry in the
@@ -2500,7 +2500,7 @@ H5C__prep_for_file_close__compute_fd_heights(const H5C_t *cache_ptr)
entry_ptr = entry_ptr->il_next;
} /* while (entry_ptr != NULL) */
- /* At present, no extenal parent or child flush dependency links
+ /* At present, no external parent or child flush dependency links
* should exist -- hence the following assertions. This will change
* if we support ageout of entries in the cache image.
*/
@@ -3381,7 +3381,7 @@ done:
* creating if specified.
*
* In general, the size and location of the cache image block
- * will be unknow at the time that the cache image superblock
+ * will be unknown at the time that the cache image superblock
* message is created. A subsequent call to this routine will
* be used to write the correct data.
*
diff --git a/src/H5Cmpio.c b/src/H5Cmpio.c
index 35d8dd0..d127961 100644
--- a/src/H5Cmpio.c
+++ b/src/H5Cmpio.c
@@ -1097,7 +1097,7 @@ done:
*
* Note that this function is a modified version of
* H5C_flush_cache() -- any changes there may need to be
- * reflected here and vise versa.
+ * reflected here and vice versa.
*
* Return: Non-negative on success/Negative on failure.
*
diff --git a/src/H5Cpkg.h b/src/H5Cpkg.h
index 99ed4f8..30b86b9 100644
--- a/src/H5Cpkg.h
+++ b/src/H5Cpkg.h
@@ -1597,7 +1597,7 @@ if ( ( (cache_ptr)->index_size != \
*
* H5C_DO_SLIST_SANITY_CHECKS
*
- * can be selected independantly. This is easy to miss as the
+ * can be selected independently. This is easy to miss as the
* two #defines are easy to confuse.
*/
diff --git a/src/H5Cprefetched.c b/src/H5Cprefetched.c
index 07fca2b..f89c233 100644
--- a/src/H5Cprefetched.c
+++ b/src/H5Cprefetched.c
@@ -302,7 +302,7 @@ H5C__prefetched_entry_free_icr(void *_thing)
HDassert(entry_ptr->fd_parent_count == 0);
if (entry_ptr->image_ptr != NULL)
- HGOTO_ERROR(H5E_CACHE, H5E_SYSTEM, FAIL, "prefetched entry image buffer still attatched?")
+ HGOTO_ERROR(H5E_CACHE, H5E_SYSTEM, FAIL, "prefetched entry image buffer still attached?")
entry_ptr = H5FL_FREE(H5C_cache_entry_t, entry_ptr);
diff --git a/src/H5Cprivate.h b/src/H5Cprivate.h
index 78d6f35..8a1043e 100644
--- a/src/H5Cprivate.h
+++ b/src/H5Cprivate.h
@@ -213,7 +213,7 @@
#define H5C_DO_TAGGING_SANITY_CHECKS 1
#define H5C_DO_EXTREME_SANITY_CHECKS 0
#else /* NDEBUG */
-/* With rare execptions, the following defines should be set
+/* With rare exceptions, the following defines should be set
* to 0 if NDEBUG is defined
*/
#define H5C_DO_SANITY_CHECKS 0
@@ -1428,7 +1428,7 @@ typedef int H5C_ring_t;
* with no flush dependency children.
*
* Since the image_fd_height is used to order entries in the
- * cache image so that fd parents preceed fd children, for
+ * cache image so that fd parents precede fd children, for
* purposes of this field, and entry is at flush dependency
* level 0 if it either has no children, or if all of its
* children are not in the cache image.
@@ -1543,7 +1543,7 @@ typedef int H5C_ring_t;
* number of times each entry is serialized during cache
* serialization. While no entry should be serialized more than
* once in any serialization call, throw an assertion if any
- * flush depencency parent is serialized more than once during
+ * flush dependency parent is serialized more than once during
* a single cache serialization.
*
* This is a debugging field, and thus is maintained only if
@@ -1734,7 +1734,7 @@ typedef struct H5C_cache_entry_t {
* with no flush dependency children.
*
* Since the image_fd_height is used to order entries in the
- * cache image so that fd parents preceed fd children, for
+ * cache image so that fd parents precede fd children, for
* purposes of this field, an entry is at flush dependency
* level 0 if it either has no children, or if all of its
* children are not in the cache image.
@@ -2177,7 +2177,7 @@ typedef struct H5C_auto_size_ctl_t {
* equivalent of H5AC__CACHE_IMAGE__ENTRY_AGEOUT__NONE.
*
* flags: Unsigned integer containing flags controlling which aspects of the
- * cache image functinality is actually executed. The primary impetus
+ * cache image functionality is actually executed. The primary impetus
* behind this field is to allow development of tests for partial
* implementations that will require little if any modification to run
* with the full implementation. In normal operation, all flags should
diff --git a/src/H5D.c b/src/H5D.c
index cfa8692..c10b079 100644
--- a/src/H5D.c
+++ b/src/H5D.c
@@ -1802,7 +1802,7 @@ H5Dset_extent(hid_t dset_id, const hsize_t size[])
FUNC_ENTER_API(FAIL)
H5TRACE2("e", "i*h", dset_id, size);
- /* Change a datset's dimenions synchronously */
+ /* Change a datset's dimensions synchronously */
if ((ret_value = H5D__set_extent_api_common(dset_id, size, NULL, NULL)) < 0)
HGOTO_ERROR(H5E_DATASET, H5E_CANTSET, FAIL, "unable to synchronously change a dataset's dimensions")
@@ -1835,7 +1835,7 @@ H5Dset_extent_async(const char *app_file, const char *app_func, unsigned app_lin
if (H5ES_NONE != es_id)
token_ptr = &token; /* Point at token for VOL connector to set up */
- /* Change a datset's dimenions asynchronously */
+ /* Change a datset's dimensions asynchronously */
if (H5D__set_extent_api_common(dset_id, size, token_ptr, &vol_obj) < 0)
HGOTO_ERROR(H5E_DATASET, H5E_CANTSET, FAIL, "unable to asynchronously change a dataset's dimensions")
diff --git a/src/H5Dchunk.c b/src/H5Dchunk.c
index fc5960e..9742688 100644
--- a/src/H5Dchunk.c
+++ b/src/H5Dchunk.c
@@ -974,7 +974,7 @@ H5D__chunk_init(H5F_t *f, const H5D_t *const dset, hid_t dapl_id)
if (!(scaled_power2up = H5VM_power2up(rdcc->scaled_dims[u])))
HGOTO_ERROR(H5E_DATASET, H5E_CANTGET, FAIL, "unable to get the next power of 2")
- /* Inital 'power2up' values for scaled dimensions */
+ /* Initial 'power2up' values for scaled dimensions */
rdcc->scaled_power2up[u] = scaled_power2up;
/* Number of bits required to encode scaled dimension size */
@@ -2550,7 +2550,7 @@ H5D__chunk_read(H5D_io_info_t *io_info, const H5D_type_info_t *type_info, hsize_
HDassert((H5F_addr_defined(udata.chunk_block.offset) && udata.chunk_block.length > 0) ||
(!H5F_addr_defined(udata.chunk_block.offset) && udata.chunk_block.length == 0));
- /* Check for non-existant chunk & skip it if appropriate */
+ /* Check for non-existent chunk & skip it if appropriate */
if (H5F_addr_defined(udata.chunk_block.offset) || UINT_MAX != udata.idx_hint ||
!skip_missing_chunks) {
H5D_io_info_t *chk_io_info; /* Pointer to I/O info object for this chunk */
@@ -3616,7 +3616,7 @@ H5D__chunk_cache_prune(const H5D_t *dset, size_t size)
* traversing the list when pointer pN reaches wN percent of the original
* list. In other words, preemption method N gets to consider entries in
* approximate least recently used order w0 percent before method N+1
- * where 100% means tha method N will run to completion before method N+1
+ * where 100% means the method N will run to completion before method N+1
* begins. The pointers participating in the list traversal are each
* given a chance at preemption before any of the pointers are advanced.
*/
@@ -5966,7 +5966,7 @@ H5D__chunk_copy_cb(const H5D_chunk_rec_t *chunk_rec, void *_udata)
size_t buf_size = udata->buf_size; /* Size of chunk buffer */
const H5O_pline_t *pline = udata->pline; /* I/O pipeline for applying filters */
- /* needed for commpressed variable length data */
+ /* needed for compressed variable length data */
hbool_t must_filter = FALSE; /* Whether chunk must be filtered during copy */
size_t nbytes; /* Size of chunk in file (in bytes) */
H5Z_cb_t filter_cb; /* Filter failure callback struct */
@@ -7526,7 +7526,7 @@ H5D__chunk_iter_cb(const H5D_chunk_rec_t *chunk_rec, void *udata)
/*-------------------------------------------------------------------------
* Function: H5D__chunk_iter
*
- * Purpose: Iterate over all the chunks in the dataset with given callbak.
+ * Purpose: Iterate over all the chunks in the dataset with given callback.
*
* Return: Success: Non-negative
* Failure: Negative
diff --git a/src/H5Dint.c b/src/H5Dint.c
index 8a09364..c9ea6bd 100644
--- a/src/H5Dint.c
+++ b/src/H5Dint.c
@@ -364,7 +364,7 @@ done:
*
* Return:
* Success: Non-negative
- * Failture: Negative
+ * Failure: Negative
*-------------------------------------------------------------------------
*/
herr_t
@@ -711,7 +711,7 @@ H5D__calculate_minimum_header_size(H5F_t *file, H5D_t *dset, H5O_t *ohdr)
H5T_t * type = NULL;
H5O_fill_t *fill_prop = NULL;
hbool_t use_at_least_v18 = FALSE;
- const char continuation[1] = ""; /* requred for work-around */
+ const char continuation[1] = ""; /* required for work-around */
size_t get_value = 0;
size_t ret_value = 0;
@@ -758,7 +758,7 @@ H5D__calculate_minimum_header_size(H5F_t *file, H5D_t *dset, H5O_t *ohdr)
HGOTO_ERROR(H5E_DATASET, H5E_CANTGET, 0, "can't get size of continuation message")
ret_value += get_value;
- /* Fill Value (backwards compatability) message size */
+ /* Fill Value (backwards compatibility) message size */
if (fill_prop->buf && !use_at_least_v18) {
H5O_fill_t old_fill_prop; /* Copy for writing "old" fill value */
@@ -847,7 +847,7 @@ H5D__prepare_minimized_oh(H5F_t *file, H5D_t *dset, H5O_loc_t *oloc)
if (ohdr_size == 0)
HGOTO_ERROR(H5E_OHDR, H5E_BADVALUE, FAIL, "computed header size is invalid")
- /* Special allocation of space for compact datsets is handled by the call here. */
+ /* Special allocation of space for compact datasets is handled by the call here. */
if (H5O_apply_ohdr(file, oh, dset->shared->dcpl_id, ohdr_size, (size_t)1, oloc) == FAIL)
HGOTO_ERROR(H5E_OHDR, H5E_BADVALUE, FAIL, "can't apply object header to file")
diff --git a/src/H5Dmpio.c b/src/H5Dmpio.c
index ddf5e06..f8bce33 100644
--- a/src/H5Dmpio.c
+++ b/src/H5Dmpio.c
@@ -365,7 +365,7 @@ H5D__mpio_opt_possible(const H5D_io_info_t *io_info, const H5S_t *file_space, co
/* Check to see if the process is reading the entire dataset */
if (H5S_GET_SELECT_TYPE(file_space) != H5S_SEL_ALL)
local_cause[1] |= H5D_MPIO_RANK0_NOT_H5S_ALL;
- /* Only perform this optimization for contigous datasets, currently */
+ /* Only perform this optimization for contiguous datasets, currently */
else if (H5D_CONTIGUOUS != io_info->dset->shared->layout.type)
/* Flag to do a MPI_Bcast of the data from one proc instead of
* having all the processes involved in the collective I/O.
@@ -996,9 +996,9 @@ H5D__link_chunk_collective_io(H5D_io_info_t *io_info, const H5D_type_info_t *typ
int sum_chunk)
{
H5D_chunk_addr_info_t *chunk_addr_info_array = NULL;
- MPI_Datatype chunk_final_mtype; /* Final memory MPI datatype for all chunks with seletion */
+ MPI_Datatype chunk_final_mtype; /* Final memory MPI datatype for all chunks with selection */
hbool_t chunk_final_mtype_is_derived = FALSE;
- MPI_Datatype chunk_final_ftype; /* Final file MPI datatype for all chunks with seletion */
+ MPI_Datatype chunk_final_ftype; /* Final file MPI datatype for all chunks with selection */
hbool_t chunk_final_ftype_is_derived = FALSE;
H5D_storage_t ctg_store; /* Storage info for "fake" contiguous dataset */
size_t total_chunks;
@@ -2432,7 +2432,7 @@ done:
*
* 1) Each process provides two piece of information for all chunks having selection
* a) chunk index
- * b) wheather this chunk is regular(for MPI derived datatype not working case)
+ * b) whether this chunk is regular(for MPI derived datatype not working case)
*
* 2) Gather all the information to the root process
*
diff --git a/src/H5Dpkg.h b/src/H5Dpkg.h
index 32f1c80..49c95a5 100644
--- a/src/H5Dpkg.h
+++ b/src/H5Dpkg.h
@@ -433,7 +433,7 @@ typedef struct H5D_rdcdc_t {
/*
* A dataset is made of two layers, an H5D_t struct that is unique to
- * each instance of an opened datset, and a shared struct that is only
+ * each instance of an opened dataset, and a shared struct that is only
* created once for a given dataset. Thus, if a dataset is opened twice,
* there will be two IDs and two H5D_t structs, both sharing one H5D_shared_t.
*/
diff --git a/src/H5Dpublic.h b/src/H5Dpublic.h
index c496414..75f4b95 100644
--- a/src/H5Dpublic.h
+++ b/src/H5Dpublic.h
@@ -300,7 +300,7 @@ extern "C" {
* caller may derive new datatypes, dataspaces, and creation and
* access properties from the old ones and reuse them in calls to
* create additional datasets. Once created, the dataset can be
- * read from or written to. Reading data from a datatset that was
+ * read from or written to. Reading data from a dataset that was
* not previously written, the HDF5 library will return default
* or user-defined fill values.
*
diff --git a/src/H5Dvirtual.c b/src/H5Dvirtual.c
index 49d1c04..b28bbbc 100644
--- a/src/H5Dvirtual.c
+++ b/src/H5Dvirtual.c
@@ -206,7 +206,7 @@ H5D_virtual_check_mapping_pre(const H5S_t *vspace, const H5S_t *src_space,
"can't get number of elements in non-unlimited dimension")
if (nenu_vs != nenu_ss)
HGOTO_ERROR(H5E_ARGS, H5E_BADVALUE, FAIL,
- "numbers of elemnts in the non-unlimited dimensions is different for source and "
+ "numbers of elements in the non-unlimited dimensions is different for source and "
"virtual spaces")
} /* end if */
/* We will handle the printf case after parsing the source names */
@@ -1492,7 +1492,7 @@ H5D__virtual_set_extent_unlim(const H5D_t *dset)
storage->list[i].source_dset.dset->shared->space) < 0)
HGOTO_ERROR(H5E_DATASET, H5E_CANTCOPY, FAIL, "can't copy source dataspace extent")
- /* Get source space dimenstions */
+ /* Get source space dimensions */
if (H5S_get_simple_extent_dims(storage->list[i].source_select, curr_dims, NULL) < 0)
HGOTO_ERROR(H5E_DATASET, H5E_CANTGET, FAIL, "can't get source space dimensions")
@@ -1970,7 +1970,7 @@ H5D__virtual_init_all(const H5D_t *dset)
storage->list[i].source_dset.dset->shared->space) < 0)
HGOTO_ERROR(H5E_DATASET, H5E_CANTCOPY, FAIL, "can't copy source dataspace extent")
- /* Get source space dimenstions */
+ /* Get source space dimensions */
if (H5S_get_simple_extent_dims(storage->list[i].source_select, source_dims, NULL) < 0)
HGOTO_ERROR(H5E_DATASET, H5E_CANTGET, FAIL, "can't get source space dimensions")
@@ -2354,7 +2354,7 @@ H5D__virtual_is_data_cached(const H5D_shared_t *shared_dset)
if (storage->list[i].psfn_nsubs || storage->list[i].psdn_nsubs) {
/* Iterate over sub-source dsets */
for (j = storage->list[i].sub_dset_io_start; j < storage->list[i].sub_dset_io_end; j++)
- /* Check for cahced data in source dset */
+ /* Check for cached data in source dset */
if (storage->list[i].sub_dset[j].dset &&
storage->list[i].sub_dset[j].dset->shared->layout.ops->is_data_cached &&
storage->list[i].sub_dset[j].dset->shared->layout.ops->is_data_cached(
@@ -2688,7 +2688,7 @@ H5D__virtual_post_io(H5O_storage_virtual_t *storage)
/*-------------------------------------------------------------------------
* Function: H5D__virtual_read_one
*
- * Purpose: Read from a singe source dataset in a virtual dataset.
+ * Purpose: Read from a single source dataset in a virtual dataset.
*
* Return: Non-negative on success/Negative on failure
*
@@ -2878,7 +2878,7 @@ done:
/*-------------------------------------------------------------------------
* Function: H5D__virtual_write_one
*
- * Purpose: Write to a singe source dataset in a virtual dataset.
+ * Purpose: Write to a single source dataset in a virtual dataset.
*
* Return: Non-negative on success/Negative on failure
*
diff --git a/src/H5E.c b/src/H5E.c
index 7f657fd..d203248 100644
--- a/src/H5E.c
+++ b/src/H5E.c
@@ -686,7 +686,7 @@ done:
/*-------------------------------------------------------------------------
* Function: H5E__close_msg
*
- * Purpose: Private function to close an error messge.
+ * Purpose: Private function to close an error message.
*
* Return: SUCCEED/FAIL
*
diff --git a/src/H5EApkg.h b/src/H5EApkg.h
index bfa8588..2212ccb 100644
--- a/src/H5EApkg.h
+++ b/src/H5EApkg.h
@@ -301,7 +301,7 @@ typedef struct H5EA_dblock_t {
/* Computed/cached values (not stored) */
size_t nelmts; /* Number of elements in block */
- size_t npages; /* Nummber of pages in a block (zero if not paged) */
+ size_t npages; /* Number of pages in a block (zero if not paged) */
} H5EA_dblock_t;
/* The extensible array data block page information */
diff --git a/src/H5Epkg.h b/src/H5Epkg.h
index 30ff084..b11ee9e 100644
--- a/src/H5Epkg.h
+++ b/src/H5Epkg.h
@@ -48,7 +48,7 @@
* each thread individually. The association of stacks to threads will
* be handled by the pthread library.
*
- * In order for this macro to work, H5E__get_my_stack() must be preceeded
+ * In order for this macro to work, H5E__get_my_stack() must be preceded
* by "H5E_t *estack =".
*/
#define H5E__get_my_stack() H5E__get_stack()
diff --git a/src/H5F.c b/src/H5F.c
index 8a28b00..09ca3fe 100644
--- a/src/H5F.c
+++ b/src/H5F.c
@@ -1306,7 +1306,7 @@ done:
/*-------------------------------------------------------------------------
* Function: H5Funmount
*
- * Purpose: Given a mount point, dissassociate the mount point's file
+ * Purpose: Given a mount point, disassociate the mount point's file
* from the file mounted there. Do not close either file.
*
* The mount point can either be the group in the parent or the
@@ -2262,7 +2262,7 @@ done:
* 1) The file being opened has v3 superblock
* 2) The file is opened with H5F_ACC_RDWR
* 3) The file is not already marked for SWMR writing
- * 4) Current implementaion for opened objects:
+ * 4) Current implementation for opened objects:
* --only allow datasets and groups without attributes
* --disallow named datatype with/without attributes
* --disallow opened attributes attached to objects
diff --git a/src/H5FA.c b/src/H5FA.c
index 23c9fe4..3e81080 100644
--- a/src/H5FA.c
+++ b/src/H5FA.c
@@ -92,7 +92,7 @@ H5FL_BLK_DEFINE(fa_native_elmt);
/*-------------------------------------------------------------------------
* Function: H5FA__new
*
- * Purpose: Allocate and initialize a new fixe array wrapper in memory
+ * Purpose: Allocate and initialize a new fixed array wrapper in memory
*
* Return: Pointer to farray wrapper success
* NULL on failure
diff --git a/src/H5FAcache.c b/src/H5FAcache.c
index 367b4b0..e6e32da 100644
--- a/src/H5FAcache.c
+++ b/src/H5FAcache.c
@@ -943,7 +943,7 @@ done:
* file space block set.
*
* This function is needed when the data block is paged, as
- * the datablock header and all its pages are allocted as a
+ * the datablock header and all its pages are allocated as a
* single contiguous chunk of file space, and must be
* deallocated the same way.
*
diff --git a/src/H5FApkg.h b/src/H5FApkg.h
index c4bf934..15f6445 100644
--- a/src/H5FApkg.h
+++ b/src/H5FApkg.h
@@ -179,8 +179,8 @@ typedef struct H5FA_dblock_t {
/* Computed/cached values (not stored) */
haddr_t addr; /* Address of this data block on disk */
hsize_t size; /* Size of data block on disk */
- size_t npages; /* Nummber of pages in data block (zero if not paged) */
- size_t last_page_nelmts; /* Nummber of elements in last page, if paged */
+ size_t npages; /* Number of pages in data block (zero if not paged) */
+ size_t last_page_nelmts; /* Number of elements in last page, if paged */
/* Fixed Array data block information (not stored) */
size_t dblk_page_nelmts; /* # of elements per data block page */
diff --git a/src/H5FD.c b/src/H5FD.c
index ce4f342..397da34 100644
--- a/src/H5FD.c
+++ b/src/H5FD.c
@@ -1367,7 +1367,7 @@ H5FD_get_fs_type_map(const H5FD_t *file, H5FD_mem_t *type_map)
HDassert(file->cls);
HDassert(type_map);
- /* Check for VFD class providing a type map retrieval rouine */
+ /* Check for VFD class providing a type map retrieval routine */
if (file->cls->get_type_map) {
/* Retrieve type mapping for this file */
if ((file->cls->get_type_map)(file, type_map) < 0)
diff --git a/src/H5FDcore.c b/src/H5FDcore.c
index 21d0324..a1750ee 100644
--- a/src/H5FDcore.c
+++ b/src/H5FDcore.c
@@ -1522,7 +1522,7 @@ done:
* Addendum -- 12/2/11
* For file images opened with the core file driver, it is
* necessary that we avoid reallocating the core file driver's
- * buffer uneccessarily.
+ * buffer unnecessarily.
*
* To this end, I have made the following functional changes
* to this function.
diff --git a/src/H5FDhdfs.c b/src/H5FDhdfs.c
index a101aa7..4927a40 100644
--- a/src/H5FDhdfs.c
+++ b/src/H5FDhdfs.c
@@ -530,16 +530,16 @@ done:
* Function: H5FD__hdfs_validate_config()
*
* Purpose: Test to see if the supplied instance of H5FD_hdfs_fapl_t
- * contains internally consistant data. Return SUCCEED if so,
+ * contains internally consistent data. Return SUCCEED if so,
* and FAIL otherwise.
*
- * Note the difference between internally consistant and
+ * Note the difference between internally consistent and
* correct. As we will have to try to access the target
* object to determine whether the supplied data is correct,
- * we will settle for internal consistancy at this point
+ * we will settle for internal consistency at this point
*
* Return: SUCCEED if instance of H5FD_hdfs_fapl_t contains internally
- * consistant data, FAIL otherwise.
+ * consistent data, FAIL otherwise.
*
* Programmer: Jacob Smith
* 9/10/17
diff --git a/src/H5FDlog.c b/src/H5FDlog.c
index 4128582..af34682 100644
--- a/src/H5FDlog.c
+++ b/src/H5FDlog.c
@@ -1543,7 +1543,7 @@ H5FD__log_write(H5FD_t *_file, H5FD_mem_t type, hid_t H5_ATTR_UNUSED dxpl_id, ha
HDfprintf(file->logfp, "%10" PRIuHADDR "-%10" PRIuHADDR " (%10zu bytes) (%s) Written", orig_addr,
(orig_addr + orig_size) - 1, orig_size, flavors[type]);
- /* Check if this is the first write into a "default" section, grabbed by the metadata agregation
+ /* Check if this is the first write into a "default" section, grabbed by the metadata aggregation
* algorithm */
if (file->fa.flags & H5FD_LOG_FLAVOR) {
if ((H5FD_mem_t)file->flavor[orig_addr] == H5FD_MEM_DEFAULT) {
diff --git a/src/H5FDmirror.c b/src/H5FDmirror.c
index 52f07e6..84c744c 100644
--- a/src/H5FDmirror.c
+++ b/src/H5FDmirror.c
@@ -1145,7 +1145,7 @@ done:
/* -------------------------------------------------------------------------
* Function: H5FD__mirror_fapl_get
*
- * Purpose: Get the file access propety list which could be used to create
+ * Purpose: Get the file access property list which could be used to create
* an identical file.
*
* Return: Success: pointer to the new file access property list value.
diff --git a/src/H5FDmirror.h b/src/H5FDmirror.h
index 2fdf6c4..b196b2b 100644
--- a/src/H5FDmirror.h
+++ b/src/H5FDmirror.h
@@ -34,7 +34,7 @@ extern "C" {
/* ---------------------------------------------------------------------------
* Structure: H5FD_mirror_fapl_t
*
- * Used to pass configuraiton information to the Mirror VFD.
+ * Used to pass configuration information to the Mirror VFD.
* Populate components as appropriate and pass structure pointer to
* `H5Pset_fapl_mirror()`.
*
diff --git a/src/H5FDmirror_priv.h b/src/H5FDmirror_priv.h
index 21de97b..6a7b13e 100644
--- a/src/H5FDmirror_priv.h
+++ b/src/H5FDmirror_priv.h
@@ -80,7 +80,7 @@ extern "C" {
*
* `magic` (uint32_t)
* A "unique" number identifying the structure and endianness of
- * transmitting maching.
+ * transmitting machine.
* Must be set to H5FD_MIRROR_XMIT_MAGIC native to the VFD "sender".
*
* `version` (uint8_t)
@@ -214,13 +214,13 @@ typedef struct H5FD_mirror_xmit_open_t {
*
* `status` (uint32_t)
* Number indicating whether the command was successful or if an
- * occured.
+ * occurred.
* Allowed values are H5FD_MIRROR_STATUS_OK and
* H5FD_MIRROR_STATUS_ERROR.
*
* `message` (char[])
* Error message. Populated if and only if there was a problem.
- * It is possible that a message may reach the end of the alloted
+ * It is possible that a message may reach the end of the allotted
* space without a NULL terminator -- the onus is on the programmer to
* handle this situation.
*
diff --git a/src/H5FDpublic.h b/src/H5FDpublic.h
index d774210..221f569 100644
--- a/src/H5FDpublic.h
+++ b/src/H5FDpublic.h
@@ -244,7 +244,7 @@
* opcodes. (known op codes should be handled as
* appropriate.)
*
- * If this flag is set for an uknown op code, that
+ * If this flag is set for an unknown op code, that
* op code should be passed to the next VFD down
* the VFD stack en-route to the terminal VFD.
* If that VFD does not support the ctl call, the
diff --git a/src/H5FDros3.c b/src/H5FDros3.c
index b2c9d21..df06526 100644
--- a/src/H5FDros3.c
+++ b/src/H5FDros3.c
@@ -354,7 +354,7 @@ H5FD__ros3_term(void)
* Function: H5Pset_fapl_ros3
*
* Purpose: Modify the file access property list to use the H5FD_ROS3
- * driver defined in this source file. All driver specfic
+ * driver defined in this source file. All driver specific
* properties are passed in as a pointer to a suitably
* initialized instance of H5FD_ros3_fapl_t
*
@@ -397,16 +397,16 @@ done:
* Function: H5FD__ros3_validate_config()
*
* Purpose: Test to see if the supplied instance of H5FD_ros3_fapl_t
- * contains internally consistant data. Return SUCCEED if so,
+ * contains internally consistent data. Return SUCCEED if so,
* and FAIL otherwise.
*
- * Note the difference between internally consistant and
+ * Note the difference between internally consistent and
* correct. As we will have to try to access the target
* object to determine whether the supplied data is correct,
- * we will settle for internal consistancy at this point
+ * we will settle for internal consistency at this point
*
* Return: SUCCEED if instance of H5FD_ros3_fapl_t contains internally
- * consistant data, FAIL otherwise.
+ * consistent data, FAIL otherwise.
*
* Programmer: Jacob Smith
* 9/10/17
diff --git a/src/H5FDs3comms.h b/src/H5FDs3comms.h
index da6a62d..b81bfae 100644
--- a/src/H5FDs3comms.h
+++ b/src/H5FDs3comms.h
@@ -179,7 +179,7 @@
* HTTP header fields, of particular use when composing an
* "S3 Canonical Request" for authentication.
*
- * - The creation of a Canoncial Request involves:
+ * - The creation of a Canonical Request involves:
* - convert field names to lower case
* - sort by this lower-case name
* - convert ": " name-value separator in HTTP string to ":"
@@ -459,7 +459,7 @@ typedef struct {
*
* Pointer to NULL-terminated string for "secret" access id to S3 resource.
*
- * Requred to authenticate.
+ * Required to authenticate.
*
* `signing_key` (unsigned char *)
*
@@ -470,7 +470,7 @@ typedef struct {
* which may be re-used for several (up to seven (7)) days from creation?
* Computed once upon file open.
*
- * Requred to authenticate.
+ * Required to authenticate.
*
*----------------------------------------------------------------------------
*/
diff --git a/src/H5FDsplitter.h b/src/H5FDsplitter.h
index 6463158..c8751c8 100644
--- a/src/H5FDsplitter.h
+++ b/src/H5FDsplitter.h
@@ -35,7 +35,7 @@
* Structure: H5FD_spliiter_vfd_config_t
*
* One-stop shopping for configuring a Splitter VFD (rather than many
- * paramaters passed into H5Pset/get functions).
+ * parameters passed into H5Pset/get functions).
*
* magic (int32_t)
* Semi-unique number, used to sanity-check that a given pointer is
diff --git a/src/H5FLprivate.h b/src/H5FLprivate.h
index 381f977..6519551 100644
--- a/src/H5FLprivate.h
+++ b/src/H5FLprivate.h
@@ -148,8 +148,8 @@ typedef struct H5FL_reg_head_t {
typedef union H5FL_blk_list_t {
size_t size; /* Size of the page */
union H5FL_blk_list_t *next; /* Pointer to next block in free list */
- double unused1; /* Unused normally, just here for aligment */
- haddr_t unused2; /* Unused normally, just here for aligment */
+ double unused1; /* Unused normally, just here for alignment */
+ haddr_t unused2; /* Unused normally, just here for alignment */
} H5FL_blk_list_t;
/* Data structure for priority queue node of block free lists */
@@ -223,8 +223,8 @@ typedef struct H5FL_blk_head_t {
typedef union H5FL_arr_list_t {
union H5FL_arr_list_t *next; /* Pointer to next block in free list */
size_t nelem; /* Number of elements in this array */
- double unused1; /* Unused normally, just here for aligment */
- haddr_t unused2; /* Unused normally, just here for aligment */
+ double unused1; /* Unused normally, just here for alignment */
+ haddr_t unused2; /* Unused normally, just here for alignment */
} H5FL_arr_list_t;
/* Data structure for each size of array element */
diff --git a/src/H5FSprivate.h b/src/H5FSprivate.h
index 718c4a7..bb4f56a 100644
--- a/src/H5FSprivate.h
+++ b/src/H5FSprivate.h
@@ -116,7 +116,7 @@ typedef struct H5FS_section_class_t {
const struct H5FS_section_class_t *, const uint8_t *, haddr_t, hsize_t,
unsigned *); /* Routine to deserialize a buffer into a "live" section */
htri_t (*can_merge)(const H5FS_section_info_t *, const H5FS_section_info_t *,
- void *); /* Routine to determine if two nodes are mergable */
+ void *); /* Routine to determine if two nodes are mergeable */
herr_t (*merge)(H5FS_section_info_t **, H5FS_section_info_t *, void *); /* Routine to merge two nodes */
htri_t (*can_shrink)(const H5FS_section_info_t *,
void *); /* Routine to determine if node can shrink container */
diff --git a/src/H5FSsection.c b/src/H5FSsection.c
index c764932..6c5a850 100644
--- a/src/H5FSsection.c
+++ b/src/H5FSsection.c
@@ -2082,11 +2082,11 @@ H5FS_sect_change_class(H5F_t *f, H5FS_t *fspace, H5FS_section_info_t *sect, uint
} /* end else */
} /* end if */
- /* Check if the section's class change will affect the mergable list */
+ /* Check if the section's class change will affect the mergeable list */
if ((old_cls->flags & H5FS_CLS_SEPAR_OBJ) != (new_cls->flags & H5FS_CLS_SEPAR_OBJ)) {
- hbool_t to_mergable; /* Flag if the section is changing to a mergable section */
+ hbool_t to_mergable; /* Flag if the section is changing to a mergeable section */
- /* Determine if this section is becoming mergable or is becoming separate */
+ /* Determine if this section is becoming mergeable or is becoming separate */
if (old_cls->flags & H5FS_CLS_SEPAR_OBJ)
to_mergable = TRUE;
else
diff --git a/src/H5Fefc.c b/src/H5Fefc.c
index 2d4eff2..2c7dd92 100644
--- a/src/H5Fefc.c
+++ b/src/H5Fefc.c
@@ -450,7 +450,7 @@ H5F__efc_release_real(H5F_efc_t *efc)
/* Sanity checks */
HDassert(efc);
- /* Lock the EFC to prevent manipulation of the EFC wile we are releasing it.
+ /* Lock the EFC to prevent manipulation of the EFC while we are releasing it.
* The EFC should never be locked when we enter this function because that
* would require a cycle, a cycle would necessarily invoke
* H5F__efc_try_close(), and that function checks the status of the lock
diff --git a/src/H5Fint.c b/src/H5Fint.c
index b9be873..05bba16 100644
--- a/src/H5Fint.c
+++ b/src/H5Fint.c
@@ -1028,7 +1028,7 @@ H5F__is_hdf5(const char *name, hid_t fapl_id)
{
H5FD_t * file = NULL; /* Low-level file struct */
H5F_shared_t *shared = NULL; /* Shared part of file */
- haddr_t sig_addr = HADDR_UNDEF; /* Addess of hdf5 file signature */
+ haddr_t sig_addr = HADDR_UNDEF; /* Address of hdf5 file signature */
htri_t ret_value = FAIL; /* Return value */
FUNC_ENTER_PACKAGE
@@ -1125,7 +1125,7 @@ H5F__new(H5F_shared_t *shared, unsigned flags, hid_t fcpl_id, hid_t fapl_id, H5F
/* Initialization for handling file space (for paged aggregation) */
f->shared->pgend_meta_thres = H5F_FILE_SPACE_PGEND_META_THRES;
- /* intialize point of no return */
+ /* initialize point of no return */
f->shared->point_of_no_return = FALSE;
/* Copy the file creation and file access property lists into the
@@ -1262,7 +1262,7 @@ H5F__new(H5F_shared_t *shared, unsigned flags, hid_t fcpl_id, hid_t fapl_id, H5F
f->shared->read_attempts = H5F_METADATA_READ_ATTEMPTS;
}
- /* Determine the # of bins for metdata read retries */
+ /* Determine the # of bins for metadata read retries */
if (H5F_set_retries(f) < 0)
HGOTO_ERROR(H5E_FILE, H5E_CANTINIT, NULL, "can't set retries and retries_nbins")
@@ -1787,7 +1787,7 @@ H5F_open(const char *name, unsigned flags, hid_t fcpl_id, hid_t fapl_id)
* or creating it) so we can compare it with files that are already
* open. If that fails then we try again with the full set of flags
* (only if they're different than the original failed attempt).
- * However, if the file driver can't distinquish between files then
+ * However, if the file driver can't distinguish between files then
* there's no reason to open the file tentatively because it's the
* application's responsibility to prevent this situation (there's no
* way for us to detect it here anyway).
@@ -3116,7 +3116,7 @@ H5F__get_file_image(H5F_t *file, void *buf_ptr, size_t buf_len, size_t *image_le
* file driver. However, this test will not work if there is some
* other file driver sitting on top of the multi file driver.
*
- * I'm not sure if this is possible at present, but in all likelyhood,
+ * I'm not sure if this is possible at present, but in all likelihood,
* it will become possible in the future. On the other hand, we may
* remove the split/multi file drivers before then.
*
@@ -3136,7 +3136,7 @@ H5F__get_file_image(H5F_t *file, void *buf_ptr, size_t buf_len, size_t *image_le
*
* While this problem is quite solvable, the required time and
* resources are lacking at present. Hence, for now, we don't
- * allow the get file image operation to be perfomed on files
+ * allow the get file image operation to be performed on files
* opened with the family file driver.
*
* Observe that the following test only looks at the top level
@@ -3567,7 +3567,7 @@ done:
* 1) The file being opened has v3 superblock
* 2) The file is opened with H5F_ACC_RDWR
* 3) The file is not already marked for SWMR writing
- * 4) Current implementaion for opened objects:
+ * 4) Current implementation for opened objects:
* --only allow datasets and groups without attributes
* --disallow named datatype with/without attributes
* --disallow opened attributes attached to objects
diff --git a/src/H5Fmount.c b/src/H5Fmount.c
index 4b90ea3..7d27234 100644
--- a/src/H5Fmount.c
+++ b/src/H5Fmount.c
@@ -104,7 +104,7 @@ H5F_mount(const H5G_loc_t *loc, const char *name, H5F_t *child, hid_t H5_ATTR_UN
H5F_t * parent = NULL; /*file containing mount point */
unsigned lt, rt, md; /*binary search indices */
int cmp; /*binary search comparison value*/
- H5G_loc_t mp_loc; /* entry of moint point to be opened */
+ H5G_loc_t mp_loc; /* entry of mount point to be opened */
H5G_name_t mp_path; /* Mount point group hier. path */
H5O_loc_t mp_oloc; /* Mount point object location */
H5G_loc_t root_loc; /* Group location of root of file to mount */
diff --git a/src/H5Fpkg.h b/src/H5Fpkg.h
index b016696..9208815 100644
--- a/src/H5Fpkg.h
+++ b/src/H5Fpkg.h
@@ -79,7 +79,7 @@
#define H5F_SUPERBLOCK_FIXED_SIZE (H5F_SIGNATURE_LEN + 1) /* superblock version */
/* The H5F_SUPERBLOCK_MINIMAL_VARLEN_SIZE is the minimal amount of super block
- * variable length data guarnateed to load the sizeof offsets and the sizeof
+ * variable length data guaranteed to load the sizeof offsets and the sizeof
* lengths fields in all versions of the superblock.
*
* This is necessary in the V3 cache, as on the initial load, we need to
@@ -352,7 +352,7 @@ struct H5F_shared_t {
/* Metadata retry info */
unsigned read_attempts; /* The # of reads to try when reading metadata with checksum */
unsigned retries_nbins; /* # of bins for each retries[] */
- uint32_t *retries[H5AC_NTYPES]; /* Track # of read retries for metdata items with checksum */
+ uint32_t *retries[H5AC_NTYPES]; /* Track # of read retries for metadata items with checksum */
/* Object flush info */
H5F_object_flush_t object_flush; /* Information for object flush callback */
diff --git a/src/H5Fsuper.c b/src/H5Fsuper.c
index 4df0064..bf02b06 100644
--- a/src/H5Fsuper.c
+++ b/src/H5Fsuper.c
@@ -375,7 +375,7 @@ H5F__super_read(H5F_t *f, H5P_genplist_t *fa_plist, hbool_t initial_read)
if (0 == mpi_rank) {
herr_t status;
- /* Try detecting file's siganture */
+ /* Try detecting file's signature */
/* (Don't leave before Bcast, to avoid hang on error) */
H5E_BEGIN_TRY
{
@@ -467,7 +467,7 @@ H5F__super_read(H5F_t *f, H5P_genplist_t *fa_plist, hbool_t initial_read)
*
* After upgrading low_bound, the library will check to ensure that the
* superblock version does not exceed the version allowed by high_bound.
- * Otherise fail file open.
+ * Otherwise fail file open.
*
* For details, please see RFC:Setting Bounds for Object Creation in HDF5 1.10.0.
*/
diff --git a/src/H5Fsuper_cache.c b/src/H5Fsuper_cache.c
index 28ff82f..b9dc74a 100644
--- a/src/H5Fsuper_cache.c
+++ b/src/H5Fsuper_cache.c
@@ -698,7 +698,7 @@ H5F__cache_superblock_serialize(const H5F_t *f, void *_image, size_t H5_ATTR_UNU
HGOTO_ERROR(H5E_RESOURCE, H5E_CANTGET, FAIL, "driver get_eoa request failed")
H5F_addr_encode(f, &image, (rel_eof + sblock->base_addr));
- /* Encode the driver informaton block address */
+ /* Encode the driver information block address */
H5F_addr_encode(f, &image, sblock->driver_addr);
/* Encode the root group object entry, including the cached stab info */
@@ -797,7 +797,7 @@ done:
/*-------------------------------------------------------------------------
* Function: H5F__cache_drvrinfo_get_initial_load_size
*
- * Purpose: Compute the intiial size of the data structure on disk.
+ * Purpose: Compute the initial size of the data structure on disk.
*
* Return: Non-negative on success/Negative on failure
*
diff --git a/src/H5Gcache.c b/src/H5Gcache.c
index 4713193..65ed1e6 100644
--- a/src/H5Gcache.c
+++ b/src/H5Gcache.c
@@ -134,7 +134,7 @@ H5G__cache_node_get_initial_load_size(void *_udata, size_t *image_len)
* Function: H5G__cache_node_deserialize
*
* Purpose: Given a buffer containing the on disk image of a symbol table
- * node, allocate an instance of H5G_node_t, load the contence of the
+ * node, allocate an instance of H5G_node_t, load the contents of the
* image into it, and return a pointer to the instance.
*
* Note that deserializing the image requires access to the file
@@ -240,7 +240,7 @@ H5G__cache_node_image_len(const void *_thing, size_t *image_len)
/*-------------------------------------------------------------------------
* Function: H5G__cache_node_serialize
*
- * Purpose: Given a correctly sized buffer and an instace of H5G_node_t,
+ * Purpose: Given a correctly sized buffer and an instance of H5G_node_t,
* serialize the contents of the instance of H5G_node_t, and write
* this data into the supplied buffer. This buffer will be written
* to disk.
diff --git a/src/H5Gcompact.c b/src/H5Gcompact.c
index 0c0b5db..4988738 100644
--- a/src/H5Gcompact.c
+++ b/src/H5Gcompact.c
@@ -419,7 +419,7 @@ done:
* Function: H5G__compact_lookup_cb
*
* Purpose: Callback routine for searching 'link' messages for a particular
- * name & gettting object location for it
+ * name & getting object location for it
*
* Return: SUCCEED/FAIL
*
diff --git a/src/H5Gnode.c b/src/H5Gnode.c
index ba29b78..d73888e 100644
--- a/src/H5Gnode.c
+++ b/src/H5Gnode.c
@@ -1210,7 +1210,7 @@ H5G__node_copy(H5F_t *f, const void H5_ATTR_UNUSED *_lt_key, haddr_t addr, const
&(sn->entry[i]); /* Convenience variable to refer to current source group entry */
H5O_link_t lnk; /* Link to insert */
const char * name; /* Name of source object */
- H5G_entry_t tmp_src_ent; /* Temperary copy. Change will not affect the cache */
+ H5G_entry_t tmp_src_ent; /* Temporary copy. Change will not affect the cache */
H5O_type_t obj_type = H5O_TYPE_UNKNOWN; /* Target object type */
H5G_copy_file_ud_t *cpy_udata; /* Copy file udata */
H5G_obj_create_t gcrt_info; /* Group creation info */
diff --git a/src/H5Gpublic.h b/src/H5Gpublic.h
index d9c29f6..ce36b84 100644
--- a/src/H5Gpublic.h
+++ b/src/H5Gpublic.h
@@ -1113,7 +1113,7 @@ H5_DLL herr_t H5Gget_objinfo(hid_t loc_id, const char *name, hbool_t follow_link
* actual object name length, the object name is truncated to
* \Code{max_size - 1} characters.
*
- * Note that if the size of the object's name is unkown, a preliminary
+ * Note that if the size of the object's name is unknown, a preliminary
* call to H5Gget_objname_by_idx() with \p name set to \c NULL will
* return the length of the object's name. A second call to
* H5Gget_objname_by_idx() can then be used to retrieve the actual
diff --git a/src/H5HF.c b/src/H5HF.c
index 8e69032..7904af9 100644
--- a/src/H5HF.c
+++ b/src/H5HF.c
@@ -355,7 +355,7 @@ H5HF_insert(H5HF_t *fh, size_t size, const void *obj, void *id /*out*/)
if (size > hdr->max_man_size) {
/* Store 'huge' object in heap
*
- * Although not ideal, we can quiet the const warning here becuase no
+ * Although not ideal, we can quiet the const warning here because no
* obj pointer that was originally const should ever arrive here.
*/
H5_GCC_CLANG_DIAG_OFF("cast-qual")
diff --git a/src/H5HFcache.c b/src/H5HFcache.c
index 5b50725..f409479 100644
--- a/src/H5HFcache.c
+++ b/src/H5HFcache.c
@@ -796,7 +796,7 @@ H5HF__cache_hdr_serialize(const H5F_t *f, void *_image, size_t H5_ATTR_NDEBUG_UN
/* Encode I/O filter information */
if (H5O_msg_encode(hdr->f, H5O_PLINE_ID, FALSE, image, &(hdr->pline)) < 0)
- HGOTO_ERROR(H5E_HEAP, H5E_CANTENCODE, FAIL, "can't encode I/O pipeline fiters")
+ HGOTO_ERROR(H5E_HEAP, H5E_CANTENCODE, FAIL, "can't encode I/O pipeline filters")
image += hdr->filter_len;
} /* end if */
diff --git a/src/H5HFdbg.c b/src/H5HFdbg.c
index 8dbb3f4..d38ae20 100644
--- a/src/H5HFdbg.c
+++ b/src/H5HFdbg.c
@@ -50,7 +50,7 @@
/* User data for direct block debugging iterator callback */
typedef struct {
FILE * stream; /* Stream for output */
- int indent; /* Indention amount */
+ int indent; /* Indentation amount */
int fwidth; /* Field width mount */
haddr_t dblock_addr; /* Direct block's address */
hsize_t dblock_size; /* Direct block's size */
@@ -63,7 +63,7 @@ typedef struct {
typedef struct {
H5FS_t *fspace; /* Free space manager */
FILE * stream; /* Stream for output */
- int indent; /* Indention amount */
+ int indent; /* Indentation amount */
int fwidth; /* Field width mount */
} H5HF_debug_iter_ud2_t;
diff --git a/src/H5HFdblock.c b/src/H5HFdblock.c
index 4adb47a..749b973 100644
--- a/src/H5HFdblock.c
+++ b/src/H5HFdblock.c
@@ -432,7 +432,7 @@ H5HF__man_dblock_protect(H5HF_hdr_t *hdr, haddr_t dblock_addr, size_t dblock_siz
H5HF_indirect_t *par_iblock, unsigned par_entry, unsigned flags)
{
H5HF_direct_t * dblock; /* Direct block from cache */
- H5HF_dblock_cache_ud_t udata; /* parent and other infor for deserializing direct block */
+ H5HF_dblock_cache_ud_t udata; /* parent and other info for deserializing direct block */
H5HF_direct_t * ret_value = NULL; /* Return value */
FUNC_ENTER_PACKAGE
diff --git a/src/H5HFspace.c b/src/H5HFspace.c
index d62267c..67f9b6f 100644
--- a/src/H5HFspace.c
+++ b/src/H5HFspace.c
@@ -383,7 +383,7 @@ H5HF__space_create_root(const H5HF_hdr_t *hdr, H5HF_indirect_t *root_iblock)
/* Only need to scan the sections if the free space has been initialized */
if (hdr->fspace)
- /* Iterate over all sections, seting the parent pointers in 'single' sections to the new indirect
+ /* Iterate over all sections, setting the parent pointers in 'single' sections to the new indirect
* block */
if (H5FS_sect_iterate(hdr->f, hdr->fspace, H5HF__space_create_root_cb, root_iblock) < 0)
HGOTO_ERROR(H5E_FSPACE, H5E_BADITER, FAIL, "can't iterate over sections to set parent pointers")
diff --git a/src/H5I.c b/src/H5I.c
index 954a86b..0a4fa10 100644
--- a/src/H5I.c
+++ b/src/H5I.c
@@ -814,7 +814,7 @@ H5I__iterate_pub_cb(void H5_ATTR_UNUSED *obj, hid_t id, void *_udata)
/*-------------------------------------------------------------------------
* Function: H5Iiterate
*
- * Purpose: Call the callback funciton op for each member of the id
+ * Purpose: Call the callback function op for each member of the id
* type type. op takes as parameters the id and a
* passthrough of op_data, and returns an herr_t. A positive
* return from op will cause the iteration to stop and
diff --git a/src/H5Ipublic.h b/src/H5Ipublic.h
index 8d4dbf8..d6892bb 100644
--- a/src/H5Ipublic.h
+++ b/src/H5Ipublic.h
@@ -508,7 +508,7 @@ H5_DLL int H5Idec_type_ref(H5I_type_t type);
*
* \brief Retrieves the reference count on an ID type
*
- * \param[in] type The identifier of the type whose reference count is to be retieved
+ * \param[in] type The identifier of the type whose reference count is to be retrieved
*
* \return Returns the current reference count on success, negative on failure.
*
diff --git a/src/H5Lpublic.h b/src/H5Lpublic.h
index 72b0182..e5a826a 100644
--- a/src/H5Lpublic.h
+++ b/src/H5Lpublic.h
@@ -462,7 +462,7 @@ H5_DLL herr_t H5Ldelete_by_idx_async(const char *app_file, const char *app_func,
*
* \return \herr_t
*
- * \details H5Lget_val() returns tha value of link \p name. For smbolic links,
+ * \details H5Lget_val() returns the value of link \p name. For smbolic links,
* this is the path to which the link points, including the null
* terminator. For external and user-defined links, it is the link
* buffer.
@@ -492,7 +492,7 @@ H5_DLL herr_t H5Ldelete_by_idx_async(const char *app_file, const char *app_func,
*
* This function should be used only after H5Lget_info() has been
* called to verify that \p name is a symbolic link. This can be
- * deteremined from the \c link_type field of the \ref H5L_info_t
+ * determined from the \c link_type field of the \ref H5L_info_t
* \c struct.
*
* \note This function will fail if called on a hard link.
@@ -625,7 +625,7 @@ H5_DLL herr_t H5Lget_val_by_idx(hid_t loc_id, const char *group_name, H5_index_t
* root group of an HDF5 file, and let \c lapl denote a valid link
* access property list identifier. A call to H5Lexists() with
* arguments c root, \c "/", and \c lapl returns a positive value;
- * in other words, \Code{H5Lexists(root, "/", lapl)} returns a postive
+ * in other words, \Code{H5Lexists(root, "/", lapl)} returns a positive
* value. In HDF5 version 1.8.16, this function returns 0.</li>
* </ol>
* Note that the function accepts link names and path names. This is
@@ -1094,7 +1094,7 @@ H5_DLL herr_t H5Lvisit2(hid_t grp_id, H5_index_t idx_type, H5_iter_order_t order
* \p idx_type specifies the index to be used. If the links have not
* been indexed by the index type, they will first be sorted by that
* index then the iteration will begin; if the links have been so
- * indexed, the sorting step will be unnecesary, so the iteration may
+ * indexed, the sorting step will be unnecessary, so the iteration may
* begin more quickly. Valid values include the following:
* \indexes
*
@@ -1853,7 +1853,7 @@ H5_DLL herr_t H5Lvisit1(hid_t grp_id, H5_index_t idx_type, H5_iter_order_t order
* \p idx_type specifies the index to be used. If the links have not
* been indexed by the index type, they will first be sorted by that
* index then the iteration will begin; if the links have been so
- * indexed, the sorting step will be unnecesary, so the iteration may
+ * indexed, the sorting step will be unnecessary, so the iteration may
* begin more quickly. Valid values include the following:
* \indexes
*
diff --git a/src/H5MFaggr.c b/src/H5MFaggr.c
index 77345c9..8ffc5c8 100644
--- a/src/H5MFaggr.c
+++ b/src/H5MFaggr.c
@@ -195,7 +195,7 @@ H5MF__aggr_alloc(H5F_t *f, H5F_blk_aggr_t *aggr, H5F_blk_aggr_t *other_aggr, H5F
haddr_t aggr_frag_addr = HADDR_UNDEF; /* Address of aggregrator fragment */
hsize_t aggr_frag_size = 0; /* Size of aggregator fragment */
hsize_t alignment; /* Alignment of this section */
- hsize_t aggr_mis_align = 0; /* Mis-alignment of aggregator */
+ hsize_t aggr_mis_align = 0; /* Misalignment of aggregator */
H5FD_mem_t alloc_type, other_alloc_type; /* Current aggregator & 'other' aggregator types */
#ifdef H5MF_AGGR_DEBUG
diff --git a/src/H5MFdbg.c b/src/H5MFdbg.c
index 6ccbe78..60d7779 100644
--- a/src/H5MFdbg.c
+++ b/src/H5MFdbg.c
@@ -50,7 +50,7 @@
typedef struct {
H5FS_t *fspace; /* Free space manager */
FILE * stream; /* Stream for output */
- int indent; /* Indention amount */
+ int indent; /* Indentation amount */
int fwidth; /* Field width amount */
} H5MF_debug_iter_ud_t;
diff --git a/src/H5Oainfo.c b/src/H5Oainfo.c
index 10c9ede..03ea3df 100644
--- a/src/H5Oainfo.c
+++ b/src/H5Oainfo.c
@@ -52,7 +52,7 @@ const H5O_msg_class_t H5O_MSG_AINFO[1] = {{
H5O_AINFO_ID, /*message id number */
"ainfo", /*message name for debugging */
sizeof(H5O_ainfo_t), /*native message size */
- 0, /* messages are sharable? */
+ 0, /* messages are shareable? */
H5O__ainfo_decode, /*decode message */
H5O__ainfo_encode, /*encode message */
H5O__ainfo_copy, /*copy the native value */
diff --git a/src/H5Oattr.c b/src/H5Oattr.c
index db53eaf..83c0370 100644
--- a/src/H5Oattr.c
+++ b/src/H5Oattr.c
@@ -65,7 +65,7 @@ const H5O_msg_class_t H5O_MSG_ATTR[1] = {{
H5O_ATTR_ID, /* message id number */
"attribute", /* message name for debugging */
sizeof(H5A_t), /* native message size */
- H5O_SHARE_IS_SHARABLE, /* messages are sharable? */
+ H5O_SHARE_IS_SHARABLE, /* messages are shareable? */
H5O__attr_shared_decode, /* decode message */
H5O__attr_shared_encode, /* encode message */
H5O__attr_copy, /* copy the native value */
diff --git a/src/H5Oattribute.c b/src/H5Oattribute.c
index 6ca4204..b83b57e 100644
--- a/src/H5Oattribute.c
+++ b/src/H5Oattribute.c
@@ -248,19 +248,19 @@ H5O__attr_create(const H5O_loc_t *loc, H5A_t *attr)
/* Check if switching to "dense" attribute storage is possible */
if (!H5F_addr_defined(ainfo.fheap_addr)) {
- htri_t sharable; /* Whether the attribute will be shared */
+ htri_t shareable; /* Whether the attribute will be shared */
size_t raw_size = 0; /* Raw size of message */
- /* Check for attribute being sharable */
- if ((sharable = H5SM_can_share(loc->file, NULL, NULL, H5O_ATTR_ID, attr)) < 0)
+ /* Check for attribute being shareable */
+ if ((shareable = H5SM_can_share(loc->file, NULL, NULL, H5O_ATTR_ID, attr)) < 0)
HGOTO_ERROR(H5E_ATTR, H5E_BADMESG, FAIL, "can't determine attribute sharing status")
- else if (sharable == FALSE) {
+ else if (shareable == FALSE) {
/* Compute the size needed to encode the attribute */
raw_size = (H5O_MSG_ATTR->raw_size)(loc->file, FALSE, attr);
} /* end if */
/* Check for condititions for switching to "dense" attribute storage are met */
- if (ainfo.nattrs == oh->max_compact || (!sharable && raw_size >= H5O_MESG_MAX_SIZE)) {
+ if (ainfo.nattrs == oh->max_compact || (!shareable && raw_size >= H5O_MESG_MAX_SIZE)) {
H5O_iter_cvt_t udata; /* User data for callback */
H5O_mesg_operator_t op; /* Wrapper for operator */
diff --git a/src/H5Obogus.c b/src/H5Obogus.c
index c7fcdb9..ea231d6 100644
--- a/src/H5Obogus.c
+++ b/src/H5Obogus.c
@@ -47,7 +47,7 @@ const H5O_msg_class_t H5O_MSG_BOGUS_VALID[1] = {{
H5O_BOGUS_VALID_ID, /*message id number */
"bogus valid", /*message name for debugging */
0, /*native message size */
- H5O_SHARE_IS_SHARABLE, /* messages are sharable? */
+ H5O_SHARE_IS_SHARABLE, /* messages are shareable? */
H5O__bogus_decode, /*decode message */
H5O__bogus_encode, /*encode message */
NULL, /*copy the native value */
@@ -71,7 +71,7 @@ const H5O_msg_class_t H5O_MSG_BOGUS_INVALID[1] = {{
H5O_BOGUS_INVALID_ID, /*message id number */
"bogus invalid", /*message name for debugging */
0, /*native message size */
- H5O_SHARE_IS_SHARABLE, /* messages are sharable? */
+ H5O_SHARE_IS_SHARABLE, /* messages are shareable? */
H5O__bogus_decode, /*decode message */
H5O__bogus_encode, /*encode message */
NULL, /*copy the native value */
diff --git a/src/H5Obtreek.c b/src/H5Obtreek.c
index a783366..1eaf3e5 100644
--- a/src/H5Obtreek.c
+++ b/src/H5Obtreek.c
@@ -37,7 +37,7 @@ const H5O_msg_class_t H5O_MSG_BTREEK[1] = {{
H5O_BTREEK_ID, /*message id number */
"v1 B-tree 'K' values", /*message name for debugging */
sizeof(H5O_btreek_t), /*native message size */
- 0, /* messages are sharable? */
+ 0, /* messages are shareable? */
H5O__btreek_decode, /*decode message */
H5O__btreek_encode, /*encode message */
H5O__btreek_copy, /*copy the native value */
diff --git a/src/H5Ocache.c b/src/H5Ocache.c
index 3aae0cb..ba47da3 100644
--- a/src/H5Ocache.c
+++ b/src/H5Ocache.c
@@ -275,7 +275,7 @@ H5O__cache_verify_chksum(const void *_image, size_t len, void *_udata)
*
* Note that the object header is read with with a speculative read.
* If the initial read is too small, make note of this fact and return
- * without error. H5C__load_entry() will note the size discrepency
+ * without error. H5C__load_entry() will note the size discrepancy
* and retry the deserialize operation with the correct size read.
*
* Return: Success: Pointer to in core representation
diff --git a/src/H5Ocache_image.c b/src/H5Ocache_image.c
index 2603f71..8f60a0e 100644
--- a/src/H5Ocache_image.c
+++ b/src/H5Ocache_image.c
@@ -51,7 +51,7 @@ const H5O_msg_class_t H5O_MSG_MDCI[1] = {{
H5O_MDCI_MSG_ID, /* message id number */
"mdci", /* message name for debugging */
sizeof(H5O_mdci_t), /* native message size */
- 0, /* messages are sharable? */
+ 0, /* messages are shareable? */
H5O__mdci_decode, /* decode message */
H5O__mdci_encode, /* encode message */
H5O__mdci_copy, /* copy method */
diff --git a/src/H5Ocont.c b/src/H5Ocont.c
index 080225a..9dd04fe 100644
--- a/src/H5Ocont.c
+++ b/src/H5Ocont.c
@@ -47,7 +47,7 @@ const H5O_msg_class_t H5O_MSG_CONT[1] = {{
H5O_CONT_ID, /*message id number */
"hdr continuation", /*message name for debugging */
sizeof(H5O_cont_t), /*native message size */
- 0, /* messages are sharable? */
+ 0, /* messages are shareable? */
H5O__cont_decode, /*decode message */
H5O__cont_encode, /*encode message */
NULL, /*no copy method */
diff --git a/src/H5Ocopy.c b/src/H5Ocopy.c
index e31db41..05dfc72 100644
--- a/src/H5Ocopy.c
+++ b/src/H5Ocopy.c
@@ -963,7 +963,7 @@ H5O__copy_header(const H5O_loc_t *oloc_src, H5O_loc_t *oloc_dst /*out */, hid_t
if (H5P_get(ocpy_plist, H5O_CPY_OPTION_NAME, &cpy_option) < 0)
HGOTO_ERROR(H5E_PLIST, H5E_CANTGET, FAIL, "can't get object copy flag")
- /* Retrieve the marge committed datatype list */
+ /* Retrieve the merge committed datatype list */
if (H5P_peek(ocpy_plist, H5O_CPY_MERGE_COMM_DT_LIST_NAME, &dt_list) < 0)
HGOTO_ERROR(H5E_PLIST, H5E_CANTGET, FAIL, "can't get merge committed datatype list")
diff --git a/src/H5Odrvinfo.c b/src/H5Odrvinfo.c
index 8bebc6d..69ab5b2 100644
--- a/src/H5Odrvinfo.c
+++ b/src/H5Odrvinfo.c
@@ -38,7 +38,7 @@ const H5O_msg_class_t H5O_MSG_DRVINFO[1] = {{
H5O_DRVINFO_ID, /*message id number */
"driver info", /*message name for debugging */
sizeof(H5O_drvinfo_t), /*native message size */
- 0, /* messages are sharable? */
+ 0, /* messages are shareable? */
H5O__drvinfo_decode, /*decode message */
H5O__drvinfo_encode, /*encode message */
H5O__drvinfo_copy, /*copy the native value */
diff --git a/src/H5Odtype.c b/src/H5Odtype.c
index 1be9522..fa49924 100644
--- a/src/H5Odtype.c
+++ b/src/H5Odtype.c
@@ -89,7 +89,7 @@ const H5O_msg_class_t H5O_MSG_DTYPE[1] = {{
H5O_DTYPE_ID, /* message id number */
"datatype", /* message name for debugging */
sizeof(H5T_t), /* native message size */
- H5O_SHARE_IS_SHARABLE | H5O_SHARE_IN_OHDR, /* messages are sharable? */
+ H5O_SHARE_IS_SHARABLE | H5O_SHARE_IN_OHDR, /* messages are shareable? */
H5O__dtype_shared_decode, /* decode message */
H5O__dtype_shared_encode, /* encode message */
H5O__dtype_copy, /* copy the native value */
diff --git a/src/H5Oefl.c b/src/H5Oefl.c
index d950249..931fe0b 100644
--- a/src/H5Oefl.c
+++ b/src/H5Oefl.c
@@ -41,7 +41,7 @@ const H5O_msg_class_t H5O_MSG_EFL[1] = {{
H5O_EFL_ID, /*message id number */
"external file list", /*message name for debugging */
sizeof(H5O_efl_t), /*native message size */
- 0, /* messages are sharable? */
+ 0, /* messages are shareable? */
H5O__efl_decode, /*decode message */
H5O__efl_encode, /*encode message */
H5O__efl_copy, /*copy native value */
diff --git a/src/H5Ofill.c b/src/H5Ofill.c
index 5068039..d905352 100644
--- a/src/H5Ofill.c
+++ b/src/H5Ofill.c
@@ -107,7 +107,7 @@ const H5O_msg_class_t H5O_MSG_FILL[1] = {{
H5O_FILL_ID, /*message id number */
"fill", /*message name for debugging */
sizeof(H5O_fill_t), /*native message size */
- H5O_SHARE_IS_SHARABLE | H5O_SHARE_IN_OHDR, /* messages are sharable? */
+ H5O_SHARE_IS_SHARABLE | H5O_SHARE_IN_OHDR, /* messages are shareable? */
H5O__fill_shared_decode, /*decode message */
H5O__fill_shared_encode, /*encode message */
H5O__fill_copy, /*copy the native value */
@@ -131,7 +131,7 @@ const H5O_msg_class_t H5O_MSG_FILL_NEW[1] = {{
H5O_FILL_NEW_ID, /*message id number */
"fill_new", /*message name for debugging */
sizeof(H5O_fill_t), /*native message size */
- H5O_SHARE_IS_SHARABLE | H5O_SHARE_IN_OHDR, /* messages are sharable? */
+ H5O_SHARE_IS_SHARABLE | H5O_SHARE_IN_OHDR, /* messages are shareable? */
H5O__fill_new_shared_decode, /*decode message */
H5O__fill_new_shared_encode, /*encode message */
H5O__fill_copy, /*copy the native value */
diff --git a/src/H5Ofsinfo.c b/src/H5Ofsinfo.c
index b60f589..7253f3e 100644
--- a/src/H5Ofsinfo.c
+++ b/src/H5Ofsinfo.c
@@ -44,7 +44,7 @@ const H5O_msg_class_t H5O_MSG_FSINFO[1] = {{
H5O_FSINFO_ID, /* message id number */
"fsinfo", /* message name for debugging */
sizeof(H5O_fsinfo_t), /* native message size */
- 0, /* messages are sharable? */
+ 0, /* messages are shareable? */
H5O__fsinfo_decode, /* decode message */
H5O__fsinfo_encode, /* encode message */
H5O__fsinfo_copy, /* copy the native value */
@@ -176,7 +176,7 @@ H5O__fsinfo_decode(H5F_t *f, H5O_t H5_ATTR_UNUSED *open_oh, unsigned H5_ATTR_UNU
H5F_DECODE_LENGTH(f, p, fsinfo->threshold); /* Free-space section threshold */
H5F_DECODE_LENGTH(f, p, fsinfo->page_size); /* File space page size */
- UINT16DECODE(p, fsinfo->pgend_meta_thres); /* Page end metdata threshold */
+ UINT16DECODE(p, fsinfo->pgend_meta_thres); /* Page end metadata threshold */
H5F_addr_decode(f, &p,
&(fsinfo->eoa_pre_fsm_fsalloc)); /* EOA before free-space header and section info */
diff --git a/src/H5Oginfo.c b/src/H5Oginfo.c
index 8540cc4..304890c 100644
--- a/src/H5Oginfo.c
+++ b/src/H5Oginfo.c
@@ -43,7 +43,7 @@ const H5O_msg_class_t H5O_MSG_GINFO[1] = {{
H5O_GINFO_ID, /*message id number */
"ginfo", /*message name for debugging */
sizeof(H5O_ginfo_t), /*native message size */
- 0, /* messages are sharable? */
+ 0, /* messages are shareable? */
H5O__ginfo_decode, /*decode message */
H5O__ginfo_encode, /*encode message */
H5O__ginfo_copy, /*copy the native value */
diff --git a/src/H5Olayout.c b/src/H5Olayout.c
index 651e317..c939e72 100644
--- a/src/H5Olayout.c
+++ b/src/H5Olayout.c
@@ -51,7 +51,7 @@ const H5O_msg_class_t H5O_MSG_LAYOUT[1] = {{
H5O_LAYOUT_ID, /* message id number */
"layout", /* message name for debugging */
sizeof(H5O_layout_t), /* native message size */
- 0, /* messages are sharable? */
+ 0, /* messages are shareable? */
H5O__layout_decode, /* decode message */
H5O__layout_encode, /* encode message */
H5O__layout_copy, /* copy the native value */
diff --git a/src/H5Olinfo.c b/src/H5Olinfo.c
index 711f199..eacc916 100644
--- a/src/H5Olinfo.c
+++ b/src/H5Olinfo.c
@@ -53,7 +53,7 @@ const H5O_msg_class_t H5O_MSG_LINFO[1] = {{
H5O_LINFO_ID, /*message id number */
"linfo", /*message name for debugging */
sizeof(H5O_linfo_t), /*native message size */
- 0, /* messages are sharable? */
+ 0, /* messages are shareable? */
H5O__linfo_decode, /*decode message */
H5O__linfo_encode, /*encode message */
H5O__linfo_copy, /*copy the native value */
diff --git a/src/H5Olink.c b/src/H5Olink.c
index 75456f4..51c44a3 100644
--- a/src/H5Olink.c
+++ b/src/H5Olink.c
@@ -56,7 +56,7 @@ const H5O_msg_class_t H5O_MSG_LINK[1] = {{
H5O_LINK_ID, /*message id number */
"link", /*message name for debugging */
sizeof(H5O_link_t), /*native message size */
- 0, /* messages are sharable? */
+ 0, /* messages are shareable? */
H5O__link_decode, /*decode message */
H5O__link_encode, /*encode message */
H5O__link_copy, /*copy the native value */
diff --git a/src/H5Omessage.c b/src/H5Omessage.c
index 15edf1e..fa20aa1 100644
--- a/src/H5Omessage.c
+++ b/src/H5Omessage.c
@@ -371,7 +371,7 @@ H5O__msg_write_real(H5F_t *f, H5O_t *oh, const H5O_msg_class_t *type, unsigned m
HDassert(((H5O_shared_t *)idx_msg->native)->type != H5O_SHARE_TYPE_COMMITTED);
/* Also, sanity check that a message doesn't switch status from being
- * shared (or sharable) to being unsharable. (Which could cause
+ * shared (or shareable) to being unshareable. (Which could cause
* a message to increase in size in the object header)
*/
HDassert(!(mesg_flags & H5O_MSG_FLAG_DONTSHARE));
@@ -1131,7 +1131,7 @@ done:
* Nov 19 2004
*
* Description:
- * This function interates over the object headers of an object
+ * This function iterates over the object headers of an object
* specified with 'loc' of type 'type_id'. For each object header of the
* object, the 'op_data' and some additional information (specified below) are
* passed to the 'op' function.
@@ -1194,7 +1194,7 @@ done:
* Sep 6 2005
*
* Description:
- * This function interates over the object headers of an object
+ * This function iterates over the object headers of an object
* specified with 'ent' of type 'type_id'. For each object header of the
* object, the 'op_data' and some additional information (specified below) are
* passed to the 'op' function.
@@ -1544,7 +1544,7 @@ H5O_msg_is_shared(unsigned type_id, const void *mesg)
HDassert(type);
HDassert(mesg);
- /* If messages in a class aren't sharable, then obviously this message isn't shared! :-) */
+ /* If messages in a class aren't shareable, then obviously this message isn't shared! :-) */
if (type->share_flags & H5O_SHARE_IS_SHARABLE)
ret_value = H5O_IS_STORED_SHARED(((const H5O_shared_t *)mesg)->type);
else
diff --git a/src/H5Omtime.c b/src/H5Omtime.c
index 163ea32..7d3c56a 100644
--- a/src/H5Omtime.c
+++ b/src/H5Omtime.c
@@ -43,7 +43,7 @@ const H5O_msg_class_t H5O_MSG_MTIME[1] = {{
H5O_MTIME_ID, /*message id number */
"mtime", /*message name for debugging */
sizeof(time_t), /*native message size */
- 0, /* messages are sharable? */
+ 0, /* messages are shareable? */
H5O__mtime_decode, /*decode message */
H5O__mtime_encode, /*encode message */
H5O__mtime_copy, /*copy the native value */
@@ -68,7 +68,7 @@ const H5O_msg_class_t H5O_MSG_MTIME_NEW[1] = {{
H5O_MTIME_NEW_ID, /*message id number */
"mtime_new", /*message name for debugging */
sizeof(time_t), /*native message size */
- 0, /* messages are sharable? */
+ 0, /* messages are shareable? */
H5O__mtime_new_decode, /*decode message */
H5O__mtime_new_encode, /*encode message */
H5O__mtime_copy, /*copy the native value */
diff --git a/src/H5Oname.c b/src/H5Oname.c
index 1636a0e..35578a9 100644
--- a/src/H5Oname.c
+++ b/src/H5Oname.c
@@ -43,7 +43,7 @@ const H5O_msg_class_t H5O_MSG_NAME[1] = {{
H5O_NAME_ID, /*message id number */
"name", /*message name for debugging */
sizeof(H5O_name_t), /*native message size */
- 0, /* messages are sharable? */
+ 0, /* messages are shareable? */
H5O__name_decode, /*decode message */
H5O__name_encode, /*encode message */
H5O__name_copy, /*copy the native value */
diff --git a/src/H5Onull.c b/src/H5Onull.c
index 9b377fb..0f3143c 100644
--- a/src/H5Onull.c
+++ b/src/H5Onull.c
@@ -31,7 +31,7 @@ const H5O_msg_class_t H5O_MSG_NULL[1] = {{
H5O_NULL_ID, /*message id number */
"null", /*message name for debugging */
0, /*native message size */
- 0, /* messages are sharable? */
+ 0, /* messages are shareable? */
NULL, /*no decode method */
NULL, /*no encode method */
NULL, /*no copy method */
diff --git a/src/H5Opkg.h b/src/H5Opkg.h
index 331fcf6..ebfe636 100644
--- a/src/H5Opkg.h
+++ b/src/H5Opkg.h
@@ -378,7 +378,7 @@ typedef struct H5O_chunk_proxy_t {
H5O_t * oh; /* Object header for this chunk */
unsigned chunkno; /* Chunk number for this chunk */
- /* Flush depencency parent information (not stored)
+ /* Flush dependency parent information (not stored)
*
* The following field is used to store a pointer
* to the in-core representation of a new chunk proxy's flush dependency
diff --git a/src/H5Opline.c b/src/H5Opline.c
index 221e323..243f454 100644
--- a/src/H5Opline.c
+++ b/src/H5Opline.c
@@ -67,7 +67,7 @@ const H5O_msg_class_t H5O_MSG_PLINE[1] = {{
H5O_PLINE_ID, /* message id number */
"filter pipeline", /* message name for debugging */
sizeof(H5O_pline_t), /* native message size */
- H5O_SHARE_IS_SHARABLE | H5O_SHARE_IN_OHDR, /* messages are sharable? */
+ H5O_SHARE_IS_SHARABLE | H5O_SHARE_IN_OHDR, /* messages are shareable? */
H5O__pline_shared_decode, /* decode message */
H5O__pline_shared_encode, /* encode message */
H5O__pline_copy, /* copy the native value */
@@ -86,7 +86,7 @@ const H5O_msg_class_t H5O_MSG_PLINE[1] = {{
H5O__pline_shared_debug /* debug the message */
}};
-/* Format version bounds for filter pipleline */
+/* Format version bounds for filter pipeline */
const unsigned H5O_pline_ver_bounds[] = {
H5O_PLINE_VERSION_1, /* H5F_LIBVER_EARLIEST */
H5O_PLINE_VERSION_2, /* H5F_LIBVER_V18 */
diff --git a/src/H5Oprivate.h b/src/H5Oprivate.h
index 050bd49..58b863d 100644
--- a/src/H5Oprivate.h
+++ b/src/H5Oprivate.h
@@ -251,7 +251,7 @@ typedef struct H5O_copy_t {
#define H5O_SHARE_TYPE_UNSHARED 0 /* Message is not shared */
#define H5O_SHARE_TYPE_SOHM 1 /* Message is stored in SOHM heap */
#define H5O_SHARE_TYPE_COMMITTED 2 /* Message is stored in another object header */
-#define H5O_SHARE_TYPE_HERE 3 /* Message is stored in this object header, but is sharable */
+#define H5O_SHARE_TYPE_HERE 3 /* Message is stored in this object header, but is shareable */
/* Detect messages that aren't stored in message's object header */
#define H5O_IS_STORED_SHARED(T) \
diff --git a/src/H5Opublic.h b/src/H5Opublic.h
index d9d0500..b05a2a8 100644
--- a/src/H5Opublic.h
+++ b/src/H5Opublic.h
@@ -80,7 +80,7 @@
#define H5O_SHMESG_MAX_LIST_SIZE 5000
/* Flags for H5Oget_info.
- * Theses flags determine which fields will be filled in in the H5O_info_t
+ * These flags determine which fields will be filled in the H5O_info_t
* struct.
*/
#define H5O_INFO_BASIC 0x0001u /**< Fill in the fileno, addr, type, and rc fields */
@@ -90,8 +90,8 @@
//! <!-- [H5O_native_info_fields_snip] -->
/**
- * Flags for H5Oget_native_info(). Theses flags determine which fields will be
- * filled in in the \ref H5O_native_info_t struct.
+ * Flags for H5Oget_native_info(). These flags determine which fields will be
+ * filled in the \ref H5O_native_info_t struct.
*/
#define H5O_NATIVE_INFO_HDR 0x0008u /**< Fill in the hdr field */
#define H5O_NATIVE_INFO_META_SIZE 0x0010u /**< Fill in the meta_size field */
@@ -1166,7 +1166,7 @@ H5_DLL ssize_t H5Oget_comment_by_name(hid_t loc_id, const char *name, char *comm
* <em>best effort</em> setting. If the application passes in
* a value indicating iteration in creation order and a group is
* encountered that was not tracked in creation order, that group
- * will be iterated over in alpha-numeric order by name, or
+ * will be iterated over in alphanumeric order by name, or
* <em>name order</em>. (<em>Name order</em> is the native order
* used by the HDF5 library and is always available.)
*
@@ -1265,7 +1265,7 @@ H5_DLL herr_t H5Ovisit3(hid_t obj_id, H5_index_t idx_type, H5_iter_order_t order
* <em>best effort</em> setting. If the application passes in a
* value indicating iteration in creation order and a group is
* encountered that was not tracked in creation order, that group
- * will be iterated over in alpha-numeric order by name, or
+ * will be iterated over in alphanumeric order by name, or
* <em>name order</em>. (<em>Name order</em> is the native order
* used by the HDF5 library and is always available.)
*
@@ -1978,7 +1978,7 @@ H5_DLL herr_t H5Oget_info2(hid_t loc_id, H5O_info1_t *oinfo, unsigned fields);
* in the H5Oget_info1() function entry.
*
* The \p fields parameter contains flags to determine which fields
- * will be filled in in the H5O_info1_t \c struct returned in
+ * will be filled in the H5O_info1_t \c struct returned in
* \p oinfo. These flags are defined in the H5Opublic.h file:
*
* \obj_info_fields
@@ -2096,7 +2096,7 @@ H5_DLL herr_t H5Oget_info_by_idx2(hid_t loc_id, const char *group_name, H5_index
* <em>best effort</em> setting. If the application passes in
* a value indicating iteration in creation order and a group is
* encountered that was not tracked in creation order, that group
- * will be iterated over in alpha-numeric order by name, or
+ * will be iterated over in alphanumeric order by name, or
* <em>name order</em>. (<em>Name order</em> is the native order
* used by the HDF5 library and is always available.)
*
@@ -2188,7 +2188,7 @@ H5_DLL herr_t H5Ovisit1(hid_t obj_id, H5_index_t idx_type, H5_iter_order_t order
* <em>best effort</em> setting. If the application passes in a
* value indicating iteration in creation order and a group is
* encountered that was not tracked in creation order, that group
- * will be iterated over in alpha-numeric order by name, or
+ * will be iterated over in alphanumeric order by name, or
* <em>name order</em>. (<em>Name order</em> is the native order
* used by the HDF5 library and is always available.)
*
@@ -2291,7 +2291,7 @@ H5_DLL herr_t H5Ovisit_by_name1(hid_t loc_id, const char *obj_name, H5_index_t i
* <em>best effort</em> setting. If the application passes in
* a value indicating iteration in creation order and a group is
* encountered that was not tracked in creation order, that group
- * will be iterated over in alpha-numeric order by name, or
+ * will be iterated over in alphanumeric order by name, or
* <em>name order</em>. (<em>Name order</em> is the native order
* used by the HDF5 library and is always available.)
*
@@ -2385,7 +2385,7 @@ H5_DLL herr_t H5Ovisit2(hid_t obj_id, H5_index_t idx_type, H5_iter_order_t order
* <em>best effort</em> setting. If the application passes in a
* value indicating iteration in creation order and a group is
* encountered that was not tracked in creation order, that group
- * will be iterated over in alpha-numeric order by name, or
+ * will be iterated over in alphanumeric order by name, or
* <em>name order</em>. (<em>Name order</em> is the native order
* used by the HDF5 library and is always available.)
*
diff --git a/src/H5Orefcount.c b/src/H5Orefcount.c
index 52eee14..500ab0f 100644
--- a/src/H5Orefcount.c
+++ b/src/H5Orefcount.c
@@ -45,7 +45,7 @@ const H5O_msg_class_t H5O_MSG_REFCOUNT[1] = {{
H5O_REFCOUNT_ID, /*message id number */
"refcount", /*message name for debugging */
sizeof(H5O_refcount_t), /*native message size */
- 0, /* messages are sharable? */
+ 0, /* messages are shareable? */
H5O__refcount_decode, /*decode message */
H5O__refcount_encode, /*encode message */
H5O__refcount_copy, /*copy the native value */
diff --git a/src/H5Osdspace.c b/src/H5Osdspace.c
index dab989f..574103d 100644
--- a/src/H5Osdspace.c
+++ b/src/H5Osdspace.c
@@ -61,7 +61,7 @@ const H5O_msg_class_t H5O_MSG_SDSPACE[1] = {{
H5O_SDSPACE_ID, /* message id number */
"dataspace", /* message name for debugging */
sizeof(H5S_extent_t), /* native message size */
- H5O_SHARE_IS_SHARABLE | H5O_SHARE_IN_OHDR, /* messages are sharable? */
+ H5O_SHARE_IS_SHARABLE | H5O_SHARE_IN_OHDR, /* messages are shareable? */
H5O__sdspace_shared_decode, /* decode message */
H5O__sdspace_shared_encode, /* encode message */
H5O__sdspace_copy, /* copy the native value */
diff --git a/src/H5Oshmesg.c b/src/H5Oshmesg.c
index 1c894a5..a4e8b8f 100644
--- a/src/H5Oshmesg.c
+++ b/src/H5Oshmesg.c
@@ -37,7 +37,7 @@ const H5O_msg_class_t H5O_MSG_SHMESG[1] = {{
H5O_SHMESG_ID, /*message id number */
"shared message table", /*message name for debugging */
sizeof(H5O_shmesg_table_t), /*native message size */
- 0, /* messages are sharable? */
+ 0, /* messages are shareable? */
H5O__shmesg_decode, /*decode message */
H5O__shmesg_encode, /*encode message */
H5O__shmesg_copy, /*copy the native value */
diff --git a/src/H5Ostab.c b/src/H5Ostab.c
index fa030bd..213b8a0 100644
--- a/src/H5Ostab.c
+++ b/src/H5Ostab.c
@@ -51,7 +51,7 @@ const H5O_msg_class_t H5O_MSG_STAB[1] = {{
H5O_STAB_ID, /*message id number */
"stab", /*message name for debugging */
sizeof(H5O_stab_t), /*native message size */
- 0, /* messages are sharable? */
+ 0, /* messages are shareable? */
H5O__stab_decode, /*decode message */
H5O__stab_encode, /*encode message */
H5O__stab_copy, /*copy the native value */
diff --git a/src/H5Ounknown.c b/src/H5Ounknown.c
index 2b8a7bb..dc43a6c 100644
--- a/src/H5Ounknown.c
+++ b/src/H5Ounknown.c
@@ -37,7 +37,7 @@ const H5O_msg_class_t H5O_MSG_UNKNOWN[1] = {{
H5O_UNKNOWN_ID, /*message id number */
"unknown", /*message name for debugging */
0, /*native message size */
- 0, /* messages are sharable? */
+ 0, /* messages are shareable? */
NULL, /*decode message */
NULL, /*encode message */
NULL, /*copy the native value */
diff --git a/src/H5P.c b/src/H5P.c
index 5717b8c..c671ea0 100644
--- a/src/H5P.c
+++ b/src/H5P.c
@@ -675,7 +675,7 @@ done:
NAME
H5Pexist
PURPOSE
- Routine to query the existance of a property in a property object.
+ Routine to query the existence of a property in a property object.
USAGE
htri_t H5P_exist(id, name)
hid_t id; IN: Property object ID to check
@@ -709,7 +709,7 @@ H5Pexist(hid_t id, const char *name)
if (!name || !*name)
HGOTO_ERROR(H5E_ARGS, H5E_BADVALUE, FAIL, "invalid property name");
- /* Check for the existance of the property in the list or class */
+ /* Check for the existence of the property in the list or class */
if (H5I_GENPROP_LST == H5I_get_type(id)) {
if (NULL == (plist = (H5P_genplist_t *)H5I_object(id)))
HGOTO_ERROR(H5E_ARGS, H5E_BADTYPE, FAIL, "not a property list");
@@ -855,7 +855,7 @@ done:
Failure: H5I_INVALID_HID (negative)
DESCRIPTION
Decodes a property list from a binary buffer. The contents of the buffer
- contain the values for the correponding properties of the plist. The decode
+ contain the values for the corresponding properties of the plist. The decode
callback of a certain property decodes its value from the buffer and sets it
in the property list.
GLOBAL VARIABLES
@@ -1296,7 +1296,7 @@ done:
Returns non-negative on success, negative on failure.
DESCRIPTION
Removes a property from a property list. Both properties which were
- in existance when the property list was created (i.e. properties registered
+ in existence when the property list was created (i.e. properties registered
with H5Pregister2) and properties added to the list after it was created
(i.e. added with H5Pinsert2) may be removed from a property list.
Properties do not need to be removed a property list before the list itself
diff --git a/src/H5PL.c b/src/H5PL.c
index 30b6c52..b052078 100644
--- a/src/H5PL.c
+++ b/src/H5PL.c
@@ -67,7 +67,7 @@
* the special "::" string.
*
* Return: Success: Non-negative
- * Failture: Negative
+ * Failure: Negative
*
*-------------------------------------------------------------------------
*/
@@ -99,7 +99,7 @@ done:
* Positive if one or more of the plugin types are enabled
*
* Return: Success: Non-negative
- * Failture: Negative
+ * Failure: Negative
*
*-------------------------------------------------------------------------
*/
@@ -128,7 +128,7 @@ done:
* Purpose: Insert a plugin search path at the end of the list.
*
* Return: Success: Non-negative
- * Failture: Negative
+ * Failure: Negative
*
*-------------------------------------------------------------------------
*/
@@ -160,7 +160,7 @@ done:
* Purpose: Insert a plugin search path at the beginning of the list.
*
* Return: Success: Non-negative
- * Failture: Negative
+ * Failure: Negative
*
*-------------------------------------------------------------------------
*/
@@ -234,7 +234,7 @@ done:
* other paths after the index.
*
* Return: Success: Non-negative
- * Failture: Negative
+ * Failure: Negative
*
*-------------------------------------------------------------------------
*/
@@ -270,11 +270,11 @@ done:
/*-------------------------------------------------------------------------
* Function: H5PLremove
*
- * Purpose: Remove the plugin path at the specifed index and compact
+ * Purpose: Remove the plugin path at the specified index and compact
* the list.
*
* Return: Success: Non-negative
- * Failture: Negative
+ * Failure: Negative
*
* Return: Non-negative or success.
*
diff --git a/src/H5PLpath.c b/src/H5PLpath.c
index 9aecad1..87ff831 100644
--- a/src/H5PLpath.c
+++ b/src/H5PLpath.c
@@ -322,7 +322,7 @@ H5PL__close_path_table(void)
* Purpose: Gets the number of plugin paths that have been stored.
*
* Return: Success: The number of paths
- * Failture: Can't fail
+ * Failure: Can't fail
*-------------------------------------------------------------------------
*/
unsigned
diff --git a/src/H5Pdxpl.c b/src/H5Pdxpl.c
index 42dd6e4..c901931 100644
--- a/src/H5Pdxpl.c
+++ b/src/H5Pdxpl.c
@@ -1991,7 +1991,7 @@ done:
/*-------------------------------------------------------------------------
* Function: H5Pget_mpio_actual_io_mode
*
- * Purpose: Retrieves the type of I/O actually preformed when collective I/O
+ * Purpose: Retrieves the type of I/O actually performed when collective I/O
* is requested.
*
* Return: Non-negative on success/Negative on failure
diff --git a/src/H5Pencdec.c b/src/H5Pencdec.c
index e2a97f8..01573d1 100644
--- a/src/H5Pencdec.c
+++ b/src/H5Pencdec.c
@@ -663,7 +663,7 @@ done:
on failure.
DESCRIPTION
Decodes a property list from a binary buffer. The contents of the buffer
- contain the values for the correponding properties of the plist. The decode
+ contain the values for the corresponding properties of the plist. The decode
callback of a certain property decodes its value from the buffer and sets it
in the property list.
GLOBAL VARIABLES
diff --git a/src/H5Pfapl.c b/src/H5Pfapl.c
index 98f9ed9..47c17db 100644
--- a/src/H5Pfapl.c
+++ b/src/H5Pfapl.c
@@ -523,7 +523,7 @@ static const size_t H5F_def_page_buf_size_g = H5F_ACS_PAGE_BUFFER_SIZE_DEF; /*
static const unsigned H5F_def_page_buf_min_meta_perc_g =
H5F_ACS_PAGE_BUFFER_MIN_META_PERC_DEF; /* Default page buffer minimum metadata size */
static const unsigned H5F_def_page_buf_min_raw_perc_g =
- H5F_ACS_PAGE_BUFFER_MIN_RAW_PERC_DEF; /* Default page buffer mininum raw data size */
+ H5F_ACS_PAGE_BUFFER_MIN_RAW_PERC_DEF; /* Default page buffer minimum raw data size */
static const hbool_t H5F_def_use_file_locking_g =
H5F_ACS_USE_FILE_LOCKING_DEF; /* Default use file locking flag */
static const hbool_t H5F_def_ignore_disabled_file_locks_g =
@@ -935,7 +935,7 @@ done:
* `driver_id`. Otherwise, `driver_id` is set to
* H5I_INVALID_HID.
*
- * Return: Non-negative on sucess/Negative on failure
+ * Return: Non-negative on success/Negative on failure
*
*-------------------------------------------------------------------------
*/
@@ -2869,7 +2869,7 @@ done:
*
* Currently, the only two valid combinations for this routine are:
* LOW = H5F_FORMAT_EARLIEST and HIGH = H5F_FORMAT_LATEST (the default
- * setting, which creates objects with the ealiest version possible for
+ * setting, which creates objects with the earliest version possible for
* each object, but no upper limit on the version allowed to be created if
* a newer version of an object's format is required to support a feature
* requested with an HDF5 library API routine), and LOW = H5F_FORMAT_LATEST
@@ -2885,7 +2885,7 @@ done:
* Note: Eventually we want to add more values to the H5F_libver_t
* enumerated type that indicate library release values where the file
* format was changed (like "H5F_FORMAT_1_2_0" for the file format changes
- * in the 1.2.x release branch and possily even "H5F_FORMAT_1_4_2" for
+ * in the 1.2.x release branch and possibly even "H5F_FORMAT_1_4_2" for
* a change mid-way through the 1.4.x release branch, etc).
*
* Adding more values will allow applications to make settings like the
@@ -3331,7 +3331,7 @@ H5Pset_file_image_callbacks(hid_t fapl_id, H5FD_file_image_callbacks_t *callback
HDassert(callbacks_ptr->udata_copy);
HDassert(callbacks_ptr->udata_free);
if ((info.callbacks.udata = callbacks_ptr->udata_copy(callbacks_ptr->udata)) == NULL)
- HGOTO_ERROR(H5E_PLIST, H5E_CANTSET, FAIL, "can't copy the suppplied udata")
+ HGOTO_ERROR(H5E_PLIST, H5E_CANTSET, FAIL, "can't copy the supplied udata")
} /* end if */
/* Set values */
diff --git a/src/H5Pint.c b/src/H5Pint.c
index 614f4d3..cc3f10b 100644
--- a/src/H5Pint.c
+++ b/src/H5Pint.c
@@ -3303,7 +3303,7 @@ done:
NAME
H5P_exist_plist
PURPOSE
- Internal routine to query the existance of a property in a property list.
+ Internal routine to query the existence of a property in a property list.
USAGE
htri_t H5P_exist_plist(plist, name)
const H5P_genplist_t *plist; IN: Property list to check
@@ -3362,7 +3362,7 @@ done:
NAME
H5P__exist_pclass
PURPOSE
- Internal routine to query the existance of a property in a property class.
+ Internal routine to query the existence of a property in a property class.
USAGE
herr_t H5P__exist_pclass(pclass, name)
H5P_genclass_t *pclass; IN: Property class to check
@@ -3877,7 +3877,7 @@ H5P__cmp_plist_cb(H5P_genprop_t *prop, void *_udata)
/* Check if the property exists in the second property list */
if ((prop2_exist = H5P_exist_plist(udata->plist2, prop->name)) < 0)
- HGOTO_ERROR(H5E_PLIST, H5E_NOTFOUND, H5_ITER_ERROR, "can't lookup existance of property?")
+ HGOTO_ERROR(H5E_PLIST, H5E_NOTFOUND, H5_ITER_ERROR, "can't lookup existence of property?")
if (prop2_exist) {
const H5P_genprop_t *prop2; /* Pointer to property in second plist */
@@ -4828,7 +4828,7 @@ done:
Returns non-negative on success, negative on failure.
DESCRIPTION
Removes a property from a property list. Both properties which were
- in existance when the property list was created (i.e. properties registered
+ in existence when the property list was created (i.e. properties registered
with H5Pregister2) and properties added to the list after it was created
(i.e. added with H5Pinsert2) may be removed from a property list.
Properties do not need to be removed a property list before the list itself
@@ -4909,7 +4909,7 @@ H5P__copy_prop_plist(hid_t dst_id, hid_t src_id, const char *name)
NULL == (dst_plist = (H5P_genplist_t *)H5I_object(dst_id)))
HGOTO_ERROR(H5E_PLIST, H5E_NOTFOUND, FAIL, "property object doesn't exist")
- /* If the property exists in the destination alread */
+ /* If the property exists in the destination already */
if (NULL != H5P__find_prop_plist(dst_plist, name)) {
/* Delete the property from the destination list, calling the 'close' callback if necessary */
if (H5P_remove(dst_plist, name) < 0)
diff --git a/src/H5Pocpl.c b/src/H5Pocpl.c
index e442030..06d9f85 100644
--- a/src/H5Pocpl.c
+++ b/src/H5Pocpl.c
@@ -482,7 +482,7 @@ done:
* transfer property list. The FLAGS argument specifies certain
* general properties of the filter and is documented below.
* The CD_VALUES is an array of CD_NELMTS integers which are
- * auxiliary data for the filter. The integer vlues will be
+ * auxiliary data for the filter. The integer values will be
* stored in the dataset object header as part of the filter
* information.
*
@@ -543,7 +543,7 @@ done:
* transfer property list. The FLAGS argument specifies certain
* general properties of the filter and is documented below.
* The CD_VALUES is an array of CD_NELMTS integers which are
- * auxiliary data for the filter. The integer vlues will be
+ * auxiliary data for the filter. The integer values will be
* stored in the dataset object header as part of the filter
* information.
*
@@ -609,7 +609,7 @@ done:
* creation property list. The FLAGS argument specifies certain
* general properties of the filter and is documented below.
* The CD_VALUES is an array of CD_NELMTS integers which are
- * auxiliary data for the filter. The integer vlues will be
+ * auxiliary data for the filter. The integer values will be
* stored in the dataset object header as part of the filter
* information.
*
@@ -671,7 +671,7 @@ done:
* creation property list. The FLAGS argument specifies certain
* general properties of the filter and is documented below.
* The CD_VALUES is an array of CD_NELMTS integers which are
- * auxiliary data for the filter. The integer vlues will be
+ * auxiliary data for the filter. The integer values will be
* stored in the dataset object header as part of the filter
* information.
*
@@ -782,7 +782,7 @@ done:
* CD_NELMTS indicates the number of entries in the CD_VALUES
* array allocated by the caller while on exit it contains the
* number of values defined by the filter. FILTER_CONFIG is a bit
- * field contaning encode/decode flags from H5Zpublic.h. The IDX
+ * field containing encode/decode flags from H5Zpublic.h. The IDX
* should be a value between zero and N-1 as described for
* H5Pget_nfilters() and the function will return failure if the
* filter number is out of range.
@@ -867,7 +867,7 @@ done:
* CD_NELMTS indicates the number of entries in the CD_VALUES
* array allocated by the caller while on exit it contains the
* number of values defined by the filter. FILTER_CONFIG is a bit
- * field contaning encode/decode flags from H5Zpublic.h. The ID
+ * field containing encode/decode flags from H5Zpublic.h. The ID
* should be the filter ID to retrieve the parameters for. If the
* filter is not set for the property list, an error will be returned.
*
@@ -916,7 +916,7 @@ done:
* CD_NELMTS indicates the number of entries in the CD_VALUES
* array allocated by the caller while on exit it contains the
* number of values defined by the filter. FILTER_CONFIG is a bit
- * field contaning encode/decode flags from H5Zpublic.h. The ID
+ * field containing encode/decode flags from H5Zpublic.h. The ID
* should be the filter ID to retrieve the parameters for. If the
* filter is not set for the property list, an error will be returned.
*
diff --git a/src/H5Ppublic.h b/src/H5Ppublic.h
index 53fe5ac..536407c 100644
--- a/src/H5Ppublic.h
+++ b/src/H5Ppublic.h
@@ -3912,13 +3912,13 @@ H5_DLL herr_t H5Pget_meta_block_size(hid_t fapl_id, hsize_t *size);
*
* The second example illustrates the two cases for retrieving the
* number of read attempts from the file access property list of a file
- * opened with SWMR acccess.
+ * opened with SWMR access.
*
* \include H5Pget_metadata_read_attempts.2.c
*
* The third example illustrates the two cases for retrieving the number
* of read attempts from the file access property list of a file opened
- * with non-SWMR acccess.
+ * with non-SWMR access.
*
* \include H5Pget_metadata_read_attempts.3.c
*
@@ -4374,7 +4374,7 @@ H5_DLL herr_t H5Pset_driver_by_value(hid_t plist_id, H5FD_class_value_t driver_v
* The <em>external link open file cache</em> holds files open after
* they have been accessed via an external link. This cache reduces
* the number of times such files are opened when external links are
- * accessed repeatedly and can siginificantly improves performance in
+ * accessed repeatedly and can significantly improves performance in
* certain heavy-use situations and when low-level file opens or closes
* are expensive.
*
diff --git a/src/H5S.c b/src/H5S.c
index 835ab81..75d3399 100644
--- a/src/H5S.c
+++ b/src/H5S.c
@@ -113,7 +113,7 @@ H5S_init(void)
if (H5I_register_type(H5I_DATASPACE_CLS) < 0)
HGOTO_ERROR(H5E_DATASPACE, H5E_CANTINIT, FAIL, "unable to initialize dataspace ID class")
- /* Initialize the ID group for the dataspace selction iterator IDs */
+ /* Initialize the ID group for the dataspace selection iterator IDs */
if (H5I_register_type(H5I_SPACE_SEL_ITER_CLS) < 0)
HGOTO_ERROR(H5E_DATASPACE, H5E_CANTINIT, FAIL,
"unable to initialize dataspace selection iterator ID class")
diff --git a/src/H5SL.c b/src/H5SL.c
index 5752fdc..e12e21a 100644
--- a/src/H5SL.c
+++ b/src/H5SL.c
@@ -239,7 +239,7 @@
}
/* Macro used to reduce the level of a node by 1. Does not update the head node
- * "current level". PREV is the previous node of the currrent height of X. */
+ * "current level". PREV is the previous node of the current height of X. */
#define H5SL_DEMOTE(X, PREV) \
{ \
size_t _lvl = X->level; \
@@ -255,7 +255,7 @@
#define H5SL_INSERT(CMP, SLIST, X, TYPE, KEY, HASHVAL) \
{ \
H5SL_node_t *_last = X; /* Lowest node in the current gap */ \
- H5SL_node_t *_next = NULL; /* Highest node in the currect gap */ \
+ H5SL_node_t *_next = NULL; /* Highest node in the current gap */ \
H5SL_node_t *_drop; /* Low node of the gap to drop into */ \
int _count; /* Number of nodes in the current gap */ \
int _i; \
@@ -311,7 +311,7 @@
{ \
H5SL_node_t *_last = X; /* Lowest node in the current gap */ \
H5SL_node_t *_llast = X; /* Lowest node in the previous gap */ \
- H5SL_node_t *_next = NULL; /* Highest node in the currect gap */ \
+ H5SL_node_t *_next = NULL; /* Highest node in the current gap */ \
H5SL_node_t *_drop = NULL; /* Low node of the gap to drop into */ \
H5SL_node_t *_ldrop = NULL; /* Low node of gap before the one to drop into */ \
H5SL_node_t *_head = SLIST->header; /* Head of the skip list */ \
@@ -1929,7 +1929,7 @@ H5SL_next(H5SL_node_t *slist_node)
NAME
H5SL_prev
PURPOSE
- Gets a pointer to the previos node in a skip list
+ Gets a pointer to the previous node in a skip list
USAGE
H5SL_node_t *H5SL_prev(slist_node)
H5SL_node_t *slist_node; IN: Pointer to skip list node
diff --git a/src/H5SM.c b/src/H5SM.c
index 45d72d4..d7ae84b 100644
--- a/src/H5SM.c
+++ b/src/H5SM.c
@@ -114,7 +114,7 @@ H5SM_init(H5F_t *f, H5P_genplist_t *fc_plist, const H5O_loc_t *ext_loc)
haddr_t table_addr = HADDR_UNDEF; /* Address of SOHM master table in file */
unsigned list_max, btree_min; /* Phase change limits for SOHM indices */
unsigned index_type_flags[H5O_SHMESG_MAX_NINDEXES]; /* Messages types stored in each index */
- unsigned minsizes[H5O_SHMESG_MAX_NINDEXES]; /* Message size sharing threshhold for each index */
+ unsigned minsizes[H5O_SHMESG_MAX_NINDEXES]; /* Message size sharing threshold for each index */
unsigned type_flags_used; /* Message type flags used, for sanity checking */
unsigned x; /* Local index variable */
herr_t ret_value = SUCCEED; /* Return value */
@@ -1596,7 +1596,7 @@ H5SM_delete(H5F_t *f, H5O_t *open_oh, H5O_shared_t *sh_mesg)
*/
if (H5SM__delete_from_index(f, open_oh, &(table->indexes[index_num]), sh_mesg, &cache_flags, &mesg_size,
&mesg_buf) < 0)
- HGOTO_ERROR(H5E_SOHM, H5E_CANTDELETE, FAIL, "unable to delete mesage from SOHM index")
+ HGOTO_ERROR(H5E_SOHM, H5E_CANTDELETE, FAIL, "unable to delete message from SOHM index")
/* Release the master SOHM table */
if (H5AC_unprotect(f, H5AC_SOHM_TABLE, H5F_SOHM_ADDR(f), table, cache_flags) < 0)
diff --git a/src/H5Shyper.c b/src/H5Shyper.c
index af2794a..0c765d7 100644
--- a/src/H5Shyper.c
+++ b/src/H5Shyper.c
@@ -2060,7 +2060,7 @@ H5S__hyper_iter_get_seq_list_opt(H5S_sel_iter_t *iter, size_t maxseq, size_t max
/* Increment the offset and count for the other dimensions */
temp_dim = (int)fast_dim - 1;
while (temp_dim >= 0) {
- /* Move to the next row in the curent dimension */
+ /* Move to the next row in the current dimension */
offset[temp_dim]++;
tmp_block[temp_dim]++;
@@ -2178,7 +2178,7 @@ H5S__hyper_iter_get_seq_list_opt(H5S_sel_iter_t *iter, size_t maxseq, size_t max
/* Increment the offset and count for the other dimensions */
temp_dim = (int)fast_dim - 1;
while (temp_dim >= 0) {
- /* Move to the next row in the curent dimension */
+ /* Move to the next row in the current dimension */
offset[temp_dim]++;
tmp_block[temp_dim]++;
@@ -2492,7 +2492,7 @@ H5S__hyper_iter_get_seq_list_single(H5S_sel_iter_t *iter, size_t maxseq, size_t
/* Increment the offset and count for the other dimensions */
while (temp_dim >= 0) {
- /* Move to the next row in the curent dimension */
+ /* Move to the next row in the current dimension */
offset[temp_dim]++;
tmp_block[temp_dim]++;
@@ -3584,7 +3584,7 @@ H5S__hyper_get_version_enc_size(H5S_t *space, hsize_t block_count, uint32_t *ver
H5F_libver_t low_bound; /* The 'low' bound of library format versions */
H5F_libver_t high_bound; /* The 'high' bound of library format versions */
htri_t is_regular; /* A regular hyperslab or not */
- uint32_t tmp_version; /* Local temporay version */
+ uint32_t tmp_version; /* Local temporary version */
unsigned u; /* Local index variable */
herr_t ret_value = SUCCEED; /* Return value */
@@ -3972,7 +3972,7 @@ H5S__hyper_serialize(H5S_t *space, uint8_t **p)
pp = (*p);
HDassert(pp);
- /* Set some convienence values */
+ /* Set some convenience values */
ndims = space->extent.rank;
diminfo = space->select.sel_info.hslab->diminfo.opt;
@@ -4075,7 +4075,7 @@ H5S__hyper_serialize(H5S_t *space, uint8_t **p)
else {
HDassert(version == H5S_HYPER_VERSION_1);
- /* Set some convienence values */
+ /* Set some convenience values */
fast_dim = ndims - 1;
/* Encode number of hyperslabs */
@@ -4225,7 +4225,7 @@ H5S__hyper_deserialize(H5S_t **space, const uint8_t **p)
{
H5S_t *tmp_space = NULL; /* Pointer to actual dataspace to use,
either *space or a newly allocated one */
- hsize_t dims[H5S_MAX_RANK]; /* Dimenion sizes */
+ hsize_t dims[H5S_MAX_RANK]; /* Dimension sizes */
hsize_t start[H5S_MAX_RANK]; /* hyperslab start information */
hsize_t block[H5S_MAX_RANK]; /* hyperslab block information */
uint32_t version; /* Version number */
@@ -4618,7 +4618,7 @@ H5S__get_select_hyper_blocklist(H5S_t *space, hsize_t startblock, hsize_t numblo
hbool_t done; /* Whether we are done with the iteration */
unsigned u; /* Counter */
- /* Set some convienence values */
+ /* Set some convenience values */
ndims = space->extent.rank;
fast_dim = ndims - 1;
@@ -5112,7 +5112,7 @@ H5S__hyper_is_contiguous(const H5S_t *space)
large_contiguous = TRUE; /* assume true and reset if the dimensions don't match */
small_contiguous = FALSE; /* assume false initially */
- /* Check for a "large contigous" block */
+ /* Check for a "large contiguous" block */
for (u = 0; u < space->extent.rank; u++) {
if (diminfo[u].count > 1) {
large_contiguous = FALSE;
@@ -11516,7 +11516,7 @@ H5S__hyper_proj_int_iterate(const H5S_hyper_span_info_t *ss_span_info,
udata->skip += (count - 1) * (udata->skip - old_skip);
} /* end if */
else {
- /* Third case: agorithm added skip and nelem (in that
+ /* Third case: algorithm added skip and nelem (in that
* order). Add the same skip and nelem once for each item
* remaining in count. */
hsize_t skip_add;
diff --git a/src/H5Smpio.c b/src/H5Smpio.c
index b626b77..aec5560 100644
--- a/src/H5Smpio.c
+++ b/src/H5Smpio.c
@@ -1145,7 +1145,7 @@ H5S__obtain_datatype(H5S_hyper_span_info_t *spans, const hsize_t *down, size_t e
/* If this is the fastest changing dimension, it is the base case for derived datatype. */
span = spans->head;
if (NULL == span->down) {
- hbool_t large_block = FALSE; /* Wether the block length is larger than 32 bit integer */
+ hbool_t large_block = FALSE; /* Whether the block length is larger than 32 bit integer */
outercount = 0;
while (span) {
diff --git a/src/H5Spkg.h b/src/H5Spkg.h
index 74013bc..afabf25 100644
--- a/src/H5Spkg.h
+++ b/src/H5Spkg.h
@@ -201,7 +201,7 @@ struct H5S_hyper_span_info_t {
typedef enum {
H5S_DIMINFO_VALID_IMPOSSIBLE, /* 0: diminfo is not valid and can never be valid with the current selection
*/
- H5S_DIMINFO_VALID_NO, /* 1: diminfo is not valid but may or may not be possible to constuct */
+ H5S_DIMINFO_VALID_NO, /* 1: diminfo is not valid but may or may not be possible to construct */
H5S_DIMINFO_VALID_YES /* 2: diminfo is valid */
} H5S_diminfo_valid_t;
diff --git a/src/H5Spoint.c b/src/H5Spoint.c
index 72f64fe..240b722 100644
--- a/src/H5Spoint.c
+++ b/src/H5Spoint.c
@@ -1505,7 +1505,7 @@ done:
dataspace they are located within. The list of points is formatted as
follows: <coordinate> followed by the next coordinate, etc. until all the
point information in the selection have been put into the user's buffer.
- The points are returned in the order they will be interated through
+ The points are returned in the order they will be iterated through
when a selection is read/written from/to disk.
GLOBAL VARIABLES
COMMENTS, BUGS, ASSUMPTIONS
@@ -1579,7 +1579,7 @@ H5S__get_select_elem_pointlist(const H5S_t *space, hsize_t startpoint, hsize_t n
dataspace they are located within. The list of points is formatted as
follows: <coordinate> followed by the next coordinate, etc. until all the
point information in the selection have been put into the user's buffer.
- The points are returned in the order they will be interated through
+ The points are returned in the order they will be iterated through
when a selection is read/written from/to disk.
GLOBAL VARIABLES
COMMENTS, BUGS, ASSUMPTIONS
diff --git a/src/H5Sselect.c b/src/H5Sselect.c
index 9b35cd3..bcf12a0 100644
--- a/src/H5Sselect.c
+++ b/src/H5Sselect.c
@@ -2246,7 +2246,7 @@ H5S_select_construct_projection(H5S_t *base_space, H5S_t **new_space_ptr, unsign
if (new_space_rank == 0) {
hssize_t npoints; /* Number of points selected */
- /* Retreve the number of elements selected */
+ /* Retrieve the number of elements selected */
if ((npoints = (hssize_t)H5S_GET_SELECT_NPOINTS(base_space)) < 0)
HGOTO_ERROR(H5E_DATASPACE, H5E_CANTGET, FAIL, "unable to get number of points selected")
HDassert(npoints <= 1);
@@ -2315,12 +2315,12 @@ H5S_select_construct_projection(H5S_t *base_space, H5S_t **new_space_ptr, unsign
* true on selections of different rank iff:
*
* 1) the selection in the lower rank dataspace matches that
- * in the dimensions with the fastest changing indicies in
+ * in the dimensions with the fastest changing indices in
* the larger rank dataspace, and
*
* 2) the selection has thickness 1 in all ranks that appear
* only in the higher rank dataspace (i.e. those with
- * more slowly changing indicies).
+ * more slowly changing indices).
*/
if (new_space_rank > base_space_rank) {
hsize_t tmp_dim_size = 1; /* Temporary dimension value, for filling arrays */
diff --git a/src/H5Stest.c b/src/H5Stest.c
index a851549..6a31f3c 100644
--- a/src/H5Stest.c
+++ b/src/H5Stest.c
@@ -185,12 +185,13 @@ H5S__check_spans_tail_ptr(const H5S_hyper_span_info_t *span_lst)
if (NULL != cur_elem->down)
if ((ret_value = H5S__check_spans_tail_ptr(cur_elem->down)) < 0)
HGOTO_ERROR(H5E_DATASPACE, H5E_INCONSISTENTSTATE, FAIL,
- "the seletion has inconsistent tail pointers")
+ "the selection has inconsistent tail pointers")
cur_elem = cur_elem->next;
} /* end while */
if (actual_tail != span_lst->tail)
- HGOTO_ERROR(H5E_DATASPACE, H5E_INCONSISTENTSTATE, FAIL, "the seletion has inconsistent tail pointers")
+ HGOTO_ERROR(H5E_DATASPACE, H5E_INCONSISTENTSTATE, FAIL,
+ "the selection has inconsistent tail pointers")
done:
FUNC_LEAVE_NOAPI(ret_value)
@@ -231,7 +232,8 @@ H5S__check_points_tail_ptr(const H5S_pnt_list_t *pnt_lst)
cur_elem = cur_elem->next;
} /* end while */
if (actual_tail != pnt_lst->tail)
- HGOTO_ERROR(H5E_DATASPACE, H5E_INCONSISTENTSTATE, FAIL, "the seletion has inconsistent tail pointers")
+ HGOTO_ERROR(H5E_DATASPACE, H5E_INCONSISTENTSTATE, FAIL,
+ "the selection has inconsistent tail pointers")
done:
FUNC_LEAVE_NOAPI(ret_value)
@@ -314,7 +316,7 @@ H5S__check_internal_consistency(const H5S_t *space)
if ((NULL != hslab) && (NULL != hslab->span_lst))
if (H5S__check_spans_tail_ptr(hslab->span_lst) < 0)
HGOTO_ERROR(H5E_DATASPACE, H5E_INCONSISTENTSTATE, FAIL,
- "the seletion has inconsistent tail pointers")
+ "the selection has inconsistent tail pointers")
} /* end if */
else if (space->select.type->type == H5S_SEL_POINTS) {
H5S_pnt_list_t *pnt_lst = space->select.sel_info.pnt_lst;
@@ -322,7 +324,7 @@ H5S__check_internal_consistency(const H5S_t *space)
if (NULL != pnt_lst)
if (H5S__check_points_tail_ptr(pnt_lst) < 0)
HGOTO_ERROR(H5E_DATASPACE, H5E_INCONSISTENTSTATE, FAIL,
- "the seletion has inconsistent tail pointers")
+ "the selection has inconsistent tail pointers")
} /* end else-if */
done:
diff --git a/src/H5T.c b/src/H5T.c
index c23f85d..bbe6ce7 100644
--- a/src/H5T.c
+++ b/src/H5T.c
@@ -4226,7 +4226,7 @@ done:
*
* Return: Success: non-negative
*
- * Failure: nagative
+ * Failure: negative
*
* Programmer: Robb Matzke
* Tuesday, December 22, 1998
diff --git a/src/H5TS.c b/src/H5TS.c
index 258c9df..20aa283 100644
--- a/src/H5TS.c
+++ b/src/H5TS.c
@@ -43,7 +43,7 @@
/* Local Typedefs */
/******************/
-/* Cancelability structure */
+/* Cancellability structure */
typedef struct H5TS_cancel_struct {
int previous_state;
unsigned int cancel_count;
@@ -685,9 +685,9 @@ H5TSmutex_release(unsigned int *lock_count)
* Creates a cancellation counter for a thread if it is the first time
* the thread is entering the library.
*
- * if counter value is zero, then set cancelability type of the thread
+ * if counter value is zero, then set cancellability type of the thread
* to PTHREAD_CANCEL_DISABLE as thread is entering the library and store
- * the previous cancelability type into cancellation counter.
+ * the previous cancellability type into cancellation counter.
* Increase the counter value by 1.
*
* PROGRAMMER: Chee Wai LEE
@@ -758,8 +758,8 @@ done:
* 0 on success and a non-zero error code on error.
*
* DESCRIPTION
- * If counter value is one, then set cancelability type of the thread
- * to the previous cancelability type stored in the cancellation counter.
+ * If counter value is one, then set cancellability type of the thread
+ * to the previous cancellability type stored in the cancellation counter.
* (the thread is leaving the library).
*
* Decrement the counter value by 1.
diff --git a/src/H5Tbit.c b/src/H5Tbit.c
index 04e2a97..20f3855 100644
--- a/src/H5Tbit.c
+++ b/src/H5Tbit.c
@@ -228,7 +228,7 @@ done:
* Purpose: Return a small bit sequence as a number. Bit vector starts
* at OFFSET and is SIZE bits long.
*
- * Return: The bit sequence interpretted as an unsigned integer
+ * Return: The bit sequence interpreted as an unsigned integer
*
*-------------------------------------------------------------------------
*/
@@ -538,7 +538,7 @@ H5T__bit_inc(uint8_t *buf, size_t start, size_t size)
/*-------------------------------------------------------------------------
* Function: H5T__bit_dec
*
- * Purpose: Decrement part of a bit field by substracting 1. The bit
+ * Purpose: Decrement part of a bit field by subtracting 1. The bit
* field starts with bit position START and is SIZE bits long.
*
* Return: The "borrow-in" value. It's TRUE if underflows, FALSE
@@ -564,9 +564,9 @@ H5T__bit_dec(uint8_t *buf, size_t start, size_t size)
if ((size + start - 1) / 8 > idx) {
/* The bit sequence doesn't end in the same byte as starts */
- /* Example: a sequence like 11000100 and start = 3. We substract 00001000 from
+ /* Example: a sequence like 11000100 and start = 3. We subtract 00001000 from
* it and get 10111100. If a sequence is 00000111, we do right shift for START
- * bits and get 00000000. So we need to borrow from higher byte when we substract
+ * bits and get 00000000. So we need to borrow from higher byte when we subtract
* 00001000.
*/
if (!(buf[idx] >> pos))
@@ -597,7 +597,7 @@ H5T__bit_dec(uint8_t *buf, size_t start, size_t size)
else {
/* The bit sequence ends in the same byte as starts */
- /* Example: a sequence like 11000100 and pos=3, size=3. We substract 00001000
+ /* Example: a sequence like 11000100 and pos=3, size=3. We subtract 00001000
* and get 10111100. A bit is borrowed from 6th bit(buf[idx]>>6=00000010, tmp>>6=00000011,
* not equal). We need to put this bit back by increment 1000000.
*/
diff --git a/src/H5Tcommit.c b/src/H5Tcommit.c
index d079e71..e99494a 100644
--- a/src/H5Tcommit.c
+++ b/src/H5Tcommit.c
@@ -1127,7 +1127,7 @@ H5T_open(const H5G_loc_t *loc)
done:
if (ret_value == NULL) {
if (dt) {
- if (shared_fo == NULL) { /* Need to free shared fo */
+ if (shared_fo == NULL) { /* Need to free shared of */
if (dt->shared->owned_vol_obj && H5VL_free_object(dt->shared->owned_vol_obj) < 0)
HDONE_ERROR(H5E_DATATYPE, H5E_CANTCLOSEOBJ, NULL, "unable to close owned VOL object")
dt->shared = H5FL_FREE(H5T_shared_t, dt->shared);
diff --git a/src/H5Tconv.c b/src/H5Tconv.c
index c796c00..94697af 100644
--- a/src/H5Tconv.c
+++ b/src/H5Tconv.c
@@ -4867,7 +4867,7 @@ H5T__conv_s_s(hid_t src_id, hid_t dst_id, H5T_cdata_t *cdata, size_t nelmts, siz
size_t olap; /*num overlapping elements */
size_t nchars = 0; /*number of characters copied */
uint8_t *s, *sp, *d, *dp; /*src and dst traversal pointers*/
- uint8_t *dbuf = NULL; /*temp buf for overlap convers. */
+ uint8_t *dbuf = NULL; /*temp buf for overlap converts. */
herr_t ret_value = SUCCEED; /* Return value */
FUNC_ENTER_PACKAGE
@@ -7145,7 +7145,7 @@ H5T__conv_schar_double(hid_t src_id, hid_t dst_id, H5T_cdata_t *cdata, size_t ne
* Return: Non-negative on success/Negative on failure
*
* Programmer: Raymond Lu
- * Tuesday, Febuary 1, 2005
+ * Tuesday, February 1, 2005
*
*-------------------------------------------------------------------------
*/
@@ -7205,7 +7205,7 @@ H5T__conv_uchar_double(hid_t src_id, hid_t dst_id, H5T_cdata_t *cdata, size_t ne
* Return: Non-negative on success/Negative on failure
*
* Programmer: Raymond Lu
- * Tuesday, Febuary 1, 2005
+ * Tuesday, February 1, 2005
*
*-------------------------------------------------------------------------
*/
@@ -7265,7 +7265,7 @@ H5T__conv_short_double(hid_t src_id, hid_t dst_id, H5T_cdata_t *cdata, size_t ne
* Return: Non-negative on success/Negative on failure
*
* Programmer: Raymond Lu
- * Tuesday, Febuary 1, 2005
+ * Tuesday, February 1, 2005
*
*-------------------------------------------------------------------------
*/
@@ -7325,7 +7325,7 @@ H5T__conv_ushort_double(hid_t src_id, hid_t dst_id, H5T_cdata_t *cdata, size_t n
* Return: Non-negative on success/Negative on failure
*
* Programmer: Raymond Lu
- * Tuesday, Febuary 1, 2005
+ * Tuesday, February 1, 2005
*
*-------------------------------------------------------------------------
*/
@@ -7385,7 +7385,7 @@ H5T__conv_int_double(hid_t src_id, hid_t dst_id, H5T_cdata_t *cdata, size_t nelm
* Return: Non-negative on success/Negative on failure
*
* Programmer: Raymond Lu
- * Tuesday, Febuary 1, 2005
+ * Tuesday, February 1, 2005
*
*-------------------------------------------------------------------------
*/
@@ -7445,7 +7445,7 @@ H5T__conv_uint_double(hid_t src_id, hid_t dst_id, H5T_cdata_t *cdata, size_t nel
* Return: Non-negative on success/Negative on failure
*
* Programmer: Raymond Lu
- * Tuesday, Febuary 1, 2005
+ * Tuesday, February 1, 2005
*
*-------------------------------------------------------------------------
*/
@@ -7505,7 +7505,7 @@ H5T__conv_long_double(hid_t src_id, hid_t dst_id, H5T_cdata_t *cdata, size_t nel
* Return: Non-negative on success/Negative on failure
*
* Programmer: Raymond Lu
- * Tuesday, Febuary 1, 2005
+ * Tuesday, February 1, 2005
*
*-------------------------------------------------------------------------
*/
@@ -7565,7 +7565,7 @@ H5T__conv_ulong_double(hid_t src_id, hid_t dst_id, H5T_cdata_t *cdata, size_t ne
* Return: Non-negative on success/Negative on failure
*
* Programmer: Raymond Lu
- * Tuesday, Febuary 1, 2005
+ * Tuesday, February 1, 2005
*
*-------------------------------------------------------------------------
*/
@@ -7625,7 +7625,7 @@ H5T__conv_llong_double(hid_t src_id, hid_t dst_id, H5T_cdata_t *cdata, size_t ne
* Return: Non-negative on success/Negative on failure
*
* Programmer: Raymond Lu
- * Tuesday, Febuary 1, 2005
+ * Tuesday, February 1, 2005
*
*-------------------------------------------------------------------------
*/
@@ -7687,7 +7687,7 @@ H5T__conv_ullong_double(hid_t src_id, hid_t dst_id, H5T_cdata_t *cdata, size_t n
* Return: Non-negative on success/Negative on failure
*
* Programmer: Raymond Lu
- * Tuesday, Febuary 1, 2005
+ * Tuesday, February 1, 2005
*
*-------------------------------------------------------------------------
*/
@@ -7797,7 +7797,7 @@ H5T__conv_double_uchar(hid_t src_id, hid_t dst_id, H5T_cdata_t *cdata, size_t ne
* Return: Non-negative on success/Negative on failure
*
* Programmer: Raymond Lu
- * Tuesday, Febuary 1, 2005
+ * Tuesday, February 1, 2005
*
*-------------------------------------------------------------------------
*/
@@ -7819,7 +7819,7 @@ H5T__conv_ldouble_schar(hid_t src_id, hid_t dst_id, H5T_cdata_t *cdata, size_t n
* Return: Non-negative on success/Negative on failure
*
* Programmer: Raymond Lu
- * Tuesday, Febuary 1, 2005
+ * Tuesday, February 1, 2005
*
*-------------------------------------------------------------------------
*/
@@ -7929,7 +7929,7 @@ H5T__conv_double_ushort(hid_t src_id, hid_t dst_id, H5T_cdata_t *cdata, size_t n
* Return: Non-negative on success/Negative on failure
*
* Programmer: Raymond Lu
- * Tuesday, Febuary 1, 2005
+ * Tuesday, February 1, 2005
*
*-------------------------------------------------------------------------
*/
@@ -7951,7 +7951,7 @@ H5T__conv_ldouble_short(hid_t src_id, hid_t dst_id, H5T_cdata_t *cdata, size_t n
* Return: Non-negative on success/Negative on failure
*
* Programmer: Raymond Lu
- * Tuesday, Febuary 1, 2005
+ * Tuesday, February 1, 2005
*
*-------------------------------------------------------------------------
*/
@@ -8061,7 +8061,7 @@ H5T__conv_double_uint(hid_t src_id, hid_t dst_id, H5T_cdata_t *cdata, size_t nel
* Return: Non-negative on success/Negative on failure
*
* Programmer: Raymond Lu
- * Tuesday, Febuary 1, 2005
+ * Tuesday, February 1, 2005
*
*-------------------------------------------------------------------------
*/
@@ -8083,7 +8083,7 @@ H5T__conv_ldouble_int(hid_t src_id, hid_t dst_id, H5T_cdata_t *cdata, size_t nel
* Return: Non-negative on success/Negative on failure
*
* Programmer: Raymond Lu
- * Tuesday, Febuary 1, 2005
+ * Tuesday, February 1, 2005
*
*-------------------------------------------------------------------------
*/
@@ -8193,7 +8193,7 @@ H5T__conv_double_ulong(hid_t src_id, hid_t dst_id, H5T_cdata_t *cdata, size_t ne
* Return: Non-negative on success/Negative on failure
*
* Programmer: Raymond Lu
- * Tuesday, Febuary 1, 2005
+ * Tuesday, February 1, 2005
*
*-------------------------------------------------------------------------
*/
@@ -8215,7 +8215,7 @@ H5T__conv_ldouble_long(hid_t src_id, hid_t dst_id, H5T_cdata_t *cdata, size_t ne
* Return: Non-negative on success/Negative on failure
*
* Programmer: Raymond Lu
- * Tuesday, Febuary 1, 2005
+ * Tuesday, February 1, 2005
*
*-------------------------------------------------------------------------
*/
@@ -8325,7 +8325,7 @@ H5T__conv_double_ullong(hid_t src_id, hid_t dst_id, H5T_cdata_t *cdata, size_t n
* Return: Non-negative on success/Negative on failure
*
* Programmer: Raymond Lu
- * Tuesday, Febuary 1, 2005
+ * Tuesday, February 1, 2005
*
*-------------------------------------------------------------------------
*/
@@ -8349,7 +8349,7 @@ H5T__conv_ldouble_llong(hid_t src_id, hid_t dst_id, H5T_cdata_t *cdata, size_t n
* Return: Non-negative on success/Negative on failure
*
* Programmer: Raymond Lu
- * Tuesday, Febuary 1, 2005
+ * Tuesday, February 1, 2005
*
*-------------------------------------------------------------------------
*/
diff --git a/src/H5Tfields.c b/src/H5Tfields.c
index baeae6d..68bc0a3 100644
--- a/src/H5Tfields.c
+++ b/src/H5Tfields.c
@@ -12,7 +12,7 @@
* * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * */
/*
- * Module Info: This module contains commond functionality for fields in
+ * Module Info: This module contains command functionality for fields in
* enumerated & compound datatypes in the H5T interface.
*/
diff --git a/src/H5Tnative.c b/src/H5Tnative.c
index c9a2907..4529e57 100644
--- a/src/H5Tnative.c
+++ b/src/H5Tnative.c
@@ -56,7 +56,7 @@ static herr_t H5T__cmp_offset(size_t *comp_size, size_t *offset, size_t elem_siz
* H5T_NATIVE_LDOUBLE
*
* Compound, array, enum, and VL types all choose among these
- * types for theire members. Time, Bifield, Opaque, Reference
+ * types for their members. Time, Bifield, Opaque, Reference
* types are only copy out.
*
* Return: Success: Returns the native data type if successful.
diff --git a/src/H5Topaque.c b/src/H5Topaque.c
index f097fca..c5c2788 100644
--- a/src/H5Topaque.c
+++ b/src/H5Topaque.c
@@ -70,7 +70,7 @@ done:
/*-------------------------------------------------------------------------
* Function: H5Tget_tag
*
- * Purpose: Get tha tag associated with an opaque datatype.
+ * Purpose: Get the tag associated with an opaque datatype.
*
* Return: A pointer to an allocated string. The caller should free
* the string. NULL is returned for errors.
diff --git a/src/H5Tpkg.h b/src/H5Tpkg.h
index 19593dd..4062cbe 100644
--- a/src/H5Tpkg.h
+++ b/src/H5Tpkg.h
@@ -123,7 +123,7 @@
#endif
/* Define an internal macro for converting unsigned long long to long double. SGI compilers give
- * some incorect conversion. 64-bit Solaris does different rounding. Windows Visual Studio 6 does
+ * some incorrect conversion. 64-bit Solaris does different rounding. Windows Visual Studio 6 does
* not support unsigned long long. For FreeBSD(sleipnir), the last 2 bytes of mantissa are lost when
* compiler tries to do the conversion. For Cygwin, compiler doesn't do rounding correctly.
* Mac OS 10.4 gives some incorrect result. */
diff --git a/src/H5Tpublic.h b/src/H5Tpublic.h
index 911efdf..3df7ca2 100644
--- a/src/H5Tpublic.h
+++ b/src/H5Tpublic.h
@@ -148,7 +148,7 @@ typedef enum H5T_pad_t {
H5T_PAD_ONE = 1, /**< always set to one */
H5T_PAD_BACKGROUND = 2, /**< set to background value */
- H5T_NPAD = 3 /**< sentinal: THIS MUST BE LAST */
+ H5T_NPAD = 3 /**< sentinel: THIS MUST BE LAST */
} H5T_pad_t;
//! <!-- [H5T_pad_t_snip] -->
@@ -1161,7 +1161,7 @@ H5_DLL herr_t H5Tlock(hid_t type_id);
* the link(s) by which the new committed datatype is accessed and
* the creation of any intermediate groups that may be missing.
*
- * Once commited, this datatype may be used to define the datatype
+ * Once committed, this datatype may be used to define the datatype
* of any other dataset or attribute in the file.
*
* This function will not accept a datatype that cannot actually hold
@@ -1171,7 +1171,7 @@ H5_DLL herr_t H5Tlock(hid_t type_id);
* Committed datatypes are sometimes referred to as named datatypes.
*
* \version 1.8.7 Function modified in this release to reject datatypes that
- * will not accomodate actual data, such as a compound datatype
+ * will not accommodate actual data, such as a compound datatype
* with no fields or an enumerated datatype with no members.
*
* \since 1.8.0
@@ -1266,7 +1266,7 @@ H5_DLL hid_t H5Topen_async(const char *app_file, const char *app_func, unsigned
* fields and enumerated datatypes with no members.
*
* \version 1.8.7 Function modified in this release to reject datatypes that
- * will not accomodate actual data, such as a compound datatype
+ * will not accommodate actual data, such as a compound datatype
* with no fields or an enumerated datatype with no members.
*
* \since 1.2.0
diff --git a/src/H5VLcallback.c b/src/H5VLcallback.c
index 2369be5..ace584e 100644
--- a/src/H5VLcallback.c
+++ b/src/H5VLcallback.c
@@ -4967,7 +4967,7 @@ done:
/*-------------------------------------------------------------------------
* Function: H5VL__link_copy
*
- * Purpose: Copys a link from src to dst.
+ * Purpose: Copies a link from src to dst.
*
* Return: Success: Non-negative
* Failure: Negative
@@ -4998,7 +4998,7 @@ done:
/*-------------------------------------------------------------------------
* Function: H5VL_link_copy
*
- * Purpose: Copys a link from src to dst.
+ * Purpose: Copies a link from src to dst.
*
* Return: Success: Non-negative
* Failure: Negative
@@ -6408,7 +6408,7 @@ done:
/*-------------------------------------------------------------------------
* Function: H5VL__request_wait
*
- * Purpose: Waits on an asychronous request through the VOL
+ * Purpose: Waits on an asynchronous request through the VOL
*
* Return: Success: Non-negative
* Failure: Negative
@@ -6442,7 +6442,7 @@ done:
/*-------------------------------------------------------------------------
* Function: H5VL_request_wait
*
- * Purpose: Waits on an asychronous request through the VOL
+ * Purpose: Waits on an asynchronous request through the VOL
*
* Return: Success: Non-negative
* Failure: Negative
diff --git a/src/H5VLint.c b/src/H5VLint.c
index 7035b83..391bd91 100644
--- a/src/H5VLint.c
+++ b/src/H5VLint.c
@@ -1336,7 +1336,7 @@ H5VL__register_connector_by_name(const char *name, hbool_t app_ref, hid_t vipl_i
if (H5I_iterate(H5I_VOL, H5VL__get_connector_cb, &op_data, app_ref) < 0)
HGOTO_ERROR(H5E_VOL, H5E_BADITER, H5I_INVALID_HID, "can't iterate over VOL ids")
- /* If connector alread registered, increment ref count on ID and return ID */
+ /* If connector already registered, increment ref count on ID and return ID */
if (op_data.found_id != H5I_INVALID_HID) {
if (H5I_inc_ref(op_data.found_id, app_ref) < 0)
HGOTO_ERROR(H5E_VOL, H5E_CANTINC, H5I_INVALID_HID,
@@ -1396,7 +1396,7 @@ H5VL__register_connector_by_value(H5VL_class_value_t value, hbool_t app_ref, hid
if (H5I_iterate(H5I_VOL, H5VL__get_connector_cb, &op_data, app_ref) < 0)
HGOTO_ERROR(H5E_VOL, H5E_BADITER, H5I_INVALID_HID, "can't iterate over VOL ids")
- /* If connector alread registered, increment ref count on ID and return ID */
+ /* If connector already registered, increment ref count on ID and return ID */
if (op_data.found_id != H5I_INVALID_HID) {
if (H5I_inc_ref(op_data.found_id, app_ref) < 0)
HGOTO_ERROR(H5E_VOL, H5E_CANTINC, H5I_INVALID_HID,
diff --git a/src/H5VLnative.c b/src/H5VLnative.c
index 18b5b9c..2a86c26 100644
--- a/src/H5VLnative.c
+++ b/src/H5VLnative.c
@@ -224,7 +224,7 @@ H5VL__native_term(void)
* Purpose: Query the connector class.
*
* Note: This routine is in this file so that it can return the address
- * of the staticly declared class struct.
+ * of the statically declared class struct.
*
* Returns: SUCCEED (Can't fail)
*
@@ -251,7 +251,7 @@ H5VL__native_introspect_get_conn_cls(void H5_ATTR_UNUSED *obj, H5VL_get_conn_lvl
* Purpose: Query the capability flags for this connector.
*
* Note: This routine is in this file so that it can return the field
- * from the staticly declared class struct.
+ * from the statically declared class struct.
*
* Returns: SUCCEED (Can't fail)
*
@@ -324,7 +324,7 @@ done:
herr_t
H5VL__native_get_file_addr_len(void *obj, H5I_type_t obj_type, size_t *addr_len)
{
- H5F_t *file = NULL; /* File stuct pointer */
+ H5F_t *file = NULL; /* File struct pointer */
herr_t ret_value = SUCCEED;
FUNC_ENTER_NOAPI(FAIL)
diff --git a/src/H5VLnative_introspect.c b/src/H5VLnative_introspect.c
index fa11bea..0cc6ee4 100644
--- a/src/H5VLnative_introspect.c
+++ b/src/H5VLnative_introspect.c
@@ -53,7 +53,7 @@
/*******************/
/* Note: H5VL__native_introspect_get_conn_cls and H5VL__native_introspect_get_cap_flags
- * are in src/H5VLnative.c so that they can work with the staticly declared
+ * are in src/H5VLnative.c so that they can work with the statically declared
* class struct.
*/
diff --git a/src/H5VLpassthru.c b/src/H5VLpassthru.c
index 681531e..6b4c59f 100644
--- a/src/H5VLpassthru.c
+++ b/src/H5VLpassthru.c
@@ -45,7 +45,7 @@
/* Macros */
/**********/
-/* Whether to display log messge when callback is invoked */
+/* Whether to display log message when callback is invoked */
/* (Uncomment to enable) */
/* #define ENABLE_PASSTHRU_LOGGING */
diff --git a/src/H5VM.c b/src/H5VM.c
index 4d65d75..209ac25 100644
--- a/src/H5VM.c
+++ b/src/H5VM.c
@@ -1244,7 +1244,7 @@ H5VM_chunk_index_scaled(unsigned ndims, const hsize_t *coord, const uint32_t *ch
* Function: H5VM_opvv
*
* Purpose: Perform an operation on a source & destination sequences
- * of offset/length pairs. Each set of sequnces has an array
+ * of offset/length pairs. Each set of sequences has an array
* of lengths, an array of offsets, the maximum number of
* sequences and the current sequence to start at in the sequence.
*
diff --git a/src/H5WB.c b/src/H5WB.c
index df7627f..84592bd 100644
--- a/src/H5WB.c
+++ b/src/H5WB.c
@@ -18,7 +18,7 @@
* Quincey Koziol
*
* Purpose: Implements the "wrapped buffer" code for wrapping
- * an existing [staticly sized] buffer, in order to
+ * an existing [statically sized] buffer, in order to
* avoid lots of memory allocation calls.
*
*-------------------------------------------------------------------------
diff --git a/src/H5Z.c b/src/H5Z.c
index 34af8eb..bcdd837 100644
--- a/src/H5Z.c
+++ b/src/H5Z.c
@@ -1501,7 +1501,7 @@ done:
/*-------------------------------------------------------------------------
* Function: H5Z_filter_in_pline
*
- * Purpose: Check wheter a filter is in the filter pipeline using the
+ * Purpose: Check whether a filter is in the filter pipeline using the
* filter ID. This function is very similar to H5Z_filter_info
*
* Return: TRUE - found filter
diff --git a/src/H5Znbit.c b/src/H5Znbit.c
index b696085..21363bc 100644
--- a/src/H5Znbit.c
+++ b/src/H5Znbit.c
@@ -1015,7 +1015,7 @@ done:
* assume one byte has 8 bit
* assume padding bit is 0
* assume size of unsigned char is one byte
- * assume one data item of certain datatype is stored continously in bytes
+ * assume one data item of certain datatype is stored continuously in bytes
* atomic datatype is treated on byte basis
*/
diff --git a/src/H5Zscaleoffset.c b/src/H5Zscaleoffset.c
index 48a29cd..9942cee 100644
--- a/src/H5Zscaleoffset.c
+++ b/src/H5Zscaleoffset.c
@@ -659,7 +659,7 @@ H5Z_class2_t H5Z_SCALEOFFSET[1] = {{
buf[i] = (type)(buf[i] + (type)(minval)); \
} while (0)
-/* Retrive minimum value of floating-point type */
+/* Retrieve minimum value of floating-point type */
#define H5Z_scaleoffset_get_min(type, minval, min) \
{ \
if (sizeof(type) <= sizeof(long long)) \
@@ -1388,7 +1388,7 @@ done:
* assume one byte has 8 bit
* assume padding bit is 0
* assume size of unsigned char is one byte
- * assume one data item of certain datatype is stored continously in bytes
+ * assume one data item of certain datatype is stored continuously in bytes
* atomic datatype is treated on byte basis
*/
diff --git a/src/H5Ztrans.c b/src/H5Ztrans.c
index ea532cf..092b289 100644
--- a/src/H5Ztrans.c
+++ b/src/H5Ztrans.c
@@ -1352,7 +1352,7 @@ H5Z__op_is_numbs2(H5Z_node *_tree)
* Function: H5Z__xform_reduce_tree
*
* Purpose: Simplifies parse tree passed in by performing any obvious
- * and trivial arithemtic calculations.
+ * and trivial arithmetic calculations.
*
* Return: None.
*
@@ -1692,7 +1692,7 @@ H5Z_xform_noop(const H5Z_data_xform_t *data_xform_prop)
if (data_xform_prop) {
ret_value = FALSE;
- /* Check for trivial data tranformation: expression = "x" */
+ /* Check for trivial data transformation: expression = "x" */
if ((HDstrlen(data_xform_prop->xform_exp) == 1) && data_xform_prop->dat_val_pointers &&
(data_xform_prop->dat_val_pointers->num_ptrs == 1)) {
ret_value = TRUE;
@@ -1723,7 +1723,7 @@ H5Z_xform_extract_xform_str(const H5Z_data_xform_t *data_xform_prop)
/* There should be no way that this can be NULL since the function
* that calls this one checks to make sure it isn't before
- * pasing them */
+ * passing them */
HDassert(data_xform_prop);
FUNC_LEAVE_NOAPI(data_xform_prop->xform_exp)
diff --git a/src/H5detect.c b/src/H5detect.c
index 022cb55..e72d7a0 100644
--- a/src/H5detect.c
+++ b/src/H5detect.c
@@ -772,7 +772,7 @@ done:\n\
fprintf(rawoutstream, "/* sigprocmask() support: no */\n");
#endif
- /* Print the statics of signal handlers called for debugging */
+ /* Print the statistics of signal handlers called for debugging */
fprintf(rawoutstream, "\n"
"/******************************/\n"
"/* signal handlers statistics */\n"
@@ -1475,7 +1475,7 @@ verify_signal_handlers(int signum, void (*handler)(int))
}
else {
if (val == signum) {
- /* return from signum handler. Record a sucess. */
+ /* return from signum handler. Record a success. */
nsuccesses++;
}
else {
diff --git a/src/H5mpi.c b/src/H5mpi.c
index 4a8aa44..aea0104 100644
--- a/src/H5mpi.c
+++ b/src/H5mpi.c
@@ -38,7 +38,7 @@ static hsize_t bigio_count_g = H5_MAX_MPI_COUNT;
/*-------------------------------------------------------------------------
* Function: H5_mpi_set_bigio_count
*
- * Purpose: Allow us to programatically change the switch point
+ * Purpose: Allow us to programmatically change the switch point
* when we utilize derived datatypes. This is of
* particular interest for allowing nightly testing
*
diff --git a/src/H5private.h b/src/H5private.h
index 1c3d358..765e7b6 100644
--- a/src/H5private.h
+++ b/src/H5private.h
@@ -1992,7 +1992,7 @@ extern H5_api_t H5_g;
#define H5_API_LOCK
#define H5_API_UNLOCK
-/* disable cancelability (sequential version) */
+/* disable cancellability (sequential version) */
#define H5_API_UNSET_CANCEL
#define H5_API_SET_CANCEL
diff --git a/src/H5public.h b/src/H5public.h
index 78d2a8f..6a3911c 100644
--- a/src/H5public.h
+++ b/src/H5public.h
@@ -46,7 +46,7 @@
#include <stddef.h>
#include <stdint.h>
-/* Unlike most sys/ headers, which are POSIX-only, sys/types.h is avaible
+/* Unlike most sys/ headers, which are POSIX-only, sys/types.h is available
* on Windows, though it doesn't necessarily contain all the POSIX types
* we need for HDF5 (e.g. ssize_t).
*/
diff --git a/test/SWMR_POSIX_Order_UG.txt b/test/SWMR_POSIX_Order_UG.txt
index 2771af1..a4190bc 100644
--- a/test/SWMR_POSIX_Order_UG.txt
+++ b/test/SWMR_POSIX_Order_UG.txt
@@ -3,7 +3,7 @@ POSIX Write Order Test Instructions
Purpose
-------
-This documents shows the requirments, implementaion design and instructions
+This documents shows the requirements, implementation design and instructions
of building and running the POSIX Write Order test. The name of the
test is twriteorder and it resides in the test/ directory.
@@ -44,7 +44,7 @@ and they are ready for the reader processes to access.
If the system, in which the writer and reader processes run, the readers
will always get all chain-linked blocks correctly. If the order of write
-is not maintained, some reader processes may found unexpect block data.
+is not maintained, some reader processes may found unexpected block data.
Building the Tests
------------------
diff --git a/test/SWMR_UseCase_UG.txt b/test/SWMR_UseCase_UG.txt
index 1e3d1e6..677fdc4 100644
--- a/test/SWMR_UseCase_UG.txt
+++ b/test/SWMR_UseCase_UG.txt
@@ -2,7 +2,7 @@
User Guide for SWMR Use Case Programs
2. Purpose:
- This is a User Guide of the SWMR Use Case programs. It descibes the use
+ This is a User Guide of the SWMR Use Case programs. It describes the use
case program and explain how to run them.
2.1. Author and Dates:
@@ -87,7 +87,7 @@
It opens the HDF5 data file without the SWMR access mode (0 means
off). This likely will result in error. This option is provided for
- users to see the effect of the neede SWMR access mode for concurrent
+ users to see the effect of the needed SWMR access mode for concurrent
access.
3.4. Test Shell Script:
@@ -176,7 +176,7 @@
It opens the HDF5 data file without the SWMR access mode (0 means
off). This likely will result in error. This option is provided for
- users to see the effect of the neede SWMR access mode for concurrent
+ users to see the effect of the needed SWMR access mode for concurrent
access.
4.4. Test Shell Script:
diff --git a/test/accum.c b/test/accum.c
index 08f86ca..080edac 100644
--- a/test/accum.c
+++ b/test/accum.c
@@ -1348,7 +1348,7 @@ test_accum_adjust(H5F_t *f)
/* ==> Accumulator will try to resize, but see that it's getting too big */
/* ==> Size of new block is less than than half maximum size of accumulator */
/* ==> New block being appended to accumulator */
- /* ==> We can slide the dirty region down, to accomodate the request */
+ /* ==> We can slide the dirty region down, to accommodate the request */
/* ==> Max Buffer Size - (dirty offset + adjust size) >= 2 * size) */
/* ==> Need to adjust location of accumulator while appending */
/* ==> Accumulator will need to be reallocated */
@@ -1394,7 +1394,7 @@ test_accum_adjust(H5F_t *f)
/* ==> Accumulator will try to resize, but see that it's getting too big */
/* ==> Size of new block is less than than half maximum size of accumulator */
/* ==> New block being appended to accumulator */
- /* ==> We can slide the dirty region down, to accomodate the request */
+ /* ==> We can slide the dirty region down, to accommodate the request */
/* ==> Max Buffer Size - (dirty offset + adjust size) < 2 * size) */
/* ==> Need to adjust location of accumulator while appending */
if (accum_write((1024 * 1024) - 5, 10, wbuf) < 0)
diff --git a/test/atomic_writer.c b/test/atomic_writer.c
index 5a6310a..33252f1 100644
--- a/test/atomic_writer.c
+++ b/test/atomic_writer.c
@@ -112,7 +112,7 @@ int
main(int argc, char *argv[])
{
int fd = -1; /* file descriptor */
- ssize_t bytes_wrote; /* the nubmer of bytes written */
+ ssize_t bytes_wrote; /* the number of bytes written */
unsigned int *buf = NULL; /* buffer to hold written data */
unsigned int n, u, i; /* local index variable */
int temp; /* temporary variable */
diff --git a/test/big.c b/test/big.c
index 6f8ce67..43fef56 100644
--- a/test/big.c
+++ b/test/big.c
@@ -43,7 +43,7 @@
* that disk space is allocated only when the contents are actually written.
* E.g., If one creates a new file, seeks forward 10 million bytes, writes
* 1 bytes and closes the file, then a sparse file, will show file size of
- * 10 million bytes but actaully uses only couple disk blocks, much smaller
+ * 10 million bytes but actually uses only couple disk blocks, much smaller
* than the formal file size.)
*
* One more consideration is that we want to distinguish an HDF5 library
@@ -70,7 +70,7 @@
#endif
/* Define Small, Large, Extra Large, Huge File which
- * corrspond to less than 2GB, 2GB, 4GB, and tens of GB file size.
+ * correspond to less than 2GB, 2GB, 4GB, and tens of GB file size.
* NO_FILE stands for "no file" to be tested.
*/
typedef enum fsizes_t { SFILE, LFILE, XLFILE, HUGEFILE, NO_FILE } fsizes_t;
@@ -607,9 +607,9 @@ usage(void)
"\t-fsize\tChange family size default to <fsize> where <fsize> is\n"
"\t\ta positive float point number. Default value is %" PRIuHSIZE ".\n"
"Examples:\n"
- "\tbig -fsize 2.1e9 \t# test with file size just under 2GB\n"
- "\tbig -fsize 2.2e9 \t# test with file size just above 2GB\n"
- "\tBe sure the file system can support the file size requested\n",
+ "\t big -fsize 2.1e9 \t# test with file size just under 2GB\n"
+ "\t big -fsize 2.2e9 \t# test with file size just above 2GB\n"
+ "\t Be sure the file system can support the file size requested\n",
(hsize_t)FAMILY_SIZE);
}
diff --git a/test/btree2.c b/test/btree2.c
index 2d34fa7..9cbb211 100644
--- a/test/btree2.c
+++ b/test/btree2.c
@@ -700,7 +700,7 @@ test_insert_basic(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_param_
if (H5B2_insert(bt2, &record) < 0)
FAIL_STACK_ERROR
- /* Attempt to find non-existant record in B-tree with 1 record */
+ /* Attempt to find non-existent record in B-tree with 1 record */
/* (Should not be found, but not fail) */
idx = 41;
found = FALSE;
@@ -717,7 +717,7 @@ test_insert_basic(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_param_
if (found)
TEST_ERROR
- /* Attempt to find existant record in B-tree with 1 record */
+ /* Attempt to find existent record in B-tree with 1 record */
idx = 42;
found = FALSE;
if (H5B2_find(bt2, &idx, &found, find_cb, &idx) < 0)
@@ -732,7 +732,7 @@ test_insert_basic(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_param_
if (!found)
TEST_ERROR
- /* Attempt to index non-existant record in B-tree with 1 record */
+ /* Attempt to index non-existent record in B-tree with 1 record */
idx = 0;
H5E_BEGIN_TRY
{
@@ -780,7 +780,7 @@ test_insert_basic(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_param_
if (H5B2_insert(bt2, &record) < 0)
FAIL_STACK_ERROR
- /* Attempt to find non-existant record in level-0 B-tree with several records */
+ /* Attempt to find non-existent record in level-0 B-tree with several records */
/* (Should not be found, but not fail) */
idx = 41;
found = FALSE;
@@ -789,7 +789,7 @@ test_insert_basic(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_param_
if (found)
TEST_ERROR
- /* Attempt to find existant record in level-0 B-tree with several record */
+ /* Attempt to find existent record in level-0 B-tree with several record */
idx = 56;
found = FALSE;
if (H5B2_find(bt2, &idx, &found, find_cb, &idx) < 0)
@@ -797,7 +797,7 @@ test_insert_basic(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_param_
if (!found)
TEST_ERROR
- /* Attempt to index non-existant record in B-tree with several records */
+ /* Attempt to index non-existent record in B-tree with several records */
idx = 0;
H5E_BEGIN_TRY
{
@@ -958,7 +958,7 @@ test_insert_split_root(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_p
if (idx != (INSERT_SPLIT_ROOT_NREC + 2))
TEST_ERROR
- /* Attempt to find non-existant record in level-1 B-tree */
+ /* Attempt to find non-existent record in level-1 B-tree */
/* (Should not be found, but not fail) */
idx = INSERT_SPLIT_ROOT_NREC + 10;
found = FALSE;
@@ -967,7 +967,7 @@ test_insert_split_root(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_p
if (found)
TEST_ERROR
- /* Attempt to find existant record in root of level-1 B-tree */
+ /* Attempt to find existent record in root of level-1 B-tree */
idx = 33;
found = FALSE;
if (H5B2_find(bt2, &idx, &found, find_cb, &idx) < 0)
@@ -975,7 +975,7 @@ test_insert_split_root(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_p
if (!found)
TEST_ERROR
- /* Attempt to find existant record in leaf of level-1 B-tree */
+ /* Attempt to find existent record in leaf of level-1 B-tree */
idx = 56;
found = FALSE;
if (H5B2_find(bt2, &idx, &found, find_cb, &idx) < 0)
@@ -983,7 +983,7 @@ test_insert_split_root(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_p
if (!found)
TEST_ERROR
- /* Attempt to index non-existant record in level-1 B-tree */
+ /* Attempt to index non-existent record in level-1 B-tree */
idx = 0;
H5E_BEGIN_TRY
{
@@ -1704,7 +1704,7 @@ test_insert_make_level2(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_
if (idx != ((INSERT_SPLIT_ROOT_NREC * 29) + 5))
TEST_ERROR
- /* Attempt to find non-existant record in level-2 B-tree */
+ /* Attempt to find non-existent record in level-2 B-tree */
/* (Should not be found, but not fail) */
idx = INSERT_SPLIT_ROOT_NREC * 30;
found = FALSE;
@@ -1713,7 +1713,7 @@ test_insert_make_level2(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_
if (found)
TEST_ERROR
- /* Attempt to find existant record in root of level-2 B-tree */
+ /* Attempt to find existent record in root of level-2 B-tree */
idx = 948;
found = FALSE;
if (H5B2_find(bt2, &idx, &found, find_cb, &idx) < 0)
@@ -1726,7 +1726,7 @@ test_insert_make_level2(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_
if (check_node_depth(bt2, &record, (unsigned)2) < 0)
TEST_ERROR
- /* Attempt to find existant record in internal node of level-2 B-tree */
+ /* Attempt to find existent record in internal node of level-2 B-tree */
idx = 505;
found = FALSE;
if (H5B2_find(bt2, &idx, &found, find_cb, &idx) < 0)
@@ -1739,7 +1739,7 @@ test_insert_make_level2(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_
if (check_node_depth(bt2, &record, (unsigned)1) < 0)
TEST_ERROR
- /* Attempt to find existant record in leaf of level-2 B-tree */
+ /* Attempt to find existent record in leaf of level-2 B-tree */
idx = 555;
found = FALSE;
if (H5B2_find(bt2, &idx, &found, find_cb, &idx) < 0)
@@ -1752,7 +1752,7 @@ test_insert_make_level2(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_
if (check_node_depth(bt2, &record, (unsigned)0) < 0)
TEST_ERROR
- /* Attempt to index non-existant record in level-2 B-tree */
+ /* Attempt to index non-existent record in level-2 B-tree */
idx = 0;
H5E_BEGIN_TRY
{
@@ -3106,7 +3106,7 @@ HDfprintf(stderr,"curr_time=%lu\n",(unsigned long)curr_time);
if (idx != INSERT_MANY)
TEST_ERROR
- /* Attempt to find non-existant record in level-4 B-tree */
+ /* Attempt to find non-existent record in level-4 B-tree */
/* (Should not be found, but not fail) */
idx = INSERT_MANY * 2;
found = FALSE;
@@ -3120,7 +3120,7 @@ HDfprintf(stderr,"curr_time=%lu\n",(unsigned long)curr_time);
/* Pick random record */
idx = (hsize_t)(HDrandom() % INSERT_MANY);
- /* Attempt to find existant record in root of level-4 B-tree */
+ /* Attempt to find existent record in root of level-4 B-tree */
found = FALSE;
if (H5B2_find(bt2, &idx, &found, find_cb, &idx) < 0)
FAIL_STACK_ERROR
@@ -3128,7 +3128,7 @@ HDfprintf(stderr,"curr_time=%lu\n",(unsigned long)curr_time);
TEST_ERROR
} /* end for */
- /* Attempt to index non-existant record in level-4 B-tree, in increasing & decreasing order */
+ /* Attempt to index non-existent record in level-4 B-tree, in increasing & decreasing order */
H5E_BEGIN_TRY
{
ret = H5B2_index(bt2, H5_ITER_INC, (hsize_t)(INSERT_MANY * 3), find_cb, NULL);
@@ -3151,12 +3151,12 @@ HDfprintf(stderr,"curr_time=%lu\n",(unsigned long)curr_time);
/* Pick random record */
idx = (hsize_t)(HDrandom() % INSERT_MANY);
- /* Attempt to find existant record in root of level-4 B-tree */
+ /* Attempt to find existent record in root of level-4 B-tree */
/* (in increasing order) */
if (H5B2_index(bt2, H5_ITER_INC, idx, find_cb, &idx) < 0)
FAIL_STACK_ERROR
- /* Attempt to find existant record in root of level-4 B-tree */
+ /* Attempt to find existent record in root of level-4 B-tree */
/* (in decreasing order) */
if (H5B2_index(bt2, H5_ITER_DEC, idx, find_dec_cb, &idx) < 0)
FAIL_STACK_ERROR
@@ -3263,7 +3263,7 @@ test_update_basic(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_param_
if (H5B2_update(bt2, &record, no_modify_cb, NULL) < 0)
FAIL_STACK_ERROR
- /* Attempt to find non-existant record in B-tree with 1 record */
+ /* Attempt to find non-existent record in B-tree with 1 record */
/* (Should not be found, but not fail) */
find.key = 10;
find.val = (hsize_t)-1;
@@ -3285,7 +3285,7 @@ test_update_basic(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_param_
if (found)
TEST_ERROR
- /* Attempt to find existant record in B-tree with 1 record */
+ /* Attempt to find existent record in B-tree with 1 record */
find.key = 42;
found = FALSE;
if (H5B2_find(bt2, &find, &found, find_rec_cb, &find) < 0)
@@ -3305,7 +3305,7 @@ test_update_basic(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_param_
if (!found)
TEST_ERROR
- /* Attempt to index non-existant record in B-tree with 1 record */
+ /* Attempt to index non-existent record in B-tree with 1 record */
H5E_BEGIN_TRY
{
ret = H5B2_index(bt2, H5_ITER_INC, (hsize_t)1, index_rec_cb, NULL);
@@ -3340,7 +3340,7 @@ test_update_basic(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_param_
if (H5B2_update(bt2, &modify, modify_rec_cb, &modify) < 0)
FAIL_STACK_ERROR
- /* Attempt to find non-existant record in B-tree with 1 record */
+ /* Attempt to find non-existent record in B-tree with 1 record */
/* (Should not be found, but not fail) */
find.key = 10;
find.val = (hsize_t)-1;
@@ -3382,7 +3382,7 @@ test_update_basic(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_param_
if (!found)
TEST_ERROR
- /* Attempt to index non-existant record in B-tree with 1 record */
+ /* Attempt to index non-existent record in B-tree with 1 record */
H5E_BEGIN_TRY
{
ret = H5B2_index(bt2, H5_ITER_INC, (hsize_t)1, index_rec_cb, NULL);
@@ -3437,7 +3437,7 @@ test_update_basic(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_param_
if (H5B2_update(bt2, &record, no_modify_cb, NULL) < 0)
FAIL_STACK_ERROR
- /* Attempt to find non-existant record in level-0 B-tree with several records */
+ /* Attempt to find non-existent record in level-0 B-tree with several records */
/* (Should not be found, but not fail) */
find.key = 10;
find.val = (hsize_t)-1;
@@ -3449,7 +3449,7 @@ test_update_basic(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_param_
if (found)
TEST_ERROR
- /* Attempt to find existant record in level-0 B-tree with several records */
+ /* Attempt to find existent record in level-0 B-tree with several records */
find.key = 56;
found = FALSE;
if (H5B2_find(bt2, &find, &found, find_rec_cb, &find) < 0)
@@ -3459,7 +3459,7 @@ test_update_basic(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_param_
if (!found)
TEST_ERROR
- /* Attempt to index non-existant record in B-tree with several records */
+ /* Attempt to index non-existent record in B-tree with several records */
H5E_BEGIN_TRY
{
ret = H5B2_index(bt2, H5_ITER_INC, (hsize_t)4, index_rec_cb, NULL);
@@ -3531,7 +3531,7 @@ test_update_basic(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_param_
if (H5B2_update(bt2, &record, modify_rec_cb, &modify) < 0)
FAIL_STACK_ERROR
- /* Attempt to find non-existant record in level-0 B-tree with several records */
+ /* Attempt to find non-existent record in level-0 B-tree with several records */
/* (Should not be found, but not fail) */
find.key = 41;
find.val = (hsize_t)-1;
@@ -3543,7 +3543,7 @@ test_update_basic(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_param_
if (found)
TEST_ERROR
- /* Attempt to find existant record in level-0 B-tree with several record */
+ /* Attempt to find existent record in level-0 B-tree with several record */
find.key = 56;
found = FALSE;
if (H5B2_find(bt2, &find, &found, find_rec_cb, &find) < 0)
@@ -3553,7 +3553,7 @@ test_update_basic(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_param_
if (!found)
TEST_ERROR
- /* Attempt to index non-existant record in B-tree with several records */
+ /* Attempt to index non-existent record in B-tree with several records */
H5E_BEGIN_TRY
{
ret = H5B2_index(bt2, H5_ITER_INC, (hsize_t)4, index_rec_cb, NULL);
@@ -3782,7 +3782,7 @@ test_update_split_root(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_p
if (idx.key != (INSERT_SPLIT_ROOT_NREC_REC + 2))
TEST_ERROR
- /* Attempt to find non-existant record in level-1 B-tree */
+ /* Attempt to find non-existent record in level-1 B-tree */
/* (Should not be found, but not fail) */
find.key = 800;
find.val = (hsize_t)-1;
@@ -3794,7 +3794,7 @@ test_update_split_root(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_p
if (found)
TEST_ERROR
- /* Attempt to find existant record in root of level-1 B-tree */
+ /* Attempt to find existent record in root of level-1 B-tree */
find.key = 33;
find.val = (hsize_t)-1;
found = FALSE;
@@ -3807,7 +3807,7 @@ test_update_split_root(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_p
if (!found)
TEST_ERROR
- /* Attempt to find existant record in leaf of level-1 B-tree */
+ /* Attempt to find existent record in leaf of level-1 B-tree */
find.key = 56;
find.val = (hsize_t)-1;
found = FALSE;
@@ -3820,7 +3820,7 @@ test_update_split_root(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_p
if (!found)
TEST_ERROR
- /* Attempt to index non-existant record in level-1 B-tree */
+ /* Attempt to index non-existent record in level-1 B-tree */
H5E_BEGIN_TRY
{
ret = H5B2_index(bt2, H5_ITER_INC, (hsize_t)(INSERT_SPLIT_ROOT_NREC_REC + 2), index_rec_cb, NULL);
@@ -4578,7 +4578,7 @@ test_update_make_level2(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_
if (idx.key != ((INSERT_SPLIT_ROOT_NREC_REC * 41) + 4))
TEST_ERROR
- /* Attempt to find non-existant record in level-2 B-tree */
+ /* Attempt to find non-existent record in level-2 B-tree */
/* (Should not be found, but not fail) */
find.key = INSERT_SPLIT_ROOT_NREC_REC * 42;
find.val = (hsize_t)-1;
@@ -4590,7 +4590,7 @@ test_update_make_level2(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_
if (found)
TEST_ERROR
- /* Attempt to find existant record in root of level-2 B-tree */
+ /* Attempt to find existent record in root of level-2 B-tree */
find.key = 1347;
find.val = (hsize_t)-1;
found = FALSE;
@@ -4606,7 +4606,7 @@ test_update_make_level2(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_
if (check_node_depth(bt2, &record, (unsigned)2) < 0)
TEST_ERROR
- /* Attempt to find existant record in internal node of level-2 B-tree */
+ /* Attempt to find existent record in internal node of level-2 B-tree */
find.key = 513;
find.val = (hsize_t)-1;
found = FALSE;
@@ -4622,7 +4622,7 @@ test_update_make_level2(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_
if (check_node_depth(bt2, &record, (unsigned)1) < 0)
TEST_ERROR
- /* Attempt to find existant record in leaf of level-2 B-tree */
+ /* Attempt to find existent record in leaf of level-2 B-tree */
find.key = 555;
find.val = (hsize_t)-1;
found = FALSE;
@@ -4638,7 +4638,7 @@ test_update_make_level2(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_
if (check_node_depth(bt2, &record, (unsigned)0) < 0)
TEST_ERROR
- /* Attempt to index non-existant record in level-2 B-tree */
+ /* Attempt to index non-existent record in level-2 B-tree */
H5E_BEGIN_TRY
{
ret = H5B2_index(bt2, H5_ITER_INC, (hsize_t)(INSERT_SPLIT_ROOT_NREC_REC * 42), index_rec_cb, NULL);
@@ -4755,7 +4755,7 @@ test_update_make_level2(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_
if (idx.key != ((INSERT_SPLIT_ROOT_NREC_REC * 41) + 4))
TEST_ERROR
- /* Attempt to find non-existant record in level-2 B-tree */
+ /* Attempt to find non-existent record in level-2 B-tree */
/* (Should not be found, but not fail) */
find.key = INSERT_SPLIT_ROOT_NREC_REC * 42;
find.val = (hsize_t)-1;
@@ -4767,7 +4767,7 @@ test_update_make_level2(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_
if (found)
TEST_ERROR
- /* Attempt to find existant record in root of level-2 B-tree */
+ /* Attempt to find existent record in root of level-2 B-tree */
find.key = 1344;
find.val = (hsize_t)-1;
found = FALSE;
@@ -4783,7 +4783,7 @@ test_update_make_level2(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_
if (check_node_depth(bt2, &record, (unsigned)2) < 0)
TEST_ERROR
- /* Attempt to find existant record in internal node of level-2 B-tree */
+ /* Attempt to find existent record in internal node of level-2 B-tree */
find.key = 512;
find.val = (hsize_t)-1;
found = FALSE;
@@ -4799,7 +4799,7 @@ test_update_make_level2(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_
if (check_node_depth(bt2, &record, (unsigned)1) < 0)
TEST_ERROR
- /* Attempt to find existant record in leaf of level-2 B-tree */
+ /* Attempt to find existent record in leaf of level-2 B-tree */
find.key = 555;
find.val = (hsize_t)-1;
found = FALSE;
@@ -4815,7 +4815,7 @@ test_update_make_level2(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_
if (check_node_depth(bt2, &record, (unsigned)0) < 0)
TEST_ERROR
- /* Attempt to index non-existant record in level-2 B-tree */
+ /* Attempt to index non-existent record in level-2 B-tree */
H5E_BEGIN_TRY
{
ret = H5B2_index(bt2, H5_ITER_INC, (hsize_t)(INSERT_SPLIT_ROOT_NREC_REC * 42), index_rec_cb, NULL);
@@ -4942,7 +4942,7 @@ test_update_make_level2(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_
if (idx.key != ((INSERT_SPLIT_ROOT_NREC_REC * 41) + 6))
TEST_ERROR
- /* Attempt to find non-existant record in level-2 B-tree */
+ /* Attempt to find non-existent record in level-2 B-tree */
/* (Should not be found, but not fail) */
find.key = INSERT_SPLIT_ROOT_NREC_REC * 42;
find.val = (hsize_t)-1;
@@ -4954,7 +4954,7 @@ test_update_make_level2(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_
if (found)
TEST_ERROR
- /* Attempt to find existant record in root of level-2 B-tree */
+ /* Attempt to find existent record in root of level-2 B-tree */
find.key = 1345;
find.val = (hsize_t)-1;
found = FALSE;
@@ -4970,7 +4970,7 @@ test_update_make_level2(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_
if (check_node_depth(bt2, &record, (unsigned)2) < 0)
TEST_ERROR
- /* Attempt to find existant record in internal node of level-2 B-tree */
+ /* Attempt to find existent record in internal node of level-2 B-tree */
find.key = 513;
find.val = (hsize_t)-1;
found = FALSE;
@@ -4986,7 +4986,7 @@ test_update_make_level2(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_
if (check_node_depth(bt2, &record, (unsigned)1) < 0)
TEST_ERROR
- /* Attempt to find existant record in leaf of level-2 B-tree */
+ /* Attempt to find existent record in leaf of level-2 B-tree */
find.key = 555;
find.val = (hsize_t)-1;
found = FALSE;
@@ -5002,7 +5002,7 @@ test_update_make_level2(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_
if (check_node_depth(bt2, &record, (unsigned)0) < 0)
TEST_ERROR
- /* Attempt to index non-existant record in level-2 B-tree */
+ /* Attempt to index non-existent record in level-2 B-tree */
H5E_BEGIN_TRY
{
ret = H5B2_index(bt2, H5_ITER_INC, (hsize_t)(INSERT_SPLIT_ROOT_NREC_REC * 42), index_rec_cb, NULL);
@@ -5188,7 +5188,7 @@ HDfprintf(stderr, "curr_time = %lu\n", (unsigned long)curr_time);
if (iter.key != INSERT_MANY_REC)
TEST_ERROR
- /* Attempt to find non-existant record in level-4 B-tree */
+ /* Attempt to find non-existent record in level-4 B-tree */
/* (Should not be found, but not fail) */
find.key = INSERT_MANY_REC * 2;
find.val = (hsize_t)-1;
@@ -5206,7 +5206,7 @@ HDfprintf(stderr, "curr_time = %lu\n", (unsigned long)curr_time);
find.key = (hsize_t)(HDrandom() % INSERT_MANY_REC);
find.val = (hsize_t)-1;
- /* Attempt to find existant record in level-4 B-tree */
+ /* Attempt to find existent record in level-4 B-tree */
found = FALSE;
if (H5B2_find(bt2, &find, &found, find_rec_cb, &find) < 0)
FAIL_STACK_ERROR
@@ -5216,7 +5216,7 @@ HDfprintf(stderr, "curr_time = %lu\n", (unsigned long)curr_time);
TEST_ERROR
} /* end for */
- /* Attempt to index non-existant record in level-4 B-tree, in increasing & decreasing order */
+ /* Attempt to index non-existent record in level-4 B-tree, in increasing & decreasing order */
H5E_BEGIN_TRY
{
ret = H5B2_index(bt2, H5_ITER_INC, (hsize_t)(INSERT_MANY_REC * 3), find_rec_cb, NULL);
@@ -5245,7 +5245,7 @@ HDfprintf(stderr, "curr_time = %lu\n", (unsigned long)curr_time);
find.key = (hsize_t)-1;
find.val = (hsize_t)-1;
- /* Attempt to find existant record in level-4 B-tree */
+ /* Attempt to find existent record in level-4 B-tree */
/* (in increasing order) */
if (H5B2_index(bt2, H5_ITER_INC, idx, index_rec_cb, &find) < 0)
FAIL_STACK_ERROR
@@ -5258,7 +5258,7 @@ HDfprintf(stderr, "curr_time = %lu\n", (unsigned long)curr_time);
find.key = (hsize_t)-1;
find.val = (hsize_t)-1;
- /* Attempt to find existant record in level-4 B-tree */
+ /* Attempt to find existent record in level-4 B-tree */
/* (in decreasing order) */
if (H5B2_index(bt2, H5_ITER_DEC, idx, index_rec_cb, &find) < 0)
FAIL_STACK_ERROR
@@ -5377,7 +5377,7 @@ test_remove_basic(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_param_
PASSED();
- TESTING("B-tree remove: non-existant record from 1 record B-tree");
+ TESTING("B-tree remove: non-existent record from 1 record B-tree");
/* Check for closing & re-opening the B-tree */
if (reopen_btree(f, &bt2, bt2_addr, tparam) < 0)
@@ -5404,7 +5404,7 @@ test_remove_basic(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_param_
if (!H5F_addr_defined(root_addr))
TEST_ERROR
- /* Attempt to remove a non-existant record from a B-tree with 1 record */
+ /* Attempt to remove a non-existent record from a B-tree with 1 record */
record = 0;
H5E_BEGIN_TRY
{
@@ -5418,7 +5418,7 @@ test_remove_basic(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_param_
PASSED();
/* Attempt to remove a record from a B-tree with 1 record */
- TESTING("B-tree remove: existant record from 1 record B-tree");
+ TESTING("B-tree remove: existent record from 1 record B-tree");
/* Check for closing & re-opening the B-tree */
if (reopen_btree(f, &bt2, bt2_addr, tparam) < 0)
@@ -5482,8 +5482,8 @@ test_remove_basic(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_param_
PASSED();
- /* Attempt to remove a non-existant record from a level-0 B-tree with mult. record */
- TESTING("B-tree remove: non-existant record from level-0 B-tree");
+ /* Attempt to remove a non-existent record from a level-0 B-tree with mult. record */
+ TESTING("B-tree remove: non-existent record from level-0 B-tree");
/* Check for closing & re-opening the B-tree */
if (reopen_btree(f, &bt2, bt2_addr, tparam) < 0)
@@ -5502,7 +5502,7 @@ test_remove_basic(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_param_
PASSED();
/* Attempt to remove a record from a level-0 B-tree with mult. record */
- TESTING("B-tree remove: mult. existant records from level-0 B-tree");
+ TESTING("B-tree remove: mult. existent records from level-0 B-tree");
/* Check for closing & re-opening the B-tree */
if (reopen_btree(f, &bt2, bt2_addr, tparam) < 0)
@@ -5660,7 +5660,7 @@ test_remove_level1_noredistrib(hid_t fapl, const H5B2_create_t *cparam, const bt
herr_t ret; /* Generic error return value */
/* B-tree record removal tests */
- TESTING("B-tree remove: non-existant record from level-1 B-tree");
+ TESTING("B-tree remove: non-existent record from level-1 B-tree");
/* Create the file for the test */
if (create_file(&file, &f, fapl) < 0)
@@ -5693,7 +5693,7 @@ test_remove_level1_noredistrib(hid_t fapl, const H5B2_create_t *cparam, const bt
if (!H5F_addr_defined(root_addr))
TEST_ERROR
- /* Attempt to remove a non-existant record from a B-tree with 1 record */
+ /* Attempt to remove a non-existent record from a B-tree with 1 record */
record = (INSERT_SPLIT_ROOT_NREC * 2) + 1;
H5E_BEGIN_TRY
{
@@ -9748,7 +9748,7 @@ test_modify(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_param_t *tpa
/*
* Test modifying records
*/
- TESTING("B-tree modify: attempt to modify non-existant record");
+ TESTING("B-tree modify: attempt to modify non-existent record");
/* Create the file for the test */
if (create_file(&file, &f, fapl) < 0)
@@ -9771,7 +9771,7 @@ test_modify(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_param_t *tpa
if (bt2_stat.depth != 2)
TEST_ERROR
- /* Attempt to modify a non-existant record */
+ /* Attempt to modify a non-existent record */
record = 3;
modify = 4;
H5E_BEGIN_TRY
diff --git a/test/cache.c b/test/cache.c
index e802a9b..e6a2ac8 100644
--- a/test/cache.c
+++ b/test/cache.c
@@ -2789,8 +2789,8 @@ write_permitted_check(int
* Test the behaviour with different flags.
*
* This test was added primarily to test basic insert
- * pinned entry functionallity, but I through in explicit
- * tests for other functionallity that is tested implicitly
+ * pinned entry functionality, but I through in explicit
+ * tests for other functionality that is tested implicitly
* elsewhere.
*
* Return: void
@@ -2908,7 +2908,7 @@ check_insert_entry(unsigned paged)
}
}
- /* Thats all we can get from H5C_get_entry_status().
+ /* That's all we can get from H5C_get_entry_status().
* Now start looking at the cache data structures directly.
*/
@@ -5146,7 +5146,7 @@ check_flush_cache__flush_ops(H5F_t *file_ptr)
*
* 10/8/07 -- JRM
* Added a resize operation to this test to satisfy the new
- * requiremnt that any resize of an entry on flush will always
+ * requirement that any resize of an entry on flush will always
* be accompanied by a resize. Note that as a result, this
* test becomes redundant with later tests.
*/
@@ -5238,7 +5238,7 @@ check_flush_cache__flush_ops(H5F_t *file_ptr)
*
* 10/8/07 -- JRM
* Added a resize operation to this test to satisfy the new
- * requiremnt that any resize of an entry on flush will always
+ * requirement that any resize of an entry on flush will always
* be accompanied by a resize. Note that as a result, this
* test becomes redundant with later tests.
*/
@@ -9603,7 +9603,7 @@ check_flush_cache__flush_op_eviction_test(H5F_t *file_ptr)
*
* (VET, 9) Y 10 KB N N - dirty (VET, 8)
*
- * Recall that in this test bed, flush operations are excuted the
+ * Recall that in this test bed, flush operations are executed the
* first time the associated entry is flushed, and are then
* deleted.
*/
@@ -10156,7 +10156,7 @@ check_flush_cache__flush_op_eviction_test(H5F_t *file_ptr)
* returns to the serialize function for (VET, 9).
*
* When the serialize function for (VET, 9) is done with (VET, 8), it
- * calls H5C_unprotect(VET, 8), which markes (VET, 8) as dirty and
+ * calls H5C_unprotect(VET, 8), which marks (VET, 8) as dirty and
* unprotected, and places it at the head of the LRU.
*
* (VET, 0) is the next item on the LRU -- it is clean and is therefore
@@ -10555,7 +10555,7 @@ check_flush_cache__flush_op_eviction_test(H5F_t *file_ptr)
* Added tests for pinned entries.
*
* JRM -- 5/17/06
- * Complete reqrite of pinned entry tests to accomodate
+ * Complete reqrite of pinned entry tests to accommodate
* the new H5C_mark_entry_dirty() call.
*
*-------------------------------------------------------------------------
@@ -12980,7 +12980,7 @@ check_expunge_entry(unsigned paged)
* Function: check_multiple_read_protect()
*
* Purpose: Verify that multiple, simultaneous read protects of a
- * single entry perform as expectd.
+ * single entry perform as expected.
*
* Return: void
*
@@ -13326,7 +13326,7 @@ check_multiple_read_protect(unsigned paged)
if (pass && (entry_ptr->header.is_pinned)) {
pass = FALSE;
- failure_mssg = "enty (0,4) still pinned.\n";
+ failure_mssg = "entry (0,4) still pinned.\n";
}
unprotect_entry(file_ptr, 0, 4, /* (0,4) unprotect */
@@ -15529,7 +15529,7 @@ check_destroy_protected_err(unsigned paged)
* Function: check_duplicate_insert_err()
*
* Purpose: Verify that an attempt to insert and entry that is
- * alread in the cache will generate an error.
+ * already in the cache will generate an error.
*
* Return: void
*
@@ -18085,7 +18085,7 @@ check_auto_cache_resize(hbool_t cork_ageout, unsigned paged)
if (show_progress)
HDfprintf(stderr, "check point %d\n", checkpoint++);
- /* eigth epoch -- force the hit rate to 100% again -- should be steady
+ /* eighth epoch -- force the hit rate to 100% again -- should be steady
* state.
*/
if (pass) {
@@ -18181,7 +18181,7 @@ check_auto_cache_resize(hbool_t cork_ageout, unsigned paged)
if (show_progress)
HDfprintf(stderr, "check point %d\n", checkpoint++);
- if (pass) { /* twelth epoch -- at minimum size so no more ageouts */
+ if (pass) { /* twelfth epoch -- at minimum size so no more ageouts */
rpt_fcn_called = FALSE;
i = 0;
while (pass && (i < 1000)) {
@@ -18462,7 +18462,7 @@ check_auto_cache_resize(hbool_t cork_ageout, unsigned paged)
if (show_progress)
HDfprintf(stderr, "check point %d\n", checkpoint++);
- /* eigth epoch -- still 100% hit rate
+ /* eighth epoch -- still 100% hit rate
*/
if (pass) {
@@ -18565,7 +18565,7 @@ check_auto_cache_resize(hbool_t cork_ageout, unsigned paged)
if (show_progress)
HDfprintf(stderr, "check point %d\n", checkpoint++);
- /* twelth epoch -- force the hit rate to 100% again -- should be steady
+ /* twelfth epoch -- force the hit rate to 100% again -- should be steady
* state.
*/
if (pass) {
@@ -18941,7 +18941,7 @@ check_auto_cache_resize(hbool_t cork_ageout, unsigned paged)
if (show_progress)
HDfprintf(stderr, "check point %d\n", checkpoint++);
- /* eigth epoch -- force the hit rate to 100% again -- should be steady
+ /* eighth epoch -- force the hit rate to 100% again -- should be steady
* state.
*/
if (pass) {
@@ -19041,7 +19041,7 @@ check_auto_cache_resize(hbool_t cork_ageout, unsigned paged)
if (show_progress)
HDfprintf(stderr, "check point %d\n", checkpoint++);
- if (pass) { /* twelth epoch -- at minimum size so no more ageouts */
+ if (pass) { /* twelfth epoch -- at minimum size so no more ageouts */
rpt_fcn_called = FALSE;
i = 0;
@@ -19692,7 +19692,7 @@ check_auto_cache_resize(hbool_t cork_ageout, unsigned paged)
if (show_progress)
HDfprintf(stderr, "check point %d\n", checkpoint++);
- /* twelth epoch -- hit rate 1.0 -- decrease as before.
+ /* twelfth epoch -- hit rate 1.0 -- decrease as before.
*/
if (pass) {
@@ -20705,7 +20705,7 @@ check_auto_cache_resize(hbool_t cork_ageout, unsigned paged)
if (show_progress)
HDfprintf(stderr, "check point %d\n", checkpoint++);
- /* protect and unprotect two more large entries -- shouldnt trigger a
+ /* protect and unprotect two more large entries -- shouldn't trigger a
* flash cache size increase.
*/
if (pass) {
@@ -23265,7 +23265,7 @@ check_auto_cache_resize_disable(unsigned paged)
* mode (add space), which is triggered whenever the size of a newly
* loaded or inserted entry, or the delta between old and new entry
* sizes exceeds some fraction of the current maximum cache size, and
- * the cache doesn't have enough free space to accomodate the new/
+ * the cache doesn't have enough free space to accommodate the new/
* resize entry without performing evictions. The range of permissible
* values for the flash_threshold (0.1 to 1.0 as of this writing), and
* for the flash_multiple (0.1 to 10.0) do not permit the facility to
@@ -23274,7 +23274,7 @@ check_auto_cache_resize_disable(unsigned paged)
* current max_cache_size equal to max_size.
*
* We have already tested the latter in check_auto_cache_resize(), so
- * we need only thest the former here. Do this by disabling flash
+ * we need only test the former here. Do this by disabling flash
* cache size increments via the flash_incr_mode, and then creating
* situations that would trigger flash cache size increases were that
* code enabled.
@@ -26347,7 +26347,7 @@ check_auto_cache_resize_input_errs(unsigned paged)
/*-------------------------------------------------------------------------
* Function: check_auto_cache_resize_aux_fcns()
*
- * Purpose: Verify that the auxilary functions associated with
+ * Purpose: Verify that the auxiliary functions associated with
* the automatic cache resize capability are operating
* correctly. These functions are:
*
@@ -26424,13 +26424,13 @@ check_auto_cache_resize_aux_fcns(unsigned paged)
};
if (paged)
- TESTING("automatic cache resize auxilary functions (paged aggregation)")
+ TESTING("automatic cache resize auxiliary functions (paged aggregation)")
else
- TESTING("automatic cache resize auxilary functions")
+ TESTING("automatic cache resize auxiliary functions")
pass = TRUE;
- /* allocate a cache, and then test the various auxilary functions.
+ /* allocate a cache, and then test the various auxiliary functions.
*/
if (pass) {
@@ -27077,7 +27077,7 @@ check_metadata_blizzard_absence(hbool_t fill_via_insertion, unsigned paged)
if (file_ptr == NULL) {
pass = FALSE;
- failure_mssg = "bad return from cache intialization.\n";
+ failure_mssg = "bad return from cache initialization.\n";
}
else
cache_ptr = file_ptr->shared->cache;
@@ -27296,7 +27296,7 @@ check_metadata_blizzard_absence(hbool_t fill_via_insertion, unsigned paged)
H5C__DIRTIED_FLAG); /* unsigned int flags */
}
- /* Expected staus is that after each insertion, the entry
+ /* Expected status is that after each insertion, the entry
* inserted 26 insertions ago has been flushed, and the
* entry currently getting inserted is now in the cache and
* dirty.
@@ -29701,7 +29701,7 @@ check_flush_deps_err(unsigned paged)
result =
H5C_destroy_flush_dependency(&((entries[entry_type])[0]), &((entries[entry_type])[1]));
if (result != FAIL)
- CACHE_ERROR("Destroying [non-existant] dependency when parent isn't protected")
+ CACHE_ERROR("Destroying [non-existent] dependency when parent isn't protected")
break;
/* Verify that parent entry has flush dependency */
@@ -33057,7 +33057,7 @@ check_metadata_cork(hbool_t fill_via_insertion, unsigned paged)
if (file_ptr == NULL) {
pass = FALSE;
- failure_mssg = "bad return from cache intialization.\n";
+ failure_mssg = "bad return from cache initialization.\n";
}
else
cache_ptr = file_ptr->shared->cache;
@@ -33567,7 +33567,7 @@ check_metadata_cork(hbool_t fill_via_insertion, unsigned paged)
* an entry that is no longer in the cache, and which may have
* been deleted.
*
- * This function contans tests for correct handling on this
+ * This function contains tests for correct handling on this
* situation.
*
* Do nothing if pass is FALSE on entry.
@@ -33798,7 +33798,7 @@ cedds__expunge_dirty_entry_in_flush_test(H5F_t *file_ptr)
*
* (HET, 3) Y 16 KB Y N - -
*
- * Recall that in this test bed, flush operations are excuted the
+ * Recall that in this test bed, flush operations are executed the
* first time the associated entry is flushed, and are then
* deleted.
*/
@@ -34048,7 +34048,7 @@ cedds__H5C_make_space_in_cache(H5F_t *file_ptr)
*
* (HET, 3) Y 16 KB N N - -
*
- * Recall that in this test bed, flush operations are excuted the
+ * Recall that in this test bed, flush operations are executed the
* first time the associated entry is flushed, and are then
* deleted.
*/
@@ -34463,7 +34463,7 @@ cedds__H5C__autoadjust__ageout__evict_aged_out_entries(H5F_t *file_ptr)
*
* (MET, 1-31) Y 64 KB N N - -
*
- * Recall that in this test bed, flush operations are excuted the
+ * Recall that in this test bed, flush operations are executed the
* first time the associated entry is flushed, and are then
* deleted.
*/
diff --git a/test/cache_api.c b/test/cache_api.c
index ff74f64..2d642a6 100644
--- a/test/cache_api.c
+++ b/test/cache_api.c
@@ -662,7 +662,7 @@ check_file_mdc_api_calls(unsigned paged, hid_t fcpl_id)
* configured as per the default both by looking at its internal
* configuration, and via the H5Fget_mdc_config() call.
*
- * Then set serveral different configurations, and verify that
+ * Then set several different configurations, and verify that
* they took as per above.
*/
@@ -1054,7 +1054,7 @@ mdc_api_call_smoke_check(int express_test, unsigned paged, hid_t fcpl_id)
}
/* set the dataset creation plist to specify that the raw data is
- * to be partioned into 10X10 element chunks.
+ * to be partitioned into 10X10 element chunks.
*/
if (pass) {
@@ -1885,7 +1885,7 @@ check_file_mdc_api_errs(unsigned paged, hid_t fcpl_id)
pass = TRUE;
- /* Create a file for test purposes, and veify that its metadata cache
+ /* Create a file for test purposes, and verify that its metadata cache
* set to the default MDC configuration.
*/
@@ -2254,7 +2254,7 @@ main(void)
} /* end if */
/* Test with paged aggregation enabled or not */
- /* The "my_fcpl" passed to each test has the paged or non-paged strategy set up accordinly */
+ /* The "my_fcpl" passed to each test has the paged or non-paged strategy set up accordingly */
for (paged = FALSE; paged <= TRUE; paged++) {
hid_t my_fcpl = fcpl_id;
diff --git a/test/cache_common.c b/test/cache_common.c
index 2d88e05..207daa1 100644
--- a/test/cache_common.c
+++ b/test/cache_common.c
@@ -1517,7 +1517,7 @@ add_flush_op(int target_type, int target_idx, int op_code, int type, int idx, hb
* Otherwise, set up a pinned entry dependency so we can
* test the pinned entry modifications to the flush routine.
*
- * Given the types and indicies of the pinned and pinning
+ * Given the types and indices of the pinned and pinning
* entries, add the pinned entry to the list of pinned
* entries in the pinning entry, increment the
* pinning reference count of the pinned entry, and
@@ -1756,7 +1756,7 @@ execute_flush_op(H5F_t *file_ptr, struct test_entry_t *entry_ptr, struct flush_o
case FLUSH_OP__EXPUNGE:
/* the expunge flush op exists to allow us to simulate the
- * case in which an entry is removed from the cashe as the
+ * case in which an entry is removed from the cache as the
* the result of the flush of a second entry. At present,
* this can only happen via the take ownership flag, but
* we will make this test feature more general to as to make
@@ -2301,7 +2301,7 @@ verify_clean(void)
* Function: verify_entry_status
*
* Purpose: Verify that a list of entries have the expected status.
- * If any discrepencies are found, set the failure message
+ * If any discrepancies are found, set the failure message
* and set pass to FALSE.
*
* Do nothing if pass is FALSE on entry.
@@ -3813,7 +3813,7 @@ row_major_scan_forward(H5F_t *file_ptr, int32_t max_index, int32_t lag, hbool_t
HDassert((!cache_ptr->slist_enabled) ||
(cache_ptr->slist_size == cache_ptr->dirty_index_size));
} /* end if */
- } /* end elsef */
+ } /* end elseif */
if (verbose)
HDfprintf(stdout, "\n");
@@ -5221,7 +5221,7 @@ resize_configs_are_equal(const H5C_auto_size_ctl_t *a, const H5C_auto_size_ctl_t
* has both internal and external configuration matching
* *config_ptr.
*
- * Do nothin on success. On failure, set pass to FALSE, and
+ * Do nothing on success. On failure, set pass to FALSE, and
* load an error message into failue_mssg. Note that
* failure_msg is assumed to be at least 128 bytes in length.
*
diff --git a/test/cache_common.h b/test/cache_common.h
index 126a8b3..3280511 100644
--- a/test/cache_common.h
+++ b/test/cache_common.h
@@ -347,7 +347,7 @@ typedef struct test_entry_t {
int pin_type[MAX_PINS]; /* array of the types of entries
* pinned by this entry.
*/
- int pin_idx[MAX_PINS]; /* array of the indicies of
+ int pin_idx[MAX_PINS]; /* array of the indices of
* entries pinned by this entry.
*/
int num_flush_ops; /* integer field containing the
@@ -393,7 +393,7 @@ typedef struct test_entry_t {
unsigned flush_dep_npar; /* Number of flush dependency parents */
unsigned flush_dep_nchd; /* Number of flush dependency children */
unsigned
- flush_dep_ndirty_chd; /* Number of dirty flush dependency children (including granchildren, etc.) */
+ flush_dep_ndirty_chd; /* Number of dirty flush dependency children (including grandchildren, etc.) */
hbool_t pinned_from_client; /* entry was pinned by client call */
hbool_t pinned_from_cache; /* entry was pinned by cache internally */
unsigned flush_order; /* Order that entry was flushed in */
diff --git a/test/cache_image.c b/test/cache_image.c
index 4100c1f..d3961a8 100644
--- a/test/cache_image.c
+++ b/test/cache_image.c
@@ -136,7 +136,7 @@ create_datasets(hid_t file_id, int min_dset, int max_dset)
}
/* set the dataset creation plist to specify that the raw data is
- * to be partioned into 10X10 element chunks.
+ * to be partitioned into 10X10 element chunks.
*/
if (pass) {
@@ -470,7 +470,7 @@ delete_datasets(hid_t file_id, int min_dset, int max_dset)
* Set pass to FALSE and issue a suitable failure
* message if either the file contains a metadata cache image
* superblock extension and mdci_sbem_expected is TRUE, or
- * vise versa.
+ * vice versa.
*
* If mdci_sbem_expected is TRUE, also verify that the metadata
* cache has been advised of this.
@@ -544,7 +544,7 @@ open_hdf5_file(hbool_t create_file, hbool_t mdci_sbem_expected, hbool_t read_onl
if (show_progress)
HDfprintf(stdout, "%s: cp = %d.\n", fcn_name, cp++);
- /* create a file access propertly list. */
+ /* create a file access property list. */
if (pass) {
fapl_id = h5_fileaccess();
@@ -879,7 +879,7 @@ attempt_swmr_open_hdf5_file(const hbool_t create_file, const hbool_t set_mdci_fa
H5AC_cache_image_config_t cache_image_config = {H5AC__CURR_CACHE_IMAGE_CONFIG_VERSION, TRUE, FALSE,
H5AC__CACHE_IMAGE__ENTRY_AGEOUT__NONE};
- /* create a file access propertly list. */
+ /* create a file access property list. */
if (pass) {
fapl_id = h5_fileaccess();
@@ -1505,7 +1505,7 @@ check_cache_image_ctl_flow_1(hbool_t single_file_vfd)
* processed as part of the first protect operation after the
* superblock is loaded.)
*
- * In this particular test, we preform the following operations:
+ * In this particular test, we perform the following operations:
*
* 1) Create a HDF5 file with the cache image FAPL entry.
*
@@ -7001,7 +7001,7 @@ cache_image_api_error_check_4(hbool_t single_file_vfd)
* image.
*
* The objective of this test is to create a test file
- * with both non-empty self referential presistant
+ * with both non-empty self referential persistent
* free space managers, and a cache image, and then
* verify that this situation is handled correctly if
* H5Fget_free_sections() is called before the metadata
diff --git a/test/cache_tagging.c b/test/cache_tagging.c
index b69a501..a335193 100644
--- a/test/cache_tagging.c
+++ b/test/cache_tagging.c
@@ -443,7 +443,7 @@ check_file_creation_tags(hid_t fcpl_id, int type)
/* Variable Declarations */
hid_t fid = -1; /* File Identifier */
#ifndef NDEBUG
- int verbose = FALSE; /* verbose test outout */
+ int verbose = FALSE; /* verbose test output */
#endif /* NDEBUG */
hid_t fapl = -1; /* File access prop list */
haddr_t root_tag = 0;
@@ -547,7 +547,7 @@ check_file_open_tags(hid_t fcpl, int type)
/* Variable Declarations */
hid_t fid = -1; /* File Identifier */
#ifndef NDEBUG
- int verbose = FALSE; /* verbose file outout */
+ int verbose = FALSE; /* verbose file output */
#endif /* NDEBUG */
hid_t fapl = -1; /* File access prop list */
haddr_t root_tag; /* Root Group Tag */
@@ -677,7 +677,7 @@ check_group_creation_tags(void)
hid_t fid = -1; /* File Identifier */
hid_t gid = -1; /* Group Identifier */
#ifndef NDEBUG
- int verbose = FALSE; /* verbose file outout */
+ int verbose = FALSE; /* verbose file output */
#endif /* NDEBUG */
hid_t fapl = -1; /* File access prop list */
haddr_t root_tag = HADDR_UNDEF; /* Root Group Tag */
@@ -800,7 +800,7 @@ check_multi_group_creation_tags(void)
hid_t fid = -1; /* File Identifier */
hid_t gid = -1; /* Group Identifier */
#ifndef NDEBUG
- int verbose = FALSE; /* verbose file outout */
+ int verbose = FALSE; /* verbose file output */
#endif /* NDEBUG */
char gname[16]; /* group name buffer */
int i = 0; /* iterator */
@@ -952,7 +952,7 @@ check_link_iteration_tags(void)
hid_t sid = -1; /* Group Identifier */
hid_t did = -1; /* Group Identifier */
#ifndef NDEBUG
- int verbose = FALSE; /* verbose file outout */
+ int verbose = FALSE; /* verbose file output */
#endif /* NDEBUG */
int i = 0; /* iterator */
haddr_t root_tag = 0; /* Root Group Tag Value */
@@ -1094,7 +1094,7 @@ check_dense_attribute_tags(void)
hid_t did = -1; /* Group Identifier */
hid_t dcpl = -1; /* Group Identifier */
#ifndef NDEBUG
- int verbose = FALSE; /* verbose file outout */
+ int verbose = FALSE; /* verbose file output */
#endif /* NDEBUG */
int i = 0; /* iterator */
hid_t fapl = -1; /* File access property list */
@@ -1455,7 +1455,7 @@ check_attribute_creation_tags(hid_t fcpl, int type)
hid_t gid = -1; /* Group Identifier */
hid_t sid = -1; /* Dataspace Identifier */
#ifndef NDEBUG
- int verbose = FALSE; /* verbose file outout */
+ int verbose = FALSE; /* verbose file output */
#endif /* NDEBUG */
hid_t fapl = -1; /* File access prop list */
haddr_t root_tag = 0; /* Root group tag */
@@ -1620,7 +1620,7 @@ check_attribute_open_tags(hid_t fcpl, int type)
hid_t gid = -1; /* Group Identifier */
hid_t sid = -1; /* Dataspace Identifier */
#ifndef NDEBUG
- int verbose = FALSE; /* verbose file outout */
+ int verbose = FALSE; /* verbose file output */
#endif /* NDEBUG */
hid_t fapl = -1; /* File access prop list */
haddr_t root_tag = 0;
@@ -1787,7 +1787,7 @@ check_attribute_rename_tags(hid_t fcpl, int type)
hid_t aid = -1; /* Attribute Identifier */
hid_t sid = -1; /* Dataset Identifier */
#ifndef NDEBUG
- int verbose = FALSE; /* verbose file outout */
+ int verbose = FALSE; /* verbose file output */
#endif /* NDEBUG */
int * data = NULL; /* data buffer */
int i, j, k = 0; /* iterators */
@@ -2000,7 +2000,7 @@ check_attribute_delete_tags(hid_t fcpl, int type)
hid_t aid = -1; /* Attribute Identifier */
hid_t sid = -1; /* Dataset Identifier */
#ifndef NDEBUG
- int verbose = FALSE; /* verbose file outout */
+ int verbose = FALSE; /* verbose file output */
#endif /* NDEBUG */
int * data = NULL; /* data buffer */
int i, j, k = 0; /* iterators */
@@ -2191,7 +2191,7 @@ check_dataset_creation_tags(hid_t fcpl, int type)
hid_t did = -1; /* Dataset Identifier */
hid_t sid = -1; /* Dataspace Identifier */
#ifndef NDEBUG
- int verbose = FALSE; /* verbose file outout */
+ int verbose = FALSE; /* verbose file output */
#endif /* NDEBUG */
hid_t dcpl = -1; /* dataset creation pl */
hsize_t cdims[2] = {1, 1}; /* chunk dimensions */
@@ -2351,7 +2351,7 @@ check_dataset_creation_earlyalloc_tags(hid_t fcpl, int type)
hid_t did = -1; /* Dataset Identifier */
hid_t sid = -1; /* Dataspace Identifier */
#ifndef NDEBUG
- int verbose = FALSE; /* verbose file outout */
+ int verbose = FALSE; /* verbose file output */
#endif /* NDEBUG */
hid_t dcpl = -1; /* dataset creation pl */
hsize_t cdims[2] = {1, 1}; /* chunk dimensions */
@@ -2517,7 +2517,7 @@ check_dataset_open_tags(void)
hid_t did = -1; /* Dataset Identifier */
hid_t sid = -1; /* Dataspace Identifier */
#ifndef NDEBUG
- int verbose = FALSE; /* verbose file outout */
+ int verbose = FALSE; /* verbose file output */
#endif /* NDEBUG */
hid_t dcpl = -1; /* dataset creation pl */
hsize_t cdims[2] = {1, 1}; /* chunk dimensions */
@@ -2669,7 +2669,7 @@ check_dataset_write_tags(void)
hid_t did = -1; /* Dataset Identifier */
hid_t sid = -1; /* Dataspace Identifier */
#ifndef NDEBUG
- int verbose = FALSE; /* verbose file outout */
+ int verbose = FALSE; /* verbose file output */
#endif /* NDEBUG */
hid_t dcpl = -1; /* dataset creation pl */
hsize_t cdims[2] = {1, 1}; /* chunk dimensions */
@@ -2836,7 +2836,7 @@ check_attribute_write_tags(hid_t fcpl, int type)
hid_t aid = -1; /* Attribute Identifier */
hid_t sid = -1; /* Dataset Identifier */
#ifndef NDEBUG
- int verbose = FALSE; /* verbose file outout */
+ int verbose = FALSE; /* verbose file output */
#endif /* NDEBUG */
int * data = NULL; /* data buffer */
int i, j, k = 0; /* iterators */
@@ -3022,7 +3022,7 @@ check_dataset_read_tags(void)
hid_t did = -1; /* Dataset Identifier */
hid_t sid = -1; /* Dataspace Identifier */
#ifndef NDEBUG
- int verbose = FALSE; /* verbose file outout */
+ int verbose = FALSE; /* verbose file output */
#endif /* NDEBUG */
hid_t dcpl = -1; /* dataset creation pl */
hsize_t cdims[2] = {1, 1}; /* chunk dimensions */
@@ -3184,7 +3184,7 @@ check_dataset_size_retrieval(void)
hid_t did = -1; /* Dataset Identifier */
hid_t sid = -1; /* Dataspace Identifier */
#ifndef NDEBUG
- int verbose = FALSE; /* verbose file outout */
+ int verbose = FALSE; /* verbose file output */
#endif /* NDEBUG */
hid_t dcpl = -1; /* dataset creation pl */
hsize_t cdims[2] = {1, 1}; /* chunk dimensions */
@@ -3348,7 +3348,7 @@ check_dataset_extend_tags(void)
hid_t did = -1; /* Dataset Identifier */
hid_t sid = -1; /* Dataspace Identifier */
#ifndef NDEBUG
- int verbose = FALSE; /* verbose file outout */
+ int verbose = FALSE; /* verbose file output */
#endif /* NDEBUG */
hid_t dcpl = -1; /* dataset creation pl */
hsize_t cdims[2] = {1, 1}; /* chunk dimensions */
@@ -3795,7 +3795,7 @@ check_link_removal_tags(hid_t fcpl, int type)
hid_t sid = -1; /* Dataspace Identifier */
hid_t gid = -1; /* Dataspace Identifier */
#ifndef NDEBUG
- int verbose = FALSE; /* verbose file outout */
+ int verbose = FALSE; /* verbose file output */
#endif /* NDEBUG */
hid_t dcpl = -1; /* dataset creation pl */
hsize_t cdims[2] = {1, 1}; /* chunk dimensions */
@@ -3984,7 +3984,7 @@ check_link_getname_tags(void)
hid_t sid = -1; /* Dataspace Identifier */
hid_t gid = -1; /* Dataspace Identifier */
#ifndef NDEBUG
- int verbose = FALSE; /* verbose file outout */
+ int verbose = FALSE; /* verbose file output */
#endif /* NDEBUG */
hid_t dcpl = -1; /* dataset creation pl */
hsize_t cdims[2] = {1, 1}; /* chunk dimensions */
@@ -4161,7 +4161,7 @@ check_external_link_creation_tags(void)
hid_t fid2 = -1; /* File Identifier */
hid_t gid = -1; /* Dataspace Identifier */
#ifndef NDEBUG
- int verbose = FALSE; /* verbose file outout */
+ int verbose = FALSE; /* verbose file output */
#endif /* NDEBUG */
hid_t fapl = -1; /* File access prop list */
haddr_t root_tag = 0;
@@ -4288,7 +4288,7 @@ check_external_link_open_tags(void)
hid_t gid = -1; /* Dataspace Identifier */
hid_t xid = -1; /* Dataspace Identifier */
#ifndef NDEBUG
- int verbose = FALSE; /* verbose file outout */
+ int verbose = FALSE; /* verbose file output */
#endif /* NDEBUG */
H5O_native_info_t ninfo; /* Native object info struct */
hid_t fapl = -1; /* File access prop list */
@@ -4585,7 +4585,7 @@ main(void)
/* Create a standard file creation property list */
fcpl_default = H5Pcreate(H5P_FILE_CREATE);
- /* Create an fcpl with shared messages and file space managment enabled */
+ /* Create an fcpl with shared messages and file space management enabled */
fcpl_shmesg_all = H5Pcreate(H5P_FILE_CREATE);
H5Pset_shared_mesg_nindexes(fcpl_shmesg_all, 1);
H5Pset_shared_mesg_index(fcpl_shmesg_all, 0, H5O_SHMESG_ALL_FLAG, 20);
diff --git a/test/chunk_info.c b/test/chunk_info.c
index a1c4160..148c577 100644
--- a/test/chunk_info.c
+++ b/test/chunk_info.c
@@ -291,7 +291,7 @@ index_type_str(H5D_chunk_index_t idx_type)
/*-------------------------------------------------------------------------
* Function: verify_selected_chunks (helper function)
*
- * Purpose: Reads the chunks within the boundery {start,end} and verify
+ * Purpose: Reads the chunks within the boundary {start,end} and verify
* the values against the populated data.
*
* Return: Success: SUCCEED
@@ -321,7 +321,7 @@ verify_selected_chunks(hid_t dset, hid_t plist, const hsize_t *start, const hsiz
for (jj = 0; jj < CHUNK_NY; jj++)
expected_buf[n][ii][jj] = (int)(ii * jj) + 1;
- /* Read each chunk within the boundery of {start,end} and verify the
+ /* Read each chunk within the boundary of {start,end} and verify the
values against the expected data */
chk_index = 0;
for (ii = start[0]; ii < end[0]; ii++)
diff --git a/test/cmpd_dset.c b/test/cmpd_dset.c
index 8777096..04a931c 100644
--- a/test/cmpd_dset.c
+++ b/test/cmpd_dset.c
@@ -713,7 +713,7 @@ test_compound(char *filename, hid_t fapl)
/*
*######################################################################
* Step 12: Write an array into the middle third of the dataset
- * initializeing only members `b' and `d' to -1.
+ * initializing only members `b' and `d' to -1.
*/
TESTING("hyperslab part initialized write");
diff --git a/test/cork.c b/test/cork.c
index 2031ac1..b1f28c1 100644
--- a/test/cork.c
+++ b/test/cork.c
@@ -101,7 +101,7 @@ verify_old_dset_cork(void)
hsize_t dims[2] = {100, 20}; /* Dataset dimension sizes */
hsize_t max_dims[2] = {100, H5S_UNLIMITED}; /* Dataset maximum dimension sizes */
hsize_t chunk_dims[2] = {2, 5}; /* Dataset chunked dimension sizes */
- int ** buf = NULL; /* Data bufer (pointers to fake 2D array) */
+ int ** buf = NULL; /* Data buffer (pointers to fake 2D array) */
int * buf_data = NULL; /* Data buffer (actual data) */
int i = 0, j = 0; /* Local index variables */
H5O_info2_t oinfo, oinfo2, oinfo3; /* Object metadata information */
@@ -513,7 +513,7 @@ verify_dset_cork(hbool_t swmr, hbool_t new_format)
hsize_t dims[2] = {100, 20}; /* Dataset dimension sizes */
hsize_t max_dims[2] = {100, H5S_UNLIMITED}; /* Dataset maximum dimension sizes */
hsize_t chunk_dims[2] = {2, 5}; /* Dataset chunked dimension sizes */
- int ** buf = NULL; /* Data bufer (pointers to fake 2D array) */
+ int ** buf = NULL; /* Data buffer (pointers to fake 2D array) */
int * buf_data = NULL; /* Data buffer (actual data) */
int i = 0, j = 0; /* Local index variables */
H5O_info2_t oinfo, oinfo2, oinfo3; /* Object metadata information */
@@ -1887,7 +1887,7 @@ test_dset_cork(hbool_t swmr, hbool_t new_format)
{
hid_t fid = H5I_INVALID_HID; /* File ID */
hid_t fapl; /* File access property list */
- hid_t gid = H5I_INVALID_HID; /* Groupd ID */
+ hid_t gid = H5I_INVALID_HID; /* Grouped ID */
hid_t did1 = H5I_INVALID_HID, did2 = H5I_INVALID_HID; /* Dataset IDs */
hid_t tid1 = H5I_INVALID_HID, tid2 = H5I_INVALID_HID; /* Datatype IDs */
hid_t sid = H5I_INVALID_HID; /* Dataspace ID */
diff --git a/test/direct_chunk.c b/test/direct_chunk.c
index 8960784..ac4ef8d 100644
--- a/test/direct_chunk.c
+++ b/test/direct_chunk.c
@@ -686,7 +686,7 @@ filter_bogus1(unsigned int flags, size_t H5_ATTR_UNUSED cd_nelmts,
size_t buf_left = *buf_size; /* Amount of data buffer left to process */
if (flags & H5Z_FLAG_REVERSE) { /* read */
- /* Substract the "add on" value to all the data values */
+ /* Subtract the "add on" value to all the data values */
while (buf_left > 0) {
*int_ptr++ -= (int)ADD_ON;
buf_left -= sizeof(int);
@@ -722,7 +722,7 @@ filter_bogus2(unsigned int flags, size_t H5_ATTR_UNUSED cd_nelmts,
size_t buf_left = *buf_size; /* Amount of data buffer left to process */
if (flags & H5Z_FLAG_REVERSE) { /* read */
- /* Substract the "add on" value to all the data values */
+ /* Subtract the "add on" value to all the data values */
while (buf_left > 0) {
*int_ptr++ /= (int)FACTOR;
buf_left -= sizeof(int);
@@ -1427,7 +1427,7 @@ error:
* Function: test_direct_chunk_read_no_cache
*
* Purpose: Test the basic functionality of H5Dread_chunk with the
- * chunk cache diabled.
+ * chunk cache disabled.
*
* Return: Success: 0
* Failure: 1
@@ -1963,7 +1963,7 @@ error:
* Function: test_read_unallocated_chunk
*
* Purpose: Tests the H5Dread_chunk and H5Dget_chunk_storage_size with valid
- * offets to chunks that have not been written to the dataset and are
+ * offsets to chunks that have not been written to the dataset and are
* not allocated in the chunk storage on disk.
*
* Return: Success: 0
@@ -2014,7 +2014,7 @@ test_read_unallocated_chunk(hid_t file)
if ((dxpl = H5Pcreate(H5P_DATASET_XFER)) < 0)
FAIL_STACK_ERROR;
- /* Write a single chunk to intialize the chunk storage */
+ /* Write a single chunk to initialize the chunk storage */
HDmemset(direct_buf, 0, CHUNK_NX * CHUNK_NY * sizeof(int));
offset[0] = 0;
offset[1] = 0;
@@ -2031,7 +2031,7 @@ test_read_unallocated_chunk(hid_t file)
offset[0] = i * CHUNK_NX;
offset[1] = j * CHUNK_NY;
- /* Read a non-existant chunk using the direct read function. */
+ /* Read a non-existent chunk using the direct read function. */
H5E_BEGIN_TRY
{
status = H5Dread_chunk(dataset, dxpl, offset, &filter_mask, &direct_buf);
@@ -2042,7 +2042,7 @@ test_read_unallocated_chunk(hid_t file)
if (status != -1)
TEST_ERROR
- /* Query the size of the non-existant chunk */
+ /* Query the size of the non-existent chunk */
direct_chunk_nbytes = ULONG_MAX;
H5E_BEGIN_TRY
{
diff --git a/test/dsets.c b/test/dsets.c
index a4b5757..690d13f 100644
--- a/test/dsets.c
+++ b/test/dsets.c
@@ -2007,7 +2007,7 @@ test_filter_internal(hid_t fid, const char *name, hid_t dcpl, int if_fletcher32,
if (status >= 0)
TEST_ERROR;
- /* Callback decides to continue inspite data is corrupted. */
+ /* Callback decides to continue in spite data is corrupted. */
if (H5Pset_filter_callback(dxpl, filter_cb_cont, NULL) < 0)
TEST_ERROR;
if (H5Dread(dataset, H5T_NATIVE_INT, H5S_ALL, H5S_ALL, dxpl, check_data) < 0)
@@ -2074,7 +2074,7 @@ test_filter_internal(hid_t fid, const char *name, hid_t dcpl, int if_fletcher32,
if (status >= 0)
TEST_ERROR;
- /* Callback decides to continue inspite data is corrupted. */
+ /* Callback decides to continue in spite data is corrupted. */
if (H5Pset_filter_callback(dxpl, filter_cb_cont, NULL) < 0)
TEST_ERROR;
if (H5Dread(dataset, H5T_NATIVE_INT, H5S_ALL, H5S_ALL, dxpl, check_data) < 0)
@@ -2138,7 +2138,7 @@ test_filter_internal(hid_t fid, const char *name, hid_t dcpl, int if_fletcher32,
if (status >= 0)
TEST_ERROR;
- /* Callback decides to continue inspite data is corrupted. */
+ /* Callback decides to continue in spite data is corrupted. */
if (H5Pset_filter_callback(dxpl, filter_cb_cont, NULL) < 0)
TEST_ERROR;
if (H5Dread(dataset, H5T_NATIVE_INT, H5S_ALL, H5S_ALL, dxpl, check_data) < 0)
@@ -2205,7 +2205,7 @@ test_filter_internal(hid_t fid, const char *name, hid_t dcpl, int if_fletcher32,
if (status >= 0)
TEST_ERROR;
- /* Callback decides to continue inspite data is corrupted. */
+ /* Callback decides to continue in spite data is corrupted. */
if (H5Pset_filter_callback(dxpl, filter_cb_cont, NULL) < 0)
TEST_ERROR;
if (H5Dread(dataset, H5T_NATIVE_INT, H5S_ALL, H5S_ALL, dxpl, check_data) < 0)
@@ -3240,7 +3240,7 @@ test_nbit_int(hid_t file)
power = HDpow(2.0, (double)(precision - 1));
orig_data[i][j] = (int)(((long long)HDrandom() % (long long)power) << offset);
- /* even-numbered values are negtive */
+ /* even-numbered values are negative */
if ((i * size[1] + j + 1) % 2 == 0)
orig_data[i][j] = -orig_data[i][j];
}
@@ -3818,7 +3818,7 @@ test_nbit_compound(hid_t file)
orig_data[i][j].s = (short)(((long long)HDrandom() % (long long)power) << offset[2]);
orig_data[i][j].f = float_val[i][j];
- /* some even-numbered integer values are negtive */
+ /* some even-numbered integer values are negative */
if ((i * size[1] + j + 1) % 2 == 0) {
orig_data[i][j].i = -orig_data[i][j].i;
orig_data[i][j].s = (short)-orig_data[i][j].s;
@@ -4532,7 +4532,7 @@ test_nbit_int_size(hid_t file)
goto error;
}
- /* Initiliaze data buffer with random data within correct range
+ /* Initialize data buffer with random data within correct range
* corresponding to the memory datatype's precision and offset.
*/
for (i = 0; i < DSET_DIM1; i++)
@@ -4745,7 +4745,7 @@ test_nbit_flt_size(hid_t file)
} /* end if */
/*
- * Initiliaze data buffer with random data
+ * Initialize data buffer with random data
*/
for (i = 0; i < DSET_DIM1; i++)
for (j = 0; j < DSET_DIM2; j++)
@@ -4904,7 +4904,7 @@ test_scaleoffset_int(hid_t file)
for (j = 0; j < (size_t)size[1]; j++) {
orig_data[i][j] = HDrandom() % 10000;
- /* even-numbered values are negtive */
+ /* even-numbered values are negative */
if ((i * size[1] + j + 1) % 2 == 0)
orig_data[i][j] = -orig_data[i][j];
}
@@ -5048,7 +5048,7 @@ test_scaleoffset_int_2(hid_t file)
for (j = 0; j < (size_t)size[1]; j++) {
orig_data[0][j] = (int)HDrandom() % 10000;
- /* even-numbered values are negtive */
+ /* even-numbered values are negative */
if ((j + 1) % 2 == 0)
orig_data[0][j] = -orig_data[0][j];
}
@@ -5170,7 +5170,7 @@ test_scaleoffset_float(hid_t file)
for (j = 0; j < (size_t)size[1]; j++) {
orig_data[i][j] = (float)(HDrandom() % 100000) / 1000.0F;
- /* even-numbered values are negtive */
+ /* even-numbered values are negative */
if ((i * size[1] + j + 1) % 2 == 0)
orig_data[i][j] = -orig_data[i][j];
}
@@ -5316,7 +5316,7 @@ test_scaleoffset_float_2(hid_t file)
for (j = 0; j < (size_t)size[1]; j++) {
orig_data[0][j] = (float)(HDrandom() % 100000) / 1000.0F;
- /* even-numbered values are negtive */
+ /* even-numbered values are negative */
if ((j + 1) % 2 == 0)
orig_data[0][j] = -orig_data[0][j];
}
@@ -5437,7 +5437,7 @@ test_scaleoffset_double(hid_t file)
for (j = 0; j < (size_t)size[1]; j++) {
orig_data[i][j] = (HDrandom() % 10000000) / 10000000.0;
- /* even-numbered values are negtive */
+ /* even-numbered values are negative */
if ((i * size[1] + j + 1) % 2 == 0)
orig_data[i][j] = -orig_data[i][j];
}
@@ -5583,7 +5583,7 @@ test_scaleoffset_double_2(hid_t file)
for (j = 0; j < (size_t)size[1]; j++) {
orig_data[0][j] = (HDrandom() % 10000000) / 10000000.0;
- /* even-numbered values are negtive */
+ /* even-numbered values are negative */
if ((j + 1) % 2 == 0)
orig_data[0][j] = -orig_data[0][j];
}
@@ -8895,7 +8895,7 @@ test_chunk_cache(hid_t fapl)
if (H5Pclose(dapl2) < 0)
FAIL_STACK_ERROR
- /* Similary, test use of H5Dcreate2 with H5P_DEFAULT */
+ /* Similarly, test use of H5Dcreate2 with H5P_DEFAULT */
if (H5Dclose(dsid) < 0)
FAIL_STACK_ERROR
if ((dsid = H5Dcreate2(fid, "dset2", H5T_NATIVE_INT, sid, H5P_DEFAULT, dcpl, H5P_DEFAULT)) < 0)
@@ -11980,7 +11980,7 @@ error:
* Purpose: Checks that a file created with either:
* (a) SWMR-write + non-latest-format
* (b) write + latest format
- * will generate datset with latest chunk indexing type.
+ * will generate dataset with latest chunk indexing type.
*
* Return: Success: 0
* Failure: -1
@@ -14714,7 +14714,7 @@ test_versionbounds(void)
hid_t dcpl = -1; /* Dataset creation property list */
hid_t srcspace = -1; /* Source dataspaces */
hid_t vspace = -1; /* Virtual dset dataspaces */
- hid_t srcdset = -1; /* Source datset */
+ hid_t srcdset = -1; /* Source dataset */
hid_t vdset = -1; /* Virtual dataset */
hsize_t dims[1] = {3}; /* Data space current size */
char srcfilename[FILENAME_BUF_SIZE];
@@ -14850,7 +14850,7 @@ error:
/*-----------------------------------------------------------------------------
* Function: test_object_header_minimization_dcpl
*
- * Purpose: Test the "datset object header minimization" property as part of
+ * Purpose: Test the "dataset object header minimization" property as part of
* the DCPL.
*
* Return: Success/pass: 0
@@ -15557,7 +15557,7 @@ main(void)
double rdcc_w0;
int nerrors = 0;
const char *envval;
- hbool_t contig_addr_vfd; /* Whether VFD used has a contigous address space */
+ hbool_t contig_addr_vfd; /* Whether VFD used has a contiguous address space */
hbool_t driver_uses_modified_filename = h5_driver_uses_modified_filename();
int i;
@@ -15566,7 +15566,7 @@ main(void)
if (envval == NULL)
envval = "nomatch";
- /* Current VFD that does not support contigous address space */
+ /* Current VFD that does not support contiguous address space */
contig_addr_vfd = (hbool_t)(HDstrcmp(envval, "split") != 0 && HDstrcmp(envval, "multi") != 0);
/* Set the random # seed */
diff --git a/test/dt_arith.c b/test/dt_arith.c
index 7e96d29..3180123 100644
--- a/test/dt_arith.c
+++ b/test/dt_arith.c
@@ -27,7 +27,7 @@
#define FP_EPSILON 0.000001F
/*
- * Offset from alinged memory returned by malloc(). This can be used to test
+ * Offset from aligned memory returned by malloc(). This can be used to test
* that type conversions handle non-aligned buffers correctly.
*/
#define ALIGNMENT 1
@@ -223,7 +223,7 @@ static int without_hardware_g = 0;
/*The number of values if multiplied by 10 for each step.*/ \
num_norm = (SRC_MAX_10_EXP - SRC_MIN_10_EXP); \
/*Reduce the number of values by 2^factor. MULTIPLY=10^(2^factor). Using this algorithm \
- *instead of arithmatic operation to avoid any conversion*/ \
+ *instead of arithmetic operation to avoid any conversion*/ \
num_norm >>= factor; \
\
/*Total number of values*/ \
@@ -640,7 +640,7 @@ test_particular_fp_integer(void)
float src_f = (float)INT_MAX;
int dst_i;
int fill_value = 13;
- int endian; /*endianess */
+ int endian; /*endianness */
unsigned int fails_this_test = 0;
size_t j;
@@ -818,7 +818,7 @@ test_derived_flt(void)
size_t src_size, dst_size;
unsigned char *buf = NULL, *saved_buf = NULL;
int * aligned = NULL;
- int endian; /*endianess */
+ int endian; /*endianness */
size_t nelmts = NTESTELEM;
unsigned int fails_this_test = 0;
const size_t max_fails = 40; /*max number of failures*/
@@ -1267,7 +1267,7 @@ test_derived_integer(void)
char filename[1024];
size_t src_size, dst_size;
unsigned char *buf = NULL, *saved_buf = NULL;
- int endian; /*endianess */
+ int endian; /*endianness */
size_t nelmts = NTESTELEM;
unsigned int fails_this_test = 0;
const size_t max_fails = 40; /*max number of failures*/
@@ -1591,7 +1591,7 @@ test_conv_int_1(const char *name, hid_t src, hid_t dst)
dtype_t src_type, dst_type; /*data types */
const char * src_type_name = NULL; /*source type name */
const char * dst_type_name = NULL; /*destination type name */
- int endian; /*machine endianess */
+ int endian; /*machine endianness */
size_t src_size, dst_size; /*type sizes */
unsigned char * buf = NULL; /*buffer for conversion */
unsigned char * saved = NULL; /*original values */
@@ -2831,8 +2831,8 @@ test_conv_flt_1(const char *name, int run_test, hid_t src, hid_t dst)
int overflow = 0; /*overflow occurred */
int uflow = 0; /*underflow debug counters*/
size_t j, k; /*counters */
- int sendian; /* source type endianess */
- int dendian; /* Destination type endianess */
+ int sendian; /* source type endianness */
+ int dendian; /* Destination type endianness */
size_t dst_ebias; /* Destination type's exponent bias */
size_t src_epos; /* Source type's exponent position */
size_t src_esize; /* Source type's exponent size */
@@ -3147,7 +3147,7 @@ test_conv_flt_1(const char *name, int run_test, hid_t src, hid_t dst)
/*
* Assume same if both results are NaN. There are many NaN bit
- * patterns and the software doesn't attemt to emulate the
+ * patterns and the software doesn't attempt to emulate the
* hardware in this regard. Instead, software uses a single bit
* pattern for NaN by setting the significand to all ones.
*/
@@ -3426,8 +3426,8 @@ test_conv_int_fp(const char *name, int run_test, hid_t src, hid_t dst)
dtype_t dst_type; /*data types */
const char * src_type_name = NULL; /*source type name */
const char * dst_type_name = NULL; /*destination type name */
- int sendian; /*source endianess */
- int dendian; /*destination endianess */
+ int sendian; /*source endianness */
+ int dendian; /*destination endianness */
size_t src_size, dst_size; /*type sizes */
unsigned char * buf = NULL; /*buffer for conversion */
unsigned char * saved = NULL; /*original values */
@@ -4485,7 +4485,7 @@ test_conv_int_fp(const char *name, int run_test, hid_t src, hid_t dst)
break;
}
- HDprintf(" ans = ");
+ HDprintf(" and = ");
for (k = 0; k < dst_size; k++)
HDprintf(" %02x", hw[ENDIAN(dst_size, k, dendian)]);
HDprintf("%*s", (int)(3 * MAX(0, (ssize_t)src_size - (ssize_t)dst_size)), "");
diff --git a/test/dtransform.c b/test/dtransform.c
index 5d1c04f..9f7e659 100644
--- a/test/dtransform.c
+++ b/test/dtransform.c
@@ -465,7 +465,7 @@ init_test(hid_t file_id)
if ((dataspace = H5Screate_simple(2, dim, NULL)) < 0)
TEST_ERROR
- TESTING("Intializing test...")
+ TESTING("Initializing test...")
if ((dset_id_int = H5Dcreate2(file_id, "/default_int", H5T_NATIVE_INT, dataspace, H5P_DEFAULT,
H5P_DEFAULT, H5P_DEFAULT)) < 0)
diff --git a/test/dtypes.c b/test/dtypes.c
index e9a2461..8ae913f 100644
--- a/test/dtypes.c
+++ b/test/dtypes.c
@@ -29,7 +29,7 @@
#define ARRAY_DIM 4
/*
- * Offset from alinged memory returned by malloc(). This can be used to test
+ * Offset from aligned memory returned by malloc(). This can be used to test
* that type conversions handle non-aligned buffers correctly.
*/
#define ALIGNMENT 1
@@ -2637,7 +2637,7 @@ test_compound_13(void)
data_out.x[u] = (unsigned char)u;
data_out.y = 99.99f;
- /* Set latest_format in access propertly list to enable the latest
+ /* Set latest_format in access property list to enable the latest
* compound datatype format.
*/
if ((fapl_id = H5Pcreate(H5P_FILE_ACCESS)) < 0)
@@ -3794,7 +3794,7 @@ error:
*
* Modifications:
* Raymond Lu
- * Wednesday, Febuary 9, 2005
+ * Wednesday, February 9, 2005
* Added test for H5Tenum_valueof, H5Tenum_nameof, and
* H5Tget_member_value.
*-------------------------------------------------------------------------
@@ -6847,7 +6847,7 @@ test_set_order(void)
{
hid_t dtype; /* Datatype ID */
H5T_order_t order; /* Byte order */
- hsize_t dims[2] = {3, 4}; /* Array dimenstions */
+ hsize_t dims[2] = {3, 4}; /* Array dimensions */
herr_t ret; /* Generic return value */
TESTING("H5Tset/get_order");
@@ -7095,7 +7095,7 @@ test_set_order_compound(hid_t fapl)
hid_t file = -1;
hid_t cmpd = -1, memb_cmpd = -1, memb_array1 = -1, memb_array2 = -1, cmpd_array = -1;
hid_t vl_id = -1;
- hsize_t dims[2] = {3, 4}; /* Array dimenstions */
+ hsize_t dims[2] = {3, 4}; /* Array dimensions */
char filename[1024];
herr_t ret; /* Generic return value */
@@ -7950,7 +7950,7 @@ test_deprec(hid_t fapl)
unsigned u; /* Local index variable */
herr_t status; /* Generic routine value */
- TESTING("deprected API routines for datatypes");
+ TESTING("deprecated API routines for datatypes");
/* Create an array datatype with an atomic base type */
/* (dimension permutations allowed, but not stored) */
diff --git a/test/earray.c b/test/earray.c
index 337e818..4763d51 100644
--- a/test/earray.c
+++ b/test/earray.c
@@ -1284,7 +1284,7 @@ typedef struct eiter_fw_t {
/*-------------------------------------------------------------------------
* Function: eiter_fw_init
*
- * Purpose: Initialize element interator (forward iteration)
+ * Purpose: Initialize element iterator (forward iteration)
*
* Return: Success: Pointer to iteration status object
* Failure: NULL
@@ -1428,7 +1428,7 @@ eiter_fw_state(void *in_eiter, const H5EA_create_t *cparam, const earray_test_pa
/*-------------------------------------------------------------------------
* Function: eiter_fw_term
*
- * Purpose: Shut down element interator (forward iteration)
+ * Purpose: Shut down element iterator (forward iteration)
*
* Return: Success: 0
* Failure: -1
@@ -1472,7 +1472,7 @@ typedef struct eiter_rv_t {
/*-------------------------------------------------------------------------
* Function: eiter_rv_init
*
- * Purpose: Initialize element interator (reverse iteration)
+ * Purpose: Initialize element iterator (reverse iteration)
*
* Return: Success: Pointer to iteration status object
* Failure: NULL
@@ -1653,7 +1653,7 @@ eiter_rv_state(void *in_eiter, const H5EA_create_t *cparam, const earray_test_pa
/*-------------------------------------------------------------------------
* Function: eiter_rv_term
*
- * Purpose: Shut down element interator (reverse iteration)
+ * Purpose: Shut down element iterator (reverse iteration)
*
* Return: Success: 0
* Failure: -1
@@ -1694,7 +1694,7 @@ typedef struct eiter_rnd_t {
/*-------------------------------------------------------------------------
* Function: eiter_rnd_init
*
- * Purpose: Initialize element interator (random iteration)
+ * Purpose: Initialize element iterator (random iteration)
*
* Return: Success: Pointer to iteration status object
* Failure: NULL
@@ -1803,7 +1803,7 @@ eiter_rnd_max(const void *in_eiter)
/*-------------------------------------------------------------------------
* Function: eiter_rnd_term
*
- * Purpose: Shut down element interator (random iteration)
+ * Purpose: Shut down element iterator (random iteration)
*
* Return: Success: 0
* Failure: -1
@@ -1843,7 +1843,7 @@ static const earray_iter_t ea_iter_rnd = {
/*-------------------------------------------------------------------------
* Function: eiter_rnd2_init
*
- * Purpose: Initialize element interator (random #2 iteration)
+ * Purpose: Initialize element iterator (random #2 iteration)
*
* Return: Success: Pointer to iteration status object
* Failure: NULL
@@ -1926,7 +1926,7 @@ typedef struct eiter_cyc_t {
/*-------------------------------------------------------------------------
* Function: eiter_cyc_init
*
- * Purpose: Initialize element interator (cyclic iteration)
+ * Purpose: Initialize element iterator (cyclic iteration)
*
* Return: Success: Pointer to iteration status object
* Failure: NULL
@@ -2019,7 +2019,7 @@ eiter_cyc_max(const void *in_eiter)
/*-------------------------------------------------------------------------
* Function: eiter_cyc_term
*
- * Purpose: Shut down element interator (cyclic iteration)
+ * Purpose: Shut down element iterator (cyclic iteration)
*
* Return: Success: 0
* Failure: -1
diff --git a/test/enc_dec_plist_cross_platform.c b/test/enc_dec_plist_cross_platform.c
index 9ea5cf9..3030d7c 100644
--- a/test/enc_dec_plist_cross_platform.c
+++ b/test/enc_dec_plist_cross_platform.c
@@ -28,7 +28,7 @@ int
main(void)
{
if (VERBOSE_MED)
- HDprintf("Encode/Decode property list endianess\n");
+ HDprintf("Encode/Decode property list endianness\n");
/******* ENCODE/DECODE DCPLS *****/
TESTING("Default DCPL Encoding/Decoding");
diff --git a/test/enum.c b/test/enum.c
index 3923892..512fd9d 100644
--- a/test/enum.c
+++ b/test/enum.c
@@ -552,7 +552,7 @@ test_value_dsnt_exist(void)
if ((datatype_id = H5Tenum_create(H5T_NATIVE_INT)) < 0)
goto error;
- /* These calls should fail, since no memebrs exist yet */
+ /* These calls should fail, since no members exist yet */
if (H5Tenum_valueof(datatype_id, "SAX", &val) >= 0)
goto error;
val = 3;
diff --git a/test/external.c b/test/external.c
index c31701d..281593c 100644
--- a/test/external.c
+++ b/test/external.c
@@ -1217,7 +1217,7 @@ test_path_relative_cwd(hid_t fapl)
}
H5E_END_TRY;
if (dset3 >= 0)
- FAIL_PUTS_ERROR("reopening the dataset with a different efile_prefix succeded");
+ FAIL_PUTS_ERROR("reopening the dataset with a different efile_prefix succeeded");
/* Read the entire dataset and compare with the original */
HDmemset(whole, 0, sizeof(whole));
@@ -1233,7 +1233,7 @@ test_path_relative_cwd(hid_t fapl)
if (H5Dclose(dset) < 0)
FAIL_STACK_ERROR
- /* Open dataset (use a differend prefix than for create.
+ /* Open dataset (use a different prefix than for create.
* This works because the dataset was closed.
*/
if (H5Pset_efile_prefix(dapl2, "${ORIGIN}/.") < 0)
@@ -1254,7 +1254,7 @@ test_path_relative_cwd(hid_t fapl)
}
H5E_END_TRY;
if (dset3 >= 0)
- FAIL_PUTS_ERROR("reopening the dataset with a different efile_prefix succeded");
+ FAIL_PUTS_ERROR("reopening the dataset with a different efile_prefix succeeded");
/* Read the entire dataset and compare with the original */
HDmemset(whole, 0, sizeof(whole));
diff --git a/test/farray.c b/test/farray.c
index 950c7a7..d31d77a 100644
--- a/test/farray.c
+++ b/test/farray.c
@@ -969,7 +969,7 @@ typedef struct fiter_fw_t {
/*-------------------------------------------------------------------------
* Function: fiter_fw_init
*
- * Purpose: Initialize element interator (forward iteration)
+ * Purpose: Initialize element iterator (forward iteration)
*
* Return: Success: Pointer to iteration status object
* Failure: NULL
@@ -1021,7 +1021,7 @@ fiter_fw_next(void *_fiter)
/*-------------------------------------------------------------------------
* Function: fiter_term
*
- * Purpose: Shut down element interator (simple iterators)
+ * Purpose: Shut down element iterator (simple iterators)
*
* Return: Success: 0
* Failure: -1
@@ -1055,7 +1055,7 @@ typedef struct fiter_rv_t {
/*-------------------------------------------------------------------------
* Function: fiter_rv_init
*
- * Purpose: Initialize element interator (reverse iteration)
+ * Purpose: Initialize element iterator (reverse iteration)
*
* Return: Success: Pointer to iteration status object
* Failure: NULL
@@ -1120,7 +1120,7 @@ typedef struct fiter_rnd_t {
/*-------------------------------------------------------------------------
* Function: fiter_rnd_init
*
- * Purpose: Initialize element interator (random iteration)
+ * Purpose: Initialize element iterator (random iteration)
*
* Return: Success: Pointer to iteration status object
* Failure: NULL
@@ -1193,7 +1193,7 @@ fiter_rnd_next(void *_fiter)
/*-------------------------------------------------------------------------
* Function: fiter_rnd_term
*
- * Purpose: Shut down element interator (random iteration)
+ * Purpose: Shut down element iterator (random iteration)
*
* Return: Success: 0
* Failure: -1
@@ -1235,7 +1235,7 @@ typedef struct fiter_cyc_t {
/*-------------------------------------------------------------------------
* Function: fiter_cyc_init
*
- * Purpose: Initialize element interator (cyclic iteration)
+ * Purpose: Initialize element iterator (cyclic iteration)
*
* Return: Success: Pointer to iteration status object
* Failure: NULL
diff --git a/test/fheap.c b/test/fheap.c
index 3c3088d..5ded20f 100644
--- a/test/fheap.c
+++ b/test/fheap.c
@@ -2469,7 +2469,7 @@ test_id_limits(hid_t fapl, H5HF_create_t *cparam, hid_t fcpl)
size_t id_len; /* Size of fractal heap IDs */
size_t tiny_max_len; /* Max. length of tiny objects */
hbool_t tiny_len_extended; /* Do tiny objects use two bytes for the length? */
- hbool_t huge_ids_direct; /* Are 'huge' objects directly acccessed? */
+ hbool_t huge_ids_direct; /* Are 'huge' objects directly accessed? */
/* Set the filename to use for this test (dependent on fapl) */
h5_fixname(FILENAME[0], fapl, filename, sizeof(filename));
@@ -13785,7 +13785,7 @@ test_filtered_huge(hid_t fapl, H5HF_create_t *cparam, fheap_test_param_t *tparam
fheap_heap_state_t state; /* State of fractal heap */
unsigned deflate_level; /* Deflation level */
size_t old_actual_id_len = 0; /* Old actual ID length */
- hbool_t huge_ids_direct; /* Are 'huge' objects directly acccessed? */
+ hbool_t huge_ids_direct; /* Are 'huge' objects directly accessed? */
hbool_t pline_init = FALSE; /* Whether the I/O pipeline has been initialized */
/* Test description */
const char *base_desc = "insert 'huge' object into heap with I/O filters, then remove %s";
@@ -16361,7 +16361,7 @@ main(void)
unsigned num_pb_fs = 1; /* The number of settings to test for page buffering and file space handling */
int ExpressMode; /* Express testing level */
const char *envval; /* Environment variable */
- hbool_t contig_addr_vfd; /* Whether VFD used has a contigous address space */
+ hbool_t contig_addr_vfd; /* Whether VFD used has a contiguous address space */
hbool_t api_ctx_pushed = FALSE; /* Whether API context pushed */
/* Don't run this test using certain file drivers */
@@ -16369,7 +16369,7 @@ main(void)
if (envval == NULL)
envval = "nomatch";
- /* Current VFD that does not support contigous address space */
+ /* Current VFD that does not support contiguous address space */
contig_addr_vfd = (hbool_t)(HDstrcmp(envval, "split") != 0 && HDstrcmp(envval, "multi") != 0);
/* Reset library */
diff --git a/test/file_image.c b/test/file_image.c
index e2b7810..7f6a83e 100644
--- a/test/file_image.c
+++ b/test/file_image.c
@@ -145,7 +145,7 @@ test_properties(void)
/* Check that sizes are the same, and that the buffers are identical but separate */
VERIFY(temp_size == size, "Sizes of buffers don't match");
- VERIFY(temp2 != NULL, "Recieved buffer not set");
+ VERIFY(temp2 != NULL, "Received buffer not set");
VERIFY(temp2 != buffer, "Retrieved buffer is the same as original");
VERIFY(temp2 != temp, "Retrieved buffer is the same as previously retrieved buffer");
VERIFY(0 == HDmemcmp(temp2, buffer, size), "Buffers contain different data");
@@ -260,7 +260,7 @@ free_cb(void *ptr, H5FD_file_image_op_t op, void *udata)
* Function: udata_copy_cb
*
* Purpose: This function allows calls to the udata_copy callback to be tracked.
- * No copying actualy takes place; it is easier to deal with only one
+ * No copying actually takes place; it is easier to deal with only one
* instance of the udata.
*
* Returns: A pointer to the same udata that was passed in.
@@ -954,7 +954,7 @@ test_get_file_image(const char *test_banner, const int file_name_num, hid_t fapl
err = H5Fclose(core_file_id);
VERIFY(err == SUCCEED, "H5Fclose(core_file_id) failed.");
- /* dicard core fapl */
+ /* discard core fapl */
err = H5Pclose(core_fapl_id);
VERIFY(err == SUCCEED, "H5Pclose(core_fapl_id) failed.");
@@ -1078,7 +1078,7 @@ test_get_file_image_error_rejection(void)
VERIFY(bytes_read < 0, "H5Fget_file_image(2 -- test 1) succeeded.");
/* Call H5Fget_file_image() with good buffer and buffer size,
- * but non-existant file_id. Should fail.
+ * but non-existent file_id. Should fail.
*/
H5E_BEGIN_TRY
{
diff --git a/test/fillval.c b/test/fillval.c
index 7faeb65..1a0c02f 100644
--- a/test/fillval.c
+++ b/test/fillval.c
@@ -1918,7 +1918,7 @@ test_extend_cases(hid_t file, hid_t _dcpl, const char *dset_name, const hsize_t
if ((fspace = H5Dget_space(dset)) < 0)
TEST_ERROR
- /* Create dataspace for single element sized bufer */
+ /* Create dataspace for single element sized buffer */
if ((mspace = H5Screate_simple(5, one, NULL)) < 0)
TEST_ERROR
@@ -2142,7 +2142,7 @@ test_extend(hid_t fapl, const char *base_name, H5D_layout_t layout)
#if 1
/*
- * Remove this when contiguous datasets can be exended to some
+ * Remove this when contiguous datasets can be extended to some
* predetermined fininte size, even if it's just in the slowest varying
* dimension. If it's removed prematurely then you'll get one of the
* errors described above or `unable to select fill value region'.
diff --git a/test/filter_fail.c b/test/filter_fail.c
index a1bcf28..64cc41a 100644
--- a/test/filter_fail.c
+++ b/test/filter_fail.c
@@ -160,7 +160,7 @@ test_filter_write(char *file_name, hid_t my_fapl, hbool_t cache_enabled)
points[i] = i;
/* Write data. If the chunk cache is enabled, H5Dwrite should succeed. If it is
- * diabled, H5Dwrite should fail. */
+ * disabled, H5Dwrite should fail. */
if (cache_enabled) {
if (H5Dwrite(dataset, H5T_NATIVE_INT, H5S_ALL, sid, H5P_DEFAULT, points) < 0)
TEST_ERROR
@@ -186,7 +186,7 @@ test_filter_write(char *file_name, hid_t my_fapl, hbool_t cache_enabled)
TEST_ERROR
/* Close dataset. If the chunk cache is enabled, the flushing of chunks should fail
- * during H5Dclose. If it is diabled, H5Dwrite should fail but H5Dclose should succeed. */
+ * during H5Dclose. If it is disabled, H5Dwrite should fail but H5Dclose should succeed. */
if (cache_enabled) {
H5E_BEGIN_TRY
{
diff --git a/test/filter_plugin1_dsets.c b/test/filter_plugin1_dsets.c
index 5d6c1ef..29b6e73 100644
--- a/test/filter_plugin1_dsets.c
+++ b/test/filter_plugin1_dsets.c
@@ -79,7 +79,7 @@ add_sub_value(unsigned int flags, size_t cd_nelmts, const unsigned int *cd_value
value = (int)cd_values[0];
if (flags & H5Z_FLAG_REVERSE) {
- /* READ - Substract the given value from all the data values */
+ /* READ - Subtract the given value from all the data values */
while (buf_left > 0) {
*int_ptr++ -= value;
buf_left -= sizeof(int);
diff --git a/test/filter_plugin3_dsets.c b/test/filter_plugin3_dsets.c
index 618ce06..1d399e5 100644
--- a/test/filter_plugin3_dsets.c
+++ b/test/filter_plugin3_dsets.c
@@ -108,7 +108,7 @@ add_sub_value_hdf5(unsigned int flags, size_t cd_nelmts, const unsigned int *cd_
value = (int)cd_values[0];
if (flags & H5Z_FLAG_REVERSE) {
- /* READ - Substract the given value from all the data values */
+ /* READ - Subtract the given value from all the data values */
while (buf_left > 0) {
*int_ptr++ -= value;
buf_left -= sizeof(int);
diff --git a/test/flushrefresh.c b/test/flushrefresh.c
index 8d2c083..2757bbf 100644
--- a/test/flushrefresh.c
+++ b/test/flushrefresh.c
@@ -63,10 +63,10 @@
#define NOT_FLUSHED "NOT_FLUSHED"
/* Error Handling */
-/* For errors occuring in the main process, use the standard TEST_ERROR macro.
+/* For errors occurring in the main process, use the standard TEST_ERROR macro.
For errors occurring in the spawned process (from the test script), use
the PROCESS_ERROR macro, which will send a signal to the main process so the
- main process can propogate errors correctly. */
+ main process can propagate errors correctly. */
FILE *errorfile;
#define ERRFILE "flushrefresh_ERROR"
#define PROCESS_ERROR \
@@ -222,7 +222,7 @@ test_flush(void)
* object, a series of verifications will occur on each object in the file.
*
* Each verification consists of spawning off a new process and determining
- * if the object can be opened and its information retreived in said
+ * if the object can be opened and its information retrieved in said
* alternate process. It reports the results, which are compared to an
* expected value (either that the object can be found on disk, or that it
* cannot).
@@ -695,7 +695,7 @@ test_refresh(void)
* that an attribute has been added, and is only visible after a
* successful call to a H5*refresh function).
*
- * As with the flush case, the implemention is a bit tricky as it's
+ * As with the flush case, the implementation is a bit tricky as it's
* dealing with signals going back and forth between the two processes
* to ensure the timing is correct, but basically, an example:
*
@@ -1166,7 +1166,7 @@ error:
* Purpose: This function is used to communicate with the verification
* process spawned by the start_refresh_verification_process
* function. It gives it the go-ahead to call H5*refresh
- * on an object and conlcude the refresh verification.
+ * on an object and conclude the refresh verification.
*
* Return: 0 on Success, 1 on Failure
*
@@ -1371,7 +1371,7 @@ error:
*
* Purpose: This function checks the status of external verification
* processes to see if they've succeeded. It checks for the
- * existance of flushrefresh_ERROR file. If present, that indicates
+ * existence of flushrefresh_ERROR file. If present, that indicates
* an external verification process has failed, and this function
* thus fails as well. If not present, then nothing else has
* failed, and this function succeeds.
diff --git a/test/gen_bounds.c b/test/gen_bounds.c
index 8148f16..87a9c18 100644
--- a/test/gen_bounds.c
+++ b/test/gen_bounds.c
@@ -858,12 +858,12 @@ gen_sel_files(const char *filename, H5F_libver_t low_bound, H5F_libver_t high_bo
if (H5Rcreate(&ref_wbuf[0], fid, DATASET, H5R_DATASET_REGION, sid) < 0)
TEST_ERROR;
- /* Create the reference datset */
+ /* Create the reference dataset */
if ((ref_did = H5Dcreate2(fid, SEL_EX_REG_DSET, H5T_STD_REF_DSETREG, ref_sid, H5P_DEFAULT,
H5P_DEFAULT, H5P_DEFAULT)) < 0)
TEST_ERROR;
- /* Write to the reference datset */
+ /* Write to the reference dataset */
if (H5Dwrite(ref_did, H5T_STD_REF_DSETREG, H5S_ALL, H5S_ALL, H5P_DEFAULT, ref_wbuf) < 0)
TEST_ERROR;
@@ -886,12 +886,12 @@ gen_sel_files(const char *filename, H5F_libver_t low_bound, H5F_libver_t high_bo
if (H5Rcreate(&ref_wbuf[0], fid, DATASET, H5R_DATASET_REGION, sid) < 0)
TEST_ERROR;
- /* Create the reference datset */
+ /* Create the reference dataset */
if ((ref_did = H5Dcreate2(fid, SEL_EX_IRR_DSET, H5T_STD_REF_DSETREG, ref_sid, H5P_DEFAULT,
H5P_DEFAULT, H5P_DEFAULT)) < 0)
TEST_ERROR;
- /* Write to the reference datset */
+ /* Write to the reference dataset */
if (H5Dwrite(ref_did, H5T_STD_REF_DSETREG, H5S_ALL, H5S_ALL, H5P_DEFAULT, ref_wbuf) < 0)
TEST_ERROR;
@@ -915,12 +915,12 @@ gen_sel_files(const char *filename, H5F_libver_t low_bound, H5F_libver_t high_bo
if (H5Rcreate(&ref_wbuf[0], fid, DATASET, H5R_DATASET_REGION, sid) < 0)
TEST_ERROR;
- /* Create the reference datset */
+ /* Create the reference dataset */
if ((ref_did = H5Dcreate2(fid, SEL_EX_PT_DSET, H5T_STD_REF_DSETREG, ref_sid, H5P_DEFAULT, H5P_DEFAULT,
H5P_DEFAULT)) < 0)
TEST_ERROR;
- /* Write to the reference datset */
+ /* Write to the reference dataset */
if (H5Dwrite(ref_did, H5T_STD_REF_DSETREG, H5S_ALL, H5S_ALL, H5P_DEFAULT, ref_wbuf) < 0)
TEST_ERROR;
diff --git a/test/gen_new_array.c b/test/gen_new_array.c
index 5811106..8161c4e 100644
--- a/test/gen_new_array.c
+++ b/test/gen_new_array.c
@@ -89,7 +89,7 @@ main(void)
if (ret < 0)
printf("field 1 insert<0!\n");
- /* Creat the array datatype */
+ /* Create the array datatype */
arr_type = H5Tarray_create2(H5T_NATIVE_FLOAT, ARRAY1_RANK, tdims1);
if (arr_type < 0)
printf("arr_type<0!\n");
@@ -104,7 +104,7 @@ main(void)
if (ret < 0)
printf("field 3 array close<0!\n");
- /* Creat the array datatype */
+ /* Create the array datatype */
arr_type = H5Tarray_create2(H5T_NATIVE_LONG, ARRAY1_RANK, tdims1);
if (arr_type < 0)
printf("arr_type<0!\n");
diff --git a/test/gen_old_group.c b/test/gen_old_group.c
index 1b4e0b2..15e67a4 100644
--- a/test/gen_old_group.c
+++ b/test/gen_old_group.c
@@ -19,7 +19,7 @@
* empty "symbol table" group.
*
* This file is used in the v1.7 branch (after the "compact group"
- * checkin) to test compatability. Compile and run this
+ * checkin) to test compatibility. Compile and run this
* program (with the 1.6.x branch), it will generate a file*
* called "group_old.h5". You need to move it to the test
* directory in the HDF5 v1.7 source tree. The test/stab.c
diff --git a/test/gen_plist.c b/test/gen_plist.c
index 10c6423..25d48cf 100644
--- a/test/gen_plist.c
+++ b/test/gen_plist.c
@@ -81,7 +81,7 @@ main(void)
H5AC_cache_image_config_t my_cache_image_config = {H5AC__CURR_CACHE_IMAGE_CONFIG_VERSION, TRUE, FALSE,
-1};
- /* check endianess */
+ /* check endianness */
{
short int word = 0x0001;
char * byte = (char *)&word;
diff --git a/test/genall5.c b/test/genall5.c
index 1da61f8..f3b0e6a 100644
--- a/test/genall5.c
+++ b/test/genall5.c
@@ -14,7 +14,7 @@
/* Programmer: John Mainzer
* 9/23/15
*
- * This file contains a heavily edited and functionaly reduce
+ * This file contains a heavily edited and functionally reduce
* version of the test code first written by Quincey in a file
* of the same name.
*/
diff --git a/test/getname.c b/test/getname.c
index 485e3c1..ba33280 100644
--- a/test/getname.c
+++ b/test/getname.c
@@ -1526,7 +1526,7 @@ test_main(hid_t file_id, hid_t fapl)
{
char name[NAME_BUF_SIZE]; /* Buffer to hold name and its size */
- /* Get name for non commited datatype, it should fail */
+ /* Get name for non committed datatype, it should fail */
H5E_BEGIN_TRY
{
if (H5Iget_name(type_id, name, NAME_BUF_SIZE) > 0)
diff --git a/test/gheap.c b/test/gheap.c
index 567fd65..f1182bd 100644
--- a/test/gheap.c
+++ b/test/gheap.c
@@ -353,7 +353,7 @@ error:
* Function: test_4
*
* Purpose: Tests the H5HG_remove() feature by writing lots of objects
- * and occassionally removing some. When we're done they're all
+ * and occasionally removing some. When we're done they're all
* removed.
*
* Return: Success: 0
diff --git a/test/h5test.c b/test/h5test.c
index e69a682..0528623 100644
--- a/test/h5test.c
+++ b/test/h5test.c
@@ -1665,7 +1665,7 @@ h5_send_message(const char *send, const char *arg1, const char *arg2)
* signal file from disk, and only continues once it has
* successfully done so (i.e., only after another process has
* called the "h5_send_message" function to write the signal file).
- * This functon will then immediately remove the file (i.e.,
+ * This function will then immediately remove the file (i.e.,
* to indicate that it has been received and can be reused),
* and then exits, allowing the calling function to continue.
*
diff --git a/test/links.c b/test/links.c
index e1a9dba..2fead45 100644
--- a/test/links.c
+++ b/test/links.c
@@ -8201,7 +8201,7 @@ external_link_dangling(hid_t fapl, hbool_t new_format)
if (status >= 0) {
H5_FAILED();
HDputs(
- " Retreiving name of object by index through dangling file external link should have failed.");
+ " Retrieving name of object by index through dangling file external link should have failed.");
} /* end if */
/* Close root group */
@@ -8674,7 +8674,7 @@ external_link_abstar(hid_t fapl, hbool_t new_format)
}
H5E_END_TRY;
- /* should be able to find the target file with abolute path */
+ /* should be able to find the target file with absolute path */
if (gid < 0) {
H5_FAILED();
HDputs(" Should have found the file in tmp_links directory.");
@@ -16598,7 +16598,7 @@ link_filters(hid_t fapl, hbool_t new_format)
TEST_ERROR
filesize_unfiltered = h5_get_file_size(filename, fapl);
- /* Set deflate fitler */
+ /* Set deflate filter */
if (H5Pset_deflate(fcpl, 6) < 0)
TEST_ERROR
@@ -22567,7 +22567,7 @@ main(void)
for (minimize_dset_oh = 0; minimize_dset_oh <= 1; minimize_dset_oh++) {
if (minimize_dset_oh) {
- HDprintf("\n-Testing with minimzed dataset object headers-\n");
+ HDprintf("\n-Testing with minimized dataset object headers-\n");
dcpl_g = H5Pcreate(H5P_DATASET_CREATE);
if (0 > dcpl_g)
TEST_ERROR
diff --git a/test/mf.c b/test/mf.c
index 5f2abc0..242e319 100644
--- a/test/mf.c
+++ b/test/mf.c
@@ -1937,7 +1937,7 @@ test_mf_fs_extend(hid_t fapl)
if (check_stats(f, f->shared->fs_man[H5FD_MEM_SUPER], &state))
TEST_ERROR
- /* Remove section A from free-space manger */
+ /* Remove section A from free-space manager */
if (H5MF__find_sect(f, H5FD_MEM_SUPER, (hsize_t)(TBLOCK_SIZE30 - 10), f->shared->fs_man[H5FD_MEM_SUPER],
&tmp) != TRUE)
TEST_ERROR
@@ -2015,7 +2015,7 @@ test_mf_fs_absorb(const char *env_h5_drvr, hid_t fapl)
haddr_t ma_addr = HADDR_UNDEF;
hsize_t ma_size = 0;
H5MF_free_section_t *sect_node = NULL;
- hbool_t contig_addr_vfd; /* Whether VFD used has a contigous address space */
+ hbool_t contig_addr_vfd; /* Whether VFD used has a contiguous address space */
TESTING("A free-space section absorbs an aggregator: test 1");
@@ -2502,7 +2502,7 @@ test_mf_aggr_alloc3(const char *env_h5_drvr, hid_t fapl)
hsize_t ma_size = 0, new_ma_size = 0;
haddr_t sdata_addr = HADDR_UNDEF;
hsize_t sdata_size = 0;
- hbool_t contig_addr_vfd; /* Whether VFD used has a contigous address space */
+ hbool_t contig_addr_vfd; /* Whether VFD used has a contiguous address space */
TESTING("H5MF_alloc() of meta/sdata aggregator: test 3");
@@ -2670,7 +2670,7 @@ test_mf_aggr_alloc4(const char *env_h5_drvr, hid_t fapl)
haddr_t addr1, addr2, saddr1, saddr2, saddr3;
haddr_t ma_addr = HADDR_UNDEF, new_ma_addr = HADDR_UNDEF, sdata_addr = HADDR_UNDEF;
hsize_t ma_size = 0, new_ma_size = 0, sdata_size = 0;
- hbool_t contig_addr_vfd; /* Whether VFD used has a contigous address space */
+ hbool_t contig_addr_vfd; /* Whether VFD used has a contiguous address space */
TESTING("H5MF_alloc() of meta/sdata aggregator:test 4");
@@ -2819,7 +2819,7 @@ test_mf_aggr_alloc5(const char *env_h5_drvr, hid_t fapl)
haddr_t addr1, addr2, addr3;
haddr_t ma_addr = HADDR_UNDEF, new_ma_addr = HADDR_UNDEF;
hsize_t ma_size = 0, new_ma_size = 0;
- hbool_t contig_addr_vfd; /* Whether VFD used has a contigous address space */
+ hbool_t contig_addr_vfd; /* Whether VFD used has a contiguous address space */
TESTING("H5MF_alloc() of meta/sdata aggregator:test 5");
@@ -2957,7 +2957,7 @@ test_mf_aggr_alloc6(const char *env_h5_drvr, hid_t fapl)
haddr_t ma_addr = HADDR_UNDEF, new_ma_addr = HADDR_UNDEF, sdata_addr = HADDR_UNDEF;
hsize_t ma_size = 0, new_ma_size = 0, sdata_size = 0;
H5FS_stat_t state;
- hbool_t contig_addr_vfd; /* Whether VFD used has a contigous address space */
+ hbool_t contig_addr_vfd; /* Whether VFD used has a contiguous address space */
TESTING("H5MF_alloc() of meta/sdata aggregator:test 6");
@@ -3131,7 +3131,7 @@ test_mf_aggr_alloc7(const char *env_h5_drvr, hid_t fapl)
haddr_t ma_addr = HADDR_UNDEF, sdata_addr = HADDR_UNDEF;
hsize_t ma_size = 0, sdata_size = 0;
H5FS_stat_t state;
- hbool_t contig_addr_vfd; /* Whether VFD used has a contigous address space */
+ hbool_t contig_addr_vfd; /* Whether VFD used has a contiguous address space */
TESTING("H5MF_alloc() of meta/sdata aggregator:test 7");
@@ -3302,7 +3302,7 @@ test_mf_aggr_extend(const char *env_h5_drvr, hid_t fapl)
haddr_t ma_addr = HADDR_UNDEF, new_ma_addr = HADDR_UNDEF, sdata_addr = HADDR_UNDEF;
hsize_t ma_size = 0, new_ma_size = 0, sdata_size = 0;
htri_t was_extended;
- hbool_t contig_addr_vfd; /* Whether VFD used has a contigous address space */
+ hbool_t contig_addr_vfd; /* Whether VFD used has a contiguous address space */
TESTING("H5MF_try_extend() of meta/sdata aggregator: test 1");
@@ -3600,7 +3600,7 @@ test_mf_aggr_absorb(const char *env_h5_drvr, hid_t fapl)
haddr_t new_sdata_addr = HADDR_UNDEF;
hsize_t ma_size = 0, new_ma_size = 0;
hsize_t sdata_size = 0, new_sdata_size = 0;
- hbool_t contig_addr_vfd; /* Whether VFD used has a contigous address space */
+ hbool_t contig_addr_vfd; /* Whether VFD used has a contiguous address space */
TESTING("H5MF_try_shrink() of meta/sdata aggregator: test 1");
@@ -5855,7 +5855,7 @@ error:
*
* Alignment = 4096 aggr->alloc_size = 2048
*
- * Allocate first block (30) from meta_aggr: (meta_aggr is emtpy)
+ * Allocate first block (30) from meta_aggr: (meta_aggr is empty)
* request-size is > meta_aggr->size and < meta_aggr->alloc_size
* Result:
* A block of meta_aggr->alloc_size is allocated from file allocation
@@ -7128,7 +7128,7 @@ test_mf_fs_gone(const char *env_h5_drvr, hid_t fapl, hbool_t new_format)
else
TESTING("File's free-space is going away with old library format")
- /* Current VFD that does not support contigous address space */
+ /* Current VFD that does not support contiguous address space */
contig_addr_vfd = (hbool_t)(HDstrcmp(env_h5_drvr, "split") != 0 && HDstrcmp(env_h5_drvr, "multi") != 0);
if (contig_addr_vfd) {
@@ -7319,7 +7319,7 @@ test_mf_strat_thres_persist(const char *env_h5_drvr, hid_t fapl, hbool_t new_for
else
TESTING("File space strategy/persisting/threshold with old library format")
- /* Current VFD that does not support contigous address space */
+ /* Current VFD that does not support contiguous address space */
contig_addr_vfd = (hbool_t)(HDstrcmp(env_h5_drvr, "split") != 0 && HDstrcmp(env_h5_drvr, "multi") != 0);
/* Set the filename to use for this test (dependent on fapl) */
@@ -7489,7 +7489,7 @@ test_mf_strat_thres_gone(const char *env_h5_drvr, hid_t fapl, hbool_t new_format
else
TESTING("File space merge/shrink for section size < threshold with old library format")
- /* Current VFD that does not support contigous address space */
+ /* Current VFD that does not support contiguous address space */
contig_addr_vfd = (hbool_t)(HDstrcmp(env_h5_drvr, "split") != 0 && HDstrcmp(env_h5_drvr, "multi") != 0);
/* Set the filename to use for this test (dependent on fapl) */
@@ -8047,7 +8047,7 @@ test_page_try_shrink(const char *env_h5_drvr, hid_t fapl)
haddr_t addr1; /* Address for small metadata block */
haddr_t saddr1; /* Address for small raw data block */
haddr_t gaddr1; /* Address for large data block */
- hbool_t contig_addr_vfd; /* Whether VFD used has a contigous address space */
+ hbool_t contig_addr_vfd; /* Whether VFD used has a contiguous address space */
htri_t status; /* status from shrinking */
h5_stat_size_t file_size; /* File size */
char filename[FILENAME_LEN]; /* Filename to use */
@@ -8177,7 +8177,7 @@ test_page_small_try_extend(const char *env_h5_drvr, hid_t fapl)
H5F_t * f = NULL; /* Internal file object pointer */
haddr_t addr1, addr2, addr3; /* Addresses for small metadata blocks */
haddr_t saddr1; /* Address for small raw data block */
- hbool_t contig_addr_vfd; /* Whether VFD used has a contigous address space */
+ hbool_t contig_addr_vfd; /* Whether VFD used has a contiguous address space */
htri_t was_extended; /* Whether the block can be extended or not */
char filename[FILENAME_LEN]; /* Filename to use */
@@ -8356,7 +8356,7 @@ test_page_large_try_extend(const char *env_h5_drvr, hid_t fapl)
hid_t fcpl = -1; /* File creation property list */
H5F_t * f = NULL; /* Internal file object pointer */
haddr_t gaddr1, gaddr2, gaddr3, gaddr4; /* Addresses for large data blocks */
- hbool_t contig_addr_vfd; /* Whether VFD used has a contigous address space */
+ hbool_t contig_addr_vfd; /* Whether VFD used has a contiguous address space */
htri_t was_extended; /* Whether the block can be extended or not */
char filename[FILENAME_LEN]; /* Filename to use */
@@ -8519,7 +8519,7 @@ test_page_large(const char *env_h5_drvr, hid_t fapl)
hid_t fcpl = -1; /* File creation property list ID */
H5F_t * f = NULL; /* Internal file object pointer */
haddr_t gaddr1, gaddr2, gaddr3, gaddr4; /* Addresses for blocks */
- hbool_t contig_addr_vfd; /* Whether VFD used has a contigous address space */
+ hbool_t contig_addr_vfd; /* Whether VFD used has a contiguous address space */
H5FS_stat_t fs_stat; /* Information for free-space manager */
h5_stat_size_t file_size; /* File size */
char filename[FILENAME_LEN]; /* Filename to use */
diff --git a/test/mirror_vfd.c b/test/mirror_vfd.c
index 3556cd2..7047146 100644
--- a/test/mirror_vfd.c
+++ b/test/mirror_vfd.c
@@ -128,7 +128,7 @@ mybzero(void *dest, size_t size)
* Function: _populate_filepath
*
* Purpose: Given a directory name and a base name, concatenate the two and
- * run h5fixname() to get the "actual" path to the intented target.
+ * run h5fixname() to get the "actual" path to the intended target.
* `h5suffix' should be FALSE to keep the base name unaltered;
* TRUE will append the '.h5' h5suffix to the basename...
* FALSE -> h5fixname_no_suffix(), TRUE -> h5fixname()
diff --git a/test/mount.c b/test/mount.c
index c5ac9b3..b23de30 100644
--- a/test/mount.c
+++ b/test/mount.c
@@ -753,7 +753,7 @@ test_move(hid_t fapl)
H5E_END_TRY;
if (status >= 0) {
H5_FAILED();
- HDputs(" Moving an object across files should't have been possible");
+ HDputs(" Moving an object across files shouldn't have been possible");
TEST_ERROR
} /* end if */
@@ -1835,7 +1835,7 @@ test_mount_after_unmount(hid_t fapl)
if (HDstrcmp(objname, "/X/M/Y") != 0)
TEST_ERROR
- /* Rename object in file #3 that is "disconnected" from name hiearchy */
+ /* Rename object in file #3 that is "disconnected" from name hierarchy */
/* (It is "disconnected" because it's parent file has been unmounted) */
if (H5Lmove(gidAMX, "M/Y", gidAMX, "M/Z", H5P_DEFAULT, H5P_DEFAULT) < 0)
TEST_ERROR
diff --git a/test/mtime.c b/test/mtime.c
index 10cafd1..8e89d58 100644
--- a/test/mtime.c
+++ b/test/mtime.c
@@ -117,7 +117,7 @@ main(void)
if (0 == oi1.ctime) {
SKIPPED();
HDputs(" The modification time could not be decoded on this OS.");
- HDputs(" Modification times will be mantained in the file but");
+ HDputs(" Modification times will be maintained in the file but");
HDputs(" cannot be queried on this system. See H5O_mtime_decode().");
return 0;
}
diff --git a/test/objcopy.c b/test/objcopy.c
index 911a45b..e23ee8e 100644
--- a/test/objcopy.c
+++ b/test/objcopy.c
@@ -2789,7 +2789,7 @@ test_copy_dataset_simple_empty(hid_t fcpl_src, hid_t fcpl_dst, hid_t src_fapl, h
char dst_filename[NAME_BUF_SIZE];
if (test_open) {
- TESTING("H5Ocopy(): empty and openend contiguous dataset");
+ TESTING("H5Ocopy(): empty and opened contiguous dataset");
}
else {
TESTING("H5Ocopy(): empty contiguous dataset");
@@ -8679,7 +8679,7 @@ test_copy_soft_link(hid_t fcpl_src, hid_t fcpl_dst, hid_t src_fapl, hid_t dst_fa
char dst_filename[NAME_BUF_SIZE];
if (test_open) {
- TESTING("H5Ocopy(): openend object through soft link");
+ TESTING("H5Ocopy(): opened object through soft link");
}
else {
TESTING("H5Ocopy(): object through soft link");
@@ -9674,7 +9674,7 @@ test_copy_dataset_contig_named_vl(hid_t fcpl_src, hid_t fcpl_dst, hid_t src_fapl
char src_filename[NAME_BUF_SIZE];
char dst_filename[NAME_BUF_SIZE];
- TESTING("H5Ocopy(): contigous dataset with named VLEN datatype");
+ TESTING("H5Ocopy(): contiguous dataset with named VLEN datatype");
/* set initial data values */
for (i = 0; i < DIM_SIZE_1; i++) {
@@ -10434,10 +10434,10 @@ test_copy_dataset_contig_vl_vl(hid_t fcpl_src, hid_t fcpl_dst, hid_t src_fapl, h
char dst_filename[NAME_BUF_SIZE];
if (test_open) {
- TESTING("H5Ocopy(): contigous and opened dataset with nested VLEN datatype");
+ TESTING("H5Ocopy(): contiguous and opened dataset with nested VLEN datatype");
}
else {
- TESTING("H5Ocopy(): contigous dataset with nested VLEN datatype");
+ TESTING("H5Ocopy(): contiguous dataset with nested VLEN datatype");
}
/* set initial data values */
@@ -13763,7 +13763,7 @@ test_copy_cdt_hier_merge(hid_t fcpl_src, hid_t fcpl_dst, hid_t src_fapl, hid_t d
TEST_ERROR
/*
- * Test 3: copy datsets in /g0 at SRC to DST group /uncopied
+ * Test 3: copy datasets in /g0 at SRC to DST group /uncopied
*/
if (H5Ocopy(fid_src, NAME_GROUP_TOP "/" SRC_DSET1, fid_dst, NAME_GROUP_UNCOPIED "/" SRC_DSET1, ocpypl_id,
H5P_DEFAULT) < 0)
diff --git a/test/ohdr.c b/test/ohdr.c
index 60d32b7..c7a13fa 100644
--- a/test/ohdr.c
+++ b/test/ohdr.c
@@ -591,7 +591,7 @@ test_unknown(unsigned bogus_id, char *filename, hid_t fapl)
TESTING("object with unknown header message & 'shareable' flag set");
- /* Open the dataset with the unknown header message, adn "shareable" flag */
+ /* Open the dataset with the unknown header message, and "shareable" flag */
if ((did = H5Dopen2(loc_bogus, "Dataset5", H5P_DEFAULT)) < 0)
FAIL_STACK_ERROR
if (H5Dclose(did) < 0)
@@ -1096,13 +1096,13 @@ test_minimized_dset_ohdr_size_comparisons(hid_t fapl_id)
/* IDs for non-minimized file open */
hid_t file_f_id = -1; /* lower 'f' for standard file setting */
hid_t dset_f_x_id = -1; /* 'x' for default */
- hid_t dset_f_N_id = -1; /* 'N' for explcit non-minimized dset */
+ hid_t dset_f_N_id = -1; /* 'N' for explicit non-minimized dset */
hid_t dset_f_Y_id = -1; /* 'Y' for minimized dset */
/* IDs for minimized file open */
hid_t file_F_id = -1; /* upper 'F' for minimized file setting */
hid_t dset_F_x_id = -1; /* 'x' for default */
- hid_t dset_F_N_id = -1; /* 'N' for explcit non-minimized dset */
+ hid_t dset_F_N_id = -1; /* 'N' for explicit non-minimized dset */
hid_t dset_F_Y_id = -1; /* 'Y' for minimized dset */
char filename_a[512] = "";
diff --git a/test/page_buffer.c b/test/page_buffer.c
index cb59fda..c078ce8 100644
--- a/test/page_buffer.c
+++ b/test/page_buffer.c
@@ -75,7 +75,7 @@ const char *FILENAME[] = {"filepaged", NULL};
* Any data mis-matches or failures reported by the HDF5
* library result in test failure.
*
- * Return: 0 if test is sucessful
+ * Return: 0 if test is successful
* 1 if test fails
*
* Programmer: unknown
@@ -215,7 +215,7 @@ error:
* Any data mis-matches or failures reported by the HDF5
* library result in test failure.
*
- * Return: 0 if test is sucessful
+ * Return: 0 if test is successful
* 1 if test fails
*
* Programmer: unknown
@@ -375,7 +375,7 @@ error:
* Any data mis-matches or unexpected failures or successes
* reported by the HDF5 library result in test failure.
*
- * Return: 0 if test is sucessful
+ * Return: 0 if test is successful
* 1 if test fails
*
* Programmer: unknown
@@ -549,7 +549,7 @@ error:
* Any data mis-matches or failures reported by the HDF5
* library result in test failure.
*
- * Return: 0 if test is sucessful
+ * Return: 0 if test is successful
* 1 if test fails
*
* Programmer: unknown
@@ -827,7 +827,7 @@ error:
* Any data mis-matches or failures reported by the HDF5
* library result in test failure.
*
- * Return: 0 if test is sucessful
+ * Return: 0 if test is successful
* 1 if test fails
*
* Programmer: unknown
@@ -1081,7 +1081,7 @@ error:
* Any data mis-matches or failures reported by the HDF5
* library result in test failure.
*
- * Return: 0 if test is sucessful
+ * Return: 0 if test is successful
* 1 if test fails
*
* Programmer: unknown
@@ -1502,7 +1502,7 @@ test_min_threshold(hid_t orig_fapl, const char *env_h5_drvr)
if (f->shared->page_buf->raw_count != 2)
TEST_ERROR;
- /* adding more meta entires should replace meta entries since raw data
+ /* adding more meta entries should replace meta entries since raw data
* is at its minimum
*/
if (H5F_block_write(f, H5FD_MEM_SUPER, meta_addr + (sizeof(int) * 600), sizeof(int) * 100, data) < 0)
@@ -1517,7 +1517,7 @@ test_min_threshold(hid_t orig_fapl, const char *env_h5_drvr)
if (f->shared->page_buf->raw_count != 2)
TEST_ERROR;
- /* bring existing raw entires up the LRU */
+ /* bring existing raw entries up the LRU */
if (H5F_block_read(f, H5FD_MEM_DRAW, raw_addr + (sizeof(int) * 750), sizeof(int) * 100, data) < 0)
FAIL_STACK_ERROR;
@@ -1533,7 +1533,7 @@ test_min_threshold(hid_t orig_fapl, const char *env_h5_drvr)
if (f->shared->page_buf->raw_count != 3)
TEST_ERROR;
- /* adding 2 meta entries should replace 2 entires at the bottom of the LRU */
+ /* adding 2 meta entries should replace 2 entries at the bottom of the LRU */
if (H5F_block_read(f, H5FD_MEM_SUPER, meta_addr + (sizeof(int) * 98), sizeof(int) * 100, data) < 0)
FAIL_STACK_ERROR;
@@ -1716,7 +1716,7 @@ error:
* Any data mis-matches or failures reported by the HDF5
* library result in test failure.
*
- * Return: 0 if test is sucessful
+ * Return: 0 if test is successful
* 1 if test fails
*
* Programmer: unknown
@@ -2019,7 +2019,7 @@ error:
* At present, page buffering should be disabled in parallel
* builds. Verify this.
*
- * Return: 0 if test is sucessful
+ * Return: 0 if test is successful
* 1 if test fails
*
* Programmer: John Mainzer
@@ -2134,7 +2134,7 @@ error:
*
* Purpose: Main function for the page buffer tests.
*
- * Return: 0 if test is sucessful
+ * Return: 0 if test is successful
* 1 if test fails
*
* Programmer: unknown
diff --git a/test/ros3.c b/test/ros3.c
index b7b27ed..a340985 100644
--- a/test/ros3.c
+++ b/test/ros3.c
@@ -1763,7 +1763,7 @@ main(void)
bucket_url_env = HDgetenv("HDF5_ROS3_TEST_BUCKET_URL");
if (bucket_url_env == NULL || bucket_url_env[0] == '\0') {
- HDprintf("WARNING: S3 bucket url is not defined in enviornment "
+ HDprintf("WARNING: S3 bucket url is not defined in environment "
"variable 'HDF5_ROS3_TEST_BUCKET_URL'!\n");
}
else {
diff --git a/test/s3comms.c b/test/s3comms.c
index cec7a27..45415cb 100644
--- a/test/s3comms.c
+++ b/test/s3comms.c
@@ -244,7 +244,7 @@ jserr_str(const char *expected, const char *actual, const char *reason)
}
#ifdef JSVERIFY_EXP_ACT
-/* VERIFY rountines with paramter order (<expected>, <actual> [, <msg> ])
+/* VERIFY routines with parameter order (<expected>, <actual> [, <msg> ])
*/
/*----------------------------------------------------------------------------
@@ -435,7 +435,7 @@ error:
*
* Purpose:
*
- * Demonstrate the construction of a Canoncial Request (and Signed Headers)
+ * Demonstrate the construction of a Canonical Request (and Signed Headers)
*
* Elided / not yet implemented:
* Query strings
@@ -730,7 +730,7 @@ test_hrb_init_request(void)
FALSE,
},
{
- "null verb substitues to GET",
+ "null verb substitutes to GET",
NULL,
"/MYPATH/MYFILE.tiff",
"/MYPATH/MYFILE.tiff",
@@ -1143,10 +1143,10 @@ test_hrb_node_set(void)
*********/
for (mock_i = 0; test->given[mock_i] != NULL; mock_i += 2) {
- const char *name = test->given[mock_i];
- const char *valu = test->given[mock_i + 1];
+ const char *name = test->given[mock_i];
+ const char *value = test->given[mock_i + 1];
- FAIL_IF(SUCCEED != H5FD_s3comms_hrb_node_set(&list, name, valu))
+ FAIL_IF(SUCCEED != H5FD_s3comms_hrb_node_set(&list, name, value))
}
/********
* TEST *
@@ -1162,11 +1162,11 @@ test_hrb_node_set(void)
node = list;
mock_i = 0;
while (test->expected[mock_i] != NULL && node != NULL) {
- const char *name = test->expected[mock_i];
- const char *valu = test->expected[mock_i + 1];
+ const char *name = test->expected[mock_i];
+ const char *value = test->expected[mock_i + 1];
JSVERIFY_STR(name, node->name, NULL)
- JSVERIFY_STR(valu, node->value, NULL)
+ JSVERIFY_STR(value, node->value, NULL)
mock_i += 2;
node = node->next;
@@ -1559,7 +1559,7 @@ test_parse_url(void)
NULL,
"a=b&d=b",
},
- "QUERY with implict PATH",
+ "QUERY with implicit PATH",
},
{
"http://[5]/path?a=b&d=b",
@@ -2014,7 +2014,7 @@ error:
* H5FD_s3comms_s3r_read << called by getsize(), multiple times working
* H5FD_s3comms_s3r_close
*
- * Shows most basic curl interation.
+ * Shows most basic curl iteration.
*
* Programmer: Jacob Smith
* 2017-10-06
@@ -2610,7 +2610,7 @@ main(void)
bucket_url_env = HDgetenv("HDF5_ROS3_TEST_BUCKET_URL");
if (bucket_url_env == NULL || bucket_url_env[0] == '\0') {
- HDprintf("WARNING: S3 bucket url is not defined in enviornment "
+ HDprintf("WARNING: S3 bucket url is not defined in environment "
"variable 'HDF5_ROS3_TEST_BUCKET_URL'!\n");
}
else {
@@ -2618,7 +2618,7 @@ main(void)
s3_test_bucket_defined = TRUE;
}
- /* tests ordered rougly by dependence */
+ /* tests ordered roughly by dependence */
nerrors += test_macro_format_credential() < 0 ? 1 : 0;
nerrors += test_trim() < 0 ? 1 : 0;
nerrors += test_nlowercase() < 0 ? 1 : 0;
diff --git a/test/set_extent.c b/test/set_extent.c
index ab2a443..4437219 100644
--- a/test/set_extent.c
+++ b/test/set_extent.c
@@ -115,12 +115,12 @@ main(void)
unsigned chunk_cache; /* Whether to enable chunk caching */
int nerrors = 0;
const char *env_h5_drvr; /* File Driver value from environment */
- hbool_t contig_addr_vfd; /* Whether VFD used has a contigous address space */
+ hbool_t contig_addr_vfd; /* Whether VFD used has a contiguous address space */
env_h5_drvr = HDgetenv(HDF5_DRIVER);
if (env_h5_drvr == NULL)
env_h5_drvr = "nomatch";
- /* Current VFD that does not support contigous address space */
+ /* Current VFD that does not support contiguous address space */
contig_addr_vfd = (hbool_t)(HDstrcmp(env_h5_drvr, "split") != 0 && HDstrcmp(env_h5_drvr, "multi") != 0);
/* Initialize random number seed */
@@ -656,7 +656,7 @@ test_rank1(hid_t fapl, hid_t dcpl, hbool_t do_fill_value, hbool_t disable_edge_f
HDprintf("\n");
#endif
- /* compare the read array with the shrinked array */
+ /* compare the read array with the shrunk array */
for (i = 0; i < (int)dims_r[0]; i++)
if (buf_s[i] != buf_o[i]) {
HDprintf("buf_s[%d] = %d\n", i, buf_s[i]);
@@ -1062,7 +1062,7 @@ test_rank2(hid_t fapl, hid_t dcpl, hbool_t do_fill_value, hbool_t disable_edge_f
}
#endif
- /* compare the read array with the shrinked array */
+ /* compare the read array with the shrunk array */
for (i = 0; i < (int)dims_r[0]; i++) {
for (j = 0; j < (int)dims_r[1]; j++) {
if (buf_s[i][j] != buf_o[i][j]) {
@@ -1543,7 +1543,7 @@ test_rank3(hid_t fapl, hid_t dcpl, hbool_t do_fill_value, hbool_t disable_edge_f
HDprintf("\n");
#endif
- /* compare the read array with the shrinked array */
+ /* compare the read array with the shrunk array */
for (i = 0; i < (int)dims_r[0]; i++) {
for (j = 0; j < (int)dims_r[1]; j++) {
for (k = 0; k < (int)dims_r[2]; k++) {
@@ -1741,7 +1741,7 @@ test_external(hid_t fapl)
hsize_t dims_r[RANK2]; /* read dimensions */
hsize_t maxdims[RANK2] = {DIME0, DIM1}; /* only the first dimension can be extendible */
int buf_o[DIM0][DIM1]; /* original buffer, for writing */
- int buf_s[DIMS0][DIMS1]; /* shrinked buffer, for reading */
+ int buf_s[DIMS0][DIMS1]; /* shrunk buffer, for reading */
int buf_e[DIME0][DIM1]; /* extended buffer, for writing, dimension 1 is the original */
int buf_ro[DIM0][DIM1]; /* original buffer for reading */
int i, j;
@@ -1953,7 +1953,7 @@ test_external(hid_t fapl)
}
#endif
- /* compare the read array with the shrinked array */
+ /* compare the read array with the shrunk array */
for (i = 0; i < (int)dims_r[0]; i++) {
for (j = 0; j < (int)dims_r[1]; j++) {
if (buf_s[i][j] != buf_o[i][j]) {
diff --git a/test/stab.c b/test/stab.c
index 31ff497..b3db76e 100644
--- a/test/stab.c
+++ b/test/stab.c
@@ -125,7 +125,7 @@ test_misc(hid_t fcpl, hid_t fapl, hbool_t new_format)
if (H5Gclose(g3) < 0)
TEST_ERROR
- /* Open all groups with absolute names to check for exsistence */
+ /* Open all groups with absolute names to check for existence */
if ((g1 = H5Gopen2(fid, "/test_1a", H5P_DEFAULT)) < 0)
TEST_ERROR
if ((g2 = H5Gopen2(fid, "/test_1a/sub_1", H5P_DEFAULT)) < 0)
@@ -1416,7 +1416,7 @@ main(void)
hid_t fcpl, fcpl2; /* File creation property list ID */
unsigned new_format; /* Whether to use the new format or not */
const char *env_h5_drvr; /* File Driver value from environment */
- hbool_t contig_addr_vfd; /* Whether VFD used has a contigous address space */
+ hbool_t contig_addr_vfd; /* Whether VFD used has a contiguous address space */
hbool_t driver_uses_modified_filename = h5_driver_uses_modified_filename();
int nerrors = 0;
@@ -1425,7 +1425,7 @@ main(void)
if (env_h5_drvr == NULL)
env_h5_drvr = "nomatch";
- /* VFD that does not support contigous address space */
+ /* VFD that does not support contiguous address space */
contig_addr_vfd = (hbool_t)(HDstrcmp(env_h5_drvr, "split") != 0 && HDstrcmp(env_h5_drvr, "multi") != 0);
/* Reset library */
diff --git a/test/swmr.c b/test/swmr.c
index bf84fcd..693ddc7 100644
--- a/test/swmr.c
+++ b/test/swmr.c
@@ -3666,7 +3666,7 @@ test_append_flush_dataset_chunked(hid_t in_fapl)
{
hid_t fid = -1; /* file ID */
hid_t fapl = -1; /* A copy of file access property */
- hid_t did1 = -1, did2 = -1; /* The datset ID */
+ hid_t did1 = -1, did2 = -1; /* The dataset ID */
hid_t sid = -1; /* The dataspace ID */
hid_t dcpl = -1; /* A copy of dataset creation property */
hid_t dapl = -1; /* A copy of dataset access property */
@@ -3889,7 +3889,7 @@ test_append_flush_dataset_fixed(hid_t in_fapl)
{
hid_t fid = -1; /* file ID */
hid_t fapl = -1; /* A copy of file access property */
- hid_t did1 = -1, did2 = -1; /* The datset ID */
+ hid_t did1 = -1, did2 = -1; /* The dataset ID */
hid_t sid = -1; /* The dataspace ID */
hid_t dapl = -1; /* A copy of dataset access property */
hid_t ddapl = -1; /* The dataset access property of the opened dataset */
@@ -4107,7 +4107,7 @@ test_append_flush_dataset_multiple(hid_t in_fapl)
{
hid_t fid = -1; /* file ID */
hid_t fapl = -1; /* A copy of file access property */
- hid_t did1 = -1, did2 = -1; /* The datset ID */
+ hid_t did1 = -1, did2 = -1; /* The dataset ID */
hid_t sid = -1; /* The dataspace ID */
hid_t dcpl = -1; /* A copy of dataset creation property */
hid_t dapl1 = -1; /* A copy of dataset access property */
@@ -7412,7 +7412,7 @@ main(void)
/*
* Modify the following routines to test for files:
* H5Fcreate(write, latest format) or H5Fcreate(SWMR write, non-latest-format)
- * --both result in v3 superblock and latest version suppport
+ * --both result in v3 superblock and latest version support
*/
nerrors += test_start_swmr_write(fapl, TRUE);
nerrors += test_start_swmr_write(fapl, FALSE);
@@ -7446,7 +7446,7 @@ main(void)
*
* Modify the following 2 routines to test for files:
* H5Fcreate(write, latest format) or H5Fcreate(SWMR write, non-latest-format)
- * --both result in v3 superblock and latest version suppport
+ * --both result in v3 superblock and latest version support
*/
nerrors += test_file_lock_swmr_same(fapl);
nerrors += test_file_lock_swmr_concur(fapl);
diff --git a/test/swmr_generator.c b/test/swmr_generator.c
index 2bff12c..e1423b6 100644
--- a/test/swmr_generator.c
+++ b/test/swmr_generator.c
@@ -221,7 +221,7 @@ gen_skeleton(const char *filename, hbool_t verbose, hbool_t swmr_write, int comp
if (verbose)
HDfprintf(stderr, "Closing objects\n");
- /* Close everythign */
+ /* Close everything */
if (H5Pclose(dcpl) < 0)
return -1;
if (H5Sclose(sid) < 0)
diff --git a/test/swmr_start_write.c b/test/swmr_start_write.c
index 2d25801..3ddd57a 100644
--- a/test/swmr_start_write.c
+++ b/test/swmr_start_write.c
@@ -125,7 +125,7 @@ create_file(const char *filename, hbool_t verbose, FILE *verbose_file, unsigned
*
* Parameters:
* fid: file ID for the SWMR test file
- * comp_level: the compresssion level
+ * comp_level: the compression level
* index_type: The chunk index type (b1 | b2 | ea | fa)
* verbose: whether verbose console output is desired.
* verbose_file: file pointer for verbose output
diff --git a/test/tattr.c b/test/tattr.c
index b807bce..d66fcc3 100644
--- a/test/tattr.c
+++ b/test/tattr.c
@@ -5886,7 +5886,7 @@ test_attr_info_by_idx(hbool_t new_format, hid_t fcpl, hid_t fapl)
is_dense = H5O__is_attr_dense_test(my_dataset);
VERIFY(is_dense, FALSE, "H5O__is_attr_dense_test");
- /* Check for query on non-existant attribute */
+ /* Check for query on non-existent attribute */
ret = H5Aget_info_by_idx(my_dataset, ".", H5_INDEX_CRT_ORDER, H5_ITER_INC, (hsize_t)0, &ainfo,
H5P_DEFAULT);
VERIFY(ret, FAIL, "H5Aget_info_by_idx");
@@ -6374,7 +6374,7 @@ test_attr_delete_by_idx(hbool_t new_format, hid_t fcpl, hid_t fapl)
is_dense = H5O__is_attr_dense_test(my_dataset);
VERIFY(is_dense, FALSE, "H5O__is_attr_dense_test");
- /* Check for deleting non-existant attribute */
+ /* Check for deleting non-existent attribute */
ret = H5Adelete_by_idx(my_dataset, ".", idx_type, order, (hsize_t)0, H5P_DEFAULT);
VERIFY(ret, FAIL, "H5Adelete_by_idx");
@@ -8026,7 +8026,7 @@ test_attr_open_by_name(hbool_t new_format, hid_t fcpl, hid_t fapl)
is_dense = H5O__is_attr_dense_test(my_dataset);
VERIFY(is_dense, FALSE, "H5O__is_attr_dense_test");
- /* Check for opening a non-existant attribute on an object with no attributes */
+ /* Check for opening a non-existent attribute on an object with no attributes */
ret_id = H5Aopen(my_dataset, "foo", H5P_DEFAULT);
VERIFY(ret_id, FAIL, "H5Aopen");
@@ -8065,7 +8065,7 @@ test_attr_open_by_name(hbool_t new_format, hid_t fcpl, hid_t fapl)
is_dense = H5O__is_attr_dense_test(my_dataset);
VERIFY(is_dense, FALSE, "H5O__is_attr_dense_test");
- /* Check for opening a non-existant attribute on an object with compact attribute storage */
+ /* Check for opening a non-existent attribute on an object with compact attribute storage */
ret_id = H5Aopen(my_dataset, "foo", H5P_DEFAULT);
VERIFY(ret_id, FAIL, "H5Aopen");
@@ -8146,7 +8146,7 @@ test_attr_open_by_name(hbool_t new_format, hid_t fcpl, hid_t fapl)
VERIFY(name_count, (max_compact * 2), "H5O__attr_dense_info_test");
} /* end if */
- /* Check for opening a non-existant attribute on an object with dense attribute storage */
+ /* Check for opening a non-existent attribute on an object with dense attribute storage */
ret_id = H5Aopen(my_dataset, "foo", H5P_DEFAULT);
VERIFY(ret_id, FAIL, "H5Aopen");
@@ -11022,7 +11022,7 @@ test_attr_bug8(hid_t fcpl, hid_t fapl)
** chunk. Do this twice: once with only attributes in the
** object header chunk and once with a (small) soft link in
** the chunk as well. In both cases, the shrunk chunk will
-** initally be too small and a new NULL message must be
+** initially be too small and a new NULL message must be
** created.
**
****************************************************************/
diff --git a/test/tcoords.c b/test/tcoords.c
index fef8689..b2d68d3 100644
--- a/test/tcoords.c
+++ b/test/tcoords.c
@@ -17,7 +17,7 @@
*
* Test the element coordinates for dataspace selection. For
* chunked dataset, when the hyperslab selection of some
- * dimensions is full, the library optimize it by "flattenning"
+ * dimensions is full, the library optimize it by "flattening"
* the fully selected dimensions. This program tests if the
* coordinates of selected elements are correctly calculated.
*
diff --git a/test/testcheck_version.sh.in b/test/testcheck_version.sh.in
index 6378ee5..273702e 100644
--- a/test/testcheck_version.sh.in
+++ b/test/testcheck_version.sh.in
@@ -36,7 +36,7 @@ RM='rm -f'
PURPOSE() {
echo "Tests for the H5check_version function."
echo "Note that abort messages may appear due to the expected termination"
- echo "of the program when it is tested with mis-matched version numnbers."
+ echo "of the program when it is tested with mis-matched version numbers."
}
# Print a line-line message left justified in a field of 70 characters.
@@ -232,7 +232,7 @@ esac
# RUNSERIAL is used. Check if it can return exit code from executalbe correctly.
if [ -n "$RUNSERIAL_NOEXITCODE" ]; then
- echo "***Warning*** Serial Exit Code is not passed back to shell corretly."
+ echo "***Warning*** Serial Exit Code is not passed back to shell correctly."
echo "***Warning*** Exit code checking is skipped."
h5haveexitcode=no
fi
diff --git a/test/testlibinfo.sh.in b/test/testlibinfo.sh.in
index 2a9a05c..d9fee48 100644
--- a/test/testlibinfo.sh.in
+++ b/test/testlibinfo.sh.in
@@ -19,7 +19,7 @@
# libhdf5.settings file.
# Part 2:
# If executable is linked with the static hdf5 library (how to determine?),
-# verify an executable indeed contains an exact copy of hte content of the
+# verify an executable indeed contains an exact copy of the content of the
# libhdf5.settings file.
#
# Programmer: Albert Cheng
diff --git a/test/testmeta.c b/test/testmeta.c
index 7cb3000..a3b5160 100644
--- a/test/testmeta.c
+++ b/test/testmeta.c
@@ -15,7 +15,7 @@
* This program illustrates assertion errors when linked
* to HDF5 1.4.1-post2 or 1.4.2-pre3 debug library.
*
- * If the assertion errors are ignored, the program eventially causes
+ * If the assertion errors are ignored, the program eventually causes
* an error in H5Gcreate2 when writing object 83381.
*
* When writing in single file mode, the assertion errors still occur
diff --git a/test/testswmr.sh.in b/test/testswmr.sh.in
index cb60fce..65bf821 100644
--- a/test/testswmr.sh.in
+++ b/test/testswmr.sh.in
@@ -198,7 +198,7 @@ do
nerrors=`expr $nerrors + 1`
fi
- # Check for error and exit if one occured
+ # Check for error and exit if one occurred
$DPRINT nerrors=$nerrors
if test $nerrors -ne 0 ; then
echo "SWMR tests failed with $nerrors errors."
@@ -258,7 +258,7 @@ do
nerrors=`expr $nerrors + 1`
fi
- # Check for error and exit if one occured
+ # Check for error and exit if one occurred
$DPRINT nerrors=$nerrors
if test $nerrors -ne 0 ; then
echo "SWMR tests failed with $nerrors errors."
@@ -330,7 +330,7 @@ do
nerrors=`expr $nerrors + 1`
fi
- # Check for error and exit if one occured
+ # Check for error and exit if one occurred
$DPRINT nerrors=$nerrors
if test $nerrors -ne 0 ; then
echo "SWMR tests failed with $nerrors errors."
@@ -393,7 +393,7 @@ do
nerrors=`expr $nerrors + 1`
fi
- # Check for error and exit if one occured
+ # Check for error and exit if one occurred
$DPRINT nerrors=$nerrors
if test $nerrors -ne 0 ; then
echo "SWMR tests failed with $nerrors errors."
@@ -474,7 +474,7 @@ do
nerrors=`expr $nerrors + 1`
fi
- # Check for error and exit if one occured
+ # Check for error and exit if one occurred
$DPRINT nerrors=$nerrors
if test $nerrors -ne 0 ; then
echo "SWMR tests failed with $nerrors errors."
@@ -544,7 +544,7 @@ do
fi
done
- # Check for error and exit if one occured
+ # Check for error and exit if one occurred
$DPRINT nerrors=$nerrors
if test $nerrors -ne 0 ; then
echo "SWMR tests failed with $nerrors errors."
diff --git a/test/testvdsswmr.sh.in b/test/testvdsswmr.sh.in
index c57cc08..9ca3689 100644
--- a/test/testvdsswmr.sh.in
+++ b/test/testvdsswmr.sh.in
@@ -166,7 +166,7 @@ if test $? -ne 0; then
nerrors=`expr $nerrors + 1`
fi
-# Check for error and exit if one occured
+# Check for error and exit if one occurred
$DPRINT nerrors=$nerrors
if test $nerrors -ne 0 ; then
echo "VDS SWMR tests failed with $nerrors errors."
@@ -221,7 +221,7 @@ for xpid in $pid_readers; do
fi
done
-# Check for error and exit if one occured
+# Check for error and exit if one occurred
$DPRINT nerrors=$nerrors
if test $nerrors -ne 0 ; then
echo "VDS SWMR tests failed with $nerrors errors."
diff --git a/test/tfile.c b/test/tfile.c
index 47d1b67..146223d 100644
--- a/test/tfile.c
+++ b/test/tfile.c
@@ -214,11 +214,11 @@ static void test_rw_noupdate(void);
static void
test_file_create(void)
{
- hid_t fid1, fid2, fid3; /* HDF5 File IDs */
- hid_t tmpl1, tmpl2; /*file creation templates */
- hsize_t ublock; /*sizeof userblock */
- size_t parm; /*file-creation parameters */
- size_t parm2; /*file-creation parameters */
+ hid_t fid1, fid2, fid3; /* HDF5 File IDs */
+ hid_t tmpl1, tmpl2; /* file creation templates */
+ hsize_t ublock; /* sizeof userblock */
+ size_t parm; /* file-creation parameters */
+ size_t parm2; /* file-creation parameters */
unsigned iparm;
unsigned iparm2;
herr_t ret; /*generic return value */
@@ -229,7 +229,7 @@ test_file_create(void)
/* First ensure the file does not exist */
HDremove(FILE1);
- /* Try opening a non-existant file */
+ /* Try opening a non-existent file */
fid1 = H5Fopen(FILE1, H5F_ACC_RDWR, H5P_DEFAULT);
VERIFY(fid1, FAIL, "H5Fopen");
@@ -1792,7 +1792,7 @@ test_file_is_accessible(const char *env_h5_drvr)
**
** test_file_ishdf5(): low-level file test routine.
** This test checks whether the H5Fis_hdf5() routine is working
-** correctly in variuous situations.
+** correctly in various situations.
**
*****************************************************************/
#ifndef H5_NO_DEPRECATED_SYMBOLS
@@ -2011,7 +2011,7 @@ test_file_delete(hid_t fapl_id)
**
** test_file_open_dot(): low-level file test routine.
** This test checks whether opening objects with "." for a name
-** works correctly in variuous situations.
+** works correctly in various situations.
**
*****************************************************************/
static void
@@ -2393,7 +2393,7 @@ test_file_double_root_open(void)
grp2_id = H5Gopen2(file2_id, "/", H5P_DEFAULT);
CHECK(grp2_id, FAIL, "H5Gopen2");
- /* Note "assymetric" close order */
+ /* Note "asymmetric" close order */
ret = H5Gclose(grp1_id);
CHECK(ret, FAIL, "H5Gclose");
ret = H5Gclose(grp2_id);
@@ -2432,7 +2432,7 @@ test_file_double_group_open(void)
grp2_id = H5Gopen2(file2_id, GRP_NAME, H5P_DEFAULT);
CHECK(grp2_id, FAIL, "H5Gopen2");
- /* Note "assymetric" close order */
+ /* Note "asymmetric" close order */
ret = H5Gclose(grp1_id);
CHECK(ret, FAIL, "H5Gclose");
ret = H5Gclose(grp2_id);
@@ -2481,7 +2481,7 @@ test_file_double_dataset_open(void)
ret = H5Sclose(space_id);
CHECK(ret, FAIL, "H5Sclose");
- /* Note "assymetric" close order */
+ /* Note "asymmetric" close order */
ret = H5Dclose(dset1_id);
CHECK(ret, FAIL, "H5Dclose");
ret = H5Dclose(dset2_id);
@@ -2882,7 +2882,7 @@ test_file_double_datatype_open(void)
type2_id = H5Topen2(file2_id, TYPE_NAME, H5P_DEFAULT);
CHECK(type2_id, FAIL, "H5Topen2");
- /* Note "assymetric" close order */
+ /* Note "asymmetric" close order */
ret = H5Tclose(type1_id);
CHECK(ret, FAIL, "H5Tclose");
ret = H5Tclose(type2_id);
@@ -2932,7 +2932,7 @@ test_userblock_file_size(const char *env_h5_drvr)
ret = H5Pset_userblock(fcpl2_id, USERBLOCK_SIZE);
CHECK(ret, FAIL, "H5Pset_userblock");
- /* Create files. Onyl file2 with have a userblock. */
+ /* Create files. Only file2 with have a userblock. */
file1_id = H5Fcreate(FILE1, H5F_ACC_TRUNC, H5P_DEFAULT, H5P_DEFAULT);
CHECK(file1_id, FAIL, "H5Fcreate");
file2_id = H5Fcreate(FILE2, H5F_ACC_TRUNC, fcpl2_id, H5P_DEFAULT);
@@ -3861,7 +3861,7 @@ test_userblock_alignment_paged(const char *env_h5_drvr)
* Strategy is H5F_FILE_SPACE_NONE; fsp_size = 1024
* H5Pset_alignment() is 16
* Outcome:
- * Should succed:
+ * Should succeed:
* userblock (512) is integral multiple of alignment (16)
*/
/* Create file creation property list with user block */
@@ -3900,7 +3900,7 @@ test_userblock_alignment_paged(const char *env_h5_drvr)
* H5Pset_alignment() is 3
* Reopen the file; H5Pset_alignment() is 1024
* Outcome:
- * Should succed:
+ * Should succeed:
* Userblock (512) is the same as alignment (512);
* The H5Pset_alignment() calls have no effect
*/
@@ -3969,7 +3969,7 @@ test_filespace_info(const char *env_h5_drvr)
hsize_t fs_threshold; /* Free-space section threshold--iteration variable */
hsize_t fsp_size; /* File space page size */
char filename[FILENAME_LEN]; /* Filename to use */
- hbool_t contig_addr_vfd; /* Whether VFD used has a contigous address space */
+ hbool_t contig_addr_vfd; /* Whether VFD used has a contiguous address space */
herr_t ret; /* Return value */
/* Output message about test being performed */
@@ -5211,7 +5211,7 @@ test_filespace_1_10_0_compatible(void)
** Verify that the trunk can open, read and modify these files--
** 1) They are initially created (via gen_filespace.c) in the trunk
** with combinations of file space strategies, default/non-default
-** threshold, and file spacing paging enabled/disbled.
+** threshold, and file spacing paging enabled/disabled.
** The library creates the file space info message with
** "mark if unknown" in these files.
** 2) They are copied to the 1.8 branch, and are opened/read/modified
@@ -6385,7 +6385,7 @@ test_libver_bounds_dataset(hid_t fapl)
VERIFY(dset->shared->dcpl_cache.fill.version, H5O_fill_ver_bounds[low], "H5O_fill_ver_bounds");
}
- /* Verify filter pipleline message version */
+ /* Verify filter pipeline message version */
VERIFY(dset->shared->dcpl_cache.pline.version, H5O_pline_ver_bounds[low], "H5O_pline_ver_bounds");
/* Close the dataset */
@@ -6438,7 +6438,7 @@ test_libver_bounds_dataset(hid_t fapl)
ret = H5Sclose(sid);
CHECK(ret, FAIL, "H5Sclose");
- /* Close the datset creation property list */
+ /* Close the dataset creation property list */
ret = H5Pclose(dcpl);
CHECK(ret, FAIL, "H5Pclose");
@@ -6458,7 +6458,7 @@ test_libver_bounds_dataset(hid_t fapl)
/* Loop through all the combinations of low/high bounds in new_fapl */
/* Open the file with the fapl and create the chunked dataset */
- /* Verify the dataset's layout, fill value and filter pipleline message versions */
+ /* Verify the dataset's layout, fill value and filter pipeline message versions */
for (low = H5F_LIBVER_EARLIEST; low < H5F_LIBVER_NBOUNDS; low++) {
for (high = H5F_LIBVER_EARLIEST; high < H5F_LIBVER_NBOUNDS; high++) {
H5E_BEGIN_TRY
@@ -6563,7 +6563,7 @@ test_libver_bounds_dataset(hid_t fapl)
** new fapl:
** --Open the same file in (a) with the fapl
** --Create a chunked dataset, a compact dataset and
-** a contigous dataset
+** a contiguous dataset
** --Verify the dataspace message version for these
** three datasets
** --Delete the three datasets and the dataspaces
@@ -6912,7 +6912,7 @@ static void
test_libver_bounds_datatype_check(hid_t fapl, hid_t tid)
{
hid_t fid = H5I_INVALID_HID; /* File ID */
- hid_t new_fapl = H5I_INVALID_HID; /* File acess property list */
+ hid_t new_fapl = H5I_INVALID_HID; /* File access property list */
hid_t dcpl = H5I_INVALID_HID; /* Dataset creation property list */
hid_t dtid = H5I_INVALID_HID; /* Datatype ID for the dataset */
hid_t str_tid = H5I_INVALID_HID; /* String datatype ID */
@@ -7172,7 +7172,7 @@ test_libver_bounds_attributes(hid_t fapl)
fid = H5Fcreate(FILE8, H5F_ACC_TRUNC, H5P_DEFAULT, fapl);
CHECK(fid, H5I_INVALID_HID, "H5Fcreate");
- /* Integer datatpye */
+ /* Integer datatype */
tid = H5Tcopy(H5T_NATIVE_INT);
CHECK(tid, H5I_INVALID_HID, "H5Tcopy");
@@ -7281,7 +7281,7 @@ test_libver_bounds_attributes(hid_t fapl)
fid = H5Fcreate(FILE8, H5F_ACC_TRUNC, fcpl, fapl);
CHECK(fid, H5I_INVALID_HID, "H5Fcreate");
- /* Create an integer datatye */
+ /* Create an integer datatype */
tid = H5Tcopy(H5T_NATIVE_INT);
CHECK(tid, H5I_INVALID_HID, "H5Tcopy");
@@ -8092,7 +8092,7 @@ test_file(void)
test_libver_macros(); /* Test the macros for library version comparison */
test_libver_macros2(); /* Test the macros for library version comparison */
test_incr_filesize(); /* Test H5Fincrement_filesize() and H5Fget_eoa() */
- test_min_dset_ohdr(); /* Test datset object header minimization */
+ test_min_dset_ohdr(); /* Test dataset object header minimization */
#ifndef H5_NO_DEPRECATED_SYMBOLS
test_file_ishdf5(env_h5_drvr); /* Test detecting HDF5 files correctly */
test_deprec(env_h5_drvr); /* Test deprecated routines */
diff --git a/test/tgenprop.c b/test/tgenprop.c
index 55e5a9b..f54fbbc 100644
--- a/test/tgenprop.c
+++ b/test/tgenprop.c
@@ -107,7 +107,7 @@ test_genprop_basic_class(void)
ret = H5Pequal(cid2, H5P_ROOT);
VERIFY(ret, 1, "H5Pequal");
- /* Make certain false postives aren't being returned */
+ /* Make certain false positives aren't being returned */
ret = H5Pequal(cid2, H5P_FILE_CREATE);
VERIFY(ret, 0, "H5Pequal");
@@ -185,7 +185,7 @@ test_genprop_basic_class_prop(void)
CHECK_I(ret, "H5Pget_nprops");
VERIFY(nprops, 0, "H5Pget_nprops");
- /* Check the existance of the first property (should fail) */
+ /* Check the existence of the first property (should fail) */
ret = H5Pexist(cid1, PROP1_NAME);
VERIFY(ret, 0, "H5Pexist");
@@ -199,7 +199,7 @@ test_genprop_basic_class_prop(void)
H5Pregister2(cid1, PROP1_NAME, PROP1_SIZE, PROP1_DEF_VALUE, NULL, NULL, NULL, NULL, NULL, NULL, NULL);
VERIFY(ret, FAIL, "H5Pregister2");
- /* Check the existance of the first property */
+ /* Check the existence of the first property */
ret = H5Pexist(cid1, PROP1_NAME);
VERIFY(ret, 1, "H5Pexist");
@@ -223,7 +223,7 @@ test_genprop_basic_class_prop(void)
H5Pregister2(cid1, PROP2_NAME, PROP2_SIZE, PROP2_DEF_VALUE, NULL, NULL, NULL, NULL, NULL, NULL, NULL);
VERIFY(ret, FAIL, "H5Pregister2");
- /* Check the existance of the second property */
+ /* Check the existence of the second property */
ret = H5Pexist(cid1, PROP2_NAME);
VERIFY(ret, 1, "H5Pexist");
@@ -242,7 +242,7 @@ test_genprop_basic_class_prop(void)
H5Pregister2(cid1, PROP3_NAME, PROP3_SIZE, PROP3_DEF_VALUE, NULL, NULL, NULL, NULL, NULL, NULL, NULL);
CHECK_I(ret, "H5Pregister2");
- /* Check the existance of the third property */
+ /* Check the existence of the third property */
ret = H5Pexist(cid1, PROP3_NAME);
VERIFY(ret, 1, "H5Pexist");
@@ -1942,7 +1942,7 @@ test_genprop_deprec_class(void)
CHECK_I(ret, "H5Pget_nprops");
VERIFY(nprops, 0, "H5Pget_nprops");
- /* Check the existance of the first property (should fail) */
+ /* Check the existence of the first property (should fail) */
ret = H5Pexist(cid1, PROP1_NAME);
VERIFY(ret, 0, "H5Pexist");
@@ -1954,7 +1954,7 @@ test_genprop_deprec_class(void)
ret = H5Pregister1(cid1, PROP1_NAME, PROP1_SIZE, PROP1_DEF_VALUE, NULL, NULL, NULL, NULL, NULL, NULL);
VERIFY(ret, FAIL, "H5Pregister1");
- /* Check the existance of the first property */
+ /* Check the existence of the first property */
ret = H5Pexist(cid1, PROP1_NAME);
VERIFY(ret, 1, "H5Pexist");
@@ -1976,7 +1976,7 @@ test_genprop_deprec_class(void)
ret = H5Pregister1(cid1, PROP2_NAME, PROP2_SIZE, PROP2_DEF_VALUE, NULL, NULL, NULL, NULL, NULL, NULL);
VERIFY(ret, FAIL, "H5Pregister1");
- /* Check the existance of the second property */
+ /* Check the existence of the second property */
ret = H5Pexist(cid1, PROP2_NAME);
VERIFY(ret, 1, "H5Pexist");
@@ -1994,7 +1994,7 @@ test_genprop_deprec_class(void)
ret = H5Pregister1(cid1, PROP3_NAME, PROP3_SIZE, PROP3_DEF_VALUE, NULL, NULL, NULL, NULL, NULL, NULL);
CHECK_I(ret, "H5Pregister1");
- /* Check the existance of the third property */
+ /* Check the existence of the third property */
ret = H5Pexist(cid1, PROP3_NAME);
VERIFY(ret, 1, "H5Pexist");
diff --git a/test/th5o.c b/test/th5o.c
index 48c0a81..9504f7a 100644
--- a/test/th5o.c
+++ b/test/th5o.c
@@ -609,7 +609,7 @@ test_h5o_refcount(void)
CHECK(ret, FAIL, "H5Oget_info_by_name3");
VERIFY(oinfo.rc, 1, "reference count in H5Oget_info_by_name3");
- /* Increment the reference counts and then close the file to make sure the increment is permanant */
+ /* Increment the reference counts and then close the file to make sure the increment is permanent */
ret = H5Oincr_refcount(grp);
CHECK(ret, FAIL, "H5Oincr_refcount");
ret = H5Oincr_refcount(dtype);
diff --git a/test/th5s.c b/test/th5s.c
index 97a6728..1bc19d4 100644
--- a/test/th5s.c
+++ b/test/th5s.c
@@ -2022,7 +2022,7 @@ test_h5s_encode_points(H5F_libver_t low, H5F_libver_t high)
{
hid_t fapl = -1; /* File access property list ID */
hid_t sid; /* Dataspace ID */
- hsize_t numparticles = 8388608; /* Used to calculate dimenion size */
+ hsize_t numparticles = 8388608; /* Used to calculate dimension size */
unsigned num_dsets = 513; /* used to calculate dimension size */
hsize_t total_particles = numparticles * num_dsets;
hsize_t vdsdims[1] = {total_particles}; /* Dimension size */
diff --git a/test/tid.c b/test/tid.c
index 0a45116..583004e 100644
--- a/test/tid.c
+++ b/test/tid.c
@@ -41,7 +41,7 @@ basic_id_test(void)
int num_ref;
hsize_t num_members;
- /* Try to register an ID with ficticious types */
+ /* Try to register an ID with fictitious types */
H5E_BEGIN_TRY
arrayID = H5Iregister((H5I_type_t)420, testObj);
H5E_END_TRY
@@ -58,7 +58,7 @@ basic_id_test(void)
if (arrayID != H5I_INVALID_HID)
goto out;
- /* Try to access IDs with ficticious types */
+ /* Try to access IDs with fictitious types */
H5E_BEGIN_TRY
testPtr = H5Iobject_verify((hid_t)100, (H5I_type_t)0);
H5E_END_TRY
diff --git a/test/tmisc.c b/test/tmisc.c
index bb8d8c8..a28e12e 100644
--- a/test/tmisc.c
+++ b/test/tmisc.c
@@ -1280,7 +1280,7 @@ test_misc8(void)
for (v = 0; v < MISC8_DIM1; v++)
*tdata++ = (int)(((u * MISC8_DIM1) + v) % 13);
- /* Create a file acccess property list */
+ /* Create a file access property list */
fapl = H5Pcreate(H5P_FILE_ACCESS);
CHECK(fapl, FAIL, "H5Pcreate");
@@ -3892,7 +3892,7 @@ test_misc21(void)
/****************************************************************
**
-** test_misc22(): Test SZIP bits-per-pixel paramter.
+** test_misc22(): Test SZIP bits-per-pixel parameter.
** This should be set according to the datatype.
** Tests for precision and offset combo's.
**
diff --git a/test/trefer.c b/test/trefer.c
index 48d3670..95e601e 100644
--- a/test/trefer.c
+++ b/test/trefer.c
@@ -2732,7 +2732,7 @@ test_reference_compat_conv(void)
hid_t sid1, sid2, sid3; /* Dataspace IDs */
hid_t tid1, tid2; /* Datatype ID */
hsize_t dims1[] = {SPACE1_DIM1}, dims2[] = {SPACE2_DIM1, SPACE2_DIM2},
- dims3[] = {SPACE1_DIM1}; /* Purposedly set dimension larger to test NULL references */
+ dims3[] = {SPACE1_DIM1}; /* Purposely set dimension larger to test NULL references */
hsize_t start[SPACE2_RANK]; /* Starting location of hyperslab */
hsize_t stride[SPACE2_RANK]; /* Stride of hyperslab */
hsize_t count[SPACE2_RANK]; /* Element count of hyperslab */
diff --git a/test/tselect.c b/test/tselect.c
index 5fec1da..d9b625b 100644
--- a/test/tselect.c
+++ b/test/tselect.c
@@ -1879,7 +1879,7 @@ verify_select_hyper_contig_dr__run_test(const uint16_t *cube_buf, size_t H5_ATTR
uint16_t expected_value; /* Expected value in dataset */
unsigned i, j, k, l, m; /* Local index variables */
size_t s; /* Local index variable */
- hbool_t mis_match; /* Flag to indicate mis-match in expected value */
+ hbool_t mis_match; /* Flag to indicate mismatch in expected value */
HDassert(cube_buf);
HDassert(cube_size > 0);
@@ -2357,7 +2357,7 @@ test_select_hyper_contig_dr__run_test(int test_num, const uint16_t *cube_buf, co
/* Now write the contents of the in memory small cube to slices of
* the on disk cube. After each write, read the on disk cube
- * into memeory, and verify that it contains the expected
+ * into memory, and verify that it contains the expected
* data. Verify that H5Sselect_shape_same() returns true on
* the memory and file selections.
*/
@@ -2742,7 +2742,7 @@ test_select_hyper_checker_board_dr__select_checker_board(hid_t tgt_n_cube_sid, u
i++;
} while ((i <= 1) && (0 >= sel_offset));
- /* Wierdness alert:
+ /* Weirdness alert:
*
* Some how, it seems that selections can extend beyond the
* boundaries of the target dataspace -- hence the following
@@ -3401,7 +3401,7 @@ test_select_hyper_checker_board_dr__run_test(int test_num, const uint16_t *cube_
/* Now write checker board selections of the entries in memory
* small cube to slices of the on disk cube. After each write,
- * read the on disk large cube * into memeory, and verify that
+ * read the on disk large cube * into memory, and verify that
* it contains the expected * data. Verify that
* H5Sselect_shape_same() returns true on the memory and file
* selections.
@@ -5256,7 +5256,7 @@ test_select_hyper_union_stagger(void)
error = H5Fclose(file_id);
CHECK(error, FAIL, "H5Fclose");
- /* Initialize intput buffer */
+ /* Initialize input buffer */
HDmemset(data_out, 0, 7 * 7 * sizeof(int));
/* Open file */
@@ -11251,7 +11251,7 @@ test_shape_same_dr__checkerboard(int test_num, int small_rank, int large_rank, i
CHECK(ret, FAIL, "H5Sselect_hyperslab");
} /* end if */
- /* Wierdness alert:
+ /* Weirdness alert:
*
* Some how, it seems that selections can extend beyond the
* boundaries of the target dataspace -- hence the following
@@ -11346,7 +11346,7 @@ test_shape_same_dr__checkerboard(int test_num, int small_rank, int large_rank, i
CHECK(ret, FAIL, "H5Sselect_hyperslab");
} /* end if */
- /* Wierdness alert:
+ /* Weirdness alert:
*
* Again, it seems that selections can extend beyond the
* boundaries of the target dataspace -- hence the following
@@ -12126,7 +12126,7 @@ test_space_rebuild(void)
ret = FAIL;
CHECK(ret, FAIL, "H5S_hyper_rebuild");
}
- /* No need to do shape comparision */
+ /* No need to do shape comparison */
MESSAGE(7, ("Testing functionality to rebuild 2-D hyperslab selection\n"));
/* Create 2-D dataspace */
@@ -12218,7 +12218,7 @@ test_space_rebuild(void)
ret = FAIL;
CHECK(ret, FAIL, "H5S_hyper_rebuild");
}
- /* No need to do shape comparision */
+ /* No need to do shape comparison */
MESSAGE(7, ("Testing functionality to rebuild 3-D hyperslab selection\n"));
@@ -12320,7 +12320,7 @@ test_space_rebuild(void)
ret = FAIL;
CHECK(ret, FAIL, "H5S_hyper_rebuild");
}
- /* No need to do shape comparision */
+ /* No need to do shape comparison */
MESSAGE(7, ("Testing functionality to rebuild 4-D hyperslab selection\n"));
@@ -12439,7 +12439,7 @@ test_space_rebuild(void)
ret = FAIL;
CHECK(ret, FAIL, "H5S_hyper_rebuild");
}
- /* No need to do shape comparision */
+ /* No need to do shape comparison */
MESSAGE(7, ("Testing functionality to rebuild 5-D hyperslab selection\n"));
@@ -12568,7 +12568,7 @@ test_space_rebuild(void)
ret = FAIL;
CHECK(ret, FAIL, "H5S_hyper_rebuild");
}
- /* No need to do shape comparision */
+ /* No need to do shape comparison */
/* We use 5-D to test a special case with
rebuilding routine TRUE, FALSE and TRUE */
@@ -12615,7 +12615,7 @@ test_space_rebuild(void)
ret = FAIL;
CHECK(ret, FAIL, "H5S_hyper_rebuild");
}
- /* No need to do shape comparision */
+ /* No need to do shape comparison */
/* Adding some selections to make it real irregular */
start5[3] = 1;
@@ -12643,7 +12643,7 @@ test_space_rebuild(void)
ret = FAIL;
CHECK(ret, FAIL, "H5S_hyper_rebuild");
}
- /* No need to do shape comparision */
+ /* No need to do shape comparison */
/* Add more selections to make it regular again */
start5[3] = 5;
@@ -12671,7 +12671,7 @@ test_space_rebuild(void)
ret = FAIL;
CHECK(ret, FAIL, "H5S_hyper_rebuild");
}
- /* No need to do shape comparision */
+ /* No need to do shape comparison */
H5Sclose(sid_reg1);
CHECK(ret, FAIL, "H5Sclose");
@@ -12977,7 +12977,7 @@ test_space_update_diminfo(void)
CHECK(ret, FAIL, "H5S_hyper_update_diminfo");
} /* end if */
- /* Add block parially overlapping first, with OR */
+ /* Add block partially overlapping first, with OR */
start1[0] = 4;
count1[0] = 1;
block1[0] = 2;
@@ -13022,7 +13022,7 @@ test_space_update_diminfo(void)
CHECK(ret, FAIL, "H5S_hyper_update_diminfo");
} /* end if */
- /* Add equally sized block parially overlapping current, with XOR */
+ /* Add equally sized block partially overlapping current, with XOR */
start1[0] = 3;
count1[0] = 1;
block1[0] = 5;
@@ -13056,7 +13056,7 @@ test_space_update_diminfo(void)
CHECK(ret, FAIL, "H5S_hyper_rebuild");
} /* end if */
- /* Add differently sized block parially overlapping current, with XOR */
+ /* Add differently sized block partially overlapping current, with XOR */
start1[0] = 4;
count1[0] = 1;
block1[0] = 5;
@@ -14232,7 +14232,7 @@ test_hyper_regular(void)
**
****************************************************************/
static void
-test_hyper_unlim_check(hid_t sid, hsize_t *dims, hssize_t enpoints, hssize_t enblocks, hsize_t *eblock1,
+test_hyper_unlim_check(hid_t sid, hsize_t *dims, hssize_t endpoints, hssize_t enblocks, hsize_t *eblock1,
hsize_t *eblock2)
{
hid_t lim_sid;
@@ -14257,7 +14257,7 @@ test_hyper_unlim_check(hid_t sid, hsize_t *dims, hssize_t enpoints, hssize_t enb
/* Check number of elements */
npoints = H5Sget_select_npoints(lim_sid);
CHECK(npoints, FAIL, "H5Sget_select_npoints");
- VERIFY(npoints, enpoints, "H5Sget_select_npoints");
+ VERIFY(npoints, endpoints, "H5Sget_select_npoints");
/* Get selection type */
sel_type = H5Sget_select_type(lim_sid);
diff --git a/test/tsohm.c b/test/tsohm.c
index 774378a..04b9522 100644
--- a/test/tsohm.c
+++ b/test/tsohm.c
@@ -636,7 +636,7 @@ size1_helper(hid_t file, const char *filename, hid_t fapl_id, hbool_t test_file_
if (GetTestExpress() > 1)
test_file_closing = FALSE;
- /* Intialize wdata */
+ /* Initialize wdata */
HDmemset(&wdata, 0, sizeof(wdata));
wdata.i1 = 11;
HDstrcpy(wdata.str, "string");
@@ -649,7 +649,7 @@ size1_helper(hid_t file, const char *filename, hid_t fapl_id, hbool_t test_file_
wdata.i8 = 88;
wdata.f1 = 0.0F;
- /* Intialize rdata */
+ /* Initialize rdata */
HDmemset(&rdata, 0, sizeof(rdata));
dtype1_id = make_dtype_1();
@@ -762,7 +762,7 @@ error:
* Function: getsize_testsize1
*
* Purpose: Creates a test file, populates it, and returns its file size.
- * Oject header information from the "first" dataset in the file
+ * Object header information from the "first" dataset in the file
* is stored in pointer `oinfo`.
*
* Programmer: Jacob Smith
@@ -1551,7 +1551,7 @@ size2_dump_struct(const char *name, size2_helper_struct *sizes)
/*-------------------------------------------------------------------------
* Function: size2_helper
*
- * Purpose: A helper functon for test_sohm_size2.
+ * Purpose: A helper function for test_sohm_size2.
*
* Creates a file using the given fcpl, then creates lots
* of different kinds of messages within the file and
@@ -1931,7 +1931,7 @@ size2_helper(hid_t fcpl_id, int test_file_closing, size2_helper_struct *ret_size
/*-------------------------------------------------------------------------
* Function: size2_verify
*
- * Purpose: A helper functon to verify the file created by size2_helper.
+ * Purpose: A helper function to verify the file created by size2_helper.
*
* Runs various tests (not exhaustive) to ensure that the
* file FILENAME actually has the structure that size2_helper
diff --git a/test/ttsafe_attr_vlen.c b/test/ttsafe_attr_vlen.c
index ecd0aa6..9fafaca 100644
--- a/test/ttsafe_attr_vlen.c
+++ b/test/ttsafe_attr_vlen.c
@@ -136,7 +136,7 @@ tts_attr_vlen_thread(void H5_ATTR_UNUSED *client_data)
gid = H5Gopen2(fid, "/", H5P_DEFAULT);
CHECK(gid, H5I_INVALID_HID, "H5Gopen");
- /* Open the attribte */
+ /* Open the attribute */
aid = H5Aopen(gid, "root_attr", H5P_DEFAULT);
CHECK(aid, H5I_INVALID_HID, "H5Aopen");
diff --git a/test/tvltypes.c b/test/tvltypes.c
index 03a8ad3..0c4cf68 100644
--- a/test/tvltypes.c
+++ b/test/tvltypes.c
@@ -131,7 +131,7 @@ test_vltypes_dataset_create(void)
dcpl = H5Pcreate(H5P_DATASET_CREATE);
CHECK(dcpl, FAIL, "H5Pcreate");
- /* Set fill value writting time to be NEVER */
+ /* Set fill value writing time to be NEVER */
ret = H5Pset_fill_time(dcpl, H5D_FILL_TIME_NEVER);
CHECK(ret, FAIL, "H5Pset_fill_time");
diff --git a/test/twriteorder.c b/test/twriteorder.c
index e792487..497542a 100644
--- a/test/twriteorder.c
+++ b/test/twriteorder.c
@@ -54,7 +54,7 @@
*
* If the system, in which the writer and reader processes run, the readers
* will always get all chain-linked blocks correctly. If the order of write
- * is not maintained, some reader processes may found unexpect block data.
+ * is not maintained, some reader processes may found unexpected block data.
*
*************************************************************/
diff --git a/test/unlink.c b/test/unlink.c
index 94bd035..19b7b37 100644
--- a/test/unlink.c
+++ b/test/unlink.c
@@ -617,7 +617,7 @@ test_filespace(hid_t fapl)
/* Create common objects for datasets */
- /* Create dataset creation property list for contigous storage */
+ /* Create dataset creation property list for contiguous storage */
if ((contig_dcpl = H5Pcreate(H5P_DATASET_CREATE)) < 0)
TEST_ERROR
@@ -1099,7 +1099,7 @@ test_filespace(hid_t fapl)
PASSED();
- /* Create simple group hiearchy, remove it & verify file size */
+ /* Create simple group hierarchy, remove it & verify file size */
TESTING(" simple group hierarchy");
/* Create file */
@@ -1138,7 +1138,7 @@ test_filespace(hid_t fapl)
PASSED();
- /* Create complex group hiearchy, remove it & verify file size */
+ /* Create complex group hierarchy, remove it & verify file size */
TESTING(" complex group hierarchy");
/* Create file */
diff --git a/test/use_common.c b/test/use_common.c
index 8f78f3b..7b5df53 100644
--- a/test/use_common.c
+++ b/test/use_common.c
@@ -298,7 +298,7 @@ write_uc_file(hbool_t tosend, hid_t file_id, options_t *opts)
return -1;
}
- /* verify chunk_dims against set paramenters */
+ /* verify chunk_dims against set parameters */
if (chunk_dims[0] != opts->chunkdims[0] || chunk_dims[1] != cz || chunk_dims[2] != cz) {
HDfprintf(stderr, "chunk size is not as expected. Got dims=(%llu,%llu,%llu)\n",
(unsigned long long)chunk_dims[0], (unsigned long long)chunk_dims[1],
@@ -533,7 +533,7 @@ read_uc_file(hbool_t towait, options_t *opts)
/* quit when all nplanes have been read */
loops_waiting_for_plane = 0;
while (nplanes_seen < opts->nplanes) {
- /* print progress message according to if new planes are availalbe */
+ /* print progress message according to if new planes are available */
if (nplanes_seen < dims[0]) {
if (loops_waiting_for_plane) {
/* end the previous message */
diff --git a/test/use_disable_mdc_flushes.c b/test/use_disable_mdc_flushes.c
index f28e574..d40e56d 100644
--- a/test/use_disable_mdc_flushes.c
+++ b/test/use_disable_mdc_flushes.c
@@ -360,7 +360,7 @@ write_file(void)
return -1;
}
- /* verify chunk_dims against set paramenters */
+ /* verify chunk_dims against set parameters */
if (chunk_dims[0] != chunkdims_g[0] || chunk_dims[1] != cz || chunk_dims[2] != cz) {
HDfprintf(stderr, "chunk size is not as expected. Got dims=(%llu,%llu,%llu)\n",
(unsigned long long)chunk_dims[0], (unsigned long long)chunk_dims[1],
diff --git a/test/vds.c b/test/vds.c
index 61b974b..4455ad4 100644
--- a/test/vds.c
+++ b/test/vds.c
@@ -196,7 +196,7 @@ vds_select_equal(hid_t space1, hid_t space2)
if (nblocks1 != nblocks2)
return FALSE;
- /* Allocate block lists. Do not return directly afer
+ /* Allocate block lists. Do not return directly after
* allocating, to make sure buffers are freed. */
if (NULL ==
(buf1 = (hsize_t *)HDmalloc((size_t)2 * (size_t)rank1 * (size_t)nblocks1 * sizeof(*buf1))))
@@ -479,7 +479,7 @@ test_api_get_ex_dcpl(test_api_config_t config, hid_t fapl, hid_t dcpl, hid_t *ex
/* Verify examination DCPL is equal to original DCPL. Do not compare the
* plist to itself, and do not do the comparison if we reopened the file,
- * because in that case the extent of the source dset will not be corrent.
+ * because in that case the extent of the source dset will not be current.
*/
if ((*ex_dcpl != dcpl) && (config != TEST_API_REOPEN_FILE)) {
if ((tri_ret = H5Pequal(dcpl, *ex_dcpl)) < 0)
@@ -1044,7 +1044,7 @@ test_api(test_api_config_t config, hid_t fapl, H5F_libver_t low)
if ((vspace[i] = H5Screate_simple(2, dims, NULL)) < 0)
TEST_ERROR
- /* Select row in virual dataspace */
+ /* Select row in virtual dataspace */
start[0] = (hsize_t)i;
if (H5Sselect_hyperslab(vspace[i], H5S_SELECT_SET, start, NULL, count, block) < 0)
TEST_ERROR
@@ -1182,7 +1182,7 @@ test_vds_prefix_first(unsigned config, hid_t vds_fapl, hid_t src_fapl)
hid_t srcspace[4] = {-1, -1, -1, -1}; /* Source dataspaces */
hid_t vspace[4] = {-1, -1, -1, -1}; /* Virtual dset dataspaces */
hid_t memspace = -1; /* Memory dataspace */
- hid_t srcdset[4] = {-1, -1, -1, -1}; /* Source datsets */
+ hid_t srcdset[4] = {-1, -1, -1, -1}; /* Source datasets */
hid_t vdset = -1; /* Virtual dataset */
hsize_t dims[4] = {10, 26, 0, 0}; /* Data space current size */
int buf[10][26]; /* Write and expected read buffer */
@@ -1461,7 +1461,7 @@ test_basic_io(unsigned config, hid_t vds_fapl, hid_t src_fapl)
hid_t srcspace[4] = {-1, -1, -1, -1}; /* Source dataspaces */
hid_t vspace[4] = {-1, -1, -1, -1}; /* Virtual dset dataspaces */
hid_t memspace = -1; /* Memory dataspace */
- hid_t srcdset[4] = {-1, -1, -1, -1}; /* Source datsets */
+ hid_t srcdset[4] = {-1, -1, -1, -1}; /* Source datasets */
hid_t vdset = -1; /* Virtual dataset */
hsize_t dims[4] = {10, 26, 0, 0}; /* Data space current size */
hsize_t start[4]; /* Hyperslab start */
@@ -4389,7 +4389,7 @@ test_unlim(unsigned config, hid_t vds_fapl, hid_t src_fapl)
hid_t vspace[4] = {-1, -1, -1, -1}; /* Virtual dset dataspaces */
hid_t memspace = -1; /* Memory dataspace */
hid_t filespace = -1; /* File dataspace */
- hid_t srcdset[4] = {-1, -1, -1, -1}; /* Source datsets */
+ hid_t srcdset[4] = {-1, -1, -1, -1}; /* Source datasets */
hid_t vdset = -1; /* Virtual dataset */
hsize_t dims[2] = {10, 10}; /* Data space current size */
hsize_t mdims[2] = {10, 20}; /* Data space maximum size */
@@ -7403,7 +7403,7 @@ test_printf(unsigned config, hid_t vds_fapl, hid_t src_fapl)
hid_t vspace[2] = {-1, -1}; /* Virtual dset dataspaces */
hid_t memspace = -1; /* Memory dataspace */
hid_t filespace = -1; /* File dataspace */
- hid_t srcdset[6] = {-1, -1, -1, -1, -1, -1}; /* Source datsets */
+ hid_t srcdset[6] = {-1, -1, -1, -1, -1, -1}; /* Source datasets */
hid_t vdset = -1; /* Virtual dataset */
hsize_t dims[2] = {10, 0}; /* Data space current size */
hsize_t mdims[2] = {10, 20}; /* Data space maximum size */
@@ -11111,7 +11111,7 @@ test_all(unsigned config, hid_t vds_fapl, hid_t src_fapl)
hid_t vspace[3] = {-1, -1, -1}; /* Virtual dset dataspaces */
hid_t memspace = -1; /* Memory dataspace */
hid_t filespace = -1; /* File dataspace */
- hid_t srcdset[5] = {-1, -1, -1, -1, -1}; /* Source datsets */
+ hid_t srcdset[5] = {-1, -1, -1, -1, -1}; /* Source datasets */
hid_t vdset = -1; /* Virtual dataset */
hsize_t dims[2] = {6, 6}; /* Data space current size */
hsize_t mdims[2] = {10, 10}; /* Data space maximum size */
diff --git a/test/vds_env.c b/test/vds_env.c
index 5020db6..17c3876 100644
--- a/test/vds_env.c
+++ b/test/vds_env.c
@@ -58,7 +58,7 @@ test_vds_prefix_second(unsigned config, hid_t fapl)
hid_t srcspace[4] = {-1, -1, -1, -1}; /* Source dataspaces */
hid_t vspace[4] = {-1, -1, -1, -1}; /* Virtual dset dataspaces */
hid_t memspace = -1; /* Memory dataspace */
- hid_t srcdset[4] = {-1, -1, -1, -1}; /* Source datsets */
+ hid_t srcdset[4] = {-1, -1, -1, -1}; /* Source datasets */
hid_t vdset = -1; /* Virtual dataset */
hsize_t dims[4] = {10, 26, 0, 0}; /* Data space current size */
int buf[10][26]; /* Write and expected read buffer */
diff --git a/test/vfd.c b/test/vfd.c
index ea526b1..4f28766 100644
--- a/test/vfd.c
+++ b/test/vfd.c
@@ -208,7 +208,7 @@ test_sec2(void)
if (os_file_handle == NULL)
FAIL_PUTS_ERROR("NULL os-specific vfd/file handle was returned from H5Fget_vfd_handle");
- /* There is no garantee the size of metadata in file is constant.
+ /* There is no guarantee the size of metadata in file is constant.
* Just try to check if it's reasonable.
*
* Currently it should be around 2 KB.
@@ -404,7 +404,7 @@ test_core(void)
if (os_file_handle == NULL)
FAIL_PUTS_ERROR("NULL os-specific vfd/file handle was returned from H5Fget_vfd_handle");
- /* There is no garantee the size of metadata in file is constant.
+ /* There is no guarantee the size of metadata in file is constant.
* Just try to check if it's reasonable.
*
* TODO: Needs justification of why is this is a reasonable size.
@@ -551,7 +551,7 @@ test_core(void)
} /* end if */
/* Check file size API.
- * There is no garantee the size of metadata in file is constant.
+ * There is no guarantee the size of metadata in file is constant.
* Just try to check if it's reasonable.
*
* TODO: Needs justification of why is this is a reasonable size.
@@ -1155,7 +1155,7 @@ error:
* See if we can open files created with v1.6 library.
* The source file was created by the test/file_handle.c
* of the v1.6 library. Then tools/misc/h5repart.c was
- * used to concantenated. The command was "h5repart -m 5k
+ * used to concatenated. The command was "h5repart -m 5k
* family_file%05d.h5 family_v16_%05d.h5".
*
* Return: Success: 0
@@ -1423,7 +1423,7 @@ H5_GCC_CLANG_DIAG_ON("format-nonliteral")
/*-------------------------------------------------------------------------
* Function: test_multi
*
- * Purpose: Tests the file handle interface for MUTLI driver
+ * Purpose: Tests the file handle interface for MULTI driver
*
* Return: SUCCEED/FAIL
*
@@ -2578,7 +2578,7 @@ done:
* Function: driver_is_splitter_compatible
*
* Purpose: Determine whether the driver set in the FAPL ID is compatible
- * with the Splitter VFD -- specificially, Write-Only channel.
+ * with the Splitter VFD -- specifically, Write-Only channel.
*
* Return: Success: 0
* Failure: -1
@@ -3151,7 +3151,7 @@ splitter_tentative_open_test(hid_t child_fapl_id)
SPLITTER_TEST_FAULT("can't close file ID\n");
}
if (!file_exists(filename_rw, child_fapl_id)) {
- SPLITTER_TEST_FAULT("R/W file mysteriously disappared\n");
+ SPLITTER_TEST_FAULT("R/W file mysteriously disappeared\n");
}
if (!file_exists(vfd_config->wo_path, child_fapl_id)) {
SPLITTER_TEST_FAULT("W/O file mysteriously disappeared\n");
@@ -3171,7 +3171,7 @@ splitter_tentative_open_test(hid_t child_fapl_id)
SPLITTER_TEST_FAULT("can't close file ID\n");
}
if (!file_exists(filename_rw, child_fapl_id)) {
- SPLITTER_TEST_FAULT("R/W file mysteriously disappared\n");
+ SPLITTER_TEST_FAULT("R/W file mysteriously disappeared\n");
}
if (!file_exists(vfd_config->wo_path, child_fapl_id)) {
SPLITTER_TEST_FAULT("W/O file mysteriously disappeared\n");
@@ -3202,7 +3202,7 @@ splitter_tentative_open_test(hid_t child_fapl_id)
SPLITTER_TEST_FAULT("can't close file ID\n");
}
if (!file_exists(filename_rw, child_fapl_id)) {
- SPLITTER_TEST_FAULT("R/W file mysteriously disappared\n");
+ SPLITTER_TEST_FAULT("R/W file mysteriously disappeared\n");
}
if (!file_exists(vfd_config->wo_path, child_fapl_id)) {
SPLITTER_TEST_FAULT("W/O file mysteriously disappeared\n");
@@ -3233,7 +3233,7 @@ splitter_tentative_open_test(hid_t child_fapl_id)
SPLITTER_TEST_FAULT("can't close file ID\n");
}
if (!file_exists(filename_rw, child_fapl_id)) {
- SPLITTER_TEST_FAULT("R/W file mysteriously disappared\n");
+ SPLITTER_TEST_FAULT("R/W file mysteriously disappeared\n");
}
if (!file_exists(vfd_config->wo_path, child_fapl_id)) {
SPLITTER_TEST_FAULT("W/O file mysteriously disappeared\n");
diff --git a/testpar/t_2Gio.c b/testpar/t_2Gio.c
index d5b9038..ac031fc 100644
--- a/testpar/t_2Gio.c
+++ b/testpar/t_2Gio.c
@@ -33,7 +33,7 @@
#include "mpi.h"
/* For this test, we don't want to inherit the RANK definition
- * from testphdf5.h. We'll define MAX_RANK to accomodate 3D arrays
+ * from testphdf5.h. We'll define MAX_RANK to accommodate 3D arrays
* and use that definition rather than RANK.
*/
#ifndef MAX_RANK
@@ -3323,7 +3323,7 @@ none_selection_chunk(void)
* Simple independent I/O. This tests that the defaults are properly set.
*
* TEST_ACTUAL_IO_RESET:
- * Perfroms collective and then independent I/O wit hthe same dxpl to
+ * Performs collective and then independent I/O with hthe same dxpl to
* make sure the peroperty is correctly reset to the default on each use.
* Specifically, this test runs TEST_ACTUAL_IO_MULTI_CHUNK_NO_OPT_MIX_DISAGREE
* (The most complex case that works on all builds) and then performs
@@ -3332,7 +3332,7 @@ none_selection_chunk(void)
* Note: DIRECT_MULTI_CHUNK_MIX and DIRECT_MULTI_CHUNK_MIX_DISAGREE
* is not needed as they are covered by DIRECT_CHUNK_MIX and
* MULTI_CHUNK_MIX_DISAGREE cases. _DIRECT_ cases are only for testing
- * path way to multi-chunk-io by H5FD_MPIO_CHUNK_MULTI_IO insted of num-threshold.
+ * path way to multi-chunk-io by H5FD_MPIO_CHUNK_MULTI_IO instead of num-threshold.
*
* Modification:
* - Refctore to remove multi-chunk-without-opimization test and update for
@@ -3520,7 +3520,7 @@ test_actual_io_mode(int selection_mode)
break;
/* RESET tests that the properties are properly reset to defaults each time I/O is
- * performed. To acheive this, we have RESET perform collective I/O (which would change
+ * performed. To achieve this, we have RESET perform collective I/O (which would change
* the values from the defaults) followed by independent I/O (which should report the
* default values). RESET doesn't need to have a unique selection, so we reuse
* MULTI_CHUMK_MIX_DISAGREE, which was chosen because it is a complex case that works
@@ -3536,7 +3536,7 @@ test_actual_io_mode(int selection_mode)
* assgigned collective I/O, while each other process gets independent I/O.
* Since the root process with only access the first chunk, it will report
* collective I/O. The subsequent processes will access the first chunk
- * collectively, and their other chunk indpendently, reporting mixed I/O.
+ * collectively, and their other chunk independently, reporting mixed I/O.
*/
if (mpi_rank == 0) {
@@ -3651,7 +3651,7 @@ test_actual_io_mode(int selection_mode)
ret = H5Pset_dxpl_mpio_chunk_opt_num(dxpl_write, (unsigned)mpi_size * 2);
VRFY((ret >= 0), "H5Pset_dxpl_mpio_chunk_opt_num succeeded");
- /* set this to manipulate testing senario about allocating processes
+ /* set this to manipulate testing scenario about allocating processes
* to chunks */
ret = H5Pset_dxpl_mpio_chunk_opt_ratio(dxpl_write, (unsigned)99);
VRFY((ret >= 0), "H5Pset_dxpl_mpio_chunk_opt_ratio succeeded");
@@ -3675,12 +3675,12 @@ test_actual_io_mode(int selection_mode)
H5Eprint2(H5E_DEFAULT, stdout);
VRFY((ret >= 0), "H5Dwrite() dataset multichunk write succeeded");
- /* Retreive Actual io valuess */
+ /* Retrieve Actual io values */
ret = H5Pget_mpio_actual_io_mode(dxpl_write, &actual_io_mode_write);
- VRFY((ret >= 0), "retriving actual io mode suceeded");
+ VRFY((ret >= 0), "retrieving actual io mode succeeded");
ret = H5Pget_mpio_actual_chunk_opt_mode(dxpl_write, &actual_chunk_opt_mode_write);
- VRFY((ret >= 0), "retriving actual chunk opt mode succeeded");
+ VRFY((ret >= 0), "retrieving actual chunk opt mode succeeded");
/* Read */
ret = H5Dread(dataset, data_type, mem_space, file_space, dxpl_read, buffer);
@@ -3688,12 +3688,12 @@ test_actual_io_mode(int selection_mode)
H5Eprint2(H5E_DEFAULT, stdout);
VRFY((ret >= 0), "H5Dread() dataset multichunk read succeeded");
- /* Retreive Actual io values */
+ /* Retrieve Actual io values */
ret = H5Pget_mpio_actual_io_mode(dxpl_read, &actual_io_mode_read);
- VRFY((ret >= 0), "retriving actual io mode succeeded");
+ VRFY((ret >= 0), "retrieving actual io mode succeeded");
ret = H5Pget_mpio_actual_chunk_opt_mode(dxpl_read, &actual_chunk_opt_mode_read);
- VRFY((ret >= 0), "retriving actual chunk opt mode succeeded");
+ VRFY((ret >= 0), "retrieving actual chunk opt mode succeeded");
/* Check write vs read */
VRFY((actual_io_mode_read == actual_io_mode_write),
@@ -3714,7 +3714,7 @@ test_actual_io_mode(int selection_mode)
actual_io_mode_write);
}
- /* To test that the property is succesfully reset to the default, we perform some
+ /* To test that the property is successfully reset to the default, we perform some
* independent I/O after the collective I/O
*/
if (selection_mode == TEST_ACTUAL_IO_RESET) {
@@ -3731,9 +3731,9 @@ test_actual_io_mode(int selection_mode)
/* Check Properties */
ret = H5Pget_mpio_actual_io_mode(dxpl_write, &actual_io_mode_write);
- VRFY((ret >= 0), "retriving actual io mode succeeded");
+ VRFY((ret >= 0), "retrieving actual io mode succeeded");
ret = H5Pget_mpio_actual_chunk_opt_mode(dxpl_write, &actual_chunk_opt_mode_write);
- VRFY((ret >= 0), "retriving actual chunk opt mode succeeded");
+ VRFY((ret >= 0), "retrieving actual chunk opt mode succeeded");
VRFY(actual_chunk_opt_mode_write == H5D_MPIO_NO_CHUNK_OPTIMIZATION,
"actual_chunk_opt_mode has correct value for reset write (independent)");
@@ -3746,9 +3746,9 @@ test_actual_io_mode(int selection_mode)
/* Check Properties */
ret = H5Pget_mpio_actual_io_mode(dxpl_read, &actual_io_mode_read);
- VRFY((ret >= 0), "retriving actual io mode succeeded");
+ VRFY((ret >= 0), "retrieving actual io mode succeeded");
ret = H5Pget_mpio_actual_chunk_opt_mode(dxpl_read, &actual_chunk_opt_mode_read);
- VRFY((ret >= 0), "retriving actual chunk opt mode succeeded");
+ VRFY((ret >= 0), "retrieving actual chunk opt mode succeeded");
VRFY(actual_chunk_opt_mode_read == H5D_MPIO_NO_CHUNK_OPTIMIZATION,
"actual_chunk_opt_mode has correct value for reset read (independent)");
@@ -3794,11 +3794,11 @@ actual_io_mode_tests(void)
test_actual_io_mode(TEST_ACTUAL_IO_MULTI_CHUNK_IND);
test_actual_io_mode(TEST_ACTUAL_IO_MULTI_CHUNK_COL);
- /* The Multi Chunk Mixed test requires atleast three processes. */
+ /* The Multi Chunk Mixed test requires at least three processes. */
if (mpi_size > 2)
test_actual_io_mode(TEST_ACTUAL_IO_MULTI_CHUNK_MIX);
else
- HDfprintf(stdout, "Multi Chunk Mixed test requires 3 proceses minimum\n");
+ HDfprintf(stdout, "Multi Chunk Mixed test requires 3 processes minimum\n");
test_actual_io_mode(TEST_ACTUAL_IO_MULTI_CHUNK_MIX_DISAGREE);
@@ -3842,7 +3842,7 @@ actual_io_mode_tests(void)
* Test for Data Type Conversion as the cause of breaking collective I/O.
*
* TEST_DATA_TRANSFORMS:
- * Test for Data Transfrom feature as the cause of breaking collective I/O.
+ * Test for Data Transform feature as the cause of breaking collective I/O.
*
* TEST_NOT_SIMPLE_OR_SCALAR_DATASPACES:
* Test for NULL dataspace as the cause of breaking collective I/O.
@@ -4004,7 +4004,7 @@ test_no_collective_cause_mode(int selection_mode)
}
if (selection_mode & TEST_DATA_TRANSFORMS) {
- test_name = "Broken Collective I/O - DATA Transfroms";
+ test_name = "Broken Collective I/O - DATA Transforms";
no_collective_cause_local_expected |= H5D_MPIO_DATA_TRANSFORMS;
no_collective_cause_global_expected |= H5D_MPIO_DATA_TRANSFORMS;
}
@@ -4102,7 +4102,7 @@ test_no_collective_cause_mode(int selection_mode)
/* Get the cause of broken collective I/O */
ret = H5Pget_mpio_no_collective_cause(dxpl_write, &no_collective_cause_local_write,
&no_collective_cause_global_write);
- VRFY((ret >= 0), "retriving no collective cause succeeded");
+ VRFY((ret >= 0), "retrieving no collective cause succeeded");
/*---------------------
* Test Read access
@@ -4122,7 +4122,7 @@ test_no_collective_cause_mode(int selection_mode)
/* Get the cause of broken collective I/O */
ret = H5Pget_mpio_no_collective_cause(dxpl_read, &no_collective_cause_local_read,
&no_collective_cause_global_read);
- VRFY((ret >= 0), "retriving no collective cause succeeded");
+ VRFY((ret >= 0), "retrieving no collective cause succeeded");
/* Check write vs read */
VRFY((no_collective_cause_local_read == no_collective_cause_local_write),
@@ -4368,7 +4368,7 @@ test_no_collective_cause_mode_filter(int selection_mode)
/* Get the cause of broken collective I/O */
ret = H5Pget_mpio_no_collective_cause (dxpl, &no_collective_cause_local_read, &no_collective_cause_global_read);
- VRFY((ret >= 0), "retriving no collective cause succeeded" );
+ VRFY((ret >= 0), "retrieving no collective cause succeeded" );
/* Test values */
HDmemset (message, 0, sizeof (message));
@@ -4547,7 +4547,7 @@ dataset_atomicity(void)
/* file locking allows only one file open (serial) for writing */
if (MAINPROCESS) {
fid = H5Fopen(filename, H5F_ACC_RDWR, H5P_DEFAULT);
- VRFY((fid >= 0), "H5Fopen succeeed");
+ VRFY((fid >= 0), "H5Fopen succeeded");
}
/* should fail */
diff --git a/testpar/t_bigio.c b/testpar/t_bigio.c
index e5654f4..4df624b 100644
--- a/testpar/t_bigio.c
+++ b/testpar/t_bigio.c
@@ -1331,7 +1331,7 @@ coll_chunk2(void)
*/
/* ------------------------------------------------------------------------
- * Descriptions for the selection: one singular selection accross many chunks
+ * Descriptions for the selection: one singular selection across many chunks
* Two dimensions, Num of chunks = 2* mpi_size
*
* dim1 = space_dim1*mpi_size
@@ -1840,7 +1840,7 @@ main(int argc, char **argv)
/* Set the bigio processing limit to be 'newsize' bytes */
hsize_t oldsize = H5_mpi_set_bigio_count(newsize);
- /* Having set the bigio handling to a size that is managable,
+ /* Having set the bigio handling to a size that is manageable,
* we'll set our 'bigcount' variable to be 2X that limit so
* that we try to ensure that our bigio handling is actually
* envoked and tested.
diff --git a/testpar/t_cache.c b/testpar/t_cache.c
index 8696092..8c96756 100644
--- a/testpar/t_cache.c
+++ b/testpar/t_cache.c
@@ -126,7 +126,7 @@ int total_writes = 0;
*
* local_pinned: Boolean flag that is set to true iff the entry
* has been pinned in the local cache, but probably not all
- * caches. Such pins will typically not be consistant across
+ * caches. Such pins will typically not be consistent across
* processes, and thus cannot be marked as dirty unless they
* happen to overlap some collective operation.
*
@@ -205,7 +205,7 @@ struct datum data[NUM_DATA_ENTRIES];
* even divisor of NUM_DATA_ENTRIES. So far, all tests have been with
* powers of 10 that meet these criteria.
*
- * Further, this value must be consistant across all processes.
+ * Further, this value must be consistent across all processes.
*/
#define STD_VIRT_NUM_DATA_ENTRIES NUM_DATA_ENTRIES
@@ -252,7 +252,7 @@ int data_index[NUM_DATA_ENTRIES];
*
* Strangely, at least on Phoenix, the first solution runs faster by a
* rather large margin. However, I can imagine this changing with
- * different OS's and MPI implementatins.
+ * different OS's and MPI implementations.
*
* Thus I have left code supporting the second solution in place.
*
@@ -1282,7 +1282,7 @@ reset_server_counters(void)
* Function: server_main()
*
* Purpose: Main function for the server process. This process exists
- * to provide an independant view of the data array.
+ * to provide an independent view of the data array.
*
* The function handles request from the other processes in
* the test until the count of done messages received equals
@@ -3962,7 +3962,7 @@ setup_cache_for_test(hid_t *fid_ptr, H5F_t **file_ptr_ptr, H5C_t **cache_ptr_ptr
* Purpose: Verify that the indicated entries have been written exactly
* once each, and that the indicated total number of writes
* has been processed by the server process. Flag an error if
- * discrepency is noted. Finally reset the counters maintained
+ * discrepancy is noted. Finally reset the counters maintained
* by the server process.
*
* This function should only be called by the metadata cache
@@ -4050,7 +4050,7 @@ verify_writes(unsigned num_writes, haddr_t *written_entries_tbl)
/* final barrier to ensure that all processes think that the server
* counters have been reset before we leave the sync point. This
- * barrier is probaby not necessary at this point in time (5/9/10),
+ * barrier is probably not necessary at this point in time (5/9/10),
* but I can think of at least one likely change to the metadata write
* strategies that will require it -- hence its insertion now.
*/
@@ -4794,7 +4794,7 @@ server_smoke_check(void)
if (!server_main()) {
- /* some error occured in the server -- report failure */
+ /* some error occurred in the server -- report failure */
nerrors++;
if (verbose) {
HDfprintf(stdout, "%d:%s: server_main() failed.\n", world_mpi_rank, __func__);
@@ -5141,7 +5141,7 @@ smoke_check_1(int metadata_write_strategy)
if (!server_main()) {
- /* some error occured in the server -- report failure */
+ /* some error occurred in the server -- report failure */
nerrors++;
if (verbose) {
HDfprintf(stdout, "%d:%s: server_main() failed.\n", world_mpi_rank, __func__);
@@ -5305,7 +5305,7 @@ smoke_check_2(int metadata_write_strategy)
if (!server_main()) {
- /* some error occured in the server -- report failure */
+ /* some error occurred in the server -- report failure */
nerrors++;
if (verbose) {
HDfprintf(stdout, "%d:%s: server_main() failed.\n", world_mpi_rank, __func__);
@@ -5453,7 +5453,7 @@ smoke_check_2(int metadata_write_strategy)
*
* Purpose: Third smoke check for the parallel cache.
*
- * Use random reads to vary the loads on the diffferent
+ * Use random reads to vary the loads on the different
* processors. Also force different cache size adjustments.
*
* In this test, load process 0 heavily, and the other
@@ -5510,7 +5510,7 @@ smoke_check_3(int metadata_write_strategy)
if (!server_main()) {
- /* some error occured in the server -- report failure */
+ /* some error occurred in the server -- report failure */
nerrors++;
if (verbose) {
HDfprintf(stdout, "%d:%s: server_main() failed.\n", world_mpi_rank, __func__);
@@ -5738,7 +5738,7 @@ smoke_check_3(int metadata_write_strategy)
*
* Purpose: Fourth smoke check for the parallel cache.
*
- * Use random reads to vary the loads on the diffferent
+ * Use random reads to vary the loads on the different
* processors. Also force different cache size adjustments.
*
* In this test, load process 0 lightly, and the other
@@ -5795,7 +5795,7 @@ smoke_check_4(int metadata_write_strategy)
if (!server_main()) {
- /* some error occured in the server -- report failure */
+ /* some error occurred in the server -- report failure */
nerrors++;
if (verbose) {
HDfprintf(stdout, "%d:%s: server_main() failed.\n", world_mpi_rank, __func__);
@@ -6070,7 +6070,7 @@ smoke_check_5(int metadata_write_strategy)
if (!server_main()) {
- /* some error occured in the server -- report failure */
+ /* some error occurred in the server -- report failure */
nerrors++;
if (verbose) {
HDfprintf(stdout, "%d:%s: server_main() failed.\n", world_mpi_rank, __func__);
@@ -6355,7 +6355,7 @@ trace_file_check(int metadata_write_strategy)
if (!server_main()) {
- /* some error occured in the server -- report failure */
+ /* some error occurred in the server -- report failure */
nerrors++;
if (verbose)
HDfprintf(stdout, "%d:%s: server_main() failed.\n", world_mpi_rank, __func__);
@@ -6660,7 +6660,7 @@ smoke_check_6(int metadata_write_strategy)
if (!server_main()) {
- /* some error occured in the server -- report failure */
+ /* some error occurred in the server -- report failure */
nerrors++;
if (verbose) {
HDfprintf(stdout, "%d:%s: server_main() failed.\n", world_mpi_rank, __func__);
diff --git a/testpar/t_cache_image.c b/testpar/t_cache_image.c
index 17fb992..feb4325 100644
--- a/testpar/t_cache_image.c
+++ b/testpar/t_cache_image.c
@@ -449,7 +449,7 @@ create_data_sets(hid_t file_id, int min_dset, int max_dset)
}
/* set the dataset creation plist to specify that the raw data is
- * to be partioned into 10X10 element chunks.
+ * to be partitioned into 10X10 element chunks.
*/
if (pass) {
@@ -795,7 +795,7 @@ delete_data_sets(hid_t file_id, int min_dset, int max_dset)
* Set pass to FALSE and issue a suitable failure
* message if either the file contains a metadata cache image
* superblock extension and mdci_sbem_expected is TRUE, or
- * vise versa.
+ * vice versa.
*
* If mdci_sbem_expected is TRUE, also verify that the metadata
* cache has been advised of this.
@@ -807,7 +807,7 @@ delete_data_sets(hid_t file_id, int min_dset, int max_dset)
* FAPL entry when opening the file, and verify that the
* metadata cache is notified.
*
- * If config_fsm is TRUE, setup the persistant free space
+ * If config_fsm is TRUE, setup the persistent free space
* manager. Note that this flag may only be set if
* create_file is also TRUE.
*
@@ -884,7 +884,7 @@ open_hdf5_file(const hbool_t create_file, const hbool_t mdci_sbem_expected, cons
if (show_progress)
HDfprintf(stdout, "%s: cp = %d, pass = %d.\n", fcn_name, cp++, pass);
- /* create a file access propertly list. */
+ /* create a file access property list. */
if (pass) {
fapl_id = H5Pcreate(H5P_FILE_ACCESS);
@@ -956,7 +956,7 @@ open_hdf5_file(const hbool_t create_file, const hbool_t mdci_sbem_expected, cons
if (show_progress)
HDfprintf(stdout, "%s: cp = %d, pass = %d.\n", fcn_name, cp++, pass);
- /* setup the persistant free space manager if indicated */
+ /* setup the persistent free space manager if indicated */
if ((pass) && (config_fsm)) {
fcpl_id = H5Pcreate(H5P_FILE_CREATE);
@@ -1138,13 +1138,13 @@ open_hdf5_file(const hbool_t create_file, const hbool_t mdci_sbem_expected, cons
if ((file_ptr->shared->page_buf) && ((!enable_page_buffer) || (l_facc_type == FACC_MPIO))) {
pass = FALSE;
- failure_mssg = "page buffer unexepectedly enabled.";
+ failure_mssg = "page buffer unexpectedly enabled.";
}
else if ((file_ptr->shared->page_buf != NULL) &&
((enable_page_buffer) || (l_facc_type != FACC_MPIO))) {
pass = FALSE;
- failure_mssg = "page buffer unexepectedly disabled.";
+ failure_mssg = "page buffer unexpectedly disabled.";
}
}
@@ -1360,7 +1360,7 @@ par_create_dataset(int dset_num, hid_t file_id, int mpi_rank, int mpi_size)
HDfprintf(stdout, "%s: cp = %d, pass = %d.\n", fcn_name, cp++, pass);
/* set the dataset creation plist to specify that the raw data is
- * to be partioned into 1X10X10 element chunks.
+ * to be partitioned into 1X10X10 element chunks.
*/
if (pass) {
@@ -2796,7 +2796,7 @@ verify_cache_image_RO(int file_name_id, int md_write_strat, int mpi_rank)
/* Verify that all other processes receive the cache image block
* from process 0.
*
- * Since we have alread verified that only process 0 has read the
+ * Since we have already verified that only process 0 has read the
* image, it is sufficient to verify that the image was loaded on
* all processes.
*/
@@ -3067,7 +3067,7 @@ verify_cache_image_RW(int file_name_id, int md_write_strat, int mpi_rank)
/* Verify that all other processes receive the cache image block
* from process 0.
*
- * Since we have alread verified that only process 0 has read the
+ * Since we have already verified that only process 0 has read the
* image, it is sufficient to verify that the image was loaded on
* all processes.
*/
@@ -3393,7 +3393,7 @@ smoke_check_1(MPI_Comm mpi_comm, MPI_Info mpi_info, int mpi_rank, int mpi_size)
/* Verify that all other processes receive the cache image block
* from process 0.
*
- * Since we have alread verified that only process 0 has read the
+ * Since we have already verified that only process 0 has read the
* image, it is sufficient to verify that the image was loaded on
* all processes.
*/
@@ -3488,7 +3488,7 @@ smoke_check_1(MPI_Comm mpi_comm, MPI_Info mpi_info, int mpi_rank, int mpi_size)
/* Verify that all other processes receive the cache image block
* from process 0.
*
- * Since we have alread verified that only process 0 has read the
+ * Since we have already verified that only process 0 has read the
* image, it is sufficient to verify that the image was loaded on
* all processes.
*/
@@ -3533,9 +3533,9 @@ smoke_check_1(MPI_Comm mpi_comm, MPI_Info mpi_info, int mpi_rank, int mpi_size)
HDfprintf(stdout, "%s: cp = %d, pass = %d.\n", fcn_name, cp++, pass);
/* 13) Get the size of the file. Verify that it is less
- * than 20 KB. Without deletions and persistant free
+ * than 20 KB. Without deletions and persistent free
* space managers, size size is about 30 MB, so this
- * is sufficient to verify that the persistant free
+ * is sufficient to verify that the persistent free
* space managers are more or less doing their job.
*
* Note that this test will have to change if we use
diff --git a/testpar/t_coll_chunk.c b/testpar/t_coll_chunk.c
index 651a392..104460a 100644
--- a/testpar/t_coll_chunk.c
+++ b/testpar/t_coll_chunk.c
@@ -155,7 +155,7 @@ coll_chunk2(void)
*/
/* ------------------------------------------------------------------------
- * Descriptions for the selection: one singular selection accross many chunks
+ * Descriptions for the selection: one singular selection across many chunks
* Two dimensions, Num of chunks = 2* mpi_size
*
* dim1 = SPACE_DIM1*mpi_size
@@ -209,7 +209,7 @@ coll_chunk3(void)
*/
/* ------------------------------------------------------------------------
- * Descriptions for the selection: one singular selection accross many chunks
+ * Descriptions for the selection: one singular selection across many chunks
* Two dimensions, Num of chunks = 2* mpi_size
*
* dim1 = SPACE_DIM1*mpi_size
@@ -261,7 +261,7 @@ coll_chunk4(void)
*/
/* ------------------------------------------------------------------------
- * Descriptions for the selection: one singular selection accross many chunks
+ * Descriptions for the selection: one singular selection across many chunks
* Two dimensions, Num of chunks = 2* mpi_size
*
* dim1 = SPACE_DIM1*mpi_size
@@ -315,7 +315,7 @@ coll_chunk5(void)
*/
/* ------------------------------------------------------------------------
- * Descriptions for the selection: one singular selection accross many chunks
+ * Descriptions for the selection: one singular selection across many chunks
* Two dimensions, Num of chunks = 2* mpi_size
*
* dim1 = SPACE_DIM1*mpi_size
@@ -367,7 +367,7 @@ coll_chunk6(void)
*/
/* ------------------------------------------------------------------------
- * Descriptions for the selection: one singular selection accross many chunks
+ * Descriptions for the selection: one singular selection across many chunks
* Two dimensions, Num of chunks = 2* mpi_size
*
* dim1 = SPACE_DIM1*mpi_size
@@ -419,7 +419,7 @@ coll_chunk7(void)
*/
/* ------------------------------------------------------------------------
- * Descriptions for the selection: one singular selection accross many chunks
+ * Descriptions for the selection: one singular selection across many chunks
* Two dimensions, Num of chunks = 2* mpi_size
*
* dim1 = SPACE_DIM1*mpi_size
@@ -471,7 +471,7 @@ coll_chunk8(void)
*/
/* ------------------------------------------------------------------------
- * Descriptions for the selection: one singular selection accross many chunks
+ * Descriptions for the selection: one singular selection across many chunks
* Two dimensions, Num of chunks = 2* mpi_size
*
* dim1 = SPACE_DIM1*mpi_size
@@ -523,7 +523,7 @@ coll_chunk9(void)
*/
/* ------------------------------------------------------------------------
- * Descriptions for the selection: one singular selection accross many chunks
+ * Descriptions for the selection: one singular selection across many chunks
* Two dimensions, Num of chunks = 2* mpi_size
*
* dim1 = SPACE_DIM1*mpi_size
diff --git a/testpar/t_coll_md_read.c b/testpar/t_coll_md_read.c
index fd62eb6..66f3151 100644
--- a/testpar/t_coll_md_read.c
+++ b/testpar/t_coll_md_read.c
@@ -351,7 +351,7 @@ test_multi_chunk_io_addrmap_issue(void)
* ranks must exceed or equal 10000. In other words, each MPI rank must be
* responsible for 10000 or more unique chunks.
*
- * Condition 3: This test will currently only be reliably reproducable for 2 or 3
+ * Condition 3: This test will currently only be reliably reproducible for 2 or 3
* MPI ranks. The threshold value calculated reduces to a constant 100 / mpi_size,
* and is compared against a default value of 30%.
*
diff --git a/testpar/t_dset.c b/testpar/t_dset.c
index bbd4b28..40268a8 100644
--- a/testpar/t_dset.c
+++ b/testpar/t_dset.c
@@ -2881,7 +2881,7 @@ none_selection_chunk(void)
* Simple independent I/O. This tests that the defaults are properly set.
*
* TEST_ACTUAL_IO_RESET:
- * Perfroms collective and then independent I/O wit hthe same dxpl to
+ * Performs collective and then independent I/O with hthe same dxpl to
* make sure the peroperty is correctly reset to the default on each use.
* Specifically, this test runs TEST_ACTUAL_IO_MULTI_CHUNK_NO_OPT_MIX_DISAGREE
* (The most complex case that works on all builds) and then performs
@@ -2890,7 +2890,7 @@ none_selection_chunk(void)
* Note: DIRECT_MULTI_CHUNK_MIX and DIRECT_MULTI_CHUNK_MIX_DISAGREE
* is not needed as they are covered by DIRECT_CHUNK_MIX and
* MULTI_CHUNK_MIX_DISAGREE cases. _DIRECT_ cases are only for testing
- * path way to multi-chunk-io by H5FD_MPIO_CHUNK_MULTI_IO insted of num-threshold.
+ * path way to multi-chunk-io by H5FD_MPIO_CHUNK_MULTI_IO instead of num-threshold.
*
* Modification:
* - Refctore to remove multi-chunk-without-opimization test and update for
@@ -3078,7 +3078,7 @@ test_actual_io_mode(int selection_mode)
break;
/* RESET tests that the properties are properly reset to defaults each time I/O is
- * performed. To acheive this, we have RESET perform collective I/O (which would change
+ * performed. To achieve this, we have RESET perform collective I/O (which would change
* the values from the defaults) followed by independent I/O (which should report the
* default values). RESET doesn't need to have a unique selection, so we reuse
* MULTI_CHUMK_MIX_DISAGREE, which was chosen because it is a complex case that works
@@ -3094,7 +3094,7 @@ test_actual_io_mode(int selection_mode)
* assgigned collective I/O, while each other process gets independent I/O.
* Since the root process with only access the first chunk, it will report
* collective I/O. The subsequent processes will access the first chunk
- * collectively, and their other chunk indpendently, reporting mixed I/O.
+ * collectively, and their other chunk independently, reporting mixed I/O.
*/
if (mpi_rank == 0) {
@@ -3209,7 +3209,7 @@ test_actual_io_mode(int selection_mode)
ret = H5Pset_dxpl_mpio_chunk_opt_num(dxpl_write, (unsigned)mpi_size * 2);
VRFY((ret >= 0), "H5Pset_dxpl_mpio_chunk_opt_num succeeded");
- /* set this to manipulate testing senario about allocating processes
+ /* set this to manipulate testing scenario about allocating processes
* to chunks */
ret = H5Pset_dxpl_mpio_chunk_opt_ratio(dxpl_write, (unsigned)99);
VRFY((ret >= 0), "H5Pset_dxpl_mpio_chunk_opt_ratio succeeded");
@@ -3233,12 +3233,12 @@ test_actual_io_mode(int selection_mode)
H5Eprint2(H5E_DEFAULT, stdout);
VRFY((ret >= 0), "H5Dwrite() dataset multichunk write succeeded");
- /* Retreive Actual io valuess */
+ /* Retrieve Actual io values */
ret = H5Pget_mpio_actual_io_mode(dxpl_write, &actual_io_mode_write);
- VRFY((ret >= 0), "retriving actual io mode suceeded");
+ VRFY((ret >= 0), "retrieving actual io mode succeeded");
ret = H5Pget_mpio_actual_chunk_opt_mode(dxpl_write, &actual_chunk_opt_mode_write);
- VRFY((ret >= 0), "retriving actual chunk opt mode succeeded");
+ VRFY((ret >= 0), "retrieving actual chunk opt mode succeeded");
/* Read */
ret = H5Dread(dataset, data_type, mem_space, file_space, dxpl_read, buffer);
@@ -3246,12 +3246,12 @@ test_actual_io_mode(int selection_mode)
H5Eprint2(H5E_DEFAULT, stdout);
VRFY((ret >= 0), "H5Dread() dataset multichunk read succeeded");
- /* Retreive Actual io values */
+ /* Retrieve Actual io values */
ret = H5Pget_mpio_actual_io_mode(dxpl_read, &actual_io_mode_read);
- VRFY((ret >= 0), "retriving actual io mode succeeded");
+ VRFY((ret >= 0), "retrieving actual io mode succeeded");
ret = H5Pget_mpio_actual_chunk_opt_mode(dxpl_read, &actual_chunk_opt_mode_read);
- VRFY((ret >= 0), "retriving actual chunk opt mode succeeded");
+ VRFY((ret >= 0), "retrieving actual chunk opt mode succeeded");
/* Check write vs read */
VRFY((actual_io_mode_read == actual_io_mode_write),
@@ -3272,7 +3272,7 @@ test_actual_io_mode(int selection_mode)
actual_io_mode_write);
}
- /* To test that the property is succesfully reset to the default, we perform some
+ /* To test that the property is successfully reset to the default, we perform some
* independent I/O after the collective I/O
*/
if (selection_mode == TEST_ACTUAL_IO_RESET) {
@@ -3289,9 +3289,9 @@ test_actual_io_mode(int selection_mode)
/* Check Properties */
ret = H5Pget_mpio_actual_io_mode(dxpl_write, &actual_io_mode_write);
- VRFY((ret >= 0), "retriving actual io mode succeeded");
+ VRFY((ret >= 0), "retrieving actual io mode succeeded");
ret = H5Pget_mpio_actual_chunk_opt_mode(dxpl_write, &actual_chunk_opt_mode_write);
- VRFY((ret >= 0), "retriving actual chunk opt mode succeeded");
+ VRFY((ret >= 0), "retrieving actual chunk opt mode succeeded");
VRFY(actual_chunk_opt_mode_write == H5D_MPIO_NO_CHUNK_OPTIMIZATION,
"actual_chunk_opt_mode has correct value for reset write (independent)");
@@ -3304,9 +3304,9 @@ test_actual_io_mode(int selection_mode)
/* Check Properties */
ret = H5Pget_mpio_actual_io_mode(dxpl_read, &actual_io_mode_read);
- VRFY((ret >= 0), "retriving actual io mode succeeded");
+ VRFY((ret >= 0), "retrieving actual io mode succeeded");
ret = H5Pget_mpio_actual_chunk_opt_mode(dxpl_read, &actual_chunk_opt_mode_read);
- VRFY((ret >= 0), "retriving actual chunk opt mode succeeded");
+ VRFY((ret >= 0), "retrieving actual chunk opt mode succeeded");
VRFY(actual_chunk_opt_mode_read == H5D_MPIO_NO_CHUNK_OPTIMIZATION,
"actual_chunk_opt_mode has correct value for reset read (independent)");
@@ -3359,11 +3359,11 @@ actual_io_mode_tests(void)
test_actual_io_mode(TEST_ACTUAL_IO_MULTI_CHUNK_IND);
test_actual_io_mode(TEST_ACTUAL_IO_MULTI_CHUNK_COL);
- /* The Multi Chunk Mixed test requires atleast three processes. */
+ /* The Multi Chunk Mixed test requires at least three processes. */
if (mpi_size > 2)
test_actual_io_mode(TEST_ACTUAL_IO_MULTI_CHUNK_MIX);
else
- HDfprintf(stdout, "Multi Chunk Mixed test requires 3 proceses minimum\n");
+ HDfprintf(stdout, "Multi Chunk Mixed test requires 3 processes minimum\n");
test_actual_io_mode(TEST_ACTUAL_IO_MULTI_CHUNK_MIX_DISAGREE);
@@ -3407,7 +3407,7 @@ actual_io_mode_tests(void)
* Test for Data Type Conversion as the cause of breaking collective I/O.
*
* TEST_DATA_TRANSFORMS:
- * Test for Data Transfrom feature as the cause of breaking collective I/O.
+ * Test for Data Transform feature as the cause of breaking collective I/O.
*
* TEST_NOT_SIMPLE_OR_SCALAR_DATASPACES:
* Test for NULL dataspace as the cause of breaking collective I/O.
@@ -3571,7 +3571,7 @@ test_no_collective_cause_mode(int selection_mode)
}
if (selection_mode & TEST_DATA_TRANSFORMS) {
- test_name = "Broken Collective I/O - DATA Transfroms";
+ test_name = "Broken Collective I/O - DATA Transforms";
no_collective_cause_local_expected |= H5D_MPIO_DATA_TRANSFORMS;
no_collective_cause_global_expected |= H5D_MPIO_DATA_TRANSFORMS;
}
@@ -3669,7 +3669,7 @@ test_no_collective_cause_mode(int selection_mode)
/* Get the cause of broken collective I/O */
ret = H5Pget_mpio_no_collective_cause(dxpl_write, &no_collective_cause_local_write,
&no_collective_cause_global_write);
- VRFY((ret >= 0), "retriving no collective cause succeeded");
+ VRFY((ret >= 0), "retrieving no collective cause succeeded");
/*---------------------
* Test Read access
@@ -3689,7 +3689,7 @@ test_no_collective_cause_mode(int selection_mode)
/* Get the cause of broken collective I/O */
ret = H5Pget_mpio_no_collective_cause(dxpl_read, &no_collective_cause_local_read,
&no_collective_cause_global_read);
- VRFY((ret >= 0), "retriving no collective cause succeeded");
+ VRFY((ret >= 0), "retrieving no collective cause succeeded");
/* Check write vs read */
VRFY((no_collective_cause_local_read == no_collective_cause_local_write),
@@ -3935,7 +3935,7 @@ test_no_collective_cause_mode_filter(int selection_mode)
/* Get the cause of broken collective I/O */
ret = H5Pget_mpio_no_collective_cause(dxpl, &no_collective_cause_local_read,
&no_collective_cause_global_read);
- VRFY((ret >= 0), "retriving no collective cause succeeded");
+ VRFY((ret >= 0), "retrieving no collective cause succeeded");
/* Test values */
HDmemset(message, 0, sizeof(message));
@@ -4114,7 +4114,7 @@ dataset_atomicity(void)
/* file locking allows only one file open (serial) for writing */
if (MAINPROCESS) {
fid = H5Fopen(filename, H5F_ACC_RDWR, H5P_DEFAULT);
- VRFY((fid >= 0), "H5Fopen succeeed");
+ VRFY((fid >= 0), "H5Fopen succeeded");
/* should fail */
H5E_BEGIN_TRY
diff --git a/testpar/t_file.c b/testpar/t_file.c
index 5348f4f..c6b9e32 100644
--- a/testpar/t_file.c
+++ b/testpar/t_file.c
@@ -179,7 +179,7 @@ test_page_buffer_access(void)
data = (int *)HDmalloc(sizeof(int) * (size_t)num_elements);
- /* intialize all the elements to have a value of -1 */
+ /* initialize all the elements to have a value of -1 */
for (i = 0; i < num_elements; i++)
data[i] = -1;
if (MAINPROCESS) {
diff --git a/testpar/t_mdset.c b/testpar/t_mdset.c
index 3041e77..97d5966 100644
--- a/testpar/t_mdset.c
+++ b/testpar/t_mdset.c
@@ -1721,7 +1721,7 @@ io_mode_confusion(void)
* Open the each of the data sets in turn. If all opens are successful,
* the test passes. Otherwise the test fails.
*
- * Note that this test will probably become irrelevent shortly, when we
+ * Note that this test will probably become irrelevant shortly, when we
* land the journaling modifications on the trunk -- at which point all
* cache clients will have to construct on disk images on demand.
*
@@ -1782,7 +1782,7 @@ rr_obj_hdr_flush_confusion(void)
mrc = MPI_Comm_split(MPI_COMM_WORLD, is_reader, mpi_rank, &comm);
VRFY((mrc == MPI_SUCCESS), "MPI_Comm_split");
- /* The reader proocesses branches off to do reading
+ /* The reader processes branches off to do reading
* while the writer processes continues to do writing
* Whenever writers finish one writing step, including a H5Fflush,
* they inform the readers, via MPI_COMM_WORLD, to verify.
@@ -2063,7 +2063,7 @@ rr_obj_hdr_flush_confusion_writer(MPI_Comm comm)
/*
* flush the metadata cache yet again to clean the object headers.
*
- * This is an attempt to crate a situation where we have dirty
+ * This is an attempt to create a situation where we have dirty
* object header continuation chunks, but clean opject headers
* to verify a speculative bug fix -- it doesn't seem to work,
* but I will leave the code in anyway, as the object header
diff --git a/testpar/t_mpi.c b/testpar/t_mpi.c
index a8a756c..96cbc4b 100644
--- a/testpar/t_mpi.c
+++ b/testpar/t_mpi.c
@@ -215,7 +215,7 @@ test_mpio_gb_file(char *filename)
* sizes.
*/
if (MAINPROCESS) { /* only process 0 needs to check it*/
- HDprintf("MPI_Offset is %s %d bytes integeral type\n", is_signed ? "signed" : "unsigned",
+ HDprintf("MPI_Offset is %s %d bytes integral type\n", is_signed ? "signed" : "unsigned",
(int)sizeof(MPI_Offset));
if (sizeof_mpi_offset <= 4 && is_signed) {
HDprintf("Skipped 2GB range test "
@@ -400,7 +400,7 @@ finish:
* MPI-IO Test: One writes, Many reads.
* Verify if only one process writes some data and then all other
* processes can read them back correctly. This tests if the
- * underlaying parallel I/O and file system supports parallel I/O
+ * underlying parallel I/O and file system supports parallel I/O
* correctly.
*
* Algorithm: Only one process (e.g., process 0) writes some data.
diff --git a/testpar/t_ph5basic.c b/testpar/t_ph5basic.c
index 05f2c77..59f14cb 100644
--- a/testpar/t_ph5basic.c
+++ b/testpar/t_ph5basic.c
@@ -56,7 +56,7 @@ test_fapl_mpio_dup(void)
HDprintf("rank/size of MPI_COMM_WORLD are %d/%d\n", mpi_rank, mpi_size);
/* Create a new communicator that has the same processes as MPI_COMM_WORLD.
- * Use MPI_Comm_split because it is simplier than MPI_Comm_create
+ * Use MPI_Comm_split because it is simpler than MPI_Comm_create
*/
mrc = MPI_Comm_split(MPI_COMM_WORLD, 0, 0, &comm);
VRFY((mrc == MPI_SUCCESS), "MPI_Comm_split");
diff --git a/testpar/t_pread.c b/testpar/t_pread.c
index 17091cb..d79130b 100644
--- a/testpar/t_pread.c
+++ b/testpar/t_pread.c
@@ -68,7 +68,7 @@ static char *test_argv0 = NULL;
* more or less in half. Each sub group will operate
* collectively on their assigned file. This split into
* subgroups validates that parallel groups can successfully
- * open and read data independantly from the other parallel
+ * open and read data independently from the other parallel
* operations taking place.
*
* Return: Success: 0
@@ -132,7 +132,7 @@ generate_test_file(MPI_Comm comm, int mpi_rank, int group_id)
}
/* Setup the file names
- * The test specfic filenames are stored as consecutive
+ * The test specific filenames are stored as consecutive
* array entries in the global 'FILENAMES' array above.
* Here, we simply decide on the starting index for
* file construction. The reading portion of the test
@@ -1043,11 +1043,11 @@ test_parallel_read(MPI_Comm comm, int mpi_rank, int mpi_size, int group_id)
* Purpose: To implement a parallel test which validates whether the
* new superblock lookup functionality is working correctly.
*
- * The test consists of creating two seperate HDF datasets
+ * The test consists of creating two separate HDF datasets
* in which random text is inserted at the start of each
* file using the 'j5jam' application. This forces the
* HDF5 file superblock to a non-zero offset.
- * Having created the two independant files, we create two
+ * Having created the two independent files, we create two
* non-overlapping MPI groups, each of which is then tasked
* with the opening and validation of the data contained
* therein.
@@ -1152,7 +1152,7 @@ main(int argc, char **argv)
goto finish;
}
- /* Now read the generated test file (stil using MPI_COMM_WORLD) */
+ /* Now read the generated test file (still using MPI_COMM_WORLD) */
nerrs += test_parallel_read(MPI_COMM_WORLD, mpi_rank, mpi_size, which_group);
if (nerrs > 0) {
diff --git a/testpar/t_shapesame.c b/testpar/t_shapesame.c
index 2dd867a..0a18781 100644
--- a/testpar/t_shapesame.c
+++ b/testpar/t_shapesame.c
@@ -11,7 +11,7 @@
* * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * */
/*
- This program will test independant and collective reads and writes between
+ This program will test independent and collective reads and writes between
selections of different rank that non-the-less are deemed as having the
same shape by H5Sselect_shape_same().
*/
@@ -415,11 +415,11 @@ hs_dr_pio_test__setup(const int test_num, const int edge_size, const int checker
/* setup selections for writing initial data to the small data set */
ret = H5Sselect_hyperslab(tv_ptr->mem_small_ds_sid, H5S_SELECT_SET, tv_ptr->start, tv_ptr->stride,
tv_ptr->count, tv_ptr->block);
- VRFY((ret >= 0), "H5Sselect_hyperslab(mem_small_ds_sid, set) suceeded");
+ VRFY((ret >= 0), "H5Sselect_hyperslab(mem_small_ds_sid, set) succeeded");
ret = H5Sselect_hyperslab(tv_ptr->file_small_ds_sid_0, H5S_SELECT_SET, tv_ptr->start, tv_ptr->stride,
tv_ptr->count, tv_ptr->block);
- VRFY((ret >= 0), "H5Sselect_hyperslab(file_small_ds_sid_0, set) suceeded");
+ VRFY((ret >= 0), "H5Sselect_hyperslab(file_small_ds_sid_0, set) succeeded");
if (MAINPROCESS) { /* add an additional slice to the selections */
@@ -427,11 +427,11 @@ hs_dr_pio_test__setup(const int test_num, const int edge_size, const int checker
ret = H5Sselect_hyperslab(tv_ptr->mem_small_ds_sid, H5S_SELECT_OR, tv_ptr->start, tv_ptr->stride,
tv_ptr->count, tv_ptr->block);
- VRFY((ret >= 0), "H5Sselect_hyperslab(mem_small_ds_sid, or) suceeded");
+ VRFY((ret >= 0), "H5Sselect_hyperslab(mem_small_ds_sid, or) succeeded");
ret = H5Sselect_hyperslab(tv_ptr->file_small_ds_sid_0, H5S_SELECT_OR, tv_ptr->start, tv_ptr->stride,
tv_ptr->count, tv_ptr->block);
- VRFY((ret >= 0), "H5Sselect_hyperslab(file_small_ds_sid_0, or) suceeded");
+ VRFY((ret >= 0), "H5Sselect_hyperslab(file_small_ds_sid_0, or) succeeded");
}
/* write the initial value of the small data set to file */
@@ -475,21 +475,21 @@ hs_dr_pio_test__setup(const int test_num, const int edge_size, const int checker
ret = H5Sselect_hyperslab(tv_ptr->mem_large_ds_sid, H5S_SELECT_SET, tv_ptr->start, tv_ptr->stride,
tv_ptr->count, tv_ptr->block);
- VRFY((ret >= 0), "H5Sselect_hyperslab(mem_large_ds_sid, set) suceeded");
+ VRFY((ret >= 0), "H5Sselect_hyperslab(mem_large_ds_sid, set) succeeded");
ret = H5Sselect_hyperslab(tv_ptr->file_large_ds_sid_0, H5S_SELECT_SET, tv_ptr->start, tv_ptr->stride,
tv_ptr->count, tv_ptr->block);
- VRFY((ret >= 0), "H5Sselect_hyperslab(file_large_ds_sid_0, set) suceeded");
+ VRFY((ret >= 0), "H5Sselect_hyperslab(file_large_ds_sid_0, set) succeeded");
/* In passing, setup the process slice dataspaces as well */
ret = H5Sselect_hyperslab(tv_ptr->mem_large_ds_process_slice_sid, H5S_SELECT_SET, tv_ptr->start,
tv_ptr->stride, tv_ptr->count, tv_ptr->block);
- VRFY((ret >= 0), "H5Sselect_hyperslab(mem_large_ds_process_slice_sid, set) suceeded");
+ VRFY((ret >= 0), "H5Sselect_hyperslab(mem_large_ds_process_slice_sid, set) succeeded");
ret = H5Sselect_hyperslab(tv_ptr->file_large_ds_process_slice_sid, H5S_SELECT_SET, tv_ptr->start,
tv_ptr->stride, tv_ptr->count, tv_ptr->block);
- VRFY((ret >= 0), "H5Sselect_hyperslab(file_large_ds_process_slice_sid, set) suceeded");
+ VRFY((ret >= 0), "H5Sselect_hyperslab(file_large_ds_process_slice_sid, set) succeeded");
if (MAINPROCESS) { /* add an additional slice to the selections */
@@ -497,11 +497,11 @@ hs_dr_pio_test__setup(const int test_num, const int edge_size, const int checker
ret = H5Sselect_hyperslab(tv_ptr->mem_large_ds_sid, H5S_SELECT_OR, tv_ptr->start, tv_ptr->stride,
tv_ptr->count, tv_ptr->block);
- VRFY((ret >= 0), "H5Sselect_hyperslab(mem_large_ds_sid, or) suceeded");
+ VRFY((ret >= 0), "H5Sselect_hyperslab(mem_large_ds_sid, or) succeeded");
ret = H5Sselect_hyperslab(tv_ptr->file_large_ds_sid_0, H5S_SELECT_OR, tv_ptr->start, tv_ptr->stride,
tv_ptr->count, tv_ptr->block);
- VRFY((ret >= 0), "H5Sselect_hyperslab(file_large_ds_sid_0, or) suceeded");
+ VRFY((ret >= 0), "H5Sselect_hyperslab(file_large_ds_sid_0, or) succeeded");
}
/* write the initial value of the large data set to file */
@@ -933,7 +933,7 @@ contig_hs_dr_pio_test__d2m_s2l(struct hs_dr_pio_test_vars_t *tv_ptr)
ret = H5Sselect_hyperslab(tv_ptr->file_small_ds_sid_0, H5S_SELECT_SET, tv_ptr->start, tv_ptr->stride,
tv_ptr->count, tv_ptr->block);
- VRFY((ret >= 0), "H5Sselect_hyperslab(file_small_ds_sid_0, set) suceeded");
+ VRFY((ret >= 0), "H5Sselect_hyperslab(file_small_ds_sid_0, set) succeeded");
#if CONTIG_HS_DR_PIO_TEST__D2M_S2L__DEBUG
HDfprintf(stdout, "%s reading slices of on disk small data set into slices of big data set.\n", fcnName);
@@ -1185,11 +1185,11 @@ contig_hs_dr_pio_test__m2d_l2s(struct hs_dr_pio_test_vars_t *tv_ptr)
ret = H5Sselect_hyperslab(tv_ptr->file_small_ds_sid_0, H5S_SELECT_SET, tv_ptr->start, tv_ptr->stride,
tv_ptr->count, tv_ptr->block);
- VRFY((ret >= 0), "H5Sselect_hyperslab(file_small_ds_sid_0, set) suceeded");
+ VRFY((ret >= 0), "H5Sselect_hyperslab(file_small_ds_sid_0, set) succeeded");
ret = H5Sselect_hyperslab(tv_ptr->mem_small_ds_sid, H5S_SELECT_SET, tv_ptr->start, tv_ptr->stride,
tv_ptr->count, tv_ptr->block);
- VRFY((ret >= 0), "H5Sselect_hyperslab(mem_small_ds_sid, set) suceeded");
+ VRFY((ret >= 0), "H5Sselect_hyperslab(mem_small_ds_sid, set) succeeded");
/* set up start, stride, count, and block -- note that we will
* change start[] so as to read slices of the large cube.
@@ -1462,7 +1462,7 @@ contig_hs_dr_pio_test__m2d_s2l(struct hs_dr_pio_test_vars_t *tv_ptr)
ret = H5Sselect_hyperslab(tv_ptr->mem_small_ds_sid, H5S_SELECT_SET, tv_ptr->start, tv_ptr->stride,
tv_ptr->count, tv_ptr->block);
- VRFY((ret >= 0), "H5Sselect_hyperslab(mem_small_ds_sid, set) suceeded");
+ VRFY((ret >= 0), "H5Sselect_hyperslab(mem_small_ds_sid, set) succeeded");
/* set up start, stride, count, and block -- note that we will
* change start[] so as to write slices of the small data set to
@@ -1572,7 +1572,7 @@ contig_hs_dr_pio_test__m2d_s2l(struct hs_dr_pio_test_vars_t *tv_ptr)
ret = H5Dwrite(tv_ptr->large_dataset, H5T_NATIVE_UINT32, tv_ptr->large_ds_slice_sid,
tv_ptr->file_large_ds_process_slice_sid, tv_ptr->xfer_plist,
tv_ptr->large_ds_buf_2);
- VRFY((ret != FAIL), "H5Dwrite() to zero large ds suceeded");
+ VRFY((ret != FAIL), "H5Dwrite() to zero large ds succeeded");
/* select the portion of the in memory large cube to which we
* are going to write data.
@@ -2008,8 +2008,8 @@ contig_hs_dr_pio_test(ShapeSameTestMethods sstest_type)
** edge_size, and a checker_edge_size, select a checker
** board selection of a sel_rank (sel_rank < tgt_rank)
** dimensional slice through the dataspace parallel to the
-** sel_rank fastest changing indicies, with origin (in the
-** higher indicies) as indicated by the start array.
+** sel_rank fastest changing indices, with origin (in the
+** higher indices) as indicated by the start array.
**
** Note that this function, like all its relatives, is
** hard coded to presume a maximum dataspace rank of 5.
@@ -2318,7 +2318,7 @@ ckrbrd_hs_dr_pio_test__slct_ckrbrd(const int mpi_rank, const hid_t tgt_sid, cons
** or writing a checker board selection of an m (1 <= m <
** rank) dimensional slice through this processes slice
** of the target data set. Also, this slice must be parallel
-** to the fastest changing indicies.
+** to the fastest changing indices.
**
** It is further presumed that the buffer was zeroed before
** the read/write, and that the full target data set (i.e.
@@ -2353,7 +2353,7 @@ ckrbrd_hs_dr_pio_test__slct_ckrbrd(const int mpi_rank, const hid_t tgt_sid, cons
** of the buffer resides either at the origin of either
** a selected or an unselected checker. (Translation:
** if partial checkers appear in the buffer, they will
-** intersect the edges of the n-cube oposite the origin.)
+** intersect the edges of the n-cube opposite the origin.)
**
****************************************************************/
@@ -3091,11 +3091,11 @@ ckrbrd_hs_dr_pio_test__m2d_l2s(struct hs_dr_pio_test_vars_t *tv_ptr)
ret = H5Sselect_hyperslab(tv_ptr->file_small_ds_sid_0, H5S_SELECT_SET, tv_ptr->start, tv_ptr->stride,
tv_ptr->count, tv_ptr->block);
- VRFY((ret >= 0), "H5Sselect_hyperslab(file_small_ds_sid_0, set) suceeded");
+ VRFY((ret >= 0), "H5Sselect_hyperslab(file_small_ds_sid_0, set) succeeded");
ret = H5Sselect_hyperslab(tv_ptr->mem_small_ds_sid, H5S_SELECT_SET, tv_ptr->start, tv_ptr->stride,
tv_ptr->count, tv_ptr->block);
- VRFY((ret >= 0), "H5Sselect_hyperslab(mem_small_ds_sid, set) suceeded");
+ VRFY((ret >= 0), "H5Sselect_hyperslab(mem_small_ds_sid, set) succeeded");
sel_start[0] = sel_start[1] = sel_start[2] = sel_start[3] = sel_start[4] = 0;
sel_start[tv_ptr->small_ds_offset] = (hsize_t)(tv_ptr->mpi_rank);
@@ -3384,11 +3384,11 @@ ckrbrd_hs_dr_pio_test__m2d_s2l(struct hs_dr_pio_test_vars_t *tv_ptr)
ret = H5Sselect_hyperslab(tv_ptr->file_large_ds_sid_0, H5S_SELECT_SET, tv_ptr->start, tv_ptr->stride,
tv_ptr->count, tv_ptr->block);
- VRFY((ret >= 0), "H5Sselect_hyperslab(file_large_ds_sid_0, set) suceeded");
+ VRFY((ret >= 0), "H5Sselect_hyperslab(file_large_ds_sid_0, set) succeeded");
ret = H5Sselect_hyperslab(tv_ptr->mem_large_ds_sid, H5S_SELECT_SET, tv_ptr->start, tv_ptr->stride,
tv_ptr->count, tv_ptr->block);
- VRFY((ret >= 0), "H5Sselect_hyperslab(tv_ptr->mem_large_ds_sid, set) suceeded");
+ VRFY((ret >= 0), "H5Sselect_hyperslab(tv_ptr->mem_large_ds_sid, set) succeeded");
/* setup a checkerboard selection of the slice of the in memory small
* data set associated with the process's mpi rank.
@@ -3496,7 +3496,7 @@ ckrbrd_hs_dr_pio_test__m2d_s2l(struct hs_dr_pio_test_vars_t *tv_ptr)
*/
ret = H5Dwrite(tv_ptr->large_dataset, H5T_NATIVE_UINT32, tv_ptr->mem_large_ds_sid,
tv_ptr->file_large_ds_sid_0, tv_ptr->xfer_plist, tv_ptr->large_ds_buf_2);
- VRFY((ret != FAIL), "H5Dwrite() to zero large ds suceeded");
+ VRFY((ret != FAIL), "H5Dwrite() to zero large ds succeeded");
/* select the portion of the in memory large cube to which we
* are going to write data.
@@ -4210,42 +4210,42 @@ create_faccess_plist(MPI_Comm comm, MPI_Info info, int l_facc_type)
return (ret_pl);
}
-/* Shape Same test using contigous hyperslab using independent IO on contigous datasets */
+/* Shape Same test using contiguous hyperslab using independent IO on contiguous datasets */
static void
sscontig1(void)
{
contig_hs_dr_pio_test(IND_CONTIG);
}
-/* Shape Same test using contigous hyperslab using collective IO on contigous datasets */
+/* Shape Same test using contiguous hyperslab using collective IO on contiguous datasets */
static void
sscontig2(void)
{
contig_hs_dr_pio_test(COL_CONTIG);
}
-/* Shape Same test using contigous hyperslab using independent IO on chunked datasets */
+/* Shape Same test using contiguous hyperslab using independent IO on chunked datasets */
static void
sscontig3(void)
{
contig_hs_dr_pio_test(IND_CHUNKED);
}
-/* Shape Same test using contigous hyperslab using collective IO on chunked datasets */
+/* Shape Same test using contiguous hyperslab using collective IO on chunked datasets */
static void
sscontig4(void)
{
contig_hs_dr_pio_test(COL_CHUNKED);
}
-/* Shape Same test using checker hyperslab using independent IO on contigous datasets */
+/* Shape Same test using checker hyperslab using independent IO on contiguous datasets */
static void
sschecker1(void)
{
ckrbrd_hs_dr_pio_test(IND_CONTIG);
}
-/* Shape Same test using checker hyperslab using collective IO on contigous datasets */
+/* Shape Same test using checker hyperslab using collective IO on contiguous datasets */
static void
sschecker2(void)
{
@@ -4305,7 +4305,7 @@ main(int argc, char **argv)
/* Initialize testing framework */
TestInit(argv[0], usage, parse_options);
- /* Shape Same tests using contigous hyperslab */
+ /* Shape Same tests using contiguous hyperslab */
AddTest("sscontig1", sscontig1, NULL, "Cntg hslab, ind IO, cntg dsets", PARATESTFILE);
AddTest("sscontig2", sscontig2, NULL, "Cntg hslab, col IO, cntg dsets", PARATESTFILE);
AddTest("sscontig3", sscontig3, NULL, "Cntg hslab, ind IO, chnk dsets", PARATESTFILE);
diff --git a/testpar/t_span_tree.c b/testpar/t_span_tree.c
index 518741d..e624a01 100644
--- a/testpar/t_span_tree.c
+++ b/testpar/t_span_tree.c
@@ -424,7 +424,7 @@ coll_write_test(int chunk_factor)
the correctedness of collective write compared with
independent write,
- In order to throughly test this feature, we choose
+ In order to thoroughly test this feature, we choose
a different selection set for reading the data out.
@@ -890,8 +890,8 @@ coll_read_test(void)
** edge_size, and a checker_edge_size, select a checker
** board selection of a sel_rank (sel_rank < tgt_rank)
** dimensional slice through the dataspace parallel to the
-** sel_rank fastest changing indicies, with origin (in the
-** higher indicies) as indicated by the start array.
+** sel_rank fastest changing indices, with origin (in the
+** higher indices) as indicated by the start array.
**
** Note that this function, is hard coded to presume a
** maximum dataspace rank of 5.
@@ -1230,7 +1230,7 @@ lower_dim_size_comp_test__select_checker_board(const int mpi_rank, const hid_t t
** or writing a checker board selection of an m (1 <= m <
** rank) dimensional slice through this processes slice
** of the target data set. Also, this slice must be parallel
-** to the fastest changing indicies.
+** to the fastest changing indices.
**
** It is further presumed that the buffer was zeroed before
** the read/write, and that the full target data set (i.e.
@@ -1265,7 +1265,7 @@ lower_dim_size_comp_test__select_checker_board(const int mpi_rank, const hid_t t
** of the buffer resides either at the origin of either
** a selected or an unselected checker. (Translation:
** if partial checkers appear in the buffer, they will
-** intersect the edges of the n-cube oposite the origin.)
+** intersect the edges of the n-cube opposite the origin.)
**
****************************************************************/
@@ -1726,7 +1726,7 @@ lower_dim_size_comp_test__run_test(const int chunk_edge_size, const hbool_t use_
if (!use_collective_io) {
ret = H5Pset_dxpl_mpio_collective_opt(xfer_plist, H5FD_MPIO_INDIVIDUAL_IO);
- VRFY((ret >= 0), "H5Pset_dxpl_mpio_collective_opt() suceeded");
+ VRFY((ret >= 0), "H5Pset_dxpl_mpio_collective_opt() succeeded");
}
/* setup selection to write initial data to the small data sets */
@@ -1758,10 +1758,10 @@ lower_dim_size_comp_test__run_test(const int chunk_edge_size, const hbool_t use_
/* setup selections for writing initial data to the small data set */
ret = H5Sselect_hyperslab(mem_small_ds_sid, H5S_SELECT_SET, start, stride, count, block);
- VRFY((ret >= 0), "H5Sselect_hyperslab(mem_small_ds_sid, set) suceeded");
+ VRFY((ret >= 0), "H5Sselect_hyperslab(mem_small_ds_sid, set) succeeded");
ret = H5Sselect_hyperslab(file_small_ds_sid, H5S_SELECT_SET, start, stride, count, block);
- VRFY((ret >= 0), "H5Sselect_hyperslab(file_small_ds_sid, set) suceeded");
+ VRFY((ret >= 0), "H5Sselect_hyperslab(file_small_ds_sid, set) succeeded");
if (MAINPROCESS) { /* add an additional slice to the selections */
@@ -1782,10 +1782,10 @@ lower_dim_size_comp_test__run_test(const int chunk_edge_size, const hbool_t use_
#endif /* LOWER_DIM_SIZE_COMP_TEST__RUN_TEST__DEBUG */
ret = H5Sselect_hyperslab(mem_small_ds_sid, H5S_SELECT_OR, start, stride, count, block);
- VRFY((ret >= 0), "H5Sselect_hyperslab(mem_small_ds_sid, or) suceeded");
+ VRFY((ret >= 0), "H5Sselect_hyperslab(mem_small_ds_sid, or) succeeded");
ret = H5Sselect_hyperslab(file_small_ds_sid, H5S_SELECT_OR, start, stride, count, block);
- VRFY((ret >= 0), "H5Sselect_hyperslab(file_small_ds_sid, or) suceeded");
+ VRFY((ret >= 0), "H5Sselect_hyperslab(file_small_ds_sid, or) succeeded");
}
check = H5Sselect_valid(mem_small_ds_sid);
@@ -1864,10 +1864,10 @@ lower_dim_size_comp_test__run_test(const int chunk_edge_size, const hbool_t use_
#endif /* LOWER_DIM_SIZE_COMP_TEST__RUN_TEST__DEBUG */
ret = H5Sselect_hyperslab(mem_large_ds_sid, H5S_SELECT_SET, start, stride, count, block);
- VRFY((ret >= 0), "H5Sselect_hyperslab(mem_large_ds_sid, set) suceeded");
+ VRFY((ret >= 0), "H5Sselect_hyperslab(mem_large_ds_sid, set) succeeded");
ret = H5Sselect_hyperslab(file_large_ds_sid, H5S_SELECT_SET, start, stride, count, block);
- VRFY((ret >= 0), "H5Sselect_hyperslab(file_large_ds_sid, set) suceeded");
+ VRFY((ret >= 0), "H5Sselect_hyperslab(file_large_ds_sid, set) succeeded");
#if LOWER_DIM_SIZE_COMP_TEST__RUN_TEST__DEBUG
if (mpi_rank == LOWER_DIM_SIZE_COMP_TEST_DEBUG_TARGET_RANK) {
@@ -1897,10 +1897,10 @@ lower_dim_size_comp_test__run_test(const int chunk_edge_size, const hbool_t use_
#endif /* LOWER_DIM_SIZE_COMP_TEST__RUN_TEST__DEBUG */
ret = H5Sselect_hyperslab(mem_large_ds_sid, H5S_SELECT_OR, start, stride, count, block);
- VRFY((ret >= 0), "H5Sselect_hyperslab(mem_large_ds_sid, or) suceeded");
+ VRFY((ret >= 0), "H5Sselect_hyperslab(mem_large_ds_sid, or) succeeded");
ret = H5Sselect_hyperslab(file_large_ds_sid, H5S_SELECT_OR, start, stride, count, block);
- VRFY((ret >= 0), "H5Sselect_hyperslab(file_large_ds_sid, or) suceeded");
+ VRFY((ret >= 0), "H5Sselect_hyperslab(file_large_ds_sid, or) succeeded");
#if LOWER_DIM_SIZE_COMP_TEST__RUN_TEST__DEBUG
if (mpi_rank == LOWER_DIM_SIZE_COMP_TEST_DEBUG_TARGET_RANK) {
@@ -2427,7 +2427,7 @@ link_chunk_collective_io_test(void)
/* select the file and mem spaces */
start[0] = (hsize_t)(mpi_rank * LINK_CHUNK_COLLECTIVE_IO_TEST_CHUNK_SIZE);
ret = H5Sselect_hyperslab(file_ds_sid, H5S_SELECT_SET, start, stride, count, block);
- VRFY((ret >= 0), "H5Sselect_hyperslab(file_ds_sid, set) suceeded");
+ VRFY((ret >= 0), "H5Sselect_hyperslab(file_ds_sid, set) succeeded");
ret = H5Sselect_all(write_mem_ds_sid);
VRFY((ret != FAIL), "H5Sselect_all(mem_ds_sid) succeeded");
diff --git a/testpar/testphdf5.h b/testpar/testphdf5.h
index 10e3027..c692287 100644
--- a/testpar/testphdf5.h
+++ b/testpar/testphdf5.h
@@ -218,8 +218,8 @@ typedef int DATATYPE;
/* Shape Same Tests Definitions */
typedef enum {
- IND_CONTIG, /* Independent IO on contigous datasets */
- COL_CONTIG, /* Collective IO on contigous datasets */
+ IND_CONTIG, /* Independent IO on contiguous datasets */
+ COL_CONTIG, /* Collective IO on contiguous datasets */
IND_CHUNKED, /* Independent IO on chunked datasets */
COL_CHUNKED /* Collective IO on chunked datasets */
} ShapeSameTestMethods;
diff --git a/tools/lib/h5diff.c b/tools/lib/h5diff.c
index 74db58f..debd54a 100644
--- a/tools/lib/h5diff.c
+++ b/tools/lib/h5diff.c
@@ -1232,7 +1232,7 @@ diff_match(hid_t file1_id, const char *grp1, trav_info_t *info1, hid_t file2_id,
H5TOOLS_DEBUG("Beginning of big else block");
/* We're in parallel mode */
/* Since the data type of diff value is hsize_t which can
- * be arbitary large such that there is no MPI type that
+ * be arbitrary large such that there is no MPI type that
* matches it, the value is passed between processes as
* an array of bytes in order to be portable. But this
* may not work in non-homogeneous MPI environments.
diff --git a/tools/lib/h5diff.h b/tools/lib/h5diff.h
index 8d7ac13..4afb3f6 100644
--- a/tools/lib/h5diff.h
+++ b/tools/lib/h5diff.h
@@ -74,7 +74,7 @@ typedef struct {
struct exclude_path_list *exclude_attr; /* keep exclude attribute list */
int count_bool; /* count, compare up to count */
hsize_t count; /* count value */
- diff_err_t err_stat; /* an error ocurred (2, error, 1, differences, 0, no error) */
+ diff_err_t err_stat; /* an error occurred (2, error, 1, differences, 0, no error) */
hsize_t nelmts; /* total number of elements */
hsize_t hs_nelmts; /* number of elements to read at a time*/
int rank; /* dimensionality */
diff --git a/tools/lib/h5tools.h b/tools/lib/h5tools.h
index dde4026..cc6183a 100644
--- a/tools/lib/h5tools.h
+++ b/tools/lib/h5tools.h
@@ -240,7 +240,7 @@ typedef struct h5tool_format_t {
*
* fmt_schar: The printf() format to use when rendering data which is
* typed `signed char'. The default is `%d'. This format is
- * used ony if the `ascii' field is zero.
+ * used only if the `ascii' field is zero.
*
* fmt_uchar: The printf() format to use when rendering data which is
* typed `unsigned char'. The default is `%u'. This format
@@ -411,9 +411,9 @@ typedef struct h5tool_format_t {
* sep: Each integer in the index list will be separated from the
* others by this string, which defaults to a comma.
*
- * fmt: After the index values are formated individually and
+ * fmt: After the index values are formatted individually and
* separated from one another by some string, the entire
- * resulting string will be formated according to this
+ * resulting string will be formatted according to this
* printf(3c) format which should include a format for a
* character string. The default is "%s".
*/
@@ -483,7 +483,7 @@ typedef struct h5tool_format_t {
const char *line_suf; /*string to append to each line */
const char *line_sep; /*separates lines */
int line_multi_new; /*split multi-line outputs? */
- const char *line_indent; /*for extra identation if we need it*/
+ const char *line_indent; /*for extra indentation if we need it*/
/*used to skip the first set of checks for line length*/
int skip_first;
@@ -606,7 +606,7 @@ typedef enum {
} driver_idx;
/* The following include, h5tools_str.h, must be after the
- * above stucts are defined. There is a dependency in the following
+ * above structs are defined. There is a dependency in the following
* include that hasn't been identified yet. */
#include "h5tools_str.h"
diff --git a/tools/lib/h5tools_dump.c b/tools/lib/h5tools_dump.c
index cb4f9d2..0fc5d2d 100644
--- a/tools/lib/h5tools_dump.c
+++ b/tools/lib/h5tools_dump.c
@@ -478,7 +478,7 @@ done:
* This is a special case subfunction to print the data in a region reference of type blocks.
*
* Return:
- * The function returns FAIL if there was an error, otherwise SUCEED
+ * The function returns FAIL if there was an error, otherwise SUCCEED
*
* Parameters Description:
* h5tools_str_t *buffer is the string into which to render
@@ -500,7 +500,7 @@ h5tools_print_region_data_blocks(hid_t region_id, FILE *stream, const h5tool_for
hsize_t * count = NULL;
hsize_t blkndx;
hsize_t total_size[H5S_MAX_RANK];
- hsize_t elmtno; /* elemnt index */
+ hsize_t elmtno; /* element index */
hsize_t curr_pos = 0;
unsigned int region_flags; /* buffer extent flags */
hsize_t numelem;
@@ -896,7 +896,7 @@ h5tools_print_region_data_points(hid_t region_space, hid_t region_id, FILE *stre
{
hbool_t dimension_break = TRUE;
hsize_t * dims1 = NULL;
- hsize_t elmtno; /* elemnt index */
+ hsize_t elmtno; /* element index */
hsize_t curr_pos = 0;
hsize_t total_size[H5S_MAX_RANK];
hsize_t jndx;
@@ -1286,7 +1286,7 @@ h5tools_print_simple_subset(FILE *stream, const h5tool_format_t *info, h5tools_c
H5TOOLS_START_DEBUG(" ");
if ((size_t)ctx->ndims > NELMTS(sm_size))
- H5TOOLS_THROW(FAIL, "ndims and sm_size comparision failed");
+ H5TOOLS_THROW(FAIL, "ndims and sm_size comparison failed");
size_row_block = ctx->sset->block.data[row_dim];
@@ -3340,7 +3340,7 @@ h5tools_dump_dcpl(FILE *stream, const h5tool_format_t *info, h5tools_context_t *
} break;
case H5D_VIRTUAL: {
- char dsetname[256]; /* virtual datset name */
+ char dsetname[256]; /* virtual dataset name */
size_t n_vmaps;
H5Pget_virtual_count(dcpl_id, &n_vmaps);
diff --git a/tools/lib/h5tools_ref.c b/tools/lib/h5tools_ref.c
index 50f283d..1c97a52 100644
--- a/tools/lib/h5tools_ref.c
+++ b/tools/lib/h5tools_ref.c
@@ -70,7 +70,7 @@ free_ref_path_info(void *item, void H5_ATTR_UNUSED *key, void H5_ATTR_UNUSED *op
/*-------------------------------------------------------------------------
* Function: init_ref_path_cb
*
- * Purpose: Called by interator to create references for
+ * Purpose: Called by iterator to create references for
* all objects and enter them in the table.
*
* Return: Error status.
@@ -123,7 +123,7 @@ ref_path_table_cmp(const void *key1, const void *key2)
/*-------------------------------------------------------------------------
* Function: init_ref_path_table
*
- * Purpose: Initalize the reference path table
+ * Purpose: Initialize the reference path table
*
* Return: Non-negative on success, negative on failure
*
@@ -377,7 +377,7 @@ lookup_ref_path(H5R_ref_t refbuf)
/*-------------------------------------------------------------------------
* Function: fill_ref_path_table
*
- * Purpose: Called by interator to create references for
+ * Purpose: Called by iterator to create references for
* all objects and enter them in the table.
*
* Return: Error status.
diff --git a/tools/lib/h5tools_str.c b/tools/lib/h5tools_str.c
index 3cd12bb..4e8dcc1 100644
--- a/tools/lib/h5tools_str.c
+++ b/tools/lib/h5tools_str.c
@@ -129,7 +129,7 @@ h5tools_str_append(h5tools_str_t *str /*in,out*/, const char *fmt, ...)
/* Note: HDvsnprintf() behaves differently on Windows as Unix, when
* buffer is smaller than source string. On Unix, this function
- * returns length of the source string and copy string upto the
+ * returns length of the source string and copy string up to the
* buffer size with NULL at the end of the buffer. However on
* Windows with the same condition, this function returns -1 and
* doesn't add NULL at the end of the buffer.
@@ -147,7 +147,7 @@ h5tools_str_append(h5tools_str_t *str /*in,out*/, const char *fmt, ...)
* following conditions, each of which indicates that the proper C99 return value probably
* should have been positive when the format string is
* something other than "%s"
- * Alocate at least twice as much space and try again.
+ * Allocate at least twice as much space and try again.
*/
size_t newsize = MAX(str->len + (size_t)nchars + 1, 2 * str->nalloc);
HDassert(newsize > str->nalloc); /*overflow*/
@@ -1246,7 +1246,7 @@ h5tools_str_sprint(h5tools_str_t *str, const h5tool_format_t *info, hid_t contai
/* Print the closing bracket */
h5tools_str_append(str, "%s", OPT(info->arr_suf, "]"));
- is_next_arry_elmt = 1; /* set for begining of next array element */
+ is_next_arry_elmt = 1; /* set for beginning of next array element */
H5Tclose(memb);
} break;
@@ -1504,7 +1504,7 @@ h5tools_str_is_zero(const void *_mem, size_t size)
* March 8, 2012
*
* Notes:
- * Applications need to call free() to free the memoery allocated for
+ * Applications need to call free() to free the memory allocated for
* the return string
*
*-------------------------------------------------------------------------
diff --git a/tools/lib/h5tools_utils.c b/tools/lib/h5tools_utils.c
index ea9812b..66d3ef6 100644
--- a/tools/lib/h5tools_utils.c
+++ b/tools/lib/h5tools_utils.c
@@ -180,7 +180,7 @@ help_ref_msg(FILE *output)
* * Last two characters in the string MUST be ")\0".
*
* * Generates a copy of the input string `start`, (src..")\0"), replacing
- * separators and close-paren with null charaters.
+ * separators and close-paren with null characters.
* * This string is allocated at runtime and should be freed when done.
* * Generates array of char pointers, and directs start of each element
* (each pointer) into this copy.
@@ -227,7 +227,7 @@ parse_tuple(const char *start, int sep, char **cpy_out, unsigned *nelems, char *
char * elem_ptr = NULL;
char * dest_ptr = NULL;
unsigned elems_count = 0;
- char ** elems = NULL; /* more like *elems[], but complier... */
+ char ** elems = NULL; /* more like *elems[], but compiler... */
char ** elems_re = NULL; /* temporary pointer, for realloc */
char * cpy = NULL;
herr_t ret_value = SUCCEED;
@@ -337,7 +337,7 @@ parse_tuple(const char *start, int sep, char **cpy_out, unsigned *nelems, char *
/* Found terminal, non-escaped close-paren. Last element.
* Write null terminator to copy.
* Advance source pointer to gently break from loop.
- * Requred to prevent ")" from always being added to last element.
+ * Required to prevent ")" from always being added to last element.
*/
start++;
}
diff --git a/tools/lib/h5tools_utils.h b/tools/lib/h5tools_utils.h
index 095ad6c..3733742 100644
--- a/tools/lib/h5tools_utils.h
+++ b/tools/lib/h5tools_utils.h
@@ -100,7 +100,7 @@ typedef enum toolname_t {
TOOL__H5DUMP /* add as necessary */
} h5tool_toolname_t;
-/* this struct can be used to differntiate among tools */
+/* this struct can be used to differentiate among tools */
typedef struct {
h5tool_toolname_t toolname;
int msg_mode;
diff --git a/tools/lib/h5trav.c b/tools/lib/h5trav.c
index d0c299f..1f4b0d6 100644
--- a/tools/lib/h5trav.c
+++ b/tools/lib/h5trav.c
@@ -534,7 +534,7 @@ trav_info_free(trav_info_t *info)
/*-------------------------------------------------------------------------
* Function: trav_table_visit_obj
*
- * Purpose: Callback for visiting object, with 'table' sructure
+ * Purpose: Callback for visiting object, with 'table' structure
*
* Return: 0 on success,
* -1 on failure
@@ -559,7 +559,7 @@ trav_table_visit_obj(const char *path, const H5O_info2_t *oinfo, const char *alr
/*-------------------------------------------------------------------------
* Function: trav_table_visit_lnk
*
- * Purpose: Callback for visiting link, with 'table' sructure
+ * Purpose: Callback for visiting link, with 'table' structure
*
* Return: 0 on success,
* -1 on failure
@@ -1127,7 +1127,7 @@ symlink_is_visited(symlink_trav_t *visited, H5L_type_t type, const char *file, c
/* Look for symlink */
for (u = 0; u < visited->nused; u++) {
/* Check for symlink values already in array */
- /* check type and path pair to distingush between symbolic links */
+ /* check type and path pair to distinguish between symbolic links */
if ((visited->objs[u].type == type) && !HDstrcmp(visited->objs[u].path, path)) {
/* if external link, file need to be matched as well */
if (visited->objs[u].type == H5L_TYPE_EXTERNAL)
diff --git a/tools/lib/h5trav.h b/tools/lib/h5trav.h
index 7a8a63d..58d347d 100644
--- a/tools/lib/h5trav.h
+++ b/tools/lib/h5trav.h
@@ -87,7 +87,7 @@ typedef struct trav_link_t {
} trav_link_t;
/*-------------------------------------------------------------------------
- * struct to store basic info needed for the h5trav table traversal algorythm
+ * struct to store basic info needed for the h5trav table traversal algorithm
*-------------------------------------------------------------------------
*/
diff --git a/tools/lib/io_timer.c b/tools/lib/io_timer.c
index aec7e9b..ce21b3e 100644
--- a/tools/lib/io_timer.c
+++ b/tools/lib/io_timer.c
@@ -11,7 +11,7 @@
* * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * */
/* changes:
* rename pio_timer.c as io_timer.c;
- * Removed pio_perf.h so that it is not dependant on it;
+ * Removed pio_perf.h so that it is not dependent on it;
* Removed set_timer_type() and get_timer_type() since no one calls them;
* Merged sio_timer.c into io_timer.c;
*/
diff --git a/tools/libtest/h5tools_test_utils.c b/tools/libtest/h5tools_test_utils.c
index 1a15a47..072179b 100644
--- a/tools/libtest/h5tools_test_utils.c
+++ b/tools/libtest/h5tools_test_utils.c
@@ -431,7 +431,7 @@ test_parse_tuple(void)
{"stuff keeps on going"},
},
{
- "4-ple, escaped seperator",
+ "4-ple, escaped separator",
"(elem0,elem1,el\\,em2,elem3)", /* "el\,em" */
',',
SUCCEED,
@@ -701,7 +701,7 @@ test_populate_ros3_fa(void)
}
/* NULL region
- * yeilds default fapl
+ * yields default fapl
*/
{
H5FD_ros3_fapl_t fa = {bad_version, FALSE, "a", "b", "c"};
@@ -720,7 +720,7 @@ test_populate_ros3_fa(void)
}
/* empty region
- * yeilds default fapl
+ * yields default fapl
*/
{
H5FD_ros3_fapl_t fa = {bad_version, FALSE, "a", "b", "c"};
@@ -739,7 +739,7 @@ test_populate_ros3_fa(void)
}
/* region overflow
- * yeilds default fapl
+ * yields default fapl
*/
{
H5FD_ros3_fapl_t fa = {bad_version, FALSE, "a", "b", "c"};
@@ -781,7 +781,7 @@ test_populate_ros3_fa(void)
}
/* empty id (non-empty region, key)
- * yeilds default fapl
+ * yields default fapl
*/
{
H5FD_ros3_fapl_t fa = {bad_version, FALSE, "a", "b", "c"};
@@ -851,7 +851,7 @@ test_populate_ros3_fa(void)
}
/* empty key (non-empty region, id)
- * yeilds authenticating fapl
+ * yields authenticating fapl
*/
{
H5FD_ros3_fapl_t fa = {bad_version, FALSE, "a", "b", "c"};
@@ -870,7 +870,7 @@ test_populate_ros3_fa(void)
}
/* empty key, region (non-empty id)
- * yeilds default fapl
+ * yields default fapl
*/
{
H5FD_ros3_fapl_t fa = {bad_version, FALSE, "a", "b", "c"};
@@ -889,7 +889,7 @@ test_populate_ros3_fa(void)
}
/* empty key, id (non-empty region)
- * yeilds default fapl
+ * yields default fapl
*/
{
H5FD_ros3_fapl_t fa = {bad_version, FALSE, "a", "b", "c"};
diff --git a/tools/src/h5diff/h5diff_common.c b/tools/src/h5diff/h5diff_common.c
index 3d89936..ec41d21 100644
--- a/tools/src/h5diff/h5diff_common.c
+++ b/tools/src/h5diff/h5diff_common.c
@@ -283,7 +283,7 @@ parse_command_line(int argc, const char *argv[], const char **fname1, const char
break;
case 'q':
- /* use quiet mode; supress the message "0 differences found" */
+ /* use quiet mode; suppress the message "0 differences found" */
opts->mode_quiet = 1;
break;
@@ -834,7 +834,7 @@ usage(void)
PRINTVALSTREAM(rawoutstream, "\n");
PRINTVALSTREAM(rawoutstream, " Modes of output:\n");
PRINTVALSTREAM(rawoutstream,
- " Default mode: print the number of differences found and where they occured\n");
+ " Default mode: print the number of differences found and where they occurred\n");
PRINTVALSTREAM(rawoutstream, " -r Report mode: print the above plus the differences\n");
PRINTVALSTREAM(rawoutstream, " -v Verbose mode: print the above plus a list of objects and warnings\n");
PRINTVALSTREAM(rawoutstream, " -q Quiet mode: do not print output\n");
diff --git a/tools/src/h5diff/h5diff_main.c b/tools/src/h5diff/h5diff_main.c
index a835953..37c215b 100644
--- a/tools/src/h5diff/h5diff_main.c
+++ b/tools/src/h5diff/h5diff_main.c
@@ -33,7 +33,7 @@
*
* Modifications: July 2004
* Introduced the four modes:
- * Normal mode: print the number of differences found and where they occured
+ * Normal mode: print the number of differences found and where they occurred
* Report mode: print the above plus the differences
* Verbose mode: print the above plus a list of objects and warnings
* Quiet mode: do not print output
diff --git a/tools/src/h5dump/h5dump.c b/tools/src/h5dump/h5dump.c
index 23498bd..a57b0e3 100644
--- a/tools/src/h5dump/h5dump.c
+++ b/tools/src/h5dump/h5dump.c
@@ -1474,7 +1474,7 @@ main(int argc, const char *argv[])
dset_table = table_list.tables[0].dset_table;
type_table = table_list.tables[0].type_table;
- /* does there exist unamed committed datatype */
+ /* does there exist unnamed committed datatype */
for (u = 0; u < type_table->nobjs; u++)
if (!type_table->objs[u].recorded) {
unamedtype = 1;
diff --git a/tools/src/h5dump/h5dump_ddl.c b/tools/src/h5dump/h5dump_ddl.c
index 638a738..8002ec2 100644
--- a/tools/src/h5dump/h5dump_ddl.c
+++ b/tools/src/h5dump/h5dump_ddl.c
@@ -829,7 +829,7 @@ dump_group(hid_t gid, const char *name)
if (!HDstrcmp(name, "/") && unamedtype) {
unsigned u; /* Local index variable */
- /* dump unamed type in root group */
+ /* dump unnamed type in root group */
for (u = 0; u < type_table->nobjs; u++)
if (!type_table->objs[u].recorded) {
char *obj_tok_str = NULL;
@@ -1283,7 +1283,7 @@ dump_fcontents(hid_t fid)
{
PRINTSTREAM(rawoutstream, "%s %s\n", FILE_CONTENTS, BEGIN);
- /* special case of unamed types in root group */
+ /* special case of unnamed types in root group */
if (unamedtype) {
unsigned u;
@@ -1959,7 +1959,7 @@ handle_datatypes(hid_t fid, const char *type, void H5_ATTR_UNUSED *data, int pe,
const char *real_name = display_name ? display_name : type;
if ((type_id = H5Topen2(fid, type, H5P_DEFAULT)) < 0) {
- /* check if type is unamed datatype */
+ /* check if type is unnamed datatype */
unsigned idx = 0;
while (idx < type_table->nobjs) {
@@ -1968,7 +1968,7 @@ handle_datatypes(hid_t fid, const char *type, void H5_ATTR_UNUSED *data, int pe,
if (!type_table->objs[idx].recorded) {
char *obj_tok_str = NULL;
- /* unamed datatype */
+ /* unnamed datatype */
H5Otoken_to_str(fid, &type_table->objs[idx].obj_token, &obj_tok_str);
HDsprintf(name, "/#%s", obj_tok_str);
H5free_memory(obj_tok_str);
diff --git a/tools/src/h5dump/h5dump_xml.c b/tools/src/h5dump/h5dump_xml.c
index 0e881df..8fec49c 100644
--- a/tools/src/h5dump/h5dump_xml.c
+++ b/tools/src/h5dump/h5dump_xml.c
@@ -23,7 +23,7 @@
const char *xmlnsprefix = "hdf5:";
/*
- * Alternative formating for data dumped to XML
+ * Alternative formatting for data dumped to XML
* In general, the numbers are the same, but separators
* except spaces are not used.
*
@@ -2793,7 +2793,7 @@ xml_dump_group(hid_t gid, const char *name)
if (isRoot && unamedtype) {
unsigned u;
- /* Very special case: dump unamed type in root group */
+ /* Very special case: dump unnamed type in root group */
for (u = 0; u < type_table->nobjs; u++) {
if (!type_table->objs[u].recorded) {
char *obj_tok_str = NULL;
@@ -2886,7 +2886,7 @@ xml_dump_group(hid_t gid, const char *name)
if (isRoot && unamedtype) {
unsigned u;
- /* Very special case: dump unamed type in root group */
+ /* Very special case: dump unnamed type in root group */
for (u = 0; u < type_table->nobjs; u++) {
if (!type_table->objs[u].recorded) {
char *obj_tok_str = NULL;
@@ -2938,7 +2938,7 @@ xml_dump_group(hid_t gid, const char *name)
/*-------------------------------------------------------------------------
* Function: xml_print_refs
*
- * Purpose: Print a path to the objects referenced by HDF5 Referneces.
+ * Purpose: Print a path to the objects referenced by HDF5 References.
*
* Return: void
*
diff --git a/tools/src/h5import/h5import.c b/tools/src/h5import/h5import.c
index 7272dff..31fa7cf 100644
--- a/tools/src/h5import/h5import.c
+++ b/tools/src/h5import/h5import.c
@@ -2126,7 +2126,7 @@ processConfigurationFile(char *infile, struct Input *in)
}
} /* while (get_next_prop) */
} /* else if(!HDstrcmp("SUBSET", key)) */
- else if (!HDstrcmp("DATA", key)) { /* FINSHED */
+ else if (!HDstrcmp("DATA", key)) { /* FINISHED */
#ifdef H5DEBUGIMPORT
HDprintf("h5dump DATA key\n");
#endif
@@ -2320,7 +2320,7 @@ processConfigurationFile(char *infile, struct Input *in)
(void)HDfprintf(stderr, err12a, infile);
goto error;
}
- /* cant appear before dimension sizes have been provided */
+ /* can't appear before dimension sizes have been provided */
if (in->configOptionVector[DIM] == 0) {
(void)HDfprintf(stderr, err12b, infile);
goto error;
@@ -2387,7 +2387,7 @@ processConfigurationFile(char *infile, struct Input *in)
(void)HDfprintf(stderr, err16a, infile);
goto error;
}
- /* cant appear before dimension sizes have been provided */
+ /* can't appear before dimension sizes have been provided */
if (in->configOptionVector[DIM] == 0) {
(void)HDfprintf(stderr, err16b, infile);
goto error;
@@ -2474,14 +2474,14 @@ validateConfigurationParameters(struct Input *in)
}
}
- /* Arch cant be STD if O/p class is FP */
+ /* Arch can't be STD if O/p class is FP */
if (in->outputArchitecture == 1)
if (in->outputClass == 1) {
(void)HDfprintf(stderr, "%s", err4a);
return (-1);
}
- /* Arch cant be IEEE if O/p class is IN */
+ /* Arch can't be IEEE if O/p class is IN */
if (in->outputArchitecture == 2)
if (in->outputClass == 0) {
(void)HDfprintf(stderr, "%s", err4b);
@@ -3725,7 +3725,7 @@ getCompressionParameter(struct Input *in, FILE *strm)
int ival;
const char *err1 = "Unable to get integer value.\n";
- const char *err2 = "Invalid value for compression paramter.\n";
+ const char *err2 = "Invalid value for compression parameter.\n";
const char *err3 = "Unsupported Compression Type.\n";
switch (in->compressionType) {
@@ -4689,7 +4689,7 @@ process(struct Options *opt)
}
if (in->configOptionVector[EXTERNALSTORE] == 1) {
- /* creating the external file if it doesnt exist */
+ /* creating the external file if it doesn't exist */
if ((extfile = HDfopen(in->externFilename, "ab")) == NULL) {
(void)HDfprintf(stderr, "%s", err4);
H5Pclose(proplist);
diff --git a/tools/src/h5jam/h5jam.c b/tools/src/h5jam/h5jam.c
index 7f3385c..07140b5 100644
--- a/tools/src/h5jam/h5jam.c
+++ b/tools/src/h5jam/h5jam.c
@@ -358,7 +358,7 @@ done:
* Purpose: Copy part of the input file to output.
* infid: fd of file to read
* outfid: fd of file to write
- * startin: offset of where to read from infid
+ * starting: offset of where to read from infid
* startout: offset of where to write to outfid
* limit: bytes to read/write
*
@@ -373,7 +373,7 @@ done:
*-------------------------------------------------------------------------
*/
hsize_t
-copy_some_to_file(int infid, int outfid, hsize_t startin, hsize_t startout, ssize_t limit)
+copy_some_to_file(int infid, int outfid, hsize_t starting, hsize_t startout, ssize_t limit)
{
char buf[1024];
h5_stat_t sbuf;
@@ -386,9 +386,9 @@ copy_some_to_file(int infid, int outfid, hsize_t startin, hsize_t startout, ssiz
ssize_t toend;
ssize_t fromend;
- if (startin > startout) {
+ if (starting > startout) {
/* this case is prohibited */
- error_msg("copy_some_to_file: panic: startin > startout?\n");
+ error_msg("copy_some_to_file: panic: starting > startout?\n");
exit(EXIT_FAILURE);
} /* end if */
@@ -409,7 +409,7 @@ copy_some_to_file(int infid, int outfid, hsize_t startin, hsize_t startout, ssiz
return 0;
toend = (ssize_t)startout + howmuch;
- fromend = (ssize_t)startin + howmuch;
+ fromend = (ssize_t)starting + howmuch;
if (howmuch > 512) {
to = toend - 512;
diff --git a/tools/src/h5ls/h5ls.c b/tools/src/h5ls/h5ls.c
index 2ec8da2..17abf43 100644
--- a/tools/src/h5ls/h5ls.c
+++ b/tools/src/h5ls/h5ls.c
@@ -32,7 +32,7 @@
#define NAME_BUF_SIZE 2048
/*
- * Alternative formating for data dumped by H5LS
+ * Alternative formatting for data dumped by H5LS
*
* This table only affects H5LS output.
*/
diff --git a/tools/src/h5perf/pio_engine.c b/tools/src/h5perf/pio_engine.c
index f890c7d..614943b 100644
--- a/tools/src/h5perf/pio_engine.c
+++ b/tools/src/h5perf/pio_engine.c
@@ -244,7 +244,7 @@ do_pio(parameters param)
if (((size_t)(snbytes / pio_mpi_nprocs_g) % buf_size) != 0) {
HDfprintf(stderr,
"Dataset size/process (%" H5_PRINTF_LL_WIDTH "d) must be a multiple of the "
- "trasfer buffer size (%zu)\n",
+ "transfer buffer size (%zu)\n",
(long long)(snbytes / pio_mpi_nprocs_g), buf_size);
GOTOERROR(FAIL);
}
@@ -253,7 +253,7 @@ do_pio(parameters param)
if (((size_t)snbytes % buf_size) != 0) {
HDfprintf(stderr,
"Dataset side size (%" H5_PRINTF_LL_WIDTH "d) must be a multiple of the "
- "trasfer buffer size (%zu)\n",
+ "transfer buffer size (%zu)\n",
(long long)snbytes, buf_size);
GOTOERROR(FAIL);
}
diff --git a/tools/src/h5perf/pio_perf.c b/tools/src/h5perf/pio_perf.c
index 41245d9..6478fe7 100644
--- a/tools/src/h5perf/pio_perf.c
+++ b/tools/src/h5perf/pio_perf.c
@@ -579,7 +579,7 @@ run_test(iotype iot, parameters parms, struct options *opts)
output_results(opts, "Raw Data Write", write_raw_mm_table, parms.num_iters, raw_size);
} /* end if */
- /* show mpi write statics */
+ /* show mpi write statistics */
if (pio_debug_level >= 3) {
/* output all of the times for all iterations */
print_indent(3);
@@ -645,7 +645,7 @@ run_test(iotype iot, parameters parms, struct options *opts)
output_results(opts, "Raw Data Read", read_raw_mm_table, parms.num_iters, raw_size);
} /* end if */
- /* show mpi read statics */
+ /* show mpi read statistics */
if (pio_debug_level >= 3) {
/* output all of the times for all iterations */
print_indent(3);
diff --git a/tools/src/h5perf/sio_engine.c b/tools/src/h5perf/sio_engine.c
index 3ebacc0..98cb920 100644
--- a/tools/src/h5perf/sio_engine.c
+++ b/tools/src/h5perf/sio_engine.c
@@ -179,7 +179,7 @@ do_sio(parameters param, results *res)
if ((param.dset_size[i] % param.buf_size[i]) != 0) {
HDfprintf(stderr,
"Dataset size[%d] (%" H5_PRINTF_LL_WIDTH "d) must be a multiple of the "
- "trasfer buffer size[%d] (%zu)\n",
+ "transfer buffer size[%d] (%zu)\n",
param.rank, (long long)param.dset_size[i], param.rank, param.buf_size[i]);
GOTOERROR(FAIL);
}
diff --git a/tools/src/h5perf/sio_perf.c b/tools/src/h5perf/sio_perf.c
index 9c7675f..0eb4508 100644
--- a/tools/src/h5perf/sio_perf.c
+++ b/tools/src/h5perf/sio_perf.c
@@ -427,7 +427,7 @@ run_test(iotype iot, parameters parms, struct options *opts)
output_results(opts, "Raw Data Write", write_raw_mm_table, parms.num_iters, raw_size);
} /* end if */
- /* show sys write statics */
+ /* show sys write statistics */
#if 0
if (sio_debug_level >= 3) {
/* output all of the times for all iterations */
@@ -473,7 +473,7 @@ run_test(iotype iot, parameters parms, struct options *opts)
output_results(opts, "Raw Data Read", read_raw_mm_table, parms.num_iters, raw_size);
} /* end if */
- /* show mpi read statics */
+ /* show mpi read statistics */
#if 0
if (sio_debug_level >= 3) {
/* output all of the times for all iterations */
@@ -1278,7 +1278,7 @@ usage(const char *prog)
HDprintf("\n");
HDprintf(" Example: 2K,2K,3K\n");
HDprintf("\n");
- HDprintf(" The example defines an object (dataset, tranfer buffer) with three\n");
+ HDprintf(" The example defines an object (dataset, transfer buffer) with three\n");
HDprintf(" dimensions. Be aware that as the number of dimensions increases, the\n");
HDprintf(" the total size of the object increases exponentially.\n");
HDprintf("\n");
diff --git a/tools/src/h5repack/h5repack.h b/tools/src/h5repack/h5repack.h
index b242a71..65e5d40 100644
--- a/tools/src/h5repack/h5repack.h
+++ b/tools/src/h5repack/h5repack.h
@@ -124,7 +124,7 @@ typedef struct {
int grp_compact; /* Set the maximum number of links to store as header messages in the group */
int grp_indexed; /* Set the minimum number of links to store in the indexed format */
int msg_size[8]; /* Minimum size of shared messages: dataspace,
- datatype, fill value, filter pipleline, attribute */
+ datatype, fill value, filter pipeline, attribute */
const char * ublock_filename; /* user block file name */
hsize_t ublock_size; /* user block size */
hsize_t meta_block_size; /* metadata aggregation block size (for H5Pset_meta_block_size) */
diff --git a/tools/src/h5repack/h5repack_copy.c b/tools/src/h5repack/h5repack_copy.c
index 0497867..c38a431 100644
--- a/tools/src/h5repack/h5repack_copy.c
+++ b/tools/src/h5repack/h5repack_copy.c
@@ -67,7 +67,7 @@ copy_objects(const char *fnamein, const char *fnameout, pack_opt_t *options)
hid_t fcpl = H5P_DEFAULT; /* file creation property list ID */
trav_table_t * travt = NULL;
hsize_t ub_size = 0; /* size of user block */
- H5F_fspace_strategy_t set_strategy; /* Strategy to be set in outupt file */
+ H5F_fspace_strategy_t set_strategy; /* Strategy to be set in output file */
hbool_t set_persist; /* Persist free-space status to be set in output file */
hsize_t set_threshold; /* Free-space section threshold to be set in output file */
hsize_t set_pagesize; /* File space page size to be set in output file */
@@ -188,7 +188,7 @@ copy_objects(const char *fnamein, const char *fnameout, pack_opt_t *options)
H5TOOLS_GOTO_ERROR((-1), "H5Pset_shared_mesg_nindexes failed to set the number of shared "
"object header message indexes");
- /* msg_size[0]=dataspace, 1=datatype, 2=file value, 3=filter pipleline, 4=attribute */
+ /* msg_size[0]=dataspace, 1=datatype, 2=file value, 3=filter pipeline, 4=attribute */
for (i = 0; i < (nindex - 1); i++)
if (H5Pset_shared_mesg_index(fcpl, i, mesg_type_flags[i], min_mesg_sizes[i]) < 0)
H5TOOLS_GOTO_ERROR((-1), "H5Pset_shared_mesg_index failed to configure the specified "
@@ -402,18 +402,18 @@ done:
/*-------------------------------------------------------------------------
* Function: get_hyperslab
*
- * Purpose: Calulate a hyperslab from a dataset for higher performance.
- * The size of hyperslab is limitted by H5TOOLS_BUFSIZE.
- * Return the hyperslab dimentions and size in byte.
+ * Purpose: Calculate a hyperslab from a dataset for higher performance.
+ * The size of hyperslab is limited by H5TOOLS_BUFSIZE.
+ * Return the hyperslab dimensions and size in byte.
*
* Return: 0 - SUCCEED, -1 FAILED
*
* Parameters:
* dcpl_id : [IN] dataset creation property.
* rank_dset : [IN] dataset rank
- * dims_dset[] : [IN] dataset dimentions
+ * dims_dset[] : [IN] dataset dimensions
* size_datum : [IN] size of a data element in byte
- * dims_hslab[] : [OUT] calculated hyperslab dimentions
+ * dims_hslab[] : [OUT] calculated hyperslab dimensions
* * hslab_nbytes_p : [OUT] total byte of the hyperslab
*
* Update:
@@ -428,7 +428,7 @@ done:
* 3. If not chunked, each data element would be a unit of collection and
* the boundary would be dataset's dims.
*
- * The calulation starts from the last dimention (h5dump dims output).
+ * The calculation starts from the last dimension (h5dump dims output).
*-----------------------------------------*/
int
@@ -442,8 +442,8 @@ get_hyperslab(hid_t dcpl_id, int rank_dset, const hsize_t dims_dset[], size_t si
hsize_t size_chunk = 1;
hsize_t nchunk_fit; /* number of chunks that fits in hyperslab buffer (H5TOOLS_BUFSIZE) */
hsize_t ndatum_fit; /* number of dataum that fits in hyperslab buffer (H5TOOLS_BUFSIZE) */
- hsize_t chunk_dims_map[H5S_MAX_RANK]; /* mapped chunk dimentions */
- hsize_t hs_dims_map[H5S_MAX_RANK]; /* mapped hyperslab dimentions */
+ hsize_t chunk_dims_map[H5S_MAX_RANK]; /* mapped chunk dimensions */
+ hsize_t hs_dims_map[H5S_MAX_RANK]; /* mapped hyperslab dimensions */
hsize_t hslab_nbytes; /* size of hyperslab in byte */
int ret_value = 0;
@@ -468,18 +468,18 @@ get_hyperslab(hid_t dcpl_id, int rank_dset, const hsize_t dims_dset[], size_t si
/* 1. if a chunk fit in hyperslab buffer */
if (nchunk_fit >= 1) {
- /* Calulate a hyperslab that contains as many chunks that can fit
+ /* Calculate a hyperslab that contains as many chunks that can fit
* in hyperslab buffer. Hyperslab will be increased starting from
- * the last dimention of the dataset (see h5dump's dims output).
+ * the last dimension of the dataset (see h5dump's dims output).
* The calculation boundary is dataset dims.
* In the loop, used mapping from a datum to a chunk to figure out
* chunk based hyperslab.
*/
for (k = rank_dset; k > 0; --k) {
- /* map dataset dimentions with a chunk dims */
+ /* map dataset dimensions with a chunk dims */
chunk_dims_map[k - 1] = dims_dset[k - 1] / dims_chunk[k - 1];
- /* if reminder exist, increse by 1 to cover partial edge chunks */
+ /* if reminder exist, increase by 1 to cover partial edge chunks */
if (dims_dset[k - 1] % dims_chunk[k - 1] > 0)
chunk_dims_map[k - 1]++;
@@ -492,7 +492,7 @@ get_hyperslab(hid_t dcpl_id, int rank_dset, const hsize_t dims_dset[], size_t si
if (nchunk_fit == 0)
nchunk_fit = 1;
- /* get hyperslab dimentions as unmapping to actual size */
+ /* get hyperslab dimensions as unmapping to actual size */
dims_hslab[k - 1] = MIN((hs_dims_map[k - 1] * dims_chunk[k - 1]), dims_dset[k - 1]);
/* calculate total size for the hyperslab */
@@ -501,9 +501,9 @@ get_hyperslab(hid_t dcpl_id, int rank_dset, const hsize_t dims_dset[], size_t si
}
/* 2. if a chunk is bigger than hyperslab buffer */
else {
- /* Calulate a hyperslab that contains as many data elements that
+ /* Calculate a hyperslab that contains as many data elements that
* can fit in hyperslab buffer. Hyperslab will be increased
- * starting from the last dimention of the chunk (see h5dump's dims
+ * starting from the last dimension of the chunk (see h5dump's dims
* output).
* The calculation boundary is a chunk dims.
*/
@@ -513,7 +513,7 @@ get_hyperslab(hid_t dcpl_id, int rank_dset, const hsize_t dims_dset[], size_t si
/* if a datum is bigger than rest of buffer */
if (ndatum_fit == 0)
ndatum_fit = 1;
- /* get hyperslab dimentions within a chunk boundary */
+ /* get hyperslab dimensions within a chunk boundary */
dims_hslab[k - 1] = MIN(dims_chunk[k - 1], ndatum_fit);
/* calculate total size for the hyperslab */
@@ -526,9 +526,9 @@ get_hyperslab(hid_t dcpl_id, int rank_dset, const hsize_t dims_dset[], size_t si
}
/* 3. if dataset is not chunked */
else {
- /* Calulate a hyperslab that contains as many data elements that can
+ /* Calculate a hyperslab that contains as many data elements that can
* fit in hyperslab buffer. Hyperslab will be increased starting from
- * the last dimention of the dataset (see h5dump's dims output).
+ * the last dimension of the dataset (see h5dump's dims output).
* The calculation boundary is dataset dims.
*/
for (k = rank_dset; k > 0; --k) {
@@ -537,7 +537,7 @@ get_hyperslab(hid_t dcpl_id, int rank_dset, const hsize_t dims_dset[], size_t si
/* if a datum is bigger than rest of buffer */
if (ndatum_fit == 0)
ndatum_fit = 1;
- /* get hyperslab dimentions within dataset boundary */
+ /* get hyperslab dimensions within dataset boundary */
dims_hslab[k - 1] = MIN(dims_dset[k - 1], ndatum_fit);
/* calculate total size for the hyperslab */
@@ -601,7 +601,7 @@ done:
* in (2) is that, when using the strip mine size, it assures that the "remaining" part
* of the dataset that does not fill an entire strip mine is processed.
*
- * 1. figure out a hyperslab (dimentions) and size (refer to get_hyperslab()).
+ * 1. figure out a hyperslab (dimensions) and size (refer to get_hyperslab()).
* 2. Calculate the hyperslab selections as the selection is moving forward.
* Selection would be same as the hyperslab except for the remaining edge portion
* of the dataset. The code take care of the remaining portion if exist.
@@ -1621,7 +1621,7 @@ copy_user_block(const char *infile, const char *outfile, hsize_t size)
/* Copy the userblock from the input file to the output file */
while (size > 0) {
- ssize_t nread, nbytes; /* # of bytes transfered, etc. */
+ ssize_t nread, nbytes; /* # of bytes transferred, etc. */
char rbuf[USERBLOCK_XFER_SIZE]; /* Buffer for reading */
const char *wbuf; /* Pointer into buffer, for writing */
diff --git a/tools/src/h5repack/h5repack_filters.c b/tools/src/h5repack/h5repack_filters.c
index 32476ce..76bac5e 100644
--- a/tools/src/h5repack/h5repack_filters.c
+++ b/tools/src/h5repack/h5repack_filters.c
@@ -231,7 +231,7 @@ aux_assign_obj(const char * name, /* object name from traverse list */
* do extra checking in the case of SZIP; delete all filters in the case
* of H5Z_FILTER_NONE present in the PACK_INFO_T filter array
*
- * Return: 0 success, -1 an error occured
+ * Return: 0 success, -1 an error occurred
*-------------------------------------------------------------------------
*/
@@ -258,7 +258,7 @@ apply_filters(const char * name, /* object name from traverse list */
H5TOOLS_GOTO_DONE(0);
/*-------------------------------------------------------------------------
- * initialize the assigment object
+ * initialize the assignment object
*-------------------------------------------------------------------------
*/
init_packobject(&obj);
diff --git a/tools/src/h5repack/h5repack_main.c b/tools/src/h5repack/h5repack_main.c
index 7c3d229..0a26500 100644
--- a/tools/src/h5repack/h5repack_main.c
+++ b/tools/src/h5repack/h5repack_main.c
@@ -39,13 +39,13 @@ static struct h5_long_options l_opts[] = {{"alignment", require_arg, 'a'},
{"file", require_arg, 'e'},
{"filter", require_arg, 'f'},
{"help", no_arg, 'h'},
- {"infile", require_arg, 'i'}, /* for backward compability */
+ {"infile", require_arg, 'i'}, /* for backward compatibility */
{"low", require_arg, 'j'},
{"high", require_arg, 'k'},
{"layout", require_arg, 'l'},
{"minimum", require_arg, 'm'},
{"native", no_arg, 'n'},
- {"outfile", require_arg, 'o'}, /* for backward compability */
+ {"outfile", require_arg, 'o'}, /* for backward compatibility */
{"sort_by", require_arg, 'q'},
{"ssize", require_arg, 's'},
{"threshold", require_arg, 't'},
@@ -399,7 +399,7 @@ read_info(const char *filename, pack_opt_t *options)
goto done;
}
- /* find begining of info */
+ /* find beginning of info */
i = 0;
c = '0';
while (c != ' ') {
diff --git a/tools/src/h5repack/h5repack_refs.c b/tools/src/h5repack/h5repack_refs.c
index 6d62e52..3259b89 100644
--- a/tools/src/h5repack/h5repack_refs.c
+++ b/tools/src/h5repack/h5repack_refs.c
@@ -539,7 +539,7 @@ copy_refs_attr(hid_t loc_in, hid_t loc_out, trav_table_t *travt, hid_t fidout) /
}
/* This line below needs to be moved in this loop instead of inserting outside. Otherwise,
ref_comp_field_n may be >0 for the next attribute, which may not be
- the reference type and will be accidently treated as the reference type.
+ the reference type and will be accidentally treated as the reference type.
It will then cause the H5Acreate2 failed since that attribute is already created.
KY 2020-02-07
*/
diff --git a/tools/src/h5repack/h5repack_verify.c b/tools/src/h5repack/h5repack_verify.c
index 8c54dde..71ee72e 100644
--- a/tools/src/h5repack/h5repack_verify.c
+++ b/tools/src/h5repack/h5repack_verify.c
@@ -397,7 +397,7 @@ h5repack_cmp_pl(const char *fname1, hid_t fname1_fapl, const char *fname2, hid_t
H5TOOLS_GOTO_ERROR((-1), "h5trav_gettable failed");
/*-------------------------------------------------------------------------
- * traverse the suppplied object list
+ * traverse the supplied object list
*-------------------------------------------------------------------------
*/
for (i = 0; i < trav->nobjs; i++) {
diff --git a/tools/src/h5stat/h5stat.c b/tools/src/h5stat/h5stat.c
index d745c0f..6b01e49 100644
--- a/tools/src/h5stat/h5stat.c
+++ b/tools/src/h5stat/h5stat.c
@@ -33,7 +33,7 @@
#define H5_NFILTERS_IMPL \
8 /* Number of currently implemented filters + one to \
accommodate for user-define filters + one \
- to accomodate datasets whithout any filters */
+ to accommodate datasets without any filters */
/* File space management strategies: see H5Fpublic.h for declarations */
const char *FS_STRATEGY_NAME[] = {"H5F_FSPACE_STRATEGY_FSM_AGGR",
diff --git a/tools/src/misc/h5debug.c b/tools/src/misc/h5debug.c
index 72f47cb..e03c72d 100644
--- a/tools/src/misc/h5debug.c
+++ b/tools/src/misc/h5debug.c
@@ -647,7 +647,7 @@ main(int argc, char *argv[])
}
else if (!HDmemcmp(sig, H5EA_HDR_MAGIC, (size_t)H5_SIZEOF_MAGIC)) {
/*
- * Debug an extensible aray header.
+ * Debug an extensible array header.
*/
const H5EA_class_t *cls = get_H5EA_class(sig);
HDassert(cls);
@@ -668,7 +668,7 @@ main(int argc, char *argv[])
}
else if (!HDmemcmp(sig, H5EA_IBLOCK_MAGIC, (size_t)H5_SIZEOF_MAGIC)) {
/*
- * Debug an extensible aray index block.
+ * Debug an extensible array index block.
*/
const H5EA_class_t *cls = get_H5EA_class(sig);
HDassert(cls);
@@ -689,7 +689,7 @@ main(int argc, char *argv[])
}
else if (!HDmemcmp(sig, H5EA_SBLOCK_MAGIC, (size_t)H5_SIZEOF_MAGIC)) {
/*
- * Debug an extensible aray super block.
+ * Debug an extensible array super block.
*/
const H5EA_class_t *cls = get_H5EA_class(sig);
HDassert(cls);
@@ -709,7 +709,7 @@ main(int argc, char *argv[])
}
else if (!HDmemcmp(sig, H5EA_DBLOCK_MAGIC, (size_t)H5_SIZEOF_MAGIC)) {
/*
- * Debug an extensible aray data block.
+ * Debug an extensible array data block.
*/
const H5EA_class_t *cls = get_H5EA_class(sig);
HDassert(cls);
diff --git a/tools/test/h5copy/CMakeTests.cmake b/tools/test/h5copy/CMakeTests.cmake
index e9b36fe..154ca15 100644
--- a/tools/test/h5copy/CMakeTests.cmake
+++ b/tools/test/h5copy/CMakeTests.cmake
@@ -211,7 +211,7 @@
endmacro ()
#
- # Similiar to ADD_H5_TEST macro. Compare to outputs from source & target
+ # Similar to ADD_H5_TEST macro. Compare to outputs from source & target
# files instead of checking with h5ls.
#
macro (ADD_H5_CMP_TEST testname resultcode infile vparam sparam srcname dparam dstname)
diff --git a/tools/test/h5copy/h5copygentest.c b/tools/test/h5copy/h5copygentest.c
index 6898839..c1f8349 100644
--- a/tools/test/h5copy/h5copygentest.c
+++ b/tools/test/h5copy/h5copygentest.c
@@ -618,7 +618,7 @@ gen_region_ref(hid_t loc_id)
goto out;
}
- /* create normal dataset which is refered */
+ /* create normal dataset which is referred */
oid2 = H5Dcreate2(loc_id, REG_REF_DS2, H5T_STD_I8LE, sid, H5P_DEFAULT, H5P_DEFAULT, H5P_DEFAULT);
if (oid2 < 0) {
HDfprintf(stderr, "Error: %s %d> H5Dcreate2 failed.\n", __func__, __LINE__);
diff --git a/tools/test/h5copy/testh5copy.sh.in b/tools/test/h5copy/testh5copy.sh.in
index 79b6fd5..6349ef4 100644
--- a/tools/test/h5copy/testh5copy.sh.in
+++ b/tools/test/h5copy/testh5copy.sh.in
@@ -88,7 +88,7 @@ test -d $TESTDIR || mkdir $TESTDIR
# RUNSERIAL is used. Check if it can return exit code from executalbe correctly.
if [ -n "$RUNSERIAL_NOEXITCODE" ]; then
- echo "***Warning*** Serial Exit Code is not passed back to shell corretly."
+ echo "***Warning*** Serial Exit Code is not passed back to shell correctly."
echo "***Warning*** Exit code checking is skipped."
h5haveexitcode=no
fi
diff --git a/tools/test/h5diff/CMakeTests.cmake b/tools/test/h5diff/CMakeTests.cmake
index 4fcfa26..8e01605 100644
--- a/tools/test/h5diff/CMakeTests.cmake
+++ b/tools/test/h5diff/CMakeTests.cmake
@@ -1278,7 +1278,7 @@ ADD_H5_TEST (h5diff_208 0 -c ${FILE19} ${FILE20})
ADD_H5_TEST (h5diff_220 1 -c non_comparables1.h5 non_comparables2.h5 /g1)
# comparable dataset with non-comparable attribute and other comparable attributes.
-# Also test non-compatible attributes with different type, dimention, rank.
+# Also test non-compatible attributes with different type, dimension, rank.
# All the rest comparables should display differences.
ADD_H5_TEST (h5diff_221 1 -c non_comparables1.h5 non_comparables2.h5 /g2)
@@ -1438,7 +1438,7 @@ ADD_H5_TEST (h5diff_475 1 -v h5diff_danglelinks1.h5 h5diff_danglelinks2.h5 /ext_
# ##############################################################################
-# # test for group diff recursivly
+# # test for group diff recursively
# ##############################################################################
# root
ADD_H5_TEST (h5diff_500 1 -v ${GRP_RECURSE_FILE1} ${GRP_RECURSE_FILE2} / /)
diff --git a/tools/test/h5diff/h5diff_plugin.sh.in b/tools/test/h5diff/h5diff_plugin.sh.in
index 0ab169a..28f723e 100644
--- a/tools/test/h5diff/h5diff_plugin.sh.in
+++ b/tools/test/h5diff/h5diff_plugin.sh.in
@@ -158,7 +158,7 @@ done
# RUNSERIAL is used. Check if it can return exit code from executalbe correctly.
if [ -n "$RUNSERIAL_NOEXITCODE" ]; then
- echo "***Warning*** Serial Exit Code is not passed back to shell corretly."
+ echo "***Warning*** Serial Exit Code is not passed back to shell correctly."
echo "***Warning*** Exit code checking is skipped."
h5haveexitcode=no
fi
@@ -236,7 +236,7 @@ TOOLTEST() {
# don't add exit code check in pmode, as it causes failure. (exit code
# is from mpirun not tool)
# if any problem occurs relate to an exit code, it will be caught in
- # serial mode, so the test is fullfilled.
+ # serial mode, so the test is fulfilled.
if test $h5haveexitcode = 'yes' -a -z "$pmode"; then
echo "EXIT CODE: $EXIT_CODE" >> $actual
fi
@@ -327,7 +327,7 @@ TOOLTEST_ERR() {
# don't add exit code check in pmode, as it causes failure. (exit code
# is from mpirun not tool)
# if any problem occurs relate to an exit code, it will be caught in
- # serial mode, so the test is fullfilled.
+ # serial mode, so the test is fulfilled.
if test $h5haveexitcode = 'yes' -a -z "$pmode"; then
echo "EXIT CODE: $EXIT_CODE" >> $actual
fi
diff --git a/tools/test/h5diff/h5diffgentest.c b/tools/test/h5diff/h5diffgentest.c
index 5a0f7bf..5e0f830 100644
--- a/tools/test/h5diff/h5diffgentest.c
+++ b/tools/test/h5diff/h5diffgentest.c
@@ -82,7 +82,7 @@ size_t H5TOOLS_MALLOCSIZE = (128 * 1024 * 1024);
#define EXCLUDE_FILE3_2 "h5diff_exclude3-2.h5"
/* compound type with multiple vlen string types */
#define COMP_VL_STRS_FILE "h5diff_comp_vl_strs.h5"
-/* attribute compre with verbose level */
+/* attribute compare with verbose level */
#define ATTR_VERBOSE_LEVEL_FILE1 "h5diff_attr_v_level1.h5"
#define ATTR_VERBOSE_LEVEL_FILE2 "h5diff_attr_v_level2.h5"
/* file containing valid/invalid enum value mix */
@@ -2904,7 +2904,7 @@ test_group_recurse2(void)
* FILE 2-3
*------------------------------------------------------------------------*/
- /* crate target file */
+ /* create target file */
fileid4 = H5Fcreate(GRP_RECURSE2_EXT3, H5F_ACC_TRUNC, H5P_DEFAULT, H5P_DEFAULT);
/*-----------------------------------------------
@@ -2938,7 +2938,7 @@ test_group_recurse2(void)
* FILE 2-2
*------------------------------------------------------------------------*/
- /* crate target file */
+ /* create target file */
fileid3 = H5Fcreate(GRP_RECURSE2_EXT2, H5F_ACC_TRUNC, H5P_DEFAULT, H5P_DEFAULT);
/*-----------------------------------------------
@@ -2990,7 +2990,7 @@ test_group_recurse2(void)
* FILE 2-1
*------------------------------------------------------------------------*/
- /* crate target file */
+ /* create target file */
fileid2 = H5Fcreate(GRP_RECURSE2_EXT1, H5F_ACC_TRUNC, H5P_DEFAULT, H5P_DEFAULT);
/*-----------------------------------------------
@@ -4856,7 +4856,7 @@ test_data_nocomparables(const char *fname, int make_diffs)
/* attr1 - non-compatible : different type */
write_attr(did2, 1, dims1_1, "attr1", tid_attr1, attr_data_ptr2);
- /* attr2 - non-compatible : same rank, different dimention */
+ /* attr2 - non-compatible : same rank, different dimension */
write_attr(did2, 1, (hsize_t *)attr2_dim_ptr, "attr2", H5T_NATIVE_INT, data3);
/* attr3 - non-compatible : different rank */
@@ -7905,7 +7905,7 @@ gen_datareg(hid_t fid, int make_diffs /* flag to modify data buffers */)
* Function: test_hyperslab
*
* Purpose: test diff by hyperslabs. create a dataset with 1GB dimensions
- * by iterating trough 1KB hyperslabs
+ * by iterating through 1KB hyperslabs
*
*-------------------------------------------------------------------------
*/
diff --git a/tools/test/h5diff/testfiles/h5diff_10.txt b/tools/test/h5diff/testfiles/h5diff_10.txt
index c8e8560..853aede 100644
--- a/tools/test/h5diff/testfiles/h5diff_10.txt
+++ b/tools/test/h5diff/testfiles/h5diff_10.txt
@@ -129,7 +129,7 @@ usage: h5diff [OPTIONS] file1 file2 [obj1[ obj2]]
This option can be used repeatedly to exclude multiple paths.
Modes of output:
- Default mode: print the number of differences found and where they occured
+ Default mode: print the number of differences found and where they occurred
-r Report mode: print the above plus the differences
-v Verbose mode: print the above plus a list of objects and warnings
-q Quiet mode: do not print output
diff --git a/tools/test/h5diff/testfiles/h5diff_600.txt b/tools/test/h5diff/testfiles/h5diff_600.txt
index c8190ce..2893b78 100644
--- a/tools/test/h5diff/testfiles/h5diff_600.txt
+++ b/tools/test/h5diff/testfiles/h5diff_600.txt
@@ -129,7 +129,7 @@ usage: h5diff [OPTIONS] file1 file2 [obj1[ obj2]]
This option can be used repeatedly to exclude multiple paths.
Modes of output:
- Default mode: print the number of differences found and where they occured
+ Default mode: print the number of differences found and where they occurred
-r Report mode: print the above plus the differences
-v Verbose mode: print the above plus a list of objects and warnings
-q Quiet mode: do not print output
diff --git a/tools/test/h5diff/testfiles/h5diff_603.txt b/tools/test/h5diff/testfiles/h5diff_603.txt
index 03c23e7..4317282 100644
--- a/tools/test/h5diff/testfiles/h5diff_603.txt
+++ b/tools/test/h5diff/testfiles/h5diff_603.txt
@@ -130,7 +130,7 @@ usage: h5diff [OPTIONS] file1 file2 [obj1[ obj2]]
This option can be used repeatedly to exclude multiple paths.
Modes of output:
- Default mode: print the number of differences found and where they occured
+ Default mode: print the number of differences found and where they occurred
-r Report mode: print the above plus the differences
-v Verbose mode: print the above plus a list of objects and warnings
-q Quiet mode: do not print output
diff --git a/tools/test/h5diff/testfiles/h5diff_606.txt b/tools/test/h5diff/testfiles/h5diff_606.txt
index 0e9d5be..c32142f 100644
--- a/tools/test/h5diff/testfiles/h5diff_606.txt
+++ b/tools/test/h5diff/testfiles/h5diff_606.txt
@@ -130,7 +130,7 @@ usage: h5diff [OPTIONS] file1 file2 [obj1[ obj2]]
This option can be used repeatedly to exclude multiple paths.
Modes of output:
- Default mode: print the number of differences found and where they occured
+ Default mode: print the number of differences found and where they occurred
-r Report mode: print the above plus the differences
-v Verbose mode: print the above plus a list of objects and warnings
-q Quiet mode: do not print output
diff --git a/tools/test/h5diff/testfiles/h5diff_612.txt b/tools/test/h5diff/testfiles/h5diff_612.txt
index ac19926..a080f29 100644
--- a/tools/test/h5diff/testfiles/h5diff_612.txt
+++ b/tools/test/h5diff/testfiles/h5diff_612.txt
@@ -130,7 +130,7 @@ usage: h5diff [OPTIONS] file1 file2 [obj1[ obj2]]
This option can be used repeatedly to exclude multiple paths.
Modes of output:
- Default mode: print the number of differences found and where they occured
+ Default mode: print the number of differences found and where they occurred
-r Report mode: print the above plus the differences
-v Verbose mode: print the above plus a list of objects and warnings
-q Quiet mode: do not print output
diff --git a/tools/test/h5diff/testfiles/h5diff_615.txt b/tools/test/h5diff/testfiles/h5diff_615.txt
index 44eba43..a9bd76c 100644
--- a/tools/test/h5diff/testfiles/h5diff_615.txt
+++ b/tools/test/h5diff/testfiles/h5diff_615.txt
@@ -130,7 +130,7 @@ usage: h5diff [OPTIONS] file1 file2 [obj1[ obj2]]
This option can be used repeatedly to exclude multiple paths.
Modes of output:
- Default mode: print the number of differences found and where they occured
+ Default mode: print the number of differences found and where they occurred
-r Report mode: print the above plus the differences
-v Verbose mode: print the above plus a list of objects and warnings
-q Quiet mode: do not print output
diff --git a/tools/test/h5diff/testfiles/h5diff_621.txt b/tools/test/h5diff/testfiles/h5diff_621.txt
index 32310eb..cf1af15 100644
--- a/tools/test/h5diff/testfiles/h5diff_621.txt
+++ b/tools/test/h5diff/testfiles/h5diff_621.txt
@@ -130,7 +130,7 @@ usage: h5diff [OPTIONS] file1 file2 [obj1[ obj2]]
This option can be used repeatedly to exclude multiple paths.
Modes of output:
- Default mode: print the number of differences found and where they occured
+ Default mode: print the number of differences found and where they occurred
-r Report mode: print the above plus the differences
-v Verbose mode: print the above plus a list of objects and warnings
-q Quiet mode: do not print output
diff --git a/tools/test/h5diff/testfiles/h5diff_622.txt b/tools/test/h5diff/testfiles/h5diff_622.txt
index 630184f..3b325de 100644
--- a/tools/test/h5diff/testfiles/h5diff_622.txt
+++ b/tools/test/h5diff/testfiles/h5diff_622.txt
@@ -130,7 +130,7 @@ usage: h5diff [OPTIONS] file1 file2 [obj1[ obj2]]
This option can be used repeatedly to exclude multiple paths.
Modes of output:
- Default mode: print the number of differences found and where they occured
+ Default mode: print the number of differences found and where they occurred
-r Report mode: print the above plus the differences
-v Verbose mode: print the above plus a list of objects and warnings
-q Quiet mode: do not print output
diff --git a/tools/test/h5diff/testfiles/h5diff_623.txt b/tools/test/h5diff/testfiles/h5diff_623.txt
index 82c4ce2..be34fb4 100644
--- a/tools/test/h5diff/testfiles/h5diff_623.txt
+++ b/tools/test/h5diff/testfiles/h5diff_623.txt
@@ -130,7 +130,7 @@ usage: h5diff [OPTIONS] file1 file2 [obj1[ obj2]]
This option can be used repeatedly to exclude multiple paths.
Modes of output:
- Default mode: print the number of differences found and where they occured
+ Default mode: print the number of differences found and where they occurred
-r Report mode: print the above plus the differences
-v Verbose mode: print the above plus a list of objects and warnings
-q Quiet mode: do not print output
diff --git a/tools/test/h5diff/testfiles/h5diff_624.txt b/tools/test/h5diff/testfiles/h5diff_624.txt
index a675ec6..ebb76af 100644
--- a/tools/test/h5diff/testfiles/h5diff_624.txt
+++ b/tools/test/h5diff/testfiles/h5diff_624.txt
@@ -130,7 +130,7 @@ usage: h5diff [OPTIONS] file1 file2 [obj1[ obj2]]
This option can be used repeatedly to exclude multiple paths.
Modes of output:
- Default mode: print the number of differences found and where they occured
+ Default mode: print the number of differences found and where they occurred
-r Report mode: print the above plus the differences
-v Verbose mode: print the above plus a list of objects and warnings
-q Quiet mode: do not print output
diff --git a/tools/test/h5diff/testh5diff.sh.in b/tools/test/h5diff/testh5diff.sh.in
index bd27dd0..6871463 100644
--- a/tools/test/h5diff/testh5diff.sh.in
+++ b/tools/test/h5diff/testh5diff.sh.in
@@ -455,7 +455,7 @@ done
# RUNSERIAL is used. Check if it can return exit code from executalbe correctly.
if [ -n "$RUNSERIAL_NOEXITCODE" ]; then
- echo "***Warning*** Serial Exit Code is not passed back to shell corretly."
+ echo "***Warning*** Serial Exit Code is not passed back to shell correctly."
echo "***Warning*** Exit code checking is skipped."
h5haveexitcode=no
fi
@@ -518,7 +518,7 @@ TOOLTEST() {
# don't add exit code check in pmode, as it causes failure. (exit code
# is from mpirun not tool)
# if any problem occurs relate to an exit code, it will be caught in
- # serial mode, so the test is fullfilled.
+ # serial mode, so the test is fulfilled.
if test $h5haveexitcode = 'yes' -a -z "$pmode"; then
echo "EXIT CODE: $EXIT_CODE" >> $actual
fi
@@ -926,7 +926,7 @@ TOOLTEST h5diff_207.txt -c h5diff_basic2.h5 h5diff_basic2.h5 g2/dset8 g2/dset9
TOOLTEST h5diff_208.txt -c h5diff_dset_zero_dim_size1.h5 h5diff_dset_zero_dim_size2.h5
# non-comparable dataset with comparable attribute, and other comparable datasets.
-# Also test non-compatible attributes with different type, dimention, rank.
+# Also test non-compatible attributes with different type, dimension, rank.
# All the comparables should display differences.
TOOLTEST h5diff_220.txt -c non_comparables1.h5 non_comparables2.h5 /g1
@@ -1089,7 +1089,7 @@ TOOLTEST h5diff_474.txt -v h5diff_danglelinks1.h5 h5diff_danglelinks2.h5 /ext_li
TOOLTEST h5diff_475.txt -v h5diff_danglelinks1.h5 h5diff_danglelinks2.h5 /ext_link1
# ##############################################################################
-# # test for group diff recursivly
+# # test for group diff recursively
# ##############################################################################
# root
TOOLTEST h5diff_500.txt -v h5diff_grp_recurse1.h5 h5diff_grp_recurse2.h5 / /
diff --git a/tools/test/h5dump/CMakeTests.cmake b/tools/test/h5dump/CMakeTests.cmake
index c0f279d..933aba9 100644
--- a/tools/test/h5dump/CMakeTests.cmake
+++ b/tools/test/h5dump/CMakeTests.cmake
@@ -863,7 +863,7 @@
# test for named data types
ADD_H5_TEST (tcomp-2 0 --enable-error-stack -t /type1 --datatype /type2 --datatype=/group1/type3 tcompound.h5)
ADD_H5_TEST_N (tcomp-2 0 --enable-error-stack -N /type1 --any_path /type2 --any_path=/group1/type3 tcompound.h5)
- # test for unamed type
+ # test for unnamed type
ADD_H5ERR_MASK_TEST (tcomp-3 0 "object '#6632' doesn't exist" "--enable-error-stack;-t;/#6632;-g;/group2;tcompound.h5")
# test complicated compound datatype
ADD_H5_TEST (tcomp-4 0 --enable-error-stack tcompound_complex.h5)
diff --git a/tools/test/h5dump/CMakeTestsPBITS.cmake b/tools/test/h5dump/CMakeTestsPBITS.cmake
index 06be332..0c3297f 100644
--- a/tools/test/h5dump/CMakeTestsPBITS.cmake
+++ b/tools/test/h5dump/CMakeTestsPBITS.cmake
@@ -158,7 +158,7 @@
# Limits:
# Maximum number of packed bits is 8 (for now).
# Maximum integer size is 8*sizeof(long long).
- # Maximun Offset is Maximum size - 1.
+ # Maximum Offset is Maximum size - 1.
# Maximum Offset+Length is Maximum size.
# Tests:
# Normal operation on both signed and unsigned int datasets.
diff --git a/tools/test/h5dump/h5dumpgentest.c b/tools/test/h5dump/h5dumpgentest.c
index b62c3a8..5258211 100644
--- a/tools/test/h5dump/h5dumpgentest.c
+++ b/tools/test/h5dump/h5dumpgentest.c
@@ -236,7 +236,7 @@ typedef struct s1_t {
#define ARRAY3_DIM2 3
/* VL string datatype name */
-/* TODO remove complier error not used, remove the link when everything is OK */
+/* TODO remove compiler error not used, remove the link when everything is OK */
/* #define VLSTR_TYPE "vl_string_type" */
/* "File 41" macros */
@@ -1128,7 +1128,7 @@ gent_compound_dt(void)
H5Dclose(dataset);
H5Gclose(group);
- /* unamed data type */
+ /* unnamed data type */
group = H5Gcreate2(fid, "/group2", H5P_DEFAULT, H5P_DEFAULT, H5P_DEFAULT);
type = H5Tcreate(H5T_COMPOUND, sizeof(dset5_t));
@@ -1312,7 +1312,7 @@ gent_compound_dt2(void)
H5Dclose(dataset);
H5Gclose(group);
- /* unamed data type */
+ /* unnamed data type */
group = H5Gcreate2(fid, "/group2", H5P_DEFAULT, H5P_DEFAULT, H5P_DEFAULT);
type = H5Tcreate(H5T_COMPOUND, sizeof(dset5_t));
@@ -1522,11 +1522,11 @@ gent_loop2(void)
fid = H5Fcreate(FILE11, H5F_ACC_TRUNC, H5P_DEFAULT, H5P_DEFAULT);
- /* create group object g1 and implcit path from root object */
+ /* create group object g1 and implicit path from root object */
group = H5Gcreate2(fid, "/g1", H5P_DEFAULT, H5P_DEFAULT, H5P_DEFAULT);
H5Gclose(group);
- /* create group object g2 and implcit path from root object */
+ /* create group object g2 and implicit path from root object */
group = H5Gcreate2(fid, "/g2", H5P_DEFAULT, H5P_DEFAULT, H5P_DEFAULT);
H5Gclose(group);
@@ -2468,7 +2468,7 @@ gent_attrreg(void)
static void
gent_nestcomp(void)
{
- /* Compound memeber of the compound datatype*/
+ /* Compound member of the compound datatype*/
typedef struct cmp_t {
char a;
float b[2];
@@ -2541,7 +2541,7 @@ gent_nestcomp(void)
H5Tinsert(s2_tid, "c_name", HOFFSET(s2_t, c), H5T_NATIVE_DOUBLE);
H5Tinsert(s2_tid, "b_name", HOFFSET(s2_t, b), H5T_NATIVE_FLOAT);
- /* Insert compound memeber created above */
+ /* Insert compound member created above */
H5Tinsert(s2_tid, "d_name", HOFFSET(s2_t, d), cmp_tid);
/*
@@ -5218,7 +5218,7 @@ gent_named_dtype_attr(void)
sid = H5Screate(H5S_SCALAR);
HDassert(sid > 0);
- /* Create attribute on commited datatype */
+ /* Create attribute on committed datatype */
aid = H5Acreate2(tid, F42_ATTRNAME, H5T_STD_I32LE, sid, H5P_DEFAULT, H5P_DEFAULT);
HDassert(aid > 0);
@@ -5735,7 +5735,7 @@ gent_filters(void)
HDassert(ret >= 0);
if (supports_comments & H5VL_OPT_QUERY_SUPPORTED) {
- ret = H5Oset_comment_by_name(fid, "mytype", "This is a commited datatype", H5P_DEFAULT);
+ ret = H5Oset_comment_by_name(fid, "mytype", "This is a committed datatype", H5P_DEFAULT);
HDassert(ret >= 0);
}
@@ -6803,7 +6803,7 @@ gent_attr_creation_order(void)
goto out;
/*-------------------------------------------------------------------------
- * create a dataset with creation order tracked for attributes and atributes in it
+ * create a dataset with creation order tracked for attributes and attributes in it
*-------------------------------------------------------------------------
*/
@@ -6827,7 +6827,7 @@ gent_attr_creation_order(void)
did = H5I_INVALID_HID;
/*-------------------------------------------------------------------------
- * create a dataset without creation order tracked for attributes and atributes in it
+ * create a dataset without creation order tracked for attributes and attributes in it
*-------------------------------------------------------------------------
*/
@@ -6851,7 +6851,7 @@ gent_attr_creation_order(void)
did = H5I_INVALID_HID;
/*-------------------------------------------------------------------------
- * create a group with creation order tracked for attributes and atributes in it
+ * create a group with creation order tracked for attributes and attributes in it
*-------------------------------------------------------------------------
*/
@@ -6874,7 +6874,7 @@ gent_attr_creation_order(void)
gid = H5I_INVALID_HID;
/*-------------------------------------------------------------------------
- * create a group without creation order tracked for attributes and atributes in it
+ * create a group without creation order tracked for attributes and attributes in it
*-------------------------------------------------------------------------
*/
@@ -6897,7 +6897,7 @@ gent_attr_creation_order(void)
gid = H5I_INVALID_HID;
/*-------------------------------------------------------------------------
- * create a named datatype with creation order tracked for attributes and atributes in it
+ * create a named datatype with creation order tracked for attributes and attributes in it
*-------------------------------------------------------------------------
*/
@@ -6923,7 +6923,7 @@ gent_attr_creation_order(void)
tid = H5I_INVALID_HID;
/*-------------------------------------------------------------------------
- * create a named datatype without creation order tracked for attributes and atributes in it
+ * create a named datatype without creation order tracked for attributes and attributes in it
*-------------------------------------------------------------------------
*/
diff --git a/tools/test/h5dump/testh5dump.sh.in b/tools/test/h5dump/testh5dump.sh.in
index 9464302..f9f3734 100644
--- a/tools/test/h5dump/testh5dump.sh.in
+++ b/tools/test/h5dump/testh5dump.sh.in
@@ -1171,7 +1171,7 @@ TOOLTEST tcomp-1.ddl --enable-error-stack tcompound.h5
# test for named data types
TOOLTEST tcomp-2.ddl --enable-error-stack -t /type1 --datatype /type2 --datatype=/group1/type3 tcompound.h5
TOOLTEST tcomp-2.ddl --enable-error-stack -N /type1 --any_path /type2 --any_path=/group1/type3 tcompound.h5
-# test for unamed type
+# test for unnamed type
TOOLTEST4 tcomp-3.ddl --enable-error-stack -t /#6632 -g /group2 tcompound.h5
# test complicated compound datatype
TOOLTEST tcomp-4.ddl --enable-error-stack tcompound_complex.h5
diff --git a/tools/test/h5dump/testh5dumppbits.sh.in b/tools/test/h5dump/testh5dumppbits.sh.in
index 6689749..eaf91fd 100644
--- a/tools/test/h5dump/testh5dumppbits.sh.in
+++ b/tools/test/h5dump/testh5dumppbits.sh.in
@@ -516,7 +516,7 @@ TOOLTEST tnofilename-with-packed-bits.ddl --enable-error-stack
# Limits:
# Maximum number of packed bits is 8 (for now).
# Maximum integer size is 64 (for now).
-# Maximun Offset is 63 (Maximum size - 1).
+# Maximum Offset is 63 (Maximum size - 1).
# Maximum Offset+Length is 64 (Maximum size).
# Tests:
# Normal operation on both signed and unsigned int datasets.
diff --git a/tools/test/h5format_convert/h5fc_gentest.c b/tools/test/h5format_convert/h5fc_gentest.c
index 09cca45..ccb64b5 100644
--- a/tools/test/h5format_convert/h5fc_gentest.c
+++ b/tools/test/h5format_convert/h5fc_gentest.c
@@ -381,7 +381,7 @@ error:
/*
* Function: gen_err_level()
*
- * Generate a file to test the situtation described in HDFFV-9434:
+ * Generate a file to test the situation described in HDFFV-9434:
* Exceed the limit of v1-btree level
*
* Create a file with H5Pset_istore_k(fcpl, 1).
diff --git a/tools/test/h5format_convert/testh5fc.sh.in b/tools/test/h5format_convert/testh5fc.sh.in
index 762f43d..df083d1 100644
--- a/tools/test/h5format_convert/testh5fc.sh.in
+++ b/tools/test/h5format_convert/testh5fc.sh.in
@@ -313,7 +313,7 @@ TOOLTEST_MASK_OUT() {
$RM $actual $actual_err $actual_sav $actual_err_sav
fi
}
-# Same as TOOLTEST_OUT except only compares error outout
+# Same as TOOLTEST_OUT except only compares error output
TOOLTEST_ERR() {
# Prepare expected and actual output
expect="$TESTDIR/$1"
diff --git a/tools/test/h5jam/testh5jam.sh.in b/tools/test/h5jam/testh5jam.sh.in
index caa3416..191fef0 100644
--- a/tools/test/h5jam/testh5jam.sh.in
+++ b/tools/test/h5jam/testh5jam.sh.in
@@ -345,7 +345,7 @@ CHECK_NOUB() {
# JAMTEST user_block file.h5 [--clobber] [ofile.h5]
#
# Test the 'jam' tool:
-# 1. figure out the input and output, and the comparision
+# 1. figure out the input and output, and the comparison
# that will be done.
# 2. call 'jam' with the appropriate arguments
# 3. check the user block is correct in the output (Check_UB)
diff --git a/tools/test/h5ls/h5ls_plugin.sh.in b/tools/test/h5ls/h5ls_plugin.sh.in
index 29e982d..0077b07 100644
--- a/tools/test/h5ls/h5ls_plugin.sh.in
+++ b/tools/test/h5ls/h5ls_plugin.sh.in
@@ -166,7 +166,7 @@ VERIFY() {
# the `nerrors' global variable and (if $verbose is set) display up to $NLINS
# lines of the actual output from the tool test. The actual output is not
# removed if $HDF5_NOCLEANUP has a non-zero value.
-# Arguemnts:
+# Arguments:
# $1 -- actual output filename to use
# $2 and on -- argument for the h5ls tool
TOOLTEST() {
diff --git a/tools/test/h5ls/testh5ls.sh.in b/tools/test/h5ls/testh5ls.sh.in
index ee0dad7..36583da 100644
--- a/tools/test/h5ls/testh5ls.sh.in
+++ b/tools/test/h5ls/testh5ls.sh.in
@@ -164,7 +164,7 @@ $SRC_H5LS_TESTFILES/tdset_idx.ls
# RUNSERIAL is used. Check if it can return exit code from executalbe correctly.
if [ -n "$RUNSERIAL_NOEXITCODE" ]; then
- echo "***Warning*** Serial Exit Code is not passed back to shell corretly."
+ echo "***Warning*** Serial Exit Code is not passed back to shell correctly."
echo "***Warning*** Exit code checking is skipped."
h5haveexitcode=no
fi
@@ -230,7 +230,7 @@ TESTING() {
# the `nerrors' global variable and (if $verbose is set) display up to $NLINS
# lines of the actual output from the tool test. The actual output is not
# removed if $HDF5_NOCLEANUP has a non-zero value.
-# Arguemnts:
+# Arguments:
# $1 -- actual output filename to use
# $2 and on -- argument for the h5ls tool
TOOLTEST() {
diff --git a/tools/test/h5ls/testh5lsvds.sh.in b/tools/test/h5ls/testh5lsvds.sh.in
index 4814f85..ab2d137 100644
--- a/tools/test/h5ls/testh5lsvds.sh.in
+++ b/tools/test/h5ls/testh5lsvds.sh.in
@@ -103,7 +103,7 @@ $SRC_H5LS_TESTFILES/vds/tvds-5.ls
# RUNSERIAL is used. Check if it can return exit code from executalbe correctly.
if [ -n "$RUNSERIAL_NOEXITCODE" ]; then
- echo "***Warning*** Serial Exit Code is not passed back to shell corretly."
+ echo "***Warning*** Serial Exit Code is not passed back to shell correctly."
echo "***Warning*** Exit code checking is skipped."
h5haveexitcode=no
fi
@@ -169,7 +169,7 @@ TESTING() {
# the `nerrors' global variable and (if $verbose is set) display up to $NLINS
# lines of the actual output from the tool test. The actual output is not
# removed if $HDF5_NOCLEANUP has a non-zero value.
-# Arguemnts:
+# Arguments:
# $1 -- actual output filename to use
# $2 and on -- argument for the h5ls tool
TOOLTEST() {
diff --git a/tools/test/h5repack/CMakeTests.cmake b/tools/test/h5repack/CMakeTests.cmake
index 397c3ac..09648ff 100644
--- a/tools/test/h5repack/CMakeTests.cmake
+++ b/tools/test/h5repack/CMakeTests.cmake
@@ -1541,7 +1541,7 @@
ADD_H5_TEST (bug1814 "TEST" ${FILE_REF})
# test attribute with various references (bug1797 / HDFFV-5932)
-# the references in attribute of compund or vlen datatype
+# the references in attribute of compound or vlen datatype
ADD_H5_TEST (HDFFV-5932 "TEST" ${FILE_ATTR_REF})
# Add test for memory leak in attribute. This test is verified by CTEST.
diff --git a/tools/test/h5repack/dynlib_rpk.c b/tools/test/h5repack/dynlib_rpk.c
index b228344..bef032a 100644
--- a/tools/test/h5repack/dynlib_rpk.c
+++ b/tools/test/h5repack/dynlib_rpk.c
@@ -83,7 +83,7 @@ H5Z_filter_dynlib1(unsigned int flags, size_t cd_nelmts, const unsigned int *cd_
add_on = (int)cd_values[0];
if (flags & H5Z_FLAG_REVERSE) { /*read*/
- /* Substract the "add on" value to all the data values */
+ /* Subtract the "add on" value to all the data values */
while (buf_left > 0) {
*int_ptr++ -= add_on;
buf_left -= sizeof(int);
diff --git a/tools/test/h5repack/dynlib_vrpk.c b/tools/test/h5repack/dynlib_vrpk.c
index 5eac4f2..a45de65 100644
--- a/tools/test/h5repack/dynlib_vrpk.c
+++ b/tools/test/h5repack/dynlib_vrpk.c
@@ -91,7 +91,7 @@ H5Z_filter_dynlib4(unsigned int flags, size_t cd_nelmts, const unsigned int *cd_
add_on = (int)cd_values[0];
if (flags & H5Z_FLAG_REVERSE) { /*read*/
- /* Substract the "add on" value to all the data values */
+ /* Subtract the "add on" value to all the data values */
while (buf_left > 0) {
*int_ptr++ -= add_on;
buf_left -= sizeof(int);
diff --git a/tools/test/h5repack/h5repack.sh.in b/tools/test/h5repack/h5repack.sh.in
index 1e54670..a241ea8 100644
--- a/tools/test/h5repack/h5repack.sh.in
+++ b/tools/test/h5repack/h5repack.sh.in
@@ -525,7 +525,7 @@ VERIFY_LAYOUT_ALL()
#---------------------------------
# check the layout from a dataset
- # check if the other layouts still exsit
+ # check if the other layouts still exist
VERIFY "layouts"
(
cd $TESTDIR
@@ -1713,10 +1713,10 @@ TOOLTEST1 family tfamily%05d.h5
TOOLTEST bug1814 h5repack_refs.h5
# test attribute with various references (bug1797 / HDFFV-5932)
-# the references in attribute of compund or vlen datatype
+# the references in attribute of compound or vlen datatype
TOOLTEST HDFFV-5932 h5repack_attr_refs.h5
-# Add test for memory leak in attirbute. This test is verified by CTEST.
+# Add test for memory leak in attribute. This test is verified by CTEST.
# 1. leak from vlen string
# 2. leak from compound type without reference member
# (HDFFV-7840, )
diff --git a/tools/test/h5repack/h5repackgentest.c b/tools/test/h5repack/h5repackgentest.c
index 576b1c7..b026cc0 100644
--- a/tools/test/h5repack/h5repackgentest.c
+++ b/tools/test/h5repack/h5repackgentest.c
@@ -27,9 +27,9 @@
* There is no restriction on the name, number, or structure of datasets and
* groups in HDF5 file.
*
- * The inluded datatypes should be more than adequate to verify the correctness
+ * The included datatypes should be more than adequate to verify the correctness
* of the behavior -- if one type can be consolidated from external storage,
- * then thay all can.
+ * then they all can.
*/
#include "hdf5.h"
diff --git a/tools/test/h5repack/h5repacktst.c b/tools/test/h5repack/h5repacktst.c
index 6ca33e2..8b114e4 100644
--- a/tools/test/h5repack/h5repacktst.c
+++ b/tools/test/h5repack/h5repacktst.c
@@ -628,7 +628,7 @@ main(void)
SKIPPED();
#endif
- TESTING(" addding shuffle filter");
+ TESTING(" adding shuffle filter");
/*-------------------------------------------------------------------------
* test an individual object option
@@ -657,7 +657,7 @@ main(void)
*-------------------------------------------------------------------------
*/
- TESTING(" addding shuffle filter to all");
+ TESTING(" adding shuffle filter to all");
if (h5repack_init(&pack_options, 0, FALSE) < 0)
GOERROR;
@@ -1602,17 +1602,17 @@ main(void)
}
/*-------------------------------------------------------------------------
- * test file with aligment
+ * test file with alignment
*-------------------------------------------------------------------------
*/
- TESTING(" file with aligment");
+ TESTING(" file with alignment");
#ifdef H5_HAVE_FILTER_DEFLATE
if (h5repack_init(&pack_options, 0, FALSE) < 0)
GOERROR;
- /* add the options for aligment */
+ /* add the options for alignment */
pack_options.alignment = 1;
pack_options.threshold = 1;
@@ -1623,7 +1623,7 @@ main(void)
if (h5repack_verify(FNAME8, FNAME8OUT, &pack_options) <= 0)
GOERROR;
- /* verify aligment */
+ /* verify alignment */
{
hsize_t threshold;
hsize_t alignment;
@@ -6347,7 +6347,7 @@ gen_refered_objs(hid_t loc_id)
goto out;
}
- /* create normal dataset which is refered */
+ /* create normal dataset which is referred */
did2 = H5Dcreate2(loc_id, NAME_OBJ_DS2, H5T_STD_I8LE, sid2, H5P_DEFAULT, H5P_DEFAULT, H5P_DEFAULT);
if (did2 < 0) {
HDfprintf(stderr, "Error: %s %d> H5Dcreate2 failed.\n", __func__, __LINE__);
@@ -6387,7 +6387,7 @@ out:
* Generate object references to objects (dataset,group and named datatype)
*
* Note:
- * copied from h5copygentest.c and upate to create named datatype
+ * copied from h5copygentest.c and update to create named datatype
*
* Programmer: Jonathan Kim (March 18, 2010)
*------------------------------------------------------------------------*/
diff --git a/tools/test/h5repack/testfiles/README b/tools/test/h5repack/testfiles/README
index 4096dee..4cb66f4 100644
--- a/tools/test/h5repack/testfiles/README
+++ b/tools/test/h5repack/testfiles/README
@@ -1,5 +1,5 @@
h5repack_nested_8bit_enum_deflated.h5:
h5repack_nested_8bit_enum.h5:
- enuberated 8bit type nested in compount type. Original file provided
+ enuberated 8bit type nested in compound type. Original file provided
by a user (HDFFV-8667) as a test file. Used h5copy to extract only the
Compound type dataset. The non-deflated version is produced by h5repack.
diff --git a/tools/test/h5stat/testh5stat.sh.in b/tools/test/h5stat/testh5stat.sh.in
index 647f06a..1b4e504 100644
--- a/tools/test/h5stat/testh5stat.sh.in
+++ b/tools/test/h5stat/testh5stat.sh.in
@@ -13,7 +13,7 @@
#
# Tests for the h5stat tool
#
-# Modifcations:
+# Modifications:
# Vailin Choi; July 2013
# Add tests for -l, -m, -a options
#
diff --git a/tools/test/misc/h5clear_gentest.c b/tools/test/misc/h5clear_gentest.c
index 923c3f2..225fe16 100644
--- a/tools/test/misc/h5clear_gentest.c
+++ b/tools/test/misc/h5clear_gentest.c
@@ -16,7 +16,7 @@
/* The HDF5 test files */
const char *FILENAME[] = {
"h5clear_sec2_v3.h5", /* 0 -- sec2 file with superblock version 3 */
- "h5clear_log_v3.h5", /* 1 -- log file with superblock veresion 3 */
+ "h5clear_log_v3.h5", /* 1 -- log file with superblock version 3 */
"h5clear_sec2_v0.h5", /* 2 -- sec2 file with superblock version 0 */
"h5clear_sec2_v2.h5" /* 3 -- sec2 file with superblock version 2 */
};
diff --git a/tools/test/misc/h5perf_gentest.c b/tools/test/misc/h5perf_gentest.c
index 332e49c..248fd92 100644
--- a/tools/test/misc/h5perf_gentest.c
+++ b/tools/test/misc/h5perf_gentest.c
@@ -255,9 +255,9 @@ create_perf_test_file(const char *fname, int ngrps, int ndsets, int nattrs, hsiz
buf_vlen_s = (char **)calloc(dim0, sizeof(char *));
buf_str = malloc(dim0 * sizeof(*buf_str));
- /* allocate array of doulbe pointers */
+ /* allocate array of double pointers */
buf_double2d = (double **)calloc(dims2d[0], sizeof(double *));
- /* allocate a contigous chunk of memory for the data */
+ /* allocate a contiguous chunk of memory for the data */
buf_double2d[0] = (double *)calloc(dims2d[0] * dims2d[1], sizeof(double));
/* assign memory city to pointer array */
for (i = 1; i < dims2d[0]; i++)
@@ -396,7 +396,7 @@ create_perf_test_file(const char *fname, int ngrps, int ndsets, int nattrs, hsiz
H5Dclose(did);
/* 6 add compound data */
- HDsprintf(name, "%05d compund data", j);
+ HDsprintf(name, "%05d compound data", j);
did = H5Dcreate(gid1, name, cmp_tid, sid_1d, H5P_DEFAULT, dcpl, H5P_DEFAULT);
H5Dwrite(did, cmp_tid, H5S_ALL, H5S_ALL, H5P_DEFAULT, buf_comp);
if (!j)
diff --git a/tools/test/perform/direct_write_perf.c b/tools/test/perform/direct_write_perf.c
index 371c4f9..2076cf4 100644
--- a/tools/test/perform/direct_write_perf.c
+++ b/tools/test/perform/direct_write_perf.c
@@ -625,7 +625,7 @@ test_unix_write(void)
TEST_ERROR;
}
else if (op_size == 0) {
- printf(" unable to write sufficent data to file because %s \n", strerror(errno));
+ printf(" unable to write sufficient data to file because %s \n", strerror(errno));
TEST_ERROR;
}
}
diff --git a/tools/test/perform/zip_perf.c b/tools/test/perform/zip_perf.c
index 30eb06e..37e69e2 100644
--- a/tools/test/perform/zip_perf.c
+++ b/tools/test/perform/zip_perf.c
@@ -267,7 +267,7 @@ usage(void)
HDfprintf(stdout, " data you want [default: 0]");
HDfprintf(stdout, " -s S, --file-size=S Maximum size of uncompressed file [default: 64M]\n");
HDfprintf(stdout, " -B S, --max-buffer_size=S Maximum size of buffer [default: 1M]\n");
- HDfprintf(stdout, " -b S, --min-buffer_size=S Minumum size of buffer [default: 128K]\n");
+ HDfprintf(stdout, " -b S, --min-buffer_size=S Minimum size of buffer [default: 128K]\n");
HDfprintf(stdout, " -p D, --prefix=D The directory prefix to place the file\n");
HDfprintf(stdout, " -r, --random-test Use random data to write to the file\n");
HDfprintf(stdout, " [default: no]\n");
@@ -551,7 +551,7 @@ main(int argc, const char *argv[])
}
if (min_buf_size > max_buf_size)
- error("minmum buffer size (%d) exceeds maximum buffer size (%d)", min_buf_size, max_buf_size);
+ error("minimum buffer size (%d) exceeds maximum buffer size (%d)", min_buf_size, max_buf_size);
HDfprintf(stdout, "Filesize: %ld\n", file_size);
diff --git a/utils/mirror_vfd/mirror_server.c b/utils/mirror_vfd/mirror_server.c
index 9de07c1..f6a8bfe 100644
--- a/utils/mirror_vfd/mirror_server.c
+++ b/utils/mirror_vfd/mirror_server.c
@@ -80,7 +80,7 @@
* line.
*
* `log_prepend_type` (int)
- * Flag that the logging messages should have the assocaited verbosity
+ * Flag that the logging messages should have the associated verbosity
* level present in the line (e.g., "WARN", "ERROR", or "INFO").
*
* `log_path` (char *)
diff --git a/utils/mirror_vfd/mirror_server_stop.c b/utils/mirror_vfd/mirror_server_stop.c
index bf306d9..024b33a 100644
--- a/utils/mirror_vfd/mirror_server_stop.c
+++ b/utils/mirror_vfd/mirror_server_stop.c
@@ -15,7 +15,7 @@
* Exists for cross-platform, optionally remote shutdown.
*/
-#include "H5private.h" /* System compatability call-wrapper macros */
+#include "H5private.h" /* System compatibility call-wrapper macros */
#ifdef H5_HAVE_MIRROR_VFD
@@ -66,7 +66,7 @@ usage(void)
"\n"
"Options:\n"
" -h | --help Print this usage message and exit.\n"
- " --ip=ADDR IP Address of remote server (defaut %s)\n"
+ " --ip=ADDR IP Address of remote server (default %s)\n"
" --port=PORT Handshake port of remote server (default %d)\n",
MSHS_DEFAULT_IP, MSHS_DEFAULT_PORTNO);
} /* end usage() */
diff --git a/utils/tools/h5dwalk/h5dwalk.1 b/utils/tools/h5dwalk/h5dwalk.1
index 60e1080..eb0e5e8 100644
--- a/utils/tools/h5dwalk/h5dwalk.1
+++ b/utils/tools/h5dwalk/h5dwalk.1
@@ -25,7 +25,7 @@ Captures the hdf5 tool output into a named file.
.B \-l
or
.B \-\-log [file]
-Captures hdf5 tool output into a individual log files. If an optional file (directory) is specified, then output from all tool instances will be written in the given file directory. Without the optional filename, each tool instance ouput will be captured in a new log file whose name is associated with the hdf5 tool that was run and is written in the current working directory.
+Captures hdf5 tool output into a individual log files. If an optional file (directory) is specified, then output from all tool instances will be written in the given file directory. Without the optional filename, each tool instance output will be captured in a new log file whose name is associated with the hdf5 tool that was run and is written in the current working directory.
.TP
.B \-E
or
diff --git a/utils/tools/h5dwalk/h5dwalk.c b/utils/tools/h5dwalk/h5dwalk.c