Merge pull request #153 from glennklockwood/docfixes

Updated documentation
master
Glenn K. Lockwood 2019-07-30 19:29:37 -05:00 committed by GitHub
commit b025d6bdb3
No known key found for this signature in database
GPG Key ID: 4AEE18F83AFDEB23
13 changed files with 372 additions and 323 deletions

View File

@ -1,31 +1,31 @@
# HPC IO Benchmark Repository [![Build Status](https://travis-ci.org/hpc/ior.svg?branch=master)](https://travis-ci.org/hpc/ior) # HPC IO Benchmark Repository [![Build Status](https://travis-ci.org/hpc/ior.svg?branch=master)](https://travis-ci.org/hpc/ior)
This repo now contains both IOR and mdtest. This repository contains the IOR and mdtest parallel I/O benchmarks. The
See also NOTES.txt [official IOR/mdtest documention][] can be found in the `docs/` subdirectory or
on Read the Docs.
# Building ## Building
0. If "configure" is missing from the top level directory, you 1. If `configure` is missing from the top level directory, you probably
probably retrieved this code directly from the repository. retrieved this code directly from the repository. Run `./bootstrap`
Run "./bootstrap". to generate the configure script. Alternatively, download an
[official IOR release][] which includes the configure script.
If your versions of the autotools are not new enough to run 1. Run `./configure`. For a full list of configuration options, use
this script, download and official tarball in which the `./configure --help`.
configure script is already provided.
1. Run "./configure" 2. Run `make`
See "./configure --help" for configuration options. 3. Optionally, run `make install`. The installation prefix
can be changed via `./configure --prefix=...`.
2. Run "make" ## Testing
3. Optionally, run "make install". The installation prefix * Run `make check` to invoke the unit tests.
can be changed as an option to the "configure" script. * More comprehensive functionality tests are included in `testing/`. These
scripts will launch IOR and mdtest via MPI.
* Docker scripts are also provided in `testing/docker/` to test various
distributions at once.
# Testing [official IOR release]: https://github.com/hpc/ior/releases
[official IOR/mdtest documention]: http://ior.readthedocs.org/
Run "make check" to invoke the unit test framework of Automake.
* To run basic functionality tests that we use for continuous integration, see ./testing/
* There are docker scripts provided to test various distributions at once.
* See ./testing/docker/

5
doc/sphinx/README.md Normal file
View File

@ -0,0 +1,5 @@
To build the documentation in this directory,
sphinx-build . _build_html
The output will be saved as html in `_build_html/`.

View File

@ -1,4 +1,4 @@
Changes in IOR Changes in IOR
***************** **************
.. .. include:: ../../ChangeLog .. include:: ../../NEWS

View File

@ -64,18 +64,30 @@ source_suffix = '.rst'
master_doc = 'index' master_doc = 'index'
# General information about the project. # General information about the project.
project = u'IOR' project = u'ior'
copyright = u'2017, IOR' copyright = u'2018, The Regents of the University of California'
author = u'IOR' author = u'The IOR Developer Community'
# The version info for the project you're documenting, acts as replacement for # The version info for the project you're documenting, acts as replacement for
# |version| and |release|, also used in various other places throughout the # |version| and |release|, also used in various other places throughout the
# built documents. # built documents.
# #
# The short X.Y version. # The short X.Y version.
version = u'3.1.0' version = u'3.2.0'
# The full version, including alpha/beta/rc tags. # The full version, including alpha/beta/rc tags.
release = u'0' release = u'3.2.0'
for line in open(os.path.join('..', '..', 'META'), 'r'):
key, value = line.split(None, 1)
if key == "Package:":
project = value
elif key == "Version:":
version = value
release = value
# We don't use the release field
# elif key == "Release:":
# release = value
# The language for content autogenerated by Sphinx. Refer to documentation # The language for content autogenerated by Sphinx. Refer to documentation
# for a list of supported languages. # for a list of supported languages.
@ -185,7 +197,7 @@ man_pages = [
# dir menu entry, description, category) # dir menu entry, description, category)
texinfo_documents = [ texinfo_documents = [
(master_doc, 'IOR', u'IOR Documentation', (master_doc, 'IOR', u'IOR Documentation',
author, 'IOR', 'One line description of project.', author, 'IOR', 'IOR/mdtest Parallel I/O Benchmarks',
'Miscellaneous'), 'Miscellaneous'),
] ]

View File

@ -1,8 +1,8 @@
Continues Integration Continuous Integration
===================== ======================
Continues Integration is used for basic sanity checking. Travis-CI provides free Continuous Integration is used for basic sanity checking. Travis-CI provides free
CI for open source github projects and is configured via a .travis.yml. CI for open source github projects and is configured via a ``.travis.yml``.
For now this is set up to compile IOR on a ubuntu 14.04 machine with gcc 4.8, For now this is set up to compile IOR on a ubuntu 14.04 machine with gcc 4.8,
openmpi and hdf5 for the backends. This is a pretty basic check and should be openmpi and hdf5 for the backends. This is a pretty basic check and should be

View File

@ -0,0 +1,38 @@
Release Process
===============
To build a new version of IOR::
$ docker run -it ubuntu bash
$ apt-get update
$ apt-get install -y git automake autoconf make gcc mpich
$ git clone -b rc https://github.com/hpc/ior
$ cd ior
$ ./travis-build.sh
To create a new release candidate from RC,
1. Disable the ``check-news`` option in ``AM_INIT_AUTOMAKE`` inside configure.ac
2. Append "rcX" to the ``Version:`` field in META where X is the release
candidate number
3. Build a release package as described above
To create a new minor release of IOR,
1. Build the rc branch as described above
2. Create a release on GitHub which creates the appropriate tag
3. Upload the source distributions generated by travis-build.sh
To create a micro branch of IOR (e.g., if a release needs a hotfix),
1. Check out the relevant release tagged in the rc branch (e.g., ``3.2.0``)
2. Create a branch with the major.minor name (e.g., ``3.2``) from that tag
3. Update the ``Version:`` in META
4. Apply hotfix(es) to that major.minor branch
5. Create the major.minor.micro release on GitHub
To initiate a feature freeze,
1. Merge the master branch into the rc branch
2. Update the ``Version:`` field in META `of the master branch` to be the `next`
release version, not the one whose features have just been frozen

View File

@ -13,7 +13,7 @@
userDoc/install userDoc/install
First Steps <userDoc/tutorial> First Steps <userDoc/tutorial>
userDoc/options userDoc/options
userDoc/skripts userDoc/scripts
userDoc/compatibility userDoc/compatibility
FAQ <userDoc/faq> FAQ <userDoc/faq>
@ -24,6 +24,7 @@
devDoc/doxygen devDoc/doxygen
devDoc/CI devDoc/CI
devDoc/release
.. toctree:: .. toctree::
:hidden: :hidden:

View File

@ -1,26 +1,22 @@
Introduction Introduction
============ ============
Welcome to the IOR documentation. IOR is a parallel IO benchmark that can be used to test the performance of
parallel storage systems using various interfaces and access patterns. The
IOR repository also includes the mdtest benchmark which specifically tests
the peak metadata rates of storage systems under different directory
structures. Both benchmarks use a common parallel I/O abstraction backend
and rely on MPI for synchronization.
**I**\ nterleaved **o**\ r **R**\ andom is a parallel IO benchmark. This documentation consists of two parts.
IOR can be used for testing performance of parallel file systems using various
interfaces and access patterns. IOR uses MPI for process synchronization.
This documentation provides information for versions 3 and higher, for other
versions check :ref:`compatibility`
This documentation consists of tow parts. **User documentation** includes installation instructions (:ref:`install`), a
beginner's tutorial (:ref:`first-steps`), and information about IOR's
runtime :ref:`options`.
The first part is a user documentation were you find instructions on compilation, a **Developer documentation** consists of code documentation generated with
beginners tutorial (:ref:`first-steps`) as well as information about all Doxygen and some notes about the contiguous integration with Travis.
available :ref:`options`.
The second part is the developer documentation. It currently only consists of a Many aspects of both IOR/mdtest user and developer documentation are incomplete,
auto generated Doxygen and some notes about the contiguous integration with travis. and contributors are encouraged to comment the code directly or expand upon this
As there are quite some people how needs to modify or extend IOR to there needs documentation.
it would be great to have documentation on what and how to alter IOR without
breaking other stuff. Currently there is neither a documentation on the overall
concept of the code nor on implementation details. If you are getting your
hands dirty in code anyways or have deeper understanding of IOR, you are more
then welcome to comment the code directly, which will result in better Doxygen
output or add your insight to this sphinx documentation.

View File

@ -1,10 +1,11 @@
.. _compatibility: .. _compatibility:
Compatibility Version Compatibility
============= =====================
IOR has a long history. Here are some hints about compatibility with older IOR has a long history and only IOR version 3 is currently supported. However,
versions. there are many forks of IOR based on earlier versions, and the following
incompatibilities are known to exist between major versions.
1) IOR version 1 (c. 1996-2002) and IOR version 2 (c. 2003-present) are 1) IOR version 1 (c. 1996-2002) and IOR version 2 (c. 2003-present) are
incompatible. Input decks from one will not work on the other. As version incompatible. Input decks from one will not work on the other. As version

View File

@ -1,3 +1,5 @@
.. _install:
Install Install
======= =======

View File

@ -16,7 +16,9 @@ normal parameters override each other, so the last one executed.
Command line options Command line options
-------------------- --------------------
These options are to be used on the command line. E.g., 'IOR -a POSIX -b 4K'.
These options are to be used on the command line (e.g., ``./ior -a POSIX -b 4K``).
-a S api -- API for I/O [POSIX|MPIIO|HDF5|HDFS|S3|S3_EMC|NCMPI|RADOS] -a S api -- API for I/O [POSIX|MPIIO|HDF5|HDFS|S3|S3_EMC|NCMPI|RADOS]
-A N refNum -- user reference number to include in long summary -A N refNum -- user reference number to include in long summary
-b N blockSize -- contiguous bytes to write per task (e.g.: 8, 4k, 2m, 1g) -b N blockSize -- contiguous bytes to write per task (e.g.: 8, 4k, 2m, 1g)
@ -69,282 +71,264 @@ These options are to be used on the command line. E.g., 'IOR -a POSIX -b 4K'.
-Z reorderTasksRandom -- changes task ordering to random ordering for readback -Z reorderTasksRandom -- changes task ordering to random ordering for readback
NOTES: * S is a string, N is an integer number. * S is a string, N is an integer number.
* For transfer and block sizes, the case-insensitive K, M, and G
* For transfer and block sizes, the case-insensitive K, M, and G
suffices are recognized. I.e., '4k' or '4K' is accepted as 4096. suffices are recognized. I.e., '4k' or '4K' is accepted as 4096.
Directive Options Directive Options
------------------ ------------------
For each of the general settings, note the default is shown in brackets.
IMPORTANT NOTE: For all true/false options below [1]=true, [0]=false
IMPORTANT NOTE: Contrary to appearance, the script options below are NOT case sensitive
For all true/false options below, [1]=true, [0]=false. All options are case-insensitive.
GENERAL: GENERAL
^^^^^^^^^^^^^^ ^^^^^^^^^^^^^^
* refNum - user supplied reference number, included in
long summary [0]
* api - must be set to one of POSIX, MPIIO, HDF5, HDFS, S3, * ``refNum`` - user supplied reference number, included in long summary
S3_EMC, or NCMPI, depending on test [POSIX] (default: 0)
* testFile - name of the output file [testFile] * ``api`` - must be set to one of POSIX, MPIIO, HDF5, HDFS, S3, S3_EMC, NCMPI,
NOTE: with filePerProc set, the tasks can round IME, MMAP, or RAODS depending on test (default: ``POSIX``)
robin across multiple file names '-o S@S@S'
* hintsFileName - name of the hints file [] * ``testFile`` - name of the output file [testFile]. With ``filePerProc`` set,
the tasks can round robin across multiple file names via ``-o S@S@S``.
If only a single file name is specified in this case, IOR appends the MPI
rank to the end of each file generated (e.g., ``testFile.00000059``)
(default: ``testFile``)
* repetitions - number of times to run each test [1] * ``hintsFileName`` - name of the hints file (default: none)
* multiFile - creates multiple files for single-shared-file or * ``repetitions`` - number of times to run each test (default: 1)
file-per-process modes; i.e. each iteration creates
a new file [0=FALSE]
* reorderTasksConstant - reorders tasks by a constant node offset for writing/reading neighbor's * ``multiFile`` - creates multiple files for single-shared-file or
data from different nodes [0=FALSE] file-per-process modes for each iteration (default: 0)
* taskPerNodeOffset - for read tests. Use with -C & -Z options. [1] * ``reorderTasksConstant`` - reorders tasks by a constant node offset for
With reorderTasks, constant N. With reordertasksrandom, >= N writing/reading neighbor's data from different nodes (default: 0)
* reorderTasksRandom - reorders tasks to random ordering for readback [0=FALSE] * ``taskPerNodeOffset`` - for read tests. Use with ``-C`` and ``-Z`` options.
With ``reorderTasks``, constant N. With ``reordertasksrandom``, >= N
(default: 1)
* reorderTasksRandomSeed - random seed for reordertasksrandom option. [0] * ``reorderTasksRandom`` - reorders tasks to random ordering for read tests
>0, same seed for all iterations. <0, different seed for each iteration (default: 0)
* quitOnError - upon error encountered on checkWrite or checkRead, * ``reorderTasksRandomSeed`` - random seed for ``reordertasksrandom`` option. (default: 0)
display current error and then stop execution; * When > 0, use the same seed for all iterations
if not set, count errors and continue [0=FALSE] * When < 0, different seed for each iteration
* numTasks - number of tasks that should participate in the test * ``quitOnError`` - upon error encountered on ``checkWrite`` or ``checkRead``,
[0] display current error and then stop execution. Otherwise, count errors and
NOTE: 0 denotes all tasks continue (default: 0)
* interTestDelay - this is the time in seconds to delay before * ``numTasks`` - number of tasks that should participate in the test. 0
beginning a write or read in a series of tests [0] denotes all tasks. (default: 0)
NOTE: it does not delay before a check write or
check read
* outlierThreshold - gives warning if any task is more than this number * ``interTestDelay`` - time (in seconds) to delay before beginning a write or
of seconds from the mean of all participating tasks. read phase in a series of tests This does not delay before check-write or
If so, the task is identified, its time (start, check-read phases. (default: 0)
elapsed create, elapsed transfer, elapsed close, or
end) is reported, as is the mean and standard
deviation for all tasks. The default for this is 0,
which turns it off. If set to a positive value, for
example 3, any task not within 3 seconds of the mean
displays its times. [0]
* intraTestBarriers - use barrier between open, write/read, and close [0=FALSE] * ``outlierThreshold`` - gives warning if any task is more than this number of
seconds from the mean of all participating tasks. The warning includes the
offending task, its timers (start, elapsed create, elapsed transfer, elapsed
close, end), and the mean and standard deviation for all tasks. When zero,
disable this feature. (default: 0)
* uniqueDir - create and use unique directory for each * ``intraTestBarriers`` - use barrier between open, write/read, and close
file-per-process [0=FALSE] phases (default: 0)
* writeFile - writes file(s), first deleting any existing file [1=TRUE] * ``uniqueDir`` - create and use unique directory for each file-per-process
NOTE: the defaults for writeFile and readFile are (default: 0)
set such that if there is not at least one of
the following -w, -r, -W, or -R, it is assumed
that -w and -r are expected and are
consequently used -- this is only true with
the command line, and may be overridden in
a script
* readFile - reads existing file(s) (from current or previous * ``writeFile`` - write file(s), first deleting any existing file.
run) [1=TRUE] The defaults for ``writeFile`` and ``readFile`` are set such that if there
NOTE: see writeFile notes is not at least one of ``-w``, ``-r``, ``-W``, or ``-R``, ``-w`` and ``-r``
are enabled. If either ``writeFile`` or ``readFile`` are explicitly
enabled, though, its complement is *not* also implicitly enabled.
* filePerProc - accesses a single file for each processor; default * ``readFile`` - reads existing file(s) as specified by the ``testFile``
is a single file accessed by all processors [0=FALSE] option. The defaults for ``writeFile`` and ``readFile`` are set such that
if there is not at least one of ``-w``, ``-r``, ``-W``, or ``-R``, ``-w``
and ``-r`` are enabled. If either ``writeFile`` or ``readFile`` are
explicitly enabled, though, its complement is *not* also implicitly enabled.
* checkWrite - read data back and check for errors against known * ``filePerProc`` - have each MPI process perform I/O to a unique file
pattern; can be used independently of writeFile [0=FALSE] (default: 0)
NOTES: - data checking is not timed and does not
affect other performance timings
- all errors tallied and returned as program
exit code, unless quitOnError set
* checkRead - reread data and check for errors between reads; can * ``checkWrite`` - read data back and check for errors against known pattern.
be used independently of readFile [0=FALSE] Can be used independently of ``writeFile``. Data checking is not timed and
NOTE: see checkWrite notes does not affect other performance timings. All errors detected are tallied
and returned as the program exit code unless ``quitOnError`` is set.
(default: 0)
* keepFile - stops removal of test file(s) on program exit [0=FALSE] * ``checkRead`` - re-read data and check for errors between reads. Can be
used independently of ``readFile``. Data checking is not timed and does not
affect other performance timings. All errors detected are tallied and
returned as the program exit code unless ``quitOnError`` is set.
(default: 0)
* keepFileWithError - ensures that with any error found in data-checking, * ``keepFile`` - do not remove test file(s) on program exit (default: 0)
the error-filled file(s) will not be deleted [0=FALSE]
* useExistingTestFile - do not remove test file before write access [0=FALSE] * ``keepFileWithError`` - do not delete any files containing errors if
detected during read-check or write-check phases. (default: 0)
* segmentCount - number of segments in file [1] * ``useExistingTestFile`` - do not remove test file(s) before write phase
NOTES: - a segment is a contiguous chunk of data (default: 0)
accessed by multiple clients each writing/
reading their own contiguous data;
comprised of blocks accessed by multiple
clients
- with HDF5 this repeats the pattern of an
entire shared dataset
* blockSize - size (in bytes) of a contiguous chunk of data * ``segmentCount`` - number of segments in file, where a segment is a
accessed by a single client; it is comprised of one contiguous chunk of data accessed by multiple clients each writing/reading
or more transfers [1048576] their own contiguous data (blocks). The exact semantics of segments
depend on the API used; for example, HDF5 repeats the pattern of an entire
shared dataset. (default: 1)
* transferSize - size (in bytes) of a single data buffer to be * ``blockSize`` - size (in bytes) of a contiguous chunk of data accessed by a
transferred in a single I/O call [262144] single client. It is comprised of one or more transfers (default: 1048576)
* verbose - output information [0] * ``transferSize`` - size (in bytes) of a single data buffer to be transferred
NOTE: this can be set to levels 0-5 on the command in a single I/O call (default: 262144)
line; repeating the -v flag will increase
verbosity level
* setTimeStampSignature - set value for time stamp signature [0] * ``verbose`` - output more information about what IOR is doing. Can be set
NOTE: used to rerun tests with the exact data to levels 0-5; repeating the -v flag will increase verbosity level.
pattern by setting data signature to contain (default: 0)
positive integer value as timestamp to be
written in data file; if set to 0, is
disabled
* showHelp - display options and help [0=FALSE] * ``setTimeStampSignature`` - Value to use for the time stamp signature. Used
to rerun tests with the exact data pattern by setting data signature to
contain positive integer value as timestamp to be written in data file; if
set to 0, is disabled (default: 0)
* storeFileOffset - use file offset as stored signature when writing * ``showHelp`` - display options and help (default: 0)
file [0=FALSE]
NOTE: this will affect performance measurements
* memoryPerNode - Allocate memory on each node to simulate real * ``storeFileOffset`` - use file offset as stored signature when writing file.
application memory usage. Accepts a percentage of This will affect performance measurements (default: 0)
node memory (e.g. "50%") on machines that support
sysconf(_SC_PHYS_PAGES) or a size. Allocation will
be split between tasks that share the node.
* memoryPerTask - Allocate secified amount of memory per task to * ``memoryPerNode`` - allocate memory on each node to simulate real
simulate real application memory usage. application memory usage or restrict page cache size. Accepts a percentage
of node memory (e.g. ``50%``) on systems that support
``sysconf(_SC_PHYS_PAGES)`` or a size. Allocation will be split between
tasks that share the node. (default: 0)
* maxTimeDuration - max time in minutes to run tests [0] * ``memoryPerTask`` - allocate specified amount of memory (in bytes) per task
NOTES: * setting this to zero (0) unsets this option to simulate real application memory usage. (default: 0)
* this option allows the current read/write
to complete without interruption
* deadlineForStonewalling - seconds before stopping write or read phase [0] * ``maxTimeDuration`` - max time (in minutes) to run all tests. Any current
NOTES: - used for measuring the amount of data moved read/write phase is not interrupted; only future I/O phases are cancelled
in a fixed time. After the barrier, each once this time is exceeded. Value of zero unsets disables. (default: 0)
task starts its own timer, begins moving
data, and the stops moving data at a pre-
arranged time. Instead of measuring the
amount of time to move a fixed amount of
data, this option measures the amount of
data moved in a fixed amount of time. The
objective is to prevent tasks slow to
complete from skewing the performance.
- setting this to zero (0) unsets this option
- this option is incompatible w/data checking
* randomOffset - access is to random, not sequential, offsets within a file [0=FALSE] * ``deadlineForStonewalling`` - seconds before stopping write or read phase.
NOTES: - this option is currently incompatible with: Used for measuring the amount of data moved in a fixed time. After the
-checkRead barrier, each task starts its own timer, begins moving data, and the stops
-storeFileOffset moving data at a pre-arranged time. Instead of measuring the amount of time
-MPIIO collective or useFileView to move a fixed amount of data, this option measures the amount of data
-HDF5 or NCMPI moved in a fixed amount of time. The objective is to prevent straggling
* summaryAlways - Always print the long summary for each test. tasks slow from skewing the performance. This option is incompatible with
Useful for long runs that may be interrupted, preventing read-check and write-check modes. Value of zero unsets this option.
the final long summary for ALL tests to be printed. (default: 0)
* ``randomOffset`` - randomize access offsets within test file(s). Currently
incompatible with ``checkRead``, ``storeFileOffset``, MPIIO ``collective``
and ``useFileView``, and HDF5 and NCMPI APIs. (default: 0)
* ``summaryAlways`` - Always print the long summary for each test even if the job is interrupted. (default: 0)
POSIX-ONLY POSIX-ONLY
^^^^^^^^^^ ^^^^^^^^^^
* useO_DIRECT - use O_DIRECT for POSIX, bypassing I/O buffers [0]
* singleXferAttempt - will not continue to retry transfer entire buffer * ``useO_DIRECT`` - use direct I/ for POSIX, bypassing I/O buffers (default: 0)
until it is transferred [0=FALSE]
NOTE: when performing a write() or read() in POSIX,
there is no guarantee that the entire
requested size of the buffer will be
transferred; this flag keeps the retrying a
single transfer until it completes or returns
an error
* fsyncPerWrite - perform fsync after each POSIX write [0=FALSE] * ``singleXferAttempt`` - do not continue to retry transfer entire buffer
* fsync - perform fsync after POSIX write close [0=FALSE] until it is transferred. When performing a write() or read() in POSIX,
there is no guarantee that the entire requested size of the buffer will be
transferred; this flag keeps the retrying a single transfer until it
completes or returns an error (default: 0)
* ``fsyncPerWrite`` - perform fsync after each POSIX write (default: 0)
* ``fsync`` - perform fsync after POSIX file close (default: 0)
MPIIO-ONLY MPIIO-ONLY
^^^^^^^^^^ ^^^^^^^^^^
* preallocate - preallocate the entire file before writing [0=FALSE]
* useFileView - use an MPI datatype for setting the file view option * ``preallocate`` - preallocate the entire file before writing (default: 0)
to use individual file pointer [0=FALSE]
NOTE: default IOR uses explicit file pointers
* useSharedFilePointer - use a shared file pointer [0=FALSE] (not working) * ``useFileView`` - use an MPI datatype for setting the file view option to
NOTE: default IOR uses explicit file pointers use individual file pointer. Default IOR uses explicit file pointers.
(default: 0)
* useStridedDatatype - create a datatype (max=2GB) for strided access; akin * ``useSharedFilePointer`` - use a shared file pointer. Default IOR uses
to MULTIBLOCK_REGION_SIZE [0] (not working) explicit file pointers. (default: 0)
* ``useStridedDatatype`` - create a datatype (max=2GB) for strided access;
akin to ``MULTIBLOCK_REGION_SIZE`` (default: 0)
HDF5-ONLY HDF5-ONLY
^^^^^^^^^ ^^^^^^^^^
* individualDataSets - within a single file each task will access its own
dataset [0=FALSE] (not working)
NOTE: default IOR creates a dataset the size of
numTasks * blockSize to be accessed by all
tasks
* noFill - no pre-filling of data in HDF5 file creation [0=FALSE] * ``individualDataSets`` - within a single file, each task will access its own
dataset. Default IOR creates a dataset the size of ``numTasks * blockSize``
to be accessed by all tasks (default: 0)
* setAlignment - HDF5 alignment in bytes (e.g.: 8, 4k, 2m, 1g) [1] * ``noFill`` - do not pre-fill data in HDF5 file creation (default: 0)
* ``setAlignment`` - set the HDF5 alignment in bytes (e.g.: 8, 4k, 2m, 1g) (default: 1)
* hdf5.collectiveMetadata - enable HDF5 collective metadata (available since HDF5-1.10.0) * hdf5.collectiveMetadata - enable HDF5 collective metadata (available since HDF5-1.10.0)
MPIIO-, HDF5-, AND NCMPI-ONLY MPIIO-, HDF5-, AND NCMPI-ONLY
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
* collective - uses collective operations for access [0=FALSE]
* showHints - show hint/value pairs attached to open file [0=FALSE] * ``collective`` - uses collective operations for access (default: 0)
NOTE: not available in NCMPI
* ``showHints`` - show hint/value pairs attached to open file. Not available
for NCMPI. (default: 0)
LUSTRE-SPECIFIC LUSTRE-SPECIFIC
^^^^^^^^^^^^^^^^^ ^^^^^^^^^^^^^^^^^
* lustreStripeCount - set the lustre stripe count for the test file(s) [0]
* lustreStripeSize - set the lustre stripe size for the test file(s) [0] * ``lustreStripeCount`` - set the Lustre stripe count for the test file(s) (default: 0)
* lustreStartOST - set the starting OST for the test file(s) [-1] * ``lustreStripeSize`` - set the Lustre stripe size for the test file(s) (default: 0)
* lustreIgnoreLocks - disable lustre range locking [0] * ``lustreStartOST`` - set the starting OST for the test file(s) (default: -1)
* ``lustreIgnoreLocks`` - disable Lustre range locking (default: 0)
GPFS-SPECIFIC GPFS-SPECIFIC
^^^^^^^^^^^^^^ ^^^^^^^^^^^^^^
* gpfsHintAccess - use gpfs_fcntl hints to pre-declare accesses
* gpfsReleaseToken - immediately after opening or creating file, release
all locks. Might help mitigate lock-revocation
traffic when many proceses write/read to same file.
* ``gpfsHintAccess`` - use ``gpfs_fcntl`` hints to pre-declare accesses (default: 0)
* ``gpfsReleaseToken`` - release all locks immediately after opening or
creating file. Might help mitigate lock-revocation traffic when many
proceses write/read to same file. (default: 0)
Verbosity levels Verbosity levels
--------------------- ----------------
The verbosity of output for IOR can be set with -v. Increasing the number of
-v instances on a command line sets the verbosity higher. The verbosity of output for IOR can be set with ``-v``. Increasing the number
of ``-v`` instances on a command line sets the verbosity higher.
Here is an overview of the information shown for different verbosity levels: Here is an overview of the information shown for different verbosity levels:
0) default; only bare essentials shown ====== ===================================
1) max clock deviation, participating tasks, free space, access pattern, Level Behavior
commence/verify access notification w/time ====== ===================================
2) rank/hostname, machine name, timer used, individual repetition 0 default; only bare essentials shown
performance results, timestamp used for data signature 1 max clock deviation, participating tasks, free space, access pattern, commence/verify access notification with time
3) full test details, transfer block/offset compared, individual data 2 rank/hostname, machine name, timer used, individual repetition performance results, timestamp used for data signature
checking errors, environment variables, task writing/reading file name, 3 full test details, transfer block/offset compared, individual data checking errors, environment variables, task writing/reading file name, all test operation times
all test operation times 4 task id and offset for each transfer
4) task id and offset for each transfer 5 each 8-byte data signature comparison (WARNING: more data to STDOUT than stored in file, use carefully)
5) each 8-byte data signature comparison (WARNING: more data to STDOUT ====== ===================================
than stored in file, use carefully)
Incompressible notes Incompressible notes
------------------------- --------------------
Please note that incompressibility is a factor of how large a block compression Please note that incompressibility is a factor of how large a block compression
algorithm uses. The incompressible buffer is filled only once before write times, algorithm uses. The incompressible buffer is filled only once before write
so if the compression algorithm takes in blocks larger than the transfer size, times, so if the compression algorithm takes in blocks larger than the transfer
there will be compression. Below are some baselines that I established for size, there will be compression. Below are some baselines for zip, gzip, and
zip, gzip, and bzip. bzip.
1) zip: For zipped files, a transfer size of 1k is sufficient. 1) zip: For zipped files, a transfer size of 1k is sufficient.
@ -354,5 +338,5 @@ zip, gzip, and bzip.
To avoid compression a transfer size of greater than the bzip block size is required To avoid compression a transfer size of greater than the bzip block size is required
(default = 900KB). I suggest a transfer size of greather than 1MB to avoid bzip2 compression. (default = 900KB). I suggest a transfer size of greather than 1MB to avoid bzip2 compression.
Be aware of the block size your compression algorithm will look at, and adjust the transfer size Be aware of the block size your compression algorithm will look at, and adjust
accordingly. the transfer size accordingly.

View File

@ -0,0 +1,82 @@
Scripting
=========
IOR can use an input script with the command line using the ``-f`` option.
**Any options on the command line set before the '-f' option is given will be
considered the default settings for running the script.** For example, ::
mpirun ./ior -W -f script
will run all tests in the script with an implicit ``-W``. The script itself can
override these settings and may be set to run many different tests of IOR under
a single execution, and it is important to note that **any command-line options
specified after ``-f`` will not be applied to the runs dictated by the script.**
For example, ::
mpirun ./ior -f script -W
will *not* run any tests with the implicit ``-W`` since that argument does not
get applied until after the ``-f`` option (and its constituent runs) are complete.
Input scripts are specified using the long-form option names that correspond to
each command-line option. In addition to long-form options,
* ``IOR START`` and ``IOR END`` mark the beginning and end of the script
* ``RUN`` dispatches the test using all of the options specified before it
* All previous set parameter stay set for the next test. They are not reset
to the default! For default the must be rest manually.
* White space is ignored in script, as are comments starting with ``#``.
* Not all test parameters need be set.
An example of a script: ::
IOR START
api=[POSIX|MPIIO|HDF5|HDFS|S3|S3_EMC|NCMPI|RADOS]
testFile=testFile
hintsFileName=hintsFile
repetitions=8
multiFile=0
interTestDelay=5
readFile=1
writeFile=1
filePerProc=0
checkWrite=0
checkRead=0
keepFile=1
quitOnError=0
segmentCount=1
blockSize=32k
outlierThreshold=0
setAlignment=1
transferSize=32
singleXferAttempt=0
individualDataSets=0
verbose=0
numTasks=32
collective=1
preallocate=0
useFileView=0
keepFileWithError=0
setTimeStampSignature=0
useSharedFilePointer=0
useStridedDatatype=0
uniqueDir=0
fsync=0
storeFileOffset=0
maxTimeDuration=60
deadlineForStonewalling=0
useExistingTestFile=0
useO_DIRECT=0
showHints=0
showHelp=0
RUN
# additional tests are optional
transferSize=64
blockSize=64k
segmentcount=2
RUN
transferSize=4K
blockSize=1M
segmentcount=1024
RUN
IOR STOP

View File

@ -1,72 +0,0 @@
Scripting
=========
IOR can use a script with the command line. Any options on the command line set
before the script will be considered the default settings for running the script.
(I.e.,'$ ./IOR -W -f script' will have all tests in the script run with the -W
option as default.)
The script itself can override these settings and may be set to run
run many different tests of IOR under a single execution.
The command line is: ::
./IOR -f script
In IOR/scripts, there are scripts of test cases for simulating I/O behavior of
various application codes. Details are included in each script as necessary.
Syntax:
* IOR START / IOR END: marks the beginning and end of the script
* RUN: Delimiter for next Test
* All previous set parameter stay set for the next test. They are not reset
to the default! For default the musst be rest manually.
* White space is ignored in script, as are comments starting with '#'.
* Not all test parameters need be set.
An example of a script: ::
IOR START
api=[POSIX|MPIIO|HDF5|HDFS|S3|S3_EMC|NCMPI|RADOS]
testFile=testFile
hintsFileName=hintsFile
repetitions=8
multiFile=0
interTestDelay=5
readFile=1
writeFile=1
filePerProc=0
checkWrite=0
checkRead=0
keepFile=1
quitOnError=0
segmentCount=1
blockSize=32k
outlierThreshold=0
setAlignment=1
transferSize=32
singleXferAttempt=0
individualDataSets=0
verbose=0
numTasks=32
collective=1
preallocate=0
useFileView=0
keepFileWithError=0
setTimeStampSignature=0
useSharedFilePointer=0
useStridedDatatype=0
uniqueDir=0
fsync=0
storeFileOffset=0
maxTimeDuration=60
deadlineForStonewalling=0
useExistingTestFile=0
useO_DIRECT=0
showHints=0
showHelp=0
RUN
# additional tests are optional
<snip>
RUN
<snip>
RUN
IOR STOP