Palmetto

Institution Clemson University
Client Procs Per Node
Client Operating System Oracle Linux
Client Operating System Version 7.5
Client Kernel Version 3.10.0-862.9.1.el7.x86_64

DATA SERVER

Storage Type HDD
Volatile Memory 256GB
Storage Interface SATA
Network InfiniBand FDR
Software Version 7.1
OS Version 7.5

INFORMATION

Client Nodes 32
Client Total Procs 64
Metadata Nodes 6
Metadata Storage Devices 2
Data Nodes 6
Data Storage Devices 10

METADATA

Easy Write 29.39 kIOP/s
Easy Stat 271.81 kIOP/s
Easy Delete 15.38 kIOP/s
Hard Write 5.11 kIOP/s
Hard Read 15.34 kIOP/s
Hard Stat 70.37 kIOP/s
Hard Delete 3.73 kIOP/s

Submitted Files

io500
#!/bin/bash
#
# INSTRUCTIONS:
# Edit this file as needed for your machine.
# This simplified version is just for running on a single node.
# It is a simplified version of the site-configs/sandia/startup.sh which include SLURM directives.
# Most of the variables set in here are needed for io500_fixed.sh which gets sourced at the end of this.
# Please also edit 'extra_description' function.
#set -x

if [ "$1" == "" ]
then
	SCALE=1
else
	SCALE=$1
fi


NP=$(( $SCALE * 16 ))

echo "$SCALE processes per node for $NP processes."

set -euo pipefail  # better error handling

export OFS_MOUNT=/scratch1/jburto2

# turn these to True successively while you debug and tune this benchmark.
# for each one that you turn to true, go and edit the appropriate function.
# to find the function name, see the 'main' function.
# These are listed in the order that they run.
io500_run_ior_easy="True" # does the write phase and enables the subsequent read
io500_run_md_easy="True"  # does the creat phase and enables the subsequent stat
io500_run_ior_hard="True" # does the write phase and enables the subsequent read
io500_run_md_hard="True"  # does the creat phase and enables the subsequent read
io500_run_find="True"     
io500_run_ior_easy_read="True"
io500_run_md_easy_stat="True"
io500_run_ior_hard_read="True"
io500_run_md_hard_stat="True"
io500_run_md_hard_read="True"  
io500_run_md_easy_delete="True" # turn this off if you want to just run find by itself
io500_run_md_hard_delete="True" # turn this off if you want to just run find by itself
io500_run_mdreal="True"  # this one is optional
io500_cleanup_workdir="False"  # this flag is currently ignored. You'll need to clean up your data files manually if you want to.
io500_stonewall_timer=300 # Stonewalling timer, stop with wearout after 300s with default test, set to 0, if you never want to abort...


# to run this benchmark, find and edit each of these functions.
# please also edit 'extra_description' function to help us collect the required data.
function main {
  setup_directories
  setup_paths    
  setup_ior_easy # required if you want a complete score
  setup_ior_hard # required if you want a complete score
  setup_mdt_easy # required if you want a complete score
  setup_mdt_hard # required if you want a complete score
  setup_find     # required if you want a complete score
  setup_mdreal   # optional
  run_benchmarks
}

function setup_directories {
  # set directories for where the benchmark files are created and where the results will go.
  # If you want to set up stripe tuning on your output directories or anything similar, then this is good place to do it. 
  timestamp=`date +%Y.%m.%d-%H.%M.%S`           # create a uniquifier
  io500_workdir=$OFS_MOUNT/io500/datafiles/io500.$timestamp # directory where the data will be stored
  io500_result_dir=$PWD/results/$timestamp      # the directory where the output results will be kept

  mkdir -p $io500_workdir $io500_result_dir
  mkdir -p ${io500_workdir}/ior_easy ${io500_workdir}/ior_hard 
  mkdir -p ${io500_workdir}/mdt_easy ${io500_workdir}/mdt_hard 
# for ior_easy 1 targets. RAID stripe is 8 x 256k = 2048k
#  beegfs-ctl --setpattern --numtargets=1 --chunksize=2048k --mount=/scratch4 ${io500_workdir}/ior_easy
# stripe across all OSTs for ior_hard, 64k chunksize
# best pattern is minimal chunksize to fit one I/O in, regardless of RAID stripe.
#  beegfs-ctl --setpattern --numtargets=6 --chunksize=64k --mount=/scratch4 ${io500_workdir}/ior_hard 
# turn off striping and use small chunks for mdtest
#  beegfs-ctl --setpattern --numtargets=1 --chunksize=64k --mount=/scratch4 ${io500_workdir}/mdt_easy 
#  beegfs-ctl --setpattern --numtargets=1 --chunksize=64k --mount=/scratch4 ${io500_workdir}/mdt_hard 
}

function setup_paths {
  # Set the paths to the binaries.  If you ran ./utilities/prepare.sh successfully, then binaries are in ./bin/
  io500_ior_cmd=$PWD/bin/ior
  io500_mdtest_cmd=$PWD/bin/mdtest
  io500_mdreal_cmd=$PWD/bin/md-real-io
  io500_mpi_prefix="/usr/lib64/openmpi"
  io500_mpi_prefix="/home/jburto2/openmpi/1.10.7"
  io500_mpi_prefix="/software/openmpi/1.10.7_gcc"
  io500_mpirun="$io500_mpi_prefix/bin/mpirun"
  io500_mpienv="-x LD_LIBRARY_PATH=/opt/pbs/14.0.1-2cu/lib:/opt/orangefs/lib"
  # Run OpenMPI over ethernet to keep the IB network clear for data. Map by node to balance processes.
  # The I/O 500 benchmarks are not heavy on interprocess communication.
  #io500_mpiargs="-np $NP --mca btl_tcp_if_exclude ib0 --mca btl ^openib --map-by node --machinefile /home/jburto2/hpccnodelistmpi --prefix $io500_mpi_prefix"
  io500_mpiargs="-np $NP  ${io500_mpienv} --mca mpi_cuda_support 0 --map-by node --machinefile /home/jburto2/phase08cmpi --prefix $io500_mpi_prefix"
}

function setup_ior_easy {
# 2M writes, 128 GB per proc, file per proc. 
  io500_ior_easy_size=$((128 * 1024 / $SCALE))
  io500_ior_easy_params="-t 2m -b ${io500_ior_easy_size}m -F -a MPIIO"
   
}

function setup_mdt_easy {
# one level, 11 directories, unique dir per thread, files only at leaves.
# BeeGFS doesn't have distributed directories, so more directories = better distribution. 
  io500_mdtest_easy_params="-u -L" 
  io500_mdtest_easy_files_per_proc=800000
}

function setup_ior_hard {
  if [ "$SCALE" == "1" ] 
  then
	# One process per node is significantly faster because of buffering.
  	io500_ior_hard_writes_per_proc=2000000
  else	
  	io500_ior_hard_writes_per_proc=$(( 2000000 / $SCALE ))
  fi

  io500_ior_hard_other_options=" -a MPIIO"

}

function setup_mdt_hard {
# Multiple directories might improve mdt_hard slightly, but this test is storage bound, not md bound.
  io500_mdtest_hard_files_per_proc="$(( 150000 / $SCALE ))"
  io500_mdtest_files_per_proc=$(( 150000 / $SCALE )) 
  io500_mdtest_hard_other_options=""
}

function setup_find {
  #
  # setup the find command. This is an area where innovation is allowed.
  #    There are three default options provided. One is a serial find, one is python
  #    parallel version, one is C parallel version.  Current default is to use serial.
  #    But it is very slow. We recommend to either customize or use the C parallel version.
  #    For GPFS, we recommend to use the provided mmfind wrapper described below.
  #    Instructions below.
  #    If a custom approach is used, please provide enough info so others can reproduce.

  # the serial version that should run (SLOWLY) without modification
  #io500_find_mpi="False"
  #io500_find_cmd=$PWD/bin/sfind.sh
  #io500_find_cmd_args=""

  # a parallel version in C, the -s adds a stonewall
  #   for a real run, turn -s (stonewall) off or set it at 300 or more
  #   to prepare this (assuming you've run ./utilities/prepare.sh already):
  #   > cd build/pfind
  #   > ./prepare.sh
  #   > ./compile.sh
  #   > cp pfind ../../bin/ 
  #   If you use io500_find_mpi="True", then this will run with the same
  #   number of MPI nodes and ranks as the other phases.
  #   If you prefer another number, and fewer might be better here,
  #   Then you can set io500_find_mpi to be "False" and write a wrapper
  #   script for this which sets up MPI as you would like.  Then change
  #   io500_find_cmd to point to your wrapper script. 
  io500_find_mpi="True"
  io500_find_cmd="$PWD/bin/pfind"
  io500_find_cmd_args="-s 10000 -r $io500_result_dir/pfind_results"
  
  # for GPFS systems, you should probably use the provided mmfind wrapper 
  # if you used ./utilities/prepare.sh, you'll find this wrapper in ./bin/mmfind.sh
  #io500_find_mpi="False"
  #io500_find_cmd="$PWD/bin/mmfind.sh"
  #io500_find_cmd_args=""
}

function setup_mdreal {
  io500_mdreal_params="-P=5000 -I=1000"
}

function run_benchmarks {
  # Important: source the io500_fixed.sh script.  Do not change it. If you discover
  # a need to change it, please email the mailing list to discuss
  source ./bin/io500_fixed.sh 2>&1 | tee $io500_result_dir/io-500-summary.$timestamp.txt
}

# Add key/value pairs defining your system 
# Feel free to add extra ones if you'd like
function extra_description {
  # top level info
  io500_info_system_name='Palmetto scratch1'      # e.g. Oakforest-PACS
  io500_info_institute_name='Clemson University'   # e.g. JCAHPC
  io500_info_storage_age_in_months='3' # not install date but age since last refresh
  io500_info_storage_install_date='08/12'  # MM/YY
  io500_info_filesysem='OrangeFS'     # e.g. BeeGFS, DataWarp, GPFS, IME, Lustre
  io500_info_filesystem_version='2.9.7'
  # client side info
  io500_info_num_client_nodes='16'
  io500_info_procs_per_node="${SCALE}"
  # server side info
  io500_info_num_metadata_server_nodes='16'
  io500_info_num_data_server_nodes='16'
  io500_info_num_data_storage_devices='32'  # if you have 5 data servers, and each has 5 drives, then this number is 25
  io500_info_num_metadata_storage_devices='128'  # if you have 2 metadata servers, and each has 5 drives, then this number is 10
  io500_info_data_storage_type='HDD' # HDD, SSD, persistent memory, etc, feel free to put specific models
  io500_info_metadata_storage_type='SSD' # HDD, SSD, persistent memory, etc, feel free to put specific models
  io500_info_storage_network='ethernet' # infiniband, omnipath, ethernet, etc
  io500_info_storage_interface='SAS' # SAS, SATA, NVMe, etc
  # miscellaneous
  io500_info_whatever='production'
}

main
ior_easy_read
IOR-3.2.0: MPI Coordinated Test of Parallel I/O
Began               : Wed Oct 24 22:47:33 2018
Command line        : /home/jburto2/io-500-dev/bin/ior -r -R -C -Q 1 -g -G 27 -k -e -t 4m -b 30720m -F -a POSIX -o /scratch4/jburto2/io500/datafiles/io500.2018.10.24-21.18.36/ior_easy/ior_file_easy -O stoneWallingStatusFile=/scratch4/jburto2/io500/datafiles/io500.2018.10.24-21.18.36/ior_easy/stonewall
Machine             : Linux ofstest008.ofsdev.clemson.edu
TestID              : 0
StartTime           : Wed Oct 24 22:47:33 2018
Path                : /scratch4/jburto2/io500/datafiles/io500.2018.10.24-21.18.36/ior_easy
FS                  : 174.6 TiB   Used FS: 26.6%   Inodes: 0.0 Mi   Used Inodes: -nan%

Options: 
api                 : POSIX
apiVersion          : 
test filename       : /scratch4/jburto2/io500/datafiles/io500.2018.10.24-21.18.36/ior_easy/ior_file_easy
access              : file-per-process
type                : independent
segments            : 1
ordering in a file  : sequential
ordering inter file : constant task offset
task offset         : 1
tasks               : 64
clients per node    : 2
repetitions         : 1
xfersize            : 4 MiB
blocksize           : 30 GiB
aggregate filesize  : 1.88 TiB

Results: 

access    bw(MiB/s)  block(KiB) xfer(KiB)  open(s)    wr/rd(s)   close(s)   total(s)   iter
------    ---------  ---------- ---------  --------   --------   --------   --------   ----
read      3194.89    31457280   4096       0.006617   615.24     0.157298   615.38     0   
Max Read:  3194.89 MiB/sec (3350.08 MB/sec)

Summary of all tests:
Operation   Max(MiB)   Min(MiB)  Mean(MiB)     StdDev   Max(OPs)   Min(OPs)  Mean(OPs)     StdDev    Mean(s) Test# #Tasks tPN reps fPP reord reordoff reordrand seed segcnt   blksiz    xsize aggs(MiB)   API RefNum
read         3194.89    3194.89    3194.89       0.00     798.72     798.72     798.72       0.00  615.38302     0     64   2    1   1     1        1         0    0      1 32212254720  4194304 1966080.0 POSIX      0
Finished            : Wed Oct 24 22:57:49 2018
ior_easy_write
IOR-3.2.0: MPI Coordinated Test of Parallel I/O
Began               : Wed Oct 24 21:18:39 2018
Command line        : /home/jburto2/io-500-dev/bin/ior -w -C -Q 1 -g -G 27 -k -e -t 4m -b 30720m -F -a POSIX -o /scratch4/jburto2/io500/datafiles/io500.2018.10.24-21.18.36/ior_easy/ior_file_easy -O stoneWallingStatusFile=/scratch4/jburto2/io500/datafiles/io500.2018.10.24-21.18.36/ior_easy/stonewall -O stoneWallingWearOut=1 -D 300
Machine             : Linux ofstest008.ofsdev.clemson.edu
TestID              : 0
StartTime           : Wed Oct 24 21:18:39 2018
Path                : /scratch4/jburto2/io500/datafiles/io500.2018.10.24-21.18.36/ior_easy
FS                  : 174.6 TiB   Used FS: 25.1%   Inodes: 0.0 Mi   Used Inodes: -nan%

Options: 
api                 : POSIX
apiVersion          : 
test filename       : /scratch4/jburto2/io500/datafiles/io500.2018.10.24-21.18.36/ior_easy/ior_file_easy
access              : file-per-process
type                : independent
segments            : 1
ordering in a file  : sequential
ordering inter file : constant task offset
task offset         : 1
tasks               : 64
clients per node    : 2
repetitions         : 1
xfersize            : 4 MiB
blocksize           : 30 GiB
aggregate filesize  : 1.88 TiB
stonewallingTime    : 300
stoneWallingWearOut : 1

Results: 

access    bw(MiB/s)  block(KiB) xfer(KiB)  open(s)    wr/rd(s)   close(s)   total(s)   iter
------    ---------  ---------- ---------  --------   --------   --------   --------   ----
stonewalling pairs accessed min: 6476 max: 7680 -- min data: 25.3 GiB mean data: 27.8 GiB time: 300.1s
write     3771       31457280   4096       0.090773   521.13     0.207138   521.38     0   
Max Write: 3770.89 MiB/sec (3954.07 MB/sec)

Summary of all tests:
Operation   Max(MiB)   Min(MiB)  Mean(MiB)     StdDev   Max(OPs)   Min(OPs)  Mean(OPs)     StdDev    Mean(s) Test# #Tasks tPN reps fPP reord reordoff reordrand seed segcnt   blksiz    xsize aggs(MiB)   API RefNum
write        3770.89    3770.89    3770.89       0.00     942.72     942.72     942.72       0.00  521.38326     0     64   2    1   1     1        1         0    0      1 32212254720  4194304 1966080.0 POSIX      0
Finished            : Wed Oct 24 21:27:21 2018
ior_hard_read
IOR-3.2.0: MPI Coordinated Test of Parallel I/O
Began               : Wed Oct 24 22:59:46 2018
Command line        : /home/jburto2/io-500-dev/bin/ior -r -R -C -Q 1 -g -G 27 -k -e -t 47008 -b 47008 -s 550000 -a POSIX -o /scratch4/jburto2/io500/datafiles/io500.2018.10.24-21.18.36/ior_hard/IOR_file -O stoneWallingStatusFile=/scratch4/jburto2/io500/datafiles/io500.2018.10.24-21.18.36/ior_hard/stonewall
Machine             : Linux ofstest008.ofsdev.clemson.edu
TestID              : 0
StartTime           : Wed Oct 24 22:59:46 2018
Path                : /scratch4/jburto2/io500/datafiles/io500.2018.10.24-21.18.36/ior_hard
FS                  : 174.6 TiB   Used FS: 26.6%   Inodes: 0.0 Mi   Used Inodes: -nan%

Options: 
api                 : POSIX
apiVersion          : 
test filename       : /scratch4/jburto2/io500/datafiles/io500.2018.10.24-21.18.36/ior_hard/IOR_file
access              : single-shared-file
type                : independent
segments            : 550000
ordering in a file  : sequential
ordering inter file : constant task offset
task offset         : 1
tasks               : 64
clients per node    : 2
repetitions         : 1
xfersize            : 47008 bytes
blocksize           : 47008 bytes
aggregate filesize  : 1.50 TiB

Results: 

access    bw(MiB/s)  block(KiB) xfer(KiB)  open(s)    wr/rd(s)   close(s)   total(s)   iter
------    ---------  ---------- ---------  --------   --------   --------   --------   ----
WARNING: Expected aggregate file size       = 1654681600000.
WARNING: Stat() of aggregate file size      = 889905815552.
WARNING: Using actual aggregate bytes moved = 889905815552.
read      878.97     45.91      45.91      0.031779   965.46     0.092257   965.54     0   
Max Read:  878.97 MiB/sec (921.67 MB/sec)

Summary of all tests:
Operation   Max(MiB)   Min(MiB)  Mean(MiB)     StdDev   Max(OPs)   Min(OPs)  Mean(OPs)     StdDev    Mean(s) Test# #Tasks tPN reps fPP reord reordoff reordrand seed segcnt   blksiz    xsize aggs(MiB)   API RefNum
read          878.97     878.97     878.97       0.00   19606.61   19606.61   19606.61       0.00  965.53897     0     64   2    1   0     1        1         0    0 550000    47008    47008  848680.3 POSIX      0
Finished            : Wed Oct 24 23:15:52 2018
ior_hard_write
IOR-3.2.0: MPI Coordinated Test of Parallel I/O
Began               : Wed Oct 24 21:44:52 2018
Command line        : /home/jburto2/io-500-dev/bin/ior -w -C -Q 1 -g -G 27 -k -e -t 47008 -b 47008 -s 550000 -a POSIX -o /scratch4/jburto2/io500/datafiles/io500.2018.10.24-21.18.36/ior_hard/IOR_file -O stoneWallingStatusFile=/scratch4/jburto2/io500/datafiles/io500.2018.10.24-21.18.36/ior_hard/stonewall -O stoneWallingWearOut=1 -D 300
Machine             : Linux ofstest008.ofsdev.clemson.edu
TestID              : 0
StartTime           : Wed Oct 24 21:44:52 2018
Path                : /scratch4/jburto2/io500/datafiles/io500.2018.10.24-21.18.36/ior_hard
FS                  : 174.6 TiB   Used FS: 26.1%   Inodes: 0.0 Mi   Used Inodes: -nan%

Options: 
api                 : POSIX
apiVersion          : 
test filename       : /scratch4/jburto2/io500/datafiles/io500.2018.10.24-21.18.36/ior_hard/IOR_file
access              : single-shared-file
type                : independent
segments            : 550000
ordering in a file  : sequential
ordering inter file : constant task offset
task offset         : 1
tasks               : 64
clients per node    : 2
repetitions         : 1
xfersize            : 47008 bytes
blocksize           : 47008 bytes
aggregate filesize  : 1.50 TiB
stonewallingTime    : 300
stoneWallingWearOut : 1

Results: 

access    bw(MiB/s)  block(KiB) xfer(KiB)  open(s)    wr/rd(s)   close(s)   total(s)   iter
------    ---------  ---------- ---------  --------   --------   --------   --------   ----
stonewalling pairs accessed min: 922 max: 295796 -- min data: 0.0 GiB mean data: 7.9 GiB time: 303.4s
WARNING: Expected aggregate file size       = 1654681600000.
WARNING: Stat() of aggregate file size      = 889905815552.
WARNING: Using actual aggregate bytes moved = 889905815552.
WARNING: maybe caused by deadlineForStonewalling
write     254.98     45.91      45.91      0.168897   3327.36    0.919621   3328.44    0   
Max Write: 254.98 MiB/sec (267.36 MB/sec)

Summary of all tests:
Operation   Max(MiB)   Min(MiB)  Mean(MiB)     StdDev   Max(OPs)   Min(OPs)  Mean(OPs)     StdDev    Mean(s) Test# #Tasks tPN reps fPP reord reordoff reordrand seed segcnt   blksiz    xsize aggs(MiB)   API RefNum
write         254.98     254.98     254.98       0.00    5687.63    5687.63    5687.63       0.00 3328.44089     0     64   2    1   0     1        1         0    0 550000    47008    47008  848680.3 POSIX      0
Finished            : Wed Oct 24 22:40:21 2018
mdtest_easy_delete
-- started at 10/24/2018 23:16:19 --

mdtest-1.9.3 was launched with 64 total task(s) on 32 node(s)
Command line used: /home/jburto2/io-500-dev/bin/mdtest "-r" "-F" "-d" "/scratch4/jburto2/io500/datafiles/io500.2018.10.24-21.18.36/mdt_easy" "-n" "800000" "-u" "-L" "-x" "/scratch4/jburto2/io500/datafiles/io500.2018.10.24-21.18.36/mdt_easy-stonewall"
Path: /scratch4/jburto2/io500/datafiles/io500.2018.10.24-21.18.36
FS: 174.6 TiB   Used FS: 26.6%   Inodes: 0.0 Mi   Used Inodes: -nan%

64 tasks, 51200000 files

SUMMARY rate: (of 1 iterations)
   Operation                      Max            Min           Mean        Std Dev
   ---------                      ---            ---           ----        -------
   File creation     :          0.000          0.000          0.000          0.000
   File stat         :          0.000          0.000          0.000          0.000
   File read         :          0.000          0.000          0.000          0.000
   File removal      :      15376.819      15376.819      15376.819          0.000
   Tree creation     :          0.000          0.000          0.000          0.000
   Tree removal      :          0.072          0.072          0.072          0.000

-- finished at 10/24/2018 23:49:55 --
mdtest_easy_stat
-- started at 10/24/2018 22:57:51 --

mdtest-1.9.3 was launched with 64 total task(s) on 32 node(s)
Command line used: /home/jburto2/io-500-dev/bin/mdtest "-T" "-F" "-d" "/scratch4/jburto2/io500/datafiles/io500.2018.10.24-21.18.36/mdt_easy" "-n" "800000" "-u" "-L" "-x" "/scratch4/jburto2/io500/datafiles/io500.2018.10.24-21.18.36/mdt_easy-stonewall"
Path: /scratch4/jburto2/io500/datafiles/io500.2018.10.24-21.18.36
FS: 174.6 TiB   Used FS: 26.6%   Inodes: 0.0 Mi   Used Inodes: -nan%

64 tasks, 51200000 files

SUMMARY rate: (of 1 iterations)
   Operation                      Max            Min           Mean        Std Dev
   ---------                      ---            ---           ----        -------
   File creation     :          0.000          0.000          0.000          0.000
   File stat         :     271810.008     271810.008     271810.008          0.000
   File read         :          0.000          0.000          0.000          0.000
   File removal      :          0.000          0.000          0.000          0.000
   Tree creation     :          0.000          0.000          0.000          0.000
   Tree removal      :          0.000          0.000          0.000          0.000

-- finished at 10/24/2018 22:59:44 --
mdtest_easy_write
-- started at 10/24/2018 21:27:23 --

mdtest-1.9.3 was launched with 64 total task(s) on 32 node(s)
Command line used: /home/jburto2/io-500-dev/bin/mdtest "-C" "-F" "-d" "/scratch4/jburto2/io500/datafiles/io500.2018.10.24-21.18.36/mdt_easy" "-n" "800000" "-u" "-L" "-x" "/scratch4/jburto2/io500/datafiles/io500.2018.10.24-21.18.36/mdt_easy-stonewall" "-W" "300"
Path: /scratch4/jburto2/io500/datafiles/io500.2018.10.24-21.18.36
FS: 174.6 TiB   Used FS: 26.1%   Inodes: 0.0 Mi   Used Inodes: -nan%

64 tasks, 51200000 files
stonewall rank 21: 436865 of 480777 
stonewall rank 58: 262393 of 480777 
stonewall rank 7: 281885 of 480777 
stonewall rank 15: 271289 of 480777 
stonewall rank 41: 332006 of 480777 
stonewall rank 1: 232415 of 480777 
stonewall rank 3: 314188 of 480777 
stonewall rank 37: 252365 of 480777 
stonewall rank 44: 293177 of 480777 
stonewall rank 32: 323771 of 480777 
stonewall rank 38: 354547 of 480777 
stonewall rank 52: 195303 of 480777 
stonewall rank 45: 344547 of 480777 
stonewall rank 61: 201169 of 480777 
stonewall rank 19: 215394 of 480777 
stonewall rank 9: 232740 of 480777 
stonewall rank 57: 188025 of 480777 
stonewall rank 24: 275604 of 480777 
stonewall rank 25: 297695 of 480777 
stonewall rank 42: 329644 of 480777 
stonewall rank 11: 242146 of 480777 
stonewall rank 53: 132772 of 480777 
stonewall rank 17: 308064 of 480777 
stonewall rank 2: 311763 of 480777 
stonewall rank 40: 348027 of 480777 
stonewall rank 33: 294010 of 480777 
stonewall rank 59: 271797 of 480777 
stonewall rank 39: 261155 of 480777 
stonewall rank 46: 296338 of 480777 
stonewall rank 49: 332164 of 480777 
stonewall rank 56: 278901 of 480777 
stonewall rank 62: 308429 of 480777 
stonewall rank 5: 264521 of 480777 
stonewall rank 8: 312935 of 480777 
stonewall rank 36: 281085 of 480777 
stonewall rank 63: 281607 of 480777 
stonewall rank 43: 334893 of 480777 
stonewall rank 28: 442619 of 480777 
stonewall rank 26: 304988 of 480777 
stonewall rank 4: 232579 of 480777 
stonewall rank 48: 352180 of 480777 
stonewall rank 54: 227373 of 480777 
stonewall rank 29: 325807 of 480777 
stonewall rank 13: 271192 of 480777 
stonewall rank 30: 324751 of 480777 
stonewall rank 47: 293594 of 480777 
stonewall rank 22: 440092 of 480777 
stonewall rank 34: 245059 of 480777 
Continue stonewall hit min: 132772 max: 480777 avg: 294080.7 
stonewall rank 0: 233046 of 480777 
stonewall rank 12: 429466 of 480777 
stonewall rank 16: 151585 of 480777 
stonewall rank 18: 264643 of 480777 
stonewall rank 20: 295524 of 480777 
stonewall rank 35: 324403 of 480777 
stonewall rank 51: 321595 of 480777 
stonewall rank 27: 308284 of 480777 
stonewall rank 55: 228498 of 480777 
stonewall rank 6: 432047 of 480777 
stonewall rank 10: 239768 of 480777 
stonewall rank 31: 296469 of 480777 
stonewall rank 14: 269464 of 480777 
stonewall rank 23: 290620 of 480777 
stonewall rank 60: 299113 of 480777 

SUMMARY rate: (of 1 iterations)
   Operation                      Max            Min           Mean        Std Dev
   ---------                      ---            ---           ----        -------
   File creation     :      48942.324      48942.324      48942.324          0.000
   File stat         :          0.000          0.000          0.000          0.000
   File read         :          0.000          0.000          0.000          0.000
   File removal      :          0.000          0.000          0.000          0.000
   Tree creation     :          1.801          1.801          1.801          0.000
   Tree removal      :          0.000          0.000          0.000          0.000

-- finished at 10/24/2018 21:44:50 --
mdtest_hard_delete
-- started at 10/24/2018 23:51:43 --

mdtest-1.9.3 was launched with 64 total task(s) on 32 node(s)
Command line used: /home/jburto2/io-500-dev/bin/mdtest "-r" "-t" "-F" "-w" "3901" "-e" "3901" "-d" "/scratch4/jburto2/io500/datafiles/io500.2018.10.24-21.18.36/mdt_hard" "-n" "37500" "-x" "/scratch4/jburto2/io500/datafiles/io500.2018.10.24-21.18.36/mdt_hard-stonewall"
Path: /scratch4/jburto2/io500/datafiles/io500.2018.10.24-21.18.36
FS: 174.6 TiB   Used FS: 26.6%   Inodes: 0.0 Mi   Used Inodes: -nan%

64 tasks, 2400000 files

SUMMARY rate: (of 1 iterations)
   Operation                      Max            Min           Mean        Std Dev
   ---------                      ---            ---           ----        -------
   File creation     :          0.000          0.000          0.000          0.000
   File stat         :          0.000          0.000          0.000          0.000
   File read         :          0.000          0.000          0.000          0.000
   File removal      :       3725.330       3725.330       3725.330          0.000
   Tree creation     :          0.000          0.000          0.000          0.000
   Tree removal      :          0.477          0.477          0.477          0.000

-- finished at 10/24/2018 23:58:53 --
mdtest_hard_read
-- started at 10/24/2018 23:49:57 --

mdtest-1.9.3 was launched with 64 total task(s) on 32 node(s)
Command line used: /home/jburto2/io-500-dev/bin/mdtest "-E" "-t" "-F" "-w" "3901" "-e" "3901" "-d" "/scratch4/jburto2/io500/datafiles/io500.2018.10.24-21.18.36/mdt_hard" "-n" "37500" "-x" "/scratch4/jburto2/io500/datafiles/io500.2018.10.24-21.18.36/mdt_hard-stonewall"
Path: /scratch4/jburto2/io500/datafiles/io500.2018.10.24-21.18.36
FS: 174.6 TiB   Used FS: 26.6%   Inodes: 0.0 Mi   Used Inodes: -nan%

64 tasks, 2400000 files

SUMMARY rate: (of 1 iterations)
   Operation                      Max            Min           Mean        Std Dev
   ---------                      ---            ---           ----        -------
   File creation     :          0.000          0.000          0.000          0.000
   File stat         :          0.000          0.000          0.000          0.000
   File read         :      15337.914      15337.914      15337.914          0.000
   File removal      :          0.000          0.000          0.000          0.000
   Tree creation     :          0.000          0.000          0.000          0.000
   Tree removal      :          0.000          0.000          0.000          0.000

-- finished at 10/24/2018 23:51:41 --
mdtest_hard_stat
-- started at 10/24/2018 23:15:54 --

mdtest-1.9.3 was launched with 64 total task(s) on 32 node(s)
Command line used: /home/jburto2/io-500-dev/bin/mdtest "-T" "-t" "-F" "-w" "3901" "-e" "3901" "-d" "/scratch4/jburto2/io500/datafiles/io500.2018.10.24-21.18.36/mdt_hard" "-n" "37500" "-x" "/scratch4/jburto2/io500/datafiles/io500.2018.10.24-21.18.36/mdt_hard-stonewall"
Path: /scratch4/jburto2/io500/datafiles/io500.2018.10.24-21.18.36
FS: 174.6 TiB   Used FS: 26.6%   Inodes: 0.0 Mi   Used Inodes: -nan%

64 tasks, 2400000 files

SUMMARY rate: (of 1 iterations)
   Operation                      Max            Min           Mean        Std Dev
   ---------                      ---            ---           ----        -------
   File creation     :          0.000          0.000          0.000          0.000
   File stat         :      70374.295      70374.295      70374.295          0.000
   File read         :          0.000          0.000          0.000          0.000
   File removal      :          0.000          0.000          0.000          0.000
   Tree creation     :          0.000          0.000          0.000          0.000
   Tree removal      :          0.000          0.000          0.000          0.000

-- finished at 10/24/2018 23:16:17 --
mdtest_hard_write
-- started at 10/24/2018 22:40:23 --

mdtest-1.9.3 was launched with 64 total task(s) on 32 node(s)
Command line used: /home/jburto2/io-500-dev/bin/mdtest "-C" "-t" "-F" "-w" "3901" "-e" "3901" "-d" "/scratch4/jburto2/io500/datafiles/io500.2018.10.24-21.18.36/mdt_hard" "-n" "37500" "-x" "/scratch4/jburto2/io500/datafiles/io500.2018.10.24-21.18.36/mdt_hard-stonewall" "-W" "300"
Path: /scratch4/jburto2/io500/datafiles/io500.2018.10.24-21.18.36
FS: 174.6 TiB   Used FS: 26.6%   Inodes: 0.0 Mi   Used Inodes: -nan%

64 tasks, 2400000 files
stonewall rank 21: 23793 of 24919 
stonewall rank 58: 23382 of 24919 
stonewall rank 38: 24746 of 24919 
stonewall rank 7: 24584 of 24919 
stonewall rank 3: 24521 of 24919 
stonewall rank 15: 23275 of 24919 
stonewall rank 37: 24833 of 24919 
stonewall rank 1: 24645 of 24919 
stonewall rank 44: 24766 of 24919 
stonewall rank 19: 23294 of 24919 
stonewall rank 39: 24757 of 24919 
stonewall rank 5: 24611 of 24919 
stonewall rank 52: 22969 of 24919 
stonewall rank 61: 23314 of 24919 
stonewall rank 32: 24787 of 24919 
stonewall rank 40: 24754 of 24919 
stonewall rank 53: 21891 of 24919 
stonewall rank 17: 23378 of 24919 
stonewall rank 24: 23715 of 24919 
stonewall rank 2: 24802 of 24919 
stonewall rank 11: 24708 of 24919 
stonewall rank 54: 23179 of 24919 
stonewall rank 41: 24825 of 24919 
stonewall rank 36: 24749 of 24919 
stonewall rank 55: 23012 of 24919 
stonewall rank 46: 24751 of 24919 
stonewall rank 56: 23160 of 24919 
stonewall rank 33: 24797 of 24919 
stonewall rank 50: 24777 of 24919 
stonewall rank 4: 24684 of 24919 
stonewall rank 48: 24717 of 24919 
stonewall rank 28: 24843 of 24919 
stonewall rank 49: 24714 of 24919 
stonewall rank 9: 24738 of 24919 
stonewall rank 35: 24727 of 24919 
stonewall rank 42: 24682 of 24919 
stonewall rank 18: 23149 of 24919 
stonewall rank 63: 23335 of 24919 
stonewall rank 45: 24754 of 24919 
stonewall rank 59: 23358 of 24919 
stonewall rank 8: 24573 of 24919 
stonewall rank 34: 24681 of 24919 
stonewall rank 57: 23032 of 24919 
stonewall rank 26: 23856 of 24919 
stonewall rank 12: 24733 of 24919 
stonewall rank 13: 24684 of 24919 
stonewall rank 43: 24772 of 24919 
stonewall rank 22: 23898 of 24919 
stonewall rank 25: 23931 of 24919 
stonewall rank 20: 23073 of 24919 
stonewall rank 30: 24756 of 24919 
stonewall rank 47: 24774 of 24919 
stonewall rank 62: 23364 of 24919 
Continue stonewall hit min: 21891 max: 24919 avg: 24192.0 
stonewall rank 0: 24682 of 24919 
stonewall rank 29: 24901 of 24919 
stonewall rank 51: 24794 of 24919 
stonewall rank 10: 24672 of 24919 
stonewall rank 6: 24794 of 24919 
stonewall rank 27: 24838 of 24919 
stonewall rank 23: 23791 of 24919 
stonewall rank 14: 24748 of 24919 
stonewall rank 60: 23444 of 24919 
stonewall rank 16: 22104 of 24919 

SUMMARY rate: (of 1 iterations)
   Operation                      Max            Min           Mean        Std Dev
   ---------                      ---            ---           ----        -------
   File creation     :       7693.147       7693.147       7693.147          0.000
   File stat         :          0.000          0.000          0.000          0.000
   File read         :          0.000          0.000          0.000          0.000
   File removal      :          0.000          0.000          0.000          0.000
   Tree creation     :         76.665         76.665         76.665          0.000
   Tree removal      :          0.000          0.000          0.000          0.000

-- finished at 10/24/2018 22:45:35 --
result_summary
[RESULT] BW   phase 1            ior_easy_write                3.683 GB/s : time 521.38 seconds
[RESULT] IOPS phase 1         mdtest_easy_write               48.942 kiops : time 1049.16 seconds
[RESULT] BW   phase 2            ior_hard_write                0.249 GB/s : time 3328.44 seconds
[RESULT] IOPS phase 2         mdtest_hard_write                7.693 kiops : time 314.21 seconds
[RESULT] IOPS phase 3                      find              279.680 kiops : time 115.72 seconds
[RESULT] BW   phase 3             ior_easy_read                3.120 GB/s : time 615.38 seconds
[RESULT] IOPS phase 4          mdtest_easy_stat              271.810 kiops : time 115.65 seconds
[RESULT] BW   phase 4             ior_hard_read                0.858 GB/s : time 965.54 seconds
[RESULT] IOPS phase 5          mdtest_hard_stat               70.374 kiops : time  25.09 seconds
[RESULT] IOPS phase 6        mdtest_easy_delete               15.377 kiops : time 2017.56 seconds
[RESULT] IOPS phase 7          mdtest_hard_read               15.338 kiops : time 106.23 seconds
[RESULT] IOPS phase 8        mdtest_hard_delete                3.725 kiops : time 432.54 seconds
[SCORE] Bandwidth 1.25184 GB/s : IOPS 33.9618 kiops : TOTAL 6.52033