[lustre-discuss] June 10th Deadline: Lustre Benchmark Submissions to IO500

OpenSFS Administration admin at opensfs.org
Thu May 30 16:19:43 PDT 2019


OpenSFS encourages the Lustre community to submit benchmarks results to the
IO500 list. Deadline is June 10th, 2019. Please see below for details and
also a talk from Andreas Dilger on IO500 titled as “IO-500 - A Storage
Benchmark for HPC” presented at LUG2019:
https://conference.cacds.uh.edu/wp-content/uploads/2019/05/LUG2019-IO500-Dil
ger.pdf

Best Regards, 

OpenSFS Administration 

 

Call for Submission

Deadline: 10 June 2019 AoE

The IO500 is now accepting and encouraging submissions for the upcoming 4th
IO500 list to be revealed at ISC-HPC 2019 in Frankfurt, Germany. Once again,
we are also accepting submissions to the 10 node I/O challenge to encourage
submission of small scale results. The new ranked lists will be announced at
our ISC19 BoF [2]. We hope to see you, and your results, there.

The benchmark suite is designed to be easy to run and the community has
multiple active support channels to help with any questions. Please submit
and we look forward to seeing many of you at ISC 2019! Please note that
submissions of all size are welcome; the site has customizable sorting so it
is possible to submit on a small system and still get a very good per-client
score for example. Additionally, the list is about much more than just the
raw rank; all submissions help the community by collecting and publishing a
wider corpus of data. More details below.

Following the success of the Top500 in collecting and analyzing historical
trends in supercomputer technology and evolution, the IO500 was created in
2017, published its first list at SC17, and has grown exponentially since
then. The need for such an initiative has long been known within
High-Performance Computing; however, defining appropriate benchmarks had
long been challenging. Despite this challenge, the community, after long and
spirited discussion, finally reached consensus on a suite of benchmarks and
a metric for resolving the scores into a single ranking.

The multi-fold goals of the benchmark suite are as follows:

        1.    Maximizing simplicity in running the benchmark suite
        2.    Encouraging complexity in tuning for performance
        3.    Allowing submitters to highlight their “hero run” performance
numbers
        4.    Forcing submitters to simultaneously report performance for
challenging IO patterns.

Specifically, the benchmark suite includes a hero-run of both IOR and mdtest
configured however possible to maximize performance and establish an
upper-bound for performance. It also includes an IOR and mdtest run with
highly prescribed parameters in an attempt to determine a lower-bound.
Finally, it includes a namespace search as this has been determined to be a
highly sought-after feature in HPC storage systems that has historically not
been well-measured. Submitters are encouraged to share their tuning insights
for publication.

The goals of the community are also multi-fold:

        1.    Gather historical data for the sake of analysis and to aid
predictions of storage futures
        2.    Collect tuning information to share valuable performance
optimizations across the community
        3.    Encourage vendors and designers to optimize for workloads
beyond “hero runs”
        4.    Establish bounded expectations for users, procurers, and
administrators
        
10 Node I/O Challenge

At ISC, we will announce our second IO-500 award for the 10 Node Challenge.
This challenge is conducted using the regular IO-500 benchmark, however,
with the rule that exactly 10 computes nodes must be used to run the
benchmark (one exception is find, which may use 1 node). You may use any
shared storage with, e.g., any number of servers. When submitting for the
IO-500 list, you can opt-in for “Participate in the 10 compute node
challenge only”, then we won't include the results into the ranked list.
Other 10 compute node submission will be included in the full list and in
the ranked list. We will announce the result in a separate derived list and
in the full list but not on the ranked IO-500 list at  <http://io500.org>
io500.org< <http://io500.org> http://io500.org>.
 
Birds-of-a-feather

Once again, we encourage you to submit [1], to join our community, and to
attend our BoF “The IO-500 and the Virtual Institute of I/O” at ISC 2019 [2]
where we will announce the fourth IO500 list and second 10 node challenge
list. The current list includes results from BeeGPFS, DataWarp, IME, Lustre,
Spectrum Scale, and WekaIO. We hope that the next list has even more.

We look forward to answering any questions or concerns you might have.

        §  [1]  <http://io500.org/submission> http://io500.org/submission
        §  [2] The BoF schedule will be announced soon”

 

-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.lustre.org/pipermail/lustre-discuss-lustre.org/attachments/20190530/83fd619b/attachment.html>


More information about the lustre-discuss mailing list