[Lustre-devel] Thoughts on benchmarking
Andreas Dilger
adilger at sun.com
Fri Feb 29 13:10:14 PST 2008
On Feb 28, 2008 20:20 -0700, Peter J. Braam wrote:
> I see some worrying dips in the graphs - can our I/O specialists comment on
> which ones are understood and which are not?
I think one of the major problems that Andrew discusses at the end of
the test runs is described in bug 7365 "Poor performance when files share
an OSC". I didn't see anywhere in the paper which version of Lustre was
being tested, but I know we did some work to improve the round-robin
allocator to make it more uniform in more recent releases, up to a
certain extent.
That said, getting completely uniform file distribution will still need
some effort, because the MDS doesn't do any correlation between create
requests (e.g. from a single job, from a single client, etc).
> On 2/25/08 2:44 PM, "Andrew C. Uselton" <acuselton at lbl.gov> wrote:
> > I'd been in conversation with Cliff White over the last few weeks, and
> > he'd expressed an interest in having me post a draft of a report I've
> > been working on. If you've already heard of it here it is. For those
> > who hadn't I'll try to describe it briefly.
> >
> > In December I assisted with some Lustre benchmark tests on the
> > Franklin Cray XT here at NERSC. Since then I've tried to summarize our
> > analysis and results. The attached pdf is a draft of that summary. The
> > introduction is almost completely useless, so feel free to skip (unless
> > you want to have a laugh at the author's expense). Section 3 has the
> > main details about what we observed and what we thought about it.
> > Section 2 may be amusing for those (like me) who care about methodology.
Cheers, Andreas
--
Andreas Dilger
Sr. Staff Engineer, Lustre Group
Sun Microsystems of Canada, Inc.
More information about the lustre-devel
mailing list