[Lustre-discuss] LUSTRE Bad Read Performance

Ben Evans Ben.Evans at terascala.com
Wed Aug 11 09:56:59 PDT 2010


My guess would be that your read pattern is very random (so read-ahead isn't helping) or that you need to tune your read-ahead parameters.

 

A simple test would be to use dd to write/read a file (make sure it's in the 10s of GB to minimize cache effects) 

 

From: lustre-discuss-bounces at lists.lustre.org [mailto:lustre-discuss-bounces at lists.lustre.org] On Behalf Of Sergius ssi. Siczek - sysGen GmbH
Sent: Wednesday, August 11, 2010 10:36 AM
To: lustre-discuss at lists.lustre.org
Cc: GmbH; Dieter Nikisch - sysGen GmbH; Rolf at lists.lustre.org
Subject: [Lustre-discuss] LUSTRE Bad Read Performance

 

Hello,

 

I've got a strange problem with lustre read perforamnce.

 

Following setup:

 

8 Core (100GB RAM) combined MGS/OSS with two OST targets server, with Infiniband

 

2x RAID6  (11x 1 TB HDD -> 8TB usable on each RAID). (/dev/sdc and /dev/sdd)

 

/dev/sdc1 = 1024GB

/dev/sdd1 = 1024GB

 

1x small HDD for journal and MDS. (/dev/sdb)

 

/dev/sdb1 = 10 GB journal

/dev/sdb2 = 10GB journal

/dev/sdb3 = 50 GB MDS

 

Procedure:

 

mkfs.lustre --reformat --fsname lustre --mdt --mgs /dev/sdb3

 

mkfs.lustre --fsname=lustre --reformat --ost --mgsnode 192.168.0.1 at o2ib --mkfsoptions="-E stride=256 -E stripe-width=9 -J device=/dev/sdb1" /dev/sdc1

mkfs.lustre --fsname=lustre --reformat --ost --mgsnode 192.168.0.1 at o2ib --mkfsoptions="-E stride=256 -E stripe-width=9 -J device=/dev/sdb2" /dev/sdd1

 

 

Client machine:

 

24 core (50 GB RAM) with infiniband.

 

Mount :

mount -t lustre 192.168.0.1 at o2ib:/lustre /mnt/client/ <mailto:192.168.0.1 at o2ib:/lustre%20/mnt/client/> 

 

After that I ran some simple throughput tests. Write performance (5 threads) is about 1072 MB/s (which is great), but read performance is very bad ~ 510 MB/s.

 

What could be wrong about this setup? Why write performance is so good, but read so bad?

 

Of course I tested the hardware itself, made a filesystem on the raids (without lustre) and ran the same tests, ~1,3GB/s write 1,7GB/s read (800 GB file)

 

The infiniband is also okay, used the same cards/cable for some HPC tests just hours before, everything good. Nevertheless used the server as a client two (just to bypass the infiniband side as a potential failure). Nearly the same results (write good, read bad). It's like there is a wall at ~510 MB/s read.

 

 

Any ideas or suggestions ? Its really urgent. Thank you a lot!

 

 

2.6.27.39+lustre1.8.3+0.credativ.lenny.4

Lustre version: 1.8.3-20100410022943-PRISTINE-2.6.27.39+lustre1.8.3+0.credativ.lenny.4

 

 

 

 

 

 

 

 

sysGen GmbH

Sergius Siczek

Am Hallacker 48 · 28327 Bremen

Tel (0421) 4 09 66-32 · Fax (0421) 4 09 66-33

ssiczek at sysgen.de · www.sysgen.de <http://www.sysgen.de/> 

Geschäftsführung: Gabriele Nikisch, Anke Heinßen · Eingetragen beim Amtsgericht Walsrode · HRB 121943

 

-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.lustre.org/pipermail/lustre-discuss-lustre.org/attachments/20100811/2e2551b4/attachment.htm>


More information about the lustre-discuss mailing list