[Lustre-discuss] OST I/O problems
Andreas Dilger
adilger at sun.com
Fri Dec 4 19:59:55 PST 2009
On 2009-12-04, at 04:59, Heiko Schröter wrote:
> we do see those messages (see below) on our OSTs when under heavy
> _read_ load (or when 60+ Jobs are trying to read data at approx the
> same time).
> The OSTs freezes and even console output is down to a few bytes the
> minute.
> After some time the OSTs do revocer.
>
> But together with the 3ware support we are pretty sure to have
> replaced all snipish disks and data transfer looks ok when not used
> by lustre (i.e. verifying up to 30-90MB/s/disk throughput).
>
> OSTs with 2GB RAM on 16port controller, 4GB RAM on 24port controller
> Raid6 (write once, read often archive system)
> lustre-1.6.6
> vanilla-kernel 2.6.22.19
> 3ware 9650se (16 and 24port) latest 9.5.3 Version
> Seagate 31000340NS disks, HITACHI 1TB disks
We've had problems in the past with 3ware controllers at other sites
in the past - the performance is not as good as expected, since they
rely heavily on readahead to get good performance.
That said:
> Dec 4 12:42:56 sadosrd24 LustreError: 4744:0:(ost_handler.c:
> 882:ost_brw_read()) @@@ timeout on bulk PUT after 100+0s
> req at ffff81007efa7e00 x7869690/t0 o3->eb2e7e64-c1d9-
> d1f6-8f9d-1ba9629ff4c0 at NET_0x20000c0a8106f_UUID:0/0 lens 384/336 e 0
> to 0 dl 1259926976 ref 1 fl Interpret:/0/0 rc 0/0
This means that the IO didn't complete before the timeout. This could
be because the OST IO is so slow that no RPC can complete before the
timeout, or because there is packet loss.
Some things to try:
- reduce the number of OSS threads via module parameter:
option ost oss_num_threads=N
- increase the lustre timeout (details in the manual)
Cheers, Andreas
--
Andreas Dilger
Sr. Staff Engineer, Lustre Group
Sun Microsystems of Canada, Inc.
More information about the lustre-discuss
mailing list