[Lustre-discuss] help

Ashok nulguda ashok0586 at gmail.com
Fri Sep 30 01:37:22 PDT 2011


Dear Sir,


Thanks for your help.

My system is ICE 8400 cluster with 30 TB of lustre of 64 node.
oss1:~ # df -h
Filesystem            Size  Used Avail Use% Mounted on
/dev/sda3             100G  5.8G   95G   6% /
tmpfs                  12G  1.1M   12G   1% /dev
tmpfs                  12G   88K   12G   1% /dev/shm
/dev/sda1            1020M  181M  840M  18% /boot
/dev/sda4             170G  6.6M  170G   1% /data1
/dev/mapper/3600a0b8000755ee0000010964dc231bc_part1
                      2.1T   74G  1.9T   4% /OST1
/dev/mapper/3600a0b8000755ed1000010614dc23425_part1
                      1.7T   67G  1.5T   5% /OST4
/dev/mapper/3600a0b8000755ee0000010a04dc23323_part1
                      2.1T   67G  1.9T   4% /OST5
/dev/mapper/3600a0b8000755f1f000011224dc239d7_part1
                      1.7T   67G  1.5T   5% /OST8
/dev/mapper/3600a0b8000755dbe000010de4dc23997_part1
                      2.1T   66G  1.9T   4% /OST9
/dev/mapper/3600a0b8000755f1f000011284dc23b5a_part1
                      1.7T   66G  1.5T   5% /OST12
/dev/mapper/3600a0b8000755eb3000011304dc23db1_part1
                      2.1T   66G  1.9T   4% /OST13
/dev/mapper/3600a0b8000755f22000011104dc23ec7_part1
                      1.7T   66G  1.5T   5% /OST16


oss1:~ # rpm -qa | grep -i lustre
kernel-default-2.6.27.39-0.3_lustre.1.8.4
kernel-ib-1.5.1-2.6.27.39_0.3_lustre.1.8.4_default
lustre-modules-1.8.4-2.6.27_39_0.3_lustre.1.8.4_default
kernel-default-base-2.6.27.39-0.3_lustre.1.8.4
lustre-1.8.4-2.6.27_39_0.3_lustre.1.8.4_default
lustre-ldiskfs-3.1.3-2.6.27_39_0.3_lustre.1.8.4_default


oss2:~ # Filesystem            Size  Used Avail Use% Mounted on
/dev/sdcw3            100G  8.3G   92G   9% /
tmpfs                  12G  1.1M   12G   1% /dev
tmpfs                  12G   88K   12G   1% /dev/shm
/dev/sdcw1           1020M  144M  876M  15% /boot
/dev/sdcw4            170G   13M  170G   1% /data1
/dev/mapper/3600a0b8000755ed10000105e4dc23397_part1
                      1.7T   69G  1.5T   5% /OST2
/dev/mapper/3600a0b8000755ee00000109b4dc232a0_part1
                      2.1T   68G  1.9T   4% /OST3
/dev/mapper/3600a0b8000755ed1000010644dc2349f_part1
                      1.7T   67G  1.5T   5% /OST6
/dev/mapper/3600a0b8000755dbe000010d94dc23873_part1
                      2.1T   67G  1.9T   4% /OST7
/dev/mapper/3600a0b8000755f1f000011254dc23add_part1
                      1.7T   66G  1.5T   5% /OST10
/dev/mapper/3600a0b8000755dbe000010e34dc23a09_part1
                      2.1T   66G  1.9T   4% /OST11
/dev/mapper/3600a0b8000755f220000110d4dc23e36_part1
                      1.7T   66G  1.5T   5% /OST14
/dev/mapper/3600a0b8000755eb3000011354dc23e39_part1
                      2.1T   66G  1.9T   4% /OST15
/dev/mapper/3600a0b8000755eb30000113a4dc23ec4_part1
                      1.4T   66G  1.3T   6% /OST17

[1]+  Done                    df -h

oss2:~ # rpm -qa | grep -i lustre
lustre-modules-1.8.4-2.6.27_39_0.3_lustre.1.8.4_default
kernel-default-base-2.6.27.39-0.3_lustre.1.8.4
kernel-default-2.6.27.39-0.3_lustre.1.8.4
kernel-ib-1.5.1-2.6.27.39_0.3_lustre.1.8.4_default
lustre-ldiskfs-3.1.3-2.6.27_39_0.3_lustre.1.8.4_default
lustre-1.8.4-2.6.27_39_0.3_lustre.1.8.4_default

mdc1:~ # Filesystem            Size  Used Avail Use% Mounted on
/dev/sde2             100G  5.2G   95G   6% /
tmpfs                  12G  184K   12G   1% /dev
tmpfs                  12G   88K   12G   1% /dev/shm
/dev/sde1            1020M  181M  840M  18% /boot
/dev/sde4             167G  196M  159G   1% /data1
/dev/mapper/3600a0b8000755f22000011134dc23f7e_part1
                      489G  2.3G  458G   1% /MDC

[1]+  Done                    df -h
mdc1:~ #


mdc1:~ # rpm -qa | grep -i lustre
lustre-1.8.4-2.6.27_39_0.3_lustre.1.8.4_default
kernel-default-2.6.27.39-0.3_lustre.1.8.4
lustre-ldiskfs-3.1.3-2.6.27_39_0.3_lustre.1.8.4_default
kernel-ib-1.5.1-2.6.27.39_0.3_lustre.1.8.4_default
lustre-modules-1.8.4-2.6.27_39_0.3_lustre.1.8.4_default
kernel-default-base-2.6.27.39-0.3_lustre.1.8.4
mdc1:~ #

mdc2:~ # Filesystem            Size  Used Avail Use% Mounted on
/dev/sde3             100G  5.0G   95G   5% /
tmpfs                  18G  184K   18G   1% /dev
tmpfs                 7.8G   88K  7.8G   1% /dev/shm
/dev/sde1            1020M  144M  876M  15% /boot
/dev/sde4             170G  6.6M  170G   1% /data1

[1]+  Done                    df -h
mdc2:~ # rpm -qqa | grep -i lustre
lustre-modules-1.8.4-2.6.27_39_0.3_lustre.1.8.4_default
kernel-default-base-2.6.27.39-0.3_lustre.1.8.4
kernel-default-2.6.27.39-0.3_lustre.1.8.4
lustre-ldiskfs-3.1.3-2.6.27_39_0.3_lustre.1.8.4_default
kernel-ib-1.5.1-2.6.27.39_0.3_lustre.1.8.4_default
lustre-1.8.4-2.6.27_39_0.3_lustre.1.8.4_default
mdc2:~ #


service0:~ # ibstat
CA 'mlx4_0'
    CA type: MT26428
    Number of ports: 2
    Firmware version: 2.7.0
    Hardware version: a0
    Node GUID: 0x0002c903000a6028
    System image GUID: 0x0002c903000a602b
    Port 1:
        State: Active
        Physical state: LinkUp
        Rate: 40
        Base lid: 9
        LMC: 0
        SM lid: 1
        Capability mask: 0x02510868
        Port GUID: 0x0002c903000a6029
    Port 2:
        State: Active
        Physical state: LinkUp
        Rate: 40
        Base lid: 10
        LMC: 0
        SM lid: 1
        Capability mask: 0x02510868
        Port GUID: 0x0002c903000a602a
service0:~ #



service0:~ # ibstatus
Infiniband device 'mlx4_0' port 1 status:
    default gid:     fec0:0000:0000:0000:0002:c903:000a:6029
    base lid:     0x9
    sm lid:         0x1
    state:         4: ACTIVE
    phys state:     5: LinkUp
    rate:         40 Gb/sec (4X QDR)

Infiniband device 'mlx4_0' port 2 status:
    default gid:     fec0:0000:0000:0000:0002:c903:000a:602a
    base lid:     0xa
    sm lid:         0x1
    state:         4: ACTIVE
    phys state:     5: LinkUp
    rate:         40 Gb/sec (4X QDR)

service0:~ #



service0:~ # ibdiagnet
Loading IBDIAGNET from: /usr/lib64/ibdiagnet1.2
-W- Topology file is not specified.
    Reports regarding cluster links will use direct routes.
Loading IBDM from: /usr/lib64/ibdm1.2
-W- A few ports of local device are up.
    Since port-num was not specified (-p option), port 1 of device 1 will be
    used as the local port.
-I- Discovering ... 88 nodes (9 Switches & 79 CA-s) discovered.


-I---------------------------------------------------
-I- Bad Guids/LIDs Info
-I---------------------------------------------------
-I- No bad Guids were found

-I---------------------------------------------------
-I- Links With Logical State = INIT
-I---------------------------------------------------
-I- No bad Links (with logical state = INIT) were found

-I---------------------------------------------------
-I- PM Counters Info
-I---------------------------------------------------
-I- No illegal PM counters values were found

-I---------------------------------------------------
-I- Fabric Partitions Report (see ibdiagnet.pkey for a full hosts list)
-I---------------------------------------------------
-I-    PKey:0x7fff Hosts:81 full:81 partial:0

-I---------------------------------------------------
-I- IPoIB Subnets Check
-I---------------------------------------------------
-I- Subnet: IPv4 PKey:0x7fff QKey:0x00000b1b MTU:2048Byte rate:10Gbps
SL:0x00
-W- Suboptimal rate for group. Lowest member rate:20Gbps > group-rate:10Gbps

-I---------------------------------------------------
-I- Bad Links Info
-I- No bad link were found
-I---------------------------------------------------
----------------------------------------------------------------
-I- Stages Status Report:
    STAGE                                    Errors Warnings
    Bad GUIDs/LIDs Check                     0      0
    Link State Active Check                  0      0
    Performance Counters Report              0      0
    Partitions Check                         0      0
    IPoIB Subnets Check                      0      1

Please see /tmp/ibdiagnet.log for complete log
----------------------------------------------------------------

-I- Done. Run time was 9 seconds.
service0:~ #


service0:~ # ibcheckerrors
#warn: counter VL15Dropped = 18584     (threshold 100) lid 1 port 1
Error check on lid 1 (r1lead HCA-1) port 1:  FAILED
#warn: counter SymbolErrors = 42829     (threshold 10) lid 9 port 1
#warn: counter RcvErrors = 9279     (threshold 10) lid 9 port 1
Error check on lid 9 (service0 HCA-1) port 1:  FAILED

## Summary: 88 nodes checked, 0 bad nodes found
##          292 ports checked, 2 ports have errors beyond threshold
service0:~ #


service0:~ # ibchecknet

# Checking Ca: nodeguid 0x0002c903000abfc2

# Checking Ca: nodeguid 0x0002c903000ac00e

# Checking Ca: nodeguid 0x0002c903000a69dc

# Checking Ca: nodeguid 0x0002c9030009cd46

# Checking Ca: nodeguid 0x003048fffff4d878

# Checking Ca: nodeguid 0x003048fffff4d880

# Checking Ca: nodeguid 0x003048fffff4d87c

# Checking Ca: nodeguid 0x003048fffff4d884

# Checking Ca: nodeguid 0x003048fffff4d888

# Checking Ca: nodeguid 0x003048fffff4d88c

# Checking Ca: nodeguid 0x003048fffff4d890

# Checking Ca: nodeguid 0x003048fffff4d894

# Checking Ca: nodeguid 0x0002c9020029fa50
#warn: counter VL15Dropped = 18617     (threshold 100) lid 1 port 1
Error check on lid 1 (r1lead HCA-1) port 1:  FAILED

# Checking Ca: nodeguid 0x0002c90300054eac

# Checking Ca: nodeguid 0x0002c9030009cebe

# Checking Ca: nodeguid 0x003048fffff4c9f8

# Checking Ca: nodeguid 0x003048fffff4db08

# Checking Ca: nodeguid 0x003048fffff4db40

# Checking Ca: nodeguid 0x003048fffff4db44

# Checking Ca: nodeguid 0x003048fffff4db48

# Checking Ca: nodeguid 0x003048fffff4db4c

# Checking Ca: nodeguid 0x003048fffff4db0c

# Checking Ca: nodeguid 0x003048fffff4dca0

# Checking Ca: nodeguid 0x0002c903000abfe2

# Checking Ca: nodeguid 0x0002c903000abfe6

# Checking Ca: nodeguid 0x0002c9030009dd28

# Checking Ca: nodeguid 0x003048fffff4db54

# Checking Ca: nodeguid 0x003048fffff4db58

# Checking Ca: nodeguid 0x003048fffff4c9f4

# Checking Ca: nodeguid 0x003048fffff4db50

# Checking Ca: nodeguid 0x003048fffff4db3c

# Checking Ca: nodeguid 0x003048fffff4db38

# Checking Ca: nodeguid 0x003048fffff4db14

# Checking Ca: nodeguid 0x003048fffff4db10

# Checking Ca: nodeguid 0x003048fffff4d8a8

# Checking Ca: nodeguid 0x003048fffff4d8ac

# Checking Ca: nodeguid 0x003048fffff4d8b4

# Checking Ca: nodeguid 0x003048fffff4d8b0

# Checking Ca: nodeguid 0x003048fffff4db70

# Checking Ca: nodeguid 0x003048fffff4db68

# Checking Ca: nodeguid 0x003048fffff4db64

# Checking Ca: nodeguid 0x003048fffff4db78

# Checking Ca: nodeguid 0x0002c903000a69f0

# Checking Ca: nodeguid 0x0002c9030006004a

# Checking Ca: nodeguid 0x0002c9030009dd2c

# Checking Ca: nodeguid 0x003048fffff4d8b8

# Checking Ca: nodeguid 0x003048fffff4d8bc

# Checking Ca: nodeguid 0x003048fffff4d8a4

# Checking Ca: nodeguid 0x003048fffff4d8a0

# Checking Ca: nodeguid 0x003048fffff4db7c

# Checking Ca: nodeguid 0x003048fffff4db80

# Checking Ca: nodeguid 0x003048fffff4db6c

# Checking Ca: nodeguid 0x003048fffff4db74

# Checking Ca: nodeguid 0x003048fffff4dcb8

# Checking Ca: nodeguid 0x003048fffff4dcd0

# Checking Ca: nodeguid 0x003048fffff4dc5c

# Checking Ca: nodeguid 0x003048fffff4dc60

# Checking Ca: nodeguid 0x003048fffff4dc54

# Checking Ca: nodeguid 0x003048fffff4dc50

# Checking Ca: nodeguid 0x003048fffff4dc4c

# Checking Ca: nodeguid 0x003048fffff4dcd4

# Checking Ca: nodeguid 0x0002c903000a6164

# Checking Ca: nodeguid 0x003048fffff4dcf0

# Checking Ca: nodeguid 0x003048fffff4db5c

# Checking Ca: nodeguid 0x003048fffff4dc90

# Checking Ca: nodeguid 0x003048fffff4dc8c

# Checking Ca: nodeguid 0x003048fffff4dc58

# Checking Ca: nodeguid 0x003048fffff4dc94

# Checking Ca: nodeguid 0x003048fffff4dc9c

# Checking Ca: nodeguid 0x003048fffff4db60

# Checking Ca: nodeguid 0x003048fffff4d89c

# Checking Ca: nodeguid 0x003048fffff4d898

# Checking Ca: nodeguid 0x003048fffff4dad8

# Checking Ca: nodeguid 0x003048fffff4dadc

# Checking Ca: nodeguid 0x003048fffff4db30

# Checking Ca: nodeguid 0x003048fffff4db34

# Checking Ca: nodeguid 0x003048fffff4d874

# Checking Ca: nodeguid 0x003048fffff4d870

# Checking Ca: nodeguid 0x0002c903000a6028
#warn: counter SymbolErrors = 44150     (threshold 10) lid 9 port 1
#warn: counter RcvErrors = 9283     (threshold 10) lid 9 port 1
Error check on lid 9 (service0 HCA-1) port 1:  FAILED

## Summary: 88 nodes checked, 0 bad nodes found
##          292 ports checked, 0 bad ports found
##          2 ports have errors beyond threshold



service0:~ # ibcheckstate

## Summary: 88 nodes checked, 0 bad nodes found
##          292 ports checked, 0 ports with bad state found
service0:~ # ibcheckwidth

## Summary: 88 nodes checked, 0 bad nodes found
##          292 ports checked, 0 ports with 1x width in error found
service0:~ #


Thanks and Regards
Ashok



On 30 September 2011 12:39, Brian O'Connor <briano at sgi.com> wrote:

>  Hello Ashok
>
> is the cluster hanging or otherwise behaving badly? The logs below show
> that the client
> lost connection to 10.148.0.106 for 10seconds or so. It should have
> recovered ok.
>
> If you want further help from the list you need to add more detail about
> the cluster i.e.
> A general description of the number of OSS/OST, clients, version of lustre
> etc, and a description
> of what is actually going wrong... ie hanging, offline etc
>
> The first thing is to check the infrastructure.. ie. in this case you
> should check your IB network for errors
>
>
>
>
> On 30-September-2011 2:39 PM, Ashok nulguda wrote:
>
> Dear All,
>
> I am having lustre error on my HPC as given below.Please any one can help
> me to resolve this problem.
> Thanks in Advance.
> Sep 30 08:40:23 service0 kernel: [343138.837222] Lustre:
> 8300:0:(client.c:1476:ptlrpc_expire_one_request()) Skipped 1 previous
> similar message
> Sep 30 08:40:23 service0 kernel: [343138.837233] Lustre:
> lustre-OST0008-osc-ffff880b272cf800: Connection to service lustre-OST0008
> via nid 10.148.0.106 at o2ib was lost; in progress operations using this
> service will wait for recovery to complete.
> Sep 30 08:40:24 service0 kernel: [343139.837260] Lustre:
> 8300:0:(client.c:1476:ptlrpc_expire_one_request()) @@@ Request
> x1380984193067288 sent from lustre-OST0006-osc-ffff880b272cf800 to NID
> 10.148.0.106 at o2ib 7s ago has timed out (7s prior to deadline).
> Sep 30 08:40:24 service0 kernel: [343139.837263]   req at ffff880a5f800c00x1380984193067288/t0 o3->
> lustre-OST0006_UUID at 10.148.0.106@o2ib:6/4 lens 448/592 e 0 to 1 dl
> 1317352224 ref 2 fl Rpc:/0/0 rc 0/0
> Sep 30 08:40:24 service0 kernel: [343139.837269] Lustre:
> 8300:0:(client.c:1476:ptlrpc_expire_one_request()) Skipped 38 previous
> similar messages
> Sep 30 08:40:24 service0 kernel: [343140.129284] LustreError:
> 9983:0:(ldlm_request.c:1025:ldlm_cli_cancel_req()) Got rc -11 from cancel
> RPC: canceling anyway
> Sep 30 08:40:24 service0 kernel: [343140.129290] LustreError:
> 9983:0:(ldlm_request.c:1025:ldlm_cli_cancel_req()) Skipped 1 previous
> similar message
> Sep 30 08:40:24 service0 kernel: [343140.129295] LustreError:
> 9983:0:(ldlm_request.c:1587:ldlm_cli_cancel_list()) ldlm_cli_cancel_list:
> -11
> Sep 30 08:40:24 service0 kernel: [343140.129299] LustreError:
> 9983:0:(ldlm_request.c:1587:ldlm_cli_cancel_list()) Skipped 1 previous
> similar message
> Sep 30 08:40:25 service0 kernel: [343140.837308] Lustre:
> 8300:0:(client.c:1476:ptlrpc_expire_one_request()) @@@ Request
> x1380984193067299 sent from lustre-OST0010-osc-ffff880b272cf800 to NID
> 10.148.0.106 at o2ib 7s ago has timed out (7s prior to deadline).
> Sep 30 08:40:25 service0 kernel: [343140.837311]   req at ffff880a557c4400x1380984193067299/t0 o3->
> lustre-OST0010_UUID at 10.148.0.106@o2ib:6/4 lens 448/592 e 0 to 1 dl
> 1317352225 ref 2 fl Rpc:/0/0 rc 0/0
> Sep 30 08:40:25 service0 kernel: [343140.837316] Lustre:
> 8300:0:(client.c:1476:ptlrpc_expire_one_request()) Skipped 4 previous
> similar messages
> Sep 30 08:40:26 service0 kernel: [343141.245365] LustreError:
> 30978:0:(ldlm_request.c:1025:ldlm_cli_cancel_req()) Got rc -11 from cancel
> RPC: canceling anyway
> Sep 30 08:40:26 service0 kernel: [343141.245371] LustreError:
> 22729:0:(ldlm_request.c:1587:ldlm_cli_cancel_list()) ldlm_cli_cancel_list:
> -11
> Sep 30 08:40:26 service0 kernel: [343141.245378] LustreError:
> 30978:0:(ldlm_request.c:1025:ldlm_cli_cancel_req()) Skipped 1 previous
> similar message
> Sep 30 08:40:33 service0 kernel: [343148.245683] Lustre:
> 22725:0:(client.c:1476:ptlrpc_expire_one_request()) @@@ Request
> x1380984193067302 sent from lustre-OST0004-osc-ffff880b272cf800 to NID
> 10.148.0.106 at o2ib 14s ago has timed out (14s prior to deadline).
> Sep 30 08:40:33 service0 kernel: [343148.245686]   req at ffff8805c879e800x1380984193067302/t0 o103->
> lustre-OST0004_UUID at 10.148.0.106@o2ib:17/18 lens 296/384 e 0 to 1 dl
> 1317352233 ref 1 fl Rpc:N/0/0 rc 0/0
> Sep 30 08:40:33 service0 kernel: [343148.245692] Lustre:
> 22725:0:(client.c:1476:ptlrpc_expire_one_request()) Skipped 2 previous
> similar messages
> Sep 30 08:40:33 service0 kernel: [343148.245708] LustreError:
> 22725:0:(ldlm_request.c:1025:ldlm_cli_cancel_req()) Got rc -11 from cancel
> RPC: canceling anyway
> Sep 30 08:40:33 service0 kernel: [343148.245714] LustreError:
> 22725:0:(ldlm_request.c:1587:ldlm_cli_cancel_list()) ldlm_cli_cancel_list:
> -11
> Sep 30 08:40:33 service0 kernel: [343148.245717] LustreError:
> 22725:0:(ldlm_request.c:1587:ldlm_cli_cancel_list()) Skipped 1 previous
> similar message
> Sep 30 08:40:36 service0 kernel: [343151.548005] LustreError: 11-0: an
> error occurred while communicating with 10.148.0.106 at o2ib. The ost_connect
> operation failed with -16
> Sep 30 08:40:36 service0 kernel: [343151.548008] LustreError: Skipped 1
> previous similar message
> Sep 30 08:40:36 service0 kernel: [343151.548024] LustreError: 167-0: This
> client was evicted by lustre-OST000b; in progress operations using this
> service will fail.
> Sep 30 08:40:36 service0 kernel: [343151.548250] LustreError:
> 30452:0:(llite_mmap.c:210:ll_tree_unlock()) couldn't unlock -5
> Sep 30 08:40:36 service0 kernel: [343151.550210] LustreError:
> 8300:0:(client.c:858:ptlrpc_import_delay_req()) @@@ IMP_INVALID
> req at ffff88049528c400 x1380984193067406/t0 o3->
> lustre-OST000b_UUID at 10.148.0.106@o2ib:6/4 lens 448/592 e 0 to 1 dl 0 ref 2
> fl Rpc:/0/0 rc 0/0
> Sep 30 08:40:36 service0 kernel: [343151.594742] Lustre:
> lustre-OST0000-osc-ffff880b272cf800: Connection restored to service
> lustre-OST0000 using nid 10.148.0.106 at o2ib.
> Sep 30 08:40:36 service0 kernel: [343151.837203] Lustre:
> lustre-OST0006-osc-ffff880b272cf800: Connection restored to service
> lustre-OST0006 using nid 10.148.0.106 at o2ib.
> Sep 30 08:40:37 service0 kernel: [343152.842631] Lustre:
> lustre-OST0003-osc-ffff880b272cf800: Connection restored to service
> lustre-OST0003 using nid 10.148.0.106 at o2ib.
> Sep 30 08:40:37 service0 kernel: [343152.842636] Lustre: Skipped 3 previous
> similar messages
>
>
> Thanks and Regards
> Ashok
>
> --
> *Ashok Nulguda
> *
> *TATA ELXSI LTD*
>  *Mb : +91 9689945767
> *
> *Email :ashokn at tataelxsi.co.in <tshrikant at tataelxsi.co.in>*
>
>
>
> _______________________________________________
> Lustre-discuss mailing listLustre-discuss at lists.lustre.orghttp://lists.lustre.org/mailman/listinfo/lustre-discuss
>
>
>
> --
> Brian O'Connor
> -------------------------------------------------
> SGI Consulting
> Email: briano at sgi.com, Mobile +61 417 746 452
> Phone: +61 3 9963 1900, Fax: +61 3 9963 1902
> 357 Camberwell Road, Camberwell, Victoria, 3124
> AUSTRALIA http://www.sgi.com/support/services
> -------------------------------------------------
>
>
>
>


-- 
*Ashok Nulguda
*
*TATA ELXSI LTD*
 *Mb : +91 9689945767
*
*Email :ashokn at tataelxsi.co.in <tshrikant at tataelxsi.co.in>*
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.lustre.org/pipermail/lustre-discuss-lustre.org/attachments/20110930/b62ae3b4/attachment.htm>


More information about the lustre-discuss mailing list