[lustre-discuss] Lustre poor performance

Dennis Nelson dnelson at ddn.com
Thu Aug 17 21:22:24 PDT 2017


It appears that you are running iozone on a single client?  What kind of network is tcp5?  Have you looked at the network to make sure it is not the bottleneck?

-- 
Dennis Nelson
Mobile: 817-233-6116
 
Applications Support Engineer
DataDirect Networks, Inc.
dnelson at ddn.com

On 8/17/17, 10:22 PM, "lustre-discuss on behalf of Riccardo Veraldi" <lustre-discuss-bounces at lists.lustre.org on behalf of Riccardo.Veraldi at cnaf.infn.it> wrote:

    Hello,
    
    I am running Lustre 2.10.0 on Centos 7.3
    I have one MDS and two OSSes, each with one OST
    each OST is a ZFS raidz1 with 6 nvme disks each.
    The configuration of ZFS is done in a way to allow maximum write
    performances:
    
    zfs set sync=disabled drpffb-ost02
    zfs set atime=off drpffb-ost02
    zfs set redundant_metadata=most drpffb-ost02
    zfs set xattr=sa drpffb-ost02
    zfs set recordsize=1M drpffb-ost02
    
    every NVMe disk has 4K byte sector, zfs -o  ashift=12
    
    In a LOCAL raidz1 configuration I get 3.6GB/sec writings and 5GB/sec
    readings.
    
    The same configuration thru Lustre has very poor performances, 1.3GB/sec
    writes and 2GB/sec reads
    
    There must be something else to look for having better performances but
    a local ZFS raidz1 is working pretty good.
    
    this is the Lustre partition client side:
    
    172.21.42.159 at tcp5:/drpffb                  10T  279G  9.8T   3% /drpffb
    
    UUID                       bytes        Used   Available Use% Mounted on
    drpffb-MDT0000_UUID        19.1G        2.1M       19.1G   0% /drpffb[MDT:0]
    drpffb-OST0001_UUID         5.0T      142.2G        4.9T   3% /drpffb[OST:1]
    drpffb-OST0002_UUID         5.0T      136.4G        4.9T   3% /drpffb[OST:2]
    
    filesystem_summary:        10.0T      278.6G        9.7T   3% /drpffb
    
    Tests both on Lustre/ZFS and local ZFS are based on 50 threads writing
    4GB of data each and 50 threads reading using iozone:
    
    iozone  -i 0 -t 50 -i 1 -t 50 -s4g
    
    I do not know what else I can do to improve performances
    
    here some details on the OSSes
    
    OSS01:
    
    NAME                 USED  AVAIL  REFER  MOUNTPOINT
    drpffb-ost01        39.4G  4.99T   153K  none
    drpffb-ost01/ost01  39.4G  4.99T  39.4G  none
    
      pool: drpffb-ost01
     state: ONLINE
      scan: none requested
    config:
    
        NAME         STATE     READ WRITE CKSUM
        drpffb-ost01  ONLINE       0     0     0
          raidz1-0   ONLINE       0     0     0
            nvme0n1  ONLINE       0     0     0
            nvme1n1  ONLINE       0     0     0
            nvme2n1  ONLINE       0     0     0
            nvme3n1  ONLINE       0     0     0
            nvme4n1  ONLINE       0     0     0
            nvme5n1  ONLINE       0     0     0
    
    OSS02:
    
    NAME                 USED  AVAIL  REFER  MOUNTPOINT
    drpffb-ost02        62.2G  4.97T   153K  none
    drpffb-ost02/ost02  62.2G  4.97T  62.2G  none
    
      pool: drpffb-ost02
     state: ONLINE
      scan: none requested
    config:
    
        NAME         STATE     READ WRITE CKSUM
        drpffb-ost02  ONLINE       0     0     0
          raidz1-0   ONLINE       0     0     0
            nvme0n1  ONLINE       0     0     0
            nvme1n1  ONLINE       0     0     0
            nvme2n1  ONLINE       0     0     0
            nvme3n1  ONLINE       0     0     0
            nvme4n1  ONLINE       0     0     0
            nvme5n1  ONLINE       0     0     0
    
    thanks to anyone who may help giving hints.
    
    Rick
    
    
    
    _______________________________________________
    lustre-discuss mailing list
    lustre-discuss at lists.lustre.org
    http://lists.lustre.org/listinfo.cgi/lustre-discuss-lustre.org
    



More information about the lustre-discuss mailing list