[lustre-discuss] Install issues on 2.10.0

Ben Evans bevans at cray.com
Tue Jul 25 10:12:30 PDT 2017


health_check moved to /sys/fs/lustre/ along with a bunch of other things.

-Ben

On 7/25/17, 12:21 PM, "lustre-discuss on behalf of John Casu"
<lustre-discuss-bounces at lists.lustre.org on behalf of
john at chiraldynamics.com> wrote:

>Just installed latest 2.10.0 Lustre over ZFS on a vanilla Centos 7.3.1611
>system, using dkms.
>ZFS is 0.6.5.11 from zfsonlinux.org, installed w. yum
>
>Not a single problem during installation, but I am having issues building
>a lustre filesystem:
>1. Building a separate mgt doesn't seem to work properly, although the
>mgt/mdt combo
>    seems to work just fine.
>2. I get spl_hostid not set warnings, which I've never seen before
>3. /proc/fs/lustre/health_check seems to be missing.
>
>thanks,
>-john c
>
>
>
>---------
>Building an mgt by itself doesn't seem to work properly:
>
>> [root at fb-lts-mds0 x86_64]# mkfs.lustre --reformat --mgs
>>--force-nohostid --servicenode=192.168.98.113 at tcp \
>>                                        --backfstype=zfs mgs/mgt
>> 
>>    Permanent disk data:
>> Target:     MGS
>> Index:      unassigned
>> Lustre FS:  
>> Mount type: zfs
>> Flags:      0x1064
>>               (MGS first_time update no_primnode )
>> Persistent mount opts:
>> Parameters: failover.node=192.168.98.113 at tcp
>> WARNING: spl_hostid not set. ZFS has no zpool import protection
>> mkfs_cmd = zfs create -o canmount=off -o xattr=sa mgs/mgt
>> WARNING: spl_hostid not set. ZFS has no zpool import protection
>> Writing mgs/mgt properties
>>   lustre:failover.node=192.168.98.113 at tcp
>>   lustre:version=1
>>   lustre:flags=4196
>>   lustre:index=65535
>>   lustre:svname=MGS
>> [root at fb-lts-mds0 x86_64]# mount.lustre mgs/mgt /mnt/mgs
>> WARNING: spl_hostid not set. ZFS has no zpool import protection
>> 
>> mount.lustre FATAL: unhandled/unloaded fs type 0 'ext3'
>
>If I build the combo mgt/mdt, things go a lot better:
>
>> 
>> [root at fb-lts-mds0 x86_64]# mkfs.lustre --reformat --mgs --mdt
>>--force-nohostid --servicenode=192.168.98.113 at tcp --backfstype=zfs
>>--index=0 --fsname=test meta/meta
>> 
>>    Permanent disk data:
>> Target:     test:MDT0000
>> Index:      0
>> Lustre FS:  test
>> Mount type: zfs
>> Flags:      0x1065
>>               (MDT MGS first_time update no_primnode )
>> Persistent mount opts:
>> Parameters: failover.node=192.168.98.113 at tcp
>> WARNING: spl_hostid not set. ZFS has no zpool import protection
>> mkfs_cmd = zfs create -o canmount=off -o xattr=sa meta/meta
>> WARNING: spl_hostid not set. ZFS has no zpool import protection
>> Writing meta/meta properties
>>   lustre:failover.node=192.168.98.113 at tcp
>>   lustre:version=1
>>   lustre:flags=4197
>>   lustre:index=0
>>   lustre:fsname=test
>>   lustre:svname=test:MDT0000
>> [root at fb-lts-mds0 x86_64]# mount.lustre meta/meta  /mnt/meta
>> WARNING: spl_hostid not set. ZFS has no zpool import protection
>> [root at fb-lts-mds0 x86_64]# df
>> Filesystem          1K-blocks    Used Available Use% Mounted on
>> /dev/mapper/cl-root  52403200 3107560  49295640   6% /
>> devtmpfs             28709656       0  28709656   0% /dev
>> tmpfs                28720660       0  28720660   0% /dev/shm
>> tmpfs                28720660   17384  28703276   1% /run
>> tmpfs                28720660       0  28720660   0% /sys/fs/cgroup
>> /dev/sdb1             1038336  195484    842852  19% /boot
>> /dev/mapper/cl-home  34418260   32944  34385316   1% /home
>> tmpfs                 5744132       0   5744132   0% /run/user/0
>> meta                 60435328     128  60435200   1% /meta
>> meta/meta            59968128    4992  59961088   1% /mnt/meta
>> [root at fb-lts-mds0 ~]# ls /proc/fs/lustre/mdt/test-MDT0000/
>> async_commit_count     hash_stats               identity_upcall
>>num_exports         sync_count
>> commit_on_sharing      hsm                      instance
>>recovery_status     sync_lock_cancel
>> enable_remote_dir      hsm_control              ir_factor
>>recovery_time_hard  uuid
>> enable_remote_dir_gid  identity_acquire_expire  job_cleanup_interval
>>recovery_time_soft
>> evict_client           identity_expire          job_stats
>>rename_stats
>> evict_tgt_nids         identity_flush           md_stats
>>root_squash
>> exports                identity_info            nosquash_nids
>>site_stats
>
>Also, there's no /proc/fs/lustre/health_check
>
>> [root at fb-lts-mds0 ~]# ls /proc/fs/lustre/
>> fld   llite  lod  lwp  mdd  mdt  mgs      osc      osp  seq
>> ldlm  lmv    lov  mdc  mds  mgc  nodemap  osd-zfs  qmt  sptlrpc
>
>
>
>
>_______________________________________________
>lustre-discuss mailing list
>lustre-discuss at lists.lustre.org
>http://lists.lustre.org/listinfo.cgi/lustre-discuss-lustre.org



More information about the lustre-discuss mailing list