[lustre-discuss] Install issues on 2.10.0
John Casu
john at chiraldynamics.com
Tue Jul 25 10:52:06 PDT 2017
Ok, so I assume this is actually a ZFS/SPL bug & not a lustre bug.
Also, thanks Ben, for the ptr.
many thanks,
-john
On 7/25/17 10:19 AM, Mannthey, Keith wrote:
> Host_id is for zpool double import protection. If a host id is set on a zpool (zfs does this automatically) then a HA server can't just import to pool (users have to use --force). This makes the system a lot safer from double zpool imports. Call 'genhostid' on your Lustre servers and the warning will go away.
>
> Thanks,
> Keith
>
>
>
> -----Original Message-----
> From: lustre-discuss [mailto:lustre-discuss-bounces at lists.lustre.org] On Behalf Of Ben Evans
> Sent: Tuesday, July 25, 2017 10:13 AM
> To: John Casu <john at chiraldynamics.com>; lustre-discuss at lists.lustre.org
> Subject: Re: [lustre-discuss] Install issues on 2.10.0
>
> health_check moved to /sys/fs/lustre/ along with a bunch of other things.
>
> -Ben
>
> On 7/25/17, 12:21 PM, "lustre-discuss on behalf of John Casu"
> <lustre-discuss-bounces at lists.lustre.org on behalf of john at chiraldynamics.com> wrote:
>
>> Just installed latest 2.10.0 Lustre over ZFS on a vanilla Centos
>> 7.3.1611 system, using dkms.
>> ZFS is 0.6.5.11 from zfsonlinux.org, installed w. yum
>>
>> Not a single problem during installation, but I am having issues
>> building a lustre filesystem:
>> 1. Building a separate mgt doesn't seem to work properly, although the
>> mgt/mdt combo
>> seems to work just fine.
>> 2. I get spl_hostid not set warnings, which I've never seen before 3.
>> /proc/fs/lustre/health_check seems to be missing.
>>
>> thanks,
>> -john c
>>
>>
>>
>> ---------
>> Building an mgt by itself doesn't seem to work properly:
>>
>>> [root at fb-lts-mds0 x86_64]# mkfs.lustre --reformat --mgs
>>> --force-nohostid --servicenode=192.168.98.113 at tcp \
>>> --backfstype=zfs mgs/mgt
>>>
>>> Permanent disk data:
>>> Target: MGS
>>> Index: unassigned
>>> Lustre FS:
>>> Mount type: zfs
>>> Flags: 0x1064
>>> (MGS first_time update no_primnode ) Persistent mount
>>> opts:
>>> Parameters: failover.node=192.168.98.113 at tcp
>>> WARNING: spl_hostid not set. ZFS has no zpool import protection
>>> mkfs_cmd = zfs create -o canmount=off -o xattr=sa mgs/mgt
>>> WARNING: spl_hostid not set. ZFS has no zpool import protection
>>> Writing mgs/mgt properties
>>> lustre:failover.node=192.168.98.113 at tcp
>>> lustre:version=1
>>> lustre:flags=4196
>>> lustre:index=65535
>>> lustre:svname=MGS
>>> [root at fb-lts-mds0 x86_64]# mount.lustre mgs/mgt /mnt/mgs
>>> WARNING: spl_hostid not set. ZFS has no zpool import protection
>>>
>>> mount.lustre FATAL: unhandled/unloaded fs type 0 'ext3'
>>
>> If I build the combo mgt/mdt, things go a lot better:
>>
>>>
>>> [root at fb-lts-mds0 x86_64]# mkfs.lustre --reformat --mgs --mdt
>>> --force-nohostid --servicenode=192.168.98.113 at tcp --backfstype=zfs
>>> --index=0 --fsname=test meta/meta
>>>
>>> Permanent disk data:
>>> Target: test:MDT0000
>>> Index: 0
>>> Lustre FS: test
>>> Mount type: zfs
>>> Flags: 0x1065
>>> (MDT MGS first_time update no_primnode ) Persistent
>>> mount opts:
>>> Parameters: failover.node=192.168.98.113 at tcp
>>> WARNING: spl_hostid not set. ZFS has no zpool import protection
>>> mkfs_cmd = zfs create -o canmount=off -o xattr=sa meta/meta
>>> WARNING: spl_hostid not set. ZFS has no zpool import protection
>>> Writing meta/meta properties
>>> lustre:failover.node=192.168.98.113 at tcp
>>> lustre:version=1
>>> lustre:flags=4197
>>> lustre:index=0
>>> lustre:fsname=test
>>> lustre:svname=test:MDT0000
>>> [root at fb-lts-mds0 x86_64]# mount.lustre meta/meta /mnt/meta
>>> WARNING: spl_hostid not set. ZFS has no zpool import protection
>>> [root at fb-lts-mds0 x86_64]# df
>>> Filesystem 1K-blocks Used Available Use% Mounted on
>>> /dev/mapper/cl-root 52403200 3107560 49295640 6% /
>>> devtmpfs 28709656 0 28709656 0% /dev
>>> tmpfs 28720660 0 28720660 0% /dev/shm
>>> tmpfs 28720660 17384 28703276 1% /run
>>> tmpfs 28720660 0 28720660 0% /sys/fs/cgroup
>>> /dev/sdb1 1038336 195484 842852 19% /boot
>>> /dev/mapper/cl-home 34418260 32944 34385316 1% /home
>>> tmpfs 5744132 0 5744132 0% /run/user/0
>>> meta 60435328 128 60435200 1% /meta
>>> meta/meta 59968128 4992 59961088 1% /mnt/meta
>>> [root at fb-lts-mds0 ~]# ls /proc/fs/lustre/mdt/test-MDT0000/
>>> async_commit_count hash_stats identity_upcall
>>> num_exports sync_count
>>> commit_on_sharing hsm instance
>>> recovery_status sync_lock_cancel
>>> enable_remote_dir hsm_control ir_factor
>>> recovery_time_hard uuid
>>> enable_remote_dir_gid identity_acquire_expire job_cleanup_interval
>>> recovery_time_soft
>>> evict_client identity_expire job_stats
>>> rename_stats
>>> evict_tgt_nids identity_flush md_stats
>>> root_squash
>>> exports identity_info nosquash_nids
>>> site_stats
>>
>> Also, there's no /proc/fs/lustre/health_check
>>
>>> [root at fb-lts-mds0 ~]# ls /proc/fs/lustre/
>>> fld llite lod lwp mdd mdt mgs osc osp seq
>>> ldlm lmv lov mdc mds mgc nodemap osd-zfs qmt sptlrpc
>>
>>
>>
>>
>> _______________________________________________
>> lustre-discuss mailing list
>> lustre-discuss at lists.lustre.org
>> http://lists.lustre.org/listinfo.cgi/lustre-discuss-lustre.org
>
> _______________________________________________
> lustre-discuss mailing list
> lustre-discuss at lists.lustre.org
> http://lists.lustre.org/listinfo.cgi/lustre-discuss-lustre.org
>
>
More information about the lustre-discuss
mailing list