[lustre-discuss] missing option mgsnode

Paul Edmon pedmon at cfa.harvard.edu
Wed Jul 20 12:18:06 PDT 2022


The individual LUN looks good but the controller is showing amber, which 
is confusing us.  However other LUN's going through that controller are 
mounting fine.

-Paul Edmon-

On 7/20/2022 3:08 PM, Colin Faber wrote:
> raid check?
>
> On Wed, Jul 20, 2022, 12:41 PM Paul Edmon <pedmon at cfa.harvard.edu> wrote:
>
>     [root at holylfs02oss06 ~]# mount -t ldiskfs /dev/mapper/mpathd
>     /mnt/holylfs2-OST001f
>     mount: wrong fs type, bad option, bad superblock on
>     /dev/mapper/mpathd,
>            missing codepage or helper program, or other error
>
>            In some cases useful info is found in syslog - try
>            dmesg | tail or so.
>
>     e2fsck did not look good:
>
>     [root at holylfs02oss06 ~]# less OST001f.out
>     ext2fs_check_desc: Corrupt group descriptor: bad block for block
>     bitmap
>     e2fsck: Group descriptors look bad... trying backup blocks...
>     MMP interval is 10 seconds and total wait time is 42 seconds.
>     Please wait...
>     Superblock needs_recovery flag is clear, but journal has data.
>     Recovery flag not set in backup superblock, so running journal anyway.
>     Clear journal? no
>
>     Block bitmap for group 8128 is not in group.  (block
>     3518518062363072290)
>     Relocate? no
>
>     Inode bitmap for group 8128 is not in group.  (block
>     12235298632209565410)
>     Relocate? no
>
>     Inode table for group 8128 is not in group.  (block
>     17751685088477790304)
>     WARNING: SEVERE DATA LOSS POSSIBLE.
>     Relocate? no
>
>     Block bitmap for group 8129 is not in group.  (block
>     2193744380193356980)
>     Relocate? no
>
>     Inode bitmap for group 8129 is not in group.  (block
>     4102707059848926418)
>     Relocate? no
>
>     It continues at length like that.
>
>     -Paul Edmon-
>
>     On 7/20/2022 2:31 PM, Colin Faber wrote:
>>     Can you mount the target directly with -t ldiskfs ?
>>
>>     Also what does e2fsck report?
>>
>>     On Wed, Jul 20, 2022, 11:48 AM Paul Edmon via lustre-discuss
>>     <lustre-discuss at lists.lustre.org> wrote:
>>
>>         We have a filesystem that we have running Lustre 2.10.4 in HA
>>         mode using
>>         IML.  One of our OST's had some disk failures and after
>>         reconstruction
>>         of the RAID set it won't remount but gives:
>>
>>         [root at holylfs02oss06 ~]# mount -t lustre /dev/mapper/mpathd
>>         /mnt/holylfs2-OST001f
>>         Failed to initialize ZFS library: 256
>>         mount.lustre: missing option mgsnode=<nid>
>>
>>         The weird thing is that we didn't build this with ZFS, the
>>         devices are
>>         all ldiskfs.  We suspect some of the data is corrupt on the
>>         disk but we
>>         were wondering if anyone had seen this error before and if
>>         there was a
>>         solution.
>>
>>         -Paul Edmon-
>>
>>         _______________________________________________
>>         lustre-discuss mailing list
>>         lustre-discuss at lists.lustre.org
>>         http://lists.lustre.org/listinfo.cgi/lustre-discuss-lustre.org
>>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.lustre.org/pipermail/lustre-discuss-lustre.org/attachments/20220720/c1816a1d/attachment.htm>


More information about the lustre-discuss mailing list