[Lustre-discuss] [wc-discuss] can't mount our lustre filesystem after tunefs.lustre --writeconf

Kit Westneat kit.westneat at nyu.edu
Sat Mar 17 23:24:13 PDT 2012


You should be able to reset the UUID by doing another writeconf with the
--fsname flag. After the writeconf, you'll have to writeconf all the OSTs
too.

It worked on my very simple test at least:
[root at mds1 tmp]# tunefs.lustre --writeconf --fsname=test1 /dev/loop0
checking for existing Lustre data: found CONFIGS/mountdata
Reading CONFIGS/mountdata

   Read previous values:
Target:     t1-MDT0000
Index:      0
Lustre FS:  t1
Mount type: ldiskfs
Flags:      0x5
              (MDT MGS )
Persistent mount opts: iopen_nopriv,user_xattr,errors=remount-ro
Parameters: mdt.group_upcall=/usr/sbin/l_getgroups


   Permanent disk data:
Target:     test1-MDT0000
Index:      0
Lustre FS:  test1
Mount type: ldiskfs
Flags:      0x105
              (MDT MGS writeconf )
Persistent mount opts: iopen_nopriv,user_xattr,errors=remount-ro
Parameters: mdt.group_upcall=/usr/sbin/l_getgroups

Writing CONFIGS/mountdata


HTH,
Kit
--
Kit Westneat
System Administrator, eSys
kit.westneat at nyu.edu
212-992-7647


On Sun, Mar 18, 2012 at 1:20 AM, Stu Midgley <sdm900 at gmail.com> wrote:

> ok, from what I can tell, the root of the problem is
>
>
> [root at mds001 CONFIGS]# hexdump -C p1-MDT0000  | grep -C 2 mds
> 00002450  0b 00 00 00 04 00 00 00  12 00 00 00 00 00 00 00
>  |................|
> 00002460  70 31 2d 4d 44 54 30 30  30 30 00 00 00 00 00 00
>  |p1-MDT0000......|
> 00002470  6d 64 73 00 00 00 00 00  70 72 6f 64 5f 6d 64 73
>  |mds.....prod_mds|
> 00002480  5f 30 30 31 5f 55 55 49  44 00 00 00 00 00 00 00
>  |_001_UUID.......|
> 00002490  78 00 00 00 07 00 00 00  88 00 00 00 08 00 00 00
>  |x...............|
> --
> 000024c0  00 00 00 00 04 00 00 00  0b 00 00 00 12 00 00 00
>  |................|
> 000024d0  02 00 00 00 0b 00 00 00  70 31 2d 4d 44 54 30 30
>  |........p1-MDT00|
> 000024e0  30 30 00 00 00 00 00 00  70 72 6f 64 5f 6d 64 73
>  |00......prod_mds|
> 000024f0  5f 30 30 31 5f 55 55 49  44 00 00 00 00 00 00 00
>  |_001_UUID.......|
> 00002500  30 00 00 00 00 00 00 00  70 31 2d 4d 44 54 30 30
>  |0.......p1-MDT00|
>
> [root at mds001 CONFIGS]#
> [root at mds001 CONFIGS]# hexdump -C /mnt/md2/CONFIGS/p1-MDT0000 | grep -C 2
> mds
> 00002450  0b 00 00 00 04 00 00 00  10 00 00 00 00 00 00 00
>  |................|
> 00002460  70 31 2d 4d 44 54 30 30  30 30 00 00 00 00 00 00
>  |p1-MDT0000......|
> 00002470  6d 64 73 00 00 00 00 00  70 31 2d 4d 44 54 30 30
>  |mds.....p1-MDT00|
> 00002480  30 30 5f 55 55 49 44 00  70 00 00 00 07 00 00 00
>  |00_UUID.p.......|
> 00002490  80 00 00 00 08 00 00 00  00 00 62 10 ff ff ff ff
>  |..........b.....|
>
>
> now if only I can get the UUID to be removed or reset...
>
>
> On Sun, Mar 18, 2012 at 1:05 PM, Dr Stuart Midgley <sdm900 at gmail.com>
> wrote:
> > hmmm… that didn't work
> >
> > # tunefs.lustre --force --fsname=p1 /dev/md2
> > checking for existing Lustre data: found CONFIGS/mountdata
> > Reading CONFIGS/mountdata
> >
> >   Read previous values:
> > Target:     p1-MDT0000
> > Index:      0
> > UUID:       prod_mds_001_UUID
> > Lustre FS:  p1
> > Mount type: ldiskfs
> > Flags:      0x405
> >              (MDT MGS )
> > Persistent mount opts: errors=remount-ro,iopen_nopriv,user_xattr
> > Parameters:
> >
> > tunefs.lustre: unrecognized option `--force'
> > tunefs.lustre: exiting with 22 (Invalid argument)
> >
> >
> >
> >
> > --
> > Dr Stuart Midgley
> > sdm900 at gmail.com
> >
> >
> >
> > On 18/03/2012, at 12:17 AM, Nathan Rutman wrote:
> >
> >> Take them all down again, use tunefs.lustre --force --fsname.
> >>
> >>
> >> On Mar 17, 2012, at 2:10 AM, "Stu Midgley" <sdm900 at gmail.com> wrote:
> >>
> >>> Afternoon
> >>>
> >>> We have a rather severe problem with our lustre file system.  We had a
> >>> full config log and the advice was to rewrite it with a new one.  So,
> >>> we unmounted our lustre file system off all clients, unmount all the
> >>> ost's and then unmounted the mds.  I then did
> >>>
> >>> mds:
> >>>  tunefs.lustre --writeconf --erase-params /dev/md2
> >>>
> >>> oss:
> >>>  tunefs.lustre --writeconf --erase-params --mgsnode=mds001 /dev/md2
> >>>
> >>>
> >>>
> >>> After the tunefs.lustre on the mds I saw
> >>>
> >>> Mar 17 14:33:02 mds001 kernel: Lustre: MGS MGS started
> >>> Mar 17 14:33:02 mds001 kernel: Lustre: MGC172.16.0.251 at tcp:
> Reactivating import
> >>> Mar 17 14:33:02 mds001 kernel: Lustre: MGS: Logs for fs p1 were
> >>> removed by user request.  All servers must be restarted in order to
> >>> regenerate the logs.
> >>> Mar 17 14:33:02 mds001 kernel: Lustre: Enabling user_xattr
> >>> Mar 17 14:33:02 mds001 kernel: Lustre: p1-MDT0000: new disk,
> initializing
> >>> Mar 17 14:33:02 mds001 kernel: Lustre: p1-MDT0000: Now serving
> >>> p1-MDT0000 on /dev/md2 with recovery enabled
> >>>
> >>> which scared me a little...
> >>>
> >>>
> >>>
> >>> the mds and the oss's mount happily BUT I can't mount the file system
> >>> on my clients... on the mds I see
> >>>
> >>>
> >>> Mar 17 16:42:11 mds001 kernel: LustreError: 137-5: UUID
> >>> 'prod_mds_001_UUID' is not available  for connect (no target)
> >>>
> >>>
> >>> On the client I see
> >>>
> >>>
> >>> Mar 17 16:00:06 host kernel: LustreError: 11-0: an error occurred
> >>> while communicating with 172.16.0.251 at tcp. The mds_connect operation
> >>> failed with -19
> >>>
> >>>
> >>> now, it appears the writeconf renamed the UUID of the mds from
> >>> prod_mds_001_UUID to p1-MDT0000_UUID but I can't work out how to get
> >>> it back...
> >>>
> >>>
> >>> for example I tried
> >>>
> >>>
> >>> # tunefs.lustre --mgs --mdt --fsname=p1 /dev/md2
> >>> checking for existing Lustre data: found CONFIGS/mountdata
> >>> Reading CONFIGS/mountdata
> >>>
> >>> Read previous values:
> >>> Target:     p1-MDT0000
> >>> Index:      0
> >>> UUID:       prod_mds_001_UUID
> >>> Lustre FS:  p1
> >>> Mount type: ldiskfs
> >>> Flags:      0x405
> >>>            (MDT MGS )
> >>> Persistent mount opts: errors=remount-ro,iopen_nopriv,user_xattr
> >>> Parameters:
> >>>
> >>> tunefs.lustre: cannot change the name of a registered target
> >>> tunefs.lustre: exiting with 1 (Operation not permitted)
> >>>
> >>>
> >>>
> >>> I'm now stuck not being able to mount a 1PB file system... which isn't
> good :(
> >>>
> >>> --
> >>> Dr Stuart Midgley
> >>> sdm900 at gmail.com
> >> ______________________________________________________________________
> >> This email may contain privileged or confidential information, which
> should only be used for the purpose for which it was sent by Xyratex. No
> further rights or licenses are granted to use such information. If you are
> not the intended recipient of this message, please notify the sender by
> return and delete it. You may not use, copy, disclose or rely on the
> information contained in it.
> >>
> >> Internet email is susceptible to data corruption, interception and
> unauthorised amendment for which Xyratex does not accept liability. While
> we have taken reasonable precautions to ensure that this email is free of
> viruses, Xyratex does not accept liability for the presence of any computer
> viruses in this email, nor for any losses caused as a result of viruses.
> >>
> >> Xyratex Technology Limited (03134912), Registered in England & Wales,
> Registered Office, Langstone Road, Havant, Hampshire, PO9 1SA.
> >>
> >> The Xyratex group of companies also includes, Xyratex Ltd, registered
> in Bermuda, Xyratex International Inc, registered in California, Xyratex
> (Malaysia) Sdn Bhd registered in Malaysia, Xyratex Technology (Wuxi) Co Ltd
> registered in The People's Republic of China and Xyratex Japan Limited
> registered in Japan.
> >> ______________________________________________________________________
> >>
> >>
> >
>
>
>
> --
> Dr Stuart Midgley
> sdm900 at gmail.com
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.lustre.org/pipermail/lustre-discuss-lustre.org/attachments/20120318/f579e3df/attachment.htm>


More information about the lustre-discuss mailing list