[lustre-discuss] Lustre and server upgrade

Colin Faber cfaber at gmail.com
Thu Nov 18 12:34:00 PST 2021


Hm.. If you install the test suite does llmount.sh succeed? This should
setup a single node cluster on whatever node you're running lustre on, I
believe it will load modules as needed (IIRC), if this test succeeds, then
you know that lustre is installed correctly (or correctly enough), if not,
I'd focus on the installation as the target issue may be a redheirring

-cf


On Thu, Nov 18, 2021 at 1:01 PM STEPHENS, DEAN - US <dean.stephens at caci.com>
wrote:

> Thanks for the fast reply.
>
> When I do the tunefs.lustre /dev/sdX command I get:
>
> Target: <name>-OST0009
>
> Index: 9
>
>
>
> Target: <name>-OST0008
>
> Index: 8
>
> I spot checked some others and they seem to be good with the exception of
> one. It shows:
>
>
>
> Target: <name>-OST000a
>
> Index: 10
>
>
>
> But since there are 11 LUNs attached that make sense to me.
>
>
>
> As far as the upgrade it was a fresh install using the legacy targets as
> the OSS and MDS nodes are virtual machine with the LUN disks attached to
> them so that Red Hat sees them as /dev/sdX devices.
>
>
>
> When I loaded Lustre on the server I did a yum install lustre and since we
> were pointed at the lustre-2.12 repo in our environment it picked up the
> following RPMs to install:
>
> Luster-resource-agents-2.12.6-1.el7.x86_64
>
> Kmod-lustre-2.12.6-1.el7.x86_64
>
> Kmod-zfs-3.10.0-1160.2.1.el7_lustre.x86_64-09.7.13-1.el7.x86_64
>
> Kmod-lustre-osd-zfs-2.12.6-1.el7.x86_64
>
> Lustre-2.12.6-1.el7.x86_64
>
> Kmod-spl-3.10.0-1160.2.1.el7_lustre.x86_64-09.7.13-1.el7.x86_64
>
> Lustre-osd-zfs-mount-2.12.6-1.el7.x86_64
>
> Lustre-osd-ldiskfs-mount-2.12.6-1.el7.x86_64
>
>
>
> Dean
>
>
>
> *From:* Colin Faber <cfaber at gmail.com>
> *Sent:* Thursday, November 18, 2021 12:35 PM
> *To:* STEPHENS, DEAN - US <dean.stephens at caci.com>
> *Cc:* lustre-discuss at lists.lustre.org
> *Subject:* Re: [lustre-discuss] Lustre and server upgrade
>
>
>
> EXTERNAL EMAIL - This email originated from outside of CACI. Do not click
> any links or attachments unless you recognize and trust the sender.
>
>
>
>
>
> Hi,
>
>
>
> I believe in 2.10 sometime (someone correct me if I'm wrong) that the
> index parameter was required and needs to be specified. On an existing
> system this should already be set, but can you check the parameters line
> with tunefs.lustre for correct index=N values across your storage nodes?
>
>
>
> Also, with your "upgrade", was this a fresh install utilizing legacy
> targets?
>
>
>
> The last thing I can think of IIRC, there was on-disk format changes
> between 2.5 and 2.12, these should be transparent to you, but it may be
> some other issue is preventing successful upgrade, though the missing
> module error really speaks to possible issues around how lustre was
> installed and loaded on the system.
>
>
>
> Cheers!
>
>
>
> -cf
>
>
>
>
>
> On Thu, Nov 18, 2021 at 12:24 PM STEPHENS, DEAN - US via lustre-discuss <
> lustre-discuss at lists.lustre.org> wrote:
>
> I am by no means a Lustre expert and am seeking some help with our system.
> I am not able to get log file to post as the servers are in the closed area
> with no access to the Internet.
>
>
>
> Here is a bit of history of our system:
>
> The OSS and MDS nodes were RHEL6 and running a Luster server the kernel
> 2.6.32-431.23.3.el6_lustre.x86_64 and the Lustre version of 2.5.3. the
> client version was 2.10. That was in a working state.
>
> We upgraded the OSS ad MDS nodes to RHEL7 and installed Lustre server 2.12
> software and kernel.
>
> The attached 11 LUNs are showing up as /dev/sdb - /dev/sdl
>
> Right now, on the OSS nodes, if I use the command tunefs.luster /dev/sdb I
> get some data back saying that Lustre data has been found but at the bottom
> of the out put it shows “tunefs.lustre: Unable to mount /dev/sdb: No such
> device” and “Is the ldiskfs module available”
>
> When I do a “modprobe -v lustre” I do not see ldiskfs.ko as being loaded
> even though there is a ldiskfs.ko file in
> /lib/modules/3.10.0-1160.2.1.el7_lustre.x86_64/extra/lustre/fs directory. I
> am not sure how to get it to load in the modprobe command.
>
> I used “insmod
> /lib/modules/3.10.0-1160.2.1.el7_lustre.x86_64/extra/lustre/fs/ ldiskfs.ko”
> and re-ran the “tunefs.luster /dev/sdb” command with the same result.
>
> If I use the same command on the MDS nodes I get “no Lustre data found and
> /dev/sdb has not been formatted with mkfs.lustre”. I am not sure that is
> what is needed here as the MDS nodes do not really have the lustre data as
> it is the meta data server.
>
> I tried to use the command “tunefs.lustre --mgs --erase_params
> --mgsnode=<IP address>@tcp --writeconf --dryrun /dev/sdb” and get the error
> “/dev/sdb has not been formatted with mkfs.lustre”.
>
>
>
> I need some help and guidance and I can provide what may be needed though
> it will need to be typed out as I am not able to get actual log files from
> the system.
>
>
>
> Dean Stephens
>
> CACI
>
> Linux System Admin
>
>
>
>
> ------------------------------
>
>
> This electronic message contains information from CACI International Inc
> or subsidiary companies, which may be company sensitive, proprietary,
> privileged or otherwise protected from disclosure. The information is
> intended to be used solely by the recipient(s) named above. If you are not
> an intended recipient, be aware that any review, disclosure, copying,
> distribution or use of this transmission or its contents is prohibited. If
> you have received this transmission in error, please notify the sender
> immediately.
>
> _______________________________________________
> lustre-discuss mailing list
> lustre-discuss at lists.lustre.org
> http://lists.lustre.org/listinfo.cgi/lustre-discuss-lustre.org
>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.lustre.org/pipermail/lustre-discuss-lustre.org/attachments/20211118/b60e5149/attachment.html>


More information about the lustre-discuss mailing list