[lustre-discuss] Getting started with Lustre on RHEL 8.8

Andreas Dilger adilger at whamcloud.com
Tue Sep 12 21:49:35 PDT 2023


On Sep 12, 2023, at 22:31, Cyberxstudio cxs <cyberxstudio.cloud at gmail.com<mailto:cyberxstudio.cloud at gmail.com>> wrote:

Hi I get this error while installing lustre and other packages

[root at localhost ~]# yum --nogpgcheck --enablerepo=lustre-server install \
> kmod-lustre-osd-ldiskfs \
> lustre-dkms \
> lustre-osd-ldiskfs-mount \
> lustre-osd-zfs-mount \
> lustre \
> lustre-resource-agents \
> zfs
Updating Subscription Management repositories.
Last metadata expiration check: 0:00:58 ago on Wed 13 Sep 2023 09:27:59 AM PKT.
Error:
 Problem: conflicting requests
  - nothing provides resource-agents needed by lustre-resource-agents-2.15.3-1.el8.x86_64
(try to add '--skip-broken' to skip uninstallable packages or '--nobest' to use not only best candidate packages)

You don't need this package to start. It is used for HA failover of storage between servers with Corosync/Pacemaker.

You also do not need the "lustre-dkms" package - that is for building Lustre clients from scratch.

You also only need one of ldiskfs or ZFS.  If you don't have RAID storage, then ZFS is probably more useful,
while ldiskfs is more of a 'traditional" filesystem (based on ext4).

Cheers, Andreas

[root at localhost ~]# dnf install resource-agents
Updating Subscription Management repositories.
Last metadata expiration check: 0:02:01 ago on Wed 13 Sep 2023 09:27:59 AM PKT.
No match for argument: resource-agents
Error: Unable to find a match: resource-agents
[root at localhost ~]#

On Wed, Sep 13, 2023 at 9:10 AM Cyberxstudio cxs <cyberxstudio.cloud at gmail.com<mailto:cyberxstudio.cloud at gmail.com>> wrote:
Thank you for the information.

On Tue, Sep 12, 2023 at 8:40 PM Andreas Dilger <adilger at whamcloud.com<mailto:adilger at whamcloud.com>> wrote:
Hello,
The preferred path to set up Lustre depends on what you are planning to do with it?  If for regular usage it is easiest to start with RPMs built for the distro from https://downloads.whamcloud.com/public/lustre/latest-release/<https://downloads.whamcloud.com/public/lustre/latest-release/el8.8/server/RPMS/x86_64/> (you can also use the server RPMs for a client if you want).

The various "client" packages for RHEL, SLES, Ubuntu can install directly on the vendor kernels, but the provided server RPMs also need the matching kernel.  You only need one of the ldiskfs (ext4) or ZFS packages, not both.

It isn't *necessary* to build/patch your kernel for the server, though the pre-built server download packages have patched the kernel to add integrated T10-PI support (which many users do not need).  You can get unpatched el8 server RPMs directly from the builders:
https://build.whamcloud.com/job/lustre-b2_15-patchless/48/arch=x86_64,build_type=server,distro=el8.7,ib_stack=inkernel/artifact/artifacts/

If you plan to run on non-standard kernels, then you can build RPMs for your particular kernel. The easiest way is to just rebuild the SPRM package:
 https://wiki.whamcloud.com/pages/viewpage.action?pageId=8556211

If you want to do Lustre development you should learn how to build from a Git checkout:
https://wiki.whamcloud.com/display/PUB/Building+Lustre+from+Source

Cheers, Andreas

On Sep 12, 2023, at 03:25, Cyberxstudio cxs via lustre-discuss <lustre-discuss at lists.lustre.org<mailto:lustre-discuss at lists.lustre.org>> wrote:


Hi, I am setting up a lab environment for lustre. I have 3 VMs of RHEL 8.8, I have studied the documentation but it does not provide detail for el 8 rather el 7. Please guide me how to start
Thank You
_______________________________________________
lustre-discuss mailing list
lustre-discuss at lists.lustre.org<mailto:lustre-discuss at lists.lustre.org>
http://lists.lustre.org/listinfo.cgi/lustre-discuss-lustre.org

Cheers, Andreas
--
Andreas Dilger
Lustre Principal Architect
Whamcloud







-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.lustre.org/pipermail/lustre-discuss-lustre.org/attachments/20230913/e9207bc3/attachment-0001.htm>


More information about the lustre-discuss mailing list