[Lustre-discuss] CentOS 5.4 (Rocks 5.3) and Lustre 1.8.2 client against HP SFS 3.2

Finn Andersen xiphias256 at gmail.com
Thu Jun 3 00:27:10 PDT 2010


Hi!

I'm trying to get Rocks 5.3 (Based on CentOS 5.4) to work with Lustre
over Infiniband. I'm almost done now... :)
>From what I have read on different mailinglist etc. the best option is
to use stock rpm's as much as possible.

>From the Lustre homepage I've used these 2 packages:
lustre-client-1.8.2-2.6.18_164.11.1.el5_lustre.1.8.2.x86_64.rpm
lustre-client-modules-1.8.2-2.6.18_164.11.1.el5_lustre.1.8.2.x86_64.rpm

I have used this kernel from CentOS: kernel-2.6.18_164.11.1.el.x86_64.rpm
My ofed install is the one that RHEL/CentOS ships with it's 5.5
release. (1.4.1-5)
All this seems to be within the support matrix on lustre.org

/etc/modprobe.conf has this entry:
options lnet networks=o2ib0

Why do I have to enter "modprobe lustre" manually before I can mount
the lustre filesystem? How can this be fixed, so that it will come up
automatically on reboot?

My /etc/fstab entry is:
10.129.2.1 at o2ib0:10.129.2.2 at o2ib0:/scratch  /scratch  lustre
_netdev,rw,flock 0 0

When I try to reboot one of the compute nodes, I get this error
message on the console:
---
failed to unload rdma_cm
failed to unload ib_cm
failed to unload iw_cm
LustreError: 131-3: Received notification of device removal
Please shutdown LNET to allow this to procceed
---
The compute node has to be reset in order to get it back up again.
Why? And how can it be fixed?

-
Finn Andersen
Norway



More information about the lustre-discuss mailing list