[Lustre-discuss] Large Corosync/Pacemaker clusters
Charles Taylor
taylor at hpc.ufl.edu
Wed Oct 24 13:32:40 PDT 2012
FWIW, we are running HA Lustre using corosync/pacemaker. We broke our OSSs and MDSs out into individual HA *pairs*. Thought about other configurations but it was our first step into corosync/pacemaker so we decided to keep it as simple as possible. Seems to work well. I'm not sure I would attempt what you are doing though it may be perfectly fine. When HA is a requirement, it probably makes sense to avoid pushing the limits of what works.
Doesn't really help you much other than to provide a data point with regard to what other sites are doing.
Good luck and report back.
Charlie Taylor
UF HPC Center
On Oct 19, 2012, at 12:52 PM, Hall, Shawn wrote:
> Hi,
>
> We’re setting up fairly large Lustre 2.1.2 filesystems, each with 18 nodes and 159 resources all in one Corosync/Pacemaker cluster as suggested by our vendor. We’re getting mixed messages on how large of a Corosync/Pacemaker cluster will work well between our vendor an others.
>
> 1. Are there Lustre Corosync/Pacemaker clusters out there of this size or larger?
> 2. If so, what tuning needed to be done to get it to work well?
> 3. Should we be looking more seriously into splitting this Corosync/Pacemaker cluster into pairs or sets of 4 nodes?
>
> Right now, our current configuration takes a long time to start/stop all resources (~30-45 mins), and failing back OSTs puts a heavy load on the cib process on every node in the cluster. Under heavy IO load, the many of the nodes will show as “unclean/offline” and many OST resources will show as inactive in crm status, despite the fact that every single MDT and OST is still mounted in the appropriate place. We are running 2 corosync rings, each on a private 1 GbE network. We have a bonded 10 GbE network for the LNET.
>
> Thanks,
> Shawn
> _______________________________________________
> Lustre-discuss mailing list
> Lustre-discuss at lists.lustre.org
> http://lists.lustre.org/mailman/listinfo/lustre-discuss
More information about the lustre-discuss
mailing list