[Lustre-discuss] MDT raid parameters, multiple MGSes

Ben Evans Ben.Evans at terascala.com
Fri Jan 21 05:55:21 PST 2011

In our lab, we've never had a problem with simply having 1 MGS per filesystem.  Mountpoints will be unique for all of them, but functionally it works just fine.

-Ben Evans

-----Original Message-----
From: lustre-discuss-bounces at lists.lustre.org on behalf of Thomas Roth
Sent: Fri 1/21/2011 6:43 AM
To: lustre-discuss at lists.lustre.org
Subject: [Lustre-discuss] MDT raid parameters, multiple MGSes
Hi all,

we have gotten new MDS hardware, and I've got two questions:

What are the recommendations for the RAID configuration and formatting 
I was following the recent discussion about these aspects on an OST: 
chunk size, strip size, stride-size, stripe-width etc. in the light of 
the 1MB chunks of Lustre ... So what about the MDT? I will have a RAID 
10 that consists of 11 RAID-1 pairs striped over. giving me roughly 3TB 
of space. What would be the correct value for <insert your favorite 
term>, the amount of data written to one disk before proceeding to the 
next disk?

Secondly, it is not yet decided whether we wouldn't use this hardware to 
set up a second Lustre cluster. The manual recommends to have only one 
MGS per site, but doesn't elaborate: what would be the drawback of 
having two MGSes, two different network addresses the clients have to 
connect to to mount the Lustres?
I know that it didn't work in Lustre 1.6.3 ;-) and there are no apparent 
issues when connecting a Lustre client to a test cluster now (version 
1.8.4), but what about production?

Lustre-discuss mailing list
Lustre-discuss at lists.lustre.org

-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.lustre.org/pipermail/lustre-discuss-lustre.org/attachments/20110121/37ba3c4f/attachment.htm>

More information about the lustre-discuss mailing list