[Lustre-discuss] More failover issues

Robert LeBlanc robert at leblancnet.us
Mon Nov 12 13:56:07 PST 2007


Yes only one MGS per site, but you should be able to specify multiple MGS
nodes. We have done it before with 1.6.0. See
http://manual.lustre.org/manual/LustreManual16_HTML/DynamicHTML-05-1.html
section 2.2.2.1.

Robert


On 11/12/07 2:48 PM, "Wojciech Turek" <wjt27 at cam.ac.uk> wrote:

> Hi,
> 
> I think this is because there can be only one MGS per lustre installation
> (this is what manual says).
> 
> Wojciech Turek
> On 12 Nov 2007, at 21:18, Robert LeBlanc wrote:
> 
>>  
>> 
>> This is what I'm getting:
>>  
>>  head2-2:~# mkfs.lustre --mkfsoptions="-O dir_index" --reformat --mdt
>> --fsname=home --mgsnode=192.168.1.252 at o2ib --mgsnode=192.168.1.253 at o2ib
>> --failnode=192.168.1.252 at o2ib /dev/mapper/ldiskd-part1
>>  
>>     Permanent disk data:
>>  Target:     home-MDTffff
>>  Index:      unassigned
>>  Lustre FS:  home
>>  Mount type: ldiskfs
>>  Flags:      0x71
>>                (MDT needs_index first_time update )
>>  Persistent mount opts: errors=remount-ro,iopen_nopriv,user_xattr
>>  Parameters:  mgsnode=192.168.1.253 at o2ib failover.node=192.168.1.252 at o2ib
>> mdt.group_upcall=/usr/sbin/l_getgroups
>>  
>>  device size = 972MB
>>  formatting backing filesystem ldiskfs on /dev/mapper/ldiskd-part1
>>          target name  home-MDTffff
>>          4k blocks     0
>>          options       -O dir_index -i 4096 -I 512 -q -F
>>  mkfs_cmd = mkfs.ext2 -j -b 4096 -L home-MDTffff -O dir_index -i 4096 -I 512
>> -q -F /dev/mapper/ldiskd-part1
>>  Writing CONFIGS/mountdata
>>  
>>  
>>  For some reason, only the last --mgsnode option is being kept.
>>  
>>  Robert
>>  
>>  
>>  -----Original Message-----
>>  From: Nathan Rutman [mailto:Nathan.Rutman at Sun.COM]
>>  Sent: Mon 11/12/2007 1:51 PM
>>  To: Robert LeBlanc
>>  Cc: lustre
>>  Subject: Re: [Lustre-discuss] More failover issues
>>  
>>  Robert LeBlanc wrote:
>>>  > In 1.6.0, when creating a MDT, you could specify multiple --mgsnode
>>> options
>>>  > and it would failover between them. 1.6.3 only seems to take the last one
>>>  > and --mgsnode=192.168.1.252 at o2ib:192.168.1.253 at o2ib doesn't seem to
>>> failover
>>>  > to the other node. Any ideas how to get around this?
>>>  >  
>>  Multiple --mgsnode parameters should work:
>>  mkfs.lustre --mkfsoptions="-O dir_index" --reformat --mdt
>>  --mgsnode=192.168.1.253 at o2ib --mgsnode=1 at elan --device-size=10000 /tmp/foo
>>  
>>     Permanent disk data:
>>  Target:     lustre-MDTffff
>>  Index:      unassigned
>>  Lustre FS:  lustre
>>  Mount type: ldiskfs
>>  Flags:      0x71
>>                (MDT needs_index first_time update )
>>  Persistent mount opts: errors=remount-ro,iopen_nopriv,user_xattr
>>  Parameters: mgsnode=192.168.1.253 at o2ib mgsnode=1 at elan
>>  
>>>  > Robert
>>>>>>  > Robert LeBlanc
>>>  > College of Life Sciences Computer Support
>>>  > Brigham Young University
>>>  > leblanc at byu.edu
>>>  > (801)422-1882
>>>  >
>>>  >
>>>  > _______________________________________________
>>>  > Lustre-discuss mailing list
>>>  > Lustre-discuss at clusterfs.com
>>>  > https://mail.clusterfs.com/mailman/listinfo/lustre-discuss
>>>  >  
>>  
>>  
>>   
>> 
>> _______________________________________________
>> Lustre-discuss mailing list
>> Lustre-discuss at clusterfs.com
>> https://mail.clusterfs.com/mailman/listinfo/lustre-discuss
>>  
>> 
>>  
>> Mr Wojciech Turek
>> Assistant System Manager
>> University of Cambridge
>> High Performance Computing service 
>> email: wjt27 at cam.ac.uk
>> tel. +441223763517
>> 
>> 
>>  
>> 
>> 
> 
>  
> Robert LeBlanc
> College of Life Sciences Computer Support
> Brigham Young University
> leblanc at byu.edu
> (801)422-1882
> 

-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.lustre.org/pipermail/lustre-discuss-lustre.org/attachments/20071112/c549be15/attachment.htm>


More information about the lustre-discuss mailing list