[Lustre-discuss] Question about failnode

Kazuki Ohara ohara at rd.scei.sony.co.jp
Fri Oct 26 00:03:56 PDT 2007


Hi Brain,
Thank you for your answer.

I understood the need of STONISH.

By the way, I doubt the need of the --failnode directive.
I walked into the source code of lustre, but I could not find
the use of the --failnode information except for building log messages.

Are there any more important reasons of the use of the --failnode directive?

Best regards,
Kazuki Ohara

Brian J. Murrell wrote:
> On Thu, 2007-10-25 at 20:58 +0900, Kazuki Ohara wrote:
>> Hi,
>> I have a question about the fainode directive of mkfs.lustre.
>> I hope someone help me.
>>
>> When a shared volume is formatted and mounted as below,
>>   [root at ossnode1 ~]# mkfs.lustre --ost --failnode=ossnode2 \
>>   > --mgsnode=mgsnode at tcp /dev/sda1
>>   [root at ossnode1 ~]# mount -t lustre /dev/sda1 /mnt/ost1
>> ossnode1 knows that sda1 can be accessed by ossnode2.
>>
>> Then, a failover occurs and ossnode2 mounts sda1,
> 
> ossnode2 should in fact, before it does the mount, make as entirely sure
> as it can that ossnode1 does not have it mounted.  The surefire way to
> do that is to kill the power to ossnode1 (assuming there is a power
> controller between the mains and ossnode1 that ossnode2 can operate).
> 
> In the failover game this is called STONITH and is an acronym for "Shoot
> The Other Node In The Head".  All of this is usually coordinated with
> something like Heartbeat.
> 
> The reason for this STONITH action is that in an HA scenario, ossnode2
> only knows that it cannot reach ossnode1.  It does not know why.  It
> could be because it's power failed, it panic'd or any number of reasons.
> Not all of those reasons imply that ossnode1 cannot (and does not) still
> have the disk mounted though.  Only by killing ossnode1 itself, can
> ossnode2 be absolutely sure that ossnode1 does not have the disk
> mounted.
> 
> More than one node mounting an ext{2,3,4} or ldiskfs (which is ext4,
> basically) filesystem is disastrous for that filesystem, so all possible
> measures necessary to prevent that need to be taken.
> 
>> I think there is no way for ossnode2 to know sda1 can be accessed by ossnode1.
>> Does this become a problem?
> 
> It does, hence the steps above.
> 
> b.
> 
> 
> _______________________________________________
> Lustre-discuss mailing list
> Lustre-discuss at clusterfs.com
> https://mail.clusterfs.com/mailman/listinfo/lustre-discuss
> 
> 


-- 
Kazuki Ohara
Sony Computer Entertainment Inc.
Computer Development Div. Distributed OS Development Dept.
Japan




More information about the lustre-discuss mailing list