[Lustre-discuss] lru_size very small

Andreas Dilger adilger at sun.com
Sat Aug 23 04:22:09 PDT 2008


On Aug 22, 2008  15:39 -0400, Brock Palen wrote:
> It looks like lru_size is not a static parameter.  While on most of  
> our hosts it starts as zero.  Once the file system is accessed some  
> the values start to rise.  The values get highest for the MDS.
> 
> cat nobackup-MDT0000-mdc-000001022c433800/lru_size
>   3877

Yes, in 1.6.5 instead of having a static LRU size it is dynamic based
on load.  This optimizes the number of locks available to nodes that
have very different workloads than others (e.g. login/build nodes vs.
compute nodes vs. backup nodes).

> So in 1.6.5.1  are lock dynamically adjusted based on ram available  
> on the MDS/OSS's?  Notice how the value above is _much_ higher than  
> the default '100' in the manual.

The total number of locks available are now a function of the RAM
on the server.  I think the maximum is 50 locks/MB, but this is
hooked into the kernel VM so that in case of too much memory pressure
then the LRU size is shrunk.

> I should point out this value was 0  till I did a   'find . | wc -l'   
> in a directory.  The same is for regular access.  users on nodes that  
> access lustre have locks.  Nodes that have not had lustre access yet  
> are still 0  (by access I mean an application that uses our lustre  
> mount vs our NFS mount.)
> 
> Any feedback on the nature of locks and lru_size?
> We are looking to do what the manual says about upping the number on  
> the login nodes.

Yes, the manual needs an update.

Cheers, Andreas
--
Andreas Dilger
Sr. Staff Engineer, Lustre Group
Sun Microsystems of Canada, Inc.




More information about the lustre-discuss mailing list