[lustre-discuss] Limiting Lustre memory use?

bill broadley bill at broadley.org
Fri Feb 18 13:42:39 PST 2022

On a cluster I managed (without Lustre), we had many problems with users running 
nodes out of ram which often killed the node.  We added cgroup support to slurm 
and those problems disappeared.  Nearly 100% of the time get a cgroup OOM 
instead of a kernel OOM and the nodes would stay up and stable. This became 
doubly important when we started allowing jobs to share nodes and didn't want 
job A to be able to crash job B.

I've tried similar on a Lustre enabled cluster and it seems like the memory used 
by Lustre (which I believe is in the kernel and outside of the job's cgroup).  I 
think part of the problem is I believe Lustre caches metadata in the linux page 
cache, but not data.  I've tried reducing the ram available to slurm, but still 
getting kernel OOMs instead of cgroup OOMs.

Anyone have a suggestion for fixing this?  Is there any way to limit Lustre's 
memory use in the kernel?  Or force that caching into userspace and inside the 
cgroup?  Or possibly out of ram and onto a client local NVMe?

More information about the lustre-discuss mailing list