[lustre-discuss] ZFS backed OSS out of memory

Carlson, Timothy S Timothy.Carlson at pnnl.gov
Thu Jun 23 09:13:51 PDT 2016


I've done my fair share of googling and run across some good information on ZFS backed Lustre tuning including this:


and various discussions around how to limit (or not) the ARC and clear it if needed.

That being said, here is my configuration.

Kernel 2.6.32-504.3.3.el6.x86_64
ZFS 0.6.3
Lustre 2.5.3 with a couple of patches
Single OST per OSS with 4 x RAIDZ2 4TB SAS drives
Log and Cache on separate SSDs
These OSSes are beefy with 128GB of memory and Dual E5-2630 v2 CPUs

 About 30 OSSes in all serving mostly a standard HPC cluster over FDR IB with a sprinkle of 10G

# more /etc/modprobe.d/lustre.conf
options lnet networks=o2ib9,tcp9(eth0)

ZFS backed MDS with same software stack.

The problem I am having is the OOM killer is whacking away at system processes on a few of the OSSes. 

"top" shows all my memory is in use with very little Cache or Buffer usage.

Tasks: 1429 total,   5 running, 1424 sleeping,   0 stopped,   0 zombie
Cpu(s):  0.0%us,  2.9%sy,  0.0%ni, 94.0%id,  3.1%wa,  0.0%hi,  0.0%si,  0.0%st
Mem:  132270088k total, 131370888k used,   899200k free,     1828k buffers
Swap: 61407100k total,     7940k used, 61399160k free,    10488k cached

   47 root      RT   0     0    0    0 S 30.0  0.0 372:57.33 migration/11

I had done zero tuning so I am getting the default ARC size of 1/2 the memory.

[root at lzfs18b ~]# arcstat.py 1
    time  read  miss  miss%  dmis  dm%  pmis  pm%  mmis  mm%  arcsz     c
09:11:50     0     0      0     0    0     0    0     0    0    63G   63G
09:11:51  6.2K  2.6K     41   206    6  2.4K   71     0    0    63G   63G
09:11:52   21K  4.0K     18   305    2  3.7K   34    18    0    63G   63G

The question is, if I have 128GB of RAM and ARC is only taking 63, where did the rest go and how can I get it back so that the OOM killer stops killing me?



More information about the lustre-discuss mailing list