[lustre-devel] how to fix unfair cpu_parttions load distribution?
liang.zhen at intel.com
Wed Jun 3 06:11:26 PDT 2015
Hi, one thing you may want to try is, turn on portal rotor by this way:
Echo 1 > /proc/sys/lnet/portal_rotor
This will round-robin dispatch incoming requests to IO services in
different partitions, the drawback of this approach is, incoming request
will lose NUMA affinity, so I suspect you will lose a little performance.
Another way to improve this is, create multiple lnet networks and evenly
distribute clients to different networks, and explicitly bind network on
CPU partition (see lustre manual for details).
On 6/3/15, 3:44 PM, "Alexander Zarochentsev"
<alexander.zarochentsev at seagate.com> wrote:
>FPP IOR tests with 8-16 clients show difference in write speed between
>for example result files after 60 sec stonewalling write test:
>-rw-r--r-- 1 root root 11G Apr 22 15:03 out.00000000.0
>-rw-r--r-- 1 root root 22G Apr 22 15:03 out.00000001.0
>-rw-r--r-- 1 root root 6.4G Apr 22 15:03 out.00000002.0
>-rw-r--r-- 1 root root 9.8G Apr 22 15:03 out.00000003.0
>-rw-r--r-- 1 root root 6.5G Apr 22 15:03 out.00000004.0
>-rw-r--r-- 1 root root 6.7G Apr 22 15:03 out.00000005.0
>-rw-r--r-- 1 root root 9.9G Apr 22 15:03 out.00000006.0
>-rw-r--r-- 1 root root 11G Apr 22 15:03 out.00000007.0
>-rw-r--r-- 1 root root 11G Apr 22 15:03 out.00000008.0
>-rw-r--r-- 1 root root 21G Apr 22 15:03 out.00000009.0
>-rw-r--r-- 1 root root 6.8G Apr 22 15:03 out.00000010.0
>-rw-r--r-- 1 root root 11G Apr 22 15:03 out.00000011.0
>-rw-r--r-- 1 root root 6.7G Apr 22 15:03 out.00000012.0
>-rw-r--r-- 1 root root 6.6G Apr 22 15:03 out.00000013.0
>-rw-r--r-- 1 root root 11G Apr 22 15:03 out.00000014.0
>-rw-r--r-- 1 root root 11G Apr 22 15:03 out.00000015.0
>the fastest client was able to write 22GB and the slowest one only 6.7GB.
>The funny thing that file size distribution depends on clients and
>sometimes (rare) all clients write at the same speed.
>LNET provides a mapping between client NIDs and CPU partitions
>calculated as a hash of 64bit NID. The mapping is often not fair for
>small number of clients and I guess may be not so good for larger
>client pool too (depends on how client nids are assigned).
>Unfair mapping causes uneven load on cpu partitions and that different
>client speed. Disabling cpu partitions in libcfs restores equal client
>write speed with some cost of performance.
>Currently there is no mechanism to balance load between CPs. NRS might
>be a solution but it is not, it works on each CP individually
>(correct?). At least no effects from non-default NRS policies were
>I think better load distribution may gave some performance gain. Also
>NRS does not work as expected with CPs.
>Replacing NID->CP mapping by RR looks not easy. Any ideas how it the
>load distribution can be improved ?
>Seagate Technology, LLC
More information about the lustre-devel