** please reply to the group, if not remember to leave out the "spam" part
in the email address **
Hi all,
I'm having a question regarding NFS, and the excessive kernel% that clients
encounter.
I have a quad CPU E4500 server which is an NFS-server, using 2 mirrored T3
trays with loads of disks configured in RAID5. Total volume size is about
200GB. I also have a Netapp F760 for testing too. These nodes are hooked up
to our network using Gigabit Ethernet.
I have several NFS-clients, ranging from T1, E220, to another E4500 (4cpu),
which all run a mail application, that access mailboxes over NFS to the
storage nodes mentioned above.
What I see is that when I stress test, that the kernel activity on the
clients goes way up high, up to 75%. It doesn't matter if the client is
having 1 or multiple CPU's... I would expect everything to go up, but not in
this proportion...
when I look at an "iostat -xn 4" on the client, the NFS mount doesn't seem
to be that busy:
extended device statistics
r/s w/s kr/s kw/s wait actv wsvc_t asvc_t %w %b
device
302.0 0.0 395.5 0.0 0.0 0.3 0.0 0.9 0 23
192.168.13.33:/global/emtc01-ds2/nfs/data
353.3 0.0 403.8 0.0 0.0 0.3 0.0 0.9 0 28
192.168.13.33:/global/emtc01-ds2/nfs/data
366.0 0.0 429.0 0.0 0.0 0.3 0.0 0.9 0 29
192.168.13.33:/global/emtc01-ds2/nfs/data
351.5 0.0 423.2 0.0 0.0 0.3 0.0 0.9 0 27
192.168.13.33:/global/emtc01-ds2/nfs/data
352.5 0.0 424.2 0.0 0.0 0.3 0.0 1.0 0 29
192.168.13.33:/global/emtc01-ds2/nfs/data
325.5 0.0 409.5 0.0 0.0 0.3 0.0 0.9 0 24
192.168.13.33:/global/emtc01-ds2/nfs/data
328.6 0.0 408.4 0.0 0.0 0.3 0.0 0.9 0 26
192.168.13.33:/global/emtc01-ds2/nfs/data
314.6 0.0 401.1 0.0 0.0 0.3 0.0 0.9 0 25
192.168.13.33:/global/emtc01-ds2/nfs/data
291.9 0.0 376.6 0.0 0.0 0.3 0.0 0.9 0 23
192.168.13.33:/global/emtc01-ds2/nfs/data
333.1 0.0 392.4 0.0 0.0 0.3 0.0 0.9 0 26
192.168.13.33:/global/emtc01-ds2/nfs/data
341.5 0.0 402.9 0.0 0.0 0.3 0.0 0.9 0 27
192.168.13.33:/global/emtc01-ds2/nfs/data
325.3 0.0 395.1 0.0 0.0 0.3 0.0 1.0 0 28
192.168.13.33:/global/emtc01-ds2/nfs/data
333.0 0.0 407.6 0.0 0.0 0.3 0.0 0.9 0 26
192.168.13.33:/global/emtc01-ds2/nfs/data
325.5 0.0 414.8 0.0 0.0 0.3 0.0 0.9 0 25
192.168.13.33:/global/emtc01-ds2/nfs/data
I'm guessing this'll probably have to do with tuning NFS and the likes, but
what ? Has any of you an explanation for this?
On the storage node side.. disks aren't busy at all! There's no iowait%,
there is some kernel activity, but only to a normal 10% of total cpu
capacity, so that seems acceptable..
extended device statistics tty
device r/s w/s kr/s kw/s wait actv svc_t %w %b
2/md120 0.0 111.0 0.0 222.0 0.0 3.0 27.4 0 5
2/md120 0.0 109.5 0.0 222.0 0.0 2.8 25.4 0 5
2/md120 0.0 110.0 0.0 220.0 0.0 2.9 26.6 0 5
2/md120 0.0 110.5 0.0 221.0 0.0 3.0 27.5 0 5
2/md120 0.0 117.0 0.0 236.3 0.0 3.2 27.1 0 5
Thanks for throwing in some tuning hints or experiences!
Regards,
Renaat Dumon