I believe we're seeing similar problems with our setup. We're having a 24-disk RAID10 setup on our box, with a 22TB-sized XFS filesystem exported over NFS(v3) to our VMWare Cluster. During initial load-testing by means of iometer and dd, we triggered strange behaviour on behalf of nfsd, which made common operations (such as readdir()) on the mounted export excruciatingly slow (we're talking more than an hour for a simple `ls` to complete from within an empty directory). Changing from the stock Lucid 2.6.32-kernel to later releases made things (seemingly) go away during load testing, but it popped up again later on, when the system was moved into semi-production as the backup storage for the aforementioned cluster. Other hardware involed is/are Intel Corporation 82598EB 10-Gigabit Ethernet adapters driven by ixgbe, two Adaptec AAC-RAID controllers driven by aacraid, on an Intel 5520-based dual-socket, fource-core (HT disabled) Nehalem machine.
This happened after copying over a few VM images from our primary to our backup storage over NFS(v3). The machine doesn't crash, but NFS performance is rather unimpressive during and after these operations.
I'll investigate if Thag's suggested workaround is applicable in our situation, and if it is, if it helps getting things to work normally. However, as we're not using multiple IPv4 addresses with our NICs afaik, I'm on the lookout for alternative solutions to the problem, or theories what may cause it.
I believe we're seeing similar problems with our setup. We're having a 24-disk RAID10 setup on our box, with a 22TB-sized XFS filesystem exported over NFS(v3) to our VMWare Cluster. During initial load-testing by means of iometer and dd, we triggered strange behaviour on behalf of nfsd, which made common operations (such as readdir()) on the mounted export excruciatingly slow (we're talking more than an hour for a simple `ls` to complete from within an empty directory). Changing from the stock Lucid 2.6.32-kernel to later releases made things (seemingly) go away during load testing, but it popped up again later on, when the system was moved into semi-production as the backup storage for the aforementioned cluster. Other hardware involed is/are Intel Corporation 82598EB 10-Gigabit Ethernet adapters driven by ixgbe, two Adaptec AAC-RAID controllers driven by aacraid, on an Intel 5520-based dual-socket, fource-core (HT disabled) Nehalem machine.
We see backtraces like the following: kernel/ hung_task_ timeout_ secs" disables this message. 838>] __mutex_ lock_slowpath+ 0xe8/0x170 47b>] mutex_lock+ 0x2b/0x50 2ff>] nfsd_unlink+ 0xaf/0x240 [nfsd] d54>] nfsd3_proc_ remove+ 0x84/0x100 [nfsd] 3fb>] nfsd_dispatch+ 0xbb/0x210 [nfsd] 625>] svc_process_ common+ 0x325/0x650 [sunrpc] a60>] ? nfsd+0x0/0x150 [nfsd] a83>] svc_process+ 0x133/0x150 [sunrpc] b1d>] nfsd+0xbd/0x150 [nfsd] 8d6>] kthread+0x96/0xa0 e64>] kernel_ thread_ helper+ 0x4/0x10 840>] ? kthread+0x0/0xa0 e60>] ? kernel_ thread_ helper+ 0x0/0x10
[150122.133802] INFO: task nfsd:2145 blocked for more than 120 seconds.
[150122.133853] "echo 0 > /proc/sys/
[150122.133934] nfsd D ffff880001e15880 0 2145 2 0x00000000
[150122.133937] ffff8806614e1cd0 0000000000000046 ffff8806614e1cd0 ffff8806614e1fd8
[150122.133940] ffff88065b7dc4a0 0000000000015880 0000000000015880 ffff8806614e1fd8
[150122.133942] 0000000000015880 ffff8806614e1fd8 0000000000015880 ffff88065b7dc4a0
[150122.133945] Call Trace:
[150122.133947] [<ffffffff81576
[150122.133949] [<ffffffff81576
[150122.133954] [<ffffffffa03e6
[150122.133960] [<ffffffffa03ed
[150122.133964] [<ffffffffa03df
[150122.133972] [<ffffffffa021d
[150122.133977] [<ffffffffa03df
[150122.133984] [<ffffffffa021d
[150122.133988] [<ffffffffa03df
[150122.133990] [<ffffffff8107f
[150122.133993] [<ffffffff8100b
[150122.133995] [<ffffffff8107f
[150122.133997] [<ffffffff8100b
This happened after copying over a few VM images from our primary to our backup storage over NFS(v3). The machine doesn't crash, but NFS performance is rather unimpressive during and after these operations.
I'll investigate if Thag's suggested workaround is applicable in our situation, and if it is, if it helps getting things to work normally. However, as we're not using multiple IPv4 addresses with our NICs afaik, I'm on the lookout for alternative solutions to the problem, or theories what may cause it.