Return-Path: linux-nfs-owner@vger.kernel.org Received: from smtp02.citrix.com ([66.165.176.63]:28984 "EHLO SMTP02.CITRIX.COM" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S932402Ab1KJLP0 (ORCPT ); Thu, 10 Nov 2011 06:15:26 -0500 Message-ID: <4EBBB247.40805@citrix.com> Date: Thu, 10 Nov 2011 11:15:19 +0000 From: Andrew Cooper MIME-Version: 1.0 To: Chuck Lever CC: linux-nfs Subject: Re: unexpected NFS timeouts, related to sync/async soft mounts over TCP References: <4EBAC88D.40902@citrix.com> In-Reply-To: Content-Type: text/plain; charset="ISO-8859-1" Sender: linux-nfs-owner@vger.kernel.org List-ID: On 09/11/11 22:36, Chuck Lever wrote: > On Nov 9, 2011, at 1:38 PM, Andrew Cooper wrote: > >> Hello, >> >> I am debugging an issue for a customer whereby an NFS mount undergoes an >> unexpected timeout and throws an EIO. >> >> Using synchronous mounts results in no problems whatsoever, whereas >> using asynchronous mounts causes the problem to occur. >> >> The problem exists only on 10GiB networks: I can't repro on slower >> networks at all, even if it is the same 10GiB network link limited to >> 1GiB. It reproduces on multiple versions of XenServer (both Xen 3 and >> Xen 4) with a 2.6.32 based dom0 kernel multiple different 10GiB network >> cards (be2net and bnx2x specifically tested). It also reproduces >> against multiple different netapp servers running different firmware >> versions. It also appears to reproduces with NFSv3 and NFSv4, although >> the customer setup uses NFSv3 >> >> The difference between mount commands are "noac" for the synchronous >> case as opposed to "actimeo=0" for the asynchronous case. >> >> The /proc/mounts entry for async is: >> 10.70.148.3:/vol/camb5e3b16tmp/82ab4a02-c10f-144a-e8c2-2ae04f213b2a >> /var/run/sr-mount/82ab4a02-c10f-144a-e8c2-2ae04f213b2a nfs >> rw,relatime,vers=3,rsize=65536,wsize=65536,namlen=255,acregmin=0,acregmax=0,acdirmin=0,acdirmax=0,soft,proto=tcp,port=65535,timeo=133,retrans=0,sec=sys,mountport=65535,addr=10.70.148.3 >> 0 0 >> >> while the entry for sync is: >> 10.70.148.3:/vol/camb5e3b15tmp/bc11aad3-b345-519f-1661-50e16bed6cdb >> /var/run/sr-mount/bc11aad3-b345-519f-1661-50e16bed6cdb nfs >> rw,sync,relatime,vers=3,rsize=65536,wsize=65536,namlen=255,acregmin=0,acregmax=0,acdirmin=0,acdirmax=0,soft,proto=tcp,port=65535,timeo=133,retrans=0,sec=sys,mountport=65535,addr=10.70.148.3 >> 0 0 > Using the "soft" mount option means that any NFS level timeout is communicated to applications as EIO. The client appears to be working as expected. Do you see this problem if you replace the "soft" mount option with the "hard" mount option? I am aware of soft mounts and EIOs. I forgot to say that it also appears with hard mounts, with the difference that the client eventually retries and continues. > Also, timeo=133 is actually a 13.3 second timeout setting. The value of the timeo= mount option is in deciseconds. This is unreasonably short for RPC over TCP, especially if the server becomes loaded. A more reasonable timeout setting for RPC over TCP is 600. I am aware of this as well. The decisions about NFS options are done by a separate team who have justified their decision to use softmounts with a short timeo. In our use case, the userland process needs to be informed very quickly if the server goes down so it can fail over to the backup. 13 seconds was considered the longest acceptable time for virtual machines to be without their disks. The problem here is that some bug in the kernel is causing userland to think the server has gone down while it is perfectly healthy and running normally. > The difference between what you describe as a "synchronous" mount and an "asynchronous" mount is that the asynchronous mount can send multiple NFS WRITE operations for one file on the wire to the server concurrently. The "synchronous" mount throttles write traffic, and requires each WRITE to complete at the server before the client sends the next WRITE. You can see that the "asynchronous" mount can send a lot more traffic per unit time, and thus may result in the client waiting longer for WRITEs to complete. I did an analysis of the latency between write requests and replies. At any one time in the stream, there are no more than 4 oustanding requests, and the maximum latency between request and reply (as seen by the client) is 50ms >> The bug itself feels very like a timing issue to me. It becomes >> substantially easier to reproduce if the kernel has a high workload: >> Different options involve disabling things like TSO offload on the >> network card or dd'ing from /dev/urandom to /tmp, but the problem cant >> be reproduced occasionally without any artificial extra workload. The >> problem itself can be reprodued with a simple dd from /dev/zero onto the >> nfs mount, which proceeds to transfer at ~220MBps until the EIO is hit. >> >> I have analyzed the TCP stream using TCPdump and Wireshark. Everything >> appears fine, with no NFS errors at all. However, the client sticks a >> TCP FIN in the final continuation of one of its 64k block writes, in the >> middle of writing the file. >> >> The server ACKs the outstanding packets, and sends a Write Reply >> (without error), but does not send its own FIN. Then, 15 seconds later, >> the client sends a TCP RST, and throws an EIO back to userspace. >> >> I am aware that sending a RST after 15 seconds is expected behavior >> following changeset 7d1e8255cf959fba7ee2317550dfde39f0b936ae, in an >> attempt to prevent deadlocks for a half closed TCP connection. I have >> searched through the commit log all the way to tip, but can not find a >> commit making any reference to problems similar to this. >> >> What I am struggling to work out is why the client is sending a FIN in >> the middle of an otherwise fine and error-free NFS stream. using >> "rpcdebug -m rpc -s trans" only says "disconnected" without giving a >> reason for disconnecting. >> >> Is there a different debug flag I can use to find out why NFS thinks it >> needs to disconnect, or does anyone have any suggestions as to where I >> should start debugging the code? I suspect the problem is in the sunrpc >> subsystem but I am not very familiar with code this high level in the >> kernel. >> >> Any help greatly appreciated >> >> -- >> Andrew Cooper - Dom0 Kernel Engineer, Citrix XenServer >> T: +44 (0)1223 225 900, http://www.citrix.com >> >> -- >> To unsubscribe from this list: send the line "unsubscribe linux-nfs" in >> the body of a message to majordomo@vger.kernel.org >> More majordomo info at http://vger.kernel.org/majordomo-info.html -- Andrew Cooper - Dom0 Kernel Engineer, Citrix XenServer T: +44 (0)1223 225 900, http://www.citrix.com