From owner-freebsd-stable@FreeBSD.ORG Tue Mar 5 15:22:54 2013 Return-Path: Delivered-To: stable@freebsd.org Received: from mx1.freebsd.org (mx1.FreeBSD.org [8.8.178.115]) by hub.freebsd.org (Postfix) with ESMTP id C4B5CFDA for ; Tue, 5 Mar 2013 15:22:54 +0000 (UTC) (envelope-from gpalmer@freebsd.org) Received: from noop.in-addr.com (mail.in-addr.com [IPv6:2001:470:8:162::1]) by mx1.freebsd.org (Postfix) with ESMTP id 81563831 for ; Tue, 5 Mar 2013 15:22:54 +0000 (UTC) Received: from gjp by noop.in-addr.com with local (Exim 4.80.1 (FreeBSD)) (envelope-from ) id 1UCthc-0003yW-Aa; Tue, 05 Mar 2013 10:22:52 -0500 Date: Tue, 5 Mar 2013 10:22:52 -0500 From: Gary Palmer To: Garrett Wollman Subject: Re: ZFS "stalls" -- and maybe we should be talking about defaults? Message-ID: <20130305152252.GA52706@in-addr.com> References: <513524B2.6020600@denninger.net> <1362449266.92708.8.camel@btw.pki2.com> <51355F64.4040409@denninger.net> <201303050540.r255ecEC083742@hergotha.csail.mit.edu> MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii Content-Disposition: inline In-Reply-To: <201303050540.r255ecEC083742@hergotha.csail.mit.edu> X-SA-Exim-Connect-IP: X-SA-Exim-Mail-From: gpalmer@freebsd.org X-SA-Exim-Scanned: No (on noop.in-addr.com); SAEximRunCond expanded to false Cc: stable@freebsd.org, killing@multiplay.co.uk X-BeenThere: freebsd-stable@freebsd.org X-Mailman-Version: 2.1.14 Precedence: list List-Id: Production branch of FreeBSD source code List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , X-List-Received-Date: Tue, 05 Mar 2013 15:22:54 -0000 On Tue, Mar 05, 2013 at 12:40:38AM -0500, Garrett Wollman wrote: > In article <8C68812328E3483BA9786EF15591124D@multiplay.co.uk>, > killing@multiplay.co.uk writes: > > >Now interesting you should say that I've seen a stall recently on ZFS > >only box running on 6 x SSD RAIDZ2. > > > >The stall was caused by fairly large mysql import, with nothing else > >running. > > > >Then it happened I thought the machine had wedged, but minutes (not > >seconds) later, everything sprung into action again. > > I have certainly seen what you might describe as "stalls", caused, so > far as I can tell, by kernel memory starvation. I've seen it take as > much as a half an hour to recover from these (which is too long for my > users). Right now I have the ARC limited to 64 GB (on a 96 GB file > server) and that has made it more stable, but it's still not behaving > quite as I would like, and I'm looking to put more memory into the > system (to be used for non-ARC functions). Looking at my munin > graphs, I find that backups in particular put very heavy pressure on, > doubling the UMA allocations over steady-state, and this takes about > four or five hours to climb back down. See > for an example. > > Some of the stalls are undoubtedly caused by internal fragmentation > rather than actual data in use. (Solaris used to have this issue, and > some hooks were added to allow some amount of garbage collection with > the cooperation of the filesystem.) Just as a note that there was a page I read in the past few months that pointed out that having a huge ARC may not always be in the best interests of the system. Some operation on the filesystem (I forget what, apologies) caused the system to churn through the ARC and discard most of it, while regular I/O was blocked Unfortunately I cannot remember where I found that page now and I don't appear to have bookmarked it >From what has been said in this thread I'm not convinced that people are hitting this issue, however I would like to raise it for consideration Regards, Gary