Date: Sun, 02 Jan 2011 09:45:55 +0100 From: Attila Nagy <bra@fsn.hu> To: "J. Hellenthal" <jhell@DataIX.net> Cc: freebsd-fs@freebsd.org, freebsd-stable@freebsd.org, Martin Matuska <mm@freebsd.org> Subject: Re: New ZFSv28 patchset for 8-STABLE Message-ID: <4D203B43.7070607@fsn.hu> In-Reply-To: <4D1FF9AC.60200@DataIX.net> References: <4D0A09AF.3040005@FreeBSD.org> <4D1F7008.3050506@fsn.hu> <4D1FF9AC.60200@DataIX.net>
next in thread | previous in thread | raw e-mail | index | archive | help
On 01/02/2011 05:06 AM, J. Hellenthal wrote: > -----BEGIN PGP SIGNED MESSAGE----- > Hash: SHA1 > > On 01/01/2011 13:18, Attila Nagy wrote: >> On 12/16/2010 01:44 PM, Martin Matuska wrote: >>> Link to the patch: >>> >>> http://people.freebsd.org/~mm/patches/zfs/v28/stable-8-zfsv28-20101215.patch.xz >>> >>> >>> >> I've used this: >> http://people.freebsd.org/~mm/patches/zfs/v28/stable-8-zfsv28-20101223-nopython.patch.xz >> >> on a server with amd64, 8 G RAM, acting as a file server on >> ftp/http/rsync, the content being read only mounted with nullfs in >> jails, and the daemons use sendfile (ftp and http). >> >> The effects can be seen here: >> http://people.fsn.hu/~bra/freebsd/20110101-zfsv28-fbsd/ >> the exact moment of the switch can be seen on zfs_mem-week.png, where >> the L2 ARC has been discarded. >> >> What I see: >> - increased CPU load >> - decreased L2 ARC hit rate, decreased SSD (ad[46]), therefore increased >> hard disk load (IOPS graph) >> >> Maybe I could accept the higher system load as normal, because there >> were a lot of things changed between v15 and v28 (but I was hoping if I >> use the same feature set, it will require less CPU), but dropping the >> L2ARC hit rate so radically seems to be a major issue somewhere. >> As you can see from the memory stats, I have enough kernel memory to >> hold the L2 headers, so the L2 devices got filled up to their maximum >> capacity. >> >> Any ideas on what could cause these? I haven't upgraded the pool version >> and nothing was changed in the pool or in the file system. >> > Running arc_summary.pl[1] -p4 should print a summary about your l2arc > and you should also notice in that section that there is a high number > of "SPA Mismatch" mine usually grew to around 172k before I would notice > a crash and I could reliably trigger this while in scrub. > > What ever is causing this needs desperate attention! > > I emailed mm@ privately off-list when I noticed this going on but have > not received any feedback as of yet. It's at zero currently (2 days of uptime): kstat.zfs.misc.arcstats.l2_write_spa_mismatch: 0
Want to link to this message? Use this URL: <https://mail-archive.FreeBSD.org/cgi/mid.cgi?4D203B43.7070607>