Date: Thu, 21 Mar 2013 16:14:46 -0000 From: "Steven Hartland" <killing@multiplay.co.uk> To: "Josh Beard" <josh@signalboxes.net>, <freebsd-fs@freebsd.org> Subject: Re: ZFS + NFS poor performance after restarting from 100 day uptime Message-ID: <D763F64A24B54755BBF716E91D646F6A@multiplay.co.uk> References: <CAHDrHSsCunt9eQKjMy9epPBYTmaGs5HNgKV2%2BUKuW0RQZPpw%2BA@mail.gmail.com>
next in thread | previous in thread | raw e-mail | index | archive | help
----- Original Message ----- From: "Josh Beard" <josh@signalboxes.net> To: <freebsd-fs@freebsd.org> Sent: Thursday, March 21, 2013 3:53 PM Subject: ZFS + NFS poor performance after restarting from 100 day uptime > Hello, > > I have a system with 12 disks spread between 2 raidz1. I'm using the > native ("new") NFS to export a pool on this. This has worked very well all > along, but since a reboot, has performed horribly - unusably under load. > > The system was running 9.1-rc3 and I upgraded it to 9.1-release-p1 (GENERIC > kernel) after ~110 days of running (with zero performance issues). After > rebooting from the upgrade, I'm finding the disks seem constantly slammed. > gstat reports 90-100% busy most of the day with only ~100-130 ops/s. > > I didn't change any settings in /etc/sysctl.conf or /boot/loader. No ZFS > tuning, etc. I've looked at the commits between 9.1-rc3 and 9.1-release-p1 > and I can't see any reason why simply upgrading it would cause this. ... > A snip of gstat: > dT: 1.002s w: 1.000s > L(q) ops/s r/s kBps ms/r w/s kBps ms/w %busy Name > 0 0 0 0 0.0 0 0 0.0 0.0| cd0 > 0 1 0 0 0.0 1 32 0.2 0.0| da0 > 0 0 0 0 0.0 0 0 0.0 0.0| da0p1 > 0 1 0 0 0.0 1 32 0.2 0.0| da0p2 > 0 0 0 0 0.0 0 0 0.0 0.0| da0p3 > 4 160 126 1319 31.3 34 100 0.1 100.3| da1 > 4 146 110 1289 33.6 36 98 0.1 97.8| da2 > 4 142 107 1370 36.1 35 101 0.2 101.9| da3 > 4 121 95 1360 35.6 26 19 0.1 95.9| da4 > 4 151 117 1409 34.0 34 102 0.1 100.1| da5 > 4 141 109 1366 35.9 32 101 0.1 97.9| da6 > 4 136 118 1207 24.6 18 13 0.1 87.0| da7 > 4 118 102 1278 32.2 16 12 0.1 89.8| da8 > 4 138 116 1240 33.4 22 55 0.1 100.0| da9 > 4 133 117 1269 27.8 16 13 0.1 86.5| da10 > 4 121 102 1302 53.1 19 51 0.1 100.0| da11 > 4 120 99 1242 40.7 21 51 0.1 99.7| da12 Your ops/s are be maxing your disks. You say "only" but the ~190 ops/s is what HD's will peak at, so whatever our machine is doing is causing it to max the available IO for your disks. If you boot back to your previous kernel does the problem go away? If so you could look at the changes between the two kernel revisions for possible causes and if needed to a binary chop with kernel builds to narrow down the cause. Regards Steve Regards Steve ================================================ This e.mail is private and confidential between Multiplay (UK) Ltd. and the person or entity to whom it is addressed. In the event of misdirection, the recipient is prohibited from using, copying, printing or otherwise disseminating it or any information contained in it. In the event of misdirection, illegible or incomplete transmission please telephone +44 845 868 1337 or return the E.mail to postmaster@multiplay.co.uk.
Want to link to this message? Use this URL: <https://mail-archive.FreeBSD.org/cgi/mid.cgi?D763F64A24B54755BBF716E91D646F6A>