Date: Wed, 24 Dec 2008 22:35:11 -0800 From: Matt Simerson <matt@corp.spry.com> To: freebsd-fs@freebsd.org Subject: Re: ZFS performance gains real or imaginary? Message-ID: <5C120CEB-6CEB-4722-BB23-7E4B83F779C2@corp.spry.com> In-Reply-To: <20081225052903.GC87625@egr.msu.edu> References: <22C8092E-210F-4E91-AA09-CFD38966975C@spry.com> <494AE6F4.30506@modulus.org> <1424BEB3-69FE-4BA2-884F-4862B3D7BCFD@corp.spry.com> <20081224034812.GR87625@egr.msu.edu> <F771A20F-384C-4435-847D-22FEFDCB1CDD@corp.spry.com> <20081225052903.GC87625@egr.msu.edu>
next in thread | previous in thread | raw e-mail | index | archive | help
On Dec 24, 2008, at 9:29 PM, Adam McDougall wrote: >> On Wed, Dec 24, 2008 at 01:00:14PM -0800, Matt Simerson wrote: >> >> On Dec 23, 2008, at 7:48 PM, Adam McDougall wrote: >> >>>> On Tue, Dec 23, 2008 at 12:43:47PM -0800, Matt Simerson wrote: >>>> >>>>> On Dec 18, 2008, at 4:12 PM, Andrew Snow wrote: >>>>> >>>>>> If so, then I really should be upgrading my production ZFS >>>>>> servers >>>>>> to the latest -HEAD. >>>>> >>>>> Thats correct, that is the only way to get the best working >>>>> version >>>>> of ZFS. Of course, then everything is unstable and broken - eg. >>>>> SMBFS became unusable for me and would crash the server. >>>> >>>> Unfortunately, the newer kernel hangs much more frequently. >>>> >>>> I have these settings in /boot/loader.conf >>>> >>>> vm.kmem_size="1536M" >>>> vm.kmem_size_max="1536M" >>>> vfs.zfs.arc_max="100M" >>>> >>>> I have also experimented with vfs.zfs.prefetch_disable, >>>> vfs.zfs.arc_min in the past, and I'm open to suggestions on what >>>> might help under this workload (multiple concurrent rsync >>>> processes from remote systems to this one). >>> >>> Can you try: >>> >>> vm.kmem_size=2G >>> vm.kmem_size_max=2G >>> vfs.zfs.arc_max=512M >>> >>> This has been working for me on one amd64 system that only >>> has 2G of ram but had similar problem frequency to yours. I >>> don't know if its coincidence with the data that I am rsyncing >>> lately, but: 10:47PM up 22 days, 7:12 >> >> I made it 23 minutes. I've reduced my rsync concurrency to 1, so I'm >> not hitting the system nearly as hard but it seems not to matter. >> >> Other workloads, like a 'make buildworld' will complete with no >> problems. For whatever reason, rsync sessions of entire unix systems >> to my backup servers are very troublesome. >> >> Matt > > Ok. Since you have 16G of ram, I suppose you could try setting both > kmem > sizes to something like 8G to see if it makes a difference? I'm > getting > a feeling that even if we don't see an outright failure, it might be > deadlocking due to a kmem shortage. back01# w 10:17PM up 40 mins, 2 users, load averages: 4.20, 3.07, 1.74 This is with: vm.kmem_size="4G" vm.kmem_size_max="4G" vfs.zfs.arc_max="512M" I'll let it trundle along with that setting and see how long it lasts. Matt PS: These settings earlier today resulted in 12+ hours of uptime, until I rebooted to test raising kmem_size to 4G. vm.kmem_size="2G" vm.kmem_size_max="2G" vfs.zfs.arc_max="512M" vfs.zfs.zil_disable=1 vfs.zfs.prefetch_disable=1 PPS: If/when it hangs with 4G, I'll raise it again to 6 or 8 GB and see how long it lasts. Whatever pattern emerges might be useful for Pawel.
Want to link to this message? Use this URL: <https://mail-archive.FreeBSD.org/cgi/mid.cgi?5C120CEB-6CEB-4722-BB23-7E4B83F779C2>