Date: Mon, 23 Aug 2010 19:10:59 -0700 From: Artem Belevich <fbsdlist@src.cx> To: jhell <jhell@dataix.net> Cc: freebsd-hackers@freebsd.org, zfs-devel@freebsd.org, Andriy Gapon <avg@freebsd.org> Subject: Re: ZFS arc_reclaim_needed: better cooperation with pagedaemon Message-ID: <AANLkTikGBWStNkyL=J3wXWzDsUftSLsum1vKryB7dU2T@mail.gmail.com> In-Reply-To: <4C72DDC3.1000006@DataIX.net> References: <4C719AB9.9020006@freebsd.org> <AANLkTinreSt_Dk_J5vpZ6xrs=snqYu8zKfO0X6H-x_n3@mail.gmail.com> <4C721161.40403@freebsd.org> <AANLkTikg3e%2BGvZNtb5Uk3yAvWYwrgfF-4Rx0dDdooyQM@mail.gmail.com> <4C72DD1A.9070204@DataIX.net> <4C72DDC3.1000006@DataIX.net>
next in thread | previous in thread | raw e-mail | index | archive | help
Could you try following experiments before and after the patch while monitoring kstat.zfs.misc.arcstats.size and vm.stats.vm.v_inactive_count. First prepare the data. * You'll need some files totalling around the amount of physical memory on your box. Multiple copies of /usr/src should do the trick. * Place one copy on UFS filesystem and another on ZFS Experiment #1: * Prime ARC by tarring dataset on ZFS into /dev/null. * Now tar both datasets in parallel with output to /dev/null Previously you would end up with ARC size shrinking down to arc_min. What I hope to see after the patch is that inactive memory and ARC reach some sort of equilibrium with neither monopolizing all available memory. #Experiment #2: If equilibrium is reached, try running some application that would allocate and use about 1/2 of your physical memory. Something like that perl one-liner used to cause memory shortage, only a bit less drastic. perl -e '$x=3D"x"x1_000_000_000'; # this should allocate about 2GB. Tune the number to suit your system. Again, in the past ARC would be the one feeing up the memory. Let's see if inactive list gives up some, too. --Artem On Mon, Aug 23, 2010 at 1:44 PM, jhell <jhell@dataix.net> wrote: > On 08/23/2010 16:42, jhell wrote: >> On 08/23/2010 03:28, Artem Belevich wrote: >>> Can anyone test the patch on a system with mix of UFS/ZFS filesystems >>> and see if the change mitigates or solves the issue with inactive >>> memory excessively backpressuring ARC. >> >> I have a system currently patched up to ZFSv15 and mm@'s metaslab patch >> running that I can test this on. Throw me a patch and some specific >> tests and I can have the results back to you in the next 2 days. >> > > Forget the patch 1 line change I can hand type that in. As for specific > tests, let me know... > > -- > > =A0jhell,v >
Want to link to this message? Use this URL: <https://mail-archive.FreeBSD.org/cgi/mid.cgi?AANLkTikGBWStNkyL=J3wXWzDsUftSLsum1vKryB7dU2T>