Date: Thu, 14 Feb 2019 13:21:59 +0700 From: Eugene Grosbein <eugen@grosbein.net> To: Mark Johnston <markj@freebsd.org> Cc: FreeBSD stable <freebsd-stable@freebsd.org> Subject: Re: 11.2-STABLE kernel wired memory leak Message-ID: <5C650907.7040703@grosbein.net> In-Reply-To: <9a263536-45e3-c690-e45e-d8ece7d1f388@grosbein.net> References: <d8c7abc0-3ba1-40e4-22b1-1b30d28ced14@grosbein.net> <20190212163446.GA29847@raichu> <9a263536-45e3-c690-e45e-d8ece7d1f388@grosbein.net>
next in thread | previous in thread | raw e-mail | index | archive | help
On 13.02.2019 19:04, Eugene Grosbein wrote: > 12.02.2019 23:34, Mark Johnston wrote: > >> On Tue, Feb 12, 2019 at 11:14:31PM +0700, Eugene Grosbein wrote: >>> Hi! >>> >>> Long story short: 11.2-STABLE/amd64 r335757 leaked over 4600MB kernel wired memory over 81 days uptime >>> out of 8GB total RAM. >>> >>> Details follow. >>> >>> I have a workstation running Xorg, Firefox, Thunderbird, LibreOffice and occasionally VirtualBox for single VM. >>> >>> It has two identical 320GB HDDs combined with single graid-based array with "Intel" >>> on-disk format having 3 volumes: >>> - one "RAID1" volume /dev/raid/r0 occupies first 10GB or each HDD; >>> - two "SINGLE" volumes /dev/raid/r1 and /dev/raid/r2 that utilize "tails" of HDDs (310GB each). >>> >>> /dev/raid/r0 (10GB) has MBR partitioning and two slices: >>> - /dev/raid/r0s1 (8GB) is used for swap; >>> - /dev/raid/r0s2 (2GB) is used by non-redundant ZFS pool named "os" that contains only >>> root file system (177M used) and /usr file system (340M used). >>> >>> There is also second pool (ZMIRROR) named "z" built directly on top of /dev/raid/r[12] volumes, >>> this pool contains all other file systems including /var, /home, /usr/ports, /usr/local, /usr/{src|obj} etc. >>> >>> # zpool list >>> NAME SIZE ALLOC FREE CKPOINT EXPANDSZ FRAG CAP DEDUP HEALTH ALTROOT >>> os 1,98G 520M 1,48G - - 55% 25% 1.00x ONLINE - >>> z 288G 79,5G 209G - - 34% 27% 1.00x ONLINE - >>> >>> This way I have swap outside of ZFS, boot blocks and partitioning mirrored by means of GEOM_RAID and >>> can use local console to break to single user mode to unmount all file system other than root and /usr >>> and can even export bigger ZFS pool "z". And I did that to see that ARC usage >>> (limited with vfs.zfs.arc_max="3G" in /boot/loader.conf) dropped from over 2500MB >>> down to 44MB but "Wired" stays high. Now after I imported "z" back and booted to multiuser mode >>> top(1) shows: >>> >>> last pid: 51242; load averages: 0.24, 0.16, 0.13 up 81+02:38:38 22:59:18 >>> 104 processes: 1 running, 103 sleeping >>> CPU: 0.0% user, 0.0% nice, 0.4% system, 0.2% interrupt, 99.4% idle >>> Mem: 84M Active, 550M Inact, 4K Laundry, 4689M Wired, 2595M Free >>> ARC: 273M Total, 86M MFU, 172M MRU, 64K Anon, 1817K Header, 12M Other >>> 117M Compressed, 333M Uncompressed, 2.83:1 Ratio >>> Swap: 8192M Total, 940K Used, 8191M Free >>> >>> I have KDB and DDB in my custom kernel also. How do I debug the leak further? >>> >>> I use nvidia-driver-340-340.107 driver for GK208 [GeForce GT 710B] video card. >>> Here are outputs of "vmstat -m": http://www.grosbein.net/freebsd/leak/vmstat-m.txt >>> and "vmstat -z": http://www.grosbein.net/freebsd/leak/vmstat-z.txt >> >> I suspect that the "leaked" memory is simply being used to cache UMA >> items. Note that the values in the FREE column of vmstat -z output are >> quite large. The cached items are reclaimed only when the page daemon >> wakes up to reclaim memory; if there are no memory shortages, large >> amounts of memory may accumulate in UMA caches. In this case, the sum >> of the product of columns 2 and 5 gives a total of roughly 4GB cached. > > After another day with mostly idle system, "Wired" increased to more than 6GB out of 8GB total. > I've tried to increase sysctl vm.v_free_min from default 12838 (50MB) > upto 1048576 (4GB) and "Wired" dropped a bit but it is still huge, 5060M: > > last pid: 61619; load averages: 1.05, 0.78, 0.40 up 81+22:33:09 18:53:49 > 119 processes: 1 running, 118 sleeping > CPU: 0.0% user, 0.0% nice, 50.0% system, 0.0% interrupt, 50.0% idle > Mem: 47M Active, 731M Inact, 4K Laundry, 5060M Wired, 2080M Free > ARC: 3049M Total, 216M MFU, 2428M MRU, 64K Anon, 80M Header, 325M Other > 2341M Compressed, 5874M Uncompressed, 2.51:1 Ratio > Swap: 8192M Total, 940K Used, 8191M Free > > # sysctl vm.v_free_min > vm.v_free_min: 1048576 > # sysctl vm.stats.vm.v_free_count > vm.stats.vm.v_free_count: 533232 > > ARC probably cached results of nightly periodic jobs traversing file system trees > and hit its limit (3G). > > Still cannot understand where have another 2G (5G-3G) of wired memory leaked to? After several more hours of pretty idle time with quoted settings memory finally got moved from "Wired" " to "Free" category. I got the system to single user mode again, exported bigger ZFS pool again and ARC dropped down to 18M and Wired to less than 500M. After re-import and boot to multiuser with Xorg running, Wired is less than 800M. So, there seems to be no leak but insanely long time for Wired memory to be freed, even with extreme value for vm.v_free_min. Even heavy memory pressure does not speed things up. It seems, page daemon is somewhat broken and fails to free wired memory timely.
Want to link to this message? Use this URL: <https://mail-archive.FreeBSD.org/cgi/mid.cgi?5C650907.7040703>