Skip site navigation (1)Skip section navigation (2)
Date:      Tue, 28 Jan 2014 20:50:34 +0200
From:      Anton Sayetsky <vsjcfm@gmail.com>
To:        freebsd-fs@freebsd.org
Subject:   Re: ZFS and Wired memory, again
Message-ID:  <CAFG2KCJUWtLwR_j2Ykr1J%2BO6PESgs3RdztS_Yx0gNJ_7UmrGJw@mail.gmail.com>
In-Reply-To: <CAFG2KC%2BZSHEVFbpPD9e1QHRdY=Sd6EuAD80vyDLDDQcpgCQNhA@mail.gmail.com>
References:  <CAFG2KC%2BZSHEVFbpPD9e1QHRdY=Sd6EuAD80vyDLDDQcpgCQNhA@mail.gmail.com>

next in thread | previous in thread | raw e-mail | index | archive | help
2013-11-22 Anton Sayetsky <vsjcfm@gmail.com>:
> Hello,
>
> I'm planning to deploy a ~150 TiB ZFS pool and when playing with ZFS
> noticed that amount of wired memory is MUCH bigger than ARC size (in
> absence of other hungry memory consumers, of course). I'm afraid that
> this strange behavior may become even worse on a machine with big pool
> and some hundreds gibibytes of RAM.
>
> So let me explain what happened.
>
> Immediately after booting system top says the following:
> =====
> Mem: 14M Active, 13M Inact, 117M Wired, 2947M Free
> ARC: 24M Total, 5360K MFU, 18M MRU, 16K Anon, 328K Header, 1096K Other
> =====
> Ok, wired mem - arc = 92 MiB
>
> Then I started to read pool (tar cpf /dev/null /).
> Memory usage when ARC size is ~1GiB
> =====
> Mem: 16M Active, 15M Inact, 1410M Wired, 1649M Free
> ARC: 1114M Total, 29M MFU, 972M MRU, 21K Anon, 18M Header, 95M Other
> =====
> 1410-1114=296 MiB
>
> Memory usage when ARC size reaches it's maximum of 2 GiB
> =====
> Mem: 16M Active, 16M Inact, 2523M Wired, 536M Free
> ARC: 2067M Total, 3255K MFU, 1821M MRU, 35K Anon, 38M Header, 204M Other
> =====
> 2523-2067=456 MiB
>
> Memory usage a few minutes later
> =====
> Mem: 10M Active, 27M Inact, 2721M Wired, 333M Free
> ARC: 2002M Total, 22M MFU, 1655M MRU, 21K Anon, 36M Header, 289M Other
> =====
> 2721-2002=719 MiB
>
> So why the wired ram on a machine with only minimal amount of services
> has grown from 92 to 719 MiB? Sometimes I can even see about a gig!
> I'm using 9.2-RELEASE-p1 amd64. Test machine has a T5450 C2D CPU and 4
> G RAM (actual available amount is 3 G). ZFS pool is configured on a
> GPT partition of a single 1 TB HDD.
> Disabling/enabling prefetch does't helps. Limiting ARC to 1 gig doesn't helps.
> When reading a pool, evict skips can increment very fast and sometimes
> arc metadata exceeds limit (2x-5x).
>
> I've attached logs with system configuration, outputs from top, ps,
> zfs-stats and vmstat.
> conf.log = system configuration, also uploaded to http://pastebin.com/NYBcJPeT
> top_ps_zfs-stats_vmstat_afterboot = memory stats immediately after
> booting system, http://pastebin.com/mudmEyG5
> top_ps_zfs-stats_vmstat_1g-arc = after ARC grown to 1 gig,
> http://pastebin.com/4AC8dn5C
> top_ps_zfs-stats_vmstat_fullmem = when ARC reached limit of 2 gigs,
> http://pastebin.com/bx7svEP0
> top_ps_zfs-stats_vmstat_fullmem_2 = few minutes later,
> http://pastebin.com/qYWFaNeA
>
> What should I do next?
BUMP



Want to link to this message? Use this URL: <https://mail-archive.FreeBSD.org/cgi/mid.cgi?CAFG2KCJUWtLwR_j2Ykr1J%2BO6PESgs3RdztS_Yx0gNJ_7UmrGJw>