Date: Mon, 29 Jun 2015 13:05:19 +0200 From: Milan Obuch <freebsd-pf@dino.sk> To: Daniel Hartmeier <daniel@benzedrine.ch> Cc: Ian FREISLICH <ian.freislich@capeaugusta.com>, freebsd-pf@freebsd.org Subject: Re: Large scale NAT with PF - some weird problem Message-ID: <20150629130519.168f0efc@zeta.dino.sk> In-Reply-To: <20150629104614.GD22693@insomnia.benzedrine.ch> References: <20150620182432.62797ec5@zeta.dino.sk> <20150619091857.304b707b@zeta.dino.sk> <14e119e8fa8.2755.abfb21602af57f30a7457738c46ad3ae@capeaugusta.com> <E1Z6dHz-0000uu-D8@clue.co.za> <20150621133236.75a4d86d@zeta.dino.sk> <20150629104614.GD22693@insomnia.benzedrine.ch>
next in thread | previous in thread | raw e-mail | index | archive | help
On Mon, 29 Jun 2015 12:46:14 +0200
Daniel Hartmeier <daniel@benzedrine.ch> wrote:
> On Sun, Jun 21, 2015 at 01:32:36PM +0200, Milan Obuch wrote:
>
> > One observation, on pfctl -vs info output - when src-limit counters
> > rises to 30 or so, I am getting first messages someone has problem.
> > Is it only coincidence or is there really some relation to my
> > problem?
>
> This might be a clue. That counter shouldn't increase. It means
> something triggered a PFRES_SRCLIMIT.
>
OK, I will keep an eye on this for some time too. I do not have much
knowledge regarding pf internals, so my observations may or may not be
relevant, just as my questions.
> Are you using source tracking for anything else besides the NAT sticky
> address feature?
>
I reviewed recently some pfctl output and I think this mechanism is
used in other scenarios as well, namely following one for ssh
protection:
block in quick on $if_ext inet proto tcp from <abusive_ips> to any port
22
pass in on $if_ext proto tcp to x.y.24.0/22 port ssh flags S/SA
keep state (max-src-conn 10, max-src-conn-rate 5/5, overload
<abusive_ips> flush)
(somewhat mail-mangled, but I am sure you know this one)
> If not, the only explanation for a PFRES_SRCLIMIT in a translation
> rule is a failure of pf.c pf_insert_src_node(), which could only be an
> allocation failure with uma_zalloc().
>
> Do you see any allocation failures? Log entries about uma, "source
> nodes limit reached"? How about vmstat -m?
>
Where should these failures come? I see nothing in /var/log/messages.
As for 'vmstat -m', I think following lines could be of some interest:
Type InUse MemUse HighUse Requests Size(s)
pf_hash 3 1728K - 3
pf_temp 0 0K - 955 32,64
pf_ifnet 21 7K - 282 128,256,2048
pf_osfp 1130 102K - 6780 32,128
pf_rule 222 129K - 468 128,1024
pf_table 9 18K - 35 2048
but no idea how to interpret this.
Regards,
Milan
Want to link to this message? Use this URL: <https://mail-archive.FreeBSD.org/cgi/mid.cgi?20150629130519.168f0efc>
