Date: Fri, 07 Mar 2014 13:43:19 +0100 From: Julien Charbon <jcharbon@verisign.com> To: freebsd-net@freebsd.org Cc: jmg@funkthat.com Subject: Re: TCP stack lock contention with short-lived connections Message-ID: <5319BEE7.607@verisign.com> In-Reply-To: <20140306215743.GB47921@funkthat.com> References: <op.w51mxed6ak5tgc@fri2jcharbon-m1.local> <op.w56mamc0ak5tgc@dul1rjacobso-l3.vcorp.ad.vrsn.com> <len481$sfv$2@ger.gmane.org> <20140306215743.GB47921@funkthat.com>
next in thread | previous in thread | raw e-mail | index | archive | help
Hi John, On 06/03/14 22:57, John-Mark Gurney wrote: > Julien Charbon wrote this message on Thu, Feb 27, 2014 at 11:32 +0100: >> [...] >> Any thoughts on this particular behavior? > > One thing that I noticed is that you now lock/unlock the tw and inp lock a > lot... Have you thought about grabing the TW lock once, grabbing some/all > of the ones necessary to process and then process them in a second step? > If the bulk processing is still an issue, then you could process them in > blocks of 50 or so, that way the number of lock/unlock cycles is reduced... First thanks, feedback are highly valuable to us. In first place, I indeed tried a kind of bulk processing enforcement with something like: tcp_tw_2msl_scan() { struct tcptw *tw; int i, end = 0, count = 100; for (;;) { INP_INFO_WLOCK(&V_tcbinfo); for (i = 0; i < count; ++i) { tw = TAILQ_FIRST(&V_twq_2msl); if (tw == NULL || (tw->tw_time - ticks) > 0)) { end = 1; break; } INP_WLOCK(tw->tw_inpcb); tcp_twclose(tw, 0); } if (end) break; INP_INFO_WUNLOCK(&V_tcbinfo); } return (NULL); } And I got best result with 'count' set to 1, this led us to current proposed patch. Thus main goal here is somehow to prioritize the NIC interruption handler calls against tcp_tw_2msl_scan() call in INP_INFO battle. As you proposed, we can add a sysctl(or a #define) to configure the maximum of tw objects to be cleaned up under a same INP_INFO_WLOCK() call, to have a finer control on how the tw objects are enforced. That said, our high level solution is to consider the NIC interruption code path (i.e. tcp_input()) as critical and remove almost most contention points on it, which is our long term goal. This change is just a step on this (long and not straightforward) path. >> +/* >> + * Drop a refcount on an tw elevated using tw_pcbref(). If it is >> + * valid, we return with the tw lock held. >> + */ > > I assume you mean that you return with the tw lock unlocked? at least > that's what the code reads to me... > >> +static int >> +tw_pcbrele(struct tcptw *tw) >> +{ >> + TW_WLOCK_ASSERT(V_tw_lock); >> + KASSERT(tw->tw_refcount > 0, ("%s: refcount 0", __func__)); >> + >> + if (!refcount_release(&tw->tw_refcount)) { >> + TW_WUNLOCK(V_tw_lock); >> + return (0); >> + } >> + >> + uma_zfree(V_tcptw_zone, tw); >> + TW_WUNLOCK(V_tw_lock); >> + return (1); >> +} You are completely right, my mistake. I will update the comment. > Otherwise looks like a good patch... Thanks again for your time. -- Julien
Want to link to this message? Use this URL: <https://mail-archive.FreeBSD.org/cgi/mid.cgi?5319BEE7.607>