Skip site navigation (1)Skip section navigation (2)
Date:      Tue, 6 Nov 2012 20:17:20 -1000 (HST)
From:      Jeff Roberson <jroberson@jroberson.net>
To:        David Xu <davidxu@freebsd.org>
Cc:        attilio@freebsd.org, freebsd-hackers <freebsd-hackers@freebsd.org>, Jeff Roberson <jeff@freebsd.org>, Andriy Gapon <avg@freebsd.org>
Subject:   Re: ule+smp: small optimization for turnstile priority lending
Message-ID:  <alpine.BSF.2.00.1211062016380.4081@desktop>
In-Reply-To: <5099C5C9.4040703@freebsd.org>
References:  <50587F8D.9060102@FreeBSD.org> <505AD2A5.6060008@freebsd.org> <CAJ-FndBGEgkrKJ9bNdq0QrdyYZb=LXUsAG3wz5Lp-HLUBd5d9w@mail.gmail.com> <5099C5C9.4040703@freebsd.org>

next in thread | previous in thread | raw e-mail | index | archive | help
On Wed, 7 Nov 2012, David Xu wrote:

> On 2012/11/06 19:03, Attilio Rao wrote:
>> On 9/20/12, David Xu <davidxu@freebsd.org> wrote:
>>> On 2012/09/18 22:05, Andriy Gapon wrote:
>>>> 
>>>> Here is a snippet that demonstrates the issue on a supposedly fully
>>>> loaded
>>>> 2-processor system:
>>>> 
>>>> 136794   0 3670427870244462 KTRGRAPH group:"thread", id:"Xorg tid
>>>> 102818",
>>>> state:"running", attributes: prio:122
>>>> 
>>>> 136793   0 3670427870241000 KTRGRAPH group:"thread", id:"cc1plus tid
>>>> 111916",
>>>> state:"yielding", attributes: prio:183, wmesg:"(null)", lockname:"(null)"
>>>> 
>>>> 136792   1 3670427870240829 KTRGRAPH group:"thread", id:"idle: cpu1 tid
>>>> 100004",
>>>> state:"running", attributes: prio:255
>>>> 
>>>> 136791   1 3670427870239520 KTRGRAPH group:"load", id:"CPU 1 load",
>>>> counter:0,
>>>> attributes: none
>>>> 
>>>> 136790   1 3670427870239248 KTRGRAPH group:"thread", id:"firefox tid
>>>> 113473",
>>>> state:"blocked", attributes: prio:122, wmesg:"(null)", lockname:"unp_mtx"
>>>> 
>>>> 136789   1 3670427870237697 KTRGRAPH group:"load", id:"CPU 0 load",
>>>> counter:2,
>>>> attributes: none
>>>> 
>>>> 136788   1 3670427870236394 KTRGRAPH group:"thread", id:"firefox tid
>>>> 113473",
>>>> point:"wokeup", attributes: linkedto:"Xorg tid 102818"
>>>> 
>>>> 136787   1 3670427870236145 KTRGRAPH group:"thread", id:"Xorg tid
>>>> 102818",
>>>> state:"runq add", attributes: prio:122, linkedto:"firefox tid 113473"
>>>> 
>>>> 136786   1 3670427870235981 KTRGRAPH group:"load", id:"CPU 1 load",
>>>> counter:1,
>>>> attributes: none
>>>> 
>>>> 136785   1 3670427870235707 KTRGRAPH group:"thread", id:"Xorg tid
>>>> 102818",
>>>> state:"runq rem", attributes: prio:176
>>>> 
>>>> 136784   1 3670427870235423 KTRGRAPH group:"thread", id:"Xorg tid
>>>> 102818",
>>>> point:"prio", attributes: prio:176, new prio:122, linkedto:"firefox tid
>>>> 113473"
>>>> 
>>>> 136783   1 3670427870202392 KTRGRAPH group:"thread", id:"firefox tid
>>>> 113473",
>>>> state:"running", attributes: prio:104
>>>> 
>>>> See how how the Xorg thread was forced from CPU 1 to CPU 0 where it
>>>> preempted
>>>> cc1plus thread (I do have preemption enabled) only to leave CPU 1 with
>>>> zero load.
>>>> 
>>>> Here is a proposed solution:
>>>>
>>>>       turnstile_wait: optimize priority lending to a thread on a runqueue
>>>>
>>>>       As the current thread is definitely going into mi_switch, it now
>>>> removes
>>>>       its load before doing priority propagation which can potentially
>>>> result
>>>>       in sched_add.  In the SMP && ULE case the latter searches for the
>>>>       least loaded CPU to place a boosted thread, which is supposedly
>>>> about
>>>>       to run.
>>>> 
>>>> diff --git a/sys/kern/sched_ule.c b/sys/kern/sched_ule.c
>>>> index 8e466cd..3299cae 100644
>>>> --- a/sys/kern/sched_ule.c
>>>> +++ b/sys/kern/sched_ule.c
>>>> @@ -1878,7 +1878,10 @@ sched_switch(struct thread *td, struct thread
>>>> *newtd, int
>>>> flags)
>>>>    		/* This thread must be going to sleep. */
>>>>    		TDQ_LOCK(tdq);
>>>>    		mtx = thread_lock_block(td);
>>>> -		tdq_load_rem(tdq, td);
>>>> +#if defined(SMP)
>>>> +		if ((flags & SW_TYPE_MASK) != SWT_TURNSTILE)
>>>> +#endif
>>>> +			tdq_load_rem(tdq, td);
>>>>    	}
>>>>    	/*
>>>>    	 * We enter here with the thread blocked and assigned to the
>>>> @@ -2412,6 +2415,21 @@ sched_rem(struct thread *td)
>>>>    		tdq_setlowpri(tdq, NULL);
>>>>    }
>>>> 
>>>> +void
>>>> +sched_load_rem(struct thread *td)
>>>> +{
>>>> +	struct tdq *tdq;
>>>> +
>>>> +	KASSERT(td == curthread,
>>>> +	    ("sched_rem_load: only curthread is supported"));
>>>> +	KASSERT(td->td_oncpu == td->td_sched->ts_cpu,
>>>> +	    ("thread running on cpu different from ts_cpu"));
>>>> +	tdq = TDQ_CPU(td->td_sched->ts_cpu);
>>>> +	TDQ_LOCK_ASSERT(tdq, MA_OWNED);
>>>> +	MPASS(td->td_lock == TDQ_LOCKPTR(tdq));
>>>> +	tdq_load_rem(tdq, td);
>>>> +}
>>>> +
>>>>    /*
>>>>     * Fetch cpu utilization information.  Updates on demand.
>>>>     */
>>>> diff --git a/sys/kern/subr_turnstile.c b/sys/kern/subr_turnstile.c
>>>> index 31d16fe..d1d68e9 100644
>>>> --- a/sys/kern/subr_turnstile.c
>>>> +++ b/sys/kern/subr_turnstile.c
>>>> @@ -731,6 +731,13 @@ turnstile_wait(struct turnstile *ts, struct thread
>>>> *owner,
>>>> int queue)
>>>>    		LIST_INSERT_HEAD(&ts->ts_free, td->td_turnstile, ts_hash);
>>>>    	}
>>>>    	thread_lock(td);
>>>> +#if defined(SCHED_ULE) && defined(SMP)
>>>> +	/*
>>>> +	 * Remove load earlier so that it does not affect cpu selection
>>>> +	 * for a thread waken up due to priority lending, if any.
>>>> +	 */
>>>> +	sched_load_rem(td);
>>>> +#endif
>>>>    	thread_lock_set(td, &ts->ts_lock);
>>>>    	td->td_turnstile = NULL;
>>>> 
>>>> diff --git a/sys/sys/sched.h b/sys/sys/sched.h
>>>> index 4b8387c..b1ead1b 100644
>>>> --- a/sys/sys/sched.h
>>>> +++ b/sys/sys/sched.h
>>>> @@ -110,6 +110,9 @@ void	sched_preempt(struct thread *td);
>>>>    void	sched_add(struct thread *td, int flags);
>>>>    void	sched_clock(struct thread *td);
>>>>    void	sched_rem(struct thread *td);
>>>> +#if defined(SCHED_ULE) && defined(SMP)
>>>> +void	sched_load_rem(struct thread *td);
>>>> +#endif
>>>>    void	sched_tick(int cnt);
>>>>    void	sched_relinquish(struct thread *td);
>>>>    struct thread *sched_choose(void);
>>>> 
>>> 
>>> I found another scenario in taskqueue, in the function
>>> taskqueue_terminate, current thread tries to wake
>>> another thread up and sleep immediately, the tq_mutex sometimes
>>> is a spinlock. So if you remove one thread load from current cpu
>>> before wakeup, the resumed thread may be put on same cpu,
>>> so it will optimize the cpu scheduling too.
>> 
>> I think that in order to fit with sched_add() modifies I have in mind
>> (see other patches within this thread) wakeup() should grow a new
>> argument, or maybe we can use wakeup_flags() new KPI.
>> If the latter is the case, I would also propose to let wakeup_one() to
>> be absorbed into wakeup_flags() with its own flag.
>> 
>
> Yes, I like the idea.

>From ~2007

http://people.freebsd.org/~jeff/wakeupflags.diff

This has some different optimizations that can be done when you have a 
hinted wakeup.

Thanks,
Jeff

>
>> Attilio
>> 
>> 
>



Want to link to this message? Use this URL: <https://mail-archive.FreeBSD.org/cgi/mid.cgi?alpine.BSF.2.00.1211062016380.4081>