Skip site navigation (1)Skip section navigation (2)
Date:      Wed, 25 Feb 2015 01:16:21 -0800
From:      Harrison Grundy <harrison.grundy@astrodoggroup.com>
To:        freebsd-arch@freebsd.org
Subject:   Re: locks and kernel randomness...
Message-ID:  <54ED92E5.4010803@astrodoggroup.com>
In-Reply-To: <20150225090638.GB74514@kib.kiev.ua>
References:  <DD06E2EA-68D6-43D7-AA17-FB230750E55A@bsdimp.com> <20150224174053.GG46794@funkthat.com> <54ECBD4B.6000007@freebsd.org> <20150224182507.GI46794@funkthat.com> <54ECEA43.2080008@freebsd.org> <20150224231921.GQ46794@funkthat.com> <1424822522.1328.11.camel@freebsd.org> <20150225002956.GT46794@funkthat.com> <2F49527F-2F58-4BD2-B8BE-1B1190CCD4D0@bsdimp.com> <54ED5656.50607@astrodoggroup.com> <20150225090638.GB74514@kib.kiev.ua>

next in thread | previous in thread | raw e-mail | index | archive | help


On 02/25/15 01:06, Konstantin Belousov wrote:
> On Tue, Feb 24, 2015 at 08:57:58PM -0800, Harrison Grundy wrote:
>> <... snip ...>
>>
>> The timing attack I talked to you about on IRC works like this:
>>
>> A userland process creates as many threads as there are CPUs, and by
>> manipulating the load they generate, gets it so they're all flagged as
>> interactive and at the same priority. (alternating spin and sleep with
>> a 2% duty cycle would work, for instance)
>>
>> It would also be possible to coerce a userland process, like apache to
>> behave this way.
>>
>> These threads now have the ability to preempt all timeshare tasks on
>> all CPUs for slice_size time, by waking up and spinning at the same
>> time. This means they can get very precise knowledge about scheduling,
>> by timing when they get to run, versus when they have to wait.
> Ok, this is definitely not impossible.
> 
>>
>> By watching CPU0, one of these threads can measure balance_ticks.
>>
>> This is important because balance_ticks directly exposes the last 7
>> bits it gets back from random(). (The value gets applied to
>> balance_interval to keep the balancer from running on exactly the same
>> interval)
>>
>> This means that if an attacker can trigger the use of random, or is
>> willing to wait long enough for a race, they can determine the value
>> of those bits that were passed along to anyone who called random() at
>> the same time.
>>
>> It also means that they can eventually discover the state of the RNG,
>> and predict future values.
>>
>> The security implications of disclosing the values this way isn't as
>> severe as it might seem, simply because random() isn't really used in
>> any cryptographically sensitive areas, but there are definite
>> consequences, like predicting firewall port values, and NFS client
>> transaction IDs.
>>
>> It is, however, surprising to learn that the balance_interval sysctl
>> has security implications.
> 
> So this is an argument to remove the current random() call from
> the sched_balance(). There is no implications for use of e.g.
> get_cyclecount() in the sched_balance(), since on x86 userspace has the
> ability to read the underlying counter directly.
> 
> On other architectures, where counter backing get_cyclecount() is not
> accessible to userspace, it is still feasible to use in sched_balance(),
> simply because counter is ticking.
> 
> Do you agree with these statements ?

Yes. sched_balance itself does not need any sort of non-public
randomness. The worst thing an attacker can do is gain a few extra
cycles on a CPU by only running on longer balance intervals. Given the
many other ways load gets transferred in ULE, there's not much utility
there.

> 
> Also, as I understand from your other responses, you did tested the
> patch to use get_cyclecount() on non-x86 machines ?  I try to understand
> what testing was done for the get_cyclecount() for sched_balance() patch,
> i.e. is it ready for commit.

I have not tested this on other arches. I spoke to some of the
committers active on them to get an idea of what get_cyclecount does.

I'm currently testing a patch that creates "sched_random()", using the
random number generator from cpu_search. This should provide good enough
jitter for the balancer, and other potential scheduler uses of random();

I'll add it to the PR, and send a note out here after I've run it for a bit.

--- Harrison

> _______________________________________________
> freebsd-arch@freebsd.org mailing list
> http://lists.freebsd.org/mailman/listinfo/freebsd-arch
> To unsubscribe, send any mail to "freebsd-arch-unsubscribe@freebsd.org"
> 



Want to link to this message? Use this URL: <https://mail-archive.FreeBSD.org/cgi/mid.cgi?54ED92E5.4010803>