Date: Thu, 5 Aug 2010 13:12:25 -0400 From: John Baldwin <jhb@freebsd.org> To: mdf@freebsd.org Cc: freebsd-hackers@freebsd.org Subject: Re: sched_pin() versus PCPU_GET Message-ID: <201008051312.25854.jhb@freebsd.org> In-Reply-To: <AANLkTikvx9c=CjMcE7WsAZrxAxfqcDQEYOa0rWRBBXA5@mail.gmail.com> References: <AANLkTikY20TxyeyqO5zP3zC-azb748kV-MdevPfm%2B8cq@mail.gmail.com> <201008041455.26066.jhb@freebsd.org> <AANLkTikvx9c=CjMcE7WsAZrxAxfqcDQEYOa0rWRBBXA5@mail.gmail.com>
next in thread | previous in thread | raw e-mail | index | archive | help
On Thursday, August 05, 2010 11:59:37 am mdf@freebsd.org wrote: > On Wed, Aug 4, 2010 at 11:55 AM, John Baldwin <jhb@freebsd.org> wrote: > > On Wednesday, August 04, 2010 12:20:31 pm mdf@freebsd.org wrote: > >> On Wed, Aug 4, 2010 at 2:26 PM, John Baldwin <jhb@freebsd.org> wrote: > >> > On Tuesday, August 03, 2010 9:46:16 pm mdf@freebsd.org wrote: > >> >> On Fri, Jul 30, 2010 at 2:31 PM, John Baldwin <jhb@freebsd.org> wro= te: > >> >> > On Friday, July 30, 2010 10:08:22 am John Baldwin wrote: > >> >> >> On Thursday, July 29, 2010 7:39:02 pm mdf@freebsd.org wrote: > >> >> >> > We've seen a few instances at work where witness_warn() in ast= () > >> >> >> > indicates the sched lock is still held, but the place it claim= s=20 it was > >> >> >> > held by is in fact sometimes not possible to keep the lock, li= ke: > >> >> >> > > >> >> >> > thread_lock(td); > >> >> >> > td->td_flags &=3D ~TDF_SELECT; > >> >> >> > thread_unlock(td); > >> >> >> > > >> >> >> > What I was wondering is, even though the assembly I see in=20 objdump -S > >> >> >> > for witness_warn has the increment of td_pinned before the=20 PCPU_GET: > >> >> >> > > >> >> >> > ffffffff802db210: 65 48 8b 1c 25 00 00 mov %gs:0x0,%rbx > >> >> >> > ffffffff802db217: 00 00 > >> >> >> > ffffffff802db219: ff 83 04 01 00 00 incl 0x104(%rbx) > >> >> >> > * Pin the thread in order to avoid problems with thread=20 migration. > >> >> >> > * Once that all verifies are passed about spinlocks=20 ownership, > >> >> >> > * the thread is in a safe path and it can be unpinned. > >> >> >> > */ > >> >> >> > sched_pin(); > >> >> >> > lock_list =3D PCPU_GET(spinlocks); > >> >> >> > ffffffff802db21f: 65 48 8b 04 25 48 00 mov %gs:0x48,%r= ax > >> >> >> > ffffffff802db226: 00 00 > >> >> >> > if (lock_list !=3D NULL && lock_list->ll_count !=3D 0) { > >> >> >> > ffffffff802db228: 48 85 c0 test %rax,%rax > >> >> >> > * Pin the thread in order to avoid problems with thread=20 migration. > >> >> >> > * Once that all verifies are passed about spinlocks=20 ownership, > >> >> >> > * the thread is in a safe path and it can be unpinned. > >> >> >> > */ > >> >> >> > sched_pin(); > >> >> >> > lock_list =3D PCPU_GET(spinlocks); > >> >> >> > ffffffff802db22b: 48 89 85 f0 fe ff ff mov =20 %rax,-0x110(%rbp) > >> >> >> > ffffffff802db232: 48 89 85 f8 fe ff ff mov =20 %rax,-0x108(%rbp) > >> >> >> > if (lock_list !=3D NULL && lock_list->ll_count !=3D 0) { > >> >> >> > ffffffff802db239: 0f 84 ff 00 00 00 je =20 ffffffff802db33e > >> >> >> > <witness_warn+0x30e> > >> >> >> > ffffffff802db23f: 44 8b 60 50 mov 0x50(%rax), %r12d > >> >> >> > > >> >> >> > is it possible for the hardware to do any re-ordering here? > >> >> >> > > >> >> >> > The reason I'm suspicious is not just that the code doesn't ha= ve=20 a > >> >> >> > lock leak at the indicated point, but in one instance I can se= e=20 in the > >> >> >> > dump that the lock_list local from witness_warn is from the pc= pu > >> >> >> > structure for CPU 0 (and I was warned about sched lock 0), but= =20 the > >> >> >> > thread id in panic_cpu is 2. So clearly the thread was being= =20 migrated > >> >> >> > right around panic time. > >> >> >> > > >> >> >> > This is the amd64 kernel on stable/7. I'm not sure exactly wh= at=20 kind > >> >> >> > of hardware; it's a 4-way Intel chip from about 3 or 4 years a= go=20 IIRC. > >> >> >> > > >> >> >> > So... do we need some kind of barrier in the code for sched_pi= n()=20 for > >> >> >> > it to really do what it claims? Could the hardware have re- ordered > >> >> >> > the "mov %gs:0x48,%rax" PCPU_GET to before the sched_pin() > >> >> >> > increment? > >> >> >> > >> >> >> Hmmm, I think it might be able to because they refer to differen= t=20 locations. > >> >> >> > >> >> >> Note this rule in section 8.2.2 of Volume 3A: > >> >> >> > >> >> >> =95 Reads may be reordered with older writes to different loca= tions=20 but not > >> >> >> with older writes to the same location. > >> >> >> > >> >> >> It is certainly true that sparc64 could reorder with RMO. I=20 believe ia64 > >> >> >> could reorder as well. Since sched_pin/unpin are frequently use= d=20 to provide > >> >> >> this sort of synchronization, we could use memory barriers in=20 pin/unpin > >> >> >> like so: > >> >> >> > >> >> >> sched_pin() > >> >> >> { > >> >> >> td->td_pinned =3D atomic_load_acq_int(&td->td_pinned) + 1; > >> >> >> } > >> >> >> > >> >> >> sched_unpin() > >> >> >> { > >> >> >> atomic_store_rel_int(&td->td_pinned, td->td_pinned - 1); > >> >> >> } > >> >> >> > >> >> >> We could also just use atomic_add_acq_int() and=20 atomic_sub_rel_int(), but they > >> >> >> are slightly more heavyweight, though it would be more clear wha= t=20 is happening > >> >> >> I think. > >> >> > > >> >> > However, to actually get a race you'd have to have an interrupt f= ire=20 and > >> >> > migrate you so that the speculative read was from the other CPU.= =20 However, I > >> >> > don't think the speculative read would be preserved in that case.= =20 The CPU > >> >> > has to return to a specific PC when it returns from the interrupt= =20 and it has > >> >> > no way of storing the state for what speculative reordering it mi= ght=20 be > >> >> > doing, so presumably it is thrown away? I suppose it is possible= =20 that it > >> >> > actually retires both instructions (but reordered) and then retur= ns=20 to the PC > >> >> > value after the read of listlocks after the interrupt. However, = in=20 that case > >> >> > the scheduler would not migrate as it would see td_pinned !=3D 0.= To=20 get the > >> >> > race you have to have the interrupt take effect prior to modifyin= g=20 td_pinned, > >> >> > so I think the processor would have to discard the reordered read= of > >> >> > listlocks so it could safely resume execution at the 'incl'=20 instruction. > >> >> > > >> >> > The other nit there on x86 at least is that the incl instruction = is=20 doing > >> >> > both a read and a write and another rule in the section 8.2.2 is= =20 this: > >> >> > > >> >> > =95 Reads are not reordered with other reads. > >> >> > > >> >> > That would seem to prevent the read of listlocks from passing the= =20 read of > >> >> > td_pinned in the incl instruction on x86. > >> >> > >> >> I wonder how that's interpreted in the microcode, though? I.e. if = the > >> >> incr instruction decodes to load, add, store, does the h/w allow the > >> >> later reads to pass the final store? > >> > > >> > Well, the architecture is defined in terms of the ISA, not the=20 microcode, per > >> > se, so I think it would have to treat the read for the incl as being= an=20 earlier > >> > read than 'spinlocks'. > >> > > >> >> I added the following: > >> >> > >> >> sched_pin(); > >> >> lock_list =3D PCPU_GET(spinlocks); > >> >> if (lock_list !=3D NULL && lock_list->ll_count !=3D 0) { > >> >> + /* XXX debug for bug 67957 */ > >> >> + mfence(); > >> >> + lle =3D PCPU_GET(spinlocks); > >> >> + if (lle !=3D lock_list) { > >> >> + panic("Bug 67957: had lock list %p, now %p\n", > >> >> + lock_list, lle); > >> >> + } > >> >> + /* XXX end debug */ > >> >> sched_unpin(); > >> >> > >> >> /* > >> >> > >> >> ... and the panic triggered. I think it's more likely that some > >> >> barrier is needed in sched_pin() than that %gs is getting corrupted > >> >> but can always be dereferenced. > >> > > >> > Actually, I would beg to differ in that case. If PCPU_GET(spinlocks) > >> > returns non-NULL, then it means that you hold a spin lock, > >> > >> ll_count is 0 for the "correct" pc_spinlocks and non-zero for the > >> "wrong" one, though. So I think it can be non-NULL but the current > >> thread/CPU doesn't hold a spinlock. > > > > Hmm, does the 'lock_list' pointer value in the dump match 'lock_list' > > from another CPU? >=20 > Yes: >=20 > (gdb) p panic_cpu > $9 =3D 2 > (gdb) p dumptid > $12 =3D 100751 > (gdb) p cpuhead.slh_first->pc_allcpu.sle_next->pc_curthread->td_tid > $14 =3D 100751 >=20 > (gdb) p *cpuhead.slh_first->pc_allcpu.sle_next > $6 =3D { > pc_curthread =3D 0xffffff00716d6960, > pc_cpuid =3D 2, > pc_spinlocks =3D 0xffffffff80803198, >=20 > (gdb) p lock_list > $2 =3D (struct lock_list_entry *) 0xffffffff80803fb0 >=20 > (gdb) p *cpuhead.slh_first->pc_allcpu.sle_next->pc_allcpu.sle_next- >pc_allcpu.sle_next > $8 =3D { > pc_curthread =3D 0xffffff0005479960, > pc_cpuid =3D 0, > pc_spinlocks =3D 0xffffffff80803fb0, >=20 > I.e. we're dumping on CPU 2, but the lock_list pointer that was saved > in the dump matches that of CPU 0. Can you print out the tid's for the two curthreads? It's not impossible th= at=20 the thread migrated after calling panic. In fact we force threads to CPU 0= =20 during shutdown. =2D-=20 John Baldwin
Want to link to this message? Use this URL: <https://mail-archive.FreeBSD.org/cgi/mid.cgi?201008051312.25854.jhb>