On Fri, Mar 08, 2019 at 03:54:29PM +0000, Mark Rutland wrote: > Hi Alex, > > On Thu, Feb 28, 2019 at 04:01:09PM +0200, Alexander Shishkin wrote: > > First, we have a race between perf_event_release_kernel() and > > perf_free_event(), which happens when parent's event is released while the > > child's fork fails (because of a fatal signal, for example), that looks > > like this: > > > > cpu X cpu Y > > ----- ----- > > copy_process() error path > > perf_release(parent) +->perf_event_free_task() > > +-> lock(child_ctx->mutex) | | > > +-> remove_from_context(child) | | > > +-> unlock(child_ctx->mutex) | | > > | | +-> lock(child_ctx->mutex) > > | | +-> unlock(child_ctx->mutex) > > | +-> free_task(child_task) > > +-> put_task_struct(child_task)
I had a wee bit of bother following that, it's this, right? close() clone() copy_process() perf_event_init_task() perf_event_init_context() mutex_lock(parent_ctx->mutex) inherit_task_group() inherit_group() inherit_event() mutex_lock(event->child_mutex) // expose event on child list list_add_tail() mutex_unlock(event->child_mutex) mutex_unlock(parent_ctx->mutex) ... goto bad_fork_* bad_fork_cleanup_perf: perf_event_free_task() perf_release() perf_event_release_kernel() list_for_each_entry() mutex_lock(ctx->mutex) mutex_lock(event->child_mutex) // event is from the failing inherit // on the other CPU perf_remove_from_context() list_move() mutex_unlock(event->child_mutex) mutex_unlock(ctx->mutex) mutex_lock(ctx->mutex) list_for_each_entry_safe() // event already stolen mutex_unlock(ctx->mutex) delayed_free_task() free_task() list_for_each_entry_safe() list_del() free_event() _free_event() // and so event->hw.target // is the already freed failed clone() if (event->hw.target) put_task_struct(event->hw.target) // WHOOPSIE, already quite dead Which puts the lie to the the comment on perf_event_free_task(); 'unexposed, unused context' my ass. Which is a 'fun' confluence of fail; copy_process() doing an unconditional free_task() and not respecting refcounts, and perf having creative locking. In particular: 82d94856fa22 ("perf/core: Fix lock inversion between perf,trace,cpuhp") seems to have overlooked this 'fun' parade. > > Technically, we're still holding a reference to the task via > > parent->hw.target, that's not stopping free_task(), so we end up poking at > > free'd memory, as is pointed out by KASAN in the syzkaller report (see Link > > below). The straightforward fix is to drop the hw.target reference while > > the task is still around. Right. > > Therein lies the second problem: the users of hw.target (uprobe) assume > > that it's around at ->destroy() callback time, where they use it for > > context. So, in order to not break the uprobe teardown and avoid leaking > > stuff, we need to call ->destroy() at the same time. > > I had not spotted that case. That's rather horrid. :/ Such joy.. > > diff --git a/kernel/events/core.c b/kernel/events/core.c > > index 36b8320590e8..640695d114f8 100644 > > --- a/kernel/events/core.c > > +++ b/kernel/events/core.c > > @@ -2105,6 +2105,27 @@ static void perf_remove_from_context(struct > > perf_event *event, unsigned long fla > > > > event_function_call(event, __perf_remove_from_context, (void *)flags); > > > > + /* > > + * This is as passable as any hw.target handling out there; > > + * hw.target implies task context, therefore, no migration. > > + * Which means that we can only get here at the teardown. > > + */ > > + if (event->hw.target) { > > + /* > > + * Now, the problem with, say uprobes, is that they > > + * use hw.target for context in their ->destroy() > > + * callbacks. Supposedly, they may need to poke at > > + * its contents, so better call it while we still > > + * have the task. > > + */ > > + if (event->destroy) { > > + event->destroy(event); > > + event->destroy = NULL; > > + } > > + put_task_struct(event->hw.target); > > + event->hw.target = NULL; > > + } > > We also use perf_remove_from_context() in perf_event_open() when we move > events from a SW context to a HW context, so we can't destroy the event > here. Also perf_pmu_migrate_context(), and yes, we must not call ->destroy() from remove_context, or rather, not unconditional. We could make it conditional on DETACH_GROUP or better add another DETACH_ flags. > I think we need something more like the below (untested), but I fear > that it's not safe to call perf_event::destroy() in this context. > ---->8---- > diff --git a/kernel/events/core.c b/kernel/events/core.c > index 26d6edab051a..b32f2cac5563 100644 > --- a/kernel/events/core.c > +++ b/kernel/events/core.c > @@ -4532,6 +4532,24 @@ static void put_event(struct perf_event *event) > _free_event(event); > } > > +void perf_event_detach_target(struct perf_event *event) > +{ > + if (!event->hw.target) > + return; > + > + /* > + * The uprobes perf_event::destroy() callback needs the target, so > call > + * that while the target is still valid. > + */ > + if (event->destroy) { > + event->destroy(event); > + event->destroy = NULL; > + } > + > + put_task_struct(event->hw.target); > + event->hw.target = NULL; > +} > + > /* > * Kill an event dead; while event:refcount will preserve the event > * object, it will not preserve its functionality. Once the last 'user' > @@ -4559,6 +4577,7 @@ int perf_event_release_kernel(struct perf_event *event) > ctx = perf_event_ctx_lock(event); > WARN_ON_ONCE(ctx->parent_ctx); > perf_remove_from_context(event, DETACH_GROUP); > + perf_event_detach_target(event); > > raw_spin_lock_irq(&ctx->lock); > /* > @@ -4614,6 +4633,7 @@ int perf_event_release_kernel(struct perf_event *event) > struct perf_event, child_list); > if (tmp == child) { > perf_remove_from_context(child, DETACH_GROUP); > + perf_event_detach_target(child); > list_move(&child->child_list, &free_list); > /* > * This matches the refcount bump in inherit_event(); And doesn't this re-introduce the problem we fixed with 82d94856fa22 ? By doing ->destroy() while holding ctx->mutex, we re-establish that #5->#0 link and close the cycle again. More thinking required...