On Fri, Mar 08, 2019 at 03:54:29PM +0000, Mark Rutland wrote:
> Hi Alex,
> 
> On Thu, Feb 28, 2019 at 04:01:09PM +0200, Alexander Shishkin wrote:
> > First, we have a race between perf_event_release_kernel() and
> > perf_free_event(), which happens when parent's event is released while the
> > child's fork fails (because of a fatal signal, for example), that looks
> > like this:
> > 
> > cpu X                            cpu Y
> > -----                            -----
> >                                  copy_process() error path
> > perf_release(parent)             +->perf_event_free_task()
> > +-> lock(child_ctx->mutex)       |  |
> > +-> remove_from_context(child)   |  |
> > +-> unlock(child_ctx->mutex)     |  |
> > |                                |  +-> lock(child_ctx->mutex)
> > |                                |  +-> unlock(child_ctx->mutex)
> > |                                +-> free_task(child_task)
> > +-> put_task_struct(child_task)

I had a wee bit of bother following that, it's this, right?

        close()                                         clone()

                                                          copy_process()
                                                            
perf_event_init_task()
                                                              
perf_event_init_context()
                                                                
mutex_lock(parent_ctx->mutex)
                                                                
inherit_task_group()
                                                                  
inherit_group()
                                                                    
inherit_event()
                                                                      
mutex_lock(event->child_mutex)
                                                                      // expose 
event on child list
                                                                      
list_add_tail()
                                                                      
mutex_unlock(event->child_mutex)
                                                                
mutex_unlock(parent_ctx->mutex)

                                                            ...
                                                            goto bad_fork_*

                                                          bad_fork_cleanup_perf:
                                                            
perf_event_free_task()

          perf_release()
            perf_event_release_kernel()
              list_for_each_entry()
                mutex_lock(ctx->mutex)
                mutex_lock(event->child_mutex)
                // event is from the failing inherit
                // on the other CPU
                perf_remove_from_context()
                list_move()
                mutex_unlock(event->child_mutex)
                mutex_unlock(ctx->mutex)

                                                              
mutex_lock(ctx->mutex)
                                                              
list_for_each_entry_safe()
                                                                // event 
already stolen
                                                              
mutex_unlock(ctx->mutex)

                                                            delayed_free_task()
                                                              free_task()

             list_for_each_entry_safe()
               list_del()
               free_event()
                 _free_event()
                   // and so event->hw.target
                   // is the already freed failed clone()
                   if (event->hw.target)
                     put_task_struct(event->hw.target)
                       // WHOOPSIE, already quite dead


Which puts the lie to the the comment on perf_event_free_task();
'unexposed, unused context' my ass.

Which is a 'fun' confluence of fail; copy_process() doing an
unconditional free_task() and not respecting refcounts, and perf having
creative locking. In particular:

  82d94856fa22 ("perf/core: Fix lock inversion between perf,trace,cpuhp")

seems to have overlooked this 'fun' parade.

> > Technically, we're still holding a reference to the task via
> > parent->hw.target, that's not stopping free_task(), so we end up poking at
> > free'd memory, as is pointed out by KASAN in the syzkaller report (see Link
> > below). The straightforward fix is to drop the hw.target reference while
> > the task is still around.

Right.

> > Therein lies the second problem: the users of hw.target (uprobe) assume
> > that it's around at ->destroy() callback time, where they use it for
> > context. So, in order to not break the uprobe teardown and avoid leaking
> > stuff, we need to call ->destroy() at the same time.
> 
> I had not spotted that case. That's rather horrid. :/

Such joy..

> > diff --git a/kernel/events/core.c b/kernel/events/core.c
> > index 36b8320590e8..640695d114f8 100644
> > --- a/kernel/events/core.c
> > +++ b/kernel/events/core.c
> > @@ -2105,6 +2105,27 @@ static void perf_remove_from_context(struct 
> > perf_event *event, unsigned long fla
> >  
> >     event_function_call(event, __perf_remove_from_context, (void *)flags);
> >  
> > +   /*
> > +    * This is as passable as any hw.target handling out there;
> > +    * hw.target implies task context, therefore, no migration.
> > +    * Which means that we can only get here at the teardown.
> > +    */
> > +   if (event->hw.target) {
> > +           /*
> > +            * Now, the problem with, say uprobes, is that they
> > +            * use hw.target for context in their ->destroy()
> > +            * callbacks. Supposedly, they may need to poke at
> > +            * its contents, so better call it while we still
> > +            * have the task.
> > +            */
> > +           if (event->destroy) {
> > +                   event->destroy(event);
> > +                   event->destroy = NULL;
> > +           }
> > +           put_task_struct(event->hw.target);
> > +           event->hw.target = NULL;
> > +   }
> 
> We also use perf_remove_from_context() in perf_event_open() when we move
> events from a SW context to a HW context, so we can't destroy the event
> here.

Also perf_pmu_migrate_context(), and yes, we must not call ->destroy()
from remove_context, or rather, not unconditional. We could make it
conditional on DETACH_GROUP or better add another DETACH_ flags.

> I think we need something more like the below (untested), but I fear
> that it's not safe to call perf_event::destroy() in this context.

> ---->8----
> diff --git a/kernel/events/core.c b/kernel/events/core.c
> index 26d6edab051a..b32f2cac5563 100644
> --- a/kernel/events/core.c
> +++ b/kernel/events/core.c
> @@ -4532,6 +4532,24 @@ static void put_event(struct perf_event *event)
>         _free_event(event);
>  }
>  
> +void perf_event_detach_target(struct perf_event *event)
> +{
> +       if (!event->hw.target)
> +               return;
> +
> +       /*
> +        * The uprobes perf_event::destroy() callback needs the target, so 
> call
> +        * that while the target is still valid.
> +        */
> +       if (event->destroy) {
> +               event->destroy(event);
> +               event->destroy = NULL;
> +       }
> +
> +       put_task_struct(event->hw.target);
> +       event->hw.target = NULL;
> +}
> +
>  /*
>   * Kill an event dead; while event:refcount will preserve the event
>   * object, it will not preserve its functionality. Once the last 'user'
> @@ -4559,6 +4577,7 @@ int perf_event_release_kernel(struct perf_event *event)
>         ctx = perf_event_ctx_lock(event);
>         WARN_ON_ONCE(ctx->parent_ctx);
>         perf_remove_from_context(event, DETACH_GROUP);
> +       perf_event_detach_target(event);
>  
>         raw_spin_lock_irq(&ctx->lock);
>         /*
> @@ -4614,6 +4633,7 @@ int perf_event_release_kernel(struct perf_event *event)
>                                                struct perf_event, child_list);
>                 if (tmp == child) {
>                         perf_remove_from_context(child, DETACH_GROUP);
> +                       perf_event_detach_target(child);
>                         list_move(&child->child_list, &free_list);
>                         /*
>                          * This matches the refcount bump in inherit_event();

And doesn't this re-introduce the problem we fixed with 82d94856fa22 ?

By doing ->destroy() while holding ctx->mutex, we re-establish that
#5->#0 link and close the cycle again.

More thinking required...

Reply via email to