On Fri, Mar 11, 2016 at 09:36:24AM +0100, Jiri Olsa wrote: > On Thu, Mar 10, 2016 at 08:25:02AM +0100, Jiri Olsa wrote: > > On Thu, Mar 10, 2016 at 09:36:37AM +0900, Namhyung Kim wrote: > > > Hi Jiri, > > > > > > On Wed, Mar 09, 2016 at 09:46:41PM +0100, Jiri Olsa wrote: > > > > Currently we check sample type for ftrace:function event > > > > even if it's not created as sampling event. That prevents > > > > creating ftrace_function event in counting mode. > > > > > > > > Making sure we check sample types only for sampling events. > > > > > > > > Before: > > > > $ sudo perf stat -e ftrace:function ls > > > > ... > > > > > > > > Performance counter stats for 'ls': > > > > > > > > <not supported> ftrace:function > > > > > > > > 0.001983662 seconds time elapsed > > > > > > > > After: > > > > $ sudo perf stat -e ftrace:function ls > > > > ... > > > > > > > > Performance counter stats for 'ls': > > > > > > > > 44,498 ftrace:function > > > > > > > > 0.037534722 seconds time elapsed > > > > > > > > Signed-off-by: Jiri Olsa <jo...@kernel.org> > > > > --- > > > > kernel/trace/trace_event_perf.c | 4 ++-- > > > > 1 file changed, 2 insertions(+), 2 deletions(-) > > > > > > > > diff --git a/kernel/trace/trace_event_perf.c > > > > b/kernel/trace/trace_event_perf.c > > > > index 00df25fd86ef..a7171ec2c1ca 100644 > > > > --- a/kernel/trace/trace_event_perf.c > > > > +++ b/kernel/trace/trace_event_perf.c > > > > @@ -52,14 +52,14 @@ static int perf_trace_event_perm(struct > > > > trace_event_call *tp_event, > > > > * event, due to issues with page faults while tracing > > > > page > > > > * fault handler and its overall trickiness nature. > > > > */ > > > > - if (!p_event->attr.exclude_callchain_user) > > > > + if (is_sampling_event(p_event) && > > > > !p_event->attr.exclude_callchain_user) > > > > return -EINVAL; > > > > > > > > /* > > > > * Same reason to disable user stack dump as for user > > > > space > > > > * callchains above. > > > > */ > > > > - if (p_event->attr.sample_type & PERF_SAMPLE_STACK_USER) > > > > + if (is_sampling_event(p_event) && > > > > p_event->attr.sample_type & PERF_SAMPLE_STACK_USER) > > > > return -EINVAL; > > > > } > > > > > > > > > > What about checking is_sampling_event() first and goto the last > > > paranoid_tracepoint_raw check instead? This way we can remove the > > > same check in the function trace case. > > > > right, will check > > hum, did you mean something like this? > > I'd rather keep it the original way.. seems more straight
Hmm.. I think I was wrong. But it seems we can simply return 0 for non sampling case. How about this? Thanks, Namhyung diff --git a/kernel/trace/trace_event_perf.c b/kernel/trace/trace_event_perf.c index 00df25fd86ef..e11108f1d197 100644 --- a/kernel/trace/trace_event_perf.c +++ b/kernel/trace/trace_event_perf.c @@ -47,6 +47,9 @@ static int perf_trace_event_perm(struct trace_event_call *tp_event, if (perf_paranoid_tracepoint_raw() && !capable(CAP_SYS_ADMIN)) return -EPERM; + if (!is_sampling_event(p_event)) + return 0; + /* * We don't allow user space callchains for function trace * event, due to issues with page faults while tracing page