Linux kernel mirror (for testing) git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
kernel os linux

tracing/perf: Avoid perf_trace_buf_*() in perf_trace_##call() when possible

perf_trace_buf_prepare() + perf_trace_buf_submit(task => NULL)
make no sense if hlist_empty(head). Change perf_trace_##call()
to check ->perf_events beforehand and do nothing if it is empty.

This removes the overhead for tasks without events associated
with them. For example, "perf record -e sched:sched_switch -p1"
attaches the counter(s) to the single task, but every task in
system will do perf_trace_buf_prepare/submit() just to realize
that it was not attached to this event.

However, we can only do this if __task == NULL, so we also add
the __builtin_constant_p(__task) check.

With this patch "perf bench sched pipe" shows approximately 4%
improvement when "perf record -p1" runs in parallel, many thanks
to Steven for the testing.

Link: http://lkml.kernel.org/r/20130806160847.GA2746@redhat.com

Tested-by: David Ahern <dsahern@gmail.com>
Acked-by: Peter Zijlstra <peterz@infradead.org>
Signed-off-by: Oleg Nesterov <oleg@redhat.com>
Signed-off-by: Steven Rostedt <rostedt@goodmis.org>

authored by

Oleg Nesterov and committed by
Steven Rostedt
d027e6a9 12473965

+6 -1
+6 -1
include/trace/ftrace.h
··· 667 667 int rctx; \ 668 668 \ 669 669 __data_size = ftrace_get_offsets_##call(&__data_offsets, args); \ 670 + \ 671 + head = this_cpu_ptr(event_call->perf_events); \ 672 + if (__builtin_constant_p(!__task) && !__task && \ 673 + hlist_empty(head)) \ 674 + return; \ 675 + \ 670 676 __entry_size = ALIGN(__data_size + sizeof(*entry) + sizeof(u32),\ 671 677 sizeof(u64)); \ 672 678 __entry_size -= sizeof(u32); \ ··· 687 681 \ 688 682 { assign; } \ 689 683 \ 690 - head = this_cpu_ptr(event_call->perf_events); \ 691 684 perf_trace_buf_submit(entry, __entry_size, rctx, __addr, \ 692 685 __count, &__regs, head, __task); \ 693 686 }