aboutsummaryrefslogtreecommitdiff
path: root/tools/perf/util/scripting-engines/trace-event-python.c
diff options
context:
space:
mode:
authorMasami Hiramatsu <[email protected]>2012-06-05 19:28:32 +0900
committerSteven Rostedt <[email protected]>2012-07-31 10:29:58 -0400
commitae6aa16fdc163afe6b04b6c073ad4ddd4663c03b (patch)
tree0a6055f56f80561c4a92df5ea50628644db73e7e /tools/perf/util/scripting-engines/trace-event-python.c
parent4dc936769e8a6382a4cc12375e8a4daa2b829fda (diff)
kprobes: introduce ftrace based optimization
Introduce function trace based kprobes optimization. With using ftrace optimization, kprobes on the mcount calling address, use ftrace's mcount call instead of breakpoint. Furthermore, this optimization works with preemptive kernel not like as current jump-based optimization. Of cource, this feature works only if the probe is on mcount call. Only if kprobe.break_handler is set, that probe is not optimized with ftrace (nor put on ftrace). The reason why this limitation comes is that this break_handler may be used only from jprobes which changes ip address (for fetching the function arguments), but function tracer ignores modified ip address. Changes in v2: - Fix ftrace_ops registering right after setting its filter. - Unregister ftrace_ops if there is no kprobe using. - Remove notrace dependency from __kprobes macro. Link: http://lkml.kernel.org/r/[email protected] Cc: Thomas Gleixner <[email protected]> Cc: Ingo Molnar <[email protected]> Cc: "H. Peter Anvin" <[email protected]> Cc: Ananth N Mavinakayanahalli <[email protected]> Cc: "Frank Ch. Eigler" <[email protected]> Cc: Andrew Morton <[email protected]> Cc: Frederic Weisbecker <[email protected]> Signed-off-by: Masami Hiramatsu <[email protected]> Signed-off-by: Steven Rostedt <[email protected]>
Diffstat (limited to 'tools/perf/util/scripting-engines/trace-event-python.c')
0 files changed, 0 insertions, 0 deletions