tracing: Remove redundant reset per-CPU buff in irqsoff tracer
authorDmitry Safonov <0x7f454c46@gmail.com>
Mon, 19 Oct 2015 18:10:26 +0000 (21:10 +0300)
committerSteven Rostedt <rostedt@goodmis.org>
Fri, 18 Mar 2016 20:39:11 +0000 (16:39 -0400)
commit741f3a69f101250dc6b171b88e14ea51b099b1a9
tree1dd4253e10c879854b82f5a9eb7c00a2d68a8e93
parent9d2099ab054558af0b3d4860b68a11aff420aa40
tracing: Remove redundant reset per-CPU buff in irqsoff tracer

  There is no reason to do it twice: from commit b6f11df26fdc28
("trace: Call tracing_reset_online_cpus before tracer->init()")
resetting of per-CPU buffers done before tracer->init() call.

tracer->init() calls {irqs,preempt,preemptirqs}off_tracer_init() and it
calls __irqsoff_tracer_init(), which resets per-CPU ringbuffer second
time.
It's slowpath, but anyway.

Link: http://lkml.kernel.org/r/1445278226-16187-1-git-send-email-0x7f454c46@gmail.com
Signed-off-by: Dmitry Safonov <0x7f454c46@gmail.com>
Signed-off-by: Steven Rostedt <rostedt@goodmis.org>
kernel/trace/trace_irqsoff.c