Merge tag 'trace-fixes-v3.10-rc1' of git://git.kernel.org/pub/scm/linux/kernel/git/rostedt/linux-trace
Pull tracing fixes from Steven Rostedt: "This includes a fix to a memory leak when adding filters to traces. Also, Masami Hiramatsu fixed up some minor bugs that were discovered by sparse." * tag 'trace-fixes-v3.10-rc1' of git://git.kernel.org/pub/scm/linux/kernel/git/rostedt/linux-trace: tracing/kprobes: Make print_*probe_event static tracing/kprobes: Fix a sparse warning for incorrect type in assignment tracing/kprobes: Use rcu_dereference_raw for tp->files tracing: Fix leaks of filter preds
This commit is contained in:
@@ -750,7 +750,11 @@ static int filter_set_pred(struct event_filter *filter,
|
|||||||
|
|
||||||
static void __free_preds(struct event_filter *filter)
|
static void __free_preds(struct event_filter *filter)
|
||||||
{
|
{
|
||||||
|
int i;
|
||||||
|
|
||||||
if (filter->preds) {
|
if (filter->preds) {
|
||||||
|
for (i = 0; i < filter->n_preds; i++)
|
||||||
|
kfree(filter->preds[i].ops);
|
||||||
kfree(filter->preds);
|
kfree(filter->preds);
|
||||||
filter->preds = NULL;
|
filter->preds = NULL;
|
||||||
}
|
}
|
||||||
|
@@ -35,7 +35,7 @@ struct trace_probe {
|
|||||||
const char *symbol; /* symbol name */
|
const char *symbol; /* symbol name */
|
||||||
struct ftrace_event_class class;
|
struct ftrace_event_class class;
|
||||||
struct ftrace_event_call call;
|
struct ftrace_event_call call;
|
||||||
struct ftrace_event_file **files;
|
struct ftrace_event_file * __rcu *files;
|
||||||
ssize_t size; /* trace entry size */
|
ssize_t size; /* trace entry size */
|
||||||
unsigned int nr_args;
|
unsigned int nr_args;
|
||||||
struct probe_arg args[];
|
struct probe_arg args[];
|
||||||
@@ -185,9 +185,14 @@ static struct trace_probe *find_trace_probe(const char *event,
|
|||||||
|
|
||||||
static int trace_probe_nr_files(struct trace_probe *tp)
|
static int trace_probe_nr_files(struct trace_probe *tp)
|
||||||
{
|
{
|
||||||
struct ftrace_event_file **file = tp->files;
|
struct ftrace_event_file **file;
|
||||||
int ret = 0;
|
int ret = 0;
|
||||||
|
|
||||||
|
/*
|
||||||
|
* Since all tp->files updater is protected by probe_enable_lock,
|
||||||
|
* we don't need to lock an rcu_read_lock.
|
||||||
|
*/
|
||||||
|
file = rcu_dereference_raw(tp->files);
|
||||||
if (file)
|
if (file)
|
||||||
while (*(file++))
|
while (*(file++))
|
||||||
ret++;
|
ret++;
|
||||||
@@ -209,9 +214,10 @@ enable_trace_probe(struct trace_probe *tp, struct ftrace_event_file *file)
|
|||||||
mutex_lock(&probe_enable_lock);
|
mutex_lock(&probe_enable_lock);
|
||||||
|
|
||||||
if (file) {
|
if (file) {
|
||||||
struct ftrace_event_file **new, **old = tp->files;
|
struct ftrace_event_file **new, **old;
|
||||||
int n = trace_probe_nr_files(tp);
|
int n = trace_probe_nr_files(tp);
|
||||||
|
|
||||||
|
old = rcu_dereference_raw(tp->files);
|
||||||
/* 1 is for new one and 1 is for stopper */
|
/* 1 is for new one and 1 is for stopper */
|
||||||
new = kzalloc((n + 2) * sizeof(struct ftrace_event_file *),
|
new = kzalloc((n + 2) * sizeof(struct ftrace_event_file *),
|
||||||
GFP_KERNEL);
|
GFP_KERNEL);
|
||||||
@@ -251,11 +257,17 @@ enable_trace_probe(struct trace_probe *tp, struct ftrace_event_file *file)
|
|||||||
static int
|
static int
|
||||||
trace_probe_file_index(struct trace_probe *tp, struct ftrace_event_file *file)
|
trace_probe_file_index(struct trace_probe *tp, struct ftrace_event_file *file)
|
||||||
{
|
{
|
||||||
|
struct ftrace_event_file **files;
|
||||||
int i;
|
int i;
|
||||||
|
|
||||||
if (tp->files) {
|
/*
|
||||||
for (i = 0; tp->files[i]; i++)
|
* Since all tp->files updater is protected by probe_enable_lock,
|
||||||
if (tp->files[i] == file)
|
* we don't need to lock an rcu_read_lock.
|
||||||
|
*/
|
||||||
|
files = rcu_dereference_raw(tp->files);
|
||||||
|
if (files) {
|
||||||
|
for (i = 0; files[i]; i++)
|
||||||
|
if (files[i] == file)
|
||||||
return i;
|
return i;
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -274,10 +286,11 @@ disable_trace_probe(struct trace_probe *tp, struct ftrace_event_file *file)
|
|||||||
mutex_lock(&probe_enable_lock);
|
mutex_lock(&probe_enable_lock);
|
||||||
|
|
||||||
if (file) {
|
if (file) {
|
||||||
struct ftrace_event_file **new, **old = tp->files;
|
struct ftrace_event_file **new, **old;
|
||||||
int n = trace_probe_nr_files(tp);
|
int n = trace_probe_nr_files(tp);
|
||||||
int i, j;
|
int i, j;
|
||||||
|
|
||||||
|
old = rcu_dereference_raw(tp->files);
|
||||||
if (n == 0 || trace_probe_file_index(tp, file) < 0) {
|
if (n == 0 || trace_probe_file_index(tp, file) < 0) {
|
||||||
ret = -EINVAL;
|
ret = -EINVAL;
|
||||||
goto out_unlock;
|
goto out_unlock;
|
||||||
@@ -872,9 +885,16 @@ __kprobe_trace_func(struct trace_probe *tp, struct pt_regs *regs,
|
|||||||
static __kprobes void
|
static __kprobes void
|
||||||
kprobe_trace_func(struct trace_probe *tp, struct pt_regs *regs)
|
kprobe_trace_func(struct trace_probe *tp, struct pt_regs *regs)
|
||||||
{
|
{
|
||||||
struct ftrace_event_file **file = tp->files;
|
/*
|
||||||
|
* Note: preempt is already disabled around the kprobe handler.
|
||||||
|
* However, we still need an smp_read_barrier_depends() corresponding
|
||||||
|
* to smp_wmb() in rcu_assign_pointer() to access the pointer.
|
||||||
|
*/
|
||||||
|
struct ftrace_event_file **file = rcu_dereference_raw(tp->files);
|
||||||
|
|
||||||
|
if (unlikely(!file))
|
||||||
|
return;
|
||||||
|
|
||||||
/* Note: preempt is already disabled around the kprobe handler */
|
|
||||||
while (*file) {
|
while (*file) {
|
||||||
__kprobe_trace_func(tp, regs, *file);
|
__kprobe_trace_func(tp, regs, *file);
|
||||||
file++;
|
file++;
|
||||||
@@ -925,9 +945,16 @@ static __kprobes void
|
|||||||
kretprobe_trace_func(struct trace_probe *tp, struct kretprobe_instance *ri,
|
kretprobe_trace_func(struct trace_probe *tp, struct kretprobe_instance *ri,
|
||||||
struct pt_regs *regs)
|
struct pt_regs *regs)
|
||||||
{
|
{
|
||||||
struct ftrace_event_file **file = tp->files;
|
/*
|
||||||
|
* Note: preempt is already disabled around the kprobe handler.
|
||||||
|
* However, we still need an smp_read_barrier_depends() corresponding
|
||||||
|
* to smp_wmb() in rcu_assign_pointer() to access the pointer.
|
||||||
|
*/
|
||||||
|
struct ftrace_event_file **file = rcu_dereference_raw(tp->files);
|
||||||
|
|
||||||
|
if (unlikely(!file))
|
||||||
|
return;
|
||||||
|
|
||||||
/* Note: preempt is already disabled around the kprobe handler */
|
|
||||||
while (*file) {
|
while (*file) {
|
||||||
__kretprobe_trace_func(tp, ri, regs, *file);
|
__kretprobe_trace_func(tp, ri, regs, *file);
|
||||||
file++;
|
file++;
|
||||||
@@ -935,7 +962,7 @@ kretprobe_trace_func(struct trace_probe *tp, struct kretprobe_instance *ri,
|
|||||||
}
|
}
|
||||||
|
|
||||||
/* Event entry printers */
|
/* Event entry printers */
|
||||||
enum print_line_t
|
static enum print_line_t
|
||||||
print_kprobe_event(struct trace_iterator *iter, int flags,
|
print_kprobe_event(struct trace_iterator *iter, int flags,
|
||||||
struct trace_event *event)
|
struct trace_event *event)
|
||||||
{
|
{
|
||||||
@@ -971,7 +998,7 @@ partial:
|
|||||||
return TRACE_TYPE_PARTIAL_LINE;
|
return TRACE_TYPE_PARTIAL_LINE;
|
||||||
}
|
}
|
||||||
|
|
||||||
enum print_line_t
|
static enum print_line_t
|
||||||
print_kretprobe_event(struct trace_iterator *iter, int flags,
|
print_kretprobe_event(struct trace_iterator *iter, int flags,
|
||||||
struct trace_event *event)
|
struct trace_event *event)
|
||||||
{
|
{
|
||||||
|
Reference in New Issue
Block a user