perf: Per cpu-context rotation timer
Give each cpu-context its own timer so that it is a self contained entity, this eases the way for per-pmu-per-cpu contexts as well as provides the basic infrastructure to allow different rotation times per pmu. Things to look at: - folding the tick and these TICK_NSEC timers - separate task context rotation Signed-off-by: Peter Zijlstra <a.p.zijlstra@chello.nl> Cc: paulus <paulus@samba.org> Cc: stephane eranian <eranian@googlemail.com> Cc: Robert Richter <robert.richter@amd.com> Cc: Frederic Weisbecker <fweisbec@gmail.com> Cc: Lin Ming <ming.m.lin@intel.com> Cc: Yanmin <yanmin_zhang@linux.intel.com> LKML-Reference: <new-submission> Signed-off-by: Ingo Molnar <mingo@elte.hu>
This commit is contained in:
committed by
Ingo Molnar
parent
b28ab83c59
commit
b5ab4cd563
@@ -861,6 +861,8 @@ struct perf_cpu_context {
|
|||||||
struct perf_event_context *task_ctx;
|
struct perf_event_context *task_ctx;
|
||||||
int active_oncpu;
|
int active_oncpu;
|
||||||
int exclusive;
|
int exclusive;
|
||||||
|
u64 timer_interval;
|
||||||
|
struct hrtimer timer;
|
||||||
};
|
};
|
||||||
|
|
||||||
struct perf_output_handle {
|
struct perf_output_handle {
|
||||||
@@ -881,7 +883,6 @@ extern void perf_pmu_unregister(struct pmu *pmu);
|
|||||||
|
|
||||||
extern void perf_event_task_sched_in(struct task_struct *task);
|
extern void perf_event_task_sched_in(struct task_struct *task);
|
||||||
extern void perf_event_task_sched_out(struct task_struct *task, struct task_struct *next);
|
extern void perf_event_task_sched_out(struct task_struct *task, struct task_struct *next);
|
||||||
extern void perf_event_task_tick(struct task_struct *task);
|
|
||||||
extern int perf_event_init_task(struct task_struct *child);
|
extern int perf_event_init_task(struct task_struct *child);
|
||||||
extern void perf_event_exit_task(struct task_struct *child);
|
extern void perf_event_exit_task(struct task_struct *child);
|
||||||
extern void perf_event_free_task(struct task_struct *task);
|
extern void perf_event_free_task(struct task_struct *task);
|
||||||
@@ -1067,8 +1068,6 @@ perf_event_task_sched_in(struct task_struct *task) { }
|
|||||||
static inline void
|
static inline void
|
||||||
perf_event_task_sched_out(struct task_struct *task,
|
perf_event_task_sched_out(struct task_struct *task,
|
||||||
struct task_struct *next) { }
|
struct task_struct *next) { }
|
||||||
static inline void
|
|
||||||
perf_event_task_tick(struct task_struct *task) { }
|
|
||||||
static inline int perf_event_init_task(struct task_struct *child) { return 0; }
|
static inline int perf_event_init_task(struct task_struct *child) { return 0; }
|
||||||
static inline void perf_event_exit_task(struct task_struct *child) { }
|
static inline void perf_event_exit_task(struct task_struct *child) { }
|
||||||
static inline void perf_event_free_task(struct task_struct *task) { }
|
static inline void perf_event_free_task(struct task_struct *task) { }
|
||||||
|
@@ -78,6 +78,25 @@ void perf_pmu_enable(struct pmu *pmu)
|
|||||||
pmu->pmu_enable(pmu);
|
pmu->pmu_enable(pmu);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
static void perf_pmu_rotate_start(void)
|
||||||
|
{
|
||||||
|
struct perf_cpu_context *cpuctx = &__get_cpu_var(perf_cpu_context);
|
||||||
|
|
||||||
|
if (hrtimer_active(&cpuctx->timer))
|
||||||
|
return;
|
||||||
|
|
||||||
|
__hrtimer_start_range_ns(&cpuctx->timer,
|
||||||
|
ns_to_ktime(cpuctx->timer_interval), 0,
|
||||||
|
HRTIMER_MODE_REL_PINNED, 0);
|
||||||
|
}
|
||||||
|
|
||||||
|
static void perf_pmu_rotate_stop(void)
|
||||||
|
{
|
||||||
|
struct perf_cpu_context *cpuctx = &__get_cpu_var(perf_cpu_context);
|
||||||
|
|
||||||
|
hrtimer_cancel(&cpuctx->timer);
|
||||||
|
}
|
||||||
|
|
||||||
static void get_ctx(struct perf_event_context *ctx)
|
static void get_ctx(struct perf_event_context *ctx)
|
||||||
{
|
{
|
||||||
WARN_ON(!atomic_inc_not_zero(&ctx->refcount));
|
WARN_ON(!atomic_inc_not_zero(&ctx->refcount));
|
||||||
@@ -281,6 +300,8 @@ list_add_event(struct perf_event *event, struct perf_event_context *ctx)
|
|||||||
}
|
}
|
||||||
|
|
||||||
list_add_rcu(&event->event_entry, &ctx->event_list);
|
list_add_rcu(&event->event_entry, &ctx->event_list);
|
||||||
|
if (!ctx->nr_events)
|
||||||
|
perf_pmu_rotate_start();
|
||||||
ctx->nr_events++;
|
ctx->nr_events++;
|
||||||
if (event->attr.inherit_stat)
|
if (event->attr.inherit_stat)
|
||||||
ctx->nr_stat++;
|
ctx->nr_stat++;
|
||||||
@@ -1383,6 +1404,12 @@ void perf_event_task_sched_in(struct task_struct *task)
|
|||||||
ctx_sched_in(ctx, cpuctx, EVENT_FLEXIBLE);
|
ctx_sched_in(ctx, cpuctx, EVENT_FLEXIBLE);
|
||||||
|
|
||||||
cpuctx->task_ctx = ctx;
|
cpuctx->task_ctx = ctx;
|
||||||
|
|
||||||
|
/*
|
||||||
|
* Since these rotations are per-cpu, we need to ensure the
|
||||||
|
* cpu-context we got scheduled on is actually rotating.
|
||||||
|
*/
|
||||||
|
perf_pmu_rotate_start();
|
||||||
}
|
}
|
||||||
|
|
||||||
#define MAX_INTERRUPTS (~0ULL)
|
#define MAX_INTERRUPTS (~0ULL)
|
||||||
@@ -1487,7 +1514,7 @@ static void perf_adjust_period(struct perf_event *event, u64 nsec, u64 count)
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
static void perf_ctx_adjust_freq(struct perf_event_context *ctx)
|
static void perf_ctx_adjust_freq(struct perf_event_context *ctx, u64 period)
|
||||||
{
|
{
|
||||||
struct perf_event *event;
|
struct perf_event *event;
|
||||||
struct hw_perf_event *hwc;
|
struct hw_perf_event *hwc;
|
||||||
@@ -1524,7 +1551,7 @@ static void perf_ctx_adjust_freq(struct perf_event_context *ctx)
|
|||||||
hwc->freq_count_stamp = now;
|
hwc->freq_count_stamp = now;
|
||||||
|
|
||||||
if (delta > 0)
|
if (delta > 0)
|
||||||
perf_adjust_period(event, TICK_NSEC, delta);
|
perf_adjust_period(event, period, delta);
|
||||||
}
|
}
|
||||||
raw_spin_unlock(&ctx->lock);
|
raw_spin_unlock(&ctx->lock);
|
||||||
}
|
}
|
||||||
@@ -1542,30 +1569,39 @@ static void rotate_ctx(struct perf_event_context *ctx)
|
|||||||
raw_spin_unlock(&ctx->lock);
|
raw_spin_unlock(&ctx->lock);
|
||||||
}
|
}
|
||||||
|
|
||||||
void perf_event_task_tick(struct task_struct *curr)
|
/*
|
||||||
|
* Cannot race with ->pmu_rotate_start() because this is ran from hardirq
|
||||||
|
* context, and ->pmu_rotate_start() is called with irqs disabled (both are
|
||||||
|
* cpu affine, so there are no SMP races).
|
||||||
|
*/
|
||||||
|
static enum hrtimer_restart perf_event_context_tick(struct hrtimer *timer)
|
||||||
{
|
{
|
||||||
|
enum hrtimer_restart restart = HRTIMER_NORESTART;
|
||||||
struct perf_cpu_context *cpuctx;
|
struct perf_cpu_context *cpuctx;
|
||||||
struct perf_event_context *ctx;
|
struct perf_event_context *ctx;
|
||||||
int rotate = 0;
|
int rotate = 0;
|
||||||
|
|
||||||
if (!atomic_read(&nr_events))
|
cpuctx = container_of(timer, struct perf_cpu_context, timer);
|
||||||
return;
|
|
||||||
|
|
||||||
cpuctx = &__get_cpu_var(perf_cpu_context);
|
if (cpuctx->ctx.nr_events) {
|
||||||
if (cpuctx->ctx.nr_events &&
|
restart = HRTIMER_RESTART;
|
||||||
cpuctx->ctx.nr_events != cpuctx->ctx.nr_active)
|
if (cpuctx->ctx.nr_events != cpuctx->ctx.nr_active)
|
||||||
rotate = 1;
|
rotate = 1;
|
||||||
|
}
|
||||||
|
|
||||||
ctx = curr->perf_event_ctxp;
|
ctx = current->perf_event_ctxp;
|
||||||
if (ctx && ctx->nr_events && ctx->nr_events != ctx->nr_active)
|
if (ctx && ctx->nr_events) {
|
||||||
|
restart = HRTIMER_RESTART;
|
||||||
|
if (ctx->nr_events != ctx->nr_active)
|
||||||
rotate = 1;
|
rotate = 1;
|
||||||
|
}
|
||||||
|
|
||||||
perf_ctx_adjust_freq(&cpuctx->ctx);
|
perf_ctx_adjust_freq(&cpuctx->ctx, cpuctx->timer_interval);
|
||||||
if (ctx)
|
if (ctx)
|
||||||
perf_ctx_adjust_freq(ctx);
|
perf_ctx_adjust_freq(ctx, cpuctx->timer_interval);
|
||||||
|
|
||||||
if (!rotate)
|
if (!rotate)
|
||||||
return;
|
goto done;
|
||||||
|
|
||||||
cpu_ctx_sched_out(cpuctx, EVENT_FLEXIBLE);
|
cpu_ctx_sched_out(cpuctx, EVENT_FLEXIBLE);
|
||||||
if (ctx)
|
if (ctx)
|
||||||
@@ -1577,7 +1613,12 @@ void perf_event_task_tick(struct task_struct *curr)
|
|||||||
|
|
||||||
cpu_ctx_sched_in(cpuctx, EVENT_FLEXIBLE);
|
cpu_ctx_sched_in(cpuctx, EVENT_FLEXIBLE);
|
||||||
if (ctx)
|
if (ctx)
|
||||||
task_ctx_sched_in(curr, EVENT_FLEXIBLE);
|
task_ctx_sched_in(current, EVENT_FLEXIBLE);
|
||||||
|
|
||||||
|
done:
|
||||||
|
hrtimer_forward_now(timer, ns_to_ktime(cpuctx->timer_interval));
|
||||||
|
|
||||||
|
return restart;
|
||||||
}
|
}
|
||||||
|
|
||||||
static int event_enable_on_exec(struct perf_event *event,
|
static int event_enable_on_exec(struct perf_event *event,
|
||||||
@@ -4786,7 +4827,7 @@ static void perf_swevent_start_hrtimer(struct perf_event *event)
|
|||||||
}
|
}
|
||||||
__hrtimer_start_range_ns(&hwc->hrtimer,
|
__hrtimer_start_range_ns(&hwc->hrtimer,
|
||||||
ns_to_ktime(period), 0,
|
ns_to_ktime(period), 0,
|
||||||
HRTIMER_MODE_REL, 0);
|
HRTIMER_MODE_REL_PINNED, 0);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -5904,6 +5945,9 @@ static void __init perf_event_init_all_cpus(void)
|
|||||||
|
|
||||||
cpuctx = &per_cpu(perf_cpu_context, cpu);
|
cpuctx = &per_cpu(perf_cpu_context, cpu);
|
||||||
__perf_event_init_context(&cpuctx->ctx, NULL);
|
__perf_event_init_context(&cpuctx->ctx, NULL);
|
||||||
|
cpuctx->timer_interval = TICK_NSEC;
|
||||||
|
hrtimer_init(&cpuctx->timer, CLOCK_MONOTONIC, HRTIMER_MODE_REL);
|
||||||
|
cpuctx->timer.function = perf_event_context_tick;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -5934,6 +5978,8 @@ static void __perf_event_exit_cpu(void *info)
|
|||||||
struct perf_event_context *ctx = &cpuctx->ctx;
|
struct perf_event_context *ctx = &cpuctx->ctx;
|
||||||
struct perf_event *event, *tmp;
|
struct perf_event *event, *tmp;
|
||||||
|
|
||||||
|
perf_pmu_rotate_stop();
|
||||||
|
|
||||||
list_for_each_entry_safe(event, tmp, &ctx->pinned_groups, group_entry)
|
list_for_each_entry_safe(event, tmp, &ctx->pinned_groups, group_entry)
|
||||||
__perf_event_remove_from_context(event);
|
__perf_event_remove_from_context(event);
|
||||||
list_for_each_entry_safe(event, tmp, &ctx->flexible_groups, group_entry)
|
list_for_each_entry_safe(event, tmp, &ctx->flexible_groups, group_entry)
|
||||||
|
@@ -3578,8 +3578,6 @@ void scheduler_tick(void)
|
|||||||
curr->sched_class->task_tick(rq, curr, 0);
|
curr->sched_class->task_tick(rq, curr, 0);
|
||||||
raw_spin_unlock(&rq->lock);
|
raw_spin_unlock(&rq->lock);
|
||||||
|
|
||||||
perf_event_task_tick(curr);
|
|
||||||
|
|
||||||
#ifdef CONFIG_SMP
|
#ifdef CONFIG_SMP
|
||||||
rq->idle_at_tick = idle_cpu(cpu);
|
rq->idle_at_tick = idle_cpu(cpu);
|
||||||
trigger_load_balance(rq, cpu);
|
trigger_load_balance(rq, cpu);
|
||||||
|
Reference in New Issue
Block a user