X-Git-Url: http://git.lttng.org./?a=blobdiff_plain;f=lttng-context-perf-counters.c;h=5327e6ef06956707a914d15c3d56e9924291d06b;hb=96ba7208aa8009345a525148e32f78cd8d4693ab;hp=20d49d580760b8c4ed8cd15ce239bb6c2d77b17b;hpb=bef96e480c8b83dc90b22803b865ad2a40bdcb67;p=lttng-modules.git diff --git a/lttng-context-perf-counters.c b/lttng-context-perf-counters.c index 20d49d58..5327e6ef 100644 --- a/lttng-context-perf-counters.c +++ b/lttng-context-perf-counters.c @@ -17,13 +17,6 @@ #include "wrapper/vmalloc.h" #include "ltt-tracer.h" -/* - * TODO: Add CPU hotplug support. - */ - -static DEFINE_MUTEX(perf_counter_mutex); -static LIST_HEAD(perf_counter_contexts); - static size_t perf_counter_get_size(size_t offset) { @@ -43,8 +36,19 @@ void perf_counter_record(struct lttng_ctx_field *field, uint64_t value; event = field->u.perf_counter.e[ctx->cpu]; - event->pmu->read(event); - value = local64_read(&event->count); + if (likely(event)) { + event->pmu->read(event); + value = local64_read(&event->count); + } else { + /* + * Perf chooses not to be clever and not to support enabling a + * perf counter before the cpu is brought up. Therefore, we need + * to support having events coming (e.g. scheduler events) + * before the counter is setup. Write an arbitrary 0 in this + * case. + */ + value = 0; + } lib_ring_buffer_align_ctx(ctx, ltt_alignof(value)); chan->ops->event_write(ctx, &value, sizeof(value)); } @@ -62,16 +66,71 @@ void lttng_destroy_perf_counter_field(struct lttng_ctx_field *field) struct perf_event **events = field->u.perf_counter.e; int cpu; - mutex_lock(&perf_counter_mutex); - list_del(&field->u.perf_counter.head); + get_online_cpus(); for_each_online_cpu(cpu) perf_event_release_kernel(events[cpu]); - mutex_unlock(&perf_counter_mutex); + put_online_cpus(); +#ifdef CONFIG_HOTPLUG_CPU + unregister_cpu_notifier(&field->u.perf_counter.nb); +#endif kfree(field->event_field.name); kfree(field->u.perf_counter.attr); kfree(events); } +#ifdef CONFIG_HOTPLUG_CPU + +/** + * lttng_perf_counter_hp_callback - CPU hotplug callback + * @nb: notifier block + * @action: hotplug action to take + * @hcpu: CPU number + * + * Returns the success/failure of the operation. (%NOTIFY_OK, %NOTIFY_BAD) + * + * We can setup perf counters when the cpu is online (up prepare seems to be too + * soon). + */ +static +int __cpuinit lttng_perf_counter_cpu_hp_callback(struct notifier_block *nb, + unsigned long action, + void *hcpu) +{ + unsigned int cpu = (unsigned long) hcpu; + struct lttng_ctx_field *field = + container_of(nb, struct lttng_ctx_field, u.perf_counter.nb); + struct perf_event **events = field->u.perf_counter.e; + struct perf_event_attr *attr = field->u.perf_counter.attr; + struct perf_event *pevent; + + if (!field->u.perf_counter.hp_enable) + return NOTIFY_OK; + + switch (action) { + case CPU_ONLINE: + case CPU_ONLINE_FROZEN: + pevent = perf_event_create_kernel_counter(attr, + cpu, NULL, overflow_callback); + if (!pevent || IS_ERR(pevent)) + return NOTIFY_BAD; + barrier(); /* Create perf counter before setting event */ + events[cpu] = pevent; + break; + case CPU_UP_CANCELED: + case CPU_UP_CANCELED_FROZEN: + case CPU_DEAD: + case CPU_DEAD_FROZEN: + pevent = events[cpu]; + events[cpu] = NULL; + barrier(); /* NULLify event before perf counter teardown */ + perf_event_release_kernel(pevent); + break; + } + return NOTIFY_OK; +} + +#endif + int lttng_add_perf_counter_to_ctx(uint32_t type, uint64_t config, const char *name, @@ -100,17 +159,6 @@ int lttng_add_perf_counter_to_ctx(uint32_t type, attr->pinned = 1; attr->disabled = 0; - mutex_lock(&perf_counter_mutex); - - for_each_online_cpu(cpu) { - events[cpu] = perf_event_create_kernel_counter(attr, - cpu, NULL, overflow_callback); - if (!events[cpu]) { - ret = -EINVAL; - goto name_alloc_error; - } - } - name_alloc = kstrdup(name, GFP_KERNEL); if (!name_alloc) { ret = -ENOMEM; @@ -120,8 +168,27 @@ int lttng_add_perf_counter_to_ctx(uint32_t type, field = lttng_append_context(ctx); if (!field) { ret = -ENOMEM; - goto error; + goto append_context_error; } + +#ifdef CONFIG_HOTPLUG_CPU + field->u.perf_counter.nb.notifier_call = + lttng_perf_counter_cpu_hp_callback; + field->u.perf_counter.nb.priority = 0; + register_cpu_notifier(&field->u.perf_counter.nb); +#endif + + get_online_cpus(); + for_each_online_cpu(cpu) { + events[cpu] = perf_event_create_kernel_counter(attr, + cpu, NULL, overflow_callback); + if (!events[cpu] || IS_ERR(events[cpu])) { + ret = -EINVAL; + goto counter_error; + } + } + put_online_cpus(); + field->destroy = lttng_destroy_perf_counter_field; field->event_field.name = name_alloc; @@ -136,21 +203,24 @@ int lttng_add_perf_counter_to_ctx(uint32_t type, field->record = perf_counter_record; field->u.perf_counter.e = events; field->u.perf_counter.attr = attr; - - list_add(&field->u.perf_counter.head, &perf_counter_contexts); - mutex_unlock(&perf_counter_mutex); + field->u.perf_counter.hp_enable = 1; wrapper_vmalloc_sync_all(); return 0; -error: - kfree(name_alloc); -name_alloc_error: +counter_error: for_each_online_cpu(cpu) { - if (events[cpu]) + if (events[cpu] && !IS_ERR(events[cpu])) perf_event_release_kernel(events[cpu]); } - mutex_unlock(&perf_counter_mutex); + put_online_cpus(); +#ifdef CONFIG_HOTPLUG_CPU + unregister_cpu_notifier(&field->u.perf_counter.nb); +#endif + lttng_remove_context_field(ctx, field); +append_context_error: + kfree(name_alloc); +name_alloc_error: kfree(attr); error_attr: kfree(events);