percpu: make percpu symbols in oprofile unique

This patch updates percpu related symbols in oprofile such that percpu
symbols are unique and don't clash with local symbols.  This serves
two purposes of decreasing the possibility of global percpu symbol
collision and allowing dropping per_cpu__ prefix from percpu symbols.

* drivers/oprofile/cpu_buffer.c: s/cpu_buffer/op_cpu_buffer/

Partly based on Rusty Russell's "alloc_percpu: rename percpu vars
which cause name clashes" patch.

Signed-off-by: Tejun Heo <tj@kernel.org>
Acked-by: Robert Richter <robert.richter@amd.com>
Cc: Rusty Russell <rusty@rustcorp.com.au>
This commit is contained in:
Tejun Heo 2009-10-29 22:34:13 +09:00
parent 9705f69ed0
commit b3e9f672b6
3 changed files with 13 additions and 14 deletions

View File

@ -47,7 +47,7 @@
*/ */
static struct ring_buffer *op_ring_buffer_read; static struct ring_buffer *op_ring_buffer_read;
static struct ring_buffer *op_ring_buffer_write; static struct ring_buffer *op_ring_buffer_write;
DEFINE_PER_CPU(struct oprofile_cpu_buffer, cpu_buffer); DEFINE_PER_CPU(struct oprofile_cpu_buffer, op_cpu_buffer);
static void wq_sync_buffer(struct work_struct *work); static void wq_sync_buffer(struct work_struct *work);
@ -61,8 +61,7 @@ unsigned long oprofile_get_cpu_buffer_size(void)
void oprofile_cpu_buffer_inc_smpl_lost(void) void oprofile_cpu_buffer_inc_smpl_lost(void)
{ {
struct oprofile_cpu_buffer *cpu_buf struct oprofile_cpu_buffer *cpu_buf = &__get_cpu_var(op_cpu_buffer);
= &__get_cpu_var(cpu_buffer);
cpu_buf->sample_lost_overflow++; cpu_buf->sample_lost_overflow++;
} }
@ -95,7 +94,7 @@ int alloc_cpu_buffers(void)
goto fail; goto fail;
for_each_possible_cpu(i) { for_each_possible_cpu(i) {
struct oprofile_cpu_buffer *b = &per_cpu(cpu_buffer, i); struct oprofile_cpu_buffer *b = &per_cpu(op_cpu_buffer, i);
b->last_task = NULL; b->last_task = NULL;
b->last_is_kernel = -1; b->last_is_kernel = -1;
@ -122,7 +121,7 @@ void start_cpu_work(void)
work_enabled = 1; work_enabled = 1;
for_each_online_cpu(i) { for_each_online_cpu(i) {
struct oprofile_cpu_buffer *b = &per_cpu(cpu_buffer, i); struct oprofile_cpu_buffer *b = &per_cpu(op_cpu_buffer, i);
/* /*
* Spread the work by 1 jiffy per cpu so they dont all * Spread the work by 1 jiffy per cpu so they dont all
@ -139,7 +138,7 @@ void end_cpu_work(void)
work_enabled = 0; work_enabled = 0;
for_each_online_cpu(i) { for_each_online_cpu(i) {
struct oprofile_cpu_buffer *b = &per_cpu(cpu_buffer, i); struct oprofile_cpu_buffer *b = &per_cpu(op_cpu_buffer, i);
cancel_delayed_work(&b->work); cancel_delayed_work(&b->work);
} }
@ -330,7 +329,7 @@ static inline void
__oprofile_add_ext_sample(unsigned long pc, struct pt_regs * const regs, __oprofile_add_ext_sample(unsigned long pc, struct pt_regs * const regs,
unsigned long event, int is_kernel) unsigned long event, int is_kernel)
{ {
struct oprofile_cpu_buffer *cpu_buf = &__get_cpu_var(cpu_buffer); struct oprofile_cpu_buffer *cpu_buf = &__get_cpu_var(op_cpu_buffer);
unsigned long backtrace = oprofile_backtrace_depth; unsigned long backtrace = oprofile_backtrace_depth;
/* /*
@ -375,7 +374,7 @@ oprofile_write_reserve(struct op_entry *entry, struct pt_regs * const regs,
{ {
struct op_sample *sample; struct op_sample *sample;
int is_kernel = !user_mode(regs); int is_kernel = !user_mode(regs);
struct oprofile_cpu_buffer *cpu_buf = &__get_cpu_var(cpu_buffer); struct oprofile_cpu_buffer *cpu_buf = &__get_cpu_var(op_cpu_buffer);
cpu_buf->sample_received++; cpu_buf->sample_received++;
@ -430,13 +429,13 @@ int oprofile_write_commit(struct op_entry *entry)
void oprofile_add_pc(unsigned long pc, int is_kernel, unsigned long event) void oprofile_add_pc(unsigned long pc, int is_kernel, unsigned long event)
{ {
struct oprofile_cpu_buffer *cpu_buf = &__get_cpu_var(cpu_buffer); struct oprofile_cpu_buffer *cpu_buf = &__get_cpu_var(op_cpu_buffer);
log_sample(cpu_buf, pc, 0, is_kernel, event); log_sample(cpu_buf, pc, 0, is_kernel, event);
} }
void oprofile_add_trace(unsigned long pc) void oprofile_add_trace(unsigned long pc)
{ {
struct oprofile_cpu_buffer *cpu_buf = &__get_cpu_var(cpu_buffer); struct oprofile_cpu_buffer *cpu_buf = &__get_cpu_var(op_cpu_buffer);
if (!cpu_buf->tracing) if (!cpu_buf->tracing)
return; return;

View File

@ -50,7 +50,7 @@ struct oprofile_cpu_buffer {
struct delayed_work work; struct delayed_work work;
}; };
DECLARE_PER_CPU(struct oprofile_cpu_buffer, cpu_buffer); DECLARE_PER_CPU(struct oprofile_cpu_buffer, op_cpu_buffer);
/* /*
* Resets the cpu buffer to a sane state. * Resets the cpu buffer to a sane state.
@ -60,7 +60,7 @@ DECLARE_PER_CPU(struct oprofile_cpu_buffer, cpu_buffer);
*/ */
static inline void op_cpu_buffer_reset(int cpu) static inline void op_cpu_buffer_reset(int cpu)
{ {
struct oprofile_cpu_buffer *cpu_buf = &per_cpu(cpu_buffer, cpu); struct oprofile_cpu_buffer *cpu_buf = &per_cpu(op_cpu_buffer, cpu);
cpu_buf->last_is_kernel = -1; cpu_buf->last_is_kernel = -1;
cpu_buf->last_task = NULL; cpu_buf->last_task = NULL;

View File

@ -23,7 +23,7 @@ void oprofile_reset_stats(void)
int i; int i;
for_each_possible_cpu(i) { for_each_possible_cpu(i) {
cpu_buf = &per_cpu(cpu_buffer, i); cpu_buf = &per_cpu(op_cpu_buffer, i);
cpu_buf->sample_received = 0; cpu_buf->sample_received = 0;
cpu_buf->sample_lost_overflow = 0; cpu_buf->sample_lost_overflow = 0;
cpu_buf->backtrace_aborted = 0; cpu_buf->backtrace_aborted = 0;
@ -51,7 +51,7 @@ void oprofile_create_stats_files(struct super_block *sb, struct dentry *root)
return; return;
for_each_possible_cpu(i) { for_each_possible_cpu(i) {
cpu_buf = &per_cpu(cpu_buffer, i); cpu_buf = &per_cpu(op_cpu_buffer, i);
snprintf(buf, 10, "cpu%d", i); snprintf(buf, 10, "cpu%d", i);
cpudir = oprofilefs_mkdir(sb, dir, buf); cpudir = oprofilefs_mkdir(sb, dir, buf);