off_t offset;
        struct op_sample *sample;
 
-       sample = cpu_buffer_read_entry(cpu);
+       sample = op_cpu_buffer_read_entry(cpu);
        if (!sample)
                goto Error;
        rip = sample->eip;
                count = IBS_OP_CODE_SIZE;       /*IBS OP is 5 int64s*/
 
        for (i = 0; i < count; i++) {
-               sample = cpu_buffer_read_entry(cpu);
+               sample = op_cpu_buffer_read_entry(cpu);
                if (!sample)
                        goto Error;
                add_event_entry(sample->eip);
 
        add_cpu_switch(cpu);
 
-       cpu_buffer_reset(cpu);
-       available = cpu_buffer_entries(cpu);
+       op_cpu_buffer_reset(cpu);
+       available = op_cpu_buffer_entries(cpu);
 
        for (i = 0; i < available; ++i) {
-               struct op_sample *s = cpu_buffer_read_entry(cpu);
+               struct op_sample *s = op_cpu_buffer_read_entry(cpu);
                if (!s)
                        break;
 
 
        struct op_entry entry;
        int ret;
 
-       ret = cpu_buffer_write_entry(&entry);
+       ret = op_cpu_buffer_write_entry(&entry);
        if (ret)
                return ret;
 
        entry.sample->eip = pc;
        entry.sample->event = event;
 
-       ret = cpu_buffer_write_commit(&entry);
+       ret = op_cpu_buffer_write_commit(&entry);
        if (ret)
                return ret;
 
 
  * reset these to invalid values; the next sample collected will
  * populate the buffer with proper values to initialize the buffer
  */
-static inline void cpu_buffer_reset(int cpu)
+static inline void op_cpu_buffer_reset(int cpu)
 {
        struct oprofile_cpu_buffer *cpu_buf = &per_cpu(cpu_buffer, cpu);
 
        cpu_buf->last_task = NULL;
 }
 
-static inline int cpu_buffer_write_entry(struct op_entry *entry)
+static inline int op_cpu_buffer_write_entry(struct op_entry *entry)
 {
        entry->event = ring_buffer_lock_reserve(op_ring_buffer_write,
                                                sizeof(struct op_sample),
        return 0;
 }
 
-static inline int cpu_buffer_write_commit(struct op_entry *entry)
+static inline int op_cpu_buffer_write_commit(struct op_entry *entry)
 {
        return ring_buffer_unlock_commit(op_ring_buffer_write, entry->event,
                                         entry->irq_flags);
 }
 
-static inline struct op_sample *cpu_buffer_read_entry(int cpu)
+static inline struct op_sample *op_cpu_buffer_read_entry(int cpu)
 {
        struct ring_buffer_event *e;
        e = ring_buffer_consume(op_ring_buffer_read, cpu, NULL);
 }
 
 /* "acquire" as many cpu buffer slots as we can */
-static inline unsigned long cpu_buffer_entries(int cpu)
+static inline unsigned long op_cpu_buffer_entries(int cpu)
 {
        return ring_buffer_entries_cpu(op_ring_buffer_read, cpu)
                + ring_buffer_entries_cpu(op_ring_buffer_write, cpu);