{
        struct ring_buffer_per_cpu *cpu_buffer;
 
-       get_online_cpus();
-
        if (!cpumask_test_cpu(cpu, buffer->cpumask))
-               goto out;
+               return;
 
        cpu_buffer = buffer->buffers[cpu];
        atomic_inc(&cpu_buffer->record_disabled);
- out:
-       put_online_cpus();
 }
 EXPORT_SYMBOL_GPL(ring_buffer_record_disable_cpu);
 
 {
        struct ring_buffer_per_cpu *cpu_buffer;
 
-       get_online_cpus();
-
        if (!cpumask_test_cpu(cpu, buffer->cpumask))
-               goto out;
+               return;
 
        cpu_buffer = buffer->buffers[cpu];
        atomic_dec(&cpu_buffer->record_disabled);
- out:
-       put_online_cpus();
 }
 EXPORT_SYMBOL_GPL(ring_buffer_record_enable_cpu);
 
 unsigned long ring_buffer_entries_cpu(struct ring_buffer *buffer, int cpu)
 {
        struct ring_buffer_per_cpu *cpu_buffer;
-       unsigned long ret = 0;
-
-       get_online_cpus();
+       unsigned long ret;
 
        if (!cpumask_test_cpu(cpu, buffer->cpumask))
-               goto out;
+               return 0;
 
        cpu_buffer = buffer->buffers[cpu];
        ret = cpu_buffer->entries;
- out:
-       put_online_cpus();
 
        return ret;
 }
 unsigned long ring_buffer_overrun_cpu(struct ring_buffer *buffer, int cpu)
 {
        struct ring_buffer_per_cpu *cpu_buffer;
-       unsigned long ret = 0;
-
-       get_online_cpus();
+       unsigned long ret;
 
        if (!cpumask_test_cpu(cpu, buffer->cpumask))
-               goto out;
+               return 0;
 
        cpu_buffer = buffer->buffers[cpu];
        ret = cpu_buffer->overrun;
- out:
-       put_online_cpus();
 
        return ret;
 }
        unsigned long entries = 0;
        int cpu;
 
-       get_online_cpus();
-
        /* if you care about this being correct, lock the buffer */
        for_each_buffer_cpu(buffer, cpu) {
                cpu_buffer = buffer->buffers[cpu];
                entries += cpu_buffer->entries;
        }
 
-       put_online_cpus();
-
        return entries;
 }
 EXPORT_SYMBOL_GPL(ring_buffer_entries);
        unsigned long overruns = 0;
        int cpu;
 
-       get_online_cpus();
-
        /* if you care about this being correct, lock the buffer */
        for_each_buffer_cpu(buffer, cpu) {
                cpu_buffer = buffer->buffers[cpu];
                overruns += cpu_buffer->overrun;
        }
 
-       put_online_cpus();
-
        return overruns;
 }
 EXPORT_SYMBOL_GPL(ring_buffer_overruns);
 ring_buffer_peek(struct ring_buffer *buffer, int cpu, u64 *ts)
 {
        struct ring_buffer_per_cpu *cpu_buffer = buffer->buffers[cpu];
-       struct ring_buffer_event *event = NULL;
+       struct ring_buffer_event *event;
        unsigned long flags;
 
-       get_online_cpus();
-
        if (!cpumask_test_cpu(cpu, buffer->cpumask))
-               goto out;
+               return NULL;
 
        spin_lock_irqsave(&cpu_buffer->reader_lock, flags);
        event = rb_buffer_peek(buffer, cpu, ts);
        spin_unlock_irqrestore(&cpu_buffer->reader_lock, flags);
 
- out:
-       put_online_cpus();
-
        return event;
 }
 
 ring_buffer_read_start(struct ring_buffer *buffer, int cpu)
 {
        struct ring_buffer_per_cpu *cpu_buffer;
-       struct ring_buffer_iter *iter = NULL;
+       struct ring_buffer_iter *iter;
        unsigned long flags;
 
-       get_online_cpus();
-
        if (!cpumask_test_cpu(cpu, buffer->cpumask))
-               goto out;
+               return NULL;
 
        iter = kmalloc(sizeof(*iter), GFP_KERNEL);
        if (!iter)
-               goto out;
+               return NULL;
 
        cpu_buffer = buffer->buffers[cpu];
 
        __raw_spin_unlock(&cpu_buffer->lock);
        spin_unlock_irqrestore(&cpu_buffer->reader_lock, flags);
 
- out:
-       put_online_cpus();
-
        return iter;
 }
 EXPORT_SYMBOL_GPL(ring_buffer_read_start);
 {
        struct ring_buffer_per_cpu *cpu_buffer = buffer->buffers[cpu];
        unsigned long flags;
-       int resched;
-
-       /* Can't use get_online_cpus because this can be in atomic */
-       resched = ftrace_preempt_disable();
 
        if (!cpumask_test_cpu(cpu, buffer->cpumask))
-               goto out;
+               return;
 
        spin_lock_irqsave(&cpu_buffer->reader_lock, flags);
 
        __raw_spin_unlock(&cpu_buffer->lock);
 
        spin_unlock_irqrestore(&cpu_buffer->reader_lock, flags);
- out:
-       ftrace_preempt_enable(resched);
 }
 EXPORT_SYMBOL_GPL(ring_buffer_reset_cpu);
 
  */
 void ring_buffer_reset(struct ring_buffer *buffer)
 {
-       int resched;
        int cpu;
 
-       /* Can't use get_online_cpus because this can be in atomic */
-       resched = ftrace_preempt_disable();
-
        for_each_buffer_cpu(buffer, cpu)
                ring_buffer_reset_cpu(buffer, cpu);
-
-       ftrace_preempt_enable(resched);
 }
 EXPORT_SYMBOL_GPL(ring_buffer_reset);
 
        struct ring_buffer_per_cpu *cpu_buffer;
        int cpu;
 
-       get_online_cpus();
-
        /* yes this is racy, but if you don't like the race, lock the buffer */
        for_each_buffer_cpu(buffer, cpu) {
                cpu_buffer = buffer->buffers[cpu];
                        return 0;
        }
 
-       put_online_cpus();
-
        return 1;
 }
 EXPORT_SYMBOL_GPL(ring_buffer_empty);
 int ring_buffer_empty_cpu(struct ring_buffer *buffer, int cpu)
 {
        struct ring_buffer_per_cpu *cpu_buffer;
-       int ret = 1;
-
-       get_online_cpus();
+       int ret;
 
        if (!cpumask_test_cpu(cpu, buffer->cpumask))
-               goto out;
+               return 1;
 
        cpu_buffer = buffer->buffers[cpu];
        ret = rb_per_cpu_empty(cpu_buffer);
 
- out:
-       put_online_cpus();
 
        return ret;
 }
        struct ring_buffer_per_cpu *cpu_buffer_b;
        int ret = -EINVAL;
 
-       get_online_cpus();
-
        if (!cpumask_test_cpu(cpu, buffer_a->cpumask) ||
            !cpumask_test_cpu(cpu, buffer_b->cpumask))
                goto out;
 
        ret = 0;
 out:
-       put_online_cpus();
-
        return ret;
 }
 EXPORT_SYMBOL_GPL(ring_buffer_swap_cpu);
        u64 save_timestamp;
        int ret = -1;
 
-       get_online_cpus();
-
        if (!cpumask_test_cpu(cpu, buffer->cpumask))
                goto out;
 
        spin_unlock_irqrestore(&cpu_buffer->reader_lock, flags);
 
  out:
-       put_online_cpus();
-
        return ret;
 }