|
@@ -1301,7 +1301,7 @@ struct ring_buffer *__ring_buffer_alloc(unsigned long size, unsigned flags,
|
|
|
* In that off case, we need to allocate for all possible cpus.
|
|
|
*/
|
|
|
#ifdef CONFIG_HOTPLUG_CPU
|
|
|
- get_online_cpus();
|
|
|
+ cpu_notifier_register_begin();
|
|
|
cpumask_copy(buffer->cpumask, cpu_online_mask);
|
|
|
#else
|
|
|
cpumask_copy(buffer->cpumask, cpu_possible_mask);
|
|
@@ -1324,10 +1324,10 @@ struct ring_buffer *__ring_buffer_alloc(unsigned long size, unsigned flags,
|
|
|
#ifdef CONFIG_HOTPLUG_CPU
|
|
|
buffer->cpu_notify.notifier_call = rb_cpu_notify;
|
|
|
buffer->cpu_notify.priority = 0;
|
|
|
- register_cpu_notifier(&buffer->cpu_notify);
|
|
|
+ __register_cpu_notifier(&buffer->cpu_notify);
|
|
|
+ cpu_notifier_register_done();
|
|
|
#endif
|
|
|
|
|
|
- put_online_cpus();
|
|
|
mutex_init(&buffer->mutex);
|
|
|
|
|
|
return buffer;
|
|
@@ -1341,7 +1341,9 @@ struct ring_buffer *__ring_buffer_alloc(unsigned long size, unsigned flags,
|
|
|
|
|
|
fail_free_cpumask:
|
|
|
free_cpumask_var(buffer->cpumask);
|
|
|
- put_online_cpus();
|
|
|
+#ifdef CONFIG_HOTPLUG_CPU
|
|
|
+ cpu_notifier_register_done();
|
|
|
+#endif
|
|
|
|
|
|
fail_free_buffer:
|
|
|
kfree(buffer);
|
|
@@ -1358,16 +1360,17 @@ ring_buffer_free(struct ring_buffer *buffer)
|
|
|
{
|
|
|
int cpu;
|
|
|
|
|
|
- get_online_cpus();
|
|
|
-
|
|
|
#ifdef CONFIG_HOTPLUG_CPU
|
|
|
- unregister_cpu_notifier(&buffer->cpu_notify);
|
|
|
+ cpu_notifier_register_begin();
|
|
|
+ __unregister_cpu_notifier(&buffer->cpu_notify);
|
|
|
#endif
|
|
|
|
|
|
for_each_buffer_cpu(buffer, cpu)
|
|
|
rb_free_cpu_buffer(buffer->buffers[cpu]);
|
|
|
|
|
|
- put_online_cpus();
|
|
|
+#ifdef CONFIG_HOTPLUG_CPU
|
|
|
+ cpu_notifier_register_done();
|
|
|
+#endif
|
|
|
|
|
|
kfree(buffer->buffers);
|
|
|
free_cpumask_var(buffer->cpumask);
|