sched_ttwu_pending();
}
+void send_call_function_single_ipi(int cpu)
+{
+ struct rq *rq = cpu_rq(cpu);
+
+ if (!set_nr_if_polling(rq->idle))
+ arch_send_call_function_single_ipi(cpu);
+ else
+ trace_sched_wake_idle_without_ipi(cpu);
+}
+
/*
* Queue a task on the target CPUs wake_list and wake the CPU via IPI if
* necessary. The wakee CPU on receipt of the IPI will queue the task
}
#endif
-#else
+extern void flush_smp_call_function_from_idle(void);
+#else /* !CONFIG_SMP: */
+static inline void flush_smp_call_function_from_idle(void) { }
static inline void sched_ttwu_pending(void) { }
-
-#endif /* CONFIG_SMP */
+#endif
#include "stats.h"
#include "autogroup.h"
static DEFINE_PER_CPU_SHARED_ALIGNED(call_single_data_t, csd_data);
+extern void send_call_function_single_ipi(int cpu);
+
/*
* Insert a previously allocated call_single_data_t element
* for execution on the given CPU. data must already have
* equipped to do the right thing...
*/
if (llist_add(&csd->llist, &per_cpu(call_single_queue, cpu)))
- arch_send_call_function_single_ipi(cpu);
+ send_call_function_single_ipi(cpu);
return 0;
}
}
}
+void flush_smp_call_function_from_idle(void)
+{
+ unsigned long flags;
+
+ if (llist_empty(this_cpu_ptr(&call_single_queue)))
+ return;
+
+ local_irq_save(flags);
+ flush_smp_call_function_queue(true);
+ local_irq_restore(flags);
+}
+
/*
* smp_call_function_single - Run a function on a specific CPU
* @func: The function to run. This must be fast and non-blocking.