Message ID | 20240613181613.4329-1-kprateek.nayak@amd.com |
---|---|
State | New |
Headers | show |
On Thu, Jun 13, 2024 at 06:15:59PM +0000, K Prateek Nayak wrote: > Effects of call_function_single_prep_ipi() > ========================================== > > To pull a TIF_POLLING thread out of idle to process an IPI, the sender > sets the TIF_NEED_RESCHED bit in the idle task's thread info in > call_function_single_prep_ipi() and avoids sending an actual IPI to the > target. As a result, the scheduler expects a task to be enqueued when > exiting the idle path. This is not the case with non-polling idle states > where the idle CPU exits the non-polling idle state to process the > interrupt, and since need_resched() returns false, soon goes back to > idle again. > > When TIF_NEED_RESCHED flag is set, do_idle() will call schedule_idle(), > a large part of which runs with local IRQ disabled. In case of ipistorm, > when measuring IPI throughput, this large IRQ disabled section delays > processing of IPIs. Further auditing revealed that in absence of any > runnable tasks, pick_next_task_fair(), which is called from the > pick_next_task() fast path, will always call newidle_balance() in this > scenario, further increasing the time spent in the IRQ disabled section. > > Following is the crude visualization of the problem with relevant > functions expanded: > -- > CPU0 CPU1 > ==== ==== > do_idle() { > __current_set_polling(); > ... > monitor(addr); > if (!need_resched()) > mwait() { > /* Waiting */ > smp_call_function_single(CPU1, func, wait = 1) { ... > ... ... > set_nr_if_polling(CPU1) { ... > /* Realizes CPU1 is polling */ ... > try_cmpxchg(addr, ... > &val, ... > val | _TIF_NEED_RESCHED); ... > } /* Does not send an IPI */ ... > ... } /* mwait exit due to write at addr */ > csd_lock_wait() { } > /* Waiting */ preempt_set_need_resched(); > ... __current_clr_polling(); > ... flush_smp_call_function_queue() { > ... func(); > } /* End of wait */ } > } schedule_idle() { > ... > local_irq_disable(); > smp_call_function_single(CPU1, func, wait = 1) { ... > ... ... > arch_send_call_function_single_ipi(CPU1); ... > \ ... > \ newidle_balance() { > \ ... > /* Delay */ ... > \ } > \ ... > \--------------> local_irq_enable(); > /* Processes the IPI */ > -- > > > Skipping newidle_balance() > ========================== > > In an earlier attempt to solve the challenge of the long IRQ disabled > section, newidle_balance() was skipped when a CPU waking up from idle > was found to have no runnable tasks, and was transitioning back to > idle [2]. Tim [3] and David [4] had pointed out that newidle_balance() > may be viable for CPUs that are idling with tick enabled, where the > newidle_balance() has the opportunity to pull tasks onto the idle CPU. I don't think we should be relying on this in any way shape or form. NOHZ can kill that tick at any time. Also, semantically, calling newidle from the idle thread is just daft. You're really not newly idle in that case. > Vincent [5] pointed out a case where the idle load kick will fail to > run on an idle CPU since the IPI handler launching the ILB will check > for need_resched(). In such cases, the idle CPU relies on > newidle_balance() to pull tasks towards itself. Is this the need_resched() in _nohz_idle_balance() ? Should we change this to 'need_resched() && (rq->nr_running || rq->ttwu_pending)' or something long those lines? I mean, it's fairly trivial to figure out if there really is going to be work there. > Using an alternate flag instead of NEED_RESCHED to indicate a pending > IPI was suggested as the correct approach to solve this problem on the > same thread. So adding per-arch changes for this seems like something we shouldn't unless there really is no other sane options. That is, I really think we should start with something like the below and then fix any fallout from that. diff --git a/kernel/sched/core.c b/kernel/sched/core.c index 0935f9d4bb7b..cfa45338ae97 100644 --- a/kernel/sched/core.c +++ b/kernel/sched/core.c @@ -5799,7 +5800,7 @@ static inline struct task_struct * __pick_next_task(struct rq *rq, struct task_struct *prev, struct rq_flags *rf) { const struct sched_class *class; - struct task_struct *p; + struct task_struct *p = NULL; /* * Optimization: we know that if all tasks are in the fair class we can @@ -5810,9 +5811,11 @@ __pick_next_task(struct rq *rq, struct task_struct *prev, struct rq_flags *rf) if (likely(!sched_class_above(prev->sched_class, &fair_sched_class) && rq->nr_running == rq->cfs.h_nr_running)) { - p = pick_next_task_fair(rq, prev, rf); - if (unlikely(p == RETRY_TASK)) - goto restart; + if (rq->nr_running) { + p = pick_next_task_fair(rq, prev, rf); + if (unlikely(p == RETRY_TASK)) + goto restart; + } /* Assume the next prioritized class is idle_sched_class */ if (!p) {
On Fri, 14 Jun 2024 at 11:28, Peter Zijlstra <peterz@infradead.org> wrote: > > On Thu, Jun 13, 2024 at 06:15:59PM +0000, K Prateek Nayak wrote: > > Effects of call_function_single_prep_ipi() > > ========================================== > > > > To pull a TIF_POLLING thread out of idle to process an IPI, the sender > > sets the TIF_NEED_RESCHED bit in the idle task's thread info in > > call_function_single_prep_ipi() and avoids sending an actual IPI to the > > target. As a result, the scheduler expects a task to be enqueued when > > exiting the idle path. This is not the case with non-polling idle states > > where the idle CPU exits the non-polling idle state to process the > > interrupt, and since need_resched() returns false, soon goes back to > > idle again. > > > > When TIF_NEED_RESCHED flag is set, do_idle() will call schedule_idle(), > > a large part of which runs with local IRQ disabled. In case of ipistorm, > > when measuring IPI throughput, this large IRQ disabled section delays > > processing of IPIs. Further auditing revealed that in absence of any > > runnable tasks, pick_next_task_fair(), which is called from the > > pick_next_task() fast path, will always call newidle_balance() in this > > scenario, further increasing the time spent in the IRQ disabled section. > > > > Following is the crude visualization of the problem with relevant > > functions expanded: > > -- > > CPU0 CPU1 > > ==== ==== > > do_idle() { > > __current_set_polling(); > > ... > > monitor(addr); > > if (!need_resched()) > > mwait() { > > /* Waiting */ > > smp_call_function_single(CPU1, func, wait = 1) { ... > > ... ... > > set_nr_if_polling(CPU1) { ... > > /* Realizes CPU1 is polling */ ... > > try_cmpxchg(addr, ... > > &val, ... > > val | _TIF_NEED_RESCHED); ... > > } /* Does not send an IPI */ ... > > ... } /* mwait exit due to write at addr */ > > csd_lock_wait() { } > > /* Waiting */ preempt_set_need_resched(); > > ... __current_clr_polling(); > > ... flush_smp_call_function_queue() { > > ... func(); > > } /* End of wait */ } > > } schedule_idle() { > > ... > > local_irq_disable(); > > smp_call_function_single(CPU1, func, wait = 1) { ... > > ... ... > > arch_send_call_function_single_ipi(CPU1); ... > > \ ... > > \ newidle_balance() { > > \ ... > > /* Delay */ ... > > \ } > > \ ... > > \--------------> local_irq_enable(); > > /* Processes the IPI */ > > -- > > > > > > Skipping newidle_balance() > > ========================== > > > > In an earlier attempt to solve the challenge of the long IRQ disabled > > section, newidle_balance() was skipped when a CPU waking up from idle > > was found to have no runnable tasks, and was transitioning back to > > idle [2]. Tim [3] and David [4] had pointed out that newidle_balance() > > may be viable for CPUs that are idling with tick enabled, where the > > newidle_balance() has the opportunity to pull tasks onto the idle CPU. > > I don't think we should be relying on this in any way shape or form. > NOHZ can kill that tick at any time. > > Also, semantically, calling newidle from the idle thread is just daft. > You're really not newly idle in that case. > > > Vincent [5] pointed out a case where the idle load kick will fail to > > run on an idle CPU since the IPI handler launching the ILB will check > > for need_resched(). In such cases, the idle CPU relies on > > newidle_balance() to pull tasks towards itself. > > Is this the need_resched() in _nohz_idle_balance() ? Should we change > this to 'need_resched() && (rq->nr_running || rq->ttwu_pending)' or > something long those lines? It's not only this but also in do_idle() as well which exits the loop to look for tasks to schedule > > I mean, it's fairly trivial to figure out if there really is going to be > work there. > > > Using an alternate flag instead of NEED_RESCHED to indicate a pending > > IPI was suggested as the correct approach to solve this problem on the > > same thread. > > So adding per-arch changes for this seems like something we shouldn't > unless there really is no other sane options. > > That is, I really think we should start with something like the below > and then fix any fallout from that. The main problem is that need_resched becomes somewhat meaningless because it doesn't only mean "I need to resched a task" and we have to add more tests around even for those not using polling > > diff --git a/kernel/sched/core.c b/kernel/sched/core.c > index 0935f9d4bb7b..cfa45338ae97 100644 > --- a/kernel/sched/core.c > +++ b/kernel/sched/core.c > @@ -5799,7 +5800,7 @@ static inline struct task_struct * > __pick_next_task(struct rq *rq, struct task_struct *prev, struct rq_flags *rf) > { > const struct sched_class *class; > - struct task_struct *p; > + struct task_struct *p = NULL; > > /* > * Optimization: we know that if all tasks are in the fair class we can > @@ -5810,9 +5811,11 @@ __pick_next_task(struct rq *rq, struct task_struct *prev, struct rq_flags *rf) > if (likely(!sched_class_above(prev->sched_class, &fair_sched_class) && > rq->nr_running == rq->cfs.h_nr_running)) { > > - p = pick_next_task_fair(rq, prev, rf); > - if (unlikely(p == RETRY_TASK)) > - goto restart; > + if (rq->nr_running) { How do you make the diff between a spurious need_resched() because of polling and a cpu becoming idle ? isn't rq->nr_running null in both cases ? In the later case, we need to call sched_balance_newidle() but not in the former > + p = pick_next_task_fair(rq, prev, rf); > + if (unlikely(p == RETRY_TASK)) > + goto restart; > + } > > /* Assume the next prioritized class is idle_sched_class */ > if (!p) {
On 2024-06-14 at 12:48:37 +0200, Vincent Guittot wrote: > On Fri, 14 Jun 2024 at 11:28, Peter Zijlstra <peterz@infradead.org> wrote: > > > > On Thu, Jun 13, 2024 at 06:15:59PM +0000, K Prateek Nayak wrote: > > > Effects of call_function_single_prep_ipi() > > > ========================================== > > > > > > To pull a TIF_POLLING thread out of idle to process an IPI, the sender > > > sets the TIF_NEED_RESCHED bit in the idle task's thread info in > > > call_function_single_prep_ipi() and avoids sending an actual IPI to the > > > target. As a result, the scheduler expects a task to be enqueued when > > > exiting the idle path. This is not the case with non-polling idle states > > > where the idle CPU exits the non-polling idle state to process the > > > interrupt, and since need_resched() returns false, soon goes back to > > > idle again. > > > > > > When TIF_NEED_RESCHED flag is set, do_idle() will call schedule_idle(), > > > a large part of which runs with local IRQ disabled. In case of ipistorm, > > > when measuring IPI throughput, this large IRQ disabled section delays > > > processing of IPIs. Further auditing revealed that in absence of any > > > runnable tasks, pick_next_task_fair(), which is called from the > > > pick_next_task() fast path, will always call newidle_balance() in this > > > scenario, further increasing the time spent in the IRQ disabled section. > > > > > > Following is the crude visualization of the problem with relevant > > > functions expanded: > > > -- > > > CPU0 CPU1 > > > ==== ==== > > > do_idle() { > > > __current_set_polling(); > > > ... > > > monitor(addr); > > > if (!need_resched()) > > > mwait() { > > > /* Waiting */ > > > smp_call_function_single(CPU1, func, wait = 1) { ... > > > ... ... > > > set_nr_if_polling(CPU1) { ... > > > /* Realizes CPU1 is polling */ ... > > > try_cmpxchg(addr, ... > > > &val, ... > > > val | _TIF_NEED_RESCHED); ... > > > } /* Does not send an IPI */ ... > > > ... } /* mwait exit due to write at addr */ > > > csd_lock_wait() { } > > > /* Waiting */ preempt_set_need_resched(); > > > ... __current_clr_polling(); > > > ... flush_smp_call_function_queue() { > > > ... func(); > > > } /* End of wait */ } > > > } schedule_idle() { > > > ... > > > local_irq_disable(); > > > smp_call_function_single(CPU1, func, wait = 1) { ... > > > ... ... > > > arch_send_call_function_single_ipi(CPU1); ... > > > \ ... > > > \ newidle_balance() { > > > \ ... > > > /* Delay */ ... > > > \ } > > > \ ... > > > \--------------> local_irq_enable(); > > > /* Processes the IPI */ > > > -- > > > > > > > > > Skipping newidle_balance() > > > ========================== > > > > > > In an earlier attempt to solve the challenge of the long IRQ disabled > > > section, newidle_balance() was skipped when a CPU waking up from idle > > > was found to have no runnable tasks, and was transitioning back to > > > idle [2]. Tim [3] and David [4] had pointed out that newidle_balance() > > > may be viable for CPUs that are idling with tick enabled, where the > > > newidle_balance() has the opportunity to pull tasks onto the idle CPU. > > > > I don't think we should be relying on this in any way shape or form. > > NOHZ can kill that tick at any time. > > > > Also, semantically, calling newidle from the idle thread is just daft. > > You're really not newly idle in that case. > > > > > Vincent [5] pointed out a case where the idle load kick will fail to > > > run on an idle CPU since the IPI handler launching the ILB will check > > > for need_resched(). In such cases, the idle CPU relies on > > > newidle_balance() to pull tasks towards itself. > > > > Is this the need_resched() in _nohz_idle_balance() ? Should we change > > this to 'need_resched() && (rq->nr_running || rq->ttwu_pending)' or > > something long those lines? > > It's not only this but also in do_idle() as well which exits the loop > to look for tasks to schedule > > > > > I mean, it's fairly trivial to figure out if there really is going to be > > work there. > > > > > Using an alternate flag instead of NEED_RESCHED to indicate a pending > > > IPI was suggested as the correct approach to solve this problem on the > > > same thread. > > > > So adding per-arch changes for this seems like something we shouldn't > > unless there really is no other sane options. > > > > That is, I really think we should start with something like the below > > and then fix any fallout from that. > > The main problem is that need_resched becomes somewhat meaningless > because it doesn't only mean "I need to resched a task" and we have > to add more tests around even for those not using polling > > > > > diff --git a/kernel/sched/core.c b/kernel/sched/core.c > > index 0935f9d4bb7b..cfa45338ae97 100644 > > --- a/kernel/sched/core.c > > +++ b/kernel/sched/core.c > > @@ -5799,7 +5800,7 @@ static inline struct task_struct * > > __pick_next_task(struct rq *rq, struct task_struct *prev, struct rq_flags *rf) > > { > > const struct sched_class *class; > > - struct task_struct *p; > > + struct task_struct *p = NULL; > > > > /* > > * Optimization: we know that if all tasks are in the fair class we can > > @@ -5810,9 +5811,11 @@ __pick_next_task(struct rq *rq, struct task_struct *prev, struct rq_flags *rf) > > if (likely(!sched_class_above(prev->sched_class, &fair_sched_class) && > > rq->nr_running == rq->cfs.h_nr_running)) { > > > > - p = pick_next_task_fair(rq, prev, rf); > > - if (unlikely(p == RETRY_TASK)) > > - goto restart; > > + if (rq->nr_running) { > > How do you make the diff between a spurious need_resched() because of > polling and a cpu becoming idle ? isn't rq->nr_running null in both > cases ? > In the later case, we need to call sched_balance_newidle() but not in the former > Not sure if I understand correctly, if the goal of smp_call_function_single() is to kick the idle CPU and do not force it to launch the schedule()->sched_balance_newidle(), can we set the _TIF_POLLING_NRFLAG rather than _TIF_NEED_RESCHED in set_nr_if_polling()? I think writing any value to the monitor address would wakeup the idle CPU. And _TIF_POLLING_NRFLAG will be cleared once that idle CPU exit the idle loop, so we don't introduce arch-wide flag. thanks, Chenyu > > + p = pick_next_task_fair(rq, prev, rf); > > + if (unlikely(p == RETRY_TASK)) > > + goto restart; > > + } > > > > /* Assume the next prioritized class is idle_sched_class */ > > if (!p) {
On Fri, Jun 14, 2024 at 12:48:37PM +0200, Vincent Guittot wrote: > On Fri, 14 Jun 2024 at 11:28, Peter Zijlstra <peterz@infradead.org> wrote: > > > Vincent [5] pointed out a case where the idle load kick will fail to > > > run on an idle CPU since the IPI handler launching the ILB will check > > > for need_resched(). In such cases, the idle CPU relies on > > > newidle_balance() to pull tasks towards itself. > > > > Is this the need_resched() in _nohz_idle_balance() ? Should we change > > this to 'need_resched() && (rq->nr_running || rq->ttwu_pending)' or > > something long those lines? > > It's not only this but also in do_idle() as well which exits the loop > to look for tasks to schedule Is that really a problem? Reading the initial email the problem seems to be newidle balance, not hitting schedule. Schedule should be fairly quick if there's nothing to do, no? > > I mean, it's fairly trivial to figure out if there really is going to be > > work there. > > > > > Using an alternate flag instead of NEED_RESCHED to indicate a pending > > > IPI was suggested as the correct approach to solve this problem on the > > > same thread. > > > > So adding per-arch changes for this seems like something we shouldn't > > unless there really is no other sane options. > > > > That is, I really think we should start with something like the below > > and then fix any fallout from that. > > The main problem is that need_resched becomes somewhat meaningless > because it doesn't only mean "I need to resched a task" and we have > to add more tests around even for those not using polling True, however we already had some of that by having the wakeup list, that made nr_running less 'reliable'. The thing is, most architectures seem to have the TIF_POLLING_NRFLAG bit, even if their main idle routine isn't actually using it, much of the idle loop until it hits the arch idle will be having it set and will thus tickle these cases *sometimes*. > > diff --git a/kernel/sched/core.c b/kernel/sched/core.c > > index 0935f9d4bb7b..cfa45338ae97 100644 > > --- a/kernel/sched/core.c > > +++ b/kernel/sched/core.c > > @@ -5799,7 +5800,7 @@ static inline struct task_struct * > > __pick_next_task(struct rq *rq, struct task_struct *prev, struct rq_flags *rf) > > { > > const struct sched_class *class; > > - struct task_struct *p; > > + struct task_struct *p = NULL; > > > > /* > > * Optimization: we know that if all tasks are in the fair class we can > > @@ -5810,9 +5811,11 @@ __pick_next_task(struct rq *rq, struct task_struct *prev, struct rq_flags *rf) > > if (likely(!sched_class_above(prev->sched_class, &fair_sched_class) && > > rq->nr_running == rq->cfs.h_nr_running)) { > > > > - p = pick_next_task_fair(rq, prev, rf); > > - if (unlikely(p == RETRY_TASK)) > > - goto restart; > > + if (rq->nr_running) { > > How do you make the diff between a spurious need_resched() because of > polling and a cpu becoming idle ? isn't rq->nr_running null in both > cases ? Bah, true. It should also check current being idle, which then makes a mess of things again. Still, we shouldn't be calling newidle from idle, that's daft. I should probably not write code at 3am, but the below horror is what I came up with. --- diff --git a/kernel/sched/core.c b/kernel/sched/core.c index 0935f9d4bb7b..cfe8d3350819 100644 --- a/kernel/sched/core.c +++ b/kernel/sched/core.c @@ -6343,19 +6344,12 @@ pick_next_task(struct rq *rq, struct task_struct *prev, struct rq_flags *rf) * Constants for the sched_mode argument of __schedule(). * * The mode argument allows RT enabled kernels to differentiate a - * preemption from blocking on an 'sleeping' spin/rwlock. Note that - * SM_MASK_PREEMPT for !RT has all bits set, which allows the compiler to - * optimize the AND operation out and just check for zero. + * preemption from blocking on an 'sleeping' spin/rwlock. */ -#define SM_NONE 0x0 -#define SM_PREEMPT 0x1 -#define SM_RTLOCK_WAIT 0x2 - -#ifndef CONFIG_PREEMPT_RT -# define SM_MASK_PREEMPT (~0U) -#else -# define SM_MASK_PREEMPT SM_PREEMPT -#endif +#define SM_IDLE (-1) +#define SM_NONE 0 +#define SM_PREEMPT 1 +#define SM_RTLOCK_WAIT 2 /* * __schedule() is the main scheduler function. @@ -6396,11 +6390,12 @@ pick_next_task(struct rq *rq, struct task_struct *prev, struct rq_flags *rf) * * WARNING: must be called with preemption disabled! */ -static void __sched notrace __schedule(unsigned int sched_mode) +static void __sched notrace __schedule(int sched_mode) { struct task_struct *prev, *next; unsigned long *switch_count; unsigned long prev_state; + bool preempt = sched_mode > 0; struct rq_flags rf; struct rq *rq; int cpu; @@ -6409,13 +6404,13 @@ static void __sched notrace __schedule(unsigned int sched_mode) rq = cpu_rq(cpu); prev = rq->curr; - schedule_debug(prev, !!sched_mode); + schedule_debug(prev, preempt); if (sched_feat(HRTICK) || sched_feat(HRTICK_DL)) hrtick_clear(rq); local_irq_disable(); - rcu_note_context_switch(!!sched_mode); + rcu_note_context_switch(preempt); /* * Make sure that signal_pending_state()->signal_pending() below @@ -6449,7 +6444,12 @@ static void __sched notrace __schedule(unsigned int sched_mode) * that we form a control dependency vs deactivate_task() below. */ prev_state = READ_ONCE(prev->__state); - if (!(sched_mode & SM_MASK_PREEMPT) && prev_state) { + if (sched_mode == SM_IDLE) { + if (!rq->nr_running) { + next = prev; + goto picked; + } + } else if (!preempt && prev_state) { if (signal_pending_state(prev_state, prev)) { WRITE_ONCE(prev->__state, TASK_RUNNING); } else { @@ -6483,6 +6483,7 @@ static void __sched notrace __schedule(unsigned int sched_mode) } next = pick_next_task(rq, prev, &rf); +picked: clear_tsk_need_resched(prev); clear_preempt_need_resched(); #ifdef CONFIG_SCHED_DEBUG @@ -6521,9 +6522,9 @@ static void __sched notrace __schedule(unsigned int sched_mode) ++*switch_count; migrate_disable_switch(rq, prev); psi_sched_switch(prev, next, !task_on_rq_queued(prev)); - trace_sched_switch(sched_mode & SM_MASK_PREEMPT, prev, next, prev_state); + trace_sched_switch(preempt, prev, next, prev_state); /* Also unlocks the rq: */ rq = context_switch(rq, prev, next, &rf); @@ -6599,7 +6601,7 @@ static void sched_update_worker(struct task_struct *tsk) } } -static __always_inline void __schedule_loop(unsigned int sched_mode) +static __always_inline void __schedule_loop(int sched_mode) { do { preempt_disable(); @@ -6644,7 +6646,7 @@ void __sched schedule_idle(void) */ WARN_ON_ONCE(current->__state); do { - __schedule(SM_NONE); + __schedule(SM_IDLE); } while (need_resched()); }
On Fri, Jun 14, 2024 at 12:48:37PM +0200, Vincent Guittot wrote: > The main problem is that need_resched becomes somewhat meaningless > because it doesn't only mean "I need to resched a task" and we have > to add more tests around even for those not using polling The converse problem is that you're adding a bunch of atomic ops that might be avoided. It might now need to set both the RESCHED and IPI flags -- and clear them again.
On Sat, Jun 15, 2024 at 03:28:14AM +0200, Peter Zijlstra wrote: > On Fri, Jun 14, 2024 at 12:48:37PM +0200, Vincent Guittot wrote: > > The main problem is that need_resched becomes somewhat meaningless > > because it doesn't only mean "I need to resched a task" and we have > > to add more tests around even for those not using polling > > True, however we already had some of that by having the wakeup list, > that made nr_running less 'reliable'. Doesn't using !idle_cpu() instead of need_resched() in those balance paths already do the right thing? Checking need_resched() as an indicator of it getting work is already a bit an assumption. Also, Ingo, idle_cpu() and friends don't really belong in syscalls.c...
On Thu, Jun 13, 2024 at 06:15:59PM +0000, K Prateek Nayak wrote: > o Dropping the ARM results since I never got my hands on the ARM64 > system I used in my last testing. If I do manage to get my hands on it > again, I'll rerun the experiments and share the results on the thread. > To test the case where TIF_NOTIFY_IPI is not enabled for a particular > architecture, I applied the series only until Patch 3 and tested the > same on my x86 machine with a WARN_ON_ONCE() in do_idle() to check if > tif_notify_ipi() ever return true and then repeated the same with > Patch 4 applied. Confused. ARM (32-bit) or ARM64? You patch 32-bit ARM, but you don't touch 64-bit Arm. "ARM" on its own in the context above to me suggests 32-bit, since you refer to ARM64 later.
On Fri, Jun 14, 2024 at 2:21 AM K Prateek Nayak <kprateek.nayak@amd.com> wrote: > > Add support for TIF_NOTIFY_IPI on C-SKY. With TIF_NOTIFY_IPI, a sender > sending an IPI to an idle CPU in TIF_POLLING mode will set the > TIF_NOTIFY_IPI flag in the target's idle tasks's thread_info to pull the > CPU out of idle, as opposed to setting TIF_NEED_RESCHED previously. This > avoids spurious calls to schedule_idle() in cases where an IPI does not > necessarily wake up a task on the idle CPU. > > Cc: Guo Ren <guoren@kernel.org> > Cc: "Rafael J. Wysocki" <rafael@kernel.org> > Cc: Daniel Lezcano <daniel.lezcano@linaro.org> > Cc: Ingo Molnar <mingo@redhat.com> > Cc: Peter Zijlstra <peterz@infradead.org> > Cc: Juri Lelli <juri.lelli@redhat.com> > Cc: Vincent Guittot <vincent.guittot@linaro.org> > Cc: Dietmar Eggemann <dietmar.eggemann@arm.com> > Cc: Steven Rostedt <rostedt@goodmis.org> > Cc: Ben Segall <bsegall@google.com> > Cc: Mel Gorman <mgorman@suse.de> > Cc: Daniel Bristot de Oliveira <bristot@redhat.com> > Cc: Valentin Schneider <vschneid@redhat.com> > Cc: K Prateek Nayak <kprateek.nayak@amd.com> > Cc: linux-csky@vger.kernel.org > Cc: linux-kernel@vger.kernel.org > Cc: linux-pm@vger.kernel.org > Acked-by: Guo Ren <guoren@kernel.org> > Signed-off-by: K Prateek Nayak <kprateek.nayak@amd.com> > --- > v1..v2: > o Collected the ack. Thank you :) > --- > arch/csky/include/asm/thread_info.h | 2 ++ > 1 file changed, 2 insertions(+) > > diff --git a/arch/csky/include/asm/thread_info.h b/arch/csky/include/asm/thread_info.h > index b5ed788f0c68..9bc7a037c476 100644 > --- a/arch/csky/include/asm/thread_info.h > +++ b/arch/csky/include/asm/thread_info.h > @@ -61,6 +61,7 @@ static inline struct thread_info *current_thread_info(void) > #define TIF_SYSCALL_TRACEPOINT 5 /* syscall tracepoint instrumentation */ > #define TIF_SYSCALL_AUDIT 6 /* syscall auditing */ > #define TIF_NOTIFY_SIGNAL 7 /* signal notifications exist */ > +#define TIF_NOTIFY_IPI 8 /* Pending IPI on TIF_POLLLING idle CPU */ > #define TIF_POLLING_NRFLAG 16 /* poll_idle() is TIF_NEED_RESCHED */ > #define TIF_MEMDIE 18 /* is terminating due to OOM killer */ > #define TIF_RESTORE_SIGMASK 20 /* restore signal mask in do_signal() */ > @@ -73,6 +74,7 @@ static inline struct thread_info *current_thread_info(void) > #define _TIF_SYSCALL_TRACEPOINT (1 << TIF_SYSCALL_TRACEPOINT) > #define _TIF_SYSCALL_AUDIT (1 << TIF_SYSCALL_AUDIT) > #define _TIF_NOTIFY_SIGNAL (1 << TIF_NOTIFY_SIGNAL) > +#define _TIF_NOTIFY_IPI (1 << TIF_NOTIFY_IPI) Acked-by: Guo Ren <guoren@kernel.org> > #define _TIF_UPROBE (1 << TIF_UPROBE) > #define _TIF_POLLING_NRFLAG (1 << TIF_POLLING_NRFLAG) > #define _TIF_MEMDIE (1 << TIF_MEMDIE) > -- > 2.34.1 >
Hello Chenyu, On 6/14/2024 10:01 PM, Chen Yu wrote: > On 2024-06-14 at 12:48:37 +0200, Vincent Guittot wrote: >> On Fri, 14 Jun 2024 at 11:28, Peter Zijlstra <peterz@infradead.org> wrote: >>> >>> On Thu, Jun 13, 2024 at 06:15:59PM +0000, K Prateek Nayak wrote: >>>> Effects of call_function_single_prep_ipi() >>>> ========================================== >>>> >>>> To pull a TIF_POLLING thread out of idle to process an IPI, the sender >>>> sets the TIF_NEED_RESCHED bit in the idle task's thread info in >>>> call_function_single_prep_ipi() and avoids sending an actual IPI to the >>>> target. As a result, the scheduler expects a task to be enqueued when >>>> exiting the idle path. This is not the case with non-polling idle states >>>> where the idle CPU exits the non-polling idle state to process the >>>> interrupt, and since need_resched() returns false, soon goes back to >>>> idle again. >>>> >>>> When TIF_NEED_RESCHED flag is set, do_idle() will call schedule_idle(), >>>> a large part of which runs with local IRQ disabled. In case of ipistorm, >>>> when measuring IPI throughput, this large IRQ disabled section delays >>>> processing of IPIs. Further auditing revealed that in absence of any >>>> runnable tasks, pick_next_task_fair(), which is called from the >>>> pick_next_task() fast path, will always call newidle_balance() in this >>>> scenario, further increasing the time spent in the IRQ disabled section. >>>> >>>> Following is the crude visualization of the problem with relevant >>>> functions expanded: >>>> -- >>>> CPU0 CPU1 >>>> ==== ==== >>>> do_idle() { >>>> __current_set_polling(); >>>> ... >>>> monitor(addr); >>>> if (!need_resched()) >>>> mwait() { >>>> /* Waiting */ >>>> smp_call_function_single(CPU1, func, wait = 1) { ... >>>> ... ... >>>> set_nr_if_polling(CPU1) { ... >>>> /* Realizes CPU1 is polling */ ... >>>> try_cmpxchg(addr, ... >>>> &val, ... >>>> val | _TIF_NEED_RESCHED); ... >>>> } /* Does not send an IPI */ ... >>>> ... } /* mwait exit due to write at addr */ >>>> csd_lock_wait() { } >>>> /* Waiting */ preempt_set_need_resched(); >>>> ... __current_clr_polling(); >>>> ... flush_smp_call_function_queue() { >>>> ... func(); >>>> } /* End of wait */ } >>>> } schedule_idle() { >>>> ... >>>> local_irq_disable(); >>>> smp_call_function_single(CPU1, func, wait = 1) { ... >>>> ... ... >>>> arch_send_call_function_single_ipi(CPU1); ... >>>> \ ... >>>> \ newidle_balance() { >>>> \ ... >>>> /* Delay */ ... >>>> \ } >>>> \ ... >>>> \--------------> local_irq_enable(); >>>> /* Processes the IPI */ >>>> -- >>>> >>>> >>>> Skipping newidle_balance() >>>> ========================== >>>> >>>> In an earlier attempt to solve the challenge of the long IRQ disabled >>>> section, newidle_balance() was skipped when a CPU waking up from idle >>>> was found to have no runnable tasks, and was transitioning back to >>>> idle [2]. Tim [3] and David [4] had pointed out that newidle_balance() >>>> may be viable for CPUs that are idling with tick enabled, where the >>>> newidle_balance() has the opportunity to pull tasks onto the idle CPU. >>> >>> I don't think we should be relying on this in any way shape or form. >>> NOHZ can kill that tick at any time. >>> >>> Also, semantically, calling newidle from the idle thread is just daft. >>> You're really not newly idle in that case. >>> >>>> Vincent [5] pointed out a case where the idle load kick will fail to >>>> run on an idle CPU since the IPI handler launching the ILB will check >>>> for need_resched(). In such cases, the idle CPU relies on >>>> newidle_balance() to pull tasks towards itself. >>> >>> Is this the need_resched() in _nohz_idle_balance() ? Should we change >>> this to 'need_resched() && (rq->nr_running || rq->ttwu_pending)' or >>> something long those lines? >> >> It's not only this but also in do_idle() as well which exits the loop >> to look for tasks to schedule >> >>> >>> I mean, it's fairly trivial to figure out if there really is going to be >>> work there. >>> >>>> Using an alternate flag instead of NEED_RESCHED to indicate a pending >>>> IPI was suggested as the correct approach to solve this problem on the >>>> same thread. >>> >>> So adding per-arch changes for this seems like something we shouldn't >>> unless there really is no other sane options. >>> >>> That is, I really think we should start with something like the below >>> and then fix any fallout from that. >> >> The main problem is that need_resched becomes somewhat meaningless >> because it doesn't only mean "I need to resched a task" and we have >> to add more tests around even for those not using polling >> >>> >>> diff --git a/kernel/sched/core.c b/kernel/sched/core.c >>> index 0935f9d4bb7b..cfa45338ae97 100644 >>> --- a/kernel/sched/core.c >>> +++ b/kernel/sched/core.c >>> @@ -5799,7 +5800,7 @@ static inline struct task_struct * >>> __pick_next_task(struct rq *rq, struct task_struct *prev, struct rq_flags *rf) >>> { >>> const struct sched_class *class; >>> - struct task_struct *p; >>> + struct task_struct *p = NULL; >>> >>> /* >>> * Optimization: we know that if all tasks are in the fair class we can >>> @@ -5810,9 +5811,11 @@ __pick_next_task(struct rq *rq, struct task_struct *prev, struct rq_flags *rf) >>> if (likely(!sched_class_above(prev->sched_class, &fair_sched_class) && >>> rq->nr_running == rq->cfs.h_nr_running)) { >>> >>> - p = pick_next_task_fair(rq, prev, rf); >>> - if (unlikely(p == RETRY_TASK)) >>> - goto restart; >>> + if (rq->nr_running) { >> >> How do you make the diff between a spurious need_resched() because of >> polling and a cpu becoming idle ? isn't rq->nr_running null in both >> cases ? >> In the later case, we need to call sched_balance_newidle() but not in the former >> > > Not sure if I understand correctly, if the goal of smp_call_function_single() is to > kick the idle CPU and do not force it to launch the schedule()->sched_balance_newidle(), > can we set the _TIF_POLLING_NRFLAG rather than _TIF_NEED_RESCHED in set_nr_if_polling()? > I think writing any value to the monitor address would wakeup the idle CPU. And _TIF_POLLING_NRFLAG > will be cleared once that idle CPU exit the idle loop, so we don't introduce arch-wide flag. Although this might work for MWAIT, there is no way for the generic idle path to know if there is a pending interrupt within a TIF_POLLING_NRFLAG section. do_idle() sets TIF_POLLING_NRFLAG and relies on a bunch of need_resched() checks along the way to bail early until finally doing a current_clr_polling_and_test() before handing off to the cpuidle driver in call_cpuidle(). I believe this section will necessarily need the sender to indicate a pending interrupt via TIF_NEED_RESCHED flag to enable the early bail out before going into the cpuidle driver since this case cannot be considered the same as a break from MWAIT. On x86, there seems to be a possibility of missing an interrupt if someone writes _TIF_POLLING_NRFLAG to thread info between the target executing MONTOR and MWAIT. AMD64 Architecture Programmer’s Manual Volume 3: "General-Purpose and System Instructions", Chapter 4. "System Instruction Reference", section "MWAIT" carries the following note in the coding requirements: "MWAIT must be conditionally executed only if the awaited store has not already occurred. (This prevents a race condition between the MONITOR instruction arming the monitoring hardware and the store intended to trigger the monitoring hardware.)" There exists a similar note in the "Example" section for "MWAIT" in Intel 64 and IA-32 Architectures Software Developer’s Manual, Vol 2B Chapter 4.3 "Instructions (M-U)" I'm not sure if one can use use _TIF_POLLING_NRFLAG alone and cover all the cases but there might be some clever trick to make it all work :) > > thanks, > Chenyu > >>> + p = pick_next_task_fair(rq, prev, rf); >>> + if (unlikely(p == RETRY_TASK)) >>> + goto restart; >>> + } >>> >>> /* Assume the next prioritized class is idle_sched_class */ >>> if (!p) {
Hello Chenyu, On 6/18/2024 1:19 PM, Chen Yu wrote: > [..snip..] >>>>> >>>>>> Vincent [5] pointed out a case where the idle load kick will fail to >>>>>> run on an idle CPU since the IPI handler launching the ILB will check >>>>>> for need_resched(). In such cases, the idle CPU relies on >>>>>> newidle_balance() to pull tasks towards itself. >>>>> >>>>> Is this the need_resched() in _nohz_idle_balance() ? Should we change >>>>> this to 'need_resched() && (rq->nr_running || rq->ttwu_pending)' or >>>>> something long those lines? >>>> >>>> It's not only this but also in do_idle() as well which exits the loop >>>> to look for tasks to schedule >>>> >>>>> >>>>> I mean, it's fairly trivial to figure out if there really is going to be >>>>> work there. >>>>> >>>>>> Using an alternate flag instead of NEED_RESCHED to indicate a pending >>>>>> IPI was suggested as the correct approach to solve this problem on the >>>>>> same thread. >>>>> >>>>> So adding per-arch changes for this seems like something we shouldn't >>>>> unless there really is no other sane options. >>>>> >>>>> That is, I really think we should start with something like the below >>>>> and then fix any fallout from that. >>>> >>>> The main problem is that need_resched becomes somewhat meaningless >>>> because it doesn't only mean "I need to resched a task" and we have >>>> to add more tests around even for those not using polling >>>> >>>>> >>>>> diff --git a/kernel/sched/core.c b/kernel/sched/core.c >>>>> index 0935f9d4bb7b..cfa45338ae97 100644 >>>>> --- a/kernel/sched/core.c >>>>> +++ b/kernel/sched/core.c >>>>> @@ -5799,7 +5800,7 @@ static inline struct task_struct * >>>>> __pick_next_task(struct rq *rq, struct task_struct *prev, struct rq_flags *rf) >>>>> { >>>>> const struct sched_class *class; >>>>> - struct task_struct *p; >>>>> + struct task_struct *p = NULL; >>>>> >>>>> /* >>>>> * Optimization: we know that if all tasks are in the fair class we can >>>>> @@ -5810,9 +5811,11 @@ __pick_next_task(struct rq *rq, struct task_struct *prev, struct rq_flags *rf) >>>>> if (likely(!sched_class_above(prev->sched_class, &fair_sched_class) && >>>>> rq->nr_running == rq->cfs.h_nr_running)) { >>>>> >>>>> - p = pick_next_task_fair(rq, prev, rf); >>>>> - if (unlikely(p == RETRY_TASK)) >>>>> - goto restart; >>>>> + if (rq->nr_running) { >>>> >>>> How do you make the diff between a spurious need_resched() because of >>>> polling and a cpu becoming idle ? isn't rq->nr_running null in both >>>> cases ? >>>> In the later case, we need to call sched_balance_newidle() but not in the former >>>> >>> >>> Not sure if I understand correctly, if the goal of smp_call_function_single() is to >>> kick the idle CPU and do not force it to launch the schedule()->sched_balance_newidle(), >>> can we set the _TIF_POLLING_NRFLAG rather than _TIF_NEED_RESCHED in set_nr_if_polling()? >>> I think writing any value to the monitor address would wakeup the idle CPU. And _TIF_POLLING_NRFLAG >>> will be cleared once that idle CPU exit the idle loop, so we don't introduce arch-wide flag. >> Although this might work for MWAIT, there is no way for the generic idle >> path to know if there is a pending interrupt within a TIF_POLLING_NRFLAG >> section. do_idle() sets TIF_POLLING_NRFLAG and relies on a bunch of >> need_resched() checks along the way to bail early until finally doing a >> current_clr_polling_and_test() before handing off to the cpuidle driver >> in call_cpuidle(). I believe this section will necessarily need the sender >> to indicate a pending interrupt via TIF_NEED_RESCHED flag to enable the >> early bail out before going into the cpuidle driver since this case cannot >> be considered the same as a break from MWAIT. >> > > I see, this is a good point. So you mean with only TIF_POLLING_NRFLAG there is > possibility that the 'ipi kick CPU out of idle' is lost after the CPU enters > do_idle() and before finally entering the idle state. While setting _TIF_NEED_RESCHED > could help the do_idle() loop to detect pending request easier. Yup, that is correct. > BTW, before the > commit b2a02fc43a1f ("smp: Optimize send_call_function_single_ipi()"), the > lost of ipi after entering do_idle() and before entering driver idle state > is also possible, right(the local irq is disabled)? From what I understand, the IPI remains pending until the interrupts are enabled again. Before the optimization, the interrupts would be disabled all the way until the instruction that is used to put the CPU to sleep which is what __sti_mwait() and native_safe_halt() does. The CPU would have received the IPI then and broke out of idle before Peter's optimization went in. There is an elaborate comment on this in do_idle() function above the call to local_irq_disable(). In commit edc8fc01f608 ("x86: Fix CPUIDLE_FLAG_IRQ_ENABLE leaking timer reprogram") Peter describes a case of actually missing the break from an interrupt as the driver enabled interrupts much earlier than executing the sleep instruction. Since the CPU was in TIF_POLLING_NRFLAG state, one could simply get away by setting TIF_NEED_RESCHED and not sending an actual IPI which the need_resched() checks in the idle path would catch and the flush_smp_call_function_queue() on the exit path would have serviced the call function. MWAIT with Interrupt Break extension (CPUID 0x5 ECX[IBE]) can break out on pending interrupts even if interrupts are disabled which is why "mwait_idle_with_hints()" now checks "ecx" to choose between "__mwait()" and "__mwait_sti()". The APM describes the extension to "allows interrupts to wake MWAIT, even when eFLAGS.IF = 0". (Vol. 3. "General-Purpose and System Instructions", Chapter 4. "System Instruction Reference", Section "MWAIT") I do hope someone corrects me if I'm wrong :) > >> On x86, there seems to be a possibility of missing an interrupt if >> someone writes _TIF_POLLING_NRFLAG to thread info between the target >> executing MONTOR and MWAIT. AMD64 Architecture Programmer’s Manual >> Volume 3: "General-Purpose and System Instructions", Chapter 4. "System >> Instruction Reference", section "MWAIT" carries the following note in >> the coding requirements: >> >> "MWAIT must be conditionally executed only if the awaited store has not >> already occurred. (This prevents a race condition between the MONITOR >> instruction arming the monitoring hardware and the store intended to >> trigger the monitoring hardware.)" >> >> There exists a similar note in the "Example" section for "MWAIT" in >> Intel 64 and IA-32 Architectures Software Developer’s Manual, Vol 2B >> Chapter 4.3 "Instructions (M-U)" >> > > Thanks for the explaination of this race condition in detail. > > thanks, > Chenyu
On 2024-06-19 at 00:03:30 +0530, K Prateek Nayak wrote: > Hello Chenyu, > > On 6/18/2024 1:19 PM, Chen Yu wrote: > > [..snip..] > > > > > > > > > > > > > Vincent [5] pointed out a case where the idle load kick will fail to > > > > > > > run on an idle CPU since the IPI handler launching the ILB will check > > > > > > > for need_resched(). In such cases, the idle CPU relies on > > > > > > > newidle_balance() to pull tasks towards itself. > > > > > > > > > > > > Is this the need_resched() in _nohz_idle_balance() ? Should we change > > > > > > this to 'need_resched() && (rq->nr_running || rq->ttwu_pending)' or > > > > > > something long those lines? > > > > > > > > > > It's not only this but also in do_idle() as well which exits the loop > > > > > to look for tasks to schedule > > > > > > > > > > > > > > > > > I mean, it's fairly trivial to figure out if there really is going to be > > > > > > work there. > > > > > > > > > > > > > Using an alternate flag instead of NEED_RESCHED to indicate a pending > > > > > > > IPI was suggested as the correct approach to solve this problem on the > > > > > > > same thread. > > > > > > > > > > > > So adding per-arch changes for this seems like something we shouldn't > > > > > > unless there really is no other sane options. > > > > > > > > > > > > That is, I really think we should start with something like the below > > > > > > and then fix any fallout from that. > > > > > > > > > > The main problem is that need_resched becomes somewhat meaningless > > > > > because it doesn't only mean "I need to resched a task" and we have > > > > > to add more tests around even for those not using polling > > > > > > > > > > > > > > > > > diff --git a/kernel/sched/core.c b/kernel/sched/core.c > > > > > > index 0935f9d4bb7b..cfa45338ae97 100644 > > > > > > --- a/kernel/sched/core.c > > > > > > +++ b/kernel/sched/core.c > > > > > > @@ -5799,7 +5800,7 @@ static inline struct task_struct * > > > > > > __pick_next_task(struct rq *rq, struct task_struct *prev, struct rq_flags *rf) > > > > > > { > > > > > > const struct sched_class *class; > > > > > > - struct task_struct *p; > > > > > > + struct task_struct *p = NULL; > > > > > > > > > > > > /* > > > > > > * Optimization: we know that if all tasks are in the fair class we can > > > > > > @@ -5810,9 +5811,11 @@ __pick_next_task(struct rq *rq, struct task_struct *prev, struct rq_flags *rf) > > > > > > if (likely(!sched_class_above(prev->sched_class, &fair_sched_class) && > > > > > > rq->nr_running == rq->cfs.h_nr_running)) { > > > > > > > > > > > > - p = pick_next_task_fair(rq, prev, rf); > > > > > > - if (unlikely(p == RETRY_TASK)) > > > > > > - goto restart; > > > > > > + if (rq->nr_running) { > > > > > > > > > > How do you make the diff between a spurious need_resched() because of > > > > > polling and a cpu becoming idle ? isn't rq->nr_running null in both > > > > > cases ? > > > > > In the later case, we need to call sched_balance_newidle() but not in the former > > > > > > > > > > > > > Not sure if I understand correctly, if the goal of smp_call_function_single() is to > > > > kick the idle CPU and do not force it to launch the schedule()->sched_balance_newidle(), > > > > can we set the _TIF_POLLING_NRFLAG rather than _TIF_NEED_RESCHED in set_nr_if_polling()? > > > > I think writing any value to the monitor address would wakeup the idle CPU. And _TIF_POLLING_NRFLAG > > > > will be cleared once that idle CPU exit the idle loop, so we don't introduce arch-wide flag. > > > Although this might work for MWAIT, there is no way for the generic idle > > > path to know if there is a pending interrupt within a TIF_POLLING_NRFLAG > > > section. do_idle() sets TIF_POLLING_NRFLAG and relies on a bunch of > > > need_resched() checks along the way to bail early until finally doing a > > > current_clr_polling_and_test() before handing off to the cpuidle driver > > > in call_cpuidle(). I believe this section will necessarily need the sender > > > to indicate a pending interrupt via TIF_NEED_RESCHED flag to enable the > > > early bail out before going into the cpuidle driver since this case cannot > > > be considered the same as a break from MWAIT. > > > > > > > I see, this is a good point. So you mean with only TIF_POLLING_NRFLAG there is > > possibility that the 'ipi kick CPU out of idle' is lost after the CPU enters > > do_idle() and before finally entering the idle state. While setting _TIF_NEED_RESCHED > > could help the do_idle() loop to detect pending request easier. > > Yup, that is correct. > > > BTW, before the > > commit b2a02fc43a1f ("smp: Optimize send_call_function_single_ipi()"), the > > lost of ipi after entering do_idle() and before entering driver idle state > > is also possible, right(the local irq is disabled)? > > From what I understand, the IPI remains pending until the interrupts > are enabled again. Before the optimization, the interrupts would be > disabled all the way until the instruction that is used to put the CPU > to sleep which is what __sti_mwait() and native_safe_halt() does. The > CPU would have received the IPI then and broke out of idle before > Peter's optimization went in. I see, once local irq is enabled, the pending ipi will be served. > There is an elaborate comment on this in > do_idle() function above the call to local_irq_disable(). In commit > edc8fc01f608 ("x86: Fix CPUIDLE_FLAG_IRQ_ENABLE leaking timer > reprogram") Peter describes a case of actually missing the break from > an interrupt as the driver enabled interrupts much earlier than > executing the sleep instruction. > Yup, the commit edc8fc01f608 deals with delay of the timer handling. If a timer queues the callback after local irq enabled and before mwait, the long sleep time after mwait might delay the handling of the callback. > Since the CPU was in TIF_POLLING_NRFLAG state, one could simply get away > by setting TIF_NEED_RESCHED and not sending an actual IPI which the > need_resched() checks in the idle path would catch and the > flush_smp_call_function_queue() on the exit path would have serviced the > call function. > > MWAIT with Interrupt Break extension (CPUID 0x5 ECX[IBE]) can break out > on pending interrupts even if interrupts are disabled which is why > "mwait_idle_with_hints()" now checks "ecx" to choose between "__mwait()" > and "__mwait_sti()". The APM describes the extension to "allows > interrupts to wake MWAIT, even when eFLAGS.IF = 0". (Vol. 3. > "General-Purpose and System Instructions", Chapter 4. "System Instruction > Reference", Section "MWAIT") > > I do hope someone corrects me if I'm wrong :) > You are right, and thanks for the description. thanks, Chenyu
diff --git a/kernel/sched/idle.c b/kernel/sched/idle.c index 31231925f1ec..735184d98c0f 100644 --- a/kernel/sched/idle.c +++ b/kernel/sched/idle.c @@ -332,11 +332,6 @@ static void do_idle(void) */ smp_mb__after_atomic(); - /* - * RCU relies on this call to be done outside of an RCU read-side - * critical section. - */ - flush_smp_call_function_queue(); schedule_idle(); if (unlikely(klp_patch_pending(current))) diff --git a/kernel/smp.c b/kernel/smp.c index f085ebcdf9e7..2ff100c41885 100644 --- a/kernel/smp.c +++ b/kernel/smp.c @@ -111,11 +111,9 @@ void __init call_function_init(void) static __always_inline void send_call_function_single_ipi(int cpu) { - if (call_function_single_prep_ipi(cpu)) { - trace_ipi_send_cpu(cpu, _RET_IP_, - generic_smp_call_function_single_interrupt); - arch_send_call_function_single_ipi(cpu); - } + trace_ipi_send_cpu(cpu, _RET_IP_, + generic_smp_call_function_single_interrupt); + arch_send_call_function_single_ipi(cpu); } static __always_inline void