mirror of
https://github.com/torvalds/linux.git
synced 2025-12-07 20:06:24 +00:00
sched: Add an initial sketch of the find_proxy_task() function
Add a find_proxy_task() function which doesn't do much. When we select a blocked task to run, we will just deactivate it and pick again. The exception being if it has become unblocked after find_proxy_task() was called. This allows us to validate keeping blocked tasks on the runqueue and later deactivating them is working ok, stressing the failure cases for when a proxy isn't found. Greatly simplified from patch by: Peter Zijlstra (Intel) <peterz@infradead.org> Juri Lelli <juri.lelli@redhat.com> Valentin Schneider <valentin.schneider@arm.com> Connor O'Brien <connoro@google.com> [jstultz: Split out from larger proxy patch and simplified for review and testing.] Signed-off-by: John Stultz <jstultz@google.com> Signed-off-by: Peter Zijlstra (Intel) <peterz@infradead.org> Tested-by: K Prateek Nayak <kprateek.nayak@amd.com> Link: https://lkml.kernel.org/r/20250712033407.2383110-7-jstultz@google.com
This commit is contained in:
committed by
Peter Zijlstra
parent
aa4f74dfd4
commit
be41bde4c3
@@ -6522,11 +6522,13 @@ pick_next_task(struct rq *rq, struct task_struct *prev, struct rq_flags *rf)
|
|||||||
/*
|
/*
|
||||||
* Helper function for __schedule()
|
* Helper function for __schedule()
|
||||||
*
|
*
|
||||||
* If a task does not have signals pending, deactivate it
|
* Tries to deactivate the task, unless the should_block arg
|
||||||
* Otherwise marks the task's __state as RUNNING
|
* is false or if a signal is pending. In the case a signal
|
||||||
|
* is pending, marks the task's __state as RUNNING (and clear
|
||||||
|
* blocked_on).
|
||||||
*/
|
*/
|
||||||
static bool try_to_block_task(struct rq *rq, struct task_struct *p,
|
static bool try_to_block_task(struct rq *rq, struct task_struct *p,
|
||||||
unsigned long *task_state_p)
|
unsigned long *task_state_p, bool should_block)
|
||||||
{
|
{
|
||||||
unsigned long task_state = *task_state_p;
|
unsigned long task_state = *task_state_p;
|
||||||
int flags = DEQUEUE_NOCLOCK;
|
int flags = DEQUEUE_NOCLOCK;
|
||||||
@@ -6537,6 +6539,16 @@ static bool try_to_block_task(struct rq *rq, struct task_struct *p,
|
|||||||
return false;
|
return false;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/*
|
||||||
|
* We check should_block after signal_pending because we
|
||||||
|
* will want to wake the task in that case. But if
|
||||||
|
* should_block is false, its likely due to the task being
|
||||||
|
* blocked on a mutex, and we want to keep it on the runqueue
|
||||||
|
* to be selectable for proxy-execution.
|
||||||
|
*/
|
||||||
|
if (!should_block)
|
||||||
|
return false;
|
||||||
|
|
||||||
p->sched_contributes_to_load =
|
p->sched_contributes_to_load =
|
||||||
(task_state & TASK_UNINTERRUPTIBLE) &&
|
(task_state & TASK_UNINTERRUPTIBLE) &&
|
||||||
!(task_state & TASK_NOLOAD) &&
|
!(task_state & TASK_NOLOAD) &&
|
||||||
@@ -6560,6 +6572,88 @@ static bool try_to_block_task(struct rq *rq, struct task_struct *p,
|
|||||||
return true;
|
return true;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
#ifdef CONFIG_SCHED_PROXY_EXEC
|
||||||
|
static inline void proxy_resched_idle(struct rq *rq)
|
||||||
|
{
|
||||||
|
put_prev_set_next_task(rq, rq->donor, rq->idle);
|
||||||
|
rq_set_donor(rq, rq->idle);
|
||||||
|
set_tsk_need_resched(rq->idle);
|
||||||
|
}
|
||||||
|
|
||||||
|
static bool __proxy_deactivate(struct rq *rq, struct task_struct *donor)
|
||||||
|
{
|
||||||
|
unsigned long state = READ_ONCE(donor->__state);
|
||||||
|
|
||||||
|
/* Don't deactivate if the state has been changed to TASK_RUNNING */
|
||||||
|
if (state == TASK_RUNNING)
|
||||||
|
return false;
|
||||||
|
/*
|
||||||
|
* Because we got donor from pick_next_task(), it is *crucial*
|
||||||
|
* that we call proxy_resched_idle() before we deactivate it.
|
||||||
|
* As once we deactivate donor, donor->on_rq is set to zero,
|
||||||
|
* which allows ttwu() to immediately try to wake the task on
|
||||||
|
* another rq. So we cannot use *any* references to donor
|
||||||
|
* after that point. So things like cfs_rq->curr or rq->donor
|
||||||
|
* need to be changed from next *before* we deactivate.
|
||||||
|
*/
|
||||||
|
proxy_resched_idle(rq);
|
||||||
|
return try_to_block_task(rq, donor, &state, true);
|
||||||
|
}
|
||||||
|
|
||||||
|
static struct task_struct *proxy_deactivate(struct rq *rq, struct task_struct *donor)
|
||||||
|
{
|
||||||
|
if (!__proxy_deactivate(rq, donor)) {
|
||||||
|
/*
|
||||||
|
* XXX: For now, if deactivation failed, set donor
|
||||||
|
* as unblocked, as we aren't doing proxy-migrations
|
||||||
|
* yet (more logic will be needed then).
|
||||||
|
*/
|
||||||
|
donor->blocked_on = NULL;
|
||||||
|
}
|
||||||
|
return NULL;
|
||||||
|
}
|
||||||
|
|
||||||
|
/*
|
||||||
|
* Initial simple sketch that just deactivates the blocked task
|
||||||
|
* chosen by pick_next_task() so we can then pick something that
|
||||||
|
* isn't blocked.
|
||||||
|
*/
|
||||||
|
static struct task_struct *
|
||||||
|
find_proxy_task(struct rq *rq, struct task_struct *donor, struct rq_flags *rf)
|
||||||
|
{
|
||||||
|
struct mutex *mutex;
|
||||||
|
|
||||||
|
mutex = donor->blocked_on;
|
||||||
|
/* Something changed in the chain, so pick again */
|
||||||
|
if (!mutex)
|
||||||
|
return NULL;
|
||||||
|
/*
|
||||||
|
* By taking mutex->wait_lock we hold off concurrent mutex_unlock()
|
||||||
|
* and ensure @owner sticks around.
|
||||||
|
*/
|
||||||
|
guard(raw_spinlock)(&mutex->wait_lock);
|
||||||
|
|
||||||
|
/* Check again that donor is blocked with blocked_lock held */
|
||||||
|
if (!task_is_blocked(donor) || mutex != __get_task_blocked_on(donor)) {
|
||||||
|
/*
|
||||||
|
* Something changed in the blocked_on chain and
|
||||||
|
* we don't know if only at this level. So, let's
|
||||||
|
* just bail out completely and let __schedule()
|
||||||
|
* figure things out (pick_again loop).
|
||||||
|
*/
|
||||||
|
return NULL; /* do pick_next_task() again */
|
||||||
|
}
|
||||||
|
return proxy_deactivate(rq, donor);
|
||||||
|
}
|
||||||
|
#else /* SCHED_PROXY_EXEC */
|
||||||
|
static struct task_struct *
|
||||||
|
find_proxy_task(struct rq *rq, struct task_struct *donor, struct rq_flags *rf)
|
||||||
|
{
|
||||||
|
WARN_ONCE(1, "This should never be called in the !SCHED_PROXY_EXEC case\n");
|
||||||
|
return donor;
|
||||||
|
}
|
||||||
|
#endif /* SCHED_PROXY_EXEC */
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* __schedule() is the main scheduler function.
|
* __schedule() is the main scheduler function.
|
||||||
*
|
*
|
||||||
@@ -6672,12 +6766,25 @@ static void __sched notrace __schedule(int sched_mode)
|
|||||||
goto picked;
|
goto picked;
|
||||||
}
|
}
|
||||||
} else if (!preempt && prev_state) {
|
} else if (!preempt && prev_state) {
|
||||||
try_to_block_task(rq, prev, &prev_state);
|
/*
|
||||||
|
* We pass task_is_blocked() as the should_block arg
|
||||||
|
* in order to keep mutex-blocked tasks on the runqueue
|
||||||
|
* for slection with proxy-exec (without proxy-exec
|
||||||
|
* task_is_blocked() will always be false).
|
||||||
|
*/
|
||||||
|
try_to_block_task(rq, prev, &prev_state,
|
||||||
|
!task_is_blocked(prev));
|
||||||
switch_count = &prev->nvcsw;
|
switch_count = &prev->nvcsw;
|
||||||
}
|
}
|
||||||
|
|
||||||
next = pick_next_task(rq, prev, &rf);
|
pick_again:
|
||||||
|
next = pick_next_task(rq, rq->donor, &rf);
|
||||||
rq_set_donor(rq, next);
|
rq_set_donor(rq, next);
|
||||||
|
if (unlikely(task_is_blocked(next))) {
|
||||||
|
next = find_proxy_task(rq, next, &rf);
|
||||||
|
if (!next)
|
||||||
|
goto pick_again;
|
||||||
|
}
|
||||||
picked:
|
picked:
|
||||||
clear_tsk_need_resched(prev);
|
clear_tsk_need_resched(prev);
|
||||||
clear_preempt_need_resched();
|
clear_preempt_need_resched();
|
||||||
|
|||||||
@@ -2262,6 +2262,14 @@ static inline int task_current_donor(struct rq *rq, struct task_struct *p)
|
|||||||
return rq->donor == p;
|
return rq->donor == p;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
static inline bool task_is_blocked(struct task_struct *p)
|
||||||
|
{
|
||||||
|
if (!sched_proxy_exec())
|
||||||
|
return false;
|
||||||
|
|
||||||
|
return !!p->blocked_on;
|
||||||
|
}
|
||||||
|
|
||||||
static inline int task_on_cpu(struct rq *rq, struct task_struct *p)
|
static inline int task_on_cpu(struct rq *rq, struct task_struct *p)
|
||||||
{
|
{
|
||||||
return p->on_cpu;
|
return p->on_cpu;
|
||||||
@@ -2459,7 +2467,7 @@ static inline void put_prev_set_next_task(struct rq *rq,
|
|||||||
struct task_struct *prev,
|
struct task_struct *prev,
|
||||||
struct task_struct *next)
|
struct task_struct *next)
|
||||||
{
|
{
|
||||||
WARN_ON_ONCE(rq->curr != prev);
|
WARN_ON_ONCE(rq->donor != prev);
|
||||||
|
|
||||||
__put_prev_set_next_dl_server(rq, prev, next);
|
__put_prev_set_next_dl_server(rq, prev, next);
|
||||||
|
|
||||||
|
|||||||
Reference in New Issue
Block a user