mirror of
https://github.com/torvalds/linux.git
synced 2026-05-12 16:18:45 +02:00
sched_ext: Introduce scx_task_sched[_rcu]()
In preparation of multiple scheduler support, add p->scx.sched which points to the scx_sched instance that the task is scheduled by, which is currently always scx_root. Add scx_task_sched[_rcu]() accessors which return the associated scx_sched of the specified task and replace the raw scx_root dereferences with it where applicable. scx_task_on_sched() is also added to test whether a given task is on the specified sched. As scx_root is still the only scheduler, this shouldn't introduce user-visible behavior changes. Signed-off-by: Tejun Heo <tj@kernel.org> Reviewed-by: Andrea Righi <arighi@nvidia.com>
This commit is contained in:
parent
ebeca1f930
commit
88234b075c
|
|
@ -165,6 +165,13 @@ struct scx_sched;
|
|||
* for a task to be scheduled by SCX.
|
||||
*/
|
||||
struct sched_ext_entity {
|
||||
#ifdef CONFIG_CGROUPS
|
||||
/*
|
||||
* Associated scx_sched. Updated either during fork or while holding
|
||||
* both p->pi_lock and rq lock.
|
||||
*/
|
||||
struct scx_sched __rcu *sched;
|
||||
#endif
|
||||
struct scx_dispatch_q *dsq;
|
||||
atomic_long_t ops_state;
|
||||
u64 ddsp_dsq_id;
|
||||
|
|
|
|||
|
|
@ -19,7 +19,7 @@ static DEFINE_RAW_SPINLOCK(scx_sched_lock);
|
|||
* are used as temporary markers to indicate that the dereferences need to be
|
||||
* updated to point to the associated scheduler instances rather than scx_root.
|
||||
*/
|
||||
static struct scx_sched __rcu *scx_root;
|
||||
struct scx_sched __rcu *scx_root;
|
||||
|
||||
/*
|
||||
* All scheds, writers must hold both scx_enable_mutex and scx_sched_lock.
|
||||
|
|
@ -304,9 +304,15 @@ static struct scx_sched *scx_next_descendant_pre(struct scx_sched *pos,
|
|||
|
||||
return NULL;
|
||||
}
|
||||
|
||||
static void scx_set_task_sched(struct task_struct *p, struct scx_sched *sch)
|
||||
{
|
||||
rcu_assign_pointer(p->scx.sched, sch);
|
||||
}
|
||||
#else /* CONFIG_EXT_SUB_SCHED */
|
||||
static struct scx_sched *scx_parent(struct scx_sched *sch) { return NULL; }
|
||||
static struct scx_sched *scx_next_descendant_pre(struct scx_sched *pos, struct scx_sched *root) { return pos ? NULL : root; }
|
||||
static void scx_set_task_sched(struct task_struct *p, struct scx_sched *sch) {}
|
||||
#endif /* CONFIG_EXT_SUB_SCHED */
|
||||
|
||||
/**
|
||||
|
|
@ -1542,7 +1548,7 @@ static bool scx_rq_online(struct rq *rq)
|
|||
static void do_enqueue_task(struct rq *rq, struct task_struct *p, u64 enq_flags,
|
||||
int sticky_cpu)
|
||||
{
|
||||
struct scx_sched *sch = scx_root;
|
||||
struct scx_sched *sch = scx_task_sched(p);
|
||||
struct task_struct **ddsp_taskp;
|
||||
struct scx_dispatch_q *dsq;
|
||||
unsigned long qseq;
|
||||
|
|
@ -1672,7 +1678,7 @@ static void clr_task_runnable(struct task_struct *p, bool reset_runnable_at)
|
|||
|
||||
static void enqueue_task_scx(struct rq *rq, struct task_struct *p, int enq_flags)
|
||||
{
|
||||
struct scx_sched *sch = scx_root;
|
||||
struct scx_sched *sch = scx_task_sched(p);
|
||||
int sticky_cpu = p->scx.sticky_cpu;
|
||||
|
||||
if (enq_flags & ENQUEUE_WAKEUP)
|
||||
|
|
@ -1723,7 +1729,7 @@ static void enqueue_task_scx(struct rq *rq, struct task_struct *p, int enq_flags
|
|||
|
||||
static void ops_dequeue(struct rq *rq, struct task_struct *p, u64 deq_flags)
|
||||
{
|
||||
struct scx_sched *sch = scx_root;
|
||||
struct scx_sched *sch = scx_task_sched(p);
|
||||
unsigned long opss;
|
||||
u64 op_deq_flags = deq_flags;
|
||||
|
||||
|
|
@ -1794,7 +1800,7 @@ static void ops_dequeue(struct rq *rq, struct task_struct *p, u64 deq_flags)
|
|||
|
||||
static bool dequeue_task_scx(struct rq *rq, struct task_struct *p, int deq_flags)
|
||||
{
|
||||
struct scx_sched *sch = scx_root;
|
||||
struct scx_sched *sch = scx_task_sched(p);
|
||||
|
||||
if (!(p->scx.flags & SCX_TASK_QUEUED)) {
|
||||
WARN_ON_ONCE(task_runnable(p));
|
||||
|
|
@ -1838,8 +1844,8 @@ static bool dequeue_task_scx(struct rq *rq, struct task_struct *p, int deq_flags
|
|||
|
||||
static void yield_task_scx(struct rq *rq)
|
||||
{
|
||||
struct scx_sched *sch = scx_root;
|
||||
struct task_struct *p = rq->donor;
|
||||
struct scx_sched *sch = scx_task_sched(p);
|
||||
|
||||
if (SCX_HAS_OP(sch, yield))
|
||||
SCX_CALL_OP_2TASKS_RET(sch, SCX_KF_REST, yield, rq, p, NULL);
|
||||
|
|
@ -1849,10 +1855,10 @@ static void yield_task_scx(struct rq *rq)
|
|||
|
||||
static bool yield_to_task_scx(struct rq *rq, struct task_struct *to)
|
||||
{
|
||||
struct scx_sched *sch = scx_root;
|
||||
struct task_struct *from = rq->donor;
|
||||
struct scx_sched *sch = scx_task_sched(from);
|
||||
|
||||
if (SCX_HAS_OP(sch, yield))
|
||||
if (SCX_HAS_OP(sch, yield) && sch == scx_task_sched(to))
|
||||
return SCX_CALL_OP_2TASKS_RET(sch, SCX_KF_REST, yield, rq,
|
||||
from, to);
|
||||
else
|
||||
|
|
@ -2517,7 +2523,7 @@ static void process_ddsp_deferred_locals(struct rq *rq)
|
|||
*/
|
||||
while ((p = list_first_entry_or_null(&rq->scx.ddsp_deferred_locals,
|
||||
struct task_struct, scx.dsq_list.node))) {
|
||||
struct scx_sched *sch = scx_root;
|
||||
struct scx_sched *sch = scx_task_sched(p);
|
||||
struct scx_dispatch_q *dsq;
|
||||
|
||||
list_del_init(&p->scx.dsq_list.node);
|
||||
|
|
@ -2531,7 +2537,7 @@ static void process_ddsp_deferred_locals(struct rq *rq)
|
|||
|
||||
static void set_next_task_scx(struct rq *rq, struct task_struct *p, bool first)
|
||||
{
|
||||
struct scx_sched *sch = scx_root;
|
||||
struct scx_sched *sch = scx_task_sched(p);
|
||||
|
||||
if (p->scx.flags & SCX_TASK_QUEUED) {
|
||||
/*
|
||||
|
|
@ -2628,7 +2634,7 @@ static void switch_class(struct rq *rq, struct task_struct *next)
|
|||
static void put_prev_task_scx(struct rq *rq, struct task_struct *p,
|
||||
struct task_struct *next)
|
||||
{
|
||||
struct scx_sched *sch = scx_root;
|
||||
struct scx_sched *sch = scx_task_sched(p);
|
||||
|
||||
/* see kick_cpus_irq_workfn() */
|
||||
smp_store_release(&rq->scx.kick_sync, rq->scx.kick_sync + 1);
|
||||
|
|
@ -2722,14 +2728,14 @@ do_pick_task_scx(struct rq *rq, struct rq_flags *rf, bool force_scx)
|
|||
if (keep_prev) {
|
||||
p = prev;
|
||||
if (!p->scx.slice)
|
||||
refill_task_slice_dfl(rcu_dereference_sched(scx_root), p);
|
||||
refill_task_slice_dfl(scx_task_sched(p), p);
|
||||
} else {
|
||||
p = first_local_task(rq);
|
||||
if (!p)
|
||||
return NULL;
|
||||
|
||||
if (unlikely(!p->scx.slice)) {
|
||||
struct scx_sched *sch = rcu_dereference_sched(scx_root);
|
||||
struct scx_sched *sch = scx_task_sched(p);
|
||||
|
||||
if (!scx_rq_bypassing(rq) && !sch->warned_zero_slice) {
|
||||
printk_deferred(KERN_WARNING "sched_ext: %s[%d] has zero slice in %s()\n",
|
||||
|
|
@ -2817,7 +2823,7 @@ bool scx_prio_less(const struct task_struct *a, const struct task_struct *b,
|
|||
|
||||
static int select_task_rq_scx(struct task_struct *p, int prev_cpu, int wake_flags)
|
||||
{
|
||||
struct scx_sched *sch = scx_root;
|
||||
struct scx_sched *sch = scx_task_sched(p);
|
||||
bool rq_bypass;
|
||||
|
||||
/*
|
||||
|
|
@ -2878,7 +2884,7 @@ static void task_woken_scx(struct rq *rq, struct task_struct *p)
|
|||
static void set_cpus_allowed_scx(struct task_struct *p,
|
||||
struct affinity_context *ac)
|
||||
{
|
||||
struct scx_sched *sch = scx_root;
|
||||
struct scx_sched *sch = scx_task_sched(p);
|
||||
|
||||
set_cpus_allowed_common(p, ac);
|
||||
|
||||
|
|
@ -3022,7 +3028,7 @@ void scx_tick(struct rq *rq)
|
|||
|
||||
static void task_tick_scx(struct rq *rq, struct task_struct *curr, int queued)
|
||||
{
|
||||
struct scx_sched *sch = scx_root;
|
||||
struct scx_sched *sch = scx_task_sched(curr);
|
||||
|
||||
update_curr_scx(rq);
|
||||
|
||||
|
|
@ -3212,11 +3218,12 @@ static void scx_disable_task(struct task_struct *p)
|
|||
|
||||
static void scx_exit_task(struct task_struct *p)
|
||||
{
|
||||
struct scx_sched *sch = scx_root;
|
||||
struct scx_sched *sch = scx_task_sched(p);
|
||||
struct scx_exit_task_args args = {
|
||||
.cancelled = false,
|
||||
};
|
||||
|
||||
lockdep_assert_held(&p->pi_lock);
|
||||
lockdep_assert_rq_held(task_rq(p));
|
||||
|
||||
switch (scx_get_task_state(p)) {
|
||||
|
|
@ -3238,6 +3245,7 @@ static void scx_exit_task(struct task_struct *p)
|
|||
if (SCX_HAS_OP(sch, exit_task))
|
||||
SCX_CALL_OP_TASK(sch, SCX_KF_REST, exit_task, task_rq(p),
|
||||
p, &args);
|
||||
scx_set_task_sched(p, NULL);
|
||||
scx_set_task_state(p, SCX_TASK_NONE);
|
||||
}
|
||||
|
||||
|
|
@ -3267,12 +3275,18 @@ void scx_pre_fork(struct task_struct *p)
|
|||
|
||||
int scx_fork(struct task_struct *p, struct kernel_clone_args *kargs)
|
||||
{
|
||||
s32 ret;
|
||||
|
||||
percpu_rwsem_assert_held(&scx_fork_rwsem);
|
||||
|
||||
if (scx_init_task_enabled)
|
||||
return scx_init_task(p, task_group(p), true);
|
||||
else
|
||||
return 0;
|
||||
if (scx_init_task_enabled) {
|
||||
ret = scx_init_task(p, task_group(p), true);
|
||||
if (!ret)
|
||||
scx_set_task_sched(p, scx_root);
|
||||
return ret;
|
||||
}
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
void scx_post_fork(struct task_struct *p)
|
||||
|
|
@ -3377,7 +3391,7 @@ void sched_ext_dead(struct task_struct *p)
|
|||
static void reweight_task_scx(struct rq *rq, struct task_struct *p,
|
||||
const struct load_weight *lw)
|
||||
{
|
||||
struct scx_sched *sch = scx_root;
|
||||
struct scx_sched *sch = scx_task_sched(p);
|
||||
|
||||
lockdep_assert_rq_held(task_rq(p));
|
||||
|
||||
|
|
@ -3396,7 +3410,7 @@ static void prio_changed_scx(struct rq *rq, struct task_struct *p, u64 oldprio)
|
|||
|
||||
static void switching_to_scx(struct rq *rq, struct task_struct *p)
|
||||
{
|
||||
struct scx_sched *sch = scx_root;
|
||||
struct scx_sched *sch = scx_task_sched(p);
|
||||
|
||||
if (task_dead_and_done(p))
|
||||
return;
|
||||
|
|
@ -4062,7 +4076,7 @@ bool scx_allow_ttwu_queue(const struct task_struct *p)
|
|||
if (!scx_enabled())
|
||||
return true;
|
||||
|
||||
sch = rcu_dereference_sched(scx_root);
|
||||
sch = scx_task_sched(p);
|
||||
if (unlikely(!sch))
|
||||
return true;
|
||||
|
||||
|
|
@ -5582,6 +5596,7 @@ static void scx_root_enable_workfn(struct kthread_work *work)
|
|||
goto err_disable_unlock_all;
|
||||
}
|
||||
|
||||
scx_set_task_sched(p, sch);
|
||||
scx_set_task_state(p, SCX_TASK_READY);
|
||||
|
||||
put_task_struct(p);
|
||||
|
|
|
|||
|
|
@ -1223,6 +1223,7 @@ enum scx_ops_state {
|
|||
#define SCX_OPSS_STATE_MASK ((1LU << SCX_OPSS_QSEQ_SHIFT) - 1)
|
||||
#define SCX_OPSS_QSEQ_MASK (~SCX_OPSS_STATE_MASK)
|
||||
|
||||
extern struct scx_sched __rcu *scx_root;
|
||||
DECLARE_PER_CPU(struct rq *, scx_locked_rq_state);
|
||||
|
||||
/*
|
||||
|
|
@ -1243,3 +1244,61 @@ static inline bool scx_rq_bypassing(struct rq *rq)
|
|||
{
|
||||
return unlikely(rq->scx.flags & SCX_RQ_BYPASSING);
|
||||
}
|
||||
|
||||
#ifdef CONFIG_EXT_SUB_SCHED
|
||||
/**
|
||||
* scx_task_sched - Find scx_sched scheduling a task
|
||||
* @p: task of interest
|
||||
*
|
||||
* Return @p's scheduler instance. Must be called with @p's pi_lock or rq lock
|
||||
* held.
|
||||
*/
|
||||
static inline struct scx_sched *scx_task_sched(const struct task_struct *p)
|
||||
{
|
||||
return rcu_dereference_protected(p->scx.sched,
|
||||
lockdep_is_held(&p->pi_lock) ||
|
||||
lockdep_is_held(__rq_lockp(task_rq(p))));
|
||||
}
|
||||
|
||||
/**
|
||||
* scx_task_sched_rcu - Find scx_sched scheduling a task
|
||||
* @p: task of interest
|
||||
*
|
||||
* Return @p's scheduler instance. The returned scx_sched is RCU protected.
|
||||
*/
|
||||
static inline struct scx_sched *scx_task_sched_rcu(const struct task_struct *p)
|
||||
{
|
||||
return rcu_dereference_all(p->scx.sched);
|
||||
}
|
||||
|
||||
/**
|
||||
* scx_task_on_sched - Is a task on the specified sched?
|
||||
* @sch: sched to test against
|
||||
* @p: task of interest
|
||||
*
|
||||
* Returns %true if @p is on @sch, %false otherwise.
|
||||
*/
|
||||
static inline bool scx_task_on_sched(struct scx_sched *sch,
|
||||
const struct task_struct *p)
|
||||
{
|
||||
return rcu_access_pointer(p->scx.sched) == sch;
|
||||
}
|
||||
#else /* CONFIG_EXT_SUB_SCHED */
|
||||
static inline struct scx_sched *scx_task_sched(const struct task_struct *p)
|
||||
{
|
||||
return rcu_dereference_protected(scx_root,
|
||||
lockdep_is_held(&p->pi_lock) ||
|
||||
lockdep_is_held(__rq_lockp(task_rq(p))));
|
||||
}
|
||||
|
||||
static inline struct scx_sched *scx_task_sched_rcu(const struct task_struct *p)
|
||||
{
|
||||
return rcu_dereference_all(scx_root);
|
||||
}
|
||||
|
||||
static inline bool scx_task_on_sched(struct scx_sched *sch,
|
||||
const struct task_struct *p)
|
||||
{
|
||||
return true;
|
||||
}
|
||||
#endif /* CONFIG_EXT_SUB_SCHED */
|
||||
|
|
|
|||
Loading…
Reference in New Issue
Block a user