| H A D | ext.c | 474 #define SCX_CALL_OP(sch, op, locked_rq, args...) \ 478 if (locked_rq) { \ 480 update_locked_rq(locked_rq); \ 483 if (locked_rq) \ 487 #define SCX_CALL_OP_RET(sch, op, locked_rq, args...) \ 492 if (locked_rq) { \ 494 update_locked_rq(locked_rq); \ 497 if (locked_rq) \ 510 * either via the @locked_rq argument here, or (for ops.select_cpu()) via @p's 518 #define SCX_CALL_OP_TASK(sch, op, locked_rq, tas 475 SCX_CALL_OP(sch,op,locked_rq,args...) global() argument 488 SCX_CALL_OP_RET(sch,op,locked_rq,args...) global() argument 519 SCX_CALL_OP_TASK(sch,op,locked_rq,task,args...) global() argument 527 SCX_CALL_OP_TASK_RET(sch,op,locked_rq,task,args...) global() argument 537 SCX_CALL_OP_2TASKS_RET(sch,op,locked_rq,task0,task1,args...) global() argument 1173 schedule_dsq_reenq(struct scx_sched * sch,struct scx_dispatch_q * dsq,u64 reenq_flags,struct rq * locked_rq) schedule_dsq_reenq() argument 2530 struct rq *locked_rq = rq; dispatch_to_local_dsq() local 4134 struct rq *locked_rq = rq; reenq_user() local 8201 struct rq *this_rq, *src_rq, *locked_rq; scx_dsq_move() local 9289 struct rq *rq = cpu_rq(cpu), *locked_rq = scx_locked_rq(); scx_bpf_cpuperf_set() local [all...] |