| H A D | ext.c | 199 static void scx_kick_cpu(struct scx_sched *sch, s32 cpu, u64 flags); 200 static bool scx_vexit(struct scx_sched *sch, enum scx_exit_kind kind, 203 static __printf(4, 5) bool scx_exit(struct scx_sched *sch, in scx_exit() argument 211 ret = scx_vexit(sch, kind, exit_code, fmt, args); in scx_exit() 217 #define scx_error(sch, fmt, args...) scx_exit((sch), SCX_EXIT_ERROR, 0, fmt, ##args) argument 218 #define scx_verror(sch, fmt, args) scx_vexit((sch), SCX_EXIT_ERROR, 0, fmt, args) argument 220 #define SCX_HAS_OP(sch, op) test_bit(SCX_OP_IDX(op), (sch) argument 248 find_global_dsq(struct scx_sched * sch,struct task_struct * p) find_global_dsq() argument 254 find_user_dsq(struct scx_sched * sch,u64 dsq_id) find_user_dsq() argument 311 SCX_CALL_OP(sch,mask,op,rq,args...) global() argument 326 SCX_CALL_OP_RET(sch,mask,op,rq,args...) global() argument 355 SCX_CALL_OP_TASK(sch,mask,op,rq,task,args...) global() argument 363 SCX_CALL_OP_TASK_RET(sch,mask,op,rq,task,args...) global() argument 373 SCX_CALL_OP_2TASKS_RET(sch,mask,op,rq,task0,task1,args...) global() argument 386 scx_kf_allowed(struct scx_sched * sch,u32 mask) scx_kf_allowed() argument 417 scx_kf_allowed_on_arg_tasks(struct scx_sched * sch,u32 mask,struct task_struct * p) scx_kf_allowed_on_arg_tasks() argument 689 scx_add_event(sch,name,cnt) global() argument 702 __scx_add_event(sch,name,cnt) global() argument 781 ops_cpu_valid(struct scx_sched * sch,s32 cpu,const char * where) ops_cpu_valid() argument 804 ops_sanitize_err(struct scx_sched * sch,const char * ops_name,s32 err) ops_sanitize_err() argument 987 refill_task_slice_dfl(struct scx_sched * sch,struct task_struct * p) refill_task_slice_dfl() argument 1017 dispatch_enqueue(struct scx_sched * sch,struct scx_dispatch_q * dsq,struct task_struct * p,u64 enq_flags) dispatch_enqueue() argument 1223 find_dsq_for_dispatch(struct scx_sched * sch,struct rq * rq,u64 dsq_id,struct task_struct * p) find_dsq_for_dispatch() argument 1255 mark_direct_dispatch(struct scx_sched * sch,struct task_struct * ddsp_task,struct task_struct * p,u64 dsq_id,u64 enq_flags) mark_direct_dispatch() argument 1286 direct_dispatch(struct scx_sched * sch,struct task_struct * p,u64 enq_flags) direct_dispatch() argument 1351 struct scx_sched *sch = scx_root; do_enqueue_task() local 1475 struct scx_sched *sch = scx_root; enqueue_task_scx() local 1525 struct scx_sched *sch = scx_root; ops_dequeue() local 1574 struct scx_sched *sch = scx_root; dequeue_task_scx() local 1618 struct scx_sched *sch = scx_root; yield_task_scx() local 1629 struct scx_sched *sch = scx_root; yield_to_task_scx() local 1715 task_can_run_on_remote_rq(struct scx_sched * sch,struct task_struct * p,struct rq * rq,bool enforce) task_can_run_on_remote_rq() argument 1844 move_task_between_dsqs(struct scx_sched * sch,struct task_struct * p,u64 enq_flags,struct scx_dispatch_q * src_dsq,struct scx_dispatch_q * dst_dsq) move_task_between_dsqs() argument 1897 consume_dispatch_q(struct scx_sched * sch,struct rq * rq,struct scx_dispatch_q * dsq) consume_dispatch_q() argument 1944 consume_global_dsq(struct scx_sched * sch,struct rq * rq) consume_global_dsq() argument 1966 dispatch_to_local_dsq(struct scx_sched * sch,struct rq * rq,struct scx_dispatch_q * dst_dsq,struct task_struct * p,u64 enq_flags) dispatch_to_local_dsq() argument 2066 finish_dispatch(struct scx_sched * sch,struct rq * rq,struct task_struct * p,unsigned long qseq_at_dispatch,u64 dsq_id,u64 enq_flags) finish_dispatch() argument 2128 flush_dispatch_buf(struct scx_sched * sch,struct rq * rq) flush_dispatch_buf() argument 2159 struct scx_sched *sch = scx_root; balance_one() local 2294 struct scx_sched *sch = scx_root; process_ddsp_deferred_locals() local 2308 struct scx_sched *sch = scx_root; set_next_task_scx() local 2364 struct scx_sched *sch = scx_root; switch_class() local 2405 struct scx_sched *sch = scx_root; put_prev_task_scx() local 2559 struct scx_sched *sch = rcu_dereference_sched(scx_root); do_pick_task_scx() local 2627 struct scx_sched *sch = scx_root; scx_prio_less() local 2647 struct scx_sched *sch = scx_root; select_task_rq_scx() local 2708 struct scx_sched *sch = scx_root; set_cpus_allowed_scx() local 2730 struct scx_sched *sch = scx_root; handle_hotplug() local 2780 struct scx_sched *sch; check_rq_for_timeouts() local 2827 struct scx_sched *sch; scx_tick() local 2852 struct scx_sched *sch = scx_root; task_tick_scx() local 2931 struct scx_sched *sch = scx_root; scx_init_task() local 2984 struct scx_sched *sch = scx_root; scx_enable_task() local 3012 struct scx_sched *sch = scx_root; scx_disable_task() local 3025 struct scx_sched *sch = scx_root; scx_exit_task() local 3190 struct scx_sched *sch = scx_root; reweight_task_scx() local 3209 struct scx_sched *sch = scx_root; switching_to_scx() local 3284 struct scx_sched *sch = scx_root; scx_tg_online() local 3313 struct scx_sched *sch = scx_root; scx_tg_offline() local 3326 struct scx_sched *sch = scx_root; scx_cgroup_can_attach() local 3375 struct scx_sched *sch = scx_root; scx_cgroup_move_task() local 3394 struct scx_sched *sch = scx_root; scx_cgroup_cancel_attach() local 3412 struct scx_sched *sch = scx_root; scx_group_set_weight() local 3428 struct scx_sched *sch = scx_root; scx_group_set_idle() local 3445 struct scx_sched *sch = scx_root; scx_group_set_bandwidth() local 3549 destroy_dsq(struct scx_sched * sch,u64 dsq_id) destroy_dsq() argument 3589 scx_cgroup_exit(struct scx_sched * sch) scx_cgroup_exit() argument 3614 scx_cgroup_init(struct scx_sched * sch) scx_cgroup_init() argument 3657 scx_cgroup_exit(struct scx_sched * sch) scx_cgroup_exit() argument 3658 scx_cgroup_init(struct scx_sched * sch) scx_cgroup_init() argument 3725 struct scx_sched *sch = container_of(rcu_work, struct scx_sched, rcu_work); scx_sched_free_rcu_work() local 3757 struct scx_sched *sch = container_of(kobj, struct scx_sched, kobj); scx_kobj_release() local 3766 struct scx_sched *sch = container_of(kobj, struct scx_sched, kobj); scx_attr_ops_show() local 3779 struct scx_sched *sch = container_of(kobj, struct scx_sched, kobj); scx_attr_events_show() local 3812 const struct scx_sched *sch = container_of(kobj, struct scx_sched, kobj); scx_uevent() local 3836 struct scx_sched *sch; scx_allow_ttwu_queue() local 3867 struct scx_sched *sch; handle_lockup() local 3947 bypass_lb_cpu(struct scx_sched * sch,struct rq * rq,struct cpumask * donee_mask,struct cpumask * resched_mask,u32 nr_donor_target,u32 nr_donee_target) bypass_lb_cpu() argument 4043 bypass_lb_node(struct scx_sched * sch,int node) bypass_lb_node() argument 4125 struct scx_sched *sch; scx_bypass_lb_timerfn() local 4178 struct scx_sched *sch; scx_bypass() local 4339 struct scx_sched *sch = container_of(work, struct scx_sched, disable_work); scx_disable_workfn() local 4492 scx_claim_exit(struct scx_sched * sch,enum scx_exit_kind kind) scx_claim_exit() argument 4512 struct scx_sched *sch; scx_disable() local 4642 struct scx_sched *sch = scx_root; scx_dump_task() local 4685 struct scx_sched *sch = scx_root; scx_dump_state() local 4831 struct scx_sched *sch = container_of(irq_work, struct scx_sched, error_irq_work); scx_error_irq_workfn() local 4840 scx_vexit(struct scx_sched * sch,enum scx_exit_kind kind,s64 exit_code,const char * fmt,va_list args) scx_vexit() argument 4898 struct scx_sched *sch; scx_alloc_and_add_sched() local 4978 check_hotplug_seq(struct scx_sched * sch,const struct sched_ext_ops * ops) check_hotplug_seq() argument 5002 validate_ops(struct scx_sched * sch,const struct sched_ext_ops * ops) validate_ops() argument 5050 struct scx_sched *sch; scx_enable_workfn() local 5470 struct scx_sched *sch = ops->priv; bpf_scx_unreg() local 5754 struct scx_sched *sch = scx_root; print_scx_info() local 5858 scx_dsq_insert_preamble(struct scx_sched * sch,struct task_struct * p,u64 enq_flags) scx_dsq_insert_preamble() argument 5879 scx_dsq_insert_commit(struct scx_sched * sch,struct task_struct * p,u64 dsq_id,u64 enq_flags) scx_dsq_insert_commit() argument 5948 struct scx_sched *sch; scx_bpf_dsq_insert___v2() local 5977 scx_dsq_insert_vtime(struct scx_sched * sch,struct task_struct * p,u64 dsq_id,u64 slice,u64 vtime,u64 enq_flags) scx_dsq_insert_vtime() argument 6037 struct scx_sched *sch; __scx_bpf_dsq_insert_vtime() local 6055 struct scx_sched *sch; scx_bpf_dsq_insert_vtime() local 6083 struct scx_sched *sch = scx_root; scx_dsq_move() local 6176 struct scx_sched *sch; scx_bpf_dispatch_nr_slots() local 6199 struct scx_sched *sch; scx_bpf_dispatch_cancel() local 6234 struct scx_sched *sch; scx_bpf_dsq_move_to_local() local 6435 struct scx_sched *sch; scx_bpf_reenqueue_local() local 6476 struct scx_sched *sch; scx_bpf_create_dsq() local 6552 scx_kick_cpu(struct scx_sched * sch,s32 cpu,u64 flags) scx_kick_cpu() argument 6617 struct scx_sched *sch; scx_bpf_kick_cpu() local 6634 struct scx_sched *sch; scx_bpf_dsq_nr_queued() local 6680 struct scx_sched *sch; scx_bpf_destroy_dsq() local 6703 struct scx_sched *sch; bpf_iter_scx_dsq_new() local 6817 struct scx_sched *sch; scx_bpf_dsq_peek() local 6840 __bstr_format(struct scx_sched * sch,u64 * data_buf,char * line_buf,size_t line_size,char * fmt,unsigned long long * data,u32 data__sz) __bstr_format() argument 6877 bstr_format(struct scx_sched * sch,struct scx_bstr_buf * buf,char * fmt,unsigned long long * data,u32 data__sz) bstr_format() argument 6899 struct scx_sched *sch; scx_bpf_exit_bstr() local 6922 struct scx_sched *sch; scx_bpf_error_bstr() local 6948 struct scx_sched *sch; scx_bpf_dump_bstr() local 7019 struct scx_sched *sch; scx_bpf_cpuperf_cap() local 7046 struct scx_sched *sch; scx_bpf_cpuperf_cur() local 7073 struct scx_sched *sch; scx_bpf_cpuperf_set() local 7190 struct scx_sched *sch; scx_bpf_cpu_rq() local 7219 struct scx_sched *sch; scx_bpf_locked_rq() local 7245 struct scx_sched *sch; scx_bpf_cpu_curr() local 7275 struct scx_sched *sch; scx_bpf_task_cgroup() local 7356 scx_read_events(struct scx_sched * sch,struct scx_event_stats * events) scx_read_events() argument 7385 struct scx_sched *sch; scx_bpf_events() local [all...] |