Lines Matching refs:__ksym

49 s32 scx_bpf_create_dsq(u64 dsq_id, s32 node) __ksym;
50 s32 scx_bpf_select_cpu_dfl(struct task_struct *p, s32 prev_cpu, u64 wake_flags, bool *is_idle) __ksym;
52 const struct cpumask *cpus_allowed, u64 flags) __ksym __weak;
53 void scx_bpf_dsq_insert(struct task_struct *p, u64 dsq_id, u64 slice, u64 enq_flags) __ksym __weak;
54 void scx_bpf_dsq_insert_vtime(struct task_struct *p, u64 dsq_id, u64 slice, u64 vtime, u64 enq_flags) __ksym __weak;
55 u32 scx_bpf_dispatch_nr_slots(void) __ksym;
56 void scx_bpf_dispatch_cancel(void) __ksym;
57 bool scx_bpf_dsq_move_to_local(u64 dsq_id) __ksym __weak;
58 void scx_bpf_dsq_move_set_slice(struct bpf_iter_scx_dsq *it__iter, u64 slice) __ksym __weak;
59 void scx_bpf_dsq_move_set_vtime(struct bpf_iter_scx_dsq *it__iter, u64 vtime) __ksym __weak;
60 bool scx_bpf_dsq_move(struct bpf_iter_scx_dsq *it__iter, struct task_struct *p, u64 dsq_id, u64 enq_flags) __ksym __weak;
61 bool scx_bpf_dsq_move_vtime(struct bpf_iter_scx_dsq *it__iter, struct task_struct *p, u64 dsq_id, u64 enq_flags) __ksym __weak;
62 u32 scx_bpf_reenqueue_local(void) __ksym;
63 void scx_bpf_kick_cpu(s32 cpu, u64 flags) __ksym;
64 s32 scx_bpf_dsq_nr_queued(u64 dsq_id) __ksym;
65 void scx_bpf_destroy_dsq(u64 dsq_id) __ksym;
66 int bpf_iter_scx_dsq_new(struct bpf_iter_scx_dsq *it, u64 dsq_id, u64 flags) __ksym __weak;
67 struct task_struct *bpf_iter_scx_dsq_next(struct bpf_iter_scx_dsq *it) __ksym __weak;
68 void bpf_iter_scx_dsq_destroy(struct bpf_iter_scx_dsq *it) __ksym __weak;
69 void scx_bpf_exit_bstr(s64 exit_code, char *fmt, unsigned long long *data, u32 data__sz) __ksym __weak;
70 void scx_bpf_error_bstr(char *fmt, unsigned long long *data, u32 data_len) __ksym;
71 void scx_bpf_dump_bstr(char *fmt, unsigned long long *data, u32 data_len) __ksym __weak;
72 u32 scx_bpf_cpuperf_cap(s32 cpu) __ksym __weak;
73 u32 scx_bpf_cpuperf_cur(s32 cpu) __ksym __weak;
74 void scx_bpf_cpuperf_set(s32 cpu, u32 perf) __ksym __weak;
75 u32 scx_bpf_nr_node_ids(void) __ksym __weak;
76 u32 scx_bpf_nr_cpu_ids(void) __ksym __weak;
77 int scx_bpf_cpu_node(s32 cpu) __ksym __weak;
78 const struct cpumask *scx_bpf_get_possible_cpumask(void) __ksym __weak;
79 const struct cpumask *scx_bpf_get_online_cpumask(void) __ksym __weak;
80 void scx_bpf_put_cpumask(const struct cpumask *cpumask) __ksym __weak;
81 const struct cpumask *scx_bpf_get_idle_cpumask_node(int node) __ksym __weak;
82 const struct cpumask *scx_bpf_get_idle_cpumask(void) __ksym;
83 const struct cpumask *scx_bpf_get_idle_smtmask_node(int node) __ksym __weak;
84 const struct cpumask *scx_bpf_get_idle_smtmask(void) __ksym;
85 void scx_bpf_put_idle_cpumask(const struct cpumask *cpumask) __ksym;
86 bool scx_bpf_test_and_clear_cpu_idle(s32 cpu) __ksym;
87 s32 scx_bpf_pick_idle_cpu_node(const cpumask_t *cpus_allowed, int node, u64 flags) __ksym __weak;
88 s32 scx_bpf_pick_idle_cpu(const cpumask_t *cpus_allowed, u64 flags) __ksym;
89 s32 scx_bpf_pick_any_cpu_node(const cpumask_t *cpus_allowed, int node, u64 flags) __ksym __weak;
90 s32 scx_bpf_pick_any_cpu(const cpumask_t *cpus_allowed, u64 flags) __ksym;
91 bool scx_bpf_task_running(const struct task_struct *p) __ksym;
92 s32 scx_bpf_task_cpu(const struct task_struct *p) __ksym;
93 struct rq *scx_bpf_cpu_rq(s32 cpu) __ksym;
94 struct cgroup *scx_bpf_task_cgroup(struct task_struct *p) __ksym __weak;
95 u64 scx_bpf_now(void) __ksym __weak;
96 void scx_bpf_events(struct scx_event_stats *events, size_t events__sz) __ksym __weak;
287 void *bpf_obj_new_impl(__u64 local_type_id, void *meta) __ksym;
288 void bpf_obj_drop_impl(void *kptr, void *meta) __ksym;
295 void *meta, __u64 off) __ksym;
300 void *meta, __u64 off) __ksym;
303 struct bpf_list_node *bpf_list_pop_front(struct bpf_list_head *head) __ksym;
304 struct bpf_list_node *bpf_list_pop_back(struct bpf_list_head *head) __ksym;
306 struct bpf_rb_node *node) __ksym;
309 void *meta, __u64 off) __ksym;
312 struct bpf_rb_node *bpf_rbtree_first(struct bpf_rb_root *root) __ksym;
314 void *bpf_refcount_acquire_impl(void *kptr, void *meta) __ksym;
318 struct task_struct *bpf_task_from_pid(s32 pid) __ksym;
319 struct task_struct *bpf_task_acquire(struct task_struct *p) __ksym;
320 void bpf_task_release(struct task_struct *p) __ksym;
323 struct cgroup *bpf_cgroup_ancestor(struct cgroup *cgrp, int level) __ksym;
324 void bpf_cgroup_release(struct cgroup *cgrp) __ksym;
325 struct cgroup *bpf_cgroup_from_id(u64 cgid) __ksym;
332 unsigned int flags) __weak __ksym;
334 bpf_iter_css_next(struct bpf_iter_css *it) __weak __ksym;
335 extern void bpf_iter_css_destroy(struct bpf_iter_css *it) __weak __ksym;
338 struct bpf_cpumask *bpf_cpumask_create(void) __ksym;
339 struct bpf_cpumask *bpf_cpumask_acquire(struct bpf_cpumask *cpumask) __ksym;
340 void bpf_cpumask_release(struct bpf_cpumask *cpumask) __ksym;
341 u32 bpf_cpumask_first(const struct cpumask *cpumask) __ksym;
342 u32 bpf_cpumask_first_zero(const struct cpumask *cpumask) __ksym;
343 void bpf_cpumask_set_cpu(u32 cpu, struct bpf_cpumask *cpumask) __ksym;
344 void bpf_cpumask_clear_cpu(u32 cpu, struct bpf_cpumask *cpumask) __ksym;
345 bool bpf_cpumask_test_cpu(u32 cpu, const struct cpumask *cpumask) __ksym;
346 bool bpf_cpumask_test_and_set_cpu(u32 cpu, struct bpf_cpumask *cpumask) __ksym;
347 bool bpf_cpumask_test_and_clear_cpu(u32 cpu, struct bpf_cpumask *cpumask) __ksym;
348 void bpf_cpumask_setall(struct bpf_cpumask *cpumask) __ksym;
349 void bpf_cpumask_clear(struct bpf_cpumask *cpumask) __ksym;
351 const struct cpumask *src2) __ksym;
353 const struct cpumask *src2) __ksym;
355 const struct cpumask *src2) __ksym;
356 bool bpf_cpumask_equal(const struct cpumask *src1, const struct cpumask *src2) __ksym;
357 bool bpf_cpumask_intersects(const struct cpumask *src1, const struct cpumask *src2) __ksym;
358 bool bpf_cpumask_subset(const struct cpumask *src1, const struct cpumask *src2) __ksym;
359 bool bpf_cpumask_empty(const struct cpumask *cpumask) __ksym;
360 bool bpf_cpumask_full(const struct cpumask *cpumask) __ksym;
361 void bpf_cpumask_copy(struct bpf_cpumask *dst, const struct cpumask *src) __ksym;
362 u32 bpf_cpumask_any_distribute(const struct cpumask *cpumask) __ksym;
364 const struct cpumask *src2) __ksym;
365 u32 bpf_cpumask_weight(const struct cpumask *cpumask) __ksym;
367 int bpf_iter_bits_new(struct bpf_iter_bits *it, const u64 *unsafe_ptr__ign, u32 nr_words) __ksym;
368 int *bpf_iter_bits_next(struct bpf_iter_bits *it) __ksym;
369 void bpf_iter_bits_destroy(struct bpf_iter_bits *it) __ksym;
447 void bpf_rcu_read_lock(void) __ksym;
448 void bpf_rcu_read_unlock(void) __ksym;