Home
last modified time | relevance | path

Searched defs:sched (Results 1 – 25 of 62) sorted by relevance

123

/linux/drivers/gpu/drm/scheduler/
H A Dsched_main.c96 static u32 drm_sched_available_credits(struct drm_gpu_scheduler *sched) in drm_sched_available_credits()
115 static bool drm_sched_can_queue(struct drm_gpu_scheduler *sched, in drm_sched_can_queue()
182 static void drm_sched_rq_init(struct drm_gpu_scheduler *sched, in drm_sched_rq_init()
256 drm_sched_rq_select_entity_rr(struct drm_gpu_scheduler *sched, in drm_sched_rq_select_entity_rr()
313 drm_sched_rq_select_entity_fifo(struct drm_gpu_scheduler *sched, in drm_sched_rq_select_entity_fifo()
345 static void drm_sched_run_job_queue(struct drm_gpu_scheduler *sched) in drm_sched_run_job_queue()
355 static void drm_sched_run_free_queue(struct drm_gpu_scheduler *sched) in drm_sched_run_free_queue()
371 struct drm_gpu_scheduler *sched = s_fence->sched; in drm_sched_job_done() local
403 static void drm_sched_start_timeout(struct drm_gpu_scheduler *sched) in drm_sched_start_timeout()
412 static void drm_sched_start_timeout_unlocked(struct drm_gpu_scheduler *sched) in drm_sched_start_timeout_unlocked()
[all …]
H A Dsched_entity.c285 struct drm_gpu_scheduler *sched; in drm_sched_entity_flush() local
402 struct drm_gpu_scheduler *sched = entity->rq->sched; in drm_sched_entity_add_dependency_cb() local
528 struct drm_gpu_scheduler *sched; in drm_sched_entity_select_rq() local
604 struct drm_gpu_scheduler *sched; in drm_sched_entity_push_job() local
/linux/net/mptcp/
H A Dsched.c55 list_for_each_entry_rcu(sched, &mptcp_sched_list, list) { in mptcp_sched_find() local
87 if (!sched->get_send) { in mptcp_register_scheduler() argument
72 struct mptcp_sched_ops *sched; mptcp_get_available_schedulers() local
104 mptcp_unregister_scheduler(struct mptcp_sched_ops * sched) mptcp_unregister_scheduler() argument
120 mptcp_init_sched(struct mptcp_sock * msk,struct mptcp_sched_ops * sched) mptcp_init_sched() argument
139 struct mptcp_sched_ops *sched = msk->sched; mptcp_release_sched() local
[all...]
/linux/drivers/gpu/drm/nouveau/
H A Dnouveau_sched.c29 struct nouveau_sched *sched = args->sched; in nouveau_job_init() local
120 struct nouveau_sched *sched = job->sched; in nouveau_job_done() local
275 struct nouveau_sched *sched = job->sched; in nouveau_job_submit() local
370 struct drm_gpu_scheduler *sched = sched_job->sched; in nouveau_sched_timedout_job() local
401 nouveau_sched_init(struct nouveau_sched *sched, struct nouveau_drm *drm, in nouveau_sched_init()
465 struct nouveau_sched *sched; in nouveau_sched_create() local
484 nouveau_sched_job_list_empty(struct nouveau_sched *sched) in nouveau_sched_job_list_empty()
496 nouveau_sched_fini(struct nouveau_sched *sched) in nouveau_sched_fini()
516 struct nouveau_sched *sched = *psched; in nouveau_sched_destroy() local
H A Dnouveau_sched.h29 struct nouveau_sched *sched; member
53 struct nouveau_sched *sched; member
H A Dnouveau_exec.h11 struct nouveau_sched *sched; member
/linux/net/netfilter/ipvs/
H A Dip_vs_sched.c61 ip_vs_unbind_scheduler(struct ip_vs_service * svc,struct ip_vs_scheduler * sched) ip_vs_unbind_scheduler() argument
81 struct ip_vs_scheduler *sched; ip_vs_sched_getbyname() local
115 struct ip_vs_scheduler *sched; ip_vs_scheduler_get() local
145 struct ip_vs_scheduler *sched = rcu_dereference(svc->scheduler); ip_vs_scheduler_err() local
169 struct ip_vs_scheduler *sched; register_ip_vs_scheduler() local
[all...]
H A Dip_vs_ctl.c1290 struct ip_vs_scheduler *sched; in __ip_vs_update_dest() local
1606 struct ip_vs_scheduler *sched; in __ip_vs_unlink_dest() local
1686 struct ip_vs_scheduler *sched = NULL; in ip_vs_add_service() local
1884 struct ip_vs_scheduler *sched = NULL, *old_sched; in ip_vs_edit_service() local
2890 struct ip_vs_scheduler *sched = rcu_dereference(svc->scheduler); in ip_vs_info_seq_show() local
3473 struct ip_vs_scheduler *sched; in ip_vs_copy_service() local
3949 struct ip_vs_scheduler *sched; in ip_vs_genl_fill_service() local
/linux/tools/perf/
H A Dbuiltin-sched.c307 static void burn_nsecs(struct perf_sched *sched, u64 nsecs) in burn_nsecs() argument
326 static void calibrate_run_measurement_overhead(struct perf_sched *sched) in calibrate_run_measurement_overhead() argument
343 calibrate_sleep_measurement_overhead(struct perf_sched * sched) calibrate_sleep_measurement_overhead() argument
389 add_sched_event_run(struct perf_sched * sched,struct task_desc * task,u64 timestamp,u64 duration) add_sched_event_run() argument
412 add_sched_event_wakeup(struct perf_sched * sched,struct task_desc * task,u64 timestamp,struct task_desc * wakee) add_sched_event_wakeup() argument
438 add_sched_event_sleep(struct perf_sched * sched,struct task_desc * task,u64 timestamp) add_sched_event_sleep() argument
448 register_pid(struct perf_sched * sched,unsigned long pid,const char * comm) register_pid() argument
494 print_task_traces(struct perf_sched * sched) print_task_traces() argument
506 add_cross_task_wakeups(struct perf_sched * sched) add_cross_task_wakeups() argument
521 perf_sched__process_event(struct perf_sched * sched,struct sched_atom * atom) perf_sched__process_event() argument
560 self_open_counters(struct perf_sched * sched,unsigned long cur_task) self_open_counters() argument
615 struct perf_sched *sched; global() member
623 struct perf_sched *sched = parms->sched; thread_func() local
660 create_tasks(struct perf_sched * sched) create_tasks() argument
690 destroy_tasks(struct perf_sched * sched) destroy_tasks() argument
710 wait_for_tasks(struct perf_sched * sched) wait_for_tasks() argument
762 run_one_test(struct perf_sched * sched) run_one_test() argument
810 test_calibrations(struct perf_sched * sched) test_calibrations() argument
828 replay_wakeup_event(struct perf_sched * sched,struct evsel * evsel,struct perf_sample * sample,struct machine * machine __maybe_unused) replay_wakeup_event() argument
849 replay_switch_event(struct perf_sched * sched,struct evsel * evsel,struct perf_sample * sample,struct machine * machine __maybe_unused) replay_switch_event() argument
894 replay_fork_event(struct perf_sched * sched,union perf_event * event,struct machine * machine) replay_fork_event() argument
1039 thread_atoms_insert(struct perf_sched * sched,struct thread * thread) thread_atoms_insert() argument
1136 latency_switch_event(struct perf_sched * sched,struct evsel * evsel,struct perf_sample * sample,struct machine * machine) latency_switch_event() argument
1206 latency_runtime_event(struct perf_sched * sched,struct evsel * evsel,struct perf_sample * sample,struct machine * machine) latency_runtime_event() argument
1241 latency_wakeup_event(struct perf_sched * sched,struct evsel * evsel,struct perf_sample * sample,struct machine * machine) latency_wakeup_event() argument
1302 latency_migrate_task_event(struct perf_sched * sched,struct evsel * evsel,struct perf_sample * sample,struct machine * machine) latency_migrate_task_event() argument
1352 output_lat_thread(struct perf_sched * sched,struct work_atoms * work_list) output_lat_thread() argument
1500 perf_sched__sort_lat(struct perf_sched * sched) perf_sched__sort_lat() argument
1526 struct perf_sched *sched = container_of(tool, struct perf_sched, tool); process_sched_wakeup_event() local
1548 map__findnew_thread(struct perf_sched * sched,struct machine * machine,pid_t pid,pid_t tid) map__findnew_thread() argument
1563 sched_match_task(struct perf_sched * sched,const char * comm_str) sched_match_task() argument
1579 print_sched_map(struct perf_sched * sched,struct perf_cpu this_cpu,int cpus_nr,const char * color,bool sched_out) print_sched_map() argument
1629 map_switch_event(struct perf_sched * sched,struct evsel * evsel,struct perf_sample * sample,struct machine * machine) map_switch_event() argument
1798 struct perf_sched *sched = container_of(tool, struct perf_sched, tool); process_sched_switch_event() local
1824 struct perf_sched *sched = container_of(tool, struct perf_sched, tool); process_sched_runtime_event() local
1837 struct perf_sched *sched = container_of(tool, struct perf_sched, tool); perf_sched__process_fork_event() local
1854 struct perf_sched *sched = container_of(tool, struct perf_sched, tool); process_sched_migrate_task_event() local
1914 perf_sched__read_events(struct perf_sched * sched) perf_sched__read_events() argument
2087 timehist_header(struct perf_sched * sched) timehist_header() argument
2163 timehist_print_sample(struct perf_sched * sched,struct evsel * evsel,struct perf_sample * sample,struct addr_location * al,struct thread * thread,u64 t,const char state) timehist_print_sample() argument
2327 save_task_callchain(struct perf_sched * sched,struct perf_sample * sample,struct evsel * evsel,struct machine * machine) save_task_callchain() argument
2486 save_idle_callchain(struct perf_sched * sched,struct idle_thread_runtime * itr,struct perf_sample * sample) save_idle_callchain() argument
2502 timehist_get_thread(struct perf_sched * sched,struct perf_sample * sample,struct machine * machine,struct evsel * evsel) timehist_get_thread() argument
2550 timehist_skip_sample(struct perf_sched * sched,struct thread * thread,struct evsel * evsel,struct perf_sample * sample) timehist_skip_sample() argument
2594 timehist_print_wakeup_event(struct perf_sched * sched,struct evsel * evsel,struct perf_sample * sample,struct machine * machine,struct thread * awakened) timehist_print_wakeup_event() argument
2646 struct perf_sched *sched = container_of(tool, struct perf_sched, tool); timehist_sched_wakeup_event() local
2674 timehist_print_migration_event(struct perf_sched * sched,struct evsel * evsel,struct perf_sample * sample,struct machine * machine,struct thread * migrated) timehist_print_migration_event() argument
2735 struct perf_sched *sched = container_of(tool, struct perf_sched, tool); timehist_migrate_task_event() local
2794 struct perf_sched *sched = container_of(tool, struct perf_sched, tool); timehist_sched_change_event() local
2992 struct perf_sched *sched; global() member
3075 timehist_print_summary(struct perf_sched * sched,struct perf_session * session) timehist_print_summary() argument
3190 struct perf_sched *sched = container_of(tool, struct perf_sched, tool); perf_timehist__process_sample() local
3208 timehist_check_attr(struct perf_sched * sched,struct evlist * evlist) timehist_check_attr() argument
3234 timehist_parse_prio_str(struct perf_sched * sched) timehist_parse_prio_str() argument
3275 perf_sched__timehist(struct perf_sched * sched) perf_sched__timehist() argument
3400 print_bad_events(struct perf_sched * sched) print_bad_events() argument
3463 perf_sched__merge_lat(struct perf_sched * sched) perf_sched__merge_lat() argument
3478 setup_cpus_switch_event(struct perf_sched * sched) setup_cpus_switch_event() argument
3498 free_cpus_switch_event(struct perf_sched * sched) free_cpus_switch_event() argument
3504 perf_sched__lat(struct perf_sched * sched) perf_sched__lat() argument
3557 setup_map_cpus(struct perf_sched * sched) setup_map_cpus() argument
3579 setup_color_pids(struct perf_sched * sched) setup_color_pids() argument
3596 setup_color_cpus(struct perf_sched * sched) setup_color_cpus() argument
3613 perf_sched__map(struct perf_sched * sched) perf_sched__map() argument
3669 perf_sched__replay(struct perf_sched * sched) perf_sched__replay() argument
3726 setup_sorting(struct perf_sched * sched,const struct option * options,const char * const usage_msg[]) setup_sorting() argument
3749 struct perf_sched *sched = container_of(tool, struct perf_sched, tool); process_synthesized_schedstat_event() local
3807 perf_sched__schedstat_record(struct perf_sched * sched,int argc,const char ** argv) perf_sched__schedstat_record() argument
4483 perf_sched__schedstat_report(struct perf_sched * sched) perf_sched__schedstat_report() argument
4535 perf_sched__schedstat_diff(struct perf_sched * sched,int argc,const char ** argv) perf_sched__schedstat_diff() argument
4637 perf_sched__schedstat_live(struct perf_sched * sched,int argc,const char ** argv) perf_sched__schedstat_live() argument
4822 struct perf_sched sched = { cmd_sched() local
[all...]
/linux/drivers/gpu/drm/panthor/
H A Dpanthor_sched.c745 #define sched_queue_work(sched, wname) \ argument
761 #define sched_queue_delayed_work(sched, wname, delay) \ argument
1480 struct panthor_scheduler *sched = ptdev->scheduler; in cs_slot_process_fatal_event_locked() local
1526 struct panthor_scheduler *sched = ptdev->scheduler; in cs_slot_process_fault_event_locked() local
1578 struct panthor_scheduler *sched = ptdev->scheduler; in group_process_tiler_oom() local
1681 struct panthor_scheduler *sched = ptdev->scheduler; in cs_slot_process_tiler_oom_event_locked() local
1732 struct panthor_scheduler *sched = ptdev->scheduler; in csg_slot_process_idle_event_locked() local
1762 struct panthor_scheduler *sched = ptdev->scheduler; in csg_slot_process_progress_timer_event_locked() local
1873 struct panthor_scheduler *sched = container_of(work, struct panthor_scheduler, in process_fw_events_work() local
1952 struct panthor_scheduler *sched = ptdev->scheduler; in csgs_upd_ctx_apply_locked() local
[all …]
/linux/tools/perf/scripts/perl/
H A Dwakeup-latency.pl28 sub sched::sched_switch subroutine
51 sub sched::sched_wakeup subroutine
/linux/net/sctp/
H A Dstream_sched.c116 void sctp_sched_ops_register(enum sctp_sched_type sched, in sctp_sched_ops_register() argument
133 const struct sctp_sched_ops *sched = sctp_sched_ops_from_stream(stream); in sctp_sched_free_sched() local
149 enum sctp_sched_type sched) in sctp_sched_set_sched() argument
266 struct sctp_sched_ops *sched = sctp_sched_ops_from_stream(stream); sctp_sched_init_sid() local
[all...]
H A Dstream.c57 const struct sctp_sched_ops *sched; in sctp_stream_free_ext() local
133 const struct sctp_sched_ops *sched = sctp_sched_ops_from_stream(stream); in sctp_stream_init() local
185 const struct sctp_sched_ops *sched = sctp_sched_ops_from_stream(stream); in sctp_stream_free() local
210 const struct sctp_sched_ops *sched = sctp_sched_ops_from_stream(stream); in sctp_stream_update() local
/linux/net/sched/
H A Dsch_taprio.c120 struct sched_gate_list *sched) in taprio_calculate_gate_durations()
170 static ktime_t sched_base_time(const struct sched_gate_list *sched) in sched_base_time()
198 struct sched_gate_list *sched = container_of(head, struct sched_gate_list, rcu); in taprio_free_sched_cb() local
224 static s32 get_cycle_time_elapsed(struct sched_gate_list *sched, ktime_t time) in get_cycle_time_elapsed()
235 static ktime_t get_interval_end_time(struct sched_gate_list *sched, in get_interval_end_time()
272 struct sched_gate_list *sched, in taprio_update_queue_max_sdu()
325 struct sched_gate_list *sched, in find_entry_to_transmit()
403 struct sched_gate_list *sched, *admin; in is_valid_interval() local
472 struct sched_gate_list *sched, *admin; in get_packet_txtime() local
545 struct sched_gate_list *sched; in taprio_skb_exceeds_queue_max_sdu() local
[all …]
/linux/drivers/gpu/drm/xe/
H A Dxe_dep_scheduler.c31 struct drm_gpu_scheduler sched; member
77 struct drm_gpu_scheduler *sched; in xe_dep_scheduler_create() local
H A Dxe_guc_submit.c1283 struct xe_gpu_scheduler *sched = &q->guc->sched; in disable_scheduling_deregister() local
1480 struct xe_gpu_scheduler *sched = &q->guc->sched; in guc_exec_queue_timedout_job() local
1961 struct xe_gpu_scheduler *sched; in guc_exec_queue_init() local
2203 struct xe_gpu_scheduler *sched = &q->guc->sched; in guc_exec_queue_suspend() local
2262 struct xe_gpu_scheduler *sched = &q->guc->sched; in guc_exec_queue_resume() local
2314 struct xe_gpu_scheduler *sched = &q->guc->sched; in guc_exec_queue_stop() local
2499 struct xe_gpu_scheduler *sched = &q->guc->sched; in guc_exec_queue_pause() local
2578 struct xe_gpu_scheduler *sched = &q->guc->sched; in guc_exec_queue_start() local
2633 struct xe_gpu_scheduler *sched = &q->guc->sched; in guc_exec_queue_unpause_prepare() local
2679 struct xe_gpu_scheduler *sched = &q->guc->sched; in guc_exec_queue_replay_pending_state_change() local
[all …]
H A Dxe_execlist_types.h37 struct drm_gpu_scheduler sched; member
/linux/drivers/gpu/drm/msm/
H A Dmsm_submitqueue.c151 struct drm_gpu_scheduler *sched = &ring->sched; in get_sched_entity() local
220 struct drm_gpu_scheduler *sched = &to_msm_vm(msm_context_vm(drm, ctx))->sched; in msm_submitqueue_create() local
/linux/drivers/slimbus/
H A Dsched.c29 struct slim_sched *sched = &ctrl->sched; in slim_ctrl_clk_pause() local
/linux/drivers/gpu/drm/msm/adreno/
H A Dadreno_device.c344 struct drm_gpu_scheduler *sched = &gpu->rb[i]->sched; in suspend_scheduler() local
355 struct drm_gpu_scheduler *sched = &gpu->rb[i]->sched; in resume_scheduler() local
/linux/drivers/net/ethernet/chelsio/cxgb/
H A Dsge.c226 struct sched { struct
231 struct sched_port p[MAX_NPORTS]; argument
232 struct tasklet_struct sched_tsk;/* tasklet used to run scheduler */ argument
236 static void restart_sched(struct tasklet_struct *t); argument
/linux/arch/x86/events/
H A Dcore.c838 static void perf_sched_init(struct perf_sched *sched, struct event_constraint **constraints, in perf_sched_init()
859 static void perf_sched_save_state(struct perf_sched *sched) in perf_sched_save_state()
868 static bool perf_sched_restore_state(struct perf_sched *sched) in perf_sched_restore_state()
890 static bool __perf_sched_find_counter(struct perf_sched *sched) in __perf_sched_find_counter()
945 static bool perf_sched_find_counter(struct perf_sched *sched) in perf_sched_find_counter()
959 static bool perf_sched_next_event(struct perf_sched *sched) in perf_sched_next_event()
990 struct perf_sched sched; in perf_assign_events() local
/linux/drivers/usb/host/
H A Dehci-sched.c1243 struct ehci_iso_sched *sched; in itd_urb_transaction() local
1387 struct ehci_iso_sched *sched, in sitd_slot_ok()
1480 struct ehci_iso_sched *sched = urb->hcpriv; in iso_stream_schedule() local
2150 struct ehci_iso_sched *sched = urb->hcpriv; in sitd_link_urb() local
/linux/drivers/gpu/drm/imagination/
H A Dpvr_queue.c811 struct drm_gpu_scheduler *sched = s_job->sched; in pvr_queue_timedout_job() local
1239 struct drm_gpu_scheduler *sched; in pvr_queue_create() local
/linux/tools/testing/selftests/bpf/progs/
H A Dxdp_features.c240 unsigned int drops, int sched, struct xdp_cpumap_stats *xdp_stats) in BPF_PROG()

123