Home
last modified time | relevance | path

Searched refs:sched_data (Results 1 – 15 of 15) sorted by relevance

/linux/drivers/gpu/drm/i915/gvt/
H A Dsched_policy.c84 vgpu_data = vgpu->sched_data; in vgpu_update_timeslice()
94 static void gvt_balance_timeslice(struct gvt_sched_data *sched_data) in gvt_balance_timeslice() argument
108 list_for_each(pos, &sched_data->lru_runq_head) { in gvt_balance_timeslice()
113 list_for_each(pos, &sched_data->lru_runq_head) { in gvt_balance_timeslice()
122 list_for_each(pos, &sched_data->lru_runq_head) { in gvt_balance_timeslice()
163 vgpu_data = scheduler->next_vgpu->sched_data; in try_to_schedule_next_vgpu()
177 static struct intel_vgpu *find_busy_vgpu(struct gvt_sched_data *sched_data) in find_busy_vgpu() argument
181 struct list_head *head = &sched_data->lru_runq_head; in find_busy_vgpu()
212 static void tbs_sched_func(struct gvt_sched_data *sched_data) in tbs_sched_func() argument
214 struct intel_gvt *gvt = sched_data->gvt; in tbs_sched_func()
[all …]
H A Dscheduler.h58 void *sched_data; member
/linux/net/netfilter/ipvs/
H A Dip_vs_rr.c27 svc->sched_data = &svc->destinations; in ip_vs_rr_init_svc()
37 p = (struct list_head *) svc->sched_data; in ip_vs_rr_del_dest()
42 svc->sched_data = p->next->prev; in ip_vs_rr_del_dest()
62 p = (struct list_head *) svc->sched_data; in ip_vs_rr_schedule()
88 svc->sched_data = &dest->n_list; in ip_vs_rr_schedule()
H A Dip_vs_dh.c160 svc->sched_data = s; in ip_vs_dh_init_svc()
174 struct ip_vs_dh_state *s = svc->sched_data; in ip_vs_dh_done_svc()
189 struct ip_vs_dh_state *s = svc->sched_data; in ip_vs_dh_dest_changed()
220 s = (struct ip_vs_dh_state *) svc->sched_data; in ip_vs_dh_schedule()
H A Dip_vs_wrr.c120 svc->sched_data = mark; in ip_vs_wrr_init_svc()
128 struct ip_vs_wrr_mark *mark = svc->sched_data; in ip_vs_wrr_done_svc()
140 struct ip_vs_wrr_mark *mark = svc->sched_data; in ip_vs_wrr_dest_changed()
163 struct ip_vs_wrr_mark *mark = svc->sched_data; in ip_vs_wrr_schedule()
H A Dip_vs_sh.c236 svc->sched_data = s; in ip_vs_sh_init_svc()
250 struct ip_vs_sh_state *s = svc->sched_data; in ip_vs_sh_done_svc()
265 struct ip_vs_sh_state *s = svc->sched_data; in ip_vs_sh_dest_changed()
323 s = (struct ip_vs_sh_state *) svc->sched_data; in ip_vs_sh_schedule()
H A Dip_vs_lblc.c229 struct ip_vs_lblc_table *tbl = svc->sched_data; in ip_vs_lblc_flush()
256 struct ip_vs_lblc_table *tbl = svc->sched_data; in ip_vs_lblc_full_check()
354 svc->sched_data = tbl; in ip_vs_lblc_init_svc()
383 struct ip_vs_lblc_table *tbl = svc->sched_data; in ip_vs_lblc_done_svc()
483 struct ip_vs_lblc_table *tbl = svc->sched_data; in ip_vs_lblc_schedule()
H A Dip_vs_mh.c412 svc->sched_data = s; in ip_vs_mh_init_svc()
418 struct ip_vs_mh_state *s = svc->sched_data; in ip_vs_mh_done_svc()
431 struct ip_vs_mh_state *s = svc->sched_data; in ip_vs_mh_dest_changed()
486 s = (struct ip_vs_mh_state *)svc->sched_data; in ip_vs_mh_schedule()
H A Dip_vs_lblcr.c395 struct ip_vs_lblcr_table *tbl = svc->sched_data; in ip_vs_lblcr_flush()
421 struct ip_vs_lblcr_table *tbl = svc->sched_data; in ip_vs_lblcr_full_check()
517 svc->sched_data = tbl; in ip_vs_lblcr_init_svc()
546 struct ip_vs_lblcr_table *tbl = svc->sched_data; in ip_vs_lblcr_done_svc()
647 struct ip_vs_lblcr_table *tbl = svc->sched_data; in ip_vs_lblcr_schedule()
/linux/block/
H A Dbfq-wf2q.c156 group_sd = next_in_service->sched_data; in bfq_update_parent_budget()
158 bfqg = container_of(group_sd, struct bfq_group, sched_data); in bfq_update_parent_budget()
223 struct bfq_sched_data *sd = entity->sched_data; in bfq_inc_active_entities()
224 struct bfq_group *bfqg = container_of(sd, struct bfq_group, sched_data); in bfq_inc_active_entities()
232 struct bfq_sched_data *sd = entity->sched_data; in bfq_dec_active_entities()
233 struct bfq_group *bfqg = container_of(sd, struct bfq_group, sched_data); in bfq_dec_active_entities()
646 entity == entity->sched_data->in_service_entity); in bfq_put_idle_entity()
676 struct bfq_sched_data *sched_data = entity->sched_data; in bfq_entity_service_tree() local
679 return sched_data->service_tree + idx; in bfq_entity_service_tree()
988 struct bfq_sched_data *sd = entity->sched_data; in __bfq_requeue_entity()
[all …]
H A Dbfq-cgroup.c433 entity->sched_data = &bfqg->sched_data; in bfq_init_entity()
538 entity->my_sched_data = &bfqg->sched_data; in bfq_pd_init()
573 entity->sched_data = &parent->sched_data; in bfq_group_set_parent()
684 entity->sched_data = &bfqg->sched_data; in bfq_bfqq_move()
713 if (sync_bfqq->entity.sched_data != &bfqg->sched_data) in bfq_sync_bfqq_move()
724 if (bfqq->entity.sched_data != &bfqg->sched_data) in bfq_sync_bfqq_move()
764 async_bfqq->entity.sched_data != &bfqg->sched_data) { in __bfq_bic_change_cgroup()
858 if (bfqg->sched_data.in_service_entity) in bfq_reparent_active_queues()
860 bfqg->sched_data.in_service_entity, in bfq_reparent_active_queues()
891 st = bfqg->sched_data.service_tree + i; in bfq_pd_offline()
[all …]
H A Dbfq-iosched.h202 struct bfq_sched_data *sched_data; member
1009 struct bfq_sched_data sched_data; member
1029 struct bfq_sched_data sched_data; member
H A Dblk-mq-sched.c694 if (e->type->ops.exit_hctx && hctx->sched_data) { in blk_mq_exit_sched()
696 hctx->sched_data = NULL; in blk_mq_exit_sched()
H A Dbfq-iosched.c585 struct bfq_sched_data *sched_data; in bfqq_request_over_limit() local
616 sched_data = entity->sched_data; in bfqq_request_over_limit()
649 sched_data->service_tree[i].wsum; in bfqq_request_over_limit()
7191 root_group->sched_data.service_tree[i] = BFQ_SERVICE_TREE_INIT; in bfq_init_root_group()
7192 root_group->sched_data.bfq_class_idle_last_service = jiffies; in bfq_init_root_group()
/linux/include/net/
H A Dip_vs.h697 void *sched_data; /* scheduler application data */ member