| /linux/kernel/sched/ |
| H A D | deadline.c | 204 __dl_overflow(struct dl_bw *dl_b, unsigned long cap, u64 old_bw, u64 new_bw) in __dl_overflow() argument 207 cap_scale(dl_b->bw, cap) < dl_b->total_bw - old_bw + new_bw; in __dl_overflow() 288 static void dl_rq_change_utilization(struct rq *rq, struct sched_dl_entity *dl_se, u64 new_bw) in dl_rq_change_utilization() argument 307 __add_rq_bw(new_bw, &rq->dl); in dl_rq_change_utilization() 333 static void dl_change_utilization(struct task_struct *p, u64 new_bw) in dl_change_utilization() argument 340 dl_rq_change_utilization(task_rq(p), &p->dl, new_bw); in dl_change_utilization() 1880 u64 new_bw = dl_se->dl_bw; in __dl_server_attach_root() local 1890 __dl_add(dl_b, new_bw, dl_bw_cpus(cpu)); in __dl_server_attach_root() 1896 u64 new_bw = to_ratio(period, runtime); in dl_server_apply_params() local 1909 if (__dl_overflow(dl_b, cap, old_bw, new_bw)) in dl_server_apply_params() [all …]
|
| /linux/drivers/gpu/drm/i915/display/ |
| H A D | intel_dp_tunnel.c | 77 int new_bw; in update_tunnel_state() local 99 new_bw = get_current_link_bw(intel_dp, &new_bw_below_dprx); in update_tunnel_state() 110 kbytes_to_mbits(old_bw), kbytes_to_mbits(new_bw)); in update_tunnel_state() 685 int new_bw; in atomic_decrease_bw() local 700 new_bw = drm_dp_tunnel_atomic_get_required_bw(new_tunnel_state); in atomic_decrease_bw() 702 if (new_bw >= old_bw) in atomic_decrease_bw() 705 drm_dp_tunnel_alloc_bw(tunnel, new_bw); in atomic_decrease_bw()
|
| /linux/net/mac80211/ |
| H A D | ht.c | 613 enum ieee80211_sta_rx_bandwidth max_bw, new_bw; in ieee80211_ht_handle_chanwidth_notif() local 626 new_bw = ieee80211_sta_cur_vht_bw(link_sta); in ieee80211_ht_handle_chanwidth_notif() 628 if (link_sta->pub->bandwidth == new_bw) in ieee80211_ht_handle_chanwidth_notif() 631 link_sta->pub->bandwidth = new_bw; in ieee80211_ht_handle_chanwidth_notif()
|
| H A D | vht.c | 651 enum ieee80211_sta_rx_bandwidth new_bw; in __ieee80211_vht_handle_opmode() local 702 new_bw = ieee80211_sta_cur_vht_bw(link_sta); in __ieee80211_vht_handle_opmode() 703 if (new_bw != link_sta->pub->bandwidth) { in __ieee80211_vht_handle_opmode() 704 link_sta->pub->bandwidth = new_bw; in __ieee80211_vht_handle_opmode()
|
| /linux/drivers/edac/ |
| H A D | edac_mc_sysfs.c | 358 int new_bw = 0; in mci_sdram_scrub_rate_store() local 363 new_bw = mci->set_sdram_scrub_rate(mci, bandwidth); in mci_sdram_scrub_rate_store() 364 if (new_bw < 0) { in mci_sdram_scrub_rate_store()
|
| H A D | i7core_edac.c | 1955 static int set_sdram_scrub_rate(struct mem_ctl_info *mci, u32 new_bw) in set_sdram_scrub_rate() argument 1969 if (new_bw == 0) { in set_sdram_scrub_rate() 1990 do_div(scrub_interval, new_bw); in set_sdram_scrub_rate() 2009 return new_bw; in set_sdram_scrub_rate()
|
| H A D | e752x_edac.c | 988 static int set_sdram_scrub_rate(struct mem_ctl_info *mci, u32 new_bw) in set_sdram_scrub_rate() argument 1005 if (scrubrates[i].bandwidth >= new_bw) in set_sdram_scrub_rate()
|
| H A D | amd64_edac.c | 189 static int __set_scrub_rate(struct amd64_pvt *pvt, u32 new_bw, u32 min_rate) in __set_scrub_rate() argument 211 if (scrubrates[i].bandwidth <= new_bw) in __set_scrub_rate()
|
| /linux/mm/ |
| H A D | mempolicy.c | 223 unsigned int *old_bw, *new_bw; in mempolicy_set_node_perf() local 228 new_bw = kcalloc(nr_node_ids, sizeof(unsigned int), GFP_KERNEL); in mempolicy_set_node_perf() 229 if (!new_bw) in mempolicy_set_node_perf() 234 kfree(new_bw); in mempolicy_set_node_perf() 250 memcpy(new_bw, old_bw, nr_node_ids * sizeof(*old_bw)); in mempolicy_set_node_perf() 251 new_bw[node] = bw_val; in mempolicy_set_node_perf() 252 node_bw_table = new_bw; in mempolicy_set_node_perf() 264 reduce_interleave_weights(new_bw, new_wi_state->iw_table); in mempolicy_set_node_perf()
|
| /linux/drivers/thunderbolt/ |
| H A D | tunnel.c | 848 u32 new_rate, new_lanes, new_bw; in tb_dp_xchg_caps() local 858 new_bw = tb_dp_bandwidth(new_rate, new_lanes); in tb_dp_xchg_caps() 861 new_rate, new_lanes, new_bw); in tb_dp_xchg_caps()
|