| /linux/net/sched/ |
| H A D | sch_plug.c | 63 bool throttled; member 105 if (q->throttled) in plug_dequeue() 113 q->throttled = true; in plug_dequeue() 144 q->throttled = true; in plug_init() 174 q->throttled = true; in plug_change() 183 q->throttled = false; in plug_change() 191 q->throttled = false; in plug_change()
|
| H A D | sch_fq.c | 185 static struct fq_flow throttled; variable 189 return f->next == &throttled; in fq_flow_is_throttled() 239 f->next = &throttled; in fq_flow_set_throttled() 1295 st.throttled = q->stat_throttled; in fq_dump_stats()
|
| /linux/drivers/usb/serial/ |
| H A D | metro-usb.c | 41 int throttled; member 113 int throttled = 0; in metrousb_read_int_callback() local 149 throttled = metro_priv->throttled; in metrousb_read_int_callback() 152 if (throttled) in metrousb_read_int_callback() 180 metro_priv->throttled = 0; in metrousb_open() 275 metro_priv->throttled = 1; in metrousb_throttle() 331 metro_priv->throttled = 0; in metrousb_unthrottle()
|
| H A D | symbolserial.c | 28 bool throttled; member 80 if (!priv->throttled) { in symbol_int_callback() 98 priv->throttled = false; in symbol_open() 122 priv->throttled = true; in symbol_throttle() 134 priv->throttled = false; in symbol_unthrottle()
|
| H A D | digi_acceleport.c | 1364 int tty_flag, throttled; in digi_read_inb_callback() local 1390 throttled = priv->dp_throttled; in digi_read_inb_callback() 1391 if (throttled) in digi_read_inb_callback() 1430 return throttled ? 1 : 0; in digi_read_inb_callback()
|
| /linux/kernel/sched/ |
| H A D | pelt.h | 163 u64 throttled; in update_idle_cfs_rq_clock_pelt() local 166 throttled = U64_MAX; in update_idle_cfs_rq_clock_pelt() 168 throttled = cfs_rq->throttled_clock_pelt_time; in update_idle_cfs_rq_clock_pelt() 170 u64_u32_store(cfs_rq->throttled_pelt_idle, throttled); in update_idle_cfs_rq_clock_pelt()
|
| H A D | fair.c | 391 * With cfs_rq being unthrottled/throttled during an enqueue, in list_del_leaf_cfs_rq() 4718 u64 throttled = 0, now, lut; in migrate_se_pelt_lag() local 4767 throttled = u64_u32_load(cfs_rq->throttled_pelt_idle); in migrate_se_pelt_lag() 4769 if (throttled == U64_MAX) in migrate_se_pelt_lag() 4782 now -= throttled; in migrate_se_pelt_lag() 5936 if (cfs_rq->throttled) in __account_cfs_rq_runtime() 5940 * hierarchy can be throttled in __account_cfs_rq_runtime() 5957 return cfs_bandwidth_used() && cfs_rq->throttled; in cfs_rq_throttled() 5965 /* check whether cfs_rq, or any parent, is throttled */ 5978 return cfs_bandwidth_used() && p->throttled; in task_is_throttled() 6378 bool throttled = false; distribute_cfs_runtime() local 6465 int throttled; do_sched_cfs_period_timer() local 9166 bool throttled; pick_task_fair() local [all...] |
| H A D | debug.c | 983 cfs_rq->throttled); in print_cfs_rq()
|
| /linux/drivers/cpufreq/ |
| H A D | powernv-cpufreq.c | 113 static bool rebooting, throttled, occ_reset; variable 136 bool throttled; member 569 if (chip->throttled) in powernv_cpufreq_throttle_check() 571 chip->throttled = true; in powernv_cpufreq_throttle_check() 583 } else if (chip->throttled) { in powernv_cpufreq_throttle_check() 584 chip->throttled = false; in powernv_cpufreq_throttle_check() 593 throttled = true; in powernv_cpufreq_throttle_check() 599 throttled = true; in powernv_cpufreq_throttle_check() 603 if (throttled) { in powernv_cpufreq_throttle_check() 751 if (!throttled) { in powernv_cpufreq_target_index() [all …]
|
| /linux/Documentation/arch/x86/ |
| H A D | intel-hfi.rst | 34 reflected in the HFI. Likewise, if the system needs to be throttled due to 67 userspace are throttled at a rate of CONFIG_HZ jiffies.
|
| /linux/arch/um/drivers/ |
| H A D | line.h | 44 int throttled; member
|
| H A D | line.c | 219 line->throttled = 1; in line_throttle() 226 line->throttled = 0; in line_unthrottle()
|
| H A D | chan_kern.c | 169 if (!line->throttled) in line_timer_cb()
|
| /linux/drivers/tty/serial/ |
| H A D | stm32-usart.h | 213 bool throttled; /* port throttled */ member
|
| /linux/block/ |
| H A D | blk-throttle.c | 1738 bool throttled = false; in __blk_throtl_bio() local 1802 throttled = true; in __blk_throtl_bio() 1821 return throttled; in __blk_throtl_bio()
|
| /linux/tools/perf/ |
| H A D | builtin-kvm.c | 1266 int i, err, throttled = 0; in perf_kvm__mmap_read() local 1287 throttled = 1; in perf_kvm__mmap_read() 1304 return throttled; in perf_kvm__mmap_read()
|
| /linux/tools/include/uapi/linux/ |
| H A D | pkt_sched.h | 797 __u64 throttled; member
|
| /linux/Documentation/networking/device_drivers/ethernet/intel/ |
| H A D | igb.rst | 149 throttled via interrupt moderation. Interrupt moderation must be used to avoid
|
| /linux/drivers/tty/ |
| H A D | n_gsm.c | 181 bool throttled; /* Private copy of throttle state */ member 548 if (dlci->throttled) in gsm_encode_modem() 4554 dlci->throttled = true; in gsmtty_throttle() 4566 dlci->throttled = false; in gsmtty_unthrottle()
|
| /linux/Documentation/PCI/ |
| H A D | pcieaer-howto.rst | 103 the console/stalling execution, messages are throttled by device and error
|
| /linux/Documentation/admin-guide/cgroup-v1/ |
| H A D | blkio-controller.rst | 43 Run dd to read a file and see if rate is throttled to 1MB/s or not::
|
| /linux/Documentation/admin-guide/thermal/ |
| H A D | intel_powerclamp.rst | 242 normal cases, such as scp a large file, applications can be throttled
|
| /linux/Documentation/scheduler/ |
| H A D | sched-util-clamp.rst | 693 aggregation rule. But since the capped p0 task was running and throttled 704 Hence lead to a frequency spike since if p0 wasn't throttled we should get:
|
| /linux/include/linux/ |
| H A D | sched.h | 891 bool throttled; member
|
| /linux/Documentation/admin-guide/ |
| H A D | cgroup-v2.rst | 1356 throttled and put under heavy reclaim pressure. 1368 job will trigger the reclaim and/or get throttled on its 1487 throttled and routed to perform direct memory reclaim 1515 this cgroup are throttled. 1825 this limit, all its further allocations will be throttled to 2278 throttled without possibly adversely affecting higher priority groups. This 2281 originating group is being throttled you will see the use_delay and delay 2288 unthrottling any peer groups that were throttled previously. If the victimized
|