| /linux/drivers/md/ |
| H A D | dm-delay.c | 88 struct dm_delay_info *delayed, *next; in flush_delayed_bios() local 99 list_for_each_entry_safe(delayed, next, &local_list, list) { in flush_delayed_bios() 101 if (flush_all || time_after_eq(jiffies, delayed->expires)) { in flush_delayed_bios() 102 struct bio *bio = dm_bio_from_per_bio_data(delayed, in flush_delayed_bios() 104 list_del(&delayed->list); in flush_delayed_bios() 106 delayed->class->ops--; in flush_delayed_bios() 113 next_expires = delayed->expires; in flush_delayed_bios() 115 next_expires = min(next_expires, delayed->expires); in flush_delayed_bios() 314 struct dm_delay_info *delayed; in delay_bio() local 320 delayed = dm_per_bio_data(bio, sizeof(struct dm_delay_info)); in delay_bio() [all …]
|
| /linux/sound/pci/mixart/ |
| H A D | mixart_core.h | 217 u32 delayed; member 230 u32 delayed; member 387 u32 delayed; member 437 u32 delayed; member 497 u32 delayed; member 542 u32 delayed; member
|
| /linux/tools/virtio/ |
| H A D | vhost_net_test.c | 276 bool delayed, int bufs) in run_tx_test() argument 330 if (delayed) { in run_tx_test() 343 bool delayed, int bufs) in run_rx_test() argument 482 bool delayed = false; in main() local 504 delayed = true; in main() 528 run_rx_test(&dev, &dev.vqs[0], delayed, nbufs); in main() 529 run_tx_test(&dev, &dev.vqs[1], delayed, nbufs); in main()
|
| H A D | virtio_test.c | 170 bool delayed, int batch, int reset_n, int bufs) in run_test() argument 262 if (delayed) { in run_test() 351 bool delayed = false; in main() local 372 delayed = true; in main() 401 run_test(&dev, &dev.vqs[0], delayed, batch, reset, 0x100000); in main()
|
| /linux/Documentation/ABI/testing/ |
| H A D | sysfs-devices-xenbus | 20 pv device has been delayed in order to avoid stalls due to 30 trigger delayed EOI processing. 37 before delayed EOI processing is triggered for a Xen pv
|
| H A D | debugfs-scmi-raw | 26 Any additional delayed response received afterwards can be read 62 Any additional delayed response received afterwards can be read 140 Any additional delayed response received afterwards can be read 195 Any additional delayed response received afterwards can be read
|
| /linux/fs/btrfs/ |
| H A D | Makefile | 31 compression.o delayed-ref.o relocation.o delayed-inode.o scrub.o \ 47 tests/raid-stripe-tree-tests.o tests/delayed-refs-tests.o
|
| /linux/fs/xfs/scrub/ |
| H A D | fscounters.c | 300 uint64_t delayed; in xchk_fscount_aggregate_agcounts() local 361 delayed = percpu_counter_sum(&mp->m_delalloc_blks); in xchk_fscount_aggregate_agcounts() 362 fsc->fdblocks -= delayed; in xchk_fscount_aggregate_agcounts() 365 delayed); in xchk_fscount_aggregate_agcounts()
|
| /linux/Documentation/networking/ |
| H A D | proc_net_tcp.rst | 40 | | | | | | (delayed ACK control data) 53 2 another timer (e.g. delayed ack or keepalive) is pending
|
| /linux/Documentation/misc-devices/ |
| H A D | bh1770glc.rst | 36 by using a delayed work. As long as there is proximity threshold above 37 interrupts the delayed work is pushed forward. So, when proximity level goes 38 below the threshold value, there is no interrupt and the delayed work will
|
| /linux/Documentation/filesystems/xfs/ |
| H A D | index.rst | 11 xfs-delayed-logging-design
|
| /linux/drivers/cpufreq/ |
| H A D | sa1110-cpufreq.c | 130 static inline void set_mdcas(u_int *mdcas, int delayed, u_int rcd) in set_mdcas() argument 135 shift = delayed + 1 + rcd; in set_mdcas()
|
| /linux/drivers/usb/serial/ |
| H A D | usb_wwan.c | 182 usb_anchor_urb(this_urb, &portdata->delayed); in usb_wwan_write() 402 urb = usb_get_from_anchor(&portdata->delayed); in usb_wwan_close() 456 init_usb_anchor(&portdata->delayed); in usb_wwan_port_probe() 578 urb = usb_get_from_anchor(&portdata->delayed); in usb_wwan_submit_delayed_urbs()
|
| H A D | usb-wwan.h | 51 struct usb_anchor delayed; member
|
| /linux/drivers/gpu/vga/ |
| H A D | Kconfig | 12 X isn't running and delayed switching until the next logoff. This
|
| /linux/drivers/crypto/intel/qat/qat_common/ |
| H A D | icp_qat_hw.h | 375 #define ICP_QAT_HW_COMPRESSION_CONFIG_BUILD(dir, delayed, \ argument 379 (((delayed) & QAT_COMPRESSION_DELAYED_MATCH_MASK) << \
|
| /linux/tools/testing/selftests/net/packetdrill/ |
| H A D | tcp_slow_start_slow-start-ack-per-2pkt-send-5pkt.pkt | 6 // approximating standard delayed ACKs.
|
| H A D | tcp_slow_start_slow-start-ack-per-2pkt-send-6pkt.pkt | 7 // approximating standard delayed ACKs.
|
| H A D | tcp_slow_start_slow-start-ack-per-2pkt.pkt | 5 // approximating standard delayed ACKs.
|
| /linux/fs/ext4/ |
| H A D | Kconfig | 16 physical block numbers. The ext4 filesystem also supports delayed 23 are some performance gains from the delayed allocation and inode
|
| /linux/net/sched/ |
| H A D | sch_fq.c | 93 struct rb_node rate_node; /* anchor in q->delayed tree */ 138 struct rb_root delayed; /* for rate limited flows */ member 215 rb_erase(&f->rate_node, &q->delayed); in fq_flow_unset_throttled() 222 struct rb_node **p = &q->delayed.rb_node, *parent = NULL; in fq_flow_set_throttled() 235 rb_insert_color(&f->rate_node, &q->delayed); in fq_flow_set_throttled() 336 /* Ordering invariants fall apart if some delayed flows in fq_fastpath_check() 630 while ((p = rb_first(&q->delayed)) != NULL) { in fq_check_throttled() 833 q->delayed = RB_ROOT; in fq_reset() 1188 q->delayed = RB_ROOT; in fq_init()
|
| /linux/drivers/usb/class/ |
| H A D | cdc-acm.h | 103 struct usb_anchor delayed; /* writes queued for a device about to be woken */ member
|
| /linux/Documentation/driver-api/mmc/ |
| H A D | mmc-async-req.rst | 86 * the transfer is delayed, guesstimate max 4k as first chunk size. 96 * before this call, the transfer is delayed.
|
| /linux/tools/testing/selftests/drivers/net/mlxsw/ |
| H A D | tc_action_hw_stats.sh | 86 skip_sw dst_ip 192.0.2.2 action drop hw_stats delayed
|
| /linux/Documentation/tools/rtla/ |
| H A D | common_osnoise_options.txt | 6 Due to implementational limitations, actions might be delayed
|