Home
last modified time | relevance | path

Searched refs:flush (Results 1 – 25 of 561) sorted by relevance

12345678910>>...23

/linux/drivers/md/dm-vdo/
H A Dflush.c6 #include "flush.h"
29 /* The current flush generation of the vdo */
31 /* The first unacknowledged flush generation */
33 /* The queue of flush requests waiting to notify other threads */
35 /* The queue of flush requests waiting for VIOs to complete */
37 /* The flush generation for which notifications are being sent */
41 /* The ID of the thread on which flush requests should be made */
43 /* The pool of flush requests */
45 /* Bios waiting for a flush request to become available */
49 /* The rotor for selecting the bio queue for submitting flush bio
103 struct vdo_flush *flush = NULL; allocate_flush() local
251 struct vdo_flush *flush = notify_flush() local
269 struct vdo_flush *flush = completion_as_vdo_flush(completion); flush_vdo() local
327 struct vdo_flush *flush = vdo_complete_flushes() local
369 initialize_flush(struct vdo_flush * flush,struct vdo * vdo) initialize_flush() argument
375 launch_flush(struct vdo_flush * flush) launch_flush() argument
398 struct vdo_flush *flush = mempool_alloc(vdo->flusher->flush_pool, GFP_NOWAIT); vdo_launch_flush() local
431 release_flush(struct vdo_flush * flush) release_flush() argument
462 struct vdo_flush *flush = completion_as_vdo_flush(completion); vdo_complete_flush_callback() local
515 vdo_complete_flush(struct vdo_flush * flush) vdo_complete_flush() argument
[all...]
/linux/arch/x86/hyperv/
H A Dnested.c23 struct hv_guest_mapping_flush *flush; in hyperv_flush_guest_mapping() local
33 flush = *this_cpu_ptr(hyperv_pcpu_input_arg); in hyperv_flush_guest_mapping()
35 if (unlikely(!flush)) { in hyperv_flush_guest_mapping()
40 flush->address_space = as; in hyperv_flush_guest_mapping()
41 flush->flags = 0; in hyperv_flush_guest_mapping()
44 flush, NULL); in hyperv_flush_guest_mapping()
57 struct hv_guest_mapping_flush_list *flush, in hyperv_fill_flush_guest_mapping_list() argument
74 flush->gpa_list[gpa_n].page.additional_pages = additional_pages; in hyperv_fill_flush_guest_mapping_list()
75 flush->gpa_list[gpa_n].page.largepage = false; in hyperv_fill_flush_guest_mapping_list()
76 flush->gpa_list[gpa_n].page.basepfn = cur; in hyperv_fill_flush_guest_mapping_list()
[all …]
/linux/fs/btrfs/
H A Dspace-info.c73 * the tail of space_info->tickets, kick async flush thread
453 enum btrfs_reserve_flush_enum flush) in calc_available_free_space()
503 * 1/2th of the space. If we can flush, don't let us overcommit in calc_available_free_space()
506 if (flush == BTRFS_RESERVE_FLUSH_ALL || flush == BTRFS_RESERVE_FLUSH_ALL_STEAL) in calc_available_free_space()
526 enum btrfs_reserve_flush_enum flush) in can_overcommit()
528 const u64 avail = calc_available_free_space(space_info, flush); in can_overcommit()
535 enum btrfs_reserve_flush_enum flush) in btrfs_can_overcommit()
541 return check_can_overcommit(space_info, space_info_used_bytes, bytes, flush); in btrfs_can_overcommit()
545 enum btrfs_reserve_flush_enum flush)
444 calc_available_free_space(const struct btrfs_space_info * space_info,enum btrfs_reserve_flush_enum flush) calc_available_free_space() argument
514 check_can_overcommit(const struct btrfs_space_info * space_info,u64 space_info_used_bytes,u64 bytes,enum btrfs_reserve_flush_enum flush) check_can_overcommit() argument
523 can_overcommit(const struct btrfs_space_info * space_info,u64 space_info_used_bytes,u64 bytes,enum btrfs_reserve_flush_enum flush) can_overcommit() argument
533 btrfs_can_overcommit(const struct btrfs_space_info * space_info,u64 bytes,enum btrfs_reserve_flush_enum flush) btrfs_can_overcommit() argument
581 enum btrfs_reserve_flush_enum flush = BTRFS_RESERVE_NO_FLUSH; btrfs_try_granting_tickets() local
1289 enum btrfs_flush_state flush; btrfs_preempt_reclaim_metadata_space() local
1655 handle_reserve_ticket(struct btrfs_space_info * space_info,struct reserve_ticket * ticket,u64 start_ns,u64 orig_bytes,enum btrfs_reserve_flush_enum flush) handle_reserve_ticket() argument
1702 is_normal_flushing(enum btrfs_reserve_flush_enum flush) is_normal_flushing() argument
1726 can_steal(enum btrfs_reserve_flush_enum flush) can_steal() argument
1736 can_ticket(enum btrfs_reserve_flush_enum flush) can_ticket() argument
1757 reserve_bytes(struct btrfs_space_info * space_info,u64 orig_bytes,enum btrfs_reserve_flush_enum flush) reserve_bytes() argument
1904 btrfs_reserve_metadata_bytes(struct btrfs_space_info * space_info,u64 orig_bytes,enum btrfs_reserve_flush_enum flush) btrfs_reserve_metadata_bytes() argument
1932 btrfs_reserve_data_bytes(struct btrfs_space_info * space_info,u64 bytes,enum btrfs_reserve_flush_enum flush) btrfs_reserve_data_bytes() argument
[all...]
/linux/net/ipv4/
H A Dtcp_offload.c296 int flush = 1; in tcp_gro_receive() local
307 flush = (__force int)((flags ^ tcp_flag_word(th2)) & in tcp_gro_receive()
309 flush |= (__force int)(th->ack_seq ^ th2->ack_seq); in tcp_gro_receive()
311 flush |= *(u32 *)((u8 *)th + i) ^ in tcp_gro_receive()
314 flush |= gro_receive_network_flush(th, th2, p); in tcp_gro_receive()
323 flush |= (mss != skb_shinfo(skb)->gso_size); in tcp_gro_receive()
325 flush |= (len - 1) >= mss; in tcp_gro_receive()
327 flush |= (ntohl(th2->seq) + skb_gro_len(p)) ^ ntohl(th->seq); in tcp_gro_receive()
328 flush |= skb_cmp_decrypted(p, skb); in tcp_gro_receive()
331 flush | in tcp_gro_receive()
[all...]
H A Dudp_offload.c27 NAPI_GRO_CB(skb)->flush = 1; in dummy_gro_rcv()
153 NAPI_GRO_CB(skb)->flush |= 1; in udp_tunnel_gro_rcv()
709 int flush; in udp_gro_receive_segment() local
713 NAPI_GRO_CB(skb)->flush = 1; in udp_gro_receive_segment()
720 NAPI_GRO_CB(skb)->flush = 1; in udp_gro_receive_segment()
739 NAPI_GRO_CB(skb)->flush = 1; in udp_gro_receive_segment()
743 flush = gro_receive_network_flush(uh, uh2, p); in udp_gro_receive_segment()
751 if (ulen > ntohs(uh2->len) || flush) { in udp_gro_receive_segment()
756 NAPI_GRO_CB(skb)->flush = 1; in udp_gro_receive_segment()
761 NAPI_GRO_CB(skb)->flush in udp_gro_receive_segment()
792 int flush = 1; udp_gro_receive() local
[all...]
/linux/lib/
H A Ddecompress_inflate.c44 long (*flush)(void*, unsigned long), in __gunzip()
53 if (flush) { in __gunzip()
82 strm->workspace = malloc(flush ? zlib_inflate_workspacesize() : in __gunzip()
138 if (!flush) { in __gunzip()
159 if (flush && strm->next_out > out_buf) { in __gunzip()
161 if (l != flush(out_buf, l)) { in __gunzip()
193 if (flush) in __gunzip()
202 long (*flush)(void*, unsigned long), in gunzip()
207 return __gunzip(buf, len, fill, flush, out_buf, 0, pos, error); in gunzip()
212 long (*flush)(void*, unsigned long), in __decompress()
[all …]
H A Ddecompress_unxz.c260 long (*flush)(void *src, unsigned long size), in unxz()
276 if (fill == NULL && flush == NULL) in unxz()
284 if (flush == NULL) { in unxz()
306 if (fill == NULL && flush == NULL) { in unxz()
332 if (flush != NULL && (b.out_pos == b.out_size in unxz()
339 if (flush(b.out, b.out_pos) != (long)b.out_pos) in unxz()
349 if (flush != NULL) in unxz()
389 if (flush != NULL) in unxz()
407 long (*flush)(void *src, unsigned long size), in __decompress()
412 return unxz(in, in_size, fill, flush, out, in_used, error); in __decompress()
H A Ddecompress_unzstd.c167 long (*flush)(void*, unsigned long), in __unzstd()
191 if (fill == NULL && flush == NULL) in __unzstd()
227 if (flush != NULL) { in __unzstd()
308 if (flush != NULL && out.pos > 0) { in __unzstd()
309 if (out.pos != flush(out.dst, out.pos)) { in __unzstd()
335 long (*flush)(void*, unsigned long), in unzstd()
340 return __unzstd(buf, len, fill, flush, out_buf, 0, pos, error); in unzstd()
345 long (*flush)(void*, unsigned long), in __decompress()
350 return __unzstd(buf, len, fill, flush, out_buf, out_len, pos, error); in __decompress()
H A Ddecompress_unlz4.c33 long (*flush)(void *, unsigned long), in unlz4()
52 } else if (!flush) { in unlz4()
174 if (flush && flush(outp, dest_len) != dest_len) in unlz4()
209 long (*flush)(void*, unsigned long), in __decompress()
215 return unlz4(buf, in_len - 4, fill, flush, output, posp, error); in __decompress()
H A Ddecompress_unlzo.c99 long (*flush)(void *, unsigned long), in unlzo()
112 } else if (!flush) { in unlzo()
243 if (flush && flush(out_buf, dst_len) != dst_len) in unlzo()
279 long (*flush)(void*, unsigned long), in __decompress()
284 return unlzo(buf, len, fill, flush, out_buf, pos, error); in __decompress()
H A Ddecompress_unlzma.c285 long (*flush)(void*, unsigned long); member
303 if (!wr->flush) { in peek_old_byte()
321 if (wr->flush && wr->buffer_pos == wr->header->dict_size) { in write_byte()
324 if (wr->flush((char *)wr->buffer, wr->header->dict_size) in write_byte()
541 long (*flush)(void*, unsigned long), in unlzma()
575 wr.flush = flush; in unlzma()
657 if (!wr.flush || wr.flush(wr.buffer, wr.buffer_pos) == wr.buffer_pos) in unlzma()
674 long (*flush)(void*, unsigned long), in __decompress()
679 return unlzma(buf, in_len - 4, fill, flush, output, posp, error); in __decompress()
/linux/tools/testing/selftests/drivers/net/netdevsim/
H A Dnexthop.sh147 $IP nexthop flush &> /dev/null
182 $IP nexthop flush &> /dev/null
202 $IP nexthop flush &> /dev/null
247 $IP nexthop flush &> /dev/null
267 $IP nexthop flush &> /dev/null
289 $IP nexthop flush &> /dev/null
314 $IP nexthop flush &> /dev/null
343 $IP nexthop flush &> /dev/null
373 $IP nexthop flush &> /dev/null
422 $IP nexthop flush &> /dev/null
[all …]
/linux/Documentation/arch/x86/
H A Dtlb.rst12 from areas other than the one we are trying to flush will be
21 1. The size of the flush being performed. A flush of the entire
25 be no collateral damage caused by doing the global flush, and
26 all of the individual flush will have ended up being wasted
29 damage we do with a full flush. So, the larger the TLB, the
30 more attractive an individual flush looks. Data and
37 especially the contents of the TLB during a given flush. The
38 sizes of the flush will vary greatly depending on the workload as
48 This will cause us to do the global flush for more cases.
53 Despite the fact that a single individual flush on x86 is
[all …]
/linux/lib/zlib_dfltcc/
H A Ddfltcc_deflate.c121 int flush, in dfltcc_deflate() argument
136 if (flush == Z_FULL_FLUSH) in dfltcc_deflate()
144 no_flush = flush == Z_NO_FLUSH; in dfltcc_deflate()
159 if (flush == Z_FINISH) in dfltcc_deflate()
162 if (flush == Z_FULL_FLUSH) in dfltcc_deflate()
217 need_empty_block = flush == Z_FINISH && param->bcf && !param->bhf; in dfltcc_deflate()
225 if (flush == Z_FINISH && !param->bcf) in dfltcc_deflate()
292 if (flush == Z_FINISH) { in dfltcc_deflate()
301 if (flush == Z_FULL_FLUSH) in dfltcc_deflate()
303 *result = flush == Z_NO_FLUSH ? need_more : block_done; in dfltcc_deflate()
H A Ddfltcc_inflate.h16 int flush, int *ret);
20 #define INFLATE_TYPEDO_HOOK(strm, flush) \ argument
25 action = dfltcc_inflate((strm), (flush), &ret); \
/linux/tools/testing/selftests/kvm/x86/
H A Dhyperv_tlb_flush.c202 struct hv_tlb_flush *flush = (struct hv_tlb_flush *)data->hcall_gva; in sender_guest_code() local
217 flush->flags = HV_FLUSH_ALL_VIRTUAL_ADDRESS_SPACES; in sender_guest_code()
218 flush->processor_mask = BIT(WORKER_VCPU_ID_1); in sender_guest_code()
229 flush->flags = HV_FLUSH_ALL_VIRTUAL_ADDRESS_SPACES; in sender_guest_code()
230 flush->processor_mask = BIT(WORKER_VCPU_ID_1); in sender_guest_code()
231 flush->gva_list[0] = (u64)data->test_pages; in sender_guest_code()
243 flush->flags = HV_FLUSH_ALL_VIRTUAL_ADDRESS_SPACES | in sender_guest_code()
245 flush->processor_mask = 0; in sender_guest_code()
256 flush->flags = HV_FLUSH_ALL_VIRTUAL_ADDRESS_SPACES | in sender_guest_code()
258 flush in sender_guest_code()
[all...]
/linux/drivers/md/
H A Ddm-delay.c45 struct delay_class flush; member
172 if (dc->flush.dev) in delay_dtr()
173 dm_put_device(ti, dc->flush.dev); in delay_dtr()
249 ret = delay_class_ctr(ti, &dc->flush, argv); in delay_ctr()
262 ret = delay_class_ctr(ti, &dc->flush, argv + 3); in delay_ctr()
268 ret = delay_class_ctr(ti, &dc->flush, argv + 6); in delay_ctr()
271 max_delay = max(max_delay, dc->flush.delay); in delay_ctr()
272 min_delay = min_not_zero(min_delay, dc->flush.delay); in delay_ctr()
372 c = &dc->flush; in delay_map()
411 DMEMIT("%u %u %u", dc->read.ops, dc->write.ops, dc->flush.ops); in delay_status()
[all …]
/linux/Documentation/block/
H A Dstat.rst44 flush I/Os requests number of flush I/Os processed
45 flush ticks milliseconds total wait time for flush requests
53 flush I/Os
56 These values increment when an flush I/O request completes.
58 Block layer combines flush requests and executes at most one at a time.
59 This counts flush requests executed by disk. Not tracked for partitions.
75 read ticks, write ticks, discard ticks, flush ticks
/linux/drivers/gpu/drm/vmwgfx/
H A Dvmwgfx_overlay.c63 SVGAEscapeVideoFlush flush; member
77 fill_escape(&cmd->escape, sizeof(cmd->flush)); in fill_flush()
78 cmd->flush.cmdType = SVGA_ESCAPE_VMWARE_VIDEO_FLUSH; in fill_flush()
79 cmd->flush.streamId = stream_id; in fill_flush()
93 struct vmw_escape_video_flush *flush; in vmw_overlay_send_put() local
117 fifo_size = sizeof(*cmds) + sizeof(*flush) + sizeof(*items) * num_items; in vmw_overlay_send_put()
125 flush = (struct vmw_escape_video_flush *)&items[num_items]; in vmw_overlay_send_put()
164 fill_flush(flush, arg->stream_id); in vmw_overlay_send_put()
184 struct vmw_escape_video_flush flush; in vmw_overlay_send_stop() member
206 fill_flush(&cmds->flush, stream_id); in vmw_overlay_send_stop()
/linux/include/net/libeth/
H A Dxsk.h290 void (*finalize)(void *xdpsq, bool sent, bool flush)) in libeth_xsk_xmit_do_bulk() argument
513 #define libeth_xsk_finalize_rx(bq, flush, finalize) \ argument
514 __libeth_xdp_finalize_rx(bq, LIBETH_XDP_TX_XSK, flush, finalize)
556 #define LIBETH_XSK_DEFINE_RUN_PROG(name, flush) \ argument
557 u32 __LIBETH_XDP_DEFINE_RUN_PROG(name, flush, xsk)
575 #define LIBETH_XSK_DEFINE_RUN(name, run, flush, populate) \ argument
576 __LIBETH_XDP_DEFINE_RUN(name, run, flush, populate, XSK)
584 #define LIBETH_XSK_DEFINE_FINALIZE(name, flush, finalize) \ argument
585 __LIBETH_XDP_DEFINE_FINALIZE(name, flush, finalize, xsk)
/linux/net/ipv6/
H A Dip6_offload.c213 u16 flush = 1; in ipv6_exthdrs_len()
224 flush += ntohs(iph->payload_len) != skb->len - hlen; in ipv6_gro_receive()
244 flush--; in ipv6_gro_receive()
278 NAPI_GRO_CB(skb)->flush |= flush; in ipv6_gro_receive()
283 flush = 1; in ipv6_gro_receive()
296 skb_gro_flush_final(skb, pp, flush); in ipv6_gro_receive()
307 NAPI_GRO_CB(skb)->flush = 1; in ipv6_gro_receive()
322 NAPI_GRO_CB(skb)->flush = 1; in sit_ip6ip6_gro_receive()
232 u16 flush = 1; ipv6_gro_receive() local
H A Dtcpv6_offload.c50 /* Don't bother verifying checksum if we're going to flush anyway. */ in tcp6_gro_receive()
51 if (!NAPI_GRO_CB(skb)->flush && in tcp6_gro_receive()
54 goto flush; in tcp6_gro_receive()
58 goto flush; in tcp6_gro_receive()
65 flush: in tcp6_gro_receive()
66 NAPI_GRO_CB(skb)->flush = 1; in tcp6_gro_receive()
/linux/drivers/i2c/busses/
H A Di2c-diolan-u2c.c156 static int diolan_write_cmd(struct i2c_diolan_u2c *dev, bool flush) in diolan_write_cmd() argument
158 if (flush || dev->olen >= DIOLAN_FLUSH_LEN) in diolan_write_cmd()
164 static int diolan_usb_cmd(struct i2c_diolan_u2c *dev, u8 command, bool flush) in diolan_usb_cmd() argument
168 return diolan_write_cmd(dev, flush); in diolan_usb_cmd()
173 bool flush) in diolan_usb_cmd_data() argument
178 return diolan_write_cmd(dev, flush); in diolan_usb_cmd_data()
183 u8 d2, bool flush) in diolan_usb_cmd_data2() argument
189 return diolan_write_cmd(dev, flush); in diolan_usb_cmd_data2()
214 dev_err(&dev->interface->dev, "Failed to flush input buffer\n"); in diolan_flush_input()
/linux/arch/arm/mm/
H A Dcache-v4.S41 mcr p15, 0, r0, c7, c7, 0 @ flush ID cache
61 mcr p15, 0, ip, c7, c7, 0 @ flush ID cache
123 mcr p15, 0, r0, c7, c7, 0 @ flush ID cache
/linux/arch/x86/kvm/mmu/
H A Dtdp_mmu.c805 * If this function should yield and flush is set, it will perform a remote
806 * TLB flush before yielding.
817 bool flush, bool shared) in tdp_mmu_iter_cond_resched() argument
824 if (flush) in tdp_mmu_iter_cond_resched()
981 gfn_t start, gfn_t end, bool can_yield, bool flush) in tdp_mmu_zap_leafs() argument
993 tdp_mmu_iter_cond_resched(kvm, &iter, flush, false)) { in tdp_mmu_zap_leafs()
994 flush = false; in tdp_mmu_zap_leafs()
1005 * Zappings SPTEs in invalid roots doesn't require a TLB flush, in tdp_mmu_zap_leafs()
1009 flush = true; in tdp_mmu_zap_leafs()
1018 return flush; in tdp_mmu_zap_leafs()
1026 kvm_tdp_mmu_zap_leafs(struct kvm * kvm,gfn_t start,gfn_t end,bool flush) kvm_tdp_mmu_zap_leafs() argument
1354 kvm_tdp_mmu_unmap_gfn_range(struct kvm * kvm,struct kvm_gfn_range * range,bool flush) kvm_tdp_mmu_unmap_gfn_range() argument
1804 bool flush = false; recover_huge_pages_range() local
[all...]

12345678910>>...23