/linux/drivers/md/dm-vdo/ |
H A D | flush.c | 103 struct vdo_flush *flush = NULL; in allocate_flush() local 106 flush = vdo_allocate_memory_nowait(sizeof(struct vdo_flush), __func__); in allocate_flush() 108 int result = vdo_allocate(1, struct vdo_flush, __func__, &flush); in allocate_flush() 114 if (flush != NULL) { in allocate_flush() 117 vdo_initialize_completion(&flush->completion, flusher->vdo, in allocate_flush() 121 return flush; in allocate_flush() 182 static void vdo_complete_flush(struct vdo_flush *flush); 251 struct vdo_flush *flush = in notify_flush() local 254 flusher->notify_generation = flush->flush_generation; in notify_flush() 269 struct vdo_flush *flush = completion_as_vdo_flush(completion); in flush_vdo() local [all …]
|
/linux/fs/btrfs/ |
H A D | space-info.c | 381 enum btrfs_reserve_flush_enum flush) in calc_available_free_space() argument 430 if (flush == BTRFS_RESERVE_FLUSH_ALL) in calc_available_free_space() 450 enum btrfs_reserve_flush_enum flush) in btrfs_can_overcommit() argument 460 avail = calc_available_free_space(fs_info, space_info, flush); in btrfs_can_overcommit() 485 enum btrfs_reserve_flush_enum flush = BTRFS_RESERVE_NO_FLUSH; in btrfs_try_granting_tickets() local 500 flush)) { in btrfs_try_granting_tickets() 513 flush = BTRFS_RESERVE_FLUSH_ALL; in btrfs_try_granting_tickets() 1111 space_info->flush = 0; in btrfs_async_reclaim_metadata_space() 1123 space_info->flush = 0; in btrfs_async_reclaim_metadata_space() 1166 space_info->flush = 0; in btrfs_async_reclaim_metadata_space() [all …]
|
H A D | delalloc-space.c | 118 enum btrfs_reserve_flush_enum flush = BTRFS_RESERVE_FLUSH_DATA; in btrfs_alloc_data_chunk_ondemand() local 124 flush = BTRFS_RESERVE_FLUSH_FREE_SPACE_INODE; in btrfs_alloc_data_chunk_ondemand() 126 return btrfs_reserve_data_bytes(fs_info, bytes, flush); in btrfs_alloc_data_chunk_ondemand() 134 enum btrfs_reserve_flush_enum flush = BTRFS_RESERVE_FLUSH_DATA; in btrfs_check_data_free_space() local 143 flush = BTRFS_RESERVE_NO_FLUSH; in btrfs_check_data_free_space() 145 flush = BTRFS_RESERVE_FLUSH_FREE_SPACE_INODE; in btrfs_check_data_free_space() 147 ret = btrfs_reserve_data_bytes(fs_info, len, flush); in btrfs_check_data_free_space() 313 enum btrfs_reserve_flush_enum flush = BTRFS_RESERVE_FLUSH_ALL; in btrfs_delalloc_reserve_metadata() local 326 flush = BTRFS_RESERVE_NO_FLUSH; in btrfs_delalloc_reserve_metadata() 329 flush = BTRFS_RESERVE_FLUSH_LIMIT; in btrfs_delalloc_reserve_metadata() [all …]
|
/linux/net/ipv4/ |
H A D | tcp_offload.c | 318 int flush = 1; in tcp_gro_receive() local 329 flush = (__force int)(flags & TCP_FLAG_CWR); in tcp_gro_receive() 330 flush |= (__force int)((flags ^ tcp_flag_word(th2)) & in tcp_gro_receive() 332 flush |= (__force int)(th->ack_seq ^ th2->ack_seq); in tcp_gro_receive() 334 flush |= *(u32 *)((u8 *)th + i) ^ in tcp_gro_receive() 337 flush |= gro_receive_network_flush(th, th2, p); in tcp_gro_receive() 346 flush |= (mss != skb_shinfo(skb)->gso_size); in tcp_gro_receive() 348 flush |= (len - 1) >= mss; in tcp_gro_receive() 350 flush |= (ntohl(th2->seq) + skb_gro_len(p)) ^ ntohl(th->seq); in tcp_gro_receive() 351 flush |= skb_cmp_decrypted(p, skb); in tcp_gro_receive() [all …]
|
H A D | udp_offload.c | 488 int flush; in udp_gro_receive_segment() local 492 NAPI_GRO_CB(skb)->flush = 1; in udp_gro_receive_segment() 499 NAPI_GRO_CB(skb)->flush = 1; in udp_gro_receive_segment() 518 NAPI_GRO_CB(skb)->flush = 1; in udp_gro_receive_segment() 522 flush = gro_receive_network_flush(uh, uh2, p); in udp_gro_receive_segment() 530 if (ulen > ntohs(uh2->len) || flush) { in udp_gro_receive_segment() 535 NAPI_GRO_CB(skb)->flush = 1; in udp_gro_receive_segment() 540 NAPI_GRO_CB(skb)->flush = 1; in udp_gro_receive_segment() 570 int flush = 1; in udp_gro_receive() local 605 flush = 0; in udp_gro_receive() [all …]
|
/linux/lib/ |
H A D | decompress_inflate.c | 44 long (*flush)(void*, unsigned long), in __gunzip() 53 if (flush) { in __gunzip() 82 strm->workspace = malloc(flush ? zlib_inflate_workspacesize() : in __gunzip() 138 if (!flush) { in __gunzip() 159 if (flush && strm->next_out > out_buf) { in __gunzip() 161 if (l != flush(out_buf, l)) { in __gunzip() 193 if (flush) in __gunzip() 202 long (*flush)(void*, unsigned long), in gunzip() 207 return __gunzip(buf, len, fill, flush, out_buf, 0, pos, error); in gunzip() 212 long (*flush)(void*, unsigned long), in __decompress() [all …]
|
H A D | decompress_unxz.c | 260 long (*flush)(void *src, unsigned long size), in unxz() 276 if (fill == NULL && flush == NULL) in unxz() 284 if (flush == NULL) { in unxz() 306 if (fill == NULL && flush == NULL) { in unxz() 332 if (flush != NULL && (b.out_pos == b.out_size in unxz() 339 if (flush(b.out, b.out_pos) != (long)b.out_pos) in unxz() 349 if (flush != NULL) in unxz() 389 if (flush != NULL) in unxz() 407 long (*flush)(void *src, unsigned long size), in __decompress() 412 return unxz(in, in_size, fill, flush, out, in_used, error); in __decompress()
|
H A D | decompress_unzstd.c | 167 long (*flush)(void*, unsigned long), in __unzstd() 191 if (fill == NULL && flush == NULL) in __unzstd() 227 if (flush != NULL) { in __unzstd() 308 if (flush != NULL && out.pos > 0) { in __unzstd() 309 if (out.pos != flush(out.dst, out.pos)) { in __unzstd() 335 long (*flush)(void*, unsigned long), in unzstd() 340 return __unzstd(buf, len, fill, flush, out_buf, 0, pos, error); in unzstd() 345 long (*flush)(void*, unsigned long), in __decompress() 350 return __unzstd(buf, len, fill, flush, out_buf, out_len, pos, error); in __decompress()
|
H A D | decompress_unlz4.c | 33 long (*flush)(void *, unsigned long), in unlz4() 52 } else if (!flush) { in unlz4() 174 if (flush && flush(outp, dest_len) != dest_len) in unlz4() 209 long (*flush)(void*, unsigned long), in __decompress() 215 return unlz4(buf, in_len - 4, fill, flush, output, posp, error); in __decompress()
|
H A D | decompress_unlzo.c | 99 long (*flush)(void *, unsigned long), in unlzo() 112 } else if (!flush) { in unlzo() 243 if (flush && flush(out_buf, dst_len) != dst_len) in unlzo() 279 long (*flush)(void*, unsigned long), in __decompress() 284 return unlzo(buf, len, fill, flush, out_buf, pos, error); in __decompress()
|
H A D | decompress_unlzma.c | 285 long (*flush)(void*, unsigned long); member 303 if (!wr->flush) { in peek_old_byte() 321 if (wr->flush && wr->buffer_pos == wr->header->dict_size) { in write_byte() 324 if (wr->flush((char *)wr->buffer, wr->header->dict_size) in write_byte() 541 long (*flush)(void*, unsigned long), in unlzma() 575 wr.flush = flush; in unlzma() 657 if (!wr.flush || wr.flush(wr.buffer, wr.buffer_pos) == wr.buffer_pos) in unlzma() 674 long (*flush)(void*, unsigned long), in __decompress() 679 return unlzma(buf, in_len - 4, fill, flush, output, posp, error); in __decompress()
|
/linux/include/net/ |
H A D | gro.h | 40 u16 flush; member 117 NAPI_GRO_CB(skb)->flush |= 1; in call_gro_receive() 132 NAPI_GRO_CB(skb)->flush |= 1; in call_gro_receive_sk() 368 static inline void skb_gro_flush_final(struct sk_buff *skb, struct sk_buff *pp, int flush) in skb_gro_flush_final() argument 371 NAPI_GRO_CB(skb)->flush |= flush; in skb_gro_flush_final() 375 int flush, in skb_gro_flush_final_remcsum() argument 379 NAPI_GRO_CB(skb)->flush |= flush; in skb_gro_flush_final_remcsum() 385 static inline void skb_gro_flush_final(struct sk_buff *skb, struct sk_buff *pp, int flush) in skb_gro_flush_final() argument 387 NAPI_GRO_CB(skb)->flush |= flush; in skb_gro_flush_final() 391 int flush, in skb_gro_flush_final_remcsum() argument [all …]
|
/linux/lib/zlib_deflate/ |
H A D | deflate.c | 60 #define DEFLATE_HOOK(strm, flush, bstate) 0 argument 69 typedef block_state (*compress_func) (deflate_state *s, int flush); 73 static block_state deflate_stored (deflate_state *s, int flush); 74 static block_state deflate_fast (deflate_state *s, int flush); 75 static block_state deflate_slow (deflate_state *s, int flush); 331 int flush in zlib_deflate() argument 338 flush > Z_FINISH || flush < 0) { in zlib_deflate() 344 (s->status == FINISH_STATE && flush != Z_FINISH)) { in zlib_deflate() 351 s->last_flush = flush; in zlib_deflate() 393 } else if (strm->avail_in == 0 && flush <= old_flush && in zlib_deflate() [all …]
|
/linux/tools/testing/selftests/drivers/net/netdevsim/ |
H A D | nexthop.sh | 147 $IP nexthop flush &> /dev/null 182 $IP nexthop flush &> /dev/null 202 $IP nexthop flush &> /dev/null 247 $IP nexthop flush &> /dev/null 267 $IP nexthop flush &> /dev/null 289 $IP nexthop flush &> /dev/null 314 $IP nexthop flush &> /dev/null 343 $IP nexthop flush &> /dev/null 373 $IP nexthop flush &> /dev/null 422 $IP nexthop flush &> /dev/null [all …]
|
/linux/Documentation/arch/x86/ |
H A D | tlb.rst | 12 from areas other than the one we are trying to flush will be 21 1. The size of the flush being performed. A flush of the entire 25 be no collateral damage caused by doing the global flush, and 26 all of the individual flush will have ended up being wasted 29 damage we do with a full flush. So, the larger the TLB, the 30 more attractive an individual flush looks. Data and 37 especially the contents of the TLB during a given flush. The 38 sizes of the flush will vary greatly depending on the workload as 48 This will cause us to do the global flush for more cases. 53 Despite the fact that a single individual flush on x86 is [all …]
|
/linux/lib/zlib_dfltcc/ |
H A D | dfltcc_deflate.c | 121 int flush, in dfltcc_deflate() argument 136 if (flush == Z_FULL_FLUSH) in dfltcc_deflate() 144 no_flush = flush == Z_NO_FLUSH; in dfltcc_deflate() 159 if (flush == Z_FINISH) in dfltcc_deflate() 162 if (flush == Z_FULL_FLUSH) in dfltcc_deflate() 217 need_empty_block = flush == Z_FINISH && param->bcf && !param->bhf; in dfltcc_deflate() 225 if (flush == Z_FINISH && !param->bcf) in dfltcc_deflate() 292 if (flush == Z_FINISH) { in dfltcc_deflate() 301 if (flush == Z_FULL_FLUSH) in dfltcc_deflate() 303 *result = flush == Z_NO_FLUSH ? need_more : block_done; in dfltcc_deflate()
|
H A D | dfltcc_inflate.h | 16 int flush, int *ret); 20 #define INFLATE_TYPEDO_HOOK(strm, flush) \ argument 25 action = dfltcc_inflate((strm), (flush), &ret); \
|
/linux/tools/testing/selftests/kvm/x86_64/ |
H A D | hyperv_tlb_flush.c |
|
/linux/drivers/md/ |
H A D | dm-delay.c | 41 struct delay_class flush; member 167 if (dc->flush.dev) in delay_dtr() 168 dm_put_device(ti, dc->flush.dev); in delay_dtr() 244 ret = delay_class_ctr(ti, &dc->flush, argv); in delay_ctr() 256 ret = delay_class_ctr(ti, &dc->flush, argv + 3); in delay_ctr() 262 ret = delay_class_ctr(ti, &dc->flush, argv + 6); in delay_ctr() 265 max_delay = max(max_delay, dc->flush.delay); in delay_ctr() 359 c = &dc->flush; in delay_map() 383 DMEMIT("%u %u %u", dc->read.ops, dc->write.ops, dc->flush.ops); in delay_status() 394 DMEMIT_DELAY_CLASS(&dc->flush); in delay_status() [all …]
|
/linux/Documentation/block/ |
H A D | stat.rst | 44 flush I/Os requests number of flush I/Os processed 45 flush ticks milliseconds total wait time for flush requests 53 flush I/Os 56 These values increment when an flush I/O request completes. 58 Block layer combines flush requests and executes at most one at a time. 59 This counts flush requests executed by disk. Not tracked for partitions. 75 read ticks, write ticks, discard ticks, flush ticks
|
/linux/drivers/gpu/drm/vmwgfx/ |
H A D | vmwgfx_overlay.c | 63 SVGAEscapeVideoFlush flush; member 77 fill_escape(&cmd->escape, sizeof(cmd->flush)); in fill_flush() 78 cmd->flush.cmdType = SVGA_ESCAPE_VMWARE_VIDEO_FLUSH; in fill_flush() 79 cmd->flush.streamId = stream_id; in fill_flush() 93 struct vmw_escape_video_flush *flush; in vmw_overlay_send_put() local 117 fifo_size = sizeof(*cmds) + sizeof(*flush) + sizeof(*items) * num_items; in vmw_overlay_send_put() 125 flush = (struct vmw_escape_video_flush *)&items[num_items]; in vmw_overlay_send_put() 164 fill_flush(flush, arg->stream_id); in vmw_overlay_send_put() 184 struct vmw_escape_video_flush flush; in vmw_overlay_send_stop() member 206 fill_flush(&cmds->flush, stream_id); in vmw_overlay_send_stop()
|
/linux/Documentation/admin-guide/device-mapper/ |
H A D | delay.rst | 15 3: apply offset and delay to read, write and flush operations on device 18 to write and flush operations on optionally different write_device with 35 # Create mapped device named "delayed" delaying read, write and flush operations for 500ms. 42 # Create mapped device delaying write and flush operations for 400ms and
|
/linux/block/ |
H A D | blk-flush.c | 105 return 1 << ffz(rq->flush.seq); in blk_flush_cur_seq() 121 rq->end_io = rq->flush.saved_end_io; in blk_flush_restore_request() 156 BUG_ON(rq->flush.seq & seq); in blk_flush_complete_seq() 157 rq->flush.seq |= seq; in blk_flush_complete_seq() 371 rq->flush.seq = 0; in blk_rq_init_flush() 373 rq->flush.saved_end_io = rq->end_io; /* Usually NULL */ in blk_rq_init_flush() 443 rq->flush.seq |= REQ_FSEQ_PREFLUSH; in blk_insert_flush()
|
/linux/net/ipv6/ |
H A D | ip6_offload.c | 36 NAPI_GRO_CB(skb)->flush |= 1, NULL : \ 230 u16 flush = 1; in ipv6_gro_receive() local 241 flush += ntohs(iph->payload_len) != skb->len - hlen; in ipv6_gro_receive() 261 flush--; in ipv6_gro_receive() 295 NAPI_GRO_CB(skb)->flush |= flush; in ipv6_gro_receive() 303 skb_gro_flush_final(skb, pp, flush); in ipv6_gro_receive() 314 NAPI_GRO_CB(skb)->flush = 1; in sit_ip6ip6_gro_receive() 329 NAPI_GRO_CB(skb)->flush = 1; in ip4ip6_gro_receive()
|
/linux/arch/arm/mm/ |
H A D | cache-v4.S | 41 mcr p15, 0, r0, c7, c7, 0 @ flush ID cache 61 mcr p15, 0, ip, c7, c7, 0 @ flush ID cache 123 mcr p15, 0, r0, c7, c7, 0 @ flush ID cache
|