| /linux/arch/x86/hyperv/ |
| H A D | nested.c | 23 struct hv_guest_mapping_flush *flush; in hyperv_flush_guest_mapping() local 33 flush = *this_cpu_ptr(hyperv_pcpu_input_arg); in hyperv_flush_guest_mapping() 35 if (unlikely(!flush)) { in hyperv_flush_guest_mapping() 40 flush->address_space = as; in hyperv_flush_guest_mapping() 41 flush->flags = 0; in hyperv_flush_guest_mapping() 44 flush, NULL); in hyperv_flush_guest_mapping() 57 struct hv_guest_mapping_flush_list *flush, in hyperv_fill_flush_guest_mapping_list() argument 74 flush->gpa_list[gpa_n].page.additional_pages = additional_pages; in hyperv_fill_flush_guest_mapping_list() 75 flush->gpa_list[gpa_n].page.largepage = false; in hyperv_fill_flush_guest_mapping_list() 76 flush->gpa_list[gpa_n].page.basepfn = cur; in hyperv_fill_flush_guest_mapping_list() [all …]
|
| /linux/drivers/md/dm-vdo/ |
| H A D | flush.c | 103 struct vdo_flush *flush = NULL; in allocate_flush() local 106 flush = vdo_allocate_memory_nowait(sizeof(struct vdo_flush), __func__); in allocate_flush() 108 int result = vdo_allocate(1, struct vdo_flush, __func__, &flush); in allocate_flush() 114 if (flush != NULL) { in allocate_flush() 117 vdo_initialize_completion(&flush->completion, flusher->vdo, in allocate_flush() 121 return flush; in allocate_flush() 182 static void vdo_complete_flush(struct vdo_flush *flush); 251 struct vdo_flush *flush = in notify_flush() local 254 flusher->notify_generation = flush->flush_generation; in notify_flush() 269 struct vdo_flush *flush = completion_as_vdo_flush(completion); in flush_vdo() local [all …]
|
| /linux/fs/btrfs/ |
| H A D | space-info.c | 444 enum btrfs_reserve_flush_enum flush) in calc_available_free_space() argument 494 if (flush == BTRFS_RESERVE_FLUSH_ALL) in calc_available_free_space() 514 enum btrfs_reserve_flush_enum flush) in check_can_overcommit() argument 516 const u64 avail = calc_available_free_space(space_info, flush); in check_can_overcommit() 523 enum btrfs_reserve_flush_enum flush) in can_overcommit() argument 529 return check_can_overcommit(space_info, space_info_used_bytes, bytes, flush); in can_overcommit() 533 enum btrfs_reserve_flush_enum flush) in btrfs_can_overcommit() argument 543 return check_can_overcommit(space_info, used, bytes, flush); in btrfs_can_overcommit() 581 enum btrfs_reserve_flush_enum flush = BTRFS_RESERVE_NO_FLUSH; in btrfs_try_granting_tickets() local 597 can_overcommit(space_info, used, ticket->bytes, flush)) { in btrfs_try_granting_tickets() [all …]
|
| /linux/net/ipv4/ |
| H A D | tcp_offload.c | 296 int flush = 1; in tcp_gro_receive() local 307 flush = (__force int)((flags ^ tcp_flag_word(th2)) & in tcp_gro_receive() 309 flush |= (__force int)(th->ack_seq ^ th2->ack_seq); in tcp_gro_receive() 311 flush |= *(u32 *)((u8 *)th + i) ^ in tcp_gro_receive() 314 flush |= gro_receive_network_flush(th, th2, p); in tcp_gro_receive() 323 flush |= (mss != skb_shinfo(skb)->gso_size); in tcp_gro_receive() 325 flush |= (len - 1) >= mss; in tcp_gro_receive() 327 flush |= (ntohl(th2->seq) + skb_gro_len(p)) ^ ntohl(th->seq); in tcp_gro_receive() 328 flush |= skb_cmp_decrypted(p, skb); in tcp_gro_receive() 331 flush | in tcp_gro_receive() [all...] |
| H A D | udp_offload.c | 27 NAPI_GRO_CB(skb)->flush = 1; in dummy_gro_rcv() 153 NAPI_GRO_CB(skb)->flush |= 1; in udp_tunnel_gro_rcv() 711 int flush; in udp_gro_receive_segment() 715 NAPI_GRO_CB(skb)->flush = 1; in udp_gro_receive_segment() 722 NAPI_GRO_CB(skb)->flush = 1; in udp_gro_receive_segment() 741 NAPI_GRO_CB(skb)->flush = 1; in udp_gro_receive_segment() 745 flush = gro_receive_network_flush(uh, uh2, p); in udp_gro_receive_segment() 753 if (ulen > ntohs(uh2->len) || flush) { in udp_gro_receive_segment() 758 NAPI_GRO_CB(skb)->flush = 1; in udp_gro_receive_segment() 763 NAPI_GRO_CB(skb)->flush in udp_gro_receive_segment() 709 int flush; udp_gro_receive_segment() local 792 int flush = 1; udp_gro_receive() local [all...] |
| /linux/lib/ |
| H A D | decompress_inflate.c | 44 long (*flush)(void*, unsigned long), in __gunzip() 53 if (flush) { in __gunzip() 82 strm->workspace = malloc(flush ? zlib_inflate_workspacesize() : in __gunzip() 138 if (!flush) { in __gunzip() 159 if (flush && strm->next_out > out_buf) { in __gunzip() 161 if (l != flush(out_buf, l)) { in __gunzip() 193 if (flush) in __gunzip() 202 long (*flush)(void*, unsigned long), in gunzip() 207 return __gunzip(buf, len, fill, flush, out_buf, 0, pos, error); in gunzip() 212 long (*flush)(void*, unsigned long), in __decompress() [all …]
|
| H A D | decompress_unxz.c | 256 * fill() and flush() won't be used. 260 long (*flush)(void *src, unsigned long size), in unxz() 276 if (fill == NULL && flush == NULL) in unxz() 284 if (flush == NULL) { in unxz() 306 if (fill == NULL && flush == NULL) { in unxz() 332 if (flush != NULL && (b.out_pos == b.out_size in unxz() 339 if (flush(b.out, b.out_pos) != (long)b.out_pos) in unxz() 349 if (flush != NULL) in unxz() 389 if (flush != NULL) in unxz() 407 long (*flush)(voi in __decompress() [all...] |
| H A D | decompress_unzstd.c | 167 long (*flush)(void*, unsigned long), in __unzstd() 191 if (fill == NULL && flush == NULL) in __unzstd() 227 if (flush != NULL) { in __unzstd() 308 if (flush != NULL && out.pos > 0) { in __unzstd() 309 if (out.pos != flush(out.dst, out.pos)) { in __unzstd() 335 long (*flush)(void*, unsigned long), in unzstd() 340 return __unzstd(buf, len, fill, flush, out_buf, 0, pos, error); in unzstd() 345 long (*flush)(void*, unsigned long), in __decompress() 350 return __unzstd(buf, len, fill, flush, out_buf, out_len, pos, error); in __decompress()
|
| H A D | decompress_unlz4.c | 33 long (*flush)(void *, unsigned long), in unlz4() 52 } else if (!flush) { in unlz4() 174 if (flush && flush(outp, dest_len) != dest_len) in unlz4() 209 long (*flush)(void*, unsigned long), in __decompress() 215 return unlz4(buf, in_len - 4, fill, flush, output, posp, error); in __decompress()
|
| H A D | decompress_unlzo.c | 99 long (*flush)(void *, unsigned long), in unlzo() 112 } else if (!flush) { in unlzo() 243 if (flush && flush(out_buf, dst_len) != dst_len) in unlzo() 279 long (*flush)(void*, unsigned long), in __decompress() 284 return unlzo(buf, len, fill, flush, out_buf, pos, error); in __decompress()
|
| H A D | decompress_unlzma.c | 285 long (*flush)(void*, unsigned long); member 303 if (!wr->flush) { in peek_old_byte() 321 if (wr->flush && wr->buffer_pos == wr->header->dict_size) { in write_byte() 324 if (wr->flush((char *)wr->buffer, wr->header->dict_size) in write_byte() 541 long (*flush)(void*, unsigned long), in unlzma() 575 wr.flush = flush; in unlzma() 657 if (!wr.flush || wr.flush(wr.buffer, wr.buffer_pos) == wr.buffer_pos) in unlzma() 674 long (*flush)(void*, unsigned long), in __decompress() 679 return unlzma(buf, in_len - 4, fill, flush, output, posp, error); in __decompress()
|
| /linux/tools/testing/selftests/drivers/net/netdevsim/ |
| H A D | nexthop.sh | 147 $IP nexthop flush &> /dev/null 182 $IP nexthop flush &> /dev/null 202 $IP nexthop flush &> /dev/null 247 $IP nexthop flush &> /dev/null 267 $IP nexthop flush &> /dev/null 289 $IP nexthop flush &> /dev/null 314 $IP nexthop flush &> /dev/null 343 $IP nexthop flush &> /dev/null 373 $IP nexthop flush &> /dev/null 422 $IP nexthop flush &> /dev/null [all …]
|
| /linux/Documentation/arch/x86/ |
| H A D | tlb.rst | 12 from areas other than the one we are trying to flush will be 21 1. The size of the flush being performed. A flush of the entire 25 be no collateral damage caused by doing the global flush, and 26 all of the individual flush will have ended up being wasted 29 damage we do with a full flush. So, the larger the TLB, the 30 more attractive an individual flush looks. Data and 37 especially the contents of the TLB during a given flush. The 38 sizes of the flush will vary greatly depending on the workload as 48 This will cause us to do the global flush for more cases. 53 Despite the fact that a single individual flush on x86 is [all …]
|
| /linux/lib/zlib_dfltcc/ |
| H A D | dfltcc_deflate.c | 121 int flush, in dfltcc_deflate() argument 136 if (flush == Z_FULL_FLUSH) in dfltcc_deflate() 144 no_flush = flush == Z_NO_FLUSH; in dfltcc_deflate() 159 if (flush == Z_FINISH) in dfltcc_deflate() 162 if (flush == Z_FULL_FLUSH) in dfltcc_deflate() 217 need_empty_block = flush == Z_FINISH && param->bcf && !param->bhf; in dfltcc_deflate() 225 if (flush == Z_FINISH && !param->bcf) in dfltcc_deflate() 292 if (flush == Z_FINISH) { in dfltcc_deflate() 301 if (flush == Z_FULL_FLUSH) in dfltcc_deflate() 303 *result = flush == Z_NO_FLUSH ? need_more : block_done; in dfltcc_deflate()
|
| H A D | dfltcc_inflate.h | 16 int flush, int *ret); 20 #define INFLATE_TYPEDO_HOOK(strm, flush) \ argument 25 action = dfltcc_inflate((strm), (flush), &ret); \
|
| /linux/tools/testing/selftests/kvm/x86/ |
| H A D | hyperv_tlb_flush.c | 202 struct hv_tlb_flush *flush = (struct hv_tlb_flush *)data->hcall_gva; in sender_guest_code() local 217 flush->flags = HV_FLUSH_ALL_VIRTUAL_ADDRESS_SPACES; in sender_guest_code() 218 flush->processor_mask = BIT(WORKER_VCPU_ID_1); in sender_guest_code() 229 flush->flags = HV_FLUSH_ALL_VIRTUAL_ADDRESS_SPACES; in sender_guest_code() 230 flush->processor_mask = BIT(WORKER_VCPU_ID_1); in sender_guest_code() 231 flush->gva_list[0] = (u64)data->test_pages; in sender_guest_code() 243 flush->flags = HV_FLUSH_ALL_VIRTUAL_ADDRESS_SPACES | in sender_guest_code() 245 flush->processor_mask = 0; in sender_guest_code() 256 flush->flags = HV_FLUSH_ALL_VIRTUAL_ADDRESS_SPACES | in sender_guest_code() 258 flush in sender_guest_code() [all...] |
| /linux/drivers/md/ |
| H A D | dm-delay.c | 45 struct delay_class flush; member 172 if (dc->flush.dev) in delay_dtr() 173 dm_put_device(ti, dc->flush.dev); in delay_dtr() 249 ret = delay_class_ctr(ti, &dc->flush, argv); in delay_ctr() 262 ret = delay_class_ctr(ti, &dc->flush, argv + 3); in delay_ctr() 268 ret = delay_class_ctr(ti, &dc->flush, argv + 6); in delay_ctr() 271 max_delay = max(max_delay, dc->flush.delay); in delay_ctr() 272 min_delay = min_not_zero(min_delay, dc->flush.delay); in delay_ctr() 284 dc->worker = kthread_run(&flush_worker_fn, dc, "dm-delay-flush-worker"); in delay_ctr() 372 c = &dc->flush; in delay_map() [all...] |
| /linux/Documentation/block/ |
| H A D | stat.rst | 44 flush I/Os requests number of flush I/Os processed 45 flush ticks milliseconds total wait time for flush requests 53 flush I/Os 56 These values increment when an flush I/O request completes. 58 Block layer combines flush requests and executes at most one at a time. 59 This counts flush requests executed by disk. Not tracked for partitions. 75 read ticks, write ticks, discard ticks, flush ticks
|
| /linux/drivers/gpu/drm/vmwgfx/ |
| H A D | vmwgfx_overlay.c | 63 SVGAEscapeVideoFlush flush; member 77 fill_escape(&cmd->escape, sizeof(cmd->flush)); in fill_flush() 78 cmd->flush.cmdType = SVGA_ESCAPE_VMWARE_VIDEO_FLUSH; in fill_flush() 79 cmd->flush.streamId = stream_id; in fill_flush() 93 struct vmw_escape_video_flush *flush; in vmw_overlay_send_put() local 117 fifo_size = sizeof(*cmds) + sizeof(*flush) + sizeof(*items) * num_items; in vmw_overlay_send_put() 125 flush = (struct vmw_escape_video_flush *)&items[num_items]; in vmw_overlay_send_put() 164 fill_flush(flush, arg->stream_id); in vmw_overlay_send_put() 184 struct vmw_escape_video_flush flush; in vmw_overlay_send_stop() member 206 fill_flush(&cmds->flush, stream_i in vmw_overlay_send_stop() [all...] |
| /linux/include/net/libeth/ |
| H A D | xsk.h | 290 void (*finalize)(void *xdpsq, bool sent, bool flush)) in libeth_xsk_xmit_do_bulk() argument 513 #define libeth_xsk_finalize_rx(bq, flush, finalize) \ argument 514 __libeth_xdp_finalize_rx(bq, LIBETH_XDP_TX_XSK, flush, finalize) 556 #define LIBETH_XSK_DEFINE_RUN_PROG(name, flush) \ argument 557 u32 __LIBETH_XDP_DEFINE_RUN_PROG(name, flush, xsk) 575 #define LIBETH_XSK_DEFINE_RUN(name, run, flush, populate) \ argument 576 __LIBETH_XDP_DEFINE_RUN(name, run, flush, populate, XSK) 584 #define LIBETH_XSK_DEFINE_FINALIZE(name, flush, finalize) \ argument 585 __LIBETH_XDP_DEFINE_FINALIZE(name, flush, finalize, xsk)
|
| /linux/net/ipv6/ |
| H A D | ip6_offload.c | 213 u16 flush = 1; in ipv6_exthdrs_len() 224 flush += ntohs(iph->payload_len) != skb->len - hlen; in ipv6_gro_receive() 244 flush--; in ipv6_gro_receive() 278 NAPI_GRO_CB(skb)->flush |= flush; in ipv6_gro_receive() 283 flush = 1; in ipv6_gro_receive() 296 skb_gro_flush_final(skb, pp, flush); in ipv6_gro_receive() 307 NAPI_GRO_CB(skb)->flush = 1; in ipv6_gro_receive() 322 NAPI_GRO_CB(skb)->flush = 1; in sit_ip6ip6_gro_receive() 232 u16 flush = 1; ipv6_gro_receive() local
|
| H A D | tcpv6_offload.c | 50 /* Don't bother verifying checksum if we're going to flush anyway. */ in tcp6_gro_receive() 51 if (!NAPI_GRO_CB(skb)->flush && in tcp6_gro_receive() 54 goto flush; in tcp6_gro_receive() 58 goto flush; in tcp6_gro_receive() 65 flush: in tcp6_gro_receive() 66 NAPI_GRO_CB(skb)->flush = 1; in tcp6_gro_receive()
|
| /linux/arch/arm/mm/ |
| H A D | cache-v4.S | 41 mcr p15, 0, r0, c7, c7, 0 @ flush ID cache 61 mcr p15, 0, ip, c7, c7, 0 @ flush ID cache 123 mcr p15, 0, r0, c7, c7, 0 @ flush ID cache
|
| /linux/drivers/i2c/busses/ |
| H A D | i2c-diolan-u2c.c | 156 static int diolan_write_cmd(struct i2c_diolan_u2c *dev, bool flush) in diolan_write_cmd() argument 158 if (flush || dev->olen >= DIOLAN_FLUSH_LEN) in diolan_write_cmd() 164 static int diolan_usb_cmd(struct i2c_diolan_u2c *dev, u8 command, bool flush) in diolan_usb_cmd() argument 168 return diolan_write_cmd(dev, flush); in diolan_usb_cmd() 173 bool flush) in diolan_usb_cmd_data() argument 178 return diolan_write_cmd(dev, flush); in diolan_usb_cmd_data() 183 u8 d2, bool flush) in diolan_usb_cmd_data2() argument 189 return diolan_write_cmd(dev, flush); in diolan_usb_cmd_data2() 214 dev_err(&dev->interface->dev, "Failed to flush input buffer\n"); in diolan_flush_input()
|
| /linux/tools/testing/selftests/drivers/net/mlxsw/ |
| H A D | fib_offload.sh | 103 ip -6 route flush 2001:db8:3::/64 dev $spine_p1 109 ip -6 route flush 2001:db8:3::/64 160 ip -6 route flush 2001:db8:3::/64 215 ip -6 route flush 2001:db8:3::/64 253 ip -6 route flush 2001:db8:3::/64 254 ip -6 route flush 2001:db8:4::/64
|