| /linux/drivers/gpu/drm/i915/gt/ |
| H A D | intel_gt_buffer_pool.c | 40 struct intel_gt_buffer_pool_node *node, *stale = NULL; in pool_free_older_than() local 68 node->free = stale; in pool_free_older_than() 69 stale = node; in pool_free_older_than() 80 while ((node = stale)) { in pool_free_older_than() 81 stale = stale->free; in pool_free_older_than()
|
| /linux/Documentation/admin-guide/hw-vuln/ |
| H A D | processor_mmio_stale_data.rst | 12 stale data into core fill buffers where the data can subsequently be inferred 21 Propagators are operations that result in stale data being copied or moved from 23 Vulnerabilities are operations that may result in stale data being directly 31 does not make stale data architecturally visible. Stale data must be propagated 36 The sideband stale data propagator (SSDP) is limited to the client (including 40 both requested data and unrequested stale data, from a transaction buffer and 41 the sideband response buffer. As a result, stale data from the sideband 46 The primary stale data propagator (PSDP) is limited to the client (including 50 buffer including both requested data and unrequested stale data. This is 51 similar to the sideband stale data propagator. [all …]
|
| H A D | gather_data_sampling.rst | 13 instruction that is transiently executed encounters a fault, stale data from 16 infer stale data using typical side channel techniques like cache timing 19 The attacker uses gather instructions to infer the stale vector register data. 29 Without mitigation, GDS can infer stale data across virtually all
|
| /linux/fs/xfs/scrub/ |
| H A D | dirtree.c | 294 dl->stale = true; in xchk_dirpath_revalidate() 381 if (dl->stale) { in xchk_dirpath_step_up() 481 if (dl->stale) { in xchk_dirpath_step_up() 582 if (!error && dl->stale) in xchk_dirpath_walk_upwards() 710 if (dl->stale || dl->aborted) in xchk_dirtree_live_update() 720 dl->stale = true; in xchk_dirtree_live_update() 750 dl->stale = false; in xchk_dirtree_reset() 840 ASSERT(dl->stale); in xchk_dirtree_find_paths_to_root() 848 } while (dl->stale); in xchk_dirtree_find_paths_to_root() 864 ASSERT(!dl->stale); in xchk_dirtree_evaluate()
|
| H A D | dirtree_repair.c | 413 if (dl->stale) { in xrep_dirtree_unlink() 543 if (!error && dl->stale) in xrep_dirtree_delete_path() 684 if (dl->stale) { in xrep_dirtree_adopt() 747 if (!error && dl->stale) in xrep_dirtree_move_to_orphanage() 805 if (!dl->stale) { in xrep_dirtree()
|
| H A D | dir.c | 636 unsigned int stale = 0; in xchk_directory_leaf1_bestfree() local 692 stale++; in xchk_directory_leaf1_bestfree() 694 if (leafhdr.stale != stale) in xchk_directory_leaf1_bestfree() 742 unsigned int stale = 0; in xchk_directory_free_bestfree() local 764 stale++; in xchk_directory_free_bestfree() 777 if (freehdr.nused + stale != freehdr.nvalid) in xchk_directory_free_bestfree()
|
| /linux/drivers/md/bcache/ |
| H A D | extents.c | 540 unsigned int i, stale; in bch_extent_bad() local 552 stale = ptr_stale(b->c, k, i); in bch_extent_bad() 554 if (stale && KEY_DIRTY(k)) { in bch_extent_bad() 557 stale, buf); in bch_extent_bad() 560 btree_bug_on(stale > BUCKET_GC_GEN_MAX, b, in bch_extent_bad() 562 stale, b->c->need_gc); in bch_extent_bad() 564 if (stale) in bch_extent_bad()
|
| H A D | btree.c | 1229 uint8_t stale = 0; in __bch_btree_mark_key() local 1239 return stale; in __bch_btree_mark_key() 1251 stale = max(stale, ptr_stale(c, k, i)); in __bch_btree_mark_key() 1275 return stale; in __bch_btree_mark_key() 1307 uint8_t stale = 0; in btree_gc_mark_node() local 1316 stale = max(stale, btree_mark_key(b, k)); in btree_gc_mark_node() 1338 if (stale > 10) in btree_gc_mark_node()
|
| /linux/kernel/configs/ |
| H A D | hardening.config | 63 # Initialize all heap variables to zero on free to reduce stale data lifetime. 69 # Wipe kernel stack after syscall completion to reduce stale data lifetime. 80 # Force IOMMU TLB invalidation so devices will never be able to access stale
|
| /linux/fs/iomap/ |
| H A D | iter.c | 63 bool stale = iter->iomap.flags & IOMAP_F_STALE; in iomap_iter() local 101 else if (iter->len == 0 || (!advanced && !stale)) in iomap_iter()
|
| /linux/drivers/gpu/drm/i915/gem/selftests/ |
| H A D | mock_context.c | 31 spin_lock_init(&ctx->stale.lock); in mock_context() 32 INIT_LIST_HEAD(&ctx->stale.engines); in mock_context()
|
| /linux/fs/xfs/ |
| H A D | xfs_buf_item.c | 505 int stale = bip->bli_flags & XFS_BLI_STALE; in xfs_buf_item_unpin() local 526 if (stale) { in xfs_buf_item_unpin() 683 bool stale = bip->bli_flags & XFS_BLI_STALE; in xfs_buf_item_release() local 701 ASSERT(!stale || (bip->__bli_format.blf_flags & XFS_BLF_CANCEL)); in xfs_buf_item_release() 723 if (stale) { in xfs_buf_item_release() 758 if (stale || hold) in xfs_buf_item_release()
|
| /linux/drivers/net/ethernet/intel/ice/ |
| H A D | ice_ptp.c | 487 if (test_and_clear_bit(idx, tx->stale)) in ice_ptp_complete_tx_single_tstamp() 648 if (test_and_clear_bit(idx, tx->stale)) in ice_ptp_process_tx_tstamp() 706 unsigned long *in_use, *stale; in ice_ptp_alloc_tx_tracker() local 711 stale = bitmap_zalloc(tx->len, GFP_KERNEL); in ice_ptp_alloc_tx_tracker() 713 if (!tstamps || !in_use || !stale) { in ice_ptp_alloc_tx_tracker() 716 bitmap_free(stale); in ice_ptp_alloc_tx_tracker() 723 tx->stale = stale; in ice_ptp_alloc_tx_tracker() 771 clear_bit(idx, tx->stale); in ice_ptp_flush_tx_tracker() 799 bitmap_or(tx->stale, tx->stale, tx->in_use, tx->len); in ice_ptp_mark_tx_tracker_stale() 846 bitmap_free(tx->stale); in ice_ptp_release_tx_tracker() [all …]
|
| H A D | ice_ptp.h | 118 unsigned long *stale; member
|
| H A D | ice_txrx.c | 2015 const skb_frag_t *frag, *stale; in __ice_chk_linearize() local 2047 for (stale = &skb_shinfo(skb)->frags[0];; stale++) { in __ice_chk_linearize() 2048 int stale_size = skb_frag_size(stale); in __ice_chk_linearize() 2059 int align_pad = -(skb_frag_off(stale)) & in __ice_chk_linearize()
|
| /linux/drivers/gpu/drm/i915/gem/ |
| H A D | i915_gem_context.c | 1084 spin_lock_irqsave(&ctx->stale.lock, flags); in engines_notify() 1086 spin_unlock_irqrestore(&ctx->stale.lock, flags); in engines_notify() 1435 spin_lock_irq(&ctx->stale.lock); in kill_context() 1437 list_for_each_entry_safe(pos, next, &ctx->stale.engines, link) { in kill_context() 1443 spin_unlock_irq(&ctx->stale.lock); in kill_context() 1448 spin_lock_irq(&ctx->stale.lock); in kill_context() 1455 spin_unlock_irq(&ctx->stale.lock); in kill_context() 1485 spin_lock_irq(&ctx->stale.lock); in engines_idle_release() 1487 list_add_tail(&engines->link, &ctx->stale.engines); in engines_idle_release() 1488 spin_unlock_irq(&ctx->stale.lock); in engines_idle_release() [all …]
|
| /linux/Documentation/translations/zh_CN/networking/ |
| H A D | mptcp-sysctl.rst | 115 用于判定子流失效(stale)的MPTCP层重传间隔次数阈值。当指定
|
| /linux/drivers/md/ |
| H A D | dm-ps-historical-service-time.c | 346 u64 *out, u64 *stale) in hst_fill_compare() argument 353 *stale = pi->stale_after; in hst_fill_compare()
|
| /linux/Documentation/scsi/ |
| H A D | hpsa.rst | 80 and not be disrupted in any way by stale commands or other stale state
|
| /linux/tools/testing/selftests/net/ |
| H A D | ndisc_unsolicited_na_test.sh | 145 to ${HOST_ADDR} dev ${ROUTER_INTF} nud stale)
|
| H A D | arp_ndisc_untracked_subnets.sh | 171 to ${HOST_ADDR_V6} dev ${ROUTER_INTF} nud stale)
|
| /linux/Documentation/arch/x86/ |
| H A D | mds.rst | 29 deallocated it can retain the stale data of the preceding operations which 37 contain stale data from a previous operation which can be forwarded to 193 that stale data from the idling CPU from spilling to the Hyper-Thread
|
| /linux/drivers/net/wireless/ath/ath9k/ |
| H A D | xmit.c | 447 tbf->bf_state.stale = false; in ath_clone_txbuf() 520 if (!bf->bf_state.stale || bf_next != NULL) in ath_tx_complete_aggr() 612 if (bf_next != NULL || !bf_last->bf_state.stale) in ath_tx_complete_aggr() 640 if (bf->bf_next == NULL && bf_last->bf_state.stale) { in ath_tx_complete_aggr() 958 bf->bf_state.stale = false; in ath_tx_get_tid_subframe() 1886 if (bf->bf_state.stale) { in ath_drain_txq_list() 2657 if (bf->bf_state.stale) { in ath_tx_processq() 2681 lastbf->bf_state.stale = true; in ath_tx_processq() 2762 if (bf->bf_state.stale) { in ath_tx_edma_tasklet() 2784 lastbf->bf_state.stale = true; in ath_tx_edma_tasklet()
|
| /linux/drivers/net/ethernet/intel/iavf/ |
| H A D | iavf_txrx.c | 2048 const skb_frag_t *frag, *stale; in __iavf_chk_linearize() local 2080 for (stale = &skb_shinfo(skb)->frags[0];; stale++) { in __iavf_chk_linearize() 2081 int stale_size = skb_frag_size(stale); in __iavf_chk_linearize() 2092 int align_pad = -(skb_frag_off(stale)) & in __iavf_chk_linearize()
|
| /linux/net/mptcp/ |
| H A D | pm.c | 228 if (!subflow->stale) { in mptcp_pm_addr_send_ack() 950 if (subflow->stale || !stale_loss_cnt || subflow->stale_count <= stale_loss_cnt) in mptcp_pm_subflows_chk_stale() 961 subflow->stale = 1; in mptcp_pm_subflows_chk_stale()
|