Home
last modified time | relevance | path

Searched refs:stale (Results 1 – 25 of 73) sorted by relevance

123

/linux/drivers/gpu/drm/i915/gt/
H A Dintel_gt_buffer_pool.c40 struct intel_gt_buffer_pool_node *node, *stale = NULL; in pool_free_older_than() local
68 node->free = stale; in pool_free_older_than()
69 stale = node; in pool_free_older_than()
80 while ((node = stale)) { in pool_free_older_than()
81 stale = stale->free; in pool_free_older_than()
/linux/Documentation/admin-guide/hw-vuln/
H A Dprocessor_mmio_stale_data.rst12 stale data into core fill buffers where the data can subsequently be inferred
21 Propagators are operations that result in stale data being copied or moved from
23 Vulnerabilities are operations that may result in stale data being directly
31 does not make stale data architecturally visible. Stale data must be propagated
36 The sideband stale data propagator (SSDP) is limited to the client (including
40 both requested data and unrequested stale data, from a transaction buffer and
41 the sideband response buffer. As a result, stale data from the sideband
46 The primary stale data propagator (PSDP) is limited to the client (including
50 buffer including both requested data and unrequested stale data. This is
51 similar to the sideband stale data propagator.
[all …]
H A Dgather_data_sampling.rst13 instruction that is transiently executed encounters a fault, stale data from
16 infer stale data using typical side channel techniques like cache timing
19 The attacker uses gather instructions to infer the stale vector register data.
29 Without mitigation, GDS can infer stale data across virtually all
/linux/fs/xfs/scrub/
H A Ddirtree.c294 dl->stale = true; in xchk_dirpath_revalidate()
381 if (dl->stale) { in xchk_dirpath_step_up()
481 if (dl->stale) { in xchk_dirpath_step_up()
582 if (!error && dl->stale) in xchk_dirpath_walk_upwards()
710 if (dl->stale || dl->aborted) in xchk_dirtree_live_update()
720 dl->stale = true; in xchk_dirtree_live_update()
750 dl->stale = false; in xchk_dirtree_reset()
840 ASSERT(dl->stale); in xchk_dirtree_find_paths_to_root()
848 } while (dl->stale); in xchk_dirtree_find_paths_to_root()
864 ASSERT(!dl->stale); in xchk_dirtree_evaluate()
H A Ddirtree_repair.c413 if (dl->stale) { in xrep_dirtree_unlink()
543 if (!error && dl->stale) in xrep_dirtree_delete_path()
684 if (dl->stale) { in xrep_dirtree_adopt()
747 if (!error && dl->stale) in xrep_dirtree_move_to_orphanage()
805 if (!dl->stale) { in xrep_dirtree()
H A Ddir.c636 unsigned int stale = 0; in xchk_directory_leaf1_bestfree() local
692 stale++; in xchk_directory_leaf1_bestfree()
694 if (leafhdr.stale != stale) in xchk_directory_leaf1_bestfree()
742 unsigned int stale = 0; in xchk_directory_free_bestfree() local
764 stale++; in xchk_directory_free_bestfree()
777 if (freehdr.nused + stale != freehdr.nvalid) in xchk_directory_free_bestfree()
/linux/drivers/md/bcache/
H A Dextents.c540 unsigned int i, stale; in bch_extent_bad() local
552 stale = ptr_stale(b->c, k, i); in bch_extent_bad()
554 if (stale && KEY_DIRTY(k)) { in bch_extent_bad()
557 stale, buf); in bch_extent_bad()
560 btree_bug_on(stale > BUCKET_GC_GEN_MAX, b, in bch_extent_bad()
562 stale, b->c->need_gc); in bch_extent_bad()
564 if (stale) in bch_extent_bad()
H A Dbtree.c1229 uint8_t stale = 0; in __bch_btree_mark_key() local
1239 return stale; in __bch_btree_mark_key()
1251 stale = max(stale, ptr_stale(c, k, i)); in __bch_btree_mark_key()
1275 return stale; in __bch_btree_mark_key()
1307 uint8_t stale = 0; in btree_gc_mark_node() local
1316 stale = max(stale, btree_mark_key(b, k)); in btree_gc_mark_node()
1338 if (stale > 10) in btree_gc_mark_node()
/linux/kernel/configs/
H A Dhardening.config63 # Initialize all heap variables to zero on free to reduce stale data lifetime.
69 # Wipe kernel stack after syscall completion to reduce stale data lifetime.
80 # Force IOMMU TLB invalidation so devices will never be able to access stale
/linux/fs/iomap/
H A Diter.c63 bool stale = iter->iomap.flags & IOMAP_F_STALE; in iomap_iter() local
101 else if (iter->len == 0 || (!advanced && !stale)) in iomap_iter()
/linux/drivers/gpu/drm/i915/gem/selftests/
H A Dmock_context.c31 spin_lock_init(&ctx->stale.lock); in mock_context()
32 INIT_LIST_HEAD(&ctx->stale.engines); in mock_context()
/linux/fs/xfs/
H A Dxfs_buf_item.c505 int stale = bip->bli_flags & XFS_BLI_STALE; in xfs_buf_item_unpin() local
526 if (stale) { in xfs_buf_item_unpin()
683 bool stale = bip->bli_flags & XFS_BLI_STALE; in xfs_buf_item_release() local
701 ASSERT(!stale || (bip->__bli_format.blf_flags & XFS_BLF_CANCEL)); in xfs_buf_item_release()
723 if (stale) { in xfs_buf_item_release()
758 if (stale || hold) in xfs_buf_item_release()
/linux/drivers/net/ethernet/intel/ice/
H A Dice_ptp.c487 if (test_and_clear_bit(idx, tx->stale)) in ice_ptp_complete_tx_single_tstamp()
648 if (test_and_clear_bit(idx, tx->stale)) in ice_ptp_process_tx_tstamp()
706 unsigned long *in_use, *stale; in ice_ptp_alloc_tx_tracker() local
711 stale = bitmap_zalloc(tx->len, GFP_KERNEL); in ice_ptp_alloc_tx_tracker()
713 if (!tstamps || !in_use || !stale) { in ice_ptp_alloc_tx_tracker()
716 bitmap_free(stale); in ice_ptp_alloc_tx_tracker()
723 tx->stale = stale; in ice_ptp_alloc_tx_tracker()
771 clear_bit(idx, tx->stale); in ice_ptp_flush_tx_tracker()
799 bitmap_or(tx->stale, tx->stale, tx->in_use, tx->len); in ice_ptp_mark_tx_tracker_stale()
846 bitmap_free(tx->stale); in ice_ptp_release_tx_tracker()
[all …]
H A Dice_ptp.h118 unsigned long *stale; member
H A Dice_txrx.c2015 const skb_frag_t *frag, *stale; in __ice_chk_linearize() local
2047 for (stale = &skb_shinfo(skb)->frags[0];; stale++) { in __ice_chk_linearize()
2048 int stale_size = skb_frag_size(stale); in __ice_chk_linearize()
2059 int align_pad = -(skb_frag_off(stale)) & in __ice_chk_linearize()
/linux/drivers/gpu/drm/i915/gem/
H A Di915_gem_context.c1084 spin_lock_irqsave(&ctx->stale.lock, flags); in engines_notify()
1086 spin_unlock_irqrestore(&ctx->stale.lock, flags); in engines_notify()
1435 spin_lock_irq(&ctx->stale.lock); in kill_context()
1437 list_for_each_entry_safe(pos, next, &ctx->stale.engines, link) { in kill_context()
1443 spin_unlock_irq(&ctx->stale.lock); in kill_context()
1448 spin_lock_irq(&ctx->stale.lock); in kill_context()
1455 spin_unlock_irq(&ctx->stale.lock); in kill_context()
1485 spin_lock_irq(&ctx->stale.lock); in engines_idle_release()
1487 list_add_tail(&engines->link, &ctx->stale.engines); in engines_idle_release()
1488 spin_unlock_irq(&ctx->stale.lock); in engines_idle_release()
[all …]
/linux/Documentation/translations/zh_CN/networking/
H A Dmptcp-sysctl.rst115 用于判定子流失效(stale)的MPTCP层重传间隔次数阈值。当指定
/linux/drivers/md/
H A Ddm-ps-historical-service-time.c346 u64 *out, u64 *stale) in hst_fill_compare() argument
353 *stale = pi->stale_after; in hst_fill_compare()
/linux/Documentation/scsi/
H A Dhpsa.rst80 and not be disrupted in any way by stale commands or other stale state
/linux/tools/testing/selftests/net/
H A Dndisc_unsolicited_na_test.sh145 to ${HOST_ADDR} dev ${ROUTER_INTF} nud stale)
H A Darp_ndisc_untracked_subnets.sh171 to ${HOST_ADDR_V6} dev ${ROUTER_INTF} nud stale)
/linux/Documentation/arch/x86/
H A Dmds.rst29 deallocated it can retain the stale data of the preceding operations which
37 contain stale data from a previous operation which can be forwarded to
193 that stale data from the idling CPU from spilling to the Hyper-Thread
/linux/drivers/net/wireless/ath/ath9k/
H A Dxmit.c447 tbf->bf_state.stale = false; in ath_clone_txbuf()
520 if (!bf->bf_state.stale || bf_next != NULL) in ath_tx_complete_aggr()
612 if (bf_next != NULL || !bf_last->bf_state.stale) in ath_tx_complete_aggr()
640 if (bf->bf_next == NULL && bf_last->bf_state.stale) { in ath_tx_complete_aggr()
958 bf->bf_state.stale = false; in ath_tx_get_tid_subframe()
1886 if (bf->bf_state.stale) { in ath_drain_txq_list()
2657 if (bf->bf_state.stale) { in ath_tx_processq()
2681 lastbf->bf_state.stale = true; in ath_tx_processq()
2762 if (bf->bf_state.stale) { in ath_tx_edma_tasklet()
2784 lastbf->bf_state.stale = true; in ath_tx_edma_tasklet()
/linux/drivers/net/ethernet/intel/iavf/
H A Diavf_txrx.c2048 const skb_frag_t *frag, *stale; in __iavf_chk_linearize() local
2080 for (stale = &skb_shinfo(skb)->frags[0];; stale++) { in __iavf_chk_linearize()
2081 int stale_size = skb_frag_size(stale); in __iavf_chk_linearize()
2092 int align_pad = -(skb_frag_off(stale)) & in __iavf_chk_linearize()
/linux/net/mptcp/
H A Dpm.c228 if (!subflow->stale) { in mptcp_pm_addr_send_ack()
950 if (subflow->stale || !stale_loss_cnt || subflow->stale_count <= stale_loss_cnt) in mptcp_pm_subflows_chk_stale()
961 subflow->stale = 1; in mptcp_pm_subflows_chk_stale()

123