/linux/kernel/time/ |
H A D | timekeeping.c | 261 } while (read_seqcount_retry(&tk_core.seq, seq)); in timekeeping_debug_get_ns() 831 } while (read_seqcount_retry(&tk_core.seq, seq)); in ktime_get_real_ts64() 852 } while (read_seqcount_retry(&tk_core.seq, seq)); in ktime_get() 869 } while (read_seqcount_retry(&tk_core.seq, seq)); in ktime_get_resolution_ns() 895 } while (read_seqcount_retry(&tk_core.seq, seq)); in ktime_get_with_offset() 916 } while (read_seqcount_retry(&tk_core.seq, seq)); in ktime_get_coarse_with_offset() 936 } while (read_seqcount_retry(&tk_core.seq, seq)); in ktime_mono_to_any() 957 } while (read_seqcount_retry(&tk_core.seq, seq)); in ktime_get_raw() 986 } while (read_seqcount_retry(&tk_core.seq, seq)); in ktime_get_ts64() 1035 } while (read_seqcount_retry(&tk_core.seq, seq)); in ktime_get_real_seconds() [all …]
|
H A D | jiffies.c | 56 } while (read_seqcount_retry(&jiffies_seq, seq)); in get_jiffies_64()
|
H A D | tick-sched.c | 84 } while (read_seqcount_retry(&jiffies_seq, seq)); in tick_do_update_jiffies64() 780 } while (read_seqcount_retry(&ts->idle_sleeptime_seq, seq)); in get_cpu_sleep_time_us() 878 } while (read_seqcount_retry(&jiffies_seq, seq)); in get_jiffies_update()
|
H A D | tick-common.c | 169 } while (read_seqcount_retry(&jiffies_seq, seq)); in tick_setup_periodic()
|
H A D | hrtimer.c | 1623 } while (read_seqcount_retry(&base->seq, seq) || in hrtimer_active()
|
/linux/fs/bcachefs/ |
H A D | btree_gc.h | 76 } while (read_seqcount_retry(&c->gc_pos_lock, seq)); in gc_visited()
|
/linux/net/netfilter/ |
H A D | nft_set_rbtree.c | 67 if (read_seqcount_retry(&priv->count, seq)) in __nft_rbtree_lookup() 125 if (ret || !read_seqcount_retry(&priv->count, seq)) in nft_rbtree_lookup() 148 if (read_seqcount_retry(&priv->count, seq)) in __nft_rbtree_get() 212 if (ret || !read_seqcount_retry(&priv->count, seq)) in nft_rbtree_get()
|
/linux/kernel/sched/ |
H A D | cputime.c | 833 } while (read_seqcount_retry(&vtime->seqcount, seq)); in task_gtime() 878 } while (read_seqcount_retry(&vtime->seqcount, seq)); in task_cputime() 968 } while (read_seqcount_retry(&vtime->seqcount, seq)); in kcpustat_field_vtime() 1055 } while (read_seqcount_retry(&vtime->seqcount, seq)); in kcpustat_cpu_fetch_vtime()
|
H A D | psi.c | 272 } while (read_seqcount_retry(&groupc->seq, seq)); in get_recent_times()
|
/linux/fs/ |
H A D | d_path.c | 246 } while (read_seqcount_retry(&fs->seq, seq)); in get_fs_root_rcu() 391 } while (read_seqcount_retry(&fs->seq, seq)); in get_fs_root_and_pwd_rcu()
|
H A D | dcache.c | 909 if (!read_seqcount_retry(&dentry->d_seq, seq)) in dget_parent() 2146 if (read_seqcount_retry(&dentry->d_seq, seq)) { in __d_lookup_rcu_op_compare() 2495 if (read_seqcount_retry(&dentry->d_seq, d_seq)) { in d_alloc_parallel() 3132 if (!read_seqcount_retry(&dentry->d_seq, seq)) in d_parent_ino()
|
/linux/include/linux/ |
H A D | u64_stats_sync.h | 177 return read_seqcount_retry(&syncp->seq, start); in __u64_stats_fetch_retry()
|
H A D | seqlock.h | 29 * read_seqcount_begin() and read_seqcount_retry(), however, there are more 35 * atomics; if there is a matching read_seqcount_retry() call, no following 269 * Return: count to be passed to read_seqcount_retry() 286 * Return: count to be passed to read_seqcount_retry() 294 * Return: count to be passed to read_seqcount_retry() 311 * Return: count to be passed to read_seqcount_retry() 329 * will fail the read_seqcount_retry() at the end of the read critical 336 * Return: count to be passed to read_seqcount_retry() 341 * If the counter is odd, let read_seqcount_retry() fail \ 352 * __read_seqcount_retry is like read_seqcount_retry, bu 397 #define read_seqcount_retry( global() macro [all...] |
/linux/lib/ |
H A D | flex_proportions.c | 159 } while (read_seqcount_retry(&p->sequence, seq)); in fprop_fraction_percpu()
|
/linux/net/core/ |
H A D | gen_estimator.c | 273 } while (read_seqcount_retry(&est->seq, seq)); in gen_estimator_read()
|
/linux/include/net/netfilter/ |
H A D | nf_conntrack.h | 348 } while (read_seqcount_retry(&nf_conntrack_generation, sequence)); in nf_conntrack_get_ht()
|
/linux/drivers/gpu/drm/i915/gt/ |
H A D | intel_gt_pm.c | 423 } while (read_seqcount_retry(>->stats.lock, seq)); in intel_gt_get_awake_time()
|
/linux/arch/x86/kernel/cpu/ |
H A D | aperfmperf.c | 516 } while (read_seqcount_retry(&s->seq, seq)); in arch_freq_get_on_cpu()
|
/linux/fs/btrfs/ |
H A D | volumes.h | 236 } while (read_seqcount_retry(&dev->data_seqcount, seq)); \
|
/linux/Documentation/locking/ |
H A D | seqlock.rst | 87 } while (read_seqcount_retry(&foo_seqcount, seq));
|
/linux/virt/kvm/ |
H A D | eventfd.c | 214 } while (read_seqcount_retry(&irqfd->irq_entry_sc, seq)); in irqfd_wakeup()
|
/linux/Documentation/filesystems/ |
H A D | path-lookup.rst | 703 ``read_seqcount_retry()`` or similar. 710 use ``read_seqcount_retry()`` to validate that copy. 712 ``read_seqcount_retry()`` not only checks the sequence number, but also 719 variables, then ``read_seqcount_retry()`` is called to confirm the two 722 instead. Notably it does *not* use ``read_seqcount_retry()``, but 724 isn't necessary. A subsequent ``read_seqcount_retry()`` will be
|
/linux/drivers/net/ethernet/fungible/funeth/ |
H A D | funeth_ethtool.c | 222 } while (read_seqcount_retry(&fp->link_seq, seq)); in fun_get_link_ksettings()
|
/linux/net/xfrm/ |
H A D | xfrm_policy.c | 1986 if (read_seqcount_retry(count, seq)) in xfrm_policy_lookup_inexact_addr() 2144 } while (read_seqcount_retry(&net->xfrm.xfrm_policy_hash_generation, sequence)); in xfrm_policy_lookup_bytype() 2178 if (read_seqcount_retry(&net->xfrm.xfrm_policy_hash_generation, sequence)) in xfrm_policy_lookup_bytype()
|
/linux/net/ipv4/netfilter/ |
H A D | arp_tables.c | 621 } while (read_seqcount_retry(s, start)); in get_counters()
|