| /linux/include/net/ |
| H A D | busy_poll.h | 42 return READ_ONCE(sysctl_net_busy_poll); in net_busy_loop_on() 47 return READ_ONCE(sk->sk_ll_usec) && !signal_pending(current); in sk_can_busy_loop() 89 unsigned long bp_usec = READ_ONCE(sysctl_net_busy_poll); in busy_loop_timeout() 105 unsigned long bp_usec = READ_ONCE(sk->sk_ll_usec); in sk_busy_loop_timeout() 120 unsigned int napi_id = READ_ONCE(sk->sk_napi_id); in sk_busy_loop() 124 READ_ONCE(sk->sk_prefer_busy_poll), in sk_busy_loop() 125 READ_ONCE(sk->sk_busy_poll_budget) ?: BUSY_POLL_BUDGET); in sk_busy_loop() 152 if (unlikely(READ_ONCE(sk->sk_napi_id) != skb->napi_id)) in sk_mark_napi_id() 174 if (!READ_ONCE(sk->sk_napi_id)) in __sk_mark_napi_id_once()
|
| /linux/io_uring/ |
| H A D | cmd_net.c | 13 struct proto *prot = READ_ONCE(sk->sk_prot); in io_uring_cmd_get_sock_ioctl() 34 level = READ_ONCE(sqe->level); in io_uring_cmd_getsockopt() 38 optval = u64_to_user_ptr(READ_ONCE(sqe->optval)); in io_uring_cmd_getsockopt() 39 optname = READ_ONCE(sqe->optname); in io_uring_cmd_getsockopt() 40 optlen = READ_ONCE(sqe->optlen); in io_uring_cmd_getsockopt() 62 optval = u64_to_user_ptr(READ_ONCE(sqe->optval)); in io_uring_cmd_setsockopt() 63 optname = READ_ONCE(sqe->optname); in io_uring_cmd_setsockopt() 64 optlen = READ_ONCE(sqe->optlen); in io_uring_cmd_setsockopt() 65 level = READ_ONCE(sqe->level); in io_uring_cmd_setsockopt() 163 uaddr = u64_to_user_ptr(READ_ONCE(sqe->addr)); in io_uring_cmd_getsockname() [all …]
|
| H A D | advise.c | 39 ma->addr = READ_ONCE(sqe->addr); in io_madvise_prep() 40 ma->len = READ_ONCE(sqe->off); in io_madvise_prep() 42 ma->len = READ_ONCE(sqe->len); in io_madvise_prep() 43 ma->advice = READ_ONCE(sqe->fadvise_advice); in io_madvise_prep() 86 fa->offset = READ_ONCE(sqe->off); in io_fadvise_prep() 87 fa->len = READ_ONCE(sqe->addr); in io_fadvise_prep() 89 fa->len = READ_ONCE(sqe->len); in io_fadvise_prep() 90 fa->advice = READ_ONCE(sqe->fadvise_advice); in io_fadvise_prep()
|
| H A D | xattr.c | 51 name = u64_to_user_ptr(READ_ONCE(sqe->addr)); in __io_getxattr_prep() 52 ix->ctx.value = u64_to_user_ptr(READ_ONCE(sqe->addr2)); in __io_getxattr_prep() 53 ix->ctx.size = READ_ONCE(sqe->len); in __io_getxattr_prep() 54 ix->ctx.flags = READ_ONCE(sqe->xattr_flags); in __io_getxattr_prep() 92 path = u64_to_user_ptr(READ_ONCE(sqe->addr3)); in io_getxattr_prep() 130 name = u64_to_user_ptr(READ_ONCE(sqe->addr)); in __io_setxattr_prep() 131 ix->ctx.cvalue = u64_to_user_ptr(READ_ONCE(sqe->addr2)); in __io_setxattr_prep() 133 ix->ctx.size = READ_ONCE(sqe->len); in __io_setxattr_prep() 134 ix->ctx.flags = READ_ONCE(sqe->xattr_flags); in __io_setxattr_prep() 164 path = u64_to_user_ptr(READ_ONCE(sqe->addr3)); in io_setxattr_prep()
|
| H A D | wait.c | 86 if (iowq->cq_min_tail != READ_ONCE(rings->cq.tail)) in io_cqring_min_timer_wakeup() 133 if (!READ_ONCE(iowq->hit_timeout)) in io_cqring_schedule_timeout() 140 return READ_ONCE(iowq->hit_timeout) ? -ETIME : 0; in io_cqring_schedule_timeout() 171 if (unlikely(READ_ONCE(ctx->check_cq))) in io_cqring_wait_schedule() 220 iowq.cq_tail = READ_ONCE(rings->cq.head) + min_events; in io_cqring_wait() 221 iowq.cq_min_tail = READ_ONCE(rings->cq.tail); in io_cqring_wait() 222 nr_wait = (int) iowq.cq_tail - READ_ONCE(rings->cq.tail); in io_cqring_wait() 292 check_cq = READ_ONCE(ctx->check_cq); in io_cqring_wait() 313 READ_ONCE(io_get_rings(ctx)->cq.tail); in io_cqring_wait() 323 return READ_ONCE(io_get_rings(ctx)->cq.head) == READ_ONCE(io_get_rings(ctx)->cq.tail) ? ret : 0; in io_cqring_wait()
|
| H A D | openclose.c | 68 open->dfd = READ_ONCE(sqe->fd); in __io_openat_prep() 69 fname = u64_to_user_ptr(READ_ONCE(sqe->addr)); in __io_openat_prep() 75 open->file_slot = READ_ONCE(sqe->file_index); in __io_openat_prep() 97 u64 mode = READ_ONCE(sqe->len); in io_openat_prep() 98 u64 flags = READ_ONCE(sqe->open_flags); in io_openat_prep() 111 how = u64_to_user_ptr(READ_ONCE(sqe->addr2)); in io_openat2_prep() 112 len = READ_ONCE(sqe->len); in io_openat2_prep() 228 close->fd = READ_ONCE(sqe->fd); in io_close_prep() 229 close->file_slot = READ_ONCE(sqe->file_index); in io_close_prep() 288 flags = READ_ONCE(sqe->install_fd_flags); in io_install_fixed_fd_prep() [all …]
|
| /linux/tools/testing/selftests/kvm/ |
| H A D | irqfd_test.c | 55 while (!READ_ONCE(done)) { in secondary_irqfd_juggler() 56 juggle_eventfd_secondary(vm1, READ_ONCE(__eventfd)); in secondary_irqfd_juggler() 57 juggle_eventfd_secondary(vm2, READ_ONCE(__eventfd)); in secondary_irqfd_juggler() 123 kvm_irqfd(vm1, 11, READ_ONCE(__eventfd), KVM_IRQFD_FLAG_DEASSIGN); in main() 124 kvm_irqfd(vm1, 12, READ_ONCE(__eventfd), KVM_IRQFD_FLAG_DEASSIGN); in main() 125 kvm_irqfd(vm1, 13, READ_ONCE(__eventfd), KVM_IRQFD_FLAG_DEASSIGN); in main() 126 kvm_irqfd(vm1, 14, READ_ONCE(__eventfd), KVM_IRQFD_FLAG_DEASSIGN); in main() 127 kvm_irqfd(vm1, 10, READ_ONCE(__eventfd), KVM_IRQFD_FLAG_DEASSIGN); in main()
|
| H A D | steal_time.c | 37 GUEST_ASSERT(!(READ_ONCE(st->version) & 1)); in check_status() 38 GUEST_ASSERT_EQ(READ_ONCE(st->flags), 0); in check_status() 39 GUEST_ASSERT_EQ(READ_ONCE(st->preempted), 0); in check_status() 54 version = READ_ONCE(st->version); in guest_code() 59 GUEST_ASSERT(version < READ_ONCE(st->version)); in guest_code() 136 GUEST_ASSERT_EQ(READ_ONCE(st->rev), 0); in check_status() 137 GUEST_ASSERT_EQ(READ_ONCE(st->attr), 0); in check_status() 269 GUEST_ASSERT(!(READ_ONCE(st->sequence) & 1)); in check_status() 270 GUEST_ASSERT(READ_ONCE(st->flags) == 0); in check_status() 271 GUEST_ASSERT(READ_ONCE(st->preempted) == 0); in check_status() [all …]
|
| /linux/security/selinux/include/ |
| H A D | security.h | 127 return READ_ONCE(selinux_state.enforcing); in enforcing_enabled() 153 return READ_ONCE(selinux_state.policycap[POLICYDB_CAP_NETPEER]); in selinux_policycap_netpeer() 158 return READ_ONCE(selinux_state.policycap[POLICYDB_CAP_OPENPERM]); in selinux_policycap_openperm() 163 return READ_ONCE(selinux_state.policycap[POLICYDB_CAP_EXTSOCKCLASS]); in selinux_policycap_extsockclass() 168 return READ_ONCE(selinux_state.policycap[POLICYDB_CAP_ALWAYSNETWORK]); in selinux_policycap_alwaysnetwork() 173 return READ_ONCE(selinux_state.policycap[POLICYDB_CAP_CGROUPSECLABEL]); in selinux_policycap_cgroupseclabel() 178 return READ_ONCE( in selinux_policycap_nnp_nosuid_transition() 184 return READ_ONCE( in selinux_policycap_genfs_seclabel_symlinks() 190 return READ_ONCE( in selinux_policycap_ioctl_skip_cloexec() 196 return READ_ONCE( in selinux_policycap_userspace_initial_context() [all …]
|
| /linux/net/ipv4/ |
| H A D | tcp_timer.c | 35 user_timeout = READ_ONCE(icsk->icsk_user_timeout); in tcp_clamp_rto_to_user_timeout() 57 user_timeout = READ_ONCE(icsk->icsk_user_timeout); in tcp_clamp_probe0_to_user_timeout() 79 tcp_done_with_error(sk, READ_ONCE(sk->sk_err_soft) ? : ETIMEDOUT); in tcp_write_err() 118 if (READ_ONCE(sk->sk_err_soft)) in tcp_out_of_resources() 152 int retries = READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_orphan_retries); /* May be zero. */ in tcp_orphan_retries() 155 if (READ_ONCE(sk->sk_err_soft) && !alive) in tcp_orphan_retries() 172 if (!READ_ONCE(net->ipv4.sysctl_tcp_mtu_probing)) in tcp_mtu_probing() 180 mss = min(READ_ONCE(net->ipv4.sysctl_tcp_base_mss), mss); in tcp_mtu_probing() 181 mss = max(mss, READ_ONCE(net->ipv4.sysctl_tcp_mtu_probe_floor)); in tcp_mtu_probing() 182 mss = max(mss, READ_ONCE(net->ipv4.sysctl_tcp_min_snd_mss)); in tcp_mtu_probing() [all …]
|
| /linux/Documentation/litmus-tests/locking/ |
| H A D | DCL-broken.litmus | 21 r0 = READ_ONCE(*flag); 24 r1 = READ_ONCE(*flag); 31 r2 = READ_ONCE(*data); 40 r0 = READ_ONCE(*flag); 43 r1 = READ_ONCE(*flag); 50 r2 = READ_ONCE(*data);
|
| /linux/arch/arm64/include/asm/ |
| H A D | preempt.h | 12 return READ_ONCE(current_thread_info()->preempt.count); in preempt_count() 46 u32 pc = READ_ONCE(current_thread_info()->preempt.count); in __preempt_count_add() 53 u32 pc = READ_ONCE(current_thread_info()->preempt.count); in __preempt_count_sub() 61 u64 pc = READ_ONCE(ti->preempt_count); in __preempt_count_dec_and_test() 73 return !pc || !READ_ONCE(ti->preempt_count); in __preempt_count_dec_and_test() 78 u64 pc = READ_ONCE(current_thread_info()->preempt_count); in should_resched()
|
| /linux/drivers/net/ethernet/mellanox/mlx4/ |
| H A D | en_port.c | 165 packets += READ_ONCE(ring->packets); in mlx4_en_fold_software_stats() 166 bytes += READ_ONCE(ring->bytes); in mlx4_en_fold_software_stats() 176 packets += READ_ONCE(ring->packets); in mlx4_en_fold_software_stats() 177 bytes += READ_ONCE(ring->bytes); in mlx4_en_fold_software_stats() 254 sw_rx_dropped += READ_ONCE(ring->dropped); in mlx4_en_DUMP_ETH_STATS() 255 priv->port_stats.rx_chksum_good += READ_ONCE(ring->csum_ok); in mlx4_en_DUMP_ETH_STATS() 256 priv->port_stats.rx_chksum_none += READ_ONCE(ring->csum_none); in mlx4_en_DUMP_ETH_STATS() 257 priv->port_stats.rx_chksum_complete += READ_ONCE(ring->csum_complete); in mlx4_en_DUMP_ETH_STATS() 258 priv->port_stats.rx_alloc_pages += READ_ONCE(ring->rx_alloc_pages); in mlx4_en_DUMP_ETH_STATS() 259 priv->xdp_stats.rx_xdp_drop += READ_ONCE(ring->xdp_drop); in mlx4_en_DUMP_ETH_STATS() [all …]
|
| /linux/kernel/rcu/ |
| H A D | srcutiny.c | 109 newval = READ_ONCE(ssp->srcu_lock_nesting[idx]) - 1; in __srcu_read_unlock() 112 if (!newval && READ_ONCE(ssp->srcu_gp_waiting) && in_task() && !irqs_disabled()) in __srcu_read_unlock() 131 if (ssp->srcu_gp_running || ULONG_CMP_GE(ssp->srcu_idx, READ_ONCE(ssp->srcu_idx_max))) { in srcu_drive_gp() 151 !READ_ONCE(ssp->srcu_lock_nesting[idx]), HZ / 10); in srcu_drive_gp() 152 } while (READ_ONCE(ssp->srcu_lock_nesting[idx])); in srcu_drive_gp() 176 idx = ULONG_CMP_LT(ssp->srcu_idx, READ_ONCE(ssp->srcu_idx_max)); in srcu_drive_gp() 203 if (ULONG_CMP_GE(READ_ONCE(ssp->srcu_idx_max), cookie)) { in srcu_gp_start_if_needed() 207 if (!READ_ONCE(ssp->srcu_gp_running)) { in srcu_gp_start_if_needed() 271 ret = (READ_ONCE(ssp->srcu_idx) + 3) & ~0x1; in get_state_synchronize_srcu() 301 unsigned long cur_s = READ_ONCE(ssp->srcu_idx); in poll_state_synchronize_srcu()
|
| H A D | rcu_segcblist.h | 15 return READ_ONCE(rclp->len); in rcu_cblist_n_cbs() 45 return !READ_ONCE(rsclp->head); in rcu_segcblist_empty() 54 return READ_ONCE(rsclp->len); in rcu_segcblist_n_cbs() 73 return READ_ONCE(rsclp->flags) & flags; in rcu_segcblist_test_flags() 105 return !READ_ONCE(*READ_ONCE(rsclp->tails[seg])); in rcu_segcblist_restempty()
|
| /linux/net/sched/ |
| H A D | sch_dualpi2.c | 292 prob = READ_ONCE(q->pi2_prob); in must_drop() 629 qc = READ_ONCE(q->c_head_ts); in get_queue_delays() 630 ql = READ_ONCE(q->l_head_ts); in get_queue_delays() 975 step_in_pkts = READ_ONCE(q->step_in_packets); in dualpi2_dump() 976 step_th = READ_ONCE(q->step_thresh); in dualpi2_dump() 983 (nla_put_u32(skb, TCA_DUALPI2_LIMIT, READ_ONCE(sch->limit)) || in dualpi2_dump() 985 READ_ONCE(q->memory_limit)) || in dualpi2_dump() 987 convert_ns_to_usec(READ_ONCE(q->pi2_target))) || in dualpi2_dump() 989 convert_ns_to_usec(READ_ONCE(q->pi2_tupdate))) || in dualpi2_dump() 991 dualpi2_unscale_alpha_beta(READ_ONCE( in dualpi2_dump() [all...] |
| H A D | sch_pie.c | 477 ((u32)PSCHED_TICKS2NS(READ_ONCE(q->params.target))) / in pie_dump() 479 nla_put_u32(skb, TCA_PIE_LIMIT, READ_ONCE(sch->limit)) || in pie_dump() 481 jiffies_to_usecs(READ_ONCE(q->params.tupdate))) || in pie_dump() 482 nla_put_u32(skb, TCA_PIE_ALPHA, READ_ONCE(q->params.alpha)) || in pie_dump() 483 nla_put_u32(skb, TCA_PIE_BETA, READ_ONCE(q->params.beta)) || in pie_dump() 486 READ_ONCE(q->params.bytemode)) || in pie_dump() 488 READ_ONCE(q->params.dq_rate_estimator))) in pie_dump() 502 .prob = READ_ONCE(q->vars.prob) << BITS_PER_BYTE, in pie_dump_stats() 503 .delay = ((u32)PSCHED_TICKS2NS(READ_ONCE(q->vars.qdelay))) / in pie_dump_stats() 505 .packets_in = READ_ONCE(q->stats.packets_in), in pie_dump_stats() [all …]
|
| H A D | sch_choke.c | 438 u8 Wlog = READ_ONCE(q->parms.Wlog); in choke_dump() 441 .limit = READ_ONCE(q->limit), in choke_dump() 442 .flags = READ_ONCE(q->flags), in choke_dump() 443 .qth_min = READ_ONCE(q->parms.qth_min) >> Wlog, in choke_dump() 444 .qth_max = READ_ONCE(q->parms.qth_max) >> Wlog, in choke_dump() 446 .Plog = READ_ONCE(q->parms.Plog), in choke_dump() 447 .Scell_log = READ_ONCE(q->parms.Scell_log), in choke_dump() 455 nla_put_u32(skb, TCA_CHOKE_MAX_P, READ_ONCE(q->parms.max_P))) in choke_dump() 468 .early = READ_ONCE(q->stats.prob_drop) + in choke_dump_stats() 469 READ_ONCE(q->stats.forced_drop), in choke_dump_stats() [all …]
|
| /linux/arch/s390/lib/ |
| H A D | spinlock.c | 169 old = READ_ONCE(lp->lock); in arch_spin_lock_queued() 207 while (READ_ONCE(node->prev) != NULL) { in arch_spin_lock_queued() 221 old = READ_ONCE(lp->lock); in arch_spin_lock_queued() 241 while ((next = READ_ONCE(node->next)) == NULL) in arch_spin_lock_queued() 257 owner = arch_spin_yield_target(READ_ONCE(lp->lock), NULL); in arch_spin_lock_classic() 297 owner = READ_ONCE(lp->lock); in arch_spin_trylock_retry() 311 while (READ_ONCE(rw->cnts) & 0x10000) in arch_read_lock_wait() 323 while (READ_ONCE(rw->cnts) & 0x10000) in arch_read_lock_wait() 340 old = READ_ONCE(rw->cnts); in arch_write_lock_wait() 356 cpu = READ_ONCE(lp->lock) & _Q_LOCK_CPU_MASK; in arch_spin_relax()
|
| /linux/tools/memory-model/Documentation/ |
| H A D | locking.txt | 43 r0 = READ_ONCE(y); 45 r1 = READ_ONCE(x); 66 r0 = READ_ONCE(y); 68 r1 = READ_ONCE(x); 101 r0 = READ_ONCE(flag); 104 r1 = READ_ONCE(flag); 111 r2 = READ_ONCE(data); 116 READ_ONCE() of "flag" and the READ_ONCE() of "data". Second, there is 129 r1 = READ_ONCE(flag); 136 r2 = READ_ONCE(data); [all …]
|
| H A D | control-dependencies.txt | 15 q = READ_ONCE(a); 17 p = READ_ONCE(b); 25 q = READ_ONCE(a); 28 p = READ_ONCE(b); 35 q = READ_ONCE(a); 40 of ordering. But please note that neither the READ_ONCE() nor the 41 WRITE_ONCE() are optional. Without the READ_ONCE(), the compiler might 55 So don't leave out either the READ_ONCE() or the WRITE_ONCE(). 56 In particular, although READ_ONCE() does force the compiler to emit a 62 q = READ_ONCE(a); [all …]
|
| /linux/include/linux/ |
| H A D | srcutiny.h | 85 idx = ((READ_ONCE(ssp->srcu_idx) + 1) & 0x2) >> 1; in __srcu_read_lock() 86 WRITE_ONCE(ssp->srcu_lock_nesting[idx], READ_ONCE(ssp->srcu_lock_nesting[idx]) + 1); in __srcu_read_lock() 148 idx = ((data_race(READ_ONCE(ssp->srcu_idx)) + 1) & 0x2) >> 1; in srcu_torture_stats_print() 151 data_race(READ_ONCE(ssp->srcu_lock_nesting[!idx])), in srcu_torture_stats_print() 152 data_race(READ_ONCE(ssp->srcu_lock_nesting[idx])), in srcu_torture_stats_print() 153 data_race(READ_ONCE(ssp->srcu_idx)), in srcu_torture_stats_print() 154 data_race(READ_ONCE(ssp->srcu_idx_max))); in srcu_torture_stats_print()
|
| /linux/kernel/cgroup/ |
| H A D | misc.c | 159 if (!(valid_type(type) && cg && READ_ONCE(misc_res_capacity[type]))) in misc_cg_try_charge() 169 if (new_usage > READ_ONCE(res->max) || in misc_cg_try_charge() 170 new_usage > READ_ONCE(misc_res_capacity[type])) { in misc_cg_try_charge() 223 if (READ_ONCE(misc_res_capacity[i])) { in misc_cg_max_show() 224 max = READ_ONCE(cg->res[i].max); in misc_cg_max_show() 289 if (READ_ONCE(misc_res_capacity[type])) in misc_cg_max_write() 313 if (READ_ONCE(misc_res_capacity[i]) || usage) in misc_cg_current_show() 336 if (READ_ONCE(misc_res_capacity[i]) || watermark) in misc_cg_peak_show() 359 cap = READ_ONCE(misc_res_capacity[i]); in misc_cg_capacity_show() 378 if (READ_ONCE(misc_res_capacity[i]) || events) in __misc_events_show()
|
| /linux/arch/x86/include/asm/ |
| H A D | resctrl.h | 103 u32 closid = READ_ONCE(state->default_closid); in __resctrl_sched_in() 104 u32 rmid = READ_ONCE(state->default_rmid); in __resctrl_sched_in() 112 tmp = READ_ONCE(tsk->closid); in __resctrl_sched_in() 118 tmp = READ_ONCE(tsk->rmid); in __resctrl_sched_in() 155 return READ_ONCE(tsk->closid) == closid; in resctrl_arch_match_closid() 161 return READ_ONCE(tsk->rmid) == rmid; in resctrl_arch_match_rmid()
|
| /linux/drivers/iommu/intel/ |
| H A D | pasid.h | 57 return READ_ONCE(pde->val) & PASID_PTE_PRESENT; in pasid_pde_is_present() 67 return phys_to_virt(READ_ONCE(pde->val) & PDE_PFN_MASK); in get_pasid_table_from_pde() 73 return READ_ONCE(pte->val[0]) & PASID_PTE_PRESENT; in pasid_pte_is_present() 79 return READ_ONCE(pte->val[0]) & PASID_PTE_FPD; in pasid_pte_is_fault_disabled() 85 return (u16)((READ_ONCE(pte->val[0]) >> 6) & 0x7); in pasid_pte_get_pgtt() 116 old = READ_ONCE(*ptr); in pasid_set_bits() 122 return READ_ONCE(*ptr); in pasid_get_bits() 141 return (u16)(READ_ONCE(pe->val[1]) & GENMASK_ULL(15, 0)); in pasid_get_domain_id()
|