Home
last modified time | relevance | path

Searched refs:READ_ONCE (Results 1 – 25 of 1264) sorted by relevance

12345678910>>...51

/linux/include/net/
H A Dbusy_poll.h42 return READ_ONCE(sysctl_net_busy_poll); in net_busy_loop_on()
47 return READ_ONCE(sk->sk_ll_usec) && !signal_pending(current); in sk_can_busy_loop()
89 unsigned long bp_usec = READ_ONCE(sysctl_net_busy_poll); in busy_loop_timeout()
105 unsigned long bp_usec = READ_ONCE(sk->sk_ll_usec); in sk_busy_loop_timeout()
120 unsigned int napi_id = READ_ONCE(sk->sk_napi_id); in sk_busy_loop()
124 READ_ONCE(sk->sk_prefer_busy_poll), in sk_busy_loop()
125 READ_ONCE(sk->sk_busy_poll_budget) ?: BUSY_POLL_BUDGET); in sk_busy_loop()
152 if (unlikely(READ_ONCE(sk->sk_napi_id) != skb->napi_id)) in sk_mark_napi_id()
174 if (!READ_ONCE(sk->sk_napi_id)) in __sk_mark_napi_id_once()
/linux/io_uring/
H A Dcmd_net.c13 struct proto *prot = READ_ONCE(sk->sk_prot); in io_uring_cmd_get_sock_ioctl()
34 level = READ_ONCE(sqe->level); in io_uring_cmd_getsockopt()
38 optval = u64_to_user_ptr(READ_ONCE(sqe->optval)); in io_uring_cmd_getsockopt()
39 optname = READ_ONCE(sqe->optname); in io_uring_cmd_getsockopt()
40 optlen = READ_ONCE(sqe->optlen); in io_uring_cmd_getsockopt()
62 optval = u64_to_user_ptr(READ_ONCE(sqe->optval)); in io_uring_cmd_setsockopt()
63 optname = READ_ONCE(sqe->optname); in io_uring_cmd_setsockopt()
64 optlen = READ_ONCE(sqe->optlen); in io_uring_cmd_setsockopt()
65 level = READ_ONCE(sqe->level); in io_uring_cmd_setsockopt()
163 uaddr = u64_to_user_ptr(READ_ONCE(sqe->addr)); in io_uring_cmd_getsockname()
[all …]
H A Dadvise.c39 ma->addr = READ_ONCE(sqe->addr); in io_madvise_prep()
40 ma->len = READ_ONCE(sqe->off); in io_madvise_prep()
42 ma->len = READ_ONCE(sqe->len); in io_madvise_prep()
43 ma->advice = READ_ONCE(sqe->fadvise_advice); in io_madvise_prep()
86 fa->offset = READ_ONCE(sqe->off); in io_fadvise_prep()
87 fa->len = READ_ONCE(sqe->addr); in io_fadvise_prep()
89 fa->len = READ_ONCE(sqe->len); in io_fadvise_prep()
90 fa->advice = READ_ONCE(sqe->fadvise_advice); in io_fadvise_prep()
H A Dxattr.c51 name = u64_to_user_ptr(READ_ONCE(sqe->addr)); in __io_getxattr_prep()
52 ix->ctx.value = u64_to_user_ptr(READ_ONCE(sqe->addr2)); in __io_getxattr_prep()
53 ix->ctx.size = READ_ONCE(sqe->len); in __io_getxattr_prep()
54 ix->ctx.flags = READ_ONCE(sqe->xattr_flags); in __io_getxattr_prep()
92 path = u64_to_user_ptr(READ_ONCE(sqe->addr3)); in io_getxattr_prep()
130 name = u64_to_user_ptr(READ_ONCE(sqe->addr)); in __io_setxattr_prep()
131 ix->ctx.cvalue = u64_to_user_ptr(READ_ONCE(sqe->addr2)); in __io_setxattr_prep()
133 ix->ctx.size = READ_ONCE(sqe->len); in __io_setxattr_prep()
134 ix->ctx.flags = READ_ONCE(sqe->xattr_flags); in __io_setxattr_prep()
164 path = u64_to_user_ptr(READ_ONCE(sqe->addr3)); in io_setxattr_prep()
H A Dwait.c86 if (iowq->cq_min_tail != READ_ONCE(rings->cq.tail)) in io_cqring_min_timer_wakeup()
133 if (!READ_ONCE(iowq->hit_timeout)) in io_cqring_schedule_timeout()
140 return READ_ONCE(iowq->hit_timeout) ? -ETIME : 0; in io_cqring_schedule_timeout()
171 if (unlikely(READ_ONCE(ctx->check_cq))) in io_cqring_wait_schedule()
220 iowq.cq_tail = READ_ONCE(rings->cq.head) + min_events; in io_cqring_wait()
221 iowq.cq_min_tail = READ_ONCE(rings->cq.tail); in io_cqring_wait()
222 nr_wait = (int) iowq.cq_tail - READ_ONCE(rings->cq.tail); in io_cqring_wait()
292 check_cq = READ_ONCE(ctx->check_cq); in io_cqring_wait()
313 READ_ONCE(io_get_rings(ctx)->cq.tail); in io_cqring_wait()
323 return READ_ONCE(io_get_rings(ctx)->cq.head) == READ_ONCE(io_get_rings(ctx)->cq.tail) ? ret : 0; in io_cqring_wait()
H A Dopenclose.c68 open->dfd = READ_ONCE(sqe->fd); in __io_openat_prep()
69 fname = u64_to_user_ptr(READ_ONCE(sqe->addr)); in __io_openat_prep()
75 open->file_slot = READ_ONCE(sqe->file_index); in __io_openat_prep()
97 u64 mode = READ_ONCE(sqe->len); in io_openat_prep()
98 u64 flags = READ_ONCE(sqe->open_flags); in io_openat_prep()
111 how = u64_to_user_ptr(READ_ONCE(sqe->addr2)); in io_openat2_prep()
112 len = READ_ONCE(sqe->len); in io_openat2_prep()
228 close->fd = READ_ONCE(sqe->fd); in io_close_prep()
229 close->file_slot = READ_ONCE(sqe->file_index); in io_close_prep()
288 flags = READ_ONCE(sqe->install_fd_flags); in io_install_fixed_fd_prep()
[all …]
/linux/tools/testing/selftests/kvm/
H A Dirqfd_test.c55 while (!READ_ONCE(done)) { in secondary_irqfd_juggler()
56 juggle_eventfd_secondary(vm1, READ_ONCE(__eventfd)); in secondary_irqfd_juggler()
57 juggle_eventfd_secondary(vm2, READ_ONCE(__eventfd)); in secondary_irqfd_juggler()
123 kvm_irqfd(vm1, 11, READ_ONCE(__eventfd), KVM_IRQFD_FLAG_DEASSIGN); in main()
124 kvm_irqfd(vm1, 12, READ_ONCE(__eventfd), KVM_IRQFD_FLAG_DEASSIGN); in main()
125 kvm_irqfd(vm1, 13, READ_ONCE(__eventfd), KVM_IRQFD_FLAG_DEASSIGN); in main()
126 kvm_irqfd(vm1, 14, READ_ONCE(__eventfd), KVM_IRQFD_FLAG_DEASSIGN); in main()
127 kvm_irqfd(vm1, 10, READ_ONCE(__eventfd), KVM_IRQFD_FLAG_DEASSIGN); in main()
H A Dsteal_time.c37 GUEST_ASSERT(!(READ_ONCE(st->version) & 1)); in check_status()
38 GUEST_ASSERT_EQ(READ_ONCE(st->flags), 0); in check_status()
39 GUEST_ASSERT_EQ(READ_ONCE(st->preempted), 0); in check_status()
54 version = READ_ONCE(st->version); in guest_code()
59 GUEST_ASSERT(version < READ_ONCE(st->version)); in guest_code()
136 GUEST_ASSERT_EQ(READ_ONCE(st->rev), 0); in check_status()
137 GUEST_ASSERT_EQ(READ_ONCE(st->attr), 0); in check_status()
269 GUEST_ASSERT(!(READ_ONCE(st->sequence) & 1)); in check_status()
270 GUEST_ASSERT(READ_ONCE(st->flags) == 0); in check_status()
271 GUEST_ASSERT(READ_ONCE(st->preempted) == 0); in check_status()
[all …]
/linux/security/selinux/include/
H A Dsecurity.h127 return READ_ONCE(selinux_state.enforcing); in enforcing_enabled()
153 return READ_ONCE(selinux_state.policycap[POLICYDB_CAP_NETPEER]); in selinux_policycap_netpeer()
158 return READ_ONCE(selinux_state.policycap[POLICYDB_CAP_OPENPERM]); in selinux_policycap_openperm()
163 return READ_ONCE(selinux_state.policycap[POLICYDB_CAP_EXTSOCKCLASS]); in selinux_policycap_extsockclass()
168 return READ_ONCE(selinux_state.policycap[POLICYDB_CAP_ALWAYSNETWORK]); in selinux_policycap_alwaysnetwork()
173 return READ_ONCE(selinux_state.policycap[POLICYDB_CAP_CGROUPSECLABEL]); in selinux_policycap_cgroupseclabel()
178 return READ_ONCE( in selinux_policycap_nnp_nosuid_transition()
184 return READ_ONCE( in selinux_policycap_genfs_seclabel_symlinks()
190 return READ_ONCE( in selinux_policycap_ioctl_skip_cloexec()
196 return READ_ONCE( in selinux_policycap_userspace_initial_context()
[all …]
/linux/net/ipv4/
H A Dtcp_timer.c35 user_timeout = READ_ONCE(icsk->icsk_user_timeout); in tcp_clamp_rto_to_user_timeout()
57 user_timeout = READ_ONCE(icsk->icsk_user_timeout); in tcp_clamp_probe0_to_user_timeout()
79 tcp_done_with_error(sk, READ_ONCE(sk->sk_err_soft) ? : ETIMEDOUT); in tcp_write_err()
118 if (READ_ONCE(sk->sk_err_soft)) in tcp_out_of_resources()
152 int retries = READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_orphan_retries); /* May be zero. */ in tcp_orphan_retries()
155 if (READ_ONCE(sk->sk_err_soft) && !alive) in tcp_orphan_retries()
172 if (!READ_ONCE(net->ipv4.sysctl_tcp_mtu_probing)) in tcp_mtu_probing()
180 mss = min(READ_ONCE(net->ipv4.sysctl_tcp_base_mss), mss); in tcp_mtu_probing()
181 mss = max(mss, READ_ONCE(net->ipv4.sysctl_tcp_mtu_probe_floor)); in tcp_mtu_probing()
182 mss = max(mss, READ_ONCE(net->ipv4.sysctl_tcp_min_snd_mss)); in tcp_mtu_probing()
[all …]
/linux/Documentation/litmus-tests/locking/
H A DDCL-broken.litmus21 r0 = READ_ONCE(*flag);
24 r1 = READ_ONCE(*flag);
31 r2 = READ_ONCE(*data);
40 r0 = READ_ONCE(*flag);
43 r1 = READ_ONCE(*flag);
50 r2 = READ_ONCE(*data);
/linux/arch/arm64/include/asm/
H A Dpreempt.h12 return READ_ONCE(current_thread_info()->preempt.count); in preempt_count()
46 u32 pc = READ_ONCE(current_thread_info()->preempt.count); in __preempt_count_add()
53 u32 pc = READ_ONCE(current_thread_info()->preempt.count); in __preempt_count_sub()
61 u64 pc = READ_ONCE(ti->preempt_count); in __preempt_count_dec_and_test()
73 return !pc || !READ_ONCE(ti->preempt_count); in __preempt_count_dec_and_test()
78 u64 pc = READ_ONCE(current_thread_info()->preempt_count); in should_resched()
/linux/drivers/net/ethernet/mellanox/mlx4/
H A Den_port.c165 packets += READ_ONCE(ring->packets); in mlx4_en_fold_software_stats()
166 bytes += READ_ONCE(ring->bytes); in mlx4_en_fold_software_stats()
176 packets += READ_ONCE(ring->packets); in mlx4_en_fold_software_stats()
177 bytes += READ_ONCE(ring->bytes); in mlx4_en_fold_software_stats()
254 sw_rx_dropped += READ_ONCE(ring->dropped); in mlx4_en_DUMP_ETH_STATS()
255 priv->port_stats.rx_chksum_good += READ_ONCE(ring->csum_ok); in mlx4_en_DUMP_ETH_STATS()
256 priv->port_stats.rx_chksum_none += READ_ONCE(ring->csum_none); in mlx4_en_DUMP_ETH_STATS()
257 priv->port_stats.rx_chksum_complete += READ_ONCE(ring->csum_complete); in mlx4_en_DUMP_ETH_STATS()
258 priv->port_stats.rx_alloc_pages += READ_ONCE(ring->rx_alloc_pages); in mlx4_en_DUMP_ETH_STATS()
259 priv->xdp_stats.rx_xdp_drop += READ_ONCE(ring->xdp_drop); in mlx4_en_DUMP_ETH_STATS()
[all …]
/linux/kernel/rcu/
H A Dsrcutiny.c109 newval = READ_ONCE(ssp->srcu_lock_nesting[idx]) - 1; in __srcu_read_unlock()
112 if (!newval && READ_ONCE(ssp->srcu_gp_waiting) && in_task() && !irqs_disabled()) in __srcu_read_unlock()
131 if (ssp->srcu_gp_running || ULONG_CMP_GE(ssp->srcu_idx, READ_ONCE(ssp->srcu_idx_max))) { in srcu_drive_gp()
151 !READ_ONCE(ssp->srcu_lock_nesting[idx]), HZ / 10); in srcu_drive_gp()
152 } while (READ_ONCE(ssp->srcu_lock_nesting[idx])); in srcu_drive_gp()
176 idx = ULONG_CMP_LT(ssp->srcu_idx, READ_ONCE(ssp->srcu_idx_max)); in srcu_drive_gp()
203 if (ULONG_CMP_GE(READ_ONCE(ssp->srcu_idx_max), cookie)) { in srcu_gp_start_if_needed()
207 if (!READ_ONCE(ssp->srcu_gp_running)) { in srcu_gp_start_if_needed()
271 ret = (READ_ONCE(ssp->srcu_idx) + 3) & ~0x1; in get_state_synchronize_srcu()
301 unsigned long cur_s = READ_ONCE(ssp->srcu_idx); in poll_state_synchronize_srcu()
H A Drcu_segcblist.h15 return READ_ONCE(rclp->len); in rcu_cblist_n_cbs()
45 return !READ_ONCE(rsclp->head); in rcu_segcblist_empty()
54 return READ_ONCE(rsclp->len); in rcu_segcblist_n_cbs()
73 return READ_ONCE(rsclp->flags) & flags; in rcu_segcblist_test_flags()
105 return !READ_ONCE(*READ_ONCE(rsclp->tails[seg])); in rcu_segcblist_restempty()
/linux/net/sched/
H A Dsch_dualpi2.c292 prob = READ_ONCE(q->pi2_prob); in must_drop()
629 qc = READ_ONCE(q->c_head_ts); in get_queue_delays()
630 ql = READ_ONCE(q->l_head_ts); in get_queue_delays()
975 step_in_pkts = READ_ONCE(q->step_in_packets); in dualpi2_dump()
976 step_th = READ_ONCE(q->step_thresh); in dualpi2_dump()
983 (nla_put_u32(skb, TCA_DUALPI2_LIMIT, READ_ONCE(sch->limit)) || in dualpi2_dump()
985 READ_ONCE(q->memory_limit)) || in dualpi2_dump()
987 convert_ns_to_usec(READ_ONCE(q->pi2_target))) || in dualpi2_dump()
989 convert_ns_to_usec(READ_ONCE(q->pi2_tupdate))) || in dualpi2_dump()
991 dualpi2_unscale_alpha_beta(READ_ONCE( in dualpi2_dump()
[all...]
H A Dsch_pie.c477 ((u32)PSCHED_TICKS2NS(READ_ONCE(q->params.target))) / in pie_dump()
479 nla_put_u32(skb, TCA_PIE_LIMIT, READ_ONCE(sch->limit)) || in pie_dump()
481 jiffies_to_usecs(READ_ONCE(q->params.tupdate))) || in pie_dump()
482 nla_put_u32(skb, TCA_PIE_ALPHA, READ_ONCE(q->params.alpha)) || in pie_dump()
483 nla_put_u32(skb, TCA_PIE_BETA, READ_ONCE(q->params.beta)) || in pie_dump()
486 READ_ONCE(q->params.bytemode)) || in pie_dump()
488 READ_ONCE(q->params.dq_rate_estimator))) in pie_dump()
502 .prob = READ_ONCE(q->vars.prob) << BITS_PER_BYTE, in pie_dump_stats()
503 .delay = ((u32)PSCHED_TICKS2NS(READ_ONCE(q->vars.qdelay))) / in pie_dump_stats()
505 .packets_in = READ_ONCE(q->stats.packets_in), in pie_dump_stats()
[all …]
H A Dsch_choke.c438 u8 Wlog = READ_ONCE(q->parms.Wlog); in choke_dump()
441 .limit = READ_ONCE(q->limit), in choke_dump()
442 .flags = READ_ONCE(q->flags), in choke_dump()
443 .qth_min = READ_ONCE(q->parms.qth_min) >> Wlog, in choke_dump()
444 .qth_max = READ_ONCE(q->parms.qth_max) >> Wlog, in choke_dump()
446 .Plog = READ_ONCE(q->parms.Plog), in choke_dump()
447 .Scell_log = READ_ONCE(q->parms.Scell_log), in choke_dump()
455 nla_put_u32(skb, TCA_CHOKE_MAX_P, READ_ONCE(q->parms.max_P))) in choke_dump()
468 .early = READ_ONCE(q->stats.prob_drop) + in choke_dump_stats()
469 READ_ONCE(q->stats.forced_drop), in choke_dump_stats()
[all …]
/linux/arch/s390/lib/
H A Dspinlock.c169 old = READ_ONCE(lp->lock); in arch_spin_lock_queued()
207 while (READ_ONCE(node->prev) != NULL) { in arch_spin_lock_queued()
221 old = READ_ONCE(lp->lock); in arch_spin_lock_queued()
241 while ((next = READ_ONCE(node->next)) == NULL) in arch_spin_lock_queued()
257 owner = arch_spin_yield_target(READ_ONCE(lp->lock), NULL); in arch_spin_lock_classic()
297 owner = READ_ONCE(lp->lock); in arch_spin_trylock_retry()
311 while (READ_ONCE(rw->cnts) & 0x10000) in arch_read_lock_wait()
323 while (READ_ONCE(rw->cnts) & 0x10000) in arch_read_lock_wait()
340 old = READ_ONCE(rw->cnts); in arch_write_lock_wait()
356 cpu = READ_ONCE(lp->lock) & _Q_LOCK_CPU_MASK; in arch_spin_relax()
/linux/tools/memory-model/Documentation/
H A Dlocking.txt43 r0 = READ_ONCE(y);
45 r1 = READ_ONCE(x);
66 r0 = READ_ONCE(y);
68 r1 = READ_ONCE(x);
101 r0 = READ_ONCE(flag);
104 r1 = READ_ONCE(flag);
111 r2 = READ_ONCE(data);
116 READ_ONCE() of "flag" and the READ_ONCE() of "data". Second, there is
129 r1 = READ_ONCE(flag);
136 r2 = READ_ONCE(data);
[all …]
H A Dcontrol-dependencies.txt15 q = READ_ONCE(a);
17 p = READ_ONCE(b);
25 q = READ_ONCE(a);
28 p = READ_ONCE(b);
35 q = READ_ONCE(a);
40 of ordering. But please note that neither the READ_ONCE() nor the
41 WRITE_ONCE() are optional. Without the READ_ONCE(), the compiler might
55 So don't leave out either the READ_ONCE() or the WRITE_ONCE().
56 In particular, although READ_ONCE() does force the compiler to emit a
62 q = READ_ONCE(a);
[all …]
/linux/include/linux/
H A Dsrcutiny.h85 idx = ((READ_ONCE(ssp->srcu_idx) + 1) & 0x2) >> 1; in __srcu_read_lock()
86 WRITE_ONCE(ssp->srcu_lock_nesting[idx], READ_ONCE(ssp->srcu_lock_nesting[idx]) + 1); in __srcu_read_lock()
148 idx = ((data_race(READ_ONCE(ssp->srcu_idx)) + 1) & 0x2) >> 1; in srcu_torture_stats_print()
151 data_race(READ_ONCE(ssp->srcu_lock_nesting[!idx])), in srcu_torture_stats_print()
152 data_race(READ_ONCE(ssp->srcu_lock_nesting[idx])), in srcu_torture_stats_print()
153 data_race(READ_ONCE(ssp->srcu_idx)), in srcu_torture_stats_print()
154 data_race(READ_ONCE(ssp->srcu_idx_max))); in srcu_torture_stats_print()
/linux/kernel/cgroup/
H A Dmisc.c159 if (!(valid_type(type) && cg && READ_ONCE(misc_res_capacity[type]))) in misc_cg_try_charge()
169 if (new_usage > READ_ONCE(res->max) || in misc_cg_try_charge()
170 new_usage > READ_ONCE(misc_res_capacity[type])) { in misc_cg_try_charge()
223 if (READ_ONCE(misc_res_capacity[i])) { in misc_cg_max_show()
224 max = READ_ONCE(cg->res[i].max); in misc_cg_max_show()
289 if (READ_ONCE(misc_res_capacity[type])) in misc_cg_max_write()
313 if (READ_ONCE(misc_res_capacity[i]) || usage) in misc_cg_current_show()
336 if (READ_ONCE(misc_res_capacity[i]) || watermark) in misc_cg_peak_show()
359 cap = READ_ONCE(misc_res_capacity[i]); in misc_cg_capacity_show()
378 if (READ_ONCE(misc_res_capacity[i]) || events) in __misc_events_show()
/linux/arch/x86/include/asm/
H A Dresctrl.h103 u32 closid = READ_ONCE(state->default_closid); in __resctrl_sched_in()
104 u32 rmid = READ_ONCE(state->default_rmid); in __resctrl_sched_in()
112 tmp = READ_ONCE(tsk->closid); in __resctrl_sched_in()
118 tmp = READ_ONCE(tsk->rmid); in __resctrl_sched_in()
155 return READ_ONCE(tsk->closid) == closid; in resctrl_arch_match_closid()
161 return READ_ONCE(tsk->rmid) == rmid; in resctrl_arch_match_rmid()
/linux/drivers/iommu/intel/
H A Dpasid.h57 return READ_ONCE(pde->val) & PASID_PTE_PRESENT; in pasid_pde_is_present()
67 return phys_to_virt(READ_ONCE(pde->val) & PDE_PFN_MASK); in get_pasid_table_from_pde()
73 return READ_ONCE(pte->val[0]) & PASID_PTE_PRESENT; in pasid_pte_is_present()
79 return READ_ONCE(pte->val[0]) & PASID_PTE_FPD; in pasid_pte_is_fault_disabled()
85 return (u16)((READ_ONCE(pte->val[0]) >> 6) & 0x7); in pasid_pte_get_pgtt()
116 old = READ_ONCE(*ptr); in pasid_set_bits()
122 return READ_ONCE(*ptr); in pasid_get_bits()
141 return (u16)(READ_ONCE(pe->val[1]) & GENMASK_ULL(15, 0)); in pasid_get_domain_id()

12345678910>>...51