| /linux/kernel/ |
| H A D | rseq.c | 136 stats.exit += data_race(per_cpu(rseq_stats.exit, cpu)); in rseq_stats_show() 137 stats.signal += data_race(per_cpu(rseq_stats.signal, cpu)); in rseq_stats_show() 138 stats.slowpath += data_race(per_cpu(rseq_stats.slowpath, cpu)); in rseq_stats_show() 139 stats.fastpath += data_race(per_cpu(rseq_stats.fastpath, cpu)); in rseq_stats_show() 140 stats.ids += data_race(per_cpu(rseq_stats.ids, cpu)); in rseq_stats_show() 141 stats.cs += data_race(per_cpu(rseq_stats.cs, cpu)); in rseq_stats_show() 142 stats.clear += data_race(per_cpu(rseq_stats.clear, cpu)); in rseq_stats_show() 143 stats.fixup += data_race(per_cpu(rseq_stats.fixup, cpu)); in rseq_stats_show() 145 stats.s_granted += data_race(per_cpu(rseq_stats.s_granted, cpu)); in rseq_stats_show() 146 stats.s_expired += data_race(per_cpu(rseq_stats.s_expired, cpu)); in rseq_stats_show() [all …]
|
| /linux/kernel/rcu/ |
| H A D | tasks.h | 235 int i = data_race(rtp->gp_state); // Let KCSAN detect update races in tasks_gp_state_getname() 299 rtp->name, data_race(rtp->percpu_enqueue_shift), data_race(rtp->percpu_enqueue_lim), in cblist_init_generic() 728 if (!data_race(rcu_segcblist_empty(&rtpcp->cblist))) in show_rcu_tasks_generic_gp_kthread() 730 if (data_race(rtpcp->urgent_gp)) in show_rcu_tasks_generic_gp_kthread() 732 if (!data_race(rcu_segcblist_empty(&rtpcp->cblist)) && data_race(rtpcp->urgent_gp)) in show_rcu_tasks_generic_gp_kthread() 739 tasks_gp_state_getname(rtp), data_race(rtp->gp_state), in show_rcu_tasks_generic_gp_kthread() 740 jiffies - data_race(rtp->gp_jiffies), in show_rcu_tasks_generic_gp_kthread() 741 data_race(rcu_seq_current(&rtp->tasks_gp_seq)), in show_rcu_tasks_generic_gp_kthread() 742 data_race(rtp->n_ipis_fails), data_race(rtp->n_ipis), in show_rcu_tasks_generic_gp_kthread() 743 ".k"[!!data_race(rtp->kthread_ptr)], in show_rcu_tasks_generic_gp_kthread() [all …]
|
| H A D | srcutree.c | 2058 u0 = data_race(atomic_long_read(&sdp->srcu_ctrs[!idx].srcu_unlocks)); in srcu_torture_stats_print() 2059 u1 = data_race(atomic_long_read(&sdp->srcu_ctrs[idx].srcu_unlocks)); in srcu_torture_stats_print() 2067 l0 = data_race(atomic_long_read(&sdp->srcu_ctrs[!idx].srcu_locks)); in srcu_torture_stats_print() 2068 l1 = data_race(atomic_long_read(&sdp->srcu_ctrs[idx].srcu_locks)); in srcu_torture_stats_print()
|
| /linux/tools/memory-model/Documentation/ |
| H A D | access-marking.txt | 20 2. Data-race marking, for example, "data_race(a = b);" 37 WRITE_ONCE(a, b + data_race(c + d) + READ_ONCE(e)); 39 Neither plain C-language accesses nor data_race() (#1 and #2 above) place 45 preferable to data_race(), which in turn is usually preferable to plain 47 data_race(READ_ONCE(a)), which will both restrict compiler optimizations 52 race with one of data_race(), READ_ONCE(), or WRITE_ONCE(), will prevent 56 ill-considered additions of data_race(), READ_ONCE(), and WRITE_ONCE() 60 data_race() and even plain C-language accesses is preferable to 64 Use of the data_race() Macro 67 Here are some situations where data_race() should be used instead of [all …]
|
| /linux/include/linux/ |
| H A D | srcutiny.h | 148 idx = ((data_race(READ_ONCE(ssp->srcu_idx)) + 1) & 0x2) >> 1; 151 data_race(READ_ONCE(ssp->srcu_lock_nesting[!idx])), 152 data_race(READ_ONCE(ssp->srcu_lock_nesting[idx])), 153 data_race(READ_ONCE(ssp->srcu_idx)), 154 data_race(READ_ONCE(ssp->srcu_idx_max)));
|
| H A D | page-flags.h | 954 return page_type_has_type(data_race(page->page_type)); in page_has_type() 960 return data_race(folio->page.page_type >> 24) == PGTY_##lname; \ 966 VM_BUG_ON_FOLIO(data_race(folio->page.page_type) != UINT_MAX, \ 982 return data_race(page->page_type >> 24) == PGTY_##lname; \ 988 VM_BUG_ON_PAGE(data_race(page->page_type) != UINT_MAX, page); \
|
| /linux/security/tomoyo/ |
| H A D | util.c | 1069 perm = data_race(container_of(ptr, struct tomoyo_path_acl, head)->perm); in tomoyo_domain_quota_is_ok() 1072 perm = data_race(container_of(ptr, struct tomoyo_path2_acl, head)->perm); in tomoyo_domain_quota_is_ok() 1075 perm = data_race(container_of(ptr, struct tomoyo_path_number_acl, head) in tomoyo_domain_quota_is_ok() 1079 perm = data_race(container_of(ptr, struct tomoyo_mkdev_acl, head)->perm); in tomoyo_domain_quota_is_ok() 1082 perm = data_race(container_of(ptr, struct tomoyo_inet_acl, head)->perm); in tomoyo_domain_quota_is_ok() 1085 perm = data_race(container_of(ptr, struct tomoyo_unix_acl, head)->perm); in tomoyo_domain_quota_is_ok()
|
| /linux/tools/virtio/linux/ |
| H A D | compiler.h | 36 #define data_race(expr) \ macro
|
| /linux/Documentation/translations/zh_CN/dev-tools/ |
| H A D | kcsan.rst | 88 * KCSAN 支持使用 ``data_race(expr)`` 注解,这个注解告诉 KCSAN 任何由访问 92 * 与 ``data_race(...)`` 相似,可以使用类型限定符 ``__data_racy`` 来标记一个变量
|
| /linux/lib/ |
| H A D | cmpxchg-emu.c | 41 ret = data_race(cmpxchg(p32, old32.w, new32.w)); // Overridden above. in cmpxchg_emu_u8()
|
| H A D | stackdepot.c | 618 if (data_race(stackdepot_memcmp(entries, stack->entries, size))) in find_stack() 859 seq_printf(seq, "pools: %d\n", data_race(pools_num)); in stats_show() 861 seq_printf(seq, "%s: %ld\n", counter_names[i], data_race(counters[i])); in stats_show()
|
| H A D | group_cpus.c | 527 cpumask_copy(npresmsk, data_race(cpu_present_mask)); in group_cpus_evenly()
|
| /linux/arch/powerpc/kernel/ |
| H A D | interrupt.c | 172 if (!data_race(warned)) { in check_return_regs_valid() 173 data_race(warned = true); in check_return_regs_valid()
|
| /linux/io_uring/ |
| H A D | refs.h | 23 WARN_ON_ONCE(!(data_race(req->flags) & REQ_F_REFCOUNT)); in req_ref_put_and_test_atomic()
|
| H A D | fdinfo.c | 87 seq_printf(m, "CachedSqHead:\t%u\n", data_race(ctx->cached_sq_head)); in __io_uring_show_fdinfo() 91 seq_printf(m, "CachedCqTail:\t%u\n", data_race(ctx->cached_cq_tail)); in __io_uring_show_fdinfo()
|
| /linux/mm/ |
| H A D | page_io.c | 454 * but that will never affect SWP_FS_OPS, so the data_race in __swap_writepage() 457 if (data_race(sis->flags & SWP_FS_OPS)) in __swap_writepage() 461 * but that will never affect SWP_SYNCHRONOUS_IO, so the data_race in __swap_writepage() 464 else if (data_race(sis->flags & SWP_SYNCHRONOUS_IO)) in __swap_writepage() 643 if (data_race(sis->flags & SWP_FS_OPS)) { in swap_read_folio()
|
| H A D | vma_init.c | 59 data_race(memcpy(&dest->shared, &src->shared, sizeof(dest->shared))); in vm_area_init_from()
|
| /linux/kernel/kcsan/ |
| H A D | debugfs.c | 157 if (data_race(report_filterlist.used == report_filterlist.size)) { in insert_report_filterlist()
|
| /linux/tools/testing/vma/include/ |
| H A D | stubs.h | 32 #define data_race(expr) expr macro
|
| /linux/kernel/locking/ |
| H A D | locktorture.c | 1006 long max = 0, min = statp ? data_race(statp[0].n_lock_acquired) : 0; in __torture_print_stats() 1011 if (data_race(statp[i].n_lock_fail)) in __torture_print_stats() 1013 cur = data_race(statp[i].n_lock_acquired); in __torture_print_stats()
|
| /linux/fs/jbd2/ |
| H A D | transaction.c | 344 if (!data_race(journal->j_running_transaction)) { in start_this_handle() 1515 if (data_race(jh->b_transaction != transaction && in jbd2_journal_dirty_metadata() 1522 if (data_race(jh->b_modified == 1)) { in jbd2_journal_dirty_metadata() 1524 if (data_race(jh->b_transaction == transaction && in jbd2_journal_dirty_metadata()
|
| /linux/kernel/irq/ |
| H A D | irqdesc.c | 1010 return data_race(desc->tot_count); in kstat_irqs_desc() 1013 sum += data_race(per_cpu(desc->kstat_irqs->cnt, cpu)); in kstat_irqs_desc()
|
| /linux/drivers/net/wireless/ath/ath9k/ |
| H A D | wmi.c | 159 if (!data_race(priv->initialized)) { in ath9k_wmi_event_tasklet()
|
| /linux/mm/kfence/ |
| H A D | core.c | 1288 distance = addr - data_race(meta->addr + meta->size); in kfence_handle_page_fault() 1294 if (!to_report || distance > data_race(meta->addr) - addr) in kfence_handle_page_fault()
|
| /linux/net/9p/ |
| H A D | trans_fd.c | 739 data_race(ts->rd->f_flags |= O_NONBLOCK); in p9_fd_open() 745 data_race(ts->wr->f_flags |= O_NONBLOCK); in p9_fd_open()
|