| /linux/fs/netfs/ |
| H A D | stats.c | 55 atomic_read(&netfs_n_rh_dio_read), in netfs_stats_show() 56 atomic_read(&netfs_n_rh_readahead), in netfs_stats_show() 57 atomic_read(&netfs_n_rh_read_folio), in netfs_stats_show() 58 atomic_read(&netfs_n_rh_read_single), in netfs_stats_show() 59 atomic_read(&netfs_n_rh_write_begin), in netfs_stats_show() 60 atomic_read(&netfs_n_rh_write_zskip)); in netfs_stats_show() 62 atomic_read(&netfs_n_wh_buffered_write), in netfs_stats_show() 63 atomic_read(&netfs_n_wh_writethrough), in netfs_stats_show() 64 atomic_read(&netfs_n_wh_dio_write), in netfs_stats_show() 65 atomic_read(&netfs_n_wh_writepages), in netfs_stats_show() [all …]
|
| H A D | fscache_stats.c | 61 atomic_read(&fscache_n_cookies), in fscache_stats_show() 62 atomic_read(&fscache_n_volumes), in fscache_stats_show() 63 atomic_read(&fscache_n_volumes_collision), in fscache_stats_show() 64 atomic_read(&fscache_n_volumes_nomem) in fscache_stats_show() 68 atomic_read(&fscache_n_acquires), in fscache_stats_show() 69 atomic_read(&fscache_n_acquires_ok), in fscache_stats_show() 70 atomic_read(&fscache_n_acquires_oom)); in fscache_stats_show() 73 atomic_read(&fscache_n_cookies_lru), in fscache_stats_show() 74 atomic_read(&fscache_n_cookies_lru_expired), in fscache_stats_show() 75 atomic_read(&fscache_n_cookies_lru_removed), in fscache_stats_show() [all …]
|
| /linux/net/rxrpc/ |
| H A D | proc.c | 183 atomic_read(&conn->active), in rxrpc_connection_seq_show() 258 atomic_read(&bundle->active), in rxrpc_bundle_seq_show() 425 atomic_read(&local->active_users), in rxrpc_local_seq_show() 481 atomic_read(&rxnet->stat_tx_data_send), in rxrpc_stats_show() 482 atomic_read(&rxnet->stat_tx_data_send_frag), in rxrpc_stats_show() 483 atomic_read(&rxnet->stat_tx_data_send_fail), in rxrpc_stats_show() 484 atomic_read(&rxnet->stat_tx_data_send_msgsize)); in rxrpc_stats_show() 487 atomic_read(&rxnet->stat_tx_data), in rxrpc_stats_show() 488 atomic_read(&rxnet->stat_tx_data_retrans), in rxrpc_stats_show() 489 atomic_read(&rxnet->stat_tx_data_underflow), in rxrpc_stats_show() [all …]
|
| /linux/net/netfilter/ipvs/ |
| H A D | ip_vs_nq.c | 44 return atomic_read(&dest->activeconns) + 1; in ip_vs_nq_dest_overhead() 76 !atomic_read(&dest->weight)) in ip_vs_nq_schedule() 82 if (atomic_read(&dest->activeconns) == 0) { in ip_vs_nq_schedule() 89 ((__s64)loh * atomic_read(&dest->weight) > in ip_vs_nq_schedule() 90 (__s64)doh * atomic_read(&least->weight))) { in ip_vs_nq_schedule() 106 atomic_read(&least->activeconns), in ip_vs_nq_schedule() 108 atomic_read(&least->weight), loh); in ip_vs_nq_schedule()
|
| H A D | ip_vs_lblcr.c | 172 if ((atomic_read(&least->weight) > 0) in ip_vs_dest_set_min() 188 if (((__s64)loh * atomic_read(&dest->weight) > in ip_vs_dest_set_min() 189 (__s64)doh * atomic_read(&least->weight)) in ip_vs_dest_set_min() 201 atomic_read(&least->activeconns), in ip_vs_dest_set_min() 203 atomic_read(&least->weight), loh); in ip_vs_dest_set_min() 221 if (atomic_read(&most->weight) > 0) { in ip_vs_dest_set_max() 234 if (((__s64)moh * atomic_read(&dest->weight) < in ip_vs_dest_set_max() 235 (__s64)doh * atomic_read(&most->weight)) in ip_vs_dest_set_max() 236 && (atomic_read(&dest->weight) > 0)) { in ip_vs_dest_set_max() 246 atomic_read(&most->activeconns), in ip_vs_dest_set_max() [all …]
|
| H A D | ip_vs_sed.c | 48 return atomic_read(&dest->activeconns) + 1; in ip_vs_sed_dest_overhead() 79 atomic_read(&dest->weight) > 0) { in ip_vs_sed_schedule() 96 if ((__s64)loh * atomic_read(&dest->weight) > in ip_vs_sed_schedule() 97 (__s64)doh * atomic_read(&least->weight)) { in ip_vs_sed_schedule() 107 atomic_read(&least->activeconns), in ip_vs_sed_schedule() 109 atomic_read(&least->weight), loh); in ip_vs_sed_schedule()
|
| H A D | ip_vs_wlc.c | 51 atomic_read(&dest->weight) > 0) { in ip_vs_wlc_schedule() 68 if ((__s64)loh * atomic_read(&dest->weight) > in ip_vs_wlc_schedule() 69 (__s64)doh * atomic_read(&least->weight)) { in ip_vs_wlc_schedule() 79 atomic_read(&least->activeconns), in ip_vs_wlc_schedule() 81 atomic_read(&least->weight), loh); in ip_vs_wlc_schedule()
|
| H A D | ip_vs_lblc.c | 310 if (atomic_read(&tbl->entries) <= tbl->max_size) { in ip_vs_lblc_check_expire() 315 goal = (atomic_read(&tbl->entries) - tbl->max_size)*4/3; in ip_vs_lblc_check_expire() 419 if (atomic_read(&dest->weight) > 0) { in __ip_vs_lblc_schedule() 436 if ((__s64)loh * atomic_read(&dest->weight) > in __ip_vs_lblc_schedule() 437 (__s64)doh * atomic_read(&least->weight)) { in __ip_vs_lblc_schedule() 447 atomic_read(&least->activeconns), in __ip_vs_lblc_schedule() 449 atomic_read(&least->weight), loh); in __ip_vs_lblc_schedule() 462 if (atomic_read(&dest->activeconns) > atomic_read(&dest->weight)) { in is_overloaded() 466 if (atomic_read(&d->activeconns)*2 in is_overloaded() 467 < atomic_read(&d->weight)) { in is_overloaded() [all …]
|
| H A D | ip_vs_fo.c | 33 atomic_read(&dest->weight) > hw) { in ip_vs_fo_schedule() 35 hw = atomic_read(&dest->weight); in ip_vs_fo_schedule() 43 atomic_read(&hweight->activeconns), in ip_vs_fo_schedule() 44 atomic_read(&hweight->weight)); in ip_vs_fo_schedule()
|
| H A D | ip_vs_ovf.c | 35 w = atomic_read(&dest->weight); in ip_vs_ovf_schedule() 37 atomic_read(&dest->activeconns) > w || in ip_vs_ovf_schedule() 50 atomic_read(&h->activeconns), in ip_vs_ovf_schedule() 51 atomic_read(&h->weight)); in ip_vs_ovf_schedule()
|
| H A D | ip_vs_twos.c | 56 weight = atomic_read(&dest->weight); in ip_vs_twos_schedule() 81 weight = atomic_read(&dest->weight); in ip_vs_twos_schedule() 110 ntohs(choice1->port), atomic_read(&choice1->activeconns), in ip_vs_twos_schedule() 112 atomic_read(&choice1->weight)); in ip_vs_twos_schedule()
|
| /linux/drivers/pinctrl/qcom/ |
| H A D | tlmm-test.c | 182 KUNIT_ASSERT_EQ(test, atomic_read(&priv->intr_count), 0); in tlmm_test_silent() 239 KUNIT_ASSERT_EQ(test, atomic_read(&priv->intr_count), 10); in tlmm_test_rising() 263 KUNIT_ASSERT_EQ(test, atomic_read(&priv->intr_count), 10); in tlmm_test_falling() 288 KUNIT_ASSERT_EQ(test, atomic_read(&priv->intr_count), 10); in tlmm_test_low() 313 KUNIT_ASSERT_EQ(test, atomic_read(&priv->intr_count), 10); in tlmm_test_high() 335 KUNIT_ASSERT_EQ(test, atomic_read(&priv->intr_count), 10); in tlmm_test_falling_in_handler() 357 KUNIT_ASSERT_EQ(test, atomic_read(&priv->intr_count), 10); in tlmm_test_rising_in_handler() 383 KUNIT_ASSERT_EQ(test, atomic_read(&priv->intr_count), 10); in tlmm_test_thread_rising() 384 KUNIT_ASSERT_EQ(test, atomic_read(&priv->thread_count), 10); in tlmm_test_thread_rising() 409 KUNIT_ASSERT_EQ(test, atomic_read(&priv->intr_count), 10); in tlmm_test_thread_falling() [all …]
|
| /linux/fs/afs/ |
| H A D | validation.c | 130 if (atomic_read(&volume->cb_v_check) != atomic_read(&volume->cb_v_break)) in afs_check_validity() 138 else if (vnode->cb_ro_snapshot != atomic_read(&volume->cb_ro_snapshot)) in afs_check_validity() 140 else if (vnode->cb_scrub != atomic_read(&volume->cb_scrub)) in afs_check_validity() 243 snap = atomic_read(&volume->cb_ro_snapshot); in afs_update_volume_creation_time() 334 unsigned int cb_v_break = atomic_read(&volume->cb_v_break); in afs_update_volume_state() 335 unsigned int cb_v_check = atomic_read(&volume->cb_v_check); in afs_update_volume_state() 421 atomic_read(&volume->cb_v_check) != atomic_read(&volume->cb_v_break)) { in afs_validate() 428 cb_ro_snapshot = atomic_read(&volume->cb_ro_snapshot); in afs_validate() 429 cb_scrub = atomic_read(&volume->cb_scrub); in afs_validate() 437 atomic_read(&volume->cb_v_check) != atomic_read(&volume->cb_v_break) || in afs_validate() [all …]
|
| /linux/drivers/infiniband/hw/vmw_pvrdma/ |
| H A D | pvrdma_ring.h | 71 const unsigned int idx = atomic_read(var); in pvrdma_idx() 80 __u32 idx = atomic_read(var) + 1; /* Increment. */ in pvrdma_idx_ring_inc() 89 const __u32 tail = atomic_read(&r->prod_tail); in pvrdma_idx_ring_has_space() 90 const __u32 head = atomic_read(&r->cons_head); in pvrdma_idx_ring_has_space() 103 const __u32 tail = atomic_read(&r->prod_tail); in pvrdma_idx_ring_has_data() 104 const __u32 head = atomic_read(&r->cons_head); in pvrdma_idx_ring_has_data()
|
| /linux/drivers/crypto/bcm/ |
| H A D | util.c | 376 atomic_read(&ipriv->session_count)); in spu_debugfs_read() 379 atomic_read(&ipriv->stream_count)); in spu_debugfs_read() 382 atomic_read(&ipriv->setkey_cnt[SPU_OP_CIPHER])); in spu_debugfs_read() 385 atomic_read(&ipriv->op_counts[SPU_OP_CIPHER])); in spu_debugfs_read() 388 op_cnt = atomic_read(&ipriv->cipher_cnt[alg][mode]); in spu_debugfs_read() 399 atomic_read(&ipriv->op_counts[SPU_OP_HASH])); in spu_debugfs_read() 401 op_cnt = atomic_read(&ipriv->hash_cnt[alg]); in spu_debugfs_read() 411 atomic_read(&ipriv->setkey_cnt[SPU_OP_HMAC])); in spu_debugfs_read() 414 atomic_read(&ipriv->op_counts[SPU_OP_HMAC])); in spu_debugfs_read() 416 op_cnt = atomic_read(&ipriv->hmac_cnt[alg]); in spu_debugfs_read() [all …]
|
| /linux/sound/core/seq/ |
| H A D | seq_lock.c | 16 if (atomic_read(lockp) < 0) { in snd_use_lock_sync_helper() 17 pr_warn("ALSA: seq_lock: lock trouble [counter = %d] in %s:%d\n", atomic_read(lockp), file, line); in snd_use_lock_sync_helper() 20 while (atomic_read(lockp) > 0) { in snd_use_lock_sync_helper() 22 pr_warn("ALSA: seq_lock: waiting [%d left] in %s:%d\n", atomic_read(lockp), file, line); in snd_use_lock_sync_helper()
|
| /linux/arch/mips/kernel/ |
| H A D | sync-r4k.c | 116 while (atomic_read(&start_count) != cpus - 1) in check_counter_sync_source() 126 while (atomic_read(&stop_count) != cpus-1) in check_counter_sync_source() 167 if (atomic_read(&test_runs) > 0) in check_counter_sync_source() 192 while (atomic_read(&start_count) != cpus) in synchronise_count_slave() 210 while (atomic_read(&stop_count) != cpus) in synchronise_count_slave() 223 if (!atomic_read(&test_runs)) { in synchronise_count_slave()
|
| /linux/include/kunit/ |
| H A D | run-in-irq-context.h | 36 task_calls = atomic_read(&state->task_func_calls); in kunit_irq_test_timer_func() 38 softirq_calls = atomic_read(&state->softirq_func_calls); in kunit_irq_test_timer_func() 128 hardirq_calls = atomic_read(&state.hardirq_func_calls); in kunit_run_irq_test() 129 softirq_calls = atomic_read(&state.softirq_func_calls); in kunit_run_irq_test() 140 KUNIT_EXPECT_GT_MSG(test, atomic_read(&state.hardirq_func_calls), 0, in kunit_run_irq_test() 142 KUNIT_EXPECT_GT_MSG(test, atomic_read(&state.softirq_func_calls), 0, in kunit_run_irq_test()
|
| /linux/arch/openrisc/kernel/ |
| H A D | sync-timer.c | 53 while (atomic_read(&count_count_start) != 1) in synchronise_count_master() 74 while (atomic_read(&count_count_stop) != 1) in synchronise_count_master() 104 while (atomic_read(&count_count_start) != 2) in synchronise_count_slave() 114 while (atomic_read(&count_count_stop) != 2) in synchronise_count_slave()
|
| /linux/net/mac80211/ |
| H A D | led.h | 16 if (!atomic_read(&local->rx_led_active)) in ieee80211_led_rx() 25 if (!atomic_read(&local->tx_led_active)) in ieee80211_led_tx() 74 if (atomic_read(&local->tpt_led_active)) in ieee80211_tpt_led_trig_tx() 83 if (atomic_read(&local->tpt_led_active)) in ieee80211_tpt_led_trig_rx()
|
| /linux/arch/x86/platform/uv/ |
| H A D | uv_nmi.c | 524 nmi = atomic_read(&hub_nmi->in_nmi); in uv_check_nmi() 552 nmi = atomic_read(&hub_nmi->in_nmi); in uv_check_nmi() 562 nmi = atomic_read(&uv_in_nmi); in uv_check_nmi() 584 if (cpu == atomic_read(&hub_nmi->cpu_owner)) { in uv_clear_nmi() 702 atomic_read(&uv_nmi_cpus_in_nmi), num_online_cpus()); in uv_nmi_wait() 770 while (atomic_read(&uv_nmi_cpus_in_nmi) > 0) in uv_nmi_sync_exit() 774 while (atomic_read(&uv_nmi_slave_continue)) in uv_nmi_sync_exit() 783 int in = atomic_read(&uv_nmi_cpus_in_nmi); in uv_nmi_action_health() 789 while (!atomic_read(&uv_nmi_slave_continue)) in uv_nmi_action_health() 805 atomic_read(&uv_nmi_cpus_in_nmi), cpu); in uv_nmi_dump_state() [all …]
|
| /linux/net/batman-adv/ |
| H A D | gateway_common.c | 31 gw_mode = atomic_read(&bat_priv->gw.mode); in batadv_gw_tvlv_container_update() 39 down = atomic_read(&bat_priv->gw.bandwidth_down); in batadv_gw_tvlv_container_update() 40 up = atomic_read(&bat_priv->gw.bandwidth_up); in batadv_gw_tvlv_container_update() 86 atomic_read(&bat_priv->gw.mode) == BATADV_GW_MODE_CLIENT) in batadv_gw_tvlv_ogm_handler_v1()
|
| /linux/drivers/base/test/ |
| H A D | test_async_driver_probe.c | 32 if (atomic_read(&timeout)) { in test_probe() 243 if (atomic_read(&async_completed) != async_id) { in test_async_probe_init() 247 } else if (!atomic_read(&errors) && !atomic_read(&warnings)) { in test_async_probe_init() 275 atomic_read(&errors), atomic_read(&warnings)); in test_async_probe_init()
|
| /linux/drivers/net/ethernet/aquantia/atlantic/ |
| H A D | aq_utils.h | 19 flags_old = atomic_read(flags); in aq_utils_obj_set() 29 flags_old = atomic_read(flags); in aq_utils_obj_clear() 36 return atomic_read(flags) & mask; in aq_utils_obj_test()
|
| /linux/drivers/s390/scsi/ |
| H A D | zfcp_qdio.c | 60 used = QDIO_MAX_BUFFERS_PER_Q - atomic_read(&qdio->req_q_free); in zfcp_qdio_account() 97 if (atomic_read(&qdio->req_q_free) < QDIO_MAX_BUFFERS_PER_Q) in zfcp_qdio_request_tasklet() 168 if (atomic_read(&qdio->req_q_free) < QDIO_MAX_BUFFERS_PER_Q) in zfcp_qdio_irq_tasklet() 267 if (atomic_read(&qdio->req_q_free) || in zfcp_qdio_sbal_check() 268 !(atomic_read(&qdio->adapter->status) & ZFCP_STATUS_ADAPTER_QDIOUP)) in zfcp_qdio_sbal_check() 290 if (!(atomic_read(&qdio->adapter->status) & ZFCP_STATUS_ADAPTER_QDIOUP)) in zfcp_qdio_sbal_get() 341 if (atomic_read(&qdio->req_q_free) <= 2 * ZFCP_QDIO_MAX_SBALS_PER_REQ) in zfcp_qdio_send() 396 if (!(atomic_read(&adapter->status) & ZFCP_STATUS_ADAPTER_QDIOUP)) in zfcp_qdio_close() 416 count = atomic_read(&qdio->req_q_free); in zfcp_qdio_close() 454 if (atomic_read(&adapter->status) & ZFCP_STATUS_ADAPTER_QDIOUP) in zfcp_qdio_open() [all …]
|