| /linux/kernel/sched/ |
| H A D | loadavg.c | 82 long nr_active, delta = 0; in calc_load_fold_active() local 84 nr_active = this_rq->nr_running - adjust; in calc_load_fold_active() 85 nr_active += (long)this_rq->nr_uninterruptible; in calc_load_fold_active() 87 if (nr_active != this_rq->calc_load_active) { in calc_load_fold_active() 88 delta = nr_active - this_rq->calc_load_active; in calc_load_fold_active() 89 this_rq->calc_load_active = nr_active; in calc_load_fold_active()
|
| /linux/arch/powerpc/platforms/cell/spufs/ |
| H A D | sched.c | 680 cbe_spu_info[node].nr_active--; in find_victim() 709 cbe_spu_info[node].nr_active++; in __spu_schedule() 750 cbe_spu_info[node].nr_active--; in spu_unschedule() 951 int nr_active = 0, node; in count_active_contexts() local 954 nr_active += cbe_spu_info[node].nr_active; in count_active_contexts() 955 nr_active += spu_prio->nr_waiting; in count_active_contexts() 957 return nr_active; in count_active_contexts()
|
| /linux/tools/sched_ext/ |
| H A D | scx_flatcg.bpf.c | 196 if (!cgc->nr_active) { in cgrp_refresh_hweight() 233 is_active = cgc->nr_active; in cgrp_refresh_hweight() 430 * ->nr_active updates, so that we don't have to grab the cgv_tree_lock in update_active_weight_sums() 477 if (!cgc->nr_active++) { in update_active_weight_sums() 485 if (!--cgc->nr_active) { in update_active_weight_sums() 603 if (pcgc && cgc->nr_active) in BPF_STRUCT_OPS()
|
| H A D | scx_flatcg.h | 40 u32 nr_active; member
|
| /linux/block/ |
| H A D | blk-mq.h | 300 atomic_add(val, &hctx->nr_active); in __blk_mq_add_active_requests() 314 atomic_sub(val, &hctx->nr_active); in __blk_mq_sub_active_requests() 352 return atomic_read(&hctx->nr_active); in __blk_mq_active_requests()
|
| H A D | blk-mq.c | 4061 atomic_set(&hctx->nr_active, 0); in blk_mq_alloc_hctx()
|
| /linux/tools/testing/selftests/bpf/progs/ |
| H A D | net_timestamping.c | 11 int nr_active; variable 225 nr_active += !bpf_test_sockopt(skops, sk, 0); in skops_sockopt()
|
| H A D | setget_sockopt.c | 18 int nr_active; variable 415 nr_active += !(bpf_test_sockopt(skops, sk) || in skops_sockopt()
|
| /linux/tools/testing/selftests/bpf/prog_tests/ |
| H A D | setget_sockopt.c | 64 ASSERT_EQ(bss->nr_active, 1, "nr_active"); in test_tcp() 144 ASSERT_EQ(bss->nr_active, 1, "nr_active"); in test_ktls()
|
| H A D | net_timestamping.c | 213 ASSERT_EQ(bss->nr_active, 1, "nr_active"); in test_tcp()
|
| /linux/fs/ceph/ |
| H A D | mdsmap.c | 450 int i, nr_active = 0; in ceph_mdsmap_is_cluster_available() local 459 nr_active++; in ceph_mdsmap_is_cluster_available() 461 return nr_active > 0; in ceph_mdsmap_is_cluster_available()
|
| /linux/arch/x86/events/amd/ |
| H A D | uncore.c | 43 int nr_active; member 107 if (!ctx->nr_active || ctx->cpu != smp_processor_id()) in amd_uncore_hrtimer() 169 if (!ctx->nr_active++) in amd_uncore_start() 195 if (!--ctx->nr_active) in amd_uncore_stop() 944 if (!ctx->nr_active++) in amd_uncore_umc_start()
|
| /linux/drivers/nvme/host/ |
| H A D | multipath.c | 174 atomic_inc(&ns->ctrl->nr_active); in nvme_failover_req() 193 atomic_dec_if_positive(&ns->ctrl->nr_active); in nvme_mpath_start_request() 408 depth = atomic_read(&ns->ctrl->nr_active); in nvme_round_robin_path() 1084 return sysfs_emit(buf, "%d\n", atomic_read(&ns->ctrl->nr_active)); in ana_state_show() 1347 atomic_set(&ctrl->nr_active, 0);
|
| H A D | nvme.h | 428 atomic_t nr_active; member
|
| /linux/kernel/ |
| H A D | workqueue.c | 280 * nr_active management and WORK_STRUCT_INACTIVE: 282 * When pwq->nr_active >= max_active, new work item is queued to 287 * nr_active and all work items in pwq->inactive_works are marked with 292 * in nr_active. For non-barrier work item, it is marked with 295 int nr_active; /* L: nr of active works */ 328 * As sharing a single nr_active across multiple sockets can be very expensive, 340 atomic_t nr; /* per-node nr_active */ 396 struct wq_node_nr_active *node_nr_active[]; /* I: per-node nr_active */ 1580 * - %NULL for per-cpu workqueues as they don't need to use shared nr_active. in wq_node_nr_active() 1704 return !pwq->nr_active in __pwq_activate_work() 286 int nr_active; /* L: nr of active works */ global() member [all...] |
| /linux/arch/powerpc/include/asm/ |
| H A D | spu.h | 183 int nr_active; member
|
| /linux/net/mptcp/ |
| H A D | protocol.c | 1497 int i, nr_active = 0; in mptcp_subflow_get_send() 1517 nr_active += !backup; in mptcp_subflow_get_send() 1536 if (!nr_active) 1461 int i, nr_active = 0; mptcp_subflow_get_send() local
|