| /linux/fs/squashfs/ |
| H A D | decompressor_multi_percpu.c | 31 struct squashfs_stream __percpu *percpu; in squashfs_decompressor_create() local 34 percpu = alloc_percpu(struct squashfs_stream); in squashfs_decompressor_create() 35 if (percpu == NULL) in squashfs_decompressor_create() 39 stream = per_cpu_ptr(percpu, cpu); in squashfs_decompressor_create() 49 return (void *)(__force unsigned long) percpu; in squashfs_decompressor_create() 53 stream = per_cpu_ptr(percpu, cpu); in squashfs_decompressor_create() 57 free_percpu(percpu); in squashfs_decompressor_create() 63 struct squashfs_stream __percpu *percpu = in squashfs_decompressor_destroy() local 70 stream = per_cpu_ptr(percpu, cpu); in squashfs_decompressor_destroy() 73 free_percpu(percpu); in squashfs_decompressor_destroy() [all …]
|
| /linux/kernel/bpf/ |
| H A D | memalloc.c | 257 static void free_one(void *obj, bool percpu) in free_one() argument 259 if (percpu) in free_one() 265 static int free_all(struct bpf_mem_cache *c, struct llist_node *llnode, bool percpu) in free_all() argument 273 free_one(pos, percpu); in free_all() 513 int bpf_mem_alloc_init(struct bpf_mem_alloc *ma, int size, bool percpu) in bpf_mem_alloc_init() argument 520 if (percpu && size == 0) in bpf_mem_alloc_init() 524 if (percpu) in bpf_mem_alloc_init() 526 ma->percpu = percpu; in bpf_mem_alloc_init() 533 if (!percpu) in bpf_mem_alloc_init() 591 ma->percpu = true; in bpf_mem_alloc_percpu_init() [all …]
|
| H A D | bpf_lru_list.h | 63 bool percpu; member 72 int bpf_lru_init(struct bpf_lru *lru, bool percpu, u32 hash_offset,
|
| H A D | hashtab.c | 409 bool percpu = (attr->map_type == BPF_MAP_TYPE_PERCPU_HASH || in htab_map_alloc_check() local 440 if (numa_node != NUMA_NO_NODE && (percpu || percpu_lru)) in htab_map_alloc_check() 459 if (percpu && round_up(attr->value_size, 8) > PCPU_MIN_UNIT_SIZE) in htab_map_alloc_check() 544 bool percpu = (attr->map_type == BPF_MAP_TYPE_PERCPU_HASH || in htab_map_alloc() local 585 if (percpu) in htab_map_alloc() 649 if (percpu) { in htab_map_alloc() 1075 bool percpu, bool onallcpus, in alloc_htab_elem() argument 1118 if (percpu) { in alloc_htab_elem() 1351 bool percpu, bool onallcpus) in htab_map_update_elem_in_place() argument 1387 if (percpu) { in htab_map_update_elem_in_place() [all …]
|
| H A D | arraymap.c | 55 bool percpu = attr->map_type == BPF_MAP_TYPE_PERCPU_ARRAY; in array_map_alloc_check() local 63 (percpu && numa_node != NUMA_NO_NODE)) in array_map_alloc_check() 78 if (percpu && round_up(attr->value_size, 8) > PCPU_MIN_UNIT_SIZE) in array_map_alloc_check() 86 bool percpu = attr->map_type == BPF_MAP_TYPE_PERCPU_ARRAY; in array_map_alloc() local 117 if (percpu) { in array_map_alloc() 153 if (percpu && bpf_array_alloc_percpu(array)) { in array_map_alloc() 783 bool percpu = map->map_type == BPF_MAP_TYPE_PERCPU_ARRAY; in array_map_mem_usage() local 788 if (percpu) { in array_map_mem_usage()
|
| /linux/arch/alpha/boot/ |
| H A D | main.c | 65 struct percpu_struct * percpu; in pal_init() local 95 percpu = (struct percpu_struct *) in pal_init() 97 rev = percpu->pal_revision = percpu->palcode_avail[2]; in pal_init()
|
| H A D | bootp.c | 71 struct percpu_struct * percpu; in pal_init() local 101 percpu = (struct percpu_struct *) in pal_init() 103 rev = percpu->pal_revision = percpu->palcode_avail[2]; in pal_init()
|
| H A D | bootpz.c | 119 struct percpu_struct * percpu; in pal_init() local 149 percpu = (struct percpu_struct *) in pal_init() 151 rev = percpu->pal_revision = percpu->palcode_avail[2]; in pal_init()
|
| /linux/Documentation/translations/zh_CN/core-api/ |
| H A D | workqueue.rst | 595 events percpu 0 2 4 6 596 events_highpri percpu 1 3 5 7 597 events_long percpu 0 2 4 6 599 events_freezable percpu 0 2 4 6 600 events_power_efficient percpu 0 2 4 6 601 events_freezable_power_ percpu 0 2 4 6 602 rcu_gp percpu 0 2 4 6 603 rcu_par_gp percpu 0 2 4 6 604 slub_flushwq percpu 0 2 4 6
|
| /linux/include/linux/ |
| H A D | bpf_mem_alloc.h | 15 bool percpu; member 31 int bpf_mem_alloc_init(struct bpf_mem_alloc *ma, int size, bool percpu); 43 int bpf_mem_alloc_check_size(bool percpu, size_t size);
|
| /linux/tools/testing/selftests/cgroup/ |
| H A D | test_kmem.c | 22 * Memory cgroup charging is performed using percpu batches 64 pages 349 * is greater than 0 and approximates matches the percpu value 356 long current, percpu; in test_percpu_basic() 381 percpu = cg_read_key_long(parent, "memory.stat", "percpu "); in test_percpu_basic() 383 if (current > 0 && percpu > 0 && labs(current - percpu) < in test_percpu_basic() 388 current, percpu); in test_percpu_basic() 359 long current, percpu; test_percpu_basic() local
|
| /linux/drivers/md/ |
| H A D | raid5.c | 1530 static struct page **to_addr_page(struct raid5_percpu *percpu, int i) in to_addr_page() argument 1532 return percpu->scribble + i * percpu->scribble_obj_size; in to_addr_page() 1537 struct raid5_percpu *percpu, int i) in to_addr_conv() argument 1539 return (void *) (to_addr_page(percpu, i) + sh->disks + 2); in to_addr_conv() 1546 to_addr_offs(struct stripe_head *sh, struct raid5_percpu *percpu) in to_addr_offs() argument 1548 return (unsigned int *) (to_addr_conv(sh, percpu, 0) + sh->disks + 2); in to_addr_offs() 1552 ops_run_compute5(struct stripe_head *sh, struct raid5_percpu *percpu) in ops_run_compute5() argument 1555 struct page **xor_srcs = to_addr_page(percpu, 0); in ops_run_compute5() 1556 unsigned int *off_srcs = to_addr_offs(sh, percpu); in ops_run_compute5() 1582 ops_complete_compute, sh, to_addr_conv(sh, percpu, 0)); in ops_run_compute5() [all …]
|
| /linux/drivers/clocksource/ |
| H A D | timer-qcom.c | 152 bool percpu) in msm_timer_init() argument 158 msm_timer_has_ppi = percpu; in msm_timer_init() 166 if (percpu) in msm_timer_init()
|
| /linux/arch/sparc/kernel/ |
| H A D | sun4m_irq.c | 107 bool percpu; member 200 if (handler_data->percpu) { in sun4m_mask_irq() 219 if (handler_data->percpu) { in sun4m_unmask_irq() 278 handler_data->percpu = real_irq < OBP_INT_LEVEL_ONBOARD; in sun4m_build_device_irq()
|
| /linux/Documentation/trace/coresight/ |
| H A D | coresight-trbe.rst | 13 Trace Buffer Extension (TRBE) is a percpu hardware which captures in system 14 memory, CPU traces generated from a corresponding percpu tracing unit. This
|
| /linux/include/asm-generic/ |
| H A D | vmlinux.lds.h | 1013 *(.data..percpu..decrypted) \ 1093 *(.data..percpu..page_aligned) \ 1096 *(SORT_BY_ALIGNMENT(.data..percpu..hot.*)) \ 1099 *(.data..percpu..read_mostly) \ 1101 *(.data..percpu) \ 1102 *(.data..percpu..shared_aligned) \ 1117 .data..percpu : AT(ADDR(.data..percpu) - LOAD_OFFSET) { \
|
| /linux/kernel/sched/ |
| H A D | cpuacct.c | 215 u64 percpu; in __cpuacct_percpu_seq_show() local 219 percpu = cpuacct_cpuusage_read(ca, i, index); in __cpuacct_percpu_seq_show() 220 seq_printf(m, "%llu ", (unsigned long long) percpu); in __cpuacct_percpu_seq_show()
|
| /linux/Documentation/RCU/ |
| H A D | rcuref.rst | 8 Please note that the percpu-ref feature is likely your first 10 include/linux/percpu-refcount.h for more information. However, in 11 those unusual cases where percpu-ref would consume too much memory,
|
| /linux/Documentation/translations/zh_CN/dev-tools/ |
| H A D | kmemleak.rst | 140 - ``kmemleak_alloc_percpu`` - 通知一个 percpu 类型的内存分配 144 - ``kmemleak_free_percpu`` - 通知一个 percpu 类型的内存释放
|
| /linux/arch/arm64/kvm/hyp/nvhe/ |
| H A D | hyp.lds.S | 25 BEGIN_HYP_SECTION(.data..percpu)
|
| /linux/arch/arm64/kernel/ |
| H A D | vmlinux.lds.S | 37 HYP_SECTION_NAME(.data..percpu) : { \ 38 *(HYP_SECTION_NAME(.data..percpu)) \
|
| /linux/Documentation/translations/zh_CN/locking/ |
| H A D | index.rst | 34 * percpu-rw-semaphore
|
| /linux/arch/um/include/asm/ |
| H A D | Kbuild | 18 generic-y += percpu.h
|
| /linux/mm/ |
| H A D | Makefile | 55 mm_init.o percpu.o slab_common.o \ 140 obj-$(CONFIG_PERCPU_STATS) += percpu-stats.o
|
| /linux/Documentation/translations/zh_CN/mm/ |
| H A D | vmalloced-kernel-stacks.rst | 129 - vmalloced堆栈的percpu缓存似乎比高阶堆栈分配要快一些,至少在缓存命中时是这样。
|