| /linux/kernel/time/ |
| H A D | tick-broadcast.c | 1233 zalloc_cpumask_var(&tick_broadcast_mask, GFP_NOWAIT); in tick_broadcast_init() 1234 zalloc_cpumask_var(&tick_broadcast_on, GFP_NOWAIT); in tick_broadcast_init() 1235 zalloc_cpumask_var(&tmpmask, GFP_NOWAIT); in tick_broadcast_init() 1237 zalloc_cpumask_var(&tick_broadcast_oneshot_mask, GFP_NOWAIT); in tick_broadcast_init() 1238 zalloc_cpumask_var(&tick_broadcast_pending_mask, GFP_NOWAIT); in tick_broadcast_init() 1239 zalloc_cpumask_var(&tick_broadcast_force_mask, GFP_NOWAIT); in tick_broadcast_init()
|
| /linux/lib/ |
| H A D | group_cpus.c | 55 if (!zalloc_cpumask_var(&masks[node], GFP_KERNEL)) in alloc_node_to_cpumask() 373 if (!zalloc_cpumask_var(&nmsk, GFP_KERNEL)) in __try_group_cluster_cpus() 499 if (!zalloc_cpumask_var(&nmsk, GFP_KERNEL)) in group_cpus_evenly() 502 if (!zalloc_cpumask_var(&npresmsk, GFP_KERNEL)) in group_cpus_evenly()
|
| H A D | cpu_rmap.c | 173 if (unlikely(!zalloc_cpumask_var(&update_mask, GFP_KERNEL))) in cpu_rmap_update()
|
| /linux/drivers/infiniband/hw/hfi1/ |
| H A D | affinity.c | 401 if (!zalloc_cpumask_var(&non_intr_cpus, GFP_KERNEL)) in _dev_comp_vect_mappings_create() 404 if (!zalloc_cpumask_var(&available_cpus, GFP_KERNEL)) { in _dev_comp_vect_mappings_create() 873 if (!zalloc_cpumask_var(&diff, GFP_KERNEL)) in get_irq_affinity() 1029 ret = zalloc_cpumask_var(&diff, GFP_KERNEL); in hfi1_get_proc_affinity() 1032 ret = zalloc_cpumask_var(&hw_thread_mask, GFP_KERNEL); in hfi1_get_proc_affinity() 1035 ret = zalloc_cpumask_var(&available_mask, GFP_KERNEL); in hfi1_get_proc_affinity() 1038 ret = zalloc_cpumask_var(&intrs_mask, GFP_KERNEL); in hfi1_get_proc_affinity()
|
| H A D | netdev_rx.c | 171 if (!zalloc_cpumask_var(&node_cpu_mask, GFP_KERNEL)) { in hfi1_num_netdev_contexts()
|
| /linux/drivers/md/ |
| H A D | dm-ps-io-affinity.c | 66 if (!zalloc_cpumask_var(&pi->cpumask, GFP_KERNEL)) { in ioa_add_path() 123 if (!zalloc_cpumask_var(&s->path_mask, GFP_KERNEL)) in ioa_create()
|
| /linux/rust/helpers/ |
| H A D | cpumask.c | 74 return zalloc_cpumask_var(mask, flags);
|
| /linux/kernel/sched/ |
| H A D | cpupri.c | 287 if (!zalloc_cpumask_var(&vec->mask, GFP_KERNEL)) in cpupri_init()
|
| H A D | topology.c | 537 if (!zalloc_cpumask_var(&rd->span, GFP_KERNEL)) in init_rootdomain() 539 if (!zalloc_cpumask_var(&rd->online, GFP_KERNEL)) in init_rootdomain() 541 if (!zalloc_cpumask_var(&rd->dlo_mask, GFP_KERNEL)) in init_rootdomain() 543 if (!zalloc_cpumask_var(&rd->rto_mask, GFP_KERNEL)) in init_rootdomain() 2765 zalloc_cpumask_var(&sched_domains_tmpmask, GFP_KERNEL); in sched_init_domains() 2766 zalloc_cpumask_var(&sched_domains_tmpmask2, GFP_KERNEL); in sched_init_domains() 2767 zalloc_cpumask_var(&fallback_doms, GFP_KERNEL); in sched_init_domains()
|
| H A D | cpudeadline.c | 259 if (!zalloc_cpumask_var(&cp->free_cpus, GFP_KERNEL)) { in cpudl_init()
|
| /linux/drivers/virt/nitro_enclaves/ |
| H A D | ne_misc_dev.c | 189 if (!zalloc_cpumask_var(&cpu_pool, GFP_KERNEL)) in ne_setup_cpu_pool() 306 if (!zalloc_cpumask_var(&ne_cpu_pool.avail_threads_per_core[i], GFP_KERNEL)) { in ne_setup_cpu_pool() 1638 if (!zalloc_cpumask_var(&ne_enclave->threads_per_core[i], GFP_KERNEL)) { in ne_create_vm_ioctl() 1644 if (!zalloc_cpumask_var(&ne_enclave->vcpu_ids, GFP_KERNEL)) { in ne_create_vm_ioctl()
|
| /linux/arch/x86/kernel/cpu/resctrl/ |
| H A D | rdtgroup.c | 150 if (!zalloc_cpumask_var(&cpu_mask, GFP_KERNEL)) in set_cache_qos_cfg()
|
| /linux/kernel/ |
| H A D | kthread.c | 358 if (!zalloc_cpumask_var(&affinity, GFP_KERNEL)) { in kthread_affine_node() 852 if (!zalloc_cpumask_var(&affinity, GFP_KERNEL)) in kthread_affine_preferred() 889 if (!zalloc_cpumask_var(&affinity, GFP_KERNEL)) in kthreads_update_affinity()
|
| H A D | compat.c | 155 if (!zalloc_cpumask_var(&mask, GFP_KERNEL)) in COMPAT_SYSCALL_DEFINE3()
|
| /linux/drivers/cpufreq/ |
| H A D | cpufreq-dt.c | 176 if (!zalloc_cpumask_var(&priv->cpus, GFP_KERNEL)) in dt_cpufreq_early_init()
|
| H A D | e_powersaver.c | 62 if (!zalloc_cpumask_var(&eps_acpi_cpu_perf->shared_cpu_map, in eps_acpi_init()
|
| H A D | scmi-cpufreq.c | 221 if (!zalloc_cpumask_var(&priv->opp_shared_cpus, GFP_KERNEL)) { in scmi_cpufreq_init()
|
| H A D | powernow-k7.c | 313 if (!zalloc_cpumask_var(&acpi_processor_perf->shared_cpu_map, in powernow_acpi_init()
|
| /linux/arch/arm64/kernel/ |
| H A D | topology.c | 267 !zalloc_cpumask_var(&amu_fie_cpus, GFP_KERNEL)) { in amu_fie_setup()
|
| /linux/drivers/soc/fsl/dpio/ |
| H A D | dpio-driver.c | 324 if (!zalloc_cpumask_var(&cpus_unused_mask, GFP_KERNEL)) in dpio_driver_init()
|
| /linux/include/linux/ |
| H A D | cpumask.h | 1056 bool zalloc_cpumask_var(cpumask_var_t *mask, gfp_t flags) in zalloc_cpumask_var() function 1087 static __always_inline bool zalloc_cpumask_var(cpumask_var_t *mask, gfp_t flags) in zalloc_cpumask_var() function
|
| /linux/drivers/thermal/intel/ |
| H A D | intel_hfi.c | 699 if (!zalloc_cpumask_var(&hfi_instance->cpus, GFP_KERNEL)) in intel_hfi_init()
|
| /linux/kernel/cgroup/ |
| H A D | cpuset.c | 519 if (!zalloc_cpumask_var(pmasks[i], GFP_KERNEL)) { in alloc_cpumasks() 3669 BUG_ON(!zalloc_cpumask_var(&subpartitions_cpus, GFP_KERNEL)); in cpuset_init() 3670 BUG_ON(!zalloc_cpumask_var(&isolated_cpus, GFP_KERNEL)); in cpuset_init() 3671 BUG_ON(!zalloc_cpumask_var(&isolated_hk_cpus, GFP_KERNEL)); in cpuset_init()
|
| /linux/drivers/target/iscsi/ |
| H A D | iscsi_target_login.c | 1049 if (!zalloc_cpumask_var(&conn->conn_cpumask, GFP_KERNEL)) { in iscsit_alloc_conn() 1054 if (!zalloc_cpumask_var(&conn->allowed_cpumask, GFP_KERNEL)) { in iscsit_alloc_conn()
|
| /linux/fs/resctrl/ |
| H A D | rdtgroup.c | 521 if (!zalloc_cpumask_var(&tmpmask, GFP_KERNEL)) in rdtgroup_cpus_write() 523 if (!zalloc_cpumask_var(&newmask, GFP_KERNEL)) { in rdtgroup_cpus_write() 527 if (!zalloc_cpumask_var(&tmpmask1, GFP_KERNEL)) { in rdtgroup_cpus_write() 4059 if (!zalloc_cpumask_var(&tmpmask, GFP_KERNEL)) in rdtgroup_rmdir() 4202 if (!zalloc_cpumask_var(&tmpmask, GFP_KERNEL)) { in rdtgroup_rename()
|