/linux/arch/mips/kernel/ |
H A D | mips-mt-fpaff.c | 69 cpumask_var_t cpus_allowed, new_mask, effective_mask; in mipsmt_sys_sched_setaffinity() local 74 if (len < sizeof(new_mask)) in mipsmt_sys_sched_setaffinity() 77 if (copy_from_user(&new_mask, user_mask_ptr, sizeof(new_mask))) in mipsmt_sys_sched_setaffinity() 98 if (!alloc_cpumask_var(&new_mask, GFP_KERNEL)) { in mipsmt_sys_sched_setaffinity() 116 cpumask_copy(&p->thread.user_cpus_allowed, new_mask); in mipsmt_sys_sched_setaffinity() 122 cpumask_intersects(new_mask, &mt_fpu_cpumask)) { in mipsmt_sys_sched_setaffinity() 123 cpumask_and(effective_mask, new_mask, &mt_fpu_cpumask); in mipsmt_sys_sched_setaffinity() 126 cpumask_copy(effective_mask, new_mask); in mipsmt_sys_sched_setaffinity() 128 retval = set_cpus_allowed_ptr(p, new_mask); in mipsmt_sys_sched_setaffinity() 139 cpumask_copy(new_mask, cpus_allowed); in mipsmt_sys_sched_setaffinity() [all …]
|
/linux/drivers/platform/x86/intel/ |
H A D | crystal_cove_charger.c | 30 u8 new_mask; member 57 if (charger->mask != charger->new_mask) { in crystal_cove_charger_irq_bus_sync_unlock() 58 regmap_write(charger->regmap, MCHGRIRQ_REG, charger->new_mask); in crystal_cove_charger_irq_bus_sync_unlock() 59 charger->mask = charger->new_mask; in crystal_cove_charger_irq_bus_sync_unlock() 69 charger->new_mask &= ~BIT(data->hwirq); in crystal_cove_charger_irq_unmask() 76 charger->new_mask |= BIT(data->hwirq); in crystal_cove_charger_irq_mask() 131 charger->mask = charger->new_mask = BIT(0); in crystal_cove_charger_probe()
|
/linux/drivers/net/wireless/ath/ath5k/ |
H A D | dma.c | 740 ath5k_hw_set_imr(struct ath5k_hw *ah, enum ath5k_int new_mask) in ath5k_hw_set_imr() argument 760 int_mask = new_mask & AR5K_INT_COMMON; in ath5k_hw_set_imr() 768 if (new_mask & AR5K_INT_FATAL) { in ath5k_hw_set_imr() 775 if (new_mask & AR5K_INT_TIM) in ath5k_hw_set_imr() 778 if (new_mask & AR5K_INT_TIM) in ath5k_hw_set_imr() 780 if (new_mask & AR5K_INT_DTIM) in ath5k_hw_set_imr() 782 if (new_mask & AR5K_INT_DTIM_SYNC) in ath5k_hw_set_imr() 784 if (new_mask & AR5K_INT_BCN_TIMEOUT) in ath5k_hw_set_imr() 786 if (new_mask & AR5K_INT_CAB_TIMEOUT) in ath5k_hw_set_imr() 790 if (new_mask & AR5K_INT_BNR) in ath5k_hw_set_imr() [all …]
|
/linux/kernel/ |
H A D | compat.c | 111 unsigned len, struct cpumask *new_mask) in compat_get_user_cpu_mask() argument 116 memset(new_mask, 0, cpumask_size()); in compat_get_user_cpu_mask() 120 k = cpumask_bits(new_mask); in compat_get_user_cpu_mask() 128 cpumask_var_t new_mask; in COMPAT_SYSCALL_DEFINE3() local 131 if (!alloc_cpumask_var(&new_mask, GFP_KERNEL)) in COMPAT_SYSCALL_DEFINE3() 134 retval = compat_get_user_cpu_mask(user_mask_ptr, len, new_mask); in COMPAT_SYSCALL_DEFINE3() 138 retval = sched_setaffinity(pid, new_mask); in COMPAT_SYSCALL_DEFINE3() 140 free_cpumask_var(new_mask); in COMPAT_SYSCALL_DEFINE3()
|
/linux/kernel/sched/ |
H A D | syscalls.c | 1156 cpumask_var_t cpus_allowed, new_mask; in __sched_setaffinity() local 1161 if (!alloc_cpumask_var(&new_mask, GFP_KERNEL)) { in __sched_setaffinity() 1167 cpumask_and(new_mask, ctx->new_mask, cpus_allowed); in __sched_setaffinity() 1169 ctx->new_mask = new_mask; in __sched_setaffinity() 1172 retval = dl_task_check_affinity(p, new_mask); in __sched_setaffinity() 1181 if (!cpumask_subset(new_mask, cpus_allowed)) { in __sched_setaffinity() 1186 cpumask_copy(new_mask, cpus_allowed); in __sched_setaffinity() 1197 bool empty = !cpumask_and(new_mask, new_mask, in __sched_setaffinity() 1201 cpumask_copy(new_mask, cpus_allowed); in __sched_setaffinity() 1208 free_cpumask_var(new_mask); in __sched_setaffinity() [all …]
|
H A D | core.c | 2349 .new_mask = cpumask_of(rq->cpu), in migrate_disable_switch() 2390 .new_mask = &p->cpus_mask, in migrate_enable() 2697 p->cpus_ptr = ctx->new_mask; in set_cpus_allowed_common() 2701 cpumask_copy(&p->cpus_mask, ctx->new_mask); in set_cpus_allowed_common() 2702 p->nr_cpus_allowed = cpumask_weight(ctx->new_mask); in set_cpus_allowed_common() 2749 mm_set_cpus_allowed(p->mm, ctx->new_mask); in __do_set_cpus_allowed() 2761 void do_set_cpus_allowed(struct task_struct *p, const struct cpumask *new_mask) in do_set_cpus_allowed() argument 2764 .new_mask = new_mask, in do_set_cpus_allowed() 3092 if (!kthread && !cpumask_subset(ctx->new_mask, cpu_allowed_mask)) { in __set_cpus_allowed_ptr_locked() 3107 if (cpumask_equal(&p->cpus_mask, ctx->new_mask)) { in __set_cpus_allowed_ptr_locked() [all …]
|
/linux/drivers/spi/ |
H A D | spi-dw.h | 252 u32 new_mask; in dw_spi_mask_intr() local 254 new_mask = dw_readl(dws, DW_SPI_IMR) & ~mask; in dw_spi_mask_intr() 255 dw_writel(dws, DW_SPI_IMR, new_mask); in dw_spi_mask_intr() 261 u32 new_mask; in dw_spi_umask_intr() local 263 new_mask = dw_readl(dws, DW_SPI_IMR) | mask; in dw_spi_umask_intr() 264 dw_writel(dws, DW_SPI_IMR, new_mask); in dw_spi_umask_intr()
|
/linux/drivers/thermal/intel/ |
H A D | intel_powerclamp.c | 156 cpumask_var_t new_mask; in cpumask_set() local 167 ret = alloc_cpumask_var(&new_mask, GFP_KERNEL); in cpumask_set() 171 ret = bitmap_parse(arg, strlen(arg), cpumask_bits(new_mask), in cpumask_set() 176 if (cpumask_empty(new_mask) || check_invalid(new_mask, max_idle)) { in cpumask_set() 188 ret = allocate_copy_idle_injection_mask(new_mask); in cpumask_set() 191 free_cpumask_var(new_mask); in cpumask_set()
|
/linux/io_uring/ |
H A D | register.c | 204 cpumask_var_t new_mask) in __io_register_iowq_aff() argument 209 ret = io_wq_cpu_affinity(current->io_uring, new_mask); in __io_register_iowq_aff() 212 ret = io_sqpoll_wq_cpu_affinity(ctx, new_mask); in __io_register_iowq_aff() 222 cpumask_var_t new_mask; in io_register_iowq_aff() local 225 if (!alloc_cpumask_var(&new_mask, GFP_KERNEL)) in io_register_iowq_aff() 228 cpumask_clear(new_mask); in io_register_iowq_aff() 234 ret = compat_get_bitmap(cpumask_bits(new_mask), in io_register_iowq_aff() 239 ret = copy_from_user(new_mask, arg, len); in io_register_iowq_aff() 242 free_cpumask_var(new_mask); in io_register_iowq_aff() 246 ret = __io_register_iowq_aff(ctx, new_mask); in io_register_iowq_aff() [all …]
|
/linux/arch/mips/cavium-octeon/ |
H A D | smp.c | 308 uint32_t mask, new_mask; in octeon_cpu_die() local 329 new_mask = labi->avail_coremask; in octeon_cpu_die() 334 new_mask = *p; in octeon_cpu_die() 337 pr_info("Reset core %d. Available Coremask = 0x%x \n", coreid, new_mask); in octeon_cpu_die()
|
/linux/drivers/infiniband/core/ |
H A D | counters.c | 15 enum rdma_nl_counter_mask new_mask) in __counter_set_mode() argument 18 if (new_mask & (~ALL_AUTO_MODE_MASKS)) in __counter_set_mode() 25 port_counter->mode.mask = new_mask; in __counter_set_mode() 80 enum rdma_nl_counter_mask new_mask) in auto_mode_init_counter() argument 85 counter->mode.mask = new_mask; in auto_mode_init_counter() 87 if (new_mask & RDMA_COUNTER_MASK_QP_TYPE) in auto_mode_init_counter()
|
/linux/kernel/trace/ |
H A D | trace_boot.c | 70 cpumask_var_t new_mask; in trace_boot_set_instance_options() local 72 if (alloc_cpumask_var(&new_mask, GFP_KERNEL)) { in trace_boot_set_instance_options() 73 if (cpumask_parse(p, new_mask) < 0 || in trace_boot_set_instance_options() 74 tracing_set_cpumask(tr, new_mask) < 0) in trace_boot_set_instance_options() 76 free_cpumask_var(new_mask); in trace_boot_set_instance_options()
|
/linux/drivers/net/ethernet/sfc/ |
H A D | ef10_sriov.c | 685 u32 old_mask, new_mask; in efx_ef10_sriov_set_privilege_mask() local 707 new_mask = old_mask & ~mask; in efx_ef10_sriov_set_privilege_mask() 708 new_mask |= value; in efx_ef10_sriov_set_privilege_mask() 710 if (new_mask == old_mask) in efx_ef10_sriov_set_privilege_mask() 713 new_mask |= MC_CMD_PRIVILEGE_MASK_IN_DO_CHANGE; in efx_ef10_sriov_set_privilege_mask() 716 MCDI_SET_DWORD(pm_inbuf, PRIVILEGE_MASK_IN_NEW_MASK, new_mask); in efx_ef10_sriov_set_privilege_mask()
|
/linux/drivers/net/wireless/marvell/libertas/ |
H A D | debugfs.c | 277 int value, freq, new_mask; in lbs_threshold_write() local 286 ret = sscanf(buf, "%d %d %d", &value, &freq, &new_mask); in lbs_threshold_write() 306 if (new_mask) in lbs_threshold_write() 307 new_mask = curr_mask | event_mask; in lbs_threshold_write() 309 new_mask = curr_mask & ~event_mask; in lbs_threshold_write() 316 events->events = cpu_to_le16(new_mask); in lbs_threshold_write()
|
/linux/drivers/gpu/drm/imagination/ |
H A D | pvr_fw_trace.c | 448 pvr_fw_trace_mask_update(struct pvr_device *pvr_dev, u32 old_mask, u32 new_mask) in pvr_fw_trace_mask_update() argument 450 if (old_mask != new_mask) in pvr_fw_trace_mask_update() 451 update_logtype(pvr_dev, new_mask); in pvr_fw_trace_mask_update()
|
H A D | pvr_fw_trace.h | 73 u32 new_mask);
|
/linux/include/linux/ |
H A D | sched.h | 1826 extern void do_set_cpus_allowed(struct task_struct *p, const struct cpumask *new_mask); 1835 extern int set_cpus_allowed_ptr(struct task_struct *p, const struct cpumask *new_mask); 1842 static inline void do_set_cpus_allowed(struct task_struct *p, const struct cpumask *new_mask) in do_set_cpus_allowed() argument 1845 static inline int set_cpus_allowed_ptr(struct task_struct *p, const struct cpumask *new_mask) in set_cpus_allowed_ptr() argument 1848 if ((*cpumask_bits(new_mask) & 1) == 0) in set_cpus_allowed_ptr() 2203 extern long sched_setaffinity(pid_t pid, const struct cpumask *new_mask);
|
/linux/drivers/gpu/drm/display/ |
H A D | drm_dp_tunnel.c | 1635 unsigned long old_mask, unsigned long new_mask) in resize_bw_array() argument 1637 unsigned long move_mask = old_mask & new_mask; in resize_bw_array() 1641 WARN_ON(!new_mask); in resize_bw_array() 1643 if (old_mask == new_mask) in resize_bw_array() 1646 new_bws = kcalloc(hweight32(new_mask), sizeof(*new_bws), GFP_KERNEL); in resize_bw_array() 1651 new_bws[stream_id_to_idx(new_mask, id)] = in resize_bw_array() 1656 tunnel_state->stream_mask = new_mask; in resize_bw_array()
|
/linux/drivers/net/fddi/ |
H A D | defza.c | 290 unsigned int old_mask, new_mask; in fza_cmd_send() local 296 new_mask = old_mask & ~FZA_MASK_STATE_CHG; in fza_cmd_send() 297 writew_u(new_mask, &fp->regs->int_mask); in fza_cmd_send() 299 fp->int_mask = new_mask; in fza_cmd_send() 1081 unsigned int old_mask, new_mask; in fza_start_xmit() local 1123 new_mask = old_mask & ~FZA_MASK_SMT_TX_POLL; in fza_start_xmit() 1124 writew_u(new_mask, &fp->regs->int_mask); in fza_start_xmit() 1126 fp->int_mask = new_mask; in fza_start_xmit()
|
/linux/drivers/net/ethernet/brocade/bna/ |
H A D | bna_hw_defs.h | 169 #define bna_intx_enable(bna, new_mask) \ argument 170 writel((new_mask), (bna)->regs.fn_int_mask)
|
/linux/drivers/gpu/drm/amd/pm/swsmu/ |
H A D | smu_cmn.h | 103 uint64_t new_mask);
|
H A D | smu_cmn.c | 840 uint64_t new_mask) in smu_cmn_set_pp_feature_mask() argument 851 feature_2_enabled = ~feature_mask & new_mask; in smu_cmn_set_pp_feature_mask() 852 feature_2_disabled = feature_mask & ~new_mask; in smu_cmn_set_pp_feature_mask()
|
/linux/drivers/gpu/drm/arm/ |
H A D | malidp_crtc.c | 425 u32 new_mask = crtc_state->connector_mask; in malidp_crtc_atomic_check() local 427 if ((old_mask ^ new_mask) == in malidp_crtc_atomic_check()
|
/linux/drivers/ptp/ |
H A D | ptp_idt82p33.c | 378 u8 new_mask = 0; in _idt82p33_gettime() local 383 new_mask = idt82p33_extts_enable_mask(channel, old_mask, false); in _idt82p33_gettime() 401 if (new_mask) in _idt82p33_gettime() 402 idt82p33_extts_enable_mask(channel, new_mask, true); in _idt82p33_gettime()
|
/linux/drivers/gpu/drm/i915/display/ |
H A D | skl_watermark.c | 257 u16 old_mask, new_mask; in icl_sagv_pre_plane_update() local 263 new_mask = old_bw_state->qgv_points_mask | new_bw_state->qgv_points_mask; in icl_sagv_pre_plane_update() 265 if (old_mask == new_mask) in icl_sagv_pre_plane_update() 271 old_mask, new_mask); in icl_sagv_pre_plane_update() 279 icl_pcode_restrict_qgv_points(i915, new_mask); in icl_sagv_pre_plane_update() 289 u16 old_mask, new_mask; in icl_sagv_post_plane_update() local 295 new_mask = new_bw_state->qgv_points_mask; in icl_sagv_post_plane_update() 297 if (old_mask == new_mask) in icl_sagv_post_plane_update() 303 old_mask, new_mask); in icl_sagv_post_plane_update() 311 icl_pcode_restrict_qgv_points(i915, new_mask); in icl_sagv_post_plane_update()
|