| /linux/arch/mips/kernel/ |
| H A D | mips-mt-fpaff.c | 69 cpumask_var_t cpus_allowed, new_mask, effective_mask; in mipsmt_sys_sched_setaffinity() local 74 if (len < sizeof(new_mask)) in mipsmt_sys_sched_setaffinity() 77 if (copy_from_user(&new_mask, user_mask_ptr, sizeof(new_mask))) in mipsmt_sys_sched_setaffinity() 98 if (!alloc_cpumask_var(&new_mask, GFP_KERNEL)) { in mipsmt_sys_sched_setaffinity() 116 cpumask_copy(&p->thread.user_cpus_allowed, new_mask); in mipsmt_sys_sched_setaffinity() 122 cpumask_intersects(new_mask, &mt_fpu_cpumask)) { in mipsmt_sys_sched_setaffinity() 123 cpumask_and(effective_mask, new_mask, &mt_fpu_cpumask); in mipsmt_sys_sched_setaffinity() 126 cpumask_copy(effective_mask, new_mask); in mipsmt_sys_sched_setaffinity() 128 retval = set_cpus_allowed_ptr(p, new_mask); in mipsmt_sys_sched_setaffinity() 139 cpumask_copy(new_mask, cpus_allowed); in mipsmt_sys_sched_setaffinity() [all …]
|
| /linux/fs/notify/dnotify/ |
| H A D | dnotify.c | 64 __u32 new_mask = 0; in dnotify_recalc_inode_mask() local 73 new_mask |= (dn->dn_mask & ~FS_DN_MULTISHOT); in dnotify_recalc_inode_mask() 74 if (fsn_mark->mask == new_mask) in dnotify_recalc_inode_mask() 76 fsn_mark->mask = new_mask; in dnotify_recalc_inode_mask() 202 __u32 new_mask = FS_EVENT_ON_CHILD; in convert_arg() local 205 new_mask |= FS_DN_MULTISHOT; in convert_arg() 207 new_mask |= (FS_DELETE | FS_MOVED_FROM); in convert_arg() 209 new_mask |= FS_MODIFY; in convert_arg() 211 new_mask |= FS_ACCESS; in convert_arg() 213 new_mask |= FS_ATTRIB; in convert_arg() [all …]
|
| /linux/drivers/platform/x86/intel/ |
| H A D | crystal_cove_charger.c | 30 u8 new_mask; member 57 if (charger->mask != charger->new_mask) { in crystal_cove_charger_irq_bus_sync_unlock() 58 regmap_write(charger->regmap, MCHGRIRQ_REG, charger->new_mask); in crystal_cove_charger_irq_bus_sync_unlock() 59 charger->mask = charger->new_mask; in crystal_cove_charger_irq_bus_sync_unlock() 69 charger->new_mask &= ~BIT(data->hwirq); in crystal_cove_charger_irq_unmask() 76 charger->new_mask |= BIT(data->hwirq); in crystal_cove_charger_irq_mask() 131 charger->mask = charger->new_mask = BIT(0); in crystal_cove_charger_probe()
|
| /linux/drivers/net/wireless/ath/ath5k/ |
| H A D | dma.c | 740 ath5k_hw_set_imr(struct ath5k_hw *ah, enum ath5k_int new_mask) in ath5k_hw_set_imr() argument 760 int_mask = new_mask & AR5K_INT_COMMON; in ath5k_hw_set_imr() 768 if (new_mask & AR5K_INT_FATAL) { in ath5k_hw_set_imr() 775 if (new_mask & AR5K_INT_TIM) in ath5k_hw_set_imr() 778 if (new_mask & AR5K_INT_TIM) in ath5k_hw_set_imr() 780 if (new_mask & AR5K_INT_DTIM) in ath5k_hw_set_imr() 782 if (new_mask & AR5K_INT_DTIM_SYNC) in ath5k_hw_set_imr() 784 if (new_mask & AR5K_INT_BCN_TIMEOUT) in ath5k_hw_set_imr() 786 if (new_mask & AR5K_INT_CAB_TIMEOUT) in ath5k_hw_set_imr() 790 if (new_mask & AR5K_INT_BNR) in ath5k_hw_set_imr() [all …]
|
| /linux/kernel/ |
| H A D | compat.c | 111 unsigned len, struct cpumask *new_mask) in compat_get_user_cpu_mask() argument 116 memset(new_mask, 0, cpumask_size()); in compat_get_user_cpu_mask() 120 k = cpumask_bits(new_mask); in compat_get_user_cpu_mask() 128 cpumask_var_t new_mask; in COMPAT_SYSCALL_DEFINE3() local 131 if (!alloc_cpumask_var(&new_mask, GFP_KERNEL)) in COMPAT_SYSCALL_DEFINE3() 134 retval = compat_get_user_cpu_mask(user_mask_ptr, len, new_mask); in COMPAT_SYSCALL_DEFINE3() 138 retval = sched_setaffinity(pid, new_mask); in COMPAT_SYSCALL_DEFINE3() 140 free_cpumask_var(new_mask); in COMPAT_SYSCALL_DEFINE3()
|
| /linux/kernel/sched/ |
| H A D | syscalls.c | 1139 cpumask_var_t cpus_allowed, new_mask; in __sched_setaffinity() local 1144 if (!alloc_cpumask_var(&new_mask, GFP_KERNEL)) { in __sched_setaffinity() 1150 cpumask_and(new_mask, ctx->new_mask, cpus_allowed); in __sched_setaffinity() 1152 ctx->new_mask = new_mask; in __sched_setaffinity() 1155 retval = dl_task_check_affinity(p, new_mask); in __sched_setaffinity() 1164 if (!cpumask_subset(new_mask, cpus_allowed)) { in __sched_setaffinity() 1169 cpumask_copy(new_mask, cpus_allowed); in __sched_setaffinity() 1180 bool empty = !cpumask_and(new_mask, new_mask, in __sched_setaffinity() 1184 cpumask_copy(new_mask, cpus_allowed); in __sched_setaffinity() 1191 free_cpumask_var(new_mask); in __sched_setaffinity() [all …]
|
| H A D | core.c | 2414 .new_mask = cpumask_of(rq->cpu), in migrate_disable_switch() 2432 .new_mask = &p->cpus_mask, in ___migrate_enable() 2737 p->cpus_ptr = ctx->new_mask; in set_cpus_allowed_common() 2741 cpumask_copy(&p->cpus_mask, ctx->new_mask); in set_cpus_allowed_common() 2742 p->nr_cpus_allowed = cpumask_weight(ctx->new_mask); in set_cpus_allowed_common() 2743 mm_update_cpus_allowed(p->mm, ctx->new_mask); in set_cpus_allowed_common() 2763 void set_cpus_allowed_force(struct task_struct *p, const struct cpumask *new_mask) in set_cpus_allowed_force() argument 2766 .new_mask = new_mask, in set_cpus_allowed_force() 3098 if (!kthread && !cpumask_subset(ctx->new_mask, cpu_allowed_mask)) { in __set_cpus_allowed_ptr_locked() 3113 if (cpumask_equal(&p->cpus_mask, ctx->new_mask)) { in __set_cpus_allowed_ptr_locked() [all …]
|
| /linux/fs/xfs/ |
| H A D | xfs_healthmon.c | 431 unsigned int new_mask) in metadata_event_mask() argument 435 return new_mask; in metadata_event_mask() 440 return new_mask; in metadata_event_mask() 443 return new_mask & ~old_mask; in metadata_event_mask() 446 return new_mask & old_mask; in metadata_event_mask() 461 unsigned int new_mask) in xfs_healthmon_report_fs() argument 472 event.fsmask = metadata_event_mask(hm, type, old_mask, new_mask) & in xfs_healthmon_report_fs() 474 trace_xfs_healthmon_report_fs(hm, old_mask, new_mask, &event); in xfs_healthmon_report_fs() 488 unsigned int new_mask) in xfs_healthmon_report_group() argument 503 new_mask) & in xfs_healthmon_report_group() [all …]
|
| H A D | xfs_trace.h | 6155 unsigned int old_mask, unsigned int new_mask, 6157 TP_ARGS(hm, old_mask, new_mask, event), 6163 __field(unsigned int, new_mask) 6171 __entry->new_mask = new_mask; 6179 __entry->new_mask, 6185 unsigned int old_mask, unsigned int new_mask, 6187 TP_ARGS(hm, old_mask, new_mask, event), 6193 __field(unsigned int, new_mask) 6202 __entry->new_mask = new_mask; 6211 __entry->new_mask, [all …]
|
| /linux/io_uring/ |
| H A D | register.c | 300 cpumask_var_t new_mask) in __io_register_iowq_aff() argument 305 ret = io_wq_cpu_affinity(current->io_uring, new_mask); in __io_register_iowq_aff() 308 ret = io_sqpoll_wq_cpu_affinity(ctx, new_mask); in __io_register_iowq_aff() 318 cpumask_var_t new_mask; in io_register_iowq_aff() local 321 if (!alloc_cpumask_var(&new_mask, GFP_KERNEL)) in io_register_iowq_aff() 324 cpumask_clear(new_mask); in io_register_iowq_aff() 330 ret = compat_get_bitmap(cpumask_bits(new_mask), in io_register_iowq_aff() 335 ret = copy_from_user(new_mask, arg, len); in io_register_iowq_aff() 338 free_cpumask_var(new_mask); in io_register_iowq_aff() 342 ret = __io_register_iowq_aff(ctx, new_mask); in io_register_iowq_aff() [all …]
|
| /linux/arch/mips/cavium-octeon/ |
| H A D | smp.c | 308 uint32_t mask, new_mask; in octeon_cpu_die() local 329 new_mask = labi->avail_coremask; in octeon_cpu_die() 334 new_mask = *p; in octeon_cpu_die() 337 pr_info("Reset core %d. Available Coremask = 0x%x\n", coreid, new_mask); in octeon_cpu_die()
|
| /linux/drivers/net/ethernet/sfc/ |
| H A D | ef10_sriov.c | 684 u32 old_mask, new_mask; in efx_ef10_sriov_set_privilege_mask() local 706 new_mask = old_mask & ~mask; in efx_ef10_sriov_set_privilege_mask() 707 new_mask |= value; in efx_ef10_sriov_set_privilege_mask() 709 if (new_mask == old_mask) in efx_ef10_sriov_set_privilege_mask() 712 new_mask |= MC_CMD_PRIVILEGE_MASK_IN_DO_CHANGE; in efx_ef10_sriov_set_privilege_mask() 715 MCDI_SET_DWORD(pm_inbuf, PRIVILEGE_MASK_IN_NEW_MASK, new_mask); in efx_ef10_sriov_set_privilege_mask()
|
| /linux/drivers/net/wireless/marvell/libertas/ |
| H A D | debugfs.c | 277 int value, freq, new_mask; in lbs_threshold_write() local 286 ret = sscanf(buf, "%d %d %d", &value, &freq, &new_mask); in lbs_threshold_write() 306 if (new_mask) in lbs_threshold_write() 307 new_mask = curr_mask | event_mask; in lbs_threshold_write() 309 new_mask = curr_mask & ~event_mask; in lbs_threshold_write() 316 events->events = cpu_to_le16(new_mask); in lbs_threshold_write()
|
| /linux/drivers/net/ethernet/brocade/bna/ |
| H A D | bna_hw_defs.h | 169 #define bna_intx_enable(bna, new_mask) \ argument 170 writel((new_mask), (bna)->regs.fn_int_mask)
|
| /linux/drivers/mmc/host/ |
| H A D | sdhci-brcmstb.c | 168 u32 new_mask = (mask & (SDHCI_RESET_CMD | SDHCI_RESET_DATA)) << 24; in brcmstb_sdhci_reset_cmd_data() local 177 new_mask |= SDHCI_CLOCK_CARD_EN | SDHCI_CLOCK_INT_EN; in brcmstb_sdhci_reset_cmd_data() 179 sdhci_writel(host, reg | new_mask, SDHCI_CLOCK_CONTROL); in brcmstb_sdhci_reset_cmd_data()
|
| /linux/drivers/irqchip/ |
| H A D | irq-sifive-plic.c | 431 u32 new_mask = j == i ? pending_irqs : 0; in cp100_isolate_pending_irq() local 433 if (new_mask != handler->enable_save[j]) in cp100_isolate_pending_irq() 434 writel_relaxed(new_mask, enable + j); in cp100_isolate_pending_irq()
|
| /linux/fs/notify/ |
| H A D | mark.c | 248 u32 new_mask = 0; in __fsnotify_recalc_mask() local 259 new_mask |= fsnotify_calc_mask(mark); in __fsnotify_recalc_mask() 268 WRITE_ONCE(*fsnotify_conn_mask_p(conn), new_mask); in __fsnotify_recalc_mask()
|
| /linux/drivers/gpu/drm/arm/ |
| H A D | malidp_crtc.c | 421 u32 new_mask = crtc_state->connector_mask; in malidp_crtc_atomic_check() local 423 if ((old_mask ^ new_mask) == in malidp_crtc_atomic_check()
|
| /linux/drivers/gpu/drm/amd/pm/swsmu/ |
| H A D | smu_cmn.c | 932 uint64_t new_mask) in smu_cmn_set_pp_feature_mask() argument 945 feature_2_enabled = ~feature_mask_u64 & new_mask; in smu_cmn_set_pp_feature_mask() 946 feature_2_disabled = feature_mask_u64 & ~new_mask; in smu_cmn_set_pp_feature_mask()
|
| H A D | smu_cmn.h | 165 uint64_t new_mask);
|
| /linux/include/linux/ |
| H A D | sched.h | 1892 extern void set_cpus_allowed_force(struct task_struct *p, const struct cpumask *new_mask); 1901 extern int set_cpus_allowed_ptr(struct task_struct *p, const struct cpumask *new_mask); 2335 extern long sched_setaffinity(pid_t pid, const struct cpumask *new_mask);
|
| /linux/drivers/gpu/drm/radeon/ |
| H A D | kv_dpm.c | 2206 u32 new_mask = (1 << level); in kv_set_enabled_level() local 2210 new_mask); in kv_set_enabled_level() 2216 u32 i, new_mask = 0; in kv_set_enabled_levels() local 2219 new_mask |= (1 << i); in kv_set_enabled_levels() 2223 new_mask); in kv_set_enabled_levels()
|
| /linux/arch/x86/events/intel/ |
| H A D | core.c | 5750 u64 new_mask, check_mask; in __intel_pmu_check_dyn_constr() local 5753 new_mask = c1->idxmsk64 & mask; in __intel_pmu_check_dyn_constr() 5754 new_weight = hweight64(new_mask); in __intel_pmu_check_dyn_constr() 5779 if (new_mask == check_mask || in __intel_pmu_check_dyn_constr() 5780 (new_mask ^ check_mask) == (new_mask | check_mask)) in __intel_pmu_check_dyn_constr() 5803 if ((new_mask | check_mask) != check_mask && in __intel_pmu_check_dyn_constr() 5804 fls64(new_mask) > fls64(check_mask)) in __intel_pmu_check_dyn_constr() 5807 if ((new_mask | check_mask) != new_mask && in __intel_pmu_check_dyn_constr() 5808 fls64(new_mask) < fls64(check_mask)) in __intel_pmu_check_dyn_constr()
|
| /linux/drivers/net/ethernet/faraday/ |
| H A D | ftgmac100.c | 1235 unsigned int status, new_mask = FTGMAC100_INT_BAD; in ftgmac100_interrupt() local 1270 new_mask &= ~status; in ftgmac100_interrupt() 1274 iowrite32(new_mask, priv->base + FTGMAC100_OFFSET_IER); in ftgmac100_interrupt()
|
| /linux/drivers/gpu/drm/amd/pm/legacy-dpm/ |
| H A D | kv_dpm.c | 2468 u32 new_mask = (1 << level); in kv_set_enabled_level() local 2472 new_mask); in kv_set_enabled_level() 2478 u32 i, new_mask = 0; in kv_set_enabled_levels() local 2481 new_mask |= (1 << i); in kv_set_enabled_levels() 2485 new_mask); in kv_set_enabled_levels()
|