Home
last modified time | relevance | path

Searched refs:new_mask (Results 1 – 25 of 32) sorted by relevance

12

/linux/arch/mips/kernel/
H A Dmips-mt-fpaff.c69 cpumask_var_t cpus_allowed, new_mask, effective_mask; in mipsmt_sys_sched_setaffinity() local
74 if (len < sizeof(new_mask)) in mipsmt_sys_sched_setaffinity()
77 if (copy_from_user(&new_mask, user_mask_ptr, sizeof(new_mask))) in mipsmt_sys_sched_setaffinity()
98 if (!alloc_cpumask_var(&new_mask, GFP_KERNEL)) { in mipsmt_sys_sched_setaffinity()
116 cpumask_copy(&p->thread.user_cpus_allowed, new_mask); in mipsmt_sys_sched_setaffinity()
122 cpumask_intersects(new_mask, &mt_fpu_cpumask)) { in mipsmt_sys_sched_setaffinity()
123 cpumask_and(effective_mask, new_mask, &mt_fpu_cpumask); in mipsmt_sys_sched_setaffinity()
126 cpumask_copy(effective_mask, new_mask); in mipsmt_sys_sched_setaffinity()
128 retval = set_cpus_allowed_ptr(p, new_mask); in mipsmt_sys_sched_setaffinity()
139 cpumask_copy(new_mask, cpus_allowed); in mipsmt_sys_sched_setaffinity()
[all …]
/linux/fs/notify/dnotify/
H A Ddnotify.c64 __u32 new_mask = 0; in dnotify_recalc_inode_mask() local
73 new_mask |= (dn->dn_mask & ~FS_DN_MULTISHOT); in dnotify_recalc_inode_mask()
74 if (fsn_mark->mask == new_mask) in dnotify_recalc_inode_mask()
76 fsn_mark->mask = new_mask; in dnotify_recalc_inode_mask()
202 __u32 new_mask = FS_EVENT_ON_CHILD; in convert_arg() local
205 new_mask |= FS_DN_MULTISHOT; in convert_arg()
207 new_mask |= (FS_DELETE | FS_MOVED_FROM); in convert_arg()
209 new_mask |= FS_MODIFY; in convert_arg()
211 new_mask |= FS_ACCESS; in convert_arg()
213 new_mask |= FS_ATTRIB; in convert_arg()
[all …]
/linux/drivers/platform/x86/intel/
H A Dcrystal_cove_charger.c30 u8 new_mask; member
57 if (charger->mask != charger->new_mask) { in crystal_cove_charger_irq_bus_sync_unlock()
58 regmap_write(charger->regmap, MCHGRIRQ_REG, charger->new_mask); in crystal_cove_charger_irq_bus_sync_unlock()
59 charger->mask = charger->new_mask; in crystal_cove_charger_irq_bus_sync_unlock()
69 charger->new_mask &= ~BIT(data->hwirq); in crystal_cove_charger_irq_unmask()
76 charger->new_mask |= BIT(data->hwirq); in crystal_cove_charger_irq_mask()
131 charger->mask = charger->new_mask = BIT(0); in crystal_cove_charger_probe()
/linux/drivers/net/wireless/ath/ath5k/
H A Ddma.c740 ath5k_hw_set_imr(struct ath5k_hw *ah, enum ath5k_int new_mask) in ath5k_hw_set_imr() argument
760 int_mask = new_mask & AR5K_INT_COMMON; in ath5k_hw_set_imr()
768 if (new_mask & AR5K_INT_FATAL) { in ath5k_hw_set_imr()
775 if (new_mask & AR5K_INT_TIM) in ath5k_hw_set_imr()
778 if (new_mask & AR5K_INT_TIM) in ath5k_hw_set_imr()
780 if (new_mask & AR5K_INT_DTIM) in ath5k_hw_set_imr()
782 if (new_mask & AR5K_INT_DTIM_SYNC) in ath5k_hw_set_imr()
784 if (new_mask & AR5K_INT_BCN_TIMEOUT) in ath5k_hw_set_imr()
786 if (new_mask & AR5K_INT_CAB_TIMEOUT) in ath5k_hw_set_imr()
790 if (new_mask & AR5K_INT_BNR) in ath5k_hw_set_imr()
[all …]
/linux/kernel/
H A Dcompat.c111 unsigned len, struct cpumask *new_mask) in compat_get_user_cpu_mask() argument
116 memset(new_mask, 0, cpumask_size()); in compat_get_user_cpu_mask()
120 k = cpumask_bits(new_mask); in compat_get_user_cpu_mask()
128 cpumask_var_t new_mask; in COMPAT_SYSCALL_DEFINE3() local
131 if (!alloc_cpumask_var(&new_mask, GFP_KERNEL)) in COMPAT_SYSCALL_DEFINE3()
134 retval = compat_get_user_cpu_mask(user_mask_ptr, len, new_mask); in COMPAT_SYSCALL_DEFINE3()
138 retval = sched_setaffinity(pid, new_mask); in COMPAT_SYSCALL_DEFINE3()
140 free_cpumask_var(new_mask); in COMPAT_SYSCALL_DEFINE3()
/linux/kernel/sched/
H A Dsyscalls.c1139 cpumask_var_t cpus_allowed, new_mask; in __sched_setaffinity() local
1144 if (!alloc_cpumask_var(&new_mask, GFP_KERNEL)) { in __sched_setaffinity()
1150 cpumask_and(new_mask, ctx->new_mask, cpus_allowed); in __sched_setaffinity()
1152 ctx->new_mask = new_mask; in __sched_setaffinity()
1155 retval = dl_task_check_affinity(p, new_mask); in __sched_setaffinity()
1164 if (!cpumask_subset(new_mask, cpus_allowed)) { in __sched_setaffinity()
1169 cpumask_copy(new_mask, cpus_allowed); in __sched_setaffinity()
1180 bool empty = !cpumask_and(new_mask, new_mask, in __sched_setaffinity()
1184 cpumask_copy(new_mask, cpus_allowed); in __sched_setaffinity()
1191 free_cpumask_var(new_mask); in __sched_setaffinity()
[all …]
H A Dcore.c2414 .new_mask = cpumask_of(rq->cpu), in migrate_disable_switch()
2432 .new_mask = &p->cpus_mask, in ___migrate_enable()
2737 p->cpus_ptr = ctx->new_mask; in set_cpus_allowed_common()
2741 cpumask_copy(&p->cpus_mask, ctx->new_mask); in set_cpus_allowed_common()
2742 p->nr_cpus_allowed = cpumask_weight(ctx->new_mask); in set_cpus_allowed_common()
2743 mm_update_cpus_allowed(p->mm, ctx->new_mask); in set_cpus_allowed_common()
2763 void set_cpus_allowed_force(struct task_struct *p, const struct cpumask *new_mask) in set_cpus_allowed_force() argument
2766 .new_mask = new_mask, in set_cpus_allowed_force()
3098 if (!kthread && !cpumask_subset(ctx->new_mask, cpu_allowed_mask)) { in __set_cpus_allowed_ptr_locked()
3113 if (cpumask_equal(&p->cpus_mask, ctx->new_mask)) { in __set_cpus_allowed_ptr_locked()
[all …]
/linux/fs/xfs/
H A Dxfs_healthmon.c431 unsigned int new_mask) in metadata_event_mask() argument
435 return new_mask; in metadata_event_mask()
440 return new_mask; in metadata_event_mask()
443 return new_mask & ~old_mask; in metadata_event_mask()
446 return new_mask & old_mask; in metadata_event_mask()
461 unsigned int new_mask) in xfs_healthmon_report_fs() argument
472 event.fsmask = metadata_event_mask(hm, type, old_mask, new_mask) & in xfs_healthmon_report_fs()
474 trace_xfs_healthmon_report_fs(hm, old_mask, new_mask, &event); in xfs_healthmon_report_fs()
488 unsigned int new_mask) in xfs_healthmon_report_group() argument
503 new_mask) & in xfs_healthmon_report_group()
[all …]
H A Dxfs_trace.h6155 unsigned int old_mask, unsigned int new_mask,
6157 TP_ARGS(hm, old_mask, new_mask, event),
6163 __field(unsigned int, new_mask)
6171 __entry->new_mask = new_mask;
6179 __entry->new_mask,
6185 unsigned int old_mask, unsigned int new_mask,
6187 TP_ARGS(hm, old_mask, new_mask, event),
6193 __field(unsigned int, new_mask)
6202 __entry->new_mask = new_mask;
6211 __entry->new_mask,
[all …]
/linux/io_uring/
H A Dregister.c300 cpumask_var_t new_mask) in __io_register_iowq_aff() argument
305 ret = io_wq_cpu_affinity(current->io_uring, new_mask); in __io_register_iowq_aff()
308 ret = io_sqpoll_wq_cpu_affinity(ctx, new_mask); in __io_register_iowq_aff()
318 cpumask_var_t new_mask; in io_register_iowq_aff() local
321 if (!alloc_cpumask_var(&new_mask, GFP_KERNEL)) in io_register_iowq_aff()
324 cpumask_clear(new_mask); in io_register_iowq_aff()
330 ret = compat_get_bitmap(cpumask_bits(new_mask), in io_register_iowq_aff()
335 ret = copy_from_user(new_mask, arg, len); in io_register_iowq_aff()
338 free_cpumask_var(new_mask); in io_register_iowq_aff()
342 ret = __io_register_iowq_aff(ctx, new_mask); in io_register_iowq_aff()
[all …]
/linux/arch/mips/cavium-octeon/
H A Dsmp.c308 uint32_t mask, new_mask; in octeon_cpu_die() local
329 new_mask = labi->avail_coremask; in octeon_cpu_die()
334 new_mask = *p; in octeon_cpu_die()
337 pr_info("Reset core %d. Available Coremask = 0x%x\n", coreid, new_mask); in octeon_cpu_die()
/linux/drivers/net/ethernet/sfc/
H A Def10_sriov.c684 u32 old_mask, new_mask; in efx_ef10_sriov_set_privilege_mask() local
706 new_mask = old_mask & ~mask; in efx_ef10_sriov_set_privilege_mask()
707 new_mask |= value; in efx_ef10_sriov_set_privilege_mask()
709 if (new_mask == old_mask) in efx_ef10_sriov_set_privilege_mask()
712 new_mask |= MC_CMD_PRIVILEGE_MASK_IN_DO_CHANGE; in efx_ef10_sriov_set_privilege_mask()
715 MCDI_SET_DWORD(pm_inbuf, PRIVILEGE_MASK_IN_NEW_MASK, new_mask); in efx_ef10_sriov_set_privilege_mask()
/linux/drivers/net/wireless/marvell/libertas/
H A Ddebugfs.c277 int value, freq, new_mask; in lbs_threshold_write() local
286 ret = sscanf(buf, "%d %d %d", &value, &freq, &new_mask); in lbs_threshold_write()
306 if (new_mask) in lbs_threshold_write()
307 new_mask = curr_mask | event_mask; in lbs_threshold_write()
309 new_mask = curr_mask & ~event_mask; in lbs_threshold_write()
316 events->events = cpu_to_le16(new_mask); in lbs_threshold_write()
/linux/drivers/net/ethernet/brocade/bna/
H A Dbna_hw_defs.h169 #define bna_intx_enable(bna, new_mask) \ argument
170 writel((new_mask), (bna)->regs.fn_int_mask)
/linux/drivers/mmc/host/
H A Dsdhci-brcmstb.c168 u32 new_mask = (mask & (SDHCI_RESET_CMD | SDHCI_RESET_DATA)) << 24; in brcmstb_sdhci_reset_cmd_data() local
177 new_mask |= SDHCI_CLOCK_CARD_EN | SDHCI_CLOCK_INT_EN; in brcmstb_sdhci_reset_cmd_data()
179 sdhci_writel(host, reg | new_mask, SDHCI_CLOCK_CONTROL); in brcmstb_sdhci_reset_cmd_data()
/linux/drivers/irqchip/
H A Dirq-sifive-plic.c431 u32 new_mask = j == i ? pending_irqs : 0; in cp100_isolate_pending_irq() local
433 if (new_mask != handler->enable_save[j]) in cp100_isolate_pending_irq()
434 writel_relaxed(new_mask, enable + j); in cp100_isolate_pending_irq()
/linux/fs/notify/
H A Dmark.c248 u32 new_mask = 0; in __fsnotify_recalc_mask() local
259 new_mask |= fsnotify_calc_mask(mark); in __fsnotify_recalc_mask()
268 WRITE_ONCE(*fsnotify_conn_mask_p(conn), new_mask); in __fsnotify_recalc_mask()
/linux/drivers/gpu/drm/arm/
H A Dmalidp_crtc.c421 u32 new_mask = crtc_state->connector_mask; in malidp_crtc_atomic_check() local
423 if ((old_mask ^ new_mask) == in malidp_crtc_atomic_check()
/linux/drivers/gpu/drm/amd/pm/swsmu/
H A Dsmu_cmn.c932 uint64_t new_mask) in smu_cmn_set_pp_feature_mask() argument
945 feature_2_enabled = ~feature_mask_u64 & new_mask; in smu_cmn_set_pp_feature_mask()
946 feature_2_disabled = feature_mask_u64 & ~new_mask; in smu_cmn_set_pp_feature_mask()
H A Dsmu_cmn.h165 uint64_t new_mask);
/linux/include/linux/
H A Dsched.h1892 extern void set_cpus_allowed_force(struct task_struct *p, const struct cpumask *new_mask);
1901 extern int set_cpus_allowed_ptr(struct task_struct *p, const struct cpumask *new_mask);
2335 extern long sched_setaffinity(pid_t pid, const struct cpumask *new_mask);
/linux/drivers/gpu/drm/radeon/
H A Dkv_dpm.c2206 u32 new_mask = (1 << level); in kv_set_enabled_level() local
2210 new_mask); in kv_set_enabled_level()
2216 u32 i, new_mask = 0; in kv_set_enabled_levels() local
2219 new_mask |= (1 << i); in kv_set_enabled_levels()
2223 new_mask); in kv_set_enabled_levels()
/linux/arch/x86/events/intel/
H A Dcore.c5750 u64 new_mask, check_mask; in __intel_pmu_check_dyn_constr() local
5753 new_mask = c1->idxmsk64 & mask; in __intel_pmu_check_dyn_constr()
5754 new_weight = hweight64(new_mask); in __intel_pmu_check_dyn_constr()
5779 if (new_mask == check_mask || in __intel_pmu_check_dyn_constr()
5780 (new_mask ^ check_mask) == (new_mask | check_mask)) in __intel_pmu_check_dyn_constr()
5803 if ((new_mask | check_mask) != check_mask && in __intel_pmu_check_dyn_constr()
5804 fls64(new_mask) > fls64(check_mask)) in __intel_pmu_check_dyn_constr()
5807 if ((new_mask | check_mask) != new_mask && in __intel_pmu_check_dyn_constr()
5808 fls64(new_mask) < fls64(check_mask)) in __intel_pmu_check_dyn_constr()
/linux/drivers/net/ethernet/faraday/
H A Dftgmac100.c1235 unsigned int status, new_mask = FTGMAC100_INT_BAD; in ftgmac100_interrupt() local
1270 new_mask &= ~status; in ftgmac100_interrupt()
1274 iowrite32(new_mask, priv->base + FTGMAC100_OFFSET_IER); in ftgmac100_interrupt()
/linux/drivers/gpu/drm/amd/pm/legacy-dpm/
H A Dkv_dpm.c2468 u32 new_mask = (1 << level); in kv_set_enabled_level() local
2472 new_mask); in kv_set_enabled_level()
2478 u32 i, new_mask = 0; in kv_set_enabled_levels() local
2481 new_mask |= (1 << i); in kv_set_enabled_levels()
2485 new_mask); in kv_set_enabled_levels()

12