| /linux/mm/ |
| H A D | mempolicy.c | 624 nodemask_t *nmask; member 644 return node_isset(nid, *qp->nmask) == !(flags & MPOL_MF_INVERT); in queue_folio_required() 983 .nmask = nodes, in queue_pages_range() 1142 static long do_get_mempolicy(int *policy, nodemask_t *nmask, in do_get_mempolicy() argument 1159 *nmask = cpuset_current_mems_allowed; in do_get_mempolicy() 1224 if (nmask) { in do_get_mempolicy() 1226 *nmask = pol->w.user_nodemask; in do_get_mempolicy() 1229 get_policy_nodemask(pol, nmask); in do_get_mempolicy() 1280 nodemask_t nmask; in migrate_to_node() local 1291 nodes_clear(nmask); in migrate_to_node() [all …]
|
| H A D | memory-tiers.c | 173 nodemask_t nmask; in nodelist_show() local 176 nmask = get_memtier_nodemask(to_memory_tier(dev)); in nodelist_show() 177 ret = sysfs_emit(buf, "%*pbl\n", nodemask_pr_args(&nmask)); in nodelist_show()
|
| H A D | hugetlb.c | 1337 int nid, nodemask_t *nmask) in dequeue_hugetlb_folio_nodemask() argument 1353 for_each_zone_zonelist_nodemask(zone, z, zonelist, gfp_zone(gfp_mask), nmask) { in dequeue_hugetlb_folio_nodemask() 1873 int nid, nodemask_t *nmask, nodemask_t *node_alloc_noretry) in alloc_buddy_hugetlb_folio() argument 1890 folio = (struct folio *)__alloc_frozen_pages(gfp_mask, order, nid, nmask); in alloc_buddy_hugetlb_folio() 1918 gfp_t gfp_mask, int nid, nodemask_t *nmask, in only_alloc_fresh_hugetlb_folio() argument 1928 folio = alloc_gigantic_folio(order, gfp_mask, nid, nmask); in only_alloc_fresh_hugetlb_folio() 1930 folio = alloc_buddy_hugetlb_folio(order, gfp_mask, nid, nmask, in only_alloc_fresh_hugetlb_folio() 1945 gfp_t gfp_mask, int nid, nodemask_t *nmask) in alloc_fresh_hugetlb_folio() argument 1949 folio = only_alloc_fresh_hugetlb_folio(h, gfp_mask, nid, nmask, NULL); in alloc_fresh_hugetlb_folio() 2157 gfp_t gfp_mask, int nid, nodemask_t *nmask) in alloc_surplus_hugetlb_folio() argument [all …]
|
| H A D | memory_hotplug.c | 1837 nodemask_t nmask = node_states[N_MEMORY]; in do_migrate_range() local 1839 .nmask = &nmask, in do_migrate_range() 1856 node_clear(mtc.nid, nmask); in do_migrate_range() 1857 if (nodes_empty(nmask)) in do_migrate_range() 1858 node_set(mtc.nid, nmask); in do_migrate_range()
|
| H A D | migrate.c | 2188 mtc->nmask, gfp_mask, in alloc_migration_target() 2205 return __folio_alloc(gfp_mask, order, nid, mtc->nmask); in alloc_migration_target()
|
| H A D | internal.h | 1341 nodemask_t *nmask; member
|
| H A D | vmscan.c | 995 allowed_mask = mtc->nmask; in alloc_demote_folio() 1005 mtc->nmask = NULL; in alloc_demote_folio() 1012 mtc->nmask = allowed_mask; in alloc_demote_folio() 1037 .nmask = &allowed_mask, in demote_folio_list()
|
| /linux/drivers/clk/ |
| H A D | clk-fractional-divider.c | 74 u32 mmask, nmask; in clk_fd_get_div() local 90 nmask = GENMASK(fd->nwidth - 1, 0) << fd->nshift; in clk_fd_get_div() 93 n = (val & nmask) >> fd->nshift; in clk_fd_get_div() 188 u32 mmask, nmask; in clk_fd_set_rate() local 206 nmask = GENMASK(fd->nwidth - 1, 0) << fd->nshift; in clk_fd_set_rate() 214 val &= ~(mmask | nmask); in clk_fd_set_rate()
|
| /linux/tools/testing/selftests/kvm/include/ |
| H A D | numaif.h | 13 KVM_SYSCALL_DEFINE(get_mempolicy, 5, int *, policy, const unsigned long *, nmask, 16 KVM_SYSCALL_DEFINE(set_mempolicy, 3, int, mode, const unsigned long *, nmask,
|
| /linux/drivers/net/ethernet/microchip/vcap/ |
| H A D | vcap_api_debugfs.c | 92 u8 nvalue[16], nmask[16]; in vcap_debugfs_show_rule_keyfield() local 96 vcap_netbytes_copy(nmask, data->u128.mask, in vcap_debugfs_show_rule_keyfield() 97 sizeof(nmask)); in vcap_debugfs_show_rule_keyfield() 98 out->prf(out->dst, "%pI6/%pI6", nvalue, nmask); in vcap_debugfs_show_rule_keyfield()
|
| /linux/include/linux/ |
| H A D | hugetlb.h | 705 nodemask_t *nmask, gfp_t gfp_mask, 708 nodemask_t *nmask, gfp_t gfp_mask); 1122 nodemask_t *nmask, gfp_t gfp_mask) 1129 nodemask_t *nmask, gfp_t gfp_mask,
|
| H A D | syscalls.h | 851 const unsigned long __user *nmask, 855 unsigned long __user *nmask, 858 asmlinkage long sys_set_mempolicy(int mode, const unsigned long __user *nmask,
|
| /linux/Documentation/admin-guide/mm/ |
| H A D | numa_memory_policy.rst | 431 long set_mempolicy(int mode, const unsigned long *nmask, 436 'nmask'. 'nmask' points to a bit mask of node ids containing at least 447 const unsigned long *nmask, unsigned long maxnode, 460 const unsigned long *nmask, unsigned long maxnode, 463 mbind() installs the policy specified by (mode, nmask, maxnodes) as a
|
| /linux/arch/arm64/kvm/ |
| H A D | config.c | 1468 u64 clear = 0, set = 0, val = m->nmask; in __compute_fgt() 1471 clear |= fgu & m->nmask; in __compute_fgt() 1476 clear |= ~nested & m->nmask; in __compute_fgt()
|
| H A D | emulate-nested.c | 2168 rmasks->nmask |= BIT(tc.bit) & ~rmasks->res0; in aggregate_fgt() 2174 wmasks->nmask |= BIT(tc.bit) & ~wmasks->res0; in aggregate_fgt() 2182 unsigned long duplicate = masks->mask & masks->nmask; in check_fgt_masks() 2197 masks->res0 = ~(masks->mask | masks->nmask); in check_fgt_masks()
|
| /linux/net/sched/ |
| H A D | sch_api.c | 685 unsigned int nsize, nmask, osize; in qdisc_class_hash_grow() local 692 nmask = nsize - 1; in qdisc_class_hash_grow() 703 h = qdisc_class_hash(cl->classid, nmask); in qdisc_class_hash_grow() 709 clhash->hashmask = nmask; in qdisc_class_hash_grow()
|
| /linux/drivers/scsi/ |
| H A D | pmcraid.c | 432 u32 nmask = gmask | GLOBAL_INTERRUPT_MASK; in pmcraid_disable_interrupts() local 435 iowrite32(nmask, pinstance->int_regs.global_interrupt_mask_reg); in pmcraid_disable_interrupts() 459 u32 nmask = gmask & (~GLOBAL_INTERRUPT_MASK); in pmcraid_enable_interrupts() local 461 iowrite32(nmask, pinstance->int_regs.global_interrupt_mask_reg); in pmcraid_enable_interrupts()
|
| /linux/arch/arm64/include/asm/ |
| H A D | kvm_host.h | 639 u64 nmask; member
|