| /linux/Documentation/translations/zh_CN/filesystems/ |
| H A D | tmpfs.rst | 76 mpol=default 采用进程分配策略 78 mpol=prefer:Node 倾向从给定的节点分配 79 mpol=bind:NodeList 只允许从指定的链表分配 80 mpol=interleave 倾向于依次从每个节点分配 81 mpol=interleave:NodeList 依次从每个节点分配 82 mpol=local 优先本地节点分配内存 86 分隔符的十进制数来表示。例如,mpol=bind0-3,5,7,9-15 104 例如,mpol=bind=staticNodeList相当于MPOL_BIND|MPOL_F_STATIC_NODES的分配策略 106 请注意,如果内核不支持NUMA,那么使用mpol选项挂载tmpfs将会失败;nodelist指定不 108 revocery内核),或者具有较少的节点在线,建议从自动模式中省略mpol选项挂载选项。 [all …]
|
| /linux/Documentation/translations/zh_TW/filesystems/ |
| H A D | tmpfs.rst | 76 mpol=default 採用進程分配策略 78 mpol=prefer:Node 傾向從給定的節點分配 79 mpol=bind:NodeList 只允許從指定的鏈表分配 80 mpol=interleave 傾向於依次從每個節點分配 81 mpol=interleave:NodeList 依次從每個節點分配 82 mpol=local 優先本地節點分配內存 86 分隔符的十進制數來表示。例如,mpol=bind0-3,5,7,9-15 104 例如,mpol=bind=staticNodeList相當於MPOL_BIND|MPOL_F_STATIC_NODES的分配策略 106 請注意,如果內核不支持NUMA,那麼使用mpol選項掛載tmpfs將會失敗;nodelist指定不 108 revocery內核),或者具有較少的節點在線,建議從自動模式中省略mpol選項掛載選項。 [all …]
|
| /linux/include/linux/ |
| H A D | mempolicy.h | 123 void mpol_shared_policy_init(struct shared_policy *sp, struct mempolicy *mpol); 125 struct vm_area_struct *vma, struct mempolicy *mpol); 144 struct mempolicy **mpol, nodemask_t **nodemask); 163 extern int mpol_parse_str(char *str, struct mempolicy **mpol); 214 struct mempolicy *mpol) in mpol_shared_policy_init() argument 260 struct mempolicy **mpol, nodemask_t **nodemask) in huge_node() argument 262 *mpol = NULL; in huge_node() 283 static inline int mpol_parse_str(char *str, struct mempolicy **mpol) in mpol_parse_str() argument
|
| H A D | gfp.h | 325 struct mempolicy *mpol, pgoff_t ilx, int nid); 338 struct mempolicy *mpol, pgoff_t ilx, int nid) in folio_alloc_mpol_noprof() argument
|
| H A D | shmem_fs.h | 87 struct mempolicy *mpol; /* default memory policy for mappings */ member
|
| /linux/mm/ |
| H A D | swap_state.c | 539 * @mpol: NUMA memory allocation policy to be applied in __swapin_nr_pages() 553 struct mempolicy *mpol, pgoff_t ilx, in __swapin_nr_pages() 571 folio = folio_alloc_mpol(gfp_mask, 0, mpol, ilx, numa_node_id()); in __swapin_nr_pages() 622 struct mempolicy *mpol; in swap_cluster_readahead() 630 mpol = get_vma_policy(vma, addr, 0, &ilx); in swap_cluster_readahead() 631 folio = swap_cache_alloc_folio(entry, gfp_mask, mpol, ilx, in swap_cluster_readahead() 633 mpol_cond_put(mpol); in swap_cluster_readahead() 707 * @mpol: NUMA memory allocation policy to be applied in swap_vma_ra_win() 722 struct mempolicy *mpol, pgoff_t ilx) 750 swp_entry(swp_type(entry), offset), gfp_mask, mpol, il in swap_vma_readahead() 405 __read_swap_cache_async(swp_entry_t entry,gfp_t gfp_mask,struct mempolicy * mpol,pgoff_t ilx,bool * new_page_allocated,bool skip_if_exists) __read_swap_cache_async() argument 518 struct mempolicy *mpol; read_swap_cache_async() local 618 swap_cluster_readahead(swp_entry_t entry,gfp_t gfp_mask,struct mempolicy * mpol,pgoff_t ilx) swap_cluster_readahead() argument 726 swap_vma_readahead(swp_entry_t targ_entry,gfp_t gfp_mask,struct mempolicy * mpol,pgoff_t targ_ilx,struct vm_fault * vmf) swap_vma_readahead() argument 813 struct mempolicy *mpol; swapin_readahead() local [all...] |
| H A D | shmem.c | 116 struct mempolicy *mpol; member 1741 static void shmem_show_mpol(struct seq_file *seq, struct mempolicy *mpol) in shmem_show_mpol() argument 1745 if (!mpol || mpol->mode == MPOL_DEFAULT) in shmem_show_mpol() 1748 mpol_to_str(buffer, sizeof(buffer), mpol); in shmem_show_mpol() 1755 struct mempolicy *mpol = NULL; in shmem_get_sbmpol() local 1756 if (sbinfo->mpol) { in shmem_get_sbmpol() 1758 mpol = sbinfo->mpol; in shmem_get_sbmpol() 1759 mpol_get(mpol); in shmem_get_sbmpol() 1762 return mpol; in shmem_get_sbmpol() 1765 static inline void shmem_show_mpol(struct seq_file *seq, struct mempolicy *mpol) in shmem_show_mpol() argument [all …]
|
| H A D | swap.h | 280 struct mempolicy *mpol, pgoff_t ilx, 297 struct mempolicy *mpol, pgoff_t ilx); in swap_zeromap_batch() 422 gfp_t gfp_mask, struct mempolicy *mpol, pgoff_t ilx) in __swap_cache_del_folio() 378 swap_cluster_readahead(swp_entry_t entry,gfp_t gfp_mask,struct mempolicy * mpol,pgoff_t ilx) swap_cluster_readahead() argument
|
| H A D | mempolicy.c | 2322 struct mempolicy **mpol, nodemask_t **nodemask) in huge_node() argument 2328 *mpol = get_vma_policy(vma, addr, hstate_vma(vma)->order, &ilx); in huge_node() 2329 *nodemask = policy_nodemask(gfp_flags, *mpol, ilx, &nid); in huge_node() 3204 void mpol_shared_policy_init(struct shared_policy *sp, struct mempolicy *mpol) in mpol_shared_policy_init() argument 3211 if (mpol) { in mpol_shared_policy_init() 3220 npol = mpol_new(mpol->mode, mpol->flags, &mpol->w.user_nodemask); in mpol_shared_policy_init() 3225 ret = mpol_set_nodemask(npol, &mpol->w.user_nodemask, scratch); in mpol_shared_policy_init() 3239 mpol_put(mpol); /* drop our incoming ref on sb mpol */ in mpol_shared_policy_init() 3413 int mpol_parse_str(char *str, struct mempolicy **mpol) in mpol_parse_str() argument 3533 *mpol = new; in mpol_parse_str()
|
| H A D | hugetlb.c | 1378 struct mempolicy *mpol; in dequeue_hugetlb_folio_vma() local 1391 nid = huge_node(vma, address, gfp_mask, &mpol, &nodemask); in dequeue_hugetlb_folio_vma() 1393 if (mpol_is_preferred_many(mpol)) { in dequeue_hugetlb_folio_vma() 1405 mpol_cond_put(mpol); in dequeue_hugetlb_folio_vma() 2212 struct mempolicy *mpol; in alloc_buddy_hugetlb_folio_with_mpol() local 2217 nid = huge_node(vma, addr, gfp_mask, &mpol, &nodemask); in alloc_buddy_hugetlb_folio_with_mpol() 2218 if (mpol_is_preferred_many(mpol)) { in alloc_buddy_hugetlb_folio_with_mpol() 2229 mpol_cond_put(mpol); in alloc_buddy_hugetlb_folio_with_mpol() 2280 struct mempolicy *mpol = get_task_policy(current); in policy_mbind_nodemask() local 2286 if (mpol->mode == MPOL_BIND && in policy_mbind_nodemask() [all …]
|
| H A D | zswap.c | 994 struct mempolicy *mpol; in zswap_writeback_entry() local 1004 mpol = get_task_policy(current); in zswap_writeback_entry() 1005 folio = swap_cache_alloc_folio(swpentry, GFP_KERNEL, mpol, in zswap_writeback_entry()
|
| H A D | slub.c | 4425 struct mempolicy *mpol = current->mempolicy; in __slab_alloc_node() local 4427 if (mpol) { in __slab_alloc_node() 4435 if (mpol->mode != MPOL_BIND || in __slab_alloc_node() 4436 !node_isset(numa_mem_id(), mpol->nodes)) in __slab_alloc_node() 4646 struct mempolicy *mpol = current->mempolicy; in alloc_from_pcs() local 4648 if (mpol) { in alloc_from_pcs() 4656 if (mpol->mode != MPOL_BIND || in alloc_from_pcs() 4657 !node_isset(numa_mem_id(), mpol->nodes)) in alloc_from_pcs()
|
| /linux/kernel/futex/ |
| H A D | core.c | 339 struct mempolicy *mpol; in __futex_key_to_node() local 345 mpol = vma_policy(vma); in __futex_key_to_node() 346 if (!mpol) in __futex_key_to_node() 349 switch (mpol->mode) { in __futex_key_to_node() 351 node = first_node(mpol->nodes); in __futex_key_to_node() 355 if (mpol->home_node != NUMA_NO_NODE) in __futex_key_to_node() 356 node = mpol->home_node; in __futex_key_to_node()
|
| /linux/ipc/ |
| H A D | shm.c | 571 static int shm_set_policy(struct vm_area_struct *vma, struct mempolicy *mpol) in shm_set_policy() argument 577 err = sfd->vm_ops->set_policy(vma, mpol); in shm_set_policy() 585 struct mempolicy *mpol = vma->vm_policy; in shm_get_policy() local 588 mpol = sfd->vm_ops->get_policy(vma, addr, ilx); in shm_get_policy() 589 return mpol; in shm_get_policy()
|
| /linux/virt/kvm/ |
| H A D | guest_memfd.c | 433 static int kvm_gmem_set_policy(struct vm_area_struct *vma, struct mempolicy *mpol) in kvm_gmem_set_policy() argument 437 return mpol_set_shared_policy(&GMEM_I(inode)->policy, vma, mpol); in kvm_gmem_set_policy()
|