| /linux/drivers/infiniband/hw/hns/ |
| H A D | hns_roce_alloc.c | 67 u32 page_shift, u32 flags) in hns_roce_buf_alloc() argument 76 if (WARN_ON(page_shift < HNS_HW_PAGE_SHIFT)) in hns_roce_buf_alloc() 84 buf->page_shift = page_shift; in hns_roce_buf_alloc() 85 page_size = 1 << buf->page_shift; in hns_roce_buf_alloc() 134 unsigned int page_shift) in hns_roce_get_kmem_bufs() argument 140 if (page_shift > buf->trunk_shift) { in hns_roce_get_kmem_bufs() 142 page_shift, buf->trunk_shift); in hns_roce_get_kmem_bufs() 150 offset += (1 << page_shift); in hns_roce_get_kmem_bufs() 157 unsigned int page_shift) in hns_roce_get_umem_bufs() argument 163 rdma_umem_for_each_dma_block(umem, &biter, 1 << page_shift) { in hns_roce_get_umem_bufs()
|
| H A D | hns_roce_trace.h | 152 __field(unsigned int, page_shift) 163 __entry->page_shift = attr->page_shift; 168 __entry->region_count, __entry->page_shift,
|
| /linux/drivers/net/ethernet/mellanox/mlx5/core/en/ |
| H A D | params.c | 48 u8 page_shift = mlx5e_mpwrq_page_shift(mdev, xsk); in mlx5e_mpwrq_umr_mode() local 53 oversized = xsk->chunk_size < (1 << page_shift); in mlx5e_mpwrq_umr_mode() 54 WARN_ON_ONCE(xsk->chunk_size > (1 << page_shift)); in mlx5e_mpwrq_umr_mode() 106 u8 mlx5e_mpwrq_log_wqe_sz(struct mlx5_core_dev *mdev, u8 page_shift, in mlx5e_mpwrq_log_wqe_sz() argument 118 max_log_wqe_size_calc = ilog2(max_pages_per_wqe) + page_shift; in mlx5e_mpwrq_log_wqe_sz() 128 u8 mlx5e_mpwrq_pages_per_wqe(struct mlx5_core_dev *mdev, u8 page_shift, in mlx5e_mpwrq_pages_per_wqe() argument 131 u8 log_wqe_sz = mlx5e_mpwrq_log_wqe_sz(mdev, page_shift, umr_mode); in mlx5e_mpwrq_pages_per_wqe() 134 pages_per_wqe = log_wqe_sz > page_shift ? (1 << (log_wqe_sz - page_shift)) : 1; in mlx5e_mpwrq_pages_per_wqe() 151 u16 mlx5e_mpwrq_umr_wqe_sz(struct mlx5_core_dev *mdev, u8 page_shift, in mlx5e_mpwrq_umr_wqe_sz() argument 154 u8 pages_per_wqe = mlx5e_mpwrq_pages_per_wqe(mdev, page_shift, umr_mode); in mlx5e_mpwrq_umr_wqe_sz() [all …]
|
| /linux/drivers/infiniband/core/ |
| H A D | umem_odp.c | 62 size_t page_size = 1UL << umem_odp->page_shift; in ib_init_umem_odp() 104 1 << umem_odp->page_shift); in ib_init_umem_odp() 150 umem_odp->page_shift = PAGE_SHIFT; in ib_umem_odp_alloc_implicit() 193 odp_data->page_shift = PAGE_SHIFT; in ib_umem_odp_alloc_child() 255 umem_odp->page_shift = PAGE_SHIFT; in ib_umem_odp_get() 258 umem_odp->page_shift = HPAGE_SHIFT; in ib_umem_odp_get() 331 unsigned int page_shift, hmm_order, pfn_start_idx; in ib_umem_odp_map_dma_and_lock() local 340 page_shift = umem_odp->page_shift; in ib_umem_odp_map_dma_and_lock() 354 range.start = ALIGN_DOWN(user_virt, 1UL << page_shift); in ib_umem_odp_map_dma_and_lock() 355 range.end = ALIGN(user_virt + bcnt, 1UL << page_shift); in ib_umem_odp_map_dma_and_lock() [all …]
|
| /linux/tools/testing/selftests/powerpc/mm/ |
| H A D | bad_accesses.c | 69 unsigned long i, j, addr, region_shift, page_shift, page_size; in test() local 84 page_shift = 16; in test() 86 page_shift = 12; in test() 103 (1 << page_shift) >> 10, in test() 121 for (j = page_shift - 1; j < 60; j++) { in test() 130 addr = (base | delta) & ~((1 << page_shift) - 1); in test()
|
| /linux/arch/powerpc/include/asm/ |
| H A D | ultravisor.h | 50 u64 page_shift) in uv_page_in() argument 53 page_shift); in uv_page_in() 57 u64 page_shift) in uv_page_out() argument 60 page_shift); in uv_page_out() 75 static inline int uv_page_inval(u64 lpid, u64 gpa, u64 page_shift) in uv_page_inval() argument 77 return ucall_norets(UV_PAGE_INVAL, lpid, gpa, page_shift); in uv_page_inval()
|
| H A D | kvm_book3s_uvmem.h | 15 unsigned long page_shift); 19 unsigned long page_shift); 54 unsigned long flags, unsigned long page_shift) in kvmppc_h_svm_page_in() argument 61 unsigned long flags, unsigned long page_shift) in kvmppc_h_svm_page_out() argument
|
| H A D | iommu.h | 171 __u32 page_shift, 176 __u32 page_shift, 299 extern int iommu_tce_check_ioba(unsigned long page_shift, 302 extern int iommu_tce_check_gpa(unsigned long page_shift,
|
| /linux/tools/testing/selftests/kvm/lib/arm64/ |
| H A D | processor.c | 31 unsigned int shift = (vm->pgtable_levels - 1) * (vm->page_shift - 3) + vm->page_shift; in pgd_index() 39 unsigned int shift = 2 * (vm->page_shift - 3) + vm->page_shift; in pud_index() 40 uint64_t mask = (1UL << (vm->page_shift - 3)) - 1; in pud_index() 50 unsigned int shift = (vm->page_shift - 3) + vm->page_shift; in pmd_index() 51 uint64_t mask = (1UL << (vm->page_shift - 3)) - 1; in pmd_index() 61 uint64_t mask = (1UL << (vm->page_shift - 3)) - 1; in pte_index() 62 return (gva >> vm->page_shift) & mask; in pte_index() 76 pte = pa & PTE_ADDR_MASK_LPA2(vm->page_shift); in addr_pte() 80 pte = pa & PTE_ADDR_MASK(vm->page_shift); in addr_pte() 81 if (vm->page_shift == 16) in addr_pte() [all …]
|
| /linux/tools/perf/bench/ |
| H A D | mem-functions.c | 84 unsigned int page_shift; member 240 p->page_shift != PAGE_SHIFT_4KB ? "has insufficient hugepages" : "is too large"); in __bench_mem_function() 284 p.page_shift = ilog2(page_size); in bench_mem_common() 345 static void *bench_mmap(size_t size, bool populate, unsigned int page_shift) in bench_mmap() argument 350 if (page_shift != PAGE_SHIFT_4KB) in bench_mmap() 351 extra |= MAP_HUGETLB | (page_shift << MAP_HUGE_SHIFT); in bench_mmap() 370 *dst = bench_mmap(p->size, true, p->page_shift); in mem_alloc() 374 *src = bench_mmap(p->size, true, p->page_shift); in mem_alloc() 483 static void mmap_page_touch(void *dst, size_t size, unsigned int page_shift, bool random) in mmap_page_touch() argument 485 unsigned long npages = size / (1 << page_shift); in mmap_page_touch() [all …]
|
| /linux/arch/powerpc/kvm/ |
| H A D | book3s_64_vio.c | 145 if ((tbltmp->it_page_shift <= stt->page_shift) && in kvm_spapr_tce_attach_iommu_group() 147 stt->offset << stt->page_shift) && in kvm_spapr_tce_attach_iommu_group() 149 stt->size << stt->page_shift)) { in kvm_spapr_tce_attach_iommu_group() 298 if (!args->size || args->page_shift < 12 || args->page_shift > 34 || in kvm_vm_ioctl_create_spapr_tce() 299 (args->offset + args->size > (ULLONG_MAX >> args->page_shift))) in kvm_vm_ioctl_create_spapr_tce() 313 stt->page_shift = args->page_shift; in kvm_vm_ioctl_create_spapr_tce() 380 if (iommu_tce_check_gpa(stt->page_shift, gpa)) in kvmppc_tce_validate() 437 unsigned long subpages = 1ULL << (stt->page_shift - tbl->it_page_shift); in kvmppc_clear_tce() 438 unsigned long io_entry = entry << (stt->page_shift - tbl->it_page_shift); in kvmppc_clear_tce() 495 unsigned long subpages = 1ULL << (stt->page_shift - tbl->it_page_shift); in kvmppc_tce_iommu_unmap() [all …]
|
| H A D | book3s_hv_uvmem.c | 515 unsigned long end, unsigned long page_shift, in __kvmppc_svm_page_out() argument 536 if (!kvmppc_gfn_is_uvmem_pfn(gpa >> page_shift, kvm, NULL)) in __kvmppc_svm_page_out() 568 ret = uv_page_out(kvm->arch.lpid, pfn << page_shift, in __kvmppc_svm_page_out() 569 gpa, 0, page_shift); in __kvmppc_svm_page_out() 588 unsigned long page_shift, in kvmppc_svm_page_out() argument 595 ret = __kvmppc_svm_page_out(vma, start, end, page_shift, kvm, gpa, in kvmppc_svm_page_out() 743 unsigned long page_shift, in kvmppc_svm_page_in() argument 780 ret = uv_page_in(kvm->arch.lpid, pfn << page_shift, in kvmppc_svm_page_in() 781 gpa, 0, page_shift); in kvmppc_svm_page_in() 878 unsigned long page_shift) in kvmppc_share_page() argument [all …]
|
| /linux/tools/testing/selftests/kvm/lib/loongarch/ |
| H A D | processor.c | 22 shift = level * (vm->page_shift - 3) + vm->page_shift; in virt_pte_index() 23 mask = (1UL << (vm->page_shift - 3)) - 1; in virt_pte_index() 29 return entry & ~((0x1UL << vm->page_shift) - 1); in pte_addr() 34 return 1 << (vm->page_shift - 3); in ptrs_per_pte() 43 ptrs_per_pte = 1 << (vm->page_shift - 3); in virt_set_pgtable() 127 (vaddr >> vm->page_shift)), in virt_arch_pg_map() 132 TEST_ASSERT((paddr >> vm->page_shift) <= vm->max_gfn, in virt_arch_pg_map() 298 width = vm->page_shift - 3; in loongarch_vcpu_setup() 303 val = (vm->page_shift + width * 2) << 20 | (width << 25); in loongarch_vcpu_setup() 307 val |= (vm->page_shift + width) << 10 | (width << 15); in loongarch_vcpu_setup() [all …]
|
| /linux/drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/ |
| H A D | vmm.c | 78 u8 page_shift = 29; /* 512MiB */ in r535_mmu_vaspace_new() local 79 const u64 page_size = BIT_ULL(page_shift); in r535_mmu_vaspace_new() 85 if (page->shift == page_shift) in r535_mmu_vaspace_new() 93 ret = nvkm_vmm_get_locked(vmm, true, false, false, page_shift, 32, page_size, in r535_mmu_vaspace_new() 116 page_shift += desc->bits; in r535_mmu_vaspace_new() 121 page_shift -= desc->bits; in r535_mmu_vaspace_new() 126 ctrl->levels[i].pageShift = page_shift; in r535_mmu_vaspace_new()
|
| /linux/tools/testing/selftests/mm/ |
| H A D | hmm-tests.c | 75 unsigned int page_shift; in FIXTURE() local 98 unsigned int page_shift; in FIXTURE() local 140 self->page_shift = ffs(self->page_size) - 1; in FIXTURE_SETUP() 151 self->page_shift = ffs(self->page_size) - 1; in FIXTURE_SETUP() 307 npages = ALIGN(HMM_BUFFER_SIZE, self->page_size) >> self->page_shift; in TEST_F() 309 size = npages << self->page_shift; in TEST_F() 370 npages = ALIGN(HMM_BUFFER_SIZE, self->page_size) >> self->page_shift; in TEST_F() 372 size = npages << self->page_shift; in TEST_F() 429 npages = ALIGN(HMM_BUFFER_SIZE, self->page_size) >> self->page_shift; in TEST_F() 431 size = npages << self->page_shift; in TEST_F() [all …]
|
| /linux/include/linux/ |
| H A D | kmsan.h | 146 unsigned int page_shift, 173 unsigned int page_shift); 352 struct page **pages, unsigned int page_shift, gfp_t gfp_mask) in kmsan_vmap_pages_range_noflush() argument 366 unsigned int page_shift) in kmsan_ioremap_page_range() argument
|
| /linux/tools/testing/selftests/kvm/lib/x86/ |
| H A D | vmx.c | 378 pte->address = paddr >> vm->page_shift; in nested_create_pte() 380 pte->address = vm_alloc_page_table(vm) >> vm->page_shift; in nested_create_pte() 414 TEST_ASSERT((nested_paddr >> vm->page_shift) <= vm->max_gfn, in __nested_pg_map() 422 TEST_ASSERT((paddr >> vm->page_shift) <= vm->max_gfn, in __nested_pg_map() 504 i = (region->region.guest_phys_addr >> vm->page_shift) - 1; in nested_map_memslot() 505 last = i + (region->region.memory_size >> vm->page_shift); in nested_map_memslot() 512 (uint64_t)i << vm->page_shift, in nested_map_memslot() 513 (uint64_t)i << vm->page_shift, in nested_map_memslot() 514 1 << vm->page_shift); in nested_map_memslot()
|
| /linux/arch/powerpc/platforms/powernv/ |
| H A D | pci-ioda-tce.c | 50 u64 dma_offset, unsigned int page_shift) in pnv_pci_setup_iommu_table() argument 54 tbl->it_page_shift = page_shift; in pnv_pci_setup_iommu_table() 291 __u32 page_shift, __u64 window_size, __u32 levels, in pnv_pci_ioda2_table_alloc_pages() argument 298 unsigned int entries_shift = window_shift - page_shift; in pnv_pci_ioda2_table_alloc_pages() 314 if ((level_shift - 3) * levels + page_shift >= 55) in pnv_pci_ioda2_table_alloc_pages() 348 page_shift); in pnv_pci_ioda2_table_alloc_pages()
|
| /linux/drivers/mtd/nand/raw/ |
| H A D | nand_bbt.c | 180 from = ((loff_t)page) << this->page_shift; in read_bbt() 396 scan_read(this, buf, (loff_t)td->pages[0] << this->page_shift, in read_abs_bbts() 405 scan_read(this, buf, (loff_t)md->pages[0] << this->page_shift, in read_abs_bbts() 561 int blocktopage = this->bbt_erase_shift - this->page_shift; in search_bbt() 666 (this->bbt_erase_shift - this->page_shift); in get_bbt_block() 694 page = block << (this->bbt_erase_shift - this->page_shift); in get_bbt_block() 797 page = block << (this->bbt_erase_shift - this->page_shift); in write_bbt() 818 to = ((loff_t)page) << this->page_shift; in write_bbt() 834 ops.ooblen = (len >> this->page_shift) * mtd->oobsize; in write_bbt() 841 pageoffs = page - (int)(to >> this->page_shift); in write_bbt() [all …]
|
| /linux/tools/testing/selftests/kvm/lib/ |
| H A D | kvm_util.c | 255 0, (1ULL << (vm->va_bits - 1)) >> vm->page_shift); in vm_vaddr_populate_bitmap() 257 (~((1ULL << (vm->va_bits - 1)) - 1)) >> vm->page_shift, in vm_vaddr_populate_bitmap() 258 (1ULL << (vm->va_bits - 1)) >> vm->page_shift); in vm_vaddr_populate_bitmap() 279 vm->page_shift = vm_guest_mode_params[vm->mode].page_shift; in ____vm_create() 967 TEST_ASSERT((((gpa >> vm->page_shift) + npages) - 1) in vm_mem_add() 1085 sparsebit_set_num(region->unused_phy_pages, gpa >> vm->page_shift, npages); in vm_mem_add() 1366 uint64_t pages = (sz + vm->page_size - 1) >> vm->page_shift; in vm_vaddr_unused_gap() 1369 uint64_t pgidx_start = (vaddr_min + vm->page_size - 1) >> vm->page_shift; in vm_vaddr_unused_gap() 1434 uint64_t pages = (sz >> vm->page_shift) + ((sz % vm->page_size) != 0); in ____vm_vaddr_alloc() 2086 base = pg = paddr_min >> vm->page_shift; in __vm_phy_pages_alloc() [all …]
|
| /linux/include/rdma/ |
| H A D | ib_umem_odp.h | 38 unsigned int page_shift; member 61 umem_odp->page_shift; in ib_umem_odp_num_pages()
|
| /linux/drivers/infiniband/sw/rxe/ |
| H A D | rxe_odp.c | 134 addr = iova & (~(BIT(umem_odp->page_shift) - 1)); in rxe_check_pagefault() 138 idx = (addr - ib_umem_start(umem_odp)) >> umem_odp->page_shift; in rxe_check_pagefault() 145 addr += BIT(umem_odp->page_shift); in rxe_check_pagefault() 152 return (iova - ib_umem_start(umem_odp)) >> umem_odp->page_shift; in rxe_odp_iova_to_index() 157 return iova & (BIT(umem_odp->page_shift) - 1); in rxe_odp_iova_to_page_offset() 210 bytes = BIT(umem_odp->page_shift) - offset; in __rxe_odp_mr_copy()
|
| /linux/tools/testing/selftests/kvm/include/arm64/ |
| H A D | processor.h | 119 #define PTE_ADDR_MASK(page_shift) GENMASK(47, (page_shift)) argument 122 #define PTE_ADDR_MASK_LPA2(page_shift) GENMASK(49, (page_shift)) argument
|
| /linux/drivers/infiniband/hw/mlx5/ |
| H A D | odp.c | 304 for (addr = start; addr < end; addr += BIT(umem_odp->page_shift)) { in mlx5_ib_invalidate_range() 305 idx = (addr - ib_umem_start(umem_odp)) >> umem_odp->page_shift; in mlx5_ib_invalidate_range() 696 int page_shift, ret, np; in pagefault_real_mr() local 709 page_shift = odp->page_shift; in pagefault_real_mr() 710 start_idx = (user_va - ib_umem_start(odp)) >> page_shift; in pagefault_real_mr() 723 ret = mlx5r_umr_update_xlt(mr, start_idx, np, page_shift, xlt_flags); in pagefault_real_mr() 734 u32 new_mappings = (np << page_shift) - in pagefault_real_mr() 735 (user_va - round_down(user_va, 1 << page_shift)); in pagefault_real_mr() 740 return np << (page_shift - PAGE_SHIFT); in pagefault_real_mr() 838 unsigned int old_page_shift = mr->page_shift; in pagefault_dmabuf_mr() [all …]
|
| /linux/drivers/net/ethernet/mellanox/mlx5/core/ |
| H A D | alloc.c | 81 buf->page_shift = PAGE_SHIFT; in mlx5_frag_buf_alloc_node() 95 if (frag->map & ((1 << buf->page_shift) - 1)) { in mlx5_frag_buf_alloc_node() 99 &frag->map, buf->page_shift); in mlx5_frag_buf_alloc_node()
|