Home
last modified time | relevance | path

Searched refs:page_shift (Results 1 – 25 of 111) sorted by relevance

12345

/linux/drivers/infiniband/hw/hns/
H A Dhns_roce_alloc.c67 u32 page_shift, u32 flags) in hns_roce_buf_alloc() argument
76 if (WARN_ON(page_shift < HNS_HW_PAGE_SHIFT)) in hns_roce_buf_alloc()
84 buf->page_shift = page_shift; in hns_roce_buf_alloc()
85 page_size = 1 << buf->page_shift; in hns_roce_buf_alloc()
134 unsigned int page_shift) in hns_roce_get_kmem_bufs() argument
140 if (page_shift > buf->trunk_shift) { in hns_roce_get_kmem_bufs()
142 page_shift, buf->trunk_shift); in hns_roce_get_kmem_bufs()
150 offset += (1 << page_shift); in hns_roce_get_kmem_bufs()
157 unsigned int page_shift) in hns_roce_get_umem_bufs() argument
163 rdma_umem_for_each_dma_block(umem, &biter, 1 << page_shift) { in hns_roce_get_umem_bufs()
H A Dhns_roce_mr.c99 buf_attr.page_shift = is_fast ? PAGE_SHIFT : in alloc_mr_pbl()
661 unsigned int page_shift) in mtr_check_direct_pages() argument
663 size_t page_size = 1 << page_shift; in mtr_check_direct_pages()
709 buf_attr->page_shift, in mtr_alloc_bufs()
750 unsigned int page_shift; in mtr_map_bufs() local
755 page_shift = need_split_huge_page(mtr) ? HNS_HW_PAGE_SHIFT : in mtr_map_bufs()
764 mtr->umem, page_shift); in mtr_map_bufs()
767 mtr->kmem, page_shift); in mtr_map_bufs()
777 ret = mtr_check_direct_pages(pages, npage, page_shift); in mtr_map_bufs()
949 buf_attr->page_shift = order_base_2(page_sz); in get_best_page_shift()
[all …]
H A Dhns_roce_device.h293 unsigned int page_shift; /* buffer page shift */ member
394 unsigned int page_shift; member
1119 return hns_roce_buf_dma_addr(buf, idx << buf->page_shift); in hns_roce_buf_page()
1129 static inline u32 to_hr_hw_page_shift(u32 page_shift) in to_hr_hw_page_shift() argument
1131 return page_shift - HNS_HW_PAGE_SHIFT; in to_hr_hw_page_shift()
1189 unsigned int page_shift, struct ib_udata *udata,
1239 u32 page_shift, u32 flags);
1243 unsigned int page_shift);
1246 unsigned int page_shift);
/linux/drivers/infiniband/core/
H A Dumem_odp.c59 size_t page_size = 1UL << umem_odp->page_shift; in ib_init_umem_odp()
73 ndmas = (end - start) >> umem_odp->page_shift; in ib_init_umem_odp()
134 umem_odp->page_shift = PAGE_SHIFT; in ib_umem_odp_alloc_implicit()
182 odp_data->page_shift = PAGE_SHIFT; in ib_umem_odp_alloc_child()
244 umem_odp->page_shift = PAGE_SHIFT; in ib_umem_odp_get()
247 umem_odp->page_shift = HPAGE_SHIFT; in ib_umem_odp_get()
315 *dma_addr = ib_dma_map_page(dev, page, 0, 1 << umem_odp->page_shift, in ib_umem_odp_map_dma_single_page()
353 unsigned int page_shift, hmm_order, pfn_start_idx; in ib_umem_odp_map_dma_and_lock() local
365 page_shift = umem_odp->page_shift; in ib_umem_odp_map_dma_and_lock()
379 range.start = ALIGN_DOWN(user_virt, 1UL << page_shift); in ib_umem_odp_map_dma_and_lock()
[all …]
/linux/drivers/net/ethernet/mellanox/mlx5/core/en/
H A Dparams.c44 u8 page_shift = mlx5e_mpwrq_page_shift(mdev, xsk); in mlx5e_mpwrq_umr_mode() local
49 oversized = xsk->chunk_size < (1 << page_shift); in mlx5e_mpwrq_umr_mode()
50 WARN_ON_ONCE(xsk->chunk_size > (1 << page_shift)); in mlx5e_mpwrq_umr_mode()
102 u8 mlx5e_mpwrq_log_wqe_sz(struct mlx5_core_dev *mdev, u8 page_shift, in mlx5e_mpwrq_log_wqe_sz() argument
113 max_log_mpwqe_size = ilog2(max_pages_per_wqe) + page_shift; in mlx5e_mpwrq_log_wqe_sz()
120 u8 mlx5e_mpwrq_pages_per_wqe(struct mlx5_core_dev *mdev, u8 page_shift, in mlx5e_mpwrq_pages_per_wqe() argument
123 u8 log_wqe_sz = mlx5e_mpwrq_log_wqe_sz(mdev, page_shift, umr_mode); in mlx5e_mpwrq_pages_per_wqe()
126 pages_per_wqe = log_wqe_sz > page_shift ? (1 << (log_wqe_sz - page_shift)) : 1; in mlx5e_mpwrq_pages_per_wqe()
143 u16 mlx5e_mpwrq_umr_wqe_sz(struct mlx5_core_dev *mdev, u8 page_shift, in mlx5e_mpwrq_umr_wqe_sz() argument
146 u8 pages_per_wqe = mlx5e_mpwrq_pages_per_wqe(mdev, page_shift, umr_mode); in mlx5e_mpwrq_umr_wqe_sz()
[all …]
H A Dparams.h63 u8 mlx5e_mpwrq_log_wqe_sz(struct mlx5_core_dev *mdev, u8 page_shift,
65 u8 mlx5e_mpwrq_pages_per_wqe(struct mlx5_core_dev *mdev, u8 page_shift,
67 u16 mlx5e_mpwrq_umr_wqe_sz(struct mlx5_core_dev *mdev, u8 page_shift,
69 u8 mlx5e_mpwrq_umr_wqebbs(struct mlx5_core_dev *mdev, u8 page_shift,
71 u8 mlx5e_mpwrq_mtts_per_wqe(struct mlx5_core_dev *mdev, u8 page_shift,
75 u8 mlx5e_mpwrq_max_log_rq_pkts(struct mlx5_core_dev *mdev, u8 page_shift,
/linux/drivers/pci/endpoint/
H A Dpci-epc-mem.c26 unsigned int page_shift = ilog2(mem->window.page_size); in pci_epc_mem_get_order() local
29 size >>= page_shift; in pci_epc_mem_get_order()
53 unsigned int page_shift; in pci_epc_multi_mem_init() local
73 page_shift = ilog2(page_size); in pci_epc_multi_mem_init()
74 pages = windows[i].size >> page_shift; in pci_epc_multi_mem_init()
183 unsigned int page_shift; in pci_epc_mem_alloc_addr() local
198 page_shift = ilog2(mem->window.page_size); in pci_epc_mem_alloc_addr()
200 ((phys_addr_t)pageno << page_shift); in pci_epc_mem_alloc_addr()
248 unsigned int page_shift; in pci_epc_mem_free_addr() local
260 page_shift = ilog2(page_size); in pci_epc_mem_free_addr()
[all …]
/linux/tools/testing/selftests/kvm/lib/aarch64/
H A Dprocessor.c30 unsigned int shift = (vm->pgtable_levels - 1) * (vm->page_shift - 3) + vm->page_shift; in pgd_index()
38 unsigned int shift = 2 * (vm->page_shift - 3) + vm->page_shift; in pud_index()
39 uint64_t mask = (1UL << (vm->page_shift - 3)) - 1; in pud_index()
49 unsigned int shift = (vm->page_shift - 3) + vm->page_shift; in pmd_index()
50 uint64_t mask = (1UL << (vm->page_shift - 3)) - 1; in pmd_index()
60 uint64_t mask = (1UL << (vm->page_shift - 3)) - 1; in pte_index()
61 return (gva >> vm->page_shift) & mask; in pte_index()
75 pte = pa & GENMASK(49, vm->page_shift); in addr_pte()
79 pte = pa & GENMASK(47, vm->page_shift); in addr_pte()
80 if (vm->page_shift == 16) in addr_pte()
[all …]
/linux/tools/testing/selftests/powerpc/mm/
H A Dbad_accesses.c69 unsigned long i, j, addr, region_shift, page_shift, page_size; in test() local
84 page_shift = 16; in test()
86 page_shift = 12; in test()
103 (1 << page_shift) >> 10, in test()
121 for (j = page_shift - 1; j < 60; j++) { in test()
130 addr = (base | delta) & ~((1 << page_shift) - 1); in test()
/linux/arch/powerpc/include/asm/
H A Dultravisor.h50 u64 page_shift) in uv_page_in() argument
53 page_shift); in uv_page_in()
57 u64 page_shift) in uv_page_out() argument
60 page_shift); in uv_page_out()
75 static inline int uv_page_inval(u64 lpid, u64 gpa, u64 page_shift) in uv_page_inval() argument
77 return ucall_norets(UV_PAGE_INVAL, lpid, gpa, page_shift); in uv_page_inval()
H A Dkvm_book3s_uvmem.h15 unsigned long page_shift);
19 unsigned long page_shift);
54 unsigned long flags, unsigned long page_shift) in kvmppc_h_svm_page_in() argument
61 unsigned long flags, unsigned long page_shift) in kvmppc_h_svm_page_out() argument
H A Diommu.h171 __u32 page_shift,
176 __u32 page_shift,
299 extern int iommu_tce_check_ioba(unsigned long page_shift,
302 extern int iommu_tce_check_gpa(unsigned long page_shift,
/linux/arch/powerpc/kvm/
H A Dbook3s_hv_uvmem.c515 unsigned long end, unsigned long page_shift, in __kvmppc_svm_page_out() argument
536 if (!kvmppc_gfn_is_uvmem_pfn(gpa >> page_shift, kvm, NULL)) in __kvmppc_svm_page_out()
568 ret = uv_page_out(kvm->arch.lpid, pfn << page_shift, in __kvmppc_svm_page_out()
569 gpa, 0, page_shift); in __kvmppc_svm_page_out()
588 unsigned long page_shift, in kvmppc_svm_page_out() argument
595 ret = __kvmppc_svm_page_out(vma, start, end, page_shift, kvm, gpa, in kvmppc_svm_page_out()
743 unsigned long page_shift, in kvmppc_svm_page_in() argument
780 ret = uv_page_in(kvm->arch.lpid, pfn << page_shift, in kvmppc_svm_page_in()
781 gpa, 0, page_shift); in kvmppc_svm_page_in()
878 unsigned long page_shift) in kvmppc_share_page() argument
[all …]
/linux/tools/testing/selftests/mm/
H A Dhmm-tests.c72 unsigned int page_shift; in FIXTURE() local
95 unsigned int page_shift; in FIXTURE() local
137 self->page_shift = ffs(self->page_size) - 1; in FIXTURE_SETUP()
148 self->page_shift = ffs(self->page_size) - 1; in FIXTURE_SETUP()
302 npages = ALIGN(HMM_BUFFER_SIZE, self->page_size) >> self->page_shift; in TEST_F()
304 size = npages << self->page_shift; in TEST_F()
365 npages = ALIGN(HMM_BUFFER_SIZE, self->page_size) >> self->page_shift; in TEST_F()
367 size = npages << self->page_shift; in TEST_F()
424 npages = ALIGN(HMM_BUFFER_SIZE, self->page_size) >> self->page_shift; in TEST_F()
426 size = npages << self->page_shift; in TEST_F()
[all …]
/linux/arch/powerpc/platforms/pseries/
H A Diommu.c564 unsigned long window_size, unsigned long page_shift, in iommu_table_setparms_common() argument
569 tbl->it_offset = win_addr >> page_shift; in iommu_table_setparms_common()
570 tbl->it_size = window_size >> page_shift; in iommu_table_setparms_common()
571 tbl->it_page_shift = page_shift; in iommu_table_setparms_common()
1242 struct ddw_create_response *create, int page_shift, in create_ddw() argument
1266 BUID_LO(buid), page_shift, window_shift); in create_ddw()
1272 BUID_LO(buid), page_shift, window_shift, ret, create->liobn, in create_ddw()
1362 u32 page_shift, u32 window_shift) in ddw_property_create() argument
1384 ddwprop->tce_shift = cpu_to_be32(page_shift); in ddw_property_create()
1407 int page_shift; in enable_ddw() local
[all …]
/linux/drivers/net/ethernet/mellanox/mlx4/
H A Dmr.c194 int mlx4_mtt_init(struct mlx4_dev *dev, int npages, int page_shift, in mlx4_mtt_init() argument
201 mtt->page_shift = MLX4_ICM_PAGE_SHIFT; in mlx4_mtt_init()
204 mtt->page_shift = page_shift; in mlx4_mtt_init()
419 int page_shift, struct mlx4_mr *mr) in mlx4_mr_alloc_reserved() argument
428 return mlx4_mtt_init(dev, npages, page_shift, &mr->mtt); in mlx4_mr_alloc_reserved()
528 int npages, int page_shift, struct mlx4_mr *mr) in mlx4_mr_alloc() argument
538 access, npages, page_shift, mr); in mlx4_mr_alloc()
591 int page_shift, struct mlx4_mpt_entry *mpt_entry) in mlx4_mr_rereg_mem_write() argument
595 err = mlx4_mtt_init(dev, npages, page_shift, &mr->mtt); in mlx4_mr_rereg_mem_write()
601 mpt_entry->entity_size = cpu_to_be32(page_shift); in mlx4_mr_rereg_mem_write()
[all …]
/linux/include/linux/
H A Dkmsan.h145 unsigned int page_shift);
171 unsigned int page_shift);
351 struct page **pages, unsigned int page_shift) in kmsan_vmap_pages_range_noflush() argument
365 unsigned int page_shift) in kmsan_ioremap_page_range() argument
/linux/tools/testing/selftests/kvm/lib/x86_64/
H A Dvmx.c380 pte->address = paddr >> vm->page_shift; in nested_create_pte()
382 pte->address = vm_alloc_page_table(vm) >> vm->page_shift; in nested_create_pte()
416 TEST_ASSERT((nested_paddr >> vm->page_shift) <= vm->max_gfn, in __nested_pg_map()
424 TEST_ASSERT((paddr >> vm->page_shift) <= vm->max_gfn, in __nested_pg_map()
506 i = (region->region.guest_phys_addr >> vm->page_shift) - 1; in nested_map_memslot()
507 last = i + (region->region.memory_size >> vm->page_shift); in nested_map_memslot()
514 (uint64_t)i << vm->page_shift, in nested_map_memslot()
515 (uint64_t)i << vm->page_shift, in nested_map_memslot()
516 1 << vm->page_shift); in nested_map_memslot()
/linux/arch/powerpc/platforms/powernv/
H A Dpci-ioda-tce.c50 u64 dma_offset, unsigned int page_shift) in pnv_pci_setup_iommu_table() argument
54 tbl->it_page_shift = page_shift; in pnv_pci_setup_iommu_table()
291 __u32 page_shift, __u64 window_size, __u32 levels, in pnv_pci_ioda2_table_alloc_pages() argument
298 unsigned int entries_shift = window_shift - page_shift; in pnv_pci_ioda2_table_alloc_pages()
314 if ((level_shift - 3) * levels + page_shift >= 55) in pnv_pci_ioda2_table_alloc_pages()
348 page_shift); in pnv_pci_ioda2_table_alloc_pages()
/linux/drivers/mtd/nand/raw/
H A Dnand_bbt.c180 from = ((loff_t)page) << this->page_shift; in read_bbt()
396 scan_read(this, buf, (loff_t)td->pages[0] << this->page_shift, in read_abs_bbts()
405 scan_read(this, buf, (loff_t)md->pages[0] << this->page_shift, in read_abs_bbts()
561 int blocktopage = this->bbt_erase_shift - this->page_shift; in search_bbt()
666 (this->bbt_erase_shift - this->page_shift); in get_bbt_block()
694 page = block << (this->bbt_erase_shift - this->page_shift); in get_bbt_block()
797 page = block << (this->bbt_erase_shift - this->page_shift); in write_bbt()
818 to = ((loff_t)page) << this->page_shift; in write_bbt()
834 ops.ooblen = (len >> this->page_shift) * mtd->oobsize; in write_bbt()
841 pageoffs = page - (int)(to >> this->page_shift); in write_bbt()
[all …]
/linux/tools/testing/selftests/bpf/progs/
H A Dbpf_iter_vma_offset.c13 __u32 page_shift = 0; variable
33 offset = address - vma->vm_start + (vma->vm_pgoff << page_shift); in get_vma_offset()
/linux/tools/testing/selftests/kvm/lib/
H A Dkvm_util.c262 0, (1ULL << (vm->va_bits - 1)) >> vm->page_shift); in vm_vaddr_populate_bitmap()
264 (~((1ULL << (vm->va_bits - 1)) - 1)) >> vm->page_shift, in vm_vaddr_populate_bitmap()
265 (1ULL << (vm->va_bits - 1)) >> vm->page_shift); in vm_vaddr_populate_bitmap()
286 vm->page_shift = vm_guest_mode_params[vm->mode].page_shift; in ____vm_create()
925 TEST_ASSERT((((guest_paddr >> vm->page_shift) + npages) - 1) in vm_mem_add()
1050 guest_paddr >> vm->page_shift, npages); in vm_mem_add()
1316 uint64_t pages = (sz + vm->page_size - 1) >> vm->page_shift; in vm_vaddr_unused_gap()
1319 uint64_t pgidx_start = (vaddr_min + vm->page_size - 1) >> vm->page_shift; in vm_vaddr_unused_gap()
1384 uint64_t pages = (sz >> vm->page_shift) + ((sz % vm->page_size) != 0); in ____vm_vaddr_alloc()
1403 sparsebit_set(vm->vpages_mapped, vaddr >> vm->page_shift); in ____vm_vaddr_alloc()
[all …]
/linux/include/rdma/
H A Dib_umem_odp.h44 unsigned int page_shift; member
67 umem_odp->page_shift; in ib_umem_odp_num_pages()
/linux/drivers/net/ethernet/mellanox/mlx5/core/
H A Dalloc.c81 buf->page_shift = PAGE_SHIFT; in mlx5_frag_buf_alloc_node()
95 if (frag->map & ((1 << buf->page_shift) - 1)) { in mlx5_frag_buf_alloc_node()
99 &frag->map, buf->page_shift); in mlx5_frag_buf_alloc_node()
/linux/mm/kmsan/
H A Dshadow.c219 unsigned int page_shift) in kmsan_vmap_pages_range_noflush() argument
250 s_pages, page_shift); in kmsan_vmap_pages_range_noflush()
256 o_pages, page_shift); in kmsan_vmap_pages_range_noflush()

12345