Home
last modified time | relevance | path

Searched refs:page_offset (Results 1 – 25 of 136) sorted by relevance

123456

/linux/drivers/hv/
H A Dmshv_regions.c44 u64 page_offset, u64 page_count, in mshv_region_process_chunk() argument
47 u64 page_offset, in mshv_region_process_chunk()
55 page = region->pages[page_offset]; in mshv_region_process_chunk()
68 page = region->pages[page_offset + count]; in mshv_region_process_chunk()
79 ret = handler(region, flags, page_offset, count); in mshv_region_process_chunk()
107 u64 page_offset, u64 page_count, in mshv_region_process_range() argument
110 u64 page_offset, in mshv_region_process_range()
115 if (page_offset + page_count > region->nr_pages) in mshv_region_process_range()
120 if (!region->pages[page_offset]) { in mshv_region_process_range()
121 page_offset++; in mshv_region_process_range()
[all …]
/linux/drivers/infiniband/sw/rxe/
H A Drxe_mr.c240 mr->page_offset = mr->ibmr.iova & (page_size - 1); in rxe_map_mr_sg()
248 unsigned int page_offset = rxe_mr_iova_to_page_offset(mr, iova); in rxe_mr_copy_xarray() local
260 mr_page_size(mr) - page_offset); in rxe_mr_copy_xarray()
263 memcpy(addr, va + page_offset, bytes); in rxe_mr_copy_xarray()
265 memcpy(va + page_offset, addr, bytes); in rxe_mr_copy_xarray()
268 page_offset = 0; in rxe_mr_copy_xarray()
280 unsigned int page_offset = dma_addr & (PAGE_SIZE - 1); in rxe_mr_copy_dma() local
288 PAGE_SIZE - page_offset); in rxe_mr_copy_dma()
292 memcpy(va + page_offset, addr, bytes); in rxe_mr_copy_dma()
294 memcpy(addr, va + page_offset, bytes); in rxe_mr_copy_dma()
[all …]
H A Drxe_odp.c113 mr->page_offset = ib_umem_offset(&umem_odp->umem); in rxe_odp_mr_init_user()
268 unsigned int page_offset; in rxe_odp_do_atomic_op() local
286 page_offset = rxe_odp_iova_to_page_offset(umem_odp, iova); in rxe_odp_do_atomic_op()
287 if (unlikely(page_offset & 0x7)) { in rxe_odp_do_atomic_op()
298 value = *orig_val = va[page_offset >> 3]; in rxe_odp_do_atomic_op()
302 va[page_offset >> 3] = swap_add; in rxe_odp_do_atomic_op()
305 va[page_offset >> 3] = value; in rxe_odp_do_atomic_op()
336 unsigned int page_offset; in rxe_odp_flush_pmem_iova() local
350 page_offset = rxe_odp_iova_to_page_offset(umem_odp, iova); in rxe_odp_flush_pmem_iova()
355 mr_page_size(mr) - page_offset); in rxe_odp_flush_pmem_iova()
[all …]
/linux/drivers/infiniband/hw/mlx5/
H A Dmem.c64 u64 page_offset; in __mlx5_umem_find_best_quantized_pgoff() local
77 page_offset = ib_umem_dma_offset(umem, page_size); in __mlx5_umem_find_best_quantized_pgoff()
78 while (page_offset & ~(u64)(page_offset_mask * (page_size / scale))) { in __mlx5_umem_find_best_quantized_pgoff()
80 page_offset = ib_umem_dma_offset(umem, page_size); in __mlx5_umem_find_best_quantized_pgoff()
91 (unsigned long)page_offset / (page_size / scale); in __mlx5_umem_find_best_quantized_pgoff()
H A Dsrq_cmd.c17 u32 page_offset = in->page_offset; in get_pas_size() local
21 u32 rq_sz_po = rq_sz + (page_offset * po_quanta); in get_pas_size()
34 MLX5_SET(wq, wq, page_offset, in->page_offset); in set_wq()
47 MLX5_SET(srqc, srqc, page_offset, in->page_offset); in set_srqc()
62 in->page_offset = MLX5_GET(wq, wq, page_offset); in get_wq()
75 in->page_offset = MLX5_GET(srqc, srqc, page_offset); in get_srqc()
111 MLX5_ADAPTER_PAGE_SHIFT, page_offset, \
112 64, &(in)->page_offset))
/linux/drivers/net/ethernet/cavium/liquidio/
H A Docteon_network.h298 pg_info->page_offset = 0; in recv_buffer_alloc()
300 skb_pg_info->page_offset = 0; in recv_buffer_alloc()
324 skb_pg_info->page_offset = 0; in recv_buffer_fast_alloc()
348 pg_info->page_offset = 0; in recv_buffer_recycle()
353 if (pg_info->page_offset == 0) in recv_buffer_recycle()
354 pg_info->page_offset = LIO_RXBUFFER_SZ; in recv_buffer_recycle()
356 pg_info->page_offset = 0; in recv_buffer_recycle()
384 skb_pg_info->page_offset = pg_info->page_offset; in recv_buffer_reuse()
398 pg_info->page_offset = 0; in recv_buffer_destroy()
415 pg_info->page_offset = 0; in recv_buffer_free()
[all …]
/linux/drivers/gpu/drm/vmwgfx/
H A Dvmwgfx_page_dirty.c400 unsigned long page_offset; in vmw_bo_vm_mkwrite() local
415 page_offset = vmf->pgoff - drm_vma_node_start(&bo->base.vma_node); in vmw_bo_vm_mkwrite()
416 if (unlikely(page_offset >= PFN_UP(bo->resource->size))) { in vmw_bo_vm_mkwrite()
422 !test_bit(page_offset, &vbo->dirty->bitmap[0])) { in vmw_bo_vm_mkwrite()
425 __set_bit(page_offset, &dirty->bitmap[0]); in vmw_bo_vm_mkwrite()
426 dirty->start = min(dirty->start, page_offset); in vmw_bo_vm_mkwrite()
427 dirty->end = max(dirty->end, page_offset + 1); in vmw_bo_vm_mkwrite()
454 unsigned long page_offset; in vmw_bo_vm_fault() local
456 page_offset = vmf->pgoff - in vmw_bo_vm_fault()
458 if (page_offset >= PFN_UP(bo->resource->size) || in vmw_bo_vm_fault()
[all …]
/linux/drivers/scsi/fnic/
H A Dfnic_trace.c69 fnic_trace_entries.page_offset[fnic_trace_entries.wr_idx]; in fnic_trace_get_buf()
122 fnic_trace_entries.page_offset[rd_idx]; in fnic_get_trace_data()
163 fnic_trace_entries.page_offset[rd_idx]; in fnic_get_trace_data()
543 fnic_trace_entries.page_offset = in fnic_trace_buf_init()
545 if (!fnic_trace_entries.page_offset) { in fnic_trace_buf_init()
564 fnic_trace_entries.page_offset[i] = fnic_buf_head; in fnic_trace_buf_init()
582 if (fnic_trace_entries.page_offset) { in fnic_trace_free()
583 vfree((void *)fnic_trace_entries.page_offset); in fnic_trace_free()
584 fnic_trace_entries.page_offset = NULL; in fnic_trace_free()
624 fc_trace_entries.page_offset = in fnic_fc_trace_init()
[all …]
/linux/drivers/vfio/pci/pds/
H A Dlm.c200 size_t page_offset; in pds_vfio_save_read() local
206 page_offset = (*pos) % PAGE_SIZE; in pds_vfio_save_read()
207 page = pds_vfio_get_file_page(lm_file, *pos - page_offset); in pds_vfio_save_read()
214 page_len = min_t(size_t, len, PAGE_SIZE - page_offset); in pds_vfio_save_read()
216 err = copy_to_user(buf, from_buff + page_offset, page_len); in pds_vfio_save_read()
298 size_t page_offset; in pds_vfio_restore_write() local
304 page_offset = (*pos) % PAGE_SIZE; in pds_vfio_restore_write()
305 page = pds_vfio_get_file_page(lm_file, *pos - page_offset); in pds_vfio_restore_write()
312 page_len = min_t(size_t, len, PAGE_SIZE - page_offset); in pds_vfio_restore_write()
314 err = copy_from_user(to_buff + page_offset, buf, page_len); in pds_vfio_restore_write()
/linux/drivers/gpu/drm/ttm/
H A Dttm_bo_vm.c87 unsigned long page_offset) in ttm_bo_io_mem_pfn() argument
92 return bdev->funcs->io_mem_pfn(bo, page_offset); in ttm_bo_io_mem_pfn()
94 return (bo->resource->bus.offset >> PAGE_SHIFT) + page_offset; in ttm_bo_io_mem_pfn()
190 unsigned long page_offset; in ttm_bo_vm_fault_reserved() local
212 page_offset = ((address - vma->vm_start) >> PAGE_SHIFT) + in ttm_bo_vm_fault_reserved()
217 if (unlikely(page_offset >= PFN_UP(bo->base.size))) in ttm_bo_vm_fault_reserved()
248 pfn = ttm_bo_io_mem_pfn(bo, page_offset); in ttm_bo_vm_fault_reserved()
250 page = ttm->pages[page_offset]; in ttm_bo_vm_fault_reserved()
278 if (unlikely(++page_offset >= page_last)) in ttm_bo_vm_fault_reserved()
/linux/scripts/
H A Dleaking_addresses.pl330 state $page_offset = get_page_offset();
336 if (hex($match) < $page_offset) {
346 my $page_offset;
354 $page_offset = get_kernel_config_option('CONFIG_PAGE_OFFSET');
355 if (!$page_offset) {
358 return $page_offset;
/linux/net/ceph/
H A Dmessenger.c740 size_t *page_offset, in ceph_msg_data_bio_next() argument
746 *page_offset = bv.bv_offset; in ceph_msg_data_bio_next()
795 size_t *page_offset, in ceph_msg_data_bvecs_next() argument
801 *page_offset = bv.bv_offset; in ceph_msg_data_bvecs_next()
845 cursor->page_offset = data->alignment & ~PAGE_MASK; in ceph_msg_data_pages_cursor_init()
849 BUG_ON(length > SIZE_MAX - cursor->page_offset); in ceph_msg_data_pages_cursor_init()
854 size_t *page_offset, size_t *length) in ceph_msg_data_pages_next() argument
861 BUG_ON(cursor->page_offset >= PAGE_SIZE); in ceph_msg_data_pages_next()
863 *page_offset = cursor->page_offset; in ceph_msg_data_pages_next()
922 ceph_msg_data_pagelist_next(struct ceph_msg_data_cursor * cursor,size_t * page_offset,size_t * length) ceph_msg_data_pagelist_next() argument
985 ceph_msg_data_iter_next(struct ceph_msg_data_cursor * cursor,size_t * page_offset,size_t * length) ceph_msg_data_iter_next() argument
1087 ceph_msg_data_next(struct ceph_msg_data_cursor * cursor,size_t * page_offset,size_t * length) ceph_msg_data_next() argument
1165 ceph_crc32c_page(u32 crc,struct page * page,unsigned int page_offset,unsigned int length) ceph_crc32c_page() argument
[all...]
/linux/drivers/mtd/tests/
H A Dnandbiterrs.c45 static unsigned page_offset; variable
46 module_param(page_offset, uint, S_IRUGO);
47 MODULE_PARM_DESC(page_offset, "Page number relative to dev start");
358 offset = (loff_t)page_offset * mtd->writesize; in mtd_nandbiterrs_init()
362 page_offset, offset, eraseblock); in mtd_nandbiterrs_init()
/linux/drivers/nvmem/
H A Drave-sp-eeprom.c161 const unsigned int page_offset = offset % RAVE_SP_EEPROM_PAGE_SIZE; in rave_sp_eeprom_page_access() local
172 if (WARN_ON(data_len > sizeof(page.data) - page_offset)) in rave_sp_eeprom_page_access()
187 memcpy(&page.data[page_offset], data, data_len); in rave_sp_eeprom_page_access()
199 memcpy(data, &page.data[page_offset], data_len); in rave_sp_eeprom_page_access()
/linux/virt/kvm/
H A Dpfncache.c258 unsigned long page_offset; in __kvm_gpc_refresh()
284 page_offset = offset_in_page(uhva); in __kvm_gpc_refresh()
295 page_offset = offset_in_page(gpa); in __kvm_gpc_refresh()
323 gpc->uhva += page_offset; in __kvm_gpc_refresh()
337 gpc->khva = old_khva + page_offset; in __kvm_gpc_refresh()
250 unsigned long page_offset; __kvm_gpc_refresh() local
/linux/drivers/net/ethernet/sfc/falcon/
H A Drx.c59 return page_address(buf->page) + buf->page_offset; in ef4_rx_buf_va()
157 unsigned int page_offset; in ef4_init_rx_buffers() local
188 page_offset = sizeof(struct ef4_rx_page_state); in ef4_init_rx_buffers()
195 rx_buf->page_offset = page_offset + efx->rx_ip_align; in ef4_init_rx_buffers()
201 page_offset += efx->rx_page_buf_step; in ef4_init_rx_buffers()
202 } while (page_offset + efx->rx_page_buf_step <= PAGE_SIZE); in ef4_init_rx_buffers()
455 rx_buf->page, rx_buf->page_offset, in ef4_rx_packet_gro()
500 rx_buf->page_offset += hdr_len; in ef4_rx_mk_skb()
505 rx_buf->page, rx_buf->page_offset, in ef4_rx_mk_skb()
587 rx_buf->page_offset += efx->rx_prefix_size; in ef4_rx_packet()
/linux/fs/jfs/
H A Djfs_metapage.c194 int page_offset = mp->data - folio_address(src); in __metapage_migrate_folio() local
196 mp->data = folio_address(dst) + page_offset; in __metapage_migrate_folio()
237 int page_offset; in __metapage_migrate_folio() local
251 page_offset = mp->data - folio_address(src); in __metapage_migrate_folio()
252 mp->data = folio_address(dst) + page_offset; in __metapage_migrate_folio()
693 unsigned long page_offset; in __get_metapage() local
701 page_offset = (lblock - (page_index << l2BlocksPerPage)) << l2bsize; in __get_metapage()
702 if ((page_offset + size) > PAGE_SIZE) { in __get_metapage()
737 mp = folio_to_mp(folio, page_offset); in __get_metapage()
770 mp->data = folio_address(folio) + page_offset; in __get_metapage()
/linux/drivers/infiniband/hw/hfi1/
H A Dpin_system.c289 unsigned int page_offset; in add_mapping_to_sdma_packet() local
310 page_offset = start - ALIGN_DOWN(start, PAGE_SIZE); in add_mapping_to_sdma_packet()
311 from_this_page = PAGE_SIZE - page_offset; in add_mapping_to_sdma_packet()
327 page_offset, from_this_page, in add_mapping_to_sdma_packet()
338 ret, page_index, page_offset, from_this_page); in add_mapping_to_sdma_packet()
/linux/drivers/iommu/iommufd/
H A Dpages.c210 return iopt_area_iova(area) - area->page_offset + index * PAGE_SIZE; in iopt_area_index_to_iova()
221 return iopt_area_iova(area) - area->page_offset + in iopt_area_index_to_iova_last()
411 unsigned int page_offset = 0; in batch_from_domain() local
417 page_offset = area->page_offset; in batch_from_domain()
424 phys = iommu_iova_to_phys(domain, iova) - page_offset; in batch_from_domain()
427 iova += PAGE_SIZE - page_offset; in batch_from_domain()
428 page_offset = 0; in batch_from_domain()
439 unsigned int page_offset = 0; in raw_pages_from_domain() local
445 page_offset = area->page_offset; in raw_pages_from_domain()
447 phys = iommu_iova_to_phys(domain, iova) - page_offset; in raw_pages_from_domain()
[all …]
/linux/drivers/net/ethernet/mellanox/mlx4/
H A Den_rx.c72 frags->page_offset = priv->rx_headroom; in mlx4_en_alloc_frags()
77 rx_desc->data[i].addr = cpu_to_be64(dma + frags->page_offset); in mlx4_en_alloc_frags()
457 dma_sync_single_range_for_cpu(priv->ddev, dma, frags->page_offset, in mlx4_en_complete_rx_desc()
460 __skb_fill_page_desc(skb, nr, page, frags->page_offset, in mlx4_en_complete_rx_desc()
467 frags->page_offset ^= PAGE_SIZE / 2; in mlx4_en_complete_rx_desc()
479 frags->page_offset += sz_align; in mlx4_en_complete_rx_desc()
480 release = frags->page_offset + frag_info->frag_size > PAGE_SIZE; in mlx4_en_complete_rx_desc()
726 va = page_address(frags[0].page) + frags[0].page_offset; in mlx4_en_process_rx_cq()
756 dma += frags[0].page_offset; in mlx4_en_process_rx_cq()
796 dma += frags[0].page_offset; in mlx4_en_process_rx_cq()
[all …]
/linux/tools/testing/selftests/powerpc/primitives/
H A Dload_unaligned_zeropad.c102 static int do_one_test(char *p, int page_offset) in do_one_test() argument
114 …printf("offset %u load_unaligned_zeropad returned 0x%lx, should be 0x%lx\n", page_offset, got, sho… in do_one_test()
/linux/drivers/net/ethernet/google/gve/
H A Dgve_rx.c150 page_info->page_offset = 0; in gve_setup_rx_buffer()
228 rx->qpl_copy_pool[j].page_offset = 0; in gve_rx_prefill_pages()
455 u32 offset = page_info->page_offset + page_info->pad; in gve_rx_add_frags()
498 page_info->page_offset ^= GVE_DEFAULT_RX_BUFFER_OFFSET; in gve_rx_flip_buff()
543 void *src = page_info->page_address + page_info->page_offset; in gve_rx_copy_to_pool()
576 alloc_page_info.page_offset = 0; in gve_rx_copy_to_pool()
593 dst = copy_page_info->page_address + copy_page_info->page_offset; in gve_rx_copy_to_pool()
603 copy_page_info->page_offset ^= GVE_DEFAULT_RX_BUFFER_OFFSET; in gve_rx_copy_to_pool()
843 va = page_info->page_address + page_info->page_offset; in gve_rx()
851 be64_to_cpu(data_slot->addr) - page_info->page_offset : in gve_rx()
[all …]
/linux/drivers/net/ethernet/sfc/siena/
H A Drx.c93 rx_buf->page_offset += hdr_len; in efx_rx_mk_skb()
98 rx_buf->page, rx_buf->page_offset, in efx_rx_mk_skb()
177 rx_buf->page_offset += efx->rx_prefix_size; in efx_siena_rx_packet()
303 rx_buf->page_offset += offset; in efx_do_xdp()
/linux/drivers/net/ethernet/aquantia/atlantic/
H A Daq_ring.c73 rxpage->pg_off = rx_ring->page_offset; in aq_alloc_rxpages()
87 u16 page_offset = self->page_offset; in aq_get_rxpages() local
96 rxbuf->rxdata.pg_off += frame_max + page_offset + in aq_get_rxpages()
115 rxbuf->rxdata.pg_off = page_offset; in aq_get_rxpages()
191 self->page_offset = AQ_XDP_HEADROOM; in aq_ring_rx_alloc()
195 self->page_offset = 0; in aq_ring_rx_alloc()
707 int frame_sz = rx_ring->page_offset + rx_ring->frame_max + in __aq_ring_xdp_clean()
789 buff->rxdata.pg_off - rx_ring->page_offset; in __aq_ring_xdp_clean()
799 xdp_prepare_buff(&xdp, hard_start, rx_ring->page_offset, in __aq_ring_xdp_clean()
/linux/drivers/net/ethernet/qlogic/qede/
H A Dqede_fp.c59 sw_rx_data->page_offset = 0; in qede_alloc_rx_buffer()
535 new_mapping = curr_prod->mapping + curr_prod->page_offset; in qede_reuse_page()
563 curr_cons->page_offset += rxq->rx_buf_seg_size; in qede_realloc_rx_buffer()
565 if (curr_cons->page_offset == PAGE_SIZE) { in qede_realloc_rx_buffer()
570 curr_cons->page_offset -= rxq->rx_buf_seg_size; in qede_realloc_rx_buffer()
680 current_bd->page_offset + rxq->rx_headroom, in qede_fill_frag_skb()
747 buf = page_address(bd->data) + bd->page_offset; in qede_build_skb()
768 bd->page_offset += rxq->rx_buf_seg_size; in qede_tpa_rx_build_skb()
770 if (bd->page_offset == PAGE_SIZE) { in qede_tpa_rx_build_skb()
774 bd->page_offset -= rxq->rx_buf_seg_size; in qede_tpa_rx_build_skb()
[all …]

123456