| /linux/drivers/hv/ |
| H A D | mshv_regions.c | 79 u64 page_offset, u64 page_count, in mshv_region_process_chunk() argument 82 u64 page_offset, in mshv_region_process_chunk() 86 u64 gfn = region->start_gfn + page_offset; in mshv_region_process_chunk() 91 page = region->mreg_pages[page_offset]; in mshv_region_process_chunk() 101 page = region->mreg_pages[page_offset + count]; in mshv_region_process_chunk() 113 ret = handler(region, flags, page_offset, count, stride > 1); in mshv_region_process_chunk() 141 u64 page_offset, u64 page_count, in mshv_region_process_range() argument 144 u64 page_offset, in mshv_region_process_range() 150 if (page_offset + page_count > region->nr_pages) in mshv_region_process_range() 155 if (!region->mreg_pages[page_offset]) { in mshv_region_process_range() [all …]
|
| /linux/drivers/infiniband/sw/rxe/ |
| H A D | rxe_mr.c | 334 unsigned int page_offset = rxe_mr_iova_to_page_offset(mr, iova); in rxe_mr_copy_xarray() local 339 page_offset += info->offset; in rxe_mr_copy_xarray() 340 bytes = min_t(unsigned int, length, PAGE_SIZE - page_offset); in rxe_mr_copy_xarray() 344 memcpy(addr, va + page_offset, bytes); in rxe_mr_copy_xarray() 346 memcpy(va + page_offset, addr, bytes); in rxe_mr_copy_xarray() 360 unsigned int page_offset = dma_addr & (PAGE_SIZE - 1); in rxe_mr_copy_dma() local 368 PAGE_SIZE - page_offset); in rxe_mr_copy_dma() 372 memcpy(va + page_offset, addr, bytes); in rxe_mr_copy_dma() 374 memcpy(addr, va + page_offset, bytes); in rxe_mr_copy_dma() 377 page_offset = 0; in rxe_mr_copy_dma() [all …]
|
| H A D | rxe_odp.c | 267 unsigned int page_offset; in rxe_odp_do_atomic_op() local 285 page_offset = rxe_odp_iova_to_page_offset(umem_odp, iova); in rxe_odp_do_atomic_op() 286 if (unlikely(page_offset & 0x7)) { in rxe_odp_do_atomic_op() 297 value = *orig_val = va[page_offset >> 3]; in rxe_odp_do_atomic_op() 301 va[page_offset >> 3] = swap_add; in rxe_odp_do_atomic_op() 304 va[page_offset >> 3] = value; in rxe_odp_do_atomic_op() 335 unsigned int page_offset; in rxe_odp_flush_pmem_iova() local 349 page_offset = rxe_odp_iova_to_page_offset(umem_odp, iova); in rxe_odp_flush_pmem_iova() 354 mr_page_size(mr) - page_offset); in rxe_odp_flush_pmem_iova() 357 arch_wb_cache_pmem(va + page_offset, bytes); in rxe_odp_flush_pmem_iova() [all …]
|
| /linux/drivers/infiniband/hw/mlx5/ |
| H A D | mem.c | 64 u64 page_offset; in __mlx5_umem_find_best_quantized_pgoff() local 77 page_offset = ib_umem_dma_offset(umem, page_size); in __mlx5_umem_find_best_quantized_pgoff() 78 while (page_offset & ~(u64)(page_offset_mask * (page_size / scale))) { in __mlx5_umem_find_best_quantized_pgoff() 80 page_offset = ib_umem_dma_offset(umem, page_size); in __mlx5_umem_find_best_quantized_pgoff() 91 (unsigned long)page_offset / (page_size / scale); in __mlx5_umem_find_best_quantized_pgoff()
|
| H A D | srq_cmd.c | 17 u32 page_offset = in->page_offset; in get_pas_size() local 21 u32 rq_sz_po = rq_sz + (page_offset * po_quanta); in get_pas_size() 34 MLX5_SET(wq, wq, page_offset, in->page_offset); in set_wq() 47 MLX5_SET(srqc, srqc, page_offset, in->page_offset); in set_srqc() 62 in->page_offset = MLX5_GET(wq, wq, page_offset); in get_wq() 75 in->page_offset = MLX5_GET(srqc, srqc, page_offset); in get_srqc() 111 MLX5_ADAPTER_PAGE_SHIFT, page_offset, \ 112 64, &(in)->page_offset))
|
| /linux/drivers/net/ethernet/cavium/liquidio/ |
| H A D | octeon_network.h | 298 pg_info->page_offset = 0; in recv_buffer_alloc() 300 skb_pg_info->page_offset = 0; in recv_buffer_alloc() 324 skb_pg_info->page_offset = 0; in recv_buffer_fast_alloc() 348 pg_info->page_offset = 0; in recv_buffer_recycle() 353 if (pg_info->page_offset == 0) in recv_buffer_recycle() 354 pg_info->page_offset = LIO_RXBUFFER_SZ; in recv_buffer_recycle() 356 pg_info->page_offset = 0; in recv_buffer_recycle() 384 skb_pg_info->page_offset = pg_info->page_offset; in recv_buffer_reuse() 398 pg_info->page_offset = 0; in recv_buffer_destroy() 415 pg_info->page_offset = 0; in recv_buffer_free() [all …]
|
| /linux/drivers/gpu/drm/vmwgfx/ |
| H A D | vmwgfx_page_dirty.c | 400 unsigned long page_offset; in vmw_bo_vm_mkwrite() local 415 page_offset = vmf->pgoff - drm_vma_node_start(&bo->base.vma_node); in vmw_bo_vm_mkwrite() 416 if (unlikely(page_offset >= PFN_UP(bo->resource->size))) { in vmw_bo_vm_mkwrite() 422 !test_bit(page_offset, &vbo->dirty->bitmap[0])) { in vmw_bo_vm_mkwrite() 425 __set_bit(page_offset, &dirty->bitmap[0]); in vmw_bo_vm_mkwrite() 426 dirty->start = min(dirty->start, page_offset); in vmw_bo_vm_mkwrite() 427 dirty->end = max(dirty->end, page_offset + 1); in vmw_bo_vm_mkwrite() 454 unsigned long page_offset; in vmw_bo_vm_fault() local 456 page_offset = vmf->pgoff - in vmw_bo_vm_fault() 458 if (page_offset >= PFN_UP(bo->resource->size) || in vmw_bo_vm_fault() [all …]
|
| /linux/drivers/scsi/fnic/ |
| H A D | fnic_trace.c | 69 fnic_trace_entries.page_offset[fnic_trace_entries.wr_idx]; in fnic_trace_get_buf() 122 fnic_trace_entries.page_offset[rd_idx]; in fnic_get_trace_data() 163 fnic_trace_entries.page_offset[rd_idx]; in fnic_get_trace_data() 543 fnic_trace_entries.page_offset = in fnic_trace_buf_init() 545 if (!fnic_trace_entries.page_offset) { in fnic_trace_buf_init() 564 fnic_trace_entries.page_offset[i] = fnic_buf_head; in fnic_trace_buf_init() 582 if (fnic_trace_entries.page_offset) { in fnic_trace_free() 583 vfree((void *)fnic_trace_entries.page_offset); in fnic_trace_free() 584 fnic_trace_entries.page_offset = NULL; in fnic_trace_free() 624 fc_trace_entries.page_offset = in fnic_fc_trace_init() [all …]
|
| /linux/drivers/vfio/pci/pds/ |
| H A D | lm.c | 200 size_t page_offset; in pds_vfio_save_read() local 206 page_offset = (*pos) % PAGE_SIZE; in pds_vfio_save_read() 207 page = pds_vfio_get_file_page(lm_file, *pos - page_offset); in pds_vfio_save_read() 214 page_len = min_t(size_t, len, PAGE_SIZE - page_offset); in pds_vfio_save_read() 216 err = copy_to_user(buf, from_buff + page_offset, page_len); in pds_vfio_save_read() 298 size_t page_offset; in pds_vfio_restore_write() local 304 page_offset = (*pos) % PAGE_SIZE; in pds_vfio_restore_write() 305 page = pds_vfio_get_file_page(lm_file, *pos - page_offset); in pds_vfio_restore_write() 312 page_len = min_t(size_t, len, PAGE_SIZE - page_offset); in pds_vfio_restore_write() 314 err = copy_from_user(to_buff + page_offset, buf, page_len); in pds_vfio_restore_write()
|
| /linux/scripts/ |
| H A D | leaking_addresses.pl | 330 state $page_offset = get_page_offset(); 336 if (hex($match) < $page_offset) { 346 my $page_offset; 354 $page_offset = get_kernel_config_option('CONFIG_PAGE_OFFSET'); 355 if (!$page_offset) { 358 return $page_offset;
|
| /linux/net/ceph/ |
| H A D | messenger.c | 740 size_t *page_offset, in ceph_msg_data_bio_next() argument 746 *page_offset = bv.bv_offset; in ceph_msg_data_bio_next() 795 size_t *page_offset, in ceph_msg_data_bvecs_next() argument 801 *page_offset = bv.bv_offset; in ceph_msg_data_bvecs_next() 845 cursor->page_offset = data->alignment & ~PAGE_MASK; in ceph_msg_data_pages_cursor_init() 849 BUG_ON(length > SIZE_MAX - cursor->page_offset); in ceph_msg_data_pages_cursor_init() 854 size_t *page_offset, size_t *length) in ceph_msg_data_pages_next() argument 861 BUG_ON(cursor->page_offset >= PAGE_SIZE); in ceph_msg_data_pages_next() 863 *page_offset = cursor->page_offset; in ceph_msg_data_pages_next() 864 *length = min_t(size_t, cursor->resid, PAGE_SIZE - *page_offset); in ceph_msg_data_pages_next() [all …]
|
| /linux/drivers/mtd/tests/ |
| H A D | nandbiterrs.c | 45 static unsigned page_offset; variable 46 module_param(page_offset, uint, S_IRUGO); 47 MODULE_PARM_DESC(page_offset, "Page number relative to dev start"); 358 offset = (loff_t)page_offset * mtd->writesize; in mtd_nandbiterrs_init() 362 page_offset, offset, eraseblock); in mtd_nandbiterrs_init()
|
| /linux/drivers/nvmem/ |
| H A D | rave-sp-eeprom.c | 161 const unsigned int page_offset = offset % RAVE_SP_EEPROM_PAGE_SIZE; in rave_sp_eeprom_page_access() local 172 if (WARN_ON(data_len > sizeof(page.data) - page_offset)) in rave_sp_eeprom_page_access() 187 memcpy(&page.data[page_offset], data, data_len); in rave_sp_eeprom_page_access() 199 memcpy(data, &page.data[page_offset], data_len); in rave_sp_eeprom_page_access()
|
| /linux/virt/kvm/ |
| H A D | pfncache.c | 258 unsigned long page_offset; in __kvm_gpc_refresh() 284 page_offset = offset_in_page(uhva); in __kvm_gpc_refresh() 295 page_offset = offset_in_page(gpa); in __kvm_gpc_refresh() 323 gpc->uhva += page_offset; in __kvm_gpc_refresh() 337 gpc->khva = old_khva + page_offset; in __kvm_gpc_refresh() 250 unsigned long page_offset; __kvm_gpc_refresh() local
|
| /linux/drivers/gpu/drm/qxl/ |
| H A D | qxl_image.c | 168 unsigned int page_base, page_offset, out_offset; in qxl_image_init_helper() local 177 page_offset = offset_in_page(out_offset); in qxl_image_init_helper() 178 size = min((int)(PAGE_SIZE - page_offset), remain); in qxl_image_init_helper() 181 k_data = ptr + page_offset; in qxl_image_init_helper()
|
| /linux/drivers/net/ethernet/huawei/hinic3/ |
| H A D | hinic3_rx.c | 100 u32 page_offset; in rx_alloc_mapped_page() local 105 page = page_pool_dev_alloc_frag(page_pool, &page_offset, buf_len); in rx_alloc_mapped_page() 110 rx_info->page_offset = page_offset; in rx_alloc_mapped_page() 162 rx_info->page_offset; in hinic3_rx_fill_buffers() 221 va = (u8 *)page_address(page) + rx_info->page_offset; in hinic3_add_rx_frag() 224 page_pool_dma_sync_for_cpu(rxq->page_pool, page, rx_info->page_offset, in hinic3_add_rx_frag() 236 rx_info->page_offset, size, rxq->buf_len); in hinic3_add_rx_frag()
|
| /linux/drivers/net/ethernet/sfc/falcon/ |
| H A D | rx.c | 59 return page_address(buf->page) + buf->page_offset; in ef4_rx_buf_va() 157 unsigned int page_offset; in ef4_init_rx_buffers() local 188 page_offset = sizeof(struct ef4_rx_page_state); in ef4_init_rx_buffers() 195 rx_buf->page_offset = page_offset + efx->rx_ip_align; in ef4_init_rx_buffers() 201 page_offset += efx->rx_page_buf_step; in ef4_init_rx_buffers() 202 } while (page_offset + efx->rx_page_buf_step <= PAGE_SIZE); in ef4_init_rx_buffers() 455 rx_buf->page, rx_buf->page_offset, in ef4_rx_packet_gro() 500 rx_buf->page_offset += hdr_len; in ef4_rx_mk_skb() 505 rx_buf->page, rx_buf->page_offset, in ef4_rx_mk_skb() 587 rx_buf->page_offset += efx->rx_prefix_size; in ef4_rx_packet()
|
| /linux/fs/jfs/ |
| H A D | jfs_metapage.c | 194 int page_offset = mp->data - folio_address(src); in __metapage_migrate_folio() local 196 mp->data = folio_address(dst) + page_offset; in __metapage_migrate_folio() 237 int page_offset; in __metapage_migrate_folio() local 251 page_offset = mp->data - folio_address(src); in __metapage_migrate_folio() 252 mp->data = folio_address(dst) + page_offset; in __metapage_migrate_folio() 693 unsigned long page_offset; in __get_metapage() local 701 page_offset = (lblock - (page_index << l2BlocksPerPage)) << l2bsize; in __get_metapage() 702 if ((page_offset + size) > PAGE_SIZE) { in __get_metapage() 737 mp = folio_to_mp(folio, page_offset); in __get_metapage() 770 mp->data = folio_address(folio) + page_offset; in __get_metapage()
|
| /linux/drivers/infiniband/hw/hfi1/ |
| H A D | pin_system.c | 289 unsigned int page_offset; in add_mapping_to_sdma_packet() local 310 page_offset = start - ALIGN_DOWN(start, PAGE_SIZE); in add_mapping_to_sdma_packet() 311 from_this_page = PAGE_SIZE - page_offset; in add_mapping_to_sdma_packet() 327 page_offset, from_this_page, in add_mapping_to_sdma_packet() 338 ret, page_index, page_offset, from_this_page); in add_mapping_to_sdma_packet()
|
| /linux/drivers/iommu/iommufd/ |
| H A D | pages.c | 210 return iopt_area_iova(area) - area->page_offset + index * PAGE_SIZE; in iopt_area_index_to_iova() 221 return iopt_area_iova(area) - area->page_offset + in iopt_area_index_to_iova_last() 412 unsigned int page_offset = 0; in batch_from_domain() local 418 page_offset = area->page_offset; in batch_from_domain() 425 phys = iommu_iova_to_phys(domain, iova) - page_offset; in batch_from_domain() 428 iova += PAGE_SIZE - page_offset; in batch_from_domain() 429 page_offset = 0; in batch_from_domain() 440 unsigned int page_offset = 0; in raw_pages_from_domain() local 446 page_offset = area->page_offset; in raw_pages_from_domain() 448 phys = iommu_iova_to_phys(domain, iova) - page_offset; in raw_pages_from_domain() [all …]
|
| H A D | io_pagetable.c | 78 unsigned long page_offset) in __alloc_iova_check_range() argument 86 aligned_start |= page_offset; in __alloc_iova_check_range() 97 unsigned long page_offset) in __alloc_iova_check_hole() argument 102 length, iova_alignment, page_offset); in __alloc_iova_check_hole() 108 unsigned long page_offset) in __alloc_iova_check_used() argument 113 length, iova_alignment, page_offset); in __alloc_iova_check_used() 123 unsigned long page_offset = addr % PAGE_SIZE; in iopt_alloc_iova() local 165 iova_alignment, page_offset)) in iopt_alloc_iova() 173 page_offset)) in iopt_alloc_iova() 220 area->page_offset = start_byte % PAGE_SIZE; in iopt_insert_area() [all …]
|
| /linux/drivers/net/ethernet/mellanox/mlx4/ |
| H A D | en_rx.c | 72 frags->page_offset = priv->rx_headroom; in mlx4_en_alloc_frags() 77 rx_desc->data[i].addr = cpu_to_be64(dma + frags->page_offset); in mlx4_en_alloc_frags() 457 dma_sync_single_range_for_cpu(priv->ddev, dma, frags->page_offset, in mlx4_en_complete_rx_desc() 460 __skb_fill_page_desc(skb, nr, page, frags->page_offset, in mlx4_en_complete_rx_desc() 467 frags->page_offset ^= PAGE_SIZE / 2; in mlx4_en_complete_rx_desc() 479 frags->page_offset += sz_align; in mlx4_en_complete_rx_desc() 480 release = frags->page_offset + frag_info->frag_size > PAGE_SIZE; in mlx4_en_complete_rx_desc() 726 va = page_address(frags[0].page) + frags[0].page_offset; in mlx4_en_process_rx_cq() 756 dma += frags[0].page_offset; in mlx4_en_process_rx_cq() 796 dma += frags[0].page_offset; in mlx4_en_process_rx_cq() [all …]
|
| /linux/tools/testing/selftests/powerpc/primitives/ |
| H A D | load_unaligned_zeropad.c | 102 static int do_one_test(char *p, int page_offset) in do_one_test() argument 114 …printf("offset %u load_unaligned_zeropad returned 0x%lx, should be 0x%lx\n", page_offset, got, sho… in do_one_test()
|
| /linux/drivers/net/ethernet/google/gve/ |
| H A D | gve_rx.c | 150 page_info->page_offset = 0; in gve_setup_rx_buffer() 228 rx->qpl_copy_pool[j].page_offset = 0; in gve_rx_prefill_pages() 454 u32 offset = page_info->page_offset + page_info->pad; in gve_rx_add_frags() 497 page_info->page_offset ^= GVE_DEFAULT_RX_BUFFER_OFFSET; in gve_rx_flip_buff() 542 void *src = page_info->page_address + page_info->page_offset; in gve_rx_copy_to_pool() 575 alloc_page_info.page_offset = 0; in gve_rx_copy_to_pool() 592 dst = copy_page_info->page_address + copy_page_info->page_offset; in gve_rx_copy_to_pool() 602 copy_page_info->page_offset ^= GVE_DEFAULT_RX_BUFFER_OFFSET; in gve_rx_copy_to_pool() 842 va = page_info->page_address + page_info->page_offset; in gve_rx() 850 be64_to_cpu(data_slot->addr) - page_info->page_offset : in gve_rx() [all …]
|
| /linux/drivers/net/ethernet/sfc/siena/ |
| H A D | rx.c | 93 rx_buf->page_offset += hdr_len; in efx_rx_mk_skb() 98 rx_buf->page, rx_buf->page_offset, in efx_rx_mk_skb() 177 rx_buf->page_offset += efx->rx_prefix_size; in efx_siena_rx_packet() 303 rx_buf->page_offset += offset; in efx_do_xdp()
|