| /linux/drivers/gpu/drm/amd/amdgpu/ |
| H A D | amdgpu_gart.c | 313 u64 page_base; in amdgpu_gart_unbind() local 326 page_base = adev->dummy_page_addr; in amdgpu_gart_unbind() 332 t, page_base, flags); in amdgpu_gart_unbind() 333 page_base += AMDGPU_GPU_PAGE_SIZE; in amdgpu_gart_unbind() 358 uint64_t page_base; in amdgpu_gart_map() local 368 page_base = dma_addr[i]; in amdgpu_gart_map() 370 amdgpu_gmc_set_pte_pde(adev, dst, t, page_base, flags); in amdgpu_gart_map() 371 page_base += AMDGPU_GPU_PAGE_SIZE; in amdgpu_gart_map() 430 uint64_t page_base; in amdgpu_gart_map_gfx9_mqd() local 445 page_base = dma_addr[i]; in amdgpu_gart_map_gfx9_mqd() [all …]
|
| /linux/net/sunrpc/xprtrdma/ |
| H A D | rpc_rdma.c | 138 offset = offset_in_page(xdr->page_base); in rpcrdma_args_inline() 190 ppages = buf->pages + (buf->page_base >> PAGE_SHIFT); in rpcrdma_alloc_sparse_pages() 552 unsigned int page_base, len, remaining; in rpcrdma_prepare_pagelist() local 556 ppages = xdr->pages + (xdr->page_base >> PAGE_SHIFT); in rpcrdma_prepare_pagelist() 557 page_base = offset_in_page(xdr->page_base); in rpcrdma_prepare_pagelist() 561 len = min_t(unsigned int, PAGE_SIZE - page_base, remaining); in rpcrdma_prepare_pagelist() 563 page_base, len, DMA_TO_DEVICE); in rpcrdma_prepare_pagelist() 573 page_base = 0; in rpcrdma_prepare_pagelist() 589 unsigned int page_base, unsigned int len) in rpcrdma_prepare_tail_iov() argument 596 sge->addr = ib_dma_map_page(rdmab_device(rb), page, page_base, len, in rpcrdma_prepare_tail_iov() [all …]
|
| H A D | frwr_ops.c | 323 unsigned int page_base, remaining; in frwr_map() local 327 page_base = offset_in_page(xdrbuf->page_base + in frwr_map() 330 ((xdrbuf->page_base + cur->xc_page_offset) in frwr_map() 337 PAGE_SIZE - page_base, remaining); in frwr_map() 339 len, page_base); in frwr_map() 346 offset_in_page(page_base + len)) in frwr_map() 348 page_base = 0; in frwr_map()
|
| H A D | svc_rdma_sendto.c | 724 ppages = xdr->pages + (xdr->page_base >> PAGE_SHIFT); 725 pageoff = offset_in_page(xdr->page_base); 769 offset = offset_in_page(xdr->page_base); 838 ppages = xdr->pages + (xdr->page_base >> PAGE_SHIFT); in svc_rdma_map_reply_msg() 839 pageoff = offset_in_page(xdr->page_base); in svc_rdma_map_reply_msg()
|
| H A D | svc_rdma_rw.c | 462 page_off = info->wi_next_off + xdr->page_base; in svc_rdma_pagelist_to_bvec()
|
| /linux/net/sunrpc/ |
| H A D | xdr.c | 103 kaddr[buf->page_base + len] = '\0'; in xdr_terminate_string() 112 return (buf->page_base + buf->page_len + PAGE_SIZE - 1) >> PAGE_SHIFT; in xdr_buf_pagecount() 163 offset = offset_in_page(xdr->page_base); in xdr_buf_to_bvec() 211 xdr->page_base = base; in xdr_inline_pages() 482 pgbase += buf->page_base; in xdr_buf_pages_zero() 515 npages = (pagelen + buf->page_base + PAGE_SIZE - 1) >> PAGE_SHIFT; in xdr_buf_pages_fill_sparse() 524 if (pagelen > buf->page_base) in xdr_buf_pages_fill_sparse() 525 buflen += pagelen - buf->page_base; in xdr_buf_pages_fill_sparse() 599 buf->page_base + base + pglen, talen); in xdr_buf_pages_copy_right() 600 _shift_data_right_pages(buf->pages, buf->page_base + to, in xdr_buf_pages_copy_right() [all …]
|
| H A D | socklib.c | 67 struct page **ppage = xdr->pages + (xdr->page_base >> PAGE_SHIFT); in xdr_partial_copy_from_skb() 68 unsigned int poff = xdr->page_base & ~PAGE_MASK; in xdr_partial_copy_from_skb() 179 xdr->page_len + xdr->page_base); in xprt_send_pagedata() 180 return xprt_sendmsg(sock, msg, base + xdr->page_base); in xprt_send_pagedata()
|
| H A D | xprtsock.c | 346 n = (buf->page_base + want + PAGE_SIZE - 1) >> PAGE_SHIFT; in xs_alloc_sparse_pages() 353 return i > buf->page_base ? i - buf->page_base : 0; in xs_alloc_sparse_pages() 509 want + buf->page_base, in xs_read_xdr_buf() 510 seek + buf->page_base); in xs_read_xdr_buf() 513 xs_flush_bvec(buf->bvec, ret, seek + buf->page_base); in xs_read_xdr_buf() 514 ret -= buf->page_base; in xs_read_xdr_buf()
|
| H A D | svc_xprt.c | 557 rqstp->rq_res.page_base = 0; in svc_xprt_release() 708 arg->page_base = 0; in svc_thread_should_sleep()
|
| H A D | svcsock.c | 674 rqstp->rq_arg.page_base = 0; in svc_udp_recvfrom() 1169 rqstp->rq_arg.page_base = 0; in svc_tcp_recvfrom()
|
| H A D | svc.c | 1628 rqstp->rq_res.page_base = 0; in svc_process()
|
| /linux/arch/x86/mm/ |
| H A D | kmmio.c | 241 unsigned long page_base = addr; in kmmio_handler() local 246 page_base &= page_level_mask(l); in kmmio_handler() 258 faultpage = get_kmmio_fault_page(page_base); in kmmio_handler() 270 if (page_base == ctx->addr) { in kmmio_handler() 298 ctx->probe = get_kmmio_probe(page_base); in kmmio_handler() 300 ctx->addr = page_base; in kmmio_handler()
|
| /linux/drivers/gpu/drm/qxl/ |
| H A D | qxl_image.c | 168 unsigned int page_base, page_offset, out_offset; in qxl_image_init_helper() local 176 page_base = out_offset & PAGE_MASK; in qxl_image_init_helper() 180 ptr = qxl_bo_kmap_atomic_page(qdev, chunk_bo, page_base); in qxl_image_init_helper()
|
| /linux/net/sunrpc/auth_gss/ |
| H A D | gss_rpc_xdr.h | 125 unsigned int page_base; /* Start of page data */ member
|
| H A D | auth_gss.c | 1816 first = snd_buf->page_base >> PAGE_SHIFT; in alloc_enc_pages() 1817 last = (snd_buf->page_base + snd_buf->page_len - 1) >> PAGE_SHIFT; in alloc_enc_pages() 1863 first = snd_buf->page_base >> PAGE_SHIFT; in gss_wrap_req_priv() 1866 snd_buf->page_base -= first << PAGE_SHIFT; in gss_wrap_req_priv()
|
| H A D | gss_krb5_crypto.c | 255 int i = (page_pos + outbuf->page_base) >> PAGE_SHIFT; in encryptor()
|
| H A D | gss_rpc_xdr.c | 56 xdr_write_pages(xdr, in->pages, in->page_base, in->page_len); in gssx_enc_in_token()
|
| H A D | svcauth_gss.c | 1075 in_token->page_base = 0; in gss_read_proxy_verf() 1091 from_offs = rqstp->rq_arg.page_base; in gss_read_proxy_verf()
|
| /linux/drivers/scsi/ |
| H A D | hpsa.c | 6912 ulong page_base = ((ulong) base) & PAGE_MASK; in remap_pci_mem() local 6913 ulong page_offs = ((ulong) base) - page_base; in remap_pci_mem() 6914 void __iomem *page_remapped = ioremap(page_base, in remap_pci_mem()
|
| /linux/fs/nfsd/ |
| H A D | nfs4xdr.c | 4782 buf->page_base, maxcount); in nfsd4_encode_splice_read()
|