| /linux/drivers/crypto/ |
| H A D | omap-crypto.c | 41 sg_set_page(tmp, sg_page(*sg), len, (*sg)->offset); in omap_crypto_copy_sg_lists() 105 if (page_zonenum(sg_page(sg)) != ZONE_DMA) in omap_crypto_check_sg() 181 srcb = kmap_atomic(sg_page(src)) + srco + src->offset; in omap_crypto_copy_data() 182 dstb = kmap_atomic(sg_page(dst)) + dsto + dst->offset; in omap_crypto_copy_data() 186 flush_dcache_page(sg_page(dst)); in omap_crypto_copy_data()
|
| H A D | hifn_795x.c | 1314 spage = sg_page(src); in hifn_setup_dma() 1328 BUG_ON(!sg_page(t)); in hifn_setup_dma() 1329 dpage = sg_page(t); in hifn_setup_dma() 1333 BUG_ON(!sg_page(dst)); in hifn_setup_dma() 1334 dpage = sg_page(dst); in hifn_setup_dma() 1384 __free_page(sg_page(s)); in hifn_cipher_walk_exit() 1611 daddr = kmap_atomic(sg_page(dst)); in skcipher_get() 1665 __func__, sg_page(t), t->length, in hifn_process_ready() 1666 sg_page(dst), dst->length, nbytes); in hifn_process_ready() 1674 saddr = kmap_atomic(sg_page(t)); in hifn_process_ready()
|
| /linux/crypto/ |
| H A D | scatterwalk.c | 115 struct page *src_page = sg_page(src); in memcpy_sglist() 116 struct page *dst_page = sg_page(dst); in memcpy_sglist() 199 sg_set_page(dst, sg_page(src), src->length - len, src->offset + len); in scatterwalk_ffwd()
|
| H A D | af_alg.c | 551 unpin_user_page(sg_page(&sgl->sgt.sgl[i])); in af_alg_free_sg() 718 struct page *page = sg_page(sg + i); in af_alg_pull_tsgl() 791 if (!sg_page(sg)) in af_alg_free_areq_sgls() 793 put_page(sg_page(sg)); in af_alg_free_areq_sgls() 1012 err = memcpy_from_msg(page_address(sg_page(sg)) + in af_alg_sendmsg() 1064 get_page(sg_page(&sg[sgl->cur])); in af_alg_sendmsg() 1085 page_address(sg_page(sg + i)), in af_alg_sendmsg() 1088 __free_page(sg_page(sg + i)); in af_alg_sendmsg()
|
| /linux/include/crypto/ |
| H A D | scatterwalk.h | 94 sg_set_page(sg_out, sg_page(walk->sg), in scatterwalk_get_sglist() 102 struct page *base_page = sg_page(walk->sg); in scatterwalk_map() 219 __scatterwalk_flush_dcache_pages(sg_page(walk->sg), in scatterwalk_done_dst()
|
| /linux/net/rds/ |
| H A D | message.c | 158 __free_page(sg_page(&rm->data.op_sg[i])); in rds_message_purge() 160 put_page(sg_page(&rm->data.op_sg[i])); in rds_message_purge() 440 put_page(sg_page(&rm->data.op_sg[i])); in rds_message_zcopy_from_user() 477 if (!sg_page(sg)) { in rds_message_copy_from_user() 490 nbytes = copy_page_from_iter(sg_page(sg), sg->offset + sg_off, in rds_message_copy_from_user() 527 ret = copy_page_to_iter(sg_page(sg), sg->offset + vec_off, in rds_message_inc_copy_to_user()
|
| H A D | page.c | 106 get_page(sg_page(scat)); in rds_page_remainder_alloc() 150 ret ? NULL : sg_page(scat), ret ? 0 : scat->offset, in rds_page_remainder_alloc()
|
| H A D | tcp_send.c | 120 bvec_set_page(&bvec, sg_page(&rm->data.op_sg[sg]), in rds_tcp_xmit() 126 rdsdebug("tcp sendpage %p:%u:%u ret %d\n", (void *)sg_page(&rm->data.op_sg[sg]), in rds_tcp_xmit()
|
| /linux/drivers/target/tcm_fc/ |
| H A D | tfc_io.c | 84 page = sg_page(sg); in ft_queue_data_in() 102 page = sg_page(sg); in ft_queue_data_in() 279 page = sg_page(sg); in ft_recv_write_data() 287 page = sg_page(sg); in ft_recv_write_data()
|
| /linux/samples/kfifo/ |
| H A D | dma-example.c | 79 i, sg_page(&sg[i]), sg[i].offset, sg[i].length); in example_init() 108 i, sg_page(&sg[i]), sg[i].offset, sg[i].length); in example_init()
|
| /linux/drivers/target/ |
| H A D | target_core_sbc.c | 409 unsigned char *addr = kmap_atomic(sg_page(sg)); in compare_and_write_do_cmp() 1224 paddr = kmap_atomic(sg_page(psg)) + psg->offset; in sbc_dif_generate() 1225 daddr = kmap_atomic(sg_page(dsg)) + dsg->offset; in sbc_dif_generate() 1240 daddr = kmap_atomic(sg_page(dsg)) + dsg->offset; in sbc_dif_generate() 1253 daddr = kmap_atomic(sg_page(dsg)) + dsg->offset; in sbc_dif_generate() 1338 paddr = kmap_atomic(sg_page(psg)) + psg->offset; in sbc_dif_copy_prot() 1342 addr = kmap_atomic(sg_page(sg)) + sg->offset + offset; in sbc_dif_copy_prot() 1381 paddr = kmap_atomic(sg_page(psg)) + psg->offset; in sbc_dif_verify() 1382 daddr = kmap_atomic(sg_page(dsg)) + dsg->offset; in sbc_dif_verify() 1398 daddr = kmap_atomic(sg_page(dsg)) + dsg->offset; in sbc_dif_verify() [all …]
|
| /linux/tools/virtio/linux/ |
| H A D | scatterlist.h | 67 static inline struct page *sg_page(struct scatterlist *sg) in sg_page() function 158 return page_to_phys(sg_page(sg)) + sg->offset; in sg_phys()
|
| /linux/drivers/gpu/drm/i915/gem/ |
| H A D | i915_gem_phys.c | 102 void *vaddr = sg_page(pages->sgl); in i915_gem_object_put_pages_phys() 142 void *vaddr = sg_page(obj->mm.pages->sgl) + args->offset; in i915_gem_object_pwrite_phys() 173 void *vaddr = sg_page(obj->mm.pages->sgl) + args->offset; in i915_gem_object_pread_phys()
|
| H A D | i915_gem_internal.c | 24 if (sg_page(sg)) in internal_free_pages() 25 __free_pages(sg_page(sg), get_order(sg->length)); in internal_free_pages()
|
| H A D | i915_gem_pages.c | 306 if (n_pages == 1 && !PageHighMem(sg_page(obj->mm.pages->sgl))) in i915_gem_object_map_page() 307 return page_address(sg_page(obj->mm.pages->sgl)); in i915_gem_object_map_page() 770 return sg_page(sg) + offset; in __i915_gem_object_get_page()
|
| /linux/drivers/crypto/qce/ |
| H A D | dma.c | 63 if (!sg_page(sg)) in qce_sgtable_add() 73 sg_set_page(sg, sg_page(new_sgl), new_len, new_sgl->offset); in qce_sgtable_add()
|
| /linux/drivers/dma-buf/heaps/ |
| H A D | system_heap.c | 66 sg_set_page(new_sg, sg_page(sg), sg->length, sg->offset); in dup_sg_table() 203 struct page *page = sg_page(sg) + pgoff; in system_heap_mmap() 297 struct page *page = sg_page(sg); in system_heap_dma_buf_release() 414 struct page *p = sg_page(sg); in system_heap_allocate()
|
| /linux/net/sunrpc/auth_gss/ |
| H A D | gss_krb5_crypto.c | 258 in_page = sg_page(sg); in encryptor() 262 sg_set_page(&desc->outfrags[desc->fragno], sg_page(sg), sg->length, in encryptor() 288 sg_set_page(&desc->outfrags[0], sg_page(sg), fraglen, in encryptor() 321 sg_set_page(&desc->frags[desc->fragno], sg_page(sg), sg->length, in decryptor() 344 sg_set_page(&desc->frags[0], sg_page(sg), fraglen, in decryptor()
|
| /linux/drivers/infiniband/hw/mthca/ |
| H A D | mthca_memfree.c | 73 __free_pages(sg_page(&chunk->mem[i]), in mthca_free_icm_pages() 83 lowmem_page_address(sg_page(&chunk->mem[i])), in mthca_free_icm_coherent() 309 page = sg_page(&chunk->mem[i]); in mthca_table_find() 495 unpin_user_page(sg_page(&db_tab->page[i].mem)); in mthca_map_user_db() 562 unpin_user_page(sg_page(&db_tab->page[i].mem)); in mthca_cleanup_user_db_tab()
|
| /linux/drivers/gpu/drm/armada/ |
| H A D | armada_gem.c | 444 if (sg_page(sg)) in armada_gem_prime_map_dma_buf() 445 put_page(sg_page(sg)); in armada_gem_prime_map_dma_buf() 467 put_page(sg_page(sg)); in armada_gem_prime_unmap_dma_buf()
|
| /linux/drivers/scsi/ |
| H A D | libiscsi_tcp.c | 131 if (!recv && sendpage_ok(sg_page(sg))) in iscsi_tcp_segment_map() 136 segment->sg_mapped = kmap_atomic(sg_page(sg)); in iscsi_tcp_segment_map() 140 segment->sg_mapped = kmap(sg_page(sg)); in iscsi_tcp_segment_map() 152 kunmap(sg_page(segment->sg)); in iscsi_tcp_segment_unmap() 207 data = kmap_local_page(sg_page(segment->sg)); in iscsi_tcp_segment_done()
|
| /linux/drivers/crypto/ccp/ |
| H A D | ccp-crypto-main.c | 308 if (!sg_page(sg)) in ccp_crypto_sg_table_add() 314 sg_set_page(sg, sg_page(sg_add), sg_add->length, in ccp_crypto_sg_table_add()
|
| /linux/net/xfrm/ |
| H A D | xfrm_ipcomp.c | 75 page = sg_page(dsg); in ipcomp_post_acomp() 88 __free_page(sg_page(dsg)); in ipcomp_post_acomp()
|
| /linux/net/core/ |
| H A D | skmsg.c | 57 sg_page(sge) == pfrag->page && in sk_msg_alloc() 114 (sg_page(sge) == sg_page(sgd)) && in sk_msg_clone() 120 sk_msg_page_add(dst, sg_page(sge), sge_len, sge_off); in sk_msg_clone() 187 put_page(sg_page(sge)); in sk_msg_free_elem() 439 page = sg_page(sge); in __sk_msg_recvmsg()
|
| /linux/include/linux/ |
| H A D | devcoredump.h | 35 page = sg_page(iter); in _devcd_free_sgtable()
|