| /linux/mm/ |
| H A D | userfaultfd.c | 496 unsigned long src_start, in mfill_atomic_hugetlb() argument 523 src_addr = src_start; in mfill_atomic_hugetlb() 567 while (src_addr < src_start + len) { in mfill_atomic_hugetlb() 655 unsigned long src_start, 706 unsigned long src_start, in mfill_atomic() argument 725 VM_WARN_ON_ONCE(src_start + len <= src_start); in mfill_atomic() 728 src_addr = src_start; in mfill_atomic() 774 src_start, len, flags); in mfill_atomic() 782 while (src_addr < src_start in mfill_atomic() 869 mfill_atomic_copy(struct userfaultfd_ctx * ctx,unsigned long dst_start,unsigned long src_start,unsigned long len,uffd_flags_t flags) mfill_atomic_copy() argument 1084 unsigned long src_start = src_addr; move_present_ptes() local 1579 find_vmas_mm_locked(struct mm_struct * mm,unsigned long dst_start,unsigned long src_start,struct vm_area_struct ** dst_vmap,struct vm_area_struct ** src_vmap) find_vmas_mm_locked() argument 1606 uffd_move_lock(struct mm_struct * mm,unsigned long dst_start,unsigned long src_start,struct vm_area_struct ** dst_vmap,struct vm_area_struct ** src_vmap) uffd_move_lock() argument 1682 uffd_move_lock(struct mm_struct * mm,unsigned long dst_start,unsigned long src_start,struct vm_area_struct ** dst_vmap,struct vm_area_struct ** src_vmap) uffd_move_lock() argument 1771 move_pages(struct userfaultfd_ctx * ctx,unsigned long dst_start,unsigned long src_start,unsigned long len,__u64 mode) move_pages() argument [all...] |
| /linux/drivers/dma/qcom/ |
| H A D | hidma_dbg.c | 20 dma_addr_t src_start; in hidma_ll_chstats() local 43 src_start = tre_local[HIDMA_TRE_SRC_LOW_IDX]; in hidma_ll_chstats() 44 src_start = ((u64) (tre_local[HIDMA_TRE_SRC_HI_IDX]) << 32) + src_start; in hidma_ll_chstats() 49 seq_printf(s, "src=%pap\n", &src_start); in hidma_ll_chstats()
|
| /linux/drivers/media/platform/ |
| H A D | m2m-deinterlace.c | 243 ctx->xt->src_start = p_in; in deinterlace_issue_dma() 250 ctx->xt->src_start = p_in + s_size / 2; in deinterlace_issue_dma() 257 ctx->xt->src_start = p_in + s_size; in deinterlace_issue_dma() 264 ctx->xt->src_start = p_in + (9 * s_size) / 8; in deinterlace_issue_dma() 271 ctx->xt->src_start = p_in + (5 * s_size) / 4; in deinterlace_issue_dma() 278 ctx->xt->src_start = p_in + (11 * s_size) / 8; in deinterlace_issue_dma() 285 ctx->xt->src_start = p_in; in deinterlace_issue_dma() 292 ctx->xt->src_start = p_in + s_size; in deinterlace_issue_dma() 299 ctx->xt->src_start = p_in + (5 * s_size) / 4; in deinterlace_issue_dma() 306 ctx->xt->src_start = p_in; in deinterlace_issue_dma() [all …]
|
| /linux/include/linux/ |
| H A D | omap-dma.h | 206 unsigned long src_start; /* source address : physical */ member 320 unsigned long src_start,
|
| H A D | userfaultfd_k.h | 120 unsigned long src_start, unsigned long len, 138 unsigned long src_start, unsigned long len, __u64 flags);
|
| H A D | dmaengine.h | 151 dma_addr_t src_start; member
|
| /linux/arch/arm64/kernel/ |
| H A D | hibernate.c | 188 static int create_safe_exec_page(void *src_start, size_t length, in create_safe_exec_page() argument 204 memcpy(page, src_start, length); in create_safe_exec_page()
|
| /linux/arch/xtensa/lib/ |
| H A D | memcopy.S | 333 # $a3:src != $a7:src_start 410 bne a3, a8, .LbackLoop1 # continue loop if a3:src != a8:src_start 491 bne a3, a10, .LbackLoop2 # continue loop if a3:src != a10:src_start
|
| /linux/arch/arm/mach-omap1/ |
| H A D | omap-dma.c | 176 unsigned long src_start, in omap_set_dma_src_params() argument 192 p->dma_write(src_start, CSSA, lch); in omap_set_dma_src_params()
|
| /linux/drivers/media/platform/ti/omap/ |
| H A D | omap_vout_vrfb.c | 256 xt->src_start = buf_phy_addr; in omap_vout_prepare_vrfb()
|
| /linux/drivers/net/ethernet/intel/ice/ |
| H A D | ice_parser.h | 117 u8 src_start; member
|
| H A D | ice_parser.c | 188 dev_info(dev, "\tsrc_start = %d\n", alu->src_start); in ice_imem_alu_dump() 341 alu->src_start = FIELD_GET(ICE_IM_ALU_SS, d64); in ice_imem_alu_init() 1122 dev_info(dev, "\tsrc_start = %d\n", alu->src_start); in ice_bst_alu_dump() 1223 alu->src_start = FIELD_GET(ICE_BST_ALU_SS, d64); in ice_bst_alu_init()
|
| H A D | ice_parser_rt.c | 499 alu->src_start, alu->src_len); in ice_alu_exe()
|
| /linux/drivers/dma/ |
| H A D | dma-axi-dmac.c | 757 if (!xt->src_inc || !axi_dmac_check_addr(chan, xt->src_start)) in axi_dmac_prep_interleaved() 790 desc->sg[0].hw->src_addr = xt->src_start; in axi_dmac_prep_interleaved()
|
| H A D | imx-dma.c | 945 imxdmac->channel, (unsigned long long)xt->src_start, in imxdma_prep_dma_interleaved() 960 desc->src = xt->src_start; in imxdma_prep_dma_interleaved()
|
| H A D | at_hdmac.c | 891 __func__, &xt->src_start, &xt->dst_start, xt->numf, in atc_prep_dma_interleaved() 915 dwidth = atc_get_xfer_width(xt->src_start, xt->dst_start, len); in atc_prep_dma_interleaved() 946 lli->saddr = xt->src_start; in atc_prep_dma_interleaved()
|
| /linux/drivers/dma/xilinx/ |
| H A D | xilinx_dpdma.c | 772 if (!IS_ALIGNED(xt->src_start, XILINX_DPDMA_ALIGN_BYTES)) { in xilinx_dpdma_chan_prep_interleaved_dma() 790 &xt->src_start, 1); in xilinx_dpdma_chan_prep_interleaved_dma()
|
| H A D | xilinx_dma.c | 2119 hw->buf_addr = lower_32_bits(xt->src_start); in xilinx_vdma_dma_prep_interleaved() 2120 hw->buf_addr_msb = upper_32_bits(xt->src_start); in xilinx_vdma_dma_prep_interleaved() 2122 hw->buf_addr = xt->src_start; in xilinx_vdma_dma_prep_interleaved()
|
| H A D | xdma.c | 756 src_addr = xt->src_start; in xdma_prep_interleaved_dma()
|
| /linux/drivers/gpu/drm/xlnx/ |
| H A D | zynqmp_disp.c | 1130 dma->xt.src_start = dma_addr; in zynqmp_disp_layer_update()
|
| /linux/drivers/dma/ti/ |
| H A D | omap-dma.c | 1308 data_type = __ffs((xt->src_start | xt->dst_start | xt->sgl[0].size)); in omap_dma_prep_dma_interleaved() 1315 d->dev_addr = xt->src_start; in omap_dma_prep_dma_interleaved()
|
| H A D | edma.c | 1278 param->src = xt->src_start; in edma_prep_dma_interleaved()
|
| /linux/drivers/dma/dw-edma/ |
| H A D | dw-edma-core.c | 427 src_addr = xfer->xfer.il->src_start; in dw_edma_device_transfer()
|