/titanic_41/usr/src/uts/sun4/io/efcode/ |
H A D | fc_ddi.c | 61 struct ddi_dma_req dmareq; in fc_ddi_dma_buf_bind_handle() local 70 dmareq.dmar_flags = flags; in fc_ddi_dma_buf_bind_handle() 71 dmareq.dmar_fp = waitfp; in fc_ddi_dma_buf_bind_handle() 72 dmareq.dmar_arg = arg; in fc_ddi_dma_buf_bind_handle() 73 dmareq.dmar_object.dmao_size = (uint_t)bp->b_bcount; in fc_ddi_dma_buf_bind_handle() 76 dmareq.dmar_object.dmao_type = DMA_OTYP_PAGES; in fc_ddi_dma_buf_bind_handle() 77 dmareq.dmar_object.dmao_obj.pp_obj.pp_pp = bp->b_pages; in fc_ddi_dma_buf_bind_handle() 78 dmareq.dmar_object.dmao_obj.pp_obj.pp_offset = in fc_ddi_dma_buf_bind_handle() 81 dmareq.dmar_object.dmao_obj.virt_obj.v_addr = bp->b_un.b_addr; in fc_ddi_dma_buf_bind_handle() 83 dmareq.dmar_object.dmao_obj.virt_obj.v_priv = in fc_ddi_dma_buf_bind_handle() [all …]
|
/titanic_41/usr/src/uts/sun4/os/ |
H A D | dvma.c | 60 struct ddi_dma_req dmareq; in dvma_reserve() local 70 bzero(&dmareq, sizeof (dmareq)); in dvma_reserve() 71 dmareq.dmar_fp = DDI_DMA_DONTWAIT; in dvma_reserve() 72 dmareq.dmar_flags = DDI_DMA_RDWR | DDI_DMA_STREAMING; in dvma_reserve() 73 dmareq.dmar_limits = &dma_lim; in dvma_reserve() 74 dmareq.dmar_object.dmao_size = pages; in dvma_reserve() 84 ret = ddi_dma_mctl(dip, dip, reqhdl, DDI_DMA_RESERVE, (off_t *)&dmareq, in dvma_reserve()
|
/titanic_41/usr/src/uts/sun4u/io/pci/ |
H A D | pci_dma.c | 358 ddi_dma_req_t *dmareq) in pci_dma_lmts2hdl() argument 365 ddi_dma_lim_t *lim_p = dmareq->dmar_limits; in pci_dma_lmts2hdl() 376 if (!(mp = pci_dma_allocmp(dip, rdip, dmareq->dmar_fp, in pci_dma_lmts2hdl() 377 dmareq->dmar_arg))) in pci_dma_lmts2hdl() 550 pci_dma_type(pci_t *pci_p, ddi_dma_req_t *dmareq, ddi_dma_impl_t *mp) in pci_dma_type() argument 553 ddi_dma_obj_t *dobj_p = &dmareq->dmar_object; in pci_dma_type() 561 mp->dmai_rflags = pci_dma_consist_check(dmareq->dmar_flags, pbm_p); in pci_dma_type() 582 int (*waitfp)(caddr_t) = dmareq->dmar_fp; in pci_dma_type() 595 dmareq->dmar_arg, in pci_dma_type() 699 pci_dma_vapfn(pci_t *pci_p, ddi_dma_req_t *dmareq, ddi_dma_impl_t *mp, in pci_dma_vapfn() argument [all …]
|
H A D | pci_fdvma.c | 179 ddi_dma_req_t *dmareq, ddi_dma_handle_t *handlep) in pci_fdvma_reserve() argument 186 ddi_dma_lim_t *lim_p = dmareq->dmar_limits; in pci_fdvma_reserve() 206 npages = dmareq->dmar_object.dmao_size; in pci_fdvma_reserve() 227 dmareq->dmar_fp == DDI_DMA_SLEEP ? VM_SLEEP : VM_NOSLEEP)); in pci_fdvma_reserve() 250 pci_dma_consist_check(dmareq->dmar_flags, pci_p->pci_pbm_p); in pci_fdvma_reserve() 251 if (!(dmareq->dmar_flags & DDI_DMA_RDWR)) in pci_fdvma_reserve() 255 mp->dmai_minxfer = dmareq->dmar_limits->dlim_minxfer; in pci_fdvma_reserve() 256 mp->dmai_burstsizes = dmareq->dmar_limits->dlim_burstsizes; in pci_fdvma_reserve()
|
H A D | pci.c | 569 pci_dma_setup(dev_info_t *dip, dev_info_t *rdip, ddi_dma_req_t *dmareq, in pci_dma_setup() argument 581 if (!(mp = pci_dma_lmts2hdl(dip, rdip, iommu_p, dmareq))) in pci_dma_setup() 585 if (ret = pci_dma_type(pci_p, dmareq, mp)) in pci_dma_setup() 587 if (ret = pci_dma_pfn(pci_p, dmareq, mp)) in pci_dma_setup() 592 if ((ret = pci_dvma_win(pci_p, dmareq, mp)) || !handlep) in pci_dma_setup() 603 if (ret = pci_dvma_map(mp, dmareq, iommu_p)) in pci_dma_setup() 607 if ((ret = pci_dma_physwin(pci_p, dmareq, mp)) || !handlep) in pci_dma_setup() 693 ddi_dma_handle_t handle, ddi_dma_req_t *dmareq, in pci_dma_bindhdl() argument 702 ddi_driver_name(rdip), ddi_get_instance(rdip), mp, dmareq); in pci_dma_bindhdl() 710 if (ret = pci_dma_type(pci_p, dmareq, mp)) in pci_dma_bindhdl() [all …]
|
/titanic_41/usr/src/uts/sun4/io/px/ |
H A D | px_dma.c | 177 ddi_dma_req_t *dmareq) in px_dma_lmts2hdl() argument 184 ddi_dma_lim_t *lim_p = dmareq->dmar_limits; in px_dma_lmts2hdl() 195 if (!(mp = px_dma_allocmp(dip, rdip, dmareq->dmar_fp, in px_dma_lmts2hdl() 196 dmareq->dmar_arg))) in px_dma_lmts2hdl() 386 px_dma_type(px_t *px_p, ddi_dma_req_t *dmareq, ddi_dma_impl_t *mp) in px_dma_type() argument 389 ddi_dma_obj_t *dobj_p = &dmareq->dmar_object; in px_dma_type() 395 mp->dmai_rflags = dmareq->dmar_flags & DMP_DDIFLAGS; in px_dma_type() 550 px_dma_pfn(px_t *px_p, ddi_dma_req_t *dmareq, ddi_dma_impl_t *mp) in px_dma_pfn() argument 553 int (*waitfp)(caddr_t) = dmareq->dmar_fp; in px_dma_pfn() 576 ddi_set_callback(waitfp, dmareq->dmar_arg, in px_dma_pfn() [all …]
|
H A D | px_fdvma.c | 132 ddi_dma_req_t *dmareq, ddi_dma_handle_t *handlep) in px_fdvma_reserve() argument 139 ddi_dma_lim_t *lim_p = dmareq->dmar_limits; in px_fdvma_reserve() 159 npages = dmareq->dmar_object.dmao_size; in px_fdvma_reserve() 179 dmareq->dmar_fp == DDI_DMA_SLEEP ? VM_SLEEP : VM_NOSLEEP)); in px_fdvma_reserve() 200 mp->dmai_burstsizes = dmareq->dmar_limits->dlim_burstsizes; in px_fdvma_reserve()
|
H A D | px.c | 839 px_dma_setup(dev_info_t *dip, dev_info_t *rdip, ddi_dma_req_t *dmareq, in px_dma_setup() argument 851 if (!(mp = px_dma_lmts2hdl(dip, rdip, mmu_p, dmareq))) in px_dma_setup() 855 if (ret = px_dma_type(px_p, dmareq, mp)) in px_dma_setup() 857 if (ret = px_dma_pfn(px_p, dmareq, mp)) in px_dma_setup() 862 if ((ret = px_dvma_win(px_p, dmareq, mp)) || !handlep) in px_dma_setup() 873 if (ret = px_dvma_map(mp, dmareq, mmu_p)) in px_dma_setup() 877 if ((ret = px_dma_physwin(px_p, dmareq, mp)) || !handlep) in px_dma_setup() 963 ddi_dma_handle_t handle, ddi_dma_req_t *dmareq, in px_dma_bindhdl() argument 972 ddi_driver_name(rdip), ddi_get_instance(rdip), mp, dmareq); in px_dma_bindhdl() 980 if (ret = px_dma_type(px_p, dmareq, mp)) in px_dma_bindhdl() [all …]
|
H A D | px_var.h | 160 ddi_dma_req_t *dmareq, ddi_dma_handle_t *handlep); 166 ddi_dma_handle_t handle, ddi_dma_req_t *dmareq,
|
H A D | px_fdvma.h | 37 struct ddi_dma_req *dmareq, ddi_dma_handle_t *handlep);
|
H A D | px_dma.h | 213 px_mmu_t *mmu_p, ddi_dma_req_t *dmareq); 220 extern int px_dvma_map(ddi_dma_impl_t *mp, ddi_dma_req_t *dmareq, 223 extern int px_dma_physwin(px_t *px_p, ddi_dma_req_t *dmareq,
|
/titanic_41/usr/src/uts/sun4u/io/ |
H A D | iommu.c | 344 #define OBJSIZE dmareq->dmar_object.dmao_size 821 check_dma_attr(struct ddi_dma_req *dmareq, ddi_dma_attr_t *dma_attr, in check_dma_attr() argument 832 if ((dmareq->dmar_flags & DDI_DMA_PARTIAL) == 0) in check_dma_attr() 839 if (!((addrlow + dmareq->dmar_object.dmao_size == 0) && in check_dma_attr() 841 if ((dmareq->dmar_flags & DDI_DMA_PARTIAL) == 0) in check_dma_attr() 851 ddi_dma_handle_t handle, struct ddi_dma_req *dmareq, in iommu_dma_bindhdl() argument 876 size = (uint32_t)dmareq->dmar_object.dmao_size; in iommu_dma_bindhdl() 878 rval = check_dma_attr(dmareq, dma_attr, &size); in iommu_dma_bindhdl() 884 mp->dmai_rflags = (dmareq->dmar_flags & DMP_DDIFLAGS) | in iommu_dma_bindhdl() 887 switch (dmareq->dmar_object.dmao_type) { in iommu_dma_bindhdl() [all …]
|
/titanic_41/usr/src/uts/intel/sys/ |
H A D | iommulib.h | 70 struct ddi_dma_req *dmareq, ddi_dma_cookie_t *cookiep, 87 struct ddi_dma_req *dmareq, ddi_dma_obj_t *dmao); 141 ddi_dma_handle_t handle, struct ddi_dma_req *dmareq, 227 ddi_dma_handle_t dma_handle, struct ddi_dma_req *dmareq, 242 ddi_dma_handle_t dma_handle, struct ddi_dma_req *dmareq, 275 ddi_dma_handle_t handle, struct ddi_dma_req *dmareq,
|
/titanic_41/usr/src/uts/i86pc/io/amd_iommu/ |
H A D | amd_iommu_page_tables.c | 1000 struct ddi_dma_req *dmareq) in init_pte() argument 1019 if (R == 0 && ((dmareq->dmar_flags & DDI_DMA_WRITE) || in init_pte() 1020 (dmareq->dmar_flags & DDI_DMA_RDWR))) { in init_pte() 1023 if (W == 0 && ((dmareq->dmar_flags & DDI_DMA_READ) || in init_pte() 1024 (dmareq->dmar_flags & DDI_DMA_RDWR))) { in init_pte() 1042 if (dmareq->dmar_flags & DDI_DMA_RDWR) { in init_pte() 1046 if (dmareq->dmar_flags & DDI_DMA_WRITE) { in init_pte() 1049 if (dmareq->dmar_flags & DDI_DMA_READ) { in init_pte() 1092 struct ddi_dma_req *dmareq, in amd_iommu_setup_1_pgtable() argument 1146 error = init_pte(pt, pa, AMD_IOMMU_VA_BITS(va, level), dmareq); in amd_iommu_setup_1_pgtable() [all …]
|
H A D | amd_iommu_page_tables.h | 119 ddi_dma_attr_t *attrp, struct ddi_dma_req *dmareq,
|
H A D | amd_iommu_impl.c | 51 struct ddi_dma_req *dmareq, ddi_dma_cookie_t *cookiep, 64 struct ddi_dma_req *dmareq, ddi_dma_obj_t *dmao); 1495 struct ddi_dma_req *dmareq, ddi_dma_cookie_t *cookie_array, uint_t ccount, in map_current_window() argument 1523 if ((error = amd_iommu_map_pa2va(iommu, rdip, attrp, dmareq, in map_current_window() 1617 struct ddi_dma_req *dmareq, ddi_dma_cookie_t *cookiep, in amd_iommu_bindhdl() argument 1634 dmareq, cookiep, ccountp); in amd_iommu_bindhdl() 1669 error = map_current_window(iommu, rdip, attrp, dmareq, in amd_iommu_bindhdl() 1895 struct ddi_dma_req *dmareq, ddi_dma_obj_t *dmao) in amd_iommu_mapobject() argument
|
/titanic_41/usr/src/uts/i86pc/io/ |
H A D | rootnex.c | 204 ddi_dma_handle_t handle, struct ddi_dma_req *dmareq, 232 ddi_dma_handle_t handle, struct ddi_dma_req *dmareq, 382 static int rootnex_valid_bind_parms(ddi_dma_req_t *dmareq, 388 static int rootnex_bind_slowpath(ddi_dma_impl_t *hp, struct ddi_dma_req *dmareq, 390 static int rootnex_setup_copybuf(ddi_dma_impl_t *hp, struct ddi_dma_req *dmareq, 1961 ddi_dma_handle_t handle, struct ddi_dma_req *dmareq, in rootnex_coredma_bindhdl() argument 1984 if (dmareq->dmar_fp == DDI_DMA_SLEEP) { in rootnex_coredma_bindhdl() 1990 hp->dmai_rflags = dmareq->dmar_flags & DMP_DDIFLAGS; in rootnex_coredma_bindhdl() 2018 e = rootnex_valid_bind_parms(dmareq, attr); in rootnex_coredma_bindhdl() 2027 dma->dp_dma = dmareq->dmar_object; in rootnex_coredma_bindhdl() [all …]
|
H A D | immu_dvma.c | 108 struct ddi_dma_req *dmareq, ddi_dma_obj_t *dmao); 378 dma_to_immu_flags(struct ddi_dma_req *dmareq) in dma_to_immu_flags() argument 382 if (dmareq->dmar_fp == DDI_DMA_SLEEP) { in dma_to_immu_flags() 398 if (dmareq->dmar_flags & DDI_DMA_READ) in dma_to_immu_flags() 401 if (dmareq->dmar_flags & DDI_DMA_WRITE) in dma_to_immu_flags() 408 if ((dmareq->dmar_flags & (DDI_DMA_READ | DDI_DMA_WRITE)) == 0) { in dma_to_immu_flags() 2551 immu_hdl_priv_t *ihp, struct ddi_dma_req *dmareq, in immu_map_dvmaseg() argument 2572 immu_flags = dma_to_immu_flags(dmareq); in immu_map_dvmaseg() 2576 dmar_object = &dmareq->dmar_object; in immu_map_dvmaseg() 2637 dmareq->dmar_fp == DDI_DMA_SLEEP ? VM_SLEEP : VM_NOSLEEP); in immu_map_dvmaseg() [all …]
|
/titanic_41/usr/src/uts/sun4u/sys/pci/ |
H A D | pci_fdvma.h | 39 struct ddi_dma_req *dmareq, ddi_dma_handle_t *handlep);
|
H A D | pci_var.h | 220 ddi_dma_req_t *dmareq, ddi_dma_handle_t *handlep); 226 ddi_dma_handle_t handle, ddi_dma_req_t *dmareq,
|
H A D | pci_dma.h | 239 iommu_t *iommu_p, ddi_dma_req_t *dmareq); 247 extern int pci_dvma_map(ddi_dma_impl_t *mp, ddi_dma_req_t *dmareq, 252 extern int pci_dma_physwin(pci_t *pci_p, ddi_dma_req_t *dmareq,
|
/titanic_41/usr/src/uts/intel/io/ |
H A D | iommulib.c | 674 ddi_dma_handle_t dma_handle, struct ddi_dma_req *dmareq, in iommulib_nexdma_bindhdl() argument 684 dmareq, cookiep, ccountp)); in iommulib_nexdma_bindhdl() 733 ddi_dma_handle_t dma_handle, struct ddi_dma_req *dmareq, in iommulib_nexdma_mapobject() argument 740 dma_handle, dmareq, dmao)); in iommulib_nexdma_mapobject() 780 ddi_dma_handle_t handle, struct ddi_dma_req *dmareq, in iommulib_iommu_dma_bindhdl() argument 786 return (nexops->nops_dma_bindhdl(dip, rdip, handle, dmareq, in iommulib_iommu_dma_bindhdl()
|
/titanic_41/usr/src/uts/common/sys/ |
H A D | devops.h | 205 struct ddi_dma_req *dmareq, 213 ddi_dma_handle_t handle, struct ddi_dma_req *dmareq, 330 struct ddi_dma_req *dmareq,
|
/titanic_41/usr/src/uts/sun4v/io/niumx/ |
H A D | niumx.c | 68 ddi_dma_handle_t handle, ddi_dma_req_t *dmareq, 723 ddi_dma_handle_t handle, ddi_dma_req_t *dmareq, in niumx_dma_bindhdl() argument 726 int (*waitfp)(caddr_t) = dmareq->dmar_fp; in niumx_dma_bindhdl() 728 ddi_dma_obj_t *dobj_p = &dmareq->dmar_object; in niumx_dma_bindhdl() 734 NIUMX_NAMEINST(rdip), mp, dmareq); in niumx_dma_bindhdl() 737 mp->dmai_rflags = dmareq->dmar_flags & DMP_DDIFLAGS | DMP_NOSYNC; in niumx_dma_bindhdl()
|
/titanic_41/usr/src/uts/common/os/ |
H A D | sunddi.c | 794 ddi_dma_handle_t handle, struct ddi_dma_req *dmareq, in ddi_dma_bindhdl() argument 804 return ((*funcp)(dip, rdip, handle, dmareq, cp, ccountp)); in ddi_dma_bindhdl() 5237 ddi_dma_handle_t handle, struct ddi_dma_req *dmareq, in ddi_no_dma_bindhdl() argument 5240 _NOTE(ARGUNUSED(dip, rdip, handle, dmareq, cp, ccountp)) in ddi_no_dma_bindhdl() 7074 struct ddi_dma_req dmareq; in ddi_dma_buf_bind_handle() local 7077 dmareq.dmar_flags = flags; in ddi_dma_buf_bind_handle() 7078 dmareq.dmar_fp = waitfp; in ddi_dma_buf_bind_handle() 7079 dmareq.dmar_arg = arg; in ddi_dma_buf_bind_handle() 7080 dmareq.dmar_object.dmao_size = (uint_t)bp->b_bcount; in ddi_dma_buf_bind_handle() 7083 dmareq.dmar_object.dmao_type = DMA_OTYP_PAGES; in ddi_dma_buf_bind_handle() [all …]
|