Home
last modified time | relevance | path

Searched refs:buf_info (Results 1 – 11 of 11) sorted by relevance

/linux/drivers/net/ethernet/pensando/ionic/
H A Dionic_txrx.c119 static void *ionic_rx_buf_va(struct ionic_buf_info *buf_info) in ionic_rx_buf_va() argument
121 return page_address(buf_info->page) + buf_info->page_offset; in ionic_rx_buf_va()
124 static dma_addr_t ionic_rx_buf_pa(struct ionic_buf_info *buf_info) in ionic_rx_buf_pa() argument
126 return page_pool_get_dma_addr(buf_info->page) + buf_info->page_offset; in ionic_rx_buf_pa()
130 struct ionic_buf_info *buf_info, in __ionic_rx_put_buf() argument
133 if (!buf_info->page) in __ionic_rx_put_buf()
136 page_pool_put_full_page(q->page_pool, buf_info->page, recycle_direct); in __ionic_rx_put_buf()
137 buf_info->page = NULL; in __ionic_rx_put_buf()
138 buf_info->len = 0; in __ionic_rx_put_buf()
139 buf_info->page_offset = 0; in __ionic_rx_put_buf()
[all …]
/linux/drivers/bus/mhi/ep/
H A Dring.c33 struct mhi_ep_buf_info buf_info = {}; in __mhi_ep_cache_ring() local
47 buf_info.size = (end - start) * sizeof(struct mhi_ring_element); in __mhi_ep_cache_ring()
48 buf_info.host_addr = ring->rbase + (start * sizeof(struct mhi_ring_element)); in __mhi_ep_cache_ring()
49 buf_info.dev_addr = &ring->ring_cache[start]; in __mhi_ep_cache_ring()
51 ret = mhi_cntrl->read_sync(mhi_cntrl, &buf_info); in __mhi_ep_cache_ring()
55 buf_info.size = (ring->ring_size - start) * sizeof(struct mhi_ring_element); in __mhi_ep_cache_ring()
56 buf_info.host_addr = ring->rbase + (start * sizeof(struct mhi_ring_element)); in __mhi_ep_cache_ring()
57 buf_info.dev_addr = &ring->ring_cache[start]; in __mhi_ep_cache_ring()
59 ret = mhi_cntrl->read_sync(mhi_cntrl, &buf_info); in __mhi_ep_cache_ring()
64 buf_info.host_addr = ring->rbase; in __mhi_ep_cache_ring()
[all …]
H A Dmain.c337 static void mhi_ep_read_completion(struct mhi_ep_buf_info *buf_info) in mhi_ep_read_completion() argument
339 struct mhi_ep_device *mhi_dev = buf_info->mhi_dev; in mhi_ep_read_completion()
348 result.buf_addr = buf_info->cb_buf; in mhi_ep_read_completion()
350 result.bytes_xferd = buf_info->size; in mhi_ep_read_completion()
360 if (buf_info->code != MHI_EV_CC_OVERFLOW) { in mhi_ep_read_completion()
398 kmem_cache_free(mhi_cntrl->tre_buf_cache, buf_info->cb_buf); in mhi_ep_read_completion()
407 struct mhi_ep_buf_info buf_info = {}; in mhi_ep_read_channel() local
440 buf_info.host_addr = mhi_chan->tre_loc + read_offset; in mhi_ep_read_channel()
441 buf_info.dev_addr = buf_addr; in mhi_ep_read_channel()
442 buf_info.size = tr_len; in mhi_ep_read_channel()
[all …]
/linux/drivers/net/caif/
H A Dcaif_virtio.c134 struct buf_info { struct
148 static void free_buf_info(struct cfv_info *cfv, struct buf_info *buf_info) in free_buf_info() argument
150 if (!buf_info) in free_buf_info()
152 gen_pool_free(cfv->genpool, (unsigned long) buf_info->vaddr, in free_buf_info()
153 buf_info->size); in free_buf_info()
154 kfree(buf_info); in free_buf_info()
169 struct buf_info *buf_info; in cfv_release_used_buf() local
173 buf_info = virtqueue_get_buf(vq_tx, &len); in cfv_release_used_buf()
177 if (!buf_info) in cfv_release_used_buf()
180 free_buf_info(cfv, buf_info); in cfv_release_used_buf()
[all …]
/linux/include/linux/
H A Dmhi_ep.h69 void (*cb)(struct mhi_ep_buf_info *buf_info);
163 int (*read_sync)(struct mhi_ep_cntrl *mhi_cntrl, struct mhi_ep_buf_info *buf_info);
164 int (*write_sync)(struct mhi_ep_cntrl *mhi_cntrl, struct mhi_ep_buf_info *buf_info);
165 int (*read_async)(struct mhi_ep_cntrl *mhi_cntrl, struct mhi_ep_buf_info *buf_info);
166 int (*write_async)(struct mhi_ep_cntrl *mhi_cntrl, struct mhi_ep_buf_info *buf_info);
/linux/drivers/acpi/
H A Dpfr_update.c469 struct pfru_com_buf_info buf_info; in pfru_write() local
476 ret = query_buffer(&buf_info, pfru_dev); in pfru_write()
480 if (len > buf_info.buf_size) { in pfru_write()
490 phy_addr = (phys_addr_t)((buf_info.addr_hi << 32) | buf_info.addr_lo); in pfru_write()
491 buf_ptr = memremap(phy_addr, buf_info.buf_size, MEMREMAP_WB); in pfru_write()
/linux/drivers/net/vmxnet3/
H A Dvmxnet3_drv.c405 BUG_ON(tq->buf_info[eop_idx].sop_idx != tq->tx_ring.next2comp); in vmxnet3_unmap_pkt()
408 tbi = &tq->buf_info[eop_idx]; in vmxnet3_unmap_pkt()
414 vmxnet3_unmap_tx_buf(tq->buf_info + tq->tx_ring.next2comp, in vmxnet3_unmap_pkt()
494 tbi = tq->buf_info + tq->tx_ring.next2comp; in vmxnet3_tq_cleanup()
513 BUG_ON(tq->buf_info[i].map_type != VMXNET3_MAP_NONE); in vmxnet3_tq_cleanup()
551 kfree(tq->buf_info); in vmxnet3_tq_destroy()
552 tq->buf_info = NULL; in vmxnet3_tq_destroy()
593 memset(tq->buf_info, 0, sizeof(tq->buf_info[0]) * tq->tx_ring.size); in vmxnet3_tq_init()
595 tq->buf_info[i].map_type = VMXNET3_MAP_NONE; in vmxnet3_tq_init()
606 tq->comp_ring.base || tq->buf_info); in vmxnet3_tq_create()
[all …]
H A Dvmxnet3_int.h261 struct vmxnet3_tx_buf_info *buf_info; member
338 struct vmxnet3_rx_buf_info *buf_info[2]; member
H A Dvmxnet3_xdp.c134 tbi = tq->buf_info + tq->tx_ring.next2fill; in vmxnet3_xdp_xmit_frame()
/linux/drivers/scsi/bfa/
H A Dbfad_bsg.c3269 struct bfad_buf_info *buf_base, *buf_info; in bfad_fcxp_map_sg() local
3283 buf_info = buf_base; in bfad_fcxp_map_sg()
3284 buf_info->size = payload_len; in bfad_fcxp_map_sg()
3285 buf_info->virt = dma_alloc_coherent(&bfad->pcidev->dev, in bfad_fcxp_map_sg()
3286 buf_info->size, &buf_info->phys, in bfad_fcxp_map_sg()
3288 if (!buf_info->virt) in bfad_fcxp_map_sg()
3292 memcpy(buf_info->virt, payload_kbuf, buf_info->size); in bfad_fcxp_map_sg()
3297 sg_table->sg_len = buf_info->size; in bfad_fcxp_map_sg()
3298 sg_table->sg_addr = (void *)(size_t) buf_info->phys; in bfad_fcxp_map_sg()
3314 struct bfad_buf_info *buf_info = buf_base; in bfad_fcxp_free_mem() local
[all …]
/linux/kernel/bpf/
H A Dverifier.c6784 const char *buf_info, in __check_buffer_access() argument
6791 regno, buf_info, off, size); in __check_buffer_access()
6829 const char *buf_info = type_is_rdonly_mem(reg->type) ? "rdonly" : "rdwr"; in check_buffer_access() local
6832 err = __check_buffer_access(env, buf_info, reg, regno, off, size); in check_buffer_access()