| /linux/sound/core/oss/ |
| H A D | io.c | 35 void **bufs = (void**)plugin->extra_data; in io_playback_transfer() local 36 if (snd_BUG_ON(!bufs)) in io_playback_transfer() 40 bufs[channel] = src_channels[channel].area.addr; in io_playback_transfer() 42 bufs[channel] = NULL; in io_playback_transfer() 44 return pcm_writev(plugin->plug, bufs, frames); in io_playback_transfer() 61 void **bufs = (void**)plugin->extra_data; in io_capture_transfer() local 62 if (snd_BUG_ON(!bufs)) in io_capture_transfer() 66 bufs[channel] = dst_channels[channel].area.addr; in io_capture_transfer() 68 bufs[channel] = NULL; in io_capture_transfer() 70 return pcm_readv(plugin->plug, bufs, frames); in io_capture_transfer()
|
| /linux/drivers/net/ethernet/cisco/enic/ |
| H A D | vnic_wq.c | 25 wq->bufs[i] = kzalloc(VNIC_WQ_BUF_BLK_SZ(count), GFP_KERNEL); in vnic_wq_alloc_bufs() 26 if (!wq->bufs[i]) in vnic_wq_alloc_bufs() 31 buf = wq->bufs[i]; in vnic_wq_alloc_bufs() 37 buf->next = wq->bufs[0]; in vnic_wq_alloc_bufs() 41 buf->next = wq->bufs[i + 1]; in vnic_wq_alloc_bufs() 51 wq->to_use = wq->to_clean = wq->bufs[0]; in vnic_wq_alloc_bufs() 66 if (wq->bufs[i]) { in vnic_wq_free() 67 kfree(wq->bufs[i]); in vnic_wq_free() 68 wq->bufs[i] = NULL; in vnic_wq_free() 140 &wq->bufs[fetch_index / VNIC_WQ_BUF_BLK_ENTRIES(count)] in enic_wq_init_start() [all …]
|
| H A D | vnic_rq.c | 25 rq->bufs[i] = kzalloc(VNIC_RQ_BUF_BLK_SZ(count), GFP_KERNEL); in vnic_rq_alloc_bufs() 26 if (!rq->bufs[i]) in vnic_rq_alloc_bufs() 31 buf = rq->bufs[i]; in vnic_rq_alloc_bufs() 37 buf->next = rq->bufs[0]; in vnic_rq_alloc_bufs() 40 buf->next = rq->bufs[i + 1]; in vnic_rq_alloc_bufs() 48 rq->to_use = rq->to_clean = rq->bufs[0]; in vnic_rq_alloc_bufs() 63 if (rq->bufs[i]) { in vnic_rq_free() 64 kfree(rq->bufs[i]); in vnic_rq_free() 65 rq->bufs[i] = NULL; in vnic_rq_free() 121 &rq->bufs[fetch_index / VNIC_RQ_BUF_BLK_ENTRIES(count)] in vnic_rq_init_start() [all …]
|
| /linux/drivers/scsi/fnic/ |
| H A D | vnic_rq.c | 22 rq->bufs[i] = kzalloc(VNIC_RQ_BUF_BLK_SZ, GFP_ATOMIC); in vnic_rq_alloc_bufs() 23 if (!rq->bufs[i]) { in vnic_rq_alloc_bufs() 30 buf = rq->bufs[i]; in vnic_rq_alloc_bufs() 36 buf->next = rq->bufs[0]; in vnic_rq_alloc_bufs() 39 buf->next = rq->bufs[i + 1]; in vnic_rq_alloc_bufs() 47 rq->to_use = rq->to_clean = rq->bufs[0]; in vnic_rq_alloc_bufs() 63 kfree(rq->bufs[i]); in vnic_rq_free() 64 rq->bufs[i] = NULL; in vnic_rq_free() 118 &rq->bufs[fetch_index / VNIC_RQ_BUF_BLK_ENTRIES] in vnic_rq_init() 174 &rq->bufs[fetch_index / VNIC_RQ_BUF_BLK_ENTRIES] in vnic_rq_clean()
|
| H A D | vnic_wq.c | 42 wq->bufs[i] = kzalloc(VNIC_WQ_BUF_BLK_SZ, GFP_ATOMIC); in vnic_wq_alloc_bufs() 43 if (!wq->bufs[i]) { in vnic_wq_alloc_bufs() 50 buf = wq->bufs[i]; in vnic_wq_alloc_bufs() 56 buf->next = wq->bufs[0]; in vnic_wq_alloc_bufs() 59 buf->next = wq->bufs[i + 1]; in vnic_wq_alloc_bufs() 67 wq->to_use = wq->to_clean = wq->bufs[0]; in vnic_wq_alloc_bufs() 82 kfree(wq->bufs[i]); in vnic_wq_free() 83 wq->bufs[i] = NULL; in vnic_wq_free() 160 &wq->bufs[fetch_index / VNIC_WQ_BUF_BLK_ENTRIES] in vnic_wq_init_start() 227 wq->to_use = wq->to_clean = wq->bufs[0]; in vnic_wq_clean()
|
| /linux/drivers/scsi/snic/ |
| H A D | vnic_wq.c | 36 wq->bufs[i] = kzalloc(VNIC_WQ_BUF_BLK_SZ, GFP_ATOMIC); in vnic_wq_alloc_bufs() 37 if (!wq->bufs[i]) { in vnic_wq_alloc_bufs() 45 buf = wq->bufs[i]; in vnic_wq_alloc_bufs() 51 buf->next = wq->bufs[0]; in vnic_wq_alloc_bufs() 54 buf->next = wq->bufs[i + 1]; in vnic_wq_alloc_bufs() 62 wq->to_use = wq->to_clean = wq->bufs[0]; in vnic_wq_alloc_bufs() 77 kfree(wq->bufs[i]); in svnic_wq_free() 78 wq->bufs[i] = NULL; in svnic_wq_free() 159 &wq->bufs[fetch_index / VNIC_WQ_BUF_BLK_ENTRIES(count)] in vnic_wq_init_start() 216 wq->to_use = wq->to_clean = wq->bufs[0]; in svnic_wq_clean()
|
| /linux/tools/virtio/ringtest/ |
| H A D | main.c | 111 int bufs = runcycles; in run_guest() local 123 if (started < bufs && in run_guest() 141 if (__builtin_expect(completed == bufs, false)) in run_guest() 148 assert(completed <= bufs); in run_guest() 149 assert(started <= bufs); in run_guest() 170 int bufs = runcycles; in run_host() local 188 if (__builtin_expect(completed == bufs, false)) in run_host() 193 assert(completed <= bufs); in run_host() 194 if (completed == bufs) in run_host()
|
| /linux/arch/riscv/kernel/ |
| H A D | unaligned_access_speed.c | 142 struct page **bufs = kzalloc_objs(*bufs, cpu_count); in check_unaligned_access_speed_all_cpus() local 144 if (!bufs) { in check_unaligned_access_speed_all_cpus() 154 bufs[cpu] = alloc_pages(GFP_KERNEL, MISALIGNED_BUFFER_ORDER); in check_unaligned_access_speed_all_cpus() 155 if (!bufs[cpu]) { in check_unaligned_access_speed_all_cpus() 162 on_each_cpu(check_unaligned_access_nonboot_cpu, bufs, 1); in check_unaligned_access_speed_all_cpus() 165 smp_call_on_cpu(0, check_unaligned_access, bufs[0], true); in check_unaligned_access_speed_all_cpus() 169 if (bufs[cpu]) in check_unaligned_access_speed_all_cpus() 170 __free_pages(bufs[cpu], MISALIGNED_BUFFER_ORDER); in check_unaligned_access_speed_all_cpus() 173 kfree(bufs); in check_unaligned_access_speed_all_cpus()
|
| /linux/drivers/scsi/arm/ |
| H A D | arm_scsi.h | 35 int bufs = SCp->buffers_residual; in copy_SCp_to_sg() local 40 BUG_ON(bufs + 1 > max); in copy_SCp_to_sg() 44 if (bufs) { in copy_SCp_to_sg() 48 for_each_sg(sg_next(SCp->buffer), src_sg, bufs, i) in copy_SCp_to_sg() 53 return bufs + 1; in copy_SCp_to_sg()
|
| /linux/drivers/net/ethernet/marvell/prestera/ |
| H A D | prestera_rxtx.c | 81 struct prestera_sdma_buf *bufs; member 86 struct prestera_sdma_buf *bufs; member 297 buf = &ring->bufs[buf_idx]; in prestera_sdma_rx_poll() 345 if (!ring->bufs) in prestera_sdma_rx_fini() 349 struct prestera_sdma_buf *buf = &ring->bufs[b]; in prestera_sdma_rx_fini() 382 ring->bufs = kmalloc_objs(*head, bnum); in prestera_sdma_rx_init() 383 if (!ring->bufs) in prestera_sdma_rx_init() 388 tail = &ring->bufs[bnum - 1]; in prestera_sdma_rx_init() 389 head = &ring->bufs[0]; in prestera_sdma_rx_init() 503 struct prestera_sdma_buf *buf = &tx_ring->bufs[b]; in prestera_sdma_tx_recycle_work_fn() [all …]
|
| /linux/tools/virtio/ |
| H A D | vhost_net_test.c | 276 bool delayed, int bufs) in run_tx_test() argument 289 while (vq->started < bufs && in run_tx_test() 306 if (vq->started >= bufs) in run_tx_test() 325 assert(vq->completed <= bufs); in run_tx_test() 326 assert(vq->started <= bufs); in run_tx_test() 327 if (vq->completed == bufs) in run_tx_test() 343 bool delayed, int bufs) in run_rx_test() argument 355 while (vq->started < bufs && in run_rx_test() 375 if (vq->started >= bufs) in run_rx_test() 400 assert(vq->completed <= bufs); in run_rx_test() [all …]
|
| H A D | virtio_test.c | 170 bool delayed, int batch, int reset_n, int bufs) in run_test() argument 195 while (started < bufs && in run_test() 218 if (started >= bufs) in run_test() 258 assert(completed <= bufs); in run_test() 259 assert(started <= bufs); in run_test() 260 if (completed == bufs) in run_test()
|
| /linux/drivers/soc/fsl/qbman/ |
| H A D | bman.c | 113 struct bm_buffer bufs[8]; member 151 struct bm_buffer bufs[8]; member 738 int bman_release(struct bman_pool *pool, const struct bm_buffer *bufs, u8 num) in bman_release() argument 772 bm_buffer_set64(r->bufs, bm_buffer_get64(bufs)); in bman_release() 773 bm_buffer_set_bpid(r->bufs, pool->bpid); in bman_release() 775 memcpy(&r->bufs[1], &bufs[1], i * sizeof(bufs[0])); in bman_release() 786 int bman_acquire(struct bman_pool *pool, struct bm_buffer *bufs, u8 num) in bman_acquire() argument 805 if (bufs) in bman_acquire() 806 memcpy(&bufs[0], &mcr->bufs[0], num * sizeof(bufs[0])); in bman_acquire()
|
| /linux/drivers/net/ethernet/fungible/funeth/ |
| H A D | funeth_rx.c | 55 c->bufs[c->prod_cnt & c->mask] = *buf; in cache_offer() 75 buf = &c->bufs[c->cons_cnt & c->mask]; in cache_get() 281 return &q->bufs[q->rq_cons & q->rq_mask]; in get_buf() 538 struct funeth_rxbuf *b = q->bufs; in fun_rxq_free_bufs() 551 struct funeth_rxbuf *b = q->bufs; in fun_rxq_alloc_bufs() 561 q->cur_buf = q->bufs; in fun_rxq_alloc_bufs() 570 c->bufs = kvzalloc_node(depth * sizeof(*c->bufs), GFP_KERNEL, node); in fun_rxq_init_cache() 571 return c->bufs ? 0 : -ENOMEM; in fun_rxq_init_cache() 577 struct funeth_rxbuf *b = q->cache.bufs; in fun_rxq_free_cache() 583 kvfree(q->cache.bufs); in fun_rxq_free_cache() [all …]
|
| /linux/tools/testing/selftests/powerpc/papr_vpd/ |
| H A D | papr_vpd.c | 198 char *bufs[2]; in papr_vpd_reread() local 200 for (size_t i = 0; i < ARRAY_SIZE(bufs); ++i) { in papr_vpd_reread() 201 bufs[i] = malloc(size); in papr_vpd_reread() 202 FAIL_IF(!bufs[i]); in papr_vpd_reread() 203 ssize_t consumed = pread(fd, bufs[i], size, 0); in papr_vpd_reread() 207 FAIL_IF(memcmp(bufs[0], bufs[1], size)); in papr_vpd_reread()
|
| /linux/drivers/net/ethernet/mellanox/mlx5/core/fpga/ |
| H A D | conn.c | 122 conn->qp.rq.bufs[ix] = buf; in mlx5_fpga_conn_post_recv() 171 conn->qp.sq.bufs[ix] = buf; in mlx5_fpga_conn_post_send() 257 buf = conn->qp.rq.bufs[ix]; in mlx5_fpga_conn_rq_cqe() 258 conn->qp.rq.bufs[ix] = NULL; in mlx5_fpga_conn_rq_cqe() 300 buf = conn->qp.sq.bufs[ix]; in mlx5_fpga_conn_sq_cqe() 301 conn->qp.sq.bufs[ix] = NULL; in mlx5_fpga_conn_sq_cqe() 534 conn->qp.rq.bufs = kvzalloc_objs(conn->qp.rq.bufs[0], conn->qp.rq.size); in mlx5_fpga_conn_create_qp() 535 if (!conn->qp.rq.bufs) { in mlx5_fpga_conn_create_qp() 540 conn->qp.sq.bufs = kvzalloc_objs(conn->qp.sq.bufs[0], conn->qp.sq.size); in mlx5_fpga_conn_create_qp() 541 if (!conn->qp.sq.bufs) { in mlx5_fpga_conn_create_qp() [all …]
|
| H A D | conn.h | 74 struct mlx5_fpga_dma_buf **bufs; member 81 struct mlx5_fpga_dma_buf **bufs; member
|
| /linux/arch/mips/include/asm/octeon/ |
| H A D | cvmx-wqe.h | 62 uint64_t bufs:8; member 193 uint64_t bufs:8; 198 uint64_t bufs:8; member 240 uint64_t bufs:8; 268 uint64_t bufs:8; member 399 uint64_t bufs:8;
|
| /linux/fs/ |
| H A D | pipe.c | 817 pipe->bufs = kzalloc_objs(struct pipe_buffer, pipe_bufs, in alloc_pipe_info() 820 if (pipe->bufs) { in alloc_pipe_info() 853 struct pipe_buffer *buf = pipe->bufs + i; in free_pipe_info() 865 kfree(pipe->bufs); in free_pipe_info() 1293 struct pipe_buffer *bufs; in pipe_resize_ring() local 1300 bufs = kzalloc_objs(*bufs, nr_slots, GFP_KERNEL_ACCOUNT | __GFP_NOWARN); in pipe_resize_ring() 1301 if (unlikely(!bufs)) in pipe_resize_ring() 1312 kfree(bufs); in pipe_resize_ring() 1324 memcpy(bufs, pipe->bufs + t, in pipe_resize_ring() 1329 memcpy(bufs + tsize, pipe->bufs, in pipe_resize_ring() [all …]
|
| /linux/drivers/infiniband/hw/hns/ |
| H A D | hns_roce_alloc.c | 132 int hns_roce_get_kmem_bufs(struct hns_roce_dev *hr_dev, dma_addr_t *bufs, in hns_roce_get_kmem_bufs() argument 149 bufs[total++] = hns_roce_buf_dma_addr(buf, offset); in hns_roce_get_kmem_bufs() 156 int hns_roce_get_umem_bufs(dma_addr_t *bufs, int buf_cnt, struct ib_umem *umem, in hns_roce_get_umem_bufs() argument 164 bufs[total++] = rdma_block_iter_dma_address(&biter); in hns_roce_get_umem_bufs()
|
| /linux/drivers/staging/rtl8723bs/os_dep/ |
| H A D | osdep_service.c | 160 cbuf->bufs[cbuf->write] = buf; in rtw_cbuf_push() 180 buf = cbuf->bufs[cbuf->read]; in rtw_cbuf_pop() 196 cbuf = kzalloc_flex(*cbuf, bufs, size); in rtw_cbuf_alloc()
|
| /linux/tools/testing/selftests/bpf/ |
| H A D | xskxceiver.c | 126 void *bufs; in ifobj_zc_avail() local 129 bufs = mmap(NULL, umem_sz, PROT_READ | PROT_WRITE, mmap_flags, -1, 0); in ifobj_zc_avail() 130 if (bufs == MAP_FAILED) in ifobj_zc_avail() 135 munmap(bufs, umem_sz); in ifobj_zc_avail() 139 ret = xsk_configure_umem(ifobject, umem, bufs, umem_sz); in ifobj_zc_avail()
|
| /linux/include/soc/fsl/ |
| H A D | bman.h | 114 int bman_release(struct bman_pool *pool, const struct bm_buffer *bufs, u8 num); 127 int bman_acquire(struct bman_pool *pool, struct bm_buffer *bufs, u8 num);
|
| /linux/drivers/md/ |
| H A D | dm-verity-fec.c | 83 return &fio->bufs[i][j * v->fec->rsn]; in fec_buffer_rs_block() 314 fio->bufs[0] = mempool_alloc(&f->prealloc_pool, GFP_NOIO); in fec_alloc_and_init_io() 318 fio->bufs[n] = kmem_cache_alloc(f->cache, GFP_NOWAIT); in fec_alloc_and_init_io() 320 if (unlikely(!fio->bufs[n])) in fec_alloc_and_init_io() 339 memset(fio->bufs[n], 0, v->fec->rsn << DM_VERITY_FEC_BUF_RS_BITS); in fec_init_bufs() 459 mempool_free(fio->bufs[0], &f->prealloc_pool); in __verity_fec_finish_io() 462 kmem_cache_free(f->cache, fio->bufs[n]); in __verity_fec_finish_io() 731 bufs, fec_max_nbufs(v))); in verity_fec_ctr()
|
| /linux/sound/core/ |
| H A D | pcm_compat.c | 319 u32 bufs; /* this is void **; */ member 347 if (get_user(buf, &data32->bufs) || in snd_pcm_ioctl_xfern_compat() 352 void __user **bufs __free(kfree) = in snd_pcm_ioctl_xfern_compat() 354 if (bufs == NULL) in snd_pcm_ioctl_xfern_compat() 360 bufs[i] = compat_ptr(ptr); in snd_pcm_ioctl_xfern_compat() 364 err = snd_pcm_lib_writev(substream, bufs, frames); in snd_pcm_ioctl_xfern_compat() 366 err = snd_pcm_lib_readv(substream, bufs, frames); in snd_pcm_ioctl_xfern_compat()
|