| /linux/drivers/net/ethernet/mellanox/mlx5/core/lib/ |
| H A D | crypto.c | 20 #define MLX5_CRYPTO_DEK_BULK_CALC_FREED(bulk) MLX5_CRYPTO_DEK_CALC_FREED(bulk) argument 22 #define MLX5_CRYPTO_DEK_BULK_IDLE(bulk) \ argument 23 ({ typeof(bulk) _bulk = (bulk); \ 80 struct mlx5_crypto_dek_bulk *bulk; member 292 struct mlx5_crypto_dek_bulk *bulk; in mlx5_crypto_dek_bulk_create() local 296 bulk = kzalloc_obj(*bulk); in mlx5_crypto_dek_bulk_create() 297 if (!bulk) in mlx5_crypto_dek_bulk_create() 301 bulk->need_sync = bitmap_zalloc(num_deks, GFP_KERNEL); in mlx5_crypto_dek_bulk_create() 302 if (!bulk->need_sync) { in mlx5_crypto_dek_bulk_create() 307 bulk->in_use = bitmap_zalloc(num_deks, GFP_KERNEL); in mlx5_crypto_dek_bulk_create() [all …]
|
| /linux/drivers/platform/raspberrypi/vchiq-interface/ |
| H A D | TESTING | 49 Testing bulk transfer for alignment. 50 Testing bulk transfer at PAGE_SIZE. 61 vchi bulk (size 0, 0 async, 0 oneway) -> 546.000000us 62 vchi bulk (size 0, 0 oneway) -> 230.000000us 65 vchi bulk (size 0, 0 async, 0 oneway) -> 296.000000us 66 vchi bulk (size 0, 0 oneway) -> 266.000000us 68 vchi bulk (size 0, 0 oneway) -> 456.000000us 70 vchi bulk (size 0, 0 oneway) -> 640.000000us 72 vchi bulk (size 0, 0 oneway) -> 2309.000000us 78 vchi bulk (size 0, 0 oneway) -> nanus [all …]
|
| H A D | vchiq_core.c | 458 struct vchiq_header *header, struct vchiq_bulk *bulk) in make_service_callback() argument 468 if (bulk) { in make_service_callback() 469 cb_data = bulk->cb_data; in make_service_callback() 470 cb_userdata = bulk->cb_userdata; in make_service_callback() 1308 get_bulk_reason(struct vchiq_bulk *bulk) in get_bulk_reason() argument 1310 if (bulk->dir == VCHIQ_BULK_TRANSMIT) { in get_bulk_reason() 1311 if (bulk->actual == VCHIQ_BULK_ACTUAL_ABORTED) in get_bulk_reason() 1317 if (bulk->actual == VCHIQ_BULK_ACTUAL_ABORTED) in get_bulk_reason() 1324 struct vchiq_bulk *bulk) in service_notify_bulk() argument 1326 if (bulk->actual != VCHIQ_BULK_ACTUAL_ABORTED) { in service_notify_bulk() [all …]
|
| H A D | vchiq_arm.c | 604 struct vchiq_bulk *bulk = waiter->bulk_waiter.bulk; in vchiq_blocking_bulk_transfer() local 606 if (bulk) { in vchiq_blocking_bulk_transfer() 609 if ((bulk->dma_addr != (dma_addr_t)(uintptr_t)bulk_params->dma_addr) || in vchiq_blocking_bulk_transfer() 610 (bulk->size != bulk_params->size)) { in vchiq_blocking_bulk_transfer() 616 bulk->waiter = NULL; in vchiq_blocking_bulk_transfer() 629 if ((ret != -EAGAIN) || fatal_signal_pending(current) || !waiter->bulk_waiter.bulk) { in vchiq_blocking_bulk_transfer() 630 struct vchiq_bulk *bulk = waiter->bulk_waiter.bulk; in vchiq_blocking_bulk_transfer() local 632 if (bulk) { in vchiq_blocking_bulk_transfer() 635 bulk->waiter = NULL; in vchiq_blocking_bulk_transfer()
|
| /linux/drivers/platform/raspberrypi/vchiq-mmal/ |
| H A D | mmal-vchiq.c | 151 } bulk; /* bulk data */ member 255 container_of(work, struct mmal_msg_context, u.bulk.work); in buffer_work_cb() 256 struct mmal_buffer *buffer = msg_context->u.bulk.buffer; in buffer_work_cb() 264 buffer->length = msg_context->u.bulk.buffer_used; in buffer_work_cb() 265 buffer->mmal_flags = msg_context->u.bulk.mmal_flags; in buffer_work_cb() 266 buffer->dts = msg_context->u.bulk.dts; in buffer_work_cb() 267 buffer->pts = msg_context->u.bulk.pts; in buffer_work_cb() 269 atomic_dec(&msg_context->u.bulk.port->buffers_with_vpu); in buffer_work_cb() 271 msg_context->u.bulk.port->buffer_cb(msg_context->u.bulk.instance, in buffer_work_cb() 272 msg_context->u.bulk.port, in buffer_work_cb() [all …]
|
| /linux/drivers/net/ethernet/intel/libeth/ |
| H A D | tx.c | 12 DEFINE_STATIC_CALL_NULL(bulk, libeth_xdp_return_buff_bulk); 27 __libeth_xdp_complete_tx(sqe, cp, static_call(bulk), in libeth_tx_complete_any() 38 static_call_update(bulk, ops ? ops->bulk : NULL); in libeth_attach_xdp()
|
| H A D | xdp.c | 126 const struct libeth_xdp_tx_frame *pos = &bq->bulk[sent]; in libeth_xdp_tx_exception() 133 memmove(bq->bulk, pos, left * sizeof(*bq->bulk)); in libeth_xdp_tx_exception() 431 .bulk = libeth_xdp_return_buff_bulk,
|
| /linux/include/net/libeth/ |
| H A D | xsk.h | 27 bq->bulk[bq->count++] = (typeof(*bq->bulk)){ in libeth_xsk_tx_queue_head() 36 bq->bulk[bq->count - 1].flags |= LIBETH_XDP_TX_MULTI; in libeth_xsk_tx_queue_head() 49 bq->bulk[bq->count++] = (typeof(*bq->bulk)){ in libeth_xsk_tx_queue_frag() 96 bq->bulk[bq->count - 1].flags |= LIBETH_XDP_TX_LAST; in libeth_xsk_tx_queue_bulk() 292 const struct libeth_xdp_tx_frame *bulk; in libeth_xsk_xmit_do_bulk() local 301 bulk = container_of(&pool->tx_descs[0], typeof(*bulk), desc); in libeth_xsk_xmit_do_bulk() 303 libeth_xdp_tx_xmit_bulk(bulk, xdpsq, n, true, in libeth_xsk_xmit_do_bulk()
|
| H A D | xdp.h | 391 struct libeth_xdp_tx_frame bulk[LIBETH_XDP_TX_BULK]; member 517 libeth_xdp_tx_xmit_bulk(const struct libeth_xdp_tx_frame *bulk, void *xdpsq, in libeth_xdp_tx_xmit_bulk() argument 551 xmit(fill(bulk[i + j], base + j, &sq, priv), in libeth_xdp_tx_xmit_bulk() 558 xmit(fill(bulk[i], ntu + i - off, &sq, priv), in libeth_xdp_tx_xmit_bulk() 603 bq->bulk[bq->count++] = (typeof(*bq->bulk)){ in libeth_xdp_tx_queue_head() 612 bq->bulk[bq->count - 1].len_fl |= LIBETH_XDP_TX_MULTI; in libeth_xdp_tx_queue_head() 625 bq->bulk[bq->count++].frag = *frag; in libeth_xdp_tx_queue_frag() 669 bq->bulk[bq->count - 1].len_fl |= LIBETH_XDP_TX_LAST; in libeth_xdp_tx_queue_bulk() 783 sent = libeth_xdp_tx_xmit_bulk(bq->bulk, bq->xdpsq, in __libeth_xdp_tx_flush_bulk() 886 bq->bulk[bq->count++] = (typeof(*bq->bulk)){ in libeth_xdp_xmit_queue_head() [all …]
|
| /linux/Documentation/w1/masters/ |
| H A D | ds2490.rst | 49 - The ds2490 specification doesn't cover short bulk in reads in 51 available, the bulk read will return an error and the hardware will 52 clear the entire bulk in buffer. It would be possible to read the 63 most of the time one of the bulk out or in, and usually the bulk in 64 would fail. qemu sets a 50ms timeout and the bulk in would timeout 65 even when the status shows data available. A bulk out write would
|
| /linux/drivers/net/ethernet/mellanox/mlx5/core/ |
| H A D | fs_counters.c | 156 kfree(counter->bulk); in mlx5_fc_free() 253 counter->bulk = fc_bulk; in mlx5_fc_single_alloc() 439 static void mlx5_fc_init(struct mlx5_fc *counter, struct mlx5_fc_bulk *bulk, in mlx5_fc_init() argument 442 counter->bulk = bulk; in mlx5_fc_init() 448 return counter->bulk->base_id; in mlx5_fc_get_base_id() 549 struct mlx5_fs_bulk *fs_bulk = &fc->bulk->fs_bulk; in mlx5_fc_pool_release_counter() 554 pool_index.index = fc->id - fc->bulk->base_id; in mlx5_fc_pool_release_counter() 588 counter->bulk = fc_bulk; in mlx5_fc_local_create() 596 kfree(counter->bulk); in mlx5_fc_local_destroy()
|
| /linux/drivers/media/usb/uvc/ |
| H A D | uvc_video.c | 1391 nbytes = min(stream->bulk.max_payload_size - stream->bulk.payload_size, in uvc_video_encode_data() 1580 if (urb->actual_length == 0 && stream->bulk.header_size == 0) in uvc_video_decode_bulk() 1585 stream->bulk.payload_size += len; in uvc_video_decode_bulk() 1591 if (stream->bulk.header_size == 0 && !stream->bulk.skip_payload) { in uvc_video_decode_bulk() 1600 stream->bulk.skip_payload = 1; in uvc_video_decode_bulk() 1602 memcpy(stream->bulk.header, mem, ret); in uvc_video_decode_bulk() 1603 stream->bulk.header_size = ret; in uvc_video_decode_bulk() 1619 if (!stream->bulk.skip_payload && buf != NULL) in uvc_video_decode_bulk() 1627 stream->bulk.payload_size >= stream->bulk.max_payload_size) { in uvc_video_decode_bulk() 1628 if (!stream->bulk.skip_payload && buf != NULL) { in uvc_video_decode_bulk() [all …]
|
| /linux/drivers/net/ethernet/mellanox/mlx5/core/steering/hws/ |
| H A D | fs_hws_pools.c | 130 pr_bulk->prs_data[i].bulk = pr_bulk; in mlx5_fs_hws_pr_bulk_create() 227 struct mlx5_fs_bulk *fs_bulk = &pr_data->bulk->fs_bulk; in mlx5_fs_hws_pr_pool_release_pr() 239 return pr_data->bulk->hws_action; in mlx5_fs_hws_pr_get_action() 286 mh_bulk->mhs_data[i].bulk = mh_bulk; in mlx5_fs_hws_mh_bulk_create() 378 struct mlx5_fs_bulk *fs_bulk = &mh_data->bulk->fs_bulk; in mlx5_fs_hws_mh_pool_release_mh() 413 struct mlx5_fc_bulk *fc_bulk = counter->bulk; in mlx5_fc_get_hws_action() 429 mlx5_fs_put_hws_action(&counter->bulk->hws_data); in mlx5_fc_put_hws_action()
|
| /linux/Documentation/ABI/testing/ |
| H A D | sysfs-driver-w1_therm | 75 * If a bulk read has been triggered, it will directly 76 return the temperature computed when the bulk read 80 * If no bulk read has been triggered, it will trigger 115 (RW) trigger a bulk read conversion. read the status 124 no bulk operation. Reading temperature will 128 'trigger': trigger a bulk read on all supporting 131 Note that if a bulk read is sent but one sensor is not read 134 of the bulk read command (not the current temperature).
|
| /linux/include/drm/ttm/ |
| H A D | ttm_resource.h | 342 struct ttm_lru_bulk_move *bulk; member 442 void ttm_lru_bulk_move_init(struct ttm_lru_bulk_move *bulk); 443 void ttm_lru_bulk_move_tail(struct ttm_lru_bulk_move *bulk); 445 struct ttm_lru_bulk_move *bulk);
|
| /linux/drivers/remoteproc/ |
| H A D | qcom_wcnss.c | 447 struct regulator_bulk_data *bulk; in wcnss_init_regulators() local 465 bulk = devm_kcalloc(wcnss->dev, in wcnss_init_regulators() 468 if (!bulk) in wcnss_init_regulators() 472 bulk[i].supply = info[i].name; in wcnss_init_regulators() 474 ret = devm_regulator_bulk_get(wcnss->dev, num_vregs, bulk); in wcnss_init_regulators() 480 regulator_set_voltage(bulk[i].consumer, in wcnss_init_regulators() 485 regulator_set_load(bulk[i].consumer, info[i].load_uA); in wcnss_init_regulators() 488 wcnss->vregs = bulk; in wcnss_init_regulators()
|
| /linux/drivers/media/usb/dvb-usb-v2/ |
| H A D | usb_urb.c | 155 stream->props.u.bulk.buffersize, in usb_urb_alloc_bulk_urbs() 260 buf_size = stream->props.u.bulk.buffersize; in usb_urb_reconfig() 281 props->u.bulk.buffersize == in usb_urb_reconfig() 282 stream->props.u.bulk.buffersize) in usb_urb_reconfig() 326 stream->props.u.bulk.buffersize); in usb_urb_initv2()
|
| /linux/Documentation/driver-api/usb/ |
| H A D | bulk-streams.rst | 1 USB bulk streams 8 device driver to overload a bulk endpoint so that multiple transfers can be 41 ID for the bulk IN and OUT endpoints used in a Bi-directional command sequence. 46 declares how many stream IDs it can support, and each bulk endpoint on a
|
| /linux/net/core/ |
| H A D | page_pool.c | 593 const int bulk = PP_ALLOC_CACHE_REFILL; in __page_pool_alloc_netmems_slow() 605 /* Don't support bulk alloc for high-order pages */ in __page_pool_alloc_netmems_slow() 614 memset(&pool->alloc.cache, 0, sizeof(void *) * bulk); in __page_pool_alloc_netmems_slow() 616 nr_pages = alloc_pages_bulk_node(gfp, pool->p.nid, bulk, in __page_pool_alloc_netmems_slow() 924 netmem_ref *bulk, in page_pool_recycle_ring_bulk() 934 if (__ptr_ring_produce(&pool->ring, (__force void *)bulk[i])) { in page_pool_recycle_ring_bulk() 953 page_pool_return_netmem(pool, bulk[i]); 983 netmem_ref bulk[XDP_BULK_QUEUE_SIZE]; in page_pool_put_netmem_bulk() 1009 /* Approved for bulk recycling in ptr_ring cache */ in page_pool_put_netmem_bulk() 1011 bulk[bulk_le in page_pool_put_netmem_bulk() 589 const int bulk = PP_ALLOC_CACHE_REFILL; __page_pool_alloc_netmems_slow() local 920 page_pool_recycle_ring_bulk(struct page_pool * pool,netmem_ref * bulk,u32 bulk_len) page_pool_recycle_ring_bulk() argument 979 netmem_ref bulk[XDP_BULK_QUEUE_SIZE]; page_pool_put_netmem_bulk() local [all...] |
| /linux/lib/ |
| H A D | test_objpool.c | 72 int bulk[2]; /* for thread and irq */ member 207 item->bulk[0] = test->bulk_normal; in ot_init_cpu_item() 208 item->bulk[1] = test->bulk_irq; in ot_init_cpu_item() 329 for (i = 0; i < item->bulk[irq]; i++) in ot_bulk_sync() 506 for (i = 0; i < item->bulk[irq]; i++) in ot_bulk_async()
|
| /linux/Documentation/usb/ |
| H A D | ehci.rst | 58 At this writing the driver should comfortably handle all control, bulk, 125 and bulk transfers. Shows each active qh and the qtds 161 good to keep in mind that bulk transfers are always in 512 byte packets, 165 So more than 50 MByte/sec is available for bulk transfers, when both 195 you issue a control or bulk request you can often expect to learn that 203 or using bulk queuing if a series of small requests needs to be issued. 213 I/O be efficient, it's better to just queue up several (bulk) requests
|
| H A D | mass-storage.rst | 15 Its requirements are modest; only a bulk-in and a bulk-out endpoint 21 a single memory/DMA buffer will be usable for bulk-in and bulk-out 140 Specifies whether the gadget is allowed to halt bulk endpoints. 242 supported. MSG always uses SCSI protocol with bulk only
|
| /linux/drivers/usb/core/ |
| H A D | devio.c | 1284 struct usbdevfs_bulktransfer *bulk) in do_proc_bulk() argument 1293 ret = findintfep(ps->dev, bulk->ep); in do_proc_bulk() 1300 len1 = bulk->len; in do_proc_bulk() 1304 if (bulk->ep & USB_DIR_IN) in do_proc_bulk() 1305 pipe = usb_rcvbulkpipe(dev, bulk->ep & 0x7f); in do_proc_bulk() 1307 pipe = usb_sndbulkpipe(dev, bulk->ep & 0x7f); in do_proc_bulk() 1336 tmo = bulk->timeout; in do_proc_bulk() 1337 if (bulk->ep & 0x80) { in do_proc_bulk() 1346 if (copy_to_user(bulk->data, tbuf, len2)) { in do_proc_bulk() 1353 if (copy_from_user(tbuf, bulk->data, len1)) { in do_proc_bulk() [all …]
|
| /linux/include/linux/raspberrypi/ |
| H A D | vchiq_core.h | 495 struct vchiq_bulk *bulk; member 556 struct vchiq_bulk *bulk); 560 struct vchiq_bulk *bulk);
|
| /linux/tools/testing/selftests/net/netfilter/ |
| H A D | conntrack_resize.sh | 59 local bulk=16 62 ip netns exec "$ns" bash -c "for i in \$(seq 1 $bulk); do \ 70 i=$((i+bulk))
|