Home
last modified time | relevance | path

Searched refs:bulk_len (Results 1 – 9 of 9) sorted by relevance

/linux/drivers/net/ethernet/mellanox/mlx5/core/steering/hws/
H A Dfs_hws_pools.c113 int bulk_len; in mlx5_fs_hws_pr_bulk_create() local
119 bulk_len = MLX5_FS_HWS_DEFAULT_BULK_LEN; in mlx5_fs_hws_pr_bulk_create()
120 pr_bulk = kvzalloc_flex(*pr_bulk, prs_data, bulk_len); in mlx5_fs_hws_pr_bulk_create()
124 mlx5_fs_bulk_init(&pr_bulk->fs_bulk, bulk_len); in mlx5_fs_hws_pr_bulk_create()
129 for (i = 0; i < bulk_len; i++) { in mlx5_fs_hws_pr_bulk_create()
153 if (mlx5_fs_bulk_get_free_amount(fs_bulk) < fs_bulk->bulk_len) { in mlx5_fs_hws_pr_bulk_destroy()
261 int bulk_len; in mlx5_fs_hws_mh_bulk_create() local
275 bulk_len = MLX5_FS_HWS_DEFAULT_BULK_LEN; in mlx5_fs_hws_mh_bulk_create()
276 mh_bulk = kvzalloc_flex(*mh_bulk, mhs_data, bulk_len); in mlx5_fs_hws_mh_bulk_create()
280 mlx5_fs_bulk_init(&mh_bulk->fs_bulk, bulk_len); in mlx5_fs_hws_mh_bulk_create()
[all …]
/linux/lib/crypto/
H A Dpoly1305.c41 unsigned int bulk_len; in poly1305_update() local
55 bulk_len = round_down(nbytes, POLY1305_BLOCK_SIZE); in poly1305_update()
58 if (bulk_len) { in poly1305_update()
59 poly1305_blocks(&desc->state, src, bulk_len, 1); in poly1305_update()
60 src += bulk_len; in poly1305_update()
/linux/drivers/net/ethernet/mellanox/mlx5/core/
H A Dfs_counters.c114 u32 bulk_len = fc_stats->bulk_query_len; in mlx5_fc_stats_query_all_counters() local
131 last_bulk_id = bulk_base_id + bulk_len; in mlx5_fc_stats_query_all_counters()
137 err = mlx5_cmd_fc_bulk_query(dev, bulk_base_id, bulk_len, data); in mlx5_fc_stats_query_all_counters()
456 int bulk_len; in mlx5_fc_bulk_create() local
461 bulk_len = alloc_bitmask > 0 ? MLX5_FC_BULK_NUM_FCS(alloc_bitmask) : 1; in mlx5_fc_bulk_create()
463 fc_bulk = kvzalloc_flex(*fc_bulk, fcs, bulk_len); in mlx5_fc_bulk_create()
467 mlx5_fs_bulk_init(&fc_bulk->fs_bulk, bulk_len); in mlx5_fc_bulk_create()
476 for (i = 0; i < bulk_len; i++) in mlx5_fc_bulk_create()
495 if (mlx5_fs_bulk_get_free_amount(fs_bulk) < fs_bulk->bulk_len) { in mlx5_fc_bulk_destroy()
H A Dfs_cmd.h118 int mlx5_cmd_fc_get_bulk_query_out_len(int bulk_len);
119 int mlx5_cmd_fc_bulk_query(struct mlx5_core_dev *dev, u32 base_id, int bulk_len,
H A Dfs_cmd.c882 int mlx5_cmd_fc_get_bulk_query_out_len(int bulk_len) in mlx5_cmd_fc_get_bulk_query_out_len() argument
885 MLX5_ST_SZ_BYTES(traffic_counter) * bulk_len; in mlx5_cmd_fc_get_bulk_query_out_len()
888 int mlx5_cmd_fc_bulk_query(struct mlx5_core_dev *dev, u32 base_id, int bulk_len, in mlx5_cmd_fc_bulk_query() argument
891 int outlen = mlx5_cmd_fc_get_bulk_query_out_len(bulk_len); in mlx5_cmd_fc_bulk_query()
897 MLX5_SET(query_flow_counter_in, in, num_of_counters, bulk_len); in mlx5_cmd_fc_bulk_query()
/linux/net/core/
H A Dpage_pool.c925 u32 bulk_len) in page_pool_recycle_ring_bulk()
933 for (i = 0; i < bulk_len; i++) { in page_pool_recycle_ring_bulk()
945 if (likely(i == bulk_len)) in page_pool_recycle_ring_bulk()
952 for (; i < bulk_len; i++)
972 u32 bulk_len = 0; in page_pool_put_netmem_bulk()
978 data[bulk_len++] = netmem; in page_pool_put_netmem_bulk()
981 count = bulk_len; in page_pool_put_netmem_bulk()
988 bulk_len = 0; in page_pool_put_netmem_bulk()
1011 bulk[bulk_len++] = netmem; in page_pool_put_netmem_bulk()
1014 if (bulk_len) in page_pool_put_netmem_bulk()
921 page_pool_recycle_ring_bulk(struct page_pool * pool,netmem_ref * bulk,u32 bulk_len) page_pool_recycle_ring_bulk() argument
968 u32 bulk_len = 0; page_pool_put_netmem_bulk() local
[all...]
/linux/crypto/
H A Dhctr2.c159 const unsigned int bulk_len = req->cryptlen - BLOCKCIPHER_BLOCK_SIZE; in hctr2_hash_message() local
166 for (i = 0; i < bulk_len; i += n) { in hctr2_hash_message()
168 n = min_t(unsigned int, miter.length, bulk_len - i); in hctr2_hash_message()
215 int bulk_len = req->cryptlen - BLOCKCIPHER_BLOCK_SIZE; in hctr2_crypt() local
254 rctx->bulk_part_dst, bulk_len, in hctr2_crypt()
/linux/drivers/gpu/drm/gud/
H A Dgud_drv.c318 string_get_size(gdrm->bulk_len, 1, STRING_UNITS_2, buf, sizeof(buf)); in gud_stats_debugfs()
399 gdrm->bulk_buf = vmalloc_32(gdrm->bulk_len); in gud_alloc_bulk_buffer()
403 num_pages = DIV_ROUND_UP(gdrm->bulk_len, PAGE_SIZE); in gud_alloc_bulk_buffer()
412 0, gdrm->bulk_len, GFP_KERNEL); in gud_alloc_bulk_buffer()
587 gdrm->bulk_len = max_buffer_size; in gud_probe()
598 gdrm->compress_buf = vmalloc(gdrm->bulk_len); in gud_probe()
H A Dgud_pipe.c165 if (len > gdrm->bulk_len) in gud_prep_flush()
342 if (gdrm->bulk_len < lines * pitch) in gud_flush_damage()
343 lines = gdrm->bulk_len / pitch; in gud_flush_damage()