Home
last modified time | relevance | path

Searched refs:num_chunks (Results 1 – 25 of 26) sorted by relevance

12

/linux/drivers/net/ethernet/intel/idpf/
H A Didpf_virtchnl.c728 u32 num_chunks; member
773 u32 num_chunks, num_msgs, buf_sz; in idpf_send_chunked_msg() local
775 u32 totqs = params->num_chunks; in idpf_send_chunked_msg()
778 num_chunks = min(IDPF_NUM_CHUNKS_PER_MSG(params->config_sz, in idpf_send_chunked_msg()
780 num_msgs = DIV_ROUND_UP(totqs, num_chunks); in idpf_send_chunked_msg()
782 buf_sz = params->config_sz + num_chunks * params->chunk_sz; in idpf_send_chunked_msg()
795 if (params->prepare_msg(vid, buf, pos, num_chunks) != buf_sz) in idpf_send_chunked_msg()
802 pos += num_chunks * params->chunk_sz; in idpf_send_chunked_msg()
803 totqs -= num_chunks; in idpf_send_chunked_msg()
805 num_chunks = min(num_chunks, totqs); in idpf_send_chunked_msg()
[all …]
H A Dvirtchnl2.h597 __le16 num_chunks; member
599 struct virtchnl2_queue_reg_chunk chunks[] __counted_by_le(num_chunks);
1202 __le16 num_chunks; member
1204 struct virtchnl2_queue_chunk chunks[] __counted_by_le(num_chunks);
H A Didpf.h586 u16 num_chunks; member
/linux/drivers/gpu/drm/xe/
H A Dxe_guc_log.c129 snapshot->num_chunks = DIV_ROUND_UP(snapshot->size, GUC_LOG_CHUNK_SIZE); in xe_guc_log_snapshot_alloc()
131 snapshot->copy = kzalloc_objs(*snapshot->copy, snapshot->num_chunks, in xe_guc_log_snapshot_alloc()
137 for (i = 0; i < snapshot->num_chunks; i++) { in xe_guc_log_snapshot_alloc()
149 for (i = 0; i < snapshot->num_chunks; i++) in xe_guc_log_snapshot_alloc()
172 for (i = 0; i < snapshot->num_chunks; i++) in xe_guc_log_snapshot_free()
205 for (i = 0; i < snapshot->num_chunks; i++) { in xe_guc_log_snapshot_capture()
253 for (i = 0; i < snapshot->num_chunks; i++) { in xe_guc_log_snapshot_print()
256 char suffix = i == snapshot->num_chunks - 1 ? '\n' : 0; in xe_guc_log_snapshot_print()
/linux/tools/testing/selftests/mm/
H A Dmremap_test.c1002 unsigned long num_chunks; in remap_region() local
1077 num_chunks = get_sqrt(threshold); in remap_region()
1078 for (unsigned long i = 0; i < num_chunks; ++i) { in remap_region()
1079 size_t chunk_size = threshold / num_chunks; in remap_region()
1102 for (t = num_chunks * (threshold / num_chunks); t < threshold; ++t) { in remap_region()
1117 num_chunks = get_sqrt(c.dest_preamble_size); in remap_region()
1119 for (unsigned long i = 0; i < num_chunks; ++i) { in remap_region()
1120 size_t chunk_size = c.dest_preamble_size / num_chunks; in remap_region()
1140 for (d = num_chunks * (c.dest_preamble_size / num_chunks); d < c.dest_preamble_size; ++d) { in remap_region()
/linux/drivers/accel/habanalabs/common/
H A Dcommand_submission.c1351 u32 cs_type_flags, num_chunks; in hl_cs_sanity_checks() local
1382 num_chunks = args->in.num_chunks_execute; in hl_cs_sanity_checks()
1393 if (!num_chunks) { in hl_cs_sanity_checks()
1397 } else if (is_sync_stream && num_chunks != 1) { in hl_cs_sanity_checks()
1409 void __user *chunks, u32 num_chunks, in hl_cs_copy_chunk_array() argument
1414 if (num_chunks > HL_MAX_JOBS_PER_CS) { in hl_cs_copy_chunk_array()
1423 *cs_chunk_array = kmalloc_objs(**cs_chunk_array, num_chunks, GFP_ATOMIC); in hl_cs_copy_chunk_array()
1425 *cs_chunk_array = kmalloc_objs(**cs_chunk_array, num_chunks); in hl_cs_copy_chunk_array()
1432 size_to_copy = num_chunks * sizeof(struct hl_cs_chunk); in hl_cs_copy_chunk_array()
1488 u32 num_chunks, u64 *cs_seq, u32 flags, in cs_ioctl_default() argument
[all …]
/linux/drivers/infiniband/sw/siw/
H A Dsiw_mem.c341 int num_pages, num_chunks, i, rv = 0; in siw_umem_get() local
348 num_chunks = (num_pages >> CHUNK_SHIFT) + 1; in siw_umem_get()
355 kzalloc_objs(struct siw_page_chunk, num_chunks); in siw_umem_get()
/linux/drivers/platform/x86/intel/ifs/
H A Difs.h189 u32 num_chunks :8; member
202 u16 num_chunks; member
/linux/drivers/gpu/drm/
H A Ddrm_gpusvm.c389 const unsigned long *chunk_sizes, int num_chunks) in drm_gpusvm_init() argument
392 if (!ops->invalidate || !num_chunks) in drm_gpusvm_init()
397 if (ops || num_chunks || mm_range || notifier_size) in drm_gpusvm_init()
409 gpusvm->num_chunks = num_chunks; in drm_gpusvm_init()
898 for (; i < gpusvm->num_chunks; ++i) { in drm_gpusvm_range_chunk_size()
909 if (i == gpusvm->num_chunks) in drm_gpusvm_range_chunk_size()
/linux/drivers/gpu/drm/lima/
H A Dlima_dump.h47 __u32 num_chunks; member
/linux/drivers/crypto/marvell/octeontx/
H A Dotx_cptvf.h37 u32 num_chunks; /* Number of command chunks */ member
H A Dotx_cptvf_main.c180 queue->num_chunks = 0; in free_command_queues()
227 i, queue->num_chunks); in alloc_command_queues()
232 if (queue->num_chunks == 0) { in alloc_command_queues()
239 queue->num_chunks++; in alloc_command_queues()
/linux/drivers/gpu/drm/radeon/
H A Dradeon_cs.c276 if (!cs->num_chunks) { in radeon_cs_parser_init()
288 p->chunks_array = kvmalloc_array(cs->num_chunks, sizeof(uint64_t), GFP_KERNEL); in radeon_cs_parser_init()
294 sizeof(uint64_t)*cs->num_chunks)) { in radeon_cs_parser_init()
298 p->nchunks = cs->num_chunks; in radeon_cs_parser_init()
/linux/drivers/crypto/intel/qat/qat_common/
H A Dqat_uclo.c117 if (suof_hdr->num_chunks <= 0x1) { in qat_uclo_check_suof_format()
450 for (i = 0; i < obj_hdr->num_chunks; i++) { in qat_uclo_find_chunk()
497 for (i = 0; i < file_hdr->num_chunks; i++) { in qat_uclo_map_chunk()
1199 suof_handle->img_table.num_simgs = suof_ptr->num_chunks - 1; in qat_uclo_map_suof()
1889 uobj_chunk_num = uobj_hdr->num_chunks; in qat_uclo_map_objs_from_mof()
1891 sobj_chunk_num = sobj_hdr->num_chunks; in qat_uclo_map_objs_from_mof()
1971 if (mof_hdr->num_chunks <= 0x1) { in qat_uclo_check_mof_format()
2014 chunks_num = mof_ptr->num_chunks; in qat_uclo_map_mof_obj()
/linux/drivers/net/dsa/sja1105/
H A Dsja1105_spi.c44 int num_chunks; in sja1105_xfer() local
47 num_chunks = DIV_ROUND_UP(len, priv->max_xfer_len); in sja1105_xfer()
56 for (i = 0; i < num_chunks; i++) { in sja1105_xfer()
/linux/drivers/vfio/pci/mlx5/
H A Dmain.c385 int num_chunks; in mlx5vf_prep_stop_copy() local
408 num_chunks = mvdev->chunk_mode ? MAX_NUM_CHUNKS : 1; in mlx5vf_prep_stop_copy()
409 for (i = 0; i < num_chunks; i++) { in mlx5vf_prep_stop_copy()
444 for (i = 0; i < num_chunks; i++) { in mlx5vf_prep_stop_copy()
/linux/drivers/mtd/nand/raw/
H A Dmxc_nand.c254 u16 num_chunks = mtd->writesize / 512; in copy_spare() local
261 oob_chunk_size = (host->used_oobsize / num_chunks) & ~1; in copy_spare()
264 for (i = 0; i < num_chunks - 1; i++) in copy_spare()
274 for (i = 0; i < num_chunks - 1; i++) in copy_spare()
/linux/drivers/media/i2c/
H A Dthp7312.c1666 unsigned int num_chunks; in thp7312_fw_load_to_ram() local
1674 num_chunks = DIV_ROUND_UP(bank_size, THP7312_FW_DOWNLOAD_UNIT); in thp7312_fw_load_to_ram()
1677 __func__, bank_size, i, num_chunks); in thp7312_fw_load_to_ram()
1679 for (j = 0 ; j < num_chunks; j++) { in thp7312_fw_load_to_ram()
/linux/drivers/gpu/drm/amd/amdgpu/
H A Damdgpu_cs.c52 if (cs->in.num_chunks == 0) in amdgpu_cs_parser_init()
189 cs->in.num_chunks, in amdgpu_cs_pass1()
194 p->nchunks = cs->in.num_chunks; in amdgpu_cs_pass1()
/linux/net/sctp/
H A Dsocket.c7013 u32 num_chunks = 0; in sctp_getsockopt_peer_auth_chunks() local
7035 num_chunks = ntohs(ch->param_hdr.length) - sizeof(struct sctp_paramhdr); in sctp_getsockopt_peer_auth_chunks()
7036 if (len < num_chunks) in sctp_getsockopt_peer_auth_chunks()
7039 if (copy_to_user(to, ch->chunks, num_chunks)) in sctp_getsockopt_peer_auth_chunks()
7042 len = sizeof(struct sctp_authchunks) + num_chunks; in sctp_getsockopt_peer_auth_chunks()
7045 if (put_user(num_chunks, &p->gauth_number_of_chunks)) in sctp_getsockopt_peer_auth_chunks()
7058 u32 num_chunks = 0; in sctp_getsockopt_local_auth_chunks() local
7085 num_chunks = ntohs(ch->param_hdr.length) - sizeof(struct sctp_paramhdr); in sctp_getsockopt_local_auth_chunks()
7086 if (len < sizeof(struct sctp_authchunks) + num_chunks) in sctp_getsockopt_local_auth_chunks()
7089 if (copy_to_user(to, ch->chunks, num_chunks)) in sctp_getsockopt_local_auth_chunks()
[all …]
/linux/include/uapi/drm/
H A Dradeon_drm.h984 __u32 num_chunks; member
H A Damdgpu_drm.h970 __u32 num_chunks; member
/linux/drivers/net/wireless/intel/ipw2x00/
H A Dipw2200.c3764 if (le32_to_cpu(bd->u.data.num_chunks) > NUM_TFD_CHUNKS) { in ipw_queue_tx_free_tfd()
3766 le32_to_cpu(bd->u.data.num_chunks)); in ipw_queue_tx_free_tfd()
3772 for (i = 0; i < le32_to_cpu(bd->u.data.num_chunks); i++) { in ipw_queue_tx_free_tfd()
10167 tfd->u.data.num_chunks = cpu_to_le32(min((u8) (NUM_TFD_CHUNKS - 2), in ipw_tx_skb()
10170 txb->nr_frags, le32_to_cpu(tfd->u.data.num_chunks)); in ipw_tx_skb()
10171 for (i = 0; i < le32_to_cpu(tfd->u.data.num_chunks); i++) { in ipw_tx_skb()
10173 i, le32_to_cpu(tfd->u.data.num_chunks), in ipw_tx_skb()
10176 i, tfd->u.data.num_chunks, in ipw_tx_skb()
10220 le32_add_cpu(&tfd->u.data.num_chunks, 1); in ipw_tx_skb()
/linux/drivers/staging/media/atomisp/pci/runtime/binary/src/
H A Dbinary.c96 info->output.num_chunks, info->pipeline.pipelining); in ia_css_binary_internal_res()
/linux/fs/btrfs/
H A Dinode.c775 atomic_t num_chunks; member
1659 if (atomic_dec_and_test(&async_cow->num_chunks)) in submit_compressed_extents()
1689 u64 num_chunks = DIV_ROUND_UP(end - start, SZ_512K); in run_delalloc_compressed() local
1695 ctx = kvmalloc_flex(*ctx, chunks, num_chunks); in run_delalloc_compressed()
1703 atomic_set(&ctx->num_chunks, num_chunks); in run_delalloc_compressed()
1705 for (i = 0; i < num_chunks; i++) { in run_delalloc_compressed()

12