| /linux/drivers/gpu/drm/xe/ |
| H A D | xe_guc_log.c | 83 snapshot->num_chunks = DIV_ROUND_UP(snapshot->size, GUC_LOG_CHUNK_SIZE); in xe_guc_log_snapshot_alloc() 85 snapshot->copy = kcalloc(snapshot->num_chunks, sizeof(*snapshot->copy), in xe_guc_log_snapshot_alloc() 91 for (i = 0; i < snapshot->num_chunks; i++) { in xe_guc_log_snapshot_alloc() 103 for (i = 0; i < snapshot->num_chunks; i++) in xe_guc_log_snapshot_alloc() 126 for (i = 0; i < snapshot->num_chunks; i++) in xe_guc_log_snapshot_free() 160 for (i = 0; i < snapshot->num_chunks; i++) { in xe_guc_log_snapshot_capture() 209 for (i = 0; i < snapshot->num_chunks; i++) { in xe_guc_log_snapshot_print() 212 char suffix = i == snapshot->num_chunks - 1 ? '\n' : 0; in xe_guc_log_snapshot_print()
|
| /linux/drivers/net/ethernet/intel/idpf/ |
| H A D | idpf_virtchnl.c | 725 u32 num_chunks; member 765 u32 num_chunks, num_msgs, buf_sz; in idpf_send_chunked_msg() local 767 u32 totqs = params->num_chunks; in idpf_send_chunked_msg() 769 num_chunks = min(IDPF_NUM_CHUNKS_PER_MSG(params->config_sz, in idpf_send_chunked_msg() 771 num_msgs = DIV_ROUND_UP(totqs, num_chunks); in idpf_send_chunked_msg() 773 buf_sz = params->config_sz + num_chunks * params->chunk_sz; in idpf_send_chunked_msg() 786 if (params->prepare_msg(vport, buf, pos, num_chunks) != buf_sz) in idpf_send_chunked_msg() 793 pos += num_chunks * params->chunk_sz; in idpf_send_chunked_msg() 794 totqs -= num_chunks; in idpf_send_chunked_msg() 796 num_chunks = min(num_chunks, totqs); in idpf_send_chunked_msg() [all …]
|
| H A D | virtchnl2.h | 597 __le16 num_chunks; member 599 struct virtchnl2_queue_reg_chunk chunks[] __counted_by_le(num_chunks); 1202 __le16 num_chunks; member 1204 struct virtchnl2_queue_chunk chunks[] __counted_by_le(num_chunks);
|
| /linux/tools/testing/selftests/mm/ |
| H A D | mremap_test.c | 1002 unsigned long num_chunks; in remap_region() local 1077 num_chunks = get_sqrt(threshold); in remap_region() 1078 for (unsigned long i = 0; i < num_chunks; ++i) { in remap_region() 1079 size_t chunk_size = threshold / num_chunks; in remap_region() 1102 for (t = num_chunks * (threshold / num_chunks); t < threshold; ++t) { in remap_region() 1117 num_chunks = get_sqrt(c.dest_preamble_size); in remap_region() 1119 for (unsigned long i = 0; i < num_chunks; ++i) { in remap_region() 1120 size_t chunk_size = c.dest_preamble_size / num_chunks; in remap_region() 1140 for (d = num_chunks * (c.dest_preamble_size / num_chunks); d < c.dest_preamble_size; ++d) { in remap_region()
|
| /linux/drivers/infiniband/sw/siw/ |
| H A D | siw_mem.c | 341 int num_pages, num_chunks, i, rv = 0; in siw_umem_get() local 348 num_chunks = (num_pages >> CHUNK_SHIFT) + 1; in siw_umem_get() 355 kcalloc(num_chunks, sizeof(struct siw_page_chunk), GFP_KERNEL); in siw_umem_get()
|
| /linux/drivers/platform/x86/intel/ifs/ |
| H A D | ifs.h | 189 u32 num_chunks :8; member 202 u16 num_chunks; member
|
| /linux/drivers/gpu/drm/lima/ |
| H A D | lima_dump.h | 47 __u32 num_chunks; member
|
| /linux/drivers/crypto/marvell/octeontx/ |
| H A D | otx_cptvf.h | 37 u32 num_chunks; /* Number of command chunks */ member
|
| H A D | otx_cptvf_main.c | 179 queue->num_chunks = 0; in free_command_queues() 226 i, queue->num_chunks); in alloc_command_queues() 231 if (queue->num_chunks == 0) { in alloc_command_queues() 238 queue->num_chunks++; in alloc_command_queues()
|
| /linux/drivers/crypto/intel/qat/qat_common/ |
| H A D | qat_uclo.c | 117 if (suof_hdr->num_chunks <= 0x1) { in qat_uclo_check_suof_format() 450 for (i = 0; i < obj_hdr->num_chunks; i++) { in qat_uclo_find_chunk() 497 for (i = 0; i < file_hdr->num_chunks; i++) { in qat_uclo_map_chunk() 1200 suof_handle->img_table.num_simgs = suof_ptr->num_chunks - 1; in qat_uclo_map_suof() 1891 uobj_chunk_num = uobj_hdr->num_chunks; in qat_uclo_map_objs_from_mof() 1893 sobj_chunk_num = sobj_hdr->num_chunks; in qat_uclo_map_objs_from_mof() 1973 if (mof_hdr->num_chunks <= 0x1) { in qat_uclo_check_mof_format() 2016 chunks_num = mof_ptr->num_chunks; in qat_uclo_map_mof_obj()
|
| /linux/drivers/net/dsa/sja1105/ |
| H A D | sja1105_spi.c | 44 int num_chunks; in sja1105_xfer() local 47 num_chunks = DIV_ROUND_UP(len, priv->max_xfer_len); in sja1105_xfer() 56 for (i = 0; i < num_chunks; i++) { in sja1105_xfer()
|
| /linux/drivers/vfio/pci/mlx5/ |
| H A D | main.c | 385 int num_chunks; in mlx5vf_prep_stop_copy() local 408 num_chunks = mvdev->chunk_mode ? MAX_NUM_CHUNKS : 1; in mlx5vf_prep_stop_copy() 409 for (i = 0; i < num_chunks; i++) { in mlx5vf_prep_stop_copy() 444 for (i = 0; i < num_chunks; i++) { in mlx5vf_prep_stop_copy()
|
| /linux/drivers/mtd/nand/raw/ |
| H A D | mxc_nand.c | 254 u16 num_chunks = mtd->writesize / 512; in copy_spare() local 261 oob_chunk_size = (host->used_oobsize / num_chunks) & ~1; in copy_spare() 264 for (i = 0; i < num_chunks - 1; i++) in copy_spare() 274 for (i = 0; i < num_chunks - 1; i++) in copy_spare()
|
| /linux/drivers/media/i2c/ |
| H A D | thp7312.c | 1666 unsigned int num_chunks; in thp7312_fw_load_to_ram() local 1674 num_chunks = DIV_ROUND_UP(bank_size, THP7312_FW_DOWNLOAD_UNIT); in thp7312_fw_load_to_ram() 1677 __func__, bank_size, i, num_chunks); in thp7312_fw_load_to_ram() 1679 for (j = 0 ; j < num_chunks; j++) { in thp7312_fw_load_to_ram()
|
| /linux/drivers/gpu/drm/amd/amdgpu/ |
| H A D | amdgpu_cs.c | 52 if (cs->in.num_chunks == 0) in amdgpu_cs_parser_init() 189 cs->in.num_chunks, in amdgpu_cs_pass1() 194 p->nchunks = cs->in.num_chunks; in amdgpu_cs_pass1()
|
| /linux/include/uapi/drm/ |
| H A D | radeon_drm.h | 984 __u32 num_chunks; member
|
| H A D | amdgpu_drm.h | 974 __u32 num_chunks; member
|
| /linux/drivers/staging/media/atomisp/pci/runtime/binary/src/ |
| H A D | binary.c | 96 info->output.num_chunks, info->pipeline.pipelining); in ia_css_binary_internal_res()
|
| /linux/drivers/staging/media/ipu3/ |
| H A D | ipu3-abi.h | 1559 u32 num_chunks; member
|