Home
last modified time | relevance | path

Searched refs:num_instances (Results 1 – 25 of 25) sorted by relevance

/linux/drivers/accel/habanalabs/common/
H A Dsecurity.c308 u32 dcore_offset, u32 num_instances, u32 instance_offset, in hl_init_pb_with_mask() argument
325 for (j = 0 ; j < num_instances ; j++) { in hl_init_pb_with_mask()
326 int seq = i * num_instances + j; in hl_init_pb_with_mask()
358 u32 num_instances, u32 instance_offset, in hl_init_pb() argument
363 num_instances, instance_offset, pb_blocks, in hl_init_pb()
386 u32 dcore_offset, u32 num_instances, u32 instance_offset, in hl_init_pb_ranges_with_mask() argument
407 for (j = 0 ; j < num_instances ; j++) { in hl_init_pb_ranges_with_mask()
408 int seq = i * num_instances + j; in hl_init_pb_ranges_with_mask()
442 u32 dcore_offset, u32 num_instances, u32 instance_offset, in hl_init_pb_ranges() argument
448 num_instances, instance_offset, pb_blocks, in hl_init_pb_ranges()
[all …]
/linux/drivers/iommu/arm/arm-smmu/
H A Darm-smmu-nvidia.c37 unsigned int num_instances; member
69 for (i = 0; i < nvidia->num_instances; i++) { in nvidia_smmu_write_reg()
90 for (i = 0; i < nvidia->num_instances; i++) { in nvidia_smmu_write_reg64()
112 for (i = 0; i < nvidia->num_instances; i++) { in nvidia_smmu_tlb_sync()
137 for (i = 0; i < nvidia->num_instances; i++) { in nvidia_smmu_reset()
182 for (inst = 0; inst < nvidia->num_instances; inst++) { in nvidia_smmu_global_fault()
230 for (inst = 0; inst < nvidia->num_instances; inst++) { in nvidia_smmu_context_fault()
323 nvidia_smmu->num_instances++; in nvidia_smmu_impl_init()
334 nvidia_smmu->num_instances++; in nvidia_smmu_impl_init()
337 if (nvidia_smmu->num_instances == 1) in nvidia_smmu_impl_init()
/linux/drivers/gpu/drm/amd/amdgpu/
H A Damdgpu_sdma.c45 for (i = 0; i < adev->sdma.num_instances; i++) in amdgpu_sdma_get_instance_from_ring()
58 for (i = 0; i < adev->sdma.num_instances; i++) { in amdgpu_sdma_get_index_from_ring()
103 for (i = 0; i < adev->sdma.num_instances; i++) { in amdgpu_sdma_ras_late_init()
193 for (i = 0; i < adev->sdma.num_instances; i++) { in amdgpu_sdma_destroy_inst_ctx()
240 for (i = 1; i < adev->sdma.num_instances; i++) in amdgpu_sdma_init_microcode()
252 for (i = 0; i < adev->sdma.num_instances; i++) { in amdgpu_sdma_init_microcode()
371 mask = BIT_ULL(adev->sdma.num_instances * num_ring) - 1; in amdgpu_debugfs_sdma_sched_mask_set()
376 for (i = 0; i < adev->sdma.num_instances; ++i) { in amdgpu_debugfs_sdma_sched_mask_set()
415 for (i = 0; i < adev->sdma.num_instances; ++i) { in amdgpu_debugfs_sdma_sched_mask_get()
450 if (!(adev->sdma.num_instances > 1)) in amdgpu_debugfs_sdma_sched_mask_init()
[all …]
H A Dsdma_v6_0.c400 for (i = 0; i < adev->sdma.num_instances; i++) { in sdma_v6_0_gfx_stop()
436 for (i = 0; i < adev->sdma.num_instances; i++) { in sdma_v6_0_ctxempty_int_enable()
466 for (i = 0; i < adev->sdma.num_instances; i++) { in sdma_v6_0_enable()
568 adev->doorbell_index.sdma_doorbell_range * adev->sdma.num_instances); in sdma_v6_0_gfx_resume_instance()
636 for (i = 0; i < adev->sdma.num_instances; i++) { in sdma_v6_0_gfx_resume()
718 for (i = 0; i < adev->sdma.num_instances; i++) { in sdma_v6_0_load_microcode()
768 for (i = 0; i < adev->sdma.num_instances; i++) { in sdma_v6_0_soft_reset()
803 for (i = 0; i < adev->sdma.num_instances; i++) { in sdma_v6_0_check_soft_reset()
1342 for (i = 0; i < adev->sdma.num_instances; i++) { in sdma_v6_0_sw_init()
1377 ptr = kcalloc(adev->sdma.num_instances * reg_count, sizeof(uint32_t), GFP_KERNEL); in sdma_v6_0_sw_init()
[all …]
H A Dsdma_v5_2.c477 for (i = 0; i < adev->sdma.num_instances; i++) { in sdma_v5_2_ctx_switch_enable()
511 inst_mask = GENMASK(adev->sdma.num_instances - 1, 0); in sdma_v5_2_enable()
518 for (i = 0; i < adev->sdma.num_instances; i++) { in sdma_v5_2_enable()
700 for (i = 0; i < adev->sdma.num_instances; i++) { in sdma_v5_2_gfx_resume()
740 for (i = 0; i < adev->sdma.num_instances; i++) { in sdma_v5_2_load_microcode()
795 for (i = 0; i < adev->sdma.num_instances; i++) { in sdma_v5_2_soft_reset()
1312 for (i = 0; i < adev->sdma.num_instances; i++) { in sdma_v5_2_sw_init()
1320 for (i = 0; i < adev->sdma.num_instances; i++) { in sdma_v5_2_sw_init()
1350 ptr = kcalloc(adev->sdma.num_instances * reg_count, sizeof(uint32_t), GFP_KERNEL); in sdma_v5_2_sw_init()
1368 for (i = 0; i < adev->sdma.num_instances; i++) in sdma_v5_2_sw_fini()
[all …]
H A Dvpe_v6_1.c78 for (i = 0; i < vpe->num_instances; i++) { in vpe_v6_1_halt()
108 for (i = 0; i < vpe->num_instances; i++) { in vpe_v6_1_set_collaborate_mode()
133 for (j = 0; j < vpe->num_instances; j++) { in vpe_v6_1_load_microcode()
183 for (j = 0; j < vpe->num_instances; j++) { in vpe_v6_1_load_microcode()
215 for (i = 0; i < vpe->num_instances; i++) { in vpe_v6_1_ring_start()
282 for (i = 0; i < vpe->num_instances; i++) { in vpe_v_6_1_ring_stop()
H A Dsdma_v5_0.c294 for (i = 0; i < adev->sdma.num_instances; i++) { in sdma_v5_0_init_microcode()
627 for (i = 0; i < adev->sdma.num_instances; i++) { in sdma_v5_0_ctx_switch_enable()
662 inst_mask = GENMASK(adev->sdma.num_instances - 1, 0); in sdma_v5_0_enable()
671 for (i = 0; i < adev->sdma.num_instances; i++) { in sdma_v5_0_enable()
853 for (i = 0; i < adev->sdma.num_instances; i++) { in sdma_v5_0_gfx_resume()
893 for (i = 0; i < adev->sdma.num_instances; i++) { in sdma_v5_0_load_microcode()
1401 for (i = 0; i < adev->sdma.num_instances; i++) { in sdma_v5_0_sw_init()
1432 ptr = kcalloc(adev->sdma.num_instances * reg_count, sizeof(uint32_t), GFP_KERNEL); in sdma_v5_0_sw_init()
1450 for (i = 0; i < adev->sdma.num_instances; i++) in sdma_v5_0_sw_fini()
1501 for (i = 0; i < adev->sdma.num_instances; i++) { in sdma_v5_0_is_idle()
[all …]
H A Dsdma_v4_4.c243 for (i = 0; i < adev->sdma.num_instances; i++) { in sdma_v4_4_reset_ras_error_count()
256 for (i = 0; i < adev->sdma.num_instances; i++) { in sdma_v4_4_query_ras_error_count()
H A Damdgpu_discovery.c1479 if (adev->sdma.num_instances < in amdgpu_discovery_reg_base_init()
1481 adev->sdma.num_instances++; in amdgpu_discovery_reg_base_init()
1486 adev->sdma.num_instances + 1, in amdgpu_discovery_reg_base_init()
1492 if (adev->vpe.num_instances < AMDGPU_MAX_VPE_INSTANCES) in amdgpu_discovery_reg_base_init()
1493 adev->vpe.num_instances++; in amdgpu_discovery_reg_base_init()
1496 adev->vpe.num_instances + 1, in amdgpu_discovery_reg_base_init()
2690 adev->sdma.num_instances = 2; in amdgpu_discovery_set_ip_blocks()
2719 adev->sdma.num_instances = 2; in amdgpu_discovery_set_ip_blocks()
2748 adev->sdma.num_instances = 1; in amdgpu_discovery_set_ip_blocks()
2796 adev->sdma.num_instances = 2; in amdgpu_discovery_set_ip_blocks()
[all …]
H A Damdgpu_kms.c324 if (query_fw->index >= adev->sdma.num_instances) in amdgpu_firmware_info()
480 for (i = 0; i < adev->sdma.num_instances; i++) in amdgpu_hw_ip_info()
752 count = adev->sdma.num_instances; in amdgpu_info_ioctl()
764 count = adev->vpe.num_instances; in amdgpu_info_ioctl()
1920 for (i = 0; i < adev->sdma.num_instances; i++) { in amdgpu_debugfs_firmware_info_show()
H A Damdgpu_dev_coredump.c159 for (int i = 0; i < adev->sdma.num_instances; i++) { in amdgpu_devcoredump_fw_info()
H A Dsoc15.c1276 for (i = 0; i < adev->sdma.num_instances; i++) { in soc15_sdma_doorbell_range_init()
H A Damdgpu_ras.c453 mask = GENMASK(adev->sdma.num_instances - 1, 0); in amdgpu_ras_instance_mask_check()
5602 void amdgpu_ras_query_boot_status(struct amdgpu_device *adev, u32 num_instances) in amdgpu_ras_query_boot_status() argument
5606 for (i = 0; i < num_instances; i++) { in amdgpu_ras_query_boot_status()
H A Dgfx_v12_0.c4126 if (adev->sdma.num_instances > 1) { in gfx_v12_0_update_coarse_grain_clock_gating()
/linux/drivers/gpu/drm/amd/include/
H A Damdgpu_reg_state.h51 uint8_t num_instances; member
/linux/drivers/media/platform/samsung/exynos4-is/
H A Dfimc-lite.h69 unsigned short num_instances; member
/linux/include/sound/
H A Dtimer.h79 int num_instances; /* current number of timer instances */ member
/linux/drivers/hwmon/
H A Dibmaem.c191 u8 num_instances; member
203 u8 num_instances; member
510 return ff_resp.num_instances; in aem_find_aem1_count()
646 fi_resp->num_instances <= instance_num) in aem_find_aem2()
/linux/sound/core/
H A Dtimer.c235 if (master->timer->num_instances >= master->timer->max_instances) in check_matching_master_slave()
238 master->timer->num_instances++; in check_matching_master_slave()
353 if (timer->num_instances >= timer->max_instances) { in snd_timer_open()
380 timer->num_instances++; in snd_timer_open()
406 timer->num_instances--; in remove_slave_links()
438 timer->num_instances--; in snd_timer_close_locked()
/linux/drivers/gpu/drm/amd/amdkfd/
H A Dkfd_device.c1500 return node->adev->sdma.num_instances/(int)node->kfd->num_nodes; in kfd_get_num_sdma_engines()
1502 return min(node->adev->sdma.num_instances/(int)node->kfd->num_nodes, 2); in kfd_get_num_sdma_engines()
1508 return node->adev->sdma.num_instances/(int)node->kfd->num_nodes - in kfd_get_num_xgmi_sdma_engines()
/linux/drivers/media/platform/ti/vpe/
H A Dvpe.c377 atomic_t num_instances; /* count of driver instances */ member
2366 if (atomic_inc_return(&dev->num_instances) == 1) in vpe_open()
2429 if (atomic_dec_return(&dev->num_instances) == 0) in vpe_release()
2535 atomic_set(&dev->num_instances, 0); in vpe_probe()
/linux/include/net/bluetooth/
H A Dmgmt.h558 __u8 num_instances; member
/linux/drivers/gpu/drm/i915/gt/
H A Dintel_engine_cs.c905 u8 class, const u8 *map, u8 num_instances) in populate_logical_ids() argument
910 for (j = 0; j < num_instances; ++j) { in populate_logical_ids()
/linux/drivers/net/ethernet/mellanox/mlx4/
H A Dresource_tracker.c426 int vf, int num_instances) in initialize_res_quotas() argument
428 res_alloc->guaranteed[vf] = num_instances / in initialize_res_quotas()
430 res_alloc->quota[vf] = (num_instances / 2) + res_alloc->guaranteed[vf]; in initialize_res_quotas()
432 res_alloc->res_free = num_instances; in initialize_res_quotas()
/linux/net/bluetooth/
H A Dmgmt.c8538 rp->num_instances = hdev->adv_instance_cnt; in read_adv_features()
8547 rp->num_instances--; in read_adv_features()