| /linux/drivers/gpu/drm/amd/amdgpu/ |
| H A D | amdgpu_nbio.c | 30 if (!adev->nbio.ras) in amdgpu_nbio_ras_sw_init() 33 ras = adev->nbio.ras; in amdgpu_nbio_ras_sw_init() 43 adev->nbio.ras_if = &ras->ras_block.ras_comm; in amdgpu_nbio_ras_sw_init() 50 if (adev->nbio.funcs && adev->nbio.funcs->get_pcie_replay_count) in amdgpu_nbio_get_pcie_replay_count() 51 return adev->nbio.funcs->get_pcie_replay_count(adev); in amdgpu_nbio_get_pcie_replay_count() 60 (!adev->nbio.funcs || !adev->nbio.funcs->get_pcie_replay_count)) in amdgpu_nbio_is_replay_cnt_supported() 74 r = amdgpu_irq_get(adev, &adev->nbio.ras_controller_irq, 0); in amdgpu_nbio_ras_late_init() 77 r = amdgpu_irq_get(adev, &adev->nbio.ras_err_event_athub_irq, 0); in amdgpu_nbio_ras_late_init()
|
| H A D | soc15.c | 342 return adev->nbio.funcs->get_memsize(adev); in soc15_get_config_memsize() 513 adev->nbio.funcs->enable_doorbell_interrupt(adev, false); in soc15_asic_baco_reset() 521 adev->nbio.funcs->enable_doorbell_interrupt(adev, true); in soc15_asic_baco_reset() 703 if (adev->nbio.funcs->program_aspm) in soc15_program_aspm() 704 adev->nbio.funcs->program_aspm(adev); in soc15_program_aspm() 963 adev->nbio.funcs->set_reg_remap(adev); in soc15_common_early_init() 1241 adev->nbio.funcs->enable_doorbell_selfring_aperture(adev, true); in soc15_common_late_init() 1277 adev->nbio.funcs->sdma_doorbell_range(adev, i, in soc15_sdma_doorbell_range_init() 1291 adev->nbio.funcs->init_registers(adev); in soc15_common_hw_init() 1296 if (adev->nbio.funcs->remap_hdp_registers && !amdgpu_sriov_vf(adev)) in soc15_common_hw_init() [all …]
|
| H A D | soc21.c | 254 return adev->nbio.funcs->get_memsize(adev); in soc21_get_config_memsize() 389 u32 memsize = adev->nbio.funcs->get_memsize(adev); 478 if (adev->nbio.funcs->program_aspm) in soc21_program_aspm() 479 adev->nbio.funcs->program_aspm(adev); in soc21_program_aspm() 591 adev->nbio.funcs->set_reg_remap(adev); in soc21_common_early_init() 900 if (adev->nbio.ras && in soc21_common_late_init() 901 adev->nbio.ras_err_event_athub_irq.funcs) in soc21_common_late_init() 906 amdgpu_irq_get(adev, &adev->nbio.ras_err_event_athub_irq, 0); in soc21_common_late_init() 912 adev->nbio.funcs->enable_doorbell_selfring_aperture(adev, true); in soc21_common_late_init() 934 adev->nbio.funcs->init_registers(adev); in soc21_common_hw_init() [all …]
|
| H A D | amdgpu_discovery.c | 3030 adev->nbio.funcs = &nbio_v6_1_funcs; in amdgpu_discovery_set_ip_blocks() 3031 adev->nbio.hdp_flush_reg = &nbio_v6_1_hdp_flush_reg; in amdgpu_discovery_set_ip_blocks() 3036 adev->nbio.funcs = &nbio_v7_0_funcs; in amdgpu_discovery_set_ip_blocks() 3037 adev->nbio.hdp_flush_reg = &nbio_v7_0_hdp_flush_reg; in amdgpu_discovery_set_ip_blocks() 3042 adev->nbio.funcs = &nbio_v7_4_funcs; in amdgpu_discovery_set_ip_blocks() 3043 adev->nbio.hdp_flush_reg = &nbio_v7_4_hdp_flush_reg; in amdgpu_discovery_set_ip_blocks() 3047 adev->nbio.funcs = &nbio_v7_9_funcs; in amdgpu_discovery_set_ip_blocks() 3048 adev->nbio.hdp_flush_reg = &nbio_v7_9_hdp_flush_reg; in amdgpu_discovery_set_ip_blocks() 3054 adev->nbio.funcs = &nbio_v7_11_funcs; in amdgpu_discovery_set_ip_blocks() 3055 adev->nbio.hdp_flush_reg = &nbio_v7_11_hdp_flush_reg; in amdgpu_discovery_set_ip_blocks() [all …]
|
| H A D | hdp_v5_2.c | 44 if (adev->nbio.funcs->get_memsize) in hdp_v5_2_flush_hdp() 45 adev->nbio.funcs->get_memsize(adev); in hdp_v5_2_flush_hdp()
|
| H A D | amdgpu_device.c | 1095 pcie_index = adev->nbio.funcs->get_pcie_index_offset(adev); in amdgpu_device_indirect_rreg() 1096 pcie_data = adev->nbio.funcs->get_pcie_data_offset(adev); in amdgpu_device_indirect_rreg() 1119 if (unlikely(!adev->nbio.funcs)) { in amdgpu_device_indirect_rreg_ext() 1123 pcie_index = adev->nbio.funcs->get_pcie_index_offset(adev); in amdgpu_device_indirect_rreg_ext() 1124 pcie_data = adev->nbio.funcs->get_pcie_data_offset(adev); in amdgpu_device_indirect_rreg_ext() 1128 if (unlikely(!adev->nbio.funcs)) in amdgpu_device_indirect_rreg_ext() 1131 pcie_index_hi = adev->nbio.funcs->get_pcie_index_hi_offset(adev); in amdgpu_device_indirect_rreg_ext() 1178 pcie_index = adev->nbio.funcs->get_pcie_index_offset(adev); in amdgpu_device_indirect_rreg64() 1179 pcie_data = adev->nbio.funcs->get_pcie_data_offset(adev); in amdgpu_device_indirect_rreg64() 1208 pcie_index = adev->nbio.funcs->get_pcie_index_offset(adev); in amdgpu_device_indirect_rreg64_ext() [all …]
|
| H A D | amdgpu_ras.c | 2395 if (adev->nbio.ras && in amdgpu_ras_interrupt_fatal_error_handler() 2396 adev->nbio.ras->handle_ras_controller_intr_no_bifring) in amdgpu_ras_interrupt_fatal_error_handler() 2397 adev->nbio.ras->handle_ras_controller_intr_no_bifring(adev); in amdgpu_ras_interrupt_fatal_error_handler() 2399 if (adev->nbio.ras && in amdgpu_ras_interrupt_fatal_error_handler() 2400 adev->nbio.ras->handle_ras_err_event_athub_intr_no_bifring) in amdgpu_ras_interrupt_fatal_error_handler() 2401 adev->nbio.ras->handle_ras_err_event_athub_intr_no_bifring(adev); in amdgpu_ras_interrupt_fatal_error_handler() 4318 adev->nbio.ras = &nbio_v7_4_ras; in amdgpu_ras_init() 4328 adev->nbio.ras = &nbio_v4_3_ras; in amdgpu_ras_init() 4339 adev->nbio.ras = &nbif_v6_3_1_ras; in amdgpu_ras_init() 4344 adev->nbio.ras = &nbio_v7_9_ras; in amdgpu_ras_init() [all …]
|
| H A D | gmc_v9_0.c | 1349 if (adev->nbio.funcs && adev->nbio.funcs->is_nps_switch_requested && in gmc_v9_0_need_reset_on_init() 1350 adev->nbio.funcs->is_nps_switch_requested(adev)) { in gmc_v9_0_need_reset_on_init() 1675 adev->nbio.funcs->get_memsize(adev) * 1024ULL * 1024ULL; in gmc_v9_0_mc_init()
|
| H A D | sdma_v6_0.c | 329 const struct nbio_hdp_flush_reg *nbio_hf_reg = adev->nbio.hdp_flush_reg; in sdma_v6_0_ring_emit_hdp_flush() 336 amdgpu_ring_write(ring, (adev->nbio.funcs->get_hdp_flush_done_offset(adev)) << 2); in sdma_v6_0_ring_emit_hdp_flush() 337 amdgpu_ring_write(ring, (adev->nbio.funcs->get_hdp_flush_req_offset(adev)) << 2); in sdma_v6_0_ring_emit_hdp_flush() 566 adev->nbio.funcs->sdma_doorbell_range(adev, i, ring->use_doorbell, in sdma_v6_0_gfx_resume_instance()
|
| H A D | sdma_v5_0.c | 493 const struct nbio_hdp_flush_reg *nbio_hf_reg = adev->nbio.hdp_flush_reg; in sdma_v5_0_ring_emit_hdp_flush() 503 amdgpu_ring_write(ring, (adev->nbio.funcs->get_hdp_flush_done_offset(adev)) << 2); in sdma_v5_0_ring_emit_hdp_flush() 504 amdgpu_ring_write(ring, (adev->nbio.funcs->get_hdp_flush_req_offset(adev)) << 2); in sdma_v5_0_ring_emit_hdp_flush() 782 adev->nbio.funcs->sdma_doorbell_range(adev, i, ring->use_doorbell, in sdma_v5_0_gfx_resume_instance()
|
| H A D | sdma_v5_2.c | 342 const struct nbio_hdp_flush_reg *nbio_hf_reg = adev->nbio.hdp_flush_reg; in sdma_v5_2_ring_emit_hdp_flush() 352 amdgpu_ring_write(ring, (adev->nbio.funcs->get_hdp_flush_done_offset(adev)) << 2); in sdma_v5_2_ring_emit_hdp_flush() 353 amdgpu_ring_write(ring, (adev->nbio.funcs->get_hdp_flush_req_offset(adev)) << 2); in sdma_v5_2_ring_emit_hdp_flush() 626 adev->nbio.funcs->sdma_doorbell_range(adev, i, ring->use_doorbell, in sdma_v5_2_gfx_resume_instance()
|
| H A D | amdgpu_gmc.c | 1549 if (adev->nbio.funcs && in amdgpu_gmc_get_memory_partition() 1550 adev->nbio.funcs->get_memory_partition_mode) in amdgpu_gmc_get_memory_partition() 1551 mode = adev->nbio.funcs->get_memory_partition_mode(adev, in amdgpu_gmc_get_memory_partition()
|
| H A D | vpe_v6_1.c | 258 adev->nbio.funcs->vpe_doorbell_range(adev, i, ring->use_doorbell, ring->doorbell_index + i*4, 4); in vpe_v6_1_ring_start()
|
| H A D | umsch_mm_v4_0.c | 225 adev->nbio.funcs->vcn_doorbell_range(adev, ring->use_doorbell, in umsch_mm_v4_0_ring_start()
|
| H A D | amdgpu.h | 1013 struct amdgpu_nbio nbio; member
|
| H A D | gfx_v12_0.c | 3683 if (adev->nbio.funcs->gc_doorbell_init) in gfx_v12_0_hw_init() 3684 adev->nbio.funcs->gc_doorbell_init(adev); in gfx_v12_0_hw_init() 4404 adev->nbio.funcs->get_hdp_flush_req_offset(adev), in gfx_v12_0_ring_emit_hdp_flush() 4405 adev->nbio.funcs->get_hdp_flush_done_offset(adev), in gfx_v12_0_ring_emit_hdp_flush()
|
| H A D | vcn_v4_0_5.c | 300 adev->nbio.funcs->vcn_doorbell_range(adev, ring->use_doorbell, in vcn_v4_0_5_hw_init()
|
| H A D | vcn_v2_5.c | 480 adev->nbio.funcs->vcn_doorbell_range(adev, ring->use_doorbell, in vcn_v2_5_hw_init()
|
| /linux/drivers/block/xen-blkback/ |
| H A D | blkback.c | 1298 int i, nbio = 0; in dispatch_rw_block_io() local 1436 biolist[nbio++] = bio; in dispatch_rw_block_io() 1451 biolist[nbio++] = bio; in dispatch_rw_block_io() 1456 atomic_set(&pending_req->pendcnt, nbio); in dispatch_rw_block_io() 1459 for (i = 0; i < nbio; i++) in dispatch_rw_block_io()
|