| /linux/arch/powerpc/platforms/powernv/ |
| H A D | vas.c | 29 static int vas_irq_fault_window_setup(struct vas_instance *vinst) in vas_irq_fault_window_setup() argument 33 rc = request_threaded_irq(vinst->virq, vas_fault_handler, in vas_irq_fault_window_setup() 34 vas_fault_thread_fn, 0, vinst->name, vinst); in vas_irq_fault_window_setup() 38 vinst->vas_id, vinst->virq, rc); in vas_irq_fault_window_setup() 42 rc = vas_setup_fault_window(vinst); in vas_irq_fault_window_setup() 44 free_irq(vinst->virq, vinst); in vas_irq_fault_window_setup() 53 struct vas_instance *vinst; in init_vas_instance() local 77 vinst = kzalloc_obj(*vinst); in init_vas_instance() 78 if (!vinst) in init_vas_instance() 81 vinst->name = kasprintf(GFP_KERNEL, "vas-%d", vasid); in init_vas_instance() [all …]
|
| H A D | vas-fault.c | 29 static void dump_fifo(struct vas_instance *vinst, void *entry) in dump_fifo() argument 31 unsigned long *end = vinst->fault_fifo + vinst->fault_fifo_size; in dump_fifo() 35 pr_err("Fault fifo size %d, Max crbs %d\n", vinst->fault_fifo_size, in dump_fifo() 36 vinst->fault_fifo_size / CRB_SIZE); in dump_fifo() 68 struct vas_instance *vinst = data; in vas_fault_thread_fn() local 97 spin_lock_irqsave(&vinst->fault_lock, flags); in vas_fault_thread_fn() 104 fifo = vinst->fault_fifo + (vinst->fault_crbs * CRB_SIZE); in vas_fault_thread_fn() 109 vinst->fifo_in_progress = 0; in vas_fault_thread_fn() 110 spin_unlock_irqrestore(&vinst->fault_lock, flags); in vas_fault_thread_fn() 114 spin_unlock_irqrestore(&vinst->fault_lock, flags); in vas_fault_thread_fn() [all …]
|
| H A D | vas-window.c | 35 base = window->vinst->paste_base_addr; in vas_win_paste_addr() 36 shift = window->vinst->paste_win_id_shift; in vas_win_paste_addr() 51 pbaddr = window->vinst->hvwc_bar_start; in get_hvwc_mmio_bar() 61 pbaddr = window->vinst->uwc_bar_start; in get_uwc_mmio_bar() 78 name = kasprintf(GFP_KERNEL, "window-v%d-w%d", txwin->vinst->vas_id, in map_paste_region() 525 struct vas_instance *vinst = window->vinst; in vas_window_free() local 534 vas_release_window_id(&vinst->ida, winid); in vas_window_free() 537 static struct pnv_vas_window *vas_window_alloc(struct vas_instance *vinst) in vas_window_alloc() argument 542 winid = vas_assign_window_id(&vinst->ida); in vas_window_alloc() 550 window->vinst = vinst; in vas_window_alloc() [all …]
|
| H A D | vas-debug.c | 123 if (!window->vinst->dbgdir) in vas_window_init_dbgdir() 132 d = debugfs_create_dir(window->vas_win.dbgname, window->vinst->dbgdir); in vas_window_init_dbgdir() 139 void vas_instance_init_dbgdir(struct vas_instance *vinst) in vas_instance_init_dbgdir() argument 145 vinst->dbgname = kzalloc(16, GFP_KERNEL); in vas_instance_init_dbgdir() 146 if (!vinst->dbgname) in vas_instance_init_dbgdir() 149 snprintf(vinst->dbgname, 16, "v%d", vinst->vas_id); in vas_instance_init_dbgdir() 151 d = debugfs_create_dir(vinst->dbgname, vas_debugfs); in vas_instance_init_dbgdir() 152 vinst->dbgdir = d; in vas_instance_init_dbgdir()
|
| H A D | vas.h | 354 struct vas_instance *vinst; member 424 extern void vas_instance_init_dbgdir(struct vas_instance *vinst); 427 extern int vas_setup_fault_window(struct vas_instance *vinst); 431 extern struct pnv_vas_window *vas_pswid_to_window(struct vas_instance *vinst,
|
| H A D | vas-trace.h | 98 __entry->vasid = win->vinst->vas_id;
|
| /linux/drivers/gpu/drm/amd/amdgpu/ |
| H A D | vcn_v5_0_2.c | 45 static int vcn_v5_0_2_set_pg_state(struct amdgpu_vcn_inst *vinst, 247 struct amdgpu_vcn_inst *vinst = &adev->vcn.inst[i]; in vcn_v5_0_2_hw_fini() local 250 if (vinst->cur_state != AMD_PG_STATE_GATE) in vcn_v5_0_2_hw_fini() 251 vinst->set_pg_state(vinst, AMD_PG_STATE_GATE); in vcn_v5_0_2_hw_fini() 295 struct amdgpu_vcn_inst *vinst = &adev->vcn.inst[i]; in vcn_v5_0_2_resume() local 298 vinst->cur_state = AMD_PG_STATE_GATE; in vcn_v5_0_2_resume() 317 static void vcn_v5_0_2_mc_resume(struct amdgpu_vcn_inst *vinst) in vcn_v5_0_2_mc_resume() argument 319 struct amdgpu_device *adev = vinst->adev; in vcn_v5_0_2_mc_resume() 320 int inst = vinst->inst; in vcn_v5_0_2_mc_resume() 381 static void vcn_v5_0_2_mc_resume_dpg_mode(struct amdgpu_vcn_inst *vinst, in vcn_v5_0_2_mc_resume_dpg_mode() argument [all …]
|
| H A D | vcn_v5_0_0.c | 81 static int vcn_v5_0_0_set_pg_state(struct amdgpu_vcn_inst *vinst, 83 static int vcn_v5_0_0_pause_dpg_mode(struct amdgpu_vcn_inst *vinst, 288 struct amdgpu_vcn_inst *vinst = &adev->vcn.inst[i]; in vcn_v5_0_0_hw_fini() local 293 cancel_delayed_work_sync(&vinst->idle_work); in vcn_v5_0_0_hw_fini() 297 (vinst->cur_state != AMD_PG_STATE_GATE && in vcn_v5_0_0_hw_fini() 299 vinst->set_pg_state(vinst, AMD_PG_STATE_GATE); in vcn_v5_0_0_hw_fini() 362 static void vcn_v5_0_0_mc_resume(struct amdgpu_vcn_inst *vinst) in vcn_v5_0_0_mc_resume() argument 364 struct amdgpu_device *adev = vinst->adev; in vcn_v5_0_0_mc_resume() 365 int inst = vinst->inst; in vcn_v5_0_0_mc_resume() 424 static void vcn_v5_0_0_mc_resume_dpg_mode(struct amdgpu_vcn_inst *vinst, in vcn_v5_0_0_mc_resume_dpg_mode() argument [all …]
|
| H A D | vcn_v4_0_5.c | 99 static int vcn_v4_0_5_set_pg_state(struct amdgpu_vcn_inst *vinst, 101 static int vcn_v4_0_5_pause_dpg_mode(struct amdgpu_vcn_inst *vinst, 324 struct amdgpu_vcn_inst *vinst = &adev->vcn.inst[i]; in vcn_v4_0_5_hw_fini() local 329 cancel_delayed_work_sync(&vinst->idle_work); in vcn_v4_0_5_hw_fini() 333 (vinst->cur_state != AMD_PG_STATE_GATE && in vcn_v4_0_5_hw_fini() 335 vinst->set_pg_state(vinst, AMD_PG_STATE_GATE); in vcn_v4_0_5_hw_fini() 398 static void vcn_v4_0_5_mc_resume(struct amdgpu_vcn_inst *vinst) in vcn_v4_0_5_mc_resume() argument 400 struct amdgpu_device *adev = vinst->adev; in vcn_v4_0_5_mc_resume() 401 int inst = vinst->inst; in vcn_v4_0_5_mc_resume() 460 static void vcn_v4_0_5_mc_resume_dpg_mode(struct amdgpu_vcn_inst *vinst, in vcn_v4_0_5_mc_resume_dpg_mode() argument [all …]
|
| H A D | vcn_v4_0_3.c | 92 static int vcn_v4_0_3_set_pg_state(struct amdgpu_vcn_inst *vinst, 94 static int vcn_v4_0_3_pause_dpg_mode(struct amdgpu_vcn_inst *vinst, 309 static int vcn_v4_0_3_hw_init_inst(struct amdgpu_vcn_inst *vinst) in vcn_v4_0_3_hw_init_inst() argument 312 struct amdgpu_device *adev = vinst->adev; in vcn_v4_0_3_hw_init_inst() 314 int inst_idx = vinst->inst; in vcn_v4_0_3_hw_init_inst() 345 struct amdgpu_vcn_inst *vinst; in vcn_v4_0_3_hw_init() local 370 vinst = &adev->vcn.inst[i]; in vcn_v4_0_3_hw_init() 371 vcn_v4_0_3_hw_init_inst(vinst); in vcn_v4_0_3_hw_init() 400 struct amdgpu_vcn_inst *vinst = &adev->vcn.inst[i]; in vcn_v4_0_3_hw_fini() local 402 cancel_delayed_work_sync(&vinst->idle_work); in vcn_v4_0_3_hw_fini() [all …]
|
| H A D | vcn_v4_0.c | 100 static int vcn_v4_0_set_pg_state(struct amdgpu_vcn_inst *vinst, 102 static int vcn_v4_0_pause_dpg_mode(struct amdgpu_vcn_inst *vinst, 371 struct amdgpu_vcn_inst *vinst = &adev->vcn.inst[i]; in vcn_v4_0_hw_fini() local 376 cancel_delayed_work_sync(&vinst->idle_work); in vcn_v4_0_hw_fini() 380 (vinst->cur_state != AMD_PG_STATE_GATE && in vcn_v4_0_hw_fini() 382 vinst->set_pg_state(vinst, AMD_PG_STATE_GATE); in vcn_v4_0_hw_fini() 386 amdgpu_irq_put(adev, &vinst->ras_poison_irq, 0); in vcn_v4_0_hw_fini() 447 static void vcn_v4_0_mc_resume(struct amdgpu_vcn_inst *vinst) in vcn_v4_0_mc_resume() argument 449 struct amdgpu_device *adev = vinst->adev; in vcn_v4_0_mc_resume() 450 int inst = vinst->inst; in vcn_v4_0_mc_resume() [all …]
|
| H A D | vcn_v5_0_1.c | 80 static int vcn_v5_0_1_set_pg_state(struct amdgpu_vcn_inst *vinst, 347 struct amdgpu_vcn_inst *vinst = &adev->vcn.inst[i]; in vcn_v5_0_1_hw_fini() local 350 if (vinst->cur_state != AMD_PG_STATE_GATE) in vcn_v5_0_1_hw_fini() 351 vinst->set_pg_state(vinst, AMD_PG_STATE_GATE); in vcn_v5_0_1_hw_fini() 398 struct amdgpu_vcn_inst *vinst = &adev->vcn.inst[i]; in vcn_v5_0_1_resume() local 401 vinst->cur_state = AMD_PG_STATE_GATE; in vcn_v5_0_1_resume() 420 static void vcn_v5_0_1_mc_resume(struct amdgpu_vcn_inst *vinst) in vcn_v5_0_1_mc_resume() argument 422 struct amdgpu_device *adev = vinst->adev; in vcn_v5_0_1_mc_resume() 423 int inst = vinst->inst; in vcn_v5_0_1_mc_resume() 484 static void vcn_v5_0_1_mc_resume_dpg_mode(struct amdgpu_vcn_inst *vinst, in vcn_v5_0_1_mc_resume_dpg_mode() argument [all …]
|
| H A D | vcn_v3_0.c | 109 static int vcn_v3_0_set_pg_state(struct amdgpu_vcn_inst *vinst, 111 static int vcn_v3_0_pause_dpg_mode(struct amdgpu_vcn_inst *vinst, 113 static int vcn_v3_0_reset(struct amdgpu_vcn_inst *vinst); 446 struct amdgpu_vcn_inst *vinst = &adev->vcn.inst[i]; in vcn_v3_0_hw_fini() local 451 cancel_delayed_work_sync(&vinst->idle_work); in vcn_v3_0_hw_fini() 455 (vinst->cur_state != AMD_PG_STATE_GATE && in vcn_v3_0_hw_fini() 457 vinst->set_pg_state(vinst, AMD_PG_STATE_GATE); in vcn_v3_0_hw_fini() 520 static void vcn_v3_0_mc_resume(struct amdgpu_vcn_inst *vinst) in vcn_v3_0_mc_resume() argument 522 struct amdgpu_device *adev = vinst->adev; in vcn_v3_0_mc_resume() 523 int inst = vinst->inst; in vcn_v3_0_mc_resume() [all …]
|
| H A D | vcn_v2_5.c | 99 static int vcn_v2_5_set_pg_state(struct amdgpu_vcn_inst *vinst, 101 static int vcn_v2_5_pause_dpg_mode(struct amdgpu_vcn_inst *vinst, 105 static int vcn_v2_5_reset(struct amdgpu_vcn_inst *vinst); 512 struct amdgpu_vcn_inst *vinst = &adev->vcn.inst[i]; in vcn_v2_5_hw_fini() local 517 cancel_delayed_work_sync(&vinst->idle_work); in vcn_v2_5_hw_fini() 520 (vinst->cur_state != AMD_PG_STATE_GATE && in vcn_v2_5_hw_fini() 522 vinst->set_pg_state(vinst, AMD_PG_STATE_GATE); in vcn_v2_5_hw_fini() 527 amdgpu_irq_put(adev, &vinst->ras_poison_irq, 0); in vcn_v2_5_hw_fini() 588 static void vcn_v2_5_mc_resume(struct amdgpu_vcn_inst *vinst) in vcn_v2_5_mc_resume() argument 590 struct amdgpu_device *adev = vinst->adev; in vcn_v2_5_mc_resume() [all …]
|
| H A D | vcn_v2_0.c | 96 static int vcn_v2_0_set_pg_state(struct amdgpu_vcn_inst *vinst, 98 static int vcn_v2_0_pause_dpg_mode(struct amdgpu_vcn_inst *vinst, 101 static int vcn_v2_0_reset(struct amdgpu_vcn_inst *vinst); 322 struct amdgpu_vcn_inst *vinst = adev->vcn.inst; in vcn_v2_0_hw_fini() local 324 cancel_delayed_work_sync(&vinst->idle_work); in vcn_v2_0_hw_fini() 327 (vinst->cur_state != AMD_PG_STATE_GATE && in vcn_v2_0_hw_fini() 329 vinst->set_pg_state(vinst, AMD_PG_STATE_GATE); in vcn_v2_0_hw_fini() 381 static void vcn_v2_0_mc_resume(struct amdgpu_vcn_inst *vinst) in vcn_v2_0_mc_resume() argument 383 struct amdgpu_device *adev = vinst->adev; in vcn_v2_0_mc_resume() 438 static void vcn_v2_0_mc_resume_dpg_mode(struct amdgpu_vcn_inst *vinst, in vcn_v2_0_mc_resume_dpg_mode() argument [all …]
|
| H A D | amdgpu_vcn.c | 1480 struct amdgpu_vcn_inst *vinst = &adev->vcn.inst[i]; in vcn_set_powergating_state() local 1482 ret |= vinst->set_pg_state(vinst, state); in vcn_set_powergating_state() 1498 struct amdgpu_vcn_inst *vinst = &adev->vcn.inst[instance_id]; in amdgpu_vcn_reset_engine() local 1501 mutex_lock(&vinst->engine_reset_mutex); in amdgpu_vcn_reset_engine() 1506 drm_sched_wqueue_stop(&vinst->ring_dec.sched); in amdgpu_vcn_reset_engine() 1507 for (i = 0; i < vinst->num_enc_rings; i++) in amdgpu_vcn_reset_engine() 1508 drm_sched_wqueue_stop(&vinst->ring_enc[i].sched); in amdgpu_vcn_reset_engine() 1511 r = vinst->reset(vinst); in amdgpu_vcn_reset_engine() 1514 r = amdgpu_ring_test_ring(&vinst->ring_dec); in amdgpu_vcn_reset_engine() 1517 for (i = 0; i < vinst->num_enc_rings; i++) { in amdgpu_vcn_reset_engine() [all …]
|
| H A D | jpeg_v5_0_1.c | 848 struct amdgpu_vcn_inst *vinst = &adev->vcn.inst[ring->me]; in jpeg_v5_0_1_ring_reset() local 852 mutex_lock(&vinst->engine_reset_mutex); in jpeg_v5_0_1_ring_reset() 858 mutex_unlock(&vinst->engine_reset_mutex); in jpeg_v5_0_1_ring_reset()
|
| H A D | jpeg_v4_0_3.c | 1189 struct amdgpu_vcn_inst *vinst = &adev->vcn.inst[ring->me]; in jpeg_v4_0_3_ring_reset() local 1193 mutex_lock(&vinst->engine_reset_mutex); in jpeg_v4_0_3_ring_reset() 1198 mutex_unlock(&vinst->engine_reset_mutex); in jpeg_v4_0_3_ring_reset()
|
| /linux/drivers/net/ethernet/microchip/sparx5/lan969x/ |
| H A D | lan969x_vcap_impl.c | 9 .vinst = 0, 21 .vinst = 1, 33 .vinst = 2, 45 .vinst = 0, 57 .vinst = 1,
|
| /linux/drivers/net/ethernet/microchip/vcap/ |
| H A D | vcap_api_debugfs.c | 282 out->prf(out->dst, "vinst: %d\n", admin->vinst); in vcap_show_admin_info() 378 if (admin->vinst) in vcap_port_debugfs_show() 450 admin->vinst); in vcap_debugfs() 459 admin->vinst); in vcap_debugfs()
|
| H A D | vcap_api_kunit.c | 77 if (admin->vinst == 0 || admin->vinst == 2) in test_add_def_fields() 1879 .vinst = 0, in vcap_api_next_lookup_basic_test() 1887 .vinst = 1, in vcap_api_next_lookup_basic_test() 1926 .vinst = 0, in vcap_api_next_lookup_advanced_test() 1933 .vinst = 1, in vcap_api_next_lookup_advanced_test() 1940 .vinst = 2, in vcap_api_next_lookup_advanced_test() 1947 .vinst = 0, in vcap_api_next_lookup_advanced_test() 1954 .vinst = 1, in vcap_api_next_lookup_advanced_test() 2194 .vinst = 0, in vcap_api_rule_chain_path_test()
|
| H A D | vcap_api_debugfs_kunit.c | 84 if (admin->vinst == 0 || admin->vinst == 2) in test_add_def_fields()
|
| H A D | vcap_api.h | 169 int vinst; /* instance number within the same type */ member
|
| H A D | vcap_api.c | 1607 int lookup_first = admin->vinst * admin->lookups_per_instance; in vcap_chain_id_to_lookup()
|
| /linux/drivers/net/ethernet/microchip/lan966x/ |
| H A D | lan966x_vcap_impl.c | 624 admin->vinst = 0; in lan966x_vcap_admin_alloc()
|