Home
last modified time | relevance | path

Searched refs:vinst (Results 1 – 25 of 25) sorted by relevance

/linux/arch/powerpc/platforms/powernv/
H A Dvas.c29 static int vas_irq_fault_window_setup(struct vas_instance *vinst) in vas_irq_fault_window_setup() argument
33 rc = request_threaded_irq(vinst->virq, vas_fault_handler, in vas_irq_fault_window_setup()
34 vas_fault_thread_fn, 0, vinst->name, vinst); in vas_irq_fault_window_setup()
38 vinst->vas_id, vinst->virq, rc); in vas_irq_fault_window_setup()
42 rc = vas_setup_fault_window(vinst); in vas_irq_fault_window_setup()
44 free_irq(vinst->virq, vinst); in vas_irq_fault_window_setup()
53 struct vas_instance *vinst; in init_vas_instance() local
77 vinst = kzalloc_obj(*vinst); in init_vas_instance()
78 if (!vinst) in init_vas_instance()
81 vinst->name = kasprintf(GFP_KERNEL, "vas-%d", vasid); in init_vas_instance()
[all …]
H A Dvas-fault.c29 static void dump_fifo(struct vas_instance *vinst, void *entry) in dump_fifo() argument
31 unsigned long *end = vinst->fault_fifo + vinst->fault_fifo_size; in dump_fifo()
35 pr_err("Fault fifo size %d, Max crbs %d\n", vinst->fault_fifo_size, in dump_fifo()
36 vinst->fault_fifo_size / CRB_SIZE); in dump_fifo()
68 struct vas_instance *vinst = data; in vas_fault_thread_fn() local
97 spin_lock_irqsave(&vinst->fault_lock, flags); in vas_fault_thread_fn()
104 fifo = vinst->fault_fifo + (vinst->fault_crbs * CRB_SIZE); in vas_fault_thread_fn()
109 vinst->fifo_in_progress = 0; in vas_fault_thread_fn()
110 spin_unlock_irqrestore(&vinst->fault_lock, flags); in vas_fault_thread_fn()
114 spin_unlock_irqrestore(&vinst->fault_lock, flags); in vas_fault_thread_fn()
[all …]
H A Dvas-window.c35 base = window->vinst->paste_base_addr; in vas_win_paste_addr()
36 shift = window->vinst->paste_win_id_shift; in vas_win_paste_addr()
51 pbaddr = window->vinst->hvwc_bar_start; in get_hvwc_mmio_bar()
61 pbaddr = window->vinst->uwc_bar_start; in get_uwc_mmio_bar()
78 name = kasprintf(GFP_KERNEL, "window-v%d-w%d", txwin->vinst->vas_id, in map_paste_region()
525 struct vas_instance *vinst = window->vinst; in vas_window_free() local
534 vas_release_window_id(&vinst->ida, winid); in vas_window_free()
537 static struct pnv_vas_window *vas_window_alloc(struct vas_instance *vinst) in vas_window_alloc() argument
542 winid = vas_assign_window_id(&vinst->ida); in vas_window_alloc()
550 window->vinst = vinst; in vas_window_alloc()
[all …]
H A Dvas-debug.c123 if (!window->vinst->dbgdir) in vas_window_init_dbgdir()
132 d = debugfs_create_dir(window->vas_win.dbgname, window->vinst->dbgdir); in vas_window_init_dbgdir()
139 void vas_instance_init_dbgdir(struct vas_instance *vinst) in vas_instance_init_dbgdir() argument
145 vinst->dbgname = kzalloc(16, GFP_KERNEL); in vas_instance_init_dbgdir()
146 if (!vinst->dbgname) in vas_instance_init_dbgdir()
149 snprintf(vinst->dbgname, 16, "v%d", vinst->vas_id); in vas_instance_init_dbgdir()
151 d = debugfs_create_dir(vinst->dbgname, vas_debugfs); in vas_instance_init_dbgdir()
152 vinst->dbgdir = d; in vas_instance_init_dbgdir()
H A Dvas.h354 struct vas_instance *vinst; member
424 extern void vas_instance_init_dbgdir(struct vas_instance *vinst);
427 extern int vas_setup_fault_window(struct vas_instance *vinst);
431 extern struct pnv_vas_window *vas_pswid_to_window(struct vas_instance *vinst,
H A Dvas-trace.h98 __entry->vasid = win->vinst->vas_id;
/linux/drivers/gpu/drm/amd/amdgpu/
H A Dvcn_v5_0_2.c45 static int vcn_v5_0_2_set_pg_state(struct amdgpu_vcn_inst *vinst,
247 struct amdgpu_vcn_inst *vinst = &adev->vcn.inst[i]; in vcn_v5_0_2_hw_fini() local
250 if (vinst->cur_state != AMD_PG_STATE_GATE) in vcn_v5_0_2_hw_fini()
251 vinst->set_pg_state(vinst, AMD_PG_STATE_GATE); in vcn_v5_0_2_hw_fini()
295 struct amdgpu_vcn_inst *vinst = &adev->vcn.inst[i]; in vcn_v5_0_2_resume() local
298 vinst->cur_state = AMD_PG_STATE_GATE; in vcn_v5_0_2_resume()
317 static void vcn_v5_0_2_mc_resume(struct amdgpu_vcn_inst *vinst) in vcn_v5_0_2_mc_resume() argument
319 struct amdgpu_device *adev = vinst->adev; in vcn_v5_0_2_mc_resume()
320 int inst = vinst->inst; in vcn_v5_0_2_mc_resume()
381 static void vcn_v5_0_2_mc_resume_dpg_mode(struct amdgpu_vcn_inst *vinst, in vcn_v5_0_2_mc_resume_dpg_mode() argument
[all …]
H A Dvcn_v5_0_0.c81 static int vcn_v5_0_0_set_pg_state(struct amdgpu_vcn_inst *vinst,
83 static int vcn_v5_0_0_pause_dpg_mode(struct amdgpu_vcn_inst *vinst,
288 struct amdgpu_vcn_inst *vinst = &adev->vcn.inst[i]; in vcn_v5_0_0_hw_fini() local
293 cancel_delayed_work_sync(&vinst->idle_work); in vcn_v5_0_0_hw_fini()
297 (vinst->cur_state != AMD_PG_STATE_GATE && in vcn_v5_0_0_hw_fini()
299 vinst->set_pg_state(vinst, AMD_PG_STATE_GATE); in vcn_v5_0_0_hw_fini()
362 static void vcn_v5_0_0_mc_resume(struct amdgpu_vcn_inst *vinst) in vcn_v5_0_0_mc_resume() argument
364 struct amdgpu_device *adev = vinst->adev; in vcn_v5_0_0_mc_resume()
365 int inst = vinst->inst; in vcn_v5_0_0_mc_resume()
424 static void vcn_v5_0_0_mc_resume_dpg_mode(struct amdgpu_vcn_inst *vinst, in vcn_v5_0_0_mc_resume_dpg_mode() argument
[all …]
H A Dvcn_v4_0_5.c99 static int vcn_v4_0_5_set_pg_state(struct amdgpu_vcn_inst *vinst,
101 static int vcn_v4_0_5_pause_dpg_mode(struct amdgpu_vcn_inst *vinst,
324 struct amdgpu_vcn_inst *vinst = &adev->vcn.inst[i]; in vcn_v4_0_5_hw_fini() local
329 cancel_delayed_work_sync(&vinst->idle_work); in vcn_v4_0_5_hw_fini()
333 (vinst->cur_state != AMD_PG_STATE_GATE && in vcn_v4_0_5_hw_fini()
335 vinst->set_pg_state(vinst, AMD_PG_STATE_GATE); in vcn_v4_0_5_hw_fini()
398 static void vcn_v4_0_5_mc_resume(struct amdgpu_vcn_inst *vinst) in vcn_v4_0_5_mc_resume() argument
400 struct amdgpu_device *adev = vinst->adev; in vcn_v4_0_5_mc_resume()
401 int inst = vinst->inst; in vcn_v4_0_5_mc_resume()
460 static void vcn_v4_0_5_mc_resume_dpg_mode(struct amdgpu_vcn_inst *vinst, in vcn_v4_0_5_mc_resume_dpg_mode() argument
[all …]
H A Dvcn_v4_0_3.c92 static int vcn_v4_0_3_set_pg_state(struct amdgpu_vcn_inst *vinst,
94 static int vcn_v4_0_3_pause_dpg_mode(struct amdgpu_vcn_inst *vinst,
309 static int vcn_v4_0_3_hw_init_inst(struct amdgpu_vcn_inst *vinst) in vcn_v4_0_3_hw_init_inst() argument
312 struct amdgpu_device *adev = vinst->adev; in vcn_v4_0_3_hw_init_inst()
314 int inst_idx = vinst->inst; in vcn_v4_0_3_hw_init_inst()
345 struct amdgpu_vcn_inst *vinst; in vcn_v4_0_3_hw_init() local
370 vinst = &adev->vcn.inst[i]; in vcn_v4_0_3_hw_init()
371 vcn_v4_0_3_hw_init_inst(vinst); in vcn_v4_0_3_hw_init()
400 struct amdgpu_vcn_inst *vinst = &adev->vcn.inst[i]; in vcn_v4_0_3_hw_fini() local
402 cancel_delayed_work_sync(&vinst->idle_work); in vcn_v4_0_3_hw_fini()
[all …]
H A Dvcn_v4_0.c100 static int vcn_v4_0_set_pg_state(struct amdgpu_vcn_inst *vinst,
102 static int vcn_v4_0_pause_dpg_mode(struct amdgpu_vcn_inst *vinst,
371 struct amdgpu_vcn_inst *vinst = &adev->vcn.inst[i]; in vcn_v4_0_hw_fini() local
376 cancel_delayed_work_sync(&vinst->idle_work); in vcn_v4_0_hw_fini()
380 (vinst->cur_state != AMD_PG_STATE_GATE && in vcn_v4_0_hw_fini()
382 vinst->set_pg_state(vinst, AMD_PG_STATE_GATE); in vcn_v4_0_hw_fini()
386 amdgpu_irq_put(adev, &vinst->ras_poison_irq, 0); in vcn_v4_0_hw_fini()
447 static void vcn_v4_0_mc_resume(struct amdgpu_vcn_inst *vinst) in vcn_v4_0_mc_resume() argument
449 struct amdgpu_device *adev = vinst->adev; in vcn_v4_0_mc_resume()
450 int inst = vinst->inst; in vcn_v4_0_mc_resume()
[all …]
H A Dvcn_v5_0_1.c80 static int vcn_v5_0_1_set_pg_state(struct amdgpu_vcn_inst *vinst,
347 struct amdgpu_vcn_inst *vinst = &adev->vcn.inst[i]; in vcn_v5_0_1_hw_fini() local
350 if (vinst->cur_state != AMD_PG_STATE_GATE) in vcn_v5_0_1_hw_fini()
351 vinst->set_pg_state(vinst, AMD_PG_STATE_GATE); in vcn_v5_0_1_hw_fini()
398 struct amdgpu_vcn_inst *vinst = &adev->vcn.inst[i]; in vcn_v5_0_1_resume() local
401 vinst->cur_state = AMD_PG_STATE_GATE; in vcn_v5_0_1_resume()
420 static void vcn_v5_0_1_mc_resume(struct amdgpu_vcn_inst *vinst) in vcn_v5_0_1_mc_resume() argument
422 struct amdgpu_device *adev = vinst->adev; in vcn_v5_0_1_mc_resume()
423 int inst = vinst->inst; in vcn_v5_0_1_mc_resume()
484 static void vcn_v5_0_1_mc_resume_dpg_mode(struct amdgpu_vcn_inst *vinst, in vcn_v5_0_1_mc_resume_dpg_mode() argument
[all …]
H A Dvcn_v3_0.c109 static int vcn_v3_0_set_pg_state(struct amdgpu_vcn_inst *vinst,
111 static int vcn_v3_0_pause_dpg_mode(struct amdgpu_vcn_inst *vinst,
113 static int vcn_v3_0_reset(struct amdgpu_vcn_inst *vinst);
446 struct amdgpu_vcn_inst *vinst = &adev->vcn.inst[i]; in vcn_v3_0_hw_fini() local
451 cancel_delayed_work_sync(&vinst->idle_work); in vcn_v3_0_hw_fini()
455 (vinst->cur_state != AMD_PG_STATE_GATE && in vcn_v3_0_hw_fini()
457 vinst->set_pg_state(vinst, AMD_PG_STATE_GATE); in vcn_v3_0_hw_fini()
520 static void vcn_v3_0_mc_resume(struct amdgpu_vcn_inst *vinst) in vcn_v3_0_mc_resume() argument
522 struct amdgpu_device *adev = vinst->adev; in vcn_v3_0_mc_resume()
523 int inst = vinst->inst; in vcn_v3_0_mc_resume()
[all …]
H A Dvcn_v2_5.c99 static int vcn_v2_5_set_pg_state(struct amdgpu_vcn_inst *vinst,
101 static int vcn_v2_5_pause_dpg_mode(struct amdgpu_vcn_inst *vinst,
105 static int vcn_v2_5_reset(struct amdgpu_vcn_inst *vinst);
512 struct amdgpu_vcn_inst *vinst = &adev->vcn.inst[i]; in vcn_v2_5_hw_fini() local
517 cancel_delayed_work_sync(&vinst->idle_work); in vcn_v2_5_hw_fini()
520 (vinst->cur_state != AMD_PG_STATE_GATE && in vcn_v2_5_hw_fini()
522 vinst->set_pg_state(vinst, AMD_PG_STATE_GATE); in vcn_v2_5_hw_fini()
527 amdgpu_irq_put(adev, &vinst->ras_poison_irq, 0); in vcn_v2_5_hw_fini()
588 static void vcn_v2_5_mc_resume(struct amdgpu_vcn_inst *vinst) in vcn_v2_5_mc_resume() argument
590 struct amdgpu_device *adev = vinst->adev; in vcn_v2_5_mc_resume()
[all …]
H A Dvcn_v2_0.c96 static int vcn_v2_0_set_pg_state(struct amdgpu_vcn_inst *vinst,
98 static int vcn_v2_0_pause_dpg_mode(struct amdgpu_vcn_inst *vinst,
101 static int vcn_v2_0_reset(struct amdgpu_vcn_inst *vinst);
322 struct amdgpu_vcn_inst *vinst = adev->vcn.inst; in vcn_v2_0_hw_fini() local
324 cancel_delayed_work_sync(&vinst->idle_work); in vcn_v2_0_hw_fini()
327 (vinst->cur_state != AMD_PG_STATE_GATE && in vcn_v2_0_hw_fini()
329 vinst->set_pg_state(vinst, AMD_PG_STATE_GATE); in vcn_v2_0_hw_fini()
381 static void vcn_v2_0_mc_resume(struct amdgpu_vcn_inst *vinst) in vcn_v2_0_mc_resume() argument
383 struct amdgpu_device *adev = vinst->adev; in vcn_v2_0_mc_resume()
438 static void vcn_v2_0_mc_resume_dpg_mode(struct amdgpu_vcn_inst *vinst, in vcn_v2_0_mc_resume_dpg_mode() argument
[all …]
H A Damdgpu_vcn.c1480 struct amdgpu_vcn_inst *vinst = &adev->vcn.inst[i]; in vcn_set_powergating_state() local
1482 ret |= vinst->set_pg_state(vinst, state); in vcn_set_powergating_state()
1498 struct amdgpu_vcn_inst *vinst = &adev->vcn.inst[instance_id]; in amdgpu_vcn_reset_engine() local
1501 mutex_lock(&vinst->engine_reset_mutex); in amdgpu_vcn_reset_engine()
1506 drm_sched_wqueue_stop(&vinst->ring_dec.sched); in amdgpu_vcn_reset_engine()
1507 for (i = 0; i < vinst->num_enc_rings; i++) in amdgpu_vcn_reset_engine()
1508 drm_sched_wqueue_stop(&vinst->ring_enc[i].sched); in amdgpu_vcn_reset_engine()
1511 r = vinst->reset(vinst); in amdgpu_vcn_reset_engine()
1514 r = amdgpu_ring_test_ring(&vinst->ring_dec); in amdgpu_vcn_reset_engine()
1517 for (i = 0; i < vinst->num_enc_rings; i++) { in amdgpu_vcn_reset_engine()
[all …]
H A Djpeg_v5_0_1.c848 struct amdgpu_vcn_inst *vinst = &adev->vcn.inst[ring->me]; in jpeg_v5_0_1_ring_reset() local
852 mutex_lock(&vinst->engine_reset_mutex); in jpeg_v5_0_1_ring_reset()
858 mutex_unlock(&vinst->engine_reset_mutex); in jpeg_v5_0_1_ring_reset()
H A Djpeg_v4_0_3.c1189 struct amdgpu_vcn_inst *vinst = &adev->vcn.inst[ring->me]; in jpeg_v4_0_3_ring_reset() local
1193 mutex_lock(&vinst->engine_reset_mutex); in jpeg_v4_0_3_ring_reset()
1198 mutex_unlock(&vinst->engine_reset_mutex); in jpeg_v4_0_3_ring_reset()
/linux/drivers/net/ethernet/microchip/sparx5/lan969x/
H A Dlan969x_vcap_impl.c9 .vinst = 0,
21 .vinst = 1,
33 .vinst = 2,
45 .vinst = 0,
57 .vinst = 1,
/linux/drivers/net/ethernet/microchip/vcap/
H A Dvcap_api_debugfs.c282 out->prf(out->dst, "vinst: %d\n", admin->vinst); in vcap_show_admin_info()
378 if (admin->vinst) in vcap_port_debugfs_show()
450 admin->vinst); in vcap_debugfs()
459 admin->vinst); in vcap_debugfs()
H A Dvcap_api_kunit.c77 if (admin->vinst == 0 || admin->vinst == 2) in test_add_def_fields()
1879 .vinst = 0, in vcap_api_next_lookup_basic_test()
1887 .vinst = 1, in vcap_api_next_lookup_basic_test()
1926 .vinst = 0, in vcap_api_next_lookup_advanced_test()
1933 .vinst = 1, in vcap_api_next_lookup_advanced_test()
1940 .vinst = 2, in vcap_api_next_lookup_advanced_test()
1947 .vinst = 0, in vcap_api_next_lookup_advanced_test()
1954 .vinst = 1, in vcap_api_next_lookup_advanced_test()
2194 .vinst = 0, in vcap_api_rule_chain_path_test()
H A Dvcap_api_debugfs_kunit.c84 if (admin->vinst == 0 || admin->vinst == 2) in test_add_def_fields()
H A Dvcap_api.h169 int vinst; /* instance number within the same type */ member
H A Dvcap_api.c1607 int lookup_first = admin->vinst * admin->lookups_per_instance; in vcap_chain_id_to_lookup()
/linux/drivers/net/ethernet/microchip/lan966x/
H A Dlan966x_vcap_impl.c624 admin->vinst = 0; in lan966x_vcap_admin_alloc()