Home
last modified time | relevance | path

Searched refs:kfd_node (Results 1 – 25 of 28) sorted by relevance

12

/linux/drivers/gpu/drm/amd/amdkfd/
H A Dkfd_priv.h221 struct kfd_node;
224 bool (*interrupt_isr)(struct kfd_node *dev,
227 void (*interrupt_wq)(struct kfd_node *dev,
247 unsigned int kfd_get_num_sdma_engines(struct kfd_node *kdev);
248 unsigned int kfd_get_num_xgmi_sdma_engines(struct kfd_node *kdev);
268 struct kfd_node { struct
372 struct kfd_node *nodes[MAX_KFD_NODES];
619 struct kfd_node *device;
757 struct kfd_node *dev;
1057 bool kfd_dev_is_large_bar(struct kfd_node *dev);
[all …]
H A Dkfd_smi_events.c40 struct kfd_node *dev;
153 struct kfd_node *dev = client->dev; in kfd_smi_ev_release()
174 static void add_event_to_kfifo(pid_t pid, struct kfd_node *dev, in add_event_to_kfifo()
199 static void kfd_smi_event_add(pid_t pid, struct kfd_node *dev, in kfd_smi_event_add()
218 void kfd_smi_event_update_gpu_reset(struct kfd_node *dev, bool post_reset, in kfd_smi_event_update_gpu_reset()
241 void kfd_smi_event_update_thermal_throttling(struct kfd_node *dev, in kfd_smi_event_update_thermal_throttling()
249 void kfd_smi_event_update_vmfault(struct kfd_node *dev, uint16_t pasid) in kfd_smi_event_update_vmfault()
263 void kfd_smi_event_page_fault_start(struct kfd_node *node, pid_t pid, in kfd_smi_event_page_fault_start()
272 void kfd_smi_event_page_fault_end(struct kfd_node *node, pid_t pid, in kfd_smi_event_page_fault_end()
280 void kfd_smi_event_migration_start(struct kfd_node *node, pid_t pid, in kfd_smi_event_migration_start()
[all …]
H A Dkfd_interrupt.c53 int kfd_interrupt_init(struct kfd_node *node) in kfd_interrupt_init()
90 void kfd_interrupt_exit(struct kfd_node *node) in kfd_interrupt_exit()
108 bool enqueue_ih_ring_entry(struct kfd_node *node, const void *ih_ring_entry) in enqueue_ih_ring_entry()
123 static bool dequeue_ih_ring_entry(struct kfd_node *node, u32 **ih_ring_entry) in dequeue_ih_ring_entry()
138 struct kfd_node *dev = container_of(work, struct kfd_node, interrupt_work); in interrupt_wq()
157 bool interrupt_is_wanted(struct kfd_node *dev, in interrupt_is_wanted()
H A Dkfd_debug.h35 bool kfd_set_dbg_ev_from_interrupt(struct kfd_node *dev,
42 struct kfd_process *process, struct kfd_node *dev,
77 static inline bool kfd_dbg_is_per_vmid_supported(struct kfd_node *dev) in kfd_dbg_is_per_vmid_supported()
100 static inline bool kfd_dbg_is_rlc_restore_supported(struct kfd_node *dev) in kfd_dbg_is_rlc_restore_supported()
106 static inline bool kfd_dbg_has_cwsr_workaround(struct kfd_node *dev) in kfd_dbg_has_cwsr_workaround()
112 static inline bool kfd_dbg_has_gws_support(struct kfd_node *dev) in kfd_dbg_has_gws_support()
132 static inline bool kfd_dbg_has_ttmps_always_setup(struct kfd_node *dev) in kfd_dbg_has_ttmps_always_setup()
H A Dkfd_device.c66 static int kfd_resume(struct kfd_node *kfd);
564 static int kfd_gws_init(struct kfd_node *node) in kfd_gws_init()
601 static void kfd_smi_init(struct kfd_node *dev) in kfd_smi_init()
607 static int kfd_init_node(struct kfd_node *node) in kfd_init_node()
657 struct kfd_node *knode; in kfd_cleanup_nodes()
680 static void kfd_setup_interrupt_bitmap(struct kfd_node *node, in kfd_setup_interrupt_bitmap()
733 struct kfd_node *node; in kgd2kfd_device_init()
867 node = kzalloc_obj(struct kfd_node); in kgd2kfd_device_init()
985 struct kfd_node *node; in kgd2kfd_pre_reset()
1013 struct kfd_node *node; in kgd2kfd_post_reset()
[all …]
H A Dkfd_mqd_manager.c52 struct kfd_node *dev = mm->dev; in allocate_hiq_mqd()
69 struct kfd_node *dev = mm->dev; in allocate_sdma_mqd()
273 uint64_t kfd_hiq_mqd_stride(struct kfd_node *dev) in kfd_hiq_mqd_stride()
278 void kfd_get_hiq_xcc_mqd(struct kfd_node *dev, struct kfd_mem_obj *mqd_mem_obj, in kfd_get_hiq_xcc_mqd()
301 bool kfd_check_hiq_mqd_doorbell_id(struct kfd_node *node, uint32_t doorbell_id, in kfd_check_hiq_mqd_doorbell_id()
H A Dkfd_crat.h298 struct kfd_node;
311 int kfd_get_gpu_cache_info(struct kfd_node *kdev, struct kfd_gpu_cache_info **pcache_info);
317 int flags, struct kfd_node *kdev,
H A Dcik_event_interrupt.c29 static bool cik_event_interrupt_isr(struct kfd_node *dev, in cik_event_interrupt_isr()
88 static void cik_event_interrupt_wq(struct kfd_node *dev, in cik_event_interrupt_wq()
H A Dkfd_process.c270 struct kfd_node *dev = NULL; in kfd_get_cu_occupancy()
720 struct kfd_node *dev = pdd->dev; in kfd_process_free_gpuvm()
742 struct kfd_node *kdev = pdd->dev; in kfd_process_alloc_gpuvm()
1075 struct kfd_node *kdev; in kfd_process_kunmap_signal_bo()
1419 struct kfd_node *dev = p->pdds[i]->dev; in kfd_process_init_cwsr_apu()
1456 struct kfd_node *dev = pdd->dev; in kfd_process_device_init_cwsr_dgpu()
1492 struct kfd_node *dev = pdd->dev; in kfd_process_device_destroy_cwsr_dgpu()
1539 struct kfd_node *dev = p->pdds[i]->dev; in kfd_process_xnack_mode()
1692 struct kfd_process_device *kfd_get_process_device_data(struct kfd_node *dev, in kfd_get_process_device_data()
1704 struct kfd_process_device *kfd_create_process_device_data(struct kfd_node *dev, in kfd_create_process_device_data()
[all …]
H A Dkfd_kernel_queue.h56 struct kfd_node *dev;
H A Dkfd_process_queue_manager.c85 struct kfd_node *dev = pdd->dev; in kfd_process_dequeue_from_device()
108 struct kfd_node *dev = NULL; in pqm_set_gws()
191 struct kfd_node *dev; in pqm_clean_queue_resource()
246 struct kfd_node *dev, struct queue **q, in init_user_queue()
310 struct kfd_node *dev, in pqm_create_queue()
497 struct kfd_node *dev; in pqm_destroy_queue()
H A Dkfd_crat.c1579 int kfd_get_gpu_cache_info(struct kfd_node *kdev, struct kfd_gpu_cache_info **pcache_info) in kfd_get_gpu_cache_info()
1984 struct kfd_node *kdev, uint8_t type, uint64_t size, in kfd_fill_gpu_memory_affinity()
2013 static void kfd_find_numa_node_in_srat(struct kfd_node *kdev) in kfd_find_numa_node_in_srat()
2111 struct kfd_node *kdev, in kfd_fill_gpu_direct_io_link_to_cpu()
2183 struct kfd_node *kdev, in kfd_fill_gpu_xgmi_link_to_gpu()
2184 struct kfd_node *peer_kdev, in kfd_fill_gpu_xgmi_link_to_gpu()
2235 size_t *size, struct kfd_node *kdev, in kfd_create_vcrat_image_gpu()
2405 int flags, struct kfd_node *kdev, in kfd_create_crat_image_virtual()
H A Dkfd_topology.c100 struct kfd_node *kfd_device_by_id(uint32_t gpu_id) in kfd_device_by_id()
1085 static uint32_t kfd_generate_gpu_id(struct kfd_node *gpu) in kfd_generate_gpu_id()
1144 static struct kfd_topology_device *kfd_assign_gpu(struct kfd_node *gpu) in kfd_assign_gpu()
1275 struct kfd_node *gpu = outbound_link->gpu; in kfd_set_recommended_sdma_engines()
1682 struct kfd_node *knode) in fill_in_l2_l3_pcache()
1782 static void kfd_fill_cache_non_crat_info(struct kfd_topology_device *dev, struct kfd_node *kdev) in kfd_fill_cache_non_crat_info()
1868 static int kfd_topology_add_device_locked(struct kfd_node *gpu, in kfd_topology_add_device_locked()
2038 int kfd_topology_add_device(struct kfd_node *gpu) in kfd_topology_add_device()
2266 int kfd_topology_remove_device(struct kfd_node *gpu) in kfd_topology_remove_device()
2306 int kfd_topology_enum_kfd_devices(uint8_t idx, struct kfd_node **kdev) in kfd_topology_enum_kfd_devices()
[all …]
H A Dkfd_kernel_queue.c42 static bool kq_initialize(struct kernel_queue *kq, struct kfd_node *dev, in kq_initialize()
307 struct kernel_queue *kernel_queue_init(struct kfd_node *dev, in kernel_queue_init()
H A Dkfd_chardev.c322 struct kfd_node *dev; in kfd_ioctl_create_queue()
905 struct kfd_node *dev; in kfd_ioctl_set_scratch_backing_va()
1024 bool kfd_dev_is_large_bar(struct kfd_node *dev) in kfd_dev_is_large_bar()
1064 struct kfd_node *dev; in kfd_ioctl_alloc_memory_of_gpu()
1265 struct kfd_node *dev; in kfd_ioctl_map_memory_to_gpu()
1484 struct kfd_node *dev; in kfd_ioctl_alloc_queue_gws()
1527 struct kfd_node *dev = NULL; in kfd_ioctl_get_dmabuf_info()
1632 struct kfd_node *dev; in kfd_ioctl_export_dmabuf()
2240 struct kfd_node *dev; in criu_restore_devices()
2414 struct kfd_node *peer; in criu_restore_bo()
[all …]
H A Dkfd_svm.c468 svm_range_validate_svm_bo(struct kfd_node *node, struct svm_range *prange) in svm_range_validate_svm_bo()
554 svm_range_vram_node_new(struct kfd_node *node, struct svm_range *prange, in svm_range_vram_node_new()
679 struct kfd_node *
696 svm_range_get_pdd_by_node(struct svm_range *prange, struct kfd_node *node) in svm_range_get_pdd_by_node()
1204 svm_nodes_in_same_hive(struct kfd_node *node_a, struct kfd_node *node_b) in svm_nodes_in_same_hive()
1211 svm_range_get_pte_flags(struct kfd_node *node, struct amdgpu_vm *vm, in svm_range_get_pte_flags()
1214 struct kfd_node *bo_node; in svm_range_get_pte_flags()
2752 struct kfd_node *node, in svm_range_best_restore_location()
2755 struct kfd_node *bo_node, *preferred_node; in svm_range_best_restore_location()
2904 svm_range *svm_range_create_unregistered_range(struct kfd_node *node, in svm_range_create_unregistered_range()
[all …]
H A Dkfd_mqd_manager_cik.c79 struct kfd_node *kfd = mm->dev; in allocate_mqd()
385 struct kfd_node *dev) in mqd_manager_init_cik()
H A Dkfd_debugfs.c56 struct kfd_node *dev; in kfd_debugfs_hang_hws_write()
H A Dkfd_mqd_manager_v12.c87 struct kfd_node *node = mm->dev; in allocate_mqd()
381 struct kfd_node *dev) in mqd_manager_init_v12()
H A Dkfd_mqd_manager_v10.c79 struct kfd_node *kfd = mm->dev; in allocate_mqd()
447 struct kfd_node *dev) in mqd_manager_init_v10()
H A Dkfd_mqd_manager_vi.c82 struct kfd_node *kfd = mm->dev; in allocate_mqd()
442 struct kfd_node *dev) in mqd_manager_init_vi()
H A Dkfd_mqd_manager_v11.c106 struct kfd_node *node = mm->dev; in allocate_mqd()
461 struct kfd_node *dev) in mqd_manager_init_v11()
H A Dkfd_doorbell.c106 int kfd_doorbell_mmap(struct kfd_node *dev, struct kfd_process *process, in kfd_doorbell_mmap()
H A Dkfd_mqd_manager_v12_1.c141 struct kfd_node *node = mm->dev; in allocate_mqd()
642 struct kfd_node *dev) in mqd_manager_init_v12_1()
H A Dkfd_events.c358 struct kfd_node *kfd; in kfd_kmap_event_page()
1245 void kfd_signal_reset_event(struct kfd_node *dev) in kfd_signal_reset_event()
1330 void kfd_signal_poison_consumed_event(struct kfd_node *dev, u32 pasid) in kfd_signal_poison_consumed_event()

12