| /linux/drivers/net/ethernet/huawei/hinic/ |
| H A D | hinic_hw_wq.c | 31 #define WQS_FREE_BLOCKS_SIZE(wqs) (WQS_MAX_NUM_BLOCKS * \ argument 32 sizeof((wqs)->free_blocks[0])) 44 #define WQ_BASE_VADDR(wqs, wq) \ argument 45 ((void *)((wqs)->page_vaddr[(wq)->page_idx]) \ 48 #define WQ_BASE_PADDR(wqs, wq) \ argument 49 ((wqs)->page_paddr[(wq)->page_idx] \ 52 #define WQ_BASE_ADDR(wqs, wq) \ argument 53 ((void *)((wqs)->shadow_page_vaddr[(wq)->page_idx]) \ 137 static int wqs_allocate_page(struct hinic_wqs *wqs, int page_idx) in wqs_allocate_page() argument 139 return queue_alloc_page(wqs->hwif, &wqs->page_vaddr[page_idx], in wqs_allocate_page() [all …]
|
| H A D | hinic_hw_wq.h | 85 int hinic_wqs_alloc(struct hinic_wqs *wqs, int num_wqs, 88 void hinic_wqs_free(struct hinic_wqs *wqs); 90 int hinic_wq_allocate(struct hinic_wqs *wqs, struct hinic_wq *wq, 94 void hinic_wq_free(struct hinic_wqs *wqs, struct hinic_wq *wq);
|
| H A D | hinic_hw_io.c | 283 err = hinic_wq_allocate(&func_to_io->wqs, &func_to_io->sq_wq[q_id], in init_qp() 291 err = hinic_wq_allocate(&func_to_io->wqs, &func_to_io->rq_wq[q_id], in init_qp() 335 hinic_wq_free(&func_to_io->wqs, &func_to_io->rq_wq[q_id]); in init_qp() 338 hinic_wq_free(&func_to_io->wqs, &func_to_io->sq_wq[q_id]); in init_qp() 357 hinic_wq_free(&func_to_io->wqs, &func_to_io->rq_wq[q_id]); in destroy_qp() 358 hinic_wq_free(&func_to_io->wqs, &func_to_io->sq_wq[q_id]); in destroy_qp() 550 err = hinic_wqs_alloc(&func_to_io->wqs, 2 * max_qps, hwif); in hinic_io_init() 606 hinic_wqs_free(&func_to_io->wqs); in hinic_io_init() 632 hinic_wqs_free(&func_to_io->wqs); in hinic_io_free()
|
| H A D | hinic_hw_io.h | 65 struct hinic_wqs wqs; member
|
| /linux/drivers/crypto/intel/iaa/ |
| H A D | iaa_crypto_main.c | 46 if (!entry->wqs[entry->cur_wq]) in wq_table_next_wq() 50 entry->cur_wq, entry->wqs[entry->cur_wq]->idxd->id, in wq_table_next_wq() 51 entry->wqs[entry->cur_wq]->id, cpu); in wq_table_next_wq() 53 return entry->wqs[entry->cur_wq]; in wq_table_next_wq() 63 entry->wqs[entry->n_wqs++] = wq; in wq_table_add() 66 entry->wqs[entry->n_wqs - 1]->idxd->id, in wq_table_add() 67 entry->wqs[entry->n_wqs - 1]->id, entry->n_wqs - 1, cpu); in wq_table_add() 74 kfree(entry->wqs); in wq_table_free_entry() 84 memset(entry->wqs, 0, entry->max_wqs * sizeof(struct idxd_wq *)); in wq_table_clear_entry() 511 INIT_LIST_HEAD(&iaa_device->wqs); in iaa_device_alloc() [all …]
|
| H A D | iaa_crypto.h | 74 struct list_head wqs; member 83 struct idxd_wq **wqs; member
|
| H A D | iaa_crypto_stats.c | 132 list_for_each_entry(iaa_wq, &iaa_device->wqs, list) in reset_device_stats() 166 list_for_each_entry(iaa_wq, &iaa_device->wqs, list) in device_stats_show()
|
| /linux/drivers/dma/idxd/ |
| H A D | init.c | 167 wq = idxd->wqs[i]; in idxd_clean_wqs() 176 kfree(idxd->wqs); in idxd_clean_wqs() 186 idxd->wqs = kcalloc_node(idxd->max_wqs, sizeof(struct idxd_wq *), in idxd_setup_wqs() 188 if (!idxd->wqs) in idxd_setup_wqs() 247 idxd->wqs[i] = wq; in idxd_setup_wqs() 254 wq = idxd->wqs[i]; in idxd_setup_wqs() 265 kfree(idxd->wqs); in idxd_setup_wqs() 937 wq = idxd->wqs[i]; in idxd_device_config_save() 1007 wq = idxd->wqs[i]; in idxd_device_config_restore() 1114 struct idxd_wq *wq = idxd->wqs[i]; in idxd_reset_done() [all …]
|
| H A D | device.c | 312 struct idxd_wq *wq = idxd->wqs[i]; in idxd_wqs_unmap_portal() 729 struct idxd_wq *wq = idxd->wqs[i]; in idxd_device_wqs_clear_state() 871 iowrite64(group->grpcfg.wqs[i], idxd->reg_base + grpcfg_offset); in idxd_group_config_write() 1021 struct idxd_wq *wq = idxd->wqs[i]; in idxd_wqs_config_write() 1093 group->grpcfg.wqs[j] = 0; in idxd_wqs_setup() 1097 wq = idxd->wqs[i]; in idxd_wqs_setup() 1109 group->grpcfg.wqs[wq->id / 64] |= BIT(wq->id % 64); in idxd_wqs_setup() 1193 group->grpcfg.wqs[i] = ioread64(idxd->reg_base + grpcfg_offset); in idxd_group_load_config() 1195 group->id, i, grpcfg_offset, group->grpcfg.wqs[i]); in idxd_group_load_config() 1209 if (group->grpcfg.wqs[i] & BIT(j)) { in idxd_group_load_config() [all …]
|
| H A D | defaults.c | 15 wq = idxd->wqs[0]; in idxd_load_iaa_device_defaults()
|
| H A D | sysfs.c | 342 struct idxd_wq *wq = idxd->wqs[i]; in group_work_queues_show() 721 struct idxd_wq *wq = idxd->wqs[i]; in total_claimed_wq_size() 1513 struct idxd_wq *wq = idxd->wqs[i]; in clients_show() 1817 kfree(idxd->wqs); in idxd_conf_device_release() 1900 wq = idxd->wqs[i]; in idxd_register_wq_devices() 1911 wq = idxd->wqs[i]; in idxd_register_wq_devices() 1916 wq = idxd->wqs[j]; in idxd_register_wq_devices() 1956 device_unregister(wq_confdev(idxd->wqs[i])); in idxd_register_devices() 1967 struct idxd_wq *wq = idxd->wqs[i]; in idxd_unregister_devices()
|
| H A D | idxd.h | 331 struct idxd_wq **wqs; member 505 return (idx == 0) ? &idxd->ie : &idxd->wqs[idx - 1]->ie; in idxd_get_ie()
|
| H A D | registers.h | 363 u64 wqs[4]; member
|
| /linux/Documentation/driver-api/crypto/iaa/ |
| H A D | iaa-crypto.rst | 364 wqs: 378 wqs: 392 wqs: 406 wqs: 545 # disable iaa wqs and devices 566 # configure iaa wqs and devices 577 # enable iaa wqs and devices 674 Once zswap is disabled and no longer using iaa_crypto, the IAA wqs and 699 # disable iaa wqs and devices
|
| /linux/tools/testing/selftests/vfio/lib/drivers/dsa/ |
| H A D | dsa.c | 180 writeq(1, dsa->grpcfg_table + offsetof(struct grpcfg, wqs[0])); in dsa_group_init()
|