Home
last modified time | relevance | path

Searched refs:aq (Results 1 – 25 of 36) sorted by relevance

12

/linux/drivers/s390/crypto/
H A Dap_queue.c25 static void __ap_flush_queue(struct ap_queue *aq);
31 static inline bool ap_q_supported_in_se(struct ap_queue *aq) in ap_q_supported_in_se() argument
33 return aq->card->hwinfo.ep11 || aq->card->hwinfo.accel; in ap_q_supported_in_se()
36 static inline bool ap_q_supports_bind(struct ap_queue *aq) in ap_q_supports_bind() argument
38 return aq->card->hwinfo.ep11 || aq->card->hwinfo.accel; in ap_q_supports_bind()
41 static inline bool ap_q_supports_assoc(struct ap_queue *aq) in ap_q_supports_assoc() argument
43 return aq->card->hwinfo.ep11; in ap_q_supports_assoc()
46 static inline bool ap_q_needs_bind(struct ap_queue *aq) in ap_q_needs_bind() argument
48 return ap_q_supports_bind(aq) && ap_sb_available(); in ap_q_needs_bind()
60 static int ap_queue_enable_irq(struct ap_queue *aq, void *ind) in ap_queue_enable_irq() argument
[all …]
H A Dap_card.c80 struct ap_queue *aq; in request_count_store() local
84 hash_for_each(ap_queues, bkt, aq, hnode) in request_count_store()
85 if (ac == aq->card) in request_count_store()
86 aq->total_request_count = 0; in request_count_store()
99 struct ap_queue *aq; in requestq_count_show() local
105 hash_for_each(ap_queues, bkt, aq, hnode) in requestq_count_show()
106 if (ac == aq->card) in requestq_count_show()
107 reqq_cnt += aq->requestq_count; in requestq_count_show()
118 struct ap_queue *aq; in pendingq_count_show() local
124 hash_for_each(ap_queues, bkt, aq, hnode) in pendingq_count_show()
[all …]
H A Dzcrypt_queue.c44 struct ap_queue *aq = to_ap_queue(dev); in online_show() local
45 int online = aq->config && !aq->chkstop && zq->online ? 1 : 0; in online_show()
55 struct ap_queue *aq = to_ap_queue(dev); in online_store() local
62 if (online && (!aq->config || !aq->card->config || in online_store()
63 aq->chkstop || aq->card->chkstop)) in online_store()
73 ap_send_online_uevent(&aq->ap_dev, online); in online_store()
/linux/drivers/infiniband/hw/ionic/
H A Dionic_admin.c31 static void ionic_admin_timedout(struct ionic_aq *aq) in ionic_admin_timedout() argument
33 struct ionic_ibdev *dev = aq->dev; in ionic_admin_timedout()
37 spin_lock_irqsave(&aq->lock, irqflags); in ionic_admin_timedout()
38 if (ionic_queue_empty(&aq->q)) in ionic_admin_timedout()
42 if (atomic_read(&aq->admin_state) < IONIC_ADMIN_KILLED) in ionic_admin_timedout()
46 aq->aqid, (u32)jiffies_to_msecs(jiffies - aq->stamp)); in ionic_admin_timedout()
48 pos = (aq->q.prod - 1) & aq->q.mask; in ionic_admin_timedout()
49 if (pos == aq->q.cons) in ionic_admin_timedout()
54 ionic_queue_at(&aq->q, pos), in ionic_admin_timedout()
55 BIT(aq->q.stride_log2), true); in ionic_admin_timedout()
[all …]
/linux/drivers/net/ethernet/intel/iavf/
H A Diavf_adminq.c18 ret_code = iavf_allocate_dma_mem(hw, &hw->aq.asq.desc_buf, in iavf_alloc_adminq_asq_ring()
20 (hw->aq.num_asq_entries * in iavf_alloc_adminq_asq_ring()
26 ret_code = iavf_allocate_virt_mem(hw, &hw->aq.asq.cmd_buf, in iavf_alloc_adminq_asq_ring()
27 (hw->aq.num_asq_entries * in iavf_alloc_adminq_asq_ring()
30 iavf_free_dma_mem(hw, &hw->aq.asq.desc_buf); in iavf_alloc_adminq_asq_ring()
45 ret_code = iavf_allocate_dma_mem(hw, &hw->aq.arq.desc_buf, in iavf_alloc_adminq_arq_ring()
47 (hw->aq.num_arq_entries * in iavf_alloc_adminq_arq_ring()
63 iavf_free_dma_mem(hw, &hw->aq.asq.desc_buf); in iavf_free_adminq_asq()
75 iavf_free_dma_mem(hw, &hw->aq.arq.desc_buf); in iavf_free_adminq_arq()
94 ret_code = iavf_allocate_virt_mem(hw, &hw->aq.arq.dma_head, in iavf_alloc_arq_bufs()
[all …]
/linux/drivers/net/ethernet/intel/i40e/
H A Di40e_adminq.c19 ret_code = i40e_allocate_dma_mem(hw, &hw->aq.asq.desc_buf, in i40e_alloc_adminq_asq_ring()
20 (hw->aq.num_asq_entries * in i40e_alloc_adminq_asq_ring()
26 ret_code = i40e_allocate_virt_mem(hw, &hw->aq.asq.cmd_buf, in i40e_alloc_adminq_asq_ring()
27 (hw->aq.num_asq_entries * in i40e_alloc_adminq_asq_ring()
30 i40e_free_dma_mem(hw, &hw->aq.asq.desc_buf); in i40e_alloc_adminq_asq_ring()
45 ret_code = i40e_allocate_dma_mem(hw, &hw->aq.arq.desc_buf, in i40e_alloc_adminq_arq_ring()
46 (hw->aq.num_arq_entries * in i40e_alloc_adminq_arq_ring()
62 i40e_free_dma_mem(hw, &hw->aq.asq.desc_buf); in i40e_free_adminq_asq()
74 i40e_free_dma_mem(hw, &hw->aq.arq.desc_buf); in i40e_free_adminq_arq()
93 ret_code = i40e_allocate_virt_mem(hw, &hw->aq.arq.dma_head, in i40e_alloc_arq_bufs()
[all …]
H A Di40e_nvm.c78 access, time_left, ret_code, hw->aq.asq_last_status); in i40e_acquire_nvm()
100 time_left, ret_code, hw->aq.asq_last_status); in i40e_acquire_nvm()
128 (total_delay < hw->aq.asq_cmd_timeout)) { in i40e_release_nvm()
896 __func__, status, hw->aq.asq_last_status); in i40e_nvmupd_nvm_erase()
897 *perrno = i40e_aq_rc_to_posix(status, hw->aq.asq_last_status); in i40e_nvmupd_nvm_erase()
939 __func__, status, hw->aq.asq_last_status); in i40e_nvmupd_nvm_write()
940 *perrno = i40e_aq_rc_to_posix(status, hw->aq.asq_last_status); in i40e_nvmupd_nvm_write()
979 __func__, status, hw->aq.asq_last_status); in i40e_nvmupd_nvm_read()
980 *perrno = i40e_aq_rc_to_posix(status, hw->aq.asq_last_status); in i40e_nvmupd_nvm_read()
1033 hw->aq.asq_buf_size); in i40e_nvmupd_exec_aq()
[all …]
H A Di40e_prototype.h470 return (hw->aq.api_maj_ver > maj || in i40e_is_aq_api_ver_ge()
471 (hw->aq.api_maj_ver == maj && hw->aq.api_min_ver >= min)); in i40e_is_aq_api_ver_ge()
497 return (hw->aq.fw_maj_ver > maj || in i40e_is_fw_ver_ge()
498 (hw->aq.fw_maj_ver == maj && hw->aq.fw_min_ver >= min)); in i40e_is_fw_ver_ge()
524 return (hw->aq.fw_maj_ver == maj && hw->aq.fw_min_ver == min); in i40e_is_fw_ver_eq()
H A Di40e_main.c1842 libie_aq_str(hw->aq.asq_last_status)); in i40e_set_mac()
1874 libie_aq_str(hw->aq.asq_last_status)); in i40e_config_rss_aq()
1886 libie_aq_str(hw->aq.asq_last_status)); in i40e_config_rss_aq()
2467 libie_aq_str(hw->aq.asq_last_status), vsi_name); in i40e_aqc_broadcast_filter()
2508 libie_aq_str(hw->aq.asq_last_status)); in i40e_set_promiscuous()
2520 libie_aq_str(hw->aq.asq_last_status)); in i40e_set_promiscuous()
2530 libie_aq_str(hw->aq.asq_last_status)); in i40e_set_promiscuous()
2650 filter_list_len = hw->aq.asq_buf_size / in i40e_sync_vsi_filters()
2711 filter_list_len = hw->aq.asq_buf_size / in i40e_sync_vsi_filters()
2833 hw->aq.asq_last_status); in i40e_sync_vsi_filters()
[all …]
H A Di40e_client.c354 cdev->lan_info.fw_maj_ver = pf->hw.aq.fw_maj_ver; in i40e_client_add_instance()
355 cdev->lan_info.fw_min_ver = pf->hw.aq.fw_min_ver; in i40e_client_add_instance()
356 cdev->lan_info.fw_build = pf->hw.aq.fw_build; in i40e_client_add_instance()
547 err, hw->aq.asq_last_status); in i40e_client_virtchnl_send()
685 ERR_PTR(err), libie_aq_str(pf->hw.aq.asq_last_status)); in i40e_client_update_vsi_ctxt()
712 libie_aq_str(pf->hw.aq.asq_last_status)); in i40e_client_update_vsi_ctxt()
H A Di40e_debugfs.c448 ring = &(hw->aq.asq); in i40e_dbg_dump_aq_desc()
461 ring = &(hw->aq.arq); in i40e_dbg_dump_aq_desc()
1003 pf->hw.aq.asq_last_status); in i40e_dbg_command_write()
1116 ret, pf->hw.aq.asq_last_status); in i40e_dbg_command_write()
1256 desc->opcode, pf->hw.aq.asq_last_status); in i40e_dbg_command_write()
1317 desc->opcode, pf->hw.aq.asq_last_status); in i40e_dbg_command_write()
1352 pf->hw.aq.asq_last_status); in i40e_dbg_command_write()
1362 __func__, pf->hw.aq.asq_last_status); in i40e_dbg_command_write()
1379 __func__, pf->hw.aq.asq_last_status); in i40e_dbg_command_write()
1387 pf->hw.aq.asq_last_status); in i40e_dbg_command_write()
[all …]
H A Di40e_dcb_nl.c139 libie_aq_str(pf->hw.aq.asq_last_status)); in i40e_dcbnl_ieee_setets()
178 libie_aq_str(pf->hw.aq.asq_last_status)); in i40e_dcbnl_ieee_setpfc()
229 libie_aq_str(pf->hw.aq.asq_last_status)); in i40e_dcbnl_ieee_setapp()
294 libie_aq_str(pf->hw.aq.asq_last_status)); in i40e_dcbnl_ieee_delapp()
H A Di40e_common.c140 if (!hw->aq.asq.count) in i40e_check_asq_alive()
1029 switch (hw->aq.asq_last_status) { in i40e_aq_get_phy_capabilities()
1043 } while ((hw->aq.asq_last_status == LIBIE_AQ_RC_EAGAIN) && in i40e_aq_get_phy_capabilities()
2512 if (!status || hw->aq.asq_last_status == LIBIE_AQ_RC_EBUSY) in i40e_aq_request_resource()
4540 if (hw->aq.asq_last_status == LIBIE_AQ_RC_EAGAIN && retry) { in i40e_read_rx_ctl()
4602 if (hw->aq.asq_last_status == LIBIE_AQ_RC_EAGAIN && retry) { in i40e_write_rx_ctl()
4856 struct i40e_profile_aq_section *aq) in i40e_ddp_exec_aq_section() argument
4863 i40e_fill_default_direct_cmd_desc(&desc, aq->opcode); in i40e_ddp_exec_aq_section()
4864 desc.flags |= cpu_to_le16(aq->flags); in i40e_ddp_exec_aq_section()
4865 memcpy(desc.params.raw, aq->param, sizeof(desc.params.raw)); in i40e_ddp_exec_aq_section()
[all …]
H A Di40e_ethtool.c1465 libie_aq_str(hw->aq.asq_last_status)); in i40e_set_link_ksettings()
1475 libie_aq_str(hw->aq.asq_last_status)); in i40e_set_link_ksettings()
1523 libie_aq_str(hw->aq.asq_last_status)); in i40e_set_fec_cfg()
1537 libie_aq_str(hw->aq.asq_last_status)); in i40e_set_fec_cfg()
1644 libie_aq_str(hw->aq.asq_last_status)); in i40e_nway_reset()
1761 libie_aq_str(hw->aq.asq_last_status)); in i40e_set_pauseparam()
1767 libie_aq_str(hw->aq.asq_last_status)); in i40e_set_pauseparam()
1773 libie_aq_str(hw->aq.asq_last_status)); in i40e_set_pauseparam()
1886 ret_val, hw->aq.asq_last_status, errno, in i40e_get_eeprom()
1904 ret_val, hw->aq.asq_last_status); in i40e_get_eeprom()
[all …]
H A Di40e_dcb.c753 if (hw->aq.asq_last_status == LIBIE_AQ_RC_ENOENT) in i40e_get_ieee_dcb_config()
802 if (hw->aq.asq_last_status == LIBIE_AQ_RC_ENOENT) in i40e_get_dcb_config()
819 if (hw->aq.asq_last_status == LIBIE_AQ_RC_ENOENT) in i40e_get_dcb_config()
928 } else if (hw->aq.asq_last_status == LIBIE_AQ_RC_ENOENT) { in i40e_get_fw_lldp_status()
932 } else if (hw->aq.asq_last_status == LIBIE_AQ_RC_EPERM) { in i40e_get_fw_lldp_status()
/linux/drivers/net/ethernet/marvell/octeontx2/nic/
H A Dotx2_common.c328 struct nix_aq_enq_req *aq; in otx2_set_rss_table() local
339 aq = otx2_mbox_alloc_msg_nix_aq_enq(mbox); in otx2_set_rss_table()
340 if (!aq) { in otx2_set_rss_table()
349 aq = otx2_mbox_alloc_msg_nix_aq_enq(mbox); in otx2_set_rss_table()
350 if (!aq) { in otx2_set_rss_table()
356 aq->rss.rq = ind_tbl[idx]; in otx2_set_rss_table()
359 aq->qidx = index + idx; in otx2_set_rss_table()
360 aq->ctype = NIX_AQ_CTYPE_RSS; in otx2_set_rss_table()
361 aq->op = NIX_AQ_INSTOP_INIT; in otx2_set_rss_table()
893 struct nix_aq_enq_req *aq; in otx2_rq_init() local
[all …]
/linux/lib/
H A Dkasprintf.c19 va_list aq; in kvasprintf() local
21 va_copy(aq, ap); in kvasprintf()
22 first = vsnprintf(NULL, 0, fmt, aq); in kvasprintf()
23 va_end(aq); in kvasprintf()
/linux/drivers/crypto/hisilicon/sec2/
H A Dsec_crypto.c1474 struct aead_request *aq = req->aead_req.aead_req; in GEN_SEC_AEAD_SETKEY_FUNC() local
1476 return sec_cipher_map(ctx, req, aq->src, aq->dst); in GEN_SEC_AEAD_SETKEY_FUNC()
1481 struct aead_request *aq = req->aead_req.aead_req; in sec_aead_sgl_unmap() local
1483 sec_cipher_unmap(ctx, req, aq->src, aq->dst); in sec_aead_sgl_unmap()
1759 struct aead_request *aq = a_req->aead_req; in sec_auth_bd_fill_xcm() local
1760 struct crypto_aead *tfm = crypto_aead_reqtfm(aq); in sec_auth_bd_fill_xcm()
1776 sec_sqe->type2.alen_ivllen = cpu_to_le32(aq->assoclen); in sec_auth_bd_fill_xcm()
1778 sec_sqe->type2.cipher_src_offset = cpu_to_le16((u16)aq->assoclen); in sec_auth_bd_fill_xcm()
1787 struct aead_request *aq = a_req->aead_req; in sec_auth_bd_fill_xcm_v3() local
1788 struct crypto_aead *tfm = crypto_aead_reqtfm(aq); in sec_auth_bd_fill_xcm_v3()
[all …]
/linux/arch/riscv/include/asm/
H A Dasm.h32 #define REG_AMOSWAP_AQ __REG_SEL(amoswap.d.aq, amoswap.w.aq)
H A Dbitops.h300 return __test_and_op_bit_ord(or, __NOP, nr, addr, .aq); in arch_test_and_set_bit_lock()
/linux/lib/tests/
H A Dprintf_kunit.c45 va_list aq; in do_test() local
51 va_copy(aq, ap); in do_test()
52 ret = vsnprintf(test_buffer, bufsize, fmt, aq); in do_test()
53 va_end(aq); in do_test()
/linux/drivers/base/
H A Ddevres.c1035 va_list aq; in devm_kvasprintf() local
1037 va_copy(aq, ap); in devm_kvasprintf()
1038 len = vsnprintf(NULL, 0, fmt, aq); in devm_kvasprintf()
1039 va_end(aq); in devm_kvasprintf()
/linux/drivers/infiniband/hw/efa/
H A Defa_main.c239 edev->aq.poll_interval = hw_hints->poll_interval; in efa_update_hw_hints()
242 edev->aq.completion_timeout = in efa_update_hw_hints()
611 edev->aq.msix_vector_idx = dev->admin_msix_vector_idx; in efa_probe_device()
/linux/drivers/net/ethernet/marvell/octeontx2/af/
H A Drvu.c1201 void rvu_aq_free(struct rvu *rvu, struct admin_queue *aq) in rvu_aq_free() argument
1203 if (!aq) in rvu_aq_free()
1206 qmem_free(rvu->dev, aq->inst); in rvu_aq_free()
1207 qmem_free(rvu->dev, aq->res); in rvu_aq_free()
1208 devm_kfree(rvu->dev, aq); in rvu_aq_free()
1214 struct admin_queue *aq; in rvu_aq_alloc() local
1217 *ad_queue = devm_kzalloc(rvu->dev, sizeof(*aq), GFP_KERNEL); in rvu_aq_alloc()
1220 aq = *ad_queue; in rvu_aq_alloc()
1223 err = qmem_alloc(rvu->dev, &aq->inst, qsize, inst_size); in rvu_aq_alloc()
1225 devm_kfree(rvu->dev, aq); in rvu_aq_alloc()
[all …]
H A Drvu_nix.c969 struct admin_queue *aq = block->aq; in nix_aq_enqueue_wait() local
975 result = (struct nix_aq_res_s *)aq->res->base; in nix_aq_enqueue_wait()
981 memcpy((void *)(aq->inst->base + (head * aq->inst->entry_sz)), in nix_aq_enqueue_wait()
982 (void *)inst, aq->inst->entry_sz); in nix_aq_enqueue_wait()
1047 struct admin_queue *aq; in rvu_nix_blk_aq_enq_inst() local
1056 aq = block->aq; in rvu_nix_blk_aq_enq_inst()
1057 if (!aq) { in rvu_nix_blk_aq_enq_inst()
1140 inst.res_addr = (u64)aq->res->iova; in rvu_nix_blk_aq_enq_inst()
1145 spin_lock(&aq->lock); in rvu_nix_blk_aq_enq_inst()
1148 memset(aq->res->base, 0, aq->res->entry_sz); in rvu_nix_blk_aq_enq_inst()
[all …]

12