/linux/drivers/net/ethernet/intel/idpf/ |
H A D | idpf_lan_txrx.h | 186 __le64 qw1; /* type_cmd_offset_bsz_l2tag1 */ member 207 __le64 qw1; /* type_cmd_tlen_mss/rt_hint */ member 236 } qw1; member 258 } qw1; member 292 } qw1; member
|
H A D | idpf_singleq_txrx.c | 238 tx_desc->qw1 = idpf_tx_singleq_build_ctob(td_cmd, in idpf_tx_singleq_map() 263 tx_desc->qw1 = idpf_tx_singleq_build_ctob(td_cmd, offsets, in idpf_tx_singleq_map() 287 tx_desc->qw1 = idpf_tx_singleq_build_ctob(td_cmd, offsets, in idpf_tx_singleq_map() 333 u64 qw1 = (u64)IDPF_TX_DESC_DTYPE_CTX; in idpf_tx_singleq_build_ctx_desc() local 336 qw1 |= IDPF_TX_CTX_DESC_TSO << IDPF_TXD_CTX_QW1_CMD_S; in idpf_tx_singleq_build_ctx_desc() 337 qw1 |= FIELD_PREP(IDPF_TXD_CTX_QW1_TSO_LEN_M, in idpf_tx_singleq_build_ctx_desc() 339 qw1 |= FIELD_PREP(IDPF_TXD_CTX_QW1_MSS_M, offload->mss); in idpf_tx_singleq_build_ctx_desc() 350 desc->qw1 = cpu_to_le64(qw1); in idpf_tx_singleq_build_ctx_desc() 471 if (!(eop_desc->qw1 & in idpf_tx_singleq_clean() 741 u64 mask, qw1; in idpf_rx_singleq_base_hash() local [all …]
|
H A D | idpf_txrx.c | 2110 desc->q.qw1.cmd_dtype = in idpf_tx_splitq_build_ctb() 2112 desc->q.qw1.cmd_dtype |= in idpf_tx_splitq_build_ctb() 2114 desc->q.qw1.buf_size = cpu_to_le16(size); in idpf_tx_splitq_build_ctb() 2115 desc->q.qw1.l2tags.l2tag1 = cpu_to_le16(params->td_tag); in idpf_tx_splitq_build_ctb() 2130 desc->flow.qw1.cmd_dtype = (u16)params->dtype | td_cmd; in idpf_tx_splitq_build_flow_desc() 2131 desc->flow.qw1.rxr_bufsize = cpu_to_le16((u16)size); in idpf_tx_splitq_build_flow_desc() 2132 desc->flow.qw1.compl_tag = cpu_to_le16(params->compl_tag); in idpf_tx_splitq_build_flow_desc() 2770 ctx_desc->tso.qw1.cmd_dtype = in idpf_tx_splitq_frame()
|
/linux/drivers/infiniband/hw/irdma/ |
H A D | uda.c | 24 u64 qw1, qw2; in irdma_sc_access_ah() local 31 qw1 = FIELD_PREP(IRDMA_UDA_CQPSQ_MAV_PDINDEXLO, info->pd_idx) | in irdma_sc_access_ah() 62 set_64bit_val(wqe, 8, qw1); in irdma_sc_access_ah()
|
H A D | ctrl.c | 305 u64 qw1 = 0; in irdma_sc_manage_qhash_table_entry() local 316 qw1 = FIELD_PREP(IRDMA_CQPSQ_QHASH_QPN, info->qp_num) | in irdma_sc_manage_qhash_table_entry() 336 qw1 |= FIELD_PREP(IRDMA_CQPSQ_QHASH_SRC_PORT, info->src_port); in irdma_sc_manage_qhash_table_entry() 350 set_64bit_val(wqe, 8, qw1); in irdma_sc_manage_qhash_table_entry()
|
/linux/drivers/iommu/intel/ |
H A D | iommu.h | 479 u64 qw1; member 1096 desc->qw1 = QI_IOTLB_ADDR(addr) | QI_IOTLB_IH(ih) in qi_desc_iotlb() 1107 desc->qw1 = QI_DEV_IOTLB_ADDR(addr) | QI_DEV_IOTLB_SIZE; in qi_desc_dev_iotlb() 1109 desc->qw1 = QI_DEV_IOTLB_ADDR(addr); in qi_desc_dev_iotlb() 1130 desc->qw1 = 0; in qi_desc_piotlb() 1142 desc->qw1 = QI_EIOTLB_ADDR(addr) | in qi_desc_piotlb() 1173 desc->qw1 = QI_DEV_EIOTLB_ADDR(addr); in qi_desc_dev_iotlb_pasid() 1181 desc->qw1 |= GENMASK_ULL(size_order + VTD_PAGE_SHIFT - 1, in qi_desc_dev_iotlb_pasid() 1184 desc->qw1 &= ~mask; in qi_desc_dev_iotlb_pasid() 1186 desc->qw1 |= QI_DEV_EIOTLB_SIZE; in qi_desc_dev_iotlb_pasid()
|
H A D | dmar.c | 1258 (unsigned long long)desc->qw1); in qi_dump_fault() 1267 (unsigned long long)desc->qw1); in qi_dump_fault() 1422 trace_qi_submit(iommu, desc[i].qw0, desc[i].qw1, in qi_submit_sync() 1431 wait_desc.qw1 = virt_to_phys(&qi->desc_status[wait_index]); in qi_submit_sync() 1504 desc.qw1 = 0; in qi_global_iec() 1519 desc.qw1 = 0; in qi_flush_context() 1577 struct qi_desc desc = {.qw1 = 0, .qw2 = 0, .qw3 = 0}; in qi_flush_dev_iotlb_pasid() 1597 struct qi_desc desc = {.qw1 = 0, .qw2 = 0, .qw3 = 0}; in qi_flush_pasid_cache()
|
H A D | svm.c | 424 desc.qw1 = QI_PGRP_IDX(req->prg_index) | in handle_bad_prq_event() 547 desc.qw1 = QI_PGRP_IDX(prm->grpid) | QI_PGRP_LPIG(last_page); in intel_svm_page_response()
|
H A D | debugfs.c | 499 index, desc->qw0, desc->qw1, in invalidation_queue_entry_show() 504 index, desc->qw0, desc->qw1, in invalidation_queue_entry_show()
|
H A D | pasid.c | 202 desc.qw1 = 0; in pasid_cache_invalidation_with_pasid()
|
H A D | irq_remapping.c | 155 desc.qw1 = 0; in qi_flush_iec()
|
/linux/Documentation/ABI/testing/ |
H A D | debugfs-intel-iommu | 132 Index qw0 qw1 qw2 146 Index qw0 qw1 status
|
/linux/drivers/infiniband/hw/hfi1/ |
H A D | sdma.c | 2239 static inline u64 add_gen(struct sdma_engine *sde, u64 qw1) in add_gen() argument 2243 qw1 &= ~SDMA_DESC1_GENERATION_SMASK; in add_gen() 2244 qw1 |= ((u64)generation & SDMA_DESC1_GENERATION_MASK) in add_gen() 2246 return qw1; in add_gen() 2282 u64 qw1; in submit_tx() local 2287 qw1 = descp->qw[1]; in submit_tx() 2291 qw1 = add_gen(sde, descp->qw[1]); in submit_tx() 2293 sde->descq[tail].qw[1] = cpu_to_le64(qw1); in submit_tx() 2294 trace_hfi1_sdma_descriptor(sde, descp->qw[0], qw1, in submit_tx()
|
/linux/drivers/net/ethernet/intel/ice/ |
H A D | ice_lan_tx_rx.h | 504 __le64 qw1; member
|
H A D | ice_txrx.c | 2387 cdesc->qw1 = cpu_to_le64(offload.cd_qw1); in ice_xmit_frame_ring()
|
/linux/drivers/scsi/hisi_sas/ |
H A D | hisi_sas.h | 541 __le64 qw1; member
|
/linux/arch/powerpc/kvm/ |
H A D | book3s_xive.c | 357 __be64 qw1 = __raw_readq(xive_tima + TM_QW1_OS); in xive_vm_h_ipoll() local 358 u8 pipr = be64_to_cpu(qw1) & 0xff; in xive_vm_h_ipoll()
|