| /linux/drivers/net/ethernet/intel/iavf/ |
| H A D | iavf_txrx.c | 24 static bool iavf_is_descriptor_done(u64 qw1, bool flex) in iavf_is_descriptor_done() argument 27 return FIELD_GET(IAVF_RXD_FLEX_DD_M, qw1); in iavf_is_descriptor_done() 29 return FIELD_GET(IAVF_RXD_LEGACY_DD_M, qw1); in iavf_is_descriptor_done() 896 rx_desc->qw1 = 0; in iavf_alloc_rx_buffers() 977 iavf_legacy_rx_csum(const struct iavf_vsi *vsi, u64 qw1, in iavf_legacy_rx_csum() argument 985 csum_bits.ipe = FIELD_GET(IAVF_RXD_LEGACY_IPE_M, qw1); in iavf_legacy_rx_csum() 986 csum_bits.eipe = FIELD_GET(IAVF_RXD_LEGACY_EIPE_M, qw1); in iavf_legacy_rx_csum() 987 csum_bits.l4e = FIELD_GET(IAVF_RXD_LEGACY_L4E_M, qw1); in iavf_legacy_rx_csum() 988 csum_bits.pprs = FIELD_GET(IAVF_RXD_LEGACY_PPRS_M, qw1); in iavf_legacy_rx_csum() 989 csum_bits.l3l4p = FIELD_GET(IAVF_RXD_LEGACY_L3L4P_M, qw1); in iavf_legacy_rx_csum() [all …]
|
| H A D | iavf_type.h | 214 aligned_le64 qw1; member
|
| /linux/drivers/iommu/intel/ |
| H A D | trace.h | 22 TP_PROTO(struct intel_iommu *iommu, u64 qw0, u64 qw1, u64 qw2, u64 qw3), 24 TP_ARGS(iommu, qw0, qw1, qw2, qw3), 28 __field(u64, qw1) 37 __entry->qw1 = qw1; 54 __entry->qw0, __entry->qw1, __entry->qw2, __entry->qw3
|
| H A D | dmar.c | 1259 (unsigned long long)desc->qw1); in qi_dump_fault() 1268 (unsigned long long)desc->qw1); in qi_dump_fault() 1422 trace_qi_submit(iommu, desc[i].qw0, desc[i].qw1, in qi_submit_sync() 1431 wait_desc.qw1 = virt_to_phys(&qi->desc_status[wait_index]); in qi_submit_sync() 1504 desc.qw1 = 0; in qi_global_iec() 1519 desc.qw1 = 0; in qi_flush_context() 1577 struct qi_desc desc = {.qw1 = 0, .qw2 = 0, .qw3 = 0}; in qi_flush_dev_iotlb_pasid() 1597 struct qi_desc desc = {.qw1 = 0, .qw2 = 0, .qw3 = 0}; in qi_flush_pasid_cache()
|
| H A D | pasid.c | 204 desc.qw1 = 0; in pasid_cache_invalidation_with_pasid()
|
| H A D | irq_remapping.c | 149 desc.qw1 = 0; in qi_flush_iec()
|
| /linux/drivers/infiniband/hw/irdma/ |
| H A D | uda.c | 24 u64 qw1, qw2; in irdma_sc_access_ah() local 31 qw1 = FIELD_PREP(IRDMA_UDA_CQPSQ_MAV_PDINDEXLO, info->pd_idx) | in irdma_sc_access_ah() 62 set_64bit_val(wqe, 8, qw1); in irdma_sc_access_ah()
|
| /linux/Documentation/ABI/testing/ |
| H A D | debugfs-intel-iommu | 132 Index qw0 qw1 qw2 146 Index qw0 qw1 status
|
| /linux/drivers/infiniband/hw/hfi1/ |
| H A D | sdma.c | 2222 static inline u64 add_gen(struct sdma_engine *sde, u64 qw1) in add_gen() argument 2226 qw1 &= ~SDMA_DESC1_GENERATION_SMASK; in add_gen() 2227 qw1 |= ((u64)generation & SDMA_DESC1_GENERATION_MASK) in add_gen() 2229 return qw1; in add_gen() 2265 u64 qw1; in submit_tx() local 2270 qw1 = descp->qw[1]; in submit_tx() 2274 qw1 = add_gen(sde, descp->qw[1]); in submit_tx() 2276 sde->descq[tail].qw[1] = cpu_to_le64(qw1); in submit_tx() 2277 trace_hfi1_sdma_descriptor(sde, descp->qw[0], qw1, in submit_tx()
|
| /linux/drivers/net/ethernet/intel/idpf/ |
| H A D | idpf_txrx.c | 2391 desc->q.qw1.cmd_dtype = in idpf_tx_splitq_build_ctb() 2393 desc->q.qw1.cmd_dtype |= in idpf_tx_splitq_build_ctb() 2395 desc->q.qw1.buf_size = cpu_to_le16(size); in idpf_tx_splitq_build_ctb() 2396 desc->q.qw1.l2tags.l2tag1 = cpu_to_le16(params->td_tag); in idpf_tx_splitq_build_ctb() 2411 *(u32 *)&desc->flow.qw1.cmd_dtype = (u8)(params->dtype | td_cmd); in idpf_tx_splitq_build_flow_desc() 2412 desc->flow.qw1.rxr_bufsize = cpu_to_le16((u16)size); in idpf_tx_splitq_build_flow_desc() 2413 desc->flow.qw1.compl_tag = cpu_to_le16(params->compl_tag); in idpf_tx_splitq_build_flow_desc() 2981 ctx_desc->tsyn.qw1 = le64_encode_bits(IDPF_TX_DESC_DTYPE_CTX, in idpf_tx_set_tstamp_desc() 3053 ctx_desc->tso.qw1.cmd_dtype = in idpf_tx_splitq_frame()
|
| /linux/drivers/scsi/hisi_sas/ |
| H A D | hisi_sas.h | 554 __le64 qw1; member
|
| /linux/arch/powerpc/kvm/ |
| H A D | book3s_xive.c | 357 __be64 qw1 = __raw_readq(xive_tima + TM_QW1_OS); in xive_vm_h_ipoll() local 358 u8 pipr = be64_to_cpu(qw1) & 0xff; in xive_vm_h_ipoll()
|
| /linux/drivers/net/ethernet/intel/ice/ |
| H A D | ice_txrx.c | 2243 cdesc->qw1 = cpu_to_le64(offload.cd_qw1); in ice_xmit_frame_ring()
|