Home
last modified time | relevance | path

Searched refs:sde (Results 1 – 25 of 27) sorted by relevance

12

/linux/drivers/infiniband/hw/hfi1/
H A Dsdma.c201 struct sdma_engine *sde,
204 struct sdma_engine *sde,
206 static void dump_sdma_state(struct sdma_engine *sde);
207 static void sdma_make_progress(struct sdma_engine *sde, u64 status);
208 static void sdma_desc_avail(struct sdma_engine *sde, uint avail);
209 static void sdma_flush_descq(struct sdma_engine *sde);
245 struct sdma_engine *sde, in write_sde_csr() argument
249 write_kctxt_csr(sde->dd, sde->this_idx, offset0, value); in write_sde_csr()
253 struct sdma_engine *sde, in read_sde_csr() argument
256 return read_kctxt_csr(sde->dd, sde->this_idx, offset0); in read_sde_csr()
[all …]
H A Dvnic_sdma.c49 static noinline int build_vnic_ulp_payload(struct sdma_engine *sde, in build_vnic_ulp_payload() argument
55 sde->dd, in build_vnic_ulp_payload()
66 ret = sdma_txadd_page(sde->dd, in build_vnic_ulp_payload()
77 ret = sdma_txadd_kvaddr(sde->dd, &tx->txreq, in build_vnic_ulp_payload()
85 static int build_vnic_tx_desc(struct sdma_engine *sde, in build_vnic_tx_desc() argument
107 sde->dd, in build_vnic_tx_desc()
115 ret = build_vnic_ulp_payload(sde, tx); in build_vnic_tx_desc()
131 struct sdma_engine *sde = vnic_sdma->sde; in hfi1_vnic_send_dma() local
138 if (unlikely(!sde || !sdma_running(sde))) in hfi1_vnic_send_dma()
151 ret = build_vnic_tx_desc(sde, tx, pbc); in hfi1_vnic_send_dma()
[all …]
H A Dsdma.h390 static inline int sdma_empty(struct sdma_engine *sde) in sdma_empty() argument
392 return sde->descq_tail == sde->descq_head; in sdma_empty()
395 static inline u16 sdma_descq_freecnt(struct sdma_engine *sde) in sdma_descq_freecnt() argument
397 return sde->descq_cnt - in sdma_descq_freecnt()
398 (sde->descq_tail - in sdma_descq_freecnt()
399 READ_ONCE(sde->descq_head)) - 1; in sdma_descq_freecnt()
402 static inline u16 sdma_descq_inprocess(struct sdma_engine *sde) in sdma_descq_inprocess() argument
404 return sde->descq_cnt - sdma_descq_freecnt(sde); in sdma_descq_inprocess()
827 int sdma_send_txreq(struct sdma_engine *sde,
831 int sdma_send_txlist(struct sdma_engine *sde,
[all …]
H A Dmsix.c184 int msix_request_sdma_irq(struct sdma_engine *sde) in msix_request_sdma_irq() argument
190 sde->dd->unit, sde->this_idx); in msix_request_sdma_irq()
191 nr = msix_request_irq(sde->dd, sde, sdma_interrupt, NULL, in msix_request_sdma_irq()
195 sde->msix_intr = nr; in msix_request_sdma_irq()
196 remap_sdma_interrupts(sde->dd, sde->this_idx, nr); in msix_request_sdma_irq()
258 struct sdma_engine *sde = &dd->per_sdma[i]; in msix_request_irqs() local
260 ret = msix_request_sdma_irq(sde); in msix_request_irqs()
263 enable_sdma_srcs(sde->dd, i); in msix_request_irqs()
H A Dsysfs.c571 ssize_t (*show)(struct sdma_engine *sde, char *buf);
572 ssize_t (*store)(struct sdma_engine *sde, const char *buf, size_t cnt);
579 struct sdma_engine *sde = in sde_show() local
585 return sde_attr->show(sde, buf); in sde_show()
593 struct sdma_engine *sde = in sde_store() local
602 return sde_attr->store(sde, buf, count); in sde_store()
618 static ssize_t sde_show_cpu_to_sde_map(struct sdma_engine *sde, char *buf) in sde_show_cpu_to_sde_map() argument
620 return sdma_get_cpu_to_sde_map(sde, buf); in sde_show_cpu_to_sde_map()
623 static ssize_t sde_store_cpu_to_sde_map(struct sdma_engine *sde, in sde_store_cpu_to_sde_map() argument
626 return sdma_set_cpu_to_sde_map(sde, buf, count); in sde_store_cpu_to_sde_map()
[all …]
H A Dqp.c26 struct sdma_engine *sde,
437 struct sdma_engine *sde, in iowait_sleep() argument
461 write_seqlock(&sde->waitlock); in iowait_sleep()
462 if (sdma_progress(sde, seq, stx)) in iowait_sleep()
472 &sde->dmawait); in iowait_sleep()
473 priv->s_iowait.lock = &sde->waitlock; in iowait_sleep()
477 write_sequnlock(&sde->waitlock); in iowait_sleep()
487 write_sequnlock(&sde->waitlock); in iowait_sleep()
542 struct sdma_engine *sde; in qp_to_sdma_engine() local
552 sde = sdma_select_engine_sc(dd, qp->ibqp.qp_num >> dd->qos_shift, sc5); in qp_to_sdma_engine()
[all …]
H A Dipoib_tx.c126 tx->txq->sde->this_idx); in hfi1_ipoib_free_tx()
389 txq->sde = in hfi1_ipoib_send_dma_common()
410 ret = sdma_send_txlist(txq->sde, in hfi1_ipoib_submit_tx_list()
444 ret = sdma_send_txreq(txq->sde, in hfi1_ipoib_submit_tx()
613 static int hfi1_ipoib_sdma_sleep(struct sdma_engine *sde, in hfi1_ipoib_sdma_sleep() argument
622 write_seqlock(&sde->waitlock); in hfi1_ipoib_sdma_sleep()
625 if (sdma_progress(sde, seq, txreq)) { in hfi1_ipoib_sdma_sleep()
626 write_sequnlock(&sde->waitlock); in hfi1_ipoib_sdma_sleep()
634 struct hfi1_ibport *ibp = &sde->ppd->ibport_data; in hfi1_ipoib_sdma_sleep()
641 iowait_queue(pkts_sent, wait->iow, &sde->dmawait); in hfi1_ipoib_sdma_sleep()
[all …]
H A Duser_sdma.c57 struct sdma_engine *sde,
65 struct sdma_engine *sde, in defer_packet_queue() argument
74 write_seqlock(&sde->waitlock); in defer_packet_queue()
75 trace_hfi1_usdma_defer(pq, sde, &pq->busy); in defer_packet_queue()
76 if (sdma_progress(sde, seq, txreq)) in defer_packet_queue()
85 pq->busy.lock = &sde->waitlock; in defer_packet_queue()
87 iowait_queue(pkts_sent, &pq->busy, &sde->dmawait); in defer_packet_queue()
89 write_sequnlock(&sde->waitlock); in defer_packet_queue()
92 write_sequnlock(&sde->waitlock); in defer_packet_queue()
512 req->sde = sdma_select_user_engine(dd, selector, vl); in hfi1_user_sdma_process_request()
[all …]
H A Daffinity.c732 struct sdma_engine *sde = msix->arg; in hfi1_update_sdma_affinity() local
733 struct hfi1_devdata *dd = sde->dd; in hfi1_update_sdma_affinity()
738 if (cpu > num_online_cpus() || cpu == sde->cpu) in hfi1_update_sdma_affinity()
746 old_cpu = sde->cpu; in hfi1_update_sdma_affinity()
747 sde->cpu = cpu; in hfi1_update_sdma_affinity()
752 sde->this_idx, cpu); in hfi1_update_sdma_affinity()
830 struct sdma_engine *sde = NULL; in get_irq_affinity() local
842 sde = (struct sdma_engine *)msix->arg; in get_irq_affinity()
843 scnprintf(extra, 64, "engine %u", sde->this_idx); in get_irq_affinity()
893 sde->cpu = cpu; in get_irq_affinity()
H A Dverbs_txreq.h23 struct sdma_engine *sde; member
50 tx->sde = priv->s_sde; in get_txreq()
H A Dmsix.h17 int msix_request_sdma_irq(struct sdma_engine *sde);
H A Dvnic.h44 struct sdma_engine *sde; member
H A Dipoib.h111 struct sdma_engine *sde; member
H A Dverbs.c655 struct sdma_engine *sde, in build_verbs_ulp_payload() argument
670 sde->dd, in build_verbs_ulp_payload()
719 struct sdma_engine *sde, in build_verbs_tx_desc() argument
753 sde->dd, in build_verbs_tx_desc()
774 ret = build_verbs_ulp_payload(sde, length, tx); in build_verbs_tx_desc()
781 ret = sdma_txadd_daddr(sde->dd, &tx->txreq, sde->dd->sdma_pad_phys, in build_verbs_tx_desc()
848 ret = build_verbs_tx_desc(tx->sde, len, tx, ahg_info, pbc); in hfi1_verbs_send_dma()
852 ret = sdma_send_txreq(tx->sde, ps->wait, &tx->txreq, ps->pkts_sent); in hfi1_verbs_send_dma()
H A Dvnic_main.c284 struct sdma_engine *sde; in hfi1_vnic_select_queue() local
287 sde = sdma_select_engine_vl(vinfo->dd, mdata->entropy, mdata->vl); in hfi1_vnic_select_queue()
288 return sde->this_idx; in hfi1_vnic_select_queue()
H A Diowait.c44 int (*sleep)(struct sdma_engine *sde, in iowait_init() argument
H A Duser_sdma.h134 struct sdma_engine *sde; member
H A Dchip.c6008 struct sdma_engine *sde; in handle_sdma_eng_err() local
6011 sde = &dd->per_sdma[source]; in handle_sdma_eng_err()
6013 dd_dev_err(sde->dd, "CONFIG SDMA(%u) %s:%d %s()\n", sde->this_idx, in handle_sdma_eng_err()
6015 dd_dev_err(sde->dd, "CONFIG SDMA(%u) source: %u status 0x%llx\n", in handle_sdma_eng_err()
6016 sde->this_idx, source, (unsigned long long)status); in handle_sdma_eng_err()
6018 sde->err_cnt++; in handle_sdma_eng_err()
6019 sdma_engine_error(sde, status); in handle_sdma_eng_err()
6038 struct sdma_engine *sde = &dd->per_sdma[source]; in is_sdma_eng_err_int() local
6040 dd_dev_err(dd, "CONFIG SDMA(%u) %s:%d %s()\n", sde->this_idx, in is_sdma_eng_err_int()
6042 dd_dev_err(dd, "CONFIG SDMA(%u) source: %u\n", sde->this_idx, in is_sdma_eng_err_int()
[all …]
H A Duc.c229 ps->s_txreq->sde = priv->s_sde; in hfi1_make_uc_req()
H A Dud.c522 ps->s_txreq->sde = priv->s_sde; in hfi1_make_ud_req()
/linux/drivers/net/fddi/skfp/
H A Dsmt.c74 static void smt_fill_sde(struct s_smc *smc, struct smt_p_sde *sde);
938 smt_fill_sde(smc,&nif->sde) ; /* set station descriptor */ in smt_send_nif()
1030 smt_fill_sde(smc,&sif->sde) ; /* set station descriptor */ in smt_send_sif_config()
1157 static void smt_fill_sde(struct s_smc *smc, struct smt_p_sde *sde) in smt_fill_sde() argument
1159 SMTSETPARA(sde,SMT_P_SDE) ; in smt_fill_sde()
1160 sde->sde_non_master = smc->mib.fddiSMTNonMaster_Ct ; in smt_fill_sde()
1161 sde->sde_master = smc->mib.fddiSMTMaster_Ct ; in smt_fill_sde()
1162 sde->sde_mac_count = NUMMACS ; /* only 1 MAC */ in smt_fill_sde()
1164 sde->sde_type = SMT_SDE_CONCENTRATOR ; in smt_fill_sde()
1166 sde->sde_type = SMT_SDE_STATION ; in smt_fill_sde()
/linux/drivers/net/fddi/skfp/h/
H A Dsmt.h727 struct smt_p_sde sde ; /* station descriptor */ member
740 struct smt_p_sde sde ; /* station descriptor */ member
/linux/arch/arm64/boot/dts/qcom/
H A Dsdm845-sony-xperia-tama.dtsi686 sde_dsi_active: sde-dsi-active-state {
693 sde_dsi_sleep: sde-dsi-sleep-state {
708 sde_te_active_sleep: sde-te-active-sleep-state {
/linux/arch/mips/include/asm/octeon/
H A Dcvmx-pciercx-defs.h264 __BITFIELD_FIELD(uint32_t sde:1,
/linux/Documentation/arch/m68k/
H A Dkernel-options.rst83 /dev/sde: -> 0x0840 (fifth SCSI disk)
114 /dev/sde are in the table above, but not /dev/sdf. Although, you can

12