| /linux/include/soc/fsl/ |
| H A D | dpaa2-global.h | 20 struct dq { struct 33 } dq; member 68 static inline u32 dpaa2_dq_flags(const struct dpaa2_dq *dq) in dpaa2_dq_flags() argument 70 return dq->dq.stat; in dpaa2_dq_flags() 80 static inline int dpaa2_dq_is_pull(const struct dpaa2_dq *dq) in dpaa2_dq_is_pull() argument 82 return (int)(dpaa2_dq_flags(dq) & DPAA2_DQ_STAT_VOLATILE); in dpaa2_dq_is_pull() 91 static inline bool dpaa2_dq_is_pull_complete(const struct dpaa2_dq *dq) in dpaa2_dq_is_pull_complete() argument 93 return !!(dpaa2_dq_flags(dq) & DPAA2_DQ_STAT_EXPIRED); in dpaa2_dq_is_pull_complete() 104 static inline u16 dpaa2_dq_seqnum(const struct dpaa2_dq *dq) in dpaa2_dq_seqnum() argument 106 return le16_to_cpu(dq->dq.seqnum); in dpaa2_dq_seqnum() [all …]
|
| /linux/lib/raid6/ |
| H A D | recov_loongarch_simd.c | 32 u8 *p, *q, *dp, *dq; in raid6_2data_recov_lsx() local 47 dq = (u8 *)ptrs[failb]; in raid6_2data_recov_lsx() 49 ptrs[disks - 1] = dq; in raid6_2data_recov_lsx() 55 ptrs[failb] = dq; in raid6_2data_recov_lsx() 81 asm volatile("vld $vr8, %0" : : "m" (dq[0])); in raid6_2data_recov_lsx() 82 asm volatile("vld $vr9, %0" : : "m" (dq[16])); in raid6_2data_recov_lsx() 83 asm volatile("vld $vr10, %0" : : "m" (dq[32])); in raid6_2data_recov_lsx() 84 asm volatile("vld $vr11, %0" : : "m" (dq[48])); in raid6_2data_recov_lsx() 161 asm volatile("vst $vr4, %0" : "=m" (dq[0])); in raid6_2data_recov_lsx() 162 asm volatile("vst $vr5, %0" : "=m" (dq[16])); in raid6_2data_recov_lsx() [all …]
|
| H A D | recov_avx2.c | 19 u8 *p, *q, *dp, *dq; in raid6_2data_recov_avx2() local 33 dq = (u8 *)ptrs[failb]; in raid6_2data_recov_avx2() 35 ptrs[disks-1] = dq; in raid6_2data_recov_avx2() 41 ptrs[failb] = dq; in raid6_2data_recov_avx2() 61 asm volatile("vpxor %0, %%ymm1, %%ymm1" : : "m" (dq[0])); in raid6_2data_recov_avx2() 62 asm volatile("vpxor %0, %%ymm9, %%ymm9" : : "m" (dq[32])); in raid6_2data_recov_avx2() 120 asm volatile("vmovdqa %%ymm1, %0" : "=m" (dq[0])); in raid6_2data_recov_avx2() 121 asm volatile("vmovdqa %%ymm13,%0" : "=m" (dq[32])); in raid6_2data_recov_avx2() 132 dq += 64; in raid6_2data_recov_avx2() 136 asm volatile("vpxor %0, %%ymm1, %%ymm1" : : "m" (*dq)); in raid6_2data_recov_avx2() [all …]
|
| H A D | recov_avx512.c | 25 u8 *p, *q, *dp, *dq; in raid6_2data_recov_avx512() local 42 dq = (u8 *)ptrs[failb]; in raid6_2data_recov_avx512() 44 ptrs[disks-1] = dq; in raid6_2data_recov_avx512() 50 ptrs[failb] = dq; in raid6_2data_recov_avx512() 76 "m" (p[64]), "m" (dq[0]), "m" (dq[64]), in raid6_2data_recov_avx512() 146 : "m" (dq[0]), "m" (dq[64])); in raid6_2data_recov_avx512() 157 dq += 128; in raid6_2data_recov_avx512() 164 : "m" (*q), "m" (*p), "m"(*dq), "m" (*dp)); in raid6_2data_recov_avx512() 207 : "m" (dq[0])); in raid6_2data_recov_avx512() 218 dq += 64; in raid6_2data_recov_avx512() [all …]
|
| H A D | recov_ssse3.c | 19 u8 *p, *q, *dp, *dq; in raid6_2data_recov_ssse3() local 35 dq = (u8 *)ptrs[failb]; in raid6_2data_recov_ssse3() 37 ptrs[disks-1] = dq; in raid6_2data_recov_ssse3() 43 ptrs[failb] = dq; in raid6_2data_recov_ssse3() 71 asm volatile("pxor %0,%%xmm1" : : "m" (dq[0])); in raid6_2data_recov_ssse3() 72 asm volatile("pxor %0,%%xmm9" : : "m" (dq[16])); in raid6_2data_recov_ssse3() 124 asm volatile("movdqa %%xmm1,%0" : "=m" (dq[0])); in raid6_2data_recov_ssse3() 125 asm volatile("movdqa %%xmm9,%0" : "=m" (dq[16])); in raid6_2data_recov_ssse3() 136 dq += 32; in raid6_2data_recov_ssse3() 140 asm volatile("pxor %0,%%xmm1" : : "m" (*dq)); in raid6_2data_recov_ssse3() [all …]
|
| H A D | recov_s390xc.c | 25 u8 *p, *q, *dp, *dq; in raid6_2data_recov_s390xc() local 39 dq = (u8 *)ptrs[failb]; in raid6_2data_recov_s390xc() 41 ptrs[disks-1] = dq; in raid6_2data_recov_s390xc() 47 ptrs[failb] = dq; in raid6_2data_recov_s390xc() 58 xor_block(dq, q); in raid6_2data_recov_s390xc() 60 dq[i] = pbmul[dp[i]] ^ qmul[dq[i]]; in raid6_2data_recov_s390xc() 61 xor_block(dp, dq); in raid6_2data_recov_s390xc() 65 dq += 256; in raid6_2data_recov_s390xc() 74 u8 *p, *q, *dq; in raid6_datap_recov_s390xc() local 83 dq = (u8 *)ptrs[faila]; in raid6_datap_recov_s390xc() [all …]
|
| H A D | recov_rvv.c | 11 u8 *dq, const u8 *pbmul, in __raid6_2data_recov_rvv() argument 74 [dq]"r"(dq), in __raid6_2data_recov_rvv() 85 dq += 16; in __raid6_2data_recov_rvv() 90 u8 *dq, const u8 *qmul) in __raid6_datap_recov_rvv() argument 133 [dq]"r"(dq), in __raid6_datap_recov_rvv() 141 dq += 16; in __raid6_datap_recov_rvv() 148 u8 *p, *q, *dp, *dq; in raid6_2data_recov_rvv() local 163 dq = (u8 *)ptrs[failb]; in raid6_2data_recov_rvv() 165 ptrs[disks - 1] = dq; in raid6_2data_recov_rvv() 171 ptrs[failb] = dq; in raid6_2data_recov_rvv() [all …]
|
| H A D | recov.c | 22 u8 *p, *q, *dp, *dq; in raid6_2data_recov_intx1() local 36 dq = (u8 *)ptrs[failb]; in raid6_2data_recov_intx1() 38 ptrs[disks-1] = dq; in raid6_2data_recov_intx1() 44 ptrs[failb] = dq; in raid6_2data_recov_intx1() 55 qx = qmul[*q ^ *dq]; in raid6_2data_recov_intx1() 56 *dq++ = db = pbmul[px] ^ qx; /* Reconstructed B */ in raid6_2data_recov_intx1() 66 u8 *p, *q, *dq; in raid6_datap_recov_intx1() local 74 dq = (u8 *)ptrs[faila]; in raid6_datap_recov_intx1() 76 ptrs[disks-1] = dq; in raid6_datap_recov_intx1() 81 ptrs[faila] = dq; in raid6_datap_recov_intx1() [all …]
|
| H A D | recov_neon.c | 25 u8 *p, *q, *dp, *dq; in raid6_2data_recov_neon() local 40 dq = (u8 *)ptrs[failb]; in raid6_2data_recov_neon() 42 ptrs[disks - 1] = dq; in raid6_2data_recov_neon() 48 ptrs[failb] = dq; in raid6_2data_recov_neon() 58 __raid6_2data_recov_neon(bytes, p, q, dp, dq, pbmul, qmul); in raid6_2data_recov_neon() 64 u8 *p, *q, *dq; in raid6_datap_recov_neon() local 74 dq = (u8 *)ptrs[faila]; in raid6_datap_recov_neon() 76 ptrs[disks - 1] = dq; in raid6_datap_recov_neon() 81 ptrs[faila] = dq; in raid6_datap_recov_neon() 88 __raid6_datap_recov_neon(bytes, p, q, dq, qmul); in raid6_datap_recov_neon()
|
| H A D | recov_neon_inner.c | 29 uint8_t *dq, const uint8_t *pbmul, in __raid6_2data_recov_neon() argument 54 vx = veorq_u8(vld1q_u8(q), vld1q_u8(dq)); in __raid6_2data_recov_neon() 67 vst1q_u8(dq, db); in __raid6_2data_recov_neon() 74 dq += 16; in __raid6_2data_recov_neon() 78 void __raid6_datap_recov_neon(int bytes, uint8_t *p, uint8_t *q, uint8_t *dq, in __raid6_datap_recov_neon() argument 95 vx = veorq_u8(vld1q_u8(q), vld1q_u8(dq)); in __raid6_datap_recov_neon() 103 vst1q_u8(dq, vx); in __raid6_datap_recov_neon() 109 dq += 16; in __raid6_datap_recov_neon()
|
| H A D | neon.h | 16 uint8_t *dq, const uint8_t *pbmul, 19 void __raid6_datap_recov_neon(int bytes, uint8_t *p, uint8_t *q, uint8_t *dq,
|
| /linux/drivers/soc/fsl/dpio/ |
| H A D | qbman-portal.h | 219 void qbman_swp_dqrr_consume(struct qbman_swp *s, const struct dpaa2_dq *dq); 221 int qbman_result_has_new_result(struct qbman_swp *p, const struct dpaa2_dq *dq); 308 static inline int qbman_result_is_DQ(const struct dpaa2_dq *dq) in qbman_result_is_DQ() argument 310 return ((dq->dq.verb & QBMAN_RESULT_MASK) == QBMAN_RESULT_DQ); in qbman_result_is_DQ() 318 static inline int qbman_result_is_SCN(const struct dpaa2_dq *dq) in qbman_result_is_SCN() argument 320 return !qbman_result_is_DQ(dq); in qbman_result_is_SCN() 324 static inline int qbman_result_is_FQDAN(const struct dpaa2_dq *dq) in qbman_result_is_FQDAN() argument 326 return ((dq->dq.verb & QBMAN_RESULT_MASK) == QBMAN_RESULT_FQDAN); in qbman_result_is_FQDAN() 330 static inline int qbman_result_is_CDAN(const struct dpaa2_dq *dq) in qbman_result_is_CDAN() argument 332 return ((dq->dq.verb & QBMAN_RESULT_MASK) == QBMAN_RESULT_CDAN); in qbman_result_is_CDAN() [all …]
|
| H A D | dpio-service.c | 228 const struct dpaa2_dq *dq; in dpaa2_io_irq() local 240 dq = qbman_swp_dqrr_next(swp); in dpaa2_io_irq() 241 while (dq) { in dpaa2_io_irq() 242 if (qbman_result_is_SCN(dq)) { in dpaa2_io_irq() 246 q64 = qbman_result_SCN_ctx(dq); in dpaa2_io_irq() 252 qbman_swp_dqrr_consume(swp, dq); in dpaa2_io_irq() 256 dq = qbman_swp_dqrr_next(swp); in dpaa2_io_irq()
|
| H A D | qbman-portal.c | 1218 verb = p->dq.verb; in qbman_swp_dqrr_next_direct() 1246 flags = p->dq.stat; in qbman_swp_dqrr_next_direct() 1310 verb = p->dq.verb; in qbman_swp_dqrr_next_mem_back() 1338 flags = p->dq.stat; in qbman_swp_dqrr_next_mem_back() 1356 void qbman_swp_dqrr_consume(struct qbman_swp *s, const struct dpaa2_dq *dq) in qbman_swp_dqrr_consume() argument 1358 qbman_write_register(s, QBMAN_CINH_SWP_DCAP, QBMAN_IDX_FROM_DQRR(dq)); in qbman_swp_dqrr_consume() 1378 int qbman_result_has_new_result(struct qbman_swp *s, const struct dpaa2_dq *dq) in qbman_result_has_new_result() argument 1380 if (dq->dq.tok != QMAN_DQ_TOKEN_VALID) in qbman_result_has_new_result() 1388 ((struct dpaa2_dq *)dq)->dq.tok = 0; in qbman_result_has_new_result() 1395 if (s->vdq.storage == dq) { in qbman_result_has_new_result()
|
| /linux/crypto/async_tx/ |
| H A D | async_raid6_recov.c | 208 struct page *p, *q, *g, *dp, *dq; in __2data_recov_5() local 244 dq = blocks[failb]; in __2data_recov_5() 250 tx = async_mult(dq, dq_off, g, g_off, in __2data_recov_5() 263 srcs[0] = dq; in __2data_recov_5() 269 tx = async_xor_offs(dq, dq_off, srcs, src_offs, 2, bytes, submit); in __2data_recov_5() 274 srcs[1] = dq; in __2data_recov_5() 279 tx = async_sum_product(dq, dq_off, srcs, src_offs, coef, bytes, submit); in __2data_recov_5() 284 srcs[1] = dq; in __2data_recov_5() 299 struct page *p, *q, *dp, *dq; in __2data_recov_n() local 323 dq = blocks[failb]; in __2data_recov_n() [all …]
|
| /linux/drivers/net/ethernet/cavium/liquidio/ |
| H A D | cn66xx_regs.h | 311 #define CN6XXX_DMA_CNT(dq) \ argument 312 (CN6XXX_DMA_CNT_START + ((dq) * CN6XXX_DMA_OFFSET)) 314 #define CN6XXX_DMA_INT_LEVEL(dq) \ argument 315 (CN6XXX_DMA_INT_LEVEL_START + ((dq) * CN6XXX_DMA_OFFSET)) 317 #define CN6XXX_DMA_PKT_INT_LEVEL(dq) \ argument 318 (CN6XXX_DMA_INT_LEVEL_START + ((dq) * CN6XXX_DMA_OFFSET)) 320 #define CN6XXX_DMA_TIME_INT_LEVEL(dq) \ argument 321 (CN6XXX_DMA_INT_LEVEL_START + 4 + ((dq) * CN6XXX_DMA_OFFSET)) 323 #define CN6XXX_DMA_TIM(dq) \ argument 324 (CN6XXX_DMA_TIM_START + ((dq) * CN6XXX_DMA_OFFSET))
|
| H A D | cn23xx_pf_regs.h | 366 #define CN23XX_DMA_CNT(dq) \ argument 367 (CN23XX_DMA_CNT_START + ((dq) * CN23XX_DMA_OFFSET)) 369 #define CN23XX_DMA_INT_LEVEL(dq) \ argument 370 (CN23XX_DMA_INT_LEVEL_START + ((dq) * CN23XX_DMA_OFFSET)) 372 #define CN23XX_DMA_PKT_INT_LEVEL(dq) \ argument 373 (CN23XX_DMA_INT_LEVEL_START + ((dq) * CN23XX_DMA_OFFSET)) 375 #define CN23XX_DMA_TIME_INT_LEVEL(dq) \ argument 376 (CN23XX_DMA_INT_LEVEL_START + 4 + ((dq) * CN23XX_DMA_OFFSET)) 378 #define CN23XX_DMA_TIM(dq) \ argument 379 (CN23XX_DMA_TIM_START + ((dq) * CN23XX_DMA_OFFSET))
|
| /linux/fs/xfs/scrub/ |
| H A D | quotacheck.c | 555 struct xfs_dquot *dq) in xqcheck_compare_dquot() argument 566 mutex_lock(&dq->q_qlock); in xqcheck_compare_dquot() 568 error = xfarray_load_sparse(counts, dq->q_id, &xcdq); in xqcheck_compare_dquot() 572 if (xcdq.icount != dq->q_ino.count) in xqcheck_compare_dquot() 573 xchk_qcheck_set_corrupt(xqc->sc, dqtype, dq->q_id); in xqcheck_compare_dquot() 575 if (xcdq.bcount != dq->q_blk.count) in xqcheck_compare_dquot() 576 xchk_qcheck_set_corrupt(xqc->sc, dqtype, dq->q_id); in xqcheck_compare_dquot() 578 if (xcdq.rtbcount != dq->q_rtb.count) in xqcheck_compare_dquot() 579 xchk_qcheck_set_corrupt(xqc->sc, dqtype, dq->q_id); in xqcheck_compare_dquot() 582 error = xfarray_store(counts, dq->q_id, &xcdq); in xqcheck_compare_dquot() [all …]
|
| /linux/fs/xfs/ |
| H A D | xfs_dquot.c | 120 struct xfs_dquot *dq) in xfs_qm_adjust_dqlimits() argument 122 struct xfs_mount *mp = dq->q_mount; in xfs_qm_adjust_dqlimits() 127 ASSERT(dq->q_id); in xfs_qm_adjust_dqlimits() 128 defq = xfs_get_defquota(q, xfs_dquot_type(dq)); in xfs_qm_adjust_dqlimits() 130 if (!dq->q_blk.softlimit) { in xfs_qm_adjust_dqlimits() 131 dq->q_blk.softlimit = defq->blk.soft; in xfs_qm_adjust_dqlimits() 134 if (!dq->q_blk.hardlimit) { in xfs_qm_adjust_dqlimits() 135 dq->q_blk.hardlimit = defq->blk.hard; in xfs_qm_adjust_dqlimits() 138 if (!dq->q_ino.softlimit) in xfs_qm_adjust_dqlimits() 139 dq->q_ino.softlimit = defq->ino.soft; in xfs_qm_adjust_dqlimits() [all …]
|
| /linux/drivers/scsi/hisi_sas/ |
| H A D | hisi_sas_main.c | 432 struct hisi_sas_dq *dq, in hisi_sas_task_deliver() argument 440 spin_lock(&dq->lock); in hisi_sas_task_deliver() 441 wr_q_index = dq->wr_point; in hisi_sas_task_deliver() 442 dq->wr_point = (dq->wr_point + 1) % HISI_SAS_QUEUE_SLOTS; in hisi_sas_task_deliver() 443 list_add_tail(&slot->delivery, &dq->list); in hisi_sas_task_deliver() 444 spin_unlock(&dq->lock); in hisi_sas_task_deliver() 449 dlvry_queue = dq->id; in hisi_sas_task_deliver() 488 spin_lock(&dq->lock); in hisi_sas_task_deliver() 489 hisi_hba->hw->start_delivery(dq); in hisi_sas_task_deliver() 490 spin_unlock(&dq->lock); in hisi_sas_task_deliver() [all …]
|
| /linux/drivers/media/common/saa7146/ |
| H A D | saa7146_vbi.c | 293 struct saa7146_dmaqueue *dq = &dev->vv_data->vbi_dmaq; in return_buffers() local 296 if (dq->curr) { in return_buffers() 297 buf = dq->curr; in return_buffers() 298 dq->curr = NULL; in return_buffers() 301 while (!list_empty(&dq->queue)) { in return_buffers() 302 buf = list_entry(dq->queue.next, struct saa7146_buf, list); in return_buffers()
|
| H A D | saa7146_video.c | 638 struct saa7146_dmaqueue *dq = &dev->vv_data->video_dmaq; in return_buffers() local 641 if (dq->curr) { in return_buffers() 642 buf = dq->curr; in return_buffers() 643 dq->curr = NULL; in return_buffers() 646 while (!list_empty(&dq->queue)) { in return_buffers() 647 buf = list_entry(dq->queue.next, struct saa7146_buf, list); in return_buffers() 669 struct saa7146_dmaqueue *dq = &dev->vv_data->video_dmaq; in stop_streaming() local 671 timer_delete(&dq->timeout); in stop_streaming()
|
| /linux/drivers/soc/fsl/qbman/ |
| H A D | qman_test_api.c | 212 const struct qm_dqrr_entry *dq, in cb_dqrr() argument 215 if (WARN_ON(fd_neq(&fd_dq, &dq->fd))) { in cb_dqrr() 220 if (!(dq->stat & QM_DQRR_STAT_UNSCHEDULED) && !fd_neq(&fd_dq, &fd)) { in cb_dqrr()
|
| /linux/tools/testing/selftests/rcutorture/bin/ |
| H A D | kvm-transform.sh | 56 awk -v dq='"' '/./ { print "\tbootarg[" NR "] = " dq $1 dq ";" }' >> $T/bootarg.awk 58 awk -v dq='"' '/./ { print "\tbootpar[" NR "] = " dq $1 dq ";" }' >> $T/bootarg.awk
|
| /linux/drivers/ata/ |
| H A D | sata_nv.c | 1687 struct defer_queue *dq = &pp->defer_queue; in nv_swncq_qc_to_dq() local 1690 WARN_ON(dq->tail - dq->head == ATA_MAX_QUEUE); in nv_swncq_qc_to_dq() 1691 dq->defer_bits |= (1 << qc->hw_tag); in nv_swncq_qc_to_dq() 1692 dq->tag[dq->tail++ & (ATA_MAX_QUEUE - 1)] = qc->hw_tag; in nv_swncq_qc_to_dq() 1698 struct defer_queue *dq = &pp->defer_queue; in nv_swncq_qc_from_dq() local 1701 if (dq->head == dq->tail) /* null queue */ in nv_swncq_qc_from_dq() 1704 tag = dq->tag[dq->head & (ATA_MAX_QUEUE - 1)]; in nv_swncq_qc_from_dq() 1705 dq->tag[dq->head++ & (ATA_MAX_QUEUE - 1)] = ATA_TAG_POISON; in nv_swncq_qc_from_dq() 1706 WARN_ON(!(dq->defer_bits & (1 << tag))); in nv_swncq_qc_from_dq() 1707 dq->defer_bits &= ~(1 << tag); in nv_swncq_qc_from_dq() [all …]
|