/linux/include/soc/fsl/ |
H A D | dpaa2-global.h | 20 struct dq { struct 33 } dq; member 66 * @dq: the dequeue result. 68 static inline u32 dpaa2_dq_flags(const struct dpaa2_dq *dq) in dpaa2_dq_flags() argument 70 return dq->dq.stat; in dpaa2_dq_flags() 74 * dpaa2_dq_is_pull() - Check whether the dq response is from a pull 76 * @dq: the dequeue result 80 static inline int dpaa2_dq_is_pull(const struct dpaa2_dq *dq) in dpaa2_dq_is_pull() argument 82 return (int)(dpaa2_dq_flags(dq) & DPAA2_DQ_STAT_VOLATILE); in dpaa2_dq_is_pull() 87 * @dq: the dequeue result [all …]
|
/linux/fs/xfs/scrub/ |
H A D | quota_repair.c | 58 struct xfs_dquot *dq, in xrep_quota_item_fill_bmap_hole() argument 74 error = xfs_bmapi_write(sc->tp, sc->ip, dq->q_fileoffset, in xrep_quota_item_fill_bmap_hole() 80 dq->q_blkno = XFS_FSB_TO_DADDR(mp, irec->br_startblock); in xrep_quota_item_fill_bmap_hole() 82 trace_xrep_dquot_item_fill_bmap_hole(sc->mp, dq->q_type, dq->q_id); in xrep_quota_item_fill_bmap_hole() 85 error = xfs_trans_get_buf(sc->tp, mp->m_ddev_targp, dq->q_blkno, in xrep_quota_item_fill_bmap_hole() 91 xfs_qm_init_dquot_blk(sc->tp, dq->q_id, dq->q_type, bp); in xrep_quota_item_fill_bmap_hole() 108 struct xfs_dquot *dq, in xrep_quota_item_bmap() argument 114 xfs_fileoff_t offset = dq->q_id / qi->qi_dqperchunk; in xrep_quota_item_bmap() 123 dq->q_fileoffset = offset; in xrep_quota_item_bmap() 131 error = xrep_quota_item_fill_bmap_hole(sc, dq, &irec); in xrep_quota_item_bmap() [all …]
|
H A D | quota.c | 23 /* Convert a scrub type code to a DQ flag, or return 0 if error. */ 84 struct xfs_dquot *dq, in xchk_quota_item_bmap() argument 97 if (dq->q_fileoffset != offset) { in xchk_quota_item_bmap() 113 if (XFS_FSB_TO_DADDR(mp, irec.br_startblock) != dq->q_blkno) in xchk_quota_item_bmap() 142 struct xfs_dquot *dq) in xchk_quota_item() argument 161 xfs_dqunlock(dq); in xchk_quota_item() 163 xfs_dqlock(dq); in xchk_quota_item() 169 offset = dq->q_id / qi->qi_dqperchunk; in xchk_quota_item() 170 if (dq->q_id && dq->q_id <= sqi->last_id) in xchk_quota_item() 173 sqi->last_id = dq->q_id; in xchk_quota_item() [all …]
|
H A D | quotacheck_repair.c | 47 struct xfs_dquot *dq) in xqcheck_commit_dquot() argument 56 xfs_dqunlock(dq); in xqcheck_commit_dquot() 58 xfs_dqlock(dq); in xqcheck_commit_dquot() 62 xfs_trans_dqjoin(xqc->sc->tp, dq); in xqcheck_commit_dquot() 70 error = xfarray_load_sparse(counts, dq->q_id, &xcdq); in xqcheck_commit_dquot() 75 delta = (int64_t)xcdq.icount - dq->q_ino.count; in xqcheck_commit_dquot() 77 dq->q_ino.reserved += delta; in xqcheck_commit_dquot() 78 dq->q_ino.count += delta; in xqcheck_commit_dquot() 82 delta = (int64_t)xcdq.bcount - dq->q_blk.count; in xqcheck_commit_dquot() 84 dq->q_blk.reserved += delta; in xqcheck_commit_dquot() [all …]
|
H A D | quotacheck.c | 555 struct xfs_dquot *dq) in xqcheck_compare_dquot() argument 567 error = xfarray_load_sparse(counts, dq->q_id, &xcdq); in xqcheck_compare_dquot() 571 if (xcdq.icount != dq->q_ino.count) in xqcheck_compare_dquot() 572 xchk_qcheck_set_corrupt(xqc->sc, dqtype, dq->q_id); in xqcheck_compare_dquot() 574 if (xcdq.bcount != dq->q_blk.count) in xqcheck_compare_dquot() 575 xchk_qcheck_set_corrupt(xqc->sc, dqtype, dq->q_id); in xqcheck_compare_dquot() 577 if (xcdq.rtbcount != dq->q_rtb.count) in xqcheck_compare_dquot() 578 xchk_qcheck_set_corrupt(xqc->sc, dqtype, dq->q_id); in xqcheck_compare_dquot() 581 error = xfarray_store(counts, dq->q_id, &xcdq); in xqcheck_compare_dquot() 616 struct xfs_dquot *dq; in xqcheck_walk_observations() local [all …]
|
H A D | dqiterate.c | 150 struct xfs_dquot *dq; in xchk_dquot_iter_advance_incore() local 156 nr_found = radix_tree_gang_lookup(tree, (void **)&dq, cursor->id, 1); in xchk_dquot_iter_advance_incore() 158 *next_incore_id = dq->q_id; in xchk_dquot_iter_advance_incore() 176 struct xfs_dquot *dq = NULL; in xchk_dquot_iter() local 204 error = xfs_qm_dqget(mp, cursor->id, cursor->dqtype, false, &dq); in xchk_dquot_iter() 208 cursor->id = dq->q_id + 1; in xchk_dquot_iter() 209 *dqpp = dq; in xchk_dquot_iter()
|
/linux/lib/raid6/ |
H A D | recov_rvv.c | 16 u8 *dq, const u8 *pbmul, in __raid6_2data_recov_rvv() argument 32 * qx = qmul[*q ^ *dq]; in __raid6_2data_recov_rvv() 33 * *dq++ = db = pbmul[px] ^ qx; in __raid6_2data_recov_rvv() 41 * v2:qx, v3:dq, in __raid6_2data_recov_rvv() 53 "vle8.v v3, (%[dq])\n" in __raid6_2data_recov_rvv() 61 "vxor.vv v2, v14, v15\n" /* v2 = qmul[*q ^ *dq] */ in __raid6_2data_recov_rvv() 72 "vse8.v v3, (%[dq])\n" in __raid6_2data_recov_rvv() 79 [dq]"r"(dq), in __raid6_2data_recov_rvv() 90 dq += 16; in __raid6_2data_recov_rvv() 95 u8 *dq, const u8 *qmul) in __raid6_datap_recov_rvv() argument [all …]
|
H A D | recov_avx512.c | 25 u8 *p, *q, *dp, *dq; in raid6_2data_recov_avx512() local 42 dq = (u8 *)ptrs[failb]; in raid6_2data_recov_avx512() 44 ptrs[disks-1] = dq; in raid6_2data_recov_avx512() 50 ptrs[failb] = dq; in raid6_2data_recov_avx512() 76 "m" (p[64]), "m" (dq[0]), "m" (dq[64]), in raid6_2data_recov_avx512() 80 * 1 = dq[0] ^ q[0] in raid6_2data_recov_avx512() 81 * 9 = dq[64] ^ q[64] in raid6_2data_recov_avx512() 138 * 1 = db = DQ in raid6_2data_recov_avx512() 139 * 13 = db[64] = DQ[64] in raid6_2data_recov_avx512() 146 : "m" (dq[0]), "m" (dq[64])); in raid6_2data_recov_avx512() [all …]
|
H A D | recov_avx2.c | 19 u8 *p, *q, *dp, *dq; in raid6_2data_recov_avx2() local 33 dq = (u8 *)ptrs[failb]; in raid6_2data_recov_avx2() 35 ptrs[disks-1] = dq; in raid6_2data_recov_avx2() 41 ptrs[failb] = dq; in raid6_2data_recov_avx2() 61 asm volatile("vpxor %0, %%ymm1, %%ymm1" : : "m" (dq[0])); in raid6_2data_recov_avx2() 62 asm volatile("vpxor %0, %%ymm9, %%ymm9" : : "m" (dq[32])); in raid6_2data_recov_avx2() 67 * 1 = dq[0] ^ q[0] in raid6_2data_recov_avx2() 68 * 9 = dq[32] ^ q[32] in raid6_2data_recov_avx2() 117 * 1 = db = DQ in raid6_2data_recov_avx2() 118 * 13 = db[32] = DQ[32] in raid6_2data_recov_avx2() [all …]
|
H A D | recov_s390xc.c | 25 u8 *p, *q, *dp, *dq; in raid6_2data_recov_s390xc() local 39 dq = (u8 *)ptrs[failb]; in raid6_2data_recov_s390xc() 41 ptrs[disks-1] = dq; in raid6_2data_recov_s390xc() 47 ptrs[failb] = dq; in raid6_2data_recov_s390xc() 58 xor_block(dq, q); in raid6_2data_recov_s390xc() 60 dq[i] = pbmul[dp[i]] ^ qmul[dq[i]]; in raid6_2data_recov_s390xc() 61 xor_block(dp, dq); in raid6_2data_recov_s390xc() 65 dq += 256; in raid6_2data_recov_s390xc() 74 u8 *p, *q, *dq; in raid6_datap_recov_s390xc() local 83 dq = (u8 *)ptrs[faila]; in raid6_datap_recov_s390xc() [all …]
|
H A D | recov_ssse3.c | 19 u8 *p, *q, *dp, *dq; in raid6_2data_recov_ssse3() local 35 dq = (u8 *)ptrs[failb]; in raid6_2data_recov_ssse3() 37 ptrs[disks-1] = dq; in raid6_2data_recov_ssse3() 43 ptrs[failb] = dq; in raid6_2data_recov_ssse3() 71 asm volatile("pxor %0,%%xmm1" : : "m" (dq[0])); in raid6_2data_recov_ssse3() 72 asm volatile("pxor %0,%%xmm9" : : "m" (dq[16])); in raid6_2data_recov_ssse3() 123 /* xmm1/9 = db = DQ */ in raid6_2data_recov_ssse3() 124 asm volatile("movdqa %%xmm1,%0" : "=m" (dq[0])); in raid6_2data_recov_ssse3() 125 asm volatile("movdqa %%xmm9,%0" : "=m" (dq[16])); in raid6_2data_recov_ssse3() 136 dq += 32; in raid6_2data_recov_ssse3() [all …]
|
H A D | recov_neon_inner.c | 29 uint8_t *dq, const uint8_t *pbmul, in __raid6_2data_recov_neon() argument 43 * qx = qmul[*q ^ *dq]; in __raid6_2data_recov_neon() 44 * *dq++ = db = pbmul[px] ^ qx; in __raid6_2data_recov_neon() 54 vx = veorq_u8(vld1q_u8(q), vld1q_u8(dq)); in __raid6_2data_recov_neon() 67 vst1q_u8(dq, db); in __raid6_2data_recov_neon() 74 dq += 16; in __raid6_2data_recov_neon() 78 void __raid6_datap_recov_neon(int bytes, uint8_t *p, uint8_t *q, uint8_t *dq, in __raid6_datap_recov_neon() argument 87 * *p++ ^= *dq = qmul[*q ^ *dq]; in __raid6_datap_recov_neon() 88 * q++; dq++; in __raid6_datap_recov_neon() 95 vx = veorq_u8(vld1q_u8(q), vld1q_u8(dq)); in __raid6_datap_recov_neon() [all …]
|
H A D | recov_loongarch_simd.c | 32 u8 *p, *q, *dp, *dq; in raid6_2data_recov_lsx() local 47 dq = (u8 *)ptrs[failb]; in raid6_2data_recov_lsx() 49 ptrs[disks - 1] = dq; in raid6_2data_recov_lsx() 55 ptrs[failb] = dq; in raid6_2data_recov_lsx() 81 asm volatile("vld $vr8, %0" : : "m" (dq[0])); in raid6_2data_recov_lsx() 82 asm volatile("vld $vr9, %0" : : "m" (dq[16])); in raid6_2data_recov_lsx() 83 asm volatile("vld $vr10, %0" : : "m" (dq[32])); in raid6_2data_recov_lsx() 84 asm volatile("vld $vr11, %0" : : "m" (dq[48])); in raid6_2data_recov_lsx() 161 asm volatile("vst $vr4, %0" : "=m" (dq[0])); in raid6_2data_recov_lsx() 162 asm volatile("vst $vr5, %0" : "=m" (dq[16])); in raid6_2data_recov_lsx() [all …]
|
H A D | recov.c | 22 u8 *p, *q, *dp, *dq; in raid6_2data_recov_intx1() local 36 dq = (u8 *)ptrs[failb]; in raid6_2data_recov_intx1() 38 ptrs[disks-1] = dq; in raid6_2data_recov_intx1() 44 ptrs[failb] = dq; in raid6_2data_recov_intx1() 55 qx = qmul[*q ^ *dq]; in raid6_2data_recov_intx1() 56 *dq++ = db = pbmul[px] ^ qx; /* Reconstructed B */ in raid6_2data_recov_intx1() 66 u8 *p, *q, *dq; in raid6_datap_recov_intx1() local 74 dq = (u8 *)ptrs[faila]; in raid6_datap_recov_intx1() 76 ptrs[disks-1] = dq; in raid6_datap_recov_intx1() 81 ptrs[faila] = dq; in raid6_datap_recov_intx1() [all …]
|
H A D | recov_neon.c | 26 u8 *p, *q, *dp, *dq; in raid6_2data_recov_neon() local 41 dq = (u8 *)ptrs[failb]; in raid6_2data_recov_neon() 43 ptrs[disks - 1] = dq; in raid6_2data_recov_neon() 49 ptrs[failb] = dq; in raid6_2data_recov_neon() 59 __raid6_2data_recov_neon(bytes, p, q, dp, dq, pbmul, qmul); in raid6_2data_recov_neon() 66 u8 *p, *q, *dq; in raid6_datap_recov_neon() local 76 dq = (u8 *)ptrs[faila]; in raid6_datap_recov_neon() 78 ptrs[disks - 1] = dq; in raid6_datap_recov_neon() 83 ptrs[faila] = dq; in raid6_datap_recov_neon() 90 __raid6_datap_recov_neon(bytes, p, q, dq, qmul); in raid6_datap_recov_neon()
|
/linux/drivers/soc/fsl/dpio/ |
H A D | qbman-portal.h | 219 void qbman_swp_dqrr_consume(struct qbman_swp *s, const struct dpaa2_dq *dq); 221 int qbman_result_has_new_result(struct qbman_swp *p, const struct dpaa2_dq *dq); 304 * @dq: the dequeue result to be checked 308 static inline int qbman_result_is_DQ(const struct dpaa2_dq *dq) in qbman_result_is_DQ() argument 310 return ((dq->dq.verb & QBMAN_RESULT_MASK) == QBMAN_RESULT_DQ); in qbman_result_is_DQ() 315 * @dq: the dequeue result to be checked 318 static inline int qbman_result_is_SCN(const struct dpaa2_dq *dq) in qbman_result_is_SCN() argument 320 return !qbman_result_is_DQ(dq); in qbman_result_is_SCN() 324 static inline int qbman_result_is_FQDAN(const struct dpaa2_dq *dq) in qbman_result_is_FQDAN() argument 326 return ((dq->dq.verb & QBMAN_RESULT_MASK) == QBMAN_RESULT_FQDAN); in qbman_result_is_FQDAN() [all …]
|
/linux/crypto/async_tx/ |
H A D | async_raid6_recov.c | 208 struct page *p, *q, *g, *dp, *dq; in __2data_recov_5() local 244 dq = blocks[failb]; in __2data_recov_5() 250 tx = async_mult(dq, dq_off, g, g_off, in __2data_recov_5() 263 srcs[0] = dq; in __2data_recov_5() 269 tx = async_xor_offs(dq, dq_off, srcs, src_offs, 2, bytes, submit); in __2data_recov_5() 274 srcs[1] = dq; in __2data_recov_5() 279 tx = async_sum_product(dq, dq_off, srcs, src_offs, coef, bytes, submit); in __2data_recov_5() 284 srcs[1] = dq; in __2data_recov_5() 299 struct page *p, *q, *dp, *dq; in __2data_recov_n() local 323 dq = blocks[failb]; in __2data_recov_n() [all …]
|
/linux/tools/memory-model/scripts/ |
H A D | runlitmushist.sh | 68 awk -v dq='"' -v hwfnseg="$hwfnseg" -v ncpu="$LKMM_JOBS" -v t="$T" ' 70 print "if test -z " dq hwfnseg dq " || scripts/simpletest.sh " dq $2 dq 72 print "\techo runtest " dq $2 dq " " hwfnseg " >> " t "/" NR % ncpu ".sh";
|
/linux/tools/testing/selftests/rcutorture/bin/ |
H A D | kvm-transform.sh | 56 awk -v dq='"' '/./ { print "\tbootarg[" NR "] = " dq $1 dq ";" }' >> $T/bootarg.awk 58 awk -v dq='"' '/./ { print "\tbootpar[" NR "] = " dq $1 dq ";" }' >> $T/bootarg.awk
|
/linux/fs/xfs/ |
H A D | xfs_dquot.c | 120 struct xfs_dquot *dq) in xfs_qm_adjust_dqlimits() argument 122 struct xfs_mount *mp = dq->q_mount; in xfs_qm_adjust_dqlimits() 127 ASSERT(dq->q_id); in xfs_qm_adjust_dqlimits() 128 defq = xfs_get_defquota(q, xfs_dquot_type(dq)); in xfs_qm_adjust_dqlimits() 130 if (!dq->q_blk.softlimit) { in xfs_qm_adjust_dqlimits() 131 dq->q_blk.softlimit = defq->blk.soft; in xfs_qm_adjust_dqlimits() 134 if (!dq->q_blk.hardlimit) { in xfs_qm_adjust_dqlimits() 135 dq->q_blk.hardlimit = defq->blk.hard; in xfs_qm_adjust_dqlimits() 138 if (!dq->q_ino.softlimit) in xfs_qm_adjust_dqlimits() 139 dq->q_ino.softlimit = defq->ino.soft; in xfs_qm_adjust_dqlimits() [all …]
|
/linux/drivers/net/ethernet/cavium/liquidio/ |
H A D | cn66xx_regs.h | 311 #define CN6XXX_DMA_CNT(dq) \ argument 312 (CN6XXX_DMA_CNT_START + ((dq) * CN6XXX_DMA_OFFSET)) 314 #define CN6XXX_DMA_INT_LEVEL(dq) \ argument 315 (CN6XXX_DMA_INT_LEVEL_START + ((dq) * CN6XXX_DMA_OFFSET)) 317 #define CN6XXX_DMA_PKT_INT_LEVEL(dq) \ argument 318 (CN6XXX_DMA_INT_LEVEL_START + ((dq) * CN6XXX_DMA_OFFSET)) 320 #define CN6XXX_DMA_TIME_INT_LEVEL(dq) \ argument 321 (CN6XXX_DMA_INT_LEVEL_START + 4 + ((dq) * CN6XXX_DMA_OFFSET)) 323 #define CN6XXX_DMA_TIM(dq) \ argument 324 (CN6XXX_DMA_TIM_START + ((dq) * CN6XXX_DMA_OFFSET))
|
H A D | cn23xx_pf_regs.h | 366 #define CN23XX_DMA_CNT(dq) \ argument 367 (CN23XX_DMA_CNT_START + ((dq) * CN23XX_DMA_OFFSET)) 369 #define CN23XX_DMA_INT_LEVEL(dq) \ argument 370 (CN23XX_DMA_INT_LEVEL_START + ((dq) * CN23XX_DMA_OFFSET)) 372 #define CN23XX_DMA_PKT_INT_LEVEL(dq) \ argument 373 (CN23XX_DMA_INT_LEVEL_START + ((dq) * CN23XX_DMA_OFFSET)) 375 #define CN23XX_DMA_TIME_INT_LEVEL(dq) \ argument 376 (CN23XX_DMA_INT_LEVEL_START + 4 + ((dq) * CN23XX_DMA_OFFSET)) 378 #define CN23XX_DMA_TIM(dq) \ argument 379 (CN23XX_DMA_TIM_START + ((dq) * CN23XX_DMA_OFFSET))
|
/linux/drivers/net/ethernet/mellanox/mlxsw/ |
H A D | cmd.h | 980 /* SW2HW_DQ - Software to Hardware DQ 982 * OpMod == 0 (send DQ) / OpMod == 1 (receive DQ) 983 * INMmod == DQ number 1051 /* HW2SW_DQ - Hardware to Software DQ 1053 * OpMod == 0 (send DQ) / OpMod == 1 (receive DQ) 1054 * INMmod == DQ number 1057 * software ownership. Incoming packets on the DQ are silently discarded, 1082 /* 2ERR_DQ - To Error DQ 1084 * OpMod == 0 (send DQ) / OpMod == 1 (receive DQ) 1085 * INMmod == DQ number [all …]
|
/linux/Documentation/devicetree/bindings/memory-controllers/ |
H A D | rockchip,rk3399-dmc.yaml | 162 When the DRAM type is DDR3, this parameter defines the PHY side DQ line 163 (including DQS/DQ/DM line) drive strength. 210 When the DRAM type is LPDDR3, this parameter defines the PHY side DQ line 211 (including DQS/DQ/DM line) drive strength. 242 DQS/DQ line strength in ohms. 273 When the DRAM type is LPDDR4, this parameter defines the PHY side DQ line 274 (including DQS/DQ/DM line) drive strength.
|
/linux/drivers/crypto/intel/qat/qat_common/ |
H A D | qat_asym_algs.c | 40 dma_addr_t dq; member 66 char *dq; member 877 qat_req->in.rsa.dec_crt.dq = ctx->dma_dq; in qat_rsa_dec() 1118 /* dq */ in qat_rsa_setkey_crt() 1119 ptr = rsa_key->dq; in qat_rsa_setkey_crt() 1124 ctx->dq = dma_alloc_coherent(dev, half_key_sz, &ctx->dma_dq, in qat_rsa_setkey_crt() 1126 if (!ctx->dq) in qat_rsa_setkey_crt() 1128 memcpy(ctx->dq + (half_key_sz - len), ptr, len); in qat_rsa_setkey_crt() 1146 memset(ctx->dq, '\0', half_key_sz); in qat_rsa_setkey_crt() 1147 dma_free_coherent(dev, half_key_sz, ctx->dq, ctx->dma_dq); in qat_rsa_setkey_crt() [all …]
|