/linux-4.19.296/include/soc/fsl/ |
D | dpaa2-global.h | 20 struct dq { struct 33 } dq; member 68 static inline u32 dpaa2_dq_flags(const struct dpaa2_dq *dq) in dpaa2_dq_flags() argument 70 return dq->dq.stat; in dpaa2_dq_flags() 80 static inline int dpaa2_dq_is_pull(const struct dpaa2_dq *dq) in dpaa2_dq_is_pull() argument 82 return (int)(dpaa2_dq_flags(dq) & DPAA2_DQ_STAT_VOLATILE); in dpaa2_dq_is_pull() 91 static inline bool dpaa2_dq_is_pull_complete(const struct dpaa2_dq *dq) in dpaa2_dq_is_pull_complete() argument 93 return !!(dpaa2_dq_flags(dq) & DPAA2_DQ_STAT_EXPIRED); in dpaa2_dq_is_pull_complete() 104 static inline u16 dpaa2_dq_seqnum(const struct dpaa2_dq *dq) in dpaa2_dq_seqnum() argument 106 return le16_to_cpu(dq->dq.seqnum); in dpaa2_dq_seqnum() [all …]
|
/linux-4.19.296/lib/raid6/ |
D | recov_avx2.c | 25 u8 *p, *q, *dp, *dq; in raid6_2data_recov_avx2() local 39 dq = (u8 *)ptrs[failb]; in raid6_2data_recov_avx2() 41 ptrs[disks-1] = dq; in raid6_2data_recov_avx2() 47 ptrs[failb] = dq; in raid6_2data_recov_avx2() 67 asm volatile("vpxor %0, %%ymm1, %%ymm1" : : "m" (dq[0])); in raid6_2data_recov_avx2() 68 asm volatile("vpxor %0, %%ymm9, %%ymm9" : : "m" (dq[32])); in raid6_2data_recov_avx2() 126 asm volatile("vmovdqa %%ymm1, %0" : "=m" (dq[0])); in raid6_2data_recov_avx2() 127 asm volatile("vmovdqa %%ymm13,%0" : "=m" (dq[32])); in raid6_2data_recov_avx2() 138 dq += 64; in raid6_2data_recov_avx2() 142 asm volatile("vpxor %0, %%ymm1, %%ymm1" : : "m" (*dq)); in raid6_2data_recov_avx2() [all …]
|
D | recov_avx512.c | 32 u8 *p, *q, *dp, *dq; in raid6_2data_recov_avx512() local 49 dq = (u8 *)ptrs[failb]; in raid6_2data_recov_avx512() 51 ptrs[disks-1] = dq; in raid6_2data_recov_avx512() 57 ptrs[failb] = dq; in raid6_2data_recov_avx512() 83 "m" (p[64]), "m" (dq[0]), "m" (dq[64]), in raid6_2data_recov_avx512() 153 : "m" (dq[0]), "m" (dq[64])); in raid6_2data_recov_avx512() 164 dq += 128; in raid6_2data_recov_avx512() 171 : "m" (*q), "m" (*p), "m"(*dq), "m" (*dp)); in raid6_2data_recov_avx512() 214 : "m" (dq[0])); in raid6_2data_recov_avx512() 225 dq += 64; in raid6_2data_recov_avx512() [all …]
|
D | recov_ssse3.c | 25 u8 *p, *q, *dp, *dq; in raid6_2data_recov_ssse3() local 41 dq = (u8 *)ptrs[failb]; in raid6_2data_recov_ssse3() 43 ptrs[disks-1] = dq; in raid6_2data_recov_ssse3() 49 ptrs[failb] = dq; in raid6_2data_recov_ssse3() 77 asm volatile("pxor %0,%%xmm1" : : "m" (dq[0])); in raid6_2data_recov_ssse3() 78 asm volatile("pxor %0,%%xmm9" : : "m" (dq[16])); in raid6_2data_recov_ssse3() 130 asm volatile("movdqa %%xmm1,%0" : "=m" (dq[0])); in raid6_2data_recov_ssse3() 131 asm volatile("movdqa %%xmm9,%0" : "=m" (dq[16])); in raid6_2data_recov_ssse3() 142 dq += 32; in raid6_2data_recov_ssse3() 146 asm volatile("pxor %0,%%xmm1" : : "m" (*dq)); in raid6_2data_recov_ssse3() [all …]
|
D | recov_s390xc.c | 26 u8 *p, *q, *dp, *dq; in raid6_2data_recov_s390xc() local 40 dq = (u8 *)ptrs[failb]; in raid6_2data_recov_s390xc() 42 ptrs[disks-1] = dq; in raid6_2data_recov_s390xc() 48 ptrs[failb] = dq; in raid6_2data_recov_s390xc() 59 xor_block(dq, q); in raid6_2data_recov_s390xc() 61 dq[i] = pbmul[dp[i]] ^ qmul[dq[i]]; in raid6_2data_recov_s390xc() 62 xor_block(dp, dq); in raid6_2data_recov_s390xc() 66 dq += 256; in raid6_2data_recov_s390xc() 75 u8 *p, *q, *dq; in raid6_datap_recov_s390xc() local 84 dq = (u8 *)ptrs[faila]; in raid6_datap_recov_s390xc() [all …]
|
D | recov.c | 28 u8 *p, *q, *dp, *dq; in raid6_2data_recov_intx1() local 42 dq = (u8 *)ptrs[failb]; in raid6_2data_recov_intx1() 44 ptrs[disks-1] = dq; in raid6_2data_recov_intx1() 50 ptrs[failb] = dq; in raid6_2data_recov_intx1() 61 qx = qmul[*q ^ *dq]; in raid6_2data_recov_intx1() 62 *dq++ = db = pbmul[px] ^ qx; /* Reconstructed B */ in raid6_2data_recov_intx1() 72 u8 *p, *q, *dq; in raid6_datap_recov_intx1() local 80 dq = (u8 *)ptrs[faila]; in raid6_datap_recov_intx1() 82 ptrs[disks-1] = dq; in raid6_datap_recov_intx1() 87 ptrs[faila] = dq; in raid6_datap_recov_intx1() [all …]
|
D | recov_neon.c | 27 uint8_t *dq, const uint8_t *pbmul, 30 void __raid6_datap_recov_neon(int bytes, uint8_t *p, uint8_t *q, uint8_t *dq, 36 u8 *p, *q, *dp, *dq; in raid6_2data_recov_neon() local 51 dq = (u8 *)ptrs[failb]; in raid6_2data_recov_neon() 53 ptrs[disks - 1] = dq; in raid6_2data_recov_neon() 59 ptrs[failb] = dq; in raid6_2data_recov_neon() 69 __raid6_2data_recov_neon(bytes, p, q, dp, dq, pbmul, qmul); in raid6_2data_recov_neon() 76 u8 *p, *q, *dq; in raid6_datap_recov_neon() local 86 dq = (u8 *)ptrs[faila]; in raid6_datap_recov_neon() 88 ptrs[disks - 1] = dq; in raid6_datap_recov_neon() [all …]
|
D | recov_neon_inner.c | 32 uint8_t *dq, const uint8_t *pbmul, in __raid6_2data_recov_neon() argument 57 vx = veorq_u8(vld1q_u8(q), vld1q_u8(dq)); in __raid6_2data_recov_neon() 70 vst1q_u8(dq, db); in __raid6_2data_recov_neon() 77 dq += 16; in __raid6_2data_recov_neon() 81 void __raid6_datap_recov_neon(int bytes, uint8_t *p, uint8_t *q, uint8_t *dq, in __raid6_datap_recov_neon() argument 98 vx = veorq_u8(vld1q_u8(q), vld1q_u8(dq)); in __raid6_datap_recov_neon() 106 vst1q_u8(dq, vx); in __raid6_datap_recov_neon() 112 dq += 16; in __raid6_datap_recov_neon()
|
/linux-4.19.296/crypto/async_tx/ |
D | async_raid6_recov.c | 203 struct page *p, *q, *g, *dp, *dq; in __2data_recov_5() local 233 dq = blocks[failb]; in __2data_recov_5() 238 tx = async_mult(dq, g, raid6_gfexp[good], bytes, submit); in __2data_recov_5() 248 srcs[0] = dq; in __2data_recov_5() 252 tx = async_xor(dq, srcs, 0, 2, bytes, submit); in __2data_recov_5() 256 srcs[1] = dq; in __2data_recov_5() 260 tx = async_sum_product(dq, srcs, coef, bytes, submit); in __2data_recov_5() 264 srcs[1] = dq; in __2data_recov_5() 277 struct page *p, *q, *dp, *dq; in __2data_recov_n() local 295 dq = blocks[failb]; in __2data_recov_n() [all …]
|
/linux-4.19.296/drivers/s390/crypto/ |
D | zcrypt_msgtype50.c | 110 unsigned char dq[64]; member 123 unsigned char dq[128]; member 136 unsigned char dq[256]; member 268 unsigned char *p, *q, *dp, *dq, *u, *inp; in ICACRT_msg_to_type50CRT_msg() local 290 dq = crb1->dq + sizeof(crb1->dq) - short_len; in ICACRT_msg_to_type50CRT_msg() 304 dq = crb2->dq + sizeof(crb2->dq) - short_len; in ICACRT_msg_to_type50CRT_msg() 319 dq = crb3->dq + sizeof(crb3->dq) - short_len; in ICACRT_msg_to_type50CRT_msg() 332 copy_from_user(dq, crt->bq_key, short_len) || in ICACRT_msg_to_type50CRT_msg()
|
D | zcrypt_cex2a.h | 82 unsigned char dq[64]; member 95 unsigned char dq[128]; member 108 unsigned char dq[256]; member
|
/linux-4.19.296/fs/xfs/ |
D | xfs_iomap.c | 296 struct xfs_dquot *dq = xfs_inode_dquot(ip, type); in xfs_quota_need_throttle() local 298 if (!dq || !xfs_this_quota_on(ip->i_mount, type)) in xfs_quota_need_throttle() 302 if (!dq->q_prealloc_hi_wmark) in xfs_quota_need_throttle() 306 if (dq->q_res_bcount + alloc_blocks < dq->q_prealloc_lo_wmark) in xfs_quota_need_throttle() 322 struct xfs_dquot *dq = xfs_inode_dquot(ip, type); in xfs_quota_calc_throttle() local 325 if (!dq || dq->q_res_bcount >= dq->q_prealloc_hi_wmark) { in xfs_quota_calc_throttle() 331 freesp = dq->q_prealloc_hi_wmark - dq->q_res_bcount; in xfs_quota_calc_throttle() 332 if (freesp < dq->q_low_space[XFS_QLOWSP_5_PCNT]) { in xfs_quota_calc_throttle() 334 if (freesp < dq->q_low_space[XFS_QLOWSP_3_PCNT]) in xfs_quota_calc_throttle() 336 if (freesp < dq->q_low_space[XFS_QLOWSP_1_PCNT]) in xfs_quota_calc_throttle()
|
D | xfs_dquot.c | 74 struct xfs_dquot *dq) in xfs_qm_adjust_dqlimits() argument 77 struct xfs_disk_dquot *d = &dq->q_core; in xfs_qm_adjust_dqlimits() 82 defq = xfs_get_defquota(dq, q); in xfs_qm_adjust_dqlimits() 102 xfs_dquot_set_prealloc_limits(dq); in xfs_qm_adjust_dqlimits() 1254 struct xfs_dquot *dq; in xfs_qm_dqiterate() local 1259 error = xfs_qm_dqget_next(mp, id, dqtype, &dq); in xfs_qm_dqiterate() 1265 error = iter_fn(dq, dqtype, priv); in xfs_qm_dqiterate() 1266 id = be32_to_cpu(dq->q_core.d_id); in xfs_qm_dqiterate() 1267 xfs_qm_dqput(dq); in xfs_qm_dqiterate()
|
D | xfs_icache.c | 1552 struct xfs_dquot *dq; in __xfs_inode_free_quota_eofblocks() local 1561 dq = xfs_inode_dquot(ip, XFS_DQ_USER); in __xfs_inode_free_quota_eofblocks() 1562 if (dq && xfs_dquot_lowsp(dq)) { in __xfs_inode_free_quota_eofblocks() 1570 dq = xfs_inode_dquot(ip, XFS_DQ_GROUP); in __xfs_inode_free_quota_eofblocks() 1571 if (dq && xfs_dquot_lowsp(dq)) { in __xfs_inode_free_quota_eofblocks()
|
D | xfs_dquot.h | 185 typedef int (*xfs_qm_dqiterate_fn)(struct xfs_dquot *dq, uint dqtype,
|
/linux-4.19.296/include/crypto/internal/ |
D | rsa.h | 43 const u8 *dq; member
|
/linux-4.19.296/fs/xfs/scrub/ |
D | quota.c | 86 struct xfs_dquot *dq, in xchk_quota_item() argument 93 struct xfs_disk_dquot *d = &dq->q_core; in xchk_quota_item()
|
/linux-4.19.296/drivers/s390/block/ |
D | dasd.c | 2791 struct dasd_queue *dq; in dasd_block_tasklet() local 2803 dq = cqr->dq; in dasd_block_tasklet() 2804 spin_lock_irq(&dq->lock); in dasd_block_tasklet() 2807 spin_unlock_irq(&dq->lock); in dasd_block_tasklet() 2839 spin_lock_irq(&cqr->dq->lock); in _dasd_requeue_request() 2842 spin_unlock_irq(&cqr->dq->lock); in _dasd_requeue_request() 2909 spin_lock_irqsave(&cqr->dq->lock, flags); in dasd_flush_block_queue() 2913 spin_unlock_irqrestore(&cqr->dq->lock, flags); in dasd_flush_block_queue() 2945 struct dasd_queue *dq = hctx->driver_data; in do_dasd_request() local 2952 spin_lock_irq(&dq->lock); in do_dasd_request() [all …]
|
D | dasd_int.h | 174 struct dasd_queue *dq; member
|
/linux-4.19.296/crypto/ |
D | rsa_helper.c | 135 key->dq = value; in rsa_get_dq()
|