/Linux-v5.4/include/soc/fsl/ |
D | dpaa2-global.h | 20 struct dq { struct 33 } dq; member 68 static inline u32 dpaa2_dq_flags(const struct dpaa2_dq *dq) in dpaa2_dq_flags() argument 70 return dq->dq.stat; in dpaa2_dq_flags() 80 static inline int dpaa2_dq_is_pull(const struct dpaa2_dq *dq) in dpaa2_dq_is_pull() argument 82 return (int)(dpaa2_dq_flags(dq) & DPAA2_DQ_STAT_VOLATILE); in dpaa2_dq_is_pull() 91 static inline bool dpaa2_dq_is_pull_complete(const struct dpaa2_dq *dq) in dpaa2_dq_is_pull_complete() argument 93 return !!(dpaa2_dq_flags(dq) & DPAA2_DQ_STAT_EXPIRED); in dpaa2_dq_is_pull_complete() 104 static inline u16 dpaa2_dq_seqnum(const struct dpaa2_dq *dq) in dpaa2_dq_seqnum() argument 106 return le16_to_cpu(dq->dq.seqnum); in dpaa2_dq_seqnum() [all …]
|
/Linux-v5.4/lib/raid6/ |
D | recov_avx2.c | 21 u8 *p, *q, *dp, *dq; in raid6_2data_recov_avx2() local 35 dq = (u8 *)ptrs[failb]; in raid6_2data_recov_avx2() 37 ptrs[disks-1] = dq; in raid6_2data_recov_avx2() 43 ptrs[failb] = dq; in raid6_2data_recov_avx2() 63 asm volatile("vpxor %0, %%ymm1, %%ymm1" : : "m" (dq[0])); in raid6_2data_recov_avx2() 64 asm volatile("vpxor %0, %%ymm9, %%ymm9" : : "m" (dq[32])); in raid6_2data_recov_avx2() 122 asm volatile("vmovdqa %%ymm1, %0" : "=m" (dq[0])); in raid6_2data_recov_avx2() 123 asm volatile("vmovdqa %%ymm13,%0" : "=m" (dq[32])); in raid6_2data_recov_avx2() 134 dq += 64; in raid6_2data_recov_avx2() 138 asm volatile("vpxor %0, %%ymm1, %%ymm1" : : "m" (*dq)); in raid6_2data_recov_avx2() [all …]
|
D | recov_avx512.c | 27 u8 *p, *q, *dp, *dq; in raid6_2data_recov_avx512() local 44 dq = (u8 *)ptrs[failb]; in raid6_2data_recov_avx512() 46 ptrs[disks-1] = dq; in raid6_2data_recov_avx512() 52 ptrs[failb] = dq; in raid6_2data_recov_avx512() 78 "m" (p[64]), "m" (dq[0]), "m" (dq[64]), in raid6_2data_recov_avx512() 148 : "m" (dq[0]), "m" (dq[64])); in raid6_2data_recov_avx512() 159 dq += 128; in raid6_2data_recov_avx512() 166 : "m" (*q), "m" (*p), "m"(*dq), "m" (*dp)); in raid6_2data_recov_avx512() 209 : "m" (dq[0])); in raid6_2data_recov_avx512() 220 dq += 64; in raid6_2data_recov_avx512() [all …]
|
D | recov_ssse3.c | 21 u8 *p, *q, *dp, *dq; in raid6_2data_recov_ssse3() local 37 dq = (u8 *)ptrs[failb]; in raid6_2data_recov_ssse3() 39 ptrs[disks-1] = dq; in raid6_2data_recov_ssse3() 45 ptrs[failb] = dq; in raid6_2data_recov_ssse3() 73 asm volatile("pxor %0,%%xmm1" : : "m" (dq[0])); in raid6_2data_recov_ssse3() 74 asm volatile("pxor %0,%%xmm9" : : "m" (dq[16])); in raid6_2data_recov_ssse3() 126 asm volatile("movdqa %%xmm1,%0" : "=m" (dq[0])); in raid6_2data_recov_ssse3() 127 asm volatile("movdqa %%xmm9,%0" : "=m" (dq[16])); in raid6_2data_recov_ssse3() 138 dq += 32; in raid6_2data_recov_ssse3() 142 asm volatile("pxor %0,%%xmm1" : : "m" (*dq)); in raid6_2data_recov_ssse3() [all …]
|
D | recov_s390xc.c | 26 u8 *p, *q, *dp, *dq; in raid6_2data_recov_s390xc() local 40 dq = (u8 *)ptrs[failb]; in raid6_2data_recov_s390xc() 42 ptrs[disks-1] = dq; in raid6_2data_recov_s390xc() 48 ptrs[failb] = dq; in raid6_2data_recov_s390xc() 59 xor_block(dq, q); in raid6_2data_recov_s390xc() 61 dq[i] = pbmul[dp[i]] ^ qmul[dq[i]]; in raid6_2data_recov_s390xc() 62 xor_block(dp, dq); in raid6_2data_recov_s390xc() 66 dq += 256; in raid6_2data_recov_s390xc() 75 u8 *p, *q, *dq; in raid6_datap_recov_s390xc() local 84 dq = (u8 *)ptrs[faila]; in raid6_datap_recov_s390xc() [all …]
|
D | recov.c | 23 u8 *p, *q, *dp, *dq; in raid6_2data_recov_intx1() local 37 dq = (u8 *)ptrs[failb]; in raid6_2data_recov_intx1() 39 ptrs[disks-1] = dq; in raid6_2data_recov_intx1() 45 ptrs[failb] = dq; in raid6_2data_recov_intx1() 56 qx = qmul[*q ^ *dq]; in raid6_2data_recov_intx1() 57 *dq++ = db = pbmul[px] ^ qx; /* Reconstructed B */ in raid6_2data_recov_intx1() 67 u8 *p, *q, *dq; in raid6_datap_recov_intx1() local 75 dq = (u8 *)ptrs[faila]; in raid6_datap_recov_intx1() 77 ptrs[disks-1] = dq; in raid6_datap_recov_intx1() 82 ptrs[faila] = dq; in raid6_datap_recov_intx1() [all …]
|
D | recov_neon.c | 23 uint8_t *dq, const uint8_t *pbmul, 26 void __raid6_datap_recov_neon(int bytes, uint8_t *p, uint8_t *q, uint8_t *dq, 32 u8 *p, *q, *dp, *dq; in raid6_2data_recov_neon() local 47 dq = (u8 *)ptrs[failb]; in raid6_2data_recov_neon() 49 ptrs[disks - 1] = dq; in raid6_2data_recov_neon() 55 ptrs[failb] = dq; in raid6_2data_recov_neon() 65 __raid6_2data_recov_neon(bytes, p, q, dp, dq, pbmul, qmul); in raid6_2data_recov_neon() 72 u8 *p, *q, *dq; in raid6_datap_recov_neon() local 82 dq = (u8 *)ptrs[faila]; in raid6_datap_recov_neon() 84 ptrs[disks - 1] = dq; in raid6_datap_recov_neon() [all …]
|
D | recov_neon_inner.c | 28 uint8_t *dq, const uint8_t *pbmul, in __raid6_2data_recov_neon() argument 53 vx = veorq_u8(vld1q_u8(q), vld1q_u8(dq)); in __raid6_2data_recov_neon() 66 vst1q_u8(dq, db); in __raid6_2data_recov_neon() 73 dq += 16; in __raid6_2data_recov_neon() 77 void __raid6_datap_recov_neon(int bytes, uint8_t *p, uint8_t *q, uint8_t *dq, in __raid6_datap_recov_neon() argument 94 vx = veorq_u8(vld1q_u8(q), vld1q_u8(dq)); in __raid6_datap_recov_neon() 102 vst1q_u8(dq, vx); in __raid6_datap_recov_neon() 108 dq += 16; in __raid6_datap_recov_neon()
|
/Linux-v5.4/drivers/soc/fsl/dpio/ |
D | qbman-portal.h | 164 void qbman_swp_dqrr_consume(struct qbman_swp *s, const struct dpaa2_dq *dq); 166 int qbman_result_has_new_result(struct qbman_swp *p, const struct dpaa2_dq *dq); 202 static inline int qbman_result_is_DQ(const struct dpaa2_dq *dq) in qbman_result_is_DQ() argument 204 return ((dq->dq.verb & QBMAN_RESULT_MASK) == QBMAN_RESULT_DQ); in qbman_result_is_DQ() 212 static inline int qbman_result_is_SCN(const struct dpaa2_dq *dq) in qbman_result_is_SCN() argument 214 return !qbman_result_is_DQ(dq); in qbman_result_is_SCN() 218 static inline int qbman_result_is_FQDAN(const struct dpaa2_dq *dq) in qbman_result_is_FQDAN() argument 220 return ((dq->dq.verb & QBMAN_RESULT_MASK) == QBMAN_RESULT_FQDAN); in qbman_result_is_FQDAN() 224 static inline int qbman_result_is_CDAN(const struct dpaa2_dq *dq) in qbman_result_is_CDAN() argument 226 return ((dq->dq.verb & QBMAN_RESULT_MASK) == QBMAN_RESULT_CDAN); in qbman_result_is_CDAN() [all …]
|
D | dpio-service.c | 188 const struct dpaa2_dq *dq; in dpaa2_io_irq() local 198 dq = qbman_swp_dqrr_next(swp); in dpaa2_io_irq() 199 while (dq) { in dpaa2_io_irq() 200 if (qbman_result_is_SCN(dq)) { in dpaa2_io_irq() 204 q64 = qbman_result_SCN_ctx(dq); in dpaa2_io_irq() 210 qbman_swp_dqrr_consume(swp, dq); in dpaa2_io_irq() 214 dq = qbman_swp_dqrr_next(swp); in dpaa2_io_irq()
|
D | qbman-portal.c | 747 verb = p->dq.verb; in qbman_swp_dqrr_next() 775 flags = p->dq.stat; in qbman_swp_dqrr_next() 793 void qbman_swp_dqrr_consume(struct qbman_swp *s, const struct dpaa2_dq *dq) in qbman_swp_dqrr_consume() argument 795 qbman_write_register(s, QBMAN_CINH_SWP_DCAP, QBMAN_IDX_FROM_DQRR(dq)); in qbman_swp_dqrr_consume() 815 int qbman_result_has_new_result(struct qbman_swp *s, const struct dpaa2_dq *dq) in qbman_result_has_new_result() argument 817 if (dq->dq.tok != QMAN_DQ_TOKEN_VALID) in qbman_result_has_new_result() 825 ((struct dpaa2_dq *)dq)->dq.tok = 0; in qbman_result_has_new_result() 832 if (s->vdq.storage == dq) { in qbman_result_has_new_result()
|
/Linux-v5.4/crypto/async_tx/ |
D | async_raid6_recov.c | 189 struct page *p, *q, *g, *dp, *dq; in __2data_recov_5() local 219 dq = blocks[failb]; in __2data_recov_5() 224 tx = async_mult(dq, g, raid6_gfexp[good], bytes, submit); in __2data_recov_5() 234 srcs[0] = dq; in __2data_recov_5() 238 tx = async_xor(dq, srcs, 0, 2, bytes, submit); in __2data_recov_5() 242 srcs[1] = dq; in __2data_recov_5() 246 tx = async_sum_product(dq, srcs, coef, bytes, submit); in __2data_recov_5() 250 srcs[1] = dq; in __2data_recov_5() 263 struct page *p, *q, *dp, *dq; in __2data_recov_n() local 281 dq = blocks[failb]; in __2data_recov_n() [all …]
|
/Linux-v5.4/drivers/net/ethernet/cavium/liquidio/ |
D | cn66xx_regs.h | 311 #define CN6XXX_DMA_CNT(dq) \ argument 312 (CN6XXX_DMA_CNT_START + ((dq) * CN6XXX_DMA_OFFSET)) 314 #define CN6XXX_DMA_INT_LEVEL(dq) \ argument 315 (CN6XXX_DMA_INT_LEVEL_START + ((dq) * CN6XXX_DMA_OFFSET)) 317 #define CN6XXX_DMA_PKT_INT_LEVEL(dq) \ argument 318 (CN6XXX_DMA_INT_LEVEL_START + ((dq) * CN6XXX_DMA_OFFSET)) 320 #define CN6XXX_DMA_TIME_INT_LEVEL(dq) \ argument 321 (CN6XXX_DMA_INT_LEVEL_START + 4 + ((dq) * CN6XXX_DMA_OFFSET)) 323 #define CN6XXX_DMA_TIM(dq) \ argument 324 (CN6XXX_DMA_TIM_START + ((dq) * CN6XXX_DMA_OFFSET))
|
D | cn23xx_pf_regs.h | 366 #define CN23XX_DMA_CNT(dq) \ argument 367 (CN23XX_DMA_CNT_START + ((dq) * CN23XX_DMA_OFFSET)) 369 #define CN23XX_DMA_INT_LEVEL(dq) \ argument 370 (CN23XX_DMA_INT_LEVEL_START + ((dq) * CN23XX_DMA_OFFSET)) 372 #define CN23XX_DMA_PKT_INT_LEVEL(dq) \ argument 373 (CN23XX_DMA_INT_LEVEL_START + ((dq) * CN23XX_DMA_OFFSET)) 375 #define CN23XX_DMA_TIME_INT_LEVEL(dq) \ argument 376 (CN23XX_DMA_INT_LEVEL_START + 4 + ((dq) * CN23XX_DMA_OFFSET)) 378 #define CN23XX_DMA_TIM(dq) \ argument 379 (CN23XX_DMA_TIM_START + ((dq) * CN23XX_DMA_OFFSET))
|
/Linux-v5.4/drivers/s390/crypto/ |
D | zcrypt_msgtype50.c | 108 unsigned char dq[64]; member 121 unsigned char dq[128]; member 134 unsigned char dq[256]; member 266 unsigned char *p, *q, *dp, *dq, *u, *inp; in ICACRT_msg_to_type50CRT_msg() local 288 dq = crb1->dq + sizeof(crb1->dq) - short_len; in ICACRT_msg_to_type50CRT_msg() 302 dq = crb2->dq + sizeof(crb2->dq) - short_len; in ICACRT_msg_to_type50CRT_msg() 317 dq = crb3->dq + sizeof(crb3->dq) - short_len; in ICACRT_msg_to_type50CRT_msg() 330 copy_from_user(dq, crt->bq_key, short_len) || in ICACRT_msg_to_type50CRT_msg()
|
D | zcrypt_cex2a.h | 80 unsigned char dq[64]; member 93 unsigned char dq[128]; member 106 unsigned char dq[256]; member
|
/Linux-v5.4/drivers/scsi/hisi_sas/ |
D | hisi_sas_main.c | 424 struct hisi_sas_dq *dq; in hisi_sas_task_prep() local 443 *dq_pointer = dq = &hisi_hba->dq[dq_index]; in hisi_sas_task_prep() 445 *dq_pointer = dq = sas_dev->dq; in hisi_sas_task_prep() 492 spin_lock_irqsave(&dq->lock, flags); in hisi_sas_task_prep() 493 wr_q_index = dq->wr_point; in hisi_sas_task_prep() 494 dq->wr_point = (dq->wr_point + 1) % HISI_SAS_QUEUE_SLOTS; in hisi_sas_task_prep() 495 list_add_tail(&slot->delivery, &dq->list); in hisi_sas_task_prep() 496 spin_unlock_irqrestore(&dq->lock, flags); in hisi_sas_task_prep() 501 dlvry_queue = dq->id; in hisi_sas_task_prep() 570 struct hisi_sas_dq *dq = NULL; in hisi_sas_task_exec() local [all …]
|
/Linux-v5.4/fs/xfs/ |
D | xfs_iomap.c | 320 struct xfs_dquot *dq = xfs_inode_dquot(ip, type); in xfs_quota_need_throttle() local 322 if (!dq || !xfs_this_quota_on(ip->i_mount, type)) in xfs_quota_need_throttle() 326 if (!dq->q_prealloc_hi_wmark) in xfs_quota_need_throttle() 330 if (dq->q_res_bcount + alloc_blocks < dq->q_prealloc_lo_wmark) in xfs_quota_need_throttle() 346 struct xfs_dquot *dq = xfs_inode_dquot(ip, type); in xfs_quota_calc_throttle() local 349 if (!dq || dq->q_res_bcount >= dq->q_prealloc_hi_wmark) { in xfs_quota_calc_throttle() 355 freesp = dq->q_prealloc_hi_wmark - dq->q_res_bcount; in xfs_quota_calc_throttle() 356 if (freesp < dq->q_low_space[XFS_QLOWSP_5_PCNT]) { in xfs_quota_calc_throttle() 358 if (freesp < dq->q_low_space[XFS_QLOWSP_3_PCNT]) in xfs_quota_calc_throttle() 360 if (freesp < dq->q_low_space[XFS_QLOWSP_1_PCNT]) in xfs_quota_calc_throttle()
|
D | xfs_dquot.c | 70 struct xfs_dquot *dq) in xfs_qm_adjust_dqlimits() argument 73 struct xfs_disk_dquot *d = &dq->q_core; in xfs_qm_adjust_dqlimits() 78 defq = xfs_get_defquota(dq, q); in xfs_qm_adjust_dqlimits() 98 xfs_dquot_set_prealloc_limits(dq); in xfs_qm_adjust_dqlimits() 1251 struct xfs_dquot *dq; in xfs_qm_dqiterate() local 1256 error = xfs_qm_dqget_next(mp, id, dqtype, &dq); in xfs_qm_dqiterate() 1262 error = iter_fn(dq, dqtype, priv); in xfs_qm_dqiterate() 1263 id = be32_to_cpu(dq->q_core.d_id); in xfs_qm_dqiterate() 1264 xfs_qm_dqput(dq); in xfs_qm_dqiterate()
|
/Linux-v5.4/drivers/soc/fsl/qbman/ |
D | qman_test_api.c | 211 const struct qm_dqrr_entry *dq) in cb_dqrr() argument 213 if (WARN_ON(fd_neq(&fd_dq, &dq->fd))) { in cb_dqrr() 218 if (!(dq->stat & QM_DQRR_STAT_UNSCHEDULED) && !fd_neq(&fd_dq, &fd)) { in cb_dqrr()
|
/Linux-v5.4/drivers/crypto/qat/qat_common/ |
D | qat_asym_algs.c | 84 dma_addr_t dq; member 110 char *dq; member 859 qat_req->in.rsa.dec_crt.dq = ctx->dma_dq; in qat_rsa_dec() 1109 ptr = rsa_key->dq; in qat_rsa_setkey_crt() 1114 ctx->dq = dma_alloc_coherent(dev, half_key_sz, &ctx->dma_dq, in qat_rsa_setkey_crt() 1116 if (!ctx->dq) in qat_rsa_setkey_crt() 1118 memcpy(ctx->dq + (half_key_sz - len), ptr, len); in qat_rsa_setkey_crt() 1136 memset(ctx->dq, '\0', half_key_sz); in qat_rsa_setkey_crt() 1137 dma_free_coherent(dev, half_key_sz, ctx->dq, ctx->dma_dq); in qat_rsa_setkey_crt() 1138 ctx->dq = NULL; in qat_rsa_setkey_crt() [all …]
|
/Linux-v5.4/drivers/ata/ |
D | sata_nv.c | 1691 struct defer_queue *dq = &pp->defer_queue; in nv_swncq_qc_to_dq() local 1694 WARN_ON(dq->tail - dq->head == ATA_MAX_QUEUE); in nv_swncq_qc_to_dq() 1695 dq->defer_bits |= (1 << qc->hw_tag); in nv_swncq_qc_to_dq() 1696 dq->tag[dq->tail++ & (ATA_MAX_QUEUE - 1)] = qc->hw_tag; in nv_swncq_qc_to_dq() 1702 struct defer_queue *dq = &pp->defer_queue; in nv_swncq_qc_from_dq() local 1705 if (dq->head == dq->tail) /* null queue */ in nv_swncq_qc_from_dq() 1708 tag = dq->tag[dq->head & (ATA_MAX_QUEUE - 1)]; in nv_swncq_qc_from_dq() 1709 dq->tag[dq->head++ & (ATA_MAX_QUEUE - 1)] = ATA_TAG_POISON; in nv_swncq_qc_from_dq() 1710 WARN_ON(!(dq->defer_bits & (1 << tag))); in nv_swncq_qc_from_dq() 1711 dq->defer_bits &= ~(1 << tag); in nv_swncq_qc_from_dq() [all …]
|
/Linux-v5.4/include/crypto/internal/ |
D | rsa.h | 38 const u8 *dq; member
|
/Linux-v5.4/drivers/atm/ |
D | firestream.c | 622 static struct FS_BPENTRY dq[60]; variable 1191 dq[qd].flags = td->flags; in fs_send() 1192 dq[qd].next = td->next; in fs_send() 1193 dq[qd].bsa = td->bsa; in fs_send() 1194 dq[qd].skb = td->skb; in fs_send() 1195 dq[qd].dev = td->dev; in fs_send() 1945 i, da[qd], dq[qd].flags, dq[qd].bsa, dq[qd].skb, dq[qd].dev); in firestream_remove_one()
|
/Linux-v5.4/drivers/s390/block/ |
D | dasd.c | 2917 struct dasd_queue *dq; in dasd_block_tasklet() local 2929 dq = cqr->dq; in dasd_block_tasklet() 2930 spin_lock_irq(&dq->lock); in dasd_block_tasklet() 2933 spin_unlock_irq(&dq->lock); in dasd_block_tasklet() 2962 spin_lock_irq(&cqr->dq->lock); in _dasd_requeue_request() 2965 spin_unlock_irq(&cqr->dq->lock); in _dasd_requeue_request() 3019 spin_lock_irqsave(&cqr->dq->lock, flags); in dasd_flush_block_queue() 3023 spin_unlock_irqrestore(&cqr->dq->lock, flags); in dasd_flush_block_queue() 3055 struct dasd_queue *dq = hctx->driver_data; in do_dasd_request() local 3062 spin_lock_irq(&dq->lock); in do_dasd_request() [all …]
|