Home
last modified time | relevance | path

Searched refs:dq (Results 1 – 25 of 62) sorted by relevance

123

/linux/lib/raid6/
H A Drecov_loongarch_simd.c32 u8 *p, *q, *dp, *dq; in raid6_2data_recov_lsx() local
47 dq = (u8 *)ptrs[failb]; in raid6_2data_recov_lsx()
49 ptrs[disks - 1] = dq; in raid6_2data_recov_lsx()
55 ptrs[failb] = dq; in raid6_2data_recov_lsx()
81 asm volatile("vld $vr8, %0" : : "m" (dq[0])); in raid6_2data_recov_lsx()
82 asm volatile("vld $vr9, %0" : : "m" (dq[16])); in raid6_2data_recov_lsx()
83 asm volatile("vld $vr10, %0" : : "m" (dq[32])); in raid6_2data_recov_lsx()
84 asm volatile("vld $vr11, %0" : : "m" (dq[48])); in raid6_2data_recov_lsx()
161 asm volatile("vst $vr4, %0" : "=m" (dq[0])); in raid6_2data_recov_lsx()
162 asm volatile("vst $vr5, %0" : "=m" (dq[16])); in raid6_2data_recov_lsx()
[all …]
H A Drecov_avx2.c19 u8 *p, *q, *dp, *dq; in raid6_2data_recov_avx2() local
33 dq = (u8 *)ptrs[failb]; in raid6_2data_recov_avx2()
35 ptrs[disks-1] = dq; in raid6_2data_recov_avx2()
41 ptrs[failb] = dq; in raid6_2data_recov_avx2()
61 asm volatile("vpxor %0, %%ymm1, %%ymm1" : : "m" (dq[0])); in raid6_2data_recov_avx2()
62 asm volatile("vpxor %0, %%ymm9, %%ymm9" : : "m" (dq[32])); in raid6_2data_recov_avx2()
120 asm volatile("vmovdqa %%ymm1, %0" : "=m" (dq[0])); in raid6_2data_recov_avx2()
121 asm volatile("vmovdqa %%ymm13,%0" : "=m" (dq[32])); in raid6_2data_recov_avx2()
132 dq += 64; in raid6_2data_recov_avx2()
136 asm volatile("vpxor %0, %%ymm1, %%ymm1" : : "m" (*dq)); in raid6_2data_recov_avx2()
[all …]
H A Drecov_ssse3.c19 u8 *p, *q, *dp, *dq; in raid6_2data_recov_ssse3() local
35 dq = (u8 *)ptrs[failb]; in raid6_2data_recov_ssse3()
37 ptrs[disks-1] = dq; in raid6_2data_recov_ssse3()
43 ptrs[failb] = dq; in raid6_2data_recov_ssse3()
71 asm volatile("pxor %0,%%xmm1" : : "m" (dq[0])); in raid6_2data_recov_ssse3()
72 asm volatile("pxor %0,%%xmm9" : : "m" (dq[16])); in raid6_2data_recov_ssse3()
124 asm volatile("movdqa %%xmm1,%0" : "=m" (dq[0])); in raid6_2data_recov_ssse3()
125 asm volatile("movdqa %%xmm9,%0" : "=m" (dq[16])); in raid6_2data_recov_ssse3()
136 dq += 32; in raid6_2data_recov_ssse3()
140 asm volatile("pxor %0,%%xmm1" : : "m" (*dq)); in raid6_2data_recov_ssse3()
[all …]
H A Drecov_avx512.c25 u8 *p, *q, *dp, *dq; in raid6_2data_recov_avx512() local
42 dq = (u8 *)ptrs[failb]; in raid6_2data_recov_avx512()
44 ptrs[disks-1] = dq; in raid6_2data_recov_avx512()
50 ptrs[failb] = dq; in raid6_2data_recov_avx512()
76 "m" (p[64]), "m" (dq[0]), "m" (dq[64]), in raid6_2data_recov_avx512()
146 : "m" (dq[0]), "m" (dq[64])); in raid6_2data_recov_avx512()
157 dq += 128; in raid6_2data_recov_avx512()
164 : "m" (*q), "m" (*p), "m"(*dq), "m" (*dp)); in raid6_2data_recov_avx512()
207 : "m" (dq[0])); in raid6_2data_recov_avx512()
218 dq += 64; in raid6_2data_recov_avx512()
[all …]
H A Drecov_s390xc.c25 u8 *p, *q, *dp, *dq; in raid6_2data_recov_s390xc() local
39 dq = (u8 *)ptrs[failb]; in raid6_2data_recov_s390xc()
41 ptrs[disks-1] = dq; in raid6_2data_recov_s390xc()
47 ptrs[failb] = dq; in raid6_2data_recov_s390xc()
58 xor_block(dq, q); in raid6_2data_recov_s390xc()
60 dq[i] = pbmul[dp[i]] ^ qmul[dq[i]]; in raid6_2data_recov_s390xc()
61 xor_block(dp, dq); in raid6_2data_recov_s390xc()
65 dq += 256; in raid6_2data_recov_s390xc()
74 u8 *p, *q, *dq; in raid6_datap_recov_s390xc() local
83 dq = (u8 *)ptrs[faila]; in raid6_datap_recov_s390xc()
[all …]
H A Drecov_rvv.c11 u8 *dq, const u8 *pbmul, in __raid6_2data_recov_rvv() argument
74 [dq]"r"(dq), in __raid6_2data_recov_rvv()
85 dq += 16; in __raid6_2data_recov_rvv()
90 u8 *dq, const u8 *qmul) in __raid6_datap_recov_rvv() argument
133 [dq]"r"(dq), in __raid6_datap_recov_rvv()
141 dq += 16; in __raid6_datap_recov_rvv()
148 u8 *p, *q, *dp, *dq; in raid6_2data_recov_rvv() local
163 dq = (u8 *)ptrs[failb]; in raid6_2data_recov_rvv()
165 ptrs[disks - 1] = dq; in raid6_2data_recov_rvv()
171 ptrs[failb] = dq; in raid6_2data_recov_rvv()
[all …]
H A Drecov.c22 u8 *p, *q, *dp, *dq; in raid6_2data_recov_intx1() local
36 dq = (u8 *)ptrs[failb]; in raid6_2data_recov_intx1()
38 ptrs[disks-1] = dq; in raid6_2data_recov_intx1()
44 ptrs[failb] = dq; in raid6_2data_recov_intx1()
55 qx = qmul[*q ^ *dq]; in raid6_2data_recov_intx1()
56 *dq++ = db = pbmul[px] ^ qx; /* Reconstructed B */ in raid6_2data_recov_intx1()
66 u8 *p, *q, *dq; in raid6_datap_recov_intx1() local
74 dq = (u8 *)ptrs[faila]; in raid6_datap_recov_intx1()
76 ptrs[disks-1] = dq; in raid6_datap_recov_intx1()
81 ptrs[faila] = dq; in raid6_datap_recov_intx1()
[all …]
H A Drecov_neon.c25 u8 *p, *q, *dp, *dq; in raid6_2data_recov_neon() local
40 dq = (u8 *)ptrs[failb]; in raid6_2data_recov_neon()
42 ptrs[disks - 1] = dq; in raid6_2data_recov_neon()
48 ptrs[failb] = dq; in raid6_2data_recov_neon()
58 __raid6_2data_recov_neon(bytes, p, q, dp, dq, pbmul, qmul); in raid6_2data_recov_neon()
64 u8 *p, *q, *dq; in raid6_datap_recov_neon() local
74 dq = (u8 *)ptrs[faila]; in raid6_datap_recov_neon()
76 ptrs[disks - 1] = dq; in raid6_datap_recov_neon()
81 ptrs[faila] = dq; in raid6_datap_recov_neon()
88 __raid6_datap_recov_neon(bytes, p, q, dq, qmul); in raid6_datap_recov_neon()
H A Drecov_neon_inner.c29 uint8_t *dq, const uint8_t *pbmul, in __raid6_2data_recov_neon() argument
54 vx = veorq_u8(vld1q_u8(q), vld1q_u8(dq)); in __raid6_2data_recov_neon()
67 vst1q_u8(dq, db); in __raid6_2data_recov_neon()
74 dq += 16; in __raid6_2data_recov_neon()
78 void __raid6_datap_recov_neon(int bytes, uint8_t *p, uint8_t *q, uint8_t *dq, in __raid6_datap_recov_neon() argument
95 vx = veorq_u8(vld1q_u8(q), vld1q_u8(dq)); in __raid6_datap_recov_neon()
103 vst1q_u8(dq, vx); in __raid6_datap_recov_neon()
109 dq += 16; in __raid6_datap_recov_neon()
H A Dneon.h16 uint8_t *dq, const uint8_t *pbmul,
19 void __raid6_datap_recov_neon(int bytes, uint8_t *p, uint8_t *q, uint8_t *dq,
/linux/include/soc/fsl/
H A Ddpaa2-global.h20 struct dq { struct
33 } dq; member
68 static inline u32 dpaa2_dq_flags(const struct dpaa2_dq *dq) in dpaa2_dq_flags() argument
70 return dq->dq.stat; in dpaa2_dq_flags()
80 static inline int dpaa2_dq_is_pull(const struct dpaa2_dq *dq) in dpaa2_dq_is_pull() argument
82 return (int)(dpaa2_dq_flags(dq) & DPAA2_DQ_STAT_VOLATILE); in dpaa2_dq_is_pull()
91 static inline bool dpaa2_dq_is_pull_complete(const struct dpaa2_dq *dq) in dpaa2_dq_is_pull_complete() argument
93 return !!(dpaa2_dq_flags(dq) & DPAA2_DQ_STAT_EXPIRED); in dpaa2_dq_is_pull_complete()
104 static inline u16 dpaa2_dq_seqnum(const struct dpaa2_dq *dq) in dpaa2_dq_seqnum() argument
106 return le16_to_cpu(dq->dq.seqnum); in dpaa2_dq_seqnum()
[all …]
/linux/fs/xfs/scrub/
H A Dquota.c84 struct xfs_dquot *dq, in xchk_quota_item_bmap() argument
97 if (dq->q_fileoffset != offset) { in xchk_quota_item_bmap()
113 if (XFS_FSB_TO_DADDR(mp, irec.br_startblock) != dq->q_blkno) in xchk_quota_item_bmap()
142 struct xfs_dquot *dq) in xchk_quota_item() argument
160 mutex_lock(&dq->q_qlock); in xchk_quota_item()
166 offset = dq->q_id / qi->qi_dqperchunk; in xchk_quota_item()
167 if (dq->q_id && dq->q_id <= sqi->last_id) in xchk_quota_item()
170 sqi->last_id = dq->q_id; in xchk_quota_item()
172 error = xchk_quota_item_bmap(sc, dq, offset); in xchk_quota_item()
175 mutex_unlock(&dq->q_qlock); in xchk_quota_item()
[all …]
H A Dquota_repair.c58 struct xfs_dquot *dq, in xrep_quota_item_fill_bmap_hole() argument
74 error = xfs_bmapi_write(sc->tp, sc->ip, dq->q_fileoffset, in xrep_quota_item_fill_bmap_hole()
80 dq->q_blkno = XFS_FSB_TO_DADDR(mp, irec->br_startblock); in xrep_quota_item_fill_bmap_hole()
82 trace_xrep_dquot_item_fill_bmap_hole(sc->mp, dq->q_type, dq->q_id); in xrep_quota_item_fill_bmap_hole()
85 error = xfs_trans_get_buf(sc->tp, mp->m_ddev_targp, dq->q_blkno, in xrep_quota_item_fill_bmap_hole()
91 xfs_qm_init_dquot_blk(sc->tp, dq->q_id, dq->q_type, bp); in xrep_quota_item_fill_bmap_hole()
108 struct xfs_dquot *dq, in xrep_quota_item_bmap() argument
114 xfs_fileoff_t offset = dq->q_id / qi->qi_dqperchunk; in xrep_quota_item_bmap()
123 dq->q_fileoffset = offset; in xrep_quota_item_bmap()
131 error = xrep_quota_item_fill_bmap_hole(sc, dq, &irec); in xrep_quota_item_bmap()
[all …]
H A Dquotacheck_repair.c47 struct xfs_dquot *dq) in xqcheck_commit_dquot() argument
59 mutex_lock(&dq->q_qlock); in xqcheck_commit_dquot()
60 xfs_trans_dqjoin(xqc->sc->tp, dq); in xqcheck_commit_dquot()
68 error = xfarray_load_sparse(counts, dq->q_id, &xcdq); in xqcheck_commit_dquot()
73 delta = (int64_t)xcdq.icount - dq->q_ino.count; in xqcheck_commit_dquot()
75 dq->q_ino.reserved += delta; in xqcheck_commit_dquot()
76 dq->q_ino.count += delta; in xqcheck_commit_dquot()
80 delta = (int64_t)xcdq.bcount - dq->q_blk.count; in xqcheck_commit_dquot()
82 dq->q_blk.reserved += delta; in xqcheck_commit_dquot()
83 dq->q_blk.count += delta; in xqcheck_commit_dquot()
[all …]
H A Dquotacheck.c555 struct xfs_dquot *dq) in xqcheck_compare_dquot() argument
566 mutex_lock(&dq->q_qlock); in xqcheck_compare_dquot()
568 error = xfarray_load_sparse(counts, dq->q_id, &xcdq); in xqcheck_compare_dquot()
572 if (xcdq.icount != dq->q_ino.count) in xqcheck_compare_dquot()
573 xchk_qcheck_set_corrupt(xqc->sc, dqtype, dq->q_id); in xqcheck_compare_dquot()
575 if (xcdq.bcount != dq->q_blk.count) in xqcheck_compare_dquot()
576 xchk_qcheck_set_corrupt(xqc->sc, dqtype, dq->q_id); in xqcheck_compare_dquot()
578 if (xcdq.rtbcount != dq->q_rtb.count) in xqcheck_compare_dquot()
579 xchk_qcheck_set_corrupt(xqc->sc, dqtype, dq->q_id); in xqcheck_compare_dquot()
582 error = xfarray_store(counts, dq->q_id, &xcdq); in xqcheck_compare_dquot()
[all …]
H A Ddqiterate.c150 struct xfs_dquot *dq; in xchk_dquot_iter_advance_incore() local
156 nr_found = radix_tree_gang_lookup(tree, (void **)&dq, cursor->id, 1); in xchk_dquot_iter_advance_incore()
158 *next_incore_id = dq->q_id; in xchk_dquot_iter_advance_incore()
176 struct xfs_dquot *dq = NULL; in xchk_dquot_iter() local
204 error = xfs_qm_dqget(mp, cursor->id, cursor->dqtype, false, &dq); in xchk_dquot_iter()
208 cursor->id = dq->q_id + 1; in xchk_dquot_iter()
209 *dqpp = dq; in xchk_dquot_iter()
/linux/drivers/soc/fsl/dpio/
H A Dqbman-portal.h219 void qbman_swp_dqrr_consume(struct qbman_swp *s, const struct dpaa2_dq *dq);
221 int qbman_result_has_new_result(struct qbman_swp *p, const struct dpaa2_dq *dq);
308 static inline int qbman_result_is_DQ(const struct dpaa2_dq *dq) in qbman_result_is_DQ() argument
310 return ((dq->dq.verb & QBMAN_RESULT_MASK) == QBMAN_RESULT_DQ); in qbman_result_is_DQ()
318 static inline int qbman_result_is_SCN(const struct dpaa2_dq *dq) in qbman_result_is_SCN() argument
320 return !qbman_result_is_DQ(dq); in qbman_result_is_SCN()
324 static inline int qbman_result_is_FQDAN(const struct dpaa2_dq *dq) in qbman_result_is_FQDAN() argument
326 return ((dq->dq.verb & QBMAN_RESULT_MASK) == QBMAN_RESULT_FQDAN); in qbman_result_is_FQDAN()
330 static inline int qbman_result_is_CDAN(const struct dpaa2_dq *dq) in qbman_result_is_CDAN() argument
332 return ((dq->dq.verb & QBMAN_RESULT_MASK) == QBMAN_RESULT_CDAN); in qbman_result_is_CDAN()
[all …]
/linux/crypto/async_tx/
H A Dasync_raid6_recov.c208 struct page *p, *q, *g, *dp, *dq; in __2data_recov_5() local
244 dq = blocks[failb]; in __2data_recov_5()
250 tx = async_mult(dq, dq_off, g, g_off, in __2data_recov_5()
263 srcs[0] = dq; in __2data_recov_5()
269 tx = async_xor_offs(dq, dq_off, srcs, src_offs, 2, bytes, submit); in __2data_recov_5()
274 srcs[1] = dq; in __2data_recov_5()
279 tx = async_sum_product(dq, dq_off, srcs, src_offs, coef, bytes, submit); in __2data_recov_5()
284 srcs[1] = dq; in __2data_recov_5()
299 struct page *p, *q, *dp, *dq; in __2data_recov_n() local
323 dq = blocks[failb]; in __2data_recov_n()
[all …]
/linux/drivers/net/ethernet/cavium/liquidio/
H A Dcn66xx_regs.h311 #define CN6XXX_DMA_CNT(dq) \ argument
312 (CN6XXX_DMA_CNT_START + ((dq) * CN6XXX_DMA_OFFSET))
314 #define CN6XXX_DMA_INT_LEVEL(dq) \ argument
315 (CN6XXX_DMA_INT_LEVEL_START + ((dq) * CN6XXX_DMA_OFFSET))
317 #define CN6XXX_DMA_PKT_INT_LEVEL(dq) \ argument
318 (CN6XXX_DMA_INT_LEVEL_START + ((dq) * CN6XXX_DMA_OFFSET))
320 #define CN6XXX_DMA_TIME_INT_LEVEL(dq) \ argument
321 (CN6XXX_DMA_INT_LEVEL_START + 4 + ((dq) * CN6XXX_DMA_OFFSET))
323 #define CN6XXX_DMA_TIM(dq) \ argument
324 (CN6XXX_DMA_TIM_START + ((dq) * CN6XXX_DMA_OFFSET))
H A Dcn23xx_pf_regs.h366 #define CN23XX_DMA_CNT(dq) \ argument
367 (CN23XX_DMA_CNT_START + ((dq) * CN23XX_DMA_OFFSET))
369 #define CN23XX_DMA_INT_LEVEL(dq) \ argument
370 (CN23XX_DMA_INT_LEVEL_START + ((dq) * CN23XX_DMA_OFFSET))
372 #define CN23XX_DMA_PKT_INT_LEVEL(dq) \ argument
373 (CN23XX_DMA_INT_LEVEL_START + ((dq) * CN23XX_DMA_OFFSET))
375 #define CN23XX_DMA_TIME_INT_LEVEL(dq) \ argument
376 (CN23XX_DMA_INT_LEVEL_START + 4 + ((dq) * CN23XX_DMA_OFFSET))
378 #define CN23XX_DMA_TIM(dq) \ argument
379 (CN23XX_DMA_TIM_START + ((dq) * CN23XX_DMA_OFFSET))
/linux/drivers/s390/crypto/
H A Dzcrypt_msgtype50.c104 unsigned char dq[64]; member
117 unsigned char dq[128]; member
130 unsigned char dq[256]; member
262 unsigned char *p, *q, *dp, *dq, *u, *inp; in ICACRT_msg_to_type50CRT_msg() local
284 dq = crb1->dq + sizeof(crb1->dq) - short_len; in ICACRT_msg_to_type50CRT_msg()
298 dq = crb2->dq + sizeof(crb2->dq) - short_len; in ICACRT_msg_to_type50CRT_msg()
313 dq = crb3->dq + sizeof(crb3->dq) - short_len; in ICACRT_msg_to_type50CRT_msg()
327 copy_from_user(dq, crt->bq_key, short_len) || in ICACRT_msg_to_type50CRT_msg()
/linux/fs/xfs/
H A Dxfs_dquot.c120 struct xfs_dquot *dq) in xfs_qm_adjust_dqlimits() argument
122 struct xfs_mount *mp = dq->q_mount; in xfs_qm_adjust_dqlimits()
127 ASSERT(dq->q_id); in xfs_qm_adjust_dqlimits()
128 defq = xfs_get_defquota(q, xfs_dquot_type(dq)); in xfs_qm_adjust_dqlimits()
130 if (!dq->q_blk.softlimit) { in xfs_qm_adjust_dqlimits()
131 dq->q_blk.softlimit = defq->blk.soft; in xfs_qm_adjust_dqlimits()
134 if (!dq->q_blk.hardlimit) { in xfs_qm_adjust_dqlimits()
135 dq->q_blk.hardlimit = defq->blk.hard; in xfs_qm_adjust_dqlimits()
138 if (!dq->q_ino.softlimit) in xfs_qm_adjust_dqlimits()
139 dq->q_ino.softlimit = defq->ino.soft; in xfs_qm_adjust_dqlimits()
[all …]
/linux/crypto/
H A Drsa.c23 MPI dq; member
96 ret = mpi_powm(m2, c, key->dq, key->q); in _rsa_dec_crt()
205 mpi_free(key->dq); in rsa_free_mpi_key()
213 key->dq = NULL; in rsa_free_mpi_key()
346 mpi_key->dq = mpi_read_raw_data(raw_key.dq, raw_key.dq_sz); in rsa_set_priv_key()
347 if (!mpi_key->dq) in rsa_set_priv_key()
/linux/drivers/scsi/hisi_sas/
H A Dhisi_sas_main.c432 struct hisi_sas_dq *dq, in hisi_sas_task_deliver() argument
440 spin_lock(&dq->lock); in hisi_sas_task_deliver()
441 wr_q_index = dq->wr_point; in hisi_sas_task_deliver()
442 dq->wr_point = (dq->wr_point + 1) % HISI_SAS_QUEUE_SLOTS; in hisi_sas_task_deliver()
443 list_add_tail(&slot->delivery, &dq->list); in hisi_sas_task_deliver()
444 spin_unlock(&dq->lock); in hisi_sas_task_deliver()
449 dlvry_queue = dq->id; in hisi_sas_task_deliver()
488 spin_lock(&dq->lock); in hisi_sas_task_deliver()
489 hisi_hba->hw->start_delivery(dq); in hisi_sas_task_deliver()
490 spin_unlock(&dq->lock); in hisi_sas_task_deliver()
[all …]
/linux/drivers/media/common/saa7146/
H A Dsaa7146_vbi.c293 struct saa7146_dmaqueue *dq = &dev->vv_data->vbi_dmaq; in return_buffers() local
296 if (dq->curr) { in return_buffers()
297 buf = dq->curr; in return_buffers()
298 dq->curr = NULL; in return_buffers()
301 while (!list_empty(&dq->queue)) { in return_buffers()
302 buf = list_entry(dq->queue.next, struct saa7146_buf, list); in return_buffers()

123