/linux/drivers/crypto/intel/qat/qat_common/ |
H A D | qat_bl.c | 111 if (unlikely(dma_mapping_error(dev, bufl->buffers[y].addr))) in __qat_bl_sgl_to_bufl() 121 if (unlikely(dma_mapping_error(dev, blp))) in __qat_bl_sgl_to_bufl() 166 if (unlikely(dma_mapping_error(dev, buffers[y].addr))) in __qat_bl_sgl_to_bufl() 184 if (unlikely(dma_mapping_error(dev, bloutp))) in __qat_bl_sgl_to_bufl() 197 if (!dma_mapping_error(dev, bloutp)) in __qat_bl_sgl_to_bufl() 204 if (!dma_mapping_error(dev, buflout->buffers[i].addr)) in __qat_bl_sgl_to_bufl() 214 if (!dma_mapping_error(dev, blp)) in __qat_bl_sgl_to_bufl() 219 if (!dma_mapping_error(dev, bufl->buffers[i].addr)) in __qat_bl_sgl_to_bufl()
|
H A D | qat_asym_algs.c | 308 if (unlikely(dma_mapping_error(dev, qat_req->in.dh.in.b))) in qat_dh_compute_value() 330 if (unlikely(dma_mapping_error(dev, qat_req->out.dh.r))) in qat_dh_compute_value() 339 if (unlikely(dma_mapping_error(dev, qat_req->phy_in))) in qat_dh_compute_value() 345 if (unlikely(dma_mapping_error(dev, qat_req->phy_out))) in qat_dh_compute_value() 361 if (!dma_mapping_error(dev, qat_req->phy_out)) in qat_dh_compute_value() 366 if (!dma_mapping_error(dev, qat_req->phy_in)) in qat_dh_compute_value() 371 if (!dma_mapping_error(dev, qat_req->out.dh.r)) in qat_dh_compute_value() 377 if (!dma_mapping_error(dev, qat_req->in.dh.in.b)) in qat_dh_compute_value() 765 if (unlikely(dma_mapping_error(dev, qat_req->in.rsa.enc.m))) in qat_rsa_enc() 780 if (unlikely(dma_mapping_error(dev, qat_req->out.rsa.enc.c))) in qat_rsa_enc() [all …]
|
/linux/drivers/crypto/caam/ |
H A D | caampkc.c | 352 if (dma_mapping_error(dev, edesc->sec4_sg_dma)) { in rsa_edesc_alloc() 416 if (dma_mapping_error(dev, pdb->n_dma)) { in set_rsa_pub_pdb() 422 if (dma_mapping_error(dev, pdb->e_dma)) { in set_rsa_pub_pdb() 461 if (dma_mapping_error(dev, pdb->n_dma)) { in set_rsa_priv_f1_pdb() 467 if (dma_mapping_error(dev, pdb->d_dma)) { in set_rsa_priv_f1_pdb() 510 if (dma_mapping_error(dev, pdb->d_dma)) { in set_rsa_priv_f2_pdb() 516 if (dma_mapping_error(dev, pdb->p_dma)) { in set_rsa_priv_f2_pdb() 522 if (dma_mapping_error(dev, pdb->q_dma)) { in set_rsa_priv_f2_pdb() 528 if (dma_mapping_error(dev, pdb->tmp1_dma)) { in set_rsa_priv_f2_pdb() 534 if (dma_mapping_error(dev, pdb->tmp2_dma)) { in set_rsa_priv_f2_pdb() [all …]
|
H A D | blob_gen.c | 93 if (dma_mapping_error(jrdev, dma_in)) { in caam_process_blob() 101 if (dma_mapping_error(jrdev, dma_out)) { in caam_process_blob()
|
H A D | caamhash.c | 153 if (dma_mapping_error(jrdev, state->ctx_dma)) { in map_seq_out_ptr_ctx() 176 if (dma_mapping_error(jrdev, state->buf_dma)) { in buf_map_to_sec4_sg() 194 if (dma_mapping_error(jrdev, state->ctx_dma)) { in ctx_map_to_sec4_sg() 380 if (dma_mapping_error(jrdev, key_dma)) { in hash_digest_key() 741 if (dma_mapping_error(ctx->jrdev, src_dma)) { in ahash_edesc_add_src() 911 if (dma_mapping_error(jrdev, edesc->sec4_sg_dma)) { in ahash_update_ctx() 984 if (dma_mapping_error(jrdev, edesc->sec4_sg_dma)) { in ahash_final_ctx() 1171 if (dma_mapping_error(jrdev, state->buf_dma)) { in ahash_final_no_ctx() 1276 if (dma_mapping_error(jrdev, edesc->sec4_sg_dma)) { in ahash_update_no_ctx() 1825 if (dma_mapping_error(ctx->jrdev, ctx->adata.key_dma)) { in caam_hash_cra_init() [all …]
|
H A D | key_gen.c | 76 if (dma_mapping_error(jrdev, dma_addr)) { in gen_split_key()
|
/linux/drivers/net/ethernet/stmicro/stmmac/ |
H A D | chain_mode.c | 39 if (dma_mapping_error(priv->device, des2)) in jumbo_frm() 57 if (dma_mapping_error(priv->device, des2)) in jumbo_frm() 70 if (dma_mapping_error(priv->device, des2)) in jumbo_frm()
|
H A D | ring_mode.c | 43 if (dma_mapping_error(priv->device, des2)) in jumbo_frm() 64 if (dma_mapping_error(priv->device, des2)) in jumbo_frm() 78 if (dma_mapping_error(priv->device, des2)) in jumbo_frm()
|
/linux/drivers/crypto/inside-secure/eip93/ |
H A D | eip93-hash.c | 222 ret = dma_mapping_error(eip93->dev, src_addr); in eip93_send_hash_req() 252 ret = dma_mapping_error(eip93->dev, rctx->sa_record_hmac_base); in eip93_send_hash_req() 405 ret = dma_mapping_error(eip93->dev, rctx->sa_state_base); in eip93_hash_update() 412 ret = dma_mapping_error(eip93->dev, rctx->sa_record_base); in eip93_hash_update() 481 ret = dma_mapping_error(eip93->dev, rctx->sa_state_base); in __eip93_hash_final() 488 ret = dma_mapping_error(eip93->dev, rctx->sa_record_base); in __eip93_hash_final() 533 ret = dma_mapping_error(eip93->dev, rctx->sa_state_base); in eip93_hash_finup() 540 ret = dma_mapping_error(eip93->dev, rctx->sa_record_base); in eip93_hash_finup()
|
/linux/drivers/spi/ |
H A D | spi-fsl-cpm.c | 139 if (dma_mapping_error(dev, mspi->tx_dma)) { in fsl_spi_cpm_bufs() 150 if (dma_mapping_error(dev, mspi->rx_dma)) { in fsl_spi_cpm_bufs() 355 if (dma_mapping_error(dev, mspi->dma_dummy_tx)) { in fsl_spi_cpm_init() 362 if (dma_mapping_error(dev, mspi->dma_dummy_rx)) { in fsl_spi_cpm_init()
|
H A D | spi-amlogic-spisg.c | 305 ret = dma_mapping_error(dev, paddr); in aml_spisg_setup_transfer() 323 ret = dma_mapping_error(dev, paddr); in aml_spisg_setup_transfer() 344 ret = dma_mapping_error(dev, paddr); in aml_spisg_setup_transfer() 362 ret = dma_mapping_error(dev, paddr); in aml_spisg_setup_transfer() 559 ret = dma_mapping_error(dev, descs_paddr); in aml_spisg_transfer_one_message()
|
/linux/drivers/crypto/allwinner/sun8i-ss/ |
H A D | sun8i-ss-prng.c | 109 if (dma_mapping_error(ss->dev, dma_iv)) { in sun8i_ss_prng_generate() 116 if (dma_mapping_error(ss->dev, dma_dst)) { in sun8i_ss_prng_generate()
|
/linux/drivers/crypto/allwinner/sun8i-ce/ |
H A D | sun8i-ce-prng.c | 98 if (dma_mapping_error(ce->dev, dma_iv)) { in sun8i_ce_prng_generate() 105 if (dma_mapping_error(ce->dev, dma_dst)) { in sun8i_ce_prng_generate()
|
H A D | sun8i-ce-trng.c | 51 if (dma_mapping_error(ce->dev, dma_dst)) { in sun8i_ce_trng_read()
|
/linux/drivers/crypto/marvell/octeontx2/ |
H A D | otx2_cpt_reqmgr.h | 234 if (unlikely(dma_mapping_error(&pdev->dev, list[i].dma_addr))) { in setup_sgio_components() 303 if (unlikely(dma_mapping_error(&pdev->dev, list[i].dma_addr))) { in sgv2io_components_setup() 425 if (unlikely(dma_mapping_error(&pdev->dev, info->dptr_baddr))) { in cn10k_sgv2_info_create() 534 if (unlikely(dma_mapping_error(&pdev->dev, info->dptr_baddr))) { in otx2_sg_info_create()
|
/linux/drivers/net/ethernet/amd/xgbe/ |
H A D | xgbe-desc.c | 210 if (dma_mapping_error(pdata->dev, pages_dma)) { in xgbe_alloc_pages() 456 if (dma_mapping_error(pdata->dev, skb_dma)) { in xgbe_map_tx_skb() 480 if (dma_mapping_error(pdata->dev, skb_dma)) { in xgbe_map_tx_skb() 512 if (dma_mapping_error(pdata->dev, skb_dma)) { in xgbe_map_tx_skb()
|
/linux/drivers/net/ethernet/synopsys/ |
H A D | dwc-xlgmac-desc.c | 352 if (dma_mapping_error(pdata->dev, pages_dma)) { in xlgmac_alloc_pages() 536 if (dma_mapping_error(pdata->dev, skb_dma)) { in xlgmac_map_tx_skb() 560 if (dma_mapping_error(pdata->dev, skb_dma)) { in xlgmac_map_tx_skb() 592 if (dma_mapping_error(pdata->dev, skb_dma)) { in xlgmac_map_tx_skb()
|
/linux/drivers/scsi/ |
H A D | a3000.c | 66 if (dma_mapping_error(hdata->dev, addr)) { in dma_setup() 107 if (dma_mapping_error(hdata->dev, addr)) { in dma_setup()
|
H A D | a2091.c | 61 if (dma_mapping_error(hdata->dev, addr)) { in dma_setup() 96 if (dma_mapping_error(hdata->dev, addr)) { in dma_setup()
|
/linux/drivers/soc/qcom/ |
H A D | qcom-geni-se.c | 709 if (dma_mapping_error(wrapper->dev, *iova)) in geni_se_tx_dma_prep() 761 if (dma_mapping_error(wrapper->dev, *iova)) in geni_se_rx_dma_prep() 781 if (!dma_mapping_error(wrapper->dev, iova)) in geni_se_tx_dma_unprep() 798 if (!dma_mapping_error(wrapper->dev, iova)) in geni_se_rx_dma_unprep()
|
/linux/drivers/crypto/cavium/cpt/ |
H A D | cptvf_reqmanager.c | 63 if (unlikely(dma_mapping_error(&pdev->dev, in setup_sgio_components() 191 if (dma_mapping_error(&pdev->dev, info->dptr_baddr)) { in setup_sgio_list() 210 if (dma_mapping_error(&pdev->dev, info->rptr_baddr)) { in setup_sgio_list() 455 if (dma_mapping_error(&pdev->dev, info->comp_baddr)) { in process_request()
|
/linux/drivers/crypto/ccree/ |
H A D | cc_buffer_mgr.c | 393 if (dma_mapping_error(dev, req_ctx->gen_ctx.iv_dma_addr)) { in cc_map_cipher_request() 568 if (dma_mapping_error(dev, areq_ctx->gen_ctx.iv_dma_addr)) { in cc_aead_chain_iv() 979 if (dma_mapping_error(dev, dma_addr)) { in cc_map_aead_request() 993 if (dma_mapping_error(dev, dma_addr)) { in cc_map_aead_request() 1011 if (dma_mapping_error(dev, dma_addr)) { in cc_map_aead_request() 1021 if (dma_mapping_error(dev, dma_addr)) { in cc_map_aead_request() 1032 if (dma_mapping_error(dev, dma_addr)) { in cc_map_aead_request() 1044 if (dma_mapping_error(dev, dma_addr)) { in cc_map_aead_request()
|
/linux/drivers/crypto/ |
H A D | mxs-dcp.c | 187 dma_err = dma_mapping_error(sdcp->dev, desc_phys); in mxs_dcp_start_dma() 241 ret = dma_mapping_error(sdcp->dev, key_phys); in mxs_dcp_run_aes() 247 ret = dma_mapping_error(sdcp->dev, src_phys); in mxs_dcp_run_aes() 253 ret = dma_mapping_error(sdcp->dev, dst_phys); in mxs_dcp_run_aes() 613 ret = dma_mapping_error(sdcp->dev, buf_phys); in mxs_dcp_run_sha() 649 ret = dma_mapping_error(sdcp->dev, digest_phys); in mxs_dcp_run_sha()
|
/linux/drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/ |
H A D | tx-gen2.c | 37 if (unlikely(dma_mapping_error(trans->dev, phys))) { in get_workaround_page() 70 if (unlikely(dma_mapping_error(trans->dev, phys))) in iwl_txq_gen2_set_tb_with_wa() 121 if (unlikely(dma_mapping_error(trans->dev, phys))) in iwl_txq_gen2_set_tb_with_wa() 333 if (unlikely(dma_mapping_error(trans->dev, tb_phys))) in iwl_txq_gen2_build_tx_amsdu() 429 if (unlikely(dma_mapping_error(trans->dev, tb_phys))) in iwl_txq_gen2_build_tx() 1380 if (dma_mapping_error(trans->dev, phys_addr)) { in iwl_pcie_gen2_enqueue_hcmd() 1402 if (dma_mapping_error(trans->dev, phys_addr)) { in iwl_pcie_gen2_enqueue_hcmd()
|
/linux/drivers/gpu/drm/amd/amdgpu/ |
H A D | amdgpu_gart.c | 84 if (dma_mapping_error(&adev->pdev->dev, adev->dummy_page_addr)) { in amdgpu_gart_dummy_page_init() 147 if (dma_mapping_error(&adev->pdev->dev, dma_addr)) { in amdgpu_gart_table_ram_alloc()
|