/linux/fs/netfs/ |
H A D | read_pgpriv2.c | 19 static void netfs_pgpriv2_copy_folio(struct netfs_io_request *creq, struct folio *folio) in netfs_pgpriv2_copy_folio() argument 21 struct netfs_io_stream *cache = &creq->io_streams[1]; in netfs_pgpriv2_copy_folio() 32 i_size = i_size_read(creq->inode); in netfs_pgpriv2_copy_folio() 41 if (fpos + fsize > creq->i_size) in netfs_pgpriv2_copy_folio() 42 creq->i_size = i_size; in netfs_pgpriv2_copy_folio() 56 if (rolling_buffer_append(&creq->buffer, folio, 0) < 0) { in netfs_pgpriv2_copy_folio() 57 clear_bit(NETFS_RREQ_FOLIO_COPY_TO_CACHE, &creq->flags); in netfs_pgpriv2_copy_folio() 73 creq->buffer.iter.iov_offset = cache->submit_off; in netfs_pgpriv2_copy_folio() 75 atomic64_set(&creq->issued_to, fpos + cache->submit_off); in netfs_pgpriv2_copy_folio() 77 part = netfs_advance_write(creq, cach in netfs_pgpriv2_copy_folio() 100 struct netfs_io_request *creq; netfs_pgpriv2_begin_copy_to_cache() local 134 struct netfs_io_request *creq = rreq->copy_to_cache; netfs_pgpriv2_copy_to_cache() local 151 struct netfs_io_request *creq = rreq->copy_to_cache; netfs_pgpriv2_end_copy_to_cache() local 171 netfs_pgpriv2_unlock_copied_folios(struct netfs_io_request * creq) netfs_pgpriv2_unlock_copied_folios() argument [all...] |
/linux/drivers/crypto/marvell/cesa/ |
H A D | hash.c | 30 struct mv_cesa_ahash_req *creq = ahash_request_ctx(req); in mv_cesa_ahash_req_iter_init() local 31 unsigned int len = req->nbytes + creq->cache_ptr; in mv_cesa_ahash_req_iter_init() 33 if (!creq->last_req) in mv_cesa_ahash_req_iter_init() 38 iter->src.op_offset = creq->cache_ptr; in mv_cesa_ahash_req_iter_init() 96 struct mv_cesa_ahash_req *creq = ahash_request_ctx(req); in mv_cesa_ahash_dma_last_cleanup() local 98 mv_cesa_ahash_dma_free_padding(&creq->req.dma); in mv_cesa_ahash_dma_last_cleanup() 103 struct mv_cesa_ahash_req *creq = ahash_request_ctx(req); in mv_cesa_ahash_dma_cleanup() local 105 dma_unmap_sg(cesa_dev->dev, req->src, creq->src_nents, DMA_TO_DEVICE); in mv_cesa_ahash_dma_cleanup() 106 mv_cesa_ahash_dma_free_cache(&creq->req.dma); in mv_cesa_ahash_dma_cleanup() 107 mv_cesa_dma_cleanup(&creq in mv_cesa_ahash_dma_cleanup() 112 struct mv_cesa_ahash_req *creq = ahash_request_ctx(req); mv_cesa_ahash_cleanup() local 123 struct mv_cesa_ahash_req *creq = ahash_request_ctx(req); mv_cesa_ahash_last_cleanup() local 129 mv_cesa_ahash_pad_len(struct mv_cesa_ahash_req * creq) mv_cesa_ahash_pad_len() argument 139 mv_cesa_ahash_pad_req(struct mv_cesa_ahash_req * creq,u8 * buf) mv_cesa_ahash_pad_req() argument 163 struct mv_cesa_ahash_req *creq = ahash_request_ctx(req); mv_cesa_ahash_std_step() local 289 struct mv_cesa_ahash_req *creq = ahash_request_ctx(req); mv_cesa_ahash_std_process() local 300 struct mv_cesa_ahash_req *creq = ahash_request_ctx(req); mv_cesa_ahash_dma_prepare() local 308 struct mv_cesa_ahash_req *creq = ahash_request_ctx(req); mv_cesa_ahash_std_prepare() local 316 struct mv_cesa_ahash_req *creq = ahash_request_ctx(req); mv_cesa_ahash_dma_step() local 336 struct mv_cesa_ahash_req *creq = ahash_request_ctx(ahashreq); mv_cesa_ahash_step() local 347 struct mv_cesa_ahash_req *creq = ahash_request_ctx(ahashreq); mv_cesa_ahash_process() local 358 struct mv_cesa_ahash_req *creq = ahash_request_ctx(ahashreq); mv_cesa_ahash_complete() local 404 struct mv_cesa_ahash_req *creq = ahash_request_ctx(ahashreq); mv_cesa_ahash_prepare() local 417 struct mv_cesa_ahash_req *creq = ahash_request_ctx(ahashreq); mv_cesa_ahash_req_cleanup() local 441 struct mv_cesa_ahash_req *creq = ahash_request_ctx(req); mv_cesa_ahash_init() local 469 struct mv_cesa_ahash_req *creq = ahash_request_ctx(req); mv_cesa_ahash_cache_req() local 519 mv_cesa_ahash_dma_add_cache(struct mv_cesa_tdma_chain * chain,struct mv_cesa_ahash_req * creq,gfp_t flags) mv_cesa_ahash_dma_add_cache() argument 545 mv_cesa_ahash_dma_last_req(struct mv_cesa_tdma_chain * chain,struct mv_cesa_ahash_dma_iter * dma_iter,struct mv_cesa_ahash_req * creq,unsigned int frag_len,gfp_t flags) mv_cesa_ahash_dma_last_req() argument 627 struct mv_cesa_ahash_req *creq = ahash_request_ctx(req); mv_cesa_ahash_dma_req_init() local 761 struct mv_cesa_ahash_req *creq = ahash_request_ctx(req); mv_cesa_ahash_req_init() local 782 struct mv_cesa_ahash_req *creq = ahash_request_ctx(req); mv_cesa_ahash_queue_req() local 807 struct mv_cesa_ahash_req *creq = ahash_request_ctx(req); mv_cesa_ahash_update() local 816 struct mv_cesa_ahash_req *creq = ahash_request_ctx(req); mv_cesa_ahash_final() local 828 struct mv_cesa_ahash_req *creq = ahash_request_ctx(req); mv_cesa_ahash_finup() local 842 struct mv_cesa_ahash_req *creq = ahash_request_ctx(req); mv_cesa_ahash_export() local 860 struct mv_cesa_ahash_req *creq = ahash_request_ctx(req); mv_cesa_ahash_import() local 892 struct mv_cesa_ahash_req *creq = ahash_request_ctx(req); mv_cesa_md5_init() local 962 struct mv_cesa_ahash_req *creq = ahash_request_ctx(req); mv_cesa_sha1_init() local 1033 struct mv_cesa_ahash_req *creq = ahash_request_ctx(req); mv_cesa_sha256_init() local [all...] |
H A D | cipher.c | 61 struct mv_cesa_skcipher_req *creq = skcipher_request_ctx(req); in mv_cesa_skcipher_dma_cleanup() local 64 dma_unmap_sg(cesa_dev->dev, req->dst, creq->dst_nents, in mv_cesa_skcipher_dma_cleanup() 66 dma_unmap_sg(cesa_dev->dev, req->src, creq->src_nents, in mv_cesa_skcipher_dma_cleanup() 69 dma_unmap_sg(cesa_dev->dev, req->src, creq->src_nents, in mv_cesa_skcipher_dma_cleanup() 72 mv_cesa_dma_cleanup(&creq->base); in mv_cesa_skcipher_dma_cleanup() 77 struct mv_cesa_skcipher_req *creq = skcipher_request_ctx(req); in mv_cesa_skcipher_cleanup() local 78 struct mv_cesa_engine *engine = creq->base.engine; in mv_cesa_skcipher_cleanup() 80 if (mv_cesa_req_get_type(&creq->base) == CESA_DMA_REQ) in mv_cesa_skcipher_cleanup() 88 struct mv_cesa_skcipher_req *creq = skcipher_request_ctx(req); in mv_cesa_skcipher_std_step() local 89 struct mv_cesa_skcipher_std_req *sreq = &creq in mv_cesa_skcipher_std_step() 129 struct mv_cesa_skcipher_req *creq = skcipher_request_ctx(req); mv_cesa_skcipher_std_process() local 149 struct mv_cesa_skcipher_req *creq = skcipher_request_ctx(skreq); mv_cesa_skcipher_process() local 161 struct mv_cesa_skcipher_req *creq = skcipher_request_ctx(skreq); mv_cesa_skcipher_step() local 172 struct mv_cesa_skcipher_req *creq = skcipher_request_ctx(req); mv_cesa_skcipher_dma_prepare() local 181 struct mv_cesa_skcipher_req *creq = skcipher_request_ctx(req); mv_cesa_skcipher_std_prepare() local 192 struct mv_cesa_skcipher_req *creq = skcipher_request_ctx(skreq); mv_cesa_skcipher_prepare() local 214 struct mv_cesa_skcipher_req *creq = skcipher_request_ctx(skreq); mv_cesa_skcipher_complete() local 317 struct mv_cesa_skcipher_req *creq = skcipher_request_ctx(req); mv_cesa_skcipher_dma_req_init() local 412 struct mv_cesa_skcipher_req *creq = skcipher_request_ctx(req); mv_cesa_skcipher_std_req_init() local 427 struct mv_cesa_skcipher_req *creq = skcipher_request_ctx(req); mv_cesa_skcipher_req_init() local 461 struct mv_cesa_skcipher_req *creq = skcipher_request_ctx(req); mv_cesa_skcipher_queue_req() local [all...] |
H A D | cesa.c | 170 struct mv_cesa_req *creq) in mv_cesa_queue_req() argument 173 struct mv_cesa_engine *engine = creq->engine; in mv_cesa_queue_req() 177 if ((mv_cesa_req_get_type(creq) == CESA_DMA_REQ) && in mv_cesa_queue_req() 179 mv_cesa_tdma_chain(engine, creq); in mv_cesa_queue_req()
|
H A D | cesa.h | 734 struct mv_cesa_req *creq);
|
/linux/drivers/infiniband/hw/bnxt_re/ |
H A D | qplib_rcfw.c | 168 bnxt_qplib_service_creq(&rcfw->creq.creq_tasklet); in __wait_for_resp() 209 bnxt_qplib_service_creq(&rcfw->creq.creq_tasklet); in __block_for_resp() 405 bnxt_qplib_service_creq(&rcfw->creq.creq_tasklet); in __poll_for_resp() 614 rc = rcfw->creq.aeq_handler(rcfw, (void *)func_event, NULL); in bnxt_qplib_process_func_event() 646 rc = rcfw->creq.aeq_handler(rcfw, qp_event, qp); in bnxt_qplib_process_qp_event() 659 * is always called with creq->lock held. Using in bnxt_qplib_process_qp_event() 733 struct bnxt_qplib_rcfw *rcfw = from_tasklet(rcfw, t, creq.creq_tasklet); in bnxt_qplib_service_creq() 734 struct bnxt_qplib_creq_ctx *creq = &rcfw->creq; in bnxt_qplib_service_creq() local 736 struct bnxt_qplib_hwq *hwq = &creq in bnxt_qplib_service_creq() 793 struct bnxt_qplib_creq_ctx *creq; bnxt_qplib_creq_irq() local 931 struct bnxt_qplib_creq_ctx *creq; bnxt_qplib_alloc_rcfw_channel() local 983 struct bnxt_qplib_creq_ctx *creq; bnxt_qplib_rcfw_stop_irq() local 1006 struct bnxt_qplib_creq_ctx *creq; bnxt_qplib_disable_rcfw_channel() local 1026 struct bnxt_qplib_creq_ctx *creq; bnxt_qplib_rcfw_start_irq() local 1133 struct bnxt_qplib_creq_ctx *creq; bnxt_qplib_start_rcfw() local 1160 struct bnxt_qplib_creq_ctx *creq; bnxt_qplib_enable_rcfw_channel() local [all...] |
H A D | main.c | 2048 bnxt_re_net_ring_free(rdev, rdev->rcfw.creq.ring_id, type); in bnxt_re_dev_uninit() 2078 struct bnxt_qplib_creq_ctx *creq; in bnxt_re_dev_init() local 2142 creq = &rdev->rcfw.creq; in bnxt_re_dev_init() 2143 rattr.dma_arr = creq->hwq.pbl[PBL_LVL_0].pg_map_arr; in bnxt_re_dev_init() 2144 rattr.pages = creq->hwq.pbl[creq->hwq.level].pg_count; in bnxt_re_dev_init() 2149 rc = bnxt_re_net_ring_alloc(rdev, &rattr, &creq->ring_id); in bnxt_re_dev_init() 2259 bnxt_re_net_ring_free(rdev, rdev->rcfw.creq.ring_id, type); in bnxt_re_dev_init()
|
H A D | qplib_rcfw.h | 225 struct bnxt_qplib_creq_ctx creq; member
|
/linux/drivers/crypto/cavium/nitrox/ |
H A D | nitrox_req.h | 213 struct se_crypto_request creq; member 623 struct se_crypto_request *creq = &nkreq->creq; in alloc_src_req_buf() local 625 nkreq->src = alloc_req_buf(nents, ivsize, creq->gfp); in alloc_src_req_buf() 648 struct se_crypto_request *creq = &nkreq->creq; in nitrox_creq_set_src_sg() local 650 creq->src = nitrox_creq_src_sg(iv, ivsize); in nitrox_creq_set_src_sg() 651 sg = creq->src; in nitrox_creq_set_src_sg() 670 struct se_crypto_request *creq = &nkreq->creq; in alloc_dst_req_buf() local 681 struct se_crypto_request *creq = &nkreq->creq; nitrox_creq_set_orh() local 689 struct se_crypto_request *creq = &nkreq->creq; nitrox_creq_set_comp() local 704 struct se_crypto_request *creq = &nkreq->creq; nitrox_creq_set_dst_sg() local [all...] |
H A D | nitrox_aead.c | 153 struct se_crypto_request *creq = &rctx->nkreq.creq; in nitrox_set_creq() local 157 creq->flags = rctx->flags; in nitrox_set_creq() 158 creq->gfp = (rctx->flags & CRYPTO_TFM_REQ_MAY_SLEEP) ? GFP_KERNEL : in nitrox_set_creq() 161 creq->ctrl.value = 0; in nitrox_set_creq() 162 creq->opcode = FLEXI_CRYPTO_ENCRYPT_HMAC; in nitrox_set_creq() 163 creq->ctrl.s.arg = rctx->ctrl_arg; in nitrox_set_creq() 165 creq->gph.param0 = cpu_to_be16(rctx->cryptlen); in nitrox_set_creq() 166 creq->gph.param1 = cpu_to_be16(rctx->cryptlen + rctx->assoclen); in nitrox_set_creq() 167 creq in nitrox_set_creq() 218 struct se_crypto_request *creq = &rctx->nkreq.creq; nitrox_aes_gcm_enc() local 252 struct se_crypto_request *creq = &rctx->nkreq.creq; nitrox_aes_gcm_dec() local 446 struct se_crypto_request *creq = &aead_rctx->nkreq.creq; nitrox_rfc4106_enc() local 478 struct se_crypto_request *creq = &aead_rctx->nkreq.creq; nitrox_rfc4106_dec() local [all...] |
H A D | nitrox_skcipher.c | 92 if (nkreq->creq.ctrl.s.arg == ENCRYPT) { in nitrox_cbc_cipher_callback() 253 struct se_crypto_request *creq; in nitrox_skcipher_crypt() local 256 creq = &nkreq->creq; in nitrox_skcipher_crypt() 257 creq->flags = skreq->base.flags; in nitrox_skcipher_crypt() 258 creq->gfp = (skreq->base.flags & CRYPTO_TFM_REQ_MAY_SLEEP) ? in nitrox_skcipher_crypt() 262 creq->ctrl.value = 0; in nitrox_skcipher_crypt() 263 creq->opcode = FLEXI_CRYPTO_ENCRYPT_HMAC; in nitrox_skcipher_crypt() 264 creq->ctrl.s.arg = (enc ? ENCRYPT : DECRYPT); in nitrox_skcipher_crypt() 266 creq in nitrox_skcipher_crypt() [all...] |
H A D | nitrox_reqmgr.c | 214 struct se_crypto_request *creq) in softreq_map_iobuf() argument 218 ret = dma_map_inbufs(sr, creq); in softreq_map_iobuf() 222 ret = dma_map_outbufs(sr, creq); in softreq_map_iobuf()
|
/linux/crypto/ |
H A D | chacha20poly1305.c | 103 struct chacha_req *creq = &rctx->u.chacha; in chacha_decrypt() local 110 chacha_iv(creq->iv, req, 1); in chacha_decrypt() 117 skcipher_request_set_callback(&creq->req, rctx->flags, in chacha_decrypt() 119 skcipher_request_set_tfm(&creq->req, ctx->chacha); in chacha_decrypt() 120 skcipher_request_set_crypt(&creq->req, src, dst, in chacha_decrypt() 121 rctx->cryptlen, creq->iv); in chacha_decrypt() 122 err = crypto_skcipher_decrypt(&creq->req); in chacha_decrypt() 207 struct chacha_req *creq = &rctx->u.chacha; in poly_genkey() local 219 sg_init_one(creq->src, rctx->key, sizeof(rctx->key)); in poly_genkey() 221 chacha_iv(creq in poly_genkey() 245 struct chacha_req *creq = &rctx->u.chacha; chacha_encrypt() local [all...] |
H A D | pcrypt.c | 96 struct aead_request *creq = pcrypt_request_ctx(preq); in pcrypt_aead_encrypt() local 110 aead_request_set_tfm(creq, ctx->child); in pcrypt_aead_encrypt() 111 aead_request_set_callback(creq, flags & ~CRYPTO_TFM_REQ_MAY_SLEEP, in pcrypt_aead_encrypt() 113 aead_request_set_crypt(creq, req->src, req->dst, in pcrypt_aead_encrypt() 115 aead_request_set_ad(creq, req->assoclen); in pcrypt_aead_encrypt() 122 return crypto_aead_encrypt(creq); in pcrypt_aead_encrypt() 147 struct aead_request *creq = pcrypt_request_ctx(preq); in pcrypt_aead_decrypt() local 161 aead_request_set_tfm(creq, ctx->child); in pcrypt_aead_decrypt() 162 aead_request_set_callback(creq, flags & ~CRYPTO_TFM_REQ_MAY_SLEEP, in pcrypt_aead_decrypt() 164 aead_request_set_crypt(creq, re in pcrypt_aead_decrypt() [all...] |
/linux/tools/net/ynl/samples/ |
H A D | ethtool.c | 13 struct ethtool_channels_get_req_dump creq = {}; in main() local 23 creq._present.header = 1; /* ethtool needs an empty nest, sigh */ in main() 24 channels = ethtool_channels_get_dump(ys, &creq); in main()
|
/linux/drivers/crypto/ccree/ |
H A D | cc_request_mgr.c | 47 struct cc_crypto_req creq; member 348 struct cc_crypto_req *creq; in cc_proc_backlog() local 362 creq = &bli->creq; in cc_proc_backlog() 363 req = creq->user_arg; in cc_proc_backlog() 370 creq->user_cb(dev, req, -EINPROGRESS); in cc_proc_backlog() 387 cc_do_send_request(drvdata, &bli->creq, bli->desc, bli->len, in cc_proc_backlog() 435 memcpy(&bli->creq, cc_req, sizeof(*cc_req)); in cc_send_request()
|
/linux/drivers/gpu/drm/vmwgfx/ |
H A D | vmwgfx_surface.c | 1451 rep->creq = rep_ext.creq.base; in vmw_gb_surface_reference_ioctl() 1725 rep->creq.base.svga3d_flags = SVGA3D_FLAGS_LOWER_32(metadata->flags); in vmw_gb_surface_reference_internal() 1726 rep->creq.base.format = metadata->format; in vmw_gb_surface_reference_internal() 1727 rep->creq.base.mip_levels = metadata->mip_levels[0]; in vmw_gb_surface_reference_internal() 1728 rep->creq.base.drm_surface_flags = 0; in vmw_gb_surface_reference_internal() 1729 rep->creq.base.multisample_count = metadata->multisample_count; in vmw_gb_surface_reference_internal() 1730 rep->creq.base.autogen_filter = metadata->autogen_filter; in vmw_gb_surface_reference_internal() 1731 rep->creq.base.array_size = metadata->array_size; in vmw_gb_surface_reference_internal() 1732 rep->creq in vmw_gb_surface_reference_internal() [all...] |
/linux/net/wireless/ |
H A D | scan.c | 3511 struct cfg80211_scan_request_int *creq; in cfg80211_wext_siwscan() local 3531 /* Determine number of channels, needed to allocate creq */ in cfg80211_wext_siwscan() 3541 creq = kzalloc(struct_size(creq, req.channels, n_channels) + in cfg80211_wext_siwscan() 3544 if (!creq) in cfg80211_wext_siwscan() 3547 creq->req.wiphy = wiphy; in cfg80211_wext_siwscan() 3548 creq->req.wdev = dev->ieee80211_ptr; in cfg80211_wext_siwscan() 3550 creq->req.ssids = (void *)creq + in cfg80211_wext_siwscan() 3551 struct_size(creq, re in cfg80211_wext_siwscan() [all...] |
/linux/include/trace/events/ |
H A D | netfs.h | 565 const struct netfs_io_request *creq), 567 TP_ARGS(rreq, creq), 571 __field(unsigned int, creq) 580 __entry->creq = creq->debug_id; 587 __entry->creq,
|
/linux/include/uapi/drm/ |
H A D | vmwgfx_drm.h | 1003 * @creq: The data used as input when the surface was created, as described 1011 struct drm_vmw_gb_surface_create_req creq; member 1210 * @creq: The data used as input when the surface was created, as described 1218 struct drm_vmw_gb_surface_create_ext_req creq; member
|
/linux/drivers/usb/gadget/function/ |
H A D | f_fs.c | 3780 const struct usb_ctrlrequest *creq) in ffs_func_setup() argument 3787 pr_vdebug("creq->bRequestType = %02x\n", creq->bRequestType); in ffs_func_setup() 3788 pr_vdebug("creq->bRequest = %02x\n", creq->bRequest); in ffs_func_setup() 3789 pr_vdebug("creq->wValue = %04x\n", le16_to_cpu(creq->wValue)); in ffs_func_setup() 3790 pr_vdebug("creq->wIndex = %04x\n", le16_to_cpu(creq->wIndex)); in ffs_func_setup() 3791 pr_vdebug("creq in ffs_func_setup() 3838 ffs_func_req_match(struct usb_function * f,const struct usb_ctrlrequest * creq,bool config0) ffs_func_req_match() argument [all...] |
/linux/drivers/net/ethernet/marvell/octeontx2/nic/ |
H A D | otx2_tc.c | 369 struct nix_mcast_grp_create_req *creq; in otx2_tc_update_mcast() local 375 creq = otx2_mbox_alloc_msg_nix_mcast_grp_create(&nic->mbox); in otx2_tc_update_mcast() 376 if (!creq) { in otx2_tc_update_mcast() 381 creq->dir = NIX_MCAST_INGRESS; in otx2_tc_update_mcast() 391 &creq->hdr); in otx2_tc_update_mcast()
|
/linux/fs/smb/client/ |
H A D | smb2ops.c | 4330 void *creq; in crypt_message() local 4357 creq = smb2_get_aead_req(tfm, rqst, num_rqst, sign, &iv, &req, &sg, in crypt_message() 4359 if (IS_ERR(creq)) in crypt_message() 4360 return PTR_ERR(creq); in crypt_message() 4388 kvfree_sensitive(creq, sensitive_size); in crypt_message()
|