Home
last modified time | relevance | path

Searched refs:creq (Results 1 – 23 of 23) sorted by relevance

/linux/fs/netfs/
H A Dread_pgpriv2.c19 static void netfs_pgpriv2_copy_folio(struct netfs_io_request *creq, struct folio *folio) in netfs_pgpriv2_copy_folio() argument
21 struct netfs_io_stream *cache = &creq->io_streams[1]; in netfs_pgpriv2_copy_folio()
32 i_size = i_size_read(creq->inode); in netfs_pgpriv2_copy_folio()
41 if (fpos + fsize > creq->i_size) in netfs_pgpriv2_copy_folio()
42 creq->i_size = i_size; in netfs_pgpriv2_copy_folio()
56 if (rolling_buffer_append(&creq->buffer, folio, 0) < 0) { in netfs_pgpriv2_copy_folio()
57 clear_bit(NETFS_RREQ_FOLIO_COPY_TO_CACHE, &creq->flags); in netfs_pgpriv2_copy_folio()
73 creq->buffer.iter.iov_offset = cache->submit_off; in netfs_pgpriv2_copy_folio()
75 atomic64_set(&creq->issued_to, fpos + cache->submit_off); in netfs_pgpriv2_copy_folio()
77 part = netfs_advance_write(creq, cach in netfs_pgpriv2_copy_folio()
100 struct netfs_io_request *creq; netfs_pgpriv2_begin_copy_to_cache() local
134 struct netfs_io_request *creq = rreq->copy_to_cache; netfs_pgpriv2_copy_to_cache() local
151 struct netfs_io_request *creq = rreq->copy_to_cache; netfs_pgpriv2_end_copy_to_cache() local
171 netfs_pgpriv2_unlock_copied_folios(struct netfs_io_request * creq) netfs_pgpriv2_unlock_copied_folios() argument
[all...]
/linux/drivers/crypto/marvell/cesa/
H A Dhash.c30 struct mv_cesa_ahash_req *creq = ahash_request_ctx(req); in mv_cesa_ahash_req_iter_init() local
31 unsigned int len = req->nbytes + creq->cache_ptr; in mv_cesa_ahash_req_iter_init()
33 if (!creq->last_req) in mv_cesa_ahash_req_iter_init()
38 iter->src.op_offset = creq->cache_ptr; in mv_cesa_ahash_req_iter_init()
96 struct mv_cesa_ahash_req *creq = ahash_request_ctx(req); in mv_cesa_ahash_dma_last_cleanup() local
98 mv_cesa_ahash_dma_free_padding(&creq->req.dma); in mv_cesa_ahash_dma_last_cleanup()
103 struct mv_cesa_ahash_req *creq = ahash_request_ctx(req); in mv_cesa_ahash_dma_cleanup() local
105 dma_unmap_sg(cesa_dev->dev, req->src, creq->src_nents, DMA_TO_DEVICE); in mv_cesa_ahash_dma_cleanup()
106 mv_cesa_ahash_dma_free_cache(&creq->req.dma); in mv_cesa_ahash_dma_cleanup()
107 mv_cesa_dma_cleanup(&creq in mv_cesa_ahash_dma_cleanup()
112 struct mv_cesa_ahash_req *creq = ahash_request_ctx(req); mv_cesa_ahash_cleanup() local
123 struct mv_cesa_ahash_req *creq = ahash_request_ctx(req); mv_cesa_ahash_last_cleanup() local
129 mv_cesa_ahash_pad_len(struct mv_cesa_ahash_req * creq) mv_cesa_ahash_pad_len() argument
139 mv_cesa_ahash_pad_req(struct mv_cesa_ahash_req * creq,u8 * buf) mv_cesa_ahash_pad_req() argument
163 struct mv_cesa_ahash_req *creq = ahash_request_ctx(req); mv_cesa_ahash_std_step() local
289 struct mv_cesa_ahash_req *creq = ahash_request_ctx(req); mv_cesa_ahash_std_process() local
300 struct mv_cesa_ahash_req *creq = ahash_request_ctx(req); mv_cesa_ahash_dma_prepare() local
308 struct mv_cesa_ahash_req *creq = ahash_request_ctx(req); mv_cesa_ahash_std_prepare() local
316 struct mv_cesa_ahash_req *creq = ahash_request_ctx(req); mv_cesa_ahash_dma_step() local
336 struct mv_cesa_ahash_req *creq = ahash_request_ctx(ahashreq); mv_cesa_ahash_step() local
347 struct mv_cesa_ahash_req *creq = ahash_request_ctx(ahashreq); mv_cesa_ahash_process() local
358 struct mv_cesa_ahash_req *creq = ahash_request_ctx(ahashreq); mv_cesa_ahash_complete() local
404 struct mv_cesa_ahash_req *creq = ahash_request_ctx(ahashreq); mv_cesa_ahash_prepare() local
417 struct mv_cesa_ahash_req *creq = ahash_request_ctx(ahashreq); mv_cesa_ahash_req_cleanup() local
441 struct mv_cesa_ahash_req *creq = ahash_request_ctx(req); mv_cesa_ahash_init() local
469 struct mv_cesa_ahash_req *creq = ahash_request_ctx(req); mv_cesa_ahash_cache_req() local
519 mv_cesa_ahash_dma_add_cache(struct mv_cesa_tdma_chain * chain,struct mv_cesa_ahash_req * creq,gfp_t flags) mv_cesa_ahash_dma_add_cache() argument
545 mv_cesa_ahash_dma_last_req(struct mv_cesa_tdma_chain * chain,struct mv_cesa_ahash_dma_iter * dma_iter,struct mv_cesa_ahash_req * creq,unsigned int frag_len,gfp_t flags) mv_cesa_ahash_dma_last_req() argument
627 struct mv_cesa_ahash_req *creq = ahash_request_ctx(req); mv_cesa_ahash_dma_req_init() local
761 struct mv_cesa_ahash_req *creq = ahash_request_ctx(req); mv_cesa_ahash_req_init() local
782 struct mv_cesa_ahash_req *creq = ahash_request_ctx(req); mv_cesa_ahash_queue_req() local
807 struct mv_cesa_ahash_req *creq = ahash_request_ctx(req); mv_cesa_ahash_update() local
816 struct mv_cesa_ahash_req *creq = ahash_request_ctx(req); mv_cesa_ahash_final() local
828 struct mv_cesa_ahash_req *creq = ahash_request_ctx(req); mv_cesa_ahash_finup() local
842 struct mv_cesa_ahash_req *creq = ahash_request_ctx(req); mv_cesa_ahash_export() local
860 struct mv_cesa_ahash_req *creq = ahash_request_ctx(req); mv_cesa_ahash_import() local
892 struct mv_cesa_ahash_req *creq = ahash_request_ctx(req); mv_cesa_md5_init() local
962 struct mv_cesa_ahash_req *creq = ahash_request_ctx(req); mv_cesa_sha1_init() local
1033 struct mv_cesa_ahash_req *creq = ahash_request_ctx(req); mv_cesa_sha256_init() local
[all...]
H A Dcipher.c61 struct mv_cesa_skcipher_req *creq = skcipher_request_ctx(req); in mv_cesa_skcipher_dma_cleanup() local
64 dma_unmap_sg(cesa_dev->dev, req->dst, creq->dst_nents, in mv_cesa_skcipher_dma_cleanup()
66 dma_unmap_sg(cesa_dev->dev, req->src, creq->src_nents, in mv_cesa_skcipher_dma_cleanup()
69 dma_unmap_sg(cesa_dev->dev, req->src, creq->src_nents, in mv_cesa_skcipher_dma_cleanup()
72 mv_cesa_dma_cleanup(&creq->base); in mv_cesa_skcipher_dma_cleanup()
77 struct mv_cesa_skcipher_req *creq = skcipher_request_ctx(req); in mv_cesa_skcipher_cleanup() local
78 struct mv_cesa_engine *engine = creq->base.engine; in mv_cesa_skcipher_cleanup()
80 if (mv_cesa_req_get_type(&creq->base) == CESA_DMA_REQ) in mv_cesa_skcipher_cleanup()
88 struct mv_cesa_skcipher_req *creq = skcipher_request_ctx(req); in mv_cesa_skcipher_std_step() local
89 struct mv_cesa_skcipher_std_req *sreq = &creq in mv_cesa_skcipher_std_step()
129 struct mv_cesa_skcipher_req *creq = skcipher_request_ctx(req); mv_cesa_skcipher_std_process() local
149 struct mv_cesa_skcipher_req *creq = skcipher_request_ctx(skreq); mv_cesa_skcipher_process() local
161 struct mv_cesa_skcipher_req *creq = skcipher_request_ctx(skreq); mv_cesa_skcipher_step() local
172 struct mv_cesa_skcipher_req *creq = skcipher_request_ctx(req); mv_cesa_skcipher_dma_prepare() local
181 struct mv_cesa_skcipher_req *creq = skcipher_request_ctx(req); mv_cesa_skcipher_std_prepare() local
192 struct mv_cesa_skcipher_req *creq = skcipher_request_ctx(skreq); mv_cesa_skcipher_prepare() local
214 struct mv_cesa_skcipher_req *creq = skcipher_request_ctx(skreq); mv_cesa_skcipher_complete() local
317 struct mv_cesa_skcipher_req *creq = skcipher_request_ctx(req); mv_cesa_skcipher_dma_req_init() local
412 struct mv_cesa_skcipher_req *creq = skcipher_request_ctx(req); mv_cesa_skcipher_std_req_init() local
427 struct mv_cesa_skcipher_req *creq = skcipher_request_ctx(req); mv_cesa_skcipher_req_init() local
461 struct mv_cesa_skcipher_req *creq = skcipher_request_ctx(req); mv_cesa_skcipher_queue_req() local
[all...]
H A Dcesa.c170 struct mv_cesa_req *creq) in mv_cesa_queue_req() argument
173 struct mv_cesa_engine *engine = creq->engine; in mv_cesa_queue_req()
177 if ((mv_cesa_req_get_type(creq) == CESA_DMA_REQ) && in mv_cesa_queue_req()
179 mv_cesa_tdma_chain(engine, creq); in mv_cesa_queue_req()
H A Dcesa.h734 struct mv_cesa_req *creq);
/linux/drivers/infiniband/hw/bnxt_re/
H A Dqplib_rcfw.c168 bnxt_qplib_service_creq(&rcfw->creq.creq_tasklet); in __wait_for_resp()
209 bnxt_qplib_service_creq(&rcfw->creq.creq_tasklet); in __block_for_resp()
405 bnxt_qplib_service_creq(&rcfw->creq.creq_tasklet); in __poll_for_resp()
614 rc = rcfw->creq.aeq_handler(rcfw, (void *)func_event, NULL); in bnxt_qplib_process_func_event()
646 rc = rcfw->creq.aeq_handler(rcfw, qp_event, qp); in bnxt_qplib_process_qp_event()
659 * is always called with creq->lock held. Using in bnxt_qplib_process_qp_event()
733 struct bnxt_qplib_rcfw *rcfw = from_tasklet(rcfw, t, creq.creq_tasklet); in bnxt_qplib_service_creq()
734 struct bnxt_qplib_creq_ctx *creq = &rcfw->creq; in bnxt_qplib_service_creq() local
736 struct bnxt_qplib_hwq *hwq = &creq in bnxt_qplib_service_creq()
793 struct bnxt_qplib_creq_ctx *creq; bnxt_qplib_creq_irq() local
931 struct bnxt_qplib_creq_ctx *creq; bnxt_qplib_alloc_rcfw_channel() local
983 struct bnxt_qplib_creq_ctx *creq; bnxt_qplib_rcfw_stop_irq() local
1006 struct bnxt_qplib_creq_ctx *creq; bnxt_qplib_disable_rcfw_channel() local
1026 struct bnxt_qplib_creq_ctx *creq; bnxt_qplib_rcfw_start_irq() local
1133 struct bnxt_qplib_creq_ctx *creq; bnxt_qplib_start_rcfw() local
1160 struct bnxt_qplib_creq_ctx *creq; bnxt_qplib_enable_rcfw_channel() local
[all...]
H A Dmain.c2048 bnxt_re_net_ring_free(rdev, rdev->rcfw.creq.ring_id, type); in bnxt_re_dev_uninit()
2078 struct bnxt_qplib_creq_ctx *creq; in bnxt_re_dev_init() local
2142 creq = &rdev->rcfw.creq; in bnxt_re_dev_init()
2143 rattr.dma_arr = creq->hwq.pbl[PBL_LVL_0].pg_map_arr; in bnxt_re_dev_init()
2144 rattr.pages = creq->hwq.pbl[creq->hwq.level].pg_count; in bnxt_re_dev_init()
2149 rc = bnxt_re_net_ring_alloc(rdev, &rattr, &creq->ring_id); in bnxt_re_dev_init()
2259 bnxt_re_net_ring_free(rdev, rdev->rcfw.creq.ring_id, type); in bnxt_re_dev_init()
H A Dqplib_rcfw.h225 struct bnxt_qplib_creq_ctx creq; member
/linux/drivers/crypto/cavium/nitrox/
H A Dnitrox_req.h213 struct se_crypto_request creq; member
623 struct se_crypto_request *creq = &nkreq->creq; in alloc_src_req_buf() local
625 nkreq->src = alloc_req_buf(nents, ivsize, creq->gfp); in alloc_src_req_buf()
648 struct se_crypto_request *creq = &nkreq->creq; in nitrox_creq_set_src_sg() local
650 creq->src = nitrox_creq_src_sg(iv, ivsize); in nitrox_creq_set_src_sg()
651 sg = creq->src; in nitrox_creq_set_src_sg()
670 struct se_crypto_request *creq = &nkreq->creq; in alloc_dst_req_buf() local
681 struct se_crypto_request *creq = &nkreq->creq; nitrox_creq_set_orh() local
689 struct se_crypto_request *creq = &nkreq->creq; nitrox_creq_set_comp() local
704 struct se_crypto_request *creq = &nkreq->creq; nitrox_creq_set_dst_sg() local
[all...]
H A Dnitrox_aead.c153 struct se_crypto_request *creq = &rctx->nkreq.creq; in nitrox_set_creq() local
157 creq->flags = rctx->flags; in nitrox_set_creq()
158 creq->gfp = (rctx->flags & CRYPTO_TFM_REQ_MAY_SLEEP) ? GFP_KERNEL : in nitrox_set_creq()
161 creq->ctrl.value = 0; in nitrox_set_creq()
162 creq->opcode = FLEXI_CRYPTO_ENCRYPT_HMAC; in nitrox_set_creq()
163 creq->ctrl.s.arg = rctx->ctrl_arg; in nitrox_set_creq()
165 creq->gph.param0 = cpu_to_be16(rctx->cryptlen); in nitrox_set_creq()
166 creq->gph.param1 = cpu_to_be16(rctx->cryptlen + rctx->assoclen); in nitrox_set_creq()
167 creq in nitrox_set_creq()
218 struct se_crypto_request *creq = &rctx->nkreq.creq; nitrox_aes_gcm_enc() local
252 struct se_crypto_request *creq = &rctx->nkreq.creq; nitrox_aes_gcm_dec() local
446 struct se_crypto_request *creq = &aead_rctx->nkreq.creq; nitrox_rfc4106_enc() local
478 struct se_crypto_request *creq = &aead_rctx->nkreq.creq; nitrox_rfc4106_dec() local
[all...]
H A Dnitrox_skcipher.c92 if (nkreq->creq.ctrl.s.arg == ENCRYPT) { in nitrox_cbc_cipher_callback()
253 struct se_crypto_request *creq; in nitrox_skcipher_crypt() local
256 creq = &nkreq->creq; in nitrox_skcipher_crypt()
257 creq->flags = skreq->base.flags; in nitrox_skcipher_crypt()
258 creq->gfp = (skreq->base.flags & CRYPTO_TFM_REQ_MAY_SLEEP) ? in nitrox_skcipher_crypt()
262 creq->ctrl.value = 0; in nitrox_skcipher_crypt()
263 creq->opcode = FLEXI_CRYPTO_ENCRYPT_HMAC; in nitrox_skcipher_crypt()
264 creq->ctrl.s.arg = (enc ? ENCRYPT : DECRYPT); in nitrox_skcipher_crypt()
266 creq in nitrox_skcipher_crypt()
[all...]
H A Dnitrox_reqmgr.c214 struct se_crypto_request *creq) in softreq_map_iobuf() argument
218 ret = dma_map_inbufs(sr, creq); in softreq_map_iobuf()
222 ret = dma_map_outbufs(sr, creq); in softreq_map_iobuf()
/linux/crypto/
H A Dchacha20poly1305.c103 struct chacha_req *creq = &rctx->u.chacha; in chacha_decrypt() local
110 chacha_iv(creq->iv, req, 1); in chacha_decrypt()
117 skcipher_request_set_callback(&creq->req, rctx->flags, in chacha_decrypt()
119 skcipher_request_set_tfm(&creq->req, ctx->chacha); in chacha_decrypt()
120 skcipher_request_set_crypt(&creq->req, src, dst, in chacha_decrypt()
121 rctx->cryptlen, creq->iv); in chacha_decrypt()
122 err = crypto_skcipher_decrypt(&creq->req); in chacha_decrypt()
207 struct chacha_req *creq = &rctx->u.chacha; in poly_genkey() local
219 sg_init_one(creq->src, rctx->key, sizeof(rctx->key)); in poly_genkey()
221 chacha_iv(creq in poly_genkey()
245 struct chacha_req *creq = &rctx->u.chacha; chacha_encrypt() local
[all...]
H A Dpcrypt.c96 struct aead_request *creq = pcrypt_request_ctx(preq); in pcrypt_aead_encrypt() local
110 aead_request_set_tfm(creq, ctx->child); in pcrypt_aead_encrypt()
111 aead_request_set_callback(creq, flags & ~CRYPTO_TFM_REQ_MAY_SLEEP, in pcrypt_aead_encrypt()
113 aead_request_set_crypt(creq, req->src, req->dst, in pcrypt_aead_encrypt()
115 aead_request_set_ad(creq, req->assoclen); in pcrypt_aead_encrypt()
122 return crypto_aead_encrypt(creq); in pcrypt_aead_encrypt()
147 struct aead_request *creq = pcrypt_request_ctx(preq); in pcrypt_aead_decrypt() local
161 aead_request_set_tfm(creq, ctx->child); in pcrypt_aead_decrypt()
162 aead_request_set_callback(creq, flags & ~CRYPTO_TFM_REQ_MAY_SLEEP, in pcrypt_aead_decrypt()
164 aead_request_set_crypt(creq, re in pcrypt_aead_decrypt()
[all...]
/linux/tools/net/ynl/samples/
H A Dethtool.c13 struct ethtool_channels_get_req_dump creq = {}; in main() local
23 creq._present.header = 1; /* ethtool needs an empty nest, sigh */ in main()
24 channels = ethtool_channels_get_dump(ys, &creq); in main()
/linux/drivers/crypto/ccree/
H A Dcc_request_mgr.c47 struct cc_crypto_req creq; member
348 struct cc_crypto_req *creq; in cc_proc_backlog() local
362 creq = &bli->creq; in cc_proc_backlog()
363 req = creq->user_arg; in cc_proc_backlog()
370 creq->user_cb(dev, req, -EINPROGRESS); in cc_proc_backlog()
387 cc_do_send_request(drvdata, &bli->creq, bli->desc, bli->len, in cc_proc_backlog()
435 memcpy(&bli->creq, cc_req, sizeof(*cc_req)); in cc_send_request()
/linux/drivers/gpu/drm/vmwgfx/
H A Dvmwgfx_surface.c1451 rep->creq = rep_ext.creq.base; in vmw_gb_surface_reference_ioctl()
1725 rep->creq.base.svga3d_flags = SVGA3D_FLAGS_LOWER_32(metadata->flags); in vmw_gb_surface_reference_internal()
1726 rep->creq.base.format = metadata->format; in vmw_gb_surface_reference_internal()
1727 rep->creq.base.mip_levels = metadata->mip_levels[0]; in vmw_gb_surface_reference_internal()
1728 rep->creq.base.drm_surface_flags = 0; in vmw_gb_surface_reference_internal()
1729 rep->creq.base.multisample_count = metadata->multisample_count; in vmw_gb_surface_reference_internal()
1730 rep->creq.base.autogen_filter = metadata->autogen_filter; in vmw_gb_surface_reference_internal()
1731 rep->creq.base.array_size = metadata->array_size; in vmw_gb_surface_reference_internal()
1732 rep->creq in vmw_gb_surface_reference_internal()
[all...]
/linux/net/wireless/
H A Dscan.c3511 struct cfg80211_scan_request_int *creq; in cfg80211_wext_siwscan() local
3531 /* Determine number of channels, needed to allocate creq */ in cfg80211_wext_siwscan()
3541 creq = kzalloc(struct_size(creq, req.channels, n_channels) + in cfg80211_wext_siwscan()
3544 if (!creq) in cfg80211_wext_siwscan()
3547 creq->req.wiphy = wiphy; in cfg80211_wext_siwscan()
3548 creq->req.wdev = dev->ieee80211_ptr; in cfg80211_wext_siwscan()
3550 creq->req.ssids = (void *)creq + in cfg80211_wext_siwscan()
3551 struct_size(creq, re in cfg80211_wext_siwscan()
[all...]
/linux/include/trace/events/
H A Dnetfs.h565 const struct netfs_io_request *creq),
567 TP_ARGS(rreq, creq),
571 __field(unsigned int, creq)
580 __entry->creq = creq->debug_id;
587 __entry->creq,
/linux/include/uapi/drm/
H A Dvmwgfx_drm.h1003 * @creq: The data used as input when the surface was created, as described
1011 struct drm_vmw_gb_surface_create_req creq; member
1210 * @creq: The data used as input when the surface was created, as described
1218 struct drm_vmw_gb_surface_create_ext_req creq; member
/linux/drivers/usb/gadget/function/
H A Df_fs.c3780 const struct usb_ctrlrequest *creq) in ffs_func_setup() argument
3787 pr_vdebug("creq->bRequestType = %02x\n", creq->bRequestType); in ffs_func_setup()
3788 pr_vdebug("creq->bRequest = %02x\n", creq->bRequest); in ffs_func_setup()
3789 pr_vdebug("creq->wValue = %04x\n", le16_to_cpu(creq->wValue)); in ffs_func_setup()
3790 pr_vdebug("creq->wIndex = %04x\n", le16_to_cpu(creq->wIndex)); in ffs_func_setup()
3791 pr_vdebug("creq in ffs_func_setup()
3838 ffs_func_req_match(struct usb_function * f,const struct usb_ctrlrequest * creq,bool config0) ffs_func_req_match() argument
[all...]
/linux/drivers/net/ethernet/marvell/octeontx2/nic/
H A Dotx2_tc.c369 struct nix_mcast_grp_create_req *creq; in otx2_tc_update_mcast() local
375 creq = otx2_mbox_alloc_msg_nix_mcast_grp_create(&nic->mbox); in otx2_tc_update_mcast()
376 if (!creq) { in otx2_tc_update_mcast()
381 creq->dir = NIX_MCAST_INGRESS; in otx2_tc_update_mcast()
391 &creq->hdr); in otx2_tc_update_mcast()
/linux/fs/smb/client/
H A Dsmb2ops.c4330 void *creq; in crypt_message() local
4357 creq = smb2_get_aead_req(tfm, rqst, num_rqst, sign, &iv, &req, &sg, in crypt_message()
4359 if (IS_ERR(creq)) in crypt_message()
4360 return PTR_ERR(creq); in crypt_message()
4388 kvfree_sensitive(creq, sensitive_size); in crypt_message()