| /linux/drivers/crypto/marvell/cesa/ |
| H A D | cipher.c | 425 struct mv_cesa_op_ctx *tmpl) in mv_cesa_skcipher_req_init() argument 446 mv_cesa_update_op_cfg(tmpl, CESA_SA_DESC_CFG_OP_CRYPT_ONLY, in mv_cesa_skcipher_req_init() 450 ret = mv_cesa_skcipher_dma_req_init(req, tmpl); in mv_cesa_skcipher_req_init() 452 ret = mv_cesa_skcipher_std_req_init(req, tmpl); in mv_cesa_skcipher_req_init() 458 struct mv_cesa_op_ctx *tmpl) in mv_cesa_skcipher_queue_req() argument 467 ret = mv_cesa_skcipher_req_init(req, tmpl); in mv_cesa_skcipher_queue_req() 483 struct mv_cesa_op_ctx *tmpl) in mv_cesa_des_op() argument 487 mv_cesa_update_op_cfg(tmpl, CESA_SA_DESC_CFG_CRYPTM_DES, in mv_cesa_des_op() 490 memcpy(tmpl->ctx.skcipher.key, ctx->key, DES_KEY_SIZE); in mv_cesa_des_op() 492 return mv_cesa_skcipher_queue_req(req, tmpl); in mv_cesa_des_op() [all …]
|
| H A D | hash.c | 439 struct mv_cesa_op_ctx *tmpl, bool algo_le) in mv_cesa_ahash_init() argument 444 mv_cesa_update_op_cfg(tmpl, in mv_cesa_ahash_init() 449 mv_cesa_set_mac_op_total_len(tmpl, 0); in mv_cesa_ahash_init() 450 mv_cesa_set_mac_op_frag_len(tmpl, 0); in mv_cesa_ahash_init() 451 creq->op_tmpl = *tmpl; in mv_cesa_ahash_init() 491 struct mv_cesa_op_ctx *tmpl, unsigned int frag_len, in mv_cesa_dma_add_frag() argument 497 op = mv_cesa_dma_add_op(chain, tmpl, false, flags); in mv_cesa_dma_add_frag() 509 if (mv_cesa_mac_op_is_first_frag(tmpl)) in mv_cesa_dma_add_frag() 510 mv_cesa_update_op_cfg(tmpl, in mv_cesa_dma_add_frag() 817 struct mv_cesa_op_ctx *tmpl = &creq->op_tmpl; in mv_cesa_ahash_final() local [all …]
|
| /linux/drivers/crypto/qce/ |
| H A D | sha.c | 42 struct qce_alg_template *tmpl = to_ahash_tmpl(async_req->tfm); in qce_ahash_done() local 43 struct qce_device *qce = tmpl->qce; in qce_ahash_done() 72 qce->async_req_done(tmpl->qce, error); in qce_ahash_done() 80 struct qce_alg_template *tmpl = to_ahash_tmpl(async_req->tfm); in qce_ahash_async_req_handle() local 81 struct qce_device *qce = tmpl->qce; in qce_ahash_async_req_handle() 118 ret = qce_start(async_req, tmpl->crypto_alg_type); in qce_ahash_async_req_handle() 136 struct qce_alg_template *tmpl = to_ahash_tmpl(req->base.tfm); in qce_ahash_init() local 137 const u32 *std_iv = tmpl->std_iv; in qce_ahash_init() 142 rctx->flags = tmpl->alg_flags; in qce_ahash_init() 187 struct qce_alg_template *tmpl = to_ahash_tmpl(req->base.tfm); in qce_ahash_update() local [all …]
|
| H A D | skcipher.c | 32 struct qce_alg_template *tmpl = to_cipher_tmpl(crypto_skcipher_reqtfm(req)); in qce_skcipher_done() local 33 struct qce_device *qce = tmpl->qce; in qce_skcipher_done() 60 qce->async_req_done(tmpl->qce, error); in qce_skcipher_done() 69 struct qce_alg_template *tmpl = to_cipher_tmpl(crypto_skcipher_reqtfm(req)); in qce_skcipher_async_req_handle() local 70 struct qce_device *qce = tmpl->qce; in qce_skcipher_async_req_handle() 152 ret = qce_start(async_req, tmpl->crypto_alg_type); in qce_skcipher_async_req_handle() 261 struct qce_alg_template *tmpl = to_cipher_tmpl(tfm); in qce_skcipher_crypt() local 266 rctx->flags = tmpl->alg_flags; in qce_skcipher_crypt() 307 return tmpl->qce->async_req_enqueue(tmpl->qce, &req->base); in qce_skcipher_crypt() 439 struct qce_alg_template *tmpl; in qce_skcipher_register_one() local [all …]
|
| H A D | aead.c | 29 struct qce_alg_template *tmpl = to_aead_tmpl(crypto_aead_reqtfm(req)); in qce_aead_done() local 30 struct qce_device *qce = tmpl->qce; in qce_aead_done() 96 struct qce_alg_template *tmpl = to_aead_tmpl(crypto_aead_reqtfm(req)); in qce_aead_prepare_result_buf() local 97 struct qce_device *qce = tmpl->qce; in qce_aead_prepare_result_buf() 116 struct qce_alg_template *tmpl = to_aead_tmpl(crypto_aead_reqtfm(req)); in qce_aead_prepare_dst_buf() local 117 struct qce_device *qce = tmpl->qce; in qce_aead_prepare_dst_buf() 304 struct qce_alg_template *tmpl = to_aead_tmpl(crypto_aead_reqtfm(req)); in qce_aead_prepare_buf() local 305 struct qce_device *qce = tmpl->qce; in qce_aead_prepare_buf() 414 struct qce_alg_template *tmpl = to_aead_tmpl(crypto_aead_reqtfm(req)); in qce_aead_async_req_handle() local 415 struct qce_device *qce = tmpl->qce; in qce_aead_async_req_handle() [all …]
|
| /linux/crypto/ |
| H A D | xor.c | 58 static void __init do_xor_register(struct xor_block_template *tmpl) in do_xor_register() argument 60 tmpl->next = template_list; in do_xor_register() 61 template_list = tmpl; in do_xor_register() 83 do_xor_speed(struct xor_block_template *tmpl, void *b1, void *b2) in do_xor_speed() argument 89 tmpl->next = template_list; in do_xor_speed() 90 template_list = tmpl; in do_xor_speed() 101 tmpl->do_2(BENCH_SIZE, b1, b2); in do_xor_speed() 110 tmpl->speed = speed; in do_xor_speed() 112 pr_info(" %-16s: %5d MB/sec\n", tmpl->name, speed); in do_xor_speed()
|
| H A D | ecb.c | 98 struct crypto_template *tmpl, struct rtattr **tb) in lskcipher_alloc_instance_simple2() argument 121 err = crypto_inst_setname(lskcipher_crypto_instance(inst), tmpl->name, in lskcipher_alloc_instance_simple2() 149 static int crypto_ecb_create2(struct crypto_template *tmpl, struct rtattr **tb) in crypto_ecb_create2() argument 154 inst = lskcipher_alloc_instance_simple2(tmpl, tb); in crypto_ecb_create2() 164 err = lskcipher_register_instance(tmpl, inst); in crypto_ecb_create2() 171 static int crypto_ecb_create(struct crypto_template *tmpl, struct rtattr **tb) in crypto_ecb_create() argument 178 inst = lskcipher_alloc_instance_simple(tmpl, tb); in crypto_ecb_create() 180 err = crypto_ecb_create2(tmpl, tb); in crypto_ecb_create() 199 err = lskcipher_register_instance(tmpl, inst); in crypto_ecb_create()
|
| H A D | dh.c | 552 struct crypto_template *tmpl, struct rtattr **tb, in __dh_safe_prime_create() argument 589 tmpl->name, &dh_alg->base); in __dh_safe_prime_create() 605 err = kpp_register_instance(tmpl, inst); in __dh_safe_prime_create() 833 static int dh_ffdhe2048_create(struct crypto_template *tmpl, in dh_ffdhe2048_create() argument 836 return __dh_safe_prime_create(tmpl, tb, &ffdhe2048_prime); in dh_ffdhe2048_create() 839 static int dh_ffdhe3072_create(struct crypto_template *tmpl, in dh_ffdhe3072_create() argument 842 return __dh_safe_prime_create(tmpl, tb, &ffdhe3072_prime); in dh_ffdhe3072_create() 845 static int dh_ffdhe4096_create(struct crypto_template *tmpl, in dh_ffdhe4096_create() argument 848 return __dh_safe_prime_create(tmpl, tb, &ffdhe4096_prime); in dh_ffdhe4096_create() 851 static int dh_ffdhe6144_create(struct crypto_template *tmpl, in dh_ffdhe6144_create() argument [all …]
|
| H A D | algboss.c | 53 struct crypto_template *tmpl; in cryptomgr_probe() local 56 tmpl = crypto_lookup_template(param->template); in cryptomgr_probe() 57 if (!tmpl) in cryptomgr_probe() 61 err = tmpl->create(tmpl, param->tb); in cryptomgr_probe() 64 crypto_tmpl_put(tmpl); in cryptomgr_probe()
|
| H A D | hmac.c | 182 static int __hmac_create_shash(struct crypto_template *tmpl, in __hmac_create_shash() argument 243 err = shash_register_instance(tmpl, inst); in __hmac_create_shash() 448 static int hmac_create_ahash(struct crypto_template *tmpl, struct rtattr **tb, in hmac_create_ahash() argument 480 err = crypto_inst_setname(ahash_crypto_instance(inst), tmpl->name, alg); in hmac_create_ahash() 511 err = ahash_register_instance(tmpl, inst); in hmac_create_ahash() 519 static int hmac_create(struct crypto_template *tmpl, struct rtattr **tb) in hmac_create() argument 532 return hmac_create_ahash(tmpl, tb, mask); in hmac_create() 538 return __hmac_create_shash(tmpl, tb, mask); in hmac_create() 541 static int hmac_create_shash(struct crypto_template *tmpl, struct rtattr **tb) in hmac_create_shash() argument 550 return __hmac_create_shash(tmpl, tb, mask); in hmac_create_shash()
|
| /linux/drivers/scsi/fnic/ |
| H A D | cq_exch_desc.h | 16 u8 tmpl; member 50 u16 tmpl; member 83 u16 *tmpl, in cq_fcp_rq_desc_dec() argument 104 *tmpl = desc_ptr->tmpl & CQ_FCP_RQ_DESC_TMPL_MASK; in cq_fcp_rq_desc_dec() 129 u16 tmpl; member 157 u16 *tmpl, in cq_sgl_desc_dec() argument 166 *tmpl = desc_ptr->tmpl & CQ_SGL_TMPL_MASK; in cq_sgl_desc_dec()
|
| /linux/include/net/netfilter/ |
| H A D | nf_conntrack_zones.h | 29 nf_ct_zone_tmpl(const struct nf_conn *tmpl, const struct sk_buff *skb, in nf_ct_zone_tmpl() argument 33 if (!tmpl) in nf_ct_zone_tmpl() 36 if (tmpl->zone.flags & NF_CT_FLAG_MARK) in nf_ct_zone_tmpl() 37 return nf_ct_zone_init(tmp, skb->mark, tmpl->zone.dir, 0); in nf_ct_zone_tmpl() 39 return nf_ct_zone(tmpl); in nf_ct_zone_tmpl()
|
| H A D | nf_conntrack_synproxy.h | 33 const struct nf_conn *tmpl) in nf_ct_add_synproxy() argument 36 if (tmpl && nfct_synproxy(tmpl)) { in nf_ct_add_synproxy()
|
| H A D | nf_tables.h | 754 static inline void nft_set_ext_prepare(struct nft_set_ext_tmpl *tmpl) in nft_set_ext_prepare() argument 756 memset(tmpl, 0, sizeof(*tmpl)); in nft_set_ext_prepare() 757 tmpl->len = sizeof(struct nft_set_ext); in nft_set_ext_prepare() 760 static inline int nft_set_ext_add_length(struct nft_set_ext_tmpl *tmpl, u8 id, in nft_set_ext_add_length() argument 763 tmpl->len = ALIGN(tmpl->len, nft_set_ext_types[id].align); in nft_set_ext_add_length() 764 if (tmpl->len > U8_MAX) in nft_set_ext_add_length() 767 tmpl->offset[id] = tmpl->len; in nft_set_ext_add_length() 768 tmpl->ext_len[id] = nft_set_ext_types[id].len + len; in nft_set_ext_add_length() 769 tmpl->len += tmpl->ext_len[id]; in nft_set_ext_add_length() 774 static inline int nft_set_ext_add(struct nft_set_ext_tmpl *tmpl, u8 id) in nft_set_ext_add() argument [all …]
|
| /linux/drivers/s390/net/ |
| H A D | fsm.c | 19 int nr_events, const fsm_node *tmpl, int tmpl_len, gfp_t order) in init_fsm() argument 58 if ((tmpl[i].cond_state >= nr_states) || in init_fsm() 59 (tmpl[i].cond_event >= nr_events) ) { in init_fsm() 62 name, i, (long)tmpl[i].cond_state, (long)f->nr_states, in init_fsm() 63 (long)tmpl[i].cond_event, (long)f->nr_events); in init_fsm() 67 m[nr_states * tmpl[i].cond_event + tmpl[i].cond_state] = in init_fsm() 68 tmpl[i].function; in init_fsm()
|
| /linux/net/netfilter/ |
| H A D | nf_conntrack_core.c | 527 struct nf_conn *tmpl, *p; in nf_ct_tmpl_alloc() local 530 tmpl = kzalloc(sizeof(*tmpl) + NFCT_INFOMASK, flags); in nf_ct_tmpl_alloc() 531 if (!tmpl) in nf_ct_tmpl_alloc() 534 p = tmpl; in nf_ct_tmpl_alloc() 535 tmpl = (struct nf_conn *)NFCT_ALIGN((unsigned long)p); in nf_ct_tmpl_alloc() 536 if (tmpl != p) in nf_ct_tmpl_alloc() 537 tmpl->proto.tmpl_padto = (char *)tmpl - (char *)p; in nf_ct_tmpl_alloc() 539 tmpl = kzalloc_obj(*tmpl, flags); in nf_ct_tmpl_alloc() 540 if (!tmpl) in nf_ct_tmpl_alloc() 544 tmpl->status = IPS_TEMPLATE; in nf_ct_tmpl_alloc() [all …]
|
| H A D | nft_dynset.c | 17 struct nft_set_ext_tmpl tmpl; member 68 elem_priv = nft_set_elem_init(set, &priv->tmpl, in nft_dynset_new() 130 nft_set_ext_add_length(&priv->tmpl, NFT_SET_EXT_EXPRESSIONS, in nft_dynset_ext_add_expr() 316 nft_set_ext_prepare(&priv->tmpl); in nft_dynset_init() 317 nft_set_ext_add_length(&priv->tmpl, NFT_SET_EXT_KEY, set->klen); in nft_dynset_init() 319 nft_set_ext_add_length(&priv->tmpl, NFT_SET_EXT_DATA, set->dlen); in nft_dynset_init() 326 nft_set_ext_add(&priv->tmpl, NFT_SET_EXT_TIMEOUT); in nft_dynset_init()
|
| /linux/drivers/media/dvb-frontends/ |
| H A D | mn88443x.c | 273 u32 cpmon, tmpu, tmpl, flg; in mn88443x_s_read_status() local 307 regmap_read(r_s, CNRDXL_S, &tmpl); in mn88443x_s_read_status() 308 x = (tmpu << 8) | tmpl; in mn88443x_s_read_status() 310 regmap_read(r_s, CNRDYL_S, &tmpl); in mn88443x_s_read_status() 311 y = (tmpu << 8) | tmpl; in mn88443x_s_read_status() 349 regmap_read(r_s, BERVRDL_S, &tmpl); in mn88443x_s_read_status() 350 bit_err = (tmpu << 8) | tmpl; in mn88443x_s_read_status() 475 u32 seqrd, st, flg, tmpu, tmpm, tmpl; in mn88443x_t_read_status() local 495 regmap_read(r_t, AGCRDL_T, &tmpl); in mn88443x_t_read_status() 496 agc = (tmpu << 8) | tmpl; in mn88443x_t_read_status() [all …]
|
| /linux/drivers/crypto/inside-secure/eip93/ |
| H A D | eip93-cipher.c | 53 struct eip93_alg_template *tmpl = container_of(tfm->__crt_alg, in eip93_skcipher_cra_init() local 61 ctx->eip93 = tmpl->eip93; in eip93_skcipher_cra_init() 62 ctx->type = tmpl->type; in eip93_skcipher_cra_init() 85 struct eip93_alg_template *tmpl = container_of(tfm->__crt_alg, in eip93_skcipher_setkey() local 90 u32 flags = tmpl->flags; in eip93_skcipher_setkey() 178 struct eip93_alg_template *tmpl = container_of(req->base.tfm->__crt_alg, in eip93_skcipher_encrypt() local 181 rctx->flags = tmpl->flags; in eip93_skcipher_encrypt() 191 struct eip93_alg_template *tmpl = container_of(req->base.tfm->__crt_alg, in eip93_skcipher_decrypt() local 196 rctx->flags = tmpl->flags; in eip93_skcipher_decrypt()
|
| /linux/drivers/mtd/nand/spi/ |
| H A D | core.c | 1417 struct spinand_mem_ops *tmpl = &spinand->ssdr_op_templates; in spinand_init_ssdr_templates() local 1419 tmpl->reset = (struct spi_mem_op)SPINAND_RESET_1S_0_0_OP; in spinand_init_ssdr_templates() 1420 tmpl->readid = (struct spi_mem_op)SPINAND_READID_1S_1S_1S_OP(0, 0, NULL, 0); in spinand_init_ssdr_templates() 1421 tmpl->wr_en = (struct spi_mem_op)SPINAND_WR_EN_1S_0_0_OP; in spinand_init_ssdr_templates() 1422 tmpl->wr_dis = (struct spi_mem_op)SPINAND_WR_DIS_1S_0_0_OP; in spinand_init_ssdr_templates() 1423 tmpl->set_feature = (struct spi_mem_op)SPINAND_SET_FEATURE_1S_1S_1S_OP(0, NULL); in spinand_init_ssdr_templates() 1424 tmpl->get_feature = (struct spi_mem_op)SPINAND_GET_FEATURE_1S_1S_1S_OP(0, NULL); in spinand_init_ssdr_templates() 1425 tmpl->blk_erase = (struct spi_mem_op)SPINAND_BLK_ERASE_1S_1S_0_OP(0); in spinand_init_ssdr_templates() 1426 tmpl->page_read = (struct spi_mem_op)SPINAND_PAGE_READ_1S_1S_0_OP(0); in spinand_init_ssdr_templates() 1427 tmpl->prog_exec = (struct spi_mem_op)SPINAND_PROG_EXEC_1S_1S_0_OP(0); in spinand_init_ssdr_templates() [all …]
|
| /linux/include/crypto/ |
| H A D | algapi.h | 62 struct crypto_template *tmpl; member 82 int (*create)(struct crypto_template *tmpl, struct rtattr **tb); 141 int crypto_register_template(struct crypto_template *tmpl); 143 void crypto_unregister_template(struct crypto_template *tmpl); 147 int crypto_register_instance(struct crypto_template *tmpl,
|
| /linux/tools/testing/selftests/net/ |
| H A D | l2tp.sh | 230 tmpl proto esp mode transport 234 tmpl proto esp mode transport 238 tmpl proto esp mode transport 242 tmpl proto esp mode transport 269 tmpl proto esp mode transport 273 tmpl proto esp mode transport 277 tmpl proto esp mode transport 281 tmpl proto esp mode transport
|
| H A D | vrf-xfrm-tests.sh | 218 tmpl src ${HOST1_4} dst ${HOST2_4} proto esp mode tunnel 223 tmpl src ${HOST1_4} dst ${HOST2_4} proto esp mode tunnel 228 tmpl src ${HOST2_4} dst ${HOST1_4} proto esp mode tunnel 233 tmpl src ${HOST2_4} dst ${HOST1_4} proto esp mode tunnel 239 tmpl src ${HOST1_6} dst ${HOST2_6} proto esp mode tunnel 244 tmpl src ${HOST1_6} dst ${HOST2_6} proto esp mode tunnel 249 tmpl src ${HOST2_6} dst ${HOST1_6} proto esp mode tunnel 254 tmpl src ${HOST2_6} dst ${HOST1_6} proto esp mode tunnel
|
| /linux/drivers/net/wireless/ti/wlcore/ |
| H A D | init.c | 147 struct wl12xx_disconn_template *tmpl; in wl1271_ap_init_deauth_template() local 151 tmpl = kzalloc_obj(*tmpl); in wl1271_ap_init_deauth_template() 152 if (!tmpl) { in wl1271_ap_init_deauth_template() 157 tmpl->header.frame_ctl = cpu_to_le16(IEEE80211_FTYPE_MGMT | in wl1271_ap_init_deauth_template() 163 tmpl, sizeof(*tmpl), 0, rate); in wl1271_ap_init_deauth_template() 166 kfree(tmpl); in wl1271_ap_init_deauth_template()
|
| /linux/net/sched/ |
| H A D | act_ct.c | 882 if (params->tmpl) { in tcf_ct_params_free() 884 nf_connlabels_put(nf_ct_net(params->tmpl)); in tcf_ct_params_free() 886 nf_ct_put(params->tmpl); in tcf_ct_params_free() 969 struct nf_conn *tmpl = NULL; in tcf_ct_act() local 985 tmpl = p->tmpl; in tcf_ct_act() 1031 if (tmpl) { in tcf_ct_act() 1033 nf_conntrack_get(&tmpl->ct_general); in tcf_ct_act() 1034 nf_ct_set(skb, tmpl, IP_CT_NEW); in tcf_ct_act() 1057 err = __nf_ct_try_assign_helper(ct, p->tmpl, GFP_ATOMIC); in tcf_ct_act() 1242 struct nf_conn *tmpl; in tcf_ct_fill_params() local [all …]
|