Lines Matching full:rx
76 STAT_BADMSGS = STAT_BADKEYS, /* rx only */
109 passive:2, /* rx only */
113 passive:2, /* rx only */
137 * @users: the number of the key users (TX/RX)
175 * struct tipc_crypto - TIPC TX/RX crypto structure
177 * @node: TIPC node (RX)
179 * @peer_rx_active: replicated peer RX active key index
180 * @key_gen: TX/RX key generation
185 * @work: delayed work sched for TX/RX
241 /* struct tipc_crypto_rx_ctx - RX context for callbacks */
281 static bool tipc_crypto_key_try_align(struct tipc_crypto *rx, u8 new_pending);
283 struct tipc_crypto *rx,
286 static void tipc_crypto_key_synch(struct tipc_crypto *rx, struct sk_buff *skb);
301 static bool tipc_crypto_key_rcv(struct tipc_crypto *rx, struct tipc_msg *hdr);
901 pr_err("RX: skb_cow_data() returned %d\n", nsg); in tipc_aead_decrypt()
917 pr_err("RX: skb_to_sgvec() returned %d, nsg %d\n", rc, nsg); in tipc_aead_decrypt()
1032 * @__rx: RX crypto handle if dest is "known"
1210 struct tipc_crypto *tx, *rx; in tipc_crypto_key_flush() local
1216 rx = c; in tipc_crypto_key_flush()
1217 tx = tipc_net(rx->net)->crypto_tx; in tipc_crypto_key_flush()
1218 if (cancel_delayed_work(&rx->work)) { in tipc_crypto_key_flush()
1219 kfree(rx->skey); in tipc_crypto_key_flush()
1220 rx->skey = NULL; in tipc_crypto_key_flush()
1221 atomic_xchg(&rx->key_distr, 0); in tipc_crypto_key_flush()
1222 tipc_node_put(rx->node); in tipc_crypto_key_flush()
1224 /* RX stopping => decrease TX key users if any */ in tipc_crypto_key_flush()
1225 k = atomic_xchg(&rx->peer_rx_active, 0); in tipc_crypto_key_flush()
1242 * tipc_crypto_key_try_align - Align RX keys if possible
1243 * @rx: RX crypto handle
1254 static bool tipc_crypto_key_try_align(struct tipc_crypto *rx, u8 new_pending) in tipc_crypto_key_try_align() argument
1262 spin_lock(&rx->lock); in tipc_crypto_key_try_align()
1263 key = rx->key; in tipc_crypto_key_try_align()
1272 if (tipc_aead_users(rx->aead[key.pending]) > 0) in tipc_crypto_key_try_align()
1276 tmp1 = tipc_aead_rcu_ptr(rx->aead[key.pending], &rx->lock); in tipc_crypto_key_try_align()
1279 rcu_assign_pointer(rx->aead[key.pending], NULL); in tipc_crypto_key_try_align()
1283 tmp2 = rcu_replace_pointer(rx->aead[key.passive], tmp2, lockdep_is_held(&rx->lock)); in tipc_crypto_key_try_align()
1289 tipc_crypto_key_set_state(rx, new_passive, 0, new_pending); in tipc_crypto_key_try_align()
1290 rcu_assign_pointer(rx->aead[new_pending], tmp1); in tipc_crypto_key_try_align()
1292 rcu_assign_pointer(rx->aead[new_passive], tmp2); in tipc_crypto_key_try_align()
1295 pr_info_ratelimited("%s: key[%d] -> key[%d]\n", rx->name, key.pending, in tipc_crypto_key_try_align()
1299 spin_unlock(&rx->lock); in tipc_crypto_key_try_align()
1306 * @rx: RX crypto handle (can be NULL)
1317 struct tipc_crypto *rx, in tipc_crypto_key_pick_tx() argument
1332 skb_cb->tx_clone_ctx.rx = rx; in tipc_crypto_key_pick_tx()
1374 * @rx: RX crypto handle
1377 * This function updates the peer node related data as the peer RX active key
1387 static void tipc_crypto_key_synch(struct tipc_crypto *rx, struct sk_buff *skb) in tipc_crypto_key_synch() argument
1390 struct tipc_crypto *tx = tipc_net(rx->net)->crypto_tx; in tipc_crypto_key_synch()
1392 u32 self = tipc_own_addr(rx->net); in tipc_crypto_key_synch()
1396 /* Update RX 'key_master' flag according to peer, also mark "legacy" if in tipc_crypto_key_synch()
1399 rx->key_master = ehdr->master_key; in tipc_crypto_key_synch()
1400 if (!rx->key_master) in tipc_crypto_key_synch()
1413 !atomic_cmpxchg(&rx->key_distr, 0, KEY_DISTR_SCHED)) { in tipc_crypto_key_synch()
1417 if (queue_delayed_work(tx->wq, &rx->work, delay)) in tipc_crypto_key_synch()
1418 tipc_node_get(rx->node); in tipc_crypto_key_synch()
1422 atomic_xchg(&rx->key_distr, 0); in tipc_crypto_key_synch()
1425 /* Case 2: Peer RX active key has changed, let's update own TX users */ in tipc_crypto_key_synch()
1426 cur = atomic_read(&rx->peer_rx_active); in tipc_crypto_key_synch()
1430 atomic_cmpxchg(&rx->peer_rx_active, cur, new) == cur) { in tipc_crypto_key_synch()
1436 atomic64_set(&rx->sndnxt, 0); in tipc_crypto_key_synch()
1441 tx->name, cur, new, rx->name); in tipc_crypto_key_synch()
1504 scnprintf(c->name, 48, "%s(%s)", (is_rx(c)) ? "RX" : "TX", in tipc_crypto_start()
1546 void tipc_crypto_timeout(struct tipc_crypto *rx) in tipc_crypto_timeout() argument
1548 struct tipc_net *tn = tipc_net(rx->net); in tipc_crypto_timeout()
1572 /* RX pending: having user -> active */ in tipc_crypto_timeout()
1573 spin_lock(&rx->lock); in tipc_crypto_timeout()
1574 key = rx->key; in tipc_crypto_timeout()
1575 if (!key.pending || tipc_aead_users(rx->aead[key.pending]) <= 0) in tipc_crypto_timeout()
1581 rx->timer2 = jiffies; in tipc_crypto_timeout()
1582 tipc_crypto_key_set_state(rx, key.passive, key.active, 0); in tipc_crypto_timeout()
1583 this_cpu_inc(rx->stats->stat[STAT_SWITCHES]); in tipc_crypto_timeout()
1584 pr_info("%s: key[%d] is activated\n", rx->name, key.pending); in tipc_crypto_timeout()
1588 /* RX pending: not working -> remove */ in tipc_crypto_timeout()
1589 if (!key.pending || tipc_aead_users(rx->aead[key.pending]) > -10) in tipc_crypto_timeout()
1592 tipc_crypto_key_set_state(rx, key.passive, key.active, 0); in tipc_crypto_timeout()
1593 tipc_crypto_key_detach(rx->aead[key.pending], &rx->lock); in tipc_crypto_timeout()
1594 pr_debug("%s: key[%d] is removed\n", rx->name, key.pending); in tipc_crypto_timeout()
1598 /* RX active: timed out or no user -> pending */ in tipc_crypto_timeout()
1601 if (time_before(jiffies, rx->timer1 + TIPC_RX_ACTIVE_LIM) && in tipc_crypto_timeout()
1602 tipc_aead_users(rx->aead[key.active]) > 0) in tipc_crypto_timeout()
1609 rx->timer2 = jiffies; in tipc_crypto_timeout()
1610 tipc_crypto_key_set_state(rx, key.passive, 0, key.pending); in tipc_crypto_timeout()
1611 tipc_aead_users_set(rx->aead[key.pending], 0); in tipc_crypto_timeout()
1612 pr_debug("%s: key[%d] is deactivated\n", rx->name, key.active); in tipc_crypto_timeout()
1616 /* RX passive: outdated or not working -> free */ in tipc_crypto_timeout()
1619 if (time_before(jiffies, rx->timer2 + TIPC_RX_PASSIVE_LIM) && in tipc_crypto_timeout()
1620 tipc_aead_users(rx->aead[key.passive]) > -10) in tipc_crypto_timeout()
1623 tipc_crypto_key_set_state(rx, 0, key.active, key.pending); in tipc_crypto_timeout()
1624 tipc_crypto_key_detach(rx->aead[key.passive], &rx->lock); in tipc_crypto_timeout()
1625 pr_debug("%s: key[%d] is freed\n", rx->name, key.passive); in tipc_crypto_timeout()
1628 spin_unlock(&rx->lock); in tipc_crypto_timeout()
1642 tipc_crypto_do_cmd(rx->net, cmd); in tipc_crypto_timeout()
1790 * @rx: RX crypto handle
1798 * Note: RX key(s) can be re-aligned, or in case of no key suitable, TX
1809 int tipc_crypto_rcv(struct net *net, struct tipc_crypto *rx, in tipc_crypto_rcv() argument
1824 if (unlikely(!rx || tx_key == KEY_MASTER)) in tipc_crypto_rcv()
1827 /* Pick RX key according to TX key if any */ in tipc_crypto_rcv()
1828 key = rx->key; in tipc_crypto_rcv()
1833 /* Unknown key, let's try to align RX key(s) */ in tipc_crypto_rcv()
1834 if (tipc_crypto_key_try_align(rx, tx_key)) in tipc_crypto_rcv()
1839 aead = tipc_crypto_key_pick_tx(tx, rx, *skb, tx_key); in tipc_crypto_rcv()
1847 aead = tipc_aead_get(rx->aead[tx_key]); in tipc_crypto_rcv()
1852 stats = ((rx) ?: tx)->stats; in tipc_crypto_rcv()
1867 if (rx) { in tipc_crypto_rcv()
1868 /* Mark rx->nokey only if we dont have a in tipc_crypto_rcv()
1873 rx->nokey = !(rx->skey || in tipc_crypto_rcv()
1874 rcu_access_pointer(rx->aead[n])); in tipc_crypto_rcv()
1876 rx->name, rx->nokey, in tipc_crypto_rcv()
1877 tx_key, rx->key.keys); in tipc_crypto_rcv()
1878 tipc_node_put(rx->node); in tipc_crypto_rcv()
1897 struct tipc_crypto *rx = aead->crypto; in tipc_crypto_rcv_complete() local
1904 rx = skb_cb->tx_clone_ctx.rx; in tipc_crypto_rcv_complete()
1905 pr_debug("TX->RX(%s): err %d, aead %p, skb->next %p, flags %x\n", in tipc_crypto_rcv_complete()
1906 (rx) ? tipc_node_get_id_str(rx->node) : "-", err, aead, in tipc_crypto_rcv_complete()
1923 if (!rx) { in tipc_crypto_rcv_complete()
1927 rx = tipc_node_crypto_rx(n); in tipc_crypto_rcv_complete()
1928 if (unlikely(!rx)) in tipc_crypto_rcv_complete()
1937 if (tipc_crypto_key_attach(rx, tmp, ehdr->tx_key, false) < 0) { in tipc_crypto_rcv_complete()
1950 /* Set the RX key's user */ in tipc_crypto_rcv_complete()
1953 /* Mark this point, RX works */ in tipc_crypto_rcv_complete()
1954 rx->timer1 = jiffies; in tipc_crypto_rcv_complete()
1960 /* Mark this point, RX passive still works */ in tipc_crypto_rcv_complete()
1961 if (rx->key.passive && ehdr->tx_key == rx->key.passive) in tipc_crypto_rcv_complete()
1962 rx->timer2 = jiffies; in tipc_crypto_rcv_complete()
1975 tipc_crypto_key_synch(rx, *skb); in tipc_crypto_rcv_complete()
1993 if (rx) in tipc_crypto_rcv_complete()
1994 tipc_node_put(rx->node); in tipc_crypto_rcv_complete()
2000 struct tipc_crypto *tx = tn->crypto_tx, *rx; in tipc_crypto_do_cmd() local
2025 rx = tipc_node_crypto_rx_by_list(p); in tipc_crypto_do_cmd()
2026 pr_info("RX(%7.7s)\n%s", tipc_node_get_id_str(rx->node), in tipc_crypto_do_cmd()
2027 tipc_crypto_key_dump(rx, buf)); in tipc_crypto_do_cmd()
2052 rx = tipc_node_crypto_rx_by_list(p); in tipc_crypto_do_cmd()
2053 j = scnprintf(buf, 200, "RX(%7.7s) ", in tipc_crypto_do_cmd()
2054 tipc_node_get_id_str(rx->node)); in tipc_crypto_do_cmd()
2057 stat = per_cpu_ptr(rx->stats, cpu)->stat[i]; in tipc_crypto_do_cmd()
2112 i += scnprintf(buf + i, 200 - i, "\tPeer RX active: %d\n", in tipc_crypto_key_dump()
2156 struct tipc_crypto *rx; in tipc_crypto_msg_rcv() local
2163 rx = tipc_node_crypto_rx_by_addr(net, msg_prevnode(hdr)); in tipc_crypto_msg_rcv()
2164 if (unlikely(!rx)) in tipc_crypto_msg_rcv()
2169 if (tipc_crypto_key_rcv(rx, hdr)) in tipc_crypto_msg_rcv()
2176 tipc_node_put(rx->node); in tipc_crypto_msg_rcv()
2268 * @rx: the RX crypto
2272 * schedules a RX work to attach the key to the corresponding RX crypto.
2277 static bool tipc_crypto_key_rcv(struct tipc_crypto *rx, struct tipc_msg *hdr) in tipc_crypto_key_rcv() argument
2279 struct tipc_crypto *tx = tipc_net(rx->net)->crypto_tx; in tipc_crypto_key_rcv()
2285 spin_lock(&rx->lock); in tipc_crypto_key_rcv()
2286 if (unlikely(rx->skey || (key_gen == rx->key_gen && rx->key.keys))) { in tipc_crypto_key_rcv()
2287 pr_err("%s: key existed <%p>, gen %d vs %d\n", rx->name, in tipc_crypto_key_rcv()
2288 rx->skey, key_gen, rx->key_gen); in tipc_crypto_key_rcv()
2295 pr_err("%s: unable to allocate memory for skey\n", rx->name); in tipc_crypto_key_rcv()
2312 rx->key_gen = key_gen; in tipc_crypto_key_rcv()
2313 rx->skey_mode = msg_key_mode(hdr); in tipc_crypto_key_rcv()
2314 rx->skey = skey; in tipc_crypto_key_rcv()
2315 rx->nokey = 0; in tipc_crypto_key_rcv()
2319 spin_unlock(&rx->lock); in tipc_crypto_key_rcv()
2322 if (likely(skey && queue_delayed_work(tx->wq, &rx->work, 0))) in tipc_crypto_key_rcv()
2329 * tipc_crypto_work_rx - Scheduled RX works handler
2330 * @work: the struct RX work
2333 * or attaching a received session key on RX crypto.
2338 struct tipc_crypto *rx = container_of(dwork, struct tipc_crypto, work); in tipc_crypto_work_rx() local
2339 struct tipc_crypto *tx = tipc_net(rx->net)->crypto_tx; in tipc_crypto_work_rx()
2346 if (atomic_cmpxchg(&rx->key_distr, in tipc_crypto_work_rx()
2351 rc = tipc_crypto_key_distr(tx, key, rx->node); in tipc_crypto_work_rx()
2354 tx->name, key, tipc_node_get_id_str(rx->node), in tipc_crypto_work_rx()
2360 atomic_cmpxchg(&rx->key_distr, KEY_DISTR_COMPL, 0); in tipc_crypto_work_rx()
2364 if (rx->skey) { in tipc_crypto_work_rx()
2365 rc = tipc_crypto_key_init(rx, rx->skey, rx->skey_mode, false); in tipc_crypto_work_rx()
2368 rx->name, rc); in tipc_crypto_work_rx()
2377 kfree(rx->skey); in tipc_crypto_work_rx()
2378 rx->skey = NULL; in tipc_crypto_work_rx()
2383 if (resched && queue_delayed_work(tx->wq, &rx->work, delay)) in tipc_crypto_work_rx()
2386 tipc_node_put(rx->node); in tipc_crypto_work_rx()