Lines Matching full:rx
76 STAT_BADMSGS = STAT_BADKEYS, /* rx only */
109 passive:2, /* rx only */
113 passive:2, /* rx only */
137 * @users: the number of the key users (TX/RX)
175 * struct tipc_crypto - TIPC TX/RX crypto structure
177 * @node: TIPC node (RX)
179 * @peer_rx_active: replicated peer RX active key index
180 * @key_gen: TX/RX key generation
185 * @work: delayed work sched for TX/RX
241 /* struct tipc_crypto_rx_ctx - RX context for callbacks */
281 static bool tipc_crypto_key_try_align(struct tipc_crypto *rx, u8 new_pending);
283 struct tipc_crypto *rx,
286 static void tipc_crypto_key_synch(struct tipc_crypto *rx, struct sk_buff *skb);
301 static bool tipc_crypto_key_rcv(struct tipc_crypto *rx, struct tipc_msg *hdr);
900 pr_err("RX: skb_cow_data() returned %d\n", nsg); in tipc_aead_decrypt()
915 pr_err("RX: skb_to_sgvec() returned %d, nsg %d\n", rc, nsg); in tipc_aead_decrypt()
1030 * @__rx: RX crypto handle if dest is "known"
1208 struct tipc_crypto *tx, *rx; in tipc_crypto_key_flush() local
1214 rx = c; in tipc_crypto_key_flush()
1215 tx = tipc_net(rx->net)->crypto_tx; in tipc_crypto_key_flush()
1216 if (cancel_delayed_work(&rx->work)) { in tipc_crypto_key_flush()
1217 kfree(rx->skey); in tipc_crypto_key_flush()
1218 rx->skey = NULL; in tipc_crypto_key_flush()
1219 atomic_xchg(&rx->key_distr, 0); in tipc_crypto_key_flush()
1220 tipc_node_put(rx->node); in tipc_crypto_key_flush()
1222 /* RX stopping => decrease TX key users if any */ in tipc_crypto_key_flush()
1223 k = atomic_xchg(&rx->peer_rx_active, 0); in tipc_crypto_key_flush()
1240 * tipc_crypto_key_try_align - Align RX keys if possible
1241 * @rx: RX crypto handle
1252 static bool tipc_crypto_key_try_align(struct tipc_crypto *rx, u8 new_pending) in tipc_crypto_key_try_align() argument
1260 spin_lock(&rx->lock); in tipc_crypto_key_try_align()
1261 key = rx->key; in tipc_crypto_key_try_align()
1270 if (tipc_aead_users(rx->aead[key.pending]) > 0) in tipc_crypto_key_try_align()
1274 tmp1 = tipc_aead_rcu_ptr(rx->aead[key.pending], &rx->lock); in tipc_crypto_key_try_align()
1277 rcu_assign_pointer(rx->aead[key.pending], NULL); in tipc_crypto_key_try_align()
1281 tmp2 = rcu_replace_pointer(rx->aead[key.passive], tmp2, lockdep_is_held(&rx->lock)); in tipc_crypto_key_try_align()
1287 tipc_crypto_key_set_state(rx, new_passive, 0, new_pending); in tipc_crypto_key_try_align()
1288 rcu_assign_pointer(rx->aead[new_pending], tmp1); in tipc_crypto_key_try_align()
1290 rcu_assign_pointer(rx->aead[new_passive], tmp2); in tipc_crypto_key_try_align()
1293 pr_info_ratelimited("%s: key[%d] -> key[%d]\n", rx->name, key.pending, in tipc_crypto_key_try_align()
1297 spin_unlock(&rx->lock); in tipc_crypto_key_try_align()
1304 * @rx: RX crypto handle (can be NULL)
1315 struct tipc_crypto *rx, in tipc_crypto_key_pick_tx() argument
1330 skb_cb->tx_clone_ctx.rx = rx; in tipc_crypto_key_pick_tx()
1372 * @rx: RX crypto handle
1375 * This function updates the peer node related data as the peer RX active key
1385 static void tipc_crypto_key_synch(struct tipc_crypto *rx, struct sk_buff *skb) in tipc_crypto_key_synch() argument
1388 struct tipc_crypto *tx = tipc_net(rx->net)->crypto_tx; in tipc_crypto_key_synch()
1390 u32 self = tipc_own_addr(rx->net); in tipc_crypto_key_synch()
1394 /* Update RX 'key_master' flag according to peer, also mark "legacy" if in tipc_crypto_key_synch()
1397 rx->key_master = ehdr->master_key; in tipc_crypto_key_synch()
1398 if (!rx->key_master) in tipc_crypto_key_synch()
1411 !atomic_cmpxchg(&rx->key_distr, 0, KEY_DISTR_SCHED)) { in tipc_crypto_key_synch()
1415 if (queue_delayed_work(tx->wq, &rx->work, delay)) in tipc_crypto_key_synch()
1416 tipc_node_get(rx->node); in tipc_crypto_key_synch()
1420 atomic_xchg(&rx->key_distr, 0); in tipc_crypto_key_synch()
1423 /* Case 2: Peer RX active key has changed, let's update own TX users */ in tipc_crypto_key_synch()
1424 cur = atomic_read(&rx->peer_rx_active); in tipc_crypto_key_synch()
1428 atomic_cmpxchg(&rx->peer_rx_active, cur, new) == cur) { in tipc_crypto_key_synch()
1434 atomic64_set(&rx->sndnxt, 0); in tipc_crypto_key_synch()
1439 tx->name, cur, new, rx->name); in tipc_crypto_key_synch()
1504 scnprintf(c->name, 48, "%s(%s)", (is_rx(c)) ? "RX" : "TX", in tipc_crypto_start()
1546 void tipc_crypto_timeout(struct tipc_crypto *rx) in tipc_crypto_timeout() argument
1548 struct tipc_net *tn = tipc_net(rx->net); in tipc_crypto_timeout()
1572 /* RX pending: having user -> active */ in tipc_crypto_timeout()
1573 spin_lock(&rx->lock); in tipc_crypto_timeout()
1574 key = rx->key; in tipc_crypto_timeout()
1575 if (!key.pending || tipc_aead_users(rx->aead[key.pending]) <= 0) in tipc_crypto_timeout()
1581 rx->timer2 = jiffies; in tipc_crypto_timeout()
1582 tipc_crypto_key_set_state(rx, key.passive, key.active, 0); in tipc_crypto_timeout()
1583 this_cpu_inc(rx->stats->stat[STAT_SWITCHES]); in tipc_crypto_timeout()
1584 pr_info("%s: key[%d] is activated\n", rx->name, key.pending); in tipc_crypto_timeout()
1588 /* RX pending: not working -> remove */ in tipc_crypto_timeout()
1589 if (!key.pending || tipc_aead_users(rx->aead[key.pending]) > -10) in tipc_crypto_timeout()
1592 tipc_crypto_key_set_state(rx, key.passive, key.active, 0); in tipc_crypto_timeout()
1593 tipc_crypto_key_detach(rx->aead[key.pending], &rx->lock); in tipc_crypto_timeout()
1594 pr_debug("%s: key[%d] is removed\n", rx->name, key.pending); in tipc_crypto_timeout()
1598 /* RX active: timed out or no user -> pending */ in tipc_crypto_timeout()
1601 if (time_before(jiffies, rx->timer1 + TIPC_RX_ACTIVE_LIM) && in tipc_crypto_timeout()
1602 tipc_aead_users(rx->aead[key.active]) > 0) in tipc_crypto_timeout()
1609 rx->timer2 = jiffies; in tipc_crypto_timeout()
1610 tipc_crypto_key_set_state(rx, key.passive, 0, key.pending); in tipc_crypto_timeout()
1611 tipc_aead_users_set(rx->aead[key.pending], 0); in tipc_crypto_timeout()
1612 pr_debug("%s: key[%d] is deactivated\n", rx->name, key.active); in tipc_crypto_timeout()
1616 /* RX passive: outdated or not working -> free */ in tipc_crypto_timeout()
1619 if (time_before(jiffies, rx->timer2 + TIPC_RX_PASSIVE_LIM) && in tipc_crypto_timeout()
1620 tipc_aead_users(rx->aead[key.passive]) > -10) in tipc_crypto_timeout()
1623 tipc_crypto_key_set_state(rx, 0, key.active, key.pending); in tipc_crypto_timeout()
1624 tipc_crypto_key_detach(rx->aead[key.passive], &rx->lock); in tipc_crypto_timeout()
1625 pr_debug("%s: key[%d] is freed\n", rx->name, key.passive); in tipc_crypto_timeout()
1628 spin_unlock(&rx->lock); in tipc_crypto_timeout()
1642 tipc_crypto_do_cmd(rx->net, cmd); in tipc_crypto_timeout()
1790 * @rx: RX crypto handle
1798 * Note: RX key(s) can be re-aligned, or in case of no key suitable, TX
1809 int tipc_crypto_rcv(struct net *net, struct tipc_crypto *rx, in tipc_crypto_rcv() argument
1824 if (unlikely(!rx || tx_key == KEY_MASTER)) in tipc_crypto_rcv()
1827 /* Pick RX key according to TX key if any */ in tipc_crypto_rcv()
1828 key = rx->key; in tipc_crypto_rcv()
1833 /* Unknown key, let's try to align RX key(s) */ in tipc_crypto_rcv()
1834 if (tipc_crypto_key_try_align(rx, tx_key)) in tipc_crypto_rcv()
1839 aead = tipc_crypto_key_pick_tx(tx, rx, *skb, tx_key); in tipc_crypto_rcv()
1847 aead = tipc_aead_get(rx->aead[tx_key]); in tipc_crypto_rcv()
1852 stats = ((rx) ?: tx)->stats; in tipc_crypto_rcv()
1867 if (rx) { in tipc_crypto_rcv()
1868 /* Mark rx->nokey only if we dont have a in tipc_crypto_rcv()
1873 rx->nokey = !(rx->skey || in tipc_crypto_rcv()
1874 rcu_access_pointer(rx->aead[n])); in tipc_crypto_rcv()
1876 rx->name, rx->nokey, in tipc_crypto_rcv()
1877 tx_key, rx->key.keys); in tipc_crypto_rcv()
1878 tipc_node_put(rx->node); in tipc_crypto_rcv()
1897 struct tipc_crypto *rx = aead->crypto; in tipc_crypto_rcv_complete() local
1904 rx = skb_cb->tx_clone_ctx.rx; in tipc_crypto_rcv_complete()
1905 pr_debug("TX->RX(%s): err %d, aead %p, skb->next %p, flags %x\n", in tipc_crypto_rcv_complete()
1906 (rx) ? tipc_node_get_id_str(rx->node) : "-", err, aead, in tipc_crypto_rcv_complete()
1923 if (!rx) { in tipc_crypto_rcv_complete()
1927 rx = tipc_node_crypto_rx(n); in tipc_crypto_rcv_complete()
1928 if (unlikely(!rx)) in tipc_crypto_rcv_complete()
1938 if (tipc_crypto_key_attach(rx, tmp, ehdr->tx_key, false) < 0) { in tipc_crypto_rcv_complete()
1951 /* Set the RX key's user */ in tipc_crypto_rcv_complete()
1954 /* Mark this point, RX works */ in tipc_crypto_rcv_complete()
1955 rx->timer1 = jiffies; in tipc_crypto_rcv_complete()
1961 /* Mark this point, RX passive still works */ in tipc_crypto_rcv_complete()
1962 if (rx->key.passive && ehdr->tx_key == rx->key.passive) in tipc_crypto_rcv_complete()
1963 rx->timer2 = jiffies; in tipc_crypto_rcv_complete()
1976 tipc_crypto_key_synch(rx, *skb); in tipc_crypto_rcv_complete()
1994 if (rx) in tipc_crypto_rcv_complete()
1995 tipc_node_put(rx->node); in tipc_crypto_rcv_complete()
2001 struct tipc_crypto *tx = tn->crypto_tx, *rx; in tipc_crypto_do_cmd() local
2026 rx = tipc_node_crypto_rx_by_list(p); in tipc_crypto_do_cmd()
2027 pr_info("RX(%7.7s)\n%s", tipc_node_get_id_str(rx->node), in tipc_crypto_do_cmd()
2028 tipc_crypto_key_dump(rx, buf)); in tipc_crypto_do_cmd()
2053 rx = tipc_node_crypto_rx_by_list(p); in tipc_crypto_do_cmd()
2054 j = scnprintf(buf, 200, "RX(%7.7s) ", in tipc_crypto_do_cmd()
2055 tipc_node_get_id_str(rx->node)); in tipc_crypto_do_cmd()
2058 stat = per_cpu_ptr(rx->stats, cpu)->stat[i]; in tipc_crypto_do_cmd()
2113 i += scnprintf(buf + i, 200 - i, "\tPeer RX active: %d\n", in tipc_crypto_key_dump()
2157 struct tipc_crypto *rx; in tipc_crypto_msg_rcv() local
2164 rx = tipc_node_crypto_rx_by_addr(net, msg_prevnode(hdr)); in tipc_crypto_msg_rcv()
2165 if (unlikely(!rx)) in tipc_crypto_msg_rcv()
2170 if (tipc_crypto_key_rcv(rx, hdr)) in tipc_crypto_msg_rcv()
2177 tipc_node_put(rx->node); in tipc_crypto_msg_rcv()
2269 * @rx: the RX crypto
2273 * schedules a RX work to attach the key to the corresponding RX crypto.
2278 static bool tipc_crypto_key_rcv(struct tipc_crypto *rx, struct tipc_msg *hdr) in tipc_crypto_key_rcv() argument
2280 struct tipc_crypto *tx = tipc_net(rx->net)->crypto_tx; in tipc_crypto_key_rcv()
2289 pr_debug("%s: message data size is too small\n", rx->name); in tipc_crypto_key_rcv()
2298 pr_debug("%s: invalid MSG_CRYPTO key size\n", rx->name); in tipc_crypto_key_rcv()
2302 spin_lock(&rx->lock); in tipc_crypto_key_rcv()
2303 if (unlikely(rx->skey || (key_gen == rx->key_gen && rx->key.keys))) { in tipc_crypto_key_rcv()
2304 pr_err("%s: key existed <%p>, gen %d vs %d\n", rx->name, in tipc_crypto_key_rcv()
2305 rx->skey, key_gen, rx->key_gen); in tipc_crypto_key_rcv()
2312 pr_err("%s: unable to allocate memory for skey\n", rx->name); in tipc_crypto_key_rcv()
2322 rx->key_gen = key_gen; in tipc_crypto_key_rcv()
2323 rx->skey_mode = msg_key_mode(hdr); in tipc_crypto_key_rcv()
2324 rx->skey = skey; in tipc_crypto_key_rcv()
2325 rx->nokey = 0; in tipc_crypto_key_rcv()
2329 spin_unlock(&rx->lock); in tipc_crypto_key_rcv()
2333 if (likely(skey && queue_delayed_work(tx->wq, &rx->work, 0))) in tipc_crypto_key_rcv()
2340 * tipc_crypto_work_rx - Scheduled RX works handler
2341 * @work: the struct RX work
2344 * or attaching a received session key on RX crypto.
2349 struct tipc_crypto *rx = container_of(dwork, struct tipc_crypto, work); in tipc_crypto_work_rx() local
2350 struct tipc_crypto *tx = tipc_net(rx->net)->crypto_tx; in tipc_crypto_work_rx()
2357 if (atomic_cmpxchg(&rx->key_distr, in tipc_crypto_work_rx()
2362 rc = tipc_crypto_key_distr(tx, key, rx->node); in tipc_crypto_work_rx()
2365 tx->name, key, tipc_node_get_id_str(rx->node), in tipc_crypto_work_rx()
2371 atomic_cmpxchg(&rx->key_distr, KEY_DISTR_COMPL, 0); in tipc_crypto_work_rx()
2375 if (rx->skey) { in tipc_crypto_work_rx()
2376 rc = tipc_crypto_key_init(rx, rx->skey, rx->skey_mode, false); in tipc_crypto_work_rx()
2379 rx->name, rc); in tipc_crypto_work_rx()
2388 kfree(rx->skey); in tipc_crypto_work_rx()
2389 rx->skey = NULL; in tipc_crypto_work_rx()
2394 if (resched && queue_delayed_work(tx->wq, &rx->work, delay)) in tipc_crypto_work_rx()
2397 tipc_node_put(rx->node); in tipc_crypto_work_rx()