Lines Matching refs:rx_sa
138 struct macsec_rx_sa *rx_sa; member
737 struct macsec_rx_sa *rx_sa = macsec_skb_cb(skb)->rx_sa; in macsec_post_decrypt() local
738 struct pcpu_rx_sc_stats *rxsc_stats = this_cpu_ptr(rx_sa->sc->stats); in macsec_post_decrypt()
742 spin_lock(&rx_sa->lock); in macsec_post_decrypt()
743 if (rx_sa->next_pn_halves.lower >= secy->replay_window) in macsec_post_decrypt()
744 lowest_pn = rx_sa->next_pn_halves.lower - secy->replay_window; in macsec_post_decrypt()
751 spin_unlock(&rx_sa->lock); in macsec_post_decrypt()
770 spin_unlock(&rx_sa->lock); in macsec_post_decrypt()
778 this_cpu_inc(rx_sa->stats->InPktsNotValid); in macsec_post_decrypt()
786 this_cpu_inc(rx_sa->stats->InPktsInvalid); in macsec_post_decrypt()
799 this_cpu_inc(rx_sa->stats->InPktsOK); in macsec_post_decrypt()
804 if (pn + 1 > rx_sa->next_pn_halves.lower) { in macsec_post_decrypt()
805 rx_sa->next_pn_halves.lower = pn + 1; in macsec_post_decrypt()
807 !pn_same_half(pn, rx_sa->next_pn_halves.lower)) { in macsec_post_decrypt()
808 rx_sa->next_pn_halves.upper++; in macsec_post_decrypt()
809 rx_sa->next_pn_halves.lower = pn + 1; in macsec_post_decrypt()
812 spin_unlock(&rx_sa->lock); in macsec_post_decrypt()
847 struct macsec_rx_sa *rx_sa = macsec_skb_cb(skb)->rx_sa; in macsec_decrypt_done() local
848 struct macsec_rx_sc *rx_sc = rx_sa->sc; in macsec_decrypt_done()
876 macsec_rxsa_put(rx_sa); in macsec_decrypt_done()
883 struct macsec_rx_sa *rx_sa, in macsec_decrypt() argument
906 req = macsec_alloc_req(rx_sa->key.tfm, &iv, &sg, ret); in macsec_decrypt()
916 pn_t recovered_pn = rx_sa->next_pn_halves; in macsec_decrypt()
919 if (hdr_pn < rx_sa->next_pn_halves.lower && in macsec_decrypt()
920 !pn_same_half(hdr_pn, rx_sa->next_pn_halves.lower)) in macsec_decrypt()
923 macsec_fill_iv_xpn(iv, rx_sa->ssci, recovered_pn.full64, in macsec_decrypt()
924 rx_sa->key.salt); in macsec_decrypt()
1132 struct macsec_rx_sa *rx_sa; in macsec_handle_frame() local
1219 rx_sa = macsec_rxsa_get(rx_sc->sa[macsec_skb_cb(skb)->assoc_num]); in macsec_handle_frame()
1220 if (!rx_sa) { in macsec_handle_frame()
1249 spin_lock(&rx_sa->lock); in macsec_handle_frame()
1250 late = rx_sa->next_pn_halves.lower >= secy->replay_window && in macsec_handle_frame()
1251 hdr_pn < (rx_sa->next_pn_halves.lower - secy->replay_window); in macsec_handle_frame()
1254 late = late && pn_same_half(rx_sa->next_pn_halves.lower, hdr_pn); in macsec_handle_frame()
1255 spin_unlock(&rx_sa->lock); in macsec_handle_frame()
1266 macsec_skb_cb(skb)->rx_sa = rx_sa; in macsec_handle_frame()
1271 skb = macsec_decrypt(skb, dev, rx_sa, sci, secy); in macsec_handle_frame()
1276 macsec_rxsa_put(rx_sa); in macsec_handle_frame()
1293 if (rx_sa) in macsec_handle_frame()
1294 macsec_rxsa_put(rx_sa); in macsec_handle_frame()
1310 macsec_rxsa_put(rx_sa); in macsec_handle_frame()
1392 static int init_rx_sa(struct macsec_rx_sa *rx_sa, char *sak, int key_len, in init_rx_sa() argument
1395 rx_sa->stats = alloc_percpu(struct macsec_rx_sa_stats); in init_rx_sa()
1396 if (!rx_sa->stats) in init_rx_sa()
1399 rx_sa->key.tfm = macsec_alloc_tfm(sak, key_len, icv_len); in init_rx_sa()
1400 if (IS_ERR(rx_sa->key.tfm)) { in init_rx_sa()
1401 free_percpu(rx_sa->stats); in init_rx_sa()
1402 return PTR_ERR(rx_sa->key.tfm); in init_rx_sa()
1405 rx_sa->ssci = MACSEC_UNDEF_SSCI; in init_rx_sa()
1406 rx_sa->active = false; in init_rx_sa()
1407 rx_sa->next_pn = 1; in init_rx_sa()
1408 refcount_set(&rx_sa->refcnt, 1); in init_rx_sa()
1409 spin_lock_init(&rx_sa->lock); in init_rx_sa()
1414 static void clear_rx_sa(struct macsec_rx_sa *rx_sa) in clear_rx_sa() argument
1416 rx_sa->active = false; in clear_rx_sa()
1418 macsec_rxsa_put(rx_sa); in clear_rx_sa()
1640 struct macsec_rx_sa *rx_sa; in get_rxsa_from_nl() local
1651 rx_sa = rtnl_dereference(rx_sc->sa[*assoc_num]); in get_rxsa_from_nl()
1652 if (!rx_sa) in get_rxsa_from_nl()
1656 return rx_sa; in get_rxsa_from_nl()
1743 struct macsec_rx_sa *rx_sa; in macsec_add_rxsa() local
1794 rx_sa = rtnl_dereference(rx_sc->sa[assoc_num]); in macsec_add_rxsa()
1795 if (rx_sa) { in macsec_add_rxsa()
1800 rx_sa = kmalloc_obj(*rx_sa); in macsec_add_rxsa()
1801 if (!rx_sa) { in macsec_add_rxsa()
1806 err = init_rx_sa(rx_sa, nla_data(tb_sa[MACSEC_SA_ATTR_KEY]), in macsec_add_rxsa()
1809 kfree(rx_sa); in macsec_add_rxsa()
1815 spin_lock_bh(&rx_sa->lock); in macsec_add_rxsa()
1816 rx_sa->next_pn = nla_get_uint(tb_sa[MACSEC_SA_ATTR_PN]); in macsec_add_rxsa()
1817 spin_unlock_bh(&rx_sa->lock); in macsec_add_rxsa()
1821 rx_sa->active = !!nla_get_u8(tb_sa[MACSEC_SA_ATTR_ACTIVE]); in macsec_add_rxsa()
1823 rx_sa->sc = rx_sc; in macsec_add_rxsa()
1826 rx_sa->ssci = nla_get_ssci(tb_sa[MACSEC_SA_ATTR_SSCI]); in macsec_add_rxsa()
1827 nla_memcpy(rx_sa->key.salt.bytes, tb_sa[MACSEC_SA_ATTR_SALT], in macsec_add_rxsa()
1843 ctx.sa.rx_sa = rx_sa; in macsec_add_rxsa()
1854 nla_memcpy(rx_sa->key.id, tb_sa[MACSEC_SA_ATTR_KEYID], MACSEC_KEYID_LEN); in macsec_add_rxsa()
1855 rcu_assign_pointer(rx_sc->sa[assoc_num], rx_sa); in macsec_add_rxsa()
1862 macsec_rxsa_put(rx_sa); in macsec_add_rxsa()
2082 struct macsec_rx_sa *rx_sa; in macsec_del_rxsa() local
2098 rx_sa = get_rxsa_from_nl(genl_info_net(info), attrs, tb_rxsc, tb_sa, in macsec_del_rxsa()
2100 if (IS_ERR(rx_sa)) { in macsec_del_rxsa()
2102 return PTR_ERR(rx_sa); in macsec_del_rxsa()
2105 if (rx_sa->active) { in macsec_del_rxsa()
2122 ctx.sa.rx_sa = rx_sa; in macsec_del_rxsa()
2131 clear_rx_sa(rx_sa); in macsec_del_rxsa()
2379 struct macsec_rx_sa *rx_sa; in macsec_upd_rxsa() local
2402 rx_sa = get_rxsa_from_nl(genl_info_net(info), attrs, tb_rxsc, tb_sa, in macsec_upd_rxsa()
2404 if (IS_ERR(rx_sa)) { in macsec_upd_rxsa()
2406 return PTR_ERR(rx_sa); in macsec_upd_rxsa()
2420 spin_lock_bh(&rx_sa->lock); in macsec_upd_rxsa()
2421 prev_pn = rx_sa->next_pn_halves; in macsec_upd_rxsa()
2422 rx_sa->next_pn = nla_get_uint(tb_sa[MACSEC_SA_ATTR_PN]); in macsec_upd_rxsa()
2423 spin_unlock_bh(&rx_sa->lock); in macsec_upd_rxsa()
2426 was_active = rx_sa->active; in macsec_upd_rxsa()
2428 rx_sa->active = nla_get_u8(tb_sa[MACSEC_SA_ATTR_ACTIVE]); in macsec_upd_rxsa()
2442 ctx.sa.rx_sa = rx_sa; in macsec_upd_rxsa()
2456 spin_lock_bh(&rx_sa->lock); in macsec_upd_rxsa()
2457 rx_sa->next_pn_halves = prev_pn; in macsec_upd_rxsa()
2458 spin_unlock_bh(&rx_sa->lock); in macsec_upd_rxsa()
2460 rx_sa->active = was_active; in macsec_upd_rxsa()
2727 struct macsec_rx_sa *rx_sa, in get_rx_sa_stats() argument
2741 ctx.sa.rx_sa = rx_sa; in get_rx_sa_stats()
2752 per_cpu_ptr(rx_sa->stats, cpu); in get_rx_sa_stats()
3203 struct macsec_rx_sa *rx_sa = rtnl_dereference(rx_sc->sa[i]); in dump_secy() local
3208 if (!rx_sa) in dump_secy()
3228 get_rx_sa_stats(dev, rx_sc, i, rx_sa, &rx_sa_stats); in dump_secy()
3239 pn = rx_sa->next_pn; in dump_secy()
3242 pn = rx_sa->next_pn_halves.lower; in dump_secy()
3248 nla_put(skb, MACSEC_SA_ATTR_KEYID, MACSEC_KEYID_LEN, rx_sa->key.id) || in dump_secy()
3249 (secy->xpn && nla_put_ssci(skb, MACSEC_SA_ATTR_SSCI, rx_sa->ssci)) || in dump_secy()
3250 nla_put_u8(skb, MACSEC_SA_ATTR_ACTIVE, rx_sa->active)) { in dump_secy()