/linux/net/sched/ |
H A D | cls_route.c | 129 struct route4_head *head = rcu_dereference_bh(tp->root); in route4_classify() 164 b = rcu_dereference_bh(head->table[h]); in route4_classify() 166 for (f = rcu_dereference_bh(b->ht[route4_hash_from(id)]); in route4_classify() 168 f = rcu_dereference_bh(f->next)) in route4_classify() 172 for (f = rcu_dereference_bh(b->ht[route4_hash_iif(iif)]); in route4_classify() 174 f = rcu_dereference_bh(f->next)) in route4_classify() 178 for (f = rcu_dereference_bh(b->ht[route4_hash_wild()]); in route4_classify() 180 f = rcu_dereference_bh(f->next)) in route4_classify()
|
H A D | cls_fw.c | 55 struct fw_head *head = rcu_dereference_bh(tp->root); in fw_classify() 63 for (f = rcu_dereference_bh(head->ht[fw_hash(id)]); f; in fw_classify() 64 f = rcu_dereference_bh(f->next)) { in fw_classify()
|
H A D | cls_cgroup.c | 30 struct cls_cgroup_head *head = rcu_dereference_bh(tp->root); in cls_cgroup_classify()
|
H A D | act_connmark.c | 47 parms = rcu_dereference_bh(ca->parms); in tcf_connmark_act()
|
H A D | act_skbmod.c | 42 p = rcu_dereference_bh(d->skbmod_p); in tcf_skbmod_act()
|
H A D | act_nat.c | 135 parms = rcu_dereference_bh(p->parms); in tcf_nat_act()
|
H A D | cls_basic.c | 45 struct basic_head *head = rcu_dereference_bh(tp->root); in basic_classify()
|
H A D | act_sample.c | 179 psample_group = rcu_dereference_bh(s->psample_group); in tcf_sample_act()
|
H A D | sch_multiq.c | 35 struct tcf_proto *fl = rcu_dereference_bh(q->filter_list); in multiq_classify()
|
/linux/include/linux/ |
H A D | rculist.h | 813 for (pos = hlist_entry_safe(rcu_dereference_bh(hlist_first_rcu(head)),\ 816 pos = hlist_entry_safe(rcu_dereference_bh(hlist_next_rcu(\ 837 for (pos = hlist_entry_safe(rcu_dereference_bh(hlist_next_rcu( \ 840 pos = hlist_entry_safe(rcu_dereference_bh(hlist_next_rcu( \
|
/linux/net/netfilter/ipset/ |
H A D | ip_set_hash_gen.h | 816 t = rcu_dereference_bh(h->table); in mtype_ext_size() 820 n = rcu_dereference_bh(hbucket(t, i)); in mtype_ext_size() 853 t = rcu_dereference_bh(h->table); in mtype_add() 876 n = rcu_dereference_bh(hbucket(t, key)); in mtype_add() 1051 t = rcu_dereference_bh(h->table); in mtype_del() 1058 n = rcu_dereference_bh(hbucket(t, key)); in mtype_del() 1164 struct htable *t = rcu_dereference_bh(h->table); in mtype_test_cidrs() 1188 n = rcu_dereference_bh(hbucket(t, key)); in mtype_test_cidrs() 1227 t = rcu_dereference_bh(h->table); in mtype_test() 1242 n = rcu_dereference_bh(hbucket(t, key)); in mtype_test() [all …]
|
/linux/drivers/net/hyperv/ |
H A D | netvsc_bpf.c | 259 nvsc_dev = rcu_dereference_bh(ndev_ctx->nvdev); in netvsc_ndoxdp_xmit() 267 vf_netdev = rcu_dereference_bh(ndev_ctx->vf_netdev); in netvsc_ndoxdp_xmit()
|
/linux/drivers/net/wireguard/ |
H A D | allowedips.c | 121 node = rcu_dereference_bh(node->bit[choose(node, key)]); in find_node() 139 node = find_node(rcu_dereference_bh(root), bits, ip); in lookup() 141 peer = wg_peer_get_maybe_zero(rcu_dereference_bh(node->peer)); in lookup()
|
H A D | send.c | 130 keypair = rcu_dereference_bh(peer->keypairs.current_keypair); in keep_key_fresh() 359 rcu_dereference_bh(peer->keypairs.current_keypair)); in wg_packet_send_staged_packets()
|
H A D | socket.c | 39 sock = rcu_dereference_bh(wg->sock4); in send4() 119 sock = rcu_dereference_bh(wg->sock6); in send6()
|
/linux/drivers/net/ethernet/aquantia/atlantic/ |
H A D | aq_macsec.c | 1074 tx_sa = rcu_dereference_bh(secy->tx_sc.sa[ctx->sa.assoc_num]); in aq_mdo_get_tx_sa_stats() 1160 rx_sa = rcu_dereference_bh(aq_rxsc->sw_rxsc->sa[ctx->sa.assoc_num]); in aq_mdo_get_rx_sa_stats() 1187 tx_sa = rcu_dereference_bh(secy->tx_sc.sa[i]); in apply_txsc_cfg() 1215 rx_sa = rcu_dereference_bh(aq_rxsc->sw_rxsc->sa[i]); in apply_rxsc_cfg() 1242 for (rx_sc = rcu_dereference_bh(secy->rx_sc); rx_sc; in aq_clear_secy() 1243 rx_sc = rcu_dereference_bh(rx_sc->next)) { in aq_clear_secy() 1268 for (rx_sc = rcu_dereference_bh(secy->rx_sc); rx_sc && rx_sc->active; in aq_apply_secy_cfg() 1269 rx_sc = rcu_dereference_bh(rx_sc->next)) { in aq_apply_secy_cfg() 1388 tx_sa = rcu_dereference_bh(secy->tx_sc.sa[an]); in aq_check_txsa_expiration()
|
/linux/drivers/net/team/ |
H A D | team_mode_activebackup.c | 40 active_port = rcu_dereference_bh(ab_priv(team)->active_port); in ab_transmit()
|
H A D | team_mode_loadbalance.c | 133 port = rcu_dereference_bh(LB_HTPM_PORT_BY_HASH(lb_priv, hash)); in lb_htpm_select_tx_port() 192 fp = rcu_dereference_bh(lb_priv->fp); in lb_get_skb_hash() 226 select_tx_port_func = rcu_dereference_bh(lb_priv->select_tx_port_func); in lb_transmit()
|
/linux/drivers/net/ipvlan/ |
H A D | ipvlan.h | 116 return rcu_dereference_bh(d->rx_handler_data); in ipvlan_port_get_rcu_bh()
|
/linux/drivers/net/ethernet/mellanox/mlx5/core/en/ |
H A D | selq.c | 200 selq = rcu_dereference_bh(priv->selq.active); in mlx5e_select_queue()
|
/linux/drivers/net/ |
H A D | net_failover.c | 104 xmit_dev = rcu_dereference_bh(nfo_info->primary_dev); in net_failover_start_xmit() 106 xmit_dev = rcu_dereference_bh(nfo_info->standby_dev); in net_failover_start_xmit()
|
/linux/net/ipv4/ |
H A D | tcp_sigpool.c | 288 c->scratch = rcu_dereference_bh(*this_cpu_ptr(&sigpool_scratch.pad)); in tcp_sigpool_start()
|
/linux/net/core/ |
H A D | netpoll.c | 185 struct netpoll_info *ni = rcu_dereference_bh(dev->npinfo); in netpoll_poll_dev() 332 npinfo = rcu_dereference_bh(dev->npinfo); in __netpoll_send_skb()
|
/linux/drivers/net/ethernet/sfc/siena/ |
H A D | rx.c | 263 xdp_prog = rcu_dereference_bh(efx->xdp_prog); in efx_do_xdp()
|
/linux/drivers/net/ethernet/qualcomm/rmnet/ |
H A D | rmnet_config.c | 414 return rcu_dereference_bh(real_dev->rx_handler_data); in rmnet_get_port_rcu()
|