Lines Matching refs:dtab
105 static inline struct hlist_head *dev_map_index_hash(struct bpf_dtab *dtab, in dev_map_index_hash() argument
108 return &dtab->dev_index_head[idx & (dtab->n_buckets - 1)]; in dev_map_index_hash()
136 static int dev_map_init_map(struct bpf_dtab *dtab, union bpf_attr *attr) in dev_map_init_map() argument
142 bpf_map_init_from_attr(&dtab->map, attr); in dev_map_init_map()
146 dtab->n_buckets = roundup_pow_of_two(dtab->map.max_entries); in dev_map_init_map()
147 dtab->dev_index_head = dev_map_create_hash(dtab->n_buckets, in dev_map_init_map()
148 dtab->map.numa_node); in dev_map_init_map()
149 if (!dtab->dev_index_head) in dev_map_init_map()
152 spin_lock_init(&dtab->index_lock); in dev_map_init_map()
154 dtab->netdev_map = bpf_map_area_alloc((u64) dtab->map.max_entries * in dev_map_init_map()
156 dtab->map.numa_node); in dev_map_init_map()
157 if (!dtab->netdev_map) in dev_map_init_map()
166 struct bpf_dtab *dtab; in dev_map_alloc() local
169 dtab = bpf_map_area_alloc(sizeof(*dtab), NUMA_NO_NODE); in dev_map_alloc()
170 if (!dtab) in dev_map_alloc()
173 err = dev_map_init_map(dtab, attr); in dev_map_alloc()
175 bpf_map_area_free(dtab); in dev_map_alloc()
180 list_add_tail_rcu(&dtab->list, &dev_map_list); in dev_map_alloc()
183 return &dtab->map; in dev_map_alloc()
188 struct bpf_dtab *dtab = container_of(map, struct bpf_dtab, map); in dev_map_free() local
202 list_del_rcu(&dtab->list); in dev_map_free()
218 if (dtab->map.map_type == BPF_MAP_TYPE_DEVMAP_HASH) { in dev_map_free()
219 for (i = 0; i < dtab->n_buckets; i++) { in dev_map_free()
224 head = dev_map_index_hash(dtab, i); in dev_map_free()
235 bpf_map_area_free(dtab->dev_index_head); in dev_map_free()
237 for (i = 0; i < dtab->map.max_entries; i++) { in dev_map_free()
240 dev = rcu_dereference_raw(dtab->netdev_map[i]); in dev_map_free()
250 bpf_map_area_free(dtab->netdev_map); in dev_map_free()
253 bpf_map_area_free(dtab); in dev_map_free()
258 struct bpf_dtab *dtab = container_of(map, struct bpf_dtab, map); in dev_map_get_next_key() local
262 if (index >= dtab->map.max_entries) { in dev_map_get_next_key()
267 if (index == dtab->map.max_entries - 1) in dev_map_get_next_key()
279 struct bpf_dtab *dtab = container_of(map, struct bpf_dtab, map); in __dev_map_hash_lookup_elem() local
280 struct hlist_head *head = dev_map_index_hash(dtab, key); in __dev_map_hash_lookup_elem()
284 lockdep_is_held(&dtab->index_lock)) in __dev_map_hash_lookup_elem()
294 struct bpf_dtab *dtab = container_of(map, struct bpf_dtab, map); in dev_map_hash_get_next_key() local
317 i = idx & (dtab->n_buckets - 1); in dev_map_hash_get_next_key()
321 for (; i < dtab->n_buckets; i++) { in dev_map_hash_get_next_key()
322 head = dev_map_index_hash(dtab, i); in dev_map_hash_get_next_key()
447 struct bpf_dtab *dtab = container_of(map, struct bpf_dtab, map); in __dev_map_lookup_elem() local
453 obj = rcu_dereference_check(dtab->netdev_map[key], in __dev_map_lookup_elem()
625 struct bpf_dtab *dtab = container_of(map, struct bpf_dtab, map); in dev_map_enqueue_multi() local
644 dst = rcu_dereference_check(dtab->netdev_map[i], in dev_map_enqueue_multi()
665 for (i = 0; i < dtab->n_buckets; i++) { in dev_map_enqueue_multi()
666 head = dev_map_index_hash(dtab, i); in dev_map_enqueue_multi()
668 lockdep_is_held(&dtab->index_lock)) { in dev_map_enqueue_multi()
746 struct bpf_dtab *dtab = container_of(map, struct bpf_dtab, map); in dev_map_redirect_multi() local
766 dst = rcu_dereference_check(dtab->netdev_map[i], in dev_map_redirect_multi()
788 for (i = 0; i < dtab->n_buckets; i++) { in dev_map_redirect_multi()
789 head = dev_map_index_hash(dtab, i); in dev_map_redirect_multi()
846 struct bpf_dtab *dtab = container_of(map, struct bpf_dtab, map); in dev_map_delete_elem() local
853 old_dev = unrcu_pointer(xchg(&dtab->netdev_map[k], NULL)); in dev_map_delete_elem()
856 atomic_dec((atomic_t *)&dtab->items); in dev_map_delete_elem()
863 struct bpf_dtab *dtab = container_of(map, struct bpf_dtab, map); in dev_map_hash_delete_elem() local
869 spin_lock_irqsave(&dtab->index_lock, flags); in dev_map_hash_delete_elem()
873 dtab->items--; in dev_map_hash_delete_elem()
878 spin_unlock_irqrestore(&dtab->index_lock, flags); in dev_map_hash_delete_elem()
884 struct bpf_dtab *dtab, in __dev_map_alloc_node() argument
891 dev = bpf_map_kmalloc_node(&dtab->map, sizeof(*dev), in __dev_map_alloc_node()
893 dtab->map.numa_node); in __dev_map_alloc_node()
907 !bpf_prog_map_compatible(&dtab->map, prog)) in __dev_map_alloc_node()
934 struct bpf_dtab *dtab = container_of(map, struct bpf_dtab, map); in __dev_map_update_elem() local
941 if (unlikely(i >= dtab->map.max_entries)) in __dev_map_update_elem()
955 dev = __dev_map_alloc_node(net, dtab, &val, i); in __dev_map_update_elem()
964 old_dev = unrcu_pointer(xchg(&dtab->netdev_map[i], RCU_INITIALIZER(dev))); in __dev_map_update_elem()
968 atomic_inc((atomic_t *)&dtab->items); in __dev_map_update_elem()
983 struct bpf_dtab *dtab = container_of(map, struct bpf_dtab, map); in __dev_map_hash_update_elem() local
996 spin_lock_irqsave(&dtab->index_lock, flags); in __dev_map_hash_update_elem()
1002 dev = __dev_map_alloc_node(net, dtab, &val, idx); in __dev_map_hash_update_elem()
1011 if (dtab->items >= dtab->map.max_entries) { in __dev_map_hash_update_elem()
1012 spin_unlock_irqrestore(&dtab->index_lock, flags); in __dev_map_hash_update_elem()
1016 dtab->items++; in __dev_map_hash_update_elem()
1020 dev_map_index_hash(dtab, idx)); in __dev_map_hash_update_elem()
1021 spin_unlock_irqrestore(&dtab->index_lock, flags); in __dev_map_hash_update_elem()
1029 spin_unlock_irqrestore(&dtab->index_lock, flags); in __dev_map_hash_update_elem()
1056 struct bpf_dtab *dtab = container_of(map, struct bpf_dtab, map); in dev_map_mem_usage() local
1060 usage += (u64)dtab->n_buckets * sizeof(struct hlist_head); in dev_map_mem_usage()
1063 usage += atomic_read((atomic_t *)&dtab->items) * in dev_map_mem_usage()
1099 static void dev_map_hash_remove_netdev(struct bpf_dtab *dtab, in dev_map_hash_remove_netdev() argument
1105 spin_lock_irqsave(&dtab->index_lock, flags); in dev_map_hash_remove_netdev()
1106 for (i = 0; i < dtab->n_buckets; i++) { in dev_map_hash_remove_netdev()
1111 head = dev_map_index_hash(dtab, i); in dev_map_hash_remove_netdev()
1117 dtab->items--; in dev_map_hash_remove_netdev()
1122 spin_unlock_irqrestore(&dtab->index_lock, flags); in dev_map_hash_remove_netdev()
1129 struct bpf_dtab *dtab; in dev_map_notification() local
1157 list_for_each_entry_rcu(dtab, &dev_map_list, list) { in dev_map_notification()
1158 if (dtab->map.map_type == BPF_MAP_TYPE_DEVMAP_HASH) { in dev_map_notification()
1159 dev_map_hash_remove_netdev(dtab, netdev); in dev_map_notification()
1163 for (i = 0; i < dtab->map.max_entries; i++) { in dev_map_notification()
1166 dev = rcu_dereference(dtab->netdev_map[i]); in dev_map_notification()
1169 odev = unrcu_pointer(cmpxchg(&dtab->netdev_map[i], RCU_INITIALIZER(dev), NULL)); in dev_map_notification()
1173 atomic_dec((atomic_t *)&dtab->items); in dev_map_notification()