| /linux/include/linux/ |
| H A D | hashtable.h | 126 #define hash_for_each(name, bkt, obj, member) \ argument 127 for ((bkt) = 0, obj = NULL; obj == NULL && (bkt) < HASH_SIZE(name);\ 128 (bkt)++)\ 129 hlist_for_each_entry(obj, &name[bkt], member) 138 #define hash_for_each_rcu(name, bkt, obj, member) \ argument 139 for ((bkt) = 0, obj = NULL; obj == NULL && (bkt) < HASH_SIZE(name);\ 140 (bkt)++)\ 141 hlist_for_each_entry_rcu(obj, &name[bkt], member) 152 #define hash_for_each_safe(name, bkt, tmp, obj, member) \ argument 153 for ((bkt) = 0, obj = NULL; obj == NULL && (bkt) < HASH_SIZE(name);\ [all …]
|
| H A D | rhashtable.h | 327 struct rhash_lock_head __rcu **bkt) in rht_lock() argument 332 bit_spin_lock(0, (unsigned long *)bkt); in rht_lock() 350 struct rhash_lock_head __rcu **bkt, in rht_unlock() argument 354 bit_spin_unlock(0, (unsigned long *)bkt); in rht_unlock() 364 struct rhash_lock_head *p, struct rhash_lock_head __rcu *const *bkt, in __rht_ptr() argument 373 (likely(p_val) ? p_val : (unsigned long)RHT_NULLS_MARKER(bkt)); in __rht_ptr() 376 (p_val ?: (unsigned long)RHT_NULLS_MARKER(bkt)); in __rht_ptr() 387 struct rhash_lock_head __rcu *const *bkt, in __rht_ptr_rcu() argument 390 return __rht_ptr(rcu_dereference_all(*bkt), bkt, freq); in __rht_ptr_rcu() 394 struct rhash_lock_head __rcu *const *bkt) in rht_ptr_rcu() argument [all …]
|
| /linux/fs/smb/client/ |
| H A D | compress.c | 63 static bool has_low_entropy(struct bucket *bkt, size_t slen) in has_low_entropy() argument 71 for (i = 0; i < 256 && bkt[i].count > 0; i++) { in has_low_entropy() 72 p = bkt[i].count; in has_low_entropy() 98 static int calc_byte_distribution(struct bucket *bkt, size_t slen) in calc_byte_distribution() argument 105 sum += bkt[i].count; in calc_byte_distribution() 110 for (; i < high && bkt[i].count > 0; i++) { in calc_byte_distribution() 111 sum += bkt[i].count; in calc_byte_distribution() 125 static bool is_mostly_ascii(const struct bucket *bkt) in is_mostly_ascii() argument 131 if (bkt[i].count > 0) in is_mostly_ascii() 197 struct bucket *bkt = NULL; in is_compressible() local [all …]
|
| /linux/tools/include/linux/ |
| H A D | hashtable.h | 105 #define hash_for_each(name, bkt, obj, member) \ argument 106 for ((bkt) = 0, obj = NULL; obj == NULL && (bkt) < HASH_SIZE(name);\ 107 (bkt)++)\ 108 hlist_for_each_entry(obj, &name[bkt], member) 119 #define hash_for_each_safe(name, bkt, tmp, obj, member) \ argument 120 for ((bkt) = 0, obj = NULL; obj == NULL && (bkt) < HASH_SIZE(name);\ 121 (bkt)++)\ 122 hlist_for_each_entry_safe(obj, tmp, &name[bkt], member)
|
| /linux/drivers/infiniband/ulp/opa_vnic/ |
| H A D | opa_vnic_internal.h | 280 #define vnic_hash_for_each_safe(name, bkt, tmp, obj, member) \ argument 281 for ((bkt) = 0, obj = NULL; \ 282 !obj && (bkt) < OPA_VNIC_MAC_TBL_SIZE; (bkt)++) \ 283 hlist_for_each_entry_safe(obj, tmp, &name[bkt], member) 289 #define vnic_hash_for_each(name, bkt, obj, member) \ argument 290 for ((bkt) = 0, obj = NULL; \ 291 !obj && (bkt) < OPA_VNIC_MAC_TBL_SIZE; (bkt)++) \ 292 hlist_for_each_entry(obj, &name[bkt], member)
|
| H A D | opa_vnic_encap.c | 107 int bkt; in opa_vnic_free_mac_tbl() local 112 vnic_hash_for_each_safe(mactbl, bkt, tmp, node, hlist) { in opa_vnic_free_mac_tbl() 157 int bkt; in opa_vnic_query_mac_tbl() local 168 vnic_hash_for_each(mactbl, bkt, node, hlist) { in opa_vnic_query_mac_tbl() 209 int i, bkt, rc = 0; in opa_vnic_update_mac_tbl() local 263 vnic_hash_for_each(old_mactbl, bkt, node, hlist) { in opa_vnic_update_mac_tbl()
|
| /linux/tools/testing/selftests/bpf/prog_tests/ |
| H A D | hashmap.c | 47 int err, bkt, found_cnt, i; in test_hashmap_generic() local 92 hashmap__for_each_entry(map, entry, bkt) { in test_hashmap_generic() 140 hashmap__for_each_entry_safe(map, entry, tmp, bkt) { in test_hashmap_generic() 197 hashmap__for_each_entry_safe(map, entry, tmp, bkt) { in test_hashmap_generic() 230 hashmap__for_each_entry(map, entry, bkt) { in test_hashmap_generic() 238 hashmap__for_each_entry(map, entry, bkt) { in test_hashmap_generic() 265 int err, i, bkt; in test_hashmap_ptr_iface() local 319 hashmap__for_each_entry(map, cur, bkt) { in test_hashmap_ptr_iface() 345 int err, bkt; in test_hashmap_multimap() local 382 hashmap__for_each_entry(map, entry, bkt) { in test_hashmap_multimap() [all …]
|
| /linux/tools/perf/util/ |
| H A D | expr.c | 79 size_t bkt; in ids__free() local 84 hashmap__for_each_entry(ids, cur, bkt) { in ids__free() 108 size_t bkt; in ids__union() local 126 hashmap__for_each_entry(ids2, cur, bkt) { in ids__union() 231 size_t bkt; in expr__subset_of_ids() local 234 hashmap__for_each_entry(needles->ids, cur, bkt) { in expr__subset_of_ids() 310 size_t bkt; in expr__ctx_clear() local 312 hashmap__for_each_entry(ctx->ids, cur, bkt) { in expr__ctx_clear() 322 size_t bkt; in expr__ctx_free() local 328 hashmap__for_each_entry(ctx->ids, cur, bkt) { in expr__ctx_free()
|
| H A D | hwmon_pmu.c | 240 size_t bkt; in hwmon_pmu__read_events() local 323 hashmap__for_each_entry_safe((&pmu->events), cur, tmp, bkt) { in hwmon_pmu__read_events() 397 size_t bkt; in hwmon_pmu__exit() local 399 hashmap__for_each_entry_safe((&hwm->events), cur, tmp, bkt) { in hwmon_pmu__exit() 463 size_t bkt; in hwmon_pmu__for_each_event() local 468 hashmap__for_each_entry((&hwm->events), cur, bkt) { in hwmon_pmu__for_each_event() 560 size_t bkt; in hwmon_pmu__have_event() local 575 hashmap__for_each_entry((&hwm->events), cur, bkt) { in hwmon_pmu__have_event() 602 size_t bkt; in hwmon_pmu__config_term() local 605 hashmap__for_each_entry((&hwm->events), cur, bkt) { in hwmon_pmu__config_term()
|
| H A D | hashmap.c | 66 size_t bkt; in hashmap__clear() local 68 hashmap__for_each_entry_safe(map, cur, tmp, bkt) { in hashmap__clear() 106 size_t h, bkt; in hashmap_grow() local 117 hashmap__for_each_entry_safe(map, cur, tmp, bkt) { in hashmap_grow()
|
| H A D | metricgroup.c | 565 size_t bkt; in metricgroup__build_event_string() local 572 hashmap__for_each_entry(ctx->ids, cur, bkt) { in metricgroup__build_event_string() 747 size_t bkt; in resolve_metric() local 763 hashmap__for_each_entry(root_metric->pctx->ids, cur, bkt) { in resolve_metric() 1218 size_t bkt; in build_combined_expr_ctx() local 1229 hashmap__for_each_entry(m->pctx->ids, cur, bkt) { in build_combined_expr_ctx()
|
| H A D | bpf-trace-summary.c | 387 size_t bkt; in trace_print_bpf_summary() local 423 hashmap__for_each_entry(&schash, entry, bkt) in trace_print_bpf_summary()
|
| /linux/lib/ |
| H A D | rhashtable.c | 229 struct rhash_lock_head __rcu **bkt, in rhashtable_rehash_one() argument 245 rht_for_each_from(entry, rht_ptr(bkt, old_tbl, old_hash), in rhashtable_rehash_one() 274 rht_assign_locked(bkt, next); in rhashtable_rehash_one() 284 struct rhash_lock_head __rcu **bkt = rht_bucket_var(old_tbl, old_hash); in rhashtable_rehash_chain() local 288 if (!bkt) in rhashtable_rehash_chain() 290 flags = rht_lock(old_tbl, bkt); in rhashtable_rehash_chain() 292 while (!(err = rhashtable_rehash_one(ht, bkt, old_hash))) in rhashtable_rehash_chain() 297 rht_unlock(old_tbl, bkt, flags); in rhashtable_rehash_chain() 495 struct rhash_lock_head __rcu **bkt, in rhashtable_lookup_one() argument 508 rht_for_each_from(head, rht_ptr(bkt, tbl, hash), tbl, hash) { in rhashtable_lookup_one() [all …]
|
| /linux/net/ipv6/ |
| H A D | calipso.c | 203 u32 bkt; in calipso_cache_check() local 212 bkt = hash & (CALIPSO_CACHE_BUCKETS - 1); in calipso_cache_check() 213 spin_lock_bh(&calipso_cache[bkt].lock); in calipso_cache_check() 214 list_for_each_entry(entry, &calipso_cache[bkt].list, list) { in calipso_cache_check() 224 spin_unlock_bh(&calipso_cache[bkt].lock); in calipso_cache_check() 239 spin_unlock_bh(&calipso_cache[bkt].lock); in calipso_cache_check() 244 spin_unlock_bh(&calipso_cache[bkt].lock); in calipso_cache_check() 268 u32 bkt; in calipso_cache_add() local 291 bkt = entry->hash & (CALIPSO_CACHE_BUCKETS - 1); in calipso_cache_add() 292 spin_lock_bh(&calipso_cache[bkt].lock); in calipso_cache_add() [all …]
|
| /linux/fs/smb/server/ |
| H A D | connection.c | 98 int bkt; in ksmbd_conn_lookup_dialect() local 102 hash_for_each(conn_list, bkt, t, hlist) { in ksmbd_conn_lookup_dialect() 163 int bkt; in ksmbd_all_conn_set_status() local 166 hash_for_each(conn_list, bkt, conn, hlist) { in ksmbd_all_conn_set_status() 182 int rcount = 1, bkt; in ksmbd_conn_wait_idle_sess_id() local 189 hash_for_each(conn_list, bkt, conn, hlist) { in ksmbd_conn_wait_idle_sess_id() 482 int bkt; in stop_sessions() local 486 hash_for_each(conn_list, bkt, conn, hlist) { in stop_sessions()
|
| /linux/tools/lib/bpf/ |
| H A D | hashmap.c | 66 size_t bkt; in hashmap__clear() local 68 hashmap__for_each_entry_safe(map, cur, tmp, bkt) { in hashmap__clear() 106 size_t h, bkt; in hashmap_grow() local 117 hashmap__for_each_entry_safe(map, cur, tmp, bkt) { in hashmap_grow()
|
| /linux/net/sched/ |
| H A D | cls_route.c | 55 struct route4_bucket *bkt; member 328 b = f->bkt; in route4_delete() 465 f->bkt = b; in route4_set_parms() 524 f->bkt = fold->bkt; in route4_change() 534 fp = &f->bkt->ht[h]; in route4_change()
|
| /linux/arch/powerpc/kvm/ |
| H A D | book3s_hv_uvmem.c | 470 int srcu_idx, bkt; in kvmppc_h_svm_init_start() local 489 kvm_for_each_memslot(memslot, bkt, slots) { in kvmppc_h_svm_init_start() 497 kvm_for_each_memslot(m, bkt, slots) { in kvmppc_h_svm_init_start() 661 int srcu_idx, bkt; in kvmppc_h_svm_init_abort() local 676 kvm_for_each_memslot(memslot, bkt, kvm_memslots(kvm)) in kvmppc_h_svm_init_abort() 834 int srcu_idx, bkt; in kvmppc_h_svm_init_done() local 843 kvm_for_each_memslot(memslot, bkt, slots) { in kvmppc_h_svm_init_done()
|
| /linux/drivers/net/ethernet/mellanox/mlx5/core/en/ |
| H A D | htb.c | 37 int bkt, err; in mlx5e_htb_enumerate_leaves() local 39 hash_for_each(htb->qos_tc2node, bkt, node, hnode) { in mlx5e_htb_enumerate_leaves() 434 int bkt; in mlx5e_htb_node_find_by_qid() local 436 hash_for_each(htb->qos_tc2node, bkt, node, hnode) in mlx5e_htb_node_find_by_qid() 627 int bkt; in mlx5e_htb_update_children() local 629 hash_for_each(htb->qos_tc2node, bkt, child, hnode) { in mlx5e_htb_update_children()
|
| /linux/arch/x86/kvm/ |
| H A D | debugfs.c | 115 int bkt; in kvm_mmu_rmaps_stat_show() local 118 kvm_for_each_memslot(slot, bkt, slots) in kvm_mmu_rmaps_stat_show()
|
| /linux/net/netlabel/ |
| H A D | netlabel_domainhash.c | 140 u32 bkt; in netlbl_domhsh_search() local 145 bkt = netlbl_domhsh_hash(domain); in netlbl_domhsh_search() 146 bkt_list = &netlbl_domhsh_rcu_deref(netlbl_domhsh)->tbl[bkt]; in netlbl_domhsh_search() 436 u32 bkt = netlbl_domhsh_hash(entry->domain); in netlbl_domhsh_add() local 438 &rcu_dereference(netlbl_domhsh)->tbl[bkt]); in netlbl_domhsh_add()
|
| /linux/tools/bpf/bpftool/ |
| H A D | pids.c | 187 size_t bkt; in delete_obj_refs_table() local 192 hashmap__for_each_entry(map, entry, bkt) { in delete_obj_refs_table()
|
| /linux/tools/perf/tests/ |
| H A D | pmu-events.c | 935 size_t bkt; in metric_parse_fake() local 957 hashmap__for_each_entry(ctx->ids, cur, bkt) in metric_parse_fake() 960 hashmap__for_each_entry(ctx->ids, cur, bkt) { in metric_parse_fake() 974 hashmap__for_each_entry(ctx->ids, cur, bkt) in metric_parse_fake()
|
| /linux/arch/x86/kvm/mmu/ |
| H A D | page_track.c | 173 int r = 0, i, bkt; in kvm_enable_external_write_tracking() local 190 kvm_for_each_memslot(slot, bkt, slots) { in kvm_enable_external_write_tracking()
|
| /linux/drivers/net/ethernet/rocker/ |
| H A D | rocker_ofdpa.c | 1953 int bkt; in ofdpa_port_fdb_flush() local 1964 hash_for_each_safe(ofdpa->fdb_tbl, bkt, tmp, found, entry) { in ofdpa_port_fdb_flush() 1994 int bkt; in ofdpa_fdb_cleanup() local 1998 hash_for_each_safe(ofdpa->fdb_tbl, bkt, tmp, entry, entry) { in ofdpa_fdb_cleanup() 2387 int bkt; in ofdpa_fini() local 2393 hash_for_each_safe(ofdpa->flow_tbl, bkt, tmp, flow_entry, entry) in ofdpa_fini() 2398 hash_for_each_safe(ofdpa->group_tbl, bkt, tmp, group_entry, entry) in ofdpa_fini() 2403 hash_for_each_safe(ofdpa->fdb_tbl, bkt, tmp, fdb_entry, entry) in ofdpa_fini() 2408 hash_for_each_safe(ofdpa->internal_vlan_tbl, bkt, in ofdpa_fini() 2414 hash_for_each_safe(ofdpa->neigh_tbl, bkt, tmp, neigh_entry, entry) in ofdpa_fini() [all …]
|