Home
last modified time | relevance | path

Searched refs:bkt (Results 1 – 25 of 69) sorted by relevance

123

/linux/include/linux/
H A Dhashtable.h126 #define hash_for_each(name, bkt, obj, member) \ argument
127 for ((bkt) = 0, obj = NULL; obj == NULL && (bkt) < HASH_SIZE(name);\
128 (bkt)++)\
129 hlist_for_each_entry(obj, &name[bkt], member)
138 #define hash_for_each_rcu(name, bkt, obj, member) \ argument
139 for ((bkt) = 0, obj = NULL; obj == NULL && (bkt) < HASH_SIZE(name);\
140 (bkt)++)\
141 hlist_for_each_entry_rcu(obj, &name[bkt], member)
152 #define hash_for_each_safe(name, bkt, tmp, obj, member) \ argument
153 for ((bkt) = 0, obj = NULL; obj == NULL && (bkt) < HASH_SIZE(name);\
[all …]
/linux/fs/smb/client/
H A Dcompress.c49 * @bkt: Bytes counts of the sample.
63 static bool has_low_entropy(struct bucket *bkt, size_t slen) in has_low_entropy() argument
71 for (i = 0; i < 256 && bkt[i].count > 0; i++) { in has_low_entropy()
72 p = bkt[i].count; in has_low_entropy()
87 * @bkt: Byte counts of the sample.
98 static int calc_byte_distribution(struct bucket *bkt, size_t slen) in calc_byte_distribution() argument
105 sum += bkt[i].count; in calc_byte_distribution()
110 for (; i < high && bkt[i].count > 0; i++) { in calc_byte_distribution()
111 sum += bkt[i].count; in calc_byte_distribution()
125 static bool is_mostly_ascii(const struct bucket *bkt) in is_mostly_ascii() argument
197 struct bucket *bkt = NULL; is_compressible() local
[all...]
/linux/tools/include/linux/
H A Dhashtable.h105 #define hash_for_each(name, bkt, obj, member) \ argument
106 for ((bkt) = 0, obj = NULL; obj == NULL && (bkt) < HASH_SIZE(name);\
107 (bkt)++)\
108 hlist_for_each_entry(obj, &name[bkt], member)
119 #define hash_for_each_safe(name, bkt, tmp, obj, member) \ argument
120 for ((bkt) = 0, obj = NULL; obj == NULL && (bkt) < HASH_SIZE(name);\
121 (bkt)++)\
122 hlist_for_each_entry_safe(obj, tmp, &name[bkt], member)
/linux/drivers/infiniband/ulp/opa_vnic/
H A Dopa_vnic_internal.h280 #define vnic_hash_for_each_safe(name, bkt, tmp, obj, member) \ argument
281 for ((bkt) = 0, obj = NULL; \
282 !obj && (bkt) < OPA_VNIC_MAC_TBL_SIZE; (bkt)++) \
283 hlist_for_each_entry_safe(obj, tmp, &name[bkt], member)
289 #define vnic_hash_for_each(name, bkt, obj, member) \ argument
290 for ((bkt) = 0, obj = NULL; \
291 !obj && (bkt) < OPA_VNIC_MAC_TBL_SIZE; (bkt)++) \
292 hlist_for_each_entry(obj, &name[bkt], member)
H A Dopa_vnic_encap.c107 int bkt; in opa_vnic_free_mac_tbl() local
112 vnic_hash_for_each_safe(mactbl, bkt, tmp, node, hlist) { in opa_vnic_free_mac_tbl()
157 int bkt; in opa_vnic_query_mac_tbl() local
168 vnic_hash_for_each(mactbl, bkt, node, hlist) { in opa_vnic_query_mac_tbl()
209 int i, bkt, rc = 0; in opa_vnic_update_mac_tbl() local
263 vnic_hash_for_each(old_mactbl, bkt, node, hlist) { in opa_vnic_update_mac_tbl()
/linux/tools/testing/selftests/bpf/prog_tests/
H A Dhashmap.c47 int err, bkt, found_cnt, i; in test_hashmap_generic() local
92 hashmap__for_each_entry(map, entry, bkt) { in test_hashmap_generic()
140 hashmap__for_each_entry_safe(map, entry, tmp, bkt) { in test_hashmap_generic()
197 hashmap__for_each_entry_safe(map, entry, tmp, bkt) { in test_hashmap_generic()
230 hashmap__for_each_entry(map, entry, bkt) { in test_hashmap_generic()
238 hashmap__for_each_entry(map, entry, bkt) { in test_hashmap_generic()
265 int err, i, bkt; in test_hashmap_ptr_iface() local
319 hashmap__for_each_entry(map, cur, bkt) { in test_hashmap_ptr_iface()
345 int err, bkt; in test_hashmap_multimap() local
382 hashmap__for_each_entry(map, entry, bkt) { in test_hashmap_multimap()
[all …]
/linux/tools/perf/util/
H A Dexpr.c79 size_t bkt; in ids__free() local
84 hashmap__for_each_entry(ids, cur, bkt) { in ids__free()
108 size_t bkt; in ids__union() local
126 hashmap__for_each_entry(ids2, cur, bkt) { in ids__union()
231 size_t bkt; in expr__subset_of_ids() local
234 hashmap__for_each_entry(needles->ids, cur, bkt) { in expr__subset_of_ids()
310 size_t bkt; in expr__ctx_clear() local
312 hashmap__for_each_entry(ctx->ids, cur, bkt) { in expr__ctx_clear()
322 size_t bkt; in expr__ctx_free() local
328 hashmap__for_each_entry(ctx->ids, cur, bkt) { in expr__ctx_free()
H A Dhwmon_pmu.c240 size_t bkt; in hwmon_pmu__read_events() local
323 hashmap__for_each_entry_safe((&pmu->events), cur, tmp, bkt) { in hwmon_pmu__read_events()
397 size_t bkt; in hwmon_pmu__exit() local
399 hashmap__for_each_entry_safe((&hwm->events), cur, tmp, bkt) { in hwmon_pmu__exit()
463 size_t bkt; in hwmon_pmu__for_each_event() local
468 hashmap__for_each_entry((&hwm->events), cur, bkt) { in hwmon_pmu__for_each_event()
560 size_t bkt; in hwmon_pmu__have_event() local
575 hashmap__for_each_entry((&hwm->events), cur, bkt) { in hwmon_pmu__have_event()
602 size_t bkt; in hwmon_pmu__config_term() local
605 hashmap__for_each_entry((&hwm->events), cur, bkt) { in hwmon_pmu__config_term()
H A Dhashmap.c66 size_t bkt; in hashmap__clear() local
68 hashmap__for_each_entry_safe(map, cur, tmp, bkt) { in hashmap__clear()
106 size_t h, bkt; in hashmap_grow() local
117 hashmap__for_each_entry_safe(map, cur, tmp, bkt) { in hashmap_grow()
/linux/lib/
H A Drhashtable.c229 struct rhash_lock_head __rcu **bkt, in rhashtable_rehash_one() argument
245 rht_for_each_from(entry, rht_ptr(bkt, old_tbl, old_hash), in rhashtable_rehash_one()
274 rht_assign_locked(bkt, next); in rhashtable_rehash_one()
284 struct rhash_lock_head __rcu **bkt = rht_bucket_var(old_tbl, old_hash); in rhashtable_rehash_chain() local
288 if (!bkt) in rhashtable_rehash_chain()
290 flags = rht_lock(old_tbl, bkt); in rhashtable_rehash_chain()
292 while (!(err = rhashtable_rehash_one(ht, bkt, old_hash))) in rhashtable_rehash_chain()
297 rht_unlock(old_tbl, bkt, flags); in rhashtable_rehash_chain()
497 struct rhash_lock_head __rcu **bkt, in rhashtable_lookup_one()
510 rht_for_each_from(head, rht_ptr(bkt, tb in rhashtable_lookup_one()
495 rhashtable_lookup_one(struct rhashtable * ht,struct rhash_lock_head __rcu ** bkt,struct bucket_table * tbl,unsigned int hash,const void * key,struct rhash_head * obj) rhashtable_lookup_one() argument
546 rhashtable_insert_one(struct rhashtable * ht,struct rhash_lock_head __rcu ** bkt,struct bucket_table * tbl,unsigned int hash,struct rhash_head * obj,void * data) rhashtable_insert_one() argument
595 struct rhash_lock_head __rcu **bkt; rhashtable_try_insert() local
[all...]
/linux/drivers/s390/crypto/
H A Dap_card.c79 int bkt; in request_count_store() local
84 hash_for_each(ap_queues, bkt, aq, hnode) in request_count_store()
98 int bkt; in requestq_count_show() local
105 hash_for_each(ap_queues, bkt, aq, hnode) in requestq_count_show()
117 int bkt; in pendingq_count_show() local
124 hash_for_each(ap_queues, bkt, aq, hnode) in pendingq_count_show()
/linux/net/ipv6/
H A Dcalipso.c202 u32 bkt; in calipso_cache_check() local
211 bkt = hash & (CALIPSO_CACHE_BUCKETS - 1); in calipso_cache_check()
212 spin_lock_bh(&calipso_cache[bkt].lock); in calipso_cache_check()
213 list_for_each_entry(entry, &calipso_cache[bkt].list, list) { in calipso_cache_check()
223 spin_unlock_bh(&calipso_cache[bkt].lock); in calipso_cache_check()
238 spin_unlock_bh(&calipso_cache[bkt].lock); in calipso_cache_check()
243 spin_unlock_bh(&calipso_cache[bkt].lock); in calipso_cache_check()
267 u32 bkt; in calipso_cache_add() local
290 bkt = entry->hash & (CALIPSO_CACHE_BUCKETS - 1); in calipso_cache_add()
291 spin_lock_bh(&calipso_cache[bkt] in calipso_cache_add()
[all...]
/linux/fs/smb/server/
H A Dconnection.c155 int bkt; in ksmbd_conn_lookup_dialect() local
159 hash_for_each(conn_list, bkt, t, hlist) { in ksmbd_conn_lookup_dialect()
220 int bkt; in ksmbd_all_conn_set_status() local
223 hash_for_each(conn_list, bkt, conn, hlist) { in ksmbd_all_conn_set_status()
239 int rcount = 1, bkt; in ksmbd_conn_wait_idle_sess_id() local
246 hash_for_each(conn_list, bkt, conn, hlist) { in ksmbd_conn_wait_idle_sess_id()
540 int bkt; in stop_sessions() local
544 hash_for_each(conn_list, bkt, conn, hlist) { in stop_sessions()
/linux/drivers/net/ethernet/intel/ice/
H A Dice_sriov.c29 unsigned int bkt; in ice_free_vf_entries() local
37 hash_for_each_safe(vfs->table, bkt, tmp, vf, entry) { in ice_free_vf_entries()
137 unsigned int bkt; in ice_free_vfs() local
156 ice_for_each_vf(pf, bkt, vf) { in ice_free_vfs()
470 unsigned int bkt, it_cnt; in ice_start_vfs() local
477 ice_for_each_vf(pf, bkt, vf) { in ice_start_vfs()
505 ice_for_each_vf(pf, bkt, vf) { in ice_start_vfs()
880 int to_remap = 0, bkt; in ice_sriov_remap_vectors() local
885 ice_for_each_vf(pf, bkt, tmp_vf) { in ice_sriov_remap_vectors()
1077 unsigned int bkt; in ice_process_vflr_event() local
1114 unsigned int bkt; ice_get_vf_from_pfq() local
1470 unsigned int bkt; ice_calc_all_vfs_min_tx_rate() local
1782 unsigned int bkt; ice_print_vfs_mdd_events() local
1823 u32 bkt; ice_restore_all_vfs_msi_state() local
[all...]
H A Dice_vf_lib.c108 unsigned int bkt; in ice_get_num_vfs() local
112 ice_for_each_vf_rcu(pf, bkt, vf) in ice_get_num_vfs()
580 unsigned int bkt; in ice_is_any_vf_in_unicast_promisc() local
583 ice_for_each_vf_rcu(pf, bkt, vf) { in ice_is_any_vf_in_unicast_promisc()
753 unsigned int bkt; in ice_reset_all_vfs() local
762 ice_for_each_vf(pf, bkt, vf) in ice_reset_all_vfs()
772 ice_for_each_vf(pf, bkt, vf) in ice_reset_all_vfs()
779 ice_for_each_vf(pf, bkt, vf) { in ice_reset_all_vfs()
791 ice_for_each_vf(pf, bkt, vf) { in ice_reset_all_vfs()
1382 unsigned int bkt; in ice_get_vf_ctrl_vsi() local
[all...]
/linux/tools/lib/bpf/
H A Dhashmap.c66 size_t bkt; in hashmap__clear() local
68 hashmap__for_each_entry_safe(map, cur, tmp, bkt) { in hashmap__clear()
106 size_t h, bkt; in hashmap_grow() local
117 hashmap__for_each_entry_safe(map, cur, tmp, bkt) { in hashmap_grow()
/linux/net/ipv4/
H A Dcipso_ipv4.c236 u32 bkt; in cipso_v4_cache_check() local
245 bkt = hash & (CIPSO_V4_CACHE_BUCKETS - 1); in cipso_v4_cache_check()
246 spin_lock_bh(&cipso_v4_cache[bkt].lock); in cipso_v4_cache_check()
247 list_for_each_entry(entry, &cipso_v4_cache[bkt].list, list) { in cipso_v4_cache_check()
257 spin_unlock_bh(&cipso_v4_cache[bkt].lock); in cipso_v4_cache_check()
272 spin_unlock_bh(&cipso_v4_cache[bkt].lock); in cipso_v4_cache_check()
277 spin_unlock_bh(&cipso_v4_cache[bkt].lock); in cipso_v4_cache_check()
300 u32 bkt; in cipso_v4_cache_add() local
323 bkt = entry->hash & (CIPSO_V4_CACHE_BUCKETS - 1); in cipso_v4_cache_add()
324 spin_lock_bh(&cipso_v4_cache[bkt] in cipso_v4_cache_add()
[all...]
/linux/net/sched/
H A Dcls_route.c55 struct route4_bucket *bkt; member
328 b = f->bkt; in route4_delete()
465 f->bkt = b; in route4_set_parms()
524 f->bkt = fold->bkt; in route4_change()
534 fp = &f->bkt->ht[h]; in route4_change()
/linux/drivers/net/ethernet/mellanox/mlx5/core/lib/
H A Dvxlan.c183 int bkt; in mlx5_vxlan_reset_to_default() local
188 hash_for_each_safe(vxlan->htable, bkt, tmp, vxlanp, hlist) { in mlx5_vxlan_reset_to_default()
/linux/arch/powerpc/kvm/
H A Dbook3s_hv_uvmem.c470 int srcu_idx, bkt; in kvmppc_h_svm_init_start() local
489 kvm_for_each_memslot(memslot, bkt, slots) { in kvmppc_h_svm_init_start()
497 kvm_for_each_memslot(m, bkt, slots) { in kvmppc_h_svm_init_start()
661 int srcu_idx, bkt; in kvmppc_h_svm_init_abort() local
676 kvm_for_each_memslot(memslot, bkt, kvm_memslots(kvm)) in kvmppc_h_svm_init_abort()
834 int srcu_idx, bkt; in kvmppc_h_svm_init_done() local
843 kvm_for_each_memslot(memslot, bkt, slots) { in kvmppc_h_svm_init_done()
/linux/drivers/net/ethernet/mellanox/mlx5/core/en/
H A Dhtb.c37 int bkt, err; in mlx5e_htb_enumerate_leaves() local
39 hash_for_each(htb->qos_tc2node, bkt, node, hnode) { in mlx5e_htb_enumerate_leaves()
434 int bkt; in mlx5e_htb_node_find_by_qid() local
436 hash_for_each(htb->qos_tc2node, bkt, node, hnode) in mlx5e_htb_node_find_by_qid()
627 int bkt; in mlx5e_htb_update_children() local
629 hash_for_each(htb->qos_tc2node, bkt, child, hnode) { in mlx5e_htb_update_children()
/linux/net/netlabel/
H A Dnetlabel_domainhash.c140 u32 bkt; in netlbl_domhsh_search() local
145 bkt = netlbl_domhsh_hash(domain); in netlbl_domhsh_search()
146 bkt_list = &netlbl_domhsh_rcu_deref(netlbl_domhsh)->tbl[bkt]; in netlbl_domhsh_search()
434 u32 bkt = netlbl_domhsh_hash(entry->domain); in netlbl_domhsh_add() local
436 &rcu_dereference(netlbl_domhsh)->tbl[bkt]); in netlbl_domhsh_add()
H A Dnetlabel_unlabeled.c204 u32 bkt; in netlbl_unlhsh_search_iface() local
208 bkt = netlbl_unlhsh_hash(ifindex); in netlbl_unlhsh_search_iface()
209 bkt_list = &netlbl_unlhsh_rcu_deref(netlbl_unlhsh)->tbl[bkt]; in netlbl_unlhsh_search_iface()
314 u32 bkt; in netlbl_unlhsh_add_iface() local
328 bkt = netlbl_unlhsh_hash(ifindex); in netlbl_unlhsh_add_iface()
332 &netlbl_unlhsh_rcu_deref(netlbl_unlhsh)->tbl[bkt]); in netlbl_unlhsh_add_iface()
/linux/arch/x86/kvm/
H A Ddebugfs.c115 int bkt; in kvm_mmu_rmaps_stat_show() local
118 kvm_for_each_memslot(slot, bkt, slots) in kvm_mmu_rmaps_stat_show()
/linux/tools/bpf/bpftool/
H A Dpids.c187 size_t bkt; in delete_obj_refs_table() local
192 hashmap__for_each_entry(map, entry, bkt) { in delete_obj_refs_table()

123