Home
last modified time | relevance | path

Searched refs:buckets (Results 1 – 25 of 74) sorted by relevance

123

/linux/drivers/s390/scsi/
H A Dzfcp_reqlist.h26 struct list_head buckets[ZFCP_REQ_LIST_BUCKETS]; member
52 INIT_LIST_HEAD(&rl->buckets[i]); in zfcp_reqlist_alloc()
68 if (!list_empty(&rl->buckets[i])) in zfcp_reqlist_isempty()
92 list_for_each_entry(req, &rl->buckets[i], list) in _zfcp_reqlist_find()
165 list_add_tail(&req->list, &rl->buckets[i]); in zfcp_reqlist_add()
182 list_splice_init(&rl->buckets[i], list); in zfcp_reqlist_move()
209 list_for_each_entry(req, &rl->buckets[i], list) in zfcp_reqlist_apply_for_all()
/linux/Documentation/networking/
H A Dnexthop-group-resilient.rst54 continuous. With a hash table, mapping between the hash table buckets and
56 the buckets that held it are simply reassigned to other next hops::
70 choose a subset of buckets that are currently not used for forwarding
72 keeping the "busy" buckets intact. This way, established flows are ideally
80 certain number of buckets, according to its weight and the number of
81 buckets in the hash table. In accordance with the source code, we will call
86 Next hops that have fewer buckets than their wants count, are called
98 buckets:
105 underweight next hops. If, after considering all buckets in this manner,
109 There may not be enough "idle" buckets to satisfy the updated wants counts
[all …]
/linux/tools/testing/selftests/bpf/
H A Dbpf_arena_htab.h14 htab_bucket_t *buckets; member
21 htab_bucket_t *b = htab->buckets; in __select_bucket()
95 void __arena *buckets = bpf_arena_alloc_pages(&arena, NULL, 2, NUMA_NO_NODE, 0); in htab_init() local
97 cast_user(buckets); in htab_init()
98 htab->buckets = buckets; in htab_init()
/linux/tools/lib/bpf/
H A Dhashmap.c45 map->buckets = NULL; in hashmap__init()
71 free(map->buckets); in hashmap__clear()
72 map->buckets = NULL; in hashmap__clear()
124 free(map->buckets); in hashmap_grow()
125 map->buckets = new_buckets; in hashmap_grow()
137 if (!map->buckets) in hashmap_find_entry()
140 for (prev_ptr = &map->buckets[hash], cur = *prev_ptr; in hashmap_find_entry()
200 hashmap_add_entry(&map->buckets[h], entry); in hashmap_insert()
/linux/tools/perf/util/
H A Dhashmap.c45 map->buckets = NULL; in hashmap__init()
71 free(map->buckets); in hashmap__clear()
72 map->buckets = NULL; in hashmap__clear()
124 free(map->buckets); in hashmap_grow()
125 map->buckets = new_buckets; in hashmap_grow()
137 if (!map->buckets) in hashmap_find_entry()
140 for (prev_ptr = &map->buckets[hash], cur = *prev_ptr; in hashmap_find_entry()
200 hashmap_add_entry(&map->buckets[h], entry); in hashmap_insert()
H A Dftrace.h57 int buckets[], struct stats *stats);
82 int buckets[] __maybe_unused, in perf_ftrace__latency_read_bpf()
H A Dbpf_ftrace.c169 int buckets[], struct stats *stats) in perf_ftrace__latency_read_bpf() argument
185 buckets[idx] = 0; in perf_ftrace__latency_read_bpf()
190 buckets[idx] += hist[i]; in perf_ftrace__latency_read_bpf()
/linux/block/
H A Dblk-stat.c83 for (bucket = 0; bucket < cb->buckets; bucket++) in blk_stat_timer_fn()
90 for (bucket = 0; bucket < cb->buckets; bucket++) { in blk_stat_timer_fn()
102 unsigned int buckets, void *data) in blk_stat_alloc_callback() argument
110 cb->stat = kmalloc_objs(struct blk_rq_stat, buckets); in blk_stat_alloc_callback()
115 cb->cpu_stat = __alloc_percpu(buckets * sizeof(struct blk_rq_stat), in blk_stat_alloc_callback()
126 cb->buckets = buckets; in blk_stat_alloc_callback()
143 for (bucket = 0; bucket < cb->buckets; bucket++) in blk_stat_add_callback()
/linux/tools/testing/selftests/drivers/net/netdevsim/
H A Dnexthop.sh213 $IP nexthop add id 10 group 1/2 type resilient buckets 4
229 $IP nexthop add id 10 group 1,3/2,2 type resilient buckets 5
259 $IP nexthop add id 10 group 1/2 type resilient buckets 4 &> /dev/null
325 $IP nexthop add id 10 group 1/2 type resilient buckets 6
353 $IP nexthop add id 10 group 1/2 type resilient buckets 6
408 $IP nexthop add id 10 group 1/2 type resilient buckets 8 idle_timer 4
434 type resilient buckets 8 idle_timer 6
469 $IP nexthop add id 10 group 1/2 type resilient buckets 8 $timer 4
504 $IP nexthop add id 10 group 1/2 type resilient buckets 8 $timer 8
535 type resilient buckets 8 $timer 4
[all …]
/linux/tools/perf/
H A Dbuiltin-ftrace.c854 static void make_histogram(struct perf_ftrace *ftrace, int buckets[], in make_histogram() argument
931 buckets[i]++; in make_histogram()
943 static void display_histogram(struct perf_ftrace *ftrace, int buckets[]) in display_histogram() argument
955 total += buckets[i]; in display_histogram()
965 bar_len = buckets[0] * bar_total / total; in display_histogram()
967 if (!ftrace->hide_empty || buckets[0]) in display_histogram()
970 buckets[0], bar_len, bar, bar_total - bar_len, ""); in display_histogram()
976 if (ftrace->hide_empty && !buckets[i]) in display_histogram()
1007 bar_len = buckets[i] * bar_total / total; in display_histogram()
1008 printf(" %s | %10d | %.*s%*s |\n", unit, buckets[i], bar_len, bar, in display_histogram()
[all …]
/linux/net/ceph/crush/
H A Dcrush.c111 if (map->buckets) { in crush_destroy()
114 if (map->buckets[b] == NULL) in crush_destroy()
116 crush_destroy_bucket(map->buckets[b]); in crush_destroy()
118 kfree(map->buckets); in crush_destroy()
H A Dmapper.c531 itemtype = map->buckets[-1-item]->type; in crush_choose_firstn()
544 in = map->buckets[-1-item]; in crush_choose_firstn()
568 map->buckets[-1-item], in crush_choose_firstn()
744 itemtype = map->buckets[-1-item]->type; in crush_choose_indep()
761 in = map->buckets[-1-item]; in crush_choose_indep()
781 map->buckets[-1-item], in crush_choose_indep()
868 if (!map->buckets[b]) in crush_init_workspace()
872 switch (map->buckets[b]->alg) { in crush_init_workspace()
880 v += map->buckets[b]->size * sizeof(__u32); in crush_init_workspace()
951 map->buckets[-1-curstep->arg1])) { in crush_do_rule()
[all …]
/linux/net/netfilter/ipvs/
H A Dip_vs_sh.c69 struct ip_vs_sh_bucket buckets[IP_VS_SH_TAB_SIZE]; member
107 struct ip_vs_dest *dest = rcu_dereference(s->buckets[hash].dest); in ip_vs_sh_get()
129 dest = rcu_dereference(s->buckets[ihash].dest); in ip_vs_sh_get_fallback()
144 dest = rcu_dereference(s->buckets[hash].dest); in ip_vs_sh_get_fallback()
171 b = &s->buckets[0]; in ip_vs_sh_reassign()
215 b = &s->buckets[0]; in ip_vs_sh_flush()
H A Dip_vs_dh.c63 struct ip_vs_dh_bucket buckets[IP_VS_DH_TAB_SIZE]; member
89 return rcu_dereference(s->buckets[ip_vs_dh_hashkey(af, addr)].dest); in ip_vs_dh_get()
105 b = &s->buckets[0]; in ip_vs_dh_reassign()
139 b = &s->buckets[0]; in ip_vs_dh_flush()
/linux/tools/testing/selftests/bpf/prog_tests/
H A Darena_htab.c20 printf("htab %p buckets %p n_buckets %d\n", htab, htab->buckets, htab->n_buckets); in test_arena_htab_common()
21 ASSERT_OK_PTR(htab->buckets, "htab->buckets shouldn't be NULL"); in test_arena_htab_common()
22 for (i = 0; htab->buckets && i < 16; i += 4) { in test_arena_htab_common()
/linux/drivers/net/ethernet/mellanox/mlx5/core/ipoib/
H A Dipoib_vlan.c45 struct hlist_head buckets[1 << MLX5I_MAX_LOG_PKEY_SUP]; member
71 static struct qpn_to_netdev *mlx5i_find_qpn_to_netdev_node(struct hlist_head *buckets, in mlx5i_find_qpn_to_netdev_node() argument
74 struct hlist_head *h = &buckets[hash_32(qpn, MLX5I_MAX_LOG_PKEY_SUP)]; in mlx5i_find_qpn_to_netdev_node()
99 hlist_add_head(&new_node->hlist, &ht->buckets[key]); in mlx5i_pkey_add_qpn()
112 node = mlx5i_find_qpn_to_netdev_node(ht->buckets, qpn); in mlx5i_pkey_del_qpn()
131 node = mlx5i_find_qpn_to_netdev_node(ipriv->qpn_htbl->buckets, qpn); in mlx5i_pkey_get_netdev()
/linux/drivers/md/
H A Ddm-region-hash.c73 struct list_head *buckets; member
209 rh->buckets = vmalloc_array(nr_buckets, sizeof(*rh->buckets)); in dm_region_hash_create()
210 if (!rh->buckets) { in dm_region_hash_create()
217 INIT_LIST_HEAD(rh->buckets + i); in dm_region_hash_create()
231 vfree(rh->buckets); in dm_region_hash_create()
247 list_for_each_entry_safe(reg, nreg, rh->buckets + h, in dm_region_hash_destroy()
258 vfree(rh->buckets); in dm_region_hash_destroy()
277 struct list_head *bucket = rh->buckets + rh_hash(rh, region); in __rh_lookup()
288 list_add(&reg->hash_list, rh->buckets + rh_hash(rh, reg->key)); in __rh_insert()
/linux/drivers/md/persistent-data/
H A Ddm-transaction-manager.c99 struct rb_root buckets[DM_HASH_SIZE]; member
113 node = &tm->buckets[bucket].rb_node; in is_shadow()
147 node = &tm->buckets[bucket].rb_node; in insert_shadow()
159 rb_insert_color(&si->node, &tm->buckets[bucket]); in insert_shadow()
170 while (!RB_EMPTY_ROOT(&tm->buckets[i])) { in wipe_shadow_table()
172 rb_entry(tm->buckets[i].rb_node, struct shadow_info, node); in wipe_shadow_table()
173 rb_erase(&si->node, &tm->buckets[i]); in wipe_shadow_table()
199 tm->buckets[i] = RB_ROOT; in dm_tm_create()
/linux/fs/nfs/
H A Dpnfs_nfs.c104 p = kmalloc_flex(*p, buckets, n, gfp_flags); in pnfs_alloc_commit_array()
111 for (b = &p->buckets[0]; n != 0; b++, n--) { in pnfs_alloc_commit_array()
261 struct pnfs_commit_bucket *buckets, in pnfs_bucket_scan_array() argument
269 cnt = pnfs_bucket_scan_ds_commit_list(&buckets[i], cinfo, max); in pnfs_bucket_scan_array()
290 cnt = pnfs_bucket_scan_array(cinfo, array->buckets, in pnfs_generic_scan_commit_lists()
306 struct pnfs_commit_bucket *buckets, in pnfs_bucket_recover_commit_reqs() argument
316 for (i = 0, b = buckets; i < nbuckets; i++, b++) { in pnfs_bucket_recover_commit_reqs()
345 array->buckets, in pnfs_generic_recover_commit_reqs()
385 static void pnfs_generic_retry_commit(struct pnfs_commit_bucket *buckets, in pnfs_generic_retry_commit() argument
394 for (bucket = buckets; idx < nbuckets; bucket++, idx++) { in pnfs_generic_retry_commit()
[all …]
/linux/kernel/bpf/
H A Dbpf_local_storage.c24 return &smap->buckets[hash_ptr(local_storage, smap->bucket_log)]; in select_bucket()
768 usage += sizeof(*smap->buckets) * (1ULL << smap->bucket_log); in bpf_local_storage_map_mem_usage()
791 smap->buckets = bpf_map_kvcalloc(&smap->map, nbuckets, in bpf_local_storage_map_alloc()
792 sizeof(*smap->buckets), GFP_USER | __GFP_NOWARN); in bpf_local_storage_map_alloc()
793 if (!smap->buckets) { in bpf_local_storage_map_alloc()
799 INIT_HLIST_HEAD(&smap->buckets[i].list); in bpf_local_storage_map_alloc()
800 raw_res_spin_lock_init(&smap->buckets[i].lock); in bpf_local_storage_map_alloc()
810 kvfree(smap->buckets); in bpf_local_storage_map_alloc()
842 b = &smap->buckets[i]; in bpf_local_storage_map_free()
875 kvfree(smap->buckets); in bpf_local_storage_map_free()
/linux/drivers/net/wireless/broadcom/brcm80211/brcmfmac/
H A Dpno.c298 struct brcmf_gscan_bucket_config **buckets, in brcmf_pno_prep_fwconfig() argument
323 *buckets = NULL; in brcmf_pno_prep_fwconfig()
355 *buckets = fw_buckets; in brcmf_pno_prep_fwconfig()
396 struct brcmf_gscan_bucket_config *buckets; in brcmf_pno_config_sched_scans() local
403 n_buckets = brcmf_pno_prep_fwconfig(pi, &pno_cfg, &buckets, in brcmf_pno_config_sched_scans()
437 memcpy(gscan_cfg->bucket, buckets, in brcmf_pno_config_sched_scans()
438 array_size(n_buckets, sizeof(*buckets))); in brcmf_pno_config_sched_scans()
463 kfree(buckets); in brcmf_pno_config_sched_scans()
/linux/net/sched/
H A Dsch_hhf.c19 * Flows are classified into two buckets: non-heavy-hitter and heavy-hitter
20 * buckets. Initially, a new flow starts as non-heavy-hitter. Once classified
22 * The buckets are dequeued by a Weighted Deficit Round Robin (WDRR) scheduler,
103 #define WDRR_BUCKET_CNT 2 /* two buckets for Weighted DRR */
128 struct wdrr_bucket buckets[WDRR_BUCKET_CNT]; member
147 struct list_head new_buckets; /* list of new buckets */
148 struct list_head old_buckets; /* list of old buckets */
245 /* Assigns packets to WDRR buckets. Implements a multi-stage filter to
356 bucket = &q->buckets[WDRR_BUCKET_FOR_HH]; in hhf_drop()
358 bucket = &q->buckets[WDRR_BUCKET_FOR_NON_H in hhf_drop()
[all...]
/linux/drivers/md/bcache/
H A Dalloc.c139 trace_bcache_invalidate(ca, b - ca->buckets); in __bch_invalidate_one_bucket()
151 fifo_push(&ca->free_inc, b - ca->buckets); in bch_invalidate_one_bucket()
220 b = ca->buckets + ca->fifo_last_bucket++; in invalidate_buckets_fifo()
246 b = ca->buckets + n; in invalidate_buckets_random()
433 b = ca->buckets + r; in bch_bucket_alloc()
495 k->ptr[0] = MAKE_PTR(ca->buckets[b].gen, in __bch_bucket_alloc_set()
/linux/net/netfilter/
H A Dnft_set_hash.c589 u32 buckets; member
610 hash = reciprocal_scale(hash, priv->buckets); in nft_hash_lookup()
634 hash = reciprocal_scale(hash, priv->buckets); in nft_hash_get()
655 hash = reciprocal_scale(hash, priv->buckets); in nft_hash_lookup_fast()
677 hash = reciprocal_scale(hash, priv->buckets); in nft_jhash()
758 for (i = 0; i < priv->buckets; i++) { in nft_hash_walk()
786 priv->buckets = nft_hash_buckets(desc->size); in nft_hash_init()
800 for (i = 0; i < priv->buckets; i++) { in nft_hash_destroy()
/linux/lib/
H A Drhashtable.c59 return bit_spin_is_locked(0, (unsigned long *)&tbl->buckets[hash]); in lockdep_rht_bucket_is_held()
72 return (void *)rcu_dereference_protected(tbl->buckets[0], 1); in nested_table_top()
159 size = sizeof(*tbl) + sizeof(tbl->buckets[0]); in nested_bucket_table_alloc()
166 if (!nested_table_alloc(ht, (union nested_table __rcu **)tbl->buckets, in nested_bucket_table_alloc()
187 kvmalloc_node_align_noprof(struct_size(tbl, buckets, nbuckets), in bucket_table_alloc()
210 INIT_RHT_NULLS_HEAD(tbl->buckets[i]); in bucket_table_alloc()
261 flags = rht_lock_nested(new_tbl, &new_tbl->buckets[new_hash], in rhashtable_rehash_one()
264 head = rht_ptr(new_tbl->buckets + new_hash, new_tbl, new_hash); in rhashtable_rehash_one()
268 rht_assign_unlock(new_tbl, &new_tbl->buckets[new_hash], entry, flags); in rhashtable_rehash_one()

123