Home
last modified time | relevance | path

Searched refs:buckets (Results 1 – 25 of 109) sorted by relevance

12345

/linux/fs/bcachefs/
H A Dmovinggc.c150 move_buckets *buckets) in bch2_copygc_get_buckets() argument
187 ret2 = darray_push(buckets, b); in bch2_copygc_get_buckets()
193 ret2 = buckets->nr >= nr_to_get; in bch2_copygc_get_buckets()
200 saw, in_flight, not_movable, buckets->nr, sectors, nr_to_get, ret); in bch2_copygc_get_buckets()
215 move_buckets buckets = { 0 }; in bch2_copygc() local
221 ret = bch2_copygc_get_buckets(ctxt, buckets_in_flight, &buckets); in bch2_copygc()
225 darray_for_each(buckets, i) { in bch2_copygc()
258 trace_and_count(c, copygc, c, buckets.nr, sectors_seen, sectors_moved); in bch2_copygc()
260 darray_exit(&buckets); in bch2_copygc()
328 struct buckets_in_flight *buckets; in bch2_copygc_thread() local
[all …]
H A Djournal_sb.c38 b[i] = le64_to_cpu(journal->buckets[i]); in bch2_sb_journal_validate()
79 prt_printf(out, " %llu", le64_to_cpu(journal->buckets[i])); in bch2_sb_journal_to_text()
192 u64 *buckets, unsigned nr) in bch2_journal_buckets_to_sb() argument
207 if (buckets[i] + 1 != buckets[i + 1]) in bch2_journal_buckets_to_sb()
217 j->d[dst].start = cpu_to_le64(buckets[0]); in bch2_journal_buckets_to_sb()
221 if (buckets[i] == buckets[i - 1] + 1) { in bch2_journal_buckets_to_sb()
225 j->d[dst].start = cpu_to_le64(buckets[i]); in bch2_journal_buckets_to_sb()
/linux/drivers/s390/scsi/
H A Dzfcp_reqlist.h26 struct list_head buckets[ZFCP_REQ_LIST_BUCKETS]; member
52 INIT_LIST_HEAD(&rl->buckets[i]); in zfcp_reqlist_alloc()
68 if (!list_empty(&rl->buckets[i])) in zfcp_reqlist_isempty()
92 list_for_each_entry(req, &rl->buckets[i], list) in _zfcp_reqlist_find()
165 list_add_tail(&req->list, &rl->buckets[i]); in zfcp_reqlist_add()
182 list_splice_init(&rl->buckets[i], list); in zfcp_reqlist_move()
209 list_for_each_entry(req, &rl->buckets[i], list) in zfcp_reqlist_apply_for_all()
/linux/Documentation/networking/
H A Dnexthop-group-resilient.rst54 continuous. With a hash table, mapping between the hash table buckets and
56 the buckets that held it are simply reassigned to other next hops::
70 choose a subset of buckets that are currently not used for forwarding
72 keeping the "busy" buckets intact. This way, established flows are ideally
80 certain number of buckets, according to its weight and the number of
81 buckets in the hash table. In accordance with the source code, we will call
86 Next hops that have fewer buckets than their wants count, are called
98 buckets:
105 underweight next hops. If, after considering all buckets in this manner,
109 There may not be enough "idle" buckets to satisfy the updated wants counts
[all …]
/linux/tools/testing/selftests/bpf/
H A Dbpf_arena_htab.h14 htab_bucket_t *buckets; member
21 htab_bucket_t *b = htab->buckets; in __select_bucket()
95 void __arena *buckets = bpf_arena_alloc_pages(&arena, NULL, 2, NUMA_NO_NODE, 0); in htab_init() local
97 cast_user(buckets); in htab_init()
98 htab->buckets = buckets; in htab_init()
/linux/tools/lib/bpf/
H A Dhashmap.c45 map->buckets = NULL; in hashmap__init()
71 free(map->buckets); in hashmap__clear()
72 map->buckets = NULL; in hashmap__clear()
124 free(map->buckets); in hashmap_grow()
125 map->buckets = new_buckets; in hashmap_grow()
137 if (!map->buckets) in hashmap_find_entry()
140 for (prev_ptr = &map->buckets[hash], cur = *prev_ptr; in hashmap_find_entry()
200 hashmap_add_entry(&map->buckets[h], entry); in hashmap_insert()
/linux/tools/perf/util/
H A Dhashmap.c45 map->buckets = NULL; in hashmap__init()
71 free(map->buckets); in hashmap__clear()
72 map->buckets = NULL; in hashmap__clear()
124 free(map->buckets); in hashmap_grow()
125 map->buckets = new_buckets; in hashmap_grow()
137 if (!map->buckets) in hashmap_find_entry()
140 for (prev_ptr = &map->buckets[hash], cur = *prev_ptr; in hashmap_find_entry()
200 hashmap_add_entry(&map->buckets[h], entry); in hashmap_insert()
H A Dftrace.h50 int buckets[], struct stats *stats);
75 int buckets[] __maybe_unused,
H A Dbpf_ftrace.c128 int buckets[], struct stats *stats) in perf_ftrace__latency_read_bpf()
144 buckets[idx] = 0; in perf_ftrace__latency_read_bpf()
149 buckets[idx] += hist[i];
122 perf_ftrace__latency_read_bpf(struct perf_ftrace * ftrace __maybe_unused,int buckets[]) perf_ftrace__latency_read_bpf() argument
/linux/block/
H A Dblk-stat.c83 for (bucket = 0; bucket < cb->buckets; bucket++) in blk_stat_timer_fn()
90 for (bucket = 0; bucket < cb->buckets; bucket++) { in blk_stat_timer_fn()
102 unsigned int buckets, void *data) in blk_stat_alloc_callback() argument
110 cb->stat = kmalloc_array(buckets, sizeof(struct blk_rq_stat), in blk_stat_alloc_callback()
116 cb->cpu_stat = __alloc_percpu(buckets * sizeof(struct blk_rq_stat), in blk_stat_alloc_callback()
127 cb->buckets = buckets; in blk_stat_alloc_callback()
144 for (bucket = 0; bucket < cb->buckets; bucket++) in blk_stat_add_callback()
/linux/drivers/md/dm-vdo/
H A Dpriority-table.c44 struct bucket buckets[]; member
69 struct bucket *bucket = &table->buckets[priority]; in vdo_make_priority_table()
116 list_del_init(&table->buckets[priority].queue); in vdo_reset_priority_table()
134 list_move_tail(entry, &table->buckets[priority].queue); in vdo_priority_table_enqueue()
173 bucket = &table->buckets[top_priority]; in vdo_priority_table_dequeue()
/linux/tools/testing/selftests/drivers/net/netdevsim/
H A Dnexthop.sh213 $IP nexthop add id 10 group 1/2 type resilient buckets 4
229 $IP nexthop add id 10 group 1,3/2,2 type resilient buckets 5
259 $IP nexthop add id 10 group 1/2 type resilient buckets 4 &> /dev/null
325 $IP nexthop add id 10 group 1/2 type resilient buckets 6
353 $IP nexthop add id 10 group 1/2 type resilient buckets 6
408 $IP nexthop add id 10 group 1/2 type resilient buckets 8 idle_timer 4
434 type resilient buckets 8 idle_timer 6
469 $IP nexthop add id 10 group 1/2 type resilient buckets 8 $timer 4
504 $IP nexthop add id 10 group 1/2 type resilient buckets 8 $timer 8
535 type resilient buckets 8 $timer 4
[all …]
/linux/lib/
H A Dhashtable_test.c189 int buckets[2]; in hashtable_test_hash_for_each_possible() local
223 buckets[y->key] = bkt; in hashtable_test_hash_for_each_possible()
230 if (buckets[0] == buckets[1]) { in hashtable_test_hash_for_each_possible()
244 int buckets[2]; in hashtable_test_hash_for_each_possible_safe() local
281 buckets[y->key] = bkt; in hashtable_test_hash_for_each_possible_safe()
288 if (buckets[0] == buckets[1]) { in hashtable_test_hash_for_each_possible_safe()
/linux/net/ceph/crush/
H A Dcrush.c111 if (map->buckets) { in crush_destroy()
114 if (map->buckets[b] == NULL) in crush_destroy()
116 crush_destroy_bucket(map->buckets[b]); in crush_destroy()
118 kfree(map->buckets); in crush_destroy()
H A Dmapper.c531 itemtype = map->buckets[-1-item]->type; in crush_choose_firstn()
544 in = map->buckets[-1-item]; in crush_choose_firstn()
568 map->buckets[-1-item], in crush_choose_firstn()
744 itemtype = map->buckets[-1-item]->type; in crush_choose_indep()
761 in = map->buckets[-1-item]; in crush_choose_indep()
781 map->buckets[-1-item], in crush_choose_indep()
868 if (!map->buckets[b]) in crush_init_workspace()
872 switch (map->buckets[b]->alg) { in crush_init_workspace()
880 v += map->buckets[b]->size * sizeof(__u32); in crush_init_workspace()
951 map->buckets[-1-curstep->arg1])) { in crush_do_rule()
[all …]
/linux/net/netfilter/ipvs/
H A Dip_vs_sh.c70 struct ip_vs_sh_bucket buckets[IP_VS_SH_TAB_SIZE]; member
108 struct ip_vs_dest *dest = rcu_dereference(s->buckets[hash].dest); in ip_vs_sh_get()
130 dest = rcu_dereference(s->buckets[ihash].dest); in ip_vs_sh_get_fallback()
145 dest = rcu_dereference(s->buckets[hash].dest); in ip_vs_sh_get_fallback()
172 b = &s->buckets[0]; in ip_vs_sh_reassign()
216 b = &s->buckets[0]; in ip_vs_sh_flush()
H A Dip_vs_dh.c64 struct ip_vs_dh_bucket buckets[IP_VS_DH_TAB_SIZE]; member
90 return rcu_dereference(s->buckets[ip_vs_dh_hashkey(af, addr)].dest); in ip_vs_dh_get()
106 b = &s->buckets[0]; in ip_vs_dh_reassign()
140 b = &s->buckets[0]; in ip_vs_dh_flush()
/linux/tools/testing/selftests/bpf/prog_tests/
H A Darena_htab.c20 printf("htab %p buckets %p n_buckets %d\n", htab, htab->buckets, htab->n_buckets); in test_arena_htab_common()
21 ASSERT_OK_PTR(htab->buckets, "htab->buckets shouldn't be NULL"); in test_arena_htab_common()
22 for (i = 0; htab->buckets && i < 16; i += 4) { in test_arena_htab_common()
/linux/drivers/md/
H A Ddm-region-hash.c73 struct list_head *buckets; member
209 rh->buckets = vmalloc(array_size(nr_buckets, sizeof(*rh->buckets))); in dm_region_hash_create()
210 if (!rh->buckets) { in dm_region_hash_create()
217 INIT_LIST_HEAD(rh->buckets + i); in dm_region_hash_create()
231 vfree(rh->buckets); in dm_region_hash_create()
247 list_for_each_entry_safe(reg, nreg, rh->buckets + h, in dm_region_hash_destroy()
258 vfree(rh->buckets); in dm_region_hash_destroy()
277 struct list_head *bucket = rh->buckets + rh_hash(rh, region); in __rh_lookup()
288 list_add(&reg->hash_list, rh->buckets + rh_hash(rh, reg->key)); in __rh_insert()
/linux/drivers/net/ethernet/mellanox/mlx5/core/ipoib/
H A Dipoib_vlan.c45 struct hlist_head buckets[1 << MLX5I_MAX_LOG_PKEY_SUP]; member
71 static struct qpn_to_netdev *mlx5i_find_qpn_to_netdev_node(struct hlist_head *buckets, in mlx5i_find_qpn_to_netdev_node() argument
74 struct hlist_head *h = &buckets[hash_32(qpn, MLX5I_MAX_LOG_PKEY_SUP)]; in mlx5i_find_qpn_to_netdev_node()
99 hlist_add_head(&new_node->hlist, &ht->buckets[key]); in mlx5i_pkey_add_qpn()
112 node = mlx5i_find_qpn_to_netdev_node(ht->buckets, qpn); in mlx5i_pkey_del_qpn()
131 node = mlx5i_find_qpn_to_netdev_node(ipriv->qpn_htbl->buckets, qpn); in mlx5i_pkey_get_netdev()
/linux/tools/perf/
H A Dbuiltin-ftrace.c731 static void make_histogram(struct perf_ftrace *ftrace, int buckets[], in make_histogram() argument
807 buckets[i]++; in make_histogram()
819 static void display_histogram(struct perf_ftrace *ftrace, int buckets[]) in display_histogram() argument
830 total += buckets[i]; in display_histogram()
840 bar_len = buckets[0] * bar_total / total; in display_histogram()
844 buckets[0], bar_len, bar, bar_total - bar_len, ""); in display_histogram()
879 bar_len = buckets[i] * bar_total / total; in display_histogram()
880 printf(" %s | %10d | %.*s%*s |\n", unit, buckets[i], bar_len, bar, in display_histogram()
884 bar_len = buckets[NUM_BUCKET - 1] * bar_total / total; in display_histogram()
900 printf(" | %10d | %.*s%*s |\n", buckets[NUM_BUCKET - 1], in display_histogram()
[all …]
/linux/fs/nfs/
H A Dpnfs_nfs.c102 p = kmalloc(struct_size(p, buckets, n), gfp_flags); in pnfs_alloc_commit_array()
109 for (b = &p->buckets[0]; n != 0; b++, n--) { in pnfs_alloc_commit_array()
259 struct pnfs_commit_bucket *buckets, in pnfs_bucket_scan_array() argument
267 cnt = pnfs_bucket_scan_ds_commit_list(&buckets[i], cinfo, max); in pnfs_bucket_scan_array()
288 cnt = pnfs_bucket_scan_array(cinfo, array->buckets, in pnfs_generic_scan_commit_lists()
304 struct pnfs_commit_bucket *buckets, in pnfs_bucket_recover_commit_reqs() argument
314 for (i = 0, b = buckets; i < nbuckets; i++, b++) { in pnfs_bucket_recover_commit_reqs()
343 array->buckets, in pnfs_generic_recover_commit_reqs()
383 static void pnfs_generic_retry_commit(struct pnfs_commit_bucket *buckets, in pnfs_generic_retry_commit() argument
392 for (bucket = buckets; idx < nbuckets; bucket++, idx++) { in pnfs_generic_retry_commit()
[all …]
/linux/kernel/bpf/
H A Dbpf_local_storage.c24 return &smap->buckets[hash_ptr(selem, smap->bucket_log)]; in select_bucket()
787 usage += sizeof(*smap->buckets) * (1ULL << smap->bucket_log); in bpf_local_storage_map_mem_usage()
821 smap->buckets = bpf_map_kvcalloc(&smap->map, nbuckets, in bpf_local_storage_map_alloc()
822 sizeof(*smap->buckets), GFP_USER | __GFP_NOWARN); in bpf_local_storage_map_alloc()
823 if (!smap->buckets) { in bpf_local_storage_map_alloc()
829 INIT_HLIST_HEAD(&smap->buckets[i].list); in bpf_local_storage_map_alloc()
830 raw_spin_lock_init(&smap->buckets[i].lock); in bpf_local_storage_map_alloc()
857 kvfree(smap->buckets); in bpf_local_storage_map_alloc()
890 b = &smap->buckets[i]; in bpf_local_storage_map_free()
928 kvfree(smap->buckets); in bpf_local_storage_map_free()
/linux/drivers/net/ethernet/mellanox/mlx5/core/lag/
H A Dport_sel.c54 ft_attr.max_fte = ldev->ports * ldev->buckets; in mlx5_lag_create_port_sel_table()
81 for (j = 0; j < ldev->buckets; j++) { in mlx5_lag_create_port_sel_table()
84 idx = i * ldev->buckets + j; in mlx5_lag_create_port_sel_table()
96 idx = k * ldev->buckets + j; in mlx5_lag_create_port_sel_table()
99 j = ldev->buckets; in mlx5_lag_create_port_sel_table()
361 for (j = 0; j < ldev->buckets; j++) { in mlx5_lag_destroy_definer()
362 idx = i * ldev->buckets + j; in mlx5_lag_destroy_definer()
592 for (j = 0; j < ldev->buckets; j++) { in __mlx5_lag_modify_definers_destinations()
593 idx = i * ldev->buckets + j; in __mlx5_lag_modify_definers_destinations()
/linux/drivers/net/wireless/broadcom/brcm80211/brcmfmac/
H A Dpno.c298 struct brcmf_gscan_bucket_config **buckets, in brcmf_pno_prep_fwconfig() argument
323 *buckets = NULL; in brcmf_pno_prep_fwconfig()
355 *buckets = fw_buckets; in brcmf_pno_prep_fwconfig()
396 struct brcmf_gscan_bucket_config *buckets; in brcmf_pno_config_sched_scans() local
403 n_buckets = brcmf_pno_prep_fwconfig(pi, &pno_cfg, &buckets, in brcmf_pno_config_sched_scans()
437 memcpy(gscan_cfg->bucket, buckets, in brcmf_pno_config_sched_scans()
438 array_size(n_buckets, sizeof(*buckets))); in brcmf_pno_config_sched_scans()
463 kfree(buckets); in brcmf_pno_config_sched_scans()

12345