Home
last modified time | relevance | path

Searched refs:bucket (Results 1 – 25 of 164) sorted by relevance

1234567

/linux/net/mptcp/
H A Dtoken.c53 /* called with bucket lock held */
66 /* called with bucket lock held */
111 struct token_bucket *bucket; in mptcp_token_new_request() local
122 bucket = token_bucket(token); in mptcp_token_new_request()
123 spin_lock_bh(&bucket->lock); in mptcp_token_new_request()
124 if (__token_bucket_busy(bucket, token)) { in mptcp_token_new_request()
125 spin_unlock_bh(&bucket->lock); in mptcp_token_new_request()
129 hlist_nulls_add_head_rcu(&subflow_req->token_node, &bucket->req_chain); in mptcp_token_new_request()
130 bucket->chain_len++; in mptcp_token_new_request()
131 spin_unlock_bh(&bucket in mptcp_token_new_request()
157 struct token_bucket *bucket; mptcp_token_new_connect() local
196 struct token_bucket *bucket; mptcp_token_accept() local
213 struct token_bucket *bucket; mptcp_token_exists() local
249 struct token_bucket *bucket; mptcp_token_get_sock() local
305 struct token_bucket *bucket = &token_hash[slot]; mptcp_token_iter_next() local
354 struct token_bucket *bucket; mptcp_token_destroy_request() local
378 struct token_bucket *bucket; mptcp_token_destroy() local
[all...]
/linux/drivers/md/dm-vdo/
H A Dpriority-table.c23 struct bucket { struct
44 struct bucket buckets[];
64 struct bucket, __func__, &table); in vdo_make_priority_table()
69 struct bucket *bucket = &table->buckets[priority]; in vdo_make_priority_table() local
71 bucket->priority = priority; in vdo_make_priority_table()
72 INIT_LIST_HEAD(&bucket->queue); in vdo_make_priority_table()
140 static inline void mark_bucket_empty(struct priority_table *table, struct bucket *bucket) in mark_bucket_empty() argument
142 table->search_vector &= ~(1ULL << bucket->priority); in mark_bucket_empty()
157 struct bucket *bucket; in vdo_priority_table_dequeue() local
173 bucket = &table->buckets[top_priority]; in vdo_priority_table_dequeue()
[all …]
/linux/net/ceph/crush/
H A Dmapper.c74 static int bucket_perm_choose(const struct crush_bucket *bucket, in bucket_perm_choose() argument
78 unsigned int pr = r % bucket->size; in bucket_perm_choose()
83 dprintk("bucket %d new x=%d\n", bucket->id, x); in bucket_perm_choose()
88 s = crush_hash32_3(bucket->hash, x, bucket->id, 0) % in bucket_perm_choose()
89 bucket->size; in bucket_perm_choose()
95 for (i = 0; i < bucket->size; i++) in bucket_perm_choose()
100 for (i = 1; i < bucket->size; i++) in bucket_perm_choose()
112 if (p < bucket->size - 1) { in bucket_perm_choose()
113 i = crush_hash32_3(bucket->hash, x, bucket->id, p) % in bucket_perm_choose()
114 (bucket->size - p); in bucket_perm_choose()
[all …]
/linux/drivers/interconnect/qcom/
H A Dbcm-voter.c65 int bucket, i; in bcm_aggregate_mask() local
67 for (bucket = 0; bucket < QCOM_ICC_NUM_BUCKETS; bucket++) { in bcm_aggregate_mask()
68 bcm->vote_x[bucket] = 0; in bcm_aggregate_mask()
69 bcm->vote_y[bucket] = 0; in bcm_aggregate_mask()
75 if (node->sum_avg[bucket] || node->max_peak[bucket]) { in bcm_aggregate_mask()
76 bcm->vote_x[bucket] = 0; in bcm_aggregate_mask()
77 bcm->vote_y[bucket] = bcm->enable_mask; in bcm_aggregate_mask()
94 size_t i, bucket; in bcm_aggregate() local
99 for (bucket = 0; bucket < QCOM_ICC_NUM_BUCKETS; bucket++) { in bcm_aggregate()
102 temp = bcm_div(node->sum_avg[bucket] * bcm->aux_data.width, in bcm_aggregate()
[all …]
/linux/block/
H A Dblk-stat.c55 int bucket, cpu; in blk_stat_add() local
66 bucket = cb->bucket_fn(rq); in blk_stat_add()
67 if (bucket < 0) in blk_stat_add()
70 stat = &per_cpu_ptr(cb->cpu_stat, cpu)[bucket]; in blk_stat_add()
80 unsigned int bucket; in blk_stat_timer_fn() local
83 for (bucket = 0; bucket < cb->buckets; bucket++) in blk_stat_timer_fn()
84 blk_rq_stat_init(&cb->stat[bucket]); in blk_stat_timer_fn()
90 for (bucket = 0; bucket < cb->buckets; bucket++) { in blk_stat_timer_fn()
91 blk_rq_stat_sum(&cb->stat[bucket], &cpu_stat[bucket]); in blk_stat_timer_fn()
92 blk_rq_stat_init(&cpu_stat[bucket]); in blk_stat_timer_fn()
[all …]
/linux/net/sched/
H A Dsch_hhf.c329 static struct sk_buff *dequeue_head(struct wdrr_bucket *bucket) in dequeue_head() argument
331 struct sk_buff *skb = bucket->head; in dequeue_head()
333 bucket->head = skb->next; in dequeue_head()
339 static void bucket_add(struct wdrr_bucket *bucket, struct sk_buff *skb) in bucket_add() argument
341 if (bucket->head == NULL) in bucket_add()
342 bucket->head = skb; in bucket_add()
344 bucket->tail->next = skb; in bucket_add()
345 bucket->tail = skb; in bucket_add()
352 struct wdrr_bucket *bucket; in hhf_drop() local
355 bucket = &q->buckets[WDRR_BUCKET_FOR_HH]; in hhf_drop()
[all …]
/linux/tools/tracing/rtla/src/
H A Dtimerlat.bpf.c86 int bucket) in update_main_hist()
92 if (bucket >= entries) in update_summary()
96 map_increment(map, bucket); in update_summary()
101 int bucket) in update_summary()
109 if (bucket >= entries) in update_summary()
144 int bucket; in handle_timerlat_sample()
151 bucket = latency / bucket_size; in handle_timerlat_sample()
154 update_main_hist(&hist_irq, bucket); in handle_timerlat_sample()
155 update_summary(&summary_irq, latency, bucket); in handle_timerlat_sample()
160 update_main_hist(&hist_thread, bucket);
75 update_main_hist(void * map,int bucket) update_main_hist() argument
90 update_summary(void * map,unsigned long long latency,int bucket) update_summary() argument
127 int bucket; handle_timerlat_sample() local
[all...]
/linux/drivers/infiniband/sw/rdmavt/
H A Dtrace_qp.h18 TP_PROTO(struct rvt_qp *qp, u32 bucket),
19 TP_ARGS(qp, bucket),
23 __field(u32, bucket)
28 __entry->bucket = bucket;
34 __entry->bucket
39 TP_PROTO(struct rvt_qp *qp, u32 bucket),
40 TP_ARGS(qp, bucket));
43 TP_PROTO(struct rvt_qp *qp, u32 bucket),
44 TP_ARGS(qp, bucket));
/linux/net/vmw_vsock/
H A Ddiag.c52 unsigned int bucket; in vsock_diag_dump() local
63 bucket = cb->args[1]; in vsock_diag_dump()
72 while (bucket < ARRAY_SIZE(vsock_bind_table)) { in vsock_diag_dump()
73 struct list_head *head = &vsock_bind_table[bucket]; in vsock_diag_dump()
94 bucket++; in vsock_diag_dump()
98 bucket = 0; in vsock_diag_dump()
102 while (bucket < ARRAY_SIZE(vsock_connected_table)) { in vsock_diag_dump()
103 struct list_head *head = &vsock_connected_table[bucket]; in vsock_diag_dump()
128 bucket++; in vsock_diag_dump()
135 cb->args[1] = bucket; in vsock_diag_dump()
/linux/kernel/dma/
H A Ddebug.c268 static void put_hash_bucket(struct hash_bucket *bucket, in put_hash_bucket() argument
270 __releases(&bucket->lock) in put_hash_bucket()
272 spin_unlock_irqrestore(&bucket->lock, flags); in put_hash_bucket()
297 static struct dma_debug_entry *__hash_bucket_find(struct hash_bucket *bucket, in __hash_bucket_find() argument
304 list_for_each_entry(entry, &bucket->list, list) { in __hash_bucket_find()
347 static struct dma_debug_entry *bucket_find_exact(struct hash_bucket *bucket, in bucket_find_exact() argument
350 return __hash_bucket_find(bucket, ref, exact_match); in bucket_find_exact()
353 static struct dma_debug_entry *bucket_find_contain(struct hash_bucket **bucket, in bucket_find_contain() argument
362 entry = __hash_bucket_find(*bucket, ref, containing_match); in bucket_find_contain()
370 put_hash_bucket(*bucket, *flags); in bucket_find_contain()
[all …]
/linux/fs/afs/
H A Ddir_search.c25 int bucket; in afs_dir_hash_name() local
29 bucket = hash & (AFS_DIR_HASHTBL_SIZE - 1); in afs_dir_hash_name()
31 bucket = AFS_DIR_HASHTBL_SIZE - bucket; in afs_dir_hash_name()
32 bucket &= (AFS_DIR_HASHTBL_SIZE - 1); in afs_dir_hash_name()
34 return bucket; in afs_dir_hash_name()
60 iter->bucket = afs_dir_hash_name(name); in afs_dir_init_iter()
134 entry = ntohs(meta->meta.hashtable[iter->bucket & (AFS_DIR_HASHTBL_SIZE - 1)]); in afs_dir_search_bucket()
135 _enter("%x,%x", iter->bucket, entry); in afs_dir_search_bucket()
148 iter->bucket, resv, slot, slot + iter->nr_slots - 1); in afs_dir_search_bucket()
169 kdebug("dir chain loop h=%x", iter->bucket); in afs_dir_search_bucket()
/linux/drivers/cpuidle/governors/
H A Dmenu.c75 unsigned int bucket; member
83 int bucket = 0; in which_bucket() local
86 return bucket; in which_bucket()
88 return bucket + 1; in which_bucket()
90 return bucket + 2; in which_bucket()
92 return bucket + 3; in which_bucket()
94 return bucket + 4; in which_bucket()
95 return bucket + 5; in which_bucket()
253 data->bucket = which_bucket(data->next_timer_ns); in menu_select()
258 data->correction_factor[data->bucket], in menu_select()
[all …]
/linux/fs/nfs/
H A Dpnfs_nfs.c65 pnfs_free_bucket_lseg(struct pnfs_commit_bucket *bucket) in pnfs_free_bucket_lseg() argument
67 if (list_empty(&bucket->committing) && list_empty(&bucket->written)) { in pnfs_free_bucket_lseg()
68 struct pnfs_layout_segment *freeme = bucket->lseg; in pnfs_free_bucket_lseg()
69 bucket->lseg = NULL; in pnfs_free_bucket_lseg()
83 struct pnfs_commit_bucket *bucket = NULL; in pnfs_generic_clear_request_commit() local
89 bucket = list_first_entry(&req->wb_list, in pnfs_generic_clear_request_commit()
93 if (bucket) in pnfs_generic_clear_request_commit()
94 pnfs_put_lseg(pnfs_free_bucket_lseg(bucket)); in pnfs_generic_clear_request_commit()
243 pnfs_bucket_scan_ds_commit_list(struct pnfs_commit_bucket *bucket, in pnfs_bucket_scan_ds_commit_list() argument
247 struct list_head *src = &bucket->written; in pnfs_bucket_scan_ds_commit_list()
[all …]
/linux/fs/ocfs2/
H A Dxattr.c125 struct ocfs2_xattr_bucket *bucket; member
279 struct ocfs2_xattr_bucket *bucket,
301 struct ocfs2_xattr_bucket *bucket,
322 struct ocfs2_xattr_bucket *bucket; in ocfs2_xattr_bucket_new() local
327 bucket = kzalloc_obj(struct ocfs2_xattr_bucket, GFP_NOFS); in ocfs2_xattr_bucket_new()
328 if (bucket) { in ocfs2_xattr_bucket_new()
329 bucket->bu_inode = inode; in ocfs2_xattr_bucket_new()
330 bucket->bu_blocks = blks; in ocfs2_xattr_bucket_new()
333 return bucket; in ocfs2_xattr_bucket_new()
336 static void ocfs2_xattr_bucket_relse(struct ocfs2_xattr_bucket *bucket) in ocfs2_xattr_bucket_relse() argument
[all …]
/linux/drivers/md/
H A Ddm-clone-target.c564 #define bucket_lock_irqsave(bucket, flags) \ argument
565 spin_lock_irqsave(&(bucket)->lock, flags)
567 #define bucket_unlock_irqrestore(bucket, flags) \ argument
568 spin_unlock_irqrestore(&(bucket)->lock, flags)
570 #define bucket_lock_irq(bucket) \ argument
571 spin_lock_irq(&(bucket)->lock)
573 #define bucket_unlock_irq(bucket) \ argument
574 spin_unlock_irq(&(bucket)->lock)
579 struct hash_table_bucket *bucket; in hash_table_init() local
588 bucket = clone->ht + i; in hash_table_init()
[all …]
/linux/Documentation/userspace-api/media/v4l/
H A Dmetafmt-vsp1-hgt.rst29 The Saturation position **n** (0 - 31) of the bucket in the matrix is
34 The Hue position **m** (0 - 5) of the bucket in the matrix depends on
102 - :cspan:`4` Histogram bucket (m=0, n=0) [31:0]
104 - :cspan:`4` Histogram bucket (m=0, n=1) [31:0]
108 - :cspan:`4` Histogram bucket (m=0, n=31) [31:0]
110 - :cspan:`4` Histogram bucket (m=1, n=0) [31:0]
114 - :cspan:`4` Histogram bucket (m=2, n=0) [31:0]
118 - :cspan:`4` Histogram bucket (m=3, n=0) [31:0]
122 - :cspan:`4` Histogram bucket (m=4, n=0) [31:0]
126 - :cspan:`4` Histogram bucket (m=5, n=0) [31:0]
[all …]
/linux/drivers/md/bcache/
H A Dalloc.c73 uint8_t bch_inc_gen(struct cache *ca, struct bucket *b) in bch_inc_gen()
86 struct bucket *b; in bch_rescale_priorities()
121 static inline bool can_inc_bucket_gen(struct bucket *b) in can_inc_bucket_gen()
126 bool bch_can_invalidate_bucket(struct cache *ca, struct bucket *b) in bch_can_invalidate_bucket()
133 void __bch_invalidate_one_bucket(struct cache *ca, struct bucket *b) in __bch_invalidate_one_bucket()
147 static void bch_invalidate_one_bucket(struct cache *ca, struct bucket *b) in bch_invalidate_one_bucket()
175 struct bucket *b; in invalidate_buckets_lru()
212 struct bucket *b; in invalidate_buckets_fifo()
235 struct bucket *b; in invalidate_buckets_random()
296 static int bch_allocator_push(struct cache *ca, long bucket) in bch_allocator_push() argument
[all …]
H A Dbcache.h197 struct bucket { struct
211 BITMASK(GC_MARK, struct bucket, gc_mark, 0, 2);
217 BITMASK(GC_SECTORS_USED, struct bucket, gc_mark, 2, GC_SECTORS_USED_SIZE);
218 BITMASK(GC_MOVE, struct bucket, gc_mark, 15, 1);
460 struct bucket *buckets;
462 DECLARE_HEAP(struct bucket *, heap);
827 static inline struct bucket *PTR_BUCKET(struct cache_set *c, in PTR_BUCKET()
923 static inline uint8_t bucket_gc_gen(struct bucket *b) in bucket_gc_gen()
987 uint8_t bch_inc_gen(struct cache *ca, struct bucket *b);
990 bool bch_can_invalidate_bucket(struct cache *ca, struct bucket *b);
[all …]
/linux/Documentation/networking/
H A Dnexthop-group-resilient.rst49 to choose a hash table bucket, then reads the next hop that this bucket
83 cause bucket allocation change, the wants counts for individual next hops
91 Each bucket maintains a last-used timer. Every time a packet is forwarded
92 through a bucket, this timer is updated to current jiffies value. One
94 amount of time that a bucket must not be hit by traffic in order for it to
104 upkeep changes the next hop that the bucket references to one of the
135 - Single-bucket notifications of the type
143 Some single-bucket notifications are forced, as indicated by the "force"
145 hop associated with the bucket was removed, and the bucket really must be
150 bucket should be migrated, but the HW discovers that the bucket has in fact
[all …]
/linux/tools/testing/selftests/drivers/net/hw/
H A Dethtool_rmon.sh35 local bucket=$1; shift
50 jq -r ".[0].rmon[\"${set}-pktsNtoM\"][$bucket].val")
58 jq -r ".[0].rmon[\"${set}-pktsNtoM\"][$bucket].val")
78 while read -r -a bucket; do
82 if ! ensure_mtu $if ${bucket[0]}; then
88 if ! bucket_test $iface $neigh $set $nbuckets ${bucket[0]}; then
/linux/include/trace/events/
H A Dbcache.h68 __field(size_t, bucket )
72 __entry->bucket = PTR_BUCKET_NR(b->c, &b->key, 0);
75 TP_printk("bucket %zu", __entry->bucket)
267 __field(size_t, bucket )
273 __entry->bucket = PTR_BUCKET_NR(b->c, &b->key, 0);
279 __entry->bucket, __entry->block, __entry->keys)
370 __field(size_t, bucket )
375 __entry->bucket = PTR_BUCKET_NR(b->c, &b->key, 0);
379 TP_printk("bucket %zu keys %u", __entry->bucket, __entry->keys)
429 TP_PROTO(struct cache *ca, size_t bucket),
[all …]
/linux/net/rxrpc/
H A Dproc.c316 unsigned int bucket, n; in rxrpc_peer_seq_start() local
326 bucket = *_pos >> shift; in rxrpc_peer_seq_start()
328 if (bucket >= HASH_SIZE(rxnet->peer_hash)) { in rxrpc_peer_seq_start()
333 if (bucket == 0) in rxrpc_peer_seq_start()
339 p = seq_hlist_start_rcu(&rxnet->peer_hash[bucket], n - 1); in rxrpc_peer_seq_start()
342 bucket++; in rxrpc_peer_seq_start()
344 *_pos = (bucket << shift) | n; in rxrpc_peer_seq_start()
351 unsigned int bucket, n; in rxrpc_peer_seq_next() local
358 bucket = *_pos >> shift; in rxrpc_peer_seq_next()
360 p = seq_hlist_next_rcu(v, &rxnet->peer_hash[bucket], _pos); in rxrpc_peer_seq_next()
[all …]
/linux/arch/sparc/kernel/
H A Dirq_64.c206 struct ino_bucket bucket; member
257 struct ino_bucket *bucket; in cookie_exists() local
268 bucket = (struct ino_bucket *) __va(cookie); in cookie_exists()
269 irq = bucket->__irq; in cookie_exists()
278 struct ino_bucket *bucket; in sysino_exists() local
281 bucket = &ivector_table[sysino]; in sysino_exists()
282 irq = bucket_get_irq(__pa(bucket)); in sysino_exists()
611 struct ino_bucket *bucket; in build_irq() local
618 bucket = &ivector_table[ino]; in build_irq()
619 irq = bucket_get_irq(__pa(bucket)); in build_irq()
[all …]
/linux/net/atm/
H A Dproc.c69 int bucket; member
78 static int __vcc_walk(struct sock **sock, int family, int *bucket, loff_t l) in __vcc_walk() argument
83 for (*bucket = 0; *bucket < VCC_HTABLE_SIZE; ++*bucket) { in __vcc_walk()
84 struct hlist_head *head = &vcc_hash[*bucket]; in __vcc_walk()
98 if (!sk && ++*bucket < VCC_HTABLE_SIZE) { in __vcc_walk()
99 sk = sk_head(&vcc_hash[*bucket]); in __vcc_walk()
113 return __vcc_walk(&state->sk, family, &state->bucket, l) ? in vcc_walk()
/linux/net/ipv4/
H A Dnexthop.c180 struct nh_res_bucket *bucket = &res_table->nh_buckets[i]; in nh_notifier_res_table_info_init() local
184 nhge = rtnl_dereference(bucket->nh_entry); in nh_notifier_res_table_info_init()
1093 static unsigned long nh_res_bucket_used_time(const struct nh_res_bucket *bucket) in nh_res_bucket_used_time() argument
1095 return (unsigned long)atomic_long_read(&bucket->used_time); in nh_res_bucket_used_time()
1100 const struct nh_res_bucket *bucket, in nh_res_bucket_idle_point() argument
1103 unsigned long time = nh_res_bucket_used_time(bucket); in nh_res_bucket_idle_point()
1106 if (time == bucket->migrated_time) in nh_res_bucket_idle_point()
1119 struct nh_res_bucket *bucket) in nh_res_bucket_set_idle() argument
1123 atomic_long_set(&bucket->used_time, (long)now); in nh_res_bucket_set_idle()
1124 bucket->migrated_time = now; in nh_res_bucket_set_idle()
[all …]

1234567