| /linux/net/mptcp/ |
| H A D | token.c | 53 /* called with bucket lock held */ 66 /* called with bucket lock held */ 111 struct token_bucket *bucket; in mptcp_token_new_request() local 122 bucket = token_bucket(token); in mptcp_token_new_request() 123 spin_lock_bh(&bucket->lock); in mptcp_token_new_request() 124 if (__token_bucket_busy(bucket, token)) { in mptcp_token_new_request() 125 spin_unlock_bh(&bucket->lock); in mptcp_token_new_request() 129 hlist_nulls_add_head_rcu(&subflow_req->token_node, &bucket->req_chain); in mptcp_token_new_request() 130 bucket->chain_len++; in mptcp_token_new_request() 131 spin_unlock_bh(&bucket in mptcp_token_new_request() 157 struct token_bucket *bucket; mptcp_token_new_connect() local 196 struct token_bucket *bucket; mptcp_token_accept() local 213 struct token_bucket *bucket; mptcp_token_exists() local 249 struct token_bucket *bucket; mptcp_token_get_sock() local 305 struct token_bucket *bucket = &token_hash[slot]; mptcp_token_iter_next() local 354 struct token_bucket *bucket; mptcp_token_destroy_request() local 378 struct token_bucket *bucket; mptcp_token_destroy() local [all...] |
| /linux/drivers/md/dm-vdo/ |
| H A D | priority-table.c | 23 struct bucket { struct 44 struct bucket buckets[]; 64 struct bucket, __func__, &table); in vdo_make_priority_table() 69 struct bucket *bucket = &table->buckets[priority]; in vdo_make_priority_table() local 71 bucket->priority = priority; in vdo_make_priority_table() 72 INIT_LIST_HEAD(&bucket->queue); in vdo_make_priority_table() 140 static inline void mark_bucket_empty(struct priority_table *table, struct bucket *bucket) in mark_bucket_empty() argument 142 table->search_vector &= ~(1ULL << bucket->priority); in mark_bucket_empty() 157 struct bucket *bucket; in vdo_priority_table_dequeue() local 173 bucket = &table->buckets[top_priority]; in vdo_priority_table_dequeue() [all …]
|
| /linux/net/ceph/crush/ |
| H A D | mapper.c | 74 static int bucket_perm_choose(const struct crush_bucket *bucket, in bucket_perm_choose() argument 78 unsigned int pr = r % bucket->size; in bucket_perm_choose() 83 dprintk("bucket %d new x=%d\n", bucket->id, x); in bucket_perm_choose() 88 s = crush_hash32_3(bucket->hash, x, bucket->id, 0) % in bucket_perm_choose() 89 bucket->size; in bucket_perm_choose() 95 for (i = 0; i < bucket->size; i++) in bucket_perm_choose() 100 for (i = 1; i < bucket->size; i++) in bucket_perm_choose() 112 if (p < bucket->size - 1) { in bucket_perm_choose() 113 i = crush_hash32_3(bucket->hash, x, bucket->id, p) % in bucket_perm_choose() 114 (bucket->size - p); in bucket_perm_choose() [all …]
|
| /linux/drivers/interconnect/qcom/ |
| H A D | bcm-voter.c | 65 int bucket, i; in bcm_aggregate_mask() local 67 for (bucket = 0; bucket < QCOM_ICC_NUM_BUCKETS; bucket++) { in bcm_aggregate_mask() 68 bcm->vote_x[bucket] = 0; in bcm_aggregate_mask() 69 bcm->vote_y[bucket] = 0; in bcm_aggregate_mask() 75 if (node->sum_avg[bucket] || node->max_peak[bucket]) { in bcm_aggregate_mask() 76 bcm->vote_x[bucket] = 0; in bcm_aggregate_mask() 77 bcm->vote_y[bucket] = bcm->enable_mask; in bcm_aggregate_mask() 94 size_t i, bucket; in bcm_aggregate() local 99 for (bucket = 0; bucket < QCOM_ICC_NUM_BUCKETS; bucket++) { in bcm_aggregate() 102 temp = bcm_div(node->sum_avg[bucket] * bcm->aux_data.width, in bcm_aggregate() [all …]
|
| /linux/block/ |
| H A D | blk-stat.c | 55 int bucket, cpu; in blk_stat_add() local 66 bucket = cb->bucket_fn(rq); in blk_stat_add() 67 if (bucket < 0) in blk_stat_add() 70 stat = &per_cpu_ptr(cb->cpu_stat, cpu)[bucket]; in blk_stat_add() 80 unsigned int bucket; in blk_stat_timer_fn() local 83 for (bucket = 0; bucket < cb->buckets; bucket++) in blk_stat_timer_fn() 84 blk_rq_stat_init(&cb->stat[bucket]); in blk_stat_timer_fn() 90 for (bucket = 0; bucket < cb->buckets; bucket++) { in blk_stat_timer_fn() 91 blk_rq_stat_sum(&cb->stat[bucket], &cpu_stat[bucket]); in blk_stat_timer_fn() 92 blk_rq_stat_init(&cpu_stat[bucket]); in blk_stat_timer_fn() [all …]
|
| /linux/tools/tracing/rtla/src/ |
| H A D | timerlat.bpf.c | 86 int bucket) in update_main_hist() 92 if (bucket >= entries) in update_summary() 96 map_increment(map, bucket); in update_summary() 101 int bucket) in update_summary() 109 if (bucket >= entries) in update_summary() 144 int bucket; in handle_timerlat_sample() 151 bucket = latency / bucket_size; in handle_timerlat_sample() 154 update_main_hist(&hist_irq, bucket); in handle_timerlat_sample() 155 update_summary(&summary_irq, latency, bucket); in handle_timerlat_sample() 160 update_main_hist(&hist_thread, bucket); 75 update_main_hist(void * map,int bucket) update_main_hist() argument 90 update_summary(void * map,unsigned long long latency,int bucket) update_summary() argument 127 int bucket; handle_timerlat_sample() local [all...] |
| /linux/net/sched/ |
| H A D | sch_hhf.c | 329 static struct sk_buff *dequeue_head(struct wdrr_bucket *bucket) in dequeue_head() argument 331 struct sk_buff *skb = bucket->head; in dequeue_head() 333 bucket->head = skb->next; in dequeue_head() 339 static void bucket_add(struct wdrr_bucket *bucket, struct sk_buff *skb) in bucket_add() argument 341 if (bucket->head == NULL) in bucket_add() 342 bucket->head = skb; in bucket_add() 344 bucket->tail->next = skb; in bucket_add() 345 bucket->tail = skb; in bucket_add() 352 struct wdrr_bucket *bucket; in hhf_drop() local 355 bucket = &q->buckets[WDRR_BUCKET_FOR_HH]; in hhf_drop() [all …]
|
| /linux/drivers/infiniband/sw/rdmavt/ |
| H A D | trace_qp.h | 18 TP_PROTO(struct rvt_qp *qp, u32 bucket), 19 TP_ARGS(qp, bucket), 23 __field(u32, bucket) 28 __entry->bucket = bucket; 34 __entry->bucket 39 TP_PROTO(struct rvt_qp *qp, u32 bucket), 40 TP_ARGS(qp, bucket)); 43 TP_PROTO(struct rvt_qp *qp, u32 bucket), 44 TP_ARGS(qp, bucket));
|
| /linux/net/vmw_vsock/ |
| H A D | diag.c | 52 unsigned int bucket; in vsock_diag_dump() local 63 bucket = cb->args[1]; in vsock_diag_dump() 72 while (bucket < ARRAY_SIZE(vsock_bind_table)) { in vsock_diag_dump() 73 struct list_head *head = &vsock_bind_table[bucket]; in vsock_diag_dump() 94 bucket++; in vsock_diag_dump() 98 bucket = 0; in vsock_diag_dump() 102 while (bucket < ARRAY_SIZE(vsock_connected_table)) { in vsock_diag_dump() 103 struct list_head *head = &vsock_connected_table[bucket]; in vsock_diag_dump() 128 bucket++; in vsock_diag_dump() 135 cb->args[1] = bucket; in vsock_diag_dump()
|
| /linux/fs/afs/ |
| H A D | dir_search.c | 25 int bucket; in afs_dir_hash_name() local 29 bucket = hash & (AFS_DIR_HASHTBL_SIZE - 1); in afs_dir_hash_name() 31 bucket = AFS_DIR_HASHTBL_SIZE - bucket; in afs_dir_hash_name() 32 bucket &= (AFS_DIR_HASHTBL_SIZE - 1); in afs_dir_hash_name() 34 return bucket; in afs_dir_hash_name() 60 iter->bucket = afs_dir_hash_name(name); in afs_dir_init_iter() 134 entry = ntohs(meta->meta.hashtable[iter->bucket & (AFS_DIR_HASHTBL_SIZE - 1)]); in afs_dir_search_bucket() 135 _enter("%x,%x", iter->bucket, entry); in afs_dir_search_bucket() 148 iter->bucket, resv, slot, slot + iter->nr_slots - 1); in afs_dir_search_bucket() 169 kdebug("dir chain loop h=%x", iter->bucket); in afs_dir_search_bucket()
|
| /linux/drivers/cpuidle/governors/ |
| H A D | menu.c | 75 unsigned int bucket; member 83 int bucket = 0; in which_bucket() local 86 return bucket; in which_bucket() 88 return bucket + 1; in which_bucket() 90 return bucket + 2; in which_bucket() 92 return bucket + 3; in which_bucket() 94 return bucket + 4; in which_bucket() 95 return bucket + 5; in which_bucket() 253 data->bucket = which_bucket(data->next_timer_ns); in menu_select() 258 data->correction_factor[data->bucket], in menu_select() [all...] |
| /linux/kernel/dma/ |
| H A D | debug.c | 251 * Request exclusive access to a hash bucket for a given dma_debug_entry. in get_hash_bucket() 266 * Give up exclusive access to the hash bucket in put_hash_bucket() argument 268 static void put_hash_bucket(struct hash_bucket *bucket, in put_hash_bucket() 270 __releases(&bucket->lock) in put_hash_bucket() 272 spin_unlock_irqrestore(&bucket->lock, flags); 295 * Search a given entry in the hash bucket list in __hash_bucket_find() argument 297 static struct dma_debug_entry *__hash_bucket_find(struct hash_bucket *bucket, in __hash_bucket_find() 304 list_for_each_entry(entry, &bucket->list, list) { in __hash_bucket_find() 347 static struct dma_debug_entry *bucket_find_exact(struct hash_bucket *bucket, in bucket_find_exact() 350 return __hash_bucket_find(bucket, re in bucket_find_exact() 345 bucket_find_exact(struct hash_bucket * bucket,struct dma_debug_entry * ref) bucket_find_exact() argument 351 bucket_find_contain(struct hash_bucket ** bucket,struct dma_debug_entry * ref,unsigned long * flags) bucket_find_contain() argument 379 hash_bucket_add(struct hash_bucket * bucket,struct dma_debug_entry * entry) hash_bucket_add() argument 525 struct hash_bucket *bucket = &dma_entry_hash[idx]; debug_dma_dump_mappings() local 557 struct hash_bucket *bucket = &dma_entry_hash[idx]; dump_show() local 586 struct hash_bucket *bucket; add_dma_entry() local 962 struct hash_bucket *bucket; check_unmap() local 1101 struct hash_bucket *bucket; check_sync() local 1248 struct hash_bucket *bucket; debug_dma_mapping_error() local 1339 struct hash_bucket *bucket; get_nr_mapped_entries() local [all...] |
| /linux/fs/nfs/ |
| H A D | pnfs_nfs.c | 65 pnfs_free_bucket_lseg(struct pnfs_commit_bucket *bucket) in pnfs_free_bucket_lseg() argument 67 if (list_empty(&bucket->committing) && list_empty(&bucket->written)) { in pnfs_free_bucket_lseg() 68 struct pnfs_layout_segment *freeme = bucket->lseg; in pnfs_free_bucket_lseg() 69 bucket->lseg = NULL; in pnfs_free_bucket_lseg() 83 struct pnfs_commit_bucket *bucket = NULL; in pnfs_generic_clear_request_commit() local 89 bucket = list_first_entry(&req->wb_list, in pnfs_generic_clear_request_commit() 93 if (bucket) in pnfs_generic_clear_request_commit() 94 pnfs_put_lseg(pnfs_free_bucket_lseg(bucket)); in pnfs_generic_clear_request_commit() 243 pnfs_bucket_scan_ds_commit_list(struct pnfs_commit_bucket *bucket, in pnfs_bucket_scan_ds_commit_list() argument 247 struct list_head *src = &bucket->written; in pnfs_bucket_scan_ds_commit_list() [all …]
|
| /linux/Documentation/userspace-api/media/v4l/ |
| H A D | metafmt-vsp1-hgt.rst | 29 The Saturation position **n** (0 - 31) of the bucket in the matrix is 34 The Hue position **m** (0 - 5) of the bucket in the matrix depends on 102 - :cspan:`4` Histogram bucket (m=0, n=0) [31:0] 104 - :cspan:`4` Histogram bucket (m=0, n=1) [31:0] 108 - :cspan:`4` Histogram bucket (m=0, n=31) [31:0] 110 - :cspan:`4` Histogram bucket (m=1, n=0) [31:0] 114 - :cspan:`4` Histogram bucket (m=2, n=0) [31:0] 118 - :cspan:`4` Histogram bucket (m=3, n=0) [31:0] 122 - :cspan:`4` Histogram bucket (m=4, n=0) [31:0] 126 - :cspan:`4` Histogram bucket (m=5, n=0) [31:0] [all …]
|
| /linux/fs/ocfs2/ |
| H A D | xattr.c | 64 /* The actual buffers that make up the bucket */ 67 /* How many blocks make up one bucket for this filesystem */ 125 struct ocfs2_xattr_bucket *bucket; member 279 struct ocfs2_xattr_bucket *bucket, 301 struct ocfs2_xattr_bucket *bucket, 322 struct ocfs2_xattr_bucket *bucket; in ocfs2_xattr_bucket_new() local 327 bucket = kzalloc(sizeof(struct ocfs2_xattr_bucket), GFP_NOFS); in ocfs2_xattr_bucket_new() 328 if (bucket) { in ocfs2_xattr_bucket_new() 329 bucket->bu_inode = inode; in ocfs2_xattr_bucket_new() 330 bucket in ocfs2_xattr_bucket_new() 336 ocfs2_xattr_bucket_relse(struct ocfs2_xattr_bucket * bucket) ocfs2_xattr_bucket_relse() argument 346 ocfs2_xattr_bucket_free(struct ocfs2_xattr_bucket * bucket) ocfs2_xattr_bucket_free() argument 361 ocfs2_init_xattr_bucket(struct ocfs2_xattr_bucket * bucket,u64 xb_blkno,int new) ocfs2_init_xattr_bucket() argument 394 ocfs2_read_xattr_bucket(struct ocfs2_xattr_bucket * bucket,u64 xb_blkno) ocfs2_read_xattr_bucket() argument 419 ocfs2_xattr_bucket_journal_access(handle_t * handle,struct ocfs2_xattr_bucket * bucket,int type) ocfs2_xattr_bucket_journal_access() argument 438 ocfs2_xattr_bucket_journal_dirty(handle_t * handle,struct ocfs2_xattr_bucket * bucket) ocfs2_xattr_bucket_journal_dirty() argument 1740 struct ocfs2_xattr_bucket *bucket = loc->xl_storage; ocfs2_xa_bucket_journal_access() local 1748 struct ocfs2_xattr_bucket *bucket = loc->xl_storage; ocfs2_xa_bucket_journal_dirty() local 1756 struct ocfs2_xattr_bucket *bucket = loc->xl_storage; ocfs2_xa_bucket_offset_pointer() local 1775 struct ocfs2_xattr_bucket *bucket = loc->xl_storage; ocfs2_xa_bucket_get_free_start() local 1897 struct ocfs2_xattr_bucket *bucket = loc->xl_storage; ocfs2_xa_bucket_fill_value_buf() local 2338 ocfs2_init_xattr_bucket_xa_loc(struct ocfs2_xa_loc * loc,struct ocfs2_xattr_bucket * bucket,struct ocfs2_xattr_entry * entry) ocfs2_init_xattr_bucket_xa_loc() argument 3781 ocfs2_find_xe_in_bucket(struct inode * inode,struct ocfs2_xattr_bucket * bucket,int name_index,const char * name,u32 name_hash,u16 * xe_index,int * found) ocfs2_find_xe_in_bucket() argument 3858 int low_bucket = 0, bucket, high_bucket; ocfs2_xattr_bucket_find() local 4008 struct ocfs2_xattr_bucket *bucket; ocfs2_iterate_xattr_buckets() local 4078 ocfs2_list_xattr_bucket(struct inode * inode,struct ocfs2_xattr_bucket * bucket,void * para) ocfs2_list_xattr_bucket() argument 4209 ocfs2_cp_xattr_block_to_bucket(struct inode * inode,struct buffer_head * xb_bh,struct ocfs2_xattr_bucket * bucket) ocfs2_cp_xattr_block_to_bucket() argument 4410 ocfs2_defrag_xattr_bucket(struct inode * inode,handle_t * handle,struct ocfs2_xattr_bucket * bucket) ocfs2_defrag_xattr_bucket() argument 5378 ocfs2_xattr_bucket_value_truncate(struct inode * inode,struct ocfs2_xattr_bucket * bucket,int xe_off,int len,struct ocfs2_xattr_set_ctxt * ctxt) ocfs2_xattr_bucket_value_truncate() argument 5542 ocfs2_check_xattr_bucket_collision(struct inode * inode,struct ocfs2_xattr_bucket * bucket,const char * name) ocfs2_check_xattr_bucket_collision() argument 5677 ocfs2_delete_xattr_in_bucket(struct inode * inode,struct ocfs2_xattr_bucket * bucket,void * para) ocfs2_delete_xattr_in_bucket() argument 5753 struct ocfs2_xattr_bucket *bucket = ocfs2_xattr_bucket_post_refcount() local 5799 struct ocfs2_xattr_bucket *bucket = NULL; ocfs2_prepare_refcount_xattr() local 6024 ocfs2_get_xattr_tree_value_root(struct super_block * sb,struct ocfs2_xattr_bucket * bucket,int offset,struct ocfs2_xattr_value_root ** xv,struct buffer_head ** bh) ocfs2_get_xattr_tree_value_root() argument 6060 ocfs2_xattr_bucket_value_refcount(struct inode * inode,struct ocfs2_xattr_bucket * bucket,void * para) ocfs2_xattr_bucket_value_refcount() argument 6706 struct ocfs2_xattr_bucket *bucket; ocfs2_get_reflink_xattr_value_root() local 6731 struct ocfs2_xattr_bucket *bucket = ocfs2_value_tree_metas_in_bucket() local 6739 ocfs2_calc_value_tree_metas(struct inode * inode,struct ocfs2_xattr_bucket * bucket,void * para) ocfs2_calc_value_tree_metas() argument [all...] |
| /linux/tools/testing/selftests/drivers/net/hw/ |
| H A D | ethtool_rmon.sh | 35 local bucket=$1; shift 50 jq -r ".[0].rmon[\"${set}-pktsNtoM\"][$bucket].val") 58 jq -r ".[0].rmon[\"${set}-pktsNtoM\"][$bucket].val") 78 while read -r -a bucket; do 82 if ! ensure_mtu $if ${bucket[0]}; then 88 if ! bucket_test $iface $neigh $set $nbuckets ${bucket[0]}; then
|
| /linux/Documentation/networking/ |
| H A D | nexthop-group-resilient.rst | 49 to choose a hash table bucket, then reads the next hop that this bucket 83 cause bucket allocation change, the wants counts for individual next hops 91 Each bucket maintains a last-used timer. Every time a packet is forwarded 92 through a bucket, this timer is updated to current jiffies value. One 94 amount of time that a bucket must not be hit by traffic in order for it to 104 upkeep changes the next hop that the bucket references to one of the 135 - Single-bucket notifications of the type 143 Some single-bucket notifications are forced, as indicated by the "force" 145 hop associated with the bucket was removed, and the bucket really must be 150 bucket should be migrated, but the HW discovers that the bucket has in fact [all …]
|
| /linux/include/trace/events/ |
| H A D | bcache.h | 68 __field(size_t, bucket ) 72 __entry->bucket = PTR_BUCKET_NR(b->c, &b->key, 0); 75 TP_printk("bucket %zu", __entry->bucket) 267 __field(size_t, bucket ) 273 __entry->bucket = PTR_BUCKET_NR(b->c, &b->key, 0); 279 __entry->bucket, __entry->block, __entry->keys) 370 __field(size_t, bucket ) 375 __entry->bucket = PTR_BUCKET_NR(b->c, &b->key, 0); 379 TP_printk("bucket %zu keys %u", __entry->bucket, __entry->keys) 429 TP_PROTO(struct cache *ca, size_t bucket), [all …]
|
| /linux/net/rxrpc/ |
| H A D | proc.c | 316 unsigned int bucket, n; in rxrpc_peer_seq_start() local 326 bucket = *_pos >> shift; in rxrpc_peer_seq_start() 328 if (bucket >= HASH_SIZE(rxnet->peer_hash)) { in rxrpc_peer_seq_start() 333 if (bucket == 0) in rxrpc_peer_seq_start() 339 p = seq_hlist_start_rcu(&rxnet->peer_hash[bucket], n - 1); in rxrpc_peer_seq_start() 342 bucket++; in rxrpc_peer_seq_start() 344 *_pos = (bucket << shift) | n; in rxrpc_peer_seq_start() 351 unsigned int bucket, n; in rxrpc_peer_seq_next() local 358 bucket = *_pos >> shift; in rxrpc_peer_seq_next() 360 p = seq_hlist_next_rcu(v, &rxnet->peer_hash[bucket], _pos); in rxrpc_peer_seq_next() [all …]
|
| /linux/net/atm/ |
| H A D | proc.c | 69 int bucket; member 78 static int __vcc_walk(struct sock **sock, int family, int *bucket, loff_t l) in __vcc_walk() argument 83 for (*bucket = 0; *bucket < VCC_HTABLE_SIZE; ++*bucket) { in __vcc_walk() 84 struct hlist_head *head = &vcc_hash[*bucket]; in __vcc_walk() 98 if (!sk && ++*bucket < VCC_HTABLE_SIZE) { in __vcc_walk() 99 sk = sk_head(&vcc_hash[*bucket]); in __vcc_walk() 113 return __vcc_walk(&state->sk, family, &state->bucket, l) ? in vcc_walk()
|
| /linux/drivers/md/bcache/ |
| H A D | bcache.h | 197 struct bucket { struct 211 BITMASK(GC_MARK, struct bucket, gc_mark, 0, 2); 217 BITMASK(GC_SECTORS_USED, struct bucket, gc_mark, 2, GC_SECTORS_USED_SIZE); 218 BITMASK(GC_MOVE, struct bucket, gc_mark, 15, 1); 460 struct bucket *buckets; 462 DECLARE_HEAP(struct bucket *, heap); 827 static inline struct bucket *PTR_BUCKET(struct cache_set *c, in PTR_BUCKET() 923 static inline uint8_t bucket_gc_gen(struct bucket *b) in bucket_gc_gen() 987 uint8_t bch_inc_gen(struct cache *ca, struct bucket *b); 990 bool bch_can_invalidate_bucket(struct cache *ca, struct bucket *b); [all …]
|
| /linux/include/net/ |
| H A D | transp_v6.h | 48 __u16 srcp, __u16 destp, int rqueue, int bucket); 51 __u16 destp, int bucket) in ip6_dgram_sock_seq_show() argument 54 bucket); in ip6_dgram_sock_seq_show()
|
| /linux/fs/smb/client/ |
| H A D | compress.c | 43 struct bucket { struct 63 static bool has_low_entropy(struct bucket *bkt, size_t slen) in has_low_entropy() argument 98 static int calc_byte_distribution(struct bucket *bkt, size_t slen) in calc_byte_distribution() 125 static bool is_mostly_ascii(const struct bucket *bkt) in is_mostly_ascii() 148 const struct bucket *a = _a, *b = _b; in cmp_bkt() 197 struct bucket *bkt = NULL; in is_compressible()
|
| /linux/drivers/net/wireguard/ |
| H A D | ratelimiter.c | 92 struct hlist_head *bucket; in wg_ratelimiter_allow() local 97 bucket = &table_v4[hsiphash_2u32(net_word, ip, &key) & in wg_ratelimiter_allow() 104 bucket = &table_v6[hsiphash_3u32(net_word, ip >> 32, ip, &key) & in wg_ratelimiter_allow() 111 hlist_for_each_entry_rcu(entry, bucket, hash) { in wg_ratelimiter_allow() 149 hlist_add_head_rcu(&entry->hash, bucket); in wg_ratelimiter_allow()
|
| /linux/drivers/net/ethernet/freescale/fman/ |
| H A D | fman_dtsec.c | 458 static void set_bucket(struct dtsec_regs __iomem *regs, int bucket, in set_bucket() argument 461 int reg_idx = (bucket >> 5) & 0xf; in set_bucket() 462 int bit_idx = bucket & 0x1f; in set_bucket() 1017 s32 bucket; in dtsec_add_hash_mac_address() local 1045 bucket = (s32)((crc >> 23) & 0x1ff); in dtsec_add_hash_mac_address() 1047 bucket = (s32)((crc >> 24) & 0xff); in dtsec_add_hash_mac_address() 1052 bucket += 0x100; in dtsec_add_hash_mac_address() 1055 set_bucket(dtsec->regs, bucket, true); in dtsec_add_hash_mac_address() 1067 &dtsec->multicast_addr_hash->lsts[bucket]); in dtsec_add_hash_mac_address() 1070 &dtsec->unicast_addr_hash->lsts[bucket]); in dtsec_add_hash_mac_address() [all …]
|