| /linux/fs/nfs/ |
| H A D | nfs4session.c | 27 static void nfs4_init_slot_table(struct nfs4_slot_table *tbl, const char *queue) in nfs4_init_slot_table() argument 29 tbl->highest_used_slotid = NFS4_NO_SLOT; in nfs4_init_slot_table() 30 spin_lock_init(&tbl->slot_tbl_lock); in nfs4_init_slot_table() 31 rpc_init_priority_wait_queue(&tbl->slot_tbl_waitq, queue); in nfs4_init_slot_table() 32 init_waitqueue_head(&tbl->slot_waitq); in nfs4_init_slot_table() 33 init_completion(&tbl->complete); in nfs4_init_slot_table() 39 static void nfs4_shrink_slot_table(struct nfs4_slot_table *tbl, u32 newsize) in nfs4_shrink_slot_table() argument 42 if (newsize >= tbl->max_slots) in nfs4_shrink_slot_table() 45 p = &tbl->slots; in nfs4_shrink_slot_table() 53 tbl->max_slots--; in nfs4_shrink_slot_table() [all …]
|
| /linux/drivers/net/ethernet/mellanox/mlx5/core/steering/sws/ |
| H A D | dr_table.c | 28 action->dest_tbl->tbl->rx.s_anchor->chunk : in dr_table_set_miss_action_nic() 29 action->dest_tbl->tbl->tx.s_anchor->chunk; in dr_table_set_miss_action_nic() 46 int mlx5dr_table_set_miss_action(struct mlx5dr_table *tbl, in mlx5dr_table_set_miss_action() argument 54 mlx5dr_domain_lock(tbl->dmn); in mlx5dr_table_set_miss_action() 56 if (tbl->dmn->type == MLX5DR_DOMAIN_TYPE_NIC_RX || in mlx5dr_table_set_miss_action() 57 tbl->dmn->type == MLX5DR_DOMAIN_TYPE_FDB) { in mlx5dr_table_set_miss_action() 58 ret = dr_table_set_miss_action_nic(tbl->dmn, &tbl->rx, action); in mlx5dr_table_set_miss_action() 63 if (tbl->dmn->type == MLX5DR_DOMAIN_TYPE_NIC_TX || in mlx5dr_table_set_miss_action() 64 tbl->dmn->type == MLX5DR_DOMAIN_TYPE_FDB) { in mlx5dr_table_set_miss_action() 65 ret = dr_table_set_miss_action_nic(tbl->dmn, &tbl->tx, action); in mlx5dr_table_set_miss_action() [all …]
|
| /linux/drivers/net/ethernet/mellanox/mlx5/core/steering/hws/ |
| H A D | table.c | 6 u32 mlx5hws_table_get_id(struct mlx5hws_table *tbl) in mlx5hws_table_get_id() argument 8 return tbl->ft_id; in mlx5hws_table_get_id() 11 static void hws_table_init_next_ft_attr(struct mlx5hws_table *tbl, in hws_table_init_next_ft_attr() argument 15 ft_attr->type = tbl->fw_ft_type; in hws_table_init_next_ft_attr() 16 if (tbl->type == MLX5HWS_TABLE_TYPE_FDB) in hws_table_init_next_ft_attr() 17 ft_attr->level = tbl->ctx->caps->fdb_ft.max_level - 1; in hws_table_init_next_ft_attr() 19 ft_attr->level = tbl->ctx->caps->nic_ft.max_level - 1; in hws_table_init_next_ft_attr() 25 static void hws_table_set_cap_attr(struct mlx5hws_table *tbl, in hws_table_set_cap_attr() argument 36 if (!MLX5_CAP_ESW_FLOWTABLE(tbl->ctx->mdev, fdb_dynamic_tunnel)) { in hws_table_set_cap_attr() 42 static int hws_table_up_default_fdb_miss_tbl(struct mlx5hws_table *tbl) in hws_table_up_default_fdb_miss_tbl() argument [all …]
|
| H A D | matcher.c | 23 mlx5hws_table_destroy_default_ft(matcher->tbl, matcher->end_ft_id); in hws_matcher_destroy_end_ft() 26 int mlx5hws_matcher_update_end_ft_isolated(struct mlx5hws_table *tbl, in mlx5hws_matcher_update_end_ft_isolated() argument 31 if (list_empty(&tbl->matchers_list)) in mlx5hws_matcher_update_end_ft_isolated() 37 list_for_each_entry(tmp_matcher, &tbl->matchers_list, list_node) in mlx5hws_matcher_update_end_ft_isolated() 40 tmp_matcher = list_last_entry(&tbl->matchers_list, in mlx5hws_matcher_update_end_ft_isolated() 44 return mlx5hws_table_ft_set_next_ft(tbl->ctx, in mlx5hws_matcher_update_end_ft_isolated() 46 tbl->fw_ft_type, in mlx5hws_matcher_update_end_ft_isolated() 52 struct mlx5hws_table *tbl = matcher->tbl; in hws_matcher_connect_end_ft_isolated() local 57 ret = mlx5hws_table_ft_set_next_rtc(tbl->ctx, in hws_matcher_connect_end_ft_isolated() 59 matcher->tbl->fw_ft_type, in hws_matcher_connect_end_ft_isolated() [all …]
|
| /linux/arch/powerpc/kernel/ |
| H A D | iommu.c | 47 struct iommu_table *tbl = data; in iommu_debugfs_weight_get() local 48 *val = bitmap_weight(tbl->it_map, tbl->it_size); in iommu_debugfs_weight_get() 53 static void iommu_debugfs_add(struct iommu_table *tbl) in iommu_debugfs_add() argument 58 sprintf(name, "%08lx", tbl->it_index); in iommu_debugfs_add() 61 debugfs_create_file_unsafe("weight", 0400, liobn_entry, tbl, &iommu_debugfs_fops_weight); in iommu_debugfs_add() 62 debugfs_create_ulong("it_size", 0400, liobn_entry, &tbl->it_size); in iommu_debugfs_add() 63 debugfs_create_ulong("it_page_shift", 0400, liobn_entry, &tbl->it_page_shift); in iommu_debugfs_add() 64 debugfs_create_ulong("it_reserved_start", 0400, liobn_entry, &tbl->it_reserved_start); in iommu_debugfs_add() 65 debugfs_create_ulong("it_reserved_end", 0400, liobn_entry, &tbl->it_reserved_end); in iommu_debugfs_add() 66 debugfs_create_ulong("it_indirect_levels", 0400, liobn_entry, &tbl->it_indirect_levels); in iommu_debugfs_add() [all …]
|
| /linux/drivers/infiniband/hw/vmw_pvrdma/ |
| H A D | pvrdma_doorbell.c | 56 struct pvrdma_id_table *tbl = &dev->uar_table.tbl; in pvrdma_uar_table_init() local 61 tbl->last = 0; in pvrdma_uar_table_init() 62 tbl->top = 0; in pvrdma_uar_table_init() 63 tbl->max = num; in pvrdma_uar_table_init() 64 tbl->mask = mask; in pvrdma_uar_table_init() 65 spin_lock_init(&tbl->lock); in pvrdma_uar_table_init() 66 tbl->table = bitmap_zalloc(num, GFP_KERNEL); in pvrdma_uar_table_init() 67 if (!tbl->table) in pvrdma_uar_table_init() 71 __set_bit(0, tbl->table); in pvrdma_uar_table_init() 78 struct pvrdma_id_table *tbl = &dev->uar_table.tbl; in pvrdma_uar_table_cleanup() local [all …]
|
| /linux/arch/powerpc/platforms/powernv/ |
| H A D | pci-ioda-tce.c | 48 void pnv_pci_setup_iommu_table(struct iommu_table *tbl, in pnv_pci_setup_iommu_table() argument 52 tbl->it_blocksize = 16; in pnv_pci_setup_iommu_table() 53 tbl->it_base = (unsigned long)tce_mem; in pnv_pci_setup_iommu_table() 54 tbl->it_page_shift = page_shift; in pnv_pci_setup_iommu_table() 55 tbl->it_offset = dma_offset >> tbl->it_page_shift; in pnv_pci_setup_iommu_table() 56 tbl->it_index = 0; in pnv_pci_setup_iommu_table() 57 tbl->it_size = tce_size >> 3; in pnv_pci_setup_iommu_table() 58 tbl->it_busno = 0; in pnv_pci_setup_iommu_table() 59 tbl->it_type = TCE_PCI; in pnv_pci_setup_iommu_table() 83 static __be64 *pnv_tce(struct iommu_table *tbl, bool user, long idx, bool alloc) in pnv_tce() argument [all …]
|
| /linux/lib/ |
| H A D | rhashtable.c | 38 const struct bucket_table *tbl, in head_hashfn() argument 41 return rht_head_hashfn(ht, tbl, he, ht->p); in head_hashfn() 53 int lockdep_rht_bucket_is_held(const struct bucket_table *tbl, u32 hash) in lockdep_rht_bucket_is_held() argument 57 if (unlikely(tbl->nest)) in lockdep_rht_bucket_is_held() 59 return bit_spin_is_locked(0, (unsigned long *)&tbl->buckets[hash]); in lockdep_rht_bucket_is_held() 67 const struct bucket_table *tbl) in nested_table_top() argument 70 * because it's set at the same time as tbl->nest. in nested_table_top() 72 return (void *)rcu_dereference_protected(tbl->buckets[0], 1); in nested_table_top() 94 static void nested_bucket_table_free(const struct bucket_table *tbl) in nested_bucket_table_free() argument 96 unsigned int size = tbl in nested_bucket_table_free() 109 bucket_table_free(const struct bucket_table * tbl) bucket_table_free() argument 153 struct bucket_table *tbl; nested_bucket_table_alloc() local 181 struct bucket_table *tbl = NULL; bucket_table_alloc() local 216 rhashtable_last_table(struct rhashtable * ht,struct bucket_table * tbl) rhashtable_last_table() argument 417 struct bucket_table *tbl; rht_deferred_worker() local 447 rhashtable_insert_rehash(struct rhashtable * ht,struct bucket_table * tbl) rhashtable_insert_rehash() argument 496 rhashtable_lookup_one(struct rhashtable * ht,struct rhash_lock_head __rcu ** bkt,struct bucket_table * tbl,unsigned int hash,const void * key,struct rhash_head * obj) rhashtable_lookup_one() argument 547 rhashtable_insert_one(struct rhashtable * ht,struct rhash_lock_head __rcu ** bkt,struct bucket_table * tbl,unsigned int hash,struct rhash_head * obj,void * data) rhashtable_insert_one() argument 594 struct bucket_table *tbl; rhashtable_try_insert() local 801 struct bucket_table *tbl = iter->walker.tbl; __rhashtable_walk_find_next() local 946 struct bucket_table *tbl = iter->walker.tbl; rhashtable_walk_stop() local 1030 struct bucket_table *tbl; rhashtable_init_noprof() local 1148 struct bucket_table *tbl, *next_tbl; rhashtable_free_and_destroy() local 1189 __rht_bucket_nested(const struct bucket_table * tbl,unsigned int hash) __rht_bucket_nested() argument 1218 rht_bucket_nested(const struct bucket_table * tbl,unsigned int hash) rht_bucket_nested() argument 1229 rht_bucket_nested_insert(struct rhashtable * ht,struct bucket_table * tbl,unsigned int hash) rht_bucket_nested_insert() argument [all...] |
| /linux/drivers/net/wireless/intel/iwlwifi/dvm/ |
| H A D | rs.c | 414 static s32 get_expected_tpt(struct iwl_scale_tbl_info *tbl, int rs_index) in get_expected_tpt() argument 416 if (tbl->expected_tpt) in get_expected_tpt() 417 return tbl->expected_tpt[rs_index]; in get_expected_tpt() 428 static int rs_collect_tx_data(struct iwl_scale_tbl_info *tbl, in rs_collect_tx_data() argument 439 window = &(tbl->win[scale_index]); in rs_collect_tx_data() 442 tpt = get_expected_tpt(tbl, scale_index); in rs_collect_tx_data() 507 struct iwl_scale_tbl_info *tbl, in rate_n_flags_from_tbl() argument 512 if (is_legacy(tbl->lq_type)) { in rate_n_flags_from_tbl() 517 } else if (is_Ht(tbl->lq_type)) { in rate_n_flags_from_tbl() 524 if (is_siso(tbl->lq_type)) in rate_n_flags_from_tbl() [all …]
|
| H A D | rs.h | 256 #define is_legacy(tbl) (((tbl) == LQ_G) || ((tbl) == LQ_A)) argument 257 #define is_siso(tbl) ((tbl) == LQ_SISO) argument 258 #define is_mimo2(tbl) ((tbl) == LQ_MIMO2) argument 259 #define is_mimo3(tbl) ((tbl) == LQ_MIMO3) argument 260 #define is_mimo(tbl) (is_mimo2(tbl) || is_mimo3(tbl)) argument 261 #define is_Ht(tbl) (is_siso(tbl) || is_mimo(tbl)) argument 262 #define is_a_band(tbl) ((tbl) == LQ_A) argument 263 #define is_g_and(tbl) ((tbl) == LQ_G) argument
|
| /linux/arch/powerpc/include/asm/ |
| H A D | iommu.h | 45 int (*set)(struct iommu_table *tbl, 56 int (*xchg_no_kill)(struct iommu_table *tbl, 61 void (*tce_kill)(struct iommu_table *tbl, 65 __be64 *(*useraddrptr)(struct iommu_table *tbl, long index, bool alloc); 67 void (*clear)(struct iommu_table *tbl, 70 unsigned long (*get)(struct iommu_table *tbl, long index); 71 void (*flush)(struct iommu_table *tbl); 72 void (*free)(struct iommu_table *tbl); 122 #define IOMMU_TABLE_USERSPACE_ENTRY_RO(tbl, entry) \ argument 123 ((tbl)->it_ops->useraddrptr((tbl), (entry), false)) [all …]
|
| /linux/net/netfilter/ipvs/ |
| H A D | ip_vs_lblc.c | 166 ip_vs_lblc_hash(struct ip_vs_lblc_table *tbl, struct ip_vs_lblc_entry *en) in ip_vs_lblc_hash() argument 170 hlist_add_head_rcu(&en->list, &tbl->bucket[hash]); in ip_vs_lblc_hash() 171 atomic_inc(&tbl->entries); in ip_vs_lblc_hash() 177 ip_vs_lblc_get(int af, struct ip_vs_lblc_table *tbl, in ip_vs_lblc_get() argument 183 hlist_for_each_entry_rcu(en, &tbl->bucket[hash], list) in ip_vs_lblc_get() 196 ip_vs_lblc_new(struct ip_vs_lblc_table *tbl, const union nf_inet_addr *daddr, in ip_vs_lblc_new() argument 201 en = ip_vs_lblc_get(af, tbl, daddr); in ip_vs_lblc_new() 218 ip_vs_lblc_hash(tbl, en); in ip_vs_lblc_new() 229 struct ip_vs_lblc_table *tbl = svc->sched_data; in ip_vs_lblc_flush() local 235 tbl->dead = true; in ip_vs_lblc_flush() [all …]
|
| H A D | ip_vs_lblcr.c | 329 ip_vs_lblcr_hash(struct ip_vs_lblcr_table *tbl, struct ip_vs_lblcr_entry *en) in ip_vs_lblcr_hash() argument 333 hlist_add_head_rcu(&en->list, &tbl->bucket[hash]); in ip_vs_lblcr_hash() 334 atomic_inc(&tbl->entries); in ip_vs_lblcr_hash() 340 ip_vs_lblcr_get(int af, struct ip_vs_lblcr_table *tbl, in ip_vs_lblcr_get() argument 346 hlist_for_each_entry_rcu(en, &tbl->bucket[hash], list) in ip_vs_lblcr_get() 359 ip_vs_lblcr_new(struct ip_vs_lblcr_table *tbl, const union nf_inet_addr *daddr, in ip_vs_lblcr_new() argument 364 en = ip_vs_lblcr_get(af, tbl, daddr); in ip_vs_lblcr_new() 380 ip_vs_lblcr_hash(tbl, en); in ip_vs_lblcr_new() 395 struct ip_vs_lblcr_table *tbl = svc->sched_data; in ip_vs_lblcr_flush() local 401 tbl->dead = true; in ip_vs_lblcr_flush() [all …]
|
| /linux/drivers/net/wireless/intel/iwlegacy/ |
| H A D | 4965-rs.c | 383 il4965_get_expected_tpt(struct il_scale_tbl_info *tbl, int rs_idx) in il4965_get_expected_tpt() argument 385 if (tbl->expected_tpt) in il4965_get_expected_tpt() 386 return tbl->expected_tpt[rs_idx]; in il4965_get_expected_tpt() 398 il4965_rs_collect_tx_data(struct il_scale_tbl_info *tbl, int scale_idx, in il4965_rs_collect_tx_data() argument 409 win = &(tbl->win[scale_idx]); in il4965_rs_collect_tx_data() 412 tpt = il4965_get_expected_tpt(tbl, scale_idx); in il4965_rs_collect_tx_data() 476 il4965_rate_n_flags_from_tbl(struct il_priv *il, struct il_scale_tbl_info *tbl, in il4965_rate_n_flags_from_tbl() argument 481 if (is_legacy(tbl->lq_type)) { in il4965_rate_n_flags_from_tbl() 486 } else if (is_Ht(tbl->lq_type)) { in il4965_rate_n_flags_from_tbl() 493 if (is_siso(tbl->lq_type)) in il4965_rate_n_flags_from_tbl() [all …]
|
| /linux/drivers/net/wireless/marvell/mwifiex/ |
| H A D | 11n_rxreorder.c | 98 struct mwifiex_rx_reorder_tbl *tbl, in mwifiex_11n_dispatch_pkt_until_start_win() argument 108 pkt_to_send = (start_win > tbl->start_win) ? in mwifiex_11n_dispatch_pkt_until_start_win() 109 min((start_win - tbl->start_win), tbl->win_size) : in mwifiex_11n_dispatch_pkt_until_start_win() 110 tbl->win_size; in mwifiex_11n_dispatch_pkt_until_start_win() 113 if (tbl->rx_reorder_ptr[i]) { in mwifiex_11n_dispatch_pkt_until_start_win() 114 skb = tbl->rx_reorder_ptr[i]; in mwifiex_11n_dispatch_pkt_until_start_win() 116 tbl->rx_reorder_ptr[i] = NULL; in mwifiex_11n_dispatch_pkt_until_start_win() 124 for (i = 0; i < tbl->win_size - pkt_to_send; ++i) { in mwifiex_11n_dispatch_pkt_until_start_win() 125 tbl->rx_reorder_ptr[i] = tbl->rx_reorder_ptr[pkt_to_send + i]; in mwifiex_11n_dispatch_pkt_until_start_win() 126 tbl->rx_reorder_ptr[pkt_to_send + i] = NULL; in mwifiex_11n_dispatch_pkt_until_start_win() [all …]
|
| /linux/net/core/ |
| H A D | neighbour.c | 56 static void pneigh_ifdown(struct neigh_table *tbl, struct net_device *dev, 141 atomic_dec(&n->tbl->gc_entries); in neigh_mark_dead() 151 spin_lock_bh(&n->tbl->lock); in neigh_update_gc_list() 166 atomic_dec(&n->tbl->gc_entries); in neigh_update_gc_list() 169 list_add_tail(&n->gc_list, &n->tbl->gc_list); in neigh_update_gc_list() 170 atomic_inc(&n->tbl->gc_entries); in neigh_update_gc_list() 174 spin_unlock_bh(&n->tbl->lock); in neigh_update_gc_list() 181 spin_lock_bh(&n->tbl->lock); in neigh_update_managed_list() 192 list_add_tail(&n->managed_list, &n->tbl->managed_list); in neigh_update_managed_list() 195 spin_unlock_bh(&n->tbl->lock); in neigh_update_managed_list() [all …]
|
| /linux/tools/perf/ |
| H A D | check-headers.sh | 199 check_2 tools/perf/arch/x86/entry/syscalls/syscall_32.tbl arch/x86/entry/syscalls/syscall_32.tbl 200 check_2 tools/perf/arch/x86/entry/syscalls/syscall_64.tbl arch/x86/entry/syscalls/syscall_64.tbl 201 check_2 tools/perf/arch/powerpc/entry/syscalls/syscall.tbl arch/powerpc/kernel/syscalls/syscall.tbl 202 check_2 tools/perf/arch/s390/entry/syscalls/syscall.tbl arch/s390/kernel/syscalls/syscall.tbl 203 …eck_2 tools/perf/arch/mips/entry/syscalls/syscall_n64.tbl arch/mips/kernel/syscalls/syscall_n64.tbl 204 check_2 tools/perf/arch/arm/entry/syscalls/syscall.tbl arch/arm/tools/syscall.tbl 205 check_2 tools/perf/arch/sh/entry/syscalls/syscall.tbl arch/sh/kernel/syscalls/syscall.tbl 206 check_2 tools/perf/arch/sparc/entry/syscalls/syscall.tbl arch/sparc/kernel/syscalls/syscall.tbl 207 check_2 tools/perf/arch/xtensa/entry/syscalls/syscall.tbl arch/xtensa/kernel/syscalls/syscall.tbl 208 check_2 tools/perf/arch/alpha/entry/syscalls/syscall.tbl arch/alpha/entry/syscalls/syscall.tbl [all …]
|
| /linux/drivers/vfio/ |
| H A D | vfio_iommu_spapr_tce.c | 209 struct iommu_table *tbl = container->tables[i]; in tce_iommu_find_table() local 211 if (tbl) { in tce_iommu_find_table() 212 unsigned long entry = ioba >> tbl->it_page_shift; in tce_iommu_find_table() 213 unsigned long start = tbl->it_offset; in tce_iommu_find_table() 214 unsigned long end = start + tbl->it_size; in tce_iommu_find_table() 217 *ptbl = tbl; in tce_iommu_find_table() 339 struct iommu_table *tbl, 342 struct iommu_table *tbl); 362 struct iommu_table *tbl = container->tables[i]; in tce_iommu_release() local 364 if (!tbl) in tce_iommu_release() [all …]
|
| /linux/net/mac80211/ |
| H A D | mesh_pathtbl.c | 20 static void mesh_path_free_rcu(struct mesh_table *tbl, struct mesh_path *mpath); 82 struct mesh_table *tbl = tblptr; in mesh_path_rht_free() local 84 mesh_path_free_rcu(tbl, mpath); in mesh_path_rht_free() 87 static void mesh_table_init(struct mesh_table *tbl) in mesh_table_init() argument 89 INIT_HLIST_HEAD(&tbl->known_gates); in mesh_table_init() 90 INIT_HLIST_HEAD(&tbl->walk_head); in mesh_table_init() 91 atomic_set(&tbl->entries, 0); in mesh_table_init() 92 spin_lock_init(&tbl->gates_lock); in mesh_table_init() 93 spin_lock_init(&tbl->walk_lock); in mesh_table_init() 98 WARN_ON(rhashtable_init(&tbl->rhead, &mesh_rht_params)); in mesh_table_init() [all …]
|
| /linux/net/netfilter/ |
| H A D | xt_repldata.h | 24 } *tbl; \ 26 size_t term_offset = (offsetof(typeof(*tbl), entries[nhooks]) + \ 28 tbl = kzalloc(term_offset + sizeof(*term), GFP_KERNEL); \ 29 if (tbl == NULL) \ 31 term = (struct type##_error *)&(((char *)tbl)[term_offset]); \ 32 strscpy(tbl->repl.name, info->name); \ 34 tbl->repl.valid_hooks = hook_mask; \ 35 tbl->repl.num_entries = nhooks + 1; \ 36 tbl->repl.size = nhooks * sizeof(struct type##_standard) + \ 41 tbl->repl.hook_entry[hooknum] = bytes; \ [all …]
|
| /linux/arch/powerpc/kvm/ |
| H A D | book3s_64_vio.c | 66 iommu_tce_table_put(stit->tbl); in kvm_spapr_tce_iommu_table_free() 98 if (table_group->tables[i] != stit->tbl) in kvm_spapr_tce_release_iommu_group() 114 struct iommu_table *tbl = NULL; in kvm_spapr_tce_attach_iommu_group() local 154 tbl = iommu_tce_table_get(tbltmp); in kvm_spapr_tce_attach_iommu_group() 158 if (!tbl) in kvm_spapr_tce_attach_iommu_group() 163 if (tbl != stit->tbl) in kvm_spapr_tce_attach_iommu_group() 168 iommu_tce_table_put(tbl); in kvm_spapr_tce_attach_iommu_group() 183 iommu_tce_table_put(tbl); in kvm_spapr_tce_attach_iommu_group() 187 stit->tbl = tbl; in kvm_spapr_tce_attach_iommu_group() 390 long shift = stit->tbl->it_page_shift; in kvmppc_tce_validate() [all …]
|
| /linux/drivers/net/ethernet/mellanox/mlx5/core/en/ |
| H A D | mod_hdr.c | 42 void mlx5e_mod_hdr_tbl_init(struct mod_hdr_tbl *tbl) in mlx5e_mod_hdr_tbl_init() argument 44 mutex_init(&tbl->lock); in mlx5e_mod_hdr_tbl_init() 45 hash_init(tbl->hlist); in mlx5e_mod_hdr_tbl_init() 48 void mlx5e_mod_hdr_tbl_destroy(struct mod_hdr_tbl *tbl) in mlx5e_mod_hdr_tbl_destroy() argument 50 WARN_ON(!hash_empty(tbl->hlist)); in mlx5e_mod_hdr_tbl_destroy() 51 mutex_destroy(&tbl->lock); in mlx5e_mod_hdr_tbl_destroy() 54 static struct mlx5e_mod_hdr_handle *mod_hdr_get(struct mod_hdr_tbl *tbl, in mod_hdr_get() argument 60 hash_for_each_possible(tbl->hlist, mh, mod_hdr_hlist, hash_key) { in mod_hdr_get() 73 struct mod_hdr_tbl *tbl, in mlx5e_mod_hdr_attach() argument 90 mutex_lock(&tbl->lock); in mlx5e_mod_hdr_attach() [all …]
|
| /linux/arch/powerpc/platforms/pseries/ |
| H A D | iommu.c | 61 struct iommu_table *tbl; in iommu_pseries_alloc_table() local 63 tbl = kzalloc_node(sizeof(struct iommu_table), GFP_KERNEL, node); in iommu_pseries_alloc_table() 64 if (!tbl) in iommu_pseries_alloc_table() 67 INIT_LIST_HEAD_RCU(&tbl->it_group_list); in iommu_pseries_alloc_table() 68 kref_init(&tbl->it_kref); in iommu_pseries_alloc_table() 69 return tbl; in iommu_pseries_alloc_table() 122 static int tce_build_pSeries(struct iommu_table *tbl, long index, in tce_build_pSeries() argument 130 const unsigned long tceshift = tbl->it_page_shift; in tce_build_pSeries() 131 const unsigned long pagesize = IOMMU_PAGE_SIZE(tbl); in tce_build_pSeries() 138 tcep = ((__be64 *)tbl->it_base) + index; in tce_build_pSeries() [all …]
|
| /linux/include/net/ |
| H A D | neighbour.h | 78 struct neigh_table *tbl; member 138 #define NEIGH_CACHE_STAT_INC(tbl, field) this_cpu_inc((tbl)->stats->field) 143 struct neigh_table *tbl; 251 return p->tbl->family; 259 return (char *)n + n->tbl->entry_size; 304 struct neigh_table *tbl, in ___neigh_lookup_noref() 312 struct neigh_hash_table *nht = rcu_dereference(tbl->nht); in ___neigh_lookup_noref() 324 static inline struct neighbour *__neigh_lookup_noref(struct neigh_table *tbl, in __neigh_lookup_noref() 328 return ___neigh_lookup_noref(tbl, tb in __neigh_lookup_noref() 136 NEIGH_CACHE_STAT_INC(tbl,field) global() argument 141 struct neigh_table *tbl; global() member 302 ___neigh_lookup_noref(struct neigh_table * tbl,bool (* key_eq)(const struct neighbour * n,const void * pkey),__u32 (* hash)(const void * pkey,const struct net_device * dev,__u32 * hash_rnd),const void * pkey,struct net_device * dev) ___neigh_lookup_noref() argument 322 __neigh_lookup_noref(struct neigh_table * tbl,const void * pkey,struct net_device * dev) __neigh_lookup_noref() argument 346 neigh_create(struct neigh_table * tbl,const void * pkey,struct net_device * dev) neigh_create() argument 405 struct neigh_table *tbl; global() member 551 __neigh_lookup(struct neigh_table * tbl,const void * pkey,struct net_device * dev,int creat) __neigh_lookup() argument 563 __neigh_lookup_errno(struct neigh_table * tbl,const void * pkey,struct net_device * dev) __neigh_lookup_errno() argument [all...] |
| /linux/arch/sparc/kernel/ |
| H A D | pci_sun4v.c | 189 struct iommu_map_table *tbl; in dma_4v_alloc_coherent() local 216 tbl = &iommu->tbl; in dma_4v_alloc_coherent() 218 tbl = &iommu->atu->tbl; in dma_4v_alloc_coherent() 220 entry = iommu_tbl_range_alloc(dev, tbl, npages, NULL, in dma_4v_alloc_coherent() 226 *dma_addrp = (tbl->table_map_base + (entry << IO_PAGE_SHIFT)); in dma_4v_alloc_coherent() 252 iommu_tbl_range_free(tbl, *dma_addrp, npages, IOMMU_ERROR_CODE); in dma_4v_alloc_coherent() 329 struct iommu_map_table *tbl; in dma_4v_free_coherent() local 341 tbl = &iommu->tbl; in dma_4v_free_coherent() 344 tbl = &atu->tbl; in dma_4v_free_coherent() 347 entry = ((dvma - tbl->table_map_base) >> IO_PAGE_SHIFT); in dma_4v_free_coherent() [all …]
|