| /linux/lib/ |
| H A D | test_maple_tree.c | 19 #define mt_dump(mt, fmt) do {} while (0) argument 20 #define mt_validate(mt) do {} while (0) argument 63 static int __init mtree_insert_index(struct maple_tree *mt, in mtree_insert_index() argument 66 return mtree_insert(mt, index, xa_mk_value(index & LONG_MAX), gfp); in mtree_insert_index() 69 static void __init mtree_erase_index(struct maple_tree *mt, unsigned long index) in mtree_erase_index() argument 71 MT_BUG_ON(mt, mtree_erase(mt, index) != xa_mk_value(index & LONG_MAX)); in mtree_erase_index() 72 MT_BUG_ON(mt, mtree_load(mt, index) != NULL); in mtree_erase_index() 75 static int __init mtree_test_insert(struct maple_tree *mt, unsigne argument 81 mtree_test_store_range(struct maple_tree * mt,unsigned long start,unsigned long end,void * ptr) mtree_test_store_range() argument 87 mtree_test_store(struct maple_tree * mt,unsigned long start,void * ptr) mtree_test_store() argument 93 mtree_test_insert_range(struct maple_tree * mt,unsigned long start,unsigned long end,void * ptr) mtree_test_insert_range() argument 99 mtree_test_load(struct maple_tree * mt,unsigned long index) mtree_test_load() argument 104 mtree_test_erase(struct maple_tree * mt,unsigned long index) mtree_test_erase() argument 110 check_mtree_alloc_range(struct maple_tree * mt,unsigned long start,unsigned long end,unsigned long size,unsigned long expected,int eret,void * ptr) check_mtree_alloc_range() argument 127 check_mtree_alloc_rrange(struct maple_tree * mt,unsigned long start,unsigned long end,unsigned long size,unsigned long expected,int eret,void * ptr) check_mtree_alloc_rrange() argument 145 check_load(struct maple_tree * mt,unsigned long index,void * ptr) check_load() argument 155 check_store_range(struct maple_tree * mt,unsigned long start,unsigned long end,void * ptr,int expected) check_store_range() argument 171 check_insert_range(struct maple_tree * mt,unsigned long start,unsigned long end,void * ptr,int expected) check_insert_range() argument 187 check_insert(struct maple_tree * mt,unsigned long index,void * ptr) check_insert() argument 196 check_dup_insert(struct maple_tree * mt,unsigned long index,void * ptr) check_dup_insert() argument 206 check_index_load(struct maple_tree * mt,unsigned long index) check_index_load() argument 227 check_rev_seq(struct maple_tree * mt,unsigned long max,bool verbose) check_rev_seq() argument 260 check_seq(struct maple_tree * mt,unsigned long max,bool verbose) check_seq() argument 289 check_lb_not_empty(struct maple_tree * mt) check_lb_not_empty() argument 308 check_lower_bound_split(struct maple_tree * mt) check_lower_bound_split() argument 314 check_upper_bound_split(struct maple_tree * mt) check_upper_bound_split() argument 339 check_mid_split(struct maple_tree * mt) check_mid_split() argument 348 check_rev_find(struct maple_tree * mt) check_rev_find() argument 387 check_find(struct maple_tree * mt) check_find() argument 604 check_find_2(struct maple_tree * mt) check_find_2() argument 649 check_alloc_rev_range(struct maple_tree * mt) check_alloc_rev_range() argument 823 check_alloc_range(struct maple_tree * mt) check_alloc_range() argument 989 check_ranges(struct maple_tree * mt) check_ranges() argument 1322 check_next_entry(struct maple_tree * mt) check_next_entry() argument 1346 check_prev_entry(struct maple_tree * mt) check_prev_entry() argument 1390 check_store_null(struct maple_tree * mt) check_store_null() argument 1476 check_root_expand(struct maple_tree * mt) check_root_expand() argument 1566 check_deficient_node(struct maple_tree * mt) check_deficient_node() argument 1590 check_gap_combining(struct maple_tree * mt) check_gap_combining() argument 1787 check_node_overwrite(struct maple_tree * mt) check_node_overwrite() argument 1800 bench_slot_store(struct maple_tree * mt) bench_slot_store() argument 1816 bench_node_store(struct maple_tree * mt) bench_node_store() argument 1835 bench_awalk(struct maple_tree * mt) bench_awalk() argument 1852 bench_walk(struct maple_tree * mt) bench_walk() argument 1869 bench_load(struct maple_tree * mt) bench_load() argument 1882 bench_mt_for_each(struct maple_tree * mt) bench_mt_for_each() argument 1906 bench_mas_for_each(struct maple_tree * mt) bench_mas_for_each() argument 1936 bench_mas_prev(struct maple_tree * mt) bench_mas_prev() argument 1968 struct maple_tree mt, newmt; check_forking() local 2010 check_iteration(struct maple_tree * mt) check_iteration() argument 2077 check_mas_store_gfp(struct maple_tree * mt) check_mas_store_gfp() argument 2112 struct maple_tree mt, newmt; bench_forking() local 2158 next_prev_test(struct maple_tree * mt) next_prev_test() argument 2340 check_spanning_relatives(struct maple_tree * mt) check_spanning_relatives() argument 2353 check_fuzzer(struct maple_tree * mt) check_fuzzer() argument 2749 check_bnode_min_spanning(struct maple_tree * mt) check_bnode_min_spanning() argument 2768 check_empty_area_window(struct maple_tree * mt) check_empty_area_window() argument 2853 check_empty_area_fill(struct maple_tree * mt) check_empty_area_fill() argument 3041 check_state_handling(struct maple_tree * mt) check_state_handling() argument 3611 alloc_cyclic_testing(struct maple_tree * mt) alloc_cyclic_testing() argument [all...] |
| H A D | maple_tree.c | 176 static void mt_set_height(struct maple_tree *mt, unsigned char height) 178 unsigned int new_flags = mt->ma_flags; in mt_free_bulk() 181 MT_BUG_ON(mt, height > MAPLE_HEIGHT_MAX); 183 mt->ma_flags = new_flags; in mt_return_sheaf() 191 static inline unsigned int mt_attr(struct maple_tree *mt) 193 return mt->ma_flags & ~MT_FLAGS_HEIGHT_MASK; in mt_refill_sheaf() 318 const enum maple_type mt) 321 RCU_INIT_POINTER(*slot, (void *)mt_mk_node(mn, mt)); 364 static __always_inline bool mt_is_alloc(struct maple_tree *mt) in mte_set_full() 366 return (mt in mte_set_full() 211 mt_set_height(struct maple_tree * mt,unsigned char height) mt_set_height() argument 226 mt_attr(struct maple_tree * mt) mt_attr() argument 392 mt_is_alloc(struct maple_tree * mt) mt_is_alloc() argument 706 ma_slots(struct maple_node * mn,enum maple_type mt) ma_slots() argument 721 mt_write_locked(const struct maple_tree * mt) mt_write_locked() argument 727 mt_locked(const struct maple_tree * mt) mt_locked() argument 733 mt_slot(const struct maple_tree * mt,void __rcu ** slots,unsigned char offset) mt_slot() argument 739 mt_slot_locked(struct maple_tree * mt,void __rcu ** slots,unsigned char offset) mt_slot_locked() argument 783 mt_root_locked(struct maple_tree * mt) mt_root_locked() argument 800 ma_meta(struct maple_node * mn,enum maple_type mt) ma_meta() argument 817 ma_set_meta(struct maple_node * mn,enum maple_type mt,unsigned char offset,unsigned char end) ma_set_meta() argument 832 mt_clear_meta(struct maple_tree * mt,struct maple_node * mn,enum maple_type type) mt_clear_meta() argument 869 ma_meta_end(struct maple_node * mn,enum maple_type mt) ma_meta_end() argument 891 ma_set_meta_gap(struct maple_node * mn,enum maple_type mt,unsigned char offset) ma_set_meta_gap() argument 1309 enum maple_type mt; mas_leaf_max_gap() local 1392 ma_max_gap(struct maple_node * node,unsigned long * gaps,enum maple_type mt,unsigned char * off) ma_max_gap() argument 1420 enum maple_type mt; mas_max_gap() local 1592 enum maple_type mt; mas_find_child() local 1741 enum maple_type mt; mas_mab_cp() local 1789 mas_leaf_set_meta(struct maple_node * node,enum maple_type mt,unsigned char end) mas_leaf_set_meta() argument 1807 enum maple_type mt = mte_node_type(mas->node); mab_mas_cp() local 4095 enum maple_type mt; mas_prev_node() local 4265 enum maple_type mt; mas_next_node() local 4739 enum maple_type mt; mas_empty_area() local 4852 mte_dead_leaves(struct maple_enode * enode,struct maple_tree * mt,void __rcu ** slots) mte_dead_leaves() argument 4948 mte_destroy_descend(struct maple_enode ** enode,struct maple_tree * mt,struct maple_enode * prev,unsigned char offset) mte_destroy_descend() argument 4977 mt_destroy_walk(struct maple_enode * enode,struct maple_tree * mt,bool free) mt_destroy_walk() argument 5041 mte_destroy_walk(struct maple_enode * enode,struct maple_tree * mt) mte_destroy_walk() argument 5353 mt_next(struct maple_tree * mt,unsigned long index,unsigned long max) mt_next() argument 5483 mt_prev(struct maple_tree * mt,unsigned long index,unsigned long min) mt_prev() argument 5882 mtree_load(struct maple_tree * mt,unsigned long index) mtree_load() argument 5924 mtree_store_range(struct maple_tree * mt,unsigned long index,unsigned long last,void * entry,gfp_t gfp) mtree_store_range() argument 5955 mtree_store(struct maple_tree * mt,unsigned long index,void * entry,gfp_t gfp) mtree_store() argument 5973 mtree_insert_range(struct maple_tree * mt,unsigned long first,unsigned long last,void * entry,gfp_t gfp) mtree_insert_range() argument 6010 mtree_insert(struct maple_tree * mt,unsigned long index,void * entry,gfp_t gfp) mtree_insert() argument 6017 mtree_alloc_range(struct maple_tree * mt,unsigned long * startp,void * entry,unsigned long size,unsigned long min,unsigned long max,gfp_t gfp) mtree_alloc_range() argument 6079 mtree_alloc_cyclic(struct maple_tree * mt,unsigned long * startp,void * entry,unsigned long range_lo,unsigned long range_hi,unsigned long * next,gfp_t gfp) mtree_alloc_cyclic() argument 6099 mtree_alloc_rrange(struct maple_tree * mt,unsigned long * startp,void * entry,unsigned long size,unsigned long min,unsigned long max,gfp_t gfp) mtree_alloc_rrange() argument 6148 mtree_erase(struct maple_tree * mt,unsigned long index) mtree_erase() argument 6376 __mt_dup(struct maple_tree * mt,struct maple_tree * new,gfp_t gfp) __mt_dup() argument 6413 mtree_dup(struct maple_tree * mt,struct maple_tree * new,gfp_t gfp) mtree_dup() argument 6440 __mt_destroy(struct maple_tree * mt) __mt_destroy() argument 6458 mtree_destroy(struct maple_tree * mt) mtree_destroy() argument 6482 mt_find(struct maple_tree * mt,unsigned long * index,unsigned long max) mt_find() argument 6542 mt_find_after(struct maple_tree * mt,unsigned long * index,unsigned long max) mt_find_after() argument 6704 mt_dump_range64(const struct maple_tree * mt,void * entry,unsigned long min,unsigned long max,unsigned int depth,enum mt_dump_format format) mt_dump_range64() argument 6757 mt_dump_arange64(const struct maple_tree * mt,void * entry,unsigned long min,unsigned long max,unsigned int depth,enum mt_dump_format format) mt_dump_arange64() argument 6816 mt_dump_node(const struct maple_tree * mt,void * entry,unsigned long min,unsigned long max,unsigned int depth,enum mt_dump_format format) mt_dump_node() argument 6851 mt_dump(const struct maple_tree * mt,enum mt_dump_format format) mt_dump() argument 6874 enum maple_type mt = mte_node_type(mas->node); mas_validate_gaps() local 7108 mt_validate_nulls(struct maple_tree * mt) mt_validate_nulls() argument 7150 mt_validate(struct maple_tree * mt) mt_validate() argument [all...] |
| /linux/drivers/input/ |
| H A D | input-mt.c | 45 if (dev->mt) in input_mt_init_slots() 46 return dev->mt->num_slots != num_slots ? -EINVAL : 0; in input_mt_init_slots() 52 struct input_mt *mt __free(kfree) = in input_mt_init_slots() 53 kzalloc_flex(*mt, slots, num_slots); in input_mt_init_slots() 54 if (!mt) in input_mt_init_slots() 57 mt->num_slots = num_slots; in input_mt_init_slots() 58 mt->flags = flags; in input_mt_init_slots() 87 mt->red = kzalloc_objs(*mt->red, n2); in input_mt_init_slots() 88 if (!mt->red) in input_mt_init_slots() 94 input_mt_set_value(&mt->slots[i], ABS_MT_TRACKING_ID, -1); in input_mt_init_slots() [all …]
|
| /linux/drivers/thermal/mediatek/ |
| H A D | auxadc_thermal.c | 307 struct mtk_thermal *mt; member 350 int (*raw_to_mcelsius)(struct mtk_thermal *mt, int sensno, s32 raw); 712 static int raw_to_mcelsius_v1(struct mtk_thermal *mt, int sensno, s32 raw) in raw_to_mcelsius_v1() argument 719 tmp /= mt->conf->cali_val + mt->o_slope; in raw_to_mcelsius_v1() 720 tmp /= 10000 + mt->adc_ge; in raw_to_mcelsius_v1() 721 tmp *= raw - mt->vts[sensno] - 3350; in raw_to_mcelsius_v1() 724 return mt->degc_cali * 500 - tmp; in raw_to_mcelsius_v1() 727 static int raw_to_mcelsius_v2(struct mtk_thermal *mt, int sensno, s32 raw) in raw_to_mcelsius_v2() argument 740 g_gain = 10000 + (((mt->adc_ge - 512) * 10000) >> 12); in raw_to_mcelsius_v2() 741 g_oe = mt->adc_oe - 512; in raw_to_mcelsius_v2() [all …]
|
| /linux/drivers/net/ethernet/microchip/vcap/ |
| H A D | vcap_tc.c | 68 struct flow_match_ipv4_addrs mt; in vcap_tc_flower_handler_ipv4_usage() local 70 flow_rule_match_ipv4_addrs(st->frule, &mt); in vcap_tc_flower_handler_ipv4_usage() 71 if (mt.mask->src) { in vcap_tc_flower_handler_ipv4_usage() 74 be32_to_cpu(mt.key->src), in vcap_tc_flower_handler_ipv4_usage() 75 be32_to_cpu(mt.mask->src)); in vcap_tc_flower_handler_ipv4_usage() 79 if (mt.mask->dst) { in vcap_tc_flower_handler_ipv4_usage() 82 be32_to_cpu(mt.key->dst), in vcap_tc_flower_handler_ipv4_usage() 83 be32_to_cpu(mt.mask->dst)); in vcap_tc_flower_handler_ipv4_usage() 104 struct flow_match_ipv6_addrs mt; in vcap_tc_flower_handler_ipv6_usage() local 108 flow_rule_match_ipv6_addrs(st->frule, &mt); in vcap_tc_flower_handler_ipv6_usage() [all …]
|
| /linux/drivers/base/regmap/ |
| H A D | regcache-maple.c | 19 struct maple_tree *mt = map->cache; in regcache_maple_read() local 20 MA_STATE(mas, mt, reg, reg); in regcache_maple_read() 41 struct maple_tree *mt = map->cache; in regcache_maple_write() local 42 MA_STATE(mas, mt, reg, reg); in regcache_maple_write() 110 struct maple_tree *mt = map->cache; in regcache_maple_drop() local 111 MA_STATE(mas, mt, min, max); in regcache_maple_drop() 241 struct maple_tree *mt = map->cache; in regcache_maple_sync() local 243 MA_STATE(mas, mt, min, max); in regcache_maple_sync() 295 struct maple_tree *mt; in regcache_maple_init() local 297 mt = kmalloc_obj(*mt, map->alloc_flags); in regcache_maple_init() [all …]
|
| /linux/drivers/video/fbdev/matrox/ |
| H A D | matroxfb_maven.c | 756 struct my_timming* mt, in maven_compute_timming() argument 772 if (maven_find_exact_clocks(mt->HTotal, mt->VTotal, m) == 0) in maven_compute_timming() 775 lmargin = mt->HTotal - mt->HSyncEnd; in maven_compute_timming() 776 slen = mt->HSyncEnd - mt->HSyncStart; in maven_compute_timming() 777 hcrt = mt->HTotal - slen - mt->delay; in maven_compute_timming() 778 umargin = mt->VTotal - mt->VSyncEnd; in maven_compute_timming() 779 vslen = mt->VSyncEnd - mt->VSyncStart; in maven_compute_timming() 781 if (m->hcorr < mt->HTotal) in maven_compute_timming() 783 if (hcrt > mt->HTotal) in maven_compute_timming() 784 hcrt -= mt->HTotal; in maven_compute_timming() [all …]
|
| H A D | matroxfb_crtc2.c | 65 struct my_timming* mt, in matroxfb_dh_restore() argument 105 if (mt->interlaced) { in matroxfb_dh_restore() 107 mt->VDisplay >>= 1; in matroxfb_dh_restore() 108 mt->VSyncStart >>= 1; in matroxfb_dh_restore() 109 mt->VSyncEnd >>= 1; in matroxfb_dh_restore() 110 mt->VTotal >>= 1; in matroxfb_dh_restore() 112 if ((mt->HTotal & 7) == 2) { in matroxfb_dh_restore() 114 mt->HTotal &= ~7; in matroxfb_dh_restore() 117 mga_outl(0x3C14, ((mt->HDisplay - 8) << 16) | (mt->HTotal - 8)); in matroxfb_dh_restore() 118 mga_outl(0x3C18, ((mt->HSyncEnd - 8) << 16) | (mt->HSyncStart - 8)); in matroxfb_dh_restore() [all …]
|
| /linux/tools/testing/radix-tree/ |
| H A D | maple.c | 24 #define RCU_MT_BUG_ON(test, y) {if (y) { test->stop = true; } MT_BUG_ON(test->mt, y); } 27 struct maple_tree *mt; member 45 struct maple_tree *mt; 64 static noinline void __init check_erase(struct maple_tree *mt, unsigned long index, in check_erase() 67 MT_BUG_ON(mt, mtree_test_erase(mt, index) != ptr); in check_erase() 70 #define erase_check_load(mt, i) check_load(mt, set[i], entry[i%2]) argument 71 #define erase_check_insert(mt, i) check_insert(mt, se argument 44 struct maple_tree *mt; global() member 63 check_erase(struct maple_tree * mt,unsigned long index,void * ptr) check_erase() argument 69 erase_check_load(mt,i) global() argument 73 check_erase_testset(struct maple_tree * mt) check_erase_testset() argument 330 erase_check_store_range(mt,a,i,ptr) global() argument 580 check_erase2_testset(struct maple_tree * mt,const unsigned long * set,unsigned long size) check_erase2_testset() argument 725 check_erase2_sets(struct maple_tree * mt) check_erase2_sets() argument 34187 rcu_stress_rev(struct maple_tree * mt,struct rcu_test_struct2 * test,int count,struct rcu_reader_struct * test_reader) rcu_stress_rev() argument 34246 rcu_stress_fwd(struct maple_tree * mt,struct rcu_test_struct2 * test,int count,struct rcu_reader_struct * test_reader) rcu_stress_fwd() argument 34314 rcu_stress(struct maple_tree * mt,bool forward) rcu_stress() argument 34391 struct maple_tree *mt; /* the maple tree */ global() member 34533 run_check_rcu(struct maple_tree * mt,struct rcu_test_struct * vals) run_check_rcu() argument 34585 run_check_rcu_slot_store(struct maple_tree * mt) run_check_rcu_slot_store() argument 34630 run_check_rcu_slowread(struct maple_tree * mt,struct rcu_test_struct * vals) run_check_rcu_slowread() argument 34672 check_rcu_simulated(struct maple_tree * mt) check_rcu_simulated() argument 34833 check_rcu_threaded(struct maple_tree * mt) check_rcu_threaded() argument 34982 check_dfs_preorder(struct maple_tree * mt) check_dfs_preorder() argument 35099 check_prealloc(struct maple_tree * mt) check_prealloc() argument 35249 check_spanning_write(struct maple_tree * mt) check_spanning_write() argument 35417 check_null_expand(struct maple_tree * mt) check_null_expand() argument 35478 check_nomem(struct maple_tree * mt) check_nomem() argument 35514 check_locky(struct maple_tree * mt) check_locky() argument 35683 build_full_tree(struct maple_tree * mt,unsigned int flags,int height) build_full_tree() argument 35736 check_mtree_dup(struct maple_tree * mt) check_mtree_dup() argument 35888 check_spanning_store_height(struct maple_tree * mt) check_spanning_store_height() argument 35913 check_collapsing_rebalance(struct maple_tree * mt) check_collapsing_rebalance() argument 35935 struct maple_tree *mt = (struct maple_tree *)maple_tree; writer2() local 35953 check_nomem_writer_race(struct maple_tree * mt) check_nomem_writer_race() argument 36001 check_vma_modification(struct maple_tree * mt) check_vma_modification() argument 36129 enum maple_type mt = mte_node_type(mas->node); get_last_index() local [all...] |
| /linux/drivers/net/ethernet/mellanox/mlx5/core/steering/hws/ |
| H A D | matcher.c | 472 struct mlx5hws_match_template *mt = matcher->mt; in hws_matcher_create_rtc() local 484 rtc_attr.is_frst_jumbo = mlx5hws_matcher_mt_is_jumbo(mt); in hws_matcher_create_rtc() 494 rtc_attr.match_definer_0 = mlx5hws_definer_get_id(mt->definer); in hws_matcher_create_rtc() 506 mlx5hws_definer_get_id(mt->definer); in hws_matcher_create_rtc() 631 bool is_jumbo = mlx5hws_matcher_mt_is_jumbo(matcher->mt); in hws_matcher_bind_at() 664 for (i = 0; i < matcher->mt->fc_sz; i++) { in hws_matcher_set_ip_version_match() 665 switch (matcher->mt->fc[i].fname) { in hws_matcher_set_ip_version_match() 693 ret = mlx5hws_definer_mt_init(ctx, matcher->mt); in hws_matcher_bind_mt() 738 mlx5hws_definer_mt_uninit(ctx, matcher->mt); in hws_matcher_bind_mt() 749 mlx5hws_definer_mt_uninit(ctx, matcher->mt); in hws_matcher_unbind_mt() [all …]
|
| H A D | definer.c | 1746 struct mlx5hws_match_template *mt, in hws_definer_conv_match_params_to_hl() argument 1760 if (mt->match_criteria_enable & MLX5HWS_DEFINER_MATCH_CRITERIA_MISC6) { in hws_definer_conv_match_params_to_hl() 1766 if (mt->match_criteria_enable & MLX5HWS_DEFINER_MATCH_CRITERIA_OUTER) { in hws_definer_conv_match_params_to_hl() 1767 ret = hws_definer_conv_outer(&cd, mt->match_param); in hws_definer_conv_match_params_to_hl() 1772 if (mt->match_criteria_enable & MLX5HWS_DEFINER_MATCH_CRITERIA_INNER) { in hws_definer_conv_match_params_to_hl() 1773 ret = hws_definer_conv_inner(&cd, mt->match_param); in hws_definer_conv_match_params_to_hl() 1778 if (mt->match_criteria_enable & MLX5HWS_DEFINER_MATCH_CRITERIA_MISC) { in hws_definer_conv_match_params_to_hl() 1779 ret = hws_definer_conv_misc(&cd, mt->match_param); in hws_definer_conv_match_params_to_hl() 1784 if (mt->match_criteria_enable & MLX5HWS_DEFINER_MATCH_CRITERIA_MISC2) { in hws_definer_conv_match_params_to_hl() 1785 ret = hws_definer_conv_misc2(&cd, mt->match_param); in hws_definer_conv_match_params_to_hl() [all …]
|
| H A D | matcher.h | 64 struct mlx5hws_match_template *mt; member 86 mlx5hws_matcher_mt_is_jumbo(struct mlx5hws_match_template *mt) in mlx5hws_matcher_mt_is_jumbo() argument 88 return mlx5hws_definer_is_jumbo(mt->definer); in mlx5hws_matcher_mt_is_jumbo()
|
| H A D | rule.c | 55 struct mlx5hws_match_template *mt, in hws_rule_init_dep_wqe() argument 165 struct mlx5hws_match_template *mt = rule->matcher->mt; in hws_rule_save_delete_info() local 166 bool is_jumbo = mlx5hws_matcher_mt_is_jumbo(mt); in hws_rule_save_delete_info() 293 struct mlx5hws_match_template *mt = &rule->matcher->mt[mt_idx]; in hws_rule_create_hws() local 294 bool is_jumbo = mlx5hws_matcher_mt_is_jumbo(mt); in hws_rule_create_hws() 323 hws_rule_init_dep_wqe(dep_wqe, rule, mt, attr); in hws_rule_create_hws() 367 mlx5hws_definer_create_tag(match_param, mt->fc, mt->fc_sz, in hws_rule_create_hws() 448 bool is_jumbo = mlx5hws_matcher_mt_is_jumbo(rule->matcher->mt); in hws_rule_destroy_hws() 574 bool is_jumbo = mlx5hws_matcher_mt_is_jumbo(rule->matcher->mt); in mlx5hws_rule_move_hws_remove() 611 bool is_jumbo = mlx5hws_matcher_mt_is_jumbo(rule->matcher->mt); in mlx5hws_rule_move_hws_add()
|
| H A D | debug.c | 51 struct mlx5hws_match_template *mt = &matcher->mt[i]; in hws_debug_dump_matcher_match_template() local 55 HWS_PTR_TO_ID(mt), in hws_debug_dump_matcher_match_template() 57 mt->fc_sz, in hws_debug_dump_matcher_match_template() 61 ret = hws_debug_dump_matcher_template_definer(f, mt, mt->definer, type); in hws_debug_dump_matcher_match_template()
|
| /linux/drivers/input/mouse/ |
| H A D | alps.c | 496 fields->mt[0] = fields->st; in alps_process_bitmap() 497 fields->mt[1] = corner[priv->second_touch]; in alps_process_bitmap() 517 input_mt_assign_slots(dev, slot, f->mt, n, 0); in alps_report_mt_data() 519 alps_set_slot(dev, slot[i], f->mt[i].x, f->mt[i].y); in alps_report_mt_data() 532 f->mt[0].x = f->st.x; in alps_report_semi_mt_data() 533 f->mt[0].y = f->st.y; in alps_report_semi_mt_data() 539 alps_set_slot(dev, 0, f->mt[0].x, f->mt[0].y); in alps_report_semi_mt_data() 541 alps_set_slot(dev, 1, f->mt[1].x, f->mt[1].y); in alps_report_semi_mt_data() 969 static void alps_get_finger_coordinate_v7(struct input_mt_pos *mt, in alps_get_finger_coordinate_v7() argument 973 mt[0].x = ((pkt[2] & 0x80) << 4); in alps_get_finger_coordinate_v7() [all …]
|
| /linux/net/netfilter/ |
| H A D | nft_set_pipapo.c | 351 * @mt: Mapping table containing bit set specifiers 364 const union nft_pipapo_map_bucket *mt, bool match_only) in pipapo_refill() argument 389 bitmap_set(dst, mt[i].to, mt[i].n); in pipapo_refill() 463 b = pipapo_refill(res_map, f->bsize, f->rules, fill_map, f->mt, in pipapo_get_slow() 476 e = f->mt[b].e; in pipapo_get_slow() 622 union nft_pipapo_map_bucket *new_mt = NULL, *old_mt = f->mt; in pipapo_realloc_mt() 666 f->mt = new_mt; in pipapo_realloc_mt() 727 goto mt; in pipapo_resize() 758 mt in pipapo_resize() 1583 pipapo_unmap(union nft_pipapo_map_bucket * mt,unsigned int rules,unsigned int start,unsigned int n,unsigned int to_offset,bool is_last) global() argument [all...] |
| H A D | nft_set_pipapo_avx2.c | 141 * @mt: Mapping table containing bit set specifiers 154 union nft_pipapo_map_bucket *mt, bool last) in nft_pipapo_avx2_refill() argument 167 nft_pipapo_avx2_fill(dst, mt[i].to, mt[i].n); \ in nft_pipapo_avx2_refill() 170 ret = mt[i].to; \ in nft_pipapo_avx2_refill() 243 b = nft_pipapo_avx2_refill(i_ul, &map[i_ul], fill, f->mt, last); in nft_pipapo_avx2_lookup_4b_2() 320 b = nft_pipapo_avx2_refill(i_ul, &map[i_ul], fill, f->mt, last); in nft_pipapo_avx2_lookup_4b_4() 415 b = nft_pipapo_avx2_refill(i_ul, &map[i_ul], fill, f->mt, last); in nft_pipapo_avx2_lookup_4b_8() 506 b = nft_pipapo_avx2_refill(i_ul, &map[i_ul], fill, f->mt, last); in nft_pipapo_avx2_lookup_4b_12() 642 b = nft_pipapo_avx2_refill(i_ul, &map[i_ul], fill, f->mt, las in nft_pipapo_avx2_lookup_4b_32() [all...] |
| /linux/mm/ |
| H A D | page_reporting.c | 116 int mt = get_pageblock_migratetype(page); in page_reporting_drain() local 119 __putback_isolated_page(page, order, mt); in page_reporting_drain() 147 unsigned int order, unsigned int mt, in page_reporting_cycle() argument 151 struct list_head *list = &area->free_list[mt]; in page_reporting_cycle() 263 unsigned int order, mt, leftover, offset = PAGE_REPORTING_CAPACITY; in page_reporting_process_zone() local 278 /* Process each free list starting from lowest order/mt */ in page_reporting_process_zone() 280 for (mt = 0; mt < MIGRATE_TYPES; mt++) { in page_reporting_process_zone() 282 if (is_migrate_isolate(mt)) in page_reporting_process_zone() [all...] |
| /linux/scripts/gdb/linux/ |
| H A D | mapletree.py | 35 def __init__(self, mt, first, end): argument 36 if mt.type == maple_tree_root_type.get_type().pointer(): 37 self.tree = mt.dereference() 38 elif mt.type != maple_tree_root_type.get_type(): 40 .format(maple_tree_root_type.get_type().pointer(), mt.type)) 41 self.tree = mt 194 def mt_slot(mt, slots, offset): argument 227 def mtree_load(mt, index): argument 230 mas = Mas(mt, index, index)
|
| /linux/include/trace/events/ |
| H A D | page_ref.h | 25 __field(int, mt) 35 __entry->mt = get_pageblock_migratetype(page); 43 __entry->mapcount, __entry->mapping, __entry->mt, 73 __field(int, mt) 84 __entry->mt = get_pageblock_migratetype(page); 93 __entry->mapcount, __entry->mapping, __entry->mt,
|
| /linux/drivers/iommu/generic_pt/ |
| H A D | kunit_iommu_pt.h | 286 static void unmap_collisions(struct kunit *test, struct maple_tree *mt, in unmap_collisions() argument 290 MA_STATE(mas, mt, start, last); in unmap_collisions() 293 mtree_lock(mt); in unmap_collisions() 301 mtree_unlock(mt); in unmap_collisions() 306 mtree_lock(mt); in unmap_collisions() 308 mtree_unlock(mt); in unmap_collisions() 333 struct maple_tree mt; in test_random_map() local 336 mt_init(&mt); in test_random_map() 375 unmap_collisions(test, &mt, start, end - 1); in test_random_map() 380 mtree_insert_range(&mt, start, end - 1, in test_random_map() [all …]
|
| /linux/drivers/slimbus/ |
| H A D | slimbus.h | 129 u8 mt; member 436 static inline bool slim_tid_txn(u8 mt, u8 mc) in slim_tid_txn() argument 438 return (mt == SLIM_MSG_MT_CORE && in slim_tid_txn() 445 static inline bool slim_ec_txn(u8 mt, u8 mc) in slim_ec_txn() argument 447 return (mt == SLIM_MSG_MT_CORE && in slim_ec_txn()
|
| H A D | qcom-ngd-ctrl.c | 94 #define SLIM_MSG_ASM_FIRST_WORD(l, mt, mc, dt, ad) \ argument 95 ((l) | ((mt) << 5) | ((mc) << 8) | ((dt) << 15) | ((ad) << 16)) 604 u8 mc, mt, len; in qcom_slim_ngd_rx() local 606 mt = SLIM_HEADER_GET_MT(buf[0]); in qcom_slim_ngd_rx() 611 mt == SLIM_MSG_MT_SRC_REFERRED_USER) in qcom_slim_ngd_rx() 616 mt == SLIM_MSG_MT_SRC_REFERRED_USER) || in qcom_slim_ngd_rx() 618 mt == SLIM_MSG_MT_SRC_REFERRED_USER)) { in qcom_slim_ngd_rx() 800 if (txn->mt == SLIM_MSG_MT_CORE && in qcom_slim_ngd_xfer_msg() 820 if (txn->mt == SLIM_MSG_MT_CORE && in qcom_slim_ngd_xfer_msg() 824 txn->mt = SLIM_MSG_MT_DEST_REFERRED_USER; in qcom_slim_ngd_xfer_msg() [all …]
|
| /linux/arch/mips/kernel/ |
| H A D | Makefile | 58 obj-$(CONFIG_MIPS_MT) += mips-mt.o 59 obj-$(CONFIG_MIPS_MT_FPAFF) += mips-mt-fpaff.o 60 obj-$(CONFIG_MIPS_MT_SMP) += smp-mt.o 66 obj-$(CONFIG_MIPS_VPE_LOADER_MT) += vpe-mt.o 68 obj-$(CONFIG_MIPS_VPE_APSP_API_MT) += rtlx-mt.o
|
| /linux/rust/helpers/ |
| H A D | maple_tree.c | 5 __rust_helper void rust_helper_mt_init_flags(struct maple_tree *mt, in rust_helper_mt_init_flags() argument 8 mt_init_flags(mt, flags); in rust_helper_mt_init_flags()
|