Home
last modified time | relevance | path

Searched refs:mt (Results 1 – 25 of 80) sorted by relevance

1234

/linux/lib/
H A Dtest_maple_tree.c19 #define mt_dump(mt, fmt) do {} while (0) argument
20 #define mt_validate(mt) do {} while (0) argument
63 static int __init mtree_insert_index(struct maple_tree *mt, in mtree_insert_index() argument
66 return mtree_insert(mt, index, xa_mk_value(index & LONG_MAX), gfp); in mtree_insert_index()
69 static void __init mtree_erase_index(struct maple_tree *mt, unsigned long index) in mtree_erase_index() argument
71 MT_BUG_ON(mt, mtree_erase(mt, index) != xa_mk_value(index & LONG_MAX)); in mtree_erase_index()
72 MT_BUG_ON(mt, mtree_load(mt, index) != NULL); in mtree_erase_index()
75 static int __init mtree_test_insert(struct maple_tree *mt, unsigne argument
81 mtree_test_store_range(struct maple_tree * mt,unsigned long start,unsigned long end,void * ptr) mtree_test_store_range() argument
87 mtree_test_store(struct maple_tree * mt,unsigned long start,void * ptr) mtree_test_store() argument
93 mtree_test_insert_range(struct maple_tree * mt,unsigned long start,unsigned long end,void * ptr) mtree_test_insert_range() argument
99 mtree_test_load(struct maple_tree * mt,unsigned long index) mtree_test_load() argument
104 mtree_test_erase(struct maple_tree * mt,unsigned long index) mtree_test_erase() argument
110 check_mtree_alloc_range(struct maple_tree * mt,unsigned long start,unsigned long end,unsigned long size,unsigned long expected,int eret,void * ptr) check_mtree_alloc_range() argument
127 check_mtree_alloc_rrange(struct maple_tree * mt,unsigned long start,unsigned long end,unsigned long size,unsigned long expected,int eret,void * ptr) check_mtree_alloc_rrange() argument
145 check_load(struct maple_tree * mt,unsigned long index,void * ptr) check_load() argument
155 check_store_range(struct maple_tree * mt,unsigned long start,unsigned long end,void * ptr,int expected) check_store_range() argument
171 check_insert_range(struct maple_tree * mt,unsigned long start,unsigned long end,void * ptr,int expected) check_insert_range() argument
187 check_insert(struct maple_tree * mt,unsigned long index,void * ptr) check_insert() argument
196 check_dup_insert(struct maple_tree * mt,unsigned long index,void * ptr) check_dup_insert() argument
206 check_index_load(struct maple_tree * mt,unsigned long index) check_index_load() argument
227 check_rev_seq(struct maple_tree * mt,unsigned long max,bool verbose) check_rev_seq() argument
260 check_seq(struct maple_tree * mt,unsigned long max,bool verbose) check_seq() argument
289 check_lb_not_empty(struct maple_tree * mt) check_lb_not_empty() argument
308 check_lower_bound_split(struct maple_tree * mt) check_lower_bound_split() argument
314 check_upper_bound_split(struct maple_tree * mt) check_upper_bound_split() argument
339 check_mid_split(struct maple_tree * mt) check_mid_split() argument
348 check_rev_find(struct maple_tree * mt) check_rev_find() argument
387 check_find(struct maple_tree * mt) check_find() argument
604 check_find_2(struct maple_tree * mt) check_find_2() argument
649 check_alloc_rev_range(struct maple_tree * mt) check_alloc_rev_range() argument
823 check_alloc_range(struct maple_tree * mt) check_alloc_range() argument
989 check_ranges(struct maple_tree * mt) check_ranges() argument
1322 check_next_entry(struct maple_tree * mt) check_next_entry() argument
1346 check_prev_entry(struct maple_tree * mt) check_prev_entry() argument
1390 check_store_null(struct maple_tree * mt) check_store_null() argument
1476 check_root_expand(struct maple_tree * mt) check_root_expand() argument
1566 check_deficient_node(struct maple_tree * mt) check_deficient_node() argument
1590 check_gap_combining(struct maple_tree * mt) check_gap_combining() argument
1787 check_node_overwrite(struct maple_tree * mt) check_node_overwrite() argument
1800 bench_slot_store(struct maple_tree * mt) bench_slot_store() argument
1816 bench_node_store(struct maple_tree * mt) bench_node_store() argument
1835 bench_awalk(struct maple_tree * mt) bench_awalk() argument
1852 bench_walk(struct maple_tree * mt) bench_walk() argument
1869 bench_load(struct maple_tree * mt) bench_load() argument
1882 bench_mt_for_each(struct maple_tree * mt) bench_mt_for_each() argument
1906 bench_mas_for_each(struct maple_tree * mt) bench_mas_for_each() argument
1936 bench_mas_prev(struct maple_tree * mt) bench_mas_prev() argument
1968 struct maple_tree mt, newmt; check_forking() local
2010 check_iteration(struct maple_tree * mt) check_iteration() argument
2077 check_mas_store_gfp(struct maple_tree * mt) check_mas_store_gfp() argument
2112 struct maple_tree mt, newmt; bench_forking() local
2158 next_prev_test(struct maple_tree * mt) next_prev_test() argument
2340 check_spanning_relatives(struct maple_tree * mt) check_spanning_relatives() argument
2353 check_fuzzer(struct maple_tree * mt) check_fuzzer() argument
2749 check_bnode_min_spanning(struct maple_tree * mt) check_bnode_min_spanning() argument
2768 check_empty_area_window(struct maple_tree * mt) check_empty_area_window() argument
2853 check_empty_area_fill(struct maple_tree * mt) check_empty_area_fill() argument
3041 check_state_handling(struct maple_tree * mt) check_state_handling() argument
3611 alloc_cyclic_testing(struct maple_tree * mt) alloc_cyclic_testing() argument
[all...]
H A Dmaple_tree.c176 static void mt_set_height(struct maple_tree *mt, unsigned char height)
178 unsigned int new_flags = mt->ma_flags; in mt_free_bulk()
181 MT_BUG_ON(mt, height > MAPLE_HEIGHT_MAX);
183 mt->ma_flags = new_flags; in mt_return_sheaf()
191 static inline unsigned int mt_attr(struct maple_tree *mt)
193 return mt->ma_flags & ~MT_FLAGS_HEIGHT_MASK; in mt_refill_sheaf()
318 const enum maple_type mt)
321 RCU_INIT_POINTER(*slot, (void *)mt_mk_node(mn, mt));
364 static __always_inline bool mt_is_alloc(struct maple_tree *mt) in mte_set_full()
366 return (mt in mte_set_full()
211 mt_set_height(struct maple_tree * mt,unsigned char height) mt_set_height() argument
226 mt_attr(struct maple_tree * mt) mt_attr() argument
392 mt_is_alloc(struct maple_tree * mt) mt_is_alloc() argument
706 ma_slots(struct maple_node * mn,enum maple_type mt) ma_slots() argument
721 mt_write_locked(const struct maple_tree * mt) mt_write_locked() argument
727 mt_locked(const struct maple_tree * mt) mt_locked() argument
733 mt_slot(const struct maple_tree * mt,void __rcu ** slots,unsigned char offset) mt_slot() argument
739 mt_slot_locked(struct maple_tree * mt,void __rcu ** slots,unsigned char offset) mt_slot_locked() argument
783 mt_root_locked(struct maple_tree * mt) mt_root_locked() argument
800 ma_meta(struct maple_node * mn,enum maple_type mt) ma_meta() argument
817 ma_set_meta(struct maple_node * mn,enum maple_type mt,unsigned char offset,unsigned char end) ma_set_meta() argument
832 mt_clear_meta(struct maple_tree * mt,struct maple_node * mn,enum maple_type type) mt_clear_meta() argument
869 ma_meta_end(struct maple_node * mn,enum maple_type mt) ma_meta_end() argument
891 ma_set_meta_gap(struct maple_node * mn,enum maple_type mt,unsigned char offset) ma_set_meta_gap() argument
1309 enum maple_type mt; mas_leaf_max_gap() local
1392 ma_max_gap(struct maple_node * node,unsigned long * gaps,enum maple_type mt,unsigned char * off) ma_max_gap() argument
1420 enum maple_type mt; mas_max_gap() local
1592 enum maple_type mt; mas_find_child() local
1741 enum maple_type mt; mas_mab_cp() local
1789 mas_leaf_set_meta(struct maple_node * node,enum maple_type mt,unsigned char end) mas_leaf_set_meta() argument
1807 enum maple_type mt = mte_node_type(mas->node); mab_mas_cp() local
4095 enum maple_type mt; mas_prev_node() local
4265 enum maple_type mt; mas_next_node() local
4739 enum maple_type mt; mas_empty_area() local
4852 mte_dead_leaves(struct maple_enode * enode,struct maple_tree * mt,void __rcu ** slots) mte_dead_leaves() argument
4948 mte_destroy_descend(struct maple_enode ** enode,struct maple_tree * mt,struct maple_enode * prev,unsigned char offset) mte_destroy_descend() argument
4977 mt_destroy_walk(struct maple_enode * enode,struct maple_tree * mt,bool free) mt_destroy_walk() argument
5041 mte_destroy_walk(struct maple_enode * enode,struct maple_tree * mt) mte_destroy_walk() argument
5353 mt_next(struct maple_tree * mt,unsigned long index,unsigned long max) mt_next() argument
5483 mt_prev(struct maple_tree * mt,unsigned long index,unsigned long min) mt_prev() argument
5882 mtree_load(struct maple_tree * mt,unsigned long index) mtree_load() argument
5924 mtree_store_range(struct maple_tree * mt,unsigned long index,unsigned long last,void * entry,gfp_t gfp) mtree_store_range() argument
5955 mtree_store(struct maple_tree * mt,unsigned long index,void * entry,gfp_t gfp) mtree_store() argument
5973 mtree_insert_range(struct maple_tree * mt,unsigned long first,unsigned long last,void * entry,gfp_t gfp) mtree_insert_range() argument
6010 mtree_insert(struct maple_tree * mt,unsigned long index,void * entry,gfp_t gfp) mtree_insert() argument
6017 mtree_alloc_range(struct maple_tree * mt,unsigned long * startp,void * entry,unsigned long size,unsigned long min,unsigned long max,gfp_t gfp) mtree_alloc_range() argument
6079 mtree_alloc_cyclic(struct maple_tree * mt,unsigned long * startp,void * entry,unsigned long range_lo,unsigned long range_hi,unsigned long * next,gfp_t gfp) mtree_alloc_cyclic() argument
6099 mtree_alloc_rrange(struct maple_tree * mt,unsigned long * startp,void * entry,unsigned long size,unsigned long min,unsigned long max,gfp_t gfp) mtree_alloc_rrange() argument
6148 mtree_erase(struct maple_tree * mt,unsigned long index) mtree_erase() argument
6376 __mt_dup(struct maple_tree * mt,struct maple_tree * new,gfp_t gfp) __mt_dup() argument
6413 mtree_dup(struct maple_tree * mt,struct maple_tree * new,gfp_t gfp) mtree_dup() argument
6440 __mt_destroy(struct maple_tree * mt) __mt_destroy() argument
6458 mtree_destroy(struct maple_tree * mt) mtree_destroy() argument
6482 mt_find(struct maple_tree * mt,unsigned long * index,unsigned long max) mt_find() argument
6542 mt_find_after(struct maple_tree * mt,unsigned long * index,unsigned long max) mt_find_after() argument
6704 mt_dump_range64(const struct maple_tree * mt,void * entry,unsigned long min,unsigned long max,unsigned int depth,enum mt_dump_format format) mt_dump_range64() argument
6757 mt_dump_arange64(const struct maple_tree * mt,void * entry,unsigned long min,unsigned long max,unsigned int depth,enum mt_dump_format format) mt_dump_arange64() argument
6816 mt_dump_node(const struct maple_tree * mt,void * entry,unsigned long min,unsigned long max,unsigned int depth,enum mt_dump_format format) mt_dump_node() argument
6851 mt_dump(const struct maple_tree * mt,enum mt_dump_format format) mt_dump() argument
6874 enum maple_type mt = mte_node_type(mas->node); mas_validate_gaps() local
7108 mt_validate_nulls(struct maple_tree * mt) mt_validate_nulls() argument
7150 mt_validate(struct maple_tree * mt) mt_validate() argument
[all...]
/linux/drivers/input/
H A Dinput-mt.c45 if (dev->mt) in input_mt_init_slots()
46 return dev->mt->num_slots != num_slots ? -EINVAL : 0; in input_mt_init_slots()
52 struct input_mt *mt __free(kfree) = in input_mt_init_slots()
53 kzalloc_flex(*mt, slots, num_slots); in input_mt_init_slots()
54 if (!mt) in input_mt_init_slots()
57 mt->num_slots = num_slots; in input_mt_init_slots()
58 mt->flags = flags; in input_mt_init_slots()
87 mt->red = kzalloc_objs(*mt->red, n2); in input_mt_init_slots()
88 if (!mt->red) in input_mt_init_slots()
94 input_mt_set_value(&mt->slots[i], ABS_MT_TRACKING_ID, -1); in input_mt_init_slots()
[all …]
/linux/drivers/thermal/mediatek/
H A Dauxadc_thermal.c307 struct mtk_thermal *mt; member
350 int (*raw_to_mcelsius)(struct mtk_thermal *mt, int sensno, s32 raw);
712 static int raw_to_mcelsius_v1(struct mtk_thermal *mt, int sensno, s32 raw) in raw_to_mcelsius_v1() argument
719 tmp /= mt->conf->cali_val + mt->o_slope; in raw_to_mcelsius_v1()
720 tmp /= 10000 + mt->adc_ge; in raw_to_mcelsius_v1()
721 tmp *= raw - mt->vts[sensno] - 3350; in raw_to_mcelsius_v1()
724 return mt->degc_cali * 500 - tmp; in raw_to_mcelsius_v1()
727 static int raw_to_mcelsius_v2(struct mtk_thermal *mt, int sensno, s32 raw) in raw_to_mcelsius_v2() argument
740 g_gain = 10000 + (((mt->adc_ge - 512) * 10000) >> 12); in raw_to_mcelsius_v2()
741 g_oe = mt->adc_oe - 512; in raw_to_mcelsius_v2()
[all …]
/linux/drivers/net/ethernet/microchip/vcap/
H A Dvcap_tc.c68 struct flow_match_ipv4_addrs mt; in vcap_tc_flower_handler_ipv4_usage() local
70 flow_rule_match_ipv4_addrs(st->frule, &mt); in vcap_tc_flower_handler_ipv4_usage()
71 if (mt.mask->src) { in vcap_tc_flower_handler_ipv4_usage()
74 be32_to_cpu(mt.key->src), in vcap_tc_flower_handler_ipv4_usage()
75 be32_to_cpu(mt.mask->src)); in vcap_tc_flower_handler_ipv4_usage()
79 if (mt.mask->dst) { in vcap_tc_flower_handler_ipv4_usage()
82 be32_to_cpu(mt.key->dst), in vcap_tc_flower_handler_ipv4_usage()
83 be32_to_cpu(mt.mask->dst)); in vcap_tc_flower_handler_ipv4_usage()
104 struct flow_match_ipv6_addrs mt; in vcap_tc_flower_handler_ipv6_usage() local
108 flow_rule_match_ipv6_addrs(st->frule, &mt); in vcap_tc_flower_handler_ipv6_usage()
[all …]
/linux/drivers/base/regmap/
H A Dregcache-maple.c19 struct maple_tree *mt = map->cache; in regcache_maple_read() local
20 MA_STATE(mas, mt, reg, reg); in regcache_maple_read()
41 struct maple_tree *mt = map->cache; in regcache_maple_write() local
42 MA_STATE(mas, mt, reg, reg); in regcache_maple_write()
110 struct maple_tree *mt = map->cache; in regcache_maple_drop() local
111 MA_STATE(mas, mt, min, max); in regcache_maple_drop()
241 struct maple_tree *mt = map->cache; in regcache_maple_sync() local
243 MA_STATE(mas, mt, min, max); in regcache_maple_sync()
295 struct maple_tree *mt; in regcache_maple_init() local
297 mt = kmalloc_obj(*mt, map->alloc_flags); in regcache_maple_init()
[all …]
/linux/drivers/video/fbdev/matrox/
H A Dmatroxfb_maven.c756 struct my_timming* mt, in maven_compute_timming() argument
772 if (maven_find_exact_clocks(mt->HTotal, mt->VTotal, m) == 0) in maven_compute_timming()
775 lmargin = mt->HTotal - mt->HSyncEnd; in maven_compute_timming()
776 slen = mt->HSyncEnd - mt->HSyncStart; in maven_compute_timming()
777 hcrt = mt->HTotal - slen - mt->delay; in maven_compute_timming()
778 umargin = mt->VTotal - mt->VSyncEnd; in maven_compute_timming()
779 vslen = mt->VSyncEnd - mt->VSyncStart; in maven_compute_timming()
781 if (m->hcorr < mt->HTotal) in maven_compute_timming()
783 if (hcrt > mt->HTotal) in maven_compute_timming()
784 hcrt -= mt->HTotal; in maven_compute_timming()
[all …]
H A Dmatroxfb_crtc2.c65 struct my_timming* mt, in matroxfb_dh_restore() argument
105 if (mt->interlaced) { in matroxfb_dh_restore()
107 mt->VDisplay >>= 1; in matroxfb_dh_restore()
108 mt->VSyncStart >>= 1; in matroxfb_dh_restore()
109 mt->VSyncEnd >>= 1; in matroxfb_dh_restore()
110 mt->VTotal >>= 1; in matroxfb_dh_restore()
112 if ((mt->HTotal & 7) == 2) { in matroxfb_dh_restore()
114 mt->HTotal &= ~7; in matroxfb_dh_restore()
117 mga_outl(0x3C14, ((mt->HDisplay - 8) << 16) | (mt->HTotal - 8)); in matroxfb_dh_restore()
118 mga_outl(0x3C18, ((mt->HSyncEnd - 8) << 16) | (mt->HSyncStart - 8)); in matroxfb_dh_restore()
[all …]
/linux/tools/testing/radix-tree/
H A Dmaple.c24 #define RCU_MT_BUG_ON(test, y) {if (y) { test->stop = true; } MT_BUG_ON(test->mt, y); }
27 struct maple_tree *mt; member
45 struct maple_tree *mt;
64 static noinline void __init check_erase(struct maple_tree *mt, unsigned long index, in check_erase()
67 MT_BUG_ON(mt, mtree_test_erase(mt, index) != ptr); in check_erase()
70 #define erase_check_load(mt, i) check_load(mt, set[i], entry[i%2]) argument
71 #define erase_check_insert(mt, i) check_insert(mt, se argument
44 struct maple_tree *mt; global() member
63 check_erase(struct maple_tree * mt,unsigned long index,void * ptr) check_erase() argument
69 erase_check_load(mt,i) global() argument
73 check_erase_testset(struct maple_tree * mt) check_erase_testset() argument
330 erase_check_store_range(mt,a,i,ptr) global() argument
580 check_erase2_testset(struct maple_tree * mt,const unsigned long * set,unsigned long size) check_erase2_testset() argument
725 check_erase2_sets(struct maple_tree * mt) check_erase2_sets() argument
34187 rcu_stress_rev(struct maple_tree * mt,struct rcu_test_struct2 * test,int count,struct rcu_reader_struct * test_reader) rcu_stress_rev() argument
34246 rcu_stress_fwd(struct maple_tree * mt,struct rcu_test_struct2 * test,int count,struct rcu_reader_struct * test_reader) rcu_stress_fwd() argument
34314 rcu_stress(struct maple_tree * mt,bool forward) rcu_stress() argument
34391 struct maple_tree *mt; /* the maple tree */ global() member
34533 run_check_rcu(struct maple_tree * mt,struct rcu_test_struct * vals) run_check_rcu() argument
34585 run_check_rcu_slot_store(struct maple_tree * mt) run_check_rcu_slot_store() argument
34630 run_check_rcu_slowread(struct maple_tree * mt,struct rcu_test_struct * vals) run_check_rcu_slowread() argument
34672 check_rcu_simulated(struct maple_tree * mt) check_rcu_simulated() argument
34833 check_rcu_threaded(struct maple_tree * mt) check_rcu_threaded() argument
34982 check_dfs_preorder(struct maple_tree * mt) check_dfs_preorder() argument
35099 check_prealloc(struct maple_tree * mt) check_prealloc() argument
35249 check_spanning_write(struct maple_tree * mt) check_spanning_write() argument
35417 check_null_expand(struct maple_tree * mt) check_null_expand() argument
35478 check_nomem(struct maple_tree * mt) check_nomem() argument
35514 check_locky(struct maple_tree * mt) check_locky() argument
35683 build_full_tree(struct maple_tree * mt,unsigned int flags,int height) build_full_tree() argument
35736 check_mtree_dup(struct maple_tree * mt) check_mtree_dup() argument
35888 check_spanning_store_height(struct maple_tree * mt) check_spanning_store_height() argument
35913 check_collapsing_rebalance(struct maple_tree * mt) check_collapsing_rebalance() argument
35935 struct maple_tree *mt = (struct maple_tree *)maple_tree; writer2() local
35953 check_nomem_writer_race(struct maple_tree * mt) check_nomem_writer_race() argument
36001 check_vma_modification(struct maple_tree * mt) check_vma_modification() argument
36129 enum maple_type mt = mte_node_type(mas->node); get_last_index() local
[all...]
/linux/drivers/net/ethernet/mellanox/mlx5/core/steering/hws/
H A Dmatcher.c472 struct mlx5hws_match_template *mt = matcher->mt; in hws_matcher_create_rtc() local
484 rtc_attr.is_frst_jumbo = mlx5hws_matcher_mt_is_jumbo(mt); in hws_matcher_create_rtc()
494 rtc_attr.match_definer_0 = mlx5hws_definer_get_id(mt->definer); in hws_matcher_create_rtc()
506 mlx5hws_definer_get_id(mt->definer); in hws_matcher_create_rtc()
631 bool is_jumbo = mlx5hws_matcher_mt_is_jumbo(matcher->mt); in hws_matcher_bind_at()
664 for (i = 0; i < matcher->mt->fc_sz; i++) { in hws_matcher_set_ip_version_match()
665 switch (matcher->mt->fc[i].fname) { in hws_matcher_set_ip_version_match()
693 ret = mlx5hws_definer_mt_init(ctx, matcher->mt); in hws_matcher_bind_mt()
738 mlx5hws_definer_mt_uninit(ctx, matcher->mt); in hws_matcher_bind_mt()
749 mlx5hws_definer_mt_uninit(ctx, matcher->mt); in hws_matcher_unbind_mt()
[all …]
H A Ddefiner.c1746 struct mlx5hws_match_template *mt, in hws_definer_conv_match_params_to_hl() argument
1760 if (mt->match_criteria_enable & MLX5HWS_DEFINER_MATCH_CRITERIA_MISC6) { in hws_definer_conv_match_params_to_hl()
1766 if (mt->match_criteria_enable & MLX5HWS_DEFINER_MATCH_CRITERIA_OUTER) { in hws_definer_conv_match_params_to_hl()
1767 ret = hws_definer_conv_outer(&cd, mt->match_param); in hws_definer_conv_match_params_to_hl()
1772 if (mt->match_criteria_enable & MLX5HWS_DEFINER_MATCH_CRITERIA_INNER) { in hws_definer_conv_match_params_to_hl()
1773 ret = hws_definer_conv_inner(&cd, mt->match_param); in hws_definer_conv_match_params_to_hl()
1778 if (mt->match_criteria_enable & MLX5HWS_DEFINER_MATCH_CRITERIA_MISC) { in hws_definer_conv_match_params_to_hl()
1779 ret = hws_definer_conv_misc(&cd, mt->match_param); in hws_definer_conv_match_params_to_hl()
1784 if (mt->match_criteria_enable & MLX5HWS_DEFINER_MATCH_CRITERIA_MISC2) { in hws_definer_conv_match_params_to_hl()
1785 ret = hws_definer_conv_misc2(&cd, mt->match_param); in hws_definer_conv_match_params_to_hl()
[all …]
H A Dmatcher.h64 struct mlx5hws_match_template *mt; member
86 mlx5hws_matcher_mt_is_jumbo(struct mlx5hws_match_template *mt) in mlx5hws_matcher_mt_is_jumbo() argument
88 return mlx5hws_definer_is_jumbo(mt->definer); in mlx5hws_matcher_mt_is_jumbo()
H A Drule.c55 struct mlx5hws_match_template *mt, in hws_rule_init_dep_wqe() argument
165 struct mlx5hws_match_template *mt = rule->matcher->mt; in hws_rule_save_delete_info() local
166 bool is_jumbo = mlx5hws_matcher_mt_is_jumbo(mt); in hws_rule_save_delete_info()
293 struct mlx5hws_match_template *mt = &rule->matcher->mt[mt_idx]; in hws_rule_create_hws() local
294 bool is_jumbo = mlx5hws_matcher_mt_is_jumbo(mt); in hws_rule_create_hws()
323 hws_rule_init_dep_wqe(dep_wqe, rule, mt, attr); in hws_rule_create_hws()
367 mlx5hws_definer_create_tag(match_param, mt->fc, mt->fc_sz, in hws_rule_create_hws()
448 bool is_jumbo = mlx5hws_matcher_mt_is_jumbo(rule->matcher->mt); in hws_rule_destroy_hws()
574 bool is_jumbo = mlx5hws_matcher_mt_is_jumbo(rule->matcher->mt); in mlx5hws_rule_move_hws_remove()
611 bool is_jumbo = mlx5hws_matcher_mt_is_jumbo(rule->matcher->mt); in mlx5hws_rule_move_hws_add()
H A Ddebug.c51 struct mlx5hws_match_template *mt = &matcher->mt[i]; in hws_debug_dump_matcher_match_template() local
55 HWS_PTR_TO_ID(mt), in hws_debug_dump_matcher_match_template()
57 mt->fc_sz, in hws_debug_dump_matcher_match_template()
61 ret = hws_debug_dump_matcher_template_definer(f, mt, mt->definer, type); in hws_debug_dump_matcher_match_template()
/linux/drivers/input/mouse/
H A Dalps.c496 fields->mt[0] = fields->st; in alps_process_bitmap()
497 fields->mt[1] = corner[priv->second_touch]; in alps_process_bitmap()
517 input_mt_assign_slots(dev, slot, f->mt, n, 0); in alps_report_mt_data()
519 alps_set_slot(dev, slot[i], f->mt[i].x, f->mt[i].y); in alps_report_mt_data()
532 f->mt[0].x = f->st.x; in alps_report_semi_mt_data()
533 f->mt[0].y = f->st.y; in alps_report_semi_mt_data()
539 alps_set_slot(dev, 0, f->mt[0].x, f->mt[0].y); in alps_report_semi_mt_data()
541 alps_set_slot(dev, 1, f->mt[1].x, f->mt[1].y); in alps_report_semi_mt_data()
969 static void alps_get_finger_coordinate_v7(struct input_mt_pos *mt, in alps_get_finger_coordinate_v7() argument
973 mt[0].x = ((pkt[2] & 0x80) << 4); in alps_get_finger_coordinate_v7()
[all …]
/linux/net/netfilter/
H A Dnft_set_pipapo.c351 * @mt: Mapping table containing bit set specifiers
364 const union nft_pipapo_map_bucket *mt, bool match_only) in pipapo_refill() argument
389 bitmap_set(dst, mt[i].to, mt[i].n); in pipapo_refill()
463 b = pipapo_refill(res_map, f->bsize, f->rules, fill_map, f->mt, in pipapo_get_slow()
476 e = f->mt[b].e; in pipapo_get_slow()
622 union nft_pipapo_map_bucket *new_mt = NULL, *old_mt = f->mt; in pipapo_realloc_mt()
666 f->mt = new_mt; in pipapo_realloc_mt()
727 goto mt; in pipapo_resize()
758 mt in pipapo_resize()
1583 pipapo_unmap(union nft_pipapo_map_bucket * mt,unsigned int rules,unsigned int start,unsigned int n,unsigned int to_offset,bool is_last) global() argument
[all...]
H A Dnft_set_pipapo_avx2.c141 * @mt: Mapping table containing bit set specifiers
154 union nft_pipapo_map_bucket *mt, bool last) in nft_pipapo_avx2_refill() argument
167 nft_pipapo_avx2_fill(dst, mt[i].to, mt[i].n); \ in nft_pipapo_avx2_refill()
170 ret = mt[i].to; \ in nft_pipapo_avx2_refill()
243 b = nft_pipapo_avx2_refill(i_ul, &map[i_ul], fill, f->mt, last); in nft_pipapo_avx2_lookup_4b_2()
320 b = nft_pipapo_avx2_refill(i_ul, &map[i_ul], fill, f->mt, last); in nft_pipapo_avx2_lookup_4b_4()
415 b = nft_pipapo_avx2_refill(i_ul, &map[i_ul], fill, f->mt, last); in nft_pipapo_avx2_lookup_4b_8()
506 b = nft_pipapo_avx2_refill(i_ul, &map[i_ul], fill, f->mt, last); in nft_pipapo_avx2_lookup_4b_12()
642 b = nft_pipapo_avx2_refill(i_ul, &map[i_ul], fill, f->mt, las in nft_pipapo_avx2_lookup_4b_32()
[all...]
/linux/mm/
H A Dpage_reporting.c116 int mt = get_pageblock_migratetype(page); in page_reporting_drain() local
119 __putback_isolated_page(page, order, mt); in page_reporting_drain()
147 unsigned int order, unsigned int mt, in page_reporting_cycle() argument
151 struct list_head *list = &area->free_list[mt]; in page_reporting_cycle()
263 unsigned int order, mt, leftover, offset = PAGE_REPORTING_CAPACITY; in page_reporting_process_zone() local
278 /* Process each free list starting from lowest order/mt */ in page_reporting_process_zone()
280 for (mt = 0; mt < MIGRATE_TYPES; mt++) { in page_reporting_process_zone()
282 if (is_migrate_isolate(mt)) in page_reporting_process_zone()
[all...]
/linux/scripts/gdb/linux/
H A Dmapletree.py35 def __init__(self, mt, first, end): argument
36 if mt.type == maple_tree_root_type.get_type().pointer():
37 self.tree = mt.dereference()
38 elif mt.type != maple_tree_root_type.get_type():
40 .format(maple_tree_root_type.get_type().pointer(), mt.type))
41 self.tree = mt
194 def mt_slot(mt, slots, offset): argument
227 def mtree_load(mt, index): argument
230 mas = Mas(mt, index, index)
/linux/include/trace/events/
H A Dpage_ref.h25 __field(int, mt)
35 __entry->mt = get_pageblock_migratetype(page);
43 __entry->mapcount, __entry->mapping, __entry->mt,
73 __field(int, mt)
84 __entry->mt = get_pageblock_migratetype(page);
93 __entry->mapcount, __entry->mapping, __entry->mt,
/linux/drivers/iommu/generic_pt/
H A Dkunit_iommu_pt.h286 static void unmap_collisions(struct kunit *test, struct maple_tree *mt, in unmap_collisions() argument
290 MA_STATE(mas, mt, start, last); in unmap_collisions()
293 mtree_lock(mt); in unmap_collisions()
301 mtree_unlock(mt); in unmap_collisions()
306 mtree_lock(mt); in unmap_collisions()
308 mtree_unlock(mt); in unmap_collisions()
333 struct maple_tree mt; in test_random_map() local
336 mt_init(&mt); in test_random_map()
375 unmap_collisions(test, &mt, start, end - 1); in test_random_map()
380 mtree_insert_range(&mt, start, end - 1, in test_random_map()
[all …]
/linux/drivers/slimbus/
H A Dslimbus.h129 u8 mt; member
436 static inline bool slim_tid_txn(u8 mt, u8 mc) in slim_tid_txn() argument
438 return (mt == SLIM_MSG_MT_CORE && in slim_tid_txn()
445 static inline bool slim_ec_txn(u8 mt, u8 mc) in slim_ec_txn() argument
447 return (mt == SLIM_MSG_MT_CORE && in slim_ec_txn()
H A Dqcom-ngd-ctrl.c94 #define SLIM_MSG_ASM_FIRST_WORD(l, mt, mc, dt, ad) \ argument
95 ((l) | ((mt) << 5) | ((mc) << 8) | ((dt) << 15) | ((ad) << 16))
604 u8 mc, mt, len; in qcom_slim_ngd_rx() local
606 mt = SLIM_HEADER_GET_MT(buf[0]); in qcom_slim_ngd_rx()
611 mt == SLIM_MSG_MT_SRC_REFERRED_USER) in qcom_slim_ngd_rx()
616 mt == SLIM_MSG_MT_SRC_REFERRED_USER) || in qcom_slim_ngd_rx()
618 mt == SLIM_MSG_MT_SRC_REFERRED_USER)) { in qcom_slim_ngd_rx()
800 if (txn->mt == SLIM_MSG_MT_CORE && in qcom_slim_ngd_xfer_msg()
820 if (txn->mt == SLIM_MSG_MT_CORE && in qcom_slim_ngd_xfer_msg()
824 txn->mt = SLIM_MSG_MT_DEST_REFERRED_USER; in qcom_slim_ngd_xfer_msg()
[all …]
/linux/arch/mips/kernel/
H A DMakefile58 obj-$(CONFIG_MIPS_MT) += mips-mt.o
59 obj-$(CONFIG_MIPS_MT_FPAFF) += mips-mt-fpaff.o
60 obj-$(CONFIG_MIPS_MT_SMP) += smp-mt.o
66 obj-$(CONFIG_MIPS_VPE_LOADER_MT) += vpe-mt.o
68 obj-$(CONFIG_MIPS_VPE_APSP_API_MT) += rtlx-mt.o
/linux/rust/helpers/
H A Dmaple_tree.c5 __rust_helper void rust_helper_mt_init_flags(struct maple_tree *mt, in rust_helper_mt_init_flags() argument
8 mt_init_flags(mt, flags); in rust_helper_mt_init_flags()

1234