Home
last modified time | relevance | path

Searched refs:order (Results 1 – 25 of 1290) sorted by relevance

12345678910>>...52

/linux/scripts/atomic/
H A Dgen-atomic-fallback.sh16 local order="$1"; shift
35 local order="$1"; shift
37 local tmpl_order=${order#_}
39 gen_template_fallback "${tmpl}" "${meta}" "${pfx}" "${name}" "${sfx}" "${order}" "$@"
49 local order="$1"; shift
51 local tmpl="$(find_fallback_template "${pfx}" "${name}" "${sfx}" "${order}")"
52 gen_template_fallback "${tmpl}" "${meta}" "${pfx}" "${name}" "${sfx}" "${order}" "$@"
62 local order="$1"; shift
66 local atomicname="${atomic}_${pfx}${name}${sfx}${order}"
69 local template="$(find_fallback_template "${pfx}" "${name}" "${sfx}" "${order}")"
[all …]
H A Datomic-tbl.sh52 local order="$1"; shift
62 for base in "${pfx}${name}${sfx}${order}" "${pfx}${name}${sfx}" "${name}"; do
194 local order="$1"; shift
198 local atomicname="${atomic}_${pfx}${name}${sfx}${order}"
208 if [ ! -z "${order}" ]; then
235 local order="$1"; shift
237 local atomicname="${atomic}_${pfx}${name}${sfx}${order}"
239 local tmpl="$(find_kerneldoc_template "${pfx}" "${name}" "${sfx}" "${order}")"
245 gen_template_kerneldoc "${tmpl}" "${class}" "${meta}" "${pfx}" "${name}" "${sfx}" "${order}" "$@"
/linux/mm/
H A Dpage_alloc.c208 static void __free_pages_ok(struct page *page, unsigned int order,
210 static void reserve_highatomic_pageblock(struct page *page, int order,
285 static bool page_contains_unaccepted(struct page *page, unsigned int order);
286 static bool cond_accept_memory(struct zone *zone, unsigned int order,
307 _deferred_grow_zone(struct zone *zone, unsigned int order) in _deferred_grow_zone() argument
309 return deferred_grow_zone(zone, order); in _deferred_grow_zone()
312 static inline bool _deferred_grow_zone(struct zone *zone, unsigned int order) in _deferred_grow_zone() argument
640 static inline unsigned int order_to_pindex(int migratetype, int order) in order_to_pindex() argument
645 if (order > PAGE_ALLOC_COSTLY_ORDER) { in order_to_pindex()
646 VM_BUG_ON(!is_pmd_order(order)); in order_to_pindex()
661 int order = pindex / MIGRATE_PCPTYPES; pindex_to_order() local
673 pcp_allowed_order(unsigned int order) pcp_allowed_order() argument
696 prep_compound_page(struct page * page,unsigned int order) prep_compound_page() argument
708 set_buddy_order(struct page * page,unsigned int order) set_buddy_order() argument
727 compaction_capture(struct capture_control * capc,struct page * page,int order,int migratetype) compaction_capture() argument
764 compaction_capture(struct capture_control * capc,struct page * page,int order,int migratetype) compaction_capture() argument
789 __add_to_free_list(struct page * page,struct zone * zone,unsigned int order,int migratetype,bool tail) __add_to_free_list() argument
815 move_to_free_list(struct page * page,struct zone * zone,unsigned int order,int old_mt,int new_mt) move_to_free_list() argument
839 __del_page_from_free_list(struct page * page,struct zone * zone,unsigned int order,int migratetype) __del_page_from_free_list() argument
861 del_page_from_free_list(struct page * page,struct zone * zone,unsigned int order,int migratetype) del_page_from_free_list() argument
884 buddy_merge_likely(unsigned long pfn,unsigned long buddy_pfn,struct page * page,unsigned int order) buddy_merge_likely() argument
936 __free_one_page(struct page * page,unsigned long pfn,struct zone * zone,unsigned int order,int migratetype,fpi_t fpi_flags) __free_one_page() argument
1304 __free_pages_prepare(struct page * page,unsigned int order,fpi_t fpi_flags) __free_pages_prepare() argument
1445 free_pages_prepare(struct page * page,unsigned int order) free_pages_prepare() argument
1460 unsigned int order; free_pcppages_bulk() local
1510 split_large_buddy(struct zone * zone,struct page * page,unsigned long pfn,int order,fpi_t fpi) split_large_buddy() argument
1533 add_page_to_zone_llist(struct zone * zone,struct page * page,unsigned int order) add_page_to_zone_llist() argument
1542 free_one_page(struct zone * zone,struct page * page,unsigned long pfn,unsigned int order,fpi_t fpi_flags) free_one_page() argument
1577 __free_pages_ok(struct page * page,unsigned int order,fpi_t fpi_flags) __free_pages_ok() argument
1587 __free_pages_core(struct page * page,unsigned int order,enum meminit_context context) __free_pages_core() argument
1766 check_new_pages(struct page * page,unsigned int order) check_new_pages() argument
1808 post_alloc_hook(struct page * page,unsigned int order,gfp_t gfp_flags) post_alloc_hook() argument
1863 prep_new_page(struct page * page,unsigned int order,gfp_t gfp_flags,unsigned int alloc_flags) prep_new_page() argument
1888 __rmqueue_smallest(struct zone * zone,unsigned int order,int migratetype) __rmqueue_smallest() argument
1928 __rmqueue_cma_fallback(struct zone * zone,unsigned int order) __rmqueue_cma_fallback() argument
1934 __rmqueue_cma_fallback(struct zone * zone,unsigned int order) __rmqueue_cma_fallback() argument
1946 unsigned int order; __move_freepages_block() local
2050 int order = start_pfn ? __ffs(start_pfn) : MAX_PAGE_ORDER; find_large_buddy() local
2123 int order = buddy_order(buddy); __move_freepages_block_isolate() local
2204 should_try_claim_block(unsigned int order,int start_mt) should_try_claim_block() argument
2252 find_suitable_fallback(struct free_area * area,unsigned int order,int migratetype,bool claimable) find_suitable_fallback() argument
2282 try_to_claim_block(struct zone * zone,struct page * page,int current_order,int order,int start_type,int block_type,unsigned int alloc_flags) try_to_claim_block() argument
2356 __rmqueue_claim(struct zone * zone,int order,int start_migratetype,unsigned int alloc_flags) __rmqueue_claim() argument
2411 __rmqueue_steal(struct zone * zone,int order,int start_migratetype) __rmqueue_steal() argument
2447 __rmqueue(struct zone * zone,unsigned int order,int migratetype,unsigned int alloc_flags,enum rmqueue_mode * mode) __rmqueue() argument
2516 rmqueue_bulk(struct zone * zone,unsigned int order,unsigned long count,struct list_head * list,int migratetype,unsigned int alloc_flags) rmqueue_bulk() argument
2827 free_frozen_page_commit(struct zone * zone,struct per_cpu_pages * pcp,struct page * page,int migratetype,unsigned int order,fpi_t fpi_flags) free_frozen_page_commit() argument
2930 __free_frozen_pages(struct page * page,unsigned int order,fpi_t fpi_flags) __free_frozen_pages() argument
2979 free_frozen_pages(struct page * page,unsigned int order) free_frozen_pages() argument
2984 free_frozen_pages_nolock(struct page * page,unsigned int order) free_frozen_pages_nolock() argument
3002 unsigned int order = folio_order(folio); free_unref_folios() local
3026 unsigned int order = (unsigned long)folio->private; free_unref_folios() local
3084 __split_page(struct page * page,unsigned int order) __split_page() argument
3101 split_page(struct page * page,unsigned int order) split_page() argument
3114 __isolate_free_page(struct page * page,unsigned int order) __isolate_free_page() argument
3164 __putback_isolated_page(struct page * page,unsigned int order,int mt) __putback_isolated_page() argument
3204 rmqueue_buddy(struct zone * preferred_zone,struct zone * zone,unsigned int order,unsigned int alloc_flags,int migratetype) rmqueue_buddy() argument
3255 nr_pcp_alloc(struct per_cpu_pages * pcp,struct zone * zone,int order) nr_pcp_alloc() argument
3307 __rmqueue_pcplist(struct zone * zone,unsigned int order,int migratetype,unsigned int alloc_flags,struct per_cpu_pages * pcp,struct list_head * list) __rmqueue_pcplist() argument
3353 rmqueue_pcplist(struct zone * preferred_zone,struct zone * zone,unsigned int order,int migratetype,unsigned int alloc_flags) rmqueue_pcplist() argument
3395 rmqueue(struct zone * preferred_zone,struct zone * zone,unsigned int order,gfp_t gfp_flags,unsigned int alloc_flags,int migratetype) rmqueue() argument
3428 reserve_highatomic_pageblock(struct page * page,int order,struct zone * zone) reserve_highatomic_pageblock() argument
3488 int order; unreserve_highatomic_pageblock() local
3560 __zone_watermark_unusable_free(struct zone * z,unsigned int order,unsigned int alloc_flags) __zone_watermark_unusable_free() argument
3586 __zone_watermark_ok(struct zone * z,unsigned int order,unsigned long mark,int highest_zoneidx,unsigned int alloc_flags,long free_pages) __zone_watermark_ok() argument
3664 zone_watermark_ok(struct zone * z,unsigned int order,unsigned long mark,int highest_zoneidx,unsigned int alloc_flags) zone_watermark_ok() argument
3671 zone_watermark_fast(struct zone * z,unsigned int order,unsigned long mark,int highest_zoneidx,unsigned int alloc_flags,gfp_t gfp_mask) zone_watermark_fast() argument
3792 get_page_from_freelist(gfp_t gfp_mask,unsigned int order,int alloc_flags,const struct alloc_context * ac) get_page_from_freelist() argument
4028 __alloc_pages_cpuset_fallback(gfp_t gfp_mask,unsigned int order,unsigned int alloc_flags,const struct alloc_context * ac) __alloc_pages_cpuset_fallback() argument
4047 __alloc_pages_may_oom(gfp_t gfp_mask,unsigned int order,const struct alloc_context * ac,unsigned long * did_some_progress) __alloc_pages_may_oom() argument
4142 __alloc_pages_direct_compact(gfp_t gfp_mask,unsigned int order,unsigned int alloc_flags,const struct alloc_context * ac,enum compact_priority prio,enum compact_result * compact_result) __alloc_pages_direct_compact() argument
4201 should_compact_retry(struct alloc_context * ac,int order,int alloc_flags,enum compact_result compact_result,enum compact_priority * compact_priority,int * compaction_retries) should_compact_retry() argument
4267 __alloc_pages_direct_compact(gfp_t gfp_mask,unsigned int order,unsigned int alloc_flags,const struct alloc_context * ac,enum compact_priority prio,enum compact_result * compact_result) __alloc_pages_direct_compact() argument
4276 should_compact_retry(struct alloc_context * ac,int order,int alloc_flags,enum compact_result compact_result,enum compact_priority * compact_priority,int * compaction_retries) should_compact_retry() argument
4388 __perform_reclaim(gfp_t gfp_mask,unsigned int order,const struct alloc_context * ac) __perform_reclaim() argument
4414 __alloc_pages_direct_reclaim(gfp_t gfp_mask,unsigned int order,unsigned int alloc_flags,const struct alloc_context * ac,unsigned long * did_some_progress) __alloc_pages_direct_reclaim() argument
4447 wake_all_kswapds(unsigned int order,gfp_t gfp_mask,const struct alloc_context * ac) wake_all_kswapds() argument
4473 gfp_to_alloc_flags(gfp_t gfp_mask,unsigned int order) gfp_to_alloc_flags() argument
4577 should_reclaim_retry(gfp_t gfp_mask,unsigned order,struct alloc_context * ac,int alloc_flags,bool did_some_progress,int * no_progress_loops) should_reclaim_retry() argument
4687 __alloc_pages_slowpath(gfp_t gfp_mask,unsigned int order,struct alloc_context * ac) __alloc_pages_slowpath() argument
4973 prepare_alloc_pages(gfp_t gfp_mask,unsigned int order,int preferred_nid,nodemask_t * nodemask,struct alloc_context * ac,gfp_t * alloc_gfp,unsigned int * alloc_flags) prepare_alloc_pages() argument
5190 __alloc_frozen_pages_noprof(gfp_t gfp,unsigned int order,int preferred_nid,nodemask_t * nodemask) __alloc_frozen_pages_noprof() argument
5255 __alloc_pages_noprof(gfp_t gfp,unsigned int order,int preferred_nid,nodemask_t * nodemask) __alloc_pages_noprof() argument
5267 __folio_alloc_noprof(gfp_t gfp,unsigned int order,int preferred_nid,nodemask_t * nodemask) __folio_alloc_noprof() argument
5281 get_free_pages_noprof(gfp_t gfp_mask,unsigned int order) get_free_pages_noprof() argument
5298 ___free_pages(struct page * page,unsigned int order,fpi_t fpi_flags) ___free_pages() argument
5343 __free_pages(struct page * page,unsigned int order) __free_pages() argument
5353 free_pages_nolock(struct page * page,unsigned int order) free_pages_nolock() argument
5367 free_pages(unsigned long addr,unsigned int order) free_pages() argument
5377 make_alloc_exact(unsigned long addr,unsigned int order,size_t size) make_alloc_exact() argument
5413 unsigned int order = get_order(size); alloc_pages_exact_noprof() local
5438 unsigned int order = get_order(size); alloc_pages_exact_nid_noprof() local
6843 int order; split_free_frozen_pages() local
6933 const unsigned int order = ilog2(end - start); alloc_contig_frozen_range_noprof() local
7133 unsigned int order; pfn_range_valid_contig() local
7285 const unsigned int order = ilog2(nr_pages); free_contig_frozen_range() local
7375 unsigned int order; __offline_isolated_pages() local
7421 unsigned int order; is_free_buddy_page() local
7437 add_to_free_list(struct page * page,struct zone * zone,unsigned int order,int migratetype,bool tail) add_to_free_list() argument
7482 unsigned int order; take_page_off_buddy() local
7565 page_contains_unaccepted(struct page * page,unsigned int order) page_contains_unaccepted() argument
7620 cond_accept_memory(struct zone * zone,unsigned int order,int alloc_flags) cond_accept_memory() argument
7679 page_contains_unaccepted(struct page * page,unsigned int order) page_contains_unaccepted() argument
7684 cond_accept_memory(struct zone * zone,unsigned int order,int alloc_flags) cond_accept_memory() argument
7698 alloc_frozen_pages_nolock_noprof(gfp_t gfp_flags,int nid,unsigned int order) alloc_frozen_pages_nolock_noprof() argument
7786 alloc_pages_nolock_noprof(gfp_t gfp_flags,int nid,unsigned int order) alloc_pages_nolock_noprof() argument
[all...]
/linux/include/trace/events/
H A Dcompaction.h168 int order,
172 TP_ARGS(order, gfp_mask, prio),
175 __field(int, order)
181 __entry->order = order;
187 __entry->order,
195 int order,
198 TP_ARGS(zone, order, ret),
203 __field(int, order)
210 __entry->order = order;
217 __entry->order,
[all …]
H A Dkmem.h141 TP_PROTO(struct page *page, unsigned int order),
143 TP_ARGS(page, order),
147 __field( unsigned int, order )
152 __entry->order = order;
158 __entry->order)
182 TP_PROTO(struct page *page, unsigned int order,
185 TP_ARGS(page, order, gfp_flags, migratetype),
189 __field( unsigned int, order )
196 __entry->order = order;
204 __entry->order,
[all …]
H A Dmigrate.h115 TP_PROTO(unsigned long addr, unsigned long pte, int order),
117 TP_ARGS(addr, pte, order),
122 __field(int, order)
128 __entry->order = order;
131 TP_printk("addr=%lx, pte=%lx order=%d", __entry->addr, __entry->pte, __entry->order)
135 TP_PROTO(unsigned long addr, unsigned long pte, int order),
136 TP_ARGS(addr, pte, order)
140 TP_PROTO(unsigned long addr, unsigned long pte, int order),
141 TP_ARGS(addr, pte, order)
H A Doom.h37 int order,
44 TP_ARGS(zoneref, order, reclaimable, available, min_wmark, no_progress_loops, wmark_check),
49 __field( int, order)
60 __entry->order = order;
70 __entry->order,
185 TP_PROTO(int order,
192 TP_ARGS(order, priority, result, retries, max_retries, ret),
195 __field( int, order)
204 __entry->order = order;
213 __entry->order,
/linux/lib/
H A Dtest_xarray.c72 unsigned order, void *entry, gfp_t gfp) in xa_store_order() argument
74 XA_STATE_ORDER(xas, xa, index, order); in xa_store_order()
177 unsigned int order; in check_xa_mark_1() local
207 for (order = 2; order < max_order; order++) { in check_xa_mark_1()
208 unsigned long base = round_down(index, 1UL << order); in check_xa_mark_1()
209 unsigned long next = base + (1UL << order); in check_xa_mark_1()
217 xa_store_order(xa, index, order, xa_mk_index(index), in check_xa_mark_1()
328 unsigned int order; in check_xa_shrink() local
353 for (order = 0; order < max_order; order++) { in check_xa_shrink()
354 unsigned long max = (1UL << order) - 1; in check_xa_shrink()
[all …]
H A Dgenalloc.c246 int order = pool->min_alloc_order; in gen_pool_destroy() local
253 end_bit = chunk_size(chunk) >> order; in gen_pool_destroy()
282 int order = pool->min_alloc_order; in gen_pool_alloc_algo_owner() local
295 nbits = (size + (1UL << order) - 1) >> order; in gen_pool_alloc_algo_owner()
302 end_bit = chunk_size(chunk) >> order; in gen_pool_alloc_algo_owner()
316 addr = chunk->start_addr + ((unsigned long)start_bit << order); in gen_pool_alloc_algo_owner()
317 size = nbits << order; in gen_pool_alloc_algo_owner()
491 int order = pool->min_alloc_order; in gen_pool_free_owner() local
501 nbits = (size + (1UL << order) - 1) >> order; in gen_pool_free_owner()
506 start_bit = (addr - chunk->start_addr) >> order; in gen_pool_free_owner()
[all …]
/linux/Documentation/trace/postprocess/
H A Dtrace-vmscan-postprocess.pl315 my $order = $1;
316 $perprocesspid{$process_pid}->{MM_VMSCAN_DIRECT_RECLAIM_BEGIN_PERORDER}[$order]++;
317 $perprocesspid{$process_pid}->{STATE_DIRECT_ORDER} = $order;
326 my $order = $perprocesspid{$process_pid}->{STATE_DIRECT_ORDER};
328 $perprocesspid{$process_pid}->{HIGH_DIRECT_RECLAIM_LATENCY}[$index] = "$order-$latency";
339 my $order = $2;
340 $perprocesspid{$process_pid}->{STATE_KSWAPD_ORDER} = $order;
345 $perprocesspid{$process_pid}->{MM_VMSCAN_KSWAPD_WAKE_PERORDER}[$order]++;
348 $perprocesspid{$process_pid}->{HIGH_KSWAPD_REWAKEUP_PERORDER}[$order]++;
358 my $order = $perprocesspid{$process_pid}->{STATE_KSWAPD_ORDER};
[all …]
/linux/mm/kmsan/
H A Dinit.c116 bool kmsan_memblock_free_pages(struct page *page, unsigned int order) in kmsan_memblock_free_pages() argument
120 if (!held_back[order].shadow) { in kmsan_memblock_free_pages()
121 held_back[order].shadow = page; in kmsan_memblock_free_pages()
124 if (!held_back[order].origin) { in kmsan_memblock_free_pages()
125 held_back[order].origin = page; in kmsan_memblock_free_pages()
128 shadow = held_back[order].shadow; in kmsan_memblock_free_pages()
129 origin = held_back[order].origin; in kmsan_memblock_free_pages()
130 kmsan_setup_meta(page, shadow, origin, order); in kmsan_memblock_free_pages()
132 held_back[order].shadow = NULL; in kmsan_memblock_free_pages()
133 held_back[order].origin = NULL; in kmsan_memblock_free_pages()
[all …]
/linux/drivers/gpu/drm/nouveau/nvkm/subdev/therm/
H A Dgk104.c34 const struct gk104_clkgate_engine_info *order = therm->clkgate_order; in gk104_clkgate_enable() local
38 for (i = 0; order[i].type != NVKM_SUBDEV_NR; i++) { in gk104_clkgate_enable()
39 if (!nvkm_device_subdev(dev, order[i].type, order[i].inst)) in gk104_clkgate_enable()
42 nvkm_mask(dev, 0x20200 + order[i].offset, 0xff00, 0x4500); in gk104_clkgate_enable()
50 for (i = 0; order[i].type != NVKM_SUBDEV_NR; i++) { in gk104_clkgate_enable()
51 if (!nvkm_device_subdev(dev, order[i].type, order[i].inst)) in gk104_clkgate_enable()
54 nvkm_mask(dev, 0x20200 + order[i].offset, 0x00ff, 0x0045); in gk104_clkgate_enable()
63 const struct gk104_clkgate_engine_info *order = therm->clkgate_order; in gk104_clkgate_fini() local
67 for (i = 0; order[i].type != NVKM_SUBDEV_NR; i++) { in gk104_clkgate_fini()
68 if (!nvkm_device_subdev(dev, order[i].type, order[i].inst)) in gk104_clkgate_fini()
[all …]
/linux/include/linux/
H A Dgfp.h223 static inline void arch_free_page(struct page *page, int order) { } in arch_free_page() argument
226 static inline void arch_alloc_page(struct page *page, int order) { } in arch_alloc_page() argument
229 struct page *__alloc_pages_noprof(gfp_t gfp, unsigned int order, int preferred_nid,
233 struct folio *__folio_alloc_noprof(gfp_t gfp, unsigned int order, int preferred_nid,
284 __alloc_pages_node_noprof(int nid, gfp_t gfp_mask, unsigned int order) in __alloc_pages_node_noprof() argument
289 return __alloc_pages_noprof(gfp_mask, order, nid, NULL); in __alloc_pages_node_noprof()
295 struct folio *__folio_alloc_node_noprof(gfp_t gfp, unsigned int order, int nid) in __folio_alloc_node_noprof() argument
300 return __folio_alloc_noprof(gfp, order, nid, NULL); in __folio_alloc_node_noprof()
311 unsigned int order) in alloc_pages_node_noprof() argument
316 return __alloc_pages_node_noprof(nid, gfp_mask, order); in alloc_pages_node_noprof()
[all …]
H A Dcompaction.h65 static inline unsigned long compact_gap(unsigned int order) in compact_gap() argument
80 return 2UL << order; in compact_gap()
90 extern unsigned int extfrag_for_order(struct zone *zone, unsigned int order);
91 extern int fragmentation_index(struct zone *zone, unsigned int order);
93 unsigned int order, unsigned int alloc_flags,
97 extern bool compaction_suitable(struct zone *zone, int order,
100 extern void compaction_defer_reset(struct zone *zone, int order,
103 bool compaction_zonelist_suitable(struct alloc_context *ac, int order,
108 extern void wakeup_kcompactd(pg_data_t *pgdat, int order, int highest_zoneidx);
115 static inline bool compaction_suitable(struct zone *zone, int order, in compaction_suitable() argument
[all …]
/linux/arch/riscv/kvm/
H A Dtlb.c25 unsigned long order) in kvm_riscv_local_hfence_gvma_vmid_gpa() argument
29 if (PTRS_PER_PTE < (gpsz >> order)) { in kvm_riscv_local_hfence_gvma_vmid_gpa()
36 for (pos = gpa; pos < (gpa + gpsz); pos += BIT(order)) in kvm_riscv_local_hfence_gvma_vmid_gpa()
41 for (pos = gpa; pos < (gpa + gpsz); pos += BIT(order)) in kvm_riscv_local_hfence_gvma_vmid_gpa()
53 unsigned long order) in kvm_riscv_local_hfence_gvma_gpa() argument
57 if (PTRS_PER_PTE < (gpsz >> order)) { in kvm_riscv_local_hfence_gvma_gpa()
64 for (pos = gpa; pos < (gpa + gpsz); pos += BIT(order)) in kvm_riscv_local_hfence_gvma_gpa()
69 for (pos = gpa; pos < (gpa + gpsz); pos += BIT(order)) in kvm_riscv_local_hfence_gvma_gpa()
84 unsigned long order) in kvm_riscv_local_hfence_vvma_asid_gva() argument
88 if (PTRS_PER_PTE < (gvsz >> order)) { in kvm_riscv_local_hfence_vvma_asid_gva()
[all …]
/linux/drivers/net/ethernet/mellanox/mlx5/core/steering/sws/
H A Ddr_buddy.c71 unsigned int *order) in dr_buddy_find_free_seg() argument
95 *order = order_iter; in dr_buddy_find_free_seg()
116 unsigned int order, in mlx5dr_buddy_alloc_mem() argument
122 err = dr_buddy_find_free_seg(buddy, order, &seg, &order_iter); in mlx5dr_buddy_alloc_mem()
133 while (order_iter > order) { in mlx5dr_buddy_alloc_mem()
140 seg <<= order; in mlx5dr_buddy_alloc_mem()
147 unsigned int seg, unsigned int order) in mlx5dr_buddy_free_mem() argument
149 seg >>= order; in mlx5dr_buddy_free_mem()
154 while (test_bit(seg ^ 1, buddy->bitmap[order])) { in mlx5dr_buddy_free_mem()
155 bitmap_clear(buddy->bitmap[order], seg ^ 1, 1); in mlx5dr_buddy_free_mem()
[all …]
/linux/drivers/net/ethernet/mellanox/mlx5/core/steering/hws/
H A Dbuddy.c85 u32 *order) in hws_buddy_find_free_seg() argument
109 *order = order_iter; in hws_buddy_find_free_seg()
113 int mlx5hws_buddy_alloc_mem(struct mlx5hws_buddy_mem *buddy, u32 order) in mlx5hws_buddy_alloc_mem() argument
117 err = hws_buddy_find_free_seg(buddy, order, &seg, &order_iter); in mlx5hws_buddy_alloc_mem()
124 while (order_iter > order) { in mlx5hws_buddy_alloc_mem()
131 seg <<= order; in mlx5hws_buddy_alloc_mem()
136 void mlx5hws_buddy_free_mem(struct mlx5hws_buddy_mem *buddy, u32 seg, u32 order) in mlx5hws_buddy_free_mem() argument
138 seg >>= order; in mlx5hws_buddy_free_mem()
140 while (test_bit(seg ^ 1, buddy->bitmap[order])) { in mlx5hws_buddy_free_mem()
141 bitmap_clear(buddy->bitmap[order], seg ^ 1, 1); in mlx5hws_buddy_free_mem()
[all …]
/linux/drivers/gpu/drm/i915/selftests/
H A Di915_syncmap.c274 unsigned int pass, order; in igt_syncmap_join_above() local
296 for (order = 0; order < 64; order += SHIFT) { in igt_syncmap_join_above()
297 u64 context = BIT_ULL(order); in igt_syncmap_join_above()
335 unsigned int step, order, idx; in igt_syncmap_join_below() local
345 for (order = 64 - SHIFT; order > 0; order -= SHIFT) { in igt_syncmap_join_below()
346 u64 context = step * BIT_ULL(order); in igt_syncmap_join_below()
354 context, order, step, sync->height, sync->prefix); in igt_syncmap_join_below()
362 for (order = SHIFT; order < 64; order += SHIFT) { in igt_syncmap_join_below()
363 u64 context = step * BIT_ULL(order); in igt_syncmap_join_below()
367 context, order, step); in igt_syncmap_join_below()
[all …]
H A Di915_random.c70 void i915_random_reorder(unsigned int *order, unsigned int count, in i915_random_reorder() argument
73 i915_prandom_shuffle(order, sizeof(*order), count, state); in i915_random_reorder()
78 unsigned int *order, i; in i915_random_order() local
80 order = kmalloc_array(count, sizeof(*order), in i915_random_order()
82 if (!order) in i915_random_order()
83 return order; in i915_random_order()
86 order[i] = i; in i915_random_order()
88 i915_random_reorder(order, count, state); in i915_random_order()
89 return order; in i915_random_order()
/linux/arch/arm/lib/
H A Dlib1funcs.S106 .macro ARM_DIV2_ORDER divisor, order argument
110 clz \order, \divisor
111 rsb \order, \order, #31
117 movhs \order, #16
118 movlo \order, #0
122 addhs \order, \order, #8
126 addhs \order, \order, #4
129 addhi \order, \order, #3
130 addls \order, \order, \divisor, lsr #1
137 .macro ARM_MOD_BODY dividend, divisor, order, spare
[all …]
/linux/tools/testing/radix-tree/
H A Diteration_check.c25 int order; in my_item_insert() local
29 for (order = max_order; order >= 0; order--) { in my_item_insert()
30 xas_set_order(&xas, index, order); in my_item_insert()
31 item->order = order; in my_item_insert()
41 if (order < 0) in my_item_insert()
165 void iteration_test(unsigned order, unsigned test_duration) in iteration_test() argument
170 order > 0 ? "multiorder " : "", test_duration); in iteration_test()
172 max_order = order; in iteration_test()
/linux/drivers/gpu/drm/
H A Ddrm_pagemap.c183 unsigned int order = 0; in drm_pagemap_migration_unlock_put_pages() local
190 order = folio_order(folio); in drm_pagemap_migration_unlock_put_pages()
196 i += NR_PAGES(order); in drm_pagemap_migration_unlock_put_pages()
210 unsigned int order, in drm_pagemap_get_devmem_page() argument
214 order); in drm_pagemap_get_devmem_page()
249 unsigned int order = 0; in drm_pagemap_migrate_map_pages() local
255 order = folio_order(folio); in drm_pagemap_migrate_map_pages()
266 num_local_pages += NR_PAGES(order); in drm_pagemap_migrate_map_pages()
268 num_peer_pages += NR_PAGES(order); in drm_pagemap_migrate_map_pages()
271 addr = dpagemap->ops->device_map(dpagemap, dev, page, order, dir); in drm_pagemap_migrate_map_pages()
[all …]
/linux/arch/riscv/include/asm/
H A Dkvm_tlb.h25 unsigned long order; member
36 unsigned long order);
39 unsigned long order);
45 unsigned long order);
50 unsigned long order);
65 unsigned long order, unsigned long vmid);
72 unsigned long order, unsigned long asid,
80 unsigned long order, unsigned long vmid);
H A Dpgtable-64.h97 #define for_each_napot_order(order) \ argument
98 for (order = NAPOT_CONT_ORDER_BASE; order < NAPOT_ORDER_MAX; order++)
99 #define for_each_napot_order_rev(order) \ argument
100 for (order = NAPOT_ORDER_MAX - 1; \
101 order >= NAPOT_CONT_ORDER_BASE; order--)
104 #define napot_cont_shift(order) ((order) + PAGE_SHIFT) argument
105 #define napot_cont_size(order) BIT(napot_cont_shift(order)) argument
106 #define napot_cont_mask(order) (~(napot_cont_size(order) - 1UL)) argument
107 #define napot_pte_num(order) BIT(order) argument
/linux/drivers/infiniband/hw/hns/
H A Dhns_roce_db.c94 struct hns_roce_db *db, int order) in hns_roce_alloc_db_from_pgdir() argument
99 for (o = order; o <= 1; ++o) { in hns_roce_alloc_db_from_pgdir()
112 if (o > order) in hns_roce_alloc_db_from_pgdir()
113 set_bit(i ^ 1, pgdir->bits[order]); in hns_roce_alloc_db_from_pgdir()
119 db->order = order; in hns_roce_alloc_db_from_pgdir()
125 int order) in hns_roce_alloc_db() argument
133 if (!hns_roce_alloc_db_from_pgdir(pgdir, db, order)) in hns_roce_alloc_db()
145 WARN_ON(hns_roce_alloc_db_from_pgdir(pgdir, db, order)); in hns_roce_alloc_db()
160 o = db->order; in hns_roce_free_db()
163 if (db->order == 0 && test_bit(i ^ 1, db->u.pgdir->order0)) { in hns_roce_free_db()

12345678910>>...52