Home
last modified time | relevance | path

Searched full:order (Results 1 – 25 of 4321) sorted by relevance

12345678910>>...173

/linux/include/trace/events/
H A Dcompaction.h168 int order,
172 TP_ARGS(order, gfp_mask, prio),
175 __field(int, order)
181 __entry->order = order;
186 TP_printk("order=%d gfp_mask=%s priority=%d",
187 __entry->order,
195 int order,
198 TP_ARGS(zone, order, ret),
203 __field(int, order)
210 __entry->order = order;
[all …]
/linux/Documentation/netlink/specs/
H A Dconntrack.yaml23 byte-order: big-endian
111 byte-order: big-endian
115 byte-order: big-endian
135 byte-order: big-endian
140 byte-order: big-endian
145 byte-order: big-endian
156 byte-order: big-endian
170 byte-order: big-endian
176 byte-order: big-endian
184 byte-order: big-endian
[all …]
H A Dnftables.yaml23 byte-order: big-endian
242 byte-order: big-endian
253 byte-order: big-endian
260 byte-order: big-endian
265 byte-order: big-endian
281 byte-order: big-endian
295 byte-order: big-endian
300 byte-order: big-endian
314 byte-order: big-endian
321 byte-order: big-endian
[all …]
H A Dovs_flow.yaml57 byte-order: big-endian
65 byte-order: big-endian
70 byte-order: big-endian
93 byte-order: big-endian
99 byte-order: big-endian
104 byte-order: big-endian
149 byte-order: big-endian
153 byte-order: big-endian
161 byte-order: big-endian
165 byte-order: big-endian
[all …]
/linux/scripts/atomic/
H A Dgen-atomic-fallback.sh8 #gen_template_fallback(template, meta, pfx, name, sfx, order, atomic, int, args...)
16 local order="$1"; shift
28 #gen_order_fallback(meta, pfx, name, sfx, order, atomic, int, args...)
35 local order="$1"; shift
37 local tmpl_order=${order#_}
39 gen_template_fallback "${tmpl}" "${meta}" "${pfx}" "${name}" "${sfx}" "${order}" "$@"
42 #gen_proto_fallback(meta, pfx, name, sfx, order, atomic, int, args...)
49 local order="$1"; shift
51 local tmpl="$(find_fallback_template "${pfx}" "${name}" "${sfx}" "${order}")"
52 gen_template_fallback "${tmpl}" "${meta}" "${pfx}" "${name}" "${sfx}" "${order}" "$@"
[all …]
H A Dgen-atomic-instrumented.sh37 local order="$1"; shift
39 if [ "${order}" = "_release" ]; then
41 elif [ -z "${order}" ] && ! meta_in "$meta" "slv"; then
52 #gen_proto_order_variant(meta, pfx, name, sfx, order, atomic, int, arg...)
59 local order="$1"; shift
63 local atomicname="${atomic}_${pfx}${name}${sfx}${order}"
67 local checks="$(gen_params_checks "${meta}" "${order}" "$@")"
71 gen_kerneldoc "" "${meta}" "${pfx}" "${name}" "${sfx}" "${order}" "${atomic}" "${int}" "$@"
88 local order="$1"; shift
92 case "$order" in
[all …]
/linux/lib/
H A Dtest_xarray.c72 unsigned order, void *entry, gfp_t gfp) in xa_store_order() argument
74 XA_STATE_ORDER(xas, xa, index, order); in xa_store_order()
177 unsigned int order; in check_xa_mark_1() local
207 for (order = 2; order < max_order; order++) { in check_xa_mark_1()
208 unsigned long base = round_down(index, 1UL << order); in check_xa_mark_1()
209 unsigned long next = base + (1UL << order); in check_xa_mark_1()
217 xa_store_order(xa, index, order, xa_mk_index(index), in check_xa_mark_1()
328 unsigned int order; in check_xa_shrink() local
353 for (order = 0; order < max_order; order++) { in check_xa_shrink()
354 unsigned long max = (1UL << order) - 1; in check_xa_shrink()
[all …]
/linux/mm/
H A Dpage_alloc.c217 static void __free_pages_ok(struct page *page, unsigned int order,
292 static bool page_contains_unaccepted(struct page *page, unsigned int order);
293 static bool cond_accept_memory(struct zone *zone, unsigned int order,
319 _deferred_grow_zone(struct zone *zone, unsigned int order) in _deferred_grow_zone() argument
321 return deferred_grow_zone(zone, order); in _deferred_grow_zone()
329 static inline bool _deferred_grow_zone(struct zone *zone, unsigned int order) in _deferred_grow_zone() argument
658 static inline unsigned int order_to_pindex(int migratetype, int order) in account_freepages()
663 if (order > PAGE_ALLOC_COSTLY_ORDER) { in __add_to_free_list() argument
664 VM_BUG_ON(order != HPAGE_PMD_ORDER); in __add_to_free_list()
671 VM_BUG_ON(order > PAGE_ALLOC_COSTLY_ORDE in __add_to_free_list()
514 order_to_pindex(int migratetype,int order) order_to_pindex() argument
535 int order = pindex / MIGRATE_PCPTYPES; pindex_to_order() local
547 pcp_allowed_order(unsigned int order) pcp_allowed_order() argument
570 prep_compound_page(struct page * page,unsigned int order) prep_compound_page() argument
582 set_buddy_order(struct page * page,unsigned int order) set_buddy_order() argument
601 compaction_capture(struct capture_control * capc,struct page * page,int order,int migratetype) compaction_capture() argument
638 compaction_capture(struct capture_control * capc,struct page * page,int order,int migratetype) compaction_capture() argument
689 move_to_free_list(struct page * page,struct zone * zone,unsigned int order,int old_mt,int new_mt) move_to_free_list() argument
713 __del_page_from_free_list(struct page * page,struct zone * zone,unsigned int order,int migratetype) __del_page_from_free_list() argument
735 del_page_from_free_list(struct page * page,struct zone * zone,unsigned int order,int migratetype) del_page_from_free_list() argument
758 buddy_merge_likely(unsigned long pfn,unsigned long buddy_pfn,struct page * page,unsigned int order) buddy_merge_likely() argument
799 __free_one_page(struct page * page,unsigned long pfn,struct zone * zone,unsigned int order,int migratetype,fpi_t fpi_flags) __free_one_page() argument
1162 free_pages_prepare(struct page * page,unsigned int order) free_pages_prepare() argument
1301 unsigned int order; free_pcppages_bulk() local
1351 split_large_buddy(struct zone * zone,struct page * page,unsigned long pfn,int order,fpi_t fpi) split_large_buddy() argument
1374 add_page_to_zone_llist(struct zone * zone,struct page * page,unsigned int order) add_page_to_zone_llist() argument
1383 free_one_page(struct zone * zone,struct page * page,unsigned long pfn,unsigned int order,fpi_t fpi_flags) free_one_page() argument
1418 __free_pages_ok(struct page * page,unsigned int order,fpi_t fpi_flags) __free_pages_ok() argument
1428 __free_pages_core(struct page * page,unsigned int order,enum meminit_context context) __free_pages_core() argument
1607 check_new_pages(struct page * page,unsigned int order) check_new_pages() argument
1649 post_alloc_hook(struct page * page,unsigned int order,gfp_t gfp_flags) post_alloc_hook() argument
1709 prep_new_page(struct page * page,unsigned int order,gfp_t gfp_flags,unsigned int alloc_flags) prep_new_page() argument
1734 __rmqueue_smallest(struct zone * zone,unsigned int order,int migratetype) __rmqueue_smallest() argument
1774 __rmqueue_cma_fallback(struct zone * zone,unsigned int order) __rmqueue_cma_fallback() argument
1780 __rmqueue_cma_fallback(struct zone * zone,unsigned int order) __rmqueue_cma_fallback() argument
1792 unsigned int order; __move_freepages_block() local
1887 int order = 0; find_large_buddy() local
1943 int order = buddy_order(buddy); move_freepages_block_isolate() local
1954 int order = buddy_order(page); move_freepages_block_isolate() local
2023 should_try_claim_block(unsigned int order,int start_mt) should_try_claim_block() argument
2071 find_suitable_fallback(struct free_area * area,unsigned int order,int migratetype,bool claimable) find_suitable_fallback() argument
2101 try_to_claim_block(struct zone * zone,struct page * page,int current_order,int order,int start_type,int block_type,unsigned int alloc_flags) try_to_claim_block() argument
2174 __rmqueue_claim(struct zone * zone,int order,int start_migratetype,unsigned int alloc_flags) __rmqueue_claim() argument
2229 __rmqueue_steal(struct zone * zone,int order,int start_migratetype) __rmqueue_steal() argument
2265 __rmqueue(struct zone * zone,unsigned int order,int migratetype,unsigned int alloc_flags,enum rmqueue_mode * mode) __rmqueue() argument
2334 rmqueue_bulk(struct zone * zone,unsigned int order,unsigned long count,struct list_head * list,int migratetype,unsigned int alloc_flags) rmqueue_bulk() argument
2635 free_frozen_page_commit(struct zone * zone,struct per_cpu_pages * pcp,struct page * page,int migratetype,unsigned int order,fpi_t fpi_flags) free_frozen_page_commit() argument
2692 __free_frozen_pages(struct page * page,unsigned int order,fpi_t fpi_flags) __free_frozen_pages() argument
2742 free_frozen_pages(struct page * page,unsigned int order) free_frozen_pages() argument
2761 unsigned int order = folio_order(folio); free_unref_folios() local
2785 unsigned int order = (unsigned long)folio->private; free_unref_folios() local
2853 split_page(struct page * page,unsigned int order) split_page() argument
2868 __isolate_free_page(struct page * page,unsigned int order) __isolate_free_page() argument
2918 __putback_isolated_page(struct page * page,unsigned int order,int mt) __putback_isolated_page() argument
2958 rmqueue_buddy(struct zone * preferred_zone,struct zone * zone,unsigned int order,unsigned int alloc_flags,int migratetype) rmqueue_buddy() argument
3002 nr_pcp_alloc(struct per_cpu_pages * pcp,struct zone * zone,int order) nr_pcp_alloc() argument
3054 __rmqueue_pcplist(struct zone * zone,unsigned int order,int migratetype,unsigned int alloc_flags,struct per_cpu_pages * pcp,struct list_head * list) __rmqueue_pcplist() argument
3086 rmqueue_pcplist(struct zone * preferred_zone,struct zone * zone,unsigned int order,int migratetype,unsigned int alloc_flags) rmqueue_pcplist() argument
3133 rmqueue(struct zone * preferred_zone,struct zone * zone,unsigned int order,gfp_t gfp_flags,unsigned int alloc_flags,int migratetype) rmqueue() argument
3166 reserve_highatomic_pageblock(struct page * page,int order,struct zone * zone) reserve_highatomic_pageblock() argument
3226 int order; unreserve_highatomic_pageblock() local
3298 __zone_watermark_unusable_free(struct zone * z,unsigned int order,unsigned int alloc_flags) __zone_watermark_unusable_free() argument
3324 __zone_watermark_ok(struct zone * z,unsigned int order,unsigned long mark,int highest_zoneidx,unsigned int alloc_flags,long free_pages) __zone_watermark_ok() argument
3402 zone_watermark_ok(struct zone * z,unsigned int order,unsigned long mark,int highest_zoneidx,unsigned int alloc_flags) zone_watermark_ok() argument
3409 zone_watermark_fast(struct zone * z,unsigned int order,unsigned long mark,int highest_zoneidx,unsigned int alloc_flags,gfp_t gfp_mask) zone_watermark_fast() argument
3530 get_page_from_freelist(gfp_t gfp_mask,unsigned int order,int alloc_flags,const struct alloc_context * ac) get_page_from_freelist() argument
3748 __alloc_pages_cpuset_fallback(gfp_t gfp_mask,unsigned int order,unsigned int alloc_flags,const struct alloc_context * ac) __alloc_pages_cpuset_fallback() argument
3767 __alloc_pages_may_oom(gfp_t gfp_mask,unsigned int order,const struct alloc_context * ac,unsigned long * did_some_progress) __alloc_pages_may_oom() argument
3862 __alloc_pages_direct_compact(gfp_t gfp_mask,unsigned int order,unsigned int alloc_flags,const struct alloc_context * ac,enum compact_priority prio,enum compact_result * compact_result) __alloc_pages_direct_compact() argument
3921 should_compact_retry(struct alloc_context * ac,int order,int alloc_flags,enum compact_result compact_result,enum compact_priority * compact_priority,int * compaction_retries) should_compact_retry() argument
3987 __alloc_pages_direct_compact(gfp_t gfp_mask,unsigned int order,unsigned int alloc_flags,const struct alloc_context * ac,enum compact_priority prio,enum compact_result * compact_result) __alloc_pages_direct_compact() argument
3996 should_compact_retry(struct alloc_context * ac,unsigned int order,int alloc_flags,enum compact_result compact_result,enum compact_priority * compact_priority,int * compaction_retries) should_compact_retry() argument
4108 __perform_reclaim(gfp_t gfp_mask,unsigned int order,const struct alloc_context * ac) __perform_reclaim() argument
4134 __alloc_pages_direct_reclaim(gfp_t gfp_mask,unsigned int order,unsigned int alloc_flags,const struct alloc_context * ac,unsigned long * did_some_progress) __alloc_pages_direct_reclaim() argument
4167 wake_all_kswapds(unsigned int order,gfp_t gfp_mask,const struct alloc_context * ac) wake_all_kswapds() argument
4193 gfp_to_alloc_flags(gfp_t gfp_mask,unsigned int order) gfp_to_alloc_flags() argument
4297 should_reclaim_retry(gfp_t gfp_mask,unsigned order,struct alloc_context * ac,int alloc_flags,bool did_some_progress,int * no_progress_loops) should_reclaim_retry() argument
4407 __alloc_pages_slowpath(gfp_t gfp_mask,unsigned int order,struct alloc_context * ac) __alloc_pages_slowpath() argument
4705 prepare_alloc_pages(gfp_t gfp_mask,unsigned int order,int preferred_nid,nodemask_t * nodemask,struct alloc_context * ac,gfp_t * alloc_gfp,unsigned int * alloc_flags) prepare_alloc_pages() argument
4923 __alloc_frozen_pages_noprof(gfp_t gfp,unsigned int order,int preferred_nid,nodemask_t * nodemask) __alloc_frozen_pages_noprof() argument
4988 __alloc_pages_noprof(gfp_t gfp,unsigned int order,int preferred_nid,nodemask_t * nodemask) __alloc_pages_noprof() argument
5000 __folio_alloc_noprof(gfp_t gfp,unsigned int order,int preferred_nid,nodemask_t * nodemask) __folio_alloc_noprof() argument
5014 get_free_pages_noprof(gfp_t gfp_mask,unsigned int order) get_free_pages_noprof() argument
5052 ___free_pages(struct page * page,unsigned int order,fpi_t fpi_flags) ___free_pages() argument
5069 __free_pages(struct page * page,unsigned int order) __free_pages() argument
5079 free_pages_nolock(struct page * page,unsigned int order) free_pages_nolock() argument
5084 free_pages(unsigned long addr,unsigned int order) free_pages() argument
5094 make_alloc_exact(unsigned long addr,unsigned int order,size_t size) make_alloc_exact() argument
5132 unsigned int order = get_order(size); alloc_pages_exact_noprof() local
5157 unsigned int order = get_order(size); alloc_pages_exact_nid_noprof() local
6595 int order; split_free_pages() local
6790 int order = ilog2(end - start); alloc_contig_range_noprof() local
6990 unsigned int order; __offline_isolated_pages() local
7036 unsigned int order; is_free_buddy_page() local
7052 add_to_free_list(struct page * page,struct zone * zone,unsigned int order,int migratetype,bool tail) add_to_free_list() argument
7097 unsigned int order; take_page_off_buddy() local
7184 page_contains_unaccepted(struct page * page,unsigned int order) page_contains_unaccepted() argument
7239 cond_accept_memory(struct zone * zone,unsigned int order,int alloc_flags) cond_accept_memory() argument
7298 page_contains_unaccepted(struct page * page,unsigned int order) page_contains_unaccepted() argument
7303 cond_accept_memory(struct zone * zone,unsigned int order,int alloc_flags) cond_accept_memory() argument
7332 alloc_pages_nolock_noprof(int nid,unsigned int order) alloc_pages_nolock_noprof() argument
[all...]
H A Dcompaction.c46 * order == -1 is expected when compacting proactively via
51 static inline bool is_via_compact_memory(int order) in is_via_compact_memory() argument
53 return order == -1; in is_via_compact_memory()
59 static inline bool is_via_compact_memory(int order) { return false; } in is_via_compact_memory() argument
67 #define block_start_pfn(pfn, order) round_down(pfn, 1UL << (order)) argument
68 #define block_end_pfn(pfn, order) ALIGN((pfn) + 1, 1UL << (order)) argument
71 * Page order with-respect-to which proactive compaction
83 static struct page *mark_allocated_noprof(struct page *page, unsigned int order, gfp_ argument
93 int order; release_free_list() local
159 defer_compaction(struct zone * zone,int order) defer_compaction() argument
174 compaction_deferred(struct zone * zone,int order) compaction_deferred() argument
197 compaction_defer_reset(struct zone * zone,int order,bool alloc_success) compaction_defer_reset() argument
211 compaction_restarting(struct zone * zone,int order) compaction_restarting() argument
601 unsigned int order; isolate_freepages_block() local
632 const unsigned int order = compound_order(page); isolate_freepages_block() local
732 int order; isolate_freepages_range() local
834 skip_isolation_on_order(int order,int target_order) skip_isolation_on_order() argument
984 const unsigned int order = compound_order(page); isolate_migratepages_block() local
1067 const unsigned int order = compound_order(page); isolate_migratepages_block() local
1420 int order = cc->order > 0 ? cc->order : pageblock_order; suitable_migration_target() local
1523 next_search_order(struct compact_control * cc,int order) next_search_order() argument
1549 int order; fast_isolate_freepages() local
1837 int order = folio_order(src); compaction_alloc_noprof() local
1894 int order = folio_order(dst); compaction_free() local
1966 int order; fast_find_migrateblock() local
2272 unsigned int order; __compact_finished() local
2393 __compaction_suitable(struct zone * zone,int order,unsigned long watermark,int highest_zoneidx,unsigned long free_pages) __compaction_suitable() argument
2423 compaction_suitable(struct zone * zone,int order,unsigned long watermark,int highest_zoneidx) compaction_suitable() argument
2468 compaction_zonelist_suitable(struct alloc_context * ac,int order,int alloc_flags) compaction_zonelist_suitable() argument
2506 compaction_suit_allocation_order(struct zone * zone,unsigned int order,int highest_zoneidx,unsigned int alloc_flags,bool async,bool kcompactd) compaction_suit_allocation_order() argument
2556 int order; compact_zone() local
2785 compact_zone_order(struct zone * zone,int order,gfp_t gfp_mask,enum compact_priority prio,unsigned int alloc_flags,int highest_zoneidx,struct page ** capture) compact_zone_order() argument
2850 try_to_compact_pages(gfp_t gfp_mask,unsigned int order,unsigned int alloc_flags,const struct alloc_context * ac,enum compact_priority prio,struct page ** capture) try_to_compact_pages() argument
3171 wakeup_kcompactd(pg_data_t * pgdat,int order,int highest_zoneidx) wakeup_kcompactd() argument
[all...]
/linux/arch/arm/lib/
H A Dlib1funcs.S106 .macro ARM_DIV2_ORDER divisor, order argument
110 clz \order, \divisor
111 rsb \order, \order, #31
117 movhs \order, #16
118 movlo \order, #0
122 addhs \order, \order, #8
126 addhs \order, \order, #4
129 addhi \order, \order, #3
130 addls \order, \order, \divisor, lsr #1
137 .macro ARM_MOD_BODY dividend, divisor, order, spare
[all …]
/linux/mm/kmsan/
H A Dinit.c107 * by their order: when kmsan_memblock_free_pages() is called for the first
108 * time with a certain order, it is reserved as a shadow block, for the second
111 * after which held_back[order] can be used again.
116 bool kmsan_memblock_free_pages(struct page *page, unsigned int order) in kmsan_memblock_free_pages() argument
120 if (!held_back[order].shadow) { in kmsan_memblock_free_pages()
121 held_back[order].shadow = page; in kmsan_memblock_free_pages()
124 if (!held_back[order].origin) { in kmsan_memblock_free_pages()
125 held_back[order].origin = page; in kmsan_memblock_free_pages()
128 shadow = held_back[order].shadow; in kmsan_memblock_free_pages()
129 origin = held_back[order].origin; in kmsan_memblock_free_pages()
[all …]
/linux/kernel/bpf/
H A Dcgroup_iter.c13 * 1. Walk the descendants of a cgroup in pre-order.
14 * 2. Walk the descendants of a cgroup in post-order.
18 * For walking descendants, cgroup_iter can walk in either pre-order or
19 * post-order. For walking ancestors, the iter walks up from a cgroup to
40 * EOPNOTSUPP. In order to work around, the user may have to update their
54 int order; member
77 if (p->order == BPF_CGROUP_ITER_DESCENDANTS_PRE) in cgroup_iter_seq_start()
79 else if (p->order == BPF_CGROUP_ITER_DESCENDANTS_POST) in cgroup_iter_seq_start()
110 if (p->order == BPF_CGROUP_ITER_DESCENDANTS_PRE) in cgroup_iter_seq_next()
112 else if (p->order == BPF_CGROUP_ITER_DESCENDANTS_POST) in cgroup_iter_seq_next()
[all …]
/linux/drivers/media/pci/cx18/
H A Dcx18-mailbox.c231 static void epu_dma_done(struct cx18 *cx, struct cx18_in_work_order *order) in epu_dma_done() argument
240 mb = &order->mb; in epu_dma_done()
247 (order->flags & CX18_F_EWO_MB_STALE_UPON_RECEIPT) ? in epu_dma_done()
253 mdl_ack = order->mdl_ack; in epu_dma_done()
277 if ((order->flags & CX18_F_EWO_MB_STALE_UPON_RECEIPT) && in epu_dma_done()
324 static void epu_debug(struct cx18 *cx, struct cx18_in_work_order *order) in epu_debug() argument
327 char *str = order->str; in epu_debug()
329 CX18_DEBUG_INFO("%x %s\n", order->mb.args[0], str); in epu_debug()
335 static void epu_cmd(struct cx18 *cx, struct cx18_in_work_order *order) in epu_cmd() argument
337 switch (order->rpu) { in epu_cmd()
[all …]
/linux/arch/riscv/kvm/
H A Dtlb.c25 unsigned long order) in kvm_riscv_local_hfence_gvma_vmid_gpa() argument
29 if (PTRS_PER_PTE < (gpsz >> order)) { in kvm_riscv_local_hfence_gvma_vmid_gpa()
36 for (pos = gpa; pos < (gpa + gpsz); pos += BIT(order)) in kvm_riscv_local_hfence_gvma_vmid_gpa()
41 for (pos = gpa; pos < (gpa + gpsz); pos += BIT(order)) in kvm_riscv_local_hfence_gvma_vmid_gpa()
53 unsigned long order) in kvm_riscv_local_hfence_gvma_gpa() argument
57 if (PTRS_PER_PTE < (gpsz >> order)) { in kvm_riscv_local_hfence_gvma_gpa()
64 for (pos = gpa; pos < (gpa + gpsz); pos += BIT(order)) in kvm_riscv_local_hfence_gvma_gpa()
69 for (pos = gpa; pos < (gpa + gpsz); pos += BIT(order)) in kvm_riscv_local_hfence_gvma_gpa()
84 unsigned long order) in kvm_riscv_local_hfence_vvma_asid_gva() argument
88 if (PTRS_PER_PTE < (gvsz >> order)) { in kvm_riscv_local_hfence_vvma_asid_gva()
[all …]
/linux/arch/riscv/mm/
H A Dhugetlbpage.c35 unsigned long order; in huge_pte_alloc() local
68 for_each_napot_order(order) { in huge_pte_alloc()
69 if (napot_cont_size(order) == sz) { in huge_pte_alloc()
70 pte = pte_alloc_huge(mm, pmd, addr & napot_cont_mask(order)); in huge_pte_alloc()
88 unsigned long order; in huge_pte_offset() local
119 for_each_napot_order(order) { in huge_pte_offset()
120 if (napot_cont_size(order) == sz) { in huge_pte_offset()
121 pte = pte_offset_huge(pmd, addr & napot_cont_mask(order)); in huge_pte_offset()
189 unsigned long order; in arch_make_huge_pte() local
191 for_each_napot_order(order) { in arch_make_huge_pte()
[all …]
/linux/drivers/gpu/drm/ttm/tests/
H A Dttm_pool_test.c14 unsigned int order; member
103 .order = 0,
107 .order = 2,
111 .order = MAX_PAGE_ORDER + 1,
115 .order = 0,
120 .order = MAX_PAGE_ORDER + 1,
143 unsigned int expected_num_pages = 1 << params->order; in ttm_pool_alloc_basic()
167 if (params->order <= MAX_PAGE_ORDER) { in ttm_pool_alloc_basic()
172 KUNIT_ASSERT_EQ(test, fst_page->private, params->order); in ttm_pool_alloc_basic()
181 * order 0 blocks in ttm_pool_alloc_basic()
[all …]
/linux/Documentation/userspace-api/media/v4l/
H A Dfield-order.rst3 .. _field-order:
6 Field Order
25 which field of a frame is older, the *temporal order*.
30 even) fields, the *spatial order*: The first line of the top field is
39 creating a natural order.
44 and spatial order of fields. The diagrams below should make this
48 bus in the same order they were captured, so if the top field was
53 order. Some drivers may permit the selection of a different order, to
76 - Applications request this field order when any field format
78 e.g. the requested image size, and return the actual field order.
[all …]
/linux/include/linux/
H A Dcompaction.h61 * Number of free order-0 pages that should be available above given watermark
65 static inline unsigned long compact_gap(unsigned int order) in compact_gap() argument
69 * free scanner may have up to 1 << order pages on its list and then in compact_gap()
70 * try to split an (order - 1) free page. At that point, a gap of in compact_gap()
71 * 1 << order might not be enough, so it's safer to require twice that in compact_gap()
80 return 2UL << order; in compact_gap()
90 extern unsigned int extfrag_for_order(struct zone *zone, unsigned int order);
91 extern int fragmentation_index(struct zone *zone, unsigned int order);
93 unsigned int order, unsigned int alloc_flags,
97 extern bool compaction_suitable(struct zone *zone, int order,
[all …]
/linux/Documentation/trace/postprocess/
H A Dtrace-vmscan-postprocess.pl31 # Per-order events
110 my $regex_direct_begin_default = 'order=([0-9]*) gfp_flags=([A-Z_|]*)';
112 my $regex_kswapd_wake_default = 'nid=([0-9]*) order=([0-9]*)';
114 my $regex_wakeup_kswapd_default = 'nid=([0-9]*) order=([0-9]*) gfp_flags=([A-Z_|]*)';
115 my $regex_lru_isolate_default = 'classzone=([0-9]*) order=([0-9]*) nr_requested=([0-9]*) nr_scanned…
166 # Verify fields are in the right order
187 "order", "gfp_flags");
195 "nid", "order");
203 "nid", "order", "gfp_flags");
207 "classzone", "order",
[all …]
/linux/tools/testing/selftests/tc-testing/tc-tests/actions/
H A Dmpls.json23 "matchPattern": "action order [0-9]+: mpls.*dec_ttl.*pipe.*index 8 ref",
50 "matchPattern": "action order [0-9]+: mpls.*dec_ttl.*pass.*index 8 ref",
77 "matchPattern": "action order [0-9]+: mpls.*dec_ttl.*drop.*index 8 ref",
104 "matchPattern": "action order [0-9]+: mpls.*dec_ttl.*reclassify.*index 8 ref",
131 "matchPattern": "action order [0-9]+: mpls.*dec_ttl.*continue.*index 8 ref",
158 "matchPattern": "action order [0-9]+: mpls.*jump 10.*index 8 ref",
185 "matchPattern": "action order [0-9]+: mpls.*dec_ttl trap.*index 8 ref",
212 "matchPattern": "action order [0-9]+: mpls.*dec_ttl pipe.*index 8 ref.*cookie aabbccddeeff",
239 …"matchPattern": "action order [0-9]+: mpls.*dec_ttl continue.*index 8 ref.*cookie aa11bb22cc33dd44…
266 "matchPattern": "action order [0-9]+: mpls.*dec_ttl.*foo.*index 8 ref",
[all …]
H A Dife.json23 …"matchPattern": "action order [0-9]*: ife encode action pass.*type 0[xX]ED3E.*allow mark.*index 2",
50 … "matchPattern": "action order [0-9]*: ife encode action pipe.*type 0[xX]ED3E.*use mark.*index 2",
77 …"matchPattern": "action order [0-9]*: ife encode action continue.*type 0[xX]ED3E.*allow mark.*inde…
104 …"matchPattern": "action order [0-9]*: ife encode action drop.*type 0[xX]ED3E.*use mark 789.*index …
131 …"matchPattern": "action order [0-9]*: ife encode action reclassify.*type 0[xX]ED3E.*use mark 65676…
158 …"matchPattern": "action order [0-9]*: ife encode action jump 1.*type 0[xX]ED3E.*use mark 65.*index…
185 …"matchPattern": "action order [0-9]*: ife encode action reclassify.*type 0[xX]ED3E.*use mark 42949…
212 …"matchPattern": "action order [0-9]*: ife encode action pipe.*type 0[xX]ED3E.*use mark 42949672959…
237 …"matchPattern": "action order [0-9]*: ife encode action pass.*type 0[xX]ED3E.*allow prio.*index 9",
264 …"matchPattern": "action order [0-9]*: ife encode action pipe.*type 0[xX]ED3E.*use prio 7.*index 9",
[all …]
H A Dsample.json23 "matchPattern": "action order [0-9]+: sample rate 1/10 group 1.*index 2 ref",
50 "matchPattern": "action order [0-9]+: sample rate 1/700 group 2 continue.*index 2 ref",
77 "matchPattern": "action order [0-9]+: sample rate 1/10000 group 11 drop.*index 22 ref",
104 … "matchPattern": "action order [0-9]+: sample rate 1/20000 group 72 reclassify.*index 100 ref",
131 "matchPattern": "action order [0-9]+: sample rate 1/20 group 2 pipe.*index 100 ref",
158 "matchPattern": "action order [0-9]+: sample rate 1/700 group 25 jump 4.*index 200 ref",
185 "matchPattern": "action order [0-9]+: sample rate 1/0 group 1.*index 2 ref",
212 "matchPattern": "action order [0-9]+: sample rate 1/200000 group 52 foo.*index 1 ref",
237 "matchPattern": "action order [0-9]+: sample.*index 1 ref",
262 "matchPattern": "action order [0-9]+: sample.*group 10.*index 1 ref",
[all …]
/linux/drivers/gpu/drm/nouveau/nvkm/subdev/therm/
H A Dgk104.c34 const struct gk104_clkgate_engine_info *order = therm->clkgate_order; in gk104_clkgate_enable() local
38 for (i = 0; order[i].type != NVKM_SUBDEV_NR; i++) { in gk104_clkgate_enable()
39 if (!nvkm_device_subdev(dev, order[i].type, order[i].inst)) in gk104_clkgate_enable()
42 nvkm_mask(dev, 0x20200 + order[i].offset, 0xff00, 0x4500); in gk104_clkgate_enable()
50 for (i = 0; order[i].type != NVKM_SUBDEV_NR; i++) { in gk104_clkgate_enable()
51 if (!nvkm_device_subdev(dev, order[i].type, order[i].inst)) in gk104_clkgate_enable()
54 nvkm_mask(dev, 0x20200 + order[i].offset, 0x00ff, 0x0045); in gk104_clkgate_enable()
63 const struct gk104_clkgate_engine_info *order = therm->clkgate_order; in gk104_clkgate_fini() local
67 for (i = 0; order[i].type != NVKM_SUBDEV_NR; i++) { in gk104_clkgate_fini()
68 if (!nvkm_device_subdev(dev, order[i].type, order[i].inst)) in gk104_clkgate_fini()
[all …]
/linux/drivers/gpu/drm/i915/selftests/
H A Di915_syncmap.c274 unsigned int pass, order; in igt_syncmap_join_above() local
296 for (order = 0; order < 64; order += SHIFT) { in igt_syncmap_join_above()
297 u64 context = BIT_ULL(order); in igt_syncmap_join_above()
335 unsigned int step, order, idx; in igt_syncmap_join_below() local
345 for (order = 64 - SHIFT; order > 0; order -= SHIFT) { in igt_syncmap_join_below()
346 u64 context = step * BIT_ULL(order); in igt_syncmap_join_below()
353 … pr_err("Inserting context=%llx (order=%d, step=%d) did not return leaf (height=%d, prefix=%llx\n", in igt_syncmap_join_below()
354 context, order, step, sync->height, sync->prefix); in igt_syncmap_join_below()
362 for (order = SHIFT; order < 64; order += SHIFT) { in igt_syncmap_join_below()
363 u64 context = step * BIT_ULL(order); in igt_syncmap_join_below()
[all …]
H A Di915_random.c70 void i915_random_reorder(unsigned int *order, unsigned int count, in i915_random_reorder() argument
73 i915_prandom_shuffle(order, sizeof(*order), count, state); in i915_random_reorder()
78 unsigned int *order, i; in i915_random_order() local
80 order = kmalloc_array(count, sizeof(*order), in i915_random_order()
82 if (!order) in i915_random_order()
83 return order; in i915_random_order()
86 order[i] = i; in i915_random_order()
88 i915_random_reorder(order, count, state); in i915_random_order()
89 return order; in i915_random_order()

12345678910>>...173