Lines Matching defs:page_start

48  * @page_start: page index of the first page to be freed
51 * Free pages [@page_start and @page_end) in @pages for all units.
55 struct page **pages, int page_start, int page_end)
61 for (i = page_start; i < page_end; i++) {
74 * @page_start: page index of the first page to be allocated
78 * Allocate pages [@page_start,@page_end) into @pages for all units.
83 struct page **pages, int page_start, int page_end,
92 for (i = page_start; i < page_end; i++) {
103 while (--i >= page_start)
109 for (i = page_start; i < page_end; i++)
118 * @page_start: page index of the first page to be flushed
121 * Pages in [@page_start,@page_end) of @chunk are about to be
128 int page_start, int page_end)
131 pcpu_chunk_addr(chunk, pcpu_low_unit_cpu, page_start),
144 * @page_start: page index of the first page to unmap
147 * For each cpu, unmap pages [@page_start,@page_end) out of @chunk.
154 struct page **pages, int page_start, int page_end)
160 for (i = page_start; i < page_end; i++) {
167 __pcpu_unmap_pages(pcpu_chunk_addr(chunk, cpu, page_start),
168 page_end - page_start);
175 * @page_start: page index of the first page to be flushed
178 * Pages [@page_start,@page_end) of @chunk have been unmapped. Flush
186 int page_start, int page_end)
189 pcpu_chunk_addr(chunk, pcpu_low_unit_cpu, page_start),
204 * @page_start: page index of the first page to map
207 * For each cpu, map pages [@page_start,@page_end) into @chunk. The
215 struct page **pages, int page_start, int page_end)
221 err = __pcpu_map_pages(pcpu_chunk_addr(chunk, cpu, page_start),
222 &pages[pcpu_page_idx(cpu, page_start)],
223 page_end - page_start);
227 for (i = page_start; i < page_end; i++)
234 __pcpu_unmap_pages(pcpu_chunk_addr(chunk, tcpu, page_start),
235 page_end - page_start);
239 pcpu_post_unmap_tlb_flush(chunk, page_start, page_end);
246 * @page_start: page index of the first page to be flushed
249 * Pages [@page_start,@page_end) of @chunk have been mapped. Flush
256 int page_start, int page_end)
259 pcpu_chunk_addr(chunk, pcpu_low_unit_cpu, page_start),
266 * @page_start: the start page
270 * For each cpu, populate and map pages [@page_start,@page_end) into
277 int page_start, int page_end, gfp_t gfp)
285 if (pcpu_alloc_pages(chunk, pages, page_start, page_end, gfp))
288 if (pcpu_map_pages(chunk, pages, page_start, page_end)) {
289 pcpu_free_pages(chunk, pages, page_start, page_end);
292 pcpu_post_map_flush(chunk, page_start, page_end);
300 * @page_start: the start page
303 * For each cpu, depopulate and unmap pages [@page_start,@page_end)
313 int page_start, int page_end)
326 pcpu_pre_unmap_flush(chunk, page_start, page_end);
328 pcpu_unmap_pages(chunk, pages, page_start, page_end);
330 pcpu_free_pages(chunk, pages, page_start, page_end);