| /linux/drivers/edac/ |
| H A D | aspeed_edac.c | 119 csrow->first_page + page, offset, syndrome, in count_rec() 140 csrow->first_page + page, offset, syndrome, in count_un_rec() 260 csrow->first_page = r.start >> PAGE_SHIFT; in init_csrows() 262 csrow->last_page = csrow->first_page + nr_pages - 1; in init_csrows() 273 csrow->first_page, nr_pages); in init_csrows()
|
| H A D | amd76x_edac.c | 148 mci->csrows[row]->first_page, 0, 0, in amd76x_process_error_info() 163 mci->csrows[row]->first_page, 0, 0, in amd76x_process_error_info() 208 csrow->first_page = mba_base >> PAGE_SHIFT; in amd76x_init_csrows() 210 csrow->last_page = csrow->first_page + dimm->nr_pages - 1; in amd76x_init_csrows()
|
| H A D | pasemi_edac.c | 102 mci->csrows[cs]->first_page, 0, 0, in pasemi_edac_process_error_info() 109 mci->csrows[cs]->first_page, 0, 0, in pasemi_edac_process_error_info() 167 csrow->first_page = last_page_in_mmc; in pasemi_edac_init_csrows() 168 csrow->last_page = csrow->first_page + dimm->nr_pages - 1; in pasemi_edac_init_csrows()
|
| H A D | cpc925_edac.c | 340 csrow->first_page = last_nr_pages; in cpc925_init_csrows() 342 csrow->last_page = csrow->first_page + nr_pages - 1; in cpc925_init_csrows() 452 if (mci->csrows[rank]->first_page == 0) { in cpc925_mc_get_pfn() 460 pa = mci->csrows[rank]->first_page << PAGE_SHIFT; in cpc925_mc_get_pfn()
|
| H A D | edac_mc.c | 112 edac_dbg(4, " csrow->first_page = 0x%lx\n", csrow->first_page); in edac_mc_dump_csrow() 773 csrow->first_page, page, csrow->last_page, in edac_mc_find_csrow_by_page() 776 if ((page >= csrow->first_page) && in edac_mc_find_csrow_by_page() 779 (csrow->first_page & csrow->page_mask))) { in edac_mc_find_csrow_by_page()
|
| H A D | i82860_edac.c | 173 csrow->first_page = last_cumul_size; in i82860_init_csrows()
|
| H A D | i3000_edac.c | 402 csrow->first_page = last_cumul_size; in i3000_probe1()
|
| H A D | i82875p_edac.c | 370 csrow->first_page = last_cumul_size; in i82875p_init_csrows()
|
| H A D | e7xxx_edac.c | 388 csrow->first_page = last_cumul_size; in e7xxx_init_csrows()
|
| H A D | i82975x_edac.c | 423 csrow->first_page = last_cumul_size; in i82975x_init_csrows()
|
| /linux/arch/sparc/kernel/ |
| H A D | iommu.c | 201 unsigned long order, first_page; in dma_4u_alloc_coherent() local 218 first_page = (unsigned long) page_address(page); in dma_4u_alloc_coherent() 219 memset((char *)first_page, 0, PAGE_SIZE << order); in dma_4u_alloc_coherent() 226 free_pages(first_page, order); in dma_4u_alloc_coherent() 232 ret = (void *) first_page; in dma_4u_alloc_coherent() 234 first_page = __pa(first_page); in dma_4u_alloc_coherent() 238 (first_page & IOPTE_PAGE)); in dma_4u_alloc_coherent() 240 first_page += IO_PAGE_SIZE; in dma_4u_alloc_coherent()
|
| H A D | pci_sun4v.c | 186 unsigned long flags, order, first_page, npages, n; in dma_4v_alloc_coherent() local 210 first_page = (unsigned long) page_address(page); in dma_4v_alloc_coherent() 211 memset((char *)first_page, 0, PAGE_SIZE << order); in dma_4v_alloc_coherent() 227 ret = (void *) first_page; in dma_4v_alloc_coherent() 228 first_page = __pa(first_page); in dma_4v_alloc_coherent() 238 long err = iommu_batch_add(first_page + (n * PAGE_SIZE), mask); in dma_4v_alloc_coherent() 255 free_pages(first_page, order); in dma_4v_alloc_coherent()
|
| /linux/drivers/platform/goldfish/ |
| H A D | goldfish_pipe.c | 256 static int goldfish_pin_pages(unsigned long first_page, in goldfish_pin_pages() argument 264 int requested_pages = ((last_page - first_page) >> PAGE_SHIFT) + 1; in goldfish_pin_pages() 273 ret = pin_user_pages_fast(first_page, requested_pages, in goldfish_pin_pages() 289 unsigned long first_page, in populate_rw_params() argument 303 int size_on_page = first_page == last_page in populate_rw_params() 333 unsigned long first_page = address & PAGE_MASK; in transfer_max_buffers() local 341 pages_count = goldfish_pin_pages(first_page, last_page, in transfer_max_buffers() 350 first_page, last_page, iter_last_page_size, is_write, in transfer_max_buffers()
|
| /linux/drivers/staging/media/meson/vdec/ |
| H A D | codec_hevc_common.c | 286 u32 first_page; in codec_hevc_fill_mmu_map() local 290 first_page = comm->fbc_buffer_paddr[vb->index] >> PAGE_SHIFT; in codec_hevc_fill_mmu_map() 292 first_page = vb2_dma_contig_plane_dma_addr(vb, 0) >> PAGE_SHIFT; in codec_hevc_fill_mmu_map() 295 mmu_map[i] = first_page + i; in codec_hevc_fill_mmu_map()
|
| /linux/mm/ |
| H A D | memremap.c | 89 struct page *first_page; in pageunmap_range() local 92 first_page = pfn_to_page(pfn_first(pgmap, range_id)); in pageunmap_range() 96 remove_pfn_range_from_zone(page_zone(first_page), PHYS_PFN(range->start), in pageunmap_range()
|
| H A D | mprotect.c | 147 struct page *first_page, bool expected_anon_exclusive) in page_anon_exclusive_sub_batch() argument 152 if (expected_anon_exclusive != PageAnonExclusive(first_page + idx)) in page_anon_exclusive_sub_batch() 173 struct folio *folio, struct page *first_page, unsigned long addr, pte_t *ptep, in commit_anon_folio_batch() argument 181 expected_anon_exclusive = PageAnonExclusive(first_page + sub_batch_idx); in commit_anon_folio_batch() 183 first_page, expected_anon_exclusive); in commit_anon_folio_batch()
|
| /linux/fs/ntfs3/ |
| H A D | fslog.c | 1033 u32 first_page; member 1092 return off >= log->l_size ? log->first_page : off; in next_page_off() 1324 log->first_page = major_ver >= 2 ? 0x22 * log->page_size : in log_init_pg_hdr() 1341 log->next_page = log->first_page; in log_create() 1380 log->total_avail_pages = log->l_size - log->first_page; in log_create() 1496 if (hdr_off == log->first_page) in next_log_lsn() 1537 log->first_page : in current_log_avail() 1547 log->next_page == log->first_page ? log->l_size : in current_log_avail() 1588 (lsn_seq == seq - 1 && log->first_page == vbo && in check_subseq_log_page() 1640 second_off = log->first_page - log->page_size; in last_log_lsn() [all …]
|
| /linux/drivers/mtd/nand/raw/ |
| H A D | nand_base.c | 254 int first_page, page_offset; in nand_block_bad() local 258 first_page = (int)(ofs >> chip->page_shift) & chip->pagemask; in nand_block_bad() 262 res = chip->ecc.read_oob(chip, first_page + page_offset); in nand_block_bad() 1231 first_lun = chip->cont_read.first_page / ppl; in rawnand_cap_cont_reads() 1240 if (chip->cont_read.first_page == chip->cont_read.pause_page) { in rawnand_cap_cont_reads() 1241 chip->cont_read.first_page++; in rawnand_cap_cont_reads() 1246 if (chip->cont_read.first_page >= chip->cont_read.last_page) in rawnand_cap_cont_reads() 1304 if (page == chip->cont_read.first_page) in nand_lp_exec_cont_read_page_op() 1317 chip->cont_read.first_page++; in nand_lp_exec_cont_read_page_op() 1326 return chip->cont_read.ongoing && page >= chip->cont_read.first_page; in rawnand_cont_read_ongoing() [all …]
|
| /linux/sound/pci/trident/ |
| H A D | trident.h | 248 short first_page, last_page; member
|
| /linux/drivers/gpu/drm/ttm/ |
| H A D | ttm_pool.c | 525 struct page **first_page = alloc->pages; in ttm_pool_restore_commit() local 530 p = first_page[i]; in ttm_pool_restore_commit() 565 first_page[i] = ttm_backup_handle_to_page_ptr(0); in ttm_pool_restore_commit()
|
| /linux/tools/testing/selftests/kvm/include/ |
| H A D | kvm_util.h | 487 uint64_t first_page, uint32_t num_pages) in kvm_vm_register_coalesced_io() 492 .first_page = first_page, in kvm_vm_register_coalesced_io() 464 kvm_vm_clear_dirty_log(struct kvm_vm * vm,int slot,void * log,uint64_t first_page,uint32_t num_pages) kvm_vm_clear_dirty_log() argument
|
| /linux/drivers/block/drbd/ |
| H A D | drbd_bitmap.c | 1514 int first_page; in _drbd_bm_set_bits() local 1536 first_page = sl >> (3 + PAGE_SHIFT); in _drbd_bm_set_bits() 1545 for (page_nr = first_page; page_nr < last_page; page_nr++) { in _drbd_bm_set_bits()
|
| /linux/include/linux/ |
| H A D | edac.h | 434 unsigned long first_page; /* first page number in csrow */ member
|
| /linux/drivers/net/ethernet/google/gve/ |
| H A D | gve_tx.c | 554 u64 first_page = iov_offset / PAGE_SIZE; in gve_dma_sync_for_device() local 557 for (page = first_page; page <= last_page; page++) in gve_dma_sync_for_device()
|
| /linux/virt/kvm/ |
| H A D | kvm_main.c | 2375 if (log->first_page & 63) in kvm_clear_dirty_log_protect() 2387 if (log->first_page > memslot->npages || in kvm_clear_dirty_log_protect() 2388 log->num_pages > memslot->npages - log->first_page || in kvm_clear_dirty_log_protect() 2389 (log->num_pages < memslot->npages - log->first_page && (log->num_pages & 63))) in kvm_clear_dirty_log_protect() 2400 for (offset = log->first_page, i = offset / BITS_PER_LONG, in kvm_clear_dirty_log_protect() 5421 __u64 first_page; member 5458 log.first_page = compat_log.first_page; in kvm_vm_compat_ioctl()
|