| /linux/drivers/net/ethernet/huawei/hinic3/ |
| H A D | hinic3_queue_common.c | 44 u32 pg_idx; in hinic3_queue_pages_alloc() local 54 for (pg_idx = 0; pg_idx < qpages->num_pages; pg_idx++) { in hinic3_queue_pages_alloc() 59 qpages->pages + pg_idx); in hinic3_queue_pages_alloc() 61 __queue_pages_free(hwdev, qpages, pg_idx); in hinic3_queue_pages_alloc()
|
| /linux/tools/testing/selftests/bpf/progs/ |
| H A D | verifier_arena_large.c | 194 int i, pg_idx; in alloc_pages() 201 pg_idx = (unsigned long) (pg - base) / PAGE_SIZE; in alloc_pages() 204 if (pg_idx != i) 208 if (pg_idx >= max_idx || (pg_idx & 1)) in big_alloc2() 211 *pg = pg_idx; in big_alloc2() 212 page[pg_idx] = pg; in big_alloc2() 181 int i, pg_idx; alloc_pages() local
|
| /linux/drivers/hwtracing/coresight/ |
| H A D | ultrasoc-smb.c | 338 long pg_idx, pg_offset; in smb_sync_perf_buffer() local 340 pg_idx = head >> PAGE_SHIFT; in smb_sync_perf_buffer() 352 memcpy(dst_pages[pg_idx] + pg_offset, in smb_sync_perf_buffer() 358 pg_idx++; in smb_sync_perf_buffer() 359 pg_idx %= buf->nr_pages; in smb_sync_perf_buffer()
|
| H A D | coresight-catu.c | 118 unsigned int table_nr, pg_idx, pg_offset; in catu_get_table() local 131 pg_idx = table_nr / CATU_PAGES_PER_SYSPAGE; in catu_get_table() 134 *daddrp = table_pages->daddrs[pg_idx] + pg_offset; in catu_get_table() 135 ptr = page_address(table_pages->pages[pg_idx]); in catu_get_table()
|
| H A D | coresight-tmc-etr.c | 399 int pg_idx = offset >> PAGE_SHIFT; in tmc_sg_table_get_data() local 412 *bufpp = page_address(data_pages->pages[pg_idx]) + pg_offset; in tmc_sg_table_get_data() 1585 long pg_idx, pg_offset; in tmc_etr_sync_perf_buffer() local 1590 pg_idx = head >> PAGE_SHIFT; in tmc_etr_sync_perf_buffer() 1611 memcpy(dst_pages[pg_idx] + pg_offset, src_buf, bytes); in tmc_etr_sync_perf_buffer() 1619 if (++pg_idx == etr_perf->nr_pages) in tmc_etr_sync_perf_buffer() 1620 pg_idx = 0; in tmc_etr_sync_perf_buffer()
|
| /linux/drivers/pci/controller/mobiveil/ |
| H A D | pcie-mobiveil.c | 28 static void mobiveil_pcie_sel_page(struct mobiveil_pcie *pcie, u8 pg_idx) in mobiveil_pcie_sel_page() argument 34 val |= (pg_idx & PAGE_SEL_MASK) << PAGE_SEL_SHIFT; in mobiveil_pcie_sel_page()
|
| /linux/drivers/net/ethernet/chelsio/libcxgb/ |
| H A D | libcxgb_ppm.h | 113 unsigned char pg_idx:2; member
|
| /linux/drivers/scsi/cxgbi/cxgb3i/ |
| H A D | cxgb3i.c | 1154 unsigned int tid, int pg_idx) in ddp_setup_conn_pgidx() argument 1159 u64 val = pg_idx < DDP_PGIDX_MAX ? pg_idx : 0; in ddp_setup_conn_pgidx() 1162 "csk 0x%p, tid %u, pg_idx %d.\n", csk, tid, pg_idx); in ddp_setup_conn_pgidx()
|
| /linux/drivers/scsi/cxgbi/cxgb4i/ |
| H A D | cxgb4i.c | 2078 int pg_idx) in ddp_setup_conn_pgidx() argument 2083 if (!pg_idx || pg_idx >= DDP_PGIDX_MAX) in ddp_setup_conn_pgidx() 2097 req->val = cpu_to_be64(pg_idx << 8); in ddp_setup_conn_pgidx() 2101 "csk 0x%p, tid 0x%x, pg_idx %u.\n", csk, csk->tid, pg_idx); in ddp_setup_conn_pgidx()
|
| /linux/drivers/infiniband/hw/irdma/ |
| H A D | verbs.c | 2819 u32 pg_idx; in irdma_check_mem_contiguous() local 2821 for (pg_idx = 0; pg_idx < npages; pg_idx++) { in irdma_check_mem_contiguous() 2822 if ((*arr + (pg_size * pg_idx)) != arr[pg_idx]) in irdma_check_mem_contiguous()
|
| /linux/drivers/target/iscsi/cxgbit/ |
| H A D | cxgbit_cm.c | 1545 int cxgbit_setup_conn_pgidx(struct cxgbit_sock *csk, u32 pg_idx) in cxgbit_setup_conn_pgidx() argument 1563 req->val = cpu_to_be64(pg_idx << 8); in cxgbit_setup_conn_pgidx()
|
| /linux/net/core/ |
| H A D | skbuff.c | 4434 unsigned int pg_idx, pg_off, pg_sz; in skb_seq_read() local 4438 pg_idx = 0; in skb_seq_read() 4443 pg_idx = (pg_off + st->frag_off) >> PAGE_SHIFT; in skb_seq_read() 4452 st->frag_data = kmap_atomic(skb_frag_page(frag) + pg_idx); in skb_seq_read()
|
| /linux/drivers/net/ethernet/meta/fbnic/ |
| H A D | fbnic_txrx.c | 1001 unsigned int pg_idx = FIELD_GET(FBNIC_RCD_AL_BUFF_PAGE_MASK, rcd); in fbnic_add_rx_frag() local 1004 netmem_ref netmem = fbnic_page_pool_get_data(qt, pg_idx); in fbnic_add_rx_frag()
|