| /linux/kernel/bpf/ |
| H A D | arena.c | 49 static void arena_free_pages(struct bpf_arena *arena, long uaddr, long page_cnt, bool sleepable); 73 u32 page_cnt; member 549 static long arena_alloc_pages(struct bpf_arena *arena, long uaddr, long page_cnt, int node_id, in arena_alloc_pages() argument 569 if (page_cnt > page_cnt_max) in arena_alloc_pages() 576 if (pgoff > page_cnt_max - page_cnt) in arena_alloc_pages() 583 alloc_pages = min(page_cnt, KMALLOC_MAX_CACHE_SIZE / sizeof(struct page *)); in arena_alloc_pages() 595 ret = is_range_tree_set(&arena->rt, pgoff, page_cnt); in arena_alloc_pages() 598 ret = range_tree_clear(&arena->rt, pgoff, page_cnt); in arena_alloc_pages() 600 ret = pgoff = range_tree_find(&arena->rt, page_cnt); in arena_alloc_pages() 602 ret = range_tree_clear(&arena->rt, pgoff, page_cnt); in arena_alloc_pages() [all …]
|
| /linux/drivers/net/ethernet/qlogic/qed/ |
| H A D | qed_chain.c | 12 u32 page_cnt) in qed_chain_init() argument 34 chain->page_cnt = page_cnt; in qed_chain_init() 35 chain->capacity = chain->usable_per_page * page_cnt; in qed_chain_init() 36 chain->size = chain->elem_per_page * page_cnt; in qed_chain_init() 80 for (i = 0; i < chain->page_cnt; i++) { in qed_chain_free_next_ptr() 114 for (i = 0; i < chain->page_cnt; i++) { in qed_chain_free_pbl() 160 u32 page_cnt) in qed_chain_alloc_sanity_check() argument 165 chain_size *= page_cnt; in qed_chain_alloc_sanity_check() 206 for (i = 0; i < chain->page_cnt; i++) { in qed_chain_alloc_next_ptr() 255 u32 page_cnt, i; in qed_chain_alloc_pbl() local [all …]
|
| H A D | qed_sp_commands.c | 311 u8 page_cnt, i; in qed_sp_pf_start() local 372 page_cnt = (u8)qed_chain_get_page_cnt(&p_hwfn->p_eq->chain); in qed_sp_pf_start() 373 p_ramrod->event_ring_num_pages = page_cnt; in qed_sp_pf_start() 378 page_cnt = (u8)qed_chain_get_page_cnt(&p_hwfn->p_consq->chain); in qed_sp_pf_start() 379 p_ramrod->consolid_q_num_pages = page_cnt; in qed_sp_pf_start()
|
| /linux/include/linux/qed/ |
| H A D | qed_chain.h | 106 u32 page_cnt; member 262 return chain->page_cnt; in qed_chain_get_page_cnt() 303 if (++(*(u16 *)page_to_inc) == p_chain->page_cnt) in qed_chain_advance_page() 307 if (++(*(u32 *)page_to_inc) == p_chain->page_cnt) in qed_chain_advance_page() 498 u32 reset_val = p_chain->page_cnt - 1; in qed_chain_reset() 556 last_page_idx = p_chain->page_cnt - 1; in qed_chain_get_last_elem() 580 u32 cur_prod, page_mask, page_cnt, page_diff; in qed_chain_set_prod() local 599 page_cnt = qed_chain_get_page_cnt(p_chain); in qed_chain_set_prod() 603 page_diff + page_cnt) % page_cnt; in qed_chain_set_prod() 607 page_diff + page_cnt) % page_cnt; in qed_chain_set_prod() [all …]
|
| /linux/tools/testing/selftests/bpf/ |
| H A D | bpf_arena_common.h | 47 void __arena* bpf_arena_alloc_pages(void *map, void __arena *addr, __u32 page_cnt, 49 int bpf_arena_reserve_pages(void *map, void __arena *addr, __u32 page_cnt) __ksym __weak; 50 void bpf_arena_free_pages(void *map, void __arena *ptr, __u32 page_cnt) __ksym __weak; 66 static inline void __arena* bpf_arena_alloc_pages(void *map, void *addr, __u32 page_cnt, in bpf_arena_alloc_pages() argument 71 static inline void bpf_arena_free_pages(void *map, void __arena *ptr, __u32 page_cnt) in bpf_arena_free_pages() argument
|
| /linux/arch/mips/cavium-octeon/executive/ |
| H A D | cvmx-helper-util.c | 133 union cvmx_ipd_portx_bp_page_cnt page_cnt; in cvmx_helper_setup_red() local 141 page_cnt.u64 = 0; in cvmx_helper_setup_red() 142 page_cnt.s.bp_enb = 0; in cvmx_helper_setup_red() 143 page_cnt.s.page_cnt = 100; in cvmx_helper_setup_red() 148 page_cnt.u64); in cvmx_helper_setup_red()
|
| /linux/drivers/base/firmware_loader/ |
| H A D | sysfs.c | 252 int page_cnt = min_t(size_t, PAGE_SIZE - page_ofs, count); in firmware_rw() local 256 page_ofs, page_cnt); in firmware_rw() 259 buffer, page_cnt); in firmware_rw() 261 buffer += page_cnt; in firmware_rw() 262 offset += page_cnt; in firmware_rw() 263 count -= page_cnt; in firmware_rw()
|
| /linux/drivers/infiniband/hw/hns/ |
| H A D | hns_roce_mr.c | 620 int page_cnt = 0; in cal_mtr_pg_cnt() local 625 page_cnt += region->count; in cal_mtr_pg_cnt() 628 return page_cnt; in cal_mtr_pg_cnt() 692 dma_addr_t *pages, unsigned int page_cnt) in hns_roce_mtr_map() argument 709 mapped_cnt < page_cnt; i++) { in hns_roce_mtr_map() 712 if (r->offset + r->count > page_cnt) { in hns_roce_mtr_map() 716 i, r->offset, r->count, page_cnt); in hns_roce_mtr_map() 721 page_cnt - mapped_cnt); in hns_roce_mtr_map() 732 if (mapped_cnt < page_cnt) { in hns_roce_mtr_map() 735 mapped_cnt, page_cnt); in hns_roce_mtr_map() [all …]
|
| /linux/tools/sched_ext/include/scx/ |
| H A D | bpf_arena_common.h | 26 static inline void __arena* bpf_arena_alloc_pages(void *map, void *addr, __u32 page_cnt, in bpf_arena_alloc_pages() argument 31 static inline void bpf_arena_free_pages(void *map, void __arena *ptr, __u32 page_cnt) in bpf_arena_free_pages() argument
|
| H A D | bpf_arena_common.bpf.h | 83 void __arena* bpf_arena_alloc_pages(void *map, void __arena *addr, __u32 page_cnt, 85 void bpf_arena_free_pages(void *map, void __arena *ptr, __u32 page_cnt) __ksym __weak; 86 int bpf_arena_reserve_pages(void *map, void __arena *ptr, __u32 page_cnt) __ksym __weak;
|
| /linux/drivers/hv/ |
| H A D | ring_buffer.c | 187 struct page *pages, u32 page_cnt, u32 max_pkt_size, in hv_ringbuffer_init() argument 199 pages_wraparound = kzalloc_objs(struct page *, page_cnt * 2 - 1); in hv_ringbuffer_init() 204 for (i = 0; i < 2 * (page_cnt - 1); i++) in hv_ringbuffer_init() 206 &pages[i % (page_cnt - 1) + 1]; in hv_ringbuffer_init() 209 vmap(pages_wraparound, page_cnt * 2 - 1, VM_MAP, in hv_ringbuffer_init() 228 ring_info->ring_size = page_cnt << PAGE_SHIFT; in hv_ringbuffer_init()
|
| /linux/arch/mips/include/asm/octeon/ |
| H A D | cvmx-ipd-defs.h | 324 uint64_t page_cnt:17; member 326 uint64_t page_cnt:17; 999 uint64_t page_cnt:17; member 1001 uint64_t page_cnt:17; 1014 uint64_t page_cnt:17; member 1016 uint64_t page_cnt:17; 1029 uint64_t page_cnt:17; member 1031 uint64_t page_cnt:17; 1387 uint64_t page_cnt:25; member 1389 uint64_t page_cnt:25;
|
| /linux/tools/testing/selftests/bpf/progs/ |
| H A D | verifier_arena_large.c | 190 __noinline int alloc_pages(int page_cnt, int pages_atonce, bool first_pass, in alloc_pages() 196 for (i = 0; i < page_cnt; i++) { in alloc_pages() 177 alloc_pages(int page_cnt,int pages_atonce,bool first_pass,int max_idx,int step) alloc_pages() argument
|
| /linux/arch/x86/kernel/cpu/sgx/ |
| H A D | ioctl.c | 29 if (!(encl->page_cnt % SGX_VA_SLOT_COUNT)) { in sgx_encl_grow() 41 WARN_ON_ONCE(encl->page_cnt % SGX_VA_SLOT_COUNT); in sgx_encl_grow() 43 encl->page_cnt++; in sgx_encl_grow() 49 encl->page_cnt--; in sgx_encl_shrink()
|
| /linux/drivers/infiniband/hw/efa/ |
| H A D | efa_verbs.c | 1285 static struct scatterlist *efa_vmalloc_buf_to_sg(u64 *buf, int page_cnt) in efa_vmalloc_buf_to_sg() argument 1291 sglist = kmalloc_objs(*sglist, page_cnt); in efa_vmalloc_buf_to_sg() 1294 sg_init_table(sglist, page_cnt); in efa_vmalloc_buf_to_sg() 1295 for (i = 0; i < page_cnt; i++) { in efa_vmalloc_buf_to_sg() 1316 int page_cnt = pbl->phys.indirect.pbl_buf_size_in_pages; in pbl_chunk_list_create() local 1327 chunk_list_size = DIV_ROUND_UP(page_cnt, EFA_PTRS_PER_CHUNK); in pbl_chunk_list_create() 1336 page_cnt); in pbl_chunk_list_create() 1347 ((page_cnt % EFA_PTRS_PER_CHUNK) * EFA_CHUNK_PAYLOAD_PTR_SIZE) + in pbl_chunk_list_create()
|
| /linux/drivers/infiniband/hw/qedr/ |
| H A D | main.c | 785 u32 page_cnt; in qedr_init_hw() local 799 page_cnt = qed_chain_get_page_cnt(&dev->cnq_array[i].pbl); in qedr_init_hw() 800 cur_pbl->num_pbl_pages = page_cnt; in qedr_init_hw()
|
| /linux/drivers/net/wireless/realtek/rtw88/ |
| H A D | fw.c | 2042 u8 page_cnt, pages; in _rtw_hw_scan_update_probe_req() local 2046 page_cnt = RTW_OLD_PROBE_PG_CNT; in _rtw_hw_scan_update_probe_req() 2048 page_cnt = RTW_PROBE_PG_CNT; in _rtw_hw_scan_update_probe_req() 2050 pages = page_offset + num_probes * page_cnt; in _rtw_hw_scan_update_probe_req() 2060 if (skb->len > page_size * page_cnt) { in _rtw_hw_scan_update_probe_req() 2070 buf_offset += page_cnt * page_size; in _rtw_hw_scan_update_probe_req() 2071 page_offset += page_cnt; in _rtw_hw_scan_update_probe_req()
|
| /linux/drivers/infiniband/hw/irdma/ |
| H A D | verbs.h | 117 u32 page_cnt; member
|
| H A D | verbs.c | 2882 status = irdma_get_pble(rf->pble_rsrc, palloc, iwmr->page_cnt, in irdma_setup_pbles() 3176 iwmr->page_cnt = max_num_sg; in irdma_alloc_mr() 3179 err_code = irdma_get_pble(iwdev->rf->pble_rsrc, palloc, iwmr->page_cnt, in irdma_alloc_mr() 3213 if (unlikely(iwmr->npages == iwmr->page_cnt)) in irdma_set_page() 3321 lvl = iwmr->page_cnt != 1 ? PBLE_LEVEL_1 | PBLE_LEVEL_2 : PBLE_LEVEL_0; in irdma_reg_user_mr_type_mem() 3397 iwmr->page_cnt = ib_umem_num_dma_blocks(region, iwmr->page_size); in irdma_alloc_iwmr() 3425 if (total > iwmr->page_cnt) in irdma_reg_user_mr_type_qp() 3457 if (total > iwmr->page_cnt) in irdma_reg_user_mr_type_srq() 3491 if (total > iwmr->page_cnt) in irdma_reg_user_mr_type_cq() 3759 iwmr->page_cnt = ib_umem_num_dma_blocks(region, iwmr->page_size); in irdma_rereg_mr_trans() [all …]
|
| /linux/drivers/net/ethernet/qlogic/qede/ |
| H A D | qede_main.c | 2171 u32 page_cnt = qed_chain_get_page_cnt(&txq->tx_pbl); in qede_start_txq() local 2192 page_cnt, &ret_params); in qede_start_txq() 2261 u32 page_cnt; in qede_start_queues() local 2277 page_cnt = qed_chain_get_page_cnt(&rxq->rx_comp_ring); in qede_start_queues() 2283 page_cnt, &ret_params); in qede_start_queues()
|
| /linux/include/linux/ |
| H A D | bpf.h | 678 void *bpf_arena_alloc_pages_non_sleepable(void *p__map, void *addr__ign, u32 page_cnt, int node_id, 680 void bpf_arena_free_pages_non_sleepable(void *p__map, void *ptr__ign, u32 page_cnt); 682 static inline void *bpf_arena_alloc_pages_non_sleepable(void *p__map, void *addr__ign, u32 page_cnt, in bpf_arena_alloc_pages_non_sleepable() argument 688 static inline void bpf_arena_free_pages_non_sleepable(void *p__map, void *ptr__ign, u32 page_cnt) in bpf_arena_free_pages_non_sleepable() argument
|
| /linux/tools/lib/bpf/ |
| H A D | libbpf.c | 13926 static struct perf_buffer *__perf_buffer__new(int map_fd, size_t page_cnt, in __perf_buffer__new() 13929 struct perf_buffer *perf_buffer__new(int map_fd, size_t page_cnt, in __perf_buffer__new() 13959 return libbpf_ptr(__perf_buffer__new(map_fd, page_cnt, &p)); 13962 struct perf_buffer *perf_buffer__new_raw(int map_fd, size_t page_cnt, 13982 return libbpf_ptr(__perf_buffer__new(map_fd, page_cnt, &p)); in perf_buffer__process_record() 13985 static struct perf_buffer *__perf_buffer__new(int map_fd, size_t page_cnt, in perf_buffer__process_record() 13995 if (page_cnt == 0 || (page_cnt & (page_cnt - 1))) { in perf_buffer__process_record() 13997 page_cnt); in perf_buffer__process_record() 13750 perf_buffer__new(int map_fd,size_t page_cnt,perf_buffer_sample_fn sample_cb,perf_buffer_lost_fn lost_cb,void * ctx,const struct perf_buffer_opts * opts) perf_buffer__new() argument 13783 perf_buffer__new_raw(int map_fd,size_t page_cnt,struct perf_event_attr * attr,perf_buffer_event_fn event_cb,void * ctx,const struct perf_buffer_raw_opts * opts) perf_buffer__new_raw() argument 13806 __perf_buffer__new(int map_fd,size_t page_cnt,struct perf_buffer_params * p) __perf_buffer__new() argument [all...] |