Home
last modified time | relevance | path

Searched refs:page_cnt (Results 1 – 22 of 22) sorted by relevance

/linux/kernel/bpf/
H A Darena.c49 static void arena_free_pages(struct bpf_arena *arena, long uaddr, long page_cnt, bool sleepable);
73 u32 page_cnt; member
549 static long arena_alloc_pages(struct bpf_arena *arena, long uaddr, long page_cnt, int node_id, in arena_alloc_pages() argument
569 if (page_cnt > page_cnt_max) in arena_alloc_pages()
576 if (pgoff > page_cnt_max - page_cnt) in arena_alloc_pages()
583 alloc_pages = min(page_cnt, KMALLOC_MAX_CACHE_SIZE / sizeof(struct page *)); in arena_alloc_pages()
595 ret = is_range_tree_set(&arena->rt, pgoff, page_cnt); in arena_alloc_pages()
598 ret = range_tree_clear(&arena->rt, pgoff, page_cnt); in arena_alloc_pages()
600 ret = pgoff = range_tree_find(&arena->rt, page_cnt); in arena_alloc_pages()
602 ret = range_tree_clear(&arena->rt, pgoff, page_cnt); in arena_alloc_pages()
[all …]
/linux/drivers/net/ethernet/qlogic/qed/
H A Dqed_chain.c12 u32 page_cnt) in qed_chain_init() argument
34 chain->page_cnt = page_cnt; in qed_chain_init()
35 chain->capacity = chain->usable_per_page * page_cnt; in qed_chain_init()
36 chain->size = chain->elem_per_page * page_cnt; in qed_chain_init()
80 for (i = 0; i < chain->page_cnt; i++) { in qed_chain_free_next_ptr()
114 for (i = 0; i < chain->page_cnt; i++) { in qed_chain_free_pbl()
160 u32 page_cnt) in qed_chain_alloc_sanity_check() argument
165 chain_size *= page_cnt; in qed_chain_alloc_sanity_check()
206 for (i = 0; i < chain->page_cnt; i++) { in qed_chain_alloc_next_ptr()
255 u32 page_cnt, i; in qed_chain_alloc_pbl() local
[all …]
H A Dqed_sp_commands.c311 u8 page_cnt, i; in qed_sp_pf_start() local
372 page_cnt = (u8)qed_chain_get_page_cnt(&p_hwfn->p_eq->chain); in qed_sp_pf_start()
373 p_ramrod->event_ring_num_pages = page_cnt; in qed_sp_pf_start()
378 page_cnt = (u8)qed_chain_get_page_cnt(&p_hwfn->p_consq->chain); in qed_sp_pf_start()
379 p_ramrod->consolid_q_num_pages = page_cnt; in qed_sp_pf_start()
/linux/include/linux/qed/
H A Dqed_chain.h106 u32 page_cnt; member
262 return chain->page_cnt; in qed_chain_get_page_cnt()
303 if (++(*(u16 *)page_to_inc) == p_chain->page_cnt) in qed_chain_advance_page()
307 if (++(*(u32 *)page_to_inc) == p_chain->page_cnt) in qed_chain_advance_page()
498 u32 reset_val = p_chain->page_cnt - 1; in qed_chain_reset()
556 last_page_idx = p_chain->page_cnt - 1; in qed_chain_get_last_elem()
580 u32 cur_prod, page_mask, page_cnt, page_diff; in qed_chain_set_prod() local
599 page_cnt = qed_chain_get_page_cnt(p_chain); in qed_chain_set_prod()
603 page_diff + page_cnt) % page_cnt; in qed_chain_set_prod()
607 page_diff + page_cnt) % page_cnt; in qed_chain_set_prod()
[all …]
/linux/tools/testing/selftests/bpf/
H A Dbpf_arena_common.h47 void __arena* bpf_arena_alloc_pages(void *map, void __arena *addr, __u32 page_cnt,
49 int bpf_arena_reserve_pages(void *map, void __arena *addr, __u32 page_cnt) __ksym __weak;
50 void bpf_arena_free_pages(void *map, void __arena *ptr, __u32 page_cnt) __ksym __weak;
66 static inline void __arena* bpf_arena_alloc_pages(void *map, void *addr, __u32 page_cnt, in bpf_arena_alloc_pages() argument
71 static inline void bpf_arena_free_pages(void *map, void __arena *ptr, __u32 page_cnt) in bpf_arena_free_pages() argument
/linux/arch/mips/cavium-octeon/executive/
H A Dcvmx-helper-util.c133 union cvmx_ipd_portx_bp_page_cnt page_cnt; in cvmx_helper_setup_red() local
141 page_cnt.u64 = 0; in cvmx_helper_setup_red()
142 page_cnt.s.bp_enb = 0; in cvmx_helper_setup_red()
143 page_cnt.s.page_cnt = 100; in cvmx_helper_setup_red()
148 page_cnt.u64); in cvmx_helper_setup_red()
/linux/drivers/base/firmware_loader/
H A Dsysfs.c252 int page_cnt = min_t(size_t, PAGE_SIZE - page_ofs, count); in firmware_rw() local
256 page_ofs, page_cnt); in firmware_rw()
259 buffer, page_cnt); in firmware_rw()
261 buffer += page_cnt; in firmware_rw()
262 offset += page_cnt; in firmware_rw()
263 count -= page_cnt; in firmware_rw()
/linux/drivers/infiniband/hw/hns/
H A Dhns_roce_mr.c620 int page_cnt = 0; in cal_mtr_pg_cnt() local
625 page_cnt += region->count; in cal_mtr_pg_cnt()
628 return page_cnt; in cal_mtr_pg_cnt()
692 dma_addr_t *pages, unsigned int page_cnt) in hns_roce_mtr_map() argument
709 mapped_cnt < page_cnt; i++) { in hns_roce_mtr_map()
712 if (r->offset + r->count > page_cnt) { in hns_roce_mtr_map()
716 i, r->offset, r->count, page_cnt); in hns_roce_mtr_map()
721 page_cnt - mapped_cnt); in hns_roce_mtr_map()
732 if (mapped_cnt < page_cnt) { in hns_roce_mtr_map()
735 mapped_cnt, page_cnt); in hns_roce_mtr_map()
[all …]
/linux/tools/sched_ext/include/scx/
H A Dbpf_arena_common.h26 static inline void __arena* bpf_arena_alloc_pages(void *map, void *addr, __u32 page_cnt, in bpf_arena_alloc_pages() argument
31 static inline void bpf_arena_free_pages(void *map, void __arena *ptr, __u32 page_cnt) in bpf_arena_free_pages() argument
H A Dbpf_arena_common.bpf.h83 void __arena* bpf_arena_alloc_pages(void *map, void __arena *addr, __u32 page_cnt,
85 void bpf_arena_free_pages(void *map, void __arena *ptr, __u32 page_cnt) __ksym __weak;
86 int bpf_arena_reserve_pages(void *map, void __arena *ptr, __u32 page_cnt) __ksym __weak;
/linux/drivers/hv/
H A Dring_buffer.c187 struct page *pages, u32 page_cnt, u32 max_pkt_size, in hv_ringbuffer_init() argument
199 pages_wraparound = kzalloc_objs(struct page *, page_cnt * 2 - 1); in hv_ringbuffer_init()
204 for (i = 0; i < 2 * (page_cnt - 1); i++) in hv_ringbuffer_init()
206 &pages[i % (page_cnt - 1) + 1]; in hv_ringbuffer_init()
209 vmap(pages_wraparound, page_cnt * 2 - 1, VM_MAP, in hv_ringbuffer_init()
228 ring_info->ring_size = page_cnt << PAGE_SHIFT; in hv_ringbuffer_init()
/linux/arch/mips/include/asm/octeon/
H A Dcvmx-ipd-defs.h324 uint64_t page_cnt:17; member
326 uint64_t page_cnt:17;
999 uint64_t page_cnt:17; member
1001 uint64_t page_cnt:17;
1014 uint64_t page_cnt:17; member
1016 uint64_t page_cnt:17;
1029 uint64_t page_cnt:17; member
1031 uint64_t page_cnt:17;
1387 uint64_t page_cnt:25; member
1389 uint64_t page_cnt:25;
/linux/tools/testing/selftests/bpf/progs/
H A Dverifier_arena_large.c190 __noinline int alloc_pages(int page_cnt, int pages_atonce, bool first_pass, in alloc_pages()
196 for (i = 0; i < page_cnt; i++) { in alloc_pages()
177 alloc_pages(int page_cnt,int pages_atonce,bool first_pass,int max_idx,int step) alloc_pages() argument
/linux/arch/x86/kernel/cpu/sgx/
H A Dioctl.c29 if (!(encl->page_cnt % SGX_VA_SLOT_COUNT)) { in sgx_encl_grow()
41 WARN_ON_ONCE(encl->page_cnt % SGX_VA_SLOT_COUNT); in sgx_encl_grow()
43 encl->page_cnt++; in sgx_encl_grow()
49 encl->page_cnt--; in sgx_encl_shrink()
/linux/drivers/infiniband/hw/efa/
H A Defa_verbs.c1285 static struct scatterlist *efa_vmalloc_buf_to_sg(u64 *buf, int page_cnt) in efa_vmalloc_buf_to_sg() argument
1291 sglist = kmalloc_objs(*sglist, page_cnt); in efa_vmalloc_buf_to_sg()
1294 sg_init_table(sglist, page_cnt); in efa_vmalloc_buf_to_sg()
1295 for (i = 0; i < page_cnt; i++) { in efa_vmalloc_buf_to_sg()
1316 int page_cnt = pbl->phys.indirect.pbl_buf_size_in_pages; in pbl_chunk_list_create() local
1327 chunk_list_size = DIV_ROUND_UP(page_cnt, EFA_PTRS_PER_CHUNK); in pbl_chunk_list_create()
1336 page_cnt); in pbl_chunk_list_create()
1347 ((page_cnt % EFA_PTRS_PER_CHUNK) * EFA_CHUNK_PAYLOAD_PTR_SIZE) + in pbl_chunk_list_create()
/linux/drivers/infiniband/hw/qedr/
H A Dmain.c785 u32 page_cnt; in qedr_init_hw() local
799 page_cnt = qed_chain_get_page_cnt(&dev->cnq_array[i].pbl); in qedr_init_hw()
800 cur_pbl->num_pbl_pages = page_cnt; in qedr_init_hw()
/linux/drivers/net/wireless/realtek/rtw88/
H A Dfw.c2042 u8 page_cnt, pages; in _rtw_hw_scan_update_probe_req() local
2046 page_cnt = RTW_OLD_PROBE_PG_CNT; in _rtw_hw_scan_update_probe_req()
2048 page_cnt = RTW_PROBE_PG_CNT; in _rtw_hw_scan_update_probe_req()
2050 pages = page_offset + num_probes * page_cnt; in _rtw_hw_scan_update_probe_req()
2060 if (skb->len > page_size * page_cnt) { in _rtw_hw_scan_update_probe_req()
2070 buf_offset += page_cnt * page_size; in _rtw_hw_scan_update_probe_req()
2071 page_offset += page_cnt; in _rtw_hw_scan_update_probe_req()
/linux/drivers/infiniband/hw/irdma/
H A Dverbs.h117 u32 page_cnt; member
H A Dverbs.c2882 status = irdma_get_pble(rf->pble_rsrc, palloc, iwmr->page_cnt, in irdma_setup_pbles()
3176 iwmr->page_cnt = max_num_sg; in irdma_alloc_mr()
3179 err_code = irdma_get_pble(iwdev->rf->pble_rsrc, palloc, iwmr->page_cnt, in irdma_alloc_mr()
3213 if (unlikely(iwmr->npages == iwmr->page_cnt)) in irdma_set_page()
3321 lvl = iwmr->page_cnt != 1 ? PBLE_LEVEL_1 | PBLE_LEVEL_2 : PBLE_LEVEL_0; in irdma_reg_user_mr_type_mem()
3397 iwmr->page_cnt = ib_umem_num_dma_blocks(region, iwmr->page_size); in irdma_alloc_iwmr()
3425 if (total > iwmr->page_cnt) in irdma_reg_user_mr_type_qp()
3457 if (total > iwmr->page_cnt) in irdma_reg_user_mr_type_srq()
3491 if (total > iwmr->page_cnt) in irdma_reg_user_mr_type_cq()
3759 iwmr->page_cnt = ib_umem_num_dma_blocks(region, iwmr->page_size); in irdma_rereg_mr_trans()
[all …]
/linux/drivers/net/ethernet/qlogic/qede/
H A Dqede_main.c2171 u32 page_cnt = qed_chain_get_page_cnt(&txq->tx_pbl); in qede_start_txq() local
2192 page_cnt, &ret_params); in qede_start_txq()
2261 u32 page_cnt; in qede_start_queues() local
2277 page_cnt = qed_chain_get_page_cnt(&rxq->rx_comp_ring); in qede_start_queues()
2283 page_cnt, &ret_params); in qede_start_queues()
/linux/include/linux/
H A Dbpf.h678 void *bpf_arena_alloc_pages_non_sleepable(void *p__map, void *addr__ign, u32 page_cnt, int node_id,
680 void bpf_arena_free_pages_non_sleepable(void *p__map, void *ptr__ign, u32 page_cnt);
682 static inline void *bpf_arena_alloc_pages_non_sleepable(void *p__map, void *addr__ign, u32 page_cnt, in bpf_arena_alloc_pages_non_sleepable() argument
688 static inline void bpf_arena_free_pages_non_sleepable(void *p__map, void *ptr__ign, u32 page_cnt) in bpf_arena_free_pages_non_sleepable() argument
/linux/tools/lib/bpf/
H A Dlibbpf.c13926 static struct perf_buffer *__perf_buffer__new(int map_fd, size_t page_cnt, in __perf_buffer__new()
13929 struct perf_buffer *perf_buffer__new(int map_fd, size_t page_cnt, in __perf_buffer__new()
13959 return libbpf_ptr(__perf_buffer__new(map_fd, page_cnt, &p));
13962 struct perf_buffer *perf_buffer__new_raw(int map_fd, size_t page_cnt,
13982 return libbpf_ptr(__perf_buffer__new(map_fd, page_cnt, &p)); in perf_buffer__process_record()
13985 static struct perf_buffer *__perf_buffer__new(int map_fd, size_t page_cnt, in perf_buffer__process_record()
13995 if (page_cnt == 0 || (page_cnt & (page_cnt - 1))) { in perf_buffer__process_record()
13997 page_cnt); in perf_buffer__process_record()
13750 perf_buffer__new(int map_fd,size_t page_cnt,perf_buffer_sample_fn sample_cb,perf_buffer_lost_fn lost_cb,void * ctx,const struct perf_buffer_opts * opts) perf_buffer__new() argument
13783 perf_buffer__new_raw(int map_fd,size_t page_cnt,struct perf_event_attr * attr,perf_buffer_event_fn event_cb,void * ctx,const struct perf_buffer_raw_opts * opts) perf_buffer__new_raw() argument
13806 __perf_buffer__new(int map_fd,size_t page_cnt,struct perf_buffer_params * p) __perf_buffer__new() argument
[all...]