/linux/security/apparmor/ |
H A D | policy_compat.c | 161 table = kvcalloc(state_count * 2, sizeof(struct aa_perms), GFP_KERNEL); in compute_fperms() 185 perms = kvcalloc(state_count, sizeof(struct aa_perms), GFP_KERNEL); in compute_xmatch_perms() 260 table = kvcalloc(state_count, sizeof(struct aa_perms), GFP_KERNEL); in compute_perms()
|
/linux/scripts/coccinelle/api/ |
H A D | kfree_mismatch.cocci | 57 E = \(kvmalloc\|kvzalloc\|kvcalloc\|kvzalloc_node\|kvmalloc_node\| 123 * E = \(kvmalloc\|kvzalloc\|kvcalloc\|kvzalloc_node\|kvmalloc_node\| 133 E = \(kvmalloc\|kvzalloc\|kvcalloc\|kvzalloc_node\|kvmalloc_node\|
|
/linux/mm/ |
H A D | swap_slots.c | 123 slots = kvcalloc(SWAP_SLOTS_CACHE_SIZE, sizeof(swp_entry_t), in alloc_swap_slot_cache() 128 slots_ret = kvcalloc(SWAP_SLOTS_CACHE_SIZE, sizeof(swp_entry_t), in alloc_swap_slot_cache()
|
H A D | gup_test.c | 115 pages = kvcalloc(nr_pages, sizeof(void *), GFP_KERNEL); in __gup_test_ioctl() 248 pages = kvcalloc(nr_pages, sizeof(void *), GFP_KERNEL); in pin_longterm_test_start()
|
/linux/drivers/net/wireguard/ |
H A D | ratelimiter.c | 179 table_v4 = kvcalloc(table_size, sizeof(*table_v4), GFP_KERNEL); in wg_ratelimiter_init() 184 table_v6 = kvcalloc(table_size, sizeof(*table_v6), GFP_KERNEL); in wg_ratelimiter_init()
|
/linux/net/xdp/ |
H A D | xsk_buff_pool.c | 47 pool->tx_descs = kvcalloc(xs->tx->nentries, sizeof(*pool->tx_descs), in xp_alloc_tx_descs() 68 pool->heads = kvcalloc(umem->chunks, sizeof(*pool->heads), GFP_KERNEL); in xp_create_and_assign_umem() 338 dma_map->dma_pages = kvcalloc(nr_pages, sizeof(*dma_map->dma_pages), GFP_KERNEL); in xp_create_dma_map() 426 pool->dma_pages = kvcalloc(dma_map->dma_pages_cnt, sizeof(*pool->dma_pages), GFP_KERNEL); in xp_init_dma_info()
|
H A D | xdp_umem.c | 100 umem->pgs = kvcalloc(umem->npgs, sizeof(*umem->pgs), GFP_KERNEL | __GFP_NOWARN); in xdp_umem_pin_pages()
|
/linux/drivers/net/ethernet/mellanox/mlx5/core/en/ |
H A D | qos.c | 90 stats_list = kvcalloc(mlx5e_qos_max_leaf_nodes(priv->mdev), in mlx5e_open_qos_sq() 279 sqs = kvcalloc(qos_sqs_size, sizeof(struct mlx5e_txqsq *), GFP_KERNEL); in mlx5e_qos_alloc_queues() 486 rl->leaves_id = kvcalloc(num_tc, sizeof(*rl->leaves_id), GFP_KERNEL); in mlx5e_mqprio_rl_init()
|
H A D | rx_res.c | 312 rx_res->rss_rqns = kvcalloc(max_nch, sizeof(*rx_res->rss_rqns), GFP_KERNEL); in mlx5e_rx_res_alloc() 319 rx_res->rss_vhca_ids = kvcalloc(max_nch, sizeof(*rx_res->rss_vhca_ids), GFP_KERNEL); in mlx5e_rx_res_alloc() 341 res->channels = kvcalloc(res->max_nch, sizeof(*res->channels), GFP_KERNEL); in mlx5e_rx_res_channels_init()
|
/linux/fs/exfat/ |
H A D | nls.c | 661 upcase_table = kvcalloc(UTBL_COUNT, sizeof(unsigned short), GFP_KERNEL); in exfat_load_upcase_table() 713 upcase_table = kvcalloc(UTBL_COUNT, sizeof(unsigned short), GFP_KERNEL); in exfat_load_default_upcase_table()
|
/linux/drivers/infiniband/core/ |
H A D | umem_odp.c | 78 umem_odp->pfn_list = kvcalloc( in ib_init_umem_odp() 83 umem_odp->dma_list = kvcalloc( in ib_init_umem_odp()
|
/linux/drivers/gpu/drm/nouveau/ |
H A D | nouveau_dmem.c | 381 src_pfns = kvcalloc(npages, sizeof(*src_pfns), GFP_KERNEL | __GFP_NOFAIL); in nouveau_dmem_evict_chunk() 382 dst_pfns = kvcalloc(npages, sizeof(*dst_pfns), GFP_KERNEL | __GFP_NOFAIL); in nouveau_dmem_evict_chunk() 383 dma_addrs = kvcalloc(npages, sizeof(*dma_addrs), GFP_KERNEL | __GFP_NOFAIL); in nouveau_dmem_evict_chunk()
|
/linux/drivers/net/ethernet/netronome/nfp/ |
H A D | nfp_net_dp.c | 321 rx_ring->xsk_rxbufs = kvcalloc(rx_ring->cnt, rxbuf_sw_desc_sz, in nfp_net_rx_ring_alloc() 326 rx_ring->rxbufs = kvcalloc(rx_ring->cnt, rxbuf_sw_desc_sz, in nfp_net_rx_ring_alloc()
|
/linux/drivers/gpu/drm/virtio/ |
H A D | virtgpu_submit.c | 107 syncobjs = kvcalloc(num_in_syncobjs, sizeof(*syncobjs), GFP_KERNEL); in virtio_gpu_parse_deps() 198 post_deps = kvcalloc(num_out_syncobjs, sizeof(*post_deps), GFP_KERNEL); in virtio_gpu_parse_post_deps()
|
/linux/fs/smb/client/compress/ |
H A D | lz77.c | 147 htable = kvcalloc(LZ77_HASH_SIZE, sizeof(*htable), GFP_KERNEL); in lz77_compress()
|
/linux/drivers/net/ethernet/netronome/nfp/abm/ |
H A D | main.c | 480 abm->thresholds = kvcalloc(abm->num_thresholds, in nfp_abm_init() 485 abm->actions = kvcalloc(abm->num_thresholds, sizeof(*abm->actions), in nfp_abm_init()
|
/linux/drivers/net/ethernet/netronome/nfp/nfdk/ |
H A D | rings.c | 108 tx_ring->ktxbufs = kvcalloc(tx_ring->cnt, sizeof(*tx_ring->ktxbufs), in nfp_nfdk_tx_ring_alloc()
|
/linux/net/sched/ |
H A D | sch_hhf.c | 605 q->hh_flows = kvcalloc(HH_FLOWS_CNT, sizeof(struct list_head), in hhf_init() 620 q->hhf_arrays[i] = kvcalloc(HHF_ARRAYS_LEN, in hhf_init()
|
H A D | sch_fq_codel.c | 498 q->flows = kvcalloc(q->flows_cnt, in fq_codel_init() 505 q->backlogs = kvcalloc(q->flows_cnt, sizeof(u32), GFP_KERNEL); in fq_codel_init()
|
/linux/drivers/net/ethernet/netronome/nfp/nfd3/ |
H A D | rings.c | 143 tx_ring->txbufs = kvcalloc(tx_ring->cnt, sizeof(*tx_ring->txbufs), in nfp_nfd3_tx_ring_alloc()
|
/linux/drivers/xen/ |
H A D | gntdev.c | 145 add->pages = kvcalloc(count, sizeof(add->pages[0]), GFP_KERNEL); in gntdev_alloc_map() 147 kvcalloc(count, sizeof(add->being_removed[0]), GFP_KERNEL); in gntdev_alloc_map() 173 add->frames = kvcalloc(count, sizeof(add->frames[0]), in gntdev_alloc_map()
|
/linux/fs/verity/ |
H A D | open.c | 235 vi->hash_block_verified = kvcalloc(BITS_TO_LONGS(num_bits), in fsverity_create_info()
|
/linux/drivers/video/fbdev/core/ |
H A D | fb_defio.c | 318 pagerefs = kvcalloc(npagerefs, sizeof(*pagerefs), GFP_KERNEL); in fb_deferred_io_init()
|
/linux/drivers/net/ethernet/google/gve/ |
H A D | gve_tx_dqo.c | 255 tx->dqo.tx_qpl_buf_next = kvcalloc(num_tx_qpl_bufs, in gve_tx_qpl_buf_init() 324 tx->dqo.pending_packets = kvcalloc(tx->dqo.num_pending_packets, in gve_tx_alloc_ring_dqo() 392 tx = kvcalloc(cfg->qcfg->max_queues, sizeof(struct gve_tx_ring), in gve_tx_alloc_rings_dqo()
|
/linux/block/ |
H A D | blk-crypto-profile.c | 96 profile->slots = kvcalloc(num_slots, sizeof(profile->slots[0]), in blk_crypto_profile_init()
|