| /linux/drivers/gpu/drm/radeon/ |
| H A D | radeon_cs.c | 292 chunk_array_ptr = (uint64_t *)(unsigned long)(cs->chunks); in radeon_cs_parser_init() 299 p->chunks = kvzalloc_objs(struct radeon_cs_chunk, p->nchunks); in radeon_cs_parser_init() 300 if (p->chunks == NULL) { in radeon_cs_parser_init() 313 p->chunks[i].length_dw = user_chunk.length_dw; in radeon_cs_parser_init() 315 p->chunk_relocs = &p->chunks[i]; in radeon_cs_parser_init() 318 p->chunk_ib = &p->chunks[i]; in radeon_cs_parser_init() 320 if (p->chunks[i].length_dw == 0) in radeon_cs_parser_init() 324 p->chunk_const_ib = &p->chunks[i]; in radeon_cs_parser_init() 326 if (p->chunks[i].length_dw == 0) in radeon_cs_parser_init() 330 p->chunk_flags = &p->chunks[i]; in radeon_cs_parser_init() [all …]
|
| /linux/drivers/net/ethernet/netronome/nfp/nfpcore/ |
| H A D | nfp_nsp.c | 505 } *chunks; in nfp_nsp_command_buf_dma_sg() local 517 chunks = kzalloc_objs(*chunks, nseg); in nfp_nsp_command_buf_dma_sg() 518 if (!chunks) in nfp_nsp_command_buf_dma_sg() 526 chunks[i].chunk = kmalloc(chunk_size, in nfp_nsp_command_buf_dma_sg() 528 if (!chunks[i].chunk) in nfp_nsp_command_buf_dma_sg() 531 chunks[i].len = min_t(u64, chunk_size, max_size - off); in nfp_nsp_command_buf_dma_sg() 536 memcpy(chunks[i].chunk, arg->in_buf + off, coff); in nfp_nsp_command_buf_dma_sg() 538 memset(chunks[i].chunk + coff, 0, chunk_size - coff); in nfp_nsp_command_buf_dma_sg() 540 off += chunks[i].len; in nfp_nsp_command_buf_dma_sg() 548 addr = dma_map_single(dev, chunks[i].chunk, chunks[i].len, in nfp_nsp_command_buf_dma_sg() [all …]
|
| /linux/drivers/comedi/drivers/ni_routing/tools/ |
| H A D | convert_csv_to_c.py | 228 chunks = [ self.output_file_top, 244 chunks.append('\t&{},'.format(dev_table_name)) 273 chunks.append('\tNULL,') # terminate list 274 chunks.append('};') 275 return '\n'.join(chunks) 416 chunks = [ self.output_file_top, 432 chunks.append('\t&{},'.format(fam_table_name)) 462 chunks.append('\tNULL,') # terminate list 463 chunks.append('};') 464 return '\n'.join(chunks)
|
| /linux/drivers/net/mctp/ |
| H A D | mctp-serial.c | 534 u8 chunks[MAX_CHUNKS]; member 552 KUNIT_EXPECT_EQ(test, next, params->chunks[i]); in test_next_chunk_len() 567 .chunks = { 3, 1, 1, 0}, 572 .chunks = { 3, 1, 1, 0}, 577 .chunks = { 1, 2, 0}, 582 .chunks = { 1, 1, 1, 0}, 587 .chunks = { 1, 1, 1, 1, 0}, 592 .chunks = { 1, 1, 1, 1, 2, 0}, 597 .chunks = { 1, 0 }, 602 .chunks = { 1, 0 }, [all …]
|
| /linux/drivers/infiniband/hw/usnic/ |
| H A D | usnic_vnic.c | 44 struct usnic_vnic_res_chunk chunks[USNIC_VNIC_RES_TYPE_MAX]; member 117 for (i = 0; i < ARRAY_SIZE(vnic->chunks); i++) { in usnic_vnic_dump() 118 chunk = &vnic->chunks[i]; in usnic_vnic_dump() 222 return vnic->chunks[type].cnt; in usnic_vnic_res_cnt() 228 return vnic->chunks[type].free_cnt; in usnic_vnic_res_free_cnt() 254 src = &vnic->chunks[type]; in usnic_vnic_get_resources() 286 vnic->chunks[res->type].free_cnt++; in usnic_vnic_put_resources() 382 &vnic->chunks[res_type]); in usnic_vnic_discover_resources() 391 usnic_vnic_free_res_chunk(&vnic->chunks[res_type]); in usnic_vnic_discover_resources() 427 usnic_vnic_free_res_chunk(&vnic->chunks[res_type]); in usnic_vnic_release_resources()
|
| /linux/scripts/gdb/linux/ |
| H A D | timerlist.py | 162 chunks = [] 168 chunks.append(buf[start:end]) 170 chunks.append(',') 174 chunks[0] = chunks[0][0] # Cut off the first 0 176 return "".join(str(chunks))
|
| /linux/net/xdp/ |
| H A D | xdp_umem.c | 164 u64 chunks, npgs; in xdp_umem_reg() local 199 chunks = div_u64_rem(size, chunk_size, &chunks_rem); in xdp_umem_reg() 200 if (!chunks || chunks > U32_MAX) in xdp_umem_reg() 218 umem->chunks = chunks; in xdp_umem_reg()
|
| /linux/drivers/gpu/drm/panthor/ |
| H A D | panthor_heap.c | 55 struct list_head chunks; member 169 if (initial_chunk && !list_empty(&heap->chunks)) { in panthor_alloc_heap_chunk() 173 prev_chunk = list_first_entry(&heap->chunks, in panthor_alloc_heap_chunk() 185 list_add(&chunk->node, &heap->chunks); in panthor_alloc_heap_chunk() 207 list_for_each_entry_safe(chunk, tmp, &heap->chunks, node) in panthor_free_heap_chunks() 313 INIT_LIST_HEAD(&heap->chunks); in panthor_heap_create() 322 first_chunk = list_first_entry(&heap->chunks, in panthor_heap_create() 393 list_for_each_entry_safe(chunk, tmp, &heap->chunks, node) { in panthor_heap_return_chunk() 480 chunk = list_first_entry(&heap->chunks, in panthor_heap_grow()
|
| /linux/drivers/md/ |
| H A D | md-llbitmap.c | 269 unsigned long chunks; member 542 unsigned int nr_pages = DIV_ROUND_UP(llbitmap->chunks + in llbitmap_cache_pages() 595 for (i = 0; i < llbitmap->chunks; i++) in llbitmap_init_state() 780 unsigned long chunks = DIV_ROUND_UP(blocks, chunksize); in llbitmap_init() local 784 while (chunks > space) { in llbitmap_init() 786 chunks = DIV_ROUND_UP_SECTOR_T(blocks, chunksize); in llbitmap_init() 792 llbitmap->chunks = chunks; in llbitmap_init() 799 llbitmap_state_machine(llbitmap, 0, llbitmap->chunks - 1, in llbitmap_init() 897 llbitmap->chunks = DIV_ROUND_UP_SECTOR_T(mddev->resync_max_sectors, chunksize); in llbitmap_read_sb() 936 end = min(llbitmap->chunks, PAGE_SIZE - BITMAP_DATA_OFFSET) - 1; in md_llbitmap_daemon_fn() [all …]
|
| H A D | md-bitmap.c | 164 unsigned long chunks; member 1014 unsigned long chunks, int with_super, in md_bitmap_storage_alloc() argument 1021 bytes = DIV_ROUND_UP(chunks, 8); in md_bitmap_storage_alloc() 1325 unsigned long chunks = bitmap->counts.chunks; in md_bitmap_init_from_disk() local 1337 for (i = 0; i < chunks ; i++) { in md_bitmap_init_from_disk() 1407 for (i = 0; i < chunks; i++) { in md_bitmap_init_from_disk() 1433 bit_cnt, chunks); in md_bitmap_init_from_disk() 1567 for (j = 0; j < counts->chunks; j++) { in bitmap_daemon_work() 2300 for (j = 0; j < counts->chunks; j++) { in bitmap_copy_from_slot() 2380 unsigned long chunks; in __bitmap_resize() local [all …]
|
| /linux/lib/ |
| H A D | genalloc.c | 160 INIT_LIST_HEAD(&pool->chunks); in gen_pool_create() 203 list_add_rcu(&chunk->next_chunk, &pool->chunks); in gen_pool_add_owner() 223 list_for_each_entry_rcu(chunk, &pool->chunks, next_chunk) { in gen_pool_virt_to_phys() 249 list_for_each_safe(_chunk, _next_chunk, &pool->chunks) { in gen_pool_destroy() 297 list_for_each_entry_rcu(chunk, &pool->chunks, next_chunk) { in gen_pool_alloc_algo_owner() 503 list_for_each_entry_rcu(chunk, &pool->chunks, next_chunk) { in gen_pool_free_owner() 538 list_for_each_entry_rcu(chunk, &(pool)->chunks, next_chunk) in gen_pool_for_each_chunk() 561 list_for_each_entry_rcu(chunk, &(pool)->chunks, next_chunk) { in gen_pool_has_addr() 586 list_for_each_entry_rcu(chunk, &pool->chunks, next_chunk) in gen_pool_avail() 605 list_for_each_entry_rcu(chunk, &pool->chunks, next_chunk) in gen_pool_size()
|
| /linux/Documentation/networking/ |
| H A D | oa-tc6-framework.rst | 49 each chunk. Ethernet frames are transferred over one or more data chunks. 59 receive (RX) chunks. Chunks in both transmit and receive directions may 69 In parallel, receive data chunks are received on MISO. Each receive data 160 the MAC-PHY will be converted into multiple transmit data chunks. Each 177 transaction. For TX data chunks, this bit shall be ’1’. 242 received from the MAC-PHY. The SPI host should not write more data chunks 248 chunks, the MAC-PHY interrupt is asserted to SPI host. On reception of the 254 host will be sent as multiple receive data chunks. Each receive data 286 data chunks of frame data that are available for 340 of transmit data chunks of frame data that the SPI host [all …]
|
| /linux/drivers/net/ethernet/intel/idpf/ |
| H A D | idpf_virtchnl.c | 727 const void *chunks; member 772 const void *pos = params->chunks; in idpf_send_chunked_msg() 1303 struct virtchnl2_queue_reg_chunk *schunk = &schunks->chunks[i]; in idpf_vport_init_queue_reg_chunks() 1325 struct virtchnl2_vector_chunks *chunks; in idpf_get_reg_intr_vecs() local 1330 chunks = &adapter->req_vec_chunks->vchunks; in idpf_get_reg_intr_vecs() 1331 num_vchunks = le16_to_cpu(chunks->num_vchunks); in idpf_get_reg_intr_vecs() 1338 chunk = &chunks->vchunks[j]; in idpf_get_reg_intr_vecs() 1375 struct idpf_queue_id_reg_info *chunks) in idpf_vport_get_q_reg() argument 1377 u16 num_chunks = chunks->num_chunks; in idpf_vport_get_q_reg() 1385 chunk = &chunks->queue_chunks[num_chunks]; in idpf_vport_get_q_reg() [all …]
|
| /linux/tools/testing/selftests/bpf/progs/ |
| H A D | dynptr_success.c | 672 int i, chunks = 200; in test_dynptr_copy_xdp() 677 bpf_ringbuf_reserve_dynptr(&ringbuf, len * chunks, 0, &ptr_buf); in test_dynptr_copy_xdp() 680 bpf_for(i, 0, chunks) { in test_dynptr_copy_xdp() 686 err = bpf_dynptr_copy(&ptr_xdp, 0, &ptr_buf, 0, len * chunks); 690 bpf_for(i, 0, chunks) { in test_dynptr_memset_zero() 701 bpf_for(i, 0, chunks) { in test_dynptr_memset_zero() 707 err = bpf_dynptr_copy(&ptr_buf, 0, &ptr_xdp, 0, len * chunks); 711 bpf_for(i, 0, chunks) { in test_dynptr_memset_notzero() 721 err = bpf_dynptr_copy(&ptr_xdp, 2, &ptr_xdp, len, len * (chunks - 1)); in test_dynptr_memset_notzero() 725 bpf_for(i, 0, chunks 617 int i, chunks = 200; test_dynptr_copy_xdp() local [all...] |
| /linux/tools/testing/selftests/drivers/net/mlxsw/spectrum/ |
| H A D | devlink_lib_spectrum.sh | 90 devlink_resource_size_set 32000 kvd linear chunks 99 devlink_resource_size_set 32000 kvd linear chunks 108 devlink_resource_size_set 49152 kvd linear chunks
|
| /linux/kernel/ |
| H A D | audit_tree.c | 17 struct list_head chunks; member 103 INIT_LIST_HEAD(&tree->chunks); in alloc_tree() 437 list_add(&chunk->owners[0].list, &tree->chunks); in create_chunk() 509 list_add(&p->list, &tree->chunks); in tag_chunk() 574 while (!list_empty(&victim->chunks)) { in prune_tree_chunks() 579 p = list_first_entry(&victim->chunks, struct audit_node, list); in prune_tree_chunks() 620 for (p = tree->chunks.next; p != &tree->chunks; p = q) { in trim_marked() 625 list_add(p, &tree->chunks); in trim_marked() 702 list_for_each_entry(node, &tree->chunks, list) { in audit_trim_trees() 851 list_for_each_entry(node, &tree->chunks, list) in audit_add_tree_rule() [all …]
|
| /linux/net/sctp/ |
| H A D | chunk.c | 43 INIT_LIST_HEAD(&msg->chunks); in sctp_datamsg_init() 65 list_for_each_entry(chunk, &msg->chunks, frag_list) in sctp_datamsg_free() 81 list_for_each_safe(pos, temp, &msg->chunks) { in sctp_datamsg_destroy() 281 list_add_tail(&chunk->frag_list, &msg->chunks); in sctp_datamsg_from_user() 290 list_for_each_safe(pos, temp, &msg->chunks) { in sctp_datamsg_from_user()
|
| H A D | auth.c | 187 struct sctp_chunks_param *chunks, in sctp_auth_make_key_vector() argument 198 if (chunks) in sctp_auth_make_key_vector() 199 chunks_len = ntohs(chunks->param_hdr.length); in sctp_auth_make_key_vector() 210 if (chunks) { in sctp_auth_make_key_vector() 211 memcpy(new->data + offset, chunks, chunks_len); in sctp_auth_make_key_vector() 565 switch (param->chunks[i]) { in __sctp_auth_cid() 573 if (param->chunks[i] == chunk) in __sctp_auth_cid() 678 p->chunks[nchunks] = chunk_id; in sctp_auth_ep_add_chunkid()
|
| /linux/drivers/gpu/drm/amd/amdgpu/ |
| H A D | amdgpu_cs.c | 188 chunk_array = memdup_array_user(u64_to_user_ptr(cs->in.chunks), in amdgpu_cs_pass1() 195 p->chunks = kvmalloc_objs(struct amdgpu_cs_chunk, p->nchunks); in amdgpu_cs_pass1() 196 if (!p->chunks) { in amdgpu_cs_pass1() 212 p->chunks[i].chunk_id = user_chunk.chunk_id; in amdgpu_cs_pass1() 213 p->chunks[i].length_dw = user_chunk.length_dw; in amdgpu_cs_pass1() 215 size = p->chunks[i].length_dw; in amdgpu_cs_pass1() 217 p->chunks[i].kdata = vmemdup_array_user(u64_to_user_ptr(user_chunk.chunk_data), in amdgpu_cs_pass1() 220 if (IS_ERR(p->chunks[i].kdata)) { in amdgpu_cs_pass1() 221 ret = PTR_ERR(p->chunks[i].kdata); in amdgpu_cs_pass1() 229 switch (p->chunks[i].chunk_id) { in amdgpu_cs_pass1() [all …]
|
| /linux/drivers/infiniband/ulp/rtrs/ |
| H A D | README | 28 session. A session is associated with a set of memory chunks reserved on the 36 chunks reserved for him on the server side. Their number, size and addresses 45 which of the memory chunks has been accessed and at which offset the message 80 the server (number of memory chunks which are going to be allocated for that 122 1. When processing a write request client selects one of the memory chunks 139 1. When processing a write request client selects one of the memory chunks 144 using the IMM field, Server invalidate rkey associated to the memory chunks 162 1. When processing a read request client selects one of the memory chunks 181 1. When processing a read request client selects one of the memory chunks 186 Server invalidate rkey associated to the memory chunks first, when it finishes,
|
| /linux/drivers/infiniband/hw/efa/ |
| H A D | efa_verbs.c | 116 struct pbl_chunk *chunks; member 1377 chunk_list->chunks = kzalloc_objs(*chunk_list->chunks, chunk_list_size); in pbl_chunk_list_create() 1378 if (!chunk_list->chunks) in pbl_chunk_list_create() 1387 chunk_list->chunks[i].buf = kzalloc(EFA_CHUNK_SIZE, GFP_KERNEL); in pbl_chunk_list_create() 1388 if (!chunk_list->chunks[i].buf) in pbl_chunk_list_create() 1391 chunk_list->chunks[i].length = EFA_CHUNK_USED_SIZE; in pbl_chunk_list_create() 1393 chunk_list->chunks[chunk_list_size - 1].length = in pbl_chunk_list_create() 1400 cur_chunk_buf = chunk_list->chunks[0].buf; in pbl_chunk_list_create() 1408 cur_chunk_buf = chunk_list->chunks[chunk_idx].buf; in pbl_chunk_list_create() 1416 chunk_list->chunks[i].buf, in pbl_chunk_list_create() [all …]
|
| /linux/drivers/net/wireless/ti/wlcore/ |
| H A D | boot.c | 237 u32 chunks, addr, len; in wlcore_boot_upload_firmware() local 242 chunks = be32_to_cpup((__be32 *) fw); in wlcore_boot_upload_firmware() 245 wl1271_debug(DEBUG_BOOT, "firmware chunks to be uploaded: %u", chunks); in wlcore_boot_upload_firmware() 247 while (chunks--) { in wlcore_boot_upload_firmware() 258 chunks, addr, len); in wlcore_boot_upload_firmware()
|
| /linux/drivers/net/wireless/intel/iwlwifi/pcie/ |
| H A D | ctxt-info-v2.c | 365 len0 = pnvm_data->chunks[0].len; in iwl_pcie_load_payloads_contig() 366 len1 = pnvm_data->chunks[1].len; in iwl_pcie_load_payloads_contig() 381 memcpy(dram->block, pnvm_data->chunks[0].data, len0); in iwl_pcie_load_payloads_contig() 382 memcpy((u8 *)dram->block + len0, pnvm_data->chunks[1].data, len1); in iwl_pcie_load_payloads_contig() 415 len = pnvm_data->chunks[i].len; in iwl_pcie_load_payloads_segments() 416 data = pnvm_data->chunks[i].data; in iwl_pcie_load_payloads_segments()
|
| /linux/tools/testing/selftests/bpf/ |
| H A D | generate_udp_fragments.py | 46 chunks = [frag[i : i + 10] for i in range(0, len(frag), 10)] 47 chunks_fmted = [", ".join([str(hex(b)) for b in chunk]) for chunk in chunks]
|
| /linux/drivers/virt/vboxguest/ |
| H A D | vboxguest_core.c | 360 u32 i, chunks; in vbg_balloon_work() local 388 chunks = req->balloon_chunks; in vbg_balloon_work() 389 if (chunks > gdev->mem_balloon.max_chunks) { in vbg_balloon_work() 391 __func__, chunks, gdev->mem_balloon.max_chunks); in vbg_balloon_work() 395 if (chunks > gdev->mem_balloon.chunks) { in vbg_balloon_work() 397 for (i = gdev->mem_balloon.chunks; i < chunks; i++) { in vbg_balloon_work() 402 gdev->mem_balloon.chunks++; in vbg_balloon_work() 406 for (i = gdev->mem_balloon.chunks; i-- > chunks;) { in vbg_balloon_work() 411 gdev->mem_balloon.chunks--; in vbg_balloon_work() 1666 balloon_info->u.out.balloon_chunks = gdev->mem_balloon.chunks; in vbg_ioctl_check_balloon()
|