| /linux/drivers/gpu/drm/panthor/ |
| H A D | panthor_heap.c | 40 * struct panthor_heap_chunk - Structure used to keep track of allocated heap chunks. 43 /** @node: Used to insert the heap chunk in panthor_heap::chunks. */ 54 /** @chunks: List containing all heap chunks allocated so far. */ 55 struct list_head chunks; member 57 /** @lock: Lock protecting insertion in the chunks list. */ 63 /** @max_chunks: Maximum number of chunks. */ 68 * we'd let the FW wait for fragment job to finish instead of allocating new chunks. 72 /** @chunk_count: Number of heap chunks currently allocated. */ 102 /** @size: Size of all chunks across all heaps in the pool. */ 169 if (initial_chunk && !list_empty(&heap->chunks)) { in panthor_alloc_heap_chunk() [all …]
|
| /linux/Documentation/networking/ |
| H A D | oa-tc6-framework.rst | 49 each chunk. Ethernet frames are transferred over one or more data chunks. 59 receive (RX) chunks. Chunks in both transmit and receive directions may 69 In parallel, receive data chunks are received on MISO. Each receive data 160 the MAC-PHY will be converted into multiple transmit data chunks. Each 177 transaction. For TX data chunks, this bit shall be ’1’. 242 received from the MAC-PHY. The SPI host should not write more data chunks 248 chunks, the MAC-PHY interrupt is asserted to SPI host. On reception of the 254 host will be sent as multiple receive data chunks. Each receive data 284 RCA (Bit 28..24) - Receive Chunks Available. The RCA field indicates to 286 data chunks of frame data that are available for [all …]
|
| /linux/drivers/gpu/drm/radeon/ |
| H A D | radeon_cs.c | 280 /* get chunks */ in radeon_cs_parser_init() 292 chunk_array_ptr = (uint64_t *)(unsigned long)(cs->chunks); in radeon_cs_parser_init() 299 p->chunks = kvzalloc_objs(struct radeon_cs_chunk, p->nchunks); in radeon_cs_parser_init() 300 if (p->chunks == NULL) { in radeon_cs_parser_init() 313 p->chunks[i].length_dw = user_chunk.length_dw; in radeon_cs_parser_init() 315 p->chunk_relocs = &p->chunks[i]; in radeon_cs_parser_init() 318 p->chunk_ib = &p->chunks[i]; in radeon_cs_parser_init() 320 if (p->chunks[i].length_dw == 0) in radeon_cs_parser_init() 324 p->chunk_const_ib = &p->chunks[i]; in radeon_cs_parser_init() 326 if (p->chunks[i].length_dw == 0) in radeon_cs_parser_init() [all …]
|
| /linux/net/sctp/ |
| H A D | chunk.c | 43 INIT_LIST_HEAD(&msg->chunks); in sctp_datamsg_init() 65 list_for_each_entry(chunk, &msg->chunks, frag_list) in sctp_datamsg_free() 81 list_for_each_safe(pos, temp, &msg->chunks) { in sctp_datamsg_destroy() 140 * down any such message into smaller chunks. Opportunistically, fragment 141 * the chunks down to the current MTU constraints. We may get refragmented 182 /* If the peer requested that we authenticate DATA chunks in sctp_datamsg_from_user() 183 * we need to account for bundling of the AUTH chunks along with in sctp_datamsg_from_user() 236 /* Create chunks for all DATA chunks. */ in sctp_datamsg_from_user() 281 list_add_tail(&chunk->frag_list, &msg->chunks); in sctp_datamsg_from_user() 290 list_for_each_safe(pos, temp, &msg->chunks) { in sctp_datamsg_from_user()
|
| H A D | outqueue.c | 11 * bundling and queueing of outgoing SCTP chunks. 204 /* Free the outqueue structure and any related pending chunks. 212 /* Throw away unacknowledged chunks. */ in __sctp_outq_teardown() 224 /* Throw away chunks that have been gap ACKed. */ in __sctp_outq_teardown() 233 /* Throw away any chunks in the retransmit queue. */ in __sctp_outq_teardown() 242 /* Throw away any chunks that are in the abandoned queue. */ in __sctp_outq_teardown() 251 /* Throw away any leftover data chunks. */ in __sctp_outq_teardown() 260 /* Throw away any leftover control chunks. */ in __sctp_outq_teardown() 273 /* Free the outqueue structure and any related pending chunks. */ 276 /* Throw away leftover chunks. */ in sctp_outq_free() [all …]
|
| H A D | auth.c | 176 * The RANDOM parameter, the CHUNKS parameter and the HMAC-ALGO 187 struct sctp_chunks_param *chunks, in sctp_auth_make_key_vector() argument 198 if (chunks) in sctp_auth_make_key_vector() 199 chunks_len = ntohs(chunks->param_hdr.length); in sctp_auth_make_key_vector() 210 if (chunks) { in sctp_auth_make_key_vector() 211 memcpy(new->data + offset, chunks, chunks_len); in sctp_auth_make_key_vector() 302 * The RANDOM parameter, the CHUNKS parameter and the HMAC-ALGO in sctp_auth_asoc_create_secret() 560 * chunks MUST NOT be listed in the CHUNKS parameter. However, if in __sctp_auth_cid() 561 * a CHUNKS parameter is received then the types for INIT, INIT-ACK, in __sctp_auth_cid() 562 * SHUTDOWN-COMPLETE and AUTH chunks MUST be ignored. in __sctp_auth_cid() [all …]
|
| H A D | ulpqueue.c | 102 /* Create a temporary list to collect chunks on. */ in sctp_ulpq_tail_data() 266 /* Helper function to store chunks that need to be reassembled. */ 412 /* The chunks are held in the reasm queue sorted by TSN. in sctp_ulpq_retrieve_reassembled() 414 * fragmented chunks that complete a datagram. in sctp_ulpq_retrieve_reassembled() 509 /* The chunks are held in the reasm queue sorted by TSN. in sctp_ulpq_retrieve_partial() 511 * sequence of fragmented chunks. in sctp_ulpq_retrieve_partial() 568 /* Helper function to reassemble chunks. Hold chunks on the reasm queue that 608 /* The chunks are held in the reasm queue sorted by TSN. in sctp_ulpq_retrieve_first() 610 * fragmented chunks that start a datagram. in sctp_ulpq_retrieve_first() 753 /* We are holding the chunks by stream, by SSN. */ in sctp_ulpq_retrieve_ordered() [all …]
|
| /linux/drivers/net/ethernet/netronome/nfp/nfpcore/ |
| H A D | nfp_nsp.c | 505 } *chunks; in nfp_nsp_command_buf_dma_sg() local 517 chunks = kzalloc_objs(*chunks, nseg); in nfp_nsp_command_buf_dma_sg() 518 if (!chunks) in nfp_nsp_command_buf_dma_sg() 526 chunks[i].chunk = kmalloc(chunk_size, in nfp_nsp_command_buf_dma_sg() 528 if (!chunks[i].chunk) in nfp_nsp_command_buf_dma_sg() 531 chunks[i].len = min_t(u64, chunk_size, max_size - off); in nfp_nsp_command_buf_dma_sg() 536 memcpy(chunks[i].chunk, arg->in_buf + off, coff); in nfp_nsp_command_buf_dma_sg() 538 memset(chunks[i].chunk + coff, 0, chunk_size - coff); in nfp_nsp_command_buf_dma_sg() 540 off += chunks[i].len; in nfp_nsp_command_buf_dma_sg() 548 addr = dma_map_single(dev, chunks[i].chunk, chunks[i].len, in nfp_nsp_command_buf_dma_sg() [all …]
|
| /linux/drivers/net/mctp/ |
| H A D | mctp-serial.c | 534 u8 chunks[MAX_CHUNKS]; member 552 KUNIT_EXPECT_EQ(test, next, params->chunks[i]); in test_next_chunk_len() 560 KUNIT_FAIL_AND_ABORT(test, "Ran out of chunks"); in test_next_chunk_len() 567 .chunks = { 3, 1, 1, 0}, 572 .chunks = { 3, 1, 1, 0}, 577 .chunks = { 1, 2, 0}, 582 .chunks = { 1, 1, 1, 0}, 587 .chunks = { 1, 1, 1, 1, 0}, 592 .chunks = { 1, 1, 1, 1, 2, 0}, 597 .chunks = { 1, 0 }, [all …]
|
| /linux/drivers/infiniband/ulp/rtrs/ |
| H A D | README | 28 session. A session is associated with a set of memory chunks reserved on the 36 chunks reserved for him on the server side. Their number, size and addresses 45 which of the memory chunks has been accessed and at which offset the message 80 the server (number of memory chunks which are going to be allocated for that 122 1. When processing a write request client selects one of the memory chunks 139 1. When processing a write request client selects one of the memory chunks 144 using the IMM field, Server invalidate rkey associated to the memory chunks 162 1. When processing a read request client selects one of the memory chunks 181 1. When processing a read request client selects one of the memory chunks 186 Server invalidate rkey associated to the memory chunks first, when it finishes,
|
| /linux/drivers/infiniband/hw/usnic/ |
| H A D | usnic_vnic.c | 44 struct usnic_vnic_res_chunk chunks[USNIC_VNIC_RES_TYPE_MAX]; member 117 for (i = 0; i < ARRAY_SIZE(vnic->chunks); i++) { in usnic_vnic_dump() 118 chunk = &vnic->chunks[i]; in usnic_vnic_dump() 222 return vnic->chunks[type].cnt; in usnic_vnic_res_cnt() 228 return vnic->chunks[type].free_cnt; in usnic_vnic_res_free_cnt() 254 src = &vnic->chunks[type]; in usnic_vnic_get_resources() 286 vnic->chunks[res->type].free_cnt++; in usnic_vnic_put_resources() 382 &vnic->chunks[res_type]); in usnic_vnic_discover_resources() 391 usnic_vnic_free_res_chunk(&vnic->chunks[res_type]); in usnic_vnic_discover_resources() 427 usnic_vnic_free_res_chunk(&vnic->chunks[res_type]); in usnic_vnic_release_resources()
|
| /linux/drivers/net/ethernet/intel/idpf/ |
| H A D | idpf_virtchnl.c | 727 const void *chunks; member 756 * idpf_send_chunked_msg - send VC message consisting of chunks 772 const void *pos = params->chunks; in idpf_send_chunked_msg() 1278 * idpf_vport_init_queue_reg_chunks - initialize queue register chunks 1280 * @schunks: source chunks to copy data from 1303 struct virtchnl2_queue_reg_chunk *schunk = &schunks->chunks[i]; in idpf_vport_init_queue_reg_chunks() 1317 * @adapter: adapter structure to get the vector chunks 1325 struct virtchnl2_vector_chunks *chunks; in idpf_get_reg_intr_vecs() local 1330 chunks = &adapter->req_vec_chunks->vchunks; in idpf_get_reg_intr_vecs() 1331 num_vchunks = le16_to_cpu(chunks->num_vchunks); in idpf_get_reg_intr_vecs() [all …]
|
| H A D | virtchnl2.h | 590 * struct virtchnl2_queue_reg_chunks - Specify several chunks of contiguous 592 * @num_chunks: Number of chunks. 594 * @chunks: Chunks of queue info. 599 struct virtchnl2_queue_reg_chunk chunks[] __counted_by_le(num_chunks); 655 * @chunks: Chunks of contiguous queues. 660 * necessary fields followed by chunks which in turn will have an array of 692 struct virtchnl2_queue_reg_chunks chunks; member 867 * @chunks: Chunks of contiguous queues. 884 struct virtchnl2_queue_reg_chunks chunks; member 928 * struct virtchnl2_vector_chunks - chunks of contiguous interrupt vectors. [all …]
|
| /linux/tools/testing/selftests/drivers/net/mlxsw/spectrum/ |
| H A D | devlink_lib_spectrum.sh | 13 KVDL_CHILDREN="singles chunks large_chunks" 90 devlink_resource_size_set 32000 kvd linear chunks 99 devlink_resource_size_set 32000 kvd linear chunks 108 devlink_resource_size_set 49152 kvd linear chunks
|
| /linux/net/xdp/ |
| H A D | xdp_umem.c | 164 u64 chunks, npgs; in xdp_umem_reg() local 199 chunks = div_u64_rem(size, chunk_size, &chunks_rem); in xdp_umem_reg() 200 if (!chunks || chunks > U32_MAX) in xdp_umem_reg() 218 umem->chunks = chunks; in xdp_umem_reg()
|
| /linux/kernel/ |
| H A D | audit_tree.c | 17 struct list_head chunks; member 68 * tree.chunks anchors chunk.owners[].list hash_lock 103 INIT_LIST_HEAD(&tree->chunks); in alloc_tree() 437 list_add(&chunk->owners[0].list, &tree->chunks); in create_chunk() 509 list_add(&p->list, &tree->chunks); in tag_chunk() 567 * Remove tree from chunks. If 'tagged' is set, remove tree only from tagged 568 * chunks. The function expects tagged chunks are all at the beginning of the 569 * chunks list. 574 while (!list_empty(&victim->chunks)) { in prune_tree_chunks() 579 p = list_first_entry(&victim->chunks, struct audit_node, list); in prune_tree_chunks() [all …]
|
| /linux/fs/ocfs2/ |
| H A D | quota.h | 40 struct list_head rc_list; /* List of chunks */ 46 struct list_head r_list[OCFS2_MAXQUOTAS]; /* List of chunks to recover */ 53 unsigned int dqi_chunks; /* Number of chunks in local quota file */ 56 struct list_head dqi_chunk; /* List of chunks */ 78 struct list_head qc_chunk; /* List of quotafile chunks */
|
| /linux/drivers/md/ |
| H A D | md-llbitmap.c | 268 /* total number of chunks */ 269 unsigned long chunks; member 542 unsigned int nr_pages = DIV_ROUND_UP(llbitmap->chunks + in llbitmap_cache_pages() 595 for (i = 0; i < llbitmap->chunks; i++) in llbitmap_init_state() 780 unsigned long chunks = DIV_ROUND_UP(blocks, chunksize); in llbitmap_init() local 784 while (chunks > space) { in llbitmap_init() 786 chunks = DIV_ROUND_UP_SECTOR_T(blocks, chunksize); in llbitmap_init() 792 llbitmap->chunks = chunks; in llbitmap_init() 799 llbitmap_state_machine(llbitmap, 0, llbitmap->chunks - 1, in llbitmap_init() 897 llbitmap->chunks = DIV_ROUND_UP_SECTOR_T(mddev->resync_max_sectors, chunksize); in llbitmap_read_sb() [all …]
|
| H A D | dm-snap-persistent.c | 35 * requires that we copy exception chunks to chunk aligned areas 47 * followed by as many exception chunks as can fit in the 114 * whole chunks worth of metadata in memory at once. 139 * When creating exceptions, all the chunks here and above are 142 * the exception store because chunks can be committed out of 146 * chunks here and above are free. It holds the value it would 147 * have held if all chunks had been committed in order of 472 * Keep track of the start of the free chunks. in insert_exceptions() 510 * Keeping reading chunks and inserting exceptions until in read_exceptions() 585 * Then there are (ps->current_area + 1) metadata chunks, each one in persistent_usage() [all …]
|
| /linux/scripts/gdb/linux/ |
| H A D | timerlist.py | 162 chunks = [] 168 chunks.append(buf[start:end]) 170 chunks.append(',') 174 chunks[0] = chunks[0][0] # Cut off the first 0 176 return "".join(str(chunks))
|
| /linux/drivers/comedi/drivers/ni_routing/tools/ |
| H A D | convert_csv_to_c.py | 228 chunks = [ self.output_file_top, 244 chunks.append('\t&{},'.format(dev_table_name)) 273 chunks.append('\tNULL,') # terminate list 274 chunks.append('};') 275 return '\n'.join(chunks) 416 chunks = [ self.output_file_top, 432 chunks.append('\t&{},'.format(fam_table_name)) 462 chunks.append('\tNULL,') # terminate list 463 chunks.append('};') 464 return '\n'.join(chunks)
|
| /linux/drivers/infiniband/hw/efa/ |
| H A D | efa_verbs.c | 116 struct pbl_chunk *chunks; member 1373 /* allocate a chunk list that consists of 4KB chunks */ in pbl_chunk_list_create() 1377 chunk_list->chunks = kzalloc_objs(*chunk_list->chunks, chunk_list_size); in pbl_chunk_list_create() 1378 if (!chunk_list->chunks) in pbl_chunk_list_create() 1387 chunk_list->chunks[i].buf = kzalloc(EFA_CHUNK_SIZE, GFP_KERNEL); in pbl_chunk_list_create() 1388 if (!chunk_list->chunks[i].buf) in pbl_chunk_list_create() 1391 chunk_list->chunks[i].length = EFA_CHUNK_USED_SIZE; in pbl_chunk_list_create() 1393 chunk_list->chunks[chunk_list_size - 1].length = in pbl_chunk_list_create() 1397 /* fill the dma addresses of sg list pages to chunks: */ in pbl_chunk_list_create() 1400 cur_chunk_buf = chunk_list->chunks[0].buf; in pbl_chunk_list_create() [all …]
|
| /linux/arch/xtensa/lib/ |
| H A D | checksum.S | 51 srli a5, a3, 5 /* 32-byte chunks */ 81 extui a5, a3, 2, 3 /* remaining 4-byte chunks */ 131 srli a5, a3, 2 /* 4-byte chunks */ 215 srli a10, a4, 5 /* 32-byte chunks */ 254 extui a10, a4, 2, 3 /* remaining 4-byte chunks */ 255 extui a4, a4, 0, 2 /* reset len for general-case, 2-byte chunks */ 286 srli a10, a4, 1 /* 2-byte chunks */
|
| /linux/arch/x86/kernel/cpu/resctrl/ |
| H A D | monitor.c | 214 u64 shift = 64 - width, chunks; in mbm_overflow_count() local 216 chunks = (cur_msr << shift) - (prev_msr << shift); in mbm_overflow_count() 217 return chunks >> shift; in mbm_overflow_count() 226 u64 chunks; in get_corrected_val() local 230 am->chunks += mbm_overflow_count(am->prev_msr, msr_val, in get_corrected_val() 232 chunks = get_corrected_mbm_count(rmid, am->chunks); in get_corrected_val() 235 chunks = msr_val; in get_corrected_val() 238 return chunks * hw_res->mon_scale; in get_corrected_val()
|
| /linux/drivers/parisc/ |
| H A D | iommu-helpers.h | 5 * iommu_fill_pdir - Insert coalesced scatter/gather chunks into the I/O Pdir. 7 * @startsg: The scatter/gather list of coalesced chunks. 11 * This function inserts the coalesced scatter/gather list chunks into the 93 ** Returns the number of DMA chunks.
|