| /linux/Documentation/networking/ | 
| H A D | oa-tc6-framework.rst | 49 each chunk. Ethernet frames are transferred over one or more data chunks.59 receive (RX) chunks. Chunks in both transmit and receive directions may
 69 In parallel, receive data chunks are received on MISO. Each receive data
 160 the MAC-PHY will be converted into multiple transmit data chunks. Each
 177                transaction. For TX data chunks, this bit shall be ’1’.
 242 received from the MAC-PHY. The SPI host should not write more data chunks
 248 chunks, the MAC-PHY interrupt is asserted to SPI host. On reception of the
 254 host will be sent as multiple receive data chunks. Each receive data
 284 RCA (Bit 28..24) - Receive Chunks Available. The RCA field indicates to
 286                    data chunks of frame data that are available for
 [all …]
 
 | 
| /linux/net/sctp/ | 
| H A D | chunk.c | 43 	INIT_LIST_HEAD(&msg->chunks); in sctp_datamsg_init() 65 	list_for_each_entry(chunk, &msg->chunks, frag_list) in sctp_datamsg_free()
 81 	list_for_each_safe(pos, temp, &msg->chunks) { in sctp_datamsg_destroy()
 140  * down any such message into smaller chunks.  Opportunistically, fragment
 141  * the chunks down to the current MTU constraints.  We may get refragmented
 182 	/* If the peer requested that we authenticate DATA chunks in sctp_datamsg_from_user()
 183 	 * we need to account for bundling of the AUTH chunks along with in sctp_datamsg_from_user()
 236 	/* Create chunks for all DATA chunks. */ in sctp_datamsg_from_user()
 281 		list_add_tail(&chunk->frag_list, &msg->chunks); in sctp_datamsg_from_user()
 [all...]
 | 
| H A D | outqueue.c | 11  * bundling and queueing of outgoing SCTP chunks.204 /* Free the outqueue structure and any related pending chunks.
 212 	/* Throw away unacknowledged chunks. */  in __sctp_outq_teardown()
 224 	/* Throw away chunks that have been gap ACKed.  */  in __sctp_outq_teardown()
 233 	/* Throw away any chunks in the retransmit queue. */  in __sctp_outq_teardown()
 242 	/* Throw away any chunks that are in the abandoned queue. */  in __sctp_outq_teardown()
 251 	/* Throw away any leftover data chunks. */  in __sctp_outq_teardown()
 260 	/* Throw away any leftover control chunks. */  in __sctp_outq_teardown()
 273 /* Free the outqueue structure and any related pending chunks.  */
 276 	/* Throw away leftover chunks. */  in sctp_outq_free()
 [all …]
 
 | 
| H A D | auth.c | 176  *    The RANDOM parameter, the CHUNKS parameter and the HMAC-ALGO187 			struct sctp_chunks_param *chunks, in sctp_auth_make_key_vector()
 198 	if (chunks) in sctp_auth_make_key_vector()
 199 		chunks_len = ntohs(chunks->param_hdr.length); in sctp_auth_make_key_vector()
 210 	if (chunks) { in sctp_auth_make_key_vector()
 211 		memcpy(new->data + offset, chunks, chunks_len); in sctp_auth_make_key_vector()
 302 	 *    The RANDOM parameter, the CHUNKS parameter and the HMAC-ALGO in sctp_auth_asoc_create_secret()
 560 	 *    chunks MUST NOT be listed in the CHUNKS parameter.  However, if in sctp_auth_asoc_get_hmac()
 561 	 *    a CHUNKS paramete in sctp_auth_asoc_get_hmac()
 186 sctp_auth_make_key_vector(struct sctp_random_param * random,struct sctp_chunks_param * chunks,struct sctp_hmac_algo_param * hmacs,gfp_t gfp) sctp_auth_make_key_vector()  argument
 [all...]
 | 
| H A D | ulpqueue.c | 102 		/* Create a temporary list to collect chunks on.  */  in sctp_ulpq_tail_data()266 /* Helper function to store chunks that need to be reassembled.  */
 412 	/* The chunks are held in the reasm queue sorted by TSN.  in sctp_ulpq_retrieve_reassembled()
 414 	 * fragmented chunks that complete a datagram.  in sctp_ulpq_retrieve_reassembled()
 509 	/* The chunks are held in the reasm queue sorted by TSN.  in sctp_ulpq_retrieve_partial()
 511 	 * sequence of fragmented chunks.  in sctp_ulpq_retrieve_partial()
 568 /* Helper function to reassemble chunks.  Hold chunks on the reasm queue that
 608 	/* The chunks are held in the reasm queue sorted by TSN.  in sctp_ulpq_retrieve_first()
 610 	 * fragmented chunks that start a datagram.  in sctp_ulpq_retrieve_first()
 753 	/* We are holding the chunks by stream, by SSN.  */  in sctp_ulpq_retrieve_ordered()
 [all …]
 
 | 
| /linux/drivers/net/ethernet/netronome/nfp/nfpcore/ | 
| H A D | nfp_nsp.c | 505 	} *chunks;  in nfp_nsp_command_buf_dma_sg()  local517 	chunks = kcalloc(nseg, sizeof(*chunks), GFP_KERNEL);  in nfp_nsp_command_buf_dma_sg()
 518 	if (!chunks)  in nfp_nsp_command_buf_dma_sg()
 526 		chunks[i].chunk = kmalloc(chunk_size,  in nfp_nsp_command_buf_dma_sg()
 528 		if (!chunks[i].chunk)  in nfp_nsp_command_buf_dma_sg()
 531 		chunks[i].len = min_t(u64, chunk_size, max_size - off);  in nfp_nsp_command_buf_dma_sg()
 536 			memcpy(chunks[i].chunk, arg->in_buf + off, coff);  in nfp_nsp_command_buf_dma_sg()
 538 		memset(chunks[i].chunk + coff, 0, chunk_size - coff);  in nfp_nsp_command_buf_dma_sg()
 540 		off += chunks[i].len;  in nfp_nsp_command_buf_dma_sg()
 548 		addr = dma_map_single(dev, chunks[i].chunk, chunks[i].len,  in nfp_nsp_command_buf_dma_sg()
 [all …]
 
 | 
| /linux/drivers/net/mctp/ | 
| H A D | mctp-serial.c | 534 	u8 chunks[MAX_CHUNKS];  member552 		KUNIT_EXPECT_EQ(test, next, params->chunks[i]);  in test_next_chunk_len()
 560 	KUNIT_FAIL_AND_ABORT(test, "Ran out of chunks");  in test_next_chunk_len()
 567 		.chunks = { 3, 1, 1, 0},
 572 		.chunks = { 3, 1, 1, 0},
 577 		.chunks = { 1, 2, 0},
 582 		.chunks = { 1, 1, 1, 0},
 587 		.chunks = { 1, 1, 1, 1, 0},
 592 		.chunks = { 1, 1, 1, 1, 2, 0},
 597 		.chunks = { 1, 0 },
 [all …]
 
 | 
| /linux/drivers/infiniband/ulp/rtrs/ | 
| H A D | README | 28 session. A session is associated with a set of memory chunks reserved on the36 chunks reserved for him on the server side. Their number, size and addresses
 45 which of the memory chunks has been accessed and at which offset the message
 80 the server (number of memory chunks which are going to be allocated for that
 122 1. When processing a write request client selects one of the memory chunks
 139 1. When processing a write request client selects one of the memory chunks
 144 using the IMM field, Server invalidate rkey associated to the memory chunks
 162 1. When processing a read request client selects one of the memory chunks
 181 1. When processing a read request client selects one of the memory chunks
 186 Server invalidate rkey associated to the memory chunks first, when it finishes,
 
 | 
| /linux/include/net/sctp/ | 
| H A D | structs.h | 366 	struct sctp_chunks_param *chunks;515 	/* Chunks waiting to be submitted to lower layer. */
 516 	struct list_head chunks;
 559 	/* This field is used by chunks that hold fragmented data.
 636 	 * spec violates the principle premis that all chunks are processed
 697 /* This structure holds lists of chunks as we are assembling for
 706 	/* This contains the payload chunks.  */
 711 	/* This is the total size of all chunks INCLUDING padding.  */
 777 		 *		chunks sent to this address is currently being
 950 	/* This is the list of transports that have chunks t
 369 struct sctp_chunks_param *chunks; global()  member
 519 struct list_head chunks; global()  member
 [all...]
 | 
| /linux/drivers/net/ethernet/intel/idpf/ | 
| H A D | idpf_virtchnl.c | 724 	const void		*chunks;  member748  * idpf_send_chunked_msg - send VC message consisting of chunks
 764 	const void *pos = params->chunks;  in idpf_send_chunked_msg()
 1270 	struct virtchnl2_vector_chunks *chunks;  in idpf_get_reg_intr_vecs()  local
 1275 	chunks = &vport->adapter->req_vec_chunks->vchunks;  in idpf_get_reg_intr_vecs()
 1276 	num_vchunks = le16_to_cpu(chunks->num_vchunks);  in idpf_get_reg_intr_vecs()
 1283 		chunk = &chunks->vchunks[j];  in idpf_get_reg_intr_vecs()
 1312  * @chunks: queue regs received over mailbox
 1320 				struct virtchnl2_queue_reg_chunks *chunks)  in idpf_vport_get_q_reg()  argument
 1322 	u16 num_chunks = le16_to_cpu(chunks->num_chunks);  in idpf_vport_get_q_reg()
 [all …]
 
 | 
| H A D | virtchnl2.h | 590  * struct virtchnl2_queue_reg_chunks - Specify several chunks of contiguous592  * @num_chunks: Number of chunks.
 594  * @chunks: Chunks of queue info.
 599 	struct virtchnl2_queue_reg_chunk chunks[] __counted_by_le(num_chunks);
 655  * @chunks: Chunks of contiguous queues.
 660  * necessary fields followed by chunks which in turn will have an array of
 692 	struct virtchnl2_queue_reg_chunks chunks;  member
 867  * @chunks: Chunks of contiguous queues.
 884 	struct virtchnl2_queue_reg_chunks chunks;  member
 928  * struct virtchnl2_vector_chunks - chunks of contiguous interrupt vectors.
 [all …]
 
 | 
| /linux/tools/testing/selftests/drivers/net/mlxsw/spectrum/ | 
| H A D | devlink_lib_spectrum.sh | 13 KVDL_CHILDREN="singles chunks large_chunks"90 		devlink_resource_size_set 32000 kvd linear chunks
 99 		devlink_resource_size_set 32000 kvd linear chunks
 108 		devlink_resource_size_set 49152 kvd linear chunks
 
 | 
| /linux/drivers/infiniband/hw/usnic/ | 
| H A D | usnic_vnic.c | 44 	struct usnic_vnic_res_chunk	chunks[USNIC_VNIC_RES_TYPE_MAX];  member117 	for (i = 0; i < ARRAY_SIZE(vnic->chunks); i++) {  in usnic_vnic_dump()
 118 		chunk = &vnic->chunks[i];  in usnic_vnic_dump()
 222 	return vnic->chunks[type].cnt;  in usnic_vnic_res_cnt()
 228 	return vnic->chunks[type].free_cnt;  in usnic_vnic_res_free_cnt()
 254 		src = &vnic->chunks[type];  in usnic_vnic_get_resources()
 286 			vnic->chunks[res->type].free_cnt++;  in usnic_vnic_put_resources()
 382 						&vnic->chunks[res_type]);  in usnic_vnic_discover_resources()
 391 		usnic_vnic_free_res_chunk(&vnic->chunks[res_type]);  in usnic_vnic_discover_resources()
 427 		usnic_vnic_free_res_chunk(&vnic->chunks[res_type]);  in usnic_vnic_release_resources()
 
 | 
| /linux/net/xdp/ | 
| H A D | xdp_umem.c | 163 	u64 chunks, npgs;  in xdp_umem_reg()  local198 	chunks = div_u64_rem(size, chunk_size, &chunks_rem);  in xdp_umem_reg()
 199 	if (!chunks || chunks > U32_MAX)  in xdp_umem_reg()
 217 	umem->chunks = chunks;  in xdp_umem_reg()
 
 | 
| /linux/fs/ocfs2/ | 
| H A D | quota.h | 40 	struct list_head rc_list;	/* List of chunks */46 	struct list_head r_list[OCFS2_MAXQUOTAS];	/* List of chunks to recover */
 53 	unsigned int dqi_chunks;	/* Number of chunks in local quota file */
 56 	struct list_head dqi_chunk;	/* List of chunks */
 78 	struct list_head qc_chunk;	/* List of quotafile chunks */
 
 | 
| /linux/kernel/ | 
| H A D | audit_tree.c | 17 	struct list_head chunks;  member68  * tree.chunks anchors chunk.owners[].list			hash_lock
 103 		INIT_LIST_HEAD(&tree->chunks);  in alloc_tree()
 437 	list_add(&chunk->owners[0].list, &tree->chunks);  in create_chunk()
 509 	list_add(&p->list, &tree->chunks);  in tag_chunk()
 567  * Remove tree from chunks. If 'tagged' is set, remove tree only from tagged
 568  * chunks. The function expects tagged chunks are all at the beginning of the
 569  * chunks list.
 574 	while (!list_empty(&victim->chunks)) {  in prune_tree_chunks()
 579 		p = list_first_entry(&victim->chunks, struct audit_node, list);  in prune_tree_chunks()
 [all …]
 
 | 
| /linux/scripts/gdb/linux/ | 
| H A D | timerlist.py | 162     chunks = []168         chunks.append(buf[start:end])
 170             chunks.append(',')
 174         chunks[0] = chunks[0][0]  # Cut off the first 0
 176     return "".join(str(chunks))
 
 | 
| /linux/drivers/comedi/drivers/ni_routing/tools/ | 
| H A D | convert_csv_to_c.py | 228     chunks = [ self.output_file_top,244       chunks.append('\t&{},'.format(dev_table_name))
 273     chunks.append('\tNULL,') # terminate list
 274     chunks.append('};')
 275     return '\n'.join(chunks)
 416     chunks = [ self.output_file_top,
 432       chunks.append('\t&{},'.format(fam_table_name))
 462     chunks.append('\tNULL,') # terminate list
 463     chunks.append('};')
 464     return '\n'.join(chunks)
 
 | 
| /linux/drivers/md/ | 
| H A D | dm-snap-persistent.c | 35  * requires that we copy exception chunks to chunk aligned areas47  * followed by as many exception chunks as can fit in the
 114 	 * whole chunks worth of metadata in memory at once.
 139 	 * When creating exceptions, all the chunks here and above are
 142 	 * the exception store because chunks can be committed out of
 146 	 * chunks here and above are free.  It holds the value it would
 147 	 * have held if all chunks had been committed in order of
 472 		 * Keep track of the start of the free chunks.  in insert_exceptions()
 510 	 * Keeping reading chunks and inserting exceptions until  in read_exceptions()
 585 	 * Then there are (ps->current_area + 1) metadata chunks, each one  in persistent_usage()
 [all …]
 
 | 
| /linux/arch/xtensa/lib/ | 
| H A D | checksum.S | 51 	srli	a5, a3, 5	/* 32-byte chunks */81 	extui	a5, a3, 2, 3	/* remaining 4-byte chunks */
 131 	srli	a5, a3, 2	/* 4-byte chunks */
 215 	srli	a10, a4, 5	/* 32-byte chunks */
 254 	extui	a10, a4, 2, 3	/* remaining 4-byte chunks */
 255 	extui	a4, a4, 0, 2	/* reset len for general-case, 2-byte chunks */
 286 	srli	a10, a4, 1	/* 2-byte chunks */
 
 | 
| /linux/drivers/parisc/ | 
| H A D | iommu-helpers.h | 5  * iommu_fill_pdir - Insert coalesced scatter/gather chunks into the I/O Pdir.7  * @startsg: The scatter/gather list of coalesced chunks.
 11  * This function inserts the coalesced scatter/gather list chunks into the
 93 ** Returns the number of DMA chunks.
 
 | 
| /linux/arch/x86/kernel/cpu/resctrl/ | 
| H A D | monitor.c | 214 	u64 shift = 64 - width, chunks;  in mbm_overflow_count()  local216 	chunks = (cur_msr << shift) - (prev_msr << shift);  in mbm_overflow_count()
 217 	return chunks >> shift;  in mbm_overflow_count()
 226 	u64 chunks;  in get_corrected_val()  local
 230 		am->chunks += mbm_overflow_count(am->prev_msr, msr_val,  in get_corrected_val()
 232 		chunks = get_corrected_mbm_count(rmid, am->chunks);  in get_corrected_val()
 235 		chunks = msr_val;  in get_corrected_val()
 238 	return chunks * hw_res->mon_scale;  in get_corrected_val()
 
 | 
| /linux/Documentation/admin-guide/device-mapper/ | 
| H A D | snapshot.rst | 15 In the first two cases, dm copies only the chunks of data that get37 A snapshot of the <origin> block device is created. Changed chunks of
 61    maps to entire chunks to will zero the corresponding exception(s) in
 81 Creates a merging snapshot that takes control of the changed chunks
 83 procedure, and merges these chunks back into the <origin>.  Once merging
 
 | 
| /linux/drivers/dma/sh/ | 
| H A D | shdma-base.c | 85 	/* Mark all chunks of this descriptor as submitted, move to the queue */  in shdma_tx_submit()88 		 * All chunks are on the global ld_free, so, we have to find  in shdma_tx_submit()
 97 		if (chunk->chunks == 1) {  in shdma_tx_submit()
 137 				/* Remove chunks from the queue and mark them as idle */  in shdma_tx_submit()
 360 		 * any chunks in a (partially) completed chain  in __ld_cleanup()
 369 		if (desc->mark == DESC_COMPLETED && desc->chunks == 1) {  in __ld_cleanup()
 385 			BUG_ON(desc->chunks != 1);  in __ld_cleanup()
 580 	int chunks = 0;  in shdma_prep_sg()  local
 585 		chunks += DIV_ROUND_UP(sg_dma_len(sg), schan->max_xfer_len);  in shdma_prep_sg()
 595 	 * if more than one chunk is needed further chunks have cookie = -EINVAL  in shdma_prep_sg()
 [all …]
 
 | 
| /linux/include/uapi/linux/ | 
| H A D | sctp.h | 822  * received only in an authenticated way.  Changes to the list of chunks954  *   number of unacked data chunks, and number of data chunks pending
 971  * 7.2.3.  Get the list of chunks the peer requires to be authenticated
 974  * This option gets a list of chunks for a specified association that
 1049 	__u64		sas_rtxchunks;   /* Retransmitted Chunks */
 1053 	__u64		sas_ouodchunks;  /* Unordered data chunks sent */
 1054 	__u64		sas_iuodchunks;  /* Unordered data chunks received */
 1055 	__u64		sas_oodchunks;	 /* Ordered data chunks sent */
 1056 	__u64		sas_iodchunks;	 /* Ordered data chunks received */
 1057 	__u64		sas_octrlchunks; /* Control chunks sent */
 [all …]
 
 |