Home
last modified time | relevance | path

Searched full:chunks (Results 1 – 25 of 572) sorted by relevance

12345678910>>...23

/linux/Documentation/networking/
H A Doa-tc6-framework.rst49 each chunk. Ethernet frames are transferred over one or more data chunks.
59 receive (RX) chunks. Chunks in both transmit and receive directions may
69 In parallel, receive data chunks are received on MISO. Each receive data
160 the MAC-PHY will be converted into multiple transmit data chunks. Each
177 transaction. For TX data chunks, this bit shall be ’1’.
242 received from the MAC-PHY. The SPI host should not write more data chunks
248 chunks, the MAC-PHY interrupt is asserted to SPI host. On reception of the
254 host will be sent as multiple receive data chunks. Each receive data
284 RCA (Bit 28..24) - Receive Chunks Available. The RCA field indicates to
286 data chunks of frame data that are available for
[all …]
/linux/net/sctp/
H A Dchunk.c43 INIT_LIST_HEAD(&msg->chunks); in sctp_datamsg_init()
65 list_for_each_entry(chunk, &msg->chunks, frag_list) in sctp_datamsg_free()
81 list_for_each_safe(pos, temp, &msg->chunks) { in sctp_datamsg_destroy()
140 * down any such message into smaller chunks. Opportunistically, fragment
141 * the chunks down to the current MTU constraints. We may get refragmented
182 /* If the peer requested that we authenticate DATA chunks in sctp_datamsg_from_user()
183 * we need to account for bundling of the AUTH chunks along with in sctp_datamsg_from_user()
236 /* Create chunks for all DATA chunks. */ in sctp_datamsg_from_user()
281 list_add_tail(&chunk->frag_list, &msg->chunks); in sctp_datamsg_from_user()
[all...]
H A Doutqueue.c11 * bundling and queueing of outgoing SCTP chunks.
204 /* Free the outqueue structure and any related pending chunks.
212 /* Throw away unacknowledged chunks. */ in __sctp_outq_teardown()
224 /* Throw away chunks that have been gap ACKed. */ in __sctp_outq_teardown()
233 /* Throw away any chunks in the retransmit queue. */ in __sctp_outq_teardown()
242 /* Throw away any chunks that are in the abandoned queue. */ in __sctp_outq_teardown()
251 /* Throw away any leftover data chunks. */ in __sctp_outq_teardown()
260 /* Throw away any leftover control chunks. */ in __sctp_outq_teardown()
273 /* Free the outqueue structure and any related pending chunks. */
276 /* Throw away leftover chunks. */ in sctp_outq_free()
[all …]
H A Dauth.c176 * The RANDOM parameter, the CHUNKS parameter and the HMAC-ALGO
187 struct sctp_chunks_param *chunks, in sctp_auth_make_key_vector()
198 if (chunks) in sctp_auth_make_key_vector()
199 chunks_len = ntohs(chunks->param_hdr.length); in sctp_auth_make_key_vector()
210 if (chunks) { in sctp_auth_make_key_vector()
211 memcpy(new->data + offset, chunks, chunks_len); in sctp_auth_make_key_vector()
302 * The RANDOM parameter, the CHUNKS parameter and the HMAC-ALGO in sctp_auth_asoc_create_secret()
560 * chunks MUST NOT be listed in the CHUNKS parameter. However, if in sctp_auth_asoc_get_hmac()
561 * a CHUNKS paramete in sctp_auth_asoc_get_hmac()
186 sctp_auth_make_key_vector(struct sctp_random_param * random,struct sctp_chunks_param * chunks,struct sctp_hmac_algo_param * hmacs,gfp_t gfp) sctp_auth_make_key_vector() argument
[all...]
H A Dulpqueue.c102 /* Create a temporary list to collect chunks on. */ in sctp_ulpq_tail_data()
266 /* Helper function to store chunks that need to be reassembled. */
412 /* The chunks are held in the reasm queue sorted by TSN. in sctp_ulpq_retrieve_reassembled()
414 * fragmented chunks that complete a datagram. in sctp_ulpq_retrieve_reassembled()
509 /* The chunks are held in the reasm queue sorted by TSN. in sctp_ulpq_retrieve_partial()
511 * sequence of fragmented chunks. in sctp_ulpq_retrieve_partial()
568 /* Helper function to reassemble chunks. Hold chunks on the reasm queue that
608 /* The chunks are held in the reasm queue sorted by TSN. in sctp_ulpq_retrieve_first()
610 * fragmented chunks that start a datagram. in sctp_ulpq_retrieve_first()
753 /* We are holding the chunks by stream, by SSN. */ in sctp_ulpq_retrieve_ordered()
[all …]
/linux/drivers/net/ethernet/netronome/nfp/nfpcore/
H A Dnfp_nsp.c505 } *chunks; in nfp_nsp_command_buf_dma_sg() local
517 chunks = kcalloc(nseg, sizeof(*chunks), GFP_KERNEL); in nfp_nsp_command_buf_dma_sg()
518 if (!chunks) in nfp_nsp_command_buf_dma_sg()
526 chunks[i].chunk = kmalloc(chunk_size, in nfp_nsp_command_buf_dma_sg()
528 if (!chunks[i].chunk) in nfp_nsp_command_buf_dma_sg()
531 chunks[i].len = min_t(u64, chunk_size, max_size - off); in nfp_nsp_command_buf_dma_sg()
536 memcpy(chunks[i].chunk, arg->in_buf + off, coff); in nfp_nsp_command_buf_dma_sg()
538 memset(chunks[i].chunk + coff, 0, chunk_size - coff); in nfp_nsp_command_buf_dma_sg()
540 off += chunks[i].len; in nfp_nsp_command_buf_dma_sg()
548 addr = dma_map_single(dev, chunks[i].chunk, chunks[i].len, in nfp_nsp_command_buf_dma_sg()
[all …]
/linux/drivers/net/mctp/
H A Dmctp-serial.c534 u8 chunks[MAX_CHUNKS]; member
552 KUNIT_EXPECT_EQ(test, next, params->chunks[i]); in test_next_chunk_len()
560 KUNIT_FAIL_AND_ABORT(test, "Ran out of chunks"); in test_next_chunk_len()
567 .chunks = { 3, 1, 1, 0},
572 .chunks = { 3, 1, 1, 0},
577 .chunks = { 1, 2, 0},
582 .chunks = { 1, 1, 1, 0},
587 .chunks = { 1, 1, 1, 1, 0},
592 .chunks = { 1, 1, 1, 1, 2, 0},
597 .chunks = { 1, 0 },
[all …]
/linux/drivers/infiniband/ulp/rtrs/
H A DREADME28 session. A session is associated with a set of memory chunks reserved on the
36 chunks reserved for him on the server side. Their number, size and addresses
45 which of the memory chunks has been accessed and at which offset the message
80 the server (number of memory chunks which are going to be allocated for that
122 1. When processing a write request client selects one of the memory chunks
139 1. When processing a write request client selects one of the memory chunks
144 using the IMM field, Server invalidate rkey associated to the memory chunks
162 1. When processing a read request client selects one of the memory chunks
181 1. When processing a read request client selects one of the memory chunks
186 Server invalidate rkey associated to the memory chunks first, when it finishes,
/linux/include/net/sctp/
H A Dstructs.h366 struct sctp_chunks_param *chunks;
515 /* Chunks waiting to be submitted to lower layer. */
516 struct list_head chunks;
559 /* This field is used by chunks that hold fragmented data.
636 * spec violates the principle premis that all chunks are processed
697 /* This structure holds lists of chunks as we are assembling for
706 /* This contains the payload chunks. */
711 /* This is the total size of all chunks INCLUDING padding. */
777 * chunks sent to this address is currently being
950 /* This is the list of transports that have chunks t
369 struct sctp_chunks_param *chunks; global() member
519 struct list_head chunks; global() member
[all...]
/linux/drivers/net/ethernet/intel/idpf/
H A Didpf_virtchnl.c724 const void *chunks; member
748 * idpf_send_chunked_msg - send VC message consisting of chunks
764 const void *pos = params->chunks; in idpf_send_chunked_msg()
1270 struct virtchnl2_vector_chunks *chunks; in idpf_get_reg_intr_vecs() local
1275 chunks = &vport->adapter->req_vec_chunks->vchunks; in idpf_get_reg_intr_vecs()
1276 num_vchunks = le16_to_cpu(chunks->num_vchunks); in idpf_get_reg_intr_vecs()
1283 chunk = &chunks->vchunks[j]; in idpf_get_reg_intr_vecs()
1312 * @chunks: queue regs received over mailbox
1320 struct virtchnl2_queue_reg_chunks *chunks) in idpf_vport_get_q_reg() argument
1322 u16 num_chunks = le16_to_cpu(chunks->num_chunks); in idpf_vport_get_q_reg()
[all …]
H A Dvirtchnl2.h590 * struct virtchnl2_queue_reg_chunks - Specify several chunks of contiguous
592 * @num_chunks: Number of chunks.
594 * @chunks: Chunks of queue info.
599 struct virtchnl2_queue_reg_chunk chunks[] __counted_by_le(num_chunks);
655 * @chunks: Chunks of contiguous queues.
660 * necessary fields followed by chunks which in turn will have an array of
692 struct virtchnl2_queue_reg_chunks chunks; member
867 * @chunks: Chunks of contiguous queues.
884 struct virtchnl2_queue_reg_chunks chunks; member
928 * struct virtchnl2_vector_chunks - chunks of contiguous interrupt vectors.
[all …]
/linux/tools/testing/selftests/drivers/net/mlxsw/spectrum/
H A Ddevlink_lib_spectrum.sh13 KVDL_CHILDREN="singles chunks large_chunks"
90 devlink_resource_size_set 32000 kvd linear chunks
99 devlink_resource_size_set 32000 kvd linear chunks
108 devlink_resource_size_set 49152 kvd linear chunks
/linux/drivers/infiniband/hw/usnic/
H A Dusnic_vnic.c44 struct usnic_vnic_res_chunk chunks[USNIC_VNIC_RES_TYPE_MAX]; member
117 for (i = 0; i < ARRAY_SIZE(vnic->chunks); i++) { in usnic_vnic_dump()
118 chunk = &vnic->chunks[i]; in usnic_vnic_dump()
222 return vnic->chunks[type].cnt; in usnic_vnic_res_cnt()
228 return vnic->chunks[type].free_cnt; in usnic_vnic_res_free_cnt()
254 src = &vnic->chunks[type]; in usnic_vnic_get_resources()
286 vnic->chunks[res->type].free_cnt++; in usnic_vnic_put_resources()
382 &vnic->chunks[res_type]); in usnic_vnic_discover_resources()
391 usnic_vnic_free_res_chunk(&vnic->chunks[res_type]); in usnic_vnic_discover_resources()
427 usnic_vnic_free_res_chunk(&vnic->chunks[res_type]); in usnic_vnic_release_resources()
/linux/net/xdp/
H A Dxdp_umem.c163 u64 chunks, npgs; in xdp_umem_reg() local
198 chunks = div_u64_rem(size, chunk_size, &chunks_rem); in xdp_umem_reg()
199 if (!chunks || chunks > U32_MAX) in xdp_umem_reg()
217 umem->chunks = chunks; in xdp_umem_reg()
/linux/fs/ocfs2/
H A Dquota.h40 struct list_head rc_list; /* List of chunks */
46 struct list_head r_list[OCFS2_MAXQUOTAS]; /* List of chunks to recover */
53 unsigned int dqi_chunks; /* Number of chunks in local quota file */
56 struct list_head dqi_chunk; /* List of chunks */
78 struct list_head qc_chunk; /* List of quotafile chunks */
/linux/kernel/
H A Daudit_tree.c17 struct list_head chunks; member
68 * tree.chunks anchors chunk.owners[].list hash_lock
103 INIT_LIST_HEAD(&tree->chunks); in alloc_tree()
437 list_add(&chunk->owners[0].list, &tree->chunks); in create_chunk()
509 list_add(&p->list, &tree->chunks); in tag_chunk()
567 * Remove tree from chunks. If 'tagged' is set, remove tree only from tagged
568 * chunks. The function expects tagged chunks are all at the beginning of the
569 * chunks list.
574 while (!list_empty(&victim->chunks)) { in prune_tree_chunks()
579 p = list_first_entry(&victim->chunks, struct audit_node, list); in prune_tree_chunks()
[all …]
/linux/scripts/gdb/linux/
H A Dtimerlist.py162 chunks = []
168 chunks.append(buf[start:end])
170 chunks.append(',')
174 chunks[0] = chunks[0][0] # Cut off the first 0
176 return "".join(str(chunks))
/linux/drivers/comedi/drivers/ni_routing/tools/
H A Dconvert_csv_to_c.py228 chunks = [ self.output_file_top,
244 chunks.append('\t&{},'.format(dev_table_name))
273 chunks.append('\tNULL,') # terminate list
274 chunks.append('};')
275 return '\n'.join(chunks)
416 chunks = [ self.output_file_top,
432 chunks.append('\t&{},'.format(fam_table_name))
462 chunks.append('\tNULL,') # terminate list
463 chunks.append('};')
464 return '\n'.join(chunks)
/linux/drivers/md/
H A Ddm-snap-persistent.c35 * requires that we copy exception chunks to chunk aligned areas
47 * followed by as many exception chunks as can fit in the
114 * whole chunks worth of metadata in memory at once.
139 * When creating exceptions, all the chunks here and above are
142 * the exception store because chunks can be committed out of
146 * chunks here and above are free. It holds the value it would
147 * have held if all chunks had been committed in order of
472 * Keep track of the start of the free chunks. in insert_exceptions()
510 * Keeping reading chunks and inserting exceptions until in read_exceptions()
585 * Then there are (ps->current_area + 1) metadata chunks, each one in persistent_usage()
[all …]
/linux/arch/xtensa/lib/
H A Dchecksum.S51 srli a5, a3, 5 /* 32-byte chunks */
81 extui a5, a3, 2, 3 /* remaining 4-byte chunks */
131 srli a5, a3, 2 /* 4-byte chunks */
215 srli a10, a4, 5 /* 32-byte chunks */
254 extui a10, a4, 2, 3 /* remaining 4-byte chunks */
255 extui a4, a4, 0, 2 /* reset len for general-case, 2-byte chunks */
286 srli a10, a4, 1 /* 2-byte chunks */
/linux/arch/x86/kernel/cpu/resctrl/
H A Dmonitor.c214 u64 shift = 64 - width, chunks; in mbm_overflow_count() local
216 chunks = (cur_msr << shift) - (prev_msr << shift); in mbm_overflow_count()
217 return chunks >> shift; in mbm_overflow_count()
226 u64 chunks; in get_corrected_val() local
230 am->chunks += mbm_overflow_count(am->prev_msr, msr_val, in get_corrected_val()
232 chunks = get_corrected_mbm_count(rmid, am->chunks); in get_corrected_val()
235 chunks = msr_val; in get_corrected_val()
238 return chunks * hw_res->mon_scale; in get_corrected_val()
/linux/drivers/parisc/
H A Diommu-helpers.h5 * iommu_fill_pdir - Insert coalesced scatter/gather chunks into the I/O Pdir.
7 * @startsg: The scatter/gather list of coalesced chunks.
11 * This function inserts the coalesced scatter/gather list chunks into the
93 ** Returns the number of DMA chunks.
/linux/Documentation/admin-guide/device-mapper/
H A Dsnapshot.rst15 In the first two cases, dm copies only the chunks of data that get
37 A snapshot of the <origin> block device is created. Changed chunks of
61 maps to entire chunks to will zero the corresponding exception(s) in
81 Creates a merging snapshot that takes control of the changed chunks
83 procedure, and merges these chunks back into the <origin>. Once merging
/linux/drivers/dma/sh/
H A Dshdma-base.c85 /* Mark all chunks of this descriptor as submitted, move to the queue */ in shdma_tx_submit()
88 * All chunks are on the global ld_free, so, we have to find in shdma_tx_submit()
97 if (chunk->chunks == 1) { in shdma_tx_submit()
137 /* Remove chunks from the queue and mark them as idle */ in shdma_tx_submit()
360 * any chunks in a (partially) completed chain in __ld_cleanup()
369 if (desc->mark == DESC_COMPLETED && desc->chunks == 1) { in __ld_cleanup()
385 BUG_ON(desc->chunks != 1); in __ld_cleanup()
580 int chunks = 0; in shdma_prep_sg() local
585 chunks += DIV_ROUND_UP(sg_dma_len(sg), schan->max_xfer_len); in shdma_prep_sg()
595 * if more than one chunk is needed further chunks have cookie = -EINVAL in shdma_prep_sg()
[all …]
/linux/include/uapi/linux/
H A Dsctp.h822 * received only in an authenticated way. Changes to the list of chunks
954 * number of unacked data chunks, and number of data chunks pending
971 * 7.2.3. Get the list of chunks the peer requires to be authenticated
974 * This option gets a list of chunks for a specified association that
1049 __u64 sas_rtxchunks; /* Retransmitted Chunks */
1053 __u64 sas_ouodchunks; /* Unordered data chunks sent */
1054 __u64 sas_iuodchunks; /* Unordered data chunks received */
1055 __u64 sas_oodchunks; /* Ordered data chunks sent */
1056 __u64 sas_iodchunks; /* Ordered data chunks received */
1057 __u64 sas_octrlchunks; /* Control chunks sent */
[all …]

12345678910>>...23