Lines Matching +full:dma +full:- +full:controller

2 DMAengine controller documentation
8 Most of the Slave DMA controllers have the same general principles of
11 They have a given number of channels to use for the DMA transfers, and
20 DMA-eligible devices to the controller itself. Whenever the device
21 will want to start a transfer, it will assert a DMA request (DRQ) by
24 A very simple DMA controller would only take into account a single
35 is why most if not all of the DMA controllers can adjust this, using a
38 Moreover, some DMA controllers, whenever the RAM is used as a source
43 reads/writes it's allowed to do without the controller splitting the
44 transfer into smaller sub-transfers.
46 Our theoretical DMA controller would then only be able to do transfers
49 non-contiguous buffers to a contiguous buffer, which is called
50 scatter-gather.
53 scatter-gather. So we're left with two cases here: either we have a
54 quite simple DMA controller that doesn't support it, and we'll have to
55 implement it in software, or we have a more advanced DMA controller,
56 that implements in hardware scatter-gather.
59 transfer, and whenever the transfer is started, the controller will go
64 or the first item of the list to one channel of the DMA controller,
69 your hardware. Each DMA controller will require a different structure,
77 whenever you're willing to use DMA.
79 These were just the general memory-to-memory (also called mem2mem) or
80 memory-to-device (mem2dev) kind of transfers. Most devices often
84 DMA Support in Linux
87 Historically, DMA controller drivers have been implemented using the
98 documentation file in Documentation/crypto/async-tx-api.rst.
104 ------------------------------------
114 - ``channels``: should be initialized as a list using the
117 - ``src_addr_widths``:
120 - ``dst_addr_widths``:
123 - ``directions``:
127 - ``residue_granularity``:
131 - Descriptor:
136 - Segment:
139 - Burst:
142 - ``dev``: should hold the pointer to the ``struct device`` associated
146 ---------------------------
161 - DMA_MEMCPY
163 - The device is able to do memory to memory copies
165 - No matter what the overall size of the combined chunks for source and
167 transmitted. That means the number and size of the scatter-gather buffers in
170 total size of the two scatter-gather list buffers.
172 - It's usually used for copying pixel data between host memory and
173 memory-mapped GPU device memory, such as found on modern PCI video graphics
178 - DMA_XOR
180 - The device is able to perform XOR operations on memory areas
182 - Used to accelerate XOR intensive tasks, such as RAID5
184 - DMA_XOR_VAL
186 - The device is able to perform parity check using the XOR
189 - DMA_PQ
191 - The device is able to perform RAID6 P+Q computations, P being a
192 simple XOR, and Q being a Reed-Solomon algorithm.
194 - DMA_PQ_VAL
196 - The device is able to perform parity check using RAID6 P+Q
199 - DMA_MEMSET
201 - The device is able to fill memory with the provided pattern
203 - The pattern is treated as a single byte signed value.
205 - DMA_INTERRUPT
207 - The device is able to trigger a dummy transfer that will
210 - Used by the client drivers to register a callback that will be
211 called on a regular basis through the DMA controller interrupt
213 - DMA_PRIVATE
215 - The devices only supports slave transfers, and as such isn't
218 - DMA_ASYNC_TX
220 - Must not be set by the device, and will be set by the framework
223 - TODO: What is it about?
225 - DMA_SLAVE
227 - The device can handle device to memory transfers, including
228 scatter-gather transfers.
230 - While in the mem2mem case we were having two distinct types to
235 - If you want to transfer a single contiguous memory buffer,
238 - DMA_CYCLIC
240 - The device can handle cyclic transfers.
242 - A cyclic transfer is a transfer where the chunk collection will
245 - It's usually used for audio transfers, where you want to operate
248 - DMA_INTERLEAVE
250 - The device supports interleaved transfer.
252 - These transfers can transfer data from a non-contiguous buffer
253 to a non-contiguous buffer, opposed to DMA_SLAVE that can
254 transfer data from a non-contiguous data set to a continuous
257 - It's usually used for 2d content transfers, in which case you
261 - DMA_COMPLETION_NO_ORDER
263 - The device does not support in order completion.
265 - The driver should return DMA_OUT_OF_ORDER for device_tx_status if
268 - All cookie tracking and checking API should be treated as invalid if
271 - At this point, this is incompatible with polling option for dmatest.
273 - If this cap is set, the user is recommended to provide an unique
274 identifier for each descriptor sent to the DMA device in order to
277 - DMA_REPEAT
279 - The device supports repeated transfers. A repeated transfer, indicated by
284 - This feature is limited to interleaved transfers, this flag should thus not
286 the current needs of DMA clients, support for additional transfer types
289 - DMA_LOAD_EOT
291 - The device supports replacing repeated transfers at end of transfer (EOT)
294 - Support for replacing a currently running transfer at another point (such
296 based on DMA clients needs, if and when the need arises.
307 -------------------------------
308 Some data movement architecture (DMA controller and peripherals) uses metadata
309 associated with a transaction. The DMA controller role is to transfer the
311 The metadata itself is not used by the DMA engine itself, but it contains
315 descriptors. Depending on the architecture the DMA driver can implement either
319 - DESC_METADATA_CLIENT
324 From the DMA driver the following is expected for this mode:
326 - DMA_MEM_TO_DEV / DEV_MEM_TO_MEM
328 The data from the provided metadata buffer should be prepared for the DMA
329 controller to be sent alongside of the payload data. Either by copying to a
332 - DMA_DEV_TO_MEM
334 On transfer completion the DMA driver must copy the metadata to the client
336 After the transfer completion, DMA drivers must not touch the metadata
339 - DESC_METADATA_ENGINE
341 The metadata buffer is allocated/managed by the DMA driver. The client driver
346 From the DMA driver the following is expected for this mode:
348 - get_metadata_ptr()
353 - set_metadata_len()
356 to let the DMA driver know the number of valid bytes provided.
359 callback (in DMA_DEV_TO_MEM case) the DMA driver must ensure that the
363 -----------------
373 - ``device_alloc_chan_resources``
375 - ``device_free_chan_resources``
377 - These functions will be called whenever a driver will call
381 - They are in charge of allocating/freeing all the needed
384 - These functions can sleep.
386 - ``device_prep_dma_*``
388 - These functions are matching the capabilities you registered
391 - These functions all take the buffer or the scatterlist relevant
395 - These functions can be called from an interrupt context
397 - Any allocation you might do should be using the GFP_NOWAIT
401 - Drivers should try to pre-allocate any memory they might need
405 - It should return a unique instance of the
409 - This structure can be initialized using the function
412 - You'll also need to set two fields in this structure:
414 - flags:
418 - tx_submit: A pointer to a function you have to implement,
422 - In this structure the function pointer callback_result can be
430 - result: This provides the transfer result defined by
433 - residue: Provides the residue bytes of the transfer for those that
436 - ``device_prep_peripheral_dma_vec``
438 - Similar to ``device_prep_slave_sg``, but it takes a pointer to a
442 - ``device_issue_pending``
444 - Takes the first transaction descriptor in the pending queue,
448 - This function can be called in an interrupt context
450 - ``device_tx_status``
452 - Should report the bytes left to go over on the given channel
454 - Should only care about the transaction descriptor passed as
457 - The tx_state argument might be NULL
459 - Should use dma_set_residue to report it
461 - In the case of a cyclic transfer, it should only take into
464 - Should return DMA_OUT_OF_ORDER if the device does not support in order
467 - This function can be called in an interrupt context.
469 - device_config
471 - Reconfigures the channel with the configuration given as argument
473 - This command should NOT perform synchronously, or on any
476 - In this case, the function will receive a ``dma_slave_config``
480 - Even though that structure contains a direction field, this
484 - This call is mandatory for slave operations only. This should NOT be
489 - device_pause
491 - Pauses a transfer on the channel
493 - This command should operate synchronously on the channel,
496 - device_resume
498 - Resumes a transfer on the channel
500 - This command should operate synchronously on the channel,
503 - device_terminate_all
505 - Aborts all the pending and ongoing transfers on the channel
507 - For aborted transfers the complete callback should not be called
509 - Can be called from atomic context or from within a complete
513 - Termination may be asynchronous. The driver does not have to
517 - device_synchronize
519 - Must synchronize the termination of a channel to the current
522 - Must make sure that memory for previously submitted
523 descriptors is no longer accessed by the DMA controller.
525 - Must make sure that all complete callbacks for previously
529 - May sleep.
540 - Should be called at the end of an async TX transfer, and can be
543 - Makes sure that dependent operations are run before marking it
548 - it's a DMA transaction ID that will increment over time.
550 - Not really relevant any more since the introduction of ``virt-dma``
555 - A small structure that contains a DMA address and length.
559 - If clear, the descriptor cannot be reused by provider until the
563 - This can be acked by invoking async_tx_ack()
565 - If set, does not mean descriptor can be reused
569 - If set, the descriptor can be reused after being completed. It should
572 - The descriptor should be prepared for reuse by invoking
575 - ``dmaengine_desc_set_reuse()`` will succeed only when channel support
578 - As a consequence, if a device driver wants to skip the
580 because the DMA'd data wasn't used, it can resubmit the transfer right after
583 - Descriptor can be freed in few ways
585 - Clearing DMA_CTRL_REUSE by invoking
588 - Explicitly invoking ``dmaengine_desc_free()``, this can succeed only
591 - Terminating the channel
593 - DMA_PREP_CMD
595 - If set, the client driver tells DMA controller that passed data in DMA
598 - Interpretation of command data is DMA controller specific. It can be
603 - DMA_PREP_REPEAT
605 - If set, the transfer will be automatically repeated when it ends until a
611 - This flag is only supported if the channel reports the DMA_REPEAT
614 - DMA_PREP_LOAD_EOT
616 - If set, the transfer will replace the transfer currently being executed at
619 - This is the default behaviour for non-repeated transfers, specifying
620 DMA_PREP_LOAD_EOT for non-repeated transfers will thus make no difference.
622 - When using repeated transfers, DMA clients will usually need to set the
628 - This flag is only supported if the channel reports the DMA_LOAD_EOT
639 This is a rather inefficient design though, because the inter-transfer
652 - Burst: A number of consecutive read or write operations that
655 - Chunk: A contiguous collection of bursts
657 - Transfer: A collection of chunks (be it contiguous or not)