| /linux/fs/netfs/ |
| H A D | rolling_buffer.c | 62 int rolling_buffer_init(struct rolling_buffer *roll, unsigned int rreq_id, in rolling_buffer_init() argument 71 roll->head = fq; in rolling_buffer_init() 72 roll->tail = fq; in rolling_buffer_init() 73 iov_iter_folio_queue(&roll->iter, direction, fq, 0, 0, 0); in rolling_buffer_init() 80 int rolling_buffer_make_space(struct rolling_buffer *roll) in rolling_buffer_make_space() argument 82 struct folio_queue *fq, *head = roll->head; in rolling_buffer_make_space() 92 roll->head = fq; in rolling_buffer_make_space() 97 if (roll->iter.folioq == head && in rolling_buffer_make_space() 98 roll->iter.folioq_slot == folioq_nr_slots(head)) { in rolling_buffer_make_space() 99 roll->iter.folioq = fq; in rolling_buffer_make_space() [all …]
|
| /linux/include/linux/ |
| H A D | rolling_buffer.h | 45 int rolling_buffer_init(struct rolling_buffer *roll, unsigned int rreq_id, 47 int rolling_buffer_make_space(struct rolling_buffer *roll); 48 ssize_t rolling_buffer_load_from_ra(struct rolling_buffer *roll, 51 ssize_t rolling_buffer_append(struct rolling_buffer *roll, struct folio *folio, 53 struct folio_queue *rolling_buffer_delete_spent(struct rolling_buffer *roll); 54 void rolling_buffer_clear(struct rolling_buffer *roll); 56 static inline void rolling_buffer_advance(struct rolling_buffer *roll, size_t amount) in rolling_buffer_advance() argument 58 iov_iter_advance(&roll->iter, amount); in rolling_buffer_advance()
|
| H A D | bitops.h | 102 * @shift: bits to roll 112 * @shift: bits to roll 122 * @shift: bits to roll 132 * @shift: bits to roll 142 * @shift: bits to roll 152 * @shift: bits to roll 162 * @shift: bits to roll 172 * @shift: bits to roll
|
| /linux/sound/soc/codecs/ |
| H A D | ak4458.c | 101 * 0, 0, 0 : Sharp Roll-Off Filter 102 * 0, 0, 1 : Slow Roll-Off Filter 103 * 0, 1, 0 : Short delay Sharp Roll-Off Filter 104 * 0, 1, 1 : Short delay Slow Roll-Off Filter 105 * 1, *, * : Super Slow Roll-Off Filter 108 "Sharp Roll-Off Filter", 109 "Slow Roll-Off Filter", 110 "Short delay Sharp Roll-Off Filter", 111 "Short delay Slow Roll-Off Filter", 112 "Super Slow Roll-Off Filter"
|
| H A D | ak4619.c | 199 "Sharp Roll-Off Filter", 200 "Slow Roll-Off Filter", 201 "Short Delay Sharp Roll-Off Filter", 202 "Short Delay Slow Roll-Off Filter", 219 "Sharp Roll-Off Filter", 220 "Slow Roll-Off Filter", 221 "Short Delay Sharp Roll-Off Filter", 222 "Short Delay Slow Roll-Off Filter"
|
| H A D | ak5558.c | 77 "Sharp Roll-Off", "Slow Roll-Off", 78 "Short Delay Sharp Roll-Off", "Short Delay Slow Roll-Off",
|
| /linux/fs/xfs/libxfs/ |
| H A D | xfs_defer.c | 40 * we can roll the transaction to adhere to AG locking order rules and 51 * roll a transaction to facilitate this, but using this facility 72 * > Roll the transaction. 80 * - Roll the deferred-op transaction as above. 92 * work items every time we roll the transaction, and that we must log 107 * a new log intent item with the unfinished work items, roll the 458 /* Roll a transaction so we can do some deferred op processing. */ 473 * Roll the transaction. Rolling always given a new transaction (even in xfs_defer_trans_roll() 528 * The caller should provide a fresh transaction and roll it after we're done. 647 * any work items that wandered in since the last transaction roll (if [all …]
|
| H A D | xfs_attr.h | 434 * calling function to roll the transaction, and then call the subroutine to 509 /* Used in xfs_attr_node_removename to roll through removing blocks */ 528 /* Used in xfs_attr_rmtval_set_blk to roll through allocating blocks */
|
| /linux/arch/x86/crypto/ |
| H A D | des3_ede-asm_64.S | |
| /linux/drivers/gpu/drm/omapdrm/ |
| H A D | omap_gem.c | 40 /** roll applied when mapping to DMM */ 41 u32 roll; member 634 int omap_gem_roll(struct drm_gem_object *obj, u32 roll) in omap_gem_roll() argument 640 if (roll > npages) { in omap_gem_roll() 641 dev_err(obj->dev->dev, "invalid roll: %d\n", roll); in omap_gem_roll() 645 omap_obj->roll = roll; in omap_gem_roll() 656 roll, true); in omap_gem_roll() 767 ret = tiler_pin(block, omap_obj->pages, npages, omap_obj->roll, true); in omap_gem_pin_tiler() 1114 omap_obj->roll, true); in omap_gem_resume() 1146 omap_obj->vaddr, omap_obj->roll); in omap_gem_describe()
|
| H A D | omap_dmm_tiler.c | 347 struct page **pages, u32 npages, u32 roll) in dmm_txn_append() argument 378 int n = i + roll; in dmm_txn_append() 462 u32 npages, u32 roll, bool wait) in fill() argument 493 dmm_txn_append(txn, &p_area, pages, npages, roll); in fill() 495 roll += tcm_sizeof(slice); in fill() 510 u32 npages, u32 roll, bool wait) in tiler_pin() argument 514 ret = fill(&block->area, pages, npages, roll, wait); in tiler_pin()
|
| H A D | omap_fbdev.c | 36 /* for deferred dmm roll when getting called in atomic ctx */ 50 /* DMM roll shifts in 4K pages: */ in pan_worker()
|
| H A D | omap_gem.h | 68 int omap_gem_roll(struct drm_gem_object *obj, u32 roll);
|
| /linux/tools/testing/selftests/bpf/prog_tests/ |
| H A D | verifier_log.c | 135 "log_%s_%d", mode ? "roll" : "fixed", i); in verif_log_subtest() 140 "log_%s_prog_load_%d", mode ? "roll" : "fixed", i); in verif_log_subtest() 148 "log_%s_strlen_%d", mode ? "roll" : "fixed", i); in verif_log_subtest() 157 "log_%s_contents_%d", mode ? "roll" : "fixed", i); in verif_log_subtest() 167 "log_%s_unused_%d", mode ? "roll" : "fixed", i); in verif_log_subtest()
|
| /linux/drivers/net/ethernet/aquantia/atlantic/macsec/ |
| H A D | macsec_struct.h | 209 * Note: if specified as 2'b11, hardware AN roll over is not 239 * automatically updated by HW, if AN roll over is enabled. 244 * When the Enable AN roll over is set, S/W does not need to 245 * program the new SA's and the H/W will automatically roll over 247 * For normal operation, Enable AN Roll over will be set to '0' 515 /*! 0: The AN number will not automatically roll over if Next_PN is 517 * 1: The AN number will automatically roll over if Next_PN is 519 * Rollover is valid only after expiry. Normal roll over between
|
| /linux/arch/x86/lib/ |
| H A D | checksum_32.S | 65 roll $8, %eax 126 roll $8, %eax 172 roll $8, %eax 244 roll $8, %eax
|
| /linux/fs/nilfs2/ |
| H A D | segment.h | 26 * @ri_lsegs_start: Region for roll-forwarding (start block number) 27 * @ri_lsegs_end: Region for roll-forwarding (end block number)
|
| /linux/arch/sh/kernel/cpu/sh3/ |
| H A D | entry.S | 300 ! - roll back gRB 302 ! k0 returns original sp (after roll back) 308 ! Check for roll back gRB (User and Kernel)
|
| /linux/tools/testing/selftests/net/tcp_ao/ |
| H A D | seq-ext.c | 18 /* Move them closer to roll-over */ 25 /* make them roll-over during quota, but on different segments */ in test_adjust_seqs()
|
| /linux/drivers/gpu/drm/i915/ |
| H A D | TODO.txt | 10 - Roll out dma_fence critical section annotations.
|
| /linux/drivers/net/ethernet/marvell/octeon_ep/ |
| H A D | octep_ctrl_mbox.c | 151 /* roll over and copy remaining w_sz */ in octep_write_mbox_data() 216 /* roll over and copy remaining r_sz */ in octep_read_mbox_data()
|
| /linux/Documentation/userspace-api/media/dvb/ |
| H A D | frontend-property-cable-systems.rst | 19 supports a subset of the Annex A modulation types, and a roll-off of
|
| /linux/Documentation/arch/powerpc/ |
| H A D | transactional_memory.rst | 18 guaranteed to either complete atomically or roll back and undo any partial 59 state will roll back to that at the 'tbegin', and control will continue from
|
| /linux/tools/testing/selftests/bpf/progs/ |
| H A D | verifier_spill_fill.c | 392 /* Roll one bit to force the verifier to track both branches. */\ in spill_32bit_of_64bit_fail() 432 /* Roll one bit to force the verifier to track both branches. */\ in spill_16bit_of_32bit_fail() 830 /* Roll one bit to make the register inexact. */\ in spill_64bit_of_64bit_ok() 864 /* Roll one bit to make the register inexact. */\ in spill_32bit_of_32bit_ok() 897 /* Roll one bit to make the register inexact. */\ in spill_16bit_of_16bit_ok() 930 /* Roll one bit to make the register inexact. */\ in spill_8bit_of_8bit_ok() 1058 /* Roll one bit to force the verifier to track both branches. */\ in fill_32bit_after_spill_64bit_clear_id()
|
| /linux/fs/xfs/ |
| H A D | xfs_drain.h | 37 * buffer locks will cycle during a transaction roll to get from one intent
|