Home
last modified time | relevance | path

Searched refs:blocks (Results 1 – 25 of 624) sorted by relevance

12345678910>>...25

/linux/Documentation/filesystems/ext4/
H A Dblockmap.rst6 | 0 to 11 | Direct map to file blocks 0 to 11. …
8 … | Indirect block: (file blocks 12 to (``$block_size`` / 4) + 11, or 12 to 1035 i…
13 … (``$block_size`` / 4) | Direct map to (``$block_size`` / 4) blocks (1024 if 4KiB blocks) | …
16 …t block: (file blocks ``$block_size``/4 + 12 to (``$block_size`` / 4) ^ 2 + (``$block_size`` / 4) …
21 …$block_size`` / 4) | Map to (``$block_size`` / 4) indirect blocks (1024 if 4KiB blocks) …
26 … (``$block_size`` / 4) | Direct map to (``$block_size`` / 4) blocks (1024 if 4KiB blocks) | …
30blocks (``$block_size`` / 4) ^ 2 + (``$block_size`` / 4) + 12 to (``$block_size`` / 4) ^ 3 + (``$b…
35 …size`` / 4) | Map to (``$block_size`` / 4) double indirect blocks (1024 if 4KiB blocks) …
40 …$block_size`` / 4) | Map to (``$block_size`` / 4) indirect blocks (1024 if 4KiB blocks) …
45 … (``$block_size`` / 4) | Direct map to (``$block_size`` / 4) blocks (1024 if 4KiB blocks) | …
/linux/crypto/async_tx/
H A Dasync_pq.c107 do_sync_gen_syndrome(struct page **blocks, unsigned int *offsets, int disks, in do_sync_gen_syndrome() argument
117 srcs = (void **) blocks; in do_sync_gen_syndrome()
120 if (blocks[i] == NULL) { in do_sync_gen_syndrome()
124 srcs[i] = page_address(blocks[i]) + offsets[i]; in do_sync_gen_syndrome()
177 async_gen_syndrome(struct page **blocks, unsigned int *offsets, int disks, in async_gen_syndrome() argument
182 &P(blocks, disks), 2, in async_gen_syndrome()
183 blocks, src_cnt, len); in async_gen_syndrome()
187 BUG_ON(disks > MAX_DISKS || !(P(blocks, disks) || Q(blocks, disks))); in async_gen_syndrome()
211 if (blocks[i] == NULL) in async_gen_syndrome()
213 unmap->addr[j] = dma_map_page(device->dev, blocks[i], in async_gen_syndrome()
[all …]
H A Dasync_raid6_recov.c154 struct page **blocks, unsigned int *offs, in __2data_recov_4() argument
168 p = blocks[disks-2]; in __2data_recov_4()
170 q = blocks[disks-1]; in __2data_recov_4()
173 a = blocks[faila]; in __2data_recov_4()
175 b = blocks[failb]; in __2data_recov_4()
204 struct page **blocks, unsigned int *offs, in __2data_recov_5() argument
222 if (blocks[i] == NULL) in __2data_recov_5()
231 p = blocks[disks-2]; in __2data_recov_5()
233 q = blocks[disks-1]; in __2data_recov_5()
235 g = blocks[good]; in __2data_recov_5()
[all …]
/linux/arch/x86/crypto/
H A Decb_cbc_helpers.h32 #define ECB_WALK_ADVANCE(blocks) do { \ argument
33 dst += (blocks) * __bsize; \
34 src += (blocks) * __bsize; \
35 nbytes -= (blocks) * __bsize; \
38 #define ECB_BLOCK(blocks, func) do { \ argument
39 const int __blocks = (blocks); \
46 ECB_WALK_ADVANCE(blocks); \
61 #define CBC_DEC_BLOCK(blocks, func) do { \ argument
62 const int __blocks = (blocks); \
68 const u8 *__iv = src + ((blocks) - 1) * __bsize; \
[all …]
/linux/arch/arm64/crypto/
H A Daes-neonbs-glue.c30 int rounds, int blocks);
32 int rounds, int blocks);
35 int rounds, int blocks, u8 iv[]);
38 int rounds, int blocks, u8 iv[]);
41 int rounds, int blocks, u8 iv[]);
43 int rounds, int blocks, u8 iv[]);
47 int rounds, int blocks);
49 int rounds, int blocks, u8 iv[]);
96 int rounds, int blocks)) in __ecb_crypt() argument
106 unsigned int blocks = walk.nbytes / AES_BLOCK_SIZE; in __ecb_crypt() local
[all …]
H A Daes-glue.c80 int rounds, int blocks);
82 int rounds, int blocks);
85 int rounds, int blocks, u8 iv[]);
87 int rounds, int blocks, u8 iv[]);
108 int rounds, int blocks, u8 iv[],
111 int rounds, int blocks, u8 iv[],
115 int blocks, u8 dg[], int enc_before,
185 unsigned int blocks; in ecb_encrypt() local
189 while ((blocks = (walk.nbytes / AES_BLOCK_SIZE))) { in ecb_encrypt()
192 ctx->key_enc, rounds, blocks); in ecb_encrypt()
[all …]
/linux/Documentation/userspace-api/media/v4l/
H A Dvidioc-g-edid.rst60 ``start_block``, ``blocks`` and ``edid`` fields, zero the ``reserved``
62 ``start_block`` and of size ``blocks`` will be placed in the memory
64 ``blocks`` * 128 bytes large (the size of one block is 128 bytes).
66 If there are fewer blocks than specified, then the driver will set
67 ``blocks`` to the actual number of blocks. If there are no EDID blocks
70 If blocks have to be retrieved from the sink, then this call will block
73 If ``start_block`` and ``blocks`` are both set to 0 when
74 :ref:`VIDIOC_G_EDID <VIDIOC_G_EDID>` is called, then the driver will set ``blocks`` to the
75 total number of available EDID blocks and it will return 0 without
76 copying any data. This is an easy way to discover how many EDID blocks
[all …]
H A Dmetafmt-c3-isp.rst32 are defined as distinct structs or "blocks" which may be added to the data
34 populating the data member with the blocks that need to be configured by the driver, but
35 need not populate it with **all** the blocks, or indeed with any at all if there
36 are no configuration changes to make. Populated blocks **must** be consecutive
38 blocks each block-specific struct embeds
41 :c:type:`c3_isp_params_block_type`. Once the blocks have been populated
42 into the data buffer, the combined size of all populated blocks shall be set in
/linux/Documentation/admin-guide/device-mapper/
H A Dwritecache.rst27 start writeback when the number of used blocks reach this
30 stop writeback when the number of used blocks drops below
33 limit the number of blocks that are in flight during
37 when the application writes this amount of blocks without
38 issuing the FLUSH request, the blocks are automatically
58 new writes (however, writes to already cached blocks are
63 blocks drops to zero, userspace can unload the
80 2. the number of blocks
81 3. the number of free blocks
82 4. the number of blocks under writeback
[all …]
H A Dera.rst9 addition it keeps track of which blocks were written within a user
14 Use cases include tracking changed blocks for backup software, and
25 origin dev device holding data blocks that may change
55 <metadata block size> <#used metadata blocks>/<#total metadata blocks>
61 #used metadata blocks Number of metadata blocks used
62 #total metadata blocks Total number of metadata blocks
64 held metadata root The location, in blocks, of the metadata root
89 - Ascertain which blocks have been written since the snapshot was taken
91 - Invalidate those blocks in the caching software
99 that it uses a few 4k blocks for updating metadata::
H A Ddm-dust.rst10 requests on specific blocks (to emulate the behavior of a hard disk
14 "dmsetup status" displays "fail_read_on_bad_block"), reads of blocks
17 Writes of blocks in the "bad block list will result in the following:
28 messages to add arbitrary bad blocks at new locations, and the
30 configured "bad blocks" will be treated as bad, or bypassed.
86 Adding and removing bad blocks
90 enabled or disabled), bad blocks may be added or removed from the
102 These bad blocks will be stored in the "bad block list".
128 ...and writing to the bad blocks will remove the blocks from the list,
157 Counting the number of bad blocks in the bad block list
[all …]
H A Dcache.rst56 3. A small metadata device - records which blocks are in the cache,
66 The origin is divided up into blocks of a fixed size. This block size
90 blocks should remain clean.
107 dirty blocks in a cache. Useful for decommissioning a cache or when
109 blocks, in the area of the cache being removed, to be clean. If the
110 area being removed from the cache still contains dirty blocks the resize
143 system crashes all cache blocks will be assumed dirty when restarted.
168 blocks. However, we allow this bitset to have a different block size
169 from the cache blocks. This is because we need to track the discard
187 cache dev fast device holding cached data blocks
[all …]
/linux/Documentation/devicetree/bindings/sifive/
H A Dsifive-blocks-ip-versioning.txt1 DT compatible string versioning for SiFive open-source IP blocks
4 strings for open-source SiFive IP blocks. HDL for these IP blocks
7 https://github.com/sifive/sifive-blocks
14 https://github.com/sifive/sifive-blocks/blob/v1.0/src/main/scala/devices/uart/UART.scala#L43
16 Until these IP blocks (or IP integration) support version
17 auto-discovery, the maintainers of these IP blocks intend to increment
19 interface to these IP blocks changes, or when the functionality of the
20 underlying IP blocks changes in a way that software should be aware of.
25 upstream sifive-blocks commits. It is expected that most drivers will
/linux/fs/xfs/libxfs/
H A Dxfs_btree_staging.c487 uint64_t *blocks, in xfs_btree_bload_level_geometry() argument
512 *blocks = div64_u64_rem(nr_this_level, desired_npb, &dontcare); in xfs_btree_bload_level_geometry()
513 *blocks = max(1ULL, *blocks); in xfs_btree_bload_level_geometry()
523 npb = div64_u64_rem(nr_this_level, *blocks, blocks_with_extra); in xfs_btree_bload_level_geometry()
525 (*blocks)++; in xfs_btree_bload_level_geometry()
526 npb = div64_u64_rem(nr_this_level, *blocks, blocks_with_extra); in xfs_btree_bload_level_geometry()
532 *avg_per_block, desired_npb, *blocks, in xfs_btree_bload_level_geometry()
685 uint64_t blocks; in xfs_btree_bload() local
702 &avg_per_block, &blocks, &blocks_with_extra); in xfs_btree_bload()
705 for (i = 0; i < blocks; i++) { in xfs_btree_bload()
[all …]
/linux/Documentation/filesystems/
H A Dqnx6.rst19 concepts of blocks, inodes and directories.
31 The space in the device or file is split up into blocks. These are a fixed
49 are done by copying all modified blocks during that specific write request
57 If the level value is 0, up to 16 direct blocks can be addressed by each
61 addressing block holds up to blocksize / 4 bytes pointers to data blocks.
63 to 16 * 256 * 256 = 1048576 blocks that can be addressed by such a tree).
66 indirect addressing blocks or inodes.
75 information (total number of filesystem blocks) or by taking the highest
86 The inode structure contains pointers to the filesystem blocks which contain
90 size, number of blocks used, access time, change time and modification time.
[all …]
H A Dnilfs2.rst67 blocks to be written to disk without making a
70 filesystem except for the updates on data blocks still
75 blocks. That means, it is guaranteed that no
84 block device when blocks are freed. This is useful
125 due to redundant move of in-use blocks.
193 of logs. Each log is composed of summary information blocks, payload
194 blocks, and an optional super root block (SR)::
209 | Summary | Payload blocks |SR|
212 The payload blocks are organized per file, and each file consists of
213 data blocks and B-tree node blocks::
[all …]
/linux/mm/
H A Ddmapool_test.c40 static int dmapool_test_alloc(struct dma_pool_pair *p, int blocks) in dmapool_test_alloc() argument
44 for (i = 0; i < blocks; i++) { in dmapool_test_alloc()
51 for (i = 0; i < blocks; i++) in dmapool_test_alloc()
64 int blocks = nr_blocks(parms->size); in dmapool_test_block() local
69 p = kcalloc(blocks, sizeof(*p), GFP_KERNEL); in dmapool_test_block()
82 ret = dmapool_test_alloc(p, blocks); in dmapool_test_block()
91 parms->size, parms->align, blocks, in dmapool_test_block()
/linux/lib/crypto/x86/
H A Dblake2s.h34 const size_t blocks = min_t(size_t, nblocks, in blake2s_compress() local
39 blake2s_compress_avx512(ctx, data, blocks, inc); in blake2s_compress()
41 blake2s_compress_ssse3(ctx, data, blocks, inc); in blake2s_compress()
44 data += blocks * BLAKE2S_BLOCK_SIZE; in blake2s_compress()
45 nblocks -= blocks; in blake2s_compress()
/linux/drivers/gpu/drm/amd/amdgpu/
H A Damdgpu_vram_mgr.c40 struct list_head blocks; member
291 list_for_each_entry(block, &vres->blocks, link) in amdgpu_vram_mgr_bo_visible_size()
307 list_for_each_entry_safe(rsv, temp, &mgr->reservations_pending, blocks) { in amdgpu_vram_mgr_do_reserve()
325 list_move(&rsv->blocks, &mgr->reserved_pages); in amdgpu_vram_mgr_do_reserve()
348 INIT_LIST_HEAD(&rsv->blocks); in amdgpu_vram_mgr_reserve_range()
354 list_add_tail(&rsv->blocks, &mgr->reservations_pending); in amdgpu_vram_mgr_reserve_range()
380 list_for_each_entry(rsv, &mgr->reservations_pending, blocks) { in amdgpu_vram_mgr_query_page_status()
388 list_for_each_entry(rsv, &mgr->reserved_pages, blocks) { in amdgpu_vram_mgr_query_page_status()
412 list_for_each_entry(block, &vres->blocks, link) { in amdgpu_vram_mgr_query_address_block_info()
493 INIT_LIST_HEAD(&vres->blocks); in amdgpu_vram_mgr_new()
[all …]
/linux/drivers/net/ethernet/microchip/sparx5/lan969x/
H A Dlan969x_vcap_impl.c16 .blocks = 1,
28 .blocks = 1,
40 .blocks = 1,
52 .blocks = 1,
64 .blocks = 1,
/linux/drivers/iio/buffer/
H A Dindustrialio-buffer-dma.c358 for (i = 0; i < ARRAY_SIZE(queue->fileio.blocks); i++) { in iio_dma_buffer_request_update()
359 block = queue->fileio.blocks[i]; in iio_dma_buffer_request_update()
375 for (i = 0; i < ARRAY_SIZE(queue->fileio.blocks); i++) { in iio_dma_buffer_request_update()
376 if (queue->fileio.blocks[i]) { in iio_dma_buffer_request_update()
377 block = queue->fileio.blocks[i]; in iio_dma_buffer_request_update()
395 queue->fileio.blocks[i] = block; in iio_dma_buffer_request_update()
430 for (i = 0; i < ARRAY_SIZE(queue->fileio.blocks); i++) { in iio_dma_buffer_fileio_free()
431 if (!queue->fileio.blocks[i]) in iio_dma_buffer_fileio_free()
433 queue->fileio.blocks[i]->state = IIO_BLOCK_STATE_DEAD; in iio_dma_buffer_fileio_free()
439 for (i = 0; i < ARRAY_SIZE(queue->fileio.blocks); i++) { in iio_dma_buffer_fileio_free()
[all …]
/linux/lib/crypto/arm/
H A Dblake2b.h24 const size_t blocks = min_t(size_t, nblocks, in blake2b_compress() local
28 blake2b_compress_neon(ctx, data, blocks, inc); in blake2b_compress()
30 data += blocks * BLAKE2B_BLOCK_SIZE; in blake2b_compress()
31 nblocks -= blocks; in blake2b_compress()
/linux/drivers/mmc/core/
H A Dsdio_ops.c115 unsigned addr, int incr_addr, u8 *buf, unsigned blocks, unsigned blksz) in mmc_io_rw_extended() argument
140 if (blocks == 0) in mmc_io_rw_extended()
143 cmd.arg |= 0x08000000 | blocks; /* block mode */ in mmc_io_rw_extended()
148 data.blocks = blocks ? blocks : 1; in mmc_io_rw_extended()
151 left_size = data.blksz * data.blocks; in mmc_io_rw_extended()
/linux/fs/xfs/scrub/
H A Drtsummary_repair.c39 unsigned long long blocks; in xrep_setup_rtsummary() local
59 blocks = mp->m_rsumblocks; in xrep_setup_rtsummary()
60 blocks += xfs_bmbt_calc_size(mp, blocks) * 2; in xrep_setup_rtsummary()
61 if (blocks > UINT_MAX) in xrep_setup_rtsummary()
64 rts->resblks += blocks; in xrep_setup_rtsummary()
/linux/drivers/md/dm-vdo/
H A Derrors.c83 struct error_block blocks[MAX_ERROR_BLOCKS]; member
87 .blocks = { {
106 for (block = registered_errors.blocks; in get_error_info()
107 block < registered_errors.blocks + registered_errors.count; in get_error_info()
294 for (block = registered_errors.blocks; in uds_register_error_block()
295 block < registered_errors.blocks + registered_errors.count; in uds_register_error_block()
305 registered_errors.blocks[registered_errors.count++] = new_block; in uds_register_error_block()

12345678910>>...25