| /linux/Documentation/filesystems/ext4/ |
| H A D | blockmap.rst | 6 | 0 to 11 | Direct map to file blocks 0 to 11. … 8 … | Indirect block: (file blocks 12 to (``$block_size`` / 4) + 11, or 12 to 1035 i… 13 … (``$block_size`` / 4) | Direct map to (``$block_size`` / 4) blocks (1024 if 4KiB blocks) | … 16 …t block: (file blocks ``$block_size``/4 + 12 to (``$block_size`` / 4) ^ 2 + (``$block_size`` / 4) … 21 …$block_size`` / 4) | Map to (``$block_size`` / 4) indirect blocks (1024 if 4KiB blocks) … 26 … (``$block_size`` / 4) | Direct map to (``$block_size`` / 4) blocks (1024 if 4KiB blocks) | … 30 …blocks (``$block_size`` / 4) ^ 2 + (``$block_size`` / 4) + 12 to (``$block_size`` / 4) ^ 3 + (``$b… 35 …size`` / 4) | Map to (``$block_size`` / 4) double indirect blocks (1024 if 4KiB blocks) … 40 …$block_size`` / 4) | Map to (``$block_size`` / 4) indirect blocks (1024 if 4KiB blocks) … 45 … (``$block_size`` / 4) | Direct map to (``$block_size`` / 4) blocks (1024 if 4KiB blocks) | …
|
| /linux/crypto/async_tx/ |
| H A D | async_pq.c | 107 do_sync_gen_syndrome(struct page **blocks, unsigned int *offsets, int disks, in do_sync_gen_syndrome() argument 117 srcs = (void **) blocks; in do_sync_gen_syndrome() 120 if (blocks[i] == NULL) { in do_sync_gen_syndrome() 124 srcs[i] = page_address(blocks[i]) + offsets[i]; in do_sync_gen_syndrome() 177 async_gen_syndrome(struct page **blocks, unsigned int *offsets, int disks, in async_gen_syndrome() argument 182 &P(blocks, disks), 2, in async_gen_syndrome() 183 blocks, src_cnt, len); in async_gen_syndrome() 187 BUG_ON(disks > MAX_DISKS || !(P(blocks, disks) || Q(blocks, disks))); in async_gen_syndrome() 211 if (blocks[i] == NULL) in async_gen_syndrome() 213 unmap->addr[j] = dma_map_page(device->dev, blocks[i], in async_gen_syndrome() [all …]
|
| H A D | async_raid6_recov.c | 154 struct page **blocks, unsigned int *offs, in __2data_recov_4() argument 168 p = blocks[disks-2]; in __2data_recov_4() 170 q = blocks[disks-1]; in __2data_recov_4() 173 a = blocks[faila]; in __2data_recov_4() 175 b = blocks[failb]; in __2data_recov_4() 204 struct page **blocks, unsigned int *offs, in __2data_recov_5() argument 222 if (blocks[i] == NULL) in __2data_recov_5() 231 p = blocks[disks-2]; in __2data_recov_5() 233 q = blocks[disks-1]; in __2data_recov_5() 235 g = blocks[good]; in __2data_recov_5() [all …]
|
| /linux/arch/x86/crypto/ |
| H A D | ecb_cbc_helpers.h | 32 #define ECB_WALK_ADVANCE(blocks) do { \ argument 33 dst += (blocks) * __bsize; \ 34 src += (blocks) * __bsize; \ 35 nbytes -= (blocks) * __bsize; \ 38 #define ECB_BLOCK(blocks, func) do { \ argument 39 const int __blocks = (blocks); \ 46 ECB_WALK_ADVANCE(blocks); \ 61 #define CBC_DEC_BLOCK(blocks, func) do { \ argument 62 const int __blocks = (blocks); \ 68 const u8 *__iv = src + ((blocks) - 1) * __bsize; \ [all …]
|
| /linux/arch/arm64/crypto/ |
| H A D | aes-neonbs-glue.c | 30 int rounds, int blocks); 32 int rounds, int blocks); 35 int rounds, int blocks, u8 iv[]); 38 int rounds, int blocks, u8 iv[]); 41 int rounds, int blocks, u8 iv[]); 43 int rounds, int blocks, u8 iv[]); 47 int rounds, int blocks); 49 int rounds, int blocks, u8 iv[]); 96 int rounds, int blocks)) in __ecb_crypt() argument 106 unsigned int blocks = walk.nbytes / AES_BLOCK_SIZE; in __ecb_crypt() local [all …]
|
| H A D | aes-glue.c | 80 int rounds, int blocks); 82 int rounds, int blocks); 85 int rounds, int blocks, u8 iv[]); 87 int rounds, int blocks, u8 iv[]); 108 int rounds, int blocks, u8 iv[], 111 int rounds, int blocks, u8 iv[], 115 int blocks, u8 dg[], int enc_before, 185 unsigned int blocks; in ecb_encrypt() local 189 while ((blocks = (walk.nbytes / AES_BLOCK_SIZE))) { in ecb_encrypt() 192 ctx->key_enc, rounds, blocks); in ecb_encrypt() [all …]
|
| /linux/Documentation/userspace-api/media/v4l/ |
| H A D | vidioc-g-edid.rst | 60 ``start_block``, ``blocks`` and ``edid`` fields, zero the ``reserved`` 62 ``start_block`` and of size ``blocks`` will be placed in the memory 64 ``blocks`` * 128 bytes large (the size of one block is 128 bytes). 66 If there are fewer blocks than specified, then the driver will set 67 ``blocks`` to the actual number of blocks. If there are no EDID blocks 70 If blocks have to be retrieved from the sink, then this call will block 73 If ``start_block`` and ``blocks`` are both set to 0 when 74 :ref:`VIDIOC_G_EDID <VIDIOC_G_EDID>` is called, then the driver will set ``blocks`` to the 75 total number of available EDID blocks and it will return 0 without 76 copying any data. This is an easy way to discover how many EDID blocks [all …]
|
| H A D | metafmt-c3-isp.rst | 32 are defined as distinct structs or "blocks" which may be added to the data 34 populating the data member with the blocks that need to be configured by the driver, but 35 need not populate it with **all** the blocks, or indeed with any at all if there 36 are no configuration changes to make. Populated blocks **must** be consecutive 38 blocks each block-specific struct embeds 41 :c:type:`c3_isp_params_block_type`. Once the blocks have been populated 42 into the data buffer, the combined size of all populated blocks shall be set in
|
| /linux/Documentation/admin-guide/device-mapper/ |
| H A D | writecache.rst | 27 start writeback when the number of used blocks reach this 30 stop writeback when the number of used blocks drops below 33 limit the number of blocks that are in flight during 37 when the application writes this amount of blocks without 38 issuing the FLUSH request, the blocks are automatically 58 new writes (however, writes to already cached blocks are 63 blocks drops to zero, userspace can unload the 80 2. the number of blocks 81 3. the number of free blocks 82 4. the number of blocks under writeback [all …]
|
| H A D | era.rst | 9 addition it keeps track of which blocks were written within a user 14 Use cases include tracking changed blocks for backup software, and 25 origin dev device holding data blocks that may change 55 <metadata block size> <#used metadata blocks>/<#total metadata blocks> 61 #used metadata blocks Number of metadata blocks used 62 #total metadata blocks Total number of metadata blocks 64 held metadata root The location, in blocks, of the metadata root 89 - Ascertain which blocks have been written since the snapshot was taken 91 - Invalidate those blocks in the caching software 99 that it uses a few 4k blocks for updating metadata::
|
| H A D | dm-dust.rst | 10 requests on specific blocks (to emulate the behavior of a hard disk 14 "dmsetup status" displays "fail_read_on_bad_block"), reads of blocks 17 Writes of blocks in the "bad block list will result in the following: 28 messages to add arbitrary bad blocks at new locations, and the 30 configured "bad blocks" will be treated as bad, or bypassed. 86 Adding and removing bad blocks 90 enabled or disabled), bad blocks may be added or removed from the 102 These bad blocks will be stored in the "bad block list". 128 ...and writing to the bad blocks will remove the blocks from the list, 157 Counting the number of bad blocks in the bad block list [all …]
|
| H A D | cache.rst | 56 3. A small metadata device - records which blocks are in the cache, 66 The origin is divided up into blocks of a fixed size. This block size 90 blocks should remain clean. 107 dirty blocks in a cache. Useful for decommissioning a cache or when 109 blocks, in the area of the cache being removed, to be clean. If the 110 area being removed from the cache still contains dirty blocks the resize 143 system crashes all cache blocks will be assumed dirty when restarted. 168 blocks. However, we allow this bitset to have a different block size 169 from the cache blocks. This is because we need to track the discard 187 cache dev fast device holding cached data blocks [all …]
|
| /linux/Documentation/devicetree/bindings/sifive/ |
| H A D | sifive-blocks-ip-versioning.txt | 1 DT compatible string versioning for SiFive open-source IP blocks 4 strings for open-source SiFive IP blocks. HDL for these IP blocks 7 https://github.com/sifive/sifive-blocks 14 https://github.com/sifive/sifive-blocks/blob/v1.0/src/main/scala/devices/uart/UART.scala#L43 16 Until these IP blocks (or IP integration) support version 17 auto-discovery, the maintainers of these IP blocks intend to increment 19 interface to these IP blocks changes, or when the functionality of the 20 underlying IP blocks changes in a way that software should be aware of. 25 upstream sifive-blocks commits. It is expected that most drivers will
|
| /linux/fs/xfs/libxfs/ |
| H A D | xfs_btree_staging.c | 487 uint64_t *blocks, in xfs_btree_bload_level_geometry() argument 512 *blocks = div64_u64_rem(nr_this_level, desired_npb, &dontcare); in xfs_btree_bload_level_geometry() 513 *blocks = max(1ULL, *blocks); in xfs_btree_bload_level_geometry() 523 npb = div64_u64_rem(nr_this_level, *blocks, blocks_with_extra); in xfs_btree_bload_level_geometry() 525 (*blocks)++; in xfs_btree_bload_level_geometry() 526 npb = div64_u64_rem(nr_this_level, *blocks, blocks_with_extra); in xfs_btree_bload_level_geometry() 532 *avg_per_block, desired_npb, *blocks, in xfs_btree_bload_level_geometry() 685 uint64_t blocks; in xfs_btree_bload() local 702 &avg_per_block, &blocks, &blocks_with_extra); in xfs_btree_bload() 705 for (i = 0; i < blocks; i++) { in xfs_btree_bload() [all …]
|
| /linux/Documentation/filesystems/ |
| H A D | qnx6.rst | 19 concepts of blocks, inodes and directories. 31 The space in the device or file is split up into blocks. These are a fixed 49 are done by copying all modified blocks during that specific write request 57 If the level value is 0, up to 16 direct blocks can be addressed by each 61 addressing block holds up to blocksize / 4 bytes pointers to data blocks. 63 to 16 * 256 * 256 = 1048576 blocks that can be addressed by such a tree). 66 indirect addressing blocks or inodes. 75 information (total number of filesystem blocks) or by taking the highest 86 The inode structure contains pointers to the filesystem blocks which contain 90 size, number of blocks used, access time, change time and modification time. [all …]
|
| H A D | nilfs2.rst | 67 blocks to be written to disk without making a 70 filesystem except for the updates on data blocks still 75 blocks. That means, it is guaranteed that no 84 block device when blocks are freed. This is useful 125 due to redundant move of in-use blocks. 193 of logs. Each log is composed of summary information blocks, payload 194 blocks, and an optional super root block (SR):: 209 | Summary | Payload blocks |SR| 212 The payload blocks are organized per file, and each file consists of 213 data blocks and B-tree node blocks:: [all …]
|
| /linux/mm/ |
| H A D | dmapool_test.c | 40 static int dmapool_test_alloc(struct dma_pool_pair *p, int blocks) in dmapool_test_alloc() argument 44 for (i = 0; i < blocks; i++) { in dmapool_test_alloc() 51 for (i = 0; i < blocks; i++) in dmapool_test_alloc() 64 int blocks = nr_blocks(parms->size); in dmapool_test_block() local 69 p = kcalloc(blocks, sizeof(*p), GFP_KERNEL); in dmapool_test_block() 82 ret = dmapool_test_alloc(p, blocks); in dmapool_test_block() 91 parms->size, parms->align, blocks, in dmapool_test_block()
|
| /linux/lib/crypto/x86/ |
| H A D | blake2s.h | 34 const size_t blocks = min_t(size_t, nblocks, in blake2s_compress() local 39 blake2s_compress_avx512(ctx, data, blocks, inc); in blake2s_compress() 41 blake2s_compress_ssse3(ctx, data, blocks, inc); in blake2s_compress() 44 data += blocks * BLAKE2S_BLOCK_SIZE; in blake2s_compress() 45 nblocks -= blocks; in blake2s_compress()
|
| /linux/drivers/gpu/drm/amd/amdgpu/ |
| H A D | amdgpu_vram_mgr.c | 40 struct list_head blocks; member 291 list_for_each_entry(block, &vres->blocks, link) in amdgpu_vram_mgr_bo_visible_size() 307 list_for_each_entry_safe(rsv, temp, &mgr->reservations_pending, blocks) { in amdgpu_vram_mgr_do_reserve() 325 list_move(&rsv->blocks, &mgr->reserved_pages); in amdgpu_vram_mgr_do_reserve() 348 INIT_LIST_HEAD(&rsv->blocks); in amdgpu_vram_mgr_reserve_range() 354 list_add_tail(&rsv->blocks, &mgr->reservations_pending); in amdgpu_vram_mgr_reserve_range() 380 list_for_each_entry(rsv, &mgr->reservations_pending, blocks) { in amdgpu_vram_mgr_query_page_status() 388 list_for_each_entry(rsv, &mgr->reserved_pages, blocks) { in amdgpu_vram_mgr_query_page_status() 412 list_for_each_entry(block, &vres->blocks, link) { in amdgpu_vram_mgr_query_address_block_info() 493 INIT_LIST_HEAD(&vres->blocks); in amdgpu_vram_mgr_new() [all …]
|
| /linux/drivers/net/ethernet/microchip/sparx5/lan969x/ |
| H A D | lan969x_vcap_impl.c | 16 .blocks = 1, 28 .blocks = 1, 40 .blocks = 1, 52 .blocks = 1, 64 .blocks = 1,
|
| /linux/drivers/iio/buffer/ |
| H A D | industrialio-buffer-dma.c | 358 for (i = 0; i < ARRAY_SIZE(queue->fileio.blocks); i++) { in iio_dma_buffer_request_update() 359 block = queue->fileio.blocks[i]; in iio_dma_buffer_request_update() 375 for (i = 0; i < ARRAY_SIZE(queue->fileio.blocks); i++) { in iio_dma_buffer_request_update() 376 if (queue->fileio.blocks[i]) { in iio_dma_buffer_request_update() 377 block = queue->fileio.blocks[i]; in iio_dma_buffer_request_update() 395 queue->fileio.blocks[i] = block; in iio_dma_buffer_request_update() 430 for (i = 0; i < ARRAY_SIZE(queue->fileio.blocks); i++) { in iio_dma_buffer_fileio_free() 431 if (!queue->fileio.blocks[i]) in iio_dma_buffer_fileio_free() 433 queue->fileio.blocks[i]->state = IIO_BLOCK_STATE_DEAD; in iio_dma_buffer_fileio_free() 439 for (i = 0; i < ARRAY_SIZE(queue->fileio.blocks); i++) { in iio_dma_buffer_fileio_free() [all …]
|
| /linux/lib/crypto/arm/ |
| H A D | blake2b.h | 24 const size_t blocks = min_t(size_t, nblocks, in blake2b_compress() local 28 blake2b_compress_neon(ctx, data, blocks, inc); in blake2b_compress() 30 data += blocks * BLAKE2B_BLOCK_SIZE; in blake2b_compress() 31 nblocks -= blocks; in blake2b_compress()
|
| /linux/drivers/mmc/core/ |
| H A D | sdio_ops.c | 115 unsigned addr, int incr_addr, u8 *buf, unsigned blocks, unsigned blksz) in mmc_io_rw_extended() argument 140 if (blocks == 0) in mmc_io_rw_extended() 143 cmd.arg |= 0x08000000 | blocks; /* block mode */ in mmc_io_rw_extended() 148 data.blocks = blocks ? blocks : 1; in mmc_io_rw_extended() 151 left_size = data.blksz * data.blocks; in mmc_io_rw_extended()
|
| /linux/fs/xfs/scrub/ |
| H A D | rtsummary_repair.c | 39 unsigned long long blocks; in xrep_setup_rtsummary() local 59 blocks = mp->m_rsumblocks; in xrep_setup_rtsummary() 60 blocks += xfs_bmbt_calc_size(mp, blocks) * 2; in xrep_setup_rtsummary() 61 if (blocks > UINT_MAX) in xrep_setup_rtsummary() 64 rts->resblks += blocks; in xrep_setup_rtsummary()
|
| /linux/drivers/md/dm-vdo/ |
| H A D | errors.c | 83 struct error_block blocks[MAX_ERROR_BLOCKS]; member 87 .blocks = { { 106 for (block = registered_errors.blocks; in get_error_info() 107 block < registered_errors.blocks + registered_errors.count; in get_error_info() 294 for (block = registered_errors.blocks; in uds_register_error_block() 295 block < registered_errors.blocks + registered_errors.count; in uds_register_error_block() 305 registered_errors.blocks[registered_errors.count++] = new_block; in uds_register_error_block()
|