| /linux/Documentation/filesystems/ext4/ |
| H A D | blockmap.rst | 6 | 0 to 11 | Direct map to file blocks 0 to 11. … 8 … | Indirect block: (file blocks 12 to (``$block_size`` / 4) + 11, or 12 to 1035 i… 13 … (``$block_size`` / 4) | Direct map to (``$block_size`` / 4) blocks (1024 if 4KiB blocks) | … 16 …t block: (file blocks ``$block_size``/4 + 12 to (``$block_size`` / 4) ^ 2 + (``$block_size`` / 4) … 21 …$block_size`` / 4) | Map to (``$block_size`` / 4) indirect blocks (1024 if 4KiB blocks) … 26 … (``$block_size`` / 4) | Direct map to (``$block_size`` / 4) blocks (1024 if 4KiB blocks) | … 30 …blocks (``$block_size`` / 4) ^ 2 + (``$block_size`` / 4) + 12 to (``$block_size`` / 4) ^ 3 + (``$b… 35 …size`` / 4) | Map to (``$block_size`` / 4) double indirect blocks (1024 if 4KiB blocks) … 40 …$block_size`` / 4) | Map to (``$block_size`` / 4) indirect blocks (1024 if 4KiB blocks) … 45 … (``$block_size`` / 4) | Direct map to (``$block_size`` / 4) blocks (1024 if 4KiB blocks) | …
|
| /linux/crypto/async_tx/ |
| H A D | async_pq.c | 107 do_sync_gen_syndrome(struct page **blocks, unsigned int *offsets, int disks, in do_sync_gen_syndrome() argument 117 srcs = (void **) blocks; in do_sync_gen_syndrome() 120 if (blocks[i] == NULL) { in do_sync_gen_syndrome() 124 srcs[i] = page_address(blocks[i]) + offsets[i]; in do_sync_gen_syndrome() 177 async_gen_syndrome(struct page **blocks, unsigned int *offsets, int disks, in async_gen_syndrome() argument 182 &P(blocks, disks), 2, in async_gen_syndrome() 183 blocks, src_cnt, len); in async_gen_syndrome() 187 BUG_ON(disks > MAX_DISKS || !(P(blocks, disks) || Q(blocks, disks))); in async_gen_syndrome() 211 if (blocks[i] == NULL) in async_gen_syndrome() 213 unmap->addr[j] = dma_map_page(device->dev, blocks[i], in async_gen_syndrome() [all …]
|
| H A D | async_raid6_recov.c | 154 struct page **blocks, unsigned int *offs, in __2data_recov_4() argument 168 p = blocks[disks-2]; in __2data_recov_4() 170 q = blocks[disks-1]; in __2data_recov_4() 173 a = blocks[faila]; in __2data_recov_4() 175 b = blocks[failb]; in __2data_recov_4() 204 struct page **blocks, unsigned int *offs, in __2data_recov_5() argument 222 if (blocks[i] == NULL) in __2data_recov_5() 231 p = blocks[disks-2]; in __2data_recov_5() 233 q = blocks[disks-1]; in __2data_recov_5() 235 g = blocks[good]; in __2data_recov_5() [all …]
|
| /linux/drivers/mtd/ |
| H A D | rfd_ftl.c | 88 struct block *blocks; member 95 struct block *block = &part->blocks[block_no]; in build_block_map() 188 part->blocks = kzalloc_objs(struct block, part->total_blocks); in scan_header() 189 if (!part->blocks) in scan_header() 234 kfree(part->blocks); in scan_header() 276 erase->addr = part->blocks[block].offset; in erase_block() 279 part->blocks[block].state = BLOCK_ERASING; in erase_block() 280 part->blocks[block].free_sectors = 0; in erase_block() 287 part->blocks[block].state = BLOCK_FAILED; in erase_block() 288 part->blocks[block].free_sectors = 0; in erase_block() [all …]
|
| /linux/arch/arm64/crypto/ |
| H A D | aes-neonbs-glue.c | 30 int rounds, int blocks); 32 int rounds, int blocks); 35 int rounds, int blocks, u8 iv[]); 38 int rounds, int blocks, u8 iv[]); 41 int rounds, int blocks, u8 iv[]); 43 int rounds, int blocks, u8 iv[]); 47 int rounds, int blocks); 49 int rounds, int blocks, u8 iv[]); 101 int rounds, int blocks)) in __ecb_crypt() argument 111 unsigned int blocks = walk.nbytes / AES_BLOCK_SIZE; in __ecb_crypt() local [all …]
|
| /linux/arch/x86/crypto/ |
| H A D | ecb_cbc_helpers.h | 32 #define ECB_WALK_ADVANCE(blocks) do { \ argument 33 dst += (blocks) * __bsize; \ 34 src += (blocks) * __bsize; \ 35 nbytes -= (blocks) * __bsize; \ 38 #define ECB_BLOCK(blocks, func) do { \ argument 39 const int __blocks = (blocks); \ 46 ECB_WALK_ADVANCE(blocks); \ 61 #define CBC_DEC_BLOCK(blocks, func) do { \ argument 62 const int __blocks = (blocks); \ 68 const u8 *__iv = src + ((blocks) - 1) * __bsize; \ [all …]
|
| /linux/Documentation/userspace-api/media/v4l/ |
| H A D | vidioc-g-edid.rst | 60 ``start_block``, ``blocks`` and ``edid`` fields, zero the ``reserved`` 62 ``start_block`` and of size ``blocks`` will be placed in the memory 64 ``blocks`` * 128 bytes large (the size of one block is 128 bytes). 66 If there are fewer blocks than specified, then the driver will set 67 ``blocks`` to the actual number of blocks. If there are no EDID blocks 70 If blocks have to be retrieved from the sink, then this call will block 73 If ``start_block`` and ``blocks`` are both set to 0 when 74 :ref:`VIDIOC_G_EDID <VIDIOC_G_EDID>` is called, then the driver will set ``blocks`` to the 75 total number of available EDID blocks and it will return 0 without 76 copying any data. This is an easy way to discover how many EDID blocks [all …]
|
| H A D | metafmt-c3-isp.rst | 32 are defined as distinct structs or "blocks" which may be added to the data 34 populating the data member with the blocks that need to be configured by the driver, but 35 need not populate it with **all** the blocks, or indeed with any at all if there 36 are no configuration changes to make. Populated blocks **must** be consecutive 38 blocks each block-specific struct embeds 41 :c:type:`c3_isp_params_block_type`. Once the blocks have been populated 42 into the data buffer, the combined size of all populated blocks shall be set in
|
| /linux/arch/m68k/emu/ |
| H A D | nfblock.c | 40 static inline s32 nfhd_get_capacity(u32 major, u32 minor, u32 *blocks, in nfhd_get_capacity() argument 44 virt_to_phys(blocks), virt_to_phys(blocksize)); in nfhd_get_capacity() 55 u32 blocks, bsize; member 84 geo->cylinders = dev->blocks >> (6 - dev->bshift); in nfhd_getgeo() 97 static int __init nfhd_init_one(int id, u32 blocks, u32 bsize) in nfhd_init_one() argument 108 blocks, bsize); in nfhd_init_one() 120 dev->blocks = blocks; in nfhd_init_one() 136 set_capacity(dev->disk, (sector_t)blocks * (bsize / 512)); in nfhd_init_one() 155 u32 blocks, bsize; in nfhd_init() local 173 if (nfhd_get_capacity(i, 0, &blocks, &bsize)) in nfhd_init() [all …]
|
| /linux/Documentation/admin-guide/device-mapper/ |
| H A D | writecache.rst | 27 start writeback when the number of used blocks reach this 30 stop writeback when the number of used blocks drops below 33 limit the number of blocks that are in flight during 37 when the application writes this amount of blocks without 38 issuing the FLUSH request, the blocks are automatically 58 new writes (however, writes to already cached blocks are 63 blocks drops to zero, userspace can unload the 80 2. the number of blocks 81 3. the number of free blocks 82 4. the number of blocks under writeback [all …]
|
| H A D | era.rst | 9 addition it keeps track of which blocks were written within a user 14 Use cases include tracking changed blocks for backup software, and 25 origin dev device holding data blocks that may change 55 <metadata block size> <#used metadata blocks>/<#total metadata blocks> 61 #used metadata blocks Number of metadata blocks used 62 #total metadata blocks Total number of metadata blocks 64 held metadata root The location, in blocks, of the metadata root 89 - Ascertain which blocks have been written since the snapshot was taken 91 - Invalidate those blocks in the caching software 99 that it uses a few 4k blocks for updating metadata::
|
| H A D | dm-dust.rst | 10 requests on specific blocks (to emulate the behavior of a hard disk 14 "dmsetup status" displays "fail_read_on_bad_block"), reads of blocks 17 Writes of blocks in the "bad block list will result in the following: 28 messages to add arbitrary bad blocks at new locations, and the 30 configured "bad blocks" will be treated as bad, or bypassed. 86 Adding and removing bad blocks 90 enabled or disabled), bad blocks may be added or removed from the 102 These bad blocks will be stored in the "bad block list". 128 ...and writing to the bad blocks will remove the blocks from the list, 157 Counting the number of bad blocks in the bad block list [all …]
|
| H A D | cache.rst | 56 3. A small metadata device - records which blocks are in the cache, 66 The origin is divided up into blocks of a fixed size. This block size 90 blocks should remain clean. 107 dirty blocks in a cache. Useful for decommissioning a cache or when 109 blocks, in the area of the cache being removed, to be clean. If the 110 area being removed from the cache still contains dirty blocks the resize 143 system crashes all cache blocks will be assumed dirty when restarted. 168 blocks. However, we allow this bitset to have a different block size 169 from the cache blocks. This is because we need to track the discard 187 cache dev fast device holding cached data blocks [all …]
|
| /linux/Documentation/devicetree/bindings/sifive/ |
| H A D | sifive-blocks-ip-versioning.txt | 1 DT compatible string versioning for SiFive open-source IP blocks 4 strings for open-source SiFive IP blocks. HDL for these IP blocks 7 https://github.com/sifive/sifive-blocks 14 https://github.com/sifive/sifive-blocks/blob/v1.0/src/main/scala/devices/uart/UART.scala#L43 16 Until these IP blocks (or IP integration) support version 17 auto-discovery, the maintainers of these IP blocks intend to increment 19 interface to these IP blocks changes, or when the functionality of the 20 underlying IP blocks changes in a way that software should be aware of. 25 upstream sifive-blocks commits. It is expected that most drivers will
|
| /linux/Documentation/filesystems/ |
| H A D | qnx6.rst | 19 concepts of blocks, inodes and directories. 31 The space in the device or file is split up into blocks. These are a fixed 49 are done by copying all modified blocks during that specific write request 57 If the level value is 0, up to 16 direct blocks can be addressed by each 61 addressing block holds up to blocksize / 4 bytes pointers to data blocks. 63 to 16 * 256 * 256 = 1048576 blocks that can be addressed by such a tree). 66 indirect addressing blocks or inodes. 75 information (total number of filesystem blocks) or by taking the highest 86 The inode structure contains pointers to the filesystem blocks which contain 90 size, number of blocks used, access time, change time and modification time. [all …]
|
| H A D | nilfs2.rst | 67 blocks to be written to disk without making a 70 filesystem except for the updates on data blocks still 75 blocks. That means, it is guaranteed that no 84 block device when blocks are freed. This is useful 125 due to redundant move of in-use blocks. 193 of logs. Each log is composed of summary information blocks, payload 194 blocks, and an optional super root block (SR):: 209 | Summary | Payload blocks |SR| 212 The payload blocks are organized per file, and each file consists of 213 data blocks and B-tree node blocks:: [all …]
|
| /linux/mm/ |
| H A D | dmapool_test.c | 41 static int dmapool_test_alloc(struct dma_pool_pair *p, int blocks) in dmapool_test_alloc() argument 45 for (i = 0; i < blocks; i++) { in dmapool_test_alloc() 52 for (i = 0; i < blocks; i++) in dmapool_test_alloc() 65 int blocks = nr_blocks(parms->size); in dmapool_test_block() local 70 p = kzalloc_objs(*p, blocks); in dmapool_test_block() 83 ret = dmapool_test_alloc(p, blocks); in dmapool_test_block() 92 parms->size, parms->align, blocks, in dmapool_test_block()
|
| /linux/drivers/gpu/drm/amd/amdgpu/ |
| H A D | amdgpu_vram_mgr.c | 40 struct list_head blocks; member 291 list_for_each_entry(block, &vres->blocks, link) in amdgpu_vram_mgr_bo_visible_size() 307 list_for_each_entry_safe(rsv, temp, &mgr->reservations_pending, blocks) { in amdgpu_vram_mgr_do_reserve() 325 list_move(&rsv->blocks, &mgr->reserved_pages); in amdgpu_vram_mgr_do_reserve() 348 INIT_LIST_HEAD(&rsv->blocks); in amdgpu_vram_mgr_reserve_range() 354 list_add_tail(&rsv->blocks, &mgr->reservations_pending); in amdgpu_vram_mgr_reserve_range() 380 list_for_each_entry(rsv, &mgr->reservations_pending, blocks) { in amdgpu_vram_mgr_query_page_status() 388 list_for_each_entry(rsv, &mgr->reserved_pages, blocks) { in amdgpu_vram_mgr_query_page_status() 412 list_for_each_entry(block, &vres->blocks, link) { in amdgpu_vram_mgr_query_address_block_info() 493 INIT_LIST_HEAD(&vres->blocks); in amdgpu_vram_mgr_new() [all …]
|
| /linux/lib/crypto/x86/ |
| H A D | blake2s.h | 34 const size_t blocks = min_t(size_t, nblocks, in blake2s_compress() local 39 blake2s_compress_avx512(ctx, data, blocks, inc); in blake2s_compress() 41 blake2s_compress_ssse3(ctx, data, blocks, inc); in blake2s_compress() 44 data += blocks * BLAKE2S_BLOCK_SIZE; in blake2s_compress() 45 nblocks -= blocks; in blake2s_compress()
|
| /linux/drivers/iio/buffer/ |
| H A D | industrialio-buffer-dma.c | 351 for (i = 0; i < ARRAY_SIZE(queue->fileio.blocks); i++) { in iio_dma_buffer_request_update() 352 block = queue->fileio.blocks[i]; in iio_dma_buffer_request_update() 368 for (i = 0; i < ARRAY_SIZE(queue->fileio.blocks); i++) { in iio_dma_buffer_request_update() 369 if (queue->fileio.blocks[i]) { in iio_dma_buffer_request_update() 370 block = queue->fileio.blocks[i]; in iio_dma_buffer_request_update() 387 queue->fileio.blocks[i] = block; in iio_dma_buffer_request_update() 419 for (i = 0; i < ARRAY_SIZE(queue->fileio.blocks); i++) { in iio_dma_buffer_fileio_free() 420 if (!queue->fileio.blocks[i]) in iio_dma_buffer_fileio_free() 422 queue->fileio.blocks[i]->state = IIO_BLOCK_STATE_DEAD; in iio_dma_buffer_fileio_free() 428 for (i = 0; i < ARRAY_SIZE(queue->fileio.blocks); i++) { in iio_dma_buffer_fileio_free() [all …]
|
| /linux/drivers/net/ethernet/microchip/sparx5/lan969x/ |
| H A D | lan969x_vcap_impl.c | 16 .blocks = 1, 28 .blocks = 1, 40 .blocks = 1, 52 .blocks = 1, 64 .blocks = 1,
|
| /linux/lib/crypto/arm/ |
| H A D | blake2b.h | 24 const size_t blocks = min_t(size_t, nblocks, in blake2b_compress() local 28 blake2b_compress_neon(ctx, data, blocks, inc); in blake2b_compress() 30 data += blocks * BLAKE2B_BLOCK_SIZE; in blake2b_compress() 31 nblocks -= blocks; in blake2b_compress()
|
| /linux/drivers/mmc/core/ |
| H A D | sdio_ops.c | 115 unsigned addr, int incr_addr, u8 *buf, unsigned blocks, unsigned blksz) in mmc_io_rw_extended() argument 140 if (blocks == 0) in mmc_io_rw_extended() 143 cmd.arg |= 0x08000000 | blocks; /* block mode */ in mmc_io_rw_extended() 148 data.blocks = blocks ? blocks : 1; in mmc_io_rw_extended() 151 left_size = data.blksz * data.blocks; in mmc_io_rw_extended()
|
| /linux/drivers/usb/storage/ |
| H A D | jumpshot.c | 478 unsigned long block, blocks; in jumpshot_transport() local 533 blocks = ((u32)(srb->cmnd[7]) << 8) | ((u32)(srb->cmnd[8])); in jumpshot_transport() 536 block, blocks); in jumpshot_transport() 537 return jumpshot_read_data(us, info, block, blocks); in jumpshot_transport() 546 blocks = ((u32)(srb->cmnd[6]) << 24) | ((u32)(srb->cmnd[7]) << 16) | in jumpshot_transport() 550 block, blocks); in jumpshot_transport() 551 return jumpshot_read_data(us, info, block, blocks); in jumpshot_transport() 558 blocks = ((u32)(srb->cmnd[7]) << 8) | ((u32)(srb->cmnd[8])); in jumpshot_transport() 561 block, blocks); in jumpshot_transport() 562 return jumpshot_write_data(us, info, block, blocks); in jumpshot_transport() [all …]
|
| /linux/drivers/md/dm-vdo/ |
| H A D | errors.c | 83 struct error_block blocks[MAX_ERROR_BLOCKS]; member 87 .blocks = { { 106 for (block = registered_errors.blocks; in get_error_info() 107 block < registered_errors.blocks + registered_errors.count; in get_error_info() 294 for (block = registered_errors.blocks; in uds_register_error_block() 295 block < registered_errors.blocks + registered_errors.count; in uds_register_error_block() 305 registered_errors.blocks[registered_errors.count++] = new_block; in uds_register_error_block()
|