| /linux/drivers/net/ethernet/marvell/prestera/ |
| H A D | prestera_counter.c | 45 prestera_counter_is_ready(struct prestera_counter_block *block, u32 id) in prestera_counter_is_ready() argument 47 return block->counter_flag[id - block->offset] == COUNTER_FLAG_READY; in prestera_counter_is_ready() 60 static void prestera_counter_block_lock(struct prestera_counter_block *block) in prestera_counter_block_lock() argument 62 mutex_lock(&block->mtx); in prestera_counter_block_lock() 65 static void prestera_counter_block_unlock(struct prestera_counter_block *block) in prestera_counter_block_unlock() argument 67 mutex_unlock(&block->mtx); in prestera_counter_block_unlock() 70 static bool prestera_counter_block_incref(struct prestera_counter_block *block) in prestera_counter_block_incref() argument 72 return refcount_inc_not_zero(&block->refcnt); in prestera_counter_block_incref() 75 static bool prestera_counter_block_decref(struct prestera_counter_block *block) in prestera_counter_block_decref() argument 77 return refcount_dec_and_test(&block->refcnt); in prestera_counter_block_decref() [all …]
|
| H A D | prestera_flow.c | 16 static int prestera_flow_block_mall_cb(struct prestera_flow_block *block, in prestera_flow_block_mall_cb() argument 21 return prestera_mall_replace(block, f); in prestera_flow_block_mall_cb() 23 prestera_mall_destroy(block); in prestera_flow_block_mall_cb() 30 static int prestera_flow_block_flower_cb(struct prestera_flow_block *block, in prestera_flow_block_flower_cb() argument 35 return prestera_flower_replace(block, f); in prestera_flow_block_flower_cb() 37 prestera_flower_destroy(block, f); in prestera_flow_block_flower_cb() 40 return prestera_flower_stats(block, f); in prestera_flow_block_flower_cb() 42 return prestera_flower_tmplt_create(block, f); in prestera_flow_block_flower_cb() 44 prestera_flower_tmplt_destroy(block, f); in prestera_flow_block_flower_cb() 54 struct prestera_flow_block *block = cb_priv; in prestera_flow_block_cb() local [all …]
|
| H A D | prestera_matchall.c | 14 static int prestera_mall_prio_check(struct prestera_flow_block *block, in prestera_mall_prio_check() argument 21 err = prestera_flower_prio_get(block, f->common.chain_index, in prestera_mall_prio_check() 32 if (f->common.prio <= flower_prio_max && !block->ingress) { in prestera_mall_prio_check() 36 if (f->common.prio >= flower_prio_min && block->ingress) { in prestera_mall_prio_check() 44 int prestera_mall_prio_get(struct prestera_flow_block *block, in prestera_mall_prio_get() argument 47 if (!block->mall.bound) in prestera_mall_prio_get() 50 *prio_min = block->mall.prio_min; in prestera_mall_prio_get() 51 *prio_max = block->mall.prio_max; in prestera_mall_prio_get() 55 static void prestera_mall_prio_update(struct prestera_flow_block *block, in prestera_mall_prio_update() argument 58 block->mall.prio_min = min(block->mall.prio_min, f->common.prio); in prestera_mall_prio_update() [all …]
|
| /linux/sound/isa/gus/ |
| H A D | gus_mem.c | 19 snd_gf1_mem_xalloc(struct snd_gf1_mem *alloc, struct snd_gf1_mem_block *block, in snd_gf1_mem_xalloc() argument 27 *nblock = *block; in snd_gf1_mem_xalloc() 60 int snd_gf1_mem_xfree(struct snd_gf1_mem * alloc, struct snd_gf1_mem_block * block) in snd_gf1_mem_xfree() argument 62 if (block->share) { /* ok.. shared block */ in snd_gf1_mem_xfree() 63 block->share--; in snd_gf1_mem_xfree() 66 if (alloc->first == block) { in snd_gf1_mem_xfree() 67 alloc->first = block->next; in snd_gf1_mem_xfree() 68 if (block->next) in snd_gf1_mem_xfree() 69 block->next->prev = NULL; in snd_gf1_mem_xfree() 71 block->prev->next = block->next; in snd_gf1_mem_xfree() [all …]
|
| /linux/drivers/iio/buffer/ |
| H A D | industrialio-buffer-dma.c | 26 * For DMA buffers the storage is sub-divided into so called blocks. Each block 27 * has its own memory buffer. The size of the block is the granularity at which 29 * basic unit of data exchange from one sample to one block decreases the 32 * sample the overhead will be x for each sample. Whereas when using a block 40 * them with data. Block on the outgoing queue have been filled with data and 43 * A block can be in one of the following states: 45 * the block. 48 * * Owned by the DMA controller: The DMA controller is processing the block 53 * * Dead: A block that is dead has been marked as to be freed. It might still 56 * incoming or outgoing queue the block will be freed. [all …]
|
| /linux/drivers/mtd/ |
| H A D | nftlmount.c | 28 unsigned int block, boot_record_count = 0; in find_boot_record() local 48 for (block = 0; block < nftl->nb_blocks; block++) { in find_boot_record() 53 ret = mtd_read(mtd, block * nftl->EraseSize, SECTORSIZE, in find_boot_record() 61 printk(KERN_WARNING "Block read at 0x%x of mtd%d failed: %d\n", in find_boot_record() 62 block * nftl->EraseSize, nftl->mbd.mtd->index, ret); in find_boot_record() 64 printk(KERN_WARNING "Further failures for this block will not be printed\n"); in find_boot_record() 73 block * nftl->EraseSize, nftl->mbd.mtd->index); in find_boot_record() 79 ret = nftl_read_oob(mtd, block * nftl->EraseSize + in find_boot_record() 84 block * nftl->EraseSize, nftl->mbd.mtd->index, ret); in find_boot_record() 94 block * nftl->EraseSize, nftl->mbd.mtd->index, in find_boot_record() [all …]
|
| H A D | inftlmount.c | 35 unsigned int i, block; in find_boot_record() local 55 for (block = 0; block < inftl->nb_blocks; block++) { in find_boot_record() 62 ret = mtd_read(mtd, block * inftl->EraseSize, SECTORSIZE, in find_boot_record() 70 printk(KERN_WARNING "INFTL: block read at 0x%x " in find_boot_record() 72 block * inftl->EraseSize, in find_boot_record() 76 "failures for this block will " in find_boot_record() 89 block * inftl->EraseSize + SECTORSIZE + 8, in find_boot_record() 94 "(err %d)\n", block * inftl->EraseSize, in find_boot_record() 107 mtd_read(mtd, block * inftl->EraseSize + 4096, SECTORSIZE, in find_boot_record() 172 block >>= mh->BlockMultiplierBits; in find_boot_record() [all …]
|
| H A D | rfd_ftl.c | 28 MODULE_PARM_DESC(block_size, "Block size to use by RFD, defaults to erase unit size"); 57 struct block { struct 80 int reserved_block; /* block next up for reclaim */ argument 81 int current_block; /* block to write to */ 88 struct block *blocks; 95 struct block *block = &part->blocks[block_no]; in build_block_map() local 98 block->offset = part->block_size * block_no; in build_block_map() 101 block->state = BLOCK_UNUSED; in build_block_map() 105 block->state = BLOCK_OK; in build_block_map() 116 block->free_sectors++; in build_block_map() [all …]
|
| /linux/drivers/gpio/ |
| H A D | gpio-sch311x.c | 39 struct sch311x_gpio_block { /* one GPIO block runtime data */ 44 spinlock_t lock; /* lock for this GPIO block */ 134 struct sch311x_gpio_block *block = gpiochip_get_data(chip); in sch311x_gpio_request() local 136 if (block->config_regs[offset] == 0) /* GPIO is not available */ in sch311x_gpio_request() 139 if (!request_region(block->runtime_reg + block->config_regs[offset], in sch311x_gpio_request() 142 block->runtime_reg + block->config_regs[offset]); in sch311x_gpio_request() 150 struct sch311x_gpio_block *block = gpiochip_get_data(chip); in sch311x_gpio_free() local 152 if (block->config_regs[offset] == 0) /* GPIO is not available */ in sch311x_gpio_free() 155 release_region(block->runtime_reg + block->config_regs[offset], 1); in sch311x_gpio_free() 160 struct sch311x_gpio_block *block = gpiochip_get_data(chip); in sch311x_gpio_get() local [all …]
|
| /linux/drivers/w1/ |
| H A D | w1_netlink.c | 41 struct w1_cb_block *block; member 49 * @block: block to calculate 55 static u16 w1_reply_len(struct w1_cb_block *block) in w1_reply_len() argument 57 if (!block->cn) in w1_reply_len() 59 return (u8 *)block->cn - (u8 *)block->first_cn + block->cn->len; in w1_reply_len() 62 static void w1_unref_block(struct w1_cb_block *block) in w1_unref_block() argument 64 if (atomic_sub_return(1, &block->refcnt) == 0) { in w1_unref_block() 65 u16 len = w1_reply_len(block); in w1_unref_block() 67 cn_netlink_send_mult(block->first_cn, len, in w1_unref_block() 68 block->portid, 0, in w1_unref_block() [all …]
|
| /linux/drivers/md/dm-vdo/ |
| H A D | recovery-journal.h | 24 * The recovery_journal provides a log of all block mapping and reference count changes which have 25 * not yet been stably written to the block map or slab journals. This log helps to reduce the 26 * write amplification of writes by providing amortization of slab journal and block map page 35 * journal. The 'head' is the oldest active block in the journal. The 'tail' is the end of the 36 * half-open interval containing the active blocks. 'active' is the number of the block actively 42 * less than the on-disk size. Each in-memory block is also a vdo_completion. Each in-memory block 43 * has a vio which is used to commit that block to disk. The vio's data is the on-disk 44 * representation of the journal block. In addition each in-memory block has a buffer which is used 45 * to accumulate entries while a partial commit of the block is in progress. In-memory blocks are 46 * kept on two lists. Free blocks live on the 'free_tail_blocks' list. When a block becomes active [all …]
|
| /linux/net/sched/ |
| H A D | cls_api.c | 181 struct tcf_block *block = chain->block; in tcf_proto_signal_destroying() local 183 mutex_lock(&block->proto_destroy_lock); in tcf_proto_signal_destroying() 184 hash_add_rcu(block->proto_destroy_ht, &tp->destroy_ht_node, in tcf_proto_signal_destroying() 186 mutex_unlock(&block->proto_destroy_lock); in tcf_proto_signal_destroying() 205 hash_for_each_possible_rcu(chain->block->proto_destroy_ht, iter, in tcf_proto_exists_destroying() 220 struct tcf_block *block = chain->block; in tcf_proto_signal_destroyed() local 222 mutex_lock(&block->proto_destroy_lock); in tcf_proto_signal_destroyed() 225 mutex_unlock(&block->proto_destroy_lock); in tcf_proto_signal_destroyed() 417 struct tcf_block *block = tp->chain->block; in tcf_proto_count_usesw() local 422 if (!atomic_dec_return(&block->useswcnt)) in tcf_proto_count_usesw() [all …]
|
| /linux/sound/pci/ctxfi/ |
| H A D | ctvmem.c | 26 * Find or create vm block based on requested @size. 32 struct ct_vm_block *block, *entry; in get_vm_block() local 46 break; /* found a block that is big enough */ in get_vm_block() 58 block = kzalloc_obj(*block); in get_vm_block() 59 if (!block) in get_vm_block() 62 block->addr = entry->addr; in get_vm_block() 63 block->size = size; in get_vm_block() 64 list_add(&block->list, &vm->used); in get_vm_block() 69 return block; in get_vm_block() 72 static void put_vm_block(struct ct_vm *vm, struct ct_vm_block *block) in put_vm_block() argument [all …]
|
| /linux/fs/afs/ |
| H A D | dir_edit.c | 18 * Find a number of contiguous clear bits in a directory block bitmask. 21 * variable. The first bit doesn't count as it corresponds to the block header 24 static int afs_find_contig_bits(union afs_xdr_dir_block *block, unsigned int nr_slots) in afs_find_contig_bits() argument 30 bitmap = (u64)block->hdr.bitmap[0] << 0 * 8; in afs_find_contig_bits() 31 bitmap |= (u64)block->hdr.bitmap[1] << 1 * 8; in afs_find_contig_bits() 32 bitmap |= (u64)block->hdr.bitmap[2] << 2 * 8; in afs_find_contig_bits() 33 bitmap |= (u64)block->hdr.bitmap[3] << 3 * 8; in afs_find_contig_bits() 34 bitmap |= (u64)block->hdr.bitmap[4] << 4 * 8; in afs_find_contig_bits() 35 bitmap |= (u64)block->hdr.bitmap[5] << 5 * 8; in afs_find_contig_bits() 36 bitmap |= (u64)block->hdr.bitmap[6] << 6 * 8; in afs_find_contig_bits() [all …]
|
| /linux/drivers/misc/ |
| H A D | sram.c | 57 static int sram_add_pool(struct sram_dev *sram, struct sram_reserve *block, in sram_add_pool() argument 63 NUMA_NO_NODE, block->label); in sram_add_pool() 68 block->size, NUMA_NO_NODE); in sram_add_pool() 77 static int sram_add_export(struct sram_dev *sram, struct sram_reserve *block, in sram_add_export() argument 90 part->battr.size = block->size; in sram_add_export() 95 static int sram_add_partition(struct sram_dev *sram, struct sram_reserve *block, in sram_add_partition() argument 107 virt_base = devm_ioremap_resource(sram->dev, &block->res); in sram_add_partition() 109 virt_base = devm_ioremap_resource_wc(sram->dev, &block->res); in sram_add_partition() 112 dev_err(sram->dev, "could not map SRAM at %pr\n", &block->res); in sram_add_partition() 118 part->base = sram->virt_base + block->start; in sram_add_partition() [all …]
|
| /linux/drivers/net/ethernet/mellanox/mlxsw/ |
| H A D | spectrum_flow.c | 15 struct mlxsw_sp_flow_block *block; in mlxsw_sp_flow_block_create() local 17 block = kzalloc_obj(*block); in mlxsw_sp_flow_block_create() 18 if (!block) in mlxsw_sp_flow_block_create() 20 INIT_LIST_HEAD(&block->binding_list); in mlxsw_sp_flow_block_create() 21 INIT_LIST_HEAD(&block->mall.list); in mlxsw_sp_flow_block_create() 22 block->mlxsw_sp = mlxsw_sp; in mlxsw_sp_flow_block_create() 23 block->net = net; in mlxsw_sp_flow_block_create() 24 return block; in mlxsw_sp_flow_block_create() 27 void mlxsw_sp_flow_block_destroy(struct mlxsw_sp_flow_block *block) in mlxsw_sp_flow_block_destroy() argument 29 WARN_ON(!list_empty(&block->binding_list)); in mlxsw_sp_flow_block_destroy() [all …]
|
| H A D | core_acl_flex_actions.c | 364 /* Block structure holds a list of action sets. One action block 375 * in this block. 381 void (*destructor)(struct mlxsw_afa_block *block, 385 static void mlxsw_afa_resource_add(struct mlxsw_afa_block *block, in mlxsw_afa_resource_add() argument 388 list_add(&resource->list, &block->resource_list); in mlxsw_afa_resource_add() 396 static void mlxsw_afa_resources_destroy(struct mlxsw_afa_block *block) in mlxsw_afa_resources_destroy() argument 400 list_for_each_entry_safe(resource, tmp, &block->resource_list, list) { in mlxsw_afa_resources_destroy() 401 resource->destructor(block, resource); in mlxsw_afa_resources_destroy() 407 struct mlxsw_afa_block *block; in mlxsw_afa_block_create() local 409 block = kzalloc_obj(*block); in mlxsw_afa_block_create() [all …]
|
| H A D | core_acl_flex_actions.h | 47 void mlxsw_afa_block_destroy(struct mlxsw_afa_block *block); 48 int mlxsw_afa_block_commit(struct mlxsw_afa_block *block); 49 char *mlxsw_afa_block_first_set(struct mlxsw_afa_block *block); 50 char *mlxsw_afa_block_cur_set(struct mlxsw_afa_block *block); 51 u32 mlxsw_afa_block_first_kvdl_index(struct mlxsw_afa_block *block); 52 int mlxsw_afa_block_activity_get(struct mlxsw_afa_block *block, bool *activity); 53 int mlxsw_afa_block_continue(struct mlxsw_afa_block *block); 54 int mlxsw_afa_block_jump(struct mlxsw_afa_block *block, u16 group_id); 55 int mlxsw_afa_block_terminate(struct mlxsw_afa_block *block); 58 int mlxsw_afa_block_append_drop(struct mlxsw_afa_block *block, bool ingress, [all …]
|
| /linux/drivers/video/fbdev/ |
| H A D | edid.h | 72 #define PIXEL_CLOCK_LO (unsigned)block[ 0 ] 73 #define PIXEL_CLOCK_HI (unsigned)block[ 1 ] 75 #define H_ACTIVE_LO (unsigned)block[ 2 ] 76 #define H_BLANKING_LO (unsigned)block[ 3 ] 77 #define H_ACTIVE_HI UPPER_NIBBLE( (unsigned)block[ 4 ] ) 79 #define H_BLANKING_HI LOWER_NIBBLE( (unsigned)block[ 4 ] ) 82 #define V_ACTIVE_LO (unsigned)block[ 5 ] 83 #define V_BLANKING_LO (unsigned)block[ 6 ] 84 #define V_ACTIVE_HI UPPER_NIBBLE( (unsigned)block[ 7 ] ) 86 #define V_BLANKING_HI LOWER_NIBBLE( (unsigned)block[ 7 ] ) [all …]
|
| /linux/drivers/mtd/nand/raw/ |
| H A D | nand_bbt.c | 4 * Bad block table support for the NAND driver 10 * When nand_scan_bbt is called, then it tries to find the bad block table 14 * Once a new bad block is discovered then the "factory" information is updated 21 * If the tables are not versioned, then we "or" the bad block information. 24 * good / bad blocks and the bad block tables are created. 29 * The auto generated bad block table is located in the last good blocks 38 * The table uses 2 bits per block 39 * 11b: block is good 40 * 00b: block is factory marked bad 41 * 01b, 10b: block is marked bad due to wear [all …]
|
| /linux/Documentation/admin-guide/device-mapper/ |
| H A D | dm-dust.rst | 15 in the "bad block list" will fail with EIO ("Input/output error"). 17 Writes of blocks in the "bad block list will result in the following: 19 1. Remove the block from the "bad block list". 40 Path to the block device. 46 Block size in bytes 59 (For a device with a block size of 512 bytes) 65 (For a device with a block size of 4096 bytes) 73 bad block additions, removals, and remaps will be verbosely logged):: 89 At any time (i.e.: whether the device has the "bad block" emulation 94 kernel: device-mapper: dust: badblock added at block 60 [all …]
|
| /linux/Documentation/filesystems/ext4/ |
| H A D | journal.rst | 45 consumes an entire block group, though mke2fs tries to put it in the 74 or a block revocation list. A finished transaction always ends with a 85 will be an ext4 super block in the usual place, with a matching UUID. 86 The journal superblock will be in the next full block after the 105 Block Header 108 Every block in the journal starts with a common 12-byte header 126 - Description of what this block contains. See the jbd2_blocktype_ table 131 - The transaction ID that goes with this block. 135 The journal block type can be any one of: 144 - Descriptor. This block precedes a series of data blocks that were [all …]
|
| /linux/arch/mips/include/asm/octeon/ |
| H A D | cvmx-bootmem.h | 45 /* Allocate from end of block instead of beginning */ 51 /* First bytes of each free physical block of memory contain this structure, 76 /* Base address of named block */ 79 * Size actually allocated for named block (may differ from 83 /* name of named block */ 122 /* address of named memory block descriptors */ 149 * Allocate a block of memory from the free list that was 155 * @size: Size in bytes of block to allocate 159 * Returns pointer to block of memory, NULL on error 165 * Frees a previously allocated named bootmem block. [all …]
|
| /linux/drivers/block/ |
| H A D | Kconfig | 3 # Block device driver configuration 7 bool "Block devices" 8 depends on BLOCK 11 Say Y here to get to see options for various different block device 19 source "drivers/block/null_blk/Kconfig" 20 source "drivers/block/rnull/Kconfig" 107 source "drivers/block/mtip32xx/Kconfig" 109 source "drivers/block/zram/Kconfig" 112 bool "Virtual block device" 116 you access arbitrary files on the host computer as block devices. [all …]
|
| /linux/drivers/net/ethernet/freescale/dpaa2/ |
| H A D | dpaa2-switch-flower.c | 160 dpaa2_switch_acl_entry_remove(struct dpaa2_switch_filter_block *block, in dpaa2_switch_acl_entry_remove() argument 165 struct ethsw_core *ethsw = block->ethsw; in dpaa2_switch_acl_entry_remove() 186 block->acl_id, acl_entry_cfg); in dpaa2_switch_acl_entry_remove() 202 dpaa2_switch_acl_entry_add_to_list(struct dpaa2_switch_filter_block *block, in dpaa2_switch_acl_entry_add_to_list() argument 209 if (list_empty(&block->acl_entries)) { in dpaa2_switch_acl_entry_add_to_list() 210 list_add(&entry->list, &block->acl_entries); in dpaa2_switch_acl_entry_add_to_list() 214 list_for_each_safe(pos, n, &block->acl_entries) { in dpaa2_switch_acl_entry_add_to_list() 225 dpaa2_switch_acl_entry_get_by_index(struct dpaa2_switch_filter_block *block, in dpaa2_switch_acl_entry_get_by_index() argument 231 list_for_each_entry(tmp, &block->acl_entries, list) { in dpaa2_switch_acl_entry_get_by_index() 241 dpaa2_switch_acl_entry_set_precedence(struct dpaa2_switch_filter_block *block, in dpaa2_switch_acl_entry_set_precedence() argument [all …]
|