Home
last modified time | relevance | path

Searched full:block (Results 1 – 25 of 3949) sorted by relevance

12345678910>>...158

/linux/drivers/net/ethernet/marvell/prestera/
H A Dprestera_counter.c45 prestera_counter_is_ready(struct prestera_counter_block *block, u32 id) in prestera_counter_is_ready() argument
47 return block->counter_flag[id - block->offset] == COUNTER_FLAG_READY; in prestera_counter_is_ready()
60 static void prestera_counter_block_lock(struct prestera_counter_block *block) in prestera_counter_block_lock() argument
62 mutex_lock(&block->mtx); in prestera_counter_block_lock()
65 static void prestera_counter_block_unlock(struct prestera_counter_block *block) in prestera_counter_block_unlock() argument
67 mutex_unlock(&block->mtx); in prestera_counter_block_unlock()
70 static bool prestera_counter_block_incref(struct prestera_counter_block *block) in prestera_counter_block_incref() argument
72 return refcount_inc_not_zero(&block->refcnt); in prestera_counter_block_incref()
75 static bool prestera_counter_block_decref(struct prestera_counter_block *block) in prestera_counter_block_decref() argument
77 return refcount_dec_and_test(&block->refcnt); in prestera_counter_block_decref()
[all …]
H A Dprestera_flow.c16 static int prestera_flow_block_mall_cb(struct prestera_flow_block *block, in prestera_flow_block_mall_cb() argument
21 return prestera_mall_replace(block, f); in prestera_flow_block_mall_cb()
23 prestera_mall_destroy(block); in prestera_flow_block_mall_cb()
30 static int prestera_flow_block_flower_cb(struct prestera_flow_block *block, in prestera_flow_block_flower_cb() argument
35 return prestera_flower_replace(block, f); in prestera_flow_block_flower_cb()
37 prestera_flower_destroy(block, f); in prestera_flow_block_flower_cb()
40 return prestera_flower_stats(block, f); in prestera_flow_block_flower_cb()
42 return prestera_flower_tmplt_create(block, f); in prestera_flow_block_flower_cb()
44 prestera_flower_tmplt_destroy(block, f); in prestera_flow_block_flower_cb()
54 struct prestera_flow_block *block = cb_priv; in prestera_flow_block_cb() local
[all …]
H A Dprestera_matchall.c14 static int prestera_mall_prio_check(struct prestera_flow_block *block, in prestera_mall_prio_check() argument
21 err = prestera_flower_prio_get(block, f->common.chain_index, in prestera_mall_prio_check()
32 if (f->common.prio <= flower_prio_max && !block->ingress) { in prestera_mall_prio_check()
36 if (f->common.prio >= flower_prio_min && block->ingress) { in prestera_mall_prio_check()
44 int prestera_mall_prio_get(struct prestera_flow_block *block, in prestera_mall_prio_get() argument
47 if (!block->mall.bound) in prestera_mall_prio_get()
50 *prio_min = block->mall.prio_min; in prestera_mall_prio_get()
51 *prio_max = block->mall.prio_max; in prestera_mall_prio_get()
55 static void prestera_mall_prio_update(struct prestera_flow_block *block, in prestera_mall_prio_update() argument
58 block->mall.prio_min = min(block->mall.prio_min, f->common.prio); in prestera_mall_prio_update()
[all …]
/linux/sound/isa/gus/
H A Dgus_mem.c19 snd_gf1_mem_xalloc(struct snd_gf1_mem *alloc, struct snd_gf1_mem_block *block, in snd_gf1_mem_xalloc() argument
27 *nblock = *block; in snd_gf1_mem_xalloc()
60 int snd_gf1_mem_xfree(struct snd_gf1_mem * alloc, struct snd_gf1_mem_block * block) in snd_gf1_mem_xfree() argument
62 if (block->share) { /* ok.. shared block */ in snd_gf1_mem_xfree()
63 block->share--; in snd_gf1_mem_xfree()
66 if (alloc->first == block) { in snd_gf1_mem_xfree()
67 alloc->first = block->next; in snd_gf1_mem_xfree()
68 if (block->next) in snd_gf1_mem_xfree()
69 block->next->prev = NULL; in snd_gf1_mem_xfree()
71 block->prev->next = block->next; in snd_gf1_mem_xfree()
[all …]
/linux/drivers/iio/buffer/
H A Dindustrialio-buffer-dma.c26 * For DMA buffers the storage is sub-divided into so called blocks. Each block
27 * has its own memory buffer. The size of the block is the granularity at which
29 * basic unit of data exchange from one sample to one block decreases the
32 * sample the overhead will be x for each sample. Whereas when using a block
40 * them with data. Block on the outgoing queue have been filled with data and
43 * A block can be in one of the following states:
45 * the block.
48 * * Owned by the DMA controller: The DMA controller is processing the block
53 * * Dead: A block that is dead has been marked as to be freed. It might still
56 * incoming or outgoing queue the block will be freed.
[all …]
/linux/drivers/mtd/
H A Dnftlmount.c28 unsigned int block, boot_record_count = 0; in find_boot_record() local
48 for (block = 0; block < nftl->nb_blocks; block++) { in find_boot_record()
53 ret = mtd_read(mtd, block * nftl->EraseSize, SECTORSIZE, in find_boot_record()
61 printk(KERN_WARNING "Block read at 0x%x of mtd%d failed: %d\n", in find_boot_record()
62 block * nftl->EraseSize, nftl->mbd.mtd->index, ret); in find_boot_record()
64 printk(KERN_WARNING "Further failures for this block will not be printed\n"); in find_boot_record()
73 block * nftl->EraseSize, nftl->mbd.mtd->index); in find_boot_record()
79 ret = nftl_read_oob(mtd, block * nftl->EraseSize + in find_boot_record()
84 block * nftl->EraseSize, nftl->mbd.mtd->index, ret); in find_boot_record()
94 block * nftl->EraseSize, nftl->mbd.mtd->index, in find_boot_record()
[all …]
H A Dinftlmount.c35 unsigned int i, block; in find_boot_record() local
55 for (block = 0; block < inftl->nb_blocks; block++) { in find_boot_record()
62 ret = mtd_read(mtd, block * inftl->EraseSize, SECTORSIZE, in find_boot_record()
70 printk(KERN_WARNING "INFTL: block read at 0x%x " in find_boot_record()
72 block * inftl->EraseSize, in find_boot_record()
76 "failures for this block will " in find_boot_record()
89 block * inftl->EraseSize + SECTORSIZE + 8, in find_boot_record()
94 "(err %d)\n", block * inftl->EraseSize, in find_boot_record()
107 mtd_read(mtd, block * inftl->EraseSize + 4096, SECTORSIZE, in find_boot_record()
172 block >>= mh->BlockMultiplierBits; in find_boot_record()
[all …]
H A Drfd_ftl.c28 MODULE_PARM_DESC(block_size, "Block size to use by RFD, defaults to erase unit size");
57 struct block { struct
80 int reserved_block; /* block next up for reclaim */ argument
81 int current_block; /* block to write to */
88 struct block *blocks;
95 struct block *block = &part->blocks[block_no]; in build_block_map() local
98 block->offset = part->block_size * block_no; in build_block_map()
101 block->state = BLOCK_UNUSED; in build_block_map()
105 block->state = BLOCK_OK; in build_block_map()
116 block->free_sectors++; in build_block_map()
[all …]
/linux/drivers/gpio/
H A Dgpio-sch311x.c39 struct sch311x_gpio_block { /* one GPIO block runtime data */
44 spinlock_t lock; /* lock for this GPIO block */
134 struct sch311x_gpio_block *block = gpiochip_get_data(chip); in sch311x_gpio_request() local
136 if (block->config_regs[offset] == 0) /* GPIO is not available */ in sch311x_gpio_request()
139 if (!request_region(block->runtime_reg + block->config_regs[offset], in sch311x_gpio_request()
142 block->runtime_reg + block->config_regs[offset]); in sch311x_gpio_request()
150 struct sch311x_gpio_block *block = gpiochip_get_data(chip); in sch311x_gpio_free() local
152 if (block->config_regs[offset] == 0) /* GPIO is not available */ in sch311x_gpio_free()
155 release_region(block->runtime_reg + block->config_regs[offset], 1); in sch311x_gpio_free()
160 struct sch311x_gpio_block *block = gpiochip_get_data(chip); in sch311x_gpio_get() local
[all …]
/linux/drivers/w1/
H A Dw1_netlink.c41 struct w1_cb_block *block; member
49 * @block: block to calculate
55 static u16 w1_reply_len(struct w1_cb_block *block) in w1_reply_len() argument
57 if (!block->cn) in w1_reply_len()
59 return (u8 *)block->cn - (u8 *)block->first_cn + block->cn->len; in w1_reply_len()
62 static void w1_unref_block(struct w1_cb_block *block) in w1_unref_block() argument
64 if (atomic_sub_return(1, &block->refcnt) == 0) { in w1_unref_block()
65 u16 len = w1_reply_len(block); in w1_unref_block()
67 cn_netlink_send_mult(block->first_cn, len, in w1_unref_block()
68 block->portid, 0, in w1_unref_block()
[all …]
/linux/drivers/md/dm-vdo/
H A Drecovery-journal.h24 * The recovery_journal provides a log of all block mapping and reference count changes which have
25 * not yet been stably written to the block map or slab journals. This log helps to reduce the
26 * write amplification of writes by providing amortization of slab journal and block map page
35 * journal. The 'head' is the oldest active block in the journal. The 'tail' is the end of the
36 * half-open interval containing the active blocks. 'active' is the number of the block actively
42 * less than the on-disk size. Each in-memory block is also a vdo_completion. Each in-memory block
43 * has a vio which is used to commit that block to disk. The vio's data is the on-disk
44 * representation of the journal block. In addition each in-memory block has a buffer which is used
45 * to accumulate entries while a partial commit of the block is in progress. In-memory blocks are
46 * kept on two lists. Free blocks live on the 'free_tail_blocks' list. When a block becomes active
[all …]
/linux/net/sched/
H A Dcls_api.c181 struct tcf_block *block = chain->block; in tcf_proto_signal_destroying() local
183 mutex_lock(&block->proto_destroy_lock); in tcf_proto_signal_destroying()
184 hash_add_rcu(block->proto_destroy_ht, &tp->destroy_ht_node, in tcf_proto_signal_destroying()
186 mutex_unlock(&block->proto_destroy_lock); in tcf_proto_signal_destroying()
205 hash_for_each_possible_rcu(chain->block->proto_destroy_ht, iter, in tcf_proto_exists_destroying()
220 struct tcf_block *block = chain->block; in tcf_proto_signal_destroyed() local
222 mutex_lock(&block->proto_destroy_lock); in tcf_proto_signal_destroyed()
225 mutex_unlock(&block->proto_destroy_lock); in tcf_proto_signal_destroyed()
417 struct tcf_block *block = tp->chain->block; in tcf_proto_count_usesw() local
422 if (!atomic_dec_return(&block->useswcnt)) in tcf_proto_count_usesw()
[all …]
/linux/sound/pci/ctxfi/
H A Dctvmem.c26 * Find or create vm block based on requested @size.
32 struct ct_vm_block *block, *entry; in get_vm_block() local
46 break; /* found a block that is big enough */ in get_vm_block()
58 block = kzalloc_obj(*block); in get_vm_block()
59 if (!block) in get_vm_block()
62 block->addr = entry->addr; in get_vm_block()
63 block->size = size; in get_vm_block()
64 list_add(&block->list, &vm->used); in get_vm_block()
69 return block; in get_vm_block()
72 static void put_vm_block(struct ct_vm *vm, struct ct_vm_block *block) in put_vm_block() argument
[all …]
/linux/fs/afs/
H A Ddir_edit.c18 * Find a number of contiguous clear bits in a directory block bitmask.
21 * variable. The first bit doesn't count as it corresponds to the block header
24 static int afs_find_contig_bits(union afs_xdr_dir_block *block, unsigned int nr_slots) in afs_find_contig_bits() argument
30 bitmap = (u64)block->hdr.bitmap[0] << 0 * 8; in afs_find_contig_bits()
31 bitmap |= (u64)block->hdr.bitmap[1] << 1 * 8; in afs_find_contig_bits()
32 bitmap |= (u64)block->hdr.bitmap[2] << 2 * 8; in afs_find_contig_bits()
33 bitmap |= (u64)block->hdr.bitmap[3] << 3 * 8; in afs_find_contig_bits()
34 bitmap |= (u64)block->hdr.bitmap[4] << 4 * 8; in afs_find_contig_bits()
35 bitmap |= (u64)block->hdr.bitmap[5] << 5 * 8; in afs_find_contig_bits()
36 bitmap |= (u64)block->hdr.bitmap[6] << 6 * 8; in afs_find_contig_bits()
[all …]
/linux/drivers/misc/
H A Dsram.c57 static int sram_add_pool(struct sram_dev *sram, struct sram_reserve *block, in sram_add_pool() argument
63 NUMA_NO_NODE, block->label); in sram_add_pool()
68 block->size, NUMA_NO_NODE); in sram_add_pool()
77 static int sram_add_export(struct sram_dev *sram, struct sram_reserve *block, in sram_add_export() argument
90 part->battr.size = block->size; in sram_add_export()
95 static int sram_add_partition(struct sram_dev *sram, struct sram_reserve *block, in sram_add_partition() argument
107 virt_base = devm_ioremap_resource(sram->dev, &block->res); in sram_add_partition()
109 virt_base = devm_ioremap_resource_wc(sram->dev, &block->res); in sram_add_partition()
112 dev_err(sram->dev, "could not map SRAM at %pr\n", &block->res); in sram_add_partition()
118 part->base = sram->virt_base + block->start; in sram_add_partition()
[all …]
/linux/drivers/net/ethernet/mellanox/mlxsw/
H A Dspectrum_flow.c15 struct mlxsw_sp_flow_block *block; in mlxsw_sp_flow_block_create() local
17 block = kzalloc_obj(*block); in mlxsw_sp_flow_block_create()
18 if (!block) in mlxsw_sp_flow_block_create()
20 INIT_LIST_HEAD(&block->binding_list); in mlxsw_sp_flow_block_create()
21 INIT_LIST_HEAD(&block->mall.list); in mlxsw_sp_flow_block_create()
22 block->mlxsw_sp = mlxsw_sp; in mlxsw_sp_flow_block_create()
23 block->net = net; in mlxsw_sp_flow_block_create()
24 return block; in mlxsw_sp_flow_block_create()
27 void mlxsw_sp_flow_block_destroy(struct mlxsw_sp_flow_block *block) in mlxsw_sp_flow_block_destroy() argument
29 WARN_ON(!list_empty(&block->binding_list)); in mlxsw_sp_flow_block_destroy()
[all …]
H A Dcore_acl_flex_actions.c364 /* Block structure holds a list of action sets. One action block
375 * in this block.
381 void (*destructor)(struct mlxsw_afa_block *block,
385 static void mlxsw_afa_resource_add(struct mlxsw_afa_block *block, in mlxsw_afa_resource_add() argument
388 list_add(&resource->list, &block->resource_list); in mlxsw_afa_resource_add()
396 static void mlxsw_afa_resources_destroy(struct mlxsw_afa_block *block) in mlxsw_afa_resources_destroy() argument
400 list_for_each_entry_safe(resource, tmp, &block->resource_list, list) { in mlxsw_afa_resources_destroy()
401 resource->destructor(block, resource); in mlxsw_afa_resources_destroy()
407 struct mlxsw_afa_block *block; in mlxsw_afa_block_create() local
409 block = kzalloc_obj(*block); in mlxsw_afa_block_create()
[all …]
H A Dcore_acl_flex_actions.h47 void mlxsw_afa_block_destroy(struct mlxsw_afa_block *block);
48 int mlxsw_afa_block_commit(struct mlxsw_afa_block *block);
49 char *mlxsw_afa_block_first_set(struct mlxsw_afa_block *block);
50 char *mlxsw_afa_block_cur_set(struct mlxsw_afa_block *block);
51 u32 mlxsw_afa_block_first_kvdl_index(struct mlxsw_afa_block *block);
52 int mlxsw_afa_block_activity_get(struct mlxsw_afa_block *block, bool *activity);
53 int mlxsw_afa_block_continue(struct mlxsw_afa_block *block);
54 int mlxsw_afa_block_jump(struct mlxsw_afa_block *block, u16 group_id);
55 int mlxsw_afa_block_terminate(struct mlxsw_afa_block *block);
58 int mlxsw_afa_block_append_drop(struct mlxsw_afa_block *block, bool ingress,
[all …]
/linux/drivers/video/fbdev/
H A Dedid.h72 #define PIXEL_CLOCK_LO (unsigned)block[ 0 ]
73 #define PIXEL_CLOCK_HI (unsigned)block[ 1 ]
75 #define H_ACTIVE_LO (unsigned)block[ 2 ]
76 #define H_BLANKING_LO (unsigned)block[ 3 ]
77 #define H_ACTIVE_HI UPPER_NIBBLE( (unsigned)block[ 4 ] )
79 #define H_BLANKING_HI LOWER_NIBBLE( (unsigned)block[ 4 ] )
82 #define V_ACTIVE_LO (unsigned)block[ 5 ]
83 #define V_BLANKING_LO (unsigned)block[ 6 ]
84 #define V_ACTIVE_HI UPPER_NIBBLE( (unsigned)block[ 7 ] )
86 #define V_BLANKING_HI LOWER_NIBBLE( (unsigned)block[ 7 ] )
[all …]
/linux/drivers/mtd/nand/raw/
H A Dnand_bbt.c4 * Bad block table support for the NAND driver
10 * When nand_scan_bbt is called, then it tries to find the bad block table
14 * Once a new bad block is discovered then the "factory" information is updated
21 * If the tables are not versioned, then we "or" the bad block information.
24 * good / bad blocks and the bad block tables are created.
29 * The auto generated bad block table is located in the last good blocks
38 * The table uses 2 bits per block
39 * 11b: block is good
40 * 00b: block is factory marked bad
41 * 01b, 10b: block is marked bad due to wear
[all …]
/linux/Documentation/admin-guide/device-mapper/
H A Ddm-dust.rst15 in the "bad block list" will fail with EIO ("Input/output error").
17 Writes of blocks in the "bad block list will result in the following:
19 1. Remove the block from the "bad block list".
40 Path to the block device.
46 Block size in bytes
59 (For a device with a block size of 512 bytes)
65 (For a device with a block size of 4096 bytes)
73 bad block additions, removals, and remaps will be verbosely logged)::
89 At any time (i.e.: whether the device has the "bad block" emulation
94 kernel: device-mapper: dust: badblock added at block 60
[all …]
/linux/Documentation/filesystems/ext4/
H A Djournal.rst45 consumes an entire block group, though mke2fs tries to put it in the
74 or a block revocation list. A finished transaction always ends with a
85 will be an ext4 super block in the usual place, with a matching UUID.
86 The journal superblock will be in the next full block after the
105 Block Header
108 Every block in the journal starts with a common 12-byte header
126 - Description of what this block contains. See the jbd2_blocktype_ table
131 - The transaction ID that goes with this block.
135 The journal block type can be any one of:
144 - Descriptor. This block precedes a series of data blocks that were
[all …]
/linux/arch/mips/include/asm/octeon/
H A Dcvmx-bootmem.h45 /* Allocate from end of block instead of beginning */
51 /* First bytes of each free physical block of memory contain this structure,
76 /* Base address of named block */
79 * Size actually allocated for named block (may differ from
83 /* name of named block */
122 /* address of named memory block descriptors */
149 * Allocate a block of memory from the free list that was
155 * @size: Size in bytes of block to allocate
159 * Returns pointer to block of memory, NULL on error
165 * Frees a previously allocated named bootmem block.
[all …]
/linux/drivers/block/
H A DKconfig3 # Block device driver configuration
7 bool "Block devices"
8 depends on BLOCK
11 Say Y here to get to see options for various different block device
19 source "drivers/block/null_blk/Kconfig"
20 source "drivers/block/rnull/Kconfig"
107 source "drivers/block/mtip32xx/Kconfig"
109 source "drivers/block/zram/Kconfig"
112 bool "Virtual block device"
116 you access arbitrary files on the host computer as block devices.
[all …]
/linux/drivers/net/ethernet/freescale/dpaa2/
H A Ddpaa2-switch-flower.c160 dpaa2_switch_acl_entry_remove(struct dpaa2_switch_filter_block *block, in dpaa2_switch_acl_entry_remove() argument
165 struct ethsw_core *ethsw = block->ethsw; in dpaa2_switch_acl_entry_remove()
186 block->acl_id, acl_entry_cfg); in dpaa2_switch_acl_entry_remove()
202 dpaa2_switch_acl_entry_add_to_list(struct dpaa2_switch_filter_block *block, in dpaa2_switch_acl_entry_add_to_list() argument
209 if (list_empty(&block->acl_entries)) { in dpaa2_switch_acl_entry_add_to_list()
210 list_add(&entry->list, &block->acl_entries); in dpaa2_switch_acl_entry_add_to_list()
214 list_for_each_safe(pos, n, &block->acl_entries) { in dpaa2_switch_acl_entry_add_to_list()
225 dpaa2_switch_acl_entry_get_by_index(struct dpaa2_switch_filter_block *block, in dpaa2_switch_acl_entry_get_by_index() argument
231 list_for_each_entry(tmp, &block->acl_entries, list) { in dpaa2_switch_acl_entry_get_by_index()
241 dpaa2_switch_acl_entry_set_precedence(struct dpaa2_switch_filter_block *block, in dpaa2_switch_acl_entry_set_precedence() argument
[all …]

12345678910>>...158