Home
last modified time | relevance | path

Searched refs:chunk_size (Results 1 – 25 of 103) sorted by relevance

12345

/linux/drivers/md/
H A Ddm-exception-store.c146 unsigned int chunk_size; in set_chunk_size() local
148 if (kstrtouint(chunk_size_arg, 10, &chunk_size)) { in set_chunk_size()
153 if (!chunk_size) { in set_chunk_size()
154 store->chunk_size = store->chunk_mask = store->chunk_shift = 0; in set_chunk_size()
158 return dm_exception_store_set_chunk_size(store, chunk_size, error); in set_chunk_size()
162 unsigned int chunk_size, in dm_exception_store_set_chunk_size() argument
166 if (!is_power_of_2(chunk_size)) { in dm_exception_store_set_chunk_size()
172 if (chunk_size % in dm_exception_store_set_chunk_size()
174 chunk_size % in dm_exception_store_set_chunk_size()
180 if (chunk_size > INT_MAX >> SECTOR_SHIFT) { in dm_exception_store_set_chunk_size()
[all …]
H A Ddm-unstripe.c22 uint32_t chunk_size; member
62 if (kstrtouint(argv[1], 10, &uc->chunk_size) || !uc->chunk_size) { in unstripe_ctr()
88 uc->unstripe_offset = (sector_t)uc->unstripe * uc->chunk_size; in unstripe_ctr()
89 uc->unstripe_width = (sector_t)(uc->stripes - 1) * uc->chunk_size; in unstripe_ctr()
90 uc->chunk_shift = is_power_of_2(uc->chunk_size) ? fls(uc->chunk_size) - 1 : 0; in unstripe_ctr()
93 if (sector_div(tmp_len, uc->chunk_size)) { in unstripe_ctr()
98 if (dm_set_target_max_io_len(ti, uc->chunk_size)) { in unstripe_ctr()
127 sector_div(tmp_sector, uc->chunk_size); in map_to_core()
157 uc->stripes, (unsigned long long)uc->chunk_size, uc->unstripe, in unstripe_status()
180 limits->chunk_sectors = uc->chunk_size; in unstripe_io_hints()
H A Ddm-stripe.c38 uint32_t chunk_size; member
93 uint32_t chunk_size; in stripe_ctr() local
107 if (kstrtouint(argv[1], 10, &chunk_size) || !chunk_size) { in stripe_ctr()
119 if (sector_div(tmp_len, chunk_size)) { in stripe_ctr()
150 r = dm_set_target_max_io_len(ti, chunk_size); in stripe_ctr()
162 sc->chunk_size = chunk_size; in stripe_ctr()
163 if (chunk_size & (chunk_size - 1)) in stripe_ctr()
166 sc->chunk_size_shift = __ffs(chunk_size); in stripe_ctr()
209 chunk_offset = sector_div(chunk, sc->chunk_size); in stripe_map_sector()
211 chunk_offset = chunk & (sc->chunk_size - 1); in stripe_map_sector()
[all …]
H A Ddm-snap-persistent.c84 __le32 chunk_size; member
175 len = ps->store->chunk_size << SECTOR_SHIFT; in alloc_area()
237 .sector = ps->store->chunk_size * chunk, in chunk_io()
238 .count = ps->store->chunk_size, in chunk_io()
297 memset(ps->area, 0, ps->store->chunk_size << SECTOR_SHIFT); in zero_memory_area()
310 unsigned int chunk_size; in read_header() local
318 if (!ps->store->chunk_size) { in read_header()
319 ps->store->chunk_size = max(DM_CHUNK_SIZE_DEFAULT_SECTORS, in read_header()
322 ps->store->chunk_mask = ps->store->chunk_size - 1; in read_header()
323 ps->store->chunk_shift = __ffs(ps->store->chunk_size); in read_header()
[all …]
H A Ddm-snap-transient.c48 if (size < (tc->next_free + store->chunk_size)) in transient_prepare_exception()
52 tc->next_free += store->chunk_size; in transient_prepare_exception()
100 DMEMIT(" N %llu", (unsigned long long)store->chunk_size); in transient_status()
/linux/tools/testing/selftests/net/
H A Dtcp_mmap.c91 static size_t chunk_size = 512*1024; variable
191 buffer = mmap_large_buffer(chunk_size, &buffer_sz); in child_thread()
197 raddr = mmap(NULL, chunk_size + map_align, PROT_READ, flags, fd, 0); in child_thread()
224 zc.length = min(chunk_size, FILE_SZ - total); in child_thread()
232 assert(zc.length <= chunk_size); in child_thread()
245 assert(zc.recv_skip_hint <= chunk_size); in child_thread()
261 while (sub < chunk_size) { in child_thread()
262 lu = read(fd, buffer + sub, min(chunk_size - sub, in child_thread()
319 munmap(raddr, chunk_size + map_align); in child_thread()
373 rcvlowat = chunk_size; in do_accept()
[all …]
/linux/net/xdp/
H A Dxdp_umem.c160 u32 chunk_size = mr->chunk_size, headroom = mr->headroom; in xdp_umem_reg() local
166 if (chunk_size < XDP_UMEM_MIN_CHUNK_SIZE || chunk_size > PAGE_SIZE) { in xdp_umem_reg()
179 if (!unaligned_chunks && !is_power_of_2(chunk_size)) in xdp_umem_reg()
198 chunks = div_u64_rem(size, chunk_size, &chunks_rem); in xdp_umem_reg()
205 if (headroom >= chunk_size - XDP_PACKET_HEADROOM) in xdp_umem_reg()
216 umem->chunk_size = chunk_size; in xdp_umem_reg()
H A Dxsk_buff_pool.c74 pool->chunk_mask = ~((u64)umem->chunk_size - 1); in xp_create_and_assign_umem()
79 pool->chunk_size = umem->chunk_size; in xp_create_and_assign_umem()
80 pool->chunk_shift = ffs(umem->chunk_size) - 1; in xp_create_and_assign_umem()
82 pool->frame_len = umem->chunk_size - umem->headroom - in xp_create_and_assign_umem()
103 xskb->xdp.frame_sz = umem->chunk_size - umem->headroom; in xp_create_and_assign_umem()
108 xp_init_xskb_addr(xskb, pool, (u64)i * pool->chunk_size); in xp_create_and_assign_umem()
487 return xp_desc_crosses_non_contig_pg(pool, addr, pool->chunk_size); in xp_dma_map()
494 *addr + pool->chunk_size > pool->addrs_cnt || in xp_addr_crosses_non_contig_pg()
/linux/tools/testing/selftests/ublk/
H A Dstripe.c291 unsigned chunk_size = ctx->stripe.chunk_size; in ublk_stripe_tgt_init() local
302 if ((chunk_size & (chunk_size - 1)) || !chunk_size) { in ublk_stripe_tgt_init()
303 ublk_err("invalid chunk size %u\n", chunk_size); in ublk_stripe_tgt_init()
307 if (chunk_size < 4096 || chunk_size > 512 * 1024) { in ublk_stripe_tgt_init()
308 ublk_err("invalid chunk size %u\n", chunk_size); in ublk_stripe_tgt_init()
312 chunk_shift = ilog2(chunk_size); in ublk_stripe_tgt_init()
[all...]
/linux/tools/lib/python/abi/
H A Dsystem_symbols.py234 def get_fileref(self, all_refs, chunk_size): argument
244 if n >= chunk_size:
251 def check_undefined_symbols(self, max_workers=None, chunk_size=50, argument
298 chunk_size = min(chunk_size, max_chunk_size)
299 chunk_size = max(1, chunk_size)
321 for refs in self.get_fileref(all_refs, chunk_size):
/linux/drivers/rtc/
H A Drtc-isl12026.c327 size_t chunk_size, num_written; in isl12026_nvm_write() local
347 chunk_size = round_down(offset, ISL12026_PAGESIZE) + in isl12026_nvm_write()
349 chunk_size = min(bytes, chunk_size); in isl12026_nvm_write()
354 memcpy(payload + 2, v + num_written, chunk_size); in isl12026_nvm_write()
357 msgs[0].len = chunk_size + 2; in isl12026_nvm_write()
367 bytes -= chunk_size; in isl12026_nvm_write()
368 offset += chunk_size; in isl12026_nvm_write()
369 num_written += chunk_size; in isl12026_nvm_write()
/linux/tools/perf/bench/
H A Dmem-functions.c82 size_t chunk_size; member
269 p.chunk_size = (size_t)perf_atoll((char *)chunk_size_str); in bench_mem_common()
270 if ((s64)p.chunk_size < 0 || (s64)p.chunk_size > (s64)p.size) { in bench_mem_common()
274 if (!p.chunk_size) in bench_mem_common()
275 p.chunk_size = p.size; in bench_mem_common()
336 for (size_t off = 0; off < p->size; off += p->chunk_size) in do_memcpy()
337 fn(dst + off, src + off, min(p->chunk_size, p->size - off)); in do_memcpy()
440 for (size_t off = 0; off < p->size; off += p->chunk_size) in do_memset()
441 fn(dst + off, i, min(p->chunk_size, p->size - off)); in do_memset()
/linux/fs/nilfs2/
H A Ddir.c113 unsigned int chunk_size = nilfs_chunk_size(dir); in nilfs_check_folio() local
121 if (limit & (chunk_size - 1)) in nilfs_check_folio()
136 if (((offs + rec_len - 1) ^ offs) & ~(chunk_size-1)) in nilfs_check_folio()
432 unsigned int chunk_size = nilfs_chunk_size(dir); in nilfs_add_link() local
461 rec_len = chunk_size; in nilfs_add_link()
462 de->rec_len = nilfs_rec_len_to_disk(chunk_size); in nilfs_add_link()
570 unsigned int chunk_size = nilfs_chunk_size(inode); in nilfs_make_empty() local
578 err = nilfs_prepare_chunk(folio, 0, chunk_size); in nilfs_make_empty()
584 memset(kaddr, 0, chunk_size); in nilfs_make_empty()
594 de->rec_len = nilfs_rec_len_to_disk(chunk_size - NILFS_DIR_REC_LEN(1)); in nilfs_make_empty()
[all …]
/linux/arch/x86/kernel/cpu/mtrr/
H A Dcleanup.c448 u64 chunk_size, u64 gran_size) in x86_setup_var_mtrrs() argument
457 var_state.chunk_sizek = chunk_size >> 10; in x86_setup_var_mtrrs()
564 mtrr_calc_range_state(u64 chunk_size, u64 gran_size, in mtrr_calc_range_state() argument
581 num_reg = x86_setup_var_mtrrs(range, nr_range, chunk_size, gran_size); in mtrr_calc_range_state()
589 result[i].chunk_sizek = chunk_size >> 10; in mtrr_calc_range_state()
661 u64 chunk_size, gran_size; in mtrr_cleanup() local
735 for (chunk_size = gran_size; chunk_size < (1ULL<<32); in mtrr_cleanup()
736 chunk_size <<= 1) { in mtrr_cleanup()
741 mtrr_calc_range_state(chunk_size, gran_size, in mtrr_cleanup()
761 chunk_size = result[i].chunk_sizek; in mtrr_cleanup()
[all …]
/linux/fs/ext2/
H A Ddir.c103 unsigned chunk_size = ext2_chunk_size(dir); in ext2_check_folio() local
112 if (limit & (chunk_size - 1)) in ext2_check_folio()
127 if (unlikely(((offs + rec_len - 1) ^ offs) & ~(chunk_size-1))) in ext2_check_folio()
481 unsigned chunk_size = ext2_chunk_size(dir); in ext2_add_link() local
510 rec_len = chunk_size; in ext2_add_link()
511 de->rec_len = ext2_rec_len_to_disk(chunk_size); in ext2_add_link()
620 unsigned chunk_size = ext2_chunk_size(inode); in ext2_make_empty() local
628 err = ext2_prepare_chunk(folio, 0, chunk_size); in ext2_make_empty()
634 memset(kaddr, 0, chunk_size); in ext2_make_empty()
644 de->rec_len = ext2_rec_len_to_disk(chunk_size - EXT2_DIR_REC_LEN(1)); in ext2_make_empty()
[all …]
/linux/drivers/rpmsg/
H A Dqcom_glink_native.c904 __le32 chunk_size; in qcom_glink_rx_data() member
907 unsigned int chunk_size; in qcom_glink_rx_data() local
920 chunk_size = le32_to_cpu(hdr.chunk_size); in qcom_glink_rx_data()
923 if (avail < sizeof(hdr) + chunk_size) { in qcom_glink_rx_data()
936 liid, chunk_size, left_size, in qcom_glink_rx_data()
952 intent->data = kmalloc(chunk_size + left_size, in qcom_glink_rx_data()
960 intent->size = chunk_size + left_size; in qcom_glink_rx_data()
981 if (intent->size - intent->offset < chunk_size) { in qcom_glink_rx_data()
989 sizeof(hdr), chunk_size); in qcom_glink_rx_data()
990 intent->offset += chunk_size; in qcom_glink_rx_data()
[all …]
H A Dqcom_glink_trace.h281 …TP_PROTO(const char *remote, const char *channel, u16 lcid, u16 rcid, u32 iid, u32 chunk_size, u32…
282 TP_ARGS(remote, channel, lcid, rcid, iid, chunk_size, left_size, cont, tx),
289 __field(u32, chunk_size)
300 __entry->chunk_size = chunk_size;
312 __entry->chunk_size,
/linux/sound/soc/sof/
H A Dipc4.c448 size_t chunk_size; in sof_ipc4_set_get_data() local
484 chunk_size = payload_limit; in sof_ipc4_set_get_data()
486 chunk_size = remaining; in sof_ipc4_set_get_data()
498 tx.data_size = chunk_size; in sof_ipc4_set_get_data()
501 tx_size = chunk_size; in sof_ipc4_set_get_data()
506 rx.data_size = chunk_size; in sof_ipc4_set_get_data()
510 rx_size = chunk_size; in sof_ipc4_set_get_data()
534 if (rx_size < chunk_size) { in sof_ipc4_set_get_data()
535 chunk_size = rx_size; in sof_ipc4_set_get_data()
542 offset += chunk_size; in sof_ipc4_set_get_data()
[all …]
/linux/Documentation/admin-guide/device-mapper/
H A Dstriped.rst36 my $chunk_size = 128 * 2;
54 $stripe_dev_size -= $stripe_dev_size % ($chunk_size * $num_devs);
56 $table = "0 $stripe_dev_size striped $num_devs $chunk_size";
/linux/mm/kmsan/
H A Dcore.c261 int chunk_size; in kmsan_internal_check_memory() local
268 chunk_size = min(size - pos, in kmsan_internal_check_memory()
284 pos += chunk_size; in kmsan_internal_check_memory()
287 for (int i = 0; i < chunk_size; i++) { in kmsan_internal_check_memory()
320 pos += chunk_size; in kmsan_internal_check_memory()
/linux/fs/nfs/blocklayout/
H A Ddev.c185 p = xdr_decode_hyper(p, &b->stripe.chunk_size); in nfs4_block_decode_volume()
263 chunk = div_u64(offset, dev->chunk_size); in bl_map_stripe()
268 __func__, chunk_idx, offset, dev->chunk_size); in bl_map_stripe()
274 offset = chunk * dev->chunk_size; in bl_map_stripe()
277 disk_offset = disk_chunk * dev->chunk_size; in bl_map_stripe()
284 map->len = dev->chunk_size; in bl_map_stripe()
507 d->chunk_size = v->stripe.chunk_size; in bl_parse_stripe()
/linux/drivers/platform/x86/intel/ifs/
H A Difs.h188 u32 chunk_size :16; member
201 u16 chunk_size; member
370 u32 chunk_size; member
/linux/drivers/net/wireless/ath/wcn36xx/
H A Ddxe.c239 int i, chunk_size = pool->chunk_size; in wcn36xx_dxe_init_tx_bd() local
250 bd_phy_addr += chunk_size; in wcn36xx_dxe_init_tx_bd()
251 bd_cpu_addr += chunk_size; in wcn36xx_dxe_init_tx_bd()
694 wcn->mgmt_mem_pool.chunk_size = WCN36XX_BD_CHUNK_SIZE + in wcn36xx_dxe_allocate_mem_pools()
697 s = wcn->mgmt_mem_pool.chunk_size * WCN36XX_DXE_CH_DESC_NUMB_TX_H; in wcn36xx_dxe_allocate_mem_pools()
709 wcn->data_mem_pool.chunk_size = WCN36XX_BD_CHUNK_SIZE + in wcn36xx_dxe_allocate_mem_pools()
712 s = wcn->data_mem_pool.chunk_size * WCN36XX_DXE_CH_DESC_NUMB_TX_L; in wcn36xx_dxe_allocate_mem_pools()
732 dma_free_coherent(wcn->dev, wcn->mgmt_mem_pool.chunk_size * in wcn36xx_dxe_free_mem_pools()
738 dma_free_coherent(wcn->dev, wcn->data_mem_pool.chunk_size * in wcn36xx_dxe_free_mem_pools()
/linux/include/uapi/linux/raid/
H A Dmd_u.h102 int chunk_size; /* 1 chunk size in bytes */ member
145 int chunk_size; /* in bytes */
143 int chunk_size; /* in bytes */ global() member
/linux/arch/x86/xen/
H A Dsetup.c792 phys_addr_t mem_end, addr, size, chunk_size; in xen_memory_setup() local
899 chunk_size = size; in xen_memory_setup()
907 chunk_size = min(size, mem_end - addr); in xen_memory_setup()
909 chunk_size = min(size, PFN_PHYS(extra_pages)); in xen_memory_setup()
911 n_pfns = PFN_DOWN(addr + chunk_size) - pfn_s; in xen_memory_setup()
920 xen_align_and_add_e820_region(addr, chunk_size, type); in xen_memory_setup()
922 addr += chunk_size; in xen_memory_setup()
923 size -= chunk_size; in xen_memory_setup()

12345