Home
last modified time | relevance | path

Searched refs:usize (Results 1 – 25 of 38) sorted by relevance

12

/freebsd/contrib/jemalloc/include/jemalloc/internal/
H A Dprof_inlines_b.h49 prof_tctx_set(tsdn_t *tsdn, const void *ptr, size_t usize, in prof_tctx_set() argument
54 arena_prof_tctx_set(tsdn, ptr, usize, alloc_ctx, tctx); in prof_tctx_set()
83 prof_sample_check(tsd_t *tsd, size_t usize, bool update) { in prof_sample_check() argument
84 ssize_t check = update ? 0 : usize; in prof_sample_check()
88 bytes_until_sample -= usize; in prof_sample_check()
101 prof_sample_accum_update(tsd_t *tsd, size_t usize, bool update, in prof_sample_accum_update() argument
108 if (likely(prof_sample_check(tsd, usize, update))) { in prof_sample_accum_update()
131 if (!booted && prof_sample_check(tsd, usize, update)) { in prof_sample_accum_update()
146 prof_alloc_prep(tsd_t *tsd, size_t usize, bool prof_active, bool update) { in prof_alloc_prep() argument
151 assert(usize == sz_s2u(usize)); in prof_alloc_prep()
[all …]
H A Dsz.h116 size_t usize = (psz + delta_mask) & ~delta_mask; in sz_psz2u() local
117 return usize; in sz_psz2u()
192 size_t usize = grp_size + mod_size; in sz_index2size_compute() local
193 return usize; in sz_index2size_compute()
233 size_t usize = (size + delta_mask) & ~delta_mask; in sz_s2u_compute() local
234 return usize; in sz_s2u_compute()
264 size_t usize; in sz_sa2u() local
284 usize = sz_s2u(ALIGNMENT_CEILING(size, alignment)); in sz_sa2u()
285 if (usize < SC_LARGE_MINCLASS) { in sz_sa2u()
286 return usize; in sz_sa2u()
[all …]
H A Dsafety_check.h9 safety_check_set_redzone(void *ptr, size_t usize, size_t bumped_usize) { in safety_check_set_redzone() argument
10 assert(usize < bumped_usize); in safety_check_set_redzone()
11 for (size_t i = usize; i < bumped_usize && i < usize + 32; ++i) { in safety_check_set_redzone()
17 safety_check_verify_redzone(const void *ptr, size_t usize, size_t bumped_usize) in safety_check_verify_redzone() argument
19 for (size_t i = usize; i < bumped_usize && i < usize + 32; ++i) { in safety_check_verify_redzone()
H A Dtcache_inlines.h47 size_t usize JEMALLOC_CC_SILENCE_INIT(0); in tcache_alloc_small()
73 usize = sz_index2size(binind); in tcache_alloc_small()
74 assert(tcache_salloc(tsd_tsdn(tsd), ret) == usize); in tcache_alloc_small()
83 memset(ret, 0, usize); in tcache_alloc_small()
90 memset(ret, 0, usize); in tcache_alloc_small()
97 tcache->prof_accumbytes += usize; in tcache_alloc_small()
129 size_t usize JEMALLOC_CC_SILENCE_INIT(0); in tcache_alloc_large()
134 usize = sz_index2size(binind); in tcache_alloc_large()
135 assert(usize <= tcache_maxclass); in tcache_alloc_large()
142 usize); in tcache_alloc_large()
[all …]
H A Djemalloc_internal_inlines_c.h67 ipallocztm(tsdn_t *tsdn, size_t usize, size_t alignment, bool zero, in ipallocztm() argument
71 assert(usize != 0); in ipallocztm()
72 assert(usize == sz_sa2u(usize, alignment)); in ipallocztm()
78 ret = arena_palloc(tsdn, arena, usize, alignment, zero, tcache); in ipallocztm()
87 ipalloct(tsdn_t *tsdn, size_t usize, size_t alignment, bool zero, in ipalloct() argument
89 return ipallocztm(tsdn, usize, alignment, zero, tcache, false, arena); in ipalloct()
93 ipalloc(tsd_t *tsd, size_t usize, size_t alignment, bool zero) { in ipalloc() argument
94 return ipallocztm(tsd_tsdn(tsd), usize, alignment, zero, in ipalloc()
141 size_t usize, copysize; in iralloct_realign() local
143 usize = sz_sa2u(size, alignment); in iralloct_realign()
[all …]
H A Dprof_inlines_a.h47 size_t usize) { in prof_accum_cancel() argument
60 a1 = (a0 >= SC_LARGE_MINCLASS - usize) in prof_accum_cancel()
61 ? a0 - (SC_LARGE_MINCLASS - usize) : 0; in prof_accum_cancel()
67 a1 = (a0 >= SC_LARGE_MINCLASS - usize) in prof_accum_cancel()
68 ? a0 - (SC_LARGE_MINCLASS - usize) : 0; in prof_accum_cancel()
H A Dlarge_externs.h6 void *large_malloc(tsdn_t *tsdn, arena_t *arena, size_t usize, bool zero);
7 void *large_palloc(tsdn_t *tsdn, arena_t *arena, size_t usize, size_t alignment,
11 void *large_ralloc(tsdn_t *tsdn, arena_t *arena, void *ptr, size_t usize,
H A Darena_externs.h36 size_t usize, size_t alignment, bool *zero);
61 void *arena_palloc(tsdn_t *tsdn, arena_t *arena, size_t usize,
63 void arena_prof_promote(tsdn_t *tsdn, void *ptr, size_t usize);
H A Dprof_externs.h47 void prof_malloc_sample_object(tsdn_t *tsdn, const void *ptr, size_t usize,
49 void prof_free_sampled_object(tsd_t *tsd, const void *ptr, size_t usize,
/freebsd/contrib/jemalloc/src/
H A Dlarge.c14 large_malloc(tsdn_t *tsdn, arena_t *arena, size_t usize, bool zero) { in large_malloc() argument
15 assert(usize == sz_s2u(usize)); in large_malloc()
17 return large_palloc(tsdn, arena, usize, CACHELINE, zero); in large_malloc()
21 large_palloc(tsdn_t *tsdn, arena_t *arena, size_t usize, size_t alignment, in large_palloc() argument
30 ausize = sz_sa2u(usize, alignment); in large_palloc()
45 arena = arena_choose_maybe_huge(tsdn_tsd(tsdn), arena, usize); in large_palloc()
48 arena, usize, alignment, &is_zeroed)) == NULL) { in large_palloc()
59 if (config_prof && arena_prof_accum(tsdn, arena, usize)) { in large_palloc()
96 large_ralloc_no_move_shrink(tsdn_t *tsdn, extent_t *extent, size_t usize) { in large_ralloc_no_move_shrink() argument
100 size_t diff = extent_size_get(extent) - (usize + sz_large_pad); in large_ralloc_no_move_shrink()
[all …]
H A Djemalloc.c1864 bool usize; member
1878 static_opts->usize = false; in static_opts_init()
1893 size_t usize; member
1905 dynamic_opts->usize = 0; in dynamic_opts_init()
1917 size_t size, size_t usize, szind_t ind) { in imalloc_no_sample() argument
1949 return ipalloct(tsd_tsdn(tsd), usize, dopts->alignment, in imalloc_no_sample()
1959 size_t usize, szind_t ind) { in imalloc_sample() argument
1967 size_t bumped_usize = usize; in imalloc_sample()
1969 if (usize <= SC_SMALL_MAXCLASS) { in imalloc_sample()
1981 arena_prof_promote(tsd_tsdn(tsd), ret, usize); in imalloc_sample()
[all …]
H A Darena.c384 arena_large_malloc_stats_update(tsdn_t *tsdn, arena_t *arena, size_t usize) { in arena_large_malloc_stats_update() argument
389 if (usize < SC_LARGE_MINCLASS) { in arena_large_malloc_stats_update()
390 usize = SC_LARGE_MINCLASS; in arena_large_malloc_stats_update()
392 index = sz_size2index(usize); in arena_large_malloc_stats_update()
400 arena_large_dalloc_stats_update(tsdn_t *tsdn, arena_t *arena, size_t usize) { in arena_large_dalloc_stats_update() argument
405 if (usize < SC_LARGE_MINCLASS) { in arena_large_dalloc_stats_update()
406 usize = SC_LARGE_MINCLASS; in arena_large_dalloc_stats_update()
408 index = sz_size2index(usize); in arena_large_dalloc_stats_update()
417 size_t usize) { in arena_large_ralloc_stats_update() argument
419 arena_large_malloc_stats_update(tsdn, arena, usize); in arena_large_ralloc_stats_update()
[all …]
H A Dckh.c274 size_t usize; in ckh_grow() local
277 usize = sz_sa2u(sizeof(ckhc_t) << lg_curcells, CACHELINE); in ckh_grow()
278 if (unlikely(usize == 0 in ckh_grow()
279 || usize > SC_LARGE_MAXCLASS)) { in ckh_grow()
283 tab = (ckhc_t *)ipallocztm(tsd_tsdn(tsd), usize, CACHELINE, in ckh_grow()
314 size_t usize; in ckh_shrink() local
323 usize = sz_sa2u(sizeof(ckhc_t) << lg_curcells, CACHELINE); in ckh_shrink()
324 if (unlikely(usize == 0 || usize > SC_LARGE_MAXCLASS)) { in ckh_shrink()
327 tab = (ckhc_t *)ipallocztm(tsd_tsdn(tsd), usize, CACHELINE, true, NULL, in ckh_shrink()
363 size_t mincells, usize; in ckh_new() local
[all …]
H A Dbase.c252 size_t usize = ALIGNMENT_CEILING(size, alignment); in base_block_alloc() local
264 + usize)); in base_block_alloc()
424 size_t usize = ALIGNMENT_CEILING(size, alignment); in base_alloc_impl() local
425 size_t asize = usize + alignment - QUANTUM; in base_alloc_impl()
438 extent = base_extent_alloc(tsdn, base, usize, alignment); in base_alloc_impl()
446 ret = base_extent_bump_alloc(base, extent, usize, alignment); in base_alloc_impl()
H A Dprof.c146 size_t usize; member
343 prof_malloc_sample_object(tsdn_t *tsdn, const void *ptr, size_t usize, in prof_malloc_sample_object() argument
345 prof_tctx_set(tsdn, ptr, usize, NULL, tctx); in prof_malloc_sample_object()
355 tctx->cnts.curbytes += usize; in prof_malloc_sample_object()
358 tctx->cnts.accumbytes += usize; in prof_malloc_sample_object()
444 prof_try_log(tsd_t *tsd, const void *ptr, size_t usize, prof_tctx_t *tctx) { in prof_try_log() argument
506 new_node->usize = usize; in prof_try_log()
521 prof_free_sampled_object(tsd_t *tsd, const void *ptr, size_t usize, in prof_free_sampled_object() argument
526 assert(tctx->cnts.curbytes >= usize); in prof_free_sampled_object()
528 tctx->cnts.curbytes -= usize; in prof_free_sampled_object()
[all …]
/freebsd/contrib/libarchive/libarchive/
H A Darchive_write_add_filter_lzop.c321 lzo_uint usize, csize; in drive_compressor()
331 usize = (lzo_uint) in drive_compressor()
337 r = lzo1x_1_compress(data->uncompressed, usize, in drive_compressor()
342 r = lzo1x_1_15_compress(data->uncompressed, usize, in drive_compressor()
347 r = lzo1x_999_compress_level(data->uncompressed, usize, in drive_compressor()
359 archive_be32enc(p + header_bytes, (uint32_t)usize); in drive_compressor()
361 checksum = lzo_adler32(1, data->uncompressed, usize); in drive_compressor()
364 if (csize < usize) { in drive_compressor()
374 archive_be32enc(p + header_bytes + 4, (uint32_t)usize); in drive_compressor()
380 usize); in drive_compressor()
322 lzo_uint usize, csize; drive_compressor() local
[all...]
/freebsd/bin/ed/
H A Dundo.c34 static long usize = 0; /* stack size variable */ variable
45 (ustack = (undo_t *) malloc((usize = USIZE) * sizeof(undo_t))) == NULL) { in push_undo_stack()
52 if (u_p < usize || in push_undo_stack()
53 (t = (undo_t *) realloc(ustack, (usize += USIZE) * sizeof(undo_t))) != NULL) { in push_undo_stack()
66 usize = 0; in push_undo_stack()
/freebsd/usr.bin/gzip/
H A Dgzip.c1750 off_t usize, gsize; in handle_stdin() local
1793 usize = cat_fd(fourbytes, sizeof fourbytes, &gsize, STDIN_FILENO); in handle_stdin()
1796 usize = gz_uncompress(STDIN_FILENO, STDOUT_FILENO, in handle_stdin()
1801 usize = unbzip2(STDIN_FILENO, STDOUT_FILENO, in handle_stdin()
1812 usize = zuncompress(in, stdout, (char *)fourbytes, in handle_stdin()
1819 usize = unpack(STDIN_FILENO, STDOUT_FILENO, in handle_stdin()
1825 usize = unxz(STDIN_FILENO, STDOUT_FILENO, in handle_stdin()
1831 usize = unlz(STDIN_FILENO, STDOUT_FILENO, in handle_stdin()
1837 usize = unzstd(STDIN_FILENO, STDOUT_FILENO, in handle_stdin()
1843 if (vflag && !tflag && usize != -1 && gsize != -1) in handle_stdin()
[all …]
/freebsd/sys/contrib/openzfs/tests/zfs-tests/tests/functional/mount/
H A Dumount_unlinked_drain.ksh54 usize=$(($nunlinks - $nunlinked))
55 if [[ $iters == $MAX_ITERS && $usize == $1 ]]; then
58 if [[ $usize == $last_usize ]]; then
63 last_usize=$usize
/freebsd/sys/compat/linux/
H A Dlinux_fork.c435 if (args->usize > PAGE_SIZE) in linux_clone3()
437 if (args->usize < LINUX_CLONE_ARGS_SIZE_VER0) in linux_clone3()
445 size = max(args->usize, sizeof(*uca)); in linux_clone3()
447 error = copyin(args->uargs, uca, args->usize); in linux_clone3()
/freebsd/contrib/llvm-project/compiler-rt/lib/sanitizer_common/
H A Dsanitizer_placement_new.h20 inline void *operator new(__sanitizer::usize sz, void *p) { return p; } in new()
H A Dsanitizer_internal_defs.h195 typedef __SIZE_TYPE__ usize; typedef
203 typedef unsigned int usize; typedef
205 typedef uptr usize; typedef
/freebsd/contrib/jemalloc/
H A DFREEBSD-diffs446 + size_t usize = je_xallocx(*ptr, size, extra, flags);
447 + ret = (usize >= size) ? ALLOCM_SUCCESS : ALLOCM_ERR_NOT_MOVED;
449 + *rsize = usize;
481 + size_t usize = je_nallocx(size, flags);
482 + if (usize == 0) {
486 + *rsize = usize;
/freebsd/sys/contrib/openzfs/cmd/
H A Ddbufstat.in403 usize = int(line[labels['usize']])
421 d[pool][objset][key]['cached'] += dbsize + usize
/freebsd/sys/dev/qat_c2xxx/
H A Dqat_ae.c1420 int ii, vali, fixup, usize = 0; in qat_ae_concat_ucode() local
1432 usize = nitems(ae_inst_1b); in qat_ae_concat_ucode()
1436 usize = nitems(ae_inst_2b); in qat_ae_concat_ucode()
1440 usize = nitems(ae_inst_3b); in qat_ae_concat_ucode()
1444 usize = nitems(ae_inst_4b); in qat_ae_concat_ucode()
1449 for (ii = 0; ii < usize; ii++) in qat_ae_concat_ucode()

12