Lines Matching refs:size
266 u_long size; in sysctl_kmem_map_size() local
268 size = uma_size(); in sysctl_kmem_map_size()
269 return (sysctl_handle_long(oidp, &size, 0, req)); in sysctl_kmem_map_size()
275 u_long size, limit; in sysctl_kmem_map_free() local
278 size = uma_size(); in sysctl_kmem_map_free()
280 if (size > limit) in sysctl_kmem_map_free()
281 size = 0; in sysctl_kmem_map_free()
283 size = limit - size; in sysctl_kmem_map_free()
284 return (sysctl_handle_long(oidp, &size, 0, req)); in sysctl_kmem_map_free()
388 malloc_type_zone_allocated(struct malloc_type *mtp, unsigned long size, in malloc_type_zone_allocated() argument
397 if (size > 0) { in malloc_type_zone_allocated()
398 mtsp->mts_memalloced += size; in malloc_type_zone_allocated()
410 (uintptr_t) mtsp, size, zindx); in malloc_type_zone_allocated()
418 malloc_type_allocated(struct malloc_type *mtp, unsigned long size) in malloc_type_allocated() argument
421 if (size > 0) in malloc_type_allocated()
422 malloc_type_zone_allocated(mtp, size, -1); in malloc_type_allocated()
432 malloc_type_freed(struct malloc_type *mtp, unsigned long size) in malloc_type_freed() argument
440 mtsp->mts_memfreed += size; in malloc_type_freed()
449 (uintptr_t) mtsp, size, 0); in malloc_type_freed()
480 contigmalloc(unsigned long size, struct malloc_type *type, int flags, in contigmalloc() argument
486 ret = (void *)kmem_alloc_contig(size, flags, low, high, alignment, in contigmalloc()
490 vsetzoneslab((uintptr_t)ret, NULL, CONTIG_MALLOC_SLAB(size)); in contigmalloc()
491 malloc_type_allocated(type, round_page(size)); in contigmalloc()
497 contigmalloc_domainset(unsigned long size, struct malloc_type *type, in contigmalloc_domainset() argument
503 ret = (void *)kmem_alloc_contig_domainset(ds, size, flags, low, high, in contigmalloc_domainset()
507 vsetzoneslab((uintptr_t)ret, NULL, CONTIG_MALLOC_SLAB(size)); in contigmalloc_domainset()
508 malloc_type_allocated(type, round_page(size)); in contigmalloc_domainset()
517 contigfree(void *addr, unsigned long size __unused, struct malloc_type *type) in contigfree()
594 malloc_large(size_t size, struct malloc_type *mtp, struct domainset *policy, in malloc_large() argument
599 size = roundup(size, PAGE_SIZE); in malloc_large()
600 va = kmem_malloc_domainset(policy, size, flags); in malloc_large()
603 vsetzoneslab((uintptr_t)va, NULL, MALLOC_LARGE_SLAB(size)); in malloc_large()
604 uma_total_inc(size); in malloc_large()
606 malloc_type_allocated(mtp, va == NULL ? 0 : size); in malloc_large()
614 kasan_mark(va, osize, size, KASAN_MALLOC_REDZONE); in malloc_large()
620 free_large(void *addr, size_t size) in free_large() argument
623 kmem_free(addr, size); in free_large()
624 uma_total_dec(size); in free_large()
638 (malloc)(size_t size, struct malloc_type *mtp, int flags)
644 unsigned long osize = size;
651 if (malloc_dbg(&va, &size, mtp, flags) != 0)
655 if (__predict_false(size > kmem_zmax))
656 return (malloc_large(size, mtp, DOMAINSET_RR(), flags
659 if (size & KMEM_ZMASK)
660 size = (size & ~KMEM_ZMASK) + KMEM_ZBASE;
661 indx = kmemsize[size >> KMEM_ZSHIFT];
665 size = zone->uz_size;
667 kmsan_mark(va, size, KMSAN_STATE_UNINIT);
668 kmsan_orig(va, size, KMSAN_TYPE_MALLOC, KMSAN_RET_ADDR);
671 malloc_type_zone_allocated(mtp, va == NULL ? 0 : size, indx);
682 kasan_mark((void *)va, osize, size, KASAN_MALLOC_REDZONE);
693 size_t size; in malloc_domain() local
696 size = *sizep; in malloc_domain()
697 KASSERT(size <= kmem_zmax && (flags & M_EXEC) == 0, in malloc_domain()
699 if (size & KMEM_ZMASK) in malloc_domain()
700 size = (size & ~KMEM_ZMASK) + KMEM_ZBASE; in malloc_domain()
701 indx = kmemsize[size >> KMEM_ZSHIFT]; in malloc_domain()
711 malloc_domainset(size_t size, struct malloc_type *mtp, struct domainset *ds, in malloc_domainset() argument
719 unsigned long osize = size; in malloc_domainset()
726 if (malloc_dbg(&va, &size, mtp, flags) != 0) in malloc_domainset()
730 if (__predict_false(size > kmem_zmax)) in malloc_domainset()
731 return (malloc_large(size, mtp, DOMAINSET_RR(), flags in malloc_domainset()
736 va = malloc_domain(&size, &indx, mtp, domain, flags); in malloc_domainset()
738 malloc_type_zone_allocated(mtp, va == NULL ? 0 : size, indx); in malloc_domainset()
749 kasan_mark((void *)va, osize, size, KASAN_MALLOC_REDZONE); in malloc_domainset()
753 kmsan_mark(va, size, KMSAN_STATE_UNINIT); in malloc_domainset()
754 kmsan_orig(va, size, KMSAN_TYPE_MALLOC, KMSAN_RET_ADDR); in malloc_domainset()
764 malloc_exec(size_t size, struct malloc_type *mtp, int flags) in malloc_exec() argument
767 return (malloc_domainset_exec(size, mtp, DOMAINSET_RR(), flags)); in malloc_exec()
771 malloc_domainset_exec(size_t size, struct malloc_type *mtp, struct domainset *ds, in malloc_domainset_exec() argument
775 unsigned long osize = size; in malloc_domainset_exec()
785 if (malloc_dbg(&va, &size, mtp, flags) != 0) in malloc_domainset_exec()
789 return (malloc_large(size, mtp, ds, flags DEBUG_REDZONE_ARG)); in malloc_domainset_exec()
793 malloc_aligned(size_t size, size_t align, struct malloc_type *type, int flags) in malloc_aligned() argument
795 return (malloc_domainset_aligned(size, align, type, DOMAINSET_RR(), in malloc_aligned()
800 malloc_domainset_aligned(size_t size, size_t align, in malloc_domainset_aligned() argument
808 align, size)); in malloc_domainset_aligned()
811 align, size)); in malloc_domainset_aligned()
821 if (size == 0) in malloc_domainset_aligned()
822 size = 1; in malloc_domainset_aligned()
823 asize = size <= align ? align : 1UL << flsl(size - 1); in malloc_domainset_aligned()
828 "allocsize %#zx align %#zx", res, size, asize, align)); in malloc_domainset_aligned()
833 mallocarray(size_t nmemb, size_t size, struct malloc_type *type, int flags) in mallocarray() argument
836 if (WOULD_OVERFLOW(nmemb, size)) in mallocarray()
837 panic("mallocarray: %zu * %zu overflowed", nmemb, size); in mallocarray()
839 return (malloc(size * nmemb, type, flags)); in mallocarray()
843 mallocarray_domainset(size_t nmemb, size_t size, struct malloc_type *type, in mallocarray_domainset() argument
847 if (WOULD_OVERFLOW(nmemb, size)) in mallocarray_domainset()
848 panic("mallocarray_domainset: %zu * %zu overflowed", nmemb, size); in mallocarray_domainset()
850 return (malloc_domainset(size * nmemb, type, ds, flags)); in mallocarray_domainset()
855 free_save_type(void *addr, struct malloc_type *mtp, u_long size) in free_save_type() argument
868 mtpp += (size - sizeof(struct malloc_type *)) / in free_save_type()
910 u_long size; in _free() local
927 size = zone->uz_size; in _free()
929 free_save_type(addr, mtp, size); in _free()
932 kasan_mark(addr, size, size, 0); in _free()
933 explicit_bzero(addr, size); in _free()
938 size = malloc_large_size(slab); in _free()
940 kasan_mark(addr, size, size, 0); in _free()
941 explicit_bzero(addr, size); in _free()
943 free_large(addr, size); in _free()
946 size = round_page(contigmalloc_size(slab)); in _free()
948 explicit_bzero(addr, size); in _free()
949 kmem_free(addr, size); in _free()
956 malloc_type_freed(mtp, size); in _free()
985 realloc(void *addr, size_t size, struct malloc_type *mtp, int flags) in realloc() argument
1001 return (malloc(size, mtp, flags)); in realloc()
1010 return (memguard_realloc(addr, size, mtp, flags)); in realloc()
1039 if (size <= alloc && in realloc()
1040 (size > (alloc >> REALLOC_FRACTION) || alloc == MINALLOCSIZE)) { in realloc()
1041 kasan_mark((void *)addr, size, alloc, KASAN_MALLOC_REDZONE); in realloc()
1047 if ((newaddr = malloc(size, mtp, flags)) == NULL) in realloc()
1055 bcopy(addr, newaddr, min(size, alloc)); in realloc()
1064 reallocf(void *addr, size_t size, struct malloc_type *mtp, int flags) in reallocf() argument
1068 if ((mem = realloc(addr, size, mtp, flags)) == NULL) in reallocf()
1078 malloc_size(size_t size) in malloc_size() argument
1082 if (size > kmem_zmax) in malloc_size()
1083 return (round_page(size)); in malloc_size()
1084 if (size & KMEM_ZMASK) in malloc_size()
1085 size = (size & ~KMEM_ZMASK) + KMEM_ZBASE; in malloc_size()
1086 indx = kmemsize[size >> KMEM_ZSHIFT]; in malloc_size()
1100 u_long size; in malloc_usable_size() local
1111 size = redzone_get_size(__DECONST(void *, addr)); in malloc_usable_size()
1120 size = zone->uz_size; in malloc_usable_size()
1123 size = malloc_large_size(slab); in malloc_usable_size()
1127 size = 0; in malloc_usable_size()
1136 kasan_mark(addr, size, size, 0); in malloc_usable_size()
1138 return (size); in malloc_usable_size()
1261 int size = kmemzones[indx].kz_size; in mallocinit() local
1267 if (powerof2(size) && size > sizeof(void *)) in mallocinit()
1268 align = MIN(size, PAGE_SIZE) - 1; in mallocinit()
1271 uma_zcreate(name, size, in mallocinit()
1279 for (;i <= size; i+= KMEM_ZBASE) in mallocinit()
1507 int64_t size; in DB_SHOW_COMMAND_FLAGS() local
1543 size = get_malloc_stats(&mtp->ks_mti, &allocs, in DB_SHOW_COMMAND_FLAGS()
1545 if (size > cur_size && size < last_size + ties) { in DB_SHOW_COMMAND_FLAGS()
1546 cur_size = size; in DB_SHOW_COMMAND_FLAGS()
1553 size = get_malloc_stats(&cur_mtype->ks_mti, &allocs, &inuse); in DB_SHOW_COMMAND_FLAGS()
1555 howmany(size, 1024), allocs); in DB_SHOW_COMMAND_FLAGS()