Lines Matching refs:bytes

238     size_t const bytes = (alignBytes - ((size_t)ptr & (alignBytesMask))) & alignBytesMask;  in ZSTD_cwksp_bytes_to_align_ptr()  local
240 assert(bytes != ZSTD_CWKSP_ALIGNMENT_BYTES); in ZSTD_cwksp_bytes_to_align_ptr()
241 return bytes; in ZSTD_cwksp_bytes_to_align_ptr()
252 ZSTD_cwksp_reserve_internal_buffer_space(ZSTD_cwksp* ws, size_t const bytes) in ZSTD_cwksp_reserve_internal_buffer_space() argument
254 void* const alloc = (BYTE*)ws->allocStart - bytes; in ZSTD_cwksp_reserve_internal_buffer_space()
257 alloc, bytes, ZSTD_cwksp_available_space(ws) - bytes); in ZSTD_cwksp_reserve_internal_buffer_space()
331 ZSTD_cwksp_reserve_internal(ZSTD_cwksp* ws, size_t bytes, ZSTD_cwksp_alloc_phase_e phase) in ZSTD_cwksp_reserve_internal() argument
334 if (ZSTD_isError(ZSTD_cwksp_internal_advance_phase(ws, phase)) || bytes == 0) { in ZSTD_cwksp_reserve_internal()
340 bytes += 2 * ZSTD_CWKSP_ASAN_REDZONE_SIZE; in ZSTD_cwksp_reserve_internal()
343 alloc = ZSTD_cwksp_reserve_internal_buffer_space(ws, bytes); in ZSTD_cwksp_reserve_internal()
351 __asan_unpoison_memory_region(alloc, bytes); in ZSTD_cwksp_reserve_internal()
362 MEM_STATIC BYTE* ZSTD_cwksp_reserve_buffer(ZSTD_cwksp* ws, size_t bytes) in ZSTD_cwksp_reserve_buffer() argument
364 return (BYTE*)ZSTD_cwksp_reserve_internal(ws, bytes, ZSTD_cwksp_alloc_buffers); in ZSTD_cwksp_reserve_buffer()
370 MEM_STATIC void* ZSTD_cwksp_reserve_aligned(ZSTD_cwksp* ws, size_t bytes) in ZSTD_cwksp_reserve_aligned() argument
372 void* ptr = ZSTD_cwksp_reserve_internal(ws, ZSTD_cwksp_align(bytes, ZSTD_CWKSP_ALIGNMENT_BYTES), in ZSTD_cwksp_reserve_aligned()
383 MEM_STATIC void* ZSTD_cwksp_reserve_table(ZSTD_cwksp* ws, size_t bytes) in ZSTD_cwksp_reserve_table() argument
394 end = (BYTE *)alloc + bytes; in ZSTD_cwksp_reserve_table()
398 alloc, bytes, ZSTD_cwksp_available_space(ws) - bytes); in ZSTD_cwksp_reserve_table()
399 assert((bytes & (sizeof(U32)-1)) == 0); in ZSTD_cwksp_reserve_table()
411 __asan_unpoison_memory_region(alloc, bytes); in ZSTD_cwksp_reserve_table()
415 assert((bytes & (ZSTD_CWKSP_ALIGNMENT_BYTES-1)) == 0); in ZSTD_cwksp_reserve_table()
424 MEM_STATIC void* ZSTD_cwksp_reserve_object(ZSTD_cwksp* ws, size_t bytes) in ZSTD_cwksp_reserve_object() argument
426 size_t const roundedBytes = ZSTD_cwksp_align(bytes, sizeof(void*)); in ZSTD_cwksp_reserve_object()
437 alloc, bytes, roundedBytes, ZSTD_cwksp_available_space(ws) - roundedBytes); in ZSTD_cwksp_reserve_object()
439 assert(bytes % ZSTD_ALIGNOF(void*) == 0); in ZSTD_cwksp_reserve_object()
456 __asan_unpoison_memory_region(alloc, bytes); in ZSTD_cwksp_reserve_object()