Searched refs:BLAKE3_CHUNK_LEN (Results 1 – 4 of 4) sorted by relevance
165 size_t full_chunks = (content_len - 1) / BLAKE3_CHUNK_LEN; in left_len()166 return round_down_to_power_of_2(full_chunks) * BLAKE3_CHUNK_LEN; in left_len()179 assert(input_len <= MAX_SIMD_DEGREE * BLAKE3_CHUNK_LEN); in compress_chunks_parallel()185 while (input_len - input_position >= BLAKE3_CHUNK_LEN) { in compress_chunks_parallel()187 input_position += BLAKE3_CHUNK_LEN; in compress_chunks_parallel()192 BLAKE3_CHUNK_LEN / BLAKE3_BLOCK_LEN, key, chunk_counter, in compress_chunks_parallel()278 if (input_len <= blake3_simd_degree() * BLAKE3_CHUNK_LEN) { in blake3_compress_subtree_wide()291 chunk_counter + (uint64_t)(left_input_len / BLAKE3_CHUNK_LEN); in blake3_compress_subtree_wide()298 if (left_input_len > BLAKE3_CHUNK_LEN && degree == 1) { in blake3_compress_subtree_wide()342 assert(input_len > BLAKE3_CHUNK_LEN); in compress_subtree_to_parent_node()[all …]
8 #define BLAKE3_CHUNK_LEN LLVM_BLAKE3_CHUNK_LEN macro
223 size_t full_chunks = (content_len - 1) / BLAKE3_CHUNK_LEN; in left_len()224 return (round_down_to_power_of_2(full_chunks) * BLAKE3_CHUNK_LEN); in left_len()240 while (input_len - input_position >= BLAKE3_CHUNK_LEN) { in compress_chunks_parallel()242 input_position += BLAKE3_CHUNK_LEN; in compress_chunks_parallel()246 ops->hash_many(chunks_array, chunks_array_len, BLAKE3_CHUNK_LEN / in compress_chunks_parallel()333 if (input_len <= (size_t)(ops->degree * BLAKE3_CHUNK_LEN)) { in blake3_compress_subtree_wide()350 (uint64_t)(left_input_len / BLAKE3_CHUNK_LEN); in blake3_compress_subtree_wide()359 if (left_input_len > BLAKE3_CHUNK_LEN && degree == 1) { in blake3_compress_subtree_wide()543 size_t take = BLAKE3_CHUNK_LEN - chunk_state_len(&ctx->chunk); in Blake3_Update2()583 while (input_len > BLAKE3_CHUNK_LEN) { in Blake3_Update2()[all …]
47 #define BLAKE3_CHUNK_LEN 1024 macro