Lines Matching refs:size_t
161 MEM_STATIC size_t ZSTD_cwksp_available_space(ZSTD_cwksp* ws);
176 MEM_STATIC size_t ZSTD_cwksp_align(size_t size, size_t const align) { in ZSTD_cwksp_align()
177 size_t const mask = align - 1; in ZSTD_cwksp_align()
194 MEM_STATIC size_t ZSTD_cwksp_alloc_size(size_t size) { in ZSTD_cwksp_alloc_size()
204 MEM_STATIC size_t ZSTD_cwksp_aligned_alloc_size(size_t size) { in ZSTD_cwksp_aligned_alloc_size()
212 MEM_STATIC size_t ZSTD_cwksp_slack_space_required(void) { in ZSTD_cwksp_slack_space_required()
220 size_t const slackSpace = ZSTD_CWKSP_ALIGNMENT_BYTES; in ZSTD_cwksp_slack_space_required()
229 MEM_STATIC size_t ZSTD_cwksp_bytes_to_align_ptr(void* ptr, const size_t alignBytes) { in ZSTD_cwksp_bytes_to_align_ptr()
230 size_t const alignBytesMask = alignBytes - 1; in ZSTD_cwksp_bytes_to_align_ptr()
231 size_t const bytes = (alignBytes - ((size_t)ptr & (alignBytesMask))) & alignBytesMask; in ZSTD_cwksp_bytes_to_align_ptr()
245 ZSTD_cwksp_reserve_internal_buffer_space(ZSTD_cwksp* ws, size_t const bytes) in ZSTD_cwksp_reserve_internal_buffer_space()
272 MEM_STATIC size_t
287 size_t const bytesToAlign = in ZSTD_cwksp_internal_advance_phase()
296 … size_t const bytesToAlign = ZSTD_cwksp_bytes_to_align_ptr(alloc, ZSTD_CWKSP_ALIGNMENT_BYTES); in ZSTD_cwksp_internal_advance_phase()
324 ZSTD_cwksp_reserve_internal(ZSTD_cwksp* ws, size_t bytes, ZSTD_cwksp_alloc_phase_e phase) in ZSTD_cwksp_reserve_internal()
341 MEM_STATIC BYTE* ZSTD_cwksp_reserve_buffer(ZSTD_cwksp* ws, size_t bytes) in ZSTD_cwksp_reserve_buffer()
349 MEM_STATIC void* ZSTD_cwksp_reserve_aligned(ZSTD_cwksp* ws, size_t bytes) in ZSTD_cwksp_reserve_aligned()
353 assert(((size_t)ptr & (ZSTD_CWKSP_ALIGNMENT_BYTES-1))== 0); in ZSTD_cwksp_reserve_aligned()
362 MEM_STATIC void* ZSTD_cwksp_reserve_table(ZSTD_cwksp* ws, size_t bytes) in ZSTD_cwksp_reserve_table()
390 assert(((size_t)alloc & (ZSTD_CWKSP_ALIGNMENT_BYTES-1))== 0); in ZSTD_cwksp_reserve_table()
398 MEM_STATIC void* ZSTD_cwksp_reserve_object(ZSTD_cwksp* ws, size_t bytes) in ZSTD_cwksp_reserve_object()
400 size_t const roundedBytes = ZSTD_cwksp_align(bytes, sizeof(void*)); in ZSTD_cwksp_reserve_object()
408 assert((size_t)alloc % ZSTD_ALIGNOF(void*) == 0); in ZSTD_cwksp_reserve_object()
494 MEM_STATIC void ZSTD_cwksp_init(ZSTD_cwksp* ws, void* start, size_t size, ZSTD_cwksp_static_alloc_e… in ZSTD_cwksp_init()
496 assert(((size_t)start & (sizeof(void*)-1)) == 0); /* ensure correct alignment */ in ZSTD_cwksp_init()
508 MEM_STATIC size_t ZSTD_cwksp_create(ZSTD_cwksp* ws, size_t size, ZSTD_customMem customMem) { in ZSTD_cwksp_create()
532 MEM_STATIC size_t ZSTD_cwksp_sizeof(const ZSTD_cwksp* ws) { in ZSTD_cwksp_sizeof()
533 return (size_t)((BYTE*)ws->workspaceEnd - (BYTE*)ws->workspace); in ZSTD_cwksp_sizeof()
536 MEM_STATIC size_t ZSTD_cwksp_used(const ZSTD_cwksp* ws) { in ZSTD_cwksp_used()
537 return (size_t)((BYTE*)ws->tableEnd - (BYTE*)ws->workspace) in ZSTD_cwksp_used()
538 + (size_t)((BYTE*)ws->workspaceEnd - (BYTE*)ws->allocStart); in ZSTD_cwksp_used()
554 … size_t const estimatedSpace, int resizedWorkspace) { in ZSTD_cwksp_estimated_space_within_bounds()
567 MEM_STATIC size_t ZSTD_cwksp_available_space(ZSTD_cwksp* ws) { in ZSTD_cwksp_available_space()
568 return (size_t)((BYTE*)ws->allocStart - (BYTE*)ws->tableEnd); in ZSTD_cwksp_available_space()
571 MEM_STATIC int ZSTD_cwksp_check_available(ZSTD_cwksp* ws, size_t additionalNeededSpace) { in ZSTD_cwksp_check_available()
575 MEM_STATIC int ZSTD_cwksp_check_too_large(ZSTD_cwksp* ws, size_t additionalNeededSpace) { in ZSTD_cwksp_check_too_large()
580 MEM_STATIC int ZSTD_cwksp_check_wasteful(ZSTD_cwksp* ws, size_t additionalNeededSpace) { in ZSTD_cwksp_check_wasteful()
586 ZSTD_cwksp* ws, size_t additionalNeededSpace) { in ZSTD_cwksp_bump_oversized_duration()