| /lib/ |
| A D | vsprintf.c | 536 if (buf < end) in number() 565 if (buf < end) in number() 571 if (buf < end) in number() 577 if (buf < end) in number() 638 if (buf < end) in widen_string() 656 if (buf < end) in string_nocheck() 1057 if (buf < end) in hex_range() 1894 if (buf < end) in date_str() 1899 if (buf < end) in date_str() 1910 if (buf < end) in time_str() [all …]
|
| A D | bitmap-str.c | 216 unsigned int end; member 289 while (start <= end && __end_of_region(*end)) in bitmap_find_region_reverse() 290 end--; in bitmap_find_region_reverse() 292 return end; in bitmap_find_region_reverse() 446 if (start > end || __end_of_region(*end)) in bitmap_get_x32_reverse() 454 return end; in bitmap_get_x32_reverse() 483 end = bitmap_find_region_reverse(start, end); in bitmap_parse() 484 if (start > end) in bitmap_parse() 491 end = bitmap_get_x32_reverse(start, end, &bitmap[chunk ^ 1]); in bitmap_parse() 493 end = bitmap_get_x32_reverse(start, end, &bitmap[chunk]); in bitmap_parse() [all …]
|
| A D | siphash.c | 52 for (; data != end; data += sizeof(u64)) { in __siphash_aligned() 71 case 1: b |= end[0]; in __siphash_aligned() 85 for (; data != end; data += sizeof(u64)) { in __siphash_unaligned() 101 case 4: b |= get_unaligned_le32(end); break; in __siphash_unaligned() 104 case 1: b |= end[0]; in __siphash_unaligned() 253 for (; data != end; data += sizeof(u64)) { in __hsiphash_aligned() 271 case 1: b |= end[0]; in __hsiphash_aligned() 286 for (; data != end; data += sizeof(u64)) { in __hsiphash_unaligned() 304 case 1: b |= end[0]; in __hsiphash_unaligned() 424 case 1: b |= end[0]; in __hsiphash_aligned() [all …]
|
| A D | string_helpers.c | 365 if (out < end) in escape_passthrough() 396 if (out < end) in escape_space() 399 if (out < end) in escape_space() 429 if (out < end) in escape_special() 432 if (out < end) in escape_special() 447 if (out < end) in escape_null() 450 if (out < end) in escape_null() 462 if (out < end) in escape_octal() 864 char *end; in strim() local 871 while (end >= s && isspace(*end)) in strim() [all …]
|
| A D | test_hmm.c | 348 for (addr = start; addr < end; addr = range.end) { in dmirror_fault() 350 range.end = min(addr + (ARRAY_SIZE(pfns) << PAGE_SHIFT), end); in dmirror_fault() 396 if (end < start) in dmirror_read() 462 if (end < start) in dmirror_write() 516 devmem->pagemap.range.end = res->end; in dmirror_allocate_chunk() 733 unsigned long end = args->end; in dmirror_migrate_finalize_and_map() local 778 if (end < start) in dmirror_exclusive() 830 unsigned long end = args->end; in dmirror_devmem_fault_alloc_and_copy() local 890 if (end < start) in dmirror_migrate_to_system() 951 if (end < start) in dmirror_migrate_to_device() [all …]
|
| A D | iomap_copy.c | 24 const u32 *end = src + count; in __iowrite32_copy() local 26 while (src < end) in __iowrite32_copy() 46 const u32 __iomem *end = src + count; in __ioread32_copy() local 48 while (src < end) in __ioread32_copy() 69 const u64 *end = src + count; in __iowrite64_copy() local 71 while (src < end) in __iowrite64_copy()
|
| A D | dynamic_queue_limits.c | 28 unsigned long hist_head, t, start, end; in dql_check_stall() local 49 end = hist_head * BITS_PER_LONG + (BITS_PER_LONG - 1); in dql_check_stall() 54 if (time_before(now, end + stall_thrs / 2)) in dql_check_stall() 55 end = now - stall_thrs / 2; in dql_check_stall() 58 for (t = start; time_before_eq(t, end); t++) in dql_check_stall() 64 if (!time_before_eq(t, end)) in dql_check_stall()
|
| A D | oid_registry.c | 132 const unsigned char *v = data, *end = v + datasize; in sprint_oid() local 138 if (v >= end) in sprint_oid() 148 while (v < end) { in sprint_oid() 155 if (v >= end) in sprint_oid()
|
| A D | maple_tree.c | 887 meta->end = end; in ma_set_meta() 924 meta->end = 0; in mt_clear_meta() 937 return meta->end; in ma_meta_end() 1771 unsigned char end; in mas_find_child() local 2003 unsigned char end; in mab_mas_cp() local 2022 end = j - 1; in mab_mas_cp() 2764 unsigned char end; in mtree_range_walk() local 2802 mas->end = end; in mtree_range_walk() 3067 tmp += end; in mas_destroy_rebalance() 4029 unsigned char end = mas->end; in mas_wr_append() local [all …]
|
| A D | bitmap.c | 420 unsigned long index, end, i; in bitmap_find_next_zero_area_off() local 427 end = index + nr; in bitmap_find_next_zero_area_off() 428 if (end > size) in bitmap_find_next_zero_area_off() 429 return end; in bitmap_find_next_zero_area_off() 430 i = find_next_bit(map, end, index); in bitmap_find_next_zero_area_off() 431 if (i < end) { in bitmap_find_next_zero_area_off() 868 const unsigned long *end = bitmap + BITS_TO_LONGS(nbits); in bitmap_to_arr64() local 870 while (bitmap < end) { in bitmap_to_arr64() 872 if (bitmap < end) in bitmap_to_arr64()
|
| A D | decompress_unlzo.c | 45 u8 *end = input + in_len; in parse_header() local 79 if (end - parse < 8 + 1 + 4) in parse_header() 89 if (end - parse < l + 4) in parse_header()
|
| A D | fw_table.c | 116 unsigned long end) in call_handler() argument 119 return proc->handler(hdr, end); in call_handler() 121 return proc->handler_arg(hdr, proc->arg, end); in call_handler()
|
| A D | dynamic_debug.c | 284 char *end; in ddebug_tokenize() local 296 for (end = buf; *end && *end != quote; end++) in ddebug_tokenize() 298 if (!*end) { in ddebug_tokenize() 303 for (end = buf; *end && !isspace(*end); end++) in ddebug_tokenize() 305 if (end == buf) { in ddebug_tokenize() 317 if (*end) in ddebug_tokenize() 318 *end++ = '\0'; /* terminate the word */ in ddebug_tokenize() 320 buf = end; in ddebug_tokenize()
|
| A D | idr.c | 79 int idr_alloc(struct idr *idr, void *ptr, int start, int end, gfp_t gfp) in idr_alloc() argument 87 ret = idr_alloc_u32(idr, ptr, &id, end > 0 ? end - 1 : INT_MAX, gfp); in idr_alloc() 117 int idr_alloc_cyclic(struct idr *idr, void *ptr, int start, int end, gfp_t gfp) in idr_alloc_cyclic() argument 120 int err, max = end > 0 ? end - 1 : INT_MAX; in idr_alloc_cyclic()
|
| /lib/kunit/ |
| A D | executor_test.c | 47 .start = subsuite, .end = &subsuite[2], in filter_suites_test() 66 KUNIT_ASSERT_EQ(test, got.end - got.start, 1); in filter_suites_test() 73 .start = subsuite, .end = &subsuite[2], in filter_suites_test_glob_test() 90 KUNIT_ASSERT_EQ(test, got.end - got.start, 1); in filter_suites_test_glob_test() 102 .start = subsuite, .end = &subsuite[2], in filter_suites_to_empty_test() 114 KUNIT_EXPECT_PTR_EQ_MSG(test, got.start, got.end, in filter_suites_to_empty_test() 154 .start = subsuite, .end = &subsuite[2], in filter_attr_test() 180 KUNIT_ASSERT_EQ(test, got.end - got.start, 1); in filter_attr_test() 192 .start = subsuite, .end = &subsuite[2], in filter_attr_empty_test() 205 KUNIT_EXPECT_PTR_EQ_MSG(test, got.start, got.end, in filter_attr_empty_test() [all …]
|
| A D | executor.c | 152 for (suites = suite_set.start; suites < suite_set.end; suites++) { in kunit_free_suite_set() 178 const size_t max = suite_set->end - suite_set->start; in kunit_filter_suites() 206 for (i = 0; &suite_set->start[i] != suite_set->end; i++) { in kunit_filter_suites() 249 filtered.end = copy; in kunit_filter_suites() 278 size_t num_suites = suite_set->end - suite_set->start; in kunit_exec_run_tests() 297 for (suites = suite_set->start; suites < suite_set->end; suites++) { in kunit_exec_list_tests() 320 init_num_suites = init_suite_set.end - init_suite_set.start; in kunit_merge_suite_sets() 321 num_suites = suite_set.end - suite_set.start; in kunit_merge_suite_sets() 338 total_suite_set.end = total_suite_start + (init_num_suites + num_suites); in kunit_merge_suite_sets() 372 size_t init_num_suites = init_suite_set.end - init_suite_set.start; in kunit_run_all_tests()
|
| /lib/dim/ |
| A D | dim.c | 58 const struct dim_sample *end, in dim_calc_stats() argument 62 u32 delta_us = ktime_us_delta(end->time, start->time); in dim_calc_stats() 63 u32 npkts = BIT_GAP(BITS_PER_TYPE(u32), end->pkt_ctr, start->pkt_ctr); in dim_calc_stats() 64 u32 nbytes = BIT_GAP(BITS_PER_TYPE(u32), end->byte_ctr, in dim_calc_stats() 66 u32 ncomps = BIT_GAP(BITS_PER_TYPE(u32), end->comp_ctr, in dim_calc_stats()
|
| /lib/zlib_inflate/ |
| A D | inftrees.c | 43 int end; /* use base and extra for symbol > end */ in zlib_inflate_table() local 171 end = 19; in zlib_inflate_table() 178 end = 256; in zlib_inflate_table() 183 end = -1; in zlib_inflate_table() 205 if ((int)(work[sym]) < end) { in zlib_inflate_table() 209 else if ((int)(work[sym]) > end) { in zlib_inflate_table()
|
| A D | inffast.c | 74 unsigned char *end; /* while out < end, enough space available */ in inflate_fast() local 101 end = out + (strm->avail_out - 257); in inflate_fast() 308 } while (in < last && out < end); in inflate_fast() 320 strm->avail_out = (unsigned)(out < end ? in inflate_fast() 321 257 + (end - out) : 257 - (out - end)); in inflate_fast()
|
| /lib/lz4/ |
| A D | lz4hc_compress.c | 66 hc4->end = start; in LZ4HC_init() 372 ctx->end += inputSize; in LZ4HC_compress_generic() 638 ctxPtr->end = (const BYTE *)dictionary + dictSize; in LZ4_loadDictHC() 649 if (ctxPtr->end >= ctxPtr->base + 4) { in LZ4HC_setExternalDict() 651 LZ4HC_Insert(ctxPtr, ctxPtr->end - 3); in LZ4HC_setExternalDict() 659 ctxPtr->dictLimit = (U32)(ctxPtr->end - ctxPtr->base); in LZ4HC_setExternalDict() 662 ctxPtr->end = newBlock; in LZ4HC_setExternalDict() 682 if ((size_t)(ctxPtr->end - ctxPtr->base) > 2 * GB) { in LZ4_compressHC_continue_generic() 692 if ((const BYTE *)source != ctxPtr->end) in LZ4_compressHC_continue_generic() 740 int const prefixSize = (int)(streamPtr->end in LZ4_saveDictHC() [all …]
|
| /lib/zstd/compress/ |
| A D | zstd_cwksp.h | 415 void* end; in ZSTD_cwksp_reserve_table() local 426 end = (BYTE *)alloc + bytes; in ZSTD_cwksp_reserve_table() 433 assert(end <= top); in ZSTD_cwksp_reserve_table() 434 if (end > top) { in ZSTD_cwksp_reserve_table() 439 ws->tableEnd = end; in ZSTD_cwksp_reserve_table() 455 void* end = (BYTE*)alloc + roundedBytes; in ZSTD_cwksp_reserve_object() local 465 if (ws->phase != ZSTD_cwksp_alloc_objects || end > ws->workspaceEnd) { in ZSTD_cwksp_reserve_object() 470 ws->objectEnd = end; in ZSTD_cwksp_reserve_object() 471 ws->tableEnd = end; in ZSTD_cwksp_reserve_object() 472 ws->tableValidEnd = end; in ZSTD_cwksp_reserve_object()
|
| A D | hist.c | 33 const BYTE* const end = ip + srcSize; in HIST_add() local 35 while (ip<end) { in HIST_add() 44 const BYTE* const end = ip + srcSize; in HIST_count_simple() local 51 while (ip<end) { in HIST_count_simple()
|
| /lib/tests/ |
| A D | usercopy_kunit.c | 49 size_t start, end, i, zero_start, zero_end; in usercopy_test_check_nonzero_user() local 93 for (end = start; end <= size; end++) { in usercopy_test_check_nonzero_user() 94 size_t len = end - start; in usercopy_test_check_nonzero_user() 100 retval, expected, start, end); in usercopy_test_check_nonzero_user()
|
| A D | fortify_kunit.c | 418 int i, end = sizeof(pad.buf) - 1; in fortify_test_strlen() local 432 pad.buf[end] = 'A'; in fortify_test_strlen() 433 end = sizeof(pad.buf); in fortify_test_strlen() 441 int i, end = sizeof(pad.buf) - 1; in fortify_test_strnlen() local 459 pad.buf[end] = 'A'; in fortify_test_strnlen() 460 end = sizeof(pad.buf); in fortify_test_strnlen() 462 KUNIT_EXPECT_EQ(test, strnlen(pad.buf, end + 1), end); in fortify_test_strnlen() 464 KUNIT_EXPECT_EQ(test, strnlen(pad.buf, end + 2), end); in fortify_test_strnlen() 468 KUNIT_EXPECT_EQ(test, strnlen(pad.buf, end), end); in fortify_test_strnlen() 471 end = sizeof(pad.buf) / 2; in fortify_test_strnlen() [all …]
|
| /lib/xz/ |
| A D | xz_dec_lzma2.c | 67 size_t end; member 305 if (dict->end - dict->pos <= out_max) in dict_limit() 306 dict->limit = dict->end; in dict_limit() 328 offset += dict->end; in dict_get() 362 back += dict->end; in dict_repeat() 366 if (back == dict->end) in dict_repeat() 387 copy_size = dict->end - dict->pos; in dict_uncompressed() 407 if (dict->pos == dict->end) in dict_uncompressed() 441 if (dict->pos == dict->end) in dict_flush() 1175 s->dict.end = s->dict.size; in xz_dec_lzma2_reset() [all …]
|