/lib/ |
A D | klist.c | 111 kref_init(&n->n_ref); in klist_node_init() 114 k->get(n); in klist_node_init() 125 add_head(k, n); in klist_add_head() 137 add_tail(k, n); in klist_add_tail() 216 knode_kill(n); in klist_put() 221 put(n); in klist_put() 230 klist_put(n, true); in klist_del() 242 waiter.node = n; in klist_remove() 249 klist_del(n); in klist_remove() 285 if (n && kref_get_unless_zero(&n->n_ref)) in klist_iter_init_node() [all …]
|
A D | sort.c | 60 *(u32 *)(a + n) = *(u32 *)(b + n); in swap_words_32() 62 } while (n); in swap_words_32() 86 *(u64 *)(a + n) = *(u64 *)(b + n); in swap_words_64() 91 *(u32 *)(a + n) = *(u32 *)(b + n); in swap_words_64() 95 *(u32 *)(a + n) = *(u32 *)(b + n); in swap_words_64() 98 } while (n); in swap_words_64() 113 ((char *)a)[n] = ((char *)b)[n]; in swap_bytes() 115 } while (n); in swap_bytes() 231 n -= size; in __sort_r() 235 n -= size; in __sort_r() [all …]
|
A D | generic-radix-tree.c | 49 new_node->children[0] = n; in __genradix_ptr_alloc() 66 n = READ_ONCE(*p); in __genradix_ptr_alloc() 67 if (!n) { in __genradix_ptr_alloc() 75 swap(n, new_node); in __genradix_ptr_alloc() 82 return &n->data[offset]; in __genradix_ptr_alloc() 91 struct genradix_node *n; in __genradix_iter_peek() local 132 n = n->children[i]; in __genradix_iter_peek() 145 struct genradix_node *n; in __genradix_iter_peek_prev() local 190 n = n->children[i]; in __genradix_iter_peek_prev() 203 if (n->children[i]) in genradix_free_recurse() [all …]
|
A D | group_cpus.c | 88 int n, nodes = 0; in get_nodes_in_cpumask() local 91 for_each_node(n) { in get_nodes_in_cpumask() 93 node_set(n, *nodemsk); in get_nodes_in_cpumask() 137 unsigned n, remaining_ncpus = 0; in alloc_nodes_groups() local 139 for (n = 0; n < nr_node_ids; n++) { in alloc_nodes_groups() 140 node_groups[n].id = n; in alloc_nodes_groups() 141 node_groups[n].ncpus = UINT_MAX; in alloc_nodes_groups() 144 for_each_node_mask(n, nodemsk) { in alloc_nodes_groups() 153 node_groups[n].ncpus = ncpus; in alloc_nodes_groups() 229 for (n = 0; n < nr_node_ids; n++) { in alloc_nodes_groups() [all …]
|
A D | dhry_run.c | 34 int i, n; in dhry_benchmark() local 37 n = dhry(iterations); in dhry_benchmark() 42 n = dhry(i); in dhry_benchmark() 43 if (n != -EAGAIN) in dhry_benchmark() 49 if (n >= 0) in dhry_benchmark() 51 n, n / DHRY_VAX); in dhry_benchmark() 52 else if (n == -EAGAIN) in dhry_benchmark() 55 pr_err("Dhrystone benchmark failed error %pe\n", ERR_PTR(n)); in dhry_benchmark()
|
A D | oid_registry.c | 134 unsigned char n; in sprint_oid() local 141 n = *v++; in sprint_oid() 142 ret = count = snprintf(buffer, bufsize, "%u.%u", n / 40, n % 40); in sprint_oid() 149 n = *v++; in sprint_oid() 150 if (!(n & 0x80)) { in sprint_oid() 151 num = n; in sprint_oid() 153 num = n & 0x7f; in sprint_oid() 157 n = *v++; in sprint_oid() 159 num |= n & 0x7f; in sprint_oid() 160 } while (n & 0x80); in sprint_oid()
|
A D | inflate.c | 234 #define NEEDBITS(n) {while(k<(n)){b|=((ulg)NEXTBYTE())<<k;k+=8;}} argument 235 #define DUMPBITS(n) {b>>=(n);k-=(n);} argument 375 p = b; i = n; in huft_build() 378 n-i, *p)); in huft_build() 645 n = t->v.n + ((unsigned)b & mask_bits[e]); in inflate_codes() 666 n -= (e = (e = WSIZE - ((d &= WSIZE-1) > w ? d : w)) > n ? n : e); in inflate_codes() 721 n = k & 7; in inflate_stored() 722 DUMPBITS(n); in inflate_stored() 736 while (n--) in inflate_stored() 912 n = nl + nd; in inflate_dynamic() [all …]
|
A D | kfifo.c | 334 unsigned int n; in setup_sgl() local 345 n += setup_sgl_buf(fifo, sgl + n, 0, nents - n, len - len_to_end, dma); in setup_sgl() 347 return n; in setup_sgl() 453 if (len > *n) in kfifo_out_copy_r() 454 len = *n; in kfifo_out_copy_r() 463 unsigned int n; in __kfifo_out_peek_r() local 488 unsigned int n; in __kfifo_out_r() local 501 unsigned int n; in __kfifo_skip_r() local 536 unsigned int n; in __kfifo_to_user_r() local 544 if (len > n) in __kfifo_to_user_r() [all …]
|
A D | bootconfig.c | 758 *k = n; in xbc_parse_key() 770 *k = n; in xbc_open_brace() 779 ret = xbc_parse_key(k, n); in xbc_close_brace() 790 struct xbc_node *n, *m; in xbc_verify_tree() local 813 n = &xbc_nodes[0]; in xbc_verify_tree() 817 while (n) { in xbc_verify_tree() 822 xbc_node_get_data(n)); in xbc_verify_tree() 826 n = m; in xbc_verify_tree() 836 n = xbc_node_get_parent(n); in xbc_verify_tree() 837 if (!n) in xbc_verify_tree() [all …]
|
A D | globtest.c | 132 unsigned n = 0; in glob_init() local 150 n++; in glob_init() 153 n -= successes; in glob_init() 154 printk(message, successes, n); in glob_init() 157 return n ? -ECANCELED : 0; in glob_init()
|
/lib/zlib_deflate/ |
A D | deftree.c | 188 for (n = 0; n < (1<<extra_lbits[code]); n++) { in tr_static_init() 231 for (n = 0; n < D_CODES; n++) { in tr_static_init() 279 for (n = 0; n < L_CODES; n++) s->dyn_ltree[n].Freq = 0; in init_block() 280 for (n = 0; n < D_CODES; n++) s->dyn_dtree[n].Freq = 0; in init_block() 281 for (n = 0; n < BL_CODES; n++) s->bl_tree[n].Freq = 0; in init_block() 465 for (n = 0; n <= max_code; n++) { in gen_codes() 472 n, (isgraph(n) ? n : ' '), len, tree[n].Code, next_code[len]-1)); in gen_codes() 502 for (n = 0; n < elems; n++) { in build_tree() 528 for (n = s->heap_len/2; n >= 1; n--) pqdownheap(s, tree, n); in build_tree() 589 for (n = 0; n <= max_code; n++) { in scan_tree() [all …]
|
/lib/crc/arm/ |
A D | crc32.h | 41 if (n) { in crc32_le_arch() 43 p += n; in crc32_le_arch() 44 len -= n; in crc32_le_arch() 46 n = round_down(len, 16); in crc32_le_arch() 50 p += n; in crc32_le_arch() 51 len -= n; in crc32_le_arch() 70 if (n) { in crc32c_arch() 72 p += n; in crc32c_arch() 73 len -= n; in crc32c_arch() 79 p += n; in crc32c_arch() [all …]
|
/lib/zstd/compress/ |
A D | huf_compress.c | 266 for (n=1; n<huffLog+1; n++) in HUF_writeCTable_wksp() 268 for (n=0; n<maxSymbolValue; n++) in HUF_writeCTable_wksp() 284 for (n=0; n<maxSymbolValue; n+=2) in HUF_writeCTable_wksp() 312 for (n=1; n<=tableLog; n++) { in HUF_readCTable() 319 { U32 n; for (n=0; n<nbSymbols; n++) { in HUF_readCTable() local 327 { U32 n; for (n=0; n<nbSymbols; n++) nbPerRank[HUF_getNbBits(ct[n])]++; } in HUF_readCTable() local 337 … { U32 n; for (n=0; n<nbSymbols; n++) HUF_setValue(ct + n, valPerRank[HUF_getNbBits(ct[n])]++); } in HUF_readCTable() local 710 for (n=0; n<=nonNullRank; n++) in HUF_buildTree() 736 for (n=0; n<=nonNullRank; n++) in HUF_buildCTableFromTree() 745 for (n=0; n<alphabetSize; n++) in HUF_buildCTableFromTree() [all …]
|
A D | zstd_preSplit.c | 62 size_t n; in addEvents_generic() local 64 for (n = 0; n < limit; n+=samplingRate) { in addEvents_generic() 99 size_t n; in fpDistance() local 101 for (n = 0; n < ((size_t)1 << hashLog); n++) { in fpDistance() 127 size_t n; in mergeEvents() local 128 for (n = 0; n < HASHTABLESIZE; n++) { in mergeEvents() 129 acc->events[n] += newfp->events[n]; in mergeEvents() 137 for (n = 0; n < HASHTABLESIZE; n++) { in flushEvents() 147 for (n = 0; n < HASHTABLESIZE; n++) { in removeEvents() 148 assert(acc->events[n] >= slice->events[n]); in removeEvents() [all …]
|
A D | zstd_ldm.c | 70 size_t n = 0; in ZSTD_ldm_gear_reset() local 101 size_t n; in ZSTD_ldm_gear_feed() local 106 n = 0; in ZSTD_ldm_gear_feed() 110 n += 1; \ in ZSTD_ldm_gear_feed() 125 while (n < size) { in ZSTD_ldm_gear_feed() 133 return n; in ZSTD_ldm_gear_feed() 304 unsigned n; in ZSTD_ldm_fillHashTable() local 309 for (n = 0; n < numSplits; n++) { in ZSTD_ldm_fillHashTable() 384 unsigned n; in ZSTD_ldm_generateSequences_internal() local 390 for (n = 0; n < numSplits; n++) { in ZSTD_ldm_generateSequences_internal() [all …]
|
/lib/crc/sparc/ |
A D | crc32.h | 23 size_t n = -(uintptr_t)data & 7; in crc32c_arch() local 28 if (n) { in crc32c_arch() 30 n = min(n, len); in crc32c_arch() 31 crc = crc32c_base(crc, data, n); in crc32c_arch() 32 data += n; in crc32c_arch() 33 len -= n; in crc32c_arch() 35 n = len & ~7U; in crc32c_arch() 36 if (n) { in crc32c_arch() 37 crc32c_sparc64(&crc, (const u64 *)data, n); in crc32c_arch() 38 data += n; in crc32c_arch() [all …]
|
/lib/crypto/mpi/ |
A D | mpi-bit.c | 43 unsigned n; in mpi_get_nbits() local 50 n = count_leading_zeros(alimb); in mpi_get_nbits() 52 n = BITS_PER_MPI_LIMB; in mpi_get_nbits() 53 n = BITS_PER_MPI_LIMB - n + (a->nlimbs - 1) * BITS_PER_MPI_LIMB; in mpi_get_nbits() 55 n = 0; in mpi_get_nbits() 56 return n; in mpi_get_nbits() 68 limbno = n / BITS_PER_MPI_LIMB; in mpi_test_bit() 69 bitno = n % BITS_PER_MPI_LIMB; in mpi_test_bit() 81 int mpi_set_bit(MPI a, unsigned int n) in mpi_set_bit() argument 86 limbno = n / BITS_PER_MPI_LIMB; in mpi_set_bit() [all …]
|
A D | mpicoder.c | 155 unsigned int n = mpi_get_size(a); in mpi_read_buffer() local 166 if (buf_len < n - lzeros) { in mpi_read_buffer() 167 *nbytes = n - lzeros; in mpi_read_buffer() 172 *nbytes = n - lzeros; in mpi_read_buffer() 207 unsigned int n; in mpi_get_buffer() local 213 n = mpi_get_size(a); in mpi_get_buffer() 215 if (!n) in mpi_get_buffer() 216 n++; in mpi_get_buffer() 218 buf = kmalloc(n, GFP_KERNEL); in mpi_get_buffer() 267 if (nbytes < n) in mpi_write_to_sgl() [all …]
|
A D | mpi-internal.h | 63 #define MPN_COPY(d, s, n) \ argument 66 for (_i = 0; _i < (n); _i++) \ 70 #define MPN_COPY_DECR(d, s, n) \ argument 73 for (_i = (n)-1; _i >= 0; _i--) \ 78 #define MPN_ZERO(d, n) \ argument 81 for (_i = 0; _i < (n); _i++) \ 85 #define MPN_NORMALIZE(d, n) \ argument 87 while ((n) > 0) { \ 88 if ((d)[(n)-1]) \ 90 (n)--; \
|
/lib/zstd/common/ |
A D | zstd_deps.h | 32 #define ZSTD_memcpy(d,s,n) __builtin_memcpy((d),(s),(n)) argument 33 #define ZSTD_memmove(d,s,n) __builtin_memmove((d),(s),(n)) argument 34 #define ZSTD_memset(d,s,n) __builtin_memset((d),(s),(n)) argument 52 #define ZSTD_calloc(n,s) ({ (void)(n); (void)(s); NULL; }) argument
|
A D | entropy_common.c | 265 { U32 n; in HUF_readStats_body() local 266 for (n=0; n<oSize; n+=2) { in HUF_readStats_body() 267 huffWeight[n] = ip[n/2] >> 4; in HUF_readStats_body() 268 huffWeight[n+1] = ip[n/2] & 15; in HUF_readStats_body() 280 { U32 n; for (n=0; n<oSize; n++) { in HUF_readStats_body() local 281 if (huffWeight[n] > HUF_TABLELOG_MAX) return ERROR(corruption_detected); in HUF_readStats_body() 282 rankStats[huffWeight[n]]++; in HUF_readStats_body() 283 weightTotal += (1 << huffWeight[n]) >> 1; in HUF_readStats_body()
|
/lib/842/ |
A D | 842_decompress.c | 72 if (n <= s) { in __split_next_bits() 77 ret = next_bits(p, &tmp, n - s); in __split_next_bits() 91 if (n > 64) { in next_bits() 118 *d &= GENMASK_ULL(n - 1, 0); in next_bits() 120 p->bit += n; in next_bits() 136 if (n > p->olen) in do_data() 139 ret = next_bits(p, &v, n * 8); in do_data() 143 switch (n) { in do_data() 157 p->out += n; in do_data() 158 p->olen -= n; in do_data() [all …]
|
A D | 842_compress.c | 122 #define find_index(p, b, n) ({ \ argument 126 if (p->data##b[n] == _n->data) { \ 127 p->index##b[n] = _n->index; \ 131 p->index##b[n] >= 0; \ 134 #define check_index(p, b, n) \ argument 136 ? find_index(p, b, n) \ 137 : (p)->index##b[n] >= 0) 158 if (n <= s) in __split_add_bits() 161 ret = add_bits(p, d >> s, n - s); in __split_add_bits() 175 if (n > 64) in add_bits() [all …]
|
/lib/math/ |
A D | div64.c | 32 uint32_t __attribute__((weak)) __div64_32(uint64_t *n, uint32_t base) in __div64_32() argument 34 uint64_t rem = *n; in __div64_32() 61 *n = res; in __div64_32() 109 int n = fls(high); in div64_u64_rem() local 110 quot = div_u64(dividend >> n, divisor >> n); in div64_u64_rem() 147 int n = fls(high); in div64_u64() local 148 quot = div_u64(dividend >> n, divisor >> n); in div64_u64() 227 u64 n = shift ? (n_lo >> shift) | (n_hi << (64 - shift)) : n_lo; in mul_u64_u64_div_u64() local 229 return div64_u64(n, c >> shift); in mul_u64_u64_div_u64()
|
A D | rational.c | 51 unsigned long n, d, n0, d0, n1, d1, n2, d2; in rational_best_approximation() local 52 n = given_numerator; in rational_best_approximation() 66 a = n / d; in rational_best_approximation() 67 d = n % d; in rational_best_approximation() 68 n = dp; in rational_best_approximation()
|