/lib/842/ |
A D | 842_compress.c | 214 p->out += p->bit / 8; in add_bits() 215 p->olen -= p->bit / 8; in add_bits() 247 ret = add_bits(p, p->data8[0], 64); in add_template() 343 ret = add_bits(p, p->in[i], 8); in add_short_data_template() 410 p->data8[0] = get_input_data(p, 0, 64); in get_next_data() 411 p->data4[0] = get_input_data(p, 0, 32); in get_next_data() 412 p->data4[1] = get_input_data(p, 4, 32); in get_next_data() 413 p->data2[0] = get_input_data(p, 0, 16); in get_next_data() 425 u64 pos = p->in - p->instart; in update_hashtables() 493 p->instart = p->in; in sw842_compress() [all …]
|
A D | 842_decompress.c | 89 u8 *in = p->in, b = p->bit, bits = b + n; in next_bits() 120 p->bit += n; in next_bits() 123 p->in += p->bit / 8; in next_bits() 124 p->ilen -= p->bit / 8; in next_bits() 125 p->bit %= 8; in next_bits() 157 p->out += n; in do_data() 158 p->olen -= n; in do_data() 205 memcpy(p->out, &p->ostart[offset], size); in __do_index() 286 p.bit = 0; in sw842_decompress() 319 memcpy(p.out, p.out - 8, 8); in sw842_decompress() [all …]
|
/lib/ |
A D | xxhash.c | 117 p += 4; in xxh32() 119 p += 4; in xxh32() 121 p += 4; in xxh32() 123 p += 4; in xxh32() 137 p += 4; in xxh32() 143 p++; in xxh32() 214 p += 8; in xxh64() 220 p += 4; in xxh64() 226 p++; in xxh64() 329 memcpy(state->mem64, p, (size_t)(b_end-p)); in xxh64_update() [all …]
|
A D | parser.c | 39 if (!p) in match_one() 48 if (strncmp(p, s, meta-p)) in match_one() 51 s += meta - p; in match_one() 52 p = meta + 1; in match_one() 55 len = simple_strtoul(p, (char **) &p, 10); in match_one() 59 p++; in match_one() 119 for (p = table; !match_one(s, p->pattern, args) ; p++) in match_token() 294 p++; in match_wildcard() 306 p++; in match_wildcard() 319 ++p; in match_wildcard() [all …]
|
A D | flex_proportions.c | 42 p->period = 0; in fprop_global_init() 47 seqcount_init(&p->sequence); in fprop_global_init() 53 percpu_counter_destroy(&p->events); in fprop_global_destroy() 74 write_seqcount_begin(&p->sequence); in fprop_new_period() 79 p->period += periods; in fprop_new_period() 80 write_seqcount_end(&p->sequence); in fprop_new_period() 111 unsigned int period = p->period; in fprop_reflect_period_percpu() 142 fprop_reflect_period_percpu(p, pl); in __fprop_add_percpu() 144 percpu_counter_add(&p->events, nr); in __fprop_add_percpu() 156 fprop_reflect_period_percpu(p, pl); in fprop_fraction_percpu() [all …]
|
A D | memcat_p.c | 12 void **p = a, **new; in __memcat_p() local 16 for (nr = 0, p = a; *p; nr++, p++) in __memcat_p() 18 for (p = b; *p; nr++, p++) in __memcat_p() 28 for (nr--; nr >= 0; nr--, p = p == b ? &a[nr] : p - 1) in __memcat_p() 29 new[nr] = *p; in __memcat_p()
|
A D | bootconfig.c | 202 p++; in xbc_node_match_prefix() 518 ret = p + strlen(p); in skip_comment() 527 while (isspace(*p) && *p != '\n') in skip_spaces_until_newline() 528 p++; in skip_spaces_until_newline() 529 return p; in skip_spaces_until_newline() 584 p = skip_spaces_until_newline(p); in __xbc_parse_value() 589 p++; in __xbc_parse_value() 601 p = skip_comment(p); in __xbc_parse_value() 678 char *p; in __xbc_parse_keys() local 687 k = p; in __xbc_parse_keys() [all …]
|
A D | earlycpio.c | 64 const char *p, *dptr, *nptr; in find_cpio_data() local 70 p = data; in find_cpio_data() 73 if (!*p) { in find_cpio_data() 75 p += 4; in find_cpio_data() 86 c = *p++; in find_cpio_data() 111 dptr = PTR_ALIGN(p + ch[C_NAMESIZE], 4); in find_cpio_data() 114 if (nptr > p + len || dptr < p || nptr < dptr) in find_cpio_data() 119 !memcmp(p, path, mypathsize)) { in find_cpio_data() 127 p, MAX_CPIO_FILE_NAME); in find_cpio_data() 135 len -= (nptr - p); in find_cpio_data() [all …]
|
/lib/tests/ |
A D | slub_kunit.c | 37 p[64] = 0x12; in test_clobber_zone() 60 p[s->offset] = ~p[s->offset]; in test_next_pointer() 100 *p = 0x78; in test_first_word() 115 p[50] = 0x9a; in test_clobber_50th_byte() 132 p[64] = 0xab; in test_clobber_redzone_free() 151 p[18] = 0xab; in test_kmalloc_redzone_access() 255 u8 *p; in test_krealloc_redzone_zeroing() local 267 p = krealloc(p, 40, GFP_KERNEL | __GFP_ZERO); in test_krealloc_redzone_zeroing() 272 p = krealloc(p, 56, GFP_KERNEL | __GFP_ZERO); in test_krealloc_redzone_zeroing() 283 p = krealloc(p, 112, GFP_KERNEL | __GFP_ZERO); in test_krealloc_redzone_zeroing() [all …]
|
/lib/crc/arm64/ |
A D | crc32.h | 24 return crc32_le_base(crc, p, len); in crc32_le_arch() 28 crc = crc32_le_arm64_4way(crc, p, len); in crc32_le_arch() 31 p += round_down(len, 64); in crc32_le_arch() 38 return crc32_le_arm64(crc, p, len); in crc32_le_arch() 44 return crc32c_base(crc, p, len); in crc32c_arch() 51 p += round_down(len, 64); in crc32c_arch() 58 return crc32c_le_arm64(crc, p, len); in crc32c_arch() 64 return crc32_be_base(crc, p, len); in crc32_be_arch() 68 crc = crc32_be_arm64_4way(crc, p, len); in crc32_be_arch() 71 p += round_down(len, 64); in crc32_be_arch() [all …]
|
/lib/raid6/ |
A D | recov_loongarch_simd.c | 32 u8 *p, *q, *dp, *dq; in raid6_2data_recov_lsx() local 36 p = (u8 *)ptrs[disks - 2]; in raid6_2data_recov_lsx() 56 ptrs[disks - 2] = p; in raid6_2data_recov_lsx() 177 p += 64; in raid6_2data_recov_lsx() 189 u8 *p, *q, *dq; in raid6_datap_recov_lsx() local 280 p += 64; in raid6_datap_recov_lsx() 306 u8 *p, *q, *dp, *dq; in raid6_2data_recov_lasx() local 330 ptrs[disks - 2] = p; in raid6_2data_recov_lasx() 416 p += 64; in raid6_2data_recov_lasx() 428 u8 *p, *q, *dq; in raid6_datap_recov_lasx() local [all …]
|
A D | avx512.c | 45 u8 *p, *q; in raid6_avx5121_gen_syndrome() local 103 u8 *p, *q; in raid6_avx5121_xor_syndrome() local 172 u8 *p, *q; in raid6_avx5122_gen_syndrome() local 223 : "m" (p[d]), "m" (p[d+64]), "m" (q[d]), in raid6_avx5122_gen_syndrome() 235 u8 *p, *q; in raid6_avx5122_xor_syndrome() local 256 "m" (p[d]), "m" (p[d+64])); in raid6_avx5122_xor_syndrome() 331 u8 *p, *q; in raid6_avx5124_gen_syndrome() local 412 : "m" (p[d]), "m" (p[d+64]), "m" (p[d+128]), in raid6_avx5124_gen_syndrome() 425 u8 *p, *q; in raid6_avx5124_xor_syndrome() local 453 "m" (p[d]), "m" (p[d+64]), "m" (p[d+128]), in raid6_avx5124_xor_syndrome() [all …]
|
A D | avx2.c | 37 u8 *p, *q; in raid6_avx21_gen_syndrome() local 41 p = dptr[z0+1]; /* XOR parity */ in raid6_avx21_gen_syndrome() 86 u8 *p, *q; in raid6_avx21_xor_syndrome() local 90 p = dptr[disks-2]; /* XOR parity */ in raid6_avx21_xor_syndrome() 144 u8 *p, *q; in raid6_avx22_gen_syndrome() local 148 p = dptr[z0+1]; /* XOR parity */ in raid6_avx22_gen_syndrome() 196 u8 *p, *q; in raid6_avx22_xor_syndrome() local 200 p = dptr[disks-2]; /* XOR parity */ in raid6_avx22_xor_syndrome() 276 u8 *p, *q; in raid6_avx24_gen_syndrome() local 280 p = dptr[z0+1]; /* XOR parity */ in raid6_avx24_gen_syndrome() [all …]
|
A D | sse2.c | 39 u8 *p, *q; in raid6_sse21_gen_syndrome() local 43 p = dptr[z0+1]; /* XOR parity */ in raid6_sse21_gen_syndrome() 91 u8 *p, *q; in raid6_sse21_xor_syndrome() local 95 p = dptr[disks-2]; /* XOR parity */ in raid6_sse21_xor_syndrome() 149 u8 *p, *q; in raid6_sse22_gen_syndrome() local 153 p = dptr[z0+1]; /* XOR parity */ in raid6_sse22_gen_syndrome() 202 u8 *p, *q; in raid6_sse22_xor_syndrome() local 206 p = dptr[disks-2]; /* XOR parity */ in raid6_sse22_xor_syndrome() 281 u8 *p, *q; in raid6_sse24_gen_syndrome() local 285 p = dptr[z0+1]; /* XOR parity */ in raid6_sse24_gen_syndrome() [all …]
|
A D | rvv.c | 29 u8 *p, *q; in raid6_rvv1_gen_syndrome_real() local 34 p = dptr[z0 + 1]; /* XOR parity */ in raid6_rvv1_gen_syndrome_real() 92 [wp0]"r"(&p[d + NSIZE * 0]), in raid6_rvv1_gen_syndrome_real() 102 u8 *p, *q; in raid6_rvv1_xor_syndrome_real() local 200 u8 *p, *q; in raid6_rvv2_gen_syndrome_real() local 205 p = dptr[z0 + 1]; /* XOR parity */ in raid6_rvv2_gen_syndrome_real() 292 u8 *p, *q; in raid6_rvv2_xor_syndrome_real() local 420 u8 *p, *q; in raid6_rvv4_gen_syndrome_real() local 546 u8 *p, *q; in raid6_rvv4_xor_syndrome_real() local 730 u8 *p, *q; in raid6_rvv8_gen_syndrome_real() local [all …]
|
A D | recov_avx2.c | 19 u8 *p, *q, *dp, *dq; in raid6_2data_recov_avx2() local 24 p = (u8 *)ptrs[disks-2]; in raid6_2data_recov_avx2() 42 ptrs[disks-2] = p; in raid6_2data_recov_avx2() 59 asm volatile("vmovdqa %0, %%ymm0" : : "m" (p[0])); in raid6_2data_recov_avx2() 129 p += 64; in raid6_2data_recov_avx2() 135 asm volatile("vmovdqa %0, %%ymm0" : : "m" (*p)); in raid6_2data_recov_avx2() 176 p += 32; in raid6_2data_recov_avx2() 189 u8 *p, *q, *dq; in raid6_datap_recov_avx2() local 193 p = (u8 *)ptrs[disks-2]; in raid6_datap_recov_avx2() 264 p += 64; in raid6_datap_recov_avx2() [all …]
|
A D | recov_avx512.c | 25 u8 *p, *q, *dp, *dq; in raid6_2data_recov_avx512() local 30 p = (u8 *)ptrs[disks-2]; in raid6_2data_recov_avx512() 51 ptrs[disks-2] = p; in raid6_2data_recov_avx512() 154 p += 128; in raid6_2data_recov_avx512() 215 p += 64; in raid6_2data_recov_avx512() 228 u8 *p, *q, *dq; in raid6_datap_recov_avx512() local 302 : "m" (p[0]), "m" (p[64])); in raid6_datap_recov_avx512() 315 "m" (p[64])); in raid6_datap_recov_avx512() 318 p += 128; in raid6_datap_recov_avx512() 348 : "m" (p[0])); in raid6_datap_recov_avx512() [all …]
|
A D | recov_ssse3.c | 19 u8 *p, *q, *dp, *dq; in raid6_2data_recov_ssse3() local 26 p = (u8 *)ptrs[disks-2]; in raid6_2data_recov_ssse3() 44 ptrs[disks-2] = p; in raid6_2data_recov_ssse3() 69 asm volatile("movdqa %0,%%xmm0" : : "m" (p[0])); in raid6_2data_recov_ssse3() 133 p += 32; in raid6_2data_recov_ssse3() 139 asm volatile("movdqa %0,%%xmm0" : : "m" (*p)); in raid6_2data_recov_ssse3() 180 p += 16; in raid6_2data_recov_ssse3() 194 u8 *p, *q, *dq; in raid6_datap_recov_ssse3() local 200 p = (u8 *)ptrs[disks-2]; in raid6_datap_recov_ssse3() 278 p += 32; in raid6_datap_recov_ssse3() [all …]
|
A D | loongarch_simd.c | 36 u8 *p, *q; in raid6_lsx_gen_syndrome() local 40 p = dptr[z0+1]; /* XOR parity */ in raid6_lsx_gen_syndrome() 118 u8 *p, *q; in raid6_lsx_xor_syndrome() local 122 p = dptr[disks-2]; /* XOR parity */ in raid6_lsx_xor_syndrome() 235 : "+m"(p[d+NSIZE*0]), "+m"(p[d+NSIZE*1]), in raid6_lsx_xor_syndrome() 236 "+m"(p[d+NSIZE*2]), "+m"(p[d+NSIZE*3]), in raid6_lsx_xor_syndrome() 267 u8 *p, *q; in raid6_lasx_gen_syndrome() local 271 p = dptr[z0+1]; /* XOR parity */ in raid6_lasx_gen_syndrome() 327 u8 *p, *q; in raid6_lasx_xor_syndrome() local 331 p = dptr[disks-2]; /* XOR parity */ in raid6_lasx_xor_syndrome() [all …]
|
/lib/crc/arm/ |
A D | crc32.h | 31 return crc32_le_base(crc, p, len); in crc32_le_scalar() 38 size_t n = -(uintptr_t)p & 15; in crc32_le_arch() 42 crc = crc32_le_scalar(crc, p, n); in crc32_le_arch() 43 p += n; in crc32_le_arch() 48 crc = crc32_pmull_le(p, n, crc); in crc32_le_arch() 50 p += n; in crc32_le_arch() 60 return crc32c_base(crc, p, len); in crc32c_scalar() 67 size_t n = -(uintptr_t)p & 15; in crc32c_arch() 71 crc = crc32c_scalar(crc, p, n); in crc32c_arch() 72 p += n; in crc32c_arch() [all …]
|
/lib/math/ |
A D | prime_numbers.c | 100 unsigned long *p, in clear_multiples() argument 111 __clear_bit(m, p); in clear_multiples() 120 const struct primes *p; in expand_to_next_prime() local 144 if (x < p->last) { in expand_to_next_prime() 154 bitmap_copy(new->primes, p->primes, p->sz); in expand_to_next_prime() 162 if (p != &small_primes) in expand_to_next_prime() 172 const struct primes *p; in free_primes() local 199 const struct primes *p; in next_prime_number() local 203 while (x >= p->last) { in next_prime_number() 212 x = find_next_bit(p->primes, p->last, x + 1); in next_prime_number() [all …]
|
/lib/crc/loongarch/ |
A D | crc32.h | 32 return crc32_le_base(crc, p, len); in crc32_le_arch() 35 u64 value = get_unaligned_le64(p); in crc32_le_arch() 38 p += sizeof(u64); in crc32_le_arch() 46 p += sizeof(u32); in crc32_le_arch() 53 p += sizeof(u16); in crc32_le_arch() 57 u8 value = *p++; in crc32_le_arch() 68 return crc32c_base(crc, p, len); in crc32c_arch() 74 p += sizeof(u64); in crc32c_arch() 82 p += sizeof(u32); in crc32c_arch() 89 p += sizeof(u16); in crc32c_arch() [all …]
|
/lib/crc/mips/ |
A D | crc32.h | 66 return crc32_le_base(crc, p, len); in crc32_le_arch() 79 p += sizeof(u32); in crc32_le_arch() 86 p += sizeof(u32); in crc32_le_arch() 91 u16 value = get_unaligned_le16(p); in crc32_le_arch() 94 p += sizeof(u16); in crc32_le_arch() 98 u8 value = *p++; in crc32_le_arch() 109 return crc32c_base(crc, p, len); in crc32c_arch() 122 p += sizeof(u32); in crc32c_arch() 129 p += sizeof(u32); in crc32c_arch() 137 p += sizeof(u16); in crc32c_arch() [all …]
|
/lib/crc/x86/ |
A D | crc32.h | 18 static inline u32 crc32_le_arch(u32 crc, const u8 *p, size_t len) in crc32_le_arch() argument 20 CRC_PCLMUL(crc, p, len, crc32_lsb, crc32_lsb_0xedb88320_consts, in crc32_le_arch() 22 return crc32_le_base(crc, p, len); in crc32_le_arch() 39 static inline u32 crc32c_arch(u32 crc, const u8 *p, size_t len) in crc32c_arch() argument 44 return crc32c_base(crc, p, len); in crc32c_arch() 72 crc = crc32_lsb_vpclmul_avx512(crc, p, len, in crc32c_arch() 75 crc = crc32c_x86_3way(crc, p, len); in crc32c_arch() 89 num_longs != 0; num_longs--, p += sizeof(unsigned long)) in crc32c_arch() 94 p += 4; in crc32c_arch() 98 p += 2; in crc32c_arch() [all …]
|
/lib/crc/ |
A D | crc32-main.c | 35 crc32_le_base(u32 crc, const u8 *p, size_t len) in crc32_le_base() argument 38 crc = (crc >> 8) ^ crc32table_le[(crc & 255) ^ *p++]; in crc32_le_base() 43 crc32_be_base(u32 crc, const u8 *p, size_t len) in crc32_be_base() argument 46 crc = (crc << 8) ^ crc32table_be[(crc >> 24) ^ *p++]; in crc32_be_base() 51 crc32c_base(u32 crc, const u8 *p, size_t len) in crc32c_base() argument 72 u32 crc32_le(u32 crc, const void *p, size_t len) in crc32_le() argument 74 return crc32_le_arch(crc, p, len); in crc32_le() 78 u32 crc32_be(u32 crc, const void *p, size_t len) in crc32_be() argument 80 return crc32_be_arch(crc, p, len); in crc32_be() 84 u32 crc32c(u32 crc, const void *p, size_t len) in crc32c() argument [all …]
|