| /include/sound/ |
| A D | pcm_params.h | 23 #define MASK_OFS(i) ((i) >> 5) argument 39 for (i = 0; i < SNDRV_MASK_SIZE; i++) in snd_mask_empty() 48 for (i = 0; i < SNDRV_MASK_SIZE; i++) { in snd_mask_min() 58 for (i = SNDRV_MASK_SIZE - 1; i >= 0; i--) { in snd_mask_max() 86 for (i = from; i <= to; i++) in snd_mask_set_range() 94 for (i = from; i <= to; i++) in snd_mask_reset_range() 110 for (i = 0; i < SNDRV_MASK_SIZE; i++) in snd_mask_intersect() 141 for (i = 0; i < SNDRV_MASK_SIZE; i++) { in snd_mask_single() 233 (i->min == i->max && (i->openmin || i->openmax))); in snd_interval_checkempty() 244 (i->min + 1 == i->max && (i->openmin || i->openmax))); in snd_interval_single() [all …]
|
| A D | simple_card_utils.h | 80 #define simple_priv_to_props(priv, i) ((priv)->dai_props + (i)) argument 84 #define simple_props_to_dlc_cpu(props, i) ((props)->cpus + i) argument 88 #define simple_props_to_dai_cpu(props, i) ((props)->cpu_dai + i) argument 103 (i)++) 108 (i)++) 113 (i)++) 119 (i)++) 125 (i)++) 130 (i)++) 252 int i; in simple_util_debug_info() local [all …]
|
| /include/linux/atomic/ |
| A D | atomic-long.h | 14 #define ATOMIC_LONG_INIT(i) ATOMIC64_INIT(i) argument 19 #define ATOMIC_LONG_INIT(i) ATOMIC_INIT(i) argument 79 raw_atomic64_set(v, i); in raw_atomic_long_set() 81 raw_atomic_set(v, i); in raw_atomic_long_set() 121 raw_atomic64_add(i, v); in raw_atomic_long_add() 123 raw_atomic_add(i, v); in raw_atomic_long_add() 310 raw_atomic64_sub(i, v); in raw_atomic_long_sub() 312 raw_atomic_sub(i, v); in raw_atomic_long_sub() 861 raw_atomic_and(i, v); in raw_atomic_long_and() 1071 raw_atomic_or(i, v); in raw_atomic_long_or() [all …]
|
| A D | atomic-instrumented.h | 68 raw_atomic_set(v, i); in atomic_set() 105 raw_atomic_add(i, v); in atomic_add() 271 raw_atomic_sub(i, v); in atomic_sub() 751 raw_atomic_and(i, v); in atomic_and() 843 raw_atomic_andnot(i, v); in atomic_andnot() 935 raw_atomic_or(i, v); in atomic_or() 1027 raw_atomic_xor(i, v); in atomic_xor() 1646 raw_atomic64_set(v, i); in atomic64_set() 1683 raw_atomic64_add(i, v); in atomic64_add() 1849 raw_atomic64_sub(i, v); in atomic64_sub() [all …]
|
| A D | atomic-arch-fallback.h | 503 arch_atomic_set(v, i); in raw_atomic_set() 527 raw_atomic_set(v, i); in raw_atomic_set_release() 546 arch_atomic_add(i, v); in raw_atomic_add() 769 arch_atomic_sub(i, v); in raw_atomic_sub() 1428 arch_atomic_and(i, v); in raw_atomic_and() 1551 raw_atomic_and(~i, v); in raw_atomic_andnot() 1672 arch_atomic_or(i, v); in raw_atomic_or() 1792 arch_atomic_xor(i, v); in raw_atomic_xor() 2629 arch_atomic64_set(v, i); in raw_atomic64_set() 2653 raw_atomic64_set(v, i); in raw_atomic64_set_release() [all …]
|
| /include/linux/ |
| A D | uio.h | 106 return i->count; in iter_iov_len() 107 return iter_iov(i)->iov_len - i->iov_offset; in iter_iov_len() 112 return i->iter_type; in iov_iter_type() 165 return iter_is_ubuf(i) || iter_is_iovec(i); in user_backed_iter() 191 struct iov_iter *i); 193 struct iov_iter *i); 316 return i->count; in iov_iter_count() 333 if (i->count > count) in iov_iter_truncate() 334 i->count = count; in iov_iter_truncate() 343 i->count = count; in iov_iter_reexpand() [all …]
|
| A D | virtio_features.h | 67 int i; in virtio_features_equal() local 69 for (i = 0; i < VIRTIO_FEATURES_DWORDS; ++i) in virtio_features_equal() 70 if (f1[i] != f2[i]) in virtio_features_equal() 82 int i; in virtio_features_andnot() local 84 for (i = 0; i < VIRTIO_FEATURES_DWORDS; i++) in virtio_features_andnot() 85 to[i] = f1[i] & ~f2[i]; in virtio_features_andnot()
|
| A D | zutil.h | 57 #define DO1(buf,i) {s1 += buf[i]; s2 += s1;} argument 58 #define DO2(buf,i) DO1(buf,i); DO1(buf,i+1); argument 59 #define DO4(buf,i) DO2(buf,i); DO2(buf,i+2); argument 60 #define DO8(buf,i) DO4(buf,i); DO4(buf,i+4); argument
|
| A D | refcount.h | 186 if (unlikely(old < 0 || old + i < 0)) in __refcount_add_not_zero() 212 return __refcount_add_not_zero(i, r, NULL); in refcount_add_not_zero() 225 if (i > limit - old) { in __refcount_add_not_zero_limited_acquire() 235 if (unlikely(old < 0 || old + i < 0)) in __refcount_add_not_zero_limited_acquire() 283 int old = atomic_fetch_add_relaxed(i, &r->refs); in __refcount_add() 290 else if (unlikely(old < 0 || old + i < 0)) in __refcount_add() 312 __refcount_add(i, r, NULL); in refcount_add() 389 int old = atomic_fetch_sub_release(i, &r->refs); in __refcount_sub_and_test() 394 if (old > 0 && old == i) { in __refcount_sub_and_test() 399 if (unlikely(old <= 0 || old - i < 0)) in __refcount_sub_and_test() [all …]
|
| A D | nvram.h | 84 loff_t i; in nvram_read_bytes() local 89 for (i = *ppos; count > 0 && i < nvram_size; ++i, ++p, --count) in nvram_read_bytes() 90 *p = nvram_read_byte(i); in nvram_read_bytes() 91 *ppos = i; in nvram_read_bytes() 98 loff_t i; in nvram_write_bytes() local 103 for (i = *ppos; count > 0 && i < nvram_size; ++i, ++p, --count) in nvram_write_bytes() 104 nvram_write_byte(*p, i); in nvram_write_bytes() 105 *ppos = i; in nvram_write_bytes()
|
| A D | rv.h | 60 for (int i = 0; i < ARRAY_SIZE(mon->states); ++i) { in rv_ltl_valid_state() local 61 if (mon->states[i]) in rv_ltl_valid_state() 69 for (int i = 0; i < ARRAY_SIZE(mon->unknown_atoms); ++i) { in rv_ltl_all_atoms_known() local 70 if (mon->unknown_atoms[i]) in rv_ltl_all_atoms_known()
|
| A D | lockdep.h | 29 int i; in lockdep_copy_map() local 40 for (i = 0; i < NR_LOCKDEP_CACHING_CLASSES; i++) in lockdep_copy_map() 513 #define spin_acquire(l, s, t, i) lock_acquire_exclusive(l, s, t, NULL, i) argument 515 #define spin_release(l, i) lock_release(l, i) argument 517 #define rwlock_acquire(l, s, t, i) lock_acquire_exclusive(l, s, t, NULL, i) argument 526 #define rwlock_release(l, i) lock_release(l, i) argument 530 #define seqcount_release(l, i) lock_release(l, i) argument 532 #define mutex_acquire(l, s, t, i) lock_acquire_exclusive(l, s, t, NULL, i) argument 534 #define mutex_release(l, i) lock_release(l, i) argument 536 #define rwsem_acquire(l, s, t, i) lock_acquire_exclusive(l, s, t, NULL, i) argument [all …]
|
| A D | blockgroup_lock.h | 29 int i; in bgl_lock_init() local 31 for (i = 0; i < NR_BG_LOCKS; i++) in bgl_lock_init() 32 spin_lock_init(&bgl->locks[i].lock); in bgl_lock_init()
|
| A D | ahci-remap.h | 19 static inline unsigned int ahci_remap_dcc(int i) in ahci_remap_dcc() argument 21 return AHCI_REMAP_N_DCC + i * 0x80; in ahci_remap_dcc() 24 static inline unsigned int ahci_remap_base(int i) in ahci_remap_base() argument 26 return AHCI_REMAP_N_OFFSET + i * AHCI_REMAP_N_SIZE; in ahci_remap_base()
|
| A D | ascii85.h | 25 int i; in ascii85_encode() local 31 for (i = 5; i--; ) { in ascii85_encode() 32 out[i] = '!' + in % 85; in ascii85_encode()
|
| A D | memblock.h | 195 for (i = 0, __next_physmem_range(&i, type, p_start, p_end); \ 196 i != (u64)ULLONG_MAX; \ 197 __next_physmem_range(&i, type, p_start, p_end)) 214 for (i = 0, __next_mem_range(&i, nid, flags, type_a, type_b, \ 216 i != (u64)ULLONG_MAX; \ 234 for (i = (u64)ULLONG_MAX, \ 237 i != (u64)ULLONG_MAX; \ 247 #define for_each_mem_range(i, p_start, p_end) \ argument 323 for (i = -1, __next_mem_pfn_range(&i, nid, p_start, p_end, p_nid); \ 324 i >= 0; __next_mem_pfn_range(&i, nid, p_start, p_end, p_nid)) [all …]
|
| /include/asm-generic/ |
| A D | local64.h | 28 #define LOCAL64_INIT(i) { LOCAL_INIT(i) } argument 31 #define local64_set(l,i) local_set((&(l)->a),(i)) argument 34 #define local64_add(i,l) local_add((i),(&(l)->a)) argument 35 #define local64_sub(i,l) local_sub((i),(&(l)->a)) argument 37 #define local64_sub_and_test(i, l) local_sub_and_test((i), (&(l)->a)) argument 41 #define local64_add_return(i, l) local_add_return((i), (&(l)->a)) argument 42 #define local64_sub_return(i, l) local_sub_return((i), (&(l)->a)) argument 75 #define LOCAL64_INIT(i) { ATOMIC_LONG_INIT(i) } argument 78 #define local64_set(l,i) atomic64_set((&(l)->a),(i)) argument 81 #define local64_add(i,l) atomic64_add((i),(&(l)->a)) argument [all …]
|
| A D | local.h | 27 #define LOCAL_INIT(i) { ATOMIC_LONG_INIT(i) } argument 30 #define local_set(l,i) atomic_long_set((&(l)->a),(i)) argument 33 #define local_add(i,l) atomic_long_add((i),(&(l)->a)) argument 34 #define local_sub(i,l) atomic_long_sub((i),(&(l)->a)) argument 36 #define local_sub_and_test(i, l) atomic_long_sub_and_test((i), (&(l)->a)) argument 39 #define local_add_negative(i, l) atomic_long_add_negative((i), (&(l)->a)) argument 40 #define local_add_return(i, l) atomic_long_add_return((i), (&(l)->a)) argument 41 #define local_sub_return(i, l) atomic_long_sub_return((i), (&(l)->a)) argument 54 #define __local_add(i,l) local_set((l), local_read(l) + (i)) argument 55 #define __local_sub(i,l) local_set((l), local_read(l) - (i)) argument
|
| A D | atomic.h | 20 static inline void generic_atomic_##op(int i, atomic_t *v) \ 25 while ((old = arch_cmpxchg(&v->counter, c, c c_op i)) != c) \ 35 while ((old = arch_cmpxchg(&v->counter, c, c c_op i)) != c) \ 38 return c c_op i; \ 42 static inline int generic_atomic_fetch_##op(int i, atomic_t *v) \ 47 while ((old = arch_cmpxchg(&v->counter, c, c c_op i)) != c) \ 58 static inline void generic_atomic_##op(int i, atomic_t *v) \ 63 v->counter = v->counter c_op i; \ 74 ret = (v->counter = v->counter c_op i); \ 88 v->counter = v->counter c_op i; \ [all …]
|
| /include/linux/mtd/ |
| A D | map.h | 263 for (i = 0; i < map_words(map); i++) \ 275 for (i = 0; i < map_words(map); i++) \ 276 r.x[i] = (val1).x[i] & (val2).x[i]; \ 284 for (i = 0; i < map_words(map); i++) \ 285 r.x[i] = (val1).x[i] & ~(val2).x[i]; \ 293 for (i = 0; i < map_words(map); i++) \ 294 r.x[i] = (val1).x[i] | (val2).x[i]; \ 301 for (i = 0; i < map_words(map); i++) { \ 313 for (i = 0; i < map_words(map); i++) { \ 353 for (i = start; i < start+len; i++) { in map_word_load_partial() [all …]
|
| /include/uapi/linux/netfilter/ |
| A D | xt_sctp.h | 73 unsigned int i; in __sctp_chunkmap_is_clear() local 74 for (i = 0; i < n; ++i) in __sctp_chunkmap_is_clear() 75 if (chunkmap[i]) in __sctp_chunkmap_is_clear() 85 unsigned int i; in __sctp_chunkmap_is_all_set() local 86 for (i = 0; i < n; ++i) in __sctp_chunkmap_is_all_set() 87 if (chunkmap[i] != ~0U) in __sctp_chunkmap_is_all_set()
|
| /include/rv/ |
| A D | ltl_monitor.h | 52 for (int i = 0; i < LTL_NUM_ATOM; ++i) in ltl_task_init() local 53 __set_bit(i, mon->unknown_atoms); in ltl_task_init() 125 int i; in ltl_trace_event() local 133 for (i = 0; i < LTL_NUM_ATOM; ++i) { in ltl_trace_event() 134 if (test_bit(i, mon->atoms)) { in ltl_trace_event() 135 seq_buf_printf(&atoms, format_str, ltl_atom_str(i)); in ltl_trace_event() 150 for (unsigned int i = 0; i < RV_NUM_BA_STATES; ++i) { in ltl_validate() local 151 if (test_bit(i, mon->states)) in ltl_validate() 152 ltl_possible_next_states(mon, i, next_states); in ltl_validate()
|
| /include/linux/byteorder/ |
| A D | generic.h | 193 size_t i; in cpu_to_be32_array() local 195 for (i = 0; i < len; i++) in cpu_to_be32_array() 196 dst[i] = cpu_to_be32(src[i]); in cpu_to_be32_array() 201 size_t i; in be32_to_cpu_array() local 203 for (i = 0; i < len; i++) in be32_to_cpu_array() 204 dst[i] = be32_to_cpu(src[i]); in be32_to_cpu_array()
|
| /include/linux/iio/ |
| A D | iio-gts-helper.h | 88 int i; in iio_gts_find_itime_by_time() local 93 for (i = 0; i < gts->num_itime; i++) in iio_gts_find_itime_by_time() 94 if (gts->itime_table[i].time_us == time) in iio_gts_find_itime_by_time() 95 return >s->itime_table[i]; in iio_gts_find_itime_by_time() 103 int i; in iio_gts_find_itime_by_sel() local 105 for (i = 0; i < gts->num_itime; i++) in iio_gts_find_itime_by_sel() 106 if (gts->itime_table[i].sel == sel) in iio_gts_find_itime_by_sel() 107 return >s->itime_table[i]; in iio_gts_find_itime_by_sel()
|
| /include/net/ |
| A D | esp.h | 23 int i; in esp_output_fill_trailer() local 24 for (i = 0; i < plen - 2; i++) in esp_output_fill_trailer() 25 tail[i] = i + 1; in esp_output_fill_trailer()
|