| /fs/bcachefs/ |
| A D | btree_journal_iter.c | 36 return idx; in idx_to_pos() 85 if (!*idx) in bch2_journal_keys_peek_max() 90 --(*idx); in bch2_journal_keys_peek_max() 93 *idx = 0; in bch2_journal_keys_peek_max() 118 (*idx)++; in bch2_journal_keys_peek_max() 121 *idx = 0; in bch2_journal_keys_peek_max() 144 if (!*idx) in bch2_journal_keys_peek_prev_min() 149 (*idx)++; in bch2_journal_keys_peek_prev_min() 152 *idx = 0; in bch2_journal_keys_peek_prev_min() 158 --(*idx); in bch2_journal_keys_peek_prev_min() [all …]
|
| A D | fast_list.c | 25 if (unlikely(idx < 0)) in fast_list_alloc_idx() 33 return idx; in fast_list_alloc_idx() 49 int idx; in fast_list_get_idx() local 60 entries[nr++] = idx; in fast_list_get_idx() 81 idx = lp->entries[--lp->nr]; in fast_list_get_idx() 84 return idx; in fast_list_get_idx() 100 if (idx < 0) in fast_list_add() 101 return idx; in fast_list_add() 104 return idx; in fast_list_add() 120 if (!idx) in fast_list_remove() [all …]
|
| A D | journal.h | 147 return j->buf + idx; in journal_cur_buf() 152 switch (idx) { in journal_state_count() 172 s->buf0_count += s->idx == 0; in journal_state_inc() 173 s->buf1_count += s->idx == 1; in journal_state_inc() 174 s->buf2_count += s->idx == 2; in journal_state_inc() 175 s->buf3_count += s->idx == 3; in journal_state_inc() 271 .buf0_count = idx == 0, in journal_state_buf_put() 272 .buf1_count = idx == 1, in journal_state_buf_put() 273 .buf2_count = idx == 2, in journal_state_buf_put() 274 .buf3_count = idx == 3, in journal_state_buf_put() [all …]
|
| A D | async_objs.h | 6 static inline void __async_object_list_del(struct fast_list *head, unsigned idx) in __async_object_list_del() argument 8 fast_list_remove(head, idx); in __async_object_list_del() 11 static inline int __async_object_list_add(struct fast_list *head, void *obj, unsigned *idx) in __async_object_list_add() argument 14 *idx = ret > 0 ? ret : 0; in __async_object_list_add() 18 #define async_object_list_del(_c, _list, idx) \ argument 19 __async_object_list_del(&(_c)->async_objs[BCH_ASYNC_OBJ_LIST_##_list].list, idx) 21 #define async_object_list_add(_c, _list, obj, idx) \ argument 22 __async_object_list_add(&(_c)->async_objs[BCH_ASYNC_OBJ_LIST_##_list].list, obj, idx) 30 #define async_object_list_del(_c, _n, idx) do {} while (0) argument 36 #define async_object_list_add(_c, _n, obj, idx) __async_object_list_add() argument
|
| A D | enumerated_ref.c | 10 void enumerated_ref_get(struct enumerated_ref *ref, unsigned idx) in enumerated_ref_get() argument 12 BUG_ON(idx >= ref->nr); in enumerated_ref_get() 13 atomic_long_inc(&ref->refs[idx]); in enumerated_ref_get() 18 BUG_ON(idx >= ref->nr); in __enumerated_ref_tryget() 19 return atomic_long_inc_not_zero(&ref->refs[idx]); in __enumerated_ref_tryget() 22 bool enumerated_ref_tryget(struct enumerated_ref *ref, unsigned idx) in enumerated_ref_tryget() argument 24 BUG_ON(idx >= ref->nr); in enumerated_ref_tryget() 26 atomic_long_inc_not_zero(&ref->refs[idx]); in enumerated_ref_tryget() 29 void enumerated_ref_put(struct enumerated_ref *ref, unsigned idx) in enumerated_ref_put() argument 31 BUG_ON(idx >= ref->nr); in enumerated_ref_put() [all …]
|
| A D | bset.c | 360 unsigned idx) in bkey_float() argument 556 idx++; in rw_aux_tree_bsearch() 558 while (idx && in rw_aux_tree_bsearch() 560 idx--; in rw_aux_tree_bsearch() 564 EBUG_ON(idx && rw_aux_tree(b, t)[idx - 1].offset >= offset); in rw_aux_tree_bsearch() 569 return idx; in rw_aux_tree_bsearch() 898 unsigned idx) in rw_aux_tree_insert_entry() argument 900 EBUG_ON(!idx || idx > t->size); in rw_aux_tree_insert_entry() 952 idx++; in bch2_bset_fix_lookup_table() 962 EBUG_ON(idx < t->size && rw_aux_tree(b, t)[idx].offset <= where); in bch2_bset_fix_lookup_table() [all …]
|
| A D | btree_iter.h | 23 unsigned idx = path - trans->paths; in __btree_path_get() local 25 EBUG_ON(idx >= trans->nr_paths); in __btree_path_get() 112 while (*idx < trans->nr_paths) { in __trans_next_path() 115 *idx += __ffs(v); in __trans_next_path() 116 return trans->paths + *idx; in __trans_next_path() 119 *idx += BITS_PER_LONG; in __trans_next_path() 120 *idx &= ~(BITS_PER_LONG - 1); in __trans_next_path() 145 return idx < trans->nr_sorted in next_btree_path() 154 return idx in prev_btree_path() 191 unsigned *idx) in __trans_next_path_with_node() argument [all …]
|
| A D | ec.c | 108 size_t idx; member 408 idx, in bch2_trigger_stripe() 517 entry->stripe_ptr.idx == idx) in extent_has_stripe_ptr() 782 ec_bio->idx = idx; in ec_block_io() 932 if (s->idx == idx) in __bch2_stripe_is_open() 950 u64 idx) in bch2_try_open_stripe() argument 959 s->idx = idx; in bch2_try_open_stripe() 969 BUG_ON(!s->idx); in bch2_stripe_close() 975 s->idx = 0; in bch2_stripe_close() 1282 if (s->idx) in bch2_ec_stripe_new_free() [all …]
|
| A D | enumerated_ref.h | 20 static inline void enumerated_ref_get(struct enumerated_ref *ref, unsigned idx) in enumerated_ref_get() argument 25 static inline bool __enumerated_ref_tryget(struct enumerated_ref *ref, unsigned idx) in __enumerated_ref_tryget() argument 30 static inline bool enumerated_ref_tryget(struct enumerated_ref *ref, unsigned idx) in enumerated_ref_tryget() argument 35 static inline void enumerated_ref_put(struct enumerated_ref *ref, unsigned idx) in enumerated_ref_put() argument
|
| /fs/ubifs/ |
| A D | commit.c | 512 struct ubifs_idx_node idx __aligned(8); 527 struct ubifs_idx_node *idx; in dbg_old_index_check_init() local 537 if (!idx) in dbg_old_index_check_init() 547 kfree(idx); in dbg_old_index_check_init() 571 struct ubifs_idx_node *idx; in dbg_check_old_index() local 607 idx = &i->idx; in dbg_check_old_index() 612 child_cnt = le16_to_cpu(idx->child_cnt); in dbg_check_old_index() 665 if (le16_to_cpu(idx->level) == 0) { in dbg_check_old_index() 677 idx = &i->idx; in dbg_check_old_index() 693 last_level = le16_to_cpu(idx->level); in dbg_check_old_index() [all …]
|
| A D | tnc_misc.c | 294 struct ubifs_idx_node *idx; in read_znode() local 296 idx = kmalloc(c->max_idx_node_sz, GFP_NOFS); in read_znode() 297 if (!idx) in read_znode() 302 kfree(idx); in read_znode() 306 err = ubifs_node_check_hash(c, idx, zzbr->hash); in read_znode() 309 kfree(idx); in read_znode() 313 znode->child_cnt = le16_to_cpu(idx->child_cnt); in read_znode() 314 znode->level = le16_to_cpu(idx->level); in read_znode() 410 kfree(idx); in read_znode() 415 ubifs_dump_node(c, idx, c->max_idx_node_sz); in read_znode() [all …]
|
| A D | tnc_commit.c | 33 idx->ch.node_type = UBIFS_IDX_NODE; in make_idx_node() 35 idx->level = cpu_to_le16(znode->level); in make_idx_node() 54 ubifs_prepare_node(c, idx, len, 0); in make_idx_node() 55 ubifs_node_calc_hash(c, idx, hash); in make_idx_node() 252 struct ubifs_idx_node *idx; in layout_leb_in_gaps() local 256 idx = snod->node; in layout_leb_in_gaps() 258 level = le16_to_cpu(idx->level); in layout_leb_in_gaps() 828 struct ubifs_idx_node *idx; in write_index() local 865 idx = c->cbuf + used; in write_index() 868 idx->ch.node_type = UBIFS_IDX_NODE; in write_index() [all …]
|
| /fs/xfs/scrub/ |
| A D | xfarray.h | 15 #define foreach_xfarray_idx(array, idx) \ argument 16 for ((idx) = XFARRAY_CURSOR_INIT; \ 17 (idx) < xfarray_length(array); \ 18 (idx)++) 43 int xfarray_load(struct xfarray *array, xfarray_idx_t idx, void *ptr); 44 int xfarray_unset(struct xfarray *array, xfarray_idx_t idx); 45 int xfarray_store(struct xfarray *array, xfarray_idx_t idx, const void *ptr); 58 uint64_t idx, in xfarray_load_sparse() argument 61 int error = xfarray_load(array, idx, rec); in xfarray_load_sparse() 89 xfarray_idx_t *idx, in xfarray_iter() argument [all …]
|
| A D | xfarray.c | 133 xfarray_idx_t idx, in xfarray_load() argument 136 if (idx >= array->nr) in xfarray_load() 169 xfarray_idx_t idx) in xfarray_unset() argument 202 xfarray_idx_t idx, in xfarray_store() argument 341 xfarray_idx_t *idx, in xfarray_load_next() argument 367 *idx = cur; in xfarray_load_next() 389 xfarray_idx_t idx, in xfarray_sort_load() argument 400 xfarray_idx_t idx, in xfarray_sort_store() argument 698 idx = *idxp; in xfarray_qsort_pivot() 820 (idx < si->first_folio_idx || idx > si->last_folio_idx)) in xfarray_sort_scan() [all …]
|
| /fs/dlm/ |
| A D | midcomms.c | 209 int idx; member 662 int ret, idx; in dlm_midcomms_receive_buffer_3_2() local 789 int idx; in dlm_midcomms_receive_buffer_3_1() local 922 int idx, ret; in dlm_midcomms_unack_msg_resend() local 1008 int idx; in dlm_midcomms_get_mhandle() local 1024 mh->idx = idx; in dlm_midcomms_get_mhandle() 1164 int i, idx; in dlm_midcomms_exit() local 1217 int idx; in dlm_midcomms_add_member() local 1262 int idx; in dlm_midcomms_remove_member() local 1394 int i, idx; in dlm_midcomms_shutdown() local [all …]
|
| A D | lowcomms.c | 368 int idx; in nodeid_to_addr() local 425 int i, idx, addr_i; in addr_to_nodeid() local 467 int idx; in dlm_lowcomms_addr() local 551 int idx; in dlm_lowcomms_connect_node() local 576 int idx; in dlm_lowcomms_nodes_set_mark() local 985 int len, idx, result, nodeid; in accept_from_sock() local 1265 int idx; in dlm_lowcomms_new_msg() local 1291 msg->idx = idx; in dlm_lowcomms_new_msg() 1447 int idx; in dlm_lowcomms_close() local 1720 int i, idx; in dlm_lowcomms_shutdown() local [all …]
|
| /fs/squashfs/ |
| A D | block.c | 88 int idx = 0; in squashfs_bio_read_cached() local 99 idx++; in squashfs_bio_read_cached() 112 if (idx == 0 && index != read_start) in squashfs_bio_read_cached() 119 cache_folios[idx] = folio; in squashfs_bio_read_cached() 122 if (!bio || idx != end_idx) { in squashfs_bio_read_cached() 134 start_idx = idx; in squashfs_bio_read_cached() 137 idx++; in squashfs_bio_read_cached() 138 end_idx = idx; in squashfs_bio_read_cached() 178 for (idx = 0; idx < page_count; idx++) { in squashfs_bio_read_cached() 179 if (!cache_folios[idx]) in squashfs_bio_read_cached() [all …]
|
| /fs/f2fs/ |
| A D | iostat.c | 90 int io, idx; in __record_iostat_latency() local 96 for (idx = 0; idx < MAX_IO_TYPE; idx++) { in __record_iostat_latency() 98 iostat_lat[idx][io].peak_lat = in __record_iostat_latency() 99 jiffies_to_msecs(io_lat->peak_lat[idx][io]); in __record_iostat_latency() 100 iostat_lat[idx][io].cnt = io_lat->bio_cnt[idx][io]; in __record_iostat_latency() 101 iostat_lat[idx][io].avg_lat = iostat_lat[idx][io].cnt ? in __record_iostat_latency() 102 jiffies_to_msecs(io_lat->sum_lat[idx][io]) / iostat_lat[idx][io].cnt : 0; in __record_iostat_latency() 103 io_lat->sum_lat[idx][io] = 0; in __record_iostat_latency() 104 io_lat->peak_lat[idx][io] = 0; in __record_iostat_latency() 105 io_lat->bio_cnt[idx][io] = 0; in __record_iostat_latency()
|
| /fs/resctrl/ |
| A D | monitor.c | 107 entry = &rmid_ptrs[idx]; in __rmid_entry() 138 u32 idx, cur_idx = 1; in __check_limbo() local 158 if (idx >= idx_limit) in __check_limbo() 161 entry = __rmid_entry(idx); in __check_limbo() 185 cur_idx = idx + 1; in __check_limbo() 293 u32 idx; in add_rmid_to_limbo() local 337 entry = __rmid_entry(idx); in free_rmid() 352 return &d->mbm_total[idx]; in get_mbm_state() 354 return &d->mbm_local[idx]; in get_mbm_state() 773 u32 idx; in dom_data_init() local [all …]
|
| /fs/btrfs/ |
| A D | accessors.c | 56 const unsigned long idx = get_eb_folio_index(eb, member_offset);\ 59 char *kaddr = folio_address(eb->folios[idx]) + oif; \ 73 kaddr = folio_address(eb->folios[idx + 1]); \ 77 folio_address(eb->folios[idx + 1]), \ 86 const unsigned long idx = get_eb_folio_index(eb, member_offset);\ 89 char *kaddr = folio_address(eb->folios[idx]) + oif; \ 105 kaddr = folio_address(eb->folios[idx + 1]); \ 109 kaddr = folio_address(eb->folios[idx + 1]); \
|
| /fs/udf/ |
| A D | unicode.c | 154 int idx, len; in udf_name_from_CS0() local 199 for (idx = ocu_len - u_ch, ext_i_len = 0; in udf_name_from_CS0() 201 idx -= u_ch, ext_i_len++) { in udf_name_from_CS0() 202 c = ocu[idx]; in udf_name_from_CS0() 204 c = (c << 8) | ocu[idx + 1]; in udf_name_from_CS0() 208 i_ext = idx; in udf_name_from_CS0() 216 idx = i_ext + u_ch; in udf_name_from_CS0() 218 ocu, ocu_len, &idx, in udf_name_from_CS0() 227 idx = 0; in udf_name_from_CS0() 229 if (translate && (idx == i_ext)) { in udf_name_from_CS0() [all …]
|
| /fs/nfs/blocklayout/ |
| A D | dev.c | 295 struct pnfs_block_volume *volumes, int idx, gfp_t gfp_mask) in bl_parse_simple() argument 297 struct pnfs_block_volume *v = &volumes[idx]; in bl_parse_simple() 388 struct pnfs_block_volume *v = &volumes[idx]; in bl_parse_scsi() 442 struct pnfs_block_volume *v = &volumes[idx]; in bl_parse_slice() 458 struct pnfs_block_volume *v = &volumes[idx]; in bl_parse_concat() 487 struct pnfs_block_volume *v = &volumes[idx]; in bl_parse_stripe() 516 d->type = volumes[idx].type; in bl_parse_deviceid() 520 return bl_parse_simple(server, d, volumes, idx, gfp_mask); in bl_parse_deviceid() 522 return bl_parse_slice(server, d, volumes, idx, gfp_mask); in bl_parse_deviceid() 524 return bl_parse_concat(server, d, volumes, idx, gfp_mask); in bl_parse_deviceid() [all …]
|
| /fs/coda/ |
| A D | inode.c | 110 int idx; member 126 int idx; in coda_set_idx() local 132 idx = iminor(inode); in coda_set_idx() 133 if (idx < 0 || idx >= MAX_CODADEVS) in coda_set_idx() 135 ctx->idx = idx; in coda_set_idx() 207 infof(fc, "coda: device index: %i\n", ctx->idx); in coda_fill_super() 209 vc = &coda_comms[ctx->idx]; in coda_fill_super()
|
| /fs/overlayfs/ |
| A D | export.c | 89 return ovl_lowerstack(oe)->layer->idx; in ovl_connectable_layer() 111 origin_layer = ovl_lowerstack(oe)->layer->idx; in ovl_connect_layer() 341 if (!idx) in ovl_dentry_real_at() 345 if (lowerstack[i].layer->idx == idx) in ovl_dentry_real_at() 377 if (ovl_dentry_real_at(connected, layer->idx) != parent) in ovl_lookup_real_one() 399 } else if (ovl_dentry_real_at(this, layer->idx) != real) { in ovl_lookup_real_one() 412 real, layer->idx, connected, err); in ovl_lookup_real_one() 437 inode = ovl_lookup_inode(sb, real, !layer->idx); in ovl_lookup_real_inode() 477 if (ovl_dentry_real_at(this, layer->idx) != real) { in ovl_lookup_real_inode() 557 layer->idx); in ovl_lookup_real() [all …]
|
| /fs/hpfs/ |
| A D | alloc.c | 492 while (tstbits(bmp, *idx, 1)) { in find_run() 493 (*idx)++; in find_run() 494 if (unlikely(*idx >= 0x4000)) in find_run() 498 while (!tstbits(bmp, *idx + len, 1)) in find_run() 529 unsigned idx, len, start_bmp, end_bmp; in hpfs_trim_fs() local 550 idx = 0; in hpfs_trim_fs() 551 while ((len = find_run(bmp, &idx)) && !err) { in hpfs_trim_fs() 553 idx += len; in hpfs_trim_fs() 571 idx = 0; in hpfs_trim_fs() 572 while ((len = find_run(bmp, &idx)) && !err) { in hpfs_trim_fs() [all …]
|