| /fs/btrfs/ |
| A D | extent-io-tree.c | 48 state->start, state->end, state->state, in btrfs_extent_state_leak_debug_check() 157 state->state = 0; in alloc_extent_state() 347 if (prev && prev->end == state->start - 1 && prev->state == state->state) { in merge_prev_state() 362 if (next && next->start == state->end + 1 && next->state == state->state) { in merge_next_state() 442 state->state == entry->state) { in insert_state() 448 state->state = 0; in insert_state() 454 state->state == entry->state) { in insert_state() 460 state->state = 0; in insert_state() 577 if (state->state == 0) { in clear_state_bit() 874 if (state->state & bits) in find_first_extent_bit_state() [all …]
|
| /fs/nfs/ |
| A D | nfs4state.c | 667 state = kzalloc(sizeof(*state), GFP_KERNEL_ACCOUNT); in nfs4_alloc_open_state() 668 if (!state) in nfs4_alloc_open_state() 675 return state; in nfs4_alloc_open_state() 681 if (state->state == fmode) in nfs4_state_set_mode_locked() 690 state->state = fmode; in nfs4_state_set_mode_locked() 725 if (state) in nfs4_get_open_state() 732 state = new; in nfs4_get_open_state() 750 return state; in nfs4_get_open_state() 1450 state = ctx->state; in nfs_inode_find_state_and_recover() 1483 if (ctx->state != state) in nfs4_state_mark_open_context_bad() [all …]
|
| A D | nfs4proc.c | 727 .state = state, in nfs4_async_handle_error() 2123 state = data->state; in nfs4_opendata_find_nfs4_state() 2189 if (ctx->state != state) in nfs4_state_find_open_context_mode() 2225 opendata->state = state; in nfs4_open_recoverdata_alloc() 3167 ctx->state = state; in _nfs4_open_and_get_state() 3277 state = ctx->state; in _nfs4_do_open() 3474 .state = state, in nfs4_do_setattr() 3505 if (state && !(state->state & FMODE_WRITE)) { in nfs4_do_setattr() 3662 .state = state, in nfs4_close_done() 7658 .state = state, in nfs4_proc_setlk() [all …]
|
| /fs/nfsd/ |
| A D | nfs4acl.c | 466 if (!state->users) in init_state() 469 if (!state->groups) { in init_state() 470 kfree(state->users); in init_state() 478 kfree(state->users); in free_state() 479 kfree(state->groups); in free_state() 508 if (!state->users->n && !state->groups->n) in posix_state_to_acl() 511 nace = 4 + state->users->n + state->groups->n; in posix_state_to_acl() 526 add_to_mask(state, &state->users->aces[i].perms); in posix_state_to_acl() 532 add_to_mask(state, &state->group); in posix_state_to_acl() 540 add_to_mask(state, &state->groups->aces[i].perms); in posix_state_to_acl() [all …]
|
| /fs/xfs/libxfs/ |
| A D | xfs_da_btree.c | 93 return state; in xfs_da_state_alloc() 117 memset((char *)state, 0, sizeof(*state)); in xfs_da_state_free() 130 state->mp = state->args->dp->i_mount; in xfs_da_state_reset() 650 xfs_da3_fixhashpath(state, &state->path); in xfs_da3_split() 905 xfs_da3_node_add(state, oldblk, &state->extrablk); in xfs_da3_node_split() 1179 xfs_da3_fixhashpath(state, &state->path); in xfs_da3_join() 1188 xfs_da3_fixhashpath(state, &state->altpath); in xfs_da3_join() 1205 xfs_da3_fixhashpath(state, &state->path); in xfs_da3_join() 1332 blk = &state->path.blk[ state->path.active-1 ]; in xfs_da3_node_toosmall() 1353 memcpy(&state->altpath, &state->path, sizeof(state->path)); in xfs_da3_node_toosmall() [all …]
|
| A D | xfs_dir2_node.c | 950 state); in xfs_dir2_leafn_lookup_int() 1517 blk = &state->path.blk[state->path.active - 1]; in xfs_dir2_leafn_toosmall() 1544 memcpy(&state->altpath, &state->path, sizeof(state->path)); in xfs_dir2_leafn_toosmall() 1606 memcpy(&state->altpath, &state->path, sizeof(state->path)); in xfs_dir2_leafn_toosmall() 2045 state->extravalid ? &state->extrablk : NULL); in xfs_dir2_node_addname() 2049 blk = &state->path.blk[state->path.active - 1]; in xfs_dir2_node_addname() 2060 xfs_da3_fixhashpath(state, &state->path); in xfs_dir2_node_addname() 2125 if (state->extravalid && state->extrablk.bp) { in xfs_dir2_node_lookup() 2163 blk = &state->path.blk[state->path.active - 1]; in xfs_dir2_node_removename() 2177 xfs_da3_fixhashpath(state, &state->path); in xfs_dir2_node_removename() [all …]
|
| A D | xfs_attr.c | 131 path = &state->path; 146 path = &state->altpath; 179 path = &state->path; 197 path = &state->altpath; 722 ASSERT(state->path.blk[state->path.active - 1].bp != NULL); in xfs_attr_node_removename_setup() 723 ASSERT(state->path.blk[state->path.active - 1].magic == in xfs_attr_node_removename_setup() 1407 blk = &state->path.blk[state->path.active-1]; in xfs_attr_node_try_addname() 1434 xfs_da3_fixhashpath(state, &state->path); in xfs_attr_node_try_addname() 1454 blk = &state->path.blk[state->path.active-1]; in xfs_attr_node_removename() 1457 xfs_da3_fixhashpath(state, &state->path); in xfs_attr_node_removename() [all …]
|
| A D | xfs_attr_leaf.c | 1379 if (state->inleaf) { in xfs_attr3_leaf_split() 1739 args = state->args; in xfs_attr3_leaf_rebalance() 1769 state->inleaf = xfs_attr3_leaf_figure_balance(state, blk1, &ichdr1, in xfs_attr3_leaf_rebalance() 1773 state->inleaf = !state->inleaf; in xfs_attr3_leaf_rebalance() 2023 blk = &state->path.blk[ state->path.active-1 ]; in xfs_attr3_leaf_toosmall() 2046 memcpy(&state->altpath, &state->path, sizeof(state->path)); in xfs_attr3_leaf_toosmall() 2047 error = xfs_da3_path_shift(state, &state->altpath, forward, in xfs_attr3_leaf_toosmall() 2076 error = xfs_attr3_leaf_read(state->args->trans, state->args->dp, in xfs_attr3_leaf_toosmall() 2103 memcpy(&state->altpath, &state->path, sizeof(state->path)); in xfs_attr3_leaf_toosmall() 2105 error = xfs_da3_path_shift(state, &state->altpath, forward, in xfs_attr3_leaf_toosmall() [all …]
|
| /fs/bcachefs/ |
| A D | six.c | 75 atomic_or(mask, &lock->state); in six_set_bitmask() 80 if (atomic_read(&lock->state) & mask) in six_clear_bitmask() 81 atomic_and(~mask, &lock->state); in six_clear_bitmask() 160 old = atomic_read(&lock->state); in __do_six_trylock() 191 old = atomic_read(&lock->state); in __do_six_trylock() 567 u32 state; in do_six_unlock_type() local 577 state = atomic_read(&lock->state); in do_six_unlock_type() 585 state = atomic_sub_return_release(v, &lock->state); in do_six_unlock_type() 765 u32 state = atomic_read(&lock->state); in six_lock_wakeup_all() local 827 atomic_add(nr, &lock->state); in six_lock_readers_add() [all …]
|
| A D | checksum.c | 35 switch (state->type) { in bch2_checksum_init() 39 state->seed = 0; in bch2_checksum_init() 42 state->seed = U32_MAX; in bch2_checksum_init() 45 state->seed = U64_MAX; in bch2_checksum_init() 57 switch (state->type) { in bch2_checksum_final() 61 return state->seed; in bch2_checksum_final() 80 state->seed = crc32c(state->seed, data, len); in bch2_checksum_update() 84 state->seed = crc64_be(state->seed, data, len); in bch2_checksum_update() 142 state.type = type; in bch2_checksum() 198 state.type = type; in __bch2_checksum_bio() [all …]
|
| A D | bkey.c | 130 EBUG_ON(state->p >= (u64 *) k->_data + state->format->key_u64s); in pack_state_finish() 132 *state->p = state->w; in pack_state_finish() 166 state->p = next_word(state->p); in get_inc_field() 167 state->w = *state->p; in get_inc_field() 168 state->bits = 64; in get_inc_field() 173 state->w <<= bits; in get_inc_field() 190 *state->p = state->w; in __set_inc_field() 191 state->p = next_word(state->p); in __set_inc_field() 192 state->w = 0; in __set_inc_field() 193 state->bits = 64; in __set_inc_field() [all …]
|
| A D | fs-io-pagecache.c | 89 switch (state) { in folio_sector_dirty() 95 return state; in folio_sector_dirty() 102 switch (state) { in folio_sector_undirty() 108 return state; in folio_sector_undirty() 115 switch (state) { in folio_sector_reserve() 121 return state; in folio_sector_reserve() 157 unsigned nr_ptrs, unsigned state) in __bch2_folio_set() argument 241 unsigned state = bkey_to_sector_state(k); in bch2_bio_page_state_set() local 247 nr_ptrs, state); in bch2_bio_page_state_set() 686 if (s->s[i].state >= SECTOR_dirty && in folio_data_offset() [all …]
|
| /fs/quota/ |
| A D | quota.c | 121 struct qc_state state; in quota_getinfo() local 131 tstate = state.s_state + type; in quota_getinfo() 357 struct qc_state state; in quota_getstate() local 380 if (state.s_state[USRQUOTA].ino) { in quota_getstate() 385 if (state.s_state[GRPQUOTA].ino) { in quota_getstate() 390 if (state.s_state[PRJQUOTA].ino) { in quota_getstate() 399 state.s_state[PRJQUOTA].blocks; in quota_getstate() 455 struct qc_state state; in quota_getstatev() local 479 if (state.s_state[USRQUOTA].ino) { in quota_getstatev() 484 if (state.s_state[GRPQUOTA].ino) { in quota_getstatev() [all …]
|
| /fs/dlm/ |
| A D | midcomms.c | 182 int state; member 234 switch (state) { in dlm_state_str() 328 node->state = DLM_CLOSED; in midcomms_node_reset() 484 switch (node->state) { in dlm_pas_fin_ack_rcv() 496 __func__, node->state); in dlm_pas_fin_ack_rcv() 541 switch (node->state) { in dlm_midcomms_receive_buffer() 677 switch (node->state) { in dlm_midcomms_receive_buffer_3_2() 1189 switch (node->state) { in dlm_act_fin_ack_rcv() 1230 switch (node->state) { in dlm_midcomms_add_member() 1293 switch (node->state) { in dlm_midcomms_remove_member() [all …]
|
| /fs/ext4/ |
| A D | fast_commit.c | 1546 state->fc_modified_inodes[state->fc_modified_inodes_used++] = ino; in ext4_fc_record_modified_inode() 1725 if (replay && state->fc_regions_used != state->fc_regions_valid) in ext4_fc_record_regions() 1726 state->fc_regions_used = state->fc_regions_valid; in ext4_fc_record_regions() 1727 if (state->fc_regions_used == state->fc_regions_size) { in ext4_fc_record_regions() 1741 region = &state->fc_regions[state->fc_regions_used++]; in ext4_fc_record_regions() 2089 state->fc_crc = 0; in ext4_fc_replay_scan() 2091 state->fc_regions_valid = state->fc_regions_used = in ext4_fc_replay_scan() 2136 state->fc_crc = ext4_chksum(state->fc_crc, cur, in ext4_fc_replay_scan() 2142 state->fc_crc = ext4_chksum(state->fc_crc, cur, in ext4_fc_replay_scan() 2148 state->fc_replay_num_tags = state->fc_cur_tag; in ext4_fc_replay_scan() [all …]
|
| /fs/netfs/ |
| A D | fscache_io.c | 26 enum fscache_cookie_state state; in fscache_wait_for_operation() local 34 state = fscache_cookie_state(cookie); in fscache_wait_for_operation() 35 _enter("c=%08x{%u},%x", cookie->debug_id, state, want_state); in fscache_wait_for_operation() 37 switch (state) { in fscache_wait_for_operation() 45 wait_var_event(&cookie->state, in fscache_wait_for_operation() 46 fscache_cookie_state(cookie) != state); in fscache_wait_for_operation() 75 enum fscache_cookie_state state; in fscache_begin_operation() local 93 state = fscache_cookie_state(cookie); in fscache_begin_operation() 96 switch (state) { in fscache_begin_operation() 126 timeo = wait_var_event_timeout(&cookie->state, in fscache_begin_operation() [all …]
|
| /fs/jfs/ |
| A D | jfs_mount.c | 410 if (state == FM_DIRTY) { in updateSuper() 411 sbi->p_state = state; in updateSuper() 413 } else if (state == FM_MOUNT) { in updateSuper() 414 sbi->p_state = sbi->state; in updateSuper() 415 state = FM_DIRTY; in updateSuper() 416 } else if (state == FM_CLEAN) { in updateSuper() 417 state = sbi->p_state; in updateSuper() 420 } else if (sbi->state == FM_DIRTY) in updateSuper() 429 sbi->state = state; in updateSuper() 431 if (state == FM_MOUNT) { in updateSuper() [all …]
|
| /fs/xfs/ |
| A D | xfs_quotaops.c | 59 struct qc_state *state) in xfs_fs_get_quota_state() argument 65 memset(state, 0, sizeof(*state)); in xfs_fs_get_quota_state() 68 state->s_incoredqs = q->qi_dquots; in xfs_fs_get_quota_state() 70 state->s_state[USRQUOTA].flags |= QCI_ACCT_ENABLED; in xfs_fs_get_quota_state() 72 state->s_state[USRQUOTA].flags |= QCI_LIMITS_ENFORCED; in xfs_fs_get_quota_state() 74 state->s_state[GRPQUOTA].flags |= QCI_ACCT_ENABLED; in xfs_fs_get_quota_state() 76 state->s_state[GRPQUOTA].flags |= QCI_LIMITS_ENFORCED; in xfs_fs_get_quota_state() 78 state->s_state[PRJQUOTA].flags |= QCI_ACCT_ENABLED; in xfs_fs_get_quota_state() 80 state->s_state[PRJQUOTA].flags |= QCI_LIMITS_ENFORCED; in xfs_fs_get_quota_state() 82 error = xfs_qm_fill_state(&state->s_state[USRQUOTA], mp, in xfs_fs_get_quota_state() [all …]
|
| /fs/btrfs/tests/ |
| A D | extent-io-tests.c | 66 if (state->state & EXTENT_##name) \ 76 PRINT_ONE_FLAG(state, dest, cur, DIRTY); in extent_flag_to_str() 77 PRINT_ONE_FLAG(state, dest, cur, LOCKED); in extent_flag_to_str() 80 PRINT_ONE_FLAG(state, dest, cur, DELALLOC); in extent_flag_to_str() 81 PRINT_ONE_FLAG(state, dest, cur, DEFRAG); in extent_flag_to_str() 82 PRINT_ONE_FLAG(state, dest, cur, BOUNDARY); in extent_flag_to_str() 83 PRINT_ONE_FLAG(state, dest, cur, NODATASUM); in extent_flag_to_str() 96 node = rb_first(&tree->state); in dump_extent_io_tree() 99 struct extent_state *state; in dump_extent_io_tree() local 102 extent_flag_to_str(state, flags_str); in dump_extent_io_tree() [all …]
|
| /fs/smb/server/ |
| A D | oplock.c | 101 lease->state = lctx->req_state; in alloc_lease() 246 lease->state = lease->new_state; in opinfo_write_to_read() 260 lease->state = lease->new_state; in opinfo_read_handle_to_read() 284 lease->state = lease->new_state; in opinfo_write_to_none() 351 lease->state = new_state; in lease_none_upgrade() 416 lease->state = lctx->req_state; in grant_write_oplock() 458 lease->state = 0; in grant_none_oplock() 519 lease->state == (lctx->req_state & lease->state)) { in same_client_has_lease() 589 if (opinfo->o_lease->state != in oplock_break_pending() 601 if (opinfo->o_lease->state != in oplock_break_pending() [all …]
|
| /fs/afs/ |
| A D | rxrpc.c | 476 call->state = AFS_CALL_COMPLETE; in afs_make_call() 518 enum afs_call_state state; in afs_deliver_to_call() local 525 while (state = READ_ONCE(call->state), in afs_deliver_to_call() 529 state == AFS_CALL_SV_AWAIT_ACK in afs_deliver_to_call() 531 if (state == AFS_CALL_SV_AWAIT_ACK) { in afs_deliver_to_call() 551 state = READ_ONCE(call->state); in afs_deliver_to_call() 583 call->debug_id, state); in afs_deliver_to_call() 591 if (state != AFS_CALL_CL_AWAIT_REPLY) in afs_deliver_to_call() 930 enum afs_call_state state; in afs_extract_data() local 944 state = READ_ONCE(call->state); in afs_extract_data() [all …]
|
| A D | flock.c | 28 vnode->lock_state = state; in afs_set_lock_state() 100 p->fl_u.afs.state = AFS_LOCK_GRANTED; in afs_grant_locks() 127 p->fl_u.afs.state = error; in afs_next_locker() 167 p->fl_u.afs.state = -ENOENT; in afs_kill_lockers_enoent() 470 fl->fl_u.afs.state = AFS_LOCK_PENDING; in afs_do_setlk() 564 fl->fl_u.afs.state = ret; in afs_do_setlk() 571 fl->fl_u.afs.state = ret; in afs_do_setlk() 578 fl->fl_u.afs.state = ret; in afs_do_setlk() 647 if (fl->fl_u.afs.state >= 0 && fl->fl_u.afs.state != AFS_LOCK_GRANTED) { in afs_do_setlk() 650 switch (fl->fl_u.afs.state) { in afs_do_setlk() [all …]
|
| /fs/ocfs2/dlm/ |
| A D | dlmdebug.c | 612 char *state; in debug_state_print() local 620 state = "NEW"; break; in debug_state_print() 719 state = "ACTIVE"; in debug_state_print() 721 state = "INACTIVE"; in debug_state_print() 738 switch (node->state) { in debug_state_print() 740 state = "INIT"; in debug_state_print() 746 state = "DEAD"; in debug_state_print() 749 state = "RECEIVING"; in debug_state_print() 752 state = "REQUESTED"; in debug_state_print() 755 state = "DONE"; in debug_state_print() [all …]
|
| A D | dlmthread.c | 51 if (res->state & flags) { in __dlm_wait_on_lockres_flags() 90 if (res->state & (DLM_LOCK_RES_RECOVERING| in __dlm_lockres_unused() 187 res->state &= ~DLM_LOCK_RES_DROPPING_REF; in __dlm_do_purge_lockres() 212 res->state |= DLM_LOCK_RES_DROPPING_REF; in dlm_purge_lockres() 269 res->state &= ~DLM_LOCK_RES_DROPPING_REF; in dlm_purge_lockres() 320 !unused, lockres->state, in dlm_run_purge_list() 516 if (res->state & (DLM_LOCK_RES_MIGRATING | in __dlm_dirty_lockres() 524 res->state |= DLM_LOCK_RES_DIRTY; in __dlm_dirty_lockres() 747 res->state &= ~DLM_LOCK_RES_DIRTY; in dlm_thread() 753 res->state); in dlm_thread() [all …]
|
| /fs/erofs/ |
| A D | decompressor_lzma.c | 7 struct xz_dec_microlzma *state; member 39 if (strm->state) in z_erofs_lzma_exit() 40 xz_dec_microlzma_end(strm->state); in z_erofs_lzma_exit() 130 if (strm->state) in z_erofs_load_lzma_config() 131 xz_dec_microlzma_end(strm->state); in z_erofs_load_lzma_config() 132 strm->state = xz_dec_microlzma_alloc(XZ_PREALLOC, dict_size); in z_erofs_load_lzma_config() 133 if (!strm->state) in z_erofs_load_lzma_config() 181 xz_dec_microlzma_reset(strm->state, rq->inputsize, rq->outputsize, in z_erofs_lzma_decompress() 203 xz_err = xz_dec_microlzma_run(strm->state, &buf); in z_erofs_lzma_decompress()
|