Lines Matching refs:delayed_refs

2004 				struct btrfs_delayed_ref_root *delayed_refs,  in btrfs_qgroup_trace_extent_nolock()  argument
2023 lockdep_assert_held(&delayed_refs->lock); in btrfs_qgroup_trace_extent_nolock()
2026 xa_lock(&delayed_refs->dirty_extents); in btrfs_qgroup_trace_extent_nolock()
2027 existing = xa_load(&delayed_refs->dirty_extents, index); in btrfs_qgroup_trace_extent_nolock()
2033 xa_unlock(&delayed_refs->dirty_extents); in btrfs_qgroup_trace_extent_nolock()
2037 ret = __xa_store(&delayed_refs->dirty_extents, index, record, GFP_ATOMIC); in btrfs_qgroup_trace_extent_nolock()
2038 xa_unlock(&delayed_refs->dirty_extents); in btrfs_qgroup_trace_extent_nolock()
2141 struct btrfs_delayed_ref_root *delayed_refs; in btrfs_qgroup_trace_extent() local
2151 if (xa_reserve(&trans->transaction->delayed_refs.dirty_extents, index, GFP_NOFS)) { in btrfs_qgroup_trace_extent()
2156 delayed_refs = &trans->transaction->delayed_refs; in btrfs_qgroup_trace_extent()
2161 spin_lock(&delayed_refs->lock); in btrfs_qgroup_trace_extent()
2162 ret = btrfs_qgroup_trace_extent_nolock(fs_info, delayed_refs, record); in btrfs_qgroup_trace_extent()
2163 spin_unlock(&delayed_refs->lock); in btrfs_qgroup_trace_extent()
2166 xa_release(&delayed_refs->dirty_extents, index); in btrfs_qgroup_trace_extent()
3033 struct btrfs_delayed_ref_root *delayed_refs; in btrfs_qgroup_account_extents() local
3043 delayed_refs = &trans->transaction->delayed_refs; in btrfs_qgroup_account_extents()
3044 qgroup_to_skip = delayed_refs->qgroup_to_skip; in btrfs_qgroup_account_extents()
3045 xa_for_each(&delayed_refs->dirty_extents, index, record) { in btrfs_qgroup_account_extents()
3111 xa_erase(&delayed_refs->dirty_extents, index); in btrfs_qgroup_account_extents()
4890 xa_for_each(&trans->delayed_refs.dirty_extents, index, entry) { in btrfs_qgroup_destroy_extent_records()
4894 xa_destroy(&trans->delayed_refs.dirty_extents); in btrfs_qgroup_destroy_extent_records()