Lines Matching refs:ptr
36 bkey_for_each_ptr(ptrs, ptr) in bkey_put_dev_refs()
37 bch2_dev_put(bch2_dev_have_ref(c, ptr->dev)); in bkey_put_dev_refs()
44 bkey_for_each_ptr(ptrs, ptr) { in bkey_get_dev_refs()
45 if (unlikely(!bch2_dev_tryget(c, ptr->dev))) { in bkey_get_dev_refs()
47 if (ptr2 == ptr) in bkey_get_dev_refs()
61 bkey_for_each_ptr(ptrs, ptr) { in bkey_nocow_unlock()
62 struct bch_dev *ca = bch2_dev_have_ref(c, ptr->dev); in bkey_nocow_unlock()
63 struct bpos bucket = PTR_BUCKET_POS(ca, ptr); in bkey_nocow_unlock()
74 bkey_for_each_ptr(ptrs, ptr) { in __bkey_nocow_lock()
75 if (ptr == start) in __bkey_nocow_lock()
78 struct bch_dev *ca = bch2_dev_have_ref(c, ptr->dev); in __bkey_nocow_lock()
79 struct bpos bucket = PTR_BUCKET_POS(ca, ptr); in __bkey_nocow_lock()
85 __bkey_for_each_ptr(start, ptrs.end, ptr) { in __bkey_nocow_lock()
86 struct bch_dev *ca = bch2_dev_have_ref(c, ptr->dev); in __bkey_nocow_lock()
87 struct bpos bucket = PTR_BUCKET_POS(ca, ptr); in __bkey_nocow_lock()
101 bkey_for_each_ptr(ptrs, ptr) { in bkey_nocow_lock()
102 struct bch_dev *ca = bch2_dev_have_ref(c, ptr->dev); in bkey_nocow_lock()
103 struct bpos bucket = PTR_BUCKET_POS(ca, ptr); in bkey_nocow_lock()
106 return __bkey_nocow_lock(c, ctxt, ptrs, ptr); in bkey_nocow_lock()
156 struct bch_extent_ptr *ptr; in trace_io_move_fail2() local
162 (ptr = bch2_extent_has_ptr(old, p, bkey_i_to_s(insert))) && in trace_io_move_fail2()
163 !ptr->cached) in trace_io_move_fail2()
282 struct bch_extent_ptr *ptr; in __bch2_data_update_index_update() local
338 (ptr = bch2_extent_has_ptr(old, p, bkey_i_to_s(insert))) && in __bch2_data_update_index_update()
339 !ptr->cached) { in __bch2_data_update_index_update()
341 bkey_i_to_s(insert), ptr); in __bch2_data_update_index_update()
359 extent_for_each_ptr(extent_i_to_s(new), ptr) in __bch2_data_update_index_update()
360 if ((ptr_c = bch2_bkey_has_device_c(bkey_i_to_s_c(insert), ptr->dev)) && in __bch2_data_update_index_update()
362 bch2_bkey_drop_ptr_noerror(bkey_i_to_s(&new->k_i), ptr); in __bch2_data_update_index_update()
373 if ((ptr = bch2_bkey_has_device(bkey_i_to_s(insert), p.ptr.dev))) in __bch2_data_update_index_update()
374 bch2_bkey_drop_ptr_noerror(bkey_i_to_s(insert), ptr); in __bch2_data_update_index_update()
385 if (!p.ptr.cached && in __bch2_data_update_index_update()
390 bkey_i_to_s(insert), &entry->ptr); in __bch2_data_update_index_update()
601 extent_for_each_ptr(extent_i_to_s(e), ptr) in bch2_update_unwritten_extent()
602 ptr->unwritten = true; in bch2_update_unwritten_extent()
705 bch2_bkey_drop_ptrs_noerror(bkey_i_to_s(n), ptr, i++ == drop); in bch2_extent_drop_ptrs()
874 if (!p.ptr.cached) { in bch2_data_update_init()
883 bch2_dev_list_add_dev(&m->op.devs_have, p.ptr.dev); in bch2_data_update_init()
902 unwritten |= p.ptr.unwritten; in bch2_data_update_init()
1013 bkey_for_each_ptr(ptrs, ptr) { in bch2_data_update_opts_normalize()
1014 if ((opts->rewrite_ptrs & ptr_bit) && ptr->cached) { in bch2_data_update_opts_normalize()