Lines Matching refs:flushing
145 EBUG_ON(!trans->c->btree_write_buffer.flushing.pin.seq); in wb_flush_one()
146 EBUG_ON(trans->c->btree_write_buffer.flushing.pin.seq > wb->journal_seq); in wb_flush_one()
229 bch2_journal_pin_add(j, wb->inc.keys.data[0].journal_seq, &wb->flushing.pin, in move_keys_from_inc_to_flushing()
232 darray_resize(&wb->flushing.keys, min_t(size_t, 1U << 20, wb->flushing.keys.nr + wb->inc.keys.nr)); in move_keys_from_inc_to_flushing()
233 darray_resize(&wb->sorted, wb->flushing.keys.size); in move_keys_from_inc_to_flushing()
235 if (!wb->flushing.keys.nr && wb->sorted.size >= wb->inc.keys.nr) { in move_keys_from_inc_to_flushing()
236 swap(wb->flushing.keys, wb->inc.keys); in move_keys_from_inc_to_flushing()
240 size_t nr = min(darray_room(wb->flushing.keys), in move_keys_from_inc_to_flushing()
241 wb->sorted.size - wb->flushing.keys.nr); in move_keys_from_inc_to_flushing()
244 memcpy(&darray_top(wb->flushing.keys), in move_keys_from_inc_to_flushing()
252 wb->flushing.keys.nr += nr; in move_keys_from_inc_to_flushing()
267 BUG_ON(wb->sorted.size < wb->flushing.keys.nr); in move_keys_from_inc_to_flushing()
307 for (size_t i = 0; i < wb->flushing.keys.nr; i++) { in bch2_btree_write_buffer_flush_locked()
309 wb->sorted.data[i].btree = wb->flushing.keys.data[i].btree; in bch2_btree_write_buffer_flush_locked()
310 memcpy(&wb->sorted.data[i].pos, &wb->flushing.keys.data[i].k.k.p, sizeof(struct bpos)); in bch2_btree_write_buffer_flush_locked()
312 wb->sorted.nr = wb->flushing.keys.nr; in bch2_btree_write_buffer_flush_locked()
331 struct btree_write_buffered_key *k = &wb->flushing.keys.data[i->idx]; in bch2_btree_write_buffer_flush_locked()
339 prefetch(&wb->flushing.keys.data[n->idx]); in bch2_btree_write_buffer_flush_locked()
351 struct btree_write_buffered_key *n = &wb->flushing.keys.data[i[1].idx]; in bch2_btree_write_buffer_flush_locked()
431 trace_and_count(c, write_buffer_flush_slowpath, trans, slowpath, wb->flushing.keys.nr); in bch2_btree_write_buffer_flush_locked()
433 sort_nonatomic(wb->flushing.keys.data, in bch2_btree_write_buffer_flush_locked()
434 wb->flushing.keys.nr, in bch2_btree_write_buffer_flush_locked()
435 sizeof(wb->flushing.keys.data[0]), in bch2_btree_write_buffer_flush_locked()
438 darray_for_each(wb->flushing.keys, i) { in bch2_btree_write_buffer_flush_locked()
449 bch2_journal_pin_update(j, i->journal_seq, &wb->flushing.pin, in bch2_btree_write_buffer_flush_locked()
486 struct btree_write_buffered_key *dst = wb->flushing.keys.data; in bch2_btree_write_buffer_flush_locked()
488 darray_for_each(wb->flushing.keys, i) in bch2_btree_write_buffer_flush_locked()
491 wb->flushing.keys.nr = dst - wb->flushing.keys.data; in bch2_btree_write_buffer_flush_locked()
496 bch2_journal_pin_drop(j, &wb->flushing.pin); in bch2_btree_write_buffer_flush_locked()
497 wb->flushing.keys.nr = 0; in bch2_btree_write_buffer_flush_locked()
501 trace_write_buffer_flush(trans, wb->flushing.keys.nr, overwritten, fast, 0); in bch2_btree_write_buffer_flush_locked()
565 *did_work |= wb->inc.keys.nr || wb->flushing.keys.nr; in btree_write_buffer_flush_seq()
571 mutex_lock(&wb->flushing.lock); in btree_write_buffer_flush_seq()
573 mutex_unlock(&wb->flushing.lock); in btree_write_buffer_flush_seq()
577 (wb->flushing.pin.seq && wb->flushing.pin.seq <= max_seq))); in btree_write_buffer_flush_seq()
622 if (mutex_trylock(&wb->flushing.lock)) { in bch2_btree_write_buffer_flush_nocheck_rw()
624 mutex_unlock(&wb->flushing.lock); in bch2_btree_write_buffer_flush_nocheck_rw()
694 mutex_lock(&wb->flushing.lock); in bch2_btree_write_buffer_flush_work()
698 mutex_unlock(&wb->flushing.lock); in bch2_btree_write_buffer_flush_work()
734 if (!ret && dst->wb == &wb->flushing) in bch2_journal_key_to_wb_slowpath()
735 ret = darray_resize(&wb->sorted, wb->flushing.keys.size); in bch2_journal_key_to_wb_slowpath()
738 if (dst->wb == &c->btree_write_buffer.flushing) { in bch2_journal_key_to_wb_slowpath()
750 if (dst->wb == &wb->flushing) in bch2_journal_key_to_wb_slowpath()
751 dst->room = min(dst->room, wb->sorted.size - wb->flushing.keys.nr); in bch2_journal_key_to_wb_slowpath()
768 if (mutex_trylock(&wb->flushing.lock)) { in bch2_journal_keys_to_write_buffer_start()
778 dst->wb = &wb->flushing; in bch2_journal_keys_to_write_buffer_start()
780 mutex_unlock(&wb->flushing.lock); in bch2_journal_keys_to_write_buffer_start()
789 if (dst->wb == &wb->flushing) in bch2_journal_keys_to_write_buffer_start()
790 dst->room = min(dst->room, wb->sorted.size - wb->flushing.keys.nr); in bch2_journal_keys_to_write_buffer_start()
833 if (dst->wb == &wb->flushing) in bch2_journal_keys_to_write_buffer_end()
834 mutex_unlock(&wb->flushing.lock); in bch2_journal_keys_to_write_buffer_end()
857 return wb_keys_resize(&wb->flushing, new_size) ?: in bch2_btree_write_buffer_resize()
865 BUG_ON((wb->inc.keys.nr || wb->flushing.keys.nr) && in bch2_fs_btree_write_buffer_exit()
870 darray_exit(&wb->flushing.keys); in bch2_fs_btree_write_buffer_exit()
879 mutex_init(&wb->flushing.lock); in bch2_fs_btree_write_buffer_init_early()
891 darray_make_room(&wb->flushing.keys, initial_size) ?: in bch2_fs_btree_write_buffer_init()