Lines Matching refs:flushing
146 EBUG_ON(!trans->c->btree_write_buffer.flushing.pin.seq); in wb_flush_one()
147 EBUG_ON(trans->c->btree_write_buffer.flushing.pin.seq > wb->journal_seq); in wb_flush_one()
228 bch2_journal_pin_add(j, wb->inc.keys.data[0].journal_seq, &wb->flushing.pin, in move_keys_from_inc_to_flushing()
231 darray_resize(&wb->flushing.keys, min_t(size_t, 1U << 20, wb->flushing.keys.nr + wb->inc.keys.nr)); in move_keys_from_inc_to_flushing()
232 darray_resize(&wb->sorted, wb->flushing.keys.size); in move_keys_from_inc_to_flushing()
234 if (!wb->flushing.keys.nr && wb->sorted.size >= wb->inc.keys.nr) { in move_keys_from_inc_to_flushing()
235 swap(wb->flushing.keys, wb->inc.keys); in move_keys_from_inc_to_flushing()
239 size_t nr = min(darray_room(wb->flushing.keys), in move_keys_from_inc_to_flushing()
240 wb->sorted.size - wb->flushing.keys.nr); in move_keys_from_inc_to_flushing()
243 memcpy(&darray_top(wb->flushing.keys), in move_keys_from_inc_to_flushing()
251 wb->flushing.keys.nr += nr; in move_keys_from_inc_to_flushing()
266 BUG_ON(wb->sorted.size < wb->flushing.keys.nr); in move_keys_from_inc_to_flushing()
291 for (size_t i = 0; i < wb->flushing.keys.nr; i++) { in bch2_btree_write_buffer_flush_locked()
293 wb->sorted.data[i].btree = wb->flushing.keys.data[i].btree; in bch2_btree_write_buffer_flush_locked()
294 memcpy(&wb->sorted.data[i].pos, &wb->flushing.keys.data[i].k.k.p, sizeof(struct bpos)); in bch2_btree_write_buffer_flush_locked()
296 wb->sorted.nr = wb->flushing.keys.nr; in bch2_btree_write_buffer_flush_locked()
315 struct btree_write_buffered_key *k = &wb->flushing.keys.data[i->idx]; in bch2_btree_write_buffer_flush_locked()
318 prefetch(&wb->flushing.keys.data[n->idx]); in bch2_btree_write_buffer_flush_locked()
330 struct btree_write_buffered_key *n = &wb->flushing.keys.data[i[1].idx]; in bch2_btree_write_buffer_flush_locked()
410 trace_and_count(c, write_buffer_flush_slowpath, trans, slowpath, wb->flushing.keys.nr); in bch2_btree_write_buffer_flush_locked()
412 sort(wb->flushing.keys.data, in bch2_btree_write_buffer_flush_locked()
413 wb->flushing.keys.nr, in bch2_btree_write_buffer_flush_locked()
414 sizeof(wb->flushing.keys.data[0]), in bch2_btree_write_buffer_flush_locked()
417 darray_for_each(wb->flushing.keys, i) { in bch2_btree_write_buffer_flush_locked()
428 bch2_journal_pin_update(j, i->journal_seq, &wb->flushing.pin, in bch2_btree_write_buffer_flush_locked()
465 struct btree_write_buffered_key *dst = wb->flushing.keys.data; in bch2_btree_write_buffer_flush_locked()
467 darray_for_each(wb->flushing.keys, i) in bch2_btree_write_buffer_flush_locked()
470 wb->flushing.keys.nr = dst - wb->flushing.keys.data; in bch2_btree_write_buffer_flush_locked()
475 bch2_journal_pin_drop(j, &wb->flushing.pin); in bch2_btree_write_buffer_flush_locked()
476 wb->flushing.keys.nr = 0; in bch2_btree_write_buffer_flush_locked()
480 trace_write_buffer_flush(trans, wb->flushing.keys.nr, overwritten, fast, 0); in bch2_btree_write_buffer_flush_locked()
510 *did_work |= wb->inc.keys.nr || wb->flushing.keys.nr; in btree_write_buffer_flush_seq()
516 mutex_lock(&wb->flushing.lock); in btree_write_buffer_flush_seq()
518 mutex_unlock(&wb->flushing.lock); in btree_write_buffer_flush_seq()
522 (wb->flushing.pin.seq && wb->flushing.pin.seq <= seq))); in btree_write_buffer_flush_seq()
567 if (mutex_trylock(&wb->flushing.lock)) { in bch2_btree_write_buffer_flush_nocheck_rw()
569 mutex_unlock(&wb->flushing.lock); in bch2_btree_write_buffer_flush_nocheck_rw()
628 mutex_lock(&wb->flushing.lock); in bch2_btree_write_buffer_flush_work()
632 mutex_unlock(&wb->flushing.lock); in bch2_btree_write_buffer_flush_work()
668 if (!ret && dst->wb == &wb->flushing) in bch2_journal_key_to_wb_slowpath()
669 ret = darray_resize(&wb->sorted, wb->flushing.keys.size); in bch2_journal_key_to_wb_slowpath()
672 if (dst->wb == &c->btree_write_buffer.flushing) { in bch2_journal_key_to_wb_slowpath()
684 if (dst->wb == &wb->flushing) in bch2_journal_key_to_wb_slowpath()
685 dst->room = min(dst->room, wb->sorted.size - wb->flushing.keys.nr); in bch2_journal_key_to_wb_slowpath()
702 if (mutex_trylock(&wb->flushing.lock)) { in bch2_journal_keys_to_write_buffer_start()
712 dst->wb = &wb->flushing; in bch2_journal_keys_to_write_buffer_start()
714 mutex_unlock(&wb->flushing.lock); in bch2_journal_keys_to_write_buffer_start()
723 if (dst->wb == &wb->flushing) in bch2_journal_keys_to_write_buffer_start()
724 dst->room = min(dst->room, wb->sorted.size - wb->flushing.keys.nr); in bch2_journal_keys_to_write_buffer_start()
767 if (dst->wb == &wb->flushing) in bch2_journal_keys_to_write_buffer_end()
768 mutex_unlock(&wb->flushing.lock); in bch2_journal_keys_to_write_buffer_end()
816 return wb_keys_resize(&wb->flushing, new_size) ?: in bch2_btree_write_buffer_resize()
824 BUG_ON((wb->inc.keys.nr || wb->flushing.keys.nr) && in bch2_fs_btree_write_buffer_exit()
829 darray_exit(&wb->flushing.keys); in bch2_fs_btree_write_buffer_exit()
838 mutex_init(&wb->flushing.lock); in bch2_fs_btree_write_buffer_init()
845 darray_make_room(&wb->flushing.keys, initial_size) ?: in bch2_fs_btree_write_buffer_init()