Lines Matching refs:sdp
35 static void gfs2_log_shutdown(struct gfs2_sbd *sdp);
48 unsigned int gfs2_struct2blk(struct gfs2_sbd *sdp, unsigned int nstruct) in gfs2_struct2blk() argument
55 first = sdp->sd_ldptrs; in gfs2_struct2blk()
59 second = sdp->sd_inptrs; in gfs2_struct2blk()
91 static int gfs2_ail1_start_one(struct gfs2_sbd *sdp, in gfs2_ail1_start_one() argument
94 __releases(&sdp->sd_ail_lock) in gfs2_ail1_start_one()
95 __acquires(&sdp->sd_ail_lock) in gfs2_ail1_start_one()
106 gfs2_assert(sdp, bd->bd_tr == tr); in gfs2_ail1_start_one()
114 if (!cmpxchg(&sdp->sd_log_error, 0, -EIO)) { in gfs2_ail1_start_one()
115 gfs2_io_error_bh(sdp, bh); in gfs2_ail1_start_one()
116 gfs2_withdraw_delayed(sdp); in gfs2_ail1_start_one()
120 if (gfs2_withdrawn(sdp)) { in gfs2_ail1_start_one()
133 spin_unlock(&sdp->sd_ail_lock); in gfs2_ail1_start_one()
140 spin_lock(&sdp->sd_ail_lock); in gfs2_ail1_start_one()
151 static void dump_ail_list(struct gfs2_sbd *sdp) in dump_ail_list() argument
157 list_for_each_entry_reverse(tr, &sdp->sd_ail1_list, tr_list) { in dump_ail_list()
161 fs_err(sdp, "bd %p: blk:0x%llx bh=%p ", bd, in dump_ail_list()
164 fs_err(sdp, "\n"); in dump_ail_list()
167 fs_err(sdp, "0x%llx up2:%d dirt:%d lkd:%d req:%d " in dump_ail_list()
192 void gfs2_ail1_flush(struct gfs2_sbd *sdp, struct writeback_control *wbc) in gfs2_ail1_flush() argument
194 struct list_head *head = &sdp->sd_ail1_list; in gfs2_ail1_flush()
200 trace_gfs2_ail_flush(sdp, wbc, 1); in gfs2_ail1_flush()
202 spin_lock(&sdp->sd_ail_lock); in gfs2_ail1_flush()
206 fs_err(sdp, "Error: In %s for ten minutes! t=%d\n", in gfs2_ail1_flush()
208 dump_ail_list(sdp); in gfs2_ail1_flush()
214 ret = gfs2_ail1_start_one(sdp, wbc, tr, &plug); in gfs2_ail1_flush()
222 spin_unlock(&sdp->sd_ail_lock); in gfs2_ail1_flush()
225 gfs2_lm(sdp, "gfs2_ail1_start_one (generic_writepages) " in gfs2_ail1_flush()
227 gfs2_withdraw(sdp); in gfs2_ail1_flush()
229 trace_gfs2_ail_flush(sdp, wbc, 0); in gfs2_ail1_flush()
237 static void gfs2_ail1_start(struct gfs2_sbd *sdp) in gfs2_ail1_start() argument
246 return gfs2_ail1_flush(sdp, &wbc); in gfs2_ail1_start()
249 static void gfs2_log_update_flush_tail(struct gfs2_sbd *sdp) in gfs2_log_update_flush_tail() argument
251 unsigned int new_flush_tail = sdp->sd_log_head; in gfs2_log_update_flush_tail()
254 if (!list_empty(&sdp->sd_ail1_list)) { in gfs2_log_update_flush_tail()
255 tr = list_last_entry(&sdp->sd_ail1_list, in gfs2_log_update_flush_tail()
259 sdp->sd_log_flush_tail = new_flush_tail; in gfs2_log_update_flush_tail()
262 static void gfs2_log_update_head(struct gfs2_sbd *sdp) in gfs2_log_update_head() argument
264 unsigned int new_head = sdp->sd_log_flush_head; in gfs2_log_update_head()
266 if (sdp->sd_log_flush_tail == sdp->sd_log_head) in gfs2_log_update_head()
267 sdp->sd_log_flush_tail = new_head; in gfs2_log_update_head()
268 sdp->sd_log_head = new_head; in gfs2_log_update_head()
275 static void gfs2_ail_empty_tr(struct gfs2_sbd *sdp, struct gfs2_trans *tr, in gfs2_ail_empty_tr() argument
283 gfs2_assert(sdp, bd->bd_tr == tr); in gfs2_ail_empty_tr()
297 static int gfs2_ail1_empty_one(struct gfs2_sbd *sdp, struct gfs2_trans *tr, in gfs2_ail1_empty_one() argument
307 gfs2_assert(sdp, bd->bd_tr == tr); in gfs2_ail1_empty_one()
317 if (!sdp->sd_log_error && buffer_busy(bh)) { in gfs2_ail1_empty_one()
322 !cmpxchg(&sdp->sd_log_error, 0, -EIO)) { in gfs2_ail1_empty_one()
323 gfs2_io_error_bh(sdp, bh); in gfs2_ail1_empty_one()
324 gfs2_withdraw_delayed(sdp); in gfs2_ail1_empty_one()
333 gfs2_add_revoke(sdp, bd); in gfs2_ail1_empty_one()
350 static int gfs2_ail1_empty(struct gfs2_sbd *sdp, int max_revokes) in gfs2_ail1_empty() argument
356 spin_lock(&sdp->sd_ail_lock); in gfs2_ail1_empty()
357 list_for_each_entry_safe_reverse(tr, s, &sdp->sd_ail1_list, tr_list) { in gfs2_ail1_empty()
358 if (!gfs2_ail1_empty_one(sdp, tr, &max_revokes) && oldest_tr) in gfs2_ail1_empty()
359 list_move(&tr->tr_list, &sdp->sd_ail2_list); in gfs2_ail1_empty()
363 gfs2_log_update_flush_tail(sdp); in gfs2_ail1_empty()
364 ret = list_empty(&sdp->sd_ail1_list); in gfs2_ail1_empty()
365 spin_unlock(&sdp->sd_ail_lock); in gfs2_ail1_empty()
367 if (test_bit(SDF_WITHDRAWING, &sdp->sd_flags)) { in gfs2_ail1_empty()
368 gfs2_lm(sdp, "fatal: I/O error(s)\n"); in gfs2_ail1_empty()
369 gfs2_withdraw(sdp); in gfs2_ail1_empty()
375 static void gfs2_ail1_wait(struct gfs2_sbd *sdp) in gfs2_ail1_wait() argument
381 spin_lock(&sdp->sd_ail_lock); in gfs2_ail1_wait()
382 list_for_each_entry_reverse(tr, &sdp->sd_ail1_list, tr_list) { in gfs2_ail1_wait()
388 spin_unlock(&sdp->sd_ail_lock); in gfs2_ail1_wait()
394 spin_unlock(&sdp->sd_ail_lock); in gfs2_ail1_wait()
397 static void __ail2_empty(struct gfs2_sbd *sdp, struct gfs2_trans *tr) in __ail2_empty() argument
399 gfs2_ail_empty_tr(sdp, tr, &tr->tr_ail2_list); in __ail2_empty()
401 gfs2_assert_warn(sdp, list_empty(&tr->tr_ail1_list)); in __ail2_empty()
402 gfs2_assert_warn(sdp, list_empty(&tr->tr_ail2_list)); in __ail2_empty()
403 gfs2_trans_free(sdp, tr); in __ail2_empty()
406 static void ail2_empty(struct gfs2_sbd *sdp, unsigned int new_tail) in ail2_empty() argument
408 struct list_head *ail2_list = &sdp->sd_ail2_list; in ail2_empty()
409 unsigned int old_tail = sdp->sd_log_tail; in ail2_empty()
412 spin_lock(&sdp->sd_ail_lock); in ail2_empty()
416 __ail2_empty(sdp, tr); in ail2_empty()
421 __ail2_empty(sdp, tr); in ail2_empty()
424 spin_unlock(&sdp->sd_ail_lock); in ail2_empty()
432 bool gfs2_log_is_empty(struct gfs2_sbd *sdp) { in gfs2_log_is_empty() argument
433 return atomic_read(&sdp->sd_log_blks_free) == sdp->sd_jdesc->jd_blocks; in gfs2_log_is_empty()
436 static bool __gfs2_log_try_reserve_revokes(struct gfs2_sbd *sdp, unsigned int revokes) in __gfs2_log_try_reserve_revokes() argument
440 available = atomic_read(&sdp->sd_log_revokes_available); in __gfs2_log_try_reserve_revokes()
442 if (atomic_try_cmpxchg(&sdp->sd_log_revokes_available, in __gfs2_log_try_reserve_revokes()
456 void gfs2_log_release_revokes(struct gfs2_sbd *sdp, unsigned int revokes) in gfs2_log_release_revokes() argument
459 atomic_add(revokes, &sdp->sd_log_revokes_available); in gfs2_log_release_revokes()
469 void gfs2_log_release(struct gfs2_sbd *sdp, unsigned int blks) in gfs2_log_release() argument
471 atomic_add(blks, &sdp->sd_log_blks_free); in gfs2_log_release()
472 trace_gfs2_log_blocks(sdp, blks); in gfs2_log_release()
473 gfs2_assert_withdraw(sdp, atomic_read(&sdp->sd_log_blks_free) <= in gfs2_log_release()
474 sdp->sd_jdesc->jd_blocks); in gfs2_log_release()
475 if (atomic_read(&sdp->sd_log_blks_needed)) in gfs2_log_release()
476 wake_up(&sdp->sd_log_waitq); in gfs2_log_release()
488 static bool __gfs2_log_try_reserve(struct gfs2_sbd *sdp, unsigned int blks, in __gfs2_log_try_reserve() argument
494 free_blocks = atomic_read(&sdp->sd_log_blks_free); in __gfs2_log_try_reserve()
496 if (atomic_try_cmpxchg(&sdp->sd_log_blks_free, &free_blocks, in __gfs2_log_try_reserve()
498 trace_gfs2_log_blocks(sdp, -blks); in __gfs2_log_try_reserve()
524 static void __gfs2_log_reserve(struct gfs2_sbd *sdp, unsigned int blks, in __gfs2_log_reserve() argument
530 atomic_add(blks, &sdp->sd_log_blks_needed); in __gfs2_log_reserve()
532 if (current != sdp->sd_logd_process) in __gfs2_log_reserve()
533 wake_up(&sdp->sd_logd_waitq); in __gfs2_log_reserve()
534 io_wait_event(sdp->sd_log_waitq, in __gfs2_log_reserve()
535 (free_blocks = atomic_read(&sdp->sd_log_blks_free), in __gfs2_log_reserve()
538 if (atomic_try_cmpxchg(&sdp->sd_log_blks_free, in __gfs2_log_reserve()
546 trace_gfs2_log_blocks(sdp, -blks); in __gfs2_log_reserve()
547 if (atomic_sub_return(blks, &sdp->sd_log_blks_needed)) in __gfs2_log_reserve()
548 wake_up(&sdp->sd_log_waitq); in __gfs2_log_reserve()
561 bool gfs2_log_try_reserve(struct gfs2_sbd *sdp, struct gfs2_trans *tr, in gfs2_log_try_reserve() argument
569 if (revokes && !__gfs2_log_try_reserve_revokes(sdp, revokes)) { in gfs2_log_try_reserve()
570 revoke_blks = DIV_ROUND_UP(revokes, sdp->sd_inptrs); in gfs2_log_try_reserve()
571 *extra_revokes = revoke_blks * sdp->sd_inptrs - revokes; in gfs2_log_try_reserve()
576 if (__gfs2_log_try_reserve(sdp, blks, GFS2_LOG_FLUSH_MIN_BLOCKS)) in gfs2_log_try_reserve()
579 gfs2_log_release_revokes(sdp, revokes); in gfs2_log_try_reserve()
592 void gfs2_log_reserve(struct gfs2_sbd *sdp, struct gfs2_trans *tr, in gfs2_log_reserve() argument
601 revoke_blks = DIV_ROUND_UP(revokes, sdp->sd_inptrs); in gfs2_log_reserve()
602 *extra_revokes = revoke_blks * sdp->sd_inptrs - revokes; in gfs2_log_reserve()
605 __gfs2_log_reserve(sdp, blks, GFS2_LOG_FLUSH_MIN_BLOCKS); in gfs2_log_reserve()
620 static inline unsigned int log_distance(struct gfs2_sbd *sdp, unsigned int newer, in log_distance() argument
627 dist += sdp->sd_jdesc->jd_blocks; in log_distance()
653 static unsigned int calc_reserved(struct gfs2_sbd *sdp) in calc_reserved() argument
657 struct gfs2_trans *tr = sdp->sd_log_tr; in calc_reserved()
661 reserved += blocks + DIV_ROUND_UP(blocks, buf_limit(sdp)); in calc_reserved()
663 reserved += blocks + DIV_ROUND_UP(blocks, databuf_limit(sdp)); in calc_reserved()
668 static void log_pull_tail(struct gfs2_sbd *sdp) in log_pull_tail() argument
670 unsigned int new_tail = sdp->sd_log_flush_tail; in log_pull_tail()
673 if (new_tail == sdp->sd_log_tail) in log_pull_tail()
675 dist = log_distance(sdp, new_tail, sdp->sd_log_tail); in log_pull_tail()
676 ail2_empty(sdp, new_tail); in log_pull_tail()
677 gfs2_log_release(sdp, dist); in log_pull_tail()
678 sdp->sd_log_tail = new_tail; in log_pull_tail()
682 void log_flush_wait(struct gfs2_sbd *sdp) in log_flush_wait() argument
686 if (atomic_read(&sdp->sd_log_in_flight)) { in log_flush_wait()
688 prepare_to_wait(&sdp->sd_log_flush_wait, &wait, in log_flush_wait()
690 if (atomic_read(&sdp->sd_log_in_flight)) in log_flush_wait()
692 } while(atomic_read(&sdp->sd_log_in_flight)); in log_flush_wait()
693 finish_wait(&sdp->sd_log_flush_wait, &wait); in log_flush_wait()
717 static void gfs2_ordered_write(struct gfs2_sbd *sdp) in gfs2_ordered_write() argument
722 spin_lock(&sdp->sd_ordered_lock); in gfs2_ordered_write()
723 list_sort(NULL, &sdp->sd_log_ordered, &ip_cmp); in gfs2_ordered_write()
724 while (!list_empty(&sdp->sd_log_ordered)) { in gfs2_ordered_write()
725 ip = list_first_entry(&sdp->sd_log_ordered, struct gfs2_inode, i_ordered); in gfs2_ordered_write()
731 spin_unlock(&sdp->sd_ordered_lock); in gfs2_ordered_write()
733 spin_lock(&sdp->sd_ordered_lock); in gfs2_ordered_write()
735 list_splice(&written, &sdp->sd_log_ordered); in gfs2_ordered_write()
736 spin_unlock(&sdp->sd_ordered_lock); in gfs2_ordered_write()
739 static void gfs2_ordered_wait(struct gfs2_sbd *sdp) in gfs2_ordered_wait() argument
743 spin_lock(&sdp->sd_ordered_lock); in gfs2_ordered_wait()
744 while (!list_empty(&sdp->sd_log_ordered)) { in gfs2_ordered_wait()
745 ip = list_first_entry(&sdp->sd_log_ordered, struct gfs2_inode, i_ordered); in gfs2_ordered_wait()
749 spin_unlock(&sdp->sd_ordered_lock); in gfs2_ordered_wait()
751 spin_lock(&sdp->sd_ordered_lock); in gfs2_ordered_wait()
753 spin_unlock(&sdp->sd_ordered_lock); in gfs2_ordered_wait()
758 struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode); in gfs2_ordered_del_inode() local
760 spin_lock(&sdp->sd_ordered_lock); in gfs2_ordered_del_inode()
762 spin_unlock(&sdp->sd_ordered_lock); in gfs2_ordered_del_inode()
765 void gfs2_add_revoke(struct gfs2_sbd *sdp, struct gfs2_bufdata *bd) in gfs2_add_revoke() argument
770 sdp->sd_log_num_revoke++; in gfs2_add_revoke()
778 list_add(&bd->bd_list, &sdp->sd_log_revokes); in gfs2_add_revoke()
801 void gfs2_flush_revokes(struct gfs2_sbd *sdp) in gfs2_flush_revokes() argument
804 unsigned int max_revokes = atomic_read(&sdp->sd_log_revokes_available); in gfs2_flush_revokes()
806 gfs2_log_lock(sdp); in gfs2_flush_revokes()
807 gfs2_ail1_empty(sdp, max_revokes); in gfs2_flush_revokes()
808 gfs2_log_unlock(sdp); in gfs2_flush_revokes()
824 void gfs2_write_log_header(struct gfs2_sbd *sdp, struct gfs2_jdesc *jd, in gfs2_write_log_header() argument
831 struct gfs2_statfs_change_host *l_sc = &sdp->sd_statfs_local; in gfs2_write_log_header()
833 struct super_block *sb = sdp->sd_vfs; in gfs2_write_log_header()
836 if (gfs2_withdrawn(sdp)) in gfs2_write_log_header()
847 lh->lh_header.mh_jid = cpu_to_be32(sdp->sd_jdesc->jd_jid); in gfs2_write_log_header()
866 if (gfs2_assert_withdraw(sdp, ret == 0)) in gfs2_write_log_header()
877 cpu_to_be64(GFS2_I(sdp->sd_sc_inode)->i_no_addr); in gfs2_write_log_header()
879 cpu_to_be64(GFS2_I(sdp->sd_qc_inode)->i_no_addr); in gfs2_write_log_header()
881 spin_lock(&sdp->sd_statfs_spin); in gfs2_write_log_header()
885 spin_unlock(&sdp->sd_statfs_spin); in gfs2_write_log_header()
894 gfs2_log_write(sdp, jd, page, sb->s_blocksize, 0, dblock); in gfs2_write_log_header()
906 static void log_write_header(struct gfs2_sbd *sdp, u32 flags) in log_write_header() argument
909 enum gfs2_freeze_state state = atomic_read(&sdp->sd_freeze_state); in log_write_header()
911 gfs2_assert_withdraw(sdp, (state != SFS_FROZEN)); in log_write_header()
913 if (test_bit(SDF_NOBARRIERS, &sdp->sd_flags)) { in log_write_header()
914 gfs2_ordered_wait(sdp); in log_write_header()
915 log_flush_wait(sdp); in log_write_header()
918 sdp->sd_log_idle = (sdp->sd_log_flush_tail == sdp->sd_log_flush_head); in log_write_header()
919 gfs2_write_log_header(sdp, sdp->sd_jdesc, sdp->sd_log_sequence++, in log_write_header()
920 sdp->sd_log_flush_tail, sdp->sd_log_flush_head, in log_write_header()
922 gfs2_log_incr_head(sdp); in log_write_header()
923 log_flush_wait(sdp); in log_write_header()
924 log_pull_tail(sdp); in log_write_header()
925 gfs2_log_update_head(sdp); in log_write_header()
932 void gfs2_ail_drain(struct gfs2_sbd *sdp) in gfs2_ail_drain() argument
936 spin_lock(&sdp->sd_ail_lock); in gfs2_ail_drain()
944 while (!list_empty(&sdp->sd_ail1_list)) { in gfs2_ail_drain()
945 tr = list_first_entry(&sdp->sd_ail1_list, struct gfs2_trans, in gfs2_ail_drain()
947 gfs2_ail_empty_tr(sdp, tr, &tr->tr_ail1_list); in gfs2_ail_drain()
948 gfs2_ail_empty_tr(sdp, tr, &tr->tr_ail2_list); in gfs2_ail_drain()
950 gfs2_trans_free(sdp, tr); in gfs2_ail_drain()
952 while (!list_empty(&sdp->sd_ail2_list)) { in gfs2_ail_drain()
953 tr = list_first_entry(&sdp->sd_ail2_list, struct gfs2_trans, in gfs2_ail_drain()
955 gfs2_ail_empty_tr(sdp, tr, &tr->tr_ail2_list); in gfs2_ail_drain()
957 gfs2_trans_free(sdp, tr); in gfs2_ail_drain()
959 gfs2_drain_revokes(sdp); in gfs2_ail_drain()
960 spin_unlock(&sdp->sd_ail_lock); in gfs2_ail_drain()
967 static void empty_ail1_list(struct gfs2_sbd *sdp) in empty_ail1_list() argument
973 fs_err(sdp, "Error: In %s for 10 minutes! t=%d\n", in empty_ail1_list()
975 dump_ail_list(sdp); in empty_ail1_list()
978 gfs2_ail1_start(sdp); in empty_ail1_list()
979 gfs2_ail1_wait(sdp); in empty_ail1_list()
980 if (gfs2_ail1_empty(sdp, 0)) in empty_ail1_list()
1027 void gfs2_log_flush(struct gfs2_sbd *sdp, struct gfs2_glock *gl, u32 flags) in gfs2_log_flush() argument
1031 enum gfs2_freeze_state state = atomic_read(&sdp->sd_freeze_state); in gfs2_log_flush()
1035 down_write(&sdp->sd_log_flush_lock); in gfs2_log_flush()
1036 trace_gfs2_log_flush(sdp, 1, flags); in gfs2_log_flush()
1043 if (gfs2_withdrawn(sdp) || !test_bit(SDF_JOURNAL_LIVE, &sdp->sd_flags)) in gfs2_log_flush()
1050 first_log_head = sdp->sd_log_head; in gfs2_log_flush()
1051 sdp->sd_log_flush_head = first_log_head; in gfs2_log_flush()
1053 tr = sdp->sd_log_tr; in gfs2_log_flush()
1054 if (tr || sdp->sd_log_num_revoke) { in gfs2_log_flush()
1056 gfs2_log_release(sdp, reserved_blocks); in gfs2_log_flush()
1057 reserved_blocks = sdp->sd_log_blks_reserved; in gfs2_log_flush()
1058 reserved_revokes = sdp->sd_log_num_revoke; in gfs2_log_flush()
1060 sdp->sd_log_tr = NULL; in gfs2_log_flush()
1063 if (gfs2_assert_withdraw_delayed(sdp, in gfs2_log_flush()
1072 if (current == sdp->sd_logd_process) in gfs2_log_flush()
1075 if (!__gfs2_log_try_reserve(sdp, reserved_blocks, taboo_blocks)) { in gfs2_log_flush()
1076 up_write(&sdp->sd_log_flush_lock); in gfs2_log_flush()
1077 __gfs2_log_reserve(sdp, reserved_blocks, taboo_blocks); in gfs2_log_flush()
1078 down_write(&sdp->sd_log_flush_lock); in gfs2_log_flush()
1081 BUG_ON(sdp->sd_log_num_revoke); in gfs2_log_flush()
1085 clear_bit(SDF_JOURNAL_LIVE, &sdp->sd_flags); in gfs2_log_flush()
1088 if (gfs2_assert_withdraw_delayed(sdp, !reserved_revokes)) in gfs2_log_flush()
1091 gfs2_ordered_write(sdp); in gfs2_log_flush()
1092 if (gfs2_withdrawn(sdp)) in gfs2_log_flush()
1094 lops_before_commit(sdp, tr); in gfs2_log_flush()
1095 if (gfs2_withdrawn(sdp)) in gfs2_log_flush()
1097 gfs2_log_submit_bio(&sdp->sd_jdesc->jd_log_bio, REQ_OP_WRITE); in gfs2_log_flush()
1098 if (gfs2_withdrawn(sdp)) in gfs2_log_flush()
1101 if (sdp->sd_log_head != sdp->sd_log_flush_head) { in gfs2_log_flush()
1102 log_write_header(sdp, flags); in gfs2_log_flush()
1103 } else if (sdp->sd_log_tail != sdp->sd_log_flush_tail && !sdp->sd_log_idle) { in gfs2_log_flush()
1104 log_write_header(sdp, flags); in gfs2_log_flush()
1106 if (gfs2_withdrawn(sdp)) in gfs2_log_flush()
1108 lops_after_commit(sdp, tr); in gfs2_log_flush()
1110 gfs2_log_lock(sdp); in gfs2_log_flush()
1111 sdp->sd_log_blks_reserved = 0; in gfs2_log_flush()
1113 spin_lock(&sdp->sd_ail_lock); in gfs2_log_flush()
1115 list_add(&tr->tr_list, &sdp->sd_ail1_list); in gfs2_log_flush()
1118 spin_unlock(&sdp->sd_ail_lock); in gfs2_log_flush()
1119 gfs2_log_unlock(sdp); in gfs2_log_flush()
1122 if (!sdp->sd_log_idle) { in gfs2_log_flush()
1123 empty_ail1_list(sdp); in gfs2_log_flush()
1124 if (gfs2_withdrawn(sdp)) in gfs2_log_flush()
1126 log_write_header(sdp, flags); in gfs2_log_flush()
1130 gfs2_log_shutdown(sdp); in gfs2_log_flush()
1132 atomic_set(&sdp->sd_freeze_state, SFS_FROZEN); in gfs2_log_flush()
1136 used_blocks = log_distance(sdp, sdp->sd_log_flush_head, first_log_head); in gfs2_log_flush()
1137 reserved_revokes += atomic_read(&sdp->sd_log_revokes_available); in gfs2_log_flush()
1138 atomic_set(&sdp->sd_log_revokes_available, sdp->sd_ldptrs); in gfs2_log_flush()
1139 gfs2_assert_withdraw(sdp, reserved_revokes % sdp->sd_inptrs == sdp->sd_ldptrs); in gfs2_log_flush()
1140 if (reserved_revokes > sdp->sd_ldptrs) in gfs2_log_flush()
1141 reserved_blocks += (reserved_revokes - sdp->sd_ldptrs) / sdp->sd_inptrs; in gfs2_log_flush()
1144 gfs2_assert_withdraw_delayed(sdp, used_blocks < reserved_blocks); in gfs2_log_flush()
1145 gfs2_log_release(sdp, reserved_blocks - used_blocks); in gfs2_log_flush()
1147 up_write(&sdp->sd_log_flush_lock); in gfs2_log_flush()
1148 gfs2_trans_free(sdp, tr); in gfs2_log_flush()
1149 if (gfs2_withdrawing(sdp)) in gfs2_log_flush()
1150 gfs2_withdraw(sdp); in gfs2_log_flush()
1151 trace_gfs2_log_flush(sdp, 0, flags); in gfs2_log_flush()
1162 spin_lock(&sdp->sd_ail_lock); in gfs2_log_flush()
1164 list_add(&tr->tr_list, &sdp->sd_ail1_list); in gfs2_log_flush()
1165 spin_unlock(&sdp->sd_ail_lock); in gfs2_log_flush()
1176 static void gfs2_merge_trans(struct gfs2_sbd *sdp, struct gfs2_trans *new) in gfs2_merge_trans() argument
1178 struct gfs2_trans *old = sdp->sd_log_tr; in gfs2_merge_trans()
1192 spin_lock(&sdp->sd_ail_lock); in gfs2_merge_trans()
1195 spin_unlock(&sdp->sd_ail_lock); in gfs2_merge_trans()
1198 static void log_refund(struct gfs2_sbd *sdp, struct gfs2_trans *tr) in log_refund() argument
1204 gfs2_log_lock(sdp); in log_refund()
1206 if (sdp->sd_log_tr) { in log_refund()
1207 gfs2_merge_trans(sdp, tr); in log_refund()
1209 gfs2_assert_withdraw(sdp, !test_bit(TR_ONSTACK, &tr->tr_flags)); in log_refund()
1210 sdp->sd_log_tr = tr; in log_refund()
1214 reserved = calc_reserved(sdp); in log_refund()
1215 maxres = sdp->sd_log_blks_reserved + tr->tr_reserved; in log_refund()
1216 gfs2_assert_withdraw(sdp, maxres >= reserved); in log_refund()
1219 gfs2_log_release(sdp, unused); in log_refund()
1220 sdp->sd_log_blks_reserved = reserved; in log_refund()
1222 gfs2_log_unlock(sdp); in log_refund()
1240 void gfs2_log_commit(struct gfs2_sbd *sdp, struct gfs2_trans *tr) in gfs2_log_commit() argument
1242 log_refund(sdp, tr); in gfs2_log_commit()
1244 if (atomic_read(&sdp->sd_log_pinned) > atomic_read(&sdp->sd_log_thresh1) || in gfs2_log_commit()
1245 ((sdp->sd_jdesc->jd_blocks - atomic_read(&sdp->sd_log_blks_free)) > in gfs2_log_commit()
1246 atomic_read(&sdp->sd_log_thresh2))) in gfs2_log_commit()
1247 wake_up(&sdp->sd_logd_waitq); in gfs2_log_commit()
1256 static void gfs2_log_shutdown(struct gfs2_sbd *sdp) in gfs2_log_shutdown() argument
1258 gfs2_assert_withdraw(sdp, !sdp->sd_log_blks_reserved); in gfs2_log_shutdown()
1259 gfs2_assert_withdraw(sdp, !sdp->sd_log_num_revoke); in gfs2_log_shutdown()
1260 gfs2_assert_withdraw(sdp, list_empty(&sdp->sd_ail1_list)); in gfs2_log_shutdown()
1262 log_write_header(sdp, GFS2_LOG_HEAD_UNMOUNT | GFS2_LFC_SHUTDOWN); in gfs2_log_shutdown()
1263 log_pull_tail(sdp); in gfs2_log_shutdown()
1265 gfs2_assert_warn(sdp, sdp->sd_log_head == sdp->sd_log_tail); in gfs2_log_shutdown()
1266 gfs2_assert_warn(sdp, list_empty(&sdp->sd_ail2_list)); in gfs2_log_shutdown()
1269 static inline int gfs2_jrnl_flush_reqd(struct gfs2_sbd *sdp) in gfs2_jrnl_flush_reqd() argument
1271 return (atomic_read(&sdp->sd_log_pinned) + in gfs2_jrnl_flush_reqd()
1272 atomic_read(&sdp->sd_log_blks_needed) >= in gfs2_jrnl_flush_reqd()
1273 atomic_read(&sdp->sd_log_thresh1)); in gfs2_jrnl_flush_reqd()
1276 static inline int gfs2_ail_flush_reqd(struct gfs2_sbd *sdp) in gfs2_ail_flush_reqd() argument
1278 unsigned int used_blocks = sdp->sd_jdesc->jd_blocks - atomic_read(&sdp->sd_log_blks_free); in gfs2_ail_flush_reqd()
1280 if (test_and_clear_bit(SDF_FORCE_AIL_FLUSH, &sdp->sd_flags)) in gfs2_ail_flush_reqd()
1283 return used_blocks + atomic_read(&sdp->sd_log_blks_needed) >= in gfs2_ail_flush_reqd()
1284 atomic_read(&sdp->sd_log_thresh2); in gfs2_ail_flush_reqd()
1297 struct gfs2_sbd *sdp = data; in gfs2_logd() local
1303 if (gfs2_withdrawn(sdp)) { in gfs2_logd()
1308 if (sdp->sd_log_error) { in gfs2_logd()
1309 gfs2_lm(sdp, in gfs2_logd()
1313 sdp->sd_fsname, sdp->sd_log_error); in gfs2_logd()
1314 gfs2_withdraw(sdp); in gfs2_logd()
1318 if (gfs2_jrnl_flush_reqd(sdp) || t == 0) { in gfs2_logd()
1319 gfs2_ail1_empty(sdp, 0); in gfs2_logd()
1320 gfs2_log_flush(sdp, NULL, GFS2_LOG_HEAD_FLUSH_NORMAL | in gfs2_logd()
1324 if (gfs2_ail_flush_reqd(sdp)) { in gfs2_logd()
1325 gfs2_ail1_start(sdp); in gfs2_logd()
1326 gfs2_ail1_wait(sdp); in gfs2_logd()
1327 gfs2_ail1_empty(sdp, 0); in gfs2_logd()
1328 gfs2_log_flush(sdp, NULL, GFS2_LOG_HEAD_FLUSH_NORMAL | in gfs2_logd()
1332 t = gfs2_tune_get(sdp, gt_logd_secs) * HZ; in gfs2_logd()
1337 prepare_to_wait(&sdp->sd_logd_waitq, &wait, in gfs2_logd()
1339 if (!gfs2_ail_flush_reqd(sdp) && in gfs2_logd()
1340 !gfs2_jrnl_flush_reqd(sdp) && in gfs2_logd()
1343 } while(t && !gfs2_ail_flush_reqd(sdp) && in gfs2_logd()
1344 !gfs2_jrnl_flush_reqd(sdp) && in gfs2_logd()
1346 finish_wait(&sdp->sd_logd_waitq, &wait); in gfs2_logd()