[PATCH] reiserfs: reiserfs hang and performance fix for data=journal mode

In data=journal mode, reiserfs writepage needs to make sure not to trigger
transactions while being run under PF_MEMALLOC.  This patch makes sure to
redirty the page instead of forcing a transaction start in this case.

Also, calling filemap_fdata* in order to trigger io on the block device can
cause lock inversions on the page lock.  Instead, do simple batching from
flush_commit_list.

Signed-off-by: Chris Mason <mason@suse.com>
Signed-off-by: Andrew Morton <akpm@osdl.org>
Signed-off-by: Linus Torvalds <torvalds@osdl.org>
This commit is contained in:
Chris Mason 2006-02-01 03:06:49 -08:00 committed by Linus Torvalds
parent fc5cd582e9
commit e0e851cf30
2 changed files with 21 additions and 5 deletions

View File

@ -2363,6 +2363,13 @@ static int reiserfs_write_full_page(struct page *page,
int bh_per_page = PAGE_CACHE_SIZE / s->s_blocksize; int bh_per_page = PAGE_CACHE_SIZE / s->s_blocksize;
th.t_trans_id = 0; th.t_trans_id = 0;
/* no logging allowed when nonblocking or from PF_MEMALLOC */
if (checked && (current->flags & PF_MEMALLOC)) {
redirty_page_for_writepage(wbc, page);
unlock_page(page);
return 0;
}
/* The page dirty bit is cleared before writepage is called, which /* The page dirty bit is cleared before writepage is called, which
* means we have to tell create_empty_buffers to make dirty buffers * means we have to tell create_empty_buffers to make dirty buffers
* The page really should be up to date at this point, so tossing * The page really should be up to date at this point, so tossing

View File

@ -988,6 +988,7 @@ static int flush_commit_list(struct super_block *s,
struct reiserfs_journal *journal = SB_JOURNAL(s); struct reiserfs_journal *journal = SB_JOURNAL(s);
int barrier = 0; int barrier = 0;
int retval = 0; int retval = 0;
int write_len;
reiserfs_check_lock_depth(s, "flush_commit_list"); reiserfs_check_lock_depth(s, "flush_commit_list");
@ -1037,16 +1038,24 @@ static int flush_commit_list(struct super_block *s,
BUG_ON(!list_empty(&jl->j_bh_list)); BUG_ON(!list_empty(&jl->j_bh_list));
/* /*
* for the description block and all the log blocks, submit any buffers * for the description block and all the log blocks, submit any buffers
* that haven't already reached the disk * that haven't already reached the disk. Try to write at least 256
* log blocks. later on, we will only wait on blocks that correspond
* to this transaction, but while we're unplugging we might as well
* get a chunk of data on there.
*/ */
atomic_inc(&journal->j_async_throttle); atomic_inc(&journal->j_async_throttle);
for (i = 0; i < (jl->j_len + 1); i++) { write_len = jl->j_len + 1;
if (write_len < 256)
write_len = 256;
for (i = 0 ; i < write_len ; i++) {
bn = SB_ONDISK_JOURNAL_1st_BLOCK(s) + (jl->j_start + i) % bn = SB_ONDISK_JOURNAL_1st_BLOCK(s) + (jl->j_start + i) %
SB_ONDISK_JOURNAL_SIZE(s); SB_ONDISK_JOURNAL_SIZE(s);
tbh = journal_find_get_block(s, bn); tbh = journal_find_get_block(s, bn);
if (buffer_dirty(tbh)) /* redundant, ll_rw_block() checks */ if (tbh) {
ll_rw_block(SWRITE, 1, &tbh); if (buffer_dirty(tbh))
put_bh(tbh); ll_rw_block(WRITE, 1, &tbh) ;
put_bh(tbh) ;
}
} }
atomic_dec(&journal->j_async_throttle); atomic_dec(&journal->j_async_throttle);