[PATCH] bounce:fix bug, avoid to flush dcache on slab page from jbd2.
Jan Kara
jack at suse.cz
Mon Mar 18 13:32:42 EDT 2013
On Fri 15-03-13 10:54:41, Darrick J. Wong wrote:
> On Fri, Mar 15, 2013 at 11:01:05AM +0100, Jan Kara wrote:
> > On Thu 14-03-13 15:42:43, Darrick J. Wong wrote:
> > > On Wed, Mar 13, 2013 at 10:02:16PM +0100, Jan Kara wrote:
> > > > On Wed 13-03-13 12:44:29, Darrick J. Wong wrote:
> > > > > On Wed, Mar 13, 2013 at 09:50:21AM +0100, Jan Kara wrote:
> > > > > > On Tue 12-03-13 18:10:20, Darrick J. Wong wrote:
> > > > > > > On Tue, Mar 12, 2013 at 03:32:21PM -0700, Andrew Morton wrote:
> > > > > > > > On Fri, 08 Mar 2013 20:37:36 +0800 Shuge <shugelinux at gmail.com> wrote:
> > > > > > > >
> > > > > > > > > The bounce accept slab pages from jbd2, and flush dcache on them.
> > > > > > > > > When enabling VM_DEBUG, it will tigger VM_BUG_ON in page_mapping().
> > > > > > > > > So, check PageSlab to avoid it in __blk_queue_bounce().
> > > > > > > > >
> > > > > > > > > Bug URL: http://lkml.org/lkml/2013/3/7/56
> > > > > > > > >
> > > > > > > > > ...
> > > > > > > > >
> > > > > > > > > --- a/mm/bounce.c
> > > > > > > > > +++ b/mm/bounce.c
> > > > > > > > > @@ -214,7 +214,8 @@ static void __blk_queue_bounce(struct request_queue
> > > > > > > > > *q, struct bio **bio_orig,
> > > > > > > > > if (rw == WRITE) {
> > > > > > > > > char *vto, *vfrom;
> > > > > > > > > - flush_dcache_page(from->bv_page);
> > > > > > > > > + if (unlikely(!PageSlab(from->bv_page)))
> > > > > > > > > + flush_dcache_page(from->bv_page);
> > > > > > > > > vto = page_address(to->bv_page) + to->bv_offset;
> > > > > > > > > vfrom = kmap(from->bv_page) + from->bv_offset;
> > > > > > > > > memcpy(vto, vfrom, to->bv_len);
> > > > > > > >
> > > > > > > > I guess this is triggered by Catalin's f1a0c4aa0937975b ("arm64: Cache
> > > > > > > > maintenance routines"), which added a page_mapping() call to arm64's
> > > > > > > > arch/arm64/mm/flush.c:flush_dcache_page().
> > > > > > > >
> > > > > > > > What's happening is that jbd2 is using kmalloc() to allocate buffer_head
> > > > > > > > data. That gets submitted down the BIO layer and __blk_queue_bounce()
> > > > > > > > calls flush_dcache_page() which in the arm64 case calls page_mapping()
> > > > > > > > and page_mapping() does VM_BUG_ON(PageSlab) and splat.
> > > > > > > >
> > > > > > > > The unusual thing about all of this is that the payload for some disk
> > > > > > > > IO is coming from kmalloc, rather than being a user page. It's oddball
> > > > > > > > but we've done this for ages and should continue to support it.
> > > > > > > >
> > > > > > > >
> > > > > > > > Now, the page from kmalloc() cannot be in highmem, so why did the
> > > > > > > > bounce code decide to bounce it?
> > > > > > > >
> > > > > > > > __blk_queue_bounce() does
> > > > > > > >
> > > > > > > > /*
> > > > > > > > * is destination page below bounce pfn?
> > > > > > > > */
> > > > > > > > if (page_to_pfn(page) <= queue_bounce_pfn(q) && !force)
> > > > > > > > continue;
> > > > > > > >
> > > > > > > > and `force' comes from must_snapshot_stable_pages(). But
> > > > > > > > must_snapshot_stable_pages() must have returned false, because if it
> > > > > > > > had returned true then it would have been must_snapshot_stable_pages()
> > > > > > > > which went BUG, because must_snapshot_stable_pages() calls page_mapping().
> > > > > > > >
> > > > > > > > So my tentative diagosis is that arm64 is fishy. A page which was
> > > > > > > > allocated via jbd2_alloc(GFP_NOFS)->kmem_cache_alloc() ended up being
> > > > > > > > above arm64's queue_bounce_pfn(). Can you please do a bit of
> > > > > > > > investigation to work out if this is what is happening? Find out why
> > > > > > > > __blk_queue_bounce() decided to bounce a page which shouldn't have been
> > > > > > > > bounced?
> > > > > > >
> > > > > > > That sure is strange. I didn't see any obvious reasons why we'd end up with a
> > > > > > > kmalloc above queue_bounce_pfn(). But then I don't have any arm64s either.
> > > > > > >
> > > > > > > > This is all terribly fragile :( afaict if someone sets
> > > > > > > > bdi_cap_stable_pages_required() against that jbd2 queue, we're going to
> > > > > > > > hit that BUG_ON() again, via must_snapshot_stable_pages()'s
> > > > > > > > page_mapping() call. (Darrick, this means you ;))
> > > > > > >
> > > > > > > Wheeee. You're right, we shouldn't be calling page_mapping on slab pages.
> > > > > > > We can keep walking the bio segments to find a non-slab page that can tell us
> > > > > > > MS_SNAP_STABLE is set, since we probably won't need the bounce buffer anyway.
> > > > > > >
> > > > > > > How does something like this look? (+ the patch above)
> > > > > > Umm, this won't quite work. We can have a bio which has just PageSlab
> > > > > > page attached and so you won't be able to get to the superblock. Heh, isn't
> > > > > > the whole page_mapping() thing in must_snapshot_stable_pages() wrong? When we
> > > > > > do direct IO, these pages come directly from userspace and hell knows where
> > > > > > they come from. Definitely their page_mapping() doesn't give us anything
> > > > > > useful... Sorry for not realizing this earlier when reviewing the patch.
> > > > > >
> > > > > > ... remembering why we need to get to sb and why ext3 needs this ... So
> > > > > > maybe a better solution would be to have a bio flag meaning that pages need
> > > > > > bouncing? And we would set it from filesystems that need it - in case of
> > > > > > ext3 only writeback of data from kjournald actually needs to bounce the
> > > > > > pages. Thoughts?
> > > > >
> > > > > What about dirty pages that don't result in journal transactions? I think
> > > > > ext3_sync_file() eventually calls ext3_ordered_writepage, which then calls
> > > > > __block_write_full_page, which in turn calls submit_bh().
> > > > So here we have two options:
> > > > Either we let ext3 wait the same way as other filesystems when stable pages
> > > > are required. Then only data IO from kjournald needs to be bounced (all
> > > > other IO is properly protected by PageWriteback bit).
> > > >
> > > > Or we won't let ext3 wait (as it is now), keep the superblock flag that fs
> > > > needs bouncing, and set the bio flag in __block_write_full_page() and
> > > > kjournald based on the sb flag.
> > > >
> > > > I think the first option is slightly better but I don't feel strongly
> > > > about that.
> > >
> > > I like that first option -- it contains the kludgery to jbd instead of
> > > spreading it around. Here's a patch that passes a quick smoke test on ext[34],
> > > xfs, and vfat. What do you think of this one? Should I create a
> > > submit_snapshot_bh() instead of letting callers stuff in arbitrary dangerous
> > > BH_ flags?
> > Thanks for writing the patch. I think _submit_bh() is OK as you did it. I
> > have just two comments below.
> >
> > > ---
> > > From: Darrick J. Wong <darrick.wong at oracle.com>
> > > Subject: [PATCH] mm: Make snapshotting pages for stable writes a per-bio operation
> > >
> > > Walking a bio's page mappings has proved problematic, so create a new bio flag
> > > to indicate that a bio's data needs to be snapshotted in order to guarantee
> > > stable pages during writeback. Next, for the one user (ext3/jbd) of
> > > snapshotting, hook all the places where writes can be initiated without
> > > PG_writeback set, and set BIO_SNAP_STABLE there. Finally, the MS_SNAP_STABLE
> > > mount flag (only used by ext3) is now superfluous, so get rid of it.
> > >
> > > Signed-off-by: Darrick J. Wong <darrick.wong at oracle.com>
> > > ---
> > > fs/buffer.c | 9 ++++++++-
> > > fs/ext3/super.c | 1 -
> > > fs/jbd/commit.c | 4 ++--
> > > include/linux/blk_types.h | 3 ++-
> > > include/linux/buffer_head.h | 1 +
> > > include/uapi/linux/fs.h | 1 -
> > > mm/bounce.c | 21 +--------------------
> > > mm/page-writeback.c | 4 ----
> > > 8 files changed, 14 insertions(+), 30 deletions(-)
> > >
> > ...
> > > diff --git a/fs/jbd/commit.c b/fs/jbd/commit.c
> > > index 86b39b1..b91b688 100644
> > > --- a/fs/jbd/commit.c
> > > +++ b/fs/jbd/commit.c
> > > @@ -163,7 +163,7 @@ static void journal_do_submit_data(struct buffer_head **wbuf, int bufs,
> > > for (i = 0; i < bufs; i++) {
> > > wbuf[i]->b_end_io = end_buffer_write_sync;
> > > /* We use-up our safety reference in submit_bh() */
> > > - submit_bh(write_op, wbuf[i]);
> > > + _submit_bh(write_op, wbuf[i], 1 << BIO_SNAP_STABLE);
> > Please add a comment here why we need BIO_SNAP_STABLE. Something like:
> > /*
> > * Here we write back pagecache data that may be mmaped. Since we cannot
> > * afford to clean the page and set PageWriteback here due to lock ordering
> > * (page lock ranks above transaction start), the data can change while IO is
> > * in flight. Tell the block layer it should bounce the bio pages if stable
> > * data during write is required.
> > */
> >
> > > }
> > > }
> > >
> > > @@ -667,7 +667,7 @@ start_journal_io:
> > > clear_buffer_dirty(bh);
> > > set_buffer_uptodate(bh);
> > > bh->b_end_io = journal_end_buffer_io_sync;
> > > - submit_bh(write_op, bh);
> > > + _submit_bh(write_op, bh, 1 << BIO_SNAP_STABLE);
> > And this isn't needed. Here we write out only metadata and JBD already
> > handles copying those / waiting for IO in flight for metadata.
>
> I think it only copies the page if either the buffer is also a part of the
> current transaction (or someone called do_get_undo_access()). Unfortunately,
> if we're in data=journal mode, dirty data pages get pushed through jbd as if
> they were fs metadata, but in the meantime other processes can still write to
> those pages. So I guess we need the journal to freeze those pages as soon as
> they come in.
So you miss the part that do_get_write_access() actually waits for buffer
if it is undergoing commit (kjournald is writing it). But you are right
that in data=journal mode if the page is mmaped, user can change it while
kjournald is doing write out. Actually the same problem is with ext4 and
data=journal mode. So for now I'd stay with your simple patch and later we
can optimize it so that we don't have to pay the penalty when the buffer is
not journalled data.
Honza
---
Jan Kara <jack at suse.cz>
SUSE Labs, CR
More information about the linux-arm-kernel
mailing list