mirror of
				https://github.com/torvalds/linux.git
				synced 2025-11-04 10:40:15 +02:00 
			
		
		
		
	mm: make snapshotting pages for stable writes a per-bio operation
Walking a bio's page mappings has proved problematic, so create a new bio flag to indicate that a bio's data needs to be snapshotted in order to guarantee stable pages during writeback. Next, for the one user (ext3/jbd) of snapshotting, hook all the places where writes can be initiated without PG_writeback set, and set BIO_SNAP_STABLE there. We must also flag journal "metadata" bios for stable writeout, since file data can be written through the journal. Finally, the MS_SNAP_STABLE mount flag (only used by ext3) is now superfluous, so get rid of it. [akpm@linux-foundation.org: rename _submit_bh()'s `flags' to `bio_flags', delobotomize the _submit_bh declaration] [akpm@linux-foundation.org: teeny cleanup] Signed-off-by: Darrick J. Wong <darrick.wong@oracle.com> Cc: Andy Lutomirski <luto@amacapital.net> Cc: Adrian Hunter <adrian.hunter@intel.com> Cc: Artem Bityutskiy <dedekind1@gmail.com> Reviewed-by: Jan Kara <jack@suse.cz> Cc: Jens Axboe <axboe@kernel.dk> Signed-off-by: Andrew Morton <akpm@linux-foundation.org> Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
This commit is contained in:
		
							parent
							
								
									106c992a5e
								
							
						
					
					
						commit
						7136851117
					
				
					 8 changed files with 34 additions and 31 deletions
				
			
		| 
						 | 
					@ -2949,7 +2949,7 @@ static void guard_bh_eod(int rw, struct bio *bio, struct buffer_head *bh)
 | 
				
			||||||
	}
 | 
						}
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
int submit_bh(int rw, struct buffer_head * bh)
 | 
					int _submit_bh(int rw, struct buffer_head *bh, unsigned long bio_flags)
 | 
				
			||||||
{
 | 
					{
 | 
				
			||||||
	struct bio *bio;
 | 
						struct bio *bio;
 | 
				
			||||||
	int ret = 0;
 | 
						int ret = 0;
 | 
				
			||||||
| 
						 | 
					@ -2984,6 +2984,7 @@ int submit_bh(int rw, struct buffer_head * bh)
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	bio->bi_end_io = end_bio_bh_io_sync;
 | 
						bio->bi_end_io = end_bio_bh_io_sync;
 | 
				
			||||||
	bio->bi_private = bh;
 | 
						bio->bi_private = bh;
 | 
				
			||||||
 | 
						bio->bi_flags |= bio_flags;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	/* Take care of bh's that straddle the end of the device */
 | 
						/* Take care of bh's that straddle the end of the device */
 | 
				
			||||||
	guard_bh_eod(rw, bio, bh);
 | 
						guard_bh_eod(rw, bio, bh);
 | 
				
			||||||
| 
						 | 
					@ -2997,6 +2998,12 @@ int submit_bh(int rw, struct buffer_head * bh)
 | 
				
			||||||
	bio_put(bio);
 | 
						bio_put(bio);
 | 
				
			||||||
	return ret;
 | 
						return ret;
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
 | 
					EXPORT_SYMBOL_GPL(_submit_bh);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					int submit_bh(int rw, struct buffer_head *bh)
 | 
				
			||||||
 | 
					{
 | 
				
			||||||
 | 
						return _submit_bh(rw, bh, 0);
 | 
				
			||||||
 | 
					}
 | 
				
			||||||
EXPORT_SYMBOL(submit_bh);
 | 
					EXPORT_SYMBOL(submit_bh);
 | 
				
			||||||
 | 
					
 | 
				
			||||||
/**
 | 
					/**
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
| 
						 | 
					@ -2067,7 +2067,6 @@ static int ext3_fill_super (struct super_block *sb, void *data, int silent)
 | 
				
			||||||
		test_opt(sb,DATA_FLAGS) == EXT3_MOUNT_JOURNAL_DATA ? "journal":
 | 
							test_opt(sb,DATA_FLAGS) == EXT3_MOUNT_JOURNAL_DATA ? "journal":
 | 
				
			||||||
		test_opt(sb,DATA_FLAGS) == EXT3_MOUNT_ORDERED_DATA ? "ordered":
 | 
							test_opt(sb,DATA_FLAGS) == EXT3_MOUNT_ORDERED_DATA ? "ordered":
 | 
				
			||||||
		"writeback");
 | 
							"writeback");
 | 
				
			||||||
	sb->s_flags |= MS_SNAP_STABLE;
 | 
					 | 
				
			||||||
 | 
					
 | 
				
			||||||
	return 0;
 | 
						return 0;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
| 
						 | 
					@ -162,8 +162,17 @@ static void journal_do_submit_data(struct buffer_head **wbuf, int bufs,
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	for (i = 0; i < bufs; i++) {
 | 
						for (i = 0; i < bufs; i++) {
 | 
				
			||||||
		wbuf[i]->b_end_io = end_buffer_write_sync;
 | 
							wbuf[i]->b_end_io = end_buffer_write_sync;
 | 
				
			||||||
		/* We use-up our safety reference in submit_bh() */
 | 
							/*
 | 
				
			||||||
		submit_bh(write_op, wbuf[i]);
 | 
							 * Here we write back pagecache data that may be mmaped. Since
 | 
				
			||||||
 | 
							 * we cannot afford to clean the page and set PageWriteback
 | 
				
			||||||
 | 
							 * here due to lock ordering (page lock ranks above transaction
 | 
				
			||||||
 | 
							 * start), the data can change while IO is in flight. Tell the
 | 
				
			||||||
 | 
							 * block layer it should bounce the bio pages if stable data
 | 
				
			||||||
 | 
							 * during write is required.
 | 
				
			||||||
 | 
							 *
 | 
				
			||||||
 | 
							 * We use up our safety reference in submit_bh().
 | 
				
			||||||
 | 
							 */
 | 
				
			||||||
 | 
							_submit_bh(write_op, wbuf[i], 1 << BIO_SNAP_STABLE);
 | 
				
			||||||
	}
 | 
						}
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
 | 
					
 | 
				
			||||||
| 
						 | 
					@ -667,7 +676,17 @@ void journal_commit_transaction(journal_t *journal)
 | 
				
			||||||
				clear_buffer_dirty(bh);
 | 
									clear_buffer_dirty(bh);
 | 
				
			||||||
				set_buffer_uptodate(bh);
 | 
									set_buffer_uptodate(bh);
 | 
				
			||||||
				bh->b_end_io = journal_end_buffer_io_sync;
 | 
									bh->b_end_io = journal_end_buffer_io_sync;
 | 
				
			||||||
				submit_bh(write_op, bh);
 | 
									/*
 | 
				
			||||||
 | 
									 * In data=journal mode, here we can end up
 | 
				
			||||||
 | 
									 * writing pagecache data that might be
 | 
				
			||||||
 | 
									 * mmapped. Since we can't afford to clean the
 | 
				
			||||||
 | 
									 * page and set PageWriteback (see the comment
 | 
				
			||||||
 | 
									 * near the other use of _submit_bh()), the
 | 
				
			||||||
 | 
									 * data can change while the write is in
 | 
				
			||||||
 | 
									 * flight.  Tell the block layer to bounce the
 | 
				
			||||||
 | 
									 * bio pages if stable pages are required.
 | 
				
			||||||
 | 
									 */
 | 
				
			||||||
 | 
									_submit_bh(write_op, bh, 1 << BIO_SNAP_STABLE);
 | 
				
			||||||
			}
 | 
								}
 | 
				
			||||||
			cond_resched();
 | 
								cond_resched();
 | 
				
			||||||
 | 
					
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
| 
						 | 
					@ -111,12 +111,13 @@ struct bio {
 | 
				
			||||||
#define BIO_FS_INTEGRITY 9	/* fs owns integrity data, not block layer */
 | 
					#define BIO_FS_INTEGRITY 9	/* fs owns integrity data, not block layer */
 | 
				
			||||||
#define BIO_QUIET	10	/* Make BIO Quiet */
 | 
					#define BIO_QUIET	10	/* Make BIO Quiet */
 | 
				
			||||||
#define BIO_MAPPED_INTEGRITY 11/* integrity metadata has been remapped */
 | 
					#define BIO_MAPPED_INTEGRITY 11/* integrity metadata has been remapped */
 | 
				
			||||||
 | 
					#define BIO_SNAP_STABLE	12	/* bio data must be snapshotted during write */
 | 
				
			||||||
 | 
					
 | 
				
			||||||
/*
 | 
					/*
 | 
				
			||||||
 * Flags starting here get preserved by bio_reset() - this includes
 | 
					 * Flags starting here get preserved by bio_reset() - this includes
 | 
				
			||||||
 * BIO_POOL_IDX()
 | 
					 * BIO_POOL_IDX()
 | 
				
			||||||
 */
 | 
					 */
 | 
				
			||||||
#define BIO_RESET_BITS	12
 | 
					#define BIO_RESET_BITS	13
 | 
				
			||||||
 | 
					
 | 
				
			||||||
#define bio_flagged(bio, flag)	((bio)->bi_flags & (1 << (flag)))
 | 
					#define bio_flagged(bio, flag)	((bio)->bi_flags & (1 << (flag)))
 | 
				
			||||||
 | 
					
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
| 
						 | 
					@ -181,6 +181,7 @@ void ll_rw_block(int, int, struct buffer_head * bh[]);
 | 
				
			||||||
int sync_dirty_buffer(struct buffer_head *bh);
 | 
					int sync_dirty_buffer(struct buffer_head *bh);
 | 
				
			||||||
int __sync_dirty_buffer(struct buffer_head *bh, int rw);
 | 
					int __sync_dirty_buffer(struct buffer_head *bh, int rw);
 | 
				
			||||||
void write_dirty_buffer(struct buffer_head *bh, int rw);
 | 
					void write_dirty_buffer(struct buffer_head *bh, int rw);
 | 
				
			||||||
 | 
					int _submit_bh(int rw, struct buffer_head *bh, unsigned long bio_flags);
 | 
				
			||||||
int submit_bh(int, struct buffer_head *);
 | 
					int submit_bh(int, struct buffer_head *);
 | 
				
			||||||
void write_boundary_block(struct block_device *bdev,
 | 
					void write_boundary_block(struct block_device *bdev,
 | 
				
			||||||
			sector_t bblock, unsigned blocksize);
 | 
								sector_t bblock, unsigned blocksize);
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
| 
						 | 
					@ -88,7 +88,6 @@ struct inodes_stat_t {
 | 
				
			||||||
#define MS_STRICTATIME	(1<<24) /* Always perform atime updates */
 | 
					#define MS_STRICTATIME	(1<<24) /* Always perform atime updates */
 | 
				
			||||||
 | 
					
 | 
				
			||||||
/* These sb flags are internal to the kernel */
 | 
					/* These sb flags are internal to the kernel */
 | 
				
			||||||
#define MS_SNAP_STABLE	(1<<27) /* Snapshot pages during writeback, if needed */
 | 
					 | 
				
			||||||
#define MS_NOSEC	(1<<28)
 | 
					#define MS_NOSEC	(1<<28)
 | 
				
			||||||
#define MS_BORN		(1<<29)
 | 
					#define MS_BORN		(1<<29)
 | 
				
			||||||
#define MS_ACTIVE	(1<<30)
 | 
					#define MS_ACTIVE	(1<<30)
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
							
								
								
									
										21
									
								
								mm/bounce.c
									
									
									
									
									
								
							
							
						
						
									
										21
									
								
								mm/bounce.c
									
									
									
									
									
								
							| 
						 | 
					@ -181,32 +181,13 @@ static void bounce_end_io_read_isa(struct bio *bio, int err)
 | 
				
			||||||
#ifdef CONFIG_NEED_BOUNCE_POOL
 | 
					#ifdef CONFIG_NEED_BOUNCE_POOL
 | 
				
			||||||
static int must_snapshot_stable_pages(struct request_queue *q, struct bio *bio)
 | 
					static int must_snapshot_stable_pages(struct request_queue *q, struct bio *bio)
 | 
				
			||||||
{
 | 
					{
 | 
				
			||||||
	struct page *page;
 | 
					 | 
				
			||||||
	struct backing_dev_info *bdi;
 | 
					 | 
				
			||||||
	struct address_space *mapping;
 | 
					 | 
				
			||||||
	struct bio_vec *from;
 | 
					 | 
				
			||||||
	int i;
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
	if (bio_data_dir(bio) != WRITE)
 | 
						if (bio_data_dir(bio) != WRITE)
 | 
				
			||||||
		return 0;
 | 
							return 0;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	if (!bdi_cap_stable_pages_required(&q->backing_dev_info))
 | 
						if (!bdi_cap_stable_pages_required(&q->backing_dev_info))
 | 
				
			||||||
		return 0;
 | 
							return 0;
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	/*
 | 
						return test_bit(BIO_SNAP_STABLE, &bio->bi_flags);
 | 
				
			||||||
	 * Based on the first page that has a valid mapping, decide whether or
 | 
					 | 
				
			||||||
	 * not we have to employ bounce buffering to guarantee stable pages.
 | 
					 | 
				
			||||||
	 */
 | 
					 | 
				
			||||||
	bio_for_each_segment(from, bio, i) {
 | 
					 | 
				
			||||||
		page = from->bv_page;
 | 
					 | 
				
			||||||
		mapping = page_mapping(page);
 | 
					 | 
				
			||||||
		if (!mapping)
 | 
					 | 
				
			||||||
			continue;
 | 
					 | 
				
			||||||
		bdi = mapping->backing_dev_info;
 | 
					 | 
				
			||||||
		return mapping->host->i_sb->s_flags & MS_SNAP_STABLE;
 | 
					 | 
				
			||||||
	}
 | 
					 | 
				
			||||||
 | 
					 | 
				
			||||||
	return 0;
 | 
					 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
#else
 | 
					#else
 | 
				
			||||||
static int must_snapshot_stable_pages(struct request_queue *q, struct bio *bio)
 | 
					static int must_snapshot_stable_pages(struct request_queue *q, struct bio *bio)
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
| 
						 | 
					@ -2311,10 +2311,6 @@ void wait_for_stable_page(struct page *page)
 | 
				
			||||||
 | 
					
 | 
				
			||||||
	if (!bdi_cap_stable_pages_required(bdi))
 | 
						if (!bdi_cap_stable_pages_required(bdi))
 | 
				
			||||||
		return;
 | 
							return;
 | 
				
			||||||
#ifdef CONFIG_NEED_BOUNCE_POOL
 | 
					 | 
				
			||||||
	if (mapping->host->i_sb->s_flags & MS_SNAP_STABLE)
 | 
					 | 
				
			||||||
		return;
 | 
					 | 
				
			||||||
#endif /* CONFIG_NEED_BOUNCE_POOL */
 | 
					 | 
				
			||||||
 | 
					
 | 
				
			||||||
	wait_on_page_writeback(page);
 | 
						wait_on_page_writeback(page);
 | 
				
			||||||
}
 | 
					}
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
		Loading…
	
		Reference in a new issue