mirror of
				https://github.com/torvalds/linux.git
				synced 2025-10-31 16:48:26 +02:00 
			
		
		
		
	 3e90417f41
			
		
	
	
		3e90417f41
		
	
	
	
	
		
			
			Since commit a2ad63daa8 ("VFS: add FMODE_CAN_ODIRECT file flag") file
systems can just set the FMODE_CAN_ODIRECT flag at open time instead of
wiring up a dummy direct_IO method to indicate support for direct I/O.
Signed-off-by: Ritesh Harjani (IBM) <ritesh.list@gmail.com>
Signed-off-by: Jan Kara <jack@suse.cz>
Message-Id: <94f78492f55c3f421359fb6e0d8fab6e79ea17b2.1709215665.git.ritesh.list@gmail.com>
		
	
			
		
			
				
	
	
		
			337 lines
		
	
	
	
		
			8.7 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			337 lines
		
	
	
	
		
			8.7 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
| // SPDX-License-Identifier: GPL-2.0
 | |
| /*
 | |
|  *  linux/fs/ext2/file.c
 | |
|  *
 | |
|  * Copyright (C) 1992, 1993, 1994, 1995
 | |
|  * Remy Card (card@masi.ibp.fr)
 | |
|  * Laboratoire MASI - Institut Blaise Pascal
 | |
|  * Universite Pierre et Marie Curie (Paris VI)
 | |
|  *
 | |
|  *  from
 | |
|  *
 | |
|  *  linux/fs/minix/file.c
 | |
|  *
 | |
|  *  Copyright (C) 1991, 1992  Linus Torvalds
 | |
|  *
 | |
|  *  ext2 fs regular file handling primitives
 | |
|  *
 | |
|  *  64-bit file support on 64-bit platforms by Jakub Jelinek
 | |
|  * 	(jj@sunsite.ms.mff.cuni.cz)
 | |
|  */
 | |
| 
 | |
| #include <linux/time.h>
 | |
| #include <linux/pagemap.h>
 | |
| #include <linux/dax.h>
 | |
| #include <linux/quotaops.h>
 | |
| #include <linux/iomap.h>
 | |
| #include <linux/uio.h>
 | |
| #include <linux/buffer_head.h>
 | |
| #include "ext2.h"
 | |
| #include "xattr.h"
 | |
| #include "acl.h"
 | |
| #include "trace.h"
 | |
| 
 | |
| #ifdef CONFIG_FS_DAX
 | |
| static ssize_t ext2_dax_read_iter(struct kiocb *iocb, struct iov_iter *to)
 | |
| {
 | |
| 	struct inode *inode = iocb->ki_filp->f_mapping->host;
 | |
| 	ssize_t ret;
 | |
| 
 | |
| 	if (!iov_iter_count(to))
 | |
| 		return 0; /* skip atime */
 | |
| 
 | |
| 	inode_lock_shared(inode);
 | |
| 	ret = dax_iomap_rw(iocb, to, &ext2_iomap_ops);
 | |
| 	inode_unlock_shared(inode);
 | |
| 
 | |
| 	file_accessed(iocb->ki_filp);
 | |
| 	return ret;
 | |
| }
 | |
| 
 | |
| static ssize_t ext2_dax_write_iter(struct kiocb *iocb, struct iov_iter *from)
 | |
| {
 | |
| 	struct file *file = iocb->ki_filp;
 | |
| 	struct inode *inode = file->f_mapping->host;
 | |
| 	ssize_t ret;
 | |
| 
 | |
| 	inode_lock(inode);
 | |
| 	ret = generic_write_checks(iocb, from);
 | |
| 	if (ret <= 0)
 | |
| 		goto out_unlock;
 | |
| 	ret = file_remove_privs(file);
 | |
| 	if (ret)
 | |
| 		goto out_unlock;
 | |
| 	ret = file_update_time(file);
 | |
| 	if (ret)
 | |
| 		goto out_unlock;
 | |
| 
 | |
| 	ret = dax_iomap_rw(iocb, from, &ext2_iomap_ops);
 | |
| 	if (ret > 0 && iocb->ki_pos > i_size_read(inode)) {
 | |
| 		i_size_write(inode, iocb->ki_pos);
 | |
| 		mark_inode_dirty(inode);
 | |
| 	}
 | |
| 
 | |
| out_unlock:
 | |
| 	inode_unlock(inode);
 | |
| 	if (ret > 0)
 | |
| 		ret = generic_write_sync(iocb, ret);
 | |
| 	return ret;
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * The lock ordering for ext2 DAX fault paths is:
 | |
|  *
 | |
|  * mmap_lock (MM)
 | |
|  *   sb_start_pagefault (vfs, freeze)
 | |
|  *     address_space->invalidate_lock
 | |
|  *       address_space->i_mmap_rwsem or page_lock (mutually exclusive in DAX)
 | |
|  *         ext2_inode_info->truncate_mutex
 | |
|  *
 | |
|  * The default page_lock and i_size verification done by non-DAX fault paths
 | |
|  * is sufficient because ext2 doesn't support hole punching.
 | |
|  */
 | |
| static vm_fault_t ext2_dax_fault(struct vm_fault *vmf)
 | |
| {
 | |
| 	struct inode *inode = file_inode(vmf->vma->vm_file);
 | |
| 	vm_fault_t ret;
 | |
| 	bool write = (vmf->flags & FAULT_FLAG_WRITE) &&
 | |
| 		(vmf->vma->vm_flags & VM_SHARED);
 | |
| 
 | |
| 	if (write) {
 | |
| 		sb_start_pagefault(inode->i_sb);
 | |
| 		file_update_time(vmf->vma->vm_file);
 | |
| 	}
 | |
| 	filemap_invalidate_lock_shared(inode->i_mapping);
 | |
| 
 | |
| 	ret = dax_iomap_fault(vmf, 0, NULL, NULL, &ext2_iomap_ops);
 | |
| 
 | |
| 	filemap_invalidate_unlock_shared(inode->i_mapping);
 | |
| 	if (write)
 | |
| 		sb_end_pagefault(inode->i_sb);
 | |
| 	return ret;
 | |
| }
 | |
| 
 | |
| static const struct vm_operations_struct ext2_dax_vm_ops = {
 | |
| 	.fault		= ext2_dax_fault,
 | |
| 	/*
 | |
| 	 * .huge_fault is not supported for DAX because allocation in ext2
 | |
| 	 * cannot be reliably aligned to huge page sizes and so pmd faults
 | |
| 	 * will always fail and fail back to regular faults.
 | |
| 	 */
 | |
| 	.page_mkwrite	= ext2_dax_fault,
 | |
| 	.pfn_mkwrite	= ext2_dax_fault,
 | |
| };
 | |
| 
 | |
| static int ext2_file_mmap(struct file *file, struct vm_area_struct *vma)
 | |
| {
 | |
| 	if (!IS_DAX(file_inode(file)))
 | |
| 		return generic_file_mmap(file, vma);
 | |
| 
 | |
| 	file_accessed(file);
 | |
| 	vma->vm_ops = &ext2_dax_vm_ops;
 | |
| 	return 0;
 | |
| }
 | |
| #else
 | |
| #define ext2_file_mmap	generic_file_mmap
 | |
| #endif
 | |
| 
 | |
| /*
 | |
|  * Called when filp is released. This happens when all file descriptors
 | |
|  * for a single struct file are closed. Note that different open() calls
 | |
|  * for the same file yield different struct file structures.
 | |
|  */
 | |
| static int ext2_release_file (struct inode * inode, struct file * filp)
 | |
| {
 | |
| 	if (filp->f_mode & FMODE_WRITE) {
 | |
| 		mutex_lock(&EXT2_I(inode)->truncate_mutex);
 | |
| 		ext2_discard_reservation(inode);
 | |
| 		mutex_unlock(&EXT2_I(inode)->truncate_mutex);
 | |
| 	}
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| int ext2_fsync(struct file *file, loff_t start, loff_t end, int datasync)
 | |
| {
 | |
| 	int ret;
 | |
| 	struct super_block *sb = file->f_mapping->host->i_sb;
 | |
| 
 | |
| 	ret = generic_buffers_fsync(file, start, end, datasync);
 | |
| 	if (ret == -EIO)
 | |
| 		/* We don't really know where the IO error happened... */
 | |
| 		ext2_error(sb, __func__,
 | |
| 			   "detected IO error when writing metadata buffers");
 | |
| 	return ret;
 | |
| }
 | |
| 
 | |
| static ssize_t ext2_dio_read_iter(struct kiocb *iocb, struct iov_iter *to)
 | |
| {
 | |
| 	struct file *file = iocb->ki_filp;
 | |
| 	struct inode *inode = file->f_mapping->host;
 | |
| 	ssize_t ret;
 | |
| 
 | |
| 	trace_ext2_dio_read_begin(iocb, to, 0);
 | |
| 	inode_lock_shared(inode);
 | |
| 	ret = iomap_dio_rw(iocb, to, &ext2_iomap_ops, NULL, 0, NULL, 0);
 | |
| 	inode_unlock_shared(inode);
 | |
| 	trace_ext2_dio_read_end(iocb, to, ret);
 | |
| 
 | |
| 	return ret;
 | |
| }
 | |
| 
 | |
| static int ext2_dio_write_end_io(struct kiocb *iocb, ssize_t size,
 | |
| 				 int error, unsigned int flags)
 | |
| {
 | |
| 	loff_t pos = iocb->ki_pos;
 | |
| 	struct inode *inode = file_inode(iocb->ki_filp);
 | |
| 
 | |
| 	if (error)
 | |
| 		goto out;
 | |
| 
 | |
| 	/*
 | |
| 	 * If we are extending the file, we have to update i_size here before
 | |
| 	 * page cache gets invalidated in iomap_dio_rw(). This prevents racing
 | |
| 	 * buffered reads from zeroing out too much from page cache pages.
 | |
| 	 * Note that all extending writes always happens synchronously with
 | |
| 	 * inode lock held by ext2_dio_write_iter(). So it is safe to update
 | |
| 	 * inode size here for extending file writes.
 | |
| 	 */
 | |
| 	pos += size;
 | |
| 	if (pos > i_size_read(inode)) {
 | |
| 		i_size_write(inode, pos);
 | |
| 		mark_inode_dirty(inode);
 | |
| 	}
 | |
| out:
 | |
| 	trace_ext2_dio_write_endio(iocb, size, error);
 | |
| 	return error;
 | |
| }
 | |
| 
 | |
| static const struct iomap_dio_ops ext2_dio_write_ops = {
 | |
| 	.end_io = ext2_dio_write_end_io,
 | |
| };
 | |
| 
 | |
| static ssize_t ext2_dio_write_iter(struct kiocb *iocb, struct iov_iter *from)
 | |
| {
 | |
| 	struct file *file = iocb->ki_filp;
 | |
| 	struct inode *inode = file->f_mapping->host;
 | |
| 	ssize_t ret;
 | |
| 	unsigned int flags = 0;
 | |
| 	unsigned long blocksize = inode->i_sb->s_blocksize;
 | |
| 	loff_t offset = iocb->ki_pos;
 | |
| 	loff_t count = iov_iter_count(from);
 | |
| 	ssize_t status = 0;
 | |
| 
 | |
| 	trace_ext2_dio_write_begin(iocb, from, 0);
 | |
| 	inode_lock(inode);
 | |
| 	ret = generic_write_checks(iocb, from);
 | |
| 	if (ret <= 0)
 | |
| 		goto out_unlock;
 | |
| 
 | |
| 	ret = kiocb_modified(iocb);
 | |
| 	if (ret)
 | |
| 		goto out_unlock;
 | |
| 
 | |
| 	/* use IOMAP_DIO_FORCE_WAIT for unaligned or extending writes */
 | |
| 	if (iocb->ki_pos + iov_iter_count(from) > i_size_read(inode) ||
 | |
| 	   (!IS_ALIGNED(iocb->ki_pos | iov_iter_alignment(from), blocksize)))
 | |
| 		flags |= IOMAP_DIO_FORCE_WAIT;
 | |
| 
 | |
| 	ret = iomap_dio_rw(iocb, from, &ext2_iomap_ops, &ext2_dio_write_ops,
 | |
| 			   flags, NULL, 0);
 | |
| 
 | |
| 	/* ENOTBLK is magic return value for fallback to buffered-io */
 | |
| 	if (ret == -ENOTBLK)
 | |
| 		ret = 0;
 | |
| 
 | |
| 	if (ret < 0 && ret != -EIOCBQUEUED)
 | |
| 		ext2_write_failed(inode->i_mapping, offset + count);
 | |
| 
 | |
| 	/* handle case for partial write and for fallback to buffered write */
 | |
| 	if (ret >= 0 && iov_iter_count(from)) {
 | |
| 		loff_t pos, endbyte;
 | |
| 		int ret2;
 | |
| 
 | |
| 		iocb->ki_flags &= ~IOCB_DIRECT;
 | |
| 		pos = iocb->ki_pos;
 | |
| 		status = generic_perform_write(iocb, from);
 | |
| 		if (unlikely(status < 0)) {
 | |
| 			ret = status;
 | |
| 			goto out_unlock;
 | |
| 		}
 | |
| 
 | |
| 		ret += status;
 | |
| 		endbyte = pos + status - 1;
 | |
| 		ret2 = filemap_write_and_wait_range(inode->i_mapping, pos,
 | |
| 						    endbyte);
 | |
| 		if (!ret2)
 | |
| 			invalidate_mapping_pages(inode->i_mapping,
 | |
| 						 pos >> PAGE_SHIFT,
 | |
| 						 endbyte >> PAGE_SHIFT);
 | |
| 		if (ret > 0)
 | |
| 			generic_write_sync(iocb, ret);
 | |
| 	}
 | |
| 
 | |
| out_unlock:
 | |
| 	inode_unlock(inode);
 | |
| 	if (status)
 | |
| 		trace_ext2_dio_write_buff_end(iocb, from, status);
 | |
| 	trace_ext2_dio_write_end(iocb, from, ret);
 | |
| 	return ret;
 | |
| }
 | |
| 
 | |
| static ssize_t ext2_file_read_iter(struct kiocb *iocb, struct iov_iter *to)
 | |
| {
 | |
| #ifdef CONFIG_FS_DAX
 | |
| 	if (IS_DAX(iocb->ki_filp->f_mapping->host))
 | |
| 		return ext2_dax_read_iter(iocb, to);
 | |
| #endif
 | |
| 	if (iocb->ki_flags & IOCB_DIRECT)
 | |
| 		return ext2_dio_read_iter(iocb, to);
 | |
| 
 | |
| 	return generic_file_read_iter(iocb, to);
 | |
| }
 | |
| 
 | |
| static ssize_t ext2_file_write_iter(struct kiocb *iocb, struct iov_iter *from)
 | |
| {
 | |
| #ifdef CONFIG_FS_DAX
 | |
| 	if (IS_DAX(iocb->ki_filp->f_mapping->host))
 | |
| 		return ext2_dax_write_iter(iocb, from);
 | |
| #endif
 | |
| 	if (iocb->ki_flags & IOCB_DIRECT)
 | |
| 		return ext2_dio_write_iter(iocb, from);
 | |
| 
 | |
| 	return generic_file_write_iter(iocb, from);
 | |
| }
 | |
| 
 | |
| static int ext2_file_open(struct inode *inode, struct file *filp)
 | |
| {
 | |
| 	filp->f_mode |= FMODE_CAN_ODIRECT;
 | |
| 	return dquot_file_open(inode, filp);
 | |
| }
 | |
| 
 | |
| const struct file_operations ext2_file_operations = {
 | |
| 	.llseek		= generic_file_llseek,
 | |
| 	.read_iter	= ext2_file_read_iter,
 | |
| 	.write_iter	= ext2_file_write_iter,
 | |
| 	.unlocked_ioctl = ext2_ioctl,
 | |
| #ifdef CONFIG_COMPAT
 | |
| 	.compat_ioctl	= ext2_compat_ioctl,
 | |
| #endif
 | |
| 	.mmap		= ext2_file_mmap,
 | |
| 	.open		= ext2_file_open,
 | |
| 	.release	= ext2_release_file,
 | |
| 	.fsync		= ext2_fsync,
 | |
| 	.get_unmapped_area = thp_get_unmapped_area,
 | |
| 	.splice_read	= filemap_splice_read,
 | |
| 	.splice_write	= iter_file_splice_write,
 | |
| };
 | |
| 
 | |
| const struct inode_operations ext2_file_inode_operations = {
 | |
| 	.listxattr	= ext2_listxattr,
 | |
| 	.getattr	= ext2_getattr,
 | |
| 	.setattr	= ext2_setattr,
 | |
| 	.get_inode_acl	= ext2_get_acl,
 | |
| 	.set_acl	= ext2_set_acl,
 | |
| 	.fiemap		= ext2_fiemap,
 | |
| 	.fileattr_get	= ext2_fileattr_get,
 | |
| 	.fileattr_set	= ext2_fileattr_set,
 | |
| };
 |