2006-10-11 10:20:50 +02:00
|
|
|
/*
|
2006-10-11 10:20:53 +02:00
|
|
|
* linux/fs/ext4/file.c
|
2006-10-11 10:20:50 +02:00
|
|
|
*
|
|
|
|
* Copyright (C) 1992, 1993, 1994, 1995
|
|
|
|
* Remy Card (card@masi.ibp.fr)
|
|
|
|
* Laboratoire MASI - Institut Blaise Pascal
|
|
|
|
* Universite Pierre et Marie Curie (Paris VI)
|
|
|
|
*
|
|
|
|
* from
|
|
|
|
*
|
|
|
|
* linux/fs/minix/file.c
|
|
|
|
*
|
|
|
|
* Copyright (C) 1991, 1992 Linus Torvalds
|
|
|
|
*
|
2006-10-11 10:20:53 +02:00
|
|
|
* ext4 fs regular file handling primitives
|
2006-10-11 10:20:50 +02:00
|
|
|
*
|
|
|
|
* 64-bit file support on 64-bit platforms by Jakub Jelinek
|
|
|
|
* (jj@sunsite.ms.mff.cuni.cz)
|
|
|
|
*/
|
|
|
|
|
|
|
|
#include <linux/time.h>
|
|
|
|
#include <linux/fs.h>
|
2009-06-13 16:09:48 +02:00
|
|
|
#include <linux/mount.h>
|
|
|
|
#include <linux/path.h>
|
2015-09-08 23:58:40 +02:00
|
|
|
#include <linux/dax.h>
|
2010-03-03 15:05:07 +01:00
|
|
|
#include <linux/quotaops.h>
|
2012-11-09 03:57:40 +01:00
|
|
|
#include <linux/pagevec.h>
|
2015-02-22 17:58:50 +01:00
|
|
|
#include <linux/uio.h>
|
2008-04-30 00:13:32 +02:00
|
|
|
#include "ext4.h"
|
|
|
|
#include "ext4_jbd2.h"
|
2006-10-11 10:20:50 +02:00
|
|
|
#include "xattr.h"
|
|
|
|
#include "acl.h"
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Called when an inode is released. Note that this is different
|
2006-10-11 10:20:53 +02:00
|
|
|
* from ext4_file_open: open gets called at every open, but release
|
2006-10-11 10:20:50 +02:00
|
|
|
* gets called only when /all/ the files are closed.
|
|
|
|
*/
|
2008-09-09 04:25:24 +02:00
|
|
|
static int ext4_release_file(struct inode *inode, struct file *filp)
|
2006-10-11 10:20:50 +02:00
|
|
|
{
|
2010-01-24 20:34:07 +01:00
|
|
|
if (ext4_test_inode_state(inode, EXT4_STATE_DA_ALLOC_CLOSE)) {
|
2009-02-24 14:21:14 +01:00
|
|
|
ext4_alloc_da_blocks(inode);
|
2010-01-24 20:34:07 +01:00
|
|
|
ext4_clear_inode_state(inode, EXT4_STATE_DA_ALLOC_CLOSE);
|
2009-02-24 14:21:14 +01:00
|
|
|
}
|
2006-10-11 10:20:50 +02:00
|
|
|
/* if we are the last writer on the inode, drop the block reservation */
|
|
|
|
if ((filp->f_mode & FMODE_WRITE) &&
|
2009-03-28 03:36:43 +01:00
|
|
|
(atomic_read(&inode->i_writecount) == 1) &&
|
|
|
|
!EXT4_I(inode)->i_reserved_data_blocks)
|
2006-10-11 10:20:50 +02:00
|
|
|
{
|
2008-01-29 05:58:26 +01:00
|
|
|
down_write(&EXT4_I(inode)->i_data_sem);
|
2008-10-10 15:40:52 +02:00
|
|
|
ext4_discard_preallocations(inode);
|
2008-01-29 05:58:26 +01:00
|
|
|
up_write(&EXT4_I(inode)->i_data_sem);
|
2006-10-11 10:20:50 +02:00
|
|
|
}
|
|
|
|
if (is_dx(inode) && filp->private_data)
|
2006-10-11 10:20:53 +02:00
|
|
|
ext4_htree_free_dir_info(filp->private_data);
|
2006-10-11 10:20:50 +02:00
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2014-05-12 16:50:23 +02:00
|
|
|
static void ext4_unwritten_wait(struct inode *inode)
|
ext4: serialize unaligned asynchronous DIO
ext4 has a data corruption case when doing non-block-aligned
asynchronous direct IO into a sparse file, as demonstrated
by xfstest 240.
The root cause is that while ext4 preallocates space in the
hole, mappings of that space still look "new" and
dio_zero_block() will zero out the unwritten portions. When
more than one AIO thread is going, they both find this "new"
block and race to zero out their portion; this is uncoordinated
and causes data corruption.
Dave Chinner fixed this for xfs by simply serializing all
unaligned asynchronous direct IO. I've done the same here.
The difference is that we only wait on conversions, not all IO.
This is a very big hammer, and I'm not very pleased with
stuffing this into ext4_file_write(). But since ext4 is
DIO_LOCKING, we need to serialize it at this high level.
I tried to move this into ext4_ext_direct_IO, but by then
we have the i_mutex already, and we will wait on the
work queue to do conversions - which must also take the
i_mutex. So that won't work.
This was originally exposed by qemu-kvm installing to
a raw disk image with a normal sector-63 alignment. I've
tested a backport of this patch with qemu, and it does
avoid the corruption. It is also quite a lot slower
(14 min for package installs, vs. 8 min for well-aligned)
but I'll take slow correctness over fast corruption any day.
Mingming suggested that we can track outstanding
conversions, and wait on those so that non-sparse
files won't be affected, and I've implemented that here;
unaligned AIO to nonsparse files won't take a perf hit.
[tytso@mit.edu: Keep the mutex as a hashed array instead
of bloating the ext4 inode]
[tytso@mit.edu: Fix up namespace issues so that global
variables are protected with an "ext4_" prefix.]
Signed-off-by: Eric Sandeen <sandeen@redhat.com>
Signed-off-by: "Theodore Ts'o" <tytso@mit.edu>
2011-02-12 14:17:34 +01:00
|
|
|
{
|
|
|
|
wait_queue_head_t *wq = ext4_ioend_wq(inode);
|
|
|
|
|
2012-09-29 05:24:52 +02:00
|
|
|
wait_event(*wq, (atomic_read(&EXT4_I(inode)->i_unwritten) == 0));
|
ext4: serialize unaligned asynchronous DIO
ext4 has a data corruption case when doing non-block-aligned
asynchronous direct IO into a sparse file, as demonstrated
by xfstest 240.
The root cause is that while ext4 preallocates space in the
hole, mappings of that space still look "new" and
dio_zero_block() will zero out the unwritten portions. When
more than one AIO thread is going, they both find this "new"
block and race to zero out their portion; this is uncoordinated
and causes data corruption.
Dave Chinner fixed this for xfs by simply serializing all
unaligned asynchronous direct IO. I've done the same here.
The difference is that we only wait on conversions, not all IO.
This is a very big hammer, and I'm not very pleased with
stuffing this into ext4_file_write(). But since ext4 is
DIO_LOCKING, we need to serialize it at this high level.
I tried to move this into ext4_ext_direct_IO, but by then
we have the i_mutex already, and we will wait on the
work queue to do conversions - which must also take the
i_mutex. So that won't work.
This was originally exposed by qemu-kvm installing to
a raw disk image with a normal sector-63 alignment. I've
tested a backport of this patch with qemu, and it does
avoid the corruption. It is also quite a lot slower
(14 min for package installs, vs. 8 min for well-aligned)
but I'll take slow correctness over fast corruption any day.
Mingming suggested that we can track outstanding
conversions, and wait on those so that non-sparse
files won't be affected, and I've implemented that here;
unaligned AIO to nonsparse files won't take a perf hit.
[tytso@mit.edu: Keep the mutex as a hashed array instead
of bloating the ext4 inode]
[tytso@mit.edu: Fix up namespace issues so that global
variables are protected with an "ext4_" prefix.]
Signed-off-by: Eric Sandeen <sandeen@redhat.com>
Signed-off-by: "Theodore Ts'o" <tytso@mit.edu>
2011-02-12 14:17:34 +01:00
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* This tests whether the IO in question is block-aligned or not.
|
|
|
|
* Ext4 utilizes unwritten extents when hole-filling during direct IO, and they
|
|
|
|
* are converted to written only after the IO is complete. Until they are
|
|
|
|
* mapped, these blocks appear as holes, so dio_zero_block() will assume that
|
|
|
|
* it needs to zero out portions of the start and/or end block. If 2 AIO
|
|
|
|
* threads are at work on the same unwritten block, they must be synchronized
|
|
|
|
* or one thread will zero the other's data, causing corruption.
|
|
|
|
*/
|
|
|
|
static int
|
2014-04-17 22:09:22 +02:00
|
|
|
ext4_unaligned_aio(struct inode *inode, struct iov_iter *from, loff_t pos)
|
ext4: serialize unaligned asynchronous DIO
ext4 has a data corruption case when doing non-block-aligned
asynchronous direct IO into a sparse file, as demonstrated
by xfstest 240.
The root cause is that while ext4 preallocates space in the
hole, mappings of that space still look "new" and
dio_zero_block() will zero out the unwritten portions. When
more than one AIO thread is going, they both find this "new"
block and race to zero out their portion; this is uncoordinated
and causes data corruption.
Dave Chinner fixed this for xfs by simply serializing all
unaligned asynchronous direct IO. I've done the same here.
The difference is that we only wait on conversions, not all IO.
This is a very big hammer, and I'm not very pleased with
stuffing this into ext4_file_write(). But since ext4 is
DIO_LOCKING, we need to serialize it at this high level.
I tried to move this into ext4_ext_direct_IO, but by then
we have the i_mutex already, and we will wait on the
work queue to do conversions - which must also take the
i_mutex. So that won't work.
This was originally exposed by qemu-kvm installing to
a raw disk image with a normal sector-63 alignment. I've
tested a backport of this patch with qemu, and it does
avoid the corruption. It is also quite a lot slower
(14 min for package installs, vs. 8 min for well-aligned)
but I'll take slow correctness over fast corruption any day.
Mingming suggested that we can track outstanding
conversions, and wait on those so that non-sparse
files won't be affected, and I've implemented that here;
unaligned AIO to nonsparse files won't take a perf hit.
[tytso@mit.edu: Keep the mutex as a hashed array instead
of bloating the ext4 inode]
[tytso@mit.edu: Fix up namespace issues so that global
variables are protected with an "ext4_" prefix.]
Signed-off-by: Eric Sandeen <sandeen@redhat.com>
Signed-off-by: "Theodore Ts'o" <tytso@mit.edu>
2011-02-12 14:17:34 +01:00
|
|
|
{
|
|
|
|
struct super_block *sb = inode->i_sb;
|
|
|
|
int blockmask = sb->s_blocksize - 1;
|
|
|
|
|
2014-04-12 18:45:25 +02:00
|
|
|
if (pos >= i_size_read(inode))
|
ext4: serialize unaligned asynchronous DIO
ext4 has a data corruption case when doing non-block-aligned
asynchronous direct IO into a sparse file, as demonstrated
by xfstest 240.
The root cause is that while ext4 preallocates space in the
hole, mappings of that space still look "new" and
dio_zero_block() will zero out the unwritten portions. When
more than one AIO thread is going, they both find this "new"
block and race to zero out their portion; this is uncoordinated
and causes data corruption.
Dave Chinner fixed this for xfs by simply serializing all
unaligned asynchronous direct IO. I've done the same here.
The difference is that we only wait on conversions, not all IO.
This is a very big hammer, and I'm not very pleased with
stuffing this into ext4_file_write(). But since ext4 is
DIO_LOCKING, we need to serialize it at this high level.
I tried to move this into ext4_ext_direct_IO, but by then
we have the i_mutex already, and we will wait on the
work queue to do conversions - which must also take the
i_mutex. So that won't work.
This was originally exposed by qemu-kvm installing to
a raw disk image with a normal sector-63 alignment. I've
tested a backport of this patch with qemu, and it does
avoid the corruption. It is also quite a lot slower
(14 min for package installs, vs. 8 min for well-aligned)
but I'll take slow correctness over fast corruption any day.
Mingming suggested that we can track outstanding
conversions, and wait on those so that non-sparse
files won't be affected, and I've implemented that here;
unaligned AIO to nonsparse files won't take a perf hit.
[tytso@mit.edu: Keep the mutex as a hashed array instead
of bloating the ext4 inode]
[tytso@mit.edu: Fix up namespace issues so that global
variables are protected with an "ext4_" prefix.]
Signed-off-by: Eric Sandeen <sandeen@redhat.com>
Signed-off-by: "Theodore Ts'o" <tytso@mit.edu>
2011-02-12 14:17:34 +01:00
|
|
|
return 0;
|
|
|
|
|
2014-04-17 22:09:22 +02:00
|
|
|
if ((pos | iov_iter_alignment(from)) & blockmask)
|
ext4: serialize unaligned asynchronous DIO
ext4 has a data corruption case when doing non-block-aligned
asynchronous direct IO into a sparse file, as demonstrated
by xfstest 240.
The root cause is that while ext4 preallocates space in the
hole, mappings of that space still look "new" and
dio_zero_block() will zero out the unwritten portions. When
more than one AIO thread is going, they both find this "new"
block and race to zero out their portion; this is uncoordinated
and causes data corruption.
Dave Chinner fixed this for xfs by simply serializing all
unaligned asynchronous direct IO. I've done the same here.
The difference is that we only wait on conversions, not all IO.
This is a very big hammer, and I'm not very pleased with
stuffing this into ext4_file_write(). But since ext4 is
DIO_LOCKING, we need to serialize it at this high level.
I tried to move this into ext4_ext_direct_IO, but by then
we have the i_mutex already, and we will wait on the
work queue to do conversions - which must also take the
i_mutex. So that won't work.
This was originally exposed by qemu-kvm installing to
a raw disk image with a normal sector-63 alignment. I've
tested a backport of this patch with qemu, and it does
avoid the corruption. It is also quite a lot slower
(14 min for package installs, vs. 8 min for well-aligned)
but I'll take slow correctness over fast corruption any day.
Mingming suggested that we can track outstanding
conversions, and wait on those so that non-sparse
files won't be affected, and I've implemented that here;
unaligned AIO to nonsparse files won't take a perf hit.
[tytso@mit.edu: Keep the mutex as a hashed array instead
of bloating the ext4 inode]
[tytso@mit.edu: Fix up namespace issues so that global
variables are protected with an "ext4_" prefix.]
Signed-off-by: Eric Sandeen <sandeen@redhat.com>
Signed-off-by: "Theodore Ts'o" <tytso@mit.edu>
2011-02-12 14:17:34 +01:00
|
|
|
return 1;
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2006-10-11 10:20:50 +02:00
|
|
|
static ssize_t
|
2014-04-17 22:09:22 +02:00
|
|
|
ext4_file_write_iter(struct kiocb *iocb, struct iov_iter *from)
|
2006-10-11 10:20:50 +02:00
|
|
|
{
|
2012-07-23 02:19:31 +02:00
|
|
|
struct file *file = iocb->ki_filp;
|
2014-04-21 20:26:57 +02:00
|
|
|
struct inode *inode = file_inode(iocb->ki_filp);
|
2014-04-21 20:36:30 +02:00
|
|
|
struct mutex *aio_mutex = NULL;
|
2012-07-23 02:19:31 +02:00
|
|
|
struct blk_plug plug;
|
2015-04-09 19:52:01 +02:00
|
|
|
int o_direct = iocb->ki_flags & IOCB_DIRECT;
|
2012-07-23 02:19:31 +02:00
|
|
|
int overwrite = 0;
|
2012-05-29 00:06:51 +02:00
|
|
|
ssize_t ret;
|
2014-04-21 20:26:28 +02:00
|
|
|
|
2014-04-21 20:37:52 +02:00
|
|
|
/*
|
|
|
|
* Unaligned direct AIO must be serialized; see comment above
|
|
|
|
* In the case of O_APPEND, assume that we must always serialize
|
|
|
|
*/
|
|
|
|
if (o_direct &&
|
|
|
|
ext4_test_inode_flag(inode, EXT4_INODE_EXTENTS) &&
|
|
|
|
!is_sync_kiocb(iocb) &&
|
2015-04-09 19:52:01 +02:00
|
|
|
(iocb->ki_flags & IOCB_APPEND ||
|
2015-04-07 20:48:22 +02:00
|
|
|
ext4_unaligned_aio(inode, from, iocb->ki_pos))) {
|
2014-04-21 20:37:52 +02:00
|
|
|
aio_mutex = ext4_aio_mutex(inode);
|
|
|
|
mutex_lock(aio_mutex);
|
|
|
|
ext4_unwritten_wait(inode);
|
|
|
|
}
|
|
|
|
|
2016-01-22 21:40:57 +01:00
|
|
|
inode_lock(inode);
|
2015-04-09 18:55:47 +02:00
|
|
|
ret = generic_write_checks(iocb, from);
|
|
|
|
if (ret <= 0)
|
2015-04-07 20:48:22 +02:00
|
|
|
goto out;
|
2014-04-21 20:37:52 +02:00
|
|
|
|
2008-01-29 05:58:27 +01:00
|
|
|
/*
|
|
|
|
* If we have encountered a bitmap-format file, the size limit
|
|
|
|
* is smaller than s_maxbytes, which is for extent-mapped files.
|
|
|
|
*/
|
2010-05-17 04:00:00 +02:00
|
|
|
if (!(ext4_test_inode_flag(inode, EXT4_INODE_EXTENTS))) {
|
2008-01-29 05:58:27 +01:00
|
|
|
struct ext4_sb_info *sbi = EXT4_SB(inode->i_sb);
|
2006-10-11 10:20:50 +02:00
|
|
|
|
2015-04-09 18:55:47 +02:00
|
|
|
if (iocb->ki_pos >= sbi->s_bitmap_maxbytes) {
|
2014-04-21 20:37:52 +02:00
|
|
|
ret = -EFBIG;
|
2015-04-07 20:48:22 +02:00
|
|
|
goto out;
|
2014-04-21 20:37:52 +02:00
|
|
|
}
|
2015-04-09 18:55:47 +02:00
|
|
|
iov_iter_truncate(from, sbi->s_bitmap_maxbytes - iocb->ki_pos);
|
2008-01-29 05:58:27 +01:00
|
|
|
}
|
|
|
|
|
2014-10-30 15:53:16 +01:00
|
|
|
iocb->private = &overwrite;
|
2014-04-21 20:36:30 +02:00
|
|
|
if (o_direct) {
|
2015-04-09 18:55:47 +02:00
|
|
|
size_t length = iov_iter_count(from);
|
|
|
|
loff_t pos = iocb->ki_pos;
|
2014-04-21 20:26:57 +02:00
|
|
|
blk_start_plug(&plug);
|
|
|
|
|
|
|
|
/* check whether we do a DIO overwrite or not */
|
2014-04-21 20:36:30 +02:00
|
|
|
if (ext4_should_dioread_nolock(inode) && !aio_mutex &&
|
2014-04-21 20:26:57 +02:00
|
|
|
!file->f_mapping->nrpages && pos + length <= i_size_read(inode)) {
|
|
|
|
struct ext4_map_blocks map;
|
|
|
|
unsigned int blkbits = inode->i_blkbits;
|
|
|
|
int err, len;
|
|
|
|
|
|
|
|
map.m_lblk = pos >> blkbits;
|
|
|
|
map.m_len = (EXT4_BLOCK_ALIGN(pos + length, blkbits) >> blkbits)
|
|
|
|
- map.m_lblk;
|
|
|
|
len = map.m_len;
|
|
|
|
|
|
|
|
err = ext4_map_blocks(NULL, inode, &map, 0);
|
|
|
|
/*
|
|
|
|
* 'err==len' means that all of blocks has
|
|
|
|
* been preallocated no matter they are
|
|
|
|
* initialized or not. For excluding
|
|
|
|
* unwritten extents, we need to check
|
|
|
|
* m_flags. There are two conditions that
|
|
|
|
* indicate for initialized extents. 1) If we
|
|
|
|
* hit extent cache, EXT4_MAP_MAPPED flag is
|
|
|
|
* returned; 2) If we do a real lookup,
|
|
|
|
* non-flags are returned. So we should check
|
|
|
|
* these two conditions.
|
|
|
|
*/
|
|
|
|
if (err == len && (map.m_flags & EXT4_MAP_MAPPED))
|
|
|
|
overwrite = 1;
|
|
|
|
}
|
2014-04-21 20:37:52 +02:00
|
|
|
}
|
2014-04-21 20:26:28 +02:00
|
|
|
|
2014-04-17 22:09:22 +02:00
|
|
|
ret = __generic_file_write_iter(iocb, from);
|
2016-01-22 21:40:57 +01:00
|
|
|
inode_unlock(inode);
|
2014-04-21 20:26:28 +02:00
|
|
|
|
2014-04-21 20:36:30 +02:00
|
|
|
if (ret > 0) {
|
|
|
|
ssize_t err;
|
|
|
|
|
|
|
|
err = generic_write_sync(file, iocb->ki_pos - ret, ret);
|
|
|
|
if (err < 0)
|
|
|
|
ret = err;
|
2014-04-21 20:26:28 +02:00
|
|
|
}
|
2014-04-21 20:36:30 +02:00
|
|
|
if (o_direct)
|
|
|
|
blk_finish_plug(&plug);
|
ext4: serialize unaligned asynchronous DIO
ext4 has a data corruption case when doing non-block-aligned
asynchronous direct IO into a sparse file, as demonstrated
by xfstest 240.
The root cause is that while ext4 preallocates space in the
hole, mappings of that space still look "new" and
dio_zero_block() will zero out the unwritten portions. When
more than one AIO thread is going, they both find this "new"
block and race to zero out their portion; this is uncoordinated
and causes data corruption.
Dave Chinner fixed this for xfs by simply serializing all
unaligned asynchronous direct IO. I've done the same here.
The difference is that we only wait on conversions, not all IO.
This is a very big hammer, and I'm not very pleased with
stuffing this into ext4_file_write(). But since ext4 is
DIO_LOCKING, we need to serialize it at this high level.
I tried to move this into ext4_ext_direct_IO, but by then
we have the i_mutex already, and we will wait on the
work queue to do conversions - which must also take the
i_mutex. So that won't work.
This was originally exposed by qemu-kvm installing to
a raw disk image with a normal sector-63 alignment. I've
tested a backport of this patch with qemu, and it does
avoid the corruption. It is also quite a lot slower
(14 min for package installs, vs. 8 min for well-aligned)
but I'll take slow correctness over fast corruption any day.
Mingming suggested that we can track outstanding
conversions, and wait on those so that non-sparse
files won't be affected, and I've implemented that here;
unaligned AIO to nonsparse files won't take a perf hit.
[tytso@mit.edu: Keep the mutex as a hashed array instead
of bloating the ext4 inode]
[tytso@mit.edu: Fix up namespace issues so that global
variables are protected with an "ext4_" prefix.]
Signed-off-by: Eric Sandeen <sandeen@redhat.com>
Signed-off-by: "Theodore Ts'o" <tytso@mit.edu>
2011-02-12 14:17:34 +01:00
|
|
|
|
2015-04-07 20:48:22 +02:00
|
|
|
if (aio_mutex)
|
|
|
|
mutex_unlock(aio_mutex);
|
|
|
|
return ret;
|
|
|
|
|
|
|
|
out:
|
2016-01-22 21:40:57 +01:00
|
|
|
inode_unlock(inode);
|
2014-04-21 20:36:30 +02:00
|
|
|
if (aio_mutex)
|
|
|
|
mutex_unlock(aio_mutex);
|
ext4: serialize unaligned asynchronous DIO
ext4 has a data corruption case when doing non-block-aligned
asynchronous direct IO into a sparse file, as demonstrated
by xfstest 240.
The root cause is that while ext4 preallocates space in the
hole, mappings of that space still look "new" and
dio_zero_block() will zero out the unwritten portions. When
more than one AIO thread is going, they both find this "new"
block and race to zero out their portion; this is uncoordinated
and causes data corruption.
Dave Chinner fixed this for xfs by simply serializing all
unaligned asynchronous direct IO. I've done the same here.
The difference is that we only wait on conversions, not all IO.
This is a very big hammer, and I'm not very pleased with
stuffing this into ext4_file_write(). But since ext4 is
DIO_LOCKING, we need to serialize it at this high level.
I tried to move this into ext4_ext_direct_IO, but by then
we have the i_mutex already, and we will wait on the
work queue to do conversions - which must also take the
i_mutex. So that won't work.
This was originally exposed by qemu-kvm installing to
a raw disk image with a normal sector-63 alignment. I've
tested a backport of this patch with qemu, and it does
avoid the corruption. It is also quite a lot slower
(14 min for package installs, vs. 8 min for well-aligned)
but I'll take slow correctness over fast corruption any day.
Mingming suggested that we can track outstanding
conversions, and wait on those so that non-sparse
files won't be affected, and I've implemented that here;
unaligned AIO to nonsparse files won't take a perf hit.
[tytso@mit.edu: Keep the mutex as a hashed array instead
of bloating the ext4 inode]
[tytso@mit.edu: Fix up namespace issues so that global
variables are protected with an "ext4_" prefix.]
Signed-off-by: Eric Sandeen <sandeen@redhat.com>
Signed-off-by: "Theodore Ts'o" <tytso@mit.edu>
2011-02-12 14:17:34 +01:00
|
|
|
return ret;
|
2006-10-11 10:20:50 +02:00
|
|
|
}
|
|
|
|
|
2015-02-17 00:59:38 +01:00
|
|
|
#ifdef CONFIG_FS_DAX
|
|
|
|
static int ext4_dax_fault(struct vm_area_struct *vma, struct vm_fault *vmf)
|
|
|
|
{
|
2015-09-08 23:59:22 +02:00
|
|
|
int result;
|
|
|
|
handle_t *handle = NULL;
|
2015-12-07 20:28:03 +01:00
|
|
|
struct inode *inode = file_inode(vma->vm_file);
|
|
|
|
struct super_block *sb = inode->i_sb;
|
2015-09-08 23:59:22 +02:00
|
|
|
bool write = vmf->flags & FAULT_FLAG_WRITE;
|
|
|
|
|
|
|
|
if (write) {
|
|
|
|
sb_start_pagefault(sb);
|
|
|
|
file_update_time(vma->vm_file);
|
2015-12-07 20:28:03 +01:00
|
|
|
down_read(&EXT4_I(inode)->i_mmap_sem);
|
2015-09-08 23:59:22 +02:00
|
|
|
handle = ext4_journal_start_sb(sb, EXT4_HT_WRITE_PAGE,
|
|
|
|
EXT4_DATA_TRANS_BLOCKS(sb));
|
2015-12-07 20:28:03 +01:00
|
|
|
} else
|
|
|
|
down_read(&EXT4_I(inode)->i_mmap_sem);
|
2015-09-08 23:59:22 +02:00
|
|
|
|
|
|
|
if (IS_ERR(handle))
|
|
|
|
result = VM_FAULT_SIGBUS;
|
|
|
|
else
|
2015-12-07 21:10:44 +01:00
|
|
|
result = __dax_fault(vma, vmf, ext4_dax_mmap_get_block, NULL);
|
2015-09-08 23:59:22 +02:00
|
|
|
|
|
|
|
if (write) {
|
|
|
|
if (!IS_ERR(handle))
|
|
|
|
ext4_journal_stop(handle);
|
2015-12-07 20:28:03 +01:00
|
|
|
up_read(&EXT4_I(inode)->i_mmap_sem);
|
2015-09-08 23:59:22 +02:00
|
|
|
sb_end_pagefault(sb);
|
2015-12-07 20:28:03 +01:00
|
|
|
} else
|
|
|
|
up_read(&EXT4_I(inode)->i_mmap_sem);
|
2015-09-08 23:59:22 +02:00
|
|
|
|
|
|
|
return result;
|
2015-02-17 00:59:38 +01:00
|
|
|
}
|
|
|
|
|
2015-09-08 23:59:03 +02:00
|
|
|
static int ext4_dax_pmd_fault(struct vm_area_struct *vma, unsigned long addr,
|
|
|
|
pmd_t *pmd, unsigned int flags)
|
|
|
|
{
|
2015-09-08 23:59:22 +02:00
|
|
|
int result;
|
|
|
|
handle_t *handle = NULL;
|
|
|
|
struct inode *inode = file_inode(vma->vm_file);
|
|
|
|
struct super_block *sb = inode->i_sb;
|
|
|
|
bool write = flags & FAULT_FLAG_WRITE;
|
|
|
|
|
|
|
|
if (write) {
|
|
|
|
sb_start_pagefault(sb);
|
|
|
|
file_update_time(vma->vm_file);
|
2015-12-07 20:28:03 +01:00
|
|
|
down_read(&EXT4_I(inode)->i_mmap_sem);
|
2015-09-08 23:59:22 +02:00
|
|
|
handle = ext4_journal_start_sb(sb, EXT4_HT_WRITE_PAGE,
|
|
|
|
ext4_chunk_trans_blocks(inode,
|
|
|
|
PMD_SIZE / PAGE_SIZE));
|
2015-12-07 20:28:03 +01:00
|
|
|
} else
|
|
|
|
down_read(&EXT4_I(inode)->i_mmap_sem);
|
2015-09-08 23:59:22 +02:00
|
|
|
|
|
|
|
if (IS_ERR(handle))
|
|
|
|
result = VM_FAULT_SIGBUS;
|
|
|
|
else
|
|
|
|
result = __dax_pmd_fault(vma, addr, pmd, flags,
|
2015-12-07 21:10:44 +01:00
|
|
|
ext4_dax_mmap_get_block, NULL);
|
2015-09-08 23:59:22 +02:00
|
|
|
|
|
|
|
if (write) {
|
|
|
|
if (!IS_ERR(handle))
|
|
|
|
ext4_journal_stop(handle);
|
2015-12-07 20:28:03 +01:00
|
|
|
up_read(&EXT4_I(inode)->i_mmap_sem);
|
2015-09-08 23:59:22 +02:00
|
|
|
sb_end_pagefault(sb);
|
2015-12-07 20:28:03 +01:00
|
|
|
} else
|
|
|
|
up_read(&EXT4_I(inode)->i_mmap_sem);
|
2015-09-08 23:59:22 +02:00
|
|
|
|
|
|
|
return result;
|
2015-09-08 23:59:03 +02:00
|
|
|
}
|
|
|
|
|
2015-02-17 00:59:38 +01:00
|
|
|
static int ext4_dax_mkwrite(struct vm_area_struct *vma, struct vm_fault *vmf)
|
|
|
|
{
|
2015-12-07 20:28:03 +01:00
|
|
|
int err;
|
|
|
|
struct inode *inode = file_inode(vma->vm_file);
|
|
|
|
|
|
|
|
sb_start_pagefault(inode->i_sb);
|
|
|
|
file_update_time(vma->vm_file);
|
|
|
|
down_read(&EXT4_I(inode)->i_mmap_sem);
|
2015-12-07 21:10:44 +01:00
|
|
|
err = __dax_mkwrite(vma, vmf, ext4_dax_mmap_get_block, NULL);
|
2015-12-07 20:28:03 +01:00
|
|
|
up_read(&EXT4_I(inode)->i_mmap_sem);
|
|
|
|
sb_end_pagefault(inode->i_sb);
|
|
|
|
|
|
|
|
return err;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Handle write fault for VM_MIXEDMAP mappings. Similarly to ext4_dax_mkwrite()
|
|
|
|
* handler we check for races agaist truncate. Note that since we cycle through
|
|
|
|
* i_mmap_sem, we are sure that also any hole punching that began before we
|
|
|
|
* were called is finished by now and so if it included part of the file we
|
|
|
|
* are working on, our pte will get unmapped and the check for pte_same() in
|
|
|
|
* wp_pfn_shared() fails. Thus fault gets retried and things work out as
|
|
|
|
* desired.
|
|
|
|
*/
|
|
|
|
static int ext4_dax_pfn_mkwrite(struct vm_area_struct *vma,
|
|
|
|
struct vm_fault *vmf)
|
|
|
|
{
|
|
|
|
struct inode *inode = file_inode(vma->vm_file);
|
|
|
|
struct super_block *sb = inode->i_sb;
|
|
|
|
loff_t size;
|
2016-01-23 00:10:53 +01:00
|
|
|
int ret;
|
2015-12-07 20:28:03 +01:00
|
|
|
|
|
|
|
sb_start_pagefault(sb);
|
|
|
|
file_update_time(vma->vm_file);
|
|
|
|
down_read(&EXT4_I(inode)->i_mmap_sem);
|
|
|
|
size = (i_size_read(inode) + PAGE_SIZE - 1) >> PAGE_SHIFT;
|
|
|
|
if (vmf->pgoff >= size)
|
|
|
|
ret = VM_FAULT_SIGBUS;
|
2016-01-23 00:10:53 +01:00
|
|
|
else
|
|
|
|
ret = dax_pfn_mkwrite(vma, vmf);
|
2015-12-07 20:28:03 +01:00
|
|
|
up_read(&EXT4_I(inode)->i_mmap_sem);
|
|
|
|
sb_end_pagefault(sb);
|
|
|
|
|
|
|
|
return ret;
|
2015-02-17 00:59:38 +01:00
|
|
|
}
|
|
|
|
|
|
|
|
static const struct vm_operations_struct ext4_dax_vm_ops = {
|
|
|
|
.fault = ext4_dax_fault,
|
2015-09-08 23:59:03 +02:00
|
|
|
.pmd_fault = ext4_dax_pmd_fault,
|
2015-02-17 00:59:38 +01:00
|
|
|
.page_mkwrite = ext4_dax_mkwrite,
|
2015-12-07 20:28:03 +01:00
|
|
|
.pfn_mkwrite = ext4_dax_pfn_mkwrite,
|
2015-02-17 00:59:38 +01:00
|
|
|
};
|
|
|
|
#else
|
|
|
|
#define ext4_dax_vm_ops ext4_file_vm_ops
|
|
|
|
#endif
|
|
|
|
|
2009-09-27 20:29:37 +02:00
|
|
|
static const struct vm_operations_struct ext4_file_vm_ops = {
|
2015-12-07 20:28:03 +01:00
|
|
|
.fault = ext4_filemap_fault,
|
2014-04-08 00:37:19 +02:00
|
|
|
.map_pages = filemap_map_pages,
|
2008-07-12 01:27:31 +02:00
|
|
|
.page_mkwrite = ext4_page_mkwrite,
|
|
|
|
};
|
|
|
|
|
|
|
|
static int ext4_file_mmap(struct file *file, struct vm_area_struct *vma)
|
|
|
|
{
|
2015-04-12 06:56:10 +02:00
|
|
|
struct inode *inode = file->f_mapping->host;
|
|
|
|
|
|
|
|
if (ext4_encrypted_inode(inode)) {
|
ext4 crypto: reorganize how we store keys in the inode
This is a pretty massive patch which does a number of different things:
1) The per-inode encryption information is now stored in an allocated
data structure, ext4_crypt_info, instead of directly in the node.
This reduces the size usage of an in-memory inode when it is not
using encryption.
2) We drop the ext4_fname_crypto_ctx entirely, and use the per-inode
encryption structure instead. This remove an unnecessary memory
allocation and free for the fname_crypto_ctx as well as allowing us
to reuse the ctfm in a directory for multiple lookups and file
creations.
3) We also cache the inode's policy information in the ext4_crypt_info
structure so we don't have to continually read it out of the
extended attributes.
4) We now keep the keyring key in the inode's encryption structure
instead of releasing it after we are done using it to derive the
per-inode key. This allows us to test to see if the key has been
revoked; if it has, we prevent the use of the derived key and free
it.
5) When an inode is released (or when the derived key is freed), we
will use memset_explicit() to zero out the derived key, so it's not
left hanging around in memory. This implies that when a user logs
out, it is important to first revoke the key, and then unlink it,
and then finally, to use "echo 3 > /proc/sys/vm/drop_caches" to
release any decrypted pages and dcache entries from the system
caches.
6) All this, and we also shrink the number of lines of code by around
100. :-)
Signed-off-by: Theodore Ts'o <tytso@mit.edu>
2015-05-18 19:17:47 +02:00
|
|
|
int err = ext4_get_encryption_info(inode);
|
2015-04-12 06:56:10 +02:00
|
|
|
if (err)
|
|
|
|
return 0;
|
2015-05-31 19:35:39 +02:00
|
|
|
if (ext4_encryption_info(inode) == NULL)
|
|
|
|
return -ENOKEY;
|
2015-04-12 06:56:10 +02:00
|
|
|
}
|
2008-07-12 01:27:31 +02:00
|
|
|
file_accessed(file);
|
2015-02-17 00:59:38 +01:00
|
|
|
if (IS_DAX(file_inode(file))) {
|
|
|
|
vma->vm_ops = &ext4_dax_vm_ops;
|
2015-09-08 23:59:03 +02:00
|
|
|
vma->vm_flags |= VM_MIXEDMAP | VM_HUGEPAGE;
|
2015-02-17 00:59:38 +01:00
|
|
|
} else {
|
|
|
|
vma->vm_ops = &ext4_file_vm_ops;
|
|
|
|
}
|
2008-07-12 01:27:31 +02:00
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2009-06-13 16:09:48 +02:00
|
|
|
static int ext4_file_open(struct inode * inode, struct file * filp)
|
|
|
|
{
|
|
|
|
struct super_block *sb = inode->i_sb;
|
|
|
|
struct ext4_sb_info *sbi = EXT4_SB(inode->i_sb);
|
|
|
|
struct vfsmount *mnt = filp->f_path.mnt;
|
2016-02-08 06:54:26 +01:00
|
|
|
struct inode *dir = filp->f_path.dentry->d_parent->d_inode;
|
2009-06-13 16:09:48 +02:00
|
|
|
struct path path;
|
|
|
|
char buf[64], *cp;
|
2015-04-12 06:56:10 +02:00
|
|
|
int ret;
|
2009-06-13 16:09:48 +02:00
|
|
|
|
|
|
|
if (unlikely(!(sbi->s_mount_flags & EXT4_MF_MNTDIR_SAMPLED) &&
|
|
|
|
!(sb->s_flags & MS_RDONLY))) {
|
|
|
|
sbi->s_mount_flags |= EXT4_MF_MNTDIR_SAMPLED;
|
|
|
|
/*
|
|
|
|
* Sample where the filesystem has been mounted and
|
|
|
|
* store it in the superblock for sysadmin convenience
|
|
|
|
* when trying to sort through large numbers of block
|
|
|
|
* devices or filesystem images.
|
|
|
|
*/
|
|
|
|
memset(buf, 0, sizeof(buf));
|
2010-01-24 02:10:29 +01:00
|
|
|
path.mnt = mnt;
|
|
|
|
path.dentry = mnt->mnt_root;
|
2009-06-13 16:09:48 +02:00
|
|
|
cp = d_path(&path, buf, sizeof(buf));
|
|
|
|
if (!IS_ERR(cp)) {
|
2012-07-23 02:31:31 +02:00
|
|
|
handle_t *handle;
|
|
|
|
int err;
|
|
|
|
|
2013-02-09 03:59:22 +01:00
|
|
|
handle = ext4_journal_start_sb(sb, EXT4_HT_MISC, 1);
|
2012-07-23 02:31:31 +02:00
|
|
|
if (IS_ERR(handle))
|
|
|
|
return PTR_ERR(handle);
|
2014-05-13 04:06:43 +02:00
|
|
|
BUFFER_TRACE(sbi->s_sbh, "get_write_access");
|
2012-07-23 02:31:31 +02:00
|
|
|
err = ext4_journal_get_write_access(handle, sbi->s_sbh);
|
|
|
|
if (err) {
|
|
|
|
ext4_journal_stop(handle);
|
|
|
|
return err;
|
|
|
|
}
|
2011-10-25 15:18:41 +02:00
|
|
|
strlcpy(sbi->s_es->s_last_mounted, cp,
|
|
|
|
sizeof(sbi->s_es->s_last_mounted));
|
2012-07-23 02:31:31 +02:00
|
|
|
ext4_handle_dirty_super(handle, sb);
|
|
|
|
ext4_journal_stop(handle);
|
2009-06-13 16:09:48 +02:00
|
|
|
}
|
|
|
|
}
|
2015-05-31 19:35:39 +02:00
|
|
|
if (ext4_encrypted_inode(inode)) {
|
|
|
|
ret = ext4_get_encryption_info(inode);
|
|
|
|
if (ret)
|
|
|
|
return -EACCES;
|
|
|
|
if (ext4_encryption_info(inode) == NULL)
|
|
|
|
return -ENOKEY;
|
|
|
|
}
|
2016-02-08 06:54:26 +01:00
|
|
|
if (ext4_encrypted_inode(dir) &&
|
|
|
|
!ext4_is_child_context_consistent_with_parent(dir, inode)) {
|
|
|
|
ext4_warning(inode->i_sb,
|
|
|
|
"Inconsistent encryption contexts: %lu/%lu\n",
|
|
|
|
(unsigned long) dir->i_ino,
|
|
|
|
(unsigned long) inode->i_ino);
|
|
|
|
return -EPERM;
|
|
|
|
}
|
2011-01-10 18:29:43 +01:00
|
|
|
/*
|
|
|
|
* Set up the jbd2_inode if we are opening the inode for
|
|
|
|
* writing and the journal is present
|
|
|
|
*/
|
2013-08-17 03:19:41 +02:00
|
|
|
if (filp->f_mode & FMODE_WRITE) {
|
2015-04-12 06:56:10 +02:00
|
|
|
ret = ext4_inode_attach_jinode(inode);
|
2013-08-17 03:19:41 +02:00
|
|
|
if (ret < 0)
|
|
|
|
return ret;
|
2011-01-10 18:29:43 +01:00
|
|
|
}
|
2015-05-31 19:35:39 +02:00
|
|
|
return dquot_file_open(inode, filp);
|
2009-06-13 16:09:48 +02:00
|
|
|
}
|
|
|
|
|
2012-11-09 03:57:40 +01:00
|
|
|
/*
|
|
|
|
* Here we use ext4_map_blocks() to get a block mapping for a extent-based
|
|
|
|
* file rather than ext4_ext_walk_space() because we can introduce
|
|
|
|
* SEEK_DATA/SEEK_HOLE for block-mapped and extent-mapped file at the same
|
|
|
|
* function. When extent status tree has been fully implemented, it will
|
|
|
|
* track all extent status for a file and we can directly use it to
|
|
|
|
* retrieve the offset for SEEK_DATA/SEEK_HOLE.
|
|
|
|
*/
|
|
|
|
|
|
|
|
/*
|
|
|
|
* When we retrieve the offset for SEEK_DATA/SEEK_HOLE, we would need to
|
|
|
|
* lookup page cache to check whether or not there has some data between
|
|
|
|
* [startoff, endoff] because, if this range contains an unwritten extent,
|
|
|
|
* we determine this extent as a data or a hole according to whether the
|
|
|
|
* page cache has data or not.
|
|
|
|
*/
|
2015-01-02 21:16:00 +01:00
|
|
|
static int ext4_find_unwritten_pgoff(struct inode *inode,
|
|
|
|
int whence,
|
|
|
|
struct ext4_map_blocks *map,
|
|
|
|
loff_t *offset)
|
2012-11-09 03:57:40 +01:00
|
|
|
{
|
|
|
|
struct pagevec pvec;
|
2015-01-02 21:16:00 +01:00
|
|
|
unsigned int blkbits;
|
2012-11-09 03:57:40 +01:00
|
|
|
pgoff_t index;
|
|
|
|
pgoff_t end;
|
2015-01-02 21:16:00 +01:00
|
|
|
loff_t endoff;
|
2012-11-09 03:57:40 +01:00
|
|
|
loff_t startoff;
|
|
|
|
loff_t lastoff;
|
|
|
|
int found = 0;
|
|
|
|
|
2015-01-02 21:16:00 +01:00
|
|
|
blkbits = inode->i_sb->s_blocksize_bits;
|
2012-11-09 03:57:40 +01:00
|
|
|
startoff = *offset;
|
|
|
|
lastoff = startoff;
|
2015-01-02 21:16:00 +01:00
|
|
|
endoff = (loff_t)(map->m_lblk + map->m_len) << blkbits;
|
2012-11-09 03:57:40 +01:00
|
|
|
|
|
|
|
index = startoff >> PAGE_CACHE_SHIFT;
|
|
|
|
end = endoff >> PAGE_CACHE_SHIFT;
|
|
|
|
|
|
|
|
pagevec_init(&pvec, 0);
|
|
|
|
do {
|
|
|
|
int i, num;
|
|
|
|
unsigned long nr_pages;
|
|
|
|
|
|
|
|
num = min_t(pgoff_t, end - index, PAGEVEC_SIZE);
|
|
|
|
nr_pages = pagevec_lookup(&pvec, inode->i_mapping, index,
|
|
|
|
(pgoff_t)num);
|
|
|
|
if (nr_pages == 0) {
|
2012-12-18 00:59:39 +01:00
|
|
|
if (whence == SEEK_DATA)
|
2012-11-09 03:57:40 +01:00
|
|
|
break;
|
|
|
|
|
2012-12-18 00:59:39 +01:00
|
|
|
BUG_ON(whence != SEEK_HOLE);
|
2012-11-09 03:57:40 +01:00
|
|
|
/*
|
|
|
|
* If this is the first time to go into the loop and
|
|
|
|
* offset is not beyond the end offset, it will be a
|
|
|
|
* hole at this offset
|
|
|
|
*/
|
|
|
|
if (lastoff == startoff || lastoff < endoff)
|
|
|
|
found = 1;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* If this is the first time to go into the loop and
|
|
|
|
* offset is smaller than the first page offset, it will be a
|
|
|
|
* hole at this offset.
|
|
|
|
*/
|
2012-12-18 00:59:39 +01:00
|
|
|
if (lastoff == startoff && whence == SEEK_HOLE &&
|
2012-11-09 03:57:40 +01:00
|
|
|
lastoff < page_offset(pvec.pages[0])) {
|
|
|
|
found = 1;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
for (i = 0; i < nr_pages; i++) {
|
|
|
|
struct page *page = pvec.pages[i];
|
|
|
|
struct buffer_head *bh, *head;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* If the current offset is not beyond the end of given
|
|
|
|
* range, it will be a hole.
|
|
|
|
*/
|
2012-12-18 00:59:39 +01:00
|
|
|
if (lastoff < endoff && whence == SEEK_HOLE &&
|
2012-11-09 03:57:40 +01:00
|
|
|
page->index > end) {
|
|
|
|
found = 1;
|
|
|
|
*offset = lastoff;
|
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
|
|
|
|
lock_page(page);
|
|
|
|
|
|
|
|
if (unlikely(page->mapping != inode->i_mapping)) {
|
|
|
|
unlock_page(page);
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (!page_has_buffers(page)) {
|
|
|
|
unlock_page(page);
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (page_has_buffers(page)) {
|
|
|
|
lastoff = page_offset(page);
|
|
|
|
bh = head = page_buffers(page);
|
|
|
|
do {
|
|
|
|
if (buffer_uptodate(bh) ||
|
|
|
|
buffer_unwritten(bh)) {
|
2012-12-18 00:59:39 +01:00
|
|
|
if (whence == SEEK_DATA)
|
2012-11-09 03:57:40 +01:00
|
|
|
found = 1;
|
|
|
|
} else {
|
2012-12-18 00:59:39 +01:00
|
|
|
if (whence == SEEK_HOLE)
|
2012-11-09 03:57:40 +01:00
|
|
|
found = 1;
|
|
|
|
}
|
|
|
|
if (found) {
|
|
|
|
*offset = max_t(loff_t,
|
|
|
|
startoff, lastoff);
|
|
|
|
unlock_page(page);
|
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
lastoff += bh->b_size;
|
|
|
|
bh = bh->b_this_page;
|
|
|
|
} while (bh != head);
|
|
|
|
}
|
|
|
|
|
|
|
|
lastoff = page_offset(page) + PAGE_SIZE;
|
|
|
|
unlock_page(page);
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* The no. of pages is less than our desired, that would be a
|
|
|
|
* hole in there.
|
|
|
|
*/
|
2012-12-18 00:59:39 +01:00
|
|
|
if (nr_pages < num && whence == SEEK_HOLE) {
|
2012-11-09 03:57:40 +01:00
|
|
|
found = 1;
|
|
|
|
*offset = lastoff;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
index = pvec.pages[i - 1]->index + 1;
|
|
|
|
pagevec_release(&pvec);
|
|
|
|
} while (index <= end);
|
|
|
|
|
|
|
|
out:
|
|
|
|
pagevec_release(&pvec);
|
|
|
|
return found;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* ext4_seek_data() retrieves the offset for SEEK_DATA.
|
|
|
|
*/
|
|
|
|
static loff_t ext4_seek_data(struct file *file, loff_t offset, loff_t maxsize)
|
|
|
|
{
|
|
|
|
struct inode *inode = file->f_mapping->host;
|
2015-01-02 21:16:00 +01:00
|
|
|
struct ext4_map_blocks map;
|
|
|
|
struct extent_status es;
|
|
|
|
ext4_lblk_t start, last, end;
|
|
|
|
loff_t dataoff, isize;
|
|
|
|
int blkbits;
|
|
|
|
int ret = 0;
|
2012-11-09 03:57:40 +01:00
|
|
|
|
2016-01-22 21:40:57 +01:00
|
|
|
inode_lock(inode);
|
2015-01-02 21:16:00 +01:00
|
|
|
|
|
|
|
isize = i_size_read(inode);
|
|
|
|
if (offset >= isize) {
|
2016-01-22 21:40:57 +01:00
|
|
|
inode_unlock(inode);
|
2012-11-09 03:57:40 +01:00
|
|
|
return -ENXIO;
|
|
|
|
}
|
2015-01-02 21:16:00 +01:00
|
|
|
|
|
|
|
blkbits = inode->i_sb->s_blocksize_bits;
|
|
|
|
start = offset >> blkbits;
|
|
|
|
last = start;
|
|
|
|
end = isize >> blkbits;
|
|
|
|
dataoff = offset;
|
|
|
|
|
|
|
|
do {
|
|
|
|
map.m_lblk = last;
|
|
|
|
map.m_len = end - last + 1;
|
|
|
|
ret = ext4_map_blocks(NULL, inode, &map, 0);
|
|
|
|
if (ret > 0 && !(map.m_flags & EXT4_MAP_UNWRITTEN)) {
|
|
|
|
if (last != start)
|
|
|
|
dataoff = (loff_t)last << blkbits;
|
2012-11-09 03:57:40 +01:00
|
|
|
break;
|
2015-01-02 21:16:00 +01:00
|
|
|
}
|
2012-11-09 03:57:40 +01:00
|
|
|
|
2015-01-02 21:16:00 +01:00
|
|
|
/*
|
|
|
|
* If there is a delay extent at this offset,
|
|
|
|
* it will be as a data.
|
|
|
|
*/
|
|
|
|
ext4_es_find_delayed_extent_range(inode, last, last, &es);
|
|
|
|
if (es.es_len != 0 && in_range(last, es.es_lblk, es.es_len)) {
|
|
|
|
if (last != start)
|
|
|
|
dataoff = (loff_t)last << blkbits;
|
2012-11-09 03:57:40 +01:00
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
2015-01-02 21:16:00 +01:00
|
|
|
/*
|
|
|
|
* If there is a unwritten extent at this offset,
|
|
|
|
* it will be as a data or a hole according to page
|
|
|
|
* cache that has data or not.
|
|
|
|
*/
|
|
|
|
if (map.m_flags & EXT4_MAP_UNWRITTEN) {
|
|
|
|
int unwritten;
|
|
|
|
unwritten = ext4_find_unwritten_pgoff(inode, SEEK_DATA,
|
|
|
|
&map, &dataoff);
|
|
|
|
if (unwritten)
|
|
|
|
break;
|
|
|
|
}
|
2012-11-09 03:57:40 +01:00
|
|
|
|
2015-01-02 21:16:00 +01:00
|
|
|
last++;
|
|
|
|
dataoff = (loff_t)last << blkbits;
|
|
|
|
} while (last <= end);
|
2012-11-09 03:57:40 +01:00
|
|
|
|
2016-01-22 21:40:57 +01:00
|
|
|
inode_unlock(inode);
|
2012-11-09 03:57:40 +01:00
|
|
|
|
2015-01-02 21:16:00 +01:00
|
|
|
if (dataoff > isize)
|
|
|
|
return -ENXIO;
|
|
|
|
|
|
|
|
return vfs_setpos(file, dataoff, maxsize);
|
2012-11-09 03:57:40 +01:00
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
2015-01-02 21:16:00 +01:00
|
|
|
* ext4_seek_hole() retrieves the offset for SEEK_HOLE.
|
2012-11-09 03:57:40 +01:00
|
|
|
*/
|
|
|
|
static loff_t ext4_seek_hole(struct file *file, loff_t offset, loff_t maxsize)
|
|
|
|
{
|
|
|
|
struct inode *inode = file->f_mapping->host;
|
2015-01-02 21:16:00 +01:00
|
|
|
struct ext4_map_blocks map;
|
|
|
|
struct extent_status es;
|
|
|
|
ext4_lblk_t start, last, end;
|
|
|
|
loff_t holeoff, isize;
|
|
|
|
int blkbits;
|
|
|
|
int ret = 0;
|
2012-11-09 03:57:40 +01:00
|
|
|
|
2016-01-22 21:40:57 +01:00
|
|
|
inode_lock(inode);
|
2015-01-02 21:16:00 +01:00
|
|
|
|
|
|
|
isize = i_size_read(inode);
|
|
|
|
if (offset >= isize) {
|
2016-01-22 21:40:57 +01:00
|
|
|
inode_unlock(inode);
|
2012-11-09 03:57:40 +01:00
|
|
|
return -ENXIO;
|
|
|
|
}
|
|
|
|
|
2015-01-02 21:16:00 +01:00
|
|
|
blkbits = inode->i_sb->s_blocksize_bits;
|
|
|
|
start = offset >> blkbits;
|
|
|
|
last = start;
|
|
|
|
end = isize >> blkbits;
|
|
|
|
holeoff = offset;
|
2012-11-09 03:57:40 +01:00
|
|
|
|
2015-01-02 21:16:00 +01:00
|
|
|
do {
|
|
|
|
map.m_lblk = last;
|
|
|
|
map.m_len = end - last + 1;
|
|
|
|
ret = ext4_map_blocks(NULL, inode, &map, 0);
|
|
|
|
if (ret > 0 && !(map.m_flags & EXT4_MAP_UNWRITTEN)) {
|
|
|
|
last += ret;
|
|
|
|
holeoff = (loff_t)last << blkbits;
|
|
|
|
continue;
|
|
|
|
}
|
2012-11-09 03:57:40 +01:00
|
|
|
|
2015-01-02 21:16:00 +01:00
|
|
|
/*
|
|
|
|
* If there is a delay extent at this offset,
|
|
|
|
* we will skip this extent.
|
|
|
|
*/
|
|
|
|
ext4_es_find_delayed_extent_range(inode, last, last, &es);
|
|
|
|
if (es.es_len != 0 && in_range(last, es.es_lblk, es.es_len)) {
|
|
|
|
last = es.es_lblk + es.es_len;
|
|
|
|
holeoff = (loff_t)last << blkbits;
|
|
|
|
continue;
|
|
|
|
}
|
2014-12-03 00:08:53 +01:00
|
|
|
|
2015-01-02 21:16:00 +01:00
|
|
|
/*
|
|
|
|
* If there is a unwritten extent at this offset,
|
|
|
|
* it will be as a data or a hole according to page
|
|
|
|
* cache that has data or not.
|
|
|
|
*/
|
|
|
|
if (map.m_flags & EXT4_MAP_UNWRITTEN) {
|
|
|
|
int unwritten;
|
|
|
|
unwritten = ext4_find_unwritten_pgoff(inode, SEEK_HOLE,
|
|
|
|
&map, &holeoff);
|
|
|
|
if (!unwritten) {
|
|
|
|
last += ret;
|
|
|
|
holeoff = (loff_t)last << blkbits;
|
2012-11-09 03:57:40 +01:00
|
|
|
continue;
|
|
|
|
}
|
2014-12-03 00:08:53 +01:00
|
|
|
}
|
2015-01-02 21:16:00 +01:00
|
|
|
|
|
|
|
/* find a hole */
|
|
|
|
break;
|
|
|
|
} while (last <= end);
|
|
|
|
|
2016-01-22 21:40:57 +01:00
|
|
|
inode_unlock(inode);
|
2012-11-09 03:57:40 +01:00
|
|
|
|
2015-01-02 21:16:00 +01:00
|
|
|
if (holeoff > isize)
|
|
|
|
holeoff = isize;
|
|
|
|
|
|
|
|
return vfs_setpos(file, holeoff, maxsize);
|
2012-11-09 03:57:40 +01:00
|
|
|
}
|
|
|
|
|
2010-10-28 03:30:06 +02:00
|
|
|
/*
|
2012-04-30 20:14:03 +02:00
|
|
|
* ext4_llseek() handles both block-mapped and extent-mapped maxbytes values
|
|
|
|
* by calling generic_file_llseek_size() with the appropriate maxbytes
|
|
|
|
* value for each.
|
2010-10-28 03:30:06 +02:00
|
|
|
*/
|
2012-12-18 00:59:39 +01:00
|
|
|
loff_t ext4_llseek(struct file *file, loff_t offset, int whence)
|
2010-10-28 03:30:06 +02:00
|
|
|
{
|
|
|
|
struct inode *inode = file->f_mapping->host;
|
|
|
|
loff_t maxbytes;
|
|
|
|
|
|
|
|
if (!(ext4_test_inode_flag(inode, EXT4_INODE_EXTENTS)))
|
|
|
|
maxbytes = EXT4_SB(inode->i_sb)->s_bitmap_maxbytes;
|
|
|
|
else
|
|
|
|
maxbytes = inode->i_sb->s_maxbytes;
|
|
|
|
|
2012-12-18 00:59:39 +01:00
|
|
|
switch (whence) {
|
2012-11-09 03:57:40 +01:00
|
|
|
case SEEK_SET:
|
|
|
|
case SEEK_CUR:
|
|
|
|
case SEEK_END:
|
2012-12-18 00:59:39 +01:00
|
|
|
return generic_file_llseek_size(file, offset, whence,
|
2012-11-09 03:57:40 +01:00
|
|
|
maxbytes, i_size_read(inode));
|
|
|
|
case SEEK_DATA:
|
|
|
|
return ext4_seek_data(file, offset, maxbytes);
|
|
|
|
case SEEK_HOLE:
|
|
|
|
return ext4_seek_hole(file, offset, maxbytes);
|
|
|
|
}
|
|
|
|
|
|
|
|
return -EINVAL;
|
2010-10-28 03:30:06 +02:00
|
|
|
}
|
|
|
|
|
2006-10-11 10:20:53 +02:00
|
|
|
const struct file_operations ext4_file_operations = {
|
2010-10-28 03:30:06 +02:00
|
|
|
.llseek = ext4_llseek,
|
2014-04-02 20:33:16 +02:00
|
|
|
.read_iter = generic_file_read_iter,
|
2014-04-17 22:09:22 +02:00
|
|
|
.write_iter = ext4_file_write_iter,
|
2008-04-30 04:03:54 +02:00
|
|
|
.unlocked_ioctl = ext4_ioctl,
|
2006-10-11 10:20:50 +02:00
|
|
|
#ifdef CONFIG_COMPAT
|
2006-10-11 10:20:53 +02:00
|
|
|
.compat_ioctl = ext4_compat_ioctl,
|
2006-10-11 10:20:50 +02:00
|
|
|
#endif
|
2008-07-12 01:27:31 +02:00
|
|
|
.mmap = ext4_file_mmap,
|
2009-06-13 16:09:48 +02:00
|
|
|
.open = ext4_file_open,
|
2006-10-11 10:20:53 +02:00
|
|
|
.release = ext4_release_file,
|
|
|
|
.fsync = ext4_sync_file,
|
2006-10-11 10:20:50 +02:00
|
|
|
.splice_read = generic_file_splice_read,
|
2014-04-05 10:27:08 +02:00
|
|
|
.splice_write = iter_file_splice_write,
|
2011-01-14 13:07:43 +01:00
|
|
|
.fallocate = ext4_fallocate,
|
2006-10-11 10:20:50 +02:00
|
|
|
};
|
|
|
|
|
2007-02-12 09:55:38 +01:00
|
|
|
const struct inode_operations ext4_file_inode_operations = {
|
2006-10-11 10:20:53 +02:00
|
|
|
.setattr = ext4_setattr,
|
2008-07-12 01:27:31 +02:00
|
|
|
.getattr = ext4_getattr,
|
2006-10-11 10:20:50 +02:00
|
|
|
.setxattr = generic_setxattr,
|
|
|
|
.getxattr = generic_getxattr,
|
2006-10-11 10:20:53 +02:00
|
|
|
.listxattr = ext4_listxattr,
|
2006-10-11 10:20:50 +02:00
|
|
|
.removexattr = generic_removexattr,
|
2011-07-23 17:37:31 +02:00
|
|
|
.get_acl = ext4_get_acl,
|
2013-12-20 14:16:44 +01:00
|
|
|
.set_acl = ext4_set_acl,
|
2008-10-07 06:46:36 +02:00
|
|
|
.fiemap = ext4_fiemap,
|
2006-10-11 10:20:50 +02:00
|
|
|
};
|
|
|
|
|