Skip to content
Projects
Groups
Snippets
Help
Loading...
Help
Support
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
L
linux
Project overview
Project overview
Details
Activity
Releases
Repository
Repository
Files
Commits
Branches
Tags
Contributors
Graph
Compare
Issues
0
Issues
0
List
Boards
Labels
Milestones
Merge Requests
0
Merge Requests
0
Analytics
Analytics
Repository
Value Stream
Wiki
Wiki
Snippets
Snippets
Members
Members
Collapse sidebar
Close sidebar
Activity
Graph
Create a new issue
Commits
Issue Boards
Open sidebar
Kirill Smelkov
linux
Commits
b47ec80b
Commit
b47ec80b
authored
Jul 20, 2016
by
Dave Chinner
Browse files
Options
Browse Files
Download
Plain Diff
Merge branch 'xfs-4.8-split-dax-dio' into for-next
parents
bbfeb614
16d4d435
Changes
8
Show whitespace changes
Inline
Side-by-side
Showing
8 changed files
with
203 additions
and
118 deletions
+203
-118
fs/xfs/xfs_aops.c
fs/xfs/xfs_aops.c
+5
-19
fs/xfs/xfs_aops.h
fs/xfs/xfs_aops.h
+3
-0
fs/xfs/xfs_file.c
fs/xfs/xfs_file.c
+176
-56
fs/xfs/xfs_inode.h
fs/xfs/xfs_inode.h
+0
-10
fs/xfs/xfs_ioctl.c
fs/xfs/xfs_ioctl.c
+8
-14
fs/xfs/xfs_ioctl.h
fs/xfs/xfs_ioctl.h
+0
-3
fs/xfs/xfs_ioctl32.c
fs/xfs/xfs_ioctl32.c
+1
-5
fs/xfs/xfs_trace.h
fs/xfs/xfs_trace.h
+10
-11
No files found.
fs/xfs/xfs_aops.c
View file @
b47ec80b
...
@@ -1303,7 +1303,7 @@ xfs_get_blocks_dax_fault(
...
@@ -1303,7 +1303,7 @@ xfs_get_blocks_dax_fault(
* whereas if we have flags set we will always be called in task context
* whereas if we have flags set we will always be called in task context
* (i.e. from a workqueue).
* (i.e. from a workqueue).
*/
*/
STATIC
int
int
xfs_end_io_direct_write
(
xfs_end_io_direct_write
(
struct
kiocb
*
iocb
,
struct
kiocb
*
iocb
,
loff_t
offset
,
loff_t
offset
,
...
@@ -1374,24 +1374,10 @@ xfs_vm_direct_IO(
...
@@ -1374,24 +1374,10 @@ xfs_vm_direct_IO(
struct
kiocb
*
iocb
,
struct
kiocb
*
iocb
,
struct
iov_iter
*
iter
)
struct
iov_iter
*
iter
)
{
{
struct
inode
*
inode
=
iocb
->
ki_filp
->
f_mapping
->
host
;
/*
dio_iodone_t
*
endio
=
NULL
;
* We just need the method present so that open/fcntl allow direct I/O.
int
flags
=
0
;
*/
struct
block_device
*
bdev
;
return
-
EINVAL
;
if
(
iov_iter_rw
(
iter
)
==
WRITE
)
{
endio
=
xfs_end_io_direct_write
;
flags
=
DIO_ASYNC_EXTEND
;
}
if
(
IS_DAX
(
inode
))
{
return
dax_do_io
(
iocb
,
inode
,
iter
,
xfs_get_blocks_direct
,
endio
,
0
);
}
bdev
=
xfs_find_bdev_for_inode
(
inode
);
return
__blockdev_direct_IO
(
iocb
,
inode
,
bdev
,
iter
,
xfs_get_blocks_direct
,
endio
,
NULL
,
flags
);
}
}
STATIC
sector_t
STATIC
sector_t
...
...
fs/xfs/xfs_aops.h
View file @
b47ec80b
...
@@ -60,6 +60,9 @@ int xfs_get_blocks_direct(struct inode *inode, sector_t offset,
...
@@ -60,6 +60,9 @@ int xfs_get_blocks_direct(struct inode *inode, sector_t offset,
int
xfs_get_blocks_dax_fault
(
struct
inode
*
inode
,
sector_t
offset
,
int
xfs_get_blocks_dax_fault
(
struct
inode
*
inode
,
sector_t
offset
,
struct
buffer_head
*
map_bh
,
int
create
);
struct
buffer_head
*
map_bh
,
int
create
);
int
xfs_end_io_direct_write
(
struct
kiocb
*
iocb
,
loff_t
offset
,
ssize_t
size
,
void
*
private
);
extern
void
xfs_count_page_state
(
struct
page
*
,
int
*
,
int
*
);
extern
void
xfs_count_page_state
(
struct
page
*
,
int
*
,
int
*
);
extern
struct
block_device
*
xfs_find_bdev_for_inode
(
struct
inode
*
);
extern
struct
block_device
*
xfs_find_bdev_for_inode
(
struct
inode
*
);
...
...
fs/xfs/xfs_file.c
View file @
b47ec80b
...
@@ -239,48 +239,35 @@ xfs_file_fsync(
...
@@ -239,48 +239,35 @@ xfs_file_fsync(
}
}
STATIC
ssize_t
STATIC
ssize_t
xfs_file_
read_iter
(
xfs_file_
dio_aio_read
(
struct
kiocb
*
iocb
,
struct
kiocb
*
iocb
,
struct
iov_iter
*
to
)
struct
iov_iter
*
to
)
{
{
struct
file
*
file
=
iocb
->
ki_filp
;
struct
address_space
*
mapping
=
iocb
->
ki_filp
->
f_mapping
;
struct
inode
*
inode
=
file
->
f_
mapping
->
host
;
struct
inode
*
inode
=
mapping
->
host
;
struct
xfs_inode
*
ip
=
XFS_I
(
inode
);
struct
xfs_inode
*
ip
=
XFS_I
(
inode
);
struct
xfs_mount
*
mp
=
ip
->
i_mount
;
loff_t
isize
=
i_size_read
(
inode
);
size_t
size
=
iov_iter_count
(
to
);
size_t
count
=
iov_iter_count
(
to
);
struct
iov_iter
data
;
struct
xfs_buftarg
*
target
;
ssize_t
ret
=
0
;
ssize_t
ret
=
0
;
int
ioflags
=
0
;
xfs_fsize_t
n
;
loff_t
pos
=
iocb
->
ki_pos
;
XFS_STATS_INC
(
mp
,
xs_read_call
s
);
trace_xfs_file_direct_read
(
ip
,
count
,
iocb
->
ki_po
s
);
if
(
unlikely
(
iocb
->
ki_flags
&
IOCB_DIRECT
))
if
(
!
count
)
ioflags
|=
XFS_IO_ISDIRECT
;
return
0
;
/* skip atime */
if
(
file
->
f_mode
&
FMODE_NOCMTIME
)
ioflags
|=
XFS_IO_INVIS
;
if
(
XFS_IS_REALTIME_INODE
(
ip
))
target
=
ip
->
i_mount
->
m_rtdev_targp
;
else
target
=
ip
->
i_mount
->
m_ddev_targp
;
if
((
ioflags
&
XFS_IO_ISDIRECT
)
&&
!
IS_DAX
(
inode
))
{
xfs_buftarg_t
*
target
=
XFS_IS_REALTIME_INODE
(
ip
)
?
mp
->
m_rtdev_targp
:
mp
->
m_ddev_targp
;
/* DIO must be aligned to device logical sector size */
/* DIO must be aligned to device logical sector size */
if
((
pos
|
size
)
&
target
->
bt_logical_sectormask
)
{
if
((
iocb
->
ki_pos
|
count
)
&
target
->
bt_logical_sectormask
)
{
if
(
pos
==
i_size_read
(
inode
)
)
if
(
iocb
->
ki_pos
==
isize
)
return
0
;
return
0
;
return
-
EINVAL
;
return
-
EINVAL
;
}
}
}
n
=
mp
->
m_super
->
s_maxbytes
-
pos
;
if
(
n
<=
0
||
size
==
0
)
return
0
;
if
(
n
<
size
)
size
=
n
;
if
(
XFS_FORCED_SHUTDOWN
(
mp
))
return
-
EIO
;
/*
/*
* Locking is a bit tricky here. If we take an exclusive lock for direct
* Locking is a bit tricky here. If we take an exclusive lock for direct
...
@@ -293,7 +280,7 @@ xfs_file_read_iter(
...
@@ -293,7 +280,7 @@ xfs_file_read_iter(
* serialisation.
* serialisation.
*/
*/
xfs_rw_ilock
(
ip
,
XFS_IOLOCK_SHARED
);
xfs_rw_ilock
(
ip
,
XFS_IOLOCK_SHARED
);
if
(
(
ioflags
&
XFS_IO_ISDIRECT
)
&&
inode
->
i_
mapping
->
nrpages
)
{
if
(
mapping
->
nrpages
)
{
xfs_rw_iunlock
(
ip
,
XFS_IOLOCK_SHARED
);
xfs_rw_iunlock
(
ip
,
XFS_IOLOCK_SHARED
);
xfs_rw_ilock
(
ip
,
XFS_IOLOCK_EXCL
);
xfs_rw_ilock
(
ip
,
XFS_IOLOCK_EXCL
);
...
@@ -308,8 +295,8 @@ xfs_file_read_iter(
...
@@ -308,8 +295,8 @@ xfs_file_read_iter(
* flush and reduce the chances of repeated iolock cycles going
* flush and reduce the chances of repeated iolock cycles going
* forward.
* forward.
*/
*/
if
(
inode
->
i_
mapping
->
nrpages
)
{
if
(
mapping
->
nrpages
)
{
ret
=
filemap_write_and_wait
(
VFS_I
(
ip
)
->
i_
mapping
);
ret
=
filemap_write_and_wait
(
mapping
);
if
(
ret
)
{
if
(
ret
)
{
xfs_rw_iunlock
(
ip
,
XFS_IOLOCK_EXCL
);
xfs_rw_iunlock
(
ip
,
XFS_IOLOCK_EXCL
);
return
ret
;
return
ret
;
...
@@ -320,20 +307,95 @@ xfs_file_read_iter(
...
@@ -320,20 +307,95 @@ xfs_file_read_iter(
* we fail to invalidate a page, but this should never
* we fail to invalidate a page, but this should never
* happen on XFS. Warn if it does fail.
* happen on XFS. Warn if it does fail.
*/
*/
ret
=
invalidate_inode_pages2
(
VFS_I
(
ip
)
->
i_
mapping
);
ret
=
invalidate_inode_pages2
(
mapping
);
WARN_ON_ONCE
(
ret
);
WARN_ON_ONCE
(
ret
);
ret
=
0
;
ret
=
0
;
}
}
xfs_rw_ilock_demote
(
ip
,
XFS_IOLOCK_EXCL
);
xfs_rw_ilock_demote
(
ip
,
XFS_IOLOCK_EXCL
);
}
}
trace_xfs_file_read
(
ip
,
size
,
pos
,
ioflags
);
data
=
*
to
;
ret
=
__blockdev_direct_IO
(
iocb
,
inode
,
target
->
bt_bdev
,
&
data
,
xfs_get_blocks_direct
,
NULL
,
NULL
,
0
);
if
(
ret
>
0
)
{
iocb
->
ki_pos
+=
ret
;
iov_iter_advance
(
to
,
ret
);
}
xfs_rw_iunlock
(
ip
,
XFS_IOLOCK_SHARED
);
file_accessed
(
iocb
->
ki_filp
);
return
ret
;
}
STATIC
ssize_t
xfs_file_dax_read
(
struct
kiocb
*
iocb
,
struct
iov_iter
*
to
)
{
struct
address_space
*
mapping
=
iocb
->
ki_filp
->
f_mapping
;
struct
inode
*
inode
=
mapping
->
host
;
struct
xfs_inode
*
ip
=
XFS_I
(
inode
);
struct
iov_iter
data
=
*
to
;
size_t
count
=
iov_iter_count
(
to
);
ssize_t
ret
=
0
;
trace_xfs_file_dax_read
(
ip
,
count
,
iocb
->
ki_pos
);
if
(
!
count
)
return
0
;
/* skip atime */
xfs_rw_ilock
(
ip
,
XFS_IOLOCK_SHARED
);
ret
=
dax_do_io
(
iocb
,
inode
,
&
data
,
xfs_get_blocks_direct
,
NULL
,
0
);
if
(
ret
>
0
)
{
iocb
->
ki_pos
+=
ret
;
iov_iter_advance
(
to
,
ret
);
}
xfs_rw_iunlock
(
ip
,
XFS_IOLOCK_SHARED
);
file_accessed
(
iocb
->
ki_filp
);
return
ret
;
}
STATIC
ssize_t
xfs_file_buffered_aio_read
(
struct
kiocb
*
iocb
,
struct
iov_iter
*
to
)
{
struct
xfs_inode
*
ip
=
XFS_I
(
file_inode
(
iocb
->
ki_filp
));
ssize_t
ret
;
trace_xfs_file_buffered_read
(
ip
,
iov_iter_count
(
to
),
iocb
->
ki_pos
);
xfs_rw_ilock
(
ip
,
XFS_IOLOCK_SHARED
);
ret
=
generic_file_read_iter
(
iocb
,
to
);
ret
=
generic_file_read_iter
(
iocb
,
to
);
xfs_rw_iunlock
(
ip
,
XFS_IOLOCK_SHARED
);
return
ret
;
}
STATIC
ssize_t
xfs_file_read_iter
(
struct
kiocb
*
iocb
,
struct
iov_iter
*
to
)
{
struct
inode
*
inode
=
file_inode
(
iocb
->
ki_filp
);
struct
xfs_mount
*
mp
=
XFS_I
(
inode
)
->
i_mount
;
ssize_t
ret
=
0
;
XFS_STATS_INC
(
mp
,
xs_read_calls
);
if
(
XFS_FORCED_SHUTDOWN
(
mp
))
return
-
EIO
;
if
(
IS_DAX
(
inode
))
ret
=
xfs_file_dax_read
(
iocb
,
to
);
else
if
(
iocb
->
ki_flags
&
IOCB_DIRECT
)
ret
=
xfs_file_dio_aio_read
(
iocb
,
to
);
else
ret
=
xfs_file_buffered_aio_read
(
iocb
,
to
);
if
(
ret
>
0
)
if
(
ret
>
0
)
XFS_STATS_ADD
(
mp
,
xs_read_bytes
,
ret
);
XFS_STATS_ADD
(
mp
,
xs_read_bytes
,
ret
);
xfs_rw_iunlock
(
ip
,
XFS_IOLOCK_SHARED
);
return
ret
;
return
ret
;
}
}
...
@@ -346,18 +408,14 @@ xfs_file_splice_read(
...
@@ -346,18 +408,14 @@ xfs_file_splice_read(
unsigned
int
flags
)
unsigned
int
flags
)
{
{
struct
xfs_inode
*
ip
=
XFS_I
(
infilp
->
f_mapping
->
host
);
struct
xfs_inode
*
ip
=
XFS_I
(
infilp
->
f_mapping
->
host
);
int
ioflags
=
0
;
ssize_t
ret
;
ssize_t
ret
;
XFS_STATS_INC
(
ip
->
i_mount
,
xs_read_calls
);
XFS_STATS_INC
(
ip
->
i_mount
,
xs_read_calls
);
if
(
infilp
->
f_mode
&
FMODE_NOCMTIME
)
ioflags
|=
XFS_IO_INVIS
;
if
(
XFS_FORCED_SHUTDOWN
(
ip
->
i_mount
))
if
(
XFS_FORCED_SHUTDOWN
(
ip
->
i_mount
))
return
-
EIO
;
return
-
EIO
;
trace_xfs_file_splice_read
(
ip
,
count
,
*
ppos
,
ioflags
);
trace_xfs_file_splice_read
(
ip
,
count
,
*
ppos
);
/*
/*
* DAX inodes cannot ues the page cache for splice, so we have to push
* DAX inodes cannot ues the page cache for splice, so we have to push
...
@@ -553,8 +611,7 @@ xfs_file_dio_aio_write(
...
@@ -553,8 +611,7 @@ xfs_file_dio_aio_write(
mp
->
m_rtdev_targp
:
mp
->
m_ddev_targp
;
mp
->
m_rtdev_targp
:
mp
->
m_ddev_targp
;
/* DIO must be aligned to device logical sector size */
/* DIO must be aligned to device logical sector size */
if
(
!
IS_DAX
(
inode
)
&&
if
((
iocb
->
ki_pos
|
count
)
&
target
->
bt_logical_sectormask
)
((
iocb
->
ki_pos
|
count
)
&
target
->
bt_logical_sectormask
))
return
-
EINVAL
;
return
-
EINVAL
;
/* "unaligned" here means not aligned to a filesystem block */
/* "unaligned" here means not aligned to a filesystem block */
...
@@ -593,7 +650,7 @@ xfs_file_dio_aio_write(
...
@@ -593,7 +650,7 @@ xfs_file_dio_aio_write(
end
=
iocb
->
ki_pos
+
count
-
1
;
end
=
iocb
->
ki_pos
+
count
-
1
;
/*
/*
* See xfs_file_
read_iter
() for why we do a full-file flush here.
* See xfs_file_
dio_aio_read
() for why we do a full-file flush here.
*/
*/
if
(
mapping
->
nrpages
)
{
if
(
mapping
->
nrpages
)
{
ret
=
filemap_write_and_wait
(
VFS_I
(
ip
)
->
i_mapping
);
ret
=
filemap_write_and_wait
(
VFS_I
(
ip
)
->
i_mapping
);
...
@@ -620,10 +677,12 @@ xfs_file_dio_aio_write(
...
@@ -620,10 +677,12 @@ xfs_file_dio_aio_write(
iolock
=
XFS_IOLOCK_SHARED
;
iolock
=
XFS_IOLOCK_SHARED
;
}
}
trace_xfs_file_direct_write
(
ip
,
count
,
iocb
->
ki_pos
,
0
);
trace_xfs_file_direct_write
(
ip
,
count
,
iocb
->
ki_pos
);
data
=
*
from
;
data
=
*
from
;
ret
=
mapping
->
a_ops
->
direct_IO
(
iocb
,
&
data
);
ret
=
__blockdev_direct_IO
(
iocb
,
inode
,
target
->
bt_bdev
,
&
data
,
xfs_get_blocks_direct
,
xfs_end_io_direct_write
,
NULL
,
DIO_ASYNC_EXTEND
);
/* see generic_file_direct_write() for why this is necessary */
/* see generic_file_direct_write() for why this is necessary */
if
(
mapping
->
nrpages
)
{
if
(
mapping
->
nrpages
)
{
...
@@ -640,10 +699,70 @@ xfs_file_dio_aio_write(
...
@@ -640,10 +699,70 @@ xfs_file_dio_aio_write(
xfs_rw_iunlock
(
ip
,
iolock
);
xfs_rw_iunlock
(
ip
,
iolock
);
/*
/*
* No fallback to buffered IO on errors for XFS. DAX can result in
* No fallback to buffered IO on errors for XFS, direct IO will either
* partial writes, but direct IO will either complete fully or fail.
* complete fully or fail.
*/
ASSERT
(
ret
<
0
||
ret
==
count
);
return
ret
;
}
STATIC
ssize_t
xfs_file_dax_write
(
struct
kiocb
*
iocb
,
struct
iov_iter
*
from
)
{
struct
address_space
*
mapping
=
iocb
->
ki_filp
->
f_mapping
;
struct
inode
*
inode
=
mapping
->
host
;
struct
xfs_inode
*
ip
=
XFS_I
(
inode
);
struct
xfs_mount
*
mp
=
ip
->
i_mount
;
ssize_t
ret
=
0
;
int
unaligned_io
=
0
;
int
iolock
;
struct
iov_iter
data
;
/* "unaligned" here means not aligned to a filesystem block */
if
((
iocb
->
ki_pos
&
mp
->
m_blockmask
)
||
((
iocb
->
ki_pos
+
iov_iter_count
(
from
))
&
mp
->
m_blockmask
))
{
unaligned_io
=
1
;
iolock
=
XFS_IOLOCK_EXCL
;
}
else
if
(
mapping
->
nrpages
)
{
iolock
=
XFS_IOLOCK_EXCL
;
}
else
{
iolock
=
XFS_IOLOCK_SHARED
;
}
xfs_rw_ilock
(
ip
,
iolock
);
ret
=
xfs_file_aio_write_checks
(
iocb
,
from
,
&
iolock
);
if
(
ret
)
goto
out
;
/*
* Yes, even DAX files can have page cache attached to them: A zeroed
* page is inserted into the pagecache when we have to serve a write
* fault on a hole. It should never be dirtied and can simply be
* dropped from the pagecache once we get real data for the page.
*/
*/
ASSERT
(
ret
<
0
||
ret
==
count
||
IS_DAX
(
VFS_I
(
ip
)));
if
(
mapping
->
nrpages
)
{
ret
=
invalidate_inode_pages2
(
mapping
);
WARN_ON_ONCE
(
ret
);
}
if
(
iolock
==
XFS_IOLOCK_EXCL
&&
!
unaligned_io
)
{
xfs_rw_ilock_demote
(
ip
,
XFS_IOLOCK_EXCL
);
iolock
=
XFS_IOLOCK_SHARED
;
}
trace_xfs_file_dax_write
(
ip
,
iov_iter_count
(
from
),
iocb
->
ki_pos
);
data
=
*
from
;
ret
=
dax_do_io
(
iocb
,
inode
,
&
data
,
xfs_get_blocks_direct
,
xfs_end_io_direct_write
,
0
);
if
(
ret
>
0
)
{
iocb
->
ki_pos
+=
ret
;
iov_iter_advance
(
from
,
ret
);
}
out:
xfs_rw_iunlock
(
ip
,
iolock
);
return
ret
;
return
ret
;
}
}
...
@@ -670,8 +789,7 @@ xfs_file_buffered_aio_write(
...
@@ -670,8 +789,7 @@ xfs_file_buffered_aio_write(
current
->
backing_dev_info
=
inode_to_bdi
(
inode
);
current
->
backing_dev_info
=
inode_to_bdi
(
inode
);
write_retry:
write_retry:
trace_xfs_file_buffered_write
(
ip
,
iov_iter_count
(
from
),
trace_xfs_file_buffered_write
(
ip
,
iov_iter_count
(
from
),
iocb
->
ki_pos
);
iocb
->
ki_pos
,
0
);
ret
=
iomap_file_buffered_write
(
iocb
,
from
,
&
xfs_iomap_ops
);
ret
=
iomap_file_buffered_write
(
iocb
,
from
,
&
xfs_iomap_ops
);
if
(
likely
(
ret
>=
0
))
if
(
likely
(
ret
>=
0
))
iocb
->
ki_pos
+=
ret
;
iocb
->
ki_pos
+=
ret
;
...
@@ -726,7 +844,9 @@ xfs_file_write_iter(
...
@@ -726,7 +844,9 @@ xfs_file_write_iter(
if
(
XFS_FORCED_SHUTDOWN
(
ip
->
i_mount
))
if
(
XFS_FORCED_SHUTDOWN
(
ip
->
i_mount
))
return
-
EIO
;
return
-
EIO
;
if
((
iocb
->
ki_flags
&
IOCB_DIRECT
)
||
IS_DAX
(
inode
))
if
(
IS_DAX
(
inode
))
ret
=
xfs_file_dax_write
(
iocb
,
from
);
else
if
(
iocb
->
ki_flags
&
IOCB_DIRECT
)
ret
=
xfs_file_dio_aio_write
(
iocb
,
from
);
ret
=
xfs_file_dio_aio_write
(
iocb
,
from
);
else
else
ret
=
xfs_file_buffered_aio_write
(
iocb
,
from
);
ret
=
xfs_file_buffered_aio_write
(
iocb
,
from
);
...
...
fs/xfs/xfs_inode.h
View file @
b47ec80b
...
@@ -473,14 +473,4 @@ do { \
...
@@ -473,14 +473,4 @@ do { \
extern
struct
kmem_zone
*
xfs_inode_zone
;
extern
struct
kmem_zone
*
xfs_inode_zone
;
/*
* Flags for read/write calls
*/
#define XFS_IO_ISDIRECT 0x00001
/* bypass page cache */
#define XFS_IO_INVIS 0x00002
/* don't update inode timestamps */
#define XFS_IO_FLAGS \
{ XFS_IO_ISDIRECT, "DIRECT" }, \
{ XFS_IO_INVIS, "INVIS"}
#endif
/* __XFS_INODE_H__ */
#endif
/* __XFS_INODE_H__ */
fs/xfs/xfs_ioctl.c
View file @
b47ec80b
...
@@ -595,13 +595,12 @@ xfs_attrmulti_by_handle(
...
@@ -595,13 +595,12 @@ xfs_attrmulti_by_handle(
int
int
xfs_ioc_space
(
xfs_ioc_space
(
struct
xfs_inode
*
ip
,
struct
inode
*
inode
,
struct
file
*
filp
,
struct
file
*
filp
,
int
ioflags
,
unsigned
int
cmd
,
unsigned
int
cmd
,
xfs_flock64_t
*
bf
)
xfs_flock64_t
*
bf
)
{
{
struct
inode
*
inode
=
file_inode
(
filp
);
struct
xfs_inode
*
ip
=
XFS_I
(
inode
);
struct
iattr
iattr
;
struct
iattr
iattr
;
enum
xfs_prealloc_flags
flags
=
0
;
enum
xfs_prealloc_flags
flags
=
0
;
uint
iolock
=
XFS_IOLOCK_EXCL
;
uint
iolock
=
XFS_IOLOCK_EXCL
;
...
@@ -626,7 +625,7 @@ xfs_ioc_space(
...
@@ -626,7 +625,7 @@ xfs_ioc_space(
if
(
filp
->
f_flags
&
O_DSYNC
)
if
(
filp
->
f_flags
&
O_DSYNC
)
flags
|=
XFS_PREALLOC_SYNC
;
flags
|=
XFS_PREALLOC_SYNC
;
if
(
ioflags
&
XFS_IO_INVIS
)
if
(
filp
->
f_mode
&
FMODE_NOCMTIME
)
flags
|=
XFS_PREALLOC_INVISIBLE
;
flags
|=
XFS_PREALLOC_INVISIBLE
;
error
=
mnt_want_write_file
(
filp
);
error
=
mnt_want_write_file
(
filp
);
...
@@ -1464,8 +1463,7 @@ xfs_getbmap_format(void **ap, struct getbmapx *bmv, int *full)
...
@@ -1464,8 +1463,7 @@ xfs_getbmap_format(void **ap, struct getbmapx *bmv, int *full)
STATIC
int
STATIC
int
xfs_ioc_getbmap
(
xfs_ioc_getbmap
(
struct
xfs_inode
*
ip
,
struct
file
*
file
,
int
ioflags
,
unsigned
int
cmd
,
unsigned
int
cmd
,
void
__user
*
arg
)
void
__user
*
arg
)
{
{
...
@@ -1479,10 +1477,10 @@ xfs_ioc_getbmap(
...
@@ -1479,10 +1477,10 @@ xfs_ioc_getbmap(
return
-
EINVAL
;
return
-
EINVAL
;
bmx
.
bmv_iflags
=
(
cmd
==
XFS_IOC_GETBMAPA
?
BMV_IF_ATTRFORK
:
0
);
bmx
.
bmv_iflags
=
(
cmd
==
XFS_IOC_GETBMAPA
?
BMV_IF_ATTRFORK
:
0
);
if
(
ioflags
&
XFS_IO_INVIS
)
if
(
file
->
f_mode
&
FMODE_NOCMTIME
)
bmx
.
bmv_iflags
|=
BMV_IF_NO_DMAPI_READ
;
bmx
.
bmv_iflags
|=
BMV_IF_NO_DMAPI_READ
;
error
=
xfs_getbmap
(
ip
,
&
bmx
,
xfs_getbmap_format
,
error
=
xfs_getbmap
(
XFS_I
(
file_inode
(
file
))
,
&
bmx
,
xfs_getbmap_format
,
(
__force
struct
getbmap
*
)
arg
+
1
);
(
__force
struct
getbmap
*
)
arg
+
1
);
if
(
error
)
if
(
error
)
return
error
;
return
error
;
...
@@ -1630,12 +1628,8 @@ xfs_file_ioctl(
...
@@ -1630,12 +1628,8 @@ xfs_file_ioctl(
struct
xfs_inode
*
ip
=
XFS_I
(
inode
);
struct
xfs_inode
*
ip
=
XFS_I
(
inode
);
struct
xfs_mount
*
mp
=
ip
->
i_mount
;
struct
xfs_mount
*
mp
=
ip
->
i_mount
;
void
__user
*
arg
=
(
void
__user
*
)
p
;
void
__user
*
arg
=
(
void
__user
*
)
p
;
int
ioflags
=
0
;
int
error
;
int
error
;
if
(
filp
->
f_mode
&
FMODE_NOCMTIME
)
ioflags
|=
XFS_IO_INVIS
;
trace_xfs_file_ioctl
(
ip
);
trace_xfs_file_ioctl
(
ip
);
switch
(
cmd
)
{
switch
(
cmd
)
{
...
@@ -1654,7 +1648,7 @@ xfs_file_ioctl(
...
@@ -1654,7 +1648,7 @@ xfs_file_ioctl(
if
(
copy_from_user
(
&
bf
,
arg
,
sizeof
(
bf
)))
if
(
copy_from_user
(
&
bf
,
arg
,
sizeof
(
bf
)))
return
-
EFAULT
;
return
-
EFAULT
;
return
xfs_ioc_space
(
ip
,
inode
,
filp
,
ioflags
,
cmd
,
&
bf
);
return
xfs_ioc_space
(
filp
,
cmd
,
&
bf
);
}
}
case
XFS_IOC_DIOINFO
:
{
case
XFS_IOC_DIOINFO
:
{
struct
dioattr
da
;
struct
dioattr
da
;
...
@@ -1713,7 +1707,7 @@ xfs_file_ioctl(
...
@@ -1713,7 +1707,7 @@ xfs_file_ioctl(
case
XFS_IOC_GETBMAP
:
case
XFS_IOC_GETBMAP
:
case
XFS_IOC_GETBMAPA
:
case
XFS_IOC_GETBMAPA
:
return
xfs_ioc_getbmap
(
ip
,
ioflags
,
cmd
,
arg
);
return
xfs_ioc_getbmap
(
filp
,
cmd
,
arg
);
case
XFS_IOC_GETBMAPX
:
case
XFS_IOC_GETBMAPX
:
return
xfs_ioc_getbmapx
(
ip
,
arg
);
return
xfs_ioc_getbmapx
(
ip
,
arg
);
...
...
fs/xfs/xfs_ioctl.h
View file @
b47ec80b
...
@@ -20,10 +20,7 @@
...
@@ -20,10 +20,7 @@
extern
int
extern
int
xfs_ioc_space
(
xfs_ioc_space
(
struct
xfs_inode
*
ip
,
struct
inode
*
inode
,
struct
file
*
filp
,
struct
file
*
filp
,
int
ioflags
,
unsigned
int
cmd
,
unsigned
int
cmd
,
xfs_flock64_t
*
bf
);
xfs_flock64_t
*
bf
);
...
...
fs/xfs/xfs_ioctl32.c
View file @
b47ec80b
...
@@ -532,12 +532,8 @@ xfs_file_compat_ioctl(
...
@@ -532,12 +532,8 @@ xfs_file_compat_ioctl(
struct
xfs_inode
*
ip
=
XFS_I
(
inode
);
struct
xfs_inode
*
ip
=
XFS_I
(
inode
);
struct
xfs_mount
*
mp
=
ip
->
i_mount
;
struct
xfs_mount
*
mp
=
ip
->
i_mount
;
void
__user
*
arg
=
(
void
__user
*
)
p
;
void
__user
*
arg
=
(
void
__user
*
)
p
;
int
ioflags
=
0
;
int
error
;
int
error
;
if
(
filp
->
f_mode
&
FMODE_NOCMTIME
)
ioflags
|=
XFS_IO_INVIS
;
trace_xfs_file_compat_ioctl
(
ip
);
trace_xfs_file_compat_ioctl
(
ip
);
switch
(
cmd
)
{
switch
(
cmd
)
{
...
@@ -589,7 +585,7 @@ xfs_file_compat_ioctl(
...
@@ -589,7 +585,7 @@ xfs_file_compat_ioctl(
if
(
xfs_compat_flock64_copyin
(
&
bf
,
arg
))
if
(
xfs_compat_flock64_copyin
(
&
bf
,
arg
))
return
-
EFAULT
;
return
-
EFAULT
;
cmd
=
_NATIVE_IOC
(
cmd
,
struct
xfs_flock64
);
cmd
=
_NATIVE_IOC
(
cmd
,
struct
xfs_flock64
);
return
xfs_ioc_space
(
ip
,
inode
,
filp
,
ioflags
,
cmd
,
&
bf
);
return
xfs_ioc_space
(
filp
,
cmd
,
&
bf
);
}
}
case
XFS_IOC_FSGEOMETRY_V1_32
:
case
XFS_IOC_FSGEOMETRY_V1_32
:
return
xfs_compat_ioc_fsgeometry_v1
(
mp
,
arg
);
return
xfs_compat_ioc_fsgeometry_v1
(
mp
,
arg
);
...
...
fs/xfs/xfs_trace.h
View file @
b47ec80b
...
@@ -1135,15 +1135,14 @@ TRACE_EVENT(xfs_log_assign_tail_lsn,
...
@@ -1135,15 +1135,14 @@ TRACE_EVENT(xfs_log_assign_tail_lsn,
)
)
DECLARE_EVENT_CLASS
(
xfs_file_class
,
DECLARE_EVENT_CLASS
(
xfs_file_class
,
TP_PROTO
(
struct
xfs_inode
*
ip
,
size_t
count
,
loff_t
offset
,
int
flags
),
TP_PROTO
(
struct
xfs_inode
*
ip
,
size_t
count
,
loff_t
offset
),
TP_ARGS
(
ip
,
count
,
offset
,
flags
),
TP_ARGS
(
ip
,
count
,
offset
),
TP_STRUCT__entry
(
TP_STRUCT__entry
(
__field
(
dev_t
,
dev
)
__field
(
dev_t
,
dev
)
__field
(
xfs_ino_t
,
ino
)
__field
(
xfs_ino_t
,
ino
)
__field
(
xfs_fsize_t
,
size
)
__field
(
xfs_fsize_t
,
size
)
__field
(
loff_t
,
offset
)
__field
(
loff_t
,
offset
)
__field
(
size_t
,
count
)
__field
(
size_t
,
count
)
__field
(
int
,
flags
)
),
),
TP_fast_assign
(
TP_fast_assign
(
__entry
->
dev
=
VFS_I
(
ip
)
->
i_sb
->
s_dev
;
__entry
->
dev
=
VFS_I
(
ip
)
->
i_sb
->
s_dev
;
...
@@ -1151,25 +1150,25 @@ DECLARE_EVENT_CLASS(xfs_file_class,
...
@@ -1151,25 +1150,25 @@ DECLARE_EVENT_CLASS(xfs_file_class,
__entry
->
size
=
ip
->
i_d
.
di_size
;
__entry
->
size
=
ip
->
i_d
.
di_size
;
__entry
->
offset
=
offset
;
__entry
->
offset
=
offset
;
__entry
->
count
=
count
;
__entry
->
count
=
count
;
__entry
->
flags
=
flags
;
),
),
TP_printk
(
"dev %d:%d ino 0x%llx size 0x%llx "
TP_printk
(
"dev %d:%d ino 0x%llx size 0x%llx offset 0x%llx count 0x%zx"
,
"offset 0x%llx count 0x%zx ioflags %s"
,
MAJOR
(
__entry
->
dev
),
MINOR
(
__entry
->
dev
),
MAJOR
(
__entry
->
dev
),
MINOR
(
__entry
->
dev
),
__entry
->
ino
,
__entry
->
ino
,
__entry
->
size
,
__entry
->
size
,
__entry
->
offset
,
__entry
->
offset
,
__entry
->
count
,
__entry
->
count
)
__print_flags
(
__entry
->
flags
,
"|"
,
XFS_IO_FLAGS
))
)
)
#define DEFINE_RW_EVENT(name) \
#define DEFINE_RW_EVENT(name) \
DEFINE_EVENT(xfs_file_class, name, \
DEFINE_EVENT(xfs_file_class, name, \
TP_PROTO(struct xfs_inode *ip, size_t count, loff_t offset, int flags), \
TP_PROTO(struct xfs_inode *ip, size_t count, loff_t offset), \
TP_ARGS(ip, count, offset, flags))
TP_ARGS(ip, count, offset))
DEFINE_RW_EVENT
(
xfs_file_read
);
DEFINE_RW_EVENT
(
xfs_file_buffered_read
);
DEFINE_RW_EVENT
(
xfs_file_direct_read
);
DEFINE_RW_EVENT
(
xfs_file_dax_read
);
DEFINE_RW_EVENT
(
xfs_file_buffered_write
);
DEFINE_RW_EVENT
(
xfs_file_buffered_write
);
DEFINE_RW_EVENT
(
xfs_file_direct_write
);
DEFINE_RW_EVENT
(
xfs_file_direct_write
);
DEFINE_RW_EVENT
(
xfs_file_dax_write
);
DEFINE_RW_EVENT
(
xfs_file_splice_read
);
DEFINE_RW_EVENT
(
xfs_file_splice_read
);
DECLARE_EVENT_CLASS
(
xfs_page_class
,
DECLARE_EVENT_CLASS
(
xfs_page_class
,
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment