Commit 6907e3c0 authored by Darrick J. Wong's avatar Darrick J. Wong Committed by Chandan Babu R

xfs: add file_{get,put}_folio

Add helper similar to file_{get,set}_page, but which deal with folios
and don't allocate new folio unless explicitly asked to, which map
to shmem_get_folio instead of calling into the aops.
Signed-off-by: default avatar"Darrick J. Wong" <djwong@kernel.org>
Signed-off-by: default avatarChristoph Hellwig <hch@lst.de>
Reviewed-by: default avatarKent Overstreet <kent.overstreet@linux.dev>
Signed-off-by: default avatarChandan Babu R <chandanbabu@kernel.org>
parent e97d70a5
...@@ -908,6 +908,8 @@ DEFINE_XFILE_EVENT(xfile_store); ...@@ -908,6 +908,8 @@ DEFINE_XFILE_EVENT(xfile_store);
DEFINE_XFILE_EVENT(xfile_seek_data); DEFINE_XFILE_EVENT(xfile_seek_data);
DEFINE_XFILE_EVENT(xfile_get_page); DEFINE_XFILE_EVENT(xfile_get_page);
DEFINE_XFILE_EVENT(xfile_put_page); DEFINE_XFILE_EVENT(xfile_put_page);
DEFINE_XFILE_EVENT(xfile_get_folio);
DEFINE_XFILE_EVENT(xfile_put_folio);
TRACE_EVENT(xfarray_create, TRACE_EVENT(xfarray_create,
TP_PROTO(struct xfarray *xfa, unsigned long long required_capacity), TP_PROTO(struct xfarray *xfa, unsigned long long required_capacity),
......
...@@ -340,3 +340,77 @@ xfile_put_page( ...@@ -340,3 +340,77 @@ xfile_put_page(
return -EIO; return -EIO;
return 0; return 0;
} }
/*
* Grab the (locked) folio for a memory object. The object cannot span a folio
* boundary. Returns the locked folio if successful, NULL if there was no
* folio or it didn't cover the range requested, or an ERR_PTR on failure.
*/
struct folio *
xfile_get_folio(
struct xfile *xf,
loff_t pos,
size_t len,
unsigned int flags)
{
struct inode *inode = file_inode(xf->file);
struct folio *folio = NULL;
unsigned int pflags;
int error;
if (inode->i_sb->s_maxbytes - pos < len)
return ERR_PTR(-ENOMEM);
trace_xfile_get_folio(xf, pos, len);
/*
* Increase the file size first so that shmem_get_folio(..., SGP_CACHE),
* actually allocates a folio instead of erroring out.
*/
if ((flags & XFILE_ALLOC) && pos + len > i_size_read(inode))
i_size_write(inode, pos + len);
pflags = memalloc_nofs_save();
error = shmem_get_folio(inode, pos >> PAGE_SHIFT, &folio,
(flags & XFILE_ALLOC) ? SGP_CACHE : SGP_READ);
memalloc_nofs_restore(pflags);
if (error)
return ERR_PTR(error);
if (!folio)
return NULL;
if (len > folio_size(folio) - offset_in_folio(folio, pos)) {
folio_unlock(folio);
folio_put(folio);
return NULL;
}
if (filemap_check_wb_err(inode->i_mapping, 0)) {
folio_unlock(folio);
folio_put(folio);
return ERR_PTR(-EIO);
}
/*
* Mark the folio dirty so that it won't be reclaimed once we drop the
* (potentially last) reference in xfile_put_folio.
*/
if (flags & XFILE_ALLOC)
folio_set_dirty(folio);
return folio;
}
/*
* Release the (locked) folio for a memory object.
*/
void
xfile_put_folio(
struct xfile *xf,
struct folio *folio)
{
trace_xfile_put_folio(xf, folio_pos(folio), folio_size(folio));
folio_unlock(folio);
folio_put(folio);
}
...@@ -39,4 +39,11 @@ int xfile_get_page(struct xfile *xf, loff_t offset, unsigned int len, ...@@ -39,4 +39,11 @@ int xfile_get_page(struct xfile *xf, loff_t offset, unsigned int len,
struct xfile_page *xbuf); struct xfile_page *xbuf);
int xfile_put_page(struct xfile *xf, struct xfile_page *xbuf); int xfile_put_page(struct xfile *xf, struct xfile_page *xbuf);
#define XFILE_MAX_FOLIO_SIZE (PAGE_SIZE << MAX_PAGECACHE_ORDER)
#define XFILE_ALLOC (1 << 0) /* allocate folio if not present */
struct folio *xfile_get_folio(struct xfile *xf, loff_t offset, size_t len,
unsigned int flags);
void xfile_put_folio(struct xfile *xf, struct folio *folio);
#endif /* __XFS_SCRUB_XFILE_H__ */ #endif /* __XFS_SCRUB_XFILE_H__ */
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment