Commit 7f348f8c authored by Matthew Wilcox (Oracle)'s avatar Matthew Wilcox (Oracle) Committed by Richard Weinberger

ubifs: Pass a folio into ubifs_bulk_read() and ubifs_do_bulk_read()

This saves a single call to compound_head().
Signed-off-by: default avatarMatthew Wilcox (Oracle) <willy@infradead.org>
Reviewed-by: default avatarZhihao Cheng <chengzhihao1@huawei.com>
Signed-off-by: default avatarRichard Weinberger <richard@nod.at>
parent 45d76698
...@@ -703,15 +703,15 @@ static int populate_page(struct ubifs_info *c, struct page *page, ...@@ -703,15 +703,15 @@ static int populate_page(struct ubifs_info *c, struct page *page,
* ubifs_do_bulk_read - do bulk-read. * ubifs_do_bulk_read - do bulk-read.
* @c: UBIFS file-system description object * @c: UBIFS file-system description object
* @bu: bulk-read information * @bu: bulk-read information
* @page1: first page to read * @folio1: first folio to read
* *
* Returns: %1 if the bulk-read is done, otherwise %0 is returned. * Returns: %1 if the bulk-read is done, otherwise %0 is returned.
*/ */
static int ubifs_do_bulk_read(struct ubifs_info *c, struct bu_info *bu, static int ubifs_do_bulk_read(struct ubifs_info *c, struct bu_info *bu,
struct page *page1) struct folio *folio1)
{ {
pgoff_t offset = page1->index, end_index; pgoff_t offset = folio1->index, end_index;
struct address_space *mapping = page1->mapping; struct address_space *mapping = folio1->mapping;
struct inode *inode = mapping->host; struct inode *inode = mapping->host;
struct ubifs_inode *ui = ubifs_inode(inode); struct ubifs_inode *ui = ubifs_inode(inode);
int err, page_idx, page_cnt, ret = 0, n = 0; int err, page_idx, page_cnt, ret = 0, n = 0;
...@@ -761,11 +761,11 @@ static int ubifs_do_bulk_read(struct ubifs_info *c, struct bu_info *bu, ...@@ -761,11 +761,11 @@ static int ubifs_do_bulk_read(struct ubifs_info *c, struct bu_info *bu,
goto out_warn; goto out_warn;
} }
err = populate_page(c, page1, bu, &n); err = populate_page(c, &folio1->page, bu, &n);
if (err) if (err)
goto out_warn; goto out_warn;
unlock_page(page1); folio_unlock(folio1);
ret = 1; ret = 1;
isize = i_size_read(inode); isize = i_size_read(inode);
...@@ -810,7 +810,7 @@ static int ubifs_do_bulk_read(struct ubifs_info *c, struct bu_info *bu, ...@@ -810,7 +810,7 @@ static int ubifs_do_bulk_read(struct ubifs_info *c, struct bu_info *bu,
/** /**
* ubifs_bulk_read - determine whether to bulk-read and, if so, do it. * ubifs_bulk_read - determine whether to bulk-read and, if so, do it.
* @page: page from which to start bulk-read. * @folio: folio from which to start bulk-read.
* *
* Some flash media are capable of reading sequentially at faster rates. UBIFS * Some flash media are capable of reading sequentially at faster rates. UBIFS
* bulk-read facility is designed to take advantage of that, by reading in one * bulk-read facility is designed to take advantage of that, by reading in one
...@@ -819,12 +819,12 @@ static int ubifs_do_bulk_read(struct ubifs_info *c, struct bu_info *bu, ...@@ -819,12 +819,12 @@ static int ubifs_do_bulk_read(struct ubifs_info *c, struct bu_info *bu,
* *
* Returns: %1 if a bulk-read is done and %0 otherwise. * Returns: %1 if a bulk-read is done and %0 otherwise.
*/ */
static int ubifs_bulk_read(struct page *page) static int ubifs_bulk_read(struct folio *folio)
{ {
struct inode *inode = page->mapping->host; struct inode *inode = folio->mapping->host;
struct ubifs_info *c = inode->i_sb->s_fs_info; struct ubifs_info *c = inode->i_sb->s_fs_info;
struct ubifs_inode *ui = ubifs_inode(inode); struct ubifs_inode *ui = ubifs_inode(inode);
pgoff_t index = page->index, last_page_read = ui->last_page_read; pgoff_t index = folio->index, last_page_read = ui->last_page_read;
struct bu_info *bu; struct bu_info *bu;
int err = 0, allocated = 0; int err = 0, allocated = 0;
...@@ -872,8 +872,8 @@ static int ubifs_bulk_read(struct page *page) ...@@ -872,8 +872,8 @@ static int ubifs_bulk_read(struct page *page)
bu->buf_len = c->max_bu_buf_len; bu->buf_len = c->max_bu_buf_len;
data_key_init(c, &bu->key, inode->i_ino, data_key_init(c, &bu->key, inode->i_ino,
page->index << UBIFS_BLOCKS_PER_PAGE_SHIFT); folio->index << UBIFS_BLOCKS_PER_PAGE_SHIFT);
err = ubifs_do_bulk_read(c, bu, page); err = ubifs_do_bulk_read(c, bu, folio);
if (!allocated) if (!allocated)
mutex_unlock(&c->bu_mutex); mutex_unlock(&c->bu_mutex);
...@@ -887,9 +887,7 @@ static int ubifs_bulk_read(struct page *page) ...@@ -887,9 +887,7 @@ static int ubifs_bulk_read(struct page *page)
static int ubifs_read_folio(struct file *file, struct folio *folio) static int ubifs_read_folio(struct file *file, struct folio *folio)
{ {
struct page *page = &folio->page; if (ubifs_bulk_read(folio))
if (ubifs_bulk_read(page))
return 0; return 0;
do_readpage(folio); do_readpage(folio);
folio_unlock(folio); folio_unlock(folio);
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment