Commit b8072d5b authored by Linus Torvalds's avatar Linus Torvalds

Merge tag 'for_v5.5-rc1' of git://git.kernel.org/pub/scm/linux/kernel/git/jack/linux-fs

Pull ext2, quota, reiserfs cleanups and fixes from Jan Kara:

 - Refactor the quota on/off kernel internal interfaces (mostly for
   ubifs quota support as ubifs does not want to have inodes holding
   quota information)

 - A few other small quota fixes and cleanups

 - Various small ext2 fixes and cleanups

 - Reiserfs xattr fix and one cleanup

* tag 'for_v5.5-rc1' of git://git.kernel.org/pub/scm/linux/kernel/git/jack/linux-fs: (28 commits)
  ext2: code cleanup for descriptor_loc()
  fs/quota: handle overflows of sysctl fs.quota.* and report as unsigned long
  ext2: fix improper function comment
  ext2: code cleanup for ext2_try_to_allocate()
  ext2: skip unnecessary operations in ext2_try_to_allocate()
  ext2: Simplify initialization in ext2_try_to_allocate()
  ext2: code cleanup by calling ext2_group_last_block_no()
  ext2: introduce new helper ext2_group_last_block_no()
  reiserfs: replace open-coded atomic_dec_and_mutex_lock()
  ext2: check err when partial != NULL
  quota: Handle quotas without quota inodes in dquot_get_state()
  quota: Make dquot_disable() work without quota inodes
  quota: Drop dquot_enable()
  fs: Use dquot_load_quota_inode() from filesystems
  quota: Rename vfs_load_quota_inode() to dquot_load_quota_inode()
  quota: Simplify dquot_resume()
  quota: Factor out setup of quota inode
  quota: Check that quota is not dirty before release
  quota: fix livelock in dquot_writeback_dquots
  ext2: don't set *count in the case of failure in ext2_try_to_allocate()
  ...
parents e2d73c30 545886fe
...@@ -269,7 +269,7 @@ goal_in_my_reservation(struct ext2_reserve_window *rsv, ext2_grpblk_t grp_goal, ...@@ -269,7 +269,7 @@ goal_in_my_reservation(struct ext2_reserve_window *rsv, ext2_grpblk_t grp_goal,
ext2_fsblk_t group_first_block, group_last_block; ext2_fsblk_t group_first_block, group_last_block;
group_first_block = ext2_group_first_block_no(sb, group); group_first_block = ext2_group_first_block_no(sb, group);
group_last_block = group_first_block + EXT2_BLOCKS_PER_GROUP(sb) - 1; group_last_block = ext2_group_last_block_no(sb, group);
if ((rsv->_rsv_start > group_last_block) || if ((rsv->_rsv_start > group_last_block) ||
(rsv->_rsv_end < group_first_block)) (rsv->_rsv_end < group_first_block))
...@@ -666,37 +666,24 @@ ext2_try_to_allocate(struct super_block *sb, int group, ...@@ -666,37 +666,24 @@ ext2_try_to_allocate(struct super_block *sb, int group,
unsigned long *count, unsigned long *count,
struct ext2_reserve_window *my_rsv) struct ext2_reserve_window *my_rsv)
{ {
ext2_fsblk_t group_first_block; ext2_fsblk_t group_first_block = ext2_group_first_block_no(sb, group);
ext2_fsblk_t group_last_block = ext2_group_last_block_no(sb, group);
ext2_grpblk_t start, end; ext2_grpblk_t start, end;
unsigned long num = 0; unsigned long num = 0;
start = 0;
end = group_last_block - group_first_block + 1;
/* we do allocation within the reservation window if we have a window */ /* we do allocation within the reservation window if we have a window */
if (my_rsv) { if (my_rsv) {
group_first_block = ext2_group_first_block_no(sb, group);
if (my_rsv->_rsv_start >= group_first_block) if (my_rsv->_rsv_start >= group_first_block)
start = my_rsv->_rsv_start - group_first_block; start = my_rsv->_rsv_start - group_first_block;
else if (my_rsv->_rsv_end < group_last_block)
/* reservation window cross group boundary */ end = my_rsv->_rsv_end - group_first_block + 1;
start = 0; if (grp_goal < start || grp_goal >= end)
end = my_rsv->_rsv_end - group_first_block + 1;
if (end > EXT2_BLOCKS_PER_GROUP(sb))
/* reservation window crosses group boundary */
end = EXT2_BLOCKS_PER_GROUP(sb);
if ((start <= grp_goal) && (grp_goal < end))
start = grp_goal;
else
grp_goal = -1; grp_goal = -1;
} else {
if (grp_goal > 0)
start = grp_goal;
else
start = 0;
end = EXT2_BLOCKS_PER_GROUP(sb);
} }
BUG_ON(start > EXT2_BLOCKS_PER_GROUP(sb)); BUG_ON(start > EXT2_BLOCKS_PER_GROUP(sb));
repeat:
if (grp_goal < 0) { if (grp_goal < 0) {
grp_goal = find_next_usable_block(start, bitmap_bh, end); grp_goal = find_next_usable_block(start, bitmap_bh, end);
if (grp_goal < 0) if (grp_goal < 0)
...@@ -711,32 +698,23 @@ ext2_try_to_allocate(struct super_block *sb, int group, ...@@ -711,32 +698,23 @@ ext2_try_to_allocate(struct super_block *sb, int group,
; ;
} }
} }
start = grp_goal;
if (ext2_set_bit_atomic(sb_bgl_lock(EXT2_SB(sb), group), grp_goal, for (; num < *count && grp_goal < end; grp_goal++) {
bitmap_bh->b_data)) { if (ext2_set_bit_atomic(sb_bgl_lock(EXT2_SB(sb), group),
/*
* The block was allocated by another thread, or it was
* allocated and then freed by another thread
*/
start++;
grp_goal++;
if (start >= end)
goto fail_access;
goto repeat;
}
num++;
grp_goal++;
while (num < *count && grp_goal < end
&& !ext2_set_bit_atomic(sb_bgl_lock(EXT2_SB(sb), group),
grp_goal, bitmap_bh->b_data)) { grp_goal, bitmap_bh->b_data)) {
if (num == 0)
continue;
break;
}
num++; num++;
grp_goal++;
} }
if (num == 0)
goto fail_access;
*count = num; *count = num;
return grp_goal - num; return grp_goal - num;
fail_access: fail_access:
*count = num;
return -1; return -1;
} }
...@@ -754,10 +732,9 @@ ext2_try_to_allocate(struct super_block *sb, int group, ...@@ -754,10 +732,9 @@ ext2_try_to_allocate(struct super_block *sb, int group,
* but we will shift to the place where start_block is, * but we will shift to the place where start_block is,
* then start from there, when looking for a reservable space. * then start from there, when looking for a reservable space.
* *
* @size: the target new reservation window size * @sb: the super block.
* *
* @group_first_block: the first block we consider to start * @start_block: the first block we consider to start the real search from
* the real search from
* *
* @last_block: * @last_block:
* the maximum block number that our goal reservable space * the maximum block number that our goal reservable space
...@@ -908,7 +885,7 @@ static int alloc_new_reservation(struct ext2_reserve_window_node *my_rsv, ...@@ -908,7 +885,7 @@ static int alloc_new_reservation(struct ext2_reserve_window_node *my_rsv,
spinlock_t *rsv_lock = &EXT2_SB(sb)->s_rsv_window_lock; spinlock_t *rsv_lock = &EXT2_SB(sb)->s_rsv_window_lock;
group_first_block = ext2_group_first_block_no(sb, group); group_first_block = ext2_group_first_block_no(sb, group);
group_end_block = group_first_block + (EXT2_BLOCKS_PER_GROUP(sb) - 1); group_end_block = ext2_group_last_block_no(sb, group);
if (grp_goal < 0) if (grp_goal < 0)
start_block = group_first_block; start_block = group_first_block;
...@@ -1115,7 +1092,7 @@ ext2_try_to_allocate_with_rsv(struct super_block *sb, unsigned int group, ...@@ -1115,7 +1092,7 @@ ext2_try_to_allocate_with_rsv(struct super_block *sb, unsigned int group,
* first block is the block number of the first block in this group * first block is the block number of the first block in this group
*/ */
group_first_block = ext2_group_first_block_no(sb, group); group_first_block = ext2_group_first_block_no(sb, group);
group_last_block = group_first_block + (EXT2_BLOCKS_PER_GROUP(sb) - 1); group_last_block = ext2_group_last_block_no(sb, group);
/* /*
* Basically we will allocate a new block from inode's reservation * Basically we will allocate a new block from inode's reservation
...@@ -1313,6 +1290,13 @@ ext2_fsblk_t ext2_new_blocks(struct inode *inode, ext2_fsblk_t goal, ...@@ -1313,6 +1290,13 @@ ext2_fsblk_t ext2_new_blocks(struct inode *inode, ext2_fsblk_t goal,
if (free_blocks > 0) { if (free_blocks > 0) {
grp_target_blk = ((goal - le32_to_cpu(es->s_first_data_block)) % grp_target_blk = ((goal - le32_to_cpu(es->s_first_data_block)) %
EXT2_BLOCKS_PER_GROUP(sb)); EXT2_BLOCKS_PER_GROUP(sb));
/*
* In case we retry allocation (due to fs reservation not
* working out or fs corruption), the bitmap_bh is non-null
* pointer and we have to release it before calling
* read_block_bitmap().
*/
brelse(bitmap_bh);
bitmap_bh = read_block_bitmap(sb, group_no); bitmap_bh = read_block_bitmap(sb, group_no);
if (!bitmap_bh) if (!bitmap_bh)
goto io_error; goto io_error;
...@@ -1404,6 +1388,7 @@ ext2_fsblk_t ext2_new_blocks(struct inode *inode, ext2_fsblk_t goal, ...@@ -1404,6 +1388,7 @@ ext2_fsblk_t ext2_new_blocks(struct inode *inode, ext2_fsblk_t goal,
* use. So we may want to selectively mark some of the blocks * use. So we may want to selectively mark some of the blocks
* as free * as free
*/ */
num = *count;
goto retry_alloc; goto retry_alloc;
} }
......
...@@ -813,6 +813,18 @@ ext2_group_first_block_no(struct super_block *sb, unsigned long group_no) ...@@ -813,6 +813,18 @@ ext2_group_first_block_no(struct super_block *sb, unsigned long group_no)
le32_to_cpu(EXT2_SB(sb)->s_es->s_first_data_block); le32_to_cpu(EXT2_SB(sb)->s_es->s_first_data_block);
} }
static inline ext2_fsblk_t
ext2_group_last_block_no(struct super_block *sb, unsigned long group_no)
{
struct ext2_sb_info *sbi = EXT2_SB(sb);
if (group_no == sbi->s_groups_count - 1)
return le32_to_cpu(sbi->s_es->s_blocks_count) - 1;
else
return ext2_group_first_block_no(sb, group_no) +
EXT2_BLOCKS_PER_GROUP(sb) - 1;
}
#define ext2_set_bit __test_and_set_bit_le #define ext2_set_bit __test_and_set_bit_le
#define ext2_clear_bit __test_and_clear_bit_le #define ext2_clear_bit __test_and_clear_bit_le
#define ext2_test_bit test_bit_le #define ext2_test_bit test_bit_le
......
...@@ -701,10 +701,13 @@ static int ext2_get_blocks(struct inode *inode, ...@@ -701,10 +701,13 @@ static int ext2_get_blocks(struct inode *inode,
if (!partial) { if (!partial) {
count++; count++;
mutex_unlock(&ei->truncate_mutex); mutex_unlock(&ei->truncate_mutex);
if (err)
goto cleanup;
goto got_it; goto got_it;
} }
if (err) {
mutex_unlock(&ei->truncate_mutex);
goto cleanup;
}
} }
/* /*
......
...@@ -145,10 +145,13 @@ long ext2_ioctl(struct file *filp, unsigned int cmd, unsigned long arg) ...@@ -145,10 +145,13 @@ long ext2_ioctl(struct file *filp, unsigned int cmd, unsigned long arg)
if (ei->i_block_alloc_info){ if (ei->i_block_alloc_info){
struct ext2_reserve_window_node *rsv = &ei->i_block_alloc_info->rsv_window_node; struct ext2_reserve_window_node *rsv = &ei->i_block_alloc_info->rsv_window_node;
rsv->rsv_goal_size = rsv_window_size; rsv->rsv_goal_size = rsv_window_size;
} else {
ret = -ENOMEM;
} }
mutex_unlock(&ei->truncate_mutex); mutex_unlock(&ei->truncate_mutex);
mnt_drop_write_file(filp); mnt_drop_write_file(filp);
return 0; return ret;
} }
default: default:
return -ENOTTY; return -ENOTTY;
......
...@@ -702,13 +702,7 @@ static int ext2_check_descriptors(struct super_block *sb) ...@@ -702,13 +702,7 @@ static int ext2_check_descriptors(struct super_block *sb)
for (i = 0; i < sbi->s_groups_count; i++) { for (i = 0; i < sbi->s_groups_count; i++) {
struct ext2_group_desc *gdp = ext2_get_group_desc(sb, i, NULL); struct ext2_group_desc *gdp = ext2_get_group_desc(sb, i, NULL);
ext2_fsblk_t first_block = ext2_group_first_block_no(sb, i); ext2_fsblk_t first_block = ext2_group_first_block_no(sb, i);
ext2_fsblk_t last_block; ext2_fsblk_t last_block = ext2_group_last_block_no(sb, i);
if (i == sbi->s_groups_count - 1)
last_block = le32_to_cpu(sbi->s_es->s_blocks_count) - 1;
else
last_block = first_block +
(EXT2_BLOCKS_PER_GROUP(sb) - 1);
if (le32_to_cpu(gdp->bg_block_bitmap) < first_block || if (le32_to_cpu(gdp->bg_block_bitmap) < first_block ||
le32_to_cpu(gdp->bg_block_bitmap) > last_block) le32_to_cpu(gdp->bg_block_bitmap) > last_block)
...@@ -806,7 +800,6 @@ static unsigned long descriptor_loc(struct super_block *sb, ...@@ -806,7 +800,6 @@ static unsigned long descriptor_loc(struct super_block *sb,
{ {
struct ext2_sb_info *sbi = EXT2_SB(sb); struct ext2_sb_info *sbi = EXT2_SB(sb);
unsigned long bg, first_meta_bg; unsigned long bg, first_meta_bg;
int has_super = 0;
first_meta_bg = le32_to_cpu(sbi->s_es->s_first_meta_bg); first_meta_bg = le32_to_cpu(sbi->s_es->s_first_meta_bg);
...@@ -814,10 +807,8 @@ static unsigned long descriptor_loc(struct super_block *sb, ...@@ -814,10 +807,8 @@ static unsigned long descriptor_loc(struct super_block *sb,
nr < first_meta_bg) nr < first_meta_bg)
return (logic_sb_block + nr + 1); return (logic_sb_block + nr + 1);
bg = sbi->s_desc_per_block * nr; bg = sbi->s_desc_per_block * nr;
if (ext2_bg_has_super(sb, bg))
has_super = 1;
return ext2_group_first_block_no(sb, bg) + has_super; return ext2_group_first_block_no(sb, bg) + ext2_bg_has_super(sb, bg);
} }
static int ext2_fill_super(struct super_block *sb, void *data, int silent) static int ext2_fill_super(struct super_block *sb, void *data, int silent)
......
...@@ -5834,7 +5834,7 @@ static int ext4_quota_enable(struct super_block *sb, int type, int format_id, ...@@ -5834,7 +5834,7 @@ static int ext4_quota_enable(struct super_block *sb, int type, int format_id,
/* Don't account quota for quota files to avoid recursion */ /* Don't account quota for quota files to avoid recursion */
qf_inode->i_flags |= S_NOQUOTA; qf_inode->i_flags |= S_NOQUOTA;
lockdep_set_quota_inode(qf_inode, I_DATA_SEM_QUOTA); lockdep_set_quota_inode(qf_inode, I_DATA_SEM_QUOTA);
err = dquot_enable(qf_inode, type, format_id, flags); err = dquot_load_quota_inode(qf_inode, type, format_id, flags);
if (err) if (err)
lockdep_set_quota_inode(qf_inode, I_DATA_SEM_NORMAL); lockdep_set_quota_inode(qf_inode, I_DATA_SEM_NORMAL);
iput(qf_inode); iput(qf_inode);
......
...@@ -1940,7 +1940,7 @@ static int f2fs_quota_enable(struct super_block *sb, int type, int format_id, ...@@ -1940,7 +1940,7 @@ static int f2fs_quota_enable(struct super_block *sb, int type, int format_id,
/* Don't account quota for quota files to avoid recursion */ /* Don't account quota for quota files to avoid recursion */
qf_inode->i_flags |= S_NOQUOTA; qf_inode->i_flags |= S_NOQUOTA;
err = dquot_enable(qf_inode, type, format_id, flags); err = dquot_load_quota_inode(qf_inode, type, format_id, flags);
iput(qf_inode); iput(qf_inode);
return err; return err;
} }
......
...@@ -728,7 +728,7 @@ static int ocfs2_release_dquot(struct dquot *dquot) ...@@ -728,7 +728,7 @@ static int ocfs2_release_dquot(struct dquot *dquot)
mutex_lock(&dquot->dq_lock); mutex_lock(&dquot->dq_lock);
/* Check whether we are not racing with some other dqget() */ /* Check whether we are not racing with some other dqget() */
if (atomic_read(&dquot->dq_count) > 1) if (dquot_is_busy(dquot))
goto out; goto out;
/* Running from downconvert thread? Postpone quota processing to wq */ /* Running from downconvert thread? Postpone quota processing to wq */
if (current == osb->dc_task) { if (current == osb->dc_task) {
......
...@@ -926,8 +926,8 @@ static int ocfs2_enable_quotas(struct ocfs2_super *osb) ...@@ -926,8 +926,8 @@ static int ocfs2_enable_quotas(struct ocfs2_super *osb)
status = -ENOENT; status = -ENOENT;
goto out_quota_off; goto out_quota_off;
} }
status = dquot_enable(inode[type], type, QFMT_OCFS2, status = dquot_load_quota_inode(inode[type], type, QFMT_OCFS2,
DQUOT_USAGE_ENABLED); DQUOT_USAGE_ENABLED);
if (status < 0) if (status < 0)
goto out_quota_off; goto out_quota_off;
} }
......
This diff is collapsed.
...@@ -60,8 +60,6 @@ static int quota_sync_all(int type) ...@@ -60,8 +60,6 @@ static int quota_sync_all(int type)
{ {
int ret; int ret;
if (type >= MAXQUOTAS)
return -EINVAL;
ret = security_quotactl(Q_SYNC, type, 0, NULL); ret = security_quotactl(Q_SYNC, type, 0, NULL);
if (!ret) if (!ret)
iterate_supers(quota_sync_one, &type); iterate_supers(quota_sync_one, &type);
...@@ -686,8 +684,6 @@ static int do_quotactl(struct super_block *sb, int type, int cmd, qid_t id, ...@@ -686,8 +684,6 @@ static int do_quotactl(struct super_block *sb, int type, int cmd, qid_t id,
{ {
int ret; int ret;
if (type >= MAXQUOTAS)
return -EINVAL;
type = array_index_nospec(type, MAXQUOTAS); type = array_index_nospec(type, MAXQUOTAS);
/* /*
* Quota not supported on this fs? Check this before s_quota_types * Quota not supported on this fs? Check this before s_quota_types
...@@ -831,6 +827,9 @@ int kernel_quotactl(unsigned int cmd, const char __user *special, ...@@ -831,6 +827,9 @@ int kernel_quotactl(unsigned int cmd, const char __user *special,
cmds = cmd >> SUBCMDSHIFT; cmds = cmd >> SUBCMDSHIFT;
type = cmd & SUBCMDMASK; type = cmd & SUBCMDMASK;
if (type >= MAXQUOTAS)
return -EINVAL;
/* /*
* As a special case Q_SYNC can be called without a specific device. * As a special case Q_SYNC can be called without a specific device.
* It will iterate all superblocks that have quota enabled and call * It will iterate all superblocks that have quota enabled and call
......
...@@ -217,7 +217,6 @@ static const struct quota_format_ops v1_format_ops = { ...@@ -217,7 +217,6 @@ static const struct quota_format_ops v1_format_ops = {
.check_quota_file = v1_check_quota_file, .check_quota_file = v1_check_quota_file,
.read_file_info = v1_read_file_info, .read_file_info = v1_read_file_info,
.write_file_info = v1_write_file_info, .write_file_info = v1_write_file_info,
.free_file_info = NULL,
.read_dqblk = v1_read_dqblk, .read_dqblk = v1_read_dqblk,
.commit_dqblk = v1_commit_dqblk, .commit_dqblk = v1_commit_dqblk,
}; };
......
...@@ -38,16 +38,10 @@ static int reiserfs_file_release(struct inode *inode, struct file *filp) ...@@ -38,16 +38,10 @@ static int reiserfs_file_release(struct inode *inode, struct file *filp)
BUG_ON(!S_ISREG(inode->i_mode)); BUG_ON(!S_ISREG(inode->i_mode));
if (atomic_add_unless(&REISERFS_I(inode)->openers, -1, 1)) if (!atomic_dec_and_mutex_lock(&REISERFS_I(inode)->openers,
&REISERFS_I(inode)->tailpack))
return 0; return 0;
mutex_lock(&REISERFS_I(inode)->tailpack);
if (!atomic_dec_and_test(&REISERFS_I(inode)->openers)) {
mutex_unlock(&REISERFS_I(inode)->tailpack);
return 0;
}
/* fast out for when nothing needs to be done */ /* fast out for when nothing needs to be done */
if ((!(REISERFS_I(inode)->i_flags & i_pack_on_close_mask) || if ((!(REISERFS_I(inode)->i_flags & i_pack_on_close_mask) ||
!tail_has_to_be_packed(inode)) && !tail_has_to_be_packed(inode)) &&
......
...@@ -2097,6 +2097,15 @@ int reiserfs_new_inode(struct reiserfs_transaction_handle *th, ...@@ -2097,6 +2097,15 @@ int reiserfs_new_inode(struct reiserfs_transaction_handle *th,
goto out_inserted_sd; goto out_inserted_sd;
} }
/*
* Mark it private if we're creating the privroot
* or something under it.
*/
if (IS_PRIVATE(dir) || dentry == REISERFS_SB(sb)->priv_root) {
inode->i_flags |= S_PRIVATE;
inode->i_opflags &= ~IOP_XATTR;
}
if (reiserfs_posixacl(inode->i_sb)) { if (reiserfs_posixacl(inode->i_sb)) {
reiserfs_write_unlock(inode->i_sb); reiserfs_write_unlock(inode->i_sb);
retval = reiserfs_inherit_default_acl(th, dir, dentry, inode); retval = reiserfs_inherit_default_acl(th, dir, dentry, inode);
...@@ -2111,8 +2120,7 @@ int reiserfs_new_inode(struct reiserfs_transaction_handle *th, ...@@ -2111,8 +2120,7 @@ int reiserfs_new_inode(struct reiserfs_transaction_handle *th,
reiserfs_warning(inode->i_sb, "jdm-13090", reiserfs_warning(inode->i_sb, "jdm-13090",
"ACLs aren't enabled in the fs, " "ACLs aren't enabled in the fs, "
"but vfs thinks they are!"); "but vfs thinks they are!");
} else if (IS_PRIVATE(dir)) }
inode->i_flags |= S_PRIVATE;
if (security->name) { if (security->name) {
reiserfs_write_unlock(inode->i_sb); reiserfs_write_unlock(inode->i_sb);
......
...@@ -377,10 +377,13 @@ static struct dentry *reiserfs_lookup(struct inode *dir, struct dentry *dentry, ...@@ -377,10 +377,13 @@ static struct dentry *reiserfs_lookup(struct inode *dir, struct dentry *dentry,
/* /*
* Propagate the private flag so we know we're * Propagate the private flag so we know we're
* in the priv tree * in the priv tree. Also clear IOP_XATTR
* since we don't have xattrs on xattr files.
*/ */
if (IS_PRIVATE(dir)) if (IS_PRIVATE(dir)) {
inode->i_flags |= S_PRIVATE; inode->i_flags |= S_PRIVATE;
inode->i_opflags &= ~IOP_XATTR;
}
} }
reiserfs_write_unlock(dir->i_sb); reiserfs_write_unlock(dir->i_sb);
if (retval == IO_ERROR) { if (retval == IO_ERROR) {
......
...@@ -1168,6 +1168,8 @@ static inline int bmap_would_wrap(unsigned bmap_nr) ...@@ -1168,6 +1168,8 @@ static inline int bmap_would_wrap(unsigned bmap_nr)
return bmap_nr > ((1LL << 16) - 1); return bmap_nr > ((1LL << 16) - 1);
} }
extern const struct xattr_handler *reiserfs_xattr_handlers[];
/* /*
* this says about version of key of all items (but stat data) the * this says about version of key of all items (but stat data) the
* object consists of * object consists of
......
...@@ -2049,6 +2049,8 @@ static int reiserfs_fill_super(struct super_block *s, void *data, int silent) ...@@ -2049,6 +2049,8 @@ static int reiserfs_fill_super(struct super_block *s, void *data, int silent)
if (replay_only(s)) if (replay_only(s))
goto error_unlocked; goto error_unlocked;
s->s_xattr = reiserfs_xattr_handlers;
if (bdev_read_only(s->s_bdev) && !sb_rdonly(s)) { if (bdev_read_only(s->s_bdev) && !sb_rdonly(s)) {
SWARN(silent, s, "clm-7000", SWARN(silent, s, "clm-7000",
"Detected readonly device, marking FS readonly"); "Detected readonly device, marking FS readonly");
......
...@@ -122,13 +122,13 @@ static struct dentry *open_xa_root(struct super_block *sb, int flags) ...@@ -122,13 +122,13 @@ static struct dentry *open_xa_root(struct super_block *sb, int flags)
struct dentry *xaroot; struct dentry *xaroot;
if (d_really_is_negative(privroot)) if (d_really_is_negative(privroot))
return ERR_PTR(-ENODATA); return ERR_PTR(-EOPNOTSUPP);
inode_lock_nested(d_inode(privroot), I_MUTEX_XATTR); inode_lock_nested(d_inode(privroot), I_MUTEX_XATTR);
xaroot = dget(REISERFS_SB(sb)->xattr_root); xaroot = dget(REISERFS_SB(sb)->xattr_root);
if (!xaroot) if (!xaroot)
xaroot = ERR_PTR(-ENODATA); xaroot = ERR_PTR(-EOPNOTSUPP);
else if (d_really_is_negative(xaroot)) { else if (d_really_is_negative(xaroot)) {
int err = -ENODATA; int err = -ENODATA;
...@@ -619,6 +619,10 @@ int reiserfs_xattr_set(struct inode *inode, const char *name, ...@@ -619,6 +619,10 @@ int reiserfs_xattr_set(struct inode *inode, const char *name,
int error, error2; int error, error2;
size_t jbegin_count = reiserfs_xattr_nblocks(inode, buffer_size); size_t jbegin_count = reiserfs_xattr_nblocks(inode, buffer_size);
/* Check before we start a transaction and then do nothing. */
if (!d_really_is_positive(REISERFS_SB(inode->i_sb)->priv_root))
return -EOPNOTSUPP;
if (!(flags & XATTR_REPLACE)) if (!(flags & XATTR_REPLACE))
jbegin_count += reiserfs_xattr_jcreate_nblocks(inode); jbegin_count += reiserfs_xattr_jcreate_nblocks(inode);
...@@ -841,8 +845,7 @@ ssize_t reiserfs_listxattr(struct dentry * dentry, char *buffer, size_t size) ...@@ -841,8 +845,7 @@ ssize_t reiserfs_listxattr(struct dentry * dentry, char *buffer, size_t size)
if (d_really_is_negative(dentry)) if (d_really_is_negative(dentry))
return -EINVAL; return -EINVAL;
if (!dentry->d_sb->s_xattr || if (get_inode_sd_version(d_inode(dentry)) == STAT_DATA_V1)
get_inode_sd_version(d_inode(dentry)) == STAT_DATA_V1)
return -EOPNOTSUPP; return -EOPNOTSUPP;
dir = open_xa_dir(d_inode(dentry), XATTR_REPLACE); dir = open_xa_dir(d_inode(dentry), XATTR_REPLACE);
...@@ -882,6 +885,7 @@ static int create_privroot(struct dentry *dentry) ...@@ -882,6 +885,7 @@ static int create_privroot(struct dentry *dentry)
} }
d_inode(dentry)->i_flags |= S_PRIVATE; d_inode(dentry)->i_flags |= S_PRIVATE;
d_inode(dentry)->i_opflags &= ~IOP_XATTR;
reiserfs_info(dentry->d_sb, "Created %s - reserved for xattr " reiserfs_info(dentry->d_sb, "Created %s - reserved for xattr "
"storage.\n", PRIVROOT_NAME); "storage.\n", PRIVROOT_NAME);
...@@ -895,7 +899,7 @@ static int create_privroot(struct dentry *dentry) { return 0; } ...@@ -895,7 +899,7 @@ static int create_privroot(struct dentry *dentry) { return 0; }
#endif #endif
/* Actual operations that are exported to VFS-land */ /* Actual operations that are exported to VFS-land */
static const struct xattr_handler *reiserfs_xattr_handlers[] = { const struct xattr_handler *reiserfs_xattr_handlers[] = {
#ifdef CONFIG_REISERFS_FS_XATTR #ifdef CONFIG_REISERFS_FS_XATTR
&reiserfs_xattr_user_handler, &reiserfs_xattr_user_handler,
&reiserfs_xattr_trusted_handler, &reiserfs_xattr_trusted_handler,
...@@ -966,8 +970,10 @@ int reiserfs_lookup_privroot(struct super_block *s) ...@@ -966,8 +970,10 @@ int reiserfs_lookup_privroot(struct super_block *s)
if (!IS_ERR(dentry)) { if (!IS_ERR(dentry)) {
REISERFS_SB(s)->priv_root = dentry; REISERFS_SB(s)->priv_root = dentry;
d_set_d_op(dentry, &xattr_lookup_poison_ops); d_set_d_op(dentry, &xattr_lookup_poison_ops);
if (d_really_is_positive(dentry)) if (d_really_is_positive(dentry)) {
d_inode(dentry)->i_flags |= S_PRIVATE; d_inode(dentry)->i_flags |= S_PRIVATE;
d_inode(dentry)->i_opflags &= ~IOP_XATTR;
}
} else } else
err = PTR_ERR(dentry); err = PTR_ERR(dentry);
inode_unlock(d_inode(s->s_root)); inode_unlock(d_inode(s->s_root));
...@@ -996,7 +1002,6 @@ int reiserfs_xattr_init(struct super_block *s, int mount_flags) ...@@ -996,7 +1002,6 @@ int reiserfs_xattr_init(struct super_block *s, int mount_flags)
} }
if (d_really_is_positive(privroot)) { if (d_really_is_positive(privroot)) {
s->s_xattr = reiserfs_xattr_handlers;
inode_lock(d_inode(privroot)); inode_lock(d_inode(privroot));
if (!REISERFS_SB(s)->xattr_root) { if (!REISERFS_SB(s)->xattr_root) {
struct dentry *dentry; struct dentry *dentry;
......
...@@ -320,10 +320,8 @@ reiserfs_inherit_default_acl(struct reiserfs_transaction_handle *th, ...@@ -320,10 +320,8 @@ reiserfs_inherit_default_acl(struct reiserfs_transaction_handle *th,
* would be useless since permissions are ignored, and a pain because * would be useless since permissions are ignored, and a pain because
* it introduces locking cycles * it introduces locking cycles
*/ */
if (IS_PRIVATE(dir)) { if (IS_PRIVATE(inode))
inode->i_flags |= S_PRIVATE;
goto apply_umask; goto apply_umask;
}
err = posix_acl_create(dir, &inode->i_mode, &default_acl, &acl); err = posix_acl_create(dir, &inode->i_mode, &default_acl, &acl);
if (err) if (err)
......
...@@ -263,7 +263,7 @@ enum { ...@@ -263,7 +263,7 @@ enum {
}; };
struct dqstats { struct dqstats {
int stat[_DQST_DQSTAT_LAST]; unsigned long stat[_DQST_DQSTAT_LAST];
struct percpu_counter counter[_DQST_DQSTAT_LAST]; struct percpu_counter counter[_DQST_DQSTAT_LAST];
}; };
......
...@@ -54,6 +54,16 @@ static inline struct dquot *dqgrab(struct dquot *dquot) ...@@ -54,6 +54,16 @@ static inline struct dquot *dqgrab(struct dquot *dquot)
atomic_inc(&dquot->dq_count); atomic_inc(&dquot->dq_count);
return dquot; return dquot;
} }
static inline bool dquot_is_busy(struct dquot *dquot)
{
if (test_bit(DQ_MOD_B, &dquot->dq_flags))
return true;
if (atomic_read(&dquot->dq_count) > 1)
return true;
return false;
}
void dqput(struct dquot *dquot); void dqput(struct dquot *dquot);
int dquot_scan_active(struct super_block *sb, int dquot_scan_active(struct super_block *sb,
int (*fn)(struct dquot *dquot, unsigned long priv), int (*fn)(struct dquot *dquot, unsigned long priv),
...@@ -87,7 +97,9 @@ int dquot_mark_dquot_dirty(struct dquot *dquot); ...@@ -87,7 +97,9 @@ int dquot_mark_dquot_dirty(struct dquot *dquot);
int dquot_file_open(struct inode *inode, struct file *file); int dquot_file_open(struct inode *inode, struct file *file);
int dquot_enable(struct inode *inode, int type, int format_id, int dquot_load_quota_sb(struct super_block *sb, int type, int format_id,
unsigned int flags);
int dquot_load_quota_inode(struct inode *inode, int type, int format_id,
unsigned int flags); unsigned int flags);
int dquot_quota_on(struct super_block *sb, int type, int format_id, int dquot_quota_on(struct super_block *sb, int type, int format_id,
const struct path *path); const struct path *path);
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment