Commit fe7257fd authored by Christoph Hellwig's avatar Christoph Hellwig Committed by Ben Myers

xfs: do not write the buffer from xfs_qm_dqflush

Instead of writing the buffer directly from inside xfs_qm_dqflush return it
to the caller and let the caller decide what to do with the buffer.  Also
remove the pincount check in xfs_qm_dqflush that all non-blocking callers
already implement and the now unused flags parameter and the XFS_DQ_IS_DIRTY
check that all callers already perform.

[ Dave Chinner: fixed build error cause by missing '{'. ]
Signed-off-by: default avatarChristoph Hellwig <hch@lst.de>
Reviewed-by: default avatarDave Chinner <dchinner@redhat.com>
Reviewed-by: default avatarMark Tinguely <tinguely@sgi.com>
Signed-off-by: default avatarBen Myers <bpm@sgi.com>
parent 4c46819a
...@@ -878,8 +878,8 @@ xfs_qm_dqflush_done( ...@@ -878,8 +878,8 @@ xfs_qm_dqflush_done(
*/ */
int int
xfs_qm_dqflush( xfs_qm_dqflush(
xfs_dquot_t *dqp, struct xfs_dquot *dqp,
uint flags) struct xfs_buf **bpp)
{ {
struct xfs_mount *mp = dqp->q_mount; struct xfs_mount *mp = dqp->q_mount;
struct xfs_buf *bp; struct xfs_buf *bp;
...@@ -891,14 +891,8 @@ xfs_qm_dqflush( ...@@ -891,14 +891,8 @@ xfs_qm_dqflush(
trace_xfs_dqflush(dqp); trace_xfs_dqflush(dqp);
/* *bpp = NULL;
* If not dirty, or it's pinned and we are not supposed to block, nada.
*/
if (!XFS_DQ_IS_DIRTY(dqp) ||
((flags & SYNC_TRYLOCK) && atomic_read(&dqp->q_pincount) > 0)) {
xfs_dqfunlock(dqp);
return 0;
}
xfs_qm_dqunpin_wait(dqp); xfs_qm_dqunpin_wait(dqp);
/* /*
...@@ -918,9 +912,8 @@ xfs_qm_dqflush( ...@@ -918,9 +912,8 @@ xfs_qm_dqflush(
xfs_trans_ail_delete(mp->m_ail, lip); xfs_trans_ail_delete(mp->m_ail, lip);
else else
spin_unlock(&mp->m_ail->xa_lock); spin_unlock(&mp->m_ail->xa_lock);
error = XFS_ERROR(EIO);
xfs_dqfunlock(dqp); goto out_unlock;
return XFS_ERROR(EIO);
} }
/* /*
...@@ -928,11 +921,8 @@ xfs_qm_dqflush( ...@@ -928,11 +921,8 @@ xfs_qm_dqflush(
*/ */
error = xfs_trans_read_buf(mp, NULL, mp->m_ddev_targp, dqp->q_blkno, error = xfs_trans_read_buf(mp, NULL, mp->m_ddev_targp, dqp->q_blkno,
mp->m_quotainfo->qi_dqchunklen, 0, &bp); mp->m_quotainfo->qi_dqchunklen, 0, &bp);
if (error) { if (error)
ASSERT(error != ENOENT); goto out_unlock;
xfs_dqfunlock(dqp);
return error;
}
/* /*
* Calculate the location of the dquot inside the buffer. * Calculate the location of the dquot inside the buffer.
...@@ -978,20 +968,13 @@ xfs_qm_dqflush( ...@@ -978,20 +968,13 @@ xfs_qm_dqflush(
xfs_log_force(mp, 0); xfs_log_force(mp, 0);
} }
if (flags & SYNC_WAIT)
error = xfs_bwrite(bp);
else
xfs_buf_delwri_queue(bp);
xfs_buf_relse(bp);
trace_xfs_dqflush_done(dqp); trace_xfs_dqflush_done(dqp);
*bpp = bp;
return 0;
/* out_unlock:
* dqp is still locked, but caller is free to unlock it now. xfs_dqfunlock(dqp);
*/ return XFS_ERROR(EIO);
return error;
} }
/* /*
......
...@@ -141,7 +141,7 @@ static inline xfs_dquot_t *xfs_inode_dquot(struct xfs_inode *ip, int type) ...@@ -141,7 +141,7 @@ static inline xfs_dquot_t *xfs_inode_dquot(struct xfs_inode *ip, int type)
extern int xfs_qm_dqread(struct xfs_mount *, xfs_dqid_t, uint, extern int xfs_qm_dqread(struct xfs_mount *, xfs_dqid_t, uint,
uint, struct xfs_dquot **); uint, struct xfs_dquot **);
extern void xfs_qm_dqdestroy(xfs_dquot_t *); extern void xfs_qm_dqdestroy(xfs_dquot_t *);
extern int xfs_qm_dqflush(xfs_dquot_t *, uint); extern int xfs_qm_dqflush(struct xfs_dquot *, struct xfs_buf **);
extern void xfs_qm_dqunpin_wait(xfs_dquot_t *); extern void xfs_qm_dqunpin_wait(xfs_dquot_t *);
extern void xfs_qm_adjust_dqtimers(xfs_mount_t *, extern void xfs_qm_adjust_dqtimers(xfs_mount_t *,
xfs_disk_dquot_t *); xfs_disk_dquot_t *);
......
...@@ -119,10 +119,12 @@ xfs_qm_dquot_logitem_push( ...@@ -119,10 +119,12 @@ xfs_qm_dquot_logitem_push(
struct xfs_log_item *lip) struct xfs_log_item *lip)
{ {
struct xfs_dquot *dqp = DQUOT_ITEM(lip)->qli_dquot; struct xfs_dquot *dqp = DQUOT_ITEM(lip)->qli_dquot;
struct xfs_buf *bp = NULL;
int error; int error;
ASSERT(XFS_DQ_IS_LOCKED(dqp)); ASSERT(XFS_DQ_IS_LOCKED(dqp));
ASSERT(!completion_done(&dqp->q_flush)); ASSERT(!completion_done(&dqp->q_flush));
ASSERT(atomic_read(&dqp->q_pincount) == 0);
/* /*
* Since we were able to lock the dquot's flush lock and * Since we were able to lock the dquot's flush lock and
...@@ -133,10 +135,16 @@ xfs_qm_dquot_logitem_push( ...@@ -133,10 +135,16 @@ xfs_qm_dquot_logitem_push(
* lock without sleeping, then there must not have been * lock without sleeping, then there must not have been
* anyone in the process of flushing the dquot. * anyone in the process of flushing the dquot.
*/ */
error = xfs_qm_dqflush(dqp, SYNC_TRYLOCK); error = xfs_qm_dqflush(dqp, &bp);
if (error) if (error) {
xfs_warn(dqp->q_mount, "%s: push error %d on dqp %p", xfs_warn(dqp->q_mount, "%s: push error %d on dqp %p",
__func__, error, dqp); __func__, error, dqp);
goto out_unlock;
}
xfs_buf_delwri_queue(bp);
xfs_buf_relse(bp);
out_unlock:
xfs_dqunlock(dqp); xfs_dqunlock(dqp);
} }
...@@ -239,6 +247,15 @@ xfs_qm_dquot_logitem_trylock( ...@@ -239,6 +247,15 @@ xfs_qm_dquot_logitem_trylock(
if (!xfs_dqlock_nowait(dqp)) if (!xfs_dqlock_nowait(dqp))
return XFS_ITEM_LOCKED; return XFS_ITEM_LOCKED;
/*
* Re-check the pincount now that we stabilized the value by
* taking the quota lock.
*/
if (atomic_read(&dqp->q_pincount) > 0) {
xfs_dqunlock(dqp);
return XFS_ITEM_PINNED;
}
if (!xfs_dqflock_nowait(dqp)) { if (!xfs_dqflock_nowait(dqp)) {
/* /*
* dquot has already been flushed to the backing buffer, * dquot has already been flushed to the backing buffer,
......
...@@ -175,16 +175,21 @@ xfs_qm_dqpurge( ...@@ -175,16 +175,21 @@ xfs_qm_dqpurge(
* we're unmounting, we do care, so we flush it and wait. * we're unmounting, we do care, so we flush it and wait.
*/ */
if (XFS_DQ_IS_DIRTY(dqp)) { if (XFS_DQ_IS_DIRTY(dqp)) {
int error; struct xfs_buf *bp = NULL;
int error;
/* /*
* We don't care about getting disk errors here. We need * We don't care about getting disk errors here. We need
* to purge this dquot anyway, so we go ahead regardless. * to purge this dquot anyway, so we go ahead regardless.
*/ */
error = xfs_qm_dqflush(dqp, SYNC_WAIT); error = xfs_qm_dqflush(dqp, &bp);
if (error) if (error) {
xfs_warn(mp, "%s: dquot %p flush failed", xfs_warn(mp, "%s: dquot %p flush failed",
__func__, dqp); __func__, dqp);
} else {
error = xfs_bwrite(bp);
xfs_buf_relse(bp);
}
xfs_dqflock(dqp); xfs_dqflock(dqp);
} }
...@@ -1200,6 +1205,7 @@ STATIC int ...@@ -1200,6 +1205,7 @@ STATIC int
xfs_qm_flush_one( xfs_qm_flush_one(
struct xfs_dquot *dqp) struct xfs_dquot *dqp)
{ {
struct xfs_buf *bp = NULL;
int error = 0; int error = 0;
xfs_dqlock(dqp); xfs_dqlock(dqp);
...@@ -1211,8 +1217,12 @@ xfs_qm_flush_one( ...@@ -1211,8 +1217,12 @@ xfs_qm_flush_one(
if (!xfs_dqflock_nowait(dqp)) if (!xfs_dqflock_nowait(dqp))
xfs_dqflock_pushbuf_wait(dqp); xfs_dqflock_pushbuf_wait(dqp);
error = xfs_qm_dqflush(dqp, 0); error = xfs_qm_dqflush(dqp, &bp);
if (error)
goto out_unlock;
xfs_buf_delwri_queue(bp);
xfs_buf_relse(bp);
out_unlock: out_unlock:
xfs_dqunlock(dqp); xfs_dqunlock(dqp);
return error; return error;
...@@ -1479,18 +1489,23 @@ xfs_qm_dqreclaim_one( ...@@ -1479,18 +1489,23 @@ xfs_qm_dqreclaim_one(
* dirty dquots. * dirty dquots.
*/ */
if (XFS_DQ_IS_DIRTY(dqp)) { if (XFS_DQ_IS_DIRTY(dqp)) {
struct xfs_buf *bp = NULL;
trace_xfs_dqreclaim_dirty(dqp); trace_xfs_dqreclaim_dirty(dqp);
/* /*
* We flush it delayed write, so don't bother releasing the * We flush it delayed write, so don't bother releasing the
* freelist lock. * freelist lock.
*/ */
error = xfs_qm_dqflush(dqp, 0); error = xfs_qm_dqflush(dqp, &bp);
if (error) { if (error) {
xfs_warn(mp, "%s: dquot %p flush failed", xfs_warn(mp, "%s: dquot %p flush failed",
__func__, dqp); __func__, dqp);
goto out_busy;
} }
xfs_buf_delwri_queue(bp);
xfs_buf_relse(bp);
/* /*
* Give the dquot another try on the freelist, as the * Give the dquot another try on the freelist, as the
* flushing will take some time. * flushing will take some time.
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment