Commit a1bb8505 authored by Dave Chinner's avatar Dave Chinner Committed by Darrick J. Wong

xfs: Fix a CIL UAF by getting get rid of the iclog callback lock

The iclog callback chain has it's own lock. That was added way back
in 2008 by myself to alleviate severe lock contention on the
icloglock in commit 114d23aa ("[XFS] Per iclog callback chain
lock"). This was long before delayed logging took the icloglock out
of the hot transaction commit path and removed all contention on it.
Hence the separate ic_callback_lock doesn't serve any scalability
purpose anymore, and hasn't for close on a decade.

Further, we only attach callbacks to iclogs in one place where we
are already taking the icloglock soon after attaching the callbacks.
We also have to drop the icloglock to run callbacks and grab it
immediately afterwards again. So given that the icloglock is no
longer hot, making it cover callbacks again doesn't really change
the locking patterns very much at all.

We also need to extend the icloglock to cover callback addition to
fix a zero-day UAF in the CIL push code. This occurs when shutdown
races with xlog_cil_push_work() and the shutdown runs the callbacks
before the push releases the iclog. This results in the CIL context
structure attached to the iclog being freed by the callback before
the CIL push has finished referencing it, leading to UAF bugs.

Hence, to avoid this UAF, we need the callback attachment to be
atomic with post processing of the commit iclog and references to
the structures being attached to the iclog. This requires holding
the icloglock as that's the only way to serialise iclog state
against a shutdown in progress.

The result is we need to be using the icloglock to protect the
callback list addition and removal and serialise them with shutdown.
That makes the ic_callback_lock redundant and so it can be removed.

Fixes: 71e330b5 ("xfs: Introduce delayed logging core code")
Signed-off-by: default avatarDave Chinner <dchinner@redhat.com>
Reviewed-by: default avatarBrian Foster <bfoster@redhat.com>
Reviewed-by: default avatarDarrick J. Wong <djwong@kernel.org>
Signed-off-by: default avatarDarrick J. Wong <djwong@kernel.org>
parent b6903358
...@@ -1484,7 +1484,6 @@ xlog_alloc_log( ...@@ -1484,7 +1484,6 @@ xlog_alloc_log(
iclog->ic_state = XLOG_STATE_ACTIVE; iclog->ic_state = XLOG_STATE_ACTIVE;
iclog->ic_log = log; iclog->ic_log = log;
atomic_set(&iclog->ic_refcnt, 0); atomic_set(&iclog->ic_refcnt, 0);
spin_lock_init(&iclog->ic_callback_lock);
INIT_LIST_HEAD(&iclog->ic_callbacks); INIT_LIST_HEAD(&iclog->ic_callbacks);
iclog->ic_datap = (char *)iclog->ic_data + log->l_iclog_hsize; iclog->ic_datap = (char *)iclog->ic_data + log->l_iclog_hsize;
...@@ -2760,32 +2759,6 @@ xlog_state_iodone_process_iclog( ...@@ -2760,32 +2759,6 @@ xlog_state_iodone_process_iclog(
} }
} }
/*
* Keep processing entries in the iclog callback list until we come around and
* it is empty. We need to atomically see that the list is empty and change the
* state to DIRTY so that we don't miss any more callbacks being added.
*
* This function is called with the icloglock held and returns with it held. We
* drop it while running callbacks, however, as holding it over thousands of
* callbacks is unnecessary and causes excessive contention if we do.
*/
static void
xlog_state_do_iclog_callbacks(
struct xlog *log,
struct xlog_in_core *iclog)
{
LIST_HEAD(tmp);
trace_xlog_iclog_callbacks_start(iclog, _RET_IP_);
spin_lock(&iclog->ic_callback_lock);
list_splice_init(&iclog->ic_callbacks, &tmp);
spin_unlock(&iclog->ic_callback_lock);
xlog_cil_process_committed(&tmp);
trace_xlog_iclog_callbacks_done(iclog, _RET_IP_);
}
STATIC void STATIC void
xlog_state_do_callback( xlog_state_do_callback(
struct xlog *log) struct xlog *log)
...@@ -2814,6 +2787,8 @@ xlog_state_do_callback( ...@@ -2814,6 +2787,8 @@ xlog_state_do_callback(
repeats++; repeats++;
do { do {
LIST_HEAD(cb_list);
if (xlog_state_iodone_process_iclog(log, iclog, if (xlog_state_iodone_process_iclog(log, iclog,
&ioerror)) &ioerror))
break; break;
...@@ -2823,9 +2798,12 @@ xlog_state_do_callback( ...@@ -2823,9 +2798,12 @@ xlog_state_do_callback(
iclog = iclog->ic_next; iclog = iclog->ic_next;
continue; continue;
} }
list_splice_init(&iclog->ic_callbacks, &cb_list);
spin_unlock(&log->l_icloglock); spin_unlock(&log->l_icloglock);
xlog_state_do_iclog_callbacks(log, iclog); trace_xlog_iclog_callbacks_start(iclog, _RET_IP_);
xlog_cil_process_committed(&cb_list);
trace_xlog_iclog_callbacks_done(iclog, _RET_IP_);
cycled_icloglock = true; cycled_icloglock = true;
spin_lock(&log->l_icloglock); spin_lock(&log->l_icloglock);
......
...@@ -873,15 +873,21 @@ xlog_cil_push_work( ...@@ -873,15 +873,21 @@ xlog_cil_push_work(
xfs_log_ticket_ungrant(log, tic); xfs_log_ticket_ungrant(log, tic);
spin_lock(&commit_iclog->ic_callback_lock); /*
* Once we attach the ctx to the iclog, a shutdown can process the
* iclog, run the callbacks and free the ctx. The only thing preventing
* this potential UAF situation here is that we are holding the
* icloglock. Hence we cannot access the ctx after we have attached the
* callbacks and dropped the icloglock.
*/
spin_lock(&log->l_icloglock);
if (commit_iclog->ic_state == XLOG_STATE_IOERROR) { if (commit_iclog->ic_state == XLOG_STATE_IOERROR) {
spin_unlock(&commit_iclog->ic_callback_lock); spin_unlock(&log->l_icloglock);
goto out_abort; goto out_abort;
} }
ASSERT_ALWAYS(commit_iclog->ic_state == XLOG_STATE_ACTIVE || ASSERT_ALWAYS(commit_iclog->ic_state == XLOG_STATE_ACTIVE ||
commit_iclog->ic_state == XLOG_STATE_WANT_SYNC); commit_iclog->ic_state == XLOG_STATE_WANT_SYNC);
list_add_tail(&ctx->iclog_entry, &commit_iclog->ic_callbacks); list_add_tail(&ctx->iclog_entry, &commit_iclog->ic_callbacks);
spin_unlock(&commit_iclog->ic_callback_lock);
/* /*
* now the checkpoint commit is complete and we've attached the * now the checkpoint commit is complete and we've attached the
...@@ -898,8 +904,10 @@ xlog_cil_push_work( ...@@ -898,8 +904,10 @@ xlog_cil_push_work(
* iclogs to complete before we submit the commit_iclog. In this case, * iclogs to complete before we submit the commit_iclog. In this case,
* the commit_iclog write needs to issue a pre-flush so that the * the commit_iclog write needs to issue a pre-flush so that the
* ordering is correctly preserved down to stable storage. * ordering is correctly preserved down to stable storage.
*
* NOTE: It is not safe to reference the ctx after this check as we drop
* the icloglock if we have to wait for completion of other iclogs.
*/ */
spin_lock(&log->l_icloglock);
if (ctx->start_lsn != commit_lsn) { if (ctx->start_lsn != commit_lsn) {
xlog_wait_on_iclog(commit_iclog->ic_prev); xlog_wait_on_iclog(commit_iclog->ic_prev);
spin_lock(&log->l_icloglock); spin_lock(&log->l_icloglock);
......
...@@ -216,9 +216,6 @@ typedef struct xlog_in_core { ...@@ -216,9 +216,6 @@ typedef struct xlog_in_core {
enum xlog_iclog_state ic_state; enum xlog_iclog_state ic_state;
unsigned int ic_flags; unsigned int ic_flags;
char *ic_datap; /* pointer to iclog data */ char *ic_datap; /* pointer to iclog data */
/* Callback structures need their own cacheline */
spinlock_t ic_callback_lock ____cacheline_aligned_in_smp;
struct list_head ic_callbacks; struct list_head ic_callbacks;
/* reference counts need their own cacheline */ /* reference counts need their own cacheline */
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment