Re: [PATCH v5 01/78] xfs: Rename xa_ elements to ail_
From: Darrick J. Wong
Date: Tue Jan 02 2018 - 20:02:00 EST
On Fri, Dec 15, 2017 at 02:03:33PM -0800, Matthew Wilcox wrote:
> From: Matthew Wilcox <mawilcox@xxxxxxxxxxxxx>
>
> This is a simple rename, except that xa_ail becomes ail_head.
>
> Signed-off-by: Matthew Wilcox <mawilcox@xxxxxxxxxxxxx>
That was an eyeful,
Reviewed-by: Darrick J. Wong <darrick.wong@xxxxxxxxxx>
> ---
> fs/xfs/xfs_buf_item.c | 10 ++--
> fs/xfs/xfs_dquot.c | 4 +-
> fs/xfs/xfs_dquot_item.c | 11 ++--
> fs/xfs/xfs_inode_item.c | 22 +++----
> fs/xfs/xfs_log.c | 6 +-
> fs/xfs/xfs_log_recover.c | 80 ++++++++++++-------------
> fs/xfs/xfs_trans.c | 18 +++---
> fs/xfs/xfs_trans_ail.c | 152 +++++++++++++++++++++++------------------------
> fs/xfs/xfs_trans_buf.c | 4 +-
> fs/xfs/xfs_trans_priv.h | 42 ++++++-------
> 10 files changed, 175 insertions(+), 174 deletions(-)
>
> diff --git a/fs/xfs/xfs_buf_item.c b/fs/xfs/xfs_buf_item.c
> index e0a0af0946f2..6c5035544a93 100644
> --- a/fs/xfs/xfs_buf_item.c
> +++ b/fs/xfs/xfs_buf_item.c
> @@ -459,7 +459,7 @@ xfs_buf_item_unpin(
> bp->b_fspriv = NULL;
> bp->b_iodone = NULL;
> } else {
> - spin_lock(&ailp->xa_lock);
> + spin_lock(&ailp->ail_lock);
> xfs_trans_ail_delete(ailp, lip, SHUTDOWN_LOG_IO_ERROR);
> xfs_buf_item_relse(bp);
> ASSERT(bp->b_fspriv == NULL);
> @@ -1056,13 +1056,13 @@ xfs_buf_do_callbacks_fail(
> struct xfs_log_item *lip = bp->b_fspriv;
> struct xfs_ail *ailp = lip->li_ailp;
>
> - spin_lock(&ailp->xa_lock);
> + spin_lock(&ailp->ail_lock);
> for (; lip; lip = next) {
> next = lip->li_bio_list;
> if (lip->li_ops->iop_error)
> lip->li_ops->iop_error(lip, bp);
> }
> - spin_unlock(&ailp->xa_lock);
> + spin_unlock(&ailp->ail_lock);
> }
>
> static bool
> @@ -1215,7 +1215,7 @@ xfs_buf_iodone(
> *
> * Either way, AIL is useless if we're forcing a shutdown.
> */
> - spin_lock(&ailp->xa_lock);
> + spin_lock(&ailp->ail_lock);
> xfs_trans_ail_delete(ailp, lip, SHUTDOWN_CORRUPT_INCORE);
> xfs_buf_item_free(BUF_ITEM(lip));
> }
> @@ -1236,7 +1236,7 @@ xfs_buf_resubmit_failed_buffers(
> /*
> * Clear XFS_LI_FAILED flag from all items before resubmit
> *
> - * XFS_LI_FAILED set/clear is protected by xa_lock, caller this
> + * XFS_LI_FAILED set/clear is protected by ail_lock, caller this
> * function already have it acquired
> */
> for (; lip; lip = next) {
> diff --git a/fs/xfs/xfs_dquot.c b/fs/xfs/xfs_dquot.c
> index f248708c10ff..e2a466df5dd1 100644
> --- a/fs/xfs/xfs_dquot.c
> +++ b/fs/xfs/xfs_dquot.c
> @@ -974,7 +974,7 @@ xfs_qm_dqflush_done(
> (lip->li_flags & XFS_LI_FAILED))) {
>
> /* xfs_trans_ail_delete() drops the AIL lock. */
> - spin_lock(&ailp->xa_lock);
> + spin_lock(&ailp->ail_lock);
> if (lip->li_lsn == qip->qli_flush_lsn) {
> xfs_trans_ail_delete(ailp, lip, SHUTDOWN_CORRUPT_INCORE);
> } else {
> @@ -984,7 +984,7 @@ xfs_qm_dqflush_done(
> */
> if (lip->li_flags & XFS_LI_FAILED)
> xfs_clear_li_failed(lip);
> - spin_unlock(&ailp->xa_lock);
> + spin_unlock(&ailp->ail_lock);
> }
> }
>
> diff --git a/fs/xfs/xfs_dquot_item.c b/fs/xfs/xfs_dquot_item.c
> index 664dea105e76..62637a226601 100644
> --- a/fs/xfs/xfs_dquot_item.c
> +++ b/fs/xfs/xfs_dquot_item.c
> @@ -160,8 +160,9 @@ xfs_dquot_item_error(
> STATIC uint
> xfs_qm_dquot_logitem_push(
> struct xfs_log_item *lip,
> - struct list_head *buffer_list) __releases(&lip->li_ailp->xa_lock)
> - __acquires(&lip->li_ailp->xa_lock)
> + struct list_head *buffer_list)
> + __releases(&lip->li_ailp->ail_lock)
> + __acquires(&lip->li_ailp->ail_lock)
> {
> struct xfs_dquot *dqp = DQUOT_ITEM(lip)->qli_dquot;
> struct xfs_buf *bp = lip->li_buf;
> @@ -208,7 +209,7 @@ xfs_qm_dquot_logitem_push(
> goto out_unlock;
> }
>
> - spin_unlock(&lip->li_ailp->xa_lock);
> + spin_unlock(&lip->li_ailp->ail_lock);
>
> error = xfs_qm_dqflush(dqp, &bp);
> if (error) {
> @@ -220,7 +221,7 @@ xfs_qm_dquot_logitem_push(
> xfs_buf_relse(bp);
> }
>
> - spin_lock(&lip->li_ailp->xa_lock);
> + spin_lock(&lip->li_ailp->ail_lock);
> out_unlock:
> xfs_dqunlock(dqp);
> return rval;
> @@ -403,7 +404,7 @@ xfs_qm_qoffend_logitem_committed(
> * Delete the qoff-start logitem from the AIL.
> * xfs_trans_ail_delete() drops the AIL lock.
> */
> - spin_lock(&ailp->xa_lock);
> + spin_lock(&ailp->ail_lock);
> xfs_trans_ail_delete(ailp, &qfs->qql_item, SHUTDOWN_LOG_IO_ERROR);
>
> kmem_free(qfs->qql_item.li_lv_shadow);
> diff --git a/fs/xfs/xfs_inode_item.c b/fs/xfs/xfs_inode_item.c
> index 6ee5c3bf19ad..071acd4249a0 100644
> --- a/fs/xfs/xfs_inode_item.c
> +++ b/fs/xfs/xfs_inode_item.c
> @@ -501,8 +501,8 @@ STATIC uint
> xfs_inode_item_push(
> struct xfs_log_item *lip,
> struct list_head *buffer_list)
> - __releases(&lip->li_ailp->xa_lock)
> - __acquires(&lip->li_ailp->xa_lock)
> + __releases(&lip->li_ailp->ail_lock)
> + __acquires(&lip->li_ailp->ail_lock)
> {
> struct xfs_inode_log_item *iip = INODE_ITEM(lip);
> struct xfs_inode *ip = iip->ili_inode;
> @@ -561,7 +561,7 @@ xfs_inode_item_push(
> ASSERT(iip->ili_fields != 0 || XFS_FORCED_SHUTDOWN(ip->i_mount));
> ASSERT(iip->ili_logged == 0 || XFS_FORCED_SHUTDOWN(ip->i_mount));
>
> - spin_unlock(&lip->li_ailp->xa_lock);
> + spin_unlock(&lip->li_ailp->ail_lock);
>
> error = xfs_iflush(ip, &bp);
> if (!error) {
> @@ -570,7 +570,7 @@ xfs_inode_item_push(
> xfs_buf_relse(bp);
> }
>
> - spin_lock(&lip->li_ailp->xa_lock);
> + spin_lock(&lip->li_ailp->ail_lock);
> out_unlock:
> xfs_iunlock(ip, XFS_ILOCK_SHARED);
> return rval;
> @@ -774,7 +774,7 @@ xfs_iflush_done(
> bool mlip_changed = false;
>
> /* this is an opencoded batch version of xfs_trans_ail_delete */
> - spin_lock(&ailp->xa_lock);
> + spin_lock(&ailp->ail_lock);
> for (blip = lip; blip; blip = blip->li_bio_list) {
> if (INODE_ITEM(blip)->ili_logged &&
> blip->li_lsn == INODE_ITEM(blip)->ili_flush_lsn)
> @@ -785,15 +785,15 @@ xfs_iflush_done(
> }
>
> if (mlip_changed) {
> - if (!XFS_FORCED_SHUTDOWN(ailp->xa_mount))
> - xlog_assign_tail_lsn_locked(ailp->xa_mount);
> - if (list_empty(&ailp->xa_ail))
> - wake_up_all(&ailp->xa_empty);
> + if (!XFS_FORCED_SHUTDOWN(ailp->ail_mount))
> + xlog_assign_tail_lsn_locked(ailp->ail_mount);
> + if (list_empty(&ailp->ail_head))
> + wake_up_all(&ailp->ail_empty);
> }
> - spin_unlock(&ailp->xa_lock);
> + spin_unlock(&ailp->ail_lock);
>
> if (mlip_changed)
> - xfs_log_space_wake(ailp->xa_mount);
> + xfs_log_space_wake(ailp->ail_mount);
> }
>
> /*
> diff --git a/fs/xfs/xfs_log.c b/fs/xfs/xfs_log.c
> index a503af96d780..7148625eebf2 100644
> --- a/fs/xfs/xfs_log.c
> +++ b/fs/xfs/xfs_log.c
> @@ -1148,7 +1148,7 @@ xlog_assign_tail_lsn_locked(
> struct xfs_log_item *lip;
> xfs_lsn_t tail_lsn;
>
> - assert_spin_locked(&mp->m_ail->xa_lock);
> + assert_spin_locked(&mp->m_ail->ail_lock);
>
> /*
> * To make sure we always have a valid LSN for the log tail we keep
> @@ -1171,9 +1171,9 @@ xlog_assign_tail_lsn(
> {
> xfs_lsn_t tail_lsn;
>
> - spin_lock(&mp->m_ail->xa_lock);
> + spin_lock(&mp->m_ail->ail_lock);
> tail_lsn = xlog_assign_tail_lsn_locked(mp);
> - spin_unlock(&mp->m_ail->xa_lock);
> + spin_unlock(&mp->m_ail->ail_lock);
>
> return tail_lsn;
> }
> diff --git a/fs/xfs/xfs_log_recover.c b/fs/xfs/xfs_log_recover.c
> index 28d1abfe835e..d871761626fb 100644
> --- a/fs/xfs/xfs_log_recover.c
> +++ b/fs/xfs/xfs_log_recover.c
> @@ -3424,7 +3424,7 @@ xlog_recover_efi_pass2(
> }
> atomic_set(&efip->efi_next_extent, efi_formatp->efi_nextents);
>
> - spin_lock(&log->l_ailp->xa_lock);
> + spin_lock(&log->l_ailp->ail_lock);
> /*
> * The EFI has two references. One for the EFD and one for EFI to ensure
> * it makes it into the AIL. Insert the EFI into the AIL directly and
> @@ -3467,7 +3467,7 @@ xlog_recover_efd_pass2(
> * Search for the EFI with the id in the EFD format structure in the
> * AIL.
> */
> - spin_lock(&ailp->xa_lock);
> + spin_lock(&ailp->ail_lock);
> lip = xfs_trans_ail_cursor_first(ailp, &cur, 0);
> while (lip != NULL) {
> if (lip->li_type == XFS_LI_EFI) {
> @@ -3477,9 +3477,9 @@ xlog_recover_efd_pass2(
> * Drop the EFD reference to the EFI. This
> * removes the EFI from the AIL and frees it.
> */
> - spin_unlock(&ailp->xa_lock);
> + spin_unlock(&ailp->ail_lock);
> xfs_efi_release(efip);
> - spin_lock(&ailp->xa_lock);
> + spin_lock(&ailp->ail_lock);
> break;
> }
> }
> @@ -3487,7 +3487,7 @@ xlog_recover_efd_pass2(
> }
>
> xfs_trans_ail_cursor_done(&cur);
> - spin_unlock(&ailp->xa_lock);
> + spin_unlock(&ailp->ail_lock);
>
> return 0;
> }
> @@ -3520,7 +3520,7 @@ xlog_recover_rui_pass2(
> }
> atomic_set(&ruip->rui_next_extent, rui_formatp->rui_nextents);
>
> - spin_lock(&log->l_ailp->xa_lock);
> + spin_lock(&log->l_ailp->ail_lock);
> /*
> * The RUI has two references. One for the RUD and one for RUI to ensure
> * it makes it into the AIL. Insert the RUI into the AIL directly and
> @@ -3560,7 +3560,7 @@ xlog_recover_rud_pass2(
> * Search for the RUI with the id in the RUD format structure in the
> * AIL.
> */
> - spin_lock(&ailp->xa_lock);
> + spin_lock(&ailp->ail_lock);
> lip = xfs_trans_ail_cursor_first(ailp, &cur, 0);
> while (lip != NULL) {
> if (lip->li_type == XFS_LI_RUI) {
> @@ -3570,9 +3570,9 @@ xlog_recover_rud_pass2(
> * Drop the RUD reference to the RUI. This
> * removes the RUI from the AIL and frees it.
> */
> - spin_unlock(&ailp->xa_lock);
> + spin_unlock(&ailp->ail_lock);
> xfs_rui_release(ruip);
> - spin_lock(&ailp->xa_lock);
> + spin_lock(&ailp->ail_lock);
> break;
> }
> }
> @@ -3580,7 +3580,7 @@ xlog_recover_rud_pass2(
> }
>
> xfs_trans_ail_cursor_done(&cur);
> - spin_unlock(&ailp->xa_lock);
> + spin_unlock(&ailp->ail_lock);
>
> return 0;
> }
> @@ -3636,7 +3636,7 @@ xlog_recover_cui_pass2(
> }
> atomic_set(&cuip->cui_next_extent, cui_formatp->cui_nextents);
>
> - spin_lock(&log->l_ailp->xa_lock);
> + spin_lock(&log->l_ailp->ail_lock);
> /*
> * The CUI has two references. One for the CUD and one for CUI to ensure
> * it makes it into the AIL. Insert the CUI into the AIL directly and
> @@ -3677,7 +3677,7 @@ xlog_recover_cud_pass2(
> * Search for the CUI with the id in the CUD format structure in the
> * AIL.
> */
> - spin_lock(&ailp->xa_lock);
> + spin_lock(&ailp->ail_lock);
> lip = xfs_trans_ail_cursor_first(ailp, &cur, 0);
> while (lip != NULL) {
> if (lip->li_type == XFS_LI_CUI) {
> @@ -3687,9 +3687,9 @@ xlog_recover_cud_pass2(
> * Drop the CUD reference to the CUI. This
> * removes the CUI from the AIL and frees it.
> */
> - spin_unlock(&ailp->xa_lock);
> + spin_unlock(&ailp->ail_lock);
> xfs_cui_release(cuip);
> - spin_lock(&ailp->xa_lock);
> + spin_lock(&ailp->ail_lock);
> break;
> }
> }
> @@ -3697,7 +3697,7 @@ xlog_recover_cud_pass2(
> }
>
> xfs_trans_ail_cursor_done(&cur);
> - spin_unlock(&ailp->xa_lock);
> + spin_unlock(&ailp->ail_lock);
>
> return 0;
> }
> @@ -3755,7 +3755,7 @@ xlog_recover_bui_pass2(
> }
> atomic_set(&buip->bui_next_extent, bui_formatp->bui_nextents);
>
> - spin_lock(&log->l_ailp->xa_lock);
> + spin_lock(&log->l_ailp->ail_lock);
> /*
> * The RUI has two references. One for the RUD and one for RUI to ensure
> * it makes it into the AIL. Insert the RUI into the AIL directly and
> @@ -3796,7 +3796,7 @@ xlog_recover_bud_pass2(
> * Search for the BUI with the id in the BUD format structure in the
> * AIL.
> */
> - spin_lock(&ailp->xa_lock);
> + spin_lock(&ailp->ail_lock);
> lip = xfs_trans_ail_cursor_first(ailp, &cur, 0);
> while (lip != NULL) {
> if (lip->li_type == XFS_LI_BUI) {
> @@ -3806,9 +3806,9 @@ xlog_recover_bud_pass2(
> * Drop the BUD reference to the BUI. This
> * removes the BUI from the AIL and frees it.
> */
> - spin_unlock(&ailp->xa_lock);
> + spin_unlock(&ailp->ail_lock);
> xfs_bui_release(buip);
> - spin_lock(&ailp->xa_lock);
> + spin_lock(&ailp->ail_lock);
> break;
> }
> }
> @@ -3816,7 +3816,7 @@ xlog_recover_bud_pass2(
> }
>
> xfs_trans_ail_cursor_done(&cur);
> - spin_unlock(&ailp->xa_lock);
> + spin_unlock(&ailp->ail_lock);
>
> return 0;
> }
> @@ -4649,9 +4649,9 @@ xlog_recover_process_efi(
> if (test_bit(XFS_EFI_RECOVERED, &efip->efi_flags))
> return 0;
>
> - spin_unlock(&ailp->xa_lock);
> + spin_unlock(&ailp->ail_lock);
> error = xfs_efi_recover(mp, efip);
> - spin_lock(&ailp->xa_lock);
> + spin_lock(&ailp->ail_lock);
>
> return error;
> }
> @@ -4667,9 +4667,9 @@ xlog_recover_cancel_efi(
>
> efip = container_of(lip, struct xfs_efi_log_item, efi_item);
>
> - spin_unlock(&ailp->xa_lock);
> + spin_unlock(&ailp->ail_lock);
> xfs_efi_release(efip);
> - spin_lock(&ailp->xa_lock);
> + spin_lock(&ailp->ail_lock);
> }
>
> /* Recover the RUI if necessary. */
> @@ -4689,9 +4689,9 @@ xlog_recover_process_rui(
> if (test_bit(XFS_RUI_RECOVERED, &ruip->rui_flags))
> return 0;
>
> - spin_unlock(&ailp->xa_lock);
> + spin_unlock(&ailp->ail_lock);
> error = xfs_rui_recover(mp, ruip);
> - spin_lock(&ailp->xa_lock);
> + spin_lock(&ailp->ail_lock);
>
> return error;
> }
> @@ -4707,9 +4707,9 @@ xlog_recover_cancel_rui(
>
> ruip = container_of(lip, struct xfs_rui_log_item, rui_item);
>
> - spin_unlock(&ailp->xa_lock);
> + spin_unlock(&ailp->ail_lock);
> xfs_rui_release(ruip);
> - spin_lock(&ailp->xa_lock);
> + spin_lock(&ailp->ail_lock);
> }
>
> /* Recover the CUI if necessary. */
> @@ -4730,9 +4730,9 @@ xlog_recover_process_cui(
> if (test_bit(XFS_CUI_RECOVERED, &cuip->cui_flags))
> return 0;
>
> - spin_unlock(&ailp->xa_lock);
> + spin_unlock(&ailp->ail_lock);
> error = xfs_cui_recover(mp, cuip, dfops);
> - spin_lock(&ailp->xa_lock);
> + spin_lock(&ailp->ail_lock);
>
> return error;
> }
> @@ -4748,9 +4748,9 @@ xlog_recover_cancel_cui(
>
> cuip = container_of(lip, struct xfs_cui_log_item, cui_item);
>
> - spin_unlock(&ailp->xa_lock);
> + spin_unlock(&ailp->ail_lock);
> xfs_cui_release(cuip);
> - spin_lock(&ailp->xa_lock);
> + spin_lock(&ailp->ail_lock);
> }
>
> /* Recover the BUI if necessary. */
> @@ -4771,9 +4771,9 @@ xlog_recover_process_bui(
> if (test_bit(XFS_BUI_RECOVERED, &buip->bui_flags))
> return 0;
>
> - spin_unlock(&ailp->xa_lock);
> + spin_unlock(&ailp->ail_lock);
> error = xfs_bui_recover(mp, buip, dfops);
> - spin_lock(&ailp->xa_lock);
> + spin_lock(&ailp->ail_lock);
>
> return error;
> }
> @@ -4789,9 +4789,9 @@ xlog_recover_cancel_bui(
>
> buip = container_of(lip, struct xfs_bui_log_item, bui_item);
>
> - spin_unlock(&ailp->xa_lock);
> + spin_unlock(&ailp->ail_lock);
> xfs_bui_release(buip);
> - spin_lock(&ailp->xa_lock);
> + spin_lock(&ailp->ail_lock);
> }
>
> /* Is this log item a deferred action intent? */
> @@ -4879,7 +4879,7 @@ xlog_recover_process_intents(
> #endif
>
> ailp = log->l_ailp;
> - spin_lock(&ailp->xa_lock);
> + spin_lock(&ailp->ail_lock);
> lip = xfs_trans_ail_cursor_first(ailp, &cur, 0);
> #if defined(DEBUG) || defined(XFS_WARN)
> last_lsn = xlog_assign_lsn(log->l_curr_cycle, log->l_curr_block);
> @@ -4933,7 +4933,7 @@ xlog_recover_process_intents(
> }
> out:
> xfs_trans_ail_cursor_done(&cur);
> - spin_unlock(&ailp->xa_lock);
> + spin_unlock(&ailp->ail_lock);
> if (error)
> xfs_defer_cancel(&dfops);
> else
> @@ -4956,7 +4956,7 @@ xlog_recover_cancel_intents(
> struct xfs_ail *ailp;
>
> ailp = log->l_ailp;
> - spin_lock(&ailp->xa_lock);
> + spin_lock(&ailp->ail_lock);
> lip = xfs_trans_ail_cursor_first(ailp, &cur, 0);
> while (lip != NULL) {
> /*
> @@ -4990,7 +4990,7 @@ xlog_recover_cancel_intents(
> }
>
> xfs_trans_ail_cursor_done(&cur);
> - spin_unlock(&ailp->xa_lock);
> + spin_unlock(&ailp->ail_lock);
> return error;
> }
>
> diff --git a/fs/xfs/xfs_trans.c b/fs/xfs/xfs_trans.c
> index a87f657f59c9..756e01999c24 100644
> --- a/fs/xfs/xfs_trans.c
> +++ b/fs/xfs/xfs_trans.c
> @@ -781,8 +781,8 @@ xfs_log_item_batch_insert(
> {
> int i;
>
> - spin_lock(&ailp->xa_lock);
> - /* xfs_trans_ail_update_bulk drops ailp->xa_lock */
> + spin_lock(&ailp->ail_lock);
> + /* xfs_trans_ail_update_bulk drops ailp->ail_lock */
> xfs_trans_ail_update_bulk(ailp, cur, log_items, nr_items, commit_lsn);
>
> for (i = 0; i < nr_items; i++) {
> @@ -825,9 +825,9 @@ xfs_trans_committed_bulk(
> struct xfs_ail_cursor cur;
> int i = 0;
>
> - spin_lock(&ailp->xa_lock);
> + spin_lock(&ailp->ail_lock);
> xfs_trans_ail_cursor_last(ailp, &cur, commit_lsn);
> - spin_unlock(&ailp->xa_lock);
> + spin_unlock(&ailp->ail_lock);
>
> /* unpin all the log items */
> for (lv = log_vector; lv; lv = lv->lv_next ) {
> @@ -847,7 +847,7 @@ xfs_trans_committed_bulk(
> * object into the AIL as we are in a shutdown situation.
> */
> if (aborted) {
> - ASSERT(XFS_FORCED_SHUTDOWN(ailp->xa_mount));
> + ASSERT(XFS_FORCED_SHUTDOWN(ailp->ail_mount));
> lip->li_ops->iop_unpin(lip, 1);
> continue;
> }
> @@ -861,11 +861,11 @@ xfs_trans_committed_bulk(
> * not affect the AIL cursor the bulk insert path is
> * using.
> */
> - spin_lock(&ailp->xa_lock);
> + spin_lock(&ailp->ail_lock);
> if (XFS_LSN_CMP(item_lsn, lip->li_lsn) > 0)
> xfs_trans_ail_update(ailp, lip, item_lsn);
> else
> - spin_unlock(&ailp->xa_lock);
> + spin_unlock(&ailp->ail_lock);
> lip->li_ops->iop_unpin(lip, 0);
> continue;
> }
> @@ -883,9 +883,9 @@ xfs_trans_committed_bulk(
> if (i)
> xfs_log_item_batch_insert(ailp, &cur, log_items, i, commit_lsn);
>
> - spin_lock(&ailp->xa_lock);
> + spin_lock(&ailp->ail_lock);
> xfs_trans_ail_cursor_done(&cur);
> - spin_unlock(&ailp->xa_lock);
> + spin_unlock(&ailp->ail_lock);
> }
>
> /*
> diff --git a/fs/xfs/xfs_trans_ail.c b/fs/xfs/xfs_trans_ail.c
> index cef89f7127d3..d4a2445215e6 100644
> --- a/fs/xfs/xfs_trans_ail.c
> +++ b/fs/xfs/xfs_trans_ail.c
> @@ -40,7 +40,7 @@ xfs_ail_check(
> {
> xfs_log_item_t *prev_lip;
>
> - if (list_empty(&ailp->xa_ail))
> + if (list_empty(&ailp->ail_head))
> return;
>
> /*
> @@ -48,11 +48,11 @@ xfs_ail_check(
> */
> ASSERT((lip->li_flags & XFS_LI_IN_AIL) != 0);
> prev_lip = list_entry(lip->li_ail.prev, xfs_log_item_t, li_ail);
> - if (&prev_lip->li_ail != &ailp->xa_ail)
> + if (&prev_lip->li_ail != &ailp->ail_head)
> ASSERT(XFS_LSN_CMP(prev_lip->li_lsn, lip->li_lsn) <= 0);
>
> prev_lip = list_entry(lip->li_ail.next, xfs_log_item_t, li_ail);
> - if (&prev_lip->li_ail != &ailp->xa_ail)
> + if (&prev_lip->li_ail != &ailp->ail_head)
> ASSERT(XFS_LSN_CMP(prev_lip->li_lsn, lip->li_lsn) >= 0);
>
>
> @@ -69,10 +69,10 @@ static xfs_log_item_t *
> xfs_ail_max(
> struct xfs_ail *ailp)
> {
> - if (list_empty(&ailp->xa_ail))
> + if (list_empty(&ailp->ail_head))
> return NULL;
>
> - return list_entry(ailp->xa_ail.prev, xfs_log_item_t, li_ail);
> + return list_entry(ailp->ail_head.prev, xfs_log_item_t, li_ail);
> }
>
> /*
> @@ -84,7 +84,7 @@ xfs_ail_next(
> struct xfs_ail *ailp,
> xfs_log_item_t *lip)
> {
> - if (lip->li_ail.next == &ailp->xa_ail)
> + if (lip->li_ail.next == &ailp->ail_head)
> return NULL;
>
> return list_first_entry(&lip->li_ail, xfs_log_item_t, li_ail);
> @@ -105,11 +105,11 @@ xfs_ail_min_lsn(
> xfs_lsn_t lsn = 0;
> xfs_log_item_t *lip;
>
> - spin_lock(&ailp->xa_lock);
> + spin_lock(&ailp->ail_lock);
> lip = xfs_ail_min(ailp);
> if (lip)
> lsn = lip->li_lsn;
> - spin_unlock(&ailp->xa_lock);
> + spin_unlock(&ailp->ail_lock);
>
> return lsn;
> }
> @@ -124,11 +124,11 @@ xfs_ail_max_lsn(
> xfs_lsn_t lsn = 0;
> xfs_log_item_t *lip;
>
> - spin_lock(&ailp->xa_lock);
> + spin_lock(&ailp->ail_lock);
> lip = xfs_ail_max(ailp);
> if (lip)
> lsn = lip->li_lsn;
> - spin_unlock(&ailp->xa_lock);
> + spin_unlock(&ailp->ail_lock);
>
> return lsn;
> }
> @@ -146,7 +146,7 @@ xfs_trans_ail_cursor_init(
> struct xfs_ail_cursor *cur)
> {
> cur->item = NULL;
> - list_add_tail(&cur->list, &ailp->xa_cursors);
> + list_add_tail(&cur->list, &ailp->ail_cursors);
> }
>
> /*
> @@ -194,7 +194,7 @@ xfs_trans_ail_cursor_clear(
> {
> struct xfs_ail_cursor *cur;
>
> - list_for_each_entry(cur, &ailp->xa_cursors, list) {
> + list_for_each_entry(cur, &ailp->ail_cursors, list) {
> if (cur->item == lip)
> cur->item = (struct xfs_log_item *)
> ((uintptr_t)cur->item | 1);
> @@ -222,7 +222,7 @@ xfs_trans_ail_cursor_first(
> goto out;
> }
>
> - list_for_each_entry(lip, &ailp->xa_ail, li_ail) {
> + list_for_each_entry(lip, &ailp->ail_head, li_ail) {
> if (XFS_LSN_CMP(lip->li_lsn, lsn) >= 0)
> goto out;
> }
> @@ -241,7 +241,7 @@ __xfs_trans_ail_cursor_last(
> {
> xfs_log_item_t *lip;
>
> - list_for_each_entry_reverse(lip, &ailp->xa_ail, li_ail) {
> + list_for_each_entry_reverse(lip, &ailp->ail_head, li_ail) {
> if (XFS_LSN_CMP(lip->li_lsn, lsn) <= 0)
> return lip;
> }
> @@ -310,7 +310,7 @@ xfs_ail_splice(
> if (lip)
> list_splice(list, &lip->li_ail);
> else
> - list_splice(list, &ailp->xa_ail);
> + list_splice(list, &ailp->ail_head);
> }
>
> /*
> @@ -335,17 +335,17 @@ xfsaild_push_item(
> * If log item pinning is enabled, skip the push and track the item as
> * pinned. This can help induce head-behind-tail conditions.
> */
> - if (XFS_TEST_ERROR(false, ailp->xa_mount, XFS_ERRTAG_LOG_ITEM_PIN))
> + if (XFS_TEST_ERROR(false, ailp->ail_mount, XFS_ERRTAG_LOG_ITEM_PIN))
> return XFS_ITEM_PINNED;
>
> - return lip->li_ops->iop_push(lip, &ailp->xa_buf_list);
> + return lip->li_ops->iop_push(lip, &ailp->ail_buf_list);
> }
>
> static long
> xfsaild_push(
> struct xfs_ail *ailp)
> {
> - xfs_mount_t *mp = ailp->xa_mount;
> + xfs_mount_t *mp = ailp->ail_mount;
> struct xfs_ail_cursor cur;
> xfs_log_item_t *lip;
> xfs_lsn_t lsn;
> @@ -360,30 +360,30 @@ xfsaild_push(
> * buffers the last time we ran, force the log first and wait for it
> * before pushing again.
> */
> - if (ailp->xa_log_flush && ailp->xa_last_pushed_lsn == 0 &&
> - (!list_empty_careful(&ailp->xa_buf_list) ||
> + if (ailp->ail_log_flush && ailp->ail_last_pushed_lsn == 0 &&
> + (!list_empty_careful(&ailp->ail_buf_list) ||
> xfs_ail_min_lsn(ailp))) {
> - ailp->xa_log_flush = 0;
> + ailp->ail_log_flush = 0;
>
> XFS_STATS_INC(mp, xs_push_ail_flush);
> xfs_log_force(mp, XFS_LOG_SYNC);
> }
>
> - spin_lock(&ailp->xa_lock);
> + spin_lock(&ailp->ail_lock);
>
> - /* barrier matches the xa_target update in xfs_ail_push() */
> + /* barrier matches the ail_target update in xfs_ail_push() */
> smp_rmb();
> - target = ailp->xa_target;
> - ailp->xa_target_prev = target;
> + target = ailp->ail_target;
> + ailp->ail_target_prev = target;
>
> - lip = xfs_trans_ail_cursor_first(ailp, &cur, ailp->xa_last_pushed_lsn);
> + lip = xfs_trans_ail_cursor_first(ailp, &cur, ailp->ail_last_pushed_lsn);
> if (!lip) {
> /*
> * If the AIL is empty or our push has reached the end we are
> * done now.
> */
> xfs_trans_ail_cursor_done(&cur);
> - spin_unlock(&ailp->xa_lock);
> + spin_unlock(&ailp->ail_lock);
> goto out_done;
> }
>
> @@ -404,7 +404,7 @@ xfsaild_push(
> XFS_STATS_INC(mp, xs_push_ail_success);
> trace_xfs_ail_push(lip);
>
> - ailp->xa_last_pushed_lsn = lsn;
> + ailp->ail_last_pushed_lsn = lsn;
> break;
>
> case XFS_ITEM_FLUSHING:
> @@ -423,7 +423,7 @@ xfsaild_push(
> trace_xfs_ail_flushing(lip);
>
> flushing++;
> - ailp->xa_last_pushed_lsn = lsn;
> + ailp->ail_last_pushed_lsn = lsn;
> break;
>
> case XFS_ITEM_PINNED:
> @@ -431,7 +431,7 @@ xfsaild_push(
> trace_xfs_ail_pinned(lip);
>
> stuck++;
> - ailp->xa_log_flush++;
> + ailp->ail_log_flush++;
> break;
> case XFS_ITEM_LOCKED:
> XFS_STATS_INC(mp, xs_push_ail_locked);
> @@ -468,10 +468,10 @@ xfsaild_push(
> lsn = lip->li_lsn;
> }
> xfs_trans_ail_cursor_done(&cur);
> - spin_unlock(&ailp->xa_lock);
> + spin_unlock(&ailp->ail_lock);
>
> - if (xfs_buf_delwri_submit_nowait(&ailp->xa_buf_list))
> - ailp->xa_log_flush++;
> + if (xfs_buf_delwri_submit_nowait(&ailp->ail_buf_list))
> + ailp->ail_log_flush++;
>
> if (!count || XFS_LSN_CMP(lsn, target) >= 0) {
> out_done:
> @@ -481,7 +481,7 @@ xfsaild_push(
> * AIL before we start the next scan from the start of the AIL.
> */
> tout = 50;
> - ailp->xa_last_pushed_lsn = 0;
> + ailp->ail_last_pushed_lsn = 0;
> } else if (((stuck + flushing) * 100) / count > 90) {
> /*
> * Either there is a lot of contention on the AIL or we are
> @@ -494,7 +494,7 @@ xfsaild_push(
> * the restart to issue a log force to unpin the stuck items.
> */
> tout = 20;
> - ailp->xa_last_pushed_lsn = 0;
> + ailp->ail_last_pushed_lsn = 0;
> } else {
> /*
> * Assume we have more work to do in a short while.
> @@ -536,26 +536,26 @@ xfsaild(
> break;
> }
>
> - spin_lock(&ailp->xa_lock);
> + spin_lock(&ailp->ail_lock);
>
> /*
> * Idle if the AIL is empty and we are not racing with a target
> * update. We check the AIL after we set the task to a sleep
> - * state to guarantee that we either catch an xa_target update
> + * state to guarantee that we either catch an ail_target update
> * or that a wake_up resets the state to TASK_RUNNING.
> * Otherwise, we run the risk of sleeping indefinitely.
> *
> - * The barrier matches the xa_target update in xfs_ail_push().
> + * The barrier matches the ail_target update in xfs_ail_push().
> */
> smp_rmb();
> if (!xfs_ail_min(ailp) &&
> - ailp->xa_target == ailp->xa_target_prev) {
> - spin_unlock(&ailp->xa_lock);
> + ailp->ail_target == ailp->ail_target_prev) {
> + spin_unlock(&ailp->ail_lock);
> freezable_schedule();
> tout = 0;
> continue;
> }
> - spin_unlock(&ailp->xa_lock);
> + spin_unlock(&ailp->ail_lock);
>
> if (tout)
> freezable_schedule_timeout(msecs_to_jiffies(tout));
> @@ -592,8 +592,8 @@ xfs_ail_push(
> xfs_log_item_t *lip;
>
> lip = xfs_ail_min(ailp);
> - if (!lip || XFS_FORCED_SHUTDOWN(ailp->xa_mount) ||
> - XFS_LSN_CMP(threshold_lsn, ailp->xa_target) <= 0)
> + if (!lip || XFS_FORCED_SHUTDOWN(ailp->ail_mount) ||
> + XFS_LSN_CMP(threshold_lsn, ailp->ail_target) <= 0)
> return;
>
> /*
> @@ -601,10 +601,10 @@ xfs_ail_push(
> * the XFS_AIL_PUSHING_BIT.
> */
> smp_wmb();
> - xfs_trans_ail_copy_lsn(ailp, &ailp->xa_target, &threshold_lsn);
> + xfs_trans_ail_copy_lsn(ailp, &ailp->ail_target, &threshold_lsn);
> smp_wmb();
>
> - wake_up_process(ailp->xa_task);
> + wake_up_process(ailp->ail_task);
> }
>
> /*
> @@ -630,18 +630,18 @@ xfs_ail_push_all_sync(
> struct xfs_log_item *lip;
> DEFINE_WAIT(wait);
>
> - spin_lock(&ailp->xa_lock);
> + spin_lock(&ailp->ail_lock);
> while ((lip = xfs_ail_max(ailp)) != NULL) {
> - prepare_to_wait(&ailp->xa_empty, &wait, TASK_UNINTERRUPTIBLE);
> - ailp->xa_target = lip->li_lsn;
> - wake_up_process(ailp->xa_task);
> - spin_unlock(&ailp->xa_lock);
> + prepare_to_wait(&ailp->ail_empty, &wait, TASK_UNINTERRUPTIBLE);
> + ailp->ail_target = lip->li_lsn;
> + wake_up_process(ailp->ail_task);
> + spin_unlock(&ailp->ail_lock);
> schedule();
> - spin_lock(&ailp->xa_lock);
> + spin_lock(&ailp->ail_lock);
> }
> - spin_unlock(&ailp->xa_lock);
> + spin_unlock(&ailp->ail_lock);
>
> - finish_wait(&ailp->xa_empty, &wait);
> + finish_wait(&ailp->ail_empty, &wait);
> }
>
> /*
> @@ -672,7 +672,7 @@ xfs_trans_ail_update_bulk(
> struct xfs_ail_cursor *cur,
> struct xfs_log_item **log_items,
> int nr_items,
> - xfs_lsn_t lsn) __releases(ailp->xa_lock)
> + xfs_lsn_t lsn) __releases(ailp->ail_lock)
> {
> xfs_log_item_t *mlip;
> int mlip_changed = 0;
> @@ -705,13 +705,13 @@ xfs_trans_ail_update_bulk(
> xfs_ail_splice(ailp, cur, &tmp, lsn);
>
> if (mlip_changed) {
> - if (!XFS_FORCED_SHUTDOWN(ailp->xa_mount))
> - xlog_assign_tail_lsn_locked(ailp->xa_mount);
> - spin_unlock(&ailp->xa_lock);
> + if (!XFS_FORCED_SHUTDOWN(ailp->ail_mount))
> + xlog_assign_tail_lsn_locked(ailp->ail_mount);
> + spin_unlock(&ailp->ail_lock);
>
> - xfs_log_space_wake(ailp->xa_mount);
> + xfs_log_space_wake(ailp->ail_mount);
> } else {
> - spin_unlock(&ailp->xa_lock);
> + spin_unlock(&ailp->ail_lock);
> }
> }
>
> @@ -756,13 +756,13 @@ void
> xfs_trans_ail_delete(
> struct xfs_ail *ailp,
> struct xfs_log_item *lip,
> - int shutdown_type) __releases(ailp->xa_lock)
> + int shutdown_type) __releases(ailp->ail_lock)
> {
> - struct xfs_mount *mp = ailp->xa_mount;
> + struct xfs_mount *mp = ailp->ail_mount;
> bool mlip_changed;
>
> if (!(lip->li_flags & XFS_LI_IN_AIL)) {
> - spin_unlock(&ailp->xa_lock);
> + spin_unlock(&ailp->ail_lock);
> if (!XFS_FORCED_SHUTDOWN(mp)) {
> xfs_alert_tag(mp, XFS_PTAG_AILDELETE,
> "%s: attempting to delete a log item that is not in the AIL",
> @@ -776,13 +776,13 @@ xfs_trans_ail_delete(
> if (mlip_changed) {
> if (!XFS_FORCED_SHUTDOWN(mp))
> xlog_assign_tail_lsn_locked(mp);
> - if (list_empty(&ailp->xa_ail))
> - wake_up_all(&ailp->xa_empty);
> + if (list_empty(&ailp->ail_head))
> + wake_up_all(&ailp->ail_empty);
> }
>
> - spin_unlock(&ailp->xa_lock);
> + spin_unlock(&ailp->ail_lock);
> if (mlip_changed)
> - xfs_log_space_wake(ailp->xa_mount);
> + xfs_log_space_wake(ailp->ail_mount);
> }
>
> int
> @@ -795,16 +795,16 @@ xfs_trans_ail_init(
> if (!ailp)
> return -ENOMEM;
>
> - ailp->xa_mount = mp;
> - INIT_LIST_HEAD(&ailp->xa_ail);
> - INIT_LIST_HEAD(&ailp->xa_cursors);
> - spin_lock_init(&ailp->xa_lock);
> - INIT_LIST_HEAD(&ailp->xa_buf_list);
> - init_waitqueue_head(&ailp->xa_empty);
> + ailp->ail_mount = mp;
> + INIT_LIST_HEAD(&ailp->ail_head);
> + INIT_LIST_HEAD(&ailp->ail_cursors);
> + spin_lock_init(&ailp->ail_lock);
> + INIT_LIST_HEAD(&ailp->ail_buf_list);
> + init_waitqueue_head(&ailp->ail_empty);
>
> - ailp->xa_task = kthread_run(xfsaild, ailp, "xfsaild/%s",
> - ailp->xa_mount->m_fsname);
> - if (IS_ERR(ailp->xa_task))
> + ailp->ail_task = kthread_run(xfsaild, ailp, "xfsaild/%s",
> + ailp->ail_mount->m_fsname);
> + if (IS_ERR(ailp->ail_task))
> goto out_free_ailp;
>
> mp->m_ail = ailp;
> @@ -821,6 +821,6 @@ xfs_trans_ail_destroy(
> {
> struct xfs_ail *ailp = mp->m_ail;
>
> - kthread_stop(ailp->xa_task);
> + kthread_stop(ailp->ail_task);
> kmem_free(ailp);
> }
> diff --git a/fs/xfs/xfs_trans_buf.c b/fs/xfs/xfs_trans_buf.c
> index 3ba7a96a8abd..b8871bcfe00b 100644
> --- a/fs/xfs/xfs_trans_buf.c
> +++ b/fs/xfs/xfs_trans_buf.c
> @@ -429,8 +429,8 @@ xfs_trans_brelse(xfs_trans_t *tp,
> * If the fs has shutdown and we dropped the last reference, it may fall
> * on us to release a (possibly dirty) bli if it never made it to the
> * AIL (e.g., the aborted unpin already happened and didn't release it
> - * due to our reference). Since we're already shutdown and need xa_lock,
> - * just force remove from the AIL and release the bli here.
> + * due to our reference). Since we're already shutdown and need
> + * ail_lock, just force remove from the AIL and release the bli here.
> */
> if (XFS_FORCED_SHUTDOWN(tp->t_mountp) && freed) {
> xfs_trans_ail_remove(&bip->bli_item, SHUTDOWN_LOG_IO_ERROR);
> diff --git a/fs/xfs/xfs_trans_priv.h b/fs/xfs/xfs_trans_priv.h
> index b317a3644c00..be24b0c8a332 100644
> --- a/fs/xfs/xfs_trans_priv.h
> +++ b/fs/xfs/xfs_trans_priv.h
> @@ -65,17 +65,17 @@ struct xfs_ail_cursor {
> * Eventually we need to drive the locking in here as well.
> */
> struct xfs_ail {
> - struct xfs_mount *xa_mount;
> - struct task_struct *xa_task;
> - struct list_head xa_ail;
> - xfs_lsn_t xa_target;
> - xfs_lsn_t xa_target_prev;
> - struct list_head xa_cursors;
> - spinlock_t xa_lock;
> - xfs_lsn_t xa_last_pushed_lsn;
> - int xa_log_flush;
> - struct list_head xa_buf_list;
> - wait_queue_head_t xa_empty;
> + struct xfs_mount *ail_mount;
> + struct task_struct *ail_task;
> + struct list_head ail_head;
> + xfs_lsn_t ail_target;
> + xfs_lsn_t ail_target_prev;
> + struct list_head ail_cursors;
> + spinlock_t ail_lock;
> + xfs_lsn_t ail_last_pushed_lsn;
> + int ail_log_flush;
> + struct list_head ail_buf_list;
> + wait_queue_head_t ail_empty;
> };
>
> /*
> @@ -84,7 +84,7 @@ struct xfs_ail {
> void xfs_trans_ail_update_bulk(struct xfs_ail *ailp,
> struct xfs_ail_cursor *cur,
> struct xfs_log_item **log_items, int nr_items,
> - xfs_lsn_t lsn) __releases(ailp->xa_lock);
> + xfs_lsn_t lsn) __releases(ailp->ail_lock);
> /*
> * Return a pointer to the first item in the AIL. If the AIL is empty, then
> * return NULL.
> @@ -93,7 +93,7 @@ static inline struct xfs_log_item *
> xfs_ail_min(
> struct xfs_ail *ailp)
> {
> - return list_first_entry_or_null(&ailp->xa_ail, struct xfs_log_item,
> + return list_first_entry_or_null(&ailp->ail_head, struct xfs_log_item,
> li_ail);
> }
>
> @@ -101,14 +101,14 @@ static inline void
> xfs_trans_ail_update(
> struct xfs_ail *ailp,
> struct xfs_log_item *lip,
> - xfs_lsn_t lsn) __releases(ailp->xa_lock)
> + xfs_lsn_t lsn) __releases(ailp->ail_lock)
> {
> xfs_trans_ail_update_bulk(ailp, NULL, &lip, 1, lsn);
> }
>
> bool xfs_ail_delete_one(struct xfs_ail *ailp, struct xfs_log_item *lip);
> void xfs_trans_ail_delete(struct xfs_ail *ailp, struct xfs_log_item *lip,
> - int shutdown_type) __releases(ailp->xa_lock);
> + int shutdown_type) __releases(ailp->ail_lock);
>
> static inline void
> xfs_trans_ail_remove(
> @@ -117,12 +117,12 @@ xfs_trans_ail_remove(
> {
> struct xfs_ail *ailp = lip->li_ailp;
>
> - spin_lock(&ailp->xa_lock);
> + spin_lock(&ailp->ail_lock);
> /* xfs_trans_ail_delete() drops the AIL lock */
> if (lip->li_flags & XFS_LI_IN_AIL)
> xfs_trans_ail_delete(ailp, lip, shutdown_type);
> else
> - spin_unlock(&ailp->xa_lock);
> + spin_unlock(&ailp->ail_lock);
> }
>
> void xfs_ail_push(struct xfs_ail *, xfs_lsn_t);
> @@ -149,9 +149,9 @@ xfs_trans_ail_copy_lsn(
> xfs_lsn_t *src)
> {
> ASSERT(sizeof(xfs_lsn_t) == 8); /* don't lock if it shrinks */
> - spin_lock(&ailp->xa_lock);
> + spin_lock(&ailp->ail_lock);
> *dst = *src;
> - spin_unlock(&ailp->xa_lock);
> + spin_unlock(&ailp->ail_lock);
> }
> #else
> static inline void
> @@ -172,7 +172,7 @@ xfs_clear_li_failed(
> struct xfs_buf *bp = lip->li_buf;
>
> ASSERT(lip->li_flags & XFS_LI_IN_AIL);
> - lockdep_assert_held(&lip->li_ailp->xa_lock);
> + lockdep_assert_held(&lip->li_ailp->ail_lock);
>
> if (lip->li_flags & XFS_LI_FAILED) {
> lip->li_flags &= ~XFS_LI_FAILED;
> @@ -186,7 +186,7 @@ xfs_set_li_failed(
> struct xfs_log_item *lip,
> struct xfs_buf *bp)
> {
> - lockdep_assert_held(&lip->li_ailp->xa_lock);
> + lockdep_assert_held(&lip->li_ailp->ail_lock);
>
> if (!(lip->li_flags & XFS_LI_FAILED)) {
> xfs_buf_hold(bp);
> --
> 2.15.1
>
> --
> To unsubscribe from this list: send the line "unsubscribe linux-xfs" in
> the body of a message to majordomo@xxxxxxxxxxxxxxx
> More majordomo info at http://vger.kernel.org/majordomo-info.html