diff mbox

[v5,01/78] xfs: Rename xa_ elements to ail_

Message ID 20171215220450.7899-2-willy@infradead.org (mailing list archive)
State New, archived
Headers show

Commit Message

Matthew Wilcox (Oracle) Dec. 15, 2017, 10:03 p.m. UTC
From: Matthew Wilcox <mawilcox@microsoft.com>

This is a simple rename, except that xa_ail becomes ail_head.

Signed-off-by: Matthew Wilcox <mawilcox@microsoft.com>
---
 fs/xfs/xfs_buf_item.c    |  10 ++--
 fs/xfs/xfs_dquot.c       |   4 +-
 fs/xfs/xfs_dquot_item.c  |  11 ++--
 fs/xfs/xfs_inode_item.c  |  22 +++----
 fs/xfs/xfs_log.c         |   6 +-
 fs/xfs/xfs_log_recover.c |  80 ++++++++++++-------------
 fs/xfs/xfs_trans.c       |  18 +++---
 fs/xfs/xfs_trans_ail.c   | 152 +++++++++++++++++++++++------------------------
 fs/xfs/xfs_trans_buf.c   |   4 +-
 fs/xfs/xfs_trans_priv.h  |  42 ++++++-------
 10 files changed, 175 insertions(+), 174 deletions(-)

Comments

Darrick J. Wong Jan. 3, 2018, 1:01 a.m. UTC | #1
On Fri, Dec 15, 2017 at 02:03:33PM -0800, Matthew Wilcox wrote:
> From: Matthew Wilcox <mawilcox@microsoft.com>
> 
> This is a simple rename, except that xa_ail becomes ail_head.
> 
> Signed-off-by: Matthew Wilcox <mawilcox@microsoft.com>

That was an eyeful,
Reviewed-by: Darrick J. Wong <darrick.wong@oracle.com>

> ---
>  fs/xfs/xfs_buf_item.c    |  10 ++--
>  fs/xfs/xfs_dquot.c       |   4 +-
>  fs/xfs/xfs_dquot_item.c  |  11 ++--
>  fs/xfs/xfs_inode_item.c  |  22 +++----
>  fs/xfs/xfs_log.c         |   6 +-
>  fs/xfs/xfs_log_recover.c |  80 ++++++++++++-------------
>  fs/xfs/xfs_trans.c       |  18 +++---
>  fs/xfs/xfs_trans_ail.c   | 152 +++++++++++++++++++++++------------------------
>  fs/xfs/xfs_trans_buf.c   |   4 +-
>  fs/xfs/xfs_trans_priv.h  |  42 ++++++-------
>  10 files changed, 175 insertions(+), 174 deletions(-)
> 
> diff --git a/fs/xfs/xfs_buf_item.c b/fs/xfs/xfs_buf_item.c
> index e0a0af0946f2..6c5035544a93 100644
> --- a/fs/xfs/xfs_buf_item.c
> +++ b/fs/xfs/xfs_buf_item.c
> @@ -459,7 +459,7 @@ xfs_buf_item_unpin(
>  			bp->b_fspriv = NULL;
>  			bp->b_iodone = NULL;
>  		} else {
> -			spin_lock(&ailp->xa_lock);
> +			spin_lock(&ailp->ail_lock);
>  			xfs_trans_ail_delete(ailp, lip, SHUTDOWN_LOG_IO_ERROR);
>  			xfs_buf_item_relse(bp);
>  			ASSERT(bp->b_fspriv == NULL);
> @@ -1056,13 +1056,13 @@ xfs_buf_do_callbacks_fail(
>  	struct xfs_log_item	*lip = bp->b_fspriv;
>  	struct xfs_ail		*ailp = lip->li_ailp;
>  
> -	spin_lock(&ailp->xa_lock);
> +	spin_lock(&ailp->ail_lock);
>  	for (; lip; lip = next) {
>  		next = lip->li_bio_list;
>  		if (lip->li_ops->iop_error)
>  			lip->li_ops->iop_error(lip, bp);
>  	}
> -	spin_unlock(&ailp->xa_lock);
> +	spin_unlock(&ailp->ail_lock);
>  }
>  
>  static bool
> @@ -1215,7 +1215,7 @@ xfs_buf_iodone(
>  	 *
>  	 * Either way, AIL is useless if we're forcing a shutdown.
>  	 */
> -	spin_lock(&ailp->xa_lock);
> +	spin_lock(&ailp->ail_lock);
>  	xfs_trans_ail_delete(ailp, lip, SHUTDOWN_CORRUPT_INCORE);
>  	xfs_buf_item_free(BUF_ITEM(lip));
>  }
> @@ -1236,7 +1236,7 @@ xfs_buf_resubmit_failed_buffers(
>  	/*
>  	 * Clear XFS_LI_FAILED flag from all items before resubmit
>  	 *
> -	 * XFS_LI_FAILED set/clear is protected by xa_lock, caller  this
> +	 * XFS_LI_FAILED set/clear is protected by ail_lock, caller  this
>  	 * function already have it acquired
>  	 */
>  	for (; lip; lip = next) {
> diff --git a/fs/xfs/xfs_dquot.c b/fs/xfs/xfs_dquot.c
> index f248708c10ff..e2a466df5dd1 100644
> --- a/fs/xfs/xfs_dquot.c
> +++ b/fs/xfs/xfs_dquot.c
> @@ -974,7 +974,7 @@ xfs_qm_dqflush_done(
>  	     (lip->li_flags & XFS_LI_FAILED))) {
>  
>  		/* xfs_trans_ail_delete() drops the AIL lock. */
> -		spin_lock(&ailp->xa_lock);
> +		spin_lock(&ailp->ail_lock);
>  		if (lip->li_lsn == qip->qli_flush_lsn) {
>  			xfs_trans_ail_delete(ailp, lip, SHUTDOWN_CORRUPT_INCORE);
>  		} else {
> @@ -984,7 +984,7 @@ xfs_qm_dqflush_done(
>  			 */
>  			if (lip->li_flags & XFS_LI_FAILED)
>  				xfs_clear_li_failed(lip);
> -			spin_unlock(&ailp->xa_lock);
> +			spin_unlock(&ailp->ail_lock);
>  		}
>  	}
>  
> diff --git a/fs/xfs/xfs_dquot_item.c b/fs/xfs/xfs_dquot_item.c
> index 664dea105e76..62637a226601 100644
> --- a/fs/xfs/xfs_dquot_item.c
> +++ b/fs/xfs/xfs_dquot_item.c
> @@ -160,8 +160,9 @@ xfs_dquot_item_error(
>  STATIC uint
>  xfs_qm_dquot_logitem_push(
>  	struct xfs_log_item	*lip,
> -	struct list_head	*buffer_list) __releases(&lip->li_ailp->xa_lock)
> -					      __acquires(&lip->li_ailp->xa_lock)
> +	struct list_head	*buffer_list)
> +		__releases(&lip->li_ailp->ail_lock)
> +		__acquires(&lip->li_ailp->ail_lock)
>  {
>  	struct xfs_dquot	*dqp = DQUOT_ITEM(lip)->qli_dquot;
>  	struct xfs_buf		*bp = lip->li_buf;
> @@ -208,7 +209,7 @@ xfs_qm_dquot_logitem_push(
>  		goto out_unlock;
>  	}
>  
> -	spin_unlock(&lip->li_ailp->xa_lock);
> +	spin_unlock(&lip->li_ailp->ail_lock);
>  
>  	error = xfs_qm_dqflush(dqp, &bp);
>  	if (error) {
> @@ -220,7 +221,7 @@ xfs_qm_dquot_logitem_push(
>  		xfs_buf_relse(bp);
>  	}
>  
> -	spin_lock(&lip->li_ailp->xa_lock);
> +	spin_lock(&lip->li_ailp->ail_lock);
>  out_unlock:
>  	xfs_dqunlock(dqp);
>  	return rval;
> @@ -403,7 +404,7 @@ xfs_qm_qoffend_logitem_committed(
>  	 * Delete the qoff-start logitem from the AIL.
>  	 * xfs_trans_ail_delete() drops the AIL lock.
>  	 */
> -	spin_lock(&ailp->xa_lock);
> +	spin_lock(&ailp->ail_lock);
>  	xfs_trans_ail_delete(ailp, &qfs->qql_item, SHUTDOWN_LOG_IO_ERROR);
>  
>  	kmem_free(qfs->qql_item.li_lv_shadow);
> diff --git a/fs/xfs/xfs_inode_item.c b/fs/xfs/xfs_inode_item.c
> index 6ee5c3bf19ad..071acd4249a0 100644
> --- a/fs/xfs/xfs_inode_item.c
> +++ b/fs/xfs/xfs_inode_item.c
> @@ -501,8 +501,8 @@ STATIC uint
>  xfs_inode_item_push(
>  	struct xfs_log_item	*lip,
>  	struct list_head	*buffer_list)
> -		__releases(&lip->li_ailp->xa_lock)
> -		__acquires(&lip->li_ailp->xa_lock)
> +		__releases(&lip->li_ailp->ail_lock)
> +		__acquires(&lip->li_ailp->ail_lock)
>  {
>  	struct xfs_inode_log_item *iip = INODE_ITEM(lip);
>  	struct xfs_inode	*ip = iip->ili_inode;
> @@ -561,7 +561,7 @@ xfs_inode_item_push(
>  	ASSERT(iip->ili_fields != 0 || XFS_FORCED_SHUTDOWN(ip->i_mount));
>  	ASSERT(iip->ili_logged == 0 || XFS_FORCED_SHUTDOWN(ip->i_mount));
>  
> -	spin_unlock(&lip->li_ailp->xa_lock);
> +	spin_unlock(&lip->li_ailp->ail_lock);
>  
>  	error = xfs_iflush(ip, &bp);
>  	if (!error) {
> @@ -570,7 +570,7 @@ xfs_inode_item_push(
>  		xfs_buf_relse(bp);
>  	}
>  
> -	spin_lock(&lip->li_ailp->xa_lock);
> +	spin_lock(&lip->li_ailp->ail_lock);
>  out_unlock:
>  	xfs_iunlock(ip, XFS_ILOCK_SHARED);
>  	return rval;
> @@ -774,7 +774,7 @@ xfs_iflush_done(
>  		bool			mlip_changed = false;
>  
>  		/* this is an opencoded batch version of xfs_trans_ail_delete */
> -		spin_lock(&ailp->xa_lock);
> +		spin_lock(&ailp->ail_lock);
>  		for (blip = lip; blip; blip = blip->li_bio_list) {
>  			if (INODE_ITEM(blip)->ili_logged &&
>  			    blip->li_lsn == INODE_ITEM(blip)->ili_flush_lsn)
> @@ -785,15 +785,15 @@ xfs_iflush_done(
>  		}
>  
>  		if (mlip_changed) {
> -			if (!XFS_FORCED_SHUTDOWN(ailp->xa_mount))
> -				xlog_assign_tail_lsn_locked(ailp->xa_mount);
> -			if (list_empty(&ailp->xa_ail))
> -				wake_up_all(&ailp->xa_empty);
> +			if (!XFS_FORCED_SHUTDOWN(ailp->ail_mount))
> +				xlog_assign_tail_lsn_locked(ailp->ail_mount);
> +			if (list_empty(&ailp->ail_head))
> +				wake_up_all(&ailp->ail_empty);
>  		}
> -		spin_unlock(&ailp->xa_lock);
> +		spin_unlock(&ailp->ail_lock);
>  
>  		if (mlip_changed)
> -			xfs_log_space_wake(ailp->xa_mount);
> +			xfs_log_space_wake(ailp->ail_mount);
>  	}
>  
>  	/*
> diff --git a/fs/xfs/xfs_log.c b/fs/xfs/xfs_log.c
> index a503af96d780..7148625eebf2 100644
> --- a/fs/xfs/xfs_log.c
> +++ b/fs/xfs/xfs_log.c
> @@ -1148,7 +1148,7 @@ xlog_assign_tail_lsn_locked(
>  	struct xfs_log_item	*lip;
>  	xfs_lsn_t		tail_lsn;
>  
> -	assert_spin_locked(&mp->m_ail->xa_lock);
> +	assert_spin_locked(&mp->m_ail->ail_lock);
>  
>  	/*
>  	 * To make sure we always have a valid LSN for the log tail we keep
> @@ -1171,9 +1171,9 @@ xlog_assign_tail_lsn(
>  {
>  	xfs_lsn_t		tail_lsn;
>  
> -	spin_lock(&mp->m_ail->xa_lock);
> +	spin_lock(&mp->m_ail->ail_lock);
>  	tail_lsn = xlog_assign_tail_lsn_locked(mp);
> -	spin_unlock(&mp->m_ail->xa_lock);
> +	spin_unlock(&mp->m_ail->ail_lock);
>  
>  	return tail_lsn;
>  }
> diff --git a/fs/xfs/xfs_log_recover.c b/fs/xfs/xfs_log_recover.c
> index 28d1abfe835e..d871761626fb 100644
> --- a/fs/xfs/xfs_log_recover.c
> +++ b/fs/xfs/xfs_log_recover.c
> @@ -3424,7 +3424,7 @@ xlog_recover_efi_pass2(
>  	}
>  	atomic_set(&efip->efi_next_extent, efi_formatp->efi_nextents);
>  
> -	spin_lock(&log->l_ailp->xa_lock);
> +	spin_lock(&log->l_ailp->ail_lock);
>  	/*
>  	 * The EFI has two references. One for the EFD and one for EFI to ensure
>  	 * it makes it into the AIL. Insert the EFI into the AIL directly and
> @@ -3467,7 +3467,7 @@ xlog_recover_efd_pass2(
>  	 * Search for the EFI with the id in the EFD format structure in the
>  	 * AIL.
>  	 */
> -	spin_lock(&ailp->xa_lock);
> +	spin_lock(&ailp->ail_lock);
>  	lip = xfs_trans_ail_cursor_first(ailp, &cur, 0);
>  	while (lip != NULL) {
>  		if (lip->li_type == XFS_LI_EFI) {
> @@ -3477,9 +3477,9 @@ xlog_recover_efd_pass2(
>  				 * Drop the EFD reference to the EFI. This
>  				 * removes the EFI from the AIL and frees it.
>  				 */
> -				spin_unlock(&ailp->xa_lock);
> +				spin_unlock(&ailp->ail_lock);
>  				xfs_efi_release(efip);
> -				spin_lock(&ailp->xa_lock);
> +				spin_lock(&ailp->ail_lock);
>  				break;
>  			}
>  		}
> @@ -3487,7 +3487,7 @@ xlog_recover_efd_pass2(
>  	}
>  
>  	xfs_trans_ail_cursor_done(&cur);
> -	spin_unlock(&ailp->xa_lock);
> +	spin_unlock(&ailp->ail_lock);
>  
>  	return 0;
>  }
> @@ -3520,7 +3520,7 @@ xlog_recover_rui_pass2(
>  	}
>  	atomic_set(&ruip->rui_next_extent, rui_formatp->rui_nextents);
>  
> -	spin_lock(&log->l_ailp->xa_lock);
> +	spin_lock(&log->l_ailp->ail_lock);
>  	/*
>  	 * The RUI has two references. One for the RUD and one for RUI to ensure
>  	 * it makes it into the AIL. Insert the RUI into the AIL directly and
> @@ -3560,7 +3560,7 @@ xlog_recover_rud_pass2(
>  	 * Search for the RUI with the id in the RUD format structure in the
>  	 * AIL.
>  	 */
> -	spin_lock(&ailp->xa_lock);
> +	spin_lock(&ailp->ail_lock);
>  	lip = xfs_trans_ail_cursor_first(ailp, &cur, 0);
>  	while (lip != NULL) {
>  		if (lip->li_type == XFS_LI_RUI) {
> @@ -3570,9 +3570,9 @@ xlog_recover_rud_pass2(
>  				 * Drop the RUD reference to the RUI. This
>  				 * removes the RUI from the AIL and frees it.
>  				 */
> -				spin_unlock(&ailp->xa_lock);
> +				spin_unlock(&ailp->ail_lock);
>  				xfs_rui_release(ruip);
> -				spin_lock(&ailp->xa_lock);
> +				spin_lock(&ailp->ail_lock);
>  				break;
>  			}
>  		}
> @@ -3580,7 +3580,7 @@ xlog_recover_rud_pass2(
>  	}
>  
>  	xfs_trans_ail_cursor_done(&cur);
> -	spin_unlock(&ailp->xa_lock);
> +	spin_unlock(&ailp->ail_lock);
>  
>  	return 0;
>  }
> @@ -3636,7 +3636,7 @@ xlog_recover_cui_pass2(
>  	}
>  	atomic_set(&cuip->cui_next_extent, cui_formatp->cui_nextents);
>  
> -	spin_lock(&log->l_ailp->xa_lock);
> +	spin_lock(&log->l_ailp->ail_lock);
>  	/*
>  	 * The CUI has two references. One for the CUD and one for CUI to ensure
>  	 * it makes it into the AIL. Insert the CUI into the AIL directly and
> @@ -3677,7 +3677,7 @@ xlog_recover_cud_pass2(
>  	 * Search for the CUI with the id in the CUD format structure in the
>  	 * AIL.
>  	 */
> -	spin_lock(&ailp->xa_lock);
> +	spin_lock(&ailp->ail_lock);
>  	lip = xfs_trans_ail_cursor_first(ailp, &cur, 0);
>  	while (lip != NULL) {
>  		if (lip->li_type == XFS_LI_CUI) {
> @@ -3687,9 +3687,9 @@ xlog_recover_cud_pass2(
>  				 * Drop the CUD reference to the CUI. This
>  				 * removes the CUI from the AIL and frees it.
>  				 */
> -				spin_unlock(&ailp->xa_lock);
> +				spin_unlock(&ailp->ail_lock);
>  				xfs_cui_release(cuip);
> -				spin_lock(&ailp->xa_lock);
> +				spin_lock(&ailp->ail_lock);
>  				break;
>  			}
>  		}
> @@ -3697,7 +3697,7 @@ xlog_recover_cud_pass2(
>  	}
>  
>  	xfs_trans_ail_cursor_done(&cur);
> -	spin_unlock(&ailp->xa_lock);
> +	spin_unlock(&ailp->ail_lock);
>  
>  	return 0;
>  }
> @@ -3755,7 +3755,7 @@ xlog_recover_bui_pass2(
>  	}
>  	atomic_set(&buip->bui_next_extent, bui_formatp->bui_nextents);
>  
> -	spin_lock(&log->l_ailp->xa_lock);
> +	spin_lock(&log->l_ailp->ail_lock);
>  	/*
>  	 * The RUI has two references. One for the RUD and one for RUI to ensure
>  	 * it makes it into the AIL. Insert the RUI into the AIL directly and
> @@ -3796,7 +3796,7 @@ xlog_recover_bud_pass2(
>  	 * Search for the BUI with the id in the BUD format structure in the
>  	 * AIL.
>  	 */
> -	spin_lock(&ailp->xa_lock);
> +	spin_lock(&ailp->ail_lock);
>  	lip = xfs_trans_ail_cursor_first(ailp, &cur, 0);
>  	while (lip != NULL) {
>  		if (lip->li_type == XFS_LI_BUI) {
> @@ -3806,9 +3806,9 @@ xlog_recover_bud_pass2(
>  				 * Drop the BUD reference to the BUI. This
>  				 * removes the BUI from the AIL and frees it.
>  				 */
> -				spin_unlock(&ailp->xa_lock);
> +				spin_unlock(&ailp->ail_lock);
>  				xfs_bui_release(buip);
> -				spin_lock(&ailp->xa_lock);
> +				spin_lock(&ailp->ail_lock);
>  				break;
>  			}
>  		}
> @@ -3816,7 +3816,7 @@ xlog_recover_bud_pass2(
>  	}
>  
>  	xfs_trans_ail_cursor_done(&cur);
> -	spin_unlock(&ailp->xa_lock);
> +	spin_unlock(&ailp->ail_lock);
>  
>  	return 0;
>  }
> @@ -4649,9 +4649,9 @@ xlog_recover_process_efi(
>  	if (test_bit(XFS_EFI_RECOVERED, &efip->efi_flags))
>  		return 0;
>  
> -	spin_unlock(&ailp->xa_lock);
> +	spin_unlock(&ailp->ail_lock);
>  	error = xfs_efi_recover(mp, efip);
> -	spin_lock(&ailp->xa_lock);
> +	spin_lock(&ailp->ail_lock);
>  
>  	return error;
>  }
> @@ -4667,9 +4667,9 @@ xlog_recover_cancel_efi(
>  
>  	efip = container_of(lip, struct xfs_efi_log_item, efi_item);
>  
> -	spin_unlock(&ailp->xa_lock);
> +	spin_unlock(&ailp->ail_lock);
>  	xfs_efi_release(efip);
> -	spin_lock(&ailp->xa_lock);
> +	spin_lock(&ailp->ail_lock);
>  }
>  
>  /* Recover the RUI if necessary. */
> @@ -4689,9 +4689,9 @@ xlog_recover_process_rui(
>  	if (test_bit(XFS_RUI_RECOVERED, &ruip->rui_flags))
>  		return 0;
>  
> -	spin_unlock(&ailp->xa_lock);
> +	spin_unlock(&ailp->ail_lock);
>  	error = xfs_rui_recover(mp, ruip);
> -	spin_lock(&ailp->xa_lock);
> +	spin_lock(&ailp->ail_lock);
>  
>  	return error;
>  }
> @@ -4707,9 +4707,9 @@ xlog_recover_cancel_rui(
>  
>  	ruip = container_of(lip, struct xfs_rui_log_item, rui_item);
>  
> -	spin_unlock(&ailp->xa_lock);
> +	spin_unlock(&ailp->ail_lock);
>  	xfs_rui_release(ruip);
> -	spin_lock(&ailp->xa_lock);
> +	spin_lock(&ailp->ail_lock);
>  }
>  
>  /* Recover the CUI if necessary. */
> @@ -4730,9 +4730,9 @@ xlog_recover_process_cui(
>  	if (test_bit(XFS_CUI_RECOVERED, &cuip->cui_flags))
>  		return 0;
>  
> -	spin_unlock(&ailp->xa_lock);
> +	spin_unlock(&ailp->ail_lock);
>  	error = xfs_cui_recover(mp, cuip, dfops);
> -	spin_lock(&ailp->xa_lock);
> +	spin_lock(&ailp->ail_lock);
>  
>  	return error;
>  }
> @@ -4748,9 +4748,9 @@ xlog_recover_cancel_cui(
>  
>  	cuip = container_of(lip, struct xfs_cui_log_item, cui_item);
>  
> -	spin_unlock(&ailp->xa_lock);
> +	spin_unlock(&ailp->ail_lock);
>  	xfs_cui_release(cuip);
> -	spin_lock(&ailp->xa_lock);
> +	spin_lock(&ailp->ail_lock);
>  }
>  
>  /* Recover the BUI if necessary. */
> @@ -4771,9 +4771,9 @@ xlog_recover_process_bui(
>  	if (test_bit(XFS_BUI_RECOVERED, &buip->bui_flags))
>  		return 0;
>  
> -	spin_unlock(&ailp->xa_lock);
> +	spin_unlock(&ailp->ail_lock);
>  	error = xfs_bui_recover(mp, buip, dfops);
> -	spin_lock(&ailp->xa_lock);
> +	spin_lock(&ailp->ail_lock);
>  
>  	return error;
>  }
> @@ -4789,9 +4789,9 @@ xlog_recover_cancel_bui(
>  
>  	buip = container_of(lip, struct xfs_bui_log_item, bui_item);
>  
> -	spin_unlock(&ailp->xa_lock);
> +	spin_unlock(&ailp->ail_lock);
>  	xfs_bui_release(buip);
> -	spin_lock(&ailp->xa_lock);
> +	spin_lock(&ailp->ail_lock);
>  }
>  
>  /* Is this log item a deferred action intent? */
> @@ -4879,7 +4879,7 @@ xlog_recover_process_intents(
>  #endif
>  
>  	ailp = log->l_ailp;
> -	spin_lock(&ailp->xa_lock);
> +	spin_lock(&ailp->ail_lock);
>  	lip = xfs_trans_ail_cursor_first(ailp, &cur, 0);
>  #if defined(DEBUG) || defined(XFS_WARN)
>  	last_lsn = xlog_assign_lsn(log->l_curr_cycle, log->l_curr_block);
> @@ -4933,7 +4933,7 @@ xlog_recover_process_intents(
>  	}
>  out:
>  	xfs_trans_ail_cursor_done(&cur);
> -	spin_unlock(&ailp->xa_lock);
> +	spin_unlock(&ailp->ail_lock);
>  	if (error)
>  		xfs_defer_cancel(&dfops);
>  	else
> @@ -4956,7 +4956,7 @@ xlog_recover_cancel_intents(
>  	struct xfs_ail		*ailp;
>  
>  	ailp = log->l_ailp;
> -	spin_lock(&ailp->xa_lock);
> +	spin_lock(&ailp->ail_lock);
>  	lip = xfs_trans_ail_cursor_first(ailp, &cur, 0);
>  	while (lip != NULL) {
>  		/*
> @@ -4990,7 +4990,7 @@ xlog_recover_cancel_intents(
>  	}
>  
>  	xfs_trans_ail_cursor_done(&cur);
> -	spin_unlock(&ailp->xa_lock);
> +	spin_unlock(&ailp->ail_lock);
>  	return error;
>  }
>  
> diff --git a/fs/xfs/xfs_trans.c b/fs/xfs/xfs_trans.c
> index a87f657f59c9..756e01999c24 100644
> --- a/fs/xfs/xfs_trans.c
> +++ b/fs/xfs/xfs_trans.c
> @@ -781,8 +781,8 @@ xfs_log_item_batch_insert(
>  {
>  	int	i;
>  
> -	spin_lock(&ailp->xa_lock);
> -	/* xfs_trans_ail_update_bulk drops ailp->xa_lock */
> +	spin_lock(&ailp->ail_lock);
> +	/* xfs_trans_ail_update_bulk drops ailp->ail_lock */
>  	xfs_trans_ail_update_bulk(ailp, cur, log_items, nr_items, commit_lsn);
>  
>  	for (i = 0; i < nr_items; i++) {
> @@ -825,9 +825,9 @@ xfs_trans_committed_bulk(
>  	struct xfs_ail_cursor	cur;
>  	int			i = 0;
>  
> -	spin_lock(&ailp->xa_lock);
> +	spin_lock(&ailp->ail_lock);
>  	xfs_trans_ail_cursor_last(ailp, &cur, commit_lsn);
> -	spin_unlock(&ailp->xa_lock);
> +	spin_unlock(&ailp->ail_lock);
>  
>  	/* unpin all the log items */
>  	for (lv = log_vector; lv; lv = lv->lv_next ) {
> @@ -847,7 +847,7 @@ xfs_trans_committed_bulk(
>  		 * object into the AIL as we are in a shutdown situation.
>  		 */
>  		if (aborted) {
> -			ASSERT(XFS_FORCED_SHUTDOWN(ailp->xa_mount));
> +			ASSERT(XFS_FORCED_SHUTDOWN(ailp->ail_mount));
>  			lip->li_ops->iop_unpin(lip, 1);
>  			continue;
>  		}
> @@ -861,11 +861,11 @@ xfs_trans_committed_bulk(
>  			 * not affect the AIL cursor the bulk insert path is
>  			 * using.
>  			 */
> -			spin_lock(&ailp->xa_lock);
> +			spin_lock(&ailp->ail_lock);
>  			if (XFS_LSN_CMP(item_lsn, lip->li_lsn) > 0)
>  				xfs_trans_ail_update(ailp, lip, item_lsn);
>  			else
> -				spin_unlock(&ailp->xa_lock);
> +				spin_unlock(&ailp->ail_lock);
>  			lip->li_ops->iop_unpin(lip, 0);
>  			continue;
>  		}
> @@ -883,9 +883,9 @@ xfs_trans_committed_bulk(
>  	if (i)
>  		xfs_log_item_batch_insert(ailp, &cur, log_items, i, commit_lsn);
>  
> -	spin_lock(&ailp->xa_lock);
> +	spin_lock(&ailp->ail_lock);
>  	xfs_trans_ail_cursor_done(&cur);
> -	spin_unlock(&ailp->xa_lock);
> +	spin_unlock(&ailp->ail_lock);
>  }
>  
>  /*
> diff --git a/fs/xfs/xfs_trans_ail.c b/fs/xfs/xfs_trans_ail.c
> index cef89f7127d3..d4a2445215e6 100644
> --- a/fs/xfs/xfs_trans_ail.c
> +++ b/fs/xfs/xfs_trans_ail.c
> @@ -40,7 +40,7 @@ xfs_ail_check(
>  {
>  	xfs_log_item_t	*prev_lip;
>  
> -	if (list_empty(&ailp->xa_ail))
> +	if (list_empty(&ailp->ail_head))
>  		return;
>  
>  	/*
> @@ -48,11 +48,11 @@ xfs_ail_check(
>  	 */
>  	ASSERT((lip->li_flags & XFS_LI_IN_AIL) != 0);
>  	prev_lip = list_entry(lip->li_ail.prev, xfs_log_item_t, li_ail);
> -	if (&prev_lip->li_ail != &ailp->xa_ail)
> +	if (&prev_lip->li_ail != &ailp->ail_head)
>  		ASSERT(XFS_LSN_CMP(prev_lip->li_lsn, lip->li_lsn) <= 0);
>  
>  	prev_lip = list_entry(lip->li_ail.next, xfs_log_item_t, li_ail);
> -	if (&prev_lip->li_ail != &ailp->xa_ail)
> +	if (&prev_lip->li_ail != &ailp->ail_head)
>  		ASSERT(XFS_LSN_CMP(prev_lip->li_lsn, lip->li_lsn) >= 0);
>  
>  
> @@ -69,10 +69,10 @@ static xfs_log_item_t *
>  xfs_ail_max(
>  	struct xfs_ail  *ailp)
>  {
> -	if (list_empty(&ailp->xa_ail))
> +	if (list_empty(&ailp->ail_head))
>  		return NULL;
>  
> -	return list_entry(ailp->xa_ail.prev, xfs_log_item_t, li_ail);
> +	return list_entry(ailp->ail_head.prev, xfs_log_item_t, li_ail);
>  }
>  
>  /*
> @@ -84,7 +84,7 @@ xfs_ail_next(
>  	struct xfs_ail  *ailp,
>  	xfs_log_item_t  *lip)
>  {
> -	if (lip->li_ail.next == &ailp->xa_ail)
> +	if (lip->li_ail.next == &ailp->ail_head)
>  		return NULL;
>  
>  	return list_first_entry(&lip->li_ail, xfs_log_item_t, li_ail);
> @@ -105,11 +105,11 @@ xfs_ail_min_lsn(
>  	xfs_lsn_t	lsn = 0;
>  	xfs_log_item_t	*lip;
>  
> -	spin_lock(&ailp->xa_lock);
> +	spin_lock(&ailp->ail_lock);
>  	lip = xfs_ail_min(ailp);
>  	if (lip)
>  		lsn = lip->li_lsn;
> -	spin_unlock(&ailp->xa_lock);
> +	spin_unlock(&ailp->ail_lock);
>  
>  	return lsn;
>  }
> @@ -124,11 +124,11 @@ xfs_ail_max_lsn(
>  	xfs_lsn_t       lsn = 0;
>  	xfs_log_item_t  *lip;
>  
> -	spin_lock(&ailp->xa_lock);
> +	spin_lock(&ailp->ail_lock);
>  	lip = xfs_ail_max(ailp);
>  	if (lip)
>  		lsn = lip->li_lsn;
> -	spin_unlock(&ailp->xa_lock);
> +	spin_unlock(&ailp->ail_lock);
>  
>  	return lsn;
>  }
> @@ -146,7 +146,7 @@ xfs_trans_ail_cursor_init(
>  	struct xfs_ail_cursor	*cur)
>  {
>  	cur->item = NULL;
> -	list_add_tail(&cur->list, &ailp->xa_cursors);
> +	list_add_tail(&cur->list, &ailp->ail_cursors);
>  }
>  
>  /*
> @@ -194,7 +194,7 @@ xfs_trans_ail_cursor_clear(
>  {
>  	struct xfs_ail_cursor	*cur;
>  
> -	list_for_each_entry(cur, &ailp->xa_cursors, list) {
> +	list_for_each_entry(cur, &ailp->ail_cursors, list) {
>  		if (cur->item == lip)
>  			cur->item = (struct xfs_log_item *)
>  					((uintptr_t)cur->item | 1);
> @@ -222,7 +222,7 @@ xfs_trans_ail_cursor_first(
>  		goto out;
>  	}
>  
> -	list_for_each_entry(lip, &ailp->xa_ail, li_ail) {
> +	list_for_each_entry(lip, &ailp->ail_head, li_ail) {
>  		if (XFS_LSN_CMP(lip->li_lsn, lsn) >= 0)
>  			goto out;
>  	}
> @@ -241,7 +241,7 @@ __xfs_trans_ail_cursor_last(
>  {
>  	xfs_log_item_t		*lip;
>  
> -	list_for_each_entry_reverse(lip, &ailp->xa_ail, li_ail) {
> +	list_for_each_entry_reverse(lip, &ailp->ail_head, li_ail) {
>  		if (XFS_LSN_CMP(lip->li_lsn, lsn) <= 0)
>  			return lip;
>  	}
> @@ -310,7 +310,7 @@ xfs_ail_splice(
>  	if (lip)
>  		list_splice(list, &lip->li_ail);
>  	else
> -		list_splice(list, &ailp->xa_ail);
> +		list_splice(list, &ailp->ail_head);
>  }
>  
>  /*
> @@ -335,17 +335,17 @@ xfsaild_push_item(
>  	 * If log item pinning is enabled, skip the push and track the item as
>  	 * pinned. This can help induce head-behind-tail conditions.
>  	 */
> -	if (XFS_TEST_ERROR(false, ailp->xa_mount, XFS_ERRTAG_LOG_ITEM_PIN))
> +	if (XFS_TEST_ERROR(false, ailp->ail_mount, XFS_ERRTAG_LOG_ITEM_PIN))
>  		return XFS_ITEM_PINNED;
>  
> -	return lip->li_ops->iop_push(lip, &ailp->xa_buf_list);
> +	return lip->li_ops->iop_push(lip, &ailp->ail_buf_list);
>  }
>  
>  static long
>  xfsaild_push(
>  	struct xfs_ail		*ailp)
>  {
> -	xfs_mount_t		*mp = ailp->xa_mount;
> +	xfs_mount_t		*mp = ailp->ail_mount;
>  	struct xfs_ail_cursor	cur;
>  	xfs_log_item_t		*lip;
>  	xfs_lsn_t		lsn;
> @@ -360,30 +360,30 @@ xfsaild_push(
>  	 * buffers the last time we ran, force the log first and wait for it
>  	 * before pushing again.
>  	 */
> -	if (ailp->xa_log_flush && ailp->xa_last_pushed_lsn == 0 &&
> -	    (!list_empty_careful(&ailp->xa_buf_list) ||
> +	if (ailp->ail_log_flush && ailp->ail_last_pushed_lsn == 0 &&
> +	    (!list_empty_careful(&ailp->ail_buf_list) ||
>  	     xfs_ail_min_lsn(ailp))) {
> -		ailp->xa_log_flush = 0;
> +		ailp->ail_log_flush = 0;
>  
>  		XFS_STATS_INC(mp, xs_push_ail_flush);
>  		xfs_log_force(mp, XFS_LOG_SYNC);
>  	}
>  
> -	spin_lock(&ailp->xa_lock);
> +	spin_lock(&ailp->ail_lock);
>  
> -	/* barrier matches the xa_target update in xfs_ail_push() */
> +	/* barrier matches the ail_target update in xfs_ail_push() */
>  	smp_rmb();
> -	target = ailp->xa_target;
> -	ailp->xa_target_prev = target;
> +	target = ailp->ail_target;
> +	ailp->ail_target_prev = target;
>  
> -	lip = xfs_trans_ail_cursor_first(ailp, &cur, ailp->xa_last_pushed_lsn);
> +	lip = xfs_trans_ail_cursor_first(ailp, &cur, ailp->ail_last_pushed_lsn);
>  	if (!lip) {
>  		/*
>  		 * If the AIL is empty or our push has reached the end we are
>  		 * done now.
>  		 */
>  		xfs_trans_ail_cursor_done(&cur);
> -		spin_unlock(&ailp->xa_lock);
> +		spin_unlock(&ailp->ail_lock);
>  		goto out_done;
>  	}
>  
> @@ -404,7 +404,7 @@ xfsaild_push(
>  			XFS_STATS_INC(mp, xs_push_ail_success);
>  			trace_xfs_ail_push(lip);
>  
> -			ailp->xa_last_pushed_lsn = lsn;
> +			ailp->ail_last_pushed_lsn = lsn;
>  			break;
>  
>  		case XFS_ITEM_FLUSHING:
> @@ -423,7 +423,7 @@ xfsaild_push(
>  			trace_xfs_ail_flushing(lip);
>  
>  			flushing++;
> -			ailp->xa_last_pushed_lsn = lsn;
> +			ailp->ail_last_pushed_lsn = lsn;
>  			break;
>  
>  		case XFS_ITEM_PINNED:
> @@ -431,7 +431,7 @@ xfsaild_push(
>  			trace_xfs_ail_pinned(lip);
>  
>  			stuck++;
> -			ailp->xa_log_flush++;
> +			ailp->ail_log_flush++;
>  			break;
>  		case XFS_ITEM_LOCKED:
>  			XFS_STATS_INC(mp, xs_push_ail_locked);
> @@ -468,10 +468,10 @@ xfsaild_push(
>  		lsn = lip->li_lsn;
>  	}
>  	xfs_trans_ail_cursor_done(&cur);
> -	spin_unlock(&ailp->xa_lock);
> +	spin_unlock(&ailp->ail_lock);
>  
> -	if (xfs_buf_delwri_submit_nowait(&ailp->xa_buf_list))
> -		ailp->xa_log_flush++;
> +	if (xfs_buf_delwri_submit_nowait(&ailp->ail_buf_list))
> +		ailp->ail_log_flush++;
>  
>  	if (!count || XFS_LSN_CMP(lsn, target) >= 0) {
>  out_done:
> @@ -481,7 +481,7 @@ xfsaild_push(
>  		 * AIL before we start the next scan from the start of the AIL.
>  		 */
>  		tout = 50;
> -		ailp->xa_last_pushed_lsn = 0;
> +		ailp->ail_last_pushed_lsn = 0;
>  	} else if (((stuck + flushing) * 100) / count > 90) {
>  		/*
>  		 * Either there is a lot of contention on the AIL or we are
> @@ -494,7 +494,7 @@ xfsaild_push(
>  		 * the restart to issue a log force to unpin the stuck items.
>  		 */
>  		tout = 20;
> -		ailp->xa_last_pushed_lsn = 0;
> +		ailp->ail_last_pushed_lsn = 0;
>  	} else {
>  		/*
>  		 * Assume we have more work to do in a short while.
> @@ -536,26 +536,26 @@ xfsaild(
>  			break;
>  		}
>  
> -		spin_lock(&ailp->xa_lock);
> +		spin_lock(&ailp->ail_lock);
>  
>  		/*
>  		 * Idle if the AIL is empty and we are not racing with a target
>  		 * update. We check the AIL after we set the task to a sleep
> -		 * state to guarantee that we either catch an xa_target update
> +		 * state to guarantee that we either catch an ail_target update
>  		 * or that a wake_up resets the state to TASK_RUNNING.
>  		 * Otherwise, we run the risk of sleeping indefinitely.
>  		 *
> -		 * The barrier matches the xa_target update in xfs_ail_push().
> +		 * The barrier matches the ail_target update in xfs_ail_push().
>  		 */
>  		smp_rmb();
>  		if (!xfs_ail_min(ailp) &&
> -		    ailp->xa_target == ailp->xa_target_prev) {
> -			spin_unlock(&ailp->xa_lock);
> +		    ailp->ail_target == ailp->ail_target_prev) {
> +			spin_unlock(&ailp->ail_lock);
>  			freezable_schedule();
>  			tout = 0;
>  			continue;
>  		}
> -		spin_unlock(&ailp->xa_lock);
> +		spin_unlock(&ailp->ail_lock);
>  
>  		if (tout)
>  			freezable_schedule_timeout(msecs_to_jiffies(tout));
> @@ -592,8 +592,8 @@ xfs_ail_push(
>  	xfs_log_item_t	*lip;
>  
>  	lip = xfs_ail_min(ailp);
> -	if (!lip || XFS_FORCED_SHUTDOWN(ailp->xa_mount) ||
> -	    XFS_LSN_CMP(threshold_lsn, ailp->xa_target) <= 0)
> +	if (!lip || XFS_FORCED_SHUTDOWN(ailp->ail_mount) ||
> +	    XFS_LSN_CMP(threshold_lsn, ailp->ail_target) <= 0)
>  		return;
>  
>  	/*
> @@ -601,10 +601,10 @@ xfs_ail_push(
>  	 * the XFS_AIL_PUSHING_BIT.
>  	 */
>  	smp_wmb();
> -	xfs_trans_ail_copy_lsn(ailp, &ailp->xa_target, &threshold_lsn);
> +	xfs_trans_ail_copy_lsn(ailp, &ailp->ail_target, &threshold_lsn);
>  	smp_wmb();
>  
> -	wake_up_process(ailp->xa_task);
> +	wake_up_process(ailp->ail_task);
>  }
>  
>  /*
> @@ -630,18 +630,18 @@ xfs_ail_push_all_sync(
>  	struct xfs_log_item	*lip;
>  	DEFINE_WAIT(wait);
>  
> -	spin_lock(&ailp->xa_lock);
> +	spin_lock(&ailp->ail_lock);
>  	while ((lip = xfs_ail_max(ailp)) != NULL) {
> -		prepare_to_wait(&ailp->xa_empty, &wait, TASK_UNINTERRUPTIBLE);
> -		ailp->xa_target = lip->li_lsn;
> -		wake_up_process(ailp->xa_task);
> -		spin_unlock(&ailp->xa_lock);
> +		prepare_to_wait(&ailp->ail_empty, &wait, TASK_UNINTERRUPTIBLE);
> +		ailp->ail_target = lip->li_lsn;
> +		wake_up_process(ailp->ail_task);
> +		spin_unlock(&ailp->ail_lock);
>  		schedule();
> -		spin_lock(&ailp->xa_lock);
> +		spin_lock(&ailp->ail_lock);
>  	}
> -	spin_unlock(&ailp->xa_lock);
> +	spin_unlock(&ailp->ail_lock);
>  
> -	finish_wait(&ailp->xa_empty, &wait);
> +	finish_wait(&ailp->ail_empty, &wait);
>  }
>  
>  /*
> @@ -672,7 +672,7 @@ xfs_trans_ail_update_bulk(
>  	struct xfs_ail_cursor	*cur,
>  	struct xfs_log_item	**log_items,
>  	int			nr_items,
> -	xfs_lsn_t		lsn) __releases(ailp->xa_lock)
> +	xfs_lsn_t		lsn) __releases(ailp->ail_lock)
>  {
>  	xfs_log_item_t		*mlip;
>  	int			mlip_changed = 0;
> @@ -705,13 +705,13 @@ xfs_trans_ail_update_bulk(
>  		xfs_ail_splice(ailp, cur, &tmp, lsn);
>  
>  	if (mlip_changed) {
> -		if (!XFS_FORCED_SHUTDOWN(ailp->xa_mount))
> -			xlog_assign_tail_lsn_locked(ailp->xa_mount);
> -		spin_unlock(&ailp->xa_lock);
> +		if (!XFS_FORCED_SHUTDOWN(ailp->ail_mount))
> +			xlog_assign_tail_lsn_locked(ailp->ail_mount);
> +		spin_unlock(&ailp->ail_lock);
>  
> -		xfs_log_space_wake(ailp->xa_mount);
> +		xfs_log_space_wake(ailp->ail_mount);
>  	} else {
> -		spin_unlock(&ailp->xa_lock);
> +		spin_unlock(&ailp->ail_lock);
>  	}
>  }
>  
> @@ -756,13 +756,13 @@ void
>  xfs_trans_ail_delete(
>  	struct xfs_ail		*ailp,
>  	struct xfs_log_item	*lip,
> -	int			shutdown_type) __releases(ailp->xa_lock)
> +	int			shutdown_type) __releases(ailp->ail_lock)
>  {
> -	struct xfs_mount	*mp = ailp->xa_mount;
> +	struct xfs_mount	*mp = ailp->ail_mount;
>  	bool			mlip_changed;
>  
>  	if (!(lip->li_flags & XFS_LI_IN_AIL)) {
> -		spin_unlock(&ailp->xa_lock);
> +		spin_unlock(&ailp->ail_lock);
>  		if (!XFS_FORCED_SHUTDOWN(mp)) {
>  			xfs_alert_tag(mp, XFS_PTAG_AILDELETE,
>  	"%s: attempting to delete a log item that is not in the AIL",
> @@ -776,13 +776,13 @@ xfs_trans_ail_delete(
>  	if (mlip_changed) {
>  		if (!XFS_FORCED_SHUTDOWN(mp))
>  			xlog_assign_tail_lsn_locked(mp);
> -		if (list_empty(&ailp->xa_ail))
> -			wake_up_all(&ailp->xa_empty);
> +		if (list_empty(&ailp->ail_head))
> +			wake_up_all(&ailp->ail_empty);
>  	}
>  
> -	spin_unlock(&ailp->xa_lock);
> +	spin_unlock(&ailp->ail_lock);
>  	if (mlip_changed)
> -		xfs_log_space_wake(ailp->xa_mount);
> +		xfs_log_space_wake(ailp->ail_mount);
>  }
>  
>  int
> @@ -795,16 +795,16 @@ xfs_trans_ail_init(
>  	if (!ailp)
>  		return -ENOMEM;
>  
> -	ailp->xa_mount = mp;
> -	INIT_LIST_HEAD(&ailp->xa_ail);
> -	INIT_LIST_HEAD(&ailp->xa_cursors);
> -	spin_lock_init(&ailp->xa_lock);
> -	INIT_LIST_HEAD(&ailp->xa_buf_list);
> -	init_waitqueue_head(&ailp->xa_empty);
> +	ailp->ail_mount = mp;
> +	INIT_LIST_HEAD(&ailp->ail_head);
> +	INIT_LIST_HEAD(&ailp->ail_cursors);
> +	spin_lock_init(&ailp->ail_lock);
> +	INIT_LIST_HEAD(&ailp->ail_buf_list);
> +	init_waitqueue_head(&ailp->ail_empty);
>  
> -	ailp->xa_task = kthread_run(xfsaild, ailp, "xfsaild/%s",
> -			ailp->xa_mount->m_fsname);
> -	if (IS_ERR(ailp->xa_task))
> +	ailp->ail_task = kthread_run(xfsaild, ailp, "xfsaild/%s",
> +			ailp->ail_mount->m_fsname);
> +	if (IS_ERR(ailp->ail_task))
>  		goto out_free_ailp;
>  
>  	mp->m_ail = ailp;
> @@ -821,6 +821,6 @@ xfs_trans_ail_destroy(
>  {
>  	struct xfs_ail	*ailp = mp->m_ail;
>  
> -	kthread_stop(ailp->xa_task);
> +	kthread_stop(ailp->ail_task);
>  	kmem_free(ailp);
>  }
> diff --git a/fs/xfs/xfs_trans_buf.c b/fs/xfs/xfs_trans_buf.c
> index 3ba7a96a8abd..b8871bcfe00b 100644
> --- a/fs/xfs/xfs_trans_buf.c
> +++ b/fs/xfs/xfs_trans_buf.c
> @@ -429,8 +429,8 @@ xfs_trans_brelse(xfs_trans_t	*tp,
>  	 * If the fs has shutdown and we dropped the last reference, it may fall
>  	 * on us to release a (possibly dirty) bli if it never made it to the
>  	 * AIL (e.g., the aborted unpin already happened and didn't release it
> -	 * due to our reference). Since we're already shutdown and need xa_lock,
> -	 * just force remove from the AIL and release the bli here.
> +	 * due to our reference). Since we're already shutdown and need
> +	 * ail_lock, just force remove from the AIL and release the bli here.
>  	 */
>  	if (XFS_FORCED_SHUTDOWN(tp->t_mountp) && freed) {
>  		xfs_trans_ail_remove(&bip->bli_item, SHUTDOWN_LOG_IO_ERROR);
> diff --git a/fs/xfs/xfs_trans_priv.h b/fs/xfs/xfs_trans_priv.h
> index b317a3644c00..be24b0c8a332 100644
> --- a/fs/xfs/xfs_trans_priv.h
> +++ b/fs/xfs/xfs_trans_priv.h
> @@ -65,17 +65,17 @@ struct xfs_ail_cursor {
>   * Eventually we need to drive the locking in here as well.
>   */
>  struct xfs_ail {
> -	struct xfs_mount	*xa_mount;
> -	struct task_struct	*xa_task;
> -	struct list_head	xa_ail;
> -	xfs_lsn_t		xa_target;
> -	xfs_lsn_t		xa_target_prev;
> -	struct list_head	xa_cursors;
> -	spinlock_t		xa_lock;
> -	xfs_lsn_t		xa_last_pushed_lsn;
> -	int			xa_log_flush;
> -	struct list_head	xa_buf_list;
> -	wait_queue_head_t	xa_empty;
> +	struct xfs_mount	*ail_mount;
> +	struct task_struct	*ail_task;
> +	struct list_head	ail_head;
> +	xfs_lsn_t		ail_target;
> +	xfs_lsn_t		ail_target_prev;
> +	struct list_head	ail_cursors;
> +	spinlock_t		ail_lock;
> +	xfs_lsn_t		ail_last_pushed_lsn;
> +	int			ail_log_flush;
> +	struct list_head	ail_buf_list;
> +	wait_queue_head_t	ail_empty;
>  };
>  
>  /*
> @@ -84,7 +84,7 @@ struct xfs_ail {
>  void	xfs_trans_ail_update_bulk(struct xfs_ail *ailp,
>  				struct xfs_ail_cursor *cur,
>  				struct xfs_log_item **log_items, int nr_items,
> -				xfs_lsn_t lsn) __releases(ailp->xa_lock);
> +				xfs_lsn_t lsn) __releases(ailp->ail_lock);
>  /*
>   * Return a pointer to the first item in the AIL.  If the AIL is empty, then
>   * return NULL.
> @@ -93,7 +93,7 @@ static inline struct xfs_log_item *
>  xfs_ail_min(
>  	struct xfs_ail  *ailp)
>  {
> -	return list_first_entry_or_null(&ailp->xa_ail, struct xfs_log_item,
> +	return list_first_entry_or_null(&ailp->ail_head, struct xfs_log_item,
>  					li_ail);
>  }
>  
> @@ -101,14 +101,14 @@ static inline void
>  xfs_trans_ail_update(
>  	struct xfs_ail		*ailp,
>  	struct xfs_log_item	*lip,
> -	xfs_lsn_t		lsn) __releases(ailp->xa_lock)
> +	xfs_lsn_t		lsn) __releases(ailp->ail_lock)
>  {
>  	xfs_trans_ail_update_bulk(ailp, NULL, &lip, 1, lsn);
>  }
>  
>  bool xfs_ail_delete_one(struct xfs_ail *ailp, struct xfs_log_item *lip);
>  void xfs_trans_ail_delete(struct xfs_ail *ailp, struct xfs_log_item *lip,
> -		int shutdown_type) __releases(ailp->xa_lock);
> +		int shutdown_type) __releases(ailp->ail_lock);
>  
>  static inline void
>  xfs_trans_ail_remove(
> @@ -117,12 +117,12 @@ xfs_trans_ail_remove(
>  {
>  	struct xfs_ail		*ailp = lip->li_ailp;
>  
> -	spin_lock(&ailp->xa_lock);
> +	spin_lock(&ailp->ail_lock);
>  	/* xfs_trans_ail_delete() drops the AIL lock */
>  	if (lip->li_flags & XFS_LI_IN_AIL)
>  		xfs_trans_ail_delete(ailp, lip, shutdown_type);
>  	else
> -		spin_unlock(&ailp->xa_lock);
> +		spin_unlock(&ailp->ail_lock);
>  }
>  
>  void			xfs_ail_push(struct xfs_ail *, xfs_lsn_t);
> @@ -149,9 +149,9 @@ xfs_trans_ail_copy_lsn(
>  	xfs_lsn_t	*src)
>  {
>  	ASSERT(sizeof(xfs_lsn_t) == 8);	/* don't lock if it shrinks */
> -	spin_lock(&ailp->xa_lock);
> +	spin_lock(&ailp->ail_lock);
>  	*dst = *src;
> -	spin_unlock(&ailp->xa_lock);
> +	spin_unlock(&ailp->ail_lock);
>  }
>  #else
>  static inline void
> @@ -172,7 +172,7 @@ xfs_clear_li_failed(
>  	struct xfs_buf	*bp = lip->li_buf;
>  
>  	ASSERT(lip->li_flags & XFS_LI_IN_AIL);
> -	lockdep_assert_held(&lip->li_ailp->xa_lock);
> +	lockdep_assert_held(&lip->li_ailp->ail_lock);
>  
>  	if (lip->li_flags & XFS_LI_FAILED) {
>  		lip->li_flags &= ~XFS_LI_FAILED;
> @@ -186,7 +186,7 @@ xfs_set_li_failed(
>  	struct xfs_log_item	*lip,
>  	struct xfs_buf		*bp)
>  {
> -	lockdep_assert_held(&lip->li_ailp->xa_lock);
> +	lockdep_assert_held(&lip->li_ailp->ail_lock);
>  
>  	if (!(lip->li_flags & XFS_LI_FAILED)) {
>  		xfs_buf_hold(bp);
> -- 
> 2.15.1
> 
> --
> To unsubscribe from this list: send the line "unsubscribe linux-xfs" in
> the body of a message to majordomo@vger.kernel.org
> More majordomo info at  http://vger.kernel.org/majordomo-info.html
--
To unsubscribe from this list: send the line "unsubscribe linux-btrfs" in
the body of a message to majordomo@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
diff mbox

Patch

diff --git a/fs/xfs/xfs_buf_item.c b/fs/xfs/xfs_buf_item.c
index e0a0af0946f2..6c5035544a93 100644
--- a/fs/xfs/xfs_buf_item.c
+++ b/fs/xfs/xfs_buf_item.c
@@ -459,7 +459,7 @@  xfs_buf_item_unpin(
 			bp->b_fspriv = NULL;
 			bp->b_iodone = NULL;
 		} else {
-			spin_lock(&ailp->xa_lock);
+			spin_lock(&ailp->ail_lock);
 			xfs_trans_ail_delete(ailp, lip, SHUTDOWN_LOG_IO_ERROR);
 			xfs_buf_item_relse(bp);
 			ASSERT(bp->b_fspriv == NULL);
@@ -1056,13 +1056,13 @@  xfs_buf_do_callbacks_fail(
 	struct xfs_log_item	*lip = bp->b_fspriv;
 	struct xfs_ail		*ailp = lip->li_ailp;
 
-	spin_lock(&ailp->xa_lock);
+	spin_lock(&ailp->ail_lock);
 	for (; lip; lip = next) {
 		next = lip->li_bio_list;
 		if (lip->li_ops->iop_error)
 			lip->li_ops->iop_error(lip, bp);
 	}
-	spin_unlock(&ailp->xa_lock);
+	spin_unlock(&ailp->ail_lock);
 }
 
 static bool
@@ -1215,7 +1215,7 @@  xfs_buf_iodone(
 	 *
 	 * Either way, AIL is useless if we're forcing a shutdown.
 	 */
-	spin_lock(&ailp->xa_lock);
+	spin_lock(&ailp->ail_lock);
 	xfs_trans_ail_delete(ailp, lip, SHUTDOWN_CORRUPT_INCORE);
 	xfs_buf_item_free(BUF_ITEM(lip));
 }
@@ -1236,7 +1236,7 @@  xfs_buf_resubmit_failed_buffers(
 	/*
 	 * Clear XFS_LI_FAILED flag from all items before resubmit
 	 *
-	 * XFS_LI_FAILED set/clear is protected by xa_lock, caller  this
+	 * XFS_LI_FAILED set/clear is protected by ail_lock, caller  this
 	 * function already have it acquired
 	 */
 	for (; lip; lip = next) {
diff --git a/fs/xfs/xfs_dquot.c b/fs/xfs/xfs_dquot.c
index f248708c10ff..e2a466df5dd1 100644
--- a/fs/xfs/xfs_dquot.c
+++ b/fs/xfs/xfs_dquot.c
@@ -974,7 +974,7 @@  xfs_qm_dqflush_done(
 	     (lip->li_flags & XFS_LI_FAILED))) {
 
 		/* xfs_trans_ail_delete() drops the AIL lock. */
-		spin_lock(&ailp->xa_lock);
+		spin_lock(&ailp->ail_lock);
 		if (lip->li_lsn == qip->qli_flush_lsn) {
 			xfs_trans_ail_delete(ailp, lip, SHUTDOWN_CORRUPT_INCORE);
 		} else {
@@ -984,7 +984,7 @@  xfs_qm_dqflush_done(
 			 */
 			if (lip->li_flags & XFS_LI_FAILED)
 				xfs_clear_li_failed(lip);
-			spin_unlock(&ailp->xa_lock);
+			spin_unlock(&ailp->ail_lock);
 		}
 	}
 
diff --git a/fs/xfs/xfs_dquot_item.c b/fs/xfs/xfs_dquot_item.c
index 664dea105e76..62637a226601 100644
--- a/fs/xfs/xfs_dquot_item.c
+++ b/fs/xfs/xfs_dquot_item.c
@@ -160,8 +160,9 @@  xfs_dquot_item_error(
 STATIC uint
 xfs_qm_dquot_logitem_push(
 	struct xfs_log_item	*lip,
-	struct list_head	*buffer_list) __releases(&lip->li_ailp->xa_lock)
-					      __acquires(&lip->li_ailp->xa_lock)
+	struct list_head	*buffer_list)
+		__releases(&lip->li_ailp->ail_lock)
+		__acquires(&lip->li_ailp->ail_lock)
 {
 	struct xfs_dquot	*dqp = DQUOT_ITEM(lip)->qli_dquot;
 	struct xfs_buf		*bp = lip->li_buf;
@@ -208,7 +209,7 @@  xfs_qm_dquot_logitem_push(
 		goto out_unlock;
 	}
 
-	spin_unlock(&lip->li_ailp->xa_lock);
+	spin_unlock(&lip->li_ailp->ail_lock);
 
 	error = xfs_qm_dqflush(dqp, &bp);
 	if (error) {
@@ -220,7 +221,7 @@  xfs_qm_dquot_logitem_push(
 		xfs_buf_relse(bp);
 	}
 
-	spin_lock(&lip->li_ailp->xa_lock);
+	spin_lock(&lip->li_ailp->ail_lock);
 out_unlock:
 	xfs_dqunlock(dqp);
 	return rval;
@@ -403,7 +404,7 @@  xfs_qm_qoffend_logitem_committed(
 	 * Delete the qoff-start logitem from the AIL.
 	 * xfs_trans_ail_delete() drops the AIL lock.
 	 */
-	spin_lock(&ailp->xa_lock);
+	spin_lock(&ailp->ail_lock);
 	xfs_trans_ail_delete(ailp, &qfs->qql_item, SHUTDOWN_LOG_IO_ERROR);
 
 	kmem_free(qfs->qql_item.li_lv_shadow);
diff --git a/fs/xfs/xfs_inode_item.c b/fs/xfs/xfs_inode_item.c
index 6ee5c3bf19ad..071acd4249a0 100644
--- a/fs/xfs/xfs_inode_item.c
+++ b/fs/xfs/xfs_inode_item.c
@@ -501,8 +501,8 @@  STATIC uint
 xfs_inode_item_push(
 	struct xfs_log_item	*lip,
 	struct list_head	*buffer_list)
-		__releases(&lip->li_ailp->xa_lock)
-		__acquires(&lip->li_ailp->xa_lock)
+		__releases(&lip->li_ailp->ail_lock)
+		__acquires(&lip->li_ailp->ail_lock)
 {
 	struct xfs_inode_log_item *iip = INODE_ITEM(lip);
 	struct xfs_inode	*ip = iip->ili_inode;
@@ -561,7 +561,7 @@  xfs_inode_item_push(
 	ASSERT(iip->ili_fields != 0 || XFS_FORCED_SHUTDOWN(ip->i_mount));
 	ASSERT(iip->ili_logged == 0 || XFS_FORCED_SHUTDOWN(ip->i_mount));
 
-	spin_unlock(&lip->li_ailp->xa_lock);
+	spin_unlock(&lip->li_ailp->ail_lock);
 
 	error = xfs_iflush(ip, &bp);
 	if (!error) {
@@ -570,7 +570,7 @@  xfs_inode_item_push(
 		xfs_buf_relse(bp);
 	}
 
-	spin_lock(&lip->li_ailp->xa_lock);
+	spin_lock(&lip->li_ailp->ail_lock);
 out_unlock:
 	xfs_iunlock(ip, XFS_ILOCK_SHARED);
 	return rval;
@@ -774,7 +774,7 @@  xfs_iflush_done(
 		bool			mlip_changed = false;
 
 		/* this is an opencoded batch version of xfs_trans_ail_delete */
-		spin_lock(&ailp->xa_lock);
+		spin_lock(&ailp->ail_lock);
 		for (blip = lip; blip; blip = blip->li_bio_list) {
 			if (INODE_ITEM(blip)->ili_logged &&
 			    blip->li_lsn == INODE_ITEM(blip)->ili_flush_lsn)
@@ -785,15 +785,15 @@  xfs_iflush_done(
 		}
 
 		if (mlip_changed) {
-			if (!XFS_FORCED_SHUTDOWN(ailp->xa_mount))
-				xlog_assign_tail_lsn_locked(ailp->xa_mount);
-			if (list_empty(&ailp->xa_ail))
-				wake_up_all(&ailp->xa_empty);
+			if (!XFS_FORCED_SHUTDOWN(ailp->ail_mount))
+				xlog_assign_tail_lsn_locked(ailp->ail_mount);
+			if (list_empty(&ailp->ail_head))
+				wake_up_all(&ailp->ail_empty);
 		}
-		spin_unlock(&ailp->xa_lock);
+		spin_unlock(&ailp->ail_lock);
 
 		if (mlip_changed)
-			xfs_log_space_wake(ailp->xa_mount);
+			xfs_log_space_wake(ailp->ail_mount);
 	}
 
 	/*
diff --git a/fs/xfs/xfs_log.c b/fs/xfs/xfs_log.c
index a503af96d780..7148625eebf2 100644
--- a/fs/xfs/xfs_log.c
+++ b/fs/xfs/xfs_log.c
@@ -1148,7 +1148,7 @@  xlog_assign_tail_lsn_locked(
 	struct xfs_log_item	*lip;
 	xfs_lsn_t		tail_lsn;
 
-	assert_spin_locked(&mp->m_ail->xa_lock);
+	assert_spin_locked(&mp->m_ail->ail_lock);
 
 	/*
 	 * To make sure we always have a valid LSN for the log tail we keep
@@ -1171,9 +1171,9 @@  xlog_assign_tail_lsn(
 {
 	xfs_lsn_t		tail_lsn;
 
-	spin_lock(&mp->m_ail->xa_lock);
+	spin_lock(&mp->m_ail->ail_lock);
 	tail_lsn = xlog_assign_tail_lsn_locked(mp);
-	spin_unlock(&mp->m_ail->xa_lock);
+	spin_unlock(&mp->m_ail->ail_lock);
 
 	return tail_lsn;
 }
diff --git a/fs/xfs/xfs_log_recover.c b/fs/xfs/xfs_log_recover.c
index 28d1abfe835e..d871761626fb 100644
--- a/fs/xfs/xfs_log_recover.c
+++ b/fs/xfs/xfs_log_recover.c
@@ -3424,7 +3424,7 @@  xlog_recover_efi_pass2(
 	}
 	atomic_set(&efip->efi_next_extent, efi_formatp->efi_nextents);
 
-	spin_lock(&log->l_ailp->xa_lock);
+	spin_lock(&log->l_ailp->ail_lock);
 	/*
 	 * The EFI has two references. One for the EFD and one for EFI to ensure
 	 * it makes it into the AIL. Insert the EFI into the AIL directly and
@@ -3467,7 +3467,7 @@  xlog_recover_efd_pass2(
 	 * Search for the EFI with the id in the EFD format structure in the
 	 * AIL.
 	 */
-	spin_lock(&ailp->xa_lock);
+	spin_lock(&ailp->ail_lock);
 	lip = xfs_trans_ail_cursor_first(ailp, &cur, 0);
 	while (lip != NULL) {
 		if (lip->li_type == XFS_LI_EFI) {
@@ -3477,9 +3477,9 @@  xlog_recover_efd_pass2(
 				 * Drop the EFD reference to the EFI. This
 				 * removes the EFI from the AIL and frees it.
 				 */
-				spin_unlock(&ailp->xa_lock);
+				spin_unlock(&ailp->ail_lock);
 				xfs_efi_release(efip);
-				spin_lock(&ailp->xa_lock);
+				spin_lock(&ailp->ail_lock);
 				break;
 			}
 		}
@@ -3487,7 +3487,7 @@  xlog_recover_efd_pass2(
 	}
 
 	xfs_trans_ail_cursor_done(&cur);
-	spin_unlock(&ailp->xa_lock);
+	spin_unlock(&ailp->ail_lock);
 
 	return 0;
 }
@@ -3520,7 +3520,7 @@  xlog_recover_rui_pass2(
 	}
 	atomic_set(&ruip->rui_next_extent, rui_formatp->rui_nextents);
 
-	spin_lock(&log->l_ailp->xa_lock);
+	spin_lock(&log->l_ailp->ail_lock);
 	/*
 	 * The RUI has two references. One for the RUD and one for RUI to ensure
 	 * it makes it into the AIL. Insert the RUI into the AIL directly and
@@ -3560,7 +3560,7 @@  xlog_recover_rud_pass2(
 	 * Search for the RUI with the id in the RUD format structure in the
 	 * AIL.
 	 */
-	spin_lock(&ailp->xa_lock);
+	spin_lock(&ailp->ail_lock);
 	lip = xfs_trans_ail_cursor_first(ailp, &cur, 0);
 	while (lip != NULL) {
 		if (lip->li_type == XFS_LI_RUI) {
@@ -3570,9 +3570,9 @@  xlog_recover_rud_pass2(
 				 * Drop the RUD reference to the RUI. This
 				 * removes the RUI from the AIL and frees it.
 				 */
-				spin_unlock(&ailp->xa_lock);
+				spin_unlock(&ailp->ail_lock);
 				xfs_rui_release(ruip);
-				spin_lock(&ailp->xa_lock);
+				spin_lock(&ailp->ail_lock);
 				break;
 			}
 		}
@@ -3580,7 +3580,7 @@  xlog_recover_rud_pass2(
 	}
 
 	xfs_trans_ail_cursor_done(&cur);
-	spin_unlock(&ailp->xa_lock);
+	spin_unlock(&ailp->ail_lock);
 
 	return 0;
 }
@@ -3636,7 +3636,7 @@  xlog_recover_cui_pass2(
 	}
 	atomic_set(&cuip->cui_next_extent, cui_formatp->cui_nextents);
 
-	spin_lock(&log->l_ailp->xa_lock);
+	spin_lock(&log->l_ailp->ail_lock);
 	/*
 	 * The CUI has two references. One for the CUD and one for CUI to ensure
 	 * it makes it into the AIL. Insert the CUI into the AIL directly and
@@ -3677,7 +3677,7 @@  xlog_recover_cud_pass2(
 	 * Search for the CUI with the id in the CUD format structure in the
 	 * AIL.
 	 */
-	spin_lock(&ailp->xa_lock);
+	spin_lock(&ailp->ail_lock);
 	lip = xfs_trans_ail_cursor_first(ailp, &cur, 0);
 	while (lip != NULL) {
 		if (lip->li_type == XFS_LI_CUI) {
@@ -3687,9 +3687,9 @@  xlog_recover_cud_pass2(
 				 * Drop the CUD reference to the CUI. This
 				 * removes the CUI from the AIL and frees it.
 				 */
-				spin_unlock(&ailp->xa_lock);
+				spin_unlock(&ailp->ail_lock);
 				xfs_cui_release(cuip);
-				spin_lock(&ailp->xa_lock);
+				spin_lock(&ailp->ail_lock);
 				break;
 			}
 		}
@@ -3697,7 +3697,7 @@  xlog_recover_cud_pass2(
 	}
 
 	xfs_trans_ail_cursor_done(&cur);
-	spin_unlock(&ailp->xa_lock);
+	spin_unlock(&ailp->ail_lock);
 
 	return 0;
 }
@@ -3755,7 +3755,7 @@  xlog_recover_bui_pass2(
 	}
 	atomic_set(&buip->bui_next_extent, bui_formatp->bui_nextents);
 
-	spin_lock(&log->l_ailp->xa_lock);
+	spin_lock(&log->l_ailp->ail_lock);
 	/*
 	 * The RUI has two references. One for the RUD and one for RUI to ensure
 	 * it makes it into the AIL. Insert the RUI into the AIL directly and
@@ -3796,7 +3796,7 @@  xlog_recover_bud_pass2(
 	 * Search for the BUI with the id in the BUD format structure in the
 	 * AIL.
 	 */
-	spin_lock(&ailp->xa_lock);
+	spin_lock(&ailp->ail_lock);
 	lip = xfs_trans_ail_cursor_first(ailp, &cur, 0);
 	while (lip != NULL) {
 		if (lip->li_type == XFS_LI_BUI) {
@@ -3806,9 +3806,9 @@  xlog_recover_bud_pass2(
 				 * Drop the BUD reference to the BUI. This
 				 * removes the BUI from the AIL and frees it.
 				 */
-				spin_unlock(&ailp->xa_lock);
+				spin_unlock(&ailp->ail_lock);
 				xfs_bui_release(buip);
-				spin_lock(&ailp->xa_lock);
+				spin_lock(&ailp->ail_lock);
 				break;
 			}
 		}
@@ -3816,7 +3816,7 @@  xlog_recover_bud_pass2(
 	}
 
 	xfs_trans_ail_cursor_done(&cur);
-	spin_unlock(&ailp->xa_lock);
+	spin_unlock(&ailp->ail_lock);
 
 	return 0;
 }
@@ -4649,9 +4649,9 @@  xlog_recover_process_efi(
 	if (test_bit(XFS_EFI_RECOVERED, &efip->efi_flags))
 		return 0;
 
-	spin_unlock(&ailp->xa_lock);
+	spin_unlock(&ailp->ail_lock);
 	error = xfs_efi_recover(mp, efip);
-	spin_lock(&ailp->xa_lock);
+	spin_lock(&ailp->ail_lock);
 
 	return error;
 }
@@ -4667,9 +4667,9 @@  xlog_recover_cancel_efi(
 
 	efip = container_of(lip, struct xfs_efi_log_item, efi_item);
 
-	spin_unlock(&ailp->xa_lock);
+	spin_unlock(&ailp->ail_lock);
 	xfs_efi_release(efip);
-	spin_lock(&ailp->xa_lock);
+	spin_lock(&ailp->ail_lock);
 }
 
 /* Recover the RUI if necessary. */
@@ -4689,9 +4689,9 @@  xlog_recover_process_rui(
 	if (test_bit(XFS_RUI_RECOVERED, &ruip->rui_flags))
 		return 0;
 
-	spin_unlock(&ailp->xa_lock);
+	spin_unlock(&ailp->ail_lock);
 	error = xfs_rui_recover(mp, ruip);
-	spin_lock(&ailp->xa_lock);
+	spin_lock(&ailp->ail_lock);
 
 	return error;
 }
@@ -4707,9 +4707,9 @@  xlog_recover_cancel_rui(
 
 	ruip = container_of(lip, struct xfs_rui_log_item, rui_item);
 
-	spin_unlock(&ailp->xa_lock);
+	spin_unlock(&ailp->ail_lock);
 	xfs_rui_release(ruip);
-	spin_lock(&ailp->xa_lock);
+	spin_lock(&ailp->ail_lock);
 }
 
 /* Recover the CUI if necessary. */
@@ -4730,9 +4730,9 @@  xlog_recover_process_cui(
 	if (test_bit(XFS_CUI_RECOVERED, &cuip->cui_flags))
 		return 0;
 
-	spin_unlock(&ailp->xa_lock);
+	spin_unlock(&ailp->ail_lock);
 	error = xfs_cui_recover(mp, cuip, dfops);
-	spin_lock(&ailp->xa_lock);
+	spin_lock(&ailp->ail_lock);
 
 	return error;
 }
@@ -4748,9 +4748,9 @@  xlog_recover_cancel_cui(
 
 	cuip = container_of(lip, struct xfs_cui_log_item, cui_item);
 
-	spin_unlock(&ailp->xa_lock);
+	spin_unlock(&ailp->ail_lock);
 	xfs_cui_release(cuip);
-	spin_lock(&ailp->xa_lock);
+	spin_lock(&ailp->ail_lock);
 }
 
 /* Recover the BUI if necessary. */
@@ -4771,9 +4771,9 @@  xlog_recover_process_bui(
 	if (test_bit(XFS_BUI_RECOVERED, &buip->bui_flags))
 		return 0;
 
-	spin_unlock(&ailp->xa_lock);
+	spin_unlock(&ailp->ail_lock);
 	error = xfs_bui_recover(mp, buip, dfops);
-	spin_lock(&ailp->xa_lock);
+	spin_lock(&ailp->ail_lock);
 
 	return error;
 }
@@ -4789,9 +4789,9 @@  xlog_recover_cancel_bui(
 
 	buip = container_of(lip, struct xfs_bui_log_item, bui_item);
 
-	spin_unlock(&ailp->xa_lock);
+	spin_unlock(&ailp->ail_lock);
 	xfs_bui_release(buip);
-	spin_lock(&ailp->xa_lock);
+	spin_lock(&ailp->ail_lock);
 }
 
 /* Is this log item a deferred action intent? */
@@ -4879,7 +4879,7 @@  xlog_recover_process_intents(
 #endif
 
 	ailp = log->l_ailp;
-	spin_lock(&ailp->xa_lock);
+	spin_lock(&ailp->ail_lock);
 	lip = xfs_trans_ail_cursor_first(ailp, &cur, 0);
 #if defined(DEBUG) || defined(XFS_WARN)
 	last_lsn = xlog_assign_lsn(log->l_curr_cycle, log->l_curr_block);
@@ -4933,7 +4933,7 @@  xlog_recover_process_intents(
 	}
 out:
 	xfs_trans_ail_cursor_done(&cur);
-	spin_unlock(&ailp->xa_lock);
+	spin_unlock(&ailp->ail_lock);
 	if (error)
 		xfs_defer_cancel(&dfops);
 	else
@@ -4956,7 +4956,7 @@  xlog_recover_cancel_intents(
 	struct xfs_ail		*ailp;
 
 	ailp = log->l_ailp;
-	spin_lock(&ailp->xa_lock);
+	spin_lock(&ailp->ail_lock);
 	lip = xfs_trans_ail_cursor_first(ailp, &cur, 0);
 	while (lip != NULL) {
 		/*
@@ -4990,7 +4990,7 @@  xlog_recover_cancel_intents(
 	}
 
 	xfs_trans_ail_cursor_done(&cur);
-	spin_unlock(&ailp->xa_lock);
+	spin_unlock(&ailp->ail_lock);
 	return error;
 }
 
diff --git a/fs/xfs/xfs_trans.c b/fs/xfs/xfs_trans.c
index a87f657f59c9..756e01999c24 100644
--- a/fs/xfs/xfs_trans.c
+++ b/fs/xfs/xfs_trans.c
@@ -781,8 +781,8 @@  xfs_log_item_batch_insert(
 {
 	int	i;
 
-	spin_lock(&ailp->xa_lock);
-	/* xfs_trans_ail_update_bulk drops ailp->xa_lock */
+	spin_lock(&ailp->ail_lock);
+	/* xfs_trans_ail_update_bulk drops ailp->ail_lock */
 	xfs_trans_ail_update_bulk(ailp, cur, log_items, nr_items, commit_lsn);
 
 	for (i = 0; i < nr_items; i++) {
@@ -825,9 +825,9 @@  xfs_trans_committed_bulk(
 	struct xfs_ail_cursor	cur;
 	int			i = 0;
 
-	spin_lock(&ailp->xa_lock);
+	spin_lock(&ailp->ail_lock);
 	xfs_trans_ail_cursor_last(ailp, &cur, commit_lsn);
-	spin_unlock(&ailp->xa_lock);
+	spin_unlock(&ailp->ail_lock);
 
 	/* unpin all the log items */
 	for (lv = log_vector; lv; lv = lv->lv_next ) {
@@ -847,7 +847,7 @@  xfs_trans_committed_bulk(
 		 * object into the AIL as we are in a shutdown situation.
 		 */
 		if (aborted) {
-			ASSERT(XFS_FORCED_SHUTDOWN(ailp->xa_mount));
+			ASSERT(XFS_FORCED_SHUTDOWN(ailp->ail_mount));
 			lip->li_ops->iop_unpin(lip, 1);
 			continue;
 		}
@@ -861,11 +861,11 @@  xfs_trans_committed_bulk(
 			 * not affect the AIL cursor the bulk insert path is
 			 * using.
 			 */
-			spin_lock(&ailp->xa_lock);
+			spin_lock(&ailp->ail_lock);
 			if (XFS_LSN_CMP(item_lsn, lip->li_lsn) > 0)
 				xfs_trans_ail_update(ailp, lip, item_lsn);
 			else
-				spin_unlock(&ailp->xa_lock);
+				spin_unlock(&ailp->ail_lock);
 			lip->li_ops->iop_unpin(lip, 0);
 			continue;
 		}
@@ -883,9 +883,9 @@  xfs_trans_committed_bulk(
 	if (i)
 		xfs_log_item_batch_insert(ailp, &cur, log_items, i, commit_lsn);
 
-	spin_lock(&ailp->xa_lock);
+	spin_lock(&ailp->ail_lock);
 	xfs_trans_ail_cursor_done(&cur);
-	spin_unlock(&ailp->xa_lock);
+	spin_unlock(&ailp->ail_lock);
 }
 
 /*
diff --git a/fs/xfs/xfs_trans_ail.c b/fs/xfs/xfs_trans_ail.c
index cef89f7127d3..d4a2445215e6 100644
--- a/fs/xfs/xfs_trans_ail.c
+++ b/fs/xfs/xfs_trans_ail.c
@@ -40,7 +40,7 @@  xfs_ail_check(
 {
 	xfs_log_item_t	*prev_lip;
 
-	if (list_empty(&ailp->xa_ail))
+	if (list_empty(&ailp->ail_head))
 		return;
 
 	/*
@@ -48,11 +48,11 @@  xfs_ail_check(
 	 */
 	ASSERT((lip->li_flags & XFS_LI_IN_AIL) != 0);
 	prev_lip = list_entry(lip->li_ail.prev, xfs_log_item_t, li_ail);
-	if (&prev_lip->li_ail != &ailp->xa_ail)
+	if (&prev_lip->li_ail != &ailp->ail_head)
 		ASSERT(XFS_LSN_CMP(prev_lip->li_lsn, lip->li_lsn) <= 0);
 
 	prev_lip = list_entry(lip->li_ail.next, xfs_log_item_t, li_ail);
-	if (&prev_lip->li_ail != &ailp->xa_ail)
+	if (&prev_lip->li_ail != &ailp->ail_head)
 		ASSERT(XFS_LSN_CMP(prev_lip->li_lsn, lip->li_lsn) >= 0);
 
 
@@ -69,10 +69,10 @@  static xfs_log_item_t *
 xfs_ail_max(
 	struct xfs_ail  *ailp)
 {
-	if (list_empty(&ailp->xa_ail))
+	if (list_empty(&ailp->ail_head))
 		return NULL;
 
-	return list_entry(ailp->xa_ail.prev, xfs_log_item_t, li_ail);
+	return list_entry(ailp->ail_head.prev, xfs_log_item_t, li_ail);
 }
 
 /*
@@ -84,7 +84,7 @@  xfs_ail_next(
 	struct xfs_ail  *ailp,
 	xfs_log_item_t  *lip)
 {
-	if (lip->li_ail.next == &ailp->xa_ail)
+	if (lip->li_ail.next == &ailp->ail_head)
 		return NULL;
 
 	return list_first_entry(&lip->li_ail, xfs_log_item_t, li_ail);
@@ -105,11 +105,11 @@  xfs_ail_min_lsn(
 	xfs_lsn_t	lsn = 0;
 	xfs_log_item_t	*lip;
 
-	spin_lock(&ailp->xa_lock);
+	spin_lock(&ailp->ail_lock);
 	lip = xfs_ail_min(ailp);
 	if (lip)
 		lsn = lip->li_lsn;
-	spin_unlock(&ailp->xa_lock);
+	spin_unlock(&ailp->ail_lock);
 
 	return lsn;
 }
@@ -124,11 +124,11 @@  xfs_ail_max_lsn(
 	xfs_lsn_t       lsn = 0;
 	xfs_log_item_t  *lip;
 
-	spin_lock(&ailp->xa_lock);
+	spin_lock(&ailp->ail_lock);
 	lip = xfs_ail_max(ailp);
 	if (lip)
 		lsn = lip->li_lsn;
-	spin_unlock(&ailp->xa_lock);
+	spin_unlock(&ailp->ail_lock);
 
 	return lsn;
 }
@@ -146,7 +146,7 @@  xfs_trans_ail_cursor_init(
 	struct xfs_ail_cursor	*cur)
 {
 	cur->item = NULL;
-	list_add_tail(&cur->list, &ailp->xa_cursors);
+	list_add_tail(&cur->list, &ailp->ail_cursors);
 }
 
 /*
@@ -194,7 +194,7 @@  xfs_trans_ail_cursor_clear(
 {
 	struct xfs_ail_cursor	*cur;
 
-	list_for_each_entry(cur, &ailp->xa_cursors, list) {
+	list_for_each_entry(cur, &ailp->ail_cursors, list) {
 		if (cur->item == lip)
 			cur->item = (struct xfs_log_item *)
 					((uintptr_t)cur->item | 1);
@@ -222,7 +222,7 @@  xfs_trans_ail_cursor_first(
 		goto out;
 	}
 
-	list_for_each_entry(lip, &ailp->xa_ail, li_ail) {
+	list_for_each_entry(lip, &ailp->ail_head, li_ail) {
 		if (XFS_LSN_CMP(lip->li_lsn, lsn) >= 0)
 			goto out;
 	}
@@ -241,7 +241,7 @@  __xfs_trans_ail_cursor_last(
 {
 	xfs_log_item_t		*lip;
 
-	list_for_each_entry_reverse(lip, &ailp->xa_ail, li_ail) {
+	list_for_each_entry_reverse(lip, &ailp->ail_head, li_ail) {
 		if (XFS_LSN_CMP(lip->li_lsn, lsn) <= 0)
 			return lip;
 	}
@@ -310,7 +310,7 @@  xfs_ail_splice(
 	if (lip)
 		list_splice(list, &lip->li_ail);
 	else
-		list_splice(list, &ailp->xa_ail);
+		list_splice(list, &ailp->ail_head);
 }
 
 /*
@@ -335,17 +335,17 @@  xfsaild_push_item(
 	 * If log item pinning is enabled, skip the push and track the item as
 	 * pinned. This can help induce head-behind-tail conditions.
 	 */
-	if (XFS_TEST_ERROR(false, ailp->xa_mount, XFS_ERRTAG_LOG_ITEM_PIN))
+	if (XFS_TEST_ERROR(false, ailp->ail_mount, XFS_ERRTAG_LOG_ITEM_PIN))
 		return XFS_ITEM_PINNED;
 
-	return lip->li_ops->iop_push(lip, &ailp->xa_buf_list);
+	return lip->li_ops->iop_push(lip, &ailp->ail_buf_list);
 }
 
 static long
 xfsaild_push(
 	struct xfs_ail		*ailp)
 {
-	xfs_mount_t		*mp = ailp->xa_mount;
+	xfs_mount_t		*mp = ailp->ail_mount;
 	struct xfs_ail_cursor	cur;
 	xfs_log_item_t		*lip;
 	xfs_lsn_t		lsn;
@@ -360,30 +360,30 @@  xfsaild_push(
 	 * buffers the last time we ran, force the log first and wait for it
 	 * before pushing again.
 	 */
-	if (ailp->xa_log_flush && ailp->xa_last_pushed_lsn == 0 &&
-	    (!list_empty_careful(&ailp->xa_buf_list) ||
+	if (ailp->ail_log_flush && ailp->ail_last_pushed_lsn == 0 &&
+	    (!list_empty_careful(&ailp->ail_buf_list) ||
 	     xfs_ail_min_lsn(ailp))) {
-		ailp->xa_log_flush = 0;
+		ailp->ail_log_flush = 0;
 
 		XFS_STATS_INC(mp, xs_push_ail_flush);
 		xfs_log_force(mp, XFS_LOG_SYNC);
 	}
 
-	spin_lock(&ailp->xa_lock);
+	spin_lock(&ailp->ail_lock);
 
-	/* barrier matches the xa_target update in xfs_ail_push() */
+	/* barrier matches the ail_target update in xfs_ail_push() */
 	smp_rmb();
-	target = ailp->xa_target;
-	ailp->xa_target_prev = target;
+	target = ailp->ail_target;
+	ailp->ail_target_prev = target;
 
-	lip = xfs_trans_ail_cursor_first(ailp, &cur, ailp->xa_last_pushed_lsn);
+	lip = xfs_trans_ail_cursor_first(ailp, &cur, ailp->ail_last_pushed_lsn);
 	if (!lip) {
 		/*
 		 * If the AIL is empty or our push has reached the end we are
 		 * done now.
 		 */
 		xfs_trans_ail_cursor_done(&cur);
-		spin_unlock(&ailp->xa_lock);
+		spin_unlock(&ailp->ail_lock);
 		goto out_done;
 	}
 
@@ -404,7 +404,7 @@  xfsaild_push(
 			XFS_STATS_INC(mp, xs_push_ail_success);
 			trace_xfs_ail_push(lip);
 
-			ailp->xa_last_pushed_lsn = lsn;
+			ailp->ail_last_pushed_lsn = lsn;
 			break;
 
 		case XFS_ITEM_FLUSHING:
@@ -423,7 +423,7 @@  xfsaild_push(
 			trace_xfs_ail_flushing(lip);
 
 			flushing++;
-			ailp->xa_last_pushed_lsn = lsn;
+			ailp->ail_last_pushed_lsn = lsn;
 			break;
 
 		case XFS_ITEM_PINNED:
@@ -431,7 +431,7 @@  xfsaild_push(
 			trace_xfs_ail_pinned(lip);
 
 			stuck++;
-			ailp->xa_log_flush++;
+			ailp->ail_log_flush++;
 			break;
 		case XFS_ITEM_LOCKED:
 			XFS_STATS_INC(mp, xs_push_ail_locked);
@@ -468,10 +468,10 @@  xfsaild_push(
 		lsn = lip->li_lsn;
 	}
 	xfs_trans_ail_cursor_done(&cur);
-	spin_unlock(&ailp->xa_lock);
+	spin_unlock(&ailp->ail_lock);
 
-	if (xfs_buf_delwri_submit_nowait(&ailp->xa_buf_list))
-		ailp->xa_log_flush++;
+	if (xfs_buf_delwri_submit_nowait(&ailp->ail_buf_list))
+		ailp->ail_log_flush++;
 
 	if (!count || XFS_LSN_CMP(lsn, target) >= 0) {
 out_done:
@@ -481,7 +481,7 @@  xfsaild_push(
 		 * AIL before we start the next scan from the start of the AIL.
 		 */
 		tout = 50;
-		ailp->xa_last_pushed_lsn = 0;
+		ailp->ail_last_pushed_lsn = 0;
 	} else if (((stuck + flushing) * 100) / count > 90) {
 		/*
 		 * Either there is a lot of contention on the AIL or we are
@@ -494,7 +494,7 @@  xfsaild_push(
 		 * the restart to issue a log force to unpin the stuck items.
 		 */
 		tout = 20;
-		ailp->xa_last_pushed_lsn = 0;
+		ailp->ail_last_pushed_lsn = 0;
 	} else {
 		/*
 		 * Assume we have more work to do in a short while.
@@ -536,26 +536,26 @@  xfsaild(
 			break;
 		}
 
-		spin_lock(&ailp->xa_lock);
+		spin_lock(&ailp->ail_lock);
 
 		/*
 		 * Idle if the AIL is empty and we are not racing with a target
 		 * update. We check the AIL after we set the task to a sleep
-		 * state to guarantee that we either catch an xa_target update
+		 * state to guarantee that we either catch an ail_target update
 		 * or that a wake_up resets the state to TASK_RUNNING.
 		 * Otherwise, we run the risk of sleeping indefinitely.
 		 *
-		 * The barrier matches the xa_target update in xfs_ail_push().
+		 * The barrier matches the ail_target update in xfs_ail_push().
 		 */
 		smp_rmb();
 		if (!xfs_ail_min(ailp) &&
-		    ailp->xa_target == ailp->xa_target_prev) {
-			spin_unlock(&ailp->xa_lock);
+		    ailp->ail_target == ailp->ail_target_prev) {
+			spin_unlock(&ailp->ail_lock);
 			freezable_schedule();
 			tout = 0;
 			continue;
 		}
-		spin_unlock(&ailp->xa_lock);
+		spin_unlock(&ailp->ail_lock);
 
 		if (tout)
 			freezable_schedule_timeout(msecs_to_jiffies(tout));
@@ -592,8 +592,8 @@  xfs_ail_push(
 	xfs_log_item_t	*lip;
 
 	lip = xfs_ail_min(ailp);
-	if (!lip || XFS_FORCED_SHUTDOWN(ailp->xa_mount) ||
-	    XFS_LSN_CMP(threshold_lsn, ailp->xa_target) <= 0)
+	if (!lip || XFS_FORCED_SHUTDOWN(ailp->ail_mount) ||
+	    XFS_LSN_CMP(threshold_lsn, ailp->ail_target) <= 0)
 		return;
 
 	/*
@@ -601,10 +601,10 @@  xfs_ail_push(
 	 * the XFS_AIL_PUSHING_BIT.
 	 */
 	smp_wmb();
-	xfs_trans_ail_copy_lsn(ailp, &ailp->xa_target, &threshold_lsn);
+	xfs_trans_ail_copy_lsn(ailp, &ailp->ail_target, &threshold_lsn);
 	smp_wmb();
 
-	wake_up_process(ailp->xa_task);
+	wake_up_process(ailp->ail_task);
 }
 
 /*
@@ -630,18 +630,18 @@  xfs_ail_push_all_sync(
 	struct xfs_log_item	*lip;
 	DEFINE_WAIT(wait);
 
-	spin_lock(&ailp->xa_lock);
+	spin_lock(&ailp->ail_lock);
 	while ((lip = xfs_ail_max(ailp)) != NULL) {
-		prepare_to_wait(&ailp->xa_empty, &wait, TASK_UNINTERRUPTIBLE);
-		ailp->xa_target = lip->li_lsn;
-		wake_up_process(ailp->xa_task);
-		spin_unlock(&ailp->xa_lock);
+		prepare_to_wait(&ailp->ail_empty, &wait, TASK_UNINTERRUPTIBLE);
+		ailp->ail_target = lip->li_lsn;
+		wake_up_process(ailp->ail_task);
+		spin_unlock(&ailp->ail_lock);
 		schedule();
-		spin_lock(&ailp->xa_lock);
+		spin_lock(&ailp->ail_lock);
 	}
-	spin_unlock(&ailp->xa_lock);
+	spin_unlock(&ailp->ail_lock);
 
-	finish_wait(&ailp->xa_empty, &wait);
+	finish_wait(&ailp->ail_empty, &wait);
 }
 
 /*
@@ -672,7 +672,7 @@  xfs_trans_ail_update_bulk(
 	struct xfs_ail_cursor	*cur,
 	struct xfs_log_item	**log_items,
 	int			nr_items,
-	xfs_lsn_t		lsn) __releases(ailp->xa_lock)
+	xfs_lsn_t		lsn) __releases(ailp->ail_lock)
 {
 	xfs_log_item_t		*mlip;
 	int			mlip_changed = 0;
@@ -705,13 +705,13 @@  xfs_trans_ail_update_bulk(
 		xfs_ail_splice(ailp, cur, &tmp, lsn);
 
 	if (mlip_changed) {
-		if (!XFS_FORCED_SHUTDOWN(ailp->xa_mount))
-			xlog_assign_tail_lsn_locked(ailp->xa_mount);
-		spin_unlock(&ailp->xa_lock);
+		if (!XFS_FORCED_SHUTDOWN(ailp->ail_mount))
+			xlog_assign_tail_lsn_locked(ailp->ail_mount);
+		spin_unlock(&ailp->ail_lock);
 
-		xfs_log_space_wake(ailp->xa_mount);
+		xfs_log_space_wake(ailp->ail_mount);
 	} else {
-		spin_unlock(&ailp->xa_lock);
+		spin_unlock(&ailp->ail_lock);
 	}
 }
 
@@ -756,13 +756,13 @@  void
 xfs_trans_ail_delete(
 	struct xfs_ail		*ailp,
 	struct xfs_log_item	*lip,
-	int			shutdown_type) __releases(ailp->xa_lock)
+	int			shutdown_type) __releases(ailp->ail_lock)
 {
-	struct xfs_mount	*mp = ailp->xa_mount;
+	struct xfs_mount	*mp = ailp->ail_mount;
 	bool			mlip_changed;
 
 	if (!(lip->li_flags & XFS_LI_IN_AIL)) {
-		spin_unlock(&ailp->xa_lock);
+		spin_unlock(&ailp->ail_lock);
 		if (!XFS_FORCED_SHUTDOWN(mp)) {
 			xfs_alert_tag(mp, XFS_PTAG_AILDELETE,
 	"%s: attempting to delete a log item that is not in the AIL",
@@ -776,13 +776,13 @@  xfs_trans_ail_delete(
 	if (mlip_changed) {
 		if (!XFS_FORCED_SHUTDOWN(mp))
 			xlog_assign_tail_lsn_locked(mp);
-		if (list_empty(&ailp->xa_ail))
-			wake_up_all(&ailp->xa_empty);
+		if (list_empty(&ailp->ail_head))
+			wake_up_all(&ailp->ail_empty);
 	}
 
-	spin_unlock(&ailp->xa_lock);
+	spin_unlock(&ailp->ail_lock);
 	if (mlip_changed)
-		xfs_log_space_wake(ailp->xa_mount);
+		xfs_log_space_wake(ailp->ail_mount);
 }
 
 int
@@ -795,16 +795,16 @@  xfs_trans_ail_init(
 	if (!ailp)
 		return -ENOMEM;
 
-	ailp->xa_mount = mp;
-	INIT_LIST_HEAD(&ailp->xa_ail);
-	INIT_LIST_HEAD(&ailp->xa_cursors);
-	spin_lock_init(&ailp->xa_lock);
-	INIT_LIST_HEAD(&ailp->xa_buf_list);
-	init_waitqueue_head(&ailp->xa_empty);
+	ailp->ail_mount = mp;
+	INIT_LIST_HEAD(&ailp->ail_head);
+	INIT_LIST_HEAD(&ailp->ail_cursors);
+	spin_lock_init(&ailp->ail_lock);
+	INIT_LIST_HEAD(&ailp->ail_buf_list);
+	init_waitqueue_head(&ailp->ail_empty);
 
-	ailp->xa_task = kthread_run(xfsaild, ailp, "xfsaild/%s",
-			ailp->xa_mount->m_fsname);
-	if (IS_ERR(ailp->xa_task))
+	ailp->ail_task = kthread_run(xfsaild, ailp, "xfsaild/%s",
+			ailp->ail_mount->m_fsname);
+	if (IS_ERR(ailp->ail_task))
 		goto out_free_ailp;
 
 	mp->m_ail = ailp;
@@ -821,6 +821,6 @@  xfs_trans_ail_destroy(
 {
 	struct xfs_ail	*ailp = mp->m_ail;
 
-	kthread_stop(ailp->xa_task);
+	kthread_stop(ailp->ail_task);
 	kmem_free(ailp);
 }
diff --git a/fs/xfs/xfs_trans_buf.c b/fs/xfs/xfs_trans_buf.c
index 3ba7a96a8abd..b8871bcfe00b 100644
--- a/fs/xfs/xfs_trans_buf.c
+++ b/fs/xfs/xfs_trans_buf.c
@@ -429,8 +429,8 @@  xfs_trans_brelse(xfs_trans_t	*tp,
 	 * If the fs has shutdown and we dropped the last reference, it may fall
 	 * on us to release a (possibly dirty) bli if it never made it to the
 	 * AIL (e.g., the aborted unpin already happened and didn't release it
-	 * due to our reference). Since we're already shutdown and need xa_lock,
-	 * just force remove from the AIL and release the bli here.
+	 * due to our reference). Since we're already shutdown and need
+	 * ail_lock, just force remove from the AIL and release the bli here.
 	 */
 	if (XFS_FORCED_SHUTDOWN(tp->t_mountp) && freed) {
 		xfs_trans_ail_remove(&bip->bli_item, SHUTDOWN_LOG_IO_ERROR);
diff --git a/fs/xfs/xfs_trans_priv.h b/fs/xfs/xfs_trans_priv.h
index b317a3644c00..be24b0c8a332 100644
--- a/fs/xfs/xfs_trans_priv.h
+++ b/fs/xfs/xfs_trans_priv.h
@@ -65,17 +65,17 @@  struct xfs_ail_cursor {
  * Eventually we need to drive the locking in here as well.
  */
 struct xfs_ail {
-	struct xfs_mount	*xa_mount;
-	struct task_struct	*xa_task;
-	struct list_head	xa_ail;
-	xfs_lsn_t		xa_target;
-	xfs_lsn_t		xa_target_prev;
-	struct list_head	xa_cursors;
-	spinlock_t		xa_lock;
-	xfs_lsn_t		xa_last_pushed_lsn;
-	int			xa_log_flush;
-	struct list_head	xa_buf_list;
-	wait_queue_head_t	xa_empty;
+	struct xfs_mount	*ail_mount;
+	struct task_struct	*ail_task;
+	struct list_head	ail_head;
+	xfs_lsn_t		ail_target;
+	xfs_lsn_t		ail_target_prev;
+	struct list_head	ail_cursors;
+	spinlock_t		ail_lock;
+	xfs_lsn_t		ail_last_pushed_lsn;
+	int			ail_log_flush;
+	struct list_head	ail_buf_list;
+	wait_queue_head_t	ail_empty;
 };
 
 /*
@@ -84,7 +84,7 @@  struct xfs_ail {
 void	xfs_trans_ail_update_bulk(struct xfs_ail *ailp,
 				struct xfs_ail_cursor *cur,
 				struct xfs_log_item **log_items, int nr_items,
-				xfs_lsn_t lsn) __releases(ailp->xa_lock);
+				xfs_lsn_t lsn) __releases(ailp->ail_lock);
 /*
  * Return a pointer to the first item in the AIL.  If the AIL is empty, then
  * return NULL.
@@ -93,7 +93,7 @@  static inline struct xfs_log_item *
 xfs_ail_min(
 	struct xfs_ail  *ailp)
 {
-	return list_first_entry_or_null(&ailp->xa_ail, struct xfs_log_item,
+	return list_first_entry_or_null(&ailp->ail_head, struct xfs_log_item,
 					li_ail);
 }
 
@@ -101,14 +101,14 @@  static inline void
 xfs_trans_ail_update(
 	struct xfs_ail		*ailp,
 	struct xfs_log_item	*lip,
-	xfs_lsn_t		lsn) __releases(ailp->xa_lock)
+	xfs_lsn_t		lsn) __releases(ailp->ail_lock)
 {
 	xfs_trans_ail_update_bulk(ailp, NULL, &lip, 1, lsn);
 }
 
 bool xfs_ail_delete_one(struct xfs_ail *ailp, struct xfs_log_item *lip);
 void xfs_trans_ail_delete(struct xfs_ail *ailp, struct xfs_log_item *lip,
-		int shutdown_type) __releases(ailp->xa_lock);
+		int shutdown_type) __releases(ailp->ail_lock);
 
 static inline void
 xfs_trans_ail_remove(
@@ -117,12 +117,12 @@  xfs_trans_ail_remove(
 {
 	struct xfs_ail		*ailp = lip->li_ailp;
 
-	spin_lock(&ailp->xa_lock);
+	spin_lock(&ailp->ail_lock);
 	/* xfs_trans_ail_delete() drops the AIL lock */
 	if (lip->li_flags & XFS_LI_IN_AIL)
 		xfs_trans_ail_delete(ailp, lip, shutdown_type);
 	else
-		spin_unlock(&ailp->xa_lock);
+		spin_unlock(&ailp->ail_lock);
 }
 
 void			xfs_ail_push(struct xfs_ail *, xfs_lsn_t);
@@ -149,9 +149,9 @@  xfs_trans_ail_copy_lsn(
 	xfs_lsn_t	*src)
 {
 	ASSERT(sizeof(xfs_lsn_t) == 8);	/* don't lock if it shrinks */
-	spin_lock(&ailp->xa_lock);
+	spin_lock(&ailp->ail_lock);
 	*dst = *src;
-	spin_unlock(&ailp->xa_lock);
+	spin_unlock(&ailp->ail_lock);
 }
 #else
 static inline void
@@ -172,7 +172,7 @@  xfs_clear_li_failed(
 	struct xfs_buf	*bp = lip->li_buf;
 
 	ASSERT(lip->li_flags & XFS_LI_IN_AIL);
-	lockdep_assert_held(&lip->li_ailp->xa_lock);
+	lockdep_assert_held(&lip->li_ailp->ail_lock);
 
 	if (lip->li_flags & XFS_LI_FAILED) {
 		lip->li_flags &= ~XFS_LI_FAILED;
@@ -186,7 +186,7 @@  xfs_set_li_failed(
 	struct xfs_log_item	*lip,
 	struct xfs_buf		*bp)
 {
-	lockdep_assert_held(&lip->li_ailp->xa_lock);
+	lockdep_assert_held(&lip->li_ailp->ail_lock);
 
 	if (!(lip->li_flags & XFS_LI_FAILED)) {
 		xfs_buf_hold(bp);