提交 03135cf7 编写于 作者: D David Chinner 提交者: Tim Shimmin

[XFS] Fix UP build breakage due to undefined m_icsb_mutex.

SGI-PV: 952227
SGI-Modid: xfs-linux-melb:xfs-kern:27692a
Signed-off-by: NDavid Chinner <dgc@sgi.com>
Signed-off-by: NLachlan McIlroy <lachlan@sgi.com>
Signed-off-by: NTim Shimmin <tes@sgi.com>
上级 20b64285
......@@ -543,11 +543,11 @@ xfs_readsb(xfs_mount_t *mp, int flags)
ASSERT(XFS_BUF_VALUSEMA(bp) <= 0);
}
mutex_lock(&mp->m_icsb_mutex);
xfs_icsb_lock(mp);
xfs_icsb_balance_counter(mp, XFS_SBS_ICOUNT, 0, 0);
xfs_icsb_balance_counter(mp, XFS_SBS_IFREE, 0, 0);
xfs_icsb_balance_counter(mp, XFS_SBS_FDBLOCKS, 0, 0);
mutex_unlock(&mp->m_icsb_mutex);
xfs_icsb_unlock(mp);
mp->m_sb_bp = bp;
xfs_buf_relse(bp);
......@@ -1736,17 +1736,17 @@ xfs_icsb_cpu_notify(
memset(cntp, 0, sizeof(xfs_icsb_cnts_t));
break;
case CPU_ONLINE:
mutex_lock(&mp->m_icsb_mutex);
xfs_icsb_lock(mp);
xfs_icsb_balance_counter(mp, XFS_SBS_ICOUNT, 0, 0);
xfs_icsb_balance_counter(mp, XFS_SBS_IFREE, 0, 0);
xfs_icsb_balance_counter(mp, XFS_SBS_FDBLOCKS, 0, 0);
mutex_unlock(&mp->m_icsb_mutex);
xfs_icsb_unlock(mp);
break;
case CPU_DEAD:
/* Disable all the counters, then fold the dead cpu's
* count into the total on the global superblock and
* re-enable the counters. */
mutex_lock(&mp->m_icsb_mutex);
xfs_icsb_lock(mp);
s = XFS_SB_LOCK(mp);
xfs_icsb_disable_counter(mp, XFS_SBS_ICOUNT);
xfs_icsb_disable_counter(mp, XFS_SBS_IFREE);
......@@ -1765,7 +1765,7 @@ xfs_icsb_cpu_notify(
xfs_icsb_balance_counter(mp, XFS_SBS_FDBLOCKS,
XFS_ICSB_SB_LOCKED, 0);
XFS_SB_UNLOCK(mp, s);
mutex_unlock(&mp->m_icsb_mutex);
xfs_icsb_unlock(mp);
break;
}
......@@ -1813,6 +1813,7 @@ xfs_icsb_destroy_counters(
unregister_hotcpu_notifier(&mp->m_icsb_notifier);
free_percpu(mp->m_sb_cnts);
}
mutex_destroy(&mp->m_icsb_mutex);
}
STATIC_INLINE void
......@@ -2156,7 +2157,7 @@ xfs_icsb_modify_counters(
* the superblock lock. We still need to hold the superblock
* lock, however, when we modify the global structures.
*/
mutex_lock(&mp->m_icsb_mutex);
xfs_icsb_lock(mp);
/*
* Now running atomically.
......@@ -2165,7 +2166,7 @@ xfs_icsb_modify_counters(
* Drop the lock and try again in the fast path....
*/
if (!(xfs_icsb_counter_disabled(mp, field))) {
mutex_unlock(&mp->m_icsb_mutex);
xfs_icsb_unlock(mp);
goto again;
}
......@@ -2192,7 +2193,7 @@ xfs_icsb_modify_counters(
*/
if (ret != ENOSPC)
xfs_icsb_balance_counter(mp, field, 0, 0);
mutex_unlock(&mp->m_icsb_mutex);
xfs_icsb_unlock(mp);
return ret;
balance_counter:
......@@ -2205,7 +2206,7 @@ xfs_icsb_modify_counters(
* do more balances than strictly necessary but it is not
* the common slowpath case.
*/
mutex_lock(&mp->m_icsb_mutex);
xfs_icsb_lock(mp);
/*
* running atomically.
......@@ -2216,7 +2217,7 @@ xfs_icsb_modify_counters(
* another balance operation being required.
*/
xfs_icsb_balance_counter(mp, field, 0, delta);
mutex_unlock(&mp->m_icsb_mutex);
xfs_icsb_unlock(mp);
goto again;
}
......
......@@ -18,6 +18,7 @@
#ifndef __XFS_MOUNT_H__
#define __XFS_MOUNT_H__
typedef struct xfs_trans_reservations {
uint tr_write; /* extent alloc trans */
uint tr_itruncate; /* truncate trans */
......@@ -563,6 +564,26 @@ xfs_daddr_to_agbno(struct xfs_mount *mp, xfs_daddr_t d)
return (xfs_agblock_t) do_div(ld, mp->m_sb.sb_agblocks);
}
/*
* Per-cpu superblock locking functions
*/
#ifdef HAVE_PERCPU_SB
STATIC_INLINE void
xfs_icsb_lock(xfs_mount_t *mp)
{
mutex_lock(&mp->m_icsb_mutex);
}
STATIC_INLINE void
xfs_icsb_unlock(xfs_mount_t *mp)
{
mutex_unlock(&mp->m_icsb_mutex);
}
#else
#define xfs_icsb_lock(mp)
#define xfs_icsb_unlock(mp)
#endif
/*
* This structure is for use by the xfs_mod_incore_sb_batch() routine.
*/
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册