/* * Copyright (C) 2017 Oracle. All Rights Reserved. * * Author: Darrick J. Wong * * This program is free software; you can redistribute it and/or * modify it under the terms of the GNU General Public License * as published by the Free Software Foundation; either version 2 * of the License, or (at your option) any later version. * * This program is distributed in the hope that it would be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the * GNU General Public License for more details. * * You should have received a copy of the GNU General Public License * along with this program; if not, write the Free Software Foundation, * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301, USA. */ #include "xfs.h" #include "xfs_fs.h" #include "xfs_shared.h" #include "xfs_format.h" #include "xfs_trans_resv.h" #include "xfs_mount.h" #include "xfs_defer.h" #include "xfs_btree.h" #include "xfs_bit.h" #include "xfs_log_format.h" #include "xfs_trans.h" #include "xfs_sb.h" #include "xfs_inode.h" #include "xfs_inode_fork.h" #include "xfs_alloc.h" #include "xfs_bmap.h" #include "xfs_quota.h" #include "xfs_qm.h" #include "xfs_dquot.h" #include "xfs_dquot_item.h" #include "scrub/xfs_scrub.h" #include "scrub/scrub.h" #include "scrub/common.h" #include "scrub/trace.h" /* Convert a scrub type code to a DQ flag, or return 0 if error. */ static inline uint xfs_scrub_quota_to_dqtype( struct xfs_scrub_context *sc) { switch (sc->sm->sm_type) { case XFS_SCRUB_TYPE_UQUOTA: return XFS_DQ_USER; case XFS_SCRUB_TYPE_GQUOTA: return XFS_DQ_GROUP; case XFS_SCRUB_TYPE_PQUOTA: return XFS_DQ_PROJ; default: return 0; } } /* Set us up to scrub a quota. */ int xfs_scrub_setup_quota( struct xfs_scrub_context *sc, struct xfs_inode *ip) { uint dqtype; dqtype = xfs_scrub_quota_to_dqtype(sc); if (dqtype == 0) return -EINVAL; if (!xfs_this_quota_on(sc->mp, dqtype)) return -ENOENT; return 0; } /* Quotas. */ struct xfs_scrub_quota_info { struct xfs_scrub_context *sc; xfs_dqid_t last_id; }; /* Scrub the fields in an individual quota item. */ STATIC int xfs_scrub_quota_item( struct xfs_dquot *dq, uint dqtype, void *priv) { struct xfs_scrub_quota_info *sqi = priv; struct xfs_scrub_context *sc = sqi->sc; struct xfs_mount *mp = sc->mp; struct xfs_disk_dquot *d = &dq->q_core; struct xfs_quotainfo *qi = mp->m_quotainfo; xfs_fileoff_t offset; unsigned long long bsoft; unsigned long long isoft; unsigned long long rsoft; unsigned long long bhard; unsigned long long ihard; unsigned long long rhard; unsigned long long bcount; unsigned long long icount; unsigned long long rcount; xfs_ino_t fs_icount; xfs_dqid_t id = be32_to_cpu(d->d_id); /* * Except for the root dquot, the actual dquot we got must either have * the same or higher id as we saw before. */ offset = id / qi->qi_dqperchunk; if (id && id <= sqi->last_id) xfs_scrub_fblock_set_corrupt(sc, XFS_DATA_FORK, offset); sqi->last_id = id; /* Did we get the dquot type we wanted? */ if (dqtype != (d->d_flags & XFS_DQ_ALLTYPES)) xfs_scrub_fblock_set_corrupt(sc, XFS_DATA_FORK, offset); if (d->d_pad0 != cpu_to_be32(0) || d->d_pad != cpu_to_be16(0)) xfs_scrub_fblock_set_corrupt(sc, XFS_DATA_FORK, offset); /* Check the limits. */ bhard = be64_to_cpu(d->d_blk_hardlimit); ihard = be64_to_cpu(d->d_ino_hardlimit); rhard = be64_to_cpu(d->d_rtb_hardlimit); bsoft = be64_to_cpu(d->d_blk_softlimit); isoft = be64_to_cpu(d->d_ino_softlimit); rsoft = be64_to_cpu(d->d_rtb_softlimit); /* * Warn if the hard limits are larger than the fs. * Administrators can do this, though in production this seems * suspect, which is why we flag it for review. * * Complain about corruption if the soft limit is greater than * the hard limit. */ if (bhard > mp->m_sb.sb_dblocks) xfs_scrub_fblock_set_warning(sc, XFS_DATA_FORK, offset); if (bsoft > bhard) xfs_scrub_fblock_set_corrupt(sc, XFS_DATA_FORK, offset); if (ihard > mp->m_maxicount) xfs_scrub_fblock_set_warning(sc, XFS_DATA_FORK, offset); if (isoft > ihard) xfs_scrub_fblock_set_corrupt(sc, XFS_DATA_FORK, offset); if (rhard > mp->m_sb.sb_rblocks) xfs_scrub_fblock_set_warning(sc, XFS_DATA_FORK, offset); if (rsoft > rhard) xfs_scrub_fblock_set_corrupt(sc, XFS_DATA_FORK, offset); /* Check the resource counts. */ bcount = be64_to_cpu(d->d_bcount); icount = be64_to_cpu(d->d_icount); rcount = be64_to_cpu(d->d_rtbcount); fs_icount = percpu_counter_sum(&mp->m_icount); /* * Check that usage doesn't exceed physical limits. However, on * a reflink filesystem we're allowed to exceed physical space * if there are no quota limits. */ if (xfs_sb_version_hasreflink(&mp->m_sb)) { if (mp->m_sb.sb_dblocks < bcount) xfs_scrub_fblock_set_warning(sc, XFS_DATA_FORK, offset); } else { if (mp->m_sb.sb_dblocks < bcount) xfs_scrub_fblock_set_corrupt(sc, XFS_DATA_FORK, offset); } if (icount > fs_icount || rcount > mp->m_sb.sb_rblocks) xfs_scrub_fblock_set_corrupt(sc, XFS_DATA_FORK, offset); /* * We can violate the hard limits if the admin suddenly sets a * lower limit than the actual usage. However, we flag it for * admin review. */ if (id != 0 && bhard != 0 && bcount > bhard) xfs_scrub_fblock_set_warning(sc, XFS_DATA_FORK, offset); if (id != 0 && ihard != 0 && icount > ihard) xfs_scrub_fblock_set_warning(sc, XFS_DATA_FORK, offset); if (id != 0 && rhard != 0 && rcount > rhard) xfs_scrub_fblock_set_warning(sc, XFS_DATA_FORK, offset); return 0; } /* Scrub all of a quota type's items. */ int xfs_scrub_quota( struct xfs_scrub_context *sc) { struct xfs_bmbt_irec irec = { 0 }; struct xfs_scrub_quota_info sqi; struct xfs_mount *mp = sc->mp; struct xfs_inode *ip; struct xfs_quotainfo *qi = mp->m_quotainfo; xfs_fileoff_t max_dqid_off; xfs_fileoff_t off = 0; uint dqtype; int nimaps; int error = 0; if (!XFS_IS_QUOTA_RUNNING(mp) || !XFS_IS_QUOTA_ON(mp)) return -ENOENT; mutex_lock(&qi->qi_quotaofflock); dqtype = xfs_scrub_quota_to_dqtype(sc); if (!xfs_this_quota_on(sc->mp, dqtype)) { error = -ENOENT; goto out_unlock_quota; } /* Attach to the quota inode and set sc->ip so that reporting works. */ ip = xfs_quota_inode(sc->mp, dqtype); sc->ip = ip; /* Look for problem extents. */ xfs_ilock(ip, XFS_ILOCK_EXCL); if (ip->i_d.di_flags & XFS_DIFLAG_REALTIME) { xfs_scrub_ino_set_corrupt(sc, sc->ip->i_ino); goto out_unlock_inode; } max_dqid_off = ((xfs_dqid_t)-1) / qi->qi_dqperchunk; while (1) { if (xfs_scrub_should_terminate(sc, &error)) break; off = irec.br_startoff + irec.br_blockcount; nimaps = 1; error = xfs_bmapi_read(ip, off, -1, &irec, &nimaps, XFS_BMAPI_ENTIRE); if (!xfs_scrub_fblock_process_error(sc, XFS_DATA_FORK, off, &error)) goto out_unlock_inode; if (!nimaps) break; if (irec.br_startblock == HOLESTARTBLOCK) continue; /* Check the extent record doesn't point to crap. */ if (irec.br_startblock + irec.br_blockcount <= irec.br_startblock) xfs_scrub_fblock_set_corrupt(sc, XFS_DATA_FORK, irec.br_startoff); if (!xfs_verify_fsbno(mp, irec.br_startblock) || !xfs_verify_fsbno(mp, irec.br_startblock + irec.br_blockcount - 1)) xfs_scrub_fblock_set_corrupt(sc, XFS_DATA_FORK, irec.br_startoff); /* * Unwritten extents or blocks mapped above the highest * quota id shouldn't happen. */ if (isnullstartblock(irec.br_startblock) || irec.br_startoff > max_dqid_off || irec.br_startoff + irec.br_blockcount > max_dqid_off + 1) xfs_scrub_fblock_set_corrupt(sc, XFS_DATA_FORK, off); } xfs_iunlock(ip, XFS_ILOCK_EXCL); if (sc->sm->sm_flags & XFS_SCRUB_OFLAG_CORRUPT) goto out; /* Check all the quota items. */ sqi.sc = sc; sqi.last_id = 0; error = xfs_qm_dqiterate(mp, dqtype, xfs_scrub_quota_item, &sqi); if (!xfs_scrub_fblock_process_error(sc, XFS_DATA_FORK, sqi.last_id * qi->qi_dqperchunk, &error)) goto out; out: /* We set sc->ip earlier, so make sure we clear it now. */ sc->ip = NULL; out_unlock_quota: mutex_unlock(&qi->qi_quotaofflock); return error; out_unlock_inode: xfs_iunlock(ip, XFS_ILOCK_EXCL); goto out; }