Author | Tokens | Token Proportion | Commits | Commit Proportion |
---|---|---|---|---|
Darrick J. Wong | 1409 | 97.92% | 28 | 75.68% |
Michal Marek | 10 | 0.69% | 1 | 2.70% |
David Chinner | 8 | 0.56% | 4 | 10.81% |
Russell Cattelan | 5 | 0.35% | 1 | 2.70% |
Christoph Hellwig | 4 | 0.28% | 1 | 2.70% |
Eric Sandeen | 3 | 0.21% | 2 | 5.41% |
Total | 1439 | 37 |
123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339
// SPDX-License-Identifier: GPL-2.0-or-later /* * Copyright (C) 2017-2023 Oracle. All Rights Reserved. * Author: Darrick J. Wong <djwong@kernel.org> */ #include "xfs.h" #include "xfs_fs.h" #include "xfs_shared.h" #include "xfs_bit.h" #include "xfs_format.h" #include "xfs_trans_resv.h" #include "xfs_mount.h" #include "xfs_log_format.h" #include "xfs_trans.h" #include "xfs_inode.h" #include "xfs_quota.h" #include "xfs_qm.h" #include "xfs_bmap.h" #include "scrub/scrub.h" #include "scrub/common.h" #include "scrub/quota.h" /* Convert a scrub type code to a DQ flag, or return 0 if error. */ xfs_dqtype_t xchk_quota_to_dqtype( struct xfs_scrub *sc) { switch (sc->sm->sm_type) { case XFS_SCRUB_TYPE_UQUOTA: return XFS_DQTYPE_USER; case XFS_SCRUB_TYPE_GQUOTA: return XFS_DQTYPE_GROUP; case XFS_SCRUB_TYPE_PQUOTA: return XFS_DQTYPE_PROJ; default: return 0; } } /* Set us up to scrub a quota. */ int xchk_setup_quota( struct xfs_scrub *sc) { xfs_dqtype_t dqtype; int error; if (!XFS_IS_QUOTA_ON(sc->mp)) return -ENOENT; dqtype = xchk_quota_to_dqtype(sc); if (dqtype == 0) return -EINVAL; if (!xfs_this_quota_on(sc->mp, dqtype)) return -ENOENT; if (xchk_need_intent_drain(sc)) xchk_fsgates_enable(sc, XCHK_FSGATES_DRAIN); error = xchk_setup_fs(sc); if (error) return error; error = xchk_install_live_inode(sc, xfs_quota_inode(sc->mp, dqtype)); if (error) return error; xchk_ilock(sc, XFS_ILOCK_EXCL); return 0; } /* Quotas. */ struct xchk_quota_info { struct xfs_scrub *sc; xfs_dqid_t last_id; }; /* There's a written block backing this dquot, right? */ STATIC int xchk_quota_item_bmap( struct xfs_scrub *sc, struct xfs_dquot *dq, xfs_fileoff_t offset) { struct xfs_bmbt_irec irec; struct xfs_mount *mp = sc->mp; int nmaps = 1; int error; if (!xfs_verify_fileoff(mp, offset)) { xchk_fblock_set_corrupt(sc, XFS_DATA_FORK, offset); return 0; } if (dq->q_fileoffset != offset) { xchk_fblock_set_corrupt(sc, XFS_DATA_FORK, offset); return 0; } error = xfs_bmapi_read(sc->ip, offset, 1, &irec, &nmaps, 0); if (error) return error; if (nmaps != 1) { xchk_fblock_set_corrupt(sc, XFS_DATA_FORK, offset); return 0; } if (!xfs_verify_fsbno(mp, irec.br_startblock)) xchk_fblock_set_corrupt(sc, XFS_DATA_FORK, offset); if (XFS_FSB_TO_DADDR(mp, irec.br_startblock) != dq->q_blkno) xchk_fblock_set_corrupt(sc, XFS_DATA_FORK, offset); if (!xfs_bmap_is_written_extent(&irec)) xchk_fblock_set_corrupt(sc, XFS_DATA_FORK, offset); return 0; } /* Complain if a quota timer is incorrectly set. */ static inline void xchk_quota_item_timer( struct xfs_scrub *sc, xfs_fileoff_t offset, const struct xfs_dquot_res *res) { if ((res->softlimit && res->count > res->softlimit) || (res->hardlimit && res->count > res->hardlimit)) { if (!res->timer) xchk_fblock_set_corrupt(sc, XFS_DATA_FORK, offset); } else { if (res->timer) xchk_fblock_set_corrupt(sc, XFS_DATA_FORK, offset); } } /* Scrub the fields in an individual quota item. */ STATIC int xchk_quota_item( struct xchk_quota_info *sqi, struct xfs_dquot *dq) { struct xfs_scrub *sc = sqi->sc; struct xfs_mount *mp = sc->mp; struct xfs_quotainfo *qi = mp->m_quotainfo; xfs_fileoff_t offset; xfs_ino_t fs_icount; int error = 0; if (xchk_should_terminate(sc, &error)) return error; /* * We want to validate the bmap record for the storage backing this * dquot, so we need to lock the dquot and the quota file. For quota * operations, the locking order is first the ILOCK and then the dquot. * However, dqiterate gave us a locked dquot, so drop the dquot lock to * get the ILOCK. */ xfs_dqunlock(dq); xchk_ilock(sc, XFS_ILOCK_SHARED); xfs_dqlock(dq); /* * Except for the root dquot, the actual dquot we got must either have * the same or higher id as we saw before. */ offset = dq->q_id / qi->qi_dqperchunk; if (dq->q_id && dq->q_id <= sqi->last_id) xchk_fblock_set_corrupt(sc, XFS_DATA_FORK, offset); sqi->last_id = dq->q_id; error = xchk_quota_item_bmap(sc, dq, offset); xchk_iunlock(sc, XFS_ILOCK_SHARED); if (!xchk_fblock_process_error(sc, XFS_DATA_FORK, offset, &error)) return error; /* * Warn if the hard limits are larger than the fs. * Administrators can do this, though in production this seems * suspect, which is why we flag it for review. * * Complain about corruption if the soft limit is greater than * the hard limit. */ if (dq->q_blk.hardlimit > mp->m_sb.sb_dblocks) xchk_fblock_set_warning(sc, XFS_DATA_FORK, offset); if (dq->q_blk.softlimit > dq->q_blk.hardlimit) xchk_fblock_set_corrupt(sc, XFS_DATA_FORK, offset); if (dq->q_ino.hardlimit > M_IGEO(mp)->maxicount) xchk_fblock_set_warning(sc, XFS_DATA_FORK, offset); if (dq->q_ino.softlimit > dq->q_ino.hardlimit) xchk_fblock_set_corrupt(sc, XFS_DATA_FORK, offset); if (dq->q_rtb.hardlimit > mp->m_sb.sb_rblocks) xchk_fblock_set_warning(sc, XFS_DATA_FORK, offset); if (dq->q_rtb.softlimit > dq->q_rtb.hardlimit) xchk_fblock_set_corrupt(sc, XFS_DATA_FORK, offset); /* Check the resource counts. */ fs_icount = percpu_counter_sum(&mp->m_icount); /* * Check that usage doesn't exceed physical limits. However, on * a reflink filesystem we're allowed to exceed physical space * if there are no quota limits. */ if (xfs_has_reflink(mp)) { if (mp->m_sb.sb_dblocks < dq->q_blk.count) xchk_fblock_set_warning(sc, XFS_DATA_FORK, offset); } else { if (mp->m_sb.sb_dblocks < dq->q_blk.count) xchk_fblock_set_corrupt(sc, XFS_DATA_FORK, offset); } if (dq->q_ino.count > fs_icount || dq->q_rtb.count > mp->m_sb.sb_rblocks) xchk_fblock_set_corrupt(sc, XFS_DATA_FORK, offset); /* * We can violate the hard limits if the admin suddenly sets a * lower limit than the actual usage. However, we flag it for * admin review. */ if (dq->q_id == 0) goto out; if (dq->q_blk.hardlimit != 0 && dq->q_blk.count > dq->q_blk.hardlimit) xchk_fblock_set_warning(sc, XFS_DATA_FORK, offset); if (dq->q_ino.hardlimit != 0 && dq->q_ino.count > dq->q_ino.hardlimit) xchk_fblock_set_warning(sc, XFS_DATA_FORK, offset); if (dq->q_rtb.hardlimit != 0 && dq->q_rtb.count > dq->q_rtb.hardlimit) xchk_fblock_set_warning(sc, XFS_DATA_FORK, offset); xchk_quota_item_timer(sc, offset, &dq->q_blk); xchk_quota_item_timer(sc, offset, &dq->q_ino); xchk_quota_item_timer(sc, offset, &dq->q_rtb); out: if (sc->sm->sm_flags & XFS_SCRUB_OFLAG_CORRUPT) return -ECANCELED; return 0; } /* Check the quota's data fork. */ STATIC int xchk_quota_data_fork( struct xfs_scrub *sc) { struct xfs_bmbt_irec irec = { 0 }; struct xfs_iext_cursor icur; struct xfs_quotainfo *qi = sc->mp->m_quotainfo; struct xfs_ifork *ifp; xfs_fileoff_t max_dqid_off; int error = 0; /* Invoke the fork scrubber. */ error = xchk_metadata_inode_forks(sc); if (error || (sc->sm->sm_flags & XFS_SCRUB_OFLAG_CORRUPT)) return error; /* Check for data fork problems that apply only to quota files. */ max_dqid_off = XFS_DQ_ID_MAX / qi->qi_dqperchunk; ifp = xfs_ifork_ptr(sc->ip, XFS_DATA_FORK); for_each_xfs_iext(ifp, &icur, &irec) { if (xchk_should_terminate(sc, &error)) break; /* * delalloc/unwritten extents or blocks mapped above the highest * quota id shouldn't happen. */ if (!xfs_bmap_is_written_extent(&irec) || irec.br_startoff > max_dqid_off || irec.br_startoff + irec.br_blockcount - 1 > max_dqid_off) { xchk_fblock_set_corrupt(sc, XFS_DATA_FORK, irec.br_startoff); break; } } return error; } /* Scrub all of a quota type's items. */ int xchk_quota( struct xfs_scrub *sc) { struct xchk_dqiter cursor = { }; struct xchk_quota_info sqi = { .sc = sc }; struct xfs_mount *mp = sc->mp; struct xfs_quotainfo *qi = mp->m_quotainfo; struct xfs_dquot *dq; xfs_dqtype_t dqtype; int error = 0; dqtype = xchk_quota_to_dqtype(sc); /* Look for problem extents. */ error = xchk_quota_data_fork(sc); if (error) goto out; if (sc->sm->sm_flags & XFS_SCRUB_OFLAG_CORRUPT) goto out; /* * Check all the quota items. Now that we've checked the quota inode * data fork we have to drop ILOCK_EXCL to use the regular dquot * functions. */ xchk_iunlock(sc, sc->ilock_flags); /* Now look for things that the quota verifiers won't complain about. */ xchk_dqiter_init(&cursor, sc, dqtype); while ((error = xchk_dquot_iter(&cursor, &dq)) == 1) { error = xchk_quota_item(&sqi, dq); xfs_qm_dqput(dq); if (error) break; } if (error == -ECANCELED) error = 0; if (!xchk_fblock_process_error(sc, XFS_DATA_FORK, sqi.last_id * qi->qi_dqperchunk, &error)) goto out; out: return error; }
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.
Created with Cregit http://github.com/cregit/cregit
Version 2.0-RC1