1 // SPDX-License-Identifier: GPL-2.0-or-later
3 * Copyright (C) 2017-2023 Oracle. All Rights Reserved.
8 #include "xfs_shared.h"
9 #include "xfs_format.h"
10 #include "xfs_trans_resv.h"
11 #include "xfs_mount.h"
12 #include "xfs_log_format.h"
13 #include "xfs_trans.h"
14 #include "xfs_inode.h"
15 #include "xfs_quota.h"
18 #include "scrub/scrub.h"
19 #include "scrub/common.h"
21 /* Convert a scrub type code to a DQ flag, or return 0 if error. */
22 static inline xfs_dqtype_t
26 switch (sc->sm->sm_type) {
27 case XFS_SCRUB_TYPE_UQUOTA:
28 return XFS_DQTYPE_USER;
29 case XFS_SCRUB_TYPE_GQUOTA:
30 return XFS_DQTYPE_GROUP;
31 case XFS_SCRUB_TYPE_PQUOTA:
32 return XFS_DQTYPE_PROJ;
38 /* Set us up to scrub a quota. */
46 if (!XFS_IS_QUOTA_ON(sc->mp))
49 dqtype = xchk_quota_to_dqtype(sc);
53 if (!xfs_this_quota_on(sc->mp, dqtype))
56 if (xchk_need_intent_drain(sc))
57 xchk_fsgates_enable(sc, XCHK_FSGATES_DRAIN);
59 error = xchk_setup_fs(sc);
62 sc->ip = xfs_quota_inode(sc->mp, dqtype);
63 xfs_ilock(sc->ip, XFS_ILOCK_EXCL);
64 sc->ilock_flags = XFS_ILOCK_EXCL;
70 struct xchk_quota_info {
75 /* Scrub the fields in an individual quota item. */
82 struct xchk_quota_info *sqi = priv;
83 struct xfs_scrub *sc = sqi->sc;
84 struct xfs_mount *mp = sc->mp;
85 struct xfs_quotainfo *qi = mp->m_quotainfo;
90 if (xchk_should_terminate(sc, &error))
94 * Except for the root dquot, the actual dquot we got must either have
95 * the same or higher id as we saw before.
97 offset = dq->q_id / qi->qi_dqperchunk;
98 if (dq->q_id && dq->q_id <= sqi->last_id)
99 xchk_fblock_set_corrupt(sc, XFS_DATA_FORK, offset);
101 sqi->last_id = dq->q_id;
104 * Warn if the hard limits are larger than the fs.
105 * Administrators can do this, though in production this seems
106 * suspect, which is why we flag it for review.
108 * Complain about corruption if the soft limit is greater than
111 if (dq->q_blk.hardlimit > mp->m_sb.sb_dblocks)
112 xchk_fblock_set_warning(sc, XFS_DATA_FORK, offset);
113 if (dq->q_blk.softlimit > dq->q_blk.hardlimit)
114 xchk_fblock_set_corrupt(sc, XFS_DATA_FORK, offset);
116 if (dq->q_ino.hardlimit > M_IGEO(mp)->maxicount)
117 xchk_fblock_set_warning(sc, XFS_DATA_FORK, offset);
118 if (dq->q_ino.softlimit > dq->q_ino.hardlimit)
119 xchk_fblock_set_corrupt(sc, XFS_DATA_FORK, offset);
121 if (dq->q_rtb.hardlimit > mp->m_sb.sb_rblocks)
122 xchk_fblock_set_warning(sc, XFS_DATA_FORK, offset);
123 if (dq->q_rtb.softlimit > dq->q_rtb.hardlimit)
124 xchk_fblock_set_corrupt(sc, XFS_DATA_FORK, offset);
126 /* Check the resource counts. */
127 fs_icount = percpu_counter_sum(&mp->m_icount);
130 * Check that usage doesn't exceed physical limits. However, on
131 * a reflink filesystem we're allowed to exceed physical space
132 * if there are no quota limits.
134 if (xfs_has_reflink(mp)) {
135 if (mp->m_sb.sb_dblocks < dq->q_blk.count)
136 xchk_fblock_set_warning(sc, XFS_DATA_FORK,
139 if (mp->m_sb.sb_dblocks < dq->q_blk.count)
140 xchk_fblock_set_corrupt(sc, XFS_DATA_FORK,
143 if (dq->q_ino.count > fs_icount || dq->q_rtb.count > mp->m_sb.sb_rblocks)
144 xchk_fblock_set_corrupt(sc, XFS_DATA_FORK, offset);
147 * We can violate the hard limits if the admin suddenly sets a
148 * lower limit than the actual usage. However, we flag it for
154 if (dq->q_blk.hardlimit != 0 &&
155 dq->q_blk.count > dq->q_blk.hardlimit)
156 xchk_fblock_set_warning(sc, XFS_DATA_FORK, offset);
158 if (dq->q_ino.hardlimit != 0 &&
159 dq->q_ino.count > dq->q_ino.hardlimit)
160 xchk_fblock_set_warning(sc, XFS_DATA_FORK, offset);
162 if (dq->q_rtb.hardlimit != 0 &&
163 dq->q_rtb.count > dq->q_rtb.hardlimit)
164 xchk_fblock_set_warning(sc, XFS_DATA_FORK, offset);
167 if (sc->sm->sm_flags & XFS_SCRUB_OFLAG_CORRUPT)
173 /* Check the quota's data fork. */
175 xchk_quota_data_fork(
176 struct xfs_scrub *sc)
178 struct xfs_bmbt_irec irec = { 0 };
179 struct xfs_iext_cursor icur;
180 struct xfs_quotainfo *qi = sc->mp->m_quotainfo;
181 struct xfs_ifork *ifp;
182 xfs_fileoff_t max_dqid_off;
185 /* Invoke the fork scrubber. */
186 error = xchk_metadata_inode_forks(sc);
187 if (error || (sc->sm->sm_flags & XFS_SCRUB_OFLAG_CORRUPT))
190 /* Check for data fork problems that apply only to quota files. */
191 max_dqid_off = ((xfs_dqid_t)-1) / qi->qi_dqperchunk;
192 ifp = xfs_ifork_ptr(sc->ip, XFS_DATA_FORK);
193 for_each_xfs_iext(ifp, &icur, &irec) {
194 if (xchk_should_terminate(sc, &error))
198 * delalloc/unwritten extents or blocks mapped above the highest
199 * quota id shouldn't happen.
201 if (!xfs_bmap_is_written_extent(&irec) ||
202 irec.br_startoff > max_dqid_off ||
203 irec.br_startoff + irec.br_blockcount - 1 > max_dqid_off) {
204 xchk_fblock_set_corrupt(sc, XFS_DATA_FORK,
213 /* Scrub all of a quota type's items. */
216 struct xfs_scrub *sc)
218 struct xchk_quota_info sqi;
219 struct xfs_mount *mp = sc->mp;
220 struct xfs_quotainfo *qi = mp->m_quotainfo;
224 dqtype = xchk_quota_to_dqtype(sc);
226 /* Look for problem extents. */
227 error = xchk_quota_data_fork(sc);
230 if (sc->sm->sm_flags & XFS_SCRUB_OFLAG_CORRUPT)
234 * Check all the quota items. Now that we've checked the quota inode
235 * data fork we have to drop ILOCK_EXCL to use the regular dquot
238 xfs_iunlock(sc->ip, sc->ilock_flags);
242 error = xfs_qm_dqiterate(mp, dqtype, xchk_quota_item, &sqi);
243 sc->ilock_flags = XFS_ILOCK_EXCL;
244 xfs_ilock(sc->ip, sc->ilock_flags);
245 if (error == -ECANCELED)
247 if (!xchk_fblock_process_error(sc, XFS_DATA_FORK,
248 sqi.last_id * qi->qi_dqperchunk, &error))