| // SPDX-License-Identifier: GPL-2.0+ | 
 | /* | 
 |  * Copyright (C) 2017 Oracle.  All Rights Reserved. | 
 |  * Author: Darrick J. Wong <darrick.wong@oracle.com> | 
 |  */ | 
 | #include "xfs.h" | 
 | #include "xfs_fs.h" | 
 | #include "xfs_shared.h" | 
 | #include "xfs_format.h" | 
 | #include "xfs_trans_resv.h" | 
 | #include "xfs_mount.h" | 
 | #include "xfs_defer.h" | 
 | #include "xfs_btree.h" | 
 | #include "xfs_bit.h" | 
 | #include "xfs_log_format.h" | 
 | #include "xfs_trans.h" | 
 | #include "xfs_sb.h" | 
 | #include "xfs_alloc.h" | 
 | #include "xfs_rmap.h" | 
 | #include "xfs_refcount.h" | 
 | #include "scrub/xfs_scrub.h" | 
 | #include "scrub/scrub.h" | 
 | #include "scrub/common.h" | 
 | #include "scrub/btree.h" | 
 | #include "scrub/trace.h" | 
 |  | 
 | /* | 
 |  * Set us up to scrub reference count btrees. | 
 |  */ | 
 | int | 
 | xchk_setup_ag_refcountbt( | 
 | 	struct xfs_scrub	*sc, | 
 | 	struct xfs_inode	*ip) | 
 | { | 
 | 	return xchk_setup_ag_btree(sc, ip, false); | 
 | } | 
 |  | 
 | /* Reference count btree scrubber. */ | 
 |  | 
 | /* | 
 |  * Confirming Reference Counts via Reverse Mappings | 
 |  * | 
 |  * We want to count the reverse mappings overlapping a refcount record | 
 |  * (bno, len, refcount), allowing for the possibility that some of the | 
 |  * overlap may come from smaller adjoining reverse mappings, while some | 
 |  * comes from single extents which overlap the range entirely.  The | 
 |  * outer loop is as follows: | 
 |  * | 
 |  * 1. For all reverse mappings overlapping the refcount extent, | 
 |  *    a. If a given rmap completely overlaps, mark it as seen. | 
 |  *    b. Otherwise, record the fragment (in agbno order) for later | 
 |  *       processing. | 
 |  * | 
 |  * Once we've seen all the rmaps, we know that for all blocks in the | 
 |  * refcount record we want to find $refcount owners and we've already | 
 |  * visited $seen extents that overlap all the blocks.  Therefore, we | 
 |  * need to find ($refcount - $seen) owners for every block in the | 
 |  * extent; call that quantity $target_nr.  Proceed as follows: | 
 |  * | 
 |  * 2. Pull the first $target_nr fragments from the list; all of them | 
 |  *    should start at or before the start of the extent. | 
 |  *    Call this subset of fragments the working set. | 
 |  * 3. Until there are no more unprocessed fragments, | 
 |  *    a. Find the shortest fragments in the set and remove them. | 
 |  *    b. Note the block number of the end of these fragments. | 
 |  *    c. Pull the same number of fragments from the list.  All of these | 
 |  *       fragments should start at the block number recorded in the | 
 |  *       previous step. | 
 |  *    d. Put those fragments in the set. | 
 |  * 4. Check that there are $target_nr fragments remaining in the list, | 
 |  *    and that they all end at or beyond the end of the refcount extent. | 
 |  * | 
 |  * If the refcount is correct, all the check conditions in the algorithm | 
 |  * should always hold true.  If not, the refcount is incorrect. | 
 |  */ | 
 | struct xchk_refcnt_frag { | 
 | 	struct list_head	list; | 
 | 	struct xfs_rmap_irec	rm; | 
 | }; | 
 |  | 
 | struct xchk_refcnt_check { | 
 | 	struct xfs_scrub	*sc; | 
 | 	struct list_head	fragments; | 
 |  | 
 | 	/* refcount extent we're examining */ | 
 | 	xfs_agblock_t		bno; | 
 | 	xfs_extlen_t		len; | 
 | 	xfs_nlink_t		refcount; | 
 |  | 
 | 	/* number of owners seen */ | 
 | 	xfs_nlink_t		seen; | 
 | }; | 
 |  | 
 | /* | 
 |  * Decide if the given rmap is large enough that we can redeem it | 
 |  * towards refcount verification now, or if it's a fragment, in | 
 |  * which case we'll hang onto it in the hopes that we'll later | 
 |  * discover that we've collected exactly the correct number of | 
 |  * fragments as the refcountbt says we should have. | 
 |  */ | 
 | STATIC int | 
 | xchk_refcountbt_rmap_check( | 
 | 	struct xfs_btree_cur		*cur, | 
 | 	struct xfs_rmap_irec		*rec, | 
 | 	void				*priv) | 
 | { | 
 | 	struct xchk_refcnt_check	*refchk = priv; | 
 | 	struct xchk_refcnt_frag		*frag; | 
 | 	xfs_agblock_t			rm_last; | 
 | 	xfs_agblock_t			rc_last; | 
 | 	int				error = 0; | 
 |  | 
 | 	if (xchk_should_terminate(refchk->sc, &error)) | 
 | 		return error; | 
 |  | 
 | 	rm_last = rec->rm_startblock + rec->rm_blockcount - 1; | 
 | 	rc_last = refchk->bno + refchk->len - 1; | 
 |  | 
 | 	/* Confirm that a single-owner refc extent is a CoW stage. */ | 
 | 	if (refchk->refcount == 1 && rec->rm_owner != XFS_RMAP_OWN_COW) { | 
 | 		xchk_btree_xref_set_corrupt(refchk->sc, cur, 0); | 
 | 		return 0; | 
 | 	} | 
 |  | 
 | 	if (rec->rm_startblock <= refchk->bno && rm_last >= rc_last) { | 
 | 		/* | 
 | 		 * The rmap overlaps the refcount record, so we can confirm | 
 | 		 * one refcount owner seen. | 
 | 		 */ | 
 | 		refchk->seen++; | 
 | 	} else { | 
 | 		/* | 
 | 		 * This rmap covers only part of the refcount record, so | 
 | 		 * save the fragment for later processing.  If the rmapbt | 
 | 		 * is healthy each rmap_irec we see will be in agbno order | 
 | 		 * so we don't need insertion sort here. | 
 | 		 */ | 
 | 		frag = kmem_alloc(sizeof(struct xchk_refcnt_frag), | 
 | 				KM_MAYFAIL); | 
 | 		if (!frag) | 
 | 			return -ENOMEM; | 
 | 		memcpy(&frag->rm, rec, sizeof(frag->rm)); | 
 | 		list_add_tail(&frag->list, &refchk->fragments); | 
 | 	} | 
 |  | 
 | 	return 0; | 
 | } | 
 |  | 
 | /* | 
 |  * Given a bunch of rmap fragments, iterate through them, keeping | 
 |  * a running tally of the refcount.  If this ever deviates from | 
 |  * what we expect (which is the refcountbt's refcount minus the | 
 |  * number of extents that totally covered the refcountbt extent), | 
 |  * we have a refcountbt error. | 
 |  */ | 
 | STATIC void | 
 | xchk_refcountbt_process_rmap_fragments( | 
 | 	struct xchk_refcnt_check	*refchk) | 
 | { | 
 | 	struct list_head		worklist; | 
 | 	struct xchk_refcnt_frag		*frag; | 
 | 	struct xchk_refcnt_frag		*n; | 
 | 	xfs_agblock_t			bno; | 
 | 	xfs_agblock_t			rbno; | 
 | 	xfs_agblock_t			next_rbno; | 
 | 	xfs_nlink_t			nr; | 
 | 	xfs_nlink_t			target_nr; | 
 |  | 
 | 	target_nr = refchk->refcount - refchk->seen; | 
 | 	if (target_nr == 0) | 
 | 		return; | 
 |  | 
 | 	/* | 
 | 	 * There are (refchk->rc.rc_refcount - refchk->nr refcount) | 
 | 	 * references we haven't found yet.  Pull that many off the | 
 | 	 * fragment list and figure out where the smallest rmap ends | 
 | 	 * (and therefore the next rmap should start).  All the rmaps | 
 | 	 * we pull off should start at or before the beginning of the | 
 | 	 * refcount record's range. | 
 | 	 */ | 
 | 	INIT_LIST_HEAD(&worklist); | 
 | 	rbno = NULLAGBLOCK; | 
 | 	nr = 1; | 
 |  | 
 | 	/* Make sure the fragments actually /are/ in agbno order. */ | 
 | 	bno = 0; | 
 | 	list_for_each_entry(frag, &refchk->fragments, list) { | 
 | 		if (frag->rm.rm_startblock < bno) | 
 | 			goto done; | 
 | 		bno = frag->rm.rm_startblock; | 
 | 	} | 
 |  | 
 | 	/* | 
 | 	 * Find all the rmaps that start at or before the refc extent, | 
 | 	 * and put them on the worklist. | 
 | 	 */ | 
 | 	list_for_each_entry_safe(frag, n, &refchk->fragments, list) { | 
 | 		if (frag->rm.rm_startblock > refchk->bno) | 
 | 			goto done; | 
 | 		bno = frag->rm.rm_startblock + frag->rm.rm_blockcount; | 
 | 		if (bno < rbno) | 
 | 			rbno = bno; | 
 | 		list_move_tail(&frag->list, &worklist); | 
 | 		if (nr == target_nr) | 
 | 			break; | 
 | 		nr++; | 
 | 	} | 
 |  | 
 | 	/* | 
 | 	 * We should have found exactly $target_nr rmap fragments starting | 
 | 	 * at or before the refcount extent. | 
 | 	 */ | 
 | 	if (nr != target_nr) | 
 | 		goto done; | 
 |  | 
 | 	while (!list_empty(&refchk->fragments)) { | 
 | 		/* Discard any fragments ending at rbno from the worklist. */ | 
 | 		nr = 0; | 
 | 		next_rbno = NULLAGBLOCK; | 
 | 		list_for_each_entry_safe(frag, n, &worklist, list) { | 
 | 			bno = frag->rm.rm_startblock + frag->rm.rm_blockcount; | 
 | 			if (bno != rbno) { | 
 | 				if (bno < next_rbno) | 
 | 					next_rbno = bno; | 
 | 				continue; | 
 | 			} | 
 | 			list_del(&frag->list); | 
 | 			kmem_free(frag); | 
 | 			nr++; | 
 | 		} | 
 |  | 
 | 		/* Try to add nr rmaps starting at rbno to the worklist. */ | 
 | 		list_for_each_entry_safe(frag, n, &refchk->fragments, list) { | 
 | 			bno = frag->rm.rm_startblock + frag->rm.rm_blockcount; | 
 | 			if (frag->rm.rm_startblock != rbno) | 
 | 				goto done; | 
 | 			list_move_tail(&frag->list, &worklist); | 
 | 			if (next_rbno > bno) | 
 | 				next_rbno = bno; | 
 | 			nr--; | 
 | 			if (nr == 0) | 
 | 				break; | 
 | 		} | 
 |  | 
 | 		/* | 
 | 		 * If we get here and nr > 0, this means that we added fewer | 
 | 		 * items to the worklist than we discarded because the fragment | 
 | 		 * list ran out of items.  Therefore, we cannot maintain the | 
 | 		 * required refcount.  Something is wrong, so we're done. | 
 | 		 */ | 
 | 		if (nr) | 
 | 			goto done; | 
 |  | 
 | 		rbno = next_rbno; | 
 | 	} | 
 |  | 
 | 	/* | 
 | 	 * Make sure the last extent we processed ends at or beyond | 
 | 	 * the end of the refcount extent. | 
 | 	 */ | 
 | 	if (rbno < refchk->bno + refchk->len) | 
 | 		goto done; | 
 |  | 
 | 	/* Actually record us having seen the remaining refcount. */ | 
 | 	refchk->seen = refchk->refcount; | 
 | done: | 
 | 	/* Delete fragments and work list. */ | 
 | 	list_for_each_entry_safe(frag, n, &worklist, list) { | 
 | 		list_del(&frag->list); | 
 | 		kmem_free(frag); | 
 | 	} | 
 | 	list_for_each_entry_safe(frag, n, &refchk->fragments, list) { | 
 | 		list_del(&frag->list); | 
 | 		kmem_free(frag); | 
 | 	} | 
 | } | 
 |  | 
 | /* Use the rmap entries covering this extent to verify the refcount. */ | 
 | STATIC void | 
 | xchk_refcountbt_xref_rmap( | 
 | 	struct xfs_scrub		*sc, | 
 | 	xfs_agblock_t			bno, | 
 | 	xfs_extlen_t			len, | 
 | 	xfs_nlink_t			refcount) | 
 | { | 
 | 	struct xchk_refcnt_check	refchk = { | 
 | 		.sc = sc, | 
 | 		.bno = bno, | 
 | 		.len = len, | 
 | 		.refcount = refcount, | 
 | 		.seen = 0, | 
 | 	}; | 
 | 	struct xfs_rmap_irec		low; | 
 | 	struct xfs_rmap_irec		high; | 
 | 	struct xchk_refcnt_frag		*frag; | 
 | 	struct xchk_refcnt_frag		*n; | 
 | 	int				error; | 
 |  | 
 | 	if (!sc->sa.rmap_cur || xchk_skip_xref(sc->sm)) | 
 | 		return; | 
 |  | 
 | 	/* Cross-reference with the rmapbt to confirm the refcount. */ | 
 | 	memset(&low, 0, sizeof(low)); | 
 | 	low.rm_startblock = bno; | 
 | 	memset(&high, 0xFF, sizeof(high)); | 
 | 	high.rm_startblock = bno + len - 1; | 
 |  | 
 | 	INIT_LIST_HEAD(&refchk.fragments); | 
 | 	error = xfs_rmap_query_range(sc->sa.rmap_cur, &low, &high, | 
 | 			&xchk_refcountbt_rmap_check, &refchk); | 
 | 	if (!xchk_should_check_xref(sc, &error, &sc->sa.rmap_cur)) | 
 | 		goto out_free; | 
 |  | 
 | 	xchk_refcountbt_process_rmap_fragments(&refchk); | 
 | 	if (refcount != refchk.seen) | 
 | 		xchk_btree_xref_set_corrupt(sc, sc->sa.rmap_cur, 0); | 
 |  | 
 | out_free: | 
 | 	list_for_each_entry_safe(frag, n, &refchk.fragments, list) { | 
 | 		list_del(&frag->list); | 
 | 		kmem_free(frag); | 
 | 	} | 
 | } | 
 |  | 
 | /* Cross-reference with the other btrees. */ | 
 | STATIC void | 
 | xchk_refcountbt_xref( | 
 | 	struct xfs_scrub	*sc, | 
 | 	xfs_agblock_t		agbno, | 
 | 	xfs_extlen_t		len, | 
 | 	xfs_nlink_t		refcount) | 
 | { | 
 | 	if (sc->sm->sm_flags & XFS_SCRUB_OFLAG_CORRUPT) | 
 | 		return; | 
 |  | 
 | 	xchk_xref_is_used_space(sc, agbno, len); | 
 | 	xchk_xref_is_not_inode_chunk(sc, agbno, len); | 
 | 	xchk_refcountbt_xref_rmap(sc, agbno, len, refcount); | 
 | } | 
 |  | 
 | /* Scrub a refcountbt record. */ | 
 | STATIC int | 
 | xchk_refcountbt_rec( | 
 | 	struct xchk_btree	*bs, | 
 | 	union xfs_btree_rec	*rec) | 
 | { | 
 | 	struct xfs_mount	*mp = bs->cur->bc_mp; | 
 | 	xfs_agblock_t		*cow_blocks = bs->private; | 
 | 	xfs_agnumber_t		agno = bs->cur->bc_private.a.agno; | 
 | 	xfs_agblock_t		bno; | 
 | 	xfs_extlen_t		len; | 
 | 	xfs_nlink_t		refcount; | 
 | 	bool			has_cowflag; | 
 | 	int			error = 0; | 
 |  | 
 | 	bno = be32_to_cpu(rec->refc.rc_startblock); | 
 | 	len = be32_to_cpu(rec->refc.rc_blockcount); | 
 | 	refcount = be32_to_cpu(rec->refc.rc_refcount); | 
 |  | 
 | 	/* Only CoW records can have refcount == 1. */ | 
 | 	has_cowflag = (bno & XFS_REFC_COW_START); | 
 | 	if ((refcount == 1 && !has_cowflag) || (refcount != 1 && has_cowflag)) | 
 | 		xchk_btree_set_corrupt(bs->sc, bs->cur, 0); | 
 | 	if (has_cowflag) | 
 | 		(*cow_blocks) += len; | 
 |  | 
 | 	/* Check the extent. */ | 
 | 	bno &= ~XFS_REFC_COW_START; | 
 | 	if (bno + len <= bno || | 
 | 	    !xfs_verify_agbno(mp, agno, bno) || | 
 | 	    !xfs_verify_agbno(mp, agno, bno + len - 1)) | 
 | 		xchk_btree_set_corrupt(bs->sc, bs->cur, 0); | 
 |  | 
 | 	if (refcount == 0) | 
 | 		xchk_btree_set_corrupt(bs->sc, bs->cur, 0); | 
 |  | 
 | 	xchk_refcountbt_xref(bs->sc, bno, len, refcount); | 
 |  | 
 | 	return error; | 
 | } | 
 |  | 
 | /* Make sure we have as many refc blocks as the rmap says. */ | 
 | STATIC void | 
 | xchk_refcount_xref_rmap( | 
 | 	struct xfs_scrub	*sc, | 
 | 	struct xfs_owner_info	*oinfo, | 
 | 	xfs_filblks_t		cow_blocks) | 
 | { | 
 | 	xfs_extlen_t		refcbt_blocks = 0; | 
 | 	xfs_filblks_t		blocks; | 
 | 	int			error; | 
 |  | 
 | 	if (!sc->sa.rmap_cur || xchk_skip_xref(sc->sm)) | 
 | 		return; | 
 |  | 
 | 	/* Check that we saw as many refcbt blocks as the rmap knows about. */ | 
 | 	error = xfs_btree_count_blocks(sc->sa.refc_cur, &refcbt_blocks); | 
 | 	if (!xchk_btree_process_error(sc, sc->sa.refc_cur, 0, &error)) | 
 | 		return; | 
 | 	error = xchk_count_rmap_ownedby_ag(sc, sc->sa.rmap_cur, oinfo, | 
 | 			&blocks); | 
 | 	if (!xchk_should_check_xref(sc, &error, &sc->sa.rmap_cur)) | 
 | 		return; | 
 | 	if (blocks != refcbt_blocks) | 
 | 		xchk_btree_xref_set_corrupt(sc, sc->sa.rmap_cur, 0); | 
 |  | 
 | 	/* Check that we saw as many cow blocks as the rmap knows about. */ | 
 | 	xfs_rmap_ag_owner(oinfo, XFS_RMAP_OWN_COW); | 
 | 	error = xchk_count_rmap_ownedby_ag(sc, sc->sa.rmap_cur, oinfo, | 
 | 			&blocks); | 
 | 	if (!xchk_should_check_xref(sc, &error, &sc->sa.rmap_cur)) | 
 | 		return; | 
 | 	if (blocks != cow_blocks) | 
 | 		xchk_btree_xref_set_corrupt(sc, sc->sa.rmap_cur, 0); | 
 | } | 
 |  | 
 | /* Scrub the refcount btree for some AG. */ | 
 | int | 
 | xchk_refcountbt( | 
 | 	struct xfs_scrub	*sc) | 
 | { | 
 | 	struct xfs_owner_info	oinfo; | 
 | 	xfs_agblock_t		cow_blocks = 0; | 
 | 	int			error; | 
 |  | 
 | 	xfs_rmap_ag_owner(&oinfo, XFS_RMAP_OWN_REFC); | 
 | 	error = xchk_btree(sc, sc->sa.refc_cur, xchk_refcountbt_rec, | 
 | 			&oinfo, &cow_blocks); | 
 | 	if (error) | 
 | 		return error; | 
 |  | 
 | 	xchk_refcount_xref_rmap(sc, &oinfo, cow_blocks); | 
 |  | 
 | 	return 0; | 
 | } | 
 |  | 
 | /* xref check that a cow staging extent is marked in the refcountbt. */ | 
 | void | 
 | xchk_xref_is_cow_staging( | 
 | 	struct xfs_scrub		*sc, | 
 | 	xfs_agblock_t			agbno, | 
 | 	xfs_extlen_t			len) | 
 | { | 
 | 	struct xfs_refcount_irec	rc; | 
 | 	bool				has_cowflag; | 
 | 	int				has_refcount; | 
 | 	int				error; | 
 |  | 
 | 	if (!sc->sa.refc_cur || xchk_skip_xref(sc->sm)) | 
 | 		return; | 
 |  | 
 | 	/* Find the CoW staging extent. */ | 
 | 	error = xfs_refcount_lookup_le(sc->sa.refc_cur, | 
 | 			agbno + XFS_REFC_COW_START, &has_refcount); | 
 | 	if (!xchk_should_check_xref(sc, &error, &sc->sa.refc_cur)) | 
 | 		return; | 
 | 	if (!has_refcount) { | 
 | 		xchk_btree_xref_set_corrupt(sc, sc->sa.refc_cur, 0); | 
 | 		return; | 
 | 	} | 
 |  | 
 | 	error = xfs_refcount_get_rec(sc->sa.refc_cur, &rc, &has_refcount); | 
 | 	if (!xchk_should_check_xref(sc, &error, &sc->sa.refc_cur)) | 
 | 		return; | 
 | 	if (!has_refcount) { | 
 | 		xchk_btree_xref_set_corrupt(sc, sc->sa.refc_cur, 0); | 
 | 		return; | 
 | 	} | 
 |  | 
 | 	/* CoW flag must be set, refcount must be 1. */ | 
 | 	has_cowflag = (rc.rc_startblock & XFS_REFC_COW_START); | 
 | 	if (!has_cowflag || rc.rc_refcount != 1) | 
 | 		xchk_btree_xref_set_corrupt(sc, sc->sa.refc_cur, 0); | 
 |  | 
 | 	/* Must be at least as long as what was passed in */ | 
 | 	if (rc.rc_blockcount < len) | 
 | 		xchk_btree_xref_set_corrupt(sc, sc->sa.refc_cur, 0); | 
 | } | 
 |  | 
 | /* | 
 |  * xref check that the extent is not shared.  Only file data blocks | 
 |  * can have multiple owners. | 
 |  */ | 
 | void | 
 | xchk_xref_is_not_shared( | 
 | 	struct xfs_scrub	*sc, | 
 | 	xfs_agblock_t		agbno, | 
 | 	xfs_extlen_t		len) | 
 | { | 
 | 	bool			shared; | 
 | 	int			error; | 
 |  | 
 | 	if (!sc->sa.refc_cur || xchk_skip_xref(sc->sm)) | 
 | 		return; | 
 |  | 
 | 	error = xfs_refcount_has_record(sc->sa.refc_cur, agbno, len, &shared); | 
 | 	if (!xchk_should_check_xref(sc, &error, &sc->sa.refc_cur)) | 
 | 		return; | 
 | 	if (shared) | 
 | 		xchk_btree_xref_set_corrupt(sc, sc->sa.refc_cur, 0); | 
 | } |