Author | Tokens | Token Proportion | Commits | Commit Proportion |
---|---|---|---|---|
Darrick J. Wong | 6544 | 91.22% | 42 | 54.55% |
David Chinner | 548 | 7.64% | 18 | 23.38% |
Christoph Hellwig | 52 | 0.72% | 9 | 11.69% |
Nathan Scott | 11 | 0.15% | 2 | 2.60% |
Michal Marek | 8 | 0.11% | 1 | 1.30% |
Russell Cattelan | 4 | 0.06% | 1 | 1.30% |
Eric Sandeen | 3 | 0.04% | 1 | 1.30% |
Brian Foster | 3 | 0.04% | 2 | 2.60% |
Chen Ni | 1 | 0.01% | 1 | 1.30% |
Total | 7174 | 77 |
// SPDX-License-Identifier: GPL-2.0-or-later /* * Copyright (C) 2018-2023 Oracle. All Rights Reserved. * Author: Darrick J. Wong <djwong@kernel.org> */ #include "xfs.h" #include "xfs_fs.h" #include "xfs_shared.h" #include "xfs_format.h" #include "xfs_trans_resv.h" #include "xfs_mount.h" #include "xfs_btree.h" #include "xfs_log_format.h" #include "xfs_trans.h" #include "xfs_sb.h" #include "xfs_alloc.h" #include "xfs_alloc_btree.h" #include "xfs_ialloc.h" #include "xfs_ialloc_btree.h" #include "xfs_rmap.h" #include "xfs_rmap_btree.h" #include "xfs_refcount_btree.h" #include "xfs_ag.h" #include "xfs_inode.h" #include "xfs_iunlink_item.h" #include "scrub/scrub.h" #include "scrub/common.h" #include "scrub/trace.h" #include "scrub/repair.h" #include "scrub/bitmap.h" #include "scrub/agb_bitmap.h" #include "scrub/agino_bitmap.h" #include "scrub/reap.h" #include "scrub/xfile.h" #include "scrub/xfarray.h" /* Superblock */ /* Repair the superblock. */ int xrep_superblock( struct xfs_scrub *sc) { struct xfs_mount *mp = sc->mp; struct xfs_buf *bp; xfs_agnumber_t agno; int error; /* Don't try to repair AG 0's sb; let xfs_repair deal with it. */ agno = sc->sm->sm_agno; if (agno == 0) return -EOPNOTSUPP; error = xfs_sb_get_secondary(mp, sc->tp, agno, &bp); if (error) return error; /* Last chance to abort before we start committing fixes. */ if (xchk_should_terminate(sc, &error)) return error; /* Copy AG 0's superblock to this one. */ xfs_buf_zero(bp, 0, BBTOB(bp->b_length)); xfs_sb_to_disk(bp->b_addr, &mp->m_sb); /* * Don't write out a secondary super with NEEDSREPAIR or log incompat * features set, since both are ignored when set on a secondary. */ if (xfs_has_crc(mp)) { struct xfs_dsb *sb = bp->b_addr; sb->sb_features_incompat &= ~cpu_to_be32(XFS_SB_FEAT_INCOMPAT_NEEDSREPAIR); sb->sb_features_log_incompat = 0; } /* Write this to disk. */ xfs_trans_buf_set_type(sc->tp, bp, XFS_BLFT_SB_BUF); xfs_trans_log_buf(sc->tp, bp, 0, BBTOB(bp->b_length) - 1); return 0; } /* AGF */ struct xrep_agf_allocbt { struct xfs_scrub *sc; xfs_agblock_t freeblks; xfs_agblock_t longest; }; /* Record free space shape information. */ STATIC int xrep_agf_walk_allocbt( struct xfs_btree_cur *cur, const struct xfs_alloc_rec_incore *rec, void *priv) { struct xrep_agf_allocbt *raa = priv; int error = 0; if (xchk_should_terminate(raa->sc, &error)) return error; raa->freeblks += rec->ar_blockcount; if (rec->ar_blockcount > raa->longest) raa->longest = rec->ar_blockcount; return error; } /* Does this AGFL block look sane? */ STATIC int xrep_agf_check_agfl_block( struct xfs_mount *mp, xfs_agblock_t agbno, void *priv) { struct xfs_scrub *sc = priv; if (!xfs_verify_agbno(sc->sa.pag, agbno)) return -EFSCORRUPTED; return 0; } /* * Offset within the xrep_find_ag_btree array for each btree type. Avoid the * XFS_BTNUM_ names here to avoid creating a sparse array. */ enum { XREP_AGF_BNOBT = 0, XREP_AGF_CNTBT, XREP_AGF_RMAPBT, XREP_AGF_REFCOUNTBT, XREP_AGF_END, XREP_AGF_MAX }; /* Check a btree root candidate. */ static inline bool xrep_check_btree_root( struct xfs_scrub *sc, struct xrep_find_ag_btree *fab) { return xfs_verify_agbno(sc->sa.pag, fab->root) && fab->height <= fab->maxlevels; } /* * Given the btree roots described by *fab, find the roots, check them for * sanity, and pass the root data back out via *fab. * * This is /also/ a chicken and egg problem because we have to use the rmapbt * (rooted in the AGF) to find the btrees rooted in the AGF. We also have no * idea if the btrees make any sense. If we hit obvious corruptions in those * btrees we'll bail out. */ STATIC int xrep_agf_find_btrees( struct xfs_scrub *sc, struct xfs_buf *agf_bp, struct xrep_find_ag_btree *fab, struct xfs_buf *agfl_bp) { struct xfs_agf *old_agf = agf_bp->b_addr; int error; /* Go find the root data. */ error = xrep_find_ag_btree_roots(sc, agf_bp, fab, agfl_bp); if (error) return error; /* We must find the bnobt, cntbt, and rmapbt roots. */ if (!xrep_check_btree_root(sc, &fab[XREP_AGF_BNOBT]) || !xrep_check_btree_root(sc, &fab[XREP_AGF_CNTBT]) || !xrep_check_btree_root(sc, &fab[XREP_AGF_RMAPBT])) return -EFSCORRUPTED; /* * We relied on the rmapbt to reconstruct the AGF. If we get a * different root then something's seriously wrong. */ if (fab[XREP_AGF_RMAPBT].root != be32_to_cpu(old_agf->agf_rmap_root)) return -EFSCORRUPTED; /* We must find the refcountbt root if that feature is enabled. */ if (xfs_has_reflink(sc->mp) && !xrep_check_btree_root(sc, &fab[XREP_AGF_REFCOUNTBT])) return -EFSCORRUPTED; return 0; } /* * Reinitialize the AGF header, making an in-core copy of the old contents so * that we know which in-core state needs to be reinitialized. */ STATIC void xrep_agf_init_header( struct xfs_scrub *sc, struct xfs_buf *agf_bp, struct xfs_agf *old_agf) { struct xfs_mount *mp = sc->mp; struct xfs_perag *pag = sc->sa.pag; struct xfs_agf *agf = agf_bp->b_addr; memcpy(old_agf, agf, sizeof(*old_agf)); memset(agf, 0, BBTOB(agf_bp->b_length)); agf->agf_magicnum = cpu_to_be32(XFS_AGF_MAGIC); agf->agf_versionnum = cpu_to_be32(XFS_AGF_VERSION); agf->agf_seqno = cpu_to_be32(pag->pag_agno); agf->agf_length = cpu_to_be32(pag->block_count); agf->agf_flfirst = old_agf->agf_flfirst; agf->agf_fllast = old_agf->agf_fllast; agf->agf_flcount = old_agf->agf_flcount; if (xfs_has_crc(mp)) uuid_copy(&agf->agf_uuid, &mp->m_sb.sb_meta_uuid); /* Mark the incore AGF data stale until we're done fixing things. */ ASSERT(xfs_perag_initialised_agf(pag)); clear_bit(XFS_AGSTATE_AGF_INIT, &pag->pag_opstate); } /* Set btree root information in an AGF. */ STATIC void xrep_agf_set_roots( struct xfs_scrub *sc, struct xfs_agf *agf, struct xrep_find_ag_btree *fab) { agf->agf_bno_root = cpu_to_be32(fab[XREP_AGF_BNOBT].root); agf->agf_bno_level = cpu_to_be32(fab[XREP_AGF_BNOBT].height); agf->agf_cnt_root = cpu_to_be32(fab[XREP_AGF_CNTBT].root); agf->agf_cnt_level = cpu_to_be32(fab[XREP_AGF_CNTBT].height); agf->agf_rmap_root = cpu_to_be32(fab[XREP_AGF_RMAPBT].root); agf->agf_rmap_level = cpu_to_be32(fab[XREP_AGF_RMAPBT].height); if (xfs_has_reflink(sc->mp)) { agf->agf_refcount_root = cpu_to_be32(fab[XREP_AGF_REFCOUNTBT].root); agf->agf_refcount_level = cpu_to_be32(fab[XREP_AGF_REFCOUNTBT].height); } } /* Update all AGF fields which derive from btree contents. */ STATIC int xrep_agf_calc_from_btrees( struct xfs_scrub *sc, struct xfs_buf *agf_bp) { struct xrep_agf_allocbt raa = { .sc = sc }; struct xfs_btree_cur *cur = NULL; struct xfs_agf *agf = agf_bp->b_addr; struct xfs_mount *mp = sc->mp; xfs_agblock_t btreeblks; xfs_agblock_t blocks; int error; /* Update the AGF counters from the bnobt. */ cur = xfs_bnobt_init_cursor(mp, sc->tp, agf_bp, sc->sa.pag); error = xfs_alloc_query_all(cur, xrep_agf_walk_allocbt, &raa); if (error) goto err; error = xfs_btree_count_blocks(cur, &blocks); if (error) goto err; xfs_btree_del_cursor(cur, error); btreeblks = blocks - 1; agf->agf_freeblks = cpu_to_be32(raa.freeblks); agf->agf_longest = cpu_to_be32(raa.longest); /* Update the AGF counters from the cntbt. */ cur = xfs_cntbt_init_cursor(mp, sc->tp, agf_bp, sc->sa.pag); error = xfs_btree_count_blocks(cur, &blocks); if (error) goto err; xfs_btree_del_cursor(cur, error); btreeblks += blocks - 1; /* Update the AGF counters from the rmapbt. */ cur = xfs_rmapbt_init_cursor(mp, sc->tp, agf_bp, sc->sa.pag); error = xfs_btree_count_blocks(cur, &blocks); if (error) goto err; xfs_btree_del_cursor(cur, error); agf->agf_rmap_blocks = cpu_to_be32(blocks); btreeblks += blocks - 1; agf->agf_btreeblks = cpu_to_be32(btreeblks); /* Update the AGF counters from the refcountbt. */ if (xfs_has_reflink(mp)) { cur = xfs_refcountbt_init_cursor(mp, sc->tp, agf_bp, sc->sa.pag); error = xfs_btree_count_blocks(cur, &blocks); if (error) goto err; xfs_btree_del_cursor(cur, error); agf->agf_refcount_blocks = cpu_to_be32(blocks); } return 0; err: xfs_btree_del_cursor(cur, error); return error; } /* Commit the new AGF and reinitialize the incore state. */ STATIC int xrep_agf_commit_new( struct xfs_scrub *sc, struct xfs_buf *agf_bp) { struct xfs_perag *pag; struct xfs_agf *agf = agf_bp->b_addr; /* Trigger fdblocks recalculation */ xfs_force_summary_recalc(sc->mp); /* Write this to disk. */ xfs_trans_buf_set_type(sc->tp, agf_bp, XFS_BLFT_AGF_BUF); xfs_trans_log_buf(sc->tp, agf_bp, 0, BBTOB(agf_bp->b_length) - 1); /* Now reinitialize the in-core counters we changed. */ pag = sc->sa.pag; pag->pagf_btreeblks = be32_to_cpu(agf->agf_btreeblks); pag->pagf_freeblks = be32_to_cpu(agf->agf_freeblks); pag->pagf_longest = be32_to_cpu(agf->agf_longest); pag->pagf_bno_level = be32_to_cpu(agf->agf_bno_level); pag->pagf_cnt_level = be32_to_cpu(agf->agf_cnt_level); pag->pagf_rmap_level = be32_to_cpu(agf->agf_rmap_level); pag->pagf_refcount_level = be32_to_cpu(agf->agf_refcount_level); set_bit(XFS_AGSTATE_AGF_INIT, &pag->pag_opstate); return xrep_roll_ag_trans(sc); } /* Repair the AGF. v5 filesystems only. */ int xrep_agf( struct xfs_scrub *sc) { struct xrep_find_ag_btree fab[XREP_AGF_MAX] = { [XREP_AGF_BNOBT] = { .rmap_owner = XFS_RMAP_OWN_AG, .buf_ops = &xfs_bnobt_buf_ops, .maxlevels = sc->mp->m_alloc_maxlevels, }, [XREP_AGF_CNTBT] = { .rmap_owner = XFS_RMAP_OWN_AG, .buf_ops = &xfs_cntbt_buf_ops, .maxlevels = sc->mp->m_alloc_maxlevels, }, [XREP_AGF_RMAPBT] = { .rmap_owner = XFS_RMAP_OWN_AG, .buf_ops = &xfs_rmapbt_buf_ops, .maxlevels = sc->mp->m_rmap_maxlevels, }, [XREP_AGF_REFCOUNTBT] = { .rmap_owner = XFS_RMAP_OWN_REFC, .buf_ops = &xfs_refcountbt_buf_ops, .maxlevels = sc->mp->m_refc_maxlevels, }, [XREP_AGF_END] = { .buf_ops = NULL, }, }; struct xfs_agf old_agf; struct xfs_mount *mp = sc->mp; struct xfs_buf *agf_bp; struct xfs_buf *agfl_bp; struct xfs_agf *agf; int error; /* We require the rmapbt to rebuild anything. */ if (!xfs_has_rmapbt(mp)) return -EOPNOTSUPP; /* * Make sure we have the AGF buffer, as scrub might have decided it * was corrupt after xfs_alloc_read_agf failed with -EFSCORRUPTED. */ error = xfs_trans_read_buf(mp, sc->tp, mp->m_ddev_targp, XFS_AG_DADDR(mp, sc->sa.pag->pag_agno, XFS_AGF_DADDR(mp)), XFS_FSS_TO_BB(mp, 1), 0, &agf_bp, NULL); if (error) return error; agf_bp->b_ops = &xfs_agf_buf_ops; agf = agf_bp->b_addr; /* * Load the AGFL so that we can screen out OWN_AG blocks that are on * the AGFL now; these blocks might have once been part of the * bno/cnt/rmap btrees but are not now. This is a chicken and egg * problem: the AGF is corrupt, so we have to trust the AGFL contents * because we can't do any serious cross-referencing with any of the * btrees rooted in the AGF. If the AGFL contents are obviously bad * then we'll bail out. */ error = xfs_alloc_read_agfl(sc->sa.pag, sc->tp, &agfl_bp); if (error) return error; /* * Spot-check the AGFL blocks; if they're obviously corrupt then * there's nothing we can do but bail out. */ error = xfs_agfl_walk(sc->mp, agf_bp->b_addr, agfl_bp, xrep_agf_check_agfl_block, sc); if (error) return error; /* * Find the AGF btree roots. This is also a chicken-and-egg situation; * see the function for more details. */ error = xrep_agf_find_btrees(sc, agf_bp, fab, agfl_bp); if (error) return error; /* Last chance to abort before we start committing fixes. */ if (xchk_should_terminate(sc, &error)) return error; /* Start rewriting the header and implant the btrees we found. */ xrep_agf_init_header(sc, agf_bp, &old_agf); xrep_agf_set_roots(sc, agf, fab); error = xrep_agf_calc_from_btrees(sc, agf_bp); if (error) goto out_revert; /* Commit the changes and reinitialize incore state. */ return xrep_agf_commit_new(sc, agf_bp); out_revert: /* Mark the incore AGF state stale and revert the AGF. */ clear_bit(XFS_AGSTATE_AGF_INIT, &sc->sa.pag->pag_opstate); memcpy(agf, &old_agf, sizeof(old_agf)); return error; } /* AGFL */ struct xrep_agfl { /* Bitmap of alleged AGFL blocks that we're not going to add. */ struct xagb_bitmap crossed; /* Bitmap of other OWN_AG metadata blocks. */ struct xagb_bitmap agmetablocks; /* Bitmap of free space. */ struct xagb_bitmap *freesp; /* rmapbt cursor for finding crosslinked blocks */ struct xfs_btree_cur *rmap_cur; struct xfs_scrub *sc; }; /* Record all OWN_AG (free space btree) information from the rmap data. */ STATIC int xrep_agfl_walk_rmap( struct xfs_btree_cur *cur, const struct xfs_rmap_irec *rec, void *priv) { struct xrep_agfl *ra = priv; int error = 0; if (xchk_should_terminate(ra->sc, &error)) return error; /* Record all the OWN_AG blocks. */ if (rec->rm_owner == XFS_RMAP_OWN_AG) { error = xagb_bitmap_set(ra->freesp, rec->rm_startblock, rec->rm_blockcount); if (error) return error; } return xagb_bitmap_set_btcur_path(&ra->agmetablocks, cur); } /* Strike out the blocks that are cross-linked according to the rmapbt. */ STATIC int xrep_agfl_check_extent( uint32_t agbno, uint32_t len, void *priv) { struct xrep_agfl *ra = priv; xfs_agblock_t last_agbno = agbno + len - 1; int error; while (agbno <= last_agbno) { bool other_owners; error = xfs_rmap_has_other_keys(ra->rmap_cur, agbno, 1, &XFS_RMAP_OINFO_AG, &other_owners); if (error) return error; if (other_owners) { error = xagb_bitmap_set(&ra->crossed, agbno, 1); if (error) return error; } if (xchk_should_terminate(ra->sc, &error)) return error; agbno++; } return 0; } /* * Map out all the non-AGFL OWN_AG space in this AG so that we can deduce * which blocks belong to the AGFL. * * Compute the set of old AGFL blocks by subtracting from the list of OWN_AG * blocks the list of blocks owned by all other OWN_AG metadata (bnobt, cntbt, * rmapbt). These are the old AGFL blocks, so return that list and the number * of blocks we're actually going to put back on the AGFL. */ STATIC int xrep_agfl_collect_blocks( struct xfs_scrub *sc, struct xfs_buf *agf_bp, struct xagb_bitmap *agfl_extents, xfs_agblock_t *flcount) { struct xrep_agfl ra; struct xfs_mount *mp = sc->mp; struct xfs_btree_cur *cur; int error; ra.sc = sc; ra.freesp = agfl_extents; xagb_bitmap_init(&ra.agmetablocks); xagb_bitmap_init(&ra.crossed); /* Find all space used by the free space btrees & rmapbt. */ cur = xfs_rmapbt_init_cursor(mp, sc->tp, agf_bp, sc->sa.pag); error = xfs_rmap_query_all(cur, xrep_agfl_walk_rmap, &ra); xfs_btree_del_cursor(cur, error); if (error) goto out_bmp; /* Find all blocks currently being used by the bnobt. */ cur = xfs_bnobt_init_cursor(mp, sc->tp, agf_bp, sc->sa.pag); error = xagb_bitmap_set_btblocks(&ra.agmetablocks, cur); xfs_btree_del_cursor(cur, error); if (error) goto out_bmp; /* Find all blocks currently being used by the cntbt. */ cur = xfs_cntbt_init_cursor(mp, sc->tp, agf_bp, sc->sa.pag); error = xagb_bitmap_set_btblocks(&ra.agmetablocks, cur); xfs_btree_del_cursor(cur, error); if (error) goto out_bmp; /* * Drop the freesp meta blocks that are in use by btrees. * The remaining blocks /should/ be AGFL blocks. */ error = xagb_bitmap_disunion(agfl_extents, &ra.agmetablocks); if (error) goto out_bmp; /* Strike out the blocks that are cross-linked. */ ra.rmap_cur = xfs_rmapbt_init_cursor(mp, sc->tp, agf_bp, sc->sa.pag); error = xagb_bitmap_walk(agfl_extents, xrep_agfl_check_extent, &ra); xfs_btree_del_cursor(ra.rmap_cur, error); if (error) goto out_bmp; error = xagb_bitmap_disunion(agfl_extents, &ra.crossed); if (error) goto out_bmp; /* * Calculate the new AGFL size. If we found more blocks than fit in * the AGFL we'll free them later. */ *flcount = min_t(uint64_t, xagb_bitmap_hweight(agfl_extents), xfs_agfl_size(mp)); out_bmp: xagb_bitmap_destroy(&ra.crossed); xagb_bitmap_destroy(&ra.agmetablocks); return error; } /* Update the AGF and reset the in-core state. */ STATIC void xrep_agfl_update_agf( struct xfs_scrub *sc, struct xfs_buf *agf_bp, xfs_agblock_t flcount) { struct xfs_agf *agf = agf_bp->b_addr; ASSERT(flcount <= xfs_agfl_size(sc->mp)); /* Trigger fdblocks recalculation */ xfs_force_summary_recalc(sc->mp); /* Update the AGF counters. */ if (xfs_perag_initialised_agf(sc->sa.pag)) { sc->sa.pag->pagf_flcount = flcount; clear_bit(XFS_AGSTATE_AGFL_NEEDS_RESET, &sc->sa.pag->pag_opstate); } agf->agf_flfirst = cpu_to_be32(0); agf->agf_flcount = cpu_to_be32(flcount); if (flcount) agf->agf_fllast = cpu_to_be32(flcount - 1); else agf->agf_fllast = cpu_to_be32(xfs_agfl_size(sc->mp) - 1); xfs_alloc_log_agf(sc->tp, agf_bp, XFS_AGF_FLFIRST | XFS_AGF_FLLAST | XFS_AGF_FLCOUNT); } struct xrep_agfl_fill { struct xagb_bitmap used_extents; struct xfs_scrub *sc; __be32 *agfl_bno; xfs_agblock_t flcount; unsigned int fl_off; }; /* Fill the AGFL with whatever blocks are in this extent. */ static int xrep_agfl_fill( uint32_t start, uint32_t len, void *priv) { struct xrep_agfl_fill *af = priv; struct xfs_scrub *sc = af->sc; xfs_agblock_t agbno = start; int error; trace_xrep_agfl_insert(sc->sa.pag, agbno, len); while (agbno < start + len && af->fl_off < af->flcount) af->agfl_bno[af->fl_off++] = cpu_to_be32(agbno++); error = xagb_bitmap_set(&af->used_extents, start, agbno - 1); if (error) return error; if (af->fl_off == af->flcount) return -ECANCELED; return 0; } /* Write out a totally new AGFL. */ STATIC int xrep_agfl_init_header( struct xfs_scrub *sc, struct xfs_buf *agfl_bp, struct xagb_bitmap *agfl_extents, xfs_agblock_t flcount) { struct xrep_agfl_fill af = { .sc = sc, .flcount = flcount, }; struct xfs_mount *mp = sc->mp; struct xfs_agfl *agfl; int error; ASSERT(flcount <= xfs_agfl_size(mp)); /* * Start rewriting the header by setting the bno[] array to * NULLAGBLOCK, then setting AGFL header fields. */ agfl = XFS_BUF_TO_AGFL(agfl_bp); memset(agfl, 0xFF, BBTOB(agfl_bp->b_length)); agfl->agfl_magicnum = cpu_to_be32(XFS_AGFL_MAGIC); agfl->agfl_seqno = cpu_to_be32(sc->sa.pag->pag_agno); uuid_copy(&agfl->agfl_uuid, &mp->m_sb.sb_meta_uuid); /* * Fill the AGFL with the remaining blocks. If agfl_extents has more * blocks than fit in the AGFL, they will be freed in a subsequent * step. */ xagb_bitmap_init(&af.used_extents); af.agfl_bno = xfs_buf_to_agfl_bno(agfl_bp); xagb_bitmap_walk(agfl_extents, xrep_agfl_fill, &af); error = xagb_bitmap_disunion(agfl_extents, &af.used_extents); if (error) return error; /* Write new AGFL to disk. */ xfs_trans_buf_set_type(sc->tp, agfl_bp, XFS_BLFT_AGFL_BUF); xfs_trans_log_buf(sc->tp, agfl_bp, 0, BBTOB(agfl_bp->b_length) - 1); xagb_bitmap_destroy(&af.used_extents); return 0; } /* Repair the AGFL. */ int xrep_agfl( struct xfs_scrub *sc) { struct xagb_bitmap agfl_extents; struct xfs_mount *mp = sc->mp; struct xfs_buf *agf_bp; struct xfs_buf *agfl_bp; xfs_agblock_t flcount; int error; /* We require the rmapbt to rebuild anything. */ if (!xfs_has_rmapbt(mp)) return -EOPNOTSUPP; xagb_bitmap_init(&agfl_extents); /* * Read the AGF so that we can query the rmapbt. We hope that there's * nothing wrong with the AGF, but all the AG header repair functions * have this chicken-and-egg problem. */ error = xfs_alloc_read_agf(sc->sa.pag, sc->tp, 0, &agf_bp); if (error) return error; /* * Make sure we have the AGFL buffer, as scrub might have decided it * was corrupt after xfs_alloc_read_agfl failed with -EFSCORRUPTED. */ error = xfs_trans_read_buf(mp, sc->tp, mp->m_ddev_targp, XFS_AG_DADDR(mp, sc->sa.pag->pag_agno, XFS_AGFL_DADDR(mp)), XFS_FSS_TO_BB(mp, 1), 0, &agfl_bp, NULL); if (error) return error; agfl_bp->b_ops = &xfs_agfl_buf_ops; /* Gather all the extents we're going to put on the new AGFL. */ error = xrep_agfl_collect_blocks(sc, agf_bp, &agfl_extents, &flcount); if (error) goto err; /* Last chance to abort before we start committing fixes. */ if (xchk_should_terminate(sc, &error)) goto err; /* * Update AGF and AGFL. We reset the global free block counter when * we adjust the AGF flcount (which can fail) so avoid updating any * buffers until we know that part works. */ xrep_agfl_update_agf(sc, agf_bp, flcount); error = xrep_agfl_init_header(sc, agfl_bp, &agfl_extents, flcount); if (error) goto err; /* * Ok, the AGFL should be ready to go now. Roll the transaction to * make the new AGFL permanent before we start using it to return * freespace overflow to the freespace btrees. */ sc->sa.agf_bp = agf_bp; error = xrep_roll_ag_trans(sc); if (error) goto err; /* Dump any AGFL overflow. */ error = xrep_reap_agblocks(sc, &agfl_extents, &XFS_RMAP_OINFO_AG, XFS_AG_RESV_AGFL); if (error) goto err; err: xagb_bitmap_destroy(&agfl_extents); return error; } /* AGI */ /* * Offset within the xrep_find_ag_btree array for each btree type. Avoid the * XFS_BTNUM_ names here to avoid creating a sparse array. */ enum { XREP_AGI_INOBT = 0, XREP_AGI_FINOBT, XREP_AGI_END, XREP_AGI_MAX }; #define XREP_AGI_LOOKUP_BATCH 32 struct xrep_agi { struct xfs_scrub *sc; /* AGI buffer, tracked separately */ struct xfs_buf *agi_bp; /* context for finding btree roots */ struct xrep_find_ag_btree fab[XREP_AGI_MAX]; /* old AGI contents in case we have to revert */ struct xfs_agi old_agi; /* bitmap of which inodes are unlinked */ struct xagino_bitmap iunlink_bmp; /* heads of the unlinked inode bucket lists */ xfs_agino_t iunlink_heads[XFS_AGI_UNLINKED_BUCKETS]; /* scratchpad for batched lookups of the radix tree */ struct xfs_inode *lookup_batch[XREP_AGI_LOOKUP_BATCH]; /* Map of ino -> next_ino for unlinked inode processing. */ struct xfarray *iunlink_next; /* Map of ino -> prev_ino for unlinked inode processing. */ struct xfarray *iunlink_prev; }; static void xrep_agi_buf_cleanup( void *buf) { struct xrep_agi *ragi = buf; xfarray_destroy(ragi->iunlink_prev); xfarray_destroy(ragi->iunlink_next); xagino_bitmap_destroy(&ragi->iunlink_bmp); } /* * Given the inode btree roots described by *fab, find the roots, check them * for sanity, and pass the root data back out via *fab. */ STATIC int xrep_agi_find_btrees( struct xrep_agi *ragi) { struct xfs_scrub *sc = ragi->sc; struct xrep_find_ag_btree *fab = ragi->fab; struct xfs_buf *agf_bp; struct xfs_mount *mp = sc->mp; int error; /* Read the AGF. */ error = xfs_alloc_read_agf(sc->sa.pag, sc->tp, 0, &agf_bp); if (error) return error; /* Find the btree roots. */ error = xrep_find_ag_btree_roots(sc, agf_bp, fab, NULL); if (error) return error; /* We must find the inobt root. */ if (!xrep_check_btree_root(sc, &fab[XREP_AGI_INOBT])) return -EFSCORRUPTED; /* We must find the finobt root if that feature is enabled. */ if (xfs_has_finobt(mp) && !xrep_check_btree_root(sc, &fab[XREP_AGI_FINOBT])) return -EFSCORRUPTED; return 0; } /* * Reinitialize the AGI header, making an in-core copy of the old contents so * that we know which in-core state needs to be reinitialized. */ STATIC void xrep_agi_init_header( struct xrep_agi *ragi) { struct xfs_scrub *sc = ragi->sc; struct xfs_buf *agi_bp = ragi->agi_bp; struct xfs_agi *old_agi = &ragi->old_agi; struct xfs_agi *agi = agi_bp->b_addr; struct xfs_perag *pag = sc->sa.pag; struct xfs_mount *mp = sc->mp; memcpy(old_agi, agi, sizeof(*old_agi)); memset(agi, 0, BBTOB(agi_bp->b_length)); agi->agi_magicnum = cpu_to_be32(XFS_AGI_MAGIC); agi->agi_versionnum = cpu_to_be32(XFS_AGI_VERSION); agi->agi_seqno = cpu_to_be32(pag->pag_agno); agi->agi_length = cpu_to_be32(pag->block_count); agi->agi_newino = cpu_to_be32(NULLAGINO); agi->agi_dirino = cpu_to_be32(NULLAGINO); if (xfs_has_crc(mp)) uuid_copy(&agi->agi_uuid, &mp->m_sb.sb_meta_uuid); /* Mark the incore AGF data stale until we're done fixing things. */ ASSERT(xfs_perag_initialised_agi(pag)); clear_bit(XFS_AGSTATE_AGI_INIT, &pag->pag_opstate); } /* Set btree root information in an AGI. */ STATIC void xrep_agi_set_roots( struct xrep_agi *ragi) { struct xfs_scrub *sc = ragi->sc; struct xfs_agi *agi = ragi->agi_bp->b_addr; struct xrep_find_ag_btree *fab = ragi->fab; agi->agi_root = cpu_to_be32(fab[XREP_AGI_INOBT].root); agi->agi_level = cpu_to_be32(fab[XREP_AGI_INOBT].height); if (xfs_has_finobt(sc->mp)) { agi->agi_free_root = cpu_to_be32(fab[XREP_AGI_FINOBT].root); agi->agi_free_level = cpu_to_be32(fab[XREP_AGI_FINOBT].height); } } /* Update the AGI counters. */ STATIC int xrep_agi_calc_from_btrees( struct xrep_agi *ragi) { struct xfs_scrub *sc = ragi->sc; struct xfs_buf *agi_bp = ragi->agi_bp; struct xfs_btree_cur *cur; struct xfs_agi *agi = agi_bp->b_addr; struct xfs_mount *mp = sc->mp; xfs_agino_t count; xfs_agino_t freecount; int error; cur = xfs_inobt_init_cursor(sc->sa.pag, sc->tp, agi_bp); error = xfs_ialloc_count_inodes(cur, &count, &freecount); if (error) goto err; if (xfs_has_inobtcounts(mp)) { xfs_agblock_t blocks; error = xfs_btree_count_blocks(cur, &blocks); if (error) goto err; agi->agi_iblocks = cpu_to_be32(blocks); } xfs_btree_del_cursor(cur, error); agi->agi_count = cpu_to_be32(count); agi->agi_freecount = cpu_to_be32(freecount); if (xfs_has_finobt(mp) && xfs_has_inobtcounts(mp)) { xfs_agblock_t blocks; cur = xfs_finobt_init_cursor(sc->sa.pag, sc->tp, agi_bp); error = xfs_btree_count_blocks(cur, &blocks); if (error) goto err; xfs_btree_del_cursor(cur, error); agi->agi_fblocks = cpu_to_be32(blocks); } return 0; err: xfs_btree_del_cursor(cur, error); return error; } /* * Record a forwards unlinked chain pointer from agino -> next_agino in our * staging information. */ static inline int xrep_iunlink_store_next( struct xrep_agi *ragi, xfs_agino_t agino, xfs_agino_t next_agino) { ASSERT(next_agino != 0); return xfarray_store(ragi->iunlink_next, agino, &next_agino); } /* * Record a backwards unlinked chain pointer from prev_ino <- agino in our * staging information. */ static inline int xrep_iunlink_store_prev( struct xrep_agi *ragi, xfs_agino_t agino, xfs_agino_t prev_agino) { ASSERT(prev_agino != 0); return xfarray_store(ragi->iunlink_prev, agino, &prev_agino); } /* * Given an @agino, look up the next inode in the iunlink bucket. Returns * NULLAGINO if we're at the end of the chain, 0 if @agino is not in memory * like it should be, or a per-AG inode number. */ static inline xfs_agino_t xrep_iunlink_next( struct xfs_scrub *sc, xfs_agino_t agino) { struct xfs_inode *ip; ip = xfs_iunlink_lookup(sc->sa.pag, agino); if (!ip) return 0; return ip->i_next_unlinked; } /* * Load the inode @agino into memory, set its i_prev_unlinked, and drop the * inode so it can be inactivated. Returns NULLAGINO if we're at the end of * the chain or if we should stop walking the chain due to corruption; or a * per-AG inode number. */ STATIC xfs_agino_t xrep_iunlink_reload_next( struct xrep_agi *ragi, xfs_agino_t prev_agino, xfs_agino_t agino) { struct xfs_scrub *sc = ragi->sc; struct xfs_inode *ip; xfs_ino_t ino; xfs_agino_t ret = NULLAGINO; int error; ino = XFS_AGINO_TO_INO(sc->mp, sc->sa.pag->pag_agno, agino); error = xchk_iget(ragi->sc, ino, &ip); if (error) return ret; trace_xrep_iunlink_reload_next(ip, prev_agino); /* If this is a linked inode, stop processing the chain. */ if (VFS_I(ip)->i_nlink != 0) { xrep_iunlink_store_next(ragi, agino, NULLAGINO); goto rele; } ip->i_prev_unlinked = prev_agino; ret = ip->i_next_unlinked; /* * Drop the inode reference that we just took. We hold the AGI, so * this inode cannot move off the unlinked list and hence cannot be * reclaimed. */ rele: xchk_irele(sc, ip); return ret; } /* * Walk an AGI unlinked bucket's list to load incore any unlinked inodes that * still existed at mount time. This can happen if iunlink processing fails * during log recovery. */ STATIC int xrep_iunlink_walk_ondisk_bucket( struct xrep_agi *ragi, unsigned int bucket) { struct xfs_scrub *sc = ragi->sc; struct xfs_agi *agi = sc->sa.agi_bp->b_addr; xfs_agino_t prev_agino = NULLAGINO; xfs_agino_t next_agino; int error = 0; next_agino = be32_to_cpu(agi->agi_unlinked[bucket]); while (next_agino != NULLAGINO) { xfs_agino_t agino = next_agino; if (xchk_should_terminate(ragi->sc, &error)) return error; trace_xrep_iunlink_walk_ondisk_bucket(sc->sa.pag, bucket, prev_agino, agino); if (bucket != agino % XFS_AGI_UNLINKED_BUCKETS) break; next_agino = xrep_iunlink_next(sc, agino); if (!next_agino) next_agino = xrep_iunlink_reload_next(ragi, prev_agino, agino); prev_agino = agino; } return 0; } /* Decide if this is an unlinked inode in this AG. */ STATIC bool xrep_iunlink_igrab( struct xfs_perag *pag, struct xfs_inode *ip) { struct xfs_mount *mp = pag->pag_mount; if (XFS_INO_TO_AGNO(mp, ip->i_ino) != pag->pag_agno) return false; if (!xfs_inode_on_unlinked_list(ip)) return false; return true; } /* * Mark the given inode in the lookup batch in our unlinked inode bitmap, and * remember if this inode is the start of the unlinked chain. */ STATIC int xrep_iunlink_visit( struct xrep_agi *ragi, unsigned int batch_idx) { struct xfs_mount *mp = ragi->sc->mp; struct xfs_inode *ip = ragi->lookup_batch[batch_idx]; xfs_agino_t agino; unsigned int bucket; int error; ASSERT(XFS_INO_TO_AGNO(mp, ip->i_ino) == ragi->sc->sa.pag->pag_agno); ASSERT(xfs_inode_on_unlinked_list(ip)); agino = XFS_INO_TO_AGINO(mp, ip->i_ino); bucket = agino % XFS_AGI_UNLINKED_BUCKETS; trace_xrep_iunlink_visit(ragi->sc->sa.pag, bucket, ragi->iunlink_heads[bucket], ip); error = xagino_bitmap_set(&ragi->iunlink_bmp, agino, 1); if (error) return error; if (ip->i_prev_unlinked == NULLAGINO) { if (ragi->iunlink_heads[bucket] == NULLAGINO) ragi->iunlink_heads[bucket] = agino; } return 0; } /* * Find all incore unlinked inodes so that we can rebuild the unlinked buckets. * We hold the AGI so there should not be any modifications to the unlinked * list. */ STATIC int xrep_iunlink_mark_incore( struct xrep_agi *ragi) { struct xfs_perag *pag = ragi->sc->sa.pag; struct xfs_mount *mp = pag->pag_mount; uint32_t first_index = 0; bool done = false; unsigned int nr_found = 0; do { unsigned int i; int error = 0; if (xchk_should_terminate(ragi->sc, &error)) return error; rcu_read_lock(); nr_found = radix_tree_gang_lookup(&pag->pag_ici_root, (void **)&ragi->lookup_batch, first_index, XREP_AGI_LOOKUP_BATCH); if (!nr_found) { rcu_read_unlock(); return 0; } for (i = 0; i < nr_found; i++) { struct xfs_inode *ip = ragi->lookup_batch[i]; if (done || !xrep_iunlink_igrab(pag, ip)) ragi->lookup_batch[i] = NULL; /* * Update the index for the next lookup. Catch * overflows into the next AG range which can occur if * we have inodes in the last block of the AG and we * are currently pointing to the last inode. * * Because we may see inodes that are from the wrong AG * due to RCU freeing and reallocation, only update the * index if it lies in this AG. It was a race that lead * us to see this inode, so another lookup from the * same index will not find it again. */ if (XFS_INO_TO_AGNO(mp, ip->i_ino) != pag->pag_agno) continue; first_index = XFS_INO_TO_AGINO(mp, ip->i_ino + 1); if (first_index < XFS_INO_TO_AGINO(mp, ip->i_ino)) done = true; } /* unlock now we've grabbed the inodes. */ rcu_read_unlock(); for (i = 0; i < nr_found; i++) { if (!ragi->lookup_batch[i]) continue; error = xrep_iunlink_visit(ragi, i); if (error) return error; } } while (!done); return 0; } /* Mark all the unlinked ondisk inodes in this inobt record in iunlink_bmp. */ STATIC int xrep_iunlink_mark_ondisk_rec( struct xfs_btree_cur *cur, const union xfs_btree_rec *rec, void *priv) { struct xfs_inobt_rec_incore irec; struct xrep_agi *ragi = priv; struct xfs_scrub *sc = ragi->sc; struct xfs_mount *mp = cur->bc_mp; xfs_agino_t agino; unsigned int i; int error = 0; xfs_inobt_btrec_to_irec(mp, rec, &irec); for (i = 0, agino = irec.ir_startino; i < XFS_INODES_PER_CHUNK; i++, agino++) { struct xfs_inode *ip; unsigned int len = 1; /* Skip free inodes */ if (XFS_INOBT_MASK(i) & irec.ir_free) continue; /* Skip inodes we've seen before */ if (xagino_bitmap_test(&ragi->iunlink_bmp, agino, &len)) continue; /* * Skip incore inodes; these were already picked up by * the _mark_incore step. */ rcu_read_lock(); ip = radix_tree_lookup(&sc->sa.pag->pag_ici_root, agino); rcu_read_unlock(); if (ip) continue; /* * Try to look up this inode. If we can't get it, just move * on because we haven't actually scrubbed the inobt or the * inodes yet. */ error = xchk_iget(ragi->sc, XFS_AGINO_TO_INO(mp, sc->sa.pag->pag_agno, agino), &ip); if (error) continue; trace_xrep_iunlink_reload_ondisk(ip); if (VFS_I(ip)->i_nlink == 0) error = xagino_bitmap_set(&ragi->iunlink_bmp, agino, 1); xchk_irele(sc, ip); if (error) break; } return error; } /* * Find ondisk inodes that are unlinked and not in cache, and mark them in * iunlink_bmp. We haven't checked the inobt yet, so we don't error out if * the btree is corrupt. */ STATIC void xrep_iunlink_mark_ondisk( struct xrep_agi *ragi) { struct xfs_scrub *sc = ragi->sc; struct xfs_buf *agi_bp = ragi->agi_bp; struct xfs_btree_cur *cur; int error; cur = xfs_inobt_init_cursor(sc->sa.pag, sc->tp, agi_bp); error = xfs_btree_query_all(cur, xrep_iunlink_mark_ondisk_rec, ragi); xfs_btree_del_cursor(cur, error); } /* * Walk an iunlink bucket's inode list. For each inode that should be on this * chain, clear its entry in in iunlink_bmp because it's ok and we don't need * to touch it further. */ STATIC int xrep_iunlink_resolve_bucket( struct xrep_agi *ragi, unsigned int bucket) { struct xfs_scrub *sc = ragi->sc; struct xfs_inode *ip; xfs_agino_t prev_agino = NULLAGINO; xfs_agino_t next_agino = ragi->iunlink_heads[bucket]; int error = 0; while (next_agino != NULLAGINO) { if (xchk_should_terminate(ragi->sc, &error)) return error; /* Find the next inode in the chain. */ ip = xfs_iunlink_lookup(sc->sa.pag, next_agino); if (!ip) { /* Inode not incore? Terminate the chain. */ trace_xrep_iunlink_resolve_uncached(sc->sa.pag, bucket, prev_agino, next_agino); next_agino = NULLAGINO; break; } if (next_agino % XFS_AGI_UNLINKED_BUCKETS != bucket) { /* * Inode is in the wrong bucket. Advance the list, * but pretend we didn't see this inode. */ trace_xrep_iunlink_resolve_wronglist(sc->sa.pag, bucket, prev_agino, next_agino); next_agino = ip->i_next_unlinked; continue; } if (!xfs_inode_on_unlinked_list(ip)) { /* * Incore inode doesn't think this inode is on an * unlinked list. This is probably because we reloaded * it from disk. Advance the list, but pretend we * didn't see this inode; we'll fix that later. */ trace_xrep_iunlink_resolve_nolist(sc->sa.pag, bucket, prev_agino, next_agino); next_agino = ip->i_next_unlinked; continue; } trace_xrep_iunlink_resolve_ok(sc->sa.pag, bucket, prev_agino, next_agino); /* * Otherwise, this inode's unlinked pointers are ok. Clear it * from the unlinked bitmap since we're done with it, and make * sure the chain is still correct. */ error = xagino_bitmap_clear(&ragi->iunlink_bmp, next_agino, 1); if (error) return error; /* Remember the previous inode's next pointer. */ if (prev_agino != NULLAGINO) { error = xrep_iunlink_store_next(ragi, prev_agino, next_agino); if (error) return error; } /* Remember this inode's previous pointer. */ error = xrep_iunlink_store_prev(ragi, next_agino, prev_agino); if (error) return error; /* Advance the list and remember this inode. */ prev_agino = next_agino; next_agino = ip->i_next_unlinked; } /* Update the previous inode's next pointer. */ if (prev_agino != NULLAGINO) { error = xrep_iunlink_store_next(ragi, prev_agino, next_agino); if (error) return error; } return 0; } /* Reinsert this unlinked inode into the head of the staged bucket list. */ STATIC int xrep_iunlink_add_to_bucket( struct xrep_agi *ragi, xfs_agino_t agino) { xfs_agino_t current_head; unsigned int bucket; int error; bucket = agino % XFS_AGI_UNLINKED_BUCKETS; /* Point this inode at the current head of the bucket list. */ current_head = ragi->iunlink_heads[bucket]; trace_xrep_iunlink_add_to_bucket(ragi->sc->sa.pag, bucket, agino, current_head); error = xrep_iunlink_store_next(ragi, agino, current_head); if (error) return error; /* Remember the head inode's previous pointer. */ if (current_head != NULLAGINO) { error = xrep_iunlink_store_prev(ragi, current_head, agino); if (error) return error; } ragi->iunlink_heads[bucket] = agino; return 0; } /* Reinsert unlinked inodes into the staged iunlink buckets. */ STATIC int xrep_iunlink_add_lost_inodes( uint32_t start, uint32_t len, void *priv) { struct xrep_agi *ragi = priv; int error; for (; len > 0; start++, len--) { error = xrep_iunlink_add_to_bucket(ragi, start); if (error) return error; } return 0; } /* * Figure out the iunlink bucket values and find inodes that need to be * reinserted into the list. */ STATIC int xrep_iunlink_rebuild_buckets( struct xrep_agi *ragi) { unsigned int i; int error; /* * Walk the ondisk AGI unlinked list to find inodes that are on the * list but aren't in memory. This can happen if a past log recovery * tried to clear the iunlinked list but failed. Our scan rebuilds the * unlinked list using incore inodes, so we must load and link them * properly. */ for (i = 0; i < XFS_AGI_UNLINKED_BUCKETS; i++) { error = xrep_iunlink_walk_ondisk_bucket(ragi, i); if (error) return error; } /* * Record all the incore unlinked inodes in iunlink_bmp that we didn't * find by walking the ondisk iunlink buckets. This shouldn't happen, * but we can't risk forgetting an inode somewhere. */ error = xrep_iunlink_mark_incore(ragi); if (error) return error; /* * If there are ondisk inodes that are unlinked and are not been loaded * into cache, record them in iunlink_bmp. */ xrep_iunlink_mark_ondisk(ragi); /* * Walk each iunlink bucket to (re)construct as much of the incore list * as would be correct. For each inode that survives this step, mark * it clear in iunlink_bmp; we're done with those inodes. */ for (i = 0; i < XFS_AGI_UNLINKED_BUCKETS; i++) { error = xrep_iunlink_resolve_bucket(ragi, i); if (error) return error; } /* * Any unlinked inodes that we didn't find through the bucket list * walk (or was ignored by the walk) must be inserted into the bucket * list. Stage this in memory for now. */ return xagino_bitmap_walk(&ragi->iunlink_bmp, xrep_iunlink_add_lost_inodes, ragi); } /* Update i_next_iunlinked for the inode @agino. */ STATIC int xrep_iunlink_relink_next( struct xrep_agi *ragi, xfarray_idx_t idx, xfs_agino_t next_agino) { struct xfs_scrub *sc = ragi->sc; struct xfs_perag *pag = sc->sa.pag; struct xfs_inode *ip; xfarray_idx_t agino = idx - 1; bool want_rele = false; int error = 0; ip = xfs_iunlink_lookup(pag, agino); if (!ip) { xfs_ino_t ino; xfs_agino_t prev_agino; /* * No inode exists in cache. Load it off the disk so that we * can reinsert it into the incore unlinked list. */ ino = XFS_AGINO_TO_INO(sc->mp, pag->pag_agno, agino); error = xchk_iget(sc, ino, &ip); if (error) return -EFSCORRUPTED; want_rele = true; /* Set the backward pointer since this just came off disk. */ error = xfarray_load(ragi->iunlink_prev, agino, &prev_agino); if (error) goto out_rele; trace_xrep_iunlink_relink_prev(ip, prev_agino); ip->i_prev_unlinked = prev_agino; } /* Update the forward pointer. */ if (ip->i_next_unlinked != next_agino) { error = xfs_iunlink_log_inode(sc->tp, ip, pag, next_agino); if (error) goto out_rele; trace_xrep_iunlink_relink_next(ip, next_agino); ip->i_next_unlinked = next_agino; } out_rele: /* * The iunlink lookup doesn't igrab because we hold the AGI buffer lock * and the inode cannot be reclaimed. However, if we used iget to load * a missing inode, we must irele it here. */ if (want_rele) xchk_irele(sc, ip); return error; } /* Update i_prev_iunlinked for the inode @agino. */ STATIC int xrep_iunlink_relink_prev( struct xrep_agi *ragi, xfarray_idx_t idx, xfs_agino_t prev_agino) { struct xfs_scrub *sc = ragi->sc; struct xfs_perag *pag = sc->sa.pag; struct xfs_inode *ip; xfarray_idx_t agino = idx - 1; bool want_rele = false; int error = 0; ASSERT(prev_agino != 0); ip = xfs_iunlink_lookup(pag, agino); if (!ip) { xfs_ino_t ino; xfs_agino_t next_agino; /* * No inode exists in cache. Load it off the disk so that we * can reinsert it into the incore unlinked list. */ ino = XFS_AGINO_TO_INO(sc->mp, pag->pag_agno, agino); error = xchk_iget(sc, ino, &ip); if (error) return -EFSCORRUPTED; want_rele = true; /* Set the forward pointer since this just came off disk. */ error = xfarray_load(ragi->iunlink_prev, agino, &next_agino); if (error) goto out_rele; error = xfs_iunlink_log_inode(sc->tp, ip, pag, next_agino); if (error) goto out_rele; trace_xrep_iunlink_relink_next(ip, next_agino); ip->i_next_unlinked = next_agino; } /* Update the backward pointer. */ if (ip->i_prev_unlinked != prev_agino) { trace_xrep_iunlink_relink_prev(ip, prev_agino); ip->i_prev_unlinked = prev_agino; } out_rele: /* * The iunlink lookup doesn't igrab because we hold the AGI buffer lock * and the inode cannot be reclaimed. However, if we used iget to load * a missing inode, we must irele it here. */ if (want_rele) xchk_irele(sc, ip); return error; } /* Log all the iunlink updates we need to finish regenerating the AGI. */ STATIC int xrep_iunlink_commit( struct xrep_agi *ragi) { struct xfs_agi *agi = ragi->agi_bp->b_addr; xfarray_idx_t idx = XFARRAY_CURSOR_INIT; xfs_agino_t agino; unsigned int i; int error; /* Fix all the forward links */ while ((error = xfarray_iter(ragi->iunlink_next, &idx, &agino)) == 1) { error = xrep_iunlink_relink_next(ragi, idx, agino); if (error) return error; } /* Fix all the back links */ idx = XFARRAY_CURSOR_INIT; while ((error = xfarray_iter(ragi->iunlink_prev, &idx, &agino)) == 1) { error = xrep_iunlink_relink_prev(ragi, idx, agino); if (error) return error; } /* Copy the staged iunlink buckets to the new AGI. */ for (i = 0; i < XFS_AGI_UNLINKED_BUCKETS; i++) { trace_xrep_iunlink_commit_bucket(ragi->sc->sa.pag, i, be32_to_cpu(ragi->old_agi.agi_unlinked[i]), ragi->iunlink_heads[i]); agi->agi_unlinked[i] = cpu_to_be32(ragi->iunlink_heads[i]); } return 0; } /* Trigger reinitialization of the in-core data. */ STATIC int xrep_agi_commit_new( struct xrep_agi *ragi) { struct xfs_scrub *sc = ragi->sc; struct xfs_buf *agi_bp = ragi->agi_bp; struct xfs_perag *pag; struct xfs_agi *agi = agi_bp->b_addr; /* Trigger inode count recalculation */ xfs_force_summary_recalc(sc->mp); /* Write this to disk. */ xfs_trans_buf_set_type(sc->tp, agi_bp, XFS_BLFT_AGI_BUF); xfs_trans_log_buf(sc->tp, agi_bp, 0, BBTOB(agi_bp->b_length) - 1); /* Now reinitialize the in-core counters if necessary. */ pag = sc->sa.pag; pag->pagi_count = be32_to_cpu(agi->agi_count); pag->pagi_freecount = be32_to_cpu(agi->agi_freecount); set_bit(XFS_AGSTATE_AGI_INIT, &pag->pag_opstate); return xrep_roll_ag_trans(sc); } /* Repair the AGI. */ int xrep_agi( struct xfs_scrub *sc) { struct xrep_agi *ragi; struct xfs_mount *mp = sc->mp; char *descr; unsigned int i; int error; /* We require the rmapbt to rebuild anything. */ if (!xfs_has_rmapbt(mp)) return -EOPNOTSUPP; sc->buf = kzalloc(sizeof(struct xrep_agi), XCHK_GFP_FLAGS); if (!sc->buf) return -ENOMEM; ragi = sc->buf; ragi->sc = sc; ragi->fab[XREP_AGI_INOBT] = (struct xrep_find_ag_btree){ .rmap_owner = XFS_RMAP_OWN_INOBT, .buf_ops = &xfs_inobt_buf_ops, .maxlevels = M_IGEO(sc->mp)->inobt_maxlevels, }; ragi->fab[XREP_AGI_FINOBT] = (struct xrep_find_ag_btree){ .rmap_owner = XFS_RMAP_OWN_INOBT, .buf_ops = &xfs_finobt_buf_ops, .maxlevels = M_IGEO(sc->mp)->inobt_maxlevels, }; ragi->fab[XREP_AGI_END] = (struct xrep_find_ag_btree){ .buf_ops = NULL, }; for (i = 0; i < XFS_AGI_UNLINKED_BUCKETS; i++) ragi->iunlink_heads[i] = NULLAGINO; xagino_bitmap_init(&ragi->iunlink_bmp); sc->buf_cleanup = xrep_agi_buf_cleanup; descr = xchk_xfile_ag_descr(sc, "iunlinked next pointers"); error = xfarray_create(descr, 0, sizeof(xfs_agino_t), &ragi->iunlink_next); kfree(descr); if (error) return error; descr = xchk_xfile_ag_descr(sc, "iunlinked prev pointers"); error = xfarray_create(descr, 0, sizeof(xfs_agino_t), &ragi->iunlink_prev); kfree(descr); if (error) return error; /* * Make sure we have the AGI buffer, as scrub might have decided it * was corrupt after xfs_ialloc_read_agi failed with -EFSCORRUPTED. */ error = xfs_trans_read_buf(mp, sc->tp, mp->m_ddev_targp, XFS_AG_DADDR(mp, sc->sa.pag->pag_agno, XFS_AGI_DADDR(mp)), XFS_FSS_TO_BB(mp, 1), 0, &ragi->agi_bp, NULL); if (error) return error; ragi->agi_bp->b_ops = &xfs_agi_buf_ops; /* Find the AGI btree roots. */ error = xrep_agi_find_btrees(ragi); if (error) return error; error = xrep_iunlink_rebuild_buckets(ragi); if (error) return error; /* Last chance to abort before we start committing fixes. */ if (xchk_should_terminate(sc, &error)) return error; /* Start rewriting the header and implant the btrees we found. */ xrep_agi_init_header(ragi); xrep_agi_set_roots(ragi); error = xrep_agi_calc_from_btrees(ragi); if (error) goto out_revert; error = xrep_iunlink_commit(ragi); if (error) goto out_revert; /* Reinitialize in-core state. */ return xrep_agi_commit_new(ragi); out_revert: /* Mark the incore AGI state stale and revert the AGI. */ clear_bit(XFS_AGSTATE_AGI_INIT, &sc->sa.pag->pag_opstate); memcpy(ragi->agi_bp->b_addr, &ragi->old_agi, sizeof(struct xfs_agi)); return error; }
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.
Created with Cregit http://github.com/cregit/cregit
Version 2.0-RC1