Contributors: 1
Author Tokens Token Proportion Commits Commit Proportion
Darrick J. Wong 269 100.00% 7 100.00%
Total 269 7


/* SPDX-License-Identifier: GPL-2.0-or-later */
/*
 * Copyright (c) 2021-2024 Oracle.  All Rights Reserved.
 * Author: Darrick J. Wong <djwong@kernel.org>
 */
#ifndef __XFS_SCRUB_ISCAN_H__
#define __XFS_SCRUB_ISCAN_H__

struct xchk_iscan {
	struct xfs_scrub	*sc;

	/* Lock to protect the scan cursor. */
	struct mutex		lock;

	/*
	 * This is the first inode in the inumber address space that we
	 * examined.  When the scan wraps around back to here, the scan is
	 * finished.
	 */
	xfs_ino_t		scan_start_ino;

	/* This is the inode that will be examined next. */
	xfs_ino_t		cursor_ino;

	/* If nonzero and non-NULL, skip this inode when scanning. */
	xfs_ino_t		skip_ino;

	/*
	 * This is the last inode that we've successfully scanned, either
	 * because the caller scanned it, or we moved the cursor past an empty
	 * part of the inode address space.  Scan callers should only use the
	 * xchk_iscan_visit function to modify this.
	 */
	xfs_ino_t		__visited_ino;

	/* Operational state of the livescan. */
	unsigned long		__opstate;

	/* Give up on iterating @cursor_ino if we can't iget it by this time. */
	unsigned long		__iget_deadline;

	/* Amount of time (in ms) that we will try to iget an inode. */
	unsigned int		iget_timeout;

	/* Wait this many ms to retry an iget. */
	unsigned int		iget_retry_delay;

	/*
	 * The scan grabs batches of inodes and stashes them here before
	 * handing them out with _iter.  Unallocated inodes are set in the
	 * mask so that all updates to that inode are selected for live
	 * update propagation.
	 */
	xfs_ino_t		__batch_ino;
	xfs_inofree_t		__skipped_inomask;
	struct xfs_inode	*__inodes[XFS_INODES_PER_CHUNK];
};

/* Set if the scan has been aborted due to some event in the fs. */
#define XCHK_ISCAN_OPSTATE_ABORTED	(1)

/* Use trylock to acquire the AGI */
#define XCHK_ISCAN_OPSTATE_TRYLOCK_AGI	(2)

static inline bool
xchk_iscan_aborted(const struct xchk_iscan *iscan)
{
	return test_bit(XCHK_ISCAN_OPSTATE_ABORTED, &iscan->__opstate);
}

static inline void
xchk_iscan_abort(struct xchk_iscan *iscan)
{
	set_bit(XCHK_ISCAN_OPSTATE_ABORTED, &iscan->__opstate);
}

static inline bool
xchk_iscan_agi_needs_trylock(const struct xchk_iscan *iscan)
{
	return test_bit(XCHK_ISCAN_OPSTATE_TRYLOCK_AGI, &iscan->__opstate);
}

static inline void
xchk_iscan_set_agi_trylock(struct xchk_iscan *iscan)
{
	set_bit(XCHK_ISCAN_OPSTATE_TRYLOCK_AGI, &iscan->__opstate);
}

void xchk_iscan_start(struct xfs_scrub *sc, unsigned int iget_timeout,
		unsigned int iget_retry_delay, struct xchk_iscan *iscan);
void xchk_iscan_finish_early(struct xchk_iscan *iscan);
void xchk_iscan_teardown(struct xchk_iscan *iscan);

int xchk_iscan_iter(struct xchk_iscan *iscan, struct xfs_inode **ipp);
void xchk_iscan_iter_finish(struct xchk_iscan *iscan);

void xchk_iscan_mark_visited(struct xchk_iscan *iscan, struct xfs_inode *ip);
bool xchk_iscan_want_live_update(struct xchk_iscan *iscan, xfs_ino_t ino);

#endif /* __XFS_SCRUB_ISCAN_H__ */