Author | Tokens | Token Proportion | Commits | Commit Proportion |
---|---|---|---|---|
Nikos Tsironis | 3883 | 99.74% | 5 | 83.33% |
Mikulas Patocka | 10 | 0.26% | 1 | 16.67% |
Total | 3893 | 6 |
// SPDX-License-Identifier: GPL-2.0-only /* * Copyright (C) 2019 Arrikto, Inc. All Rights Reserved. */ #include <linux/mm.h> #include <linux/err.h> #include <linux/slab.h> #include <linux/rwsem.h> #include <linux/bitops.h> #include <linux/bitmap.h> #include <linux/device-mapper.h> #include "persistent-data/dm-bitset.h" #include "persistent-data/dm-space-map.h" #include "persistent-data/dm-block-manager.h" #include "persistent-data/dm-transaction-manager.h" #include "dm-clone-metadata.h" #define DM_MSG_PREFIX "clone metadata" #define SUPERBLOCK_LOCATION 0 #define SUPERBLOCK_MAGIC 0x8af27f64 #define SUPERBLOCK_CSUM_XOR 257649492 #define DM_CLONE_MAX_CONCURRENT_LOCKS 5 #define UUID_LEN 16 /* Min and max dm-clone metadata versions supported */ #define DM_CLONE_MIN_METADATA_VERSION 1 #define DM_CLONE_MAX_METADATA_VERSION 1 /* * On-disk metadata layout */ struct superblock_disk { __le32 csum; __le32 flags; __le64 blocknr; __u8 uuid[UUID_LEN]; __le64 magic; __le32 version; __u8 metadata_space_map_root[SPACE_MAP_ROOT_SIZE]; __le64 region_size; __le64 target_size; __le64 bitset_root; } __packed; /* * Region and Dirty bitmaps. * * dm-clone logically splits the source and destination devices in regions of * fixed size. The destination device's regions are gradually hydrated, i.e., * we copy (clone) the source's regions to the destination device. Eventually, * all regions will get hydrated and all I/O will be served from the * destination device. * * We maintain an on-disk bitmap which tracks the state of each of the * destination device's regions, i.e., whether they are hydrated or not. * * To save constantly doing look ups on disk we keep an in core copy of the * on-disk bitmap, the region_map. * * In order to track which regions are hydrated during a metadata transaction, * we use a second set of bitmaps, the dmap (dirty bitmap), which includes two * bitmaps, namely dirty_regions and dirty_words. The dirty_regions bitmap * tracks the regions that got hydrated during the current metadata * transaction. The dirty_words bitmap tracks the dirty words, i.e. longs, of * the dirty_regions bitmap. * * This allows us to precisely track the regions that were hydrated during the * current metadata transaction and update the metadata accordingly, when we * commit the current transaction. This is important because dm-clone should * only commit the metadata of regions that were properly flushed to the * destination device beforehand. Otherwise, in case of a crash, we could end * up with a corrupted dm-clone device. * * When a region finishes hydrating dm-clone calls * dm_clone_set_region_hydrated(), or for discard requests * dm_clone_cond_set_range(), which sets the corresponding bits in region_map * and dmap. * * During a metadata commit we scan dmap->dirty_words and dmap->dirty_regions * and update the on-disk metadata accordingly. Thus, we don't have to flush to * disk the whole region_map. We can just flush the dirty region_map bits. * * We use the helper dmap->dirty_words bitmap, which is smaller than the * original region_map, to reduce the amount of memory accesses during a * metadata commit. Moreover, as dm-bitset also accesses the on-disk bitmap in * 64-bit word granularity, the dirty_words bitmap helps us avoid useless disk * accesses. * * We could update directly the on-disk bitmap, when dm-clone calls either * dm_clone_set_region_hydrated() or dm_clone_cond_set_range(), buts this * inserts significant metadata I/O overhead in dm-clone's I/O path. Also, as * these two functions don't block, we can call them in interrupt context, * e.g., in a hooked overwrite bio's completion routine, and further reduce the * I/O completion latency. * * We maintain two dirty bitmap sets. During a metadata commit we atomically * swap the currently used dmap with the unused one. This allows the metadata * update functions to run concurrently with an ongoing commit. */ struct dirty_map { unsigned long *dirty_words; unsigned long *dirty_regions; unsigned int changed; }; struct dm_clone_metadata { /* The metadata block device */ struct block_device *bdev; sector_t target_size; sector_t region_size; unsigned long nr_regions; unsigned long nr_words; /* Spinlock protecting the region and dirty bitmaps. */ spinlock_t bitmap_lock; struct dirty_map dmap[2]; struct dirty_map *current_dmap; /* Protected by lock */ struct dirty_map *committing_dmap; /* * In core copy of the on-disk bitmap to save constantly doing look ups * on disk. */ unsigned long *region_map; /* Protected by bitmap_lock */ unsigned int read_only; struct dm_block_manager *bm; struct dm_space_map *sm; struct dm_transaction_manager *tm; struct rw_semaphore lock; struct dm_disk_bitset bitset_info; dm_block_t bitset_root; /* * Reading the space map root can fail, so we read it into this * buffer before the superblock is locked and updated. */ __u8 metadata_space_map_root[SPACE_MAP_ROOT_SIZE]; bool hydration_done:1; bool fail_io:1; }; /*---------------------------------------------------------------------------*/ /* * Superblock validation. */ static void sb_prepare_for_write(struct dm_block_validator *v, struct dm_block *b, size_t sb_block_size) { struct superblock_disk *sb; u32 csum; sb = dm_block_data(b); sb->blocknr = cpu_to_le64(dm_block_location(b)); csum = dm_bm_checksum(&sb->flags, sb_block_size - sizeof(__le32), SUPERBLOCK_CSUM_XOR); sb->csum = cpu_to_le32(csum); } static int sb_check(struct dm_block_validator *v, struct dm_block *b, size_t sb_block_size) { struct superblock_disk *sb; u32 csum, metadata_version; sb = dm_block_data(b); if (dm_block_location(b) != le64_to_cpu(sb->blocknr)) { DMERR("Superblock check failed: blocknr %llu, expected %llu", le64_to_cpu(sb->blocknr), (unsigned long long)dm_block_location(b)); return -ENOTBLK; } if (le64_to_cpu(sb->magic) != SUPERBLOCK_MAGIC) { DMERR("Superblock check failed: magic %llu, expected %llu", le64_to_cpu(sb->magic), (unsigned long long)SUPERBLOCK_MAGIC); return -EILSEQ; } csum = dm_bm_checksum(&sb->flags, sb_block_size - sizeof(__le32), SUPERBLOCK_CSUM_XOR); if (sb->csum != cpu_to_le32(csum)) { DMERR("Superblock check failed: checksum %u, expected %u", csum, le32_to_cpu(sb->csum)); return -EILSEQ; } /* Check metadata version */ metadata_version = le32_to_cpu(sb->version); if (metadata_version < DM_CLONE_MIN_METADATA_VERSION || metadata_version > DM_CLONE_MAX_METADATA_VERSION) { DMERR("Clone metadata version %u found, but only versions between %u and %u supported.", metadata_version, DM_CLONE_MIN_METADATA_VERSION, DM_CLONE_MAX_METADATA_VERSION); return -EINVAL; } return 0; } static struct dm_block_validator sb_validator = { .name = "superblock", .prepare_for_write = sb_prepare_for_write, .check = sb_check }; /* * Check if the superblock is formatted or not. We consider the superblock to * be formatted in case we find non-zero bytes in it. */ static int __superblock_all_zeroes(struct dm_block_manager *bm, bool *formatted) { int r; unsigned int i, nr_words; struct dm_block *sblock; __le64 *data_le, zero = cpu_to_le64(0); /* * We don't use a validator here because the superblock could be all * zeroes. */ r = dm_bm_read_lock(bm, SUPERBLOCK_LOCATION, NULL, &sblock); if (r) { DMERR("Failed to read_lock superblock"); return r; } data_le = dm_block_data(sblock); *formatted = false; /* This assumes that the block size is a multiple of 8 bytes */ BUG_ON(dm_bm_block_size(bm) % sizeof(__le64)); nr_words = dm_bm_block_size(bm) / sizeof(__le64); for (i = 0; i < nr_words; i++) { if (data_le[i] != zero) { *formatted = true; break; } } dm_bm_unlock(sblock); return 0; } /*---------------------------------------------------------------------------*/ /* * Low-level metadata handling. */ static inline int superblock_read_lock(struct dm_clone_metadata *cmd, struct dm_block **sblock) { return dm_bm_read_lock(cmd->bm, SUPERBLOCK_LOCATION, &sb_validator, sblock); } static inline int superblock_write_lock_zero(struct dm_clone_metadata *cmd, struct dm_block **sblock) { return dm_bm_write_lock_zero(cmd->bm, SUPERBLOCK_LOCATION, &sb_validator, sblock); } static int __copy_sm_root(struct dm_clone_metadata *cmd) { int r; size_t root_size; r = dm_sm_root_size(cmd->sm, &root_size); if (r) return r; return dm_sm_copy_root(cmd->sm, &cmd->metadata_space_map_root, root_size); } /* Save dm-clone metadata in superblock */ static void __prepare_superblock(struct dm_clone_metadata *cmd, struct superblock_disk *sb) { sb->flags = cpu_to_le32(0UL); /* FIXME: UUID is currently unused */ memset(sb->uuid, 0, sizeof(sb->uuid)); sb->magic = cpu_to_le64(SUPERBLOCK_MAGIC); sb->version = cpu_to_le32(DM_CLONE_MAX_METADATA_VERSION); /* Save the metadata space_map root */ memcpy(&sb->metadata_space_map_root, &cmd->metadata_space_map_root, sizeof(cmd->metadata_space_map_root)); sb->region_size = cpu_to_le64(cmd->region_size); sb->target_size = cpu_to_le64(cmd->target_size); sb->bitset_root = cpu_to_le64(cmd->bitset_root); } static int __open_metadata(struct dm_clone_metadata *cmd) { int r; struct dm_block *sblock; struct superblock_disk *sb; r = superblock_read_lock(cmd, &sblock); if (r) { DMERR("Failed to read_lock superblock"); return r; } sb = dm_block_data(sblock); /* Verify that target_size and region_size haven't changed. */ if (cmd->region_size != le64_to_cpu(sb->region_size) || cmd->target_size != le64_to_cpu(sb->target_size)) { DMERR("Region and/or target size don't match the ones in metadata"); r = -EINVAL; goto out_with_lock; } r = dm_tm_open_with_sm(cmd->bm, SUPERBLOCK_LOCATION, sb->metadata_space_map_root, sizeof(sb->metadata_space_map_root), &cmd->tm, &cmd->sm); if (r) { DMERR("dm_tm_open_with_sm failed"); goto out_with_lock; } dm_disk_bitset_init(cmd->tm, &cmd->bitset_info); cmd->bitset_root = le64_to_cpu(sb->bitset_root); out_with_lock: dm_bm_unlock(sblock); return r; } static int __format_metadata(struct dm_clone_metadata *cmd) { int r; struct dm_block *sblock; struct superblock_disk *sb; r = dm_tm_create_with_sm(cmd->bm, SUPERBLOCK_LOCATION, &cmd->tm, &cmd->sm); if (r) { DMERR("Failed to create transaction manager"); return r; } dm_disk_bitset_init(cmd->tm, &cmd->bitset_info); r = dm_bitset_empty(&cmd->bitset_info, &cmd->bitset_root); if (r) { DMERR("Failed to create empty on-disk bitset"); goto err_with_tm; } r = dm_bitset_resize(&cmd->bitset_info, cmd->bitset_root, 0, cmd->nr_regions, false, &cmd->bitset_root); if (r) { DMERR("Failed to resize on-disk bitset to %lu entries", cmd->nr_regions); goto err_with_tm; } /* Flush to disk all blocks, except the superblock */ r = dm_tm_pre_commit(cmd->tm); if (r) { DMERR("dm_tm_pre_commit failed"); goto err_with_tm; } r = __copy_sm_root(cmd); if (r) { DMERR("__copy_sm_root failed"); goto err_with_tm; } r = superblock_write_lock_zero(cmd, &sblock); if (r) { DMERR("Failed to write_lock superblock"); goto err_with_tm; } sb = dm_block_data(sblock); __prepare_superblock(cmd, sb); r = dm_tm_commit(cmd->tm, sblock); if (r) { DMERR("Failed to commit superblock"); goto err_with_tm; } return 0; err_with_tm: dm_sm_destroy(cmd->sm); dm_tm_destroy(cmd->tm); return r; } static int __open_or_format_metadata(struct dm_clone_metadata *cmd, bool may_format_device) { int r; bool formatted = false; r = __superblock_all_zeroes(cmd->bm, &formatted); if (r) return r; if (!formatted) return may_format_device ? __format_metadata(cmd) : -EPERM; return __open_metadata(cmd); } static int __create_persistent_data_structures(struct dm_clone_metadata *cmd, bool may_format_device) { int r; /* Create block manager */ cmd->bm = dm_block_manager_create(cmd->bdev, DM_CLONE_METADATA_BLOCK_SIZE << SECTOR_SHIFT, DM_CLONE_MAX_CONCURRENT_LOCKS); if (IS_ERR(cmd->bm)) { DMERR("Failed to create block manager"); return PTR_ERR(cmd->bm); } r = __open_or_format_metadata(cmd, may_format_device); if (r) dm_block_manager_destroy(cmd->bm); return r; } static void __destroy_persistent_data_structures(struct dm_clone_metadata *cmd) { dm_sm_destroy(cmd->sm); dm_tm_destroy(cmd->tm); dm_block_manager_destroy(cmd->bm); } /*---------------------------------------------------------------------------*/ static int __dirty_map_init(struct dirty_map *dmap, unsigned long nr_words, unsigned long nr_regions) { dmap->changed = 0; dmap->dirty_words = kvzalloc(bitmap_size(nr_words), GFP_KERNEL); if (!dmap->dirty_words) return -ENOMEM; dmap->dirty_regions = kvzalloc(bitmap_size(nr_regions), GFP_KERNEL); if (!dmap->dirty_regions) { kvfree(dmap->dirty_words); return -ENOMEM; } return 0; } static void __dirty_map_exit(struct dirty_map *dmap) { kvfree(dmap->dirty_words); kvfree(dmap->dirty_regions); } static int dirty_map_init(struct dm_clone_metadata *cmd) { if (__dirty_map_init(&cmd->dmap[0], cmd->nr_words, cmd->nr_regions)) { DMERR("Failed to allocate dirty bitmap"); return -ENOMEM; } if (__dirty_map_init(&cmd->dmap[1], cmd->nr_words, cmd->nr_regions)) { DMERR("Failed to allocate dirty bitmap"); __dirty_map_exit(&cmd->dmap[0]); return -ENOMEM; } cmd->current_dmap = &cmd->dmap[0]; cmd->committing_dmap = NULL; return 0; } static void dirty_map_exit(struct dm_clone_metadata *cmd) { __dirty_map_exit(&cmd->dmap[0]); __dirty_map_exit(&cmd->dmap[1]); } static int __load_bitset_in_core(struct dm_clone_metadata *cmd) { int r; unsigned long i; struct dm_bitset_cursor c; /* Flush bitset cache */ r = dm_bitset_flush(&cmd->bitset_info, cmd->bitset_root, &cmd->bitset_root); if (r) return r; r = dm_bitset_cursor_begin(&cmd->bitset_info, cmd->bitset_root, cmd->nr_regions, &c); if (r) return r; for (i = 0; ; i++) { if (dm_bitset_cursor_get_value(&c)) __set_bit(i, cmd->region_map); else __clear_bit(i, cmd->region_map); if (i >= (cmd->nr_regions - 1)) break; r = dm_bitset_cursor_next(&c); if (r) break; } dm_bitset_cursor_end(&c); return r; } struct dm_clone_metadata *dm_clone_metadata_open(struct block_device *bdev, sector_t target_size, sector_t region_size) { int r; struct dm_clone_metadata *cmd; cmd = kzalloc(sizeof(*cmd), GFP_KERNEL); if (!cmd) { DMERR("Failed to allocate memory for dm-clone metadata"); return ERR_PTR(-ENOMEM); } cmd->bdev = bdev; cmd->target_size = target_size; cmd->region_size = region_size; cmd->nr_regions = dm_sector_div_up(cmd->target_size, cmd->region_size); cmd->nr_words = BITS_TO_LONGS(cmd->nr_regions); init_rwsem(&cmd->lock); spin_lock_init(&cmd->bitmap_lock); cmd->read_only = 0; cmd->fail_io = false; cmd->hydration_done = false; cmd->region_map = kvmalloc(bitmap_size(cmd->nr_regions), GFP_KERNEL); if (!cmd->region_map) { DMERR("Failed to allocate memory for region bitmap"); r = -ENOMEM; goto out_with_md; } r = __create_persistent_data_structures(cmd, true); if (r) goto out_with_region_map; r = __load_bitset_in_core(cmd); if (r) { DMERR("Failed to load on-disk region map"); goto out_with_pds; } r = dirty_map_init(cmd); if (r) goto out_with_pds; if (bitmap_full(cmd->region_map, cmd->nr_regions)) cmd->hydration_done = true; return cmd; out_with_pds: __destroy_persistent_data_structures(cmd); out_with_region_map: kvfree(cmd->region_map); out_with_md: kfree(cmd); return ERR_PTR(r); } void dm_clone_metadata_close(struct dm_clone_metadata *cmd) { if (!cmd->fail_io) __destroy_persistent_data_structures(cmd); dirty_map_exit(cmd); kvfree(cmd->region_map); kfree(cmd); } bool dm_clone_is_hydration_done(struct dm_clone_metadata *cmd) { return cmd->hydration_done; } bool dm_clone_is_region_hydrated(struct dm_clone_metadata *cmd, unsigned long region_nr) { return dm_clone_is_hydration_done(cmd) || test_bit(region_nr, cmd->region_map); } bool dm_clone_is_range_hydrated(struct dm_clone_metadata *cmd, unsigned long start, unsigned long nr_regions) { unsigned long bit; if (dm_clone_is_hydration_done(cmd)) return true; bit = find_next_zero_bit(cmd->region_map, cmd->nr_regions, start); return (bit >= (start + nr_regions)); } unsigned int dm_clone_nr_of_hydrated_regions(struct dm_clone_metadata *cmd) { return bitmap_weight(cmd->region_map, cmd->nr_regions); } unsigned long dm_clone_find_next_unhydrated_region(struct dm_clone_metadata *cmd, unsigned long start) { return find_next_zero_bit(cmd->region_map, cmd->nr_regions, start); } static int __update_metadata_word(struct dm_clone_metadata *cmd, unsigned long *dirty_regions, unsigned long word) { int r; unsigned long index = word * BITS_PER_LONG; unsigned long max_index = min(cmd->nr_regions, (word + 1) * BITS_PER_LONG); while (index < max_index) { if (test_bit(index, dirty_regions)) { r = dm_bitset_set_bit(&cmd->bitset_info, cmd->bitset_root, index, &cmd->bitset_root); if (r) { DMERR("dm_bitset_set_bit failed"); return r; } __clear_bit(index, dirty_regions); } index++; } return 0; } static int __metadata_commit(struct dm_clone_metadata *cmd) { int r; struct dm_block *sblock; struct superblock_disk *sb; /* Flush bitset cache */ r = dm_bitset_flush(&cmd->bitset_info, cmd->bitset_root, &cmd->bitset_root); if (r) { DMERR("dm_bitset_flush failed"); return r; } /* Flush to disk all blocks, except the superblock */ r = dm_tm_pre_commit(cmd->tm); if (r) { DMERR("dm_tm_pre_commit failed"); return r; } /* Save the space map root in cmd->metadata_space_map_root */ r = __copy_sm_root(cmd); if (r) { DMERR("__copy_sm_root failed"); return r; } /* Lock the superblock */ r = superblock_write_lock_zero(cmd, &sblock); if (r) { DMERR("Failed to write_lock superblock"); return r; } /* Save the metadata in superblock */ sb = dm_block_data(sblock); __prepare_superblock(cmd, sb); /* Unlock superblock and commit it to disk */ r = dm_tm_commit(cmd->tm, sblock); if (r) { DMERR("Failed to commit superblock"); return r; } /* * FIXME: Find a more efficient way to check if the hydration is done. */ if (bitmap_full(cmd->region_map, cmd->nr_regions)) cmd->hydration_done = true; return 0; } static int __flush_dmap(struct dm_clone_metadata *cmd, struct dirty_map *dmap) { int r; unsigned long word; word = 0; do { word = find_next_bit(dmap->dirty_words, cmd->nr_words, word); if (word == cmd->nr_words) break; r = __update_metadata_word(cmd, dmap->dirty_regions, word); if (r) return r; __clear_bit(word, dmap->dirty_words); word++; } while (word < cmd->nr_words); r = __metadata_commit(cmd); if (r) return r; /* Update the changed flag */ spin_lock_irq(&cmd->bitmap_lock); dmap->changed = 0; spin_unlock_irq(&cmd->bitmap_lock); return 0; } int dm_clone_metadata_pre_commit(struct dm_clone_metadata *cmd) { int r = 0; struct dirty_map *dmap, *next_dmap; down_write(&cmd->lock); if (cmd->fail_io || dm_bm_is_read_only(cmd->bm)) { r = -EPERM; goto out; } /* Get current dirty bitmap */ dmap = cmd->current_dmap; /* Get next dirty bitmap */ next_dmap = (dmap == &cmd->dmap[0]) ? &cmd->dmap[1] : &cmd->dmap[0]; /* * The last commit failed, so we don't have a clean dirty-bitmap to * use. */ if (WARN_ON(next_dmap->changed || cmd->committing_dmap)) { r = -EINVAL; goto out; } /* Swap dirty bitmaps */ spin_lock_irq(&cmd->bitmap_lock); cmd->current_dmap = next_dmap; spin_unlock_irq(&cmd->bitmap_lock); /* Set old dirty bitmap as currently committing */ cmd->committing_dmap = dmap; out: up_write(&cmd->lock); return r; } int dm_clone_metadata_commit(struct dm_clone_metadata *cmd) { int r = -EPERM; down_write(&cmd->lock); if (cmd->fail_io || dm_bm_is_read_only(cmd->bm)) goto out; if (WARN_ON(!cmd->committing_dmap)) { r = -EINVAL; goto out; } r = __flush_dmap(cmd, cmd->committing_dmap); if (!r) { /* Clear committing dmap */ cmd->committing_dmap = NULL; } out: up_write(&cmd->lock); return r; } int dm_clone_set_region_hydrated(struct dm_clone_metadata *cmd, unsigned long region_nr) { int r = 0; struct dirty_map *dmap; unsigned long word, flags; if (unlikely(region_nr >= cmd->nr_regions)) { DMERR("Region %lu out of range (total number of regions %lu)", region_nr, cmd->nr_regions); return -ERANGE; } word = region_nr / BITS_PER_LONG; spin_lock_irqsave(&cmd->bitmap_lock, flags); if (cmd->read_only) { r = -EPERM; goto out; } dmap = cmd->current_dmap; __set_bit(word, dmap->dirty_words); __set_bit(region_nr, dmap->dirty_regions); __set_bit(region_nr, cmd->region_map); dmap->changed = 1; out: spin_unlock_irqrestore(&cmd->bitmap_lock, flags); return r; } int dm_clone_cond_set_range(struct dm_clone_metadata *cmd, unsigned long start, unsigned long nr_regions) { int r = 0; struct dirty_map *dmap; unsigned long word, region_nr; if (unlikely(start >= cmd->nr_regions || (start + nr_regions) < start || (start + nr_regions) > cmd->nr_regions)) { DMERR("Invalid region range: start %lu, nr_regions %lu (total number of regions %lu)", start, nr_regions, cmd->nr_regions); return -ERANGE; } spin_lock_irq(&cmd->bitmap_lock); if (cmd->read_only) { r = -EPERM; goto out; } dmap = cmd->current_dmap; for (region_nr = start; region_nr < (start + nr_regions); region_nr++) { if (!test_bit(region_nr, cmd->region_map)) { word = region_nr / BITS_PER_LONG; __set_bit(word, dmap->dirty_words); __set_bit(region_nr, dmap->dirty_regions); __set_bit(region_nr, cmd->region_map); dmap->changed = 1; } } out: spin_unlock_irq(&cmd->bitmap_lock); return r; } /* * WARNING: This must not be called concurrently with either * dm_clone_set_region_hydrated() or dm_clone_cond_set_range(), as it changes * cmd->region_map without taking the cmd->bitmap_lock spinlock. The only * exception is after setting the metadata to read-only mode, using * dm_clone_metadata_set_read_only(). * * We don't take the spinlock because __load_bitset_in_core() does I/O, so it * may block. */ int dm_clone_reload_in_core_bitset(struct dm_clone_metadata *cmd) { int r = -EINVAL; down_write(&cmd->lock); if (cmd->fail_io) goto out; r = __load_bitset_in_core(cmd); out: up_write(&cmd->lock); return r; } bool dm_clone_changed_this_transaction(struct dm_clone_metadata *cmd) { bool r; unsigned long flags; spin_lock_irqsave(&cmd->bitmap_lock, flags); r = cmd->dmap[0].changed || cmd->dmap[1].changed; spin_unlock_irqrestore(&cmd->bitmap_lock, flags); return r; } int dm_clone_metadata_abort(struct dm_clone_metadata *cmd) { int r = -EPERM; down_write(&cmd->lock); if (cmd->fail_io || dm_bm_is_read_only(cmd->bm)) goto out; __destroy_persistent_data_structures(cmd); r = __create_persistent_data_structures(cmd, false); if (r) { /* If something went wrong we can neither write nor read the metadata */ cmd->fail_io = true; } out: up_write(&cmd->lock); return r; } void dm_clone_metadata_set_read_only(struct dm_clone_metadata *cmd) { down_write(&cmd->lock); spin_lock_irq(&cmd->bitmap_lock); cmd->read_only = 1; spin_unlock_irq(&cmd->bitmap_lock); if (!cmd->fail_io) dm_bm_set_read_only(cmd->bm); up_write(&cmd->lock); } void dm_clone_metadata_set_read_write(struct dm_clone_metadata *cmd) { down_write(&cmd->lock); spin_lock_irq(&cmd->bitmap_lock); cmd->read_only = 0; spin_unlock_irq(&cmd->bitmap_lock); if (!cmd->fail_io) dm_bm_set_read_write(cmd->bm); up_write(&cmd->lock); } int dm_clone_get_free_metadata_block_count(struct dm_clone_metadata *cmd, dm_block_t *result) { int r = -EINVAL; down_read(&cmd->lock); if (!cmd->fail_io) r = dm_sm_get_nr_free(cmd->sm, result); up_read(&cmd->lock); return r; } int dm_clone_get_metadata_dev_size(struct dm_clone_metadata *cmd, dm_block_t *result) { int r = -EINVAL; down_read(&cmd->lock); if (!cmd->fail_io) r = dm_sm_get_nr_blocks(cmd->sm, result); up_read(&cmd->lock); return r; }
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.
Created with Cregit http://github.com/cregit/cregit
Version 2.0-RC1