Author | Tokens | Token Proportion | Commits | Commit Proportion |
---|---|---|---|---|
Filipe David Borba Manana | 1095 | 28.50% | 1 | 2.33% |
Josef Bacik | 1036 | 26.97% | 7 | 16.28% |
Josef Whiter | 698 | 18.17% | 4 | 9.30% |
Li Dongyang | 382 | 9.94% | 1 | 2.33% |
David Sterba | 214 | 5.57% | 11 | 25.58% |
Byongho Lee | 146 | 3.80% | 1 | 2.33% |
Li Zefan | 106 | 2.76% | 2 | 4.65% |
Feifei Xu | 80 | 2.08% | 3 | 6.98% |
Jeff Mahoney | 32 | 0.83% | 2 | 4.65% |
Chris Mason | 24 | 0.62% | 3 | 6.98% |
Dan Carpenter | 12 | 0.31% | 1 | 2.33% |
Omar Sandoval | 5 | 0.13% | 1 | 2.33% |
Linda Knippers | 4 | 0.10% | 1 | 2.33% |
Dan Magenheimer | 2 | 0.05% | 1 | 2.33% |
Nicholas D Steeves | 2 | 0.05% | 1 | 2.33% |
Linus Torvalds (pre-git) | 2 | 0.05% | 1 | 2.33% |
Linus Torvalds | 1 | 0.03% | 1 | 2.33% |
David Woodhouse | 1 | 0.03% | 1 | 2.33% |
Total | 3842 | 43 |
// SPDX-License-Identifier: GPL-2.0 /* * Copyright (C) 2013 Fusion IO. All rights reserved. */ #include <linux/slab.h> #include "btrfs-tests.h" #include "../ctree.h" #include "../disk-io.h" #include "../free-space-cache.h" #include "../block-group.h" #define BITS_PER_BITMAP (PAGE_SIZE * 8UL) /* * This test just does basic sanity checking, making sure we can add an extent * entry and remove space from either end and the middle, and make sure we can * remove space that covers adjacent extent entries. */ static int test_extents(struct btrfs_block_group *cache) { int ret = 0; test_msg("running extent only tests"); /* First just make sure we can remove an entire entry */ ret = btrfs_add_free_space(cache, 0, SZ_4M); if (ret) { test_err("error adding initial extents %d", ret); return ret; } ret = btrfs_remove_free_space(cache, 0, SZ_4M); if (ret) { test_err("error removing extent %d", ret); return ret; } if (test_check_exists(cache, 0, SZ_4M)) { test_err("full remove left some lingering space"); return -1; } /* Ok edge and middle cases now */ ret = btrfs_add_free_space(cache, 0, SZ_4M); if (ret) { test_err("error adding half extent %d", ret); return ret; } ret = btrfs_remove_free_space(cache, 3 * SZ_1M, SZ_1M); if (ret) { test_err("error removing tail end %d", ret); return ret; } ret = btrfs_remove_free_space(cache, 0, SZ_1M); if (ret) { test_err("error removing front end %d", ret); return ret; } ret = btrfs_remove_free_space(cache, SZ_2M, 4096); if (ret) { test_err("error removing middle piece %d", ret); return ret; } if (test_check_exists(cache, 0, SZ_1M)) { test_err("still have space at the front"); return -1; } if (test_check_exists(cache, SZ_2M, 4096)) { test_err("still have space in the middle"); return -1; } if (test_check_exists(cache, 3 * SZ_1M, SZ_1M)) { test_err("still have space at the end"); return -1; } /* Cleanup */ btrfs_remove_free_space_cache(cache); return 0; } static int test_bitmaps(struct btrfs_block_group *cache, u32 sectorsize) { u64 next_bitmap_offset; int ret; test_msg("running bitmap only tests"); ret = test_add_free_space_entry(cache, 0, SZ_4M, 1); if (ret) { test_err("couldn't create a bitmap entry %d", ret); return ret; } ret = btrfs_remove_free_space(cache, 0, SZ_4M); if (ret) { test_err("error removing bitmap full range %d", ret); return ret; } if (test_check_exists(cache, 0, SZ_4M)) { test_err("left some space in bitmap"); return -1; } ret = test_add_free_space_entry(cache, 0, SZ_4M, 1); if (ret) { test_err("couldn't add to our bitmap entry %d", ret); return ret; } ret = btrfs_remove_free_space(cache, SZ_1M, SZ_2M); if (ret) { test_err("couldn't remove middle chunk %d", ret); return ret; } /* * The first bitmap we have starts at offset 0 so the next one is just * at the end of the first bitmap. */ next_bitmap_offset = (u64)(BITS_PER_BITMAP * sectorsize); /* Test a bit straddling two bitmaps */ ret = test_add_free_space_entry(cache, next_bitmap_offset - SZ_2M, SZ_4M, 1); if (ret) { test_err("couldn't add space that straddles two bitmaps %d", ret); return ret; } ret = btrfs_remove_free_space(cache, next_bitmap_offset - SZ_1M, SZ_2M); if (ret) { test_err("couldn't remove overlapping space %d", ret); return ret; } if (test_check_exists(cache, next_bitmap_offset - SZ_1M, SZ_2M)) { test_err("left some space when removing overlapping"); return -1; } btrfs_remove_free_space_cache(cache); return 0; } /* This is the high grade jackassery */ static int test_bitmaps_and_extents(struct btrfs_block_group *cache, u32 sectorsize) { u64 bitmap_offset = (u64)(BITS_PER_BITMAP * sectorsize); int ret; test_msg("running bitmap and extent tests"); /* * First let's do something simple, an extent at the same offset as the * bitmap, but the free space completely in the extent and then * completely in the bitmap. */ ret = test_add_free_space_entry(cache, SZ_4M, SZ_1M, 1); if (ret) { test_err("couldn't create bitmap entry %d", ret); return ret; } ret = test_add_free_space_entry(cache, 0, SZ_1M, 0); if (ret) { test_err("couldn't add extent entry %d", ret); return ret; } ret = btrfs_remove_free_space(cache, 0, SZ_1M); if (ret) { test_err("couldn't remove extent entry %d", ret); return ret; } if (test_check_exists(cache, 0, SZ_1M)) { test_err("left remnants after our remove"); return -1; } /* Now to add back the extent entry and remove from the bitmap */ ret = test_add_free_space_entry(cache, 0, SZ_1M, 0); if (ret) { test_err("couldn't re-add extent entry %d", ret); return ret; } ret = btrfs_remove_free_space(cache, SZ_4M, SZ_1M); if (ret) { test_err("couldn't remove from bitmap %d", ret); return ret; } if (test_check_exists(cache, SZ_4M, SZ_1M)) { test_err("left remnants in the bitmap"); return -1; } /* * Ok so a little more evil, extent entry and bitmap at the same offset, * removing an overlapping chunk. */ ret = test_add_free_space_entry(cache, SZ_1M, SZ_4M, 1); if (ret) { test_err("couldn't add to a bitmap %d", ret); return ret; } ret = btrfs_remove_free_space(cache, SZ_512K, 3 * SZ_1M); if (ret) { test_err("couldn't remove overlapping space %d", ret); return ret; } if (test_check_exists(cache, SZ_512K, 3 * SZ_1M)) { test_err("left over pieces after removing overlapping"); return -1; } btrfs_remove_free_space_cache(cache); /* Now with the extent entry offset into the bitmap */ ret = test_add_free_space_entry(cache, SZ_4M, SZ_4M, 1); if (ret) { test_err("couldn't add space to the bitmap %d", ret); return ret; } ret = test_add_free_space_entry(cache, SZ_2M, SZ_2M, 0); if (ret) { test_err("couldn't add extent to the cache %d", ret); return ret; } ret = btrfs_remove_free_space(cache, 3 * SZ_1M, SZ_4M); if (ret) { test_err("problem removing overlapping space %d", ret); return ret; } if (test_check_exists(cache, 3 * SZ_1M, SZ_4M)) { test_err("left something behind when removing space"); return -1; } /* * This has blown up in the past, the extent entry starts before the * bitmap entry, but we're trying to remove an offset that falls * completely within the bitmap range and is in both the extent entry * and the bitmap entry, looks like this * * [ extent ] * [ bitmap ] * [ del ] */ btrfs_remove_free_space_cache(cache); ret = test_add_free_space_entry(cache, bitmap_offset + SZ_4M, SZ_4M, 1); if (ret) { test_err("couldn't add bitmap %d", ret); return ret; } ret = test_add_free_space_entry(cache, bitmap_offset - SZ_1M, 5 * SZ_1M, 0); if (ret) { test_err("couldn't add extent entry %d", ret); return ret; } ret = btrfs_remove_free_space(cache, bitmap_offset + SZ_1M, 5 * SZ_1M); if (ret) { test_err("failed to free our space %d", ret); return ret; } if (test_check_exists(cache, bitmap_offset + SZ_1M, 5 * SZ_1M)) { test_err("left stuff over"); return -1; } btrfs_remove_free_space_cache(cache); /* * This blew up before, we have part of the free space in a bitmap and * then the entirety of the rest of the space in an extent. This used * to return -EAGAIN back from btrfs_remove_extent, make sure this * doesn't happen. */ ret = test_add_free_space_entry(cache, SZ_1M, SZ_2M, 1); if (ret) { test_err("couldn't add bitmap entry %d", ret); return ret; } ret = test_add_free_space_entry(cache, 3 * SZ_1M, SZ_1M, 0); if (ret) { test_err("couldn't add extent entry %d", ret); return ret; } ret = btrfs_remove_free_space(cache, SZ_1M, 3 * SZ_1M); if (ret) { test_err("error removing bitmap and extent overlapping %d", ret); return ret; } btrfs_remove_free_space_cache(cache); return 0; } /* Used by test_steal_space_from_bitmap_to_extent(). */ static bool test_use_bitmap(struct btrfs_free_space_ctl *ctl, struct btrfs_free_space *info) { return ctl->free_extents > 0; } /* Used by test_steal_space_from_bitmap_to_extent(). */ static int check_num_extents_and_bitmaps(const struct btrfs_block_group *cache, const int num_extents, const int num_bitmaps) { if (cache->free_space_ctl->free_extents != num_extents) { test_err( "incorrect # of extent entries in the cache: %d, expected %d", cache->free_space_ctl->free_extents, num_extents); return -EINVAL; } if (cache->free_space_ctl->total_bitmaps != num_bitmaps) { test_err( "incorrect # of extent entries in the cache: %d, expected %d", cache->free_space_ctl->total_bitmaps, num_bitmaps); return -EINVAL; } return 0; } /* Used by test_steal_space_from_bitmap_to_extent(). */ static int check_cache_empty(struct btrfs_block_group *cache) { u64 offset; u64 max_extent_size; /* * Now lets confirm that there's absolutely no free space left to * allocate. */ if (cache->free_space_ctl->free_space != 0) { test_err("cache free space is not 0"); return -EINVAL; } /* And any allocation request, no matter how small, should fail now. */ offset = btrfs_find_space_for_alloc(cache, 0, 4096, 0, &max_extent_size); if (offset != 0) { test_err("space allocation did not fail, returned offset: %llu", offset); return -EINVAL; } /* And no extent nor bitmap entries in the cache anymore. */ return check_num_extents_and_bitmaps(cache, 0, 0); } /* * Before we were able to steal free space from a bitmap entry to an extent * entry, we could end up with 2 entries representing a contiguous free space. * One would be an extent entry and the other a bitmap entry. Since in order * to allocate space to a caller we use only 1 entry, we couldn't return that * whole range to the caller if it was requested. This forced the caller to * either assume ENOSPC or perform several smaller space allocations, which * wasn't optimal as they could be spread all over the block group while under * concurrency (extra overhead and fragmentation). * * This stealing approach is beneficial, since we always prefer to allocate * from extent entries, both for clustered and non-clustered allocation * requests. */ static int test_steal_space_from_bitmap_to_extent(struct btrfs_block_group *cache, u32 sectorsize) { int ret; u64 offset; u64 max_extent_size; const struct btrfs_free_space_op test_free_space_ops = { .use_bitmap = test_use_bitmap, }; const struct btrfs_free_space_op *orig_free_space_ops; test_msg("running space stealing from bitmap to extent tests"); /* * For this test, we want to ensure we end up with an extent entry * immediately adjacent to a bitmap entry, where the bitmap starts * at an offset where the extent entry ends. We keep adding and * removing free space to reach into this state, but to get there * we need to reach a point where marking new free space doesn't * result in adding new extent entries or merging the new space * with existing extent entries - the space ends up being marked * in an existing bitmap that covers the new free space range. * * To get there, we need to reach the threshold defined set at * cache->free_space_ctl->extents_thresh, which currently is * 256 extents on a x86_64 system at least, and a few other * conditions (check free_space_cache.c). Instead of making the * test much longer and complicated, use a "use_bitmap" operation * that forces use of bitmaps as soon as we have at least 1 * extent entry. */ orig_free_space_ops = cache->free_space_ctl->op; cache->free_space_ctl->op = &test_free_space_ops; /* * Extent entry covering free space range [128Mb - 256Kb, 128Mb - 128Kb[ */ ret = test_add_free_space_entry(cache, SZ_128M - SZ_256K, SZ_128K, 0); if (ret) { test_err("couldn't add extent entry %d", ret); return ret; } /* Bitmap entry covering free space range [128Mb + 512Kb, 256Mb[ */ ret = test_add_free_space_entry(cache, SZ_128M + SZ_512K, SZ_128M - SZ_512K, 1); if (ret) { test_err("couldn't add bitmap entry %d", ret); return ret; } ret = check_num_extents_and_bitmaps(cache, 2, 1); if (ret) return ret; /* * Now make only the first 256Kb of the bitmap marked as free, so that * we end up with only the following ranges marked as free space: * * [128Mb - 256Kb, 128Mb - 128Kb[ * [128Mb + 512Kb, 128Mb + 768Kb[ */ ret = btrfs_remove_free_space(cache, SZ_128M + 768 * SZ_1K, SZ_128M - 768 * SZ_1K); if (ret) { test_err("failed to free part of bitmap space %d", ret); return ret; } /* Confirm that only those 2 ranges are marked as free. */ if (!test_check_exists(cache, SZ_128M - SZ_256K, SZ_128K)) { test_err("free space range missing"); return -ENOENT; } if (!test_check_exists(cache, SZ_128M + SZ_512K, SZ_256K)) { test_err("free space range missing"); return -ENOENT; } /* * Confirm that the bitmap range [128Mb + 768Kb, 256Mb[ isn't marked * as free anymore. */ if (test_check_exists(cache, SZ_128M + 768 * SZ_1K, SZ_128M - 768 * SZ_1K)) { test_err("bitmap region not removed from space cache"); return -EINVAL; } /* * Confirm that the region [128Mb + 256Kb, 128Mb + 512Kb[, which is * covered by the bitmap, isn't marked as free. */ if (test_check_exists(cache, SZ_128M + SZ_256K, SZ_256K)) { test_err("invalid bitmap region marked as free"); return -EINVAL; } /* * Confirm that the region [128Mb, 128Mb + 256Kb[, which is covered * by the bitmap too, isn't marked as free either. */ if (test_check_exists(cache, SZ_128M, SZ_256K)) { test_err("invalid bitmap region marked as free"); return -EINVAL; } /* * Now lets mark the region [128Mb, 128Mb + 512Kb[ as free too. But, * lets make sure the free space cache marks it as free in the bitmap, * and doesn't insert a new extent entry to represent this region. */ ret = btrfs_add_free_space(cache, SZ_128M, SZ_512K); if (ret) { test_err("error adding free space: %d", ret); return ret; } /* Confirm the region is marked as free. */ if (!test_check_exists(cache, SZ_128M, SZ_512K)) { test_err("bitmap region not marked as free"); return -ENOENT; } /* * Confirm that no new extent entries or bitmap entries were added to * the cache after adding that free space region. */ ret = check_num_extents_and_bitmaps(cache, 2, 1); if (ret) return ret; /* * Now lets add a small free space region to the right of the previous * one, which is not contiguous with it and is part of the bitmap too. * The goal is to test that the bitmap entry space stealing doesn't * steal this space region. */ ret = btrfs_add_free_space(cache, SZ_128M + SZ_16M, sectorsize); if (ret) { test_err("error adding free space: %d", ret); return ret; } /* * Confirm that no new extent entries or bitmap entries were added to * the cache after adding that free space region. */ ret = check_num_extents_and_bitmaps(cache, 2, 1); if (ret) return ret; /* * Now mark the region [128Mb - 128Kb, 128Mb[ as free too. This will * expand the range covered by the existing extent entry that represents * the free space [128Mb - 256Kb, 128Mb - 128Kb[. */ ret = btrfs_add_free_space(cache, SZ_128M - SZ_128K, SZ_128K); if (ret) { test_err("error adding free space: %d", ret); return ret; } /* Confirm the region is marked as free. */ if (!test_check_exists(cache, SZ_128M - SZ_128K, SZ_128K)) { test_err("extent region not marked as free"); return -ENOENT; } /* * Confirm that our extent entry didn't stole all free space from the * bitmap, because of the small 4Kb free space region. */ ret = check_num_extents_and_bitmaps(cache, 2, 1); if (ret) return ret; /* * So now we have the range [128Mb - 256Kb, 128Mb + 768Kb[ as free * space. Without stealing bitmap free space into extent entry space, * we would have all this free space represented by 2 entries in the * cache: * * extent entry covering range: [128Mb - 256Kb, 128Mb[ * bitmap entry covering range: [128Mb, 128Mb + 768Kb[ * * Attempting to allocate the whole free space (1Mb) would fail, because * we can't allocate from multiple entries. * With the bitmap free space stealing, we get a single extent entry * that represents the 1Mb free space, and therefore we're able to * allocate the whole free space at once. */ if (!test_check_exists(cache, SZ_128M - SZ_256K, SZ_1M)) { test_err("expected region not marked as free"); return -ENOENT; } if (cache->free_space_ctl->free_space != (SZ_1M + sectorsize)) { test_err("cache free space is not 1Mb + %u", sectorsize); return -EINVAL; } offset = btrfs_find_space_for_alloc(cache, 0, SZ_1M, 0, &max_extent_size); if (offset != (SZ_128M - SZ_256K)) { test_err( "failed to allocate 1Mb from space cache, returned offset is: %llu", offset); return -EINVAL; } /* * All that remains is a sectorsize free space region in a bitmap. * Confirm. */ ret = check_num_extents_and_bitmaps(cache, 1, 1); if (ret) return ret; if (cache->free_space_ctl->free_space != sectorsize) { test_err("cache free space is not %u", sectorsize); return -EINVAL; } offset = btrfs_find_space_for_alloc(cache, 0, sectorsize, 0, &max_extent_size); if (offset != (SZ_128M + SZ_16M)) { test_err("failed to allocate %u, returned offset : %llu", sectorsize, offset); return -EINVAL; } ret = check_cache_empty(cache); if (ret) return ret; btrfs_remove_free_space_cache(cache); /* * Now test a similar scenario, but where our extent entry is located * to the right of the bitmap entry, so that we can check that stealing * space from a bitmap to the front of an extent entry works. */ /* * Extent entry covering free space range [128Mb + 128Kb, 128Mb + 256Kb[ */ ret = test_add_free_space_entry(cache, SZ_128M + SZ_128K, SZ_128K, 0); if (ret) { test_err("couldn't add extent entry %d", ret); return ret; } /* Bitmap entry covering free space range [0, 128Mb - 512Kb[ */ ret = test_add_free_space_entry(cache, 0, SZ_128M - SZ_512K, 1); if (ret) { test_err("couldn't add bitmap entry %d", ret); return ret; } ret = check_num_extents_and_bitmaps(cache, 2, 1); if (ret) return ret; /* * Now make only the last 256Kb of the bitmap marked as free, so that * we end up with only the following ranges marked as free space: * * [128Mb + 128b, 128Mb + 256Kb[ * [128Mb - 768Kb, 128Mb - 512Kb[ */ ret = btrfs_remove_free_space(cache, 0, SZ_128M - 768 * SZ_1K); if (ret) { test_err("failed to free part of bitmap space %d", ret); return ret; } /* Confirm that only those 2 ranges are marked as free. */ if (!test_check_exists(cache, SZ_128M + SZ_128K, SZ_128K)) { test_err("free space range missing"); return -ENOENT; } if (!test_check_exists(cache, SZ_128M - 768 * SZ_1K, SZ_256K)) { test_err("free space range missing"); return -ENOENT; } /* * Confirm that the bitmap range [0, 128Mb - 768Kb[ isn't marked * as free anymore. */ if (test_check_exists(cache, 0, SZ_128M - 768 * SZ_1K)) { test_err("bitmap region not removed from space cache"); return -EINVAL; } /* * Confirm that the region [128Mb - 512Kb, 128Mb[, which is * covered by the bitmap, isn't marked as free. */ if (test_check_exists(cache, SZ_128M - SZ_512K, SZ_512K)) { test_err("invalid bitmap region marked as free"); return -EINVAL; } /* * Now lets mark the region [128Mb - 512Kb, 128Mb[ as free too. But, * lets make sure the free space cache marks it as free in the bitmap, * and doesn't insert a new extent entry to represent this region. */ ret = btrfs_add_free_space(cache, SZ_128M - SZ_512K, SZ_512K); if (ret) { test_err("error adding free space: %d", ret); return ret; } /* Confirm the region is marked as free. */ if (!test_check_exists(cache, SZ_128M - SZ_512K, SZ_512K)) { test_err("bitmap region not marked as free"); return -ENOENT; } /* * Confirm that no new extent entries or bitmap entries were added to * the cache after adding that free space region. */ ret = check_num_extents_and_bitmaps(cache, 2, 1); if (ret) return ret; /* * Now lets add a small free space region to the left of the previous * one, which is not contiguous with it and is part of the bitmap too. * The goal is to test that the bitmap entry space stealing doesn't * steal this space region. */ ret = btrfs_add_free_space(cache, SZ_32M, 2 * sectorsize); if (ret) { test_err("error adding free space: %d", ret); return ret; } /* * Now mark the region [128Mb, 128Mb + 128Kb[ as free too. This will * expand the range covered by the existing extent entry that represents * the free space [128Mb + 128Kb, 128Mb + 256Kb[. */ ret = btrfs_add_free_space(cache, SZ_128M, SZ_128K); if (ret) { test_err("error adding free space: %d", ret); return ret; } /* Confirm the region is marked as free. */ if (!test_check_exists(cache, SZ_128M, SZ_128K)) { test_err("extent region not marked as free"); return -ENOENT; } /* * Confirm that our extent entry didn't stole all free space from the * bitmap, because of the small 2 * sectorsize free space region. */ ret = check_num_extents_and_bitmaps(cache, 2, 1); if (ret) return ret; /* * So now we have the range [128Mb - 768Kb, 128Mb + 256Kb[ as free * space. Without stealing bitmap free space into extent entry space, * we would have all this free space represented by 2 entries in the * cache: * * extent entry covering range: [128Mb, 128Mb + 256Kb[ * bitmap entry covering range: [128Mb - 768Kb, 128Mb[ * * Attempting to allocate the whole free space (1Mb) would fail, because * we can't allocate from multiple entries. * With the bitmap free space stealing, we get a single extent entry * that represents the 1Mb free space, and therefore we're able to * allocate the whole free space at once. */ if (!test_check_exists(cache, SZ_128M - 768 * SZ_1K, SZ_1M)) { test_err("expected region not marked as free"); return -ENOENT; } if (cache->free_space_ctl->free_space != (SZ_1M + 2 * sectorsize)) { test_err("cache free space is not 1Mb + %u", 2 * sectorsize); return -EINVAL; } offset = btrfs_find_space_for_alloc(cache, 0, SZ_1M, 0, &max_extent_size); if (offset != (SZ_128M - 768 * SZ_1K)) { test_err( "failed to allocate 1Mb from space cache, returned offset is: %llu", offset); return -EINVAL; } /* * All that remains is 2 * sectorsize free space region * in a bitmap. Confirm. */ ret = check_num_extents_and_bitmaps(cache, 1, 1); if (ret) return ret; if (cache->free_space_ctl->free_space != 2 * sectorsize) { test_err("cache free space is not %u", 2 * sectorsize); return -EINVAL; } offset = btrfs_find_space_for_alloc(cache, 0, 2 * sectorsize, 0, &max_extent_size); if (offset != SZ_32M) { test_err("failed to allocate %u, offset: %llu", 2 * sectorsize, offset); return -EINVAL; } ret = check_cache_empty(cache); if (ret) return ret; cache->free_space_ctl->op = orig_free_space_ops; btrfs_remove_free_space_cache(cache); return 0; } static bool bytes_index_use_bitmap(struct btrfs_free_space_ctl *ctl, struct btrfs_free_space *info) { return true; } static int test_bytes_index(struct btrfs_block_group *cache, u32 sectorsize) { const struct btrfs_free_space_op test_free_space_ops = { .use_bitmap = bytes_index_use_bitmap, }; const struct btrfs_free_space_op *orig_free_space_ops; struct btrfs_free_space_ctl *ctl = cache->free_space_ctl; struct btrfs_free_space *entry; struct rb_node *node; u64 offset, max_extent_size, bytes; int ret, i; test_msg("running bytes index tests"); /* First just validate that it does everything in order. */ offset = 0; for (i = 0; i < 10; i++) { bytes = (i + 1) * SZ_1M; ret = test_add_free_space_entry(cache, offset, bytes, 0); if (ret) { test_err("couldn't add extent entry %d\n", ret); return ret; } offset += bytes + sectorsize; } for (node = rb_first_cached(&ctl->free_space_bytes), i = 9; node; node = rb_next(node), i--) { entry = rb_entry(node, struct btrfs_free_space, bytes_index); bytes = (i + 1) * SZ_1M; if (entry->bytes != bytes) { test_err("invalid bytes index order, found %llu expected %llu", entry->bytes, bytes); return -EINVAL; } } /* Now validate bitmaps do the correct thing. */ btrfs_remove_free_space_cache(cache); for (i = 0; i < 2; i++) { offset = i * BITS_PER_BITMAP * sectorsize; bytes = (i + 1) * SZ_1M; ret = test_add_free_space_entry(cache, offset, bytes, 1); if (ret) { test_err("couldn't add bitmap entry"); return ret; } } for (node = rb_first_cached(&ctl->free_space_bytes), i = 1; node; node = rb_next(node), i--) { entry = rb_entry(node, struct btrfs_free_space, bytes_index); bytes = (i + 1) * SZ_1M; if (entry->bytes != bytes) { test_err("invalid bytes index order, found %llu expected %llu", entry->bytes, bytes); return -EINVAL; } } /* Now validate bitmaps with different ->max_extent_size. */ btrfs_remove_free_space_cache(cache); orig_free_space_ops = cache->free_space_ctl->op; cache->free_space_ctl->op = &test_free_space_ops; ret = test_add_free_space_entry(cache, 0, sectorsize, 1); if (ret) { test_err("couldn't add bitmap entry"); return ret; } offset = BITS_PER_BITMAP * sectorsize; ret = test_add_free_space_entry(cache, offset, sectorsize, 1); if (ret) { test_err("couldn't add bitmap_entry"); return ret; } /* * Now set a bunch of sectorsize extents in the first entry so it's * ->bytes is large. */ for (i = 2; i < 20; i += 2) { offset = sectorsize * i; ret = btrfs_add_free_space(cache, offset, sectorsize); if (ret) { test_err("error populating sparse bitmap %d", ret); return ret; } } /* * Now set a contiguous extent in the second bitmap so its * ->max_extent_size is larger than the first bitmaps. */ offset = (BITS_PER_BITMAP * sectorsize) + sectorsize; ret = btrfs_add_free_space(cache, offset, sectorsize); if (ret) { test_err("error adding contiguous extent %d", ret); return ret; } /* * Since we don't set ->max_extent_size unless we search everything * should be indexed on bytes. */ entry = rb_entry(rb_first_cached(&ctl->free_space_bytes), struct btrfs_free_space, bytes_index); if (entry->bytes != (10 * sectorsize)) { test_err("error, wrong entry in the first slot in bytes_index"); return -EINVAL; } max_extent_size = 0; offset = btrfs_find_space_for_alloc(cache, cache->start, sectorsize * 3, 0, &max_extent_size); if (offset != 0) { test_err("found space to alloc even though we don't have enough space"); return -EINVAL; } if (max_extent_size != (2 * sectorsize)) { test_err("got the wrong max_extent size %llu expected %llu", max_extent_size, (unsigned long long)(2 * sectorsize)); return -EINVAL; } /* * The search should have re-arranged the bytes index to use the * ->max_extent_size, validate it's now what we expect it to be. */ entry = rb_entry(rb_first_cached(&ctl->free_space_bytes), struct btrfs_free_space, bytes_index); if (entry->bytes != (2 * sectorsize)) { test_err("error, the bytes index wasn't recalculated properly"); return -EINVAL; } /* Add another sectorsize to re-arrange the tree back to ->bytes. */ offset = (BITS_PER_BITMAP * sectorsize) - sectorsize; ret = btrfs_add_free_space(cache, offset, sectorsize); if (ret) { test_err("error adding extent to the sparse entry %d", ret); return ret; } entry = rb_entry(rb_first_cached(&ctl->free_space_bytes), struct btrfs_free_space, bytes_index); if (entry->bytes != (11 * sectorsize)) { test_err("error, wrong entry in the first slot in bytes_index"); return -EINVAL; } /* * Now make sure we find our correct entry after searching that will * result in a re-arranging of the tree. */ max_extent_size = 0; offset = btrfs_find_space_for_alloc(cache, cache->start, sectorsize * 2, 0, &max_extent_size); if (offset != (BITS_PER_BITMAP * sectorsize)) { test_err("error, found %llu instead of %llu for our alloc", offset, (unsigned long long)(BITS_PER_BITMAP * sectorsize)); return -EINVAL; } cache->free_space_ctl->op = orig_free_space_ops; btrfs_remove_free_space_cache(cache); return 0; } int btrfs_test_free_space_cache(u32 sectorsize, u32 nodesize) { struct btrfs_fs_info *fs_info; struct btrfs_block_group *cache; struct btrfs_root *root = NULL; int ret = -ENOMEM; test_msg("running btrfs free space cache tests"); fs_info = btrfs_alloc_dummy_fs_info(nodesize, sectorsize); if (!fs_info) { test_std_err(TEST_ALLOC_FS_INFO); return -ENOMEM; } /* * For ppc64 (with 64k page size), bytes per bitmap might be * larger than 1G. To make bitmap test available in ppc64, * alloc dummy block group whose size cross bitmaps. */ cache = btrfs_alloc_dummy_block_group(fs_info, BITS_PER_BITMAP * sectorsize + PAGE_SIZE); if (!cache) { test_std_err(TEST_ALLOC_BLOCK_GROUP); btrfs_free_dummy_fs_info(fs_info); return 0; } root = btrfs_alloc_dummy_root(fs_info); if (IS_ERR(root)) { test_std_err(TEST_ALLOC_ROOT); ret = PTR_ERR(root); goto out; } root->root_key.objectid = BTRFS_EXTENT_TREE_OBJECTID; root->root_key.type = BTRFS_ROOT_ITEM_KEY; root->root_key.offset = 0; btrfs_global_root_insert(root); ret = test_extents(cache); if (ret) goto out; ret = test_bitmaps(cache, sectorsize); if (ret) goto out; ret = test_bitmaps_and_extents(cache, sectorsize); if (ret) goto out; ret = test_steal_space_from_bitmap_to_extent(cache, sectorsize); if (ret) goto out; ret = test_bytes_index(cache, sectorsize); out: btrfs_free_dummy_block_group(cache); btrfs_free_dummy_root(root); btrfs_free_dummy_fs_info(fs_info); return ret; }
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.
Created with Cregit http://github.com/cregit/cregit
Version 2.0-RC1