cregit-Linux how code gets into the kernel

Release 4.10 block/blk-wbt.c

Directory: block
/*
 * buffered writeback throttling. loosely based on CoDel. We can't drop
 * packets for IO scheduling, so the logic is something like this:
 *
 * - Monitor latencies in a defined window of time.
 * - If the minimum latency in the above window exceeds some target, increment
 *   scaling step and scale down queue depth by a factor of 2x. The monitoring
 *   window is then shrunk to 100 / sqrt(scaling step + 1).
 * - For any window where we don't have solid data on what the latencies
 *   look like, retain status quo.
 * - If latencies look good, decrement scaling step.
 * - If we're only doing writes, allow the scaling step to go negative. This
 *   will temporarily boost write performance, snapping back to a stable
 *   scaling step of 0 if reads show up or the heavy writers finish. Unlike
 *   positive scaling steps where we shrink the monitoring window, a negative
 *   scaling step retains the default step==0 window size.
 *
 * Copyright (C) 2016 Jens Axboe
 *
 */
#include <linux/kernel.h>
#include <linux/blk_types.h>
#include <linux/slab.h>
#include <linux/backing-dev.h>
#include <linux/swap.h>

#include "blk-wbt.h"


#define CREATE_TRACE_POINTS
#include <trace/events/wbt.h>

enum {
	/*
         * Default setting, we'll scale up (to 75% of QD max) or down (min 1)
         * from here depending on device stats
         */
	
RWB_DEF_DEPTH	= 16,

	/*
         * 100msec window
         */
	
RWB_WINDOW_NSEC		= 100 * 1000 * 1000ULL,

	/*
         * Disregard stats, if we don't meet this minimum
         */
	
RWB_MIN_WRITE_SAMPLES	= 3,

	/*
         * If we have this number of consecutive windows with not enough
         * information to scale up or down, scale up.
         */
	
RWB_UNKNOWN_BUMP	= 5,
};


static inline bool rwb_enabled(struct rq_wb *rwb) { return rwb && rwb->wb_normal != 0; }

Contributors

PersonTokensPropCommitsCommitProp
jens axboejens axboe21100.00%1100.00%
Total21100.00%1100.00%

/* * Increment 'v', if 'v' is below 'below'. Returns true if we succeeded, * false if 'v' + 1 would be bigger than 'below'. */
static bool atomic_inc_below(atomic_t *v, int below) { int cur = atomic_read(v); for (;;) { int old; if (cur >= below) return false; old = atomic_cmpxchg(v, cur, cur + 1); if (old == cur) break; cur = old; } return true; }

Contributors

PersonTokensPropCommitsCommitProp
jens axboejens axboe67100.00%1100.00%
Total67100.00%1100.00%


static void wb_timestamp(struct rq_wb *rwb, unsigned long *var) { if (rwb_enabled(rwb)) { const unsigned long cur = jiffies; if (cur != *var) *var = cur; } }

Contributors

PersonTokensPropCommitsCommitProp
jens axboejens axboe44100.00%1100.00%
Total44100.00%1100.00%

/* * If a task was rate throttled in balance_dirty_pages() within the last * second or so, use that to indicate a higher cleaning rate. */
static bool wb_recent_wait(struct rq_wb *rwb) { struct bdi_writeback *wb = &rwb->queue->backing_dev_info.wb; return time_before(jiffies, wb->dirty_sleep + HZ); }

Contributors

PersonTokensPropCommitsCommitProp
jens axboejens axboe37100.00%2100.00%
Total37100.00%2100.00%


static inline struct rq_wait *get_rq_wait(struct rq_wb *rwb, bool is_kswapd) { return &rwb->rq_wait[is_kswapd]; }

Contributors

PersonTokensPropCommitsCommitProp
jens axboejens axboe26100.00%1100.00%
Total26100.00%1100.00%


static void rwb_wake_all(struct rq_wb *rwb) { int i; for (i = 0; i < WBT_NUM_RWQ; i++) { struct rq_wait *rqw = &rwb->rq_wait[i]; if (waitqueue_active(&rqw->wait)) wake_up_all(&rqw->wait); } }

Contributors

PersonTokensPropCommitsCommitProp
jens axboejens axboe60100.00%1100.00%
Total60100.00%1100.00%


void __wbt_done(struct rq_wb *rwb, enum wbt_flags wb_acct) { struct rq_wait *rqw; int inflight, limit; if (!(wb_acct & WBT_TRACKED)) return; rqw = get_rq_wait(rwb, wb_acct & WBT_KSWAPD); inflight = atomic_dec_return(&rqw->inflight); /* * wbt got disabled with IO in flight. Wake up any potential * waiters, we don't have to do more than that. */ if (unlikely(!rwb_enabled(rwb))) { rwb_wake_all(rwb); return; } /* * If the device does write back caching, drop further down * before we wake people up. */ if (rwb->wc && !wb_recent_wait(rwb)) limit = 0; else limit = rwb->wb_normal; /* * Don't wake anyone up if we are above the normal limit. */ if (inflight && inflight >= limit) return; if (waitqueue_active(&rqw->wait)) { int diff = limit - inflight; if (!inflight || diff >= rwb->wb_background / 2) wake_up_all(&rqw->wait); } }

Contributors

PersonTokensPropCommitsCommitProp
jens axboejens axboe149100.00%1100.00%
Total149100.00%1100.00%

/* * Called on completion of a request. Note that it's also called when * a request is merged, when the request gets freed. */
void wbt_done(struct rq_wb *rwb, struct blk_issue_stat *stat) { if (!rwb) return; if (!wbt_is_tracked(stat)) { if (rwb->sync_cookie == stat) { rwb->sync_issue = 0; rwb->sync_cookie = NULL; } if (wbt_is_read(stat)) wb_timestamp(rwb, &rwb->last_comp); wbt_clear_state(stat); } else { WARN_ON_ONCE(stat == rwb->sync_cookie); __wbt_done(rwb, wbt_stat_to_mask(stat)); wbt_clear_state(stat); } }

Contributors

PersonTokensPropCommitsCommitProp
jens axboejens axboe102100.00%1100.00%
Total102100.00%1100.00%

/* * Return true, if we can't increase the depth further by scaling */
static bool calc_wb_limits(struct rq_wb *rwb) { unsigned int depth; bool ret = false; if (!rwb->min_lat_nsec) { rwb->wb_max = rwb->wb_normal = rwb->wb_background = 0; return false; } /* * For QD=1 devices, this is a special case. It's important for those * to have one request ready when one completes, so force a depth of * 2 for those devices. On the backend, it'll be a depth of 1 anyway, * since the device can't have more than that in flight. If we're * scaling down, then keep a setting of 1/1/1. */ if (rwb->queue_depth == 1) { if (rwb->scale_step > 0) rwb->wb_max = rwb->wb_normal = 1; else { rwb->wb_max = rwb->wb_normal = 2; ret = true; } rwb->wb_background = 1; } else { /* * scale_step == 0 is our default state. If we have suffered * latency spikes, step will be > 0, and we shrink the * allowed write depths. If step is < 0, we're only doing * writes, and we allow a temporarily higher depth to * increase performance. */ depth = min_t(unsigned int, RWB_DEF_DEPTH, rwb->queue_depth); if (rwb->scale_step > 0) depth = 1 + ((depth - 1) >> min(31, rwb->scale_step)); else if (rwb->scale_step < 0) { unsigned int maxd = 3 * rwb->queue_depth / 4; depth = 1 + ((depth - 1) << -rwb->scale_step); if (depth > maxd) { depth = maxd; ret = true; } } /* * Set our max/normal/bg queue depths based on how far * we have scaled down (->scale_step). */ rwb->wb_max = depth; rwb->wb_normal = (rwb->wb_max + 1) / 2; rwb->wb_background = (rwb->wb_max + 3) / 4; } return ret; }

Contributors

PersonTokensPropCommitsCommitProp
jens axboejens axboe236100.00%1100.00%
Total236100.00%1100.00%


static inline bool stat_sample_valid(struct blk_rq_stat *stat) { /* * We need at least one read sample, and a minimum of * RWB_MIN_WRITE_SAMPLES. We require some write samples to know * that it's writes impacting us, and not just some sole read on * a device that is in a lower power state. */ return stat[BLK_STAT_READ].nr_samples >= 1 && stat[BLK_STAT_WRITE].nr_samples >= RWB_MIN_WRITE_SAMPLES; }

Contributors

PersonTokensPropCommitsCommitProp
jens axboejens axboe3196.88%266.67%
arnd bergmannarnd bergmann13.12%133.33%
Total32100.00%3100.00%


static u64 rwb_sync_issue_lat(struct rq_wb *rwb) { u64 now, issue = ACCESS_ONCE(rwb->sync_issue); if (!issue || !rwb->sync_cookie) return 0; now = ktime_to_ns(ktime_get()); return now - issue; }

Contributors

PersonTokensPropCommitsCommitProp
jens axboejens axboe49100.00%1100.00%
Total49100.00%1100.00%

enum { LAT_OK = 1, LAT_UNKNOWN, LAT_UNKNOWN_WRITES, LAT_EXCEEDED, };
static int __latency_exceeded(struct rq_wb *rwb, struct blk_rq_stat *stat) { struct backing_dev_info *bdi = &rwb->queue->backing_dev_info; u64 thislat; /* * If our stored sync issue exceeds the window size, or it * exceeds our min target AND we haven't logged any entries, * flag the latency as exceeded. wbt works off completion latencies, * but for a flooded device, a single sync IO can take a long time * to complete after being issued. If this time exceeds our * monitoring window AND we didn't see any other completions in that * window, then count that sync IO as a violation of the latency. */ thislat = rwb_sync_issue_lat(rwb); if (thislat > rwb->cur_win_nsec || (thislat > rwb->min_lat_nsec && !stat[BLK_STAT_READ].nr_samples)) { trace_wbt_lat(bdi, thislat); return LAT_EXCEEDED; } /* * No read/write mix, if stat isn't valid */ if (!stat_sample_valid(stat)) { /* * If we had writes in this stat window and the window is * current, we're only doing writes. If a task recently * waited or still has writes in flights, consider us doing * just writes as well. */ if ((stat[BLK_STAT_WRITE].nr_samples && blk_stat_is_current(stat)) || wb_recent_wait(rwb) || wbt_inflight(rwb)) return LAT_UNKNOWN_WRITES; return LAT_UNKNOWN; } /* * If the 'min' latency exceeds our target, step down. */ if (stat[BLK_STAT_READ].min > rwb->min_lat_nsec) { trace_wbt_lat(bdi, stat[BLK_STAT_READ].min); trace_wbt_stat(bdi, stat); return LAT_EXCEEDED; } if (rwb->scale_step) trace_wbt_stat(bdi, stat); return LAT_OK; }

Contributors

PersonTokensPropCommitsCommitProp
jens axboejens axboe173100.00%4100.00%
Total173100.00%4100.00%


static int latency_exceeded(struct rq_wb *rwb) { struct blk_rq_stat stat[2]; blk_queue_stat_get(rwb->queue, stat); return __latency_exceeded(rwb, stat); }

Contributors

PersonTokensPropCommitsCommitProp
jens axboejens axboe35100.00%2100.00%
Total35100.00%2100.00%


static void rwb_trace_step(struct rq_wb *rwb, const char *msg) { struct backing_dev_info *bdi = &rwb->queue->backing_dev_info; trace_wbt_step(bdi, msg, rwb->scale_step, rwb->cur_win_nsec, rwb->wb_background, rwb->wb_normal, rwb->wb_max); }

Contributors

PersonTokensPropCommitsCommitProp
jens axboejens axboe55100.00%2100.00%
Total55100.00%2100.00%


static void scale_up(struct rq_wb *rwb) { /* * Hit max in previous round, stop here */ if (rwb->scaled_max) return; rwb->scale_step--; rwb->unknown_cnt = 0; blk_stat_clear(rwb->queue); rwb->scaled_max = calc_wb_limits(rwb); rwb_wake_all(rwb); rwb_trace_step(rwb, "step up"); }

Contributors

PersonTokensPropCommitsCommitProp
jens axboejens axboe58100.00%2100.00%
Total58100.00%2100.00%

/* * Scale rwb down. If 'hard_throttle' is set, do it quicker, since we * had a latency violation. */
static void scale_down(struct rq_wb *rwb, bool hard_throttle) { /* * Stop scaling down when we've hit the limit. This also prevents * ->scale_step from going to crazy values, if the device can't * keep up. */ if (rwb->wb_max == 1) return; if (rwb->scale_step < 0 && hard_throttle) rwb->scale_step = 0; else rwb->scale_step++; rwb->scaled_max = false; rwb->unknown_cnt = 0; blk_stat_clear(rwb->queue); calc_wb_limits(rwb); rwb_trace_step(rwb, "step down"); }

Contributors

PersonTokensPropCommitsCommitProp
jens axboejens axboe77100.00%2100.00%
Total77100.00%2100.00%


static void rwb_arm_timer(struct rq_wb *rwb) { unsigned long expires; if (rwb->scale_step > 0) { /* * We should speed this up, using some variant of a fast * integer inverse square root calculation. Since we only do * this for every window expiration, it's not a huge deal, * though. */ rwb->cur_win_nsec = div_u64(rwb->win_nsec << 4, int_sqrt((rwb->scale_step + 1) << 8)); } else { /* * For step < 0, we don't want to increase/decrease the * window size. */ rwb->cur_win_nsec = rwb->win_nsec; } expires = jiffies + nsecs_to_jiffies(rwb->cur_win_nsec); mod_timer(&rwb->window_timer, expires); }

Contributors

PersonTokensPropCommitsCommitProp
jens axboejens axboe85100.00%1100.00%
Total85100.00%1100.00%


static void wb_timer_fn(unsigned long data) { struct rq_wb *rwb = (struct rq_wb *) data; unsigned int inflight = wbt_inflight(rwb); int status; status = latency_exceeded(rwb); trace_wbt_timer(&rwb->queue->backing_dev_info, status, rwb->scale_step, inflight); /* * If we exceeded the latency target, step down. If we did not, * step one level up. If we don't know enough to say either exceeded * or ok, then don't do anything. */ switch (status) { case LAT_EXCEEDED: scale_down(rwb, true); break; case LAT_OK: scale_up(rwb); break; case LAT_UNKNOWN_WRITES: /* * We started a the center step, but don't have a valid * read/write sample, but we do have writes going on. * Allow step to go negative, to increase write perf. */ scale_up(rwb); break; case LAT_UNKNOWN: if (++rwb->unknown_cnt < RWB_UNKNOWN_BUMP) break; /* * We get here when previously scaled reduced depth, and we * currently don't have a valid read/write sample. For that * case, slowly return to center state (step == 0). */ if (rwb->scale_step > 0) scale_up(rwb); else if (rwb->scale_step < 0) scale_down(rwb, false); break; default: break; } /* * Re-arm timer, if we have IO in flight */ if (rwb->scale_step || inflight) rwb_arm_timer(rwb); }

Contributors

PersonTokensPropCommitsCommitProp
jens axboejens axboe156100.00%2100.00%
Total156100.00%2100.00%


void wbt_update_limits(struct rq_wb *rwb) { rwb->scale_step = 0; rwb->scaled_max = false; calc_wb_limits(rwb); rwb_wake_all(rwb); }

Contributors

PersonTokensPropCommitsCommitProp
jens axboejens axboe32100.00%1100.00%
Total32100.00%1100.00%


static bool close_io(struct rq_wb *rwb) { const unsigned long now = jiffies; return time_before(now, rwb->last_issue + HZ / 10) || time_before(now, rwb->last_comp + HZ / 10); }

Contributors

PersonTokensPropCommitsCommitProp
jens axboejens axboe45100.00%1100.00%
Total45100.00%1100.00%

#define REQ_HIPRIO (REQ_SYNC | REQ_META | REQ_PRIO)
static inline unsigned int get_limit(struct rq_wb *rwb, unsigned long rw) { unsigned int limit; /* * At this point we know it's a buffered write. If this is * kswapd trying to free memory, or REQ_SYNC is set, set, then * it's WB_SYNC_ALL writeback, and we'll use the max limit for * that. If the write is marked as a background write, then use * the idle limit, or go to normal if we haven't had competing * IO for a bit. */ if ((rw & REQ_HIPRIO) || wb_recent_wait(rwb) || current_is_kswapd()) limit = rwb->wb_max; else if ((rw & REQ_BACKGROUND) || close_io(rwb)) { /* * If less than 100ms since we completed unrelated IO, * limit us to half the depth for background writeback. */ limit = rwb->wb_background; } else limit = rwb->wb_normal; return limit; }

Contributors

PersonTokensPropCommitsCommitProp
jens axboejens axboe77100.00%1100.00%
Total77100.00%1100.00%


static inline bool may_queue(struct rq_wb *rwb, struct rq_wait *rqw, wait_queue_t *wait, unsigned long rw) { /* * inc it here even if disabled, since we'll dec it at completion. * this only happens if the task was sleeping in __wbt_wait(), * and someone turned it off at the same time. */ if (!rwb_enabled(rwb)) { atomic_inc(&rqw->inflight); return true; } /* * If the waitqueue is already active and we are not the next * in line to be woken up, wait for our turn. */ if (waitqueue_active(&rqw->wait) && rqw->wait.task_list.next != &wait->task_list) return false; return atomic_inc_below(&rqw->inflight, get_limit(rwb, rw)); }

Contributors

PersonTokensPropCommitsCommitProp
jens axboejens axboe90100.00%1100.00%
Total90100.00%1100.00%

/* * Block if we will exceed our limit, or if we are currently waiting for * the timer to kick off queuing again. */
static void __wbt_wait(struct rq_wb *rwb, unsigned long rw, spinlock_t *lock) __releases (lock) __acquires(lock) { struct rq_wait *rqw = get_rq_wait(rwb, current_is_kswapd()); DEFINE_WAIT(wait); if (may_queue(rwb, rqw, &wait, rw)) return; do { prepare_to_wait_exclusive(&rqw->wait, &wait, TASK_UNINTERRUPTIBLE); if (may_queue(rwb, rqw, &wait, rw)) break; if (lock) { spin_unlock_irq(lock); io_schedule(); spin_lock_irq(lock); } else io_schedule(); } while (1); finish_wait(&rqw->wait, &wait); }

Contributors

PersonTokensPropCommitsCommitProp
jens axboejens axboe11689.23%150.00%
bart van asschebart van assche1410.77%150.00%
Total130100.00%2100.00%


static inline bool wbt_should_throttle(struct rq_wb *rwb, struct bio *bio) { const int op = bio_op(bio); /* * If not a WRITE, do nothing */ if (op != REQ_OP_WRITE) return false; /* * Don't throttle WRITE_ODIRECT */ if ((bio->bi_opf & (REQ_SYNC | REQ_IDLE)) == (REQ_SYNC | REQ_IDLE)) return false; return true; }

Contributors

PersonTokensPropCommitsCommitProp
jens axboejens axboe6196.83%150.00%
christoph hellwigchristoph hellwig23.17%150.00%
Total63100.00%2100.00%

/* * Returns true if the IO request should be accounted, false if not. * May sleep, if we have exceeded the writeback limits. Caller can pass * in an irq held spinlock, if it holds one when calling this function. * If we do sleep, we'll release and re-grab it. */
enum wbt_flags wbt_wait(struct rq_wb *rwb, struct bio *bio, spinlock_t *lock) { unsigned int ret = 0; if (!rwb_enabled(rwb)) return 0; if (bio_op(bio) == REQ_OP_READ) ret = WBT_READ; if (!wbt_should_throttle(rwb, bio)) { if (ret & WBT_READ) wb_timestamp(rwb, &rwb->last_issue); return ret; } __wbt_wait(rwb, bio->bi_opf, lock); if (!timer_pending(&rwb->window_timer)) rwb_arm_timer(rwb); if (current_is_kswapd()) ret |= WBT_KSWAPD; return ret | WBT_TRACKED; }

Contributors

PersonTokensPropCommitsCommitProp
jens axboejens axboe12098.36%150.00%
bart van asschebart van assche21.64%150.00%
Total122100.00%2100.00%


void wbt_issue(struct rq_wb *rwb, struct blk_issue_stat *stat) { if (!rwb_enabled(rwb)) return; /* * Track sync issue, in case it takes a long time to complete. Allows * us to react quicker, if a sync IO takes a long time to complete. * Note that this is just a hint. 'stat' can go away when the * request completes, so it's important we never dereference it. We * only use the address to compare with, which is why we store the * sync_issue time locally. */ if (wbt_is_read(stat) && !rwb->sync_issue) { rwb->sync_cookie = stat; rwb->sync_issue = blk_stat_time(stat); } }

Contributors

PersonTokensPropCommitsCommitProp
jens axboejens axboe54100.00%1100.00%
Total54100.00%1100.00%


void wbt_requeue(struct rq_wb *rwb, struct blk_issue_stat *stat) { if (!rwb_enabled(rwb)) return; if (stat == rwb->sync_cookie) { rwb->sync_issue = 0; rwb->sync_cookie = NULL; } }

Contributors

PersonTokensPropCommitsCommitProp
jens axboejens axboe46100.00%1100.00%
Total46100.00%1100.00%


void wbt_set_queue_depth(struct rq_wb *rwb, unsigned int depth) { if (rwb) { rwb->queue_depth = depth; wbt_update_limits(rwb); } }

Contributors

PersonTokensPropCommitsCommitProp
jens axboejens axboe31100.00%1100.00%
Total31100.00%1100.00%


void wbt_set_write_cache(struct rq_wb *rwb, bool write_cache_on) { if (rwb) rwb->wc = write_cache_on; }

Contributors

PersonTokensPropCommitsCommitProp
jens axboejens axboe23100.00%1100.00%
Total23100.00%1100.00%

/* * Disable wbt, if enabled by default. Only called from CFQ, if we have * cgroups enabled */
void wbt_disable_default(struct request_queue *q) { struct rq_wb *rwb = q->rq_wb; if (rwb && rwb->enable_state == WBT_STATE_ON_DEFAULT) { del_timer_sync(&rwb->window_timer); rwb->win_nsec = rwb->min_lat_nsec = 0; wbt_update_limits(rwb); } }

Contributors

PersonTokensPropCommitsCommitProp
jens axboejens axboe54100.00%3100.00%
Total54100.00%3100.00%

EXPORT_SYMBOL_GPL(wbt_disable_default);
u64 wbt_default_latency_nsec(struct request_queue *q) { /* * We default to 2msec for non-rotational storage, and 75msec * for rotational storage. */ if (blk_queue_nonrot(q)) return 2000000ULL; else return 75000000ULL; }

Contributors

PersonTokensPropCommitsCommitProp
jens axboejens axboe25100.00%1100.00%
Total25100.00%1100.00%


int wbt_init(struct request_queue *q) { struct rq_wb *rwb; int i; /* * For now, we depend on the stats window being larger than * our monitoring window. Ensure that this isn't inadvertently * violated. */ BUILD_BUG_ON(RWB_WINDOW_NSEC > BLK_STAT_NSEC); BUILD_BUG_ON(WBT_NR_BITS > BLK_STAT_RES_BITS); rwb = kzalloc(sizeof(*rwb), GFP_KERNEL); if (!rwb) return -ENOMEM; for (i = 0; i < WBT_NUM_RWQ; i++) { atomic_set(&rwb->rq_wait[i].inflight, 0); init_waitqueue_head(&rwb->rq_wait[i].wait); } setup_timer(&rwb->window_timer, wb_timer_fn, (unsigned long) rwb); rwb->wc = 1; rwb->queue_depth = RWB_DEF_DEPTH; rwb->last_comp = rwb->last_issue = jiffies; rwb->queue = q; rwb->win_nsec = RWB_WINDOW_NSEC; rwb->enable_state = WBT_STATE_ON_DEFAULT; wbt_update_limits(rwb); /* * Assign rwb, and turn on stats tracking for this queue */ q->rq_wb = rwb; blk_stat_enable(q); rwb->min_lat_nsec = wbt_default_latency_nsec(q); wbt_set_queue_depth(rwb, blk_queue_depth(q)); wbt_set_write_cache(rwb, test_bit(QUEUE_FLAG_WC, &q->queue_flags)); return 0; }

Contributors

PersonTokensPropCommitsCommitProp
jens axboejens axboe208100.00%4100.00%
Total208100.00%4100.00%


void wbt_exit(struct request_queue *q) { struct rq_wb *rwb = q->rq_wb; if (rwb) { del_timer_sync(&rwb->window_timer); q->rq_wb = NULL; kfree(rwb); } }

Contributors

PersonTokensPropCommitsCommitProp
jens axboejens axboe44100.00%1100.00%
Total44100.00%1100.00%


Overall Contributors

PersonTokensPropCommitsCommitProp
jens axboejens axboe256999.27%763.64%
bart van asschebart van assche160.62%218.18%
christoph hellwigchristoph hellwig20.08%19.09%
arnd bergmannarnd bergmann10.04%19.09%
Total2588100.00%11100.00%
Directory: block
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.