Author | Tokens | Token Proportion | Commits | Commit Proportion |
---|---|---|---|---|
Steven Rostedt | 592 | 97.69% | 6 | 60.00% |
Wei Yongjun | 7 | 1.16% | 1 | 10.00% |
Chunyan Zhang | 4 | 0.66% | 1 | 10.00% |
Paul E. McKenney | 2 | 0.33% | 1 | 10.00% |
Greg Kroah-Hartman | 1 | 0.17% | 1 | 10.00% |
Total | 606 | 10 |
// SPDX-License-Identifier: GPL-2.0 #include <linux/delay.h> #include <linux/module.h> #include <linux/kthread.h> #include <linux/trace_clock.h> #define CREATE_TRACE_POINTS #include "trace_benchmark.h" static struct task_struct *bm_event_thread; static char bm_str[BENCHMARK_EVENT_STRLEN] = "START"; static u64 bm_total; static u64 bm_totalsq; static u64 bm_last; static u64 bm_max; static u64 bm_min; static u64 bm_first; static u64 bm_cnt; static u64 bm_stddev; static unsigned int bm_avg; static unsigned int bm_std; static bool ok_to_run; /* * This gets called in a loop recording the time it took to write * the tracepoint. What it writes is the time statistics of the last * tracepoint write. As there is nothing to write the first time * it simply writes "START". As the first write is cold cache and * the rest is hot, we save off that time in bm_first and it is * reported as "first", which is shown in the second write to the * tracepoint. The "first" field is writen within the statics from * then on but never changes. */ static void trace_do_benchmark(void) { u64 start; u64 stop; u64 delta; u64 stddev; u64 seed; u64 last_seed; unsigned int avg; unsigned int std = 0; /* Only run if the tracepoint is actually active */ if (!trace_benchmark_event_enabled() || !tracing_is_on()) return; local_irq_disable(); start = trace_clock_local(); trace_benchmark_event(bm_str); stop = trace_clock_local(); local_irq_enable(); bm_cnt++; delta = stop - start; /* * The first read is cold cached, keep it separate from the * other calculations. */ if (bm_cnt == 1) { bm_first = delta; scnprintf(bm_str, BENCHMARK_EVENT_STRLEN, "first=%llu [COLD CACHED]", bm_first); return; } bm_last = delta; if (delta > bm_max) bm_max = delta; if (!bm_min || delta < bm_min) bm_min = delta; /* * When bm_cnt is greater than UINT_MAX, it breaks the statistics * accounting. Freeze the statistics when that happens. * We should have enough data for the avg and stddev anyway. */ if (bm_cnt > UINT_MAX) { scnprintf(bm_str, BENCHMARK_EVENT_STRLEN, "last=%llu first=%llu max=%llu min=%llu ** avg=%u std=%d std^2=%lld", bm_last, bm_first, bm_max, bm_min, bm_avg, bm_std, bm_stddev); return; } bm_total += delta; bm_totalsq += delta * delta; if (bm_cnt > 1) { /* * Apply Welford's method to calculate standard deviation: * s^2 = 1 / (n * (n-1)) * (n * \Sum (x_i)^2 - (\Sum x_i)^2) */ stddev = (u64)bm_cnt * bm_totalsq - bm_total * bm_total; do_div(stddev, (u32)bm_cnt); do_div(stddev, (u32)bm_cnt - 1); } else stddev = 0; delta = bm_total; do_div(delta, bm_cnt); avg = delta; if (stddev > 0) { int i = 0; /* * stddev is the square of standard deviation but * we want the actualy number. Use the average * as our seed to find the std. * * The next try is: * x = (x + N/x) / 2 * * Where N is the squared number to find the square * root of. */ seed = avg; do { last_seed = seed; seed = stddev; if (!last_seed) break; do_div(seed, last_seed); seed += last_seed; do_div(seed, 2); } while (i++ < 10 && last_seed != seed); std = seed; } scnprintf(bm_str, BENCHMARK_EVENT_STRLEN, "last=%llu first=%llu max=%llu min=%llu avg=%u std=%d std^2=%lld", bm_last, bm_first, bm_max, bm_min, avg, std, stddev); bm_std = std; bm_avg = avg; bm_stddev = stddev; } static int benchmark_event_kthread(void *arg) { /* sleep a bit to make sure the tracepoint gets activated */ msleep(100); while (!kthread_should_stop()) { trace_do_benchmark(); /* * We don't go to sleep, but let others run as well. * This is bascially a "yield()" to let any task that * wants to run, schedule in, but if the CPU is idle, * we'll keep burning cycles. * * Note the tasks_rcu_qs() version of cond_resched() will * notify synchronize_rcu_tasks() that this thread has * passed a quiescent state for rcu_tasks. Otherwise * this thread will never voluntarily schedule which would * block synchronize_rcu_tasks() indefinitely. */ cond_resched_tasks_rcu_qs(); } return 0; } /* * When the benchmark tracepoint is enabled, it calls this * function and the thread that calls the tracepoint is created. */ int trace_benchmark_reg(void) { if (!ok_to_run) { pr_warning("trace benchmark cannot be started via kernel command line\n"); return -EBUSY; } bm_event_thread = kthread_run(benchmark_event_kthread, NULL, "event_benchmark"); if (IS_ERR(bm_event_thread)) { pr_warning("trace benchmark failed to create kernel thread\n"); return PTR_ERR(bm_event_thread); } return 0; } /* * When the benchmark tracepoint is disabled, it calls this * function and the thread that calls the tracepoint is deleted * and all the numbers are reset. */ void trace_benchmark_unreg(void) { if (!bm_event_thread) return; kthread_stop(bm_event_thread); bm_event_thread = NULL; strcpy(bm_str, "START"); bm_total = 0; bm_totalsq = 0; bm_last = 0; bm_max = 0; bm_min = 0; bm_cnt = 0; /* These don't need to be reset but reset them anyway */ bm_first = 0; bm_std = 0; bm_avg = 0; bm_stddev = 0; } static __init int ok_to_run_trace_benchmark(void) { ok_to_run = true; return 0; } early_initcall(ok_to_run_trace_benchmark);
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.
Created with Cregit http://github.com/cregit/cregit
Version 2.0-RC1