Release 4.15 kernel/trace/trace_output.c
/*
* trace_output.c
*
* Copyright (C) 2008 Red Hat Inc, Steven Rostedt <srostedt@redhat.com>
*
*/
#include <linux/module.h>
#include <linux/mutex.h>
#include <linux/ftrace.h>
#include <linux/sched/clock.h>
#include <linux/sched/mm.h>
#include "trace_output.h"
/* must be a power of 2 */
#define EVENT_HASHSIZE 128
DECLARE_RWSEM(trace_event_sem);
static struct hlist_head event_hash[EVENT_HASHSIZE] __read_mostly;
static int next_event_type = __TRACE_LAST_TYPE + 1;
enum print_line_t trace_print_bputs_msg_only(struct trace_iterator *iter)
{
struct trace_seq *s = &iter->seq;
struct trace_entry *entry = iter->ent;
struct bputs_entry *field;
trace_assign_type(field, entry);
trace_seq_puts(s, field->str);
return trace_handle_return(s);
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Steven Rostedt | 55 | 96.49% | 2 | 66.67% |
Eduard - Gabriel Munteanu | 2 | 3.51% | 1 | 33.33% |
Total | 57 | 100.00% | 3 | 100.00% |
enum print_line_t trace_print_bprintk_msg_only(struct trace_iterator *iter)
{
struct trace_seq *s = &iter->seq;
struct trace_entry *entry = iter->ent;
struct bprint_entry *field;
trace_assign_type(field, entry);
trace_seq_bprintf(s, field->fmt, field->buf);
return trace_handle_return(s);
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Steven Rostedt | 57 | 93.44% | 3 | 75.00% |
Johannes Berg | 4 | 6.56% | 1 | 25.00% |
Total | 61 | 100.00% | 4 | 100.00% |
enum print_line_t trace_print_printk_msg_only(struct trace_iterator *iter)
{
struct trace_seq *s = &iter->seq;
struct trace_entry *entry = iter->ent;
struct print_entry *field;
trace_assign_type(field, entry);
trace_seq_puts(s, field->buf);
return trace_handle_return(s);
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Steven Rostedt | 57 | 100.00% | 3 | 100.00% |
Total | 57 | 100.00% | 3 | 100.00% |
const char *
trace_print_flags_seq(struct trace_seq *p, const char *delim,
unsigned long flags,
const struct trace_print_flags *flag_array)
{
unsigned long mask;
const char *str;
const char *ret = trace_seq_buffer_ptr(p);
int i, first = 1;
for (i = 0; flag_array[i].name && flags; i++) {
mask = flag_array[i].mask;
if ((flags & mask) != mask)
continue;
str = flag_array[i].name;
flags &= ~mask;
if (!first && delim)
trace_seq_puts(p, delim);
else
first = 0;
trace_seq_puts(p, str);
}
/* check for left over flags */
if (flags) {
if (!first && delim)
trace_seq_puts(p, delim);
trace_seq_printf(p, "0x%lx", flags);
}
trace_seq_putc(p, 0);
return ret;
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Steven Rostedt | 154 | 89.02% | 4 | 66.67% |
Andrey Vagin | 11 | 6.36% | 1 | 16.67% |
Steven Whitehouse | 8 | 4.62% | 1 | 16.67% |
Total | 173 | 100.00% | 6 | 100.00% |
EXPORT_SYMBOL(trace_print_flags_seq);
const char *
trace_print_symbols_seq(struct trace_seq *p, unsigned long val,
const struct trace_print_flags *symbol_array)
{
int i;
const char *ret = trace_seq_buffer_ptr(p);
for (i = 0; symbol_array[i].name; i++) {
if (val != symbol_array[i].mask)
continue;
trace_seq_puts(p, symbol_array[i].name);
break;
}
if (ret == (const char *)(trace_seq_buffer_ptr(p)))
trace_seq_printf(p, "0x%lx", val);
trace_seq_putc(p, 0);
return ret;
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Steven Rostedt | 106 | 93.81% | 4 | 80.00% |
Steven Whitehouse | 7 | 6.19% | 1 | 20.00% |
Total | 113 | 100.00% | 5 | 100.00% |
EXPORT_SYMBOL(trace_print_symbols_seq);
#if BITS_PER_LONG == 32
const char *
trace_print_flags_seq_u64(struct trace_seq *p, const char *delim,
unsigned long long flags,
const struct trace_print_flags_u64 *flag_array)
{
unsigned long long mask;
const char *str;
const char *ret = trace_seq_buffer_ptr(p);
int i, first = 1;
for (i = 0; flag_array[i].name && flags; i++) {
mask = flag_array[i].mask;
if ((flags & mask) != mask)
continue;
str = flag_array[i].name;
flags &= ~mask;
if (!first && delim)
trace_seq_puts(p, delim);
else
first = 0;
trace_seq_puts(p, str);
}
/* check for left over flags */
if (flags) {
if (!first && delim)
trace_seq_puts(p, delim);
trace_seq_printf(p, "0x%llx", flags);
}
trace_seq_putc(p, 0);
return ret;
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Ross Zwisler | 172 | 98.29% | 1 | 50.00% |
Liu Bo | 3 | 1.71% | 1 | 50.00% |
Total | 175 | 100.00% | 2 | 100.00% |
EXPORT_SYMBOL(trace_print_flags_seq_u64);
const char *
trace_print_symbols_seq_u64(struct trace_seq *p, unsigned long long val,
const struct trace_print_flags_u64 *symbol_array)
{
int i;
const char *ret = trace_seq_buffer_ptr(p);
for (i = 0; symbol_array[i].name; i++) {
if (val != symbol_array[i].mask)
continue;
trace_seq_puts(p, symbol_array[i].name);
break;
}
if (ret == (const char *)(trace_seq_buffer_ptr(p)))
trace_seq_printf(p, "0x%llx", val);
trace_seq_putc(p, 0);
return ret;
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Liu Bo | 94 | 82.46% | 1 | 20.00% |
Steven Rostedt | 17 | 14.91% | 3 | 60.00% |
Ross Zwisler | 3 | 2.63% | 1 | 20.00% |
Total | 114 | 100.00% | 5 | 100.00% |
EXPORT_SYMBOL(trace_print_symbols_seq_u64);
#endif
const char *
trace_print_bitmask_seq(struct trace_seq *p, void *bitmask_ptr,
unsigned int bitmask_size)
{
const char *ret = trace_seq_buffer_ptr(p);
trace_seq_bitmask(p, bitmask_ptr, bitmask_size * 8);
trace_seq_putc(p, 0);
return ret;
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Steven Rostedt | 48 | 94.12% | 3 | 75.00% |
Keiichiro Tokunaga | 3 | 5.88% | 1 | 25.00% |
Total | 51 | 100.00% | 4 | 100.00% |
EXPORT_SYMBOL_GPL(trace_print_bitmask_seq);
/**
* trace_print_hex_seq - print buffer as hex sequence
* @p: trace seq struct to write to
* @buf: The buffer to print
* @buf_len: Length of @buf in bytes
* @concatenate: Print @buf as single hex string or with spacing
*
* Prints the passed buffer as a hex sequence either as a whole,
* single hex string if @concatenate is true or with spacing after
* each byte in case @concatenate is false.
*/
const char *
trace_print_hex_seq(struct trace_seq *p, const unsigned char *buf, int buf_len,
bool concatenate)
{
int i;
const char *ret = trace_seq_buffer_ptr(p);
for (i = 0; i < buf_len; i++)
trace_seq_printf(p, "%s%2.2x", concatenate || i == 0 ? "" : " ",
buf[i]);
trace_seq_putc(p, 0);
return ret;
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Keiichiro Tokunaga | 70 | 85.37% | 1 | 16.67% |
Steven Rostedt | 7 | 8.54% | 3 | 50.00% |
Daniel Borkmann | 5 | 6.10% | 2 | 33.33% |
Total | 82 | 100.00% | 6 | 100.00% |
EXPORT_SYMBOL(trace_print_hex_seq);
const char *
trace_print_array_seq(struct trace_seq *p, const void *buf, int count,
size_t el_size)
{
const char *ret = trace_seq_buffer_ptr(p);
const char *prefix = "";
void *ptr = (void *)buf;
size_t buf_len = count * el_size;
trace_seq_putc(p, '{');
while (ptr < buf + buf_len) {
switch (el_size) {
case 1:
trace_seq_printf(p, "%s0x%x", prefix,
*(u8 *)ptr);
break;
case 2:
trace_seq_printf(p, "%s0x%x", prefix,
*(u16 *)ptr);
break;
case 4:
trace_seq_printf(p, "%s0x%x", prefix,
*(u32 *)ptr);
break;
case 8:
trace_seq_printf(p, "%s0x%llx", prefix,
*(u64 *)ptr);
break;
default:
trace_seq_printf(p, "BAD SIZE:%zu 0x%x", el_size,
*(u8 *)ptr);
el_size = 1;
}
prefix = ",";
ptr += el_size;
}
trace_seq_putc(p, '}');
trace_seq_putc(p, 0);
return ret;
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Dave P Martin | 197 | 95.63% | 1 | 33.33% |
Alex Bennée | 8 | 3.88% | 1 | 33.33% |
Steven Rostedt | 1 | 0.49% | 1 | 33.33% |
Total | 206 | 100.00% | 3 | 100.00% |
EXPORT_SYMBOL(trace_print_array_seq);
int trace_raw_output_prep(struct trace_iterator *iter,
struct trace_event *trace_event)
{
struct trace_event_call *event;
struct trace_seq *s = &iter->seq;
struct trace_seq *p = &iter->tmp_seq;
struct trace_entry *entry;
event = container_of(trace_event, struct trace_event_call, event);
entry = iter->ent;
if (entry->type != event->event.type) {
WARN_ON_ONCE(1);
return TRACE_TYPE_UNHANDLED;
}
trace_seq_init(p);
trace_seq_printf(s, "%s: ", trace_event_name(event));
return trace_handle_return(s);
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Li Zefan | 97 | 89.81% | 1 | 14.29% |
Steven Rostedt | 9 | 8.33% | 5 | 71.43% |
Mathieu Desnoyers | 2 | 1.85% | 1 | 14.29% |
Total | 108 | 100.00% | 7 | 100.00% |
EXPORT_SYMBOL(trace_raw_output_prep);
static int trace_output_raw(struct trace_iterator *iter, char *name,
char *fmt, va_list ap)
{
struct trace_seq *s = &iter->seq;
trace_seq_printf(s, "%s: ", name);
trace_seq_vprintf(s, fmt, ap);
return trace_handle_return(s);
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Steven Rostedt | 56 | 100.00% | 3 | 100.00% |
Total | 56 | 100.00% | 3 | 100.00% |
int trace_output_call(struct trace_iterator *iter, char *name, char *fmt, ...)
{
va_list ap;
int ret;
va_start(ap, fmt);
ret = trace_output_raw(iter, name, fmt, ap);
va_end(ap);
return ret;
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Steven Rostedt | 54 | 100.00% | 2 | 100.00% |
Total | 54 | 100.00% | 2 | 100.00% |
EXPORT_SYMBOL_GPL(trace_output_call);
#ifdef CONFIG_KRETPROBES
static inline const char *kretprobed(const char *name)
{
static const char tramp_name[] = "kretprobe_trampoline";
int size = sizeof(tramp_name);
if (strncmp(tramp_name, name, size) == 0)
return "[unknown/kretprobe'd]";
return name;
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Steven Rostedt | 49 | 100.00% | 1 | 100.00% |
Total | 49 | 100.00% | 1 | 100.00% |
#else
static inline const char *kretprobed(const char *name)
{
return name;
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Steven Rostedt | 17 | 100.00% | 1 | 100.00% |
Total | 17 | 100.00% | 1 | 100.00% |
#endif /* CONFIG_KRETPROBES */
static void
seq_print_sym_short(struct trace_seq *s, const char *fmt, unsigned long address)
{
char str[KSYM_SYMBOL_LEN];
#ifdef CONFIG_KALLSYMS
const char *name;
kallsyms_lookup(address, NULL, NULL, NULL, str);
name = kretprobed(str);
if (name && strlen(name)) {
trace_seq_printf(s, fmt, name);
return;
}
#endif
snprintf(str, KSYM_SYMBOL_LEN, "0x%08lx", address);
trace_seq_printf(s, fmt, str);
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Steven Rostedt | 97 | 100.00% | 3 | 100.00% |
Total | 97 | 100.00% | 3 | 100.00% |
static void
seq_print_sym_offset(struct trace_seq *s, const char *fmt,
unsigned long address)
{
char str[KSYM_SYMBOL_LEN];
#ifdef CONFIG_KALLSYMS
const char *name;
sprint_symbol(str, address);
name = kretprobed(str);
if (name && strlen(name)) {
trace_seq_printf(s, fmt, name);
return;
}
#endif
snprintf(str, KSYM_SYMBOL_LEN, "0x%08lx", address);
trace_seq_printf(s, fmt, str);
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Steven Rostedt | 91 | 100.00% | 3 | 100.00% |
Total | 91 | 100.00% | 3 | 100.00% |
#ifndef CONFIG_64BIT
# define IP_FMT "%08lx"
#else
# define IP_FMT "%016lx"
#endif
static int seq_print_user_ip(struct trace_seq *s, struct mm_struct *mm,
unsigned long ip, unsigned long sym_flags)
{
struct file *file = NULL;
unsigned long vmstart = 0;
int ret = 1;
if (s->full)
return 0;
if (mm) {
const struct vm_area_struct *vma;
down_read(&mm->mmap_sem);
vma = find_vma(mm, ip);
if (vma) {
file = vma->vm_file;
vmstart = vma->vm_start;
}
if (file) {
ret = trace_seq_path(s, &file->f_path);
if (ret)
trace_seq_printf(s, "[+0x%lx]",
ip - vmstart);
}
up_read(&mm->mmap_sem);
}
if (ret && ((sym_flags & TRACE_ITER_SYM_ADDR) || !file))
trace_seq_printf(s, " <" IP_FMT ">", ip);
return !trace_seq_has_overflowed(s);
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Steven Rostedt | 163 | 94.77% | 3 | 75.00% |
Johannes Berg | 9 | 5.23% | 1 | 25.00% |
Total | 172 | 100.00% | 4 | 100.00% |
int
seq_print_ip_sym(struct trace_seq *s, unsigned long ip, unsigned long sym_flags)
{
if (!ip) {
trace_seq_putc(s, '0');
goto out;
}
if (sym_flags & TRACE_ITER_SYM_OFFSET)
seq_print_sym_offset(s, "%s", ip);
else
seq_print_sym_short(s, "%s", ip);
if (sym_flags & TRACE_ITER_SYM_ADDR)
trace_seq_printf(s, " <" IP_FMT ">", ip);
out:
return !trace_seq_has_overflowed(s);
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Steven Rostedt | 84 | 97.67% | 2 | 66.67% |
Jovi Zhangwei | 2 | 2.33% | 1 | 33.33% |
Total | 86 | 100.00% | 3 | 100.00% |
/**
* trace_print_lat_fmt - print the irq, preempt and lockdep fields
* @s: trace seq struct to write to
* @entry: The trace entry field from the ring buffer
*
* Prints the generic fields of irqs off, in hard or softirq, preempt
* count.
*/
int trace_print_lat_fmt(struct trace_seq *s, struct trace_entry *entry)
{
char hardsoft_irq;
char need_resched;
char irqs_off;
int hardirq;
int softirq;
int nmi;
nmi = entry->flags & TRACE_FLAG_NMI;
hardirq = entry->flags & TRACE_FLAG_HARDIRQ;
softirq = entry->flags & TRACE_FLAG_SOFTIRQ;
irqs_off =
(entry->flags & TRACE_FLAG_IRQS_OFF) ? 'd' :
(entry->flags & TRACE_FLAG_IRQS_NOSUPPORT) ? 'X' :
'.';
switch (entry->flags & (TRACE_FLAG_NEED_RESCHED |
TRACE_FLAG_PREEMPT_RESCHED)) {
case TRACE_FLAG_NEED_RESCHED | TRACE_FLAG_PREEMPT_RESCHED:
need_resched = 'N';
break;
case TRACE_FLAG_NEED_RESCHED:
need_resched = 'n';
break;
case TRACE_FLAG_PREEMPT_RESCHED:
need_resched = 'p';
break;
default:
need_resched = '.';
break;
}
hardsoft_irq =
(nmi && hardirq) ? 'Z' :
nmi ? 'z' :
(hardirq && softirq) ? 'H' :
hardirq ? 'h' :
softirq ? 's' :
'.' ;
trace_seq_printf(s, "%c%c%c",
irqs_off, need_resched, hardsoft_irq);
if (entry->preempt_count)
trace_seq_printf(s, "%x", entry->preempt_count);
else
trace_seq_putc(s, '.');
return !trace_seq_has_overflowed(s);
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Frédéric Weisbecker | 60 | 29.41% | 1 | 11.11% |
Peter Zijlstra | 60 | 29.41% | 2 | 22.22% |
David Sharp | 29 | 14.22% | 1 | 11.11% |
Steven Rostedt | 28 | 13.73% | 4 | 44.44% |
Arnaldo Carvalho de Melo | 27 | 13.24% | 1 | 11.11% |
Total | 204 | 100.00% | 9 | 100.00% |
static int
lat_print_generic(struct trace_seq *s, struct trace_entry *entry, int cpu)
{
char comm[TASK_COMM_LEN];
trace_find_cmdline(entry->pid, comm);
trace_seq_printf(s, "%8.8s-%-5d %3d",
comm, entry->pid, cpu);
return trace_print_lat_fmt(s, entry);
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Steven Rostedt | 57 | 100.00% | 2 | 100.00% |
Total | 57 | 100.00% | 2 | 100.00% |
#undef MARK
#define MARK(v, s) {.val = v, .sym = s}
/* trace overhead mark */
static const struct trace_mark {
unsigned long long val; /* unit: nsec */
char sym;
} mark[] = {
MARK(1000000000ULL , '$'), /* 1 sec */
MARK(100000000ULL , '@'), /* 100 msec */
MARK(10000000ULL , '*'), /* 10 msec */
MARK(1000000ULL , '#'), /* 1000 usecs */
MARK(100000ULL , '!'), /* 100 usecs */
MARK(10000ULL , '+'), /* 10 usecs */
};
#undef MARK
char trace_find_mark(unsigned long long d)
{
int i;
int size = ARRAY_SIZE(mark);
for (i = 0; i < size; i++) {
if (d > mark[i].val)
break;
}
return (i == size) ? ' ' : mark[i].sym;
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Byungchul Park | 62 | 96.88% | 1 | 33.33% |
Frédéric Weisbecker | 1 | 1.56% | 1 | 33.33% |
Jungseok Lee | 1 | 1.56% | 1 | 33.33% |
Total | 64 | 100.00% | 3 | 100.00% |
static int
lat_print_timestamp(struct trace_iterator *iter, u64 next_ts)
{
struct trace_array *tr = iter->tr;
unsigned long verbose = tr->trace_flags & TRACE_ITER_VERBOSE;
unsigned long in_ns = iter->iter_flags & TRACE_FILE_TIME_IN_NS;
unsigned long long abs_ts = iter->ts - iter->trace_buffer->time_start;
unsigned long long rel_ts = next_ts - iter->ts;
struct trace_seq *s = &iter->seq;
if (in_ns) {
abs_ts = ns2usecs(abs_ts);
rel_ts = ns2usecs(rel_ts);
}
if (verbose && in_ns) {
unsigned long abs_usec = do_div(abs_ts, USEC_PER_MSEC);
unsigned long abs_msec = (unsigned long)abs_ts;
unsigned long rel_usec = do_div(rel_ts, USEC_PER_MSEC);
unsigned long rel_msec = (unsigned long)rel_ts;
trace_seq_printf(
s, "[%08llx] %ld.%03ldms (+%ld.%03ldms): ",
ns2usecs(iter->ts),
abs_msec, abs_usec,
rel_msec, rel_usec);
} else if (verbose && !in_ns) {
trace_seq_printf(
s, "[%016llx] %lld (+%lld): ",
iter->ts, abs_ts, rel_ts);
} else if (!verbose && in_ns) {
trace_seq_printf(
s, " %4lldus%c: ",
abs_ts,
trace_find_mark(rel_ts * NSEC_PER_USEC));
} else { /* !verbose && !in_ns */
trace_seq_printf(s, " %4lld: ", abs_ts);
}
return !trace_seq_has_overflowed(s);
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
David Sharp | 194 | 80.17% | 1 | 14.29% |
Steven Rostedt | 23 | 9.50% | 3 | 42.86% |
Frédéric Weisbecker | 17 | 7.02% | 1 | 14.29% |
Byungchul Park | 5 | 2.07% | 1 | 14.29% |
Arnaldo Carvalho de Melo | 3 | 1.24% | 1 | 14.29% |
Total | 242 | 100.00% | 7 | 100.00% |
int trace_print_context(struct trace_iterator *iter)
{
struct trace_array *tr = iter->tr;
struct trace_seq *s = &iter->seq;
struct trace_entry *entry = iter->ent;
unsigned long long t;
unsigned long secs, usec_rem;
char comm[TASK_COMM_LEN];
trace_find_cmdline(entry->pid, comm);
trace_seq_printf(s, "%16s-%-5d [%03d] ",
comm, entry->pid, iter->cpu);
if (tr->trace_flags & TRACE_ITER_RECORD_TGID) {
unsigned int tgid = trace_find_tgid(entry->pid);
if (!tgid)
trace_seq_printf(s, "(-----) ");
else
trace_seq_printf(s, "(%5d) ", tgid);
}
if (tr->trace_flags & TRACE_ITER_IRQ_INFO)
trace_print_lat_fmt(s, entry);
if (iter->iter_flags & TRACE_FILE_TIME_IN_NS) {
t = ns2usecs(iter->ts);
usec_rem = do_div(t, USEC_PER_SEC);
secs = (unsigned long)t;
trace_seq_printf(s, " %5lu.%06lu: ", secs, usec_rem);
} else
trace_seq_printf(s, " %12llu: ", iter->ts);
return !trace_seq_has_overflowed(s);
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Steven Rostedt | 58 | 28.29% | 5 | 62.50% |
Frédéric Weisbecker | 56 | 27.32% | 1 | 12.50% |
David Sharp | 48 | 23.41% | 1 | 12.50% |
Joel Fernandes | 43 | 20.98% | 1 | 12.50% |
Total | 205 | 100.00% | 8 | 100.00% |
int trace_print_lat_context(struct trace_iterator *iter)
{
struct trace_array *tr = iter->tr;
/* trace_find_next_entry will reset ent_size */
int ent_size = iter->ent_size;
struct trace_seq *s = &iter->seq;
u64 next_ts;
struct trace_entry *entry = iter->ent,
*next_entry = trace_find_next_entry(iter, NULL,
&next_ts);
unsigned long verbose = (tr->trace_flags & TRACE_ITER_VERBOSE);
/* Restore the original ent_size */
iter->ent_size = ent_size;
if (!next_entry)
next_ts = iter->ts;
if (verbose) {
char comm[TASK_COMM_LEN];
trace_find_cmdline(entry->pid, comm);
trace_seq_printf(
s, "%16s %5d %3d %d %08x %08lx ",
comm, entry->pid, iter->cpu, entry->flags,
entry->preempt_count, iter->idx);
} else {
lat_print_generic(s, entry, iter->cpu);
}
lat_print_timestamp(iter, next_ts);
return !trace_seq_has_overflowed(s);
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Frédéric Weisbecker | 124 | 72.94% | 1 | 14.29% |
Steven Rostedt | 42 | 24.71% | 5 | 71.43% |
David Sharp | 4 | 2.35% | 1 | 14.29% |
Total | 170 | 100.00% | 7 | 100.00% |
/**
* ftrace_find_event - find a registered event
* @type: the type of event to look for
*
* Returns an event of type @type otherwise NULL
* Called with trace_event_read_lock() held.
*/
struct trace_event *ftrace_find_event(int type)
{
struct trace_event *event;
unsigned key;
key = type & (EVENT_HASHSIZE - 1);
hlist_for_each_entry(event, &event_hash[key], node) {
if (event->type == type)
return event;
}
return NULL;
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Steven Rostedt | 51 | 98.08% | 1 | 50.00% |
Lai Jiangshan | 1 | 1.92% | 1 | 50.00% |
Total | 52 | 100.00% | 2 | 100.00% |
static LIST_HEAD(ftrace_event_list);
static int trace_search_list(struct list_head **list)
{
struct trace_event *e;
int last = __TRACE_LAST_TYPE;
if (list_empty(&ftrace_event_list)) {
*list = &ftrace_event_list;
return last + 1;
}
/*
* We used up all possible max events,
* lets see if somebody freed one.
*/
list_for_each_entry(e, &ftrace_event_list, list) {
if (e->type != last + 1)
break;
last++;
}
/* Did we used up all 65 thousand events??? */
if ((last + 1) > TRACE_EVENT_TYPE_MAX)
return 0;
*list = &e->list;
return last + 1;
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Steven Rostedt | 95 | 100.00% | 2 | 100.00% |
Total | 95 | 100.00% | 2 | 100.00% |
void trace_event_read_lock(void)
{
down_read(&trace_event_sem);
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Lai Jiangshan | 12 | 92.31% | 1 | 50.00% |
Jovi Zhangwei | 1 | 7.69% | 1 | 50.00% |
Total | 13 | 100.00% | 2 | 100.00% |
void trace_event_read_unlock(void)
{
up_read(&trace_event_sem);
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Lai Jiangshan | 12 | 92.31% | 1 | 50.00% |
Jovi Zhangwei | 1 | 7.69% | 1 | 50.00% |
Total | 13 | 100.00% | 2 | 100.00% |
/**
* register_trace_event - register output for an event type
* @event: the event type to register
*
* Event types are stored in a hash and this hash is used to
* find a way to print an event. If the @event->type is set
* then it will use that type, otherwise it will assign a
* type to use.
*
* If you assign your own type, please make sure it is added
* to the trace_type enum in trace.h, to avoid collisions
* with the dynamic types.
*
* Returns the event type number or zero on error.
*/
int register_trace_event(struct trace_event *event)
{
unsigned key;
int ret = 0;
down_write(&trace_event_sem);
if (WARN_ON(!event))
goto out;
if (WARN_ON(!event->funcs))
goto out;
INIT_LIST_HEAD(&event->list);
if (!event->type) {
struct list_head *list = NULL;
if (next_event_type > TRACE_EVENT_TYPE_MAX) {
event->type = trace_search_list(&list);
if (!event->type)
goto out;
} else {
event->type = next_event_type++;
list = &ftrace_event_list;
}
if (WARN_ON(ftrace_find_event(event->type)))
goto out;
list_add_tail(&event->list, list);
} else if (event->type > __TRACE_LAST_TYPE) {
printk(KERN_WARNING "Need to add type to trace.h\n");
WARN_ON(1);
goto out;
} else {
/* Is this event already used */
if (ftrace_find_event(event->type))
goto out;
}
if (event->funcs->trace == NULL)
event->funcs->trace = trace_nop_print;
if (event->funcs->raw == NULL)
event->funcs->raw = trace_nop_print;
if (event->funcs->hex == NULL)
event->funcs->hex = trace_nop_print;
if (event->funcs->binary == NULL)
event->funcs->binary = trace_nop_print;
key = event->type & (EVENT_HASHSIZE - 1);
hlist_add_head(&event->node, &event_hash[key]);
ret = event->type;
out:
up_write(&trace_event_sem);
return ret;
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Steven Rostedt | 225 | 76.01% | 5 | 50.00% |
Arnaldo Carvalho de Melo | 56 | 18.92% | 1 | 10.00% |
Peter Zijlstra | 8 | 2.70% | 1 | 10.00% |
Lai Jiangshan | 3 | 1.01% | 1 | 10.00% |
Jovi Zhangwei | 2 | 0.68% | 1 | 10.00% |
Jaswinder Singh Rajput | 2 | 0.68% | 1 | 10.00% |
Total | 296 | 100.00% | 10 | 100.00% |
EXPORT_SYMBOL_GPL(register_trace_event);
/*
* Used by module code with the trace_event_sem held for write.
*/
int __unregister_trace_event(struct trace_event *event)
{
hlist_del(&event->node);
list_del(&event->list);
return 0;
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Steven Rostedt | 29 | 100.00% | 4 | 100.00% |
Total | 29 | 100.00% | 4 | 100.00% |
/**
* unregister_trace_event - remove a no longer used event
* @event: the event to remove
*/
int unregister_trace_event(struct trace_event *event)
{
down_write(&trace_event_sem);
__unregister_trace_event(event);
up_write(&trace_event_sem);
return 0;
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Steven Rostedt | 27 | 90.00% | 3 | 60.00% |
Jovi Zhangwei | 2 | 6.67% | 1 | 20.00% |
Lai Jiangshan | 1 | 3.33% | 1 | 20.00% |
Total | 30 | 100.00% | 5 | 100.00% |
EXPORT_SYMBOL_GPL(unregister_trace_event);
/*
* Standard events
*/
enum print_line_t trace_nop_print(struct trace_iterator *iter, int flags,
struct trace_event *event)
{
trace_seq_printf(&iter->seq, "type: %d\n", iter->ent->type);
return trace_handle_return(&iter->seq);
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Steven Rostedt | 25 | 56.82% | 3 | 50.00% |
Jiri Olsa | 15 | 34.09% | 1 | 16.67% |
Arnaldo Carvalho de Melo | 4 | 9.09% | 2 | 33.33% |
Total | 44 | 100.00% | 6 | 100.00% |
/* TRACE_FN */
static enum print_line_t trace_fn_trace(struct trace_iterator *iter, int flags,
struct trace_event *event)
{
struct ftrace_entry *field;
struct trace_seq *s = &iter->seq;
trace_assign_type(field, iter->ent);
seq_print_ip_sym(s, field->ip, flags);
if ((flags & TRACE_ITER_PRINT_PARENT) && field->parent_ip) {
trace_seq_puts(s, " <-");
seq_print_ip_sym(s, field->parent_ip, flags);
}
trace_seq_putc(s, '\n');
return trace_handle_return(s);
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Steven Rostedt | 80 | 80.00% | 3 | 50.00% |
Arnaldo Carvalho de Melo | 17 | 17.00% | 2 | 33.33% |
Jovi Zhangwei | 3 | 3.00% | 1 | 16.67% |
Total | 100 | 100.00% | 6 | 100.00% |
static enum print_line_t trace_fn_raw(struct trace_iterator *iter, int flags,
struct trace_event *event)
{
struct ftrace_entry *field;
trace_assign_type(field, iter->ent);
trace_seq_printf(&iter->seq, "%lx %lx\n",
field->ip,
field->parent_ip);
return trace_handle_return(&iter->seq);
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Steven Rostedt | 50 | 81.97% | 4 | 66.67% |
Arnaldo Carvalho de Melo | 11 | 18.03% | 2 | 33.33% |
Total | 61 | 100.00% | 6 | 100.00% |
static enum print_line_t trace_fn_hex(struct trace_iterator *iter, int flags,
struct trace_event *event)
{
struct ftrace_entry *field;
struct trace_seq *s = &iter->seq;
trace_assign_type(field, iter->ent);
SEQ_PUT_HEX_FIELD(s, field->ip);
SEQ_PUT_HEX_FIELD(s, field->parent_ip);
return trace_handle_return(s);
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Steven Rostedt | 51 | 75.00% | 3 | 60.00% |
Arnaldo Carvalho de Melo | 17 | 25.00% | 2 | 40.00% |
Total | 68 | 100.00% | 5 | 100.00% |
static enum print_line_t trace_fn_bin(struct trace_iterator *iter, int flags,
struct trace_event *event)
{
struct ftrace_entry *field;
struct trace_seq *s = &iter->seq;
trace_assign_type(field, iter->ent);
SEQ_PUT_FIELD(s, field->ip);
SEQ_PUT_FIELD(s, field->parent_ip);
return trace_handle_return(s);
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Steven Rostedt | 51 | 75.00% | 3 | 60.00% |
Arnaldo Carvalho de Melo | 17 | 25.00% | 2 | 40.00% |
Total | 68 | 100.00% | 5 | 100.00% |
static struct trace_event_functions trace_fn_funcs = {
.trace = trace_fn_trace,
.raw = trace_fn_raw,
.hex = trace_fn_hex,
.binary = trace_fn_bin,
};
static struct trace_event trace_fn_event = {
.type = TRACE_FN,
.funcs = &trace_fn_funcs,
};
/* TRACE_CTX an TRACE_WAKE */
static enum print_line_t trace_ctxwake_print(struct trace_iterator *iter,
char *delim)
{
struct ctx_switch_entry *field;
char comm[TASK_COMM_LEN];
int S, T;
trace_assign_type(field, iter->ent);
T = task_index_to_char(field->next_state);
S = task_index_to_char(field->prev_state);
trace_find_cmdline(field->next_pid, comm);
trace_seq_printf(&iter->seq,
" %5d:%3d:%c %s [%03d] %5d:%3d:%c %s\n",
field->prev_pid,
field->prev_prio,
S, delim,
field->next_cpu,
field->next_pid,
field->next_prio,
T, comm);
return trace_handle_return(&iter->seq);
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Steven Rostedt | 102 | 88.70% | 3 | 50.00% |
Arnaldo Carvalho de Melo | 11 | 9.57% | 2 | 33.33% |
Peter Zijlstra | 2 | 1.74% | 1 | 16.67% |
Total | 115 | 100.00% | 6 | 100.00% |
static enum print_line_t trace_ctx_print(struct trace_iterator *iter, int flags,
struct trace_event *event)
{
return trace_ctxwake_print(iter, "==>");
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Steven Rostedt | 23 | 82.14% | 2 | 50.00% |
Arnaldo Carvalho de Melo | 5 | 17.86% | 2 | 50.00% |
Total | 28 | 100.00% | 4 | 100.00% |
static enum print_line_t trace_wake_print(struct trace_iterator *iter,
int flags, struct trace_event *event)
{
return trace_ctxwake_print(iter, " +");
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Steven Rostedt | 23 | 82.14% | 2 | 50.00% |
Arnaldo Carvalho de Melo | 5 | 17.86% | 2 | 50.00% |
Total | 28 | 100.00% | 4 | 100.00% |
static int trace_ctxwake_raw(struct trace_iterator *iter, char S)
{
struct ctx_switch_entry *field;
int T;
trace_assign_type(field, iter->ent);
if (!S)
S = task_index_to_char(field->prev_state);
T = task_index_to_char(field->next_state);
trace_seq_printf(&iter->seq, "%d %d %c %d %d %d %c\n",
field->prev_pid,
field->prev_prio,
S,
field->next_cpu,
field->next_pid,
field->next_prio,
T);
return trace_handle_return(&iter->seq);
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Steven Rostedt | 84 | 86.60% | 2 | 40.00% |
Arnaldo Carvalho de Melo | 9 | 9.28% | 1 | 20.00% |
Peter Zijlstra | 2 | 2.06% | 1 | 20.00% |
Hiroshi Shimamoto | 2 | 2.06% | 1 | 20.00% |
Total | 97 | 100.00% | 5 | 100.00% |
static enum print_line_t trace_ctx_raw(struct trace_iterator *iter, int flags,
struct trace_event *event)
{
return trace_ctxwake_raw(iter, 0);
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Steven Rostedt | 23 | 82.14% | 2 | 50.00% |
Arnaldo Carvalho de Melo | 5 | 17.86% | 2 | 50.00% |
Total | 28 | 100.00% | 4 | 100.00% |
static enum print_line_t trace_wake_raw(struct trace_iterator *iter, int flags,
struct trace_event *event)
{
return trace_ctxwake_raw(iter, '+');
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Steven Rostedt | 23 | 82.14% | 2 | 50.00% |
Arnaldo Carvalho de Melo | 5 | 17.86% | 2 | 50.00% |
Total | 28 | 100.00% | 4 | 100.00% |
static int trace_ctxwake_hex(struct trace_iterator *iter, char S)
{
struct ctx_switch_entry *field;
struct trace_seq *s = &iter->seq;
int T;
trace_assign_type(field, iter->ent);
if (!S)
S = task_index_to_char(field->prev_state);
T = task_index_to_char(field->next_state);
SEQ_PUT_HEX_FIELD(s, field->prev_pid);
SEQ_PUT_HEX_FIELD(s, field->prev_prio);
SEQ_PUT_HEX_FIELD(s, S);
SEQ_PUT_HEX_FIELD(s, field->next_cpu);
SEQ_PUT_HEX_FIELD(s, field->next_pid);
SEQ_PUT_HEX_FIELD(s, field->next_prio);
SEQ_PUT_HEX_FIELD(s, T);
return trace_handle_return(s);
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Steven Rostedt | 110 | 85.27% | 2 | 40.00% |
Arnaldo Carvalho de Melo | 15 | 11.63% | 1 | 20.00% |
Peter Zijlstra | 2 | 1.55% | 1 | 20.00% |
Hiroshi Shimamoto | 2 | 1.55% | 1 | 20.00% |
Total | 129 | 100.00% | 5 | 100.00% |
static enum print_line_t trace_ctx_hex(struct trace_iterator *iter, int flags,
struct trace_event *event)
{
return trace_ctxwake_hex(iter, 0);
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Steven Rostedt | 23 | 82.14% | 2 | 50.00% |
Arnaldo Carvalho de Melo | 5 | 17.86% | 2 | 50.00% |
Total | 28 | 100.00% | 4 | 100.00% |
static enum print_line_t trace_wake_hex(struct trace_iterator *iter, int flags,
struct trace_event *event)
{
return trace_ctxwake_hex(iter, '+');
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Steven Rostedt | 23 | 82.14% | 2 | 50.00% |
Arnaldo Carvalho de Melo | 5 | 17.86% | 2 | 50.00% |
Total | 28 | 100.00% | 4 | 100.00% |
static enum print_line_t trace_ctxwake_bin(struct trace_iterator *iter,
int flags, struct trace_event *event)
{
struct ctx_switch_entry *field;
struct trace_seq *s = &iter->seq;
trace_assign_type(field, iter->ent);
SEQ_PUT_FIELD(s, field->prev_pid);
SEQ_PUT_FIELD(s, field->prev_prio);
SEQ_PUT_FIELD(s, field->prev_state);
SEQ_PUT_FIELD(s, field->next_cpu);
SEQ_PUT_FIELD(s, field->next_pid);
SEQ_PUT_FIELD(s, field->next_prio);
SEQ_PUT_FIELD(s, field->next_state);
return trace_handle_return(s);
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Steven Rostedt | 88 | 77.88% | 3 | 50.00% |
Arnaldo Carvalho de Melo | 17 | 15.04% | 2 | 33.33% |
Jiang Liu | 8 | 7.08% | 1 | 16.67% |
Total | 113 | 100.00% | 6 | 100.00% |
static struct trace_event_functions trace_ctx_funcs = {
.trace = trace_ctx_print,
.raw = trace_ctx_raw,
.hex = trace_ctx_hex,
.binary = trace_ctxwake_bin,
};
static struct trace_event trace_ctx_event = {
.type = TRACE_CTX,
.funcs = &trace_ctx_funcs,
};
static struct trace_event_functions trace_wake_funcs = {
.trace = trace_wake_print,
.raw = trace_wake_raw,
.hex = trace_wake_hex,
.binary = trace_ctxwake_bin,
};
static struct trace_event trace_wake_event = {
.type = TRACE_WAKE,
.funcs = &trace_wake_funcs,
};
/* TRACE_STACK */
static enum print_line_t trace_stack_print(struct trace_iterator *iter,
int flags, struct trace_event *event)
{
struct stack_entry *field;
struct trace_seq *s = &iter->seq;
unsigned long *p;
unsigned long *end;
trace_assign_type(field, iter->ent);
end = (unsigned long *)((long)iter->ent + iter->ent_size);
trace_seq_puts(s, "<stack trace>\n");
for (p = field->caller; p && *p != ULONG_MAX && p < end; p++) {
if (trace_seq_has_overflowed(s))
break;
trace_seq_puts(s, " => ");
seq_print_ip_sym(s, *p, flags);
trace_seq_putc(s, '\n');
}
return trace_handle_return(s);
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Steven Rostedt | 118 | 82.52% | 5 | 55.56% |
Arnaldo Carvalho de Melo | 17 | 11.89% | 2 | 22.22% |
Liming Wang | 6 | 4.20% | 1 | 11.11% |
Jovi Zhangwei | 2 | 1.40% | 1 | 11.11% |
Total | 143 | 100.00% | 9 | 100.00% |
static struct trace_event_functions trace_stack_funcs = {
.trace = trace_stack_print,
};
static struct trace_event trace_stack_event = {
.type = TRACE_STACK,
.funcs = &trace_stack_funcs,
};
/* TRACE_USER_STACK */
static enum print_line_t trace_user_stack_print(struct trace_iterator *iter,
int flags, struct trace_event *event)
{
struct trace_array *tr = iter->tr;
struct userstack_entry *field;
struct trace_seq *s = &iter->seq;
struct mm_struct *mm = NULL;
unsigned int i;
trace_assign_type(field, iter->ent);
trace_seq_puts(s, "<user stack trace>\n");
if (tr->trace_flags & TRACE_ITER_SYM_USEROBJ) {
struct task_struct *task;
/*
* we do the lookup on the thread group leader,
* since individual threads might have already quit!
*/
rcu_read_lock();
task = find_task_by_vpid(field->tgid);
if (task)
mm = get_task_mm(task);
rcu_read_unlock();
}
for (i = 0; i < FTRACE_STACK_ENTRIES; i++) {
unsigned long ip = field->caller[i];
if (ip == ULONG_MAX || trace_seq_has_overflowed(s))
break;
trace_seq_puts(s, " => ");
if (!ip) {
trace_seq_puts(s, "??");
trace_seq_putc(s, '\n');
continue;
}
seq_print_user_ip(s, mm, ip, flags);
trace_seq_putc(s, '\n');
}
if (mm)
mmput(mm);
return trace_handle_return(s);
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Steven Rostedt | 193 | 90.61% | 6 | 66.67% |
Arnaldo Carvalho de Melo | 17 | 7.98% | 2 | 22.22% |
Liming Wang | 3 | 1.41% | 1 | 11.11% |
Total | 213 | 100.00% | 9 | 100.00% |
static struct trace_event_functions trace_user_stack_funcs = {
.trace = trace_user_stack_print,
};
static struct trace_event trace_user_stack_event = {
.type = TRACE_USER_STACK,
.funcs = &trace_user_stack_funcs,
};
/* TRACE_HWLAT */
static enum print_line_t
trace_hwlat_print(struct trace_iterator *iter, int flags,
struct trace_event *event)
{
struct trace_entry *entry = iter->ent;
struct trace_seq *s = &iter->seq;
struct hwlat_entry *field;
trace_assign_type(field, entry);
trace_seq_printf(s, "#%-5u inner/outer(us): %4llu/%-5llu ts:%lld.%09ld",
field->seqnum,
field->duration,
field->outer_duration,
(long long)field->timestamp.tv_sec,
field->timestamp.tv_nsec);
if (field->nmi_count) {
/*
* The generic sched_clock() is not NMI safe, thus
* we only record the count and not the time.
*/
if (!IS_ENABLED(CONFIG_GENERIC_SCHED_CLOCK))
trace_seq_printf(s, " nmi-total:%llu",
field->nmi_total_ts);
trace_seq_printf(s, " nmi-count:%u",
field->nmi_count);
}
trace_seq_putc(s, '\n');
return trace_handle_return(s);
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Steven Rostedt | 133 | 96.38% | 2 | 66.67% |
Deepa Dinamani | 5 | 3.62% | 1 | 33.33% |
Total | 138 | 100.00% | 3 | 100.00% |
static enum print_line_t
trace_hwlat_raw(struct trace_iterator *iter, int flags,
struct trace_event *event)
{
struct hwlat_entry *field;
struct trace_seq *s = &iter->seq;
trace_assign_type(field, iter->ent);
trace_seq_printf(s, "%llu %lld %lld %09ld %u\n",
field->duration,
field->outer_duration,
(long long)field->timestamp.tv_sec,
field->timestamp.tv_nsec,
field->seqnum);
return trace_handle_return(s);
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Steven Rostedt | 80 | 94.12% | 1 | 50.00% |
Deepa Dinamani | 5 | 5.88% | 1 | 50.00% |
Total | 85 | 100.00% | 2 | 100.00% |
static struct trace_event_functions trace_hwlat_funcs = {
.trace = trace_hwlat_print,
.raw = trace_hwlat_raw,
};
static struct trace_event trace_hwlat_event = {
.type = TRACE_HWLAT,
.funcs = &trace_hwlat_funcs,
};
/* TRACE_BPUTS */
static enum print_line_t
trace_bputs_print(struct trace_iterator *iter, int flags,
struct trace_event *event)
{
struct trace_entry *entry = iter->ent;
struct trace_seq *s = &iter->seq;
struct bputs_entry *field;
trace_assign_type(field, entry);
seq_print_ip_sym(s, field->ip, flags);
trace_seq_puts(s, ": ");
trace_seq_puts(s, field->str);
return trace_handle_return(s);
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Steven Rostedt | 84 | 100.00% | 2 | 100.00% |
Total | 84 | 100.00% | 2 | 100.00% |
static enum print_line_t
trace_bputs_raw(struct trace_iterator *iter, int flags,
struct trace_event *event)
{
struct bputs_entry *field;
struct trace_seq *s = &iter->seq;
trace_assign_type(field, iter->ent);
trace_seq_printf(s, ": %lx : ", field->ip);
trace_seq_puts(s, field->str);
return trace_handle_return(s);
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Steven Rostedt | 70 | 100.00% | 2 | 100.00% |
Total | 70 | 100.00% | 2 | 100.00% |
static struct trace_event_functions trace_bputs_funcs = {
.trace = trace_bputs_print,
.raw = trace_bputs_raw,
};
static struct trace_event trace_bputs_event = {
.type = TRACE_BPUTS,
.funcs = &trace_bputs_funcs,
};
/* TRACE_BPRINT */
static enum print_line_t
trace_bprint_print(struct trace_iterator *iter, int flags,
struct trace_event *event)
{
struct trace_entry *entry = iter->ent;
struct trace_seq *s = &iter->seq;
struct bprint_entry *field;
trace_assign_type(field, entry);
seq_print_ip_sym(s, field->ip, flags);
trace_seq_puts(s, ": ");
trace_seq_bprintf(s, field->fmt, field->buf);
return trace_handle_return(s);
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Lai Jiangshan | 71 | 80.68% | 1 | 16.67% |
Steven Rostedt | 13 | 14.77% | 3 | 50.00% |
Arnaldo Carvalho de Melo | 2 | 2.27% | 1 | 16.67% |
Frédéric Weisbecker | 2 | 2.27% | 1 | 16.67% |
Total | 88 | 100.00% | 6 | 100.00% |
static enum print_line_t
trace_bprint_raw(struct trace_iterator *iter, int flags,
struct trace_event *event)
{
struct bprint_entry *field;
struct trace_seq *s = &iter->seq;
trace_assign_type(field, iter->ent);
trace_seq_printf(s, ": %lx : ", field->ip);
trace_seq_bprintf(s, field->fmt, field->buf);
return trace_handle_return(s);
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Lai Jiangshan | 57 | 77.03% | 1 | 20.00% |
Steven Rostedt | 11 | 14.86% | 2 | 40.00% |
Frédéric Weisbecker | 6 | 8.11% | 2 | 40.00% |
Total | 74 | 100.00% | 5 | 100.00% |
static struct trace_event_functions trace_bprint_funcs = {
.trace = trace_bprint_print,
.raw = trace_bprint_raw,
};
static struct trace_event trace_bprint_event = {
.type = TRACE_BPRINT,
.funcs = &trace_bprint_funcs,
};
/* TRACE_PRINT */
static enum print_line_t trace_print_print(struct trace_iterator *iter,
int flags, struct trace_event *event)
{
struct print_entry *field;
struct trace_seq *s = &iter->seq;
trace_assign_type(field, iter->ent);
seq_print_ip_sym(s, field->ip, flags);
trace_seq_printf(s, ": %s", field->buf);
return trace_handle_return(s);
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Frédéric Weisbecker | 61 | 84.72% | 1 | 33.33% |
Steven Rostedt | 11 | 15.28% | 2 | 66.67% |
Total | 72 | 100.00% | 3 | 100.00% |
static enum print_line_t trace_print_raw(struct trace_iterator *iter, int flags,
struct trace_event *event)
{
struct print_entry *field;
trace_assign_type(field, iter->ent);
trace_seq_printf(&iter->seq, "# %lx %s", field->ip, field->buf);
return trace_handle_return(&iter->seq);
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Frédéric Weisbecker | 48 | 78.69% | 1 | 33.33% |
Steven Rostedt | 13 | 21.31% | 2 | 66.67% |
Total | 61 | 100.00% | 3 | 100.00% |
static struct trace_event_functions trace_print_funcs = {
.trace = trace_print_print,
.raw = trace_print_raw,
};
static struct trace_event trace_print_event = {
.type = TRACE_PRINT,
.funcs = &trace_print_funcs,
};
static enum print_line_t trace_raw_data(struct trace_iterator *iter, int flags,
struct trace_event *event)
{
struct raw_data_entry *field;
int i;
trace_assign_type(field, iter->ent);
trace_seq_printf(&iter->seq, "# %x buf:", field->id);
for (i = 0; i < iter->ent_size - offsetof(struct raw_data_entry, buf); i++)
trace_seq_printf(&iter->seq, " %02x",
(unsigned char)field->buf[i]);
trace_seq_putc(&iter->seq, '\n');
return trace_handle_return(&iter->seq);
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Steven Rostedt | 114 | 100.00% | 1 | 100.00% |
Total | 114 | 100.00% | 1 | 100.00% |
static struct trace_event_functions trace_raw_data_funcs = {
.trace = trace_raw_data,
.raw = trace_raw_data,
};
static struct trace_event trace_raw_data_event = {
.type = TRACE_RAW_DATA,
.funcs = &trace_raw_data_funcs,
};
static struct trace_event *events[] __initdata = {
&trace_fn_event,
&trace_ctx_event,
&trace_wake_event,
&trace_stack_event,
&trace_user_stack_event,
&trace_bputs_event,
&trace_bprint_event,
&trace_print_event,
&trace_hwlat_event,
&trace_raw_data_event,
NULL
};
__init static int init_events(void)
{
struct trace_event *event;
int i, ret;
for (i = 0; events[i]; i++) {
event = events[i];
ret = register_trace_event(event);
if (!ret) {
printk(KERN_WARNING "event %d failed to register\n",
event->type);
WARN_ON_ONCE(1);
}
}
return 0;
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Steven Rostedt | 74 | 100.00% | 2 | 100.00% |
Total | 74 | 100.00% | 2 | 100.00% |
early_initcall(init_events);
Overall Contributors
Person | Tokens | Prop | Commits | CommitProp |
Steven Rostedt | 3937 | 64.31% | 42 | 49.41% |
Frédéric Weisbecker | 401 | 6.55% | 3 | 3.53% |
David Sharp | 275 | 4.49% | 2 | 2.35% |
Arnaldo Carvalho de Melo | 270 | 4.41% | 4 | 4.71% |
Dave P Martin | 201 | 3.28% | 1 | 1.18% |
Ross Zwisler | 180 | 2.94% | 1 | 1.18% |
Lai Jiangshan | 172 | 2.81% | 2 | 2.35% |
Byungchul Park | 132 | 2.16% | 1 | 1.18% |
Liu Bo | 108 | 1.76% | 1 | 1.18% |
Li Zefan | 101 | 1.65% | 1 | 1.18% |
Keiichiro Tokunaga | 77 | 1.26% | 1 | 1.18% |
Peter Zijlstra | 74 | 1.21% | 4 | 4.71% |
Joel Fernandes | 43 | 0.70% | 1 | 1.18% |
Steven Whitehouse | 23 | 0.38% | 2 | 2.35% |
Jungseok Lee | 17 | 0.28% | 1 | 1.18% |
Jovi Zhangwei | 15 | 0.25% | 2 | 2.35% |
Jiri Olsa | 15 | 0.25% | 1 | 1.18% |
Johannes Berg | 13 | 0.21% | 1 | 1.18% |
Andrey Vagin | 11 | 0.18% | 1 | 1.18% |
Deepa Dinamani | 10 | 0.16% | 1 | 1.18% |
Liming Wang | 9 | 0.15% | 2 | 2.35% |
Jiang Liu | 8 | 0.13% | 1 | 1.18% |
Alex Bennée | 8 | 0.13% | 1 | 1.18% |
Ingo Molnar | 6 | 0.10% | 2 | 2.35% |
Daniel Borkmann | 6 | 0.10% | 2 | 2.35% |
Hiroshi Shimamoto | 4 | 0.07% | 1 | 1.18% |
Eduard - Gabriel Munteanu | 2 | 0.03% | 1 | 1.18% |
Mathieu Desnoyers | 2 | 0.03% | 1 | 1.18% |
Jaswinder Singh Rajput | 2 | 0.03% | 1 | 1.18% |
Total | 6122 | 100.00% | 85 | 100.00% |
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.