License cleanup: add SPDX GPL-2.0 license identifier to files with no license
Many source files in the tree are missing licensing information, which
makes it harder for compliance tools to determine the correct license.
By default all files without license information are under the default
license of the kernel, which is GPL version 2.
Update the files which contain no license information with the 'GPL-2.0'
SPDX license identifier. The SPDX identifier is a legally binding
shorthand, which can be used instead of the full boiler plate text.
This patch is based on work done by Thomas Gleixner and Kate Stewart and
Philippe Ombredanne.
How this work was done:
Patches were generated and checked against linux-4.14-rc6 for a subset of
the use cases:
- file had no licensing information it it.
- file was a */uapi/* one with no licensing information in it,
- file was a */uapi/* one with existing licensing information,
Further patches will be generated in subsequent months to fix up cases
where non-standard license headers were used, and references to license
had to be inferred by heuristics based on keywords.
The analysis to determine which SPDX License Identifier to be applied to
a file was done in a spreadsheet of side by side results from of the
output of two independent scanners (ScanCode & Windriver) producing SPDX
tag:value files created by Philippe Ombredanne. Philippe prepared the
base worksheet, and did an initial spot review of a few 1000 files.
The 4.13 kernel was the starting point of the analysis with 60,537 files
assessed. Kate Stewart did a file by file comparison of the scanner
results in the spreadsheet to determine which SPDX license identifier(s)
to be applied to the file. She confirmed any determination that was not
immediately clear with lawyers working with the Linux Foundation.
Criteria used to select files for SPDX license identifier tagging was:
- Files considered eligible had to be source code files.
- Make and config files were included as candidates if they contained >5
lines of source
- File already had some variant of a license header in it (even if <5
lines).
All documentation files were explicitly excluded.
The following heuristics were used to determine which SPDX license
identifiers to apply.
- when both scanners couldn't find any license traces, file was
considered to have no license information in it, and the top level
COPYING file license applied.
For non */uapi/* files that summary was:
SPDX license identifier # files
---------------------------------------------------|-------
GPL-2.0 11139
and resulted in the first patch in this series.
If that file was a */uapi/* path one, it was "GPL-2.0 WITH
Linux-syscall-note" otherwise it was "GPL-2.0". Results of that was:
SPDX license identifier # files
---------------------------------------------------|-------
GPL-2.0 WITH Linux-syscall-note 930
and resulted in the second patch in this series.
- if a file had some form of licensing information in it, and was one
of the */uapi/* ones, it was denoted with the Linux-syscall-note if
any GPL family license was found in the file or had no licensing in
it (per prior point). Results summary:
SPDX license identifier # files
---------------------------------------------------|------
GPL-2.0 WITH Linux-syscall-note 270
GPL-2.0+ WITH Linux-syscall-note 169
((GPL-2.0 WITH Linux-syscall-note) OR BSD-2-Clause) 21
((GPL-2.0 WITH Linux-syscall-note) OR BSD-3-Clause) 17
LGPL-2.1+ WITH Linux-syscall-note 15
GPL-1.0+ WITH Linux-syscall-note 14
((GPL-2.0+ WITH Linux-syscall-note) OR BSD-3-Clause) 5
LGPL-2.0+ WITH Linux-syscall-note 4
LGPL-2.1 WITH Linux-syscall-note 3
((GPL-2.0 WITH Linux-syscall-note) OR MIT) 3
((GPL-2.0 WITH Linux-syscall-note) AND MIT) 1
and that resulted in the third patch in this series.
- when the two scanners agreed on the detected license(s), that became
the concluded license(s).
- when there was disagreement between the two scanners (one detected a
license but the other didn't, or they both detected different
licenses) a manual inspection of the file occurred.
- In most cases a manual inspection of the information in the file
resulted in a clear resolution of the license that should apply (and
which scanner probably needed to revisit its heuristics).
- When it was not immediately clear, the license identifier was
confirmed with lawyers working with the Linux Foundation.
- If there was any question as to the appropriate license identifier,
the file was flagged for further research and to be revisited later
in time.
In total, over 70 hours of logged manual review was done on the
spreadsheet to determine the SPDX license identifiers to apply to the
source files by Kate, Philippe, Thomas and, in some cases, confirmation
by lawyers working with the Linux Foundation.
Kate also obtained a third independent scan of the 4.13 code base from
FOSSology, and compared selected files where the other two scanners
disagreed against that SPDX file, to see if there was new insights. The
Windriver scanner is based on an older version of FOSSology in part, so
they are related.
Thomas did random spot checks in about 500 files from the spreadsheets
for the uapi headers and agreed with SPDX license identifier in the
files he inspected. For the non-uapi files Thomas did random spot checks
in about 15000 files.
In initial set of patches against 4.14-rc6, 3 files were found to have
copy/paste license identifier errors, and have been fixed to reflect the
correct identifier.
Additionally Philippe spent 10 hours this week doing a detailed manual
inspection and review of the 12,461 patched files from the initial patch
version early this week with:
- a full scancode scan run, collecting the matched texts, detected
license ids and scores
- reviewing anything where there was a license detected (about 500+
files) to ensure that the applied SPDX license was correct
- reviewing anything where there was no detection but the patch license
was not GPL-2.0 WITH Linux-syscall-note to ensure that the applied
SPDX license was correct
This produced a worksheet with 20 files needing minor correction. This
worksheet was then exported into 3 different .csv files for the
different types of files to be modified.
These .csv files were then reviewed by Greg. Thomas wrote a script to
parse the csv files and add the proper SPDX tag to the file, in the
format that the file expected. This script was further refined by Greg
based on the output to detect more types of files automatically and to
distinguish between header and source .c files (which need different
comment types.) Finally Greg ran the script using the .csv files to
generate the patches.
Reviewed-by: Kate Stewart <kstewart@linuxfoundation.org>
Reviewed-by: Philippe Ombredanne <pombredanne@nexb.com>
Reviewed-by: Thomas Gleixner <tglx@linutronix.de>
Signed-off-by: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
2017-11-01 22:07:57 +08:00
|
|
|
// SPDX-License-Identifier: GPL-2.0
|
2012-08-20 12:52:05 +08:00
|
|
|
#include <stdio.h>
|
2019-08-31 01:45:20 +08:00
|
|
|
#include <stdlib.h>
|
2017-07-21 02:27:39 +08:00
|
|
|
#include <linux/string.h>
|
2012-08-20 12:52:05 +08:00
|
|
|
|
2019-01-29 18:11:04 +08:00
|
|
|
#include "../../util/callchain.h"
|
2019-08-23 04:11:39 +08:00
|
|
|
#include "../../util/debug.h"
|
2019-09-18 22:36:13 +08:00
|
|
|
#include "../../util/event.h"
|
2012-08-20 12:52:05 +08:00
|
|
|
#include "../../util/hist.h"
|
2019-01-27 18:55:39 +08:00
|
|
|
#include "../../util/map.h"
|
2019-11-26 09:24:10 +08:00
|
|
|
#include "../../util/maps.h"
|
2019-01-28 07:03:34 +08:00
|
|
|
#include "../../util/symbol.h"
|
2012-08-20 12:52:05 +08:00
|
|
|
#include "../../util/sort.h"
|
2013-01-22 17:09:37 +08:00
|
|
|
#include "../../util/evsel.h"
|
2017-04-18 03:30:49 +08:00
|
|
|
#include "../../util/srcline.h"
|
2017-04-18 03:51:59 +08:00
|
|
|
#include "../../util/string2.h"
|
2017-04-20 08:34:35 +08:00
|
|
|
#include "../../util/thread.h"
|
2019-11-07 15:47:18 +08:00
|
|
|
#include "../../util/block-info.h"
|
tools perf: Move from sane_ctype.h obtained from git to the Linux's original
We got the sane_ctype.h headers from git and kept using it so far, but
since that code originally came from the kernel sources to the git
sources, perhaps its better to just use the one in the kernel, so that
we can leverage tools/perf/check_headers.sh to be notified when our copy
gets out of sync, i.e. when fixes or goodies are added to the code we've
copied.
This will help with things like tools/lib/string.c where we want to have
more things in common with the kernel, such as strim(), skip_spaces(),
etc so as to go on removing the things that we have in tools/perf/util/
and instead using the code in the kernel, indirectly and removing things
like EXPORT_SYMBOL(), etc, getting notified when fixes and improvements
are made to the original code.
Hopefully this also should help with reducing the difference of code
hosted in tools/ to the one in the kernel proper.
Cc: Adrian Hunter <adrian.hunter@intel.com>
Cc: Jiri Olsa <jolsa@kernel.org>
Cc: Namhyung Kim <namhyung@kernel.org>
Link: https://lkml.kernel.org/n/tip-7k9868l713wqtgo01xxygn12@git.kernel.org
Signed-off-by: Arnaldo Carvalho de Melo <acme@redhat.com>
2019-06-26 04:27:31 +08:00
|
|
|
#include <linux/ctype.h>
|
2019-07-04 22:32:27 +08:00
|
|
|
#include <linux/zalloc.h>
|
2012-08-20 12:52:05 +08:00
|
|
|
|
|
|
|
static size_t callchain__fprintf_left_margin(FILE *fp, int left_margin)
|
|
|
|
{
|
|
|
|
int i;
|
|
|
|
int ret = fprintf(fp, " ");
|
|
|
|
|
|
|
|
for (i = 0; i < left_margin; i++)
|
|
|
|
ret += fprintf(fp, " ");
|
|
|
|
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
|
|
|
static size_t ipchain__fprintf_graph_line(FILE *fp, int depth, int depth_mask,
|
|
|
|
int left_margin)
|
|
|
|
{
|
|
|
|
int i;
|
|
|
|
size_t ret = callchain__fprintf_left_margin(fp, left_margin);
|
|
|
|
|
|
|
|
for (i = 0; i < depth; i++)
|
|
|
|
if (depth_mask & (1 << i))
|
|
|
|
ret += fprintf(fp, "| ");
|
|
|
|
else
|
|
|
|
ret += fprintf(fp, " ");
|
|
|
|
|
|
|
|
ret += fprintf(fp, "\n");
|
|
|
|
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
2015-11-09 13:45:39 +08:00
|
|
|
static size_t ipchain__fprintf_graph(FILE *fp, struct callchain_node *node,
|
|
|
|
struct callchain_list *chain,
|
2012-08-20 12:52:05 +08:00
|
|
|
int depth, int depth_mask, int period,
|
2015-11-09 13:45:39 +08:00
|
|
|
u64 total_samples, int left_margin)
|
2012-08-20 12:52:05 +08:00
|
|
|
{
|
|
|
|
int i;
|
|
|
|
size_t ret = 0;
|
2016-10-31 09:19:52 +08:00
|
|
|
char bf[1024], *alloc_str = NULL;
|
|
|
|
char buf[64];
|
|
|
|
const char *str;
|
2012-08-20 12:52:05 +08:00
|
|
|
|
|
|
|
ret += callchain__fprintf_left_margin(fp, left_margin);
|
|
|
|
for (i = 0; i < depth; i++) {
|
|
|
|
if (depth_mask & (1 << i))
|
|
|
|
ret += fprintf(fp, "|");
|
|
|
|
else
|
|
|
|
ret += fprintf(fp, " ");
|
|
|
|
if (!period && i == depth - 1) {
|
2015-11-09 13:45:39 +08:00
|
|
|
ret += fprintf(fp, "--");
|
|
|
|
ret += callchain_node__fprintf_value(node, fp, total_samples);
|
|
|
|
ret += fprintf(fp, "--");
|
2012-08-20 12:52:05 +08:00
|
|
|
} else
|
|
|
|
ret += fprintf(fp, "%s", " ");
|
|
|
|
}
|
2016-10-31 09:19:52 +08:00
|
|
|
|
|
|
|
str = callchain_list__sym_name(chain, bf, sizeof(bf), false);
|
|
|
|
|
|
|
|
if (symbol_conf.show_branchflag_count) {
|
2017-08-07 21:05:15 +08:00
|
|
|
callchain_list_counts__printf_value(chain, NULL,
|
|
|
|
buf, sizeof(buf));
|
2016-10-31 09:19:52 +08:00
|
|
|
|
|
|
|
if (asprintf(&alloc_str, "%s%s", str, buf) < 0)
|
|
|
|
str = "Not enough memory!";
|
|
|
|
else
|
|
|
|
str = alloc_str;
|
|
|
|
}
|
|
|
|
|
|
|
|
fputs(str, fp);
|
2014-11-13 10:05:23 +08:00
|
|
|
fputc('\n', fp);
|
2016-10-31 09:19:52 +08:00
|
|
|
free(alloc_str);
|
2017-03-26 04:34:28 +08:00
|
|
|
|
2012-08-20 12:52:05 +08:00
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
|
|
|
static struct symbol *rem_sq_bracket;
|
|
|
|
static struct callchain_list rem_hits;
|
|
|
|
|
|
|
|
static void init_rem_hits(void)
|
|
|
|
{
|
|
|
|
rem_sq_bracket = malloc(sizeof(*rem_sq_bracket) + 6);
|
|
|
|
if (!rem_sq_bracket) {
|
|
|
|
fprintf(stderr, "Not enough memory to display remaining hits\n");
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
strcpy(rem_sq_bracket->name, "[...]");
|
|
|
|
rem_hits.ms.sym = rem_sq_bracket;
|
|
|
|
}
|
|
|
|
|
|
|
|
static size_t __callchain__fprintf_graph(FILE *fp, struct rb_root *root,
|
|
|
|
u64 total_samples, int depth,
|
|
|
|
int depth_mask, int left_margin)
|
|
|
|
{
|
|
|
|
struct rb_node *node, *next;
|
2015-11-09 13:45:41 +08:00
|
|
|
struct callchain_node *child = NULL;
|
2012-08-20 12:52:05 +08:00
|
|
|
struct callchain_list *chain;
|
|
|
|
int new_depth_mask = depth_mask;
|
|
|
|
u64 remaining;
|
|
|
|
size_t ret = 0;
|
|
|
|
int i;
|
|
|
|
uint entries_printed = 0;
|
2015-11-09 13:45:41 +08:00
|
|
|
int cumul_count = 0;
|
2012-08-20 12:52:05 +08:00
|
|
|
|
|
|
|
remaining = total_samples;
|
|
|
|
|
|
|
|
node = rb_first(root);
|
|
|
|
while (node) {
|
|
|
|
u64 new_total;
|
|
|
|
u64 cumul;
|
|
|
|
|
|
|
|
child = rb_entry(node, struct callchain_node, rb_node);
|
|
|
|
cumul = callchain_cumul_hits(child);
|
|
|
|
remaining -= cumul;
|
2015-11-09 13:45:41 +08:00
|
|
|
cumul_count += callchain_cumul_counts(child);
|
2012-08-20 12:52:05 +08:00
|
|
|
|
|
|
|
/*
|
|
|
|
* The depth mask manages the output of pipes that show
|
|
|
|
* the depth. We don't want to keep the pipes of the current
|
|
|
|
* level for the last child of this depth.
|
|
|
|
* Except if we have remaining filtered hits. They will
|
|
|
|
* supersede the last child
|
|
|
|
*/
|
|
|
|
next = rb_next(node);
|
|
|
|
if (!next && (callchain_param.mode != CHAIN_GRAPH_REL || !remaining))
|
|
|
|
new_depth_mask &= ~(1 << (depth - 1));
|
|
|
|
|
|
|
|
/*
|
|
|
|
* But we keep the older depth mask for the line separator
|
|
|
|
* to keep the level link until we reach the last child
|
|
|
|
*/
|
|
|
|
ret += ipchain__fprintf_graph_line(fp, depth, depth_mask,
|
|
|
|
left_margin);
|
|
|
|
i = 0;
|
|
|
|
list_for_each_entry(chain, &child->val, list) {
|
2015-11-09 13:45:39 +08:00
|
|
|
ret += ipchain__fprintf_graph(fp, child, chain, depth,
|
2012-08-20 12:52:05 +08:00
|
|
|
new_depth_mask, i++,
|
|
|
|
total_samples,
|
|
|
|
left_margin);
|
|
|
|
}
|
|
|
|
|
|
|
|
if (callchain_param.mode == CHAIN_GRAPH_REL)
|
|
|
|
new_total = child->children_hit;
|
|
|
|
else
|
|
|
|
new_total = total_samples;
|
|
|
|
|
|
|
|
ret += __callchain__fprintf_graph(fp, &child->rb_root, new_total,
|
|
|
|
depth + 1,
|
|
|
|
new_depth_mask | (1 << depth),
|
|
|
|
left_margin);
|
|
|
|
node = next;
|
|
|
|
if (++entries_printed == callchain_param.print_limit)
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (callchain_param.mode == CHAIN_GRAPH_REL &&
|
|
|
|
remaining && remaining != total_samples) {
|
2015-11-09 13:45:39 +08:00
|
|
|
struct callchain_node rem_node = {
|
|
|
|
.hit = remaining,
|
|
|
|
};
|
2012-08-20 12:52:05 +08:00
|
|
|
|
|
|
|
if (!rem_sq_bracket)
|
|
|
|
return ret;
|
|
|
|
|
2015-11-09 13:45:41 +08:00
|
|
|
if (callchain_param.value == CCVAL_COUNT && child && child->parent) {
|
|
|
|
rem_node.count = child->parent->children_count - cumul_count;
|
|
|
|
if (rem_node.count <= 0)
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
2012-08-20 12:52:05 +08:00
|
|
|
new_depth_mask &= ~(1 << (depth - 1));
|
2015-11-09 13:45:39 +08:00
|
|
|
ret += ipchain__fprintf_graph(fp, &rem_node, &rem_hits, depth,
|
2012-08-20 12:52:05 +08:00
|
|
|
new_depth_mask, 0, total_samples,
|
2015-11-09 13:45:39 +08:00
|
|
|
left_margin);
|
2012-08-20 12:52:05 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
2016-01-27 23:40:53 +08:00
|
|
|
/*
|
|
|
|
* If have one single callchain root, don't bother printing
|
|
|
|
* its percentage (100 % in fractal mode and the same percentage
|
|
|
|
* than the hist in graph mode). This also avoid one level of column.
|
|
|
|
*
|
|
|
|
* However when percent-limit applied, it's possible that single callchain
|
|
|
|
* node have different (non-100% in fractal mode) percentage.
|
|
|
|
*/
|
|
|
|
static bool need_percent_display(struct rb_node *node, u64 parent_samples)
|
|
|
|
{
|
|
|
|
struct callchain_node *cnode;
|
|
|
|
|
|
|
|
if (rb_next(node))
|
|
|
|
return true;
|
|
|
|
|
|
|
|
cnode = rb_entry(node, struct callchain_node, rb_node);
|
|
|
|
return callchain_cumul_hits(cnode) != parent_samples;
|
|
|
|
}
|
|
|
|
|
2012-08-20 12:52:05 +08:00
|
|
|
static size_t callchain__fprintf_graph(FILE *fp, struct rb_root *root,
|
2016-01-27 23:40:52 +08:00
|
|
|
u64 total_samples, u64 parent_samples,
|
|
|
|
int left_margin)
|
2012-08-20 12:52:05 +08:00
|
|
|
{
|
|
|
|
struct callchain_node *cnode;
|
|
|
|
struct callchain_list *chain;
|
|
|
|
u32 entries_printed = 0;
|
|
|
|
bool printed = false;
|
|
|
|
struct rb_node *node;
|
|
|
|
int i = 0;
|
|
|
|
int ret = 0;
|
2014-11-13 10:05:23 +08:00
|
|
|
char bf[1024];
|
2012-08-20 12:52:05 +08:00
|
|
|
|
|
|
|
node = rb_first(root);
|
2016-01-27 23:40:53 +08:00
|
|
|
if (node && !need_percent_display(node, parent_samples)) {
|
2012-08-20 12:52:05 +08:00
|
|
|
cnode = rb_entry(node, struct callchain_node, rb_node);
|
|
|
|
list_for_each_entry(chain, &cnode->val, list) {
|
|
|
|
/*
|
|
|
|
* If we sort by symbol, the first entry is the same than
|
|
|
|
* the symbol. No need to print it otherwise it appears as
|
|
|
|
* displayed twice.
|
|
|
|
*/
|
2014-05-19 13:19:30 +08:00
|
|
|
if (!i++ && field_order == NULL &&
|
2017-07-21 02:27:39 +08:00
|
|
|
sort_order && strstarts(sort_order, "sym"))
|
2012-08-20 12:52:05 +08:00
|
|
|
continue;
|
2017-03-26 04:34:28 +08:00
|
|
|
|
2012-08-20 12:52:05 +08:00
|
|
|
if (!printed) {
|
|
|
|
ret += callchain__fprintf_left_margin(fp, left_margin);
|
|
|
|
ret += fprintf(fp, "|\n");
|
|
|
|
ret += callchain__fprintf_left_margin(fp, left_margin);
|
|
|
|
ret += fprintf(fp, "---");
|
|
|
|
left_margin += 3;
|
|
|
|
printed = true;
|
|
|
|
} else
|
|
|
|
ret += callchain__fprintf_left_margin(fp, left_margin);
|
|
|
|
|
2016-10-31 09:19:52 +08:00
|
|
|
ret += fprintf(fp, "%s",
|
|
|
|
callchain_list__sym_name(chain, bf,
|
|
|
|
sizeof(bf),
|
|
|
|
false));
|
|
|
|
|
|
|
|
if (symbol_conf.show_branchflag_count)
|
|
|
|
ret += callchain_list_counts__printf_value(
|
2017-08-07 21:05:15 +08:00
|
|
|
chain, fp, NULL, 0);
|
2016-10-31 09:19:52 +08:00
|
|
|
ret += fprintf(fp, "\n");
|
2012-08-20 12:52:05 +08:00
|
|
|
|
|
|
|
if (++entries_printed == callchain_param.print_limit)
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
root = &cnode->rb_root;
|
|
|
|
}
|
|
|
|
|
2016-01-27 23:40:52 +08:00
|
|
|
if (callchain_param.mode == CHAIN_GRAPH_REL)
|
|
|
|
total_samples = parent_samples;
|
|
|
|
|
2012-08-20 12:52:05 +08:00
|
|
|
ret += __callchain__fprintf_graph(fp, root, total_samples,
|
|
|
|
1, 1, left_margin);
|
2016-01-28 20:24:54 +08:00
|
|
|
if (ret) {
|
|
|
|
/* do not add a blank line if it printed nothing */
|
|
|
|
ret += fprintf(fp, "\n");
|
|
|
|
}
|
2012-08-20 12:52:05 +08:00
|
|
|
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
2013-11-06 02:32:36 +08:00
|
|
|
static size_t __callchain__fprintf_flat(FILE *fp, struct callchain_node *node,
|
2012-08-20 12:52:05 +08:00
|
|
|
u64 total_samples)
|
|
|
|
{
|
|
|
|
struct callchain_list *chain;
|
|
|
|
size_t ret = 0;
|
2014-11-13 10:05:23 +08:00
|
|
|
char bf[1024];
|
2012-08-20 12:52:05 +08:00
|
|
|
|
2013-11-06 02:32:36 +08:00
|
|
|
if (!node)
|
2012-08-20 12:52:05 +08:00
|
|
|
return 0;
|
|
|
|
|
2013-11-06 02:32:36 +08:00
|
|
|
ret += __callchain__fprintf_flat(fp, node->parent, total_samples);
|
2012-08-20 12:52:05 +08:00
|
|
|
|
|
|
|
|
2013-11-06 02:32:36 +08:00
|
|
|
list_for_each_entry(chain, &node->val, list) {
|
2012-08-20 12:52:05 +08:00
|
|
|
if (chain->ip >= PERF_CONTEXT_MAX)
|
|
|
|
continue;
|
2014-11-13 10:05:23 +08:00
|
|
|
ret += fprintf(fp, " %s\n", callchain_list__sym_name(chain,
|
|
|
|
bf, sizeof(bf), false));
|
2012-08-20 12:52:05 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
2013-11-06 02:32:36 +08:00
|
|
|
static size_t callchain__fprintf_flat(FILE *fp, struct rb_root *tree,
|
2012-08-20 12:52:05 +08:00
|
|
|
u64 total_samples)
|
|
|
|
{
|
|
|
|
size_t ret = 0;
|
|
|
|
u32 entries_printed = 0;
|
|
|
|
struct callchain_node *chain;
|
2013-11-06 02:32:36 +08:00
|
|
|
struct rb_node *rb_node = rb_first(tree);
|
2012-08-20 12:52:05 +08:00
|
|
|
|
|
|
|
while (rb_node) {
|
|
|
|
chain = rb_entry(rb_node, struct callchain_node, rb_node);
|
|
|
|
|
2015-11-09 13:45:39 +08:00
|
|
|
ret += fprintf(fp, " ");
|
|
|
|
ret += callchain_node__fprintf_value(chain, fp, total_samples);
|
|
|
|
ret += fprintf(fp, "\n");
|
2012-08-20 12:52:05 +08:00
|
|
|
ret += __callchain__fprintf_flat(fp, chain, total_samples);
|
|
|
|
ret += fprintf(fp, "\n");
|
|
|
|
if (++entries_printed == callchain_param.print_limit)
|
|
|
|
break;
|
|
|
|
|
|
|
|
rb_node = rb_next(rb_node);
|
|
|
|
}
|
|
|
|
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
2015-11-09 13:45:37 +08:00
|
|
|
static size_t __callchain__fprintf_folded(FILE *fp, struct callchain_node *node)
|
|
|
|
{
|
|
|
|
const char *sep = symbol_conf.field_sep ?: ";";
|
|
|
|
struct callchain_list *chain;
|
|
|
|
size_t ret = 0;
|
|
|
|
char bf[1024];
|
|
|
|
bool first;
|
|
|
|
|
|
|
|
if (!node)
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
ret += __callchain__fprintf_folded(fp, node->parent);
|
|
|
|
|
|
|
|
first = (ret == 0);
|
|
|
|
list_for_each_entry(chain, &node->val, list) {
|
|
|
|
if (chain->ip >= PERF_CONTEXT_MAX)
|
|
|
|
continue;
|
|
|
|
ret += fprintf(fp, "%s%s", first ? "" : sep,
|
|
|
|
callchain_list__sym_name(chain,
|
|
|
|
bf, sizeof(bf), false));
|
|
|
|
first = false;
|
|
|
|
}
|
|
|
|
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
|
|
|
static size_t callchain__fprintf_folded(FILE *fp, struct rb_root *tree,
|
|
|
|
u64 total_samples)
|
|
|
|
{
|
|
|
|
size_t ret = 0;
|
|
|
|
u32 entries_printed = 0;
|
|
|
|
struct callchain_node *chain;
|
|
|
|
struct rb_node *rb_node = rb_first(tree);
|
|
|
|
|
|
|
|
while (rb_node) {
|
|
|
|
|
|
|
|
chain = rb_entry(rb_node, struct callchain_node, rb_node);
|
|
|
|
|
2015-11-09 13:45:39 +08:00
|
|
|
ret += callchain_node__fprintf_value(chain, fp, total_samples);
|
|
|
|
ret += fprintf(fp, " ");
|
2015-11-09 13:45:37 +08:00
|
|
|
ret += __callchain__fprintf_folded(fp, chain);
|
|
|
|
ret += fprintf(fp, "\n");
|
|
|
|
if (++entries_printed == callchain_param.print_limit)
|
|
|
|
break;
|
|
|
|
|
|
|
|
rb_node = rb_next(rb_node);
|
|
|
|
}
|
|
|
|
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
2012-08-20 12:52:05 +08:00
|
|
|
static size_t hist_entry_callchain__fprintf(struct hist_entry *he,
|
|
|
|
u64 total_samples, int left_margin,
|
|
|
|
FILE *fp)
|
|
|
|
{
|
2016-01-27 23:40:52 +08:00
|
|
|
u64 parent_samples = he->stat.period;
|
|
|
|
|
|
|
|
if (symbol_conf.cumulate_callchain)
|
|
|
|
parent_samples = he->stat_acc->period;
|
|
|
|
|
2012-08-20 12:52:05 +08:00
|
|
|
switch (callchain_param.mode) {
|
|
|
|
case CHAIN_GRAPH_REL:
|
2016-01-27 23:40:52 +08:00
|
|
|
return callchain__fprintf_graph(fp, &he->sorted_chain, total_samples,
|
|
|
|
parent_samples, left_margin);
|
2012-08-20 12:52:05 +08:00
|
|
|
break;
|
|
|
|
case CHAIN_GRAPH_ABS:
|
|
|
|
return callchain__fprintf_graph(fp, &he->sorted_chain, total_samples,
|
2016-01-27 23:40:52 +08:00
|
|
|
parent_samples, left_margin);
|
2012-08-20 12:52:05 +08:00
|
|
|
break;
|
|
|
|
case CHAIN_FLAT:
|
|
|
|
return callchain__fprintf_flat(fp, &he->sorted_chain, total_samples);
|
|
|
|
break;
|
2015-11-09 13:45:37 +08:00
|
|
|
case CHAIN_FOLDED:
|
|
|
|
return callchain__fprintf_folded(fp, &he->sorted_chain, total_samples);
|
|
|
|
break;
|
2012-08-20 12:52:05 +08:00
|
|
|
case CHAIN_NONE:
|
|
|
|
break;
|
|
|
|
default:
|
|
|
|
pr_err("Bad callchain mode\n");
|
|
|
|
}
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2016-09-22 23:36:36 +08:00
|
|
|
int __hist_entry__snprintf(struct hist_entry *he, struct perf_hpp *hpp,
|
|
|
|
struct perf_hpp_list *hpp_list)
|
2013-02-04 23:33:19 +08:00
|
|
|
{
|
|
|
|
const char *sep = symbol_conf.field_sep;
|
|
|
|
struct perf_hpp_fmt *fmt;
|
|
|
|
char *start = hpp->buf;
|
|
|
|
int ret;
|
|
|
|
bool first = true;
|
|
|
|
|
|
|
|
if (symbol_conf.exclude_other && !he->parent)
|
|
|
|
return 0;
|
|
|
|
|
2016-09-22 23:36:29 +08:00
|
|
|
perf_hpp_list__for_each_format(hpp_list, fmt) {
|
2015-12-23 01:07:08 +08:00
|
|
|
if (perf_hpp__should_skip(fmt, he->hists))
|
2014-03-18 12:00:59 +08:00
|
|
|
continue;
|
|
|
|
|
2013-02-04 23:33:19 +08:00
|
|
|
/*
|
|
|
|
* If there's no field_sep, we still need
|
|
|
|
* to display initial ' '.
|
|
|
|
*/
|
|
|
|
if (!sep || !first) {
|
|
|
|
ret = scnprintf(hpp->buf, hpp->size, "%s", sep ?: " ");
|
|
|
|
advance_hpp(hpp, ret);
|
|
|
|
} else
|
|
|
|
first = false;
|
|
|
|
|
2013-10-25 19:24:53 +08:00
|
|
|
if (perf_hpp__use_color() && fmt->color)
|
2013-02-04 23:33:19 +08:00
|
|
|
ret = fmt->color(fmt, hpp, he);
|
|
|
|
else
|
|
|
|
ret = fmt->entry(fmt, hpp, he);
|
|
|
|
|
perf hists: Do column alignment on the format iterator
We were doing column alignment in the format function for each cell,
returning a string padded with spaces so that when the next column is
printed the cursor is at its column alignment.
This ends up needlessly printing trailing spaces, do it at the format
iterator, that is where we know if it is needed, i.e. if there is more
columns to be printed.
This eliminates the need for triming lines when doing a dump using 'P'
in the TUI browser and also produces far saner results with things like
piping 'perf report' to 'less'.
Right now only the formatters for sym->name and the 'locked' column
(perf mem report), that are the ones that end up at the end of lines
in the default 'perf report', 'perf top' and 'perf mem report' tools,
the others will be done in a subsequent patch.
In the end the 'width' parameter for the formatters now mean, in
'printf' terms, the 'precision', where before it was the field 'width'.
Reported-by: Dave Jones <davej@codemonkey.org.uk>
Cc: Jiri Olsa <jolsa@kernel.org>
Cc: Namhyung Kim <namhyung@kernel.org>
Link: http://lkml.kernel.org/n/tip-s7iwl2gj23w92l6tibnrcqzr@git.kernel.org
Signed-off-by: Arnaldo Carvalho de Melo <acme@redhat.com>
2016-02-12 04:14:13 +08:00
|
|
|
ret = hist_entry__snprintf_alignment(he, hpp, fmt, ret);
|
2013-02-04 23:33:19 +08:00
|
|
|
advance_hpp(hpp, ret);
|
|
|
|
}
|
|
|
|
|
|
|
|
return hpp->buf - start;
|
|
|
|
}
|
|
|
|
|
2016-09-22 23:36:29 +08:00
|
|
|
static int hist_entry__snprintf(struct hist_entry *he, struct perf_hpp *hpp)
|
|
|
|
{
|
|
|
|
return __hist_entry__snprintf(he, hpp, he->hists->hpp_list);
|
|
|
|
}
|
|
|
|
|
2016-02-24 23:13:41 +08:00
|
|
|
static int hist_entry__hierarchy_fprintf(struct hist_entry *he,
|
|
|
|
struct perf_hpp *hpp,
|
2016-03-08 03:44:48 +08:00
|
|
|
struct hists *hists,
|
2016-02-24 23:13:41 +08:00
|
|
|
FILE *fp)
|
|
|
|
{
|
|
|
|
const char *sep = symbol_conf.field_sep;
|
|
|
|
struct perf_hpp_fmt *fmt;
|
2016-03-08 03:44:49 +08:00
|
|
|
struct perf_hpp_list_node *fmt_node;
|
2016-02-24 23:13:41 +08:00
|
|
|
char *buf = hpp->buf;
|
2016-02-27 02:52:45 +08:00
|
|
|
size_t size = hpp->size;
|
2016-02-24 23:13:41 +08:00
|
|
|
int ret, printed = 0;
|
|
|
|
bool first = true;
|
|
|
|
|
|
|
|
if (symbol_conf.exclude_other && !he->parent)
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
ret = scnprintf(hpp->buf, hpp->size, "%*s", he->depth * HIERARCHY_INDENT, "");
|
|
|
|
advance_hpp(hpp, ret);
|
|
|
|
|
2016-03-08 03:44:49 +08:00
|
|
|
/* the first hpp_list_node is for overhead columns */
|
|
|
|
fmt_node = list_first_entry(&hists->hpp_formats,
|
|
|
|
struct perf_hpp_list_node, list);
|
|
|
|
perf_hpp_list__for_each_format(&fmt_node->hpp, fmt) {
|
2016-02-24 23:13:41 +08:00
|
|
|
/*
|
|
|
|
* If there's no field_sep, we still need
|
|
|
|
* to display initial ' '.
|
|
|
|
*/
|
|
|
|
if (!sep || !first) {
|
|
|
|
ret = scnprintf(hpp->buf, hpp->size, "%s", sep ?: " ");
|
|
|
|
advance_hpp(hpp, ret);
|
|
|
|
} else
|
|
|
|
first = false;
|
|
|
|
|
|
|
|
if (perf_hpp__use_color() && fmt->color)
|
|
|
|
ret = fmt->color(fmt, hpp, he);
|
|
|
|
else
|
|
|
|
ret = fmt->entry(fmt, hpp, he);
|
|
|
|
|
|
|
|
ret = hist_entry__snprintf_alignment(he, hpp, fmt, ret);
|
|
|
|
advance_hpp(hpp, ret);
|
|
|
|
}
|
|
|
|
|
2016-03-08 03:44:46 +08:00
|
|
|
if (!sep)
|
2016-02-24 23:13:41 +08:00
|
|
|
ret = scnprintf(hpp->buf, hpp->size, "%*s",
|
2016-03-08 03:44:48 +08:00
|
|
|
(hists->nr_hpp_node - 2) * HIERARCHY_INDENT, "");
|
2016-02-24 23:13:41 +08:00
|
|
|
advance_hpp(hpp, ret);
|
|
|
|
|
2016-02-27 02:52:45 +08:00
|
|
|
printed += fprintf(fp, "%s", buf);
|
|
|
|
|
2016-03-08 03:44:46 +08:00
|
|
|
perf_hpp_list__for_each_format(he->hpp_list, fmt) {
|
|
|
|
hpp->buf = buf;
|
|
|
|
hpp->size = size;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* No need to call hist_entry__snprintf_alignment() since this
|
|
|
|
* fmt is always the last column in the hierarchy mode.
|
|
|
|
*/
|
|
|
|
if (perf_hpp__use_color() && fmt->color)
|
|
|
|
fmt->color(fmt, hpp, he);
|
|
|
|
else
|
|
|
|
fmt->entry(fmt, hpp, he);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* dynamic entries are right-aligned but we want left-aligned
|
|
|
|
* in the hierarchy mode
|
|
|
|
*/
|
2019-06-26 22:42:03 +08:00
|
|
|
printed += fprintf(fp, "%s%s", sep ?: " ", skip_spaces(buf));
|
2016-03-08 03:44:46 +08:00
|
|
|
}
|
|
|
|
printed += putc('\n', fp);
|
2016-02-24 23:13:41 +08:00
|
|
|
|
perf hists: Check if a hist_entry has callchains before using them
So far if we use 'perf record -g' this will make
symbol_conf.use_callchain 'true' and logic will assume that all events
have callchains enabled, but ever since we added the possibility of
setting up callchains for some events (e.g.: -e
cycles/call-graph=dwarf/) while not for others, we limit usage scenarios
by looking at that symbol_conf.use_callchain global boolean, we better
look at each event attributes.
On the road to that we need to look if a hist_entry has callchains, that
is, to go from hist_entry->hists to the evsel that contains it, to then
look at evsel->sample_type for PERF_SAMPLE_CALLCHAIN.
The next step is to add a symbol_conf.ignore_callchains global, to use
in the places where what we really want to know is if callchains should
be ignored, even if present.
Then -g will mean just to select a callchain mode to be applied to all
events not explicitely setting some other callchain mode, i.e. a default
callchain mode, and --no-call-graph will set
symbol_conf.ignore_callchains with that clear intention.
That too will at some point become a per evsel thing, that tools can set
for all or just a few of its evsels.
Cc: Adrian Hunter <adrian.hunter@intel.com>
Cc: David Ahern <dsahern@gmail.com>
Cc: Jiri Olsa <jolsa@kernel.org>
Cc: Namhyung Kim <namhyung@kernel.org>
Cc: Wang Nan <wangnan0@huawei.com>
Link: https://lkml.kernel.org/n/tip-0sas5cm4dsw2obn75g7ruz69@git.kernel.org
Signed-off-by: Arnaldo Carvalho de Melo <acme@redhat.com>
2018-05-30 00:59:24 +08:00
|
|
|
if (he->leaf && hist_entry__has_callchains(he) && symbol_conf.use_callchain) {
|
2016-02-24 23:13:41 +08:00
|
|
|
u64 total = hists__total_period(hists);
|
|
|
|
|
|
|
|
printed += hist_entry_callchain__fprintf(he, total, 0, fp);
|
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
|
|
|
|
out:
|
|
|
|
return printed;
|
|
|
|
}
|
|
|
|
|
2019-06-28 17:23:03 +08:00
|
|
|
static int hist_entry__block_fprintf(struct hist_entry *he,
|
|
|
|
char *bf, size_t size,
|
|
|
|
FILE *fp)
|
|
|
|
{
|
|
|
|
struct block_hist *bh = container_of(he, struct block_hist, he);
|
|
|
|
int ret = 0;
|
|
|
|
|
|
|
|
for (unsigned int i = 0; i < bh->block_hists.nr_entries; i++) {
|
|
|
|
struct perf_hpp hpp = {
|
|
|
|
.buf = bf,
|
|
|
|
.size = size,
|
|
|
|
.skip = false,
|
|
|
|
};
|
|
|
|
|
|
|
|
bh->block_idx = i;
|
|
|
|
hist_entry__snprintf(he, &hpp);
|
|
|
|
|
|
|
|
if (!hpp.skip)
|
|
|
|
ret += fprintf(fp, "%s\n", bf);
|
|
|
|
}
|
|
|
|
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
perf report: Sort by sampled cycles percent per block for stdio
It would be useful to support sorting for all blocks by the sampled
cycles percent per block. This is useful to concentrate on the globally
hottest blocks.
This patch implements a new option "--total-cycles" which sorts all
blocks by 'Sampled Cycles%'. The 'Sampled Cycles%' is the percent:
percent = block sampled cycles aggregation / total sampled cycles
Note that, this patch only supports "--stdio" mode.
For example,
# perf record -b ./div
# perf report --total-cycles --stdio
# To display the perf.data header info, please use --header/--header-only options.
#
# Total Lost Samples: 0
#
# Samples: 2M of event 'cycles'
# Event count (approx.): 2753248
#
# Sampled Cycles% Sampled Cycles Avg Cycles% Avg Cycles [Program Block Range] Shared Object
# ............... .............. ........... .......... ................................................ .................
#
26.04% 2.8M 0.40% 18 [div.c:42 -> div.c:39] div
15.17% 1.2M 0.16% 7 [random_r.c:357 -> random_r.c:380] libc-2.27.so
5.11% 402.0K 0.04% 2 [div.c:27 -> div.c:28] div
4.87% 381.6K 0.04% 2 [random.c:288 -> random.c:291] libc-2.27.so
4.53% 381.0K 0.04% 2 [div.c:40 -> div.c:40] div
3.85% 300.9K 0.02% 1 [div.c:22 -> div.c:25] div
3.08% 241.1K 0.02% 1 [rand.c:26 -> rand.c:27] libc-2.27.so
3.06% 240.0K 0.02% 1 [random.c:291 -> random.c:291] libc-2.27.so
2.78% 215.7K 0.02% 1 [random.c:298 -> random.c:298] libc-2.27.so
2.52% 198.3K 0.02% 1 [random.c:293 -> random.c:293] libc-2.27.so
2.36% 184.8K 0.02% 1 [rand.c:28 -> rand.c:28] libc-2.27.so
2.33% 180.5K 0.02% 1 [random.c:295 -> random.c:295] libc-2.27.so
2.28% 176.7K 0.02% 1 [random.c:295 -> random.c:295] libc-2.27.so
2.20% 168.8K 0.02% 1 [rand@plt+0 -> rand@plt+0] div
1.98% 158.2K 0.02% 1 [random_r.c:388 -> random_r.c:388] libc-2.27.so
1.57% 123.3K 0.02% 1 [div.c:42 -> div.c:44] div
1.44% 116.0K 0.42% 19 [random_r.c:357 -> random_r.c:394] libc-2.27.so
0.25% 182.5K 0.02% 1 [random_r.c:388 -> random_r.c:391] libc-2.27.so
0.00% 48 1.07% 48 [x86_pmu_enable+284 -> x86_pmu_enable+298] [kernel.kallsyms]
0.00% 74 1.64% 74 [vm_mmap_pgoff+0 -> vm_mmap_pgoff+92] [kernel.kallsyms]
0.00% 73 1.62% 73 [vm_mmap+0 -> vm_mmap+48] [kernel.kallsyms]
0.00% 63 0.69% 31 [up_write+0 -> up_write+34] [kernel.kallsyms]
0.00% 13 0.29% 13 [setup_arg_pages+396 -> setup_arg_pages+413] [kernel.kallsyms]
0.00% 3 0.07% 3 [setup_arg_pages+418 -> setup_arg_pages+450] [kernel.kallsyms]
0.00% 616 6.84% 308 [security_mmap_file+0 -> security_mmap_file+72] [kernel.kallsyms]
0.00% 23 0.51% 23 [security_mmap_file+77 -> security_mmap_file+87] [kernel.kallsyms]
0.00% 4 0.02% 1 [sched_clock+0 -> sched_clock+4] [kernel.kallsyms]
0.00% 4 0.02% 1 [sched_clock+9 -> sched_clock+12] [kernel.kallsyms]
0.00% 1 0.02% 1 [rcu_nmi_exit+0 -> rcu_nmi_exit+9] [kernel.kallsyms]
Committer testing:
This should provide material for hours of endless joy, both from looking
for suspicious things in the implementation of this patch, such as the
top one:
# Sampled Cycles% Sampled Cycles Avg Cycles% Avg Cycles [Program Block Range] Shared Object
2.17% 1.7M 0.08% 607 [compiler.h:199 -> common.c:221] [kernel.vmlinux]
As well from things that look legit:
# Sampled Cycles% Sampled Cycles Avg Cycles% Avg Cycles [Program Block Range] Shared Object
0.16% 123.0K 0.60% 4.7K [nospec-branch.h:265 -> nospec-branch.h:278] [kernel.vmlinux]
:-)
Very short system wide taken branches session:
# perf record -h -b
Usage: perf record [<options>] [<command>]
or: perf record [<options>] -- <command> [<options>]
-b, --branch-any sample any taken branches
#
# perf record -b
^C[ perf record: Woken up 595 times to write data ]
[ perf record: Captured and wrote 156.672 MB perf.data (196873 samples) ]
#
# perf evlist -v
cycles: size: 112, { sample_period, sample_freq }: 4000, sample_type: IP|TID|TIME|CPU|PERIOD|BRANCH_STACK, read_format: ID, disabled: 1, inherit: 1, mmap: 1, comm: 1, freq: 1, task: 1, precise_ip: 3, sample_id_all: 1, exclude_guest: 1, mmap2: 1, comm_exec: 1, ksymbol: 1, bpf_event: 1, branch_sample_type: ANY
#
# perf report --total-cycles --stdio
# To display the perf.data header info, please use --header/--header-only options.
#
# Total Lost Samples: 0
#
# Samples: 6M of event 'cycles'
# Event count (approx.): 6299936
#
# Sampled Cycles% Sampled Cycles Avg Cycles% Avg Cycles [Program Block Range] Shared Object
# ............... .............. ........... .......... ...................................................................... ....................
#
2.17% 1.7M 0.08% 607 [compiler.h:199 -> common.c:221] [kernel.vmlinux]
1.75% 1.3M 8.34% 65.5K [memset-vec-unaligned-erms.S:147 -> memset-vec-unaligned-erms.S:151] libc-2.29.so
0.72% 544.5K 0.03% 230 [entry_64.S:657 -> entry_64.S:662] [kernel.vmlinux]
0.56% 541.8K 0.09% 672 [compiler.h:199 -> common.c:300] [kernel.vmlinux]
0.39% 293.2K 0.01% 104 [list_debug.c:43 -> list_debug.c:61] [kernel.vmlinux]
0.36% 278.6K 0.03% 272 [entry_64.S:1289 -> entry_64.S:1308] [kernel.vmlinux]
0.30% 260.8K 0.07% 564 [clear_page_64.S:47 -> clear_page_64.S:50] [kernel.vmlinux]
0.28% 215.3K 0.05% 369 [traps.c:623 -> traps.c:628] [kernel.vmlinux]
0.23% 178.1K 0.04% 278 [entry_64.S:271 -> entry_64.S:275] [kernel.vmlinux]
0.20% 152.6K 0.09% 706 [paravirt.c:177 -> paravirt.c:179] [kernel.vmlinux]
0.20% 155.8K 0.05% 373 [entry_64.S:153 -> entry_64.S:175] [kernel.vmlinux]
0.18% 136.6K 0.03% 222 [msr.h:105 -> msr.h:166] [kernel.vmlinux]
0.16% 123.0K 0.60% 4.7K [nospec-branch.h:265 -> nospec-branch.h:278] [kernel.vmlinux]
0.16% 118.3K 0.01% 44 [entry_64.S:632 -> entry_64.S:657] [kernel.vmlinux]
0.14% 104.5K 0.00% 28 [rwsem.c:1541 -> rwsem.c:1544] [kernel.vmlinux]
0.13% 99.2K 0.01% 53 [spinlock.c:150 -> spinlock.c:152] [kernel.vmlinux]
0.13% 95.5K 0.00% 35 [swap.c:456 -> swap.c:471] [kernel.vmlinux]
0.12% 96.2K 0.05% 407 [copy_user_64.S:175 -> copy_user_64.S:209] [kernel.vmlinux]
0.11% 85.9K 0.00% 31 [swap.c:400 -> page-flags.h:188] [kernel.vmlinux]
0.10% 73.0K 0.01% 52 [paravirt.h:763 -> list.h:131] [kernel.vmlinux]
0.07% 56.2K 0.03% 214 [filemap.c:1524 -> filemap.c:1557] [kernel.vmlinux]
0.07% 54.2K 0.02% 145 [memory.c:1032 -> memory.c:1049] [kernel.vmlinux]
0.07% 50.3K 0.00% 39 [mmzone.c:49 -> mmzone.c:69] [kernel.vmlinux]
0.06% 48.3K 0.01% 40 [paravirt.h:768 -> page_alloc.c:3304] [kernel.vmlinux]
0.06% 46.7K 0.02% 155 [memory.c:1032 -> memory.c:1056] [kernel.vmlinux]
0.06% 46.9K 0.01% 103 [swap.c:867 -> swap.c:902] [kernel.vmlinux]
0.06% 47.8K 0.00% 34 [entry_64.S:1201 -> entry_64.S:1202] [kernel.vmlinux]
-----------------------------------------------------------
v7:
---
Use use_browser in report__browse_block_hists for supporting
stdio and potential tui mode.
v6:
---
Create report__browse_block_hists in block-info.c (codes are
moved from builtin-report.c). It's called from
perf_evlist__tty_browse_hists.
v5:
---
1. Move all block functions to block-info.c
2. Move the code of setting ms in block hist_entry to
other patch.
v4:
---
1. Use new option '--total-cycles' to replace
'-s total_cycles' in v3.
2. Move block info collection out of block info
printing.
v3:
---
1. Use common function block_info__process_sym to
process the blocks per symbol.
2. Remove the nasty hack for skipping calculation
of column length
3. Some minor cleanup
Signed-off-by: Jin Yao <yao.jin@linux.intel.com>
Reviewed-by: Jiri Olsa <jolsa@kernel.org>
Tested-by: Arnaldo Carvalho de Melo <acme@redhat.com>
Cc: Alexander Shishkin <alexander.shishkin@linux.intel.com>
Cc: Andi Kleen <ak@linux.intel.com>
Cc: Jin Yao <yao.jin@intel.com>
Cc: Kan Liang <kan.liang@linux.intel.com>
Cc: Peter Zijlstra <peterz@infradead.org>
Link: http://lore.kernel.org/lkml/20191107074719.26139-6-yao.jin@linux.intel.com
Signed-off-by: Arnaldo Carvalho de Melo <acme@redhat.com>
2019-11-07 15:47:17 +08:00
|
|
|
static int hist_entry__individual_block_fprintf(struct hist_entry *he,
|
|
|
|
char *bf, size_t size,
|
|
|
|
FILE *fp)
|
|
|
|
{
|
|
|
|
int ret = 0;
|
|
|
|
|
|
|
|
struct perf_hpp hpp = {
|
|
|
|
.buf = bf,
|
|
|
|
.size = size,
|
|
|
|
.skip = false,
|
|
|
|
};
|
|
|
|
|
|
|
|
hist_entry__snprintf(he, &hpp);
|
|
|
|
if (!hpp.skip)
|
|
|
|
ret += fprintf(fp, "%s\n", bf);
|
|
|
|
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
2012-08-20 12:52:05 +08:00
|
|
|
static int hist_entry__fprintf(struct hist_entry *he, size_t size,
|
2016-06-15 02:19:18 +08:00
|
|
|
char *bf, size_t bfsz, FILE *fp,
|
2018-06-21 02:58:20 +08:00
|
|
|
bool ignore_callchains)
|
2012-08-20 12:52:05 +08:00
|
|
|
{
|
|
|
|
int ret;
|
2017-03-26 04:34:28 +08:00
|
|
|
int callchain_ret = 0;
|
2012-09-03 10:53:06 +08:00
|
|
|
struct perf_hpp hpp = {
|
|
|
|
.buf = bf,
|
|
|
|
.size = size,
|
|
|
|
};
|
2016-06-15 02:19:17 +08:00
|
|
|
struct hists *hists = he->hists;
|
2016-01-27 23:40:51 +08:00
|
|
|
u64 total_period = hists->stats.total_period;
|
2012-08-20 12:52:05 +08:00
|
|
|
|
2013-09-06 02:39:12 +08:00
|
|
|
if (size == 0 || size > bfsz)
|
|
|
|
size = hpp.size = bfsz;
|
2012-08-20 12:52:05 +08:00
|
|
|
|
2016-03-08 03:44:48 +08:00
|
|
|
if (symbol_conf.report_hierarchy)
|
|
|
|
return hist_entry__hierarchy_fprintf(he, &hpp, hists, fp);
|
2016-02-24 23:13:41 +08:00
|
|
|
|
2019-06-28 17:23:03 +08:00
|
|
|
if (symbol_conf.report_block)
|
|
|
|
return hist_entry__block_fprintf(he, bf, size, fp);
|
|
|
|
|
perf report: Sort by sampled cycles percent per block for stdio
It would be useful to support sorting for all blocks by the sampled
cycles percent per block. This is useful to concentrate on the globally
hottest blocks.
This patch implements a new option "--total-cycles" which sorts all
blocks by 'Sampled Cycles%'. The 'Sampled Cycles%' is the percent:
percent = block sampled cycles aggregation / total sampled cycles
Note that, this patch only supports "--stdio" mode.
For example,
# perf record -b ./div
# perf report --total-cycles --stdio
# To display the perf.data header info, please use --header/--header-only options.
#
# Total Lost Samples: 0
#
# Samples: 2M of event 'cycles'
# Event count (approx.): 2753248
#
# Sampled Cycles% Sampled Cycles Avg Cycles% Avg Cycles [Program Block Range] Shared Object
# ............... .............. ........... .......... ................................................ .................
#
26.04% 2.8M 0.40% 18 [div.c:42 -> div.c:39] div
15.17% 1.2M 0.16% 7 [random_r.c:357 -> random_r.c:380] libc-2.27.so
5.11% 402.0K 0.04% 2 [div.c:27 -> div.c:28] div
4.87% 381.6K 0.04% 2 [random.c:288 -> random.c:291] libc-2.27.so
4.53% 381.0K 0.04% 2 [div.c:40 -> div.c:40] div
3.85% 300.9K 0.02% 1 [div.c:22 -> div.c:25] div
3.08% 241.1K 0.02% 1 [rand.c:26 -> rand.c:27] libc-2.27.so
3.06% 240.0K 0.02% 1 [random.c:291 -> random.c:291] libc-2.27.so
2.78% 215.7K 0.02% 1 [random.c:298 -> random.c:298] libc-2.27.so
2.52% 198.3K 0.02% 1 [random.c:293 -> random.c:293] libc-2.27.so
2.36% 184.8K 0.02% 1 [rand.c:28 -> rand.c:28] libc-2.27.so
2.33% 180.5K 0.02% 1 [random.c:295 -> random.c:295] libc-2.27.so
2.28% 176.7K 0.02% 1 [random.c:295 -> random.c:295] libc-2.27.so
2.20% 168.8K 0.02% 1 [rand@plt+0 -> rand@plt+0] div
1.98% 158.2K 0.02% 1 [random_r.c:388 -> random_r.c:388] libc-2.27.so
1.57% 123.3K 0.02% 1 [div.c:42 -> div.c:44] div
1.44% 116.0K 0.42% 19 [random_r.c:357 -> random_r.c:394] libc-2.27.so
0.25% 182.5K 0.02% 1 [random_r.c:388 -> random_r.c:391] libc-2.27.so
0.00% 48 1.07% 48 [x86_pmu_enable+284 -> x86_pmu_enable+298] [kernel.kallsyms]
0.00% 74 1.64% 74 [vm_mmap_pgoff+0 -> vm_mmap_pgoff+92] [kernel.kallsyms]
0.00% 73 1.62% 73 [vm_mmap+0 -> vm_mmap+48] [kernel.kallsyms]
0.00% 63 0.69% 31 [up_write+0 -> up_write+34] [kernel.kallsyms]
0.00% 13 0.29% 13 [setup_arg_pages+396 -> setup_arg_pages+413] [kernel.kallsyms]
0.00% 3 0.07% 3 [setup_arg_pages+418 -> setup_arg_pages+450] [kernel.kallsyms]
0.00% 616 6.84% 308 [security_mmap_file+0 -> security_mmap_file+72] [kernel.kallsyms]
0.00% 23 0.51% 23 [security_mmap_file+77 -> security_mmap_file+87] [kernel.kallsyms]
0.00% 4 0.02% 1 [sched_clock+0 -> sched_clock+4] [kernel.kallsyms]
0.00% 4 0.02% 1 [sched_clock+9 -> sched_clock+12] [kernel.kallsyms]
0.00% 1 0.02% 1 [rcu_nmi_exit+0 -> rcu_nmi_exit+9] [kernel.kallsyms]
Committer testing:
This should provide material for hours of endless joy, both from looking
for suspicious things in the implementation of this patch, such as the
top one:
# Sampled Cycles% Sampled Cycles Avg Cycles% Avg Cycles [Program Block Range] Shared Object
2.17% 1.7M 0.08% 607 [compiler.h:199 -> common.c:221] [kernel.vmlinux]
As well from things that look legit:
# Sampled Cycles% Sampled Cycles Avg Cycles% Avg Cycles [Program Block Range] Shared Object
0.16% 123.0K 0.60% 4.7K [nospec-branch.h:265 -> nospec-branch.h:278] [kernel.vmlinux]
:-)
Very short system wide taken branches session:
# perf record -h -b
Usage: perf record [<options>] [<command>]
or: perf record [<options>] -- <command> [<options>]
-b, --branch-any sample any taken branches
#
# perf record -b
^C[ perf record: Woken up 595 times to write data ]
[ perf record: Captured and wrote 156.672 MB perf.data (196873 samples) ]
#
# perf evlist -v
cycles: size: 112, { sample_period, sample_freq }: 4000, sample_type: IP|TID|TIME|CPU|PERIOD|BRANCH_STACK, read_format: ID, disabled: 1, inherit: 1, mmap: 1, comm: 1, freq: 1, task: 1, precise_ip: 3, sample_id_all: 1, exclude_guest: 1, mmap2: 1, comm_exec: 1, ksymbol: 1, bpf_event: 1, branch_sample_type: ANY
#
# perf report --total-cycles --stdio
# To display the perf.data header info, please use --header/--header-only options.
#
# Total Lost Samples: 0
#
# Samples: 6M of event 'cycles'
# Event count (approx.): 6299936
#
# Sampled Cycles% Sampled Cycles Avg Cycles% Avg Cycles [Program Block Range] Shared Object
# ............... .............. ........... .......... ...................................................................... ....................
#
2.17% 1.7M 0.08% 607 [compiler.h:199 -> common.c:221] [kernel.vmlinux]
1.75% 1.3M 8.34% 65.5K [memset-vec-unaligned-erms.S:147 -> memset-vec-unaligned-erms.S:151] libc-2.29.so
0.72% 544.5K 0.03% 230 [entry_64.S:657 -> entry_64.S:662] [kernel.vmlinux]
0.56% 541.8K 0.09% 672 [compiler.h:199 -> common.c:300] [kernel.vmlinux]
0.39% 293.2K 0.01% 104 [list_debug.c:43 -> list_debug.c:61] [kernel.vmlinux]
0.36% 278.6K 0.03% 272 [entry_64.S:1289 -> entry_64.S:1308] [kernel.vmlinux]
0.30% 260.8K 0.07% 564 [clear_page_64.S:47 -> clear_page_64.S:50] [kernel.vmlinux]
0.28% 215.3K 0.05% 369 [traps.c:623 -> traps.c:628] [kernel.vmlinux]
0.23% 178.1K 0.04% 278 [entry_64.S:271 -> entry_64.S:275] [kernel.vmlinux]
0.20% 152.6K 0.09% 706 [paravirt.c:177 -> paravirt.c:179] [kernel.vmlinux]
0.20% 155.8K 0.05% 373 [entry_64.S:153 -> entry_64.S:175] [kernel.vmlinux]
0.18% 136.6K 0.03% 222 [msr.h:105 -> msr.h:166] [kernel.vmlinux]
0.16% 123.0K 0.60% 4.7K [nospec-branch.h:265 -> nospec-branch.h:278] [kernel.vmlinux]
0.16% 118.3K 0.01% 44 [entry_64.S:632 -> entry_64.S:657] [kernel.vmlinux]
0.14% 104.5K 0.00% 28 [rwsem.c:1541 -> rwsem.c:1544] [kernel.vmlinux]
0.13% 99.2K 0.01% 53 [spinlock.c:150 -> spinlock.c:152] [kernel.vmlinux]
0.13% 95.5K 0.00% 35 [swap.c:456 -> swap.c:471] [kernel.vmlinux]
0.12% 96.2K 0.05% 407 [copy_user_64.S:175 -> copy_user_64.S:209] [kernel.vmlinux]
0.11% 85.9K 0.00% 31 [swap.c:400 -> page-flags.h:188] [kernel.vmlinux]
0.10% 73.0K 0.01% 52 [paravirt.h:763 -> list.h:131] [kernel.vmlinux]
0.07% 56.2K 0.03% 214 [filemap.c:1524 -> filemap.c:1557] [kernel.vmlinux]
0.07% 54.2K 0.02% 145 [memory.c:1032 -> memory.c:1049] [kernel.vmlinux]
0.07% 50.3K 0.00% 39 [mmzone.c:49 -> mmzone.c:69] [kernel.vmlinux]
0.06% 48.3K 0.01% 40 [paravirt.h:768 -> page_alloc.c:3304] [kernel.vmlinux]
0.06% 46.7K 0.02% 155 [memory.c:1032 -> memory.c:1056] [kernel.vmlinux]
0.06% 46.9K 0.01% 103 [swap.c:867 -> swap.c:902] [kernel.vmlinux]
0.06% 47.8K 0.00% 34 [entry_64.S:1201 -> entry_64.S:1202] [kernel.vmlinux]
-----------------------------------------------------------
v7:
---
Use use_browser in report__browse_block_hists for supporting
stdio and potential tui mode.
v6:
---
Create report__browse_block_hists in block-info.c (codes are
moved from builtin-report.c). It's called from
perf_evlist__tty_browse_hists.
v5:
---
1. Move all block functions to block-info.c
2. Move the code of setting ms in block hist_entry to
other patch.
v4:
---
1. Use new option '--total-cycles' to replace
'-s total_cycles' in v3.
2. Move block info collection out of block info
printing.
v3:
---
1. Use common function block_info__process_sym to
process the blocks per symbol.
2. Remove the nasty hack for skipping calculation
of column length
3. Some minor cleanup
Signed-off-by: Jin Yao <yao.jin@linux.intel.com>
Reviewed-by: Jiri Olsa <jolsa@kernel.org>
Tested-by: Arnaldo Carvalho de Melo <acme@redhat.com>
Cc: Alexander Shishkin <alexander.shishkin@linux.intel.com>
Cc: Andi Kleen <ak@linux.intel.com>
Cc: Jin Yao <yao.jin@intel.com>
Cc: Kan Liang <kan.liang@linux.intel.com>
Cc: Peter Zijlstra <peterz@infradead.org>
Link: http://lore.kernel.org/lkml/20191107074719.26139-6-yao.jin@linux.intel.com
Signed-off-by: Arnaldo Carvalho de Melo <acme@redhat.com>
2019-11-07 15:47:17 +08:00
|
|
|
if (symbol_conf.report_individual_block)
|
|
|
|
return hist_entry__individual_block_fprintf(he, bf, size, fp);
|
|
|
|
|
2014-03-03 15:16:20 +08:00
|
|
|
hist_entry__snprintf(he, &hpp);
|
2012-08-20 12:52:05 +08:00
|
|
|
|
2012-08-20 12:52:06 +08:00
|
|
|
ret = fprintf(fp, "%s\n", bf);
|
2012-08-20 12:52:05 +08:00
|
|
|
|
2018-06-21 02:58:20 +08:00
|
|
|
if (hist_entry__has_callchains(he) && !ignore_callchains)
|
2017-03-26 04:34:28 +08:00
|
|
|
callchain_ret = hist_entry_callchain__fprintf(he, total_period,
|
|
|
|
0, fp);
|
|
|
|
|
2017-10-10 04:32:55 +08:00
|
|
|
ret += callchain_ret;
|
2012-08-20 12:52:05 +08:00
|
|
|
|
2012-08-20 12:52:06 +08:00
|
|
|
return ret;
|
2012-08-20 12:52:05 +08:00
|
|
|
}
|
|
|
|
|
2016-03-08 03:44:48 +08:00
|
|
|
static int print_hierarchy_indent(const char *sep, int indent,
|
2016-02-24 23:13:42 +08:00
|
|
|
const char *line, FILE *fp)
|
|
|
|
{
|
2019-06-26 03:24:20 +08:00
|
|
|
int width;
|
|
|
|
|
2016-03-08 03:44:48 +08:00
|
|
|
if (sep != NULL || indent < 2)
|
2016-02-24 23:13:42 +08:00
|
|
|
return 0;
|
|
|
|
|
2019-06-26 03:24:20 +08:00
|
|
|
width = (indent - 2) * HIERARCHY_INDENT;
|
|
|
|
|
|
|
|
return fprintf(fp, "%-*.*s", width, width, line);
|
2016-02-24 23:13:42 +08:00
|
|
|
}
|
|
|
|
|
2016-09-13 15:45:50 +08:00
|
|
|
static int hists__fprintf_hierarchy_headers(struct hists *hists,
|
|
|
|
struct perf_hpp *hpp, FILE *fp)
|
2016-02-24 23:13:42 +08:00
|
|
|
{
|
2016-03-08 03:44:49 +08:00
|
|
|
bool first_node, first_col;
|
2016-03-08 03:44:48 +08:00
|
|
|
int indent;
|
2016-02-27 02:52:45 +08:00
|
|
|
int depth;
|
2016-02-24 23:13:42 +08:00
|
|
|
unsigned width = 0;
|
|
|
|
unsigned header_width = 0;
|
|
|
|
struct perf_hpp_fmt *fmt;
|
2016-03-08 03:44:49 +08:00
|
|
|
struct perf_hpp_list_node *fmt_node;
|
2016-09-13 15:45:50 +08:00
|
|
|
const char *sep = symbol_conf.field_sep;
|
2016-02-24 23:13:42 +08:00
|
|
|
|
2016-03-08 03:44:48 +08:00
|
|
|
indent = hists->nr_hpp_node;
|
2016-02-24 23:13:42 +08:00
|
|
|
|
|
|
|
/* preserve max indent depth for column headers */
|
2019-06-26 03:24:20 +08:00
|
|
|
print_hierarchy_indent(sep, indent, " ", fp);
|
2016-02-24 23:13:42 +08:00
|
|
|
|
2016-03-08 03:44:49 +08:00
|
|
|
/* the first hpp_list_node is for overhead columns */
|
|
|
|
fmt_node = list_first_entry(&hists->hpp_formats,
|
|
|
|
struct perf_hpp_list_node, list);
|
2016-02-24 23:13:42 +08:00
|
|
|
|
2016-03-08 03:44:49 +08:00
|
|
|
perf_hpp_list__for_each_format(&fmt_node->hpp, fmt) {
|
2016-08-07 23:28:30 +08:00
|
|
|
fmt->header(fmt, hpp, hists, 0, NULL);
|
2016-03-08 03:44:49 +08:00
|
|
|
fprintf(fp, "%s%s", hpp->buf, sep ?: " ");
|
2016-02-24 23:13:42 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
/* combine sort headers with ' / ' */
|
2016-03-08 03:44:49 +08:00
|
|
|
first_node = true;
|
|
|
|
list_for_each_entry_continue(fmt_node, &hists->hpp_formats, list) {
|
|
|
|
if (!first_node)
|
2016-02-24 23:13:42 +08:00
|
|
|
header_width += fprintf(fp, " / ");
|
2016-03-08 03:44:49 +08:00
|
|
|
first_node = false;
|
2016-02-24 23:13:42 +08:00
|
|
|
|
2016-03-08 03:44:49 +08:00
|
|
|
first_col = true;
|
|
|
|
perf_hpp_list__for_each_format(&fmt_node->hpp, fmt) {
|
|
|
|
if (perf_hpp__should_skip(fmt, hists))
|
|
|
|
continue;
|
|
|
|
|
|
|
|
if (!first_col)
|
|
|
|
header_width += fprintf(fp, "+");
|
|
|
|
first_col = false;
|
|
|
|
|
2016-08-07 23:28:30 +08:00
|
|
|
fmt->header(fmt, hpp, hists, 0, NULL);
|
2016-02-24 23:13:42 +08:00
|
|
|
|
2019-06-26 23:06:20 +08:00
|
|
|
header_width += fprintf(fp, "%s", strim(hpp->buf));
|
2016-03-08 03:44:49 +08:00
|
|
|
}
|
2016-02-24 23:13:42 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
fprintf(fp, "\n# ");
|
|
|
|
|
|
|
|
/* preserve max indent depth for initial dots */
|
2016-03-08 03:44:48 +08:00
|
|
|
print_hierarchy_indent(sep, indent, dots, fp);
|
2016-02-24 23:13:42 +08:00
|
|
|
|
2016-03-08 03:44:49 +08:00
|
|
|
/* the first hpp_list_node is for overhead columns */
|
|
|
|
fmt_node = list_first_entry(&hists->hpp_formats,
|
|
|
|
struct perf_hpp_list_node, list);
|
2016-02-24 23:13:42 +08:00
|
|
|
|
2016-03-08 03:44:49 +08:00
|
|
|
first_col = true;
|
|
|
|
perf_hpp_list__for_each_format(&fmt_node->hpp, fmt) {
|
|
|
|
if (!first_col)
|
|
|
|
fprintf(fp, "%s", sep ?: "..");
|
|
|
|
first_col = false;
|
2016-02-24 23:13:42 +08:00
|
|
|
|
2016-06-15 02:19:20 +08:00
|
|
|
width = fmt->width(fmt, hpp, hists);
|
2016-02-24 23:13:42 +08:00
|
|
|
fprintf(fp, "%.*s", width, dots);
|
|
|
|
}
|
|
|
|
|
2016-02-27 02:52:45 +08:00
|
|
|
depth = 0;
|
2016-03-08 03:44:49 +08:00
|
|
|
list_for_each_entry_continue(fmt_node, &hists->hpp_formats, list) {
|
|
|
|
first_col = true;
|
|
|
|
width = depth * HIERARCHY_INDENT;
|
2016-02-24 23:13:42 +08:00
|
|
|
|
2016-03-08 03:44:49 +08:00
|
|
|
perf_hpp_list__for_each_format(&fmt_node->hpp, fmt) {
|
|
|
|
if (perf_hpp__should_skip(fmt, hists))
|
|
|
|
continue;
|
|
|
|
|
|
|
|
if (!first_col)
|
|
|
|
width++; /* for '+' sign between column header */
|
|
|
|
first_col = false;
|
|
|
|
|
2016-06-15 02:19:20 +08:00
|
|
|
width += fmt->width(fmt, hpp, hists);
|
2016-03-08 03:44:49 +08:00
|
|
|
}
|
2016-02-27 02:52:45 +08:00
|
|
|
|
2016-02-24 23:13:42 +08:00
|
|
|
if (width > header_width)
|
|
|
|
header_width = width;
|
2016-02-27 02:52:45 +08:00
|
|
|
|
|
|
|
depth++;
|
2016-02-24 23:13:42 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
fprintf(fp, "%s%-.*s", sep ?: " ", header_width, dots);
|
|
|
|
|
|
|
|
fprintf(fp, "\n#\n");
|
|
|
|
|
|
|
|
return 2;
|
|
|
|
}
|
|
|
|
|
2016-08-07 23:28:29 +08:00
|
|
|
static void fprintf_line(struct hists *hists, struct perf_hpp *hpp,
|
|
|
|
int line, FILE *fp)
|
2012-08-20 12:52:05 +08:00
|
|
|
{
|
2012-10-13 06:06:16 +08:00
|
|
|
struct perf_hpp_fmt *fmt;
|
2012-08-20 12:52:05 +08:00
|
|
|
const char *sep = symbol_conf.field_sep;
|
2012-10-04 20:49:37 +08:00
|
|
|
bool first = true;
|
2016-08-07 23:28:30 +08:00
|
|
|
int span = 0;
|
2014-03-03 15:16:20 +08:00
|
|
|
|
2016-01-18 17:24:23 +08:00
|
|
|
hists__for_each_format(hists, fmt) {
|
2015-12-23 01:07:08 +08:00
|
|
|
if (perf_hpp__should_skip(fmt, hists))
|
2014-03-18 12:00:59 +08:00
|
|
|
continue;
|
|
|
|
|
2016-08-07 23:28:30 +08:00
|
|
|
if (!first && !span)
|
2014-03-03 15:16:20 +08:00
|
|
|
fprintf(fp, "%s", sep ?: " ");
|
|
|
|
else
|
|
|
|
first = false;
|
|
|
|
|
2016-08-07 23:28:30 +08:00
|
|
|
fmt->header(fmt, hpp, hists, line, &span);
|
|
|
|
|
|
|
|
if (!span)
|
|
|
|
fprintf(fp, "%s", hpp->buf);
|
2012-08-20 12:52:05 +08:00
|
|
|
}
|
2016-08-07 23:28:29 +08:00
|
|
|
}
|
2012-08-20 12:52:05 +08:00
|
|
|
|
2016-08-07 23:28:29 +08:00
|
|
|
static int
|
|
|
|
hists__fprintf_standard_headers(struct hists *hists,
|
|
|
|
struct perf_hpp *hpp,
|
|
|
|
FILE *fp)
|
|
|
|
{
|
|
|
|
struct perf_hpp_list *hpp_list = hists->hpp_list;
|
|
|
|
struct perf_hpp_fmt *fmt;
|
|
|
|
unsigned int width;
|
|
|
|
const char *sep = symbol_conf.field_sep;
|
|
|
|
bool first = true;
|
|
|
|
int line;
|
|
|
|
|
|
|
|
for (line = 0; line < hpp_list->nr_header_lines; line++) {
|
|
|
|
/* first # is displayed one level up */
|
|
|
|
if (line)
|
|
|
|
fprintf(fp, "# ");
|
|
|
|
fprintf_line(hists, hpp, line, fp);
|
|
|
|
fprintf(fp, "\n");
|
|
|
|
}
|
2012-08-20 12:52:05 +08:00
|
|
|
|
|
|
|
if (sep)
|
2016-08-07 23:28:29 +08:00
|
|
|
return hpp_list->nr_header_lines;
|
2012-08-20 12:52:05 +08:00
|
|
|
|
2012-10-04 20:49:37 +08:00
|
|
|
first = true;
|
|
|
|
|
2012-09-03 10:53:06 +08:00
|
|
|
fprintf(fp, "# ");
|
|
|
|
|
2016-01-18 17:24:23 +08:00
|
|
|
hists__for_each_format(hists, fmt) {
|
2012-10-13 06:06:16 +08:00
|
|
|
unsigned int i;
|
2012-09-03 10:53:06 +08:00
|
|
|
|
2015-12-23 01:07:08 +08:00
|
|
|
if (perf_hpp__should_skip(fmt, hists))
|
2014-03-18 12:00:59 +08:00
|
|
|
continue;
|
|
|
|
|
2012-10-04 20:49:37 +08:00
|
|
|
if (!first)
|
2012-09-03 10:53:06 +08:00
|
|
|
fprintf(fp, "%s", sep ?: " ");
|
2012-10-04 20:49:37 +08:00
|
|
|
else
|
|
|
|
first = false;
|
2012-09-03 10:53:06 +08:00
|
|
|
|
2016-06-15 02:19:20 +08:00
|
|
|
width = fmt->width(fmt, hpp, hists);
|
2012-09-03 10:53:06 +08:00
|
|
|
for (i = 0; i < width; i++)
|
|
|
|
fprintf(fp, ".");
|
2012-08-20 12:52:05 +08:00
|
|
|
}
|
2012-09-03 10:53:06 +08:00
|
|
|
|
2012-08-20 12:52:05 +08:00
|
|
|
fprintf(fp, "\n");
|
|
|
|
fprintf(fp, "#\n");
|
2016-08-07 23:28:29 +08:00
|
|
|
return hpp_list->nr_header_lines + 2;
|
2016-06-15 02:19:14 +08:00
|
|
|
}
|
|
|
|
|
2016-09-22 23:36:37 +08:00
|
|
|
int hists__fprintf_headers(struct hists *hists, FILE *fp)
|
2016-06-15 02:19:16 +08:00
|
|
|
{
|
2016-09-19 21:09:13 +08:00
|
|
|
char bf[1024];
|
2016-06-15 02:19:16 +08:00
|
|
|
struct perf_hpp dummy_hpp = {
|
|
|
|
.buf = bf,
|
|
|
|
.size = sizeof(bf),
|
|
|
|
};
|
|
|
|
|
|
|
|
fprintf(fp, "# ");
|
|
|
|
|
|
|
|
if (symbol_conf.report_hierarchy)
|
|
|
|
return hists__fprintf_hierarchy_headers(hists, &dummy_hpp, fp);
|
|
|
|
else
|
|
|
|
return hists__fprintf_standard_headers(hists, &dummy_hpp, fp);
|
|
|
|
|
|
|
|
}
|
|
|
|
|
2016-06-15 02:19:14 +08:00
|
|
|
size_t hists__fprintf(struct hists *hists, bool show_header, int max_rows,
|
2016-06-15 02:19:18 +08:00
|
|
|
int max_cols, float min_pcnt, FILE *fp,
|
2018-06-21 02:58:20 +08:00
|
|
|
bool ignore_callchains)
|
2016-06-15 02:19:14 +08:00
|
|
|
{
|
|
|
|
struct rb_node *nd;
|
|
|
|
size_t ret = 0;
|
|
|
|
const char *sep = symbol_conf.field_sep;
|
|
|
|
int nr_rows = 0;
|
|
|
|
size_t linesz;
|
|
|
|
char *line = NULL;
|
|
|
|
unsigned indent;
|
|
|
|
|
|
|
|
init_rem_hits();
|
|
|
|
|
2016-09-20 13:30:24 +08:00
|
|
|
hists__reset_column_width(hists);
|
2016-06-15 02:19:14 +08:00
|
|
|
|
|
|
|
if (symbol_conf.col_width_list_str)
|
|
|
|
perf_hpp__set_user_width(symbol_conf.col_width_list_str);
|
|
|
|
|
|
|
|
if (show_header)
|
|
|
|
nr_rows += hists__fprintf_headers(hists, fp);
|
|
|
|
|
|
|
|
if (max_rows && nr_rows >= max_rows)
|
2012-08-20 12:52:05 +08:00
|
|
|
goto out;
|
|
|
|
|
2013-09-06 02:39:12 +08:00
|
|
|
linesz = hists__sort_list_width(hists) + 3 + 1;
|
2013-10-25 19:24:53 +08:00
|
|
|
linesz += perf_hpp__color_overhead();
|
2013-09-06 02:39:12 +08:00
|
|
|
line = malloc(linesz);
|
|
|
|
if (line == NULL) {
|
|
|
|
ret = -1;
|
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
|
2016-02-26 20:13:17 +08:00
|
|
|
indent = hists__overhead_width(hists) + 4;
|
|
|
|
|
2018-12-07 03:18:18 +08:00
|
|
|
for (nd = rb_first_cached(&hists->entries); nd;
|
|
|
|
nd = __rb_hierarchy_next(nd, HMD_FORCE_CHILD)) {
|
2012-08-20 12:52:05 +08:00
|
|
|
struct hist_entry *h = rb_entry(nd, struct hist_entry, rb_node);
|
2013-10-31 09:17:39 +08:00
|
|
|
float percent;
|
2012-08-20 12:52:05 +08:00
|
|
|
|
|
|
|
if (h->filtered)
|
|
|
|
continue;
|
|
|
|
|
2019-11-07 15:47:18 +08:00
|
|
|
if (symbol_conf.report_individual_block)
|
|
|
|
percent = block_info__total_cycles_percent(h);
|
|
|
|
else
|
|
|
|
percent = hist_entry__get_percent_limit(h);
|
|
|
|
|
2013-05-14 10:09:04 +08:00
|
|
|
if (percent < min_pcnt)
|
|
|
|
continue;
|
|
|
|
|
2018-06-21 02:58:20 +08:00
|
|
|
ret += hist_entry__fprintf(h, max_cols, line, linesz, fp, ignore_callchains);
|
2012-08-20 12:52:05 +08:00
|
|
|
|
|
|
|
if (max_rows && ++nr_rows >= max_rows)
|
2013-09-06 02:39:12 +08:00
|
|
|
break;
|
2012-08-20 12:52:05 +08:00
|
|
|
|
2016-02-26 20:13:17 +08:00
|
|
|
/*
|
|
|
|
* If all children are filtered out or percent-limited,
|
|
|
|
* display "no entry >= x.xx%" message.
|
|
|
|
*/
|
|
|
|
if (!h->leaf && !hist_entry__has_hierarchy_children(h, min_pcnt)) {
|
2016-03-08 03:44:49 +08:00
|
|
|
int depth = hists->nr_hpp_node + h->depth + 1;
|
2016-02-26 20:13:17 +08:00
|
|
|
|
2019-06-26 03:24:20 +08:00
|
|
|
print_hierarchy_indent(sep, depth, " ", fp);
|
2016-02-26 20:13:17 +08:00
|
|
|
fprintf(fp, "%*sno entry >= %.2f%%\n", indent, "", min_pcnt);
|
|
|
|
|
|
|
|
if (max_rows && ++nr_rows >= max_rows)
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
2012-08-20 12:52:05 +08:00
|
|
|
if (h->ms.map == NULL && verbose > 1) {
|
2019-11-26 09:07:43 +08:00
|
|
|
maps__fprintf(h->thread->maps, fp);
|
2012-08-20 12:52:05 +08:00
|
|
|
fprintf(fp, "%.10s end\n", graph_dotted_line);
|
|
|
|
}
|
|
|
|
}
|
2013-09-06 02:39:12 +08:00
|
|
|
|
|
|
|
free(line);
|
2012-08-20 12:52:05 +08:00
|
|
|
out:
|
2013-12-28 03:55:14 +08:00
|
|
|
zfree(&rem_sq_bracket);
|
2012-08-20 12:52:05 +08:00
|
|
|
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
2012-12-19 03:02:17 +08:00
|
|
|
size_t events_stats__fprintf(struct events_stats *stats, FILE *fp)
|
2012-08-20 12:52:05 +08:00
|
|
|
{
|
|
|
|
int i;
|
|
|
|
size_t ret = 0;
|
|
|
|
|
|
|
|
for (i = 0; i < PERF_RECORD_HEADER_MAX; ++i) {
|
|
|
|
const char *name;
|
|
|
|
|
|
|
|
name = perf_event__name(i);
|
|
|
|
if (!strcmp(name, "UNKNOWN"))
|
|
|
|
continue;
|
|
|
|
|
2018-03-07 23:24:30 +08:00
|
|
|
ret += fprintf(fp, "%16s events: %10d\n", name, stats->nr_events[i]);
|
2012-08-20 12:52:05 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
return ret;
|
|
|
|
}
|