* [PATCH 01/12] perf tools: Separate generic code in dso__data_file_size
2019-04-16 16:01 [PATCH 00/12] perf tools: Display eBPF code in intel_pt trace Jiri Olsa
@ 2019-04-16 16:01 ` Jiri Olsa
2019-04-16 16:01 ` [PATCH 02/12] perf tools: Separate generic code in dso_cache__read Jiri Olsa
` (10 subsequent siblings)
11 siblings, 0 replies; 31+ messages in thread
From: Jiri Olsa @ 2019-04-16 16:01 UTC (permalink / raw)
To: Arnaldo Carvalho de Melo
Cc: lkml, Ingo Molnar, Namhyung Kim, Alexander Shishkin,
Peter Zijlstra, Andi Kleen, Adrian Hunter, Song Liu,
Alexei Starovoitov, Daniel Borkmann
Moving file specific code in dso__data_file_size function
into separate file_size function. I'll add bpf specific
code in following patches.
Link: http://lkml.kernel.org/n/tip-rkcsft4a0f8sw33p67llxf0d@git.kernel.org
Signed-off-by: Jiri Olsa <jolsa@kernel.org>
---
tools/perf/util/dso.c | 19 ++++++++++++-------
1 file changed, 12 insertions(+), 7 deletions(-)
diff --git a/tools/perf/util/dso.c b/tools/perf/util/dso.c
index e059976d9d93..cb6199c1390a 100644
--- a/tools/perf/util/dso.c
+++ b/tools/perf/util/dso.c
@@ -898,18 +898,12 @@ static ssize_t cached_read(struct dso *dso, struct machine *machine,
return r;
}
-int dso__data_file_size(struct dso *dso, struct machine *machine)
+static int file_size(struct dso *dso, struct machine *machine)
{
int ret = 0;
struct stat st;
char sbuf[STRERR_BUFSIZE];
- if (dso->data.file_size)
- return 0;
-
- if (dso->data.status == DSO_DATA_STATUS_ERROR)
- return -1;
-
pthread_mutex_lock(&dso__data_open_lock);
/*
@@ -938,6 +932,17 @@ int dso__data_file_size(struct dso *dso, struct machine *machine)
return ret;
}
+int dso__data_file_size(struct dso *dso, struct machine *machine)
+{
+ if (dso->data.file_size)
+ return 0;
+
+ if (dso->data.status == DSO_DATA_STATUS_ERROR)
+ return -1;
+
+ return file_size(dso, machine);
+}
+
/**
* dso__data_size - Return dso data size
* @dso: dso object
--
2.17.2
^ permalink raw reply related [flat|nested] 31+ messages in thread
* [PATCH 02/12] perf tools: Separate generic code in dso_cache__read
2019-04-16 16:01 [PATCH 00/12] perf tools: Display eBPF code in intel_pt trace Jiri Olsa
2019-04-16 16:01 ` [PATCH 01/12] perf tools: Separate generic code in dso__data_file_size Jiri Olsa
@ 2019-04-16 16:01 ` Jiri Olsa
2019-04-16 17:17 ` Stanislav Fomichev
2019-04-16 16:01 ` [PATCH 03/12] perf tools: Simplify dso_cache__read function Jiri Olsa
` (9 subsequent siblings)
11 siblings, 1 reply; 31+ messages in thread
From: Jiri Olsa @ 2019-04-16 16:01 UTC (permalink / raw)
To: Arnaldo Carvalho de Melo
Cc: lkml, Ingo Molnar, Namhyung Kim, Alexander Shishkin,
Peter Zijlstra, Andi Kleen, Adrian Hunter, Song Liu,
Alexei Starovoitov, Daniel Borkmann
Moving file specific code in dso_cache__read function
into separate file_read function. I'll add bpf specific
code in following patches.
Link: http://lkml.kernel.org/n/tip-7f7d717uzrqt5ka2xp29ige3@git.kernel.org
Signed-off-by: Jiri Olsa <jolsa@kernel.org>
---
tools/perf/util/dso.c | 47 ++++++++++++++++++++++++-------------------
1 file changed, 26 insertions(+), 21 deletions(-)
diff --git a/tools/perf/util/dso.c b/tools/perf/util/dso.c
index cb6199c1390a..6baad22ec8a9 100644
--- a/tools/perf/util/dso.c
+++ b/tools/perf/util/dso.c
@@ -794,6 +794,30 @@ dso_cache__memcpy(struct dso_cache *cache, u64 offset,
return cache_size;
}
+static ssize_t file_read(struct dso *dso, struct machine *machine,
+ u64 offset, char *data)
+{
+ ssize_t ret;
+
+ pthread_mutex_lock(&dso__data_open_lock);
+
+ /*
+ * dso->data.fd might be closed if other thread opened another
+ * file (dso) due to open file limit (RLIMIT_NOFILE).
+ */
+ try_to_open_dso(dso, machine);
+
+ if (dso->data.fd < 0) {
+ dso->data.status = DSO_DATA_STATUS_ERROR;
+ return -errno;
+ }
+
+ ret = pread(dso->data.fd, data, DSO__DATA_CACHE_SIZE, offset);
+ pthread_mutex_unlock(&dso__data_open_lock);
+
+ return ret;
+}
+
static ssize_t
dso_cache__read(struct dso *dso, struct machine *machine,
u64 offset, u8 *data, ssize_t size)
@@ -803,37 +827,18 @@ dso_cache__read(struct dso *dso, struct machine *machine,
ssize_t ret;
do {
- u64 cache_offset;
+ u64 cache_offset = offset & DSO__DATA_CACHE_MASK;
cache = zalloc(sizeof(*cache) + DSO__DATA_CACHE_SIZE);
if (!cache)
return -ENOMEM;
- pthread_mutex_lock(&dso__data_open_lock);
-
- /*
- * dso->data.fd might be closed if other thread opened another
- * file (dso) due to open file limit (RLIMIT_NOFILE).
- */
- try_to_open_dso(dso, machine);
-
- if (dso->data.fd < 0) {
- ret = -errno;
- dso->data.status = DSO_DATA_STATUS_ERROR;
- break;
- }
-
- cache_offset = offset & DSO__DATA_CACHE_MASK;
-
- ret = pread(dso->data.fd, cache->data, DSO__DATA_CACHE_SIZE, cache_offset);
- if (ret <= 0)
- break;
+ ret = file_read(dso, machine, cache_offset, cache->data);
cache->offset = cache_offset;
cache->size = ret;
} while (0);
- pthread_mutex_unlock(&dso__data_open_lock);
if (ret > 0) {
old = dso_cache__insert(dso, cache);
--
2.17.2
^ permalink raw reply related [flat|nested] 31+ messages in thread
* Re: [PATCH 02/12] perf tools: Separate generic code in dso_cache__read
2019-04-16 16:01 ` [PATCH 02/12] perf tools: Separate generic code in dso_cache__read Jiri Olsa
@ 2019-04-16 17:17 ` Stanislav Fomichev
2019-04-16 18:21 ` Jiri Olsa
0 siblings, 1 reply; 31+ messages in thread
From: Stanislav Fomichev @ 2019-04-16 17:17 UTC (permalink / raw)
To: Jiri Olsa
Cc: Arnaldo Carvalho de Melo, lkml, Ingo Molnar, Namhyung Kim,
Alexander Shishkin, Peter Zijlstra, Andi Kleen, Adrian Hunter,
Song Liu, Alexei Starovoitov, Daniel Borkmann
On 04/16, Jiri Olsa wrote:
> Moving file specific code in dso_cache__read function
> into separate file_read function. I'll add bpf specific
> code in following patches.
>
> Link: http://lkml.kernel.org/n/tip-7f7d717uzrqt5ka2xp29ige3@git.kernel.org
> Signed-off-by: Jiri Olsa <jolsa@kernel.org>
> ---
> tools/perf/util/dso.c | 47 ++++++++++++++++++++++++-------------------
> 1 file changed, 26 insertions(+), 21 deletions(-)
>
> diff --git a/tools/perf/util/dso.c b/tools/perf/util/dso.c
> index cb6199c1390a..6baad22ec8a9 100644
> --- a/tools/perf/util/dso.c
> +++ b/tools/perf/util/dso.c
> @@ -794,6 +794,30 @@ dso_cache__memcpy(struct dso_cache *cache, u64 offset,
> return cache_size;
> }
>
> +static ssize_t file_read(struct dso *dso, struct machine *machine,
> + u64 offset, char *data)
> +{
> + ssize_t ret;
> +
> + pthread_mutex_lock(&dso__data_open_lock);
> +
> + /*
> + * dso->data.fd might be closed if other thread opened another
> + * file (dso) due to open file limit (RLIMIT_NOFILE).
> + */
> + try_to_open_dso(dso, machine);
> +
> + if (dso->data.fd < 0) {
> + dso->data.status = DSO_DATA_STATUS_ERROR;
pthread_mutex_unlock(&dso__data_open_lock) here?
> + return -errno;
> + }
> +
> + ret = pread(dso->data.fd, data, DSO__DATA_CACHE_SIZE, offset);
> + pthread_mutex_unlock(&dso__data_open_lock);
> +
> + return ret;
> +}
> +
> static ssize_t
> dso_cache__read(struct dso *dso, struct machine *machine,
> u64 offset, u8 *data, ssize_t size)
> @@ -803,37 +827,18 @@ dso_cache__read(struct dso *dso, struct machine *machine,
> ssize_t ret;
>
> do {
> - u64 cache_offset;
> + u64 cache_offset = offset & DSO__DATA_CACHE_MASK;
>
> cache = zalloc(sizeof(*cache) + DSO__DATA_CACHE_SIZE);
> if (!cache)
> return -ENOMEM;
>
> - pthread_mutex_lock(&dso__data_open_lock);
> -
> - /*
> - * dso->data.fd might be closed if other thread opened another
> - * file (dso) due to open file limit (RLIMIT_NOFILE).
> - */
> - try_to_open_dso(dso, machine);
> -
> - if (dso->data.fd < 0) {
> - ret = -errno;
> - dso->data.status = DSO_DATA_STATUS_ERROR;
> - break;
> - }
> -
> - cache_offset = offset & DSO__DATA_CACHE_MASK;
> -
> - ret = pread(dso->data.fd, cache->data, DSO__DATA_CACHE_SIZE, cache_offset);
> - if (ret <= 0)
> - break;
> + ret = file_read(dso, machine, cache_offset, cache->data);
>
> cache->offset = cache_offset;
> cache->size = ret;
> } while (0);
>
> - pthread_mutex_unlock(&dso__data_open_lock);
>
> if (ret > 0) {
> old = dso_cache__insert(dso, cache);
> --
> 2.17.2
>
^ permalink raw reply [flat|nested] 31+ messages in thread
* Re: [PATCH 02/12] perf tools: Separate generic code in dso_cache__read
2019-04-16 17:17 ` Stanislav Fomichev
@ 2019-04-16 18:21 ` Jiri Olsa
0 siblings, 0 replies; 31+ messages in thread
From: Jiri Olsa @ 2019-04-16 18:21 UTC (permalink / raw)
To: Stanislav Fomichev
Cc: Jiri Olsa, Arnaldo Carvalho de Melo, lkml, Ingo Molnar,
Namhyung Kim, Alexander Shishkin, Peter Zijlstra, Andi Kleen,
Adrian Hunter, Song Liu, Alexei Starovoitov, Daniel Borkmann
On Tue, Apr 16, 2019 at 10:17:13AM -0700, Stanislav Fomichev wrote:
> On 04/16, Jiri Olsa wrote:
> > Moving file specific code in dso_cache__read function
> > into separate file_read function. I'll add bpf specific
> > code in following patches.
> >
> > Link: http://lkml.kernel.org/n/tip-7f7d717uzrqt5ka2xp29ige3@git.kernel.org
> > Signed-off-by: Jiri Olsa <jolsa@kernel.org>
> > ---
> > tools/perf/util/dso.c | 47 ++++++++++++++++++++++++-------------------
> > 1 file changed, 26 insertions(+), 21 deletions(-)
> >
> > diff --git a/tools/perf/util/dso.c b/tools/perf/util/dso.c
> > index cb6199c1390a..6baad22ec8a9 100644
> > --- a/tools/perf/util/dso.c
> > +++ b/tools/perf/util/dso.c
> > @@ -794,6 +794,30 @@ dso_cache__memcpy(struct dso_cache *cache, u64 offset,
> > return cache_size;
> > }
> >
> > +static ssize_t file_read(struct dso *dso, struct machine *machine,
> > + u64 offset, char *data)
> > +{
> > + ssize_t ret;
> > +
> > + pthread_mutex_lock(&dso__data_open_lock);
> > +
> > + /*
> > + * dso->data.fd might be closed if other thread opened another
> > + * file (dso) due to open file limit (RLIMIT_NOFILE).
> > + */
> > + try_to_open_dso(dso, machine);
> > +
> > + if (dso->data.fd < 0) {
> > + dso->data.status = DSO_DATA_STATUS_ERROR;
> pthread_mutex_unlock(&dso__data_open_lock) here?
oops, yea.. thanks
jirka
^ permalink raw reply [flat|nested] 31+ messages in thread
* [PATCH 03/12] perf tools: Simplify dso_cache__read function
2019-04-16 16:01 [PATCH 00/12] perf tools: Display eBPF code in intel_pt trace Jiri Olsa
2019-04-16 16:01 ` [PATCH 01/12] perf tools: Separate generic code in dso__data_file_size Jiri Olsa
2019-04-16 16:01 ` [PATCH 02/12] perf tools: Separate generic code in dso_cache__read Jiri Olsa
@ 2019-04-16 16:01 ` Jiri Olsa
2019-04-16 16:01 ` [PATCH 04/12] perf tools: Add bpf dso read and size hooks Jiri Olsa
` (8 subsequent siblings)
11 siblings, 0 replies; 31+ messages in thread
From: Jiri Olsa @ 2019-04-16 16:01 UTC (permalink / raw)
To: Arnaldo Carvalho de Melo
Cc: lkml, Ingo Molnar, Namhyung Kim, Alexander Shishkin,
Peter Zijlstra, Andi Kleen, Adrian Hunter, Song Liu,
Alexei Starovoitov, Daniel Borkmann
There's no need for the while loop now, also we can
connect two (ret > 0) condition legs together.
Link: http://lkml.kernel.org/n/tip-2vg28bak8euojuvq33cy2hl5@git.kernel.org
Signed-off-by: Jiri Olsa <jolsa@kernel.org>
---
tools/perf/util/dso.c | 17 ++++++-----------
1 file changed, 6 insertions(+), 11 deletions(-)
diff --git a/tools/perf/util/dso.c b/tools/perf/util/dso.c
index 6baad22ec8a9..6fdff723e55a 100644
--- a/tools/perf/util/dso.c
+++ b/tools/perf/util/dso.c
@@ -822,25 +822,20 @@ static ssize_t
dso_cache__read(struct dso *dso, struct machine *machine,
u64 offset, u8 *data, ssize_t size)
{
+ u64 cache_offset = offset & DSO__DATA_CACHE_MASK;
struct dso_cache *cache;
struct dso_cache *old;
ssize_t ret;
- do {
- u64 cache_offset = offset & DSO__DATA_CACHE_MASK;
-
- cache = zalloc(sizeof(*cache) + DSO__DATA_CACHE_SIZE);
- if (!cache)
- return -ENOMEM;
-
- ret = file_read(dso, machine, cache_offset, cache->data);
+ cache = zalloc(sizeof(*cache) + DSO__DATA_CACHE_SIZE);
+ if (!cache)
+ return -ENOMEM;
+ ret = file_read(dso, machine, cache_offset, cache->data);
+ if (ret > 0) {
cache->offset = cache_offset;
cache->size = ret;
- } while (0);
-
- if (ret > 0) {
old = dso_cache__insert(dso, cache);
if (old) {
/* we lose the race */
--
2.17.2
^ permalink raw reply related [flat|nested] 31+ messages in thread
* [PATCH 04/12] perf tools: Add bpf dso read and size hooks
2019-04-16 16:01 [PATCH 00/12] perf tools: Display eBPF code in intel_pt trace Jiri Olsa
` (2 preceding siblings ...)
2019-04-16 16:01 ` [PATCH 03/12] perf tools: Simplify dso_cache__read function Jiri Olsa
@ 2019-04-16 16:01 ` Jiri Olsa
2019-04-16 16:01 ` [PATCH 05/12] perf tools: Read also the end of the kernel Jiri Olsa
` (7 subsequent siblings)
11 siblings, 0 replies; 31+ messages in thread
From: Jiri Olsa @ 2019-04-16 16:01 UTC (permalink / raw)
To: Arnaldo Carvalho de Melo
Cc: lkml, Ingo Molnar, Namhyung Kim, Alexander Shishkin,
Peter Zijlstra, Andi Kleen, Adrian Hunter, Song Liu,
Alexei Starovoitov, Daniel Borkmann
Adding bpf related code into dso reading paths.
Link: http://lkml.kernel.org/n/tip-ql6jpegvv5823yc87u0hlzfa@git.kernel.org
Signed-off-by: Jiri Olsa <jolsa@kernel.org>
---
tools/perf/util/dso.c | 49 ++++++++++++++++++++++++++++++++++++++++++-
1 file changed, 48 insertions(+), 1 deletion(-)
diff --git a/tools/perf/util/dso.c b/tools/perf/util/dso.c
index 6fdff723e55a..0cc6702c6442 100644
--- a/tools/perf/util/dso.c
+++ b/tools/perf/util/dso.c
@@ -9,6 +9,8 @@
#include <errno.h>
#include <fcntl.h>
#include <libgen.h>
+#include <bpf/libbpf.h>
+#include "bpf-event.h"
#include "compress.h"
#include "namespaces.h"
#include "path.h"
@@ -706,6 +708,44 @@ bool dso__data_status_seen(struct dso *dso, enum dso_data_status_seen by)
return false;
}
+static ssize_t bpf_read(struct dso *dso, u64 offset, char *data)
+{
+ struct bpf_prog_info_node *node;
+ ssize_t size = DSO__DATA_CACHE_SIZE;
+ u64 len;
+ u8 *buf;
+
+ node = perf_env__find_bpf_prog_info(dso->bpf_prog.env, dso->bpf_prog.id);
+ if (!node || !node->info_linear) {
+ dso->data.status = DSO_DATA_STATUS_ERROR;
+ return -1;
+ }
+
+ len = node->info_linear->info.jited_prog_len;
+ buf = (u8 *) node->info_linear->info.jited_prog_insns;
+
+ if (offset >= len)
+ return -1;
+
+ size = (ssize_t) min(len - offset, (u64) size);
+ memcpy(data, buf + offset, size);
+ return size;
+}
+
+static int bpf_size(struct dso *dso)
+{
+ struct bpf_prog_info_node *node;
+
+ node = perf_env__find_bpf_prog_info(dso->bpf_prog.env, dso->bpf_prog.id);
+ if (!node || !node->info_linear) {
+ dso->data.status = DSO_DATA_STATUS_ERROR;
+ return -1;
+ }
+
+ dso->data.file_size = node->info_linear->info.jited_prog_len;
+ return 0;
+}
+
static void
dso_cache__free(struct dso *dso)
{
@@ -831,7 +871,11 @@ dso_cache__read(struct dso *dso, struct machine *machine,
if (!cache)
return -ENOMEM;
- ret = file_read(dso, machine, cache_offset, cache->data);
+ if (dso->binary_type == DSO_BINARY_TYPE__BPF_PROG_INFO)
+ ret = bpf_read(dso, cache_offset, cache->data);
+ else
+ ret = file_read(dso, machine, cache_offset, cache->data);
+
if (ret > 0) {
cache->offset = cache_offset;
cache->size = ret;
@@ -940,6 +984,9 @@ int dso__data_file_size(struct dso *dso, struct machine *machine)
if (dso->data.status == DSO_DATA_STATUS_ERROR)
return -1;
+ if (dso->binary_type == DSO_BINARY_TYPE__BPF_PROG_INFO)
+ return bpf_size(dso);
+
return file_size(dso, machine);
}
--
2.17.2
^ permalink raw reply related [flat|nested] 31+ messages in thread
* [PATCH 05/12] perf tools: Read also the end of the kernel
2019-04-16 16:01 [PATCH 00/12] perf tools: Display eBPF code in intel_pt trace Jiri Olsa
` (3 preceding siblings ...)
2019-04-16 16:01 ` [PATCH 04/12] perf tools: Add bpf dso read and size hooks Jiri Olsa
@ 2019-04-16 16:01 ` Jiri Olsa
2019-04-16 16:01 ` [PATCH 06/12] perf tools: Do not erase uncovered maps by kcore Jiri Olsa
` (6 subsequent siblings)
11 siblings, 0 replies; 31+ messages in thread
From: Jiri Olsa @ 2019-04-16 16:01 UTC (permalink / raw)
To: Arnaldo Carvalho de Melo
Cc: lkml, Ingo Molnar, Namhyung Kim, Alexander Shishkin,
Peter Zijlstra, Andi Kleen, Adrian Hunter, Song Liu,
Alexei Starovoitov, Daniel Borkmann
We mark the end of kernel based on the first module,
but that could cover some bpf program maps. Reading
_etext symbol if it's present to get precise kernel
map end.
Link: http://lkml.kernel.org/n/tip-ynut991ttyyhvo1sbhlm4c42@git.kernel.org
Signed-off-by: Jiri Olsa <jolsa@kernel.org>
---
tools/perf/util/machine.c | 27 ++++++++++++++++++---------
1 file changed, 18 insertions(+), 9 deletions(-)
diff --git a/tools/perf/util/machine.c b/tools/perf/util/machine.c
index 3c520baa198c..ad0205fbb506 100644
--- a/tools/perf/util/machine.c
+++ b/tools/perf/util/machine.c
@@ -924,7 +924,8 @@ const char *ref_reloc_sym_names[] = {"_text", "_stext", NULL};
* symbol_name if it's not that important.
*/
static int machine__get_running_kernel_start(struct machine *machine,
- const char **symbol_name, u64 *start)
+ const char **symbol_name,
+ u64 *start, u64 *end)
{
char filename[PATH_MAX];
int i, err = -1;
@@ -949,6 +950,11 @@ static int machine__get_running_kernel_start(struct machine *machine,
*symbol_name = name;
*start = addr;
+
+ err = kallsyms__get_function_start(filename, "_etext", &addr);
+ if (!err)
+ *end = addr;
+
return 0;
}
@@ -1440,7 +1446,7 @@ int machine__create_kernel_maps(struct machine *machine)
struct dso *kernel = machine__get_kernel(machine);
const char *name = NULL;
struct map *map;
- u64 addr = 0;
+ u64 start = 0, end = ~0ULL;
int ret;
if (kernel == NULL)
@@ -1459,9 +1465,9 @@ int machine__create_kernel_maps(struct machine *machine)
"continuing anyway...\n", machine->pid);
}
- if (!machine__get_running_kernel_start(machine, &name, &addr)) {
+ if (!machine__get_running_kernel_start(machine, &name, &start, &end)) {
if (name &&
- map__set_kallsyms_ref_reloc_sym(machine->vmlinux_map, name, addr)) {
+ map__set_kallsyms_ref_reloc_sym(machine->vmlinux_map, name, start)) {
machine__destroy_kernel_maps(machine);
ret = -1;
goto out_put;
@@ -1471,16 +1477,19 @@ int machine__create_kernel_maps(struct machine *machine)
* we have a real start address now, so re-order the kmaps
* assume it's the last in the kmaps
*/
- machine__update_kernel_mmap(machine, addr, ~0ULL);
+ machine__update_kernel_mmap(machine, start, end);
}
if (machine__create_extra_kernel_maps(machine, kernel))
pr_debug("Problems creating extra kernel maps, continuing anyway...\n");
- /* update end address of the kernel map using adjacent module address */
- map = map__next(machine__kernel_map(machine));
- if (map)
- machine__set_kernel_mmap(machine, addr, map->start);
+ if (end == ~0ULL) {
+ /* update end address of the kernel map using adjacent module address */
+ map = map__next(machine__kernel_map(machine));
+ if (map)
+ machine__set_kernel_mmap(machine, start, map->start);
+ }
+
out_put:
dso__put(kernel);
return ret;
--
2.17.2
^ permalink raw reply related [flat|nested] 31+ messages in thread
* [PATCH 06/12] perf tools: Do not erase uncovered maps by kcore
2019-04-16 16:01 [PATCH 00/12] perf tools: Display eBPF code in intel_pt trace Jiri Olsa
` (4 preceding siblings ...)
2019-04-16 16:01 ` [PATCH 05/12] perf tools: Read also the end of the kernel Jiri Olsa
@ 2019-04-16 16:01 ` Jiri Olsa
2019-04-23 9:32 ` Adrian Hunter
2019-04-16 16:01 ` [PATCH 07/12] perf tools: Check maps for bpf programs Jiri Olsa
` (5 subsequent siblings)
11 siblings, 1 reply; 31+ messages in thread
From: Jiri Olsa @ 2019-04-16 16:01 UTC (permalink / raw)
To: Arnaldo Carvalho de Melo
Cc: lkml, Ingo Molnar, Namhyung Kim, Alexander Shishkin,
Peter Zijlstra, Andi Kleen, Adrian Hunter, Song Liu,
Alexei Starovoitov, Daniel Borkmann
Maps in kcore do not cover bpf maps, so we can't just
remove everything. Keeping all kernel maps, which are
not covered by kcore maps.
Link: http://lkml.kernel.org/n/tip-9eytka8wofp0a047ul6lmejk@git.kernel.org
Signed-off-by: Jiri Olsa <jolsa@kernel.org>
---
tools/perf/util/symbol.c | 14 +++++++++++++-
1 file changed, 13 insertions(+), 1 deletion(-)
diff --git a/tools/perf/util/symbol.c b/tools/perf/util/symbol.c
index 5cbad55cd99d..96738a7a8c14 100644
--- a/tools/perf/util/symbol.c
+++ b/tools/perf/util/symbol.c
@@ -1166,6 +1166,18 @@ static int kcore_mapfn(u64 start, u64 len, u64 pgoff, void *data)
return 0;
}
+static bool in_kcore(struct kcore_mapfn_data *md, struct map *map)
+{
+ struct map *iter;
+
+ list_for_each_entry(iter, &md->maps, node) {
+ if ((map->start >= iter->start) && (map->start < iter->end))
+ return true;
+ }
+
+ return false;
+}
+
static int dso__load_kcore(struct dso *dso, struct map *map,
const char *kallsyms_filename)
{
@@ -1222,7 +1234,7 @@ static int dso__load_kcore(struct dso *dso, struct map *map,
while (old_map) {
struct map *next = map_groups__next(old_map);
- if (old_map != map)
+ if (old_map != map && !in_kcore(&md, old_map))
map_groups__remove(kmaps, old_map);
old_map = next;
}
--
2.17.2
^ permalink raw reply related [flat|nested] 31+ messages in thread
* Re: [PATCH 06/12] perf tools: Do not erase uncovered maps by kcore
2019-04-16 16:01 ` [PATCH 06/12] perf tools: Do not erase uncovered maps by kcore Jiri Olsa
@ 2019-04-23 9:32 ` Adrian Hunter
2019-04-23 14:55 ` Jiri Olsa
0 siblings, 1 reply; 31+ messages in thread
From: Adrian Hunter @ 2019-04-23 9:32 UTC (permalink / raw)
To: Jiri Olsa, Arnaldo Carvalho de Melo
Cc: lkml, Ingo Molnar, Namhyung Kim, Alexander Shishkin,
Peter Zijlstra, Andi Kleen, Song Liu, Alexei Starovoitov,
Daniel Borkmann
On 16/04/19 7:01 PM, Jiri Olsa wrote:
> Maps in kcore do not cover bpf maps, so we can't just
> remove everything. Keeping all kernel maps, which are
> not covered by kcore maps.
Memory for jited-bpf is allocated from the same area that is used for
modules. In the case of /proc/kcore, that entire area is mapped, so there
won't be any bpf-maps that are not covered. For copies of kcore made by
'perf buildid-cache' the same would be true for any bpf that got allocated
in between modules.
But shouldn't the bpf map supersede the kcore map for the address range that
it maps? I guess that would mean splitting the kcore map, truncating the
first piece and inserting the bpf map in between.
>
> Link: http://lkml.kernel.org/n/tip-9eytka8wofp0a047ul6lmejk@git.kernel.org
> Signed-off-by: Jiri Olsa <jolsa@kernel.org>
> ---
> tools/perf/util/symbol.c | 14 +++++++++++++-
> 1 file changed, 13 insertions(+), 1 deletion(-)
>
> diff --git a/tools/perf/util/symbol.c b/tools/perf/util/symbol.c
> index 5cbad55cd99d..96738a7a8c14 100644
> --- a/tools/perf/util/symbol.c
> +++ b/tools/perf/util/symbol.c
> @@ -1166,6 +1166,18 @@ static int kcore_mapfn(u64 start, u64 len, u64 pgoff, void *data)
> return 0;
> }
>
> +static bool in_kcore(struct kcore_mapfn_data *md, struct map *map)
> +{
> + struct map *iter;
> +
> + list_for_each_entry(iter, &md->maps, node) {
> + if ((map->start >= iter->start) && (map->start < iter->end))
> + return true;
> + }
> +
> + return false;
> +}
> +
> static int dso__load_kcore(struct dso *dso, struct map *map,
> const char *kallsyms_filename)
> {
> @@ -1222,7 +1234,7 @@ static int dso__load_kcore(struct dso *dso, struct map *map,
> while (old_map) {
> struct map *next = map_groups__next(old_map);
>
> - if (old_map != map)
> + if (old_map != map && !in_kcore(&md, old_map))
> map_groups__remove(kmaps, old_map);
> old_map = next;
> }
>
^ permalink raw reply [flat|nested] 31+ messages in thread
* Re: [PATCH 06/12] perf tools: Do not erase uncovered maps by kcore
2019-04-23 9:32 ` Adrian Hunter
@ 2019-04-23 14:55 ` Jiri Olsa
0 siblings, 0 replies; 31+ messages in thread
From: Jiri Olsa @ 2019-04-23 14:55 UTC (permalink / raw)
To: Adrian Hunter
Cc: Jiri Olsa, Arnaldo Carvalho de Melo, lkml, Ingo Molnar,
Namhyung Kim, Alexander Shishkin, Peter Zijlstra, Andi Kleen,
Song Liu, Alexei Starovoitov, Daniel Borkmann
On Tue, Apr 23, 2019 at 12:32:12PM +0300, Adrian Hunter wrote:
> On 16/04/19 7:01 PM, Jiri Olsa wrote:
> > Maps in kcore do not cover bpf maps, so we can't just
> > remove everything. Keeping all kernel maps, which are
> > not covered by kcore maps.
>
> Memory for jited-bpf is allocated from the same area that is used for
> modules. In the case of /proc/kcore, that entire area is mapped, so there
> won't be any bpf-maps that are not covered. For copies of kcore made by
> 'perf buildid-cache' the same would be true for any bpf that got allocated
> in between modules.
>
> But shouldn't the bpf map supersede the kcore map for the address range that
> it maps? I guess that would mean splitting the kcore map, truncating the
> first piece and inserting the bpf map in between.
I haven't considered it could get in between modules,
I think you're right and we need to cut kcore maps
in case bpf is mapped within.. I'll submit new version
thanks,
jirka
>
> >
> > Link: http://lkml.kernel.org/n/tip-9eytka8wofp0a047ul6lmejk@git.kernel.org
> > Signed-off-by: Jiri Olsa <jolsa@kernel.org>
> > ---
> > tools/perf/util/symbol.c | 14 +++++++++++++-
> > 1 file changed, 13 insertions(+), 1 deletion(-)
> >
> > diff --git a/tools/perf/util/symbol.c b/tools/perf/util/symbol.c
> > index 5cbad55cd99d..96738a7a8c14 100644
> > --- a/tools/perf/util/symbol.c
> > +++ b/tools/perf/util/symbol.c
> > @@ -1166,6 +1166,18 @@ static int kcore_mapfn(u64 start, u64 len, u64 pgoff, void *data)
> > return 0;
> > }
> >
> > +static bool in_kcore(struct kcore_mapfn_data *md, struct map *map)
> > +{
> > + struct map *iter;
> > +
> > + list_for_each_entry(iter, &md->maps, node) {
> > + if ((map->start >= iter->start) && (map->start < iter->end))
> > + return true;
> > + }
> > +
> > + return false;
> > +}
> > +
> > static int dso__load_kcore(struct dso *dso, struct map *map,
> > const char *kallsyms_filename)
> > {
> > @@ -1222,7 +1234,7 @@ static int dso__load_kcore(struct dso *dso, struct map *map,
> > while (old_map) {
> > struct map *next = map_groups__next(old_map);
> >
> > - if (old_map != map)
> > + if (old_map != map && !in_kcore(&md, old_map))
> > map_groups__remove(kmaps, old_map);
> > old_map = next;
> > }
> >
>
^ permalink raw reply [flat|nested] 31+ messages in thread
* [PATCH 07/12] perf tools: Check maps for bpf programs
2019-04-16 16:01 [PATCH 00/12] perf tools: Display eBPF code in intel_pt trace Jiri Olsa
` (5 preceding siblings ...)
2019-04-16 16:01 ` [PATCH 06/12] perf tools: Do not erase uncovered maps by kcore Jiri Olsa
@ 2019-04-16 16:01 ` Jiri Olsa
2019-04-16 19:31 ` Arnaldo Carvalho de Melo
2019-04-19 17:18 ` [tip:perf/urgent] " tip-bot for Song Liu
2019-04-16 16:01 ` [PATCH 08/12] perf tools: Fix side band thread draining Jiri Olsa
` (4 subsequent siblings)
11 siblings, 2 replies; 31+ messages in thread
From: Jiri Olsa @ 2019-04-16 16:01 UTC (permalink / raw)
To: Arnaldo Carvalho de Melo
Cc: Adrian Hunter, Alexei Starovoitov, Andrii Nakryiko,
Daniel Borkmann, Martin KaFai Lau, Namhyung Kim, Peter Zijlstra,
Song Liu, Yonghong Song, Arnaldo Carvalho de Melo, lkml,
Ingo Molnar, Alexander Shishkin, Peter Zijlstra, Andi Kleen
From: Song Liu <songliubraving@fb.com>
As reported by Jiri Olsa in:
"[BUG] perf: intel_pt won't display kernel function"
https://lore.kernel.org/lkml/20190403143738.GB32001@krava
Recent changes to support PERF_RECORD_KSYMBOL and PERF_RECORD_BPF_EVENT
broke --kallsyms option. This is because it broke test __map__is_kmodule.
This patch fixes this by adding check for bpf program, so that these maps
are not mistaken as kernel modules.
Reported-by: Jiri Olsa <jolsa@kernel.org>
Cc: Adrian Hunter <adrian.hunter@intel.com>
Cc: Alexei Starovoitov <ast@kernel.org>
Cc: Andrii Nakryiko <andrii.nakryiko@gmail.com>
Cc: Daniel Borkmann <daniel@iogearbox.net>
Cc: Martin KaFai Lau <kafai@fb.com>
Cc: Namhyung Kim <namhyung@kernel.org>
Cc: Peter Zijlstra <peterz@infradead.org>
Cc: Song Liu <songliubraving@fb.com>
Cc: Yonghong Song <yhs@fb.com>
Cc: Arnaldo Carvalho de Melo <acme@redhat.com>
Fixes: 76193a94522f ("perf, bpf: Introduce PERF_RECORD_KSYMBOL")
Link: https://lore.kernel.org/lkml/20190403145353.GE32553@kernel.org
Signed-off-by: Song Liu <songliubraving@fb.com>
Link: http://lkml.kernel.org/n/tip-gnffkcoc4nauqcpgvsmwka2w@git.kernel.org
Signed-off-by: Jiri Olsa <jolsa@kernel.org>
---
tools/perf/util/map.c | 16 ++++++++++++++++
tools/perf/util/map.h | 4 +++-
2 files changed, 19 insertions(+), 1 deletion(-)
diff --git a/tools/perf/util/map.c b/tools/perf/util/map.c
index e32628cd20a7..28d484ef74ae 100644
--- a/tools/perf/util/map.c
+++ b/tools/perf/util/map.c
@@ -261,6 +261,22 @@ bool __map__is_extra_kernel_map(const struct map *map)
return kmap && kmap->name[0];
}
+bool __map__is_bpf_prog(const struct map *map)
+{
+ const char *name;
+
+ if (map->dso->binary_type == DSO_BINARY_TYPE__BPF_PROG_INFO)
+ return true;
+
+ /*
+ * If PERF_RECORD_BPF_EVENT is not included, the dso will not have
+ * type of DSO_BINARY_TYPE__BPF_PROG_INFO. In such cases, we can
+ * guess the type based on name.
+ */
+ name = map->dso->short_name;
+ return name && (strstr(name, "bpf_prog_") == name);
+}
+
bool map__has_symbols(const struct map *map)
{
return dso__has_symbols(map->dso);
diff --git a/tools/perf/util/map.h b/tools/perf/util/map.h
index 0e20749f2c55..dc93787c74f0 100644
--- a/tools/perf/util/map.h
+++ b/tools/perf/util/map.h
@@ -159,10 +159,12 @@ int map__set_kallsyms_ref_reloc_sym(struct map *map, const char *symbol_name,
bool __map__is_kernel(const struct map *map);
bool __map__is_extra_kernel_map(const struct map *map);
+bool __map__is_bpf_prog(const struct map *map);
static inline bool __map__is_kmodule(const struct map *map)
{
- return !__map__is_kernel(map) && !__map__is_extra_kernel_map(map);
+ return !__map__is_kernel(map) && !__map__is_extra_kernel_map(map) &&
+ !__map__is_bpf_prog(map);
}
bool map__has_symbols(const struct map *map);
--
2.17.2
^ permalink raw reply related [flat|nested] 31+ messages in thread
* Re: [PATCH 07/12] perf tools: Check maps for bpf programs
2019-04-16 16:01 ` [PATCH 07/12] perf tools: Check maps for bpf programs Jiri Olsa
@ 2019-04-16 19:31 ` Arnaldo Carvalho de Melo
2019-04-19 17:18 ` [tip:perf/urgent] " tip-bot for Song Liu
1 sibling, 0 replies; 31+ messages in thread
From: Arnaldo Carvalho de Melo @ 2019-04-16 19:31 UTC (permalink / raw)
To: Jiri Olsa
Cc: Arnaldo Carvalho de Melo, Adrian Hunter, Alexei Starovoitov,
Andrii Nakryiko, Daniel Borkmann, Martin KaFai Lau, Namhyung Kim,
Peter Zijlstra, Song Liu, Yonghong Song, lkml, Ingo Molnar,
Alexander Shishkin, Peter Zijlstra, Andi Kleen
Em Tue, Apr 16, 2019 at 06:01:22PM +0200, Jiri Olsa escreveu:
> From: Song Liu <songliubraving@fb.com>
>
> As reported by Jiri Olsa in:
>
> "[BUG] perf: intel_pt won't display kernel function"
> https://lore.kernel.org/lkml/20190403143738.GB32001@krava
>
> Recent changes to support PERF_RECORD_KSYMBOL and PERF_RECORD_BPF_EVENT
> broke --kallsyms option. This is because it broke test __map__is_kmodule.
>
> This patch fixes this by adding check for bpf program, so that these maps
> are not mistaken as kernel modules.
Thanks, applied to perf/urgent.
- Arnaldo
^ permalink raw reply [flat|nested] 31+ messages in thread
* [tip:perf/urgent] perf tools: Check maps for bpf programs
2019-04-16 16:01 ` [PATCH 07/12] perf tools: Check maps for bpf programs Jiri Olsa
2019-04-16 19:31 ` Arnaldo Carvalho de Melo
@ 2019-04-19 17:18 ` tip-bot for Song Liu
1 sibling, 0 replies; 31+ messages in thread
From: tip-bot for Song Liu @ 2019-04-19 17:18 UTC (permalink / raw)
To: linux-tip-commits
Cc: andrii.nakryiko, tglx, yhs, namhyung, alexander.shishkin,
adrian.hunter, mingo, ast, daniel, linux-kernel, kafai,
songliubraving, acme, hpa, peterz, jolsa, ak
Commit-ID: a93e0b2365e81e5a5b61f25e269b5dc73d242cba
Gitweb: https://git.kernel.org/tip/a93e0b2365e81e5a5b61f25e269b5dc73d242cba
Author: Song Liu <songliubraving@fb.com>
AuthorDate: Tue, 16 Apr 2019 18:01:22 +0200
Committer: Arnaldo Carvalho de Melo <acme@redhat.com>
CommitDate: Wed, 17 Apr 2019 14:30:11 -0300
perf tools: Check maps for bpf programs
As reported by Jiri Olsa in:
"[BUG] perf: intel_pt won't display kernel function"
https://lore.kernel.org/lkml/20190403143738.GB32001@krava
Recent changes to support PERF_RECORD_KSYMBOL and PERF_RECORD_BPF_EVENT
broke --kallsyms option. This is because it broke test __map__is_kmodule.
This patch fixes this by adding check for bpf program, so that these maps
are not mistaken as kernel modules.
Signed-off-by: Song Liu <songliubraving@fb.com>
Reported-by: Jiri Olsa <jolsa@kernel.org>
Cc: Adrian Hunter <adrian.hunter@intel.com>
Cc: Alexander Shishkin <alexander.shishkin@linux.intel.com>
Cc: Alexei Starovoitov <ast@kernel.org>
Cc: Andi Kleen <ak@linux.intel.com>
Cc: Andrii Nakryiko <andrii.nakryiko@gmail.com>
Cc: Daniel Borkmann <daniel@iogearbox.net>
Cc: Martin KaFai Lau <kafai@fb.com>
Cc: Namhyung Kim <namhyung@kernel.org>
Cc: Peter Zijlstra <peterz@infradead.org>
Cc: Yonghong Song <yhs@fb.com>
Link: http://lkml.kernel.org/r/20190416160127.30203-8-jolsa@kernel.org
Fixes: 76193a94522f ("perf, bpf: Introduce PERF_RECORD_KSYMBOL")
Signed-off-by: Jiri Olsa <jolsa@kernel.org>
Signed-off-by: Arnaldo Carvalho de Melo <acme@redhat.com>
---
tools/perf/util/map.c | 16 ++++++++++++++++
tools/perf/util/map.h | 4 +++-
2 files changed, 19 insertions(+), 1 deletion(-)
diff --git a/tools/perf/util/map.c b/tools/perf/util/map.c
index e32628cd20a7..28d484ef74ae 100644
--- a/tools/perf/util/map.c
+++ b/tools/perf/util/map.c
@@ -261,6 +261,22 @@ bool __map__is_extra_kernel_map(const struct map *map)
return kmap && kmap->name[0];
}
+bool __map__is_bpf_prog(const struct map *map)
+{
+ const char *name;
+
+ if (map->dso->binary_type == DSO_BINARY_TYPE__BPF_PROG_INFO)
+ return true;
+
+ /*
+ * If PERF_RECORD_BPF_EVENT is not included, the dso will not have
+ * type of DSO_BINARY_TYPE__BPF_PROG_INFO. In such cases, we can
+ * guess the type based on name.
+ */
+ name = map->dso->short_name;
+ return name && (strstr(name, "bpf_prog_") == name);
+}
+
bool map__has_symbols(const struct map *map)
{
return dso__has_symbols(map->dso);
diff --git a/tools/perf/util/map.h b/tools/perf/util/map.h
index 0e20749f2c55..dc93787c74f0 100644
--- a/tools/perf/util/map.h
+++ b/tools/perf/util/map.h
@@ -159,10 +159,12 @@ int map__set_kallsyms_ref_reloc_sym(struct map *map, const char *symbol_name,
bool __map__is_kernel(const struct map *map);
bool __map__is_extra_kernel_map(const struct map *map);
+bool __map__is_bpf_prog(const struct map *map);
static inline bool __map__is_kmodule(const struct map *map)
{
- return !__map__is_kernel(map) && !__map__is_extra_kernel_map(map);
+ return !__map__is_kernel(map) && !__map__is_extra_kernel_map(map) &&
+ !__map__is_bpf_prog(map);
}
bool map__has_symbols(const struct map *map);
^ permalink raw reply related [flat|nested] 31+ messages in thread
* [PATCH 08/12] perf tools: Fix side band thread draining
2019-04-16 16:01 [PATCH 00/12] perf tools: Display eBPF code in intel_pt trace Jiri Olsa
` (6 preceding siblings ...)
2019-04-16 16:01 ` [PATCH 07/12] perf tools: Check maps for bpf programs Jiri Olsa
@ 2019-04-16 16:01 ` Jiri Olsa
2019-04-16 19:33 ` Arnaldo Carvalho de Melo
` (2 more replies)
2019-04-16 16:01 ` [PATCH 09/12] perf tools: Fix map reference counting Jiri Olsa
` (3 subsequent siblings)
11 siblings, 3 replies; 31+ messages in thread
From: Jiri Olsa @ 2019-04-16 16:01 UTC (permalink / raw)
To: Arnaldo Carvalho de Melo
Cc: lkml, Ingo Molnar, Namhyung Kim, Alexander Shishkin,
Peter Zijlstra, Andi Kleen, Adrian Hunter, Song Liu,
Alexei Starovoitov, Daniel Borkmann
Current perf_evlist__poll_thread code could finish
without draining the data. Adding the logic that
makes sure we won't finish before the drain.
Fixes: 657ee5531903 ("perf evlist: Introduce side band thread")
Link: http://lkml.kernel.org/n/tip-41i888xyim9n5ceyr44jb468@git.kernel.org
Signed-off-by: Jiri Olsa <jolsa@kernel.org>
---
tools/perf/util/evlist.c | 14 +++++++++-----
1 file changed, 9 insertions(+), 5 deletions(-)
diff --git a/tools/perf/util/evlist.c b/tools/perf/util/evlist.c
index f2bbae38278d..4b6783ff5813 100644
--- a/tools/perf/util/evlist.c
+++ b/tools/perf/util/evlist.c
@@ -1868,12 +1868,12 @@ static void *perf_evlist__poll_thread(void *arg)
{
struct perf_evlist *evlist = arg;
bool draining = false;
- int i;
+ int i, done = 0;
+
+ while (!done) {
+ bool got_data = false;
- while (draining || !(evlist->thread.done)) {
- if (draining)
- draining = false;
- else if (evlist->thread.done)
+ if (evlist->thread.done)
draining = true;
if (!draining)
@@ -1894,9 +1894,13 @@ static void *perf_evlist__poll_thread(void *arg)
pr_warning("cannot locate proper evsel for the side band event\n");
perf_mmap__consume(map);
+ got_data = true;
}
perf_mmap__read_done(map);
}
+
+ if (draining && !got_data)
+ break;
}
return NULL;
}
--
2.17.2
^ permalink raw reply related [flat|nested] 31+ messages in thread
* Re: [PATCH 08/12] perf tools: Fix side band thread draining
2019-04-16 16:01 ` [PATCH 08/12] perf tools: Fix side band thread draining Jiri Olsa
@ 2019-04-16 19:33 ` Arnaldo Carvalho de Melo
2019-04-16 20:41 ` Song Liu
2019-04-19 17:19 ` [tip:perf/urgent] perf evlist: " tip-bot for Jiri Olsa
2 siblings, 0 replies; 31+ messages in thread
From: Arnaldo Carvalho de Melo @ 2019-04-16 19:33 UTC (permalink / raw)
To: Jiri Olsa
Cc: lkml, Ingo Molnar, Namhyung Kim, Alexander Shishkin,
Peter Zijlstra, Andi Kleen, Adrian Hunter, Song Liu,
Alexei Starovoitov, Daniel Borkmann
Em Tue, Apr 16, 2019 at 06:01:23PM +0200, Jiri Olsa escreveu:
> Current perf_evlist__poll_thread code could finish
> without draining the data. Adding the logic that
> makes sure we won't finish before the drain.
>
> Fixes: 657ee5531903 ("perf evlist: Introduce side band thread")
Thanks, applied to perf/urgent.
- Arnaldo
> Link: http://lkml.kernel.org/n/tip-41i888xyim9n5ceyr44jb468@git.kernel.org
> Signed-off-by: Jiri Olsa <jolsa@kernel.org>
> ---
> tools/perf/util/evlist.c | 14 +++++++++-----
> 1 file changed, 9 insertions(+), 5 deletions(-)
>
> diff --git a/tools/perf/util/evlist.c b/tools/perf/util/evlist.c
> index f2bbae38278d..4b6783ff5813 100644
> --- a/tools/perf/util/evlist.c
> +++ b/tools/perf/util/evlist.c
> @@ -1868,12 +1868,12 @@ static void *perf_evlist__poll_thread(void *arg)
> {
> struct perf_evlist *evlist = arg;
> bool draining = false;
> - int i;
> + int i, done = 0;
> +
> + while (!done) {
> + bool got_data = false;
>
> - while (draining || !(evlist->thread.done)) {
> - if (draining)
> - draining = false;
> - else if (evlist->thread.done)
> + if (evlist->thread.done)
> draining = true;
>
> if (!draining)
> @@ -1894,9 +1894,13 @@ static void *perf_evlist__poll_thread(void *arg)
> pr_warning("cannot locate proper evsel for the side band event\n");
>
> perf_mmap__consume(map);
> + got_data = true;
> }
> perf_mmap__read_done(map);
> }
> +
> + if (draining && !got_data)
> + break;
> }
> return NULL;
> }
> --
> 2.17.2
--
- Arnaldo
^ permalink raw reply [flat|nested] 31+ messages in thread
* Re: [PATCH 08/12] perf tools: Fix side band thread draining
2019-04-16 16:01 ` [PATCH 08/12] perf tools: Fix side band thread draining Jiri Olsa
2019-04-16 19:33 ` Arnaldo Carvalho de Melo
@ 2019-04-16 20:41 ` Song Liu
2019-04-19 17:19 ` [tip:perf/urgent] perf evlist: " tip-bot for Jiri Olsa
2 siblings, 0 replies; 31+ messages in thread
From: Song Liu @ 2019-04-16 20:41 UTC (permalink / raw)
To: Jiri Olsa
Cc: Arnaldo Carvalho de Melo, lkml, Ingo Molnar, Namhyung Kim,
Alexander Shishkin, Peter Zijlstra, Andi Kleen, Adrian Hunter,
Alexei Starovoitov, Daniel Borkmann
> On Apr 16, 2019, at 9:01 AM, Jiri Olsa <jolsa@kernel.org> wrote:
>
> Current perf_evlist__poll_thread code could finish
> without draining the data. Adding the logic that
> makes sure we won't finish before the drain.
>
> Fixes: 657ee5531903 ("perf evlist: Introduce side band thread")
> Link: http://lkml.kernel.org/n/tip-41i888xyim9n5ceyr44jb468@git.kernel.org
> Signed-off-by: Jiri Olsa <jolsa@kernel.org>
Thanks for the fix!
Acked-by: Song Liu <songliubraving@fb.com>
> ---
> tools/perf/util/evlist.c | 14 +++++++++-----
> 1 file changed, 9 insertions(+), 5 deletions(-)
>
> diff --git a/tools/perf/util/evlist.c b/tools/perf/util/evlist.c
> index f2bbae38278d..4b6783ff5813 100644
> --- a/tools/perf/util/evlist.c
> +++ b/tools/perf/util/evlist.c
> @@ -1868,12 +1868,12 @@ static void *perf_evlist__poll_thread(void *arg)
> {
> struct perf_evlist *evlist = arg;
> bool draining = false;
> - int i;
> + int i, done = 0;
> +
> + while (!done) {
> + bool got_data = false;
>
> - while (draining || !(evlist->thread.done)) {
> - if (draining)
> - draining = false;
> - else if (evlist->thread.done)
> + if (evlist->thread.done)
> draining = true;
>
> if (!draining)
> @@ -1894,9 +1894,13 @@ static void *perf_evlist__poll_thread(void *arg)
> pr_warning("cannot locate proper evsel for the side band event\n");
>
> perf_mmap__consume(map);
> + got_data = true;
> }
> perf_mmap__read_done(map);
> }
> +
> + if (draining && !got_data)
> + break;
> }
> return NULL;
> }
> --
> 2.17.2
>
^ permalink raw reply [flat|nested] 31+ messages in thread
* [tip:perf/urgent] perf evlist: Fix side band thread draining
2019-04-16 16:01 ` [PATCH 08/12] perf tools: Fix side band thread draining Jiri Olsa
2019-04-16 19:33 ` Arnaldo Carvalho de Melo
2019-04-16 20:41 ` Song Liu
@ 2019-04-19 17:19 ` tip-bot for Jiri Olsa
2 siblings, 0 replies; 31+ messages in thread
From: tip-bot for Jiri Olsa @ 2019-04-19 17:19 UTC (permalink / raw)
To: linux-tip-commits
Cc: acme, namhyung, jolsa, daniel, ast, alexander.shishkin, peterz,
songliubraving, hpa, mingo, adrian.hunter, tglx, ak,
linux-kernel
Commit-ID: adc6257c4a6f23ff97dca8314fcd33828e2d8db5
Gitweb: https://git.kernel.org/tip/adc6257c4a6f23ff97dca8314fcd33828e2d8db5
Author: Jiri Olsa <jolsa@kernel.org>
AuthorDate: Tue, 16 Apr 2019 18:01:23 +0200
Committer: Arnaldo Carvalho de Melo <acme@redhat.com>
CommitDate: Wed, 17 Apr 2019 14:30:11 -0300
perf evlist: Fix side band thread draining
Current perf_evlist__poll_thread() code could finish without draining
the data. Adding the logic that makes sure we won't finish before the
drain.
Signed-off-by: Jiri Olsa <jolsa@kernel.org>
Cc: Adrian Hunter <adrian.hunter@intel.com>
Cc: Alexander Shishkin <alexander.shishkin@linux.intel.com>
Cc: Alexei Starovoitov <ast@kernel.org>
Cc: Andi Kleen <ak@linux.intel.com>
Cc: Daniel Borkmann <daniel@iogearbox.net>
Cc: Namhyung Kim <namhyung@kernel.org>
Cc: Peter Zijlstra <peterz@infradead.org>
Cc: Song Liu <songliubraving@fb.com>
Fixes: 657ee5531903 ("perf evlist: Introduce side band thread")
Link: http://lkml.kernel.org/r/20190416160127.30203-9-jolsa@kernel.org
Signed-off-by: Arnaldo Carvalho de Melo <acme@redhat.com>
---
tools/perf/util/evlist.c | 14 +++++++++-----
1 file changed, 9 insertions(+), 5 deletions(-)
diff --git a/tools/perf/util/evlist.c b/tools/perf/util/evlist.c
index 6689378ee577..51ead577533f 100644
--- a/tools/perf/util/evlist.c
+++ b/tools/perf/util/evlist.c
@@ -1868,12 +1868,12 @@ static void *perf_evlist__poll_thread(void *arg)
{
struct perf_evlist *evlist = arg;
bool draining = false;
- int i;
+ int i, done = 0;
+
+ while (!done) {
+ bool got_data = false;
- while (draining || !(evlist->thread.done)) {
- if (draining)
- draining = false;
- else if (evlist->thread.done)
+ if (evlist->thread.done)
draining = true;
if (!draining)
@@ -1894,9 +1894,13 @@ static void *perf_evlist__poll_thread(void *arg)
pr_warning("cannot locate proper evsel for the side band event\n");
perf_mmap__consume(map);
+ got_data = true;
}
perf_mmap__read_done(map);
}
+
+ if (draining && !got_data)
+ break;
}
return NULL;
}
^ permalink raw reply related [flat|nested] 31+ messages in thread
* [PATCH 09/12] perf tools: Fix map reference counting
2019-04-16 16:01 [PATCH 00/12] perf tools: Display eBPF code in intel_pt trace Jiri Olsa
` (7 preceding siblings ...)
2019-04-16 16:01 ` [PATCH 08/12] perf tools: Fix side band thread draining Jiri Olsa
@ 2019-04-16 16:01 ` Jiri Olsa
2019-04-16 19:36 ` Arnaldo Carvalho de Melo
2019-04-19 17:20 ` [tip:perf/urgent] " tip-bot for Jiri Olsa
2019-04-16 16:01 ` [PATCH 10/12] perf tools: Keep zero in pgoff bpf map Jiri Olsa
` (2 subsequent siblings)
11 siblings, 2 replies; 31+ messages in thread
From: Jiri Olsa @ 2019-04-16 16:01 UTC (permalink / raw)
To: Arnaldo Carvalho de Melo
Cc: Song Liu, lkml, Ingo Molnar, Namhyung Kim, Alexander Shishkin,
Peter Zijlstra, Andi Kleen, Adrian Hunter, Alexei Starovoitov,
Daniel Borkmann
By calling maps__insert we assume to get 2 references
on the map, which we relese within maps__remove call.
However if there's already same map name, we currently
don't bump the reference and can crash, like:
Program received signal SIGABRT, Aborted.
0x00007ffff75e60f5 in raise () from /lib64/libc.so.6
(gdb) bt
#0 0x00007ffff75e60f5 in raise () from /lib64/libc.so.6
#1 0x00007ffff75d0895 in abort () from /lib64/libc.so.6
#2 0x00007ffff75d0769 in __assert_fail_base.cold () from /lib64/libc.so.6
#3 0x00007ffff75de596 in __assert_fail () from /lib64/libc.so.6
#4 0x00000000004fc006 in refcount_sub_and_test (i=1, r=0x1224e88) at tools/include/linux/refcount.h:131
#5 refcount_dec_and_test (r=0x1224e88) at tools/include/linux/refcount.h:148
#6 map__put (map=0x1224df0) at util/map.c:299
#7 0x00000000004fdb95 in __maps__remove (map=0x1224df0, maps=0xb17d80) at util/map.c:953
#8 maps__remove (maps=0xb17d80, map=0x1224df0) at util/map.c:959
#9 0x00000000004f7d8a in map_groups__remove (map=<optimized out>, mg=<optimized out>) at util/map_groups.h:65
#10 machine__process_ksymbol_unregister (sample=<optimized out>, event=0x7ffff7279670, machine=<optimized out>) at util/machine.c:728
#11 machine__process_ksymbol (machine=<optimized out>, event=0x7ffff7279670, sample=<optimized out>) at util/machine.c:741
#12 0x00000000004fffbb in perf_session__deliver_event (session=0xb11390, event=0x7ffff7279670, tool=0x7fffffffc7b0, file_offset=13936) at util/session.c:1362
#13 0x00000000005039bb in do_flush (show_progress=false, oe=0xb17e80) at util/ordered-events.c:243
#14 __ordered_events__flush (oe=0xb17e80, how=OE_FLUSH__ROUND, timestamp=<optimized out>) at util/ordered-events.c:322
#15 0x00000000005005e4 in perf_session__process_user_event (session=session@entry=0xb11390, event=event@entry=0x7ffff72a4af8,
...
Adding the map on the list and getting the reference event
if we find the map with same name.
Cc: Song Liu <songliubraving@fb.com>
Link: http://lkml.kernel.org/n/tip-38t1ihcy32lvu4xfpm0p7yex@git.kernel.org
Signed-off-by: Jiri Olsa <jolsa@kernel.org>
---
tools/perf/util/map.c | 4 +---
1 file changed, 1 insertion(+), 3 deletions(-)
diff --git a/tools/perf/util/map.c b/tools/perf/util/map.c
index 28d484ef74ae..ee71efb9db62 100644
--- a/tools/perf/util/map.c
+++ b/tools/perf/util/map.c
@@ -926,10 +926,8 @@ static void __maps__insert_name(struct maps *maps, struct map *map)
rc = strcmp(m->dso->short_name, map->dso->short_name);
if (rc < 0)
p = &(*p)->rb_left;
- else if (rc > 0)
- p = &(*p)->rb_right;
else
- return;
+ p = &(*p)->rb_right;
}
rb_link_node(&map->rb_node_name, parent, p);
rb_insert_color(&map->rb_node_name, &maps->names);
--
2.17.2
^ permalink raw reply related [flat|nested] 31+ messages in thread
* Re: [PATCH 09/12] perf tools: Fix map reference counting
2019-04-16 16:01 ` [PATCH 09/12] perf tools: Fix map reference counting Jiri Olsa
@ 2019-04-16 19:36 ` Arnaldo Carvalho de Melo
2019-04-19 17:20 ` [tip:perf/urgent] " tip-bot for Jiri Olsa
1 sibling, 0 replies; 31+ messages in thread
From: Arnaldo Carvalho de Melo @ 2019-04-16 19:36 UTC (permalink / raw)
To: Jiri Olsa
Cc: Song Liu, lkml, Ingo Molnar, Namhyung Kim, Alexander Shishkin,
Peter Zijlstra, Andi Kleen, Adrian Hunter, Alexei Starovoitov,
Daniel Borkmann, Eric Saint-Etienne
Em Tue, Apr 16, 2019 at 06:01:24PM +0200, Jiri Olsa escreveu:
> By calling maps__insert we assume to get 2 references
> on the map, which we relese within maps__remove call.
>
> However if there's already same map name, we currently
> don't bump the reference and can crash, like:
>
> Program received signal SIGABRT, Aborted.
> 0x00007ffff75e60f5 in raise () from /lib64/libc.so.6
>
> (gdb) bt
> #0 0x00007ffff75e60f5 in raise () from /lib64/libc.so.6
> #1 0x00007ffff75d0895 in abort () from /lib64/libc.so.6
> #2 0x00007ffff75d0769 in __assert_fail_base.cold () from /lib64/libc.so.6
> #3 0x00007ffff75de596 in __assert_fail () from /lib64/libc.so.6
> #4 0x00000000004fc006 in refcount_sub_and_test (i=1, r=0x1224e88) at tools/include/linux/refcount.h:131
> #5 refcount_dec_and_test (r=0x1224e88) at tools/include/linux/refcount.h:148
> #6 map__put (map=0x1224df0) at util/map.c:299
> #7 0x00000000004fdb95 in __maps__remove (map=0x1224df0, maps=0xb17d80) at util/map.c:953
> #8 maps__remove (maps=0xb17d80, map=0x1224df0) at util/map.c:959
> #9 0x00000000004f7d8a in map_groups__remove (map=<optimized out>, mg=<optimized out>) at util/map_groups.h:65
> #10 machine__process_ksymbol_unregister (sample=<optimized out>, event=0x7ffff7279670, machine=<optimized out>) at util/machine.c:728
> #11 machine__process_ksymbol (machine=<optimized out>, event=0x7ffff7279670, sample=<optimized out>) at util/machine.c:741
> #12 0x00000000004fffbb in perf_session__deliver_event (session=0xb11390, event=0x7ffff7279670, tool=0x7fffffffc7b0, file_offset=13936) at util/session.c:1362
> #13 0x00000000005039bb in do_flush (show_progress=false, oe=0xb17e80) at util/ordered-events.c:243
> #14 __ordered_events__flush (oe=0xb17e80, how=OE_FLUSH__ROUND, timestamp=<optimized out>) at util/ordered-events.c:322
> #15 0x00000000005005e4 in perf_session__process_user_event (session=session@entry=0xb11390, event=event@entry=0x7ffff72a4af8,
> ...
>
> Adding the map on the list and getting the reference event
> if we find the map with same name.
Added:
Fixes: 1e6285699b30 ("perf symbols: Fix slowness due to -ffunction-section")
Ccing Eric Saint-Etienne <eric.saint.etienne@oracle.com>, the author of
the patch that introduced that function.
And applied to perf/urgent
> Cc: Song Liu <songliubraving@fb.com>
> Link: http://lkml.kernel.org/n/tip-38t1ihcy32lvu4xfpm0p7yex@git.kernel.org
> Signed-off-by: Jiri Olsa <jolsa@kernel.org>
> ---
> tools/perf/util/map.c | 4 +---
> 1 file changed, 1 insertion(+), 3 deletions(-)
>
> diff --git a/tools/perf/util/map.c b/tools/perf/util/map.c
> index 28d484ef74ae..ee71efb9db62 100644
> --- a/tools/perf/util/map.c
> +++ b/tools/perf/util/map.c
> @@ -926,10 +926,8 @@ static void __maps__insert_name(struct maps *maps, struct map *map)
> rc = strcmp(m->dso->short_name, map->dso->short_name);
> if (rc < 0)
> p = &(*p)->rb_left;
> - else if (rc > 0)
> - p = &(*p)->rb_right;
> else
> - return;
> + p = &(*p)->rb_right;
> }
> rb_link_node(&map->rb_node_name, parent, p);
> rb_insert_color(&map->rb_node_name, &maps->names);
> --
> 2.17.2
--
- Arnaldo
^ permalink raw reply [flat|nested] 31+ messages in thread
* [tip:perf/urgent] perf tools: Fix map reference counting
2019-04-16 16:01 ` [PATCH 09/12] perf tools: Fix map reference counting Jiri Olsa
2019-04-16 19:36 ` Arnaldo Carvalho de Melo
@ 2019-04-19 17:20 ` tip-bot for Jiri Olsa
1 sibling, 0 replies; 31+ messages in thread
From: tip-bot for Jiri Olsa @ 2019-04-19 17:20 UTC (permalink / raw)
To: linux-tip-commits
Cc: songliubraving, tglx, adrian.hunter, eric.saint.etienne, jolsa,
hpa, peterz, daniel, namhyung, alexander.shishkin, ak, acme,
linux-kernel, ast, mingo
Commit-ID: b9abbdfa88024d52c8084d8f46ea4f161606c692
Gitweb: https://git.kernel.org/tip/b9abbdfa88024d52c8084d8f46ea4f161606c692
Author: Jiri Olsa <jolsa@kernel.org>
AuthorDate: Tue, 16 Apr 2019 18:01:24 +0200
Committer: Arnaldo Carvalho de Melo <acme@redhat.com>
CommitDate: Wed, 17 Apr 2019 14:30:11 -0300
perf tools: Fix map reference counting
By calling maps__insert() we assume to get 2 references on the map,
which we relese within maps__remove call.
However if there's already same map name, we currently don't bump the
reference and can crash, like:
Program received signal SIGABRT, Aborted.
0x00007ffff75e60f5 in raise () from /lib64/libc.so.6
(gdb) bt
#0 0x00007ffff75e60f5 in raise () from /lib64/libc.so.6
#1 0x00007ffff75d0895 in abort () from /lib64/libc.so.6
#2 0x00007ffff75d0769 in __assert_fail_base.cold () from /lib64/libc.so.6
#3 0x00007ffff75de596 in __assert_fail () from /lib64/libc.so.6
#4 0x00000000004fc006 in refcount_sub_and_test (i=1, r=0x1224e88) at tools/include/linux/refcount.h:131
#5 refcount_dec_and_test (r=0x1224e88) at tools/include/linux/refcount.h:148
#6 map__put (map=0x1224df0) at util/map.c:299
#7 0x00000000004fdb95 in __maps__remove (map=0x1224df0, maps=0xb17d80) at util/map.c:953
#8 maps__remove (maps=0xb17d80, map=0x1224df0) at util/map.c:959
#9 0x00000000004f7d8a in map_groups__remove (map=<optimized out>, mg=<optimized out>) at util/map_groups.h:65
#10 machine__process_ksymbol_unregister (sample=<optimized out>, event=0x7ffff7279670, machine=<optimized out>) at util/machine.c:728
#11 machine__process_ksymbol (machine=<optimized out>, event=0x7ffff7279670, sample=<optimized out>) at util/machine.c:741
#12 0x00000000004fffbb in perf_session__deliver_event (session=0xb11390, event=0x7ffff7279670, tool=0x7fffffffc7b0, file_offset=13936) at util/session.c:1362
#13 0x00000000005039bb in do_flush (show_progress=false, oe=0xb17e80) at util/ordered-events.c:243
#14 __ordered_events__flush (oe=0xb17e80, how=OE_FLUSH__ROUND, timestamp=<optimized out>) at util/ordered-events.c:322
#15 0x00000000005005e4 in perf_session__process_user_event (session=session@entry=0xb11390, event=event@entry=0x7ffff72a4af8,
...
Add the map to the list and getting the reference event if we find the
map with same name.
Signed-off-by: Jiri Olsa <jolsa@kernel.org>
Cc: Adrian Hunter <adrian.hunter@intel.com>
Cc: Alexander Shishkin <alexander.shishkin@linux.intel.com>
Cc: Alexei Starovoitov <ast@kernel.org>
Cc: Andi Kleen <ak@linux.intel.com>
Cc: Daniel Borkmann <daniel@iogearbox.net>
Cc: Eric Saint-Etienne <eric.saint.etienne@oracle.com>
Cc: Namhyung Kim <namhyung@kernel.org>
Cc: Peter Zijlstra <peterz@infradead.org>
Cc: Song Liu <songliubraving@fb.com>
Fixes: 1e6285699b30 ("perf symbols: Fix slowness due to -ffunction-section")
Link: http://lkml.kernel.org/r/20190416160127.30203-10-jolsa@kernel.org
Signed-off-by: Arnaldo Carvalho de Melo <acme@redhat.com>
---
tools/perf/util/map.c | 4 +---
1 file changed, 1 insertion(+), 3 deletions(-)
diff --git a/tools/perf/util/map.c b/tools/perf/util/map.c
index 28d484ef74ae..ee71efb9db62 100644
--- a/tools/perf/util/map.c
+++ b/tools/perf/util/map.c
@@ -926,10 +926,8 @@ static void __maps__insert_name(struct maps *maps, struct map *map)
rc = strcmp(m->dso->short_name, map->dso->short_name);
if (rc < 0)
p = &(*p)->rb_left;
- else if (rc > 0)
- p = &(*p)->rb_right;
else
- return;
+ p = &(*p)->rb_right;
}
rb_link_node(&map->rb_node_name, parent, p);
rb_insert_color(&map->rb_node_name, &maps->names);
^ permalink raw reply related [flat|nested] 31+ messages in thread
* [PATCH 10/12] perf tools: Keep zero in pgoff bpf map
2019-04-16 16:01 [PATCH 00/12] perf tools: Display eBPF code in intel_pt trace Jiri Olsa
` (8 preceding siblings ...)
2019-04-16 16:01 ` [PATCH 09/12] perf tools: Fix map reference counting Jiri Olsa
@ 2019-04-16 16:01 ` Jiri Olsa
2019-04-16 16:01 ` [PATCH 11/12] perf tools: Reuse shared eBPF dso objects Jiri Olsa
2019-04-16 16:01 ` [PATCH 12/12] perf script: Pad dso name for --call-trace Jiri Olsa
11 siblings, 0 replies; 31+ messages in thread
From: Jiri Olsa @ 2019-04-16 16:01 UTC (permalink / raw)
To: Arnaldo Carvalho de Melo
Cc: lkml, Ingo Molnar, Namhyung Kim, Alexander Shishkin,
Peter Zijlstra, Andi Kleen, Adrian Hunter, Song Liu,
Alexei Starovoitov, Daniel Borkmann
With pgoff set to zero, the map__map_ip function
will return bpf address based from 0, which is
what we need when we read the data from bpf dso.
Adding bpf symbols with mapped ip addresses as well.
Link: http://lkml.kernel.org/n/tip-nqgyzbqgekvxqc5tjmsb3da2@git.kernel.org
Signed-off-by: Jiri Olsa <jolsa@kernel.org>
---
tools/perf/util/machine.c | 4 ++--
1 file changed, 2 insertions(+), 2 deletions(-)
diff --git a/tools/perf/util/machine.c b/tools/perf/util/machine.c
index ad0205fbb506..d4aa44489011 100644
--- a/tools/perf/util/machine.c
+++ b/tools/perf/util/machine.c
@@ -704,12 +704,12 @@ static int machine__process_ksymbol_register(struct machine *machine,
return -ENOMEM;
map->start = event->ksymbol_event.addr;
- map->pgoff = map->start;
map->end = map->start + event->ksymbol_event.len;
map_groups__insert(&machine->kmaps, map);
}
- sym = symbol__new(event->ksymbol_event.addr, event->ksymbol_event.len,
+ sym = symbol__new(map->map_ip(map, map->start),
+ event->ksymbol_event.len,
0, 0, event->ksymbol_event.name);
if (!sym)
return -ENOMEM;
--
2.17.2
^ permalink raw reply related [flat|nested] 31+ messages in thread
* [PATCH 11/12] perf tools: Reuse shared eBPF dso objects
2019-04-16 16:01 [PATCH 00/12] perf tools: Display eBPF code in intel_pt trace Jiri Olsa
` (9 preceding siblings ...)
2019-04-16 16:01 ` [PATCH 10/12] perf tools: Keep zero in pgoff bpf map Jiri Olsa
@ 2019-04-16 16:01 ` Jiri Olsa
2019-04-17 6:35 ` Adrian Hunter
2019-04-16 16:01 ` [PATCH 12/12] perf script: Pad dso name for --call-trace Jiri Olsa
11 siblings, 1 reply; 31+ messages in thread
From: Jiri Olsa @ 2019-04-16 16:01 UTC (permalink / raw)
To: Arnaldo Carvalho de Melo
Cc: lkml, Ingo Molnar, Namhyung Kim, Alexander Shishkin,
Peter Zijlstra, Andi Kleen, Adrian Hunter, Song Liu,
Alexei Starovoitov, Daniel Borkmann
The eBPF program can be loaded multiple times
with the same name (tag). We can share dso
objects for those programs.
Link: http://lkml.kernel.org/n/tip-3damf8vq1dryhtpbk5b06jpt@git.kernel.org
Signed-off-by: Jiri Olsa <jolsa@kernel.org>
---
tools/perf/util/machine.c | 14 +++++++++++---
1 file changed, 11 insertions(+), 3 deletions(-)
diff --git a/tools/perf/util/machine.c b/tools/perf/util/machine.c
index d4aa44489011..a60653827891 100644
--- a/tools/perf/util/machine.c
+++ b/tools/perf/util/machine.c
@@ -23,6 +23,7 @@
#include "linux/hash.h"
#include "asm/bug.h"
#include "bpf-event.h"
+#include "dso.h"
#include "sane_ctype.h"
#include <symbol/kallsyms.h>
@@ -699,11 +700,18 @@ static int machine__process_ksymbol_register(struct machine *machine,
map = map_groups__find(&machine->kmaps, event->ksymbol_event.addr);
if (!map) {
- map = dso__new_map(event->ksymbol_event.name);
- if (!map)
+ struct dso *dso;
+
+ dso = dsos__findnew(&machine->dsos, event->ksymbol_event.name);
+ if (!dso)
return -ENOMEM;
- map->start = event->ksymbol_event.addr;
+ map = map__new2(event->ksymbol_event.addr, dso);
+ if (!map) {
+ dso__put(dso);
+ return -ENOMEM;
+ }
+
map->end = map->start + event->ksymbol_event.len;
map_groups__insert(&machine->kmaps, map);
}
--
2.17.2
^ permalink raw reply related [flat|nested] 31+ messages in thread
* Re: [PATCH 11/12] perf tools: Reuse shared eBPF dso objects
2019-04-16 16:01 ` [PATCH 11/12] perf tools: Reuse shared eBPF dso objects Jiri Olsa
@ 2019-04-17 6:35 ` Adrian Hunter
2019-04-17 6:51 ` Jiri Olsa
0 siblings, 1 reply; 31+ messages in thread
From: Adrian Hunter @ 2019-04-17 6:35 UTC (permalink / raw)
To: Jiri Olsa, Arnaldo Carvalho de Melo
Cc: lkml, Ingo Molnar, Namhyung Kim, Alexander Shishkin,
Peter Zijlstra, Andi Kleen, Song Liu, Alexei Starovoitov,
Daniel Borkmann
On 16/04/19 7:01 PM, Jiri Olsa wrote:
> The eBPF program can be loaded multiple times
> with the same name (tag). We can share dso
> objects for those programs.
Doesn't a eBPF program get recompiled differently every time it is loaded?
>
> Link: http://lkml.kernel.org/n/tip-3damf8vq1dryhtpbk5b06jpt@git.kernel.org
> Signed-off-by: Jiri Olsa <jolsa@kernel.org>
> ---
> tools/perf/util/machine.c | 14 +++++++++++---
> 1 file changed, 11 insertions(+), 3 deletions(-)
>
> diff --git a/tools/perf/util/machine.c b/tools/perf/util/machine.c
> index d4aa44489011..a60653827891 100644
> --- a/tools/perf/util/machine.c
> +++ b/tools/perf/util/machine.c
> @@ -23,6 +23,7 @@
> #include "linux/hash.h"
> #include "asm/bug.h"
> #include "bpf-event.h"
> +#include "dso.h"
>
> #include "sane_ctype.h"
> #include <symbol/kallsyms.h>
> @@ -699,11 +700,18 @@ static int machine__process_ksymbol_register(struct machine *machine,
>
> map = map_groups__find(&machine->kmaps, event->ksymbol_event.addr);
> if (!map) {
> - map = dso__new_map(event->ksymbol_event.name);
> - if (!map)
> + struct dso *dso;
> +
> + dso = dsos__findnew(&machine->dsos, event->ksymbol_event.name);
> + if (!dso)
> return -ENOMEM;
>
> - map->start = event->ksymbol_event.addr;
> + map = map__new2(event->ksymbol_event.addr, dso);
> + if (!map) {
> + dso__put(dso);
> + return -ENOMEM;
> + }
> +
> map->end = map->start + event->ksymbol_event.len;
> map_groups__insert(&machine->kmaps, map);
> }
>
^ permalink raw reply [flat|nested] 31+ messages in thread
* Re: [PATCH 11/12] perf tools: Reuse shared eBPF dso objects
2019-04-17 6:35 ` Adrian Hunter
@ 2019-04-17 6:51 ` Jiri Olsa
2019-04-17 6:55 ` Adrian Hunter
0 siblings, 1 reply; 31+ messages in thread
From: Jiri Olsa @ 2019-04-17 6:51 UTC (permalink / raw)
To: Adrian Hunter
Cc: Jiri Olsa, Arnaldo Carvalho de Melo, lkml, Ingo Molnar,
Namhyung Kim, Alexander Shishkin, Peter Zijlstra, Andi Kleen,
Song Liu, Alexei Starovoitov, Daniel Borkmann
On Wed, Apr 17, 2019 at 09:35:32AM +0300, Adrian Hunter wrote:
> On 16/04/19 7:01 PM, Jiri Olsa wrote:
> > The eBPF program can be loaded multiple times
> > with the same name (tag). We can share dso
> > objects for those programs.
>
> Doesn't a eBPF program get recompiled differently every time it is loaded?
yes, but those with same tag are identical
jirka
>
> >
> > Link: http://lkml.kernel.org/n/tip-3damf8vq1dryhtpbk5b06jpt@git.kernel.org
> > Signed-off-by: Jiri Olsa <jolsa@kernel.org>
> > ---
> > tools/perf/util/machine.c | 14 +++++++++++---
> > 1 file changed, 11 insertions(+), 3 deletions(-)
> >
> > diff --git a/tools/perf/util/machine.c b/tools/perf/util/machine.c
> > index d4aa44489011..a60653827891 100644
> > --- a/tools/perf/util/machine.c
> > +++ b/tools/perf/util/machine.c
> > @@ -23,6 +23,7 @@
> > #include "linux/hash.h"
> > #include "asm/bug.h"
> > #include "bpf-event.h"
> > +#include "dso.h"
> >
> > #include "sane_ctype.h"
> > #include <symbol/kallsyms.h>
> > @@ -699,11 +700,18 @@ static int machine__process_ksymbol_register(struct machine *machine,
> >
> > map = map_groups__find(&machine->kmaps, event->ksymbol_event.addr);
> > if (!map) {
> > - map = dso__new_map(event->ksymbol_event.name);
> > - if (!map)
> > + struct dso *dso;
> > +
> > + dso = dsos__findnew(&machine->dsos, event->ksymbol_event.name);
> > + if (!dso)
> > return -ENOMEM;
> >
> > - map->start = event->ksymbol_event.addr;
> > + map = map__new2(event->ksymbol_event.addr, dso);
> > + if (!map) {
> > + dso__put(dso);
> > + return -ENOMEM;
> > + }
> > +
> > map->end = map->start + event->ksymbol_event.len;
> > map_groups__insert(&machine->kmaps, map);
> > }
> >
>
^ permalink raw reply [flat|nested] 31+ messages in thread
* Re: [PATCH 11/12] perf tools: Reuse shared eBPF dso objects
2019-04-17 6:51 ` Jiri Olsa
@ 2019-04-17 6:55 ` Adrian Hunter
2019-04-17 7:32 ` Jiri Olsa
0 siblings, 1 reply; 31+ messages in thread
From: Adrian Hunter @ 2019-04-17 6:55 UTC (permalink / raw)
To: Jiri Olsa
Cc: Jiri Olsa, Arnaldo Carvalho de Melo, lkml, Ingo Molnar,
Namhyung Kim, Alexander Shishkin, Peter Zijlstra, Andi Kleen,
Song Liu, Alexei Starovoitov, Daniel Borkmann
On 17/04/19 9:51 AM, Jiri Olsa wrote:
> On Wed, Apr 17, 2019 at 09:35:32AM +0300, Adrian Hunter wrote:
>> On 16/04/19 7:01 PM, Jiri Olsa wrote:
>>> The eBPF program can be loaded multiple times
>>> with the same name (tag). We can share dso
>>> objects for those programs.
>>
>> Doesn't a eBPF program get recompiled differently every time it is loaded?
>
> yes, but those with same tag are identical
But won't recompiled eBPF programs be different due to blinded constants?
Or perhaps in the future other code randomization?
>
> jirka
>
>>
>>>
>>> Link: http://lkml.kernel.org/n/tip-3damf8vq1dryhtpbk5b06jpt@git.kernel.org
>>> Signed-off-by: Jiri Olsa <jolsa@kernel.org>
>>> ---
>>> tools/perf/util/machine.c | 14 +++++++++++---
>>> 1 file changed, 11 insertions(+), 3 deletions(-)
>>>
>>> diff --git a/tools/perf/util/machine.c b/tools/perf/util/machine.c
>>> index d4aa44489011..a60653827891 100644
>>> --- a/tools/perf/util/machine.c
>>> +++ b/tools/perf/util/machine.c
>>> @@ -23,6 +23,7 @@
>>> #include "linux/hash.h"
>>> #include "asm/bug.h"
>>> #include "bpf-event.h"
>>> +#include "dso.h"
>>>
>>> #include "sane_ctype.h"
>>> #include <symbol/kallsyms.h>
>>> @@ -699,11 +700,18 @@ static int machine__process_ksymbol_register(struct machine *machine,
>>>
>>> map = map_groups__find(&machine->kmaps, event->ksymbol_event.addr);
>>> if (!map) {
>>> - map = dso__new_map(event->ksymbol_event.name);
>>> - if (!map)
>>> + struct dso *dso;
>>> +
>>> + dso = dsos__findnew(&machine->dsos, event->ksymbol_event.name);
>>> + if (!dso)
>>> return -ENOMEM;
>>>
>>> - map->start = event->ksymbol_event.addr;
>>> + map = map__new2(event->ksymbol_event.addr, dso);
>>> + if (!map) {
>>> + dso__put(dso);
>>> + return -ENOMEM;
>>> + }
>>> +
>>> map->end = map->start + event->ksymbol_event.len;
>>> map_groups__insert(&machine->kmaps, map);
>>> }
>>>
>>
>
^ permalink raw reply [flat|nested] 31+ messages in thread
* Re: [PATCH 11/12] perf tools: Reuse shared eBPF dso objects
2019-04-17 6:55 ` Adrian Hunter
@ 2019-04-17 7:32 ` Jiri Olsa
2019-04-17 16:56 ` Song Liu
0 siblings, 1 reply; 31+ messages in thread
From: Jiri Olsa @ 2019-04-17 7:32 UTC (permalink / raw)
To: Adrian Hunter
Cc: Jiri Olsa, Arnaldo Carvalho de Melo, lkml, Ingo Molnar,
Namhyung Kim, Alexander Shishkin, Peter Zijlstra, Andi Kleen,
Song Liu, Alexei Starovoitov, Daniel Borkmann
On Wed, Apr 17, 2019 at 09:55:25AM +0300, Adrian Hunter wrote:
> On 17/04/19 9:51 AM, Jiri Olsa wrote:
> > On Wed, Apr 17, 2019 at 09:35:32AM +0300, Adrian Hunter wrote:
> >> On 16/04/19 7:01 PM, Jiri Olsa wrote:
> >>> The eBPF program can be loaded multiple times
> >>> with the same name (tag). We can share dso
> >>> objects for those programs.
> >>
> >> Doesn't a eBPF program get recompiled differently every time it is loaded?
> >
> > yes, but those with same tag are identical
>
> But won't recompiled eBPF programs be different due to blinded constants?
> Or perhaps in the future other code randomization?
ah right.. that can happen, let's skip this one then
perhaps we could add bpf prog id to be part of the name
to make it unique.. I'll check
thanks,
jirka
>
> >
> > jirka
> >
> >>
> >>>
> >>> Link: http://lkml.kernel.org/n/tip-3damf8vq1dryhtpbk5b06jpt@git.kernel.org
> >>> Signed-off-by: Jiri Olsa <jolsa@kernel.org>
> >>> ---
> >>> tools/perf/util/machine.c | 14 +++++++++++---
> >>> 1 file changed, 11 insertions(+), 3 deletions(-)
> >>>
> >>> diff --git a/tools/perf/util/machine.c b/tools/perf/util/machine.c
> >>> index d4aa44489011..a60653827891 100644
> >>> --- a/tools/perf/util/machine.c
> >>> +++ b/tools/perf/util/machine.c
> >>> @@ -23,6 +23,7 @@
> >>> #include "linux/hash.h"
> >>> #include "asm/bug.h"
> >>> #include "bpf-event.h"
> >>> +#include "dso.h"
> >>>
> >>> #include "sane_ctype.h"
> >>> #include <symbol/kallsyms.h>
> >>> @@ -699,11 +700,18 @@ static int machine__process_ksymbol_register(struct machine *machine,
> >>>
> >>> map = map_groups__find(&machine->kmaps, event->ksymbol_event.addr);
> >>> if (!map) {
> >>> - map = dso__new_map(event->ksymbol_event.name);
> >>> - if (!map)
> >>> + struct dso *dso;
> >>> +
> >>> + dso = dsos__findnew(&machine->dsos, event->ksymbol_event.name);
> >>> + if (!dso)
> >>> return -ENOMEM;
> >>>
> >>> - map->start = event->ksymbol_event.addr;
> >>> + map = map__new2(event->ksymbol_event.addr, dso);
> >>> + if (!map) {
> >>> + dso__put(dso);
> >>> + return -ENOMEM;
> >>> + }
> >>> +
> >>> map->end = map->start + event->ksymbol_event.len;
> >>> map_groups__insert(&machine->kmaps, map);
> >>> }
> >>>
> >>
> >
>
^ permalink raw reply [flat|nested] 31+ messages in thread
* Re: [PATCH 11/12] perf tools: Reuse shared eBPF dso objects
2019-04-17 7:32 ` Jiri Olsa
@ 2019-04-17 16:56 ` Song Liu
0 siblings, 0 replies; 31+ messages in thread
From: Song Liu @ 2019-04-17 16:56 UTC (permalink / raw)
To: Jiri Olsa
Cc: Adrian Hunter, Jiri Olsa, Arnaldo Carvalho de Melo, lkml,
Ingo Molnar, Namhyung Kim, Alexander Shishkin, Peter Zijlstra,
Andi Kleen, Alexei Starovoitov, Daniel Borkmann
> On Apr 17, 2019, at 12:32 AM, Jiri Olsa <jolsa@redhat.com> wrote:
>
> On Wed, Apr 17, 2019 at 09:55:25AM +0300, Adrian Hunter wrote:
>> On 17/04/19 9:51 AM, Jiri Olsa wrote:
>>> On Wed, Apr 17, 2019 at 09:35:32AM +0300, Adrian Hunter wrote:
>>>> On 16/04/19 7:01 PM, Jiri Olsa wrote:
>>>>> The eBPF program can be loaded multiple times
>>>>> with the same name (tag). We can share dso
>>>>> objects for those programs.
>>>>
>>>> Doesn't a eBPF program get recompiled differently every time it is loaded?
>>>
>>> yes, but those with same tag are identical
>>
>> But won't recompiled eBPF programs be different due to blinded constants?
>> Or perhaps in the future other code randomization?
>
> ah right.. that can happen, let's skip this one then
>
> perhaps we could add bpf prog id to be part of the name
> to make it unique.. I'll check
>
> thanks,
> jirka
I thought about similar optimizations. However, this is not easy.
Tag of a BPF program is calculated _before_ JiT, so we really cannot
say two programs with same tag are identical. Current implementation
iterates over all bpf program IDs, so same program id will not appear
twice in the same perf session.
I think the more realistic opportunity of optimizations comes from
sharing the dsos cross multiple perf.data files.
Thanks,
Song
>>
>>>
>>> jirka
>>>
>>>>
>>>>>
>>>>> Link: http://lkml.kernel.org/n/tip-3damf8vq1dryhtpbk5b06jpt@git.kernel.org
>>>>> Signed-off-by: Jiri Olsa <jolsa@kernel.org>
>>>>> ---
>>>>> tools/perf/util/machine.c | 14 +++++++++++---
>>>>> 1 file changed, 11 insertions(+), 3 deletions(-)
>>>>>
>>>>> diff --git a/tools/perf/util/machine.c b/tools/perf/util/machine.c
>>>>> index d4aa44489011..a60653827891 100644
>>>>> --- a/tools/perf/util/machine.c
>>>>> +++ b/tools/perf/util/machine.c
>>>>> @@ -23,6 +23,7 @@
>>>>> #include "linux/hash.h"
>>>>> #include "asm/bug.h"
>>>>> #include "bpf-event.h"
>>>>> +#include "dso.h"
>>>>>
>>>>> #include "sane_ctype.h"
>>>>> #include <symbol/kallsyms.h>
>>>>> @@ -699,11 +700,18 @@ static int machine__process_ksymbol_register(struct machine *machine,
>>>>>
>>>>> map = map_groups__find(&machine->kmaps, event->ksymbol_event.addr);
>>>>> if (!map) {
>>>>> - map = dso__new_map(event->ksymbol_event.name);
>>>>> - if (!map)
>>>>> + struct dso *dso;
>>>>> +
>>>>> + dso = dsos__findnew(&machine->dsos, event->ksymbol_event.name);
>>>>> + if (!dso)
>>>>> return -ENOMEM;
>>>>>
>>>>> - map->start = event->ksymbol_event.addr;
>>>>> + map = map__new2(event->ksymbol_event.addr, dso);
>>>>> + if (!map) {
>>>>> + dso__put(dso);
>>>>> + return -ENOMEM;
>>>>> + }
>>>>> +
>>>>> map->end = map->start + event->ksymbol_event.len;
>>>>> map_groups__insert(&machine->kmaps, map);
>>>>> }
^ permalink raw reply [flat|nested] 31+ messages in thread
* [PATCH 12/12] perf script: Pad dso name for --call-trace
2019-04-16 16:01 [PATCH 00/12] perf tools: Display eBPF code in intel_pt trace Jiri Olsa
` (10 preceding siblings ...)
2019-04-16 16:01 ` [PATCH 11/12] perf tools: Reuse shared eBPF dso objects Jiri Olsa
@ 2019-04-16 16:01 ` Jiri Olsa
11 siblings, 0 replies; 31+ messages in thread
From: Jiri Olsa @ 2019-04-16 16:01 UTC (permalink / raw)
To: Arnaldo Carvalho de Melo
Cc: lkml, Ingo Molnar, Namhyung Kim, Alexander Shishkin,
Peter Zijlstra, Andi Kleen, Adrian Hunter, Song Liu,
Alexei Starovoitov, Daniel Borkmann
so we don't have the indent screwed by different dso name lenghts,
as now for kernel there's also bpf code displayed.
# perf-with-kcore record pt -e intel_pt//ku -- sleep 1
# perf-core/perf-with-kcore script pt --call-trace
Before:
sleep 36660 [016] 1057036.806464404: ([kernel.kallsyms]) kretprobe_perf_func
sleep 36660 [016] 1057036.806464404: ([kernel.kallsyms]) trace_call_bpf
sleep 36660 [016] 1057036.806464404: ([kernel.kallsyms]) __x86_indirect_thunk_rax
sleep 36660 [016] 1057036.806464404: ([kernel.kallsyms]) __x86_indirect_thunk_rax
sleep 36660 [016] 1057036.806464725: (bpf_prog_da4fe6b3d2c29b25_trace_return) bpf_get_current_pid_tgid
sleep 36660 [016] 1057036.806464725: (bpf_prog_da4fe6b3d2c29b25_trace_return) bpf_ktime_get_ns
sleep 36660 [016] 1057036.806464725: ([kernel.kallsyms]) __x86_indirect_thunk_rax
sleep 36660 [016] 1057036.806464725: ([kernel.kallsyms]) __x86_indirect_thunk_rax
sleep 36660 [016] 1057036.806465045: (bpf_prog_da4fe6b3d2c29b25_trace_return) __htab_map_lookup_elem
sleep 36660 [016] 1057036.806465366: ([kernel.kallsyms]) memcmp
sleep 36660 [016] 1057036.806465687: (bpf_prog_da4fe6b3d2c29b25_trace_return) bpf_probe_read
sleep 36660 [016] 1057036.806465687: ([kernel.kallsyms]) probe_kernel_read
sleep 36660 [016] 1057036.806465687: ([kernel.kallsyms]) __check_object_size
sleep 36660 [016] 1057036.806465687: ([kernel.kallsyms]) check_stack_object
sleep 36660 [016] 1057036.806465687: ([kernel.kallsyms]) copy_user_enhanced_fast_string
sleep 36660 [016] 1057036.806465687: (bpf_prog_da4fe6b3d2c29b25_trace_return) bpf_probe_read
sleep 36660 [016] 1057036.806465687: ([kernel.kallsyms]) probe_kernel_read
sleep 36660 [016] 1057036.806465687: ([kernel.kallsyms]) __check_object_size
sleep 36660 [016] 1057036.806465687: ([kernel.kallsyms]) check_stack_object
sleep 36660 [016] 1057036.806465687: ([kernel.kallsyms]) copy_user_enhanced_fast_string
sleep 36660 [016] 1057036.806466008: (bpf_prog_da4fe6b3d2c29b25_trace_return) bpf_get_current_uid_gid
sleep 36660 [016] 1057036.806466008: ([kernel.kallsyms]) from_kgid
sleep 36660 [016] 1057036.806466008: ([kernel.kallsyms]) from_kuid
sleep 36660 [016] 1057036.806466008: (bpf_prog_da4fe6b3d2c29b25_trace_return) bpf_perf_event_output
sleep 36660 [016] 1057036.806466328: ([kernel.kallsyms]) perf_event_output
sleep 36660 [016] 1057036.806466328: ([kernel.kallsyms]) perf_prepare_sample
sleep 36660 [016] 1057036.806466328: ([kernel.kallsyms]) perf_misc_flags
sleep 36660 [016] 1057036.806466328: ([kernel.kallsyms]) __x86_indirect_thunk_rax
sleep 36660 [016] 1057036.806466328: ([kernel.kallsyms]) __x86_indirect_thunk_rax
sleep 36660 [016] 1057036.806466328: ([kvm]) kvm_is_in_guest
sleep 36660 [016] 1057036.806466649: ([kernel.kallsyms]) __perf_event_header__init_id.isra.0
sleep 36660 [016] 1057036.806466649: ([kernel.kallsyms]) perf_output_begin
After:
sleep 36660 [016] 1057036.806464404: ([kernel.kallsyms] ) kretprobe_perf_func
sleep 36660 [016] 1057036.806464404: ([kernel.kallsyms] ) trace_call_bpf
sleep 36660 [016] 1057036.806464404: ([kernel.kallsyms] ) __x86_indirect_thunk_rax
sleep 36660 [016] 1057036.806464404: ([kernel.kallsyms] ) __x86_indirect_thunk_rax
sleep 36660 [016] 1057036.806464725: (bpf_prog_da4fe6b3d2c29b25_trace_return ) bpf_get_current_pid_tgid
sleep 36660 [016] 1057036.806464725: (bpf_prog_da4fe6b3d2c29b25_trace_return ) bpf_ktime_get_ns
sleep 36660 [016] 1057036.806464725: ([kernel.kallsyms] ) __x86_indirect_thunk_rax
sleep 36660 [016] 1057036.806464725: ([kernel.kallsyms] ) __x86_indirect_thunk_rax
sleep 36660 [016] 1057036.806465045: (bpf_prog_da4fe6b3d2c29b25_trace_return ) __htab_map_lookup_elem
sleep 36660 [016] 1057036.806465366: ([kernel.kallsyms] ) memcmp
sleep 36660 [016] 1057036.806465687: (bpf_prog_da4fe6b3d2c29b25_trace_return ) bpf_probe_read
sleep 36660 [016] 1057036.806465687: ([kernel.kallsyms] ) probe_kernel_read
sleep 36660 [016] 1057036.806465687: ([kernel.kallsyms] ) __check_object_size
sleep 36660 [016] 1057036.806465687: ([kernel.kallsyms] ) check_stack_object
sleep 36660 [016] 1057036.806465687: ([kernel.kallsyms] ) copy_user_enhanced_fast_string
sleep 36660 [016] 1057036.806465687: (bpf_prog_da4fe6b3d2c29b25_trace_return ) bpf_probe_read
sleep 36660 [016] 1057036.806465687: ([kernel.kallsyms] ) probe_kernel_read
sleep 36660 [016] 1057036.806465687: ([kernel.kallsyms] ) __check_object_size
sleep 36660 [016] 1057036.806465687: ([kernel.kallsyms] ) check_stack_object
sleep 36660 [016] 1057036.806465687: ([kernel.kallsyms] ) copy_user_enhanced_fast_string
sleep 36660 [016] 1057036.806466008: (bpf_prog_da4fe6b3d2c29b25_trace_return ) bpf_get_current_uid_gid
sleep 36660 [016] 1057036.806466008: ([kernel.kallsyms] ) from_kgid
sleep 36660 [016] 1057036.806466008: ([kernel.kallsyms] ) from_kuid
sleep 36660 [016] 1057036.806466008: (bpf_prog_da4fe6b3d2c29b25_trace_return ) bpf_perf_event_output
sleep 36660 [016] 1057036.806466328: ([kernel.kallsyms] ) perf_event_output
sleep 36660 [016] 1057036.806466328: ([kernel.kallsyms] ) perf_prepare_sample
sleep 36660 [016] 1057036.806466328: ([kernel.kallsyms] ) perf_misc_flags
sleep 36660 [016] 1057036.806466328: ([kernel.kallsyms] ) __x86_indirect_thunk_rax
sleep 36660 [016] 1057036.806466328: ([kernel.kallsyms] ) __x86_indirect_thunk_rax
Link: http://lkml.kernel.org/n/tip-99g9rg4p20a1o99vr0nkjhq8@git.kernel.org
Signed-off-by: Jiri Olsa <jolsa@kernel.org>
---
tools/include/linux/kernel.h | 1 +
tools/lib/vsprintf.c | 19 +++++++++++++++++++
tools/perf/builtin-script.c | 1 +
tools/perf/util/map.c | 6 ++++++
tools/perf/util/symbol_conf.h | 1 +
5 files changed, 28 insertions(+)
diff --git a/tools/include/linux/kernel.h b/tools/include/linux/kernel.h
index 857d9e22826e..cba226948a0c 100644
--- a/tools/include/linux/kernel.h
+++ b/tools/include/linux/kernel.h
@@ -102,6 +102,7 @@
int vscnprintf(char *buf, size_t size, const char *fmt, va_list args);
int scnprintf(char * buf, size_t size, const char * fmt, ...);
+int scnprintf_pad(char * buf, size_t size, const char * fmt, ...);
#define ARRAY_SIZE(arr) (sizeof(arr) / sizeof((arr)[0]) + __must_be_array(arr))
diff --git a/tools/lib/vsprintf.c b/tools/lib/vsprintf.c
index e08ee147eab4..149a15013b23 100644
--- a/tools/lib/vsprintf.c
+++ b/tools/lib/vsprintf.c
@@ -23,3 +23,22 @@ int scnprintf(char * buf, size_t size, const char * fmt, ...)
return (i >= ssize) ? (ssize - 1) : i;
}
+
+int scnprintf_pad(char * buf, size_t size, const char * fmt, ...)
+{
+ ssize_t ssize = size;
+ va_list args;
+ int i;
+
+ va_start(args, fmt);
+ i = vsnprintf(buf, size, fmt, args);
+ va_end(args);
+
+ if (i < (int) size) {
+ for (; i < (int) size; i++)
+ buf[i] = ' ';
+ buf[i] = 0x0;
+ }
+
+ return (i >= ssize) ? (ssize - 1) : i;
+}
diff --git a/tools/perf/builtin-script.c b/tools/perf/builtin-script.c
index 61cfd8f70989..7adaa6c63a0b 100644
--- a/tools/perf/builtin-script.c
+++ b/tools/perf/builtin-script.c
@@ -3297,6 +3297,7 @@ static int parse_call_trace(const struct option *opt __maybe_unused,
parse_output_fields(NULL, "-ip,-addr,-event,-period,+callindent", 0);
itrace_parse_synth_opts(opt, "cewp", 0);
symbol_conf.nanosecs = true;
+ symbol_conf.pad_output_len_dso = 50;
return 0;
}
diff --git a/tools/perf/util/map.c b/tools/perf/util/map.c
index ee71efb9db62..c3fbd6e556b0 100644
--- a/tools/perf/util/map.c
+++ b/tools/perf/util/map.c
@@ -405,6 +405,7 @@ size_t map__fprintf(struct map *map, FILE *fp)
size_t map__fprintf_dsoname(struct map *map, FILE *fp)
{
+ char buf[PATH_MAX];
const char *dsoname = "[unknown]";
if (map && map->dso) {
@@ -414,6 +415,11 @@ size_t map__fprintf_dsoname(struct map *map, FILE *fp)
dsoname = map->dso->name;
}
+ if (symbol_conf.pad_output_len_dso) {
+ scnprintf_pad(buf, symbol_conf.pad_output_len_dso, "%s", dsoname);
+ dsoname = buf;
+ }
+
return fprintf(fp, "%s", dsoname);
}
diff --git a/tools/perf/util/symbol_conf.h b/tools/perf/util/symbol_conf.h
index 6c55fa6fccec..382ba63fc554 100644
--- a/tools/perf/util/symbol_conf.h
+++ b/tools/perf/util/symbol_conf.h
@@ -69,6 +69,7 @@ struct symbol_conf {
*tid_list;
const char *symfs;
int res_sample;
+ int pad_output_len_dso;
};
extern struct symbol_conf symbol_conf;
--
2.17.2
^ permalink raw reply related [flat|nested] 31+ messages in thread