forked from luck/tmp_suning_uos_patched
43b987d23d
test_maps: Tests that queue/stack maps are behaving correctly even in corner cases test_progs: Tests new ebpf helpers Signed-off-by: Mauricio Vasquez B <mauricio.vasquez@polito.it> Acked-by: Song Liu <songliubraving@fb.com> Signed-off-by: Alexei Starovoitov <ast@kernel.org>
1863 lines
51 KiB
C
1863 lines
51 KiB
C
/* Copyright (c) 2017 Facebook
|
|
*
|
|
* This program is free software; you can redistribute it and/or
|
|
* modify it under the terms of version 2 of the GNU General Public
|
|
* License as published by the Free Software Foundation.
|
|
*/
|
|
#include <stdio.h>
|
|
#include <unistd.h>
|
|
#include <errno.h>
|
|
#include <string.h>
|
|
#include <assert.h>
|
|
#include <stdlib.h>
|
|
#include <time.h>
|
|
|
|
#include <linux/types.h>
|
|
typedef __u16 __sum16;
|
|
#include <arpa/inet.h>
|
|
#include <linux/if_ether.h>
|
|
#include <linux/if_packet.h>
|
|
#include <linux/ip.h>
|
|
#include <linux/ipv6.h>
|
|
#include <linux/tcp.h>
|
|
#include <linux/filter.h>
|
|
#include <linux/perf_event.h>
|
|
#include <linux/unistd.h>
|
|
|
|
#include <sys/ioctl.h>
|
|
#include <sys/wait.h>
|
|
#include <sys/types.h>
|
|
#include <fcntl.h>
|
|
|
|
#include <linux/bpf.h>
|
|
#include <linux/err.h>
|
|
#include <bpf/bpf.h>
|
|
#include <bpf/libbpf.h>
|
|
|
|
#include "test_iptunnel_common.h"
|
|
#include "bpf_util.h"
|
|
#include "bpf_endian.h"
|
|
#include "bpf_rlimit.h"
|
|
#include "trace_helpers.h"
|
|
|
|
static int error_cnt, pass_cnt;
|
|
static bool jit_enabled;
|
|
|
|
#define MAGIC_BYTES 123
|
|
|
|
/* ipv4 test vector */
|
|
static struct {
|
|
struct ethhdr eth;
|
|
struct iphdr iph;
|
|
struct tcphdr tcp;
|
|
} __packed pkt_v4 = {
|
|
.eth.h_proto = bpf_htons(ETH_P_IP),
|
|
.iph.ihl = 5,
|
|
.iph.protocol = 6,
|
|
.iph.tot_len = bpf_htons(MAGIC_BYTES),
|
|
.tcp.urg_ptr = 123,
|
|
};
|
|
|
|
/* ipv6 test vector */
|
|
static struct {
|
|
struct ethhdr eth;
|
|
struct ipv6hdr iph;
|
|
struct tcphdr tcp;
|
|
} __packed pkt_v6 = {
|
|
.eth.h_proto = bpf_htons(ETH_P_IPV6),
|
|
.iph.nexthdr = 6,
|
|
.iph.payload_len = bpf_htons(MAGIC_BYTES),
|
|
.tcp.urg_ptr = 123,
|
|
};
|
|
|
|
#define CHECK(condition, tag, format...) ({ \
|
|
int __ret = !!(condition); \
|
|
if (__ret) { \
|
|
error_cnt++; \
|
|
printf("%s:FAIL:%s ", __func__, tag); \
|
|
printf(format); \
|
|
} else { \
|
|
pass_cnt++; \
|
|
printf("%s:PASS:%s %d nsec\n", __func__, tag, duration);\
|
|
} \
|
|
__ret; \
|
|
})
|
|
|
|
static int bpf_find_map(const char *test, struct bpf_object *obj,
|
|
const char *name)
|
|
{
|
|
struct bpf_map *map;
|
|
|
|
map = bpf_object__find_map_by_name(obj, name);
|
|
if (!map) {
|
|
printf("%s:FAIL:map '%s' not found\n", test, name);
|
|
error_cnt++;
|
|
return -1;
|
|
}
|
|
return bpf_map__fd(map);
|
|
}
|
|
|
|
static void test_pkt_access(void)
|
|
{
|
|
const char *file = "./test_pkt_access.o";
|
|
struct bpf_object *obj;
|
|
__u32 duration, retval;
|
|
int err, prog_fd;
|
|
|
|
err = bpf_prog_load(file, BPF_PROG_TYPE_SCHED_CLS, &obj, &prog_fd);
|
|
if (err) {
|
|
error_cnt++;
|
|
return;
|
|
}
|
|
|
|
err = bpf_prog_test_run(prog_fd, 100000, &pkt_v4, sizeof(pkt_v4),
|
|
NULL, NULL, &retval, &duration);
|
|
CHECK(err || retval, "ipv4",
|
|
"err %d errno %d retval %d duration %d\n",
|
|
err, errno, retval, duration);
|
|
|
|
err = bpf_prog_test_run(prog_fd, 100000, &pkt_v6, sizeof(pkt_v6),
|
|
NULL, NULL, &retval, &duration);
|
|
CHECK(err || retval, "ipv6",
|
|
"err %d errno %d retval %d duration %d\n",
|
|
err, errno, retval, duration);
|
|
bpf_object__close(obj);
|
|
}
|
|
|
|
static void test_xdp(void)
|
|
{
|
|
struct vip key4 = {.protocol = 6, .family = AF_INET};
|
|
struct vip key6 = {.protocol = 6, .family = AF_INET6};
|
|
struct iptnl_info value4 = {.family = AF_INET};
|
|
struct iptnl_info value6 = {.family = AF_INET6};
|
|
const char *file = "./test_xdp.o";
|
|
struct bpf_object *obj;
|
|
char buf[128];
|
|
struct ipv6hdr *iph6 = (void *)buf + sizeof(struct ethhdr);
|
|
struct iphdr *iph = (void *)buf + sizeof(struct ethhdr);
|
|
__u32 duration, retval, size;
|
|
int err, prog_fd, map_fd;
|
|
|
|
err = bpf_prog_load(file, BPF_PROG_TYPE_XDP, &obj, &prog_fd);
|
|
if (err) {
|
|
error_cnt++;
|
|
return;
|
|
}
|
|
|
|
map_fd = bpf_find_map(__func__, obj, "vip2tnl");
|
|
if (map_fd < 0)
|
|
goto out;
|
|
bpf_map_update_elem(map_fd, &key4, &value4, 0);
|
|
bpf_map_update_elem(map_fd, &key6, &value6, 0);
|
|
|
|
err = bpf_prog_test_run(prog_fd, 1, &pkt_v4, sizeof(pkt_v4),
|
|
buf, &size, &retval, &duration);
|
|
|
|
CHECK(err || retval != XDP_TX || size != 74 ||
|
|
iph->protocol != IPPROTO_IPIP, "ipv4",
|
|
"err %d errno %d retval %d size %d\n",
|
|
err, errno, retval, size);
|
|
|
|
err = bpf_prog_test_run(prog_fd, 1, &pkt_v6, sizeof(pkt_v6),
|
|
buf, &size, &retval, &duration);
|
|
CHECK(err || retval != XDP_TX || size != 114 ||
|
|
iph6->nexthdr != IPPROTO_IPV6, "ipv6",
|
|
"err %d errno %d retval %d size %d\n",
|
|
err, errno, retval, size);
|
|
out:
|
|
bpf_object__close(obj);
|
|
}
|
|
|
|
static void test_xdp_adjust_tail(void)
|
|
{
|
|
const char *file = "./test_adjust_tail.o";
|
|
struct bpf_object *obj;
|
|
char buf[128];
|
|
__u32 duration, retval, size;
|
|
int err, prog_fd;
|
|
|
|
err = bpf_prog_load(file, BPF_PROG_TYPE_XDP, &obj, &prog_fd);
|
|
if (err) {
|
|
error_cnt++;
|
|
return;
|
|
}
|
|
|
|
err = bpf_prog_test_run(prog_fd, 1, &pkt_v4, sizeof(pkt_v4),
|
|
buf, &size, &retval, &duration);
|
|
|
|
CHECK(err || retval != XDP_DROP,
|
|
"ipv4", "err %d errno %d retval %d size %d\n",
|
|
err, errno, retval, size);
|
|
|
|
err = bpf_prog_test_run(prog_fd, 1, &pkt_v6, sizeof(pkt_v6),
|
|
buf, &size, &retval, &duration);
|
|
CHECK(err || retval != XDP_TX || size != 54,
|
|
"ipv6", "err %d errno %d retval %d size %d\n",
|
|
err, errno, retval, size);
|
|
bpf_object__close(obj);
|
|
}
|
|
|
|
|
|
|
|
#define MAGIC_VAL 0x1234
|
|
#define NUM_ITER 100000
|
|
#define VIP_NUM 5
|
|
|
|
static void test_l4lb(const char *file)
|
|
{
|
|
unsigned int nr_cpus = bpf_num_possible_cpus();
|
|
struct vip key = {.protocol = 6};
|
|
struct vip_meta {
|
|
__u32 flags;
|
|
__u32 vip_num;
|
|
} value = {.vip_num = VIP_NUM};
|
|
__u32 stats_key = VIP_NUM;
|
|
struct vip_stats {
|
|
__u64 bytes;
|
|
__u64 pkts;
|
|
} stats[nr_cpus];
|
|
struct real_definition {
|
|
union {
|
|
__be32 dst;
|
|
__be32 dstv6[4];
|
|
};
|
|
__u8 flags;
|
|
} real_def = {.dst = MAGIC_VAL};
|
|
__u32 ch_key = 11, real_num = 3;
|
|
__u32 duration, retval, size;
|
|
int err, i, prog_fd, map_fd;
|
|
__u64 bytes = 0, pkts = 0;
|
|
struct bpf_object *obj;
|
|
char buf[128];
|
|
u32 *magic = (u32 *)buf;
|
|
|
|
err = bpf_prog_load(file, BPF_PROG_TYPE_SCHED_CLS, &obj, &prog_fd);
|
|
if (err) {
|
|
error_cnt++;
|
|
return;
|
|
}
|
|
|
|
map_fd = bpf_find_map(__func__, obj, "vip_map");
|
|
if (map_fd < 0)
|
|
goto out;
|
|
bpf_map_update_elem(map_fd, &key, &value, 0);
|
|
|
|
map_fd = bpf_find_map(__func__, obj, "ch_rings");
|
|
if (map_fd < 0)
|
|
goto out;
|
|
bpf_map_update_elem(map_fd, &ch_key, &real_num, 0);
|
|
|
|
map_fd = bpf_find_map(__func__, obj, "reals");
|
|
if (map_fd < 0)
|
|
goto out;
|
|
bpf_map_update_elem(map_fd, &real_num, &real_def, 0);
|
|
|
|
err = bpf_prog_test_run(prog_fd, NUM_ITER, &pkt_v4, sizeof(pkt_v4),
|
|
buf, &size, &retval, &duration);
|
|
CHECK(err || retval != 7/*TC_ACT_REDIRECT*/ || size != 54 ||
|
|
*magic != MAGIC_VAL, "ipv4",
|
|
"err %d errno %d retval %d size %d magic %x\n",
|
|
err, errno, retval, size, *magic);
|
|
|
|
err = bpf_prog_test_run(prog_fd, NUM_ITER, &pkt_v6, sizeof(pkt_v6),
|
|
buf, &size, &retval, &duration);
|
|
CHECK(err || retval != 7/*TC_ACT_REDIRECT*/ || size != 74 ||
|
|
*magic != MAGIC_VAL, "ipv6",
|
|
"err %d errno %d retval %d size %d magic %x\n",
|
|
err, errno, retval, size, *magic);
|
|
|
|
map_fd = bpf_find_map(__func__, obj, "stats");
|
|
if (map_fd < 0)
|
|
goto out;
|
|
bpf_map_lookup_elem(map_fd, &stats_key, stats);
|
|
for (i = 0; i < nr_cpus; i++) {
|
|
bytes += stats[i].bytes;
|
|
pkts += stats[i].pkts;
|
|
}
|
|
if (bytes != MAGIC_BYTES * NUM_ITER * 2 || pkts != NUM_ITER * 2) {
|
|
error_cnt++;
|
|
printf("test_l4lb:FAIL:stats %lld %lld\n", bytes, pkts);
|
|
}
|
|
out:
|
|
bpf_object__close(obj);
|
|
}
|
|
|
|
static void test_l4lb_all(void)
|
|
{
|
|
const char *file1 = "./test_l4lb.o";
|
|
const char *file2 = "./test_l4lb_noinline.o";
|
|
|
|
test_l4lb(file1);
|
|
test_l4lb(file2);
|
|
}
|
|
|
|
static void test_xdp_noinline(void)
|
|
{
|
|
const char *file = "./test_xdp_noinline.o";
|
|
unsigned int nr_cpus = bpf_num_possible_cpus();
|
|
struct vip key = {.protocol = 6};
|
|
struct vip_meta {
|
|
__u32 flags;
|
|
__u32 vip_num;
|
|
} value = {.vip_num = VIP_NUM};
|
|
__u32 stats_key = VIP_NUM;
|
|
struct vip_stats {
|
|
__u64 bytes;
|
|
__u64 pkts;
|
|
} stats[nr_cpus];
|
|
struct real_definition {
|
|
union {
|
|
__be32 dst;
|
|
__be32 dstv6[4];
|
|
};
|
|
__u8 flags;
|
|
} real_def = {.dst = MAGIC_VAL};
|
|
__u32 ch_key = 11, real_num = 3;
|
|
__u32 duration, retval, size;
|
|
int err, i, prog_fd, map_fd;
|
|
__u64 bytes = 0, pkts = 0;
|
|
struct bpf_object *obj;
|
|
char buf[128];
|
|
u32 *magic = (u32 *)buf;
|
|
|
|
err = bpf_prog_load(file, BPF_PROG_TYPE_XDP, &obj, &prog_fd);
|
|
if (err) {
|
|
error_cnt++;
|
|
return;
|
|
}
|
|
|
|
map_fd = bpf_find_map(__func__, obj, "vip_map");
|
|
if (map_fd < 0)
|
|
goto out;
|
|
bpf_map_update_elem(map_fd, &key, &value, 0);
|
|
|
|
map_fd = bpf_find_map(__func__, obj, "ch_rings");
|
|
if (map_fd < 0)
|
|
goto out;
|
|
bpf_map_update_elem(map_fd, &ch_key, &real_num, 0);
|
|
|
|
map_fd = bpf_find_map(__func__, obj, "reals");
|
|
if (map_fd < 0)
|
|
goto out;
|
|
bpf_map_update_elem(map_fd, &real_num, &real_def, 0);
|
|
|
|
err = bpf_prog_test_run(prog_fd, NUM_ITER, &pkt_v4, sizeof(pkt_v4),
|
|
buf, &size, &retval, &duration);
|
|
CHECK(err || retval != 1 || size != 54 ||
|
|
*magic != MAGIC_VAL, "ipv4",
|
|
"err %d errno %d retval %d size %d magic %x\n",
|
|
err, errno, retval, size, *magic);
|
|
|
|
err = bpf_prog_test_run(prog_fd, NUM_ITER, &pkt_v6, sizeof(pkt_v6),
|
|
buf, &size, &retval, &duration);
|
|
CHECK(err || retval != 1 || size != 74 ||
|
|
*magic != MAGIC_VAL, "ipv6",
|
|
"err %d errno %d retval %d size %d magic %x\n",
|
|
err, errno, retval, size, *magic);
|
|
|
|
map_fd = bpf_find_map(__func__, obj, "stats");
|
|
if (map_fd < 0)
|
|
goto out;
|
|
bpf_map_lookup_elem(map_fd, &stats_key, stats);
|
|
for (i = 0; i < nr_cpus; i++) {
|
|
bytes += stats[i].bytes;
|
|
pkts += stats[i].pkts;
|
|
}
|
|
if (bytes != MAGIC_BYTES * NUM_ITER * 2 || pkts != NUM_ITER * 2) {
|
|
error_cnt++;
|
|
printf("test_xdp_noinline:FAIL:stats %lld %lld\n", bytes, pkts);
|
|
}
|
|
out:
|
|
bpf_object__close(obj);
|
|
}
|
|
|
|
static void test_tcp_estats(void)
|
|
{
|
|
const char *file = "./test_tcp_estats.o";
|
|
int err, prog_fd;
|
|
struct bpf_object *obj;
|
|
__u32 duration = 0;
|
|
|
|
err = bpf_prog_load(file, BPF_PROG_TYPE_TRACEPOINT, &obj, &prog_fd);
|
|
CHECK(err, "", "err %d errno %d\n", err, errno);
|
|
if (err) {
|
|
error_cnt++;
|
|
return;
|
|
}
|
|
|
|
bpf_object__close(obj);
|
|
}
|
|
|
|
static inline __u64 ptr_to_u64(const void *ptr)
|
|
{
|
|
return (__u64) (unsigned long) ptr;
|
|
}
|
|
|
|
static bool is_jit_enabled(void)
|
|
{
|
|
const char *jit_sysctl = "/proc/sys/net/core/bpf_jit_enable";
|
|
bool enabled = false;
|
|
int sysctl_fd;
|
|
|
|
sysctl_fd = open(jit_sysctl, 0, O_RDONLY);
|
|
if (sysctl_fd != -1) {
|
|
char tmpc;
|
|
|
|
if (read(sysctl_fd, &tmpc, sizeof(tmpc)) == 1)
|
|
enabled = (tmpc != '0');
|
|
close(sysctl_fd);
|
|
}
|
|
|
|
return enabled;
|
|
}
|
|
|
|
static void test_bpf_obj_id(void)
|
|
{
|
|
const __u64 array_magic_value = 0xfaceb00c;
|
|
const __u32 array_key = 0;
|
|
const int nr_iters = 2;
|
|
const char *file = "./test_obj_id.o";
|
|
const char *expected_prog_name = "test_obj_id";
|
|
const char *expected_map_name = "test_map_id";
|
|
const __u64 nsec_per_sec = 1000000000;
|
|
|
|
struct bpf_object *objs[nr_iters];
|
|
int prog_fds[nr_iters], map_fds[nr_iters];
|
|
/* +1 to test for the info_len returned by kernel */
|
|
struct bpf_prog_info prog_infos[nr_iters + 1];
|
|
struct bpf_map_info map_infos[nr_iters + 1];
|
|
/* Each prog only uses one map. +1 to test nr_map_ids
|
|
* returned by kernel.
|
|
*/
|
|
__u32 map_ids[nr_iters + 1];
|
|
char jited_insns[128], xlated_insns[128], zeros[128];
|
|
__u32 i, next_id, info_len, nr_id_found, duration = 0;
|
|
struct timespec real_time_ts, boot_time_ts;
|
|
int err = 0;
|
|
__u64 array_value;
|
|
uid_t my_uid = getuid();
|
|
time_t now, load_time;
|
|
|
|
err = bpf_prog_get_fd_by_id(0);
|
|
CHECK(err >= 0 || errno != ENOENT,
|
|
"get-fd-by-notexist-prog-id", "err %d errno %d\n", err, errno);
|
|
|
|
err = bpf_map_get_fd_by_id(0);
|
|
CHECK(err >= 0 || errno != ENOENT,
|
|
"get-fd-by-notexist-map-id", "err %d errno %d\n", err, errno);
|
|
|
|
for (i = 0; i < nr_iters; i++)
|
|
objs[i] = NULL;
|
|
|
|
/* Check bpf_obj_get_info_by_fd() */
|
|
bzero(zeros, sizeof(zeros));
|
|
for (i = 0; i < nr_iters; i++) {
|
|
now = time(NULL);
|
|
err = bpf_prog_load(file, BPF_PROG_TYPE_SOCKET_FILTER,
|
|
&objs[i], &prog_fds[i]);
|
|
/* test_obj_id.o is a dumb prog. It should never fail
|
|
* to load.
|
|
*/
|
|
if (err)
|
|
error_cnt++;
|
|
assert(!err);
|
|
|
|
/* Insert a magic value to the map */
|
|
map_fds[i] = bpf_find_map(__func__, objs[i], "test_map_id");
|
|
assert(map_fds[i] >= 0);
|
|
err = bpf_map_update_elem(map_fds[i], &array_key,
|
|
&array_magic_value, 0);
|
|
assert(!err);
|
|
|
|
/* Check getting map info */
|
|
info_len = sizeof(struct bpf_map_info) * 2;
|
|
bzero(&map_infos[i], info_len);
|
|
err = bpf_obj_get_info_by_fd(map_fds[i], &map_infos[i],
|
|
&info_len);
|
|
if (CHECK(err ||
|
|
map_infos[i].type != BPF_MAP_TYPE_ARRAY ||
|
|
map_infos[i].key_size != sizeof(__u32) ||
|
|
map_infos[i].value_size != sizeof(__u64) ||
|
|
map_infos[i].max_entries != 1 ||
|
|
map_infos[i].map_flags != 0 ||
|
|
info_len != sizeof(struct bpf_map_info) ||
|
|
strcmp((char *)map_infos[i].name, expected_map_name),
|
|
"get-map-info(fd)",
|
|
"err %d errno %d type %d(%d) info_len %u(%Zu) key_size %u value_size %u max_entries %u map_flags %X name %s(%s)\n",
|
|
err, errno,
|
|
map_infos[i].type, BPF_MAP_TYPE_ARRAY,
|
|
info_len, sizeof(struct bpf_map_info),
|
|
map_infos[i].key_size,
|
|
map_infos[i].value_size,
|
|
map_infos[i].max_entries,
|
|
map_infos[i].map_flags,
|
|
map_infos[i].name, expected_map_name))
|
|
goto done;
|
|
|
|
/* Check getting prog info */
|
|
info_len = sizeof(struct bpf_prog_info) * 2;
|
|
bzero(&prog_infos[i], info_len);
|
|
bzero(jited_insns, sizeof(jited_insns));
|
|
bzero(xlated_insns, sizeof(xlated_insns));
|
|
prog_infos[i].jited_prog_insns = ptr_to_u64(jited_insns);
|
|
prog_infos[i].jited_prog_len = sizeof(jited_insns);
|
|
prog_infos[i].xlated_prog_insns = ptr_to_u64(xlated_insns);
|
|
prog_infos[i].xlated_prog_len = sizeof(xlated_insns);
|
|
prog_infos[i].map_ids = ptr_to_u64(map_ids + i);
|
|
prog_infos[i].nr_map_ids = 2;
|
|
err = clock_gettime(CLOCK_REALTIME, &real_time_ts);
|
|
assert(!err);
|
|
err = clock_gettime(CLOCK_BOOTTIME, &boot_time_ts);
|
|
assert(!err);
|
|
err = bpf_obj_get_info_by_fd(prog_fds[i], &prog_infos[i],
|
|
&info_len);
|
|
load_time = (real_time_ts.tv_sec - boot_time_ts.tv_sec)
|
|
+ (prog_infos[i].load_time / nsec_per_sec);
|
|
if (CHECK(err ||
|
|
prog_infos[i].type != BPF_PROG_TYPE_SOCKET_FILTER ||
|
|
info_len != sizeof(struct bpf_prog_info) ||
|
|
(jit_enabled && !prog_infos[i].jited_prog_len) ||
|
|
(jit_enabled &&
|
|
!memcmp(jited_insns, zeros, sizeof(zeros))) ||
|
|
!prog_infos[i].xlated_prog_len ||
|
|
!memcmp(xlated_insns, zeros, sizeof(zeros)) ||
|
|
load_time < now - 60 || load_time > now + 60 ||
|
|
prog_infos[i].created_by_uid != my_uid ||
|
|
prog_infos[i].nr_map_ids != 1 ||
|
|
*(int *)prog_infos[i].map_ids != map_infos[i].id ||
|
|
strcmp((char *)prog_infos[i].name, expected_prog_name),
|
|
"get-prog-info(fd)",
|
|
"err %d errno %d i %d type %d(%d) info_len %u(%Zu) jit_enabled %d jited_prog_len %u xlated_prog_len %u jited_prog %d xlated_prog %d load_time %lu(%lu) uid %u(%u) nr_map_ids %u(%u) map_id %u(%u) name %s(%s)\n",
|
|
err, errno, i,
|
|
prog_infos[i].type, BPF_PROG_TYPE_SOCKET_FILTER,
|
|
info_len, sizeof(struct bpf_prog_info),
|
|
jit_enabled,
|
|
prog_infos[i].jited_prog_len,
|
|
prog_infos[i].xlated_prog_len,
|
|
!!memcmp(jited_insns, zeros, sizeof(zeros)),
|
|
!!memcmp(xlated_insns, zeros, sizeof(zeros)),
|
|
load_time, now,
|
|
prog_infos[i].created_by_uid, my_uid,
|
|
prog_infos[i].nr_map_ids, 1,
|
|
*(int *)prog_infos[i].map_ids, map_infos[i].id,
|
|
prog_infos[i].name, expected_prog_name))
|
|
goto done;
|
|
}
|
|
|
|
/* Check bpf_prog_get_next_id() */
|
|
nr_id_found = 0;
|
|
next_id = 0;
|
|
while (!bpf_prog_get_next_id(next_id, &next_id)) {
|
|
struct bpf_prog_info prog_info = {};
|
|
__u32 saved_map_id;
|
|
int prog_fd;
|
|
|
|
info_len = sizeof(prog_info);
|
|
|
|
prog_fd = bpf_prog_get_fd_by_id(next_id);
|
|
if (prog_fd < 0 && errno == ENOENT)
|
|
/* The bpf_prog is in the dead row */
|
|
continue;
|
|
if (CHECK(prog_fd < 0, "get-prog-fd(next_id)",
|
|
"prog_fd %d next_id %d errno %d\n",
|
|
prog_fd, next_id, errno))
|
|
break;
|
|
|
|
for (i = 0; i < nr_iters; i++)
|
|
if (prog_infos[i].id == next_id)
|
|
break;
|
|
|
|
if (i == nr_iters)
|
|
continue;
|
|
|
|
nr_id_found++;
|
|
|
|
/* Negative test:
|
|
* prog_info.nr_map_ids = 1
|
|
* prog_info.map_ids = NULL
|
|
*/
|
|
prog_info.nr_map_ids = 1;
|
|
err = bpf_obj_get_info_by_fd(prog_fd, &prog_info, &info_len);
|
|
if (CHECK(!err || errno != EFAULT,
|
|
"get-prog-fd-bad-nr-map-ids", "err %d errno %d(%d)",
|
|
err, errno, EFAULT))
|
|
break;
|
|
bzero(&prog_info, sizeof(prog_info));
|
|
info_len = sizeof(prog_info);
|
|
|
|
saved_map_id = *(int *)(prog_infos[i].map_ids);
|
|
prog_info.map_ids = prog_infos[i].map_ids;
|
|
prog_info.nr_map_ids = 2;
|
|
err = bpf_obj_get_info_by_fd(prog_fd, &prog_info, &info_len);
|
|
prog_infos[i].jited_prog_insns = 0;
|
|
prog_infos[i].xlated_prog_insns = 0;
|
|
CHECK(err || info_len != sizeof(struct bpf_prog_info) ||
|
|
memcmp(&prog_info, &prog_infos[i], info_len) ||
|
|
*(int *)prog_info.map_ids != saved_map_id,
|
|
"get-prog-info(next_id->fd)",
|
|
"err %d errno %d info_len %u(%Zu) memcmp %d map_id %u(%u)\n",
|
|
err, errno, info_len, sizeof(struct bpf_prog_info),
|
|
memcmp(&prog_info, &prog_infos[i], info_len),
|
|
*(int *)prog_info.map_ids, saved_map_id);
|
|
close(prog_fd);
|
|
}
|
|
CHECK(nr_id_found != nr_iters,
|
|
"check total prog id found by get_next_id",
|
|
"nr_id_found %u(%u)\n",
|
|
nr_id_found, nr_iters);
|
|
|
|
/* Check bpf_map_get_next_id() */
|
|
nr_id_found = 0;
|
|
next_id = 0;
|
|
while (!bpf_map_get_next_id(next_id, &next_id)) {
|
|
struct bpf_map_info map_info = {};
|
|
int map_fd;
|
|
|
|
info_len = sizeof(map_info);
|
|
|
|
map_fd = bpf_map_get_fd_by_id(next_id);
|
|
if (map_fd < 0 && errno == ENOENT)
|
|
/* The bpf_map is in the dead row */
|
|
continue;
|
|
if (CHECK(map_fd < 0, "get-map-fd(next_id)",
|
|
"map_fd %d next_id %u errno %d\n",
|
|
map_fd, next_id, errno))
|
|
break;
|
|
|
|
for (i = 0; i < nr_iters; i++)
|
|
if (map_infos[i].id == next_id)
|
|
break;
|
|
|
|
if (i == nr_iters)
|
|
continue;
|
|
|
|
nr_id_found++;
|
|
|
|
err = bpf_map_lookup_elem(map_fd, &array_key, &array_value);
|
|
assert(!err);
|
|
|
|
err = bpf_obj_get_info_by_fd(map_fd, &map_info, &info_len);
|
|
CHECK(err || info_len != sizeof(struct bpf_map_info) ||
|
|
memcmp(&map_info, &map_infos[i], info_len) ||
|
|
array_value != array_magic_value,
|
|
"check get-map-info(next_id->fd)",
|
|
"err %d errno %d info_len %u(%Zu) memcmp %d array_value %llu(%llu)\n",
|
|
err, errno, info_len, sizeof(struct bpf_map_info),
|
|
memcmp(&map_info, &map_infos[i], info_len),
|
|
array_value, array_magic_value);
|
|
|
|
close(map_fd);
|
|
}
|
|
CHECK(nr_id_found != nr_iters,
|
|
"check total map id found by get_next_id",
|
|
"nr_id_found %u(%u)\n",
|
|
nr_id_found, nr_iters);
|
|
|
|
done:
|
|
for (i = 0; i < nr_iters; i++)
|
|
bpf_object__close(objs[i]);
|
|
}
|
|
|
|
static void test_pkt_md_access(void)
|
|
{
|
|
const char *file = "./test_pkt_md_access.o";
|
|
struct bpf_object *obj;
|
|
__u32 duration, retval;
|
|
int err, prog_fd;
|
|
|
|
err = bpf_prog_load(file, BPF_PROG_TYPE_SCHED_CLS, &obj, &prog_fd);
|
|
if (err) {
|
|
error_cnt++;
|
|
return;
|
|
}
|
|
|
|
err = bpf_prog_test_run(prog_fd, 10, &pkt_v4, sizeof(pkt_v4),
|
|
NULL, NULL, &retval, &duration);
|
|
CHECK(err || retval, "",
|
|
"err %d errno %d retval %d duration %d\n",
|
|
err, errno, retval, duration);
|
|
|
|
bpf_object__close(obj);
|
|
}
|
|
|
|
static void test_obj_name(void)
|
|
{
|
|
struct {
|
|
const char *name;
|
|
int success;
|
|
int expected_errno;
|
|
} tests[] = {
|
|
{ "", 1, 0 },
|
|
{ "_123456789ABCDE", 1, 0 },
|
|
{ "_123456789ABCDEF", 0, EINVAL },
|
|
{ "_123456789ABCD\n", 0, EINVAL },
|
|
};
|
|
struct bpf_insn prog[] = {
|
|
BPF_ALU64_IMM(BPF_MOV, BPF_REG_0, 0),
|
|
BPF_EXIT_INSN(),
|
|
};
|
|
__u32 duration = 0;
|
|
int i;
|
|
|
|
for (i = 0; i < sizeof(tests) / sizeof(tests[0]); i++) {
|
|
size_t name_len = strlen(tests[i].name) + 1;
|
|
union bpf_attr attr;
|
|
size_t ncopy;
|
|
int fd;
|
|
|
|
/* test different attr.prog_name during BPF_PROG_LOAD */
|
|
ncopy = name_len < sizeof(attr.prog_name) ?
|
|
name_len : sizeof(attr.prog_name);
|
|
bzero(&attr, sizeof(attr));
|
|
attr.prog_type = BPF_PROG_TYPE_SCHED_CLS;
|
|
attr.insn_cnt = 2;
|
|
attr.insns = ptr_to_u64(prog);
|
|
attr.license = ptr_to_u64("");
|
|
memcpy(attr.prog_name, tests[i].name, ncopy);
|
|
|
|
fd = syscall(__NR_bpf, BPF_PROG_LOAD, &attr, sizeof(attr));
|
|
CHECK((tests[i].success && fd < 0) ||
|
|
(!tests[i].success && fd != -1) ||
|
|
(!tests[i].success && errno != tests[i].expected_errno),
|
|
"check-bpf-prog-name",
|
|
"fd %d(%d) errno %d(%d)\n",
|
|
fd, tests[i].success, errno, tests[i].expected_errno);
|
|
|
|
if (fd != -1)
|
|
close(fd);
|
|
|
|
/* test different attr.map_name during BPF_MAP_CREATE */
|
|
ncopy = name_len < sizeof(attr.map_name) ?
|
|
name_len : sizeof(attr.map_name);
|
|
bzero(&attr, sizeof(attr));
|
|
attr.map_type = BPF_MAP_TYPE_ARRAY;
|
|
attr.key_size = 4;
|
|
attr.value_size = 4;
|
|
attr.max_entries = 1;
|
|
attr.map_flags = 0;
|
|
memcpy(attr.map_name, tests[i].name, ncopy);
|
|
fd = syscall(__NR_bpf, BPF_MAP_CREATE, &attr, sizeof(attr));
|
|
CHECK((tests[i].success && fd < 0) ||
|
|
(!tests[i].success && fd != -1) ||
|
|
(!tests[i].success && errno != tests[i].expected_errno),
|
|
"check-bpf-map-name",
|
|
"fd %d(%d) errno %d(%d)\n",
|
|
fd, tests[i].success, errno, tests[i].expected_errno);
|
|
|
|
if (fd != -1)
|
|
close(fd);
|
|
}
|
|
}
|
|
|
|
static void test_tp_attach_query(void)
|
|
{
|
|
const int num_progs = 3;
|
|
int i, j, bytes, efd, err, prog_fd[num_progs], pmu_fd[num_progs];
|
|
__u32 duration = 0, info_len, saved_prog_ids[num_progs];
|
|
const char *file = "./test_tracepoint.o";
|
|
struct perf_event_query_bpf *query;
|
|
struct perf_event_attr attr = {};
|
|
struct bpf_object *obj[num_progs];
|
|
struct bpf_prog_info prog_info;
|
|
char buf[256];
|
|
|
|
snprintf(buf, sizeof(buf),
|
|
"/sys/kernel/debug/tracing/events/sched/sched_switch/id");
|
|
efd = open(buf, O_RDONLY, 0);
|
|
if (CHECK(efd < 0, "open", "err %d errno %d\n", efd, errno))
|
|
return;
|
|
bytes = read(efd, buf, sizeof(buf));
|
|
close(efd);
|
|
if (CHECK(bytes <= 0 || bytes >= sizeof(buf),
|
|
"read", "bytes %d errno %d\n", bytes, errno))
|
|
return;
|
|
|
|
attr.config = strtol(buf, NULL, 0);
|
|
attr.type = PERF_TYPE_TRACEPOINT;
|
|
attr.sample_type = PERF_SAMPLE_RAW | PERF_SAMPLE_CALLCHAIN;
|
|
attr.sample_period = 1;
|
|
attr.wakeup_events = 1;
|
|
|
|
query = malloc(sizeof(*query) + sizeof(__u32) * num_progs);
|
|
for (i = 0; i < num_progs; i++) {
|
|
err = bpf_prog_load(file, BPF_PROG_TYPE_TRACEPOINT, &obj[i],
|
|
&prog_fd[i]);
|
|
if (CHECK(err, "prog_load", "err %d errno %d\n", err, errno))
|
|
goto cleanup1;
|
|
|
|
bzero(&prog_info, sizeof(prog_info));
|
|
prog_info.jited_prog_len = 0;
|
|
prog_info.xlated_prog_len = 0;
|
|
prog_info.nr_map_ids = 0;
|
|
info_len = sizeof(prog_info);
|
|
err = bpf_obj_get_info_by_fd(prog_fd[i], &prog_info, &info_len);
|
|
if (CHECK(err, "bpf_obj_get_info_by_fd", "err %d errno %d\n",
|
|
err, errno))
|
|
goto cleanup1;
|
|
saved_prog_ids[i] = prog_info.id;
|
|
|
|
pmu_fd[i] = syscall(__NR_perf_event_open, &attr, -1 /* pid */,
|
|
0 /* cpu 0 */, -1 /* group id */,
|
|
0 /* flags */);
|
|
if (CHECK(pmu_fd[i] < 0, "perf_event_open", "err %d errno %d\n",
|
|
pmu_fd[i], errno))
|
|
goto cleanup2;
|
|
err = ioctl(pmu_fd[i], PERF_EVENT_IOC_ENABLE, 0);
|
|
if (CHECK(err, "perf_event_ioc_enable", "err %d errno %d\n",
|
|
err, errno))
|
|
goto cleanup3;
|
|
|
|
if (i == 0) {
|
|
/* check NULL prog array query */
|
|
query->ids_len = num_progs;
|
|
err = ioctl(pmu_fd[i], PERF_EVENT_IOC_QUERY_BPF, query);
|
|
if (CHECK(err || query->prog_cnt != 0,
|
|
"perf_event_ioc_query_bpf",
|
|
"err %d errno %d query->prog_cnt %u\n",
|
|
err, errno, query->prog_cnt))
|
|
goto cleanup3;
|
|
}
|
|
|
|
err = ioctl(pmu_fd[i], PERF_EVENT_IOC_SET_BPF, prog_fd[i]);
|
|
if (CHECK(err, "perf_event_ioc_set_bpf", "err %d errno %d\n",
|
|
err, errno))
|
|
goto cleanup3;
|
|
|
|
if (i == 1) {
|
|
/* try to get # of programs only */
|
|
query->ids_len = 0;
|
|
err = ioctl(pmu_fd[i], PERF_EVENT_IOC_QUERY_BPF, query);
|
|
if (CHECK(err || query->prog_cnt != 2,
|
|
"perf_event_ioc_query_bpf",
|
|
"err %d errno %d query->prog_cnt %u\n",
|
|
err, errno, query->prog_cnt))
|
|
goto cleanup3;
|
|
|
|
/* try a few negative tests */
|
|
/* invalid query pointer */
|
|
err = ioctl(pmu_fd[i], PERF_EVENT_IOC_QUERY_BPF,
|
|
(struct perf_event_query_bpf *)0x1);
|
|
if (CHECK(!err || errno != EFAULT,
|
|
"perf_event_ioc_query_bpf",
|
|
"err %d errno %d\n", err, errno))
|
|
goto cleanup3;
|
|
|
|
/* no enough space */
|
|
query->ids_len = 1;
|
|
err = ioctl(pmu_fd[i], PERF_EVENT_IOC_QUERY_BPF, query);
|
|
if (CHECK(!err || errno != ENOSPC || query->prog_cnt != 2,
|
|
"perf_event_ioc_query_bpf",
|
|
"err %d errno %d query->prog_cnt %u\n",
|
|
err, errno, query->prog_cnt))
|
|
goto cleanup3;
|
|
}
|
|
|
|
query->ids_len = num_progs;
|
|
err = ioctl(pmu_fd[i], PERF_EVENT_IOC_QUERY_BPF, query);
|
|
if (CHECK(err || query->prog_cnt != (i + 1),
|
|
"perf_event_ioc_query_bpf",
|
|
"err %d errno %d query->prog_cnt %u\n",
|
|
err, errno, query->prog_cnt))
|
|
goto cleanup3;
|
|
for (j = 0; j < i + 1; j++)
|
|
if (CHECK(saved_prog_ids[j] != query->ids[j],
|
|
"perf_event_ioc_query_bpf",
|
|
"#%d saved_prog_id %x query prog_id %x\n",
|
|
j, saved_prog_ids[j], query->ids[j]))
|
|
goto cleanup3;
|
|
}
|
|
|
|
i = num_progs - 1;
|
|
for (; i >= 0; i--) {
|
|
cleanup3:
|
|
ioctl(pmu_fd[i], PERF_EVENT_IOC_DISABLE);
|
|
cleanup2:
|
|
close(pmu_fd[i]);
|
|
cleanup1:
|
|
bpf_object__close(obj[i]);
|
|
}
|
|
free(query);
|
|
}
|
|
|
|
static int compare_map_keys(int map1_fd, int map2_fd)
|
|
{
|
|
__u32 key, next_key;
|
|
char val_buf[PERF_MAX_STACK_DEPTH *
|
|
sizeof(struct bpf_stack_build_id)];
|
|
int err;
|
|
|
|
err = bpf_map_get_next_key(map1_fd, NULL, &key);
|
|
if (err)
|
|
return err;
|
|
err = bpf_map_lookup_elem(map2_fd, &key, val_buf);
|
|
if (err)
|
|
return err;
|
|
|
|
while (bpf_map_get_next_key(map1_fd, &key, &next_key) == 0) {
|
|
err = bpf_map_lookup_elem(map2_fd, &next_key, val_buf);
|
|
if (err)
|
|
return err;
|
|
|
|
key = next_key;
|
|
}
|
|
if (errno != ENOENT)
|
|
return -1;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int compare_stack_ips(int smap_fd, int amap_fd, int stack_trace_len)
|
|
{
|
|
__u32 key, next_key, *cur_key_p, *next_key_p;
|
|
char *val_buf1, *val_buf2;
|
|
int i, err = 0;
|
|
|
|
val_buf1 = malloc(stack_trace_len);
|
|
val_buf2 = malloc(stack_trace_len);
|
|
cur_key_p = NULL;
|
|
next_key_p = &key;
|
|
while (bpf_map_get_next_key(smap_fd, cur_key_p, next_key_p) == 0) {
|
|
err = bpf_map_lookup_elem(smap_fd, next_key_p, val_buf1);
|
|
if (err)
|
|
goto out;
|
|
err = bpf_map_lookup_elem(amap_fd, next_key_p, val_buf2);
|
|
if (err)
|
|
goto out;
|
|
for (i = 0; i < stack_trace_len; i++) {
|
|
if (val_buf1[i] != val_buf2[i]) {
|
|
err = -1;
|
|
goto out;
|
|
}
|
|
}
|
|
key = *next_key_p;
|
|
cur_key_p = &key;
|
|
next_key_p = &next_key;
|
|
}
|
|
if (errno != ENOENT)
|
|
err = -1;
|
|
|
|
out:
|
|
free(val_buf1);
|
|
free(val_buf2);
|
|
return err;
|
|
}
|
|
|
|
static void test_stacktrace_map()
|
|
{
|
|
int control_map_fd, stackid_hmap_fd, stackmap_fd, stack_amap_fd;
|
|
const char *file = "./test_stacktrace_map.o";
|
|
int bytes, efd, err, pmu_fd, prog_fd, stack_trace_len;
|
|
struct perf_event_attr attr = {};
|
|
__u32 key, val, duration = 0;
|
|
struct bpf_object *obj;
|
|
char buf[256];
|
|
|
|
err = bpf_prog_load(file, BPF_PROG_TYPE_TRACEPOINT, &obj, &prog_fd);
|
|
if (CHECK(err, "prog_load", "err %d errno %d\n", err, errno))
|
|
return;
|
|
|
|
/* Get the ID for the sched/sched_switch tracepoint */
|
|
snprintf(buf, sizeof(buf),
|
|
"/sys/kernel/debug/tracing/events/sched/sched_switch/id");
|
|
efd = open(buf, O_RDONLY, 0);
|
|
if (CHECK(efd < 0, "open", "err %d errno %d\n", efd, errno))
|
|
goto close_prog;
|
|
|
|
bytes = read(efd, buf, sizeof(buf));
|
|
close(efd);
|
|
if (bytes <= 0 || bytes >= sizeof(buf))
|
|
goto close_prog;
|
|
|
|
/* Open the perf event and attach bpf progrram */
|
|
attr.config = strtol(buf, NULL, 0);
|
|
attr.type = PERF_TYPE_TRACEPOINT;
|
|
attr.sample_type = PERF_SAMPLE_RAW | PERF_SAMPLE_CALLCHAIN;
|
|
attr.sample_period = 1;
|
|
attr.wakeup_events = 1;
|
|
pmu_fd = syscall(__NR_perf_event_open, &attr, -1 /* pid */,
|
|
0 /* cpu 0 */, -1 /* group id */,
|
|
0 /* flags */);
|
|
if (CHECK(pmu_fd < 0, "perf_event_open", "err %d errno %d\n",
|
|
pmu_fd, errno))
|
|
goto close_prog;
|
|
|
|
err = ioctl(pmu_fd, PERF_EVENT_IOC_ENABLE, 0);
|
|
if (err)
|
|
goto disable_pmu;
|
|
|
|
err = ioctl(pmu_fd, PERF_EVENT_IOC_SET_BPF, prog_fd);
|
|
if (err)
|
|
goto disable_pmu;
|
|
|
|
/* find map fds */
|
|
control_map_fd = bpf_find_map(__func__, obj, "control_map");
|
|
if (control_map_fd < 0)
|
|
goto disable_pmu;
|
|
|
|
stackid_hmap_fd = bpf_find_map(__func__, obj, "stackid_hmap");
|
|
if (stackid_hmap_fd < 0)
|
|
goto disable_pmu;
|
|
|
|
stackmap_fd = bpf_find_map(__func__, obj, "stackmap");
|
|
if (stackmap_fd < 0)
|
|
goto disable_pmu;
|
|
|
|
stack_amap_fd = bpf_find_map(__func__, obj, "stack_amap");
|
|
if (stack_amap_fd < 0)
|
|
goto disable_pmu;
|
|
|
|
/* give some time for bpf program run */
|
|
sleep(1);
|
|
|
|
/* disable stack trace collection */
|
|
key = 0;
|
|
val = 1;
|
|
bpf_map_update_elem(control_map_fd, &key, &val, 0);
|
|
|
|
/* for every element in stackid_hmap, we can find a corresponding one
|
|
* in stackmap, and vise versa.
|
|
*/
|
|
err = compare_map_keys(stackid_hmap_fd, stackmap_fd);
|
|
if (CHECK(err, "compare_map_keys stackid_hmap vs. stackmap",
|
|
"err %d errno %d\n", err, errno))
|
|
goto disable_pmu_noerr;
|
|
|
|
err = compare_map_keys(stackmap_fd, stackid_hmap_fd);
|
|
if (CHECK(err, "compare_map_keys stackmap vs. stackid_hmap",
|
|
"err %d errno %d\n", err, errno))
|
|
goto disable_pmu_noerr;
|
|
|
|
stack_trace_len = PERF_MAX_STACK_DEPTH * sizeof(__u64);
|
|
err = compare_stack_ips(stackmap_fd, stack_amap_fd, stack_trace_len);
|
|
if (CHECK(err, "compare_stack_ips stackmap vs. stack_amap",
|
|
"err %d errno %d\n", err, errno))
|
|
goto disable_pmu_noerr;
|
|
|
|
goto disable_pmu_noerr;
|
|
disable_pmu:
|
|
error_cnt++;
|
|
disable_pmu_noerr:
|
|
ioctl(pmu_fd, PERF_EVENT_IOC_DISABLE);
|
|
close(pmu_fd);
|
|
close_prog:
|
|
bpf_object__close(obj);
|
|
}
|
|
|
|
static void test_stacktrace_map_raw_tp()
|
|
{
|
|
int control_map_fd, stackid_hmap_fd, stackmap_fd;
|
|
const char *file = "./test_stacktrace_map.o";
|
|
int efd, err, prog_fd;
|
|
__u32 key, val, duration = 0;
|
|
struct bpf_object *obj;
|
|
|
|
err = bpf_prog_load(file, BPF_PROG_TYPE_RAW_TRACEPOINT, &obj, &prog_fd);
|
|
if (CHECK(err, "prog_load raw tp", "err %d errno %d\n", err, errno))
|
|
return;
|
|
|
|
efd = bpf_raw_tracepoint_open("sched_switch", prog_fd);
|
|
if (CHECK(efd < 0, "raw_tp_open", "err %d errno %d\n", efd, errno))
|
|
goto close_prog;
|
|
|
|
/* find map fds */
|
|
control_map_fd = bpf_find_map(__func__, obj, "control_map");
|
|
if (control_map_fd < 0)
|
|
goto close_prog;
|
|
|
|
stackid_hmap_fd = bpf_find_map(__func__, obj, "stackid_hmap");
|
|
if (stackid_hmap_fd < 0)
|
|
goto close_prog;
|
|
|
|
stackmap_fd = bpf_find_map(__func__, obj, "stackmap");
|
|
if (stackmap_fd < 0)
|
|
goto close_prog;
|
|
|
|
/* give some time for bpf program run */
|
|
sleep(1);
|
|
|
|
/* disable stack trace collection */
|
|
key = 0;
|
|
val = 1;
|
|
bpf_map_update_elem(control_map_fd, &key, &val, 0);
|
|
|
|
/* for every element in stackid_hmap, we can find a corresponding one
|
|
* in stackmap, and vise versa.
|
|
*/
|
|
err = compare_map_keys(stackid_hmap_fd, stackmap_fd);
|
|
if (CHECK(err, "compare_map_keys stackid_hmap vs. stackmap",
|
|
"err %d errno %d\n", err, errno))
|
|
goto close_prog;
|
|
|
|
err = compare_map_keys(stackmap_fd, stackid_hmap_fd);
|
|
if (CHECK(err, "compare_map_keys stackmap vs. stackid_hmap",
|
|
"err %d errno %d\n", err, errno))
|
|
goto close_prog;
|
|
|
|
goto close_prog_noerr;
|
|
close_prog:
|
|
error_cnt++;
|
|
close_prog_noerr:
|
|
bpf_object__close(obj);
|
|
}
|
|
|
|
static int extract_build_id(char *build_id, size_t size)
|
|
{
|
|
FILE *fp;
|
|
char *line = NULL;
|
|
size_t len = 0;
|
|
|
|
fp = popen("readelf -n ./urandom_read | grep 'Build ID'", "r");
|
|
if (fp == NULL)
|
|
return -1;
|
|
|
|
if (getline(&line, &len, fp) == -1)
|
|
goto err;
|
|
fclose(fp);
|
|
|
|
if (len > size)
|
|
len = size;
|
|
memcpy(build_id, line, len);
|
|
build_id[len] = '\0';
|
|
return 0;
|
|
err:
|
|
fclose(fp);
|
|
return -1;
|
|
}
|
|
|
|
static void test_stacktrace_build_id(void)
|
|
{
|
|
int control_map_fd, stackid_hmap_fd, stackmap_fd, stack_amap_fd;
|
|
const char *file = "./test_stacktrace_build_id.o";
|
|
int bytes, efd, err, pmu_fd, prog_fd, stack_trace_len;
|
|
struct perf_event_attr attr = {};
|
|
__u32 key, previous_key, val, duration = 0;
|
|
struct bpf_object *obj;
|
|
char buf[256];
|
|
int i, j;
|
|
struct bpf_stack_build_id id_offs[PERF_MAX_STACK_DEPTH];
|
|
int build_id_matches = 0;
|
|
|
|
err = bpf_prog_load(file, BPF_PROG_TYPE_TRACEPOINT, &obj, &prog_fd);
|
|
if (CHECK(err, "prog_load", "err %d errno %d\n", err, errno))
|
|
goto out;
|
|
|
|
/* Get the ID for the sched/sched_switch tracepoint */
|
|
snprintf(buf, sizeof(buf),
|
|
"/sys/kernel/debug/tracing/events/random/urandom_read/id");
|
|
efd = open(buf, O_RDONLY, 0);
|
|
if (CHECK(efd < 0, "open", "err %d errno %d\n", efd, errno))
|
|
goto close_prog;
|
|
|
|
bytes = read(efd, buf, sizeof(buf));
|
|
close(efd);
|
|
if (CHECK(bytes <= 0 || bytes >= sizeof(buf),
|
|
"read", "bytes %d errno %d\n", bytes, errno))
|
|
goto close_prog;
|
|
|
|
/* Open the perf event and attach bpf progrram */
|
|
attr.config = strtol(buf, NULL, 0);
|
|
attr.type = PERF_TYPE_TRACEPOINT;
|
|
attr.sample_type = PERF_SAMPLE_RAW | PERF_SAMPLE_CALLCHAIN;
|
|
attr.sample_period = 1;
|
|
attr.wakeup_events = 1;
|
|
pmu_fd = syscall(__NR_perf_event_open, &attr, -1 /* pid */,
|
|
0 /* cpu 0 */, -1 /* group id */,
|
|
0 /* flags */);
|
|
if (CHECK(pmu_fd < 0, "perf_event_open", "err %d errno %d\n",
|
|
pmu_fd, errno))
|
|
goto close_prog;
|
|
|
|
err = ioctl(pmu_fd, PERF_EVENT_IOC_ENABLE, 0);
|
|
if (CHECK(err, "perf_event_ioc_enable", "err %d errno %d\n",
|
|
err, errno))
|
|
goto close_pmu;
|
|
|
|
err = ioctl(pmu_fd, PERF_EVENT_IOC_SET_BPF, prog_fd);
|
|
if (CHECK(err, "perf_event_ioc_set_bpf", "err %d errno %d\n",
|
|
err, errno))
|
|
goto disable_pmu;
|
|
|
|
/* find map fds */
|
|
control_map_fd = bpf_find_map(__func__, obj, "control_map");
|
|
if (CHECK(control_map_fd < 0, "bpf_find_map control_map",
|
|
"err %d errno %d\n", err, errno))
|
|
goto disable_pmu;
|
|
|
|
stackid_hmap_fd = bpf_find_map(__func__, obj, "stackid_hmap");
|
|
if (CHECK(stackid_hmap_fd < 0, "bpf_find_map stackid_hmap",
|
|
"err %d errno %d\n", err, errno))
|
|
goto disable_pmu;
|
|
|
|
stackmap_fd = bpf_find_map(__func__, obj, "stackmap");
|
|
if (CHECK(stackmap_fd < 0, "bpf_find_map stackmap", "err %d errno %d\n",
|
|
err, errno))
|
|
goto disable_pmu;
|
|
|
|
stack_amap_fd = bpf_find_map(__func__, obj, "stack_amap");
|
|
if (CHECK(stack_amap_fd < 0, "bpf_find_map stack_amap",
|
|
"err %d errno %d\n", err, errno))
|
|
goto disable_pmu;
|
|
|
|
assert(system("dd if=/dev/urandom of=/dev/zero count=4 2> /dev/null")
|
|
== 0);
|
|
assert(system("./urandom_read") == 0);
|
|
/* disable stack trace collection */
|
|
key = 0;
|
|
val = 1;
|
|
bpf_map_update_elem(control_map_fd, &key, &val, 0);
|
|
|
|
/* for every element in stackid_hmap, we can find a corresponding one
|
|
* in stackmap, and vise versa.
|
|
*/
|
|
err = compare_map_keys(stackid_hmap_fd, stackmap_fd);
|
|
if (CHECK(err, "compare_map_keys stackid_hmap vs. stackmap",
|
|
"err %d errno %d\n", err, errno))
|
|
goto disable_pmu;
|
|
|
|
err = compare_map_keys(stackmap_fd, stackid_hmap_fd);
|
|
if (CHECK(err, "compare_map_keys stackmap vs. stackid_hmap",
|
|
"err %d errno %d\n", err, errno))
|
|
goto disable_pmu;
|
|
|
|
err = extract_build_id(buf, 256);
|
|
|
|
if (CHECK(err, "get build_id with readelf",
|
|
"err %d errno %d\n", err, errno))
|
|
goto disable_pmu;
|
|
|
|
err = bpf_map_get_next_key(stackmap_fd, NULL, &key);
|
|
if (CHECK(err, "get_next_key from stackmap",
|
|
"err %d, errno %d\n", err, errno))
|
|
goto disable_pmu;
|
|
|
|
do {
|
|
char build_id[64];
|
|
|
|
err = bpf_map_lookup_elem(stackmap_fd, &key, id_offs);
|
|
if (CHECK(err, "lookup_elem from stackmap",
|
|
"err %d, errno %d\n", err, errno))
|
|
goto disable_pmu;
|
|
for (i = 0; i < PERF_MAX_STACK_DEPTH; ++i)
|
|
if (id_offs[i].status == BPF_STACK_BUILD_ID_VALID &&
|
|
id_offs[i].offset != 0) {
|
|
for (j = 0; j < 20; ++j)
|
|
sprintf(build_id + 2 * j, "%02x",
|
|
id_offs[i].build_id[j] & 0xff);
|
|
if (strstr(buf, build_id) != NULL)
|
|
build_id_matches = 1;
|
|
}
|
|
previous_key = key;
|
|
} while (bpf_map_get_next_key(stackmap_fd, &previous_key, &key) == 0);
|
|
|
|
if (CHECK(build_id_matches < 1, "build id match",
|
|
"Didn't find expected build ID from the map\n"))
|
|
goto disable_pmu;
|
|
|
|
stack_trace_len = PERF_MAX_STACK_DEPTH
|
|
* sizeof(struct bpf_stack_build_id);
|
|
err = compare_stack_ips(stackmap_fd, stack_amap_fd, stack_trace_len);
|
|
CHECK(err, "compare_stack_ips stackmap vs. stack_amap",
|
|
"err %d errno %d\n", err, errno);
|
|
|
|
disable_pmu:
|
|
ioctl(pmu_fd, PERF_EVENT_IOC_DISABLE);
|
|
|
|
close_pmu:
|
|
close(pmu_fd);
|
|
|
|
close_prog:
|
|
bpf_object__close(obj);
|
|
|
|
out:
|
|
return;
|
|
}
|
|
|
|
static void test_stacktrace_build_id_nmi(void)
|
|
{
|
|
int control_map_fd, stackid_hmap_fd, stackmap_fd, stack_amap_fd;
|
|
const char *file = "./test_stacktrace_build_id.o";
|
|
int err, pmu_fd, prog_fd;
|
|
struct perf_event_attr attr = {
|
|
.sample_freq = 5000,
|
|
.freq = 1,
|
|
.type = PERF_TYPE_HARDWARE,
|
|
.config = PERF_COUNT_HW_CPU_CYCLES,
|
|
};
|
|
__u32 key, previous_key, val, duration = 0;
|
|
struct bpf_object *obj;
|
|
char buf[256];
|
|
int i, j;
|
|
struct bpf_stack_build_id id_offs[PERF_MAX_STACK_DEPTH];
|
|
int build_id_matches = 0;
|
|
|
|
err = bpf_prog_load(file, BPF_PROG_TYPE_PERF_EVENT, &obj, &prog_fd);
|
|
if (CHECK(err, "prog_load", "err %d errno %d\n", err, errno))
|
|
return;
|
|
|
|
pmu_fd = syscall(__NR_perf_event_open, &attr, -1 /* pid */,
|
|
0 /* cpu 0 */, -1 /* group id */,
|
|
0 /* flags */);
|
|
if (CHECK(pmu_fd < 0, "perf_event_open",
|
|
"err %d errno %d. Does the test host support PERF_COUNT_HW_CPU_CYCLES?\n",
|
|
pmu_fd, errno))
|
|
goto close_prog;
|
|
|
|
err = ioctl(pmu_fd, PERF_EVENT_IOC_ENABLE, 0);
|
|
if (CHECK(err, "perf_event_ioc_enable", "err %d errno %d\n",
|
|
err, errno))
|
|
goto close_pmu;
|
|
|
|
err = ioctl(pmu_fd, PERF_EVENT_IOC_SET_BPF, prog_fd);
|
|
if (CHECK(err, "perf_event_ioc_set_bpf", "err %d errno %d\n",
|
|
err, errno))
|
|
goto disable_pmu;
|
|
|
|
/* find map fds */
|
|
control_map_fd = bpf_find_map(__func__, obj, "control_map");
|
|
if (CHECK(control_map_fd < 0, "bpf_find_map control_map",
|
|
"err %d errno %d\n", err, errno))
|
|
goto disable_pmu;
|
|
|
|
stackid_hmap_fd = bpf_find_map(__func__, obj, "stackid_hmap");
|
|
if (CHECK(stackid_hmap_fd < 0, "bpf_find_map stackid_hmap",
|
|
"err %d errno %d\n", err, errno))
|
|
goto disable_pmu;
|
|
|
|
stackmap_fd = bpf_find_map(__func__, obj, "stackmap");
|
|
if (CHECK(stackmap_fd < 0, "bpf_find_map stackmap", "err %d errno %d\n",
|
|
err, errno))
|
|
goto disable_pmu;
|
|
|
|
stack_amap_fd = bpf_find_map(__func__, obj, "stack_amap");
|
|
if (CHECK(stack_amap_fd < 0, "bpf_find_map stack_amap",
|
|
"err %d errno %d\n", err, errno))
|
|
goto disable_pmu;
|
|
|
|
assert(system("dd if=/dev/urandom of=/dev/zero count=4 2> /dev/null")
|
|
== 0);
|
|
assert(system("taskset 0x1 ./urandom_read 100000") == 0);
|
|
/* disable stack trace collection */
|
|
key = 0;
|
|
val = 1;
|
|
bpf_map_update_elem(control_map_fd, &key, &val, 0);
|
|
|
|
/* for every element in stackid_hmap, we can find a corresponding one
|
|
* in stackmap, and vise versa.
|
|
*/
|
|
err = compare_map_keys(stackid_hmap_fd, stackmap_fd);
|
|
if (CHECK(err, "compare_map_keys stackid_hmap vs. stackmap",
|
|
"err %d errno %d\n", err, errno))
|
|
goto disable_pmu;
|
|
|
|
err = compare_map_keys(stackmap_fd, stackid_hmap_fd);
|
|
if (CHECK(err, "compare_map_keys stackmap vs. stackid_hmap",
|
|
"err %d errno %d\n", err, errno))
|
|
goto disable_pmu;
|
|
|
|
err = extract_build_id(buf, 256);
|
|
|
|
if (CHECK(err, "get build_id with readelf",
|
|
"err %d errno %d\n", err, errno))
|
|
goto disable_pmu;
|
|
|
|
err = bpf_map_get_next_key(stackmap_fd, NULL, &key);
|
|
if (CHECK(err, "get_next_key from stackmap",
|
|
"err %d, errno %d\n", err, errno))
|
|
goto disable_pmu;
|
|
|
|
do {
|
|
char build_id[64];
|
|
|
|
err = bpf_map_lookup_elem(stackmap_fd, &key, id_offs);
|
|
if (CHECK(err, "lookup_elem from stackmap",
|
|
"err %d, errno %d\n", err, errno))
|
|
goto disable_pmu;
|
|
for (i = 0; i < PERF_MAX_STACK_DEPTH; ++i)
|
|
if (id_offs[i].status == BPF_STACK_BUILD_ID_VALID &&
|
|
id_offs[i].offset != 0) {
|
|
for (j = 0; j < 20; ++j)
|
|
sprintf(build_id + 2 * j, "%02x",
|
|
id_offs[i].build_id[j] & 0xff);
|
|
if (strstr(buf, build_id) != NULL)
|
|
build_id_matches = 1;
|
|
}
|
|
previous_key = key;
|
|
} while (bpf_map_get_next_key(stackmap_fd, &previous_key, &key) == 0);
|
|
|
|
if (CHECK(build_id_matches < 1, "build id match",
|
|
"Didn't find expected build ID from the map\n"))
|
|
goto disable_pmu;
|
|
|
|
/*
|
|
* We intentionally skip compare_stack_ips(). This is because we
|
|
* only support one in_nmi() ips-to-build_id translation per cpu
|
|
* at any time, thus stack_amap here will always fallback to
|
|
* BPF_STACK_BUILD_ID_IP;
|
|
*/
|
|
|
|
disable_pmu:
|
|
ioctl(pmu_fd, PERF_EVENT_IOC_DISABLE);
|
|
|
|
close_pmu:
|
|
close(pmu_fd);
|
|
|
|
close_prog:
|
|
bpf_object__close(obj);
|
|
}
|
|
|
|
#define MAX_CNT_RAWTP 10ull
|
|
#define MAX_STACK_RAWTP 100
|
|
struct get_stack_trace_t {
|
|
int pid;
|
|
int kern_stack_size;
|
|
int user_stack_size;
|
|
int user_stack_buildid_size;
|
|
__u64 kern_stack[MAX_STACK_RAWTP];
|
|
__u64 user_stack[MAX_STACK_RAWTP];
|
|
struct bpf_stack_build_id user_stack_buildid[MAX_STACK_RAWTP];
|
|
};
|
|
|
|
static int get_stack_print_output(void *data, int size)
|
|
{
|
|
bool good_kern_stack = false, good_user_stack = false;
|
|
const char *nonjit_func = "___bpf_prog_run";
|
|
struct get_stack_trace_t *e = data;
|
|
int i, num_stack;
|
|
static __u64 cnt;
|
|
struct ksym *ks;
|
|
|
|
cnt++;
|
|
|
|
if (size < sizeof(struct get_stack_trace_t)) {
|
|
__u64 *raw_data = data;
|
|
bool found = false;
|
|
|
|
num_stack = size / sizeof(__u64);
|
|
/* If jit is enabled, we do not have a good way to
|
|
* verify the sanity of the kernel stack. So we
|
|
* just assume it is good if the stack is not empty.
|
|
* This could be improved in the future.
|
|
*/
|
|
if (jit_enabled) {
|
|
found = num_stack > 0;
|
|
} else {
|
|
for (i = 0; i < num_stack; i++) {
|
|
ks = ksym_search(raw_data[i]);
|
|
if (strcmp(ks->name, nonjit_func) == 0) {
|
|
found = true;
|
|
break;
|
|
}
|
|
}
|
|
}
|
|
if (found) {
|
|
good_kern_stack = true;
|
|
good_user_stack = true;
|
|
}
|
|
} else {
|
|
num_stack = e->kern_stack_size / sizeof(__u64);
|
|
if (jit_enabled) {
|
|
good_kern_stack = num_stack > 0;
|
|
} else {
|
|
for (i = 0; i < num_stack; i++) {
|
|
ks = ksym_search(e->kern_stack[i]);
|
|
if (strcmp(ks->name, nonjit_func) == 0) {
|
|
good_kern_stack = true;
|
|
break;
|
|
}
|
|
}
|
|
}
|
|
if (e->user_stack_size > 0 && e->user_stack_buildid_size > 0)
|
|
good_user_stack = true;
|
|
}
|
|
if (!good_kern_stack || !good_user_stack)
|
|
return LIBBPF_PERF_EVENT_ERROR;
|
|
|
|
if (cnt == MAX_CNT_RAWTP)
|
|
return LIBBPF_PERF_EVENT_DONE;
|
|
|
|
return LIBBPF_PERF_EVENT_CONT;
|
|
}
|
|
|
|
static void test_get_stack_raw_tp(void)
|
|
{
|
|
const char *file = "./test_get_stack_rawtp.o";
|
|
int i, efd, err, prog_fd, pmu_fd, perfmap_fd;
|
|
struct perf_event_attr attr = {};
|
|
struct timespec tv = {0, 10};
|
|
__u32 key = 0, duration = 0;
|
|
struct bpf_object *obj;
|
|
|
|
err = bpf_prog_load(file, BPF_PROG_TYPE_RAW_TRACEPOINT, &obj, &prog_fd);
|
|
if (CHECK(err, "prog_load raw tp", "err %d errno %d\n", err, errno))
|
|
return;
|
|
|
|
efd = bpf_raw_tracepoint_open("sys_enter", prog_fd);
|
|
if (CHECK(efd < 0, "raw_tp_open", "err %d errno %d\n", efd, errno))
|
|
goto close_prog;
|
|
|
|
perfmap_fd = bpf_find_map(__func__, obj, "perfmap");
|
|
if (CHECK(perfmap_fd < 0, "bpf_find_map", "err %d errno %d\n",
|
|
perfmap_fd, errno))
|
|
goto close_prog;
|
|
|
|
err = load_kallsyms();
|
|
if (CHECK(err < 0, "load_kallsyms", "err %d errno %d\n", err, errno))
|
|
goto close_prog;
|
|
|
|
attr.sample_type = PERF_SAMPLE_RAW;
|
|
attr.type = PERF_TYPE_SOFTWARE;
|
|
attr.config = PERF_COUNT_SW_BPF_OUTPUT;
|
|
pmu_fd = syscall(__NR_perf_event_open, &attr, getpid()/*pid*/, -1/*cpu*/,
|
|
-1/*group_fd*/, 0);
|
|
if (CHECK(pmu_fd < 0, "perf_event_open", "err %d errno %d\n", pmu_fd,
|
|
errno))
|
|
goto close_prog;
|
|
|
|
err = bpf_map_update_elem(perfmap_fd, &key, &pmu_fd, BPF_ANY);
|
|
if (CHECK(err < 0, "bpf_map_update_elem", "err %d errno %d\n", err,
|
|
errno))
|
|
goto close_prog;
|
|
|
|
err = ioctl(pmu_fd, PERF_EVENT_IOC_ENABLE, 0);
|
|
if (CHECK(err < 0, "ioctl PERF_EVENT_IOC_ENABLE", "err %d errno %d\n",
|
|
err, errno))
|
|
goto close_prog;
|
|
|
|
err = perf_event_mmap(pmu_fd);
|
|
if (CHECK(err < 0, "perf_event_mmap", "err %d errno %d\n", err, errno))
|
|
goto close_prog;
|
|
|
|
/* trigger some syscall action */
|
|
for (i = 0; i < MAX_CNT_RAWTP; i++)
|
|
nanosleep(&tv, NULL);
|
|
|
|
err = perf_event_poller(pmu_fd, get_stack_print_output);
|
|
if (CHECK(err < 0, "perf_event_poller", "err %d errno %d\n", err, errno))
|
|
goto close_prog;
|
|
|
|
goto close_prog_noerr;
|
|
close_prog:
|
|
error_cnt++;
|
|
close_prog_noerr:
|
|
bpf_object__close(obj);
|
|
}
|
|
|
|
static void test_task_fd_query_rawtp(void)
|
|
{
|
|
const char *file = "./test_get_stack_rawtp.o";
|
|
__u64 probe_offset, probe_addr;
|
|
__u32 len, prog_id, fd_type;
|
|
struct bpf_object *obj;
|
|
int efd, err, prog_fd;
|
|
__u32 duration = 0;
|
|
char buf[256];
|
|
|
|
err = bpf_prog_load(file, BPF_PROG_TYPE_RAW_TRACEPOINT, &obj, &prog_fd);
|
|
if (CHECK(err, "prog_load raw tp", "err %d errno %d\n", err, errno))
|
|
return;
|
|
|
|
efd = bpf_raw_tracepoint_open("sys_enter", prog_fd);
|
|
if (CHECK(efd < 0, "raw_tp_open", "err %d errno %d\n", efd, errno))
|
|
goto close_prog;
|
|
|
|
/* query (getpid(), efd) */
|
|
len = sizeof(buf);
|
|
err = bpf_task_fd_query(getpid(), efd, 0, buf, &len, &prog_id,
|
|
&fd_type, &probe_offset, &probe_addr);
|
|
if (CHECK(err < 0, "bpf_task_fd_query", "err %d errno %d\n", err,
|
|
errno))
|
|
goto close_prog;
|
|
|
|
err = fd_type == BPF_FD_TYPE_RAW_TRACEPOINT &&
|
|
strcmp(buf, "sys_enter") == 0;
|
|
if (CHECK(!err, "check_results", "fd_type %d tp_name %s\n",
|
|
fd_type, buf))
|
|
goto close_prog;
|
|
|
|
/* test zero len */
|
|
len = 0;
|
|
err = bpf_task_fd_query(getpid(), efd, 0, buf, &len, &prog_id,
|
|
&fd_type, &probe_offset, &probe_addr);
|
|
if (CHECK(err < 0, "bpf_task_fd_query (len = 0)", "err %d errno %d\n",
|
|
err, errno))
|
|
goto close_prog;
|
|
err = fd_type == BPF_FD_TYPE_RAW_TRACEPOINT &&
|
|
len == strlen("sys_enter");
|
|
if (CHECK(!err, "check_results", "fd_type %d len %u\n", fd_type, len))
|
|
goto close_prog;
|
|
|
|
/* test empty buffer */
|
|
len = sizeof(buf);
|
|
err = bpf_task_fd_query(getpid(), efd, 0, 0, &len, &prog_id,
|
|
&fd_type, &probe_offset, &probe_addr);
|
|
if (CHECK(err < 0, "bpf_task_fd_query (buf = 0)", "err %d errno %d\n",
|
|
err, errno))
|
|
goto close_prog;
|
|
err = fd_type == BPF_FD_TYPE_RAW_TRACEPOINT &&
|
|
len == strlen("sys_enter");
|
|
if (CHECK(!err, "check_results", "fd_type %d len %u\n", fd_type, len))
|
|
goto close_prog;
|
|
|
|
/* test smaller buffer */
|
|
len = 3;
|
|
err = bpf_task_fd_query(getpid(), efd, 0, buf, &len, &prog_id,
|
|
&fd_type, &probe_offset, &probe_addr);
|
|
if (CHECK(err >= 0 || errno != ENOSPC, "bpf_task_fd_query (len = 3)",
|
|
"err %d errno %d\n", err, errno))
|
|
goto close_prog;
|
|
err = fd_type == BPF_FD_TYPE_RAW_TRACEPOINT &&
|
|
len == strlen("sys_enter") &&
|
|
strcmp(buf, "sy") == 0;
|
|
if (CHECK(!err, "check_results", "fd_type %d len %u\n", fd_type, len))
|
|
goto close_prog;
|
|
|
|
goto close_prog_noerr;
|
|
close_prog:
|
|
error_cnt++;
|
|
close_prog_noerr:
|
|
bpf_object__close(obj);
|
|
}
|
|
|
|
static void test_task_fd_query_tp_core(const char *probe_name,
|
|
const char *tp_name)
|
|
{
|
|
const char *file = "./test_tracepoint.o";
|
|
int err, bytes, efd, prog_fd, pmu_fd;
|
|
struct perf_event_attr attr = {};
|
|
__u64 probe_offset, probe_addr;
|
|
__u32 len, prog_id, fd_type;
|
|
struct bpf_object *obj;
|
|
__u32 duration = 0;
|
|
char buf[256];
|
|
|
|
err = bpf_prog_load(file, BPF_PROG_TYPE_TRACEPOINT, &obj, &prog_fd);
|
|
if (CHECK(err, "bpf_prog_load", "err %d errno %d\n", err, errno))
|
|
goto close_prog;
|
|
|
|
snprintf(buf, sizeof(buf),
|
|
"/sys/kernel/debug/tracing/events/%s/id", probe_name);
|
|
efd = open(buf, O_RDONLY, 0);
|
|
if (CHECK(efd < 0, "open", "err %d errno %d\n", efd, errno))
|
|
goto close_prog;
|
|
bytes = read(efd, buf, sizeof(buf));
|
|
close(efd);
|
|
if (CHECK(bytes <= 0 || bytes >= sizeof(buf), "read",
|
|
"bytes %d errno %d\n", bytes, errno))
|
|
goto close_prog;
|
|
|
|
attr.config = strtol(buf, NULL, 0);
|
|
attr.type = PERF_TYPE_TRACEPOINT;
|
|
attr.sample_type = PERF_SAMPLE_RAW;
|
|
attr.sample_period = 1;
|
|
attr.wakeup_events = 1;
|
|
pmu_fd = syscall(__NR_perf_event_open, &attr, -1 /* pid */,
|
|
0 /* cpu 0 */, -1 /* group id */,
|
|
0 /* flags */);
|
|
if (CHECK(err, "perf_event_open", "err %d errno %d\n", err, errno))
|
|
goto close_pmu;
|
|
|
|
err = ioctl(pmu_fd, PERF_EVENT_IOC_ENABLE, 0);
|
|
if (CHECK(err, "perf_event_ioc_enable", "err %d errno %d\n", err,
|
|
errno))
|
|
goto close_pmu;
|
|
|
|
err = ioctl(pmu_fd, PERF_EVENT_IOC_SET_BPF, prog_fd);
|
|
if (CHECK(err, "perf_event_ioc_set_bpf", "err %d errno %d\n", err,
|
|
errno))
|
|
goto close_pmu;
|
|
|
|
/* query (getpid(), pmu_fd) */
|
|
len = sizeof(buf);
|
|
err = bpf_task_fd_query(getpid(), pmu_fd, 0, buf, &len, &prog_id,
|
|
&fd_type, &probe_offset, &probe_addr);
|
|
if (CHECK(err < 0, "bpf_task_fd_query", "err %d errno %d\n", err,
|
|
errno))
|
|
goto close_pmu;
|
|
|
|
err = (fd_type == BPF_FD_TYPE_TRACEPOINT) && !strcmp(buf, tp_name);
|
|
if (CHECK(!err, "check_results", "fd_type %d tp_name %s\n",
|
|
fd_type, buf))
|
|
goto close_pmu;
|
|
|
|
close(pmu_fd);
|
|
goto close_prog_noerr;
|
|
|
|
close_pmu:
|
|
close(pmu_fd);
|
|
close_prog:
|
|
error_cnt++;
|
|
close_prog_noerr:
|
|
bpf_object__close(obj);
|
|
}
|
|
|
|
static void test_task_fd_query_tp(void)
|
|
{
|
|
test_task_fd_query_tp_core("sched/sched_switch",
|
|
"sched_switch");
|
|
test_task_fd_query_tp_core("syscalls/sys_enter_read",
|
|
"sys_enter_read");
|
|
}
|
|
|
|
static void test_reference_tracking()
|
|
{
|
|
const char *file = "./test_sk_lookup_kern.o";
|
|
struct bpf_object *obj;
|
|
struct bpf_program *prog;
|
|
__u32 duration;
|
|
int err = 0;
|
|
|
|
obj = bpf_object__open(file);
|
|
if (IS_ERR(obj)) {
|
|
error_cnt++;
|
|
return;
|
|
}
|
|
|
|
bpf_object__for_each_program(prog, obj) {
|
|
const char *title;
|
|
|
|
/* Ignore .text sections */
|
|
title = bpf_program__title(prog, false);
|
|
if (strstr(title, ".text") != NULL)
|
|
continue;
|
|
|
|
bpf_program__set_type(prog, BPF_PROG_TYPE_SCHED_CLS);
|
|
|
|
/* Expect verifier failure if test name has 'fail' */
|
|
if (strstr(title, "fail") != NULL) {
|
|
libbpf_set_print(NULL, NULL, NULL);
|
|
err = !bpf_program__load(prog, "GPL", 0);
|
|
libbpf_set_print(printf, printf, NULL);
|
|
} else {
|
|
err = bpf_program__load(prog, "GPL", 0);
|
|
}
|
|
CHECK(err, title, "\n");
|
|
}
|
|
bpf_object__close(obj);
|
|
}
|
|
|
|
enum {
|
|
QUEUE,
|
|
STACK,
|
|
};
|
|
|
|
static void test_queue_stack_map(int type)
|
|
{
|
|
const int MAP_SIZE = 32;
|
|
__u32 vals[MAP_SIZE], duration, retval, size, val;
|
|
int i, err, prog_fd, map_in_fd, map_out_fd;
|
|
char file[32], buf[128];
|
|
struct bpf_object *obj;
|
|
struct iphdr *iph = (void *)buf + sizeof(struct ethhdr);
|
|
|
|
/* Fill test values to be used */
|
|
for (i = 0; i < MAP_SIZE; i++)
|
|
vals[i] = rand();
|
|
|
|
if (type == QUEUE)
|
|
strncpy(file, "./test_queue_map.o", sizeof(file));
|
|
else if (type == STACK)
|
|
strncpy(file, "./test_stack_map.o", sizeof(file));
|
|
else
|
|
return;
|
|
|
|
err = bpf_prog_load(file, BPF_PROG_TYPE_SCHED_CLS, &obj, &prog_fd);
|
|
if (err) {
|
|
error_cnt++;
|
|
return;
|
|
}
|
|
|
|
map_in_fd = bpf_find_map(__func__, obj, "map_in");
|
|
if (map_in_fd < 0)
|
|
goto out;
|
|
|
|
map_out_fd = bpf_find_map(__func__, obj, "map_out");
|
|
if (map_out_fd < 0)
|
|
goto out;
|
|
|
|
/* Push 32 elements to the input map */
|
|
for (i = 0; i < MAP_SIZE; i++) {
|
|
err = bpf_map_update_elem(map_in_fd, NULL, &vals[i], 0);
|
|
if (err) {
|
|
error_cnt++;
|
|
goto out;
|
|
}
|
|
}
|
|
|
|
/* The eBPF program pushes iph.saddr in the output map,
|
|
* pops the input map and saves this value in iph.daddr
|
|
*/
|
|
for (i = 0; i < MAP_SIZE; i++) {
|
|
if (type == QUEUE) {
|
|
val = vals[i];
|
|
pkt_v4.iph.saddr = vals[i] * 5;
|
|
} else if (type == STACK) {
|
|
val = vals[MAP_SIZE - 1 - i];
|
|
pkt_v4.iph.saddr = vals[MAP_SIZE - 1 - i] * 5;
|
|
}
|
|
|
|
err = bpf_prog_test_run(prog_fd, 1, &pkt_v4, sizeof(pkt_v4),
|
|
buf, &size, &retval, &duration);
|
|
if (err || retval || size != sizeof(pkt_v4) ||
|
|
iph->daddr != val)
|
|
break;
|
|
}
|
|
|
|
CHECK(err || retval || size != sizeof(pkt_v4) || iph->daddr != val,
|
|
"bpf_map_pop_elem",
|
|
"err %d errno %d retval %d size %d iph->daddr %u\n",
|
|
err, errno, retval, size, iph->daddr);
|
|
|
|
/* Queue is empty, program should return TC_ACT_SHOT */
|
|
err = bpf_prog_test_run(prog_fd, 1, &pkt_v4, sizeof(pkt_v4),
|
|
buf, &size, &retval, &duration);
|
|
CHECK(err || retval != 2 /* TC_ACT_SHOT */|| size != sizeof(pkt_v4),
|
|
"check-queue-stack-map-empty",
|
|
"err %d errno %d retval %d size %d\n",
|
|
err, errno, retval, size);
|
|
|
|
/* Check that the program pushed elements correctly */
|
|
for (i = 0; i < MAP_SIZE; i++) {
|
|
err = bpf_map_lookup_and_delete_elem(map_out_fd, NULL, &val);
|
|
if (err || val != vals[i] * 5)
|
|
break;
|
|
}
|
|
|
|
CHECK(i != MAP_SIZE && (err || val != vals[i] * 5),
|
|
"bpf_map_push_elem", "err %d value %u\n", err, val);
|
|
|
|
out:
|
|
pkt_v4.iph.saddr = 0;
|
|
bpf_object__close(obj);
|
|
}
|
|
|
|
int main(void)
|
|
{
|
|
srand(time(NULL));
|
|
|
|
jit_enabled = is_jit_enabled();
|
|
|
|
test_pkt_access();
|
|
test_xdp();
|
|
test_xdp_adjust_tail();
|
|
test_l4lb_all();
|
|
test_xdp_noinline();
|
|
test_tcp_estats();
|
|
test_bpf_obj_id();
|
|
test_pkt_md_access();
|
|
test_obj_name();
|
|
test_tp_attach_query();
|
|
test_stacktrace_map();
|
|
test_stacktrace_build_id();
|
|
test_stacktrace_build_id_nmi();
|
|
test_stacktrace_map_raw_tp();
|
|
test_get_stack_raw_tp();
|
|
test_task_fd_query_rawtp();
|
|
test_task_fd_query_tp();
|
|
test_reference_tracking();
|
|
test_queue_stack_map(QUEUE);
|
|
test_queue_stack_map(STACK);
|
|
|
|
printf("Summary: %d PASSED, %d FAILED\n", pass_cnt, error_cnt);
|
|
return error_cnt ? EXIT_FAILURE : EXIT_SUCCESS;
|
|
}
|