162 lines
4.4 KiB
C
162 lines
4.4 KiB
C
|
// SPDX-License-Identifier: GPL-2.0
|
||
|
#include <test_progs.h>
|
||
|
#include <network_helpers.h>
|
||
|
#include <net/if.h>
|
||
|
#include "test_xdp.skel.h"
|
||
|
#include "test_xdp_bpf2bpf.skel.h"
|
||
|
|
||
|
struct meta {
|
||
|
int ifindex;
|
||
|
int pkt_len;
|
||
|
};
|
||
|
|
||
|
struct test_ctx_s {
|
||
|
bool passed;
|
||
|
int pkt_size;
|
||
|
};
|
||
|
|
||
|
struct test_ctx_s test_ctx;
|
||
|
|
||
|
static void on_sample(void *ctx, int cpu, void *data, __u32 size)
|
||
|
{
|
||
|
struct meta *meta = (struct meta *)data;
|
||
|
struct ipv4_packet *trace_pkt_v4 = data + sizeof(*meta);
|
||
|
unsigned char *raw_pkt = data + sizeof(*meta);
|
||
|
struct test_ctx_s *tst_ctx = ctx;
|
||
|
|
||
|
ASSERT_GE(size, sizeof(pkt_v4) + sizeof(*meta), "check_size");
|
||
|
ASSERT_EQ(meta->ifindex, if_nametoindex("lo"), "check_meta_ifindex");
|
||
|
ASSERT_EQ(meta->pkt_len, tst_ctx->pkt_size, "check_meta_pkt_len");
|
||
|
ASSERT_EQ(memcmp(trace_pkt_v4, &pkt_v4, sizeof(pkt_v4)), 0,
|
||
|
"check_packet_content");
|
||
|
|
||
|
if (meta->pkt_len > sizeof(pkt_v4)) {
|
||
|
for (int i = 0; i < meta->pkt_len - sizeof(pkt_v4); i++)
|
||
|
ASSERT_EQ(raw_pkt[i + sizeof(pkt_v4)], (unsigned char)i,
|
||
|
"check_packet_content");
|
||
|
}
|
||
|
|
||
|
tst_ctx->passed = true;
|
||
|
}
|
||
|
|
||
|
#define BUF_SZ 9000
|
||
|
|
||
|
static void run_xdp_bpf2bpf_pkt_size(int pkt_fd, struct perf_buffer *pb,
|
||
|
struct test_xdp_bpf2bpf *ftrace_skel,
|
||
|
int pkt_size)
|
||
|
{
|
||
|
__u8 *buf, *buf_in;
|
||
|
int err;
|
||
|
LIBBPF_OPTS(bpf_test_run_opts, topts);
|
||
|
|
||
|
if (!ASSERT_LE(pkt_size, BUF_SZ, "pkt_size") ||
|
||
|
!ASSERT_GE(pkt_size, sizeof(pkt_v4), "pkt_size"))
|
||
|
return;
|
||
|
|
||
|
buf_in = malloc(BUF_SZ);
|
||
|
if (!ASSERT_OK_PTR(buf_in, "buf_in malloc()"))
|
||
|
return;
|
||
|
|
||
|
buf = malloc(BUF_SZ);
|
||
|
if (!ASSERT_OK_PTR(buf, "buf malloc()")) {
|
||
|
free(buf_in);
|
||
|
return;
|
||
|
}
|
||
|
|
||
|
test_ctx.passed = false;
|
||
|
test_ctx.pkt_size = pkt_size;
|
||
|
|
||
|
memcpy(buf_in, &pkt_v4, sizeof(pkt_v4));
|
||
|
if (pkt_size > sizeof(pkt_v4)) {
|
||
|
for (int i = 0; i < (pkt_size - sizeof(pkt_v4)); i++)
|
||
|
buf_in[i + sizeof(pkt_v4)] = i;
|
||
|
}
|
||
|
|
||
|
/* Run test program */
|
||
|
topts.data_in = buf_in;
|
||
|
topts.data_size_in = pkt_size;
|
||
|
topts.data_out = buf;
|
||
|
topts.data_size_out = BUF_SZ;
|
||
|
|
||
|
err = bpf_prog_test_run_opts(pkt_fd, &topts);
|
||
|
|
||
|
ASSERT_OK(err, "ipv4");
|
||
|
ASSERT_EQ(topts.retval, XDP_PASS, "ipv4 retval");
|
||
|
ASSERT_EQ(topts.data_size_out, pkt_size, "ipv4 size");
|
||
|
|
||
|
/* Make sure bpf_xdp_output() was triggered and it sent the expected
|
||
|
* data to the perf ring buffer.
|
||
|
*/
|
||
|
err = perf_buffer__poll(pb, 100);
|
||
|
|
||
|
ASSERT_GE(err, 0, "perf_buffer__poll");
|
||
|
ASSERT_TRUE(test_ctx.passed, "test passed");
|
||
|
/* Verify test results */
|
||
|
ASSERT_EQ(ftrace_skel->bss->test_result_fentry, if_nametoindex("lo"),
|
||
|
"fentry result");
|
||
|
ASSERT_EQ(ftrace_skel->bss->test_result_fexit, XDP_PASS, "fexit result");
|
||
|
|
||
|
free(buf);
|
||
|
free(buf_in);
|
||
|
}
|
||
|
|
||
|
void test_xdp_bpf2bpf(void)
|
||
|
{
|
||
|
int err, pkt_fd, map_fd;
|
||
|
int pkt_sizes[] = {sizeof(pkt_v4), 1024, 4100, 8200};
|
||
|
struct iptnl_info value4 = {.family = AF_INET6};
|
||
|
struct test_xdp *pkt_skel = NULL;
|
||
|
struct test_xdp_bpf2bpf *ftrace_skel = NULL;
|
||
|
struct vip key4 = {.protocol = 6, .family = AF_INET};
|
||
|
struct bpf_program *prog;
|
||
|
struct perf_buffer *pb = NULL;
|
||
|
|
||
|
/* Load XDP program to introspect */
|
||
|
pkt_skel = test_xdp__open_and_load();
|
||
|
if (!ASSERT_OK_PTR(pkt_skel, "test_xdp__open_and_load"))
|
||
|
return;
|
||
|
|
||
|
pkt_fd = bpf_program__fd(pkt_skel->progs._xdp_tx_iptunnel);
|
||
|
|
||
|
map_fd = bpf_map__fd(pkt_skel->maps.vip2tnl);
|
||
|
bpf_map_update_elem(map_fd, &key4, &value4, 0);
|
||
|
|
||
|
/* Load trace program */
|
||
|
ftrace_skel = test_xdp_bpf2bpf__open();
|
||
|
if (!ASSERT_OK_PTR(ftrace_skel, "test_xdp_bpf2bpf__open"))
|
||
|
goto out;
|
||
|
|
||
|
/* Demonstrate the bpf_program__set_attach_target() API rather than
|
||
|
* the load with options, i.e. opts.attach_prog_fd.
|
||
|
*/
|
||
|
prog = ftrace_skel->progs.trace_on_entry;
|
||
|
bpf_program__set_expected_attach_type(prog, BPF_TRACE_FENTRY);
|
||
|
bpf_program__set_attach_target(prog, pkt_fd, "_xdp_tx_iptunnel");
|
||
|
|
||
|
prog = ftrace_skel->progs.trace_on_exit;
|
||
|
bpf_program__set_expected_attach_type(prog, BPF_TRACE_FEXIT);
|
||
|
bpf_program__set_attach_target(prog, pkt_fd, "_xdp_tx_iptunnel");
|
||
|
|
||
|
err = test_xdp_bpf2bpf__load(ftrace_skel);
|
||
|
if (!ASSERT_OK(err, "test_xdp_bpf2bpf__load"))
|
||
|
goto out;
|
||
|
|
||
|
err = test_xdp_bpf2bpf__attach(ftrace_skel);
|
||
|
if (!ASSERT_OK(err, "test_xdp_bpf2bpf__attach"))
|
||
|
goto out;
|
||
|
|
||
|
/* Set up perf buffer */
|
||
|
pb = perf_buffer__new(bpf_map__fd(ftrace_skel->maps.perf_buf_map), 8,
|
||
|
on_sample, NULL, &test_ctx, NULL);
|
||
|
if (!ASSERT_OK_PTR(pb, "perf_buf__new"))
|
||
|
goto out;
|
||
|
|
||
|
for (int i = 0; i < ARRAY_SIZE(pkt_sizes); i++)
|
||
|
run_xdp_bpf2bpf_pkt_size(pkt_fd, pb, ftrace_skel,
|
||
|
pkt_sizes[i]);
|
||
|
out:
|
||
|
perf_buffer__free(pb);
|
||
|
test_xdp__destroy(pkt_skel);
|
||
|
test_xdp_bpf2bpf__destroy(ftrace_skel);
|
||
|
}
|