root/tools/testing/selftests/bpf/prog_tests/map_btf.c
// SPDX-License-Identifier: GPL-2.0
/* Copyright (C) 2023. Huawei Technologies Co., Ltd */
#include <test_progs.h>

#include "normal_map_btf.skel.h"
#include "map_in_map_btf.skel.h"

static void do_test_normal_map_btf(void)
{
        struct normal_map_btf *skel;
        int i, err, new_fd = -1;
        int map_fd_arr[64];

        skel = normal_map_btf__open_and_load();
        if (!ASSERT_OK_PTR(skel, "open_load"))
                return;

        err = normal_map_btf__attach(skel);
        if (!ASSERT_OK(err, "attach"))
                goto out;

        skel->bss->pid = getpid();
        usleep(1);
        ASSERT_TRUE(skel->bss->done, "done");

        /* Use percpu_array to slow bpf_map_free_deferred() down.
         * The memory allocation may fail, so doesn't check the returned fd.
         */
        for (i = 0; i < ARRAY_SIZE(map_fd_arr); i++)
                map_fd_arr[i] = bpf_map_create(BPF_MAP_TYPE_PERCPU_ARRAY, NULL, 4, 4, 256, NULL);

        /* Close array fd later */
        new_fd = dup(bpf_map__fd(skel->maps.array));
out:
        normal_map_btf__destroy(skel);
        if (new_fd < 0)
                return;
        /* Use kern_sync_rcu() to wait for the start of the free of the bpf
         * program and use an assumed delay to wait for the release of the map
         * btf which is held by other maps (e.g, bss). After that, array map
         * holds the last reference of map btf.
         */
        kern_sync_rcu();
        usleep(4000);
        /* Spawn multiple kworkers to delay the invocation of
         * bpf_map_free_deferred() for array map.
         */
        for (i = 0; i < ARRAY_SIZE(map_fd_arr); i++) {
                if (map_fd_arr[i] < 0)
                        continue;
                close(map_fd_arr[i]);
        }
        close(new_fd);
}

static void do_test_map_in_map_btf(void)
{
        int err, zero = 0, new_fd = -1;
        struct map_in_map_btf *skel;

        skel = map_in_map_btf__open_and_load();
        if (!ASSERT_OK_PTR(skel, "open_load"))
                return;

        err = map_in_map_btf__attach(skel);
        if (!ASSERT_OK(err, "attach"))
                goto out;

        skel->bss->pid = getpid();
        usleep(1);
        ASSERT_TRUE(skel->bss->done, "done");

        /* Close inner_array fd later */
        new_fd = dup(bpf_map__fd(skel->maps.inner_array));
        /* Defer the free of inner_array */
        err = bpf_map__delete_elem(skel->maps.outer_array, &zero, sizeof(zero), 0);
        ASSERT_OK(err, "delete inner map");
out:
        map_in_map_btf__destroy(skel);
        if (new_fd < 0)
                return;
        /* Use kern_sync_rcu() to wait for the start of the free of the bpf
         * program and use an assumed delay to wait for the free of the outer
         * map and the release of map btf. After that, inner map holds the last
         * reference of map btf.
         */
        kern_sync_rcu();
        usleep(10000);
        close(new_fd);
}

void test_map_btf(void)
{
        if (test__start_subtest("array_btf"))
                do_test_normal_map_btf();
        if (test__start_subtest("inner_array_btf"))
                do_test_map_in_map_btf();
}