btf__type_cnt
if (root_type_ids[i] >= btf__type_cnt(btf)) {
if (v->type < btf__type_cnt(btf)) {
int cnt = btf__type_cnt(btf);
start_id = btf__type_cnt(base);
__u32 cnt = btf__type_cnt(btf), i, j;
n = btf__type_cnt(btf);
int cnt = btf__type_cnt(btf);
info.attach_btf_id < btf__type_cnt(btf_vmlinux)) {
int n = btf__type_cnt(btf), i;
unsigned int i, n = btf__type_cnt(info->marked_btf);
for (i = 1; i < btf__type_cnt(btf_new); i++) {
start_id = btf__type_cnt(btf__base_btf(btf));
nr_types = btf__type_cnt(btf) - start_id;
nr_types = btf__type_cnt(btf);
const u32 type_cnt = btf__type_cnt(ctx->btf);
btf->start_id = btf__type_cnt(base_btf);
btf->start_id = btf__type_cnt(base_btf);
cnt = btf__type_cnt(src_btf) - 1;
return btf_type_by_id(btf, btf__type_cnt(btf) - 1);
type_cnt = btf__type_cnt(btf);
__u32 nr_types = btf__type_cnt(d->btf);
int id = btf__type_cnt(dist->pipe.dst);
unsigned int n = btf__type_cnt(src_btf);
dist.split_start_id = btf__type_cnt(old_base);
dist.diff_id = dist.split_start_id - btf__type_cnt(new_base);
n = btf__type_cnt(new_split);
btf->start_id = btf__type_cnt(base_btf);
if (new_id >= btf__type_cnt(p->btf))
n = btf__type_cnt(btf);
if (id < btf->start_id || id >= btf__type_cnt(btf)) {
__u32 i, n = btf__type_cnt(btf);
n = btf__type_cnt(btf);
n = btf__type_cnt(btf);
r = btf__type_cnt(btf) - 1;
return btf__type_cnt(btf);
__u32 nr_types = btf__type_cnt(btf);
LIBBPF_API __u32 btf__type_cnt(const struct btf *btf);
int err, last_id = btf__type_cnt(d->btf) - 1;
if (id >= btf__type_cnt(d->btf))
int i, j, n = btf__type_cnt(d->btf);
unsigned int nr_types = btf__type_cnt(btf);
r.nr_dist_base_types = btf__type_cnt(r.dist_base_btf);
r.nr_base_types = btf__type_cnt(base_btf);
nr_types = btf__type_cnt(obj->btf);
for (i = 1; i < btf__type_cnt(btf); i++) {
n = btf__type_cnt(obj->btf);
n = btf__type_cnt(obj->btf);
n = btf__type_cnt(btf);
n = btf__type_cnt(btf);
n = btf__type_cnt(btf);
n = btf__type_cnt(targ_btf);
btf__type_cnt(obj->btf_vmlinux),
n = btf__type_cnt(obj->btf);
for (i = 1, n = btf__type_cnt(btf); i < n; i++) {
if (*type_id >= btf__type_cnt(btf))
n = btf__type_cnt(obj->btf);
n = btf__type_cnt(obj->btf);
n = btf__type_cnt(obj->btf);
start_id = btf__type_cnt(linker->btf);
n = btf__type_cnt(obj->btf);
n = btf__type_cnt(linker->btf);
if (btf__type_cnt(linker->btf) == 1)
ASSERT_EQ(btf__type_cnt(btf) - 1, nr_types, "btf_nr_types");
for (i = 1; i < btf__type_cnt(btf); i++) {
test_nr_types = btf__type_cnt(test_btf);
expect_nr_types = btf__type_cnt(expect_btf);
nr_base_types = btf__type_cnt(vmlinux_btf);
!ASSERT_EQ(8, btf__type_cnt(btf3), "distilled_base_type_cnt"))
!ASSERT_EQ(3, btf__type_cnt(btf3), "distilled_base_type_cnt"))
!ASSERT_EQ(3, btf__type_cnt(btf3), "distilled_base_type_cnt"))
!ASSERT_EQ(3, btf__type_cnt(btf3), "distilled_base_type_cnt"))
!ASSERT_EQ(2, btf__type_cnt(btf3), "distilled_base_type_cnt"))
!ASSERT_EQ(2, btf__type_cnt(new_base), "distilled_base_type_cnt"))
!ASSERT_EQ(2, btf__type_cnt(btf3), "distilled_base_type_cnt"))
for (i = 1; i < btf__type_cnt(t->btf); i++) {
size_t type_cnt = btf__type_cnt(btf);
ASSERT_EQ(btf__type_cnt(swap_btf), btf__type_cnt(btf), "nr_types");
ASSERT_EQ(btf__type_cnt(swap_btf), btf__type_cnt(btf), "nr_types");
for (id = 1; id < btf__type_cnt(btf); id++) {
start_id = btf__type_cnt(base_btf);
ASSERT_EQ(btf__type_cnt(btf3), 6, "split_type_cnt");
for (i = 1; i < btf__type_cnt(btf3); i++) {
if (!ASSERT_EQ(btf__type_cnt(btf3), btf__type_cnt(btf6), "cmp_type_cnt"))
for (i = 1; i < btf__type_cnt(btf6); i++) {
for (i = 1; i < btf__type_cnt(local_btf); i++)
nr = btf__type_cnt(btf);
for (i = 1; i < btf__type_cnt(btf); i++) {
for (i = 1, type_cnt = btf__type_cnt(btf); i < type_cnt; i++) {
cnt = btf__type_cnt(btf);
cnt = btf__type_cnt(btf);