perf_session__env
ret = perf_env__lookup_objdump(perf_session__env(session),
ret = symbol__init(perf_session__env(annotate.session));
if (setup_sorting(/*evlist=*/NULL, perf_session__env(annotate.session)) < 0)
if (symbol__init(session ? perf_session__env(session) : NULL) < 0)
struct perf_env *env = perf_session__env(session);
print_pareto(out, perf_session__env(session));
env = perf_session__env(session);
err = c2c_hists__init(&c2c.hists, "dcacheline", 2, perf_session__env(session));
c2c_hists__reinit(&c2c.hists, output_str, sort_str, perf_session__env(session));
hists__iterate_cb(&c2c.hists.hists, resort_cl_cb, perf_session__env(session));
if (setup_sorting(/*evlist=*/NULL, perf_session__env(data__files[0].session)) < 0)
ret = symbol__init(perf_session__env(inject.session));
symbol__init(perf_session__env(session));
cpuid = perf_session__env(kvm->session)->cpuid;
symbol__init(perf_session__env(kvm->session));
symbol__init(perf_session__env(session));
symbol__init(perf_session__env(session));
symbol__init(perf_session__env(session));
ret = symbol__init(perf_session__env(session));
if (evlist__add_bpf_sb_event(rec->sb_evlist, perf_session__env(rec->session))) {
struct perf_env *env = perf_session__env(session);
env = perf_session__env(session);
perf_session__env(rec->session),
perf_session__env(rec->session),
(setup_sorting(session->evlist, perf_session__env(session)) < 0)) {
if (symbol__init(perf_session__env(session)) < 0)
perf_session__env(rep->session));
perf_session__env(session), true);
symbol__init(perf_session__env(session));
env = perf_session__env(session);
env = perf_session__env(session);
if (tod && !perf_session__env(session)->clock.enabled) {
env = perf_session__env(script->session);
return perf_env__get_socket_aggr_by_cpu(cpu, perf_session__env(perf_stat.session));
return perf_env__get_die_aggr_by_cpu(cpu, perf_session__env(perf_stat.session));
return perf_env__get_cluster_aggr_by_cpu(cpu, perf_session__env(perf_stat.session));
return perf_env__get_cache_aggr_by_cpu(cpu, perf_session__env(perf_stat.session));
return perf_env__get_core_aggr_by_cpu(cpu, perf_session__env(perf_stat.session));
return perf_env__get_cpu_aggr_by_cpu(cpu, perf_session__env(perf_stat.session));
return perf_env__get_node_aggr_by_cpu(cpu, perf_session__env(perf_stat.session));
return perf_env__get_global_aggr_by_cpu(cpu, perf_session__env(perf_stat.session));
struct perf_env *env = perf_session__env(st->session);
struct perf_env *env = perf_session__env(session);
symbol__init(perf_session__env(session));
ret = perf_env__lookup_objdump(perf_session__env(top->session),
ret = perf_env__read_cpu_topology_map(perf_session__env(top->session));
if (setup_sorting(top.evlist, perf_session__env(top.session)) < 0) {
perf_session__env(top->session),
if (symbol__init(perf_session__env(session)) < 0)
env = perf_session__env(session);
struct perf_env *env = perf_session__env(evlist->session);
cpuid = perf_env__cpuid(perf_session__env(spe->session));
env = perf_session__env(session);
struct perf_env *env = perf_session__env(session);
struct perf_env *env = perf_session__env(session);
struct perf_env *env = perf_session__env(session);
const struct perf_env *env = perf_session__env(session);
if (symbol__init(perf_session__env(session)) < 0) {
return session ? perf_session__env(session) : NULL;
sf->machine_type = s390_cpumsf_get_type(perf_session__env(session)->cpuid);
int nr_cpus = min(perf_session__env(session)->nr_cpus_avail, MAX_NR_CPUS);
env = perf_session__env(session);
struct perf_env *perf_session__env(struct perf_session *session);
size_t mmap_len, decomp_len = perf_session__env(session)->comp_mmap_len;