static_branch_unlikely
return static_branch_unlikely(&arm64_mismatched_32bit_el0) ||
(static_branch_unlikely(&kvm_protected_mode_initialized) ? \
!static_branch_unlikely(&arm64_mismatched_32bit_el0))
if (!static_branch_unlikely(&arm64_mismatched_32bit_el0))
return static_branch_unlikely(&mte_async_or_asymm_mode);
return static_branch_unlikely(&rsi_present);
return static_branch_unlikely(&arm_si_l1_workaround_4311569);
if (static_branch_unlikely(&arm64_mismatched_32bit_el0))
return !static_branch_unlikely(&arm64_mismatched_32bit_el0) ||
if (static_branch_unlikely(&arm64_mismatched_32bit_el0))
} else if (static_branch_unlikely(&arm64_mismatched_32bit_el0)) {
!static_branch_unlikely(&has_gic_active_state))
if (static_branch_unlikely(&vgic_v3_cpuif_trap) &&
if (static_branch_unlikely(&vgic_v3_cpuif_trap) &&
if (static_branch_unlikely(&vgic_v2_cpuif_trap)) {
if (static_branch_unlikely(&kvm_protected_mode_initialized))
if (static_branch_unlikely(&kvm_vgic_global_state.gicv3_cpuif)) {
if (static_branch_unlikely(&kvm_vgic_global_state.gicv3_cpuif)) {
if (static_branch_unlikely(&vgic_v3_has_v2_compat)) {
if (static_branch_unlikely(&vgic_v3_cpuif_trap) ||
if (static_branch_unlikely(&vgic_v3_has_v2_compat)) {
if (static_branch_unlikely(&vgic_v3_cpuif_trap) ||
if (!static_branch_unlikely(&vgic_v2_cpuif_trap)) {
return !static_branch_unlikely(&kvm_vgic_global_state.gicv3_cpuif) || has_vhe();
if (!static_branch_unlikely(&kvm_vgic_global_state.gicv3_cpuif))
if (!static_branch_unlikely(&kvm_vgic_global_state.gicv3_cpuif))
if (has_vhe() && static_branch_unlikely(&kvm_vgic_global_state.gicv3_cpuif))
if (!static_branch_unlikely(&kvm_vgic_global_state.gicv3_cpuif))
if (has_vhe() && static_branch_unlikely(&kvm_vgic_global_state.gicv3_cpuif))
if (!static_branch_unlikely(&kvm_vgic_global_state.gicv3_cpuif))
if (!static_branch_unlikely(&kvm_vgic_global_state.gicv3_cpuif)) {
if (static_branch_unlikely(&arm64_ptdump_lock_key) && acquire_mmap_lock) {
if (static_branch_unlikely(&arm64_ptdump_lock_key)) {
if (!static_branch_unlikely(&virt_spin_lock_key))
if (!static_branch_unlikely(&virt_preempt_key))
if (static_branch_unlikely(&ftrace_graph_enable)) {
if (static_branch_unlikely(&uaccess_flush_key))
if (static_branch_unlikely(&uaccess_flush_key))
if (static_branch_unlikely(&uaccess_flush_key) && flags == AMR_KUAP_BLOCKED)
return static_branch_unlikely(&__kvmhv_is_nestedv2);
return static_branch_unlikely(&kvm_guest);
return static_branch_unlikely(&shared_processor);
(static_branch_unlikely(&sk_dynamic_irqentry_exit_cond_resched))
return static_branch_unlikely(&interrupt_exit_not_reentrant);
if (static_branch_unlikely(&splpar_asym_pack))
if (static_branch_unlikely(&splpar_asym_pack))
if (static_branch_unlikely(&splpar_asym_pack))
return static_branch_unlikely(&stress_slb_key);
return static_branch_unlikely(&stress_hpt_key);
return static_branch_unlikely(&no_slb_preload_key);
#define DO_TRACE (static_branch_unlikely(&opal_tracepoint_key))
static_branch_unlikely(&kvm_riscv_aia_available)
static_branch_unlikely(&kvm_riscv_nacl_available)
static_branch_unlikely(&kvm_riscv_nacl_sync_csr_available)
static_branch_unlikely(&kvm_riscv_nacl_sync_hfence_available)
static_branch_unlikely(&kvm_riscv_nacl_sync_sret_available)
static_branch_unlikely(&kvm_riscv_nacl_autoswap_csr_available)
static_branch_unlikely(&riscv_sbi_for_rfence)
if (static_branch_unlikely(&qspinlock_key)) \
if (static_branch_unlikely(&kvm_riscv_vsstage_tlb_no_gpa))
if (static_branch_unlikely(&use_asid_allocator)) {
if (!static_branch_unlikely(&pai_key))
if (!static_branch_unlikely(&pai_key))
if (!static_branch_unlikely(&have_mio))
if (!static_branch_unlikely(&have_mio))
if (!static_branch_unlikely(&have_mio))
if (!static_branch_unlikely(&have_mio))
if (static_branch_unlikely(&lbr_from_quirk_key)) {
if (static_branch_unlikely(&lbr_from_quirk_key)) {
#define is_hybrid() static_branch_unlikely(&perf_is_hybrid)
return static_branch_unlikely(&isolation_type_snp);
return static_branch_unlikely(&isolation_type_tdx);
return static_branch_unlikely(&__fpu_state_size_dynamic);
if (static_branch_unlikely(&kvm_async_pf_enabled))
return static_branch_unlikely(µcode_nmi_handler_enable);
if (static_branch_unlikely(&arch_scale_freq_key)) {
if (static_branch_unlikely(&arch_hybrid_cap_scale_key)) {
if (static_branch_unlikely(&arch_hybrid_cap_scale_key))
if (static_branch_unlikely(&arch_hybrid_cap_scale_key))
if (!static_branch_unlikely(&switch_mm_cond_l1d_flush))
if (!static_branch_unlikely(&switch_mm_cond_l1d_flush))
static_branch_unlikely(&switch_to_cond_stibp)) {
WARN_ON(static_branch_unlikely(&apic_hw_disabled.key));
WARN_ON(static_branch_unlikely(&apic_sw_disabled.key));
if (static_branch_unlikely(&kvm_has_noapic_vcpu))
if (static_branch_unlikely(&apic_hw_disabled.key))
if (static_branch_unlikely(&apic_sw_disabled.key))
if (!static_branch_unlikely(&vmx_l1d_should_flush))
return static_branch_unlikely(&__kvm_is_using_evmcs);
WARN_ON(static_branch_unlikely(&kvm_xen_enabled.key));
WARN_ON_ONCE(static_branch_unlikely(&kvm_has_noapic_vcpu));
if (static_branch_unlikely(&kvm_xen_enabled.key) &&
return static_branch_unlikely(&kvm_xen_enabled.key) &&
return static_branch_unlikely(&kvm_xen_enabled.key) &&
if (!static_branch_unlikely(&kvm_xen_enabled.key))
return static_branch_unlikely(&kvm_xen_enabled.key) &&
if (static_branch_unlikely(&kvm_xen_enabled.key) &&
return static_branch_unlikely(&kvm_xen_enabled.key) &&
#define copy_mc_fragile_enabled (static_branch_unlikely(©_mc_fragile_key))
if (static_branch_unlikely(&switch_mm_always_ibpb)) {
if (static_branch_unlikely(&switch_mm_cond_l1d_flush)) {
if (static_branch_unlikely(&rdpmc_always_available_key) ||
(!static_branch_unlikely(&rdpmc_never_available_key) &&
if (static_branch_unlikely(&(scmi_quirk_ ## _qn))) \
if (static_branch_unlikely(&i2c_trace_msg_key)) {
if (static_branch_unlikely(&i2c_trace_msg_key)) {
if (static_branch_unlikely(&iommu_deferred_attach_enabled)) {
if (static_branch_unlikely(&iommu_deferred_attach_enabled) &&
if (static_branch_unlikely(&iommu_deferred_attach_enabled) &&
if (static_branch_unlikely(&iommu_deferred_attach_enabled) &&
if (static_branch_unlikely(&iommu_debug_initialized))
if (static_branch_unlikely(&iommu_debug_initialized))
if (static_branch_unlikely(&iommu_debug_initialized))
if (!static_branch_unlikely(&gic_nvidia_t241_erratum)) {
if (static_branch_unlikely(&gic_nvidia_t241_erratum)) {
if (!static_branch_unlikely(&gic_arm64_2941627_erratum))
if (static_branch_unlikely(&frankengic_key))
if (static_branch_unlikely(&needs_rmw_access))
if (static_branch_unlikely(&legacy_bindings)) {
!static_branch_unlikely(&legacy_bindings)))
if (static_branch_unlikely(&legacy_bindings))
unsigned int param_count = static_branch_unlikely(&legacy_bindings) ? 3 : 2;
if (static_branch_unlikely(&legacy_bindings)) {
if (static_branch_unlikely(&xintc_is_be))
if (static_branch_unlikely(&xintc_is_be))
if (static_branch_unlikely(&no_sleep_enabled) && c->no_sleep)
if (static_branch_unlikely(&no_sleep_enabled) && c->no_sleep)
if (static_branch_unlikely(&no_sleep_enabled) && c->no_sleep &&
if (static_branch_unlikely(&no_sleep_enabled) && c->no_sleep)
if (static_branch_unlikely(&no_sleep_enabled) && b->c->no_sleep) {
if (static_branch_unlikely(&no_sleep_enabled) && bc->no_sleep)
if (static_branch_unlikely(&no_sleep_enabled) && bc->no_sleep)
if (static_branch_unlikely(&no_sleep_enabled) && bc->no_sleep)
if (static_branch_unlikely(&no_sleep_enabled) && bc->no_sleep)
if (static_branch_unlikely(&no_sleep_enabled) && lh->cache->no_sleep)
if (static_branch_unlikely(&no_sleep_enabled) && lh->cache->no_sleep)
if (static_branch_unlikely(&no_sleep_enabled) && lh->cache->no_sleep)
if (static_branch_unlikely(&no_sleep_enabled) && lh->cache->no_sleep)
if (static_branch_unlikely(&use_bh_wq_enabled) && io->in_bh) {
else if (static_branch_unlikely(&use_bh_wq_enabled) && io->in_bh) {
if (static_branch_unlikely(&use_bh_wq_enabled) && io->in_bh) {
if (static_branch_unlikely(&use_bh_wq_enabled) && io->in_bh) {
if (static_branch_unlikely(&use_bh_wq_enabled) && io->v->use_bh_wq &&
if (static_branch_unlikely(&zoned_enabled) &&
if (static_branch_unlikely(&zoned_enabled)) {
if (static_branch_unlikely(&swap_bios_enabled) &&
if (static_branch_unlikely(&zoned_enabled) &&
if (static_branch_unlikely(&swap_bios_enabled) &&
if (static_branch_unlikely(&swap_bios_enabled) &&
if (static_branch_unlikely(&zoned_enabled) &&
if (static_branch_unlikely(&zoned_enabled)) {
if (static_branch_unlikely(&zoned_enabled) && dm_zone_plug_bio(md, bio))
if (static_branch_unlikely(&zoned_enabled) &&
if (static_branch_unlikely(&stats_enabled) &&
if (static_branch_unlikely(&stats_enabled) &&
static_branch_unlikely(&balloon_stat_enabled);
return static_branch_unlikely(&brcmnand_soc_has_ops_key);
if (!static_branch_unlikely(&bond_bcast_neigh_enabled))
if (static_branch_unlikely(&aq_xdp_locking_key))
if (static_branch_unlikely(&bnxt_xdp_locking_key))
if (static_branch_unlikely(&bnxt_xdp_locking_key))
if (static_branch_unlikely(&enetc_has_err050089))
if (static_branch_unlikely(&enetc_has_err050089))
if (static_branch_unlikely(&enetc_has_err050089))
if (static_branch_unlikely(&enetc_has_err050089))
if (static_branch_unlikely(&enetc_has_err050089)) {
if (static_branch_unlikely(&enetc_has_err050089)) {
if (static_branch_unlikely(&ice_xdp_locking_key))
if (static_branch_unlikely(&ice_xdp_locking_key))
if (static_branch_unlikely(&ice_xdp_locking_key)) {
if (static_branch_unlikely(&ice_xdp_locking_key))
if (static_branch_unlikely(&ice_xdp_locking_key))
if (static_branch_unlikely(&ice_xdp_locking_key))
return static_branch_unlikely(&igc_fpe_enabled) &&
if (static_branch_unlikely(&ixgbe_xdp_locking_key))
if (static_branch_unlikely(&ixgbe_xdp_locking_key))
if (static_branch_unlikely(&ixgbe_xdp_locking_key))
if (static_branch_unlikely(&ixgbe_xdp_locking_key))
if (static_branch_unlikely(&ixgbe_xdp_locking_key))
if (static_branch_unlikely(&ixgbe_xdp_locking_key))
if (static_branch_unlikely(&ixgbe_xdp_locking_key))
if (static_branch_unlikely(&ixgbe_xdp_locking_key))
if (static_branch_unlikely(&cn10k_ipsec_sa_enabled) &&
if (static_branch_unlikely(&cn10k_ipsec_sa_enabled) &&
if (static_branch_unlikely(&cn10k_ipsec_sa_enabled) &&
if (static_branch_unlikely(&ocelot_fdma_enabled))
if (tun->numqueues == 1 && static_branch_unlikely(&rps_needed)) {
static_branch_unlikely(&sbi_pmu_snapshot_available)
if (static_branch_unlikely(&cio_inject_enabled)) {
static_branch_unlikely(&xfs_defer_drain_waiter_gate) &&
# define xfs_hooks_switched_on(name) static_branch_unlikely(name)
if (!static_branch_unlikely(&kunit_running))
if (static_branch_unlikely(&kunit_running)) { \
return static_branch_unlikely(&broken_cntvoff_key);
#define cgroup_bpf_enabled(atype) static_branch_unlikely(&cgroup_bpf_enabled_key[atype])
return static_branch_unlikely(&context_tracking_key);
if (!static_branch_unlikely(&cpusets_pre_enable_key))
if (!static_branch_unlikely(&cpusets_enabled_key))
return static_branch_unlikely(&cpusets_enabled_key);
return static_branch_unlikely(&cpusets_insane_config_key);
if (!static_branch_unlikely(&delayacct_key))
if (!static_branch_unlikely(&delayacct_key))
if (!static_branch_unlikely(&delayacct_key))
if (!static_branch_unlikely(&delayacct_key))
if (!static_branch_unlikely(&delayacct_key))
if (!static_branch_unlikely(&delayacct_key))
if (!static_branch_unlikely(&delayacct_key))
if (!static_branch_unlikely(&delayacct_key))
if (!static_branch_unlikely(&delayacct_key))
if (!static_branch_unlikely(&delayacct_key))
if (!static_branch_unlikely(&delayacct_key))
if (!static_branch_unlikely(&delayacct_key))
if (!static_branch_unlikely(&delayacct_key))
static_branch_unlikely(&descriptor.key.dd_key_false)
if (static_branch_unlikely(&bpf_stats_enabled_key)) {
if (static_branch_unlikely(&freezer_active))
return static_branch_unlikely(&kill_ftrace_graph);
# define force_irqthreads() (static_branch_unlikely(&force_irqthreads_key))
if (static_branch_unlikely(&iommu_debug_initialized))
: static_branch_unlikely(x))
if (!static_branch_unlikely(&kfence_allocation_key))
if (static_branch_unlikely(&klp_sched_try_switch_key) &&
#define mem_cgroup_sockets_enabled static_branch_unlikely(&memcg_sockets_enabled_key)
return static_branch_unlikely(&_page_poisoning_enabled);
return static_branch_unlikely(&_debug_pagealloc_enabled);
return static_branch_unlikely(&_debug_guardpage_enabled);
return static_branch_unlikely(&lru_gen_caps[LRU_GEN_CORE]);
if (static_branch_unlikely(&___once_key)) { \
if (static_branch_unlikely(&___once_key)) { \
if (!static_branch_unlikely(&hugetlb_optimize_vmemmap_key))
if (!static_branch_unlikely(&hugetlb_optimize_vmemmap_key))
if (static_branch_unlikely(&page_owner_inited))
if (static_branch_unlikely(&page_owner_inited))
if (static_branch_unlikely(&page_owner_inited))
if (static_branch_unlikely(&page_owner_inited))
if (static_branch_unlikely(&page_owner_inited))
if (static_branch_unlikely(&page_owner_inited))
if (static_branch_unlikely(&perf_sched_events))
if (static_branch_unlikely(&perf_sched_events))
if (static_branch_unlikely(&rseq_debug_enabled))
if (static_branch_unlikely(&rseq_debug_enabled)) {
if (!static_branch_unlikely(&rseq_debug_enabled)) {
if (static_branch_unlikely(&rseq_debug_enabled) || unlikely(csaddr)) {
if (static_branch_unlikely(&rseq_debug_enabled))
if (static_branch_unlikely(&rseq_debug_enabled))
if (static_branch_unlikely(&housekeeping_overridden))
if (static_branch_unlikely(&netstamp_needed_key))
if (static_branch_unlikely(&netstamp_needed_key) || cond)
static_branch_unlikely(&(__tracepoint_##tp).key)
return static_branch_unlikely(&__tracepoint_##name.key);\
if (static_branch_unlikely(&__tracepoint_##name.key)) \
if (static_branch_unlikely(&__tracepoint_##name.key)) \
if (static_branch_unlikely(&udp_encap_needed_key))
if (static_branch_unlikely(&udpv6_encap_needed_key))
return static_branch_unlikely(&ip_tunnel_metadata_cnt);
if (static_branch_unlikely(&ipv6_flowlabel_exclusive.key) &&
if (static_branch_unlikely(&libeth_xdpsq_share) &&
if (static_branch_unlikely(&libeth_xdpsq_share) && lock->share)
if (static_branch_unlikely(&libeth_xdpsq_share) && lock->share)
if (static_branch_unlikely(&libeth_xdpsq_share) && lock->share)
#define tc_skb_ext_tc_enabled() static_branch_unlikely(&tc_skb_ext_tc)
return static_branch_unlikely(&rfs_needed);
return static_branch_unlikely(&memalloc_socks_key);
if (!static_branch_unlikely(&tcp_md5_needed.key))
if (!static_branch_unlikely(&tcp_md5_needed.key))
if (static_branch_unlikely(&tcp_ao_needed.key)) {
if (static_branch_unlikely(&tcp_md5_needed.key) &&
if (static_branch_unlikely(&tcp_tx_delay_enabled))
if (static_branch_unlikely(&tcp_tx_delay_enabled)) {
if (!static_branch_unlikely(&tcp_ao_needed.key))
#define static_branch_tcp_md5() static_branch_unlikely(&tcp_md5_needed.key)
#define static_branch_tcp_ao() static_branch_unlikely(&tcp_ao_needed.key)
if (static_branch_unlikely(&bpf_master_redirect_enabled_key)) {
if (static_branch_unlikely(&io_key_has_sqarray.key) &&
if (static_branch_unlikely(&bpf_stats_enabled_key)) {
if (static_branch_unlikely(&bpf_stats_enabled_key))
if (!static_branch_unlikely(&sk_dynamic_irqentry_exit_cond_resched))
if (static_branch_unlikely(&irqhandler_duration_check_enabled)) {
WARN_ON(!static_branch_unlikely(&sk_true));
WARN_ON(static_branch_unlikely(&sk_false));
WARN_ON(static_branch_unlikely(&sk_true));
WARN_ON(!static_branch_unlikely(&sk_false));
int state = !static_branch_unlikely(&stack_erasing_bypass);
#define skip_erasing() static_branch_unlikely(&stack_erasing_bypass)
bool on = static_branch_unlikely(&rseq_debug_enabled);
if (static_branch_unlikely(&rseq_debug_enabled))
return static_branch_unlikely(&sk_dynamic_preempt_lazy);
if (!static_branch_unlikely(&preempt_notifier_key))
if (static_branch_unlikely(&preempt_notifier_key))
if (static_branch_unlikely(&preempt_notifier_key))
if (!static_branch_unlikely(&sk_dynamic_preempt_schedule))
if (!static_branch_unlikely(&sk_dynamic_preempt_schedule_notrace))
if (!static_branch_unlikely(&sk_dynamic_cond_resched))
if (!static_branch_unlikely(&sk_dynamic_might_resched))
if (static_branch_unlikely(&scx_builtin_idle_per_node)) {
if (static_branch_unlikely(&scx_builtin_idle_per_node)) {
if (static_branch_unlikely(&sched_numa_balancing))
if (static_branch_unlikely(&sched_cluster_active)) {
if (!static_branch_unlikely(&sched_cluster_active) ||
if (!static_branch_unlikely(&sched_cluster_active) ||
if (static_branch_unlikely(&arch_needs_tick_broadcast))
if (static_branch_unlikely(&arch_needs_tick_broadcast))
if (static_branch_unlikely(&housekeeping_overridden))
if (static_branch_unlikely(&housekeeping_overridden) &&
if (static_branch_unlikely(&housekeeping_overridden)) {
if (static_branch_unlikely(&housekeeping_overridden)) {
return static_branch_unlikely(&__sched_core_enabled) && rq->core_enabled;
return !static_branch_unlikely(&__sched_core_enabled);
return static_branch_unlikely(&rt_group_sched);
if (static_branch_unlikely(&sched_smt_present))
#define scx_enabled() static_branch_unlikely(&__scx_enabled)
#define scx_switched_all() static_branch_unlikely(&__scx_switched_all)
return static_branch_unlikely(&sched_asym_cpucapacity);
return static_branch_unlikely(&sched_energy_present);
#define schedstat_enabled() static_branch_unlikely(&sched_schedstats)
if (static_branch_unlikely(&sched_cluster_active))
state = static_branch_unlikely(&sched_energy_present);
if (!has_eas && static_branch_unlikely(&sched_energy_present)) {
} else if (has_eas && !static_branch_unlikely(&sched_energy_present)) {
if (static_branch_unlikely(&csdlock_debug_enabled))
if (static_branch_unlikely(&csdlock_debug_enabled)) {
return static_branch_unlikely(&timers_nohz_active);
if (!static_branch_unlikely(&tmigr_exclude_isolated))
if (static_branch_unlikely(&trace_event_exports_enabled))
if (static_branch_unlikely(&trace_function_exports_enabled))
if (static_branch_unlikely(&trace_marker_exports_enabled))
test_key_func(base_true_key, static_branch_unlikely)
test_key_func(base_inv_true_key, static_branch_unlikely)
test_key_func(base_false_key, static_branch_unlikely)
test_key_func(base_inv_false_key, static_branch_unlikely)
test_key_func(true_key, static_branch_unlikely)
test_key_func(false_key, static_branch_unlikely)
return static_branch_unlikely(&kasan_flag_stacktrace);
return static_branch_unlikely(&check_pages_enabled);
return static_branch_unlikely(&deferred_pages);
if (static_branch_unlikely(&page_owner_inited))
if (!static_branch_unlikely(&page_owner_inited))
if (!static_branch_unlikely(&page_owner_inited)) {
return static_branch_unlikely(&page_reporting_enabled) &&
if (!static_branch_unlikely(&page_reporting_enabled))
if (!static_branch_unlikely(&page_alloc_shuffle_key))
if (!static_branch_unlikely(&page_alloc_shuffle_key))
if (!static_branch_unlikely(&page_alloc_shuffle_key))
return static_branch_unlikely(&slub_debug_enabled);
if ((static_branch_unlikely(&init_on_alloc) ||
static_branch_unlikely(&init_on_free)) &&
if (static_branch_unlikely(&strict_numa) &&
if (static_branch_unlikely(&strict_numa) &&
#define get_cap(cap) static_branch_unlikely(&lru_gen_caps[cap])
if (!static_branch_unlikely(&page_owner_inited))
return static_branch_unlikely(&br_mst_used) &&
if (!static_branch_unlikely(&br_switchdev_tx_fwd_offload))
if (!static_branch_unlikely(&br_switchdev_tx_fwd_offload))
return static_branch_unlikely(&use_backlog_threads_key);
if (static_branch_unlikely(&netstamp_needed_key))
if (static_branch_unlikely(&netstamp_needed_key)) { \
if (static_branch_unlikely(&tcx_needed_key)) {
if (static_branch_unlikely(&tcx_needed_key)) {
if (static_branch_unlikely(&egress_needed_key)) {
if (static_branch_unlikely(&rps_needed)) {
if (static_branch_unlikely(&generic_xdp_needed_key)) {
if (static_branch_unlikely(&ingress_needed_key)) {
if (static_branch_unlikely(&rps_needed)) {
if (static_branch_unlikely(&rps_needed)) {
if (static_branch_unlikely(&page_pool_mem_providers) && pool->mp_ops)
if (static_branch_unlikely(&page_pool_mem_providers) && pool->mp_ops)
!static_branch_unlikely(&net_high_order_alloc_disable_key)) {
if (static_branch_unlikely(&bpf_sk_lookup_enabled)) {
if (val && !static_branch_unlikely(&tcp_tx_delay_enabled)) {
if (!static_branch_unlikely(&tcp_md5_needed.key))
if (!static_branch_unlikely(&tcp_ao_needed.key))
if (!static_branch_unlikely(&tcp_ao_needed.key))
if (!static_branch_unlikely(&tcp_ao_needed.key))
if (!static_branch_unlikely(&tcp_ao_needed.key))
if (static_branch_unlikely(&clean_acked_data_enabled.key))
if (static_branch_unlikely(&tcp_have_smc)) {
if (static_branch_unlikely(&tcp_have_smc)) {
if (static_branch_unlikely(&tcp_ao_needed.key)) {
if (static_branch_unlikely(&tcp_ao_needed.key) &&
if (static_branch_unlikely(&ip4_min_ttl)) {
if (static_branch_unlikely(&ip4_min_ttl)) {
if (!static_branch_unlikely(&tcp_md5_needed.key))
if (static_branch_unlikely(&tcp_md5_needed.key)) {
if (static_branch_unlikely(&tcp_have_smc)) {
if (static_branch_unlikely(&tcp_tx_delay_enabled) &&
if (static_branch_unlikely(&tcp_have_smc)) {
if (static_branch_unlikely(&tcp_have_smc) && tp->syn_smc) {
if (static_branch_unlikely(&tcp_have_smc) && tp->syn_smc && ireq->smc_ok) {
if (static_branch_unlikely(&udp_encap_needed_key) &&
if (static_branch_unlikely(&udp_encap_needed_key)) {
if (static_branch_unlikely(&bpf_sk_lookup_enabled) &&
if (static_branch_unlikely(&udp_encap_needed_key)) {
if (static_branch_unlikely(&udp_encap_needed_key))
if (static_branch_unlikely(&bpf_sk_lookup_enabled)) {
if (static_branch_unlikely(&nf_hooks_lwtunnel_enabled))
if (static_branch_unlikely(&nf_hooks_lwtunnel_enabled))
if (static_branch_unlikely(&nf_hooks_lwtunnel_enabled))
if (static_branch_unlikely(&nf_hooks_lwtunnel_enabled))
if (static_branch_unlikely(&nf_hooks_lwtunnel_enabled))
if (static_branch_unlikely(&nf_hooks_lwtunnel_enabled))
if (static_branch_unlikely(&nf_hooks_lwtunnel_enabled))
if (static_branch_unlikely(&tcp_ao_needed.key)) {
} else if (static_branch_unlikely(&tcp_md5_needed.key)) {
if (static_branch_unlikely(&tcp_ao_needed.key) &&
} else if (static_branch_unlikely(&tcp_md5_needed.key)) {
if (static_branch_unlikely(&ip6_min_hopcount)) {
if (static_branch_unlikely(&ip6_min_hopcount)) {
if (static_branch_unlikely(&udpv6_encap_needed_key)) {
if (static_branch_unlikely(&bpf_sk_lookup_enabled) &&
if (static_branch_unlikely(&udpv6_encap_needed_key)) {
if (static_branch_unlikely(&udpv6_encap_needed_key) &&
if (static_branch_unlikely(&udpv6_encap_needed_key))
if (static_branch_unlikely(&aql_disable))
if (static_branch_unlikely(&nf_hooks_lwtunnel_enabled))
if (static_branch_unlikely(&nf_hooks_lwtunnel_enabled)) {
if (static_branch_unlikely(&nft_trace_enabled))
if (static_branch_unlikely(&nft_trace_enabled))
if (static_branch_unlikely(&nft_counters_enabled))
if (static_branch_unlikely(&nft_trace_enabled)) {
if (static_branch_unlikely(&nft_trace_enabled))
if (static_branch_unlikely(&ovs_ct_limit_enabled)) {
if (static_branch_unlikely(&tcf_frag_xmit_count))
if (static_branch_unlikely(&taprio_have_broken_mqprio) &&
!static_branch_unlikely(&taprio_have_broken_mqprio)) {
if (static_branch_unlikely(&SECURITY_HOOK_ACTIVE_KEY(HOOK, NUM))) { \
if (static_branch_unlikely(&SECURITY_HOOK_ACTIVE_KEY(HOOK, NUM))) { \