__cold
void __noreturn __cold nvhe_hyp_panic_handler(u64 esr, u64 spsr, u64 elr_virt,
void __noreturn __cold nvhe_hyp_panic_handler(u64 esr, u64 spsr,
union ieee754dp __cold ieee754dp_nanxcpt(union ieee754dp r)
extern union ieee754dp __cold ieee754dp_nanxcpt(union ieee754dp);
union ieee754sp __cold ieee754sp_nanxcpt(union ieee754sp r)
extern union ieee754sp __cold ieee754sp_nanxcpt(union ieee754sp);
int code, unsigned long offset) __noreturn __cold;
void notrace __noreturn __cold toc_intr(struct pt_regs *regs)
void panic(const char *fmt, ...) __noreturn __cold;
void __cold
int __cold unregister_random_vmfork_notifier(struct notifier_block *nb)
int __cold random_online_cpu(unsigned int cpu)
static void __cold crng_set_ready(struct work_struct *work)
void __cold rand_initialize_disk(struct gendisk *disk)
static void __cold entropy_timer(struct timer_list *timer)
static void __cold try_to_generate_entropy(void)
int __cold execute_with_initialized_rng(struct notifier_block *nb)
int __cold random_prepare_cpu(unsigned int cpu)
static void __cold _credit_init_bits(size_t bits)
void __cold add_vmfork_randomness(const void *unique_vm_id, size_t len)
int __cold register_random_vmfork_notifier(struct notifier_block *nb)
static int __cold dpaa2_dpseci_dpio_setup(struct dpaa2_caam_priv *priv)
static void __cold dpaa2_dpseci_dpio_free(struct dpaa2_caam_priv *priv)
static int __cold dpaa2_dpseci_setup(struct fsl_mc_device *ls_dev)
static int __cold dpaa2_dpseci_disable(struct dpaa2_caam_priv *priv)
static void __cold dpaa2_caam_remove(struct fsl_mc_device *ls_dev)
static int __cold dpaa2_qdma_setup(struct fsl_mc_device *ls_dev)
static int __cold dpaa2_qdma_dpio_setup(struct dpaa2_qdma_priv *priv)
static int __cold dpaa2_dpdmai_bind(struct dpaa2_qdma_priv *priv)
static int __cold dpaa2_dpdmai_dpio_unbind(struct dpaa2_qdma_priv *priv)
void __cold libeth_rx_recycle_slow(netmem_ref netmem)
static void __cold libeth_trace_xdp_exception(const struct net_device *dev,
void __cold libeth_xdp_tx_exception(struct libeth_xdp_tx_bulk *bq, u32 sent,
u32 __cold libeth_xdp_xmit_return_bulk(const struct libeth_xdp_tx_frame *bq,
void __cold libeth_xdp_return_buff_slow(struct libeth_xdp_buff *xdp)
u32 __cold libeth_xdp_prog_exception(const struct libeth_xdp_tx_bulk *bq,
static void __cold
static void __cold
u32 __cold libeth_xsk_prog_exception(struct libeth_xdp_buff *xdp,
void __cold libeth_xsk_tx_return_bulk(const struct libeth_xdp_tx_frame *bq,
void __printf(4, 5) __cold qed_hw_err_notify(struct qed_hwfn *p_hwfn,
static __printf(1,2) __cold void vringh_bad(const char *fmt, ...)
static u16 __cold return_from_indirect(const struct vringh *vrh, int *up_next,
static void __cold report_setget_bounds(const struct extent_buffer *eb,
void __cold btrfs_prelim_ref_exit(void)
void __cold btrfs_prelim_ref_exit(void);
void __cold btrfs_bioset_exit(void)
void __cold btrfs_bioset_exit(void);
void __cold btrfs_destroy_cachep(void);
void __cold btrfs_exit_compress(void)
void __cold btrfs_exit_compress(void);
void __cold btrfs_ctree_exit(void)
void __cold btrfs_ctree_exit(void);
void __cold btrfs_auto_defrag_exit(void)
void __cold btrfs_auto_defrag_exit(void);
void __cold btrfs_delayed_inode_exit(void)
void __cold btrfs_delayed_inode_exit(void);
void __cold btrfs_delayed_ref_exit(void)
void __cold btrfs_delayed_ref_exit(void);
void __cold btrfs_destroy_dio(void)
void __cold btrfs_destroy_dio(void);
static int __cold init_tree_roots(struct btrfs_fs_info *fs_info)
int __cold open_ctree(struct super_block *sb, struct btrfs_fs_devices *fs_devices)
void __cold close_ctree(struct btrfs_fs_info *fs_info)
int __cold open_ctree(struct super_block *sb, struct btrfs_fs_devices *fs_devices);
void __cold close_ctree(struct btrfs_fs_info *fs_info);
void __cold btrfs_extent_state_free_cachep(void)
static void __cold extent_io_tree_panic(const struct extent_io_tree *tree,
void __cold btrfs_extent_state_free_cachep(void);
void __cold extent_buffer_free_cachep(void)
void __cold extent_buffer_free_cachep(void);
void __cold btrfs_extent_map_exit(void)
void __cold btrfs_extent_map_exit(void);
void __cold btrfs_free_space_exit(void)
void __cold btrfs_free_space_exit(void);
static void __cold btrfs_print_data_csum_error(struct btrfs_inode *inode,
void __cold btrfs_destroy_cachep(void)
__cold
__printf(3, 4) __cold
void __cold btrfs_warn_32bit_limit(struct btrfs_fs_info *fs_info)
void __cold btrfs_err_32bit_limit(struct btrfs_fs_info *fs_info)
__cold
static inline __printf(2, 3) __cold
__cold
__cold
__printf(3, 4) __cold
void __cold ordered_data_exit(void)
void __cold ordered_data_exit(void);
__cold
__cold void btrfs_dump_space_info_for_trans_abort(struct btrfs_fs_info *fs_info)
void __cold btrfs_subpage_dump_bitmap(const struct btrfs_fs_info *fs_info,
void __cold btrfs_subpage_dump_bitmap(const struct btrfs_fs_info *fs_info,
static __cold void btrfs_interface_exit(void)
void __cold btrfs_exit_sysfs(void)
void __cold btrfs_exit_sysfs(void);
void __cold __btrfs_abort_transaction(struct btrfs_trans_handle *trans,
void __cold btrfs_transaction_exit(void)
void __cold __btrfs_abort_transaction(struct btrfs_trans_handle *trans,
void __cold btrfs_transaction_exit(void);
__cold
__cold
__cold
__cold
__cold
__cold
__cold
void __cold btrfs_assign_next_active_device(struct btrfs_device *device,
void __printf(3, 4) __cold
__printf(3, 4) __cold;
extern __printf(3, 4) __cold
__printf(3, 4) __cold
void __printf(3, 4) __cold
__printf(3, 0) __cold
__printf(3, 4) __cold
__printf(3, 4) __cold
__printf(2, 3) __cold
__printf(2, 3) __cold
__printf(2, 3) __cold
__printf(2, 3) __cold
__printf(2, 3) __cold
__printf(2, 3) __cold
__printf(2, 3) __cold
void __fortify_panic(const u8 reason, const size_t avail, const size_t size) __cold __noreturn;
#define __init __section(".init.text") __cold __latent_entropy \
#define __exit __section(".exit.text") __exitused __cold notrace
# define __list_valid_slowpath __cold __preserve_most
void panic(const char *fmt, ...) __noreturn __cold;
void vpanic(const char *fmt, va_list args) __noreturn __cold;
static inline __printf(1, 2) __cold
asmlinkage __printf(1, 2) __cold
__printf(1, 2) __cold int _printk_deferred(const char *fmt, ...);
extern asmlinkage void dump_stack_lvl(const char *log_lvl) __cold;
extern asmlinkage void dump_stack(void) __cold;
static inline __printf(1, 2) __cold
static inline __printf(1, 2) __cold
__printf(3, 4) __cold
__printf(2, 3) __cold
__printf(2, 3) __cold
__printf(2, 3) __cold
__printf(2, 3) __cold
__printf(2, 3) __cold
__printf(2, 3) __cold
__printf(2, 3) __cold
__printf(4, 5) __cold
__printf(4, 5) __cold
static inline __printf(4, 5) __cold
static inline __printf(4, 5) __cold
__printf(2, 3) __cold
__printf(2, 3) __cold
__printf(2, 3) __cold
__printf(2, 3) __cold
__printf(2, 3) __cold
__printf(2, 3) __cold
__printf(2, 3) __cold
__printf(2, 3) __cold
__printf(2, 3) __cold
static __cold bool io_cancel_defer_files(struct io_ring_ctx *ctx,
__cold bool io_cancel_ctx_cb(struct io_wq_work *work, void *data)
static __cold bool io_uring_try_cancel_iowq(struct io_ring_ctx *ctx)
__cold bool io_uring_try_cancel_requests(struct io_ring_ctx *ctx,
__cold void io_uring_cancel_generic(bool cancel_all, struct io_sq_data *sqd)
__cold bool io_uring_try_cancel_requests(struct io_ring_ctx *ctx,
__cold void io_uring_cancel_generic(bool cancel_all, struct io_sq_data *sqd);
__cold bool io_cancel_ctx_cb(struct io_wq_work *work, void *data);
static __cold void common_tracking_show_fdinfo(struct io_ring_ctx *ctx,
__cold void io_uring_show_fdinfo(struct seq_file *m, struct file *file)
static __cold void napi_show_fdinfo(struct io_ring_ctx *ctx,
__cold void io_iopoll_try_reap_events(struct io_ring_ctx *ctx)
static __cold void io_drain_req(struct io_kiocb *req)
static __cold int io_init_fail_req(struct io_kiocb *req, int err)
static __cold int io_submit_fail_init(const struct io_uring_sqe *sqe,
static __cold void io_ring_ctx_ref_free(struct percpu_ref *ref)
static __cold void __io_req_caches_free(struct io_ring_ctx *ctx)
static __cold void io_req_caches_free(struct io_ring_ctx *ctx)
static __cold void io_ring_ctx_free(struct io_ring_ctx *ctx)
static __cold void io_activate_pollwq_cb(struct callback_head *cb)
__cold void io_activate_pollwq(struct io_ring_ctx *ctx)
static __cold struct io_ring_ctx *io_ring_ctx_alloc(struct io_uring_params *p)
static __cold void io_tctx_exit_cb(struct callback_head *cb)
static __cold void io_ring_exit_work(struct work_struct *work)
static __cold void io_ring_ctx_wait_and_kill(struct io_ring_ctx *ctx)
static __cold int io_allocate_scq_urings(struct io_ring_ctx *ctx,
static __cold int io_uring_create(struct io_ctx_config *config)
static __cold noinline void io_queue_deferred(struct io_ring_ctx *ctx)
__cold void io_uring_drop_tctx_refs(struct task_struct *task)
static __cold bool io_cqring_add_overflow(struct io_ring_ctx *ctx,
static __cold void io_cqe_overflow(struct io_ring_ctx *ctx, struct io_cqe *cqe,
static __cold bool io_cqe_overflow_locked(struct io_ring_ctx *ctx,
__cold bool __io_alloc_req_refill(struct io_ring_ctx *ctx)
__cold void io_free_req(struct io_kiocb *req)
__cold void io_uring_drop_tctx_refs(struct task_struct *task);
__cold void io_iopoll_try_reap_events(struct io_ring_ctx *ctx);
__cold int io_uring_mmap(struct file *file, struct vm_area_struct *vma)
static __cold int io_pollfree_wake(struct io_kiocb *req, struct io_poll *poll)
__cold bool io_poll_remove_all(struct io_ring_ctx *ctx, struct io_uring_task *tctx,
static __cold int io_parse_restrictions(void __user *arg, unsigned int nr_args,
static __cold int io_register_restrictions(struct io_ring_ctx *ctx,
static __cold int __io_register_iowq_aff(struct io_ring_ctx *ctx,
static __cold int io_register_iowq_aff(struct io_ring_ctx *ctx,
static __cold int io_unregister_iowq_aff(struct io_ring_ctx *ctx)
static __cold int io_register_iowq_max_workers(struct io_ring_ctx *ctx,
static __cold int io_probe(struct io_ring_ctx *ctx, void __user *arg,
__cold void io_rsrc_data_free(struct io_ring_ctx *ctx,
__cold int io_rsrc_data_alloc(struct io_rsrc_data *data, unsigned nr)
__cold int io_register_rsrc(struct io_ring_ctx *ctx, void __user *arg,
__cold int io_sq_offload_create(struct io_ring_ctx *ctx,
__cold int io_sqpoll_wq_cpu_affinity(struct io_ring_ctx *ctx,
static __cold void io_sqd_update_thread_idle(struct io_sq_data *sqd)
__cold void io_uring_del_tctx_node(unsigned long index)
__cold void io_uring_clean_tctx(struct io_uring_task *tctx)
__cold int io_uring_alloc_task_context(struct task_struct *task,
__cold void io_flush_timeouts(struct io_ring_ctx *ctx)
__cold bool io_kill_timeouts(struct io_ring_ctx *ctx, struct io_uring_task *tctx,
static __cold bool io_flush_killed_timeouts(struct list_head *list, int err)
__cold void io_flush_timeouts(struct io_ring_ctx *ctx);
__cold bool io_kill_timeouts(struct io_ring_ctx *ctx, struct io_uring_task *tctx,
void __cold io_move_task_work_from_local(struct io_ring_ctx *ctx)
static __cold void __io_fallback_tw(struct llist_node *node, bool sync)
__cold void io_fallback_req_func(struct work_struct *work);
__cold void io_move_task_work_from_local(struct io_ring_ctx *ctx);
static __printf(3, 4) __cold
static __cold struct net_device_core_stats __percpu *netdev_core_stats_alloc(
static __cold void tcp_gro_dev_warn(const struct sock *sk, const struct sk_buff *skb,
static __cold noinline int
static unsigned int __cold
static void __cold nf_tcp_handle_invalid(struct nf_conn *ct,