arch/arc/include/asm/entry-arcv2.h
52
; 1. SP auto-switched to kernel mode stack
arch/arc/include/asm/entry-arcv2.h
81
; 1. SP auto-switched to kernel mode stack
arch/arm/kernel/sys_oabi-compat.c
200
static int get_oabi_flock(struct flock64 *kernel, struct oabi_flock64 __user *arg)
arch/arm/kernel/sys_oabi-compat.c
208
kernel->l_type = user.l_type;
arch/arm/kernel/sys_oabi-compat.c
209
kernel->l_whence = user.l_whence;
arch/arm/kernel/sys_oabi-compat.c
210
kernel->l_start = user.l_start;
arch/arm/kernel/sys_oabi-compat.c
211
kernel->l_len = user.l_len;
arch/arm/kernel/sys_oabi-compat.c
212
kernel->l_pid = user.l_pid;
arch/arm/kernel/sys_oabi-compat.c
217
static int put_oabi_flock(struct flock64 *kernel, struct oabi_flock64 __user *arg)
arch/arm/kernel/sys_oabi-compat.c
221
user.l_type = kernel->l_type;
arch/arm/kernel/sys_oabi-compat.c
222
user.l_whence = kernel->l_whence;
arch/arm/kernel/sys_oabi-compat.c
223
user.l_start = kernel->l_start;
arch/arm/kernel/sys_oabi-compat.c
224
user.l_len = kernel->l_len;
arch/arm/kernel/sys_oabi-compat.c
225
user.l_pid = kernel->l_pid;
arch/arm/kernel/sys_oabi-compat.c
287
struct epoll_event kernel;
arch/arm/kernel/sys_oabi-compat.c
293
kernel.events = user.events;
arch/arm/kernel/sys_oabi-compat.c
294
kernel.data = user.data;
arch/arm/kernel/sys_oabi-compat.c
296
return do_epoll_ctl(epfd, op, fd, &kernel, false);
arch/arm64/kernel/kexec_image.c
37
char *kernel, unsigned long kernel_len,
arch/arm64/kernel/kexec_image.c
54
h = (struct arm64_image_header *)kernel;
arch/arm64/kernel/kexec_image.c
80
kbuf.buffer = kernel;
arch/loongarch/kernel/kexec_efi.c
39
char *kernel, unsigned long kernel_len,
arch/loongarch/kernel/kexec_efi.c
49
h = (struct loongarch_image_header *)kernel;
arch/loongarch/kernel/kexec_efi.c
61
kbuf.buffer = kernel;
arch/loongarch/kernel/kexec_elf.c
55
char *kernel, unsigned long kernel_len,
arch/loongarch/kernel/kexec_elf.c
66
ret = kexec_build_elf_info(kernel, kernel_len, &ehdr, &elf_info);
arch/mips/include/asm/mach-cavium-octeon/kernel-entry-init.h
141
# Someone tried to boot SMP with a non SMP kernel. All extra cores
arch/mips/include/asm/mach-cavium-octeon/kernel-entry-init.h
22
# by the kernel
arch/mips/include/asm/mach-ip27/kernel-entry-init.h
39
or t1, t1, t0 # Physical load address of kernel text
arch/mips/include/asm/mach-ip27/kernel-entry-init.h
40
or t2, t2, t0 # Physical load address of kernel data
arch/mips/include/asm/unaligned-emul.h
763
#define LoadHWU(addr, value, res) _LoadHWU(addr, value, res, kernel)
arch/mips/include/asm/unaligned-emul.h
765
#define LoadWU(addr, value, res) _LoadWU(addr, value, res, kernel)
arch/mips/include/asm/unaligned-emul.h
767
#define LoadHW(addr, value, res) _LoadHW(addr, value, res, kernel)
arch/mips/include/asm/unaligned-emul.h
769
#define LoadW(addr, value, res) _LoadW(addr, value, res, kernel)
arch/mips/include/asm/unaligned-emul.h
773
#define StoreHW(addr, value, res) _StoreHW(addr, value, res, kernel)
arch/mips/include/asm/unaligned-emul.h
775
#define StoreW(addr, value, res) _StoreW(addr, value, res, kernel)
arch/powerpc/include/asm/pgalloc.h
23
pte_t *pte_fragment_alloc(struct mm_struct *mm, int kernel);
arch/powerpc/include/asm/pgalloc.h
36
void pte_fragment_free(unsigned long *table, int kernel);
arch/powerpc/mm/book3s64/slb.c
551
static enum slb_index alloc_slb_index(bool kernel)
arch/powerpc/mm/book3s64/slb.c
574
if (kernel)
arch/powerpc/mm/book3s64/slb.c
585
if (kernel)
arch/powerpc/mm/book3s64/slb.c
597
unsigned long flags, int ssize, bool kernel)
arch/powerpc/mm/book3s64/slb.c
618
index = alloc_slb_index(kernel);
arch/powerpc/mm/book3s64/slb.c
640
if (!kernel || slb_cache_index == 3) {
arch/powerpc/mm/book3s64/slb.c
648
if (kernel)
arch/powerpc/mm/book3s64/slb.c
656
if (!kernel)
arch/powerpc/mm/pgtable-frag.c
104
return __alloc_for_ptecache(mm, kernel);
arch/powerpc/mm/pgtable-frag.c
116
void pte_fragment_free(unsigned long *table, int kernel)
arch/powerpc/mm/pgtable-frag.c
125
if (kernel || !folio_test_clear_active(ptdesc_folio(ptdesc)))
arch/powerpc/mm/pgtable-frag.c
55
static pte_t *__alloc_for_ptecache(struct mm_struct *mm, int kernel)
arch/powerpc/mm/pgtable-frag.c
61
if (!kernel)
arch/powerpc/mm/pgtable-frag.c
96
pte_t *pte_fragment_alloc(struct mm_struct *mm, int kernel)
arch/riscv/kernel/kexec_image.c
37
char *kernel, unsigned long kernel_len,
arch/riscv/kernel/kexec_image.c
48
h = (struct riscv_image_header *)kernel;
arch/riscv/kernel/kexec_image.c
69
kbuf.buffer = kernel;
arch/s390/include/asm/kexec.h
72
int s390_verify_sig(const char *kernel, unsigned long kernel_len);
arch/s390/kernel/kexec_elf.c
23
void *kernel;
arch/s390/kernel/kexec_elf.c
26
kernel = image->kernel_buf;
arch/s390/kernel/kexec_elf.c
27
ehdr = (Elf_Ehdr *)kernel;
arch/s390/kernel/kexec_elf.c
39
buf.buffer = kernel + phdr->p_offset;
arch/s390/kernel/kexec_elf.c
69
char *kernel, unsigned long kernel_len,
arch/s390/kernel/kexec_elf.c
79
ehdr = (Elf_Ehdr *)kernel;
arch/s390/kernel/kexec_image.c
46
char *kernel, unsigned long kernel_len,
arch/s390/kernel/machine_kexec_file.c
29
int s390_verify_sig(const char *kernel, unsigned long kernel_len)
arch/s390/kernel/machine_kexec_file.c
43
if (memcmp(kernel + kernel_len - marker_len, MODULE_SIG_STRING,
arch/s390/kernel/machine_kexec_file.c
48
ms = (void *)kernel + kernel_len - sizeof(*ms);
arch/s390/kernel/machine_kexec_file.c
69
ret = verify_pkcs7_signature(kernel, kernel_len,
arch/s390/kernel/machine_kexec_file.c
70
kernel + kernel_len, sig_len,
arch/s390/kernel/machine_kexec_file.c
75
ret = verify_pkcs7_signature(kernel, kernel_len,
arch/s390/kernel/machine_kexec_file.c
76
kernel + kernel_len, sig_len,
arch/s390/kernel/perf_pai.c
192
static u64 pai_getdata(struct perf_event *event, bool kernel)
arch/s390/kernel/perf_pai.c
204
kernel ? pp->kernel_offset : 0);
arch/s390/kernel/perf_pai.c
209
kernel ? pp->kernel_offset : 0);
arch/x86/kernel/kexec-bzimage64.c
475
static void *bzImage64_load(struct kimage *image, char *kernel,
arch/x86/kernel/kexec-bzimage64.c
496
header = (struct setup_header *)(kernel + setup_hdr_offset);
arch/x86/kernel/kexec-bzimage64.c
597
setup_header_size = 0x0202 + kernel[0x0201] - setup_hdr_offset;
arch/x86/kernel/kexec-bzimage64.c
600
memcpy(¶ms->hdr, (kernel + setup_hdr_offset), setup_header_size);
arch/x86/kernel/kexec-bzimage64.c
614
kbuf.buffer = kernel + kern16_size;
arch/x86/xen/mmu_pv.c
1374
static void __xen_write_cr3(bool kernel, unsigned long cr3)
arch/x86/xen/mmu_pv.c
1379
trace_xen_mmu_write_cr3(kernel, cr3);
arch/x86/xen/mmu_pv.c
1386
WARN_ON(mfn == 0 && kernel);
arch/x86/xen/mmu_pv.c
1388
op.cmd = kernel ? MMUEXT_NEW_BASEPTR : MMUEXT_NEW_USER_BASEPTR;
arch/x86/xen/mmu_pv.c
1393
if (kernel) {
arch/xtensa/mm/tlb.c
243
bool kernel = tlb_asid == 1;
arch/xtensa/mm/tlb.c
246
if (tlb_asid > 0 && ((vpn < TASK_SIZE) == kernel)) {
arch/xtensa/mm/tlb.c
249
kernel ? "kernel" : "user");
drivers/gpu/drm/i915/gt/gen7_renderclear.c
209
const struct cb_kernel *kernel,
drivers/gpu/drm/i915/gt/gen7_renderclear.c
213
gen7_fill_kernel_data(state, kernel->data, kernel->size);
drivers/gpu/drm/i915/gt/intel_engine_cs.c
1407
static struct lock_class_key kernel;
drivers/gpu/drm/i915/gt/intel_engine_cs.c
1415
&kernel, "ggtt_bind_context");
drivers/gpu/drm/i915/gt/intel_engine_cs.c
1421
static struct lock_class_key kernel;
drivers/gpu/drm/i915/gt/intel_engine_cs.c
1425
&kernel, "kernel_context");
drivers/gpu/drm/loongson/lsdc_ttm.c
433
bool kernel,
drivers/gpu/drm/loongson/lsdc_ttm.c
466
if (kernel)
drivers/gpu/drm/loongson/lsdc_ttm.h
68
bool kernel,
drivers/gpu/drm/qxl/qxl_drv.h
302
bool discardable, bool kernel,
drivers/gpu/drm/qxl/qxl_gem.c
48
bool discardable, bool kernel,
drivers/gpu/drm/qxl/qxl_gem.c
59
r = qxl_bo_create(qdev, size, kernel, false, initial_domain, 0, surf, &qbo);
drivers/gpu/drm/qxl/qxl_object.c
105
bool kernel, bool pinned, u32 domain, u32 priority,
drivers/gpu/drm/qxl/qxl_object.c
109
struct ttm_operation_ctx ctx = { !kernel, false };
drivers/gpu/drm/qxl/qxl_object.c
114
if (kernel)
drivers/gpu/drm/qxl/qxl_object.h
58
bool kernel, bool pinned, u32 domain,
drivers/gpu/drm/radeon/radeon.h
537
u32 flags, bool kernel,
drivers/gpu/drm/radeon/radeon_gem.c
119
r = radeon_bo_create(rdev, size, alignment, kernel, initial_domain,
drivers/gpu/drm/radeon/radeon_gem.c
95
u32 flags, bool kernel,
drivers/gpu/drm/radeon/radeon_object.c
129
unsigned long size, int byte_align, bool kernel,
drivers/gpu/drm/radeon/radeon_object.c
141
if (kernel) {
drivers/gpu/drm/radeon/radeon_object.c
205
&bo->placement, page_align, !kernel, sg, resv,
drivers/gpu/drm/radeon/radeon_object.h
138
bool kernel, u32 domain, u32 flags,
drivers/gpu/drm/vc4/vc4_hvs.c
389
const u32 *kernel)
drivers/gpu/drm/vc4/vc4_hvs.c
410
writel(kernel[i], &dst_kernel[i]);
drivers/gpu/drm/vc4/vc4_hvs.c
412
writel(kernel[VC4_KERNEL_DWORDS - i - 1],
drivers/gpu/drm/vc4/vc4_plane.c
1659
u32 kernel = VC4_SET_FIELD(vc4->hvs->mitchell_netravali_filter.start,
drivers/gpu/drm/vc4/vc4_plane.c
1663
vc4_dlist_write(vc4_state, kernel);
drivers/gpu/drm/vc4/vc4_plane.c
1665
vc4_dlist_write(vc4_state, kernel);
drivers/gpu/drm/vc4/vc4_plane.c
1667
vc4_dlist_write(vc4_state, kernel);
drivers/gpu/drm/vc4/vc4_plane.c
1669
vc4_dlist_write(vc4_state, kernel);
drivers/gpu/drm/vc4/vc4_plane.c
2071
u32 kernel =
drivers/gpu/drm/vc4/vc4_plane.c
2076
vc4_dlist_write(vc4_state, kernel);
drivers/gpu/drm/vc4/vc4_plane.c
2078
vc4_dlist_write(vc4_state, kernel);
drivers/gpu/drm/vc4/vc4_plane.c
2080
vc4_dlist_write(vc4_state, kernel);
drivers/gpu/drm/vc4/vc4_plane.c
2082
vc4_dlist_write(vc4_state, kernel);
drivers/gpu/drm/vmwgfx/vmwgfx_drv.h
395
bool kernel;
drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c
3236
if (unlikely(!sw_context->kernel)) {
drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c
3689
if (unlikely(!entry->user_allow && !sw_context->kernel))
drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c
4125
sw_context->kernel = false;
drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c
4141
sw_context->kernel = true;
drivers/misc/fastrpc.c
1084
if (!kernel) {
drivers/misc/fastrpc.c
1116
u32 kernel)
drivers/misc/fastrpc.c
1140
if (!kernel) {
drivers/misc/fastrpc.c
1165
u32 kernel, uint32_t handle)
drivers/misc/fastrpc.c
1176
if (kernel)
drivers/misc/fastrpc.c
1195
static int fastrpc_internal_invoke(struct fastrpc_user *fl, u32 kernel,
drivers/misc/fastrpc.c
1210
if (handle == FASTRPC_INIT_HANDLE && !kernel) {
drivers/misc/fastrpc.c
1215
ctx = fastrpc_context_alloc(fl, kernel, sc, args);
drivers/misc/fastrpc.c
1219
err = fastrpc_get_args(kernel, ctx);
drivers/misc/fastrpc.c
1226
err = fastrpc_invoke_send(fl->sctx, ctx, kernel, handle);
drivers/misc/fastrpc.c
1230
if (kernel) {
drivers/misc/fastrpc.c
1243
err = fastrpc_put_args(ctx, kernel);
drivers/misc/fastrpc.c
595
struct fastrpc_user *user, u32 kernel, u32 sc,
drivers/misc/fastrpc.c
986
static int fastrpc_get_args(u32 kernel, struct fastrpc_invoke_ctx *ctx)
drivers/misc/ocxl/link.c
515
static u64 calculate_cfg_state(bool kernel)
drivers/misc/ocxl/link.c
527
if (kernel) {
drivers/net/ethernet/ibm/ehea/ehea.h
164
struct h_epa kernel; /* kernel space accessible resource,
drivers/net/ethernet/ibm/ehea/ehea_hw.h
206
struct h_epa epa = qp->epas.kernel;
drivers/net/ethernet/ibm/ehea/ehea_hw.h
213
struct h_epa epa = qp->epas.kernel;
drivers/net/ethernet/ibm/ehea/ehea_hw.h
220
struct h_epa epa = qp->epas.kernel;
drivers/net/ethernet/ibm/ehea/ehea_hw.h
227
struct h_epa epa = qp->epas.kernel;
drivers/net/ethernet/ibm/ehea/ehea_hw.h
234
struct h_epa epa = cq->epas.kernel;
drivers/net/ethernet/ibm/ehea/ehea_hw.h
241
struct h_epa epa = cq->epas.kernel;
drivers/net/ethernet/ibm/ehea/ehea_hw.h
248
struct h_epa epa = my_cq->epas.kernel;
drivers/net/ethernet/ibm/ehea/ehea_phyp.h
53
epas->kernel.addr = ioremap((paddr_kernel & PAGE_MASK), PAGE_SIZE) +
drivers/net/ethernet/ibm/ehea/ehea_phyp.h
60
if (epas->kernel.addr)
drivers/net/ethernet/ibm/ehea/ehea_phyp.h
61
iounmap((void __iomem *)((u64)epas->kernel.addr & PAGE_MASK));
drivers/net/ethernet/ibm/ehea/ehea_phyp.h
64
epas->kernel.addr = 0;
drivers/net/wireless/mediatek/mt76/mt7915/coredump.c
332
strscpy(dump->kernel, init_utsname()->release, sizeof(dump->kernel));
drivers/net/wireless/mediatek/mt76/mt7915/coredump.h
26
char kernel[64];
drivers/net/wireless/mediatek/mt76/mt7996/coredump.c
192
strscpy(dump->kernel, init_utsname()->release, sizeof(dump->kernel));
drivers/net/wireless/mediatek/mt76/mt7996/coredump.h
21
char kernel[64];
drivers/perf/apple_m1_cpu_pmu.c
345
bool kernel, bool host)
drivers/perf/apple_m1_cpu_pmu.c
368
if (kernel)
drivers/perf/apple_m1_cpu_pmu.c
410
bool kernel = config_base & M1_PMU_CFG_COUNT_KERNEL;
drivers/perf/apple_m1_cpu_pmu.c
416
__m1_pmu_configure_event_filter(index, user && host, kernel && host, true);
drivers/perf/apple_m1_cpu_pmu.c
417
__m1_pmu_configure_event_filter(index, user && guest, kernel && guest, false);
drivers/perf/apple_m1_cpu_pmu.c
424
bool user, kernel;
drivers/perf/apple_m1_cpu_pmu.c
429
kernel = event->hw.config_base & M1_PMU_CFG_COUNT_KERNEL;
drivers/staging/media/atomisp/pci/runtime/debug/src/ia_css_debug.c
855
#define FIND_DMEM_PARAMS_TYPE(stream, kernel, type) \
drivers/staging/media/atomisp/pci/runtime/debug/src/ia_css_debug.c
857
findf_dmem_params(stream, offsetof(struct ia_css_memory_offsets, dmem.kernel))
drivers/staging/media/atomisp/pci/runtime/debug/src/ia_css_debug.c
859
#define FIND_DMEM_PARAMS(stream, kernel) FIND_DMEM_PARAMS_TYPE(stream, kernel, kernel)
include/linux/bpfptr.h
18
return (bpfptr_t) { .kernel = p, .is_kernel = true };
include/linux/bpfptr.h
37
return !bpfptr.kernel;
include/linux/bpfptr.h
44
bpfptr->kernel += val;
include/linux/bpfptr.h
54
return copy_from_kernel_nofault(dst, src.kernel + offset, size);
include/linux/bpfptr.h
85
return strncpy_from_kernel_nofault(dst, src.kernel, count);
include/linux/kernel_read_file.h
12
id(MODULE, kernel-module) \
include/linux/kernel_read_file.h
17
id(MODULE_COMPRESSED, kernel-module-compressed) \
include/linux/kexec.h
255
int kexec_kernel_verify_pe_sig(const char *kernel, unsigned long kernel_len);
include/linux/lsm_hook_defs.h
431
LSM_HOOK(int, 0, bpf, int cmd, union bpf_attr *attr, unsigned int size, bool kernel)
include/linux/lsm_hook_defs.h
435
struct bpf_token *token, bool kernel)
include/linux/lsm_hook_defs.h
438
struct bpf_token *token, bool kernel)
include/linux/netfilter/x_tables.h
458
} kernel;
include/linux/netfilter/x_tables.h
474
} kernel;
include/linux/perf_event.h
1738
get_perf_callchain(struct pt_regs *regs, bool kernel, bool user,
include/linux/security.h
2267
extern int security_bpf(int cmd, union bpf_attr *attr, unsigned int size, bool kernel);
include/linux/security.h
2271
struct bpf_token *token, bool kernel);
include/linux/security.h
2274
struct bpf_token *token, bool kernel);
include/linux/security.h
2283
unsigned int size, bool kernel)
include/linux/security.h
2299
struct bpf_token *token, bool kernel)
include/linux/security.h
2308
struct bpf_token *token, bool kernel)
include/linux/sockptr.h
106
memcpy(dst, src.kernel, size);
include/linux/sockptr.h
115
memcpy(dst.kernel + offset, src, size);
include/linux/sockptr.h
156
size_t len = min(strnlen(src.kernel, count - 1) + 1, count);
include/linux/sockptr.h
158
memcpy(dst, src.kernel, len);
include/linux/sockptr.h
16
void *kernel;
include/linux/sockptr.h
169
return memchr_inv(src.kernel + offset, 0, size) == NULL;
include/linux/sockptr.h
29
return (sockptr_t) { .kernel = p, .is_kernel = true };
include/linux/sockptr.h
40
return !sockptr.kernel;
include/linux/sockptr.h
49
memcpy(dst, src.kernel + offset, size);
include/linux/sockptr.h
99
char *p = src.kernel;
include/trace/events/xen.h
368
TP_PROTO(bool kernel, unsigned long cr3),
include/trace/events/xen.h
369
TP_ARGS(kernel, cr3),
include/trace/events/xen.h
371
__field(bool, kernel)
include/trace/events/xen.h
374
TP_fast_assign(__entry->kernel = kernel;
include/trace/events/xen.h
377
__entry->kernel ? "kernel" : "user", __entry->cr3)
include/uapi/linux/netfilter/x_tables.h
25
} kernel;
include/uapi/linux/netfilter/x_tables.h
48
} kernel;
include/uapi/sound/asequencer.h
495
void *kernel; /* reserved for kernel use (must be NULL) */
kernel/bpf/stackmap.c
329
bool kernel = !user;
kernel/bpf/stackmap.c
337
trace = get_perf_callchain(regs, kernel, user, max_depth,
kernel/bpf/stackmap.c
373
bool kernel, user;
kernel/bpf/stackmap.c
387
kernel = !user;
kernel/bpf/stackmap.c
396
if (kernel) {
kernel/bpf/stackmap.c
435
bool kernel = !user;
kernel/bpf/stackmap.c
442
if (kernel && user_build_id)
kernel/bpf/stackmap.c
469
} else if (kernel && task) {
kernel/bpf/stackmap.c
472
trace = get_perf_callchain(regs, kernel, user, max_depth,
kernel/bpf/stackmap.c
603
bool kernel, user;
kernel/bpf/stackmap.c
615
kernel = !user;
kernel/bpf/stackmap.c
624
if (kernel) {
kernel/bpf/syscall.c
104
res = memchr_inv(uaddr.kernel + expected_size, 0,
kernel/events/callchain.c
220
get_perf_callchain(struct pt_regs *regs, bool kernel, bool user,
kernel/events/callchain.c
228
if (crosstask && user && !kernel)
kernel/events/callchain.c
241
if (kernel && !user_mode(regs)) {
kernel/events/core.c
11945
unsigned int kernel = 0;
kernel/events/core.c
11984
kernel = 1;
kernel/events/core.c
12038
if (!kernel) {
kernel/events/core.c
12074
kernel = 0;
kernel/events/core.c
8506
bool kernel = !event->attr.exclude_callchain_kernel;
kernel/events/core.c
8520
if (!kernel && !user)
kernel/events/core.c
8527
callchain = get_perf_callchain(regs, kernel, user, max_stack,
kernel/kexec_file.c
147
int kexec_kernel_verify_pe_sig(const char *kernel, unsigned long kernel_len)
kernel/kexec_file.c
151
ret = verify_pefile_signature(kernel, kernel_len,
kernel/kexec_file.c
155
ret = verify_pefile_signature(kernel, kernel_len,
net/ipv4/ip_sockglue.c
1653
msg.msg_control = optval.kernel;
net/ipv4/netfilter/arp_tables.c
1070
module_put(t->u.kernel.target->me);
net/ipv4/netfilter/arp_tables.c
1112
t->u.kernel.target = target;
net/ipv4/netfilter/arp_tables.c
1123
module_put(t->u.kernel.target->me);
net/ipv4/netfilter/arp_tables.c
236
if (!t->u.kernel.target->target) {
net/ipv4/netfilter/arp_tables.c
268
acpar.target = t->u.kernel.target;
net/ipv4/netfilter/arp_tables.c
270
verdict = t->u.kernel.target->target(skb, &acpar);
net/ipv4/netfilter/arp_tables.c
395
.target = t->u.kernel.target,
net/ipv4/netfilter/arp_tables.c
423
t->u.kernel.target = target;
net/ipv4/netfilter/arp_tables.c
430
module_put(t->u.kernel.target->me);
net/ipv4/netfilter/arp_tables.c
507
par.target = t->u.kernel.target;
net/ipv4/netfilter/arp_tables.c
748
off += xt_compat_target_offset(t->u.kernel.target);
net/ipv4/netfilter/ip_tables.c
1268
m->u.kernel.match = match;
net/ipv4/netfilter/ip_tables.c
1280
module_put(ematch->u.kernel.match->me);
net/ipv4/netfilter/ip_tables.c
1282
module_put(t->u.kernel.target->me);
net/ipv4/netfilter/ip_tables.c
1333
t->u.kernel.target = target;
net/ipv4/netfilter/ip_tables.c
1344
module_put(t->u.kernel.target->me);
net/ipv4/netfilter/ip_tables.c
1349
module_put(ematch->u.kernel.match->me);
net/ipv4/netfilter/ip_tables.c
162
if (strcmp(t->target.u.kernel.target->name, XT_ERROR_TARGET) == 0) {
net/ipv4/netfilter/ip_tables.c
170
strcmp(t->target.u.kernel.target->name,
net/ipv4/netfilter/ip_tables.c
291
acpar.match = ematch->u.kernel.match;
net/ipv4/netfilter/ip_tables.c
301
WARN_ON(!t->u.kernel.target);
net/ipv4/netfilter/ip_tables.c
310
if (!t->u.kernel.target->target) {
net/ipv4/netfilter/ip_tables.c
342
acpar.target = t->u.kernel.target;
net/ipv4/netfilter/ip_tables.c
345
verdict = t->u.kernel.target->target(skb, &acpar);
net/ipv4/netfilter/ip_tables.c
455
par.match = m->u.kernel.match;
net/ipv4/netfilter/ip_tables.c
468
par->match = m->u.kernel.match;
net/ipv4/netfilter/ip_tables.c
485
m->u.kernel.match = match;
net/ipv4/netfilter/ip_tables.c
493
module_put(m->u.kernel.match->me);
net/ipv4/netfilter/ip_tables.c
504
.target = t->u.kernel.target,
net/ipv4/netfilter/ip_tables.c
550
t->u.kernel.target = target;
net/ipv4/netfilter/ip_tables.c
558
module_put(t->u.kernel.target->me);
net/ipv4/netfilter/ip_tables.c
648
par.target = t->u.kernel.target;
net/ipv4/netfilter/ip_tables.c
901
off += xt_compat_match_offset(ematch->u.kernel.match);
net/ipv4/netfilter/ip_tables.c
903
off += xt_compat_target_offset(t->u.kernel.target);
net/ipv6/ipv6_sockglue.c
1122
msg.msg_control = optval.kernel;
net/ipv6/netfilter/ip6_tables.c
1284
m->u.kernel.match = match;
net/ipv6/netfilter/ip6_tables.c
1296
module_put(ematch->u.kernel.match->me);
net/ipv6/netfilter/ip6_tables.c
1298
module_put(t->u.kernel.target->me);
net/ipv6/netfilter/ip6_tables.c
1349
t->u.kernel.target = target;
net/ipv6/netfilter/ip6_tables.c
1360
module_put(t->u.kernel.target->me);
net/ipv6/netfilter/ip6_tables.c
1365
module_put(ematch->u.kernel.match->me);
net/ipv6/netfilter/ip6_tables.c
187
if (strcmp(t->target.u.kernel.target->name, XT_ERROR_TARGET) == 0) {
net/ipv6/netfilter/ip6_tables.c
195
strcmp(t->target.u.kernel.target->name,
net/ipv6/netfilter/ip6_tables.c
314
acpar.match = ematch->u.kernel.match;
net/ipv6/netfilter/ip6_tables.c
324
WARN_ON(!t->u.kernel.target);
net/ipv6/netfilter/ip6_tables.c
333
if (!t->u.kernel.target->target) {
net/ipv6/netfilter/ip6_tables.c
363
acpar.target = t->u.kernel.target;
net/ipv6/netfilter/ip6_tables.c
366
verdict = t->u.kernel.target->target(skb, &acpar);
net/ipv6/netfilter/ip6_tables.c
473
par.match = m->u.kernel.match;
net/ipv6/netfilter/ip6_tables.c
485
par->match = m->u.kernel.match;
net/ipv6/netfilter/ip6_tables.c
503
m->u.kernel.match = match;
net/ipv6/netfilter/ip6_tables.c
511
module_put(m->u.kernel.match->me);
net/ipv6/netfilter/ip6_tables.c
522
.target = t->u.kernel.target,
net/ipv6/netfilter/ip6_tables.c
569
t->u.kernel.target = target;
net/ipv6/netfilter/ip6_tables.c
576
module_put(t->u.kernel.target->me);
net/ipv6/netfilter/ip6_tables.c
665
par.target = t->u.kernel.target;
net/ipv6/netfilter/ip6_tables.c
917
off += xt_compat_match_offset(ematch->u.kernel.match);
net/ipv6/netfilter/ip6_tables.c
919
off += xt_compat_target_offset(t->u.kernel.target);
net/netfilter/x_tables.c
1157
const struct xt_target *target = t->u.kernel.target;
net/netfilter/x_tables.c
1185
const struct xt_target *target = t->u.kernel.target;
net/netfilter/x_tables.c
309
U->u.user.name, K->u.kernel.TYPE->name, \
net/netfilter/x_tables.c
310
&U->u.user.revision, K->u.kernel.TYPE->revision)
net/netfilter/x_tables.c
328
K->u.kernel.TYPE->usersize, \
net/netfilter/x_tables.c
329
K->u.kernel.TYPE->TYPE##size, \
net/netfilter/x_tables.c
330
XT_ALIGN(K->u.kernel.TYPE->TYPE##size))
net/netfilter/x_tables.c
765
const struct xt_match *match = m->u.kernel.match;
net/netfilter/x_tables.c
791
K->u.kernel.TYPE->usersize, \
net/netfilter/x_tables.c
798
const struct xt_match *match = m->u.kernel.match;
net/netfilter/xt_TCPMSS.c
255
if (strcmp(m->u.kernel.match->name, "tcp") == 0 &&
net/rxrpc/af_rxrpc.c
373
p.kernel = true;
net/rxrpc/ar-internal.h
899
bool kernel; /* T if kernel is making the call */
net/rxrpc/call_object.c
214
if (p->kernel)
net/rxrpc/call_object.c
265
if (p->kernel)
security/security.c
5175
int security_bpf(int cmd, union bpf_attr *attr, unsigned int size, bool kernel)
security/security.c
5177
return call_int_hook(bpf, cmd, attr, size, kernel);
security/security.c
5222
struct bpf_token *token, bool kernel)
security/security.c
5230
rc = call_int_hook(bpf_map_create, map, attr, token, kernel);
security/security.c
5250
struct bpf_token *token, bool kernel)
security/security.c
5258
rc = call_int_hook(bpf_prog_load, prog, attr, token, kernel);
security/selinux/hooks.c
7046
unsigned int size, bool kernel)
security/selinux/hooks.c
7154
struct bpf_token *token, bool kernel)
security/selinux/hooks.c
7172
struct bpf_token *token, bool kernel)
sound/core/seq/oss/seq_oss_init.c
313
port.kernel = &callback;
sound/core/seq/oss/seq_oss_init.c
92
port->kernel = &port_callback;
sound/core/seq/seq_clientmgr.c
1199
info->card = cptr->data.kernel.card ? cptr->data.kernel.card->number : -1;
sound/core/seq/seq_clientmgr.c
1278
if (client->type == USER_CLIENT && info->kernel)
sound/core/seq/seq_clientmgr.c
1295
callback = info->kernel;
sound/core/seq/seq_clientmgr.c
2304
client->data.kernel.card = card;
sound/core/seq/seq_clientmgr.h
60
struct snd_seq_kernel_client kernel;
sound/core/seq/seq_compat.c
25
u32 kernel; /* reserved for kernel use (must be NULL) */
sound/core/seq/seq_compat.c
45
data->kernel = NULL;
sound/core/seq/seq_dummy.c
145
pinfo.kernel = &pcb;
sound/core/seq/seq_midi.c
386
port->kernel = &pcallbacks;
sound/core/seq/seq_ports.c
686
portinfo.kernel = pcbp;
sound/core/seq/seq_system.c
152
port->kernel = &pcallbacks;
sound/core/seq/seq_system.c
168
port->kernel = &pcallbacks;
sound/core/seq/seq_ump_client.c
237
port->kernel = &pcallbacks;
sound/core/seq/seq_ump_client.c
326
port->kernel = &pcallbacks;
sound/core/seq/seq_virmidi.c
402
pinfo->kernel = &pcallbacks;
tools/perf/arch/x86/util/machine.c
71
struct dso *kernel)
tools/perf/arch/x86/util/machine.c
94
ret = machine__create_extra_kernel_map(machine, kernel, xm);
tools/perf/builtin-script.c
1103
bool kernel;
tools/perf/builtin-script.c
1110
kernel = machine__kernel_ip(machine, start);
tools/perf/builtin-script.c
1111
if (kernel)
tools/perf/builtin-script.c
1122
if (kernel != machine__kernel_ip(machine, end)) {
tools/perf/util/dso.c
1647
dso->kernel = DSO_SPACE__USER;
tools/perf/util/dso.h
293
enum dso_space_type kernel:2;
tools/perf/util/dso.h
533
return RC_CHK_ACCESS(dso)->kernel;
tools/perf/util/dso.h
536
static inline void dso__set_kernel(struct dso *dso, enum dso_space_type kernel)
tools/perf/util/dso.h
538
RC_CHK_ACCESS(dso)->kernel = kernel;
tools/perf/util/evsel.c
717
MOD_PRINT(kernel, 'k');
tools/perf/util/machine.c
1039
struct dso *kernel,
tools/perf/util/machine.c
1046
map = map__new2(xm->start, kernel);
tools/perf/util/machine.c
1125
struct dso *kernel)
tools/perf/util/machine.c
1143
pgoff = find_entry_trampoline(kernel);
tools/perf/util/machine.c
1162
if (machine__create_extra_kernel_map(machine, kernel, &xm) < 0)
tools/perf/util/machine.c
1172
struct dso *kernel __maybe_unused)
tools/perf/util/machine.c
1178
__machine__create_kernel_maps(struct machine *machine, struct dso *kernel)
tools/perf/util/machine.c
1184
machine->vmlinux_map = map__new2(0, kernel);
tools/perf/util/machine.c
1545
struct dso *kernel = machine__get_kernel(machine);
tools/perf/util/machine.c
1550
if (kernel == NULL)
tools/perf/util/machine.c
1553
ret = __machine__create_kernel_maps(machine, kernel);
tools/perf/util/machine.c
1583
if (machine__create_extra_kernel_maps(machine, kernel))
tools/perf/util/machine.c
1600
dso__put(kernel);
tools/perf/util/machine.c
1624
struct dso *kernel = machine__kernel_dso(machine);
tools/perf/util/machine.c
1626
if (kernel == NULL)
tools/perf/util/machine.c
1629
return machine__create_extra_kernel_map(machine, kernel, xm);
tools/perf/util/machine.c
1678
struct dso *kernel = dsos__find_kernel_dso(&machine->dsos);
tools/perf/util/machine.c
1680
if (kernel == NULL)
tools/perf/util/machine.c
1681
kernel = machine__findnew_dso(machine, machine->mmap_name);
tools/perf/util/machine.c
1682
if (kernel == NULL)
tools/perf/util/machine.c
1685
dso__set_kernel(kernel, dso_space);
tools/perf/util/machine.c
1686
if (__machine__create_kernel_maps(machine, kernel) < 0) {
tools/perf/util/machine.c
1687
dso__put(kernel);
tools/perf/util/machine.c
1691
if (strstr(dso__long_name(kernel), "vmlinux"))
tools/perf/util/machine.c
1692
dso__set_short_name(kernel, "[kernel.vmlinux]", false);
tools/perf/util/machine.c
1695
dso__put(kernel);
tools/perf/util/machine.c
1700
dso__set_build_id(kernel, bid);
tools/perf/util/machine.c
1717
dso__load(kernel, machine__kernel_map(machine));
tools/perf/util/machine.c
1719
dso__put(kernel);
tools/perf/util/machine.c
963
struct dso *kernel;
tools/perf/util/machine.c
969
kernel = machine__findnew_kernel(machine, vmlinux_name,
tools/perf/util/machine.c
975
kernel = machine__findnew_kernel(machine, vmlinux_name,
tools/perf/util/machine.c
980
if (kernel != NULL && (!dso__has_build_id(kernel)))
tools/perf/util/machine.c
981
dso__read_running_kernel_build_id(kernel, machine);
tools/perf/util/machine.c
983
return kernel;
tools/perf/util/machine.h
314
struct dso *kernel);
tools/perf/util/machine.h
325
struct dso *kernel,
tools/perf/util/machine.h
329
struct dso *kernel);
tools/perf/util/parse-events.c
1821
if (mod.kernel) {
tools/perf/util/parse-events.h
208
bool kernel : 1; /* 'k' */
tools/perf/util/symbol.c
361
struct symbol *sym, bool kernel)
tools/perf/util/symbol.c
369
if (kernel) {
tools/perf/util/symbol.h
188
bool kernel);
tools/perf/util/synthetic-events.c
147
*kernel = true;
tools/perf/util/synthetic-events.c
149
*kernel = false;
tools/perf/util/synthetic-events.c
156
pid_t *tgid, pid_t *ppid, bool *kernel)
tools/perf/util/synthetic-events.c
167
tgid, ppid, kernel) != 0) {
tools/perf/util/synthetic-events.c
74
pid_t *tgid, pid_t *ppid, bool *kernel)
tools/testing/selftests/bpf/prog_tests/kprobe_multi_test.c
455
static void test_kprobe_multi_bench_attach(bool kernel)
tools/testing/selftests/bpf/prog_tests/kprobe_multi_test.c
461
if (!ASSERT_OK(bpf_get_ksyms(&ksyms, kernel), "bpf_get_ksyms"))
tools/testing/selftests/bpf/prog_tests/kprobe_multi_test.c
478
static void test_kprobe_multi_bench_attach_addr(bool kernel)
tools/testing/selftests/bpf/prog_tests/kprobe_multi_test.c
486
err = bpf_get_addrs(&addrs, &cnt, kernel);
tools/testing/selftests/bpf/progs/rcu_read_lock.c
257
bool kernel)
tools/testing/selftests/bpf/progs/test_cgroup1_hierarchy.c
54
int BPF_PROG(lsm_run, int cmd, union bpf_attr *attr, unsigned int size, bool kernel)
tools/testing/selftests/bpf/progs/test_cgroup1_hierarchy.c
60
int BPF_PROG(lsm_s_run, int cmd, union bpf_attr *attr, unsigned int size, bool kernel)
tools/testing/selftests/bpf/progs/test_kernel_flag.c
19
int BPF_PROG(bpf, int cmd, union bpf_attr *attr, unsigned int size, bool kernel)
tools/testing/selftests/bpf/progs/test_kernel_flag.c
24
if (!kernel || tid != monitored_tid)
tools/testing/selftests/bpf/progs/test_kfunc_dynptr_param.c
39
int BPF_PROG(not_valid_dynptr, int cmd, union bpf_attr *attr, unsigned int size, bool kernel)
tools/testing/selftests/bpf/progs/test_kfunc_dynptr_param.c
49
int BPF_PROG(not_ptr_to_stack, int cmd, union bpf_attr *attr, unsigned int size, bool kernel)
tools/testing/selftests/bpf/progs/test_kfunc_dynptr_param.c
57
int BPF_PROG(dynptr_data_null, int cmd, union bpf_attr *attr, unsigned int size, bool kernel)
tools/testing/selftests/bpf/progs/test_lookup_key.c
26
int BPF_PROG(bpf, int cmd, union bpf_attr *attr, unsigned int size, bool kernel)
tools/testing/selftests/bpf/progs/test_ptr_untrusted.c
10
int BPF_PROG(lsm_run, int cmd, union bpf_attr *attr, unsigned int size, bool kernel)
tools/testing/selftests/bpf/progs/test_task_under_cgroup.c
52
int BPF_PROG(lsm_run, int cmd, union bpf_attr *attr, unsigned int size, bool kernel)
tools/testing/selftests/bpf/progs/test_verify_pkcs7_sig.c
40
int BPF_PROG(bpf, int cmd, union bpf_attr *attr, unsigned int size, bool kernel)
tools/testing/selftests/bpf/trace_helpers.c
531
static bool is_invalid_entry(char *buf, bool kernel)
tools/testing/selftests/bpf/trace_helpers.c
533
if (kernel && strchr(buf, '['))
tools/testing/selftests/bpf/trace_helpers.c
535
if (!kernel && !strchr(buf, '['))
tools/testing/selftests/bpf/trace_helpers.c
609
int bpf_get_ksyms(struct ksyms **ksymsp, bool kernel)
tools/testing/selftests/bpf/trace_helpers.c
648
if (is_invalid_entry(buf, kernel))
tools/testing/selftests/bpf/trace_helpers.c
695
int bpf_get_addrs(unsigned long **addrsp, size_t *cntp, bool kernel)
tools/testing/selftests/bpf/trace_helpers.c
725
if (is_invalid_entry(buf, kernel))
tools/testing/selftests/bpf/trace_helpers.h
63
int bpf_get_ksyms(struct ksyms **ksymsp, bool kernel);
tools/testing/selftests/bpf/trace_helpers.h
64
int bpf_get_addrs(unsigned long **addrsp, size_t *cntp, bool kernel);
tools/testing/selftests/cgroup/test_kmem.c
170
long current, anon, file, kernel, sock, sum;
tools/testing/selftests/cgroup/test_kmem.c
190
kernel = cg_read_key_long(parent, "memory.stat", "kernel ");
tools/testing/selftests/cgroup/test_kmem.c
192
if (current < 0 || anon < 0 || file < 0 || kernel < 0 || sock < 0)
tools/testing/selftests/cgroup/test_kmem.c
195
sum = anon + file + kernel + sock;
tools/testing/selftests/cgroup/test_kmem.c
203
printf("kernel = %ld\n", kernel);