Symbol: e
arch/alpha/include/asm/pgtable.h
355
#define pte_ERROR(e) \
arch/alpha/include/asm/pgtable.h
356
printk("%s:%d: bad pte %016lx.\n", __FILE__, __LINE__, pte_val(e))
arch/alpha/include/asm/pgtable.h
357
#define pmd_ERROR(e) \
arch/alpha/include/asm/pgtable.h
358
printk("%s:%d: bad pmd %016lx.\n", __FILE__, __LINE__, pmd_val(e))
arch/alpha/include/asm/pgtable.h
359
#define pgd_ERROR(e) \
arch/alpha/include/asm/pgtable.h
360
printk("%s:%d: bad pgd %016lx.\n", __FILE__, __LINE__, pgd_val(e))
arch/arc/include/asm/disasm.h
32
#define BITS(word, s, e) (((word) >> (s)) & (~((-2) << ((e) - (s)))))
arch/arc/include/asm/linkage.h
17
.macro ST2 e, o, off
arch/arc/include/asm/linkage.h
26
.macro LD2 e, o, off
arch/arc/include/asm/pgtable-levels.h
101
#define pgd_ERROR(e) \
arch/arc/include/asm/pgtable-levels.h
102
pr_crit("%s:%d: bad pgd %08lx.\n", __FILE__, __LINE__, pgd_val(e))
arch/arc/include/asm/pgtable-levels.h
118
#define pud_ERROR(e) \
arch/arc/include/asm/pgtable-levels.h
119
pr_crit("%s:%d: bad pud %08lx.\n", __FILE__, __LINE__, pud_val(e))
arch/arc/include/asm/pgtable-levels.h
140
#define pmd_ERROR(e) \
arch/arc/include/asm/pgtable-levels.h
141
pr_crit("%s:%d: bad pmd %08lx.\n", __FILE__, __LINE__, pmd_val(e))
arch/arc/include/asm/pgtable-levels.h
168
#define pte_ERROR(e) \
arch/arc/include/asm/pgtable-levels.h
169
pr_crit("%s:%d: bad pte %08lx.\n", __FILE__, __LINE__, pte_val(e))
arch/arc/include/asm/tlbflush.h
23
#define flush_tlb_range(vma, s, e) local_flush_tlb_range(vma, s, e)
arch/arc/include/asm/tlbflush.h
25
#define flush_tlb_kernel_range(s, e) local_flush_tlb_kernel_range(s, e)
arch/arc/include/asm/tlbflush.h
29
#define flush_pmd_tlb_range(vma, s, e) local_flush_pmd_tlb_range(vma, s, e)
arch/arc/mm/cache.c
305
unsigned int s, e;
arch/arc/mm/cache.c
310
e = ARC_REG_IC_ENDR;
arch/arc/mm/cache.c
313
e = ARC_REG_DC_ENDR;
arch/arc/mm/cache.c
337
write_aux_reg(e, paddr + sz); /* ENDR is exclusive */
arch/arm/include/asm/cacheflush.h
268
#define flush_icache_user_range(s,e) __cpuc_coherent_user_range(s,e)
arch/arm/include/asm/cacheflush.h
274
#define flush_icache_range(s,e) __cpuc_coherent_kern_range(s,e)
arch/arm/include/asm/tlbflush.h
598
#define local_flush_tlb_kernel_range(s,e) __cpu_flush_kern_tlb_range(s,e)
arch/arm/mm/dma-mapping.c
145
struct page *page, *p, *e;
arch/arm/mm/dma-mapping.c
155
for (p = page + (size >> PAGE_SHIFT), e = page + (1 << order); p < e; p++)
arch/arm/mm/dma-mapping.c
168
struct page *e = page + (size >> PAGE_SHIFT);
arch/arm/mm/dma-mapping.c
170
while (page < e) {
arch/arm/probes/kprobes/test-arm.c
1252
COPROCESSOR_INSTRUCTIONS_ST_LD("",e)
arch/arm/probes/kprobes/test-arm.c
1254
COPROCESSOR_INSTRUCTIONS_MC_MR("",e)
arch/arm64/include/asm/esr.h
292
#define esr_sys64_to_sysreg(e) \
arch/arm64/include/asm/esr.h
293
sys_reg((((e) & ESR_ELx_SYS64_ISS_OP0_MASK) >> \
arch/arm64/include/asm/esr.h
295
(((e) & ESR_ELx_SYS64_ISS_OP1_MASK) >> \
arch/arm64/include/asm/esr.h
297
(((e) & ESR_ELx_SYS64_ISS_CRN_MASK) >> \
arch/arm64/include/asm/esr.h
299
(((e) & ESR_ELx_SYS64_ISS_CRM_MASK) >> \
arch/arm64/include/asm/esr.h
301
(((e) & ESR_ELx_SYS64_ISS_OP2_MASK) >> \
arch/arm64/include/asm/esr.h
304
#define esr_cp15_to_sysreg(e) \
arch/arm64/include/asm/esr.h
306
(((e) & ESR_ELx_SYS64_ISS_OP1_MASK) >> \
arch/arm64/include/asm/esr.h
308
(((e) & ESR_ELx_SYS64_ISS_CRN_MASK) >> \
arch/arm64/include/asm/esr.h
310
(((e) & ESR_ELx_SYS64_ISS_CRM_MASK) >> \
arch/arm64/include/asm/esr.h
312
(((e) & ESR_ELx_SYS64_ISS_OP2_MASK) >> \
arch/arm64/include/asm/kvm_emulate.h
607
#define kvm_pend_exception(v, e) \
arch/arm64/include/asm/kvm_emulate.h
611
vcpu_set_flag((v), e); \
arch/arm64/include/asm/pgtable.h
1086
#define p4d_ERROR(e) \
arch/arm64/include/asm/pgtable.h
1087
pr_err("%s:%d: bad p4d %016llx.\n", __FILE__, __LINE__, p4d_val(e))
arch/arm64/include/asm/pgtable.h
116
#define pte_ERROR(e) \
arch/arm64/include/asm/pgtable.h
117
pr_err("%s:%d: bad pte %016llx.\n", __FILE__, __LINE__, pte_val(e))
arch/arm64/include/asm/pgtable.h
1215
#define pgd_ERROR(e) \
arch/arm64/include/asm/pgtable.h
1216
pr_err("%s:%d: bad pgd %016llx.\n", __FILE__, __LINE__, pgd_val(e))
arch/arm64/include/asm/pgtable.h
868
#define pmd_ERROR(e) \
arch/arm64/include/asm/pgtable.h
869
pr_err("%s:%d: bad pmd %016llx.\n", __FILE__, __LINE__, pmd_val(e))
arch/arm64/include/asm/pgtable.h
958
#define pud_ERROR(e) \
arch/arm64/include/asm/pgtable.h
959
pr_err("%s:%d: bad pud %016llx.\n", __FILE__, __LINE__, pud_val(e))
arch/arm64/kvm/vgic/vgic-irqfd.c
105
int kvm_arch_set_irq_inatomic(struct kvm_kernel_irq_routing_entry *e,
arch/arm64/kvm/vgic/vgic-irqfd.c
112
switch (e->type) {
arch/arm64/kvm/vgic/vgic-irqfd.c
119
kvm_populate_msi(e, &msi);
arch/arm64/kvm/vgic/vgic-irqfd.c
130
return vgic_irqfd_set_irq(e, kvm, irq_source_id, 1, line_status);
arch/arm64/kvm/vgic/vgic-irqfd.c
18
static int vgic_irqfd_set_irq(struct kvm_kernel_irq_routing_entry *e,
arch/arm64/kvm/vgic/vgic-irqfd.c
22
unsigned int spi_id = e->irqchip.pin + VGIC_NR_PRIVATE_IRQS;
arch/arm64/kvm/vgic/vgic-irqfd.c
39
struct kvm_kernel_irq_routing_entry *e,
arch/arm64/kvm/vgic/vgic-irqfd.c
46
e->set = vgic_irqfd_set_irq;
arch/arm64/kvm/vgic/vgic-irqfd.c
47
e->irqchip.irqchip = ue->u.irqchip.irqchip;
arch/arm64/kvm/vgic/vgic-irqfd.c
48
e->irqchip.pin = ue->u.irqchip.pin;
arch/arm64/kvm/vgic/vgic-irqfd.c
49
if ((e->irqchip.pin >= KVM_IRQCHIP_NUM_PINS) ||
arch/arm64/kvm/vgic/vgic-irqfd.c
50
(e->irqchip.irqchip >= KVM_NR_IRQCHIPS))
arch/arm64/kvm/vgic/vgic-irqfd.c
54
e->set = kvm_set_msi;
arch/arm64/kvm/vgic/vgic-irqfd.c
55
e->msi.address_lo = ue->u.msi.address_lo;
arch/arm64/kvm/vgic/vgic-irqfd.c
56
e->msi.address_hi = ue->u.msi.address_hi;
arch/arm64/kvm/vgic/vgic-irqfd.c
57
e->msi.data = ue->u.msi.data;
arch/arm64/kvm/vgic/vgic-irqfd.c
58
e->msi.flags = ue->flags;
arch/arm64/kvm/vgic/vgic-irqfd.c
59
e->msi.devid = ue->u.msi.devid;
arch/arm64/kvm/vgic/vgic-irqfd.c
69
static void kvm_populate_msi(struct kvm_kernel_irq_routing_entry *e,
arch/arm64/kvm/vgic/vgic-irqfd.c
72
msi->address_lo = e->msi.address_lo;
arch/arm64/kvm/vgic/vgic-irqfd.c
73
msi->address_hi = e->msi.address_hi;
arch/arm64/kvm/vgic/vgic-irqfd.c
74
msi->data = e->msi.data;
arch/arm64/kvm/vgic/vgic-irqfd.c
75
msi->flags = e->msi.flags;
arch/arm64/kvm/vgic/vgic-irqfd.c
76
msi->devid = e->msi.devid;
arch/arm64/kvm/vgic/vgic-irqfd.c
86
int kvm_set_msi(struct kvm_kernel_irq_routing_entry *e,
arch/arm64/kvm/vgic/vgic-irqfd.c
98
kvm_populate_msi(e, &msi);
arch/csky/include/asm/futex.h
111
: [ov] "Jr" (oldval), [nv] "Jr" (newval), [e] "Jr" (-EFAULT)
arch/csky/include/asm/futex.h
35
: [op] "Jr" (oparg), [e] "jr" (-EFAULT) \
arch/csky/include/asm/pgtable.h
26
#define pte_ERROR(e) \
arch/csky/include/asm/pgtable.h
27
pr_err("%s:%d: bad pte %08lx.\n", __FILE__, __LINE__, (e).pte_low)
arch/csky/include/asm/pgtable.h
28
#define pgd_ERROR(e) \
arch/csky/include/asm/pgtable.h
29
pr_err("%s:%d: bad pgd %08lx.\n", __FILE__, __LINE__, pgd_val(e))
arch/hexagon/include/asm/pgtable.h
100
#define pgd_ERROR(e) \
arch/hexagon/include/asm/pgtable.h
102
pgd_val(e))
arch/loongarch/include/asm/loongarch.h
298
#define CSR_INVALID_ENTRY(e) (CSR_TLBIDX_EHINV | e)
arch/loongarch/include/asm/pgtable.h
127
#define pte_ERROR(e) \
arch/loongarch/include/asm/pgtable.h
128
pr_err("%s:%d: bad pte %016lx.\n", __FILE__, __LINE__, pte_val(e))
arch/loongarch/include/asm/pgtable.h
130
#define pmd_ERROR(e) \
arch/loongarch/include/asm/pgtable.h
131
pr_err("%s:%d: bad pmd %016lx.\n", __FILE__, __LINE__, pmd_val(e))
arch/loongarch/include/asm/pgtable.h
134
#define pud_ERROR(e) \
arch/loongarch/include/asm/pgtable.h
135
pr_err("%s:%d: bad pud %016lx.\n", __FILE__, __LINE__, pud_val(e))
arch/loongarch/include/asm/pgtable.h
137
#define pgd_ERROR(e) \
arch/loongarch/include/asm/pgtable.h
138
pr_err("%s:%d: bad pgd %016lx.\n", __FILE__, __LINE__, pgd_val(e))
arch/loongarch/kvm/irqfd.c
10
static int kvm_set_pic_irq(struct kvm_kernel_irq_routing_entry *e,
arch/loongarch/kvm/irqfd.c
14
pch_pic_set_irq(kvm->arch.pch_pic, e->irqchip.pin, level);
arch/loongarch/kvm/irqfd.c
26
int kvm_set_msi(struct kvm_kernel_irq_routing_entry *e,
arch/loongarch/kvm/irqfd.c
32
pch_msi_set_irq(kvm, e->msi.data, level);
arch/loongarch/kvm/irqfd.c
47
struct kvm_kernel_irq_routing_entry *e,
arch/loongarch/kvm/irqfd.c
52
e->set = kvm_set_pic_irq;
arch/loongarch/kvm/irqfd.c
53
e->irqchip.irqchip = ue->u.irqchip.irqchip;
arch/loongarch/kvm/irqfd.c
54
e->irqchip.pin = ue->u.irqchip.pin;
arch/loongarch/kvm/irqfd.c
56
if (e->irqchip.pin >= KVM_IRQCHIP_NUM_PINS)
arch/loongarch/kvm/irqfd.c
61
e->set = kvm_set_msi;
arch/loongarch/kvm/irqfd.c
62
e->msi.address_lo = ue->u.msi.address_lo;
arch/loongarch/kvm/irqfd.c
63
e->msi.address_hi = ue->u.msi.address_hi;
arch/loongarch/kvm/irqfd.c
64
e->msi.data = ue->u.msi.data;
arch/loongarch/kvm/irqfd.c
71
int kvm_arch_set_irq_inatomic(struct kvm_kernel_irq_routing_entry *e,
arch/loongarch/kvm/irqfd.c
74
switch (e->type) {
arch/loongarch/kvm/irqfd.c
76
pch_pic_set_irq(kvm->arch.pch_pic, e->irqchip.pin, level);
arch/loongarch/kvm/irqfd.c
79
pch_msi_set_irq(kvm, e->msi.data, level);
arch/m68k/include/asm/bootstd.h
116
#define _bsc5(type,name,atype,a,btype,b,ctype,c,dtype,d,etype,e) \
arch/m68k/include/asm/bootstd.h
117
type name(atype a, btype b, ctype c, dtype d, etype e) \
arch/m68k/include/asm/bootstd.h
124
register long __e __asm__ ("%d5") = (long)e; \
arch/m68k/include/asm/mcf_pgtable.h
140
#define pte_ERROR(e) \
arch/m68k/include/asm/mcf_pgtable.h
142
__FILE__, __LINE__, pte_val(e))
arch/m68k/include/asm/mcf_pgtable.h
143
#define pgd_ERROR(e) \
arch/m68k/include/asm/mcf_pgtable.h
145
__FILE__, __LINE__, pgd_val(e))
arch/m68k/include/asm/motorola_pgtable.h
134
#define pte_ERROR(e) \
arch/m68k/include/asm/motorola_pgtable.h
135
printk("%s:%d: bad pte %08lx.\n", __FILE__, __LINE__, pte_val(e))
arch/m68k/include/asm/motorola_pgtable.h
136
#define pmd_ERROR(e) \
arch/m68k/include/asm/motorola_pgtable.h
137
printk("%s:%d: bad pmd %08lx.\n", __FILE__, __LINE__, pmd_val(e))
arch/m68k/include/asm/motorola_pgtable.h
138
#define pgd_ERROR(e) \
arch/m68k/include/asm/motorola_pgtable.h
139
printk("%s:%d: bad pgd %08lx.\n", __FILE__, __LINE__, pgd_val(e))
arch/m68k/include/asm/sun3_pgtable.h
123
#define pte_ERROR(e) \
arch/m68k/include/asm/sun3_pgtable.h
124
pr_err("%s:%d: bad pte %08lx.\n", __FILE__, __LINE__, pte_val(e))
arch/m68k/include/asm/sun3_pgtable.h
125
#define pgd_ERROR(e) \
arch/m68k/include/asm/sun3_pgtable.h
126
pr_err("%s:%d: bad pgd %08lx.\n", __FILE__, __LINE__, pgd_val(e))
arch/microblaze/include/asm/pgtable.h
106
#define pte_ERROR(e) \
arch/microblaze/include/asm/pgtable.h
108
__FILE__, __LINE__, pte_val(e))
arch/microblaze/include/asm/pgtable.h
109
#define pgd_ERROR(e) \
arch/microblaze/include/asm/pgtable.h
111
__FILE__, __LINE__, pgd_val(e))
arch/mips/boot/elf2ecoff.c
169
static void convert_elf_hdr(Elf32_Ehdr * e)
arch/mips/boot/elf2ecoff.c
171
e->e_type = swab16(e->e_type);
arch/mips/boot/elf2ecoff.c
172
e->e_machine = swab16(e->e_machine);
arch/mips/boot/elf2ecoff.c
173
e->e_version = swab32(e->e_version);
arch/mips/boot/elf2ecoff.c
174
e->e_entry = swab32(e->e_entry);
arch/mips/boot/elf2ecoff.c
175
e->e_phoff = swab32(e->e_phoff);
arch/mips/boot/elf2ecoff.c
176
e->e_shoff = swab32(e->e_shoff);
arch/mips/boot/elf2ecoff.c
177
e->e_flags = swab32(e->e_flags);
arch/mips/boot/elf2ecoff.c
178
e->e_ehsize = swab16(e->e_ehsize);
arch/mips/boot/elf2ecoff.c
179
e->e_phentsize = swab16(e->e_phentsize);
arch/mips/boot/elf2ecoff.c
180
e->e_phnum = swab16(e->e_phnum);
arch/mips/boot/elf2ecoff.c
181
e->e_shentsize = swab16(e->e_shentsize);
arch/mips/boot/elf2ecoff.c
182
e->e_shnum = swab16(e->e_shnum);
arch/mips/boot/elf2ecoff.c
183
e->e_shstrndx = swab16(e->e_shstrndx);
arch/mips/include/asm/pgtable-32.h
108
#define pte_ERROR(e) \
arch/mips/include/asm/pgtable-32.h
109
printk("%s:%d: bad pte %016Lx.\n", __FILE__, __LINE__, pte_val(e))
arch/mips/include/asm/pgtable-32.h
111
#define pte_ERROR(e) \
arch/mips/include/asm/pgtable-32.h
112
printk("%s:%d: bad pte %08lx.\n", __FILE__, __LINE__, pte_val(e))
arch/mips/include/asm/pgtable-32.h
114
#define pgd_ERROR(e) \
arch/mips/include/asm/pgtable-32.h
115
printk("%s:%d: bad pgd %08lx.\n", __FILE__, __LINE__, pgd_val(e))
arch/mips/include/asm/pgtable-64.h
154
#define pte_ERROR(e) \
arch/mips/include/asm/pgtable-64.h
155
printk("%s:%d: bad pte %016lx.\n", __FILE__, __LINE__, pte_val(e))
arch/mips/include/asm/pgtable-64.h
157
#define pmd_ERROR(e) \
arch/mips/include/asm/pgtable-64.h
158
printk("%s:%d: bad pmd %016lx.\n", __FILE__, __LINE__, pmd_val(e))
arch/mips/include/asm/pgtable-64.h
161
#define pud_ERROR(e) \
arch/mips/include/asm/pgtable-64.h
162
printk("%s:%d: bad pud %016lx.\n", __FILE__, __LINE__, pud_val(e))
arch/mips/include/asm/pgtable-64.h
164
#define pgd_ERROR(e) \
arch/mips/include/asm/pgtable-64.h
165
printk("%s:%d: bad pgd %016lx.\n", __FILE__, __LINE__, pgd_val(e))
arch/mips/kernel/jump_label.c
38
void arch_jump_label_transform(struct jump_entry *e,
arch/mips/kernel/jump_label.c
45
insn_p = (union mips_instruction *)msk_isa16_mode(e->code);
arch/mips/kernel/jump_label.c
48
BUG_ON((e->target & J_ALIGN_MASK) != J_ISA_BIT);
arch/mips/kernel/jump_label.c
52
offset = e->target - ((unsigned long)insn_p + 4);
arch/mips/kernel/jump_label.c
69
WARN_ON((e->target & ~J_RANGE_MASK) !=
arch/mips/kernel/jump_label.c
70
((e->code + 4) & ~J_RANGE_MASK));
arch/mips/kernel/jump_label.c
73
insn.j_format.target = e->target >> J_RANGE_SHIFT;
arch/mips/kernel/module.c
396
const struct exception_table_entry *e = NULL;
arch/mips/kernel/module.c
401
e = search_extable(dbe->dbe_start,
arch/mips/kernel/module.c
403
if (e)
arch/mips/kernel/module.c
410
return e;
arch/mips/kernel/relocate.c
181
unsigned long *etable_start, *etable_end, *e;
arch/mips/kernel/relocate.c
186
for (e = etable_start; e < etable_end; e++)
arch/mips/kernel/relocate.c
187
*e += offset;
arch/mips/kernel/traps.c
440
const struct exception_table_entry *e;
arch/mips/kernel/traps.c
442
e = search_extable(__start___dbe_table,
arch/mips/kernel/traps.c
444
if (!e)
arch/mips/kernel/traps.c
445
e = search_module_dbetables(addr);
arch/mips/kernel/traps.c
446
return e;
arch/mips/math-emu/ieee754dp.h
50
#define DPDNORMx(m,e) \
arch/mips/math-emu/ieee754dp.h
51
while ((m >> DP_FBITS) == 0) { m <<= 1; e--; }
arch/mips/math-emu/ieee754sp.h
54
#define SPDNORMx(m,e) \
arch/mips/math-emu/ieee754sp.h
55
while ((m >> SP_FBITS) == 0) { m <<= 1; e--; }
arch/mips/mm/uasm-micromips.c
32
#define M(a, b, c, d, e, f) \
arch/mips/mm/uasm-micromips.c
37
| (e) << RE_SH \
arch/mips/mm/uasm-mips.c
32
#define M(a, b, c, d, e, f) \
arch/mips/mm/uasm-mips.c
37
| (e) << RE_SH \
arch/mips/mm/uasm-mips.c
41
#define M6(a, b, c, d, e) \
arch/mips/mm/uasm-mips.c
46
| (e) << FUNC_SH)
arch/mips/sgi-ip27/ip27-init.c
107
e = LOCAL_HUB_L(PI_CPU_ENABLE_A) & 1;
arch/mips/sgi-ip27/ip27-init.c
110
e ? ", CPU is running" : "");
arch/mips/sgi-ip27/ip27-init.c
113
e = LOCAL_HUB_L(PI_CPU_ENABLE_B) & 1;
arch/mips/sgi-ip27/ip27-init.c
116
e ? ", CPU is running" : "");
arch/mips/sgi-ip27/ip27-init.c
93
u64 p, e, n_mode;
arch/nios2/include/asm/pgtable.h
233
#define pte_ERROR(e) \
arch/nios2/include/asm/pgtable.h
235
__FILE__, __LINE__, pte_val(e))
arch/nios2/include/asm/pgtable.h
236
#define pgd_ERROR(e) \
arch/nios2/include/asm/pgtable.h
238
__FILE__, __LINE__, pgd_val(e))
arch/openrisc/include/asm/pgtable.h
345
#define pte_ERROR(e) \
arch/openrisc/include/asm/pgtable.h
347
__FILE__, __LINE__, &(e), pte_val(e))
arch/openrisc/include/asm/pgtable.h
348
#define pgd_ERROR(e) \
arch/openrisc/include/asm/pgtable.h
350
__FILE__, __LINE__, &(e), pgd_val(e))
arch/openrisc/mm/init.c
102
for (j = 0; p < e && j < PTRS_PER_PTE;
arch/openrisc/mm/init.c
61
unsigned long v, p, e;
arch/openrisc/mm/init.c
77
e = (u32) end;
arch/openrisc/mm/init.c
82
while (p < e) {
arch/parisc/include/asm/cacheflush.h
65
#define flush_icache_range(s,e) do { \
arch/parisc/include/asm/cacheflush.h
66
flush_kernel_dcache_range_asm(s,e); \
arch/parisc/include/asm/cacheflush.h
67
flush_kernel_icache_range_asm(s,e); \
arch/parisc/include/asm/pgtable.h
78
#define pte_ERROR(e) \
arch/parisc/include/asm/pgtable.h
79
printk("%s:%d: bad pte %08lx.\n", __FILE__, __LINE__, pte_val(e))
arch/parisc/include/asm/pgtable.h
81
#define pmd_ERROR(e) \
arch/parisc/include/asm/pgtable.h
82
printk("%s:%d: bad pmd %08lx.\n", __FILE__, __LINE__, (unsigned long)pmd_val(e))
arch/parisc/include/asm/pgtable.h
84
#define pgd_ERROR(e) \
arch/parisc/include/asm/pgtable.h
85
printk("%s:%d: bad pgd %08lx.\n", __FILE__, __LINE__, (unsigned long)pgd_val(e))
arch/parisc/include/asm/psw.h
71
unsigned int e:1;
arch/parisc/kernel/unwind.c
101
return e;
arch/parisc/kernel/unwind.c
156
struct unwind_table_entry *e = (struct unwind_table_entry *)end;
arch/parisc/kernel/unwind.c
158
unwind_table_sort(s, e);
arch/parisc/kernel/unwind.c
281
const struct unwind_table_entry *e;
arch/parisc/kernel/unwind.c
287
e = find_unwind_entry(info->ip);
arch/parisc/kernel/unwind.c
288
if (e == NULL) {
arch/parisc/kernel/unwind.c
330
e->region_start, e->region_end, e->Save_SP, e->Save_RP,
arch/parisc/kernel/unwind.c
331
e->Millicode, e->Total_frame_size);
arch/parisc/kernel/unwind.c
333
looking_for_rp = e->Save_RP;
arch/parisc/kernel/unwind.c
335
for (npc = e->region_start;
arch/parisc/kernel/unwind.c
336
(frame_size < (e->Total_frame_size << 3) ||
arch/parisc/kernel/unwind.c
371
if (frame_size > e->Total_frame_size << 3)
arch/parisc/kernel/unwind.c
372
frame_size = e->Total_frame_size << 3;
arch/parisc/kernel/unwind.c
374
if (!unwind_special(info, e->region_start, frame_size)) {
arch/parisc/kernel/unwind.c
376
if (e->Millicode)
arch/parisc/kernel/unwind.c
55
const struct unwind_table_entry *e = NULL;
arch/parisc/kernel/unwind.c
63
e = &table->table[mid];
arch/parisc/kernel/unwind.c
64
if (addr < e->region_start)
arch/parisc/kernel/unwind.c
66
else if (addr > e->region_end)
arch/parisc/kernel/unwind.c
69
return e;
arch/parisc/kernel/unwind.c
79
const struct unwind_table_entry *e = NULL;
arch/parisc/kernel/unwind.c
83
e = find_unwind_entry_in_table(&kernel_unwind_table, addr);
arch/parisc/kernel/unwind.c
91
e = find_unwind_entry_in_table(table, addr);
arch/parisc/kernel/unwind.c
92
if (e) {
arch/parisc/mm/init.c
751
unsigned long s, t, e;
arch/parisc/mm/init.c
763
e = (uintptr_t) dereference_function_descriptor(&_etext);
arch/parisc/mm/init.c
765
BUG_ON(t != e);
arch/parisc/mm/init.c
769
alloc_btlb(s, e, &slot, 0x13800000);
arch/parisc/mm/init.c
773
e = (uintptr_t) dereference_function_descriptor(&__bss_start);
arch/parisc/mm/init.c
774
BUG_ON(t != e);
arch/parisc/mm/init.c
778
e = (uintptr_t) dereference_function_descriptor(&__bss_stop);
arch/parisc/mm/init.c
779
alloc_btlb(s, e, &slot, 0x11800000);
arch/parisc/net/bpf_jit_comp32.c
1121
int s, e, paoff, i = insn - ctx->prog->insnsi;
arch/parisc/net/bpf_jit_comp32.c
1383
e = ctx->ninsns;
arch/parisc/net/bpf_jit_comp32.c
1384
paoff -= (e - s);
arch/parisc/net/bpf_jit_comp32.c
702
int e, s = ctx->ninsns;
arch/parisc/net/bpf_jit_comp32.c
782
e = ctx->ninsns;
arch/parisc/net/bpf_jit_comp32.c
784
paoff -= (e - s);
arch/parisc/net/bpf_jit_comp32.c
791
int e, s;
arch/parisc/net/bpf_jit_comp32.c
858
e = ctx->ninsns;
arch/parisc/net/bpf_jit_comp32.c
860
paoff -= (e - s);
arch/parisc/net/bpf_jit_comp32.c
869
int e, s = ctx->ninsns;
arch/parisc/net/bpf_jit_comp32.c
876
e = ctx->ninsns;
arch/parisc/net/bpf_jit_comp32.c
878
paoff -= (e - s);
arch/parisc/net/bpf_jit_comp64.c
251
int e, s;
arch/parisc/net/bpf_jit_comp64.c
320
e = ctx->ninsns;
arch/parisc/net/bpf_jit_comp64.c
322
paoff -= (e - s);
arch/parisc/net/bpf_jit_comp64.c
597
int s, e, ret, i = insn - ctx->prog->insnsi;
arch/parisc/net/bpf_jit_comp64.c
894
e = ctx->ninsns;
arch/parisc/net/bpf_jit_comp64.c
897
paoff -= (e - s);
arch/parisc/net/bpf_jit_comp64.c
945
e = ctx->ninsns;
arch/parisc/net/bpf_jit_comp64.c
948
paoff -= (e - s);
arch/parisc/net/bpf_jit_comp64.c
963
e = ctx->ninsns;
arch/parisc/net/bpf_jit_comp64.c
964
paoff -= (e - s);
arch/powerpc/boot/rs6000.h
142
} e;
arch/powerpc/boot/rs6000.h
143
} e;
arch/powerpc/include/asm/book3s/32/pgtable.h
206
#define pgd_ERROR(e) \
arch/powerpc/include/asm/book3s/32/pgtable.h
207
pr_err("%s:%d: bad pgd %08lx.\n", __FILE__, __LINE__, pgd_val(e))
arch/powerpc/include/asm/book3s/64/pgtable.h
994
#define pmd_ERROR(e) \
arch/powerpc/include/asm/book3s/64/pgtable.h
995
pr_err("%s:%d: bad pmd %08lx.\n", __FILE__, __LINE__, pmd_val(e))
arch/powerpc/include/asm/book3s/64/pgtable.h
996
#define pud_ERROR(e) \
arch/powerpc/include/asm/book3s/64/pgtable.h
997
pr_err("%s:%d: bad pud %08lx.\n", __FILE__, __LINE__, pud_val(e))
arch/powerpc/include/asm/book3s/64/pgtable.h
998
#define pgd_ERROR(e) \
arch/powerpc/include/asm/book3s/64/pgtable.h
999
pr_err("%s:%d: bad pgd %08lx.\n", __FILE__, __LINE__, pgd_val(e))
arch/powerpc/include/asm/nohash/32/pgtable.h
54
#define pgd_ERROR(e) \
arch/powerpc/include/asm/nohash/32/pgtable.h
55
pr_err("%s:%d: bad pgd %08llx.\n", __FILE__, __LINE__, (unsigned long long)pgd_val(e))
arch/powerpc/include/asm/nohash/64/pgtable-4k.h
85
#define pud_ERROR(e) \
arch/powerpc/include/asm/nohash/64/pgtable-4k.h
86
pr_err("%s:%d: bad pud %08lx.\n", __FILE__, __LINE__, pud_val(e))
arch/powerpc/include/asm/nohash/64/pgtable.h
162
#define pmd_ERROR(e) \
arch/powerpc/include/asm/nohash/64/pgtable.h
163
pr_err("%s:%d: bad pmd %08lx.\n", __FILE__, __LINE__, pmd_val(e))
arch/powerpc/include/asm/nohash/64/pgtable.h
164
#define pgd_ERROR(e) \
arch/powerpc/include/asm/nohash/64/pgtable.h
165
pr_err("%s:%d: bad pgd %08lx.\n", __FILE__, __LINE__, pgd_val(e))
arch/powerpc/include/asm/uaccess.h
539
#define arch_unsafe_get_user(x, p, e) do { \
arch/powerpc/include/asm/uaccess.h
543
__get_user_size_goto(__gu_val, __gu_addr, sizeof(*(p)), e); \
arch/powerpc/include/asm/uaccess.h
547
#define arch_unsafe_put_user(x, p, e) \
arch/powerpc/include/asm/uaccess.h
548
__put_user_size_goto((__typeof__(*(p)))(x), (p), sizeof(*(p)), e)
arch/powerpc/include/asm/uaccess.h
550
#define unsafe_copy_from_user(d, s, l, e) \
arch/powerpc/include/asm/uaccess.h
558
unsafe_get_user(*(u64 *)(_dst + _i), (u64 __user *)(_src + _i), e); \
arch/powerpc/include/asm/uaccess.h
560
unsafe_get_user(*(u32 *)(_dst + _i), (u32 __user *)(_src + _i), e); \
arch/powerpc/include/asm/uaccess.h
564
unsafe_get_user(*(u16 *)(_dst + _i), (u16 __user *)(_src + _i), e); \
arch/powerpc/include/asm/uaccess.h
568
unsafe_get_user(*(u8 *)(_dst + _i), (u8 __user *)(_src + _i), e); \
arch/powerpc/include/asm/uaccess.h
571
#define unsafe_copy_to_user(d, s, l, e) \
arch/powerpc/include/asm/uaccess.h
579
unsafe_put_user(*(u64 *)(_src + _i), (u64 __user *)(_dst + _i), e); \
arch/powerpc/include/asm/uaccess.h
581
unsafe_put_user(*(u32*)(_src + _i), (u32 __user *)(_dst + _i), e); \
arch/powerpc/include/asm/uaccess.h
585
unsafe_put_user(*(u16*)(_src + _i), (u16 __user *)(_dst + _i), e); \
arch/powerpc/include/asm/uaccess.h
589
unsafe_put_user(*(u8*)(_src + _i), (u8 __user *)(_dst + _i), e); \
arch/powerpc/include/uapi/asm/kvm.h
260
} e;
arch/powerpc/kvm/book3s.c
1040
static int kvmppc_book3s_set_irq(struct kvm_kernel_irq_routing_entry *e,
arch/powerpc/kvm/book3s.c
1044
return kvm_set_irq(kvm, irq_source_id, e->gsi, level, line_status);
arch/powerpc/kvm/booke.c
1514
sregs->u.e.features |= KVM_SREGS_E_BASE;
arch/powerpc/kvm/booke.c
1516
sregs->u.e.csrr0 = vcpu->arch.csrr0;
arch/powerpc/kvm/booke.c
1517
sregs->u.e.csrr1 = vcpu->arch.csrr1;
arch/powerpc/kvm/booke.c
1518
sregs->u.e.mcsr = vcpu->arch.mcsr;
arch/powerpc/kvm/booke.c
1519
sregs->u.e.esr = kvmppc_get_esr(vcpu);
arch/powerpc/kvm/booke.c
1520
sregs->u.e.dear = kvmppc_get_dar(vcpu);
arch/powerpc/kvm/booke.c
1521
sregs->u.e.tsr = vcpu->arch.tsr;
arch/powerpc/kvm/booke.c
1522
sregs->u.e.tcr = vcpu->arch.tcr;
arch/powerpc/kvm/booke.c
1523
sregs->u.e.dec = kvmppc_get_dec(vcpu, tb);
arch/powerpc/kvm/booke.c
1524
sregs->u.e.tb = tb;
arch/powerpc/kvm/booke.c
1525
sregs->u.e.vrsave = vcpu->arch.vrsave;
arch/powerpc/kvm/booke.c
1531
if (!(sregs->u.e.features & KVM_SREGS_E_BASE))
arch/powerpc/kvm/booke.c
1534
vcpu->arch.csrr0 = sregs->u.e.csrr0;
arch/powerpc/kvm/booke.c
1535
vcpu->arch.csrr1 = sregs->u.e.csrr1;
arch/powerpc/kvm/booke.c
1536
vcpu->arch.mcsr = sregs->u.e.mcsr;
arch/powerpc/kvm/booke.c
1537
kvmppc_set_esr(vcpu, sregs->u.e.esr);
arch/powerpc/kvm/booke.c
1538
kvmppc_set_dar(vcpu, sregs->u.e.dear);
arch/powerpc/kvm/booke.c
1539
vcpu->arch.vrsave = sregs->u.e.vrsave;
arch/powerpc/kvm/booke.c
1540
kvmppc_set_tcr(vcpu, sregs->u.e.tcr);
arch/powerpc/kvm/booke.c
1542
if (sregs->u.e.update_special & KVM_SREGS_E_UPDATE_DEC) {
arch/powerpc/kvm/booke.c
1543
vcpu->arch.dec = sregs->u.e.dec;
arch/powerpc/kvm/booke.c
1547
if (sregs->u.e.update_special & KVM_SREGS_E_UPDATE_TSR)
arch/powerpc/kvm/booke.c
1548
kvmppc_set_tsr(vcpu, sregs->u.e.tsr);
arch/powerpc/kvm/booke.c
1556
sregs->u.e.features |= KVM_SREGS_E_ARCH206;
arch/powerpc/kvm/booke.c
1558
sregs->u.e.pir = vcpu->vcpu_id;
arch/powerpc/kvm/booke.c
1559
sregs->u.e.mcsrr0 = vcpu->arch.mcsrr0;
arch/powerpc/kvm/booke.c
1560
sregs->u.e.mcsrr1 = vcpu->arch.mcsrr1;
arch/powerpc/kvm/booke.c
1561
sregs->u.e.decar = vcpu->arch.decar;
arch/powerpc/kvm/booke.c
1562
sregs->u.e.ivpr = vcpu->arch.ivpr;
arch/powerpc/kvm/booke.c
1568
if (!(sregs->u.e.features & KVM_SREGS_E_ARCH206))
arch/powerpc/kvm/booke.c
1571
if (sregs->u.e.pir != vcpu->vcpu_id)
arch/powerpc/kvm/booke.c
1574
vcpu->arch.mcsrr0 = sregs->u.e.mcsrr0;
arch/powerpc/kvm/booke.c
1575
vcpu->arch.mcsrr1 = sregs->u.e.mcsrr1;
arch/powerpc/kvm/booke.c
1576
vcpu->arch.decar = sregs->u.e.decar;
arch/powerpc/kvm/booke.c
1577
vcpu->arch.ivpr = sregs->u.e.ivpr;
arch/powerpc/kvm/booke.c
1584
sregs->u.e.features |= KVM_SREGS_E_IVOR;
arch/powerpc/kvm/booke.c
1586
sregs->u.e.ivor_low[0] = vcpu->arch.ivor[BOOKE_IRQPRIO_CRITICAL];
arch/powerpc/kvm/booke.c
1587
sregs->u.e.ivor_low[1] = vcpu->arch.ivor[BOOKE_IRQPRIO_MACHINE_CHECK];
arch/powerpc/kvm/booke.c
1588
sregs->u.e.ivor_low[2] = vcpu->arch.ivor[BOOKE_IRQPRIO_DATA_STORAGE];
arch/powerpc/kvm/booke.c
1589
sregs->u.e.ivor_low[3] = vcpu->arch.ivor[BOOKE_IRQPRIO_INST_STORAGE];
arch/powerpc/kvm/booke.c
1590
sregs->u.e.ivor_low[4] = vcpu->arch.ivor[BOOKE_IRQPRIO_EXTERNAL];
arch/powerpc/kvm/booke.c
1591
sregs->u.e.ivor_low[5] = vcpu->arch.ivor[BOOKE_IRQPRIO_ALIGNMENT];
arch/powerpc/kvm/booke.c
1592
sregs->u.e.ivor_low[6] = vcpu->arch.ivor[BOOKE_IRQPRIO_PROGRAM];
arch/powerpc/kvm/booke.c
1593
sregs->u.e.ivor_low[7] = vcpu->arch.ivor[BOOKE_IRQPRIO_FP_UNAVAIL];
arch/powerpc/kvm/booke.c
1594
sregs->u.e.ivor_low[8] = vcpu->arch.ivor[BOOKE_IRQPRIO_SYSCALL];
arch/powerpc/kvm/booke.c
1595
sregs->u.e.ivor_low[9] = vcpu->arch.ivor[BOOKE_IRQPRIO_AP_UNAVAIL];
arch/powerpc/kvm/booke.c
1596
sregs->u.e.ivor_low[10] = vcpu->arch.ivor[BOOKE_IRQPRIO_DECREMENTER];
arch/powerpc/kvm/booke.c
1597
sregs->u.e.ivor_low[11] = vcpu->arch.ivor[BOOKE_IRQPRIO_FIT];
arch/powerpc/kvm/booke.c
1598
sregs->u.e.ivor_low[12] = vcpu->arch.ivor[BOOKE_IRQPRIO_WATCHDOG];
arch/powerpc/kvm/booke.c
1599
sregs->u.e.ivor_low[13] = vcpu->arch.ivor[BOOKE_IRQPRIO_DTLB_MISS];
arch/powerpc/kvm/booke.c
1600
sregs->u.e.ivor_low[14] = vcpu->arch.ivor[BOOKE_IRQPRIO_ITLB_MISS];
arch/powerpc/kvm/booke.c
1601
sregs->u.e.ivor_low[15] = vcpu->arch.ivor[BOOKE_IRQPRIO_DEBUG];
arch/powerpc/kvm/booke.c
1607
if (!(sregs->u.e.features & KVM_SREGS_E_IVOR))
arch/powerpc/kvm/booke.c
1610
vcpu->arch.ivor[BOOKE_IRQPRIO_CRITICAL] = sregs->u.e.ivor_low[0];
arch/powerpc/kvm/booke.c
1611
vcpu->arch.ivor[BOOKE_IRQPRIO_MACHINE_CHECK] = sregs->u.e.ivor_low[1];
arch/powerpc/kvm/booke.c
1612
vcpu->arch.ivor[BOOKE_IRQPRIO_DATA_STORAGE] = sregs->u.e.ivor_low[2];
arch/powerpc/kvm/booke.c
1613
vcpu->arch.ivor[BOOKE_IRQPRIO_INST_STORAGE] = sregs->u.e.ivor_low[3];
arch/powerpc/kvm/booke.c
1614
vcpu->arch.ivor[BOOKE_IRQPRIO_EXTERNAL] = sregs->u.e.ivor_low[4];
arch/powerpc/kvm/booke.c
1615
vcpu->arch.ivor[BOOKE_IRQPRIO_ALIGNMENT] = sregs->u.e.ivor_low[5];
arch/powerpc/kvm/booke.c
1616
vcpu->arch.ivor[BOOKE_IRQPRIO_PROGRAM] = sregs->u.e.ivor_low[6];
arch/powerpc/kvm/booke.c
1617
vcpu->arch.ivor[BOOKE_IRQPRIO_FP_UNAVAIL] = sregs->u.e.ivor_low[7];
arch/powerpc/kvm/booke.c
1618
vcpu->arch.ivor[BOOKE_IRQPRIO_SYSCALL] = sregs->u.e.ivor_low[8];
arch/powerpc/kvm/booke.c
1619
vcpu->arch.ivor[BOOKE_IRQPRIO_AP_UNAVAIL] = sregs->u.e.ivor_low[9];
arch/powerpc/kvm/booke.c
1620
vcpu->arch.ivor[BOOKE_IRQPRIO_DECREMENTER] = sregs->u.e.ivor_low[10];
arch/powerpc/kvm/booke.c
1621
vcpu->arch.ivor[BOOKE_IRQPRIO_FIT] = sregs->u.e.ivor_low[11];
arch/powerpc/kvm/booke.c
1622
vcpu->arch.ivor[BOOKE_IRQPRIO_WATCHDOG] = sregs->u.e.ivor_low[12];
arch/powerpc/kvm/booke.c
1623
vcpu->arch.ivor[BOOKE_IRQPRIO_DTLB_MISS] = sregs->u.e.ivor_low[13];
arch/powerpc/kvm/booke.c
1624
vcpu->arch.ivor[BOOKE_IRQPRIO_ITLB_MISS] = sregs->u.e.ivor_low[14];
arch/powerpc/kvm/booke.c
1625
vcpu->arch.ivor[BOOKE_IRQPRIO_DEBUG] = sregs->u.e.ivor_low[15];
arch/powerpc/kvm/e500.c
366
sregs->u.e.features |= KVM_SREGS_E_ARCH206_MMU | KVM_SREGS_E_SPE |
arch/powerpc/kvm/e500.c
368
sregs->u.e.impl_id = KVM_SREGS_E_IMPL_FSL;
arch/powerpc/kvm/e500.c
370
sregs->u.e.impl.fsl.features = 0;
arch/powerpc/kvm/e500.c
371
sregs->u.e.impl.fsl.svr = vcpu_e500->svr;
arch/powerpc/kvm/e500.c
372
sregs->u.e.impl.fsl.hid0 = vcpu_e500->hid0;
arch/powerpc/kvm/e500.c
373
sregs->u.e.impl.fsl.mcar = vcpu_e500->mcar;
arch/powerpc/kvm/e500.c
375
sregs->u.e.ivor_high[0] = vcpu->arch.ivor[BOOKE_IRQPRIO_SPE_UNAVAIL];
arch/powerpc/kvm/e500.c
376
sregs->u.e.ivor_high[1] = vcpu->arch.ivor[BOOKE_IRQPRIO_SPE_FP_DATA];
arch/powerpc/kvm/e500.c
377
sregs->u.e.ivor_high[2] = vcpu->arch.ivor[BOOKE_IRQPRIO_SPE_FP_ROUND];
arch/powerpc/kvm/e500.c
378
sregs->u.e.ivor_high[3] =
arch/powerpc/kvm/e500.c
392
if (sregs->u.e.impl_id == KVM_SREGS_E_IMPL_FSL) {
arch/powerpc/kvm/e500.c
393
vcpu_e500->svr = sregs->u.e.impl.fsl.svr;
arch/powerpc/kvm/e500.c
394
vcpu_e500->hid0 = sregs->u.e.impl.fsl.hid0;
arch/powerpc/kvm/e500.c
395
vcpu_e500->mcar = sregs->u.e.impl.fsl.mcar;
arch/powerpc/kvm/e500.c
402
if (!(sregs->u.e.features & KVM_SREGS_E_IVOR))
arch/powerpc/kvm/e500.c
405
if (sregs->u.e.features & KVM_SREGS_E_SPE) {
arch/powerpc/kvm/e500.c
407
sregs->u.e.ivor_high[0];
arch/powerpc/kvm/e500.c
409
sregs->u.e.ivor_high[1];
arch/powerpc/kvm/e500.c
411
sregs->u.e.ivor_high[2];
arch/powerpc/kvm/e500.c
414
if (sregs->u.e.features & KVM_SREGS_E_PM) {
arch/powerpc/kvm/e500.c
416
sregs->u.e.ivor_high[3];
arch/powerpc/kvm/e500_mmu.c
569
sregs->u.e.mas0 = vcpu->arch.shared->mas0;
arch/powerpc/kvm/e500_mmu.c
570
sregs->u.e.mas1 = vcpu->arch.shared->mas1;
arch/powerpc/kvm/e500_mmu.c
571
sregs->u.e.mas2 = vcpu->arch.shared->mas2;
arch/powerpc/kvm/e500_mmu.c
572
sregs->u.e.mas7_3 = vcpu->arch.shared->mas7_3;
arch/powerpc/kvm/e500_mmu.c
573
sregs->u.e.mas4 = vcpu->arch.shared->mas4;
arch/powerpc/kvm/e500_mmu.c
574
sregs->u.e.mas6 = vcpu->arch.shared->mas6;
arch/powerpc/kvm/e500_mmu.c
576
sregs->u.e.mmucfg = vcpu->arch.mmucfg;
arch/powerpc/kvm/e500_mmu.c
577
sregs->u.e.tlbcfg[0] = vcpu->arch.tlbcfg[0];
arch/powerpc/kvm/e500_mmu.c
578
sregs->u.e.tlbcfg[1] = vcpu->arch.tlbcfg[1];
arch/powerpc/kvm/e500_mmu.c
579
sregs->u.e.tlbcfg[2] = 0;
arch/powerpc/kvm/e500_mmu.c
580
sregs->u.e.tlbcfg[3] = 0;
arch/powerpc/kvm/e500_mmu.c
585
if (sregs->u.e.features & KVM_SREGS_E_ARCH206_MMU) {
arch/powerpc/kvm/e500_mmu.c
586
vcpu->arch.shared->mas0 = sregs->u.e.mas0;
arch/powerpc/kvm/e500_mmu.c
587
vcpu->arch.shared->mas1 = sregs->u.e.mas1;
arch/powerpc/kvm/e500_mmu.c
588
vcpu->arch.shared->mas2 = sregs->u.e.mas2;
arch/powerpc/kvm/e500_mmu.c
589
vcpu->arch.shared->mas7_3 = sregs->u.e.mas7_3;
arch/powerpc/kvm/e500_mmu.c
590
vcpu->arch.shared->mas4 = sregs->u.e.mas4;
arch/powerpc/kvm/e500_mmu.c
591
vcpu->arch.shared->mas6 = sregs->u.e.mas6;
arch/powerpc/kvm/e500mc.c
224
sregs->u.e.features |= KVM_SREGS_E_ARCH206_MMU | KVM_SREGS_E_PM |
arch/powerpc/kvm/e500mc.c
226
sregs->u.e.impl_id = KVM_SREGS_E_IMPL_FSL;
arch/powerpc/kvm/e500mc.c
228
sregs->u.e.impl.fsl.features = 0;
arch/powerpc/kvm/e500mc.c
229
sregs->u.e.impl.fsl.svr = vcpu_e500->svr;
arch/powerpc/kvm/e500mc.c
230
sregs->u.e.impl.fsl.hid0 = vcpu_e500->hid0;
arch/powerpc/kvm/e500mc.c
231
sregs->u.e.impl.fsl.mcar = vcpu_e500->mcar;
arch/powerpc/kvm/e500mc.c
235
sregs->u.e.ivor_high[3] =
arch/powerpc/kvm/e500mc.c
237
sregs->u.e.ivor_high[4] = vcpu->arch.ivor[BOOKE_IRQPRIO_DBELL];
arch/powerpc/kvm/e500mc.c
238
sregs->u.e.ivor_high[5] = vcpu->arch.ivor[BOOKE_IRQPRIO_DBELL_CRIT];
arch/powerpc/kvm/e500mc.c
249
if (sregs->u.e.impl_id == KVM_SREGS_E_IMPL_FSL) {
arch/powerpc/kvm/e500mc.c
250
vcpu_e500->svr = sregs->u.e.impl.fsl.svr;
arch/powerpc/kvm/e500mc.c
251
vcpu_e500->hid0 = sregs->u.e.impl.fsl.hid0;
arch/powerpc/kvm/e500mc.c
252
vcpu_e500->mcar = sregs->u.e.impl.fsl.mcar;
arch/powerpc/kvm/e500mc.c
259
if (!(sregs->u.e.features & KVM_SREGS_E_IVOR))
arch/powerpc/kvm/e500mc.c
262
if (sregs->u.e.features & KVM_SREGS_E_PM) {
arch/powerpc/kvm/e500mc.c
264
sregs->u.e.ivor_high[3];
arch/powerpc/kvm/e500mc.c
267
if (sregs->u.e.features & KVM_SREGS_E_PC) {
arch/powerpc/kvm/e500mc.c
269
sregs->u.e.ivor_high[4];
arch/powerpc/kvm/e500mc.c
271
sregs->u.e.ivor_high[5];
arch/powerpc/kvm/mpic.c
1790
static int mpic_set_irq(struct kvm_kernel_irq_routing_entry *e,
arch/powerpc/kvm/mpic.c
1794
u32 irq = e->irqchip.pin;
arch/powerpc/kvm/mpic.c
1806
int kvm_set_msi(struct kvm_kernel_irq_routing_entry *e,
arch/powerpc/kvm/mpic.c
1818
openpic_msi_write(kvm->arch.mpic, MSIIR_OFFSET, e->msi.data);
arch/powerpc/kvm/mpic.c
1826
struct kvm_kernel_irq_routing_entry *e,
arch/powerpc/kvm/mpic.c
1833
e->set = mpic_set_irq;
arch/powerpc/kvm/mpic.c
1834
e->irqchip.irqchip = ue->u.irqchip.irqchip;
arch/powerpc/kvm/mpic.c
1835
e->irqchip.pin = ue->u.irqchip.pin;
arch/powerpc/kvm/mpic.c
1836
if (e->irqchip.pin >= KVM_IRQCHIP_NUM_PINS)
arch/powerpc/kvm/mpic.c
1840
e->set = kvm_set_msi;
arch/powerpc/kvm/mpic.c
1841
e->msi.address_lo = ue->u.msi.address_lo;
arch/powerpc/kvm/mpic.c
1842
e->msi.address_hi = ue->u.msi.address_hi;
arch/powerpc/kvm/mpic.c
1843
e->msi.data = ue->u.msi.data;
arch/powerpc/kvm/test-guest-state-buffer.c
217
struct kvmppc_gs_buff_info e;
arch/powerpc/kvm/test-guest-state-buffer.c
256
kvmppc_gse_put_buff_info(gsb, KVMPPC_GSID_RUN_INPUT, data->e);
arch/powerpc/lib/rheap.c
158
unsigned long s, e, bs, be;
arch/powerpc/lib/rheap.c
164
e = s + size;
arch/powerpc/lib/rheap.c
184
if (e == bs)
arch/powerpc/lib/rheap.c
196
if (after && e != after->start)
arch/powerpc/lib/rheap.c
30
static inline void fixup(unsigned long s, unsigned long e, int d,
arch/powerpc/lib/rheap.c
332
unsigned long s, e, m;
arch/powerpc/lib/rheap.c
337
e = s + size;
arch/powerpc/lib/rheap.c
344
e = e & ~m;
arch/powerpc/lib/rheap.c
346
if (IS_ERR_VALUE(e) || (e < s))
arch/powerpc/lib/rheap.c
351
size = e - s;
arch/powerpc/lib/rheap.c
36
if (*pp >= s && *pp < e)
arch/powerpc/lib/rheap.c
374
unsigned long s, e, m, bs, be;
arch/powerpc/lib/rheap.c
382
e = s + size;
arch/powerpc/lib/rheap.c
389
e = e & ~m;
arch/powerpc/lib/rheap.c
40
if (*pp >= s && *pp < e)
arch/powerpc/lib/rheap.c
400
if (s >= bs && e <= be)
arch/powerpc/lib/rheap.c
409
if (bs == s && be == e) {
arch/powerpc/lib/rheap.c
417
if (bs == s || be == e) {
arch/powerpc/lib/rheap.c
428
newblk->start = e;
arch/powerpc/lib/rheap.c
429
newblk->size = be - e;
arch/powerpc/lib/rheap.c
531
unsigned long s, e, m, bs = 0, be = 0;
arch/powerpc/lib/rheap.c
539
e = s + size;
arch/powerpc/lib/rheap.c
546
e = e & ~m;
arch/powerpc/lib/rheap.c
557
if (s >= bs && e <= be)
arch/powerpc/lib/rheap.c
566
if (bs == s && be == e) {
arch/powerpc/lib/rheap.c
579
if (bs == s || be == e) {
arch/powerpc/lib/rheap.c
590
newblk2->start = e;
arch/powerpc/lib/rheap.c
591
newblk2->size = be - e;
arch/powerpc/lib/rheap.c
598
newblk1->size = e - s;
arch/powerpc/mm/book3s64/slb.c
214
unsigned long e, v;
arch/powerpc/mm/book3s64/slb.c
223
asm volatile("slbmfee %0,%1" : "=r" (e) : "r" (i));
arch/powerpc/mm/book3s64/slb.c
225
slb_ptr->esid = e;
arch/powerpc/mm/book3s64/slb.c
234
unsigned long e, v;
arch/powerpc/mm/book3s64/slb.c
243
e = slb_ptr->esid;
arch/powerpc/mm/book3s64/slb.c
247
if (!e && !v)
arch/powerpc/mm/book3s64/slb.c
250
pr_err("%02d %016lx %016lx %s\n", i, e, v,
arch/powerpc/mm/book3s64/slb.c
251
(e & SLB_ESID_V) ? "VALID" : "NOT VALID");
arch/powerpc/mm/book3s64/slb.c
253
if (!(e & SLB_ESID_V))
arch/powerpc/mm/book3s64/slb.c
259
GET_ESID_1T(e),
arch/powerpc/mm/book3s64/slb.c
263
GET_ESID(e),
arch/powerpc/platforms/pasemi/setup.c
382
unsigned long e, v;
arch/powerpc/platforms/pasemi/setup.c
387
asm volatile("slbmfee %0,%1" : "=r" (e) : "r" (i));
arch/powerpc/platforms/pasemi/setup.c
389
pr_err("%02d %016lx %016lx\n", i, e, v);
arch/powerpc/platforms/powernv/opal-irqchip.c
44
u64 e;
arch/powerpc/platforms/powernv/opal-irqchip.c
46
e = READ_ONCE(last_outstanding_events) & opal_event_irqchip.mask;
arch/powerpc/platforms/powernv/opal-irqchip.c
48
while (e) {
arch/powerpc/platforms/powernv/opal-irqchip.c
51
hwirq = fls64(e) - 1;
arch/powerpc/platforms/powernv/opal-irqchip.c
52
e &= ~BIT_ULL(hwirq);
arch/powerpc/platforms/powernv/opal-irqchip.c
65
e = be64_to_cpu(events) & opal_event_irqchip.mask;
arch/powerpc/platforms/powernv/opal-irqchip.c
66
if (e)
arch/riscv/include/asm/pgtable-64.h
267
#define pmd_ERROR(e) \
arch/riscv/include/asm/pgtable-64.h
268
pr_err("%s:%d: bad pmd %016lx.\n", __FILE__, __LINE__, pmd_val(e))
arch/riscv/include/asm/pgtable-64.h
270
#define pud_ERROR(e) \
arch/riscv/include/asm/pgtable-64.h
271
pr_err("%s:%d: bad pud %016lx.\n", __FILE__, __LINE__, pud_val(e))
arch/riscv/include/asm/pgtable-64.h
273
#define p4d_ERROR(e) \
arch/riscv/include/asm/pgtable-64.h
274
pr_err("%s:%d: bad p4d %016lx.\n", __FILE__, __LINE__, p4d_val(e))
arch/riscv/include/asm/pgtable.h
567
#define pgd_ERROR(e) \
arch/riscv/include/asm/pgtable.h
568
pr_err("%s:%d: bad pgd " PTE_FMT ".\n", __FILE__, __LINE__, pgd_val(e))
arch/riscv/include/asm/sbi.h
572
#define sbi_ecall(e, f, a0, a1, a2, a3, a4, a5) \
arch/riscv/include/asm/sbi.h
573
__sbi_ecall(a0, a1, a2, a3, a4, a5, f, e)
arch/riscv/kvm/vm.c
120
struct kvm_kernel_irq_routing_entry *e,
arch/riscv/kvm/vm.c
127
e->set = kvm_riscv_set_irq;
arch/riscv/kvm/vm.c
128
e->irqchip.irqchip = ue->u.irqchip.irqchip;
arch/riscv/kvm/vm.c
129
e->irqchip.pin = ue->u.irqchip.pin;
arch/riscv/kvm/vm.c
130
if ((e->irqchip.pin >= KVM_IRQCHIP_NUM_PINS) ||
arch/riscv/kvm/vm.c
131
(e->irqchip.irqchip >= KVM_NR_IRQCHIPS))
arch/riscv/kvm/vm.c
135
e->set = kvm_set_msi;
arch/riscv/kvm/vm.c
136
e->msi.address_lo = ue->u.msi.address_lo;
arch/riscv/kvm/vm.c
137
e->msi.address_hi = ue->u.msi.address_hi;
arch/riscv/kvm/vm.c
138
e->msi.data = ue->u.msi.data;
arch/riscv/kvm/vm.c
139
e->msi.flags = ue->flags;
arch/riscv/kvm/vm.c
140
e->msi.devid = ue->u.msi.devid;
arch/riscv/kvm/vm.c
150
int kvm_arch_set_irq_inatomic(struct kvm_kernel_irq_routing_entry *e,
arch/riscv/kvm/vm.c
157
switch (e->type) {
arch/riscv/kvm/vm.c
159
return kvm_set_msi(e, kvm, irq_source_id, level, line_status);
arch/riscv/kvm/vm.c
162
return kvm_riscv_set_irq(e, kvm, irq_source_id,
arch/riscv/kvm/vm.c
68
int kvm_set_msi(struct kvm_kernel_irq_routing_entry *e,
arch/riscv/kvm/vm.c
77
msi.address_lo = e->msi.address_lo;
arch/riscv/kvm/vm.c
78
msi.address_hi = e->msi.address_hi;
arch/riscv/kvm/vm.c
79
msi.data = e->msi.data;
arch/riscv/kvm/vm.c
80
msi.flags = e->msi.flags;
arch/riscv/kvm/vm.c
81
msi.devid = e->msi.devid;
arch/riscv/kvm/vm.c
86
static int kvm_riscv_set_irq(struct kvm_kernel_irq_routing_entry *e,
arch/riscv/kvm/vm.c
90
return kvm_riscv_aia_inject_irq(kvm, e->irqchip.pin, level);
arch/riscv/net/bpf_jit_comp32.c
1213
e = ctx->ninsns;
arch/riscv/net/bpf_jit_comp32.c
1214
rvoff -= ninsns_rvoff(e - s);
arch/riscv/net/bpf_jit_comp32.c
567
int e, s = ctx->ninsns;
arch/riscv/net/bpf_jit_comp32.c
645
e = ctx->ninsns;
arch/riscv/net/bpf_jit_comp32.c
647
rvoff -= ninsns_rvoff(e - s);
arch/riscv/net/bpf_jit_comp32.c
654
int e, s = ctx->ninsns;
arch/riscv/net/bpf_jit_comp32.c
714
e = ctx->ninsns;
arch/riscv/net/bpf_jit_comp32.c
716
rvoff -= ninsns_rvoff(e - s);
arch/riscv/net/bpf_jit_comp32.c
725
int e, s = ctx->ninsns;
arch/riscv/net/bpf_jit_comp32.c
732
e = ctx->ninsns;
arch/riscv/net/bpf_jit_comp32.c
734
rvoff -= ninsns_rvoff(e - s);
arch/riscv/net/bpf_jit_comp32.c
958
int s, e, rvoff, i = insn - ctx->prog->insnsi;
arch/riscv/net/bpf_jit_comp64.c
1374
int s, e, rvoff, ret, i = insn - ctx->prog->insnsi;
arch/riscv/net/bpf_jit_comp64.c
1709
e = ctx->ninsns;
arch/riscv/net/bpf_jit_comp64.c
1712
rvoff -= ninsns_rvoff(e - s);
arch/riscv/net/bpf_jit_comp64.c
1761
e = ctx->ninsns;
arch/riscv/net/bpf_jit_comp64.c
1764
rvoff -= ninsns_rvoff(e - s);
arch/riscv/net/bpf_jit_comp64.c
1784
e = ctx->ninsns;
arch/riscv/net/bpf_jit_comp64.c
1785
rvoff -= ninsns_rvoff(e - s);
arch/s390/include/asm/pci.h
276
static inline void zpci_event_error(void *e) {}
arch/s390/include/asm/pci.h
277
static inline void zpci_event_availability(void *e) {}
arch/s390/include/asm/pgtable.h
69
#define pte_ERROR(e) \
arch/s390/include/asm/pgtable.h
70
pr_err("%s:%d: bad pte %016lx.\n", __FILE__, __LINE__, pte_val(e))
arch/s390/include/asm/pgtable.h
71
#define pmd_ERROR(e) \
arch/s390/include/asm/pgtable.h
72
pr_err("%s:%d: bad pmd %016lx.\n", __FILE__, __LINE__, pmd_val(e))
arch/s390/include/asm/pgtable.h
73
#define pud_ERROR(e) \
arch/s390/include/asm/pgtable.h
74
pr_err("%s:%d: bad pud %016lx.\n", __FILE__, __LINE__, pud_val(e))
arch/s390/include/asm/pgtable.h
75
#define p4d_ERROR(e) \
arch/s390/include/asm/pgtable.h
76
pr_err("%s:%d: bad p4d %016lx.\n", __FILE__, __LINE__, p4d_val(e))
arch/s390/include/asm/pgtable.h
77
#define pgd_ERROR(e) \
arch/s390/include/asm/pgtable.h
78
pr_err("%s:%d: bad pgd %016lx.\n", __FILE__, __LINE__, pgd_val(e))
arch/s390/include/uapi/asm/guarded_storage.h
20
__u8 e : 1;
arch/s390/include/uapi/asm/runtime_instr.h
42
__u32 e : 1;
arch/s390/kernel/cert_store.c
179
static void pr_dbf_vce(const struct vce *e)
arch/s390/kernel/cert_store.c
185
pr_dbf_msg("vce_hdr.vce_length: %d", e->vce_hdr.vce_length);
arch/s390/kernel/cert_store.c
186
pr_dbf_msg("vce_hdr.flags: %d", e->vce_hdr.flags);
arch/s390/kernel/cert_store.c
187
pr_dbf_msg("vce_hdr.key_type: %d", e->vce_hdr.key_type);
arch/s390/kernel/cert_store.c
188
pr_dbf_msg("vce_hdr.vc_index: %d", e->vce_hdr.vc_index);
arch/s390/kernel/cert_store.c
189
pr_dbf_msg("vce_hdr.vc_format: %d", e->vce_hdr.vc_format);
arch/s390/kernel/cert_store.c
190
pr_dbf_msg("vce_hdr.key_id_length: %d", e->vce_hdr.key_id_length);
arch/s390/kernel/cert_store.c
191
pr_dbf_msg("vce_hdr.vc_hash_type: %d", e->vce_hdr.vc_hash_type);
arch/s390/kernel/cert_store.c
192
pr_dbf_msg("vce_hdr.vc_hash_length: %d", e->vce_hdr.vc_hash_length);
arch/s390/kernel/cert_store.c
193
pr_dbf_msg("vce_hdr.vc_hash_offset: %d", e->vce_hdr.vc_hash_offset);
arch/s390/kernel/cert_store.c
194
pr_dbf_msg("vce_hdr.vc_length: %d", e->vce_hdr.vc_length);
arch/s390/kernel/cert_store.c
195
pr_dbf_msg("vce_hdr.vc_offset: %d", e->vce_hdr.vc_offset);
arch/s390/kernel/cert_store.c
198
memcpy(vc_name, e->vce_hdr.vc_name, VC_NAME_LEN_BYTES);
arch/s390/kernel/cert_store.c
204
e->vce_hdr.vc_index, vc_name);
arch/s390/kernel/cert_store.c
209
debug_event(cert_store_hexdump, 3, (u8 *)e->cert_data_buf, 128);
arch/s390/kernel/cert_store.c
212
(u8 *)e->cert_data_buf + e->vce_hdr.vce_length - 128, 128);
arch/s390/kvm/interrupt.c
2794
static int set_adapter_int(struct kvm_kernel_irq_routing_entry *e,
arch/s390/kvm/interrupt.c
2804
adapter = get_io_adapter(kvm, e->adapter.adapter_id);
arch/s390/kvm/interrupt.c
2807
ret = adapter_indicators_set(kvm, adapter, &e->adapter);
arch/s390/kvm/interrupt.c
2861
struct kvm_kernel_irq_routing_entry *e,
arch/s390/kvm/interrupt.c
2873
e->set = set_adapter_int;
arch/s390/kvm/interrupt.c
2890
e->adapter.summary_addr = uaddr_s;
arch/s390/kvm/interrupt.c
2891
e->adapter.summary_gaddr = ue->u.adapter.summary_addr;
arch/s390/kvm/interrupt.c
2892
e->adapter.ind_addr = uaddr_i;
arch/s390/kvm/interrupt.c
2893
e->adapter.ind_gaddr = ue->u.adapter.ind_addr;
arch/s390/kvm/interrupt.c
2894
e->adapter.summary_offset = ue->u.adapter.summary_offset;
arch/s390/kvm/interrupt.c
2895
e->adapter.ind_offset = ue->u.adapter.ind_offset;
arch/s390/kvm/interrupt.c
2896
e->adapter.adapter_id = ue->u.adapter.adapter_id;
arch/s390/kvm/interrupt.c
2903
int kvm_set_msi(struct kvm_kernel_irq_routing_entry *e, struct kvm *kvm,
arch/s390/lib/string.c
25
unsigned long e = 0;
arch/s390/lib/string.c
31
: [e] "+&a" (e), [s] "+&a" (s)
arch/s390/lib/string.c
34
return (char *)e;
arch/sh/include/asm/pgtable-3level.h
28
#define pmd_ERROR(e) \
arch/sh/include/asm/pgtable-3level.h
29
printk("%s:%d: bad pmd %016llx.\n", __FILE__, __LINE__, pmd_val(e))
arch/sh/include/asm/pgtable_32.h
404
#define pte_ERROR(e) \
arch/sh/include/asm/pgtable_32.h
406
&(e), (e).pte_high, (e).pte_low)
arch/sh/include/asm/pgtable_32.h
407
#define pgd_ERROR(e) \
arch/sh/include/asm/pgtable_32.h
408
printk("%s:%d: bad pgd %016llx.\n", __FILE__, __LINE__, pgd_val(e))
arch/sh/include/asm/pgtable_32.h
410
#define pte_ERROR(e) \
arch/sh/include/asm/pgtable_32.h
411
printk("%s:%d: bad pte %08lx.\n", __FILE__, __LINE__, pte_val(e))
arch/sh/include/asm/pgtable_32.h
412
#define pgd_ERROR(e) \
arch/sh/include/asm/pgtable_32.h
413
printk("%s:%d: bad pgd %08lx.\n", __FILE__, __LINE__, pgd_val(e))
arch/sparc/crypto/des_glue.c
33
static void encrypt_to_decrypt(u64 *d, const u64 *e)
arch/sparc/crypto/des_glue.c
35
const u64 *s = e + (DES_EXPKEY_WORDS / 2) - 1;
arch/sparc/include/asm/leon_amba.h
136
struct leon3_gptimerelem_regs_map e[8];
arch/sparc/include/asm/pgtable_32.h
43
#define pte_ERROR(e) __builtin_trap()
arch/sparc/include/asm/pgtable_32.h
44
#define pmd_ERROR(e) __builtin_trap()
arch/sparc/include/asm/pgtable_32.h
45
#define pgd_ERROR(e) __builtin_trap()
arch/sparc/include/asm/pgtable_64.h
101
__FILE__, __LINE__, &(e), pmd_val(e), __builtin_return_address(0))
arch/sparc/include/asm/pgtable_64.h
102
#define pud_ERROR(e) \
arch/sparc/include/asm/pgtable_64.h
104
__FILE__, __LINE__, &(e), pud_val(e), __builtin_return_address(0))
arch/sparc/include/asm/pgtable_64.h
105
#define pgd_ERROR(e) \
arch/sparc/include/asm/pgtable_64.h
107
__FILE__, __LINE__, &(e), pgd_val(e), __builtin_return_address(0))
arch/sparc/include/asm/pgtable_64.h
99
#define pmd_ERROR(e) \
arch/sparc/kernel/leon_kernel.c
264
rld = LEON3_BYPASS_LOAD_PA(&leon3_gptimer_regs->e[leon3_gptimer_idx].rld);
arch/sparc/kernel/leon_kernel.c
265
val = LEON3_BYPASS_LOAD_PA(&leon3_gptimer_regs->e[leon3_gptimer_idx].val);
arch/sparc/kernel/leon_kernel.c
266
ctrl = LEON3_BYPASS_LOAD_PA(&leon3_gptimer_regs->e[leon3_gptimer_idx].ctrl);
arch/sparc/kernel/leon_kernel.c
268
val = LEON3_BYPASS_LOAD_PA(&leon3_gptimer_regs->e[leon3_gptimer_idx].val);
arch/sparc/kernel/leon_kernel.c
385
ctrl = LEON3_BYPASS_LOAD_PA(&leon3_gptimer_regs->e[leon3_gptimer_idx].ctrl);
arch/sparc/kernel/leon_kernel.c
386
LEON3_BYPASS_STORE_PA(&leon3_gptimer_regs->e[leon3_gptimer_idx].ctrl,
arch/sparc/kernel/leon_kernel.c
388
ctrl = LEON3_BYPASS_LOAD_PA(&leon3_gptimer_regs->e[leon3_gptimer_idx].ctrl);
arch/sparc/kernel/leon_kernel.c
395
LEON3_BYPASS_STORE_PA(&leon3_gptimer_regs->e[leon3_gptimer_idx].val, 0);
arch/sparc/kernel/leon_kernel.c
396
LEON3_BYPASS_STORE_PA(&leon3_gptimer_regs->e[leon3_gptimer_idx].rld,
arch/sparc/kernel/leon_kernel.c
399
&leon3_gptimer_regs->e[leon3_gptimer_idx].ctrl, 0);
arch/sparc/kernel/leon_kernel.c
459
LEON3_BYPASS_STORE_PA(&leon3_gptimer_regs->e[leon3_gptimer_idx].ctrl,
arch/sparc/kernel/leon_kernel.c
475
ctrl = LEON3_BYPASS_LOAD_PA(&leon3_gptimer_regs->e[leon3_gptimer_idx].ctrl);
arch/sparc/kernel/leon_kernel.c
476
LEON3_BYPASS_STORE_PA(&leon3_gptimer_regs->e[leon3_gptimer_idx].ctrl,
arch/sparc/vdso/vma.c
102
static Elf64_Sym *find_sym64(const struct vdso_elfinfo64 *e, const char *name)
arch/sparc/vdso/vma.c
106
for (i = 0; i < (e->dynsymsize / sizeof(Elf64_Sym)); i++) {
arch/sparc/vdso/vma.c
107
Elf64_Sym *s = &e->dynsym[i];
arch/sparc/vdso/vma.c
110
if (!strcmp(e->dynstr + s->st_name, name))
arch/sparc/vdso/vma.c
119
struct vdso_elfinfo64 *e = &_e->u.elf64;
arch/sparc/vdso/vma.c
120
Elf64_Sym *osym = find_sym64(e, orig);
arch/sparc/vdso/vma.c
121
Elf64_Sym *nsym = find_sym64(e, new);
arch/sparc/vdso/vma.c
136
static void *one_section32(struct vdso_elfinfo32 *e, const char *name,
arch/sparc/vdso/vma.c
143
shdrs = (void *)e->hdr + e->hdr->e_shoff;
arch/sparc/vdso/vma.c
144
snames = (void *)e->hdr + shdrs[e->hdr->e_shstrndx].sh_offset;
arch/sparc/vdso/vma.c
145
for (i = 1; i < e->hdr->e_shnum; i++) {
arch/sparc/vdso/vma.c
149
return (void *)e->hdr + shdrs[i].sh_offset;
arch/sparc/vdso/vma.c
157
struct vdso_elfinfo32 *e = &_e->u.elf32;
arch/sparc/vdso/vma.c
159
e->hdr = image->data;
arch/sparc/vdso/vma.c
160
e->dynsym = one_section32(e, ".dynsym", &e->dynsymsize);
arch/sparc/vdso/vma.c
161
e->dynstr = one_section32(e, ".dynstr", NULL);
arch/sparc/vdso/vma.c
163
if (!e->dynsym || !e->dynstr) {
arch/sparc/vdso/vma.c
170
static Elf32_Sym *find_sym32(const struct vdso_elfinfo32 *e, const char *name)
arch/sparc/vdso/vma.c
174
for (i = 0; i < (e->dynsymsize / sizeof(Elf32_Sym)); i++) {
arch/sparc/vdso/vma.c
175
Elf32_Sym *s = &e->dynsym[i];
arch/sparc/vdso/vma.c
178
if (!strcmp(e->dynstr + s->st_name, name))
arch/sparc/vdso/vma.c
187
struct vdso_elfinfo32 *e = &_e->u.elf32;
arch/sparc/vdso/vma.c
188
Elf32_Sym *osym = find_sym32(e, orig);
arch/sparc/vdso/vma.c
189
Elf32_Sym *nsym = find_sym32(e, new);
arch/sparc/vdso/vma.c
204
static int find_sections(const struct vdso_image *image, struct vdso_elfinfo *e,
arch/sparc/vdso/vma.c
208
return find_sections64(image, e);
arch/sparc/vdso/vma.c
210
return find_sections32(image, e);
arch/sparc/vdso/vma.c
213
static int patch_one_symbol(struct vdso_elfinfo *e, const char *orig,
arch/sparc/vdso/vma.c
217
return patchsym64(e, orig, new_target);
arch/sparc/vdso/vma.c
219
return patchsym32(e, orig, new_target);
arch/sparc/vdso/vma.c
222
static int stick_patch(const struct vdso_image *image, struct vdso_elfinfo *e, bool elf64)
arch/sparc/vdso/vma.c
226
err = find_sections(image, e, elf64);
arch/sparc/vdso/vma.c
230
err = patch_one_symbol(e,
arch/sparc/vdso/vma.c
236
return patch_one_symbol(e,
arch/sparc/vdso/vma.c
68
static void *one_section64(struct vdso_elfinfo64 *e, const char *name,
arch/sparc/vdso/vma.c
75
shdrs = (void *)e->hdr + e->hdr->e_shoff;
arch/sparc/vdso/vma.c
76
snames = (void *)e->hdr + shdrs[e->hdr->e_shstrndx].sh_offset;
arch/sparc/vdso/vma.c
77
for (i = 1; i < e->hdr->e_shnum; i++) {
arch/sparc/vdso/vma.c
81
return (void *)e->hdr + shdrs[i].sh_offset;
arch/sparc/vdso/vma.c
89
struct vdso_elfinfo64 *e = &_e->u.elf64;
arch/sparc/vdso/vma.c
91
e->hdr = image->data;
arch/sparc/vdso/vma.c
92
e->dynsym = one_section64(e, ".dynsym", &e->dynsymsize);
arch/sparc/vdso/vma.c
93
e->dynstr = one_section64(e, ".dynstr", NULL);
arch/sparc/vdso/vma.c
95
if (!e->dynsym || !e->dynstr) {
arch/um/include/asm/pgtable-2level.h
27
#define pte_ERROR(e) \
arch/um/include/asm/pgtable-2level.h
28
printk("%s:%d: bad pte %p(%08lx).\n", __FILE__, __LINE__, &(e), \
arch/um/include/asm/pgtable-2level.h
29
pte_val(e))
arch/um/include/asm/pgtable-2level.h
30
#define pgd_ERROR(e) \
arch/um/include/asm/pgtable-2level.h
31
printk("%s:%d: bad pgd %p(%08lx).\n", __FILE__, __LINE__, &(e), \
arch/um/include/asm/pgtable-2level.h
32
pgd_val(e))
arch/um/include/asm/pgtable-4level.h
45
#define pte_ERROR(e) \
arch/um/include/asm/pgtable-4level.h
46
printk("%s:%d: bad pte %p(%016lx).\n", __FILE__, __LINE__, &(e), \
arch/um/include/asm/pgtable-4level.h
47
pte_val(e))
arch/um/include/asm/pgtable-4level.h
48
#define pmd_ERROR(e) \
arch/um/include/asm/pgtable-4level.h
49
printk("%s:%d: bad pmd %p(%016lx).\n", __FILE__, __LINE__, &(e), \
arch/um/include/asm/pgtable-4level.h
50
pmd_val(e))
arch/um/include/asm/pgtable-4level.h
51
#define pud_ERROR(e) \
arch/um/include/asm/pgtable-4level.h
52
printk("%s:%d: bad pud %p(%016lx).\n", __FILE__, __LINE__, &(e), \
arch/um/include/asm/pgtable-4level.h
53
pud_val(e))
arch/um/include/asm/pgtable-4level.h
54
#define pgd_ERROR(e) \
arch/um/include/asm/pgtable-4level.h
55
printk("%s:%d: bad pgd %p(%016lx).\n", __FILE__, __LINE__, &(e), \
arch/um/include/asm/pgtable-4level.h
56
pgd_val(e))
arch/um/include/linux/time-internal.h
27
time_travel_set_event_fn(struct time_travel_event *e,
arch/um/include/linux/time-internal.h
30
e->fn = fn;
arch/um/include/linux/time-internal.h
49
void time_travel_add_irq_event(struct time_travel_event *e);
arch/um/include/linux/time-internal.h
50
void time_travel_add_event_rel(struct time_travel_event *e,
arch/um/include/linux/time-internal.h
52
bool time_travel_del_event(struct time_travel_event *e);
arch/um/include/linux/time-internal.h
62
#define time_travel_set_event_fn(e, fn) do {} while (0)
arch/um/include/linux/time-internal.h
72
static inline void time_travel_add_irq_event(struct time_travel_event *e)
arch/um/include/shared/os.h
129
flags.e = 1;
arch/um/include/shared/os.h
72
unsigned int e : 1; /* O_EXCL */
arch/um/include/shared/os.h
77
.t = 0, .a = 0, .e = 0, .cl = 0 })
arch/um/kernel/time.c
399
static void __time_travel_add_event(struct time_travel_event *e,
arch/um/kernel/time.c
406
if (e->pending)
arch/um/kernel/time.c
409
e->pending = true;
arch/um/kernel/time.c
410
e->time = time;
arch/um/kernel/time.c
421
if ((tmp->time > e->time) ||
arch/um/kernel/time.c
422
(tmp->time == e->time && tmp->onstack && e->onstack)) {
arch/um/kernel/time.c
423
list_add_tail(&e->list, &tmp->list);
arch/um/kernel/time.c
430
list_add_tail(&e->list, &time_travel_events);
arch/um/kernel/time.c
438
static void time_travel_add_event(struct time_travel_event *e,
arch/um/kernel/time.c
441
if (WARN_ON(!e->fn))
arch/um/kernel/time.c
444
__time_travel_add_event(e, time);
arch/um/kernel/time.c
447
void time_travel_add_event_rel(struct time_travel_event *e,
arch/um/kernel/time.c
450
time_travel_add_event(e, time_travel_time + delay_ns);
arch/um/kernel/time.c
453
static void time_travel_periodic_timer(struct time_travel_event *e)
arch/um/kernel/time.c
467
struct time_travel_event *e;
arch/um/kernel/time.c
480
while ((e = list_first_entry_or_null(&time_travel_irqs,
arch/um/kernel/time.c
483
list_del(&e->list);
arch/um/kernel/time.c
484
e->pending = false;
arch/um/kernel/time.c
485
e->fn(e);
arch/um/kernel/time.c
491
static void time_travel_deliver_event(struct time_travel_event *e)
arch/um/kernel/time.c
493
if (e == &time_travel_timer_event) {
arch/um/kernel/time.c
498
e->fn(e);
arch/um/kernel/time.c
500
list_add_tail(&e->list, &time_travel_irqs);
arch/um/kernel/time.c
506
e->pending = true;
arch/um/kernel/time.c
512
e->fn(e);
arch/um/kernel/time.c
518
bool time_travel_del_event(struct time_travel_event *e)
arch/um/kernel/time.c
522
if (!e->pending)
arch/um/kernel/time.c
525
list_del(&e->list);
arch/um/kernel/time.c
526
e->pending = false;
arch/um/kernel/time.c
536
struct time_travel_event *e;
arch/um/kernel/time.c
543
e = time_travel_first_event();
arch/um/kernel/time.c
545
BUG_ON(!e);
arch/um/kernel/time.c
546
__time_travel_update_time(e->time, idle);
arch/um/kernel/time.c
549
if (e == time_travel_first_event()) {
arch/um/kernel/time.c
550
BUG_ON(!time_travel_del_event(e));
arch/um/kernel/time.c
551
BUG_ON(time_travel_time != e->time);
arch/um/kernel/time.c
553
if (e == &ne) {
arch/um/kernel/time.c
556
if (e->onstack)
arch/um/kernel/time.c
558
time_travel_time, e->time, e);
arch/um/kernel/time.c
559
time_travel_deliver_event(e);
arch/um/kernel/time.c
563
e = time_travel_first_event();
arch/um/kernel/time.c
564
if (e)
arch/um/kernel/time.c
565
time_travel_ext_update_request(e->time);
arch/um/kernel/time.c
597
void time_travel_add_irq_event(struct time_travel_event *e)
arch/um/kernel/time.c
607
time_travel_add_event(e, time_travel_time);
arch/um/kernel/time.c
611
static void time_travel_oneshot_timer(struct time_travel_event *e)
arch/um/kernel/time.c
756
#define time_travel_add_event(e, time) do { } while (0)
arch/um/kernel/time.c
759
#define time_travel_del_event(e) do { } while (0)
arch/um/os-Linux/file.c
183
if (flags.e)
arch/x86/boot/compressed/kaslr.c
680
struct efi_info *e = &boot_params_ptr->efi_info;
arch/x86/boot/compressed/kaslr.c
689
signature = (char *)&e->efi_loader_signature;
arch/x86/boot/compressed/kaslr.c
696
if (e->efi_memmap_hi) {
arch/x86/boot/compressed/kaslr.c
700
pmap = e->efi_memmap;
arch/x86/boot/compressed/kaslr.c
702
pmap = (e->efi_memmap | ((__u64)e->efi_memmap_hi << 32));
arch/x86/boot/compressed/kaslr.c
705
nr_desc = e->efi_memmap_size / e->efi_memdesc_size;
arch/x86/boot/compressed/kaslr.c
707
md = efi_early_memdesc_ptr(pmap, e->efi_memdesc_size, i);
arch/x86/boot/compressed/kaslr.c
715
md = efi_early_memdesc_ptr(pmap, e->efi_memdesc_size, i);
arch/x86/boot/cpu.c
38
u32 e = err_flags[i];
arch/x86/boot/cpu.c
47
if (e & 1) {
arch/x86/boot/cpu.c
55
e >>= 1;
arch/x86/boot/early_serial_console.c
54
char *e;
arch/x86/boot/early_serial_console.c
71
port = simple_strtoull(arg + pos, &e, 16);
arch/x86/boot/early_serial_console.c
72
if (port == 0 || arg + pos == e)
arch/x86/boot/early_serial_console.c
75
pos = e - arg;
arch/x86/boot/early_serial_console.c
92
baud = simple_strtoull(arg + pos, &e, 0);
arch/x86/boot/early_serial_console.c
93
if (baud == 0 || arg + pos == e)
arch/x86/boot/startup/sev-shared.c
233
const struct snp_cpuid_fn *e = &cpuid_table->fn[i];
arch/x86/boot/startup/sev-shared.c
235
if (!(e->eax_in == 0xD && e->ecx_in > 1 && e->ecx_in < 64))
arch/x86/boot/startup/sev-shared.c
237
if (!(xfeatures_en & (BIT_ULL(e->ecx_in))))
arch/x86/boot/startup/sev-shared.c
239
if (xfeatures_found & (BIT_ULL(e->ecx_in)))
arch/x86/boot/startup/sev-shared.c
242
xfeatures_found |= (BIT_ULL(e->ecx_in));
arch/x86/boot/startup/sev-shared.c
245
xsave_size += e->eax;
arch/x86/boot/startup/sev-shared.c
247
xsave_size = max(xsave_size, e->eax + e->ebx);
arch/x86/boot/startup/sev-shared.c
268
const struct snp_cpuid_fn *e = &cpuid_table->fn[i];
arch/x86/boot/startup/sev-shared.c
270
if (e->eax_in != leaf->fn)
arch/x86/boot/startup/sev-shared.c
273
if (cpuid_function_is_indexed(leaf->fn) && e->ecx_in != leaf->subfn)
arch/x86/boot/startup/sev-shared.c
282
if (e->eax_in == 0xD && (e->ecx_in == 0 || e->ecx_in == 1))
arch/x86/boot/startup/sev-shared.c
283
if (!(e->xcr0_in == 1 || e->xcr0_in == 3) || e->xss_in)
arch/x86/boot/startup/sev-shared.c
286
leaf->eax = e->eax;
arch/x86/boot/startup/sev-shared.c
287
leaf->ebx = e->ebx;
arch/x86/boot/startup/sev-shared.c
288
leaf->ecx = e->ecx;
arch/x86/boot/startup/sev-shared.c
289
leaf->edx = e->edx;
arch/x86/coco/sev/core.c
190
struct psc_entry *e;
arch/x86/coco/sev/core.c
199
e = &desc->entries[i];
arch/x86/coco/sev/core.c
201
pfn = e->gfn;
arch/x86/coco/sev/core.c
203
size = e->pagesize ? RMP_PG_SIZE_2M : RMP_PG_SIZE_4K;
arch/x86/coco/sev/core.c
204
validate = e->operation == SNP_PAGE_STATE_PRIVATE;
arch/x86/coco/sev/core.c
226
struct psc_entry *e;
arch/x86/coco/sev/core.c
242
e = &desc->entries[i];
arch/x86/coco/sev/core.c
248
if (e->operation == SNP_PAGE_STATE_PRIVATE)
arch/x86/coco/sev/core.c
249
sev_evict_cache(pfn_to_kaddr(e->gfn), e->pagesize ? 512 : 1);
arch/x86/coco/sev/core.c
325
struct psc_entry *e;
arch/x86/coco/sev/core.c
332
e = data->entries;
arch/x86/coco/sev/core.c
348
e->gfn = pfn;
arch/x86/coco/sev/core.c
349
e->operation = op;
arch/x86/coco/sev/core.c
353
e->pagesize = RMP_PG_SIZE_2M;
arch/x86/coco/sev/core.c
356
e->pagesize = RMP_PG_SIZE_4K;
arch/x86/coco/sev/core.c
360
e++;
arch/x86/coco/sev/svsm.c
130
struct psc_entry *e;
arch/x86/coco/sev/svsm.c
137
e = &desc->entries[desc_entry];
arch/x86/coco/sev/svsm.c
140
pe->page_size = e->pagesize ? RMP_PG_SIZE_2M : RMP_PG_SIZE_4K;
arch/x86/coco/sev/svsm.c
141
pe->action = e->operation == SNP_PAGE_STATE_PRIVATE;
arch/x86/coco/sev/svsm.c
144
pe->pfn = e->gfn;
arch/x86/coco/sev/svsm.c
147
e++;
arch/x86/coco/tdx/tdx.c
31
#define VE_IS_IO_IN(e) ((e) & BIT(3))
arch/x86/coco/tdx/tdx.c
32
#define VE_GET_IO_SIZE(e) (((e) & GENMASK(2, 0)) + 1)
arch/x86/coco/tdx/tdx.c
33
#define VE_GET_PORT_NUM(e) ((e) >> 16)
arch/x86/coco/tdx/tdx.c
34
#define VE_IS_IO_STRING(e) ((e) & BIT(4))
arch/x86/events/amd/brs.c
96
static inline int amd_is_brs_event(struct perf_event *e)
arch/x86/events/amd/brs.c
98
return (e->hw.config & AMD64_RAW_EVENT_MASK) == AMD_FAM19H_BRS_EVENT;
arch/x86/events/core.c
1008
struct perf_event *e;
arch/x86/events/core.c
1126
e = cpuc->event_list[i];
arch/x86/events/core.c
1131
static_call_cond(x86_pmu_put_event_constraints)(cpuc, e);
arch/x86/events/intel/lbr.c
904
struct perf_branch_entry *e;
arch/x86/events/intel/lbr.c
911
e = &cpuc->lbr_entries[i];
arch/x86/events/intel/lbr.c
923
perf_clear_branch_entry_bitfields(e);
arch/x86/events/intel/lbr.c
925
e->from = from;
arch/x86/events/intel/lbr.c
926
e->to = to;
arch/x86/events/intel/lbr.c
927
e->mispred = get_lbr_mispred(info);
arch/x86/events/intel/lbr.c
928
e->predicted = !e->mispred;
arch/x86/events/intel/lbr.c
929
e->in_tx = !!(info & LBR_INFO_IN_TX);
arch/x86/events/intel/lbr.c
930
e->abort = !!(info & LBR_INFO_ABORT);
arch/x86/events/intel/lbr.c
931
e->cycles = get_lbr_cycles(info);
arch/x86/events/intel/lbr.c
932
e->type = get_lbr_br_type(info);
arch/x86/events/intel/lbr.c
940
e->reserved = (info >> LBR_INFO_BR_CNTR_OFFSET) & LBR_INFO_BR_CNTR_FULL_MASK;
arch/x86/events/intel/uncore.h
36
#define UNCORE_EVENT_CONSTRAINT_RANGE(c, e, n) \
arch/x86/events/intel/uncore.h
37
EVENT_CONSTRAINT_RANGE(c, e, n, 0xff)
arch/x86/events/intel/uncore_snbep.c
104
#define SNBEP_CBO_EVENT_EXTRA_REG(e, m, i) { \
arch/x86/events/intel/uncore_snbep.c
105
.event = (e), \
arch/x86/events/perf_event.h
389
#define __EVENT_CONSTRAINT_RANGE(c, e, n, m, w, o, f) { \
arch/x86/events/perf_event.h
392
.size = (e) - (c), \
arch/x86/events/perf_event.h
409
#define EVENT_CONSTRAINT_RANGE(c, e, n, m) \
arch/x86/events/perf_event.h
410
__EVENT_CONSTRAINT_RANGE(c, e, n, m, HWEIGHT(n), 0, 0)
arch/x86/events/perf_event.h
449
#define INTEL_EVENT_CONSTRAINT_RANGE(c, e, n) \
arch/x86/events/perf_event.h
450
EVENT_CONSTRAINT_RANGE(c, e, n, ARCH_PERFMON_EVENTSEL_EVENT)
arch/x86/events/perf_event.h
528
#define INTEL_FLAGS_EVENT_CONSTRAINT_RANGE(c, e, n) \
arch/x86/events/perf_event.h
529
EVENT_CONSTRAINT_RANGE(c, e, n, ARCH_PERFMON_EVENTSEL_EVENT|X86_ALL_EVENT_FLAGS)
arch/x86/events/perf_event.h
601
#define for_each_event_constraint(e, c) \
arch/x86/events/perf_event.h
602
for ((e) = (c); (e)->weight != -1; (e)++)
arch/x86/events/perf_event.h
623
#define EVENT_EXTRA_REG(e, ms, m, vm, i) { \
arch/x86/events/perf_event.h
624
.event = (e), \
arch/x86/include/asm/asm.h
34
#define __ASM_REG(reg) __ASM_SEL_RAW(e##reg, r##reg)
arch/x86/include/asm/atomic.h
47
return GEN_BINARY_RMWcc(LOCK_PREFIX "subl", v->counter, e, "er", i);
arch/x86/include/asm/atomic.h
67
return GEN_UNARY_RMWcc(LOCK_PREFIX "decl", v->counter, e);
arch/x86/include/asm/atomic.h
73
return GEN_UNARY_RMWcc(LOCK_PREFIX "incl", v->counter, e);
arch/x86/include/asm/atomic64_64.h
39
return GEN_BINARY_RMWcc(LOCK_PREFIX "subq", v->counter, e, "er", i);
arch/x86/include/asm/atomic64_64.h
61
return GEN_UNARY_RMWcc(LOCK_PREFIX "decq", v->counter, e);
arch/x86/include/asm/atomic64_64.h
67
return GEN_UNARY_RMWcc(LOCK_PREFIX "incq", v->counter, e);
arch/x86/include/asm/kvm_host.h
2182
#define EMULTYPE_GET_SOFT_INT_VECTOR(e) (((e) >> 16) & 0xff)
arch/x86/include/asm/local.h
56
return GEN_BINARY_RMWcc(_ASM_SUB, l->a.counter, e, "er", i);
arch/x86/include/asm/local.h
69
return GEN_UNARY_RMWcc(_ASM_DEC, l->a.counter, e);
arch/x86/include/asm/local.h
82
return GEN_UNARY_RMWcc(_ASM_INC, l->a.counter, e);
arch/x86/include/asm/pgtable-2level.h
5
#define pte_ERROR(e) \
arch/x86/include/asm/pgtable-2level.h
6
pr_err("%s:%d: bad pte %08lx\n", __FILE__, __LINE__, (e).pte_low)
arch/x86/include/asm/pgtable-2level.h
7
#define pgd_ERROR(e) \
arch/x86/include/asm/pgtable-2level.h
8
pr_err("%s:%d: bad pgd %08lx\n", __FILE__, __LINE__, pgd_val(e))
arch/x86/include/asm/pgtable-3level.h
12
#define pte_ERROR(e) \
arch/x86/include/asm/pgtable-3level.h
14
__FILE__, __LINE__, &(e), (e).pte_high, (e).pte_low)
arch/x86/include/asm/pgtable-3level.h
15
#define pmd_ERROR(e) \
arch/x86/include/asm/pgtable-3level.h
17
__FILE__, __LINE__, &(e), pmd_val(e))
arch/x86/include/asm/pgtable-3level.h
18
#define pgd_ERROR(e) \
arch/x86/include/asm/pgtable-3level.h
20
__FILE__, __LINE__, &(e), pgd_val(e))
arch/x86/include/asm/pgtable_64.h
32
#define pte_ERROR(e) \
arch/x86/include/asm/pgtable_64.h
34
__FILE__, __LINE__, &(e), pte_val(e))
arch/x86/include/asm/pgtable_64.h
35
#define pmd_ERROR(e) \
arch/x86/include/asm/pgtable_64.h
37
__FILE__, __LINE__, &(e), pmd_val(e))
arch/x86/include/asm/pgtable_64.h
38
#define pud_ERROR(e) \
arch/x86/include/asm/pgtable_64.h
40
__FILE__, __LINE__, &(e), pud_val(e))
arch/x86/include/asm/pgtable_64.h
42
#define p4d_ERROR(e) \
arch/x86/include/asm/pgtable_64.h
44
__FILE__, __LINE__, &(e), p4d_val(e))
arch/x86/include/asm/pgtable_64.h
46
#define pgd_ERROR(e) \
arch/x86/include/asm/pgtable_64.h
48
__FILE__, __LINE__, &(e), pgd_val(e))
arch/x86/include/asm/preempt.h
95
return GEN_UNARY_RMWcc("decl", __my_cpu_var(__preempt_count), e,
arch/x86/include/asm/xen/interface_64.h
91
uint64_t r ## name, e ## name; \
arch/x86/kernel/apic/io_apic.c
1587
struct IO_APIC_route_entry e;
arch/x86/kernel/apic/io_apic.c
1591
e.w1 = io_apic_read(entry->apic, 0x10 + pin*2);
arch/x86/kernel/apic/io_apic.c
1593
if (e.irr)
arch/x86/kernel/apic/io_apic.c
310
static void __ioapic_write_entry(int apic, int pin, struct IO_APIC_route_entry e)
arch/x86/kernel/apic/io_apic.c
312
io_apic_write(apic, 0x11 + 2*pin, e.w2);
arch/x86/kernel/apic/io_apic.c
313
io_apic_write(apic, 0x10 + 2*pin, e.w1);
arch/x86/kernel/apic/io_apic.c
316
static void ioapic_write_entry(int apic, int pin, struct IO_APIC_route_entry e)
arch/x86/kernel/apic/io_apic.c
319
__ioapic_write_entry(apic, pin, e);
arch/x86/kernel/apic/io_apic.c
329
struct IO_APIC_route_entry e = { .masked = true };
arch/x86/kernel/apic/io_apic.c
332
io_apic_write(apic, 0x10 + 2*pin, e.w1);
arch/x86/kernel/apic/io_apic.c
333
io_apic_write(apic, 0x11 + 2*pin, e.w2);
arch/x86/kernel/cpu/microcode/amd.c
357
struct equiv_cpu_entry *e = &et->entry[i];
arch/x86/kernel/cpu/microcode/amd.c
359
if (sig == e->installed_cpu)
arch/x86/kernel/cpu/microcode/amd.c
360
return e->equiv_cpu;
arch/x86/kernel/cpu/resctrl/intel_aet.c
167
static bool skip_telem_region(struct telemetry_region *tr, struct event_group *e)
arch/x86/kernel/cpu/resctrl/intel_aet.c
169
if (tr->guid != e->guid)
arch/x86/kernel/cpu/resctrl/intel_aet.c
176
if (tr->size != e->mmio_size) {
arch/x86/kernel/cpu/resctrl/intel_aet.c
178
tr->size, e->guid, e->mmio_size);
arch/x86/kernel/cpu/resctrl/intel_aet.c
185
static bool group_has_usable_regions(struct event_group *e, struct pmt_feature_group *p)
arch/x86/kernel/cpu/resctrl/intel_aet.c
190
if (skip_telem_region(&p->regions[i], e)) {
arch/x86/kernel/cpu/resctrl/intel_aet.c
209
static bool all_regions_have_sufficient_rmid(struct event_group *e, struct pmt_feature_group *p)
arch/x86/kernel/cpu/resctrl/intel_aet.c
217
if (tr->num_rmids < e->num_rmid) {
arch/x86/kernel/cpu/resctrl/intel_aet.c
218
e->force_off = true;
arch/x86/kernel/cpu/resctrl/intel_aet.c
226
static bool enable_events(struct event_group *e, struct pmt_feature_group *p)
arch/x86/kernel/cpu/resctrl/intel_aet.c
231
if (e->force_off)
arch/x86/kernel/cpu/resctrl/intel_aet.c
234
if (!group_has_usable_regions(e, p))
arch/x86/kernel/cpu/resctrl/intel_aet.c
241
if (!all_regions_have_sufficient_rmid(e, p) && !e->force_on) {
arch/x86/kernel/cpu/resctrl/intel_aet.c
243
r->name, e->pfname, e->guid);
arch/x86/kernel/cpu/resctrl/intel_aet.c
255
e->num_rmid = min(e->num_rmid, p->regions[i].num_rmids);
arch/x86/kernel/cpu/resctrl/intel_aet.c
258
for (int j = 0; j < e->num_events; j++) {
arch/x86/kernel/cpu/resctrl/intel_aet.c
259
if (!resctrl_enable_mon_event(e->evts[j].id, true,
arch/x86/kernel/cpu/resctrl/intel_aet.c
260
e->evts[j].bin_bits, &e->evts[j]))
arch/x86/kernel/cpu/resctrl/intel_aet.c
263
if (e->num_events == skipped_events) {
arch/x86/kernel/cpu/resctrl/intel_aet.c
264
pr_info("No events enabled in %s %s:0x%x\n", r->name, e->pfname, e->guid);
arch/x86/kernel/cpu/resctrl/intel_aet.c
269
r->mon.num_rmid = min(r->mon.num_rmid, e->num_rmid);
arch/x86/kernel/cpu/resctrl/intel_aet.c
271
r->mon.num_rmid = e->num_rmid;
arch/x86/kernel/cpu/resctrl/intel_aet.c
275
e->pfname, e->guid, skipped_events);
arch/x86/kernel/cpu/resctrl/intel_aet.c
277
pr_info("%s %s:0x%x monitoring detected\n", r->name, e->pfname, e->guid);
arch/x86/kernel/cpu/resctrl/intel_aet.c
352
struct event_group *e;
arch/x86/kernel/cpu/resctrl/intel_aet.c
360
e = container_of(pevt0, struct event_group, evts);
arch/x86/kernel/cpu/resctrl/intel_aet.c
361
idx = rmid * e->num_events;
arch/x86/kernel/cpu/resctrl/intel_aet.c
364
if (idx * sizeof(u64) + sizeof(u64) > e->mmio_size) {
arch/x86/kernel/cpu/resctrl/intel_aet.c
369
for (int i = 0; i < e->pfg->count; i++) {
arch/x86/kernel/cpu/resctrl/intel_aet.c
370
if (!e->pfg->regions[i].addr)
arch/x86/kernel/cpu/resctrl/intel_aet.c
372
if (e->pfg->regions[i].plat_info.package_id != domid)
arch/x86/kernel/cpu/resctrl/intel_aet.c
374
evtcount = readq(e->pfg->regions[i].addr + idx * sizeof(u64));
arch/x86/kernel/early_printk.c
161
char *e;
arch/x86/kernel/early_printk.c
169
early_serial_base = simple_strtoul(s, &e, 16);
arch/x86/kernel/early_printk.c
175
port = simple_strtoul(s, &e, 10);
arch/x86/kernel/early_printk.c
176
if (port > 1 || s == e)
arch/x86/kernel/early_printk.c
186
baud = simple_strtoull(s, &e, 0);
arch/x86/kernel/early_printk.c
188
if (baud == 0 || s == e)
arch/x86/kernel/early_printk.c
223
char *e;
arch/x86/kernel/early_printk.c
230
membase = simple_strtoul(s, &e, 16);
arch/x86/kernel/early_printk.c
244
baudrate = simple_strtoul(s, &e, 0);
arch/x86/kernel/early_printk.c
245
if (baudrate == 0 || s == e)
arch/x86/kernel/early_printk.c
269
char *e;
arch/x86/kernel/early_printk.c
287
bus = (u8)simple_strtoul(s, &e, 16);
arch/x86/kernel/early_printk.c
288
s = e;
arch/x86/kernel/early_printk.c
292
slot = (u8)simple_strtoul(s, &e, 16);
arch/x86/kernel/early_printk.c
293
s = e;
arch/x86/kernel/early_printk.c
297
func = (u8)simple_strtoul(s, &e, 16);
arch/x86/kernel/early_printk.c
298
s = e;
arch/x86/kernel/kvm.c
120
struct kvm_task_sleep_node *e;
arch/x86/kernel/kvm.c
123
e = _find_apf_task(b, token);
arch/x86/kernel/kvm.c
124
if (e) {
arch/x86/kernel/kvm.c
132
if (e->dummy) {
arch/x86/kernel/kvm.c
133
hlist_del(&e->link);
arch/x86/kernel/kvm.c
134
dummy = e;
arch/x86/kvm/cpuid.c
106
e = &entries[i];
arch/x86/kvm/cpuid.c
108
if (e->function != function)
arch/x86/kvm/cpuid.c
116
if (!(e->flags & KVM_CPUID_FLAG_SIGNIFCANT_INDEX) || e->index == index)
arch/x86/kvm/cpuid.c
117
return e;
arch/x86/kvm/cpuid.c
131
return e;
arch/x86/kvm/cpuid.c
594
struct kvm_cpuid_entry *e = NULL;
arch/x86/kvm/cpuid.c
601
e = vmemdup_array_user(entries, cpuid->nent, sizeof(*e));
arch/x86/kvm/cpuid.c
602
if (IS_ERR(e))
arch/x86/kvm/cpuid.c
603
return PTR_ERR(e);
arch/x86/kvm/cpuid.c
612
e2[i].function = e[i].function;
arch/x86/kvm/cpuid.c
613
e2[i].eax = e[i].eax;
arch/x86/kvm/cpuid.c
614
e2[i].ebx = e[i].ebx;
arch/x86/kvm/cpuid.c
615
e2[i].ecx = e[i].ecx;
arch/x86/kvm/cpuid.c
616
e2[i].edx = e[i].edx;
arch/x86/kvm/cpuid.c
629
kvfree(e);
arch/x86/kvm/cpuid.c
90
struct kvm_cpuid_entry2 *e;
arch/x86/kvm/hyperv.c
500
int kvm_hv_synic_set_irq(struct kvm_kernel_irq_routing_entry *e, struct kvm *kvm,
arch/x86/kvm/hyperv.c
508
synic = synic_get(kvm, e->hv_sint.vcpu);
arch/x86/kvm/hyperv.c
512
return synic_set_irq(synic, e->hv_sint.sint);
arch/x86/kvm/hyperv.c
545
struct kvm_kernel_irq_routing_entry *e;
arch/x86/kvm/hyperv.c
552
hlist_for_each_entry(e, &irq_rt->map[gsi], link) {
arch/x86/kvm/hyperv.c
553
if (e->type == KVM_IRQ_ROUTING_HV_SINT)
arch/x86/kvm/hyperv.c
554
kvm_hv_set_sint_gsi(kvm, e->hv_sint.vcpu,
arch/x86/kvm/hyperv.c
555
e->hv_sint.sint, gsi);
arch/x86/kvm/hyperv.h
106
int kvm_hv_synic_set_irq(struct kvm_kernel_irq_routing_entry *e, struct kvm *kvm,
arch/x86/kvm/i8259.c
190
int kvm_pic_set_irq(struct kvm_kernel_irq_routing_entry *e, struct kvm *kvm,
arch/x86/kvm/i8259.c
194
int irq = e->irqchip.pin;
arch/x86/kvm/ioapic.c
100
e->fields.dest_id,
arch/x86/kvm/ioapic.c
101
kvm_lapic_irq_dest_mode(!!e->fields.dest_mode)))
arch/x86/kvm/ioapic.c
104
new_val = kvm_apic_pending_eoi(vcpu, e->fields.vector);
arch/x86/kvm/ioapic.c
112
status->vectors[vcpu->vcpu_id] = e->fields.vector;
arch/x86/kvm/ioapic.c
263
union kvm_ioapic_redirect_entry *e;
arch/x86/kvm/ioapic.c
274
e = &ioapic->redirtbl[index];
arch/x86/kvm/ioapic.c
275
if (e->fields.trig_mode == IOAPIC_LEVEL_TRIG ||
arch/x86/kvm/ioapic.c
278
u16 dm = kvm_lapic_irq_dest_mode(!!e->fields.dest_mode);
arch/x86/kvm/ioapic.c
280
kvm_scan_ioapic_irq(vcpu, e->fields.dest_id, dm,
arch/x86/kvm/ioapic.c
281
e->fields.vector, ioapic_handled_vectors);
arch/x86/kvm/ioapic.c
335
union kvm_ioapic_redirect_entry *e;
arch/x86/kvm/ioapic.c
357
e = &ioapic->redirtbl[index];
arch/x86/kvm/ioapic.c
358
mask_before = e->fields.mask;
arch/x86/kvm/ioapic.c
360
old_remote_irr = e->fields.remote_irr;
arch/x86/kvm/ioapic.c
361
old_delivery_status = e->fields.delivery_status;
arch/x86/kvm/ioapic.c
362
old_dest_id = e->fields.dest_id;
arch/x86/kvm/ioapic.c
363
old_dest_mode = e->fields.dest_mode;
arch/x86/kvm/ioapic.c
365
e->bits &= 0xffffffff;
arch/x86/kvm/ioapic.c
366
e->bits |= (u64) val << 32;
arch/x86/kvm/ioapic.c
368
e->bits &= ~0xffffffffULL;
arch/x86/kvm/ioapic.c
369
e->bits |= (u32) val;
arch/x86/kvm/ioapic.c
371
e->fields.remote_irr = old_remote_irr;
arch/x86/kvm/ioapic.c
372
e->fields.delivery_status = old_delivery_status;
arch/x86/kvm/ioapic.c
380
if (e->fields.trig_mode == IOAPIC_EDGE_TRIG)
arch/x86/kvm/ioapic.c
381
e->fields.remote_irr = 0;
arch/x86/kvm/ioapic.c
383
mask_after = e->fields.mask;
arch/x86/kvm/ioapic.c
386
if (e->fields.trig_mode == IOAPIC_LEVEL_TRIG &&
arch/x86/kvm/ioapic.c
387
ioapic->irr & (1 << index) && !e->fields.mask && !e->fields.remote_irr) {
arch/x86/kvm/ioapic.c
419
if (e->fields.delivery_mode == APIC_DM_FIXED) {
arch/x86/kvm/ioapic.c
422
irq.vector = e->fields.vector;
arch/x86/kvm/ioapic.c
423
irq.delivery_mode = e->fields.delivery_mode << 8;
arch/x86/kvm/ioapic.c
425
kvm_lapic_irq_dest_mode(!!e->fields.dest_mode);
arch/x86/kvm/ioapic.c
427
irq.trig_mode = e->fields.trig_mode;
arch/x86/kvm/ioapic.c
429
irq.dest_id = e->fields.dest_id;
arch/x86/kvm/ioapic.c
434
if (old_dest_mode != e->fields.dest_mode ||
arch/x86/kvm/ioapic.c
435
old_dest_id != e->fields.dest_id) {
arch/x86/kvm/ioapic.c
444
!!e->fields.dest_mode);
arch/x86/kvm/ioapic.c
500
int kvm_ioapic_set_irq(struct kvm_kernel_irq_routing_entry *e, struct kvm *kvm,
arch/x86/kvm/ioapic.c
504
int irq = e->irqchip.pin;
arch/x86/kvm/ioapic.c
96
union kvm_ioapic_redirect_entry *e;
arch/x86/kvm/ioapic.c
98
e = &ioapic->redirtbl[RTC_GSI];
arch/x86/kvm/ioapic.h
109
int kvm_ioapic_set_irq(struct kvm_kernel_irq_routing_entry *e, struct kvm *kvm,
arch/x86/kvm/irq.c
199
struct kvm_kernel_irq_routing_entry *e,
arch/x86/kvm/irq.c
202
struct msi_msg msg = { .address_lo = e->msi.address_lo,
arch/x86/kvm/irq.c
203
.address_hi = e->msi.address_hi,
arch/x86/kvm/irq.c
204
.data = e->msi.data };
arch/x86/kvm/irq.c
220
struct kvm_kernel_irq_routing_entry *e)
arch/x86/kvm/irq.c
222
return kvm->arch.x2apic_format && (e->msi.address_hi & 0xff);
arch/x86/kvm/irq.c
225
int kvm_set_msi(struct kvm_kernel_irq_routing_entry *e,
arch/x86/kvm/irq.c
230
if (kvm_msi_route_invalid(kvm, e))
arch/x86/kvm/irq.c
236
kvm_msi_to_lapic_irq(kvm, e, &irq);
arch/x86/kvm/irq.c
241
int kvm_arch_set_irq_inatomic(struct kvm_kernel_irq_routing_entry *e,
arch/x86/kvm/irq.c
248
switch (e->type) {
arch/x86/kvm/irq.c
251
return kvm_hv_synic_set_irq(e, kvm, irq_source_id, level,
arch/x86/kvm/irq.c
256
if (kvm_msi_route_invalid(kvm, e))
arch/x86/kvm/irq.c
259
kvm_msi_to_lapic_irq(kvm, e, &irq);
arch/x86/kvm/irq.c
270
return kvm_xen_set_evtchn_fast(&e->xen_evtchn, kvm);
arch/x86/kvm/irq.c
297
struct kvm_kernel_irq_routing_entry *e,
arch/x86/kvm/irq.c
309
e->irqchip.pin = ue->u.irqchip.pin;
arch/x86/kvm/irq.c
312
e->irqchip.pin += PIC_NUM_PINS / 2;
arch/x86/kvm/irq.c
317
e->set = kvm_pic_set_irq;
arch/x86/kvm/irq.c
322
e->set = kvm_ioapic_set_irq;
arch/x86/kvm/irq.c
327
e->irqchip.irqchip = ue->u.irqchip.irqchip;
arch/x86/kvm/irq.c
331
e->set = kvm_set_msi;
arch/x86/kvm/irq.c
332
e->msi.address_lo = ue->u.msi.address_lo;
arch/x86/kvm/irq.c
333
e->msi.address_hi = ue->u.msi.address_hi;
arch/x86/kvm/irq.c
334
e->msi.data = ue->u.msi.data;
arch/x86/kvm/irq.c
336
if (kvm_msi_route_invalid(kvm, e))
arch/x86/kvm/irq.c
341
e->set = kvm_hv_synic_set_irq;
arch/x86/kvm/irq.c
342
e->hv_sint.vcpu = ue->u.hv_sint.vcpu;
arch/x86/kvm/irq.c
343
e->hv_sint.sint = ue->u.hv_sint.sint;
arch/x86/kvm/irq.c
348
return kvm_xen_setup_evtchn(kvm, e, ue);
arch/x86/kvm/irq.h
68
int kvm_pic_set_irq(struct kvm_kernel_irq_routing_entry *e, struct kvm *kvm,
arch/x86/kvm/mmu.h
46
static __always_inline u64 rsvd_bits(int s, int e)
arch/x86/kvm/mmu.h
48
BUILD_BUG_ON(__builtin_constant_p(e) && __builtin_constant_p(s) && e < s);
arch/x86/kvm/mmu.h
50
if (__builtin_constant_p(e))
arch/x86/kvm/mmu.h
51
BUILD_BUG_ON(e > 63);
arch/x86/kvm/mmu.h
53
e &= 63;
arch/x86/kvm/mmu.h
55
if (e < s)
arch/x86/kvm/mmu.h
58
return ((2ULL << (e - s)) - 1) << s;
arch/x86/kvm/trace.h
275
TP_PROTO(__u64 e, int pin, bool coalesced),
arch/x86/kvm/trace.h
276
TP_ARGS(e, pin, coalesced),
arch/x86/kvm/trace.h
279
__field( __u64, e )
arch/x86/kvm/trace.h
285
__entry->e = e;
arch/x86/kvm/trace.h
291
__entry->pin, (u8)(__entry->e >> 56), (u8)__entry->e,
arch/x86/kvm/trace.h
292
__print_symbolic((__entry->e >> 8 & 0x7), kvm_deliver_mode),
arch/x86/kvm/trace.h
293
(__entry->e & (1<<11)) ? "logical" : "physical",
arch/x86/kvm/trace.h
294
(__entry->e & (1<<15)) ? "level" : "edge",
arch/x86/kvm/trace.h
295
(__entry->e & (1<<16)) ? "|masked" : "",
arch/x86/kvm/trace.h
300
TP_PROTO(__u64 e),
arch/x86/kvm/trace.h
301
TP_ARGS(e),
arch/x86/kvm/trace.h
304
__field( __u64, e )
arch/x86/kvm/trace.h
308
__entry->e = e;
arch/x86/kvm/trace.h
312
(u8)(__entry->e >> 56), (u8)__entry->e,
arch/x86/kvm/trace.h
313
__print_symbolic((__entry->e >> 8 & 0x7), kvm_deliver_mode),
arch/x86/kvm/trace.h
314
(__entry->e & (1<<11)) ? "logical" : "physical",
arch/x86/kvm/trace.h
315
(__entry->e & (1<<15)) ? "level" : "edge",
arch/x86/kvm/trace.h
316
(__entry->e & (1<<16)) ? "|masked" : "")
arch/x86/kvm/vmx/nested.c
1001
if (CC(e->reserved != 0))
arch/x86/kvm/vmx/nested.c
1007
struct vmx_msr_entry *e)
arch/x86/kvm/vmx/nested.c
1009
if (CC(e->index == MSR_FS_BASE) ||
arch/x86/kvm/vmx/nested.c
1010
CC(e->index == MSR_GS_BASE) ||
arch/x86/kvm/vmx/nested.c
1011
CC(e->index == MSR_IA32_SMM_MONITOR_CTL) || /* SMM is not supported */
arch/x86/kvm/vmx/nested.c
1012
nested_vmx_msr_check_common(vcpu, e))
arch/x86/kvm/vmx/nested.c
1018
struct vmx_msr_entry *e)
arch/x86/kvm/vmx/nested.c
1020
if (CC(e->index == MSR_IA32_SMBASE) || /* SMM is not supported */
arch/x86/kvm/vmx/nested.c
1021
nested_vmx_msr_check_common(vcpu, e))
arch/x86/kvm/vmx/nested.c
1038
struct vmx_msr_entry e;
arch/x86/kvm/vmx/nested.c
1045
if (kvm_vcpu_read_guest(vcpu, gpa + i * sizeof(e),
arch/x86/kvm/vmx/nested.c
1046
&e, sizeof(e))) {
arch/x86/kvm/vmx/nested.c
1049
__func__, i, gpa + i * sizeof(e));
arch/x86/kvm/vmx/nested.c
1052
if (nested_vmx_load_msr_check(vcpu, &e)) {
arch/x86/kvm/vmx/nested.c
1055
__func__, i, e.index, e.reserved);
arch/x86/kvm/vmx/nested.c
1058
if (kvm_emulate_msr_write(vcpu, e.index, e.value)) {
arch/x86/kvm/vmx/nested.c
1061
__func__, i, e.index, e.value);
arch/x86/kvm/vmx/nested.c
1099
struct vmx_msr_entry *e)
arch/x86/kvm/vmx/nested.c
1102
gpa + i * sizeof(*e),
arch/x86/kvm/vmx/nested.c
1103
e, 2 * sizeof(u32))) {
arch/x86/kvm/vmx/nested.c
1106
__func__, i, gpa + i * sizeof(*e));
arch/x86/kvm/vmx/nested.c
1109
if (nested_vmx_store_msr_check(vcpu, e)) {
arch/x86/kvm/vmx/nested.c
1112
__func__, i, e->index, e->reserved);
arch/x86/kvm/vmx/nested.c
1122
struct vmx_msr_entry e;
arch/x86/kvm/vmx/nested.c
1129
if (!read_and_check_msr_entry(vcpu, gpa, i, &e))
arch/x86/kvm/vmx/nested.c
1132
if (!nested_vmx_get_vmexit_msr_value(vcpu, e.index, &data))
arch/x86/kvm/vmx/nested.c
1136
gpa + i * sizeof(e) +
arch/x86/kvm/vmx/nested.c
1141
__func__, i, e.index, data);
arch/x86/kvm/vmx/nested.c
1153
struct vmx_msr_entry e;
arch/x86/kvm/vmx/nested.c
1157
if (!read_and_check_msr_entry(vcpu, gpa, i, &e))
arch/x86/kvm/vmx/nested.c
1160
if (e.index == msr_index)
arch/x86/kvm/vmx/nested.c
5347
struct x86_exception e;
arch/x86/kvm/vmx/nested.c
5357
r = kvm_read_guest_virt(vcpu, gva, vmpointer, sizeof(*vmpointer), &e);
arch/x86/kvm/vmx/nested.c
5359
*ret = kvm_handle_memory_failure(vcpu, r, &e);
arch/x86/kvm/vmx/nested.c
5638
struct x86_exception e;
arch/x86/kvm/vmx/nested.c
5705
r = kvm_write_guest_virt_system(vcpu, gva, &value, len, &e);
arch/x86/kvm/vmx/nested.c
5707
return kvm_handle_memory_failure(vcpu, r, &e);
arch/x86/kvm/vmx/nested.c
5744
struct x86_exception e;
arch/x86/kvm/vmx/nested.c
5778
r = kvm_read_guest_virt(vcpu, gva, &value, len, &e);
arch/x86/kvm/vmx/nested.c
5780
return kvm_handle_memory_failure(vcpu, r, &e);
arch/x86/kvm/vmx/nested.c
5933
struct x86_exception e;
arch/x86/kvm/vmx/nested.c
5948
sizeof(gpa_t), &e);
arch/x86/kvm/vmx/nested.c
5950
return kvm_handle_memory_failure(vcpu, r, &e);
arch/x86/kvm/vmx/nested.c
5963
struct x86_exception e;
arch/x86/kvm/vmx/nested.c
5994
r = kvm_read_guest_virt(vcpu, gva, &operand, sizeof(operand), &e);
arch/x86/kvm/vmx/nested.c
5996
return kvm_handle_memory_failure(vcpu, r, &e);
arch/x86/kvm/vmx/nested.c
6042
struct x86_exception e;
arch/x86/kvm/vmx/nested.c
6077
r = kvm_read_guest_virt(vcpu, gva, &operand, sizeof(operand), &e);
arch/x86/kvm/vmx/nested.c
6079
return kvm_handle_memory_failure(vcpu, r, &e);
arch/x86/kvm/vmx/nested.c
993
struct vmx_msr_entry *e)
arch/x86/kvm/vmx/nested.c
996
if (CC(vcpu->arch.apic_base & X2APIC_ENABLE && e->index >> 8 == 0x8))
arch/x86/kvm/vmx/nested.c
998
if (CC(e->index == MSR_IA32_UCODE_WRITE) || /* SDM Table 35-2 */
arch/x86/kvm/vmx/nested.c
999
CC(e->index == MSR_IA32_UCODE_REV))
arch/x86/kvm/vmx/vmx.c
6569
struct vmx_msr_entry *e;
arch/x86/kvm/vmx/vmx.c
6572
for (i = 0, e = m->val; i < m->nr; ++i, ++e)
arch/x86/kvm/vmx/vmx.c
6573
pr_err(" %2d: msr=0x%08x value=0x%016llx\n", i, e->index, e->value);
arch/x86/kvm/x86.c
104
#define ERR_PTR_USR(e) ((void __user *)ERR_PTR(e))
arch/x86/kvm/x86.c
14141
struct x86_exception *e)
arch/x86/kvm/x86.c
14144
if (KVM_BUG_ON(!e, vcpu->kvm))
arch/x86/kvm/x86.c
14147
kvm_inject_emulated_page_fault(vcpu, e);
arch/x86/kvm/x86.c
14167
struct x86_exception e;
arch/x86/kvm/x86.c
14174
r = kvm_read_guest_virt(vcpu, gva, &operand, sizeof(operand), &e);
arch/x86/kvm/x86.c
14176
return kvm_handle_memory_failure(vcpu, r, &e);
arch/x86/kvm/x86.c
8032
struct x86_exception e;
arch/x86/kvm/x86.c
8041
sig, sizeof(sig), &e) == 0 &&
arch/x86/kvm/x86.h
655
struct x86_exception *e);
arch/x86/kvm/xen.c
111
struct kvm_xen_evtchn e;
arch/x86/kvm/xen.c
113
e.vcpu_id = vcpu->vcpu_id;
arch/x86/kvm/xen.c
114
e.vcpu_idx = vcpu->vcpu_idx;
arch/x86/kvm/xen.c
115
e.port = vcpu->arch.xen.timer_virq;
arch/x86/kvm/xen.c
116
e.priority = KVM_IRQ_ROUTING_XEN_EVTCHN_PRIO_2LEVEL;
arch/x86/kvm/xen.c
118
kvm_xen_set_evtchn(&e, vcpu->kvm);
arch/x86/kvm/xen.c
129
struct kvm_xen_evtchn e;
arch/x86/kvm/xen.c
135
e.vcpu_id = vcpu->vcpu_id;
arch/x86/kvm/xen.c
136
e.vcpu_idx = vcpu->vcpu_idx;
arch/x86/kvm/xen.c
137
e.port = vcpu->arch.xen.timer_virq;
arch/x86/kvm/xen.c
138
e.priority = KVM_IRQ_ROUTING_XEN_EVTCHN_PRIO_2LEVEL;
arch/x86/kvm/xen.c
140
rc = kvm_xen_set_evtchn_fast(&e, vcpu->kvm);
arch/x86/kvm/xen.c
1476
struct x86_exception e;
arch/x86/kvm/xen.c
1489
if (kvm_read_guest_virt(vcpu, param, &sp32, sizeof(sp32), &e)) {
arch/x86/kvm/xen.c
1504
sizeof(sched_poll), &e)) {
arch/x86/kvm/xen.c
1526
sched_poll.nr_ports * sizeof(*ports), &e)) {
arch/x86/kvm/xen.c
1608
struct x86_exception e;
arch/x86/kvm/xen.c
1637
sizeof(struct compat_vcpu_set_singleshot_timer), &e)) {
arch/x86/kvm/xen.c
1953
static int evtchn_set_fn(struct kvm_kernel_irq_routing_entry *e, struct kvm *kvm,
arch/x86/kvm/xen.c
1959
return kvm_xen_set_evtchn(&e->xen_evtchn, kvm);
arch/x86/kvm/xen.c
1967
struct kvm_kernel_irq_routing_entry *e,
arch/x86/kvm/xen.c
2001
e->xen_evtchn.vcpu_idx = vcpu->vcpu_idx;
arch/x86/kvm/xen.c
2003
e->xen_evtchn.vcpu_idx = -1;
arch/x86/kvm/xen.c
2005
e->xen_evtchn.port = ue->u.xen_evtchn.port;
arch/x86/kvm/xen.c
2006
e->xen_evtchn.vcpu_id = ue->u.xen_evtchn.vcpu;
arch/x86/kvm/xen.c
2007
e->xen_evtchn.priority = ue->u.xen_evtchn.priority;
arch/x86/kvm/xen.c
2008
e->set = evtchn_set_fn;
arch/x86/kvm/xen.c
2018
struct kvm_xen_evtchn e;
arch/x86/kvm/xen.c
2028
e.port = uxe->port;
arch/x86/kvm/xen.c
2029
e.vcpu_id = uxe->vcpu;
arch/x86/kvm/xen.c
2030
e.vcpu_idx = -1;
arch/x86/kvm/xen.c
2031
e.priority = uxe->priority;
arch/x86/kvm/xen.c
2033
ret = kvm_xen_set_evtchn(&e, kvm);
arch/x86/kvm/xen.c
2266
struct x86_exception e;
arch/x86/kvm/xen.c
2270
if (kvm_read_guest_virt(vcpu, param, &send, sizeof(send), &e)) {
arch/x86/kvm/xen.h
37
struct kvm_kernel_irq_routing_entry *e,
arch/x86/math-emu/fpu_trig.c
1031
int e, tag;
arch/x86/math-emu/fpu_trig.c
1049
e = exponent16(st0_ptr);
arch/x86/math-emu/fpu_trig.c
1050
if (e >= 0) {
arch/x86/math-emu/fpu_trig.c
1051
exponent.sigh = e;
arch/x86/math-emu/fpu_trig.c
1054
exponent.sigh = -e;
arch/x86/math-emu/fpu_trig.c
369
long e;
arch/x86/math-emu/fpu_trig.c
378
e = exponent16(st_new_ptr);
arch/x86/math-emu/fpu_trig.c
379
convert_l2reg(&e, 1);
arch/x86/math-emu/reg_constant.c
20
#define MAKE_REG(s, e, l, h) { l, h, \
arch/x86/math-emu/reg_constant.c
21
(u16)((EXTENDED_Ebias+(e)) | ((SIGN_##s != 0)*0x8000)) }
arch/x86/mm/extable.c
294
const struct exception_table_entry *e = search_exception_tables(ip);
arch/x86/mm/extable.c
296
return e ? FIELD_GET(EX_DATA_TYPE_MASK, e->data) : EX_TYPE_NONE;
arch/x86/mm/extable.c
302
const struct exception_table_entry *e;
arch/x86/mm/extable.c
319
e = search_exception_tables(regs->ip);
arch/x86/mm/extable.c
320
if (!e)
arch/x86/mm/extable.c
323
type = FIELD_GET(EX_DATA_TYPE_MASK, e->data);
arch/x86/mm/extable.c
324
reg = FIELD_GET(EX_DATA_REG_MASK, e->data);
arch/x86/mm/extable.c
325
imm = FIELD_GET(EX_DATA_IMM_MASK, e->data);
arch/x86/mm/extable.c
33
static bool ex_handler_default(const struct exception_table_entry *e,
arch/x86/mm/extable.c
330
return ex_handler_default(e, regs);
arch/x86/mm/extable.c
333
return ex_handler_fault(e, regs, trapnr);
arch/x86/mm/extable.c
335
return ex_handler_uaccess(e, regs, trapnr, fault_addr);
arch/x86/mm/extable.c
337
return ex_handler_clear_fs(e, regs);
arch/x86/mm/extable.c
339
return ex_handler_fprestore(e, regs);
arch/x86/mm/extable.c
341
return ex_handler_bpf(e, regs);
arch/x86/mm/extable.c
343
return ex_handler_msr(e, regs, true, false, reg);
arch/x86/mm/extable.c
345
return ex_handler_msr(e, regs, false, false, reg);
arch/x86/mm/extable.c
347
return ex_handler_msr(e, regs, true, true, reg);
arch/x86/mm/extable.c
349
return ex_handler_msr(e, regs, false, true, reg);
arch/x86/mm/extable.c
36
if (e->data & EX_FLAG_CLEAR_AX)
arch/x86/mm/extable.c
360
return ex_handler_imm_reg(e, regs, reg, imm);
arch/x86/mm/extable.c
362
return ex_handler_sgx(e, regs, trapnr);
arch/x86/mm/extable.c
364
return ex_handler_ucopy_len(e, regs, trapnr, fault_addr, reg, imm);
arch/x86/mm/extable.c
366
return ex_handler_zeropad(e, regs, fault_addr);
arch/x86/mm/extable.c
369
return ex_handler_eretu(e, regs, error_code);
arch/x86/mm/extable.c
38
if (e->data & EX_FLAG_CLEAR_DX)
arch/x86/mm/extable.c
41
regs->ip = ex_fixup_addr(e);
arch/x86/mm/extable.c
57
static bool ex_handler_zeropad(const struct exception_table_entry *e,
arch/x86/mm/extable.c
66
next_ip = ex_fixup_addr(e);
arch/x86/mm/extable.c
95
return ex_handler_default(e, regs);
arch/x86/pci/irq.c
213
struct irq_info *e;
arch/x86/pci/irq.c
217
e = &rt->slots[i];
arch/x86/pci/irq.c
222
e->bus, e->devfn / 8, e->devfn % 8, e->slot);
arch/x86/pci/irq.c
224
DBG(" %d:%02x/%04x", j, e->irq[j].link, e->irq[j].bitmap);
arch/x86/pci/irq.c
228
busmap[e->bus] = 1;
arch/x86/platform/efi/efi.c
194
struct efi_info *e = &boot_params.efi_info;
arch/x86/platform/efi/efi.c
203
if (IS_ENABLED(CONFIG_X86_32) && e->efi_memmap_hi > 0) {
arch/x86/platform/efi/efi.c
207
pmap = (phys_addr_t)(e->efi_memmap | ((u64)e->efi_memmap_hi << 32));
arch/x86/platform/efi/efi.c
210
data.size = e->efi_memmap_size;
arch/x86/platform/efi/efi.c
211
data.desc_size = e->efi_memdesc_size;
arch/x86/platform/efi/efi.c
212
data.desc_version = e->efi_memdesc_version;
arch/x86/virt/svm/sev.c
701
static int get_rmpentry(u64 pfn, struct rmpentry *e)
arch/x86/virt/svm/sev.c
711
: "a" (pfn << PAGE_SHIFT), "c" (e)
arch/x86/virt/svm/sev.c
727
memset(e, 0, sizeof(*e));
arch/x86/virt/svm/sev.c
728
e->gpa = e_raw->gpa << PAGE_SHIFT;
arch/x86/virt/svm/sev.c
729
e->asid = e_raw->asid;
arch/x86/virt/svm/sev.c
730
e->assigned = e_raw->assigned;
arch/x86/virt/svm/sev.c
731
e->pagesize = e_raw->pagesize;
arch/x86/virt/svm/sev.c
732
e->immutable = e_raw->immutable;
arch/x86/virt/svm/sev.c
737
static int __snp_lookup_rmpentry(u64 pfn, struct rmpentry *e, int *level)
arch/x86/virt/svm/sev.c
745
ret = get_rmpentry(pfn, e);
arch/x86/virt/svm/sev.c
765
struct rmpentry e;
arch/x86/virt/svm/sev.c
768
ret = __snp_lookup_rmpentry(pfn, &e, level);
arch/x86/virt/svm/sev.c
772
*assigned = !!e.assigned;
arch/x86/virt/svm/sev.c
787
struct rmpentry e;
arch/x86/virt/svm/sev.c
790
ret = __snp_lookup_rmpentry(pfn, &e, &level);
arch/x86/virt/svm/sev.c
797
if (e.assigned) {
arch/x86/xen/enlighten_pvh.c
100
if (pages != (PFN_DOWN(e->addr + e->size) - PFN_UP(e->addr))) {
arch/x86/xen/enlighten_pvh.c
109
next = e + 1;
arch/x86/xen/enlighten_pvh.c
110
memmove(next, e,
arch/x86/xen/enlighten_pvh.c
111
(bootp->e820_entries - i) * sizeof(*e));
arch/x86/xen/enlighten_pvh.c
113
next->addr = PAGE_ALIGN(e->addr) + PFN_PHYS(pages);
arch/x86/xen/enlighten_pvh.c
114
e->size = next->addr - e->addr;
arch/x86/xen/enlighten_pvh.c
115
next->size -= e->size;
arch/x86/xen/enlighten_pvh.c
117
e->type = E820_TYPE_RAM;
arch/x86/xen/enlighten_pvh.c
120
xen_add_extra_mem(PFN_UP(e->addr), pages);
arch/x86/xen/enlighten_pvh.c
76
struct boot_e820_entry *e = &bootp->e820_table[i];
arch/x86/xen/enlighten_pvh.c
78
if (e->type != E820_TYPE_RAM)
arch/x86/xen/enlighten_pvh.c
80
ram_pages += PFN_DOWN(e->addr + e->size) - PFN_UP(e->addr);
arch/x86/xen/enlighten_pvh.c
91
struct boot_e820_entry *e = &bootp->e820_table[i];
arch/x86/xen/enlighten_pvh.c
94
if (e->type != E820_TYPE_UNUSABLE)
arch/x86/xen/enlighten_pvh.c
98
PFN_DOWN(e->addr + e->size) - PFN_UP(e->addr));
arch/xtensa/include/asm/pgtable.h
207
#define pte_ERROR(e) \
arch/xtensa/include/asm/pgtable.h
208
printk("%s:%d: bad pte %08lx.\n", __FILE__, __LINE__, pte_val(e))
arch/xtensa/include/asm/pgtable.h
209
#define pgd_ERROR(e) \
arch/xtensa/include/asm/pgtable.h
210
printk("%s:%d: bad pgd entry %08lx.\n", __FILE__, __LINE__, pgd_val(e))
arch/xtensa/kernel/jump_label.c
74
void arch_jump_label_transform(struct jump_entry *e,
arch/xtensa/kernel/jump_label.c
77
u32 d = (jump_entry_target(e) - (jump_entry_code(e) + 4));
arch/xtensa/kernel/jump_label.c
94
patch_text(jump_entry_code(e), &insn, JUMP_LABEL_NOP_SIZE);
arch/xtensa/mm/tlb.c
232
static int check_tlb_entry(unsigned w, unsigned e, bool dtlb)
arch/xtensa/mm/tlb.c
234
unsigned tlbidx = w | (e << PAGE_SHIFT);
arch/xtensa/mm/tlb.c
239
unsigned vpn = (r0 & PAGE_MASK) | (e << PAGE_SHIFT);
arch/xtensa/mm/tlb.c
248
dtlb ? 'D' : 'I', w, e, vpn,
arch/xtensa/mm/tlb.c
256
dtlb ? 'D' : 'I', w, e, r0, r1, pte);
arch/xtensa/mm/tlb.c
278
unsigned w, e;
arch/xtensa/mm/tlb.c
283
for (e = 0; e < (1 << XCHAL_DTLB_ARF_ENTRIES_LOG2); ++e)
arch/xtensa/mm/tlb.c
284
bug |= check_tlb_entry(w, e, true);
arch/xtensa/mm/tlb.c
286
for (e = 0; e < (1 << XCHAL_ITLB_ARF_ENTRIES_LOG2); ++e)
arch/xtensa/mm/tlb.c
287
bug |= check_tlb_entry(w, e, false);
arch/xtensa/mm/tlb.c
31
int e = w + (i << PAGE_SHIFT);
arch/xtensa/mm/tlb.c
32
invalidate_itlb_entry_no_isync(e);
arch/xtensa/mm/tlb.c
44
int e = w + (i << PAGE_SHIFT);
arch/xtensa/mm/tlb.c
45
invalidate_dtlb_entry_no_isync(e);
block/bfq-iosched.c
7141
static void bfq_exit_queue(struct elevator_queue *e)
block/bfq-iosched.c
7143
struct bfq_data *bfqd = e->elevator_data;
block/bfq-iosched.c
7397
static ssize_t __FUNC(struct elevator_queue *e, char *page) \
block/bfq-iosched.c
7399
struct bfq_data *bfqd = e->elevator_data; \
block/bfq-iosched.c
7419
static ssize_t __FUNC(struct elevator_queue *e, char *page) \
block/bfq-iosched.c
7421
struct bfq_data *bfqd = e->elevator_data; \
block/bfq-iosched.c
7431
__FUNC(struct elevator_queue *e, const char *page, size_t count) \
block/bfq-iosched.c
7433
struct bfq_data *bfqd = e->elevator_data; \
block/bfq-iosched.c
7463
static ssize_t __FUNC(struct elevator_queue *e, const char *page, size_t count)\
block/bfq-iosched.c
7465
struct bfq_data *bfqd = e->elevator_data; \
block/bfq-iosched.c
7483
static ssize_t bfq_max_budget_store(struct elevator_queue *e,
block/bfq-iosched.c
7486
struct bfq_data *bfqd = e->elevator_data;
block/bfq-iosched.c
7511
static ssize_t bfq_timeout_sync_store(struct elevator_queue *e,
block/bfq-iosched.c
7514
struct bfq_data *bfqd = e->elevator_data;
block/bfq-iosched.c
7534
static ssize_t bfq_strict_guarantees_store(struct elevator_queue *e,
block/bfq-iosched.c
7537
struct bfq_data *bfqd = e->elevator_data;
block/bfq-iosched.c
7556
static ssize_t bfq_low_latency_store(struct elevator_queue *e,
block/bfq-iosched.c
7559
struct bfq_data *bfqd = e->elevator_data;
block/blk-mq-debugfs.c
711
struct elevator_type *e = q->elevator->type;
block/blk-mq-debugfs.c
722
if (!e->queue_debugfs_attrs)
block/blk-mq-debugfs.c
727
debugfs_create_files(q, q->sched_debugfs_dir, q, e->queue_debugfs_attrs);
block/blk-mq-debugfs.c
787
struct elevator_type *e = q->elevator->type;
block/blk-mq-debugfs.c
799
if (!e->hctx_debugfs_attrs)
block/blk-mq-debugfs.c
805
e->hctx_debugfs_attrs);
block/blk-mq-sched.c
104
if (e->type->ops.has_work && !e->type->ops.has_work(hctx))
block/blk-mq-sched.c
116
rq = e->type->ops.dispatch_request(hctx);
block/blk-mq-sched.c
338
struct elevator_queue *e = q->elevator;
block/blk-mq-sched.c
344
if (e && e->type->ops.bio_merge) {
block/blk-mq-sched.c
345
ret = e->type->ops.bio_merge(q, bio, nr_segs);
block/blk-mq-sched.c
614
int blk_mq_init_sched(struct request_queue *q, struct elevator_type *e,
block/blk-mq-sched.c
624
eq = elevator_alloc(q, e, res);
block/blk-mq-sched.c
643
ret = e->ops.init_sched(q, eq);
block/blk-mq-sched.c
648
if (e->ops.init_hctx) {
block/blk-mq-sched.c
649
ret = e->ops.init_hctx(hctx, i);
block/blk-mq-sched.c
687
void blk_mq_exit_sched(struct request_queue *q, struct elevator_queue *e)
block/blk-mq-sched.c
694
if (e->type->ops.exit_hctx && hctx->sched_data) {
block/blk-mq-sched.c
695
e->type->ops.exit_hctx(hctx, i);
block/blk-mq-sched.c
701
if (e->type->ops.exit_sched)
block/blk-mq-sched.c
702
e->type->ops.exit_sched(e);
block/blk-mq-sched.c
88
struct elevator_queue *e = q->elevator;
block/blk-mq-sched.h
107
struct elevator_queue *e = q->elevator;
block/blk-mq-sched.h
109
if (e->type->ops.requeue_request)
block/blk-mq-sched.h
110
e->type->ops.requeue_request(rq);
block/blk-mq-sched.h
116
struct elevator_queue *e = hctx->queue->elevator;
block/blk-mq-sched.h
118
if (e && e->type->ops.has_work)
block/blk-mq-sched.h
119
return e->type->ops.has_work(hctx);
block/blk-mq-sched.h
21
int blk_mq_init_sched(struct request_queue *q, struct elevator_type *e,
block/blk-mq-sched.h
23
void blk_mq_exit_sched(struct request_queue *q, struct elevator_queue *e);
block/blk-mq-sched.h
52
struct elevator_type *e)
block/blk-mq-sched.h
56
if (!e || !e->ops.alloc_sched_data)
block/blk-mq-sched.h
59
sched_data = e->ops.alloc_sched_data(q);
block/blk-mq-sched.h
63
static inline void blk_mq_free_sched_data(struct elevator_type *e, void *data)
block/blk-mq-sched.h
65
if (e && e->ops.free_sched_data)
block/blk-mq-sched.h
66
e->ops.free_sched_data(data);
block/blk-mq-sched.h
85
struct elevator_queue *e = q->elevator;
block/blk-mq-sched.h
87
if (e->type->ops.allow_merge)
block/blk-mq-sched.h
88
return e->type->ops.allow_merge(q, rq, bio);
block/blk-mq-sched.h
96
struct elevator_queue *e = rq->q->elevator;
block/blk-mq-sched.h
98
if (e->type->ops.completed_request)
block/blk-mq-sched.h
99
e->type->ops.completed_request(rq, now);
block/blk-mq.c
451
struct elevator_queue *e = data->q->elevator;
block/blk-mq.c
456
if (e->type->ops.prepare_request)
block/blk-mq.c
457
e->type->ops.prepare_request(rq);
block/elevator.c
101
struct elevator_type *e;
block/elevator.c
103
list_for_each_entry(e, &elv_list, list)
block/elevator.c
104
if (elevator_match(e, name))
block/elevator.c
105
return e;
block/elevator.c
111
struct elevator_type *e;
block/elevator.c
114
e = __elevator_find(name);
block/elevator.c
115
if (e && (!elevator_tryget(e)))
block/elevator.c
116
e = NULL;
block/elevator.c
118
return e;
block/elevator.c
124
struct elevator_type *e, struct elevator_resources *res)
block/elevator.c
132
__elevator_get(e);
block/elevator.c
133
eq->type = e;
block/elevator.c
145
struct elevator_queue *e;
block/elevator.c
147
e = container_of(kobj, struct elevator_queue, kobj);
block/elevator.c
148
elevator_put(e->type);
block/elevator.c
149
kfree(e);
block/elevator.c
154
struct elevator_queue *e = q->elevator;
block/elevator.c
160
mutex_lock(&e->sysfs_lock);
block/elevator.c
161
blk_mq_exit_sched(q, e);
block/elevator.c
162
mutex_unlock(&e->sysfs_lock);
block/elevator.c
180
struct elevator_queue *e = q->elevator;
block/elevator.c
183
hash_add(e->hash, &rq->hash, rq_hash_key(rq));
block/elevator.c
196
struct elevator_queue *e = q->elevator;
block/elevator.c
200
hash_for_each_possible_safe(e->hash, rq, next, hash, offset) {
block/elevator.c
271
struct elevator_queue *e = q->elevator;
block/elevator.c
310
if (e->type->ops.request_merge)
block/elevator.c
311
return e->type->ops.request_merge(q, req, bio);
block/elevator.c
365
struct elevator_queue *e = q->elevator;
block/elevator.c
367
if (e->type->ops.request_merged)
block/elevator.c
368
e->type->ops.request_merged(q, rq, type);
block/elevator.c
379
struct elevator_queue *e = q->elevator;
block/elevator.c
381
if (e->type->ops.requests_merged)
block/elevator.c
382
e->type->ops.requests_merged(q, rq, next);
block/elevator.c
390
struct elevator_queue *e = q->elevator;
block/elevator.c
392
if (e->type->ops.next_request)
block/elevator.c
393
return e->type->ops.next_request(q, rq);
block/elevator.c
400
struct elevator_queue *e = q->elevator;
block/elevator.c
402
if (e->type->ops.former_request)
block/elevator.c
403
return e->type->ops.former_request(q, rq);
block/elevator.c
414
struct elevator_queue *e;
block/elevator.c
420
e = container_of(kobj, struct elevator_queue, kobj);
block/elevator.c
421
mutex_lock(&e->sysfs_lock);
block/elevator.c
422
if (!test_bit(ELEVATOR_FLAG_DYING, &e->flags))
block/elevator.c
423
error = entry->show(e, page);
block/elevator.c
424
mutex_unlock(&e->sysfs_lock);
block/elevator.c
433
struct elevator_queue *e;
block/elevator.c
439
e = container_of(kobj, struct elevator_queue, kobj);
block/elevator.c
440
mutex_lock(&e->sysfs_lock);
block/elevator.c
441
if (!test_bit(ELEVATOR_FLAG_DYING, &e->flags))
block/elevator.c
442
error = entry->store(e, page, length);
block/elevator.c
443
mutex_unlock(&e->sysfs_lock);
block/elevator.c
458
struct elevator_queue *e,
block/elevator.c
463
error = kobject_add(&e->kobj, &q->disk->queue_kobj, "iosched");
block/elevator.c
465
const struct elv_fs_entry *attr = e->type->elevator_attrs;
block/elevator.c
468
if (sysfs_create_file(&e->kobj, &attr->attr))
block/elevator.c
474
kobject_uevent(&e->kobj, KOBJ_ADD);
block/elevator.c
481
set_bit(ELEVATOR_FLAG_REGISTERED, &e->flags);
block/elevator.c
487
struct elevator_queue *e)
block/elevator.c
489
if (e && test_and_clear_bit(ELEVATOR_FLAG_REGISTERED, &e->flags)) {
block/elevator.c
490
kobject_uevent(&e->kobj, KOBJ_REMOVE);
block/elevator.c
491
kobject_del(&e->kobj);
block/elevator.c
498
int elv_register(struct elevator_type *e)
block/elevator.c
501
if (WARN_ON_ONCE(!e->ops.finish_request))
block/elevator.c
504
if (WARN_ON_ONCE(!e->ops.insert_requests || !e->ops.dispatch_request))
block/elevator.c
508
if (e->icq_size) {
block/elevator.c
509
if (WARN_ON(e->icq_size < sizeof(struct io_cq)) ||
block/elevator.c
510
WARN_ON(e->icq_align < __alignof__(struct io_cq)))
block/elevator.c
513
snprintf(e->icq_cache_name, sizeof(e->icq_cache_name),
block/elevator.c
514
"%s_io_cq", e->elevator_name);
block/elevator.c
515
e->icq_cache = kmem_cache_create(e->icq_cache_name, e->icq_size,
block/elevator.c
516
e->icq_align, 0, NULL);
block/elevator.c
517
if (!e->icq_cache)
block/elevator.c
523
if (__elevator_find(e->elevator_name)) {
block/elevator.c
525
kmem_cache_destroy(e->icq_cache);
block/elevator.c
528
list_add_tail(&e->list, &elv_list);
block/elevator.c
531
printk(KERN_INFO "io scheduler %s registered\n", e->elevator_name);
block/elevator.c
537
void elv_unregister(struct elevator_type *e)
block/elevator.c
541
list_del_init(&e->list);
block/elevator.c
548
if (e->icq_cache) {
block/elevator.c
550
kmem_cache_destroy(e->icq_cache);
block/elevator.c
551
e->icq_cache = NULL;
block/elevator.c
612
struct elevator_queue *e;
block/elevator.c
617
e = q->elevator;
block/elevator.c
621
if (e) {
block/elevator.c
623
kobject_put(&e->kobj);
block/elevator.c
63
struct elevator_queue *e = q->elevator;
block/elevator.c
65
if (e->type->ops.allow_merge)
block/elevator.c
66
return e->type->ops.allow_merge(q, rq, bio);
block/elevator.c
838
struct elevator_type *cur = NULL, *e;
block/elevator.c
850
list_for_each_entry(e, &elv_list, list) {
block/elevator.c
851
if (e == cur)
block/elevator.c
852
len += sprintf(name+len, "[%s] ", e->elevator_name);
block/elevator.c
854
len += sprintf(name+len, "%s ", e->elevator_name);
block/elevator.c
93
static bool elevator_match(const struct elevator_type *e, const char *name)
block/elevator.c
95
return !strcmp(e->elevator_name, name) ||
block/elevator.c
96
(e->elevator_alias && !strcmp(e->elevator_alias, name));
block/elevator.h
121
static inline bool elevator_tryget(struct elevator_type *e)
block/elevator.h
123
return try_module_get(e->elevator_owner);
block/elevator.h
126
static inline void __elevator_get(struct elevator_type *e)
block/elevator.h
128
__module_get(e->elevator_owner);
block/elevator.h
131
static inline void elevator_put(struct elevator_type *e)
block/elevator.h
133
module_put(e->elevator_owner);
block/kyber-iosched.c
427
static void kyber_exit_sched(struct elevator_queue *e)
block/kyber-iosched.c
429
struct kyber_queue_data *kqd = e->elevator_data;
block/kyber-iosched.c
850
static ssize_t kyber_##name##_lat_show(struct elevator_queue *e, \
block/kyber-iosched.c
853
struct kyber_queue_data *kqd = e->elevator_data; \
block/kyber-iosched.c
858
static ssize_t kyber_##name##_lat_store(struct elevator_queue *e, \
block/kyber-iosched.c
861
struct kyber_queue_data *kqd = e->elevator_data; \
block/mq-deadline.c
500
static void dd_exit_sched(struct elevator_queue *e)
block/mq-deadline.c
502
struct deadline_data *dd = e->elevator_data;
block/mq-deadline.c
738
static ssize_t __FUNC(struct elevator_queue *e, char *page) \
block/mq-deadline.c
740
struct deadline_data *dd = e->elevator_data; \
block/mq-deadline.c
755
static ssize_t __FUNC(struct elevator_queue *e, const char *page, size_t count) \
block/mq-deadline.c
757
struct deadline_data *dd = e->elevator_data; \
certs/extract-cert.c
107
e = ENGINE_by_id("pkcs11");
certs/extract-cert.c
108
ERR(!e, "Load PKCS#11 ENGINE");
certs/extract-cert.c
109
if (ENGINE_init(e))
certs/extract-cert.c
114
ERR(!ENGINE_ctrl_cmd_string(e, "PIN", key_pass, 0), "Set PKCS#11 PIN");
certs/extract-cert.c
115
ENGINE_ctrl_cmd(e, "LOAD_CERT_CTRL", 0, &parms, NULL, 1);
certs/extract-cert.c
96
ENGINE *e;
crypto/ecrdsa.c
115
vli_from_le64(e, digest, ndigits);
crypto/ecrdsa.c
116
if (vli_cmp(e, ctx->curve->n, ndigits) >= 0)
crypto/ecrdsa.c
117
vli_sub(e, e, ctx->curve->n, ndigits);
crypto/ecrdsa.c
118
if (vli_is_zero(e, ndigits))
crypto/ecrdsa.c
119
e[0] = 1;
crypto/ecrdsa.c
122
vli_mod_inv(v, e, ctx->curve->n, ndigits);
crypto/ecrdsa.c
79
u64 e[ECRDSA_MAX_DIGITS]; /* h \mod q */
crypto/ecrdsa.c
80
u64 *v = e; /* e^{-1} \mod q */
crypto/ecrdsa.c
83
struct ecc_point cc = ECC_POINT_INIT(s, e, ndigits); /* reuse s, e */
crypto/rmd160.c
39
#define ROUND(a, b, c, d, e, f, k, x, s) { \
crypto/rmd160.c
41
(a) = rol32((a), (s)) + (e); \
crypto/rsa.c
131
if (unlikely(!pkey->n || !pkey->e)) {
crypto/rsa.c
18
MPI e;
crypto/rsa.c
200
mpi_free(key->e);
crypto/rsa.c
208
key->e = NULL;
crypto/rsa.c
235
static int rsa_check_exponent_fips(MPI e)
crypto/rsa.c
241
if (!mpi_test_bit(e, 0)) {
crypto/rsa.c
246
if (mpi_cmp_ui(e, 65536) <= 0) {
crypto/rsa.c
260
if (mpi_cmp(e, e_max) >= 0) {
crypto/rsa.c
283
mpi_key->e = mpi_read_raw_data(raw_key.e, raw_key.e_sz);
crypto/rsa.c
284
if (!mpi_key->e)
crypto/rsa.c
296
if (fips_enabled && rsa_check_exponent_fips(mpi_key->e)) {
crypto/rsa.c
326
mpi_key->e = mpi_read_raw_data(raw_key.e, raw_key.e_sz);
crypto/rsa.c
327
if (!mpi_key->e)
crypto/rsa.c
359
if (fips_enabled && rsa_check_exponent_fips(mpi_key->e)) {
crypto/rsa.c
62
return mpi_powm(c, m, key->e, key->n);
crypto/rsa_helper.c
55
key->e = value;
crypto/tcrypt.c
1069
const char *e;
crypto/tcrypt.c
1074
e = "encryption";
crypto/tcrypt.c
1076
e = "decryption";
crypto/tcrypt.c
1115
get_driver_name(crypto_skcipher, tfm), e);
crypto/tcrypt.c
1192
pr_err("%s() failed flags=%x\n", e,
crypto/tcrypt.c
1304
const char *e;
crypto/tcrypt.c
1307
e = "encryption";
crypto/tcrypt.c
1309
e = "decryption";
crypto/tcrypt.c
1322
algo, get_driver_name(crypto_skcipher, tfm), e);
crypto/tcrypt.c
1407
pr_err("%s() failed flags=%x\n", e,
crypto/tcrypt.c
253
const char *e;
crypto/tcrypt.c
269
e = "encryption";
crypto/tcrypt.c
271
e = "decryption";
crypto/tcrypt.c
331
get_driver_name(crypto_aead, tfm), e);
crypto/tcrypt.c
424
pr_err("%s() failed return code=%d\n", e, ret);
crypto/tcrypt.c
530
const char *e;
crypto/tcrypt.c
550
e = "encryption";
crypto/tcrypt.c
552
e = "decryption";
crypto/tcrypt.c
582
get_driver_name(crypto_aead, tfm), e);
crypto/tcrypt.c
676
pr_err("%s() failed return code=%d\n", e, ret);
crypto/testmgr.c
2720
const char *e;
crypto/testmgr.c
2730
e = "encryption";
crypto/testmgr.c
2732
e = "decryption";
crypto/testmgr.c
2785
"on %s for %s\n", j, e, algo);
drivers/accel/amdxdna/aie2_error.c
269
struct async_event *e = handle;
drivers/accel/amdxdna/aie2_error.c
272
e->resp.type = readl(data + offsetof(struct async_event_msg_resp, type));
drivers/accel/amdxdna/aie2_error.c
274
e->resp.status = readl(data + offsetof(struct async_event_msg_resp, status));
drivers/accel/amdxdna/aie2_error.c
276
queue_work(e->wq, &e->work);
drivers/accel/amdxdna/aie2_error.c
280
static int aie2_error_event_send(struct async_event *e)
drivers/accel/amdxdna/aie2_error.c
282
drm_clflush_virt_range(e->buf, e->size); /* device can access */
drivers/accel/amdxdna/aie2_error.c
283
return aie2_register_asyn_event_msg(e->ndev, e->addr, e->size, e,
drivers/accel/amdxdna/aie2_error.c
291
struct async_event *e;
drivers/accel/amdxdna/aie2_error.c
295
e = container_of(err_work, struct async_event, work);
drivers/accel/amdxdna/aie2_error.c
297
xdna = e->ndev->xdna;
drivers/accel/amdxdna/aie2_error.c
299
if (e->resp.status == MAX_AIE2_STATUS_CODE)
drivers/accel/amdxdna/aie2_error.c
302
e->resp.status = MAX_AIE2_STATUS_CODE;
drivers/accel/amdxdna/aie2_error.c
305
e->buf, 0x100, false);
drivers/accel/amdxdna/aie2_error.c
307
info = (struct aie_err_info *)e->buf;
drivers/accel/amdxdna/aie2_error.c
310
max_err = (e->size - sizeof(*info)) / sizeof(struct aie_error);
drivers/accel/amdxdna/aie2_error.c
315
err_col = aie2_error_backtrack(e->ndev, info->payload, info->err_cnt);
drivers/accel/amdxdna/aie2_error.c
322
aie2_update_last_async_error(e->ndev, info->payload, info->err_cnt);
drivers/accel/amdxdna/aie2_error.c
325
if (aie2_error_event_send(e))
drivers/accel/amdxdna/aie2_error.c
373
struct async_event *e = &events->event[i];
drivers/accel/amdxdna/aie2_error.c
376
e->ndev = ndev;
drivers/accel/amdxdna/aie2_error.c
377
e->wq = events->wq;
drivers/accel/amdxdna/aie2_error.c
378
e->buf = &events->buf[offset];
drivers/accel/amdxdna/aie2_error.c
379
e->addr = events->addr + offset;
drivers/accel/amdxdna/aie2_error.c
380
e->size = ASYNC_BUF_SIZE;
drivers/accel/amdxdna/aie2_error.c
381
e->resp.status = MAX_AIE2_STATUS_CODE;
drivers/accel/amdxdna/aie2_error.c
382
INIT_WORK(&e->work, aie2_error_worker);
drivers/accel/amdxdna/aie2_error.c
384
ret = aie2_error_event_send(e);
drivers/accel/habanalabs/common/device.c
349
void hl_engine_data_sprintf(struct engines_data *e, const char *fmt, ...)
drivers/accel/habanalabs/common/device.c
361
if ((e->actual_size + str_size) < e->allocated_buf_size) {
drivers/accel/habanalabs/common/device.c
363
vsnprintf(e->buf + e->actual_size, str_size, fmt, args);
drivers/accel/habanalabs/common/device.c
370
e->actual_size += str_size;
drivers/accel/habanalabs/common/habanalabs.h
1710
struct engines_data *e);
drivers/accel/habanalabs/common/habanalabs.h
4120
__printf(2, 3) void hl_engine_data_sprintf(struct engines_data *e, const char *fmt, ...);
drivers/accel/habanalabs/gaudi/gaudi.c
8026
struct engines_data *e)
drivers/accel/habanalabs/gaudi/gaudi.c
8038
if (e)
drivers/accel/habanalabs/gaudi/gaudi.c
8039
hl_engine_data_sprintf(e,
drivers/accel/habanalabs/gaudi/gaudi.c
8056
if (e)
drivers/accel/habanalabs/gaudi/gaudi.c
8057
hl_engine_data_sprintf(e, fmt, dma_id,
drivers/accel/habanalabs/gaudi/gaudi.c
8062
if (e)
drivers/accel/habanalabs/gaudi/gaudi.c
8063
hl_engine_data_sprintf(e,
drivers/accel/habanalabs/gaudi/gaudi.c
8078
if (e)
drivers/accel/habanalabs/gaudi/gaudi.c
8079
hl_engine_data_sprintf(e, fmt, i,
drivers/accel/habanalabs/gaudi/gaudi.c
8084
if (e)
drivers/accel/habanalabs/gaudi/gaudi.c
8085
hl_engine_data_sprintf(e,
drivers/accel/habanalabs/gaudi/gaudi.c
8106
if (e) {
drivers/accel/habanalabs/gaudi/gaudi.c
8108
hl_engine_data_sprintf(e, fmt, i,
drivers/accel/habanalabs/gaudi/gaudi.c
8112
hl_engine_data_sprintf(e, mme_slave_fmt, i,
drivers/accel/habanalabs/gaudi/gaudi.c
8118
if (e)
drivers/accel/habanalabs/gaudi/gaudi.c
8119
hl_engine_data_sprintf(e,
drivers/accel/habanalabs/gaudi/gaudi.c
8134
if (e)
drivers/accel/habanalabs/gaudi/gaudi.c
8135
hl_engine_data_sprintf(e, nic_fmt, port,
drivers/accel/habanalabs/gaudi/gaudi.c
8149
if (e)
drivers/accel/habanalabs/gaudi/gaudi.c
8150
hl_engine_data_sprintf(e, nic_fmt, port,
drivers/accel/habanalabs/gaudi/gaudi.c
8156
if (e)
drivers/accel/habanalabs/gaudi/gaudi.c
8157
hl_engine_data_sprintf(e, "\n");
drivers/accel/habanalabs/gaudi2/gaudi2.c
2539
struct engines_data *e;
drivers/accel/habanalabs/gaudi2/gaudi2.c
2560
struct engines_data *e);
drivers/accel/habanalabs/gaudi2/gaudi2.c
2562
struct engines_data *e);
drivers/accel/habanalabs/gaudi2/gaudi2.c
2564
struct engines_data *e);
drivers/accel/habanalabs/gaudi2/gaudi2.c
7504
struct engines_data *e)
drivers/accel/habanalabs/gaudi2/gaudi2.c
7514
if (e)
drivers/accel/habanalabs/gaudi2/gaudi2.c
7515
hl_engine_data_sprintf(e,
drivers/accel/habanalabs/gaudi2/gaudi2.c
7544
if (e)
drivers/accel/habanalabs/gaudi2/gaudi2.c
7545
hl_engine_data_sprintf(e, edma_fmt, i, j, is_eng_idle ? "Y" : "N",
drivers/accel/habanalabs/gaudi2/gaudi2.c
7554
struct engines_data *e)
drivers/accel/habanalabs/gaudi2/gaudi2.c
7563
if (e)
drivers/accel/habanalabs/gaudi2/gaudi2.c
7564
hl_engine_data_sprintf(e,
drivers/accel/habanalabs/gaudi2/gaudi2.c
7585
if (e)
drivers/accel/habanalabs/gaudi2/gaudi2.c
7586
hl_engine_data_sprintf(e, pdma_fmt, i, is_eng_idle ? "Y" : "N",
drivers/accel/habanalabs/gaudi2/gaudi2.c
7594
struct engines_data *e)
drivers/accel/habanalabs/gaudi2/gaudi2.c
7604
if (e && hdev->nic_ports_mask)
drivers/accel/habanalabs/gaudi2/gaudi2.c
7605
hl_engine_data_sprintf(e,
drivers/accel/habanalabs/gaudi2/gaudi2.c
7631
if (e)
drivers/accel/habanalabs/gaudi2/gaudi2.c
7632
hl_engine_data_sprintf(e, nic_fmt, i, is_eng_idle ? "Y" : "N",
drivers/accel/habanalabs/gaudi2/gaudi2.c
7640
struct engines_data *e)
drivers/accel/habanalabs/gaudi2/gaudi2.c
7649
if (e)
drivers/accel/habanalabs/gaudi2/gaudi2.c
7650
hl_engine_data_sprintf(e,
drivers/accel/habanalabs/gaudi2/gaudi2.c
7669
if (e)
drivers/accel/habanalabs/gaudi2/gaudi2.c
7670
hl_engine_data_sprintf(e, mme_fmt, i, "N",
drivers/accel/habanalabs/gaudi2/gaudi2.c
7708
if (idle_data->e)
drivers/accel/habanalabs/gaudi2/gaudi2.c
7709
hl_engine_data_sprintf(idle_data->e,
drivers/accel/habanalabs/gaudi2/gaudi2.c
7716
struct engines_data *e)
drivers/accel/habanalabs/gaudi2/gaudi2.c
7724
.e = e,
drivers/accel/habanalabs/gaudi2/gaudi2.c
7733
if (e && prop->tpc_enabled_mask)
drivers/accel/habanalabs/gaudi2/gaudi2.c
7734
hl_engine_data_sprintf(e,
drivers/accel/habanalabs/gaudi2/gaudi2.c
7744
struct engines_data *e)
drivers/accel/habanalabs/gaudi2/gaudi2.c
7756
if (e && (prop->decoder_enabled_mask & (~PCIE_DEC_EN_MASK)))
drivers/accel/habanalabs/gaudi2/gaudi2.c
7757
hl_engine_data_sprintf(e,
drivers/accel/habanalabs/gaudi2/gaudi2.c
7778
if (e)
drivers/accel/habanalabs/gaudi2/gaudi2.c
7779
hl_engine_data_sprintf(e, dec_fmt, i, j,
drivers/accel/habanalabs/gaudi2/gaudi2.c
7784
if (e && (prop->decoder_enabled_mask & PCIE_DEC_EN_MASK))
drivers/accel/habanalabs/gaudi2/gaudi2.c
7785
hl_engine_data_sprintf(e,
drivers/accel/habanalabs/gaudi2/gaudi2.c
7804
if (e)
drivers/accel/habanalabs/gaudi2/gaudi2.c
7805
hl_engine_data_sprintf(e, pcie_dec_fmt, i,
drivers/accel/habanalabs/gaudi2/gaudi2.c
7813
struct engines_data *e)
drivers/accel/habanalabs/gaudi2/gaudi2.c
7822
if (e)
drivers/accel/habanalabs/gaudi2/gaudi2.c
7823
hl_engine_data_sprintf(e,
drivers/accel/habanalabs/gaudi2/gaudi2.c
7842
if (e)
drivers/accel/habanalabs/gaudi2/gaudi2.c
7843
hl_engine_data_sprintf(e, rot_fmt, i, 0, is_eng_idle ? "Y" : "N",
drivers/accel/habanalabs/gaudi2/gaudi2.c
7851
struct engines_data *e)
drivers/accel/habanalabs/gaudi2/gaudi2.c
7855
is_idle &= gaudi2_get_edma_idle_status(hdev, mask_arr, mask_len, e);
drivers/accel/habanalabs/gaudi2/gaudi2.c
7856
is_idle &= gaudi2_get_pdma_idle_status(hdev, mask_arr, mask_len, e);
drivers/accel/habanalabs/gaudi2/gaudi2.c
7857
is_idle &= gaudi2_get_nic_idle_status(hdev, mask_arr, mask_len, e);
drivers/accel/habanalabs/gaudi2/gaudi2.c
7858
is_idle &= gaudi2_get_mme_idle_status(hdev, mask_arr, mask_len, e);
drivers/accel/habanalabs/gaudi2/gaudi2.c
7859
is_idle &= gaudi2_get_tpc_idle_status(hdev, mask_arr, mask_len, e);
drivers/accel/habanalabs/gaudi2/gaudi2.c
7860
is_idle &= gaudi2_get_decoder_idle_status(hdev, mask_arr, mask_len, e);
drivers/accel/habanalabs/gaudi2/gaudi2.c
7861
is_idle &= gaudi2_get_rotator_idle_status(hdev, mask_arr, mask_len, e);
drivers/accel/habanalabs/goya/goya.c
5120
struct engines_data *e)
drivers/accel/habanalabs/goya/goya.c
5131
if (e)
drivers/accel/habanalabs/goya/goya.c
5132
hl_engine_data_sprintf(e, "\nDMA is_idle QM_GLBL_STS0 DMA_CORE_STS0\n"
drivers/accel/habanalabs/goya/goya.c
5146
if (e)
drivers/accel/habanalabs/goya/goya.c
5147
hl_engine_data_sprintf(e, dma_fmt, i, is_eng_idle ? "Y" : "N",
drivers/accel/habanalabs/goya/goya.c
5151
if (e)
drivers/accel/habanalabs/goya/goya.c
5152
hl_engine_data_sprintf(e,
drivers/accel/habanalabs/goya/goya.c
5169
if (e)
drivers/accel/habanalabs/goya/goya.c
5170
hl_engine_data_sprintf(e, fmt, i, is_eng_idle ? "Y" : "N",
drivers/accel/habanalabs/goya/goya.c
5174
if (e)
drivers/accel/habanalabs/goya/goya.c
5175
hl_engine_data_sprintf(e,
drivers/accel/habanalabs/goya/goya.c
5189
if (e) {
drivers/accel/habanalabs/goya/goya.c
5190
hl_engine_data_sprintf(e, fmt, 0, is_eng_idle ? "Y" : "N", qm_glbl_sts0,
drivers/accel/habanalabs/goya/goya.c
5192
hl_engine_data_sprintf(e, "\n");
drivers/acpi/acpi_mrrm.c
101
e->local_region_id = mre_entry->local_region_id;
drivers/acpi/acpi_mrrm.c
103
e->local_region_id = -1;
drivers/acpi/acpi_mrrm.c
105
e->remote_region_id = mre_entry->remote_region_id;
drivers/acpi/acpi_mrrm.c
107
e->remote_region_id = -1;
drivers/acpi/acpi_mrrm.c
37
static int get_node_num(struct mrrm_mem_range_entry *e)
drivers/acpi/acpi_mrrm.c
47
if (zone_intersects(zone, PHYS_PFN(e->base), PHYS_PFN(e->length)))
drivers/acpi/acpi_mrrm.c
91
struct mrrm_mem_range_entry *e;
drivers/acpi/acpi_mrrm.c
94
e = mrrm_mem_range_entry + mrrm_mem_entry_num;
drivers/acpi/acpi_mrrm.c
96
e->base = mre_entry->addr_base;
drivers/acpi/acpi_mrrm.c
97
e->length = mre_entry->addr_len;
drivers/acpi/acpi_mrrm.c
98
e->node = get_node_num(e);
drivers/acpi/acpica/acmacros.h
402
#define ARGI_LIST5(a, b, c, d, e) (ARG_1(e)|ARG_2(d)|ARG_3(c)|ARG_4(b)|ARG_5(a))
drivers/acpi/acpica/acmacros.h
403
#define ARGI_LIST6(a, b, c, d, e, f) (ARG_1(f)|ARG_2(e)|ARG_3(d)|ARG_4(c)|ARG_5(b)|ARG_6(a))
drivers/acpi/acpica/acmacros.h
409
#define ARGP_LIST5(a, b, c, d, e) (ARG_1(a)|ARG_2(b)|ARG_3(c)|ARG_4(d)|ARG_5(e))
drivers/acpi/acpica/acmacros.h
410
#define ARGP_LIST6(a, b, c, d, e, f) (ARG_1(a)|ARG_2(b)|ARG_3(c)|ARG_4(d)|ARG_5(e)|ARG_6(f))
drivers/acpi/acpica/acmacros.h
424
#define ACPI_ERROR_NAMESPACE(s, p, e) acpi_ut_prefixed_namespace_error (AE_INFO, s, p, e);
drivers/acpi/acpica/acmacros.h
425
#define ACPI_ERROR_METHOD(s, n, p, e) acpi_ut_method_error (AE_INFO, s, n, p, e);
drivers/acpi/acpica/acmacros.h
435
#define ACPI_ERROR_NAMESPACE(s, p, e)
drivers/acpi/acpica/acmacros.h
436
#define ACPI_ERROR_METHOD(s, n, p, e)
drivers/acpi/acpica/acpredef.h
125
#define PACKAGE_INFO(a,b,c,d,e,f) {{{(a),(b),(c),(d)}, ((((u16)(f)) << 8) | (e)), 0}}
drivers/acpi/apei/einj-core.c
1000
byte[0] = e > s ? *--e : '0';
drivers/acpi/apei/einj-core.c
1004
if (e <= s)
drivers/acpi/apei/einj-core.c
287
static u32 einjv2_init(struct einjv2_extension_struct *e)
drivers/acpi/apei/einj-core.c
289
if (e->revision != 1) {
drivers/acpi/apei/einj-core.c
290
pr_info("Unknown v2 extension revision %u\n", e->revision);
drivers/acpi/apei/einj-core.c
293
if (e->length < sizeof(*e) || e->length > PAGE_SIZE) {
drivers/acpi/apei/einj-core.c
294
pr_info(FW_BUG "Bad1 v2 extension length %u\n", e->length);
drivers/acpi/apei/einj-core.c
297
if ((e->length - sizeof(*e)) % sizeof(e->component_arr[0])) {
drivers/acpi/apei/einj-core.c
298
pr_info(FW_BUG "Bad2 v2 extension length %u\n", e->length);
drivers/acpi/apei/einj-core.c
302
return (e->length - sizeof(*e)) / sizeof(e->component_arr[0]);
drivers/acpi/apei/einj-core.c
336
struct einjv2_extension_struct *e;
drivers/acpi/apei/einj-core.c
338
e = &v5param.einjv2_struct;
drivers/acpi/apei/einj-core.c
339
max_nr_components = einjv2_init(e);
drivers/acpi/apei/einj-core.c
343
v66param_size = v5param_size - sizeof(*e) + e->length;
drivers/acpi/apei/einj-core.c
970
char *s, *e;
drivers/acpi/apei/einj-core.c
996
e = input + c - 1;
drivers/acpi/apei/einj-core.c
999
byte[1] = *--e;
drivers/acpi/pci_mcfg.c
251
struct mcfg_entry *e;
drivers/acpi/pci_mcfg.c
261
list_for_each_entry(e, &pci_mcfg_list, list) {
drivers/acpi/pci_mcfg.c
262
if (e->segment == seg && e->bus_start <= bus_res->start &&
drivers/acpi/pci_mcfg.c
263
e->bus_end >= bus_res->end) {
drivers/acpi/pci_mcfg.c
264
root->mcfg_addr = e->addr;
drivers/acpi/pci_mcfg.c
296
struct mcfg_entry *e, *arr;
drivers/acpi/pci_mcfg.c
311
for (i = 0, e = arr; i < n; i++, mptr++, e++) {
drivers/acpi/pci_mcfg.c
312
e->segment = mptr->pci_segment;
drivers/acpi/pci_mcfg.c
313
e->addr = mptr->address;
drivers/acpi/pci_mcfg.c
314
e->bus_start = mptr->start_bus_number;
drivers/acpi/pci_mcfg.c
315
e->bus_end = mptr->end_bus_number;
drivers/acpi/pci_mcfg.c
316
list_add(&e->list, &pci_mcfg_list);
drivers/acpi/power.c
113
struct acpi_power_resource_entry *e;
drivers/acpi/power.c
115
list_for_each_entry(e, list, node)
drivers/acpi/power.c
116
if (e->resource->order > resource->order) {
drivers/acpi/power.c
117
list_add_tail(&entry->node, &e->node);
drivers/acpi/power.c
127
struct acpi_power_resource_entry *entry, *e;
drivers/acpi/power.c
129
list_for_each_entry_safe(entry, e, list, node) {
drivers/android/binder.c
236
struct binder_transaction_log_entry *e;
drivers/android/binder.c
241
e = &log->entry[cur % ARRAY_SIZE(log->entry)];
drivers/android/binder.c
242
WRITE_ONCE(e->debug_id_done, 0);
drivers/android/binder.c
249
memset(e, 0, sizeof(*e));
drivers/android/binder.c
250
return e;
drivers/android/binder.c
3073
struct binder_transaction_log_entry *e;
drivers/android/binder.c
3090
e = binder_transaction_log_add(&binder_transaction_log);
drivers/android/binder.c
3091
e->debug_id = t_debug_id;
drivers/android/binder.c
3092
e->call_type = reply ? 2 : !!(tr->flags & TF_ONE_WAY);
drivers/android/binder.c
3093
e->from_proc = proc->pid;
drivers/android/binder.c
3094
e->from_thread = thread->pid;
drivers/android/binder.c
3095
e->target_handle = tr->target.handle;
drivers/android/binder.c
3096
e->data_size = tr->data_size;
drivers/android/binder.c
3097
e->offsets_size = tr->offsets_size;
drivers/android/binder.c
3098
strscpy(e->context_name, proc->context->name, BINDERFS_MAX_NAME);
drivers/android/binder.c
3239
e->to_node = target_node->debug_id;
drivers/android/binder.c
3320
e->to_thread = target_thread->pid;
drivers/android/binder.c
3321
e->to_proc = target_proc->pid;
drivers/android/binder.c
3823
WRITE_ONCE(e->debug_id_done, t_debug_id);
drivers/android/binder.c
3890
e->return_error = return_error;
drivers/android/binder.c
3891
e->return_error_param = return_error_param;
drivers/android/binder.c
3892
e->return_error_line = return_error_line;
drivers/android/binder.c
3894
*fe = *e;
drivers/android/binder.c
3900
WRITE_ONCE(e->debug_id_done, t_debug_id);
drivers/android/binder.c
4819
struct binder_error *e = container_of(
drivers/android/binder.c
4822
WARN_ON(e->cmd == BR_OK);
drivers/android/binder.c
4824
if (put_user(e->cmd, (uint32_t __user *)ptr))
drivers/android/binder.c
4826
cmd = e->cmd;
drivers/android/binder.c
4827
e->cmd = BR_OK;
drivers/android/binder.c
5201
struct binder_error *e = container_of(
drivers/android/binder.c
5206
e->cmd);
drivers/android/binder.c
6447
struct binder_error *e = container_of(
drivers/android/binder.c
6451
prefix, e->cmd);
drivers/android/binder.c
6940
struct binder_transaction_log_entry *e)
drivers/android/binder.c
6942
int debug_id = READ_ONCE(e->debug_id_done);
drivers/android/binder.c
6950
e->debug_id, (e->call_type == 2) ? "reply" :
drivers/android/binder.c
6951
((e->call_type == 1) ? "async" : "call "), e->from_proc,
drivers/android/binder.c
6952
e->from_thread, e->to_proc, e->to_thread, e->context_name,
drivers/android/binder.c
6953
e->to_node, e->target_handle, e->data_size, e->offsets_size,
drivers/android/binder.c
6954
e->return_error, e->return_error_param,
drivers/android/binder.c
6955
e->return_error_line);
drivers/android/binder.c
6961
seq_printf(m, debug_id && debug_id == READ_ONCE(e->debug_id_done) ?
drivers/ata/pata_parport/bpck.c
301
int i, e, l, h, om;
drivers/ata/pata_parport/bpck.c
373
e = 0;
drivers/ata/pata_parport/bpck.c
376
e++;
drivers/ata/pata_parport/bpck.c
379
return e;
drivers/ata/pata_parport/epat.c
275
int e[2] = { 0, 0 };
drivers/ata/pata_parport/epat.c
289
e[j]++;
drivers/ata/pata_parport/epat.c
309
pi->port, pi->mode, cc, e[0], e[1], f);
drivers/ata/pata_parport/epat.c
311
return (e[0] && e[1]) || f;
drivers/ata/pata_parport/epia.c
246
int e[2] = { 0, 0 };
drivers/ata/pata_parport/epia.c
256
e[j]++;
drivers/ata/pata_parport/epia.c
276
pi->port, pi->mode, e[0], e[1], f);
drivers/ata/pata_parport/epia.c
278
return (e[0] && e[1]) || f;
drivers/ata/pata_parport/friq.c
183
int e[2] = { 0, 0 };
drivers/ata/pata_parport/friq.c
198
e[j]++;
drivers/ata/pata_parport/friq.c
214
pi->port, pi->mode, e[0], e[1], r);
drivers/ata/pata_parport/friq.c
216
return r || (e[0] && e[1]);
drivers/ata/pata_parport/frpw.c
222
int e[2] = { 0, 0 };
drivers/ata/pata_parport/frpw.c
246
e[j]++;
drivers/ata/pata_parport/frpw.c
262
pi->port, (pi->private%2), pi->mode, e[0], e[1], r);
drivers/ata/pata_parport/frpw.c
264
return r || (e[0] && e[1]);
drivers/ata/pata_parport/pata_parport.c
356
int e[2] = { 0, 0 };
drivers/ata/pata_parport/pata_parport.c
366
e[j]++;
drivers/ata/pata_parport/pata_parport.c
372
pi->proto->name, pi->port, pi->mode, e[0], e[1]);
drivers/ata/pata_parport/pata_parport.c
374
return e[0] && e[1]; /* not here if both > 0 */
drivers/ata/pata_parport/pata_parport.c
421
int max, s, e;
drivers/ata/pata_parport/pata_parport.c
424
e = s + 1;
drivers/ata/pata_parport/pata_parport.c
428
e = pi->proto->max_units;
drivers/ata/pata_parport/pata_parport.c
441
for (pi->unit = s; pi->unit < e; pi->unit++) {
drivers/atm/idt77252.c
1540
int e;
drivers/atm/idt77252.c
1564
for (e = 0; e < card->tst_size - 2; e++) {
drivers/atm/idt77252.c
1565
if (card->soft_tst[e].tste & TSTE_PUSH_IDLE) {
drivers/atm/idt77252.c
1566
write_sram(card, idle + e,
drivers/atm/idt77252.c
1567
card->soft_tst[e].tste & TSTE_MASK);
drivers/atm/idt77252.c
1568
card->soft_tst[e].tste &= ~(TSTE_PUSH_IDLE);
drivers/atm/idt77252.c
1575
for (e = 0; e < card->tst_size - 2; e++) {
drivers/atm/idt77252.c
1576
if (card->soft_tst[e].tste & TSTE_PUSH_ACTIVE) {
drivers/atm/idt77252.c
1577
write_sram(card, idle + e,
drivers/atm/idt77252.c
1578
card->soft_tst[e].tste & TSTE_MASK);
drivers/atm/idt77252.c
1579
card->soft_tst[e].tste &= ~(TSTE_PUSH_ACTIVE);
drivers/atm/idt77252.c
1580
card->soft_tst[e].tste |= TSTE_PUSH_IDLE;
drivers/atm/idt77252.c
1602
int e, r;
drivers/atm/idt77252.c
1606
for (e = 0; e < avail; e++) {
drivers/atm/idt77252.c
1607
if (card->soft_tst[e].vc == NULL)
drivers/atm/idt77252.c
1610
if (e >= avail) {
drivers/atm/idt77252.c
1616
card->name, vc ? vc->index : -1, e);
drivers/atm/idt77252.c
1630
if ((cl >= avail) && (card->soft_tst[e].vc == NULL)) {
drivers/atm/idt77252.c
1632
card->soft_tst[e].vc = vc;
drivers/atm/idt77252.c
1634
card->soft_tst[e].vc = (void *)-1;
drivers/atm/idt77252.c
1636
card->soft_tst[e].tste = data;
drivers/atm/idt77252.c
1638
card->soft_tst[e].tste |= TSTE_PUSH_ACTIVE;
drivers/atm/idt77252.c
1640
write_sram(card, idle + e, data);
drivers/atm/idt77252.c
1641
card->soft_tst[e].tste |= TSTE_PUSH_IDLE;
drivers/atm/idt77252.c
1648
if (++e == avail)
drivers/atm/idt77252.c
1649
e = 0;
drivers/atm/idt77252.c
1678
int e;
drivers/atm/idt77252.c
1682
for (e = 0; e < card->tst_size - 2; e++) {
drivers/atm/idt77252.c
1683
if (card->soft_tst[e].vc == vc) {
drivers/atm/idt77252.c
1684
card->soft_tst[e].vc = NULL;
drivers/atm/idt77252.c
1686
card->soft_tst[e].tste = TSTE_OPC_VAR;
drivers/atm/idt77252.c
1688
card->soft_tst[e].tste |= TSTE_PUSH_ACTIVE;
drivers/atm/idt77252.c
1690
write_sram(card, idle + e, TSTE_OPC_VAR);
drivers/atm/idt77252.c
1691
card->soft_tst[e].tste |= TSTE_PUSH_IDLE;
drivers/atm/idt77252.c
2048
u16 m, e;
drivers/atm/idt77252.c
2052
e = idt77252_fls(rate) - 1;
drivers/atm/idt77252.c
2053
if (e < 9)
drivers/atm/idt77252.c
2054
m = (rate - (1 << e)) << (9 - e);
drivers/atm/idt77252.c
2055
else if (e == 9)
drivers/atm/idt77252.c
2056
m = (rate - (1 << e));
drivers/atm/idt77252.c
2058
m = (rate - (1 << e)) >> (e - 9);
drivers/atm/idt77252.c
2059
return 0x4000 | (e << 9) | m;
drivers/atm/lanai.c
1083
e(STATUS_SOOL, "SOOL");
drivers/atm/lanai.c
1084
e(STATUS_LOCD, "LOCD");
drivers/atm/lanai.c
1085
e(STATUS_LED, "LED");
drivers/atm/lanai.c
1086
e(STATUS_GPIN, "GPIN");
drivers/atm/lanai.c
1121
e(PCI_STATUS_DETECTED_PARITY, "parity", parity_detect);
drivers/atm/lanai.c
1122
e(PCI_STATUS_SIG_SYSTEM_ERROR, "signalled system", serr_set);
drivers/atm/lanai.c
1123
e(PCI_STATUS_REC_MASTER_ABORT, "master", master_abort);
drivers/atm/lanai.c
1124
e(PCI_STATUS_REC_TARGET_ABORT, "master target", m_target_abort);
drivers/atm/lanai.c
1125
e(PCI_STATUS_SIG_TARGET_ABORT, "slave", s_target_abort);
drivers/atm/lanai.c
1126
e(PCI_STATUS_PARITY, "master parity", master_parity);
drivers/atm/lanai.c
1199
unsigned char *e;
drivers/atm/lanai.c
1201
e = ((unsigned char *) lvcc->tx.buf.ptr) + n;
drivers/atm/lanai.c
1202
m = e - (unsigned char *) lvcc->tx.buf.end;
drivers/atm/lanai.c
1208
e = ((unsigned char *) lvcc->tx.buf.start) + m;
drivers/atm/lanai.c
1210
lvcc->tx.buf.ptr = (u32 *) e;
drivers/atm/lanai.c
1215
unsigned char *e;
drivers/atm/lanai.c
1219
e = ((unsigned char *) lvcc->tx.buf.ptr) + n;
drivers/atm/lanai.c
1220
m = e - (unsigned char *) lvcc->tx.buf.end;
drivers/atm/lanai.c
1226
e = ((unsigned char *) lvcc->tx.buf.start) + m;
drivers/atm/lanai.c
1228
lvcc->tx.buf.ptr = (u32 *) e;
drivers/atm/lanai.c
947
const u8 *e = lanai->eeprom;
drivers/atm/lanai.c
952
if (e[i] < 0x20 || e[i] > 0x7E)
drivers/atm/lanai.c
955
i != EEPROM_COPYRIGHT + EEPROM_COPYRIGHT_LEN && e[i] == '\0')
drivers/atm/lanai.c
957
(char *) &e[EEPROM_COPYRIGHT]);
drivers/atm/lanai.c
963
s += e[i];
drivers/atm/lanai.c
965
if (s != e[EEPROM_CHECKSUM]) {
drivers/atm/lanai.c
968
(unsigned int) s, (unsigned int) e[EEPROM_CHECKSUM]);
drivers/atm/lanai.c
972
if (s != e[EEPROM_CHECKSUM_REV]) {
drivers/atm/lanai.c
975
(unsigned int) s, (unsigned int) e[EEPROM_CHECKSUM_REV]);
drivers/atm/lanai.c
980
if ((e[EEPROM_MAC + i] ^ e[EEPROM_MAC_REV + i]) != 0xFF) {
drivers/atm/lanai.c
984
(unsigned int) e[EEPROM_MAC + i],
drivers/atm/lanai.c
985
(unsigned int) e[EEPROM_MAC_REV + i]);
drivers/atm/lanai.c
988
DPRINTK("eeprom: MAC address = %pM\n", &e[EEPROM_MAC]);
drivers/atm/nicstar.c
1576
int e, r;
drivers/atm/nicstar.c
1587
for (e = 0; e < NS_TST_NUM_ENTRIES; e++) {
drivers/atm/nicstar.c
1588
if (card->tste2vc[e] == NULL)
drivers/atm/nicstar.c
1591
if (e == NS_TST_NUM_ENTRIES) {
drivers/atm/nicstar.c
1601
if (cl >= NS_TST_NUM_ENTRIES && card->tste2vc[e] == NULL) {
drivers/atm/nicstar.c
1602
card->tste2vc[e] = vc;
drivers/atm/nicstar.c
1603
ns_write_sram(card, new_tst + e, &data, 1);
drivers/atm/nicstar.c
1608
if (++e == NS_TST_NUM_ENTRIES) {
drivers/atm/nicstar.c
1609
e = 0;
drivers/auxdisplay/panel.c
1584
lcd.pins.e = lcd_e_pin;
drivers/auxdisplay/panel.c
209
int e;
drivers/auxdisplay/panel.c
855
lcd.pins.e = PIN_STROBE;
drivers/auxdisplay/panel.c
880
lcd.pins.e = PIN_AUTOLF;
drivers/auxdisplay/panel.c
900
lcd.pins.e = PIN_STROBE;
drivers/auxdisplay/panel.c
924
lcd.pins.e = lcd_e_pin;
drivers/auxdisplay/panel.c
961
if (lcd.pins.e == PIN_NOT_SET)
drivers/auxdisplay/panel.c
962
lcd.pins.e = DEFAULT_LCD_PIN_E;
drivers/auxdisplay/panel.c
976
if (lcd.pins.e == PIN_NOT_SET)
drivers/auxdisplay/panel.c
977
lcd.pins.e = PIN_NONE;
drivers/auxdisplay/panel.c
997
pin_to_bits(lcd.pins.e, lcd_bits[LCD_PORT_D][LCD_BIT_E],
drivers/bcma/driver_chipcommon_sflash.c
110
for (e = bcma_sflash_sst_tbl; e->name; e++) {
drivers/bcma/driver_chipcommon_sflash.c
111
if (e->id == id2)
drivers/bcma/driver_chipcommon_sflash.c
118
for (e = bcma_sflash_st_tbl; e->name; e++) {
drivers/bcma/driver_chipcommon_sflash.c
119
if (e->id == id)
drivers/bcma/driver_chipcommon_sflash.c
124
if (!e->name) {
drivers/bcma/driver_chipcommon_sflash.c
134
for (e = bcma_sflash_at_tbl; e->name; e++) {
drivers/bcma/driver_chipcommon_sflash.c
135
if (e->id == id)
drivers/bcma/driver_chipcommon_sflash.c
138
if (!e->name) {
drivers/bcma/driver_chipcommon_sflash.c
149
sflash->blocksize = e->blocksize;
drivers/bcma/driver_chipcommon_sflash.c
150
sflash->numblocks = e->numblocks;
drivers/bcma/driver_chipcommon_sflash.c
155
e->name, sflash->size / 1024, sflash->blocksize,
drivers/bcma/driver_chipcommon_sflash.c
93
const struct bcma_sflash_tbl_e *e;
drivers/block/aoe/aoeblk.c
63
struct aoeif *ifp, *e;
drivers/block/aoe/aoeblk.c
73
e = ifp + NAOEIFS;
drivers/block/aoe/aoeblk.c
74
for (; ifp < e && ifp->nd; ifp++) {
drivers/block/aoe/aoecmd.c
1014
struct aoetgt **t, **e;
drivers/block/aoe/aoecmd.c
1017
e = t + d->ntargets;
drivers/block/aoe/aoecmd.c
1018
for (; t < e && *t; t++)
drivers/block/aoe/aoecmd.c
1481
struct aoetgt **t, **e;
drivers/block/aoe/aoecmd.c
1485
e = t + d->ntargets;
drivers/block/aoe/aoecmd.c
1486
for (; t < e && *t; t++)
drivers/block/aoe/aoecmd.c
1500
struct aoeif *p, *e;
drivers/block/aoe/aoecmd.c
1506
e = p + NAOEIFS;
drivers/block/aoe/aoecmd.c
1507
for (; p < e; p++) {
drivers/block/aoe/aoecmd.c
1517
if (p == e) {
drivers/block/aoe/aoecmd.c
533
struct aoeif *p, *e;
drivers/block/aoe/aoecmd.c
536
e = p + NAOEIFS;
drivers/block/aoe/aoecmd.c
537
for (; p < e; p++)
drivers/block/aoe/aoecmd.c
546
struct aoeif *e;
drivers/block/aoe/aoecmd.c
551
e = t->ifs + NAOEIFS - 1;
drivers/block/aoe/aoecmd.c
552
n = (e - ifp) * sizeof *ifp;
drivers/block/aoe/aoecmd.c
554
e->nd = NULL;
drivers/block/aoe/aoedev.c
274
struct aoetgt **t, **e;
drivers/block/aoe/aoedev.c
296
e = t + d->ntargets;
drivers/block/aoe/aoedev.c
297
for (; t < e && *t; t++)
drivers/block/drbd/drbd_actlog.c
1016
struct lc_element *e;
drivers/block/drbd/drbd_actlog.c
1044
e = lc_find(device->resync, device->resync_wenr);
drivers/block/drbd/drbd_actlog.c
1045
bm_ext = e ? lc_entry(e, struct bm_extent, lce) : NULL;
drivers/block/drbd/drbd_actlog.c
1061
e = lc_try_get(device->resync, enr);
drivers/block/drbd/drbd_actlog.c
1062
bm_ext = e ? lc_entry(e, struct bm_extent, lce) : NULL;
drivers/block/drbd/drbd_actlog.c
1082
e = lc_get(device->resync, enr);
drivers/block/drbd/drbd_actlog.c
1083
bm_ext = e ? lc_entry(e, struct bm_extent, lce) : NULL;
drivers/block/drbd/drbd_actlog.c
1137
struct lc_element *e;
drivers/block/drbd/drbd_actlog.c
1142
e = lc_find(device->resync, enr);
drivers/block/drbd/drbd_actlog.c
1143
bm_ext = e ? lc_entry(e, struct bm_extent, lce) : NULL;
drivers/block/drbd/drbd_actlog.c
1195
struct lc_element *e;
drivers/block/drbd/drbd_actlog.c
1204
e = lc_element_by_index(device->resync, i);
drivers/block/drbd/drbd_actlog.c
1205
bm_ext = lc_entry(e, struct bm_extent, lce);
drivers/block/drbd/drbd_actlog.c
318
struct lc_element *e;
drivers/block/drbd/drbd_actlog.c
338
list_for_each_entry(e, &device->act_log->to_be_changed, list) {
drivers/block/drbd/drbd_actlog.c
343
buffer->update_slot_nr[i] = cpu_to_be16(e->lc_index);
drivers/block/drbd/drbd_actlog.c
344
buffer->update_extent_nr[i] = cpu_to_be32(e->lc_new_number);
drivers/block/drbd/drbd_actlog.c
345
if (e->lc_number != LC_FREE)
drivers/block/drbd/drbd_actlog.c
347
al_extent_to_bm_page(e->lc_number));
drivers/block/drbd/drbd_actlog.c
648
struct lc_element *e;
drivers/block/drbd/drbd_actlog.c
660
e = lc_find(device->resync, enr);
drivers/block/drbd/drbd_actlog.c
662
e = lc_get(device->resync, enr);
drivers/block/drbd/drbd_actlog.c
663
if (e) {
drivers/block/drbd/drbd_actlog.c
664
struct bm_extent *ext = lc_entry(e, struct bm_extent, lce);
drivers/block/drbd/drbd_actlog.c
899
struct lc_element *e;
drivers/block/drbd/drbd_actlog.c
909
e = lc_get(device->resync, enr);
drivers/block/drbd/drbd_actlog.c
910
bm_ext = e ? lc_entry(e, struct bm_extent, lce) : NULL;
drivers/block/drbd/drbd_bitmap.c
1387
unsigned long e, int val)
drivers/block/drbd/drbd_bitmap.c
1396
if (e >= b->bm_bits) {
drivers/block/drbd/drbd_bitmap.c
1398
s, e, b->bm_bits);
drivers/block/drbd/drbd_bitmap.c
1399
e = b->bm_bits ? b->bm_bits -1 : 0;
drivers/block/drbd/drbd_bitmap.c
1401
for (bitnr = s; bitnr <= e; bitnr++) {
drivers/block/drbd/drbd_bitmap.c
1436
const unsigned long e, int val)
drivers/block/drbd/drbd_bitmap.c
1451
c = __bm_change_bits_to(device, s, e, val);
drivers/block/drbd/drbd_bitmap.c
1458
int drbd_bm_set_bits(struct drbd_device *device, const unsigned long s, const unsigned long e)
drivers/block/drbd/drbd_bitmap.c
1460
return bm_change_bits_to(device, s, e, 1);
drivers/block/drbd/drbd_bitmap.c
1464
int drbd_bm_clear_bits(struct drbd_device *device, const unsigned long s, const unsigned long e)
drivers/block/drbd/drbd_bitmap.c
1466
return -bm_change_bits_to(device, s, e, 0);
drivers/block/drbd/drbd_bitmap.c
1501
void _drbd_bm_set_bits(struct drbd_device *device, const unsigned long s, const unsigned long e)
drivers/block/drbd/drbd_bitmap.c
1513
unsigned long el = (e+1) & ~((unsigned long)BITS_PER_LONG-1);
drivers/block/drbd/drbd_bitmap.c
1520
if (e - s <= 3*BITS_PER_LONG) {
drivers/block/drbd/drbd_bitmap.c
1523
__bm_change_bits_to(device, s, e, 1);
drivers/block/drbd/drbd_bitmap.c
1569
if (el <= e)
drivers/block/drbd/drbd_bitmap.c
1570
__bm_change_bits_to(device, el, e, 1);
drivers/block/drbd/drbd_bitmap.c
1612
int drbd_bm_count_bits(struct drbd_device *device, const unsigned long s, const unsigned long e)
drivers/block/drbd/drbd_bitmap.c
1633
for (bitnr = s; bitnr <= e; bitnr++) {
drivers/block/drbd/drbd_bitmap.c
1670
int count, s, e;
drivers/block/drbd/drbd_bitmap.c
1684
e = min((size_t)S2W(enr+1), b->bm_words);
drivers/block/drbd/drbd_bitmap.c
1687
int n = e-s;
drivers/block/drbd/drbd_debugfs.c
649
static void resync_dump_detail(struct seq_file *m, struct lc_element *e)
drivers/block/drbd/drbd_debugfs.c
651
struct bm_extent *bme = lc_entry(e, struct bm_extent, lce);
drivers/block/drbd/drbd_int.h
1283
struct drbd_device *device, unsigned long s, unsigned long e);
drivers/block/drbd/drbd_int.h
1285
struct drbd_device *device, unsigned long s, unsigned long e);
drivers/block/drbd/drbd_int.h
1287
struct drbd_device *device, const unsigned long s, const unsigned long e);
drivers/block/drbd/drbd_int.h
1291
const unsigned long s, const unsigned long e);
drivers/block/drbd/drbd_int.h
1680
#define drbd_chk_io_error(m,e,f) drbd_chk_io_error_(m,e,f, __func__)
drivers/block/drbd/drbd_nl.c
1151
struct lc_element *e;
drivers/block/drbd/drbd_nl.c
1171
e = lc_element_by_index(t, i);
drivers/block/drbd/drbd_nl.c
1172
if (e->refcnt)
drivers/block/drbd/drbd_nl.c
1174
e->lc_number, e->refcnt);
drivers/block/drbd/drbd_nl.c
1175
in_use += e->refcnt;
drivers/block/drbd/drbd_receiver.c
4511
unsigned long e;
drivers/block/drbd/drbd_receiver.c
4528
e = s + rl -1;
drivers/block/drbd/drbd_receiver.c
4529
if (e >= c->bm_bits) {
drivers/block/drbd/drbd_receiver.c
4530
drbd_err(peer_device, "bitmap overflow (e:%lu) while decoding bm RLE packet\n", e);
drivers/block/drbd/drbd_receiver.c
4533
_drbd_bm_set_bits(peer_device->device, s, e);
drivers/char/ipmi/ipmi_msghandler.c
82
int e;
drivers/char/ipmi/ipmi_msghandler.c
85
e = match_string(ipmi_panic_event_str, -1, strstrip(valcp));
drivers/char/ipmi/ipmi_msghandler.c
86
if (e < 0)
drivers/char/ipmi/ipmi_msghandler.c
87
return e;
drivers/char/ipmi/ipmi_msghandler.c
89
ipmi_send_panic_event = e;
drivers/char/ipmi/ipmi_si_intf.c
1879
struct smi_info *e;
drivers/char/ipmi/ipmi_si_intf.c
1881
list_for_each_entry(e, &smi_infos, link) {
drivers/char/ipmi/ipmi_si_intf.c
1882
if (e->io.addr_space != info->io.addr_space)
drivers/char/ipmi/ipmi_si_intf.c
1884
if (e->io.addr_data == info->io.addr_data) {
drivers/char/ipmi/ipmi_si_intf.c
1890
if (info->io.slave_addr && !e->io.slave_addr)
drivers/char/ipmi/ipmi_si_intf.c
1891
e->io.slave_addr = info->io.slave_addr;
drivers/char/ipmi/ipmi_si_intf.c
1892
return e;
drivers/char/ipmi/ipmi_si_intf.c
2137
struct smi_info *e, *e2;
drivers/char/ipmi/ipmi_si_intf.c
2161
list_for_each_entry(e, &smi_infos, link) {
drivers/char/ipmi/ipmi_si_intf.c
2165
if (!e->io.irq)
drivers/char/ipmi/ipmi_si_intf.c
2173
if (e2 == e)
drivers/char/ipmi/ipmi_si_intf.c
2175
if (e2->io.irq && ipmi_smi_info_same(e, e2)) {
drivers/char/ipmi/ipmi_si_intf.c
2181
try_smi_init(e);
drivers/char/ipmi/ipmi_si_intf.c
2187
list_for_each_entry(e, &smi_infos, link) {
drivers/char/ipmi/ipmi_si_intf.c
2190
if (e->io.irq)
drivers/char/ipmi/ipmi_si_intf.c
2201
if (ipmi_smi_info_same(e, e2)) {
drivers/char/ipmi/ipmi_si_intf.c
2207
if (e2 == e)
drivers/char/ipmi/ipmi_si_intf.c
2209
if (ipmi_smi_info_same(e, e2)) {
drivers/char/ipmi/ipmi_si_intf.c
2215
try_smi_init(e);
drivers/char/ipmi/ipmi_si_intf.c
2320
struct smi_info *e;
drivers/char/ipmi/ipmi_si_intf.c
2323
list_for_each_entry(e, &smi_infos, link) {
drivers/char/ipmi/ipmi_si_intf.c
2324
if (e->io.dev == dev) {
drivers/char/ipmi/ipmi_si_intf.c
2325
cleanup_one_si(e);
drivers/char/ipmi/ipmi_si_intf.c
2336
struct smi_info *e, *tmp_e;
drivers/char/ipmi/ipmi_si_intf.c
2340
list_for_each_entry_safe(e, tmp_e, &smi_infos, link) {
drivers/char/ipmi/ipmi_si_intf.c
2341
if (e->io.addr_space != addr_space)
drivers/char/ipmi/ipmi_si_intf.c
2343
if (e->io.si_info->type != si_type)
drivers/char/ipmi/ipmi_si_intf.c
2345
if (e->io.addr_data == addr) {
drivers/char/ipmi/ipmi_si_intf.c
2346
dev = get_device(e->io.dev);
drivers/char/ipmi/ipmi_si_intf.c
2347
cleanup_one_si(e);
drivers/char/ipmi/ipmi_si_intf.c
2357
struct smi_info *e, *tmp_e;
drivers/char/ipmi/ipmi_si_intf.c
2371
list_for_each_entry_safe(e, tmp_e, &smi_infos, link)
drivers/char/ipmi/ipmi_si_intf.c
2372
cleanup_one_si(e);
drivers/clk/meson/axg-audio.c
464
AUD_MST_SCLK_PRE_EN(e, AUDIO_MST_E_SCLK_CTRL0);
drivers/clk/meson/axg-audio.c
477
AUD_MST_SCLK_DIV(e, AUDIO_MST_E_SCLK_CTRL0);
drivers/clk/meson/axg-audio.c
490
AUD_MST_SCLK_POST_EN(e, AUDIO_MST_E_SCLK_CTRL0);
drivers/clk/meson/axg-audio.c
503
AUD_MST_SCLK(e, AUDIO_MST_E_SCLK_CTRL1);
drivers/clk/meson/axg-audio.c
516
AUD_MST_LRCLK_DIV(e, AUDIO_MST_E_SCLK_CTRL0);
drivers/clk/meson/axg-audio.c
529
AUD_MST_LRCLK(e, AUDIO_MST_E_SCLK_CTRL1);
drivers/clocksource/arm_arch_timer_mmio.c
55
#define evt_to_arch_timer(e) container_of(e, struct arch_timer, evt)
drivers/connector/cn_proc.c
191
ev->event_data.id.e.euid = from_kuid_munged(&init_user_ns, cred->euid);
drivers/connector/cn_proc.c
194
ev->event_data.id.e.egid = from_kgid_munged(&init_user_ns, cred->egid);
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-cipher.c
383
int e;
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-cipher.c
389
e = sun8i_ce_get_engine_number(op->ce);
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-cipher.c
390
rctx->flow = e;
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-cipher.c
391
engine = op->ce->chanlist[e].engine;
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-cipher.c
402
int e;
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-cipher.c
408
e = sun8i_ce_get_engine_number(op->ce);
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-cipher.c
409
rctx->flow = e;
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-cipher.c
410
engine = op->ce->chanlist[e].engine;
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-hash.c
246
int e;
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-hash.c
251
e = sun8i_ce_get_engine_number(ce);
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-hash.c
252
rctx->flow = e;
drivers/crypto/allwinner/sun8i-ce/sun8i-ce-hash.c
253
engine = ce->chanlist[e].engine;
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-cipher.c
351
int e;
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-cipher.c
359
e = sun8i_ss_get_engine_number(op->ss);
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-cipher.c
360
engine = op->ss->flows[e].engine;
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-cipher.c
361
rctx->flow = e;
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-cipher.c
372
int e;
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-cipher.c
380
e = sun8i_ss_get_engine_number(op->ss);
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-cipher.c
381
engine = op->ss->flows[e].engine;
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-cipher.c
382
rctx->flow = e;
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-hash.c
402
int e;
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-hash.c
410
e = sun8i_ss_get_engine_number(ss);
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-hash.c
411
rctx->flow = e;
drivers/crypto/allwinner/sun8i-ss/sun8i-ss-hash.c
412
engine = ss->flows[e].engine;
drivers/crypto/amlogic/amlogic-gxl-cipher.c
280
int e;
drivers/crypto/amlogic/amlogic-gxl-cipher.c
285
e = get_engine_number(op->mc);
drivers/crypto/amlogic/amlogic-gxl-cipher.c
286
engine = op->mc->chanlist[e].engine;
drivers/crypto/amlogic/amlogic-gxl-cipher.c
287
rctx->flow = e;
drivers/crypto/amlogic/amlogic-gxl-cipher.c
298
int e;
drivers/crypto/amlogic/amlogic-gxl-cipher.c
303
e = get_engine_number(op->mc);
drivers/crypto/amlogic/amlogic-gxl-cipher.c
304
engine = op->mc->chanlist[e].engine;
drivers/crypto/amlogic/amlogic-gxl-cipher.c
305
rctx->flow = e;
drivers/crypto/aspeed/aspeed-acry.c
118
u8 *e;
drivers/crypto/aspeed/aspeed-acry.c
368
if (!ctx->e || !ctx->e_sz) {
drivers/crypto/aspeed/aspeed-acry.c
375
ctx->e, ctx->e_sz,
drivers/crypto/aspeed/aspeed-acry.c
457
ctx->e = aspeed_rsa_key_copy(value, len);
drivers/crypto/aspeed/aspeed-acry.c
458
if (!ctx->e)
drivers/crypto/aspeed/aspeed-acry.c
478
kfree_sensitive(ctx->e);
drivers/crypto/aspeed/aspeed-acry.c
513
ret = aspeed_rsa_set_e(ctx, (u8 *)ctx->key.e, ctx->key.e_sz);
drivers/crypto/caam/caampkc.c
1069
rsa_key->e = kmemdup(raw_key.e, raw_key.e_sz, GFP_KERNEL);
drivers/crypto/caam/caampkc.c
1070
if (!rsa_key->e)
drivers/crypto/caam/caampkc.c
421
pdb->e_dma = dma_map_single(dev, key->e, key->e_sz, DMA_TO_DEVICE);
drivers/crypto/caam/caampkc.c
723
if (unlikely(!key->n || !key->e))
drivers/crypto/caam/caampkc.c
876
kfree(key->e);
drivers/crypto/caam/caampkc.c
959
rsa_key->e = kmemdup(raw_key.e, raw_key.e_sz, GFP_KERNEL);
drivers/crypto/caam/caampkc.c
960
if (!rsa_key->e)
drivers/crypto/caam/caampkc.h
71
u8 *e;
drivers/crypto/cavium/cpt/cptvf_algs.c
121
fctx->enc.enc_ctrl.e.enc_cipher = ctx->cipher_type;
drivers/crypto/cavium/cpt/cptvf_algs.c
122
fctx->enc.enc_ctrl.e.aes_key = ctx->key_type;
drivers/crypto/cavium/cpt/cptvf_algs.c
123
fctx->enc.enc_ctrl.e.iv_source = FROM_DPTR;
drivers/crypto/cavium/cpt/cptvf_algs.h
74
} e;
drivers/crypto/ccp/ccp-crypto-rsa.c
157
raw_key.e, raw_key.e_sz);
drivers/crypto/ccp/ccp-dev.c
102
if (e < ARRAY_SIZE(ccp_error_codes))
drivers/crypto/ccp/ccp-dev.c
103
dev_err(d->dev, "CCP error %d: %s\n", e, ccp_error_codes[e]);
drivers/crypto/ccp/ccp-dev.c
105
dev_err(d->dev, "CCP error %d: Unknown Error\n", e);
drivers/crypto/ccp/ccp-dev.c
97
void ccp_log_error(struct ccp_device *d, unsigned int e)
drivers/crypto/ccp/ccp-dev.c
99
if (WARN_ON(e >= CCP_MAX_ERROR_CODE))
drivers/crypto/hisilicon/hpre/hpre_crypto.c
1069
ret = hpre_rsa_set_e(ctx, rsa_key.e, rsa_key.e_sz);
drivers/crypto/intel/qat/qat_common/qat_asym_algs.c
1027
ctx->e = NULL;
drivers/crypto/intel/qat/qat_common/qat_asym_algs.c
1031
ctx->e = dma_alloc_coherent(dev, ctx->key_sz, &ctx->dma_e, GFP_KERNEL);
drivers/crypto/intel/qat/qat_common/qat_asym_algs.c
1032
if (!ctx->e)
drivers/crypto/intel/qat/qat_common/qat_asym_algs.c
1035
memcpy(ctx->e + (ctx->key_sz - vlen), ptr, vlen);
drivers/crypto/intel/qat/qat_common/qat_asym_algs.c
1172
if (ctx->e)
drivers/crypto/intel/qat/qat_common/qat_asym_algs.c
1173
dma_free_coherent(dev, ctx->key_sz, ctx->e, ctx->dma_e);
drivers/crypto/intel/qat/qat_common/qat_asym_algs.c
1200
ctx->e = NULL;
drivers/crypto/intel/qat/qat_common/qat_asym_algs.c
1231
ret = qat_rsa_set_e(ctx, rsa_key.e, rsa_key.e_sz);
drivers/crypto/intel/qat/qat_common/qat_asym_algs.c
1241
if (!ctx->n || !ctx->e) {
drivers/crypto/intel/qat/qat_common/qat_asym_algs.c
27
dma_addr_t e;
drivers/crypto/intel/qat/qat_common/qat_asym_algs.c
61
char *e;
drivers/crypto/intel/qat/qat_common/qat_asym_algs.c
711
if (unlikely(!ctx->n || !ctx->e))
drivers/crypto/intel/qat/qat_common/qat_asym_algs.c
737
qat_req->in.rsa.enc.e = ctx->dma_e;
drivers/crypto/marvell/octeontx/otx_cptvf_algs.c
273
fctx->enc.enc_ctrl.e.enc_cipher = ctx->cipher_type;
drivers/crypto/marvell/octeontx/otx_cptvf_algs.c
274
fctx->enc.enc_ctrl.e.aes_key = ctx->key_type;
drivers/crypto/marvell/octeontx/otx_cptvf_algs.c
275
fctx->enc.enc_ctrl.e.iv_source = OTX_CPT_FROM_CPTR;
drivers/crypto/marvell/octeontx/otx_cptvf_algs.c
896
rctx->ctrl_word.e.enc_data_offset = req->assoclen;
drivers/crypto/marvell/octeontx/otx_cptvf_algs.c
900
fctx->enc.enc_ctrl.e.iv_source = OTX_CPT_FROM_CPTR;
drivers/crypto/marvell/octeontx/otx_cptvf_algs.c
911
memcpy(fctx->hmac.e.ipad, ctx->ipad, ds);
drivers/crypto/marvell/octeontx/otx_cptvf_algs.c
913
memcpy(fctx->hmac.e.opad, ctx->opad, ds);
drivers/crypto/marvell/octeontx/otx_cptvf_algs.c
917
fctx->enc.enc_ctrl.e.iv_source = OTX_CPT_FROM_DPTR;
drivers/crypto/marvell/octeontx/otx_cptvf_algs.c
924
rctx->ctrl_word.e.iv_offset = req->assoclen - AES_GCM_IV_OFFSET;
drivers/crypto/marvell/octeontx/otx_cptvf_algs.c
947
fctx->enc.enc_ctrl.e.enc_cipher = ctx->cipher_type;
drivers/crypto/marvell/octeontx/otx_cptvf_algs.c
948
fctx->enc.enc_ctrl.e.aes_key = ctx->key_type;
drivers/crypto/marvell/octeontx/otx_cptvf_algs.c
949
fctx->enc.enc_ctrl.e.mac_type = ctx->mac_type;
drivers/crypto/marvell/octeontx/otx_cptvf_algs.c
950
fctx->enc.enc_ctrl.e.mac_len = mac_len;
drivers/crypto/marvell/octeontx/otx_cptvf_algs.h
117
} e;
drivers/crypto/marvell/octeontx/otx_cptvf_algs.h
156
} e;
drivers/crypto/marvell/octeontx/otx_cptvf_algs.h
99
} e;
drivers/crypto/marvell/octeontx2/otx2_cptpf_ucode.c
463
int ret, e, i;
drivers/crypto/marvell/octeontx2/otx2_cptpf_ucode.c
467
for (e = 1; e < OTX2_CPT_MAX_ENG_TYPES; e++) {
drivers/crypto/marvell/octeontx2/otx2_cptpf_ucode.c
468
strscpy(eng_type, get_eng_type_str(e));
drivers/crypto/marvell/octeontx2/otx2_cptvf_algs.c
1006
fctx->enc.enc_ctrl.e.iv_source = OTX2_CPT_FROM_DPTR;
drivers/crypto/marvell/octeontx2/otx2_cptvf_algs.c
1013
rctx->ctrl_word.e.iv_offset = req->assoclen - AES_GCM_IV_OFFSET;
drivers/crypto/marvell/octeontx2/otx2_cptvf_algs.c
1036
fctx->enc.enc_ctrl.e.enc_cipher = ctx->cipher_type;
drivers/crypto/marvell/octeontx2/otx2_cptvf_algs.c
1037
fctx->enc.enc_ctrl.e.aes_key = ctx->key_type;
drivers/crypto/marvell/octeontx2/otx2_cptvf_algs.c
1038
fctx->enc.enc_ctrl.e.mac_type = ctx->mac_type;
drivers/crypto/marvell/octeontx2/otx2_cptvf_algs.c
1039
fctx->enc.enc_ctrl.e.mac_len = mac_len;
drivers/crypto/marvell/octeontx2/otx2_cptvf_algs.c
259
fctx->enc.enc_ctrl.e.enc_cipher = ctx->cipher_type;
drivers/crypto/marvell/octeontx2/otx2_cptvf_algs.c
260
fctx->enc.enc_ctrl.e.aes_key = ctx->key_type;
drivers/crypto/marvell/octeontx2/otx2_cptvf_algs.c
261
fctx->enc.enc_ctrl.e.iv_source = OTX2_CPT_FROM_CPTR;
drivers/crypto/marvell/octeontx2/otx2_cptvf_algs.c
979
rctx->ctrl_word.e.enc_data_offset = req->assoclen;
drivers/crypto/marvell/octeontx2/otx2_cptvf_algs.c
986
fctx->enc.enc_ctrl.e.iv_source = OTX2_CPT_FROM_CPTR;
drivers/crypto/marvell/octeontx2/otx2_cptvf_algs.c
997
memcpy(fctx->hmac.e.ipad, ctx->ipad, ds);
drivers/crypto/marvell/octeontx2/otx2_cptvf_algs.c
999
memcpy(fctx->hmac.e.opad, ctx->opad, ds);
drivers/crypto/marvell/octeontx2/otx2_cptvf_algs.h
108
} e;
drivers/crypto/marvell/octeontx2/otx2_cptvf_algs.h
145
} e;
drivers/crypto/marvell/octeontx2/otx2_cptvf_algs.h
90
} e;
drivers/crypto/starfive/jh7110-cryp.h
139
u8 *e;
drivers/crypto/starfive/jh7110-rsa.c
267
ret = starfive_rsa_start(ctx, rctx->rsa_data, key->e,
drivers/crypto/starfive/jh7110-rsa.c
302
if (unlikely(!key->n || !key->e))
drivers/crypto/starfive/jh7110-rsa.c
393
rsa_key->e = NULL;
drivers/crypto/starfive/jh7110-rsa.c
397
rsa_key->e = kzalloc(rsa_key->key_sz, GFP_KERNEL);
drivers/crypto/starfive/jh7110-rsa.c
398
if (!rsa_key->e)
drivers/crypto/starfive/jh7110-rsa.c
408
memcpy(rsa_key->e + (rsa_key->key_sz - vlen), ptr, vlen);
drivers/crypto/starfive/jh7110-rsa.c
476
ret = starfive_rsa_set_e(rsa_key, raw_key.e, raw_key.e_sz);
drivers/crypto/starfive/jh7110-rsa.c
486
if (!rsa_key->n || !rsa_key->e) {
drivers/crypto/starfive/jh7110-rsa.c
51
kfree_sensitive(key->e);
drivers/dma/amd/ptdma/ptdma-dev.c
52
static void pt_log_error(struct pt_device *d, int e)
drivers/dma/amd/ptdma/ptdma-dev.c
54
dev_err(d->dev, "PTDMA error: %s (0x%x)\n", pt_error_codes[e], e);
drivers/dma/amd/ptdma/ptdma-dmaengine.c
28
static void ae4_log_error(struct pt_device *d, int e)
drivers/dma/amd/ptdma/ptdma-dmaengine.c
31
if (e <= 7)
drivers/dma/amd/ptdma/ptdma-dmaengine.c
32
dev_info(d->dev, "AE4DMA error: %s (0x%x)\n", ae4_error_codes[e], e);
drivers/dma/amd/ptdma/ptdma-dmaengine.c
34
else if (e > 7 && e <= 15)
drivers/dma/amd/ptdma/ptdma-dmaengine.c
35
dev_info(d->dev, "AE4DMA error: %s (0x%x)\n", "INVALID DESCRIPTOR", e);
drivers/dma/amd/ptdma/ptdma-dmaengine.c
37
else if (e > 15 && e <= 31)
drivers/dma/amd/ptdma/ptdma-dmaengine.c
38
dev_info(d->dev, "AE4DMA error: %s (0x%x)\n", "FIRMWARE ERROR", e);
drivers/dma/amd/ptdma/ptdma-dmaengine.c
40
else if (e > 31 && e <= 63)
drivers/dma/amd/ptdma/ptdma-dmaengine.c
41
dev_info(d->dev, "AE4DMA error: %s (0x%x)\n", "FATAL ERROR", e);
drivers/dma/amd/ptdma/ptdma-dmaengine.c
43
else if (e > 63 && e <= 255)
drivers/dma/amd/ptdma/ptdma-dmaengine.c
44
dev_info(d->dev, "AE4DMA error: %s (0x%x)\n", "PTE ERROR", e);
drivers/dma/idxd/debugfs.c
27
if (!entry->e.desc_valid)
drivers/dma/idxd/debugfs.c
37
entry->e.desc_valid, entry->e.wq_idx_valid,
drivers/dma/idxd/debugfs.c
38
entry->e.batch, entry->e.fault_rw, entry->e.priv,
drivers/dma/idxd/debugfs.c
39
entry->e.error, entry->e.wq_idx, entry->e.operation,
drivers/dma/idxd/debugfs.c
40
entry->e.pasid, entry->e.batch_idx, entry->e.fault_addr);
drivers/dma/idxd/registers.h
657
struct __evl_entry e;
drivers/dma/idxd/registers.h
662
struct __evl_entry e;
drivers/edac/edac_mc.c
1015
p = e->label;
drivers/edac/edac_mc.c
1017
end = p + sizeof(e->label);
drivers/edac/edac_mc.c
1029
if (dimm->grain > e->grain)
drivers/edac/edac_mc.c
1030
e->grain = dimm->grain;
drivers/edac/edac_mc.c
1043
p = e->label;
drivers/edac/edac_mc.c
1069
strscpy(e->label, "any memory", sizeof(e->label));
drivers/edac/edac_mc.c
1070
else if (!*e->label)
drivers/edac/edac_mc.c
1071
strscpy(e->label, "unknown memory", sizeof(e->label));
drivers/edac/edac_mc.c
1073
edac_inc_csrow(e, row, chan);
drivers/edac/edac_mc.c
1076
p = e->location;
drivers/edac/edac_mc.c
1077
end = p + sizeof(e->location);
drivers/edac/edac_mc.c
1089
edac_raw_mc_handle_error(e);
drivers/edac/edac_mc.c
56
static struct mem_ctl_info *error_desc_to_mci(struct edac_raw_error_desc *e)
drivers/edac/edac_mc.c
58
return container_of(e, struct mem_ctl_info, error_desc);
drivers/edac/edac_mc.c
803
static void edac_inc_ce_error(struct edac_raw_error_desc *e)
drivers/edac/edac_mc.c
805
int pos[EDAC_MAX_LAYERS] = { e->top_layer, e->mid_layer, e->low_layer };
drivers/edac/edac_mc.c
806
struct mem_ctl_info *mci = error_desc_to_mci(e);
drivers/edac/edac_mc.c
809
mci->ce_mc += e->error_count;
drivers/edac/edac_mc.c
812
dimm->ce_count += e->error_count;
drivers/edac/edac_mc.c
814
mci->ce_noinfo_count += e->error_count;
drivers/edac/edac_mc.c
817
static void edac_inc_ue_error(struct edac_raw_error_desc *e)
drivers/edac/edac_mc.c
819
int pos[EDAC_MAX_LAYERS] = { e->top_layer, e->mid_layer, e->low_layer };
drivers/edac/edac_mc.c
820
struct mem_ctl_info *mci = error_desc_to_mci(e);
drivers/edac/edac_mc.c
823
mci->ue_mc += e->error_count;
drivers/edac/edac_mc.c
826
dimm->ue_count += e->error_count;
drivers/edac/edac_mc.c
828
mci->ue_noinfo_count += e->error_count;
drivers/edac/edac_mc.c
831
static void edac_ce_error(struct edac_raw_error_desc *e)
drivers/edac/edac_mc.c
833
struct mem_ctl_info *mci = error_desc_to_mci(e);
drivers/edac/edac_mc.c
839
e->error_count, e->msg,
drivers/edac/edac_mc.c
840
*e->msg ? " " : "",
drivers/edac/edac_mc.c
841
e->label, e->location, e->page_frame_number, e->offset_in_page,
drivers/edac/edac_mc.c
842
e->grain, e->syndrome,
drivers/edac/edac_mc.c
843
*e->other_detail ? " - " : "",
drivers/edac/edac_mc.c
844
e->other_detail);
drivers/edac/edac_mc.c
847
edac_inc_ce_error(e);
drivers/edac/edac_mc.c
862
mci->ctl_page_to_phys(mci, e->page_frame_number) :
drivers/edac/edac_mc.c
863
e->page_frame_number;
drivers/edac/edac_mc.c
865
edac_mc_scrub_block(remapped_page, e->offset_in_page, e->grain);
drivers/edac/edac_mc.c
869
static void edac_ue_error(struct edac_raw_error_desc *e)
drivers/edac/edac_mc.c
871
struct mem_ctl_info *mci = error_desc_to_mci(e);
drivers/edac/edac_mc.c
876
e->error_count, e->msg,
drivers/edac/edac_mc.c
877
*e->msg ? " " : "",
drivers/edac/edac_mc.c
878
e->label, e->location, e->page_frame_number, e->offset_in_page,
drivers/edac/edac_mc.c
879
e->grain,
drivers/edac/edac_mc.c
880
*e->other_detail ? " - " : "",
drivers/edac/edac_mc.c
881
e->other_detail);
drivers/edac/edac_mc.c
884
edac_inc_ue_error(e);
drivers/edac/edac_mc.c
888
e->msg,
drivers/edac/edac_mc.c
889
*e->msg ? " " : "",
drivers/edac/edac_mc.c
890
e->label, e->location, e->page_frame_number, e->offset_in_page,
drivers/edac/edac_mc.c
891
e->grain,
drivers/edac/edac_mc.c
892
*e->other_detail ? " - " : "",
drivers/edac/edac_mc.c
893
e->other_detail);
drivers/edac/edac_mc.c
897
static void edac_inc_csrow(struct edac_raw_error_desc *e, int row, int chan)
drivers/edac/edac_mc.c
899
struct mem_ctl_info *mci = error_desc_to_mci(e);
drivers/edac/edac_mc.c
900
enum hw_event_mc_err_type type = e->type;
drivers/edac/edac_mc.c
901
u16 count = e->error_count;
drivers/edac/edac_mc.c
917
void edac_raw_mc_handle_error(struct edac_raw_error_desc *e)
drivers/edac/edac_mc.c
919
struct mem_ctl_info *mci = error_desc_to_mci(e);
drivers/edac/edac_mc.c
923
if (WARN_ON_ONCE(!e->grain))
drivers/edac/edac_mc.c
924
e->grain = 1;
drivers/edac/edac_mc.c
926
grain_bits = fls_long(e->grain - 1);
drivers/edac/edac_mc.c
930
trace_mc_event(e->type, e->msg, e->label, e->error_count,
drivers/edac/edac_mc.c
931
mci->mc_idx, e->top_layer, e->mid_layer,
drivers/edac/edac_mc.c
932
e->low_layer,
drivers/edac/edac_mc.c
933
(e->page_frame_number << PAGE_SHIFT) | e->offset_in_page,
drivers/edac/edac_mc.c
934
grain_bits, e->syndrome, e->other_detail);
drivers/edac/edac_mc.c
936
if (e->type == HW_EVENT_ERR_CORRECTED)
drivers/edac/edac_mc.c
937
edac_ce_error(e);
drivers/edac/edac_mc.c
939
edac_ue_error(e);
drivers/edac/edac_mc.c
960
struct edac_raw_error_desc *e = &mci->error_desc;
drivers/edac/edac_mc.c
967
memset(e, 0, sizeof (*e));
drivers/edac/edac_mc.c
968
e->error_count = error_count;
drivers/edac/edac_mc.c
969
e->type = type;
drivers/edac/edac_mc.c
970
e->top_layer = top_layer;
drivers/edac/edac_mc.c
971
e->mid_layer = mid_layer;
drivers/edac/edac_mc.c
972
e->low_layer = low_layer;
drivers/edac/edac_mc.c
973
e->page_frame_number = page_frame_number;
drivers/edac/edac_mc.c
974
e->offset_in_page = offset_in_page;
drivers/edac/edac_mc.c
975
e->syndrome = syndrome;
drivers/edac/edac_mc.c
977
e->msg = msg ?: "";
drivers/edac/edac_mc.c
978
e->other_detail = other_detail ?: "";
drivers/edac/edac_mc.h
221
void edac_raw_mc_handle_error(struct edac_raw_error_desc *e);
drivers/edac/ghes_edac.c
275
struct edac_raw_error_desc *e;
drivers/edac/ghes_edac.c
297
e = &mci->error_desc;
drivers/edac/ghes_edac.c
300
memset(e, 0, sizeof (*e));
drivers/edac/ghes_edac.c
301
e->error_count = 1;
drivers/edac/ghes_edac.c
302
e->grain = 1;
drivers/edac/ghes_edac.c
303
e->msg = pvt->msg;
drivers/edac/ghes_edac.c
304
e->other_detail = pvt->other_detail;
drivers/edac/ghes_edac.c
305
e->top_layer = -1;
drivers/edac/ghes_edac.c
306
e->mid_layer = -1;
drivers/edac/ghes_edac.c
307
e->low_layer = -1;
drivers/edac/ghes_edac.c
313
e->type = HW_EVENT_ERR_CORRECTED;
drivers/edac/ghes_edac.c
316
e->type = HW_EVENT_ERR_UNCORRECTED;
drivers/edac/ghes_edac.c
319
e->type = HW_EVENT_ERR_FATAL;
drivers/edac/ghes_edac.c
323
e->type = HW_EVENT_ERR_INFO;
drivers/edac/ghes_edac.c
341
e->page_frame_number = PHYS_PFN(mem_err->physical_addr);
drivers/edac/ghes_edac.c
342
e->offset_in_page = offset_in_page(mem_err->physical_addr);
drivers/edac/ghes_edac.c
347
e->grain = ~mem_err->physical_addr_mask + 1;
drivers/edac/ghes_edac.c
350
p = e->location;
drivers/edac/ghes_edac.c
360
e->top_layer = dimm->idx;
drivers/edac/ghes_edac.c
361
strscpy(e->label, dimm->label);
drivers/edac/ghes_edac.c
364
if (p > e->location)
drivers/edac/ghes_edac.c
367
if (!*e->label)
drivers/edac/ghes_edac.c
368
strscpy(e->label, "unknown memory");
drivers/edac/ghes_edac.c
372
p += print_mem_error_other_detail(mem_err, p, e->location, OTHER_DETAIL_LEN);
drivers/edac/ghes_edac.c
376
edac_raw_mc_handle_error(e);
drivers/firewire/core-cdev.c
1000
sizeof(e->interrupt) + header_length, NULL, 0);
drivers/firewire/core-cdev.c
1007
struct iso_interrupt_mc_event *e;
drivers/firewire/core-cdev.c
1009
e = kmalloc_obj(*e);
drivers/firewire/core-cdev.c
1010
if (e == NULL)
drivers/firewire/core-cdev.c
1013
e->interrupt.type = FW_CDEV_EVENT_ISO_INTERRUPT_MULTICHANNEL;
drivers/firewire/core-cdev.c
1014
e->interrupt.closure = client->iso_closure;
drivers/firewire/core-cdev.c
1015
e->interrupt.completed = fw_iso_buffer_lookup(&client->buffer,
drivers/firewire/core-cdev.c
1017
queue_event(client, &e->event, &e->interrupt,
drivers/firewire/core-cdev.c
1018
sizeof(e->interrupt), NULL, 0);
drivers/firewire/core-cdev.c
1298
struct iso_resource_event *e;
drivers/firewire/core-cdev.c
1367
e = r->e_alloc;
drivers/firewire/core-cdev.c
1370
e = r->e_dealloc;
drivers/firewire/core-cdev.c
1373
e->iso_resource.handle = r->resource.handle;
drivers/firewire/core-cdev.c
1374
e->iso_resource.channel = channel;
drivers/firewire/core-cdev.c
1375
e->iso_resource.bandwidth = bandwidth;
drivers/firewire/core-cdev.c
1377
queue_event(client, &e->event,
drivers/firewire/core-cdev.c
1378
&e->iso_resource, sizeof(e->iso_resource), NULL, 0);
drivers/firewire/core-cdev.c
1539
struct outbound_phy_packet_event *e =
drivers/firewire/core-cdev.c
1541
struct client *e_client = e->client;
drivers/firewire/core-cdev.c
1573
switch (e->phy_packet.without_tstamp.type) {
drivers/firewire/core-cdev.c
1576
struct fw_cdev_event_phy_packet *pp = &e->phy_packet.without_tstamp;
drivers/firewire/core-cdev.c
1580
queue_event(e->client, &e->event, &e->phy_packet, sizeof(*pp) + pp->length,
drivers/firewire/core-cdev.c
1586
struct fw_cdev_event_phy_packet2 *pp = &e->phy_packet.with_tstamp;
drivers/firewire/core-cdev.c
1590
queue_event(e->client, &e->event, &e->phy_packet, sizeof(*pp) + pp->length,
drivers/firewire/core-cdev.c
1606
struct outbound_phy_packet_event *e;
drivers/firewire/core-cdev.c
1612
e = kzalloc(sizeof(*e) + sizeof(a->data), GFP_KERNEL);
drivers/firewire/core-cdev.c
1613
if (e == NULL)
drivers/firewire/core-cdev.c
1617
e->client = client;
drivers/firewire/core-cdev.c
1618
e->p.speed = SCODE_100;
drivers/firewire/core-cdev.c
1619
e->p.generation = a->generation;
drivers/firewire/core-cdev.c
1620
async_header_set_tcode(e->p.header, TCODE_LINK_INTERNAL);
drivers/firewire/core-cdev.c
1621
e->p.header[1] = a->data[0];
drivers/firewire/core-cdev.c
1622
e->p.header[2] = a->data[1];
drivers/firewire/core-cdev.c
1623
e->p.header_length = 12;
drivers/firewire/core-cdev.c
1624
e->p.callback = outbound_phy_packet_callback;
drivers/firewire/core-cdev.c
1627
struct fw_cdev_event_phy_packet *pp = &e->phy_packet.without_tstamp;
drivers/firewire/core-cdev.c
1634
struct fw_cdev_event_phy_packet2 *pp = &e->phy_packet.with_tstamp;
drivers/firewire/core-cdev.c
1644
trace_async_phy_outbound_initiate((uintptr_t)&e->p, card->index, e->p.generation,
drivers/firewire/core-cdev.c
1645
e->p.header[1], e->p.header[2]);
drivers/firewire/core-cdev.c
1647
card->driver->send_request(card, &e->p);
drivers/firewire/core-cdev.c
1679
struct inbound_phy_packet_event *e;
drivers/firewire/core-cdev.c
1684
e = kmalloc(sizeof(*e) + 8, GFP_ATOMIC);
drivers/firewire/core-cdev.c
1685
if (e == NULL)
drivers/firewire/core-cdev.c
1689
struct fw_cdev_event_phy_packet *pp = &e->phy_packet.without_tstamp;
drivers/firewire/core-cdev.c
1697
queue_event(client, &e->event, &e->phy_packet, sizeof(*pp) + 8, NULL, 0);
drivers/firewire/core-cdev.c
1699
struct fw_cdev_event_phy_packet2 *pp = &e->phy_packet.with_tstamp;
drivers/firewire/core-cdev.c
1701
pp = &e->phy_packet.with_tstamp;
drivers/firewire/core-cdev.c
1709
queue_event(client, &e->event, &e->phy_packet, sizeof(*pp) + 8, NULL, 0);
drivers/firewire/core-cdev.c
411
struct bus_reset_event *e;
drivers/firewire/core-cdev.c
415
e = kzalloc_obj(*e);
drivers/firewire/core-cdev.c
416
if (e == NULL)
drivers/firewire/core-cdev.c
419
fill_bus_reset_event(&e->reset, client);
drivers/firewire/core-cdev.c
421
queue_event(client, &e->event,
drivers/firewire/core-cdev.c
422
&e->reset, sizeof(e->reset), NULL, 0);
drivers/firewire/core-cdev.c
573
struct outbound_transaction_event *e = data;
drivers/firewire/core-cdev.c
574
struct client *client = e->client;
drivers/firewire/core-cdev.c
575
unsigned long index = e->r.resource.handle;
drivers/firewire/core-cdev.c
583
switch (e->rsp.without_tstamp.type) {
drivers/firewire/core-cdev.c
586
struct fw_cdev_event_response *rsp = &e->rsp.without_tstamp;
drivers/firewire/core-cdev.c
601
queue_event(client, &e->event, rsp, sizeof(*rsp), rsp->data, rsp->length);
drivers/firewire/core-cdev.c
603
queue_event(client, &e->event, rsp, sizeof(*rsp) + rsp->length, NULL, 0);
drivers/firewire/core-cdev.c
609
struct fw_cdev_event_response2 *rsp = &e->rsp.with_tstamp;
drivers/firewire/core-cdev.c
620
queue_event(client, &e->event, rsp, sizeof(*rsp) + rsp->length, NULL, 0);
drivers/firewire/core-cdev.c
637
struct outbound_transaction_event *e;
drivers/firewire/core-cdev.c
649
e = kmalloc(sizeof(*e) + request->length, GFP_KERNEL);
drivers/firewire/core-cdev.c
650
if (e == NULL)
drivers/firewire/core-cdev.c
652
e->client = client;
drivers/firewire/core-cdev.c
655
struct fw_cdev_event_response *rsp = &e->rsp.without_tstamp;
drivers/firewire/core-cdev.c
662
struct fw_cdev_event_response2 *rsp = &e->rsp.with_tstamp;
drivers/firewire/core-cdev.c
675
e->r.resource.release = release_transaction;
drivers/firewire/core-cdev.c
676
ret = add_client_resource(client, &e->r.resource, GFP_KERNEL);
drivers/firewire/core-cdev.c
680
fw_send_request_with_tstamp(client->device->card, &e->r.transaction, request->tcode,
drivers/firewire/core-cdev.c
682
payload, request->length, complete_transaction, e);
drivers/firewire/core-cdev.c
686
kfree(e);
drivers/firewire/core-cdev.c
736
struct inbound_transaction_event *e;
drivers/firewire/core-cdev.c
749
e = kmalloc_obj(*e, GFP_ATOMIC);
drivers/firewire/core-cdev.c
750
if (r == NULL || e == NULL)
drivers/firewire/core-cdev.c
765
struct fw_cdev_event_request *req = &e->req.request;
drivers/firewire/core-cdev.c
778
struct fw_cdev_event_request2 *req = &e->req.request2;
drivers/firewire/core-cdev.c
792
struct fw_cdev_event_request3 *req = &e->req.with_tstamp;
drivers/firewire/core-cdev.c
808
queue_event(handler->client, &e->event,
drivers/firewire/core-cdev.c
809
&e->req, event_size0, r->data, length);
drivers/firewire/core-cdev.c
814
kfree(e);
drivers/firewire/core-cdev.c
988
struct iso_interrupt_event *e;
drivers/firewire/core-cdev.c
990
e = kmalloc(sizeof(*e) + header_length, GFP_KERNEL);
drivers/firewire/core-cdev.c
991
if (e == NULL)
drivers/firewire/core-cdev.c
994
e->interrupt.type = FW_CDEV_EVENT_ISO_INTERRUPT;
drivers/firewire/core-cdev.c
995
e->interrupt.closure = client->iso_closure;
drivers/firewire/core-cdev.c
996
e->interrupt.cycle = cycle;
drivers/firewire/core-cdev.c
997
e->interrupt.header_length = header_length;
drivers/firewire/core-cdev.c
998
memcpy(e->interrupt.header, header, header_length);
drivers/firewire/core-cdev.c
999
queue_event(client, &e->event, &e->interrupt,
drivers/firmware/arm_scmi/notify.c
108
#define MAKE_HASH_KEY(p, e, s) \
drivers/firmware/arm_scmi/notify.c
110
FIELD_PREP(EVT_ID_MASK, (e)) | \
drivers/firmware/arm_scmi/notify.c
113
#define MAKE_ALL_SRCS_KEY(p, e) MAKE_HASH_KEY((p), (e), SRC_ID_MASK)
drivers/firmware/arm_sdei.c
168
struct sdei_event *e, *found = NULL;
drivers/firmware/arm_sdei.c
173
list_for_each_entry(e, &sdei_list, list) {
drivers/firmware/arm_sdei.c
174
if (e->event_num == event_num) {
drivers/firmware/arm_sdei.c
175
found = e;
drivers/firmware/dmi_scan.c
1039
char *e;
drivers/firmware/dmi_scan.c
1057
year = simple_strtoul(y, &e, 10);
drivers/firmware/dmi_scan.c
1058
if (y != e && year < 100) { /* 2-digit year */
drivers/firmware/dmi_scan.c
1067
month = simple_strtoul(s, &e, 10);
drivers/firmware/dmi_scan.c
1068
if (s == e || *e != '/' || !month || month > 12) {
drivers/firmware/dmi_scan.c
1073
s = e + 1;
drivers/firmware/dmi_scan.c
1074
day = simple_strtoul(s, &e, 10);
drivers/firmware/dmi_scan.c
1075
if (s == y || s == e || *e != '/' || day > 31)
drivers/firmware/qcom/qcom_scm.h
29
#define QCOM_SCM_ARGS_IMPL(num, a, b, c, d, e, f, g, h, i, j, ...) (\
drivers/firmware/qcom/qcom_scm.h
34
(((e) & 0x3) << 12) | \
drivers/firmware/qemu_fw_cfg.c
405
static ssize_t fw_cfg_sysfs_show_size(struct fw_cfg_sysfs_entry *e, char *buf)
drivers/firmware/qemu_fw_cfg.c
407
return sprintf(buf, "%u\n", e->size);
drivers/firmware/qemu_fw_cfg.c
410
static ssize_t fw_cfg_sysfs_show_key(struct fw_cfg_sysfs_entry *e, char *buf)
drivers/firmware/qemu_fw_cfg.c
412
return sprintf(buf, "%u\n", e->select);
drivers/firmware/qemu_fw_cfg.c
415
static ssize_t fw_cfg_sysfs_show_name(struct fw_cfg_sysfs_entry *e, char *buf)
drivers/firmware/qemu_fw_cfg.c
417
return sprintf(buf, "%s\n", e->name);
drivers/gpu/drm/amd/amdgpu/amdgpu.h
1479
#define amdgpu_asic_query_video_codecs(adev, e, c) (adev)->asic_funcs->query_video_codecs((adev), (e), (c))
drivers/gpu/drm/amd/amdgpu/amdgpu_bo_list.c
53
struct amdgpu_bo_list_entry *e;
drivers/gpu/drm/amd/amdgpu/amdgpu_bo_list.c
55
amdgpu_bo_list_for_each_entry(e, list)
drivers/gpu/drm/amd/amdgpu/amdgpu_bo_list.c
56
amdgpu_bo_unref(&e->bo);
drivers/gpu/drm/amd/amdgpu/amdgpu_bo_list.h
73
#define amdgpu_bo_list_for_each_entry(e, list) \
drivers/gpu/drm/amd/amdgpu/amdgpu_bo_list.h
74
for (e = list->entries; \
drivers/gpu/drm/amd/amdgpu/amdgpu_bo_list.h
75
e != &list->entries[list->num_entries]; \
drivers/gpu/drm/amd/amdgpu/amdgpu_bo_list.h
76
++e)
drivers/gpu/drm/amd/amdgpu/amdgpu_bo_list.h
78
#define amdgpu_bo_list_for_each_userptr_entry(e, list) \
drivers/gpu/drm/amd/amdgpu/amdgpu_bo_list.h
79
for (e = &list->entries[list->first_userptr]; \
drivers/gpu/drm/amd/amdgpu/amdgpu_bo_list.h
80
e != &list->entries[list->num_entries]; \
drivers/gpu/drm/amd/amdgpu/amdgpu_bo_list.h
81
++e)
drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c
1099
struct amdgpu_bo_list_entry *e;
drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c
1153
amdgpu_bo_list_for_each_entry(e, p->bo_list) {
drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c
1154
bo_va = e->bo_va;
drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c
1191
amdgpu_bo_list_for_each_entry(e, p->bo_list) {
drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c
1192
struct amdgpu_bo *bo = e->bo;
drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c
1288
struct amdgpu_bo_list_entry *e;
drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c
1328
amdgpu_bo_list_for_each_userptr_entry(e, p->bo_list) {
drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c
1329
r |= !amdgpu_hmm_range_valid(e->range);
drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c
1330
amdgpu_hmm_range_free(e->range);
drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c
1331
e->range = NULL;
drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c
858
struct amdgpu_bo_list_entry *e;
drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c
887
amdgpu_bo_list_for_each_userptr_entry(e, p->bo_list) {
drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c
889
struct amdgpu_bo *bo = e->bo;
drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c
891
e->range = amdgpu_hmm_range_alloc(NULL);
drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c
892
if (unlikely(!e->range)) {
drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c
897
r = amdgpu_ttm_tt_get_user_pages(bo, e->range);
drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c
903
hmm_pfn_to_page(e->range->hmm_range.hmm_pfns[i])) {
drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c
908
e->user_invalidated = userpage_invalidated;
drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c
917
amdgpu_bo_list_for_each_entry(e, p->bo_list) {
drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c
919
r = drm_exec_prepare_obj(&p->exec, &e->bo->tbo.base,
drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c
925
e->bo_va = amdgpu_vm_bo_find(vm, e->bo);
drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c
937
amdgpu_bo_list_for_each_userptr_entry(e, p->bo_list) {
drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c
940
usermm = amdgpu_ttm_tt_get_usermm(e->bo->tbo.ttm);
drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c
946
if (amdgpu_ttm_tt_is_userptr(e->bo->tbo.ttm) &&
drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c
947
e->user_invalidated) {
drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c
948
amdgpu_bo_placement_from_domain(e->bo,
drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c
950
r = ttm_bo_validate(&e->bo->tbo, &e->bo->placement,
drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c
955
amdgpu_ttm_tt_set_user_pages(e->bo->tbo.ttm,
drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c
956
e->range);
drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c
996
amdgpu_bo_list_for_each_userptr_entry(e, p->bo_list) {
drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c
997
amdgpu_hmm_range_free(e->range);
drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c
998
e->range = NULL;
drivers/gpu/drm/amd/amdgpu/amdgpu_ctx.c
32
#define to_amdgpu_ctx_entity(e) \
drivers/gpu/drm/amd/amdgpu/amdgpu_ctx.c
33
container_of((e), struct amdgpu_ctx_entity, entity)
drivers/gpu/drm/amd/amdgpu/amdgpu_display.h
29
#define amdgpu_display_backlight_set_level(adev, e, l) (adev)->mode_info.funcs->backlight_set_level((e), (l))
drivers/gpu/drm/amd/amdgpu/amdgpu_display.h
30
#define amdgpu_display_backlight_get_level(adev, e) (adev)->mode_info.funcs->backlight_get_level((e))
drivers/gpu/drm/amd/amdgpu/amdgpu_display.h
37
#define amdgpu_display_add_encoder(adev, e, s, c) (adev)->mode_info.funcs->add_encoder((adev), (e), (s), (c))
drivers/gpu/drm/amd/amdgpu/amdgpu_eviction_fence.c
29
#define evf_mgr_to_fpriv(e) container_of(e, struct amdgpu_fpriv, evf_mgr)
drivers/gpu/drm/amd/amdgpu/amdgpu_ring_mux.c
103
list_for_each_entry(chunk, &e->list, entry) {
drivers/gpu/drm/amd/amdgpu/amdgpu_ring_mux.c
105
amdgpu_fence_update_start_timestamp(e->ring,
drivers/gpu/drm/amd/amdgpu/amdgpu_ring_mux.c
109
le32_to_cpu(*(e->ring->fence_drv.cpu_addr + 2))) {
drivers/gpu/drm/amd/amdgpu/amdgpu_ring_mux.c
110
if (chunk->cntl_offset <= e->ring->buf_mask)
drivers/gpu/drm/amd/amdgpu/amdgpu_ring_mux.c
111
amdgpu_ring_patch_cntl(e->ring,
drivers/gpu/drm/amd/amdgpu/amdgpu_ring_mux.c
113
if (chunk->ce_offset <= e->ring->buf_mask)
drivers/gpu/drm/amd/amdgpu/amdgpu_ring_mux.c
114
amdgpu_ring_patch_ce(e->ring, chunk->ce_offset);
drivers/gpu/drm/amd/amdgpu/amdgpu_ring_mux.c
115
if (chunk->de_offset <= e->ring->buf_mask)
drivers/gpu/drm/amd/amdgpu/amdgpu_ring_mux.c
116
amdgpu_ring_patch_de(e->ring, chunk->de_offset);
drivers/gpu/drm/amd/amdgpu/amdgpu_ring_mux.c
118
amdgpu_ring_mux_copy_pkt_from_sw_ring(mux, e->ring,
drivers/gpu/drm/amd/amdgpu/amdgpu_ring_mux.c
177
struct amdgpu_mux_entry *e;
drivers/gpu/drm/amd/amdgpu/amdgpu_ring_mux.c
182
e = &mux->ring_entry[i];
drivers/gpu/drm/amd/amdgpu/amdgpu_ring_mux.c
183
list_for_each_entry_safe(chunk, chunk2, &e->list, entry) {
drivers/gpu/drm/amd/amdgpu/amdgpu_ring_mux.c
197
struct amdgpu_mux_entry *e;
drivers/gpu/drm/amd/amdgpu/amdgpu_ring_mux.c
204
e = &mux->ring_entry[mux->num_ring_entries];
drivers/gpu/drm/amd/amdgpu/amdgpu_ring_mux.c
206
e->ring = ring;
drivers/gpu/drm/amd/amdgpu/amdgpu_ring_mux.c
208
INIT_LIST_HEAD(&e->list);
drivers/gpu/drm/amd/amdgpu/amdgpu_ring_mux.c
215
struct amdgpu_mux_entry *e;
drivers/gpu/drm/amd/amdgpu/amdgpu_ring_mux.c
222
e = amdgpu_ring_mux_sw_entry(mux, ring);
drivers/gpu/drm/amd/amdgpu/amdgpu_ring_mux.c
223
if (!e) {
drivers/gpu/drm/amd/amdgpu/amdgpu_ring_mux.c
235
e->sw_cptr = e->sw_wptr;
drivers/gpu/drm/amd/amdgpu/amdgpu_ring_mux.c
237
if (ring->hw_prio <= AMDGPU_RING_PRIO_DEFAULT && e->sw_cptr < mux->wptr_resubmit)
drivers/gpu/drm/amd/amdgpu/amdgpu_ring_mux.c
238
e->sw_cptr = mux->wptr_resubmit;
drivers/gpu/drm/amd/amdgpu/amdgpu_ring_mux.c
239
e->sw_wptr = wptr;
drivers/gpu/drm/amd/amdgpu/amdgpu_ring_mux.c
240
e->start_ptr_in_hw_ring = mux->real_ring->wptr;
drivers/gpu/drm/amd/amdgpu/amdgpu_ring_mux.c
244
amdgpu_ring_mux_copy_pkt_from_sw_ring(mux, ring, e->sw_cptr, wptr);
drivers/gpu/drm/amd/amdgpu/amdgpu_ring_mux.c
245
e->end_ptr_in_hw_ring = mux->real_ring->wptr;
drivers/gpu/drm/amd/amdgpu/amdgpu_ring_mux.c
248
e->end_ptr_in_hw_ring = mux->real_ring->wptr;
drivers/gpu/drm/amd/amdgpu/amdgpu_ring_mux.c
255
struct amdgpu_mux_entry *e;
drivers/gpu/drm/amd/amdgpu/amdgpu_ring_mux.c
257
e = amdgpu_ring_mux_sw_entry(mux, ring);
drivers/gpu/drm/amd/amdgpu/amdgpu_ring_mux.c
258
if (!e) {
drivers/gpu/drm/amd/amdgpu/amdgpu_ring_mux.c
263
return e->sw_wptr;
drivers/gpu/drm/amd/amdgpu/amdgpu_ring_mux.c
284
struct amdgpu_mux_entry *e;
drivers/gpu/drm/amd/amdgpu/amdgpu_ring_mux.c
287
e = amdgpu_ring_mux_sw_entry(mux, ring);
drivers/gpu/drm/amd/amdgpu/amdgpu_ring_mux.c
288
if (!e) {
drivers/gpu/drm/amd/amdgpu/amdgpu_ring_mux.c
295
start = e->start_ptr_in_hw_ring & mux->real_ring->buf_mask;
drivers/gpu/drm/amd/amdgpu/amdgpu_ring_mux.c
296
end = e->end_ptr_in_hw_ring & mux->real_ring->buf_mask;
drivers/gpu/drm/amd/amdgpu/amdgpu_ring_mux.c
305
e->sw_rptr = (e->sw_cptr + offset) & ring->buf_mask;
drivers/gpu/drm/amd/amdgpu/amdgpu_ring_mux.c
307
e->sw_rptr = e->sw_cptr;
drivers/gpu/drm/amd/amdgpu/amdgpu_ring_mux.c
310
e->sw_rptr = e->sw_wptr;
drivers/gpu/drm/amd/amdgpu/amdgpu_ring_mux.c
313
return e->sw_rptr;
drivers/gpu/drm/amd/amdgpu/amdgpu_ring_mux.c
435
struct amdgpu_mux_entry *e;
drivers/gpu/drm/amd/amdgpu/amdgpu_ring_mux.c
442
e = amdgpu_ring_mux_sw_entry(mux, ring);
drivers/gpu/drm/amd/amdgpu/amdgpu_ring_mux.c
443
if (!e) {
drivers/gpu/drm/amd/amdgpu/amdgpu_ring_mux.c
459
list_add_tail(&chunk->entry, &e->list);
drivers/gpu/drm/amd/amdgpu/amdgpu_ring_mux.c
465
struct amdgpu_mux_entry *e;
drivers/gpu/drm/amd/amdgpu/amdgpu_ring_mux.c
468
e = amdgpu_ring_mux_sw_entry(mux, ring);
drivers/gpu/drm/amd/amdgpu/amdgpu_ring_mux.c
469
if (!e) {
drivers/gpu/drm/amd/amdgpu/amdgpu_ring_mux.c
476
list_for_each_entry_safe(chunk, tmp, &e->list, entry) {
drivers/gpu/drm/amd/amdgpu/amdgpu_ring_mux.c
488
struct amdgpu_mux_entry *e;
drivers/gpu/drm/amd/amdgpu/amdgpu_ring_mux.c
491
e = amdgpu_ring_mux_sw_entry(mux, ring);
drivers/gpu/drm/amd/amdgpu/amdgpu_ring_mux.c
492
if (!e) {
drivers/gpu/drm/amd/amdgpu/amdgpu_ring_mux.c
497
chunk = list_last_entry(&e->list, struct amdgpu_mux_chunk, entry);
drivers/gpu/drm/amd/amdgpu/amdgpu_ring_mux.c
521
struct amdgpu_mux_entry *e;
drivers/gpu/drm/amd/amdgpu/amdgpu_ring_mux.c
524
e = amdgpu_ring_mux_sw_entry(mux, ring);
drivers/gpu/drm/amd/amdgpu/amdgpu_ring_mux.c
525
if (!e) {
drivers/gpu/drm/amd/amdgpu/amdgpu_ring_mux.c
530
chunk = list_last_entry(&e->list, struct amdgpu_mux_chunk, entry);
drivers/gpu/drm/amd/amdgpu/amdgpu_ring_mux.c
544
struct amdgpu_mux_entry *e;
drivers/gpu/drm/amd/amdgpu/amdgpu_ring_mux.c
555
e = &mux->ring_entry[i];
drivers/gpu/drm/amd/amdgpu/amdgpu_ring_mux.c
556
if (e->ring->hw_prio <= AMDGPU_RING_PRIO_DEFAULT) {
drivers/gpu/drm/amd/amdgpu/amdgpu_ring_mux.c
557
ring = e->ring;
drivers/gpu/drm/amd/amdgpu/amdgpu_ring_mux.c
78
struct amdgpu_mux_entry *e = NULL;
drivers/gpu/drm/amd/amdgpu/amdgpu_ring_mux.c
89
e = &mux->ring_entry[i];
drivers/gpu/drm/amd/amdgpu/amdgpu_ring_mux.c
94
if (!e) {
drivers/gpu/drm/amd/amdgpu/amdgpu_ring_mux.c
99
last_seq = atomic_read(&e->ring->fence_drv.last_seq);
drivers/gpu/drm/amd/amdgpu/amdgpu_sync.c
135
struct amdgpu_sync_entry *e;
drivers/gpu/drm/amd/amdgpu/amdgpu_sync.c
137
hash_for_each_possible(sync->fences, e, node, f->context) {
drivers/gpu/drm/amd/amdgpu/amdgpu_sync.c
138
if (dma_fence_is_signaled(e->fence)) {
drivers/gpu/drm/amd/amdgpu/amdgpu_sync.c
139
dma_fence_put(e->fence);
drivers/gpu/drm/amd/amdgpu/amdgpu_sync.c
140
e->fence = dma_fence_get(f);
drivers/gpu/drm/amd/amdgpu/amdgpu_sync.c
144
if (likely(e->fence->context == f->context)) {
drivers/gpu/drm/amd/amdgpu/amdgpu_sync.c
145
amdgpu_sync_keep_later(&e->fence, f);
drivers/gpu/drm/amd/amdgpu/amdgpu_sync.c
164
struct amdgpu_sync_entry *e;
drivers/gpu/drm/amd/amdgpu/amdgpu_sync.c
172
e = kmem_cache_alloc(amdgpu_sync_slab, flags);
drivers/gpu/drm/amd/amdgpu/amdgpu_sync.c
173
if (!e)
drivers/gpu/drm/amd/amdgpu/amdgpu_sync.c
176
hash_add(sync->fences, &e->node, f->context);
drivers/gpu/drm/amd/amdgpu/amdgpu_sync.c
177
e->fence = dma_fence_get(f);
drivers/gpu/drm/amd/amdgpu/amdgpu_sync.c
300
static void amdgpu_sync_entry_free(struct amdgpu_sync_entry *e)
drivers/gpu/drm/amd/amdgpu/amdgpu_sync.c
302
hash_del(&e->node);
drivers/gpu/drm/amd/amdgpu/amdgpu_sync.c
303
dma_fence_put(e->fence);
drivers/gpu/drm/amd/amdgpu/amdgpu_sync.c
304
kmem_cache_free(amdgpu_sync_slab, e);
drivers/gpu/drm/amd/amdgpu/amdgpu_sync.c
319
struct amdgpu_sync_entry *e;
drivers/gpu/drm/amd/amdgpu/amdgpu_sync.c
323
hash_for_each_safe(sync->fences, i, tmp, e, node) {
drivers/gpu/drm/amd/amdgpu/amdgpu_sync.c
324
struct dma_fence *f = e->fence;
drivers/gpu/drm/amd/amdgpu/amdgpu_sync.c
328
amdgpu_sync_entry_free(e);
drivers/gpu/drm/amd/amdgpu/amdgpu_sync.c
358
struct amdgpu_sync_entry *e;
drivers/gpu/drm/amd/amdgpu/amdgpu_sync.c
363
hash_for_each_safe(sync->fences, i, tmp, e, node) {
drivers/gpu/drm/amd/amdgpu/amdgpu_sync.c
365
f = e->fence;
drivers/gpu/drm/amd/amdgpu/amdgpu_sync.c
367
hash_del(&e->node);
drivers/gpu/drm/amd/amdgpu/amdgpu_sync.c
368
kmem_cache_free(amdgpu_sync_slab, e);
drivers/gpu/drm/amd/amdgpu/amdgpu_sync.c
389
struct amdgpu_sync_entry *e;
drivers/gpu/drm/amd/amdgpu/amdgpu_sync.c
394
hash_for_each_safe(source->fences, i, tmp, e, node) {
drivers/gpu/drm/amd/amdgpu/amdgpu_sync.c
395
f = e->fence;
drivers/gpu/drm/amd/amdgpu/amdgpu_sync.c
401
amdgpu_sync_entry_free(e);
drivers/gpu/drm/amd/amdgpu/amdgpu_sync.c
436
struct amdgpu_sync_entry *e;
drivers/gpu/drm/amd/amdgpu/amdgpu_sync.c
441
hash_for_each_safe(sync->fences, i, tmp, e, node) {
drivers/gpu/drm/amd/amdgpu/amdgpu_sync.c
442
f = e->fence;
drivers/gpu/drm/amd/amdgpu/amdgpu_sync.c
444
amdgpu_sync_entry_free(e);
drivers/gpu/drm/amd/amdgpu/amdgpu_sync.c
460
struct amdgpu_sync_entry *e;
drivers/gpu/drm/amd/amdgpu/amdgpu_sync.c
464
hash_for_each_safe(sync->fences, i, tmp, e, node) {
drivers/gpu/drm/amd/amdgpu/amdgpu_sync.c
465
r = dma_fence_wait(e->fence, intr);
drivers/gpu/drm/amd/amdgpu/amdgpu_sync.c
469
amdgpu_sync_entry_free(e);
drivers/gpu/drm/amd/amdgpu/amdgpu_sync.c
484
struct amdgpu_sync_entry *e;
drivers/gpu/drm/amd/amdgpu/amdgpu_sync.c
488
hash_for_each_safe(sync->fences, i, tmp, e, node)
drivers/gpu/drm/amd/amdgpu/amdgpu_sync.c
489
amdgpu_sync_entry_free(e);
drivers/gpu/drm/amd/amdgpu/cikd.h
495
#define SDMA_PACKET(op, sub_op, e) ((((e) & 0xFFFF) << 16) | \
drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c
518
#define AMDGPU_RAS_SUB_BLOCK(subblock, a, b, c, d, e, f, g, h) \
drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c
523
(((e) << 1) | ((f) << 3) | (g) | ((h) << 2)), \
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
1772
unsigned long e, s;
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
1778
e = min(end, prange->last);
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
1779
if (e >= s)
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
1780
r = svm_range_unmap_from_gpus(prange, s, e,
drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c
438
struct drm_pending_vblank_event *e;
drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c
463
e = amdgpu_crtc->event;
drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c
466
WARN_ON(!e);
drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c
484
if (e) {
drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c
485
drm_crtc_send_vblank_event(&amdgpu_crtc->base, e);
drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c
490
} else if (e) {
drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c
505
e->sequence = drm_crtc_vblank_count(&amdgpu_crtc->base);
drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c
506
e->pipe = amdgpu_crtc->crtc_id;
drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c
508
list_add_tail(&e->base.link, &adev_to_drm(adev)->vblank_event_list);
drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c
509
e = NULL;
drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c
525
amdgpu_crtc->crtc_id, amdgpu_crtc, vrr_active, (int)!e);
drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c
954
static const char *dmub_notification_type_str(enum dmub_notification_type e)
drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c
956
switch (e) {
drivers/gpu/drm/amd/display/dc/dce/dce_clock_source.c
1075
const struct pixel_rate_range_table_entry *e =
drivers/gpu/drm/amd/display/dc/dce/dce_clock_source.c
1091
if (e) {
drivers/gpu/drm/amd/display/dc/dce/dce_clock_source.c
1092
dto_params.pixclk_hz = e->target_pixel_rate_khz;
drivers/gpu/drm/amd/display/dc/dce/dce_clock_source.c
1093
dto_params.pixclk_hz *= e->mult_factor;
drivers/gpu/drm/amd/display/dc/dce/dce_clock_source.c
1095
dto_params.refclk_hz *= e->div_factor;
drivers/gpu/drm/amd/display/dc/dce/dce_clock_source.c
1271
const struct pixel_rate_range_table_entry *e = &video_optimized_pixel_rates[i];
drivers/gpu/drm/amd/display/dc/dce/dce_clock_source.c
1273
if (e->range_min_khz <= pixel_rate_khz && pixel_rate_khz <= e->range_max_khz) {
drivers/gpu/drm/amd/display/dc/dce/dce_clock_source.c
1274
return e;
drivers/gpu/drm/amd/display/dc/dce/dce_clock_source.c
1336
const struct pixel_rate_range_table_entry *e =
drivers/gpu/drm/amd/display/dc/dce/dce_clock_source.c
1341
if (e) {
drivers/gpu/drm/amd/display/dc/dce/dce_clock_source.c
1343
REG_WRITE(PHASE[inst], e->target_pixel_rate_khz * e->mult_factor);
drivers/gpu/drm/amd/display/dc/dce/dce_clock_source.c
1344
REG_WRITE(MODULO[inst], dp_dto_ref_khz * e->div_factor);
drivers/gpu/drm/amd/display/dc/dce/dce_clock_source.c
974
const struct pixel_rate_range_table_entry *e =
drivers/gpu/drm/amd/display/dc/dce/dce_clock_source.c
987
if (e) {
drivers/gpu/drm/amd/display/dc/dce/dce_clock_source.c
989
REG_WRITE(PHASE[inst], e->target_pixel_rate_khz * e->mult_factor);
drivers/gpu/drm/amd/display/dc/dce/dce_clock_source.c
990
REG_WRITE(MODULO[inst], dp_dto_ref_khz * e->div_factor);
drivers/gpu/drm/amd/display/dc/dml/dml_inline_defs.h
62
static inline double dml_max5(double a, double b, double c, double d, double e)
drivers/gpu/drm/amd/display/dc/dml/dml_inline_defs.h
64
return dml_max(dml_max4(a, b, c, d), e);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_util.c
165
dml_float_t dml_max5(dml_float_t a, dml_float_t b, dml_float_t c, dml_float_t d, dml_float_t e)
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_util.c
167
return dml_max(dml_max4(a, b, c, d), e);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_util.h
45
__DML_DLL_EXPORT__ dml_float_t dml_max5(dml_float_t a, dml_float_t b, dml_float_t c, dml_float_t d, dml_float_t e);
drivers/gpu/drm/amd/pm/amdgpu_dpm.c
36
#define amdgpu_dpm_enable_bapm(adev, e) \
drivers/gpu/drm/amd/pm/amdgpu_dpm.c
37
((adev)->powerplay.pp_funcs->enable_bapm((adev)->powerplay.pp_handle, (e)))
drivers/gpu/drm/amd/ras/rascore/ras_gfx_v9_0.c
237
#define RAS_GFX_SUB_BLOCK(subblock, a, b, c, d, e, f, g, h) \
drivers/gpu/drm/amd/ras/rascore/ras_gfx_v9_0.c
242
(((e) << 1) | ((f) << 3) | (g) | ((h) << 2)), \
drivers/gpu/drm/bridge/cadence/cdns-dsi-core.c
393
#define STS_CTL_EDGE(e) ((e) << 16)
drivers/gpu/drm/bridge/parade-ps8640.c
147
static inline struct ps8640 *bridge_to_ps8640(struct drm_bridge *e)
drivers/gpu/drm/bridge/parade-ps8640.c
149
return container_of(e, struct ps8640, bridge);
drivers/gpu/drm/drm_atomic_uapi.c
1089
struct drm_pending_vblank_event *e = NULL;
drivers/gpu/drm/drm_atomic_uapi.c
1091
e = kzalloc_obj(*e);
drivers/gpu/drm/drm_atomic_uapi.c
1092
if (!e)
drivers/gpu/drm/drm_atomic_uapi.c
1095
e->event.base.type = DRM_EVENT_FLIP_COMPLETE;
drivers/gpu/drm/drm_atomic_uapi.c
1096
e->event.base.length = sizeof(e->event);
drivers/gpu/drm/drm_atomic_uapi.c
1097
e->event.vbl.crtc_id = crtc->base.id;
drivers/gpu/drm/drm_atomic_uapi.c
1098
e->event.vbl.user_data = user_data;
drivers/gpu/drm/drm_atomic_uapi.c
1100
return e;
drivers/gpu/drm/drm_atomic_uapi.c
1394
struct drm_pending_vblank_event *e;
drivers/gpu/drm/drm_atomic_uapi.c
1396
e = create_vblank_event(crtc, arg->user_data);
drivers/gpu/drm/drm_atomic_uapi.c
1397
if (!e)
drivers/gpu/drm/drm_atomic_uapi.c
1400
crtc_state->event = e;
drivers/gpu/drm/drm_atomic_uapi.c
1404
struct drm_pending_vblank_event *e = crtc_state->event;
drivers/gpu/drm/drm_atomic_uapi.c
1409
ret = drm_event_reserve_init(dev, file_priv, &e->base,
drivers/gpu/drm/drm_atomic_uapi.c
1410
&e->event.base);
drivers/gpu/drm/drm_atomic_uapi.c
1412
kfree(e);
drivers/gpu/drm/drm_file.c
202
struct drm_pending_event *e, *et;
drivers/gpu/drm/drm_file.c
208
list_for_each_entry_safe(e, et, &file_priv->pending_event_list,
drivers/gpu/drm/drm_file.c
210
list_del(&e->pending_link);
drivers/gpu/drm/drm_file.c
211
e->file_priv = NULL;
drivers/gpu/drm/drm_file.c
215
list_for_each_entry_safe(e, et, &file_priv->event_list, link) {
drivers/gpu/drm/drm_file.c
216
list_del(&e->link);
drivers/gpu/drm/drm_file.c
217
kfree(e);
drivers/gpu/drm/drm_file.c
552
struct drm_pending_event *e = NULL;
drivers/gpu/drm/drm_file.c
556
e = list_first_entry(&file_priv->event_list,
drivers/gpu/drm/drm_file.c
558
file_priv->event_space += e->event->length;
drivers/gpu/drm/drm_file.c
559
list_del(&e->link);
drivers/gpu/drm/drm_file.c
563
if (e == NULL) {
drivers/gpu/drm/drm_file.c
580
unsigned length = e->event->length;
drivers/gpu/drm/drm_file.c
586
list_add(&e->link, &file_priv->event_list);
drivers/gpu/drm/drm_file.c
593
if (copy_to_user(buffer + ret, e->event, length)) {
drivers/gpu/drm/drm_file.c
600
kfree(e);
drivers/gpu/drm/drm_file.c
664
struct drm_event *e)
drivers/gpu/drm/drm_file.c
666
if (file_priv->event_space < e->length)
drivers/gpu/drm/drm_file.c
669
file_priv->event_space -= e->length;
drivers/gpu/drm/drm_file.c
671
p->event = e;
drivers/gpu/drm/drm_file.c
705
struct drm_event *e)
drivers/gpu/drm/drm_file.c
711
ret = drm_event_reserve_init_locked(dev, file_priv, p, e);
drivers/gpu/drm/drm_file.c
747
struct drm_pending_event *e, ktime_t timestamp)
drivers/gpu/drm/drm_file.c
751
if (e->completion) {
drivers/gpu/drm/drm_file.c
752
complete_all(e->completion);
drivers/gpu/drm/drm_file.c
753
e->completion_release(e->completion);
drivers/gpu/drm/drm_file.c
754
e->completion = NULL;
drivers/gpu/drm/drm_file.c
757
if (e->fence) {
drivers/gpu/drm/drm_file.c
759
dma_fence_signal_timestamp(e->fence, timestamp);
drivers/gpu/drm/drm_file.c
761
dma_fence_signal(e->fence);
drivers/gpu/drm/drm_file.c
762
dma_fence_put(e->fence);
drivers/gpu/drm/drm_file.c
765
if (!e->file_priv) {
drivers/gpu/drm/drm_file.c
766
kfree(e);
drivers/gpu/drm/drm_file.c
770
list_del(&e->pending_link);
drivers/gpu/drm/drm_file.c
771
list_add_tail(&e->link,
drivers/gpu/drm/drm_file.c
772
&e->file_priv->event_list);
drivers/gpu/drm/drm_file.c
773
wake_up_interruptible_poll(&e->file_priv->event_wait,
drivers/gpu/drm/drm_file.c
794
struct drm_pending_event *e, ktime_t timestamp)
drivers/gpu/drm/drm_file.c
796
drm_send_event_helper(dev, e, timestamp);
drivers/gpu/drm/drm_file.c
814
void drm_send_event_locked(struct drm_device *dev, struct drm_pending_event *e)
drivers/gpu/drm/drm_file.c
816
drm_send_event_helper(dev, e, 0);
drivers/gpu/drm/drm_file.c
835
void drm_send_event(struct drm_device *dev, struct drm_pending_event *e)
drivers/gpu/drm/drm_file.c
840
drm_send_event_helper(dev, e, 0);
drivers/gpu/drm/drm_plane.c
1388
struct drm_pending_vblank_event *e = NULL;
drivers/gpu/drm/drm_plane.c
1517
e = kzalloc_obj(*e);
drivers/gpu/drm/drm_plane.c
1518
if (!e) {
drivers/gpu/drm/drm_plane.c
1523
e->event.base.type = DRM_EVENT_FLIP_COMPLETE;
drivers/gpu/drm/drm_plane.c
1524
e->event.base.length = sizeof(e->event);
drivers/gpu/drm/drm_plane.c
1525
e->event.vbl.user_data = page_flip->user_data;
drivers/gpu/drm/drm_plane.c
1526
e->event.vbl.crtc_id = crtc->base.id;
drivers/gpu/drm/drm_plane.c
1528
ret = drm_event_reserve_init(dev, file_priv, &e->base, &e->event.base);
drivers/gpu/drm/drm_plane.c
1530
kfree(e);
drivers/gpu/drm/drm_plane.c
1531
e = NULL;
drivers/gpu/drm/drm_plane.c
1538
ret = crtc->funcs->page_flip_target(crtc, fb, e,
drivers/gpu/drm/drm_plane.c
1543
ret = crtc->funcs->page_flip(crtc, fb, e, page_flip->flags,
drivers/gpu/drm/drm_plane.c
1547
drm_event_cancel_free(dev, &e->base);
drivers/gpu/drm/drm_vblank.c
1040
struct drm_pending_vblank_event *e,
drivers/gpu/drm/drm_vblank.c
1045
switch (e->event.base.type) {
drivers/gpu/drm/drm_vblank.c
1049
e->event.vbl.sequence = seq;
drivers/gpu/drm/drm_vblank.c
1055
e->event.vbl.tv_sec = tv.tv_sec;
drivers/gpu/drm/drm_vblank.c
1056
e->event.vbl.tv_usec = tv.tv_nsec / 1000;
drivers/gpu/drm/drm_vblank.c
1060
e->event.seq.sequence = seq;
drivers/gpu/drm/drm_vblank.c
1061
e->event.seq.time_ns = ktime_to_ns(now);
drivers/gpu/drm/drm_vblank.c
1064
trace_drm_vblank_event_delivered(e->base.file_priv, e->pipe, seq);
drivers/gpu/drm/drm_vblank.c
1072
drm_send_event_timestamp_locked(dev, &e->base, now);
drivers/gpu/drm/drm_vblank.c
1114
struct drm_pending_vblank_event *e)
drivers/gpu/drm/drm_vblank.c
1121
e->pipe = pipe;
drivers/gpu/drm/drm_vblank.c
1122
e->sequence = drm_crtc_accurate_vblank_count(crtc) + 1;
drivers/gpu/drm/drm_vblank.c
1123
list_add_tail(&e->base.link, &dev->vblank_event_list);
drivers/gpu/drm/drm_vblank.c
1139
struct drm_pending_vblank_event *e)
drivers/gpu/drm/drm_vblank.c
1153
e->pipe = pipe;
drivers/gpu/drm/drm_vblank.c
1154
send_vblank_event(dev, e, seq, now);
drivers/gpu/drm/drm_vblank.c
1344
struct drm_pending_vblank_event *e, *t;
drivers/gpu/drm/drm_vblank.c
1381
list_for_each_entry_safe(e, t, &dev->vblank_event_list, base.link) {
drivers/gpu/drm/drm_vblank.c
1382
if (e->pipe != pipe)
drivers/gpu/drm/drm_vblank.c
1386
e->sequence, seq);
drivers/gpu/drm/drm_vblank.c
1387
list_del(&e->base.link);
drivers/gpu/drm/drm_vblank.c
1389
send_vblank_event(dev, e, seq, now);
drivers/gpu/drm/drm_vblank.c
1613
struct drm_pending_vblank_event *e;
drivers/gpu/drm/drm_vblank.c
1618
e = kzalloc_obj(*e);
drivers/gpu/drm/drm_vblank.c
1619
if (e == NULL) {
drivers/gpu/drm/drm_vblank.c
1624
e->pipe = pipe;
drivers/gpu/drm/drm_vblank.c
1625
e->event.base.type = DRM_EVENT_VBLANK;
drivers/gpu/drm/drm_vblank.c
1626
e->event.base.length = sizeof(e->event.vbl);
drivers/gpu/drm/drm_vblank.c
1627
e->event.vbl.user_data = vblwait->request.signal;
drivers/gpu/drm/drm_vblank.c
1628
e->event.vbl.crtc_id = 0;
drivers/gpu/drm/drm_vblank.c
1633
e->event.vbl.crtc_id = crtc->base.id;
drivers/gpu/drm/drm_vblank.c
1649
ret = drm_event_reserve_init_locked(dev, file_priv, &e->base,
drivers/gpu/drm/drm_vblank.c
1650
&e->event.base);
drivers/gpu/drm/drm_vblank.c
1662
e->sequence = req_seq;
drivers/gpu/drm/drm_vblank.c
1665
send_vblank_event(dev, e, seq, now);
drivers/gpu/drm/drm_vblank.c
1669
list_add_tail(&e->base.link, &dev->vblank_event_list);
drivers/gpu/drm/drm_vblank.c
1679
kfree(e);
drivers/gpu/drm/drm_vblank.c
1880
struct drm_pending_vblank_event *e, *t;
drivers/gpu/drm/drm_vblank.c
1888
list_for_each_entry_safe(e, t, &dev->vblank_event_list, base.link) {
drivers/gpu/drm/drm_vblank.c
1889
if (e->pipe != pipe)
drivers/gpu/drm/drm_vblank.c
1891
if (!drm_vblank_passed(seq, e->sequence))
drivers/gpu/drm/drm_vblank.c
1895
e->sequence, seq);
drivers/gpu/drm/drm_vblank.c
1897
list_del(&e->base.link);
drivers/gpu/drm/drm_vblank.c
1899
send_vblank_event(dev, e, seq, now);
drivers/gpu/drm/drm_vblank.c
2071
struct drm_pending_vblank_event *e;
drivers/gpu/drm/drm_vblank.c
2097
e = kzalloc_obj(*e);
drivers/gpu/drm/drm_vblank.c
2098
if (e == NULL)
drivers/gpu/drm/drm_vblank.c
2118
e->pipe = pipe;
drivers/gpu/drm/drm_vblank.c
2119
e->event.base.type = DRM_EVENT_CRTC_SEQUENCE;
drivers/gpu/drm/drm_vblank.c
2120
e->event.base.length = sizeof(e->event.seq);
drivers/gpu/drm/drm_vblank.c
2121
e->event.seq.user_data = queue_seq->user_data;
drivers/gpu/drm/drm_vblank.c
2136
ret = drm_event_reserve_init_locked(dev, file_priv, &e->base,
drivers/gpu/drm/drm_vblank.c
2137
&e->event.base);
drivers/gpu/drm/drm_vblank.c
2142
e->sequence = req_seq;
drivers/gpu/drm/drm_vblank.c
2146
send_vblank_event(dev, e, seq, now);
drivers/gpu/drm/drm_vblank.c
2150
list_add_tail(&e->base.link, &dev->vblank_event_list);
drivers/gpu/drm/drm_vblank.c
2161
kfree(e);
drivers/gpu/drm/etnaviv/etnaviv_flop_reset.c
148
const struct etnaviv_flop_reset_entry *e = etnaviv_flop_reset_db;
drivers/gpu/drm/etnaviv/etnaviv_flop_reset.c
150
for (int i = 0; i < ARRAY_SIZE(etnaviv_flop_reset_db); ++i, ++e) {
drivers/gpu/drm/etnaviv/etnaviv_flop_reset.c
151
if (chip_id->model == e->chip_model &&
drivers/gpu/drm/etnaviv/etnaviv_flop_reset.c
152
chip_id->revision == e->revision)
drivers/gpu/drm/exynos/exynos5433_drm_decon.c
397
#define BIT_VAL(x, e, s) (((x) & ((1 << ((e) - (s) + 1)) - 1)) << (s))
drivers/gpu/drm/exynos/exynos_drm_dpi.c
38
static inline struct exynos_dpi *encoder_to_dpi(struct drm_encoder *e)
drivers/gpu/drm/exynos/exynos_drm_dpi.c
40
return container_of(e, struct exynos_dpi, encoder);
drivers/gpu/drm/exynos/exynos_drm_g2d.c
1147
struct drm_exynos_pending_g2d_event *e;
drivers/gpu/drm/exynos/exynos_drm_g2d.c
1171
e = kzalloc_obj(*node->event);
drivers/gpu/drm/exynos/exynos_drm_g2d.c
1172
if (!e) {
drivers/gpu/drm/exynos/exynos_drm_g2d.c
1177
e->event.base.type = DRM_EXYNOS_G2D_EVENT;
drivers/gpu/drm/exynos/exynos_drm_g2d.c
1178
e->event.base.length = sizeof(e->event);
drivers/gpu/drm/exynos/exynos_drm_g2d.c
1179
e->event.user_data = req->user_data;
drivers/gpu/drm/exynos/exynos_drm_g2d.c
1181
ret = drm_event_reserve_init(drm_dev, file, &e->base, &e->event.base);
drivers/gpu/drm/exynos/exynos_drm_g2d.c
1183
kfree(e);
drivers/gpu/drm/exynos/exynos_drm_g2d.c
1187
node->event = e;
drivers/gpu/drm/exynos/exynos_drm_g2d.c
916
struct drm_exynos_pending_g2d_event *e;
drivers/gpu/drm/exynos/exynos_drm_g2d.c
922
e = list_first_entry(&runqueue_node->event_list,
drivers/gpu/drm/exynos/exynos_drm_g2d.c
926
e->event.tv_sec = now.tv_sec;
drivers/gpu/drm/exynos/exynos_drm_g2d.c
927
e->event.tv_usec = now.tv_nsec / NSEC_PER_USEC;
drivers/gpu/drm/exynos/exynos_drm_g2d.c
928
e->event.cmdlist_no = cmdlist_no;
drivers/gpu/drm/exynos/exynos_drm_g2d.c
930
drm_send_event(drm_dev, &e->base);
drivers/gpu/drm/exynos/exynos_drm_ipp.c
699
struct drm_pending_exynos_ipp_event *e = NULL;
drivers/gpu/drm/exynos/exynos_drm_ipp.c
702
e = kzalloc_obj(*e);
drivers/gpu/drm/exynos/exynos_drm_ipp.c
703
if (!e)
drivers/gpu/drm/exynos/exynos_drm_ipp.c
706
e->event.base.type = DRM_EXYNOS_IPP_EVENT;
drivers/gpu/drm/exynos/exynos_drm_ipp.c
707
e->event.base.length = sizeof(e->event);
drivers/gpu/drm/exynos/exynos_drm_ipp.c
708
e->event.user_data = user_data;
drivers/gpu/drm/exynos/exynos_drm_ipp.c
710
ret = drm_event_reserve_init(task->ipp->drm_dev, file_priv, &e->base,
drivers/gpu/drm/exynos/exynos_drm_ipp.c
711
&e->event.base);
drivers/gpu/drm/exynos/exynos_drm_ipp.c
715
task->event = e;
drivers/gpu/drm/exynos/exynos_drm_ipp.c
718
kfree(e);
drivers/gpu/drm/exynos/exynos_drm_mic.c
310
struct drm_encoder *e, *encoder = NULL;
drivers/gpu/drm/exynos/exynos_drm_mic.c
312
drm_for_each_encoder(e, drm_dev)
drivers/gpu/drm/exynos/exynos_drm_mic.c
313
if (e->possible_crtcs == drm_crtc_mask(&crtc->base))
drivers/gpu/drm/exynos/exynos_drm_mic.c
314
encoder = e;
drivers/gpu/drm/exynos/exynos_drm_vidi.c
53
static inline struct vidi_context *encoder_to_vidi(struct drm_encoder *e)
drivers/gpu/drm/exynos/exynos_drm_vidi.c
55
return container_of(e, struct vidi_context, encoder);
drivers/gpu/drm/exynos/exynos_hdmi.c
148
static inline struct hdmi_context *encoder_to_hdmi(struct drm_encoder *e)
drivers/gpu/drm/exynos/exynos_hdmi.c
150
return container_of(e, struct hdmi_context, encoder);
drivers/gpu/drm/i915/gem/i915_gem_context.c
1009
static void __unpin_engines(struct i915_gem_engines *e, unsigned int count)
drivers/gpu/drm/i915/gem/i915_gem_context.c
1012
struct intel_context *ce = e->engines[count], *child;
drivers/gpu/drm/i915/gem/i915_gem_context.c
1023
static void unpin_engines(struct i915_gem_engines *e)
drivers/gpu/drm/i915/gem/i915_gem_context.c
1025
__unpin_engines(e, e->num_engines);
drivers/gpu/drm/i915/gem/i915_gem_context.c
1028
static void __free_engines(struct i915_gem_engines *e, unsigned int count)
drivers/gpu/drm/i915/gem/i915_gem_context.c
1031
if (!e->engines[count])
drivers/gpu/drm/i915/gem/i915_gem_context.c
1034
intel_context_put(e->engines[count]);
drivers/gpu/drm/i915/gem/i915_gem_context.c
1036
kfree(e);
drivers/gpu/drm/i915/gem/i915_gem_context.c
1039
static void free_engines(struct i915_gem_engines *e)
drivers/gpu/drm/i915/gem/i915_gem_context.c
1041
__free_engines(e, e->num_engines);
drivers/gpu/drm/i915/gem/i915_gem_context.c
1104
struct i915_gem_engines *e;
drivers/gpu/drm/i915/gem/i915_gem_context.c
1106
e = kzalloc_flex(*e, engines, count);
drivers/gpu/drm/i915/gem/i915_gem_context.c
1107
if (!e)
drivers/gpu/drm/i915/gem/i915_gem_context.c
1110
i915_sw_fence_init(&e->fence, engines_notify);
drivers/gpu/drm/i915/gem/i915_gem_context.c
1111
return e;
drivers/gpu/drm/i915/gem/i915_gem_context.c
1119
struct i915_gem_engines *e, *err;
drivers/gpu/drm/i915/gem/i915_gem_context.c
1121
e = alloc_engines(max);
drivers/gpu/drm/i915/gem/i915_gem_context.c
1122
if (!e)
drivers/gpu/drm/i915/gem/i915_gem_context.c
1134
GEM_BUG_ON(e->engines[engine->legacy_idx]);
drivers/gpu/drm/i915/gem/i915_gem_context.c
1142
e->engines[engine->legacy_idx] = ce;
drivers/gpu/drm/i915/gem/i915_gem_context.c
1143
e->num_engines = max(e->num_engines, engine->legacy_idx + 1);
drivers/gpu/drm/i915/gem/i915_gem_context.c
1156
return e;
drivers/gpu/drm/i915/gem/i915_gem_context.c
1159
free_engines(e);
drivers/gpu/drm/i915/gem/i915_gem_context.c
1201
struct i915_gem_engines *e, *err;
drivers/gpu/drm/i915/gem/i915_gem_context.c
1204
e = alloc_engines(num_engines);
drivers/gpu/drm/i915/gem/i915_gem_context.c
1205
if (!e)
drivers/gpu/drm/i915/gem/i915_gem_context.c
1207
e->num_engines = num_engines;
drivers/gpu/drm/i915/gem/i915_gem_context.c
1240
e->engines[n] = ce;
drivers/gpu/drm/i915/gem/i915_gem_context.c
1273
return e;
drivers/gpu/drm/i915/gem/i915_gem_context.c
1276
free_engines(e);
drivers/gpu/drm/i915/gem/i915_gem_context.c
1608
struct i915_gem_engines *e;
drivers/gpu/drm/i915/gem/i915_gem_context.c
1650
e = user_engines(ctx, pc->num_user_engines, pc->user_engines);
drivers/gpu/drm/i915/gem/i915_gem_context.c
1653
e = default_engines(ctx, pc->legacy_rcs_sseu);
drivers/gpu/drm/i915/gem/i915_gem_context.c
1655
if (IS_ERR(e)) {
drivers/gpu/drm/i915/gem/i915_gem_context.c
1656
err = PTR_ERR(e);
drivers/gpu/drm/i915/gem/i915_gem_context.c
1659
RCU_INIT_POINTER(ctx->engines, e);
drivers/gpu/drm/i915/gem/i915_gem_context.c
1690
free_engines(e);
drivers/gpu/drm/i915/gem/i915_gem_context.c
2671
const struct i915_gem_engines *e = it->engines;
drivers/gpu/drm/i915/gem/i915_gem_context.c
2674
if (unlikely(!e))
drivers/gpu/drm/i915/gem/i915_gem_context.c
2678
if (it->idx >= e->num_engines)
drivers/gpu/drm/i915/gem/i915_gem_context.c
2681
ctx = e->engines[it->idx++];
drivers/gpu/drm/i915/gem/i915_gem_context.h
210
struct i915_gem_engines *e = rcu_dereference(ctx->engines);
drivers/gpu/drm/i915/gem/i915_gem_context.h
211
if (unlikely(!e)) /* context was closed! */
drivers/gpu/drm/i915/gem/i915_gem_context.h
213
else if (likely(idx < e->num_engines && e->engines[idx]))
drivers/gpu/drm/i915/gem/i915_gem_context.h
214
ce = intel_context_get(e->engines[idx]);
drivers/gpu/drm/i915/gem/selftests/mock_context.c
17
struct i915_gem_engines *e;
drivers/gpu/drm/i915/gem/selftests/mock_context.c
49
e = default_engines(ctx, null_sseu);
drivers/gpu/drm/i915/gem/selftests/mock_context.c
50
if (IS_ERR(e))
drivers/gpu/drm/i915/gem/selftests/mock_context.c
52
RCU_INIT_POINTER(ctx->engines, e);
drivers/gpu/drm/i915/gt/intel_engine.h
35
#define ENGINE_TRACE(e, fmt, ...) do { \
drivers/gpu/drm/i915/gt/intel_engine.h
36
const struct intel_engine_cs *e__ __maybe_unused = (e); \
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
4529
struct intel_engine_cs *e;
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
4532
for_each_engine_masked(e, engine->gt, mask, tmp)
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
4533
e->serial++;
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
5189
struct intel_engine_cs *e;
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
5193
for_each_engine_masked(e, ce->engine->gt, virtual_mask, tmp) {
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
5194
bool match = intel_guc_capture_is_matching_engine(gt, ce, e);
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
5197
intel_engine_set_hung_context(e, ce);
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
5198
engine_mask |= e->mask;
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
5200
e);
drivers/gpu/drm/i915/gvt/cmd_parser.c
2727
static void add_cmd_entry(struct intel_gvt *gvt, struct cmd_entry *e)
drivers/gpu/drm/i915/gvt/cmd_parser.c
2729
hash_add(gvt->cmd_table, &e->hlist, e->info->opcode);
drivers/gpu/drm/i915/gvt/cmd_parser.c
3224
struct cmd_entry *e;
drivers/gpu/drm/i915/gvt/cmd_parser.c
3229
e = kzalloc_obj(*e);
drivers/gpu/drm/i915/gvt/cmd_parser.c
3230
if (!e)
drivers/gpu/drm/i915/gvt/cmd_parser.c
3233
e->info = &cmd_info[i];
drivers/gpu/drm/i915/gvt/cmd_parser.c
3237
INIT_HLIST_NODE(&e->hlist);
drivers/gpu/drm/i915/gvt/cmd_parser.c
3238
add_cmd_entry(gvt, e);
drivers/gpu/drm/i915/gvt/cmd_parser.c
3240
e->info->name, e->info->opcode, e->info->flag,
drivers/gpu/drm/i915/gvt/cmd_parser.c
3241
e->info->devices, e->info->rings);
drivers/gpu/drm/i915/gvt/cmd_parser.c
3250
struct cmd_entry *e;
drivers/gpu/drm/i915/gvt/cmd_parser.c
3253
hash_for_each_safe(gvt->cmd_table, i, tmp, e, hlist)
drivers/gpu/drm/i915/gvt/cmd_parser.c
3254
kfree(e);
drivers/gpu/drm/i915/gvt/cmd_parser.c
675
struct cmd_entry *e;
drivers/gpu/drm/i915/gvt/cmd_parser.c
677
hash_for_each_possible(gvt->cmd_table, e, hlist, opcode) {
drivers/gpu/drm/i915/gvt/cmd_parser.c
678
if (opcode == e->info->opcode &&
drivers/gpu/drm/i915/gvt/cmd_parser.c
679
e->info->rings & engine->mask)
drivers/gpu/drm/i915/gvt/cmd_parser.c
680
return e->info;
drivers/gpu/drm/i915/gvt/execlist.c
42
#define execlist_ring_mmio(e, offset) ((e)->mmio_base + (offset))
drivers/gpu/drm/i915/gvt/gtt.c
1980
struct intel_gvt_gtt_entry *e, unsigned long index, bool guest)
drivers/gpu/drm/i915/gvt/gtt.c
1986
s = intel_vgpu_find_spt_by_mfn(vgpu, ops->get_pfn(e));
drivers/gpu/drm/i915/gvt/gtt.c
1991
ppgtt_get_shadow_entry(s, e, index);
drivers/gpu/drm/i915/gvt/gtt.c
1993
ppgtt_get_guest_entry(s, e, index);
drivers/gpu/drm/i915/gvt/gtt.c
2016
struct intel_gvt_gtt_entry e;
drivers/gpu/drm/i915/gvt/gtt.c
2027
ggtt_get_guest_entry(mm, &e,
drivers/gpu/drm/i915/gvt/gtt.c
2030
gpa = (pte_ops->get_pfn(&e) << I915_GTT_PAGE_SHIFT)
drivers/gpu/drm/i915/gvt/gtt.c
2037
ppgtt_get_shadow_root_entry(mm, &e, 0);
drivers/gpu/drm/i915/gvt/gtt.c
2046
ppgtt_get_shadow_root_entry(mm, &e,
drivers/gpu/drm/i915/gvt/gtt.c
2059
ret = ppgtt_get_next_level_entry(mm, &e, gma_index[i],
drivers/gpu/drm/i915/gvt/gtt.c
2064
if (!pte_ops->test_present(&e)) {
drivers/gpu/drm/i915/gvt/gtt.c
2070
gpa = (pte_ops->get_pfn(&e) << I915_GTT_PAGE_SHIFT) +
drivers/gpu/drm/i915/gvt/gtt.c
2089
struct intel_gvt_gtt_entry e;
drivers/gpu/drm/i915/gvt/gtt.c
2102
ggtt_get_guest_entry(ggtt_mm, &e, index);
drivers/gpu/drm/i915/gvt/gtt.c
2103
memcpy(p_data, (void *)&e.val64 + (off & (info->gtt_entry_size - 1)),
drivers/gpu/drm/i915/gvt/gtt.c
2154
struct intel_gvt_gtt_entry e = {.val64 = 0, .type = GTT_TYPE_GGTT_PTE};
drivers/gpu/drm/i915/gvt/gtt.c
2170
e.type = GTT_TYPE_GGTT_PTE;
drivers/gpu/drm/i915/gvt/gtt.c
2171
memcpy((void *)&e.val64 + (off & (info->gtt_entry_size - 1)), p_data,
drivers/gpu/drm/i915/gvt/gtt.c
2190
memcpy((void *)&e.val64 + last_off,
drivers/gpu/drm/i915/gvt/gtt.c
2201
pos->data = e.val64;
drivers/gpu/drm/i915/gvt/gtt.c
2202
ggtt_set_guest_entry(ggtt_mm, &e, g_gtt_index);
drivers/gpu/drm/i915/gvt/gtt.c
2213
partial_pte->data = e.val64;
drivers/gpu/drm/i915/gvt/gtt.c
2220
if (!partial_update && (ops->test_present(&e))) {
drivers/gpu/drm/i915/gvt/gtt.c
2221
gfn = ops->get_pfn(&e);
drivers/gpu/drm/i915/gvt/gtt.c
2222
m.val64 = e.val64;
drivers/gpu/drm/i915/gvt/gtt.c
2223
m.type = e.type;
drivers/gpu/drm/i915/gvt/gtt.c
2241
ggtt_set_guest_entry(ggtt_mm, &e, g_gtt_index);
drivers/gpu/drm/i915/gvt/gtt.c
2243
ggtt_get_host_entry(ggtt_mm, &e, g_gtt_index);
drivers/gpu/drm/i915/gvt/gtt.c
2244
ggtt_invalidate_pte(vgpu, &e);
drivers/gpu/drm/i915/gvt/gtt.c
241
struct intel_gvt_gtt_entry *e,
drivers/gpu/drm/i915/gvt/gtt.c
254
&e->val64, 8);
drivers/gpu/drm/i915/gvt/gtt.c
258
e->val64 = read_pte64(vgpu->gvt->gt->ggtt, index);
drivers/gpu/drm/i915/gvt/gtt.c
260
e->val64 = *((u64 *)pt + index);
drivers/gpu/drm/i915/gvt/gtt.c
266
struct intel_gvt_gtt_entry *e,
drivers/gpu/drm/i915/gvt/gtt.c
279
&e->val64, 8);
drivers/gpu/drm/i915/gvt/gtt.c
283
write_pte64(vgpu->gvt->gt->ggtt, index, e->val64);
drivers/gpu/drm/i915/gvt/gtt.c
285
*((u64 *)pt + index) = e->val64;
drivers/gpu/drm/i915/gvt/gtt.c
302
static unsigned long gen8_gtt_get_pfn(struct intel_gvt_gtt_entry *e)
drivers/gpu/drm/i915/gvt/gtt.c
306
if (e->type == GTT_TYPE_PPGTT_PTE_1G_ENTRY)
drivers/gpu/drm/i915/gvt/gtt.c
307
pfn = (e->val64 & ADDR_1G_MASK) >> PAGE_SHIFT;
drivers/gpu/drm/i915/gvt/gtt.c
308
else if (e->type == GTT_TYPE_PPGTT_PTE_2M_ENTRY)
drivers/gpu/drm/i915/gvt/gtt.c
309
pfn = (e->val64 & ADDR_2M_MASK) >> PAGE_SHIFT;
drivers/gpu/drm/i915/gvt/gtt.c
310
else if (e->type == GTT_TYPE_PPGTT_PTE_64K_ENTRY)
drivers/gpu/drm/i915/gvt/gtt.c
311
pfn = (e->val64 & ADDR_64K_MASK) >> PAGE_SHIFT;
drivers/gpu/drm/i915/gvt/gtt.c
313
pfn = (e->val64 & ADDR_4K_MASK) >> PAGE_SHIFT;
drivers/gpu/drm/i915/gvt/gtt.c
317
static void gen8_gtt_set_pfn(struct intel_gvt_gtt_entry *e, unsigned long pfn)
drivers/gpu/drm/i915/gvt/gtt.c
319
if (e->type == GTT_TYPE_PPGTT_PTE_1G_ENTRY) {
drivers/gpu/drm/i915/gvt/gtt.c
320
e->val64 &= ~ADDR_1G_MASK;
drivers/gpu/drm/i915/gvt/gtt.c
322
} else if (e->type == GTT_TYPE_PPGTT_PTE_2M_ENTRY) {
drivers/gpu/drm/i915/gvt/gtt.c
323
e->val64 &= ~ADDR_2M_MASK;
drivers/gpu/drm/i915/gvt/gtt.c
325
} else if (e->type == GTT_TYPE_PPGTT_PTE_64K_ENTRY) {
drivers/gpu/drm/i915/gvt/gtt.c
326
e->val64 &= ~ADDR_64K_MASK;
drivers/gpu/drm/i915/gvt/gtt.c
329
e->val64 &= ~ADDR_4K_MASK;
drivers/gpu/drm/i915/gvt/gtt.c
333
e->val64 |= (pfn << PAGE_SHIFT);
drivers/gpu/drm/i915/gvt/gtt.c
336
static bool gen8_gtt_test_pse(struct intel_gvt_gtt_entry *e)
drivers/gpu/drm/i915/gvt/gtt.c
338
return !!(e->val64 & _PAGE_PSE);
drivers/gpu/drm/i915/gvt/gtt.c
341
static void gen8_gtt_clear_pse(struct intel_gvt_gtt_entry *e)
drivers/gpu/drm/i915/gvt/gtt.c
343
if (gen8_gtt_test_pse(e)) {
drivers/gpu/drm/i915/gvt/gtt.c
344
switch (e->type) {
drivers/gpu/drm/i915/gvt/gtt.c
346
e->val64 &= ~_PAGE_PSE;
drivers/gpu/drm/i915/gvt/gtt.c
347
e->type = GTT_TYPE_PPGTT_PDE_ENTRY;
drivers/gpu/drm/i915/gvt/gtt.c
350
e->type = GTT_TYPE_PPGTT_PDP_ENTRY;
drivers/gpu/drm/i915/gvt/gtt.c
351
e->val64 &= ~_PAGE_PSE;
drivers/gpu/drm/i915/gvt/gtt.c
359
static bool gen8_gtt_test_ips(struct intel_gvt_gtt_entry *e)
drivers/gpu/drm/i915/gvt/gtt.c
361
if (GEM_WARN_ON(e->type != GTT_TYPE_PPGTT_PDE_ENTRY))
drivers/gpu/drm/i915/gvt/gtt.c
364
return !!(e->val64 & GEN8_PDE_IPS_64K);
drivers/gpu/drm/i915/gvt/gtt.c
367
static void gen8_gtt_clear_ips(struct intel_gvt_gtt_entry *e)
drivers/gpu/drm/i915/gvt/gtt.c
369
if (GEM_WARN_ON(e->type != GTT_TYPE_PPGTT_PDE_ENTRY))
drivers/gpu/drm/i915/gvt/gtt.c
372
e->val64 &= ~GEN8_PDE_IPS_64K;
drivers/gpu/drm/i915/gvt/gtt.c
375
static bool gen8_gtt_test_present(struct intel_gvt_gtt_entry *e)
drivers/gpu/drm/i915/gvt/gtt.c
382
if (e->type == GTT_TYPE_PPGTT_ROOT_L3_ENTRY
drivers/gpu/drm/i915/gvt/gtt.c
383
|| e->type == GTT_TYPE_PPGTT_ROOT_L4_ENTRY)
drivers/gpu/drm/i915/gvt/gtt.c
384
return (e->val64 != 0);
drivers/gpu/drm/i915/gvt/gtt.c
386
return (e->val64 & GEN8_PAGE_PRESENT);
drivers/gpu/drm/i915/gvt/gtt.c
389
static void gtt_entry_clear_present(struct intel_gvt_gtt_entry *e)
drivers/gpu/drm/i915/gvt/gtt.c
391
e->val64 &= ~GEN8_PAGE_PRESENT;
drivers/gpu/drm/i915/gvt/gtt.c
394
static void gtt_entry_set_present(struct intel_gvt_gtt_entry *e)
drivers/gpu/drm/i915/gvt/gtt.c
396
e->val64 |= GEN8_PAGE_PRESENT;
drivers/gpu/drm/i915/gvt/gtt.c
399
static bool gen8_gtt_test_64k_splited(struct intel_gvt_gtt_entry *e)
drivers/gpu/drm/i915/gvt/gtt.c
401
return !!(e->val64 & GTT_SPTE_FLAG_64K_SPLITED);
drivers/gpu/drm/i915/gvt/gtt.c
404
static void gen8_gtt_set_64k_splited(struct intel_gvt_gtt_entry *e)
drivers/gpu/drm/i915/gvt/gtt.c
406
e->val64 |= GTT_SPTE_FLAG_64K_SPLITED;
drivers/gpu/drm/i915/gvt/gtt.c
409
static void gen8_gtt_clear_64k_splited(struct intel_gvt_gtt_entry *e)
drivers/gpu/drm/i915/gvt/gtt.c
411
e->val64 &= ~GTT_SPTE_FLAG_64K_SPLITED;
drivers/gpu/drm/i915/gvt/gtt.c
591
struct intel_gvt_gtt_entry *e, unsigned long index,
drivers/gpu/drm/i915/gvt/gtt.c
598
e->type = get_entry_type(type);
drivers/gpu/drm/i915/gvt/gtt.c
600
if (WARN(!gtt_type_is_entry(e->type), "invalid entry type\n"))
drivers/gpu/drm/i915/gvt/gtt.c
603
ret = ops->get_entry(page_table, e, index, guest,
drivers/gpu/drm/i915/gvt/gtt.c
609
update_entry_type_for_real(ops, e, guest ?
drivers/gpu/drm/i915/gvt/gtt.c
613
type, e->type, index, e->val64);
drivers/gpu/drm/i915/gvt/gtt.c
620
struct intel_gvt_gtt_entry *e, unsigned long index,
drivers/gpu/drm/i915/gvt/gtt.c
626
if (WARN(!gtt_type_is_entry(e->type), "invalid entry type\n"))
drivers/gpu/drm/i915/gvt/gtt.c
630
type, e->type, index, e->val64);
drivers/gpu/drm/i915/gvt/gtt.c
632
return ops->set_entry(page_table, e, index, guest,
drivers/gpu/drm/i915/gvt/gtt.c
637
#define ppgtt_get_guest_entry(spt, e, index) \
drivers/gpu/drm/i915/gvt/gtt.c
639
spt->guest_page.type, e, index, true)
drivers/gpu/drm/i915/gvt/gtt.c
641
#define ppgtt_set_guest_entry(spt, e, index) \
drivers/gpu/drm/i915/gvt/gtt.c
643
spt->guest_page.type, e, index, true)
drivers/gpu/drm/i915/gvt/gtt.c
645
#define ppgtt_get_shadow_entry(spt, e, index) \
drivers/gpu/drm/i915/gvt/gtt.c
647
spt->shadow_page.type, e, index, false)
drivers/gpu/drm/i915/gvt/gtt.c
649
#define ppgtt_set_shadow_entry(spt, e, index) \
drivers/gpu/drm/i915/gvt/gtt.c
651
spt->shadow_page.type, e, index, false)
drivers/gpu/drm/i915/gvt/gtt.c
848
#define for_each_present_guest_entry(spt, e, i) \
drivers/gpu/drm/i915/gvt/gtt.c
851
if (!ppgtt_get_guest_entry(spt, e, i) && \
drivers/gpu/drm/i915/gvt/gtt.c
852
spt->vgpu->gvt->gtt.pte_ops->test_present(e))
drivers/gpu/drm/i915/gvt/gtt.c
854
#define for_each_present_shadow_entry(spt, e, i) \
drivers/gpu/drm/i915/gvt/gtt.c
857
if (!ppgtt_get_shadow_entry(spt, e, i) && \
drivers/gpu/drm/i915/gvt/gtt.c
858
spt->vgpu->gvt->gtt.pte_ops->test_present(e))
drivers/gpu/drm/i915/gvt/gtt.c
860
#define for_each_shadow_entry(spt, e, i) \
drivers/gpu/drm/i915/gvt/gtt.c
863
if (!ppgtt_get_shadow_entry(spt, e, i))
drivers/gpu/drm/i915/gvt/gtt.c
884
struct intel_gvt_gtt_entry *e)
drivers/gpu/drm/i915/gvt/gtt.c
891
GEM_BUG_ON(!gtt_type_is_pt(get_next_pt_type(e->type)));
drivers/gpu/drm/i915/gvt/gtt.c
893
if (e->type != GTT_TYPE_PPGTT_ROOT_L3_ENTRY
drivers/gpu/drm/i915/gvt/gtt.c
894
&& e->type != GTT_TYPE_PPGTT_ROOT_L4_ENTRY) {
drivers/gpu/drm/i915/gvt/gtt.c
895
cur_pt_type = get_next_pt_type(e->type);
drivers/gpu/drm/i915/gvt/gtt.c
907
if (ops->get_pfn(e) ==
drivers/gpu/drm/i915/gvt/gtt.c
91
#define gtt_init_entry(e, t, p, v) do { \
drivers/gpu/drm/i915/gvt/gtt.c
911
s = intel_vgpu_find_spt_by_mfn(vgpu, ops->get_pfn(e));
drivers/gpu/drm/i915/gvt/gtt.c
914
ops->get_pfn(e));
drivers/gpu/drm/i915/gvt/gtt.c
92
(e)->type = t; \
drivers/gpu/drm/i915/gvt/gtt.c
93
(e)->pdev = p; \
drivers/gpu/drm/i915/gvt/gtt.c
94
memcpy(&(e)->val64, &v, sizeof(v)); \
drivers/gpu/drm/i915/gvt/gtt.c
941
struct intel_gvt_gtt_entry e;
drivers/gpu/drm/i915/gvt/gtt.c
951
for_each_present_shadow_entry(spt, &e, index) {
drivers/gpu/drm/i915/gvt/gtt.c
952
switch (e.type) {
drivers/gpu/drm/i915/gvt/gtt.c
955
ppgtt_invalidate_pte(spt, &e);
drivers/gpu/drm/i915/gvt/gtt.c
972
spt->vgpu, &e);
drivers/gpu/drm/i915/gvt/gtt.c
987
spt, e.val64, e.type);
drivers/gpu/drm/i915/gvt/gtt.h
59
struct intel_gvt_gtt_entry *e,
drivers/gpu/drm/i915/gvt/gtt.h
65
struct intel_gvt_gtt_entry *e,
drivers/gpu/drm/i915/gvt/gtt.h
70
bool (*test_present)(struct intel_gvt_gtt_entry *e);
drivers/gpu/drm/i915/gvt/gtt.h
71
void (*clear_present)(struct intel_gvt_gtt_entry *e);
drivers/gpu/drm/i915/gvt/gtt.h
72
void (*set_present)(struct intel_gvt_gtt_entry *e);
drivers/gpu/drm/i915/gvt/gtt.h
73
bool (*test_pse)(struct intel_gvt_gtt_entry *e);
drivers/gpu/drm/i915/gvt/gtt.h
74
void (*clear_pse)(struct intel_gvt_gtt_entry *e);
drivers/gpu/drm/i915/gvt/gtt.h
75
bool (*test_ips)(struct intel_gvt_gtt_entry *e);
drivers/gpu/drm/i915/gvt/gtt.h
76
void (*clear_ips)(struct intel_gvt_gtt_entry *e);
drivers/gpu/drm/i915/gvt/gtt.h
77
bool (*test_64k_splited)(struct intel_gvt_gtt_entry *e);
drivers/gpu/drm/i915/gvt/gtt.h
78
void (*clear_64k_splited)(struct intel_gvt_gtt_entry *e);
drivers/gpu/drm/i915/gvt/gtt.h
79
void (*set_64k_splited)(struct intel_gvt_gtt_entry *e);
drivers/gpu/drm/i915/gvt/gtt.h
80
void (*set_pfn)(struct intel_gvt_gtt_entry *e, unsigned long pfn);
drivers/gpu/drm/i915/gvt/gtt.h
81
unsigned long (*get_pfn)(struct intel_gvt_gtt_entry *e);
drivers/gpu/drm/i915/gvt/handlers.c
121
struct intel_gvt_mmio_info *e;
drivers/gpu/drm/i915/gvt/handlers.c
123
hash_for_each_possible(gvt->mmio.mmio_info_table, e, node, offset) {
drivers/gpu/drm/i915/gvt/handlers.c
124
if (e->offset == offset)
drivers/gpu/drm/i915/gvt/handlers.c
125
return e;
drivers/gpu/drm/i915/gvt/handlers.c
2856
struct intel_gvt_mmio_info *e;
drivers/gpu/drm/i915/gvt/handlers.c
2859
hash_for_each_safe(gvt->mmio.mmio_info_table, i, tmp, e, node)
drivers/gpu/drm/i915/gvt/handlers.c
2860
kfree(e);
drivers/gpu/drm/i915/gvt/handlers.c
3052
struct intel_gvt_mmio_info *e;
drivers/gpu/drm/i915/gvt/handlers.c
3055
hash_for_each(gvt->mmio.mmio_info_table, i, e, node) {
drivers/gpu/drm/i915/gvt/handlers.c
3056
ret = handler(gvt, e->offset, data);
drivers/gpu/drm/i915/gvt/interrupt.c
558
#define SET_BIT_INFO(s, b, e, i) \
drivers/gpu/drm/i915/gvt/interrupt.c
560
s->events[e].bit = b; \
drivers/gpu/drm/i915/gvt/interrupt.c
561
s->events[e].info = s->info[i]; \
drivers/gpu/drm/i915/gvt/interrupt.c
562
s->info[i]->bit_to_event[b] = e;\
drivers/gpu/drm/i915/gvt/interrupt.c
68
#define get_event_virt_handler(irq, e) (irq->events[e].v_handler)
drivers/gpu/drm/i915/gvt/interrupt.c
69
#define get_irq_info(irq, e) (irq->events[e].info)
drivers/gpu/drm/i915/gvt/scheduler.h
134
#define workload_q_head(vgpu, e) \
drivers/gpu/drm/i915/gvt/scheduler.h
135
(&(vgpu)->submission.workload_q_head[(e)->id])
drivers/gpu/drm/i915/i915_gpu_error.c
100
e->cur->offset = 0;
drivers/gpu/drm/i915/i915_gpu_error.c
101
e->cur->length = 0;
drivers/gpu/drm/i915/i915_gpu_error.c
102
e->cur->page_link =
drivers/gpu/drm/i915/i915_gpu_error.c
105
e->sgl = sgl;
drivers/gpu/drm/i915/i915_gpu_error.c
108
e->cur = sgl;
drivers/gpu/drm/i915/i915_gpu_error.c
109
e->end = sgl + SG_MAX_SINGLE_ALLOC - 1;
drivers/gpu/drm/i915/i915_gpu_error.c
112
e->size = ALIGN(len + 1, SZ_64K);
drivers/gpu/drm/i915/i915_gpu_error.c
113
e->buf = kmalloc(e->size, ALLOW_FAIL);
drivers/gpu/drm/i915/i915_gpu_error.c
114
if (!e->buf) {
drivers/gpu/drm/i915/i915_gpu_error.c
115
e->size = PAGE_ALIGN(len + 1);
drivers/gpu/drm/i915/i915_gpu_error.c
116
e->buf = kmalloc(e->size, GFP_KERNEL);
drivers/gpu/drm/i915/i915_gpu_error.c
118
if (!e->buf) {
drivers/gpu/drm/i915/i915_gpu_error.c
119
e->err = -ENOMEM;
drivers/gpu/drm/i915/i915_gpu_error.c
127
static void i915_error_vprintf(struct drm_i915_error_state_buf *e,
drivers/gpu/drm/i915/i915_gpu_error.c
133
if (e->err)
drivers/gpu/drm/i915/i915_gpu_error.c
140
e->err = len;
drivers/gpu/drm/i915/i915_gpu_error.c
144
if (!__i915_error_grow(e, len))
drivers/gpu/drm/i915/i915_gpu_error.c
1450
static bool record_context(struct i915_gem_context_coredump *e,
drivers/gpu/drm/i915/i915_gpu_error.c
1468
strscpy(e->comm, task->comm);
drivers/gpu/drm/i915/i915_gpu_error.c
1469
e->pid = task->pid;
drivers/gpu/drm/i915/i915_gpu_error.c
147
GEM_BUG_ON(e->bytes >= e->size);
drivers/gpu/drm/i915/i915_gpu_error.c
1473
e->sched_attr = ctx->sched;
drivers/gpu/drm/i915/i915_gpu_error.c
1474
e->guilty = atomic_read(&ctx->guilty_count);
drivers/gpu/drm/i915/i915_gpu_error.c
1475
e->active = atomic_read(&ctx->active_count);
drivers/gpu/drm/i915/i915_gpu_error.c
1476
e->hwsp_seqno = (ce->timeline && ce->timeline->hwsp_seqno) ?
drivers/gpu/drm/i915/i915_gpu_error.c
1479
e->total_runtime = intel_context_get_total_runtime_ns(ce);
drivers/gpu/drm/i915/i915_gpu_error.c
148
len = vscnprintf(e->buf + e->bytes, e->size - e->bytes, fmt, args);
drivers/gpu/drm/i915/i915_gpu_error.c
1480
e->avg_runtime = intel_context_get_avg_runtime_ns(ce);
drivers/gpu/drm/i915/i915_gpu_error.c
150
e->err = len;
drivers/gpu/drm/i915/i915_gpu_error.c
153
e->bytes += len;
drivers/gpu/drm/i915/i915_gpu_error.c
156
static void i915_error_puts(struct drm_i915_error_state_buf *e, const char *str)
drivers/gpu/drm/i915/i915_gpu_error.c
160
if (e->err || !str)
drivers/gpu/drm/i915/i915_gpu_error.c
164
if (!__i915_error_grow(e, len))
drivers/gpu/drm/i915/i915_gpu_error.c
167
GEM_BUG_ON(e->bytes + len > e->size);
drivers/gpu/drm/i915/i915_gpu_error.c
168
memcpy(e->buf + e->bytes, str, len);
drivers/gpu/drm/i915/i915_gpu_error.c
169
e->bytes += len;
drivers/gpu/drm/i915/i915_gpu_error.c
172
#define err_printf(e, ...) i915_error_printf(e, __VA_ARGS__)
drivers/gpu/drm/i915/i915_gpu_error.c
173
#define err_puts(e, s) i915_error_puts(e, s)
drivers/gpu/drm/i915/i915_gpu_error.c
181
i915_error_printer(struct drm_i915_error_state_buf *e)
drivers/gpu/drm/i915/i915_gpu_error.c
185
.arg = e,
drivers/gpu/drm/i915/i915_gpu_error.c
604
void i915_error_printf(struct drm_i915_error_state_buf *e, const char *f, ...)
drivers/gpu/drm/i915/i915_gpu_error.c
609
i915_error_vprintf(e, f, args);
drivers/gpu/drm/i915/i915_gpu_error.c
75
static bool __i915_error_grow(struct drm_i915_error_state_buf *e, size_t len)
drivers/gpu/drm/i915/i915_gpu_error.c
80
if (e->bytes + len + 1 <= e->size)
drivers/gpu/drm/i915/i915_gpu_error.c
83
if (e->bytes) {
drivers/gpu/drm/i915/i915_gpu_error.c
84
__sg_set_buf(e->cur++, e->buf, e->bytes, e->iter);
drivers/gpu/drm/i915/i915_gpu_error.c
85
e->iter += e->bytes;
drivers/gpu/drm/i915/i915_gpu_error.c
86
e->buf = NULL;
drivers/gpu/drm/i915/i915_gpu_error.c
87
e->bytes = 0;
drivers/gpu/drm/i915/i915_gpu_error.c
90
if (e->cur == e->end) {
drivers/gpu/drm/i915/i915_gpu_error.c
95
e->err = -ENOMEM;
drivers/gpu/drm/i915/i915_gpu_error.c
99
if (e->cur) {
drivers/gpu/drm/i915/i915_gpu_error.h
278
void i915_error_printf(struct drm_i915_error_state_buf *e, const char *f, ...);
drivers/gpu/drm/i915/i915_gpu_error.h
338
i915_error_printf(struct drm_i915_error_state_buf *e, const char *f, ...)
drivers/gpu/drm/i915/intel_uncore.c
1208
#define GEN_FW_RANGE(s, e, d) \
drivers/gpu/drm/i915/intel_uncore.c
1209
{ .start = (s), .end = (e), .domains = (d) }
drivers/gpu/drm/imagination/pvr_rogue_fwif_sf.h
58
#define ROGUE_FW_LOG_CREATESFID(a, b, e) ((u32)(a) | ((u32)(b) << 12) | ((u32)(e) << 16) | \
drivers/gpu/drm/imx/ipuv3/dw_hdmi-imx.c
41
static inline struct imx_hdmi *enc_to_imx_hdmi(struct drm_encoder *e)
drivers/gpu/drm/imx/ipuv3/dw_hdmi-imx.c
43
return container_of(e, struct imx_hdmi_encoder, encoder)->hdmi;
drivers/gpu/drm/imx/ipuv3/imx-ldb.c
71
static inline struct imx_ldb_channel *enc_to_imx_ldb_ch(struct drm_encoder *e)
drivers/gpu/drm/imx/ipuv3/imx-ldb.c
73
return container_of(e, struct imx_ldb_encoder, encoder)->channel;
drivers/gpu/drm/imx/ipuv3/imx-tve.c
129
static inline struct imx_tve *enc_to_tve(struct drm_encoder *e)
drivers/gpu/drm/imx/ipuv3/imx-tve.c
131
return container_of(e, struct imx_tve_encoder, encoder)->tve;
drivers/gpu/drm/msm/adreno/adreno_gpu.c
971
struct msm_gem_vm_log_entry *e = &state->vm_logs[i];
drivers/gpu/drm/msm/adreno/adreno_gpu.c
973
e->op, e->queue_id, e->iova,
drivers/gpu/drm/msm/adreno/adreno_gpu.c
974
e->iova + e->range);
drivers/gpu/drm/msm/disp/dpu1/dpu_encoder.c
39
#define DPU_DEBUG_ENC(e, fmt, ...) DRM_DEBUG_ATOMIC("enc%d " fmt,\
drivers/gpu/drm/msm/disp/dpu1/dpu_encoder.c
40
(e) ? (e)->base.base.id : -1, ##__VA_ARGS__)
drivers/gpu/drm/msm/disp/dpu1/dpu_encoder.c
42
#define DPU_ERROR_ENC(e, fmt, ...) DPU_ERROR("enc%d " fmt,\
drivers/gpu/drm/msm/disp/dpu1/dpu_encoder.c
43
(e) ? (e)->base.base.id : -1, ##__VA_ARGS__)
drivers/gpu/drm/msm/disp/dpu1/dpu_encoder.c
45
#define DPU_ERROR_ENC_RATELIMITED(e, fmt, ...) DPU_ERROR_RATELIMITED("enc%d " fmt,\
drivers/gpu/drm/msm/disp/dpu1/dpu_encoder.c
46
(e) ? (e)->base.base.id : -1, ##__VA_ARGS__)
drivers/gpu/drm/msm/disp/dpu1/dpu_encoder_phys_cmd.c
19
#define DPU_DEBUG_CMDENC(e, fmt, ...) DPU_DEBUG("enc%d intf%d " fmt, \
drivers/gpu/drm/msm/disp/dpu1/dpu_encoder_phys_cmd.c
20
(e) && (e)->base.parent ? \
drivers/gpu/drm/msm/disp/dpu1/dpu_encoder_phys_cmd.c
21
(e)->base.parent->base.id : -1, \
drivers/gpu/drm/msm/disp/dpu1/dpu_encoder_phys_cmd.c
22
(e) ? (e)->base.hw_intf->idx - INTF_0 : -1, ##__VA_ARGS__)
drivers/gpu/drm/msm/disp/dpu1/dpu_encoder_phys_cmd.c
24
#define DPU_ERROR_CMDENC(e, fmt, ...) DPU_ERROR("enc%d intf%d " fmt, \
drivers/gpu/drm/msm/disp/dpu1/dpu_encoder_phys_cmd.c
25
(e) && (e)->base.parent ? \
drivers/gpu/drm/msm/disp/dpu1/dpu_encoder_phys_cmd.c
26
(e)->base.parent->base.id : -1, \
drivers/gpu/drm/msm/disp/dpu1/dpu_encoder_phys_cmd.c
27
(e) ? (e)->base.hw_intf->idx - INTF_0 : -1, ##__VA_ARGS__)
drivers/gpu/drm/msm/disp/dpu1/dpu_encoder_phys_vid.c
17
#define DPU_DEBUG_VIDENC(e, fmt, ...) DPU_DEBUG("enc%d intf%d " fmt, \
drivers/gpu/drm/msm/disp/dpu1/dpu_encoder_phys_vid.c
18
(e) && (e)->parent ? \
drivers/gpu/drm/msm/disp/dpu1/dpu_encoder_phys_vid.c
19
(e)->parent->base.id : -1, \
drivers/gpu/drm/msm/disp/dpu1/dpu_encoder_phys_vid.c
20
(e) && (e)->hw_intf ? \
drivers/gpu/drm/msm/disp/dpu1/dpu_encoder_phys_vid.c
21
(e)->hw_intf->idx - INTF_0 : -1, ##__VA_ARGS__)
drivers/gpu/drm/msm/disp/dpu1/dpu_encoder_phys_vid.c
23
#define DPU_ERROR_VIDENC(e, fmt, ...) DPU_ERROR("enc%d intf%d " fmt, \
drivers/gpu/drm/msm/disp/dpu1/dpu_encoder_phys_vid.c
24
(e) && (e)->parent ? \
drivers/gpu/drm/msm/disp/dpu1/dpu_encoder_phys_vid.c
25
(e)->parent->base.id : -1, \
drivers/gpu/drm/msm/disp/dpu1/dpu_encoder_phys_vid.c
26
(e) && (e)->hw_intf ? \
drivers/gpu/drm/msm/disp/dpu1/dpu_encoder_phys_vid.c
27
(e)->hw_intf->idx - INTF_0 : -1, ##__VA_ARGS__)
drivers/gpu/drm/msm/msm_gem_vma.c
215
struct msm_gem_vm_log_entry *e = &vm->log[idx];
drivers/gpu/drm/msm/msm_gem_vma.c
217
e->op, e->queue_id, e->iova,
drivers/gpu/drm/msm/msm_gem_vma.c
218
e->iova + e->range);
drivers/gpu/drm/nouveau/include/nvhw/drf.h
126
#define DRF_RD_X(e,p,o,dr) e((p), (o), dr)
drivers/gpu/drm/nouveau/include/nvhw/drf.h
127
#define DRF_RD_N(X,e,p,o,d,r ) DRF_RD_X(e, (p), (o), d##_##r)
drivers/gpu/drm/nouveau/include/nvhw/drf.h
128
#define DRF_RD_I(X,e,p,o,d,r,i) DRF_RD_X(e, (p), (o), d##_##r(i))
drivers/gpu/drm/nouveau/include/nvhw/drf.h
133
#define DRF_WR_X(e,p,o,dr,v) e((p), (o), dr, (v))
drivers/gpu/drm/nouveau/include/nvhw/drf.h
134
#define DRF_WR_N(X,e,p,o,d,r, v) DRF_WR_X(e, (p), (o), d##_##r , (v))
drivers/gpu/drm/nouveau/include/nvhw/drf.h
135
#define DRF_WR_I(X,e,p,o,d,r,i,v) DRF_WR_X(e, (p), (o), d##_##r(i), (v))
drivers/gpu/drm/nouveau/include/nvhw/drf.h
151
#define DRF_RV_X(e,p,o,dr,drf) NVVAL_GET_X(DRF_RD_X(e, (p), (o), dr), drf)
drivers/gpu/drm/nouveau/include/nvhw/drf.h
152
#define DRF_RV_N(X,e,p,o,d,r, f) DRF_RV_X(e, (p), (o), d##_##r , d##_##r##_##f)
drivers/gpu/drm/nouveau/include/nvhw/drf.h
153
#define DRF_RV_I(X,e,p,o,d,r,i,f) DRF_RV_X(e, (p), (o), d##_##r(i), d##_##r##_##f)
drivers/gpu/drm/nouveau/include/nvhw/drf.h
158
#define DRF_WV_N(X,e,p,o,d,r, f,v) \
drivers/gpu/drm/nouveau/include/nvhw/drf.h
159
DRF_WR_X(e, (p), (o), d##_##r , NVVAL_X(d##_##r##_##f, (v)))
drivers/gpu/drm/nouveau/include/nvhw/drf.h
160
#define DRF_WV_I(X,e,p,o,d,r,i,f,v) \
drivers/gpu/drm/nouveau/include/nvhw/drf.h
161
DRF_WR_X(e, (p), (o), d##_##r(i), NVVAL_X(d##_##r##_##f, (v)))
drivers/gpu/drm/nouveau/include/nvhw/drf.h
166
#define DRF_WD_N(X,e,p,o,d,r, f,v) \
drivers/gpu/drm/nouveau/include/nvhw/drf.h
167
DRF_WR_X(e, (p), (o), d##_##r , NVVAL_X(d##_##r##_##f, d##_##r##_##f##_##v))
drivers/gpu/drm/nouveau/include/nvhw/drf.h
168
#define DRF_WD_I(X,e,p,o,d,r,i,f,v) \
drivers/gpu/drm/nouveau/include/nvhw/drf.h
169
DRF_WR_X(e, (p), (o), d##_##r(i), NVVAL_X(d##_##r##_##f, d##_##r##_##f##_##v))
drivers/gpu/drm/nouveau/include/nvhw/drf.h
194
#define DRF_TV_N(X,e,p,o,d,r, f,cmp,v) \
drivers/gpu/drm/nouveau/include/nvhw/drf.h
195
NVVAL_TEST_X(DRF_RD_X(e, (p), (o), d##_##r ), d##_##r##_##f, cmp, (v))
drivers/gpu/drm/nouveau/include/nvhw/drf.h
196
#define DRF_TV_I(X,e,p,o,d,r,i,f,cmp,v) \
drivers/gpu/drm/nouveau/include/nvhw/drf.h
197
NVVAL_TEST_X(DRF_RD_X(e, (p), (o), d##_##r(i)), d##_##r##_##f, cmp, (v))
drivers/gpu/drm/nouveau/include/nvhw/drf.h
202
#define DRF_TD_N(X,e,p,o,d,r, f,cmp,v) \
drivers/gpu/drm/nouveau/include/nvhw/drf.h
203
NVVAL_TEST_X(DRF_RD_X(e, (p), (o), d##_##r ), d##_##r##_##f, cmp, d##_##r##_##f##_##v)
drivers/gpu/drm/nouveau/include/nvhw/drf.h
204
#define DRF_TD_I(X,e,p,o,d,r,i,f,cmp,v) \
drivers/gpu/drm/nouveau/include/nvhw/drf.h
205
NVVAL_TEST_X(DRF_RD_X(e, (p), (o), d##_##r(i)), d##_##r##_##f, cmp, d##_##r##_##f##_##v)
drivers/gpu/drm/nouveau/nvkm/core/mm.c
117
u32 s, e;
drivers/gpu/drm/nouveau/nvkm/core/mm.c
126
e = this->offset + this->length;
drivers/gpu/drm/nouveau/nvkm/core/mm.c
135
e = rounddown(e, mm->block_size);
drivers/gpu/drm/nouveau/nvkm/core/mm.c
138
e &= ~mask;
drivers/gpu/drm/nouveau/nvkm/core/mm.c
139
if (s > e || e - s < size_min)
drivers/gpu/drm/nouveau/nvkm/core/mm.c
146
this = region_head(mm, this, min(size_max, e - s));
drivers/gpu/drm/nouveau/nvkm/core/mm.c
195
u32 e = this->offset + this->length;
drivers/gpu/drm/nouveau/nvkm/core/mm.c
209
e = rounddown(e, mm->block_size);
drivers/gpu/drm/nouveau/nvkm/core/mm.c
210
c = next->offset - e;
drivers/gpu/drm/nouveau/nvkm/core/mm.c
214
a = e - s;
drivers/gpu/drm/nouveau/nvkm/core/mm.c
215
if (s > e || a < size_min)
drivers/gpu/drm/nouveau/nvkm/core/mm.c
219
s = (e - a) & ~mask;
drivers/gpu/drm/nouveau/nvkm/core/mm.c
220
c += (e - s) - a;
drivers/gpu/drm/nouveau/nvkm/engine/fifo/runl.h
40
#define ENGN_PRINT(e,l,p,f,a...) \
drivers/gpu/drm/nouveau/nvkm/engine/fifo/runl.h
41
RUNL_PRINT((e)->runl, l, p, "%02d[%8s]:"f, (e)->id, (e)->engine->subdev.name, ##a)
drivers/gpu/drm/nouveau/nvkm/engine/fifo/runl.h
42
#define ENGN_DEBUG(e,f,a...) ENGN_PRINT((e), DEBUG, info, " "f"\n", ##a)
drivers/gpu/drm/nouveau/nvkm/subdev/bios/vpstate.c
77
u8 idx, struct nvbios_vpstate_entry *e)
drivers/gpu/drm/nouveau/nvkm/subdev/bios/vpstate.c
81
if (!e || !h || idx > h->ecount)
drivers/gpu/drm/nouveau/nvkm/subdev/bios/vpstate.c
85
e->pstate = nvbios_rd08(b, offset);
drivers/gpu/drm/nouveau/nvkm/subdev/bios/vpstate.c
86
e->clock_mhz = nvbios_rd16(b, offset + 0x5);
drivers/gpu/drm/nouveau/nvkm/subdev/clk/seq.h
7
#define clk_exec(s,e) hwsq_exec(&(s)->base, (e))
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramfuc.h
165
#define ram_exec(s,e) ramfuc_exec(&(s)->base, (e))
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramseq.h
7
#define ram_exec(s,e) hwsq_exec(&(s)->base, (e))
drivers/gpu/drm/omapdrm/dss/dsi.c
692
u32 e;
drivers/gpu/drm/omapdrm/dss/dsi.c
695
e = dsi->errors;
drivers/gpu/drm/omapdrm/dss/dsi.c
698
return e;
drivers/gpu/drm/radeon/cikd.h
2003
#define SDMA_PACKET(op, sub_op, e) ((((e) & 0xFFFF) << 16) | \
drivers/gpu/drm/radeon/radeon.h
2701
#define radeon_gart_set_page(rdev, i, e) (rdev)->asic->gart.set_page((rdev), (i), (e))
drivers/gpu/drm/radeon/radeon.h
2721
#define radeon_set_backlight_level(rdev, e, l) (rdev)->asic->display.set_backlight_level((e), (l))
drivers/gpu/drm/radeon/radeon.h
2722
#define radeon_get_backlight_level(rdev, e) (rdev)->asic->display.get_backlight_level((e))
drivers/gpu/drm/radeon/radeon.h
2723
#define radeon_hdmi_enable(rdev, e, b) (rdev)->asic->display.hdmi_enable((e), (b))
drivers/gpu/drm/radeon/radeon.h
2724
#define radeon_hdmi_setmode(rdev, e, m) (rdev)->asic->display.hdmi_setmode((e), (m))
drivers/gpu/drm/radeon/radeon.h
2734
#define radeon_set_engine_clock(rdev, e) (rdev)->asic->pm.set_engine_clock((rdev), (e))
drivers/gpu/drm/radeon/radeon.h
2736
#define radeon_set_memory_clock(rdev, e) (rdev)->asic->pm.set_memory_clock((rdev), (e))
drivers/gpu/drm/radeon/radeon.h
2739
#define radeon_set_clock_gating(rdev, e) (rdev)->asic->pm.set_clock_gating((rdev), (e))
drivers/gpu/drm/radeon/radeon.h
2780
#define radeon_dpm_enable_bapm(rdev, e) rdev->asic->dpm.enable_bapm((rdev), (e))
drivers/gpu/drm/renesas/rcar-du/rcar_du_encoder.h
22
#define to_rcar_encoder(e) \
drivers/gpu/drm/renesas/rcar-du/rcar_du_encoder.h
23
container_of(e, struct rcar_du_encoder, base)
drivers/gpu/drm/renesas/rcar-du/rcar_lvds.c
219
unsigned int e;
drivers/gpu/drm/renesas/rcar-du/rcar_lvds.c
231
for (e = e_min; e < 3; ++e) {
drivers/gpu/drm/renesas/rcar-du/rcar_lvds.c
241
fout = fvco / (1 << e) / div7;
drivers/gpu/drm/renesas/rcar-du/rcar_lvds.c
249
pll->pll_e = e;
drivers/gpu/drm/renesas/rz-du/rzg2l_du_encoder.h
23
static inline struct rzg2l_du_encoder *to_rzg2l_encoder(struct drm_encoder *e)
drivers/gpu/drm/renesas/rz-du/rzg2l_du_encoder.h
25
return container_of(e, struct rzg2l_du_encoder, base);
drivers/gpu/drm/sun4i/sun4i_hdmi_enc.c
37
#define drm_encoder_to_sun4i_hdmi(e) \
drivers/gpu/drm/sun4i/sun4i_hdmi_enc.c
38
container_of_const(e, struct sun4i_hdmi, encoder)
drivers/gpu/drm/tegra/drm.h
148
static inline struct tegra_output *encoder_to_output(struct drm_encoder *e)
drivers/gpu/drm/tegra/drm.h
150
return container_of(e, struct tegra_output, encoder);
drivers/gpu/drm/tests/drm_hdmi_state_helper_test.c
45
#define encoder_to_priv(e) \
drivers/gpu/drm/tests/drm_hdmi_state_helper_test.c
46
container_of_const(e, struct drm_atomic_helper_connector_hdmi_priv, encoder)
drivers/gpu/drm/virtio/virtgpu_drv.h
153
struct virtio_gpu_fence_event *e;
drivers/gpu/drm/virtio/virtgpu_fence.c
138
if (curr->e) {
drivers/gpu/drm/virtio/virtgpu_fence.c
139
drm_send_event(vgdev->ddev, &curr->e->base);
drivers/gpu/drm/virtio/virtgpu_fence.c
140
curr->e = NULL;
drivers/gpu/drm/virtio/virtgpu_fence.c
148
if (signaled->e) {
drivers/gpu/drm/virtio/virtgpu_fence.c
149
drm_send_event(vgdev->ddev, &signaled->e->base);
drivers/gpu/drm/virtio/virtgpu_fence.c
150
signaled->e = NULL;
drivers/gpu/drm/virtio/virtgpu_submit.c
277
struct virtio_gpu_fence_event *e = NULL;
drivers/gpu/drm/virtio/virtgpu_submit.c
280
e = kzalloc_obj(*e);
drivers/gpu/drm/virtio/virtgpu_submit.c
281
if (!e)
drivers/gpu/drm/virtio/virtgpu_submit.c
284
e->event.type = VIRTGPU_EVENT_FENCE_SIGNALED;
drivers/gpu/drm/virtio/virtgpu_submit.c
285
e->event.length = sizeof(e->event);
drivers/gpu/drm/virtio/virtgpu_submit.c
287
ret = drm_event_reserve_init(dev, file, &e->base, &e->event);
drivers/gpu/drm/virtio/virtgpu_submit.c
289
kfree(e);
drivers/gpu/drm/virtio/virtgpu_submit.c
293
fence->e = e;
drivers/gpu/drm/xe/tests/xe_args_test.c
135
count = COUNT_ARGS(a, b, c, d, e);
drivers/gpu/drm/xe/tests/xe_args_test.c
138
count = COUNT_ARGS(a, b, c, d, e, f, g, h, i, j, k, l);
drivers/gpu/drm/xe/xe_exec_queue.h
133
void xe_exec_queue_last_fence_put(struct xe_exec_queue *e, struct xe_vm *vm);
drivers/gpu/drm/xe/xe_exec_queue.h
134
void xe_exec_queue_last_fence_put_unlocked(struct xe_exec_queue *e);
drivers/gpu/drm/xe/xe_exec_queue.h
135
struct dma_fence *xe_exec_queue_last_fence_get(struct xe_exec_queue *e,
drivers/gpu/drm/xe/xe_exec_queue.h
137
struct dma_fence *xe_exec_queue_last_fence_get_for_resume(struct xe_exec_queue *e,
drivers/gpu/drm/xe/xe_exec_queue.h
139
void xe_exec_queue_last_fence_set(struct xe_exec_queue *e, struct xe_vm *vm,
drivers/gpu/drm/xe/xe_gt_sriov_pf_monitor.c
122
int e;
drivers/gpu/drm/xe/xe_gt_sriov_pf_monitor.c
129
for (e = 0; e < XE_GUC_KLV_NUM_THRESHOLDS; e++)
drivers/gpu/drm/xe/xe_gt_sriov_pf_monitor.c
130
if (data->guc.events[e])
drivers/gpu/drm/xe/xe_gt_sriov_pf_monitor.c
134
if (e >= XE_GUC_KLV_NUM_THRESHOLDS &&
drivers/gpu/drm/xe/xe_gt_sriov_pf_monitor.c
26
int e;
drivers/gpu/drm/xe/xe_gt_sriov_pf_monitor.c
31
for (e = 0; e < XE_GUC_KLV_NUM_THRESHOLDS; e++)
drivers/gpu/drm/xe/xe_gt_sriov_pf_monitor.c
32
gt->sriov.pf.vfs[vfid].monitor.guc.events[e] = 0;
drivers/gpu/drm/xe/xe_gt_sriov_pf_monitor.c
36
enum xe_guc_klv_threshold_index e)
drivers/gpu/drm/xe/xe_gt_sriov_pf_monitor.c
39
xe_gt_assert(gt, e < XE_GUC_KLV_NUM_THRESHOLDS);
drivers/gpu/drm/xe/xe_gt_sriov_pf_monitor.c
41
gt->sriov.pf.vfs[vfid].monitor.guc.events[e]++;
drivers/gpu/drm/xe/xe_gt_sriov_pf_monitor.c
47
int e;
drivers/gpu/drm/xe/xe_gt_sriov_pf_monitor.c
49
e = xe_guc_klv_threshold_key_to_index(threshold);
drivers/gpu/drm/xe/xe_gt_sriov_pf_monitor.c
53
if (unlikely(e < 0)) {
drivers/gpu/drm/xe/xe_gt_sriov_pf_monitor.c
60
origin, xe_gt_sriov_pf_config_get_threshold(gt, vfid, e),
drivers/gpu/drm/xe/xe_gt_sriov_pf_monitor.c
63
pf_update_event_counter(gt, vfid, e);
drivers/gpu/drm/xe/xe_guc_ads.c
744
} *e, extra_regs[] = {
drivers/gpu/drm/xe/xe_guc_ads.c
759
for (e = extra_regs; e < extra_regs + ARRAY_SIZE(extra_regs); e++) {
drivers/gpu/drm/xe/xe_guc_ads.c
760
if (e->skip)
drivers/gpu/drm/xe/xe_guc_ads.c
763
guc_mmio_regset_write_one(ads, regset_map, e->reg, count++);
drivers/gpu/drm/xe/xe_reg_sr.c
110
idx, e->clr_bits, e->set_bits,
drivers/gpu/drm/xe/xe_reg_sr.c
111
str_yes_no(e->reg.masked),
drivers/gpu/drm/xe/xe_reg_sr.c
112
str_yes_no(e->reg.mcr),
drivers/gpu/drm/xe/xe_reg_sr.c
72
const struct xe_reg_sr_entry *e,
drivers/gpu/drm/xe/xe_reg_sr.c
75
unsigned long idx = e->reg.addr;
drivers/gpu/drm/xe/xe_reg_sr.c
80
if (!compatible_entries(pentry, e)) {
drivers/gpu/drm/xe/xe_reg_sr.c
85
pentry->clr_bits |= e->clr_bits;
drivers/gpu/drm/xe/xe_reg_sr.c
86
pentry->set_bits |= e->set_bits;
drivers/gpu/drm/xe/xe_reg_sr.c
87
pentry->read_mask |= e->read_mask;
drivers/gpu/drm/xe/xe_reg_sr.c
98
*pentry = *e;
drivers/gpu/drm/xe/xe_reg_sr.h
23
int xe_reg_sr_add(struct xe_reg_sr *sr, const struct xe_reg_sr_entry *e,
drivers/gpu/drm/xe/xe_vm.c
1911
struct xe_sync_entry *e = &syncs[i];
drivers/gpu/drm/xe/xe_vm.c
1913
if (xe_sync_is_ufence(e))
drivers/gpu/drm/xe/xe_vm.c
1914
return xe_sync_ufence_get(e);
drivers/hid/bpf/hid_bpf_dispatch.c
106
list_for_each_entry_srcu(e, &hdev->bpf.prog_list, list,
drivers/hid/bpf/hid_bpf_dispatch.c
108
if (!e->hid_hw_request)
drivers/hid/bpf/hid_bpf_dispatch.c
111
ret = e->hid_hw_request(&ctx_kern.ctx, reportnum, rtype, reqtype, source);
drivers/hid/bpf/hid_bpf_dispatch.c
136
struct hid_bpf_ops *e;
drivers/hid/bpf/hid_bpf_dispatch.c
143
list_for_each_entry_srcu(e, &hdev->bpf.prog_list, list,
drivers/hid/bpf/hid_bpf_dispatch.c
145
if (!e->hid_hw_output_report)
drivers/hid/bpf/hid_bpf_dispatch.c
148
ret = e->hid_hw_output_report(&ctx_kern.ctx, source);
drivers/hid/bpf/hid_bpf_dispatch.c
38
struct hid_bpf_ops *e;
drivers/hid/bpf/hid_bpf_dispatch.c
55
list_for_each_entry_rcu(e, &hdev->bpf.prog_list, list) {
drivers/hid/bpf/hid_bpf_dispatch.c
56
if (e->hid_device_event) {
drivers/hid/bpf/hid_bpf_dispatch.c
57
ret = e->hid_device_event(&ctx_kern.ctx, type, source);
drivers/hid/bpf/hid_bpf_dispatch.c
603
struct hid_bpf_ops *e;
drivers/hid/bpf/hid_bpf_dispatch.c
606
list_for_each_entry_rcu(e, &hdev->bpf.prog_list, list) {
drivers/hid/bpf/hid_bpf_dispatch.c
607
if (e->hid_device_event) {
drivers/hid/bpf/hid_bpf_dispatch.c
96
struct hid_bpf_ops *e;
drivers/hid/bpf/hid_bpf_struct_ops.c
312
struct hid_bpf_ops *e;
drivers/hid/bpf/hid_bpf_struct_ops.c
315
list_for_each_entry_rcu(e, &hdev->bpf.prog_list, list) {
drivers/hid/bpf/hid_bpf_struct_ops.c
317
e->hdev = NULL;
drivers/hid/bpf/progs/hid_bpf_async.h
100
struct hid_bpf_async_map_elem *e = (struct hid_bpf_async_map_elem *)value;
drivers/hid/bpf/progs/hid_bpf_async.h
102
bpf_wq_start(&e->wq, 0);
drivers/hid/bpf/progs/hid_bpf_async.h
51
struct hid_bpf_async_map_elem *e; \
drivers/hid/bpf/progs/hid_bpf_async.h
54
e = (struct hid_bpf_async_map_elem *)value; \
drivers/hid/bpf/progs/hid_bpf_async.h
55
ctx = hid_bpf_allocate_context(e->hid); \
drivers/hid/bpf/progs/hid_bpf_async.h
59
e->state = HID_BPF_ASYNC_STATE_RUNNING; \
drivers/hid/bpf/progs/hid_bpf_async.h
63
e->state = HID_BPF_ASYNC_STATE_INITIALIZED; \
drivers/hid/bpf/progs/hid_report_helpers.h
25
#define must_be(e, msg_) \
drivers/hid/bpf/progs/hid_report_helpers.h
29
_Static_assert(e, msg_); \
drivers/hid/hid-ntrig.c
106
return sprintf(buf, "%u.%u.%u.%u.%u", a, b, c, d, e);
drivers/hid/hid-ntrig.c
99
__u8 e = raw[2] & 0x07;
drivers/hwmon/pmbus/ltc2978.c
228
s16 e = ((s16)data) >> 11;
drivers/hwmon/pmbus/ltc2978.c
235
e += 6;
drivers/hwmon/pmbus/ltc2978.c
236
return (e < 0 ? m >> -e : m << e);
drivers/hwmon/pmbus/stpddc60.c
66
s16 e, m, d;
drivers/hwmon/pmbus/stpddc60.c
68
e = ((s16)word) >> 11;
drivers/hwmon/pmbus/stpddc60.c
70
d = e - fixed;
drivers/iio/industrialio-core.c
527
const struct iio_enum *e = (const struct iio_enum *)priv;
drivers/iio/industrialio-core.c
531
if (!e->num_items)
drivers/iio/industrialio-core.c
534
for (i = 0; i < e->num_items; ++i) {
drivers/iio/industrialio-core.c
535
if (!e->items[i])
drivers/iio/industrialio-core.c
537
len += sysfs_emit_at(buf, len, "%s ", e->items[i]);
drivers/iio/industrialio-core.c
550
const struct iio_enum *e = (const struct iio_enum *)priv;
drivers/iio/industrialio-core.c
553
if (!e->get)
drivers/iio/industrialio-core.c
556
i = e->get(indio_dev, chan);
drivers/iio/industrialio-core.c
559
if (i >= e->num_items || !e->items[i])
drivers/iio/industrialio-core.c
562
return sysfs_emit(buf, "%s\n", e->items[i]);
drivers/iio/industrialio-core.c
570
const struct iio_enum *e = (const struct iio_enum *)priv;
drivers/iio/industrialio-core.c
573
if (!e->set)
drivers/iio/industrialio-core.c
576
ret = __sysfs_match_string(e->items, e->num_items, buf);
drivers/iio/industrialio-core.c
580
ret = e->set(indio_dev, chan, ret);
drivers/infiniband/core/iwcm.c
160
struct list_head *e, *tmp;
drivers/infiniband/core/iwcm.c
162
list_for_each_safe(e, tmp, &cm_id_priv->work_free_list) {
drivers/infiniband/core/iwcm.c
163
list_del(e);
drivers/infiniband/core/iwcm.c
164
kfree(list_entry(e, struct iwcm_work, free_list));
drivers/infiniband/core/restrack.c
68
struct rdma_restrack_entry *e;
drivers/infiniband/core/restrack.c
73
xas_for_each(&xas, e, U32_MAX) {
drivers/infiniband/core/restrack.c
74
if (xa_get_mark(&rt->xa, e->id, RESTRACK_DD) && !show_details)
drivers/infiniband/core/sa_query.c
2410
int s, e, i;
drivers/infiniband/core/sa_query.c
2415
e = rdma_end_port(device);
drivers/infiniband/core/sa_query.c
2417
sa_dev = kzalloc_flex(*sa_dev, port, size_add(size_sub(e, s), 1));
drivers/infiniband/core/sa_query.c
2422
sa_dev->end_port = e;
drivers/infiniband/core/sa_query.c
2424
for (i = 0; i <= e - s; ++i) {
drivers/infiniband/core/sa_query.c
2469
for (i = 0; i <= e - s; ++i) {
drivers/infiniband/core/user_mad.c
1392
int s, e, i;
drivers/infiniband/core/user_mad.c
1397
e = rdma_end_port(device);
drivers/infiniband/core/user_mad.c
1399
umad_dev = kzalloc_flex(*umad_dev, ports, size_add(size_sub(e, s), 1));
drivers/infiniband/core/user_mad.c
1404
for (i = s; i <= e; ++i) {
drivers/infiniband/core/uverbs_ioctl.c
250
struct uverbs_attr *e = &bundle->attrs[attr_bkey];
drivers/infiniband/core/uverbs_ioctl.c
268
e->ptr_attr.enum_id = uattr->attr_data.enum_data.elem_id;
drivers/infiniband/core/uverbs_ioctl.c
292
e->ptr_attr.uattr_idx = uattr - pbundle->uattrs;
drivers/infiniband/core/uverbs_ioctl.c
293
e->ptr_attr.len = uattr->len;
drivers/infiniband/core/uverbs_ioctl.c
295
if (val_spec->alloc_and_copy && !uverbs_attr_ptr_is_inline(e)) {
drivers/infiniband/core/uverbs_ioctl.c
302
e->ptr_attr.ptr = p;
drivers/infiniband/core/uverbs_ioctl.c
308
e->ptr_attr.data = uattr->data;
drivers/infiniband/core/uverbs_ioctl.c
320
o_attr = &e->obj_attr;
drivers/infiniband/core/uverbs_ioctl.c
352
e->ptr_attr.data = uattr->data_s64;
drivers/infiniband/core/uverbs_ioctl.c
357
&e->objs_arr_attr, uattr,
drivers/infiniband/hw/cxgb4/cm.c
4071
struct l2t_entry *e;
drivers/infiniband/hw/cxgb4/cm.c
4161
e = cxgb4_l2t_get(dev->rdev.lldi.l2t, neigh,
drivers/infiniband/hw/cxgb4/cm.c
4167
e = cxgb4_l2t_get(dev->rdev.lldi.l2t, neigh,
drivers/infiniband/hw/cxgb4/cm.c
4172
if (!e) {
drivers/infiniband/hw/cxgb4/cm.c
4185
e));
drivers/infiniband/hw/cxgb4/cm.c
4196
cxgb4_l2t_release(e);
drivers/infiniband/hw/hfi1/init.c
1086
u32 e;
drivers/infiniband/hw/hfi1/init.c
1107
for (e = 0; e < rcd->egrbufs.alloced; e++) {
drivers/infiniband/hw/hfi1/init.c
1108
if (rcd->egrbufs.buffers[e].addr)
drivers/infiniband/hw/hfi1/init.c
1110
rcd->egrbufs.buffers[e].len,
drivers/infiniband/hw/hfi1/init.c
1111
rcd->egrbufs.buffers[e].addr,
drivers/infiniband/hw/hfi1/init.c
1112
rcd->egrbufs.buffers[e].dma);
drivers/infiniband/hw/hfi1/opfn.c
135
void opfn_conn_response(struct rvt_qp *qp, struct rvt_ack_entry *e,
drivers/infiniband/hw/hfi1/opfn.c
153
e->atomic_data = capcode;
drivers/infiniband/hw/hfi1/opfn.c
170
e->atomic_data = (data & ~0xf) | capcode;
drivers/infiniband/hw/hfi1/opfn.h
78
void opfn_conn_response(struct rvt_qp *qp, struct rvt_ack_entry *e,
drivers/infiniband/hw/hfi1/pio.c
1753
struct pio_map_elem *e;
drivers/infiniband/hw/hfi1/pio.c
1772
e = m->map[vl & m->mask];
drivers/infiniband/hw/hfi1/pio.c
1773
rval = e->ksc[selector & e->mask];
drivers/infiniband/hw/hfi1/qp.c
605
struct rvt_ack_entry *e = NULL;
drivers/infiniband/hw/hfi1/qp.c
613
e = &qp->s_ack_queue[qp->s_tail_ack_queue];
drivers/infiniband/hw/hfi1/qp.c
659
e ? e->opcode : 0,
drivers/infiniband/hw/hfi1/qp.c
660
e ? e->psn : 0,
drivers/infiniband/hw/hfi1/qp.c
661
e ? e->lpsn : 0,
drivers/infiniband/hw/hfi1/rc.c
100
e = &qp->s_ack_queue[qp->s_tail_ack_queue];
drivers/infiniband/hw/hfi1/rc.c
101
release_rdma_sge_mr(e);
drivers/infiniband/hw/hfi1/rc.c
115
e = &qp->s_ack_queue[qp->s_tail_ack_queue];
drivers/infiniband/hw/hfi1/rc.c
116
if (e->opcode != TID_OP(WRITE_REQ) &&
drivers/infiniband/hw/hfi1/rc.c
120
trace_hfi1_rsp_make_rc_ack(qp, e->psn);
drivers/infiniband/hw/hfi1/rc.c
131
e = &qp->s_ack_queue[qp->s_tail_ack_queue];
drivers/infiniband/hw/hfi1/rc.c
134
hfi1_tid_rdma_ack_interlock(qp, e)) {
drivers/infiniband/hw/hfi1/rc.c
138
if (e->opcode == OP(RDMA_READ_REQUEST)) {
drivers/infiniband/hw/hfi1/rc.c
145
len = e->rdma_sge.sge_length;
drivers/infiniband/hw/hfi1/rc.c
146
if (len && !e->rdma_sge.mr) {
drivers/infiniband/hw/hfi1/rc.c
155
ps->s_txreq->mr = e->rdma_sge.mr;
drivers/infiniband/hw/hfi1/rc.c
158
qp->s_ack_rdma_sge.sge = e->rdma_sge;
drivers/infiniband/hw/hfi1/rc.c
166
e->sent = 1;
drivers/infiniband/hw/hfi1/rc.c
170
qp->s_ack_rdma_psn = e->psn;
drivers/infiniband/hw/hfi1/rc.c
172
} else if (e->opcode == TID_OP(WRITE_REQ)) {
drivers/infiniband/hw/hfi1/rc.c
180
req = ack_to_tid_req(e);
drivers/infiniband/hw/hfi1/rc.c
185
qp->s_ack_rdma_psn = mask_psn(e->psn + req->cur_seg);
drivers/infiniband/hw/hfi1/rc.c
187
} else if (e->opcode == TID_OP(READ_REQ)) {
drivers/infiniband/hw/hfi1/rc.c
194
len = e->rdma_sge.sge_length;
drivers/infiniband/hw/hfi1/rc.c
195
if (len && !e->rdma_sge.mr) {
drivers/infiniband/hw/hfi1/rc.c
20
struct rvt_ack_entry *e = NULL;
drivers/infiniband/hw/hfi1/rc.c
204
ps->s_txreq->mr = e->rdma_sge.mr;
drivers/infiniband/hw/hfi1/rc.c
207
qp->s_ack_rdma_sge.sge = e->rdma_sge;
drivers/infiniband/hw/hfi1/rc.c
217
ib_u64_put(e->atomic_data, &ohdr->u.at.atomic_ack_eth);
drivers/infiniband/hw/hfi1/rc.c
219
bth2 = mask_psn(e->psn);
drivers/infiniband/hw/hfi1/rc.c
220
e->sent = 1;
drivers/infiniband/hw/hfi1/rc.c
242
e = &qp->s_ack_queue[qp->s_tail_ack_queue];
drivers/infiniband/hw/hfi1/rc.c
243
e->sent = 1;
drivers/infiniband/hw/hfi1/rc.c
2515
struct rvt_ack_entry *e;
drivers/infiniband/hw/hfi1/rc.c
2559
e = NULL;
drivers/infiniband/hw/hfi1/rc.c
2565
e = find_prev_entry(qp, psn, &prev, &mra, &old_req);
drivers/infiniband/hw/hfi1/rc.c
2577
if (!e || e->opcode != OP(RDMA_READ_REQUEST))
drivers/infiniband/hw/hfi1/rc.c
2588
offset = delta_psn(psn, e->psn) * qp->pmtu;
drivers/infiniband/hw/hfi1/rc.c
2590
if (unlikely(offset + len != e->rdma_sge.sge_length))
drivers/infiniband/hw/hfi1/rc.c
2592
release_rdma_sge_mr(e);
drivers/infiniband/hw/hfi1/rc.c
2598
ok = rvt_rkey_ok(qp, &e->rdma_sge, len, vaddr, rkey,
drivers/infiniband/hw/hfi1/rc.c
2603
e->rdma_sge.vaddr = NULL;
drivers/infiniband/hw/hfi1/rc.c
2604
e->rdma_sge.length = 0;
drivers/infiniband/hw/hfi1/rc.c
2605
e->rdma_sge.sge_length = 0;
drivers/infiniband/hw/hfi1/rc.c
2607
e->psn = psn;
drivers/infiniband/hw/hfi1/rc.c
2623
if (!e || e->opcode != (u8)opcode || old_req)
drivers/infiniband/hw/hfi1/rc.c
270
e = &qp->s_ack_queue[qp->s_tail_ack_queue];
drivers/infiniband/hw/hfi1/rc.c
271
req = ack_to_tid_req(e);
drivers/infiniband/hw/hfi1/rc.c
286
hdrlen = hfi1_build_tid_rdma_write_resp(qp, e, ohdr, &bth1,
drivers/infiniband/hw/hfi1/rc.c
295
trace_hfi1_tid_req_make_rc_ack_write(qp, 0, e->opcode, e->psn,
drivers/infiniband/hw/hfi1/rc.c
296
e->lpsn, req);
drivers/infiniband/hw/hfi1/rc.c
300
e->sent = 1;
drivers/infiniband/hw/hfi1/rc.c
3016
struct rvt_ack_entry *e;
drivers/infiniband/hw/hfi1/rc.c
3032
e = &qp->s_ack_queue[qp->r_head_ack_queue];
drivers/infiniband/hw/hfi1/rc.c
3033
release_rdma_sge_mr(e);
drivers/infiniband/hw/hfi1/rc.c
3042
ok = rvt_rkey_ok(qp, &e->rdma_sge, len, vaddr,
drivers/infiniband/hw/hfi1/rc.c
3052
e->rdma_sge.mr = NULL;
drivers/infiniband/hw/hfi1/rc.c
3053
e->rdma_sge.vaddr = NULL;
drivers/infiniband/hw/hfi1/rc.c
3054
e->rdma_sge.length = 0;
drivers/infiniband/hw/hfi1/rc.c
3055
e->rdma_sge.sge_length = 0;
drivers/infiniband/hw/hfi1/rc.c
3057
e->opcode = opcode;
drivers/infiniband/hw/hfi1/rc.c
3058
e->sent = 0;
drivers/infiniband/hw/hfi1/rc.c
3059
e->psn = psn;
drivers/infiniband/hw/hfi1/rc.c
3060
e->lpsn = qp->r_psn;
drivers/infiniband/hw/hfi1/rc.c
307
e = &qp->s_ack_queue[qp->s_tail_ack_queue];
drivers/infiniband/hw/hfi1/rc.c
3089
struct rvt_ack_entry *e;
drivers/infiniband/hw/hfi1/rc.c
309
delta = hfi1_build_tid_rdma_read_resp(qp, e, ohdr, &bth0,
drivers/infiniband/hw/hfi1/rc.c
3107
e = &qp->s_ack_queue[qp->r_head_ack_queue];
drivers/infiniband/hw/hfi1/rc.c
3108
release_rdma_sge_mr(e);
drivers/infiniband/hw/hfi1/rc.c
3111
opfn_conn_response(qp, e, ateth);
drivers/infiniband/hw/hfi1/rc.c
3125
e->atomic_data = (opcode == OP(FETCH_ADD)) ?
drivers/infiniband/hw/hfi1/rc.c
3133
e->opcode = opcode;
drivers/infiniband/hw/hfi1/rc.c
3134
e->sent = 0;
drivers/infiniband/hw/hfi1/rc.c
3135
e->psn = psn;
drivers/infiniband/hw/hfi1/rc.c
3136
e->lpsn = psn;
drivers/infiniband/hw/hfi1/rc.c
316
e->sent = 1;
drivers/infiniband/hw/hfi1/rc.c
32
e = NULL;
drivers/infiniband/hw/hfi1/rc.c
35
e = &qp->s_ack_queue[p];
drivers/infiniband/hw/hfi1/rc.c
36
if (!e->opcode) {
drivers/infiniband/hw/hfi1/rc.c
37
e = NULL;
drivers/infiniband/hw/hfi1/rc.c
40
if (cmp_psn(psn, e->psn) >= 0) {
drivers/infiniband/hw/hfi1/rc.c
42
cmp_psn(psn, e->lpsn) <= 0)
drivers/infiniband/hw/hfi1/rc.c
53
return e;
drivers/infiniband/hw/hfi1/rc.c
71
struct rvt_ack_entry *e;
drivers/infiniband/hw/hfi1/rc.h
44
static inline void release_rdma_sge_mr(struct rvt_ack_entry *e)
drivers/infiniband/hw/hfi1/rc.h
46
if (e->rdma_sge.mr) {
drivers/infiniband/hw/hfi1/rc.h
47
rvt_put_mr(e->rdma_sge.mr);
drivers/infiniband/hw/hfi1/rc.h
48
e->rdma_sge.mr = NULL;
drivers/infiniband/hw/hfi1/sdma.c
746
struct sdma_map_elem *e;
drivers/infiniband/hw/hfi1/sdma.c
764
e = m->map[vl & m->mask];
drivers/infiniband/hw/hfi1/sdma.c
765
rval = e->sde[selector & e->mask];
drivers/infiniband/hw/hfi1/tid_rdma.c
1890
struct rvt_ack_entry *e,
drivers/infiniband/hw/hfi1/tid_rdma.c
1900
req = ack_to_tid_req(e);
drivers/infiniband/hw/hfi1/tid_rdma.c
1964
e->opcode = (bth0 >> 24) & 0xff;
drivers/infiniband/hw/hfi1/tid_rdma.c
1965
e->psn = psn;
drivers/infiniband/hw/hfi1/tid_rdma.c
1966
e->lpsn = psn + flow->npkts - 1;
drivers/infiniband/hw/hfi1/tid_rdma.c
1967
e->sent = 0;
drivers/infiniband/hw/hfi1/tid_rdma.c
1978
req->r_flow_psn = e->psn;
drivers/infiniband/hw/hfi1/tid_rdma.c
1980
trace_hfi1_tid_req_rcv_read_req(qp, 0, e->opcode, e->psn, e->lpsn,
drivers/infiniband/hw/hfi1/tid_rdma.c
1993
struct rvt_ack_entry *e;
drivers/infiniband/hw/hfi1/tid_rdma.c
2015
e = find_prev_entry(qp, psn, &prev, NULL, &old_req);
drivers/infiniband/hw/hfi1/tid_rdma.c
2016
if (!e || (e->opcode != TID_OP(READ_REQ) &&
drivers/infiniband/hw/hfi1/tid_rdma.c
2017
e->opcode != TID_OP(WRITE_REQ)))
drivers/infiniband/hw/hfi1/tid_rdma.c
2020
req = ack_to_tid_req(e);
drivers/infiniband/hw/hfi1/tid_rdma.c
2022
trace_hfi1_tid_req_rcv_err(qp, 0, e->opcode, e->psn, e->lpsn, req);
drivers/infiniband/hw/hfi1/tid_rdma.c
2023
if (e->opcode == TID_OP(READ_REQ)) {
drivers/infiniband/hw/hfi1/tid_rdma.c
2037
if (psn != e->psn || len != req->total_len)
drivers/infiniband/hw/hfi1/tid_rdma.c
2040
release_rdma_sge_mr(e);
drivers/infiniband/hw/hfi1/tid_rdma.c
2046
ok = rvt_rkey_ok(qp, &e->rdma_sge, len, vaddr, rkey,
drivers/infiniband/hw/hfi1/tid_rdma.c
2062
if (tid_rdma_rcv_read_request(qp, e, packet, ohdr, bth0, psn,
drivers/infiniband/hw/hfi1/tid_rdma.c
2098
e = &qp->s_ack_queue[i];
drivers/infiniband/hw/hfi1/tid_rdma.c
2099
req = ack_to_tid_req(e);
drivers/infiniband/hw/hfi1/tid_rdma.c
2100
if (e->opcode == TID_OP(WRITE_REQ) &&
drivers/infiniband/hw/hfi1/tid_rdma.c
2146
req->cur_seg = delta_psn(psn, e->psn);
drivers/infiniband/hw/hfi1/tid_rdma.c
2160
e = &qp->s_ack_queue[i];
drivers/infiniband/hw/hfi1/tid_rdma.c
2161
req = ack_to_tid_req(e);
drivers/infiniband/hw/hfi1/tid_rdma.c
2162
trace_hfi1_tid_req_rcv_err(qp, 0, e->opcode, e->psn,
drivers/infiniband/hw/hfi1/tid_rdma.c
2163
e->lpsn, req);
drivers/infiniband/hw/hfi1/tid_rdma.c
2164
if (e->opcode != TID_OP(WRITE_REQ) ||
drivers/infiniband/hw/hfi1/tid_rdma.c
2201
qp->r_psn = e->lpsn + 1;
drivers/infiniband/hw/hfi1/tid_rdma.c
2205
qp->r_state = e->opcode;
drivers/infiniband/hw/hfi1/tid_rdma.c
2234
struct rvt_ack_entry *e;
drivers/infiniband/hw/hfi1/tid_rdma.c
2284
e = &qp->s_ack_queue[qp->r_head_ack_queue];
drivers/infiniband/hw/hfi1/tid_rdma.c
2285
release_rdma_sge_mr(e);
drivers/infiniband/hw/hfi1/tid_rdma.c
2290
if (unlikely(!rvt_rkey_ok(qp, &e->rdma_sge, qp->r_len, vaddr,
drivers/infiniband/hw/hfi1/tid_rdma.c
2295
if (tid_rdma_rcv_read_request(qp, e, packet, ohdr, bth0, psn, vaddr,
drivers/infiniband/hw/hfi1/tid_rdma.c
2299
qp->r_state = e->opcode;
drivers/infiniband/hw/hfi1/tid_rdma.c
2307
qp->r_psn += e->lpsn - e->psn + 1;
drivers/infiniband/hw/hfi1/tid_rdma.c
2344
u32 hfi1_build_tid_rdma_read_resp(struct rvt_qp *qp, struct rvt_ack_entry *e,
drivers/infiniband/hw/hfi1/tid_rdma.c
2348
struct hfi1_ack_priv *epriv = e->priv;
drivers/infiniband/hw/hfi1/tid_rdma.c
2501
len = restart_sge(&ss, req->e.swqe, ipsn, pmtu);
drivers/infiniband/hw/hfi1/tid_rdma.c
2528
trace_hfi1_tid_req_rcv_read_resp(qp, 0, req->e.swqe->wr.opcode,
drivers/infiniband/hw/hfi1/tid_rdma.c
2529
req->e.swqe->psn, req->e.swqe->lpsn,
drivers/infiniband/hw/hfi1/tid_rdma.c
2864
struct rvt_ack_entry *e;
drivers/infiniband/hw/hfi1/tid_rdma.c
2939
e = &qp->s_ack_queue[qpriv->r_tid_tail];
drivers/infiniband/hw/hfi1/tid_rdma.c
2940
if (e->opcode != TID_OP(WRITE_REQ))
drivers/infiniband/hw/hfi1/tid_rdma.c
2942
req = ack_to_tid_req(e);
drivers/infiniband/hw/hfi1/tid_rdma.c
2949
trace_hfi1_tid_req_handle_kdeth_eflags(qp, 0, e->opcode, e->psn,
drivers/infiniband/hw/hfi1/tid_rdma.c
2950
e->lpsn, req);
drivers/infiniband/hw/hfi1/tid_rdma.c
3195
struct rvt_ack_entry *e = &qp->s_ack_queue[i];
drivers/infiniband/hw/hfi1/tid_rdma.c
3200
if (e->opcode != TID_OP(WRITE_REQ))
drivers/infiniband/hw/hfi1/tid_rdma.c
3203
struct hfi1_ack_priv *priv = e->priv;
drivers/infiniband/hw/hfi1/tid_rdma.c
3470
struct rvt_ack_entry *e;
drivers/infiniband/hw/hfi1/tid_rdma.c
3506
e = &qp->s_ack_queue[qpriv->r_tid_alloc];
drivers/infiniband/hw/hfi1/tid_rdma.c
3507
if (e->opcode != TID_OP(WRITE_REQ))
drivers/infiniband/hw/hfi1/tid_rdma.c
3509
req = ack_to_tid_req(e);
drivers/infiniband/hw/hfi1/tid_rdma.c
3510
trace_hfi1_tid_req_write_alloc_res(qp, 0, e->opcode, e->psn,
drivers/infiniband/hw/hfi1/tid_rdma.c
3511
e->lpsn, req);
drivers/infiniband/hw/hfi1/tid_rdma.c
3602
qp->r_psn = e->psn + req->alloc_seg;
drivers/infiniband/hw/hfi1/tid_rdma.c
3663
struct rvt_ack_entry *e;
drivers/infiniband/hw/hfi1/tid_rdma.c
3719
e = &qp->s_ack_queue[qp->r_head_ack_queue];
drivers/infiniband/hw/hfi1/tid_rdma.c
3720
req = ack_to_tid_req(e);
drivers/infiniband/hw/hfi1/tid_rdma.c
3727
qp->r_psn = e->lpsn + 1;
drivers/infiniband/hw/hfi1/tid_rdma.c
3732
release_rdma_sge_mr(e);
drivers/infiniband/hw/hfi1/tid_rdma.c
3741
if (e->opcode == TID_OP(WRITE_REQ) &&
drivers/infiniband/hw/hfi1/tid_rdma.c
3746
if (unlikely(!rvt_rkey_ok(qp, &e->rdma_sge, qp->r_len, vaddr,
drivers/infiniband/hw/hfi1/tid_rdma.c
3752
e->opcode = (bth0 >> 24) & 0xff;
drivers/infiniband/hw/hfi1/tid_rdma.c
3753
e->psn = psn;
drivers/infiniband/hw/hfi1/tid_rdma.c
3754
e->lpsn = qp->r_psn;
drivers/infiniband/hw/hfi1/tid_rdma.c
3755
e->sent = 0;
drivers/infiniband/hw/hfi1/tid_rdma.c
3767
req->r_flow_psn = e->psn;
drivers/infiniband/hw/hfi1/tid_rdma.c
3768
req->ss.sge = e->rdma_sge;
drivers/infiniband/hw/hfi1/tid_rdma.c
3775
qp->r_state = e->opcode;
drivers/infiniband/hw/hfi1/tid_rdma.c
3785
trace_hfi1_tid_req_rcv_write_req(qp, 0, e->opcode, e->psn, e->lpsn,
drivers/infiniband/hw/hfi1/tid_rdma.c
3793
e = &qp->s_ack_queue[qpriv->r_tid_tail];
drivers/infiniband/hw/hfi1/tid_rdma.c
3794
ptr = ack_to_tid_req(e);
drivers/infiniband/hw/hfi1/tid_rdma.c
3796
if (e->opcode != TID_OP(WRITE_REQ) ||
drivers/infiniband/hw/hfi1/tid_rdma.c
3835
u32 hfi1_build_tid_rdma_write_resp(struct rvt_qp *qp, struct rvt_ack_entry *e,
drivers/infiniband/hw/hfi1/tid_rdma.c
3840
struct hfi1_ack_priv *epriv = e->priv;
drivers/infiniband/hw/hfi1/tid_rdma.c
3848
trace_hfi1_tid_req_build_write_resp(qp, 0, e->opcode, e->psn, e->lpsn,
drivers/infiniband/hw/hfi1/tid_rdma.c
386
priv->tid_req.e.swqe = wqe;
drivers/infiniband/hw/hfi1/tid_rdma.c
398
priv->tid_req.e.ack = &qp->s_ack_queue[i];
drivers/infiniband/hw/hfi1/tid_rdma.c
4275
struct rvt_ack_entry *e;
drivers/infiniband/hw/hfi1/tid_rdma.c
4293
e = &qp->s_ack_queue[priv->r_tid_tail];
drivers/infiniband/hw/hfi1/tid_rdma.c
4294
req = ack_to_tid_req(e);
drivers/infiniband/hw/hfi1/tid_rdma.c
4333
ss.sge = e->rdma_sge;
drivers/infiniband/hw/hfi1/tid_rdma.c
4361
trace_hfi1_tid_req_rcv_write_data(qp, 0, e->opcode, e->psn, e->lpsn,
drivers/infiniband/hw/hfi1/tid_rdma.c
4367
release_rdma_sge_mr(e);
drivers/infiniband/hw/hfi1/tid_rdma.c
4373
e = &qp->s_ack_queue[next];
drivers/infiniband/hw/hfi1/tid_rdma.c
4374
if (e->opcode == TID_OP(WRITE_REQ))
drivers/infiniband/hw/hfi1/tid_rdma.c
4426
u32 hfi1_build_tid_rdma_write_ack(struct rvt_qp *qp, struct rvt_ack_entry *e,
drivers/infiniband/hw/hfi1/tid_rdma.c
4432
struct tid_rdma_request *req = ack_to_tid_req(e);
drivers/infiniband/hw/hfi1/tid_rdma.c
4877
struct rvt_ack_entry *e;
drivers/infiniband/hw/hfi1/tid_rdma.c
4934
e = &qp->s_ack_queue[idx];
drivers/infiniband/hw/hfi1/tid_rdma.c
4935
if (e->opcode == TID_OP(WRITE_REQ)) {
drivers/infiniband/hw/hfi1/tid_rdma.c
4936
req = ack_to_tid_req(e);
drivers/infiniband/hw/hfi1/tid_rdma.c
4937
trace_hfi1_tid_req_rcv_resync(qp, 0, e->opcode, e->psn,
drivers/infiniband/hw/hfi1/tid_rdma.c
4938
e->lpsn, req);
drivers/infiniband/hw/hfi1/tid_rdma.c
5017
struct rvt_ack_entry *e = &qp->s_ack_queue[qp->s_tail_ack_queue];
drivers/infiniband/hw/hfi1/tid_rdma.c
5018
struct tid_rdma_request *req = ack_to_tid_req(e);
drivers/infiniband/hw/hfi1/tid_rdma.c
5032
(e->opcode == TID_OP(WRITE_REQ) && req->cur_seg < req->alloc_seg &&
drivers/infiniband/hw/hfi1/tid_rdma.c
5191
struct rvt_ack_entry *e;
drivers/infiniband/hw/hfi1/tid_rdma.c
5209
e = &qp->s_ack_queue[qpriv->r_tid_ack];
drivers/infiniband/hw/hfi1/tid_rdma.c
5210
req = ack_to_tid_req(e);
drivers/infiniband/hw/hfi1/tid_rdma.c
5228
e = &qp->s_ack_queue[qpriv->r_tid_ack];
drivers/infiniband/hw/hfi1/tid_rdma.c
5229
req = ack_to_tid_req(e);
drivers/infiniband/hw/hfi1/tid_rdma.c
5232
trace_hfi1_rsp_make_tid_ack(qp, e->psn);
drivers/infiniband/hw/hfi1/tid_rdma.c
5233
trace_hfi1_tid_req_make_tid_ack(qp, 0, e->opcode, e->psn, e->lpsn,
drivers/infiniband/hw/hfi1/tid_rdma.c
5278
e = &qp->s_ack_queue[qpriv->r_tid_ack];
drivers/infiniband/hw/hfi1/tid_rdma.c
5279
req = ack_to_tid_req(e);
drivers/infiniband/hw/hfi1/tid_rdma.c
5299
e = &qp->s_ack_queue[qpriv->r_tid_ack];
drivers/infiniband/hw/hfi1/tid_rdma.c
5300
req = ack_to_tid_req(e);
drivers/infiniband/hw/hfi1/tid_rdma.c
5307
trace_hfi1_tid_req_make_tid_ack(qp, 0, e->opcode, e->psn, e->lpsn,
drivers/infiniband/hw/hfi1/tid_rdma.c
5309
hwords += hfi1_build_tid_rdma_write_ack(qp, e, ohdr, flow, &bth1,
drivers/infiniband/hw/hfi1/tid_rdma.c
5467
bool hfi1_tid_rdma_ack_interlock(struct rvt_qp *qp, struct rvt_ack_entry *e)
drivers/infiniband/hw/hfi1/tid_rdma.c
5479
if ((e->opcode == TID_OP(READ_REQ) ||
drivers/infiniband/hw/hfi1/tid_rdma.c
5480
e->opcode == OP(RDMA_READ_REQUEST)) &&
drivers/infiniband/hw/hfi1/tid_rdma.h
103
} e;
drivers/infiniband/hw/hfi1/tid_rdma.h
251
u32 hfi1_build_tid_rdma_read_resp(struct rvt_qp *qp, struct rvt_ack_entry *e,
drivers/infiniband/hw/hfi1/tid_rdma.h
280
u32 hfi1_build_tid_rdma_write_resp(struct rvt_qp *qp, struct rvt_ack_entry *e,
drivers/infiniband/hw/hfi1/tid_rdma.h
295
u32 hfi1_build_tid_rdma_write_ack(struct rvt_qp *qp, struct rvt_ack_entry *e,
drivers/infiniband/hw/hfi1/tid_rdma.h
317
bool hfi1_tid_rdma_ack_interlock(struct rvt_qp *qp, struct rvt_ack_entry *e);
drivers/infiniband/hw/hfi1/trace_tx.h
113
u16 e,
drivers/infiniband/hw/hfi1/trace_tx.h
115
TP_ARGS(sde, desc0, desc1, e, descp),
drivers/infiniband/hw/hfi1/trace_tx.h
120
__field(u16, e)
drivers/infiniband/hw/hfi1/trace_tx.h
128
__entry->e = e;
drivers/infiniband/hw/hfi1/trace_tx.h
144
__entry->e
drivers/infiniband/hw/hfi1/verbs.h
338
static inline struct tid_rdma_request *ack_to_tid_req(struct rvt_ack_entry *e)
drivers/infiniband/hw/hfi1/verbs.h
340
return &((struct hfi1_ack_priv *)e->priv)->tid_req;
drivers/infiniband/hw/mlx5/srq.h
48
void (*event)(struct mlx5_core_srq *srq, enum mlx5_event e);
drivers/infiniband/hw/vmw_pvrdma/pvrdma_main.c
335
struct ib_event e;
drivers/infiniband/hw/vmw_pvrdma/pvrdma_main.c
337
e.device = ibqp->device;
drivers/infiniband/hw/vmw_pvrdma/pvrdma_main.c
338
e.element.qp = ibqp;
drivers/infiniband/hw/vmw_pvrdma/pvrdma_main.c
339
e.event = type; /* 1:1 mapping for now. */
drivers/infiniband/hw/vmw_pvrdma/pvrdma_main.c
340
ibqp->event_handler(&e, ibqp->qp_context);
drivers/infiniband/hw/vmw_pvrdma/pvrdma_main.c
361
struct ib_event e;
drivers/infiniband/hw/vmw_pvrdma/pvrdma_main.c
363
e.device = ibcq->device;
drivers/infiniband/hw/vmw_pvrdma/pvrdma_main.c
364
e.element.cq = ibcq;
drivers/infiniband/hw/vmw_pvrdma/pvrdma_main.c
365
e.event = type; /* 1:1 mapping for now. */
drivers/infiniband/hw/vmw_pvrdma/pvrdma_main.c
366
ibcq->event_handler(&e, ibcq->cq_context);
drivers/infiniband/hw/vmw_pvrdma/pvrdma_main.c
390
struct ib_event e;
drivers/infiniband/hw/vmw_pvrdma/pvrdma_main.c
392
e.device = ibsrq->device;
drivers/infiniband/hw/vmw_pvrdma/pvrdma_main.c
393
e.element.srq = ibsrq;
drivers/infiniband/hw/vmw_pvrdma/pvrdma_main.c
394
e.event = type; /* 1:1 mapping for now. */
drivers/infiniband/hw/vmw_pvrdma/pvrdma_main.c
395
ibsrq->event_handler(&e, ibsrq->srq_context);
drivers/infiniband/sw/rdmavt/qp.c
604
struct rvt_ack_entry *e = &qp->s_ack_queue[n];
drivers/infiniband/sw/rdmavt/qp.c
606
if (e->rdma_sge.mr) {
drivers/infiniband/sw/rdmavt/qp.c
607
rvt_put_mr(e->rdma_sge.mr);
drivers/infiniband/sw/rdmavt/qp.c
608
e->rdma_sge.mr = NULL;
drivers/infiniband/sw/rdmavt/qp.c
668
struct rvt_ack_entry *e = &qp->s_ack_queue[i];
drivers/infiniband/sw/rdmavt/qp.c
670
if (rvt_mr_has_lkey(e->rdma_sge.mr, lkey))
drivers/infiniband/ulp/ipoib/ipoib_cm.c
1462
int e = skb_queue_empty(&priv->cm.skb_queue);
drivers/infiniband/ulp/ipoib/ipoib_cm.c
1467
if (e)
drivers/infiniband/ulp/isert/ib_isert.c
83
isert_qp_event_callback(struct ib_event *e, void *context)
drivers/infiniband/ulp/isert/ib_isert.c
88
ib_event_msg(e->event), e->event, isert_conn);
drivers/infiniband/ulp/isert/ib_isert.c
90
switch (e->event) {
drivers/infiniband/ulp/srpt/ib_srpt.c
119
struct srpt_memory_cache_entry *e;
drivers/infiniband/ulp/srpt/ib_srpt.c
124
e = xa_load(&srpt_memory_caches, object_size);
drivers/infiniband/ulp/srpt/ib_srpt.c
125
if (e) {
drivers/infiniband/ulp/srpt/ib_srpt.c
126
refcount_inc(&e->ref);
drivers/infiniband/ulp/srpt/ib_srpt.c
127
return e->c;
drivers/infiniband/ulp/srpt/ib_srpt.c
130
e = kmalloc_obj(*e);
drivers/infiniband/ulp/srpt/ib_srpt.c
131
if (!e)
drivers/infiniband/ulp/srpt/ib_srpt.c
133
refcount_set(&e->ref, 1);
drivers/infiniband/ulp/srpt/ib_srpt.c
134
e->c = kmem_cache_create(name, object_size, /*align=*/512, 0, NULL);
drivers/infiniband/ulp/srpt/ib_srpt.c
135
if (!e->c)
drivers/infiniband/ulp/srpt/ib_srpt.c
137
res = xa_store(&srpt_memory_caches, object_size, e, GFP_KERNEL);
drivers/infiniband/ulp/srpt/ib_srpt.c
140
return e->c;
drivers/infiniband/ulp/srpt/ib_srpt.c
143
kmem_cache_destroy(e->c);
drivers/infiniband/ulp/srpt/ib_srpt.c
146
kfree(e);
drivers/infiniband/ulp/srpt/ib_srpt.c
152
struct srpt_memory_cache_entry *e = NULL;
drivers/infiniband/ulp/srpt/ib_srpt.c
156
xa_for_each(&srpt_memory_caches, object_size, e)
drivers/infiniband/ulp/srpt/ib_srpt.c
157
if (e->c == c)
drivers/infiniband/ulp/srpt/ib_srpt.c
159
if (WARN_ON_ONCE(!e))
drivers/infiniband/ulp/srpt/ib_srpt.c
161
if (!refcount_dec_and_test(&e->ref))
drivers/infiniband/ulp/srpt/ib_srpt.c
163
WARN_ON_ONCE(xa_erase(&srpt_memory_caches, object_size) != e);
drivers/infiniband/ulp/srpt/ib_srpt.c
164
kmem_cache_destroy(e->c);
drivers/infiniband/ulp/srpt/ib_srpt.c
165
kfree(e);
drivers/input/gameport/gameport.c
295
struct gameport_event *e, *next;
drivers/input/gameport/gameport.c
300
list_for_each_entry_safe(e, next, &gameport_event_list, node) {
drivers/input/gameport/gameport.c
301
if (event->object == e->object) {
drivers/input/gameport/gameport.c
307
if (event->type != e->type)
drivers/input/gameport/gameport.c
310
list_del_init(&e->node);
drivers/input/gameport/gameport.c
311
gameport_free_event(e);
drivers/input/misc/yealink.c
61
#define _SEG(t, a, am, b, bm, c, cm, d, dm, e, em, f, fm, g, gm) \
drivers/input/misc/yealink.c
64
_LOC(d, dm), _LOC(e, em), _LOC(g, gm), \
drivers/input/mouse/cypress_ps2.h
23
#define CYTP_CMD_SET_PALM_GEOMETRY(e) ENCODE_CMD(1, 2, 1, (e))
drivers/input/serio/serio.c
166
struct serio_event *e, *next;
drivers/input/serio/serio.c
170
list_for_each_entry_safe(e, next, &serio_event_list, node) {
drivers/input/serio/serio.c
171
if (object == e->object) {
drivers/input/serio/serio.c
177
if (type != e->type)
drivers/input/serio/serio.c
180
list_del_init(&e->node);
drivers/input/serio/serio.c
181
serio_free_event(e);
drivers/input/sparse-keymap.c
172
const struct key_entry *e;
drivers/input/sparse-keymap.c
177
for (e = keymap; e->type != KE_END; e++)
drivers/iommu/amd/init.c
1232
struct ivhd_dte_flags *e;
drivers/iommu/amd/init.c
1236
for_each_ivhd_dte_flags(e) {
drivers/iommu/amd/init.c
1241
if ((e->segid == segid) &&
drivers/iommu/amd/init.c
1242
(e->devid_first <= devid) && (devid <= e->devid_last)) {
drivers/iommu/amd/init.c
1243
unsigned int len = e->devid_last - e->devid_first;
drivers/iommu/amd/init.c
1246
dte = &(e->dte);
drivers/iommu/amd/init.c
1256
struct ivhd_dte_flags *e;
drivers/iommu/amd/init.c
1258
for_each_ivhd_dte_flags(e) {
drivers/iommu/amd/init.c
1259
if ((e->segid == segid) &&
drivers/iommu/amd/init.c
1260
(e->devid_first == first) &&
drivers/iommu/amd/init.c
1261
(e->devid_last == last))
drivers/iommu/amd/init.c
1452
struct ivhd_entry *e;
drivers/iommu/amd/init.c
1484
e = (struct ivhd_entry *)p;
drivers/iommu/amd/init.c
1487
switch (e->type) {
drivers/iommu/amd/init.c
1490
DUMP_printk(" DEV_ALL\t\t\tsetting: %#02x\n", e->flags);
drivers/iommu/amd/init.c
1491
set_dev_entry_from_acpi_range(iommu, 0, pci_seg->last_bdf, e->flags, 0);
drivers/iommu/amd/init.c
1496
seg_id, PCI_BUS_NUM(e->devid),
drivers/iommu/amd/init.c
1497
PCI_SLOT(e->devid),
drivers/iommu/amd/init.c
1498
PCI_FUNC(e->devid),
drivers/iommu/amd/init.c
1499
e->flags);
drivers/iommu/amd/init.c
1501
devid = e->devid;
drivers/iommu/amd/init.c
1502
set_dev_entry_from_acpi(iommu, devid, e->flags, 0);
drivers/iommu/amd/init.c
1507
seg_id, PCI_BUS_NUM(e->devid),
drivers/iommu/amd/init.c
1508
PCI_SLOT(e->devid),
drivers/iommu/amd/init.c
1509
PCI_FUNC(e->devid),
drivers/iommu/amd/init.c
1510
e->flags);
drivers/iommu/amd/init.c
1512
devid_start = e->devid;
drivers/iommu/amd/init.c
1513
flags = e->flags;
drivers/iommu/amd/init.c
1520
seg_id, PCI_BUS_NUM(e->devid),
drivers/iommu/amd/init.c
1521
PCI_SLOT(e->devid),
drivers/iommu/amd/init.c
1522
PCI_FUNC(e->devid),
drivers/iommu/amd/init.c
1523
e->flags,
drivers/iommu/amd/init.c
1524
PCI_BUS_NUM(e->ext >> 8),
drivers/iommu/amd/init.c
1525
PCI_SLOT(e->ext >> 8),
drivers/iommu/amd/init.c
1526
PCI_FUNC(e->ext >> 8));
drivers/iommu/amd/init.c
1528
devid = e->devid;
drivers/iommu/amd/init.c
1529
devid_to = e->ext >> 8;
drivers/iommu/amd/init.c
1530
set_dev_entry_from_acpi(iommu, devid , e->flags, 0);
drivers/iommu/amd/init.c
1531
set_dev_entry_from_acpi(iommu, devid_to, e->flags, 0);
drivers/iommu/amd/init.c
1537
seg_id, PCI_BUS_NUM(e->devid),
drivers/iommu/amd/init.c
1538
PCI_SLOT(e->devid),
drivers/iommu/amd/init.c
1539
PCI_FUNC(e->devid),
drivers/iommu/amd/init.c
1540
e->flags,
drivers/iommu/amd/init.c
1541
seg_id, PCI_BUS_NUM(e->ext >> 8),
drivers/iommu/amd/init.c
1542
PCI_SLOT(e->ext >> 8),
drivers/iommu/amd/init.c
1543
PCI_FUNC(e->ext >> 8));
drivers/iommu/amd/init.c
1545
devid_start = e->devid;
drivers/iommu/amd/init.c
1546
flags = e->flags;
drivers/iommu/amd/init.c
1547
devid_to = e->ext >> 8;
drivers/iommu/amd/init.c
1554
seg_id, PCI_BUS_NUM(e->devid),
drivers/iommu/amd/init.c
1555
PCI_SLOT(e->devid),
drivers/iommu/amd/init.c
1556
PCI_FUNC(e->devid),
drivers/iommu/amd/init.c
1557
e->flags, e->ext);
drivers/iommu/amd/init.c
1559
devid = e->devid;
drivers/iommu/amd/init.c
1560
set_dev_entry_from_acpi(iommu, devid, e->flags,
drivers/iommu/amd/init.c
1561
e->ext);
drivers/iommu/amd/init.c
1566
seg_id, PCI_BUS_NUM(e->devid),
drivers/iommu/amd/init.c
1567
PCI_SLOT(e->devid),
drivers/iommu/amd/init.c
1568
PCI_FUNC(e->devid),
drivers/iommu/amd/init.c
1569
e->flags, e->ext);
drivers/iommu/amd/init.c
1571
devid_start = e->devid;
drivers/iommu/amd/init.c
1572
flags = e->flags;
drivers/iommu/amd/init.c
1573
ext_flags = e->ext;
drivers/iommu/amd/init.c
1579
seg_id, PCI_BUS_NUM(e->devid),
drivers/iommu/amd/init.c
1580
PCI_SLOT(e->devid),
drivers/iommu/amd/init.c
1581
PCI_FUNC(e->devid));
drivers/iommu/amd/init.c
1583
devid = e->devid;
drivers/iommu/amd/init.c
1597
handle = e->ext & 0xff;
drivers/iommu/amd/init.c
1598
devid = PCI_SEG_DEVID_TO_SBDF(seg_id, (e->ext >> 8));
drivers/iommu/amd/init.c
1599
type = (e->ext >> 24) & 0xff;
drivers/iommu/amd/init.c
1613
e->flags);
drivers/iommu/amd/init.c
1624
set_dev_entry_from_acpi(iommu, devid, e->flags, 0);
drivers/iommu/amd/init.c
1636
e->type);
drivers/iommu/amd/init.c
1640
BUILD_BUG_ON(sizeof(e->ext_hid) != ACPIHID_HID_LEN - 1);
drivers/iommu/amd/init.c
1641
memcpy(hid, &e->ext_hid, ACPIHID_HID_LEN - 1);
drivers/iommu/amd/init.c
1650
switch (e->uidf) {
drivers/iommu/amd/init.c
1653
if (e->uidl != 0)
drivers/iommu/amd/init.c
1659
sprintf(uid, "%d", e->uid);
drivers/iommu/amd/init.c
1664
memcpy(uid, &e->uid, e->uidl);
drivers/iommu/amd/init.c
1665
uid[e->uidl] = '\0';
drivers/iommu/amd/init.c
1672
devid = PCI_SEG_DEVID_TO_SBDF(seg_id, e->devid);
drivers/iommu/amd/init.c
1678
e->flags);
drivers/iommu/amd/init.c
1680
flags = e->flags;
drivers/iommu/amd/init.c
1691
set_dev_entry_from_acpi(iommu, devid, e->flags, 0);
drivers/iommu/amd/init.c
2634
struct unity_map_entry *e = NULL;
drivers/iommu/amd/init.c
2642
e = kzalloc_obj(*e);
drivers/iommu/amd/init.c
2643
if (e == NULL)
drivers/iommu/amd/init.c
2648
kfree(e);
drivers/iommu/amd/init.c
2652
e->devid_start = e->devid_end = m->devid;
drivers/iommu/amd/init.c
2656
e->devid_start = 0;
drivers/iommu/amd/init.c
2657
e->devid_end = pci_seg->last_bdf;
drivers/iommu/amd/init.c
2661
e->devid_start = m->devid;
drivers/iommu/amd/init.c
2662
e->devid_end = m->aux;
drivers/iommu/amd/init.c
2665
e->address_start = PAGE_ALIGN(m->range_start);
drivers/iommu/amd/init.c
2666
e->address_end = e->address_start + PAGE_ALIGN(m->range_length);
drivers/iommu/amd/init.c
2667
e->prot = m->flags >> 1;
drivers/iommu/amd/init.c
2677
e->prot = (IVMD_FLAG_IW | IVMD_FLAG_IR) >> 1;
drivers/iommu/amd/init.c
2682
PCI_BUS_NUM(e->devid_start), PCI_SLOT(e->devid_start),
drivers/iommu/amd/init.c
2683
PCI_FUNC(e->devid_start), m->pci_seg,
drivers/iommu/amd/init.c
2684
PCI_BUS_NUM(e->devid_end),
drivers/iommu/amd/init.c
2685
PCI_SLOT(e->devid_end), PCI_FUNC(e->devid_end),
drivers/iommu/amd/init.c
2686
e->address_start, e->address_end, m->flags);
drivers/iommu/amd/init.c
2688
list_add_tail(&e->list, &pci_seg->unity_map);
drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3.c
1954
#define ARM_SMMU_EVT_KNOWN(e) ((e)->id < ARRAY_SIZE(event_str) && event_str[(e)->id])
drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3.c
1955
#define ARM_SMMU_LOG_EVT_STR(e) ARM_SMMU_EVT_KNOWN(e) ? event_str[(e)->id] : "UNKNOWN"
drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3.c
1956
#define ARM_SMMU_LOG_CLIENT(e) (e)->dev ? dev_name((e)->dev) : "(unassigned sid)"
drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3.c
4744
struct iommu_resv_region *e;
drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3.c
4749
list_for_each_entry(e, &rmr_list, list) {
drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3.c
4753
rmr = container_of(e, struct iommu_iort_rmr_data, rr);
drivers/iommu/arm/arm-smmu/arm-smmu.c
2076
struct iommu_resv_region *e;
drivers/iommu/arm/arm-smmu/arm-smmu.c
2093
list_for_each_entry(e, &rmr_list, list) {
drivers/iommu/arm/arm-smmu/arm-smmu.c
2097
rmr = container_of(e, struct iommu_iort_rmr_data, rr);
drivers/iommu/hyperv-iommu.c
199
struct IO_APIC_route_entry e;
drivers/iommu/hyperv-iommu.c
229
e.w1 = entry.ioapic_rte.low_uint32;
drivers/iommu/hyperv-iommu.c
230
e.w2 = entry.ioapic_rte.high_uint32;
drivers/iommu/hyperv-iommu.c
233
msg->arch_data.vector = e.vector;
drivers/iommu/hyperv-iommu.c
234
msg->arch_data.delivery_mode = e.delivery_mode;
drivers/iommu/hyperv-iommu.c
235
msg->arch_addr_lo.dest_mode_logical = e.dest_mode_logical;
drivers/iommu/hyperv-iommu.c
236
msg->arch_addr_lo.dmar_format = e.ir_format;
drivers/iommu/hyperv-iommu.c
237
msg->arch_addr_lo.dmar_index_0_14 = e.ir_index_0_14;
drivers/iommu/hyperv-iommu.c
306
struct hv_interrupt_entry *e;
drivers/iommu/hyperv-iommu.c
314
e = &data->entry;
drivers/iommu/hyperv-iommu.c
316
if (e->source == HV_DEVICE_TYPE_IOAPIC &&
drivers/iommu/hyperv-iommu.c
317
e->ioapic_rte.as_uint64)
drivers/iommu/intel/iommu.h
194
#define ecap_pms(e) (((e) >> 51) & 0x1)
drivers/iommu/intel/iommu.h
195
#define ecap_rps(e) (((e) >> 49) & 0x1)
drivers/iommu/intel/iommu.h
196
#define ecap_smpwc(e) (((e) >> 48) & 0x1)
drivers/iommu/intel/iommu.h
197
#define ecap_flts(e) (((e) >> 47) & 0x1)
drivers/iommu/intel/iommu.h
198
#define ecap_slts(e) (((e) >> 46) & 0x1)
drivers/iommu/intel/iommu.h
199
#define ecap_slads(e) (((e) >> 45) & 0x1)
drivers/iommu/intel/iommu.h
200
#define ecap_smts(e) (((e) >> 43) & 0x1)
drivers/iommu/intel/iommu.h
201
#define ecap_dit(e) (((e) >> 41) & 0x1)
drivers/iommu/intel/iommu.h
202
#define ecap_pds(e) (((e) >> 42) & 0x1)
drivers/iommu/intel/iommu.h
203
#define ecap_pasid(e) (((e) >> 40) & 0x1)
drivers/iommu/intel/iommu.h
204
#define ecap_pss(e) (((e) >> 35) & 0x1f)
drivers/iommu/intel/iommu.h
205
#define ecap_eafs(e) (((e) >> 34) & 0x1)
drivers/iommu/intel/iommu.h
206
#define ecap_nwfs(e) (((e) >> 33) & 0x1)
drivers/iommu/intel/iommu.h
207
#define ecap_srs(e) (((e) >> 31) & 0x1)
drivers/iommu/intel/iommu.h
208
#define ecap_ers(e) (((e) >> 30) & 0x1)
drivers/iommu/intel/iommu.h
209
#define ecap_prs(e) (((e) >> 29) & 0x1)
drivers/iommu/intel/iommu.h
210
#define ecap_broken_pasid(e) (((e) >> 28) & 0x1)
drivers/iommu/intel/iommu.h
211
#define ecap_dis(e) (((e) >> 27) & 0x1)
drivers/iommu/intel/iommu.h
212
#define ecap_nest(e) (((e) >> 26) & 0x1)
drivers/iommu/intel/iommu.h
213
#define ecap_mts(e) (((e) >> 25) & 0x1)
drivers/iommu/intel/iommu.h
214
#define ecap_iotlb_offset(e) ((((e) >> 8) & 0x3ff) * 16)
drivers/iommu/intel/iommu.h
215
#define ecap_max_iotlb_offset(e) (ecap_iotlb_offset(e) + 16)
drivers/iommu/intel/iommu.h
216
#define ecap_coherent(e) ((e) & 0x1)
drivers/iommu/intel/iommu.h
217
#define ecap_qis(e) ((e) & 0x2)
drivers/iommu/intel/iommu.h
218
#define ecap_pass_through(e) (((e) >> 6) & 0x1)
drivers/iommu/intel/iommu.h
219
#define ecap_eim_support(e) (((e) >> 4) & 0x1)
drivers/iommu/intel/iommu.h
220
#define ecap_ir_support(e) (((e) >> 3) & 0x1)
drivers/iommu/intel/iommu.h
221
#define ecap_dev_iotlb_support(e) (((e) >> 2) & 0x1)
drivers/iommu/intel/iommu.h
222
#define ecap_max_handle_mask(e) (((e) >> 20) & 0xf)
drivers/iommu/intel/iommu.h
223
#define ecap_sc_support(e) (((e) >> 7) & 0x1) /* Snooping Control */
drivers/iommu/intel/perfmon.c
88
#define iommu_pmu_en_requester_id(e) ((e) & 0x1)
drivers/iommu/intel/perfmon.c
89
#define iommu_pmu_en_domain(e) (((e) >> 1) & 0x1)
drivers/iommu/intel/perfmon.c
90
#define iommu_pmu_en_pasid(e) (((e) >> 2) & 0x1)
drivers/iommu/intel/perfmon.c
91
#define iommu_pmu_en_ats(e) (((e) >> 3) & 0x1)
drivers/iommu/intel/perfmon.c
92
#define iommu_pmu_en_page_table(e) (((e) >> 4) & 0x1)
drivers/iommu/omap-iommu.c
1282
static u32 iotlb_init_entry(struct iotlb_entry *e, u32 da, u32 pa, int pgsz)
drivers/iommu/omap-iommu.c
1284
memset(e, 0, sizeof(*e));
drivers/iommu/omap-iommu.c
1286
e->da = da;
drivers/iommu/omap-iommu.c
1287
e->pa = pa;
drivers/iommu/omap-iommu.c
1288
e->valid = MMU_CAM_V;
drivers/iommu/omap-iommu.c
1289
e->pgsz = pgsz;
drivers/iommu/omap-iommu.c
1290
e->endian = MMU_RAM_ENDIAN_LITTLE;
drivers/iommu/omap-iommu.c
1291
e->elsz = MMU_RAM_ELSZ_8;
drivers/iommu/omap-iommu.c
1292
e->mixed = 0;
drivers/iommu/omap-iommu.c
1294
return iopgsz_to_bytes(e->pgsz);
drivers/iommu/omap-iommu.c
1305
struct iotlb_entry e;
drivers/iommu/omap-iommu.c
1318
iotlb_init_entry(&e, da, pa, omap_pgsz);
drivers/iommu/omap-iommu.c
1323
ret = omap_iopgtable_store_entry(oiommu, &e);
drivers/iommu/omap-iommu.c
219
static u32 get_iopte_attr(struct iotlb_entry *e)
drivers/iommu/omap-iommu.c
223
attr = e->mixed << 5;
drivers/iommu/omap-iommu.c
224
attr |= e->endian;
drivers/iommu/omap-iommu.c
225
attr |= e->elsz >> 3;
drivers/iommu/omap-iommu.c
226
attr <<= (((e->pgsz == MMU_CAM_PGSZ_4K) ||
drivers/iommu/omap-iommu.c
227
(e->pgsz == MMU_CAM_PGSZ_64K)) ? 0 : 6);
drivers/iommu/omap-iommu.c
301
struct iotlb_entry *e)
drivers/iommu/omap-iommu.c
305
if (!e)
drivers/iommu/omap-iommu.c
308
if (e->da & ~(get_cam_va_mask(e->pgsz))) {
drivers/iommu/omap-iommu.c
310
e->da);
drivers/iommu/omap-iommu.c
318
cr->cam = (e->da & MMU_CAM_VATAG_MASK) | e->prsvd | e->pgsz | e->valid;
drivers/iommu/omap-iommu.c
319
cr->ram = e->pa | e->endian | e->elsz | e->mixed;
drivers/iommu/omap-iommu.c
329
static int load_iotlb_entry(struct omap_iommu *obj, struct iotlb_entry *e)
drivers/iommu/omap-iommu.c
335
if (!obj || !obj->nr_tlb_entries || !e)
drivers/iommu/omap-iommu.c
346
if (!e->prsvd) {
drivers/iommu/omap-iommu.c
366
cr = iotlb_alloc_cr(obj, e);
drivers/iommu/omap-iommu.c
375
if (e->prsvd)
drivers/iommu/omap-iommu.c
388
static int load_iotlb_entry(struct omap_iommu *obj, struct iotlb_entry *e)
drivers/iommu/omap-iommu.c
395
static int prefetch_iotlb_entry(struct omap_iommu *obj, struct iotlb_entry *e)
drivers/iommu/omap-iommu.c
397
return load_iotlb_entry(obj, e);
drivers/iommu/omap-iommu.c
621
iopgtable_store_entry_core(struct omap_iommu *obj, struct iotlb_entry *e)
drivers/iommu/omap-iommu.c
627
if (!obj || !e)
drivers/iommu/omap-iommu.c
630
switch (e->pgsz) {
drivers/iommu/omap-iommu.c
651
prot = get_iopte_attr(e);
drivers/iommu/omap-iommu.c
654
err = fn(obj, e->da, e->pa, prot);
drivers/iommu/omap-iommu.c
666
omap_iopgtable_store_entry(struct omap_iommu *obj, struct iotlb_entry *e)
drivers/iommu/omap-iommu.c
670
flush_iotlb_page(obj, e->da);
drivers/iommu/omap-iommu.c
671
err = iopgtable_store_entry_core(obj, e);
drivers/iommu/omap-iommu.c
673
prefetch_iotlb_entry(obj, e);
drivers/irqchip/irq-ti-sci-inta.c
115
#define to_vint_desc(e, i) container_of(e, struct ti_sci_inta_vint_desc, \
drivers/isdn/mISDN/dsp_audio.c
90
short mu, e, f, y;
drivers/isdn/mISDN/dsp_audio.c
94
e = (mu & 0x70) / 16;
drivers/isdn/mISDN/dsp_audio.c
96
y = f * (1 << (e + 3));
drivers/isdn/mISDN/dsp_audio.c
97
y += etab[e];
drivers/leds/flash/leds-aat1290.c
290
int i, b = 0, e = AAT1290_MM_CURRENT_SCALE_SIZE;
drivers/leds/flash/leds-aat1290.c
292
while (e - b > 1) {
drivers/leds/flash/leds-aat1290.c
293
i = b + (e - b) / 2;
drivers/leds/flash/leds-aat1290.c
295
e = i;
drivers/mailbox/arm_mhuv3.c
381
struct mhuv3_extension *e = mhu->ext[DBE_EXT];
drivers/mailbox/arm_mhuv3.c
389
spin_lock_irqsave(&e->pending_lock, flags);
drivers/mailbox/arm_mhuv3.c
390
e->pending_db[priv->ch_idx] = 0;
drivers/mailbox/arm_mhuv3.c
391
spin_unlock_irqrestore(&e->pending_lock, flags);
drivers/mailbox/arm_mhuv3.c
430
struct mhuv3_extension *e = mhu->ext[DBE_EXT];
drivers/mailbox/arm_mhuv3.c
434
spin_lock_irqsave(&e->pending_lock, flags);
drivers/mailbox/arm_mhuv3.c
435
e->pending_db[priv->ch_idx] &= ~BIT(priv->doorbell);
drivers/mailbox/arm_mhuv3.c
436
spin_unlock_irqrestore(&e->pending_lock, flags);
drivers/mailbox/arm_mhuv3.c
446
struct mhuv3_extension *e = mhu->ext[DBE_EXT];
drivers/mailbox/arm_mhuv3.c
448
scoped_guard(spinlock_irqsave, &e->pending_lock) {
drivers/mailbox/arm_mhuv3.c
450
if (e->pending_db[priv->ch_idx] & BIT(priv->doorbell))
drivers/mailbox/arm_mhuv3.c
453
e->pending_db[priv->ch_idx] |= BIT(priv->doorbell);
drivers/mailbox/arm_mhuv3.c
558
struct mhuv3_extension *e = mhu->ext[DBE_EXT];
drivers/mailbox/arm_mhuv3.c
562
if (channel >= e->num_chans || doorbell >= MHUV3_FLAG_BITS) {
drivers/mailbox/arm_mhuv3.c
568
return &chans[e->base_ch_idx + channel * MHUV3_FLAG_BITS + doorbell];
drivers/mailbox/arm_mhuv3.c
573
struct mhuv3_extension *e = mhu->ext[DBE_EXT];
drivers/mailbox/arm_mhuv3.c
579
for (i = 0; i < e->num_chans; i++) {
drivers/mailbox/arm_mhuv3.c
587
for (i = 0; i < e->num_chans; i++) {
drivers/mailbox/arm_mhuv3.c
597
struct mhuv3_extension *e = mhu->ext[DBE_EXT];
drivers/mailbox/arm_mhuv3.c
603
e->base_ch_idx = mbox->num_chans;
drivers/mailbox/arm_mhuv3.c
604
for (i = 0; i < e->num_chans; i++) {
drivers/mailbox/arm_mhuv3.c
621
spin_lock_init(&e->pending_lock);
drivers/mailbox/arm_mhuv3.c
629
struct mhuv3_extension *e = mhu->ext[DBE_EXT];
drivers/mailbox/arm_mhuv3.c
642
scoped_guard(spinlock_irqsave, &e->pending_lock) {
drivers/mailbox/arm_mhuv3.c
643
fired_dbs = e->pending_db[channel] & ~active_dbs;
drivers/mailbox/arm_mhuv3.c
648
e->pending_db[channel] &= ~BIT(*db);
drivers/mailbox/arm_mhuv3.c
675
struct mhuv3_extension *e = mhu->ext[DBE_EXT];
drivers/mailbox/arm_mhuv3.c
688
if (channel >= e->num_chans) {
drivers/mailbox/arm_mhuv3.c
709
struct mhuv3_extension *e;
drivers/mailbox/arm_mhuv3.c
716
e = devm_kzalloc(dev, sizeof(*e), GFP_KERNEL);
drivers/mailbox/arm_mhuv3.c
717
if (!e)
drivers/mailbox/arm_mhuv3.c
720
e->type = DBE_EXT;
drivers/mailbox/arm_mhuv3.c
722
e->num_chans =
drivers/mailbox/arm_mhuv3.c
724
e->mbox_of_xlate = mhuv3_dbe_mbox_of_xlate;
drivers/mailbox/arm_mhuv3.c
725
e->combined_irq_setup = mhuv3_dbe_combined_irq_setup;
drivers/mailbox/arm_mhuv3.c
726
e->channels_init = mhuv3_dbe_channels_init;
drivers/mailbox/arm_mhuv3.c
727
e->chan_from_comb_irq_get = mhuv3_dbe_chan_from_comb_irq_get;
drivers/mailbox/arm_mhuv3.c
729
mhu->num_chans += e->num_chans * MHUV3_FLAG_BITS;
drivers/mailbox/arm_mhuv3.c
730
mhu->ext[DBE_EXT] = e;
drivers/mailbox/arm_mhuv3.c
733
mhuv3_str[mhu->frame], e->num_chans);
drivers/md/bcache/util.c
27
char *e; \
drivers/md/bcache/util.c
28
type i = simple_ ## name(cp, &e, 10); \
drivers/md/bcache/util.c
30
switch (tolower(*e)) { \
drivers/md/bcache/util.c
54
if (e++ == cp) \
drivers/md/bcache/util.c
59
if (*e == '\n') \
drivers/md/bcache/util.c
60
e++; \
drivers/md/bcache/util.c
63
if (*e) \
drivers/md/dm-bufio.c
143
it->e = lru->cursor ? to_le(lru->cursor) : NULL;
drivers/md/dm-bufio.c
166
struct lru_entry *e;
drivers/md/dm-bufio.c
168
while (it->e) {
drivers/md/dm-bufio.c
169
e = it->e;
drivers/md/dm-bufio.c
172
if (it->e == it->stop)
drivers/md/dm-bufio.c
173
it->e = NULL;
drivers/md/dm-bufio.c
175
it->e = to_le(it->e->list.next);
drivers/md/dm-bufio.c
177
if (pred(e, context))
drivers/md/dm-bufio.c
178
return e;
drivers/md/dm-bufio.c
188
static void lru_iter_invalidate(struct lru *lru, struct lru_entry *e)
drivers/md/dm-bufio.c
194
if (it->e == e) {
drivers/md/dm-bufio.c
195
it->e = to_le(it->e->list.next);
drivers/md/dm-bufio.c
196
if (it->e == e)
drivers/md/dm-bufio.c
197
it->e = NULL;
drivers/md/dm-bufio.c
201
if (it->stop == e) {
drivers/md/dm-bufio.c
203
if (it->stop == e)
drivers/md/dm-bufio.c
2140
static bool is_writing(struct lru_entry *e, void *context)
drivers/md/dm-bufio.c
2142
struct dm_buffer *b = le_to_buffer(e);
drivers/md/dm-bufio.c
2151
struct lru_entry *e;
drivers/md/dm-bufio.c
2164
while ((e = lru_iter_next(&it, is_writing, c))) {
drivers/md/dm-bufio.c
2165
struct dm_buffer *b = le_to_buffer(e);
drivers/md/dm-bufio.c
81
struct lru_entry *e;
drivers/md/dm-cache-policy-smq.c
1000
static void requeue(struct smq_policy *mq, struct entry *e)
drivers/md/dm-cache-policy-smq.c
1005
if (e->pending_work)
drivers/md/dm-cache-policy-smq.c
1008
if (!test_and_set_bit(from_cblock(infer_cblock(mq, e)), mq->cache_hit_bits)) {
drivers/md/dm-cache-policy-smq.c
1009
if (!e->dirty) {
drivers/md/dm-cache-policy-smq.c
1010
q_requeue(&mq->clean, e, 1u, NULL, NULL);
drivers/md/dm-cache-policy-smq.c
1014
q_requeue(&mq->dirty, e, 1u,
drivers/md/dm-cache-policy-smq.c
1015
get_sentinel(&mq->writeback_sentinel_alloc, e->level, !mq->current_writeback_sentinels),
drivers/md/dm-cache-policy-smq.c
1016
get_sentinel(&mq->writeback_sentinel_alloc, e->level, mq->current_writeback_sentinels));
drivers/md/dm-cache-policy-smq.c
1171
static void mark_pending(struct smq_policy *mq, struct entry *e)
drivers/md/dm-cache-policy-smq.c
1173
BUG_ON(e->sentinel);
drivers/md/dm-cache-policy-smq.c
1174
BUG_ON(!e->allocated);
drivers/md/dm-cache-policy-smq.c
1175
BUG_ON(e->pending_work);
drivers/md/dm-cache-policy-smq.c
1176
e->pending_work = true;
drivers/md/dm-cache-policy-smq.c
1179
static void clear_pending(struct smq_policy *mq, struct entry *e)
drivers/md/dm-cache-policy-smq.c
1181
BUG_ON(!e->pending_work);
drivers/md/dm-cache-policy-smq.c
1182
e->pending_work = false;
drivers/md/dm-cache-policy-smq.c
1189
struct entry *e;
drivers/md/dm-cache-policy-smq.c
1191
e = q_peek(&mq->dirty, mq->dirty.nr_levels, idle);
drivers/md/dm-cache-policy-smq.c
1192
if (e) {
drivers/md/dm-cache-policy-smq.c
1193
mark_pending(mq, e);
drivers/md/dm-cache-policy-smq.c
1194
q_del(&mq->dirty, e);
drivers/md/dm-cache-policy-smq.c
1197
work.oblock = e->oblock;
drivers/md/dm-cache-policy-smq.c
1198
work.cblock = infer_cblock(mq, e);
drivers/md/dm-cache-policy-smq.c
1202
clear_pending(mq, e);
drivers/md/dm-cache-policy-smq.c
1203
q_push_front(&mq->dirty, e);
drivers/md/dm-cache-policy-smq.c
1212
struct entry *e;
drivers/md/dm-cache-policy-smq.c
1217
e = q_peek(&mq->clean, mq->clean.nr_levels / 2, true);
drivers/md/dm-cache-policy-smq.c
1218
if (!e) {
drivers/md/dm-cache-policy-smq.c
1224
mark_pending(mq, e);
drivers/md/dm-cache-policy-smq.c
1225
q_del(&mq->clean, e);
drivers/md/dm-cache-policy-smq.c
1228
work.oblock = e->oblock;
drivers/md/dm-cache-policy-smq.c
1229
work.cblock = infer_cblock(mq, e);
drivers/md/dm-cache-policy-smq.c
1232
clear_pending(mq, e);
drivers/md/dm-cache-policy-smq.c
1233
q_push_front(&mq->clean, e);
drivers/md/dm-cache-policy-smq.c
1241
struct entry *e;
drivers/md/dm-cache-policy-smq.c
1264
e = alloc_entry(&mq->cache_alloc);
drivers/md/dm-cache-policy-smq.c
1265
BUG_ON(!e);
drivers/md/dm-cache-policy-smq.c
1266
e->pending_work = true;
drivers/md/dm-cache-policy-smq.c
1269
work.cblock = infer_cblock(mq, e);
drivers/md/dm-cache-policy-smq.c
1272
free_entry(&mq->cache_alloc, e);
drivers/md/dm-cache-policy-smq.c
1314
struct entry *e = h_lookup(&mq->hotspot_table, hb);
drivers/md/dm-cache-policy-smq.c
1316
if (e) {
drivers/md/dm-cache-policy-smq.c
1317
stats_level_accessed(&mq->hotspot_stats, e->level);
drivers/md/dm-cache-policy-smq.c
1319
hi = get_index(&mq->hotspot_alloc, e);
drivers/md/dm-cache-policy-smq.c
1320
q_requeue(&mq->hotspot, e,
drivers/md/dm-cache-policy-smq.c
1328
e = alloc_entry(&mq->hotspot_alloc);
drivers/md/dm-cache-policy-smq.c
1329
if (!e) {
drivers/md/dm-cache-policy-smq.c
133
static struct entry *l_next(struct entry_space *es, struct entry *e)
drivers/md/dm-cache-policy-smq.c
1330
e = q_pop(&mq->hotspot);
drivers/md/dm-cache-policy-smq.c
1331
if (e) {
drivers/md/dm-cache-policy-smq.c
1332
h_remove(&mq->hotspot_table, e);
drivers/md/dm-cache-policy-smq.c
1333
hi = get_index(&mq->hotspot_alloc, e);
drivers/md/dm-cache-policy-smq.c
1339
if (e) {
drivers/md/dm-cache-policy-smq.c
1340
e->oblock = hb;
drivers/md/dm-cache-policy-smq.c
1341
q_push(&mq->hotspot, e);
drivers/md/dm-cache-policy-smq.c
1342
h_insert(&mq->hotspot_table, e);
drivers/md/dm-cache-policy-smq.c
1346
return e;
drivers/md/dm-cache-policy-smq.c
135
return to_entry(es, e->next);
drivers/md/dm-cache-policy-smq.c
138
static struct entry *l_prev(struct entry_space *es, struct entry *e)
drivers/md/dm-cache-policy-smq.c
1380
struct entry *e, *hs_e;
drivers/md/dm-cache-policy-smq.c
1385
e = h_lookup(&mq->table, oblock);
drivers/md/dm-cache-policy-smq.c
1386
if (e) {
drivers/md/dm-cache-policy-smq.c
1387
stats_level_accessed(&mq->cache_stats, e->level);
drivers/md/dm-cache-policy-smq.c
1389
requeue(mq, e);
drivers/md/dm-cache-policy-smq.c
1390
*cblock = infer_cblock(mq, e);
drivers/md/dm-cache-policy-smq.c
140
return to_entry(es, e->prev);
drivers/md/dm-cache-policy-smq.c
1473
struct entry *e = get_entry(&mq->cache_alloc,
drivers/md/dm-cache-policy-smq.c
1479
clear_pending(mq, e);
drivers/md/dm-cache-policy-smq.c
148
static void l_add_head(struct entry_space *es, struct ilist *l, struct entry *e)
drivers/md/dm-cache-policy-smq.c
1481
e->oblock = work->oblock;
drivers/md/dm-cache-policy-smq.c
1482
e->level = NR_CACHE_LEVELS - 1;
drivers/md/dm-cache-policy-smq.c
1483
push(mq, e);
drivers/md/dm-cache-policy-smq.c
1486
free_entry(&mq->cache_alloc, e);
drivers/md/dm-cache-policy-smq.c
1494
h_remove(&mq->table, e);
drivers/md/dm-cache-policy-smq.c
1495
free_entry(&mq->cache_alloc, e);
drivers/md/dm-cache-policy-smq.c
1498
clear_pending(mq, e);
drivers/md/dm-cache-policy-smq.c
1499
push_queue(mq, e);
drivers/md/dm-cache-policy-smq.c
1506
clear_pending(mq, e);
drivers/md/dm-cache-policy-smq.c
1507
push_queue(mq, e);
drivers/md/dm-cache-policy-smq.c
152
e->next = l->head;
drivers/md/dm-cache-policy-smq.c
153
e->prev = INDEXER_NULL;
drivers/md/dm-cache-policy-smq.c
1530
struct entry *e = get_entry(&mq->cache_alloc, from_cblock(cblock));
drivers/md/dm-cache-policy-smq.c
1532
if (e->pending_work)
drivers/md/dm-cache-policy-smq.c
1533
e->dirty = set;
drivers/md/dm-cache-policy-smq.c
1535
del_queue(mq, e);
drivers/md/dm-cache-policy-smq.c
1536
e->dirty = set;
drivers/md/dm-cache-policy-smq.c
1537
push_queue(mq, e);
drivers/md/dm-cache-policy-smq.c
156
head->prev = l->head = to_index(es, e);
drivers/md/dm-cache-policy-smq.c
1571
struct entry *e;
drivers/md/dm-cache-policy-smq.c
1573
e = alloc_particular_entry(&mq->cache_alloc, from_cblock(cblock));
drivers/md/dm-cache-policy-smq.c
1574
e->oblock = oblock;
drivers/md/dm-cache-policy-smq.c
1575
e->dirty = dirty;
drivers/md/dm-cache-policy-smq.c
1576
e->level = hint_valid ? min(hint, NR_CACHE_LEVELS - 1) : random_level(cblock);
drivers/md/dm-cache-policy-smq.c
1577
e->pending_work = false;
drivers/md/dm-cache-policy-smq.c
158
l->head = l->tail = to_index(es, e);
drivers/md/dm-cache-policy-smq.c
1583
push_front(mq, e);
drivers/md/dm-cache-policy-smq.c
1591
struct entry *e = get_entry(&mq->cache_alloc, from_cblock(cblock));
drivers/md/dm-cache-policy-smq.c
1593
if (!e->allocated)
drivers/md/dm-cache-policy-smq.c
1597
del_queue(mq, e);
drivers/md/dm-cache-policy-smq.c
1598
h_remove(&mq->table, e);
drivers/md/dm-cache-policy-smq.c
1599
free_entry(&mq->cache_alloc, e);
drivers/md/dm-cache-policy-smq.c
160
if (!e->sentinel)
drivers/md/dm-cache-policy-smq.c
1606
struct entry *e = get_entry(&mq->cache_alloc, from_cblock(cblock));
drivers/md/dm-cache-policy-smq.c
1608
if (!e->allocated)
drivers/md/dm-cache-policy-smq.c
1611
return e->level;
drivers/md/dm-cache-policy-smq.c
164
static void l_add_tail(struct entry_space *es, struct ilist *l, struct entry *e)
drivers/md/dm-cache-policy-smq.c
168
e->next = INDEXER_NULL;
drivers/md/dm-cache-policy-smq.c
169
e->prev = l->tail;
drivers/md/dm-cache-policy-smq.c
172
tail->next = l->tail = to_index(es, e);
drivers/md/dm-cache-policy-smq.c
174
l->head = l->tail = to_index(es, e);
drivers/md/dm-cache-policy-smq.c
176
if (!e->sentinel)
drivers/md/dm-cache-policy-smq.c
181
struct entry *old, struct entry *e)
drivers/md/dm-cache-policy-smq.c
186
l_add_head(es, l, e);
drivers/md/dm-cache-policy-smq.c
189
e->prev = old->prev;
drivers/md/dm-cache-policy-smq.c
190
e->next = to_index(es, old);
drivers/md/dm-cache-policy-smq.c
191
prev->next = old->prev = to_index(es, e);
drivers/md/dm-cache-policy-smq.c
193
if (!e->sentinel)
drivers/md/dm-cache-policy-smq.c
198
static void l_del(struct entry_space *es, struct ilist *l, struct entry *e)
drivers/md/dm-cache-policy-smq.c
200
struct entry *prev = l_prev(es, e);
drivers/md/dm-cache-policy-smq.c
201
struct entry *next = l_next(es, e);
drivers/md/dm-cache-policy-smq.c
204
prev->next = e->next;
drivers/md/dm-cache-policy-smq.c
206
l->head = e->next;
drivers/md/dm-cache-policy-smq.c
209
next->prev = e->prev;
drivers/md/dm-cache-policy-smq.c
211
l->tail = e->prev;
drivers/md/dm-cache-policy-smq.c
213
if (!e->sentinel)
drivers/md/dm-cache-policy-smq.c
219
struct entry *e;
drivers/md/dm-cache-policy-smq.c
221
for (e = l_head(es, l); e; e = l_next(es, e))
drivers/md/dm-cache-policy-smq.c
222
if (!e->sentinel) {
drivers/md/dm-cache-policy-smq.c
223
l_del(es, l, e);
drivers/md/dm-cache-policy-smq.c
224
return e;
drivers/md/dm-cache-policy-smq.c
232
struct entry *e;
drivers/md/dm-cache-policy-smq.c
234
for (e = l_tail(es, l); e; e = l_prev(es, e))
drivers/md/dm-cache-policy-smq.c
235
if (!e->sentinel) {
drivers/md/dm-cache-policy-smq.c
236
l_del(es, l, e);
drivers/md/dm-cache-policy-smq.c
237
return e;
drivers/md/dm-cache-policy-smq.c
296
static void q_push(struct queue *q, struct entry *e)
drivers/md/dm-cache-policy-smq.c
298
BUG_ON(e->pending_work);
drivers/md/dm-cache-policy-smq.c
300
if (!e->sentinel)
drivers/md/dm-cache-policy-smq.c
303
l_add_tail(q->es, q->qs + e->level, e);
drivers/md/dm-cache-policy-smq.c
306
static void q_push_front(struct queue *q, struct entry *e)
drivers/md/dm-cache-policy-smq.c
308
BUG_ON(e->pending_work);
drivers/md/dm-cache-policy-smq.c
310
if (!e->sentinel)
drivers/md/dm-cache-policy-smq.c
313
l_add_head(q->es, q->qs + e->level, e);
drivers/md/dm-cache-policy-smq.c
316
static void q_push_before(struct queue *q, struct entry *old, struct entry *e)
drivers/md/dm-cache-policy-smq.c
318
BUG_ON(e->pending_work);
drivers/md/dm-cache-policy-smq.c
320
if (!e->sentinel)
drivers/md/dm-cache-policy-smq.c
323
l_add_before(q->es, q->qs + e->level, old, e);
drivers/md/dm-cache-policy-smq.c
326
static void q_del(struct queue *q, struct entry *e)
drivers/md/dm-cache-policy-smq.c
328
l_del(q->es, q->qs + e->level, e);
drivers/md/dm-cache-policy-smq.c
329
if (!e->sentinel)
drivers/md/dm-cache-policy-smq.c
339
struct entry *e;
drivers/md/dm-cache-policy-smq.c
344
for (e = l_head(q->es, q->qs + level); e; e = l_next(q->es, e)) {
drivers/md/dm-cache-policy-smq.c
345
if (e->sentinel) {
drivers/md/dm-cache-policy-smq.c
352
return e;
drivers/md/dm-cache-policy-smq.c
360
struct entry *e = q_peek(q, q->nr_levels, true);
drivers/md/dm-cache-policy-smq.c
362
if (e)
drivers/md/dm-cache-policy-smq.c
363
q_del(q, e);
drivers/md/dm-cache-policy-smq.c
365
return e;
drivers/md/dm-cache-policy-smq.c
375
struct entry *e;
drivers/md/dm-cache-policy-smq.c
378
for (e = l_head(q->es, q->qs + level); e; e = l_next(q->es, e))
drivers/md/dm-cache-policy-smq.c
379
if (!e->sentinel) {
drivers/md/dm-cache-policy-smq.c
380
l_del(q->es, q->qs + e->level, e);
drivers/md/dm-cache-policy-smq.c
381
return e;
drivers/md/dm-cache-policy-smq.c
433
struct entry *e;
drivers/md/dm-cache-policy-smq.c
445
e = __redist_pop_from(q, level + 1u);
drivers/md/dm-cache-policy-smq.c
446
if (!e) {
drivers/md/dm-cache-policy-smq.c
451
e->level = level;
drivers/md/dm-cache-policy-smq.c
452
l_add_tail(q->es, l, e);
drivers/md/dm-cache-policy-smq.c
460
e = l_pop_tail(q->es, l);
drivers/md/dm-cache-policy-smq.c
462
if (!e)
drivers/md/dm-cache-policy-smq.c
466
e->level = level + 1u;
drivers/md/dm-cache-policy-smq.c
467
l_add_tail(q->es, l_above, e);
drivers/md/dm-cache-policy-smq.c
472
static void q_requeue(struct queue *q, struct entry *e, unsigned int extra_levels,
drivers/md/dm-cache-policy-smq.c
477
unsigned int new_level = min(q->nr_levels - 1u, e->level + extra_levels);
drivers/md/dm-cache-policy-smq.c
480
if (extra_levels && (e->level < q->nr_levels - 1u)) {
drivers/md/dm-cache-policy-smq.c
486
de->level = e->level;
drivers/md/dm-cache-policy-smq.c
505
q_del(q, e);
drivers/md/dm-cache-policy-smq.c
506
e->level = new_level;
drivers/md/dm-cache-policy-smq.c
507
q_push(q, e);
drivers/md/dm-cache-policy-smq.c
613
static struct entry *h_next(struct smq_hash_table *ht, struct entry *e)
drivers/md/dm-cache-policy-smq.c
615
return to_entry(ht->es, e->hash_next);
drivers/md/dm-cache-policy-smq.c
618
static void __h_insert(struct smq_hash_table *ht, unsigned int bucket, struct entry *e)
drivers/md/dm-cache-policy-smq.c
620
e->hash_next = ht->buckets[bucket];
drivers/md/dm-cache-policy-smq.c
621
ht->buckets[bucket] = to_index(ht->es, e);
drivers/md/dm-cache-policy-smq.c
624
static void h_insert(struct smq_hash_table *ht, struct entry *e)
drivers/md/dm-cache-policy-smq.c
626
unsigned int h = hash_64(from_oblock(e->oblock), ht->hash_bits);
drivers/md/dm-cache-policy-smq.c
628
__h_insert(ht, h, e);
drivers/md/dm-cache-policy-smq.c
634
struct entry *e;
drivers/md/dm-cache-policy-smq.c
637
for (e = h_head(ht, h); e; e = h_next(ht, e)) {
drivers/md/dm-cache-policy-smq.c
638
if (e->oblock == oblock)
drivers/md/dm-cache-policy-smq.c
639
return e;
drivers/md/dm-cache-policy-smq.c
641
*prev = e;
drivers/md/dm-cache-policy-smq.c
648
struct entry *e, struct entry *prev)
drivers/md/dm-cache-policy-smq.c
651
prev->hash_next = e->hash_next;
drivers/md/dm-cache-policy-smq.c
653
ht->buckets[h] = e->hash_next;
drivers/md/dm-cache-policy-smq.c
661
struct entry *e, *prev;
drivers/md/dm-cache-policy-smq.c
664
e = __h_lookup(ht, h, oblock, &prev);
drivers/md/dm-cache-policy-smq.c
665
if (e && prev) {
drivers/md/dm-cache-policy-smq.c
670
__h_unlink(ht, h, e, prev);
drivers/md/dm-cache-policy-smq.c
671
__h_insert(ht, h, e);
drivers/md/dm-cache-policy-smq.c
674
return e;
drivers/md/dm-cache-policy-smq.c
677
static void h_remove(struct smq_hash_table *ht, struct entry *e)
drivers/md/dm-cache-policy-smq.c
679
unsigned int h = hash_64(from_oblock(e->oblock), ht->hash_bits);
drivers/md/dm-cache-policy-smq.c
686
e = __h_lookup(ht, h, e->oblock, &prev);
drivers/md/dm-cache-policy-smq.c
687
if (e)
drivers/md/dm-cache-policy-smq.c
688
__h_unlink(ht, h, e, prev);
drivers/md/dm-cache-policy-smq.c
715
static void init_entry(struct entry *e)
drivers/md/dm-cache-policy-smq.c
721
e->hash_next = INDEXER_NULL;
drivers/md/dm-cache-policy-smq.c
722
e->next = INDEXER_NULL;
drivers/md/dm-cache-policy-smq.c
723
e->prev = INDEXER_NULL;
drivers/md/dm-cache-policy-smq.c
724
e->level = 0u;
drivers/md/dm-cache-policy-smq.c
725
e->dirty = true; /* FIXME: audit */
drivers/md/dm-cache-policy-smq.c
726
e->allocated = true;
drivers/md/dm-cache-policy-smq.c
727
e->sentinel = false;
drivers/md/dm-cache-policy-smq.c
728
e->pending_work = false;
drivers/md/dm-cache-policy-smq.c
733
struct entry *e;
drivers/md/dm-cache-policy-smq.c
738
e = l_pop_head(ea->es, &ea->free);
drivers/md/dm-cache-policy-smq.c
739
init_entry(e);
drivers/md/dm-cache-policy-smq.c
742
return e;
drivers/md/dm-cache-policy-smq.c
750
struct entry *e = __get_entry(ea->es, ea->begin + i);
drivers/md/dm-cache-policy-smq.c
752
BUG_ON(e->allocated);
drivers/md/dm-cache-policy-smq.c
754
l_del(ea->es, &ea->free, e);
drivers/md/dm-cache-policy-smq.c
755
init_entry(e);
drivers/md/dm-cache-policy-smq.c
758
return e;
drivers/md/dm-cache-policy-smq.c
761
static void free_entry(struct entry_alloc *ea, struct entry *e)
drivers/md/dm-cache-policy-smq.c
764
BUG_ON(!e->allocated);
drivers/md/dm-cache-policy-smq.c
767
e->allocated = false;
drivers/md/dm-cache-policy-smq.c
768
l_add_tail(ea->es, &ea->free, e);
drivers/md/dm-cache-policy-smq.c
776
static unsigned int get_index(struct entry_alloc *ea, struct entry *e)
drivers/md/dm-cache-policy-smq.c
778
return to_index(ea->es, e) - ea->begin;
drivers/md/dm-cache-policy-smq.c
88
struct entry *e;
drivers/md/dm-cache-policy-smq.c
90
e = es->begin + block;
drivers/md/dm-cache-policy-smq.c
91
BUG_ON(e >= es->end);
drivers/md/dm-cache-policy-smq.c
93
return e;
drivers/md/dm-cache-policy-smq.c
959
static void del_queue(struct smq_policy *mq, struct entry *e)
drivers/md/dm-cache-policy-smq.c
96
static unsigned int to_index(struct entry_space *es, struct entry *e)
drivers/md/dm-cache-policy-smq.c
961
q_del(e->dirty ? &mq->dirty : &mq->clean, e);
drivers/md/dm-cache-policy-smq.c
964
static void push_queue(struct smq_policy *mq, struct entry *e)
drivers/md/dm-cache-policy-smq.c
966
if (e->dirty)
drivers/md/dm-cache-policy-smq.c
967
q_push(&mq->dirty, e);
drivers/md/dm-cache-policy-smq.c
969
q_push(&mq->clean, e);
drivers/md/dm-cache-policy-smq.c
973
static void push(struct smq_policy *mq, struct entry *e)
drivers/md/dm-cache-policy-smq.c
975
h_insert(&mq->table, e);
drivers/md/dm-cache-policy-smq.c
976
if (!e->pending_work)
drivers/md/dm-cache-policy-smq.c
977
push_queue(mq, e);
drivers/md/dm-cache-policy-smq.c
98
BUG_ON(e < es->begin || e >= es->end);
drivers/md/dm-cache-policy-smq.c
980
static void push_queue_front(struct smq_policy *mq, struct entry *e)
drivers/md/dm-cache-policy-smq.c
982
if (e->dirty)
drivers/md/dm-cache-policy-smq.c
983
q_push_front(&mq->dirty, e);
drivers/md/dm-cache-policy-smq.c
985
q_push_front(&mq->clean, e);
drivers/md/dm-cache-policy-smq.c
988
static void push_front(struct smq_policy *mq, struct entry *e)
drivers/md/dm-cache-policy-smq.c
99
return e - es->begin;
drivers/md/dm-cache-policy-smq.c
990
h_insert(&mq->table, e);
drivers/md/dm-cache-policy-smq.c
991
if (!e->pending_work)
drivers/md/dm-cache-policy-smq.c
992
push_queue_front(mq, e);
drivers/md/dm-cache-policy-smq.c
995
static dm_cblock_t infer_cblock(struct smq_policy *mq, struct entry *e)
drivers/md/dm-cache-policy-smq.c
997
return to_cblock(get_index(&mq->cache_alloc, e));
drivers/md/dm-cache-target.c
1015
dm_dblock_t *b, dm_dblock_t *e)
drivers/md/dm-cache-target.c
1023
*e = *b;
drivers/md/dm-cache-target.c
1025
*e = to_dblock(block_div(se, cache->discard_block_size));
drivers/md/dm-cache-target.c
1791
dm_dblock_t b, e;
drivers/md/dm-cache-target.c
1798
calc_discard_block_range(cache, bio, &b, &e);
drivers/md/dm-cache-target.c
1799
while (b != e) {
drivers/md/dm-cache-target.c
2865
sector_t b, e;
drivers/md/dm-cache-target.c
2874
e = li->discard_end * li->block_size;
drivers/md/dm-cache-target.c
2880
sector_div(e, li->cache->discard_block_size);
drivers/md/dm-cache-target.c
2886
if (e > from_dblock(li->cache->discard_nr_blocks))
drivers/md/dm-cache-target.c
2887
e = from_dblock(li->cache->discard_nr_blocks);
drivers/md/dm-cache-target.c
2889
for (; b < e; b++)
drivers/md/dm-cache-target.c
3282
uint64_t b, e;
drivers/md/dm-cache-target.c
3288
r = sscanf(str, "%llu-%llu%c", &b, &e, &dummy);
drivers/md/dm-cache-target.c
3292
result->end = to_cblock(e);
drivers/md/dm-cache-target.c
3314
uint64_t e = from_cblock(range->end);
drivers/md/dm-cache-target.c
3323
if (e > n) {
drivers/md/dm-cache-target.c
3325
cache_device_name(cache), e, n);
drivers/md/dm-cache-target.c
3329
if (b >= e) {
drivers/md/dm-cache-target.c
3331
cache_device_name(cache), b, e);
drivers/md/dm-era-target.c
718
unsigned int b, e = min(d->current_bit + INSERTS_PER_STEP, d->nr_bits);
drivers/md/dm-era-target.c
720
for (b = d->current_bit; b < e; b++) {
drivers/md/dm-exception-store.h
148
static inline unsigned int dm_consecutive_chunk_count(struct dm_exception *e)
drivers/md/dm-exception-store.h
150
return e->new_chunk >> DM_CHUNK_NUMBER_BITS;
drivers/md/dm-exception-store.h
153
static inline void dm_consecutive_chunk_count_inc(struct dm_exception *e)
drivers/md/dm-exception-store.h
155
e->new_chunk += (1ULL << DM_CHUNK_NUMBER_BITS);
drivers/md/dm-exception-store.h
157
BUG_ON(!dm_consecutive_chunk_count(e));
drivers/md/dm-exception-store.h
160
static inline void dm_consecutive_chunk_count_dec(struct dm_exception *e)
drivers/md/dm-exception-store.h
162
BUG_ON(!dm_consecutive_chunk_count(e));
drivers/md/dm-exception-store.h
164
e->new_chunk -= (1ULL << DM_CHUNK_NUMBER_BITS);
drivers/md/dm-exception-store.h
68
struct dm_exception *e);
drivers/md/dm-exception-store.h
74
struct dm_exception *e, int valid,
drivers/md/dm-integrity.c
714
bool e, const char *function)
drivers/md/dm-integrity.c
717
unsigned int limit = e ? ic->journal_section_entries : ic->journal_section_sectors;
drivers/md/dm-snap-persistent.c
420
uint32_t index, struct core_exception *e)
drivers/md/dm-snap-persistent.c
425
de->old_chunk = cpu_to_le64(e->old_chunk);
drivers/md/dm-snap-persistent.c
426
de->new_chunk = cpu_to_le64(e->new_chunk);
drivers/md/dm-snap-persistent.c
451
struct core_exception e;
drivers/md/dm-snap-persistent.c
457
read_exception(ps, ps_area, i, &e);
drivers/md/dm-snap-persistent.c
465
if (e.new_chunk == 0LL) {
drivers/md/dm-snap-persistent.c
474
if (ps->next_free <= e.new_chunk)
drivers/md/dm-snap-persistent.c
475
ps->next_free = e.new_chunk + 1;
drivers/md/dm-snap-persistent.c
480
r = callback(callback_context, e.old_chunk, e.new_chunk);
drivers/md/dm-snap-persistent.c
674
struct dm_exception *e)
drivers/md/dm-snap-persistent.c
683
e->new_chunk = ps->next_free;
drivers/md/dm-snap-persistent.c
697
struct dm_exception *e, int valid,
drivers/md/dm-snap-persistent.c
709
ce.old_chunk = e->old_chunk;
drivers/md/dm-snap-persistent.c
710
ce.new_chunk = e->new_chunk;
drivers/md/dm-snap-transient.c
43
struct dm_exception *e)
drivers/md/dm-snap-transient.c
51
e->new_chunk = sector_to_chunk(store, tc->next_free);
drivers/md/dm-snap-transient.c
58
struct dm_exception *e, int valid,
drivers/md/dm-snap.c
1637
struct dm_exception *e;
drivers/md/dm-snap.c
1645
dm_exception_table_lock_init(s, pe->e.old_chunk, &lock);
drivers/md/dm-snap.c
1656
e = alloc_completed_exception(GFP_NOIO);
drivers/md/dm-snap.c
1657
if (!e) {
drivers/md/dm-snap.c
1664
*e = pe->e;
drivers/md/dm-snap.c
1670
free_completed_exception(e);
drivers/md/dm-snap.c
1683
dm_insert_exception(&s->complete, e);
drivers/md/dm-snap.c
1687
if (__chunk_is_tracked(s, pe->e.old_chunk)) {
drivers/md/dm-snap.c
1689
__check_for_conflicting_io(s, pe->e.old_chunk);
drivers/md/dm-snap.c
1695
dm_remove_exception(&pe->e);
drivers/md/dm-snap.c
1727
s->store->type->commit_exception(s->store, &pe->e, !pe->copy_error,
drivers/md/dm-snap.c
1795
src.sector = chunk_to_sector(s->store, pe->e.old_chunk);
drivers/md/dm-snap.c
1799
dest.sector = chunk_to_sector(s->store, pe->e.new_chunk);
drivers/md/dm-snap.c
1836
struct dm_exception *e = dm_lookup_exception(&s->pending, chunk);
drivers/md/dm-snap.c
1838
if (!e)
drivers/md/dm-snap.c
1841
return container_of(e, struct dm_snap_pending_exception, e);
drivers/md/dm-snap.c
1854
pe->e.old_chunk = chunk;
drivers/md/dm-snap.c
1861
if (s->store->type->prepare_exception(s->store, &pe->e)) {
drivers/md/dm-snap.c
1870
dm_insert_exception(&s->pending, &pe->e);
drivers/md/dm-snap.c
1898
static void remap_exception(struct dm_snapshot *s, struct dm_exception *e,
drivers/md/dm-snap.c
1903
chunk_to_sector(s->store, dm_chunk_number(e->new_chunk) +
drivers/md/dm-snap.c
1904
(chunk - e->old_chunk)) +
drivers/md/dm-snap.c
1918
static void zero_exception(struct dm_snapshot *s, struct dm_exception *e,
drivers/md/dm-snap.c
1941
struct dm_exception *e;
drivers/md/dm-snap.c
1993
e = dm_lookup_exception(&s->complete, chunk);
drivers/md/dm-snap.c
1994
if (e) {
drivers/md/dm-snap.c
1995
remap_exception(s, e, bio, chunk);
drivers/md/dm-snap.c
2000
zero_exception(s, e, bio, chunk);
drivers/md/dm-snap.c
2029
e = dm_lookup_exception(&s->complete, chunk);
drivers/md/dm-snap.c
2030
if (e) {
drivers/md/dm-snap.c
2032
remap_exception(s, e, bio, chunk);
drivers/md/dm-snap.c
205
struct dm_exception e;
drivers/md/dm-snap.c
2057
remap_exception(s, &pe->e, bio, chunk);
drivers/md/dm-snap.c
2109
struct dm_exception *e;
drivers/md/dm-snap.c
2139
e = dm_lookup_exception(&s->complete, chunk);
drivers/md/dm-snap.c
2140
if (e) {
drivers/md/dm-snap.c
2152
remap_exception(s, e, bio, chunk);
drivers/md/dm-snap.c
2430
struct dm_exception *e;
drivers/md/dm-snap.c
2471
e = dm_lookup_exception(&snap->complete, chunk);
drivers/md/dm-snap.c
2472
if (e)
drivers/md/dm-snap.c
2482
e = dm_lookup_exception(&snap->complete, chunk);
drivers/md/dm-snap.c
2483
if (e) {
drivers/md/dm-snap.c
705
static void dm_remove_exception(struct dm_exception *e)
drivers/md/dm-snap.c
707
hlist_del(&e->hash_list);
drivers/md/dm-snap.c
718
struct dm_exception *e;
drivers/md/dm-snap.c
721
hlist_for_each_entry(e, slot, hash_list)
drivers/md/dm-snap.c
722
if (chunk >= e->old_chunk &&
drivers/md/dm-snap.c
723
chunk <= e->old_chunk + dm_consecutive_chunk_count(e))
drivers/md/dm-snap.c
724
return e;
drivers/md/dm-snap.c
731
struct dm_exception *e;
drivers/md/dm-snap.c
733
e = kmem_cache_alloc(exception_cache, gfp);
drivers/md/dm-snap.c
734
if (!e && gfp == GFP_NOIO)
drivers/md/dm-snap.c
735
e = kmem_cache_alloc(exception_cache, GFP_ATOMIC);
drivers/md/dm-snap.c
737
return e;
drivers/md/dm-snap.c
740
static void free_completed_exception(struct dm_exception *e)
drivers/md/dm-snap.c
742
kmem_cache_free(exception_cache, e);
drivers/md/dm-snap.c
769
struct dm_exception *e = NULL;
drivers/md/dm-snap.c
778
hlist_for_each_entry(e, l, hash_list) {
drivers/md/dm-snap.c
780
if (new_e->old_chunk == (e->old_chunk +
drivers/md/dm-snap.c
781
dm_consecutive_chunk_count(e) + 1) &&
drivers/md/dm-snap.c
782
new_e->new_chunk == (dm_chunk_number(e->new_chunk) +
drivers/md/dm-snap.c
783
dm_consecutive_chunk_count(e) + 1)) {
drivers/md/dm-snap.c
784
dm_consecutive_chunk_count_inc(e);
drivers/md/dm-snap.c
790
if (new_e->old_chunk == (e->old_chunk - 1) &&
drivers/md/dm-snap.c
791
new_e->new_chunk == (dm_chunk_number(e->new_chunk) - 1)) {
drivers/md/dm-snap.c
792
dm_consecutive_chunk_count_inc(e);
drivers/md/dm-snap.c
793
e->old_chunk--;
drivers/md/dm-snap.c
794
e->new_chunk--;
drivers/md/dm-snap.c
799
if (new_e->old_chunk < e->old_chunk)
drivers/md/dm-snap.c
804
if (!e) {
drivers/md/dm-snap.c
810
} else if (new_e->old_chunk < e->old_chunk) {
drivers/md/dm-snap.c
812
hlist_add_before(&new_e->hash_list, &e->hash_list);
drivers/md/dm-snap.c
815
hlist_add_behind(&new_e->hash_list, &e->hash_list);
drivers/md/dm-snap.c
826
struct dm_exception *e;
drivers/md/dm-snap.c
828
e = alloc_completed_exception(GFP_KERNEL);
drivers/md/dm-snap.c
829
if (!e)
drivers/md/dm-snap.c
832
e->old_chunk = old;
drivers/md/dm-snap.c
835
e->new_chunk = new;
drivers/md/dm-snap.c
837
dm_insert_exception(&s->complete, e);
drivers/md/dm-snap.c
933
struct dm_exception *e;
drivers/md/dm-snap.c
935
e = dm_lookup_exception(&s->complete, old_chunk);
drivers/md/dm-snap.c
936
if (!e) {
drivers/md/dm-snap.c
945
if (!dm_consecutive_chunk_count(e)) {
drivers/md/dm-snap.c
946
dm_remove_exception(e);
drivers/md/dm-snap.c
947
free_completed_exception(e);
drivers/md/dm-snap.c
959
if (old_chunk == e->old_chunk) {
drivers/md/dm-snap.c
960
e->old_chunk++;
drivers/md/dm-snap.c
961
e->new_chunk++;
drivers/md/dm-snap.c
962
} else if (old_chunk != e->old_chunk +
drivers/md/dm-snap.c
963
dm_consecutive_chunk_count(e)) {
drivers/md/dm-snap.c
966
(unsigned long long)e->old_chunk,
drivers/md/dm-snap.c
968
e->old_chunk + dm_consecutive_chunk_count(e));
drivers/md/dm-snap.c
972
dm_consecutive_chunk_count_dec(e);
drivers/md/dm-thin-metadata.c
1789
int dm_pool_inc_data_range(struct dm_pool_metadata *pmd, dm_block_t b, dm_block_t e)
drivers/md/dm-thin-metadata.c
1795
r = dm_sm_inc_blocks(pmd->data_sm, b, e);
drivers/md/dm-thin-metadata.c
1801
int dm_pool_dec_data_range(struct dm_pool_metadata *pmd, dm_block_t b, dm_block_t e)
drivers/md/dm-thin-metadata.c
1807
r = dm_sm_dec_blocks(pmd->data_sm, b, e);
drivers/md/dm-thin-metadata.h
200
int dm_pool_inc_data_range(struct dm_pool_metadata *pmd, dm_block_t b, dm_block_t e);
drivers/md/dm-thin-metadata.h
201
int dm_pool_dec_data_range(struct dm_pool_metadata *pmd, dm_block_t b, dm_block_t e);
drivers/md/dm-thin.c
1088
dm_block_t b = m->data_block, e, end = m->data_block + m->virt_end - m->virt_begin;
drivers/md/dm-thin.c
1107
for (e = b + 1; e != end; e++) {
drivers/md/dm-thin.c
1108
r = dm_pool_block_is_shared(pool->pmd, e, &shared);
drivers/md/dm-thin.c
1116
issue_discard(&op, b, e);
drivers/md/dm-thin.c
1118
b = e;
drivers/md/dm-thin.c
122
dm_block_t b, dm_block_t e, struct dm_cell_key *key)
drivers/md/dm-thin.c
127
key->block_end = e;
drivers/md/dm-thin.c
693
sector_t e = b + (bio->bi_iter.bi_size >> SECTOR_SHIFT);
drivers/md/dm-thin.c
699
e >>= pool->sectors_per_block_shift;
drivers/md/dm-thin.c
702
(void) sector_div(e, pool->sectors_per_block);
drivers/md/dm-thin.c
705
if (e < b) {
drivers/md/dm-thin.c
707
e = b;
drivers/md/dm-thin.c
711
*end = e;
drivers/md/dm-writecache.c
1041
struct wc_entry *e = &wc->entries[b];
drivers/md/dm-writecache.c
1045
e->original_sector = -1;
drivers/md/dm-writecache.c
1046
e->seq_count = -1;
drivers/md/dm-writecache.c
1049
r = copy_mc_to_kernel(&wme, memory_entry(wc, e),
drivers/md/dm-writecache.c
1054
e->original_sector = -1;
drivers/md/dm-writecache.c
1055
e->seq_count = -1;
drivers/md/dm-writecache.c
1057
e->original_sector = le64_to_cpu(wme.original_sector);
drivers/md/dm-writecache.c
1058
e->seq_count = le64_to_cpu(wme.seq_count);
drivers/md/dm-writecache.c
1064
struct wc_entry *e = &wc->entries[b];
drivers/md/dm-writecache.c
1066
if (!writecache_entry_is_committed(wc, e)) {
drivers/md/dm-writecache.c
1067
if (read_seq_count(wc, e) != -1) {
drivers/md/dm-writecache.c
1069
clear_seq_count(wc, e);
drivers/md/dm-writecache.c
1072
writecache_add_to_freelist(wc, e);
drivers/md/dm-writecache.c
1076
old = writecache_find_entry(wc, read_original_sector(wc, e), 0);
drivers/md/dm-writecache.c
1078
writecache_insert_entry(wc, e);
drivers/md/dm-writecache.c
1080
if (read_seq_count(wc, old) == read_seq_count(wc, e)) {
drivers/md/dm-writecache.c
1083
(unsigned long long)b, (unsigned long long)read_original_sector(wc, e),
drivers/md/dm-writecache.c
1084
(unsigned long long)read_seq_count(wc, e));
drivers/md/dm-writecache.c
1086
if (read_seq_count(wc, old) > read_seq_count(wc, e)) {
drivers/md/dm-writecache.c
1090
writecache_insert_entry(wc, e);
drivers/md/dm-writecache.c
1339
struct wc_entry *e)
drivers/md/dm-writecache.c
1341
if (e) {
drivers/md/dm-writecache.c
1343
read_original_sector(wc, e) - bio->bi_iter.bi_sector;
drivers/md/dm-writecache.c
1352
struct wc_entry *e;
drivers/md/dm-writecache.c
1356
e = writecache_find_entry(wc, bio->bi_iter.bi_sector, WFE_RETURN_FOLLOWING);
drivers/md/dm-writecache.c
1357
if (e && read_original_sector(wc, e) == bio->bi_iter.bi_sector) {
drivers/md/dm-writecache.c
1360
bio_copy_block(wc, bio, memory_data(wc, e));
drivers/md/dm-writecache.c
1367
bio->bi_iter.bi_sector = cache_sector(wc, e);
drivers/md/dm-writecache.c
1368
if (!writecache_entry_is_committed(wc, e))
drivers/md/dm-writecache.c
1373
writecache_map_remap_origin(wc, bio, e);
drivers/md/dm-writecache.c
1382
struct wc_entry *e, bool search_used)
drivers/md/dm-writecache.c
1385
sector_t start_cache_sec = cache_sector(wc, e);
drivers/md/dm-writecache.c
1400
struct rb_node *next = rb_next(&e->rb_node);
drivers/md/dm-writecache.c
1405
if (f != e + 1)
drivers/md/dm-writecache.c
1408
read_original_sector(wc, e) + (wc->block_size >> SECTOR_SHIFT))
drivers/md/dm-writecache.c
1414
e = f;
drivers/md/dm-writecache.c
1437
struct wc_entry *e;
drivers/md/dm-writecache.c
1447
e = writecache_find_entry(wc, bio->bi_iter.bi_sector, 0);
drivers/md/dm-writecache.c
1448
if (e) {
drivers/md/dm-writecache.c
1449
if (!writecache_entry_is_committed(wc, e)) {
drivers/md/dm-writecache.c
1455
if (!WC_MODE_PMEM(wc) && !e->write_in_progress) {
drivers/md/dm-writecache.c
1466
e = writecache_pop_from_freelist(wc, (sector_t)-1);
drivers/md/dm-writecache.c
1467
if (unlikely(!e)) {
drivers/md/dm-writecache.c
1470
e = writecache_find_entry(wc, bio->bi_iter.bi_sector, WFE_RETURN_FOLLOWING);
drivers/md/dm-writecache.c
1471
writecache_map_remap_origin(wc, bio, e);
drivers/md/dm-writecache.c
1480
write_original_sector_seq_count(wc, e, bio->bi_iter.bi_sector, wc->seq_count);
drivers/md/dm-writecache.c
1481
writecache_insert_entry(wc, e);
drivers/md/dm-writecache.c
1486
bio_copy_block(wc, bio, memory_data(wc, e));
drivers/md/dm-writecache.c
1489
writecache_bio_copy_ssd(wc, bio, e, search_used);
drivers/md/dm-writecache.c
1685
struct wc_entry *e;
drivers/md/dm-writecache.c
1697
e = wb->wc_list[i];
drivers/md/dm-writecache.c
1698
BUG_ON(!e->write_in_progress);
drivers/md/dm-writecache.c
1699
e->write_in_progress = false;
drivers/md/dm-writecache.c
1700
INIT_LIST_HEAD(&e->lru);
drivers/md/dm-writecache.c
1702
writecache_free_entry(wc, e);
drivers/md/dm-writecache.c
1723
struct wc_entry *e;
drivers/md/dm-writecache.c
1732
e = c->e;
drivers/md/dm-writecache.c
1734
BUG_ON(!e->write_in_progress);
drivers/md/dm-writecache.c
1735
e->write_in_progress = false;
drivers/md/dm-writecache.c
1736
INIT_LIST_HEAD(&e->lru);
drivers/md/dm-writecache.c
1738
writecache_free_entry(wc, e);
drivers/md/dm-writecache.c
1742
e++;
drivers/md/dm-writecache.c
1796
static bool wc_add_block(struct writeback_struct *wb, struct wc_entry *e)
drivers/md/dm-writecache.c
1800
void *address = memory_data(wc, e);
drivers/md/dm-writecache.c
1831
struct wc_entry *e, *f;
drivers/md/dm-writecache.c
1838
e = container_of(wbl->list.prev, struct wc_entry, lru);
drivers/md/dm-writecache.c
1839
list_del(&e->lru);
drivers/md/dm-writecache.c
1841
max_pages = e->wc_list_contiguous;
drivers/md/dm-writecache.c
1848
bio->bi_iter.bi_sector = read_original_sector(wc, e);
drivers/md/dm-writecache.c
1859
BUG_ON(!wc_add_block(wb, e));
drivers/md/dm-writecache.c
1861
wb->wc_list[0] = e;
drivers/md/dm-writecache.c
1867
read_original_sector(wc, e) + (wc->block_size >> SECTOR_SHIFT))
drivers/md/dm-writecache.c
1874
e = f;
drivers/md/dm-writecache.c
1894
struct wc_entry *e, *f;
drivers/md/dm-writecache.c
1902
e = container_of(wbl->list.prev, struct wc_entry, lru);
drivers/md/dm-writecache.c
1903
list_del(&e->lru);
drivers/md/dm-writecache.c
1905
n_sectors = e->wc_list_contiguous << (wc->block_size_bits - SECTOR_SHIFT);
drivers/md/dm-writecache.c
1908
from.sector = cache_sector(wc, e);
drivers/md/dm-writecache.c
1911
to.sector = read_original_sector(wc, e);
drivers/md/dm-writecache.c
1916
c->e = e;
drivers/md/dm-writecache.c
1917
c->n_entries = e->wc_list_contiguous;
drivers/md/dm-writecache.c
1922
BUG_ON(f != e + 1);
drivers/md/dm-writecache.c
1924
e = f;
drivers/md/dm-writecache.c
1945
struct wc_entry *f, *g, *e = NULL;
drivers/md/dm-writecache.c
2007
if (unlikely(!e)) {
drivers/md/dm-writecache.c
2009
e = container_of(rb_first(&wc->tree), struct wc_entry, rb_node);
drivers/md/dm-writecache.c
2011
e = g;
drivers/md/dm-writecache.c
2013
e = container_of(wc->lru.prev, struct wc_entry, lru);
drivers/md/dm-writecache.c
2014
BUG_ON(e->write_in_progress);
drivers/md/dm-writecache.c
2015
if (unlikely(!writecache_entry_is_committed(wc, e)))
drivers/md/dm-writecache.c
2018
node = rb_prev(&e->rb_node);
drivers/md/dm-writecache.c
2022
read_original_sector(wc, e))) {
drivers/md/dm-writecache.c
2024
list_move(&e->lru, &skipped);
drivers/md/dm-writecache.c
2030
list_move(&e->lru, &wbl.list);
drivers/md/dm-writecache.c
2032
e->write_in_progress = true;
drivers/md/dm-writecache.c
2033
e->wc_list_contiguous = 1;
drivers/md/dm-writecache.c
2035
f = e;
drivers/md/dm-writecache.c
2070
e->wc_list_contiguous++;
drivers/md/dm-writecache.c
2071
if (unlikely(e->wc_list_contiguous == BIO_MAX_VECS)) {
drivers/md/dm-writecache.c
2116
struct wc_entry e;
drivers/md/dm-writecache.c
2136
e.index = n_blocks;
drivers/md/dm-writecache.c
2137
if (e.index != n_blocks)
drivers/md/dm-writecache.c
234
struct wc_entry *e;
drivers/md/dm-writecache.c
392
static struct wc_memory_entry *memory_entry(struct dm_writecache *wc, struct wc_entry *e)
drivers/md/dm-writecache.c
394
return &sb(wc)->entries[e->index];
drivers/md/dm-writecache.c
397
static void *memory_data(struct dm_writecache *wc, struct wc_entry *e)
drivers/md/dm-writecache.c
399
return (char *)wc->block_start + (e->index << wc->block_size_bits);
drivers/md/dm-writecache.c
402
static sector_t cache_sector(struct dm_writecache *wc, struct wc_entry *e)
drivers/md/dm-writecache.c
405
((sector_t)e->index << (wc->block_size_bits - SECTOR_SHIFT));
drivers/md/dm-writecache.c
408
static uint64_t read_original_sector(struct dm_writecache *wc, struct wc_entry *e)
drivers/md/dm-writecache.c
411
return e->original_sector;
drivers/md/dm-writecache.c
413
return le64_to_cpu(memory_entry(wc, e)->original_sector);
drivers/md/dm-writecache.c
417
static uint64_t read_seq_count(struct dm_writecache *wc, struct wc_entry *e)
drivers/md/dm-writecache.c
420
return e->seq_count;
drivers/md/dm-writecache.c
422
return le64_to_cpu(memory_entry(wc, e)->seq_count);
drivers/md/dm-writecache.c
426
static void clear_seq_count(struct dm_writecache *wc, struct wc_entry *e)
drivers/md/dm-writecache.c
429
e->seq_count = -1;
drivers/md/dm-writecache.c
431
pmem_assign(memory_entry(wc, e)->seq_count, cpu_to_le64(-1));
drivers/md/dm-writecache.c
434
static void write_original_sector_seq_count(struct dm_writecache *wc, struct wc_entry *e,
drivers/md/dm-writecache.c
439
e->original_sector = original_sector;
drivers/md/dm-writecache.c
440
e->seq_count = seq_count;
drivers/md/dm-writecache.c
444
pmem_assign(*memory_entry(wc, e), me);
drivers/md/dm-writecache.c
609
struct wc_entry *e;
drivers/md/dm-writecache.c
616
e = container_of(node, struct wc_entry, rb_node);
drivers/md/dm-writecache.c
617
if (read_original_sector(wc, e) == block)
drivers/md/dm-writecache.c
620
node = (read_original_sector(wc, e) >= block ?
drivers/md/dm-writecache.c
621
e->rb_node.rb_left : e->rb_node.rb_right);
drivers/md/dm-writecache.c
625
if (read_original_sector(wc, e) >= block)
drivers/md/dm-writecache.c
626
return e;
drivers/md/dm-writecache.c
628
node = rb_next(&e->rb_node);
drivers/md/dm-writecache.c
632
e = container_of(node, struct wc_entry, rb_node);
drivers/md/dm-writecache.c
633
return e;
drivers/md/dm-writecache.c
641
node = rb_prev(&e->rb_node);
drivers/md/dm-writecache.c
643
node = rb_next(&e->rb_node);
drivers/md/dm-writecache.c
645
return e;
drivers/md/dm-writecache.c
648
return e;
drivers/md/dm-writecache.c
649
e = e2;
drivers/md/dm-writecache.c
655
struct wc_entry *e;
drivers/md/dm-writecache.c
659
e = container_of(*node, struct wc_entry, rb_node);
drivers/md/dm-writecache.c
660
parent = &e->rb_node;
drivers/md/dm-writecache.c
661
if (read_original_sector(wc, e) > read_original_sector(wc, ins))
drivers/md/dm-writecache.c
672
static void writecache_unlink(struct dm_writecache *wc, struct wc_entry *e)
drivers/md/dm-writecache.c
674
list_del(&e->lru);
drivers/md/dm-writecache.c
675
rb_erase(&e->rb_node, &wc->tree);
drivers/md/dm-writecache.c
678
static void writecache_add_to_freelist(struct dm_writecache *wc, struct wc_entry *e)
drivers/md/dm-writecache.c
684
wc->current_free = e;
drivers/md/dm-writecache.c
687
if (&e->rb_node < *node)
drivers/md/dm-writecache.c
692
rb_link_node(&e->rb_node, parent, node);
drivers/md/dm-writecache.c
693
rb_insert_color(&e->rb_node, &wc->freetree);
drivers/md/dm-writecache.c
695
list_add_tail(&e->lru, &wc->freelist);
drivers/md/dm-writecache.c
718
struct wc_entry *e;
drivers/md/dm-writecache.c
725
e = wc->current_free;
drivers/md/dm-writecache.c
726
if (expected_sector != (sector_t)-1 && unlikely(cache_sector(wc, e) != expected_sector))
drivers/md/dm-writecache.c
728
next = rb_next(&e->rb_node);
drivers/md/dm-writecache.c
729
rb_erase(&e->rb_node, &wc->freetree);
drivers/md/dm-writecache.c
736
e = container_of(wc->freelist.next, struct wc_entry, lru);
drivers/md/dm-writecache.c
737
if (expected_sector != (sector_t)-1 && unlikely(cache_sector(wc, e) != expected_sector))
drivers/md/dm-writecache.c
739
list_del(&e->lru);
drivers/md/dm-writecache.c
745
return e;
drivers/md/dm-writecache.c
748
static void writecache_free_entry(struct dm_writecache *wc, struct wc_entry *e)
drivers/md/dm-writecache.c
750
writecache_unlink(wc, e);
drivers/md/dm-writecache.c
751
writecache_add_to_freelist(wc, e);
drivers/md/dm-writecache.c
752
clear_seq_count(wc, e);
drivers/md/dm-writecache.c
753
writecache_flush_region(wc, memory_entry(wc, e), sizeof(struct wc_memory_entry));
drivers/md/dm-writecache.c
781
static void writecache_flush_entry(struct dm_writecache *wc, struct wc_entry *e)
drivers/md/dm-writecache.c
783
writecache_flush_region(wc, memory_entry(wc, e), sizeof(struct wc_memory_entry));
drivers/md/dm-writecache.c
785
writecache_flush_region(wc, memory_data(wc, e), wc->block_size);
drivers/md/dm-writecache.c
788
static bool writecache_entry_is_committed(struct dm_writecache *wc, struct wc_entry *e)
drivers/md/dm-writecache.c
790
return read_seq_count(wc, e) < wc->seq_count;
drivers/md/dm-writecache.c
795
struct wc_entry *e, *e2;
drivers/md/dm-writecache.c
804
e = container_of(wc->lru.next, struct wc_entry, lru);
drivers/md/dm-writecache.c
805
if (writecache_entry_is_committed(wc, e)) {
drivers/md/dm-writecache.c
814
writecache_flush_entry(wc, e);
drivers/md/dm-writecache.c
815
if (unlikely(e->lru.next == &wc->lru))
drivers/md/dm-writecache.c
817
e2 = container_of(e->lru.next, struct wc_entry, lru);
drivers/md/dm-writecache.c
820
e = e2;
drivers/md/dm-writecache.c
837
struct rb_node *rb_node = rb_prev(&e->rb_node);
drivers/md/dm-writecache.c
841
if (read_original_sector(wc, e2) == read_original_sector(wc, e) &&
drivers/md/dm-writecache.c
847
if (unlikely(e->lru.prev == &wc->lru))
drivers/md/dm-writecache.c
849
e = container_of(e->lru.prev, struct wc_entry, lru);
drivers/md/dm-writecache.c
882
struct wc_entry *e;
drivers/md/dm-writecache.c
885
e = writecache_find_entry(wc, start, WFE_RETURN_FOLLOWING | WFE_LOWEST_SEQ);
drivers/md/dm-writecache.c
886
if (unlikely(!e))
drivers/md/dm-writecache.c
889
while (read_original_sector(wc, e) < end) {
drivers/md/dm-writecache.c
890
struct rb_node *node = rb_next(&e->rb_node);
drivers/md/dm-writecache.c
892
if (likely(!e->write_in_progress)) {
drivers/md/dm-writecache.c
900
if (!writecache_entry_is_committed(wc, e))
drivers/md/dm-writecache.c
902
writecache_free_entry(wc, e);
drivers/md/dm-writecache.c
908
e = container_of(node, struct wc_entry, rb_node);
drivers/md/dm-writecache.c
968
struct wc_entry *e = &wc->entries[b];
drivers/md/dm-writecache.c
970
e->index = b;
drivers/md/dm-writecache.c
971
e->write_in_progress = false;
drivers/md/dm-zoned-metadata.c
1684
unsigned int i = 0, e = 0, chunk = 0;
drivers/md/dm-zoned-metadata.c
1703
e = 0;
drivers/md/dm-zoned-metadata.c
1707
dzone_id = le32_to_cpu(dmap[e].dzone_id);
drivers/md/dm-zoned-metadata.c
1735
bzone_id = le32_to_cpu(dmap[e].bzone_id);
drivers/md/dm-zoned-metadata.c
1769
e++;
drivers/md/dm-zoned-metadata.c
1770
if (e >= DMZ_MAP_ENTRIES)
drivers/md/md-bitmap.c
1959
unsigned long e)
drivers/md/md-bitmap.c
1967
for (chunk = s; chunk <= e; chunk++) {
drivers/md/md-bitmap.h
95
unsigned long e);
drivers/md/md-llbitmap.c
1340
unsigned long e)
drivers/md/md-llbitmap.c
1342
llbitmap_state_machine(mddev->bitmap, s, e, BitmapActionStartwrite);
drivers/md/md.c
4807
char *e;
drivers/md/md.c
4808
int major = simple_strtoul(buf, &e, 10);
drivers/md/md.c
4814
if (!*buf || *e != ':' || !e[1] || e[1] == '\n')
drivers/md/md.c
4816
minor = simple_strtoul(e+1, &e, 10);
drivers/md/md.c
4817
if (*e && *e != '\n')
drivers/md/md.c
4968
char *e;
drivers/md/md.c
5006
major = simple_strtoul(buf, &e, 10);
drivers/md/md.c
5008
if (e==buf || *e != '.')
drivers/md/md.c
5010
buf = e+1;
drivers/md/md.c
5011
minor = simple_strtoul(buf, &e, 10);
drivers/md/md.c
5012
if (e==buf || (*e && *e != '\n') )
drivers/md/persistent-data/dm-space-map-common.c
716
static int __sm_ll_inc(struct ll_disk *ll, dm_block_t b, dm_block_t e,
drivers/md/persistent-data/dm-space-map-common.c
735
bit_end = min(bit + (e - b), (dm_block_t) ll->entries_per_block);
drivers/md/persistent-data/dm-space-map-common.c
746
int sm_ll_inc(struct ll_disk *ll, dm_block_t b, dm_block_t e,
drivers/md/persistent-data/dm-space-map-common.c
750
while (b != e) {
drivers/md/persistent-data/dm-space-map-common.c
751
int r = __sm_ll_inc(ll, b, e, nr_allocations, &b);
drivers/md/persistent-data/dm-space-map-common.c
899
static int __sm_ll_dec(struct ll_disk *ll, dm_block_t b, dm_block_t e,
drivers/md/persistent-data/dm-space-map-common.c
918
bit_end = min(bit + (e - b), (dm_block_t) ll->entries_per_block);
drivers/md/persistent-data/dm-space-map-common.c
928
int sm_ll_dec(struct ll_disk *ll, dm_block_t b, dm_block_t e,
drivers/md/persistent-data/dm-space-map-common.c
932
while (b != e) {
drivers/md/persistent-data/dm-space-map-common.c
933
int r = __sm_ll_dec(ll, b, e, nr_allocations, &b);
drivers/md/persistent-data/dm-space-map-common.h
132
int sm_ll_inc(struct ll_disk *ll, dm_block_t b, dm_block_t e, int32_t *nr_allocations);
drivers/md/persistent-data/dm-space-map-common.h
133
int sm_ll_dec(struct ll_disk *ll, dm_block_t b, dm_block_t e, int32_t *nr_allocations);
drivers/md/persistent-data/dm-space-map-disk.c
104
static int sm_disk_inc_blocks(struct dm_space_map *sm, dm_block_t b, dm_block_t e)
drivers/md/persistent-data/dm-space-map-disk.c
110
r = sm_ll_inc(&smd->ll, b, e, &nr_allocations);
drivers/md/persistent-data/dm-space-map-disk.c
117
static int sm_disk_dec_blocks(struct dm_space_map *sm, dm_block_t b, dm_block_t e)
drivers/md/persistent-data/dm-space-map-disk.c
123
r = sm_ll_dec(&smd->ll, b, e, &nr_allocations);
drivers/md/persistent-data/dm-space-map-metadata.c
122
enum block_op_type type, dm_block_t b, dm_block_t e)
drivers/md/persistent-data/dm-space-map-metadata.c
137
bop->e = e;
drivers/md/persistent-data/dm-space-map-metadata.c
183
static int add_bop(struct sm_metadata *smm, enum block_op_type type, dm_block_t b, dm_block_t e)
drivers/md/persistent-data/dm-space-map-metadata.c
185
int r = brb_push(&smm->uncommitted, type, b, e);
drivers/md/persistent-data/dm-space-map-metadata.c
202
r = sm_ll_inc(&smm->ll, op->b, op->e, &nr_allocations);
drivers/md/persistent-data/dm-space-map-metadata.c
206
r = sm_ll_dec(&smm->ll, op->b, op->e, &nr_allocations);
drivers/md/persistent-data/dm-space-map-metadata.c
319
if (b < op->b || b >= op->e)
drivers/md/persistent-data/dm-space-map-metadata.c
360
if (b < op->b || b >= op->e)
drivers/md/persistent-data/dm-space-map-metadata.c
413
static int sm_metadata_inc_blocks(struct dm_space_map *sm, dm_block_t b, dm_block_t e)
drivers/md/persistent-data/dm-space-map-metadata.c
420
r = add_bop(smm, BOP_INC, b, e);
drivers/md/persistent-data/dm-space-map-metadata.c
425
r = sm_ll_inc(&smm->ll, b, e, &nr_allocations);
drivers/md/persistent-data/dm-space-map-metadata.c
432
static int sm_metadata_dec_blocks(struct dm_space_map *sm, dm_block_t b, dm_block_t e)
drivers/md/persistent-data/dm-space-map-metadata.c
439
r = add_bop(smm, BOP_DEC, b, e);
drivers/md/persistent-data/dm-space-map-metadata.c
442
r = sm_ll_dec(&smm->ll, b, e, &nr_allocations);
drivers/md/persistent-data/dm-space-map-metadata.c
656
static int sm_bootstrap_inc_blocks(struct dm_space_map *sm, dm_block_t b, dm_block_t e)
drivers/md/persistent-data/dm-space-map-metadata.c
661
r = add_bop(smm, BOP_INC, b, e);
drivers/md/persistent-data/dm-space-map-metadata.c
668
static int sm_bootstrap_dec_blocks(struct dm_space_map *sm, dm_block_t b, dm_block_t e)
drivers/md/persistent-data/dm-space-map-metadata.c
673
r = add_bop(smm, BOP_DEC, b, e);
drivers/md/persistent-data/dm-space-map-metadata.c
94
dm_block_t e;
drivers/md/persistent-data/dm-space-map.h
122
static inline int dm_sm_inc_blocks(struct dm_space_map *sm, dm_block_t b, dm_block_t e)
drivers/md/persistent-data/dm-space-map.h
124
return sm->inc_blocks(sm, b, e);
drivers/md/persistent-data/dm-space-map.h
132
static inline int dm_sm_dec_blocks(struct dm_space_map *sm, dm_block_t b, dm_block_t e)
drivers/md/persistent-data/dm-space-map.h
134
return sm->dec_blocks(sm, b, e);
drivers/md/persistent-data/dm-space-map.h
50
int (*inc_blocks)(struct dm_space_map *sm, dm_block_t b, dm_block_t e);
drivers/md/persistent-data/dm-space-map.h
51
int (*dec_blocks)(struct dm_space_map *sm, dm_block_t b, dm_block_t e);
drivers/md/persistent-data/dm-transaction-manager.c
387
void dm_tm_inc_range(struct dm_transaction_manager *tm, dm_block_t b, dm_block_t e)
drivers/md/persistent-data/dm-transaction-manager.c
394
dm_sm_inc_blocks(tm->sm, b, e);
drivers/md/persistent-data/dm-transaction-manager.c
409
void dm_tm_dec_range(struct dm_transaction_manager *tm, dm_block_t b, dm_block_t e)
drivers/md/persistent-data/dm-transaction-manager.c
416
dm_sm_dec_blocks(tm->sm, b, e);
drivers/md/persistent-data/dm-transaction-manager.h
104
void dm_tm_inc_range(struct dm_transaction_manager *tm, dm_block_t b, dm_block_t e);
drivers/md/persistent-data/dm-transaction-manager.h
106
void dm_tm_dec_range(struct dm_transaction_manager *tm, dm_block_t b, dm_block_t e);
drivers/md/raid5-ppl.c
1019
ret = ppl_recover_entry(log, e, ppl_sector);
drivers/md/raid5-ppl.c
270
struct ppl_header_entry *e = NULL;
drivers/md/raid5-ppl.c
333
e = last;
drivers/md/raid5-ppl.c
336
if (!e) {
drivers/md/raid5-ppl.c
337
e = &pplhdr->entries[io->entries_count++];
drivers/md/raid5-ppl.c
338
e->data_sector = cpu_to_le64(data_sector);
drivers/md/raid5-ppl.c
339
e->parity_disk = cpu_to_le32(sh->pd_idx);
drivers/md/raid5-ppl.c
340
e->checksum = cpu_to_le32(~0);
drivers/md/raid5-ppl.c
343
le32_add_cpu(&e->data_size, data_disks << PAGE_SHIFT);
drivers/md/raid5-ppl.c
347
le32_add_cpu(&e->pp_size, PAGE_SIZE);
drivers/md/raid5-ppl.c
349
e->checksum = cpu_to_le32(crc32c(le32_to_cpu(e->checksum),
drivers/md/raid5-ppl.c
445
struct ppl_header_entry *e = &pplhdr->entries[i];
drivers/md/raid5-ppl.c
448
__func__, io->seq, i, le64_to_cpu(e->data_sector),
drivers/md/raid5-ppl.c
449
le32_to_cpu(e->pp_size), le32_to_cpu(e->data_size));
drivers/md/raid5-ppl.c
451
e->data_sector = cpu_to_le64(le64_to_cpu(e->data_sector) >>
drivers/md/raid5-ppl.c
453
e->checksum = cpu_to_le32(~le32_to_cpu(e->checksum));
drivers/md/raid5-ppl.c
784
static int ppl_recover_entry(struct ppl_log *log, struct ppl_header_entry *e,
drivers/md/raid5-ppl.c
799
unsigned int pp_size = le32_to_cpu(e->pp_size);
drivers/md/raid5-ppl.c
800
unsigned int data_size = le32_to_cpu(e->data_size);
drivers/md/raid5-ppl.c
810
r_sector_first = le64_to_cpu(e->data_sector) * (block_size >> 9);
drivers/md/raid5-ppl.c
933
BUG_ON(sh.pd_idx != le32_to_cpu(e->parity_disk));
drivers/md/raid5-ppl.c
977
struct ppl_header_entry *e = &pplhdr->entries[i];
drivers/md/raid5-ppl.c
978
u32 pp_size = le32_to_cpu(e->pp_size);
drivers/md/raid5-ppl.c
988
crc_stored = le32_to_cpu(e->checksum);
drivers/media/cec/core/cec-pin-error-inj.c
358
u64 e = pin->error_inj[i];
drivers/media/cec/core/cec-pin-error-inj.c
368
mode = (e >> mode_offset) & CEC_ERROR_INJ_MODE_MASK;
drivers/media/cec/core/cec-pin.c
189
u64 e = pin->error_inj[cmd];
drivers/media/cec/core/cec-pin.c
190
unsigned int mode = (e >> mode_offset) & CEC_ERROR_INJ_MODE_MASK;
drivers/media/cec/core/cec-pin.c
251
u64 e = pin->error_inj[cmd];
drivers/media/cec/core/cec-pin.c
252
unsigned int mode = (e >> mode_offset) & CEC_ERROR_INJ_MODE_MASK;
drivers/media/dvb-core/dvb_frontend.c
243
struct dvb_frontend_event *e;
drivers/media/dvb-core/dvb_frontend.c
259
e = &events->events[events->eventw];
drivers/media/dvb-core/dvb_frontend.c
260
e->status = status;
drivers/media/dvb-core/dvb_frontend.c
261
e->parameters = fepriv->parameters_out;
drivers/media/dvb-frontends/drx39xyj/drxj.c
9473
u32 e = 0; /* exponent value used for QAM BER/SER */
drivers/media/dvb-frontends/drx39xyj/drxj.c
9562
e = (qsym_err_vd & QAM_VD_NR_QSYM_ERRORS_EXP__M) >>
drivers/media/dvb-frontends/drx39xyj/drxj.c
9567
if ((m << e) >> 3 > 549752)
drivers/media/dvb-frontends/drx39xyj/drxj.c
9568
qam_vd_ser = 500000 * vd_bit_cnt * ((e > 2) ? 1 : 8) / 8;
drivers/media/dvb-frontends/drx39xyj/drxj.c
9570
qam_vd_ser = m << ((e > 2) ? (e - 3) : e);
drivers/media/dvb-frontends/drx39xyj/drxj.c
9583
e = (pre_bit_err_rs & FEC_RS_NR_BIT_ERRORS_EXP__M) >>
drivers/media/dvb-frontends/drx39xyj/drxj.c
9588
ber_cnt = m << e;
drivers/media/dvb-frontends/drx39xyj/drxj.c
9591
if (m > (rs_bit_cnt >> (e + 1)) || (rs_bit_cnt >> e) == 0)
drivers/media/dvb-frontends/drx39xyj/drxj.c
9592
qam_pre_rs_ber = 500000 * rs_bit_cnt >> e;
drivers/media/dvb-frontends/drx39xyj/drxj.c
9610
e = post_bit_err_rs * 742686;
drivers/media/dvb-frontends/drx39xyj/drxj.c
9612
qam_post_rs_ber = e / m;
drivers/media/dvb-frontends/drx39xyj/drxj.c
9626
p->pre_bit_count.stat[0].uvalue += vd_bit_cnt * ((e > 2) ? 1 : 8) / 8;
drivers/media/dvb-frontends/drx39xyj/drxj.c
9629
p->pre_bit_count.stat[0].uvalue += rs_bit_cnt >> e;
drivers/media/dvb-frontends/drx39xyj/drxj.c
9633
p->post_bit_count.stat[0].uvalue += rs_bit_cnt >> e;
drivers/media/i2c/adv7604.c
1891
#define _SEL(a,b,c,d,e,f) { \
drivers/media/i2c/adv7604.c
1893
ADV76XX_OP_CH_SEL_##d, ADV76XX_OP_CH_SEL_##e, ADV76XX_OP_CH_SEL_##f }
drivers/media/i2c/adv7842.c
2043
#define _SEL(a, b, c, d, e, f) { \
drivers/media/i2c/adv7842.c
2045
ADV7842_OP_CH_SEL_##d, ADV7842_OP_CH_SEL_##e, ADV7842_OP_CH_SEL_##f }
drivers/media/i2c/adv7842.c
2509
static int adv7842_set_edid(struct v4l2_subdev *sd, struct v4l2_edid *e)
drivers/media/i2c/adv7842.c
2512
unsigned int max_blocks = e->pad == ADV7842_EDID_PORT_VGA ? 1 : 4;
drivers/media/i2c/adv7842.c
2515
memset(e->reserved, 0, sizeof(e->reserved));
drivers/media/i2c/adv7842.c
2517
if (e->pad > ADV7842_EDID_PORT_VGA)
drivers/media/i2c/adv7842.c
2519
if (e->start_block != 0)
drivers/media/i2c/adv7842.c
2521
if (e->pad < ADV7842_EDID_PORT_VGA && state->vga_edid.blocks)
drivers/media/i2c/adv7842.c
2523
if (e->pad == ADV7842_EDID_PORT_VGA && state->hdmi_edid.blocks > 2)
drivers/media/i2c/adv7842.c
2525
if (e->blocks > max_blocks) {
drivers/media/i2c/adv7842.c
2526
e->blocks = max_blocks;
drivers/media/i2c/adv7842.c
2531
if (e->blocks)
drivers/media/i2c/adv7842.c
2532
state->aspect_ratio = v4l2_calc_aspect_ratio(e->edid[0x15],
drivers/media/i2c/adv7842.c
2533
e->edid[0x16]);
drivers/media/i2c/adv7842.c
2535
switch (e->pad) {
drivers/media/i2c/adv7842.c
2538
state->vga_edid.blocks = e->blocks;
drivers/media/i2c/adv7842.c
2539
state->vga_edid.present = e->blocks ? 0x1 : 0x0;
drivers/media/i2c/adv7842.c
2540
if (e->blocks)
drivers/media/i2c/adv7842.c
2541
memcpy(state->vga_edid.edid, e->edid, 128);
drivers/media/i2c/adv7842.c
2547
state->hdmi_edid.blocks = e->blocks;
drivers/media/i2c/adv7842.c
2548
if (e->blocks) {
drivers/media/i2c/adv7842.c
2549
state->hdmi_edid.present |= 0x04 << e->pad;
drivers/media/i2c/adv7842.c
2550
memcpy(state->hdmi_edid.edid, e->edid, 128 * e->blocks);
drivers/media/i2c/adv7842.c
2552
state->hdmi_edid.present &= ~(0x04 << e->pad);
drivers/media/i2c/adv7842.c
2555
err = edid_write_hdmi_segment(sd, e->pad);
drivers/media/i2c/adv7842.c
2561
v4l2_err(sd, "error %d writing edid on port %d\n", err, e->pad);
drivers/media/pci/ivtv/ivtv-fileops.c
175
struct v4l2_enc_idx_entry *e = itv->pgm_info + idx;
drivers/media/pci/ivtv/ivtv-fileops.c
181
e->offset = read_enc(addr + 4) + ((u64)read_enc(addr + 8) << 32);
drivers/media/pci/ivtv/ivtv-fileops.c
182
if (e->offset > itv->mpg_data_received) {
drivers/media/pci/ivtv/ivtv-fileops.c
185
e->offset += itv->vbi_data_inserted;
drivers/media/pci/ivtv/ivtv-fileops.c
186
e->length = read_enc(addr);
drivers/media/pci/ivtv/ivtv-fileops.c
187
e->pts = read_enc(addr + 16) + ((u64)(read_enc(addr + 20) & 1) << 32);
drivers/media/pci/ivtv/ivtv-fileops.c
188
e->flags = mapping[read_enc(addr + 12) & 7];
drivers/media/pci/ivtv/ivtv-ioctl.c
1265
struct v4l2_enc_idx_entry *e = idx->entry;
drivers/media/pci/ivtv/ivtv-ioctl.c
1278
*e = itv->pgm_info[(itv->pgm_info_read_idx + i) % IVTV_MAX_PGM_INDEX];
drivers/media/pci/ivtv/ivtv-ioctl.c
1279
if ((e->flags & V4L2_ENC_IDX_FRAME_MASK) <= V4L2_ENC_IDX_FRAME_B) {
drivers/media/pci/ivtv/ivtv-ioctl.c
1281
e++;
drivers/media/platform/mediatek/mdp3/mtk-mdp3-comp.h
27
typeof(evt) (e) = (evt); \
drivers/media/platform/mediatek/mdp3/mtk-mdp3-comp.h
28
cmdq_pkt_wfe(&((c)->pkt), (e), true); \
drivers/media/platform/mediatek/mdp3/mtk-mdp3-comp.h
34
typeof(evt) (e) = (evt); \
drivers/media/platform/mediatek/mdp3/mtk-mdp3-comp.h
35
cmdq_pkt_wfe(&((c)->pkt), (e), false); \
drivers/media/platform/mediatek/mdp3/mtk-mdp3-comp.h
41
typeof(evt) (e) = (evt); \
drivers/media/platform/mediatek/mdp3/mtk-mdp3-comp.h
42
cmdq_pkt_clear_event(&((c)->pkt), (e)); \
drivers/media/platform/mediatek/mdp3/mtk-mdp3-comp.h
48
typeof(evt) (e) = (evt); \
drivers/media/platform/mediatek/mdp3/mtk-mdp3-comp.h
49
cmdq_pkt_set_event(&((c)->pkt), (e)); \
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_av1_req_lat_if.c
1085
int e;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_av1_req_lat_if.c
1089
e = D - ((u32)1 << *shift);
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_av1_req_lat_if.c
1092
f = AV1_DIV_ROUND_UP_POW2(e, *shift - DIV_LUT_BITS);
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_av1_req_lat_if.c
1094
f = e << (DIV_LUT_BITS - *shift);
drivers/media/platform/renesas/vsp1/vsp1_drm.c
28
#define BRX_NAME(e) (e)->type == VSP1_ENTITY_BRU ? "BRU" : "BRS"
drivers/media/platform/renesas/vsp1/vsp1_video.c
470
struct vsp1_entity *e;
drivers/media/platform/renesas/vsp1/vsp1_video.c
476
e = to_vsp1_entity(subdev);
drivers/media/platform/renesas/vsp1/vsp1_video.c
477
list_add_tail(&e->list_pipe, &pipe->entities);
drivers/media/platform/renesas/vsp1/vsp1_video.c
478
e->pipe = pipe;
drivers/media/platform/renesas/vsp1/vsp1_video.c
480
switch (e->type) {
drivers/media/platform/renesas/vsp1/vsp1_video.c
494
pipe->lif = e;
drivers/media/platform/renesas/vsp1/vsp1_video.c
499
pipe->brx = e;
drivers/media/platform/renesas/vsp1/vsp1_video.c
503
pipe->hgo = e;
drivers/media/platform/renesas/vsp1/vsp1_video.c
507
pipe->hgt = e;
drivers/media/platform/verisilicon/rockchip_vpu981_hw_av1_dec.c
464
u64 e;
drivers/media/platform/verisilicon/rockchip_vpu981_hw_av1_dec.c
468
e = d - ((u32)1 << *shift);
drivers/media/platform/verisilicon/rockchip_vpu981_hw_av1_dec.c
471
f = AV1_DIV_ROUND_UP_POW2(e, *shift - DIV_LUT_BITS);
drivers/media/platform/verisilicon/rockchip_vpu981_hw_av1_dec.c
473
f = e << (DIV_LUT_BITS - *shift);
drivers/media/rc/ir-imon-decoder.c
193
struct ir_raw_event *e = events;
drivers/media/rc/ir-imon-decoder.c
198
init_ir_raw_event_duration(e, 1, IMON_UNIT);
drivers/media/rc/ir-imon-decoder.c
206
if (pulse == e->pulse) {
drivers/media/rc/ir-imon-decoder.c
207
e->duration += IMON_UNIT;
drivers/media/rc/ir-imon-decoder.c
211
init_ir_raw_event_duration(++e, pulse, IMON_UNIT);
drivers/media/rc/ir-imon-decoder.c
216
if (pulse == e->pulse) {
drivers/media/rc/ir-imon-decoder.c
217
e->duration += IMON_UNIT;
drivers/media/rc/ir-imon-decoder.c
221
init_ir_raw_event_duration(++e, pulse, IMON_UNIT);
drivers/media/rc/ir-imon-decoder.c
225
if (e->pulse)
drivers/media/rc/ir-imon-decoder.c
226
e++;
drivers/media/rc/ir-imon-decoder.c
228
return e - events;
drivers/media/rc/ir-jvc-decoder.c
191
struct ir_raw_event *e = events;
drivers/media/rc/ir-jvc-decoder.c
196
ret = ir_raw_gen_pd(&e, max, &ir_jvc_timings, JVC_NBITS, raw);
drivers/media/rc/ir-jvc-decoder.c
200
return e - events;
drivers/media/rc/ir-mce_kbd-decoder.c
403
struct ir_raw_event *e = events;
drivers/media/rc/ir-mce_kbd-decoder.c
417
ret = ir_raw_gen_manchester(&e, max, &ir_mce_kbd_timings, len, raw);
drivers/media/rc/ir-mce_kbd-decoder.c
421
return e - events;
drivers/media/rc/ir-nec-decoder.c
235
struct ir_raw_event *e = events;
drivers/media/rc/ir-nec-decoder.c
243
ret = ir_raw_gen_pd(&e, max, &ir_nec_timings, NEC_NBITS, raw);
drivers/media/rc/ir-nec-decoder.c
247
return e - events;
drivers/media/rc/ir-rc5-decoder.c
213
struct ir_raw_event *e = events;
drivers/media/rc/ir-rc5-decoder.c
226
ret = ir_raw_gen_manchester(&e, max, &ir_rc5_timings,
drivers/media/rc/ir-rc5-decoder.c
242
ret = ir_raw_gen_manchester(&e, max, &ir_rc5x_timings[0],
drivers/media/rc/ir-rc5-decoder.c
247
ret = ir_raw_gen_manchester(&e, max - (e - events),
drivers/media/rc/ir-rc5-decoder.c
256
ret = ir_raw_gen_manchester(&e, max, &ir_rc5_sz_timings,
drivers/media/rc/ir-rc5-decoder.c
265
return e - events;
drivers/media/rc/ir-rc6-decoder.c
313
struct ir_raw_event *e = events;
drivers/media/rc/ir-rc6-decoder.c
317
ret = ir_raw_gen_manchester(&e, max - (e - events),
drivers/media/rc/ir-rc6-decoder.c
324
ret = ir_raw_gen_manchester(&e, max - (e - events),
drivers/media/rc/ir-rc6-decoder.c
330
ret = ir_raw_gen_manchester(&e, max - (e - events),
drivers/media/rc/ir-rc6-decoder.c
355
ret = ir_raw_gen_manchester(&e, max - (e - events),
drivers/media/rc/ir-rc6-decoder.c
362
ret = ir_raw_gen_manchester(&e, max - (e - events),
drivers/media/rc/ir-rc6-decoder.c
368
ret = ir_raw_gen_manchester(&e, max - (e - events),
drivers/media/rc/ir-rc6-decoder.c
376
return e - events;
drivers/media/rc/ir-rcmm-decoder.c
204
struct ir_raw_event *e = events;
drivers/media/rc/ir-rcmm-decoder.c
209
ret = ir_rcmm_rawencoder(&e, max, 32, scancode);
drivers/media/rc/ir-rcmm-decoder.c
212
ret = ir_rcmm_rawencoder(&e, max, 24, scancode);
drivers/media/rc/ir-rcmm-decoder.c
215
ret = ir_rcmm_rawencoder(&e, max, 12, scancode);
drivers/media/rc/ir-rcmm-decoder.c
224
return e - events;
drivers/media/rc/ir-sanyo-decoder.c
192
struct ir_raw_event *e = events;
drivers/media/rc/ir-sanyo-decoder.c
201
ret = ir_raw_gen_pd(&e, max, &ir_sanyo_timings, SANYO_NBITS, raw);
drivers/media/rc/ir-sanyo-decoder.c
205
return e - events;
drivers/media/rc/ir-sharp-decoder.c
195
struct ir_raw_event *e = events;
drivers/media/rc/ir-sharp-decoder.c
201
ret = ir_raw_gen_pd(&e, max, &ir_sharp_timings, SHARP_NBITS,
drivers/media/rc/ir-sharp-decoder.c
210
ret = ir_raw_gen_pd(&e, max, &ir_sharp_timings, SHARP_NBITS,
drivers/media/rc/ir-sharp-decoder.c
215
return e - events;
drivers/media/rc/ir-sony-decoder.c
190
struct ir_raw_event *e = events;
drivers/media/rc/ir-sony-decoder.c
206
ret = ir_raw_gen_pl(&e, max, &ir_sony_timings, len, raw);
drivers/media/rc/ir-sony-decoder.c
210
return e - events;
drivers/media/rc/rc-main.c
503
const struct rc_map_table *e = elt;
drivers/media/rc/rc-main.c
505
if (*scancode < e->scancode)
drivers/media/rc/rc-main.c
507
else if (*scancode > e->scancode)
drivers/media/test-drivers/vidtv/vidtv_channel.c
34
static void vidtv_channel_encoder_destroy(struct vidtv_encoder *e)
drivers/media/test-drivers/vidtv/vidtv_channel.c
365
vidtv_channel_destroy_service_list(struct vidtv_psi_desc_service_list_entry *e)
drivers/media/test-drivers/vidtv/vidtv_channel.c
369
while (e) {
drivers/media/test-drivers/vidtv/vidtv_channel.c
37
struct vidtv_encoder *curr = e;
drivers/media/test-drivers/vidtv/vidtv_channel.c
370
tmp = e;
drivers/media/test-drivers/vidtv/vidtv_channel.c
371
e = e->next;
drivers/media/test-drivers/vidtv/vidtv_encoder.h
150
void *(*encode)(struct vidtv_encoder *e);
drivers/media/test-drivers/vidtv/vidtv_encoder.h
152
u32 (*clear)(struct vidtv_encoder *e);
drivers/media/test-drivers/vidtv/vidtv_encoder.h
160
void (*destroy)(struct vidtv_encoder *e);
drivers/media/test-drivers/vidtv/vidtv_mux.c
273
struct vidtv_encoder *e)
drivers/media/test-drivers/vidtv/vidtv_mux.c
278
.pid = be16_to_cpu(e->es_pid),
drivers/media/test-drivers/vidtv/vidtv_mux.c
279
.encoder_id = e->id,
drivers/media/test-drivers/vidtv/vidtv_mux.c
280
.stream_id = be16_to_cpu(e->stream_id),
drivers/media/test-drivers/vidtv/vidtv_mux.c
284
struct vidtv_access_unit *au = e->access_units;
drivers/media/test-drivers/vidtv/vidtv_mux.c
296
pid_ctx = vidtv_mux_create_pid_ctx_once(m, be16_to_cpu(e->es_pid));
drivers/media/test-drivers/vidtv/vidtv_mux.c
300
buf = e->encoder_buf + au->offset;
drivers/media/test-drivers/vidtv/vidtv_mux.c
316
e->clear(e);
drivers/media/test-drivers/vidtv/vidtv_mux.c
325
struct vidtv_encoder *e = NULL;
drivers/media/test-drivers/vidtv/vidtv_mux.c
330
e = cur_chnl->encoders;
drivers/media/test-drivers/vidtv/vidtv_mux.c
332
while (e) {
drivers/media/test-drivers/vidtv/vidtv_mux.c
333
e->encode(e);
drivers/media/test-drivers/vidtv/vidtv_mux.c
335
au_nbytes = vidtv_mux_packetize_access_units(m, e);
drivers/media/test-drivers/vidtv/vidtv_mux.c
339
e = e->next;
drivers/media/test-drivers/vidtv/vidtv_psi.c
1803
struct vidtv_psi_table_eit_event *e = eit->event;
drivers/media/test-drivers/vidtv/vidtv_psi.c
1813
while (e) {
drivers/media/test-drivers/vidtv/vidtv_psi.c
1819
desc_loop_len = vidtv_psi_desc_comp_loop_len(e->descriptor);
drivers/media/test-drivers/vidtv/vidtv_psi.c
1820
vidtv_psi_set_desc_loop_len(&e->bitfield, desc_loop_len, 12);
drivers/media/test-drivers/vidtv/vidtv_psi.c
1824
e = e->next;
drivers/media/test-drivers/vidtv/vidtv_psi.c
1833
struct vidtv_psi_table_eit_event *e)
drivers/media/test-drivers/vidtv/vidtv_psi.c
1836
if (e == eit->event)
drivers/media/test-drivers/vidtv/vidtv_psi.c
1839
eit->event = e;
drivers/media/test-drivers/vidtv/vidtv_psi.c
1842
e = NULL;
drivers/media/test-drivers/vidtv/vidtv_psi.c
1977
struct vidtv_psi_table_eit_event *e;
drivers/media/test-drivers/vidtv/vidtv_psi.c
1983
e = kzalloc_obj(*e);
drivers/media/test-drivers/vidtv/vidtv_psi.c
1984
if (!e)
drivers/media/test-drivers/vidtv/vidtv_psi.c
1987
e->event_id = cpu_to_be16(event_id);
drivers/media/test-drivers/vidtv/vidtv_psi.c
2008
memcpy(e->start_time, &mjd_be, sizeof(mjd_be));
drivers/media/test-drivers/vidtv/vidtv_psi.c
2009
e->start_time[2] = bin2bcd(time.tm_hour);
drivers/media/test-drivers/vidtv/vidtv_psi.c
2010
e->start_time[3] = 0;
drivers/media/test-drivers/vidtv/vidtv_psi.c
2011
e->start_time[4] = 0;
drivers/media/test-drivers/vidtv/vidtv_psi.c
2020
memcpy(e->duration, DURATION, sizeof(e->duration));
drivers/media/test-drivers/vidtv/vidtv_psi.c
2022
e->bitfield = cpu_to_be16(RUNNING << 13);
drivers/media/test-drivers/vidtv/vidtv_psi.c
2028
head->next = e;
drivers/media/test-drivers/vidtv/vidtv_psi.c
2031
return e;
drivers/media/test-drivers/vidtv/vidtv_psi.c
2034
void vidtv_psi_eit_event_destroy(struct vidtv_psi_table_eit_event *e)
drivers/media/test-drivers/vidtv/vidtv_psi.c
2037
struct vidtv_psi_table_eit_event *curr_e = e;
drivers/media/test-drivers/vidtv/vidtv_psi.h
802
struct vidtv_psi_table_eit_event *e);
drivers/media/test-drivers/vidtv/vidtv_psi.h
807
void vidtv_psi_eit_event_destroy(struct vidtv_psi_table_eit_event *e);
drivers/media/test-drivers/vidtv/vidtv_s302m.c
165
static void vidtv_s302m_access_unit_destroy(struct vidtv_encoder *e)
drivers/media/test-drivers/vidtv/vidtv_s302m.c
167
struct vidtv_access_unit *head = e->access_units;
drivers/media/test-drivers/vidtv/vidtv_s302m.c
176
e->access_units = NULL;
drivers/media/test-drivers/vidtv/vidtv_s302m.c
179
static void vidtv_s302m_alloc_au(struct vidtv_encoder *e)
drivers/media/test-drivers/vidtv/vidtv_s302m.c
184
if (e->sync && e->sync->is_video_encoder) {
drivers/media/test-drivers/vidtv/vidtv_s302m.c
185
sync_au = e->sync->access_units;
drivers/media/test-drivers/vidtv/vidtv_s302m.c
188
temp = vidtv_s302m_access_unit_init(e->access_units);
drivers/media/test-drivers/vidtv/vidtv_s302m.c
189
if (!e->access_units)
drivers/media/test-drivers/vidtv/vidtv_s302m.c
190
e->access_units = temp;
drivers/media/test-drivers/vidtv/vidtv_s302m.c
198
e->access_units = vidtv_s302m_access_unit_init(NULL);
drivers/media/test-drivers/vidtv/vidtv_s302m.c
202
vidtv_s302m_compute_sample_count_from_video(struct vidtv_encoder *e)
drivers/media/test-drivers/vidtv/vidtv_s302m.c
204
struct vidtv_access_unit *sync_au = e->sync->access_units;
drivers/media/test-drivers/vidtv/vidtv_s302m.c
205
struct vidtv_access_unit *au = e->access_units;
drivers/media/test-drivers/vidtv/vidtv_s302m.c
210
vau_duration_usecs = USEC_PER_SEC / e->sync->sampling_rate_hz;
drivers/media/test-drivers/vidtv/vidtv_s302m.c
211
sample_duration_usecs = USEC_PER_SEC / e->sampling_rate_hz;
drivers/media/test-drivers/vidtv/vidtv_s302m.c
221
static void vidtv_s302m_compute_pts_from_video(struct vidtv_encoder *e)
drivers/media/test-drivers/vidtv/vidtv_s302m.c
223
struct vidtv_access_unit *au = e->access_units;
drivers/media/test-drivers/vidtv/vidtv_s302m.c
224
struct vidtv_access_unit *sync_au = e->sync->access_units;
drivers/media/test-drivers/vidtv/vidtv_s302m.c
234
static u16 vidtv_s302m_get_sample(struct vidtv_encoder *e)
drivers/media/test-drivers/vidtv/vidtv_s302m.c
238
struct vidtv_s302m_ctx *ctx = e->ctx;
drivers/media/test-drivers/vidtv/vidtv_s302m.c
240
if (!e->src_buf) {
drivers/media/test-drivers/vidtv/vidtv_s302m.c
246
if (e->src_buf_offset >= ARRAY_SIZE(beethoven_fur_elise))
drivers/media/test-drivers/vidtv/vidtv_s302m.c
247
e->src_buf_offset = 0;
drivers/media/test-drivers/vidtv/vidtv_s302m.c
249
ctx->last_tone = beethoven_fur_elise[e->src_buf_offset].note;
drivers/media/test-drivers/vidtv/vidtv_s302m.c
250
ctx->last_duration = beethoven_fur_elise[e->src_buf_offset].duration *
drivers/media/test-drivers/vidtv/vidtv_s302m.c
252
e->src_buf_offset++;
drivers/media/test-drivers/vidtv/vidtv_s302m.c
269
if (e->src_buf_offset > e->src_buf_sz) {
drivers/media/test-drivers/vidtv/vidtv_s302m.c
271
e->src_buf_offset,
drivers/media/test-drivers/vidtv/vidtv_s302m.c
272
e->src_buf_sz);
drivers/media/test-drivers/vidtv/vidtv_s302m.c
274
e->src_buf_offset = 0;
drivers/media/test-drivers/vidtv/vidtv_s302m.c
277
if (e->src_buf_offset >= e->src_buf_sz) {
drivers/media/test-drivers/vidtv/vidtv_s302m.c
279
if (e->last_sample_cb)
drivers/media/test-drivers/vidtv/vidtv_s302m.c
280
e->last_sample_cb(e->sample_count);
drivers/media/test-drivers/vidtv/vidtv_s302m.c
282
e->src_buf_offset = 0;
drivers/media/test-drivers/vidtv/vidtv_s302m.c
285
sample = *(u16 *)(e->src_buf + e->src_buf_offset);
drivers/media/test-drivers/vidtv/vidtv_s302m.c
290
static u32 vidtv_s302m_write_frame(struct vidtv_encoder *e,
drivers/media/test-drivers/vidtv/vidtv_s302m.c
293
struct vidtv_s302m_ctx *ctx = e->ctx;
drivers/media/test-drivers/vidtv/vidtv_s302m.c
313
nbytes += vidtv_memcpy(e->encoder_buf,
drivers/media/test-drivers/vidtv/vidtv_s302m.c
314
e->encoder_buf_offset,
drivers/media/test-drivers/vidtv/vidtv_s302m.c
319
e->encoder_buf_offset += nbytes;
drivers/media/test-drivers/vidtv/vidtv_s302m.c
328
static u32 vidtv_s302m_write_h(struct vidtv_encoder *e, u32 p_sz)
drivers/media/test-drivers/vidtv/vidtv_s302m.c
336
nbytes += vidtv_memcpy(e->encoder_buf,
drivers/media/test-drivers/vidtv/vidtv_s302m.c
337
e->encoder_buf_offset,
drivers/media/test-drivers/vidtv/vidtv_s302m.c
338
e->encoder_buf_sz,
drivers/media/test-drivers/vidtv/vidtv_s302m.c
342
e->encoder_buf_offset += nbytes;
drivers/media/test-drivers/vidtv/vidtv_s302m.c
346
static void vidtv_s302m_write_frames(struct vidtv_encoder *e)
drivers/media/test-drivers/vidtv/vidtv_s302m.c
348
struct vidtv_access_unit *au = e->access_units;
drivers/media/test-drivers/vidtv/vidtv_s302m.c
349
struct vidtv_s302m_ctx *ctx = e->ctx;
drivers/media/test-drivers/vidtv/vidtv_s302m.c
360
nbytes_per_unit = vidtv_s302m_write_h(e, au_sz);
drivers/media/test-drivers/vidtv/vidtv_s302m.c
363
sample = vidtv_s302m_get_sample(e);
drivers/media/test-drivers/vidtv/vidtv_s302m.c
364
nbytes_per_unit += vidtv_s302m_write_frame(e, sample);
drivers/media/test-drivers/vidtv/vidtv_s302m.c
366
if (e->src_buf)
drivers/media/test-drivers/vidtv/vidtv_s302m.c
367
e->src_buf_offset += sizeof(u16);
drivers/media/test-drivers/vidtv/vidtv_s302m.c
369
e->sample_count++;
drivers/media/test-drivers/vidtv/vidtv_s302m.c
390
static void *vidtv_s302m_encode(struct vidtv_encoder *e)
drivers/media/test-drivers/vidtv/vidtv_s302m.c
392
struct vidtv_s302m_ctx *ctx = e->ctx;
drivers/media/test-drivers/vidtv/vidtv_s302m.c
407
vidtv_s302m_access_unit_destroy(e);
drivers/media/test-drivers/vidtv/vidtv_s302m.c
408
vidtv_s302m_alloc_au(e);
drivers/media/test-drivers/vidtv/vidtv_s302m.c
410
if (e->sync && e->sync->is_video_encoder) {
drivers/media/test-drivers/vidtv/vidtv_s302m.c
411
vidtv_s302m_compute_sample_count_from_video(e);
drivers/media/test-drivers/vidtv/vidtv_s302m.c
412
vidtv_s302m_compute_pts_from_video(e);
drivers/media/test-drivers/vidtv/vidtv_s302m.c
414
e->access_units->num_samples = FF_S302M_DEFAULT_NUM_FRAMES;
drivers/media/test-drivers/vidtv/vidtv_s302m.c
415
e->access_units->pts = (ctx->au_count * FF_S302M_DEFAULT_PTS_INCREMENT) +
drivers/media/test-drivers/vidtv/vidtv_s302m.c
419
vidtv_s302m_write_frames(e);
drivers/media/test-drivers/vidtv/vidtv_s302m.c
421
return e->encoder_buf;
drivers/media/test-drivers/vidtv/vidtv_s302m.c
424
static u32 vidtv_s302m_clear(struct vidtv_encoder *e)
drivers/media/test-drivers/vidtv/vidtv_s302m.c
426
struct vidtv_access_unit *au = e->access_units;
drivers/media/test-drivers/vidtv/vidtv_s302m.c
434
vidtv_s302m_access_unit_destroy(e);
drivers/media/test-drivers/vidtv/vidtv_s302m.c
435
memset(e->encoder_buf, 0, VIDTV_S302M_BUF_SZ);
drivers/media/test-drivers/vidtv/vidtv_s302m.c
436
e->encoder_buf_offset = 0;
drivers/media/test-drivers/vidtv/vidtv_s302m.c
446
struct vidtv_encoder *e;
drivers/media/test-drivers/vidtv/vidtv_s302m.c
448
e = kzalloc_obj(*e);
drivers/media/test-drivers/vidtv/vidtv_s302m.c
449
if (!e)
drivers/media/test-drivers/vidtv/vidtv_s302m.c
452
e->id = S302M;
drivers/media/test-drivers/vidtv/vidtv_s302m.c
455
e->name = kstrdup(args.name, GFP_KERNEL);
drivers/media/test-drivers/vidtv/vidtv_s302m.c
457
e->encoder_buf = vzalloc(VIDTV_S302M_BUF_SZ);
drivers/media/test-drivers/vidtv/vidtv_s302m.c
458
if (!e->encoder_buf)
drivers/media/test-drivers/vidtv/vidtv_s302m.c
461
e->encoder_buf_sz = VIDTV_S302M_BUF_SZ;
drivers/media/test-drivers/vidtv/vidtv_s302m.c
462
e->encoder_buf_offset = 0;
drivers/media/test-drivers/vidtv/vidtv_s302m.c
464
e->sample_count = 0;
drivers/media/test-drivers/vidtv/vidtv_s302m.c
466
e->src_buf = (args.src_buf) ? args.src_buf : NULL;
drivers/media/test-drivers/vidtv/vidtv_s302m.c
467
e->src_buf_sz = (args.src_buf) ? args.src_buf_sz : 0;
drivers/media/test-drivers/vidtv/vidtv_s302m.c
468
e->src_buf_offset = 0;
drivers/media/test-drivers/vidtv/vidtv_s302m.c
470
e->is_video_encoder = false;
drivers/media/test-drivers/vidtv/vidtv_s302m.c
476
e->ctx = ctx;
drivers/media/test-drivers/vidtv/vidtv_s302m.c
479
e->encode = vidtv_s302m_encode;
drivers/media/test-drivers/vidtv/vidtv_s302m.c
480
e->clear = vidtv_s302m_clear;
drivers/media/test-drivers/vidtv/vidtv_s302m.c
482
e->es_pid = cpu_to_be16(args.es_pid);
drivers/media/test-drivers/vidtv/vidtv_s302m.c
483
e->stream_id = cpu_to_be16(PES_PRIVATE_STREAM_1);
drivers/media/test-drivers/vidtv/vidtv_s302m.c
485
e->sync = args.sync;
drivers/media/test-drivers/vidtv/vidtv_s302m.c
486
e->sampling_rate_hz = S302M_SAMPLING_RATE_HZ;
drivers/media/test-drivers/vidtv/vidtv_s302m.c
488
e->last_sample_cb = args.last_sample_cb;
drivers/media/test-drivers/vidtv/vidtv_s302m.c
490
e->destroy = vidtv_s302m_encoder_destroy;
drivers/media/test-drivers/vidtv/vidtv_s302m.c
496
args.head->next = e;
drivers/media/test-drivers/vidtv/vidtv_s302m.c
499
e->next = NULL;
drivers/media/test-drivers/vidtv/vidtv_s302m.c
501
return e;
drivers/media/test-drivers/vidtv/vidtv_s302m.c
504
vfree(e->encoder_buf);
drivers/media/test-drivers/vidtv/vidtv_s302m.c
507
kfree(e->name);
drivers/media/test-drivers/vidtv/vidtv_s302m.c
508
kfree(e);
drivers/media/test-drivers/vidtv/vidtv_s302m.c
512
void vidtv_s302m_encoder_destroy(struct vidtv_encoder *e)
drivers/media/test-drivers/vidtv/vidtv_s302m.c
514
if (e->id != S302M) {
drivers/media/test-drivers/vidtv/vidtv_s302m.c
519
vidtv_s302m_access_unit_destroy(e);
drivers/media/test-drivers/vidtv/vidtv_s302m.c
520
kfree(e->name);
drivers/media/test-drivers/vidtv/vidtv_s302m.c
521
vfree(e->encoder_buf);
drivers/media/test-drivers/vidtv/vidtv_s302m.c
522
kfree(e->ctx);
drivers/media/test-drivers/vidtv/vidtv_s302m.c
523
kfree(e);
drivers/media/test-drivers/visl/visl-trace-h264.h
272
TP_PROTO(const struct v4l2_h264_dpb_entry *e, int i),
drivers/media/test-drivers/visl/visl-trace-h264.h
273
TP_ARGS(e, i),
drivers/media/test-drivers/visl/visl-trace-h264.h
274
TP_STRUCT__entry(__field_struct(struct v4l2_h264_dpb_entry, e)
drivers/media/test-drivers/visl/visl-trace-h264.h
276
TP_fast_assign(__entry->e = *e; __entry->i = i;),
drivers/media/test-drivers/visl/visl-trace-h264.h
280
__entry->e.reference_ts,
drivers/media/test-drivers/visl/visl-trace-h264.h
281
__entry->e.pic_num,
drivers/media/test-drivers/visl/visl-trace-h264.h
282
__entry->e.frame_num,
drivers/media/test-drivers/visl/visl-trace-h264.h
283
__print_flags(__entry->e.fields, "|",
drivers/media/test-drivers/visl/visl-trace-h264.h
287
__entry->e.top_field_order_cnt,
drivers/media/test-drivers/visl/visl-trace-h264.h
288
__entry->e.bottom_field_order_cnt,
drivers/media/test-drivers/visl/visl-trace-h264.h
289
__print_flags(__entry->e.flags, "|",
drivers/media/test-drivers/visl/visl-trace-h264.h
339
TP_PROTO(const struct v4l2_h264_dpb_entry *e, int i),
drivers/media/test-drivers/visl/visl-trace-h264.h
340
TP_ARGS(e, i)
drivers/media/test-drivers/visl/visl-trace-hevc.h
397
TP_PROTO(const struct v4l2_hevc_dpb_entry *e),
drivers/media/test-drivers/visl/visl-trace-hevc.h
398
TP_ARGS(e),
drivers/media/test-drivers/visl/visl-trace-hevc.h
399
TP_STRUCT__entry(__field_struct(struct v4l2_hevc_dpb_entry, e)),
drivers/media/test-drivers/visl/visl-trace-hevc.h
400
TP_fast_assign(__entry->e = *e),
drivers/media/test-drivers/visl/visl-trace-hevc.h
405
__entry->e.timestamp,
drivers/media/test-drivers/visl/visl-trace-hevc.h
406
__print_flags(__entry->e.flags, "|",
drivers/media/test-drivers/visl/visl-trace-hevc.h
409
__entry->e.field_pic,
drivers/media/test-drivers/visl/visl-trace-hevc.h
410
__entry->e.pic_order_cnt_val
drivers/media/test-drivers/visl/visl-trace-hevc.h
454
TP_PROTO(const struct v4l2_hevc_dpb_entry *e),
drivers/media/test-drivers/visl/visl-trace-hevc.h
455
TP_ARGS(e)
drivers/media/tuners/tda9887.c
146
.e = ( cGating_36 |
drivers/media/tuners/tda9887.c
157
.e = ( cGating_36 |
drivers/media/tuners/tda9887.c
168
.e = ( cGating_36 |
drivers/media/tuners/tda9887.c
179
.e = ( cGating_36 |
drivers/media/tuners/tda9887.c
188
.e = ( cAudioIF_5_5 |
drivers/media/tuners/tda9887.c
196
.e = ( cGating_36 |
drivers/media/tuners/tda9887.c
206
.e = ( cGating_36 |
drivers/media/tuners/tda9887.c
217
.e = ( cGating_36 |
drivers/media/tuners/tda9887.c
228
.e = ( cGating_36 |
drivers/media/tuners/tda9887.c
239
.e = ( cGating_36 |
drivers/media/tuners/tda9887.c
252
.e = ( cTunerGainLow |
drivers/media/tuners/tda9887.c
264
.e = ( cTunerGainLow |
drivers/media/tuners/tda9887.c
429
buf[3] = norm->e;
drivers/media/tuners/tda9887.c
53
unsigned char e;
drivers/media/usb/cx231xx/cx231xx-cards.c
1298
struct eeprom *e = kzalloc_obj(*e);
drivers/media/usb/cx231xx/cx231xx-cards.c
1300
if (e == NULL) {
drivers/media/usb/cx231xx/cx231xx-cards.c
1305
e->client.adapter = cx231xx_get_i2c_adap(dev, I2C_1_MUX_1);
drivers/media/usb/cx231xx/cx231xx-cards.c
1306
e->client.addr = 0xa0 >> 1;
drivers/media/usb/cx231xx/cx231xx-cards.c
1308
read_eeprom(dev, &e->client, e->eeprom, sizeof(e->eeprom));
drivers/media/usb/cx231xx/cx231xx-cards.c
1309
tveeprom_hauppauge_analog(&e->tvee, e->eeprom + 0xc0);
drivers/media/usb/cx231xx/cx231xx-cards.c
1310
kfree(e);
drivers/media/usb/dvb-usb/dib0700_core.c
811
const struct usb_endpoint_descriptor *e;
drivers/media/usb/dvb-usb/dib0700_core.c
841
e = &intf->cur_altsetting->endpoint[rc_ep].desc;
drivers/media/usb/dvb-usb/dib0700_core.c
842
if (usb_endpoint_dir_in(e)) {
drivers/media/usb/dvb-usb/dib0700_core.c
843
if (usb_endpoint_xfer_bulk(e)) {
drivers/media/usb/dvb-usb/dib0700_core.c
850
} else if (usb_endpoint_xfer_int(e)) {
drivers/media/usb/em28xx/em28xx-audio.c
648
struct usb_endpoint_descriptor *e)
drivers/media/usb/em28xx/em28xx-audio.c
650
int size = le16_to_cpu(e->wMaxPacketSize);
drivers/media/usb/em28xx/em28xx-audio.c
661
struct usb_endpoint_descriptor *e, *ep = NULL;
drivers/media/usb/em28xx/em28xx-audio.c
681
e = &intf->altsetting[alt].endpoint[i].desc;
drivers/media/usb/em28xx/em28xx-audio.c
682
if (!usb_endpoint_dir_in(e))
drivers/media/usb/em28xx/em28xx-audio.c
684
if (e->bEndpointAddress == EM28XX_EP_AUDIO) {
drivers/media/usb/em28xx/em28xx-audio.c
685
ep = e;
drivers/media/usb/em28xx/em28xx-cards.c
3768
const struct usb_endpoint_descriptor *e;
drivers/media/usb/em28xx/em28xx-cards.c
3796
e = &intf->altsetting[alt].endpoint[ep].desc;
drivers/media/usb/em28xx/em28xx-cards.c
3798
if (!usb_endpoint_dir_in(e))
drivers/media/usb/em28xx/em28xx-cards.c
3801
sizedescr = le16_to_cpu(e->wMaxPacketSize);
drivers/media/usb/em28xx/em28xx-cards.c
3809
switch (e->bEndpointAddress) {
drivers/media/usb/em28xx/em28xx-cards.c
3812
if (usb_endpoint_xfer_isoc(e)) {
drivers/media/usb/em28xx/em28xx-cards.c
3813
dev->analog_ep_isoc = e->bEndpointAddress;
drivers/media/usb/em28xx/em28xx-cards.c
3815
} else if (usb_endpoint_xfer_bulk(e)) {
drivers/media/usb/em28xx/em28xx-cards.c
3816
dev->analog_ep_bulk = e->bEndpointAddress;
drivers/media/usb/em28xx/em28xx-cards.c
3820
if (usb_endpoint_xfer_isoc(e))
drivers/media/usb/em28xx/em28xx-cards.c
3827
if (*has_video && (usb_endpoint_xfer_bulk(e))) {
drivers/media/usb/em28xx/em28xx-cards.c
3828
dev->analog_ep_bulk = e->bEndpointAddress;
drivers/media/usb/em28xx/em28xx-cards.c
3830
if (usb_endpoint_xfer_isoc(e)) {
drivers/media/usb/em28xx/em28xx-cards.c
3840
dev->dvb_ep_isoc = e->bEndpointAddress;
drivers/media/usb/em28xx/em28xx-cards.c
3846
dev->dvb_ep_bulk = e->bEndpointAddress;
drivers/media/usb/em28xx/em28xx-cards.c
3851
if (usb_endpoint_xfer_isoc(e)) {
drivers/media/usb/em28xx/em28xx-cards.c
3853
dev->dvb_ep_isoc_ts2 = e->bEndpointAddress;
drivers/media/usb/em28xx/em28xx-cards.c
3858
dev->dvb_ep_bulk_ts2 = e->bEndpointAddress;
drivers/media/usb/gspca/cpia1.c
563
u8 e, u8 f, u8 g, u8 h,
drivers/media/usb/gspca/cpia1.c
576
gspca_dev->usb_buf[0] = e;
drivers/media/usb/stk1160/stk1160-core.c
177
int i, e, sizedescr, size, ifnum;
drivers/media/usb/stk1160/stk1160-core.c
188
for (e = 0; e < intf->altsetting[i].desc.bNumEndpoints; e++) {
drivers/media/usb/stk1160/stk1160-core.c
191
desc = &intf->altsetting[i].endpoint[e].desc;
drivers/mfd/twl-core.c
613
int e = 0;
drivers/mfd/twl-core.c
615
e = twl_i2c_write_u8(TWL_MODULE_PM_MASTER, 0,
drivers/mfd/twl-core.c
617
return e;
drivers/mfd/twl-core.c
622
int e = 0;
drivers/mfd/twl-core.c
624
e |= twl_i2c_write_u8(TWL_MODULE_PM_MASTER, TWL4030_PM_MASTER_KEY_CFG1,
drivers/mfd/twl-core.c
626
e |= twl_i2c_write_u8(TWL_MODULE_PM_MASTER, TWL4030_PM_MASTER_KEY_CFG2,
drivers/mfd/twl-core.c
629
return e;
drivers/mfd/twl-core.c
634
int e = 0;
drivers/mfd/twl-core.c
663
e |= unprotect_pm_master();
drivers/mfd/twl-core.c
665
e |= twl_i2c_write_u8(TWL_MODULE_PM_MASTER, ctrl, R_CFG_BOOT);
drivers/mfd/twl-core.c
666
e |= protect_pm_master();
drivers/mfd/twl-core.c
668
if (e < 0)
drivers/mfd/twl-core.c
669
pr_err("%s: clock init err [%d]\n", DRIVER_NAME, e);
drivers/misc/bcm-vk/bcm_vk_dev.c
1270
unsigned long e, void *p)
drivers/misc/bcm-vk/bcm_vk_tty.c
32
#define VK_BAR_CHAN(v, DIR, e) ((v)->DIR##_offset \
drivers/misc/bcm-vk/bcm_vk_tty.c
33
+ offsetof(struct bcm_vk_tty_chan, e))
drivers/misc/genwqe/card_base.c
385
unsigned int type, e = 0;
drivers/misc/genwqe/card_base.c
390
e = genwqe_ffdc_buff_size(cd, 0);
drivers/misc/genwqe/card_base.c
393
e = genwqe_ffdc_buff_size(cd, 1);
drivers/misc/genwqe/card_base.c
396
e = genwqe_ffdc_buff_size(cd, 2);
drivers/misc/genwqe/card_base.c
399
e = GENWQE_FFDC_REGS;
drivers/misc/genwqe/card_base.c
404
cd->ffdc[type].entries = e;
drivers/misc/genwqe/card_base.c
406
kmalloc_objs(struct genwqe_reg, e);
drivers/misc/genwqe/card_utils.c
893
u64 eevptr, e, val, addr;
drivers/misc/genwqe/card_utils.c
901
e = __genwqe_readq(cd, l_addr);
drivers/misc/genwqe/card_utils.c
902
if ((e == 0x0) || (e == 0xffffffffffffffffull))
drivers/misc/genwqe/card_utils.c
905
d_addr = (e & 0x0000000000ffffffull); /* 23:0 */
drivers/misc/genwqe/card_utils.c
906
d_len = (e & 0x0000007fff000000ull) >> 24; /* 38:24 */
drivers/misc/genwqe/card_utils.c
907
d_type = (e & 0x0000008000000000ull) >> 36; /* 39 */
drivers/misc/ocxl/config.c
9
#define EXTRACT_BITS(val, s, e) ((val & GENMASK(e, s)) >> s)
drivers/misc/vmw_vmci/vmci_event.c
46
int e;
drivers/misc/vmw_vmci/vmci_event.c
49
for (e = 0; e < VMCI_EVENT_MAX; e++) {
drivers/misc/vmw_vmci/vmci_event.c
51
list_for_each_entry_safe(cur, p2, &subscriber_array[e], node) {
drivers/misc/vmw_vmci/vmci_event.c
70
int e;
drivers/misc/vmw_vmci/vmci_event.c
72
for (e = 0; e < VMCI_EVENT_MAX; e++) {
drivers/misc/vmw_vmci/vmci_event.c
74
list_for_each_entry(cur, &subscriber_array[e], node) {
drivers/mmc/core/mmc.c
153
unsigned int e, m, a, b;
drivers/mmc/core/mmc.c
170
e = unstuff_bits(resp, 112, 3);
drivers/mmc/core/mmc.c
171
csd->taac_ns = (taac_exp[e] * taac_mant[m] + 9) / 10;
drivers/mmc/core/mmc.c
175
e = unstuff_bits(resp, 96, 3);
drivers/mmc/core/mmc.c
176
csd->max_dtr = tran_exp[e] * tran_mant[m];
drivers/mmc/core/mmc.c
179
e = unstuff_bits(resp, 47, 3);
drivers/mmc/core/mmc.c
181
csd->capacity = (1 + m) << (e + 2);
drivers/mmc/core/sd.c
110
unsigned int e, m, csd_struct;
drivers/mmc/core/sd.c
118
e = unstuff_bits(resp, 112, 3);
drivers/mmc/core/sd.c
119
csd->taac_ns = (taac_exp[e] * taac_mant[m] + 9) / 10;
drivers/mmc/core/sd.c
123
e = unstuff_bits(resp, 96, 3);
drivers/mmc/core/sd.c
124
csd->max_dtr = tran_exp[e] * tran_mant[m];
drivers/mmc/core/sd.c
127
e = unstuff_bits(resp, 47, 3);
drivers/mmc/core/sd.c
129
csd->capacity = (1 + m) << (e + 2);
drivers/mmc/core/sd.c
164
e = unstuff_bits(resp, 96, 3);
drivers/mmc/core/sd.c
165
csd->max_dtr = tran_exp[e] * tran_mant[m];
drivers/mmc/host/cb710-mmc.c
152
u32 e, x;
drivers/mmc/host/cb710-mmc.c
153
e = cb710_read_port_32(slot, CB710_MMC_STATUS_PORT);
drivers/mmc/host/cb710-mmc.c
173
limit, what, e, x);
drivers/mmc/host/cb710-mmc.c
185
u32 e, x;
drivers/mmc/host/cb710-mmc.c
186
e = cb710_read_port_32(slot, CB710_MMC_STATUS_PORT);
drivers/mmc/host/cb710-mmc.c
206
limit, mask, e, x);
drivers/mtd/nand/raw/rockchip-nand-controller.c
64
#define ECC_ERR_CNT(x, e) ((((x) >> (e).low) & (e).low_mask) |\
drivers/mtd/nand/raw/rockchip-nand-controller.c
65
(((x) >> (e).high) & (e).high_mask) << (e).low_bn)
drivers/mtd/ubi/fastmap-wl.c
189
struct ubi_wl_entry *e;
drivers/mtd/ubi/fastmap-wl.c
226
e = wl_get_wle(ubi);
drivers/mtd/ubi/fastmap-wl.c
227
if (!e)
drivers/mtd/ubi/fastmap-wl.c
230
pool->pebs[pool->size] = e->pnum;
drivers/mtd/ubi/fastmap-wl.c
241
e = find_wl_entry(ubi, &ubi->free, WL_FREE_MAX_DIFF,
drivers/mtd/ubi/fastmap-wl.c
243
self_check_in_wl_tree(ubi, e, &ubi->free);
drivers/mtd/ubi/fastmap-wl.c
244
rb_erase(&e->u.rb, &ubi->free);
drivers/mtd/ubi/fastmap-wl.c
247
wl_pool->pebs[wl_pool->size] = e->pnum;
drivers/mtd/ubi/fastmap-wl.c
29
struct ubi_wl_entry *e, *victim = NULL;
drivers/mtd/ubi/fastmap-wl.c
32
ubi_rb_for_each_entry(p, e, root, u.rb) {
drivers/mtd/ubi/fastmap-wl.c
33
if (e->pnum < UBI_FM_MAX_START && e->ec < max_ec) {
drivers/mtd/ubi/fastmap-wl.c
34
victim = e;
drivers/mtd/ubi/fastmap-wl.c
35
max_ec = e->ec;
drivers/mtd/ubi/fastmap-wl.c
386
struct ubi_wl_entry *e;
drivers/mtd/ubi/fastmap-wl.c
391
e = next_peb_for_wl(ubi, false);
drivers/mtd/ubi/fastmap-wl.c
392
if (!e) {
drivers/mtd/ubi/fastmap-wl.c
395
e = find_wl_entry(ubi, &ubi->free, WL_FREE_MAX_DIFF, 0);
drivers/mtd/ubi/fastmap-wl.c
396
ec = e->ec;
drivers/mtd/ubi/fastmap-wl.c
398
ec = e->ec;
drivers/mtd/ubi/fastmap-wl.c
400
e = find_wl_entry(ubi, &ubi->free, WL_FREE_MAX_DIFF, 0);
drivers/mtd/ubi/fastmap-wl.c
401
ec = max(ec, e->ec);
drivers/mtd/ubi/fastmap-wl.c
404
e = rb_entry(rb_first(&ubi->used), struct ubi_wl_entry, u.rb);
drivers/mtd/ubi/fastmap-wl.c
406
return ec - e->ec >= UBI_WL_THRESHOLD;
drivers/mtd/ubi/fastmap-wl.c
43
struct ubi_wl_entry *e)
drivers/mtd/ubi/fastmap-wl.c
45
wl_tree_add(e, &ubi->free);
drivers/mtd/ubi/fastmap-wl.c
495
struct ubi_wl_entry *e;
drivers/mtd/ubi/fastmap-wl.c
504
e = ubi->lookuptbl[pnum];
drivers/mtd/ubi/fastmap-wl.c
510
if (!e) {
drivers/mtd/ubi/fastmap-wl.c
511
e = fm_e;
drivers/mtd/ubi/fastmap-wl.c
512
ubi_assert(e->ec >= 0);
drivers/mtd/ubi/fastmap-wl.c
513
ubi->lookuptbl[pnum] = e;
drivers/mtd/ubi/fastmap-wl.c
519
return schedule_erase(ubi, e, vol_id, lnum, torture, true);
drivers/mtd/ubi/fastmap-wl.c
553
struct ubi_wl_entry *e,
drivers/mtd/ubi/fastmap-wl.c
555
if (e && !ubi->fm_disabled && !ubi->fm && !ubi->fm_anchor &&
drivers/mtd/ubi/fastmap-wl.c
556
e->pnum < UBI_FM_MAX_START)
drivers/mtd/ubi/fastmap-wl.c
557
e = rb_entry(rb_next(root->rb_node),
drivers/mtd/ubi/fastmap-wl.c
560
return e;
drivers/mtd/ubi/fastmap-wl.c
58
struct ubi_wl_entry *e;
drivers/mtd/ubi/fastmap-wl.c
61
e = ubi->lookuptbl[pool->pebs[i]];
drivers/mtd/ubi/fastmap-wl.c
62
return_unused_peb(ubi, e);
drivers/mtd/ubi/fastmap-wl.c
77
struct ubi_wl_entry *e = NULL;
drivers/mtd/ubi/fastmap-wl.c
83
e = find_anchor_wl_entry(&ubi->free);
drivers/mtd/ubi/fastmap-wl.c
85
e = find_mean_wl_entry(ubi, &ubi->free);
drivers/mtd/ubi/fastmap-wl.c
87
if (!e)
drivers/mtd/ubi/fastmap-wl.c
90
self_check_in_wl_tree(ubi, e, &ubi->free);
drivers/mtd/ubi/fastmap-wl.c
94
rb_erase(&e->u.rb, &ubi->free);
drivers/mtd/ubi/fastmap-wl.c
97
return e;
drivers/mtd/ubi/fastmap.c
1061
struct ubi_wl_entry *e;
drivers/mtd/ubi/fastmap.c
1063
e = kmem_cache_alloc(ubi_wl_entry_slab, GFP_KERNEL);
drivers/mtd/ubi/fastmap.c
1064
if (!e) {
drivers/mtd/ubi/fastmap.c
1066
kmem_cache_free(ubi_wl_entry_slab, fm->e[i]);
drivers/mtd/ubi/fastmap.c
1072
e->pnum = be32_to_cpu(fmsb2->block_loc[i]);
drivers/mtd/ubi/fastmap.c
1073
e->ec = be32_to_cpu(fmsb2->block_ec[i]);
drivers/mtd/ubi/fastmap.c
1074
fm->e[i] = e;
drivers/mtd/ubi/fastmap.c
1275
wl_e = ubi_wrk->e;
drivers/mtd/ubi/fastmap.c
1336
dbg_bld("writing fastmap SB to PEB %i", new_fm->e[0]->pnum);
drivers/mtd/ubi/fastmap.c
1337
ret = ubi_io_write_vid_hdr(ubi, new_fm->e[0]->pnum, avbuf);
drivers/mtd/ubi/fastmap.c
1344
fmsb->block_loc[i] = cpu_to_be32(new_fm->e[i]->pnum);
drivers/mtd/ubi/fastmap.c
1345
set_seen(ubi, new_fm->e[i]->pnum, seen_pebs);
drivers/mtd/ubi/fastmap.c
1346
fmsb->block_ec[i] = cpu_to_be32(new_fm->e[i]->ec);
drivers/mtd/ubi/fastmap.c
1357
new_fm->e[i]->pnum, be64_to_cpu(dvhdr->sqnum));
drivers/mtd/ubi/fastmap.c
1358
ret = ubi_io_write_vid_hdr(ubi, new_fm->e[i]->pnum, dvbuf);
drivers/mtd/ubi/fastmap.c
1361
new_fm->e[i]->pnum);
drivers/mtd/ubi/fastmap.c
1368
new_fm->e[i]->pnum, 0, ubi->leb_size);
drivers/mtd/ubi/fastmap.c
1371
new_fm->e[i]->pnum);
drivers/mtd/ubi/fastmap.c
1409
struct ubi_wl_entry *e;
drivers/mtd/ubi/fastmap.c
1430
e = ubi_wl_get_fm_peb(ubi, 1);
drivers/mtd/ubi/fastmap.c
1431
if (!e)
drivers/mtd/ubi/fastmap.c
1439
ret = ubi_io_write_vid_hdr(ubi, e->pnum, vb);
drivers/mtd/ubi/fastmap.c
1441
ubi_wl_put_fm_peb(ubi, e, 0, 0);
drivers/mtd/ubi/fastmap.c
1446
fm->e[0] = e;
drivers/mtd/ubi/fastmap.c
1474
if (fm->e[i]) {
drivers/mtd/ubi/fastmap.c
1475
ubi_wl_put_fm_peb(ubi, fm->e[i], i,
drivers/mtd/ubi/fastmap.c
1477
fm->e[i] = NULL;
drivers/mtd/ubi/fastmap.c
1528
if (old_fm && old_fm->e[i]) {
drivers/mtd/ubi/fastmap.c
1529
ret = ubi_sync_erase(ubi, old_fm->e[i], 0);
drivers/mtd/ubi/fastmap.c
1534
ubi_wl_put_fm_peb(ubi, new_fm->e[j],
drivers/mtd/ubi/fastmap.c
1536
new_fm->e[j] = NULL;
drivers/mtd/ubi/fastmap.c
1540
new_fm->e[i] = old_fm->e[i];
drivers/mtd/ubi/fastmap.c
1541
old_fm->e[i] = NULL;
drivers/mtd/ubi/fastmap.c
1546
ubi_wl_put_fm_peb(ubi, new_fm->e[j], j, 0);
drivers/mtd/ubi/fastmap.c
1547
new_fm->e[j] = NULL;
drivers/mtd/ubi/fastmap.c
1554
new_fm->e[i] = tmp_e;
drivers/mtd/ubi/fastmap.c
1556
if (old_fm && old_fm->e[i]) {
drivers/mtd/ubi/fastmap.c
1557
ubi_wl_put_fm_peb(ubi, old_fm->e[i], i,
drivers/mtd/ubi/fastmap.c
1559
old_fm->e[i] = NULL;
drivers/mtd/ubi/fastmap.c
1567
ubi_wl_put_fm_peb(ubi, old_fm->e[i], i,
drivers/mtd/ubi/fastmap.c
1569
old_fm->e[i] = NULL;
drivers/mtd/ubi/fastmap.c
1581
ret = ubi_sync_erase(ubi, old_fm->e[0], 0);
drivers/mtd/ubi/fastmap.c
1586
ubi_wl_put_fm_peb(ubi, new_fm->e[i],
drivers/mtd/ubi/fastmap.c
1588
new_fm->e[i] = NULL;
drivers/mtd/ubi/fastmap.c
1592
new_fm->e[0] = old_fm->e[0];
drivers/mtd/ubi/fastmap.c
1593
old_fm->e[0] = NULL;
drivers/mtd/ubi/fastmap.c
1596
ubi_wl_put_fm_peb(ubi, old_fm->e[0], 0,
drivers/mtd/ubi/fastmap.c
1598
new_fm->e[0] = tmp_e;
drivers/mtd/ubi/fastmap.c
1599
old_fm->e[0] = NULL;
drivers/mtd/ubi/fastmap.c
1606
ubi_wl_put_fm_peb(ubi, new_fm->e[i], i, 0);
drivers/mtd/ubi/fastmap.c
1607
new_fm->e[i] = NULL;
drivers/mtd/ubi/fastmap.c
1613
new_fm->e[0] = tmp_e;
drivers/mtd/ubi/ubi.h
1015
#define ubi_for_each_free_peb(ubi, e, tmp_rb) \
drivers/mtd/ubi/ubi.h
1016
ubi_rb_for_each_entry((tmp_rb), (e), &(ubi)->free, u.rb)
drivers/mtd/ubi/ubi.h
1024
#define ubi_for_each_used_peb(ubi, e, tmp_rb) \
drivers/mtd/ubi/ubi.h
1025
ubi_rb_for_each_entry((tmp_rb), (e), &(ubi)->used, u.rb)
drivers/mtd/ubi/ubi.h
1033
#define ubi_for_each_scrub_peb(ubi, e, tmp_rb) \
drivers/mtd/ubi/ubi.h
1034
ubi_rb_for_each_entry((tmp_rb), (e), &(ubi)->scrub, u.rb)
drivers/mtd/ubi/ubi.h
1042
#define ubi_for_each_protected_peb(ubi, i, e) \
drivers/mtd/ubi/ubi.h
1044
list_for_each_entry((e), &(ubi->pq[(i)]), u.list)
drivers/mtd/ubi/ubi.h
1239
if (ubi->fm->e[i]->pnum == pnum)
drivers/mtd/ubi/ubi.h
1240
return ubi->fm->e[i];
drivers/mtd/ubi/ubi.h
233
struct ubi_wl_entry *e[UBI_FM_MAX_BLOCKS];
drivers/mtd/ubi/ubi.h
805
struct ubi_wl_entry *e;
drivers/mtd/ubi/ubi.h
904
int ubi_sync_erase(struct ubi_device *ubi, struct ubi_wl_entry *e, int torture);
drivers/mtd/ubi/ubi.h
978
kmem_cache_free(ubi_wl_entry_slab, ubi->fm->e[i]);
drivers/mtd/ubi/wl.c
1107
struct ubi_wl_entry *e = wl_wrk->e;
drivers/mtd/ubi/wl.c
1108
int pnum = e->pnum;
drivers/mtd/ubi/wl.c
1114
pnum, e->ec, wl_wrk->vol_id, wl_wrk->lnum);
drivers/mtd/ubi/wl.c
1116
err = ubi_sync_erase(ubi, e, wl_wrk->torture);
drivers/mtd/ubi/wl.c
1121
e->pnum < UBI_FM_MAX_START) {
drivers/mtd/ubi/wl.c
1126
ubi->fm_anchor = e;
drivers/mtd/ubi/wl.c
1129
wl_tree_add(e, &ubi->free);
drivers/mtd/ubi/wl.c
1153
err1 = schedule_erase(ubi, e, vol_id, lnum, 0, true);
drivers/mtd/ubi/wl.c
1156
wl_entry_destroy(ubi, e);
drivers/mtd/ubi/wl.c
1165
wl_entry_destroy(ubi, e);
drivers/mtd/ubi/wl.c
1241
struct ubi_wl_entry *e = wl_wrk->e;
drivers/mtd/ubi/wl.c
1243
dbg_wl("cancel erasure of PEB %d EC %d", e->pnum, e->ec);
drivers/mtd/ubi/wl.c
1245
wl_entry_destroy(ubi, e);
drivers/mtd/ubi/wl.c
127
struct ubi_wl_entry *e, struct rb_root *root);
drivers/mtd/ubi/wl.c
1271
struct ubi_wl_entry *e;
drivers/mtd/ubi/wl.c
1281
e = ubi->lookuptbl[pnum];
drivers/mtd/ubi/wl.c
1282
if (!e) {
drivers/mtd/ubi/wl.c
129
struct ubi_wl_entry *e);
drivers/mtd/ubi/wl.c
1294
if (e == ubi->move_from) {
drivers/mtd/ubi/wl.c
1307
} else if (e == ubi->move_to) {
drivers/mtd/ubi/wl.c
1324
if (in_wl_tree(e, &ubi->used)) {
drivers/mtd/ubi/wl.c
1325
self_check_in_wl_tree(ubi, e, &ubi->used);
drivers/mtd/ubi/wl.c
1326
rb_erase(&e->u.rb, &ubi->used);
drivers/mtd/ubi/wl.c
1327
} else if (in_wl_tree(e, &ubi->scrub)) {
drivers/mtd/ubi/wl.c
1328
self_check_in_wl_tree(ubi, e, &ubi->scrub);
drivers/mtd/ubi/wl.c
1329
rb_erase(&e->u.rb, &ubi->scrub);
drivers/mtd/ubi/wl.c
1330
} else if (in_wl_tree(e, &ubi->erroneous)) {
drivers/mtd/ubi/wl.c
1331
self_check_in_wl_tree(ubi, e, &ubi->erroneous);
drivers/mtd/ubi/wl.c
1332
rb_erase(&e->u.rb, &ubi->erroneous);
drivers/mtd/ubi/wl.c
1338
err = prot_queue_del(ubi, e->pnum);
drivers/mtd/ubi/wl.c
1350
err = schedule_erase(ubi, e, vol_id, lnum, torture, false);
drivers/mtd/ubi/wl.c
1353
wl_tree_add(e, &ubi->used);
drivers/mtd/ubi/wl.c
1373
struct ubi_wl_entry *e;
drivers/mtd/ubi/wl.c
1379
e = ubi->lookuptbl[pnum];
drivers/mtd/ubi/wl.c
1380
if (e == ubi->move_from || in_wl_tree(e, &ubi->scrub) ||
drivers/mtd/ubi/wl.c
1381
in_wl_tree(e, &ubi->erroneous)) {
drivers/mtd/ubi/wl.c
1386
if (e == ubi->move_to) {
drivers/mtd/ubi/wl.c
139
static void wl_tree_add(struct ubi_wl_entry *e, struct rb_root *root)
drivers/mtd/ubi/wl.c
1399
if (in_wl_tree(e, &ubi->used)) {
drivers/mtd/ubi/wl.c
1400
self_check_in_wl_tree(ubi, e, &ubi->used);
drivers/mtd/ubi/wl.c
1401
rb_erase(&e->u.rb, &ubi->used);
drivers/mtd/ubi/wl.c
1405
err = prot_queue_del(ubi, e->pnum);
drivers/mtd/ubi/wl.c
1414
wl_tree_add(e, &ubi->scrub);
drivers/mtd/ubi/wl.c
1487
static bool scrub_possible(struct ubi_device *ubi, struct ubi_wl_entry *e)
drivers/mtd/ubi/wl.c
1489
if (in_wl_tree(e, &ubi->scrub))
drivers/mtd/ubi/wl.c
1491
else if (in_wl_tree(e, &ubi->erroneous))
drivers/mtd/ubi/wl.c
1493
else if (ubi->move_from == e)
drivers/mtd/ubi/wl.c
1495
else if (ubi->move_to == e)
drivers/mtd/ubi/wl.c
150
if (e->ec < e1->ec)
drivers/mtd/ubi/wl.c
152
else if (e->ec > e1->ec)
drivers/mtd/ubi/wl.c
1523
struct ubi_wl_entry *e;
drivers/mtd/ubi/wl.c
1541
e = ubi->lookuptbl[pnum];
drivers/mtd/ubi/wl.c
1542
if (!e) {
drivers/mtd/ubi/wl.c
155
ubi_assert(e->pnum != e1->pnum);
drivers/mtd/ubi/wl.c
1551
if (!scrub_possible(ubi, e)) {
drivers/mtd/ubi/wl.c
156
if (e->pnum < e1->pnum)
drivers/mtd/ubi/wl.c
1574
e = ubi->lookuptbl[pnum];
drivers/mtd/ubi/wl.c
1575
if (!e) {
drivers/mtd/ubi/wl.c
1584
if (!scrub_possible(ubi, e)) {
drivers/mtd/ubi/wl.c
1590
if (in_pq(ubi, e)) {
drivers/mtd/ubi/wl.c
1591
prot_queue_del(ubi, e->pnum);
drivers/mtd/ubi/wl.c
1592
wl_tree_add(e, &ubi->scrub);
drivers/mtd/ubi/wl.c
1596
} else if (in_wl_tree(e, &ubi->used)) {
drivers/mtd/ubi/wl.c
1597
rb_erase(&e->u.rb, &ubi->used);
drivers/mtd/ubi/wl.c
1598
wl_tree_add(e, &ubi->scrub);
drivers/mtd/ubi/wl.c
1602
} else if (in_wl_tree(e, &ubi->free)) {
drivers/mtd/ubi/wl.c
1603
rb_erase(&e->u.rb, &ubi->free);
drivers/mtd/ubi/wl.c
1611
err = schedule_erase(ubi, e, UBI_UNKNOWN, UBI_UNKNOWN,
drivers/mtd/ubi/wl.c
163
rb_link_node(&e->u.rb, parent, p);
drivers/mtd/ubi/wl.c
1639
struct ubi_wl_entry *e;
drivers/mtd/ubi/wl.c
164
rb_insert_color(&e->u.rb, root);
drivers/mtd/ubi/wl.c
1648
e = rb_entry(rb, struct ubi_wl_entry, u.rb);
drivers/mtd/ubi/wl.c
1652
if (rb->rb_left == &e->u.rb)
drivers/mtd/ubi/wl.c
1658
wl_entry_destroy(ubi, e);
drivers/mtd/ubi/wl.c
175
static void wl_entry_destroy(struct ubi_device *ubi, struct ubi_wl_entry *e)
drivers/mtd/ubi/wl.c
1759
struct ubi_wl_entry *e;
drivers/mtd/ubi/wl.c
1762
e = kmem_cache_alloc(ubi_wl_entry_slab, GFP_KERNEL);
drivers/mtd/ubi/wl.c
1763
if (!e)
drivers/mtd/ubi/wl.c
1766
e->pnum = aeb->pnum;
drivers/mtd/ubi/wl.c
1767
e->ec = aeb->ec;
drivers/mtd/ubi/wl.c
1768
ubi->lookuptbl[e->pnum] = e;
drivers/mtd/ubi/wl.c
177
ubi->lookuptbl[e->pnum] = NULL;
drivers/mtd/ubi/wl.c
1771
err = ubi_sync_erase(ubi, e, false);
drivers/mtd/ubi/wl.c
1775
wl_tree_add(e, &ubi->free);
drivers/mtd/ubi/wl.c
1778
err = schedule_erase(ubi, e, aeb->vol_id, aeb->lnum, 0, false);
drivers/mtd/ubi/wl.c
178
kmem_cache_free(ubi_wl_entry_slab, e);
drivers/mtd/ubi/wl.c
1786
wl_entry_destroy(ubi, e);
drivers/mtd/ubi/wl.c
1805
struct ubi_wl_entry *e;
drivers/mtd/ubi/wl.c
1839
e = kmem_cache_alloc(ubi_wl_entry_slab, GFP_KERNEL);
drivers/mtd/ubi/wl.c
1840
if (!e) {
drivers/mtd/ubi/wl.c
1845
e->pnum = aeb->pnum;
drivers/mtd/ubi/wl.c
1846
e->ec = aeb->ec;
drivers/mtd/ubi/wl.c
1847
ubi_assert(e->ec >= 0);
drivers/mtd/ubi/wl.c
1849
wl_tree_add(e, &ubi->free);
drivers/mtd/ubi/wl.c
1852
ubi->lookuptbl[e->pnum] = e;
drivers/mtd/ubi/wl.c
1861
e = kmem_cache_alloc(ubi_wl_entry_slab, GFP_KERNEL);
drivers/mtd/ubi/wl.c
1862
if (!e) {
drivers/mtd/ubi/wl.c
1867
e->pnum = aeb->pnum;
drivers/mtd/ubi/wl.c
1868
e->ec = aeb->ec;
drivers/mtd/ubi/wl.c
1869
ubi->lookuptbl[e->pnum] = e;
drivers/mtd/ubi/wl.c
1873
e->pnum, e->ec);
drivers/mtd/ubi/wl.c
1874
wl_tree_add(e, &ubi->used);
drivers/mtd/ubi/wl.c
1877
e->pnum, e->ec);
drivers/mtd/ubi/wl.c
1878
wl_tree_add(e, &ubi->scrub);
drivers/mtd/ubi/wl.c
1888
e = ubi_find_fm_block(ubi, aeb->pnum);
drivers/mtd/ubi/wl.c
1890
if (e) {
drivers/mtd/ubi/wl.c
1891
ubi_assert(!ubi->lookuptbl[e->pnum]);
drivers/mtd/ubi/wl.c
1892
ubi->lookuptbl[e->pnum] = e;
drivers/mtd/ubi/wl.c
1971
struct ubi_wl_entry *e, *tmp;
drivers/mtd/ubi/wl.c
1974
list_for_each_entry_safe(e, tmp, &ubi->pq[i], u.list) {
drivers/mtd/ubi/wl.c
1975
list_del(&e->u.list);
drivers/mtd/ubi/wl.c
1976
wl_entry_destroy(ubi, e);
drivers/mtd/ubi/wl.c
2052
struct ubi_wl_entry *e, struct rb_root *root)
drivers/mtd/ubi/wl.c
2057
if (in_wl_tree(e, root))
drivers/mtd/ubi/wl.c
2061
e->pnum, e->ec, root);
drivers/mtd/ubi/wl.c
2075
struct ubi_wl_entry *e)
drivers/mtd/ubi/wl.c
2080
if (in_pq(ubi, e))
drivers/mtd/ubi/wl.c
2084
e->pnum, e->ec);
drivers/mtd/ubi/wl.c
2091
struct ubi_wl_entry *e;
drivers/mtd/ubi/wl.c
2093
e = find_wl_entry(ubi, &ubi->free, WL_FREE_MAX_DIFF, 0);
drivers/mtd/ubi/wl.c
2094
self_check_in_wl_tree(ubi, e, &ubi->free);
drivers/mtd/ubi/wl.c
2097
rb_erase(&e->u.rb, &ubi->free);
drivers/mtd/ubi/wl.c
2099
return e;
drivers/mtd/ubi/wl.c
2140
struct ubi_wl_entry *e;
drivers/mtd/ubi/wl.c
2163
e = wl_get_wle(ubi);
drivers/mtd/ubi/wl.c
2164
prot_queue_add(ubi, e);
drivers/mtd/ubi/wl.c
2167
err = ubi_self_check_all_ff(ubi, e->pnum, ubi->vid_hdr_aloffset,
drivers/mtd/ubi/wl.c
2170
ubi_err(ubi, "new PEB %d does not contain all 0xFF bytes", e->pnum);
drivers/mtd/ubi/wl.c
2174
return e->pnum;
drivers/mtd/ubi/wl.c
242
static int in_wl_tree(struct ubi_wl_entry *e, struct rb_root *root)
drivers/mtd/ubi/wl.c
252
if (e->pnum == e1->pnum) {
drivers/mtd/ubi/wl.c
253
ubi_assert(e == e1);
drivers/mtd/ubi/wl.c
257
if (e->ec < e1->ec)
drivers/mtd/ubi/wl.c
259
else if (e->ec > e1->ec)
drivers/mtd/ubi/wl.c
262
ubi_assert(e->pnum != e1->pnum);
drivers/mtd/ubi/wl.c
263
if (e->pnum < e1->pnum)
drivers/mtd/ubi/wl.c
281
static inline int in_pq(const struct ubi_device *ubi, struct ubi_wl_entry *e)
drivers/mtd/ubi/wl.c
288
if (p == e)
drivers/mtd/ubi/wl.c
304
static void prot_queue_add(struct ubi_device *ubi, struct ubi_wl_entry *e)
drivers/mtd/ubi/wl.c
311
list_add_tail(&e->u.list, &ubi->pq[pq_tail]);
drivers/mtd/ubi/wl.c
312
dbg_wl("added PEB %d EC %d to the protection queue", e->pnum, e->ec);
drivers/mtd/ubi/wl.c
330
struct ubi_wl_entry *e;
drivers/mtd/ubi/wl.c
333
e = rb_entry(rb_first(root), struct ubi_wl_entry, u.rb);
drivers/mtd/ubi/wl.c
334
max = e->ec + diff;
drivers/mtd/ubi/wl.c
343
e = e1;
drivers/mtd/ubi/wl.c
347
e = e1;
drivers/mtd/ubi/wl.c
351
return e;
drivers/mtd/ubi/wl.c
366
struct ubi_wl_entry *e, *first, *last;
drivers/mtd/ubi/wl.c
372
e = rb_entry(root->rb_node, struct ubi_wl_entry, u.rb);
drivers/mtd/ubi/wl.c
380
e = may_reserve_for_fm(ubi, e, root);
drivers/mtd/ubi/wl.c
382
e = find_wl_entry(ubi, root, WL_FREE_MAX_DIFF/2, 0);
drivers/mtd/ubi/wl.c
384
return e;
drivers/mtd/ubi/wl.c
397
struct ubi_wl_entry *e;
drivers/mtd/ubi/wl.c
399
e = find_mean_wl_entry(ubi, &ubi->free);
drivers/mtd/ubi/wl.c
400
if (!e) {
drivers/mtd/ubi/wl.c
405
self_check_in_wl_tree(ubi, e, &ubi->free);
drivers/mtd/ubi/wl.c
411
rb_erase(&e->u.rb, &ubi->free);
drivers/mtd/ubi/wl.c
413
dbg_wl("PEB %d EC %d", e->pnum, e->ec);
drivers/mtd/ubi/wl.c
415
return e;
drivers/mtd/ubi/wl.c
428
struct ubi_wl_entry *e;
drivers/mtd/ubi/wl.c
430
e = ubi->lookuptbl[pnum];
drivers/mtd/ubi/wl.c
431
if (!e)
drivers/mtd/ubi/wl.c
434
if (self_check_in_pq(ubi, e))
drivers/mtd/ubi/wl.c
437
list_del(&e->u.list);
drivers/mtd/ubi/wl.c
438
dbg_wl("deleted PEB %d from the protection queue", e->pnum);
drivers/mtd/ubi/wl.c
451
int ubi_sync_erase(struct ubi_device *ubi, struct ubi_wl_entry *e, int torture)
drivers/mtd/ubi/wl.c
455
unsigned long long ec = e->ec;
drivers/mtd/ubi/wl.c
457
dbg_wl("erase PEB %d, old EC %llu", e->pnum, ec);
drivers/mtd/ubi/wl.c
459
err = self_check_ec(ubi, e->pnum, e->ec);
drivers/mtd/ubi/wl.c
467
err = ubi_io_sync_erase(ubi, e->pnum, torture);
drivers/mtd/ubi/wl.c
478
e->pnum, ec);
drivers/mtd/ubi/wl.c
483
dbg_wl("erased PEB %d, new EC %llu", e->pnum, ec);
drivers/mtd/ubi/wl.c
487
err = ubi_io_write_ec_hdr(ubi, e->pnum, ec_hdr);
drivers/mtd/ubi/wl.c
491
e->ec = ec;
drivers/mtd/ubi/wl.c
493
if (e->ec > ubi->max_ec)
drivers/mtd/ubi/wl.c
494
ubi->max_ec = e->ec;
drivers/mtd/ubi/wl.c
512
struct ubi_wl_entry *e, *tmp;
drivers/mtd/ubi/wl.c
522
list_for_each_entry_safe(e, tmp, &ubi->pq[ubi->pq_head], u.list) {
drivers/mtd/ubi/wl.c
524
e->pnum, e->ec);
drivers/mtd/ubi/wl.c
526
list_del(&e->u.list);
drivers/mtd/ubi/wl.c
527
wl_tree_add(e, &ubi->used);
drivers/mtd/ubi/wl.c
595
static int schedule_erase(struct ubi_device *ubi, struct ubi_wl_entry *e,
drivers/mtd/ubi/wl.c
600
ubi_assert(e);
drivers/mtd/ubi/wl.c
603
e->pnum, e->ec, torture);
drivers/mtd/ubi/wl.c
610
wl_wrk->e = e;
drivers/mtd/ubi/wl.c
632
static int do_sync_erase(struct ubi_device *ubi, struct ubi_wl_entry *e,
drivers/mtd/ubi/wl.c
637
dbg_wl("sync erase of PEB %i", e->pnum);
drivers/mtd/ubi/wl.c
639
wl_wrk.e = e;
drivers/mtd/ubi/wl.h
21
struct ubi_wl_entry *e,
drivers/mtd/ubi/wl.h
28
struct ubi_wl_entry *e,
drivers/mtd/ubi/wl.h
30
return e;
drivers/net/dsa/b53/b53_common.c
2628
int b53_set_mac_eee(struct dsa_switch *ds, int port, struct ethtool_keee *e)
drivers/net/dsa/b53/b53_common.c
2633
p->eee_enabled = e->eee_enabled;
drivers/net/dsa/b53/b53_common.c
2634
b53_eee_enable_set(ds, port, e->eee_enabled);
drivers/net/dsa/b53/b53_priv.h
546
int b53_set_mac_eee(struct dsa_switch *ds, int port, struct ethtool_keee *e);
drivers/net/dsa/lantiq/lantiq_gswip_common.c
1560
struct ethtool_keee *e)
drivers/net/dsa/lantiq/lantiq_gswip_common.c
1562
if (e->tx_lpi_timer > 0x7f)
drivers/net/dsa/microchip/ksz9477_acl.c
474
struct ksz9477_acl_entry *e;
drivers/net/dsa/microchip/ksz9477_acl.c
478
e = &acles->entries[i];
drivers/net/dsa/microchip/ksz9477_acl.c
479
*e = acles->entries[i + num_entries_to_move];
drivers/net/dsa/microchip/ksz9477_acl.c
481
ret = ksz9477_acl_update_linkage(dev, &e->entry[0],
drivers/net/dsa/microchip/ksz9477_acl.c
512
struct ksz9477_acl_entry *e;
drivers/net/dsa/microchip/ksz9477_acl.c
518
e = &acles->entries[b];
drivers/net/dsa/microchip/ksz9477_acl.c
519
*e = acles->entries[i - 1];
drivers/net/dsa/microchip/ksz9477_acl.c
521
ret = ksz9477_acl_update_linkage(dev, &e->entry[0], i - 1, b);
drivers/net/dsa/microchip/ksz_common.c
3737
struct ethtool_keee *e)
drivers/net/dsa/microchip/ksz_common.c
3741
if (!e->tx_lpi_enabled) {
drivers/net/dsa/microchip/ksz_common.c
3746
if (e->tx_lpi_timer) {
drivers/net/dsa/mt7530.c
3144
struct ethtool_keee *e)
drivers/net/dsa/mt7530.c
3146
if (e->tx_lpi_timer > 0xFFF)
drivers/net/dsa/mt7530.h
911
static inline void mt7530_hw_vlan_entry_init(struct mt7530_hw_vlan_entry *e,
drivers/net/dsa/mt7530.h
914
e->port = port;
drivers/net/dsa/mt7530.h
915
e->untagged = untagged;
drivers/net/dsa/mv88e6xxx/chip.c
1517
struct ethtool_keee *e)
drivers/net/dsa/ocelot/felix_vsc9959.c
2122
struct action_gate_entry *e;
drivers/net/dsa/ocelot/felix_vsc9959.c
2167
e = sgi->entries;
drivers/net/dsa/ocelot/felix_vsc9959.c
2169
u32 ips = (e[i].ipv < 0) ? 0 : (e[i].ipv + 8);
drivers/net/dsa/ocelot/felix_vsc9959.c
2172
(e[i].gate_state ?
drivers/net/dsa/ocelot/felix_vsc9959.c
2176
interval_sum += e[i].interval;
drivers/net/dsa/sja1105/sja1105_tas.c
323
struct sja1105_gate_entry *e;
drivers/net/dsa/sja1105/sja1105_tas.c
343
list_for_each_entry(e, &gating_cfg->entries, list) {
drivers/net/dsa/sja1105/sja1105_tas.c
344
schedule[k].delta = ns_to_sja1105_delta(e->interval);
drivers/net/dsa/sja1105/sja1105_tas.c
345
schedule[k].destports = e->rule->vl.destports;
drivers/net/dsa/sja1105/sja1105_tas.c
348
schedule[k].vlindex = e->rule->vl.sharindx;
drivers/net/dsa/sja1105/sja1105_tas.c
349
schedule[k].winstindex = e->rule->vl.sharindx;
drivers/net/dsa/sja1105/sja1105_tas.c
350
if (e->gate_state) /* Gate open */
drivers/net/dsa/sja1105/sja1105_tas.c
473
struct sja1105_gate_entry *e;
drivers/net/dsa/sja1105/sja1105_tas.c
490
list_for_each_entry(e, &gating_cfg->entries, list)
drivers/net/dsa/sja1105/sja1105_tas.c
491
dummy->entries[i++].interval = e->interval;
drivers/net/dsa/sja1105/sja1105_vl.c
16
struct sja1105_gate_entry *e;
drivers/net/dsa/sja1105/sja1105_vl.c
19
e = kzalloc_obj(*e);
drivers/net/dsa/sja1105/sja1105_vl.c
20
if (!e)
drivers/net/dsa/sja1105/sja1105_vl.c
23
e->rule = rule;
drivers/net/dsa/sja1105/sja1105_vl.c
24
e->gate_state = gate_state;
drivers/net/dsa/sja1105/sja1105_vl.c
25
e->interval = entry_time;
drivers/net/dsa/sja1105/sja1105_vl.c
28
list_add(&e->list, &gating_cfg->entries);
drivers/net/dsa/sja1105/sja1105_vl.c
33
if (p->interval == e->interval) {
drivers/net/dsa/sja1105/sja1105_vl.c
40
if (e->interval < p->interval)
drivers/net/dsa/sja1105/sja1105_vl.c
43
list_add(&e->list, p->list.prev);
drivers/net/dsa/sja1105/sja1105_vl.c
50
kfree(e);
drivers/net/dsa/sja1105/sja1105_vl.c
62
struct sja1105_gate_entry *e;
drivers/net/dsa/sja1105/sja1105_vl.c
65
list_for_each_entry(e, &gating_cfg->entries, list) {
drivers/net/dsa/sja1105/sja1105_vl.c
68
prev = e->list.prev;
drivers/net/dsa/sja1105/sja1105_vl.c
74
p->interval = e->interval - p->interval;
drivers/net/dsa/sja1105/sja1105_vl.c
83
struct sja1105_gate_entry *e, *n;
drivers/net/dsa/sja1105/sja1105_vl.c
85
list_for_each_entry_safe(e, n, &gating_cfg->entries, list) {
drivers/net/dsa/sja1105/sja1105_vl.c
86
list_del(&e->list);
drivers/net/dsa/sja1105/sja1105_vl.c
87
kfree(e);
drivers/net/dsa/yt921x.c
1004
res = yt921x_set_eee(priv, port, e);
drivers/net/dsa/yt921x.c
958
yt921x_set_eee(struct yt921x_priv *priv, int port, struct ethtool_keee *e)
drivers/net/dsa/yt921x.c
962
bool enable = e->eee_enabled;
drivers/net/dsa/yt921x.c
998
yt921x_dsa_set_mac_eee(struct dsa_switch *ds, int port, struct ethtool_keee *e)
drivers/net/ethernet/airoha/airoha_eth.c
1048
struct airoha_queue_entry *e = &q->entry[i];
drivers/net/ethernet/airoha/airoha_eth.c
1050
if (!e->dma_addr)
drivers/net/ethernet/airoha/airoha_eth.c
1053
dma_unmap_single(eth->dev, e->dma_addr, e->dma_len,
drivers/net/ethernet/airoha/airoha_eth.c
1055
dev_kfree_skb_any(e->skb);
drivers/net/ethernet/airoha/airoha_eth.c
1056
e->dma_addr = 0;
drivers/net/ethernet/airoha/airoha_eth.c
1057
e->skb = NULL;
drivers/net/ethernet/airoha/airoha_eth.c
1058
list_add_tail(&e->list, &q->tx_list);
drivers/net/ethernet/airoha/airoha_eth.c
1924
struct airoha_queue_entry *e;
drivers/net/ethernet/airoha/airoha_eth.c
1983
e = list_first_entry(&q->tx_list, struct airoha_queue_entry,
drivers/net/ethernet/airoha/airoha_eth.c
1985
index = e - q->entry;
drivers/net/ethernet/airoha/airoha_eth.c
1998
list_move_tail(&e->list, &tx_list);
drivers/net/ethernet/airoha/airoha_eth.c
1999
e->skb = i ? NULL : skb;
drivers/net/ethernet/airoha/airoha_eth.c
2000
e->dma_addr = addr;
drivers/net/ethernet/airoha/airoha_eth.c
2001
e->dma_len = len;
drivers/net/ethernet/airoha/airoha_eth.c
2003
e = list_first_entry(&q->tx_list, struct airoha_queue_entry,
drivers/net/ethernet/airoha/airoha_eth.c
2005
index = e - q->entry;
drivers/net/ethernet/airoha/airoha_eth.c
2040
e = list_first_entry(&tx_list, struct airoha_queue_entry,
drivers/net/ethernet/airoha/airoha_eth.c
2042
dma_unmap_single(dev->dev.parent, e->dma_addr, e->dma_len,
drivers/net/ethernet/airoha/airoha_eth.c
2044
e->dma_addr = 0;
drivers/net/ethernet/airoha/airoha_eth.c
2045
list_move_tail(&e->list, &q->tx_list);
drivers/net/ethernet/airoha/airoha_eth.c
547
struct airoha_queue_entry *e = &q->entry[q->head];
drivers/net/ethernet/airoha/airoha_eth.c
562
e->buf = page_address(page) + offset;
drivers/net/ethernet/airoha/airoha_eth.c
563
e->dma_addr = page_pool_get_dma_addr(page) + offset;
drivers/net/ethernet/airoha/airoha_eth.c
564
e->dma_len = SKB_WITH_OVERHEAD(q->buf_size);
drivers/net/ethernet/airoha/airoha_eth.c
566
val = FIELD_PREP(QDMA_DESC_LEN_MASK, e->dma_len);
drivers/net/ethernet/airoha/airoha_eth.c
568
WRITE_ONCE(desc->addr, cpu_to_le32(e->dma_addr));
drivers/net/ethernet/airoha/airoha_eth.c
613
struct airoha_queue_entry *e = &q->entry[q->tail];
drivers/net/ethernet/airoha/airoha_eth.c
616
struct page *page = virt_to_head_page(e->buf);
drivers/net/ethernet/airoha/airoha_eth.c
627
dma_sync_single_for_cpu(eth->dev, e->dma_addr,
drivers/net/ethernet/airoha/airoha_eth.c
642
q->skb = napi_build_skb(e->buf, q->buf_size);
drivers/net/ethernet/airoha/airoha_eth.c
660
e->buf - page_address(page), len,
drivers/net/ethernet/airoha/airoha_eth.c
801
struct airoha_queue_entry *e = &q->entry[q->tail];
drivers/net/ethernet/airoha/airoha_eth.c
803
struct page *page = virt_to_head_page(e->buf);
drivers/net/ethernet/airoha/airoha_eth.c
805
dma_sync_single_for_cpu(eth->dev, e->dma_addr, e->dma_len,
drivers/net/ethernet/airoha/airoha_eth.c
868
struct airoha_queue_entry *e;
drivers/net/ethernet/airoha/airoha_eth.c
905
e = &q->entry[index];
drivers/net/ethernet/airoha/airoha_eth.c
906
skb = e->skb;
drivers/net/ethernet/airoha/airoha_eth.c
908
dma_unmap_single(eth->dev, e->dma_addr, e->dma_len,
drivers/net/ethernet/airoha/airoha_eth.c
910
e->dma_addr = 0;
drivers/net/ethernet/airoha/airoha_eth.c
911
list_add_tail(&e->list, &q->tx_list);
drivers/net/ethernet/airoha/airoha_ppe.c
1001
return airoha_ppe_get_entry_idle_time(ppe, e->data.ib1);
drivers/net/ethernet/airoha/airoha_ppe.c
1008
struct airoha_flow_table_entry *e;
drivers/net/ethernet/airoha/airoha_ppe.c
1178
e = kzalloc_obj(*e);
drivers/net/ethernet/airoha/airoha_ppe.c
1179
if (!e)
drivers/net/ethernet/airoha/airoha_ppe.c
1182
e->cookie = f->cookie;
drivers/net/ethernet/airoha/airoha_ppe.c
1183
memcpy(&e->data, &hwe, sizeof(e->data));
drivers/net/ethernet/airoha/airoha_ppe.c
1185
err = airoha_ppe_foe_flow_commit_entry(eth->ppe, e);
drivers/net/ethernet/airoha/airoha_ppe.c
1189
err = rhashtable_insert_fast(&eth->flow_table, &e->node,
drivers/net/ethernet/airoha/airoha_ppe.c
1197
airoha_ppe_foe_flow_remove_entry(eth->ppe, e);
drivers/net/ethernet/airoha/airoha_ppe.c
1199
kfree(e);
drivers/net/ethernet/airoha/airoha_ppe.c
1207
struct airoha_flow_table_entry *e;
drivers/net/ethernet/airoha/airoha_ppe.c
1209
e = rhashtable_lookup(&eth->flow_table, &f->cookie,
drivers/net/ethernet/airoha/airoha_ppe.c
1211
if (!e)
drivers/net/ethernet/airoha/airoha_ppe.c
1214
airoha_ppe_foe_flow_remove_entry(eth->ppe, e);
drivers/net/ethernet/airoha/airoha_ppe.c
1215
rhashtable_remove_fast(&eth->flow_table, &e->node,
drivers/net/ethernet/airoha/airoha_ppe.c
1217
kfree(e);
drivers/net/ethernet/airoha/airoha_ppe.c
1260
struct airoha_flow_table_entry *e;
drivers/net/ethernet/airoha/airoha_ppe.c
1263
e = rhashtable_lookup(&eth->flow_table, &f->cookie,
drivers/net/ethernet/airoha/airoha_ppe.c
1265
if (!e)
drivers/net/ethernet/airoha/airoha_ppe.c
1268
idle = airoha_ppe_entry_idle_time(eth->ppe, e);
drivers/net/ethernet/airoha/airoha_ppe.c
1271
if (e->hash != 0xffff) {
drivers/net/ethernet/airoha/airoha_ppe.c
1274
airoha_ppe_foe_entry_get_stats(eth->ppe, e->hash, &stats);
drivers/net/ethernet/airoha/airoha_ppe.c
1275
f->stats.pkts += (stats.packets - e->stats.packets);
drivers/net/ethernet/airoha/airoha_ppe.c
1276
f->stats.bytes += (stats.bytes - e->stats.bytes);
drivers/net/ethernet/airoha/airoha_ppe.c
1277
e->stats = stats;
drivers/net/ethernet/airoha/airoha_ppe.c
648
static bool airoha_ppe_foe_compare_entry(struct airoha_flow_table_entry *e,
drivers/net/ethernet/airoha/airoha_ppe.c
651
int type = FIELD_GET(AIROHA_FOE_IB1_BIND_PACKET_TYPE, e->data.ib1);
drivers/net/ethernet/airoha/airoha_ppe.c
654
if ((hwe->ib1 ^ e->data.ib1) & AIROHA_FOE_IB1_BIND_UDP)
drivers/net/ethernet/airoha/airoha_ppe.c
662
return !memcmp(&e->data.d, &hwe->d, len - sizeof(hwe->ib1));
drivers/net/ethernet/airoha/airoha_ppe.c
687
struct airoha_foe_entry *e,
drivers/net/ethernet/airoha/airoha_ppe.c
697
memcpy(&hwe->d, &e->d, sizeof(*hwe) - sizeof(hwe->ib1));
drivers/net/ethernet/airoha/airoha_ppe.c
700
e->ib1 &= ~AIROHA_FOE_IB1_BIND_TIMESTAMP;
drivers/net/ethernet/airoha/airoha_ppe.c
701
e->ib1 |= FIELD_PREP(AIROHA_FOE_IB1_BIND_TIMESTAMP, ts);
drivers/net/ethernet/airoha/airoha_ppe.c
702
hwe->ib1 = e->ib1;
drivers/net/ethernet/airoha/airoha_ppe.c
724
struct airoha_flow_table_entry *e)
drivers/net/ethernet/airoha/airoha_ppe.c
728
hlist_del_init(&e->list);
drivers/net/ethernet/airoha/airoha_ppe.c
729
if (e->hash != 0xffff) {
drivers/net/ethernet/airoha/airoha_ppe.c
730
e->data.ib1 &= ~AIROHA_FOE_IB1_BIND_STATE;
drivers/net/ethernet/airoha/airoha_ppe.c
731
e->data.ib1 |= FIELD_PREP(AIROHA_FOE_IB1_BIND_STATE,
drivers/net/ethernet/airoha/airoha_ppe.c
733
airoha_ppe_foe_commit_entry(ppe, &e->data, e->hash, false);
drivers/net/ethernet/airoha/airoha_ppe.c
734
e->hash = 0xffff;
drivers/net/ethernet/airoha/airoha_ppe.c
736
if (e->type == FLOW_TYPE_L2_SUBFLOW) {
drivers/net/ethernet/airoha/airoha_ppe.c
737
hlist_del_init(&e->l2_subflow_node);
drivers/net/ethernet/airoha/airoha_ppe.c
738
kfree(e);
drivers/net/ethernet/airoha/airoha_ppe.c
743
struct airoha_flow_table_entry *e)
drivers/net/ethernet/airoha/airoha_ppe.c
745
struct hlist_head *head = &e->l2_flows;
drivers/net/ethernet/airoha/airoha_ppe.c
750
rhashtable_remove_fast(&ppe->l2_flows, &e->l2_node,
drivers/net/ethernet/airoha/airoha_ppe.c
752
hlist_for_each_entry_safe(e, n, head, l2_subflow_node)
drivers/net/ethernet/airoha/airoha_ppe.c
753
airoha_ppe_foe_remove_flow(ppe, e);
drivers/net/ethernet/airoha/airoha_ppe.c
757
struct airoha_flow_table_entry *e)
drivers/net/ethernet/airoha/airoha_ppe.c
761
if (e->type == FLOW_TYPE_L2)
drivers/net/ethernet/airoha/airoha_ppe.c
762
airoha_ppe_foe_remove_l2_flow(ppe, e);
drivers/net/ethernet/airoha/airoha_ppe.c
764
airoha_ppe_foe_remove_flow(ppe, e);
drivers/net/ethernet/airoha/airoha_ppe.c
771
struct airoha_flow_table_entry *e,
drivers/net/ethernet/airoha/airoha_ppe.c
787
hlist_add_head(&f->l2_subflow_node, &e->l2_flows);
drivers/net/ethernet/airoha/airoha_ppe.c
792
hwe.ib1 = (hwe.ib1 & mask) | (e->data.ib1 & ~mask);
drivers/net/ethernet/airoha/airoha_ppe.c
796
memcpy(&hwe.ipv6.l2, &e->data.bridge.l2, sizeof(hwe.ipv6.l2));
drivers/net/ethernet/airoha/airoha_ppe.c
797
hwe.ipv6.ib2 = e->data.bridge.ib2;
drivers/net/ethernet/airoha/airoha_ppe.c
804
memcpy(&hwe.bridge.l2, &e->data.bridge.l2,
drivers/net/ethernet/airoha/airoha_ppe.c
806
hwe.bridge.ib2 = e->data.bridge.ib2;
drivers/net/ethernet/airoha/airoha_ppe.c
812
hwe.bridge.data = e->data.bridge.data;
drivers/net/ethernet/airoha/airoha_ppe.c
822
struct airoha_flow_table_entry *e;
drivers/net/ethernet/airoha/airoha_ppe.c
840
hlist_for_each_entry_safe(e, n, &ppe->foe_flow[index], list) {
drivers/net/ethernet/airoha/airoha_ppe.c
841
if (e->type == FLOW_TYPE_L2_SUBFLOW) {
drivers/net/ethernet/airoha/airoha_ppe.c
844
e->hash = 0xffff;
drivers/net/ethernet/airoha/airoha_ppe.c
845
airoha_ppe_foe_remove_flow(ppe, e);
drivers/net/ethernet/airoha/airoha_ppe.c
850
if (!airoha_ppe_foe_compare_entry(e, hwe))
drivers/net/ethernet/airoha/airoha_ppe.c
853
airoha_ppe_foe_commit_entry(ppe, &e->data, hash, rx_wlan);
drivers/net/ethernet/airoha/airoha_ppe.c
855
e->hash = hash;
drivers/net/ethernet/airoha/airoha_ppe.c
862
e = rhashtable_lookup_fast(&ppe->l2_flows, &br,
drivers/net/ethernet/airoha/airoha_ppe.c
864
if (e)
drivers/net/ethernet/airoha/airoha_ppe.c
865
airoha_ppe_foe_commit_subflow_entry(ppe, e, hash, rx_wlan);
drivers/net/ethernet/airoha/airoha_ppe.c
872
struct airoha_flow_table_entry *e)
drivers/net/ethernet/airoha/airoha_ppe.c
876
e->type = FLOW_TYPE_L2;
drivers/net/ethernet/airoha/airoha_ppe.c
877
prev = rhashtable_lookup_get_insert_fast(&ppe->l2_flows, &e->l2_node,
drivers/net/ethernet/airoha/airoha_ppe.c
886
&e->l2_node,
drivers/net/ethernet/airoha/airoha_ppe.c
891
struct airoha_flow_table_entry *e)
drivers/net/ethernet/airoha/airoha_ppe.c
893
int type = FIELD_GET(AIROHA_FOE_IB1_BIND_PACKET_TYPE, e->data.ib1);
drivers/net/ethernet/airoha/airoha_ppe.c
897
return airoha_ppe_foe_l2_flow_commit_entry(ppe, e);
drivers/net/ethernet/airoha/airoha_ppe.c
899
hash = airoha_ppe_foe_get_entry_hash(ppe, &e->data);
drivers/net/ethernet/airoha/airoha_ppe.c
900
e->type = FLOW_TYPE_L4;
drivers/net/ethernet/airoha/airoha_ppe.c
901
e->hash = 0xffff;
drivers/net/ethernet/airoha/airoha_ppe.c
904
hlist_add_head(&e->list, &ppe->foe_flow[hash]);
drivers/net/ethernet/airoha/airoha_ppe.c
931
struct airoha_flow_table_entry *e)
drivers/net/ethernet/airoha/airoha_ppe.c
933
int min_idle = airoha_ppe_get_entry_idle_time(ppe, e->data.ib1);
drivers/net/ethernet/airoha/airoha_ppe.c
939
hlist_for_each_entry_safe(iter, n, &e->l2_flows, l2_subflow_node) {
drivers/net/ethernet/airoha/airoha_ppe.c
961
e->data.ib1 &= ~AIROHA_FOE_IB1_BIND_TIMESTAMP;
drivers/net/ethernet/airoha/airoha_ppe.c
962
e->data.ib1 |= ib1 & AIROHA_FOE_IB1_BIND_TIMESTAMP;
drivers/net/ethernet/airoha/airoha_ppe.c
967
struct airoha_flow_table_entry *e)
drivers/net/ethernet/airoha/airoha_ppe.c
973
if (e->type == FLOW_TYPE_L2) {
drivers/net/ethernet/airoha/airoha_ppe.c
974
airoha_ppe_foe_flow_l2_entry_update(ppe, e);
drivers/net/ethernet/airoha/airoha_ppe.c
978
if (e->hash == 0xffff)
drivers/net/ethernet/airoha/airoha_ppe.c
981
hwe_p = airoha_ppe_foe_get_entry_locked(ppe, e->hash);
drivers/net/ethernet/airoha/airoha_ppe.c
986
if (!airoha_ppe_foe_compare_entry(e, &hwe)) {
drivers/net/ethernet/airoha/airoha_ppe.c
987
e->hash = 0xffff;
drivers/net/ethernet/airoha/airoha_ppe.c
991
e->data.ib1 = hwe.ib1;
drivers/net/ethernet/airoha/airoha_ppe.c
997
struct airoha_flow_table_entry *e)
drivers/net/ethernet/airoha/airoha_ppe.c
999
airoha_ppe_foe_flow_entry_update(ppe, e);
drivers/net/ethernet/broadcom/asp2/bcmasp_ethtool.c
353
static int bcmasp_get_eee(struct net_device *dev, struct ethtool_keee *e)
drivers/net/ethernet/broadcom/asp2/bcmasp_ethtool.c
358
return phy_ethtool_get_eee(dev->phydev, e);
drivers/net/ethernet/broadcom/asp2/bcmasp_ethtool.c
361
static int bcmasp_set_eee(struct net_device *dev, struct ethtool_keee *e)
drivers/net/ethernet/broadcom/asp2/bcmasp_ethtool.c
366
return phy_ethtool_set_eee(dev->phydev, e);
drivers/net/ethernet/broadcom/genet/bcmgenet.c
1386
static int bcmgenet_get_eee(struct net_device *dev, struct ethtool_keee *e)
drivers/net/ethernet/broadcom/genet/bcmgenet.c
1397
ret = phy_ethtool_get_eee(dev->phydev, e);
drivers/net/ethernet/broadcom/genet/bcmgenet.c
1404
e->tx_lpi_timer = bcmgenet_umac_readl(priv, UMAC_EEE_LPI_TIMER);
drivers/net/ethernet/broadcom/genet/bcmgenet.c
1409
static int bcmgenet_set_eee(struct net_device *dev, struct ethtool_keee *e)
drivers/net/ethernet/broadcom/genet/bcmgenet.c
1419
bcmgenet_umac_writel(priv, e->tx_lpi_timer, UMAC_EEE_LPI_TIMER);
drivers/net/ethernet/broadcom/genet/bcmgenet.c
1421
return phy_ethtool_set_eee(dev->phydev, e);
drivers/net/ethernet/brocade/bna/bfa_cs.h
22
#define BFA_SM_TABLE(n, s, e, t) \
drivers/net/ethernet/brocade/bna/bfa_cs.h
24
enum e; \
drivers/net/ethernet/brocade/bna/bfa_cs.h
25
typedef void (*t)(struct s *, enum e); \
drivers/net/ethernet/brocade/bna/bfa_ioc.h
153
void (*fsm)(struct bfa_iocpf *s, enum iocpf_event e);
drivers/net/ethernet/brocade/bna/bfa_ioc.h
163
void (*fsm)(struct bfa_ioc *s, enum ioc_event e);
drivers/net/ethernet/brocade/bna/bfa_msgq.h
61
void (*fsm)(struct bfa_msgq_cmdq *s, enum cmdq_event e);
drivers/net/ethernet/brocade/bna/bfa_msgq.h
89
void (*fsm)(struct bfa_msgq_rspq *s, enum rspq_event e);
drivers/net/ethernet/brocade/bna/bna_types.h
318
void (*fsm)(struct bna_ioceth *s, enum bna_ioceth_event e);
drivers/net/ethernet/brocade/bna/bna_types.h
342
void (*fsm)(struct bna_enet *s, enum bna_enet_event e);
drivers/net/ethernet/brocade/bna/bna_types.h
370
void (*fsm)(struct bna_ethport *s, enum bna_ethport_event e);
drivers/net/ethernet/brocade/bna/bna_types.h
472
void (*fsm)(struct bna_tx *s, enum bna_tx_event e);
drivers/net/ethernet/brocade/bna/bna_types.h
714
void (*fsm)(struct bna_rxf *s, enum bna_rxf_event e);
drivers/net/ethernet/brocade/bna/bna_types.h
793
void (*fsm)(struct bna_rx *s, enum bna_rx_event e);
drivers/net/ethernet/chelsio/cxgb/cxgb2.c
523
const struct espi_intr_counts *e;
drivers/net/ethernet/chelsio/cxgb/cxgb2.c
525
e = t1_espi_get_intr_counts(adapter->espi);
drivers/net/ethernet/chelsio/cxgb/cxgb2.c
526
*data++ = e->DIP2_parity_err;
drivers/net/ethernet/chelsio/cxgb/cxgb2.c
527
*data++ = e->DIP4_err;
drivers/net/ethernet/chelsio/cxgb/cxgb2.c
528
*data++ = e->rx_drops;
drivers/net/ethernet/chelsio/cxgb/cxgb2.c
529
*data++ = e->tx_drops;
drivers/net/ethernet/chelsio/cxgb/cxgb2.c
530
*data++ = e->rx_ovflw;
drivers/net/ethernet/chelsio/cxgb/cxgb2.c
531
*data++ = e->parity_err;
drivers/net/ethernet/chelsio/cxgb/cxgb2.c
713
static void get_sge_param(struct net_device *dev, struct ethtool_ringparam *e,
drivers/net/ethernet/chelsio/cxgb/cxgb2.c
720
e->rx_max_pending = MAX_RX_BUFFERS;
drivers/net/ethernet/chelsio/cxgb/cxgb2.c
721
e->rx_jumbo_max_pending = MAX_RX_JUMBO_BUFFERS;
drivers/net/ethernet/chelsio/cxgb/cxgb2.c
722
e->tx_max_pending = MAX_CMDQ_ENTRIES;
drivers/net/ethernet/chelsio/cxgb/cxgb2.c
724
e->rx_pending = adapter->params.sge.freelQ_size[!jumbo_fl];
drivers/net/ethernet/chelsio/cxgb/cxgb2.c
725
e->rx_jumbo_pending = adapter->params.sge.freelQ_size[jumbo_fl];
drivers/net/ethernet/chelsio/cxgb/cxgb2.c
726
e->tx_pending = adapter->params.sge.cmdQ_size[0];
drivers/net/ethernet/chelsio/cxgb/cxgb2.c
729
static int set_sge_param(struct net_device *dev, struct ethtool_ringparam *e,
drivers/net/ethernet/chelsio/cxgb/cxgb2.c
736
if (e->rx_pending > MAX_RX_BUFFERS || e->rx_mini_pending ||
drivers/net/ethernet/chelsio/cxgb/cxgb2.c
737
e->rx_jumbo_pending > MAX_RX_JUMBO_BUFFERS ||
drivers/net/ethernet/chelsio/cxgb/cxgb2.c
738
e->tx_pending > MAX_CMDQ_ENTRIES ||
drivers/net/ethernet/chelsio/cxgb/cxgb2.c
739
e->rx_pending < MIN_FL_ENTRIES ||
drivers/net/ethernet/chelsio/cxgb/cxgb2.c
740
e->rx_jumbo_pending < MIN_FL_ENTRIES ||
drivers/net/ethernet/chelsio/cxgb/cxgb2.c
741
e->tx_pending < (adapter->params.nports + 1) * (MAX_SKB_FRAGS + 1))
drivers/net/ethernet/chelsio/cxgb/cxgb2.c
747
adapter->params.sge.freelQ_size[!jumbo_fl] = e->rx_pending;
drivers/net/ethernet/chelsio/cxgb/cxgb2.c
748
adapter->params.sge.freelQ_size[jumbo_fl] = e->rx_jumbo_pending;
drivers/net/ethernet/chelsio/cxgb/cxgb2.c
749
adapter->params.sge.cmdQ_size[0] = e->tx_pending;
drivers/net/ethernet/chelsio/cxgb/cxgb2.c
750
adapter->params.sge.cmdQ_size[1] = e->tx_pending > MAX_CMDQ1_ENTRIES ?
drivers/net/ethernet/chelsio/cxgb/cxgb2.c
751
MAX_CMDQ1_ENTRIES : e->tx_pending;
drivers/net/ethernet/chelsio/cxgb/cxgb2.c
790
static int get_eeprom(struct net_device *dev, struct ethtool_eeprom *e,
drivers/net/ethernet/chelsio/cxgb/cxgb2.c
797
e->magic = EEPROM_MAGIC(adapter);
drivers/net/ethernet/chelsio/cxgb/cxgb2.c
798
for (i = e->offset & ~3; i < e->offset + e->len; i += sizeof(u32))
drivers/net/ethernet/chelsio/cxgb/cxgb2.c
800
memcpy(data, buf + e->offset, e->len);
drivers/net/ethernet/chelsio/cxgb/sge.c
1142
static inline void write_tx_desc(struct cmdQ_e *e, dma_addr_t mapping,
drivers/net/ethernet/chelsio/cxgb/sge.c
1148
e->addr_lo = (u32)mapping;
drivers/net/ethernet/chelsio/cxgb/sge.c
1149
e->addr_hi = (u64)mapping >> 32;
drivers/net/ethernet/chelsio/cxgb/sge.c
1150
e->len_gen = V_CMD_LEN(len) | V_CMD_GEN1(gen);
drivers/net/ethernet/chelsio/cxgb/sge.c
1151
e->flags = F_CMD_DATAVALID | V_CMD_EOP(eop) | V_CMD_GEN2(gen);
drivers/net/ethernet/chelsio/cxgb/sge.c
1161
struct cmdQ_e **e,
drivers/net/ethernet/chelsio/cxgb/sge.c
1170
struct cmdQ_e *e1 = *e;
drivers/net/ethernet/chelsio/cxgb/sge.c
1191
*e = e1;
drivers/net/ethernet/chelsio/cxgb/sge.c
1206
struct cmdQ_e *e, *e1;
drivers/net/ethernet/chelsio/cxgb/sge.c
1211
e = e1 = &q->entries[pidx];
drivers/net/ethernet/chelsio/cxgb/sge.c
1225
e->addr_lo = (u32)desc_mapping;
drivers/net/ethernet/chelsio/cxgb/sge.c
1226
e->addr_hi = (u64)desc_mapping >> 32;
drivers/net/ethernet/chelsio/cxgb/sge.c
1227
e->len_gen = V_CMD_LEN(first_desc_len) | V_CMD_GEN1(gen);
drivers/net/ethernet/chelsio/cxgb/sge.c
1284
e->flags = flags;
drivers/net/ethernet/chelsio/cxgb/sge.c
1473
struct respQ_e *e = &q->entries[q->cidx];
drivers/net/ethernet/chelsio/cxgb/sge.c
1478
while (done < budget && e->GenerationBit == q->genbit) {
drivers/net/ethernet/chelsio/cxgb/sge.c
1479
flags |= e->Qsleeping;
drivers/net/ethernet/chelsio/cxgb/sge.c
1481
cmdq_processed[0] += e->Cmdq0CreditReturn;
drivers/net/ethernet/chelsio/cxgb/sge.c
1482
cmdq_processed[1] += e->Cmdq1CreditReturn;
drivers/net/ethernet/chelsio/cxgb/sge.c
1498
if (likely(e->DataValid)) {
drivers/net/ethernet/chelsio/cxgb/sge.c
1499
struct freelQ *fl = &sge->freelQ[e->FreelistQid];
drivers/net/ethernet/chelsio/cxgb/sge.c
1501
BUG_ON(!e->Sop || !e->Eop);
drivers/net/ethernet/chelsio/cxgb/sge.c
1502
if (unlikely(e->Offload))
drivers/net/ethernet/chelsio/cxgb/sge.c
1505
sge_rx(sge, fl, e->BufferLength);
drivers/net/ethernet/chelsio/cxgb/sge.c
1523
e++;
drivers/net/ethernet/chelsio/cxgb/sge.c
1527
e = q->entries;
drivers/net/ethernet/chelsio/cxgb/sge.c
1529
prefetch(e);
drivers/net/ethernet/chelsio/cxgb/sge.c
1546
const struct respQ_e *e = &Q->entries[Q->cidx];
drivers/net/ethernet/chelsio/cxgb/sge.c
1548
return e->GenerationBit == Q->genbit;
drivers/net/ethernet/chelsio/cxgb/sge.c
1563
struct respQ_e *e = &q->entries[q->cidx];
drivers/net/ethernet/chelsio/cxgb/sge.c
1564
const struct freelQ *fl = &sge->freelQ[e->FreelistQid];
drivers/net/ethernet/chelsio/cxgb/sge.c
1569
if (e->DataValid)
drivers/net/ethernet/chelsio/cxgb/sge.c
1573
flags |= e->Qsleeping;
drivers/net/ethernet/chelsio/cxgb/sge.c
1575
cmdq_processed[0] += e->Cmdq0CreditReturn;
drivers/net/ethernet/chelsio/cxgb/sge.c
1576
cmdq_processed[1] += e->Cmdq1CreditReturn;
drivers/net/ethernet/chelsio/cxgb/sge.c
1578
e++;
drivers/net/ethernet/chelsio/cxgb/sge.c
1582
e = q->entries;
drivers/net/ethernet/chelsio/cxgb/sge.c
1584
prefetch(e);
drivers/net/ethernet/chelsio/cxgb/sge.c
1591
} while (e->GenerationBit == q->genbit && !e->DataValid);
drivers/net/ethernet/chelsio/cxgb/sge.c
1596
return e->GenerationBit == q->genbit;
drivers/net/ethernet/chelsio/cxgb/sge.c
1800
const struct ethhdr *e = data;
drivers/net/ethernet/chelsio/cxgb/sge.c
1802
return e->h_proto == htons(ETH_P_8021Q) ? VLAN_ETH_HLEN : ETH_HLEN;
drivers/net/ethernet/chelsio/cxgb/sge.c
825
struct freelQ_e *e = &q->entries[q->pidx];
drivers/net/ethernet/chelsio/cxgb/sge.c
844
e->addr_lo = (u32)mapping;
drivers/net/ethernet/chelsio/cxgb/sge.c
845
e->addr_hi = (u64)mapping >> 32;
drivers/net/ethernet/chelsio/cxgb/sge.c
846
e->len_gen = V_CMD_LEN(dma_len) | V_CMD_GEN1(q->genbit);
drivers/net/ethernet/chelsio/cxgb/sge.c
848
e->gen2 = V_CMD_GEN2(q->genbit);
drivers/net/ethernet/chelsio/cxgb/sge.c
850
e++;
drivers/net/ethernet/chelsio/cxgb/sge.c
856
e = q->entries;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_defs.h
101
e = atid2entry(t, tid);
drivers/net/ethernet/chelsio/cxgb3/cxgb3_defs.h
102
if ((void *)e->next >= (void *)t->tid_tab &&
drivers/net/ethernet/chelsio/cxgb3/cxgb3_defs.h
103
(void *)e->next < (void *)&t->atid_tab[t->natids])
drivers/net/ethernet/chelsio/cxgb3/cxgb3_defs.h
106
return &e->t3c_tid;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_defs.h
77
union listen_entry *e;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_defs.h
82
e = stid2entry(t, tid);
drivers/net/ethernet/chelsio/cxgb3/cxgb3_defs.h
83
if ((void *)e->next >= (void *)t->tid_tab &&
drivers/net/ethernet/chelsio/cxgb3/cxgb3_defs.h
84
(void *)e->next < (void *)&t->atid_tab[t->natids])
drivers/net/ethernet/chelsio/cxgb3/cxgb3_defs.h
87
return &e->t3c_tid;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_defs.h
96
union active_open_entry *e;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
1950
static void get_sge_param(struct net_device *dev, struct ethtool_ringparam *e,
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
1958
e->rx_max_pending = MAX_RX_BUFFERS;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
1959
e->rx_jumbo_max_pending = MAX_RX_JUMBO_BUFFERS;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
1960
e->tx_max_pending = MAX_TXQ_ENTRIES;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
1962
e->rx_pending = q->fl_size;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
1963
e->rx_mini_pending = q->rspq_size;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
1964
e->rx_jumbo_pending = q->jumbo_size;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
1965
e->tx_pending = q->txq_size[0];
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
1968
static int set_sge_param(struct net_device *dev, struct ethtool_ringparam *e,
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
1977
if (e->rx_pending > MAX_RX_BUFFERS ||
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
1978
e->rx_jumbo_pending > MAX_RX_JUMBO_BUFFERS ||
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
1979
e->tx_pending > MAX_TXQ_ENTRIES ||
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
1980
e->rx_mini_pending > MAX_RSPQ_ENTRIES ||
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
1981
e->rx_mini_pending < MIN_RSPQ_ENTRIES ||
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
1982
e->rx_pending < MIN_FL_ENTRIES ||
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
1983
e->rx_jumbo_pending < MIN_FL_ENTRIES ||
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
1984
e->tx_pending < adapter->params.nports * MIN_TXQ_ENTRIES)
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
1992
q->rspq_size = e->rx_mini_pending;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
1993
q->fl_size = e->rx_pending;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
1994
q->jumbo_size = e->rx_jumbo_pending;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
1995
q->txq_size[0] = e->tx_pending;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
1996
q->txq_size[1] = e->tx_pending;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
1997
q->txq_size[2] = e->tx_pending;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2037
static int get_eeprom(struct net_device *dev, struct ethtool_eeprom *e,
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2044
e->magic = EEPROM_MAGIC;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2045
cnt = pci_read_vpd(adapter->pdev, e->offset, e->len, data);
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2049
e->len = cnt;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
1049
static void set_l2t_ix(struct t3cdev *tdev, u32 tid, struct l2t_entry *e)
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
1067
req->val = cpu_to_be64(V_TCB_L2T_IX(e->idx));
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
1080
struct l2t_entry *e;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
1091
e = t3_l2t_get(tdev, new, dev, daddr);
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
1092
if (!e) {
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
1103
update_tcb = te->client->redirect(te->ctx, old, new, e);
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
1106
l2t_hold(L2DATA(tdev), e);
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
1108
set_l2t_ix(tdev, tid, e);
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
1112
l2t_release(tdev, e);
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
461
struct ch_embedded_info *e = data;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
464
t3_get_fw_version(adapter, &e->fw_vers);
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
465
t3_get_tp_version(adapter, &e->tp_vers);
drivers/net/ethernet/chelsio/cxgb3/l2t.c
101
OPCODE_TID(req) = htonl(MK_OPCODE_TID(CPL_L2T_WRITE_REQ, e->idx));
drivers/net/ethernet/chelsio/cxgb3/l2t.c
102
req->params = htonl(V_L2T_W_IDX(e->idx) | V_L2T_W_IFF(e->smt_idx) |
drivers/net/ethernet/chelsio/cxgb3/l2t.c
103
V_L2T_W_VLAN(e->vlan & VLAN_VID_MASK) |
drivers/net/ethernet/chelsio/cxgb3/l2t.c
104
V_L2T_W_PRIO(vlan_prio(e)));
drivers/net/ethernet/chelsio/cxgb3/l2t.c
105
memcpy(e->dmac, e->neigh->ha, sizeof(e->dmac));
drivers/net/ethernet/chelsio/cxgb3/l2t.c
106
memcpy(req->dst_mac, e->dmac, sizeof(req->dst_mac));
drivers/net/ethernet/chelsio/cxgb3/l2t.c
110
skb_queue_walk_safe(&e->arpq, skb, tmp) {
drivers/net/ethernet/chelsio/cxgb3/l2t.c
111
__skb_unlink(skb, &e->arpq);
drivers/net/ethernet/chelsio/cxgb3/l2t.c
114
e->state = L2T_STATE_VALID;
drivers/net/ethernet/chelsio/cxgb3/l2t.c
123
static inline void arpq_enqueue(struct l2t_entry *e, struct sk_buff *skb)
drivers/net/ethernet/chelsio/cxgb3/l2t.c
125
__skb_queue_tail(&e->arpq, skb);
drivers/net/ethernet/chelsio/cxgb3/l2t.c
129
struct l2t_entry *e)
drivers/net/ethernet/chelsio/cxgb3/l2t.c
132
switch (e->state) {
drivers/net/ethernet/chelsio/cxgb3/l2t.c
134
neigh_event_send(e->neigh, NULL);
drivers/net/ethernet/chelsio/cxgb3/l2t.c
135
spin_lock_bh(&e->lock);
drivers/net/ethernet/chelsio/cxgb3/l2t.c
136
if (e->state == L2T_STATE_STALE)
drivers/net/ethernet/chelsio/cxgb3/l2t.c
137
e->state = L2T_STATE_VALID;
drivers/net/ethernet/chelsio/cxgb3/l2t.c
138
spin_unlock_bh(&e->lock);
drivers/net/ethernet/chelsio/cxgb3/l2t.c
143
spin_lock_bh(&e->lock);
drivers/net/ethernet/chelsio/cxgb3/l2t.c
144
if (e->state != L2T_STATE_RESOLVING) {
drivers/net/ethernet/chelsio/cxgb3/l2t.c
146
spin_unlock_bh(&e->lock);
drivers/net/ethernet/chelsio/cxgb3/l2t.c
149
arpq_enqueue(e, skb);
drivers/net/ethernet/chelsio/cxgb3/l2t.c
150
spin_unlock_bh(&e->lock);
drivers/net/ethernet/chelsio/cxgb3/l2t.c
160
if (!neigh_event_send(e->neigh, NULL)) {
drivers/net/ethernet/chelsio/cxgb3/l2t.c
166
spin_lock_bh(&e->lock);
drivers/net/ethernet/chelsio/cxgb3/l2t.c
167
if (!skb_queue_empty(&e->arpq))
drivers/net/ethernet/chelsio/cxgb3/l2t.c
168
setup_l2e_send_pending(dev, skb, e);
drivers/net/ethernet/chelsio/cxgb3/l2t.c
171
spin_unlock_bh(&e->lock);
drivers/net/ethernet/chelsio/cxgb3/l2t.c
184
struct l2t_entry *end, *e, **p;
drivers/net/ethernet/chelsio/cxgb3/l2t.c
190
for (e = d->rover, end = &d->l2tab[d->nentries]; e != end; ++e)
drivers/net/ethernet/chelsio/cxgb3/l2t.c
191
if (atomic_read(&e->refcnt) == 0)
drivers/net/ethernet/chelsio/cxgb3/l2t.c
194
for (e = &d->l2tab[1]; atomic_read(&e->refcnt); ++e) ;
drivers/net/ethernet/chelsio/cxgb3/l2t.c
196
d->rover = e + 1;
drivers/net/ethernet/chelsio/cxgb3/l2t.c
203
if (e->state != L2T_STATE_UNUSED) {
drivers/net/ethernet/chelsio/cxgb3/l2t.c
204
int hash = arp_hash(e->addr, e->ifindex, d);
drivers/net/ethernet/chelsio/cxgb3/l2t.c
207
if (*p == e) {
drivers/net/ethernet/chelsio/cxgb3/l2t.c
208
*p = e->next;
drivers/net/ethernet/chelsio/cxgb3/l2t.c
211
e->state = L2T_STATE_UNUSED;
drivers/net/ethernet/chelsio/cxgb3/l2t.c
213
return e;
drivers/net/ethernet/chelsio/cxgb3/l2t.c
227
void t3_l2e_free(struct l2t_data *d, struct l2t_entry *e)
drivers/net/ethernet/chelsio/cxgb3/l2t.c
229
spin_lock_bh(&e->lock);
drivers/net/ethernet/chelsio/cxgb3/l2t.c
230
if (atomic_read(&e->refcnt) == 0) { /* hasn't been recycled */
drivers/net/ethernet/chelsio/cxgb3/l2t.c
231
if (e->neigh) {
drivers/net/ethernet/chelsio/cxgb3/l2t.c
232
neigh_release(e->neigh);
drivers/net/ethernet/chelsio/cxgb3/l2t.c
233
e->neigh = NULL;
drivers/net/ethernet/chelsio/cxgb3/l2t.c
236
spin_unlock_bh(&e->lock);
drivers/net/ethernet/chelsio/cxgb3/l2t.c
246
static inline void reuse_entry(struct l2t_entry *e, struct neighbour *neigh)
drivers/net/ethernet/chelsio/cxgb3/l2t.c
250
spin_lock(&e->lock); /* avoid race with t3_l2t_free */
drivers/net/ethernet/chelsio/cxgb3/l2t.c
252
if (neigh != e->neigh)
drivers/net/ethernet/chelsio/cxgb3/l2t.c
253
neigh_replace(e, neigh);
drivers/net/ethernet/chelsio/cxgb3/l2t.c
255
if (memcmp(e->dmac, neigh->ha, sizeof(e->dmac)) ||
drivers/net/ethernet/chelsio/cxgb3/l2t.c
257
e->state = L2T_STATE_RESOLVING;
drivers/net/ethernet/chelsio/cxgb3/l2t.c
259
e->state = L2T_STATE_VALID;
drivers/net/ethernet/chelsio/cxgb3/l2t.c
261
e->state = L2T_STATE_STALE;
drivers/net/ethernet/chelsio/cxgb3/l2t.c
262
spin_unlock(&e->lock);
drivers/net/ethernet/chelsio/cxgb3/l2t.c
268
struct l2t_entry *e = NULL;
drivers/net/ethernet/chelsio/cxgb3/l2t.c
297
for (e = d->l2tab[hash].first; e; e = e->next)
drivers/net/ethernet/chelsio/cxgb3/l2t.c
298
if (e->addr == addr && e->ifindex == ifidx &&
drivers/net/ethernet/chelsio/cxgb3/l2t.c
299
e->smt_idx == smt_idx) {
drivers/net/ethernet/chelsio/cxgb3/l2t.c
300
l2t_hold(d, e);
drivers/net/ethernet/chelsio/cxgb3/l2t.c
301
if (atomic_read(&e->refcnt) == 1)
drivers/net/ethernet/chelsio/cxgb3/l2t.c
302
reuse_entry(e, neigh);
drivers/net/ethernet/chelsio/cxgb3/l2t.c
307
e = alloc_l2e(d);
drivers/net/ethernet/chelsio/cxgb3/l2t.c
308
if (e) {
drivers/net/ethernet/chelsio/cxgb3/l2t.c
309
spin_lock(&e->lock); /* avoid race with t3_l2t_free */
drivers/net/ethernet/chelsio/cxgb3/l2t.c
310
e->next = d->l2tab[hash].first;
drivers/net/ethernet/chelsio/cxgb3/l2t.c
311
d->l2tab[hash].first = e;
drivers/net/ethernet/chelsio/cxgb3/l2t.c
312
e->state = L2T_STATE_RESOLVING;
drivers/net/ethernet/chelsio/cxgb3/l2t.c
313
e->addr = addr;
drivers/net/ethernet/chelsio/cxgb3/l2t.c
314
e->ifindex = ifidx;
drivers/net/ethernet/chelsio/cxgb3/l2t.c
315
e->smt_idx = smt_idx;
drivers/net/ethernet/chelsio/cxgb3/l2t.c
316
atomic_set(&e->refcnt, 1);
drivers/net/ethernet/chelsio/cxgb3/l2t.c
317
neigh_replace(e, neigh);
drivers/net/ethernet/chelsio/cxgb3/l2t.c
319
e->vlan = vlan_dev_vlan_id(neigh->dev);
drivers/net/ethernet/chelsio/cxgb3/l2t.c
321
e->vlan = VLAN_NONE;
drivers/net/ethernet/chelsio/cxgb3/l2t.c
322
spin_unlock(&e->lock);
drivers/net/ethernet/chelsio/cxgb3/l2t.c
330
return e;
drivers/net/ethernet/chelsio/cxgb3/l2t.c
365
struct l2t_entry *e;
drivers/net/ethernet/chelsio/cxgb3/l2t.c
372
for (e = d->l2tab[hash].first; e; e = e->next)
drivers/net/ethernet/chelsio/cxgb3/l2t.c
373
if (e->addr == addr && e->ifindex == ifidx) {
drivers/net/ethernet/chelsio/cxgb3/l2t.c
374
spin_lock(&e->lock);
drivers/net/ethernet/chelsio/cxgb3/l2t.c
384
if (atomic_read(&e->refcnt)) {
drivers/net/ethernet/chelsio/cxgb3/l2t.c
385
if (neigh != e->neigh)
drivers/net/ethernet/chelsio/cxgb3/l2t.c
386
neigh_replace(e, neigh);
drivers/net/ethernet/chelsio/cxgb3/l2t.c
388
if (e->state == L2T_STATE_RESOLVING) {
drivers/net/ethernet/chelsio/cxgb3/l2t.c
390
skb_queue_splice_init(&e->arpq, &arpq);
drivers/net/ethernet/chelsio/cxgb3/l2t.c
392
setup_l2e_send_pending(dev, NULL, e);
drivers/net/ethernet/chelsio/cxgb3/l2t.c
394
e->state = neigh->nud_state & NUD_CONNECTED ?
drivers/net/ethernet/chelsio/cxgb3/l2t.c
396
if (!ether_addr_equal(e->dmac, neigh->ha))
drivers/net/ethernet/chelsio/cxgb3/l2t.c
397
setup_l2e_send_pending(dev, NULL, e);
drivers/net/ethernet/chelsio/cxgb3/l2t.c
400
spin_unlock_bh(&e->lock);
drivers/net/ethernet/chelsio/cxgb3/l2t.c
63
static inline unsigned int vlan_prio(const struct l2t_entry *e)
drivers/net/ethernet/chelsio/cxgb3/l2t.c
65
return e->vlan >> 13;
drivers/net/ethernet/chelsio/cxgb3/l2t.c
74
static inline void neigh_replace(struct l2t_entry *e, struct neighbour *n)
drivers/net/ethernet/chelsio/cxgb3/l2t.c
77
if (e->neigh)
drivers/net/ethernet/chelsio/cxgb3/l2t.c
78
neigh_release(e->neigh);
drivers/net/ethernet/chelsio/cxgb3/l2t.c
79
e->neigh = n;
drivers/net/ethernet/chelsio/cxgb3/l2t.c
88
struct l2t_entry *e)
drivers/net/ethernet/chelsio/cxgb3/l2t.h
110
void t3_l2e_free(struct l2t_data *d, struct l2t_entry *e);
drivers/net/ethernet/chelsio/cxgb3/l2t.h
115
struct l2t_entry *e);
drivers/net/ethernet/chelsio/cxgb3/l2t.h
121
struct l2t_entry *e)
drivers/net/ethernet/chelsio/cxgb3/l2t.h
123
if (likely(e->state == L2T_STATE_VALID))
drivers/net/ethernet/chelsio/cxgb3/l2t.h
125
return t3_l2t_send_slow(dev, skb, e);
drivers/net/ethernet/chelsio/cxgb3/l2t.h
128
static inline void l2t_release(struct t3cdev *t, struct l2t_entry *e)
drivers/net/ethernet/chelsio/cxgb3/l2t.h
135
if (atomic_dec_and_test(&e->refcnt) && d)
drivers/net/ethernet/chelsio/cxgb3/l2t.h
136
t3_l2e_free(d, e);
drivers/net/ethernet/chelsio/cxgb3/l2t.h
141
static inline void l2t_hold(struct l2t_data *d, struct l2t_entry *e)
drivers/net/ethernet/chelsio/cxgb3/l2t.h
143
if (d && atomic_add_return(1, &e->refcnt) == 1) /* 0 -> 1 transition */
drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.h
278
struct cudbg_qdesc_entry *cudbg_next_qdesc(struct cudbg_qdesc_entry *e)
drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.h
281
((u8 *)e + sizeof(*e) + e->data_size);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_cudbg.c
115
const struct cxgb4_collect_entity *e = &e_arr[i];
drivers/net/ethernet/chelsio/cxgb4/cxgb4_cudbg.c
117
entity_hdr = cudbg_get_entity_hdr(buf, e->entity);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_cudbg.c
118
entity_hdr->entity_type = e->entity;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_cudbg.c
121
ret = e->collect_cb(pdbg_init, dbg_buff, &cudbg_err);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
1002
struct fw_devlog_e *e;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
1010
e = &dinfo->log[index];
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
1011
if (e->timestamp == 0)
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
1019
be32_to_cpu(e->seqno),
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
1020
be64_to_cpu(e->timestamp),
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
1021
(e->level < ARRAY_SIZE(devlog_level_strings)
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
1022
? devlog_level_strings[e->level]
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
1024
(e->facility < ARRAY_SIZE(devlog_facility_strings)
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
1025
? devlog_facility_strings[e->facility]
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
1027
seq_printf(seq, e->fmt,
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
1028
be32_to_cpu(e->params[0]),
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
1029
be32_to_cpu(e->params[1]),
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
1030
be32_to_cpu(e->params[2]),
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
1031
be32_to_cpu(e->params[3]),
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
1032
be32_to_cpu(e->params[4]),
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
1033
be32_to_cpu(e->params[5]),
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
1034
be32_to_cpu(e->params[6]),
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
1035
be32_to_cpu(e->params[7]));
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
1121
struct fw_devlog_e *e = &dinfo->log[index];
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
1124
if (e->timestamp == 0)
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
1127
seqno = be32_to_cpu(e->seqno);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c
1211
static int get_eeprom(struct net_device *dev, struct ethtool_eeprom *e,
drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c
1221
e->magic = EEPROM_MAGIC;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c
1222
for (i = e->offset & ~3; !err && i < e->offset + e->len; i += 4)
drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c
1226
memcpy(data, buf + e->offset, e->len);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c
893
static void get_sge_param(struct net_device *dev, struct ethtool_ringparam *e,
drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c
900
e->rx_max_pending = MAX_RX_BUFFERS;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c
901
e->rx_mini_max_pending = MAX_RSPQ_ENTRIES;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c
902
e->rx_jumbo_max_pending = 0;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c
903
e->tx_max_pending = MAX_TXQ_ENTRIES;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c
905
e->rx_pending = s->ethrxq[pi->first_qset].fl.size - 8;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c
906
e->rx_mini_pending = s->ethrxq[pi->first_qset].rspq.size;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c
907
e->rx_jumbo_pending = 0;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c
908
e->tx_pending = s->ethtxq[pi->first_qset].q.size;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c
911
static int set_sge_param(struct net_device *dev, struct ethtool_ringparam *e,
drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c
920
if (e->rx_pending > MAX_RX_BUFFERS || e->rx_jumbo_pending ||
drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c
921
e->tx_pending > MAX_TXQ_ENTRIES ||
drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c
922
e->rx_mini_pending > MAX_RSPQ_ENTRIES ||
drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c
923
e->rx_mini_pending < MIN_RSPQ_ENTRIES ||
drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c
924
e->rx_pending < MIN_FL_ENTRIES || e->tx_pending < MIN_TXQ_ENTRIES)
drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c
931
s->ethtxq[pi->first_qset + i].q.size = e->tx_pending;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c
932
s->ethrxq[pi->first_qset + i].fl.size = e->rx_pending + 8;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c
933
s->ethrxq[pi->first_qset + i].rspq.size = e->rx_mini_pending;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
3488
struct ch_sched_class *e;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
3506
e = cxgb4_sched_queue_lookup(dev, &qe);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
3507
if (e && e->info.u.params.level != SCHED_CLASS_LEVEL_CL_RL) {
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
3510
index, e->idx, e->info.u.params.level);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
3555
e = cxgb4_sched_class_alloc(dev, &p);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
3556
if (!e)
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
3562
qe.class = e->idx;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_matchall.c
113
e = cxgb4_sched_queue_lookup(dev, &qe);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_matchall.c
114
if (e && e->info.u.params.level != SCHED_CLASS_LEVEL_CH_RL) {
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_matchall.c
183
struct ch_sched_class *e;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_matchall.c
200
e = cxgb4_sched_class_alloc(dev, &p);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_matchall.c
201
if (!e) {
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_matchall.c
207
ret = cxgb4_matchall_tc_bind_queues(dev, e->idx);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_matchall.c
214
tc_port_matchall->egress.hwtc = e->idx;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_matchall.c
220
cxgb4_sched_class_free(dev, e->idx);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_matchall.c
59
struct ch_sched_class *e;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_mqprio.c
330
struct ch_sched_class *e;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_mqprio.c
348
e = cxgb4_sched_class_alloc(dev, &p);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_mqprio.c
349
if (!e) {
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_mqprio.c
354
tc_port_mqprio->tc_hwtc_map[i] = e->idx;
drivers/net/ethernet/chelsio/cxgb4/l2t.c
118
static int addreq(const struct l2t_entry *e, const u32 *addr)
drivers/net/ethernet/chelsio/cxgb4/l2t.c
120
if (e->v6)
drivers/net/ethernet/chelsio/cxgb4/l2t.c
121
return (e->addr[0] ^ addr[0]) | (e->addr[1] ^ addr[1]) |
drivers/net/ethernet/chelsio/cxgb4/l2t.c
122
(e->addr[2] ^ addr[2]) | (e->addr[3] ^ addr[3]);
drivers/net/ethernet/chelsio/cxgb4/l2t.c
123
return e->addr[0] ^ addr[0];
drivers/net/ethernet/chelsio/cxgb4/l2t.c
126
static void neigh_replace(struct l2t_entry *e, struct neighbour *n)
drivers/net/ethernet/chelsio/cxgb4/l2t.c
129
if (e->neigh)
drivers/net/ethernet/chelsio/cxgb4/l2t.c
130
neigh_release(e->neigh);
drivers/net/ethernet/chelsio/cxgb4/l2t.c
131
e->neigh = n;
drivers/net/ethernet/chelsio/cxgb4/l2t.c
138
static int write_l2e(struct adapter *adap, struct l2t_entry *e, int sync)
drivers/net/ethernet/chelsio/cxgb4/l2t.c
141
unsigned int l2t_idx = e->idx + d->l2t_start;
drivers/net/ethernet/chelsio/cxgb4/l2t.c
155
req->params = htons(L2T_W_PORT_V(e->lport) | L2T_W_NOREPLY_V(!sync));
drivers/net/ethernet/chelsio/cxgb4/l2t.c
157
req->vlan = htons(e->vlan);
drivers/net/ethernet/chelsio/cxgb4/l2t.c
158
if (e->neigh && !(e->neigh->dev->flags & IFF_LOOPBACK))
drivers/net/ethernet/chelsio/cxgb4/l2t.c
159
memcpy(e->dmac, e->neigh->ha, sizeof(e->dmac));
drivers/net/ethernet/chelsio/cxgb4/l2t.c
160
memcpy(req->dst_mac, e->dmac, sizeof(req->dst_mac));
drivers/net/ethernet/chelsio/cxgb4/l2t.c
164
if (sync && e->state != L2T_STATE_SWITCHING)
drivers/net/ethernet/chelsio/cxgb4/l2t.c
165
e->state = L2T_STATE_SYNC_WRITE;
drivers/net/ethernet/chelsio/cxgb4/l2t.c
173
static void send_pending(struct adapter *adap, struct l2t_entry *e)
drivers/net/ethernet/chelsio/cxgb4/l2t.c
177
while ((skb = __skb_dequeue(&e->arpq)) != NULL)
drivers/net/ethernet/chelsio/cxgb4/l2t.c
200
struct l2t_entry *e = &d->l2tab[l2t_idx - d->l2t_start];
drivers/net/ethernet/chelsio/cxgb4/l2t.c
202
spin_lock(&e->lock);
drivers/net/ethernet/chelsio/cxgb4/l2t.c
203
if (e->state != L2T_STATE_SWITCHING) {
drivers/net/ethernet/chelsio/cxgb4/l2t.c
204
send_pending(adap, e);
drivers/net/ethernet/chelsio/cxgb4/l2t.c
205
e->state = (e->neigh->nud_state & NUD_STALE) ?
drivers/net/ethernet/chelsio/cxgb4/l2t.c
208
spin_unlock(&e->lock);
drivers/net/ethernet/chelsio/cxgb4/l2t.c
216
static inline void arpq_enqueue(struct l2t_entry *e, struct sk_buff *skb)
drivers/net/ethernet/chelsio/cxgb4/l2t.c
218
__skb_queue_tail(&e->arpq, skb);
drivers/net/ethernet/chelsio/cxgb4/l2t.c
222
struct l2t_entry *e)
drivers/net/ethernet/chelsio/cxgb4/l2t.c
227
switch (e->state) {
drivers/net/ethernet/chelsio/cxgb4/l2t.c
229
neigh_event_send(e->neigh, NULL);
drivers/net/ethernet/chelsio/cxgb4/l2t.c
230
spin_lock_bh(&e->lock);
drivers/net/ethernet/chelsio/cxgb4/l2t.c
231
if (e->state == L2T_STATE_STALE)
drivers/net/ethernet/chelsio/cxgb4/l2t.c
232
e->state = L2T_STATE_VALID;
drivers/net/ethernet/chelsio/cxgb4/l2t.c
233
spin_unlock_bh(&e->lock);
drivers/net/ethernet/chelsio/cxgb4/l2t.c
239
spin_lock_bh(&e->lock);
drivers/net/ethernet/chelsio/cxgb4/l2t.c
240
if (e->state != L2T_STATE_SYNC_WRITE &&
drivers/net/ethernet/chelsio/cxgb4/l2t.c
241
e->state != L2T_STATE_RESOLVING) {
drivers/net/ethernet/chelsio/cxgb4/l2t.c
242
spin_unlock_bh(&e->lock);
drivers/net/ethernet/chelsio/cxgb4/l2t.c
245
arpq_enqueue(e, skb);
drivers/net/ethernet/chelsio/cxgb4/l2t.c
246
spin_unlock_bh(&e->lock);
drivers/net/ethernet/chelsio/cxgb4/l2t.c
248
if (e->state == L2T_STATE_RESOLVING &&
drivers/net/ethernet/chelsio/cxgb4/l2t.c
249
!neigh_event_send(e->neigh, NULL)) {
drivers/net/ethernet/chelsio/cxgb4/l2t.c
250
spin_lock_bh(&e->lock);
drivers/net/ethernet/chelsio/cxgb4/l2t.c
251
if (e->state == L2T_STATE_RESOLVING &&
drivers/net/ethernet/chelsio/cxgb4/l2t.c
252
!skb_queue_empty(&e->arpq))
drivers/net/ethernet/chelsio/cxgb4/l2t.c
253
write_l2e(adap, e, 1);
drivers/net/ethernet/chelsio/cxgb4/l2t.c
254
spin_unlock_bh(&e->lock);
drivers/net/ethernet/chelsio/cxgb4/l2t.c
266
struct l2t_entry *end, *e, **p;
drivers/net/ethernet/chelsio/cxgb4/l2t.c
272
for (e = d->rover, end = &d->l2tab[d->l2t_size]; e != end; ++e)
drivers/net/ethernet/chelsio/cxgb4/l2t.c
273
if (atomic_read(&e->refcnt) == 0)
drivers/net/ethernet/chelsio/cxgb4/l2t.c
276
for (e = d->l2tab; atomic_read(&e->refcnt); ++e)
drivers/net/ethernet/chelsio/cxgb4/l2t.c
279
d->rover = e + 1;
drivers/net/ethernet/chelsio/cxgb4/l2t.c
286
if (e->state < L2T_STATE_SWITCHING)
drivers/net/ethernet/chelsio/cxgb4/l2t.c
287
for (p = &d->l2tab[e->hash].first; *p; p = &(*p)->next)
drivers/net/ethernet/chelsio/cxgb4/l2t.c
288
if (*p == e) {
drivers/net/ethernet/chelsio/cxgb4/l2t.c
289
*p = e->next;
drivers/net/ethernet/chelsio/cxgb4/l2t.c
290
e->next = NULL;
drivers/net/ethernet/chelsio/cxgb4/l2t.c
294
e->state = L2T_STATE_UNUSED;
drivers/net/ethernet/chelsio/cxgb4/l2t.c
295
return e;
drivers/net/ethernet/chelsio/cxgb4/l2t.c
301
struct l2t_entry *end, *e, **p;
drivers/net/ethernet/chelsio/cxgb4/l2t.c
304
for (e = &d->l2tab[0], end = &d->l2tab[d->l2t_size]; e != end; ++e) {
drivers/net/ethernet/chelsio/cxgb4/l2t.c
305
if (atomic_read(&e->refcnt) == 0) {
drivers/net/ethernet/chelsio/cxgb4/l2t.c
307
first_free = e;
drivers/net/ethernet/chelsio/cxgb4/l2t.c
309
if (e->state == L2T_STATE_SWITCHING) {
drivers/net/ethernet/chelsio/cxgb4/l2t.c
310
if (ether_addr_equal(e->dmac, dmac) &&
drivers/net/ethernet/chelsio/cxgb4/l2t.c
311
(e->vlan == vlan) && (e->lport == port))
drivers/net/ethernet/chelsio/cxgb4/l2t.c
318
e = first_free;
drivers/net/ethernet/chelsio/cxgb4/l2t.c
328
if (e->state < L2T_STATE_SWITCHING)
drivers/net/ethernet/chelsio/cxgb4/l2t.c
329
for (p = &d->l2tab[e->hash].first; *p; p = &(*p)->next)
drivers/net/ethernet/chelsio/cxgb4/l2t.c
330
if (*p == e) {
drivers/net/ethernet/chelsio/cxgb4/l2t.c
331
*p = e->next;
drivers/net/ethernet/chelsio/cxgb4/l2t.c
332
e->next = NULL;
drivers/net/ethernet/chelsio/cxgb4/l2t.c
335
e->state = L2T_STATE_UNUSED;
drivers/net/ethernet/chelsio/cxgb4/l2t.c
338
return e;
drivers/net/ethernet/chelsio/cxgb4/l2t.c
351
static void _t4_l2e_free(struct l2t_entry *e)
drivers/net/ethernet/chelsio/cxgb4/l2t.c
355
if (atomic_read(&e->refcnt) == 0) { /* hasn't been recycled */
drivers/net/ethernet/chelsio/cxgb4/l2t.c
356
if (e->neigh) {
drivers/net/ethernet/chelsio/cxgb4/l2t.c
357
neigh_release(e->neigh);
drivers/net/ethernet/chelsio/cxgb4/l2t.c
358
e->neigh = NULL;
drivers/net/ethernet/chelsio/cxgb4/l2t.c
360
__skb_queue_purge(&e->arpq);
drivers/net/ethernet/chelsio/cxgb4/l2t.c
363
d = container_of(e, struct l2t_data, l2tab[e->idx]);
drivers/net/ethernet/chelsio/cxgb4/l2t.c
368
static void t4_l2e_free(struct l2t_entry *e)
drivers/net/ethernet/chelsio/cxgb4/l2t.c
372
spin_lock_bh(&e->lock);
drivers/net/ethernet/chelsio/cxgb4/l2t.c
373
if (atomic_read(&e->refcnt) == 0) { /* hasn't been recycled */
drivers/net/ethernet/chelsio/cxgb4/l2t.c
374
if (e->neigh) {
drivers/net/ethernet/chelsio/cxgb4/l2t.c
375
neigh_release(e->neigh);
drivers/net/ethernet/chelsio/cxgb4/l2t.c
376
e->neigh = NULL;
drivers/net/ethernet/chelsio/cxgb4/l2t.c
378
__skb_queue_purge(&e->arpq);
drivers/net/ethernet/chelsio/cxgb4/l2t.c
380
spin_unlock_bh(&e->lock);
drivers/net/ethernet/chelsio/cxgb4/l2t.c
382
d = container_of(e, struct l2t_data, l2tab[e->idx]);
drivers/net/ethernet/chelsio/cxgb4/l2t.c
386
void cxgb4_l2t_release(struct l2t_entry *e)
drivers/net/ethernet/chelsio/cxgb4/l2t.c
388
if (atomic_dec_and_test(&e->refcnt))
drivers/net/ethernet/chelsio/cxgb4/l2t.c
389
t4_l2e_free(e);
drivers/net/ethernet/chelsio/cxgb4/l2t.c
397
static void reuse_entry(struct l2t_entry *e, struct neighbour *neigh)
drivers/net/ethernet/chelsio/cxgb4/l2t.c
401
spin_lock(&e->lock); /* avoid race with t4_l2t_free */
drivers/net/ethernet/chelsio/cxgb4/l2t.c
402
if (neigh != e->neigh)
drivers/net/ethernet/chelsio/cxgb4/l2t.c
403
neigh_replace(e, neigh);
drivers/net/ethernet/chelsio/cxgb4/l2t.c
405
if (memcmp(e->dmac, neigh->ha, sizeof(e->dmac)) ||
drivers/net/ethernet/chelsio/cxgb4/l2t.c
407
e->state = L2T_STATE_RESOLVING;
drivers/net/ethernet/chelsio/cxgb4/l2t.c
409
e->state = L2T_STATE_VALID;
drivers/net/ethernet/chelsio/cxgb4/l2t.c
411
e->state = L2T_STATE_STALE;
drivers/net/ethernet/chelsio/cxgb4/l2t.c
412
spin_unlock(&e->lock);
drivers/net/ethernet/chelsio/cxgb4/l2t.c
421
struct l2t_entry *e;
drivers/net/ethernet/chelsio/cxgb4/l2t.c
440
for (e = d->l2tab[hash].first; e; e = e->next)
drivers/net/ethernet/chelsio/cxgb4/l2t.c
441
if (!addreq(e, addr) && e->ifindex == ifidx &&
drivers/net/ethernet/chelsio/cxgb4/l2t.c
442
e->vlan == vlan && e->lport == lport) {
drivers/net/ethernet/chelsio/cxgb4/l2t.c
443
l2t_hold(d, e);
drivers/net/ethernet/chelsio/cxgb4/l2t.c
444
if (atomic_read(&e->refcnt) == 1)
drivers/net/ethernet/chelsio/cxgb4/l2t.c
445
reuse_entry(e, neigh);
drivers/net/ethernet/chelsio/cxgb4/l2t.c
450
e = alloc_l2e(d);
drivers/net/ethernet/chelsio/cxgb4/l2t.c
451
if (e) {
drivers/net/ethernet/chelsio/cxgb4/l2t.c
452
spin_lock(&e->lock); /* avoid race with t4_l2t_free */
drivers/net/ethernet/chelsio/cxgb4/l2t.c
453
e->state = L2T_STATE_RESOLVING;
drivers/net/ethernet/chelsio/cxgb4/l2t.c
455
memcpy(e->dmac, physdev->dev_addr, sizeof(e->dmac));
drivers/net/ethernet/chelsio/cxgb4/l2t.c
456
memcpy(e->addr, addr, addr_len);
drivers/net/ethernet/chelsio/cxgb4/l2t.c
457
e->ifindex = ifidx;
drivers/net/ethernet/chelsio/cxgb4/l2t.c
458
e->hash = hash;
drivers/net/ethernet/chelsio/cxgb4/l2t.c
459
e->lport = lport;
drivers/net/ethernet/chelsio/cxgb4/l2t.c
460
e->v6 = addr_len == 16;
drivers/net/ethernet/chelsio/cxgb4/l2t.c
461
atomic_set(&e->refcnt, 1);
drivers/net/ethernet/chelsio/cxgb4/l2t.c
462
neigh_replace(e, neigh);
drivers/net/ethernet/chelsio/cxgb4/l2t.c
463
e->vlan = vlan;
drivers/net/ethernet/chelsio/cxgb4/l2t.c
464
e->next = d->l2tab[hash].first;
drivers/net/ethernet/chelsio/cxgb4/l2t.c
465
d->l2tab[hash].first = e;
drivers/net/ethernet/chelsio/cxgb4/l2t.c
466
spin_unlock(&e->lock);
drivers/net/ethernet/chelsio/cxgb4/l2t.c
470
return e;
drivers/net/ethernet/chelsio/cxgb4/l2t.c
516
struct l2t_entry *e;
drivers/net/ethernet/chelsio/cxgb4/l2t.c
520
for (e = d->l2tab[hash].first; e; e = e->next)
drivers/net/ethernet/chelsio/cxgb4/l2t.c
521
if (!addreq(e, addr) && e->ifindex == ifidx) {
drivers/net/ethernet/chelsio/cxgb4/l2t.c
522
spin_lock(&e->lock);
drivers/net/ethernet/chelsio/cxgb4/l2t.c
523
if (atomic_read(&e->refcnt))
drivers/net/ethernet/chelsio/cxgb4/l2t.c
525
spin_unlock(&e->lock);
drivers/net/ethernet/chelsio/cxgb4/l2t.c
534
if (neigh != e->neigh)
drivers/net/ethernet/chelsio/cxgb4/l2t.c
535
neigh_replace(e, neigh);
drivers/net/ethernet/chelsio/cxgb4/l2t.c
537
if (e->state == L2T_STATE_RESOLVING) {
drivers/net/ethernet/chelsio/cxgb4/l2t.c
539
arpq = &e->arpq;
drivers/net/ethernet/chelsio/cxgb4/l2t.c
541
!skb_queue_empty(&e->arpq)) {
drivers/net/ethernet/chelsio/cxgb4/l2t.c
542
write_l2e(adap, e, 1);
drivers/net/ethernet/chelsio/cxgb4/l2t.c
545
e->state = neigh->nud_state & NUD_CONNECTED ?
drivers/net/ethernet/chelsio/cxgb4/l2t.c
547
if (memcmp(e->dmac, neigh->ha, sizeof(e->dmac)))
drivers/net/ethernet/chelsio/cxgb4/l2t.c
548
write_l2e(adap, e, 0);
drivers/net/ethernet/chelsio/cxgb4/l2t.c
559
while ((skb = __skb_dequeue(&e->arpq)) != NULL) {
drivers/net/ethernet/chelsio/cxgb4/l2t.c
562
spin_unlock(&e->lock);
drivers/net/ethernet/chelsio/cxgb4/l2t.c
567
spin_lock(&e->lock);
drivers/net/ethernet/chelsio/cxgb4/l2t.c
570
spin_unlock_bh(&e->lock);
drivers/net/ethernet/chelsio/cxgb4/l2t.c
581
struct l2t_entry *e;
drivers/net/ethernet/chelsio/cxgb4/l2t.c
585
e = find_or_alloc_l2e(d, vlan, port, eth_addr);
drivers/net/ethernet/chelsio/cxgb4/l2t.c
586
if (e) {
drivers/net/ethernet/chelsio/cxgb4/l2t.c
587
spin_lock(&e->lock); /* avoid race with t4_l2t_free */
drivers/net/ethernet/chelsio/cxgb4/l2t.c
588
if (!atomic_read(&e->refcnt)) {
drivers/net/ethernet/chelsio/cxgb4/l2t.c
589
e->state = L2T_STATE_SWITCHING;
drivers/net/ethernet/chelsio/cxgb4/l2t.c
590
e->vlan = vlan;
drivers/net/ethernet/chelsio/cxgb4/l2t.c
591
e->lport = port;
drivers/net/ethernet/chelsio/cxgb4/l2t.c
592
ether_addr_copy(e->dmac, eth_addr);
drivers/net/ethernet/chelsio/cxgb4/l2t.c
593
atomic_set(&e->refcnt, 1);
drivers/net/ethernet/chelsio/cxgb4/l2t.c
594
ret = write_l2e(adap, e, 0);
drivers/net/ethernet/chelsio/cxgb4/l2t.c
596
_t4_l2e_free(e);
drivers/net/ethernet/chelsio/cxgb4/l2t.c
597
spin_unlock(&e->lock);
drivers/net/ethernet/chelsio/cxgb4/l2t.c
602
atomic_inc(&e->refcnt);
drivers/net/ethernet/chelsio/cxgb4/l2t.c
605
spin_unlock(&e->lock);
drivers/net/ethernet/chelsio/cxgb4/l2t.c
608
return e;
drivers/net/ethernet/chelsio/cxgb4/l2t.c
65
static inline unsigned int vlan_prio(const struct l2t_entry *e)
drivers/net/ethernet/chelsio/cxgb4/l2t.c
667
static char l2e_state(const struct l2t_entry *e)
drivers/net/ethernet/chelsio/cxgb4/l2t.c
669
switch (e->state) {
drivers/net/ethernet/chelsio/cxgb4/l2t.c
67
return e->vlan >> VLAN_PRIO_SHIFT;
drivers/net/ethernet/chelsio/cxgb4/l2t.c
674
return skb_queue_empty(&e->arpq) ? 'R' : 'A';
drivers/net/ethernet/chelsio/cxgb4/l2t.c
681
bool cxgb4_check_l2t_valid(struct l2t_entry *e)
drivers/net/ethernet/chelsio/cxgb4/l2t.c
685
spin_lock(&e->lock);
drivers/net/ethernet/chelsio/cxgb4/l2t.c
686
valid = (e->state == L2T_STATE_VALID);
drivers/net/ethernet/chelsio/cxgb4/l2t.c
687
spin_unlock(&e->lock);
drivers/net/ethernet/chelsio/cxgb4/l2t.c
70
static inline void l2t_hold(struct l2t_data *d, struct l2t_entry *e)
drivers/net/ethernet/chelsio/cxgb4/l2t.c
700
struct l2t_entry *e = v;
drivers/net/ethernet/chelsio/cxgb4/l2t.c
702
spin_lock_bh(&e->lock);
drivers/net/ethernet/chelsio/cxgb4/l2t.c
703
if (e->state == L2T_STATE_SWITCHING)
drivers/net/ethernet/chelsio/cxgb4/l2t.c
706
sprintf(ip, e->v6 ? "%pI6c" : "%pI4", e->addr);
drivers/net/ethernet/chelsio/cxgb4/l2t.c
708
e->idx + d->l2t_start, ip, e->dmac,
drivers/net/ethernet/chelsio/cxgb4/l2t.c
709
e->vlan & VLAN_VID_MASK, vlan_prio(e), e->lport,
drivers/net/ethernet/chelsio/cxgb4/l2t.c
710
l2e_state(e), atomic_read(&e->refcnt),
drivers/net/ethernet/chelsio/cxgb4/l2t.c
711
e->neigh ? e->neigh->dev->name : "");
drivers/net/ethernet/chelsio/cxgb4/l2t.c
712
spin_unlock_bh(&e->lock);
drivers/net/ethernet/chelsio/cxgb4/l2t.c
72
if (atomic_add_return(1, &e->refcnt) == 1) /* 0 -> 1 transition */
drivers/net/ethernet/chelsio/cxgb4/l2t.h
110
void cxgb4_l2t_release(struct l2t_entry *e);
drivers/net/ethernet/chelsio/cxgb4/l2t.h
112
struct l2t_entry *e);
drivers/net/ethernet/chelsio/cxgb4/l2t.h
123
bool cxgb4_check_l2t_valid(struct l2t_entry *e);
drivers/net/ethernet/chelsio/cxgb4/sched.c
125
struct ch_sched_class *e, *end;
drivers/net/ethernet/chelsio/cxgb4/sched.c
130
for (e = &s->tab[0]; e != end; ++e) {
drivers/net/ethernet/chelsio/cxgb4/sched.c
131
if (e->state == SCHED_STATE_UNUSED ||
drivers/net/ethernet/chelsio/cxgb4/sched.c
132
e->bind_type != type)
drivers/net/ethernet/chelsio/cxgb4/sched.c
139
list_for_each_entry(qe, &e->entry_list, list) {
drivers/net/ethernet/chelsio/cxgb4/sched.c
150
list_for_each_entry(fe, &e->entry_list, list) {
drivers/net/ethernet/chelsio/cxgb4/sched.c
190
struct ch_sched_class *e;
drivers/net/ethernet/chelsio/cxgb4/sched.c
206
e = &pi->sched_tbl->tab[qe->param.class];
drivers/net/ethernet/chelsio/cxgb4/sched.c
209
if (atomic_dec_and_test(&e->refcnt))
drivers/net/ethernet/chelsio/cxgb4/sched.c
210
cxgb4_sched_class_free(adap->port[pi->port_id], e->idx);
drivers/net/ethernet/chelsio/cxgb4/sched.c
221
struct ch_sched_class *e;
drivers/net/ethernet/chelsio/cxgb4/sched.c
244
e = &s->tab[qe->param.class];
drivers/net/ethernet/chelsio/cxgb4/sched.c
249
list_add_tail(&qe->list, &e->entry_list);
drivers/net/ethernet/chelsio/cxgb4/sched.c
250
e->bind_type = SCHED_QUEUE;
drivers/net/ethernet/chelsio/cxgb4/sched.c
251
atomic_inc(&e->refcnt);
drivers/net/ethernet/chelsio/cxgb4/sched.c
263
struct ch_sched_class *e;
drivers/net/ethernet/chelsio/cxgb4/sched.c
277
e = &pi->sched_tbl->tab[fe->param.class];
drivers/net/ethernet/chelsio/cxgb4/sched.c
280
if (atomic_dec_and_test(&e->refcnt))
drivers/net/ethernet/chelsio/cxgb4/sched.c
281
cxgb4_sched_class_free(adap->port[pi->port_id], e->idx);
drivers/net/ethernet/chelsio/cxgb4/sched.c
291
struct ch_sched_class *e;
drivers/net/ethernet/chelsio/cxgb4/sched.c
309
e = &s->tab[fe->param.class];
drivers/net/ethernet/chelsio/cxgb4/sched.c
314
list_add_tail(&fe->list, &e->entry_list);
drivers/net/ethernet/chelsio/cxgb4/sched.c
315
e->bind_type = SCHED_FLOWC;
drivers/net/ethernet/chelsio/cxgb4/sched.c
316
atomic_inc(&e->refcnt);
drivers/net/ethernet/chelsio/cxgb4/sched.c
325
struct ch_sched_class *e,
drivers/net/ethernet/chelsio/cxgb4/sched.c
328
if (!e)
drivers/net/ethernet/chelsio/cxgb4/sched.c
335
list_for_each_entry(qe, &e->entry_list, list)
drivers/net/ethernet/chelsio/cxgb4/sched.c
342
list_for_each_entry(fe, &e->entry_list, list)
drivers/net/ethernet/chelsio/cxgb4/sched.c
47
struct ch_sched_class *e;
drivers/net/ethernet/chelsio/cxgb4/sched.c
484
struct ch_sched_class *e, *end;
drivers/net/ethernet/chelsio/cxgb4/sched.c
489
for (e = &s->tab[0]; e != end; ++e) {
drivers/net/ethernet/chelsio/cxgb4/sched.c
490
if (e->state == SCHED_STATE_UNUSED) {
drivers/net/ethernet/chelsio/cxgb4/sched.c
491
found = e;
drivers/net/ethernet/chelsio/cxgb4/sched.c
50
e = &s->tab[p->u.params.class];
drivers/net/ethernet/chelsio/cxgb4/sched.c
505
for (e = &s->tab[0]; e != end; ++e) {
drivers/net/ethernet/chelsio/cxgb4/sched.c
506
if (e->state == SCHED_STATE_UNUSED)
drivers/net/ethernet/chelsio/cxgb4/sched.c
509
memcpy(&info, &e->info, sizeof(info));
drivers/net/ethernet/chelsio/cxgb4/sched.c
516
found = e;
drivers/net/ethernet/chelsio/cxgb4/sched.c
528
struct ch_sched_class *e = NULL;
drivers/net/ethernet/chelsio/cxgb4/sched.c
548
e = t4_sched_class_lookup(pi, p);
drivers/net/ethernet/chelsio/cxgb4/sched.c
550
if (!e) {
drivers/net/ethernet/chelsio/cxgb4/sched.c
554
e = t4_sched_class_lookup(pi, NULL);
drivers/net/ethernet/chelsio/cxgb4/sched.c
555
if (!e)
drivers/net/ethernet/chelsio/cxgb4/sched.c
559
np.u.params.class = e->idx;
drivers/net/ethernet/chelsio/cxgb4/sched.c
564
memcpy(&e->info, &np, sizeof(e->info));
drivers/net/ethernet/chelsio/cxgb4/sched.c
565
atomic_set(&e->refcnt, 0);
drivers/net/ethernet/chelsio/cxgb4/sched.c
566
e->state = SCHED_STATE_ACTIVE;
drivers/net/ethernet/chelsio/cxgb4/sched.c
569
return e;
drivers/net/ethernet/chelsio/cxgb4/sched.c
58
p->u.params.channel, e->idx,
drivers/net/ethernet/chelsio/cxgb4/sched.c
610
struct ch_sched_class *e;
drivers/net/ethernet/chelsio/cxgb4/sched.c
614
e = &s->tab[classid];
drivers/net/ethernet/chelsio/cxgb4/sched.c
615
if (!atomic_read(&e->refcnt) && e->state != SCHED_STATE_UNUSED) {
drivers/net/ethernet/chelsio/cxgb4/sched.c
621
memcpy(&p, &e->info, sizeof(p));
drivers/net/ethernet/chelsio/cxgb4/sched.c
638
e->state = SCHED_STATE_UNUSED;
drivers/net/ethernet/chelsio/cxgb4/sched.c
639
memset(&e->info, 0, sizeof(e->info));
drivers/net/ethernet/chelsio/cxgb4/sched.c
643
static void t4_sched_class_free(struct net_device *dev, struct ch_sched_class *e)
drivers/net/ethernet/chelsio/cxgb4/sched.c
647
t4_sched_class_unbind_all(pi, e, e->bind_type);
drivers/net/ethernet/chelsio/cxgb4/sched.c
648
cxgb4_sched_class_free(dev, e->idx);
drivers/net/ethernet/chelsio/cxgb4/sched.c
685
struct ch_sched_class *e;
drivers/net/ethernet/chelsio/cxgb4/sched.c
687
e = &s->tab[i];
drivers/net/ethernet/chelsio/cxgb4/sched.c
688
if (e->state == SCHED_STATE_ACTIVE)
drivers/net/ethernet/chelsio/cxgb4/sched.c
689
t4_sched_class_free(adap->port[j], e);
drivers/net/ethernet/chelsio/cxgb4/smt.c
100
if (e->refcnt == 0) { /* hasn't been recycled */
drivers/net/ethernet/chelsio/cxgb4/smt.c
101
e->state = SMT_STATE_UNUSED;
drivers/net/ethernet/chelsio/cxgb4/smt.c
111
void cxgb4_smt_release(struct smt_entry *e)
drivers/net/ethernet/chelsio/cxgb4/smt.c
113
spin_lock_bh(&e->lock);
drivers/net/ethernet/chelsio/cxgb4/smt.c
114
if ((--e->refcnt) == 0)
drivers/net/ethernet/chelsio/cxgb4/smt.c
115
t4_smte_free(e);
drivers/net/ethernet/chelsio/cxgb4/smt.c
116
spin_unlock_bh(&e->lock);
drivers/net/ethernet/chelsio/cxgb4/smt.c
126
struct smt_entry *e = &s->smtab[smtidx];
drivers/net/ethernet/chelsio/cxgb4/smt.c
131
spin_lock(&e->lock);
drivers/net/ethernet/chelsio/cxgb4/smt.c
132
e->state = SMT_STATE_ERROR;
drivers/net/ethernet/chelsio/cxgb4/smt.c
133
spin_unlock(&e->lock);
drivers/net/ethernet/chelsio/cxgb4/smt.c
138
static int write_smt_entry(struct adapter *adapter, struct smt_entry *e)
drivers/net/ethernet/chelsio/cxgb4/smt.c
161
row = (e->idx >> 1);
drivers/net/ethernet/chelsio/cxgb4/smt.c
162
if (e->idx & 1) {
drivers/net/ethernet/chelsio/cxgb4/smt.c
164
memcpy(req->src_mac1, e->src_mac, ETH_ALEN);
drivers/net/ethernet/chelsio/cxgb4/smt.c
170
memcpy(req->src_mac0, s->smtab[e->idx - 1].src_mac,
drivers/net/ethernet/chelsio/cxgb4/smt.c
174
memcpy(req->src_mac0, e->src_mac, ETH_ALEN);
drivers/net/ethernet/chelsio/cxgb4/smt.c
180
memcpy(req->src_mac1, s->smtab[e->idx + 1].src_mac,
drivers/net/ethernet/chelsio/cxgb4/smt.c
195
memcpy(req->src_mac0, s->smtab[e->idx].src_mac, ETH_ALEN);
drivers/net/ethernet/chelsio/cxgb4/smt.c
196
row = e->idx;
drivers/net/ethernet/chelsio/cxgb4/smt.c
200
htonl(MK_OPCODE_TID(CPL_SMT_WRITE_REQ, e->idx |
drivers/net/ethernet/chelsio/cxgb4/smt.c
213
struct smt_entry *e;
drivers/net/ethernet/chelsio/cxgb4/smt.c
216
e = find_or_alloc_smte(s, smac);
drivers/net/ethernet/chelsio/cxgb4/smt.c
217
if (e) {
drivers/net/ethernet/chelsio/cxgb4/smt.c
218
spin_lock(&e->lock);
drivers/net/ethernet/chelsio/cxgb4/smt.c
219
if (!e->refcnt) {
drivers/net/ethernet/chelsio/cxgb4/smt.c
220
e->refcnt = 1;
drivers/net/ethernet/chelsio/cxgb4/smt.c
221
e->state = SMT_STATE_SWITCHING;
drivers/net/ethernet/chelsio/cxgb4/smt.c
222
e->pfvf = pfvf;
drivers/net/ethernet/chelsio/cxgb4/smt.c
223
memcpy(e->src_mac, smac, ETH_ALEN);
drivers/net/ethernet/chelsio/cxgb4/smt.c
224
write_smt_entry(adap, e);
drivers/net/ethernet/chelsio/cxgb4/smt.c
226
++e->refcnt;
drivers/net/ethernet/chelsio/cxgb4/smt.c
228
spin_unlock(&e->lock);
drivers/net/ethernet/chelsio/cxgb4/smt.c
231
return e;
drivers/net/ethernet/chelsio/cxgb4/smt.c
68
struct smt_entry *e, *end;
drivers/net/ethernet/chelsio/cxgb4/smt.c
70
for (e = &s->smtab[0], end = &s->smtab[s->smt_size]; e != end; ++e) {
drivers/net/ethernet/chelsio/cxgb4/smt.c
71
if (e->refcnt == 0) {
drivers/net/ethernet/chelsio/cxgb4/smt.c
73
first_free = e;
drivers/net/ethernet/chelsio/cxgb4/smt.c
75
if (e->state == SMT_STATE_SWITCHING) {
drivers/net/ethernet/chelsio/cxgb4/smt.c
79
if (memcmp(e->src_mac, smac, ETH_ALEN) == 0)
drivers/net/ethernet/chelsio/cxgb4/smt.c
86
e = first_free;
drivers/net/ethernet/chelsio/cxgb4/smt.c
92
e->state = SMT_STATE_UNUSED;
drivers/net/ethernet/chelsio/cxgb4/smt.c
95
return e;
drivers/net/ethernet/chelsio/cxgb4/smt.c
98
static void t4_smte_free(struct smt_entry *e)
drivers/net/ethernet/chelsio/cxgb4/smt.h
74
void cxgb4_smt_release(struct smt_entry *e);
drivers/net/ethernet/chelsio/cxgb4/srq.c
59
struct srq_entry *e;
drivers/net/ethernet/chelsio/cxgb4/srq.c
69
e = s->entryp;
drivers/net/ethernet/chelsio/cxgb4/srq.c
70
e->valid = 1;
drivers/net/ethernet/chelsio/cxgb4/srq.c
71
e->idx = idx;
drivers/net/ethernet/chelsio/cxgb4/srq.c
72
e->pdid = SRQT_PDID_G(be64_to_cpu(rpl->rsvd_pdid));
drivers/net/ethernet/chelsio/cxgb4/srq.c
73
e->qlen = SRQT_QLEN_G(be32_to_cpu(rpl->qlen_qbase));
drivers/net/ethernet/chelsio/cxgb4/srq.c
74
e->qbase = SRQT_QBASE_G(be32_to_cpu(rpl->qlen_qbase));
drivers/net/ethernet/chelsio/cxgb4/srq.c
75
e->cur_msn = be16_to_cpu(rpl->cur_msn);
drivers/net/ethernet/chelsio/cxgb4/srq.c
76
e->max_msn = be16_to_cpu(rpl->max_msn);
drivers/net/ethernet/chelsio/inline_crypto/chtls/chtls_cm.c
2041
struct l2t_entry *e = csk->l2t_entry;
drivers/net/ethernet/chelsio/inline_crypto/chtls/chtls_cm.c
2043
if (e && sk->sk_state != TCP_SYN_RECV) {
drivers/net/ethernet/chelsio/inline_crypto/chtls/chtls_cm.c
2044
cxgb4_l2t_send(csk->egress_dev, reply_skb, e);
drivers/net/ethernet/freescale/enetc/enetc_qos.c
1136
struct action_gate_entry *e;
drivers/net/ethernet/freescale/enetc/enetc_qos.c
1259
e = sgi->entries;
drivers/net/ethernet/freescale/enetc/enetc_qos.c
1261
e[i].gate_state = entryg->gate.entries[i].gate_state;
drivers/net/ethernet/freescale/enetc/enetc_qos.c
1262
e[i].interval = entryg->gate.entries[i].interval;
drivers/net/ethernet/freescale/enetc/enetc_qos.c
1263
e[i].ipv = entryg->gate.entries[i].ipv;
drivers/net/ethernet/freescale/enetc/enetc_qos.c
1264
e[i].maxoctets = entryg->gate.entries[i].maxoctets;
drivers/net/ethernet/intel/i40e/i40e_adminq.c
1046
struct i40e_arq_event_info *e,
drivers/net/ethernet/intel/i40e/i40e_adminq.c
1059
memset(&e->desc, 0, sizeof(e->desc));
drivers/net/ethernet/intel/i40e/i40e_adminq.c
1094
e->desc = *desc;
drivers/net/ethernet/intel/i40e/i40e_adminq.c
1096
e->msg_len = min(datalen, e->buf_len);
drivers/net/ethernet/intel/i40e/i40e_adminq.c
1097
if (e->msg_buf != NULL && (e->msg_len != 0))
drivers/net/ethernet/intel/i40e/i40e_adminq.c
1098
memcpy(e->msg_buf, hw->aq.arq.r.arq_bi[desc_idx].va,
drivers/net/ethernet/intel/i40e/i40e_adminq.c
1099
e->msg_len);
drivers/net/ethernet/intel/i40e/i40e_adminq.c
1102
i40e_debug_aq(hw, I40E_DEBUG_AQ_COMMAND, (void *)desc, e->msg_buf,
drivers/net/ethernet/intel/i40e/i40e_adminq.c
1128
i40e_nvmupd_check_wait_event(hw, le16_to_cpu(e->desc.opcode), &e->desc);
drivers/net/ethernet/intel/i40e/i40e_main.c
10086
struct i40e_arq_event_info *e)
drivers/net/ethernet/intel/i40e/i40e_main.c
10088
struct i40e_aqc_get_link_status *status = libie_aq_raw(&e->desc);
drivers/net/ethernet/intel/i40e/i40e_main.c
9428
struct i40e_arq_event_info *e)
drivers/net/ethernet/intel/i40e/i40e_main.c
9430
struct i40e_aqc_lldp_get_mib *mib = libie_aq_raw(&e->desc);
drivers/net/ethernet/intel/i40e/i40e_main.c
9566
struct i40e_arq_event_info *e)
drivers/net/ethernet/intel/i40e/i40e_main.c
9568
struct i40e_aqc_lan_overflow *data = libie_aq_raw(&e->desc);
drivers/net/ethernet/intel/i40e/i40e_prototype.h
23
struct i40e_arq_event_info *e,
drivers/net/ethernet/intel/iavf/iavf_adminq.c
831
struct iavf_arq_event_info *e,
drivers/net/ethernet/intel/iavf/iavf_adminq.c
844
memset(&e->desc, 0, sizeof(e->desc));
drivers/net/ethernet/intel/iavf/iavf_adminq.c
879
e->desc = *desc;
drivers/net/ethernet/intel/iavf/iavf_adminq.c
881
e->msg_len = min(datalen, e->buf_len);
drivers/net/ethernet/intel/iavf/iavf_adminq.c
882
if (e->msg_buf && (e->msg_len != 0))
drivers/net/ethernet/intel/iavf/iavf_adminq.c
883
memcpy(e->msg_buf, hw->aq.arq.r.arq_bi[desc_idx].va,
drivers/net/ethernet/intel/iavf/iavf_adminq.c
884
e->msg_len);
drivers/net/ethernet/intel/iavf/iavf_adminq.c
887
iavf_debug_aq(hw, IAVF_DEBUG_AQ_COMMAND, (void *)desc, e->msg_buf,
drivers/net/ethernet/intel/iavf/iavf_prototype.h
22
struct iavf_arq_event_info *e,
drivers/net/ethernet/intel/ice/ice_adminq_cmd.h
207
} e;
drivers/net/ethernet/intel/ice/ice_arfs.c
108
struct ice_arfs_entry *e;
drivers/net/ethernet/intel/ice/ice_arfs.c
114
hlist_for_each_entry_safe(e, n, del_list_head, list_entry) {
drivers/net/ethernet/intel/ice/ice_arfs.c
117
result = ice_fdir_write_fltr(vsi->back, &e->fltr_info, false,
drivers/net/ethernet/intel/ice/ice_arfs.c
120
ice_arfs_update_active_fltr_cntrs(vsi, e, false);
drivers/net/ethernet/intel/ice/ice_arfs.c
123
result, e->fltr_state, e->fltr_info.fltr_id,
drivers/net/ethernet/intel/ice/ice_arfs.c
124
e->flow_id, e->fltr_info.q_index);
drivers/net/ethernet/intel/ice/ice_arfs.c
127
hlist_del(&e->list_entry);
drivers/net/ethernet/intel/ice/ice_arfs.c
128
devm_kfree(dev, e);
drivers/net/ethernet/intel/ice/ice_arfs.c
219
struct ice_arfs_entry *e;
drivers/net/ethernet/intel/ice/ice_arfs.c
226
hlist_for_each_entry_safe(e, n, &vsi->arfs_fltr_list[idx], list_entry)
drivers/net/ethernet/intel/ice/ice_arfs.c
228
if (e->fltr_state == ICE_ARFS_INACTIVE) {
drivers/net/ethernet/intel/ice/ice_arfs.c
229
enum ice_fltr_ptype flow_type = e->fltr_info.flow_type;
drivers/net/ethernet/intel/ice/ice_arfs.c
237
ep->arfs_entry = e;
drivers/net/ethernet/intel/ice/ice_arfs.c
239
e->fltr_state = ICE_ARFS_ACTIVE;
drivers/net/ethernet/intel/ice/ice_arfs.c
243
e->time_activated = get_jiffies_64();
drivers/net/ethernet/intel/ice/ice_arfs.c
244
} else if (e->fltr_state == ICE_ARFS_ACTIVE) {
drivers/net/ethernet/intel/ice/ice_arfs.c
246
if (ice_arfs_is_flow_expired(vsi, e)) {
drivers/net/ethernet/intel/ice/ice_arfs.c
251
hlist_del(&e->list_entry);
drivers/net/ethernet/intel/ice/ice_arfs.c
252
e->fltr_state = ICE_ARFS_TODEL;
drivers/net/ethernet/intel/ice/ice_arfs.c
254
hlist_add_head(&e->list_entry, del_list);
drivers/net/ethernet/intel/ice/ice_common.h
93
struct ice_rq_event_info *e, u16 *pending);
drivers/net/ethernet/intel/ice/ice_controlq.c
1174
struct ice_rq_event_info *e, u16 *pending)
drivers/net/ethernet/intel/ice/ice_controlq.c
1187
memset(&e->desc, 0, sizeof(e->desc));
drivers/net/ethernet/intel/ice/ice_controlq.c
1218
memcpy(&e->desc, desc, sizeof(e->desc));
drivers/net/ethernet/intel/ice/ice_controlq.c
1220
e->msg_len = min_t(u16, datalen, e->buf_len);
drivers/net/ethernet/intel/ice/ice_controlq.c
1221
if (e->msg_buf && e->msg_len)
drivers/net/ethernet/intel/ice/ice_controlq.c
1222
memcpy(e->msg_buf, cq->rq.r.rq_bi[desc_idx].va, e->msg_len);
drivers/net/ethernet/intel/ice/ice_controlq.c
1226
ice_debug_cq(hw, cq, (void *)desc, e->msg_buf, cq->rq_buf_size, true);
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
2056
struct ice_flow_entry *e, *t;
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
2058
list_for_each_entry_safe(e, t, &p->entries, l_entry)
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
2060
ICE_FLOW_ENTRY_HNDL(e));
drivers/net/ethernet/intel/ice/ice_flow.c
1545
struct ice_flow_entry *e, *t;
drivers/net/ethernet/intel/ice/ice_flow.c
1549
list_for_each_entry_safe(e, t, &prof->entries, l_entry) {
drivers/net/ethernet/intel/ice/ice_flow.c
1550
status = ice_flow_rem_entry_sync(hw, blk, e);
drivers/net/ethernet/intel/ice/ice_flow.c
1802
struct ice_flow_entry *e = NULL;
drivers/net/ethernet/intel/ice/ice_flow.c
1822
e = devm_kzalloc(ice_hw_to_dev(hw), sizeof(*e), GFP_KERNEL);
drivers/net/ethernet/intel/ice/ice_flow.c
1823
if (!e)
drivers/net/ethernet/intel/ice/ice_flow.c
1833
e->id = entry_id;
drivers/net/ethernet/intel/ice/ice_flow.c
1834
e->vsi_handle = vsi_handle;
drivers/net/ethernet/intel/ice/ice_flow.c
1835
e->prof = prof;
drivers/net/ethernet/intel/ice/ice_flow.c
1836
e->priority = prio;
drivers/net/ethernet/intel/ice/ice_flow.c
1848
list_add(&e->l_entry, &prof->entries);
drivers/net/ethernet/intel/ice/ice_flow.c
1851
*entry_h = ICE_FLOW_ENTRY_HNDL(e);
drivers/net/ethernet/intel/ice/ice_flow.c
1855
devm_kfree(ice_hw_to_dev(hw), e);
drivers/net/ethernet/intel/ice/ice_flow.c
2024
struct ice_flow_entry *e, *t;
drivers/net/ethernet/intel/ice/ice_flow.c
2027
list_for_each_entry_safe(e, t, &prof->entries, l_entry) {
drivers/net/ethernet/intel/ice/ice_flow.c
2028
if (e->vsi_handle != vsi_handle)
drivers/net/ethernet/intel/ice/ice_flow.c
2031
status = ice_flow_rem_entry_sync(hw, ICE_BLK_FD, e);
drivers/net/ethernet/intel/ice/ice_flow.h
465
#define ICE_FLOW_ENTRY_HNDL(e) ((u64)(uintptr_t)e)
drivers/net/ethernet/intel/ice/ice_fltr.c
17
struct ice_fltr_list_entry *e, *tmp;
drivers/net/ethernet/intel/ice/ice_fltr.c
19
list_for_each_entry_safe(e, tmp, h, list_entry) {
drivers/net/ethernet/intel/ice/ice_fltr.c
20
list_del(&e->list_entry);
drivers/net/ethernet/intel/ice/ice_fltr.c
21
devm_kfree(dev, e);
drivers/net/ethernet/intel/ice/ice_lag.c
1484
buf->elem[0].e.sw_resp = cpu_to_le16(local_lag->bond_swid);
drivers/net/ethernet/intel/ice/ice_lag.c
1497
buf->elem[0].e.sw_resp = cpu_to_le16(local_lag->bond_swid);
drivers/net/ethernet/intel/ice/ice_lag.c
1499
buf->elem[0].e.sw_resp =
drivers/net/ethernet/intel/ice/ice_switch.c
1918
sw_buf->elem[0].e.sw_resp = cpu_to_le16(*vsi_list_id);
drivers/net/ethernet/intel/ice/ice_switch.c
1926
*vsi_list_id = le16_to_cpu(vsi_ele->e.sw_resp);
drivers/net/ethernet/intel/ice/ice_switch.c
2196
*rid = le16_to_cpu(sw_buf->elem[0].e.sw_resp);
drivers/net/ethernet/intel/ice/ice_switch.c
4577
*counter_id = le16_to_cpu(buf->elem[0].e.sw_resp);
drivers/net/ethernet/intel/ice/ice_switch.c
4600
buf->elem[0].e.sw_resp = cpu_to_le16(counter_id);
drivers/net/ethernet/intel/ice/ice_switch.c
4634
buf->elem[0].e.sw_resp = cpu_to_le16(res_id);
drivers/net/ethernet/intel/ice/ice_switch.c
5279
sw_buf->elem[0].e.sw_resp = cpu_to_le16(rid);
drivers/net/ethernet/intel/igc/igc_main.c
6377
const struct tc_taprio_sched_entry *e, *prev;
drivers/net/ethernet/intel/igc/igc_main.c
6381
e = &qopt->entries[n];
drivers/net/ethernet/intel/igc/igc_main.c
6386
if (e->command != TC_TAPRIO_CMD_SET_GATES)
drivers/net/ethernet/intel/igc/igc_main.c
6390
if (e->gate_mask & BIT(i)) {
drivers/net/ethernet/intel/igc/igc_main.c
6523
struct tc_taprio_sched_entry *e = &qopt->entries[n];
drivers/net/ethernet/intel/igc/igc_main.c
6525
end_time += e->interval;
drivers/net/ethernet/intel/igc/igc_main.c
6545
if (!(e->gate_mask & BIT(i)))
drivers/net/ethernet/intel/igc/igc_main.c
6562
start_time += e->interval;
drivers/net/ethernet/intel/ixgbe/ixgbe_e610.c
286
int ixgbe_aci_get_event(struct ixgbe_hw *hw, struct ixgbe_aci_event *e,
drivers/net/ethernet/intel/ixgbe/ixgbe_e610.c
292
if (!e || (!e->msg_buf && e->buf_len))
drivers/net/ethernet/intel/ixgbe/ixgbe_e610.c
305
err = ixgbe_aci_send_cmd_execute(hw, &desc, e->msg_buf, e->buf_len);
drivers/net/ethernet/intel/ixgbe/ixgbe_e610.c
316
e->msg_len = min_t(u16, le16_to_cpu(desc.datalen), e->buf_len);
drivers/net/ethernet/intel/ixgbe/ixgbe_e610.c
318
memcpy(&e->desc, &desc, sizeof(e->desc));
drivers/net/ethernet/intel/ixgbe/ixgbe_e610.h
12
int ixgbe_aci_get_event(struct ixgbe_hw *hw, struct ixgbe_aci_event *e,
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
3282
struct ixgbe_aci_event *e)
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
3288
link_data = (struct ixgbe_aci_cmd_get_link_status_data *)e->msg_buf;
drivers/net/ethernet/marvell/octeontx2/af/rvu_cpt.c
32
u32 e, i; \
drivers/net/ethernet/marvell/octeontx2/af/rvu_cpt.c
34
for (e = (e_min), i = 0; e < (e_max); e++, i++) { \
drivers/net/ethernet/marvell/octeontx2/af/rvu_cpt.c
35
reg = rvu_read64(rvu, blkaddr, CPT_AF_EXEX_STS(e)); \
drivers/net/ethernet/marvell/octeontx2/af/rvu_debugfs.c
3766
u32 e_min = 0, e_max = 0, e, i;
drivers/net/ethernet/marvell/octeontx2/af/rvu_debugfs.c
3794
for (e = e_min, i = 0; e < e_max; e++, i++) {
drivers/net/ethernet/marvell/octeontx2/af/rvu_debugfs.c
3795
reg = rvu_read64(rvu, blkaddr, CPT_AF_EXEX_STS(e));
drivers/net/ethernet/marvell/octeontx2/af/rvu_debugfs.c
3835
u32 e_max, e;
drivers/net/ethernet/marvell/octeontx2/af/rvu_debugfs.c
3846
for (e = 0; e < e_max; e++) {
drivers/net/ethernet/marvell/octeontx2/af/rvu_debugfs.c
3847
reg = rvu_read64(rvu, blkaddr, CPT_AF_EXEX_CTL2(e));
drivers/net/ethernet/marvell/octeontx2/af/rvu_debugfs.c
3848
seq_printf(filp, "CPT Engine[%u] Group Enable 0x%02llx\n", e,
drivers/net/ethernet/marvell/octeontx2/af/rvu_debugfs.c
3850
reg = rvu_read64(rvu, blkaddr, CPT_AF_EXEX_ACTIVE(e));
drivers/net/ethernet/marvell/octeontx2/af/rvu_debugfs.c
3851
seq_printf(filp, "CPT Engine[%u] Active Info 0x%llx\n", e,
drivers/net/ethernet/marvell/octeontx2/af/rvu_debugfs.c
3853
reg = rvu_read64(rvu, blkaddr, CPT_AF_EXEX_CTL(e));
drivers/net/ethernet/marvell/octeontx2/af/rvu_debugfs.c
3854
seq_printf(filp, "CPT Engine[%u] Control 0x%llx\n", e,
drivers/net/ethernet/marvell/prestera/prestera_acl.c
579
struct prestera_acl_rule_entry *e)
drivers/net/ethernet/marvell/prestera/prestera_acl.c
581
return prestera_hw_vtcam_rule_del(sw, e->vtcam_id, e->hw_id);
drivers/net/ethernet/marvell/prestera/prestera_acl.c
585
struct prestera_acl_rule_entry *e)
drivers/net/ethernet/marvell/prestera/prestera_acl.c
594
if (e->accept.valid) {
drivers/net/ethernet/marvell/prestera/prestera_acl.c
599
if (e->drop.valid) {
drivers/net/ethernet/marvell/prestera/prestera_acl.c
604
if (e->trap.valid) {
drivers/net/ethernet/marvell/prestera/prestera_acl.c
609
if (e->police.valid) {
drivers/net/ethernet/marvell/prestera/prestera_acl.c
611
act_hw[act_num].police = e->police.i;
drivers/net/ethernet/marvell/prestera/prestera_acl.c
615
if (e->jump.valid) {
drivers/net/ethernet/marvell/prestera/prestera_acl.c
617
act_hw[act_num].jump = e->jump.i;
drivers/net/ethernet/marvell/prestera/prestera_acl.c
621
if (e->counter.block) {
drivers/net/ethernet/marvell/prestera/prestera_acl.c
623
act_hw[act_num].count.id = e->counter.id;
drivers/net/ethernet/marvell/prestera/prestera_acl.c
627
return prestera_hw_vtcam_rule_add(sw, e->vtcam_id, e->key.prio,
drivers/net/ethernet/marvell/prestera/prestera_acl.c
628
e->key.match.key, e->key.match.mask,
drivers/net/ethernet/marvell/prestera/prestera_acl.c
629
act_hw, act_num, &e->hw_id);
drivers/net/ethernet/marvell/prestera/prestera_acl.c
634
struct prestera_acl_rule_entry *e)
drivers/net/ethernet/marvell/prestera/prestera_acl.c
637
prestera_counter_put(sw->counter, e->counter.block, e->counter.id);
drivers/net/ethernet/marvell/prestera/prestera_acl.c
639
if (e->police.valid)
drivers/net/ethernet/marvell/prestera/prestera_acl.c
640
prestera_hw_policer_release(sw, e->police.i.id);
drivers/net/ethernet/marvell/prestera/prestera_acl.c
644
struct prestera_acl_rule_entry *e)
drivers/net/ethernet/marvell/prestera/prestera_acl.c
648
rhashtable_remove_fast(&acl->acl_rule_entry_ht, &e->ht_node,
drivers/net/ethernet/marvell/prestera/prestera_acl.c
651
ret = __prestera_acl_rule_entry2hw_del(acl->sw, e);
drivers/net/ethernet/marvell/prestera/prestera_acl.c
654
__prestera_acl_rule_entry_act_destruct(acl->sw, e);
drivers/net/ethernet/marvell/prestera/prestera_acl.c
655
kfree(e);
drivers/net/ethernet/marvell/prestera/prestera_acl.c
660
struct prestera_acl_rule_entry *e,
drivers/net/ethernet/marvell/prestera/prestera_acl.c
666
e->accept.valid = arg->accept.valid;
drivers/net/ethernet/marvell/prestera/prestera_acl.c
668
e->drop.valid = arg->drop.valid;
drivers/net/ethernet/marvell/prestera/prestera_acl.c
670
e->trap.valid = arg->trap.valid;
drivers/net/ethernet/marvell/prestera/prestera_acl.c
672
e->jump.valid = arg->jump.valid;
drivers/net/ethernet/marvell/prestera/prestera_acl.c
673
e->jump.i = arg->jump.i;
drivers/net/ethernet/marvell/prestera/prestera_acl.c
679
err = prestera_hw_policer_create(sw, type, &e->police.i.id);
drivers/net/ethernet/marvell/prestera/prestera_acl.c
683
err = prestera_hw_policer_sr_tcm_set(sw, e->police.i.id,
drivers/net/ethernet/marvell/prestera/prestera_acl.c
687
prestera_hw_policer_release(sw, e->police.i.id);
drivers/net/ethernet/marvell/prestera/prestera_acl.c
690
e->police.valid = arg->police.valid;
drivers/net/ethernet/marvell/prestera/prestera_acl.c
695
&e->counter.block,
drivers/net/ethernet/marvell/prestera/prestera_acl.c
696
&e->counter.id);
drivers/net/ethernet/marvell/prestera/prestera_acl.c
704
__prestera_acl_rule_entry_act_destruct(sw, e);
drivers/net/ethernet/marvell/prestera/prestera_acl.c
713
struct prestera_acl_rule_entry *e;
drivers/net/ethernet/marvell/prestera/prestera_acl.c
716
e = kzalloc_obj(*e);
drivers/net/ethernet/marvell/prestera/prestera_acl.c
717
if (!e)
drivers/net/ethernet/marvell/prestera/prestera_acl.c
720
memcpy(&e->key, key, sizeof(*key));
drivers/net/ethernet/marvell/prestera/prestera_acl.c
721
e->vtcam_id = arg->vtcam_id;
drivers/net/ethernet/marvell/prestera/prestera_acl.c
722
err = __prestera_acl_rule_entry_act_construct(acl->sw, e, arg);
drivers/net/ethernet/marvell/prestera/prestera_acl.c
726
err = __prestera_acl_rule_entry2hw_add(acl->sw, e);
drivers/net/ethernet/marvell/prestera/prestera_acl.c
730
err = rhashtable_insert_fast(&acl->acl_rule_entry_ht, &e->ht_node,
drivers/net/ethernet/marvell/prestera/prestera_acl.c
735
return e;
drivers/net/ethernet/marvell/prestera/prestera_acl.c
738
WARN_ON(__prestera_acl_rule_entry2hw_del(acl->sw, e));
drivers/net/ethernet/marvell/prestera/prestera_acl.c
740
__prestera_acl_rule_entry_act_destruct(acl->sw, e);
drivers/net/ethernet/marvell/prestera/prestera_acl.c
742
kfree(e);
drivers/net/ethernet/marvell/prestera/prestera_acl.h
175
struct prestera_acl_rule_entry *e);
drivers/net/ethernet/marvell/prestera/prestera_router_hw.c
232
struct prestera_rif_entry *e)
drivers/net/ethernet/marvell/prestera/prestera_router_hw.c
236
list_del(&e->router_node);
drivers/net/ethernet/marvell/prestera/prestera_router_hw.c
238
memcpy(&iface, &e->key.iface, sizeof(iface));
drivers/net/ethernet/marvell/prestera/prestera_router_hw.c
239
iface.vr_id = e->vr->hw_vr_id;
drivers/net/ethernet/marvell/prestera/prestera_router_hw.c
240
prestera_hw_rif_delete(sw, e->hw_id, &iface);
drivers/net/ethernet/marvell/prestera/prestera_router_hw.c
242
prestera_vr_put(sw, e->vr);
drivers/net/ethernet/marvell/prestera/prestera_router_hw.c
243
kfree(e);
drivers/net/ethernet/marvell/prestera/prestera_router_hw.c
252
struct prestera_rif_entry *e;
drivers/net/ethernet/marvell/prestera/prestera_router_hw.c
255
e = kzalloc_obj(*e);
drivers/net/ethernet/marvell/prestera/prestera_router_hw.c
256
if (!e)
drivers/net/ethernet/marvell/prestera/prestera_router_hw.c
259
if (__prestera_rif_entry_key_copy(k, &e->key))
drivers/net/ethernet/marvell/prestera/prestera_router_hw.c
262
e->vr = prestera_vr_get(sw, tb_id, NULL);
drivers/net/ethernet/marvell/prestera/prestera_router_hw.c
263
if (IS_ERR(e->vr))
drivers/net/ethernet/marvell/prestera/prestera_router_hw.c
266
memcpy(&e->addr, addr, sizeof(e->addr));
drivers/net/ethernet/marvell/prestera/prestera_router_hw.c
269
memcpy(&iface, &e->key.iface, sizeof(iface));
drivers/net/ethernet/marvell/prestera/prestera_router_hw.c
270
iface.vr_id = e->vr->hw_vr_id;
drivers/net/ethernet/marvell/prestera/prestera_router_hw.c
271
err = prestera_hw_rif_create(sw, &iface, e->addr, &e->hw_id);
drivers/net/ethernet/marvell/prestera/prestera_router_hw.c
275
list_add(&e->router_node, &sw->router->rif_entry_list);
drivers/net/ethernet/marvell/prestera/prestera_router_hw.c
277
return e;
drivers/net/ethernet/marvell/prestera/prestera_router_hw.c
280
prestera_vr_put(sw, e->vr);
drivers/net/ethernet/marvell/prestera/prestera_router_hw.c
283
kfree(e);
drivers/net/ethernet/marvell/prestera/prestera_router_hw.h
126
struct prestera_rif_entry *e);
drivers/net/ethernet/marvell/skge.c
1000
dma_unmap_len(e, maplen),
drivers/net/ethernet/marvell/skge.c
1002
dev_kfree_skb(e->skb);
drivers/net/ethernet/marvell/skge.c
1003
e->skb = NULL;
drivers/net/ethernet/marvell/skge.c
1005
} while ((e = e->next) != ring->start);
drivers/net/ethernet/marvell/skge.c
1016
struct skge_element *e;
drivers/net/ethernet/marvell/skge.c
1018
e = ring->start;
drivers/net/ethernet/marvell/skge.c
1028
if (skge_rx_setup(skge, e, skb, skge->rx_buf_size) < 0) {
drivers/net/ethernet/marvell/skge.c
1032
} while ((e = e->next) != ring->start);
drivers/net/ethernet/marvell/skge.c
2516
const struct skge_element *e)
drivers/net/ethernet/marvell/skge.c
2520
u64 base = skge->dma + (e->desc - skge->mem);
drivers/net/ethernet/marvell/skge.c
2743
struct skge_element *e;
drivers/net/ethernet/marvell/skge.c
2755
e = skge->tx_ring.to_use;
drivers/net/ethernet/marvell/skge.c
2756
td = e->desc;
drivers/net/ethernet/marvell/skge.c
2758
e->skb = skb;
drivers/net/ethernet/marvell/skge.c
2764
dma_unmap_addr_set(e, mapaddr, map);
drivers/net/ethernet/marvell/skge.c
2765
dma_unmap_len_set(e, maplen, len);
drivers/net/ethernet/marvell/skge.c
2802
e = e->next;
drivers/net/ethernet/marvell/skge.c
2803
e->skb = skb;
drivers/net/ethernet/marvell/skge.c
2804
tf = e->desc;
drivers/net/ethernet/marvell/skge.c
2809
dma_unmap_addr_set(e, mapaddr, map);
drivers/net/ethernet/marvell/skge.c
2810
dma_unmap_len_set(e, maplen, skb_frag_size(frag));
drivers/net/ethernet/marvell/skge.c
2827
e - skge->tx_ring.start, skb->len);
drivers/net/ethernet/marvell/skge.c
2829
skge->tx_ring.to_use = e->next;
drivers/net/ethernet/marvell/skge.c
2840
e = skge->tx_ring.to_use;
drivers/net/ethernet/marvell/skge.c
2841
dma_unmap_single(&hw->pdev->dev, dma_unmap_addr(e, mapaddr),
drivers/net/ethernet/marvell/skge.c
2842
dma_unmap_len(e, maplen), DMA_TO_DEVICE);
drivers/net/ethernet/marvell/skge.c
2844
e = e->next;
drivers/net/ethernet/marvell/skge.c
2845
dma_unmap_page(&hw->pdev->dev, dma_unmap_addr(e, mapaddr),
drivers/net/ethernet/marvell/skge.c
2846
dma_unmap_len(e, maplen), DMA_TO_DEVICE);
drivers/net/ethernet/marvell/skge.c
2858
static inline void skge_tx_unmap(struct pci_dev *pdev, struct skge_element *e,
drivers/net/ethernet/marvell/skge.c
2863
dma_unmap_single(&pdev->dev, dma_unmap_addr(e, mapaddr),
drivers/net/ethernet/marvell/skge.c
2864
dma_unmap_len(e, maplen), DMA_TO_DEVICE);
drivers/net/ethernet/marvell/skge.c
2866
dma_unmap_page(&pdev->dev, dma_unmap_addr(e, mapaddr),
drivers/net/ethernet/marvell/skge.c
2867
dma_unmap_len(e, maplen), DMA_TO_DEVICE);
drivers/net/ethernet/marvell/skge.c
2874
struct skge_element *e;
drivers/net/ethernet/marvell/skge.c
2876
for (e = skge->tx_ring.to_clean; e != skge->tx_ring.to_use; e = e->next) {
drivers/net/ethernet/marvell/skge.c
2877
struct skge_tx_desc *td = e->desc;
drivers/net/ethernet/marvell/skge.c
2879
skge_tx_unmap(skge->hw->pdev, e, td->control);
drivers/net/ethernet/marvell/skge.c
2882
dev_kfree_skb(e->skb);
drivers/net/ethernet/marvell/skge.c
2887
skge->tx_ring.to_clean = e;
drivers/net/ethernet/marvell/skge.c
3050
struct skge_element *e,
drivers/net/ethernet/marvell/skge.c
3059
e - skge->rx_ring.start, status, len);
drivers/net/ethernet/marvell/skge.c
3079
dma_unmap_addr(e, mapaddr),
drivers/net/ethernet/marvell/skge.c
3080
dma_unmap_len(e, maplen),
drivers/net/ethernet/marvell/skge.c
3082
skb_copy_from_linear_data(e->skb, skb->data, len);
drivers/net/ethernet/marvell/skge.c
3084
dma_unmap_addr(e, mapaddr),
drivers/net/ethernet/marvell/skge.c
3085
dma_unmap_len(e, maplen),
drivers/net/ethernet/marvell/skge.c
3087
skge_rx_reuse(e, skge->rx_buf_size);
drivers/net/ethernet/marvell/skge.c
3096
ee = *e;
drivers/net/ethernet/marvell/skge.c
3101
if (skge_rx_setup(skge, e, nskb, skge->rx_buf_size) < 0) {
drivers/net/ethernet/marvell/skge.c
3125
e - skge->rx_ring.start, control, status);
drivers/net/ethernet/marvell/skge.c
3144
skge_rx_reuse(e, skge->rx_buf_size);
drivers/net/ethernet/marvell/skge.c
3153
struct skge_element *e;
drivers/net/ethernet/marvell/skge.c
3158
for (e = ring->to_clean; e != ring->to_use; e = e->next) {
drivers/net/ethernet/marvell/skge.c
3159
u32 control = ((const struct skge_tx_desc *) e->desc)->control;
drivers/net/ethernet/marvell/skge.c
3164
skge_tx_unmap(skge->hw->pdev, e, control);
drivers/net/ethernet/marvell/skge.c
3169
e - skge->tx_ring.start);
drivers/net/ethernet/marvell/skge.c
3172
bytes_compl += e->skb->len;
drivers/net/ethernet/marvell/skge.c
3174
dev_consume_skb_any(e->skb);
drivers/net/ethernet/marvell/skge.c
3178
skge->tx_ring.to_clean = e;
drivers/net/ethernet/marvell/skge.c
3201
struct skge_element *e;
drivers/net/ethernet/marvell/skge.c
3208
for (e = ring->to_clean; prefetch(e->next), work_done < budget; e = e->next) {
drivers/net/ethernet/marvell/skge.c
3209
struct skge_rx_desc *rd = e->desc;
drivers/net/ethernet/marvell/skge.c
3218
skb = skge_rx_get(dev, e, control, rd->status, rd->csum2);
drivers/net/ethernet/marvell/skge.c
3224
ring->to_clean = e;
drivers/net/ethernet/marvell/skge.c
3696
const struct skge_element *e;
drivers/net/ethernet/marvell/skge.c
3705
for (e = skge->tx_ring.to_clean; e != skge->tx_ring.to_use; e = e->next) {
drivers/net/ethernet/marvell/skge.c
3706
const struct skge_tx_desc *t = e->desc;
drivers/net/ethernet/marvell/skge.c
3713
for (e = skge->rx_ring.to_clean; ; e = e->next) {
drivers/net/ethernet/marvell/skge.c
3714
const struct skge_rx_desc *r = e->desc;
drivers/net/ethernet/marvell/skge.c
918
struct skge_element *e;
drivers/net/ethernet/marvell/skge.c
921
ring->start = kzalloc_objs(*e, ring->count);
drivers/net/ethernet/marvell/skge.c
925
for (i = 0, e = ring->start, d = vaddr; i < ring->count; i++, e++, d++) {
drivers/net/ethernet/marvell/skge.c
926
e->desc = d;
drivers/net/ethernet/marvell/skge.c
928
e->next = ring->start;
drivers/net/ethernet/marvell/skge.c
931
e->next = e + 1;
drivers/net/ethernet/marvell/skge.c
941
static int skge_rx_setup(struct skge_port *skge, struct skge_element *e,
drivers/net/ethernet/marvell/skge.c
944
struct skge_rx_desc *rd = e->desc;
drivers/net/ethernet/marvell/skge.c
955
e->skb = skb;
drivers/net/ethernet/marvell/skge.c
964
dma_unmap_addr_set(e, mapaddr, map);
drivers/net/ethernet/marvell/skge.c
965
dma_unmap_len_set(e, maplen, bufsize);
drivers/net/ethernet/marvell/skge.c
973
static inline void skge_rx_reuse(struct skge_element *e, unsigned int size)
drivers/net/ethernet/marvell/skge.c
975
struct skge_rx_desc *rd = e->desc;
drivers/net/ethernet/marvell/skge.c
991
struct skge_element *e;
drivers/net/ethernet/marvell/skge.c
993
e = ring->start;
drivers/net/ethernet/marvell/skge.c
995
struct skge_rx_desc *rd = e->desc;
drivers/net/ethernet/marvell/skge.c
997
if (e->skb) {
drivers/net/ethernet/marvell/skge.c
999
dma_unmap_addr(e, mapaddr),
drivers/net/ethernet/mediatek/mtk_ppe.c
144
static u32 mtk_ppe_hash_entry(struct mtk_eth *eth, struct mtk_foe_entry *e)
drivers/net/ethernet/mediatek/mtk_ppe.c
149
switch (mtk_get_ib1_pkt_type(eth, e->ib1)) {
drivers/net/ethernet/mediatek/mtk_ppe.c
152
hv1 = e->ipv4.orig.ports;
drivers/net/ethernet/mediatek/mtk_ppe.c
153
hv2 = e->ipv4.orig.dest_ip;
drivers/net/ethernet/mediatek/mtk_ppe.c
154
hv3 = e->ipv4.orig.src_ip;
drivers/net/ethernet/mediatek/mtk_ppe.c
158
hv1 = e->ipv6.src_ip[3] ^ e->ipv6.dest_ip[3];
drivers/net/ethernet/mediatek/mtk_ppe.c
159
hv1 ^= e->ipv6.ports;
drivers/net/ethernet/mediatek/mtk_ppe.c
161
hv2 = e->ipv6.src_ip[2] ^ e->ipv6.dest_ip[2];
drivers/net/ethernet/mediatek/mtk_ppe.c
162
hv2 ^= e->ipv6.dest_ip[0];
drivers/net/ethernet/mediatek/mtk_ppe.c
164
hv3 = e->ipv6.src_ip[1] ^ e->ipv6.dest_ip[1];
drivers/net/ethernet/mediatek/mtk_ppe.c
165
hv3 ^= e->ipv6.src_ip[0];
drivers/net/ethernet/mellanox/mlx5/core/en/rep/neigh.c
132
struct mlx5e_encap_entry *e = NULL;
drivers/net/ethernet/mellanox/mlx5/core/en/rep/neigh.c
161
while ((e = mlx5e_get_next_init_encap(nhe, e)) != NULL)
drivers/net/ethernet/mellanox/mlx5/core/en/rep/neigh.c
162
mlx5e_rep_update_flows(netdev_priv(e->out_dev), e, neigh_connected, ha);
drivers/net/ethernet/mellanox/mlx5/core/en/rep/tc.c
102
encap_connected = !!(e->flags & MLX5_ENCAP_ENTRY_VALID);
drivers/net/ethernet/mellanox/mlx5/core/en/rep/tc.c
103
if (encap_connected == neigh_connected && ether_addr_equal(e->h_dest, ha))
drivers/net/ethernet/mellanox/mlx5/core/en/rep/tc.c
106
mlx5e_take_all_encap_flows(e, &flow_list);
drivers/net/ethernet/mellanox/mlx5/core/en/rep/tc.c
108
if ((e->flags & MLX5_ENCAP_ENTRY_VALID) &&
drivers/net/ethernet/mellanox/mlx5/core/en/rep/tc.c
109
(!neigh_connected || !ether_addr_equal(e->h_dest, ha)))
drivers/net/ethernet/mellanox/mlx5/core/en/rep/tc.c
110
mlx5e_tc_encap_flows_del(priv, e, &flow_list);
drivers/net/ethernet/mellanox/mlx5/core/en/rep/tc.c
112
if (neigh_connected && !(e->flags & MLX5_ENCAP_ENTRY_VALID)) {
drivers/net/ethernet/mellanox/mlx5/core/en/rep/tc.c
115
ether_addr_copy(e->h_dest, ha);
drivers/net/ethernet/mellanox/mlx5/core/en/rep/tc.c
120
route_dev = __dev_get_by_index(dev_net(priv->netdev), e->route_dev_ifindex);
drivers/net/ethernet/mellanox/mlx5/core/en/rep/tc.c
124
mlx5e_tc_encap_flows_add(priv, e, &flow_list);
drivers/net/ethernet/mellanox/mlx5/core/en/rep/tc.c
34
struct mlx5e_encap_entry *e,
drivers/net/ethernet/mellanox/mlx5/core/en/rep/tc.c
44
err = mlx5_tun_entropy_refcount_inc(tun_entropy, e->reformat_type);
drivers/net/ethernet/mellanox/mlx5/core/en/rep/tc.c
55
e->reformat_type);
drivers/net/ethernet/mellanox/mlx5/core/en/rep/tc.c
60
e->nhe = nhe;
drivers/net/ethernet/mellanox/mlx5/core/en/rep/tc.c
62
list_add_rcu(&e->encap_list, &nhe->encap_list);
drivers/net/ethernet/mellanox/mlx5/core/en/rep/tc.c
71
struct mlx5e_encap_entry *e)
drivers/net/ethernet/mellanox/mlx5/core/en/rep/tc.c
77
if (!e->nhe)
drivers/net/ethernet/mellanox/mlx5/core/en/rep/tc.c
80
spin_lock(&e->nhe->encap_list_lock);
drivers/net/ethernet/mellanox/mlx5/core/en/rep/tc.c
81
list_del_rcu(&e->encap_list);
drivers/net/ethernet/mellanox/mlx5/core/en/rep/tc.c
82
spin_unlock(&e->nhe->encap_list_lock);
drivers/net/ethernet/mellanox/mlx5/core/en/rep/tc.c
84
mlx5e_rep_neigh_entry_release(e->nhe);
drivers/net/ethernet/mellanox/mlx5/core/en/rep/tc.c
85
e->nhe = NULL;
drivers/net/ethernet/mellanox/mlx5/core/en/rep/tc.c
86
mlx5_tun_entropy_refcount_dec(tun_entropy, e->reformat_type);
drivers/net/ethernet/mellanox/mlx5/core/en/rep/tc.c
90
struct mlx5e_encap_entry *e,
drivers/net/ethernet/mellanox/mlx5/core/en/rep/tc.c
94
struct ethhdr *eth = (struct ethhdr *)e->encap_header;
drivers/net/ethernet/mellanox/mlx5/core/en/rep/tc.h
25
struct mlx5e_encap_entry *e,
drivers/net/ethernet/mellanox/mlx5/core/en/rep/tc.h
30
struct mlx5e_encap_entry *e,
drivers/net/ethernet/mellanox/mlx5/core/en/rep/tc.h
34
struct mlx5e_encap_entry *e);
drivers/net/ethernet/mellanox/mlx5/core/en/tc/meter.c
19
#define MLX5_CALC_CIR(m, e) ((MLX5_CONST_CIR * (m)) >> (e))
drivers/net/ethernet/mellanox/mlx5/core/en/tc/meter.c
23
#define MLX5_CALC_CBS(m, e) ((m) << (e))
drivers/net/ethernet/mellanox/mlx5/core/en/tc/meter.c
56
u8 e, _man = 0, _exp = 0;
drivers/net/ethernet/mellanox/mlx5/core/en/tc/meter.c
59
for (e = 0; e <= 0x1F; e++) { /* exp width 5bit */
drivers/net/ethernet/mellanox/mlx5/core/en/tc/meter.c
60
m = cir << e;
drivers/net/ethernet/mellanox/mlx5/core/en/tc/meter.c
66
_cir = MLX5_CALC_CIR(m, e);
drivers/net/ethernet/mellanox/mlx5/core/en/tc/meter.c
70
_exp = e;
drivers/net/ethernet/mellanox/mlx5/core/en/tc/meter.c
86
u8 e, _man = 0, _exp = 0;
drivers/net/ethernet/mellanox/mlx5/core/en/tc/meter.c
89
for (e = 0; e <= 0x1F; e++) { /* exp width 5bit */
drivers/net/ethernet/mellanox/mlx5/core/en/tc/meter.c
90
m = cbs >> e;
drivers/net/ethernet/mellanox/mlx5/core/en/tc/meter.c
93
_cbs = MLX5_CALC_CBS(m, e);
drivers/net/ethernet/mellanox/mlx5/core/en/tc/meter.c
97
_exp = e;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_priv.h
66
struct mlx5e_encap_entry *e; /* attached encap instance */
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
187
struct mlx5e_encap_entry *e)
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
189
if (!e->tunnel) {
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
194
return e->tunnel->generate_ip_tun_hdr(buf, ip_proto, e);
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
198
struct mlx5e_encap_entry *e,
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
204
ether_addr_copy(eth->h_dest, e->h_dest);
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
223
struct mlx5e_encap_entry *e)
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
226
const struct ip_tunnel_key *tun_key = &e->tun_info->key;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
249
e->tunnel->calc_hlen(e);
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
266
e->out_dev = attr.out_dev;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
267
e->route_dev_ifindex = attr.route_dev->ifindex;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
274
err = mlx5e_rep_encap_entry_attach(netdev_priv(attr.out_dev), e, &m_neigh, attr.n->dev);
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
280
ether_addr_copy(e->h_dest, attr.n->ha);
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
284
ip = (struct iphdr *)gen_eth_tnl_hdr(encap_header, attr.route_dev, e,
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
297
&ip->protocol, e);
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
301
e->encap_size = ipv4_encap_size;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
302
e->encap_header = encap_header;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
314
reformat_params.type = e->reformat_type;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
315
reformat_params.size = e->encap_size;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
316
reformat_params.data = e->encap_header;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
317
e->pkt_reformat = mlx5_packet_reformat_alloc(priv->mdev, &reformat_params,
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
319
if (IS_ERR(e->pkt_reformat)) {
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
320
err = PTR_ERR(e->pkt_reformat);
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
324
e->flags |= MLX5_ENCAP_ENTRY_VALID;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
330
mlx5e_rep_encap_entry_detach(netdev_priv(e->out_dev), e);
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
340
struct mlx5e_encap_entry *e)
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
343
const struct ip_tunnel_key *tun_key = &e->tun_info->key;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
365
e->tunnel->calc_hlen(e);
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
380
e->route_dev_ifindex = attr.route_dev->ifindex;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
384
ether_addr_copy(e->h_dest, attr.n->ha);
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
385
WRITE_ONCE(e->nhe->neigh_dev, attr.n->dev);
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
389
ip = (struct iphdr *)gen_eth_tnl_hdr(encap_header, attr.route_dev, e,
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
402
&ip->protocol, e);
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
406
e->encap_size = ipv4_encap_size;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
407
kfree(e->encap_header);
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
408
e->encap_header = encap_header;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
420
reformat_params.type = e->reformat_type;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
421
reformat_params.size = e->encap_size;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
422
reformat_params.data = e->encap_header;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
423
e->pkt_reformat = mlx5_packet_reformat_alloc(priv->mdev, &reformat_params,
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
425
if (IS_ERR(e->pkt_reformat)) {
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
426
err = PTR_ERR(e->pkt_reformat);
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
430
e->flags |= MLX5_ENCAP_ENTRY_VALID;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
495
struct mlx5e_encap_entry *e)
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
498
const struct ip_tunnel_key *tun_key = &e->tun_info->key;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
520
e->tunnel->calc_hlen(e);
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
537
e->out_dev = attr.out_dev;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
538
e->route_dev_ifindex = attr.route_dev->ifindex;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
545
err = mlx5e_rep_encap_entry_attach(netdev_priv(attr.out_dev), e, &m_neigh, attr.n->dev);
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
551
ether_addr_copy(e->h_dest, attr.n->ha);
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
555
ip6h = (struct ipv6hdr *)gen_eth_tnl_hdr(encap_header, attr.route_dev, e,
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
567
&ip6h->nexthdr, e);
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
571
e->encap_size = ipv6_encap_size;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
572
e->encap_header = encap_header;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
584
reformat_params.type = e->reformat_type;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
585
reformat_params.size = e->encap_size;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
586
reformat_params.data = e->encap_header;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
587
e->pkt_reformat = mlx5_packet_reformat_alloc(priv->mdev, &reformat_params,
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
589
if (IS_ERR(e->pkt_reformat)) {
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
590
err = PTR_ERR(e->pkt_reformat);
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
594
e->flags |= MLX5_ENCAP_ENTRY_VALID;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
600
mlx5e_rep_encap_entry_detach(netdev_priv(e->out_dev), e);
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
610
struct mlx5e_encap_entry *e)
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
613
const struct ip_tunnel_key *tun_key = &e->tun_info->key;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
635
e->tunnel->calc_hlen(e);
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
650
e->route_dev_ifindex = attr.route_dev->ifindex;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
654
ether_addr_copy(e->h_dest, attr.n->ha);
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
655
WRITE_ONCE(e->nhe->neigh_dev, attr.n->dev);
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
659
ip6h = (struct ipv6hdr *)gen_eth_tnl_hdr(encap_header, attr.route_dev, e,
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
671
&ip6h->nexthdr, e);
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
675
e->encap_size = ipv6_encap_size;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
676
kfree(e->encap_header);
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
677
e->encap_header = encap_header;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
689
reformat_params.type = e->reformat_type;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
690
reformat_params.size = e->encap_size;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
691
reformat_params.data = e->encap_header;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
692
e->pkt_reformat = mlx5_packet_reformat_alloc(priv->mdev, &reformat_params,
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
694
if (IS_ERR(e->pkt_reformat)) {
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
695
err = PTR_ERR(e->pkt_reformat);
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
699
e->flags |= MLX5_ENCAP_ENTRY_VALID;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
785
struct mlx5e_encap_entry *e,
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
791
e->reformat_type = -1;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
795
return tunnel->init_encap_attr(tunnel_dev, priv, e, extack);
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.h
34
int (*calc_hlen)(struct mlx5e_encap_entry *e);
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.h
37
struct mlx5e_encap_entry *e,
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.h
41
struct mlx5e_encap_entry *e);
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.h
66
struct mlx5e_encap_entry *e,
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.h
71
struct mlx5e_encap_entry *e);
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.h
74
struct mlx5e_encap_entry *e);
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.h
79
struct mlx5e_encap_entry *e);
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.h
82
struct mlx5e_encap_entry *e);
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.h
87
struct mlx5e_encap_entry *e)
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.h
92
struct mlx5e_encap_entry *e)
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
1350
struct mlx5e_encap_entry *e,
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
1378
err = mlx5e_set_vf_tunnel(esw, attr, &parse_attr->mod_hdr_acts, e->out_dev,
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
1379
e->route_dev_ifindex, out_index);
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
1401
list_add(&e->route_list, &r->encap_entries);
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
1435
struct mlx5e_encap_entry *e, *tmp;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
1447
list_for_each_entry_safe(e, tmp, &r->encap_entries, route_list)
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
1448
list_del_init(&e->route_list);
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
1456
struct mlx5e_encap_entry *e,
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
1487
e->flags |= MLX5_ENCAP_ENTRY_NO_ROUTE;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
1488
if (e->flags & MLX5_ENCAP_ENTRY_VALID) {
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
1489
e->flags &= ~MLX5_ENCAP_ENTRY_VALID;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
1490
mlx5_packet_reformat_dealloc(priv->mdev, e->pkt_reformat);
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
1491
e->pkt_reformat = NULL;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
1497
struct mlx5e_encap_entry *e,
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
1504
err = ip_tunnel_info_af(e->tun_info) == AF_INET ?
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
1505
mlx5e_tc_tun_update_header_ipv4(priv, tunnel_dev, e) :
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
1506
mlx5e_tc_tun_update_header_ipv6(priv, tunnel_dev, e);
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
1509
e->flags &= ~MLX5_ENCAP_ENTRY_NO_ROUTE;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
152
struct mlx5e_encap_entry *e,
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
1528
e->out_dev, e->route_dev_ifindex,
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
1542
if (e->flags & MLX5_ENCAP_ENTRY_VALID) {
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
1543
esw_attr->dests[flow->tmp_entry_index].pkt_reformat = e->pkt_reformat;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
1590
struct mlx5e_encap_entry *e;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
1596
list_for_each_entry(e, &r->encap_entries, route_list) {
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
1599
mlx5e_take_all_encap_flows(e, &encap_flows);
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
1604
mlx5e_invalidate_encap(priv, e, &encap_flows);
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
1611
mlx5e_reoffload_encap(priv, tunnel_dev, e, &encap_flows);
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
164
if (e->flags & MLX5_ENCAP_ENTRY_NO_ROUTE)
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
168
reformat_params.type = e->reformat_type;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
169
reformat_params.size = e->encap_size;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
170
reformat_params.data = e->encap_header;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
171
e->pkt_reformat = mlx5_packet_reformat_alloc(priv->mdev,
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
174
if (IS_ERR(e->pkt_reformat)) {
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
176
e->pkt_reformat);
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
179
e->flags |= MLX5_ENCAP_ENTRY_VALID;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
190
esw_attr->dests[flow->tmp_entry_index].pkt_reformat = e->pkt_reformat;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
20
struct mlx5e_encap_entry *e,
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
222
struct mlx5e_encap_entry *e,
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
26
route_dev = dev_get_by_index(dev_net(e->out_dev), e->route_dev_ifindex);
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
269
e->flags &= ~MLX5_ENCAP_ENTRY_VALID;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
270
mlx5_packet_reformat_dealloc(priv->mdev, e->pkt_reformat);
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
271
e->pkt_reformat = NULL;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
295
void mlx5e_take_all_encap_flows(struct mlx5e_encap_entry *e, struct list_head *flow_list)
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
300
list_for_each_entry(efi, &e->flows, list) {
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
322
struct mlx5e_encap_entry *e,
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
331
for (next = e ?
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
333
&e->encap_list,
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
350
if (e)
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
351
mlx5e_encap_put(netdev_priv(e->out_dev), e);
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
359
e = next;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
366
static bool mlx5e_encap_valid(struct mlx5e_encap_entry *e)
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
368
return e->flags & MLX5_ENCAP_ENTRY_VALID;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
373
struct mlx5e_encap_entry *e)
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
375
return mlx5e_get_next_matching_encap(nhe, e, mlx5e_encap_valid);
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
378
static bool mlx5e_encap_initialized(struct mlx5e_encap_entry *e)
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
380
return e->compl_result >= 0;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
385
struct mlx5e_encap_entry *e)
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
387
return mlx5e_get_next_matching_encap(nhe, e, mlx5e_encap_initialized);
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
393
struct mlx5e_encap_entry *e = NULL;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
40
err = mlx5e_set_fwd_to_int_port_actions(priv, attr, e->route_dev_ifindex,
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
413
while ((e = mlx5e_get_next_valid_encap(nhe, e)) != NULL) {
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
414
struct mlx5e_priv *priv = netdev_priv(e->out_dev);
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
421
list_for_each_entry_safe(efi, tmp, &e->flows, list) {
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
442
mlx5e_encap_put(priv, e);
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
464
static void mlx5e_encap_dealloc(struct mlx5e_priv *priv, struct mlx5e_encap_entry *e)
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
466
WARN_ON(!list_empty(&e->flows));
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
468
if (e->compl_result > 0) {
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
469
mlx5e_rep_encap_entry_detach(netdev_priv(e->out_dev), e);
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
471
if (e->flags & MLX5_ENCAP_ENTRY_VALID)
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
472
mlx5_packet_reformat_dealloc(priv->mdev, e->pkt_reformat);
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
475
kfree(e->tun_info);
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
476
kfree(e->encap_header);
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
477
kfree_rcu(e, rcu);
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
491
void mlx5e_encap_put(struct mlx5e_priv *priv, struct mlx5e_encap_entry *e)
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
495
if (!refcount_dec_and_mutex_lock(&e->refcnt, &esw->offloads.encap_tbl_lock))
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
497
list_del(&e->route_list);
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
498
hash_del_rcu(&e->encap_hlist);
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
501
mlx5e_encap_dealloc(priv, e);
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
504
static void mlx5e_encap_put_locked(struct mlx5e_priv *priv, struct mlx5e_encap_entry *e)
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
510
if (!refcount_dec_and_test(&e->refcnt))
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
512
list_del(&e->route_list);
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
513
hash_del_rcu(&e->encap_hlist);
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
514
mlx5e_encap_dealloc(priv, e);
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
538
struct mlx5e_encap_entry *e = flow->encaps[out_index].e;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
549
if (!e)
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
554
flow->encaps[out_index].e = NULL;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
555
if (!refcount_dec_and_test(&e->refcnt)) {
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
559
list_del(&e->route_list);
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
560
hash_del_rcu(&e->encap_hlist);
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
563
mlx5e_encap_dealloc(priv, e);
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
644
bool mlx5e_encap_take(struct mlx5e_encap_entry *e)
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
646
return refcount_inc_not_zero(&e->refcnt);
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
649
static bool mlx5e_decap_take(struct mlx5e_decap_entry *e)
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
651
return refcount_inc_not_zero(&e->refcnt);
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
660
struct mlx5e_encap_entry *e;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
662
hash_for_each_possible_rcu(esw->offloads.encap_tbl, e,
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
664
e_key.ip_tun_key = &e->tun_info->key;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
665
e_key.tc_tunnel = e->tunnel;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
666
if (e->tunnel->encap_info_equal(&e_key, key) &&
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
667
mlx5e_encap_take(e))
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
668
return e;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
680
struct mlx5e_decap_entry *e;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
682
hash_for_each_possible_rcu(esw->offloads.decap_tbl, e,
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
684
r_key = e->key;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
686
mlx5e_decap_take(e))
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
687
return e;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
702
struct mlx5e_encap_entry *e,
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
708
if (flow->encaps[i].e != e)
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
814
struct mlx5e_encap_entry *e,
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
832
struct mlx5e_encap_entry *e;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
854
e = mlx5e_encap_get(priv, &key, hash_key);
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
857
if (e) {
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
859
if (is_duplicated_encap_entry(priv, flow, out_index, e, extack)) {
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
867
e = kzalloc_obj(*e);
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
868
if (!e) {
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
873
refcount_set(&e->refcnt, 1);
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
874
init_completion(&e->res_ready);
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
876
INIT_LIST_HEAD(&e->route_list);
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
883
e->tun_info = tun_info;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
884
memcpy(&e->mpls_info, mpls_info, sizeof(*mpls_info));
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
885
err = mlx5e_tc_tun_init_encap_attr(mirred_dev, priv, e, extack);
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
889
INIT_LIST_HEAD(&e->flows);
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
890
hash_add_rcu(esw->offloads.encap_tbl, &e->encap_hlist, hash_key);
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
894
err = mlx5e_tc_tun_create_header_ipv4(priv, mirred_dev, e);
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
896
err = mlx5e_tc_tun_create_header_ipv6(priv, mirred_dev, e);
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
898
complete_all(&e->res_ready);
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
900
e->compl_result = err;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
903
e->compl_result = 1;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
906
err = mlx5e_attach_encap_route(priv, flow, attr, e, entry_created,
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
911
err = mlx5e_set_int_port_tunnel(priv, attr, e, out_index);
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
922
flow->encaps[out_index].e = e;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
923
list_add(&flow->encaps[out_index].list, &e->flows);
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
925
*encap_dev = e->out_dev;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
926
if (e->flags & MLX5_ENCAP_ENTRY_VALID) {
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
927
attr->esw_attr->dests[out_index].pkt_reformat = e->pkt_reformat;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
936
if (e)
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
937
mlx5e_encap_put_locked(priv, e);
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_encap.c
942
kfree(e);
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_geneve.c
15
static int mlx5e_tc_tun_calc_hlen_geneve(struct mlx5e_encap_entry *e)
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_geneve.c
19
e->tun_info->options_len;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_geneve.c
66
struct mlx5e_encap_entry *e,
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_geneve.c
69
e->tunnel = &geneve_tunnel;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_geneve.c
75
e->reformat_type = MLX5_REFORMAT_TYPE_L2_TO_VXLAN;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_geneve.c
94
struct mlx5e_encap_entry *e)
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_geneve.c
96
const struct ip_tunnel_info *tun_info = e->tun_info;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_gre.c
12
static int mlx5e_tc_tun_calc_hlen_gretap(struct mlx5e_encap_entry *e)
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_gre.c
14
return gre_calc_hlen(e->tun_info->key.tun_flags);
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_gre.c
19
struct mlx5e_encap_entry *e,
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_gre.c
22
e->tunnel = &gre_tunnel;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_gre.c
23
e->reformat_type = MLX5_REFORMAT_TYPE_L2_TO_NVGRE;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_gre.c
29
struct mlx5e_encap_entry *e)
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_gre.c
31
const struct ip_tunnel_key *tun_key = &e->tun_info->key;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_gre.c
49
hdr_len = mlx5e_tc_tun_calc_hlen_gretap(e);
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_mplsoudp.c
13
static int calc_hlen(struct mlx5e_encap_entry *e)
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_mplsoudp.c
20
struct mlx5e_encap_entry *e,
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_mplsoudp.c
23
e->tunnel = &mplsoudp_tunnel;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_mplsoudp.c
24
e->reformat_type = MLX5_REFORMAT_TYPE_L2_TO_L3_TUNNEL;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_vxlan.c
103
md = ip_tunnel_info_opts(e->tun_info);
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_vxlan.c
14
static int mlx5e_tc_tun_calc_hlen_vxlan(struct mlx5e_encap_entry *e)
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_vxlan.c
63
struct mlx5e_encap_entry *e,
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_vxlan.c
66
int dst_port = be16_to_cpu(e->tun_info->key.tp_dst);
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_vxlan.c
68
e->tunnel = &vxlan_tunnel;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_vxlan.c
79
e->reformat_type = MLX5_REFORMAT_TYPE_L2_TO_VXLAN;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_vxlan.c
85
struct mlx5e_encap_entry *e)
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_vxlan.c
87
const struct ip_tunnel_key *tun_key = &e->tun_info->key;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun_vxlan.c
94
e->tun_info->options_len != sizeof(*md))
drivers/net/ethernet/mellanox/mlx5/core/en_tc.h
210
struct mlx5e_encap_entry *e,
drivers/net/ethernet/mellanox/mlx5/core/en_tc.h
213
struct mlx5e_encap_entry *e,
drivers/net/ethernet/mellanox/mlx5/core/en_tc.h
215
bool mlx5e_encap_take(struct mlx5e_encap_entry *e);
drivers/net/ethernet/mellanox/mlx5/core/en_tc.h
216
void mlx5e_encap_put(struct mlx5e_priv *priv, struct mlx5e_encap_entry *e);
drivers/net/ethernet/mellanox/mlx5/core/en_tc.h
218
void mlx5e_take_all_encap_flows(struct mlx5e_encap_entry *e, struct list_head *flow_list);
drivers/net/ethernet/mellanox/mlx5/core/en_tc.h
224
struct mlx5e_encap_entry *e);
drivers/net/ethernet/mellanox/mlx5/core/esw/indir_table.c
109
if (e->recirc_rule) {
drivers/net/ethernet/mellanox/mlx5/core/esw/indir_table.c
110
refcount_inc(&e->recirc_rule->refcnt);
drivers/net/ethernet/mellanox/mlx5/core/esw/indir_table.c
139
flow_act.fg = e->recirc_grp;
drivers/net/ethernet/mellanox/mlx5/core/esw/indir_table.c
146
handle = mlx5_add_flow_rules(e->ft, NULL, &flow_act, &dest, 1);
drivers/net/ethernet/mellanox/mlx5/core/esw/indir_table.c
156
e->recirc_rule = rule;
drivers/net/ethernet/mellanox/mlx5/core/esw/indir_table.c
172
struct mlx5_esw_indir_table_entry *e)
drivers/net/ethernet/mellanox/mlx5/core/esw/indir_table.c
174
struct mlx5_esw_indir_table_rule *rule = e->recirc_rule;
drivers/net/ethernet/mellanox/mlx5/core/esw/indir_table.c
187
e->recirc_rule = NULL;
drivers/net/ethernet/mellanox/mlx5/core/esw/indir_table.c
190
static int mlx5_create_indir_recirc_group(struct mlx5_esw_indir_table_entry *e)
drivers/net/ethernet/mellanox/mlx5/core/esw/indir_table.c
201
e->recirc_grp = mlx5_create_flow_group(e->ft, in);
drivers/net/ethernet/mellanox/mlx5/core/esw/indir_table.c
202
if (IS_ERR(e->recirc_grp))
drivers/net/ethernet/mellanox/mlx5/core/esw/indir_table.c
203
err = PTR_ERR(e->recirc_grp);
drivers/net/ethernet/mellanox/mlx5/core/esw/indir_table.c
210
struct mlx5_esw_indir_table_entry *e)
drivers/net/ethernet/mellanox/mlx5/core/esw/indir_table.c
224
e->fwd_grp = mlx5_create_flow_group(e->ft, in);
drivers/net/ethernet/mellanox/mlx5/core/esw/indir_table.c
225
if (IS_ERR(e->fwd_grp)) {
drivers/net/ethernet/mellanox/mlx5/core/esw/indir_table.c
226
err = PTR_ERR(e->fwd_grp);
drivers/net/ethernet/mellanox/mlx5/core/esw/indir_table.c
231
flow_act.fg = e->fwd_grp;
drivers/net/ethernet/mellanox/mlx5/core/esw/indir_table.c
233
dest.vport.num = e->vport;
drivers/net/ethernet/mellanox/mlx5/core/esw/indir_table.c
236
e->fwd_rule = mlx5_add_flow_rules(e->ft, NULL, &flow_act, &dest, 1);
drivers/net/ethernet/mellanox/mlx5/core/esw/indir_table.c
237
if (IS_ERR(e->fwd_rule)) {
drivers/net/ethernet/mellanox/mlx5/core/esw/indir_table.c
238
mlx5_destroy_flow_group(e->fwd_grp);
drivers/net/ethernet/mellanox/mlx5/core/esw/indir_table.c
239
err = PTR_ERR(e->fwd_rule);
drivers/net/ethernet/mellanox/mlx5/core/esw/indir_table.c
253
struct mlx5_esw_indir_table_entry *e;
drivers/net/ethernet/mellanox/mlx5/core/esw/indir_table.c
261
e = kzalloc_obj(*e);
drivers/net/ethernet/mellanox/mlx5/core/esw/indir_table.c
262
if (!e)
drivers/net/ethernet/mellanox/mlx5/core/esw/indir_table.c
275
e->ft = ft;
drivers/net/ethernet/mellanox/mlx5/core/esw/indir_table.c
276
e->vport = vport;
drivers/net/ethernet/mellanox/mlx5/core/esw/indir_table.c
277
e->fwd_ref = !decap;
drivers/net/ethernet/mellanox/mlx5/core/esw/indir_table.c
279
err = mlx5_create_indir_recirc_group(e);
drivers/net/ethernet/mellanox/mlx5/core/esw/indir_table.c
284
err = mlx5_esw_indir_table_rule_get(esw, attr, e);
drivers/net/ethernet/mellanox/mlx5/core/esw/indir_table.c
289
err = mlx5_create_indir_fwd_group(esw, e);
drivers/net/ethernet/mellanox/mlx5/core/esw/indir_table.c
293
hash_add(esw->fdb_table.offloads.indir->table, &e->hlist,
drivers/net/ethernet/mellanox/mlx5/core/esw/indir_table.c
296
return e;
drivers/net/ethernet/mellanox/mlx5/core/esw/indir_table.c
300
mlx5_esw_indir_table_rule_put(esw, e);
drivers/net/ethernet/mellanox/mlx5/core/esw/indir_table.c
302
mlx5_destroy_flow_group(e->recirc_grp);
drivers/net/ethernet/mellanox/mlx5/core/esw/indir_table.c
304
mlx5_destroy_flow_table(e->ft);
drivers/net/ethernet/mellanox/mlx5/core/esw/indir_table.c
306
kfree(e);
drivers/net/ethernet/mellanox/mlx5/core/esw/indir_table.c
313
struct mlx5_esw_indir_table_entry *e;
drivers/net/ethernet/mellanox/mlx5/core/esw/indir_table.c
316
hash_for_each_possible(esw->fdb_table.offloads.indir->table, e, hlist, key)
drivers/net/ethernet/mellanox/mlx5/core/esw/indir_table.c
317
if (e->vport == vport)
drivers/net/ethernet/mellanox/mlx5/core/esw/indir_table.c
318
return e;
drivers/net/ethernet/mellanox/mlx5/core/esw/indir_table.c
327
struct mlx5_esw_indir_table_entry *e;
drivers/net/ethernet/mellanox/mlx5/core/esw/indir_table.c
331
e = mlx5_esw_indir_table_entry_lookup(esw, vport);
drivers/net/ethernet/mellanox/mlx5/core/esw/indir_table.c
332
if (e) {
drivers/net/ethernet/mellanox/mlx5/core/esw/indir_table.c
334
e->fwd_ref++;
drivers/net/ethernet/mellanox/mlx5/core/esw/indir_table.c
336
err = mlx5_esw_indir_table_rule_get(esw, attr, e);
drivers/net/ethernet/mellanox/mlx5/core/esw/indir_table.c
341
e = mlx5_esw_indir_table_entry_create(esw, attr, vport, decap);
drivers/net/ethernet/mellanox/mlx5/core/esw/indir_table.c
342
if (IS_ERR(e)) {
drivers/net/ethernet/mellanox/mlx5/core/esw/indir_table.c
343
err = PTR_ERR(e);
drivers/net/ethernet/mellanox/mlx5/core/esw/indir_table.c
349
return e->ft;
drivers/net/ethernet/mellanox/mlx5/core/esw/indir_table.c
359
struct mlx5_esw_indir_table_entry *e;
drivers/net/ethernet/mellanox/mlx5/core/esw/indir_table.c
362
e = mlx5_esw_indir_table_entry_lookup(esw, vport);
drivers/net/ethernet/mellanox/mlx5/core/esw/indir_table.c
363
if (!e)
drivers/net/ethernet/mellanox/mlx5/core/esw/indir_table.c
367
e->fwd_ref--;
drivers/net/ethernet/mellanox/mlx5/core/esw/indir_table.c
369
mlx5_esw_indir_table_rule_put(esw, e);
drivers/net/ethernet/mellanox/mlx5/core/esw/indir_table.c
371
if (e->fwd_ref || e->recirc_rule)
drivers/net/ethernet/mellanox/mlx5/core/esw/indir_table.c
374
hash_del(&e->hlist);
drivers/net/ethernet/mellanox/mlx5/core/esw/indir_table.c
375
mlx5_destroy_flow_group(e->recirc_grp);
drivers/net/ethernet/mellanox/mlx5/core/esw/indir_table.c
376
mlx5_del_flow_rules(e->fwd_rule);
drivers/net/ethernet/mellanox/mlx5/core/esw/indir_table.c
377
mlx5_destroy_flow_group(e->fwd_grp);
drivers/net/ethernet/mellanox/mlx5/core/esw/indir_table.c
378
mlx5_destroy_flow_table(e->ft);
drivers/net/ethernet/mellanox/mlx5/core/esw/indir_table.c
379
kfree(e);
drivers/net/ethernet/mellanox/mlx5/core/esw/indir_table.c
97
struct mlx5_esw_indir_table_entry *e)
drivers/net/ethernet/mellanox/mlx5/core/esw/vporttbl.c
100
if (!e) {
drivers/net/ethernet/mellanox/mlx5/core/esw/vporttbl.c
116
e->fdb = fdb;
drivers/net/ethernet/mellanox/mlx5/core/esw/vporttbl.c
117
e->num_rules = 1;
drivers/net/ethernet/mellanox/mlx5/core/esw/vporttbl.c
118
e->key = skey;
drivers/net/ethernet/mellanox/mlx5/core/esw/vporttbl.c
119
hash_add(esw->fdb_table.offloads.vports.table, &e->hlist, hkey);
drivers/net/ethernet/mellanox/mlx5/core/esw/vporttbl.c
122
return e->fdb;
drivers/net/ethernet/mellanox/mlx5/core/esw/vporttbl.c
125
kfree(e);
drivers/net/ethernet/mellanox/mlx5/core/esw/vporttbl.c
134
struct mlx5_vport_table *e;
drivers/net/ethernet/mellanox/mlx5/core/esw/vporttbl.c
141
e = esw_vport_tbl_lookup(esw, &key, hkey);
drivers/net/ethernet/mellanox/mlx5/core/esw/vporttbl.c
142
if (!e || --e->num_rules)
drivers/net/ethernet/mellanox/mlx5/core/esw/vporttbl.c
145
hash_del(&e->hlist);
drivers/net/ethernet/mellanox/mlx5/core/esw/vporttbl.c
146
mlx5_destroy_flow_table(e->fdb);
drivers/net/ethernet/mellanox/mlx5/core/esw/vporttbl.c
147
kfree(e);
drivers/net/ethernet/mellanox/mlx5/core/esw/vporttbl.c
71
struct mlx5_vport_table *e;
drivers/net/ethernet/mellanox/mlx5/core/esw/vporttbl.c
73
hash_for_each_possible(esw->fdb_table.offloads.vports.table, e, hlist, key)
drivers/net/ethernet/mellanox/mlx5/core/esw/vporttbl.c
74
if (!memcmp(&e->key, skey, sizeof(*skey)))
drivers/net/ethernet/mellanox/mlx5/core/esw/vporttbl.c
75
return e;
drivers/net/ethernet/mellanox/mlx5/core/esw/vporttbl.c
86
struct mlx5_vport_table *e;
drivers/net/ethernet/mellanox/mlx5/core/esw/vporttbl.c
93
e = esw_vport_tbl_lookup(esw, &skey, hkey);
drivers/net/ethernet/mellanox/mlx5/core/esw/vporttbl.c
94
if (e) {
drivers/net/ethernet/mellanox/mlx5/core/esw/vporttbl.c
95
e->num_rules++;
drivers/net/ethernet/mellanox/mlx5/core/esw/vporttbl.c
99
e = kzalloc_obj(*e);
drivers/net/ethernet/mellanox/mlx5/core/port.c
1019
MLX5_SET16(qpdpm_dscp_reg, qpdpm_dscp, e, 1);
drivers/net/ethernet/mellanox/mlxsw/pci_hw.h
203
MLXSW_ITEM32(pci, cqe0, e, 0x0C, 7, 1);
drivers/net/ethernet/mellanox/mlxsw/pci_hw.h
204
MLXSW_ITEM32(pci, cqe12, e, 0x00, 27, 1);
drivers/net/ethernet/mellanox/mlxsw/pci_hw.h
205
mlxsw_pci_cqe_item_helpers(e, 0, 12, 12);
drivers/net/ethernet/mellanox/mlxsw/reg.h
10006
u16 system_port, bool e,
drivers/net/ethernet/mellanox/mlxsw/reg.h
10012
mlxsw_reg_mpat_e_set(payload, e);
drivers/net/ethernet/mellanox/mlxsw/reg.h
10746
MLXSW_ITEM32(reg, mpsc, e, 0x04, 30, 1);
drivers/net/ethernet/mellanox/mlxsw/reg.h
10757
static inline void mlxsw_reg_mpsc_pack(char *payload, u16 local_port, bool e,
drivers/net/ethernet/mellanox/mlxsw/reg.h
10762
mlxsw_reg_mpsc_e_set(payload, e);
drivers/net/ethernet/mellanox/mlxsw/reg.h
2606
MLXSW_ITEM32(reg, ppbt, e, 0x00, 31, 1);
drivers/net/ethernet/mellanox/mlxsw/reg.h
2639
static inline void mlxsw_reg_ppbt_pack(char *payload, enum mlxsw_reg_pxbt_e e,
drivers/net/ethernet/mellanox/mlxsw/reg.h
2644
mlxsw_reg_ppbt_e_set(payload, e);
drivers/net/ethernet/mellanox/mlxsw/reg.h
5031
MLXSW_ITEM32(reg, paos, e, 0x04, 0, 2);
drivers/net/ethernet/mellanox/mlxsw/reg.h
6014
MLXSW_ITEM32(reg, pmaos, e, 0x04, 0, 2);
drivers/net/ethernet/mellanox/mlxsw/reg.h
6172
MLXSW_ITEM32(reg, pmecr, e, 0x04, 0, 2);
drivers/net/ethernet/mellanox/mlxsw/reg.h
6175
enum mlxsw_reg_pmecr_e e)
drivers/net/ethernet/mellanox/mlxsw/reg.h
6179
mlxsw_reg_pmecr_e_set(payload, e);
drivers/net/ethernet/mellanox/mlxsw/reg.h
9869
MLXSW_ITEM32(reg, mpat, e, 0x04, 31, 1);
drivers/net/ethernet/microchip/sparx5/sparx5_switchdev.c
109
struct sparx5_mdb_entry *e;
drivers/net/ethernet/microchip/sparx5/sparx5_switchdev.c
120
list_for_each_entry(e, &sparx5->mdb_entries, list) {
drivers/net/ethernet/microchip/sparx5/sparx5_switchdev.c
121
if (!test_bit(port->portno, e->port_mask) &&
drivers/net/ethernet/microchip/sparx5/sparx5_switchdev.c
122
ether_addr_is_ip_mcast(e->addr))
drivers/net/ethernet/microchip/sparx5/sparx5_switchdev.c
123
sparx5_pgid_update_mask(port, e->pgid_idx, enable);
drivers/net/ethernet/microchip/sparx5/sparx5_switchdev.c
516
struct sparx5_mdb_entry *e, *found = NULL;
drivers/net/ethernet/microchip/sparx5/sparx5_switchdev.c
519
list_for_each_entry(e, &sparx5->mdb_entries, list) {
drivers/net/ethernet/microchip/sparx5/sparx5_switchdev.c
520
if (ether_addr_equal(e->addr, addr) && e->vid == vid) {
drivers/net/ethernet/microchip/sparx5/sparx5_switchdev.c
521
found = e;
drivers/net/ethernet/microsoft/mana/gdma_main.c
334
union gdma_doorbell_entry e = {};
drivers/net/ethernet/microsoft/mana/gdma_main.c
338
e.eq.id = qid;
drivers/net/ethernet/microsoft/mana/gdma_main.c
339
e.eq.tail_ptr = tail_ptr;
drivers/net/ethernet/microsoft/mana/gdma_main.c
340
e.eq.arm = num_req;
drivers/net/ethernet/microsoft/mana/gdma_main.c
346
e.cq.id = qid;
drivers/net/ethernet/microsoft/mana/gdma_main.c
347
e.cq.tail_ptr = tail_ptr;
drivers/net/ethernet/microsoft/mana/gdma_main.c
348
e.cq.arm = num_req;
drivers/net/ethernet/microsoft/mana/gdma_main.c
354
e.rq.id = qid;
drivers/net/ethernet/microsoft/mana/gdma_main.c
355
e.rq.tail_ptr = tail_ptr;
drivers/net/ethernet/microsoft/mana/gdma_main.c
356
e.rq.wqe_cnt = num_req;
drivers/net/ethernet/microsoft/mana/gdma_main.c
362
e.sq.id = qid;
drivers/net/ethernet/microsoft/mana/gdma_main.c
363
e.sq.tail_ptr = tail_ptr;
drivers/net/ethernet/microsoft/mana/gdma_main.c
376
writeq(e.as_uint64, addr);
drivers/net/ethernet/realtek/r8169_main.c
2963
const struct ephy_info *e, int len)
drivers/net/ethernet/realtek/r8169_main.c
2968
w = (rtl_ephy_read(tp, e->offset) & ~e->mask) | e->bits;
drivers/net/ethernet/realtek/r8169_main.c
2969
rtl_ephy_write(tp, e->offset, w);
drivers/net/ethernet/realtek/r8169_main.c
2970
e++;
drivers/net/ethernet/sun/cassini.c
2997
const unsigned char *e = &cp->dev->dev_addr[0];
drivers/net/ethernet/sun/cassini.c
3045
writel((e[5] | (e[4] << 8)) & 0x3ff, cp->regs + REG_MAC_RANDOM_SEED);
drivers/net/ethernet/sun/cassini.c
3057
writel((e[4] << 8) | e[5], cp->regs + REG_MAC_ADDRN(0));
drivers/net/ethernet/sun/cassini.c
3058
writel((e[2] << 8) | e[3], cp->regs + REG_MAC_ADDRN(1));
drivers/net/ethernet/sun/cassini.c
3059
writel((e[0] << 8) | e[1], cp->regs + REG_MAC_ADDRN(2));
drivers/net/ethernet/sun/sunbmac.c
626
const unsigned char *e = &bp->dev->dev_addr[0];
drivers/net/ethernet/sun/sunbmac.c
647
sbus_writel(((e[4] << 8) | e[5]), bregs + BMAC_MACADDR2);
drivers/net/ethernet/sun/sunbmac.c
648
sbus_writel(((e[2] << 8) | e[3]), bregs + BMAC_MACADDR1);
drivers/net/ethernet/sun/sunbmac.c
649
sbus_writel(((e[0] << 8) | e[1]), bregs + BMAC_MACADDR0);
drivers/net/ethernet/sun/sunbmac.c
666
sbus_writel(((e[5] | e[4] << 8) & 0x3ff),
drivers/net/ethernet/sun/sungem.c
1802
const unsigned char *e = &gp->dev->dev_addr[0];
drivers/net/ethernet/sun/sungem.c
1820
writel((e[5] | (e[4] << 8)) & 0x3ff, gp->regs + MAC_RANDSEED);
drivers/net/ethernet/sun/sungem.c
1822
writel((e[4] << 8) | e[5], gp->regs + MAC_ADDR0);
drivers/net/ethernet/sun/sungem.c
1823
writel((e[2] << 8) | e[3], gp->regs + MAC_ADDR1);
drivers/net/ethernet/sun/sungem.c
1824
writel((e[0] << 8) | e[1], gp->regs + MAC_ADDR2);
drivers/net/ethernet/sun/sungem.c
2079
const unsigned char *e = &gp->dev->dev_addr[0];
drivers/net/ethernet/sun/sungem.c
2085
writel((e[4] << 8) | e[5], gp->regs + WOL_MATCH0);
drivers/net/ethernet/sun/sungem.c
2086
writel((e[2] << 8) | e[3], gp->regs + WOL_MATCH1);
drivers/net/ethernet/sun/sungem.c
2087
writel((e[0] << 8) | e[1], gp->regs + WOL_MATCH2);
drivers/net/ethernet/sun/sungem.c
2423
const unsigned char *e = &dev->dev_addr[0];
drivers/net/ethernet/sun/sungem.c
2439
writel((e[4] << 8) | e[5], gp->regs + MAC_ADDR0);
drivers/net/ethernet/sun/sungem.c
2440
writel((e[2] << 8) | e[3], gp->regs + MAC_ADDR1);
drivers/net/ethernet/sun/sungem.c
2441
writel((e[0] << 8) | e[1], gp->regs + MAC_ADDR2);
drivers/net/ethernet/sun/sunhme.c
1258
const unsigned char *e = &hp->dev->dev_addr[0];
drivers/net/ethernet/sun/sunhme.c
1338
hme_write32(hp, bregs + BMAC_RSEED, ((e[5] | e[4]<<8)&0x3ff));
drivers/net/ethernet/sun/sunhme.c
1340
hme_write32(hp, bregs + BMAC_MACADDR2, ((e[4] << 8) | e[5]));
drivers/net/ethernet/sun/sunhme.c
1341
hme_write32(hp, bregs + BMAC_MACADDR1, ((e[2] << 8) | e[3]));
drivers/net/ethernet/sun/sunhme.c
1342
hme_write32(hp, bregs + BMAC_MACADDR0, ((e[0] << 8) | e[1]));
drivers/net/ethernet/sun/sunqe.c
147
const unsigned char *e = &qep->dev->dev_addr[0];
drivers/net/ethernet/sun/sunqe.c
206
sbus_writeb(e[0], mregs + MREGS_ETHADDR);
drivers/net/ethernet/sun/sunqe.c
207
sbus_writeb(e[1], mregs + MREGS_ETHADDR);
drivers/net/ethernet/sun/sunqe.c
208
sbus_writeb(e[2], mregs + MREGS_ETHADDR);
drivers/net/ethernet/sun/sunqe.c
209
sbus_writeb(e[3], mregs + MREGS_ETHADDR);
drivers/net/ethernet/sun/sunqe.c
210
sbus_writeb(e[4], mregs + MREGS_ETHADDR);
drivers/net/ethernet/sun/sunqe.c
211
sbus_writeb(e[5], mregs + MREGS_ETHADDR);
drivers/net/ethernet/ti/icssm/icssm_prueth_switch.c
313
struct fdb_mac_tbl_entry __iomem *e;
drivers/net/ethernet/ti/icssm/icssm_prueth_switch.c
322
e = &mac_tbl->mac_tbl_entry[mac_tbl_idx];
drivers/net/ethernet/ti/icssm/icssm_prueth_switch.c
323
memcpy_fromio(mac_from_hw, e->mac, ETH_ALEN);
drivers/net/ethernet/ti/icssm/icssm_prueth_switch.c
328
if (readb(&e->port) != port) {
drivers/net/ethernet/ti/icssm/icssm_prueth_switch.c
334
writeb(port, &e->port);
drivers/net/ethernet/ti/icssm/icssm_prueth_switch.c
338
writew(0, &e->age);
drivers/net/ethernet/toshiba/ps3_gelic_wireless.c
319
range->freq[chs].e = 6;
drivers/net/ethernet/toshiba/ps3_gelic_wireless.c
592
iwe.u.freq.e = 0; /* table value in MHz */
drivers/net/fddi/skfp/ecm.c
101
smc->e.ecm_line_state = FALSE ;
drivers/net/fddi/skfp/ecm.c
154
smc->e.DisconnectFlag = FALSE ;
drivers/net/fddi/skfp/ecm.c
157
smc->e.DisconnectFlag = TRUE ;
drivers/net/fddi/skfp/ecm.c
165
smc->e.path_test = PT_PASSED ;
drivers/net/fddi/skfp/ecm.c
166
smc->e.ecm_line_state = FALSE ;
drivers/net/fddi/skfp/ecm.c
173
&& smc->e.path_test==PT_PASSED) {
drivers/net/fddi/skfp/ecm.c
178
else if (cmd == EC_CONNECT && (smc->e.path_test==PT_PASSED) &&
drivers/net/fddi/skfp/ecm.c
187
smc->e.trace_prop = 0 ;
drivers/net/fddi/skfp/ecm.c
221
smc->e.path_test = PT_EXITING ;
drivers/net/fddi/skfp/ecm.c
226
else if (smc->e.path_test == PT_PENDING) {
drivers/net/fddi/skfp/ecm.c
237
smc->e.path_test = PT_PENDING ;
drivers/net/fddi/skfp/ecm.c
251
(smc->e.path_test != PT_PENDING)) {
drivers/net/fddi/skfp/ecm.c
257
(smc->e.path_test == PT_PENDING)) {
drivers/net/fddi/skfp/ecm.c
262
else if (cmd == EC_CONNECT && smc->e.path_test == PT_PASSED) {
drivers/net/fddi/skfp/ecm.c
268
smc->e.path_test == PT_PENDING) {
drivers/net/fddi/skfp/ecm.c
269
smc->e.path_test = PT_EXITING ;
drivers/net/fddi/skfp/ecm.c
277
smc->e.path_test != PT_PENDING) {
drivers/net/fddi/skfp/ecm.c
284
smc->e.path_test = PT_TESTING ;
drivers/net/fddi/skfp/ecm.c
292
smc->e.path_test = PT_PASSED ;
drivers/net/fddi/skfp/ecm.c
294
if (smc->e.path_test == PT_FAILED)
drivers/net/fddi/skfp/ecm.c
298
if (smc->e.path_test == PT_FAILED &&
drivers/net/fddi/skfp/ecm.c
310
else if (smc->e.path_test == PT_PASSED) {
drivers/net/fddi/skfp/ecm.c
315
else if (smc->e.path_test == PT_FAILED &&
drivers/net/fddi/skfp/ecm.c
350
smc->e.ecm_line_state = TRUE ; /* flag to pcm: report Q/HLS */
drivers/net/fddi/skfp/ecm.c
360
smc->e.sb_flag = FALSE ;
drivers/net/fddi/skfp/ecm.c
361
smc->e.ecm_line_state = FALSE ;
drivers/net/fddi/skfp/ecm.c
366
else if (!smc->e.sb_flag &&
drivers/net/fddi/skfp/ecm.c
369
smc->e.sb_flag = TRUE ;
drivers/net/fddi/skfp/ecm.c
377
smc->e.ecm_line_state = FALSE ;
drivers/net/fddi/skfp/ecm.c
400
else if (cmd == EC_CONNECT && smc->e.path_test == PT_PASSED) {
drivers/net/fddi/skfp/ecm.c
434
DB_ECM("ECM : prop_actions - trace_prop %lu", smc->e.trace_prop);
drivers/net/fddi/skfp/ecm.c
437
if (smc->e.trace_prop & ENTITY_BIT(ENTITY_MAC)) {
drivers/net/fddi/skfp/ecm.c
442
else if ((smc->e.trace_prop & ENTITY_BIT(ENTITY_PHY(PA))) &&
drivers/net/fddi/skfp/ecm.c
448
else if ((smc->e.trace_prop & ENTITY_BIT(ENTITY_PHY(PB))) &&
drivers/net/fddi/skfp/ecm.c
457
smc->e.path_test = PT_PENDING ;
drivers/net/fddi/skfp/ecm.c
459
smc->e.trace_prop = 0 ;
drivers/net/fddi/skfp/ecm.c
472
while (smc->e.trace_prop) {
drivers/net/fddi/skfp/ecm.c
474
smc->e.trace_prop);
drivers/net/fddi/skfp/ecm.c
476
if (smc->e.trace_prop & ENTITY_BIT(ENTITY_MAC)) {
drivers/net/fddi/skfp/ecm.c
478
smc->e.trace_prop &= ~ENTITY_BIT(ENTITY_MAC) ;
drivers/net/fddi/skfp/ecm.c
483
if (smc->e.trace_prop &
drivers/net/fddi/skfp/ecm.c
488
smc->e.trace_prop &= ~ENTITY_BIT(ENTITY_PHY(p)) ;
drivers/net/fddi/skfp/ecm.c
495
smc->e.path_test = PT_PENDING ;
drivers/net/fddi/skfp/ecm.c
513
smt_timer_start(smc,&smc->e.ecm_timer,value,EV_TOKEN(EVENT_ECM,event));
drivers/net/fddi/skfp/ecm.c
522
if (smc->e.ecm_timer.tm_active)
drivers/net/fddi/skfp/ecm.c
523
smt_timer_stop(smc,&smc->e.ecm_timer) ;
drivers/net/fddi/skfp/ecm.c
97
smc->e.path_test = PT_PASSED ;
drivers/net/fddi/skfp/ecm.c
98
smc->e.trace_prop = 0 ;
drivers/net/fddi/skfp/ecm.c
99
smc->e.sb_flag = 0 ;
drivers/net/fddi/skfp/h/smc.h
440
struct s_ecm e ; /* ecm */
drivers/net/fddi/skfp/pcmplc.c
1764
DB_PCMN(1, "PLC %d: MDcF = %x", np, smc->e.DisconnectFlag);
drivers/net/fddi/skfp/pcmplc.c
1765
if (smc->e.DisconnectFlag == FALSE) {
drivers/net/fddi/skfp/pcmplc.c
1795
smc->e.trace_prop |= ENTITY_BIT(ENTITY_PHY(np)) ;
drivers/net/fddi/skfp/pcmplc.c
1805
if (smc->e.path_test == PT_PASSED) {
drivers/net/fddi/skfp/pcmplc.c
1810
smc->e.path_test = PT_PENDING ;
drivers/net/fddi/skfp/pmf.c
39
#define MOFFSS(e) offsetof(struct fddi_mib, e)
drivers/net/fddi/skfp/pmf.c
40
#define MOFFMS(e) offsetof(struct fddi_mib_m, e)
drivers/net/fddi/skfp/pmf.c
41
#define MOFFAS(e) offsetof(struct fddi_mib_a, e)
drivers/net/fddi/skfp/pmf.c
42
#define MOFFPS(e) offsetof(struct fddi_mib_p, e)
drivers/net/fddi/skfp/rmt.c
504
smc->e.trace_prop |= ENTITY_BIT(ENTITY_MAC) ;
drivers/net/fddi/skfp/smt.c
1787
static int entity_to_index(struct s_smc *smc, int e)
drivers/net/fddi/skfp/smt.c
1789
if (e == ENTITY_MAC)
drivers/net/fddi/skfp/smt.c
1792
return phy_index(smc, e - ENTITY_PHY(0));
drivers/net/ieee802154/mac802154_hwsim.c
1003
list_for_each_entry_rcu(e, &phy->edges, list) {
drivers/net/ieee802154/mac802154_hwsim.c
1004
list_del_rcu(&e->list);
drivers/net/ieee802154/mac802154_hwsim.c
1005
hwsim_free_edge(e);
drivers/net/ieee802154/mac802154_hwsim.c
261
struct hwsim_edge *e;
drivers/net/ieee802154/mac802154_hwsim.c
267
list_for_each_entry_rcu(e, &current_phy->edges, list) {
drivers/net/ieee802154/mac802154_hwsim.c
273
if (e->endpoint->suspended)
drivers/net/ieee802154/mac802154_hwsim.c
276
endpoint_pib = rcu_dereference(e->endpoint->pib);
drivers/net/ieee802154/mac802154_hwsim.c
281
einfo = rcu_dereference(e->info);
drivers/net/ieee802154/mac802154_hwsim.c
283
hwsim_hw_receive(e->endpoint->hw, newskb, einfo->lqi);
drivers/net/ieee802154/mac802154_hwsim.c
372
struct hwsim_edge *e;
drivers/net/ieee802154/mac802154_hwsim.c
392
list_for_each_entry_rcu(e, &phy->edges, list) {
drivers/net/ieee802154/mac802154_hwsim.c
402
e->endpoint->idx);
drivers/net/ieee802154/mac802154_hwsim.c
410
einfo = rcu_dereference(e->info);
drivers/net/ieee802154/mac802154_hwsim.c
546
struct hwsim_edge *e;
drivers/net/ieee802154/mac802154_hwsim.c
548
e = kzalloc_obj(*e);
drivers/net/ieee802154/mac802154_hwsim.c
549
if (!e)
drivers/net/ieee802154/mac802154_hwsim.c
554
kfree(e);
drivers/net/ieee802154/mac802154_hwsim.c
559
rcu_assign_pointer(e->info, einfo);
drivers/net/ieee802154/mac802154_hwsim.c
560
e->endpoint = endpoint;
drivers/net/ieee802154/mac802154_hwsim.c
562
return e;
drivers/net/ieee802154/mac802154_hwsim.c
565
static void hwsim_free_edge(struct hwsim_edge *e)
drivers/net/ieee802154/mac802154_hwsim.c
570
einfo = rcu_dereference(e->info);
drivers/net/ieee802154/mac802154_hwsim.c
574
kfree_rcu(e, rcu);
drivers/net/ieee802154/mac802154_hwsim.c
581
struct hwsim_edge *e;
drivers/net/ieee802154/mac802154_hwsim.c
614
list_for_each_entry_rcu(e, &phy_v0->edges, list) {
drivers/net/ieee802154/mac802154_hwsim.c
615
if (e->endpoint->idx == v1) {
drivers/net/ieee802154/mac802154_hwsim.c
623
e = hwsim_alloc_edge(phy_v1, 0xff);
drivers/net/ieee802154/mac802154_hwsim.c
624
if (!e) {
drivers/net/ieee802154/mac802154_hwsim.c
628
list_add_rcu(&e->list, &phy_v0->edges);
drivers/net/ieee802154/mac802154_hwsim.c
643
struct hwsim_edge *e;
drivers/net/ieee802154/mac802154_hwsim.c
667
list_for_each_entry_rcu(e, &phy_v0->edges, list) {
drivers/net/ieee802154/mac802154_hwsim.c
668
if (e->endpoint->idx == v1) {
drivers/net/ieee802154/mac802154_hwsim.c
670
list_del_rcu(&e->list);
drivers/net/ieee802154/mac802154_hwsim.c
671
hwsim_free_edge(e);
drivers/net/ieee802154/mac802154_hwsim.c
690
struct hwsim_edge *e;
drivers/net/ieee802154/mac802154_hwsim.c
723
list_for_each_entry_rcu(e, &phy_v0->edges, list) {
drivers/net/ieee802154/mac802154_hwsim.c
724
if (e->endpoint->idx == v1) {
drivers/net/ieee802154/mac802154_hwsim.c
726
einfo_old = rcu_replace_pointer(e->info, einfo,
drivers/net/ieee802154/mac802154_hwsim.c
844
struct hwsim_edge *e;
drivers/net/ieee802154/mac802154_hwsim.c
849
list_for_each_entry_rcu(e, &tmp->edges, list) {
drivers/net/ieee802154/mac802154_hwsim.c
850
if (e->endpoint->idx == phy->idx) {
drivers/net/ieee802154/mac802154_hwsim.c
851
list_del_rcu(&e->list);
drivers/net/ieee802154/mac802154_hwsim.c
852
hwsim_free_edge(e);
drivers/net/ieee802154/mac802154_hwsim.c
864
struct hwsim_edge *e;
drivers/net/ieee802154/mac802154_hwsim.c
867
e = hwsim_alloc_edge(sub, 0xff);
drivers/net/ieee802154/mac802154_hwsim.c
868
if (!e)
drivers/net/ieee802154/mac802154_hwsim.c
871
list_add_rcu(&e->list, &phy->edges);
drivers/net/ieee802154/mac802154_hwsim.c
875
e = hwsim_alloc_edge(phy, 0xff);
drivers/net/ieee802154/mac802154_hwsim.c
876
if (!e)
drivers/net/ieee802154/mac802154_hwsim.c
879
list_add_rcu(&e->list, &sub->edges);
drivers/net/ieee802154/mac802154_hwsim.c
888
list_for_each_entry_rcu(e, &phy->edges, list) {
drivers/net/ieee802154/mac802154_hwsim.c
889
list_del_rcu(&e->list);
drivers/net/ieee802154/mac802154_hwsim.c
890
hwsim_free_edge(e);
drivers/net/ieee802154/mac802154_hwsim.c
996
struct hwsim_edge *e;
drivers/net/netdevsim/bpf.c
335
nsim_map_key_match(struct bpf_map *map, struct nsim_map_entry *e, void *key)
drivers/net/netdevsim/bpf.c
337
return e->key && !memcmp(key, e->key, map->key_size);
drivers/net/tun.c
308
struct tun_flow_entry *e;
drivers/net/tun.c
310
hlist_for_each_entry_rcu(e, head, hash_link) {
drivers/net/tun.c
311
if (e->rxhash == rxhash)
drivers/net/tun.c
312
return e;
drivers/net/tun.c
321
struct tun_flow_entry *e = kmalloc_obj(*e, GFP_ATOMIC);
drivers/net/tun.c
323
if (e) {
drivers/net/tun.c
327
e->updated = jiffies;
drivers/net/tun.c
328
e->rxhash = rxhash;
drivers/net/tun.c
329
e->rps_rxhash = 0;
drivers/net/tun.c
330
e->queue_index = queue_index;
drivers/net/tun.c
331
e->tun = tun;
drivers/net/tun.c
332
hlist_add_head_rcu(&e->hash_link, head);
drivers/net/tun.c
335
return e;
drivers/net/tun.c
338
static void tun_flow_delete(struct tun_struct *tun, struct tun_flow_entry *e)
drivers/net/tun.c
341
e->rxhash, e->queue_index);
drivers/net/tun.c
342
hlist_del_rcu(&e->hash_link);
drivers/net/tun.c
343
kfree_rcu(e, rcu);
drivers/net/tun.c
353
struct tun_flow_entry *e;
drivers/net/tun.c
356
hlist_for_each_entry_safe(e, n, &tun->flows[i], hash_link)
drivers/net/tun.c
357
tun_flow_delete(tun, e);
drivers/net/tun.c
368
struct tun_flow_entry *e;
drivers/net/tun.c
371
hlist_for_each_entry_safe(e, n, &tun->flows[i], hash_link) {
drivers/net/tun.c
372
if (e->queue_index == queue_index)
drivers/net/tun.c
373
tun_flow_delete(tun, e);
drivers/net/tun.c
389
struct tun_flow_entry *e;
drivers/net/tun.c
392
hlist_for_each_entry_safe(e, n, &tun->flows[i], hash_link) {
drivers/net/tun.c
395
this_timer = e->updated + delay;
drivers/net/tun.c
397
tun_flow_delete(tun, e);
drivers/net/tun.c
415
struct tun_flow_entry *e;
drivers/net/tun.c
423
e = tun_flow_find(head, rxhash);
drivers/net/tun.c
424
if (likely(e)) {
drivers/net/tun.c
426
if (READ_ONCE(e->queue_index) != queue_index)
drivers/net/tun.c
427
WRITE_ONCE(e->queue_index, queue_index);
drivers/net/tun.c
428
if (e->updated != jiffies)
drivers/net/tun.c
429
e->updated = jiffies;
drivers/net/tun.c
430
sock_rps_record_flow_hash(e->rps_rxhash);
drivers/net/tun.c
449
static inline void tun_flow_save_rps_rxhash(struct tun_flow_entry *e, u32 hash)
drivers/net/tun.c
451
if (unlikely(e->rps_rxhash != hash))
drivers/net/tun.c
452
e->rps_rxhash = hash;
drivers/net/tun.c
463
struct tun_flow_entry *e;
drivers/net/tun.c
469
e = tun_flow_find(&tun->flows[tun_hashfn(txq)], txq);
drivers/net/tun.c
470
if (e) {
drivers/net/tun.c
471
tun_flow_save_rps_rxhash(e, txq);
drivers/net/tun.c
472
txq = e->queue_index;
drivers/net/tun.c
979
struct tun_flow_entry *e;
drivers/net/tun.c
983
e = tun_flow_find(&tun->flows[tun_hashfn(rxhash)], rxhash);
drivers/net/tun.c
984
if (e)
drivers/net/tun.c
985
tun_flow_save_rps_rxhash(e, rxhash);
drivers/net/usb/cdc_ncm.c
728
struct usb_host_endpoint *e, *in = NULL, *out = NULL;
drivers/net/usb/cdc_ncm.c
732
e = intf->cur_altsetting->endpoint + ep;
drivers/net/usb/cdc_ncm.c
735
if (!usb_endpoint_maxp(&e->desc))
drivers/net/usb/cdc_ncm.c
738
switch (e->desc.bmAttributes & USB_ENDPOINT_XFERTYPE_MASK) {
drivers/net/usb/cdc_ncm.c
740
if (usb_endpoint_dir_in(&e->desc)) {
drivers/net/usb/cdc_ncm.c
742
dev->status = e;
drivers/net/usb/cdc_ncm.c
747
if (usb_endpoint_dir_in(&e->desc)) {
drivers/net/usb/cdc_ncm.c
749
in = e;
drivers/net/usb/cdc_ncm.c
752
out = e;
drivers/net/usb/usbnet.c
108
struct usb_host_endpoint *e;
drivers/net/usb/usbnet.c
111
e = alt->endpoint + ep;
drivers/net/usb/usbnet.c
114
if (!usb_endpoint_maxp(&e->desc))
drivers/net/usb/usbnet.c
117
switch (e->desc.bmAttributes) {
drivers/net/usb/usbnet.c
119
if (!usb_endpoint_dir_in(&e->desc))
drivers/net/usb/usbnet.c
128
if (usb_endpoint_dir_in(&e->desc)) {
drivers/net/usb/usbnet.c
130
in = e;
drivers/net/usb/usbnet.c
132
status = e;
drivers/net/usb/usbnet.c
135
out = e;
drivers/net/vxlan/vxlan_mdb.c
100
struct br_mdb_entry *e)
drivers/net/vxlan/vxlan_mdb.c
104
memset(e, 0, sizeof(*e));
drivers/net/vxlan/vxlan_mdb.c
105
e->ifindex = vxlan->dev->ifindex;
drivers/net/vxlan/vxlan_mdb.c
106
e->state = MDB_PERMANENT;
drivers/net/vxlan/vxlan_mdb.c
109
e->flags |= MDB_FLAGS_BLOCKED;
drivers/net/vxlan/vxlan_mdb.c
113
e->addr.u.ip4 = dst->sin.sin_addr.s_addr;
drivers/net/vxlan/vxlan_mdb.c
114
e->addr.proto = htons(ETH_P_IP);
drivers/net/vxlan/vxlan_mdb.c
118
e->addr.u.ip6 = dst->sin6.sin6_addr;
drivers/net/vxlan/vxlan_mdb.c
119
e->addr.proto = htons(ETH_P_IPV6);
drivers/net/vxlan/vxlan_mdb.c
168
struct br_mdb_entry e;
drivers/net/vxlan/vxlan_mdb.c
175
vxlan_br_mdb_entry_fill(vxlan, mdb_entry, remote, &e);
drivers/net/vxlan/vxlan_mdb.c
177
if (nla_put_nohdr(skb, sizeof(e), &e) ||
drivers/net/wireguard/noise.c
594
u8 e[NOISE_PUBLIC_KEY_LEN];
drivers/net/wireguard/noise.c
605
message_ephemeral(e, src->unencrypted_ephemeral, chaining_key, hash);
drivers/net/wireguard/noise.c
608
if (!mix_dh(chaining_key, key, wg->static_identity.static_private, e))
drivers/net/wireguard/noise.c
644
memcpy(handshake->remote_ephemeral, e, NOISE_PUBLIC_KEY_LEN);
drivers/net/wireguard/noise.c
737
u8 e[NOISE_PUBLIC_KEY_LEN];
drivers/net/wireguard/noise.c
767
message_ephemeral(e, src->unencrypted_ephemeral, chaining_key, hash);
drivers/net/wireguard/noise.c
770
if (!mix_dh(chaining_key, NULL, ephemeral_private, e))
drivers/net/wireguard/noise.c
774
if (!mix_dh(chaining_key, NULL, wg->static_identity.static_private, e))
drivers/net/wireguard/noise.c
794
memcpy(handshake->remote_ephemeral, e, NOISE_PUBLIC_KEY_LEN);
drivers/net/wireguard/selftest/allowedips.c
497
*d = init_peer(), *e = init_peer(), *f = init_peer(),
drivers/net/wireguard/selftest/allowedips.c
511
if (!a || !b || !c || !d || !e || !f || !g || !h) {
drivers/net/wireguard/selftest/allowedips.c
524
insert(4, e, 0, 0, 0, 0, 0);
drivers/net/wireguard/selftest/allowedips.c
525
insert(6, e, 0, 0, 0, 0, 0);
drivers/net/wireguard/selftest/allowedips.c
556
test(4, e, 192, 95, 5, 96);
drivers/net/wireguard/selftest/allowedips.c
712
kfree(e);
drivers/net/wireless/ath/ath9k/common-debug.c
127
#define RXS_ERR(s, e) \
drivers/net/wireless/ath/ath9k/common-debug.c
131
rxstats->e); \
drivers/net/wireless/ath/dfs_pri_detector.h
62
(*add_pulse)(struct pri_detector *de, struct pulse_event *e);
drivers/net/wireless/broadcom/b43/debugfs.c
645
struct b43_dfsentry *e = dev->dfsentry;
drivers/net/wireless/broadcom/b43/debugfs.c
648
e->dyn_debug[id] = (initstate); \
drivers/net/wireless/broadcom/b43/debugfs.c
649
debugfs_create_bool(name, 0600, e->subdir, \
drivers/net/wireless/broadcom/b43/debugfs.c
650
&(e->dyn_debug[id])); \
drivers/net/wireless/broadcom/b43/debugfs.c
668
struct b43_dfsentry *e;
drivers/net/wireless/broadcom/b43/debugfs.c
673
e = kzalloc_obj(*e);
drivers/net/wireless/broadcom/b43/debugfs.c
674
if (!e) {
drivers/net/wireless/broadcom/b43/debugfs.c
678
e->dev = dev;
drivers/net/wireless/broadcom/b43/debugfs.c
679
log = &e->txstatlog;
drivers/net/wireless/broadcom/b43/debugfs.c
683
kfree(e);
drivers/net/wireless/broadcom/b43/debugfs.c
688
dev->dfsentry = e;
drivers/net/wireless/broadcom/b43/debugfs.c
691
e->subdir = debugfs_create_dir(devdir, rootdir);
drivers/net/wireless/broadcom/b43/debugfs.c
693
e->mmio16read_next = 0xFFFF; /* invalid address */
drivers/net/wireless/broadcom/b43/debugfs.c
694
e->mmio32read_next = 0xFFFF; /* invalid address */
drivers/net/wireless/broadcom/b43/debugfs.c
695
e->shm16read_routing_next = 0xFFFFFFFF; /* invalid routing */
drivers/net/wireless/broadcom/b43/debugfs.c
696
e->shm16read_addr_next = 0xFFFFFFFF; /* invalid address */
drivers/net/wireless/broadcom/b43/debugfs.c
697
e->shm32read_routing_next = 0xFFFFFFFF; /* invalid routing */
drivers/net/wireless/broadcom/b43/debugfs.c
698
e->shm32read_addr_next = 0xFFFFFFFF; /* invalid address */
drivers/net/wireless/broadcom/b43/debugfs.c
703
mode, e->subdir, dev, \
drivers/net/wireless/broadcom/b43/debugfs.c
727
struct b43_dfsentry *e;
drivers/net/wireless/broadcom/b43/debugfs.c
731
e = dev->dfsentry;
drivers/net/wireless/broadcom/b43/debugfs.c
732
if (!e)
drivers/net/wireless/broadcom/b43/debugfs.c
735
debugfs_remove(e->subdir);
drivers/net/wireless/broadcom/b43/debugfs.c
736
kfree(e->txstatlog.log);
drivers/net/wireless/broadcom/b43/debugfs.c
737
kfree(e);
drivers/net/wireless/broadcom/b43/debugfs.c
743
struct b43_dfsentry *e = dev->dfsentry;
drivers/net/wireless/broadcom/b43/debugfs.c
748
if (!e)
drivers/net/wireless/broadcom/b43/debugfs.c
750
log = &e->txstatlog;
drivers/net/wireless/broadcom/b43/phy_ht.c
36
const struct b43_phy_ht_channeltab_e_radio2059 *e)
drivers/net/wireless/broadcom/b43/phy_ht.c
42
b43_radio_write(dev, 0x16, e->radio_syn16);
drivers/net/wireless/broadcom/b43/phy_ht.c
43
b43_radio_write(dev, 0x17, e->radio_syn17);
drivers/net/wireless/broadcom/b43/phy_ht.c
44
b43_radio_write(dev, 0x22, e->radio_syn22);
drivers/net/wireless/broadcom/b43/phy_ht.c
45
b43_radio_write(dev, 0x25, e->radio_syn25);
drivers/net/wireless/broadcom/b43/phy_ht.c
46
b43_radio_write(dev, 0x27, e->radio_syn27);
drivers/net/wireless/broadcom/b43/phy_ht.c
47
b43_radio_write(dev, 0x28, e->radio_syn28);
drivers/net/wireless/broadcom/b43/phy_ht.c
48
b43_radio_write(dev, 0x29, e->radio_syn29);
drivers/net/wireless/broadcom/b43/phy_ht.c
49
b43_radio_write(dev, 0x2c, e->radio_syn2c);
drivers/net/wireless/broadcom/b43/phy_ht.c
50
b43_radio_write(dev, 0x2d, e->radio_syn2d);
drivers/net/wireless/broadcom/b43/phy_ht.c
51
b43_radio_write(dev, 0x37, e->radio_syn37);
drivers/net/wireless/broadcom/b43/phy_ht.c
52
b43_radio_write(dev, 0x41, e->radio_syn41);
drivers/net/wireless/broadcom/b43/phy_ht.c
53
b43_radio_write(dev, 0x43, e->radio_syn43);
drivers/net/wireless/broadcom/b43/phy_ht.c
54
b43_radio_write(dev, 0x47, e->radio_syn47);
drivers/net/wireless/broadcom/b43/phy_ht.c
58
b43_radio_write(dev, r | 0x4a, e->radio_rxtx4a);
drivers/net/wireless/broadcom/b43/phy_ht.c
59
b43_radio_write(dev, r | 0x58, e->radio_rxtx58);
drivers/net/wireless/broadcom/b43/phy_ht.c
60
b43_radio_write(dev, r | 0x5a, e->radio_rxtx5a);
drivers/net/wireless/broadcom/b43/phy_ht.c
61
b43_radio_write(dev, r | 0x6a, e->radio_rxtx6a);
drivers/net/wireless/broadcom/b43/phy_ht.c
62
b43_radio_write(dev, r | 0x6d, e->radio_rxtx6d);
drivers/net/wireless/broadcom/b43/phy_ht.c
63
b43_radio_write(dev, r | 0x6e, e->radio_rxtx6e);
drivers/net/wireless/broadcom/b43/phy_ht.c
64
b43_radio_write(dev, r | 0x92, e->radio_rxtx92);
drivers/net/wireless/broadcom/b43/phy_ht.c
65
b43_radio_write(dev, r | 0x98, e->radio_rxtx98);
drivers/net/wireless/broadcom/b43/phy_ht.c
764
const struct b43_phy_ht_channeltab_e_phy *e,
drivers/net/wireless/broadcom/b43/phy_ht.c
782
b43_phy_write(dev, B43_PHY_HT_BW1, e->bw1);
drivers/net/wireless/broadcom/b43/phy_ht.c
783
b43_phy_write(dev, B43_PHY_HT_BW2, e->bw2);
drivers/net/wireless/broadcom/b43/phy_ht.c
784
b43_phy_write(dev, B43_PHY_HT_BW3, e->bw3);
drivers/net/wireless/broadcom/b43/phy_ht.c
785
b43_phy_write(dev, B43_PHY_HT_BW4, e->bw4);
drivers/net/wireless/broadcom/b43/phy_ht.c
786
b43_phy_write(dev, B43_PHY_HT_BW5, e->bw5);
drivers/net/wireless/broadcom/b43/phy_ht.c
787
b43_phy_write(dev, B43_PHY_HT_BW6, e->bw6);
drivers/net/wireless/broadcom/b43/phy_lp.c
653
const struct lpphy_stx_table_entry *e;
drivers/net/wireless/broadcom/b43/phy_lp.c
658
e = &lpphy_stx_table[i];
drivers/net/wireless/broadcom/b43/phy_lp.c
659
tmp = b43_radio_read(dev, e->rf_addr);
drivers/net/wireless/broadcom/b43/phy_lp.c
660
tmp >>= e->rf_shift;
drivers/net/wireless/broadcom/b43/phy_lp.c
661
tmp <<= e->phy_shift;
drivers/net/wireless/broadcom/b43/phy_lp.c
662
b43_phy_maskset(dev, B43_PHY_OFDM(0xF2 + e->phy_offset),
drivers/net/wireless/broadcom/b43/phy_lp.c
663
~(e->mask << e->phy_shift), tmp);
drivers/net/wireless/broadcom/b43/phy_n.c
1046
const struct b43_nphy_channeltab_entry_rev3 *e)
drivers/net/wireless/broadcom/b43/phy_n.c
1048
b43_radio_write(dev, B2056_SYN_PLL_VCOCAL1, e->radio_syn_pll_vcocal1);
drivers/net/wireless/broadcom/b43/phy_n.c
1049
b43_radio_write(dev, B2056_SYN_PLL_VCOCAL2, e->radio_syn_pll_vcocal2);
drivers/net/wireless/broadcom/b43/phy_n.c
1050
b43_radio_write(dev, B2056_SYN_PLL_REFDIV, e->radio_syn_pll_refdiv);
drivers/net/wireless/broadcom/b43/phy_n.c
1051
b43_radio_write(dev, B2056_SYN_PLL_MMD2, e->radio_syn_pll_mmd2);
drivers/net/wireless/broadcom/b43/phy_n.c
1052
b43_radio_write(dev, B2056_SYN_PLL_MMD1, e->radio_syn_pll_mmd1);
drivers/net/wireless/broadcom/b43/phy_n.c
1054
e->radio_syn_pll_loopfilter1);
drivers/net/wireless/broadcom/b43/phy_n.c
1056
e->radio_syn_pll_loopfilter2);
drivers/net/wireless/broadcom/b43/phy_n.c
1058
e->radio_syn_pll_loopfilter3);
drivers/net/wireless/broadcom/b43/phy_n.c
1060
e->radio_syn_pll_loopfilter4);
drivers/net/wireless/broadcom/b43/phy_n.c
1062
e->radio_syn_pll_loopfilter5);
drivers/net/wireless/broadcom/b43/phy_n.c
1064
e->radio_syn_reserved_addr27);
drivers/net/wireless/broadcom/b43/phy_n.c
1066
e->radio_syn_reserved_addr28);
drivers/net/wireless/broadcom/b43/phy_n.c
1068
e->radio_syn_reserved_addr29);
drivers/net/wireless/broadcom/b43/phy_n.c
1070
e->radio_syn_logen_vcobuf1);
drivers/net/wireless/broadcom/b43/phy_n.c
1071
b43_radio_write(dev, B2056_SYN_LOGEN_MIXER2, e->radio_syn_logen_mixer2);
drivers/net/wireless/broadcom/b43/phy_n.c
1072
b43_radio_write(dev, B2056_SYN_LOGEN_BUF3, e->radio_syn_logen_buf3);
drivers/net/wireless/broadcom/b43/phy_n.c
1073
b43_radio_write(dev, B2056_SYN_LOGEN_BUF4, e->radio_syn_logen_buf4);
drivers/net/wireless/broadcom/b43/phy_n.c
1076
e->radio_rx0_lnaa_tune);
drivers/net/wireless/broadcom/b43/phy_n.c
1078
e->radio_rx0_lnag_tune);
drivers/net/wireless/broadcom/b43/phy_n.c
1081
e->radio_tx0_intpaa_boost_tune);
drivers/net/wireless/broadcom/b43/phy_n.c
1083
e->radio_tx0_intpag_boost_tune);
drivers/net/wireless/broadcom/b43/phy_n.c
1085
e->radio_tx0_pada_boost_tune);
drivers/net/wireless/broadcom/b43/phy_n.c
1087
e->radio_tx0_padg_boost_tune);
drivers/net/wireless/broadcom/b43/phy_n.c
1089
e->radio_tx0_pgaa_boost_tune);
drivers/net/wireless/broadcom/b43/phy_n.c
1091
e->radio_tx0_pgag_boost_tune);
drivers/net/wireless/broadcom/b43/phy_n.c
1093
e->radio_tx0_mixa_boost_tune);
drivers/net/wireless/broadcom/b43/phy_n.c
1095
e->radio_tx0_mixg_boost_tune);
drivers/net/wireless/broadcom/b43/phy_n.c
1098
e->radio_rx1_lnaa_tune);
drivers/net/wireless/broadcom/b43/phy_n.c
1100
e->radio_rx1_lnag_tune);
drivers/net/wireless/broadcom/b43/phy_n.c
1103
e->radio_tx1_intpaa_boost_tune);
drivers/net/wireless/broadcom/b43/phy_n.c
1105
e->radio_tx1_intpag_boost_tune);
drivers/net/wireless/broadcom/b43/phy_n.c
1107
e->radio_tx1_pada_boost_tune);
drivers/net/wireless/broadcom/b43/phy_n.c
1109
e->radio_tx1_padg_boost_tune);
drivers/net/wireless/broadcom/b43/phy_n.c
1111
e->radio_tx1_pgaa_boost_tune);
drivers/net/wireless/broadcom/b43/phy_n.c
1113
e->radio_tx1_pgag_boost_tune);
drivers/net/wireless/broadcom/b43/phy_n.c
1115
e->radio_tx1_mixa_boost_tune);
drivers/net/wireless/broadcom/b43/phy_n.c
1117
e->radio_tx1_mixg_boost_tune);
drivers/net/wireless/broadcom/b43/phy_n.c
1122
const struct b43_nphy_channeltab_entry_rev3 *e)
drivers/net/wireless/broadcom/b43/phy_n.c
1142
b43_chantab_radio_2056_upload(dev, e);
drivers/net/wireless/broadcom/b43/phy_n.c
1373
const struct b43_nphy_channeltab_entry_rev2 *e)
drivers/net/wireless/broadcom/b43/phy_n.c
1375
b43_radio_write(dev, B2055_PLL_REF, e->radio_pll_ref);
drivers/net/wireless/broadcom/b43/phy_n.c
1376
b43_radio_write(dev, B2055_RF_PLLMOD0, e->radio_rf_pllmod0);
drivers/net/wireless/broadcom/b43/phy_n.c
1377
b43_radio_write(dev, B2055_RF_PLLMOD1, e->radio_rf_pllmod1);
drivers/net/wireless/broadcom/b43/phy_n.c
1378
b43_radio_write(dev, B2055_VCO_CAPTAIL, e->radio_vco_captail);
drivers/net/wireless/broadcom/b43/phy_n.c
1381
b43_radio_write(dev, B2055_VCO_CAL1, e->radio_vco_cal1);
drivers/net/wireless/broadcom/b43/phy_n.c
1382
b43_radio_write(dev, B2055_VCO_CAL2, e->radio_vco_cal2);
drivers/net/wireless/broadcom/b43/phy_n.c
1383
b43_radio_write(dev, B2055_PLL_LFC1, e->radio_pll_lfc1);
drivers/net/wireless/broadcom/b43/phy_n.c
1384
b43_radio_write(dev, B2055_PLL_LFR1, e->radio_pll_lfr1);
drivers/net/wireless/broadcom/b43/phy_n.c
1387
b43_radio_write(dev, B2055_PLL_LFC2, e->radio_pll_lfc2);
drivers/net/wireless/broadcom/b43/phy_n.c
1388
b43_radio_write(dev, B2055_LGBUF_CENBUF, e->radio_lgbuf_cenbuf);
drivers/net/wireless/broadcom/b43/phy_n.c
1389
b43_radio_write(dev, B2055_LGEN_TUNE1, e->radio_lgen_tune1);
drivers/net/wireless/broadcom/b43/phy_n.c
1390
b43_radio_write(dev, B2055_LGEN_TUNE2, e->radio_lgen_tune2);
drivers/net/wireless/broadcom/b43/phy_n.c
1393
b43_radio_write(dev, B2055_C1_LGBUF_ATUNE, e->radio_c1_lgbuf_atune);
drivers/net/wireless/broadcom/b43/phy_n.c
1394
b43_radio_write(dev, B2055_C1_LGBUF_GTUNE, e->radio_c1_lgbuf_gtune);
drivers/net/wireless/broadcom/b43/phy_n.c
1395
b43_radio_write(dev, B2055_C1_RX_RFR1, e->radio_c1_rx_rfr1);
drivers/net/wireless/broadcom/b43/phy_n.c
1396
b43_radio_write(dev, B2055_C1_TX_PGAPADTN, e->radio_c1_tx_pgapadtn);
drivers/net/wireless/broadcom/b43/phy_n.c
1399
b43_radio_write(dev, B2055_C1_TX_MXBGTRIM, e->radio_c1_tx_mxbgtrim);
drivers/net/wireless/broadcom/b43/phy_n.c
1400
b43_radio_write(dev, B2055_C2_LGBUF_ATUNE, e->radio_c2_lgbuf_atune);
drivers/net/wireless/broadcom/b43/phy_n.c
1401
b43_radio_write(dev, B2055_C2_LGBUF_GTUNE, e->radio_c2_lgbuf_gtune);
drivers/net/wireless/broadcom/b43/phy_n.c
1402
b43_radio_write(dev, B2055_C2_RX_RFR1, e->radio_c2_rx_rfr1);
drivers/net/wireless/broadcom/b43/phy_n.c
1405
b43_radio_write(dev, B2055_C2_TX_PGAPADTN, e->radio_c2_tx_pgapadtn);
drivers/net/wireless/broadcom/b43/phy_n.c
1406
b43_radio_write(dev, B2055_C2_TX_MXBGTRIM, e->radio_c2_tx_mxbgtrim);
drivers/net/wireless/broadcom/b43/phy_n.c
1411
const struct b43_nphy_channeltab_entry_rev2 *e)
drivers/net/wireless/broadcom/b43/phy_n.c
1415
b43_chantab_radio_upload(dev, e);
drivers/net/wireless/broadcom/b43/phy_n.c
155
const struct nphy_rf_control_override_rev7 *e;
drivers/net/wireless/broadcom/b43/phy_n.c
170
e = b43_nphy_get_rf_ctl_over_rev7(dev, field, override);
drivers/net/wireless/broadcom/b43/phy_n.c
179
if (e)
drivers/net/wireless/broadcom/b43/phy_n.c
180
val_addr = (i == 0) ? e->val_addr_core0 : e->val_addr_core1;
drivers/net/wireless/broadcom/b43/phy_n.c
184
if (e) /* Do it safer, better than wl */
drivers/net/wireless/broadcom/b43/phy_n.c
185
b43_phy_mask(dev, val_addr, ~e->val_mask);
drivers/net/wireless/broadcom/b43/phy_n.c
189
if (e)
drivers/net/wireless/broadcom/b43/phy_n.c
190
b43_phy_maskset(dev, val_addr, ~e->val_mask, (value << e->val_shift));
drivers/net/wireless/broadcom/b43/phy_n.c
2492
struct nphy_gain_ctl_workaround_entry *e;
drivers/net/wireless/broadcom/b43/phy_n.c
2501
e = b43_nphy_get_gain_ctl_workaround_ent(dev, ghz5, ext_lna);
drivers/net/wireless/broadcom/b43/phy_n.c
2532
b43_ntab_write_bulk(dev, B43_NTAB8(0, 8), 4, e->lna1_gain);
drivers/net/wireless/broadcom/b43/phy_n.c
2533
b43_ntab_write_bulk(dev, B43_NTAB8(1, 8), 4, e->lna1_gain);
drivers/net/wireless/broadcom/b43/phy_n.c
2534
b43_ntab_write_bulk(dev, B43_NTAB8(0, 16), 4, e->lna2_gain);
drivers/net/wireless/broadcom/b43/phy_n.c
2535
b43_ntab_write_bulk(dev, B43_NTAB8(1, 16), 4, e->lna2_gain);
drivers/net/wireless/broadcom/b43/phy_n.c
2536
b43_ntab_write_bulk(dev, B43_NTAB8(0, 32), 10, e->gain_db);
drivers/net/wireless/broadcom/b43/phy_n.c
2537
b43_ntab_write_bulk(dev, B43_NTAB8(1, 32), 10, e->gain_db);
drivers/net/wireless/broadcom/b43/phy_n.c
2538
b43_ntab_write_bulk(dev, B43_NTAB8(2, 32), 10, e->gain_bits);
drivers/net/wireless/broadcom/b43/phy_n.c
2539
b43_ntab_write_bulk(dev, B43_NTAB8(3, 32), 10, e->gain_bits);
drivers/net/wireless/broadcom/b43/phy_n.c
2545
b43_phy_write(dev, B43_NPHY_REV3_C1_INITGAIN_A, e->init_gain);
drivers/net/wireless/broadcom/b43/phy_n.c
2546
b43_phy_write(dev, B43_NPHY_REV3_C2_INITGAIN_A, e->init_gain);
drivers/net/wireless/broadcom/b43/phy_n.c
2549
e->rfseq_init);
drivers/net/wireless/broadcom/b43/phy_n.c
2551
b43_phy_write(dev, B43_NPHY_REV3_C1_CLIP_HIGAIN_A, e->cliphi_gain);
drivers/net/wireless/broadcom/b43/phy_n.c
2552
b43_phy_write(dev, B43_NPHY_REV3_C2_CLIP_HIGAIN_A, e->cliphi_gain);
drivers/net/wireless/broadcom/b43/phy_n.c
2553
b43_phy_write(dev, B43_NPHY_REV3_C1_CLIP_MEDGAIN_A, e->clipmd_gain);
drivers/net/wireless/broadcom/b43/phy_n.c
2554
b43_phy_write(dev, B43_NPHY_REV3_C2_CLIP_MEDGAIN_A, e->clipmd_gain);
drivers/net/wireless/broadcom/b43/phy_n.c
2555
b43_phy_write(dev, B43_NPHY_REV3_C1_CLIP_LOGAIN_A, e->cliplo_gain);
drivers/net/wireless/broadcom/b43/phy_n.c
2556
b43_phy_write(dev, B43_NPHY_REV3_C2_CLIP_LOGAIN_A, e->cliplo_gain);
drivers/net/wireless/broadcom/b43/phy_n.c
2558
b43_phy_maskset(dev, B43_NPHY_CRSMINPOWER0, 0xFF00, e->crsmin);
drivers/net/wireless/broadcom/b43/phy_n.c
2559
b43_phy_maskset(dev, B43_NPHY_CRSMINPOWERL0, 0xFF00, e->crsminl);
drivers/net/wireless/broadcom/b43/phy_n.c
2560
b43_phy_maskset(dev, B43_NPHY_CRSMINPOWERU0, 0xFF00, e->crsminu);
drivers/net/wireless/broadcom/b43/phy_n.c
2561
b43_phy_write(dev, B43_NPHY_C1_NBCLIPTHRES, e->nbclip);
drivers/net/wireless/broadcom/b43/phy_n.c
2562
b43_phy_write(dev, B43_NPHY_C2_NBCLIPTHRES, e->nbclip);
drivers/net/wireless/broadcom/b43/phy_n.c
2564
~B43_NPHY_C1_CLIPWBTHRES_CLIP2, e->wlclip);
drivers/net/wireless/broadcom/b43/phy_n.c
2566
~B43_NPHY_C2_CLIPWBTHRES_CLIP2, e->wlclip);
drivers/net/wireless/broadcom/b43/phy_n.c
6193
const struct b43_phy_n_sfo_cfg *e)
drivers/net/wireless/broadcom/b43/phy_n.c
6195
b43_phy_write(dev, B43_NPHY_BW1A, e->phy_bw1a);
drivers/net/wireless/broadcom/b43/phy_n.c
6196
b43_phy_write(dev, B43_NPHY_BW2, e->phy_bw2);
drivers/net/wireless/broadcom/b43/phy_n.c
6197
b43_phy_write(dev, B43_NPHY_BW3, e->phy_bw3);
drivers/net/wireless/broadcom/b43/phy_n.c
6198
b43_phy_write(dev, B43_NPHY_BW4, e->phy_bw4);
drivers/net/wireless/broadcom/b43/phy_n.c
6199
b43_phy_write(dev, B43_NPHY_BW5, e->phy_bw5);
drivers/net/wireless/broadcom/b43/phy_n.c
6200
b43_phy_write(dev, B43_NPHY_BW6, e->phy_bw6);
drivers/net/wireless/broadcom/b43/phy_n.c
6224
const struct b43_phy_n_sfo_cfg *e,
drivers/net/wireless/broadcom/b43/phy_n.c
6253
b43_chantab_phy_upload(dev, e);
drivers/net/wireless/broadcom/b43/radio_2055.c
1289
const struct b2055_inittab_entry *e;
drivers/net/wireless/broadcom/b43/radio_2055.c
1294
e = &(b2055_inittab[i]);
drivers/net/wireless/broadcom/b43/radio_2055.c
1295
if (!(e->flags & B2055_INITTAB_ENTRY_OK))
drivers/net/wireless/broadcom/b43/radio_2055.c
1297
if ((e->flags & B2055_INITTAB_UPLOAD) || ignore_uploadflag) {
drivers/net/wireless/broadcom/b43/radio_2055.c
1299
value = e->ghz5;
drivers/net/wireless/broadcom/b43/radio_2055.c
1301
value = e->ghz2;
drivers/net/wireless/broadcom/b43/radio_2055.c
1312
const struct b43_nphy_channeltab_entry_rev2 *e;
drivers/net/wireless/broadcom/b43/radio_2055.c
1316
e = &(b43_nphy_channeltab_rev2[i]);
drivers/net/wireless/broadcom/b43/radio_2055.c
1317
if (e->channel == channel)
drivers/net/wireless/broadcom/b43/radio_2055.c
1318
return e;
drivers/net/wireless/broadcom/b43/radio_2056.c
10196
const struct b2056_inittab_entry *e,
drivers/net/wireless/broadcom/b43/radio_2056.c
10202
for (i = 0; i < length; i++, e++) {
drivers/net/wireless/broadcom/b43/radio_2056.c
10203
if (!(e->flags & B2056_INITTAB_ENTRY_OK))
drivers/net/wireless/broadcom/b43/radio_2056.c
10205
if ((e->flags & B2056_INITTAB_UPLOAD) || ignore_uploadflag) {
drivers/net/wireless/broadcom/b43/radio_2056.c
10207
value = e->ghz5;
drivers/net/wireless/broadcom/b43/radio_2056.c
10209
value = e->ghz2;
drivers/net/wireless/broadcom/b43/radio_2056.c
10241
const struct b2056_inittab_entry *e;
drivers/net/wireless/broadcom/b43/radio_2056.c
10249
e = &pts->syn[B2056_SYN_PLL_CP2];
drivers/net/wireless/broadcom/b43/radio_2056.c
10251
b43_radio_write(dev, B2056_SYN_PLL_CP2, ghz5 ? e->ghz5 : e->ghz2);
drivers/net/wireless/broadcom/b43/radio_2056.c
10258
const struct b43_nphy_channeltab_entry_rev3 *e;
drivers/net/wireless/broadcom/b43/radio_2056.c
10263
e = b43_nphy_channeltab_phy_rev3;
drivers/net/wireless/broadcom/b43/radio_2056.c
10267
e = b43_nphy_channeltab_phy_rev4;
drivers/net/wireless/broadcom/b43/radio_2056.c
10273
e = b43_nphy_channeltab_radio_rev5;
drivers/net/wireless/broadcom/b43/radio_2056.c
10277
e = b43_nphy_channeltab_radio_rev6;
drivers/net/wireless/broadcom/b43/radio_2056.c
10282
e = b43_nphy_channeltab_radio_rev7_9;
drivers/net/wireless/broadcom/b43/radio_2056.c
10286
e = b43_nphy_channeltab_radio_rev8;
drivers/net/wireless/broadcom/b43/radio_2056.c
10290
e = b43_nphy_channeltab_radio_rev11;
drivers/net/wireless/broadcom/b43/radio_2056.c
10299
for (i = 0; i < length; i++, e++) {
drivers/net/wireless/broadcom/b43/radio_2056.c
10300
if (e->freq == freq)
drivers/net/wireless/broadcom/b43/radio_2056.c
10301
return e;
drivers/net/wireless/broadcom/b43/radio_2059.c
341
const struct b43_phy_ht_channeltab_e_radio2059 *e;
drivers/net/wireless/broadcom/b43/radio_2059.c
344
e = b43_phy_ht_channeltab_radio2059;
drivers/net/wireless/broadcom/b43/radio_2059.c
345
for (i = 0; i < ARRAY_SIZE(b43_phy_ht_channeltab_radio2059); i++, e++) {
drivers/net/wireless/broadcom/b43/radio_2059.c
346
if (e->freq == freq)
drivers/net/wireless/broadcom/b43/radio_2059.c
347
return e;
drivers/net/wireless/broadcom/b43/tables_lpphy.c
545
const struct b206x_init_tab_entry *e;
drivers/net/wireless/broadcom/b43/tables_lpphy.c
549
e = &b2062_init_tab[i];
drivers/net/wireless/broadcom/b43/tables_lpphy.c
551
if (!(e->flags & B206X_FLAG_G))
drivers/net/wireless/broadcom/b43/tables_lpphy.c
553
b43_radio_write(dev, e->offset, e->value_g);
drivers/net/wireless/broadcom/b43/tables_lpphy.c
555
if (!(e->flags & B206X_FLAG_A))
drivers/net/wireless/broadcom/b43/tables_lpphy.c
557
b43_radio_write(dev, e->offset, e->value_a);
drivers/net/wireless/broadcom/b43/tables_lpphy.c
564
const struct b206x_init_tab_entry *e;
drivers/net/wireless/broadcom/b43/tables_lpphy.c
568
e = &b2063_init_tab[i];
drivers/net/wireless/broadcom/b43/tables_lpphy.c
570
if (!(e->flags & B206X_FLAG_G))
drivers/net/wireless/broadcom/b43/tables_lpphy.c
572
b43_radio_write(dev, e->offset, e->value_g);
drivers/net/wireless/broadcom/b43/tables_lpphy.c
574
if (!(e->flags & B206X_FLAG_A))
drivers/net/wireless/broadcom/b43/tables_lpphy.c
576
b43_radio_write(dev, e->offset, e->value_a);
drivers/net/wireless/broadcom/b43/tables_nphy.c
3769
struct nphy_gain_ctl_workaround_entry *e;
drivers/net/wireless/broadcom/b43/tables_nphy.c
3784
e = &nphy_gain_ctl_workaround[ghz5][phy_idx];
drivers/net/wireless/broadcom/b43/tables_nphy.c
3798
e->cliplo_gain = gain_data[tr_iso];
drivers/net/wireless/broadcom/b43/tables_nphy.c
3804
e->cliplo_gain = gain_data[tr_iso];
drivers/net/wireless/broadcom/b43/tables_nphy.c
3808
e->rfseq_init[0] &= ~0x4000;
drivers/net/wireless/broadcom/b43/tables_nphy.c
3809
e->rfseq_init[1] &= ~0x4000;
drivers/net/wireless/broadcom/b43/tables_nphy.c
3810
e->rfseq_init[2] &= ~0x4000;
drivers/net/wireless/broadcom/b43/tables_nphy.c
3811
e->rfseq_init[3] &= ~0x4000;
drivers/net/wireless/broadcom/b43/tables_nphy.c
3812
e->init_gain &= ~0x4000;
drivers/net/wireless/broadcom/b43/tables_nphy.c
3817
e->crsminu = 0x2d;
drivers/net/wireless/broadcom/b43/tables_nphy.c
3819
e->rfseq_init[0] &= ~0x4000;
drivers/net/wireless/broadcom/b43/tables_nphy.c
3820
e->rfseq_init[1] &= ~0x4000;
drivers/net/wireless/broadcom/b43/tables_nphy.c
3821
e->rfseq_init[2] &= ~0x4000;
drivers/net/wireless/broadcom/b43/tables_nphy.c
3822
e->rfseq_init[3] &= ~0x4000;
drivers/net/wireless/broadcom/b43/tables_nphy.c
3823
e->init_gain &= ~0x4000;
drivers/net/wireless/broadcom/b43/tables_nphy.c
3824
e->rfseq_init[0] |= 0x1000;
drivers/net/wireless/broadcom/b43/tables_nphy.c
3825
e->rfseq_init[1] |= 0x1000;
drivers/net/wireless/broadcom/b43/tables_nphy.c
3826
e->rfseq_init[2] |= 0x1000;
drivers/net/wireless/broadcom/b43/tables_nphy.c
3827
e->rfseq_init[3] |= 0x1000;
drivers/net/wireless/broadcom/b43/tables_nphy.c
3828
e->init_gain |= 0x1000;
drivers/net/wireless/broadcom/b43/tables_nphy.c
3832
return e;
drivers/net/wireless/broadcom/b43/tables_nphy.c
3838
const struct nphy_rf_control_override_rev7 *e;
drivers/net/wireless/broadcom/b43/tables_nphy.c
3843
e = tbl_rf_control_override_rev7_over0;
drivers/net/wireless/broadcom/b43/tables_nphy.c
3847
e = tbl_rf_control_override_rev7_over1;
drivers/net/wireless/broadcom/b43/tables_nphy.c
3851
e = tbl_rf_control_override_rev7_over2;
drivers/net/wireless/broadcom/b43/tables_nphy.c
3860
if (e[i].field == field)
drivers/net/wireless/broadcom/b43/tables_nphy.c
3861
return &e[i];
drivers/net/wireless/broadcom/b43/xmit.c
170
const struct b43_tx_legacy_rate_phy_ctl_entry *e;
drivers/net/wireless/broadcom/b43/xmit.c
184
e = b43_tx_legacy_rate_phy_ctl_ent(bitrate);
drivers/net/wireless/broadcom/b43/xmit.c
185
if (e) {
drivers/net/wireless/broadcom/b43/xmit.c
186
control |= e->coding_rate;
drivers/net/wireless/broadcom/b43/xmit.c
187
control |= e->modulation;
drivers/net/wireless/broadcom/b43/xmit.c
40
const struct b43_tx_legacy_rate_phy_ctl_entry *e;
drivers/net/wireless/broadcom/b43/xmit.c
44
e = &(b43_tx_legacy_rate_phy_ctl[i]);
drivers/net/wireless/broadcom/b43/xmit.c
45
if (e->bitrate == bitrate)
drivers/net/wireless/broadcom/b43/xmit.c
46
return e;
drivers/net/wireless/broadcom/b43legacy/debugfs.c
337
struct b43legacy_dfsentry *e = dev->dfsentry;
drivers/net/wireless/broadcom/b43legacy/debugfs.c
340
e->dyn_debug[id] = (initstate); \
drivers/net/wireless/broadcom/b43legacy/debugfs.c
341
debugfs_create_bool(name, 0600, e->subdir, \
drivers/net/wireless/broadcom/b43legacy/debugfs.c
342
&(e->dyn_debug[id])); \
drivers/net/wireless/broadcom/b43legacy/debugfs.c
356
struct b43legacy_dfsentry *e;
drivers/net/wireless/broadcom/b43legacy/debugfs.c
361
e = kzalloc_obj(*e);
drivers/net/wireless/broadcom/b43legacy/debugfs.c
362
if (!e) {
drivers/net/wireless/broadcom/b43legacy/debugfs.c
366
e->dev = dev;
drivers/net/wireless/broadcom/b43legacy/debugfs.c
367
log = &e->txstatlog;
drivers/net/wireless/broadcom/b43legacy/debugfs.c
372
kfree(e);
drivers/net/wireless/broadcom/b43legacy/debugfs.c
378
dev->dfsentry = e;
drivers/net/wireless/broadcom/b43legacy/debugfs.c
381
e->subdir = debugfs_create_dir(devdir, rootdir);
drivers/net/wireless/broadcom/b43legacy/debugfs.c
386
e->subdir, dev, \
drivers/net/wireless/broadcom/b43legacy/debugfs.c
404
struct b43legacy_dfsentry *e;
drivers/net/wireless/broadcom/b43legacy/debugfs.c
408
e = dev->dfsentry;
drivers/net/wireless/broadcom/b43legacy/debugfs.c
409
if (!e)
drivers/net/wireless/broadcom/b43legacy/debugfs.c
412
debugfs_remove(e->subdir);
drivers/net/wireless/broadcom/b43legacy/debugfs.c
413
kfree(e->txstatlog.log);
drivers/net/wireless/broadcom/b43legacy/debugfs.c
414
kfree(e);
drivers/net/wireless/broadcom/b43legacy/debugfs.c
420
struct b43legacy_dfsentry *e = dev->dfsentry;
drivers/net/wireless/broadcom/b43legacy/debugfs.c
425
if (!e)
drivers/net/wireless/broadcom/b43legacy/debugfs.c
427
log = &e->txstatlog;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
1691
static u16 brcmf_map_fw_linkdown_reason(const struct brcmf_event_msg *e)
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
1695
switch (e->event_code) {
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
1699
reason = e->reason;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
3627
const struct brcmf_event_msg *e, void *data)
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
3641
status = e->status;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
3654
if (e->datalen < sizeof(*escan_result_le)) {
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
3665
escan_buflen > e->datalen ||
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
3864
const struct brcmf_event_msg *e, void *data)
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
3880
if (e->datalen < (sizeof(*pfn_result) + sizeof(*netinfo))) {
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
3885
if (e->event_code == BRCMF_E_PFN_NET_LOST) {
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
3905
datalen = e->datalen - ((void *)netinfo_start - (void *)pfn_result);
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
4038
brcmf_wowl_nd_results(struct brcmf_if *ifp, const struct brcmf_event_msg *e,
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
4048
if (e->datalen < (sizeof(*pfn_result) + sizeof(*netinfo))) {
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
4055
if (e->event_code == BRCMF_E_PFN_NET_LOST) {
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
5809
const struct brcmf_event_msg *e, void *data)
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
5811
switch (e->reason) {
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
5817
brcmf_proto_add_tdls_peer(ifp->drvr, ifp->ifidx, (u8 *)e->addr);
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
5821
brcmf_proto_delete_peer(ifp->drvr, ifp->ifidx, (u8 *)e->addr);
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
6098
const struct brcmf_event_msg *e)
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
6100
u32 event = e->event_code;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
6101
u32 status = e->status;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
6110
memcpy(vif->profile.bssid, e->addr, ETH_ALEN);
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
6128
const struct brcmf_event_msg *e)
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
6130
u32 event = e->event_code;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
6131
u16 flags = e->flags;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
6145
const struct brcmf_event_msg *e)
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
6147
u32 event = e->event_code;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
6148
u32 status = e->status;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
6152
e->flags & BRCMF_EVENT_MSG_LINK ? "up" : "down");
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
6405
const struct brcmf_event_msg *e)
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
6423
memcpy(profile->bssid, e->addr, ETH_ALEN);
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
6477
struct net_device *ndev, const struct brcmf_event_msg *e,
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
6519
const struct brcmf_event_msg *e, void *data)
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
6523
u32 event = e->event_code;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
6524
u32 reason = e->reason;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
6548
sinfo->assoc_req_ies_len = e->datalen;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
6551
cfg80211_new_sta(ndev, e->addr, sinfo, GFP_KERNEL);
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
6557
cfg80211_del_sta(ndev, e->addr, GFP_KERNEL);
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
6564
const struct brcmf_event_msg *e, void *data)
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
6572
if ((e->event_code == BRCMF_E_DEAUTH) ||
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
6573
(e->event_code == BRCMF_E_DEAUTH_IND) ||
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
6574
(e->event_code == BRCMF_E_DISASSOC_IND) ||
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
6575
((e->event_code == BRCMF_E_LINK) && (!e->flags))) {
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
6576
brcmf_proto_delete_peer(ifp->drvr, ifp->ifidx, (u8 *)e->addr);
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
6580
err = brcmf_notify_connect_status_ap(cfg, ndev, e, data);
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
6581
} else if (brcmf_is_linkup(ifp->vif, e)) {
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
6584
brcmf_inform_ibss(cfg, ndev, e->addr);
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
6586
memcpy(profile->bssid, e->addr, ETH_ALEN);
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
6587
cfg80211_ibss_joined(ndev, e->addr, chan, GFP_KERNEL);
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
6593
brcmf_bss_connect_done(cfg, ndev, e, true);
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
6595
} else if (brcmf_is_linkdown(ifp->vif, e)) {
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
6604
memcmp(profile->bssid, e->addr, ETH_ALEN))
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
6607
brcmf_bss_connect_done(cfg, ndev, e, false);
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
6609
brcmf_map_fw_linkdown_reason(e),
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
6610
e->event_code &
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
6619
} else if (brcmf_is_nonetwork(cfg, e)) {
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
6624
brcmf_bss_connect_done(cfg, ndev, e, false);
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
6632
const struct brcmf_event_msg *e, void *data)
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
6635
u32 event = e->event_code;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
6636
u32 status = e->status;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
6641
brcmf_bss_roaming_done(cfg, ifp->ndev, e);
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
6643
brcmf_bss_connect_done(cfg, ifp->ndev, e, true);
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
6653
const struct brcmf_event_msg *e, void *data)
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
6655
u16 flags = e->flags;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
6663
cfg80211_michael_mic_failure(ifp->ndev, (u8 *)&e->addr, key_type, -1,
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
6670
const struct brcmf_event_msg *e, void *data)
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
6677
if (e->datalen >= sizeof(*info)) {
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
6681
} else if (e->datalen >= sizeof(rssi)) {
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
6713
const struct brcmf_event_msg *e, void *data)
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cyw/core.c
240
const struct brcmf_event_msg *e, void *data)
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cyw/core.c
249
brcmf_fweh_event_name(e->event_code), e->event_code);
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cyw/core.c
251
if (e->datalen < sizeof(*auth_req)) {
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cyw/core.c
253
brcmf_fweh_event_name(e->event_code), e->event_code);
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cyw/core.c
275
const struct brcmf_event_msg *e, void *data)
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cyw/core.c
280
u32 mgmt_frame_len = e->datalen - sizeof(struct brcmf_rx_mgmt_data);
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cyw/core.c
288
brcmf_fweh_event_name(e->event_code), e->event_code);
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cyw/core.c
290
if (e->datalen < sizeof(*rxframe)) {
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cyw/core.c
292
brcmf_fweh_event_name(e->event_code), e->event_code);
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cyw/core.c
308
memcpy(mgmt_frame->sa, e->addr, ETH_ALEN);
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cyw/core.c
328
const struct brcmf_event_msg *e, void *data)
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cyw/core.c
334
brcmf_fweh_event_name(e->event_code), e->event_code,
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cyw/core.c
335
e->status);
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cyw/core.c
341
if (e->event_code == BRCMF_E_MGMT_FRAME_TXSTATUS) {
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cyw/core.c
342
if (e->status == BRCMF_E_STATUS_SUCCESS)
drivers/net/wireless/broadcom/brcm80211/brcmfmac/feature.c
109
const struct brcmf_feat_fwfeat *e;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/feature.c
114
e = &brcmf_feat_fwfeat_map[i];
drivers/net/wireless/broadcom/brcm80211/brcmfmac/feature.c
115
if (!strcmp(e->fwid, drv->fwver)) {
drivers/net/wireless/broadcom/brcm80211/brcmfmac/feature.c
116
feat_flags = e->feat_flags;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/feature.c
145
const struct brcmf_feat_wlcfeat *e;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/feature.c
160
e = &brcmf_feat_wlcfeat_map[i];
drivers/net/wireless/broadcom/brcm80211/brcmfmac/feature.c
161
if (major > e->min_ver_major ||
drivers/net/wireless/broadcom/brcm80211/brcmfmac/feature.c
162
(major == e->min_ver_major &&
drivers/net/wireless/broadcom/brcm80211/brcmfmac/feature.c
163
minor >= e->min_ver_minor)) {
drivers/net/wireless/broadcom/brcm80211/brcmfmac/feature.c
164
feat_flags |= e->feat_flags;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c
1596
const struct brcmf_event_msg *e,
drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c
1604
if (e->datalen < BRCMF_FWS_FIFO_COUNT) {
drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c
1605
bphy_err(drvr, "event payload too small (%d)\n", e->datalen);
drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c
1630
const struct brcmf_event_msg *e,
drivers/net/wireless/broadcom/brcm80211/brcmfmac/p2p.c
1012
const struct brcmf_event_msg *e,
drivers/net/wireless/broadcom/brcm80211/brcmfmac/p2p.c
1386
const struct brcmf_event_msg *e,
drivers/net/wireless/broadcom/brcm80211/brcmfmac/p2p.c
1394
u32 mgmt_frame_len = e->datalen - sizeof(struct brcmf_rx_mgmt_data);
drivers/net/wireless/broadcom/brcm80211/brcmfmac/p2p.c
1405
if (e->datalen < sizeof(*rxframe)) {
drivers/net/wireless/broadcom/brcm80211/brcmfmac/p2p.c
1425
(brcmf_p2p_gon_req_collision(p2p, (u8 *)e->addr))) {
drivers/net/wireless/broadcom/brcm80211/brcmfmac/p2p.c
1428
(ether_addr_equal(afx_hdl->tx_dst_addr, e->addr))) {
drivers/net/wireless/broadcom/brcm80211/brcmfmac/p2p.c
1467
memcpy(mgmt_frame->sa, e->addr, ETH_ALEN);
drivers/net/wireless/broadcom/brcm80211/brcmfmac/p2p.c
1494
const struct brcmf_event_msg *e,
drivers/net/wireless/broadcom/brcm80211/brcmfmac/p2p.c
1501
e->event_code == BRCMF_E_ACTION_FRAME_OFF_CHAN_COMPLETE ?
drivers/net/wireless/broadcom/brcm80211/brcmfmac/p2p.c
1503
e->status);
drivers/net/wireless/broadcom/brcm80211/brcmfmac/p2p.c
1508
if (e->event_code == BRCMF_E_ACTION_FRAME_COMPLETE) {
drivers/net/wireless/broadcom/brcm80211/brcmfmac/p2p.c
1509
if (e->status == BRCMF_E_STATUS_SUCCESS) {
drivers/net/wireless/broadcom/brcm80211/brcmfmac/p2p.c
1925
const struct brcmf_event_msg *e,
drivers/net/wireless/broadcom/brcm80211/brcmfmac/p2p.c
1939
brcmf_dbg(INFO, "Enter: event %d reason %d\n", e->event_code,
drivers/net/wireless/broadcom/brcm80211/brcmfmac/p2p.c
1940
e->reason);
drivers/net/wireless/broadcom/brcm80211/brcmfmac/p2p.c
1942
if (e->datalen < sizeof(*rxframe)) {
drivers/net/wireless/broadcom/brcm80211/brcmfmac/p2p.c
1951
(ether_addr_equal(afx_hdl->tx_dst_addr, e->addr))) {
drivers/net/wireless/broadcom/brcm80211/brcmfmac/p2p.c
1960
if (e->bsscfgidx == 0)
drivers/net/wireless/broadcom/brcm80211/brcmfmac/p2p.c
1976
mgmt_frame_len = e->datalen - sizeof(*rxframe);
drivers/net/wireless/broadcom/brcm80211/brcmfmac/p2p.c
1985
mgmt_frame_len, e->datalen, ch.chspec, freq);
drivers/net/wireless/broadcom/brcm80211/brcmfmac/p2p.h
162
const struct brcmf_event_msg *e,
drivers/net/wireless/broadcom/brcm80211/brcmfmac/p2p.h
166
const struct brcmf_event_msg *e,
drivers/net/wireless/broadcom/brcm80211/brcmfmac/p2p.h
169
const struct brcmf_event_msg *e,
drivers/net/wireless/broadcom/brcm80211/brcmfmac/p2p.h
176
const struct brcmf_event_msg *e,
drivers/net/wireless/intel/ipw2x00/ipw2100.c
2796
int e, i;
drivers/net/wireless/intel/ipw2x00/ipw2100.c
2812
e = txq->oldest;
drivers/net/wireless/intel/ipw2x00/ipw2100.c
2819
e = txq->oldest + frag_num;
drivers/net/wireless/intel/ipw2x00/ipw2100.c
2820
e %= txq->entries;
drivers/net/wireless/intel/ipw2x00/ipw2100.c
2863
if (!((r <= w && (e < r || e >= w)) || (e < r && e >= w))) {
drivers/net/wireless/intel/ipw2x00/ipw2100.c
2953
txq->oldest = (e + 1) % txq->entries;
drivers/net/wireless/intel/ipw2x00/ipw2100.c
6591
if (fwrq->e == 1) {
drivers/net/wireless/intel/ipw2x00/ipw2100.c
6601
fwrq->e = 0;
drivers/net/wireless/intel/ipw2x00/ipw2100.c
6606
if (fwrq->e > 0 || fwrq->m > 1000) {
drivers/net/wireless/intel/ipw2x00/ipw2100.c
6629
wrqu->freq.e = 0;
drivers/net/wireless/intel/ipw2x00/ipw2100.c
6822
range->freq[val].e = 1;
drivers/net/wireless/intel/ipw2x00/ipw2200.c
8619
if (fwrq->e == 1) {
drivers/net/wireless/intel/ipw2x00/ipw2200.c
8655
wrqu->freq.e = 0;
drivers/net/wireless/intel/ipw2x00/ipw2200.c
8666
wrqu->freq.e = 1;
drivers/net/wireless/intel/ipw2x00/ipw2200.c
8824
range->freq[i].e = 1;
drivers/net/wireless/intel/ipw2x00/ipw2200.c
8837
range->freq[i].e = 1;
drivers/net/wireless/intel/ipw2x00/libipw_module.c
255
struct proc_dir_entry *e;
drivers/net/wireless/intel/ipw2x00/libipw_module.c
264
e = proc_create("debug_level", 0644, libipw_proc,
drivers/net/wireless/intel/ipw2x00/libipw_module.c
266
if (!e) {
drivers/net/wireless/intel/ipw2x00/libipw_wx.c
91
iwe.u.freq.e = 6;
drivers/net/wireless/intel/iwlegacy/3945.h
290
s32 a, b, c, d, e; /* coefficients for voltage->power
drivers/net/wireless/intel/iwlegacy/common.c
701
__le16 *e;
drivers/net/wireless/intel/iwlegacy/common.c
714
e = (__le16 *) il->eeprom;
drivers/net/wireless/intel/iwlegacy/common.c
750
e[addr / 2] = cpu_to_le16(r >> 16);
drivers/net/wireless/intel/iwlwifi/dvm/eeprom.c
1003
e[addr / 2] = cpu_to_le16(r >> 16);
drivers/net/wireless/intel/iwlwifi/dvm/eeprom.c
1013
*eeprom = (u8 *)e;
drivers/net/wireless/intel/iwlwifi/dvm/eeprom.c
1019
kfree(e);
drivers/net/wireless/intel/iwlwifi/dvm/eeprom.c
920
__le16 *e;
drivers/net/wireless/intel/iwlwifi/dvm/eeprom.c
939
e = kmalloc(sz, GFP_KERNEL);
drivers/net/wireless/intel/iwlwifi/dvm/eeprom.c
940
if (!e)
drivers/net/wireless/intel/iwlwifi/dvm/eeprom.c
983
e[cache_addr / 2] = eeprom_data;
drivers/net/wireless/intel/iwlwifi/mvm/ftm-initiator.c
51
struct iwl_mvm_loc_entry *e, *t;
drivers/net/wireless/intel/iwlwifi/mvm/ftm-initiator.c
58
list_for_each_entry_safe(e, t, &mvm->ftm_initiator.loc_list, list) {
drivers/net/wireless/intel/iwlwifi/mvm/ftm-initiator.c
59
list_del(&e->list);
drivers/net/wireless/intel/iwlwifi/mvm/ftm-initiator.c
60
kfree(e);
drivers/net/wireless/mediatek/mt76/dma.c
379
struct mt76_queue_entry *e = &q->entry[idx];
drivers/net/wireless/mediatek/mt76/dma.c
381
if (!e->skip_buf0)
drivers/net/wireless/mediatek/mt76/dma.c
382
dma_unmap_single(dev->dma_dev, e->dma_addr[0], e->dma_len[0],
drivers/net/wireless/mediatek/mt76/dma.c
385
if (!e->skip_buf1)
drivers/net/wireless/mediatek/mt76/dma.c
386
dma_unmap_single(dev->dma_dev, e->dma_addr[1], e->dma_len[1],
drivers/net/wireless/mediatek/mt76/dma.c
389
if (e->txwi == DMA_DUMMY_DATA)
drivers/net/wireless/mediatek/mt76/dma.c
390
e->txwi = NULL;
drivers/net/wireless/mediatek/mt76/dma.c
392
*prev_e = *e;
drivers/net/wireless/mediatek/mt76/dma.c
393
memset(e, 0, sizeof(*e));
drivers/net/wireless/mediatek/mt76/dma.c
451
struct mt76_queue_entry *e = &q->entry[idx];
drivers/net/wireless/mediatek/mt76/dma.c
452
struct mt76_rro_rxdmad_c *dmad = e->buf;
drivers/net/wireless/mediatek/mt76/dma.c
494
struct mt76_queue_entry *e = &q->entry[idx];
drivers/net/wireless/mediatek/mt76/dma.c
497
void *buf = e->buf;
drivers/net/wireless/mediatek/mt76/dma.c
537
dma_sync_single_for_cpu(dev->dma_dev, e->dma_addr[0],
drivers/net/wireless/mediatek/mt76/dma.c
543
e->buf = NULL;
drivers/net/wireless/mediatek/mt76/mt76.h
1775
struct mt76_queue_entry *e);
drivers/net/wireless/mediatek/mt76/mt76.h
556
struct mt76_queue_entry *e);
drivers/net/wireless/mediatek/mt76/mt7603/mac.c
1292
void mt7603_tx_complete_skb(struct mt76_dev *mdev, struct mt76_queue_entry *e)
drivers/net/wireless/mediatek/mt76/mt7603/mac.c
1295
struct sk_buff *skb = e->skb;
drivers/net/wireless/mediatek/mt76/mt7603/mac.c
1297
if (!e->txwi) {
drivers/net/wireless/mediatek/mt76/mt7603/mac.c
1303
mt76_tx_complete_skb(mdev, e->wcid, skb);
drivers/net/wireless/mediatek/mt76/mt7603/mt7603.h
241
void mt7603_tx_complete_skb(struct mt76_dev *mdev, struct mt76_queue_entry *e);
drivers/net/wireless/mediatek/mt76/mt7615/mt7615.h
550
struct mt76_queue_entry *e);
drivers/net/wireless/mediatek/mt76/mt7615/usb_sdio.c
169
struct mt76_queue_entry *e)
drivers/net/wireless/mediatek/mt76/mt7615/usb_sdio.c
175
skb_pull(e->skb, headroom);
drivers/net/wireless/mediatek/mt76/mt7615/usb_sdio.c
177
mt76_tx_complete_skb(mdev, e->wcid, e->skb);
drivers/net/wireless/mediatek/mt76/mt76_connac.h
423
struct mt76_queue_entry *e);
drivers/net/wireless/mediatek/mt76/mt76_connac_mac.c
150
struct mt76_queue_entry *e)
drivers/net/wireless/mediatek/mt76/mt76_connac_mac.c
152
if (!e->txwi) {
drivers/net/wireless/mediatek/mt76/mt76_connac_mac.c
153
dev_kfree_skb_any(e->skb);
drivers/net/wireless/mediatek/mt76/mt76_connac_mac.c
157
if (e->skb)
drivers/net/wireless/mediatek/mt76/mt76_connac_mac.c
158
mt76_tx_complete_skb(mdev, e->wcid, e->skb);
drivers/net/wireless/mediatek/mt76/mt76x02_mac.c
906
void mt76x02_tx_complete_skb(struct mt76_dev *mdev, struct mt76_queue_entry *e)
drivers/net/wireless/mediatek/mt76/mt76x02_mac.c
912
if (!e->txwi) {
drivers/net/wireless/mediatek/mt76/mt76x02_mac.c
913
dev_kfree_skb_any(e->skb);
drivers/net/wireless/mediatek/mt76/mt76x02_mac.c
919
txwi_ptr = mt76_get_txwi_ptr(mdev, e->txwi);
drivers/net/wireless/mediatek/mt76/mt76x02_mac.c
923
mt76_tx_complete_skb(mdev, e->wcid, e->skb);
drivers/net/wireless/mediatek/mt76/mt76x02_mac.h
197
void mt76x02_tx_complete_skb(struct mt76_dev *mdev, struct mt76_queue_entry *e);
drivers/net/wireless/mediatek/mt76/mt76x02_usb.h
22
void mt76x02u_tx_complete_skb(struct mt76_dev *mdev, struct mt76_queue_entry *e);
drivers/net/wireless/mediatek/mt76/mt76x02_usb_core.c
18
void mt76x02u_tx_complete_skb(struct mt76_dev *mdev, struct mt76_queue_entry *e)
drivers/net/wireless/mediatek/mt76/mt76x02_usb_core.c
20
mt76x02u_remove_dma_hdr(e->skb);
drivers/net/wireless/mediatek/mt76/mt76x02_usb_core.c
21
mt76_tx_complete_skb(mdev, e->wcid, e->skb);
drivers/net/wireless/mediatek/mt76/mt7915/mcu.c
2550
struct edca *e = &req.edca[ac];
drivers/net/wireless/mediatek/mt76/mt7915/mcu.c
2552
e->set = WMM_PARAM_SET;
drivers/net/wireless/mediatek/mt76/mt7915/mcu.c
2553
e->queue = ac + mvif->mt76.wmm_idx * MT76_CONNAC_MAX_WMM_SETS;
drivers/net/wireless/mediatek/mt76/mt7915/mcu.c
2554
e->aifs = q->aifs;
drivers/net/wireless/mediatek/mt76/mt7915/mcu.c
2555
e->txop = cpu_to_le16(q->txop);
drivers/net/wireless/mediatek/mt76/mt7915/mcu.c
2558
e->cw_min = fls(q->cw_min);
drivers/net/wireless/mediatek/mt76/mt7915/mcu.c
2560
e->cw_min = 5;
drivers/net/wireless/mediatek/mt76/mt7915/mcu.c
2563
e->cw_max = cpu_to_le16(fls(q->cw_max));
drivers/net/wireless/mediatek/mt76/mt7915/mcu.c
2565
e->cw_max = cpu_to_le16(10);
drivers/net/wireless/mediatek/mt76/mt7915/testmode.c
176
struct edca *e = &req.edca[0];
drivers/net/wireless/mediatek/mt76/mt7915/testmode.c
178
e->queue = qid + mvif->mt76.wmm_idx * MT76_CONNAC_MAX_WMM_SETS;
drivers/net/wireless/mediatek/mt76/mt7915/testmode.c
179
e->set = WMM_PARAM_SET;
drivers/net/wireless/mediatek/mt76/mt7915/testmode.c
181
e->aifs = aifs;
drivers/net/wireless/mediatek/mt76/mt7915/testmode.c
182
e->cw_min = cw_min;
drivers/net/wireless/mediatek/mt76/mt7915/testmode.c
183
e->cw_max = cpu_to_le16(cw_max);
drivers/net/wireless/mediatek/mt76/mt7915/testmode.c
184
e->txop = cpu_to_le16(txop);
drivers/net/wireless/mediatek/mt76/mt7921/mac.c
812
struct mt76_queue_entry *e)
drivers/net/wireless/mediatek/mt76/mt7921/mac.c
814
__le32 *txwi = (__le32 *)(e->skb->data + MT_SDIO_HDR_SIZE);
drivers/net/wireless/mediatek/mt76/mt7921/mac.c
824
if (sta && likely(e->skb->protocol != cpu_to_be16(ETH_P_PAE)))
drivers/net/wireless/mediatek/mt76/mt7921/mac.c
827
skb_pull(e->skb, headroom);
drivers/net/wireless/mediatek/mt76/mt7921/mac.c
828
mt76_tx_complete_skb(mdev, e->wcid, e->skb);
drivers/net/wireless/mediatek/mt76/mt7921/mcu.c
722
struct edca *e = &req.edca[to_aci[ac]];
drivers/net/wireless/mediatek/mt76/mt7921/mcu.c
724
e->aifs = cpu_to_le16(q->aifs);
drivers/net/wireless/mediatek/mt76/mt7921/mcu.c
725
e->txop = cpu_to_le16(q->txop);
drivers/net/wireless/mediatek/mt76/mt7921/mcu.c
728
e->cw_min = cpu_to_le16(q->cw_min);
drivers/net/wireless/mediatek/mt76/mt7921/mcu.c
730
e->cw_min = cpu_to_le16(5);
drivers/net/wireless/mediatek/mt76/mt7921/mcu.c
733
e->cw_max = cpu_to_le16(q->cw_max);
drivers/net/wireless/mediatek/mt76/mt7921/mcu.c
735
e->cw_max = cpu_to_le16(10);
drivers/net/wireless/mediatek/mt76/mt7921/mcu.c
748
struct mu_edca *e;
drivers/net/wireless/mediatek/mt76/mt7921/mcu.c
754
e = &(req_mu.edca[to_aci[ac]]);
drivers/net/wireless/mediatek/mt76/mt7921/mcu.c
756
e->cw_min = q->ecw_min_max & 0xf;
drivers/net/wireless/mediatek/mt76/mt7921/mcu.c
757
e->cw_max = (q->ecw_min_max & 0xf0) >> 4;
drivers/net/wireless/mediatek/mt76/mt7921/mcu.c
758
e->aifsn = q->aifsn;
drivers/net/wireless/mediatek/mt76/mt7921/mcu.c
759
e->timer = q->mu_edca_timer;
drivers/net/wireless/mediatek/mt76/mt7921/mt7921.h
318
struct mt76_queue_entry *e);
drivers/net/wireless/mediatek/mt76/mt7925/mac.c
1442
struct mt76_queue_entry *e)
drivers/net/wireless/mediatek/mt76/mt7925/mac.c
1444
__le32 *txwi = (__le32 *)(e->skb->data + MT_SDIO_HDR_SIZE);
drivers/net/wireless/mediatek/mt76/mt7925/mac.c
1454
if (sta && likely(e->skb->protocol != cpu_to_be16(ETH_P_PAE)))
drivers/net/wireless/mediatek/mt76/mt7925/mac.c
1455
mt7925_tx_check_aggr(sta, e->skb, wcid);
drivers/net/wireless/mediatek/mt76/mt7925/mac.c
1457
skb_pull(e->skb, headroom);
drivers/net/wireless/mediatek/mt76/mt7925/mac.c
1458
mt76_tx_complete_skb(mdev, e->wcid, e->skb);
drivers/net/wireless/mediatek/mt76/mt7925/mcu.c
1164
struct edca *e;
drivers/net/wireless/mediatek/mt76/mt7925/mcu.c
1167
tlv = mt76_connac_mcu_add_tlv(skb, MCU_EDCA_AC_PARAM, sizeof(*e));
drivers/net/wireless/mediatek/mt76/mt7925/mcu.c
1169
e = (struct edca *)tlv;
drivers/net/wireless/mediatek/mt76/mt7925/mcu.c
1170
e->set = WMM_PARAM_SET;
drivers/net/wireless/mediatek/mt76/mt7925/mcu.c
1171
e->queue = ac;
drivers/net/wireless/mediatek/mt76/mt7925/mcu.c
1172
e->aifs = q->aifs;
drivers/net/wireless/mediatek/mt76/mt7925/mcu.c
1173
e->txop = cpu_to_le16(q->txop);
drivers/net/wireless/mediatek/mt76/mt7925/mcu.c
1176
e->cw_min = fls(q->cw_min);
drivers/net/wireless/mediatek/mt76/mt7925/mcu.c
1178
e->cw_min = 5;
drivers/net/wireless/mediatek/mt76/mt7925/mcu.c
1181
e->cw_max = fls(q->cw_max);
drivers/net/wireless/mediatek/mt76/mt7925/mcu.c
1183
e->cw_max = 10;
drivers/net/wireless/mediatek/mt76/mt7925/mt7925.h
339
struct mt76_queue_entry *e);
drivers/net/wireless/mediatek/mt76/mt7996/init.c
1077
struct mt7996_wed_rro_session_id *e;
drivers/net/wireless/mediatek/mt76/mt7996/init.c
1080
e = list_first_entry(&list, struct mt7996_wed_rro_session_id,
drivers/net/wireless/mediatek/mt76/mt7996/init.c
1082
list_del_init(&e->list);
drivers/net/wireless/mediatek/mt76/mt7996/init.c
1092
if (e->id == MT7996_RRO_MAX_SESSION)
drivers/net/wireless/mediatek/mt76/mt7996/init.c
1095
idx = e->id / MT7996_RRO_BA_BITMAP_SESSION_SIZE;
drivers/net/wireless/mediatek/mt76/mt7996/init.c
1101
(e->id % MT7996_RRO_BA_BITMAP_SESSION_SIZE) *
drivers/net/wireless/mediatek/mt76/mt7996/init.c
1108
mt7996_mcu_wed_rro_reset_sessions(dev, e->id);
drivers/net/wireless/mediatek/mt76/mt7996/init.c
1110
kfree(e);
drivers/net/wireless/mediatek/mt76/mt7996/mac.c
1909
struct mt7996_wed_rro_addr *e;
drivers/net/wireless/mediatek/mt76/mt7996/mac.c
1924
e = mt7996_rro_addr_elem_get(dev, seq_id, seq_num);
drivers/net/wireless/mediatek/mt76/mt7996/mac.c
1925
data = le32_to_cpu(e->data);
drivers/net/wireless/mediatek/mt76/mt7996/mac.c
1931
e->data |= cpu_to_le32(val);
drivers/net/wireless/mediatek/mt76/mt7996/mac.c
1939
dma_addr |= le32_to_cpu(e->head_low);
drivers/net/wireless/mediatek/mt76/mt7996/mac.c
2483
struct mt7996_wed_rro_session_id *e;
drivers/net/wireless/mediatek/mt76/mt7996/mac.c
2485
e = list_first_entry(&list, struct mt7996_wed_rro_session_id,
drivers/net/wireless/mediatek/mt76/mt7996/mac.c
2487
list_del_init(&e->list);
drivers/net/wireless/mediatek/mt76/mt7996/mac.c
2488
kfree(e);
drivers/net/wireless/mediatek/mt76/mt7996/mcu.c
3455
struct edca *e;
drivers/net/wireless/mediatek/mt76/mt7996/mcu.c
3458
tlv = mt7996_mcu_add_uni_tlv(skb, MCU_EDCA_AC_PARAM, sizeof(*e));
drivers/net/wireless/mediatek/mt76/mt7996/mcu.c
3460
e = (struct edca *)tlv;
drivers/net/wireless/mediatek/mt76/mt7996/mcu.c
3461
e->set = WMM_PARAM_SET;
drivers/net/wireless/mediatek/mt76/mt7996/mcu.c
3462
e->queue = ac;
drivers/net/wireless/mediatek/mt76/mt7996/mcu.c
3463
e->aifs = q->aifs;
drivers/net/wireless/mediatek/mt76/mt7996/mcu.c
3464
e->txop = cpu_to_le16(q->txop);
drivers/net/wireless/mediatek/mt76/mt7996/mcu.c
3467
e->cw_min = fls(q->cw_min);
drivers/net/wireless/mediatek/mt76/mt7996/mcu.c
3469
e->cw_min = 5;
drivers/net/wireless/mediatek/mt76/mt7996/mcu.c
3472
e->cw_max = fls(q->cw_max);
drivers/net/wireless/mediatek/mt76/mt7996/mcu.c
3474
e->cw_max = 10;
drivers/net/wireless/mediatek/mt76/mt7996/mcu.c
697
struct mt7996_mcu_wed_rro_ba_event *e;
drivers/net/wireless/mediatek/mt76/mt7996/mcu.c
699
while (skb->len >= sizeof(*e)) {
drivers/net/wireless/mediatek/mt76/mt7996/mcu.c
704
e = (void *)skb->data;
drivers/net/wireless/mediatek/mt76/mt7996/mcu.c
705
idx = le16_to_cpu(e->wlan_id);
drivers/net/wireless/mediatek/mt76/mt7996/mcu.c
710
if (e->tid >= ARRAY_SIZE(wcid->aggr))
drivers/net/wireless/mediatek/mt76/mt7996/mcu.c
713
tid = rcu_dereference(wcid->aggr[e->tid]);
drivers/net/wireless/mediatek/mt76/mt7996/mcu.c
717
tid->id = le16_to_cpu(e->id);
drivers/net/wireless/mediatek/mt76/mt7996/mcu.c
718
skb_pull(skb, sizeof(*e));
drivers/net/wireless/mediatek/mt76/mt7996/mcu.c
723
struct mt7996_mcu_wed_rro_ba_delete_event *e;
drivers/net/wireless/mediatek/mt76/mt7996/mcu.c
725
while (skb->len >= sizeof(*e)) {
drivers/net/wireless/mediatek/mt76/mt7996/mcu.c
728
e = (void *)skb->data;
drivers/net/wireless/mediatek/mt76/mt7996/mcu.c
733
session->id = le16_to_cpu(e->session_id);
drivers/net/wireless/mediatek/mt76/mt7996/mcu.c
740
skb_pull(skb, sizeof(*e));
drivers/net/wireless/mediatek/mt76/npu.c
102
e->buf - page_address(page),
drivers/net/wireless/mediatek/mt76/npu.c
23
struct mt76_queue_entry *e = &q->entry[q->head];
drivers/net/wireless/mediatek/mt76/npu.c
27
e->buf = mt76_get_page_pool_buf(q, &offset, q->buf_size);
drivers/net/wireless/mediatek/mt76/npu.c
28
if (!e->buf)
drivers/net/wireless/mediatek/mt76/npu.c
31
e->dma_len[0] = SKB_WITH_OVERHEAD(q->buf_size);
drivers/net/wireless/mediatek/mt76/npu.c
32
page = virt_to_head_page(e->buf);
drivers/net/wireless/mediatek/mt76/npu.c
33
e->dma_addr[0] = page_pool_get_dma_addr(page) + offset;
drivers/net/wireless/mediatek/mt76/npu.c
36
desc[q->head].addr = e->dma_addr[0];
drivers/net/wireless/mediatek/mt76/npu.c
50
struct mt76_queue_entry *e = &q->entry[q->tail];
drivers/net/wireless/mediatek/mt76/npu.c
52
dma_sync_single_for_cpu(dev->dma_dev, e->dma_addr[0],
drivers/net/wireless/mediatek/mt76/npu.c
53
e->dma_len[0],
drivers/net/wireless/mediatek/mt76/npu.c
55
mt76_put_page_pool_buf(e->buf, false);
drivers/net/wireless/mediatek/mt76/npu.c
74
struct mt76_queue_entry *e = &q->entry[index];
drivers/net/wireless/mediatek/mt76/npu.c
83
dma_sync_single_for_cpu(dev->dma_dev, e->dma_addr[0],
drivers/net/wireless/mediatek/mt76/npu.c
84
e->dma_len[0],
drivers/net/wireless/mediatek/mt76/npu.c
88
skb = napi_build_skb(e->buf, q->buf_size);
drivers/net/wireless/mediatek/mt76/npu.c
97
struct page *page = virt_to_head_page(e->buf);
drivers/net/wireless/mediatek/mt76/sdio.c
369
struct mt76_queue_entry *e = NULL;
drivers/net/wireless/mediatek/mt76/sdio.c
373
e = &q->entry[q->tail];
drivers/net/wireless/mediatek/mt76/sdio.c
379
return e;
drivers/net/wireless/mediatek/mt76/sdio.c
389
struct mt76_queue_entry *e;
drivers/net/wireless/mediatek/mt76/sdio.c
394
e = mt76s_get_next_rx_entry(q);
drivers/net/wireless/mediatek/mt76/sdio.c
395
if (!e || !e->skb)
drivers/net/wireless/mediatek/mt76/sdio.c
398
dev->drv->rx_skb(dev, MT_RXQ_MAIN, e->skb, NULL);
drivers/net/wireless/mediatek/mt76/sdio.c
399
e->skb = NULL;
drivers/net/wireless/mediatek/mt76/sdio.c
623
struct mt76_queue_entry *e = &q->entry[j];
drivers/net/wireless/mediatek/mt76/sdio.c
625
if (!e->skb)
drivers/net/wireless/mediatek/mt76/sdio.c
628
dev_kfree_skb(e->skb);
drivers/net/wireless/mediatek/mt76/sdio.c
629
e->skb = NULL;
drivers/net/wireless/mediatek/mt76/sdio_txrx.c
125
struct mt76_queue_entry *e = &q->entry[index];
drivers/net/wireless/mediatek/mt76/sdio_txrx.c
133
e->skb = mt76s_build_rx_skb(buf, len,
drivers/net/wireless/mediatek/mt76/sdio_txrx.c
135
if (!e->skb)
drivers/net/wireless/mediatek/mt76/sdio_txrx.c
253
struct mt76_queue_entry *e = &q->entry[q->first];
drivers/net/wireless/mediatek/mt76/sdio_txrx.c
262
__skb_put_zero(e->skb, 4);
drivers/net/wireless/mediatek/mt76/sdio_txrx.c
263
err = __skb_grow(e->skb, roundup(e->skb->len,
drivers/net/wireless/mediatek/mt76/sdio_txrx.c
267
err = __mt76s_xmit_queue(dev, e->skb->data,
drivers/net/wireless/mediatek/mt76/sdio_txrx.c
268
e->skb->len);
drivers/net/wireless/mediatek/mt76/sdio_txrx.c
275
pad = roundup(e->skb->len, 4) - e->skb->len;
drivers/net/wireless/mediatek/mt76/sdio_txrx.c
276
if (len + e->skb->len + pad + 4 > dev->sdio.xmit_buf_sz)
drivers/net/wireless/mediatek/mt76/sdio_txrx.c
279
if (mt76s_tx_pick_quota(sdio, mcu, e->buf_sz, &pse_sz,
drivers/net/wireless/mediatek/mt76/sdio_txrx.c
283
memcpy(sdio->xmit_buf + len, e->skb->data, skb_headlen(e->skb));
drivers/net/wireless/mediatek/mt76/sdio_txrx.c
284
len += skb_headlen(e->skb);
drivers/net/wireless/mediatek/mt76/sdio_txrx.c
287
skb_walk_frags(e->skb, iter) {
drivers/net/wireless/mediatek/mt76/sdio_txrx.c
299
e->done = true;
drivers/net/wireless/mediatek/mt76/tx.c
806
struct mt76_queue_entry *e)
drivers/net/wireless/mediatek/mt76/tx.c
808
if (e->skb)
drivers/net/wireless/mediatek/mt76/tx.c
809
dev->drv->tx_complete_skb(dev, e);
drivers/net/wireless/mediatek/mt76/usb.c
369
mt76u_urb_alloc(struct mt76_dev *dev, struct mt76_queue_entry *e,
drivers/net/wireless/mediatek/mt76/usb.c
377
e->urb = kzalloc(size, GFP_KERNEL);
drivers/net/wireless/mediatek/mt76/usb.c
378
if (!e->urb)
drivers/net/wireless/mediatek/mt76/usb.c
381
usb_init_urb(e->urb);
drivers/net/wireless/mediatek/mt76/usb.c
384
e->urb->sg = (struct scatterlist *)(e->urb + 1);
drivers/net/wireless/mediatek/mt76/usb.c
391
struct mt76_queue_entry *e)
drivers/net/wireless/mediatek/mt76/usb.c
397
err = mt76u_urb_alloc(dev, e, sg_size);
drivers/net/wireless/mediatek/mt76/usb.c
401
return mt76u_refill_rx(dev, q, e->urb, sg_size);
drivers/net/wireless/mediatek/mt76/usb.c
824
struct mt76_queue_entry *e = urb->context;
drivers/net/wireless/mediatek/mt76/usb.c
828
e->done = true;
drivers/net/wireless/mediatek/mt7601u/dma.c
12
struct mt7601u_dma_buf_rx *e, gfp_t gfp);
drivers/net/wireless/mediatek/mt7601u/dma.c
134
mt7601u_rx_process_entry(struct mt7601u_dev *dev, struct mt7601u_dma_buf_rx *e)
drivers/net/wireless/mediatek/mt7601u/dma.c
136
u32 seg_len, data_len = e->urb->actual_length;
drivers/net/wireless/mediatek/mt7601u/dma.c
137
u8 *data = page_address(e->p);
drivers/net/wireless/mediatek/mt7601u/dma.c
151
new_p ? e->p : NULL, &list);
drivers/net/wireless/mediatek/mt7601u/dma.c
165
put_page(e->p);
drivers/net/wireless/mediatek/mt7601u/dma.c
166
e->p = new_p;
drivers/net/wireless/mediatek/mt7601u/dma.c
182
buf = &q->e[q->start];
drivers/net/wireless/mediatek/mt7601u/dma.c
215
if (WARN_ONCE(q->e[q->end].urb != urb, "RX urb mismatch"))
drivers/net/wireless/mediatek/mt7601u/dma.c
228
struct mt7601u_dma_buf_rx *e;
drivers/net/wireless/mediatek/mt7601u/dma.c
230
while ((e = mt7601u_rx_get_pending_entry(dev))) {
drivers/net/wireless/mediatek/mt7601u/dma.c
231
if (e->urb->status)
drivers/net/wireless/mediatek/mt7601u/dma.c
234
mt7601u_rx_process_entry(dev, e);
drivers/net/wireless/mediatek/mt7601u/dma.c
235
mt7601u_submit_rx_buf(dev, e, GFP_ATOMIC);
drivers/net/wireless/mediatek/mt7601u/dma.c
261
if (WARN_ONCE(q->e[q->start].urb != urb, "TX urb mismatch"))
drivers/net/wireless/mediatek/mt7601u/dma.c
264
skb = q->e[q->start].skb;
drivers/net/wireless/mediatek/mt7601u/dma.c
265
q->e[q->start].skb = NULL;
drivers/net/wireless/mediatek/mt7601u/dma.c
311
struct mt7601u_dma_buf_tx *e;
drivers/net/wireless/mediatek/mt7601u/dma.c
323
e = &q->e[q->end];
drivers/net/wireless/mediatek/mt7601u/dma.c
324
usb_fill_bulk_urb(e->urb, usb_dev, snd_pipe, skb->data, skb->len,
drivers/net/wireless/mediatek/mt7601u/dma.c
326
ret = usb_submit_urb(e->urb, GFP_ATOMIC);
drivers/net/wireless/mediatek/mt7601u/dma.c
341
e->skb = skb;
drivers/net/wireless/mediatek/mt7601u/dma.c
395
usb_poison_urb(dev->rx_q.e[i].urb);
drivers/net/wireless/mediatek/mt7601u/dma.c
399
struct mt7601u_dma_buf_rx *e, gfp_t gfp)
drivers/net/wireless/mediatek/mt7601u/dma.c
402
u8 *buf = page_address(e->p);
drivers/net/wireless/mediatek/mt7601u/dma.c
408
usb_fill_bulk_urb(e->urb, usb_dev, pipe, buf, MT_RX_URB_SIZE,
drivers/net/wireless/mediatek/mt7601u/dma.c
411
trace_mt_submit_urb(dev, e->urb);
drivers/net/wireless/mediatek/mt7601u/dma.c
412
ret = usb_submit_urb(e->urb, gfp);
drivers/net/wireless/mediatek/mt7601u/dma.c
424
ret = mt7601u_submit_rx_buf(dev, &dev->rx_q.e[i], GFP_KERNEL);
drivers/net/wireless/mediatek/mt7601u/dma.c
437
__free_pages(dev->rx_q.e[i].p, MT_RX_ORDER);
drivers/net/wireless/mediatek/mt7601u/dma.c
438
usb_free_urb(dev->rx_q.e[i].urb);
drivers/net/wireless/mediatek/mt7601u/dma.c
451
dev->rx_q.e[i].urb = usb_alloc_urb(0, GFP_KERNEL);
drivers/net/wireless/mediatek/mt7601u/dma.c
452
dev->rx_q.e[i].p = dev_alloc_pages(MT_RX_ORDER);
drivers/net/wireless/mediatek/mt7601u/dma.c
454
if (!dev->rx_q.e[i].urb || !dev->rx_q.e[i].p)
drivers/net/wireless/mediatek/mt7601u/dma.c
466
usb_poison_urb(q->e[i].urb);
drivers/net/wireless/mediatek/mt7601u/dma.c
467
if (q->e[i].skb)
drivers/net/wireless/mediatek/mt7601u/dma.c
468
mt7601u_tx_status(q->dev, q->e[i].skb);
drivers/net/wireless/mediatek/mt7601u/dma.c
469
usb_free_urb(q->e[i].urb);
drivers/net/wireless/mediatek/mt7601u/dma.c
493
q->e[i].urb = usb_alloc_urb(0, GFP_KERNEL);
drivers/net/wireless/mediatek/mt7601u/dma.c
494
if (!q->e[i].urb)
drivers/net/wireless/mediatek/mt7601u/mt7601u.h
73
} e[N_RX_ENTRIES];
drivers/net/wireless/mediatek/mt7601u/mt7601u.h
89
} e[N_TX_ENTRIES];
drivers/net/wireless/microchip/wilc1000/cfg80211.c
939
struct wilc_attr_entry *e;
drivers/net/wireless/microchip/wilc1000/cfg80211.c
949
while (index + sizeof(*e) <= len) {
drivers/net/wireless/microchip/wilc1000/cfg80211.c
952
e = (struct wilc_attr_entry *)&buf[index];
drivers/net/wireless/microchip/wilc1000/cfg80211.c
953
attr_size = le16_to_cpu(e->attr_len);
drivers/net/wireless/microchip/wilc1000/cfg80211.c
955
if (index + sizeof(*e) + attr_size > len)
drivers/net/wireless/microchip/wilc1000/cfg80211.c
958
if (e->attr_type == IEEE80211_P2P_ATTR_CHANNEL_LIST &&
drivers/net/wireless/microchip/wilc1000/cfg80211.c
959
attr_size >= (sizeof(struct wilc_attr_ch_list) - sizeof(*e)))
drivers/net/wireless/microchip/wilc1000/cfg80211.c
961
else if (e->attr_type == IEEE80211_P2P_ATTR_OPER_CHANNEL &&
drivers/net/wireless/microchip/wilc1000/cfg80211.c
962
attr_size == (sizeof(struct wilc_attr_oper_ch) - sizeof(*e)))
drivers/net/wireless/microchip/wilc1000/cfg80211.c
968
index += sizeof(*e) + attr_size;
drivers/net/wireless/microchip/wilc1000/cfg80211.c
979
struct wilc_ch_list_elem *e;
drivers/net/wireless/microchip/wilc1000/cfg80211.c
981
e = (struct wilc_ch_list_elem *)(ch_list->elem + i);
drivers/net/wireless/microchip/wilc1000/cfg80211.c
983
i += sizeof(*e);
drivers/net/wireless/microchip/wilc1000/cfg80211.c
987
i += e->no_of_channels;
drivers/net/wireless/microchip/wilc1000/cfg80211.c
991
if (e->op_class == WILC_WLAN_OPERATING_CLASS_2_4GHZ) {
drivers/net/wireless/microchip/wilc1000/cfg80211.c
992
memset(e->ch_list, sta_ch, e->no_of_channels);
drivers/net/wireless/realtek/rtlwifi/debug.c
146
RTL_DEBUG_IMPL_BB_SERIES(e, 0x0e00);
drivers/net/wireless/realtek/rtw88/reg.h
342
#define BIT_RQPN_NE(n, e) (BIT_RQPN_NPQ(n) | BIT_RQPN_EPQ(e))
drivers/net/wireless/realtek/rtw88/rtw8723x.c
229
ether_addr_copy(efuse->addr, map->e.mac_addr);
drivers/net/wireless/realtek/rtw88/rtw8723x.c
79
rtw_dbg(rtwdev, RTW_DBG_EFUSE, "mac_addr=%pM\n", map->e.mac_addr);
drivers/net/wireless/realtek/rtw88/rtw8723x.c
80
DBG_EFUSE_2BYTE(rtwdev, map, e.vendor_id);
drivers/net/wireless/realtek/rtw88/rtw8723x.c
81
DBG_EFUSE_2BYTE(rtwdev, map, e.device_id);
drivers/net/wireless/realtek/rtw88/rtw8723x.c
82
DBG_EFUSE_2BYTE(rtwdev, map, e.sub_vendor_id);
drivers/net/wireless/realtek/rtw88/rtw8723x.c
83
DBG_EFUSE_2BYTE(rtwdev, map, e.sub_device_id);
drivers/net/wireless/realtek/rtw88/rtw8723x.h
95
struct rtw8723xe_efuse e;
drivers/net/wireless/realtek/rtw88/rtw8814a.c
171
ether_addr_copy(efuse->addr, map->e.mac_addr);
drivers/net/wireless/realtek/rtw88/rtw8814a.h
53
struct rtw8814ae_efuse e;
drivers/net/wireless/realtek/rtw88/rtw8821c.c
26
ether_addr_copy(efuse->addr, map->e.mac_addr);
drivers/net/wireless/realtek/rtw88/rtw8821c.h
101
struct rtw8821ce_efuse e;
drivers/net/wireless/realtek/rtw88/rtw8822b.c
26
ether_addr_copy(efuse->addr, map->e.mac_addr);
drivers/net/wireless/realtek/rtw88/rtw8822b.h
103
struct rtw8822be_efuse e;
drivers/net/wireless/realtek/rtw88/rtw8822c.c
29
ether_addr_copy(efuse->addr, map->e.mac_addr);
drivers/net/wireless/realtek/rtw88/rtw8822c.h
102
struct rtw8822ce_efuse e;
drivers/net/wireless/realtek/rtw89/coex.c
297
#define CASE_BTC_MLME_STATE(e) case MLME_##e: return #e
drivers/net/wireless/realtek/rtw89/coex.c
730
#define CASE_BTC_WL_LINK_MODE(e) case BTC_WLINK_## e: return #e
drivers/net/wireless/realtek/rtw89/coex.c
9250
#define CASE_BTC_RSN_STR(e) case BTC_RSN_ ## e: return #e
drivers/net/wireless/realtek/rtw89/coex.c
9251
#define CASE_BTC_ACT_STR(e) case BTC_ACT_ ## e | BTC_ACT_EXT_BIT: return #e
drivers/net/wireless/realtek/rtw89/coex.c
9252
#define CASE_BTC_POLICY_STR(e) \
drivers/net/wireless/realtek/rtw89/coex.c
9253
case BTC_CXP_ ## e | BTC_POLICY_EXT_BIT: return #e
drivers/net/wireless/realtek/rtw89/coex.c
9254
#define CASE_BTC_SLOT_STR(e) case CXST_ ## e: return #e
drivers/net/wireless/realtek/rtw89/coex.c
9255
#define CASE_BTC_EVT_STR(e) case CXEVNT_## e: return #e
drivers/net/wireless/realtek/rtw89/coex.c
9256
#define CASE_BTC_INIT(e) case BTC_MODE_## e: return #e
drivers/net/wireless/realtek/rtw89/coex.c
9257
#define CASE_BTC_ANTPATH_STR(e) case BTC_ANT_##e: return #e
drivers/net/wireless/realtek/rtw89/coex.c
9258
#define CASE_BTC_POLUT_STR(e) case BTC_PLT_## e: return #e
drivers/net/wireless/realtek/rtw89/coex.c
9259
#define CASE_BTC_REGTYPE_STR(e) case REG_## e: return #e
drivers/net/wireless/realtek/rtw89/coex.c
9260
#define CASE_BTC_GDBG_STR(e) case BTC_DBG_## e: return #e
drivers/net/wireless/realtek/rtw89/fw.c
10725
#define __fw_txpwr_entry_acceptable(e, cursor, ent_sz) \
drivers/net/wireless/realtek/rtw89/fw.c
10727
u8 __var_sz = sizeof(*(e)); \
drivers/net/wireless/realtek/rtw89/fw.c
10738
fw_txpwr_byrate_entry_valid(const struct rtw89_fw_txpwr_byrate_entry *e,
drivers/net/wireless/realtek/rtw89/fw.c
10742
if (!__fw_txpwr_entry_acceptable(e, cursor, conf->ent_sz))
drivers/net/wireless/realtek/rtw89/fw.c
10745
if (e->band >= RTW89_BAND_NUM || e->bw >= RTW89_BYR_BW_NUM)
drivers/net/wireless/realtek/rtw89/fw.c
10748
switch (e->rs) {
drivers/net/wireless/realtek/rtw89/fw.c
10750
if (e->shf + e->len > RTW89_RATE_CCK_NUM)
drivers/net/wireless/realtek/rtw89/fw.c
10754
if (e->shf + e->len > RTW89_RATE_OFDM_NUM)
drivers/net/wireless/realtek/rtw89/fw.c
10758
if (e->shf + e->len > __RTW89_RATE_MCS_NUM ||
drivers/net/wireless/realtek/rtw89/fw.c
10759
e->nss >= RTW89_NSS_NUM ||
drivers/net/wireless/realtek/rtw89/fw.c
10760
e->ofdma >= RTW89_OFDMA_NUM)
drivers/net/wireless/realtek/rtw89/fw.c
10764
if (e->shf + e->len > RTW89_RATE_HEDCM_NUM ||
drivers/net/wireless/realtek/rtw89/fw.c
10765
e->nss >= RTW89_NSS_HEDCM_NUM ||
drivers/net/wireless/realtek/rtw89/fw.c
10766
e->ofdma >= RTW89_OFDMA_NUM)
drivers/net/wireless/realtek/rtw89/fw.c
10770
if (e->shf + e->len > __RTW89_RATE_OFFSET_NUM)
drivers/net/wireless/realtek/rtw89/fw.c
10812
fw_txpwr_lmt_2ghz_entry_valid(const struct rtw89_fw_txpwr_lmt_2ghz_entry *e,
drivers/net/wireless/realtek/rtw89/fw.c
10816
if (!__fw_txpwr_entry_acceptable(e, cursor, conf->ent_sz))
drivers/net/wireless/realtek/rtw89/fw.c
10819
if (e->bw >= RTW89_2G_BW_NUM)
drivers/net/wireless/realtek/rtw89/fw.c
10821
if (e->nt >= RTW89_NTX_NUM)
drivers/net/wireless/realtek/rtw89/fw.c
10823
if (e->rs >= RTW89_RS_LMT_NUM)
drivers/net/wireless/realtek/rtw89/fw.c
10825
if (e->bf >= RTW89_BF_NUM)
drivers/net/wireless/realtek/rtw89/fw.c
10827
if (e->regd >= RTW89_REGD_NUM)
drivers/net/wireless/realtek/rtw89/fw.c
10829
if (e->ch_idx >= RTW89_2G_CH_NUM)
drivers/net/wireless/realtek/rtw89/fw.c
10852
fw_txpwr_lmt_5ghz_entry_valid(const struct rtw89_fw_txpwr_lmt_5ghz_entry *e,
drivers/net/wireless/realtek/rtw89/fw.c
10856
if (!__fw_txpwr_entry_acceptable(e, cursor, conf->ent_sz))
drivers/net/wireless/realtek/rtw89/fw.c
10859
if (e->bw >= RTW89_5G_BW_NUM)
drivers/net/wireless/realtek/rtw89/fw.c
10861
if (e->nt >= RTW89_NTX_NUM)
drivers/net/wireless/realtek/rtw89/fw.c
10863
if (e->rs >= RTW89_RS_LMT_NUM)
drivers/net/wireless/realtek/rtw89/fw.c
10865
if (e->bf >= RTW89_BF_NUM)
drivers/net/wireless/realtek/rtw89/fw.c
10867
if (e->regd >= RTW89_REGD_NUM)
drivers/net/wireless/realtek/rtw89/fw.c
10869
if (e->ch_idx >= RTW89_5G_CH_NUM)
drivers/net/wireless/realtek/rtw89/fw.c
10892
fw_txpwr_lmt_6ghz_entry_valid(const struct rtw89_fw_txpwr_lmt_6ghz_entry *e,
drivers/net/wireless/realtek/rtw89/fw.c
10896
if (!__fw_txpwr_entry_acceptable(e, cursor, conf->ent_sz))
drivers/net/wireless/realtek/rtw89/fw.c
10899
if (e->bw >= RTW89_6G_BW_NUM)
drivers/net/wireless/realtek/rtw89/fw.c
10901
if (e->nt >= RTW89_NTX_NUM)
drivers/net/wireless/realtek/rtw89/fw.c
10903
if (e->rs >= RTW89_RS_LMT_NUM)
drivers/net/wireless/realtek/rtw89/fw.c
10905
if (e->bf >= RTW89_BF_NUM)
drivers/net/wireless/realtek/rtw89/fw.c
10907
if (e->regd >= RTW89_REGD_NUM)
drivers/net/wireless/realtek/rtw89/fw.c
10909
if (e->reg_6ghz_power >= NUM_OF_RTW89_REG_6GHZ_POWER)
drivers/net/wireless/realtek/rtw89/fw.c
10911
if (e->ch_idx >= RTW89_6G_CH_NUM)
drivers/net/wireless/realtek/rtw89/fw.c
10934
fw_txpwr_lmt_ru_2ghz_entry_valid(const struct rtw89_fw_txpwr_lmt_ru_2ghz_entry *e,
drivers/net/wireless/realtek/rtw89/fw.c
10938
if (!__fw_txpwr_entry_acceptable(e, cursor, conf->ent_sz))
drivers/net/wireless/realtek/rtw89/fw.c
10941
if (e->ru >= RTW89_RU_NUM)
drivers/net/wireless/realtek/rtw89/fw.c
10943
if (e->nt >= RTW89_NTX_NUM)
drivers/net/wireless/realtek/rtw89/fw.c
10945
if (e->regd >= RTW89_REGD_NUM)
drivers/net/wireless/realtek/rtw89/fw.c
10947
if (e->ch_idx >= RTW89_2G_CH_NUM)
drivers/net/wireless/realtek/rtw89/fw.c
10969
fw_txpwr_lmt_ru_5ghz_entry_valid(const struct rtw89_fw_txpwr_lmt_ru_5ghz_entry *e,
drivers/net/wireless/realtek/rtw89/fw.c
10973
if (!__fw_txpwr_entry_acceptable(e, cursor, conf->ent_sz))
drivers/net/wireless/realtek/rtw89/fw.c
10976
if (e->ru >= RTW89_RU_NUM)
drivers/net/wireless/realtek/rtw89/fw.c
10978
if (e->nt >= RTW89_NTX_NUM)
drivers/net/wireless/realtek/rtw89/fw.c
10980
if (e->regd >= RTW89_REGD_NUM)
drivers/net/wireless/realtek/rtw89/fw.c
10982
if (e->ch_idx >= RTW89_5G_CH_NUM)
drivers/net/wireless/realtek/rtw89/fw.c
11004
fw_txpwr_lmt_ru_6ghz_entry_valid(const struct rtw89_fw_txpwr_lmt_ru_6ghz_entry *e,
drivers/net/wireless/realtek/rtw89/fw.c
11008
if (!__fw_txpwr_entry_acceptable(e, cursor, conf->ent_sz))
drivers/net/wireless/realtek/rtw89/fw.c
11011
if (e->ru >= RTW89_RU_NUM)
drivers/net/wireless/realtek/rtw89/fw.c
11013
if (e->nt >= RTW89_NTX_NUM)
drivers/net/wireless/realtek/rtw89/fw.c
11015
if (e->regd >= RTW89_REGD_NUM)
drivers/net/wireless/realtek/rtw89/fw.c
11017
if (e->reg_6ghz_power >= NUM_OF_RTW89_REG_6GHZ_POWER)
drivers/net/wireless/realtek/rtw89/fw.c
11019
if (e->ch_idx >= RTW89_6G_CH_NUM)
drivers/net/wireless/realtek/rtw89/fw.c
11042
fw_tx_shape_lmt_entry_valid(const struct rtw89_fw_tx_shape_lmt_entry *e,
drivers/net/wireless/realtek/rtw89/fw.c
11046
if (!__fw_txpwr_entry_acceptable(e, cursor, conf->ent_sz))
drivers/net/wireless/realtek/rtw89/fw.c
11049
if (e->band >= RTW89_BAND_NUM)
drivers/net/wireless/realtek/rtw89/fw.c
11051
if (e->tx_shape_rs >= RTW89_RS_TX_SHAPE_NUM)
drivers/net/wireless/realtek/rtw89/fw.c
11053
if (e->regd >= RTW89_REGD_NUM)
drivers/net/wireless/realtek/rtw89/fw.c
11075
fw_tx_shape_lmt_ru_entry_valid(const struct rtw89_fw_tx_shape_lmt_ru_entry *e,
drivers/net/wireless/realtek/rtw89/fw.c
11079
if (!__fw_txpwr_entry_acceptable(e, cursor, conf->ent_sz))
drivers/net/wireless/realtek/rtw89/fw.c
11082
if (e->band >= RTW89_BAND_NUM)
drivers/net/wireless/realtek/rtw89/fw.c
11084
if (e->regd >= RTW89_REGD_NUM)
drivers/net/wireless/realtek/rtw89/rtw8851b.c
624
ether_addr_copy(efuse->addr, map->e.mac_addr);
drivers/net/wireless/realtek/rtw89/rtw8851b.h
70
struct rtw8851be_efuse e;
drivers/net/wireless/realtek/rtw89/rtw8852a.c
669
ether_addr_copy(efuse->addr, map->e.mac_addr);
drivers/net/wireless/realtek/rtw89/rtw8852a.h
80
struct rtw8852ae_efuse e;
drivers/net/wireless/realtek/rtw89/rtw8852b_common.c
256
ether_addr_copy(efuse->addr, map->e.mac_addr);
drivers/net/wireless/realtek/rtw89/rtw8852b_common.h
91
struct rtw8852bx_e_efuse e;
drivers/net/wireless/realtek/rtw89/rtw8852c.c
591
ether_addr_copy(efuse->addr, map->e.mac_addr);
drivers/net/wireless/realtek/rtw89/rtw8852c.h
97
struct rtw8852c_e_efuse e;
drivers/net/xen-netback/xenbus.c
468
char *s, *e;
drivers/net/xen-netback/xenbus.c
481
b = simple_strtoul(s, &e, 10);
drivers/net/xen-netback/xenbus.c
482
if ((s == e) || (*e != ','))
drivers/net/xen-netback/xenbus.c
485
s = e + 1;
drivers/net/xen-netback/xenbus.c
486
u = simple_strtoul(s, &e, 10);
drivers/net/xen-netback/xenbus.c
487
if ((s == e) || (*e != '\0'))
drivers/net/xen-netback/xenbus.c
503
char *s, *e, *macstr;
drivers/net/xen-netback/xenbus.c
511
mac[i] = simple_strtoul(s, &e, 16);
drivers/net/xen-netback/xenbus.c
512
if ((s == e) || (*e != ((i == ETH_ALEN-1) ? '\0' : ':'))) {
drivers/net/xen-netback/xenbus.c
516
s = e+1;
drivers/net/xen-netfront.c
1885
char *s, *e, *macstr;
drivers/net/xen-netfront.c
1893
mac[i] = simple_strtoul(s, &e, 16);
drivers/net/xen-netfront.c
1894
if ((s == e) || (*e != ((i == ETH_ALEN-1) ? '\0' : ':'))) {
drivers/net/xen-netfront.c
1898
s = e+1;
drivers/nvdimm/label.c
1015
struct nd_label_ent *label_ent, *e;
drivers/nvdimm/label.c
1030
list_for_each_entry_safe(label_ent, e, &nd_mapping->labels, list) {
drivers/nvdimm/namespace_devs.c
1929
struct nd_label_ent *label_ent, *e;
drivers/nvdimm/namespace_devs.c
1939
list_for_each_entry_safe(label_ent, e, &nd_mapping->labels, list) {
drivers/nvdimm/namespace_devs.c
1998
struct list_head *l, *e;
drivers/nvdimm/namespace_devs.c
2009
list_for_each_safe(l, e, &nd_mapping->labels) {
drivers/nvdimm/region_devs.c
861
struct nd_label_ent *label_ent, *e;
drivers/nvdimm/region_devs.c
864
list_for_each_entry_safe(label_ent, e, &nd_mapping->labels, list) {
drivers/nvme/target/discovery.c
109
struct nvmf_disc_rsp_page_entry *e = &hdr->entries[numrec];
drivers/nvme/target/discovery.c
111
e->trtype = port->disc_addr.trtype;
drivers/nvme/target/discovery.c
112
e->adrfam = port->disc_addr.adrfam;
drivers/nvme/target/discovery.c
113
e->treq = port->disc_addr.treq;
drivers/nvme/target/discovery.c
114
e->portid = port->disc_addr.portid;
drivers/nvme/target/discovery.c
116
e->cntlid = cpu_to_le16(NVME_CNTLID_DYNAMIC);
drivers/nvme/target/discovery.c
117
e->asqsz = cpu_to_le16(NVME_AQ_DEPTH);
drivers/nvme/target/discovery.c
118
e->subtype = type;
drivers/nvme/target/discovery.c
119
memcpy(e->trsvcid, port->disc_addr.trsvcid, NVMF_TRSVCID_SIZE);
drivers/nvme/target/discovery.c
120
memcpy(e->traddr, traddr, NVMF_TRADDR_SIZE);
drivers/nvme/target/discovery.c
121
memcpy(e->tsas.common, port->disc_addr.tsas.common, NVMF_TSAS_SIZE);
drivers/nvme/target/discovery.c
122
strscpy(e->subnqn, subsys_nqn, NVMF_NQN_SIZE);
drivers/pci/pcie/aer.c
439
#define AER_GET_AGENT(t, e) \
drivers/pci/pcie/aer.c
440
((e & AER_AGENT_COMPLETER_MASK(t)) ? AER_AGENT_COMPLETER : \
drivers/pci/pcie/aer.c
441
(e & AER_AGENT_REQUESTER_MASK(t)) ? AER_AGENT_REQUESTER : \
drivers/pci/pcie/aer.c
442
(e & AER_AGENT_TRANSMITTER_MASK(t)) ? AER_AGENT_TRANSMITTER : \
drivers/pci/pcie/aer.c
457
#define AER_GET_LAYER_ERROR(t, e) \
drivers/pci/pcie/aer.c
458
((e & AER_PHYSICAL_LAYER_ERROR_MASK(t)) ? AER_PHYSICAL_LAYER_ERROR : \
drivers/pci/pcie/aer.c
459
(e & AER_DATA_LINK_LAYER_ERROR_MASK(t)) ? AER_DATA_LINK_LAYER_ERROR : \
drivers/pci/proc.c
422
struct proc_dir_entry *e;
drivers/pci/proc.c
441
e = proc_create_data(name, S_IFREG | S_IRUGO | S_IWUSR, bus->procdir,
drivers/pci/proc.c
443
if (!e)
drivers/pci/proc.c
445
proc_set_size(e, dev->cfg_size);
drivers/pci/proc.c
446
dev->procent = e;
drivers/perf/arm-ccn.c
779
struct arm_ccn_pmu_event *e = &arm_ccn_pmu_events[i];
drivers/perf/arm-ccn.c
783
if (!arm_ccn_pmu_type_eq(type, e->type))
drivers/perf/arm-ccn.c
785
if (event_id != e->event)
drivers/perf/arm-ccn.c
787
if (e->num_ports && port >= e->num_ports) {
drivers/perf/arm-ccn.c
792
if (e->num_vcs && vc >= e->num_vcs) {
drivers/perf/arm-cmn.c
1754
struct arm_cmn_event_attr *e;
drivers/perf/arm-cmn.c
1758
e = container_of(arm_cmn_event_attrs[i], typeof(*e), attr.attr);
drivers/perf/arm-cmn.c
1759
if (e->model & model && e->type == type && e->eventid == eventid)
drivers/perf/arm-cmn.c
1760
return e->fsel;
drivers/pinctrl/intel/pinctrl-alderlake.c
30
#define ADL_N_COMMUNITY(b, s, e, g) \
drivers/pinctrl/intel/pinctrl-alderlake.c
31
INTEL_COMMUNITY_GPPS(b, s, e, g, ADL_N)
drivers/pinctrl/intel/pinctrl-alderlake.c
33
#define ADL_S_COMMUNITY(b, s, e, g) \
drivers/pinctrl/intel/pinctrl-alderlake.c
34
INTEL_COMMUNITY_GPPS(b, s, e, g, ADL_S)
drivers/pinctrl/intel/pinctrl-broxton.c
24
#define BXT_COMMUNITY(b, s, e) \
drivers/pinctrl/intel/pinctrl-broxton.c
25
INTEL_COMMUNITY_SIZE(b, s, e, 32, 4, BXT)
drivers/pinctrl/intel/pinctrl-cannonlake.c
31
#define CNL_LP_COMMUNITY(b, s, e, g) \
drivers/pinctrl/intel/pinctrl-cannonlake.c
32
INTEL_COMMUNITY_GPPS(b, s, e, g, CNL_LP)
drivers/pinctrl/intel/pinctrl-cannonlake.c
34
#define CNL_H_COMMUNITY(b, s, e, g) \
drivers/pinctrl/intel/pinctrl-cannonlake.c
35
INTEL_COMMUNITY_GPPS(b, s, e, g, CNL_H)
drivers/pinctrl/intel/pinctrl-cedarfork.c
24
#define CDF_COMMUNITY(b, s, e, g) \
drivers/pinctrl/intel/pinctrl-cedarfork.c
25
INTEL_COMMUNITY_GPPS(b, s, e, g, CDF)
drivers/pinctrl/intel/pinctrl-denverton.c
24
#define DNV_COMMUNITY(b, s, e, g) \
drivers/pinctrl/intel/pinctrl-denverton.c
25
INTEL_COMMUNITY_GPPS(b, s, e, g, DNV)
drivers/pinctrl/intel/pinctrl-elkhartlake.c
24
#define EHL_COMMUNITY(b, s, e, g) \
drivers/pinctrl/intel/pinctrl-elkhartlake.c
25
INTEL_COMMUNITY_GPPS(b, s, e, g, EHL)
drivers/pinctrl/intel/pinctrl-emmitsburg.c
24
#define EBG_COMMUNITY(b, s, e, g) \
drivers/pinctrl/intel/pinctrl-emmitsburg.c
25
INTEL_COMMUNITY_GPPS(b, s, e, g, EBG)
drivers/pinctrl/intel/pinctrl-geminilake.c
24
#define GLK_COMMUNITY(b, s, e) \
drivers/pinctrl/intel/pinctrl-geminilake.c
25
INTEL_COMMUNITY_SIZE(b, s, e, 32, 4, GLK)
drivers/pinctrl/intel/pinctrl-icelake.c
31
#define ICL_LP_COMMUNITY(b, s, e, g) \
drivers/pinctrl/intel/pinctrl-icelake.c
32
INTEL_COMMUNITY_GPPS(b, s, e, g, ICL_LP)
drivers/pinctrl/intel/pinctrl-icelake.c
34
#define ICL_N_COMMUNITY(b, s, e, g) \
drivers/pinctrl/intel/pinctrl-icelake.c
35
INTEL_COMMUNITY_GPPS(b, s, e, g, ICL_N)
drivers/pinctrl/intel/pinctrl-intel.h
154
#define __INTEL_COMMUNITY(b, s, e, g, n, gs, gn, soc) \
drivers/pinctrl/intel/pinctrl-intel.h
165
.npins = ((e) - (s) + 1), \
drivers/pinctrl/intel/pinctrl-intel.h
170
#define INTEL_COMMUNITY_GPPS(b, s, e, g, soc) \
drivers/pinctrl/intel/pinctrl-intel.h
171
__INTEL_COMMUNITY(b, s, e, g, ARRAY_SIZE(g), 0, 0, soc)
drivers/pinctrl/intel/pinctrl-intel.h
173
#define INTEL_COMMUNITY_SIZE(b, s, e, gs, gn, soc) \
drivers/pinctrl/intel/pinctrl-intel.h
174
__INTEL_COMMUNITY(b, s, e, NULL, 0, gs, gn, soc)
drivers/pinctrl/intel/pinctrl-intel.h
80
#define INTEL_GPP(r, s, e, g) \
drivers/pinctrl/intel/pinctrl-intel.h
84
.size = ((e) - (s) + 1), \
drivers/pinctrl/intel/pinctrl-jasperlake.c
24
#define JSL_COMMUNITY(b, s, e, g) \
drivers/pinctrl/intel/pinctrl-jasperlake.c
25
INTEL_COMMUNITY_GPPS(b, s, e, g, JSL)
drivers/pinctrl/intel/pinctrl-lakefield.c
24
#define LKF_COMMUNITY(b, s, e, g) \
drivers/pinctrl/intel/pinctrl-lakefield.c
25
INTEL_COMMUNITY_GPPS(b, s, e, g, LKF)
drivers/pinctrl/intel/pinctrl-lewisburg.c
24
#define LBG_COMMUNITY(b, s, e) \
drivers/pinctrl/intel/pinctrl-lewisburg.c
25
INTEL_COMMUNITY_SIZE(b, s, e, 24, 3, LBG)
drivers/pinctrl/intel/pinctrl-meteorlake.c
30
#define MTL_P_COMMUNITY(b, s, e, g) \
drivers/pinctrl/intel/pinctrl-meteorlake.c
31
INTEL_COMMUNITY_GPPS(b, s, e, g, MTL_P)
drivers/pinctrl/intel/pinctrl-meteorlake.c
33
#define MTL_S_COMMUNITY(b, s, e, g) \
drivers/pinctrl/intel/pinctrl-meteorlake.c
34
INTEL_COMMUNITY_GPPS(b, s, e, g, MTL_S)
drivers/pinctrl/intel/pinctrl-meteorpoint.c
24
#define MTP_COMMUNITY(b, s, e, g) \
drivers/pinctrl/intel/pinctrl-meteorpoint.c
25
INTEL_COMMUNITY_GPPS(b, s, e, g, MTP)
drivers/pinctrl/intel/pinctrl-sunrisepoint.c
31
#define SPT_H_COMMUNITY(b, s, e, g) \
drivers/pinctrl/intel/pinctrl-sunrisepoint.c
32
INTEL_COMMUNITY_GPPS(b, s, e, g, SPT_H)
drivers/pinctrl/intel/pinctrl-sunrisepoint.c
34
#define SPT_LP_COMMUNITY(b, s, e) \
drivers/pinctrl/intel/pinctrl-sunrisepoint.c
35
INTEL_COMMUNITY_SIZE(b, s, e, 24, 4, SPT_LP)
drivers/pinctrl/intel/pinctrl-tangier.h
43
#define TNG_FAMILY(b, s, e) \
drivers/pinctrl/intel/pinctrl-tangier.h
47
.npins = (e) - (s) + 1, \
drivers/pinctrl/intel/pinctrl-tangier.h
50
#define TNG_FAMILY_PROTECTED(b, s, e) \
drivers/pinctrl/intel/pinctrl-tangier.h
54
.npins = (e) - (s) + 1, \
drivers/pinctrl/intel/pinctrl-tigerlake.c
31
#define TGL_LP_COMMUNITY(b, s, e, g) \
drivers/pinctrl/intel/pinctrl-tigerlake.c
32
INTEL_COMMUNITY_GPPS(b, s, e, g, TGL_LP)
drivers/pinctrl/intel/pinctrl-tigerlake.c
34
#define TGL_H_COMMUNITY(b, s, e, g) \
drivers/pinctrl/intel/pinctrl-tigerlake.c
35
INTEL_COMMUNITY_GPPS(b, s, e, g, TGL_H)
drivers/pinctrl/nuvoton/pinctrl-npcm7xx.c
873
#define NPCM7XX_PINCFG(a, b, c, d, e, f, g, h, i, j, k) \
drivers/pinctrl/nuvoton/pinctrl-npcm7xx.c
875
.fn1 = fn_ ## e, .reg1 = NPCM7XX_GCR_ ## f, .bit1 = g, \
drivers/pinctrl/nuvoton/pinctrl-npcm8xx.c
1272
#define NPCM8XX_PINCFG(a, b, c, d, e, f, g, h, i, j, k, l, m, n, o, p, q) \
drivers/pinctrl/nuvoton/pinctrl-npcm8xx.c
1276
.fn1 = fn_ ## e, .reg1 = NPCM8XX_GCR_ ## f, .bit1 = g, \
drivers/pinctrl/nuvoton/pinctrl-wpcm450.c
618
#define WPCM450_PINCFG(a, b, c, d, e, f, g) \
drivers/pinctrl/nuvoton/pinctrl-wpcm450.c
620
.fn1 = fn_ ## e, .reg1 = WPCM450_GCR_ ## f, .bit1 = g }
drivers/pinctrl/pinctrl-lpc18xx.c
386
LPC_P(e,0, R, R, R, EMC, GPIO, CAN1, R, R, 0, ND);
drivers/pinctrl/pinctrl-lpc18xx.c
387
LPC_P(e,1, R, R, R, EMC, GPIO, CAN1, R, R, 0, ND);
drivers/pinctrl/pinctrl-lpc18xx.c
388
LPC_P(e,2,ADCTRIG, CAN0, R, EMC, GPIO, R, R, R, 0, ND);
drivers/pinctrl/pinctrl-lpc18xx.c
389
LPC_P(e,3, R, CAN0,ADCTRIG, EMC, GPIO, R, R, R, 0, ND);
drivers/pinctrl/pinctrl-lpc18xx.c
390
LPC_P(e,4, R, NMI, R, EMC, GPIO, R, R, R, 0, ND);
drivers/pinctrl/pinctrl-lpc18xx.c
391
LPC_P(e,5, R, CTOUT, UART1, EMC, GPIO, R, R, R, 0, ND);
drivers/pinctrl/pinctrl-lpc18xx.c
392
LPC_P(e,6, R, CTOUT, UART1, EMC, GPIO, R, R, R, 0, ND);
drivers/pinctrl/pinctrl-lpc18xx.c
393
LPC_P(e,7, R, CTOUT, UART1, EMC, GPIO, R, R, R, 0, ND);
drivers/pinctrl/pinctrl-lpc18xx.c
394
LPC_P(e,8, R, CTOUT, UART1, EMC, GPIO, R, R, R, 0, ND);
drivers/pinctrl/pinctrl-lpc18xx.c
395
LPC_P(e,9, R, CTIN, UART1, EMC, GPIO, R, R, R, 0, ND);
drivers/pinctrl/pinctrl-lpc18xx.c
396
LPC_P(e,10, R, CTIN, UART1, EMC, GPIO, R, R, R, 0, ND);
drivers/pinctrl/pinctrl-lpc18xx.c
397
LPC_P(e,11, R, CTOUT, UART1, EMC, GPIO, R, R, R, 0, ND);
drivers/pinctrl/pinctrl-lpc18xx.c
398
LPC_P(e,12, R, CTOUT, UART1, EMC, GPIO, R, R, R, 0, ND);
drivers/pinctrl/pinctrl-lpc18xx.c
399
LPC_P(e,13, R, CTOUT, I2C1, EMC, GPIO, R, R, R, 0, ND);
drivers/pinctrl/pinctrl-lpc18xx.c
400
LPC_P(e,14, R, R, R, EMC, GPIO, R, R, R, 0, ND);
drivers/pinctrl/pinctrl-lpc18xx.c
401
LPC_P(e,15, R, CTOUT, I2C1, EMC, GPIO, R, R, R, 0, ND);
drivers/pinctrl/pinctrl-lpc18xx.c
596
LPC18XX_PIN_P(e,0),
drivers/pinctrl/pinctrl-lpc18xx.c
597
LPC18XX_PIN_P(e,1),
drivers/pinctrl/pinctrl-lpc18xx.c
598
LPC18XX_PIN_P(e,2),
drivers/pinctrl/pinctrl-lpc18xx.c
599
LPC18XX_PIN_P(e,3),
drivers/pinctrl/pinctrl-lpc18xx.c
600
LPC18XX_PIN_P(e,4),
drivers/pinctrl/pinctrl-lpc18xx.c
601
LPC18XX_PIN_P(e,5),
drivers/pinctrl/pinctrl-lpc18xx.c
602
LPC18XX_PIN_P(e,6),
drivers/pinctrl/pinctrl-lpc18xx.c
603
LPC18XX_PIN_P(e,7),
drivers/pinctrl/pinctrl-lpc18xx.c
604
LPC18XX_PIN_P(e,8),
drivers/pinctrl/pinctrl-lpc18xx.c
605
LPC18XX_PIN_P(e,9),
drivers/pinctrl/pinctrl-lpc18xx.c
606
LPC18XX_PIN_P(e,10),
drivers/pinctrl/pinctrl-lpc18xx.c
607
LPC18XX_PIN_P(e,11),
drivers/pinctrl/pinctrl-lpc18xx.c
608
LPC18XX_PIN_P(e,12),
drivers/pinctrl/pinctrl-lpc18xx.c
609
LPC18XX_PIN_P(e,13),
drivers/pinctrl/pinctrl-lpc18xx.c
610
LPC18XX_PIN_P(e,14),
drivers/pinctrl/pinctrl-lpc18xx.c
611
LPC18XX_PIN_P(e,15),
drivers/pinctrl/uniphier/pinctrl-uniphier.h
152
#define UNIPHIER_PINCTRL_PIN(a, b, c, d, e, f, g) \
drivers/pinctrl/uniphier/pinctrl-uniphier.h
156
.drv_data = (void *)UNIPHIER_PIN_ATTR_PACKED(c, d, e, f, g), \
drivers/platform/chrome/cros_ec_chardev.c
255
struct ec_event *event, *e;
drivers/platform/chrome/cros_ec_chardev.c
260
list_for_each_entry_safe(event, e, &priv->events, node) {
drivers/platform/surface/aggregator/controller.c
2574
struct ssam_nf_refcount_entry *e;
drivers/platform/surface/aggregator/controller.c
2576
e = rb_entry(n, struct ssam_nf_refcount_entry, node);
drivers/platform/surface/aggregator/controller.c
2577
status = ssam_ssh_event_disable(ctrl, e->key.reg,
drivers/platform/surface/aggregator/controller.c
2578
e->key.id, e->flags);
drivers/platform/surface/aggregator/controller.c
2588
struct ssam_nf_refcount_entry *e;
drivers/platform/surface/aggregator/controller.c
2590
e = rb_entry(n, struct ssam_nf_refcount_entry, node);
drivers/platform/surface/aggregator/controller.c
2591
ssam_ssh_event_enable(ctrl, e->key.reg, e->key.id, e->flags);
drivers/platform/surface/aggregator/controller.c
2618
struct ssam_nf_refcount_entry *e;
drivers/platform/surface/aggregator/controller.c
2620
e = rb_entry(n, struct ssam_nf_refcount_entry, node);
drivers/platform/surface/aggregator/controller.c
2623
ssam_ssh_event_enable(ctrl, e->key.reg, e->key.id, e->flags);
drivers/platform/surface/aggregator/controller.c
2659
struct ssam_nf_refcount_entry *e, *n;
drivers/platform/surface/aggregator/controller.c
2662
rbtree_postorder_for_each_entry_safe(e, n, &nf->refcount, node) {
drivers/platform/surface/aggregator/controller.c
2664
ssam_ssh_event_disable(ctrl, e->key.reg, e->key.id, e->flags);
drivers/platform/surface/aggregator/controller.c
2665
kfree(e);
drivers/platform/surface/surface_dtx.c
578
struct sdtx_event e;
drivers/platform/surface/surface_dtx.c
583
struct sdtx_event e;
drivers/platform/surface/surface_dtx.c
670
event.base.e.length = sizeof(struct sdtx_base_info);
drivers/platform/surface/surface_dtx.c
671
event.base.e.code = SDTX_EVENT_BASE_CONNECTION;
drivers/platform/surface/surface_dtx.c
682
event.status.e.length = sizeof(u16);
drivers/platform/surface/surface_dtx.c
683
event.status.e.code = SDTX_EVENT_CANCEL;
drivers/platform/surface/surface_dtx.c
696
event.status.e.length = sizeof(u16);
drivers/platform/surface/surface_dtx.c
697
event.status.e.code = SDTX_EVENT_LATCH_STATUS;
drivers/platform/surface/surface_dtx.c
773
event.e.length = sizeof(u16);
drivers/platform/surface/surface_dtx.c
774
event.e.code = SDTX_EVENT_DEVICE_MODE;
drivers/platform/surface/surface_dtx.c
777
sdtx_push_event(ddev, &event.e);
drivers/platform/surface/surface_dtx.c
807
event.e.length = sizeof(struct sdtx_base_info);
drivers/platform/surface/surface_dtx.c
808
event.e.code = SDTX_EVENT_BASE_CONNECTION;
drivers/platform/surface/surface_dtx.c
812
sdtx_push_event(ddev, &event.e);
drivers/platform/surface/surface_dtx.c
842
event.e.length = sizeof(u16);
drivers/platform/surface/surface_dtx.c
843
event.e.code = SDTX_EVENT_DEVICE_MODE;
drivers/platform/surface/surface_dtx.c
846
sdtx_push_event(ddev, &event.e);
drivers/platform/surface/surface_dtx.c
867
event.e.length = sizeof(struct sdtx_base_info);
drivers/platform/surface/surface_dtx.c
868
event.e.code = SDTX_EVENT_BASE_CONNECTION;
drivers/platform/surface/surface_dtx.c
871
sdtx_push_event(ddev, &event.e);
drivers/pmdomain/bcm/bcm-pmb.c
277
const struct bcm_pmb_pd_data *e;
drivers/pmdomain/bcm/bcm-pmb.c
301
for (e = table; e->name; e++)
drivers/pmdomain/bcm/bcm-pmb.c
302
max_id = max(max_id, e->id);
drivers/pmdomain/bcm/bcm-pmb.c
311
for (e = table; e->name; e++) {
drivers/pmdomain/bcm/bcm-pmb.c
318
pd->data = e;
drivers/pmdomain/bcm/bcm-pmb.c
319
pd->genpd.name = e->name;
drivers/pmdomain/bcm/bcm-pmb.c
324
pmb->genpd_onecell_data.domains[e->id] = &pd->genpd;
drivers/pmdomain/mediatek/mtk-mfg-pmdomain.c
540
struct mtk_mfg_opp_entry e = {};
drivers/pmdomain/mediatek/mtk-mfg-pmdomain.c
571
memcpy_fromio(&e, mfg->shared_mem + GF_REG_OPP_TABLE_GPU + i * sizeof(e),
drivers/pmdomain/mediatek/mtk-mfg-pmdomain.c
572
sizeof(e));
drivers/pmdomain/mediatek/mtk-mfg-pmdomain.c
573
if (mem_is_zero(&e, sizeof(e))) {
drivers/pmdomain/mediatek/mtk-mfg-pmdomain.c
578
mfg->gpu_opps[i].freq = e.freq_khz * HZ_PER_KHZ;
drivers/pmdomain/mediatek/mtk-mfg-pmdomain.c
579
mfg->gpu_opps[i].u_volt = e.voltage_core * 10;
drivers/pmdomain/mediatek/mtk-mfg-pmdomain.c
586
memcpy_fromio(&e, mfg->shared_mem + GF_REG_OPP_TABLE_STK + i * sizeof(e),
drivers/pmdomain/mediatek/mtk-mfg-pmdomain.c
587
sizeof(e));
drivers/pmdomain/mediatek/mtk-mfg-pmdomain.c
588
if (mem_is_zero(&e, sizeof(e))) {
drivers/pmdomain/mediatek/mtk-mfg-pmdomain.c
593
mfg->stack_opps[i].freq = e.freq_khz * HZ_PER_KHZ;
drivers/pmdomain/mediatek/mtk-mfg-pmdomain.c
594
mfg->stack_opps[i].u_volt = e.voltage_core * 10;
drivers/ras/cec.c
89
#define PFN(e) ((e) >> PAGE_SHIFT)
drivers/ras/cec.c
90
#define DECAY(e) (((e) >> COUNT_BITS) & DECAY_MASK)
drivers/ras/cec.c
91
#define COUNT(e) ((unsigned int)(e) & COUNT_MASK)
drivers/ras/cec.c
92
#define FULL_COUNT(e) ((e) & (PAGE_SIZE - 1))
drivers/regulator/s2dos05-regulator.c
35
#define BUCK_DESC(_name, _id, _ops, m, s, v, e, em, t, a) { \
drivers/regulator/s2dos05-regulator.c
49
.enable_reg = e, \
drivers/regulator/s2dos05-regulator.c
58
#define LDO_DESC(_name, _id, _ops, m, s, v, e, em, t, a) { \
drivers/regulator/s2dos05-regulator.c
72
.enable_reg = e, \
drivers/rpmsg/qcom_smd.c
174
#define to_smd_endpoint(e) container_of(e, struct qcom_smd_endpoint, ept)
drivers/s390/char/sclp.c
89
struct sclp_trace_entry e;
drivers/s390/char/sclp.c
91
memset(&e, 0, sizeof(e));
drivers/s390/char/sclp.c
92
strtomem(e.id, id);
drivers/s390/char/sclp.c
93
e.a = a;
drivers/s390/char/sclp.c
94
e.b = b;
drivers/s390/char/sclp.c
95
debug_event(&sclp_debug, prio, &e, sizeof(e));
drivers/s390/char/sclp.c
97
debug_event(&sclp_debug_err, 0, &e, sizeof(e));
drivers/s390/cio/chsc.c
875
u32 e : 1;
drivers/s390/cio/chsc.c
898
secm_area->e = 1;
drivers/s390/cio/device_fsm.c
412
static void ccw_device_request_event(struct ccw_device *cdev, enum dev_event e)
drivers/s390/cio/device_fsm.c
414
switch (e) {
drivers/s390/net/fsm.c
102
if (e == -1)
drivers/s390/net/fsm.c
108
fi->f->event_names[e]);
drivers/s390/net/fsm.c
99
int e = fi->history[idx].event;
drivers/s390/net/ism.h
220
u32 e : 1; /* event bit */
drivers/s390/net/ism_drv.c
528
if (ism->sba->e) {
drivers/s390/net/ism_drv.c
529
ism->sba->e = 0;
drivers/s390/net/qeth_core_main.c
3791
int e = 0;
drivers/s390/net/qeth_core_main.c
3793
while ((e < QDIO_MAX_ELEMENTS_PER_BUFFER) &&
drivers/s390/net/qeth_core_main.c
3794
buffer->element[e].addr) {
drivers/s390/net/qeth_core_main.c
3795
dma64_t phys_aob_addr = buffer->element[e].addr;
drivers/s390/net/qeth_core_main.c
3798
++e;
drivers/s390/net/qeth_l3_main.c
1303
int e;
drivers/s390/net/qeth_l3_main.c
1330
for (e = 0; e < qdata->no_entries; ++e) {
drivers/scsi/cxgbi/cxgb3i/cxgb3i.c
160
const struct l2t_entry *e)
drivers/scsi/cxgbi/cxgb3i/cxgb3i.c
176
V_L2T_IDX(e->idx) | V_TX_CHANNEL(e->smt_idx));
drivers/scsi/cxgbi/cxgb3i/cxgb3i.c
185
csk->mss_idx, e->idx, e->smt_idx);
drivers/scsi/cxgbi/cxgb4i/cxgb4i.c
209
struct l2t_entry *e)
drivers/scsi/cxgbi/cxgb4i/cxgb4i.c
328
struct l2t_entry *e)
drivers/scsi/elx/efct/efct_hw.c
164
efct_hw_cb_link(void *ctx, void *e)
drivers/scsi/elx/efct/efct_hw.c
167
struct sli4_link_event *event = e;
drivers/scsi/isci/request.c
105
static void init_sgl_element(struct scu_sgl_element *e, struct scatterlist *sg)
drivers/scsi/isci/request.c
107
e->length = sg_dma_len(sg);
drivers/scsi/isci/request.c
108
e->address_upper = upper_32_bits(sg_dma_address(sg));
drivers/scsi/isci/request.c
109
e->address_lower = lower_32_bits(sg_dma_address(sg));
drivers/scsi/isci/request.c
110
e->address_modifier = 0;
drivers/scsi/pm8001/pm8001_hwi.c
3898
buf_prd->im_len.e = 0;
drivers/scsi/pm8001/pm8001_hwi.c
4592
flash_update_info.sgl.im_len.e = 0;
drivers/scsi/pm8001/pm8001_sas.h
282
__le32 e;
drivers/scsi/qla2xxx/qla_edif.c
1915
struct enode *e, *tmp;
drivers/scsi/qla2xxx/qla_edif.c
1925
list_for_each_entry_safe(e, tmp, &vha->pur_cinfo.head, list) {
drivers/scsi/qla2xxx/qla_edif.c
1926
purex = &e->u.purexinfo;
drivers/scsi/qla2xxx/qla_edif.c
1934
list_del_init(&e->list);
drivers/scsi/qla2xxx/qla_edif.c
1935
list_add_tail(&e->list, &enode_list);
drivers/scsi/qla2xxx/qla_edif.c
1940
list_for_each_entry_safe(e, tmp, &enode_list, list) {
drivers/scsi/qla2xxx/qla_edif.c
1941
list_del_init(&e->list);
drivers/scsi/qla2xxx/qla_edif.c
1942
qla_enode_free(vha, e);
drivers/scsi/qla2xxx/qla_edif.c
207
struct qla_work_evt *e;
drivers/scsi/qla2xxx/qla_edif.c
209
e = qla2x00_alloc_work(vha, QLA_EVT_SA_REPLACE);
drivers/scsi/qla2xxx/qla_edif.c
210
if (!e)
drivers/scsi/qla2xxx/qla_edif.c
2109
struct edb_node *e, *tmp;
drivers/scsi/qla2xxx/qla_edif.c
2122
list_for_each_entry_safe(e, tmp, &vha->e_dbell.head, list) {
drivers/scsi/qla2xxx/qla_edif.c
2123
switch (e->ntype) {
drivers/scsi/qla2xxx/qla_edif.c
2126
sid = e->u.plogi_did;
drivers/scsi/qla2xxx/qla_edif.c
2129
sid = e->u.els_sid;
drivers/scsi/qla2xxx/qla_edif.c
213
e->u.sa_update.fcport = fcport;
drivers/scsi/qla2xxx/qla_edif.c
2136
"%s unknown node type: %x\n", __func__, e->ntype);
drivers/scsi/qla2xxx/qla_edif.c
214
e->u.sa_update.sa_ctl = sa_ctl;
drivers/scsi/qla2xxx/qla_edif.c
2143
__func__, e->ntype, e);
drivers/scsi/qla2xxx/qla_edif.c
2144
list_del_init(&e->list);
drivers/scsi/qla2xxx/qla_edif.c
2145
list_add_tail(&e->list, &edb_list);
drivers/scsi/qla2xxx/qla_edif.c
215
e->u.sa_update.nport_handle = nport_handle;
drivers/scsi/qla2xxx/qla_edif.c
2150
list_for_each_entry_safe(e, tmp, &edb_list, list)
drivers/scsi/qla2xxx/qla_edif.c
2151
qla_edb_node_free(vha, e);
drivers/scsi/qla2xxx/qla_edif.c
217
return qla2x00_post_work(vha, e);
drivers/scsi/qla2xxx/qla_edif.c
2350
qla24xx_issue_sa_replace_iocb(scsi_qla_host_t *vha, struct qla_work_evt *e)
drivers/scsi/qla2xxx/qla_edif.c
2356
struct edif_sa_ctl *sa_ctl = e->u.sa_update.sa_ctl;
drivers/scsi/qla2xxx/qla_edif.c
2357
uint16_t nport_handle = e->u.sa_update.nport_handle;
drivers/scsi/qla2xxx/qla_edif.c
253
p = &req->e;
drivers/scsi/qla2xxx/qla_edif.c
3525
if (p->e.sub_cmd == SEND_ELS_REPLY) {
drivers/scsi/qla2xxx/qla_edif.c
3526
a->control_flags = p->e.extra_control_flags << 13;
drivers/scsi/qla2xxx/qla_edif.c
3527
a->rx_xchg_address = cpu_to_le32(p->e.extra_rx_xchg_address);
drivers/scsi/qla2xxx/qla_edif.c
3528
if (p->e.extra_control_flags == BSG_CTL_FLAG_LS_ACC)
drivers/scsi/qla2xxx/qla_edif.c
3530
else if (p->e.extra_control_flags == BSG_CTL_FLAG_LS_RJT)
drivers/scsi/qla2xxx/qla_edif.c
3648
__func__, sc_to_str(p->e.sub_cmd), fcport->port_name,
drivers/scsi/qla2xxx/qla_edif.c
3649
p->e.extra_rx_xchg_address, p->e.extra_control_flags,
drivers/scsi/qla2xxx/qla_edif.c
45
struct els_sub_cmd *e;
drivers/scsi/qla2xxx/qla_edif.c
48
e = sc_str + i;
drivers/scsi/qla2xxx/qla_edif.c
49
if (cmd == e->cmd)
drivers/scsi/qla2xxx/qla_edif.c
50
return e->str;
drivers/scsi/qla2xxx/qla_edif_bsg.h
40
struct extra_auth_els e;
drivers/scsi/qla2xxx/qla_gbl.h
331
struct qla_work_evt *e);
drivers/scsi/qla2xxx/qla_gbl.h
77
int qla2x00_post_work(struct scsi_qla_host *vha, struct qla_work_evt *e);
drivers/scsi/qla2xxx/qla_gs.c
2714
struct qla_work_evt *e;
drivers/scsi/qla2xxx/qla_gs.c
2716
e = qla2x00_alloc_work(vha, QLA_EVT_GPSC);
drivers/scsi/qla2xxx/qla_gs.c
2717
if (!e)
drivers/scsi/qla2xxx/qla_gs.c
2720
e->u.fcport.fcport = fcport;
drivers/scsi/qla2xxx/qla_gs.c
2721
return qla2x00_post_work(vha, e);
drivers/scsi/qla2xxx/qla_gs.c
3319
struct qla_work_evt *e;
drivers/scsi/qla2xxx/qla_gs.c
3321
e = qla2x00_alloc_work(vha, cmd);
drivers/scsi/qla2xxx/qla_gs.c
3322
if (!e)
drivers/scsi/qla2xxx/qla_gs.c
3325
e->u.iosb.sp = sp;
drivers/scsi/qla2xxx/qla_gs.c
3327
return qla2x00_post_work(vha, e);
drivers/scsi/qla2xxx/qla_gs.c
3874
struct qla_work_evt *e;
drivers/scsi/qla2xxx/qla_gs.c
3882
e = qla2x00_alloc_work(vha, QLA_EVT_GFPNID);
drivers/scsi/qla2xxx/qla_gs.c
3883
if (!e)
drivers/scsi/qla2xxx/qla_gs.c
3886
e->u.fcport.fcport = fcport;
drivers/scsi/qla2xxx/qla_gs.c
3887
return qla2x00_post_work(vha, e);
drivers/scsi/qla2xxx/qla_gs.c
507
struct qla_work_evt *e;
drivers/scsi/qla2xxx/qla_gs.c
528
e = qla2x00_alloc_work(vha, QLA_EVT_SP_RETRY);
drivers/scsi/qla2xxx/qla_gs.c
529
if (!e)
drivers/scsi/qla2xxx/qla_gs.c
532
e->u.iosb.sp = sp;
drivers/scsi/qla2xxx/qla_gs.c
533
qla2x00_post_work(vha, e);
drivers/scsi/qla2xxx/qla_gs.c
538
e = qla2x00_alloc_work(vha, QLA_EVT_UNMAP);
drivers/scsi/qla2xxx/qla_gs.c
540
if (!e) {
drivers/scsi/qla2xxx/qla_gs.c
563
e->u.iosb.sp = sp;
drivers/scsi/qla2xxx/qla_gs.c
564
qla2x00_post_work(vha, e);
drivers/scsi/qla2xxx/qla_init.c
1030
struct get_name_list_extended *e;
drivers/scsi/qla2xxx/qla_init.c
1054
e = &vha->gnl.l[i];
drivers/scsi/qla2xxx/qla_init.c
1055
loop_id = le16_to_cpu(e->nport_handle);
drivers/scsi/qla2xxx/qla_init.c
1059
wwn = wwn_to_u64(e->port_name);
drivers/scsi/qla2xxx/qla_init.c
1063
__func__, &wwn, e->port_id[2], e->port_id[1],
drivers/scsi/qla2xxx/qla_init.c
1064
e->port_id[0], e->current_login_state, e->last_login_state,
drivers/scsi/qla2xxx/qla_init.c
1091
e = &vha->gnl.l[i];
drivers/scsi/qla2xxx/qla_init.c
1092
wwn = wwn_to_u64(e->port_name);
drivers/scsi/qla2xxx/qla_init.c
1103
id.b.domain = e->port_id[2];
drivers/scsi/qla2xxx/qla_init.c
1104
id.b.area = e->port_id[1];
drivers/scsi/qla2xxx/qla_init.c
1105
id.b.al_pa = e->port_id[0];
drivers/scsi/qla2xxx/qla_init.c
1112
wwnn = wwn_to_u64(e->node_name);
drivers/scsi/qla2xxx/qla_init.c
1210
struct qla_work_evt *e;
drivers/scsi/qla2xxx/qla_init.c
1212
e = qla2x00_alloc_work(vha, QLA_EVT_GNL);
drivers/scsi/qla2xxx/qla_init.c
1213
if (!e)
drivers/scsi/qla2xxx/qla_init.c
1216
e->u.fcport.fcport = fcport;
drivers/scsi/qla2xxx/qla_init.c
1218
return qla2x00_post_work(vha, e);
drivers/scsi/qla2xxx/qla_init.c
1253
struct qla_work_evt *e;
drivers/scsi/qla2xxx/qla_init.c
1258
e = qla2x00_alloc_work(vha, QLA_EVT_PRLI);
drivers/scsi/qla2xxx/qla_init.c
1259
if (!e)
drivers/scsi/qla2xxx/qla_init.c
1262
e->u.fcport.fcport = fcport;
drivers/scsi/qla2xxx/qla_init.c
1264
return qla2x00_post_work(vha, e);
drivers/scsi/qla2xxx/qla_init.c
1361
struct qla_work_evt *e;
drivers/scsi/qla2xxx/qla_init.c
1363
e = qla2x00_alloc_work(vha, QLA_EVT_GPDB);
drivers/scsi/qla2xxx/qla_init.c
1364
if (!e)
drivers/scsi/qla2xxx/qla_init.c
1367
e->u.fcport.fcport = fcport;
drivers/scsi/qla2xxx/qla_init.c
1368
e->u.fcport.opt = opt;
drivers/scsi/qla2xxx/qla_init.c
1370
return qla2x00_post_work(vha, e);
drivers/scsi/qla2xxx/qla_init.c
1829
struct qla_work_evt *e;
drivers/scsi/qla2xxx/qla_init.c
1831
e = qla2x00_alloc_work(vha, QLA_EVT_NEW_SESS);
drivers/scsi/qla2xxx/qla_init.c
1832
if (!e)
drivers/scsi/qla2xxx/qla_init.c
1835
e->u.new_sess.id = *id;
drivers/scsi/qla2xxx/qla_init.c
1836
e->u.new_sess.pla = pla;
drivers/scsi/qla2xxx/qla_init.c
1837
e->u.new_sess.fc4_type = fc4_type;
drivers/scsi/qla2xxx/qla_init.c
1838
memcpy(e->u.new_sess.port_name, port_name, WWN_SIZE);
drivers/scsi/qla2xxx/qla_init.c
1840
memcpy(e->u.new_sess.node_name, node_name, WWN_SIZE);
drivers/scsi/qla2xxx/qla_init.c
1842
return qla2x00_post_work(vha, e);
drivers/scsi/qla2xxx/qla_init.c
559
struct qla_work_evt *e;
drivers/scsi/qla2xxx/qla_init.c
561
e = qla2x00_alloc_work(vha, QLA_EVT_ELS_PLOGI);
drivers/scsi/qla2xxx/qla_init.c
562
if (!e)
drivers/scsi/qla2xxx/qla_init.c
565
e->u.fcport.fcport = fcport;
drivers/scsi/qla2xxx/qla_init.c
568
return qla2x00_post_work(vha, e);
drivers/scsi/qla2xxx/qla_init.c
6122
struct qla_work_evt *e;
drivers/scsi/qla2xxx/qla_init.c
6124
e = qla2x00_alloc_work(vha, QLA_EVT_IIDMA);
drivers/scsi/qla2xxx/qla_init.c
6125
if (!e)
drivers/scsi/qla2xxx/qla_init.c
6128
e->u.fcport.fcport = fcport;
drivers/scsi/qla2xxx/qla_init.c
6129
return qla2x00_post_work(vha, e);
drivers/scsi/qla2xxx/qla_init.c
725
struct get_name_list_extended *e;
drivers/scsi/qla2xxx/qla_init.c
773
e = &vha->gnl.l[i];
drivers/scsi/qla2xxx/qla_init.c
774
wwn = wwn_to_u64(e->port_name);
drivers/scsi/qla2xxx/qla_init.c
775
id.b.domain = e->port_id[2];
drivers/scsi/qla2xxx/qla_init.c
776
id.b.area = e->port_id[1];
drivers/scsi/qla2xxx/qla_init.c
777
id.b.al_pa = e->port_id[0];
drivers/scsi/qla2xxx/qla_init.c
788
loop_id = le16_to_cpu(e->nport_handle);
drivers/scsi/qla2xxx/qla_init.c
790
nvme_cls = e->current_login_state >> 4;
drivers/scsi/qla2xxx/qla_init.c
791
current_login_state = e->current_login_state & 0xf;
drivers/scsi/qla2xxx/qla_init.c
805
e->current_login_state, fcport->fw_login_state,
drivers/scsi/qla2xxx/qla_init.c
856
if ((e->prli_svc_param_word_3[0] & BIT_4) == 0)
drivers/scsi/qla2xxx/qla_init.c
901
if ((e->prli_svc_param_word_3[0] & BIT_4) == 0)
drivers/scsi/qla2xxx/qla_init.c
949
e = &vha->gnl.l[i];
drivers/scsi/qla2xxx/qla_init.c
950
id.b.domain = e->port_id[0];
drivers/scsi/qla2xxx/qla_init.c
951
id.b.area = e->port_id[1];
drivers/scsi/qla2xxx/qla_init.c
952
id.b.al_pa = e->port_id[2];
drivers/scsi/qla2xxx/qla_init.c
954
loop_id = le16_to_cpu(e->nport_handle);
drivers/scsi/qla2xxx/qla_init.c
959
e->port_name, 0);
drivers/scsi/qla2xxx/qla_iocb.c
2911
struct qla_work_evt *e;
drivers/scsi/qla2xxx/qla_iocb.c
3027
e = qla2x00_alloc_work(vha, QLA_EVT_UNMAP);
drivers/scsi/qla2xxx/qla_iocb.c
3028
if (!e) {
drivers/scsi/qla2xxx/qla_iocb.c
3036
e->u.iosb.sp = sp;
drivers/scsi/qla2xxx/qla_iocb.c
3037
qla2x00_post_work(vha, e);
drivers/scsi/qla2xxx/qla_isr.c
2324
struct els_sts_entry_24xx *e = (struct els_sts_entry_24xx *)pkt;
drivers/scsi/qla2xxx/qla_isr.c
2352
__func__, sc_to_str(p->e.sub_cmd),
drivers/scsi/qla2xxx/qla_isr.c
2353
e->d_id[2], e->d_id[1], e->d_id[0],
drivers/scsi/qla2xxx/qla_isr.c
2354
comp_status, p->e.extra_rx_xchg_address, bsg_job);
drivers/scsi/qla2xxx/qla_isr.c
2419
__func__, e->s_id[0], e->s_id[2], e->s_id[1],
drivers/scsi/qla2xxx/qla_isr.c
2420
e->d_id[2], e->d_id[1], e->d_id[0]);
drivers/scsi/qla2xxx/qla_isr.c
2449
e->s_id[0], e->s_id[2], e->s_id[1],
drivers/scsi/qla2xxx/qla_isr.c
2450
e->d_id[2], e->d_id[1], e->d_id[0]);
drivers/scsi/qla2xxx/qla_isr.c
2467
e->s_id[0], e->s_id[2], e->s_id[1],
drivers/scsi/qla2xxx/qla_isr.c
2468
e->d_id[2], e->d_id[1], e->d_id[0]);
drivers/scsi/qla2xxx/qla_mbx.c
31
struct mb_cmd_name *e;
drivers/scsi/qla2xxx/qla_mbx.c
34
e = mb_str + i;
drivers/scsi/qla2xxx/qla_mbx.c
35
if (cmd == e->cmd)
drivers/scsi/qla2xxx/qla_mbx.c
36
return e->str;
drivers/scsi/qla2xxx/qla_os.c
5127
struct qla_work_evt *e;
drivers/scsi/qla2xxx/qla_os.c
5135
e = kzalloc_obj(struct qla_work_evt, GFP_ATOMIC);
drivers/scsi/qla2xxx/qla_os.c
5136
if (!e) {
drivers/scsi/qla2xxx/qla_os.c
5141
INIT_LIST_HEAD(&e->list);
drivers/scsi/qla2xxx/qla_os.c
5142
e->type = type;
drivers/scsi/qla2xxx/qla_os.c
5143
e->flags = QLA_EVT_FLAG_FREE;
drivers/scsi/qla2xxx/qla_os.c
5144
return e;
drivers/scsi/qla2xxx/qla_os.c
5148
qla2x00_post_work(struct scsi_qla_host *vha, struct qla_work_evt *e)
drivers/scsi/qla2xxx/qla_os.c
5154
list_add_tail(&e->list, &vha->work_list);
drivers/scsi/qla2xxx/qla_os.c
5171
struct qla_work_evt *e;
drivers/scsi/qla2xxx/qla_os.c
5173
e = qla2x00_alloc_work(vha, QLA_EVT_AEN);
drivers/scsi/qla2xxx/qla_os.c
5174
if (!e)
drivers/scsi/qla2xxx/qla_os.c
5177
e->u.aen.code = code;
drivers/scsi/qla2xxx/qla_os.c
5178
e->u.aen.data = data;
drivers/scsi/qla2xxx/qla_os.c
5179
return qla2x00_post_work(vha, e);
drivers/scsi/qla2xxx/qla_os.c
5185
struct qla_work_evt *e;
drivers/scsi/qla2xxx/qla_os.c
5187
e = qla2x00_alloc_work(vha, QLA_EVT_IDC_ACK);
drivers/scsi/qla2xxx/qla_os.c
5188
if (!e)
drivers/scsi/qla2xxx/qla_os.c
5191
memcpy(e->u.idc_ack.mb, mb, QLA_IDC_ACK_REGS * sizeof(uint16_t));
drivers/scsi/qla2xxx/qla_os.c
5192
return qla2x00_post_work(vha, e);
drivers/scsi/qla2xxx/qla_os.c
5200
struct qla_work_evt *e; \
drivers/scsi/qla2xxx/qla_os.c
5202
e = qla2x00_alloc_work(vha, type); \
drivers/scsi/qla2xxx/qla_os.c
5203
if (!e) \
drivers/scsi/qla2xxx/qla_os.c
5206
e->u.logio.fcport = fcport; \
drivers/scsi/qla2xxx/qla_os.c
5208
e->u.logio.data[0] = data[0]; \
drivers/scsi/qla2xxx/qla_os.c
5209
e->u.logio.data[1] = data[1]; \
drivers/scsi/qla2xxx/qla_os.c
5212
return qla2x00_post_work(vha, e); \
drivers/scsi/qla2xxx/qla_os.c
5224
struct qla_work_evt *e;
drivers/scsi/qla2xxx/qla_os.c
5226
e = qla2x00_alloc_work(vha, QLA_EVT_UEVENT);
drivers/scsi/qla2xxx/qla_os.c
5227
if (!e)
drivers/scsi/qla2xxx/qla_os.c
5230
e->u.uevent.code = code;
drivers/scsi/qla2xxx/qla_os.c
5231
return qla2x00_post_work(vha, e);
drivers/scsi/qla2xxx/qla_os.c
5256
struct qla_work_evt *e;
drivers/scsi/qla2xxx/qla_os.c
5258
e = qla2x00_alloc_work(vha, QLA_EVT_AENFX);
drivers/scsi/qla2xxx/qla_os.c
5259
if (!e)
drivers/scsi/qla2xxx/qla_os.c
5262
e->u.aenfx.evtcode = evtcode;
drivers/scsi/qla2xxx/qla_os.c
5263
e->u.aenfx.count = cnt;
drivers/scsi/qla2xxx/qla_os.c
5264
memcpy(e->u.aenfx.mbx, data, sizeof(*data) * cnt);
drivers/scsi/qla2xxx/qla_os.c
5265
return qla2x00_post_work(vha, e);
drivers/scsi/qla2xxx/qla_os.c
5290
void qla24xx_create_new_sess(struct scsi_qla_host *vha, struct qla_work_evt *e)
drivers/scsi/qla2xxx/qla_os.c
5295
(struct qlt_plogi_ack_t *)e->u.new_sess.pla;
drivers/scsi/qla2xxx/qla_os.c
5300
__func__, __LINE__, e->u.new_sess.port_name);
drivers/scsi/qla2xxx/qla_os.c
5303
fcport = qla2x00_find_fcport_by_wwpn(vha, e->u.new_sess.port_name, 1);
drivers/scsi/qla2xxx/qla_os.c
5305
fcport->d_id = e->u.new_sess.id;
drivers/scsi/qla2xxx/qla_os.c
5321
fcport->d_id = e->u.new_sess.id;
drivers/scsi/qla2xxx/qla_os.c
5326
memcpy(fcport->port_name, e->u.new_sess.port_name,
drivers/scsi/qla2xxx/qla_os.c
5329
fcport->fc4_type = e->u.new_sess.fc4_type;
drivers/scsi/qla2xxx/qla_os.c
5335
if (e->u.new_sess.fc4_type & FS_FCP_IS_N2N) {
drivers/scsi/qla2xxx/qla_os.c
5347
__func__, e->u.new_sess.port_name);
drivers/scsi/qla2xxx/qla_os.c
5359
e->u.new_sess.port_name, 1);
drivers/scsi/qla2xxx/qla_os.c
5384
memcpy(fcport->node_name, e->u.new_sess.node_name, WWN_SIZE);
drivers/scsi/qla2xxx/qla_os.c
5414
&e->u.new_sess.id, 1);
drivers/scsi/qla2xxx/qla_os.c
5468
static void qla_sp_retry(struct scsi_qla_host *vha, struct qla_work_evt *e)
drivers/scsi/qla2xxx/qla_os.c
5470
struct srb *sp = e->u.iosb.sp;
drivers/scsi/qla2xxx/qla_os.c
5485
struct qla_work_evt *e, *tmp;
drivers/scsi/qla2xxx/qla_os.c
5494
list_for_each_entry_safe(e, tmp, &work, list) {
drivers/scsi/qla2xxx/qla_os.c
5496
switch (e->type) {
drivers/scsi/qla2xxx/qla_os.c
5499
e->u.aen.code, e->u.aen.data);
drivers/scsi/qla2xxx/qla_os.c
5502
qla81xx_idc_ack(vha, e->u.idc_ack.mb);
drivers/scsi/qla2xxx/qla_os.c
5505
qla2x00_async_login(vha, e->u.logio.fcport,
drivers/scsi/qla2xxx/qla_os.c
5506
e->u.logio.data);
drivers/scsi/qla2xxx/qla_os.c
5509
rc = qla2x00_async_logout(vha, e->u.logio.fcport);
drivers/scsi/qla2xxx/qla_os.c
5512
qla2x00_async_adisc(vha, e->u.logio.fcport,
drivers/scsi/qla2xxx/qla_os.c
5513
e->u.logio.data);
drivers/scsi/qla2xxx/qla_os.c
5516
qla2x00_uevent_emit(vha, e->u.uevent.code);
drivers/scsi/qla2xxx/qla_os.c
5519
qlafx00_process_aen(vha, e);
drivers/scsi/qla2xxx/qla_os.c
5522
qla24xx_sp_unmap(vha, e->u.iosb.sp);
drivers/scsi/qla2xxx/qla_os.c
5528
qla24xx_create_new_sess(vha, e);
drivers/scsi/qla2xxx/qla_os.c
5531
qla24xx_async_gpdb(vha, e->u.fcport.fcport,
drivers/scsi/qla2xxx/qla_os.c
5532
e->u.fcport.opt);
drivers/scsi/qla2xxx/qla_os.c
5535
qla24xx_async_prli(vha, e->u.fcport.fcport);
drivers/scsi/qla2xxx/qla_os.c
5538
qla24xx_async_gpsc(vha, e->u.fcport.fcport);
drivers/scsi/qla2xxx/qla_os.c
5541
qla24xx_async_gnl(vha, e->u.fcport.fcport);
drivers/scsi/qla2xxx/qla_os.c
5544
qla24xx_do_nack_work(vha, e);
drivers/scsi/qla2xxx/qla_os.c
5547
rc = qla2x00_async_prlo(vha, e->u.logio.fcport);
drivers/scsi/qla2xxx/qla_os.c
5550
qla2x00_async_prlo_done(vha, e->u.logio.fcport,
drivers/scsi/qla2xxx/qla_os.c
5551
e->u.logio.data);
drivers/scsi/qla2xxx/qla_os.c
5554
qla_fab_async_scan(vha, e->u.iosb.sp);
drivers/scsi/qla2xxx/qla_os.c
5557
qla_fab_scan_finish(vha, e->u.iosb.sp);
drivers/scsi/qla2xxx/qla_os.c
5560
qla24xx_async_gfpnid(vha, e->u.fcport.fcport);
drivers/scsi/qla2xxx/qla_os.c
5563
qla_sp_retry(vha, e);
drivers/scsi/qla2xxx/qla_os.c
5566
qla_do_iidma_work(vha, e->u.fcport.fcport);
drivers/scsi/qla2xxx/qla_os.c
5570
e->u.fcport.fcport);
drivers/scsi/qla2xxx/qla_os.c
5573
rc = qla24xx_issue_sa_replace_iocb(vha, e);
drivers/scsi/qla2xxx/qla_os.c
5584
list_del_init(&e->list);
drivers/scsi/qla2xxx/qla_os.c
5585
if (e->flags & QLA_EVT_FLAG_FREE)
drivers/scsi/qla2xxx/qla_os.c
5586
kfree(e);
drivers/scsi/qla2xxx/qla_os.c
5595
struct qla_work_evt *e;
drivers/scsi/qla2xxx/qla_os.c
5597
e = qla2x00_alloc_work(vha, QLA_EVT_RELOGIN);
drivers/scsi/qla2xxx/qla_os.c
5599
if (!e) {
drivers/scsi/qla2xxx/qla_os.c
5604
return qla2x00_post_work(vha, e);
drivers/scsi/qla2xxx/qla_target.c
517
struct qla_work_evt *e;
drivers/scsi/qla2xxx/qla_target.c
519
e = qla2x00_alloc_work(vha, QLA_EVT_NACK);
drivers/scsi/qla2xxx/qla_target.c
520
if (!e)
drivers/scsi/qla2xxx/qla_target.c
523
e->u.nack.fcport = fcport;
drivers/scsi/qla2xxx/qla_target.c
524
e->u.nack.type = type;
drivers/scsi/qla2xxx/qla_target.c
525
memcpy(e->u.nack.iocb, ntfy, sizeof(struct imm_ntfy_from_isp));
drivers/scsi/qla2xxx/qla_target.c
526
return qla2x00_post_work(vha, e);
drivers/scsi/qla2xxx/qla_target.c
651
void qla24xx_do_nack_work(struct scsi_qla_host *vha, struct qla_work_evt *e)
drivers/scsi/qla2xxx/qla_target.c
655
switch (e->u.nack.type) {
drivers/scsi/qla2xxx/qla_target.c
657
t = e->u.nack.fcport;
drivers/scsi/qla2xxx/qla_target.c
661
t = qlt_create_sess(vha, e->u.nack.fcport, 0);
drivers/scsi/qla2xxx/qla_target.c
667
vha->hw->tgt.tgt_ops->put_sess(e->u.nack.fcport);
drivers/scsi/qla2xxx/qla_target.c
671
qla24xx_async_notify_ack(vha, e->u.nack.fcport,
drivers/scsi/qla2xxx/qla_target.c
672
(struct imm_ntfy_from_isp *)e->u.nack.iocb, e->u.nack.type);
drivers/scsi/qla4xxx/ql4_os.c
5207
struct qla4_work_evt *e;
drivers/scsi/qla4xxx/ql4_os.c
5210
e = kzalloc(size, GFP_ATOMIC);
drivers/scsi/qla4xxx/ql4_os.c
5211
if (!e)
drivers/scsi/qla4xxx/ql4_os.c
5214
INIT_LIST_HEAD(&e->list);
drivers/scsi/qla4xxx/ql4_os.c
5215
e->type = type;
drivers/scsi/qla4xxx/ql4_os.c
5216
return e;
drivers/scsi/qla4xxx/ql4_os.c
5220
struct qla4_work_evt *e)
drivers/scsi/qla4xxx/ql4_os.c
5225
list_add_tail(&e->list, &ha->work_list);
drivers/scsi/qla4xxx/ql4_os.c
5234
struct qla4_work_evt *e;
drivers/scsi/qla4xxx/ql4_os.c
5236
e = qla4xxx_alloc_work(ha, data_size, QLA4_EVENT_AEN);
drivers/scsi/qla4xxx/ql4_os.c
5237
if (!e)
drivers/scsi/qla4xxx/ql4_os.c
5240
e->u.aen.code = aen_code;
drivers/scsi/qla4xxx/ql4_os.c
5241
e->u.aen.data_size = data_size;
drivers/scsi/qla4xxx/ql4_os.c
5242
memcpy(e->u.aen.data, data, data_size);
drivers/scsi/qla4xxx/ql4_os.c
5244
qla4xxx_post_work(ha, e);
drivers/scsi/qla4xxx/ql4_os.c
5253
struct qla4_work_evt *e;
drivers/scsi/qla4xxx/ql4_os.c
5255
e = qla4xxx_alloc_work(ha, data_size, QLA4_EVENT_PING_STATUS);
drivers/scsi/qla4xxx/ql4_os.c
5256
if (!e)
drivers/scsi/qla4xxx/ql4_os.c
5259
e->u.ping.status = status;
drivers/scsi/qla4xxx/ql4_os.c
5260
e->u.ping.pid = pid;
drivers/scsi/qla4xxx/ql4_os.c
5261
e->u.ping.data_size = data_size;
drivers/scsi/qla4xxx/ql4_os.c
5262
memcpy(e->u.ping.data, data, data_size);
drivers/scsi/qla4xxx/ql4_os.c
5264
qla4xxx_post_work(ha, e);
drivers/scsi/qla4xxx/ql4_os.c
5271
struct qla4_work_evt *e, *tmp;
drivers/scsi/qla4xxx/ql4_os.c
5279
list_for_each_entry_safe(e, tmp, &work, list) {
drivers/scsi/qla4xxx/ql4_os.c
5280
list_del_init(&e->list);
drivers/scsi/qla4xxx/ql4_os.c
5282
switch (e->type) {
drivers/scsi/qla4xxx/ql4_os.c
5286
e->u.aen.code,
drivers/scsi/qla4xxx/ql4_os.c
5287
e->u.aen.data_size,
drivers/scsi/qla4xxx/ql4_os.c
5288
e->u.aen.data);
drivers/scsi/qla4xxx/ql4_os.c
5293
e->u.ping.status,
drivers/scsi/qla4xxx/ql4_os.c
5294
e->u.ping.pid,
drivers/scsi/qla4xxx/ql4_os.c
5295
e->u.ping.data_size,
drivers/scsi/qla4xxx/ql4_os.c
5296
e->u.ping.data);
drivers/scsi/qla4xxx/ql4_os.c
5300
"supported", e->type);
drivers/scsi/qla4xxx/ql4_os.c
5302
kfree(e);
drivers/scsi/scsi_proc.c
104
struct scsi_proc_entry *e;
drivers/scsi/scsi_proc.c
108
list_for_each_entry(e, &scsi_proc_list, entry)
drivers/scsi/scsi_proc.c
109
if (e->sht == sht)
drivers/scsi/scsi_proc.c
110
return e;
drivers/scsi/scsi_proc.c
118
struct scsi_proc_entry *e;
drivers/scsi/scsi_proc.c
121
e = __scsi_lookup_proc_entry(sht);
drivers/scsi/scsi_proc.c
124
return e;
drivers/scsi/scsi_proc.c
134
struct scsi_proc_entry *e = scsi_lookup_proc_entry(sht);
drivers/scsi/scsi_proc.c
136
return e ? e->proc_dir : NULL;
drivers/scsi/scsi_proc.c
156
struct scsi_proc_entry *e;
drivers/scsi/scsi_proc.c
163
e = __scsi_lookup_proc_entry(sht);
drivers/scsi/scsi_proc.c
164
if (!e) {
drivers/scsi/scsi_proc.c
165
e = kzalloc_obj(*e);
drivers/scsi/scsi_proc.c
166
if (!e) {
drivers/scsi/scsi_proc.c
171
if (e->present++)
drivers/scsi/scsi_proc.c
173
e->proc_dir = proc_mkdir(sht->proc_name, proc_scsi);
drivers/scsi/scsi_proc.c
174
if (!e->proc_dir) {
drivers/scsi/scsi_proc.c
180
e->sht = sht;
drivers/scsi/scsi_proc.c
181
list_add_tail(&e->entry, &scsi_proc_list);
drivers/scsi/scsi_proc.c
183
e = NULL;
drivers/scsi/scsi_proc.c
188
kfree(e);
drivers/scsi/scsi_proc.c
198
struct scsi_proc_entry *e;
drivers/scsi/scsi_proc.c
204
e = __scsi_lookup_proc_entry(sht);
drivers/scsi/scsi_proc.c
205
if (e && !--e->present) {
drivers/scsi/scsi_proc.c
207
list_del(&e->entry);
drivers/scsi/scsi_proc.c
208
kfree(e);
drivers/scsi/scsi_proc.c
221
struct scsi_proc_entry *e;
drivers/scsi/scsi_proc.c
228
e = scsi_lookup_proc_entry(sht);
drivers/scsi/scsi_proc.c
229
if (!e)
drivers/scsi/scsi_proc.c
233
p = proc_create_data(name, S_IRUGO | S_IWUSR, e->proc_dir,
drivers/scsi/scsi_proc.c
242
e ? "proc_create_data()" : "scsi_proc_hostdir_add()");
drivers/scsi/scsi_proc.c
252
struct scsi_proc_entry *e;
drivers/scsi/scsi_proc.c
258
e = scsi_lookup_proc_entry(sht);
drivers/scsi/scsi_proc.c
259
if (!e)
drivers/scsi/scsi_proc.c
263
remove_proc_entry(name, e->proc_dir);
drivers/scsi/vmw_pvscsi.c
1034
const struct PVSCSIRingMsgDesc *e)
drivers/scsi/vmw_pvscsi.c
1041
e->type, s->msgProdIdx, s->msgConsIdx, s->msgNumEntriesLog2);
drivers/scsi/vmw_pvscsi.c
1045
if (e->type == PVSCSI_MSG_DEV_ADDED) {
drivers/scsi/vmw_pvscsi.c
1047
desc = (struct PVSCSIMsgDescDevStatusChanged *)e;
drivers/scsi/vmw_pvscsi.c
1066
} else if (e->type == PVSCSI_MSG_DEV_REMOVED) {
drivers/scsi/vmw_pvscsi.c
1068
desc = (struct PVSCSIMsgDescDevStatusChanged *)e;
drivers/scsi/vmw_pvscsi.c
1105
struct PVSCSIRingMsgDesc *e = ring + (s->msgConsIdx &
drivers/scsi/vmw_pvscsi.c
1109
pvscsi_process_msg(adapter, e);
drivers/scsi/vmw_pvscsi.c
349
struct PVSCSIRingReqDesc *e)
drivers/scsi/vmw_pvscsi.c
355
e->dataLen = bufflen;
drivers/scsi/vmw_pvscsi.c
356
e->dataAddr = 0;
drivers/scsi/vmw_pvscsi.c
372
e->flags |= PVSCSI_FLAG_CMD_WITH_SG_LIST;
drivers/scsi/vmw_pvscsi.c
382
e->dataAddr = ctx->sglPA;
drivers/scsi/vmw_pvscsi.c
384
e->dataAddr = sg_dma_address(sg);
drivers/scsi/vmw_pvscsi.c
397
e->dataAddr = ctx->dataPA;
drivers/scsi/vmw_pvscsi.c
544
const struct PVSCSIRingCmpDesc *e)
drivers/scsi/vmw_pvscsi.c
549
u32 btstat = e->hostStatus;
drivers/scsi/vmw_pvscsi.c
550
u32 sdstat = e->scsiStatus;
drivers/scsi/vmw_pvscsi.c
552
ctx = pvscsi_get_context(adapter, e->context);
drivers/scsi/vmw_pvscsi.c
593
if (e->dataLen && (e->dataLen < scsi_bufflen(cmd)))
drivers/scsi/vmw_pvscsi.c
594
scsi_set_resid(cmd, scsi_bufflen(cmd) - e->dataLen);
drivers/scsi/vmw_pvscsi.c
601
scsi_set_resid(cmd, scsi_bufflen(cmd) - e->dataLen);
drivers/scsi/vmw_pvscsi.c
666
struct PVSCSIRingCmpDesc *e = ring + (s->cmpConsIdx &
drivers/scsi/vmw_pvscsi.c
675
pvscsi_complete_request(adapter, e);
drivers/scsi/vmw_pvscsi.c
694
struct PVSCSIRingReqDesc *e;
drivers/scsi/vmw_pvscsi.c
717
e = adapter->req_ring + (s->reqProdIdx & MASK(req_entries));
drivers/scsi/vmw_pvscsi.c
719
e->bus = sdev->channel;
drivers/scsi/vmw_pvscsi.c
720
e->target = sdev->id;
drivers/scsi/vmw_pvscsi.c
721
memset(e->lun, 0, sizeof(e->lun));
drivers/scsi/vmw_pvscsi.c
722
e->lun[1] = sdev->lun;
drivers/scsi/vmw_pvscsi.c
734
e->senseAddr = ctx->sensePA;
drivers/scsi/vmw_pvscsi.c
735
e->senseLen = SCSI_SENSE_BUFFERSIZE;
drivers/scsi/vmw_pvscsi.c
737
e->senseLen = 0;
drivers/scsi/vmw_pvscsi.c
738
e->senseAddr = 0;
drivers/scsi/vmw_pvscsi.c
740
e->cdbLen = cmd->cmd_len;
drivers/scsi/vmw_pvscsi.c
741
e->vcpuHint = smp_processor_id();
drivers/scsi/vmw_pvscsi.c
742
memcpy(e->cdb, cmd->cmnd, e->cdbLen);
drivers/scsi/vmw_pvscsi.c
744
e->tag = SIMPLE_QUEUE_TAG;
drivers/scsi/vmw_pvscsi.c
747
e->flags = PVSCSI_FLAG_CMD_DIR_TOHOST;
drivers/scsi/vmw_pvscsi.c
749
e->flags = PVSCSI_FLAG_CMD_DIR_TODEVICE;
drivers/scsi/vmw_pvscsi.c
751
e->flags = PVSCSI_FLAG_CMD_DIR_NONE;
drivers/scsi/vmw_pvscsi.c
753
e->flags = 0;
drivers/scsi/vmw_pvscsi.c
755
if (pvscsi_map_buffers(adapter, ctx, cmd, e) != 0) {
drivers/scsi/vmw_pvscsi.c
765
e->context = pvscsi_map_context(adapter, ctx);
drivers/soc/fsl/qbman/bman.c
287
static int rcr_ptr2idx(struct bm_rcr_entry *e)
drivers/soc/fsl/qbman/bman.c
289
return ((uintptr_t)e >> RCR_SHIFT) & (BM_RCR_SIZE - 1);
drivers/soc/fsl/qbman/qman.c
1321
const union qm_mr_entry *e = qm_mr_current(p);
drivers/soc/fsl/qbman/qman.c
1324
e->verb, e->ern.rc, qm_fd_addr_get64(&e->ern.fd));
drivers/soc/fsl/qbman/qman.c
405
static int eqcr_ptr2idx(struct qm_eqcr_entry *e)
drivers/soc/fsl/qbman/qman.c
407
return ((uintptr_t)e >> EQCR_SHIFT) & (QM_EQCR_SIZE - 1);
drivers/soc/fsl/qbman/qman.c
585
static inline int dqrr_ptr2idx(const struct qm_dqrr_entry *e)
drivers/soc/fsl/qbman/qman.c
587
return ((uintptr_t)e >> DQRR_SHIFT) & (QM_DQRR_SIZE - 1);
drivers/soc/fsl/qbman/qman.c
590
static const struct qm_dqrr_entry *dqrr_inc(const struct qm_dqrr_entry *e)
drivers/soc/fsl/qbman/qman.c
592
return dqrr_carryclear(e + 1);
drivers/soc/fsl/qbman/qman.c
752
static inline int mr_ptr2idx(const union qm_mr_entry *e)
drivers/soc/fsl/qbman/qman.c
754
return ((uintptr_t)e >> MR_SHIFT) & (QM_MR_SIZE - 1);
drivers/soc/fsl/qbman/qman.c
757
static inline union qm_mr_entry *mr_inc(union qm_mr_entry *e)
drivers/soc/fsl/qbman/qman.c
759
return mr_carryclear(e + 1);
drivers/soc/qcom/smem.c
304
struct smem_private_entry *e;
drivers/soc/qcom/smem.c
306
return p + le32_to_cpu(phdr->size) - ALIGN(sizeof(*e), cacheline);
drivers/soc/qcom/smem.c
326
uncached_entry_next(struct smem_private_entry *e)
drivers/soc/qcom/smem.c
328
void *p = e;
drivers/soc/qcom/smem.c
330
return p + sizeof(*e) + le16_to_cpu(e->padding_hdr) +
drivers/soc/qcom/smem.c
331
le32_to_cpu(e->size);
drivers/soc/qcom/smem.c
335
cached_entry_next(struct smem_private_entry *e, size_t cacheline)
drivers/soc/qcom/smem.c
337
void *p = e;
drivers/soc/qcom/smem.c
339
return p - le32_to_cpu(e->size) - ALIGN(sizeof(*e), cacheline);
drivers/soc/qcom/smem.c
342
static void *uncached_entry_to_item(struct smem_private_entry *e)
drivers/soc/qcom/smem.c
344
void *p = e;
drivers/soc/qcom/smem.c
346
return p + sizeof(*e) + le16_to_cpu(e->padding_hdr);
drivers/soc/qcom/smem.c
349
static void *cached_entry_to_item(struct smem_private_entry *e)
drivers/soc/qcom/smem.c
351
void *p = e;
drivers/soc/qcom/smem.c
353
return p - le32_to_cpu(e->size);
drivers/soc/qcom/smem.c
593
struct smem_private_entry *e, *end;
drivers/soc/qcom/smem.c
602
e = phdr_to_first_uncached_entry(phdr);
drivers/soc/qcom/smem.c
605
while (e < end) {
drivers/soc/qcom/smem.c
606
if (e->canary != SMEM_PRIVATE_CANARY)
drivers/soc/qcom/smem.c
609
if (le16_to_cpu(e->item) == item) {
drivers/soc/qcom/smem.c
611
e_size = le32_to_cpu(e->size);
drivers/soc/qcom/smem.c
612
padding_data = le16_to_cpu(e->padding_data);
drivers/soc/qcom/smem.c
620
item_ptr = uncached_entry_to_item(e);
drivers/soc/qcom/smem.c
627
e = uncached_entry_next(e);
drivers/soc/qcom/smem.c
630
if (WARN_ON((void *)e > p_end))
drivers/soc/qcom/smem.c
635
e = phdr_to_first_cached_entry(phdr, part->cacheline);
drivers/soc/qcom/smem.c
638
if (WARN_ON((void *)e < (void *)phdr || (void *)end > p_end))
drivers/soc/qcom/smem.c
641
while (e > end) {
drivers/soc/qcom/smem.c
642
if (e->canary != SMEM_PRIVATE_CANARY)
drivers/soc/qcom/smem.c
645
if (le16_to_cpu(e->item) == item) {
drivers/soc/qcom/smem.c
647
e_size = le32_to_cpu(e->size);
drivers/soc/qcom/smem.c
648
padding_data = le16_to_cpu(e->padding_data);
drivers/soc/qcom/smem.c
656
item_ptr = cached_entry_to_item(e);
drivers/soc/qcom/smem.c
663
e = cached_entry_next(e, part->cacheline);
drivers/soc/qcom/smem.c
666
if (WARN_ON((void *)e < (void *)phdr))
drivers/ssb/driver_chipcommon_pmu.c
100
if (!e)
drivers/ssb/driver_chipcommon_pmu.c
101
e = pmu0_plltab_find_entry(SSB_PMU0_DEFAULT_XTALFREQ);
drivers/ssb/driver_chipcommon_pmu.c
102
BUG_ON(!e);
drivers/ssb/driver_chipcommon_pmu.c
103
crystalfreq = e->freq;
drivers/ssb/driver_chipcommon_pmu.c
104
cc->pmu.crystalfreq = e->freq;
drivers/ssb/driver_chipcommon_pmu.c
108
if (((pmuctl & SSB_CHIPCO_PMU_CTL_XTALFREQ) >> SSB_CHIPCO_PMU_CTL_XTALFREQ_SHIFT) == e->xf) {
drivers/ssb/driver_chipcommon_pmu.c
155
pllctl |= ((u32)e->wb_int << SSB_PMU0_PLLCTL1_WILD_IMSK_SHIFT) & SSB_PMU0_PLLCTL1_WILD_IMSK;
drivers/ssb/driver_chipcommon_pmu.c
156
pllctl |= ((u32)e->wb_frac << SSB_PMU0_PLLCTL1_WILD_FMSK_SHIFT) & SSB_PMU0_PLLCTL1_WILD_FMSK;
drivers/ssb/driver_chipcommon_pmu.c
157
if (e->wb_frac == 0)
drivers/ssb/driver_chipcommon_pmu.c
164
pllctl |= (((u32)e->wb_int >> 4) << SSB_PMU0_PLLCTL2_WILD_IMSKHI_SHIFT) & SSB_PMU0_PLLCTL2_WILD_IMSKHI;
drivers/ssb/driver_chipcommon_pmu.c
173
pmuctl |= ((u32)e->xf << SSB_CHIPCO_PMU_CTL_XTALFREQ_SHIFT) & SSB_CHIPCO_PMU_CTL_XTALFREQ;
drivers/ssb/driver_chipcommon_pmu.c
208
const struct pmu1_plltab_entry *e;
drivers/ssb/driver_chipcommon_pmu.c
212
e = &pmu1_plltab[i];
drivers/ssb/driver_chipcommon_pmu.c
213
if (e->freq == crystalfreq)
drivers/ssb/driver_chipcommon_pmu.c
214
return e;
drivers/ssb/driver_chipcommon_pmu.c
225
const struct pmu1_plltab_entry *e = NULL;
drivers/ssb/driver_chipcommon_pmu.c
238
e = pmu1_plltab_find_entry(crystalfreq);
drivers/ssb/driver_chipcommon_pmu.c
239
if (!e)
drivers/ssb/driver_chipcommon_pmu.c
240
e = pmu1_plltab_find_entry(SSB_PMU1_DEFAULT_XTALFREQ);
drivers/ssb/driver_chipcommon_pmu.c
241
BUG_ON(!e);
drivers/ssb/driver_chipcommon_pmu.c
242
crystalfreq = e->freq;
drivers/ssb/driver_chipcommon_pmu.c
243
cc->pmu.crystalfreq = e->freq;
drivers/ssb/driver_chipcommon_pmu.c
247
if (((pmuctl & SSB_CHIPCO_PMU_CTL_XTALFREQ) >> SSB_CHIPCO_PMU_CTL_XTALFREQ_SHIFT) == e->xf) {
drivers/ssb/driver_chipcommon_pmu.c
283
pllctl |= ((u32)e->p1div << SSB_PMU1_PLLCTL0_P1DIV_SHIFT) & SSB_PMU1_PLLCTL0_P1DIV;
drivers/ssb/driver_chipcommon_pmu.c
284
pllctl |= ((u32)e->p2div << SSB_PMU1_PLLCTL0_P2DIV_SHIFT) & SSB_PMU1_PLLCTL0_P2DIV;
drivers/ssb/driver_chipcommon_pmu.c
290
pllctl |= ((u32)e->ndiv_int << SSB_PMU1_PLLCTL2_NDIVINT_SHIFT) & SSB_PMU1_PLLCTL2_NDIVINT;
drivers/ssb/driver_chipcommon_pmu.c
297
pllctl |= ((u32)e->ndiv_frac << SSB_PMU1_PLLCTL3_NDIVFRAC_SHIFT) & SSB_PMU1_PLLCTL3_NDIVFRAC;
drivers/ssb/driver_chipcommon_pmu.c
311
pmuctl |= ((((u32)e->freq + 127) / 128 - 1) << SSB_CHIPCO_PMU_CTL_ILP_DIV_SHIFT)
drivers/ssb/driver_chipcommon_pmu.c
313
pmuctl |= ((u32)e->xf << SSB_CHIPCO_PMU_CTL_XTALFREQ_SHIFT) & SSB_CHIPCO_PMU_CTL_XTALFREQ;
drivers/ssb/driver_chipcommon_pmu.c
622
const struct pmu0_plltab_entry *e = NULL;
drivers/ssb/driver_chipcommon_pmu.c
626
e = pmu0_plltab_find_entry(crystalfreq);
drivers/ssb/driver_chipcommon_pmu.c
627
BUG_ON(!e);
drivers/ssb/driver_chipcommon_pmu.c
628
return e->freq * 1000;
drivers/ssb/driver_chipcommon_pmu.c
77
const struct pmu0_plltab_entry *e;
drivers/ssb/driver_chipcommon_pmu.c
81
e = &pmu0_plltab[i];
drivers/ssb/driver_chipcommon_pmu.c
82
if (e->freq == crystalfreq)
drivers/ssb/driver_chipcommon_pmu.c
83
return e;
drivers/ssb/driver_chipcommon_pmu.c
94
const struct pmu0_plltab_entry *e = NULL;
drivers/ssb/driver_chipcommon_pmu.c
99
e = pmu0_plltab_find_entry(crystalfreq);
drivers/ssb/driver_chipcommon_sflash.c
107
for (e = ssb_sflash_sst_tbl; e->name; e++) {
drivers/ssb/driver_chipcommon_sflash.c
108
if (e->id == id2)
drivers/ssb/driver_chipcommon_sflash.c
115
for (e = ssb_sflash_st_tbl; e->name; e++) {
drivers/ssb/driver_chipcommon_sflash.c
116
if (e->id == id)
drivers/ssb/driver_chipcommon_sflash.c
121
if (!e->name) {
drivers/ssb/driver_chipcommon_sflash.c
132
for (e = ssb_sflash_at_tbl; e->name; e++) {
drivers/ssb/driver_chipcommon_sflash.c
133
if (e->id == id)
drivers/ssb/driver_chipcommon_sflash.c
136
if (!e->name) {
drivers/ssb/driver_chipcommon_sflash.c
149
sflash->blocksize = e->blocksize;
drivers/ssb/driver_chipcommon_sflash.c
150
sflash->numblocks = e->numblocks;
drivers/ssb/driver_chipcommon_sflash.c
155
e->name, sflash->size / 1024, e->blocksize, e->numblocks);
drivers/ssb/driver_chipcommon_sflash.c
90
const struct ssb_sflash_tbl_e *e;
drivers/staging/greybus/audio_topology.c
548
struct soc_enum *e = (struct soc_enum *)kcontrol->private_value;
drivers/staging/greybus/audio_topology.c
580
if (e->shift_l != e->shift_r)
drivers/staging/greybus/audio_topology.c
593
struct soc_enum *e = (struct soc_enum *)kcontrol->private_value;
drivers/staging/greybus/audio_topology.c
606
if (ucontrol->value.enumerated.item[0] > e->items - 1)
drivers/staging/greybus/audio_topology.c
611
if (e->shift_l != e->shift_r) {
drivers/staging/greybus/audio_topology.c
612
if (ucontrol->value.enumerated.item[1] > e->items - 1)
drivers/staging/greybus/audio_topology.c
717
struct soc_enum *e = (struct soc_enum *)kcontrol->private_value;
drivers/staging/greybus/audio_topology.c
746
if (e->shift_l != e->shift_r)
drivers/staging/greybus/audio_topology.c
764
struct soc_enum *e = (struct soc_enum *)kcontrol->private_value;
drivers/staging/greybus/audio_topology.c
767
if (ucontrol->value.enumerated.item[0] > e->items - 1)
drivers/staging/greybus/audio_topology.c
797
val = mux << e->shift_l;
drivers/staging/greybus/audio_topology.c
806
if (e->shift_l != e->shift_r) {
drivers/staging/greybus/audio_topology.c
807
if (ucontrol->value.enumerated.item[1] > e->items - 1)
drivers/staging/greybus/audio_topology.c
809
val |= ucontrol->value.enumerated.item[1] << e->shift_r;
drivers/staging/greybus/audio_topology.c
836
val, e, NULL);
drivers/staging/media/ipu3/ipu3-css.c
2200
unsigned int e = IPU3_CSS_RECT_ENVELOPE;
drivers/staging/media/ipu3/ipu3-css.c
2213
css_pipe->rect[e].width,
drivers/staging/media/ipu3/ipu3-css.c
2214
css_pipe->rect[e].height);
drivers/staging/rtl8723bs/core/rtw_security.c
1389
size_t i, e, left, total_len;
drivers/staging/rtl8723bs/core/rtw_security.c
1398
for (e = 0; e < num_elem; e++)
drivers/staging/rtl8723bs/core/rtw_security.c
1399
total_len += len[e];
drivers/staging/rtl8723bs/core/rtw_security.c
1402
e = 0;
drivers/staging/rtl8723bs/core/rtw_security.c
1410
e++;
drivers/staging/rtl8723bs/core/rtw_security.c
1411
pos = addr[e];
drivers/staging/rtl8723bs/core/rtw_security.c
1412
end = pos + len[e];
drivers/staging/rtl8723bs/core/rtw_security.c
1428
e++;
drivers/staging/rtl8723bs/core/rtw_security.c
1429
pos = addr[e];
drivers/staging/rtl8723bs/core/rtw_security.c
1430
end = pos + len[e];
drivers/tty/vt/selection.c
55
static inline void highlight(const int s, const int e)
drivers/tty/vt/selection.c
57
invert_screen(vc_sel.cons, s, e-s+2, true);
drivers/ufs/core/ufs-debugfs.c
37
struct ufs_event_hist *e = hba->ufs_stats.event;
drivers/ufs/core/ufs-debugfs.c
40
seq_printf(s, fmt, e[UFS_EVT_ ## typ].cnt)
drivers/ufs/core/ufshcd.c
5019
struct ufs_event_hist *e;
drivers/ufs/core/ufshcd.c
5024
e = &hba->ufs_stats.event[id];
drivers/ufs/core/ufshcd.c
5025
e->val[e->pos] = val;
drivers/ufs/core/ufshcd.c
5026
e->tstamp[e->pos] = local_clock();
drivers/ufs/core/ufshcd.c
5027
e->cnt += 1;
drivers/ufs/core/ufshcd.c
5028
e->pos = (e->pos + 1) % UFS_EVENT_HIST_LENGTH;
drivers/ufs/core/ufshcd.c
551
const struct ufs_event_hist *e;
drivers/ufs/core/ufshcd.c
556
e = &hba->ufs_stats.event[id];
drivers/ufs/core/ufshcd.c
559
int p = (i + e->pos) % UFS_EVENT_HIST_LENGTH;
drivers/ufs/core/ufshcd.c
561
if (e->tstamp[p] == 0)
drivers/ufs/core/ufshcd.c
564
e->val[p], div_u64(e->tstamp[p], 1000));
drivers/ufs/core/ufshcd.c
571
dev_err(hba->dev, "%s: total cnt=%llu\n", err_name, e->cnt);
drivers/usb/core/devio.c
858
unsigned int i, j, e;
drivers/usb/core/devio.c
871
for (e = 0; e < alts->desc.bNumEndpoints; e++) {
drivers/usb/core/devio.c
872
endpt = &alts->endpoint[e].desc;
drivers/usb/early/ehci-dbgp.c
832
char *e;
drivers/usb/early/ehci-dbgp.c
841
dbgp_num = simple_strtoul(s, &e, 10);
drivers/usb/early/xhci-dbc.c
605
char *e;
drivers/usb/early/xhci-dbc.c
617
dbgp_num = simple_strtoul(s, &e, 10);
drivers/usb/early/xhci-dbc.c
618
if (s == e)
drivers/usb/fotg210/fotg210-hcd.c
1337
unsigned e;
drivers/usb/fotg210/fotg210-hcd.c
1350
for_each_set_bit(e, &events, FOTG210_HRTIMER_NUM_EVENTS) {
drivers/usb/fotg210/fotg210-hcd.c
1351
if (ktime_compare(now, fotg210->hr_timeouts[e]) >= 0)
drivers/usb/fotg210/fotg210-hcd.c
1352
event_handlers[e](fotg210);
drivers/usb/fotg210/fotg210-hcd.c
1354
fotg210_enable_event(fotg210, e, false);
drivers/usb/fotg210/fotg210-hcd.h
634
#define fotg210_has_fsl_portno_bug(e) (0)
drivers/usb/fotg210/fotg210-hcd.h
647
#define fotg210_big_endian_mmio(e) 0
drivers/usb/fotg210/fotg210-hcd.h
648
#define fotg210_big_endian_capbase(e) 0
drivers/usb/gadget/legacy/hid.c
100
list_for_each_entry(e, &hidg_func_list, node) {
drivers/usb/gadget/legacy/hid.c
101
e->f = usb_get_function(e->fi);
drivers/usb/gadget/legacy/hid.c
102
if (IS_ERR(e->f)) {
drivers/usb/gadget/legacy/hid.c
103
status = PTR_ERR(e->f);
drivers/usb/gadget/legacy/hid.c
106
status = usb_add_function(c, e->f);
drivers/usb/gadget/legacy/hid.c
108
usb_put_function(e->f);
drivers/usb/gadget/legacy/hid.c
116
if (n == e)
drivers/usb/gadget/legacy/hid.c
242
struct hidg_func_node *e, *n;
drivers/usb/gadget/legacy/hid.c
244
list_for_each_entry_safe(e, n, &hidg_func_list, node) {
drivers/usb/gadget/legacy/hid.c
245
list_del(&e->node);
drivers/usb/gadget/legacy/hid.c
246
kfree(e);
drivers/usb/gadget/legacy/hid.c
92
struct hidg_func_node *e, *n;
drivers/usb/gadget/udc/aspeed_udc.c
247
#define EP_DBG(e, fmt, ...) \
drivers/usb/gadget/udc/aspeed_udc.c
248
dev_dbg(&(e)->udc->pdev->dev, "%s():%s " fmt, __func__, \
drivers/usb/gadget/udc/aspeed_udc.c
249
(e)->ep.name, ##__VA_ARGS__)
drivers/usb/gadget/udc/bdc/bdc.h
262
#define to_bdc_ep(e) container_of(e, struct bdc_ep, usb_ep)
drivers/usb/gadget/udc/lpc32xx_udc.c
271
#define USBD_RX_EP_SEL(e) (1 << ((e) << 1))
drivers/usb/gadget/udc/lpc32xx_udc.c
274
#define USBD_TX_EP_SEL(e) (1 << (((e) << 1) + 1))
drivers/usb/gadget/udc/lpc32xx_udc.c
285
#define USBD_EP_SEL(e) (1 << (e))
drivers/usb/gadget/udc/lpc32xx_udc.c
304
#define USBD_LOG_ENDPOINT(e) ((e) << 2)
drivers/usb/gadget/udc/lpc32xx_udc.c
317
#define USBD_DMAEP(e) (1 << (e))
drivers/usb/gadget/udc/max3420_udc.c
186
#define to_max3420_ep(e) container_of((e), struct max3420_ep, ep_usb)
drivers/usb/gadget/udc/net2280.c
2866
struct net2280_ep *e;
drivers/usb/gadget/udc/net2280.c
2894
e = get_ep_by_addr(dev, w_index);
drivers/usb/gadget/udc/net2280.c
2895
if (!e)
drivers/usb/gadget/udc/net2280.c
2897
status = readl(&e->regs->ep_rsp) &
drivers/usb/gadget/udc/net2280.c
2954
e = get_ep_by_addr(dev, w_index);
drivers/usb/gadget/udc/net2280.c
2955
if (!e)
drivers/usb/gadget/udc/net2280.c
2959
ep_vdbg(dev, "%s clear halt\n", e->ep.name);
drivers/usb/gadget/udc/net2280.c
2964
ep_clear_seqnum(e);
drivers/usb/gadget/udc/net2280.c
2965
clear_halt(e);
drivers/usb/gadget/udc/net2280.c
2966
if (!list_empty(&e->queue) && e->td_dma)
drivers/usb/gadget/udc/net2280.c
2967
restart_dma(e);
drivers/usb/gadget/udc/net2280.c
3020
e = get_ep_by_addr(dev, w_index);
drivers/usb/gadget/udc/net2280.c
3021
if (!e || (w_value != USB_ENDPOINT_HALT))
drivers/usb/gadget/udc/net2280.c
3214
struct net2280_ep *e;
drivers/usb/gadget/udc/net2280.c
3220
e = get_ep_by_addr(dev, w_index);
drivers/usb/gadget/udc/net2280.c
3221
if (!e || w_length > 2)
drivers/usb/gadget/udc/net2280.c
3224
if (readl(&e->regs->ep_rsp) & BIT(SET_ENDPOINT_HALT))
drivers/usb/gadget/udc/net2280.c
3239
struct net2280_ep *e;
drivers/usb/gadget/udc/net2280.c
3246
e = get_ep_by_addr(dev, w_index);
drivers/usb/gadget/udc/net2280.c
3247
if (!e)
drivers/usb/gadget/udc/net2280.c
3249
if (e->wedged) {
drivers/usb/gadget/udc/net2280.c
3253
ep_vdbg(dev, "%s clear halt\n", e->ep.name);
drivers/usb/gadget/udc/net2280.c
3254
clear_halt(e);
drivers/usb/gadget/udc/net2280.c
3256
!list_empty(&e->queue) && e->td_dma)
drivers/usb/gadget/udc/net2280.c
3257
restart_dma(e);
drivers/usb/gadget/udc/net2280.c
3264
struct net2280_ep *e;
drivers/usb/gadget/udc/net2280.c
3271
e = get_ep_by_addr(dev, w_index);
drivers/usb/gadget/udc/net2280.c
3272
if (!e)
drivers/usb/gadget/udc/net2280.c
3274
if (e->ep.name == ep0name)
drivers/usb/gadget/udc/net2280.c
3276
set_halt(e);
drivers/usb/gadget/udc/net2280.c
3277
if ((dev->quirks & PLX_PCIE) && e->dma)
drivers/usb/gadget/udc/net2280.c
3278
abort_dma(e);
drivers/usb/host/ehci-timer.c
404
unsigned e;
drivers/usb/host/ehci-timer.c
417
for_each_set_bit(e, &events, EHCI_HRTIMER_NUM_EVENTS) {
drivers/usb/host/ehci-timer.c
418
if (ktime_compare(now, ehci->hr_timeouts[e]) >= 0)
drivers/usb/host/ehci-timer.c
419
event_handlers[e](ehci);
drivers/usb/host/ehci-timer.c
421
ehci_enable_event(ehci, e, false);
drivers/usb/host/ehci.h
661
#define ehci_is_TDI(e) (ehci_to_hcd(e)->has_tt)
drivers/usb/host/ehci.h
683
#define ehci_is_TDI(e) (0)
drivers/usb/host/ehci.h
694
#define ehci_has_fsl_portno_bug(e) ((e)->has_fsl_port_bug)
drivers/usb/host/ehci.h
696
#define ehci_has_fsl_portno_bug(e) (0)
drivers/usb/host/ehci.h
705
#define ehci_has_fsl_hs_errata(e) ((e)->has_fsl_hs_errata)
drivers/usb/host/ehci.h
707
#define ehci_has_fsl_hs_errata(e) (0)
drivers/usb/host/ehci.h
715
#define ehci_has_fsl_susp_errata(e) ((e)->has_fsl_susp_errata)
drivers/usb/host/ehci.h
722
#define ehci_has_ci_pec_bug(e, portsc) \
drivers/usb/host/ehci.h
723
((e)->has_ci_pec_bug && ((e)->command & CMD_PSE) \
drivers/usb/host/ehci.h
741
#define ehci_big_endian_mmio(e) ((e)->big_endian_mmio)
drivers/usb/host/ehci.h
742
#define ehci_big_endian_capbase(e) ((e)->big_endian_capbase)
drivers/usb/host/ehci.h
744
#define ehci_big_endian_mmio(e) 0
drivers/usb/host/ehci.h
745
#define ehci_big_endian_capbase(e) 0
drivers/usb/host/ehci.h
830
#define ehci_big_endian_desc(e) ((e)->big_endian_desc)
drivers/usb/misc/usbtest.c
131
struct usb_host_endpoint *e)
drivers/usb/misc/usbtest.c
135
*in = e;
drivers/usb/misc/usbtest.c
138
*out = e;
drivers/usb/misc/usbtest.c
168
struct usb_host_endpoint *e;
drivers/usb/misc/usbtest.c
171
e = alt->endpoint + ep;
drivers/usb/misc/usbtest.c
172
edi = usb_endpoint_dir_in(&e->desc);
drivers/usb/misc/usbtest.c
174
switch (usb_endpoint_type(&e->desc)) {
drivers/usb/misc/usbtest.c
176
endpoint_update(edi, &in, &out, e);
drivers/usb/misc/usbtest.c
180
endpoint_update(edi, &int_in, &int_out, e);
drivers/usb/misc/usbtest.c
184
endpoint_update(edi, &iso_in, &iso_out, e);
drivers/usb/musb/musb_gadget_ep0.c
1002
ep = to_musb_ep(e);
drivers/usb/musb/musb_gadget_ep0.c
896
static int musb_g_ep0_disable(struct usb_ep *e)
drivers/usb/musb/musb_gadget_ep0.c
903
musb_g_ep0_queue(struct usb_ep *e, struct usb_request *r, gfp_t gfp_flags)
drivers/usb/musb/musb_gadget_ep0.c
912
if (!e || !r)
drivers/usb/musb/musb_gadget_ep0.c
915
ep = to_musb_ep(e);
drivers/usb/musb/musb_gadget_ep0.c
990
static int musb_g_ep0_halt(struct usb_ep *e, int value)
drivers/usb/musb/musb_gadget_ep0.c
999
if (!e || !value)
drivers/usb/renesas_usbhs/fifo.c
216
#define usbhsf_irq_empty_ctrl(p, e) usbhsf_irq_callback_ctrl(p, irq_bempsts, e)
drivers/usb/renesas_usbhs/fifo.c
217
#define usbhsf_irq_ready_ctrl(p, e) usbhsf_irq_callback_ctrl(p, irq_brdysts, e)
drivers/usb/renesas_usbhs/mod_gadget.c
93
#define usbhsg_ep_to_uep(e) container_of(e, struct usbhsg_uep, ep)
drivers/usb/typec/tcpm/tcpm.c
193
#define GENERATE_ENUM(e) e
drivers/vdpa/mlx5/core/mr.c
150
struct mlx5_vdpa_direct_mr *e;
drivers/vdpa/mlx5/core/mr.c
156
e = list_last_entry(&mkey->head, struct mlx5_vdpa_direct_mr, list);
drivers/vdpa/mlx5/core/mr.c
158
return e->end - s->start;
drivers/video/fbdev/i810/i810-i2c.c
164
const u8 *e = fb_firmware_edid(info->device);
drivers/video/fbdev/i810/i810-i2c.c
166
if (e != NULL) {
drivers/video/fbdev/i810/i810-i2c.c
168
edid = kmemdup(e, EDID_LENGTH, GFP_KERNEL);
drivers/video/fbdev/nvidia/nv_i2c.c
162
const u8 *e = fb_firmware_edid(info->device);
drivers/video/fbdev/nvidia/nv_i2c.c
164
if (e != NULL)
drivers/video/fbdev/nvidia/nv_i2c.c
165
edid = kmemdup(e, EDID_LENGTH, GFP_KERNEL);
drivers/video/fbdev/omap2/omapfb/dss/apply.c
1449
bool e;
drivers/video/fbdev/omap2/omapfb/dss/apply.c
1453
e = op->enabled;
drivers/video/fbdev/omap2/omapfb/dss/apply.c
1457
return e;
drivers/video/fbdev/omap2/omapfb/dss/dsi.c
1125
u32 e;
drivers/video/fbdev/omap2/omapfb/dss/dsi.c
1127
e = dsi->errors;
drivers/video/fbdev/omap2/omapfb/dss/dsi.c
1130
return e;
drivers/video/fbdev/savage/savagefb-i2c.c
230
const u8 *e = fb_firmware_edid(info->device);
drivers/video/fbdev/savage/savagefb-i2c.c
232
if (e)
drivers/video/fbdev/savage/savagefb-i2c.c
233
edid = kmemdup(e, EDID_LENGTH, GFP_KERNEL);
drivers/virt/coco/efi_secret/efi_secret.c
100
seq_write(file, e->data, secret_entry_data_len(e));
drivers/virt/coco/efi_secret/efi_secret.c
121
struct secret_entry *e = (struct secret_entry *)inode->i_private;
drivers/virt/coco/efi_secret/efi_secret.c
123
if (e) {
drivers/virt/coco/efi_secret/efi_secret.c
125
wipe_memory(e->data, secret_entry_data_len(e));
drivers/virt/coco/efi_secret/efi_secret.c
126
e->guid = NULL_GUID;
drivers/virt/coco/efi_secret/efi_secret.c
194
struct secret_entry *e;
drivers/virt/coco/efi_secret/efi_secret.c
237
while (bytes_left >= (int)sizeof(*e) && i < EFI_SECRET_NUM_FILES) {
drivers/virt/coco/efi_secret/efi_secret.c
238
e = (struct secret_entry *)ptr;
drivers/virt/coco/efi_secret/efi_secret.c
239
if (e->len < sizeof(*e) || e->len > (unsigned int)bytes_left) {
drivers/virt/coco/efi_secret/efi_secret.c
246
if (efi_guidcmp(e->guid, NULL_GUID)) {
drivers/virt/coco/efi_secret/efi_secret.c
247
efi_guid_to_str(&e->guid, guid_str);
drivers/virt/coco/efi_secret/efi_secret.c
249
dent = securityfs_create_file(guid_str, 0440, dir, (void *)e,
drivers/virt/coco/efi_secret/efi_secret.c
258
ptr += e->len;
drivers/virt/coco/efi_secret/efi_secret.c
259
bytes_left -= e->len;
drivers/virt/coco/efi_secret/efi_secret.c
83
static size_t secret_entry_data_len(struct secret_entry *e)
drivers/virt/coco/efi_secret/efi_secret.c
85
return e->len - sizeof(*e);
drivers/virt/coco/efi_secret/efi_secret.c
97
struct secret_entry *e = file->private;
drivers/virt/coco/efi_secret/efi_secret.c
99
if (e)
drivers/xen/events/events_base.c
179
#define EVTCHN_ROW(e) (e / (PAGE_SIZE/sizeof(**evtchn_to_irq)))
drivers/xen/events/events_base.c
180
#define EVTCHN_COL(e) (e % (PAGE_SIZE/sizeof(**evtchn_to_irq)))
drivers/xen/xen-pciback/passthrough.c
122
struct pci_dev_entry *dev_entry, *e;
drivers/xen/xen-pciback/passthrough.c
136
list_for_each_entry(e, &dev_data->dev_list, list) {
drivers/xen/xen-pciback/passthrough.c
137
if (dev == e->dev) {
drivers/xen/xen-pciback/vpci.c
160
struct pci_dev_entry *e;
drivers/xen/xen-pciback/vpci.c
162
list_for_each_entry(e, &vpci_dev->dev_list[slot], list) {
drivers/xen/xen-pciback/vpci.c
163
if (e->dev == dev) {
drivers/xen/xen-pciback/vpci.c
164
list_del(&e->list);
drivers/xen/xen-pciback/vpci.c
165
found_dev = e->dev;
drivers/xen/xen-pciback/vpci.c
166
kfree(e);
drivers/xen/xen-pciback/vpci.c
216
struct pci_dev_entry *e, *tmp;
drivers/xen/xen-pciback/vpci.c
217
list_for_each_entry_safe(e, tmp, &vpci_dev->dev_list[slot],
drivers/xen/xen-pciback/vpci.c
219
struct pci_dev *dev = e->dev;
drivers/xen/xen-pciback/vpci.c
220
list_del(&e->list);
drivers/xen/xen-pciback/vpci.c
224
kfree(e);
fs/afs/misc.c
143
void afs_prioritise_error(struct afs_error *e, int error, u32 abort_code)
fs/afs/misc.c
147
e->aborted = false;
fs/afs/misc.c
148
e->error = 0;
fs/afs/misc.c
151
if (e->error == -ETIMEDOUT ||
fs/afs/misc.c
152
e->error == -ETIME)
fs/afs/misc.c
157
if (e->error == -ENOMEM ||
fs/afs/misc.c
158
e->error == -ENONET)
fs/afs/misc.c
163
if (e->error == -ERFKILL)
fs/afs/misc.c
167
if (e->error == -EADDRNOTAVAIL)
fs/afs/misc.c
171
if (e->error == -ENETUNREACH)
fs/afs/misc.c
175
if (e->error == -EHOSTUNREACH)
fs/afs/misc.c
179
if (e->error == -EHOSTDOWN)
fs/afs/misc.c
183
if (e->error == -ECONNREFUSED)
fs/afs/misc.c
187
if (e->error == -ECONNRESET)
fs/afs/misc.c
191
if (e->responded)
fs/afs/misc.c
193
e->error = error;
fs/afs/misc.c
194
e->aborted = false;
fs/afs/misc.c
198
e->error = afs_abort_to_error(abort_code);
fs/afs/misc.c
199
e->aborted = true;
fs/afs/misc.c
200
e->responded = true;
fs/afs/misc.c
203
e->aborted = false;
fs/afs/misc.c
204
e->responded = true;
fs/afs/misc.c
205
e->error = error;
fs/afs/rotate.c
742
const struct afs_endpoint_state *e =
fs/afs/rotate.c
744
const struct afs_addr_list *a = e->addresses;
fs/afs/rotate.c
749
e->probe_seq, e->responsive_set, e->failed_set);
fs/afs/rotate.c
755
if (a == e->addresses)
fs/afs/vl_probe.c
208
struct afs_error e = {};
fs/afs/vl_probe.c
218
afs_do_probe_vlserver(net, server, key, i, &e))
fs/afs/vl_probe.c
222
return in_progress ? 0 : e.error;
fs/binfmt_misc.c
103
if (!test_bit(Enabled, &e->flags))
fs/binfmt_misc.c
107
if (!test_bit(Magic, &e->flags)) {
fs/binfmt_misc.c
108
if (p && !strcmp(e->magic, p + 1))
fs/binfmt_misc.c
109
return e;
fs/binfmt_misc.c
114
s = bprm->buf + e->offset;
fs/binfmt_misc.c
115
if (e->mask) {
fs/binfmt_misc.c
116
for (j = 0; j < e->size; j++)
fs/binfmt_misc.c
117
if ((*s++ ^ e->magic[j]) & e->mask[j])
fs/binfmt_misc.c
120
for (j = 0; j < e->size; j++)
fs/binfmt_misc.c
121
if ((*s++ ^ e->magic[j]))
fs/binfmt_misc.c
124
if (j == e->size)
fs/binfmt_misc.c
125
return e;
fs/binfmt_misc.c
144
Node *e;
fs/binfmt_misc.c
147
e = search_binfmt_handler(misc, bprm);
fs/binfmt_misc.c
148
if (e)
fs/binfmt_misc.c
149
refcount_inc(&e->users);
fs/binfmt_misc.c
151
return e;
fs/binfmt_misc.c
162
static void put_binfmt_handler(Node *e)
fs/binfmt_misc.c
164
if (refcount_dec_and_test(&e->users)) {
fs/binfmt_misc.c
165
if (e->flags & MISC_FMT_OPEN_FILE)
fs/binfmt_misc.c
166
filp_close(e->interp_file, NULL);
fs/binfmt_misc.c
167
kfree(e);
fs/binfmt_misc.c
306
static char *check_special_flags(char *sfs, Node *e)
fs/binfmt_misc.c
317
e->flags |= MISC_FMT_PRESERVE_ARGV0;
fs/binfmt_misc.c
322
e->flags |= MISC_FMT_OPEN_BINARY;
fs/binfmt_misc.c
329
e->flags |= (MISC_FMT_CREDENTIALS |
fs/binfmt_misc.c
335
e->flags |= MISC_FMT_OPEN_FILE;
fs/binfmt_misc.c
352
Node *e;
fs/binfmt_misc.c
366
e = kmalloc(memsize, GFP_KERNEL_ACCOUNT);
fs/binfmt_misc.c
367
if (!e)
fs/binfmt_misc.c
370
p = buf = (char *)e + sizeof(Node);
fs/binfmt_misc.c
372
memset(e, 0, sizeof(Node));
fs/binfmt_misc.c
384
e->name = p;
fs/binfmt_misc.c
389
if (!e->name[0] ||
fs/binfmt_misc.c
390
!strcmp(e->name, ".") ||
fs/binfmt_misc.c
391
!strcmp(e->name, "..") ||
fs/binfmt_misc.c
392
strchr(e->name, '/'))
fs/binfmt_misc.c
395
pr_debug("register: name: {%s}\n", e->name);
fs/binfmt_misc.c
401
e->flags = 1 << Enabled;
fs/binfmt_misc.c
405
e->flags = (1 << Enabled) | (1 << Magic);
fs/binfmt_misc.c
413
if (test_bit(Magic, &e->flags)) {
fs/binfmt_misc.c
423
int r = kstrtoint(p, 10, &e->offset);
fs/binfmt_misc.c
424
if (r != 0 || e->offset < 0)
fs/binfmt_misc.c
430
pr_debug("register: offset: %#x\n", e->offset);
fs/binfmt_misc.c
433
e->magic = p;
fs/binfmt_misc.c
437
if (!e->magic[0])
fs/binfmt_misc.c
442
DUMP_PREFIX_NONE, e->magic, p - e->magic);
fs/binfmt_misc.c
445
e->mask = p;
fs/binfmt_misc.c
449
if (!e->mask[0]) {
fs/binfmt_misc.c
450
e->mask = NULL;
fs/binfmt_misc.c
455
DUMP_PREFIX_NONE, e->mask, p - e->mask);
fs/binfmt_misc.c
463
e->size = string_unescape_inplace(e->magic, UNESCAPE_HEX);
fs/binfmt_misc.c
464
if (e->mask &&
fs/binfmt_misc.c
465
string_unescape_inplace(e->mask, UNESCAPE_HEX) != e->size)
fs/binfmt_misc.c
467
if (e->size > BINPRM_BUF_SIZE ||
fs/binfmt_misc.c
468
BINPRM_BUF_SIZE - e->size < e->offset)
fs/binfmt_misc.c
470
pr_debug("register: magic/mask length: %i\n", e->size);
fs/binfmt_misc.c
474
DUMP_PREFIX_NONE, e->magic, e->size);
fs/binfmt_misc.c
476
if (e->mask) {
fs/binfmt_misc.c
478
char *masked = kmalloc(e->size, GFP_KERNEL_ACCOUNT);
fs/binfmt_misc.c
482
DUMP_PREFIX_NONE, e->mask, e->size);
fs/binfmt_misc.c
485
for (i = 0; i < e->size; ++i)
fs/binfmt_misc.c
486
masked[i] = e->magic[i] & e->mask[i];
fs/binfmt_misc.c
489
DUMP_PREFIX_NONE, masked, e->size);
fs/binfmt_misc.c
505
e->magic = p;
fs/binfmt_misc.c
510
if (!e->magic[0] || strchr(e->magic, '/'))
fs/binfmt_misc.c
512
pr_debug("register: extension: {%s}\n", e->magic);
fs/binfmt_misc.c
522
e->interpreter = p;
fs/binfmt_misc.c
527
if (!e->interpreter[0])
fs/binfmt_misc.c
529
pr_debug("register: interpreter: {%s}\n", e->interpreter);
fs/binfmt_misc.c
532
p = check_special_flags(p, e);
fs/binfmt_misc.c
538
return e;
fs/binfmt_misc.c
544
kfree(e);
fs/binfmt_misc.c
547
kfree(e);
fs/binfmt_misc.c
578
static void entry_status(Node *e, char *page)
fs/binfmt_misc.c
583
if (test_bit(Enabled, &e->flags))
fs/binfmt_misc.c
591
dp += sprintf(dp, "%s\ninterpreter %s\n", status, e->interpreter);
fs/binfmt_misc.c
595
if (e->flags & MISC_FMT_PRESERVE_ARGV0)
fs/binfmt_misc.c
597
if (e->flags & MISC_FMT_OPEN_BINARY)
fs/binfmt_misc.c
599
if (e->flags & MISC_FMT_CREDENTIALS)
fs/binfmt_misc.c
601
if (e->flags & MISC_FMT_OPEN_FILE)
fs/binfmt_misc.c
605
if (!test_bit(Magic, &e->flags)) {
fs/binfmt_misc.c
606
sprintf(dp, "extension .%s\n", e->magic);
fs/binfmt_misc.c
608
dp += sprintf(dp, "offset %i\nmagic ", e->offset);
fs/binfmt_misc.c
609
dp = bin2hex(dp, e->magic, e->size);
fs/binfmt_misc.c
610
if (e->mask) {
fs/binfmt_misc.c
612
dp = bin2hex(dp, e->mask, e->size);
fs/binfmt_misc.c
662
Node *e = inode->i_private;
fs/binfmt_misc.c
666
if (e) {
fs/binfmt_misc.c
671
if (!list_empty(&e->list))
fs/binfmt_misc.c
672
list_del_init(&e->list);
fs/binfmt_misc.c
674
put_binfmt_handler(e);
fs/binfmt_misc.c
690
static void remove_binfmt_handler(struct binfmt_misc *misc, Node *e)
fs/binfmt_misc.c
693
list_del_init(&e->list);
fs/binfmt_misc.c
695
locked_recursive_removal(e->dentry, NULL);
fs/binfmt_misc.c
703
Node *e = file_inode(file)->i_private;
fs/binfmt_misc.c
711
entry_status(e, page);
fs/binfmt_misc.c
723
Node *e = inode->i_private;
fs/binfmt_misc.c
729
clear_bit(Enabled, &e->flags);
fs/binfmt_misc.c
733
set_bit(Enabled, &e->flags);
fs/binfmt_misc.c
749
if (!list_empty(&e->list))
fs/binfmt_misc.c
750
remove_binfmt_handler(i_binfmt_misc(inode), e);
fs/binfmt_misc.c
770
static int add_entry(Node *e, struct super_block *sb)
fs/binfmt_misc.c
772
struct dentry *dentry = simple_start_creating(sb->s_root, e->name);
fs/binfmt_misc.c
785
refcount_set(&e->users, 1);
fs/binfmt_misc.c
786
e->dentry = dentry;
fs/binfmt_misc.c
787
inode->i_private = e;
fs/binfmt_misc.c
793
list_add(&e->list, &misc->entries);
fs/binfmt_misc.c
802
Node *e;
fs/binfmt_misc.c
807
e = create_entry(buffer, count);
fs/binfmt_misc.c
809
if (IS_ERR(e))
fs/binfmt_misc.c
810
return PTR_ERR(e);
fs/binfmt_misc.c
812
if (e->flags & MISC_FMT_OPEN_FILE) {
fs/binfmt_misc.c
821
f = open_exec(e->interpreter);
fs/binfmt_misc.c
824
e->interpreter);
fs/binfmt_misc.c
825
kfree(e);
fs/binfmt_misc.c
828
e->interp_file = f;
fs/binfmt_misc.c
831
err = add_entry(e, sb);
fs/binfmt_misc.c
837
kfree(e);
fs/binfmt_misc.c
866
Node *e, *next;
fs/binfmt_misc.c
893
list_for_each_entry_safe(e, next, &misc->entries, list)
fs/binfmt_misc.c
894
remove_binfmt_handler(misc, e);
fs/binfmt_misc.c
95
Node *e;
fs/binfmt_misc.c
98
list_for_each_entry(e, &misc->entries, list) {
fs/btrfs/backref.c
43
struct extent_inode_elem *e;
fs/btrfs/backref.c
81
e = kmalloc_obj(*e, GFP_NOFS);
fs/btrfs/backref.c
82
if (!e)
fs/btrfs/backref.c
85
e->next = *eie;
fs/btrfs/backref.c
86
e->inum = key->objectid;
fs/btrfs/backref.c
87
e->offset = offset;
fs/btrfs/backref.c
88
e->num_bytes = data_len;
fs/btrfs/backref.c
89
*eie = e;
fs/btrfs/extent-io-tree.c
80
#define btrfs_debug_check_extent_io_range(c, s, e) do {} while (0)
fs/btrfs/file-item.h
44
const struct btrfs_file_extent_item *e)
fs/btrfs/file-item.h
46
return (unsigned long)e + BTRFS_FILE_EXTENT_INLINE_DATA_START;
fs/btrfs/free-space-cache.c
1097
struct btrfs_free_space *e;
fs/btrfs/free-space-cache.c
1099
e = rb_entry(node, struct btrfs_free_space, offset_index);
fs/btrfs/free-space-cache.c
1102
ret = io_ctl_add_entry(io_ctl, e->offset, e->bytes,
fs/btrfs/free-space-cache.c
1103
e->bitmap);
fs/btrfs/free-space-cache.c
1107
if (e->bitmap) {
fs/btrfs/free-space-cache.c
1108
list_add_tail(&e->list, bitmap_list);
fs/btrfs/free-space-cache.c
651
struct btrfs_free_space_entry *e;
fs/btrfs/free-space-cache.c
660
e = io_ctl->cur;
fs/btrfs/free-space-cache.c
661
entry->offset = get_unaligned_le64(&e->offset);
fs/btrfs/free-space-cache.c
662
entry->bytes = get_unaligned_le64(&e->bytes);
fs/btrfs/free-space-cache.c
663
*type = e->type;
fs/btrfs/free-space-cache.c
743
struct btrfs_free_space *e, *n;
fs/btrfs/free-space-cache.c
813
e = kmem_cache_zalloc(btrfs_free_space_cachep,
fs/btrfs/free-space-cache.c
815
if (!e) {
fs/btrfs/free-space-cache.c
820
ret = io_ctl_read_entry(&io_ctl, e, &type);
fs/btrfs/free-space-cache.c
822
kmem_cache_free(btrfs_free_space_cachep, e);
fs/btrfs/free-space-cache.c
826
if (!e->bytes) {
fs/btrfs/free-space-cache.c
828
kmem_cache_free(btrfs_free_space_cachep, e);
fs/btrfs/free-space-cache.c
834
ret = link_free_space(ctl, e);
fs/btrfs/free-space-cache.c
839
kmem_cache_free(btrfs_free_space_cachep, e);
fs/btrfs/free-space-cache.c
845
e->bitmap = kmem_cache_zalloc(
fs/btrfs/free-space-cache.c
847
if (!e->bitmap) {
fs/btrfs/free-space-cache.c
850
btrfs_free_space_cachep, e);
fs/btrfs/free-space-cache.c
854
ret = link_free_space(ctl, e);
fs/btrfs/free-space-cache.c
859
kmem_cache_free(btrfs_free_space_bitmap_cachep, e->bitmap);
fs/btrfs/free-space-cache.c
860
kmem_cache_free(btrfs_free_space_cachep, e);
fs/btrfs/free-space-cache.c
866
list_add_tail(&e->list, &bitmaps);
fs/btrfs/free-space-cache.c
878
list_for_each_entry_safe(e, n, &bitmaps, list) {
fs/btrfs/free-space-cache.c
879
list_del_init(&e->list);
fs/btrfs/free-space-cache.c
880
ret = io_ctl_read_bitmap(&io_ctl, e);
fs/ceph/snap.c
1028
void *e = p + msg->front.iov_len;
fs/ceph/snap.c
1070
ceph_decode_need(&p, e, sizeof(*ri), bad);
fs/ceph/snap.c
1155
if (ceph_update_snap_trace(mdsc, p, e,
fs/ceph/snap.c
780
void *p, void *e, bool deletion,
fs/ceph/snap.c
802
ceph_decode_need(&p, e, sizeof(*ri), bad);
fs/ceph/snap.c
805
ceph_decode_need(&p, e, sizeof(u64)*(le32_to_cpu(ri->num_snaps) +
fs/ceph/snap.c
861
realm, rebuild_snapcs, p, e);
fs/ceph/snap.c
872
if (realm_to_rebuild && p >= e)
fs/ceph/snap.c
880
if (p < e)
fs/ceph/super.h
1020
void *p, void *e, bool deletion,
fs/dlm/lowcomms.c
1112
static void writequeue_entry_complete(struct writequeue_entry *e, int completed)
fs/dlm/lowcomms.c
1114
e->offset += completed;
fs/dlm/lowcomms.c
1115
e->len -= completed;
fs/dlm/lowcomms.c
1117
e->dirty = true;
fs/dlm/lowcomms.c
1119
if (e->len == 0 && e->users == 0)
fs/dlm/lowcomms.c
1120
free_entry(e);
fs/dlm/lowcomms.c
1193
struct writequeue_entry *e;
fs/dlm/lowcomms.c
1197
e = list_last_entry(&con->writequeue, struct writequeue_entry, list);
fs/dlm/lowcomms.c
1198
if (DLM_WQ_REMAIN_BYTES(e) >= len) {
fs/dlm/lowcomms.c
1199
kref_get(&e->ref);
fs/dlm/lowcomms.c
1201
*ppc = page_address(e->page) + e->end;
fs/dlm/lowcomms.c
1205
e->end += len;
fs/dlm/lowcomms.c
1206
e->users++;
fs/dlm/lowcomms.c
1211
e = new_writequeue_entry(con);
fs/dlm/lowcomms.c
1212
if (!e)
fs/dlm/lowcomms.c
1215
kref_get(&e->ref);
fs/dlm/lowcomms.c
1216
*ppc = page_address(e->page);
fs/dlm/lowcomms.c
1217
e->end += len;
fs/dlm/lowcomms.c
122
#define DLM_WQ_REMAIN_BYTES(e) (PAGE_SIZE - e->end)
fs/dlm/lowcomms.c
1221
list_add_tail(&e->list, &con->writequeue);
fs/dlm/lowcomms.c
1225
return e;
fs/dlm/lowcomms.c
123
#define DLM_WQ_LENGTH_BYTES(e) (e->end - e->offset)
fs/dlm/lowcomms.c
1232
struct writequeue_entry *e;
fs/dlm/lowcomms.c
1241
e = new_wq_entry(con, len, ppc, cb, data);
fs/dlm/lowcomms.c
1242
if (!e) {
fs/dlm/lowcomms.c
1251
msg->entry = e;
fs/dlm/lowcomms.c
1298
struct writequeue_entry *e = msg->entry;
fs/dlm/lowcomms.c
1299
struct connection *con = e->con;
fs/dlm/lowcomms.c
1304
list_add(&msg->list, &e->msgs);
fs/dlm/lowcomms.c
1306
users = --e->users;
fs/dlm/lowcomms.c
1310
e->len = DLM_WQ_LENGTH_BYTES(e);
fs/dlm/lowcomms.c
1365
struct writequeue_entry *e;
fs/dlm/lowcomms.c
1373
e = con_next_wq(con);
fs/dlm/lowcomms.c
1374
if (!e) {
fs/dlm/lowcomms.c
1380
len = e->len;
fs/dlm/lowcomms.c
1381
offset = e->offset;
fs/dlm/lowcomms.c
1382
WARN_ON_ONCE(len == 0 && e->users == 0);
fs/dlm/lowcomms.c
1385
bvec_set_page(&bvec, e->page, len, offset);
fs/dlm/lowcomms.c
1416
writequeue_entry_complete(e, ret);
fs/dlm/lowcomms.c
1424
struct writequeue_entry *e, *safe;
fs/dlm/lowcomms.c
1427
list_for_each_entry_safe(e, safe, &con->writequeue, list) {
fs/dlm/lowcomms.c
1428
free_entry(e);
fs/dlm/lowcomms.c
257
struct writequeue_entry *e;
fs/dlm/lowcomms.c
259
e = list_first_entry_or_null(&con->writequeue, struct writequeue_entry,
fs/dlm/lowcomms.c
264
if (!e || e->users || e->len == 0)
fs/dlm/lowcomms.c
267
return e;
fs/dlm/lowcomms.c
684
struct writequeue_entry *e = container_of(kref, struct writequeue_entry,
fs/dlm/lowcomms.c
687
__free_page(e->page);
fs/dlm/lowcomms.c
688
dlm_free_writequeue(e);
fs/dlm/lowcomms.c
699
static void free_entry(struct writequeue_entry *e)
fs/dlm/lowcomms.c
703
list_for_each_entry_safe(msg, tmp, &e->msgs, list) {
fs/dlm/lowcomms.c
713
list_del(&e->list);
fs/dlm/lowcomms.c
714
kref_put(&e->ref, dlm_page_release);
fs/dlm/lowcomms.c
758
struct writequeue_entry *e;
fs/dlm/lowcomms.c
784
e = list_first_entry(&con->writequeue, struct writequeue_entry,
fs/dlm/lowcomms.c
786
if (e->dirty)
fs/dlm/lowcomms.c
787
free_entry(e);
fs/dlm/requestqueue.c
136
struct rq_entry *e, *safe;
fs/dlm/requestqueue.c
139
list_for_each_entry_safe(e, safe, &ls->ls_requestqueue, list) {
fs/dlm/requestqueue.c
140
ms = &e->request;
fs/dlm/requestqueue.c
142
if (purge_request(ls, ms, e->nodeid)) {
fs/dlm/requestqueue.c
143
list_del(&e->list);
fs/dlm/requestqueue.c
144
kfree(e);
fs/dlm/requestqueue.c
36
struct rq_entry *e;
fs/dlm/requestqueue.c
40
e = kmalloc(sizeof(struct rq_entry) + length, GFP_ATOMIC);
fs/dlm/requestqueue.c
41
if (!e) {
fs/dlm/requestqueue.c
46
e->recover_seq = ls->ls_recover_seq & 0xFFFFFFFF;
fs/dlm/requestqueue.c
47
e->nodeid = nodeid;
fs/dlm/requestqueue.c
48
memcpy(&e->request, ms, sizeof(*ms));
fs/dlm/requestqueue.c
49
memcpy(&e->request.m_extra, ms->m_extra, length);
fs/dlm/requestqueue.c
51
list_add_tail(&e->list, &ls->ls_requestqueue);
fs/dlm/requestqueue.c
67
struct rq_entry *e;
fs/dlm/requestqueue.c
78
e = list_first_entry(&ls->ls_requestqueue, struct rq_entry, list);
fs/dlm/requestqueue.c
80
ms = &e->request;
fs/dlm/requestqueue.c
88
e->recover_seq);
fs/dlm/requestqueue.c
90
dlm_receive_message_saved(ls, &e->request, e->recover_seq);
fs/dlm/requestqueue.c
91
list_del(&e->list);
fs/dlm/requestqueue.c
92
kfree(e);
fs/erofs/decompressor_crypto.c
127
struct z_erofs_crypto_engine *e;
fs/erofs/decompressor_crypto.c
133
for (e = z_erofs_crypto[alg]; e->crypto_name; ++e) {
fs/erofs/decompressor_crypto.c
134
if (!strncmp(name, e->crypto_name, len)) {
fs/erofs/decompressor_crypto.c
135
if (e->tfm)
fs/erofs/decompressor_crypto.c
137
tfm = crypto_alloc_acomp(e->crypto_name, 0, 0);
fs/erofs/decompressor_crypto.c
142
e->tfm = tfm;
fs/erofs/decompressor_crypto.c
153
struct z_erofs_crypto_engine *e;
fs/erofs/decompressor_crypto.c
158
for (e = z_erofs_crypto[alg]; e->crypto_name; ++e) {
fs/erofs/decompressor_crypto.c
159
if (!e->tfm)
fs/erofs/decompressor_crypto.c
161
crypto_free_acomp(e->tfm);
fs/erofs/decompressor_crypto.c
162
e->tfm = NULL;
fs/erofs/decompressor_crypto.c
170
struct z_erofs_crypto_engine *e;
fs/erofs/decompressor_crypto.c
174
for (e = z_erofs_crypto[alg]; e->crypto_name; ++e) {
fs/erofs/decompressor_crypto.c
175
if (!e->tfm)
fs/erofs/decompressor_crypto.c
178
e->crypto_name, sep);
fs/erofs/decompressor_crypto.c
84
struct z_erofs_crypto_engine *e;
fs/erofs/decompressor_crypto.c
86
for (e = z_erofs_crypto[alg]; e->crypto_name; ++e)
fs/erofs/decompressor_crypto.c
87
if (e->tfm)
fs/erofs/decompressor_crypto.c
88
return e->tfm;
fs/erofs/erofs_fs.h
258
static inline unsigned int erofs_xattr_entry_size(struct erofs_xattr_entry *e)
fs/erofs/erofs_fs.h
261
e->e_name_len + le16_to_cpu(e->e_value_size));
fs/ext2/acl.c
104
e = (char *)ext_acl + sizeof(ext2_acl_header);
fs/ext2/acl.c
107
ext2_acl_entry *entry = (ext2_acl_entry *)e;
fs/ext2/acl.c
114
e += sizeof(ext2_acl_entry);
fs/ext2/acl.c
119
e += sizeof(ext2_acl_entry);
fs/ext2/acl.c
126
e += sizeof(ext2_acl_entry_short);
fs/ext2/acl.c
95
char *e;
fs/ext4/acl.c
103
e = (char *)ext_acl + sizeof(ext4_acl_header);
fs/ext4/acl.c
106
ext4_acl_entry *entry = (ext4_acl_entry *)e;
fs/ext4/acl.c
113
e += sizeof(ext4_acl_entry);
fs/ext4/acl.c
118
e += sizeof(ext4_acl_entry);
fs/ext4/acl.c
125
e += sizeof(ext4_acl_entry_short);
fs/ext4/acl.c
94
char *e;
fs/ext4/xattr.c
195
struct ext4_xattr_entry *e = entry;
fs/ext4/xattr.c
227
while (!IS_LAST_ENTRY(e)) {
fs/ext4/xattr.c
228
struct ext4_xattr_entry *next = EXT4_XATTR_NEXT(e);
fs/ext4/xattr.c
233
if (strnlen(e->e_name, e->e_name_len) != e->e_name_len) {
fs/ext4/xattr.c
237
e = next;
fs/ext4/xattr.c
278
if (value < (void *)e + sizeof(u32) ||
fs/f2fs/checkpoint.c
713
struct ino_entry *e = NULL, *new = NULL;
fs/f2fs/checkpoint.c
718
e = radix_tree_lookup(&im->ino_root, ino);
fs/f2fs/checkpoint.c
723
if (!e)
fs/f2fs/checkpoint.c
731
e = radix_tree_lookup(&im->ino_root, ino);
fs/f2fs/checkpoint.c
732
if (!e) {
fs/f2fs/checkpoint.c
738
e = new;
fs/f2fs/checkpoint.c
739
if (unlikely(radix_tree_insert(&im->ino_root, ino, e)))
fs/f2fs/checkpoint.c
742
memset(e, 0, sizeof(struct ino_entry));
fs/f2fs/checkpoint.c
743
e->ino = ino;
fs/f2fs/checkpoint.c
745
list_add_tail(&e->list, &im->ino_list);
fs/f2fs/checkpoint.c
751
f2fs_set_bit(devidx, (char *)&e->dirty_device);
fs/f2fs/checkpoint.c
756
if (new && e != new)
fs/f2fs/checkpoint.c
763
struct ino_entry *e;
fs/f2fs/checkpoint.c
766
e = radix_tree_lookup(&im->ino_root, ino);
fs/f2fs/checkpoint.c
767
if (e) {
fs/f2fs/checkpoint.c
768
list_del(&e->list);
fs/f2fs/checkpoint.c
772
kmem_cache_free(ino_entry_slab, e);
fs/f2fs/checkpoint.c
794
struct ino_entry *e;
fs/f2fs/checkpoint.c
797
e = radix_tree_lookup(&im->ino_root, ino);
fs/f2fs/checkpoint.c
799
return e ? true : false;
fs/f2fs/checkpoint.c
804
struct ino_entry *e, *tmp;
fs/f2fs/checkpoint.c
811
list_for_each_entry_safe(e, tmp, &im->ino_list, list) {
fs/f2fs/checkpoint.c
812
list_del(&e->list);
fs/f2fs/checkpoint.c
813
radix_tree_delete(&im->ino_root, e->ino);
fs/f2fs/checkpoint.c
814
kmem_cache_free(ino_entry_slab, e);
fs/f2fs/checkpoint.c
831
struct ino_entry *e;
fs/f2fs/checkpoint.c
835
e = radix_tree_lookup(&im->ino_root, ino);
fs/f2fs/checkpoint.c
836
if (e && f2fs_test_bit(devidx, (char *)&e->dirty_device))
fs/f2fs/node.c
186
static void __free_nat_entry(struct nat_entry *e)
fs/f2fs/node.c
188
kmem_cache_free(nat_entry_slab, e);
fs/f2fs/node.c
2399
struct free_nid *i, *e;
fs/f2fs/node.c
244
static void __del_from_nat_cache(struct f2fs_nm_info *nm_i, struct nat_entry *e)
fs/f2fs/node.c
2449
e = __lookup_free_nid_list(nm_i, nid);
fs/f2fs/node.c
2450
if (e) {
fs/f2fs/node.c
2451
if (e->state == FREE_NID)
fs/f2fs/node.c
246
radix_tree_delete(&nm_i->nat_root, nat_get_nid(e));
fs/f2fs/node.c
249
__free_nat_entry(e);
fs/f2fs/node.c
397
struct nat_entry *e;
fs/f2fs/node.c
401
e = __lookup_nat_cache(nm_i, nid, false);
fs/f2fs/node.c
402
if (e) {
fs/f2fs/node.c
403
if (!get_nat_flag(e, IS_CHECKPOINTED) &&
fs/f2fs/node.c
404
!get_nat_flag(e, HAS_FSYNCED_INODE))
fs/f2fs/node.c
414
struct nat_entry *e;
fs/f2fs/node.c
418
e = __lookup_nat_cache(nm_i, nid, false);
fs/f2fs/node.c
419
if (e && !get_nat_flag(e, IS_CHECKPOINTED))
fs/f2fs/node.c
428
struct nat_entry *e;
fs/f2fs/node.c
432
e = __lookup_nat_cache(nm_i, ino, false);
fs/f2fs/node.c
433
if (e && get_nat_flag(e, HAS_LAST_FSYNC) &&
fs/f2fs/node.c
434
(get_nat_flag(e, IS_CHECKPOINTED) ||
fs/f2fs/node.c
435
get_nat_flag(e, HAS_FSYNCED_INODE)))
fs/f2fs/node.c
446
struct nat_entry *new, *e;
fs/f2fs/node.c
457
e = __lookup_nat_cache(nm_i, nid, false);
fs/f2fs/node.c
458
if (!e)
fs/f2fs/node.c
459
e = __init_nat_entry(nm_i, new, ne, false, false);
fs/f2fs/node.c
461
f2fs_bug_on(sbi, nat_get_ino(e) != le32_to_cpu(ne->ino) ||
fs/f2fs/node.c
462
nat_get_blkaddr(e) !=
fs/f2fs/node.c
464
nat_get_version(e) != ne->version);
fs/f2fs/node.c
466
if (e != new)
fs/f2fs/node.c
474
struct nat_entry *e;
fs/f2fs/node.c
479
e = __lookup_nat_cache(nm_i, ni->nid, true);
fs/f2fs/node.c
480
if (!e) {
fs/f2fs/node.c
482
e = __init_nat_entry(nm_i, new, NULL, true, true);
fs/f2fs/node.c
483
copy_node_info(&e->ni, ni);
fs/f2fs/node.c
491
copy_node_info(&e->ni, ni);
fs/f2fs/node.c
495
if (e != new)
fs/f2fs/node.c
499
f2fs_bug_on(sbi, nat_get_blkaddr(e) != ni->blk_addr);
fs/f2fs/node.c
500
f2fs_bug_on(sbi, nat_get_blkaddr(e) == NULL_ADDR &&
fs/f2fs/node.c
502
f2fs_bug_on(sbi, nat_get_blkaddr(e) == NEW_ADDR &&
fs/f2fs/node.c
504
f2fs_bug_on(sbi, __is_valid_data_blkaddr(nat_get_blkaddr(e)) &&
fs/f2fs/node.c
508
if (nat_get_blkaddr(e) != NEW_ADDR && new_blkaddr == NULL_ADDR) {
fs/f2fs/node.c
509
unsigned char version = nat_get_version(e);
fs/f2fs/node.c
511
nat_set_version(e, inc_node_version(version));
fs/f2fs/node.c
515
nat_set_blkaddr(e, new_blkaddr);
fs/f2fs/node.c
517
set_nat_flag(e, IS_CHECKPOINTED, false);
fs/f2fs/node.c
518
__set_nat_cache_dirty(nm_i, e, init_dirty);
fs/f2fs/node.c
522
e = __lookup_nat_cache(nm_i, ni->ino, false);
fs/f2fs/node.c
523
if (e) {
fs/f2fs/node.c
525
set_nat_flag(e, HAS_FSYNCED_INODE, true);
fs/f2fs/node.c
526
set_nat_flag(e, HAS_LAST_FSYNC, fsync_done);
fs/f2fs/node.c
572
struct nat_entry *e;
fs/f2fs/node.c
582
e = __lookup_nat_cache(nm_i, nid, false);
fs/f2fs/node.c
583
if (e) {
fs/f2fs/node.c
584
ni->ino = nat_get_ino(e);
fs/f2fs/node.c
585
ni->blk_addr = nat_get_blkaddr(e);
fs/f2fs/node.c
586
ni->version = nat_get_version(e);
fs/hfs/bfind.c
63
int b, e;
fs/hfs/bfind.c
67
e = bnode->num_recs - 1;
fs/hfs/bfind.c
70
rec = (e + b) / 2;
fs/hfs/bfind.c
80
e = rec;
fs/hfs/bfind.c
87
e = rec - 1;
fs/hfs/bfind.c
88
} while (b <= e);
fs/hfs/bfind.c
89
if (rec != e && e >= 0) {
fs/hfs/bfind.c
90
len = hfs_brec_lenoff(bnode, e, &off);
fs/hfs/bfind.c
91
keylen = hfs_brec_keylen(bnode, e);
fs/hfs/bfind.c
99
fd->record = e;
fs/hfsplus/bfind.c
108
int b, e;
fs/hfsplus/bfind.c
113
e = bnode->num_recs - 1;
fs/hfsplus/bfind.c
116
rec = (e + b) / 2;
fs/hfsplus/bfind.c
124
if (rec_found(bnode, fd, &b, &e, &rec)) {
fs/hfsplus/bfind.c
128
} while (b <= e);
fs/hfsplus/bfind.c
130
if (rec != e && e >= 0) {
fs/hfsplus/bfind.c
131
len = hfs_brec_lenoff(bnode, e, &off);
fs/hfsplus/bfind.c
132
keylen = hfs_brec_keylen(bnode, e);
fs/hfsplus/bfind.c
141
fd->record = e;
fs/hfsplus/unicode.c
132
int i, s, e;
fs/hfsplus/unicode.c
135
e = p[1];
fs/hfsplus/unicode.c
136
if (!e || cc < p[s * 2] || cc > p[e * 2])
fs/hfsplus/unicode.c
139
i = (s + e) / 2;
fs/hfsplus/unicode.c
143
e = i - 1;
fs/hfsplus/unicode.c
146
} while (s <= e);
fs/hpfs/dir.c
103
e = 1;
fs/hpfs/dir.c
107
if (e) {
fs/hpfs/dir.c
92
int e = 0;
fs/hpfs/dir.c
98
e = 1;
fs/jffs2/acl.c
132
void *e;
fs/jffs2/acl.c
140
e = header + 1;
fs/jffs2/acl.c
143
entry = e;
fs/jffs2/acl.c
150
e += sizeof(struct jffs2_acl_entry);
fs/jffs2/acl.c
155
e += sizeof(struct jffs2_acl_entry);
fs/jffs2/acl.c
162
e += sizeof(struct jffs2_acl_entry_short);
fs/jffs2/os-linux.h
87
#define jffs2_flash_writev(a,b,c,d,e,f) jffs2_flash_direct_writev(a,b,c,d,e)
fs/jffs2/summary.h
209
#define jffs2_sum_scan_sumnode(a,b,c,d,e) (0)
fs/nfsd/nfs4idmap.c
100
struct ent *e = kmalloc_obj(*e);
fs/nfsd/nfs4idmap.c
101
if (e)
fs/nfsd/nfs4idmap.c
102
return &e->h;
fs/ntfs3/bitmap.c
1022
e = NULL;
fs/ntfs3/bitmap.c
1031
e = rb_entry(cr, struct e_node, start.node);
fs/ntfs3/bitmap.c
1033
if (e->start.key == hint)
fs/ntfs3/bitmap.c
1036
if (e->start.key < hint) {
fs/ntfs3/bitmap.c
1046
e = pr ? rb_entry(pr, struct e_node, start.node) : NULL;
fs/ntfs3/bitmap.c
1051
if (!e)
fs/ntfs3/bitmap.c
1054
if (e->start.key + e->count.key > hint) {
fs/ntfs3/bitmap.c
1056
size_t len = e->start.key + e->count.key - hint;
fs/ntfs3/bitmap.c
1077
e = rb_entry(rb_first(&wnd->count_tree), struct e_node, count.node);
fs/ntfs3/bitmap.c
1078
if (e->count.key != wnd->extent_max)
fs/ntfs3/bitmap.c
1079
wnd->extent_max = e->count.key;
fs/ntfs3/bitmap.c
1081
if (e->count.key < max_alloc) {
fs/ntfs3/bitmap.c
1082
if (e->count.key >= to_alloc) {
fs/ntfs3/bitmap.c
1085
if (e->count.key < to_alloc0) {
fs/ntfs3/bitmap.c
1089
to_alloc = e->count.key;
fs/ntfs3/bitmap.c
1091
to_alloc = e->count.key;
fs/ntfs3/bitmap.c
1101
max_check = e->start.key + to_alloc;
fs/ntfs3/bitmap.c
1104
for (op = e->start.key + e->count.key; op < max_check;
fs/ntfs3/bitmap.c
1111
to_alloc = op - e->start.key;
fs/ntfs3/bitmap.c
1115
fnd = e->start.key;
fs/ntfs3/bitmap.c
1116
if (e->start.key + to_alloc > max_alloc)
fs/ntfs3/bitmap.c
1117
to_alloc = max_alloc - e->start.key;
fs/ntfs3/bitmap.c
1126
b_len = e->count.key;
fs/ntfs3/bitmap.c
1127
b_pos = e->start.key;
fs/ntfs3/bitmap.c
167
static inline bool rb_insert_count(struct rb_root *root, struct e_node *e)
fs/ntfs3/bitmap.c
171
size_t e_ckey = e->count.key;
fs/ntfs3/bitmap.c
172
size_t e_skey = e->start.key;
fs/ntfs3/bitmap.c
192
rb_link_node(&e->count.node, parent, p);
fs/ntfs3/bitmap.c
193
rb_insert_color(&e->count.node, root);
fs/ntfs3/bitmap.c
200
static inline bool rb_insert_start(struct rb_root *root, struct e_node *e)
fs/ntfs3/bitmap.c
204
size_t e_skey = e->start.key;
fs/ntfs3/bitmap.c
222
rb_link_node(&e->start.node, parent, p);
fs/ntfs3/bitmap.c
223
rb_insert_color(&e->start.node, root);
fs/ntfs3/bitmap.c
234
struct e_node *e, *e0 = NULL;
fs/ntfs3/bitmap.c
252
e = rb_entry(n, struct e_node, start.node);
fs/ntfs3/bitmap.c
254
if (e->start.key + e->count.key == bit) {
fs/ntfs3/bitmap.c
256
bit = e->start.key;
fs/ntfs3/bitmap.c
257
len += e->count.key;
fs/ntfs3/bitmap.c
258
rb_erase(&e->start.node, &wnd->start_tree);
fs/ntfs3/bitmap.c
259
rb_erase(&e->count.node, &wnd->count_tree);
fs/ntfs3/bitmap.c
261
e0 = e;
fs/ntfs3/bitmap.c
268
e = rb_entry(n, struct e_node, start.node);
fs/ntfs3/bitmap.c
269
next_end = e->start.key + e->count.key;
fs/ntfs3/bitmap.c
270
if (e->start.key > end_in)
fs/ntfs3/bitmap.c
277
rb_erase(&e->start.node, &wnd->start_tree);
fs/ntfs3/bitmap.c
278
rb_erase(&e->count.node, &wnd->count_tree);
fs/ntfs3/bitmap.c
282
e0 = e;
fs/ntfs3/bitmap.c
284
kmem_cache_free(ntfs_enode_cachep, e);
fs/ntfs3/bitmap.c
320
e = rb_entry(n, struct e_node, count.node);
fs/ntfs3/bitmap.c
321
if (len <= e->count.key)
fs/ntfs3/bitmap.c
334
rb_erase(&e->start.node, &wnd->start_tree);
fs/ntfs3/bitmap.c
335
rb_erase(&e->count.node, &wnd->count_tree);
fs/ntfs3/bitmap.c
338
e = e0 ? e0 : kmem_cache_alloc(ntfs_enode_cachep, GFP_ATOMIC);
fs/ntfs3/bitmap.c
339
if (!e) {
fs/ntfs3/bitmap.c
347
e->start.key = bit;
fs/ntfs3/bitmap.c
348
e->count.key = len;
fs/ntfs3/bitmap.c
352
rb_insert_start(&wnd->start_tree, e);
fs/ntfs3/bitmap.c
353
rb_insert_count(&wnd->count_tree, e);
fs/ntfs3/bitmap.c
365
struct e_node *e, *e3;
fs/ntfs3/bitmap.c
375
e = rb_entry(n, struct e_node, start.node);
fs/ntfs3/bitmap.c
376
end = e->start.key + e->count.key;
fs/ntfs3/bitmap.c
379
len = e->count.key;
fs/ntfs3/bitmap.c
382
if (e->start.key > bit)
fs/ntfs3/bitmap.c
388
len = bit - e->start.key;
fs/ntfs3/bitmap.c
426
if (e->count.key != wnd->extent_max) {
fs/ntfs3/bitmap.c
428
} else if (rb_prev(&e->count.node)) {
fs/ntfs3/bitmap.c
431
n3 = rb_next(&e->count.node);
fs/ntfs3/bitmap.c
443
e->start.key = new_key;
fs/ntfs3/bitmap.c
444
rb_erase(&e->count.node, &wnd->count_tree);
fs/ntfs3/bitmap.c
445
e->count.key = new_len;
fs/ntfs3/bitmap.c
446
rb_insert_count(&wnd->count_tree, e);
fs/ntfs3/bitmap.c
448
rb_erase(&e->start.node, &wnd->start_tree);
fs/ntfs3/bitmap.c
449
rb_erase(&e->count.node, &wnd->count_tree);
fs/ntfs3/bitmap.c
451
kmem_cache_free(ntfs_enode_cachep, e);
fs/ntfs3/bitmap.c
455
rb_erase(&e->count.node, &wnd->count_tree);
fs/ntfs3/bitmap.c
456
e->count.key = len;
fs/ntfs3/bitmap.c
457
rb_insert_count(&wnd->count_tree, e);
fs/ntfs3/bitmap.c
466
e = rb_entry(rb_last(&wnd->count_tree), struct e_node,
fs/ntfs3/bitmap.c
468
if (e->count.key > new_len)
fs/ntfs3/bitmap.c
472
rb_erase(&e->start.node, &wnd->start_tree);
fs/ntfs3/bitmap.c
473
rb_erase(&e->count.node, &wnd->count_tree);
fs/ntfs3/bitmap.c
476
e = kmem_cache_alloc(ntfs_enode_cachep, GFP_ATOMIC);
fs/ntfs3/bitmap.c
477
if (!e)
fs/ntfs3/bitmap.c
481
if (e) {
fs/ntfs3/bitmap.c
482
e->start.key = new_key;
fs/ntfs3/bitmap.c
483
e->count.key = new_len;
fs/ntfs3/bitmap.c
484
rb_insert_start(&wnd->start_tree, e);
fs/ntfs3/bitmap.c
485
rb_insert_count(&wnd->count_tree, e);
fs/ntfs3/bitmap.c
892
struct e_node *e;
fs/ntfs3/bitmap.c
901
e = rb_entry(n, struct e_node, start.node);
fs/ntfs3/bitmap.c
903
end = e->start.key + e->count.key;
fs/ntfs3/bitmap.c
929
struct e_node *e;
fs/ntfs3/bitmap.c
939
e = rb_entry(n, struct e_node, start.node);
fs/ntfs3/bitmap.c
940
if (e->start.key + e->count.key > bit)
fs/ntfs3/bitmap.c
984
const struct e_node *e;
fs/ntfs3/dir.c
241
struct NTFS_DE *e;
fs/ntfs3/dir.c
255
err = indx_find(&ni->dir, ni, NULL, uni, 0, sbi, &diff, &e, fnd);
fs/ntfs3/dir.c
265
inode = ntfs_iget5(sb, &e->ref, uni);
fs/ntfs3/dir.c
280
struct ntfs_inode *ni, const struct NTFS_DE *e,
fs/ntfs3/dir.c
288
fname = Add2Ptr(e, sizeof(struct NTFS_DE));
fs/ntfs3/dir.c
296
ino = ino_get(&e->ref);
fs/ntfs3/dir.c
308
if (fname->name_len + sizeof(struct NTFS_DE) > le16_to_cpu(e->size))
fs/ntfs3/dir.c
337
struct inode *inode = ntfs_iget5(sbi->sb, &e->ref, NULL);
fs/ntfs3/dir.c
358
const struct NTFS_DE *e;
fs/ntfs3/dir.c
367
e = Add2Ptr(hdr, off);
fs/ntfs3/dir.c
368
e_size = le16_to_cpu(e->size);
fs/ntfs3/dir.c
372
if (de_is_last(e))
fs/ntfs3/dir.c
379
if (le16_to_cpu(e->key_size) < SIZEOF_ATTRIBUTE_FILENAME)
fs/ntfs3/dir.c
385
if (!ntfs_dir_emit(sbi, ni, e, name, ctx)) {
fs/ntfs3/dir.c
572
struct NTFS_DE *e = NULL;
fs/ntfs3/dir.c
595
e = Add2Ptr(hdr, off);
fs/ntfs3/dir.c
596
e_size = le16_to_cpu(e->size);
fs/ntfs3/dir.c
603
if (de_is_last(e))
fs/ntfs3/dir.c
606
fname = de_get_fname(e);
fs/ntfs3/fslog.c
2604
const struct NTFS_DE *e;
fs/ntfs3/fslog.c
2623
e = Add2Ptr(hdr, de_off);
fs/ntfs3/fslog.c
2625
u16 esize = le16_to_cpu(e->size);
fs/ntfs3/fslog.c
2626
struct NTFS_DE *next = Add2Ptr(e, esize);
fs/ntfs3/fslog.c
2629
(e->flags & NTFS_IE_HAS_SUBNODES) != mask) {
fs/ntfs3/fslog.c
2633
if (de_is_last(e))
fs/ntfs3/fslog.c
2636
e = next;
fs/ntfs3/fslog.c
2889
const struct NTFS_DE *e = Add2Ptr(hdr, de_off);
fs/ntfs3/fslog.c
2898
esize = le16_to_cpu(e->size);
fs/ntfs3/fslog.c
2903
e = Add2Ptr(e, esize);
fs/ntfs3/fslog.c
2914
const struct NTFS_DE *e = Add2Ptr(hdr, de_off);
fs/ntfs3/fslog.c
2923
esize = le16_to_cpu(e->size);
fs/ntfs3/fslog.c
2929
e = Add2Ptr(e, esize);
fs/ntfs3/fslog.c
3071
struct NTFS_DE *e, *e1, *e2;
fs/ntfs3/fslog.c
3393
e = (struct NTFS_DE *)data;
fs/ntfs3/fslog.c
3394
esize = le16_to_cpu(e->size);
fs/ntfs3/fslog.c
3412
memmove(e1, e, esize);
fs/ntfs3/fslog.c
3458
e = Add2Ptr(attr, le16_to_cpu(lrh->attr_off));
fs/ntfs3/fslog.c
3460
de_set_vbn_le(e, *(__le64 *)data);
fs/ntfs3/fslog.c
3473
e = Add2Ptr(attr, le16_to_cpu(lrh->attr_off));
fs/ntfs3/fslog.c
3474
fname = (struct ATTR_FILE_NAME *)(e + 1);
fs/ntfs3/fslog.c
3488
e = Add2Ptr(attr, le16_to_cpu(lrh->attr_off));
fs/ntfs3/fslog.c
3490
memmove(Add2Ptr(e, le16_to_cpu(e->view.data_off)), data, dlen);
fs/ntfs3/fslog.c
3517
e = data;
fs/ntfs3/fslog.c
3518
esize = le16_to_cpu(e->size);
fs/ntfs3/fslog.c
3530
Add2Ptr(e, esize) > Add2Ptr(lrh, rec_len) ||
fs/ntfs3/fslog.c
3537
memcpy(e1, e, esize);
fs/ntfs3/fslog.c
3549
e = Add2Ptr(ib, aoff);
fs/ntfs3/fslog.c
3550
esize = le16_to_cpu(e->size);
fs/ntfs3/fslog.c
3562
e1 = Add2Ptr(e, esize);
fs/ntfs3/fslog.c
3566
memmove(e, e1, PtrOffset(e1, Add2Ptr(hdr, used)));
fs/ntfs3/fslog.c
3578
e = Add2Ptr(ib, aoff);
fs/ntfs3/fslog.c
3591
hdr->used = cpu_to_le32(dlen + PtrOffset(hdr, e));
fs/ntfs3/fslog.c
3592
memmove(e, data, dlen);
fs/ntfs3/fslog.c
3601
e = Add2Ptr(ib, aoff);
fs/ntfs3/fslog.c
3613
de_set_vbn_le(e, *(__le64 *)data);
fs/ntfs3/fslog.c
3622
e = Add2Ptr(ib, aoff);
fs/ntfs3/fslog.c
3634
fname = (struct ATTR_FILE_NAME *)(e + 1);
fs/ntfs3/fslog.c
3670
e = Add2Ptr(ib, aoff);
fs/ntfs3/fslog.c
3682
memmove(Add2Ptr(e, le16_to_cpu(e->view.data_off)), data, dlen);
fs/ntfs3/fslog.c
623
__le32 *e;
fs/ntfs3/fslog.c
636
e = Add2Ptr(t, sizeof(struct RESTART_TABLE));
fs/ntfs3/fslog.c
638
e = Add2Ptr(c, rsize);
fs/ntfs3/fslog.c
642
for (bprt = bytes_per_rt(t); PtrOffset(t, e) < bprt;
fs/ntfs3/fslog.c
643
e = Add2Ptr(e, rsize)) {
fs/ntfs3/fslog.c
644
if (*e == RESTART_ENTRY_ALLOCATED_LE)
fs/ntfs3/fslog.c
645
return e;
fs/ntfs3/fslog.c
786
__le32 *e;
fs/ntfs3/fslog.c
790
e = Add2Ptr(rt, off);
fs/ntfs3/fslog.c
793
*e = rt->first_free;
fs/ntfs3/fslog.c
804
*e = 0;
fs/ntfs3/fslog.c
812
__le32 *e, *last_free;
fs/ntfs3/fslog.c
827
e = (__le32 *)(t + 1);
fs/ntfs3/fslog.c
830
for (off = sizeof(struct RESTART_TABLE) + esize; e < last_free;
fs/ntfs3/fslog.c
831
e = Add2Ptr(e, esize), off += esize) {
fs/ntfs3/fslog.c
832
*e = cpu_to_le32(off);
fs/ntfs3/fslog.c
877
__le32 *e;
fs/ntfs3/fslog.c
889
e = Add2Ptr(t, off);
fs/ntfs3/fslog.c
891
t->first_free = *e;
fs/ntfs3/fslog.c
893
memset(e, 0, le16_to_cpu(t->size));
fs/ntfs3/fslog.c
895
*e = RESTART_ENTRY_ALLOCATED_LE;
fs/ntfs3/fslog.c
914
__le32 *e;
fs/ntfs3/fslog.c
937
e = Add2Ptr(rt, vbo);
fs/ntfs3/fslog.c
939
if (*e == RESTART_ENTRY_ALLOCATED_LE)
fs/ntfs3/fslog.c
940
return e;
fs/ntfs3/fslog.c
947
e = Add2Ptr(rt, off);
fs/ntfs3/fslog.c
951
rt->first_free = *e;
fs/ntfs3/fslog.c
962
__le32 *last_e = e;
fs/ntfs3/fslog.c
968
e = Add2Ptr(rt, off);
fs/ntfs3/fslog.c
972
*last_e = *e;
fs/ntfs3/fslog.c
990
memset(e, 0, esize);
fs/ntfs3/fslog.c
991
*e = RESTART_ENTRY_ALLOCATED_LE;
fs/ntfs3/fslog.c
995
return e;
fs/ntfs3/fsntfs.c
2089
struct NTFS_DE_SDH *e;
fs/ntfs3/fsntfs.c
2142
&d_security->key.sec_id, &diff, (struct NTFS_DE **)&e,
fs/ntfs3/fsntfs.c
2147
while (e) {
fs/ntfs3/fsntfs.c
2148
if (le32_to_cpu(e->sec_hdr.size) == new_sec_size) {
fs/ntfs3/fsntfs.c
2150
le64_to_cpu(e->sec_hdr.off),
fs/ntfs3/fsntfs.c
2166
(struct NTFS_DE **)&e, fnd_sdh);
fs/ntfs3/fsntfs.c
2170
if (!e || e->key.hash != hash_key.hash)
fs/ntfs3/index.c
1142
struct NTFS_DE *e;
fs/ntfs3/index.c
1154
e = fnd->level ? fnd->de[fnd->level - 1] : fnd->root_de;
fs/ntfs3/index.c
1155
if (e && !de_is_last(e) &&
fs/ntfs3/index.c
1156
!(*indx->cmp)(key, key_len, e + 1, le16_to_cpu(e->key_size), ctx)) {
fs/ntfs3/index.c
1157
*entry = e;
fs/ntfs3/index.c
1166
e = hdr_find_e(indx, &root->ihdr, key, key_len, ctx, diff);
fs/ntfs3/index.c
1167
if (!e)
fs/ntfs3/index.c
1170
fnd->root_de = e;
fs/ntfs3/index.c
1174
if (*diff >= 0 || !de_has_vcn_ex(e))
fs/ntfs3/index.c
1178
err = indx_read(indx, ni, de_get_vbn(e), &node);
fs/ntfs3/index.c
1185
e = hdr_find_e(indx, &node->index->ihdr, key, key_len, ctx,
fs/ntfs3/index.c
1187
if (!e) {
fs/ntfs3/index.c
1192
err = fnd_push(fnd, node, e);
fs/ntfs3/index.c
1200
*entry = e;
fs/ntfs3/index.c
1210
struct NTFS_DE *e;
fs/ntfs3/index.c
1216
e = hdr_first_de(&root->ihdr);
fs/ntfs3/index.c
1217
if (!e)
fs/ntfs3/index.c
1220
fnd->root_de = e;
fs/ntfs3/index.c
1227
e = hdr_next_de(&root->ihdr, fnd->root_de);
fs/ntfs3/index.c
1228
if (!e)
fs/ntfs3/index.c
1230
fnd->root_de = e;
fs/ntfs3/index.c
1233
e = fnd->de[level - 1];
fs/ntfs3/index.c
1235
if (de_is_last(e))
fs/ntfs3/index.c
1238
e = hdr_next_de(&n->index->ihdr, e);
fs/ntfs3/index.c
1239
if (!e)
fs/ntfs3/index.c
1242
fnd->de[level - 1] = e;
fs/ntfs3/index.c
1250
while (de_has_vcn_ex(e)) {
fs/ntfs3/index.c
1251
if (le16_to_cpu(e->size) <
fs/ntfs3/index.c
1261
err = indx_read(indx, ni, de_get_vbn(e), &n);
fs/ntfs3/index.c
1266
e = hdr_first_de(&n->index->ihdr);
fs/ntfs3/index.c
1267
if (!e) {
fs/ntfs3/index.c
1272
fnd_push(fnd, n, e);
fs/ntfs3/index.c
1275
if (le16_to_cpu(e->size) > sizeof(struct NTFS_DE)) {
fs/ntfs3/index.c
1276
*entry = e;
fs/ntfs3/index.c
1282
if (!de_is_last(e))
fs/ntfs3/index.c
1295
e = fnd->de[level - 1];
fs/ntfs3/index.c
1298
e = fnd->root_de;
fs/ntfs3/index.c
1305
if (le16_to_cpu(e->size) > sizeof(struct NTFS_DE)) {
fs/ntfs3/index.c
1306
*entry = e;
fs/ntfs3/index.c
1308
fnd->root_de = e;
fs/ntfs3/index.c
1320
struct NTFS_DE *e = NULL;
fs/ntfs3/index.c
1330
e = hdr_first_de(&root->ihdr);
fs/ntfs3/index.c
1331
if (!e)
fs/ntfs3/index.c
1334
fnd->root_de = e;
fs/ntfs3/index.c
1351
if (e && le16_to_cpu(e->size) > sizeof(struct NTFS_DE))
fs/ntfs3/index.c
1357
e = hdr_next_de(&root->ihdr, fnd->root_de);
fs/ntfs3/index.c
1358
if (!e)
fs/ntfs3/index.c
1360
fnd->root_de = e;
fs/ntfs3/index.c
1373
e = hdr_next_de(&n->index->ihdr, e2);
fs/ntfs3/index.c
1374
if (!e)
fs/ntfs3/index.c
1376
fnd->de[fnd->level - 1] = e;
fs/ntfs3/index.c
1409
e = hdr_first_de(&n->index->ihdr);
fs/ntfs3/index.c
1410
fnd_push(fnd, n, e);
fs/ntfs3/index.c
1411
if (!e)
fs/ntfs3/index.c
1419
*off = PtrOffset(&root->ihdr, e);
fs/ntfs3/index.c
1423
record_size + PtrOffset(&n->index->ihdr, e);
fs/ntfs3/index.c
1426
*entry = e;
fs/ntfs3/index.c
1591
struct NTFS_DE *e, *e0, *re;
fs/ntfs3/index.c
1629
e = hdr_insert_de(indx, hdr, new_de, root_de, ctx);
fs/ntfs3/index.c
1630
WARN_ON(!e);
fs/ntfs3/index.c
1632
fnd->root_de = e;
fs/ntfs3/index.c
1650
for (e = e0;; e = hdr_next_de(hdr, e)) {
fs/ntfs3/index.c
1651
if (!e) {
fs/ntfs3/index.c
1656
if (de_is_last(e))
fs/ntfs3/index.c
1658
to_move += le16_to_cpu(e->size);
fs/ntfs3/index.c
1672
if (de_has_vcn(e)) {
fs/ntfs3/index.c
1673
t_vbn = de_get_vbn_le(e);
fs/ntfs3/index.c
1691
e = (struct NTFS_DE *)(root + 1);
fs/ntfs3/index.c
1692
memset(e, 0, sizeof(struct NTFS_DE));
fs/ntfs3/index.c
1693
e->size = cpu_to_le16(sizeof(struct NTFS_DE) + sizeof(u64));
fs/ntfs3/index.c
1694
e->flags = NTFS_IE_HAS_SUBNODES | NTFS_IE_LAST;
fs/ntfs3/index.c
1728
e = (struct NTFS_DE *)(root + 1);
fs/ntfs3/index.c
1729
*(__le64 *)(e + 1) = cpu_to_le64(new_vbn);
fs/ntfs3/index.c
1770
e = hdr_insert_de(indx, hdr, new_de, NULL, ctx);
fs/ntfs3/index.c
1771
if (!e) {
fs/ntfs3/index.c
1775
fnd_push(fnd, n, e);
fs/ntfs3/index.c
1804
struct NTFS_DE *e, *de_t, *up_e;
fs/ntfs3/index.c
1816
e = fnd->level - 1 == level ? fnd->de[level] : NULL;
fs/ntfs3/index.c
1817
e = hdr_insert_de(indx, hdr1, new_de, e, ctx);
fs/ntfs3/index.c
1818
fnd->de[level] = e;
fs/ntfs3/index.c
1819
if (e) {
fs/ntfs3/index.c
1952
struct NTFS_DE *e;
fs/ntfs3/index.c
1977
le16_to_cpu(new_de->key_size), ctx, &diff, &e,
fs/ntfs3/index.c
2019
const struct NTFS_DE *e;
fs/ntfs3/index.c
2024
for (e = hdr_first_de(hdr);; e = hdr_next_de(hdr, e)) {
fs/ntfs3/index.c
2025
if (!e)
fs/ntfs3/index.c
2028
if (de_has_vcn(e) && vbn == de_get_vbn_le(e))
fs/ntfs3/index.c
2031
if (de_is_last(e))
fs/ntfs3/index.c
2036
e = Add2Ptr(hdr, le32_to_cpu(hdr->de_off));
fs/ntfs3/index.c
2038
if (de_has_vcn_ex(e)) {
fs/ntfs3/index.c
2039
err = indx_read(indx, ni, de_get_vbn(e), &n);
fs/ntfs3/index.c
2048
if (de_is_last(e))
fs/ntfs3/index.c
2051
e = Add2Ptr(e, le16_to_cpu(e->size));
fs/ntfs3/index.c
2125
const struct NTFS_DE *e, bool trim)
fs/ntfs3/index.c
2130
CLST vbn = de_get_vbn(e);
fs/ntfs3/index.c
2140
for (e = hdr_first_de(hdr); e; e = hdr_next_de(hdr, e)) {
fs/ntfs3/index.c
2141
indx_free_children(indx, ni, e, false);
fs/ntfs3/index.c
2142
if (de_is_last(e))
fs/ntfs3/index.c
2183
struct NTFS_DE *e, *te, *re;
fs/ntfs3/index.c
2197
e = hdr_first_de(&n->index->ihdr);
fs/ntfs3/index.c
2198
fnd_push(fnd, n, e);
fs/ntfs3/index.c
2199
if (!e) {
fs/ntfs3/index.c
2204
if (!de_is_last(e)) {
fs/ntfs3/index.c
2212
if (!de_has_vcn(e))
fs/ntfs3/index.c
2216
vbn = de_get_vbn(e);
fs/ntfs3/index.c
2259
e = hdr_delete_de(&ib->ihdr, te);
fs/ntfs3/index.c
2261
fnd->de[level] = e;
fs/ntfs3/index.c
2285
struct NTFS_DE *e, *re, *next, *prev, *me;
fs/ntfs3/index.c
2314
err = indx_find(indx, ni, root, key, key_len, ctx, &diff, &e, fnd);
fs/ntfs3/index.c
2318
if (!e || diff) {
fs/ntfs3/index.c
2327
e = fnd->de[level - 1];
fs/ntfs3/index.c
2332
e = fnd->root_de;
fs/ntfs3/index.c
2337
e_size = le16_to_cpu(e->size);
fs/ntfs3/index.c
2339
if (!de_has_vcn_ex(e)) {
fs/ntfs3/index.c
2341
hdr_delete_de(hdr, e);
fs/ntfs3/index.c
2359
fnd_push(fnd2, n, e);
fs/ntfs3/index.c
2366
next = de_get_next(e);
fs/ntfs3/index.c
2373
de_set_vbn_le(re, de_get_vbn_le(e));
fs/ntfs3/index.c
2374
hdr_delete_de(hdr, e);
fs/ntfs3/index.c
2380
indx_insert_into_root(indx, ni, re, e,
fs/ntfs3/index.c
2396
de_set_vbn_le(next, de_get_vbn_le(e));
fs/ntfs3/index.c
2397
hdr_delete_de(hdr, e);
fs/ntfs3/index.c
2428
for (e = hdr_first_de(hdr);; e = hdr_next_de(hdr, e)) {
fs/ntfs3/index.c
2429
if (!e) {
fs/ntfs3/index.c
2434
if (de_has_vcn(e) && sub_vbn == de_get_vbn_le(e))
fs/ntfs3/index.c
2437
if (de_is_last(e)) {
fs/ntfs3/index.c
2438
e = NULL;
fs/ntfs3/index.c
2443
if (!e) {
fs/ntfs3/index.c
2484
e = hdr_first_de(hdr);
fs/ntfs3/index.c
2485
if (!e) {
fs/ntfs3/index.c
2490
if (hdr != &root->ihdr || !de_is_last(e)) {
fs/ntfs3/index.c
2492
while (!de_is_last(e)) {
fs/ntfs3/index.c
2493
if (de_has_vcn(e) && sub_vbn == de_get_vbn_le(e))
fs/ntfs3/index.c
2495
prev = e;
fs/ntfs3/index.c
2496
e = hdr_next_de(hdr, e);
fs/ntfs3/index.c
2497
if (!e) {
fs/ntfs3/index.c
2503
if (sub_vbn != de_get_vbn_le(e)) {
fs/ntfs3/index.c
2512
if (de_is_last(e)) {
fs/ntfs3/index.c
2526
de_set_vbn_le(e, de_get_vbn_le(prev));
fs/ntfs3/index.c
2527
} else if (de_has_vcn(e)) {
fs/ntfs3/index.c
2528
le16_sub_cpu(&e->size, sizeof(u64));
fs/ntfs3/index.c
2529
e->flags &= ~NTFS_IE_HAS_SUBNODES;
fs/ntfs3/index.c
2532
e = prev;
fs/ntfs3/index.c
2540
e_size = le16_to_cpu(e->size);
fs/ntfs3/index.c
2541
me = kmemdup(e, e_size, GFP_NOFS);
fs/ntfs3/index.c
2552
hdr_delete_de(hdr, e);
fs/ntfs3/index.c
2636
e = (struct NTFS_DE *)(root + 1);
fs/ntfs3/index.c
2637
e->ref.low = 0;
fs/ntfs3/index.c
2638
e->ref.high = 0;
fs/ntfs3/index.c
2639
e->ref.seq = 0;
fs/ntfs3/index.c
2640
e->size = cpu_to_le16(sizeof(struct NTFS_DE));
fs/ntfs3/index.c
2641
e->flags = NTFS_IE_LAST; // 0x02
fs/ntfs3/index.c
2642
e->key_size = 0;
fs/ntfs3/index.c
2643
e->res = 0;
fs/ntfs3/index.c
2670
struct NTFS_DE *e = NULL;
fs/ntfs3/index.c
2689
&diff, &e, fnd);
fs/ntfs3/index.c
2693
if (!e) {
fs/ntfs3/index.c
2703
e_fname = (struct ATTR_FILE_NAME *)(e + 1);
fs/ntfs3/index.c
560
const struct NTFS_DE *e = hdr_first_de(hdr);
fs/ntfs3/index.c
564
if (!e || de_is_last(e))
fs/ntfs3/index.c
567
esize = le16_to_cpu(e->size);
fs/ntfs3/index.c
569
const struct NTFS_DE *p = e;
fs/ntfs3/index.c
571
e = Add2Ptr(hdr, o);
fs/ntfs3/index.c
574
if (de_is_last(e))
fs/ntfs3/index.c
577
esize = le16_to_cpu(e->size);
fs/ntfs3/index.c
580
return e;
fs/ntfs3/index.c
592
struct NTFS_DE *e = hdr_first_de(hdr);
fs/ntfs3/index.c
595
if (!e)
fs/ntfs3/index.c
600
memmove(Add2Ptr(e, ins_bytes), e, to_move);
fs/ntfs3/index.c
601
memcpy(e, ins, ins_bytes);
fs/ntfs3/index.c
604
return e;
fs/ntfs3/index.c
672
struct NTFS_DE *e)
fs/ntfs3/index.c
679
fnd->de[i] = e;
fs/ntfs3/index.c
719
struct NTFS_DE *e, *found = NULL;
fs/ntfs3/index.c
740
e = Add2Ptr(hdr, off);
fs/ntfs3/index.c
741
e_size = le16_to_cpu(e->size);
fs/ntfs3/index.c
746
if (!de_is_last(e)) {
fs/ntfs3/index.c
758
e_key_len = le16_to_cpu(e->key_size);
fs/ntfs3/index.c
760
diff2 = (*cmp)(key, key_len, e + 1, e_key_len, ctx);
fs/ntfs3/index.c
765
if (de_is_last(e))
fs/ntfs3/index.c
778
found = e;
fs/ntfs3/index.c
781
return e;
fs/ntfs3/index.c
790
e = Add2Ptr(hdr, offs[mid_idx]);
fs/ntfs3/index.c
934
struct NTFS_DE *e;
fs/ntfs3/index.c
971
e = Add2Ptr(hdr, eo);
fs/ntfs3/index.c
974
e->flags = NTFS_IE_LAST | NTFS_IE_HAS_SUBNODES;
fs/ntfs3/index.c
975
e->size = cpu_to_le16(sizeof(struct NTFS_DE) + sizeof(u64));
fs/ntfs3/index.c
978
de_set_vbn_le(e, *sub_vbn);
fs/ntfs3/index.c
981
e->size = cpu_to_le16(sizeof(struct NTFS_DE));
fs/ntfs3/index.c
983
e->flags = NTFS_IE_LAST;
fs/ntfs3/inode.c
1260
struct NTFS_DE *e, *new_de = NULL;
fs/ntfs3/inode.c
1497
e = Add2Ptr(root, sizeof(struct INDEX_ROOT));
fs/ntfs3/inode.c
1498
e->size = cpu_to_le16(sizeof(struct NTFS_DE));
fs/ntfs3/inode.c
1499
e->flags = NTFS_IE_LAST;
fs/ntfs3/namei.c
24
struct NTFS_DE *e = buf;
fs/ntfs3/namei.c
26
struct ATTR_FILE_NAME *fname = (struct ATTR_FILE_NAME *)(e + 1);
fs/ntfs3/namei.c
29
e->ref.high = fname->home.high = 0;
fs/ntfs3/namei.c
56
e->size = cpu_to_le16(ALIGN(data_size, 8) + sizeof(struct NTFS_DE));
fs/ntfs3/namei.c
57
e->key_size = cpu_to_le16(data_size);
fs/ntfs3/namei.c
58
e->flags = 0;
fs/ntfs3/namei.c
59
e->res = 0;
fs/ntfs3/ntfs.h
639
static inline void de_set_vbn_le(struct NTFS_DE *e, __le64 vcn)
fs/ntfs3/ntfs.h
641
__le64 *v = Add2Ptr(e, le16_to_cpu(e->size) - sizeof(__le64));
fs/ntfs3/ntfs.h
646
static inline void de_set_vbn(struct NTFS_DE *e, CLST vcn)
fs/ntfs3/ntfs.h
648
__le64 *v = Add2Ptr(e, le16_to_cpu(e->size) - sizeof(__le64));
fs/ntfs3/ntfs.h
653
static inline __le64 de_get_vbn_le(const struct NTFS_DE *e)
fs/ntfs3/ntfs.h
655
return *(__le64 *)Add2Ptr(e, le16_to_cpu(e->size) - sizeof(__le64));
fs/ntfs3/ntfs.h
658
static inline CLST de_get_vbn(const struct NTFS_DE *e)
fs/ntfs3/ntfs.h
660
__le64 *v = Add2Ptr(e, le16_to_cpu(e->size) - sizeof(__le64));
fs/ntfs3/ntfs.h
665
static inline struct NTFS_DE *de_get_next(const struct NTFS_DE *e)
fs/ntfs3/ntfs.h
667
return Add2Ptr(e, le16_to_cpu(e->size));
fs/ntfs3/ntfs.h
670
static inline struct ATTR_FILE_NAME *de_get_fname(const struct NTFS_DE *e)
fs/ntfs3/ntfs.h
672
return le16_to_cpu(e->key_size) >= SIZEOF_ATTRIBUTE_FILENAME ?
fs/ntfs3/ntfs.h
673
Add2Ptr(e, sizeof(struct NTFS_DE)) :
fs/ntfs3/ntfs.h
677
static inline bool de_is_last(const struct NTFS_DE *e)
fs/ntfs3/ntfs.h
679
return e->flags & NTFS_IE_LAST;
fs/ntfs3/ntfs.h
682
static inline bool de_has_vcn(const struct NTFS_DE *e)
fs/ntfs3/ntfs.h
684
return e->flags & NTFS_IE_HAS_SUBNODES;
fs/ntfs3/ntfs.h
687
static inline bool de_has_vcn_ex(const struct NTFS_DE *e)
fs/ntfs3/ntfs.h
689
return (e->flags & NTFS_IE_HAS_SUBNODES) &&
fs/ntfs3/ntfs.h
690
(u64)(-1) != *((u64 *)Add2Ptr(e, le16_to_cpu(e->size) -
fs/ntfs3/ntfs.h
720
struct NTFS_DE *e;
fs/ntfs3/ntfs.h
726
e = Add2Ptr(hdr, de_off);
fs/ntfs3/ntfs.h
727
esize = le16_to_cpu(e->size);
fs/ntfs3/ntfs.h
731
return e;
fs/ntfs3/ntfs.h
735
const struct NTFS_DE *e)
fs/ntfs3/ntfs.h
737
size_t off = PtrOffset(hdr, e);
fs/ntfs3/ntfs.h
744
esize = le16_to_cpu(e->size);
fs/ntfs3/ntfs.h
750
return Add2Ptr(e, esize);
fs/ntfs3/run.c
1258
struct ntfs_run *r, *e, *eat_start, *eat_end;
fs/ntfs3/run.c
1273
e = run->runs + run->count;
fs/ntfs3/run.c
1297
for (; r < e; r++) {
fs/ntfs3/run.c
1317
memmove(eat_start, eat_end, (e - eat_end) * sizeof(*r));
fs/ntfs3/run.c
1325
const struct ntfs_run *r, *e;
fs/ntfs3/run.c
1328
for (r = run->runs, e = r + run->count; r < e; r++) {
fs/ntfs3/run.c
494
struct ntfs_run *r, *e, *eat_start, *eat_end;
fs/ntfs3/run.c
501
e = run->runs + run->count;
fs/ntfs3/run.c
525
for (; r < e; r++) {
fs/ntfs3/run.c
547
memmove(eat_start, eat_end, (e - eat_end) * sizeof(*r));
fs/ntfs3/run.c
551
e -= eat;
fs/ntfs3/run.c
552
for (r = run->runs; r < e; r++) {
fs/ntfs3/run.c
568
struct ntfs_run *r, *e;
fs/ntfs3/run.c
573
e = run->runs + run->count;
fs/ntfs3/run.c
579
for (; r < e; r++)
fs/ntfs3/run.c
609
struct ntfs_run *r, *r0 = NULL, *e = run->runs + run->count;
fs/ntfs3/run.c
612
for (r = run->runs; r < e; r++) {
fs/ntfs3/super.c
585
struct proc_dir_entry *e;
fs/ntfs3/super.c
590
e = proc_mkdir(sb->s_id, proc_info_root);
fs/ntfs3/super.c
591
if (e) {
fs/ntfs3/super.c
594
proc_create_data("volinfo", 0444, e, &ntfs3_volinfo_fops, sb);
fs/ntfs3/super.c
595
proc_create_data("label", 0644, e, &ntfs3_label_fops, sb);
fs/ntfs3/super.c
596
sbi->procdir = e;
fs/ocfs2/cluster/tcp.c
201
# define o2net_init_nst(a, b, c, d, e)
fs/ocfs2/dlmglue.h
143
#define ocfs2_inode_lock_full(i, r, e, f)\
fs/ocfs2/dlmglue.h
144
ocfs2_inode_lock_full_nested(i, r, e, f, OI_LS_NORMAL)
fs/ocfs2/dlmglue.h
145
#define ocfs2_inode_lock_nested(i, b, e, s)\
fs/ocfs2/dlmglue.h
146
ocfs2_inode_lock_full_nested(i, b, e, 0, s)
fs/ocfs2/dlmglue.h
149
#define ocfs2_inode_lock(i, b, e) ocfs2_inode_lock_full_nested(i, b, e, 0, OI_LS_NORMAL)
fs/ocfs2/dlmglue.h
150
#define ocfs2_try_inode_lock(i, b, e)\
fs/ocfs2/dlmglue.h
151
ocfs2_inode_lock_full_nested(i, b, e, OCFS2_META_LOCK_NOQUEUE,\
fs/overlayfs/inode.c
351
struct posix_acl_entry *e = &acl->a_entries[i];
fs/overlayfs/inode.c
352
switch (e->e_tag) {
fs/overlayfs/inode.c
354
vfsuid = make_vfsuid(idmap, fs_userns, e->e_uid);
fs/overlayfs/inode.c
355
e->e_uid = vfsuid_into_kuid(vfsuid);
fs/overlayfs/inode.c
358
vfsgid = make_vfsgid(idmap, fs_userns, e->e_gid);
fs/overlayfs/inode.c
359
e->e_gid = vfsgid_into_kgid(vfsgid);
fs/smb/client/dfs_cache.c
274
#define dump_tgts(e)
fs/smb/client/dfs_cache.c
275
#define dump_ce(e)
fs/smb/client/dfs_cache.c
559
const char *s = path, *e;
fs/smb/client/dfs_cache.c
579
e = path + strlen(path) - 1;
fs/smb/client/dfs_cache.c
580
while (e > s) {
fs/smb/client/dfs_cache.c
584
while (e > s && *e == sep)
fs/smb/client/dfs_cache.c
585
e--;
fs/smb/client/dfs_cache.c
586
if (e == s)
fs/smb/client/dfs_cache.c
589
len = e + 1 - path;
fs/smb/client/dfs_cache.c
598
while (e > s && *e != sep)
fs/smb/client/dfs_cache.c
599
e--;
fs/ubifs/recovery.c
1250
struct size_entry *e;
fs/ubifs/recovery.c
1254
e = rb_entry(parent, struct size_entry, rb);
fs/ubifs/recovery.c
1255
if (inum < e->inum)
fs/ubifs/recovery.c
1261
e = kzalloc_obj(struct size_entry);
fs/ubifs/recovery.c
1262
if (!e)
fs/ubifs/recovery.c
1265
e->inum = inum;
fs/ubifs/recovery.c
1266
e->i_size = i_size;
fs/ubifs/recovery.c
1267
e->d_size = d_size;
fs/ubifs/recovery.c
1268
e->exists = exists;
fs/ubifs/recovery.c
1270
rb_link_node(&e->rb, parent, p);
fs/ubifs/recovery.c
1271
rb_insert_color(&e->rb, &c->size_tree);
fs/ubifs/recovery.c
1284
struct size_entry *e;
fs/ubifs/recovery.c
1287
e = rb_entry(p, struct size_entry, rb);
fs/ubifs/recovery.c
1288
if (inum < e->inum)
fs/ubifs/recovery.c
1290
else if (inum > e->inum)
fs/ubifs/recovery.c
1293
return e;
fs/ubifs/recovery.c
1305
struct size_entry *e = find_ino(c, inum);
fs/ubifs/recovery.c
1307
if (!e)
fs/ubifs/recovery.c
1309
rb_erase(&e->rb, &c->size_tree);
fs/ubifs/recovery.c
1310
kfree(e);
fs/ubifs/recovery.c
1319
struct size_entry *e, *n;
fs/ubifs/recovery.c
1321
rbtree_postorder_for_each_entry_safe(e, n, &c->size_tree, rb) {
fs/ubifs/recovery.c
1322
iput(e->inode);
fs/ubifs/recovery.c
1323
kfree(e);
fs/ubifs/recovery.c
1358
struct size_entry *e;
fs/ubifs/recovery.c
1366
e = find_ino(c, inum);
fs/ubifs/recovery.c
1367
if (e) {
fs/ubifs/recovery.c
1368
e->i_size = new_size;
fs/ubifs/recovery.c
1369
e->exists = 1;
fs/ubifs/recovery.c
1378
e = find_ino(c, inum);
fs/ubifs/recovery.c
1379
if (e) {
fs/ubifs/recovery.c
1380
if (new_size > e->d_size)
fs/ubifs/recovery.c
1381
e->d_size = new_size;
fs/ubifs/recovery.c
1389
e = find_ino(c, inum);
fs/ubifs/recovery.c
1390
if (e)
fs/ubifs/recovery.c
1391
e->d_size = new_size;
fs/ubifs/recovery.c
1402
static int fix_size_in_place(struct ubifs_info *c, struct size_entry *e)
fs/ubifs/recovery.c
1411
ino_key_init(c, &key, e->inum);
fs/ubifs/recovery.c
1420
if (i_size >= e->d_size)
fs/ubifs/recovery.c
1428
ino->size = cpu_to_le64(e->d_size);
fs/ubifs/recovery.c
1442
(unsigned long)e->inum, lnum, offs, i_size, e->d_size);
fs/ubifs/recovery.c
1447
(unsigned long)e->inum, e->i_size, e->d_size, err);
fs/ubifs/recovery.c
1456
static int inode_fix_size(struct ubifs_info *c, struct size_entry *e)
fs/ubifs/recovery.c
1463
ubifs_assert(c, !e->inode);
fs/ubifs/recovery.c
1465
if (e->inode) {
fs/ubifs/recovery.c
1467
inode = e->inode;
fs/ubifs/recovery.c
1469
inode = ubifs_iget(c->vfs_sb, e->inum);
fs/ubifs/recovery.c
1473
if (inode->i_size >= e->d_size) {
fs/ubifs/recovery.c
1483
(unsigned long)e->inum,
fs/ubifs/recovery.c
1484
inode->i_size, e->d_size);
fs/ubifs/recovery.c
1488
inode->i_size = e->d_size;
fs/ubifs/recovery.c
1489
ui->ui_size = e->d_size;
fs/ubifs/recovery.c
1490
ui->synced_i_size = e->d_size;
fs/ubifs/recovery.c
1492
e->inode = inode;
fs/ubifs/recovery.c
1510
rb_erase(&e->rb, &c->size_tree);
fs/ubifs/recovery.c
1511
kfree(e);
fs/ubifs/recovery.c
1531
struct size_entry *e;
fs/ubifs/recovery.c
1534
e = rb_entry(this, struct size_entry, rb);
fs/ubifs/recovery.c
1538
if (!e->exists) {
fs/ubifs/recovery.c
1541
ino_key_init(c, &key, e->inum);
fs/ubifs/recovery.c
1548
(unsigned long)e->inum);
fs/ubifs/recovery.c
1549
err = ubifs_tnc_remove_ino(c, e->inum);
fs/ubifs/recovery.c
1555
e->exists = 1;
fs/ubifs/recovery.c
1556
e->i_size = le64_to_cpu(ino->size);
fs/ubifs/recovery.c
1560
if (e->exists && e->i_size < e->d_size) {
fs/ubifs/recovery.c
1569
err = fix_size_in_place(c, e);
fs/ubifs/recovery.c
1572
iput(e->inode);
fs/ubifs/recovery.c
1574
err = inode_fix_size(c, e);
fs/ubifs/recovery.c
1581
rb_erase(&e->rb, &c->size_tree);
fs/ubifs/recovery.c
1582
kfree(e);
fs/xfs/libxfs/xfs_fs.h
1134
} e;
fs/xfs/xfs_attr_item.c
469
static inline struct xfs_attr_intent *attri_entry(const struct list_head *e)
fs/xfs/xfs_attr_item.c
471
return list_entry(e, struct xfs_attr_intent, xattri_list);
fs/xfs/xfs_bmap_item.c
233
static inline struct xfs_bmap_intent *bi_entry(const struct list_head *e)
fs/xfs/xfs_bmap_item.c
235
return list_entry(e, struct xfs_bmap_intent, bi_list);
fs/xfs/xfs_bmap_util.c
690
xfs_fileoff_t s, e;
fs/xfs/xfs_bmap_util.c
701
e = startoffset_fsb + allocatesize_fsb;
fs/xfs/xfs_bmap_util.c
704
e += temp;
fs/xfs/xfs_bmap_util.c
705
div_u64_rem(e, extsz, &temp);
fs/xfs/xfs_bmap_util.c
707
e += extsz - temp;
fs/xfs/xfs_bmap_util.c
710
e = allocatesize_fsb;
fs/xfs/xfs_bmap_util.c
721
resblks = min_t(xfs_fileoff_t, (e - s),
fs/xfs/xfs_error.h
23
#define XFS_ERROR_REPORT(e, lvl, mp) \
fs/xfs/xfs_error.h
24
xfs_error_report(e, lvl, mp, __FILE__, __LINE__, __return_address)
fs/xfs/xfs_error.h
25
#define XFS_CORRUPTION_ERROR(e, lvl, mp, buf, bufsize) \
fs/xfs/xfs_error.h
26
xfs_corruption_error(e, lvl, mp, buf, bufsize, \
fs/xfs/xfs_exchmaps_item.c
273
static inline struct xfs_exchmaps_intent *xmi_entry(const struct list_head *e)
fs/xfs/xfs_exchmaps_item.c
275
return list_entry(e, struct xfs_exchmaps_intent, xmi_list);
fs/xfs/xfs_extfree_item.c
323
static inline struct xfs_extent_free_item *xefi_entry(const struct list_head *e)
fs/xfs/xfs_extfree_item.c
325
return list_entry(e, struct xfs_extent_free_item, xefi_list);
fs/xfs/xfs_healthmon.c
765
hme.e.lost.count = event->lostcount;
fs/xfs/xfs_healthmon.c
768
hme.e.shutdown.reasons = shutdown_mask(event->flags);
fs/xfs/xfs_healthmon.c
775
hme.e.fs.mask = xfs_healthmon_fs_mask(event->fsmask);
fs/xfs/xfs_healthmon.c
778
hme.e.group.mask = xfs_healthmon_rtgroup_mask(event->grpmask);
fs/xfs/xfs_healthmon.c
779
hme.e.group.gno = event->group;
fs/xfs/xfs_healthmon.c
782
hme.e.group.mask = xfs_healthmon_perag_mask(event->grpmask);
fs/xfs/xfs_healthmon.c
783
hme.e.group.gno = event->group;
fs/xfs/xfs_healthmon.c
786
hme.e.inode.mask = xfs_healthmon_inode_mask(event->imask);
fs/xfs/xfs_healthmon.c
787
hme.e.inode.ino = event->ino;
fs/xfs/xfs_healthmon.c
788
hme.e.inode.gen = event->gen;
fs/xfs/xfs_healthmon.c
793
hme.e.media.daddr = event->daddr;
fs/xfs/xfs_healthmon.c
794
hme.e.media.bbcount = event->bbcount;
fs/xfs/xfs_healthmon.c
797
hme.e.filerange.ino = event->fino;
fs/xfs/xfs_healthmon.c
798
hme.e.filerange.gen = event->fgen;
fs/xfs/xfs_healthmon.c
799
hme.e.filerange.pos = event->fpos;
fs/xfs/xfs_healthmon.c
800
hme.e.filerange.len = event->flen;
fs/xfs/xfs_healthmon.c
801
hme.e.filerange.error = abs(event->error);
fs/xfs/xfs_refcount_item.c
246
static inline struct xfs_refcount_intent *ci_entry(const struct list_head *e)
fs/xfs/xfs_refcount_item.c
248
return list_entry(e, struct xfs_refcount_intent, ri_list);
fs/xfs/xfs_rmap_item.c
247
static inline struct xfs_rmap_intent *ri_entry(const struct list_head *e)
fs/xfs/xfs_rmap_item.c
249
return list_entry(e, struct xfs_rmap_intent, ri_list);
include/asm-generic/vmlinux.lds.h
126
.text.e[_0-9A-Za-wy-z]* .text.e .text.e.* \
include/crypto/internal/rsa.h
34
const u8 *e;
include/drm/drm_edid.h
349
#define EDID_PRODUCT_ID(e) ((e)->prod_code[0] | ((e)->prod_code[1] << 8))
include/drm/drm_file.h
483
struct drm_event *e);
include/drm/drm_file.h
487
struct drm_event *e);
include/drm/drm_file.h
490
void drm_send_event_locked(struct drm_device *dev, struct drm_pending_event *e);
include/drm/drm_file.h
491
void drm_send_event(struct drm_device *dev, struct drm_pending_event *e);
include/drm/drm_file.h
493
struct drm_pending_event *e,
include/drm/drm_vblank.h
295
struct drm_pending_vblank_event *e);
include/drm/drm_vblank.h
297
struct drm_pending_vblank_event *e);
include/drm/drm_vblank.h
298
void drm_vblank_set_event(struct drm_pending_vblank_event *e,
include/linux/average.h
32
static inline void ewma_##name##_init(struct ewma_##name *e) \
include/linux/average.h
42
e->internal = 0; \
include/linux/average.h
45
ewma_##name##_read(struct ewma_##name *e) \
include/linux/average.h
51
return e->internal >> (_precision); \
include/linux/average.h
53
static inline void ewma_##name##_add(struct ewma_##name *e, \
include/linux/average.h
56
unsigned long internal = READ_ONCE(e->internal); \
include/linux/average.h
65
WRITE_ONCE(e->internal, internal ? \
include/linux/btf.h
321
static inline u64 btf_enum64_value(const struct btf_enum64 *e)
include/linux/btf.h
323
return ((u64)e->val_hi32 << 32) | e->val_lo32;
include/linux/build_bug.h
16
#define BUILD_BUG_ON_ZERO(e, ...) \
include/linux/build_bug.h
17
__BUILD_BUG_ON_ZERO_MSG(e, ##__VA_ARGS__, #e " is true")
include/linux/build_bug.h
30
#define BUILD_BUG_ON_INVALID(e) ((void)(sizeof((__force long)(e))))
include/linux/compiler.h
201
#define __BUILD_BUG_ON_ZERO_MSG(e, msg, ...) (0)
include/linux/compiler.h
203
#define __BUILD_BUG_ON_ZERO_MSG(e, msg, ...) ((int)sizeof(struct {_Static_assert(!(e), msg);}))
include/linux/crush/hash.h
22
__u32 e);
include/linux/find.h
623
#define for_each_set_bitrange(b, e, addr, size) \
include/linux/find.h
626
(e) = find_next_zero_bit((addr), (size), (b) + 1), \
include/linux/find.h
628
(b) = (e) + 1)
include/linux/find.h
637
#define for_each_set_bitrange_from(b, e, addr, size) \
include/linux/find.h
640
(e) = find_next_zero_bit((addr), (size), (b) + 1), \
include/linux/find.h
642
(b) = (e) + 1)
include/linux/find.h
651
#define for_each_clear_bitrange(b, e, addr, size) \
include/linux/find.h
654
(e) = find_next_bit((addr), (size), (b) + 1), \
include/linux/find.h
656
(b) = (e) + 1)
include/linux/find.h
665
#define for_each_clear_bitrange_from(b, e, addr, size) \
include/linux/find.h
668
(e) = find_next_bit((addr), (size), (b) + 1), \
include/linux/find.h
670
(b) = (e) + 1)
include/linux/kvm_host.h
1791
int kvm_arch_set_irq_inatomic(struct kvm_kernel_irq_routing_entry *e,
include/linux/kvm_host.h
2194
struct kvm_kernel_irq_routing_entry *e,
include/linux/kvm_host.h
669
int (*set)(struct kvm_kernel_irq_routing_entry *e,
include/linux/lru_cache.h
249
extern unsigned int lc_put(struct lru_cache *lc, struct lc_element *e);
include/linux/lwq.h
118
struct llist_node *e = n;
include/linux/lwq.h
121
return llist_add_batch(llist_reverse_order(n), e, &q->new) &&
include/linux/mlx5/mlx5_ifc.h
10571
u8 e[0x1];
include/linux/mlx5/mlx5_ifc.h
10604
u8 e[0x2];
include/linux/mlx5/mlx5_ifc.h
11071
u8 e[0x2];
include/linux/mlx5/mlx5_ifc.h
11994
u8 e[0x1];
include/linux/netdevice.h
127
#define net_xmit_eval(e) ((e) == NET_XMIT_CN ? 0 : (e))
include/linux/netdevice.h
128
#define net_xmit_errno(e) ((e) != NET_XMIT_CN ? -ENOBUFS : 0)
include/linux/netfilter.h
144
static inline struct nf_hook_ops **nf_hook_entries_get_hook_ops(const struct nf_hook_entries *e)
include/linux/netfilter.h
146
unsigned int n = e->num_hook_entries;
include/linux/netfilter.h
149
hook_end = &e->hooks[n]; /* this is *past* ->hooks[]! */
include/linux/netfilter.h
216
const struct nf_hook_entries *e, unsigned int i);
include/linux/netfilter.h
219
const struct nf_hook_entries *e);
include/linux/netfilter/ipset/ip_set.h
127
#define ext_timeout(e, s) \
include/linux/netfilter/ipset/ip_set.h
128
((unsigned long *)(((void *)(e)) + (s)->offset[IPSET_EXT_ID_TIMEOUT]))
include/linux/netfilter/ipset/ip_set.h
129
#define ext_counter(e, s) \
include/linux/netfilter/ipset/ip_set.h
130
((struct ip_set_counter *)(((void *)(e)) + (s)->offset[IPSET_EXT_ID_COUNTER]))
include/linux/netfilter/ipset/ip_set.h
131
#define ext_comment(e, s) \
include/linux/netfilter/ipset/ip_set.h
132
((struct ip_set_comment *)(((void *)(e)) + (s)->offset[IPSET_EXT_ID_COMMENT]))
include/linux/netfilter/ipset/ip_set.h
133
#define ext_skbinfo(e, s) \
include/linux/netfilter/ipset/ip_set.h
134
((struct ip_set_skbinfo *)(((void *)(e)) + (s)->offset[IPSET_EXT_ID_SKBINFO]))
include/linux/netfilter/ipset/ip_set.h
340
const void *e, bool active);
include/linux/netfilter_arp/arp_tables.h
73
compat_arpt_get_target(struct compat_arpt_entry *e)
include/linux/netfilter_arp/arp_tables.h
75
return (void *)e + e->target_offset;
include/linux/netfilter_bridge/ebtables.h
75
struct ebt_entry *e; /* pointer to entry data */
include/linux/netfilter_ipv4/ip_tables.h
85
compat_ipt_get_target(struct compat_ipt_entry *e)
include/linux/netfilter_ipv4/ip_tables.h
87
return (void *)e + e->target_offset;
include/linux/netfilter_ipv6/ip6_tables.h
49
compat_ip6t_get_target(struct compat_ip6t_entry *e)
include/linux/netfilter_ipv6/ip6_tables.h
51
return (void *)e + e->target_offset;
include/linux/netfilter_netdev.h
108
ret = nf_hook_slow(skb, &state, e, 0);
include/linux/netfilter_netdev.h
21
struct nf_hook_entries *e = rcu_dereference(skb->dev->nf_hooks_ingress);
include/linux/netfilter_netdev.h
28
if (unlikely(!e))
include/linux/netfilter_netdev.h
34
ret = nf_hook_slow(skb, &state, e, 0);
include/linux/netfilter_netdev.h
89
struct nf_hook_entries *e;
include/linux/netfilter_netdev.h
98
e = rcu_dereference_check(dev->nf_hooks_egress, rcu_read_lock_bh_held());
include/linux/netfilter_netdev.h
99
if (!e)
include/linux/pinctrl/pinconf-generic.h
192
#define PCONFDUMP_WITH_VALUES(a, b, c, d, e, f) { \
include/linux/pinctrl/pinconf-generic.h
194
.values = e, .num_values = f \
include/linux/platform_data/cros_ec_commands.h
5808
#define EC_MKBP_PCHG_EVENT_TO_PORT(e) (((e) >> EC_MKBP_PCHG_PORT_SHIFT) & 0xf)
include/linux/platform_data/cros_ec_commands.h
5810
#define EC_MKBP_PCHG_EVENT_MASK(e) ((e) \
include/linux/suspend.h
383
extern void register_nosave_region(unsigned long b, unsigned long e);
include/linux/suspend.h
406
static inline void register_nosave_region(unsigned long b, unsigned long e) {}
include/linux/uaccess.h
630
#define unsafe_get_user(x,p,e) unsafe_op_wrap(__get_user(x,p),e)
include/linux/uaccess.h
631
#define unsafe_put_user(x,p,e) unsafe_op_wrap(__put_user(x,p),e)
include/linux/uaccess.h
632
#define unsafe_copy_to_user(d,s,l,e) unsafe_op_wrap(__copy_to_user(d,s,l),e)
include/linux/uaccess.h
633
#define unsafe_copy_from_user(d,s,l,e) unsafe_op_wrap(__copy_from_user(d,s,l),e)
include/media/v4l2-ioctl.h
413
struct v4l2_exportbuffer *e);
include/net/bluetooth/hci_core.h
946
static inline long inquiry_entry_age(struct inquiry_entry *e)
include/net/bluetooth/hci_core.h
948
return jiffies - e->timestamp;
include/net/dsa.h
998
struct ethtool_keee *e);
include/net/fou.h
12
size_t fou_encap_hlen(struct ip_tunnel_encap *e);
include/net/fou.h
13
size_t gue_encap_hlen(struct ip_tunnel_encap *e);
include/net/fou.h
15
int __fou_build_header(struct sk_buff *skb, struct ip_tunnel_encap *e,
include/net/fou.h
17
int __gue_build_header(struct sk_buff *skb, struct ip_tunnel_encap *e,
include/net/ip6_tunnel.h
101
ops = rcu_dereference(ip6tun_encaps[e->type]);
include/net/ip6_tunnel.h
103
hlen = ops->encap_hlen(e);
include/net/ip6_tunnel.h
70
size_t (*encap_hlen)(struct ip_tunnel_encap *e);
include/net/ip6_tunnel.h
71
int (*build_header)(struct sk_buff *skb, struct ip_tunnel_encap *e,
include/net/ip6_tunnel.h
89
static inline int ip6_encap_hlen(struct ip_tunnel_encap *e)
include/net/ip6_tunnel.h
94
if (e->type == TUNNEL_ENCAP_NONE)
include/net/ip6_tunnel.h
97
if (e->type >= MAX_IPTUN_ENCAP_OPS)
include/net/ip_tunnels.h
444
size_t (*encap_hlen)(struct ip_tunnel_encap *e);
include/net/ip_tunnels.h
445
int (*build_header)(struct sk_buff *skb, struct ip_tunnel_encap *e,
include/net/ip_tunnels.h
526
static inline int ip_encap_hlen(struct ip_tunnel_encap *e)
include/net/ip_tunnels.h
531
if (e->type == TUNNEL_ENCAP_NONE)
include/net/ip_tunnels.h
534
if (e->type >= MAX_IPTUN_ENCAP_OPS)
include/net/ip_tunnels.h
538
ops = rcu_dereference(iptun_encaps[e->type]);
include/net/ip_tunnels.h
540
hlen = ops->encap_hlen(e);
include/net/ip_tunnels.h
547
struct ip_tunnel_encap *e,
include/net/ip_tunnels.h
553
if (e->type == TUNNEL_ENCAP_NONE)
include/net/ip_tunnels.h
556
if (e->type >= MAX_IPTUN_ENCAP_OPS)
include/net/ip_tunnels.h
560
ops = rcu_dereference(iptun_encaps[e->type]);
include/net/ip_tunnels.h
562
ret = ops->build_header(skb, e, protocol, fl4);
include/net/mana/gdma.h
295
struct gdma_event *e);
include/net/netfilter/nf_conntrack_ecache.h
106
struct nf_conntrack_ecache *e;
include/net/netfilter/nf_conntrack_ecache.h
111
e = nf_ct_ecache_find(ct);
include/net/netfilter/nf_conntrack_ecache.h
112
if (e == NULL)
include/net/netfilter/nf_conntrack_ecache.h
119
if (local64_read(&e->timestamp) && READ_ONCE(e->cache) == 0)
include/net/netfilter/nf_conntrack_ecache.h
120
local64_set(&e->timestamp, ktime_get_real_ns());
include/net/netfilter/nf_conntrack_ecache.h
123
set_bit(event, &e->cache);
include/net/netfilter/nf_conntrack_ecache.h
166
static inline void nf_ct_expect_event_report(enum ip_conntrack_expect_events e,
include/net/netfilter/nf_conntrack_expect.h
132
void nf_ct_expect_iterate_destroy(bool (*iter)(struct nf_conntrack_expect *e, void *data), void *data);
include/net/netfilter/nf_conntrack_expect.h
134
bool (*iter)(struct nf_conntrack_expect *e, void *data),
include/net/sch_generic.h
906
#define net_xmit_drop_count(e) ((e) & __NET_XMIT_STOLEN ? 0 : 1)
include/net/sch_generic.h
908
#define net_xmit_drop_count(e) (1)
include/soc/fsl/qman.h
431
u32 e = 0;
include/soc/fsl/qman.h
440
e++;
include/soc/fsl/qman.h
446
td |= (e & QM_FQD_TD_EXP_MASK);
include/soc/fsl/qman.h
588
u32 e = 0;
include/soc/fsl/qman.h
594
e++;
include/soc/fsl/qman.h
598
th->word = cpu_to_be16(((val & 0xff) << 5) | (e & 0x1f));
include/sound/emu8000_reg.h
15
#define EMU8000_DATA0(e) ((e)->port1)
include/sound/emu8000_reg.h
16
#define EMU8000_DATA1(e) ((e)->port2)
include/sound/emu8000_reg.h
17
#define EMU8000_DATA2(e) ((e)->port2+2)
include/sound/emu8000_reg.h
18
#define EMU8000_DATA3(e) ((e)->port3)
include/sound/emu8000_reg.h
19
#define EMU8000_PTR(e) ((e)->port3+2)
include/sound/pcm_drm_eld.h
89
int snd_parse_eld(struct device *dev, struct snd_parsed_hdmi_eld *e,
include/sound/pcm_drm_eld.h
91
void snd_show_eld(struct device *dev, struct snd_parsed_hdmi_eld *e);
include/sound/soc-dapm.h
398
#define SND_SOC_DAPM_EVENT_ON(e) (e & (SND_SOC_DAPM_PRE_PMU | SND_SOC_DAPM_POST_PMU))
include/sound/soc-dapm.h
399
#define SND_SOC_DAPM_EVENT_OFF(e) (e & (SND_SOC_DAPM_PRE_PMD | SND_SOC_DAPM_POST_PMD))
include/sound/soc-dapm.h
669
struct snd_kcontrol *kcontrol, int mux, struct soc_enum *e,
include/sound/soc.h
1299
static inline unsigned int snd_soc_enum_val_to_item(const struct soc_enum *e,
include/sound/soc.h
1304
if (!e->values)
include/sound/soc.h
1307
for (i = 0; i < e->items; i++)
include/sound/soc.h
1308
if (val == e->values[i])
include/sound/soc.h
1314
static inline unsigned int snd_soc_enum_item_to_val(const struct soc_enum *e,
include/sound/soc.h
1317
if (!e->values)
include/sound/soc.h
1320
return e->values[item];
include/uapi/linux/cn_proc.h
120
} e;
include/uapi/linux/coff.h
250
} e;
include/uapi/linux/coff.h
251
} e;
include/uapi/linux/map_to_14segment.h
107
#define _SEG14(sym, a, b, c, d, e, f, g1, g2, h, j, k, l, m, n) \
include/uapi/linux/map_to_14segment.h
110
e << BIT_SEG14_E | f << BIT_SEG14_F | \
include/uapi/linux/map_to_7segment.h
89
#define _SEG7(l,a,b,c,d,e,f,g) \
include/uapi/linux/map_to_7segment.h
91
e<<BIT_SEG7_E | f<<BIT_SEG7_F | g<<BIT_SEG7_G )
include/uapi/linux/netfilter/x_tables.h
129
#define XT_MATCH_ITERATE(type, e, fn, args...) \
include/uapi/linux/netfilter/x_tables.h
136
__i < (e)->target_offset; \
include/uapi/linux/netfilter/x_tables.h
138
__m = (void *)e + __i; \
include/uapi/linux/netfilter_arp/arp_tables.h
200
static __inline__ struct xt_entry_target *arpt_get_target(struct arpt_entry *e)
include/uapi/linux/netfilter_arp/arp_tables.h
202
return (struct xt_entry_target *)((char *)e + e->target_offset);
include/uapi/linux/netfilter_bridge/ebtables.h
197
ebt_get_target(struct ebt_entry *e)
include/uapi/linux/netfilter_bridge/ebtables.h
199
return (struct ebt_entry_target *)((char *)e + e->target_offset);
include/uapi/linux/netfilter_bridge/ebtables.h
218
#define EBT_MATCH_ITERATE(e, fn, args...) \
include/uapi/linux/netfilter_bridge/ebtables.h
225
__i < (e)->watchers_offset; \
include/uapi/linux/netfilter_bridge/ebtables.h
228
__match = (void *)(e) + __i; \
include/uapi/linux/netfilter_bridge/ebtables.h
235
if (__i != (e)->watchers_offset) \
include/uapi/linux/netfilter_bridge/ebtables.h
241
#define EBT_WATCHER_ITERATE(e, fn, args...) \
include/uapi/linux/netfilter_bridge/ebtables.h
247
for (__i = e->watchers_offset; \
include/uapi/linux/netfilter_bridge/ebtables.h
248
__i < (e)->target_offset; \
include/uapi/linux/netfilter_bridge/ebtables.h
251
__watcher = (void *)(e) + __i; \
include/uapi/linux/netfilter_bridge/ebtables.h
258
if (__i != (e)->target_offset) \
include/uapi/linux/netfilter_ipv4/ip_tables.h
223
ipt_get_target(struct ipt_entry *e)
include/uapi/linux/netfilter_ipv4/ip_tables.h
225
return (struct xt_entry_target *)((char *)e + e->target_offset);
include/uapi/linux/netfilter_ipv4/ip_tables.h
62
#define IPT_MATCH_ITERATE(e, fn, args...) \
include/uapi/linux/netfilter_ipv4/ip_tables.h
63
XT_MATCH_ITERATE(struct ipt_entry, e, fn, ## args)
include/uapi/linux/netfilter_ipv6/ip6_tables.h
263
ip6t_get_target(struct ip6t_entry *e)
include/uapi/linux/netfilter_ipv6/ip6_tables.h
265
return (struct xt_entry_target *)((char *)e + e->target_offset);
include/uapi/linux/netfilter_ipv6/ip6_tables.h
57
#define IP6T_MATCH_ITERATE(e, fn, args...) \
include/uapi/linux/netfilter_ipv6/ip6_tables.h
58
XT_MATCH_ITERATE(struct ip6t_entry, e, fn, ## args)
include/uapi/linux/wireless.h
708
__s16 e; /* Exponent */
include/video/pm3fb.h
560
#define PM3FBDestReadEnables_E(e) ((e) & 0xff)
io_uring/napi.c
100
hash_del_rcu(&e->node);
io_uring/napi.c
101
kfree_rcu(e, rcu);
io_uring/napi.c
107
struct io_napi_entry *e;
io_uring/napi.c
116
list_for_each_entry(e, &ctx->napi_list, list) {
io_uring/napi.c
117
if (time_after(jiffies, READ_ONCE(e->timeout))) {
io_uring/napi.c
118
list_del_rcu(&e->list);
io_uring/napi.c
119
hash_del_rcu(&e->node);
io_uring/napi.c
120
kfree_rcu(e, rcu);
io_uring/napi.c
167
struct io_napi_entry *e;
io_uring/napi.c
169
list_for_each_entry_rcu(e, &ctx->napi_list, list)
io_uring/napi.c
170
napi_busy_loop_rcu(e->napi_id, loop_end, loop_end_arg,
io_uring/napi.c
180
struct io_napi_entry *e;
io_uring/napi.c
183
list_for_each_entry_rcu(e, &ctx->napi_list, list) {
io_uring/napi.c
184
napi_busy_loop_rcu(e->napi_id, loop_end, loop_end_arg,
io_uring/napi.c
187
if (time_after(jiffies, READ_ONCE(e->timeout)))
io_uring/napi.c
24
struct io_napi_entry *e;
io_uring/napi.c
256
struct io_napi_entry *e;
io_uring/napi.c
259
list_for_each_entry(e, &ctx->napi_list, list) {
io_uring/napi.c
26
hlist_for_each_entry_rcu(e, hash_list, node) {
io_uring/napi.c
260
hash_del_rcu(&e->node);
io_uring/napi.c
261
kfree_rcu(e, rcu);
io_uring/napi.c
27
if (e->napi_id != napi_id)
io_uring/napi.c
29
return e;
io_uring/napi.c
44
struct io_napi_entry *e;
io_uring/napi.c
53
e = io_napi_hash_find(hash_list, napi_id);
io_uring/napi.c
54
if (e) {
io_uring/napi.c
55
WRITE_ONCE(e->timeout, jiffies + NAPI_TIMEOUT);
io_uring/napi.c
60
e = kmalloc(sizeof(*e), GFP_NOWAIT);
io_uring/napi.c
61
if (!e)
io_uring/napi.c
64
e->napi_id = napi_id;
io_uring/napi.c
65
e->timeout = jiffies + NAPI_TIMEOUT;
io_uring/napi.c
74
kfree(e);
io_uring/napi.c
78
hlist_add_tail_rcu(&e->node, hash_list);
io_uring/napi.c
79
list_add_tail_rcu(&e->list, &ctx->napi_list);
io_uring/napi.c
87
struct io_napi_entry *e;
io_uring/napi.c
95
e = io_napi_hash_find(hash_list, napi_id);
io_uring/napi.c
96
if (!e)
io_uring/napi.c
99
list_del_rcu(&e->list);
io_uring/query.c
20
struct io_uring_query_opcode *e = &data->opcodes;
io_uring/query.c
22
e->nr_request_opcodes = IORING_OP_LAST;
io_uring/query.c
23
e->nr_register_opcodes = IORING_REGISTER_LAST;
io_uring/query.c
24
e->feature_flags = IORING_FEAT_FLAGS;
io_uring/query.c
25
e->ring_setup_flags = IORING_SETUP_FLAGS;
io_uring/query.c
26
e->enter_flags = IORING_ENTER_FLAGS;
io_uring/query.c
27
e->sqe_flags = SQE_VALID_FLAGS;
io_uring/query.c
28
e->nr_query_opcodes = __IO_URING_QUERY_MAX;
io_uring/query.c
29
e->__pad = 0;
io_uring/query.c
30
return sizeof(*e);
io_uring/query.c
35
struct io_uring_query_zcrx *e = &data->zcrx;
io_uring/query.c
37
e->register_flags = ZCRX_REG_IMPORT;
io_uring/query.c
38
e->area_flags = IORING_ZCRX_AREA_DMABUF;
io_uring/query.c
39
e->nr_ctrl_opcodes = __ZCRX_CTRL_LAST;
io_uring/query.c
40
e->rq_hdr_size = sizeof(struct io_uring);
io_uring/query.c
41
e->rq_hdr_alignment = L1_CACHE_BYTES;
io_uring/query.c
42
e->features = ZCRX_FEATURE_RX_PAGE_SIZE;
io_uring/query.c
43
e->__resv2 = 0;
io_uring/query.c
44
return sizeof(*e);
io_uring/query.c
49
struct io_uring_query_scq *e = &data->scq;
io_uring/query.c
51
e->hdr_size = sizeof(struct io_rings);
io_uring/query.c
52
e->hdr_alignment = SMP_CACHE_BYTES;
io_uring/query.c
53
return sizeof(*e);
kernel/audit_watch.c
326
struct audit_entry *e;
kernel/audit_watch.c
331
e = container_of(r, struct audit_entry, rule);
kernel/audit_watch.c
333
if (e->rule.exe)
kernel/audit_watch.c
334
audit_remove_mark(e->rule.exe);
kernel/audit_watch.c
337
list_del_rcu(&e->list);
kernel/audit_watch.c
338
call_rcu(&e->rcu, audit_free_rule_rcu);
kernel/auditfilter.c
101
struct audit_entry *e = container_of(head, struct audit_entry, rcu);
kernel/auditfilter.c
102
audit_free_rule(e);
kernel/auditfilter.c
1025
struct audit_entry *e;
kernel/auditfilter.c
1042
e = audit_find_rule(entry, &list);
kernel/auditfilter.c
1043
if (!e) {
kernel/auditfilter.c
1048
if (e->rule.watch)
kernel/auditfilter.c
1049
audit_remove_watch_rule(&e->rule);
kernel/auditfilter.c
1051
if (e->rule.tree)
kernel/auditfilter.c
1052
audit_remove_tree_rule(&e->rule);
kernel/auditfilter.c
1054
if (e->rule.exe)
kernel/auditfilter.c
1055
audit_remove_mark_rule(&e->rule);
kernel/auditfilter.c
1065
list_del_rcu(&e->list);
kernel/auditfilter.c
1066
list_del(&e->rule.list);
kernel/auditfilter.c
1067
call_rcu(&e->rcu, audit_free_rule_rcu);
kernel/auditfilter.c
1339
struct audit_entry *e;
kernel/auditfilter.c
1343
list_for_each_entry_rcu(e, &audit_filter_list[listtype], list) {
kernel/auditfilter.c
1346
for (i = 0; i < e->rule.field_count; i++) {
kernel/auditfilter.c
1347
struct audit_field *f = &e->rule.fields[i];
kernel/auditfilter.c
1386
result = audit_exe_compare(current, e->rule.exe);
kernel/auditfilter.c
1399
if (e->rule.action == AUDIT_NEVER || listtype == AUDIT_FILTER_EXCLUDE)
kernel/auditfilter.c
83
static inline void audit_free_rule(struct audit_entry *e)
kernel/auditfilter.c
86
struct audit_krule *erule = &e->rule;
kernel/auditfilter.c
903
struct audit_entry *e, *found = NULL;
kernel/auditfilter.c
914
list_for_each_entry(e, list, list)
kernel/auditfilter.c
915
if (!audit_compare_rule(&entry->rule, &e->rule)) {
kernel/auditfilter.c
916
found = e;
kernel/auditfilter.c
925
list_for_each_entry(e, list, list)
kernel/auditfilter.c
926
if (!audit_compare_rule(&entry->rule, &e->rule)) {
kernel/auditfilter.c
927
found = e;
kernel/auditfilter.c
941
struct audit_entry *e;
kernel/auditfilter.c
959
e = audit_find_rule(entry, &list);
kernel/auditfilter.c
96
kfree(e);
kernel/auditfilter.c
960
if (e) {
kernel/auditsc.c
2252
struct audit_entry *e;
kernel/auditsc.c
2260
list_for_each_entry_rcu(e, list, list) {
kernel/auditsc.c
2261
for (i = 0; i < e->rule.field_count; i++) {
kernel/auditsc.c
2262
struct audit_field *f = &e->rule.fields[i];
kernel/auditsc.c
2267
&& e->rule.action == AUDIT_NEVER) {
kernel/auditsc.c
2370
struct audit_entry *e;
kernel/auditsc.c
2378
list_for_each_entry_rcu(e, list, list) {
kernel/auditsc.c
2379
for (i = 0; i < e->rule.field_count; i++) {
kernel/auditsc.c
2380
struct audit_field *f = &e->rule.fields[i];
kernel/auditsc.c
2385
&& e->rule.action == AUDIT_NEVER) {
kernel/auditsc.c
780
struct audit_entry *e;
kernel/auditsc.c
784
list_for_each_entry_rcu(e, &audit_filter_list[AUDIT_FILTER_TASK], list) {
kernel/auditsc.c
785
if (audit_filter_rules(tsk, &e->rule, NULL, NULL,
kernel/auditsc.c
788
*key = kstrdup(e->rule.filterkey, GFP_ATOMIC);
kernel/auditsc.c
833
struct audit_entry *e;
kernel/auditsc.c
836
list_for_each_entry_rcu(e, list, list) {
kernel/auditsc.c
837
if (audit_in_mask(&e->rule, op) &&
kernel/auditsc.c
838
audit_filter_rules(tsk, &e->rule, ctx, name,
kernel/bpf/core.c
785
const struct exception_table_entry *e = NULL;
kernel/bpf/core.c
795
e = search_extable(prog->aux->extable, prog->aux->num_exentries, addr);
kernel/bpf/core.c
798
return e;
kernel/bpf/inode.c
640
const struct btf_enum *e;
kernel/bpf/inode.c
650
e = &btf_enum(enum_t)[i];
kernel/bpf/inode.c
652
name = btf_name_by_offset(btf, e->name_off);
kernel/bpf/inode.c
658
*value = e->val;
kernel/bpf/inode.c
673
const struct btf_enum *e;
kernel/bpf/inode.c
692
e = &btf_enum(enum_t)[i];
kernel/bpf/inode.c
693
name = btf_name_by_offset(btf, e->name_off);
kernel/bpf/inode.c
696
msk = 1ULL << e->val;
kernel/bpf/verifier.c
17372
struct linked_reg *e;
kernel/bpf/verifier.c
17377
e = linked_regs_push(reg_set);
kernel/bpf/verifier.c
17378
if (e) {
kernel/bpf/verifier.c
17379
e->frameno = frameno;
kernel/bpf/verifier.c
17380
e->is_reg = is_reg;
kernel/bpf/verifier.c
17381
e->regno = spi_or_reg;
kernel/bpf/verifier.c
17429
struct linked_reg *e;
kernel/bpf/verifier.c
17433
e = &linked_regs->entries[i];
kernel/bpf/verifier.c
17434
reg = e->is_reg ? &vstate->frame[e->frameno]->regs[e->regno]
kernel/bpf/verifier.c
17435
: &vstate->frame[e->frameno]->stack[e->spi].spilled_ptr;
kernel/bpf/verifier.c
17477
if (e->is_reg)
kernel/bpf/verifier.c
17478
mark_reg_scratched(env, e->regno);
kernel/bpf/verifier.c
17480
mark_stack_slot_scratched(env, e->spi);
kernel/bpf/verifier.c
18248
static int push_insn(int t, int w, int e, struct bpf_verifier_env *env)
kernel/bpf/verifier.c
18253
if (e == FALLTHROUGH && insn_state[t] >= (DISCOVERED | FALLTHROUGH))
kernel/bpf/verifier.c
18256
if (e == BRANCH && insn_state[t] >= (DISCOVERED | BRANCH))
kernel/bpf/verifier.c
18265
if (e == BRANCH) {
kernel/bpf/verifier.c
18273
insn_state[t] = DISCOVERED | e;
kernel/bpf/verifier.c
18288
insn_state[t] = DISCOVERED | e;
kernel/bpf/verifier.c
4053
struct linked_reg *e = &s->entries[i];
kernel/bpf/verifier.c
4056
tmp |= e->frameno;
kernel/bpf/verifier.c
4057
tmp |= e->spi << LR_SPI_OFF;
kernel/bpf/verifier.c
4058
tmp |= (e->is_reg ? 1 : 0) << LR_IS_REG_OFF;
kernel/bpf/verifier.c
4076
struct linked_reg *e = &s->entries[i];
kernel/bpf/verifier.c
4078
e->frameno = val & LR_FRAMENO_MASK;
kernel/bpf/verifier.c
4079
e->spi = (val >> LR_SPI_OFF) & LR_SPI_MASK;
kernel/bpf/verifier.c
4080
e->is_reg = (val >> LR_IS_REG_OFF) & 0x1;
kernel/bpf/verifier.c
4359
struct linked_reg *e = &linked_regs.entries[i];
kernel/bpf/verifier.c
4361
if ((e->is_reg && bt_is_frame_reg_set(bt, e->frameno, e->regno)) ||
kernel/bpf/verifier.c
4362
(!e->is_reg && bt_is_frame_slot_set(bt, e->frameno, e->spi))) {
kernel/bpf/verifier.c
4372
struct linked_reg *e = &linked_regs.entries[i];
kernel/bpf/verifier.c
4374
if (e->is_reg)
kernel/bpf/verifier.c
4375
bt_set_frame_reg(bt, e->frameno, e->regno);
kernel/bpf/verifier.c
4377
bt_set_frame_slot(bt, e->frameno, e->spi);
kernel/debug/kdb/kdb_main.c
238
char *e = *ep++;
kernel/debug/kdb/kdb_main.c
240
if (!e)
kernel/debug/kdb/kdb_main.c
243
if ((strncmp(match, e, matchlen) == 0)
kernel/debug/kdb/kdb_main.c
244
&& ((e[matchlen] == '\0')
kernel/debug/kdb/kdb_main.c
245
|| (e[matchlen] == '='))) {
kernel/debug/kdb/kdb_main.c
246
char *cp = strchr(e, '=');
kernel/events/core.c
1829
struct perf_event *e = __node_2_pe(a);
kernel/events/core.c
1830
return perf_event_groups_cmp(e->cpu, e->pmu_ctx->pmu, event_cgroup(e),
kernel/events/core.c
1831
e->group_index, __node_2_pe(b)) < 0;
kernel/extable.c
56
const struct exception_table_entry *e;
kernel/extable.c
58
e = search_kernel_exception_table(addr);
kernel/extable.c
59
if (!e)
kernel/extable.c
60
e = search_module_extables(addr);
kernel/extable.c
61
if (!e)
kernel/extable.c
62
e = search_bpf_extables(addr);
kernel/extable.c
63
return e;
kernel/locking/lockdep.c
1000
if (e == f)
kernel/locking/lockdep.c
1011
static bool in_any_class_list(struct list_head *e)
kernel/locking/lockdep.c
1018
if (in_list(e, &class->locks_after) ||
kernel/locking/lockdep.c
1019
in_list(e, &class->locks_before))
kernel/locking/lockdep.c
1027
struct lock_list *e;
kernel/locking/lockdep.c
1029
list_for_each_entry(e, h, entry) {
kernel/locking/lockdep.c
1030
if (e->links_to != c) {
kernel/locking/lockdep.c
1033
(unsigned long)(e - list_entries),
kernel/locking/lockdep.c
1034
e->links_to && e->links_to->name ?
kernel/locking/lockdep.c
1035
e->links_to->name : "(?)",
kernel/locking/lockdep.c
1036
e->class && e->class->name ? e->class->name :
kernel/locking/lockdep.c
1089
struct lock_list *e;
kernel/locking/lockdep.c
1127
e = list_entries + i;
kernel/locking/lockdep.c
1128
if (!in_any_class_list(&e->entry)) {
kernel/locking/lockdep.c
1130
(unsigned int)(e - list_entries),
kernel/locking/lockdep.c
1131
e->class->name ? : "(?)",
kernel/locking/lockdep.c
1132
e->links_to->name ? : "(?)");
kernel/locking/lockdep.c
1142
e = list_entries + i;
kernel/locking/lockdep.c
1143
if (in_any_class_list(&e->entry)) {
kernel/locking/lockdep.c
1145
(unsigned int)(e - list_entries),
kernel/locking/lockdep.c
1146
e->class && e->class->name ? e->class->name :
kernel/locking/lockdep.c
1148
e->links_to && e->links_to->name ?
kernel/locking/lockdep.c
1149
e->links_to->name : "(?)");
kernel/locking/lockdep.c
995
static bool in_list(struct list_head *e, struct list_head *h)
kernel/printk/printk.c
1114
struct prb_reserved_entry e;
kernel/printk/printk.c
1119
if (!prb_reserve(&e, rb, &dest_r))
kernel/printk/printk.c
1131
prb_final_commit(&e);
kernel/printk/printk.c
1133
return prb_record_text_space(&e);
kernel/printk/printk.c
2273
struct prb_reserved_entry e;
kernel/printk/printk.c
2328
if (prb_reserve_in_last(&e, prb, &r, caller_id, PRINTKRB_RECORD_MAX)) {
kernel/printk/printk.c
2338
prb_final_commit(&e);
kernel/printk/printk.c
2340
prb_commit(&e);
kernel/printk/printk.c
2354
if (!prb_reserve(&e, prb, &r)) {
kernel/printk/printk.c
2359
if (!prb_reserve(&e, prb, &r))
kernel/printk/printk.c
2379
prb_commit(&e);
kernel/printk/printk.c
2381
prb_final_commit(&e);
kernel/printk/printk.c
635
static void append_char(char **pp, char *e, char c)
kernel/printk/printk.c
637
if (*pp < e)
kernel/printk/printk.c
666
char *p = buf, *e = buf + size;
kernel/printk/printk.c
674
p += scnprintf(p, e - p, "\\x%02x", c);
kernel/printk/printk.c
676
append_char(&p, e, c);
kernel/printk/printk.c
678
append_char(&p, e, endc);
kernel/printk/printk_ringbuffer.c
1421
bool prb_reserve_in_last(struct prb_reserved_entry *e, struct printk_ringbuffer *rb,
kernel/printk/printk_ringbuffer.c
1430
local_irq_save(e->irqflags);
kernel/printk/printk_ringbuffer.c
1435
local_irq_restore(e->irqflags);
kernel/printk/printk_ringbuffer.c
1447
e->rb = rb;
kernel/printk/printk_ringbuffer.c
1448
e->id = id;
kernel/printk/printk_ringbuffer.c
1503
e->text_space = space_used(&rb->text_data_ring, &d->text_blk_lpos);
kernel/printk/printk_ringbuffer.c
1507
prb_commit(e);
kernel/printk/printk_ringbuffer.c
1654
bool prb_reserve(struct prb_reserved_entry *e, struct printk_ringbuffer *rb,
kernel/printk/printk_ringbuffer.c
1672
local_irq_save(e->irqflags);
kernel/printk/printk_ringbuffer.c
1677
local_irq_restore(e->irqflags);
kernel/printk/printk_ringbuffer.c
1696
e->rb = rb;
kernel/printk/printk_ringbuffer.c
1697
e->id = id;
kernel/printk/printk_ringbuffer.c
1727
prb_commit(e);
kernel/printk/printk_ringbuffer.c
1735
e->text_space = space_used(&rb->text_data_ring, &d->text_blk_lpos);
kernel/printk/printk_ringbuffer.c
1746
static void _prb_commit(struct prb_reserved_entry *e, unsigned long state_val)
kernel/printk/printk_ringbuffer.c
1748
struct prb_desc_ring *desc_ring = &e->rb->desc_ring;
kernel/printk/printk_ringbuffer.c
1749
struct prb_desc *d = to_desc(desc_ring, e->id);
kernel/printk/printk_ringbuffer.c
1750
unsigned long prev_state_val = DESC_SV(e->id, desc_reserved);
kernel/printk/printk_ringbuffer.c
1778
DESC_SV(e->id, state_val))) { /* LMM(_prb_commit:B) */
kernel/printk/printk_ringbuffer.c
1783
local_irq_restore(e->irqflags);
kernel/printk/printk_ringbuffer.c
1802
void prb_commit(struct prb_reserved_entry *e)
kernel/printk/printk_ringbuffer.c
1804
struct prb_desc_ring *desc_ring = &e->rb->desc_ring;
kernel/printk/printk_ringbuffer.c
1807
_prb_commit(e, desc_committed);
kernel/printk/printk_ringbuffer.c
1815
if (head_id != e->id)
kernel/printk/printk_ringbuffer.c
1816
desc_make_final(e->rb, e->id);
kernel/printk/printk_ringbuffer.c
1835
void prb_final_commit(struct prb_reserved_entry *e)
kernel/printk/printk_ringbuffer.c
1837
_prb_commit(e, desc_finalized);
kernel/printk/printk_ringbuffer.c
1839
desc_update_last_finalized(e->rb);
kernel/printk/printk_ringbuffer.c
2412
unsigned int prb_record_text_space(struct prb_reserved_entry *e)
kernel/printk/printk_ringbuffer.c
2414
return e->text_space;
kernel/printk/printk_ringbuffer.h
327
bool prb_reserve(struct prb_reserved_entry *e, struct printk_ringbuffer *rb,
kernel/printk/printk_ringbuffer.h
329
bool prb_reserve_in_last(struct prb_reserved_entry *e, struct printk_ringbuffer *rb,
kernel/printk/printk_ringbuffer.h
331
void prb_commit(struct prb_reserved_entry *e);
kernel/printk/printk_ringbuffer.h
332
void prb_final_commit(struct prb_reserved_entry *e);
kernel/printk/printk_ringbuffer.h
338
unsigned int prb_record_text_space(struct prb_reserved_entry *e);
kernel/printk/printk_ringbuffer_kunit_test.c
111
struct prb_reserved_entry e;
kernel/printk/printk_ringbuffer_kunit_test.c
146
if (prb_reserve(&e, tr->test_data->ringbuffer, &r)) {
kernel/printk/printk_ringbuffer_kunit_test.c
154
prb_commit(&e);
kernel/trace/ring_buffer.c
1853
int e;
kernel/trace/ring_buffer.c
1860
for (e = 0; e < tail; e += len) {
kernel/trace/ring_buffer.c
1862
event = (struct ring_buffer_event *)(dpage->data + e);
kernel/trace/ring_buffer.c
1864
if (len <= 0 || len > tail - e)
kernel/trace/ring_buffer.c
4333
int e;
kernel/trace/ring_buffer.c
4338
for (e = 0; e < tail; e += rb_event_length(event)) {
kernel/trace/ring_buffer.c
4340
event = (struct ring_buffer_event *)(bpage->data + e);
kernel/trace/ring_buffer.c
4348
e, ts, delta);
kernel/trace/ring_buffer.c
4355
e, ts, delta);
kernel/trace/ring_buffer.c
4361
e, ts, event->time_delta);
kernel/trace/ring_buffer.c
4367
e, ts, event->time_delta,
kernel/trace/ring_buffer.c
4375
pr_warn("expected end:0x%lx last event actually ended at:0x%x\n", tail, e);
kernel/trace/trace.c
3318
unsigned long t, e;
kernel/trace/trace.c
3325
get_total_entries_cpu(buf, &t, &e, cpu);
kernel/trace/trace.c
3327
*entries += e;
kernel/trace/trace.h
2373
#define FGRAPH_ENTRY_ARGS(e) \
kernel/trace/trace.h
2376
struct ftrace_graph_ent_entry *_e = e; \
kernel/trace/trace.h
2379
e->ent.type == TRACE_GRAPH_RETADDR_ENT) { \
kernel/trace/trace_events.c
304
const char *r, *e, *a;
kernel/trace/trace_events.c
306
e = fmt + len;
kernel/trace/trace_events.c
310
if (r && r < e) {
kernel/trace/trace_events.c
319
} else if (find_print_string(fmt, "__get_dynamic_array(", e)) {
kernel/trace/trace_events.c
321
} else if (find_print_string(fmt, "__get_rel_dynamic_array(", e)) {
kernel/trace/trace_events.c
323
} else if (find_print_string(fmt, "__get_dynamic_array_len(", e)) {
kernel/trace/trace_events.c
325
} else if (find_print_string(fmt, "__get_rel_dynamic_array_len(", e)) {
kernel/trace/trace_events.c
327
} else if (find_print_string(fmt, "__get_sockaddr(", e)) {
kernel/trace/trace_events.c
329
} else if (find_print_string(fmt, "__get_rel_sockaddr(", e)) {
kernel/trace/trace_events.c
339
const char *r, *e, *s;
kernel/trace/trace_events.c
341
e = fmt + len;
kernel/trace/trace_events.c
352
if (!r || r >= e)
kernel/trace/trace_events.c
365
} while (s < e);
kernel/trace/trace_events.c
373
if (r && r < e) {
kernel/trace/trace_events.c
375
if (r && r < e)
kernel/trace/trace_events.c
384
if (find_print_string(fmt, "\"", e))
kernel/trace/trace_events.c
433
int i, e;
kernel/trace/trace_events.c
561
e = i;
kernel/trace/trace_events.c
581
e - start_arg,
kernel/trace/trace_events_filter.c
2753
.e = ve, .f = vf, .g = vg, .h = vh }, \
kernel/trace/trace_events_filter_test.h
12
TP_PROTO(int a, int b, int c, int d, int e, int f, int g, int h),
kernel/trace/trace_events_filter_test.h
14
TP_ARGS(a, b, c, d, e, f, g, h),
kernel/trace/trace_events_filter_test.h
21
__field(int, e)
kernel/trace/trace_events_filter_test.h
32
__entry->e = e;
kernel/trace/trace_events_filter_test.h
40
__entry->e, __entry->f, __entry->g, __entry->h)
kernel/trace/trace_events_hist.c
3187
int e;
kernel/trace/trace_events_hist.c
3189
e = stack_trace_save((void *)stack_start,
kernel/trace/trace_events_hist.c
3192
if (e < HIST_STACKTRACE_DEPTH - 1)
kernel/trace/trace_events_hist.c
3193
((unsigned long *)stack_start)[e] = 0;
kernel/trace/trace_events_hist.c
3194
*((unsigned long *)str) = e;
kernel/trace/trace_events_hist.c
5223
int e;
kernel/trace/trace_events_hist.c
5225
e = stack_trace_save((void *)stack_start,
kernel/trace/trace_events_hist.c
5228
if (e < HIST_STACKTRACE_DEPTH - 1)
kernel/trace/trace_events_hist.c
5229
((unsigned long *)stack_start)[e] = 0;
kernel/trace/trace_events_hist.c
5230
*((unsigned long *)str) = e;
kernel/trace/trace_events_user.c
131
#define ENABLE_BITOPS(e) (&(e)->values)
kernel/trace/trace_events_user.c
133
#define ENABLE_BIT(e) ((int)((e)->values & ENABLE_VAL_BIT_MASK))
kernel/user_namespace.c
1064
struct uid_gid_extent *e;
kernel/user_namespace.c
1068
e = &new_map.extent[idx];
kernel/user_namespace.c
1070
e = &new_map.forward[idx];
kernel/user_namespace.c
1073
e->lower_first,
kernel/user_namespace.c
1074
e->count);
kernel/user_namespace.c
1082
e->lower_first = lower_first;
kernel/user_namespace.c
251
static int cmp_map_id(const void *k, const void *e)
kernel/user_namespace.c
255
const struct uid_gid_extent *el = e;
lib/bch.c
669
unsigned int a, b, c, d, e = 0, f, a2, b2, c2, e4;
lib/bch.c
689
e = a_pow(bch, l/2);
lib/bch.c
698
b = gf_mul(bch, a, e)^b;
lib/bch.c
719
roots[i] = a_ilog(bch, f^e);
lib/crypto/curve25519-fiat32.c
764
u8 e[32];
lib/crypto/curve25519-fiat32.c
766
memcpy(e, scalar, 32);
lib/crypto/curve25519-fiat32.c
767
curve25519_clamp_secret(e);
lib/crypto/curve25519-fiat32.c
814
unsigned b = 1 & (e[pos / 8] >> (pos & 7));
lib/crypto/curve25519-fiat32.c
863
memzero_explicit(&e, sizeof(e));
lib/crypto/curve25519-hacl64.c
769
u8 e[32] __aligned(32) = { 0 };
lib/crypto/curve25519-hacl64.c
771
memcpy(e, secret, 32);
lib/crypto/curve25519-hacl64.c
772
curve25519_clamp_secret(e);
lib/crypto/curve25519-hacl64.c
773
scalar = e;
lib/crypto/curve25519-hacl64.c
783
memzero_explicit(e, sizeof(e));
lib/crypto/mldsa.c
211
s64 e = c - (s64)d * Q;
lib/crypto/mldsa.c
214
return e >> 32;
lib/crypto/mpi/mpi-pow.c
154
mpi_limb_t e;
lib/crypto/mpi/mpi-pow.c
164
e = ep[i];
lib/crypto/mpi/mpi-pow.c
165
c = count_leading_zeros(e);
lib/crypto/mpi/mpi-pow.c
166
e = (e << c) << 1; /* shift the exp bits to the left, lose msb */
lib/crypto/mpi/mpi-pow.c
214
if ((mpi_limb_signed_t) e < 0) {
lib/crypto/mpi/mpi-pow.c
240
e <<= 1;
lib/crypto/mpi/mpi-pow.c
248
e = ep[i];
lib/crypto/powerpc/curve25519.h
138
fe51 a, b, c, d, e;
lib/crypto/powerpc/curve25519.h
157
fsub(e, aa, bb); // E = AA - BB
lib/crypto/powerpc/curve25519.h
162
fmul121666(z3, e); // 121666 * E
lib/crypto/powerpc/curve25519.h
167
fmul(z2, e, b); // z2 = e * (BB + (DA + CB)^2)
lib/crypto/sha256.c
118
e = state->h[4];
lib/crypto/sha256.c
125
SHA256_ROUND(i + 0, a, b, c, d, e, f, g, h);
lib/crypto/sha256.c
126
SHA256_ROUND(i + 1, h, a, b, c, d, e, f, g);
lib/crypto/sha256.c
127
SHA256_ROUND(i + 2, g, h, a, b, c, d, e, f);
lib/crypto/sha256.c
128
SHA256_ROUND(i + 3, f, g, h, a, b, c, d, e);
lib/crypto/sha256.c
129
SHA256_ROUND(i + 4, e, f, g, h, a, b, c, d);
lib/crypto/sha256.c
130
SHA256_ROUND(i + 5, d, e, f, g, h, a, b, c);
lib/crypto/sha256.c
131
SHA256_ROUND(i + 6, c, d, e, f, g, h, a, b);
lib/crypto/sha256.c
132
SHA256_ROUND(i + 7, b, c, d, e, f, g, h, a);
lib/crypto/sha256.c
139
state->h[4] += e;
lib/crypto/sha256.c
74
#define SHA256_ROUND(i, a, b, c, d, e, f, g, h) \
lib/crypto/sha256.c
77
t1 = h + e1(e) + Ch(e, f, g) + sha256_K[i] + W[i]; \
lib/crypto/sha256.c
86
u32 a, b, c, d, e, f, g, h;
lib/crypto/sha512.c
100
t1 = g + e1(d) + Ch(d, e, f) + sha512_K[i+1] + W[(i & 15) + 1];
lib/crypto/sha512.c
102
t1 = f + e1(c) + Ch(c, d, e) + sha512_K[i+2] + W[(i & 15) + 2];
lib/crypto/sha512.c
104
t1 = e + e1(b) + Ch(b, c, d) + sha512_K[i+3] + W[(i & 15) + 3];
lib/crypto/sha512.c
105
t2 = e0(f) + Maj(f, g, h); a += t1; e = t1 + t2;
lib/crypto/sha512.c
107
t2 = e0(e) + Maj(e, f, g); h += t1; d = t1 + t2;
lib/crypto/sha512.c
109
t2 = e0(d) + Maj(d, e, f); g += t1; c = t1 + t2;
lib/crypto/sha512.c
111
t2 = e0(c) + Maj(c, d, e); f += t1; b = t1 + t2;
lib/crypto/sha512.c
113
t2 = e0(b) + Maj(b, c, d); e += t1; a = t1 + t2;
lib/crypto/sha512.c
120
state->h[4] += e;
lib/crypto/sha512.c
80
u64 e = state->h[4];
lib/crypto/sha512.c
98
t1 = h + e1(e) + Ch(e, f, g) + sha512_K[i] + W[(i & 15)];
lib/crypto/sm3.c
100
R1(a, b, c, d, e, f, g, h, K[12], W1(12), W2(16));
lib/crypto/sm3.c
101
R1(d, a, b, c, h, e, f, g, K[13], W1(13), W2(17));
lib/crypto/sm3.c
102
R1(c, d, a, b, g, h, e, f, K[14], W1(14), W2(18));
lib/crypto/sm3.c
103
R1(b, c, d, a, f, g, h, e, K[15], W1(15), W2(19));
lib/crypto/sm3.c
105
R2(a, b, c, d, e, f, g, h, K[16], W1(16), W2(20));
lib/crypto/sm3.c
106
R2(d, a, b, c, h, e, f, g, K[17], W1(17), W2(21));
lib/crypto/sm3.c
107
R2(c, d, a, b, g, h, e, f, K[18], W1(18), W2(22));
lib/crypto/sm3.c
108
R2(b, c, d, a, f, g, h, e, K[19], W1(19), W2(23));
lib/crypto/sm3.c
109
R2(a, b, c, d, e, f, g, h, K[20], W1(20), W2(24));
lib/crypto/sm3.c
110
R2(d, a, b, c, h, e, f, g, K[21], W1(21), W2(25));
lib/crypto/sm3.c
111
R2(c, d, a, b, g, h, e, f, K[22], W1(22), W2(26));
lib/crypto/sm3.c
112
R2(b, c, d, a, f, g, h, e, K[23], W1(23), W2(27));
lib/crypto/sm3.c
113
R2(a, b, c, d, e, f, g, h, K[24], W1(24), W2(28));
lib/crypto/sm3.c
114
R2(d, a, b, c, h, e, f, g, K[25], W1(25), W2(29));
lib/crypto/sm3.c
115
R2(c, d, a, b, g, h, e, f, K[26], W1(26), W2(30));
lib/crypto/sm3.c
116
R2(b, c, d, a, f, g, h, e, K[27], W1(27), W2(31));
lib/crypto/sm3.c
117
R2(a, b, c, d, e, f, g, h, K[28], W1(28), W2(32));
lib/crypto/sm3.c
118
R2(d, a, b, c, h, e, f, g, K[29], W1(29), W2(33));
lib/crypto/sm3.c
119
R2(c, d, a, b, g, h, e, f, K[30], W1(30), W2(34));
lib/crypto/sm3.c
120
R2(b, c, d, a, f, g, h, e, K[31], W1(31), W2(35));
lib/crypto/sm3.c
122
R2(a, b, c, d, e, f, g, h, K[32], W1(32), W2(36));
lib/crypto/sm3.c
123
R2(d, a, b, c, h, e, f, g, K[33], W1(33), W2(37));
lib/crypto/sm3.c
124
R2(c, d, a, b, g, h, e, f, K[34], W1(34), W2(38));
lib/crypto/sm3.c
125
R2(b, c, d, a, f, g, h, e, K[35], W1(35), W2(39));
lib/crypto/sm3.c
126
R2(a, b, c, d, e, f, g, h, K[36], W1(36), W2(40));
lib/crypto/sm3.c
127
R2(d, a, b, c, h, e, f, g, K[37], W1(37), W2(41));
lib/crypto/sm3.c
128
R2(c, d, a, b, g, h, e, f, K[38], W1(38), W2(42));
lib/crypto/sm3.c
129
R2(b, c, d, a, f, g, h, e, K[39], W1(39), W2(43));
lib/crypto/sm3.c
130
R2(a, b, c, d, e, f, g, h, K[40], W1(40), W2(44));
lib/crypto/sm3.c
131
R2(d, a, b, c, h, e, f, g, K[41], W1(41), W2(45));
lib/crypto/sm3.c
132
R2(c, d, a, b, g, h, e, f, K[42], W1(42), W2(46));
lib/crypto/sm3.c
133
R2(b, c, d, a, f, g, h, e, K[43], W1(43), W2(47));
lib/crypto/sm3.c
134
R2(a, b, c, d, e, f, g, h, K[44], W1(44), W2(48));
lib/crypto/sm3.c
135
R2(d, a, b, c, h, e, f, g, K[45], W1(45), W2(49));
lib/crypto/sm3.c
136
R2(c, d, a, b, g, h, e, f, K[46], W1(46), W2(50));
lib/crypto/sm3.c
137
R2(b, c, d, a, f, g, h, e, K[47], W1(47), W2(51));
lib/crypto/sm3.c
139
R2(a, b, c, d, e, f, g, h, K[48], W1(48), W2(52));
lib/crypto/sm3.c
140
R2(d, a, b, c, h, e, f, g, K[49], W1(49), W2(53));
lib/crypto/sm3.c
141
R2(c, d, a, b, g, h, e, f, K[50], W1(50), W2(54));
lib/crypto/sm3.c
142
R2(b, c, d, a, f, g, h, e, K[51], W1(51), W2(55));
lib/crypto/sm3.c
143
R2(a, b, c, d, e, f, g, h, K[52], W1(52), W2(56));
lib/crypto/sm3.c
144
R2(d, a, b, c, h, e, f, g, K[53], W1(53), W2(57));
lib/crypto/sm3.c
145
R2(c, d, a, b, g, h, e, f, K[54], W1(54), W2(58));
lib/crypto/sm3.c
146
R2(b, c, d, a, f, g, h, e, K[55], W1(55), W2(59));
lib/crypto/sm3.c
147
R2(a, b, c, d, e, f, g, h, K[56], W1(56), W2(60));
lib/crypto/sm3.c
148
R2(d, a, b, c, h, e, f, g, K[57], W1(57), W2(61));
lib/crypto/sm3.c
149
R2(c, d, a, b, g, h, e, f, K[58], W1(58), W2(62));
lib/crypto/sm3.c
150
R2(b, c, d, a, f, g, h, e, K[59], W1(59), W2(63));
lib/crypto/sm3.c
151
R2(a, b, c, d, e, f, g, h, K[60], W1(60), W2(64));
lib/crypto/sm3.c
152
R2(d, a, b, c, h, e, f, g, K[61], W1(61), W2(65));
lib/crypto/sm3.c
153
R2(c, d, a, b, g, h, e, f, K[62], W1(62), W2(66));
lib/crypto/sm3.c
154
R2(b, c, d, a, f, g, h, e, K[63], W1(63), W2(67));
lib/crypto/sm3.c
160
sctx->state[4] ^= e;
lib/crypto/sm3.c
41
#define R(i, a, b, c, d, e, f, g, h, t, w1, w2) \
lib/crypto/sm3.c
43
ss1 = rol32((rol32((a), 12) + (e) + (t)), 7); \
lib/crypto/sm3.c
46
h += GG ## i(e, f, g) + ss1 + (w1); \
lib/crypto/sm3.c
52
#define R1(a, b, c, d, e, f, g, h, t, w1, w2) \
lib/crypto/sm3.c
53
R(1, a, b, c, d, e, f, g, h, t, w1, w2)
lib/crypto/sm3.c
54
#define R2(a, b, c, d, e, f, g, h, t, w1, w2) \
lib/crypto/sm3.c
55
R(2, a, b, c, d, e, f, g, h, t, w1, w2)
lib/crypto/sm3.c
77
u32 a, b, c, d, e, f, g, h, ss1, ss2;
lib/crypto/sm3.c
83
e = sctx->state[4];
lib/crypto/sm3.c
88
R1(a, b, c, d, e, f, g, h, K[0], I(0), I(4));
lib/crypto/sm3.c
89
R1(d, a, b, c, h, e, f, g, K[1], I(1), I(5));
lib/crypto/sm3.c
90
R1(c, d, a, b, g, h, e, f, K[2], I(2), I(6));
lib/crypto/sm3.c
91
R1(b, c, d, a, f, g, h, e, K[3], I(3), I(7));
lib/crypto/sm3.c
92
R1(a, b, c, d, e, f, g, h, K[4], W1(4), I(8));
lib/crypto/sm3.c
93
R1(d, a, b, c, h, e, f, g, K[5], W1(5), I(9));
lib/crypto/sm3.c
94
R1(c, d, a, b, g, h, e, f, K[6], W1(6), I(10));
lib/crypto/sm3.c
95
R1(b, c, d, a, f, g, h, e, K[7], W1(7), I(11));
lib/crypto/sm3.c
96
R1(a, b, c, d, e, f, g, h, K[8], W1(8), I(12));
lib/crypto/sm3.c
97
R1(d, a, b, c, h, e, f, g, K[9], W1(9), I(13));
lib/crypto/sm3.c
98
R1(c, d, a, b, g, h, e, f, K[10], W1(10), I(14));
lib/crypto/sm3.c
99
R1(b, c, d, a, f, g, h, e, K[11], W1(11), I(15));
lib/decompress_unxz.c
202
uint8_t *e = b + size;
lib/decompress_unxz.c
204
while (b != e)
lib/inflate.c
1035
int *e /* last block flag */
lib/inflate.c
1052
*e = (int)b & 1;
lib/inflate.c
1088
int e; /* last block flag */
lib/inflate.c
1105
r = inflate_block(&e);
lib/inflate.c
1110
} while (!e);
lib/inflate.c
1152
unsigned long e; /* polynomial exclusive-or pattern */
lib/inflate.c
1160
e = 0;
lib/inflate.c
1162
e |= 1L << (31 - p[i]);
lib/inflate.c
1171
c = c & 1 ? (c >> 1) ^ e : c >> 1;
lib/inflate.c
1173
c ^= e;
lib/inflate.c
140
uch e; /* number of extra bits or operation */
lib/inflate.c
327
const ush *e, /* list of extra bits for non-simple codes */
lib/inflate.c
508
r.e = (uch)(16 + j); /* bits in this table */
lib/inflate.c
520
r.e = 99; /* out of values--invalid code */
lib/inflate.c
523
r.e = (uch)(*p < 256 ? 16 : 15); /* 256 is end-of-block code */
lib/inflate.c
529
r.e = (uch)e[*p - s]; /* non-simple--look up in lists */
lib/inflate.c
598
register unsigned e; /* table entry flag/number of extra bits */
lib/inflate.c
618
if ((e = (t = tl + ((unsigned)b & ml))->e) > 16)
lib/inflate.c
620
if (e == 99)
lib/inflate.c
623
e -= 16;
lib/inflate.c
624
NEEDBITS(e)
lib/inflate.c
625
} while ((e = (t = t->v.t + ((unsigned)b & mask_bits[e]))->e) > 16);
lib/inflate.c
627
if (e == 16) /* then it's a literal */
lib/inflate.c
640
if (e == 15)
lib/inflate.c
644
NEEDBITS(e)
lib/inflate.c
645
n = t->v.n + ((unsigned)b & mask_bits[e]);
lib/inflate.c
646
DUMPBITS(e);
lib/inflate.c
650
if ((e = (t = td + ((unsigned)b & md))->e) > 16)
lib/inflate.c
652
if (e == 99)
lib/inflate.c
655
e -= 16;
lib/inflate.c
656
NEEDBITS(e)
lib/inflate.c
657
} while ((e = (t = t->v.t + ((unsigned)b & mask_bits[e]))->e) > 16);
lib/inflate.c
659
NEEDBITS(e)
lib/inflate.c
660
d = w - t->v.n - ((unsigned)b & mask_bits[e]);
lib/inflate.c
661
DUMPBITS(e)
lib/inflate.c
666
n -= (e = (e = WSIZE - ((d &= WSIZE-1) > w ? d : w)) > n ? n : e);
lib/inflate.c
668
if (w - d >= e) /* (this test assumes unsigned comparison) */
lib/inflate.c
670
memcpy(slide + w, slide + d, e);
lib/inflate.c
671
w += e;
lib/inflate.c
672
d += e;
lib/inflate.c
679
} while (--e);
lib/locking-selftest.c
1495
#define dotest_rt(fn, e, m) dotest((fn), (e), (m))
lib/locking-selftest.c
1497
#define dotest_rt(fn, e, m)
lib/lru_cache.c
128
e = p + e_off;
lib/lru_cache.c
129
e->lc_index = i;
lib/lru_cache.c
130
e->lc_number = LC_FREE;
lib/lru_cache.c
131
e->lc_new_number = LC_FREE;
lib/lru_cache.c
132
list_add(&e->list, &lc->free);
lib/lru_cache.c
133
element[i] = e;
lib/lru_cache.c
202
struct lc_element *e = lc->lc_element[i];
lib/lru_cache.c
203
void *p = e;
lib/lru_cache.c
207
e->lc_index = i;
lib/lru_cache.c
208
e->lc_number = LC_FREE;
lib/lru_cache.c
209
e->lc_new_number = LC_FREE;
lib/lru_cache.c
210
list_add(&e->list, &lc->free);
lib/lru_cache.c
242
struct lc_element *e;
lib/lru_cache.c
246
hlist_for_each_entry(e, lc_hash_slot(lc, enr), collision) {
lib/lru_cache.c
250
if (e->lc_new_number != enr)
lib/lru_cache.c
252
if (e->lc_new_number == e->lc_number || include_changing)
lib/lru_cache.c
253
return e;
lib/lru_cache.c
287
struct lc_element *e = __lc_find(lc, enr, 1);
lib/lru_cache.c
288
return e && e->refcnt;
lib/lru_cache.c
299
void lc_del(struct lru_cache *lc, struct lc_element *e)
lib/lru_cache.c
302
PARANOIA_LC_ELEMENT(lc, e);
lib/lru_cache.c
303
BUG_ON(e->refcnt);
lib/lru_cache.c
305
e->lc_number = e->lc_new_number = LC_FREE;
lib/lru_cache.c
306
hlist_del_init(&e->collision);
lib/lru_cache.c
307
list_move(&e->list, &lc->free);
lib/lru_cache.c
314
struct lc_element *e;
lib/lru_cache.c
323
e = list_entry(n, struct lc_element, list);
lib/lru_cache.c
324
PARANOIA_LC_ELEMENT(lc, e);
lib/lru_cache.c
326
e->lc_new_number = new_number;
lib/lru_cache.c
327
if (!hlist_unhashed(&e->collision))
lib/lru_cache.c
328
__hlist_del(&e->collision);
lib/lru_cache.c
329
hlist_add_head(&e->collision, lc_hash_slot(lc, new_number));
lib/lru_cache.c
330
list_move(&e->list, &lc->to_be_changed);
lib/lru_cache.c
332
return e;
lib/lru_cache.c
353
struct lc_element *e;
lib/lru_cache.c
361
e = __lc_find(lc, enr, 1);
lib/lru_cache.c
366
if (e) {
lib/lru_cache.c
367
if (e->lc_new_number != e->lc_number) {
lib/lru_cache.c
376
++e->refcnt;
lib/lru_cache.c
378
RETURN(e);
lib/lru_cache.c
382
if (e->refcnt++ == 0)
lib/lru_cache.c
384
list_move(&e->list, &lc->in_use); /* Not evictable... */
lib/lru_cache.c
385
RETURN(e);
lib/lru_cache.c
39
#define PARANOIA_LC_ELEMENT(lc, e) do { \
lib/lru_cache.c
41
struct lc_element *e_ = (e); \
lib/lru_cache.c
419
e = lc_prepare_for_change(lc, enr);
lib/lru_cache.c
420
BUG_ON(!e);
lib/lru_cache.c
423
BUG_ON(++e->refcnt != 1);
lib/lru_cache.c
427
RETURN(e);
lib/lru_cache.c
526
struct lc_element *e, *tmp;
lib/lru_cache.c
529
list_for_each_entry_safe(e, tmp, &lc->to_be_changed, list) {
lib/lru_cache.c
532
e->lc_number = e->lc_new_number;
lib/lru_cache.c
533
list_move(&e->list, &lc->in_use);
lib/lru_cache.c
549
unsigned int lc_put(struct lru_cache *lc, struct lc_element *e)
lib/lru_cache.c
552
PARANOIA_LC_ELEMENT(lc, e);
lib/lru_cache.c
553
BUG_ON(e->refcnt == 0);
lib/lru_cache.c
554
BUG_ON(e->lc_number != e->lc_new_number);
lib/lru_cache.c
555
if (--e->refcnt == 0) {
lib/lru_cache.c
557
list_move(&e->list, &lc->lru);
lib/lru_cache.c
561
RETURN(e->refcnt);
lib/lru_cache.c
590
struct lc_element *e;
lib/lru_cache.c
595
e = lc_element_by_index(lc, i);
lib/lru_cache.c
596
if (e->lc_number != e->lc_new_number)
lib/lru_cache.c
598
i, e->lc_number, e->lc_new_number, e->refcnt);
lib/lru_cache.c
601
i, e->lc_number, "-\"-", e->refcnt);
lib/lru_cache.c
603
detail(seq, e);
lib/lru_cache.c
84
struct lc_element *e;
lib/lz4/lz4defs.h
177
BYTE *const e = (BYTE *)dstEnd;
lib/lz4/lz4defs.h
183
} while (d < e);
lib/reed_solomon/test_rslib.c
442
static int run_exercise(struct etab *e)
lib/reed_solomon/test_rslib.c
444
int nn = (1 << e->symsize) - 1;
lib/reed_solomon/test_rslib.c
445
int kk = nn - e->nroots;
lib/reed_solomon/test_rslib.c
453
rsc = init_rs(e->symsize, e->genpoly, e->fcs, e->prim, e->nroots);
lib/reed_solomon/test_rslib.c
475
retval |= exercise_rs(rsc, ws, len, e->ntrials);
lib/reed_solomon/test_rslib.c
477
retval |= exercise_rs_bc(rsc, ws, len, e->ntrials);
lib/test_bitmap.c
1001
for_each_clear_bitrange_from(s, e, orig, 500)
lib/test_bitmap.c
1002
bitmap_clear(copy, s, e - s);
lib/test_bitmap.c
910
unsigned int s, e;
lib/test_bitmap.c
922
for_each_set_bitrange(s, e, orig, 500)
lib/test_bitmap.c
923
bitmap_set(copy, s, e-s);
lib/test_bitmap.c
932
unsigned int s, e;
lib/test_bitmap.c
944
for_each_clear_bitrange(s, e, orig, 500)
lib/test_bitmap.c
945
bitmap_clear(copy, s, e-s);
lib/test_bitmap.c
954
unsigned int wr, s, e;
lib/test_bitmap.c
971
for_each_set_bitrange_from(s, e, orig, 500)
lib/test_bitmap.c
972
bitmap_set(copy, s, e - s);
lib/test_bitmap.c
984
unsigned int wr, s, e;
lib/test_bpf.c
15291
long e = 4;
lib/test_bpf.c
15298
a, b, c, d, e, f, g, h, (int)arg);
lib/test_fpu_impl.c
15
volatile double a, b, c, d, e, f, g;
lib/test_fpu_impl.c
25
e = a + b / 2;
lib/test_fpu_impl.c
33
if (d > a && e > a && g > a)
lib/test_hexdump.c
167
int ae, he, e, f, r;
lib/test_hexdump.c
184
e = ae;
lib/test_hexdump.c
186
e = he;
lib/test_hexdump.c
188
f = min_t(int, e + 1, buflen);
lib/test_hexdump.c
195
a = r == e && !memcmp(test, buf, TEST_HEXDUMP_BUF_SIZE);
lib/test_hexdump.c
203
pr_err("Expect: %d '%s'\n", e, test);
lib/tests/cmdline_kunit.c
108
unsigned int n, const int *e)
lib/tests/cmdline_kunit.c
116
KUNIT_EXPECT_EQ_MSG(test, r[0], e[0], "in test %u (parsed) expected %d numbers, got %d",
lib/tests/cmdline_kunit.c
117
n, e[0], r[0]);
lib/tests/cmdline_kunit.c
119
KUNIT_EXPECT_EQ_MSG(test, r[i], e[i], "in test %u at %u", n, i);
lib/tests/cmdline_kunit.c
123
KUNIT_EXPECT_EQ_MSG(test, r[0], e[0], "in test %u (validated) expected %d numbers, got %d",
lib/tests/cmdline_kunit.c
124
n, e[0], r[0]);
lib/tests/cmdline_kunit.c
136
const int *e = cmdline_test_range_values[i];
lib/tests/cmdline_kunit.c
138
cmdline_do_one_range_test(test, str, i, e);
lib/tests/randstruct_kunit.c
236
.e = 11, \
lib/tests/randstruct_kunit.c
25
macro(e, args) \
lib/tests/randstruct_kunit.c
73
char e; \
lib/zstd/common/error_private.h
50
#define CHECK_V_F(e, f) \
lib/zstd/common/error_private.h
51
size_t const e = f; \
lib/zstd/common/error_private.h
53
if (ERR_isError(e)) \
lib/zstd/common/error_private.h
54
return e; \
mm/kmsan/kmsan_test.c
151
#define EXPECTATION_ETYPE_FN(e, reason, fn) \
mm/kmsan/kmsan_test.c
152
struct expect_report e = { \
mm/kmsan/kmsan_test.c
157
#define EXPECTATION_NO_REPORT(e) EXPECTATION_ETYPE_FN(e, NULL, NULL)
mm/kmsan/kmsan_test.c
158
#define EXPECTATION_UNINIT_VALUE_FN(e, fn) \
mm/kmsan/kmsan_test.c
159
EXPECTATION_ETYPE_FN(e, "uninit-value", fn)
mm/kmsan/kmsan_test.c
160
#define EXPECTATION_UNINIT_VALUE(e) EXPECTATION_UNINIT_VALUE_FN(e, __func__)
mm/kmsan/kmsan_test.c
161
#define EXPECTATION_USE_AFTER_FREE(e) \
mm/kmsan/kmsan_test.c
162
EXPECTATION_ETYPE_FN(e, "use-after-free", __func__)
mm/vmalloc.c
2976
unsigned long s, e;
mm/vmalloc.c
2979
e = va_start + (vb->dirty_max << PAGE_SHIFT);
mm/vmalloc.c
2982
end = max(e, end);
net/atm/clip.c
820
static struct clip_vcc *clip_seq_next_vcc(struct atmarp_entry *e,
net/atm/clip.c
824
curr = e->vccs;
net/atm/clip.c
838
struct atmarp_entry *e, loff_t * pos)
net/atm/clip.c
842
vcc = clip_seq_next_vcc(e, vcc);
net/atm/clip.c
845
vcc = clip_seq_next_vcc(e, vcc);
net/atm/lec.c
856
struct hlist_node *e = state->node;
net/atm/lec.c
858
if (!e)
net/atm/lec.c
859
e = tbl->first;
net/atm/lec.c
860
if (e == SEQ_START_TOKEN) {
net/atm/lec.c
861
e = tbl->first;
net/atm/lec.c
865
for (; e; e = e->next) {
net/atm/lec.c
869
state->node = e;
net/bluetooth/hci_core.c
170
struct inquiry_entry *e;
net/bluetooth/hci_core.c
174
list_for_each_entry(e, &cache->all, all) {
net/bluetooth/hci_core.c
175
if (!bacmp(&e->data.bdaddr, bdaddr))
net/bluetooth/hci_core.c
176
return e;
net/bluetooth/hci_core.c
186
struct inquiry_entry *e;
net/bluetooth/hci_core.c
190
list_for_each_entry(e, &cache->unknown, list) {
net/bluetooth/hci_core.c
191
if (!bacmp(&e->data.bdaddr, bdaddr))
net/bluetooth/hci_core.c
192
return e;
net/bluetooth/hci_core.c
203
struct inquiry_entry *e;
net/bluetooth/hci_core.c
207
list_for_each_entry(e, &cache->resolve, list) {
net/bluetooth/hci_core.c
208
if (!bacmp(bdaddr, BDADDR_ANY) && e->name_state == state)
net/bluetooth/hci_core.c
209
return e;
net/bluetooth/hci_core.c
210
if (!bacmp(&e->data.bdaddr, bdaddr))
net/bluetooth/hci_core.c
211
return e;
net/bluetooth/hci_core.c
302
struct inquiry_entry *e;
net/bluetooth/hci_core.c
305
list_for_each_entry(e, &cache->all, all) {
net/bluetooth/hci_core.c
306
struct inquiry_data *data = &e->data;
net/bluetooth/hci_debugfs.c
364
struct inquiry_entry *e;
net/bluetooth/hci_debugfs.c
368
list_for_each_entry(e, &cache->all, all) {
net/bluetooth/hci_debugfs.c
369
struct inquiry_data *data = &e->data;
net/bluetooth/hci_debugfs.c
376
data->rssi, data->ssp_mode, e->timestamp);
net/bluetooth/hci_event.c
2397
struct inquiry_entry *e)
net/bluetooth/hci_event.c
2403
bacpy(&cp.bdaddr, &e->data.bdaddr);
net/bluetooth/hci_event.c
2404
cp.pscan_rep_mode = e->data.pscan_rep_mode;
net/bluetooth/hci_event.c
2405
cp.pscan_mode = e->data.pscan_mode;
net/bluetooth/hci_event.c
2406
cp.clock_offset = e->data.clock_offset;
net/bluetooth/hci_event.c
2414
struct inquiry_entry *e;
net/bluetooth/hci_event.c
2425
e = hci_inquiry_cache_lookup_resolve(hdev, BDADDR_ANY, NAME_NEEDED);
net/bluetooth/hci_event.c
2426
if (!e)
net/bluetooth/hci_event.c
2429
if (hci_resolve_name(hdev, e) == 0) {
net/bluetooth/hci_event.c
2430
e->name_state = NAME_PENDING;
net/bluetooth/hci_event.c
2441
struct inquiry_entry *e;
net/bluetooth/hci_event.c
2460
e = hci_inquiry_cache_lookup_resolve(hdev, bdaddr, NAME_PENDING);
net/bluetooth/hci_event.c
2465
if (!e)
net/bluetooth/hci_event.c
2468
list_del(&e->list);
net/bluetooth/hci_event.c
2470
e->name_state = name ? NAME_KNOWN : NAME_NOT_KNOWN;
net/bluetooth/hci_event.c
2471
mgmt_remote_name(hdev, bdaddr, ACL_LINK, 0x00, e->data.rssi,
net/bluetooth/hci_event.c
2982
struct inquiry_entry *e;
net/bluetooth/hci_event.c
3014
e = hci_inquiry_cache_lookup_resolve(hdev, BDADDR_ANY, NAME_NEEDED);
net/bluetooth/hci_event.c
3015
if (e && hci_resolve_name(hdev, e) == 0) {
net/bluetooth/hci_event.c
3016
e->name_state = NAME_PENDING;
net/bluetooth/hci_sync.c
5515
struct inquiry_entry *e;
net/bluetooth/hci_sync.c
5551
e = hci_inquiry_cache_lookup_resolve(hdev, BDADDR_ANY,
net/bluetooth/hci_sync.c
5553
if (!e)
net/bluetooth/hci_sync.c
5559
hci_remote_name_cancel_sync(hdev, &e->data.bdaddr);
net/bluetooth/mgmt.c
6270
struct inquiry_entry *e;
net/bluetooth/mgmt.c
6284
e = hci_inquiry_cache_lookup_unknown(hdev, &cp->addr.bdaddr);
net/bluetooth/mgmt.c
6285
if (!e) {
net/bluetooth/mgmt.c
6293
e->name_state = NAME_KNOWN;
net/bluetooth/mgmt.c
6294
list_del(&e->list);
net/bluetooth/mgmt.c
6296
e->name_state = NAME_NEEDED;
net/bluetooth/mgmt.c
6297
hci_inquiry_cache_update_resolve(hdev, e);
net/bluetooth/smp.c
1474
local_addr, remote_addr, check.e);
net/bluetooth/smp.c
2872
u8 io_cap[3], r[16], e[16];
net/bluetooth/smp.c
2903
io_cap, remote_addr, local_addr, e);
net/bluetooth/smp.c
2907
if (crypto_memneq(check->e, e, 16))
net/bluetooth/smp.h
117
__u8 e[16];
net/bpf/test_run.c
531
noinline int bpf_fentry_test5(u64 a, void *b, short c, int d, u64 e)
net/bpf/test_run.c
533
return a + (long)b + c + d + e;
net/bpf/test_run.c
536
noinline int bpf_fentry_test6(u64 a, void *b, short c, int d, void *e, u64 f)
net/bpf/test_run.c
538
return a + (long)b + c + d + (long)e + f;
net/bpf/test_run.c
577
void *e, char f, int g)
net/bpf/test_run.c
580
return a + *b + c + d + (long)e + f + g;
net/bridge/br_input.c
270
struct nf_hook_entries *e = NULL;
net/bridge/br_input.c
282
e = rcu_dereference(net->nf.hooks_bridge[NF_BR_PRE_ROUTING]);
net/bridge/br_input.c
283
if (!e)
net/bridge/br_input.c
290
for (i = 0; i < e->num_hook_entries; i++) {
net/bridge/br_input.c
291
verdict = nf_hook_entry_hookfn(&e->hooks[i], skb, &state);
net/bridge/br_mdb.c
135
static void __mdb_entry_fill_flags(struct br_mdb_entry *e, unsigned char flags)
net/bridge/br_mdb.c
137
e->state = flags & MDB_PG_FLAGS_PERMANENT;
net/bridge/br_mdb.c
138
e->flags = 0;
net/bridge/br_mdb.c
140
e->flags |= MDB_FLAGS_OFFLOAD;
net/bridge/br_mdb.c
142
e->flags |= MDB_FLAGS_FAST_LEAVE;
net/bridge/br_mdb.c
144
e->flags |= MDB_FLAGS_STAR_EXCL;
net/bridge/br_mdb.c
146
e->flags |= MDB_FLAGS_BLOCKED;
net/bridge/br_mdb.c
148
e->flags |= MDB_FLAGS_OFFLOAD_FAILED;
net/bridge/br_mdb.c
239
struct br_mdb_entry e;
net/bridge/br_mdb.c
243
memset(&e, 0, sizeof(e));
net/bridge/br_mdb.c
253
__mdb_entry_fill_flags(&e, flags);
net/bridge/br_mdb.c
254
e.ifindex = ifindex;
net/bridge/br_mdb.c
255
e.vid = mp->addr.vid;
net/bridge/br_mdb.c
257
e.addr.u.ip4 = mp->addr.dst.ip4;
net/bridge/br_mdb.c
260
e.addr.u.ip6 = mp->addr.dst.ip6;
net/bridge/br_mdb.c
263
ether_addr_copy(e.addr.u.mac_addr, mp->addr.dst.mac_addr);
net/bridge/br_mdb.c
264
e.state = MDB_PERMANENT;
net/bridge/br_mdb.c
266
e.addr.proto = mp->addr.proto;
net/bridge/br_mdb.c
272
if (nla_put_nohdr(skb, sizeof(e), &e) ||
net/bridge/br_mdb.c
300
ether_addr_copy(e.addr.u.mac_addr, mp->addr.dst.mac_addr);
net/bridge/br_netfilter_hooks.c
1132
const struct nf_hook_entries *e;
net/bridge/br_netfilter_hooks.c
1138
e = rcu_dereference(net->nf.hooks_bridge[hook]);
net/bridge/br_netfilter_hooks.c
1139
if (!e)
net/bridge/br_netfilter_hooks.c
1142
ops = nf_hook_entries_get_hook_ops(e);
net/bridge/br_netfilter_hooks.c
1143
for (i = 0; i < e->num_hook_entries; i++) {
net/bridge/br_netfilter_hooks.c
1165
ret = nf_hook_slow(skb, &state, e, i);
net/bridge/netfilter/ebt_arp.c
104
const struct ebt_entry *e = par->entryinfo;
net/bridge/netfilter/ebt_arp.c
106
if ((e->ethproto != htons(ETH_P_ARP) &&
net/bridge/netfilter/ebt_arp.c
107
e->ethproto != htons(ETH_P_RARP)) ||
net/bridge/netfilter/ebt_arp.c
108
e->invflags & EBT_IPROTO)
net/bridge/netfilter/ebt_arpreply.c
65
const struct ebt_entry *e = par->entryinfo;
net/bridge/netfilter/ebt_arpreply.c
69
if (e->ethproto != htons(ETH_P_ARP) ||
net/bridge/netfilter/ebt_arpreply.c
70
e->invflags & EBT_IPROTO)
net/bridge/netfilter/ebt_ip.c
107
const struct ebt_entry *e = par->entryinfo;
net/bridge/netfilter/ebt_ip.c
109
if (e->ethproto != htons(ETH_P_IP) ||
net/bridge/netfilter/ebt_ip.c
110
e->invflags & EBT_IPROTO)
net/bridge/netfilter/ebt_ip6.c
108
const struct ebt_entry *e = par->entryinfo;
net/bridge/netfilter/ebt_ip6.c
111
if (e->ethproto != htons(ETH_P_IPV6) || e->invflags & EBT_IPROTO)
net/bridge/netfilter/ebt_stp.c
156
const struct ebt_entry *e = par->entryinfo;
net/bridge/netfilter/ebt_stp.c
163
(!ether_addr_equal(e->destmac, eth_stp_addr) ||
net/bridge/netfilter/ebt_stp.c
164
!(e->bitmask & EBT_DESTMAC) ||
net/bridge/netfilter/ebt_stp.c
165
!is_broadcast_ether_addr(e->destmsk)))
net/bridge/netfilter/ebt_vlan.c
80
const struct ebt_entry *e = par->entryinfo;
net/bridge/netfilter/ebt_vlan.c
83
if (e->ethproto != htons(ETH_P_8021Q)) {
net/bridge/netfilter/ebt_vlan.c
85
ntohs(e->ethproto));
net/bridge/netfilter/ebtables.c
131
ebt_basic_match(const struct ebt_entry *e, const struct sk_buff *skb,
net/bridge/netfilter/ebtables.c
143
if (e->bitmask & EBT_802_3) {
net/bridge/netfilter/ebtables.c
144
if (NF_INVF(e, EBT_IPROTO, eth_proto_is_802_3(ethproto)))
net/bridge/netfilter/ebtables.c
146
} else if (!(e->bitmask & EBT_NOPROTO) &&
net/bridge/netfilter/ebtables.c
147
NF_INVF(e, EBT_IPROTO, e->ethproto != ethproto))
net/bridge/netfilter/ebtables.c
1478
static inline int ebt_entry_to_user(struct ebt_entry *e, const char *base,
net/bridge/netfilter/ebtables.c
1485
if (e->bitmask == 0) {
net/bridge/netfilter/ebtables.c
1487
if (copy_to_user(ubase + ((char *)e - base), e,
net/bridge/netfilter/ebtables.c
1493
if (copy_to_user(ubase + ((char *)e - base), e, sizeof(*e)))
net/bridge/netfilter/ebtables.c
1496
hlp = ubase + (((char *)e + e->target_offset) - base);
net/bridge/netfilter/ebtables.c
1497
t = ebt_get_target_c(e);
net/bridge/netfilter/ebtables.c
1499
ret = EBT_MATCH_ITERATE(e, ebt_match_to_user, base, ubase);
net/bridge/netfilter/ebtables.c
150
if (NF_INVF(e, EBT_IIN, ebt_dev_check(e->in, in)))
net/bridge/netfilter/ebtables.c
1502
ret = EBT_WATCHER_ITERATE(e, ebt_watcher_to_user, base, ubase);
net/bridge/netfilter/ebtables.c
152
if (NF_INVF(e, EBT_IOUT, ebt_dev_check(e->out, out)))
net/bridge/netfilter/ebtables.c
156
NF_INVF(e, EBT_ILOGICALIN,
net/bridge/netfilter/ebtables.c
157
ebt_dev_check(e->logical_in, p->br->dev)))
net/bridge/netfilter/ebtables.c
160
NF_INVF(e, EBT_ILOGICALOUT,
net/bridge/netfilter/ebtables.c
161
ebt_dev_check(e->logical_out, p->br->dev)))
net/bridge/netfilter/ebtables.c
164
if (e->bitmask & EBT_SOURCEMAC) {
net/bridge/netfilter/ebtables.c
165
if (NF_INVF(e, EBT_ISOURCE,
net/bridge/netfilter/ebtables.c
166
!ether_addr_equal_masked(h->h_source, e->sourcemac,
net/bridge/netfilter/ebtables.c
167
e->sourcemsk)))
net/bridge/netfilter/ebtables.c
170
if (e->bitmask & EBT_DESTMAC) {
net/bridge/netfilter/ebtables.c
171
if (NF_INVF(e, EBT_IDEST,
net/bridge/netfilter/ebtables.c
1711
static int compat_copy_entry_to_user(struct ebt_entry *e, void __user **dstptr,
net/bridge/netfilter/ebtables.c
172
!ether_addr_equal_masked(h->h_dest, e->destmac,
net/bridge/netfilter/ebtables.c
1720
if (e->bitmask == 0) {
net/bridge/netfilter/ebtables.c
1723
if (copy_to_user(*dstptr, e, sizeof(struct ebt_entries)))
net/bridge/netfilter/ebtables.c
173
e->destmsk)))
net/bridge/netfilter/ebtables.c
1735
if (copy_to_user(ce, e, sizeof(*ce)))
net/bridge/netfilter/ebtables.c
1741
ret = EBT_MATCH_ITERATE(e, compat_match_to_user, dstptr, size);
net/bridge/netfilter/ebtables.c
1744
watchers_offset = e->watchers_offset - (origsize - *size);
net/bridge/netfilter/ebtables.c
1746
ret = EBT_WATCHER_ITERATE(e, compat_watcher_to_user, dstptr, size);
net/bridge/netfilter/ebtables.c
1749
target_offset = e->target_offset - (origsize - *size);
net/bridge/netfilter/ebtables.c
1751
t = ebt_get_target(e);
net/bridge/netfilter/ebtables.c
1756
next_offset = e->next_offset - (origsize - *size);
net/bridge/netfilter/ebtables.c
1781
static int compat_calc_entry(const struct ebt_entry *e,
net/bridge/netfilter/ebtables.c
1790
if (e->bitmask == 0)
net/bridge/netfilter/ebtables.c
1794
entry_offset = (void *)e - base;
net/bridge/netfilter/ebtables.c
1796
EBT_MATCH_ITERATE(e, compat_calc_match, &off);
net/bridge/netfilter/ebtables.c
1797
EBT_WATCHER_ITERATE(e, compat_calc_watcher, &off);
net/bridge/netfilter/ebtables.c
1799
t = ebt_get_target_c(e);
net/bridge/netfilter/ebtables.c
1813
(e < (struct ebt_entry *)(base - hookptr))) {
net/bridge/netfilter/ebtables.c
186
ebt_get_target_c(const struct ebt_entry *e)
net/bridge/netfilter/ebtables.c
188
return ebt_get_target((struct ebt_entry *)e);
net/bridge/netfilter/ebtables.c
273
point = cs[sp].e;
net/bridge/netfilter/ebtables.c
289
cs[sp].e = ebt_next_entry(point);
net/bridge/netfilter/ebtables.c
396
const struct ebt_entry *e = par->entryinfo;
net/bridge/netfilter/ebtables.c
398
size_t left = ((char *)e + e->watchers_offset) - (char *)m;
net/bridge/netfilter/ebtables.c
419
ntohs(e->ethproto), e->invflags & EBT_IPROTO);
net/bridge/netfilter/ebtables.c
433
const struct ebt_entry *e = par->entryinfo;
net/bridge/netfilter/ebtables.c
435
size_t left = ((char *)e + e->target_offset) - (char *)w;
net/bridge/netfilter/ebtables.c
456
ntohs(e->ethproto), e->invflags & EBT_IPROTO);
net/bridge/netfilter/ebtables.c
482
struct ebt_entry *e = (void *)newinfo->entries + offset;
net/bridge/netfilter/ebtables.c
495
if (i != NF_BR_NUMHOOKS || !(e->bitmask & EBT_ENTRY_OR_ENTRIES)) {
net/bridge/netfilter/ebtables.c
496
if (e->bitmask != 0) {
net/bridge/netfilter/ebtables.c
503
newinfo->hook_entry[i] = (struct ebt_entries *)e;
net/bridge/netfilter/ebtables.c
510
if (left < e->next_offset)
net/bridge/netfilter/ebtables.c
512
if (e->next_offset < sizeof(struct ebt_entry))
net/bridge/netfilter/ebtables.c
514
offset += e->next_offset;
net/bridge/netfilter/ebtables.c
533
ebt_check_entry_size_and_hooks(const struct ebt_entry *e,
net/bridge/netfilter/ebtables.c
541
if ((void *)e == (void *)newinfo->hook_entry[i])
net/bridge/netfilter/ebtables.c
547
if (i != NF_BR_NUMHOOKS || !e->bitmask) {
net/bridge/netfilter/ebtables.c
554
if (((struct ebt_entries *)e)->policy != EBT_DROP &&
net/bridge/netfilter/ebtables.c
555
((struct ebt_entries *)e)->policy != EBT_ACCEPT) {
net/bridge/netfilter/ebtables.c
558
((struct ebt_entries *)e)->policy != EBT_RETURN)
net/bridge/netfilter/ebtables.c
563
if (((struct ebt_entries *)e)->counter_offset != *totalcnt)
net/bridge/netfilter/ebtables.c
565
*n = ((struct ebt_entries *)e)->nentries;
net/bridge/netfilter/ebtables.c
570
if (sizeof(struct ebt_entry) > e->watchers_offset ||
net/bridge/netfilter/ebtables.c
571
e->watchers_offset > e->target_offset ||
net/bridge/netfilter/ebtables.c
572
e->target_offset >= e->next_offset)
net/bridge/netfilter/ebtables.c
576
if (e->next_offset - e->target_offset < sizeof(struct ebt_entry_target))
net/bridge/netfilter/ebtables.c
594
ebt_get_udc_positions(struct ebt_entry *e, struct ebt_table_info *newinfo,
net/bridge/netfilter/ebtables.c
600
if (e->bitmask)
net/bridge/netfilter/ebtables.c
603
if (newinfo->hook_entry[i] == (struct ebt_entries *)e)
net/bridge/netfilter/ebtables.c
610
udc[*n].cs.chaininfo = (struct ebt_entries *)e;
net/bridge/netfilter/ebtables.c
656
ebt_cleanup_entry(struct ebt_entry *e, struct net *net, unsigned int *cnt)
net/bridge/netfilter/ebtables.c
661
if (e->bitmask == 0)
net/bridge/netfilter/ebtables.c
666
EBT_WATCHER_ITERATE(e, ebt_cleanup_watcher, net, NULL);
net/bridge/netfilter/ebtables.c
667
EBT_MATCH_ITERATE(e, ebt_cleanup_match, net, NULL);
net/bridge/netfilter/ebtables.c
668
t = ebt_get_target(e);
net/bridge/netfilter/ebtables.c
681
ebt_check_entry(struct ebt_entry *e, struct net *net,
net/bridge/netfilter/ebtables.c
695
if (e->bitmask == 0)
net/bridge/netfilter/ebtables.c
698
if (e->bitmask & ~EBT_F_MASK)
net/bridge/netfilter/ebtables.c
701
if (e->invflags & ~EBT_INV_MASK)
net/bridge/netfilter/ebtables.c
704
if ((e->bitmask & EBT_NOPROTO) && (e->bitmask & EBT_802_3))
net/bridge/netfilter/ebtables.c
711
if ((char *)newinfo->hook_entry[i] < (char *)e)
net/bridge/netfilter/ebtables.c
723
if ((char *)(cl_s[i].cs.chaininfo) > (char *)e)
net/bridge/netfilter/ebtables.c
736
mtpar.entryinfo = tgpar.entryinfo = e;
net/bridge/netfilter/ebtables.c
739
ret = EBT_MATCH_ITERATE(e, ebt_check_match, &mtpar, &i);
net/bridge/netfilter/ebtables.c
743
ret = EBT_WATCHER_ITERATE(e, ebt_check_watcher, &tgpar, &j);
net/bridge/netfilter/ebtables.c
746
t = ebt_get_target(e);
net/bridge/netfilter/ebtables.c
747
gap = e->next_offset - e->target_offset;
net/bridge/netfilter/ebtables.c
782
ntohs(e->ethproto), e->invflags & EBT_IPROTO);
net/bridge/netfilter/ebtables.c
790
EBT_WATCHER_ITERATE(e, ebt_cleanup_watcher, net, &j);
net/bridge/netfilter/ebtables.c
792
EBT_MATCH_ITERATE(e, ebt_cleanup_match, net, &i);
net/bridge/netfilter/ebtables.c
804
const struct ebt_entry *e = (struct ebt_entry *)chain->data;
net/bridge/netfilter/ebtables.c
811
e = cl_s[chain_nr].cs.e;
net/bridge/netfilter/ebtables.c
824
t = ebt_get_target_c(e);
net/bridge/netfilter/ebtables.c
827
if (e->target_offset + sizeof(struct ebt_standard_target) >
net/bridge/netfilter/ebtables.c
828
e->next_offset)
net/bridge/netfilter/ebtables.c
850
cl_s[i].cs.e = ebt_next_entry(e);
net/bridge/netfilter/ebtables.c
851
e = (struct ebt_entry *)(hlp2->data);
net/bridge/netfilter/ebtables.c
860
e = ebt_next_entry(e);
net/caif/caif_dev.c
74
static void caifd_put(struct caif_device_entry *e)
net/caif/caif_dev.c
76
this_cpu_dec(*e->pcpu_refcnt);
net/caif/caif_dev.c
79
static void caifd_hold(struct caif_device_entry *e)
net/caif/caif_dev.c
81
this_cpu_inc(*e->pcpu_refcnt);
net/caif/caif_dev.c
84
static int caifd_refcnt_read(struct caif_device_entry *e)
net/caif/caif_dev.c
88
refcnt += *per_cpu_ptr(e->pcpu_refcnt, i);
net/ceph/crush/hash.c
134
__u32 crush_hash32_5(int type, __u32 a, __u32 b, __u32 c, __u32 d, __u32 e)
net/ceph/crush/hash.c
138
return crush_hash32_rjenkins1_5(a, b, c, d, e);
net/ceph/crush/hash.c
77
__u32 e)
net/ceph/crush/hash.c
79
__u32 hash = crush_hash_seed ^ a ^ b ^ c ^ d ^ e;
net/ceph/crush/hash.c
84
crush_hashmix(e, x, hash);
net/ceph/crush/hash.c
89
crush_hashmix(y, e, hash);
net/core/gen_estimator.c
62
static void est_fetch_counters(struct net_rate_estimator *e,
net/core/gen_estimator.c
66
if (e->stats_lock)
net/core/gen_estimator.c
67
spin_lock(e->stats_lock);
net/core/gen_estimator.c
69
gnet_stats_add_basic(b, e->cpu_bstats, e->bstats, e->running);
net/core/gen_estimator.c
71
if (e->stats_lock)
net/core/gen_estimator.c
72
spin_unlock(e->stats_lock);
net/dsa/port.c
33
static int dsa_port_notify(const struct dsa_port *dp, unsigned long e, void *v)
net/dsa/port.c
35
return dsa_tree_notify(dp->ds->dst, e, v);
net/dsa/switch.c
1084
int dsa_tree_notify(struct dsa_switch_tree *dst, unsigned long e, void *v)
net/dsa/switch.c
1089
err = raw_notifier_call_chain(nh, e, v);
net/dsa/switch.c
1106
int dsa_broadcast(unsigned long e, void *v)
net/dsa/switch.c
1112
err = dsa_tree_notify(dst, e, v);
net/dsa/switch.h
117
int dsa_tree_notify(struct dsa_switch_tree *dst, unsigned long e, void *v);
net/dsa/switch.h
118
int dsa_broadcast(unsigned long e, void *v);
net/dsa/user.c
1222
static int dsa_user_set_eee(struct net_device *dev, struct ethtool_keee *e)
net/dsa/user.c
1243
ret = ds->ops->set_mac_eee(ds, dp->index, e);
net/dsa/user.c
1247
ret = ds->ops->set_mac_eee(ds, dp->index, e);
net/dsa/user.c
1252
return phylink_ethtool_set_eee(dp->pl, e);
net/dsa/user.c
1255
static int dsa_user_get_eee(struct net_device *dev, struct ethtool_keee *e)
net/dsa/user.c
1268
return phylink_ethtool_get_eee(dp->pl, e);
net/ipv4/fib_frontend.c
1028
unsigned int e = 0, s_e;
net/ipv4/fib_frontend.c
1067
e = 0;
net/ipv4/fib_frontend.c
1070
if (e < s_e)
net/ipv4/fib_frontend.c
1080
e++;
net/ipv4/fib_frontend.c
1085
cb->args[1] = e;
net/ipv4/fou_core.c
1034
static void fou_build_udp(struct sk_buff *skb, struct ip_tunnel_encap *e,
net/ipv4/fou_core.c
1044
uh->dest = e->dport;
net/ipv4/fou_core.c
1047
udp_set_csum(!(e->flags & TUNNEL_ENCAP_FLAG_CSUM), skb,
net/ipv4/fou_core.c
1053
static int fou_build_header(struct sk_buff *skb, struct ip_tunnel_encap *e,
net/ipv4/fou_core.c
1056
int type = e->flags & TUNNEL_ENCAP_FLAG_CSUM ? SKB_GSO_UDP_TUNNEL_CSUM :
net/ipv4/fou_core.c
1061
err = __fou_build_header(skb, e, protocol, &sport, type);
net/ipv4/fou_core.c
1065
fou_build_udp(skb, e, fl4, protocol, sport);
net/ipv4/fou_core.c
1070
static int gue_build_header(struct sk_buff *skb, struct ip_tunnel_encap *e,
net/ipv4/fou_core.c
1073
int type = e->flags & TUNNEL_ENCAP_FLAG_CSUM ? SKB_GSO_UDP_TUNNEL_CSUM :
net/ipv4/fou_core.c
1078
err = __gue_build_header(skb, e, protocol, &sport, type);
net/ipv4/fou_core.c
1082
fou_build_udp(skb, e, fl4, protocol, sport);
net/ipv4/fou_core.c
919
size_t fou_encap_hlen(struct ip_tunnel_encap *e)
net/ipv4/fou_core.c
925
size_t gue_encap_hlen(struct ip_tunnel_encap *e)
net/ipv4/fou_core.c
932
if (e->flags & TUNNEL_ENCAP_FLAG_REMCSUM) {
net/ipv4/fou_core.c
943
int __fou_build_header(struct sk_buff *skb, struct ip_tunnel_encap *e,
net/ipv4/fou_core.c
952
*sport = e->sport ? : udp_flow_src_port(dev_net(skb->dev),
net/ipv4/fou_core.c
959
int __gue_build_header(struct sk_buff *skb, struct ip_tunnel_encap *e,
net/ipv4/fou_core.c
968
if ((e->flags & TUNNEL_ENCAP_FLAG_REMCSUM) &&
net/ipv4/fou_core.c
982
*sport = e->sport ? : udp_flow_src_port(dev_net(skb->dev),
net/ipv4/ipconfig.c
694
u8 *e = options;
net/ipv4/ipconfig.c
699
memcpy(e, ic_bootp_cookie, 4); /* RFC1048 Magic Cookie */
net/ipv4/ipconfig.c
700
e += 4;
net/ipv4/ipconfig.c
702
*e++ = 53; /* DHCP message type */
net/ipv4/ipconfig.c
703
*e++ = 1;
net/ipv4/ipconfig.c
704
*e++ = mt;
net/ipv4/ipconfig.c
707
*e++ = 54; /* Server ID (IP address) */
net/ipv4/ipconfig.c
708
*e++ = 4;
net/ipv4/ipconfig.c
709
memcpy(e, &ic_servaddr, 4);
net/ipv4/ipconfig.c
710
e += 4;
net/ipv4/ipconfig.c
712
*e++ = 50; /* Requested IP address */
net/ipv4/ipconfig.c
713
*e++ = 4;
net/ipv4/ipconfig.c
714
memcpy(e, &ic_myaddr, 4);
net/ipv4/ipconfig.c
715
e += 4;
net/ipv4/ipconfig.c
718
*e++ = 55; /* Parameter request list */
net/ipv4/ipconfig.c
719
*e++ = sizeof(ic_req_params);
net/ipv4/ipconfig.c
720
memcpy(e, ic_req_params, sizeof(ic_req_params));
net/ipv4/ipconfig.c
721
e += sizeof(ic_req_params);
net/ipv4/ipconfig.c
724
*e++ = 12; /* host-name */
net/ipv4/ipconfig.c
726
*e++ = len;
net/ipv4/ipconfig.c
727
memcpy(e, utsname()->nodename, len);
net/ipv4/ipconfig.c
728
e += len;
net/ipv4/ipconfig.c
733
*e++ = 60; /* Class-identifier */
net/ipv4/ipconfig.c
735
*e++ = len;
net/ipv4/ipconfig.c
736
memcpy(e, vendor_class_identifier, len);
net/ipv4/ipconfig.c
737
e += len;
net/ipv4/ipconfig.c
743
if (len >= 1 && len < 312 - (e - options) - 1) {
net/ipv4/ipconfig.c
744
*e++ = 61;
net/ipv4/ipconfig.c
745
*e++ = len + 1;
net/ipv4/ipconfig.c
746
memcpy(e, dhcp_client_identifier, len + 1);
net/ipv4/ipconfig.c
747
e += len + 1;
net/ipv4/ipconfig.c
750
*e++ = 255; /* End of the list */
net/ipv4/ipconfig.c
755
static void __init ic_bootp_init_ext(u8 *e)
net/ipv4/ipconfig.c
757
memcpy(e, ic_bootp_cookie, 4); /* RFC1048 Magic Cookie */
net/ipv4/ipconfig.c
758
e += 4;
net/ipv4/ipconfig.c
759
*e++ = 1; /* Subnet mask request */
net/ipv4/ipconfig.c
760
*e++ = 4;
net/ipv4/ipconfig.c
761
e += 4;
net/ipv4/ipconfig.c
762
*e++ = 3; /* Default gateway request */
net/ipv4/ipconfig.c
763
*e++ = 4;
net/ipv4/ipconfig.c
764
e += 4;
net/ipv4/ipconfig.c
766
*e++ = 6; /* (DNS) name server request */
net/ipv4/ipconfig.c
767
*e++ = 4 * CONF_NAMESERVERS_MAX;
net/ipv4/ipconfig.c
768
e += 4 * CONF_NAMESERVERS_MAX;
net/ipv4/ipconfig.c
770
*e++ = 12; /* Host name request */
net/ipv4/ipconfig.c
771
*e++ = 32;
net/ipv4/ipconfig.c
772
e += 32;
net/ipv4/ipconfig.c
773
*e++ = 40; /* NIS Domain name request */
net/ipv4/ipconfig.c
774
*e++ = 32;
net/ipv4/ipconfig.c
775
e += 32;
net/ipv4/ipconfig.c
776
*e++ = 17; /* Boot path */
net/ipv4/ipconfig.c
777
*e++ = 40;
net/ipv4/ipconfig.c
778
e += 40;
net/ipv4/ipconfig.c
780
*e++ = 57; /* set extension buffer size for reply */
net/ipv4/ipconfig.c
781
*e++ = 2;
net/ipv4/ipconfig.c
782
*e++ = 1; /* 128+236+8+20+14, see dhcpd sources */
net/ipv4/ipconfig.c
783
*e++ = 150;
net/ipv4/ipconfig.c
785
*e++ = 255; /* End of the list */
net/ipv4/ipmr.c
1002
struct nlmsgerr *e;
net/ipv4/ipmr.c
1018
e = nlmsg_data(nlh);
net/ipv4/ipmr.c
1019
e->error = -EMSGSIZE;
net/ipv4/ipmr.c
1020
memset(&e->msg, 0, sizeof(e->msg));
net/ipv4/ipmr.c
2982
unsigned int e = 0, s_e;
net/ipv4/ipmr.c
3030
if (e < s_e)
net/ipv4/ipmr.c
3039
e++;
net/ipv4/ipmr.c
3042
e = 0;
net/ipv4/ipmr.c
3051
cb->args[1] = e;
net/ipv4/ipmr.c
742
struct nlmsgerr *e;
net/ipv4/ipmr.c
753
e = nlmsg_data(nlh);
net/ipv4/ipmr.c
754
e->error = -ETIMEDOUT;
net/ipv4/ipmr.c
755
memset(&e->msg, 0, sizeof(e->msg));
net/ipv4/ipmr_base.c
305
unsigned int e = 0, s_e = cb->args[1];
net/ipv4/ipmr_base.c
315
if (e < s_e)
net/ipv4/ipmr_base.c
326
e++;
net/ipv4/ipmr_base.c
331
if (e < s_e)
net/ipv4/ipmr_base.c
341
e++;
net/ipv4/ipmr_base.c
346
cb->args[1] = e;
net/ipv4/netfilter/arp_tables.c
1065
static inline void compat_release_entry(struct compat_arpt_entry *e)
net/ipv4/netfilter/arp_tables.c
1069
t = compat_arpt_get_target(e);
net/ipv4/netfilter/arp_tables.c
1074
check_compat_entry_size_and_hooks(struct compat_arpt_entry *e,
net/ipv4/netfilter/arp_tables.c
1085
if ((unsigned long)e % __alignof__(struct compat_arpt_entry) != 0 ||
net/ipv4/netfilter/arp_tables.c
1086
(unsigned char *)e + sizeof(struct compat_arpt_entry) >= limit ||
net/ipv4/netfilter/arp_tables.c
1087
(unsigned char *)e + e->next_offset > limit)
net/ipv4/netfilter/arp_tables.c
1090
if (e->next_offset < sizeof(struct compat_arpt_entry) +
net/ipv4/netfilter/arp_tables.c
1094
if (!arp_checkentry(&e->arp))
net/ipv4/netfilter/arp_tables.c
1097
ret = xt_compat_check_entry_offsets(e, e->elems, e->target_offset,
net/ipv4/netfilter/arp_tables.c
1098
e->next_offset);
net/ipv4/netfilter/arp_tables.c
1103
entry_offset = (void *)e - (void *)base;
net/ipv4/netfilter/arp_tables.c
1105
t = compat_arpt_get_target(e);
net/ipv4/netfilter/arp_tables.c
1129
compat_copy_entry_from_user(struct compat_arpt_entry *e, void **dstptr,
net/ipv4/netfilter/arp_tables.c
1140
memcpy(de, e, sizeof(struct arpt_entry));
net/ipv4/netfilter/arp_tables.c
1141
memcpy(&de->counters, &e->counters, sizeof(e->counters));
net/ipv4/netfilter/arp_tables.c
1146
de->target_offset = e->target_offset - (origsize - *size);
net/ipv4/netfilter/arp_tables.c
1147
t = compat_arpt_get_target(e);
net/ipv4/netfilter/arp_tables.c
1150
de->next_offset = e->next_offset - (origsize - *size);
net/ipv4/netfilter/arp_tables.c
1305
static int compat_copy_entry_to_user(struct arpt_entry *e, void __user **dstptr,
net/ipv4/netfilter/arp_tables.c
1318
if (copy_to_user(ce, e, sizeof(struct arpt_entry)) != 0 ||
net/ipv4/netfilter/arp_tables.c
1326
target_offset = e->target_offset - (origsize - *size);
net/ipv4/netfilter/arp_tables.c
1328
t = arpt_get_target(e);
net/ipv4/netfilter/arp_tables.c
1332
next_offset = e->next_offset - (origsize - *size);
net/ipv4/netfilter/arp_tables.c
165
arpt_get_target_c(const struct arpt_entry *e)
net/ipv4/netfilter/arp_tables.c
167
return arpt_get_target((struct arpt_entry *)e);
net/ipv4/netfilter/arp_tables.c
191
struct arpt_entry *e, **jumpstack;
net/ipv4/netfilter/arp_tables.c
215
e = get_entry(table_base, private->hook_entry[hook]);
net/ipv4/netfilter/arp_tables.c
225
if (!arp_packet_match(arp, skb->dev, indev, outdev, &e->arp)) {
net/ipv4/netfilter/arp_tables.c
226
e = arpt_next_entry(e);
net/ipv4/netfilter/arp_tables.c
230
counter = xt_get_this_cpu_counter(&e->counters);
net/ipv4/netfilter/arp_tables.c
233
t = arpt_get_target_c(e);
net/ipv4/netfilter/arp_tables.c
247
e = get_entry(table_base,
net/ipv4/netfilter/arp_tables.c
250
e = jumpstack[--stackidx];
net/ipv4/netfilter/arp_tables.c
251
e = arpt_next_entry(e);
net/ipv4/netfilter/arp_tables.c
256
!= arpt_next_entry(e)) {
net/ipv4/netfilter/arp_tables.c
261
jumpstack[stackidx++] = e;
net/ipv4/netfilter/arp_tables.c
264
e = get_entry(table_base, v);
net/ipv4/netfilter/arp_tables.c
275
e = arpt_next_entry(e);
net/ipv4/netfilter/arp_tables.c
291
static inline bool unconditional(const struct arpt_entry *e)
net/ipv4/netfilter/arp_tables.c
295
return e->target_offset == sizeof(struct arpt_entry) &&
net/ipv4/netfilter/arp_tables.c
296
memcmp(&e->arp, &uncond, sizeof(uncond)) == 0;
net/ipv4/netfilter/arp_tables.c
313
struct arpt_entry *e = entry0 + pos;
net/ipv4/netfilter/arp_tables.c
319
e->counters.pcnt = pos;
net/ipv4/netfilter/arp_tables.c
323
= (void *)arpt_get_target_c(e);
net/ipv4/netfilter/arp_tables.c
324
int visited = e->comefrom & (1 << hook);
net/ipv4/netfilter/arp_tables.c
326
if (e->comefrom & (1 << NF_ARP_NUMHOOKS))
net/ipv4/netfilter/arp_tables.c
329
e->comefrom
net/ipv4/netfilter/arp_tables.c
333
if ((unconditional(e) &&
net/ipv4/netfilter/arp_tables.c
343
e->comefrom ^= (1<<NF_ARP_NUMHOOKS);
net/ipv4/netfilter/arp_tables.c
345
pos = e->counters.pcnt;
net/ipv4/netfilter/arp_tables.c
346
e->counters.pcnt = 0;
net/ipv4/netfilter/arp_tables.c
352
e = entry0 + pos;
net/ipv4/netfilter/arp_tables.c
353
} while (oldpos == pos + e->next_offset);
net/ipv4/netfilter/arp_tables.c
356
size = e->next_offset;
net/ipv4/netfilter/arp_tables.c
357
e = entry0 + pos + size;
net/ipv4/netfilter/arp_tables.c
360
e->counters.pcnt = pos;
net/ipv4/netfilter/arp_tables.c
374
newpos = pos + e->next_offset;
net/ipv4/netfilter/arp_tables.c
378
e = entry0 + newpos;
net/ipv4/netfilter/arp_tables.c
379
e->counters.pcnt = pos;
net/ipv4/netfilter/arp_tables.c
388
static int check_target(struct arpt_entry *e, struct net *net, const char *name)
net/ipv4/netfilter/arp_tables.c
390
struct xt_entry_target *t = arpt_get_target(e);
net/ipv4/netfilter/arp_tables.c
394
.entryinfo = e,
net/ipv4/netfilter/arp_tables.c
397
.hook_mask = e->comefrom,
net/ipv4/netfilter/arp_tables.c
405
find_check_entry(struct arpt_entry *e, struct net *net, const char *name,
net/ipv4/netfilter/arp_tables.c
413
if (!xt_percpu_counter_alloc(alloc_state, &e->counters))
net/ipv4/netfilter/arp_tables.c
416
t = arpt_get_target(e);
net/ipv4/netfilter/arp_tables.c
425
ret = check_target(e, net, name);
net/ipv4/netfilter/arp_tables.c
432
xt_percpu_counter_free(&e->counters);
net/ipv4/netfilter/arp_tables.c
437
static bool check_underflow(const struct arpt_entry *e)
net/ipv4/netfilter/arp_tables.c
442
if (!unconditional(e))
net/ipv4/netfilter/arp_tables.c
444
t = arpt_get_target_c(e);
net/ipv4/netfilter/arp_tables.c
452
static inline int check_entry_size_and_hooks(struct arpt_entry *e,
net/ipv4/netfilter/arp_tables.c
463
if ((unsigned long)e % __alignof__(struct arpt_entry) != 0 ||
net/ipv4/netfilter/arp_tables.c
464
(unsigned char *)e + sizeof(struct arpt_entry) >= limit ||
net/ipv4/netfilter/arp_tables.c
465
(unsigned char *)e + e->next_offset > limit)
net/ipv4/netfilter/arp_tables.c
468
if (e->next_offset
net/ipv4/netfilter/arp_tables.c
472
if (!arp_checkentry(&e->arp))
net/ipv4/netfilter/arp_tables.c
475
err = xt_check_entry_offsets(e, e->elems, e->target_offset,
net/ipv4/netfilter/arp_tables.c
476
e->next_offset);
net/ipv4/netfilter/arp_tables.c
484
if ((unsigned char *)e - base == hook_entries[h])
net/ipv4/netfilter/arp_tables.c
486
if ((unsigned char *)e - base == underflows[h]) {
net/ipv4/netfilter/arp_tables.c
487
if (!check_underflow(e))
net/ipv4/netfilter/arp_tables.c
495
e->counters = ((struct xt_counters) { 0, 0 });
net/ipv4/netfilter/arp_tables.c
496
e->comefrom = 0;
net/ipv4/netfilter/arp_tables.c
500
static void cleanup_entry(struct arpt_entry *e, struct net *net)
net/ipv4/netfilter/arp_tables.c
505
t = arpt_get_target(e);
net/ipv4/netfilter/arp_tables.c
513
xt_percpu_counter_free(&e->counters);
net/ipv4/netfilter/arp_tables.c
675
const struct arpt_entry *e;
net/ipv4/netfilter/arp_tables.c
689
for (off = 0, num = 0; off < total_size; off += e->next_offset, num++){
net/ipv4/netfilter/arp_tables.c
692
e = loc_cpu_entry + off;
net/ipv4/netfilter/arp_tables.c
693
if (copy_to_user(userptr + off, e, sizeof(*e))) {
net/ipv4/netfilter/arp_tables.c
705
t = arpt_get_target_c(e);
net/ipv4/netfilter/arp_tables.c
706
if (xt_target_to_user(t, userptr + off + e->target_offset)) {
net/ipv4/netfilter/arp_tables.c
736
static int compat_calc_entry(const struct arpt_entry *e,
net/ipv4/netfilter/arp_tables.c
745
entry_offset = (void *)e - base;
net/ipv4/netfilter/arp_tables.c
747
t = arpt_get_target_c(e);
net/ipv4/netfilter/arp_tables.c
756
(e < (struct arpt_entry *)(base + info->hook_entry[i])))
net/ipv4/netfilter/arp_tables.c
759
(e < (struct arpt_entry *)(base + info->underflow[i])))
net/ipv4/netfilter/ip_tables.c
108
static inline bool unconditional(const struct ipt_entry *e)
net/ipv4/netfilter/ip_tables.c
112
return e->target_offset == sizeof(struct ipt_entry) &&
net/ipv4/netfilter/ip_tables.c
113
memcmp(&e->ip, &uncond, sizeof(uncond)) == 0;
net/ipv4/netfilter/ip_tables.c
118
ipt_get_target_c(const struct ipt_entry *e)
net/ipv4/netfilter/ip_tables.c
120
return ipt_get_target((struct ipt_entry *)e);
net/ipv4/netfilter/ip_tables.c
1218
compat_copy_entry_to_user(struct ipt_entry *e, void __user **dstptr,
net/ipv4/netfilter/ip_tables.c
1231
if (copy_to_user(ce, e, sizeof(struct ipt_entry)) != 0 ||
net/ipv4/netfilter/ip_tables.c
1239
xt_ematch_foreach(ematch, e) {
net/ipv4/netfilter/ip_tables.c
1244
target_offset = e->target_offset - (origsize - *size);
net/ipv4/netfilter/ip_tables.c
1245
t = ipt_get_target(e);
net/ipv4/netfilter/ip_tables.c
1249
next_offset = e->next_offset - (origsize - *size);
net/ipv4/netfilter/ip_tables.c
1273
static void compat_release_entry(struct compat_ipt_entry *e)
net/ipv4/netfilter/ip_tables.c
1279
xt_ematch_foreach(ematch, e)
net/ipv4/netfilter/ip_tables.c
1281
t = compat_ipt_get_target(e);
net/ipv4/netfilter/ip_tables.c
1286
check_compat_entry_size_and_hooks(struct compat_ipt_entry *e,
net/ipv4/netfilter/ip_tables.c
1299
if ((unsigned long)e % __alignof__(struct compat_ipt_entry) != 0 ||
net/ipv4/netfilter/ip_tables.c
1300
(unsigned char *)e + sizeof(struct compat_ipt_entry) >= limit ||
net/ipv4/netfilter/ip_tables.c
1301
(unsigned char *)e + e->next_offset > limit)
net/ipv4/netfilter/ip_tables.c
1304
if (e->next_offset < sizeof(struct compat_ipt_entry) +
net/ipv4/netfilter/ip_tables.c
1308
if (!ip_checkentry(&e->ip))
net/ipv4/netfilter/ip_tables.c
1311
ret = xt_compat_check_entry_offsets(e, e->elems,
net/ipv4/netfilter/ip_tables.c
1312
e->target_offset, e->next_offset);
net/ipv4/netfilter/ip_tables.c
1317
entry_offset = (void *)e - (void *)base;
net/ipv4/netfilter/ip_tables.c
1319
xt_ematch_foreach(ematch, e) {
net/ipv4/netfilter/ip_tables.c
1320
ret = compat_find_calc_match(ematch, &e->ip, &off);
net/ipv4/netfilter/ip_tables.c
1326
t = compat_ipt_get_target(e);
net/ipv4/netfilter/ip_tables.c
1346
xt_ematch_foreach(ematch, e) {
net/ipv4/netfilter/ip_tables.c
1355
compat_copy_entry_from_user(struct compat_ipt_entry *e, void **dstptr,
net/ipv4/netfilter/ip_tables.c
1367
memcpy(de, e, sizeof(struct ipt_entry));
net/ipv4/netfilter/ip_tables.c
1368
memcpy(&de->counters, &e->counters, sizeof(e->counters));
net/ipv4/netfilter/ip_tables.c
1373
xt_ematch_foreach(ematch, e)
net/ipv4/netfilter/ip_tables.c
1376
de->target_offset = e->target_offset - (origsize - *size);
net/ipv4/netfilter/ip_tables.c
1377
t = compat_ipt_get_target(e);
net/ipv4/netfilter/ip_tables.c
1380
de->next_offset = e->next_offset - (origsize - *size);
net/ipv4/netfilter/ip_tables.c
156
get_chainname_rulenum(const struct ipt_entry *s, const struct ipt_entry *e,
net/ipv4/netfilter/ip_tables.c
166
} else if (s == e) {
net/ipv4/netfilter/ip_tables.c
192
const struct ipt_entry *e)
net/ipv4/netfilter/ip_tables.c
205
if (get_chainname_rulenum(iter, e, hookname,
net/ipv4/netfilter/ip_tables.c
235
struct ipt_entry *e, **jumpstack;
net/ipv4/netfilter/ip_tables.c
275
e = get_entry(table_base, private->hook_entry[hook]);
net/ipv4/netfilter/ip_tables.c
282
WARN_ON(!e);
net/ipv4/netfilter/ip_tables.c
284
&e->ip, acpar.fragoff)) {
net/ipv4/netfilter/ip_tables.c
286
e = ipt_next_entry(e);
net/ipv4/netfilter/ip_tables.c
290
xt_ematch_foreach(ematch, e) {
net/ipv4/netfilter/ip_tables.c
297
counter = xt_get_this_cpu_counter(&e->counters);
net/ipv4/netfilter/ip_tables.c
300
t = ipt_get_target_c(e);
net/ipv4/netfilter/ip_tables.c
307
state->out, table->name, private, e);
net/ipv4/netfilter/ip_tables.c
321
e = get_entry(table_base,
net/ipv4/netfilter/ip_tables.c
324
e = jumpstack[--stackidx];
net/ipv4/netfilter/ip_tables.c
325
e = ipt_next_entry(e);
net/ipv4/netfilter/ip_tables.c
329
if (table_base + v != ipt_next_entry(e) &&
net/ipv4/netfilter/ip_tables.c
330
!(e->ip.flags & IPT_F_GOTO)) {
net/ipv4/netfilter/ip_tables.c
335
jumpstack[stackidx++] = e;
net/ipv4/netfilter/ip_tables.c
338
e = get_entry(table_base, v);
net/ipv4/netfilter/ip_tables.c
349
e = ipt_next_entry(e);
net/ipv4/netfilter/ip_tables.c
377
struct ipt_entry *e = entry0 + pos;
net/ipv4/netfilter/ip_tables.c
383
e->counters.pcnt = pos;
net/ipv4/netfilter/ip_tables.c
387
= (void *)ipt_get_target_c(e);
net/ipv4/netfilter/ip_tables.c
388
int visited = e->comefrom & (1 << hook);
net/ipv4/netfilter/ip_tables.c
390
if (e->comefrom & (1 << NF_INET_NUMHOOKS))
net/ipv4/netfilter/ip_tables.c
393
e->comefrom |= ((1 << hook) | (1 << NF_INET_NUMHOOKS));
net/ipv4/netfilter/ip_tables.c
396
if ((unconditional(e) &&
net/ipv4/netfilter/ip_tables.c
405
e->comefrom ^= (1<<NF_INET_NUMHOOKS);
net/ipv4/netfilter/ip_tables.c
407
pos = e->counters.pcnt;
net/ipv4/netfilter/ip_tables.c
408
e->counters.pcnt = 0;
net/ipv4/netfilter/ip_tables.c
414
e = entry0 + pos;
net/ipv4/netfilter/ip_tables.c
415
} while (oldpos == pos + e->next_offset);
net/ipv4/netfilter/ip_tables.c
418
size = e->next_offset;
net/ipv4/netfilter/ip_tables.c
419
e = entry0 + pos + size;
net/ipv4/netfilter/ip_tables.c
422
e->counters.pcnt = pos;
net/ipv4/netfilter/ip_tables.c
436
newpos = pos + e->next_offset;
net/ipv4/netfilter/ip_tables.c
440
e = entry0 + newpos;
net/ipv4/netfilter/ip_tables.c
441
e->counters.pcnt = pos;
net/ipv4/netfilter/ip_tables.c
497
static int check_target(struct ipt_entry *e, struct net *net, const char *name)
net/ipv4/netfilter/ip_tables.c
499
struct xt_entry_target *t = ipt_get_target(e);
net/ipv4/netfilter/ip_tables.c
503
.entryinfo = e,
net/ipv4/netfilter/ip_tables.c
506
.hook_mask = e->comefrom,
net/ipv4/netfilter/ip_tables.c
511
e->ip.proto, e->ip.invflags & IPT_INV_PROTO);
net/ipv4/netfilter/ip_tables.c
515
find_check_entry(struct ipt_entry *e, struct net *net, const char *name,
net/ipv4/netfilter/ip_tables.c
526
if (!xt_percpu_counter_alloc(alloc_state, &e->counters))
net/ipv4/netfilter/ip_tables.c
533
mtpar.entryinfo = &e->ip;
net/ipv4/netfilter/ip_tables.c
534
mtpar.hook_mask = e->comefrom;
net/ipv4/netfilter/ip_tables.c
536
xt_ematch_foreach(ematch, e) {
net/ipv4/netfilter/ip_tables.c
543
t = ipt_get_target(e);
net/ipv4/netfilter/ip_tables.c
552
ret = check_target(e, net, name);
net/ipv4/netfilter/ip_tables.c
560
xt_ematch_foreach(ematch, e) {
net/ipv4/netfilter/ip_tables.c
566
xt_percpu_counter_free(&e->counters);
net/ipv4/netfilter/ip_tables.c
571
static bool check_underflow(const struct ipt_entry *e)
net/ipv4/netfilter/ip_tables.c
576
if (!unconditional(e))
net/ipv4/netfilter/ip_tables.c
578
t = ipt_get_target_c(e);
net/ipv4/netfilter/ip_tables.c
587
check_entry_size_and_hooks(struct ipt_entry *e,
net/ipv4/netfilter/ip_tables.c
598
if ((unsigned long)e % __alignof__(struct ipt_entry) != 0 ||
net/ipv4/netfilter/ip_tables.c
599
(unsigned char *)e + sizeof(struct ipt_entry) >= limit ||
net/ipv4/netfilter/ip_tables.c
600
(unsigned char *)e + e->next_offset > limit)
net/ipv4/netfilter/ip_tables.c
603
if (e->next_offset
net/ipv4/netfilter/ip_tables.c
607
if (!ip_checkentry(&e->ip))
net/ipv4/netfilter/ip_tables.c
610
err = xt_check_entry_offsets(e, e->elems, e->target_offset,
net/ipv4/netfilter/ip_tables.c
611
e->next_offset);
net/ipv4/netfilter/ip_tables.c
619
if ((unsigned char *)e - base == hook_entries[h])
net/ipv4/netfilter/ip_tables.c
621
if ((unsigned char *)e - base == underflows[h]) {
net/ipv4/netfilter/ip_tables.c
622
if (!check_underflow(e))
net/ipv4/netfilter/ip_tables.c
630
e->counters = ((struct xt_counters) { 0, 0 });
net/ipv4/netfilter/ip_tables.c
631
e->comefrom = 0;
net/ipv4/netfilter/ip_tables.c
636
cleanup_entry(struct ipt_entry *e, struct net *net)
net/ipv4/netfilter/ip_tables.c
643
xt_ematch_foreach(ematch, e)
net/ipv4/netfilter/ip_tables.c
645
t = ipt_get_target(e);
net/ipv4/netfilter/ip_tables.c
654
xt_percpu_counter_free(&e->counters);
net/ipv4/netfilter/ip_tables.c
815
const struct ipt_entry *e;
net/ipv4/netfilter/ip_tables.c
829
for (off = 0, num = 0; off < total_size; off += e->next_offset, num++){
net/ipv4/netfilter/ip_tables.c
834
e = loc_cpu_entry + off;
net/ipv4/netfilter/ip_tables.c
835
if (copy_to_user(userptr + off, e, sizeof(*e))) {
net/ipv4/netfilter/ip_tables.c
848
i < e->target_offset;
net/ipv4/netfilter/ip_tables.c
850
m = (void *)e + i;
net/ipv4/netfilter/ip_tables.c
858
t = ipt_get_target_c(e);
net/ipv4/netfilter/ip_tables.c
859
if (xt_target_to_user(t, userptr + off + e->target_offset)) {
net/ipv4/netfilter/ip_tables.c
889
static int compat_calc_entry(const struct ipt_entry *e,
net/ipv4/netfilter/ip_tables.c
899
entry_offset = (void *)e - base;
net/ipv4/netfilter/ip_tables.c
900
xt_ematch_foreach(ematch, e)
net/ipv4/netfilter/ip_tables.c
902
t = ipt_get_target_c(e);
net/ipv4/netfilter/ip_tables.c
911
(e < (struct ipt_entry *)(base + info->hook_entry[i])))
net/ipv4/netfilter/ip_tables.c
914
(e < (struct ipt_entry *)(base + info->underflow[i])))
net/ipv4/netfilter/ipt_ECN.c
105
(e->ip.proto != IPPROTO_TCP || (e->ip.invflags & XT_INV_PROTO))) {
net/ipv4/netfilter/ipt_ECN.c
96
const struct ipt_entry *e = par->entryinfo;
net/ipv4/netfilter/ipt_REJECT.c
72
const struct ipt_entry *e = par->entryinfo;
net/ipv4/netfilter/ipt_REJECT.c
79
if (e->ip.proto != IPPROTO_TCP ||
net/ipv4/netfilter/ipt_REJECT.c
80
(e->ip.invflags & XT_INV_PROTO)) {
net/ipv4/netfilter/ipt_SYNPROXY.c
67
const struct ipt_entry *e = par->entryinfo;
net/ipv4/netfilter/ipt_SYNPROXY.c
70
if (e->ip.proto != IPPROTO_TCP ||
net/ipv4/netfilter/ipt_SYNPROXY.c
71
e->ip.invflags & XT_INV_PROTO)
net/ipv4/tcp_cong.c
28
struct tcp_congestion_ops *e;
net/ipv4/tcp_cong.c
30
list_for_each_entry_rcu(e, &tcp_cong_list, list) {
net/ipv4/tcp_cong.c
31
if (strcmp(e->name, name) == 0)
net/ipv4/tcp_cong.c
32
return e;
net/ipv4/tcp_cong.c
68
struct tcp_congestion_ops *e;
net/ipv4/tcp_cong.c
70
list_for_each_entry_rcu(e, &tcp_cong_list, list) {
net/ipv4/tcp_cong.c
71
if (e->key == key)
net/ipv4/tcp_cong.c
72
return e;
net/ipv4/tcp_sigpool.c
120
static int __cpool_alloc_ahash(struct sigpool_entry *e, const char *alg)
net/ipv4/tcp_sigpool.c
125
e->alg = kstrdup(alg, GFP_KERNEL);
net/ipv4/tcp_sigpool.c
126
if (!e->alg)
net/ipv4/tcp_sigpool.c
135
e->needs_key = crypto_ahash_get_flags(cpu0_hash) & CRYPTO_TFM_NEED_KEY;
net/ipv4/tcp_sigpool.c
140
e->hash = cpu0_hash;
net/ipv4/tcp_sigpool.c
141
kref_init(&e->kref);
net/ipv4/tcp_sigpool.c
147
kfree(e->alg);
net/ipv4/tcp_sigpool.c
148
e->alg = NULL;
net/ipv4/tcp_sigpool.c
200
static void __cpool_free_entry(struct sigpool_entry *e)
net/ipv4/tcp_sigpool.c
202
crypto_free_ahash(e->hash);
net/ipv4/tcp_sigpool.c
203
kfree(e->alg);
net/ipv4/tcp_sigpool.c
204
memset(e, 0, sizeof(*e));
net/ipv4/tcp_ulp.c
23
struct tcp_ulp_ops *e;
net/ipv4/tcp_ulp.c
25
list_for_each_entry_rcu(e, &tcp_ulp_list, list,
net/ipv4/tcp_ulp.c
27
if (strcmp(e->name, name) == 0)
net/ipv4/tcp_ulp.c
28
return e;
net/ipv6/fou6.c
21
static void fou6_build_udp(struct sk_buff *skb, struct ip_tunnel_encap *e,
net/ipv6/fou6.c
31
uh->dest = e->dport;
net/ipv6/fou6.c
34
udp6_set_csum(!(e->flags & TUNNEL_ENCAP_FLAG_CSUM6), skb,
net/ipv6/fou6.c
40
static int fou6_build_header(struct sk_buff *skb, struct ip_tunnel_encap *e,
net/ipv6/fou6.c
45
int type = e->flags & TUNNEL_ENCAP_FLAG_CSUM6 ?
net/ipv6/fou6.c
48
err = __fou_build_header(skb, e, protocol, &sport, type);
net/ipv6/fou6.c
52
fou6_build_udp(skb, e, fl6, protocol, sport);
net/ipv6/fou6.c
57
static int gue6_build_header(struct sk_buff *skb, struct ip_tunnel_encap *e,
net/ipv6/fou6.c
62
int type = e->flags & TUNNEL_ENCAP_FLAG_CSUM6 ?
net/ipv6/fou6.c
65
err = __gue_build_header(skb, e, protocol, &sport, type);
net/ipv6/fou6.c
69
fou6_build_udp(skb, e, fl6, protocol, sport);
net/ipv6/ip6_fib.c
637
unsigned int e = 0, s_e;
net/ipv6/ip6_fib.c
705
e = 0;
net/ipv6/ip6_fib.c
708
if (e < s_e)
net/ipv6/ip6_fib.c
714
e++;
net/ipv6/ip6_fib.c
718
cb->args[1] = e;
net/ipv6/mcast.c
114
#define mc_dereference(e, idev) \
net/ipv6/mcast.c
115
rcu_dereference_protected(e, lockdep_is_held(&(idev)->mc_lock))
net/ipv6/mcast.c
117
#define sock_dereference(e, sk) \
net/ipv6/mcast.c
118
rcu_dereference_protected(e, lockdep_sock_is_held(sk))
net/ipv6/netfilter/ip6_tables.c
1234
compat_copy_entry_to_user(struct ip6t_entry *e, void __user **dstptr,
net/ipv6/netfilter/ip6_tables.c
1247
if (copy_to_user(ce, e, sizeof(struct ip6t_entry)) != 0 ||
net/ipv6/netfilter/ip6_tables.c
1255
xt_ematch_foreach(ematch, e) {
net/ipv6/netfilter/ip6_tables.c
1260
target_offset = e->target_offset - (origsize - *size);
net/ipv6/netfilter/ip6_tables.c
1261
t = ip6t_get_target(e);
net/ipv6/netfilter/ip6_tables.c
1265
next_offset = e->next_offset - (origsize - *size);
net/ipv6/netfilter/ip6_tables.c
1289
static void compat_release_entry(struct compat_ip6t_entry *e)
net/ipv6/netfilter/ip6_tables.c
1295
xt_ematch_foreach(ematch, e)
net/ipv6/netfilter/ip6_tables.c
1297
t = compat_ip6t_get_target(e);
net/ipv6/netfilter/ip6_tables.c
1302
check_compat_entry_size_and_hooks(struct compat_ip6t_entry *e,
net/ipv6/netfilter/ip6_tables.c
1315
if ((unsigned long)e % __alignof__(struct compat_ip6t_entry) != 0 ||
net/ipv6/netfilter/ip6_tables.c
1316
(unsigned char *)e + sizeof(struct compat_ip6t_entry) >= limit ||
net/ipv6/netfilter/ip6_tables.c
1317
(unsigned char *)e + e->next_offset > limit)
net/ipv6/netfilter/ip6_tables.c
1320
if (e->next_offset < sizeof(struct compat_ip6t_entry) +
net/ipv6/netfilter/ip6_tables.c
1324
if (!ip6_checkentry(&e->ipv6))
net/ipv6/netfilter/ip6_tables.c
1327
ret = xt_compat_check_entry_offsets(e, e->elems,
net/ipv6/netfilter/ip6_tables.c
1328
e->target_offset, e->next_offset);
net/ipv6/netfilter/ip6_tables.c
133
static inline bool unconditional(const struct ip6t_entry *e)
net/ipv6/netfilter/ip6_tables.c
1333
entry_offset = (void *)e - (void *)base;
net/ipv6/netfilter/ip6_tables.c
1335
xt_ematch_foreach(ematch, e) {
net/ipv6/netfilter/ip6_tables.c
1336
ret = compat_find_calc_match(ematch, &e->ipv6, &off);
net/ipv6/netfilter/ip6_tables.c
1342
t = compat_ip6t_get_target(e);
net/ipv6/netfilter/ip6_tables.c
1362
xt_ematch_foreach(ematch, e) {
net/ipv6/netfilter/ip6_tables.c
137
return e->target_offset == sizeof(struct ip6t_entry) &&
net/ipv6/netfilter/ip6_tables.c
1371
compat_copy_entry_from_user(struct compat_ip6t_entry *e, void **dstptr,
net/ipv6/netfilter/ip6_tables.c
138
memcmp(&e->ipv6, &uncond, sizeof(uncond)) == 0;
net/ipv6/netfilter/ip6_tables.c
1383
memcpy(de, e, sizeof(struct ip6t_entry));
net/ipv6/netfilter/ip6_tables.c
1384
memcpy(&de->counters, &e->counters, sizeof(e->counters));
net/ipv6/netfilter/ip6_tables.c
1389
xt_ematch_foreach(ematch, e)
net/ipv6/netfilter/ip6_tables.c
1392
de->target_offset = e->target_offset - (origsize - *size);
net/ipv6/netfilter/ip6_tables.c
1393
t = compat_ip6t_get_target(e);
net/ipv6/netfilter/ip6_tables.c
1396
de->next_offset = e->next_offset - (origsize - *size);
net/ipv6/netfilter/ip6_tables.c
142
ip6t_get_target_c(const struct ip6t_entry *e)
net/ipv6/netfilter/ip6_tables.c
144
return ip6t_get_target((struct ip6t_entry *)e);
net/ipv6/netfilter/ip6_tables.c
181
get_chainname_rulenum(const struct ip6t_entry *s, const struct ip6t_entry *e,
net/ipv6/netfilter/ip6_tables.c
191
} else if (s == e) {
net/ipv6/netfilter/ip6_tables.c
217
const struct ip6t_entry *e)
net/ipv6/netfilter/ip6_tables.c
230
if (get_chainname_rulenum(iter, e, hookname,
net/ipv6/netfilter/ip6_tables.c
258
struct ip6t_entry *e, **jumpstack;
net/ipv6/netfilter/ip6_tables.c
297
e = get_entry(table_base, private->hook_entry[hook]);
net/ipv6/netfilter/ip6_tables.c
304
WARN_ON(!e);
net/ipv6/netfilter/ip6_tables.c
306
if (!ip6_packet_match(skb, indev, outdev, &e->ipv6,
net/ipv6/netfilter/ip6_tables.c
309
e = ip6t_next_entry(e);
net/ipv6/netfilter/ip6_tables.c
313
xt_ematch_foreach(ematch, e) {
net/ipv6/netfilter/ip6_tables.c
320
counter = xt_get_this_cpu_counter(&e->counters);
net/ipv6/netfilter/ip6_tables.c
323
t = ip6t_get_target_c(e);
net/ipv6/netfilter/ip6_tables.c
330
state->out, table->name, private, e);
net/ipv6/netfilter/ip6_tables.c
344
e = get_entry(table_base,
net/ipv6/netfilter/ip6_tables.c
347
e = ip6t_next_entry(jumpstack[--stackidx]);
net/ipv6/netfilter/ip6_tables.c
350
if (table_base + v != ip6t_next_entry(e) &&
net/ipv6/netfilter/ip6_tables.c
351
!(e->ipv6.flags & IP6T_F_GOTO)) {
net/ipv6/netfilter/ip6_tables.c
356
jumpstack[stackidx++] = e;
net/ipv6/netfilter/ip6_tables.c
359
e = get_entry(table_base, v);
net/ipv6/netfilter/ip6_tables.c
368
e = ip6t_next_entry(e);
net/ipv6/netfilter/ip6_tables.c
395
struct ip6t_entry *e = entry0 + pos;
net/ipv6/netfilter/ip6_tables.c
401
e->counters.pcnt = pos;
net/ipv6/netfilter/ip6_tables.c
405
= (void *)ip6t_get_target_c(e);
net/ipv6/netfilter/ip6_tables.c
406
int visited = e->comefrom & (1 << hook);
net/ipv6/netfilter/ip6_tables.c
408
if (e->comefrom & (1 << NF_INET_NUMHOOKS))
net/ipv6/netfilter/ip6_tables.c
411
e->comefrom |= ((1 << hook) | (1 << NF_INET_NUMHOOKS));
net/ipv6/netfilter/ip6_tables.c
414
if ((unconditional(e) &&
net/ipv6/netfilter/ip6_tables.c
423
e->comefrom ^= (1<<NF_INET_NUMHOOKS);
net/ipv6/netfilter/ip6_tables.c
425
pos = e->counters.pcnt;
net/ipv6/netfilter/ip6_tables.c
426
e->counters.pcnt = 0;
net/ipv6/netfilter/ip6_tables.c
432
e = entry0 + pos;
net/ipv6/netfilter/ip6_tables.c
433
} while (oldpos == pos + e->next_offset);
net/ipv6/netfilter/ip6_tables.c
436
size = e->next_offset;
net/ipv6/netfilter/ip6_tables.c
437
e = entry0 + pos + size;
net/ipv6/netfilter/ip6_tables.c
440
e->counters.pcnt = pos;
net/ipv6/netfilter/ip6_tables.c
454
newpos = pos + e->next_offset;
net/ipv6/netfilter/ip6_tables.c
458
e = entry0 + newpos;
net/ipv6/netfilter/ip6_tables.c
459
e->counters.pcnt = pos;
net/ipv6/netfilter/ip6_tables.c
515
static int check_target(struct ip6t_entry *e, struct net *net, const char *name)
net/ipv6/netfilter/ip6_tables.c
517
struct xt_entry_target *t = ip6t_get_target(e);
net/ipv6/netfilter/ip6_tables.c
521
.entryinfo = e,
net/ipv6/netfilter/ip6_tables.c
524
.hook_mask = e->comefrom,
net/ipv6/netfilter/ip6_tables.c
529
e->ipv6.proto,
net/ipv6/netfilter/ip6_tables.c
530
e->ipv6.invflags & IP6T_INV_PROTO);
net/ipv6/netfilter/ip6_tables.c
534
find_check_entry(struct ip6t_entry *e, struct net *net, const char *name,
net/ipv6/netfilter/ip6_tables.c
545
if (!xt_percpu_counter_alloc(alloc_state, &e->counters))
net/ipv6/netfilter/ip6_tables.c
552
mtpar.entryinfo = &e->ipv6;
net/ipv6/netfilter/ip6_tables.c
553
mtpar.hook_mask = e->comefrom;
net/ipv6/netfilter/ip6_tables.c
555
xt_ematch_foreach(ematch, e) {
net/ipv6/netfilter/ip6_tables.c
562
t = ip6t_get_target(e);
net/ipv6/netfilter/ip6_tables.c
571
ret = check_target(e, net, name);
net/ipv6/netfilter/ip6_tables.c
578
xt_ematch_foreach(ematch, e) {
net/ipv6/netfilter/ip6_tables.c
584
xt_percpu_counter_free(&e->counters);
net/ipv6/netfilter/ip6_tables.c
589
static bool check_underflow(const struct ip6t_entry *e)
net/ipv6/netfilter/ip6_tables.c
594
if (!unconditional(e))
net/ipv6/netfilter/ip6_tables.c
596
t = ip6t_get_target_c(e);
net/ipv6/netfilter/ip6_tables.c
605
check_entry_size_and_hooks(struct ip6t_entry *e,
net/ipv6/netfilter/ip6_tables.c
616
if ((unsigned long)e % __alignof__(struct ip6t_entry) != 0 ||
net/ipv6/netfilter/ip6_tables.c
617
(unsigned char *)e + sizeof(struct ip6t_entry) >= limit ||
net/ipv6/netfilter/ip6_tables.c
618
(unsigned char *)e + e->next_offset > limit)
net/ipv6/netfilter/ip6_tables.c
621
if (e->next_offset
net/ipv6/netfilter/ip6_tables.c
625
if (!ip6_checkentry(&e->ipv6))
net/ipv6/netfilter/ip6_tables.c
628
err = xt_check_entry_offsets(e, e->elems, e->target_offset,
net/ipv6/netfilter/ip6_tables.c
629
e->next_offset);
net/ipv6/netfilter/ip6_tables.c
637
if ((unsigned char *)e - base == hook_entries[h])
net/ipv6/netfilter/ip6_tables.c
639
if ((unsigned char *)e - base == underflows[h]) {
net/ipv6/netfilter/ip6_tables.c
640
if (!check_underflow(e))
net/ipv6/netfilter/ip6_tables.c
648
e->counters = ((struct xt_counters) { 0, 0 });
net/ipv6/netfilter/ip6_tables.c
649
e->comefrom = 0;
net/ipv6/netfilter/ip6_tables.c
653
static void cleanup_entry(struct ip6t_entry *e, struct net *net)
net/ipv6/netfilter/ip6_tables.c
660
xt_ematch_foreach(ematch, e)
net/ipv6/netfilter/ip6_tables.c
662
t = ip6t_get_target(e);
net/ipv6/netfilter/ip6_tables.c
671
xt_percpu_counter_free(&e->counters);
net/ipv6/netfilter/ip6_tables.c
831
const struct ip6t_entry *e;
net/ipv6/netfilter/ip6_tables.c
845
for (off = 0, num = 0; off < total_size; off += e->next_offset, num++){
net/ipv6/netfilter/ip6_tables.c
850
e = loc_cpu_entry + off;
net/ipv6/netfilter/ip6_tables.c
851
if (copy_to_user(userptr + off, e, sizeof(*e))) {
net/ipv6/netfilter/ip6_tables.c
864
i < e->target_offset;
net/ipv6/netfilter/ip6_tables.c
866
m = (void *)e + i;
net/ipv6/netfilter/ip6_tables.c
874
t = ip6t_get_target_c(e);
net/ipv6/netfilter/ip6_tables.c
875
if (xt_target_to_user(t, userptr + off + e->target_offset)) {
net/ipv6/netfilter/ip6_tables.c
905
static int compat_calc_entry(const struct ip6t_entry *e,
net/ipv6/netfilter/ip6_tables.c
915
entry_offset = (void *)e - base;
net/ipv6/netfilter/ip6_tables.c
916
xt_ematch_foreach(ematch, e)
net/ipv6/netfilter/ip6_tables.c
918
t = ip6t_get_target_c(e);
net/ipv6/netfilter/ip6_tables.c
927
(e < (struct ip6t_entry *)(base + info->hook_entry[i])))
net/ipv6/netfilter/ip6_tables.c
930
(e < (struct ip6t_entry *)(base + info->underflow[i])))
net/ipv6/netfilter/ip6t_REJECT.c
81
const struct ip6t_entry *e = par->entryinfo;
net/ipv6/netfilter/ip6t_REJECT.c
88
if (!(e->ipv6.flags & IP6T_F_PROTO) ||
net/ipv6/netfilter/ip6t_REJECT.c
89
e->ipv6.proto != IPPROTO_TCP ||
net/ipv6/netfilter/ip6t_REJECT.c
90
(e->ipv6.invflags & XT_INV_PROTO)) {
net/ipv6/netfilter/ip6t_SYNPROXY.c
69
const struct ip6t_entry *e = par->entryinfo;
net/ipv6/netfilter/ip6t_SYNPROXY.c
72
if (!(e->ipv6.flags & IP6T_F_PROTO) ||
net/ipv6/netfilter/ip6t_SYNPROXY.c
73
e->ipv6.proto != IPPROTO_TCP ||
net/ipv6/netfilter/ip6t_SYNPROXY.c
74
e->ipv6.invflags & XT_INV_PROTO)
net/key/af_key.c
3851
struct proc_dir_entry *e;
net/key/af_key.c
3853
e = proc_create_net("pfkey", 0, net->proc_net, &pfkey_seq_ops,
net/key/af_key.c
3855
if (e == NULL)
net/l2tp/trace.h
12
#define encap_type_name(e) { L2TP_ENCAPTYPE_##e, #e }
net/mpls/af_mpls.c
45
static size_t ipgre_mpls_encap_hlen(struct ip_tunnel_encap *e)
net/mptcp/pm_userspace.c
49
struct mptcp_pm_addr_entry *e;
net/mptcp/pm_userspace.c
57
mptcp_for_each_userspace_pm_addr(msk, e) {
net/mptcp/pm_userspace.c
58
addr_match = mptcp_addresses_equal(&e->addr, &entry->addr, true);
net/mptcp/pm_userspace.c
60
entry->addr.id = e->addr.id;
net/mptcp/pm_userspace.c
61
id_match = (e->addr.id == entry->addr.id);
net/mptcp/pm_userspace.c
64
__set_bit(e->addr.id, id_bitmap);
net/mptcp/pm_userspace.c
71
e = sock_kmemdup(sk, entry, sizeof(*entry), GFP_ATOMIC);
net/mptcp/pm_userspace.c
72
if (!e) {
net/mptcp/pm_userspace.c
77
if (!e->addr.id && needs_id)
net/mptcp/pm_userspace.c
78
e->addr.id = find_next_zero_bit(id_bitmap,
net/mptcp/pm_userspace.c
81
list_add_tail_rcu(&e->list, &msk->pm.userspace_pm_local_addr_list);
net/mptcp/pm_userspace.c
83
ret = e->addr.id;
net/mptcp/syncookies.c
100
e = &join_entries[i];
net/mptcp/syncookies.c
104
if (e->valid == 0) {
net/mptcp/syncookies.c
109
e->valid = 0;
net/mptcp/syncookies.c
111
msk = mptcp_token_get_sock(net, e->token);
net/mptcp/syncookies.c
117
subflow_req->remote_nonce = e->remote_nonce;
net/mptcp/syncookies.c
118
subflow_req->local_nonce = e->local_nonce;
net/mptcp/syncookies.c
119
subflow_req->backup = e->backup;
net/mptcp/syncookies.c
120
subflow_req->remote_id = e->join_id;
net/mptcp/syncookies.c
121
subflow_req->token = e->token;
net/mptcp/syncookies.c
98
struct join_entry *e;
net/netfilter/core.c
44
#define nf_entry_dereference(e) \
net/netfilter/core.c
45
rcu_dereference_protected(e, lockdep_is_held(&nf_hook_mutex))
net/netfilter/core.c
49
struct nf_hook_entries *e;
net/netfilter/core.c
50
size_t alloc = sizeof(*e) +
net/netfilter/core.c
58
e = kvzalloc(alloc, GFP_KERNEL_ACCOUNT);
net/netfilter/core.c
59
if (e)
net/netfilter/core.c
60
e->num_hook_entries = num;
net/netfilter/core.c
61
return e;
net/netfilter/core.c
617
const struct nf_hook_entries *e, unsigned int s)
net/netfilter/core.c
622
for (; s < e->num_hook_entries; s++) {
net/netfilter/core.c
623
verdict = nf_hook_entry_hookfn(&e->hooks[s], skb, state);
net/netfilter/core.c
652
const struct nf_hook_entries *e)
net/netfilter/core.c
660
ret = nf_hook_slow(skb, state, e, 0);
net/netfilter/core.c
72
static void nf_hook_entries_free(struct nf_hook_entries *e)
net/netfilter/core.c
763
__netfilter_net_init(struct nf_hook_entries __rcu **e, int max)
net/netfilter/core.c
768
RCU_INIT_POINTER(e[h], NULL);
net/netfilter/core.c
78
if (!e)
net/netfilter/core.c
81
num = e->num_hook_entries;
net/netfilter/core.c
82
ops = nf_hook_entries_get_hook_ops(e);
net/netfilter/core.c
84
head->allocation = e;
net/netfilter/ipset/ip_set_bitmap_gen.h
120
const struct mtype_adt_elem *e = value;
net/netfilter/ipset/ip_set_bitmap_gen.h
121
void *x = get_ext(set, map, e->id);
net/netfilter/ipset/ip_set_bitmap_gen.h
122
int ret = mtype_do_test(e, map, set->dsize);
net/netfilter/ipset/ip_set_bitmap_gen.h
134
const struct mtype_adt_elem *e = value;
net/netfilter/ipset/ip_set_bitmap_gen.h
135
void *x = get_ext(set, map, e->id);
net/netfilter/ipset/ip_set_bitmap_gen.h
136
int ret = mtype_do_add(e, map, flags, set->dsize);
net/netfilter/ipset/ip_set_bitmap_gen.h
144
set_bit(e->id, map->members);
net/netfilter/ipset/ip_set_bitmap_gen.h
155
mtype_add_timeout(ext_timeout(x, set), e, ext, set, map, ret);
net/netfilter/ipset/ip_set_bitmap_gen.h
168
set_bit(e->id, map->members);
net/netfilter/ipset/ip_set_bitmap_gen.h
179
const struct mtype_adt_elem *e = value;
net/netfilter/ipset/ip_set_bitmap_gen.h
180
void *x = get_ext(set, map, e->id);
net/netfilter/ipset/ip_set_bitmap_gen.h
182
if (mtype_do_del(e, map))
net/netfilter/ipset/ip_set_bitmap_ip.c
116
struct bitmap_ip_adt_elem e = { .id = 0 };
net/netfilter/ipset/ip_set_bitmap_ip.c
124
e.id = ip_to_id(map, ip);
net/netfilter/ipset/ip_set_bitmap_ip.c
126
return adtfn(set, &e, &ext, &opt->ext, opt->cmdflags);
net/netfilter/ipset/ip_set_bitmap_ip.c
136
struct bitmap_ip_adt_elem e = { .id = 0 };
net/netfilter/ipset/ip_set_bitmap_ip.c
158
e.id = ip_to_id(map, ip);
net/netfilter/ipset/ip_set_bitmap_ip.c
159
return adtfn(set, &e, &ext, &ext, flags);
net/netfilter/ipset/ip_set_bitmap_ip.c
182
e.id = ip_to_id(map, ip);
net/netfilter/ipset/ip_set_bitmap_ip.c
183
ret = adtfn(set, &e, &ext, &ext, flags);
net/netfilter/ipset/ip_set_bitmap_ip.c
67
bitmap_ip_do_test(const struct bitmap_ip_adt_elem *e,
net/netfilter/ipset/ip_set_bitmap_ip.c
70
return !!test_bit(e->id, map->members);
net/netfilter/ipset/ip_set_bitmap_ip.c
80
bitmap_ip_do_add(const struct bitmap_ip_adt_elem *e, struct bitmap_ip *map,
net/netfilter/ipset/ip_set_bitmap_ip.c
83
return !!test_bit(e->id, map->members);
net/netfilter/ipset/ip_set_bitmap_ip.c
87
bitmap_ip_do_del(const struct bitmap_ip_adt_elem *e, struct bitmap_ip *map)
net/netfilter/ipset/ip_set_bitmap_ip.c
89
return !test_and_clear_bit(e->id, map->members);
net/netfilter/ipset/ip_set_bitmap_ipmac.c
117
const struct bitmap_ipmac_adt_elem *e,
net/netfilter/ipset/ip_set_bitmap_ipmac.c
134
if (e->add_mac)
net/netfilter/ipset/ip_set_bitmap_ipmac.c
143
bitmap_ipmac_do_add(const struct bitmap_ipmac_adt_elem *e,
net/netfilter/ipset/ip_set_bitmap_ipmac.c
148
elem = get_elem(map->extensions, e->id, dsize);
net/netfilter/ipset/ip_set_bitmap_ipmac.c
149
if (test_bit(e->id, map->members)) {
net/netfilter/ipset/ip_set_bitmap_ipmac.c
151
if (e->add_mac &&
net/netfilter/ipset/ip_set_bitmap_ipmac.c
153
!ether_addr_equal(e->ether, elem->ether)) {
net/netfilter/ipset/ip_set_bitmap_ipmac.c
155
clear_bit(e->id, map->members);
net/netfilter/ipset/ip_set_bitmap_ipmac.c
157
ether_addr_copy(elem->ether, e->ether);
net/netfilter/ipset/ip_set_bitmap_ipmac.c
160
} else if (!e->add_mac)
net/netfilter/ipset/ip_set_bitmap_ipmac.c
164
clear_bit(e->id, map->members);
net/netfilter/ipset/ip_set_bitmap_ipmac.c
166
ether_addr_copy(elem->ether, e->ether);
net/netfilter/ipset/ip_set_bitmap_ipmac.c
169
} else if (e->add_mac) {
net/netfilter/ipset/ip_set_bitmap_ipmac.c
171
ether_addr_copy(elem->ether, e->ether);
net/netfilter/ipset/ip_set_bitmap_ipmac.c
181
bitmap_ipmac_do_del(const struct bitmap_ipmac_adt_elem *e,
net/netfilter/ipset/ip_set_bitmap_ipmac.c
184
return !test_and_clear_bit(e->id, map->members);
net/netfilter/ipset/ip_set_bitmap_ipmac.c
214
struct bitmap_ipmac_adt_elem e = { .id = 0, .add_mac = 1 };
net/netfilter/ipset/ip_set_bitmap_ipmac.c
227
e.id = ip_to_id(map, ip);
net/netfilter/ipset/ip_set_bitmap_ipmac.c
230
ether_addr_copy(e.ether, eth_hdr(skb)->h_source);
net/netfilter/ipset/ip_set_bitmap_ipmac.c
232
ether_addr_copy(e.ether, eth_hdr(skb)->h_dest);
net/netfilter/ipset/ip_set_bitmap_ipmac.c
234
if (is_zero_ether_addr(e.ether))
net/netfilter/ipset/ip_set_bitmap_ipmac.c
237
return adtfn(set, &e, &ext, &opt->ext, opt->cmdflags);
net/netfilter/ipset/ip_set_bitmap_ipmac.c
246
struct bitmap_ipmac_adt_elem e = { .id = 0 };
net/netfilter/ipset/ip_set_bitmap_ipmac.c
268
e.id = ip_to_id(map, ip);
net/netfilter/ipset/ip_set_bitmap_ipmac.c
272
memcpy(e.ether, nla_data(tb[IPSET_ATTR_ETHER]), ETH_ALEN);
net/netfilter/ipset/ip_set_bitmap_ipmac.c
273
e.add_mac = 1;
net/netfilter/ipset/ip_set_bitmap_ipmac.c
275
ret = adtfn(set, &e, &ext, &ext, flags);
net/netfilter/ipset/ip_set_bitmap_ipmac.c
83
bitmap_ipmac_do_test(const struct bitmap_ipmac_adt_elem *e,
net/netfilter/ipset/ip_set_bitmap_ipmac.c
88
if (!test_bit(e->id, map->members))
net/netfilter/ipset/ip_set_bitmap_ipmac.c
90
elem = get_const_elem(map->extensions, e->id, dsize);
net/netfilter/ipset/ip_set_bitmap_ipmac.c
91
if (e->add_mac && elem->filled == MAC_FILLED)
net/netfilter/ipset/ip_set_bitmap_ipmac.c
92
return ether_addr_equal(e->ether, elem->ether);
net/netfilter/ipset/ip_set_bitmap_port.c
133
struct bitmap_port_adt_elem e = { .id = 0 };
net/netfilter/ipset/ip_set_bitmap_port.c
147
e.id = port_to_id(map, port);
net/netfilter/ipset/ip_set_bitmap_port.c
149
return adtfn(set, &e, &ext, &opt->ext, opt->cmdflags);
net/netfilter/ipset/ip_set_bitmap_port.c
158
struct bitmap_port_adt_elem e = { .id = 0 };
net/netfilter/ipset/ip_set_bitmap_port.c
179
e.id = port_to_id(map, port);
net/netfilter/ipset/ip_set_bitmap_port.c
180
return adtfn(set, &e, &ext, &ext, flags);
net/netfilter/ipset/ip_set_bitmap_port.c
198
e.id = port_to_id(map, port);
net/netfilter/ipset/ip_set_bitmap_port.c
199
ret = adtfn(set, &e, &ext, &ext, flags);
net/netfilter/ipset/ip_set_bitmap_port.c
58
bitmap_port_do_test(const struct bitmap_port_adt_elem *e,
net/netfilter/ipset/ip_set_bitmap_port.c
61
return !!test_bit(e->id, map->members);
net/netfilter/ipset/ip_set_bitmap_port.c
71
bitmap_port_do_add(const struct bitmap_port_adt_elem *e,
net/netfilter/ipset/ip_set_bitmap_port.c
74
return !!test_bit(e->id, map->members);
net/netfilter/ipset/ip_set_bitmap_port.c
78
bitmap_port_do_del(const struct bitmap_port_adt_elem *e,
net/netfilter/ipset/ip_set_bitmap_port.c
81
return !test_and_clear_bit(e->id, map->members);
net/netfilter/ipset/ip_set_core.c
565
const void *e, bool active)
net/netfilter/ipset/ip_set_core.c
568
unsigned long *timeout = ext_timeout(e, set);
net/netfilter/ipset/ip_set_core.c
576
ip_set_put_counter(skb, ext_counter(e, set)))
net/netfilter/ipset/ip_set_core.c
579
ip_set_put_comment(skb, ext_comment(e, set)))
net/netfilter/ipset/ip_set_core.c
582
ip_set_put_skbinfo(skb, ext_skbinfo(e, set)))
net/netfilter/ipset/ip_set_hash_gen.h
1358
const struct mtype_elem *e;
net/netfilter/ipset/ip_set_hash_gen.h
1384
e = ahash_data(n, i, set->dsize);
net/netfilter/ipset/ip_set_hash_gen.h
1385
if (SET_ELEM_EXPIRED(set, e))
net/netfilter/ipset/ip_set_hash_gen.h
1388
cb->args[IPSET_CB_ARG0], n, i, e);
net/netfilter/ipset/ip_set_hash_gen.h
1398
if (mtype_data_list(skb, e))
net/netfilter/ipset/ip_set_hash_gen.h
1400
if (ip_set_put_extensions(skb, set, e, true))
net/netfilter/ipset/ip_set_hash_gen.h
860
u32 e;
net/netfilter/ipset/ip_set_hash_gen.h
868
for (e = 0; e < ahash_numof_locks(t->htable_bits); e++)
net/netfilter/ipset/ip_set_hash_gen.h
869
elements += t->hregion[e].elements;
net/netfilter/ipset/ip_set_hash_ip.c
105
struct hash_ip4_elem e = { 0 };
net/netfilter/ipset/ip_set_hash_ip.c
125
e.ip = htonl(ip);
net/netfilter/ipset/ip_set_hash_ip.c
126
if (e.ip == 0)
net/netfilter/ipset/ip_set_hash_ip.c
130
return adtfn(set, &e, &ext, &ext, flags);
net/netfilter/ipset/ip_set_hash_ip.c
155
e.ip = htonl(ip);
net/netfilter/ipset/ip_set_hash_ip.c
157
hash_ip4_data_next(&h->next, &e);
net/netfilter/ipset/ip_set_hash_ip.c
160
ret = adtfn(set, &e, &ext, &ext, flags);
net/netfilter/ipset/ip_set_hash_ip.c
191
hash_ip6_data_list(struct sk_buff *skb, const struct hash_ip6_elem *e)
net/netfilter/ipset/ip_set_hash_ip.c
193
if (nla_put_ipaddr6(skb, IPSET_ATTR_IP, &e->ip.in6))
net/netfilter/ipset/ip_set_hash_ip.c
202
hash_ip6_data_next(struct hash_ip6_elem *next, const struct hash_ip6_elem *e)
net/netfilter/ipset/ip_set_hash_ip.c
222
struct hash_ip6_elem e = { { .all = { 0 } } };
net/netfilter/ipset/ip_set_hash_ip.c
225
ip6addrptr(skb, opt->flags & IPSET_DIM_ONE_SRC, &e.ip.in6);
net/netfilter/ipset/ip_set_hash_ip.c
226
nf_inet_addr_mask_inplace(&e.ip, &h->bitmask);
net/netfilter/ipset/ip_set_hash_ip.c
227
if (ipv6_addr_any(&e.ip.in6))
net/netfilter/ipset/ip_set_hash_ip.c
230
return adtfn(set, &e, &ext, &opt->ext, opt->cmdflags);
net/netfilter/ipset/ip_set_hash_ip.c
239
struct hash_ip6_elem e = { { .all = { 0 } } };
net/netfilter/ipset/ip_set_hash_ip.c
257
ret = ip_set_get_ipaddr6(tb[IPSET_ATTR_IP], &e.ip);
net/netfilter/ipset/ip_set_hash_ip.c
265
nf_inet_addr_mask_inplace(&e.ip, &h->bitmask);
net/netfilter/ipset/ip_set_hash_ip.c
266
if (ipv6_addr_any(&e.ip.in6))
net/netfilter/ipset/ip_set_hash_ip.c
269
ret = adtfn(set, &e, &ext, &ext, flags);
net/netfilter/ipset/ip_set_hash_ip.c
59
hash_ip4_data_list(struct sk_buff *skb, const struct hash_ip4_elem *e)
net/netfilter/ipset/ip_set_hash_ip.c
61
if (nla_put_ipaddr4(skb, IPSET_ATTR_IP, e->ip))
net/netfilter/ipset/ip_set_hash_ip.c
70
hash_ip4_data_next(struct hash_ip4_elem *next, const struct hash_ip4_elem *e)
net/netfilter/ipset/ip_set_hash_ip.c
72
next->ip = e->ip;
net/netfilter/ipset/ip_set_hash_ip.c
86
struct hash_ip4_elem e = { 0 };
net/netfilter/ipset/ip_set_hash_ip.c
95
e.ip = ip;
net/netfilter/ipset/ip_set_hash_ip.c
96
return adtfn(set, &e, &ext, &opt->ext, opt->cmdflags);
net/netfilter/ipset/ip_set_hash_ipmac.c
101
if (is_zero_ether_addr(e.ether))
net/netfilter/ipset/ip_set_hash_ipmac.c
104
ip4addrptr(skb, opt->flags & IPSET_DIM_ONE_SRC, &e.ip);
net/netfilter/ipset/ip_set_hash_ipmac.c
106
return adtfn(set, &e, &ext, &opt->ext, opt->cmdflags);
net/netfilter/ipset/ip_set_hash_ipmac.c
114
struct hash_ipmac4_elem e = { .ip = 0, { .foo[0] = 0, .foo[1] = 0 } };
net/netfilter/ipset/ip_set_hash_ipmac.c
132
ret = ip_set_get_ipaddr4(tb[IPSET_ATTR_IP], &e.ip) ||
net/netfilter/ipset/ip_set_hash_ipmac.c
136
memcpy(e.ether, nla_data(tb[IPSET_ATTR_ETHER]), ETH_ALEN);
net/netfilter/ipset/ip_set_hash_ipmac.c
137
if (is_zero_ether_addr(e.ether))
net/netfilter/ipset/ip_set_hash_ipmac.c
140
return adtfn(set, &e, &ext, &ext, flags);
net/netfilter/ipset/ip_set_hash_ipmac.c
167
hash_ipmac6_data_list(struct sk_buff *skb, const struct hash_ipmac6_elem *e)
net/netfilter/ipset/ip_set_hash_ipmac.c
169
if (nla_put_ipaddr6(skb, IPSET_ATTR_IP, &e->ip.in6) ||
net/netfilter/ipset/ip_set_hash_ipmac.c
170
nla_put(skb, IPSET_ATTR_ETHER, ETH_ALEN, e->ether))
net/netfilter/ipset/ip_set_hash_ipmac.c
180
const struct hash_ipmac6_elem *e)
net/netfilter/ipset/ip_set_hash_ipmac.c
202
struct hash_ipmac6_elem e = {
net/netfilter/ipset/ip_set_hash_ipmac.c
213
ether_addr_copy(e.ether, eth_hdr(skb)->h_source);
net/netfilter/ipset/ip_set_hash_ipmac.c
215
ether_addr_copy(e.ether, eth_hdr(skb)->h_dest);
net/netfilter/ipset/ip_set_hash_ipmac.c
217
if (is_zero_ether_addr(e.ether))
net/netfilter/ipset/ip_set_hash_ipmac.c
220
ip6addrptr(skb, opt->flags & IPSET_DIM_ONE_SRC, &e.ip.in6);
net/netfilter/ipset/ip_set_hash_ipmac.c
222
return adtfn(set, &e, &ext, &opt->ext, opt->cmdflags);
net/netfilter/ipset/ip_set_hash_ipmac.c
230
struct hash_ipmac6_elem e = {
net/netfilter/ipset/ip_set_hash_ipmac.c
251
ret = ip_set_get_ipaddr6(tb[IPSET_ATTR_IP], &e.ip) ||
net/netfilter/ipset/ip_set_hash_ipmac.c
256
memcpy(e.ether, nla_data(tb[IPSET_ATTR_ETHER]), ETH_ALEN);
net/netfilter/ipset/ip_set_hash_ipmac.c
257
if (is_zero_ether_addr(e.ether))
net/netfilter/ipset/ip_set_hash_ipmac.c
260
return adtfn(set, &e, &ext, &ext, flags);
net/netfilter/ipset/ip_set_hash_ipmac.c
59
hash_ipmac4_data_list(struct sk_buff *skb, const struct hash_ipmac4_elem *e)
net/netfilter/ipset/ip_set_hash_ipmac.c
61
if (nla_put_ipaddr4(skb, IPSET_ATTR_IP, e->ip) ||
net/netfilter/ipset/ip_set_hash_ipmac.c
62
nla_put(skb, IPSET_ATTR_ETHER, ETH_ALEN, e->ether))
net/netfilter/ipset/ip_set_hash_ipmac.c
72
const struct hash_ipmac4_elem *e)
net/netfilter/ipset/ip_set_hash_ipmac.c
74
next->ip = e->ip;
net/netfilter/ipset/ip_set_hash_ipmac.c
89
struct hash_ipmac4_elem e = { .ip = 0, { .foo[0] = 0, .foo[1] = 0 } };
net/netfilter/ipset/ip_set_hash_ipmac.c
97
ether_addr_copy(e.ether, eth_hdr(skb)->h_source);
net/netfilter/ipset/ip_set_hash_ipmac.c
99
ether_addr_copy(e.ether, eth_hdr(skb)->h_dest);
net/netfilter/ipset/ip_set_hash_ipmark.c
102
struct hash_ipmark4_elem e = { };
net/netfilter/ipset/ip_set_hash_ipmark.c
114
ret = ip_set_get_ipaddr4(tb[IPSET_ATTR_IP], &e.ip);
net/netfilter/ipset/ip_set_hash_ipmark.c
122
e.mark = ntohl(nla_get_be32(tb[IPSET_ATTR_MARK]));
net/netfilter/ipset/ip_set_hash_ipmark.c
123
e.mark &= h->markmask;
net/netfilter/ipset/ip_set_hash_ipmark.c
124
if (e.mark == 0 && e.ip == 0)
net/netfilter/ipset/ip_set_hash_ipmark.c
129
ret = adtfn(set, &e, &ext, &ext, flags);
net/netfilter/ipset/ip_set_hash_ipmark.c
133
ip_to = ip = ntohl(e.ip);
net/netfilter/ipset/ip_set_hash_ipmark.c
139
if (e.mark == 0 && ip_to == 0)
net/netfilter/ipset/ip_set_hash_ipmark.c
154
e.ip = htonl(ip);
net/netfilter/ipset/ip_set_hash_ipmark.c
156
hash_ipmark4_data_next(&h->next, &e);
net/netfilter/ipset/ip_set_hash_ipmark.c
159
ret = adtfn(set, &e, &ext, &ext, flags);
net/netfilter/ipset/ip_set_hash_ipmark.c
221
struct hash_ipmark6_elem e = { };
net/netfilter/ipset/ip_set_hash_ipmark.c
224
e.mark = skb->mark;
net/netfilter/ipset/ip_set_hash_ipmark.c
225
e.mark &= h->markmask;
net/netfilter/ipset/ip_set_hash_ipmark.c
227
ip6addrptr(skb, opt->flags & IPSET_DIM_ONE_SRC, &e.ip.in6);
net/netfilter/ipset/ip_set_hash_ipmark.c
228
return adtfn(set, &e, &ext, &opt->ext, opt->cmdflags);
net/netfilter/ipset/ip_set_hash_ipmark.c
237
struct hash_ipmark6_elem e = { };
net/netfilter/ipset/ip_set_hash_ipmark.c
256
ret = ip_set_get_ipaddr6(tb[IPSET_ATTR_IP], &e.ip);
net/netfilter/ipset/ip_set_hash_ipmark.c
264
e.mark = ntohl(nla_get_be32(tb[IPSET_ATTR_MARK]));
net/netfilter/ipset/ip_set_hash_ipmark.c
265
e.mark &= h->markmask;
net/netfilter/ipset/ip_set_hash_ipmark.c
268
ret = adtfn(set, &e, &ext, &ext, flags);
net/netfilter/ipset/ip_set_hash_ipmark.c
272
ret = adtfn(set, &e, &ext, &ext, flags);
net/netfilter/ipset/ip_set_hash_ipmark.c
86
struct hash_ipmark4_elem e = { };
net/netfilter/ipset/ip_set_hash_ipmark.c
89
e.mark = skb->mark;
net/netfilter/ipset/ip_set_hash_ipmark.c
90
e.mark &= h->markmask;
net/netfilter/ipset/ip_set_hash_ipmark.c
92
ip4addrptr(skb, opt->flags & IPSET_DIM_ONE_SRC, &e.ip);
net/netfilter/ipset/ip_set_hash_ipmark.c
93
return adtfn(set, &e, &ext, &opt->ext, opt->cmdflags);
net/netfilter/ipset/ip_set_hash_ipport.c
101
&e.port, &e.proto))
net/netfilter/ipset/ip_set_hash_ipport.c
104
ip4addrptr(skb, opt->flags & IPSET_DIM_ONE_SRC, &e.ip);
net/netfilter/ipset/ip_set_hash_ipport.c
105
e.ip &= h->bitmask.ip;
net/netfilter/ipset/ip_set_hash_ipport.c
106
if (e.ip == 0)
net/netfilter/ipset/ip_set_hash_ipport.c
108
return adtfn(set, &e, &ext, &opt->ext, opt->cmdflags);
net/netfilter/ipset/ip_set_hash_ipport.c
117
struct hash_ipport4_elem e = { .ip = 0 };
net/netfilter/ipset/ip_set_hash_ipport.c
131
ret = ip_set_get_ipaddr4(tb[IPSET_ATTR_IP], &e.ip);
net/netfilter/ipset/ip_set_hash_ipport.c
139
e.ip &= h->bitmask.ip;
net/netfilter/ipset/ip_set_hash_ipport.c
140
if (e.ip == 0)
net/netfilter/ipset/ip_set_hash_ipport.c
143
e.port = nla_get_be16(tb[IPSET_ATTR_PORT]);
net/netfilter/ipset/ip_set_hash_ipport.c
146
e.proto = nla_get_u8(tb[IPSET_ATTR_PROTO]);
net/netfilter/ipset/ip_set_hash_ipport.c
147
with_ports = ip_set_proto_with_ports(e.proto);
net/netfilter/ipset/ip_set_hash_ipport.c
149
if (e.proto == 0)
net/netfilter/ipset/ip_set_hash_ipport.c
155
if (!(with_ports || e.proto == IPPROTO_ICMP))
net/netfilter/ipset/ip_set_hash_ipport.c
156
e.port = 0;
net/netfilter/ipset/ip_set_hash_ipport.c
161
ret = adtfn(set, &e, &ext, &ext, flags);
net/netfilter/ipset/ip_set_hash_ipport.c
165
ip_to = ip = ntohl(e.ip);
net/netfilter/ipset/ip_set_hash_ipport.c
180
port_to = port = ntohs(e.port);
net/netfilter/ipset/ip_set_hash_ipport.c
193
e.ip = htonl(ip);
net/netfilter/ipset/ip_set_hash_ipport.c
194
e.port = htons(p);
net/netfilter/ipset/ip_set_hash_ipport.c
196
hash_ipport4_data_next(&h->next, &e);
net/netfilter/ipset/ip_set_hash_ipport.c
199
ret = adtfn(set, &e, &ext, &ext, flags);
net/netfilter/ipset/ip_set_hash_ipport.c
266
struct hash_ipport6_elem e = { .ip = { .all = { 0 } } };
net/netfilter/ipset/ip_set_hash_ipport.c
271
&e.port, &e.proto))
net/netfilter/ipset/ip_set_hash_ipport.c
274
ip6addrptr(skb, opt->flags & IPSET_DIM_ONE_SRC, &e.ip.in6);
net/netfilter/ipset/ip_set_hash_ipport.c
275
nf_inet_addr_mask_inplace(&e.ip, &h->bitmask);
net/netfilter/ipset/ip_set_hash_ipport.c
276
if (ipv6_addr_any(&e.ip.in6))
net/netfilter/ipset/ip_set_hash_ipport.c
279
return adtfn(set, &e, &ext, &opt->ext, opt->cmdflags);
net/netfilter/ipset/ip_set_hash_ipport.c
288
struct hash_ipport6_elem e = { .ip = { .all = { 0 } } };
net/netfilter/ipset/ip_set_hash_ipport.c
310
ret = ip_set_get_ipaddr6(tb[IPSET_ATTR_IP], &e.ip);
net/netfilter/ipset/ip_set_hash_ipport.c
318
nf_inet_addr_mask_inplace(&e.ip, &h->bitmask);
net/netfilter/ipset/ip_set_hash_ipport.c
319
if (ipv6_addr_any(&e.ip.in6))
net/netfilter/ipset/ip_set_hash_ipport.c
322
e.port = nla_get_be16(tb[IPSET_ATTR_PORT]);
net/netfilter/ipset/ip_set_hash_ipport.c
325
e.proto = nla_get_u8(tb[IPSET_ATTR_PROTO]);
net/netfilter/ipset/ip_set_hash_ipport.c
326
with_ports = ip_set_proto_with_ports(e.proto);
net/netfilter/ipset/ip_set_hash_ipport.c
328
if (e.proto == 0)
net/netfilter/ipset/ip_set_hash_ipport.c
334
if (!(with_ports || e.proto == IPPROTO_ICMPV6))
net/netfilter/ipset/ip_set_hash_ipport.c
335
e.port = 0;
net/netfilter/ipset/ip_set_hash_ipport.c
338
ret = adtfn(set, &e, &ext, &ext, flags);
net/netfilter/ipset/ip_set_hash_ipport.c
342
port = ntohs(e.port);
net/netfilter/ipset/ip_set_hash_ipport.c
350
e.port = htons(port);
net/netfilter/ipset/ip_set_hash_ipport.c
351
ret = adtfn(set, &e, &ext, &ext, flags);
net/netfilter/ipset/ip_set_hash_ipport.c
96
struct hash_ipport4_elem e = { .ip = 0 };
net/netfilter/ipset/ip_set_hash_ipportip.c
102
ip4addrptr(skb, opt->flags & IPSET_DIM_ONE_SRC, &e.ip);
net/netfilter/ipset/ip_set_hash_ipportip.c
103
ip4addrptr(skb, opt->flags & IPSET_DIM_THREE_SRC, &e.ip2);
net/netfilter/ipset/ip_set_hash_ipportip.c
104
return adtfn(set, &e, &ext, &opt->ext, opt->cmdflags);
net/netfilter/ipset/ip_set_hash_ipportip.c
113
struct hash_ipportip4_elem e = { .ip = 0 };
net/netfilter/ipset/ip_set_hash_ipportip.c
127
ret = ip_set_get_ipaddr4(tb[IPSET_ATTR_IP], &e.ip);
net/netfilter/ipset/ip_set_hash_ipportip.c
135
ret = ip_set_get_ipaddr4(tb[IPSET_ATTR_IP2], &e.ip2);
net/netfilter/ipset/ip_set_hash_ipportip.c
139
e.port = nla_get_be16(tb[IPSET_ATTR_PORT]);
net/netfilter/ipset/ip_set_hash_ipportip.c
142
e.proto = nla_get_u8(tb[IPSET_ATTR_PROTO]);
net/netfilter/ipset/ip_set_hash_ipportip.c
143
with_ports = ip_set_proto_with_ports(e.proto);
net/netfilter/ipset/ip_set_hash_ipportip.c
145
if (e.proto == 0)
net/netfilter/ipset/ip_set_hash_ipportip.c
151
if (!(with_ports || e.proto == IPPROTO_ICMP))
net/netfilter/ipset/ip_set_hash_ipportip.c
152
e.port = 0;
net/netfilter/ipset/ip_set_hash_ipportip.c
157
ret = adtfn(set, &e, &ext, &ext, flags);
net/netfilter/ipset/ip_set_hash_ipportip.c
161
ip_to = ip = ntohl(e.ip);
net/netfilter/ipset/ip_set_hash_ipportip.c
176
port_to = port = ntohs(e.port);
net/netfilter/ipset/ip_set_hash_ipportip.c
189
e.ip = htonl(ip);
net/netfilter/ipset/ip_set_hash_ipportip.c
190
e.port = htons(p);
net/netfilter/ipset/ip_set_hash_ipportip.c
192
hash_ipportip4_data_next(&h->next, &e);
net/netfilter/ipset/ip_set_hash_ipportip.c
195
ret = adtfn(set, &e, &ext, &ext, flags);
net/netfilter/ipset/ip_set_hash_ipportip.c
265
struct hash_ipportip6_elem e = { .ip = { .all = { 0 } } };
net/netfilter/ipset/ip_set_hash_ipportip.c
269
&e.port, &e.proto))
net/netfilter/ipset/ip_set_hash_ipportip.c
272
ip6addrptr(skb, opt->flags & IPSET_DIM_ONE_SRC, &e.ip.in6);
net/netfilter/ipset/ip_set_hash_ipportip.c
273
ip6addrptr(skb, opt->flags & IPSET_DIM_THREE_SRC, &e.ip2.in6);
net/netfilter/ipset/ip_set_hash_ipportip.c
274
return adtfn(set, &e, &ext, &opt->ext, opt->cmdflags);
net/netfilter/ipset/ip_set_hash_ipportip.c
283
struct hash_ipportip6_elem e = { .ip = { .all = { 0 } } };
net/netfilter/ipset/ip_set_hash_ipportip.c
305
ret = ip_set_get_ipaddr6(tb[IPSET_ATTR_IP], &e.ip);
net/netfilter/ipset/ip_set_hash_ipportip.c
313
ret = ip_set_get_ipaddr6(tb[IPSET_ATTR_IP2], &e.ip2);
net/netfilter/ipset/ip_set_hash_ipportip.c
317
e.port = nla_get_be16(tb[IPSET_ATTR_PORT]);
net/netfilter/ipset/ip_set_hash_ipportip.c
320
e.proto = nla_get_u8(tb[IPSET_ATTR_PROTO]);
net/netfilter/ipset/ip_set_hash_ipportip.c
321
with_ports = ip_set_proto_with_ports(e.proto);
net/netfilter/ipset/ip_set_hash_ipportip.c
323
if (e.proto == 0)
net/netfilter/ipset/ip_set_hash_ipportip.c
329
if (!(with_ports || e.proto == IPPROTO_ICMPV6))
net/netfilter/ipset/ip_set_hash_ipportip.c
330
e.port = 0;
net/netfilter/ipset/ip_set_hash_ipportip.c
333
ret = adtfn(set, &e, &ext, &ext, flags);
net/netfilter/ipset/ip_set_hash_ipportip.c
337
port = ntohs(e.port);
net/netfilter/ipset/ip_set_hash_ipportip.c
345
e.port = htons(port);
net/netfilter/ipset/ip_set_hash_ipportip.c
346
ret = adtfn(set, &e, &ext, &ext, flags);
net/netfilter/ipset/ip_set_hash_ipportip.c
95
struct hash_ipportip4_elem e = { .ip = 0 };
net/netfilter/ipset/ip_set_hash_ipportip.c
99
&e.port, &e.proto))
net/netfilter/ipset/ip_set_hash_ipportnet.c
140
struct hash_ipportnet4_elem e = {
net/netfilter/ipset/ip_set_hash_ipportnet.c
146
e.cidr = HOST_MASK - 1;
net/netfilter/ipset/ip_set_hash_ipportnet.c
149
&e.port, &e.proto))
net/netfilter/ipset/ip_set_hash_ipportnet.c
152
ip4addrptr(skb, opt->flags & IPSET_DIM_ONE_SRC, &e.ip);
net/netfilter/ipset/ip_set_hash_ipportnet.c
153
ip4addrptr(skb, opt->flags & IPSET_DIM_THREE_SRC, &e.ip2);
net/netfilter/ipset/ip_set_hash_ipportnet.c
154
e.ip2 &= ip_set_netmask(e.cidr + 1);
net/netfilter/ipset/ip_set_hash_ipportnet.c
156
return adtfn(set, &e, &ext, &opt->ext, opt->cmdflags);
net/netfilter/ipset/ip_set_hash_ipportnet.c
165
struct hash_ipportnet4_elem e = { .cidr = HOST_MASK - 1 };
net/netfilter/ipset/ip_set_hash_ipportnet.c
198
e.cidr = cidr - 1;
net/netfilter/ipset/ip_set_hash_ipportnet.c
201
e.port = nla_get_be16(tb[IPSET_ATTR_PORT]);
net/netfilter/ipset/ip_set_hash_ipportnet.c
204
e.proto = nla_get_u8(tb[IPSET_ATTR_PROTO]);
net/netfilter/ipset/ip_set_hash_ipportnet.c
205
with_ports = ip_set_proto_with_ports(e.proto);
net/netfilter/ipset/ip_set_hash_ipportnet.c
207
if (e.proto == 0)
net/netfilter/ipset/ip_set_hash_ipportnet.c
213
if (!(with_ports || e.proto == IPPROTO_ICMP))
net/netfilter/ipset/ip_set_hash_ipportnet.c
214
e.port = 0;
net/netfilter/ipset/ip_set_hash_ipportnet.c
227
e.ip = htonl(ip);
net/netfilter/ipset/ip_set_hash_ipportnet.c
228
e.ip2 = htonl(ip2_from & ip_set_hostmask(e.cidr + 1));
net/netfilter/ipset/ip_set_hash_ipportnet.c
229
ret = adtfn(set, &e, &ext, &ext, flags);
net/netfilter/ipset/ip_set_hash_ipportnet.c
249
port_to = port = ntohs(e.port);
net/netfilter/ipset/ip_set_hash_ipportnet.c
266
ip_set_mask_from_to(ip2_from, ip2_to, e.cidr + 1);
net/netfilter/ipset/ip_set_hash_ipportnet.c
278
e.ip = htonl(ip);
net/netfilter/ipset/ip_set_hash_ipportnet.c
280
e.port = htons(p);
net/netfilter/ipset/ip_set_hash_ipportnet.c
283
e.ip2 = htonl(ip2);
net/netfilter/ipset/ip_set_hash_ipportnet.c
285
e.cidr = cidr - 1;
net/netfilter/ipset/ip_set_hash_ipportnet.c
288
&e);
net/netfilter/ipset/ip_set_hash_ipportnet.c
291
ret = adtfn(set, &e, &ext, &ext, flags);
net/netfilter/ipset/ip_set_hash_ipportnet.c
397
struct hash_ipportnet6_elem e = {
net/netfilter/ipset/ip_set_hash_ipportnet.c
403
e.cidr = HOST_MASK - 1;
net/netfilter/ipset/ip_set_hash_ipportnet.c
406
&e.port, &e.proto))
net/netfilter/ipset/ip_set_hash_ipportnet.c
409
ip6addrptr(skb, opt->flags & IPSET_DIM_ONE_SRC, &e.ip.in6);
net/netfilter/ipset/ip_set_hash_ipportnet.c
410
ip6addrptr(skb, opt->flags & IPSET_DIM_THREE_SRC, &e.ip2.in6);
net/netfilter/ipset/ip_set_hash_ipportnet.c
411
ip6_netmask(&e.ip2, e.cidr + 1);
net/netfilter/ipset/ip_set_hash_ipportnet.c
413
return adtfn(set, &e, &ext, &opt->ext, opt->cmdflags);
net/netfilter/ipset/ip_set_hash_ipportnet.c
422
struct hash_ipportnet6_elem e = { .cidr = HOST_MASK - 1 };
net/netfilter/ipset/ip_set_hash_ipportnet.c
446
ret = ip_set_get_ipaddr6(tb[IPSET_ATTR_IP], &e.ip);
net/netfilter/ipset/ip_set_hash_ipportnet.c
454
ret = ip_set_get_ipaddr6(tb[IPSET_ATTR_IP2], &e.ip2);
net/netfilter/ipset/ip_set_hash_ipportnet.c
462
e.cidr = cidr - 1;
net/netfilter/ipset/ip_set_hash_ipportnet.c
465
ip6_netmask(&e.ip2, e.cidr + 1);
net/netfilter/ipset/ip_set_hash_ipportnet.c
467
e.port = nla_get_be16(tb[IPSET_ATTR_PORT]);
net/netfilter/ipset/ip_set_hash_ipportnet.c
470
e.proto = nla_get_u8(tb[IPSET_ATTR_PROTO]);
net/netfilter/ipset/ip_set_hash_ipportnet.c
471
with_ports = ip_set_proto_with_ports(e.proto);
net/netfilter/ipset/ip_set_hash_ipportnet.c
473
if (e.proto == 0)
net/netfilter/ipset/ip_set_hash_ipportnet.c
479
if (!(with_ports || e.proto == IPPROTO_ICMPV6))
net/netfilter/ipset/ip_set_hash_ipportnet.c
480
e.port = 0;
net/netfilter/ipset/ip_set_hash_ipportnet.c
490
ret = adtfn(set, &e, &ext, &ext, flags);
net/netfilter/ipset/ip_set_hash_ipportnet.c
495
port = ntohs(e.port);
net/netfilter/ipset/ip_set_hash_ipportnet.c
503
e.port = htons(port);
net/netfilter/ipset/ip_set_hash_ipportnet.c
504
ret = adtfn(set, &e, &ext, &ext, flags);
net/netfilter/ipset/ip_set_hash_mac.c
113
ether_addr_copy(e.ether, nla_data(tb[IPSET_ATTR_ETHER]));
net/netfilter/ipset/ip_set_hash_mac.c
114
if (is_zero_ether_addr(e.ether))
net/netfilter/ipset/ip_set_hash_mac.c
117
return adtfn(set, &e, &ext, &ext, flags);
net/netfilter/ipset/ip_set_hash_mac.c
49
hash_mac4_data_list(struct sk_buff *skb, const struct hash_mac4_elem *e)
net/netfilter/ipset/ip_set_hash_mac.c
51
if (nla_put(skb, IPSET_ATTR_ETHER, ETH_ALEN, e->ether))
net/netfilter/ipset/ip_set_hash_mac.c
61
const struct hash_mac4_elem *e)
net/netfilter/ipset/ip_set_hash_mac.c
77
struct hash_mac4_elem e = { { .foo[0] = 0, .foo[1] = 0 } };
net/netfilter/ipset/ip_set_hash_mac.c
85
ether_addr_copy(e.ether, eth_hdr(skb)->h_source);
net/netfilter/ipset/ip_set_hash_mac.c
87
ether_addr_copy(e.ether, eth_hdr(skb)->h_dest);
net/netfilter/ipset/ip_set_hash_mac.c
89
if (is_zero_ether_addr(e.ether))
net/netfilter/ipset/ip_set_hash_mac.c
91
return adtfn(set, &e, &ext, &opt->ext, opt->cmdflags);
net/netfilter/ipset/ip_set_hash_mac.c
99
struct hash_mac4_elem e = { { .foo[0] = 0, .foo[1] = 0 } };
net/netfilter/ipset/ip_set_hash_net.c
119
struct hash_net4_elem e = {
net/netfilter/ipset/ip_set_hash_net.c
124
if (e.cidr == 0)
net/netfilter/ipset/ip_set_hash_net.c
127
e.cidr = HOST_MASK;
net/netfilter/ipset/ip_set_hash_net.c
129
ip4addrptr(skb, opt->flags & IPSET_DIM_ONE_SRC, &e.ip);
net/netfilter/ipset/ip_set_hash_net.c
130
e.ip &= ip_set_netmask(e.cidr);
net/netfilter/ipset/ip_set_hash_net.c
132
return adtfn(set, &e, &ext, &opt->ext, opt->cmdflags);
net/netfilter/ipset/ip_set_hash_net.c
141
struct hash_net4_elem e = { .cidr = HOST_MASK };
net/netfilter/ipset/ip_set_hash_net.c
162
e.cidr = nla_get_u8(tb[IPSET_ATTR_CIDR]);
net/netfilter/ipset/ip_set_hash_net.c
163
if (!e.cidr || e.cidr > HOST_MASK)
net/netfilter/ipset/ip_set_hash_net.c
175
e.ip = htonl(ip & ip_set_hostmask(e.cidr));
net/netfilter/ipset/ip_set_hash_net.c
176
ret = adtfn(set, &e, &ext, &ext, flags);
net/netfilter/ipset/ip_set_hash_net.c
196
e.ip = htonl(ip);
net/netfilter/ipset/ip_set_hash_net.c
198
hash_net4_data_next(&h->next, &e);
net/netfilter/ipset/ip_set_hash_net.c
201
ip = ip_set_range_to_cidr(ip, ip_to, &e.cidr);
net/netfilter/ipset/ip_set_hash_net.c
202
ret = adtfn(set, &e, &ext, &ext, flags);
net/netfilter/ipset/ip_set_hash_net.c
293
struct hash_net6_elem e = {
net/netfilter/ipset/ip_set_hash_net.c
298
if (e.cidr == 0)
net/netfilter/ipset/ip_set_hash_net.c
301
e.cidr = HOST_MASK;
net/netfilter/ipset/ip_set_hash_net.c
303
ip6addrptr(skb, opt->flags & IPSET_DIM_ONE_SRC, &e.ip.in6);
net/netfilter/ipset/ip_set_hash_net.c
304
ip6_netmask(&e.ip, e.cidr);
net/netfilter/ipset/ip_set_hash_net.c
306
return adtfn(set, &e, &ext, &opt->ext, opt->cmdflags);
net/netfilter/ipset/ip_set_hash_net.c
314
struct hash_net6_elem e = { .cidr = HOST_MASK };
net/netfilter/ipset/ip_set_hash_net.c
327
ret = ip_set_get_ipaddr6(tb[IPSET_ATTR_IP], &e.ip);
net/netfilter/ipset/ip_set_hash_net.c
336
e.cidr = nla_get_u8(tb[IPSET_ATTR_CIDR]);
net/netfilter/ipset/ip_set_hash_net.c
337
if (!e.cidr || e.cidr > HOST_MASK)
net/netfilter/ipset/ip_set_hash_net.c
341
ip6_netmask(&e.ip, e.cidr);
net/netfilter/ipset/ip_set_hash_net.c
350
ret = adtfn(set, &e, &ext, &ext, flags);
net/netfilter/ipset/ip_set_hash_netiface.c
163
struct hash_netiface4_elem e = {
net/netfilter/ipset/ip_set_hash_netiface.c
170
e.cidr = HOST_MASK;
net/netfilter/ipset/ip_set_hash_netiface.c
172
ip4addrptr(skb, opt->flags & IPSET_DIM_ONE_SRC, &e.ip);
net/netfilter/ipset/ip_set_hash_netiface.c
173
e.ip &= ip_set_netmask(e.cidr);
net/netfilter/ipset/ip_set_hash_netiface.c
185
STRSCPY(e.iface, eiface);
net/netfilter/ipset/ip_set_hash_netiface.c
186
e.physdev = 1;
net/netfilter/ipset/ip_set_hash_netiface.c
189
STRSCPY(e.iface, SRCDIR ? IFACE(in) : IFACE(out));
net/netfilter/ipset/ip_set_hash_netiface.c
192
if (strlen(e.iface) == 0)
net/netfilter/ipset/ip_set_hash_netiface.c
194
return adtfn(set, &e, &ext, &opt->ext, opt->cmdflags);
net/netfilter/ipset/ip_set_hash_netiface.c
203
struct hash_netiface4_elem e = { .cidr = HOST_MASK, .elem = 1 };
net/netfilter/ipset/ip_set_hash_netiface.c
225
e.cidr = nla_get_u8(tb[IPSET_ATTR_CIDR]);
net/netfilter/ipset/ip_set_hash_netiface.c
226
if (e.cidr > HOST_MASK)
net/netfilter/ipset/ip_set_hash_netiface.c
229
nla_strscpy(e.iface, tb[IPSET_ATTR_IFACE], IFNAMSIZ);
net/netfilter/ipset/ip_set_hash_netiface.c
235
e.physdev = 1;
net/netfilter/ipset/ip_set_hash_netiface.c
239
e.wildcard = 1;
net/netfilter/ipset/ip_set_hash_netiface.c
242
e.ip = htonl(ip & ip_set_hostmask(e.cidr));
net/netfilter/ipset/ip_set_hash_netiface.c
243
ret = adtfn(set, &e, &ext, &ext, flags);
net/netfilter/ipset/ip_set_hash_netiface.c
257
ip_set_mask_from_to(ip, ip_to, e.cidr);
net/netfilter/ipset/ip_set_hash_netiface.c
264
e.ip = htonl(ip);
net/netfilter/ipset/ip_set_hash_netiface.c
266
hash_netiface4_data_next(&h->next, &e);
net/netfilter/ipset/ip_set_hash_netiface.c
269
ip = ip_set_range_to_cidr(ip, ip_to, &e.cidr);
net/netfilter/ipset/ip_set_hash_netiface.c
270
ret = adtfn(set, &e, &ext, &ext, flags);
net/netfilter/ipset/ip_set_hash_netiface.c
384
struct hash_netiface6_elem e = {
net/netfilter/ipset/ip_set_hash_netiface.c
391
e.cidr = HOST_MASK;
net/netfilter/ipset/ip_set_hash_netiface.c
393
ip6addrptr(skb, opt->flags & IPSET_DIM_ONE_SRC, &e.ip.in6);
net/netfilter/ipset/ip_set_hash_netiface.c
394
ip6_netmask(&e.ip, e.cidr);
net/netfilter/ipset/ip_set_hash_netiface.c
403
STRSCPY(e.iface, eiface);
net/netfilter/ipset/ip_set_hash_netiface.c
404
e.physdev = 1;
net/netfilter/ipset/ip_set_hash_netiface.c
407
STRSCPY(e.iface, SRCDIR ? IFACE(in) : IFACE(out));
net/netfilter/ipset/ip_set_hash_netiface.c
410
if (strlen(e.iface) == 0)
net/netfilter/ipset/ip_set_hash_netiface.c
413
return adtfn(set, &e, &ext, &opt->ext, opt->cmdflags);
net/netfilter/ipset/ip_set_hash_netiface.c
421
struct hash_netiface6_elem e = { .cidr = HOST_MASK, .elem = 1 };
net/netfilter/ipset/ip_set_hash_netiface.c
435
ret = ip_set_get_ipaddr6(tb[IPSET_ATTR_IP], &e.ip);
net/netfilter/ipset/ip_set_hash_netiface.c
444
e.cidr = nla_get_u8(tb[IPSET_ATTR_CIDR]);
net/netfilter/ipset/ip_set_hash_netiface.c
445
if (e.cidr > HOST_MASK)
net/netfilter/ipset/ip_set_hash_netiface.c
449
ip6_netmask(&e.ip, e.cidr);
net/netfilter/ipset/ip_set_hash_netiface.c
451
nla_strscpy(e.iface, tb[IPSET_ATTR_IFACE], IFNAMSIZ);
net/netfilter/ipset/ip_set_hash_netiface.c
457
e.physdev = 1;
net/netfilter/ipset/ip_set_hash_netiface.c
461
e.wildcard = 1;
net/netfilter/ipset/ip_set_hash_netiface.c
464
ret = adtfn(set, &e, &ext, &ext, flags);
net/netfilter/ipset/ip_set_hash_netnet.c
136
hash_netnet4_init(struct hash_netnet4_elem *e)
net/netfilter/ipset/ip_set_hash_netnet.c
138
e->cidr[0] = HOST_MASK;
net/netfilter/ipset/ip_set_hash_netnet.c
139
e->cidr[1] = HOST_MASK;
net/netfilter/ipset/ip_set_hash_netnet.c
149
struct hash_netnet4_elem e = { };
net/netfilter/ipset/ip_set_hash_netnet.c
152
e.cidr[0] = INIT_CIDR(h->nets[0].cidr[0], HOST_MASK);
net/netfilter/ipset/ip_set_hash_netnet.c
153
e.cidr[1] = INIT_CIDR(h->nets[0].cidr[1], HOST_MASK);
net/netfilter/ipset/ip_set_hash_netnet.c
155
e.ccmp = (HOST_MASK << (sizeof(e.cidr[0]) * 8)) | HOST_MASK;
net/netfilter/ipset/ip_set_hash_netnet.c
157
ip4addrptr(skb, opt->flags & IPSET_DIM_ONE_SRC, &e.ip[0]);
net/netfilter/ipset/ip_set_hash_netnet.c
158
ip4addrptr(skb, opt->flags & IPSET_DIM_TWO_SRC, &e.ip[1]);
net/netfilter/ipset/ip_set_hash_netnet.c
159
e.ip[0] &= (ip_set_netmask(e.cidr[0]) & h->bitmask.ip);
net/netfilter/ipset/ip_set_hash_netnet.c
160
e.ip[1] &= (ip_set_netmask(e.cidr[1]) & h->bitmask.ip);
net/netfilter/ipset/ip_set_hash_netnet.c
162
return adtfn(set, &e, &ext, &opt->ext, opt->cmdflags);
net/netfilter/ipset/ip_set_hash_netnet.c
171
struct hash_netnet4_elem e = { };
net/netfilter/ipset/ip_set_hash_netnet.c
180
hash_netnet4_init(&e);
net/netfilter/ipset/ip_set_hash_netnet.c
198
e.cidr[0] = nla_get_u8(tb[IPSET_ATTR_CIDR]);
net/netfilter/ipset/ip_set_hash_netnet.c
199
if (!e.cidr[0] || e.cidr[0] > HOST_MASK)
net/netfilter/ipset/ip_set_hash_netnet.c
204
e.cidr[1] = nla_get_u8(tb[IPSET_ATTR_CIDR2]);
net/netfilter/ipset/ip_set_hash_netnet.c
205
if (!e.cidr[1] || e.cidr[1] > HOST_MASK)
net/netfilter/ipset/ip_set_hash_netnet.c
218
e.ip[0] = htonl(ip & ntohl(h->bitmask.ip) & ip_set_hostmask(e.cidr[0]));
net/netfilter/ipset/ip_set_hash_netnet.c
219
e.ip[1] = htonl(ip2_from & ntohl(h->bitmask.ip) & ip_set_hostmask(e.cidr[1]));
net/netfilter/ipset/ip_set_hash_netnet.c
220
ret = adtfn(set, &e, &ext, &ext, flags);
net/netfilter/ipset/ip_set_hash_netnet.c
235
ip_set_mask_from_to(ip, ip_to, e.cidr[0]);
net/netfilter/ipset/ip_set_hash_netnet.c
248
ip_set_mask_from_to(ip2_from, ip2_to, e.cidr[1]);
net/netfilter/ipset/ip_set_hash_netnet.c
259
e.ip[0] = htonl(ip);
net/netfilter/ipset/ip_set_hash_netnet.c
260
ip = ip_set_range_to_cidr(ip, ip_to, &e.cidr[0]);
net/netfilter/ipset/ip_set_hash_netnet.c
263
e.ip[1] = htonl(ip2);
net/netfilter/ipset/ip_set_hash_netnet.c
265
hash_netnet4_data_next(&h->next, &e);
net/netfilter/ipset/ip_set_hash_netnet.c
268
ip2 = ip_set_range_to_cidr(ip2, ip2_to, &e.cidr[1]);
net/netfilter/ipset/ip_set_hash_netnet.c
269
ret = adtfn(set, &e, &ext, &ext, flags);
net/netfilter/ipset/ip_set_hash_netnet.c
375
hash_netnet6_init(struct hash_netnet6_elem *e)
net/netfilter/ipset/ip_set_hash_netnet.c
377
e->cidr[0] = HOST_MASK;
net/netfilter/ipset/ip_set_hash_netnet.c
378
e->cidr[1] = HOST_MASK;
net/netfilter/ipset/ip_set_hash_netnet.c
388
struct hash_netnet6_elem e = { };
net/netfilter/ipset/ip_set_hash_netnet.c
391
e.cidr[0] = INIT_CIDR(h->nets[0].cidr[0], HOST_MASK);
net/netfilter/ipset/ip_set_hash_netnet.c
392
e.cidr[1] = INIT_CIDR(h->nets[0].cidr[1], HOST_MASK);
net/netfilter/ipset/ip_set_hash_netnet.c
394
e.ccmp = (HOST_MASK << (sizeof(u8) * 8)) | HOST_MASK;
net/netfilter/ipset/ip_set_hash_netnet.c
396
ip6addrptr(skb, opt->flags & IPSET_DIM_ONE_SRC, &e.ip[0].in6);
net/netfilter/ipset/ip_set_hash_netnet.c
397
ip6addrptr(skb, opt->flags & IPSET_DIM_TWO_SRC, &e.ip[1].in6);
net/netfilter/ipset/ip_set_hash_netnet.c
398
ip6_netmask(&e.ip[0], e.cidr[0]);
net/netfilter/ipset/ip_set_hash_netnet.c
399
ip6_netmask(&e.ip[1], e.cidr[1]);
net/netfilter/ipset/ip_set_hash_netnet.c
401
nf_inet_addr_mask_inplace(&e.ip[0], &h->bitmask);
net/netfilter/ipset/ip_set_hash_netnet.c
402
nf_inet_addr_mask_inplace(&e.ip[1], &h->bitmask);
net/netfilter/ipset/ip_set_hash_netnet.c
403
if (e.cidr[0] == HOST_MASK && ipv6_addr_any(&e.ip[0].in6))
net/netfilter/ipset/ip_set_hash_netnet.c
406
return adtfn(set, &e, &ext, &opt->ext, opt->cmdflags);
net/netfilter/ipset/ip_set_hash_netnet.c
414
struct hash_netnet6_elem e = { };
net/netfilter/ipset/ip_set_hash_netnet.c
422
hash_netnet6_init(&e);
net/netfilter/ipset/ip_set_hash_netnet.c
429
ret = ip_set_get_ipaddr6(tb[IPSET_ATTR_IP], &e.ip[0]);
net/netfilter/ipset/ip_set_hash_netnet.c
433
ret = ip_set_get_ipaddr6(tb[IPSET_ATTR_IP2], &e.ip[1]);
net/netfilter/ipset/ip_set_hash_netnet.c
442
e.cidr[0] = nla_get_u8(tb[IPSET_ATTR_CIDR]);
net/netfilter/ipset/ip_set_hash_netnet.c
443
if (!e.cidr[0] || e.cidr[0] > HOST_MASK)
net/netfilter/ipset/ip_set_hash_netnet.c
448
e.cidr[1] = nla_get_u8(tb[IPSET_ATTR_CIDR2]);
net/netfilter/ipset/ip_set_hash_netnet.c
449
if (!e.cidr[1] || e.cidr[1] > HOST_MASK)
net/netfilter/ipset/ip_set_hash_netnet.c
453
ip6_netmask(&e.ip[0], e.cidr[0]);
net/netfilter/ipset/ip_set_hash_netnet.c
454
ip6_netmask(&e.ip[1], e.cidr[1]);
net/netfilter/ipset/ip_set_hash_netnet.c
456
nf_inet_addr_mask_inplace(&e.ip[0], &h->bitmask);
net/netfilter/ipset/ip_set_hash_netnet.c
457
nf_inet_addr_mask_inplace(&e.ip[1], &h->bitmask);
net/netfilter/ipset/ip_set_hash_netnet.c
458
if (e.cidr[0] == HOST_MASK && ipv6_addr_any(&e.ip[0].in6))
net/netfilter/ipset/ip_set_hash_netnet.c
468
ret = adtfn(set, &e, &ext, &ext, flags);
net/netfilter/ipset/ip_set_hash_netport.c
135
struct hash_netport4_elem e = {
net/netfilter/ipset/ip_set_hash_netport.c
141
e.cidr = HOST_MASK - 1;
net/netfilter/ipset/ip_set_hash_netport.c
144
&e.port, &e.proto))
net/netfilter/ipset/ip_set_hash_netport.c
147
ip4addrptr(skb, opt->flags & IPSET_DIM_ONE_SRC, &e.ip);
net/netfilter/ipset/ip_set_hash_netport.c
148
e.ip &= ip_set_netmask(e.cidr + 1);
net/netfilter/ipset/ip_set_hash_netport.c
150
return adtfn(set, &e, &ext, &opt->ext, opt->cmdflags);
net/netfilter/ipset/ip_set_hash_netport.c
159
struct hash_netport4_elem e = { .cidr = HOST_MASK - 1 };
net/netfilter/ipset/ip_set_hash_netport.c
187
e.cidr = cidr - 1;
net/netfilter/ipset/ip_set_hash_netport.c
190
e.port = nla_get_be16(tb[IPSET_ATTR_PORT]);
net/netfilter/ipset/ip_set_hash_netport.c
193
e.proto = nla_get_u8(tb[IPSET_ATTR_PROTO]);
net/netfilter/ipset/ip_set_hash_netport.c
194
with_ports = ip_set_proto_with_ports(e.proto);
net/netfilter/ipset/ip_set_hash_netport.c
196
if (e.proto == 0)
net/netfilter/ipset/ip_set_hash_netport.c
202
if (!(with_ports || e.proto == IPPROTO_ICMP))
net/netfilter/ipset/ip_set_hash_netport.c
203
e.port = 0;
net/netfilter/ipset/ip_set_hash_netport.c
215
e.ip = htonl(ip & ip_set_hostmask(e.cidr + 1));
net/netfilter/ipset/ip_set_hash_netport.c
216
ret = adtfn(set, &e, &ext, &ext, flags);
net/netfilter/ipset/ip_set_hash_netport.c
221
port = port_to = ntohs(e.port);
net/netfilter/ipset/ip_set_hash_netport.c
236
ip_set_mask_from_to(ip, ip_to, e.cidr + 1);
net/netfilter/ipset/ip_set_hash_netport.c
246
e.ip = htonl(ip);
net/netfilter/ipset/ip_set_hash_netport.c
248
e.cidr = cidr - 1;
net/netfilter/ipset/ip_set_hash_netport.c
250
e.port = htons(p);
net/netfilter/ipset/ip_set_hash_netport.c
252
hash_netport4_data_next(&h->next, &e);
net/netfilter/ipset/ip_set_hash_netport.c
255
ret = adtfn(set, &e, &ext, &ext, flags);
net/netfilter/ipset/ip_set_hash_netport.c
355
struct hash_netport6_elem e = {
net/netfilter/ipset/ip_set_hash_netport.c
361
e.cidr = HOST_MASK - 1;
net/netfilter/ipset/ip_set_hash_netport.c
364
&e.port, &e.proto))
net/netfilter/ipset/ip_set_hash_netport.c
367
ip6addrptr(skb, opt->flags & IPSET_DIM_ONE_SRC, &e.ip.in6);
net/netfilter/ipset/ip_set_hash_netport.c
368
ip6_netmask(&e.ip, e.cidr + 1);
net/netfilter/ipset/ip_set_hash_netport.c
370
return adtfn(set, &e, &ext, &opt->ext, opt->cmdflags);
net/netfilter/ipset/ip_set_hash_netport.c
379
struct hash_netport6_elem e = { .cidr = HOST_MASK - 1 };
net/netfilter/ipset/ip_set_hash_netport.c
397
ret = ip_set_get_ipaddr6(tb[IPSET_ATTR_IP], &e.ip);
net/netfilter/ipset/ip_set_hash_netport.c
409
e.cidr = cidr - 1;
net/netfilter/ipset/ip_set_hash_netport.c
411
ip6_netmask(&e.ip, e.cidr + 1);
net/netfilter/ipset/ip_set_hash_netport.c
413
e.port = nla_get_be16(tb[IPSET_ATTR_PORT]);
net/netfilter/ipset/ip_set_hash_netport.c
416
e.proto = nla_get_u8(tb[IPSET_ATTR_PROTO]);
net/netfilter/ipset/ip_set_hash_netport.c
417
with_ports = ip_set_proto_with_ports(e.proto);
net/netfilter/ipset/ip_set_hash_netport.c
419
if (e.proto == 0)
net/netfilter/ipset/ip_set_hash_netport.c
425
if (!(with_ports || e.proto == IPPROTO_ICMPV6))
net/netfilter/ipset/ip_set_hash_netport.c
426
e.port = 0;
net/netfilter/ipset/ip_set_hash_netport.c
436
ret = adtfn(set, &e, &ext, &ext, flags);
net/netfilter/ipset/ip_set_hash_netport.c
441
port = ntohs(e.port);
net/netfilter/ipset/ip_set_hash_netport.c
449
e.port = htons(port);
net/netfilter/ipset/ip_set_hash_netport.c
450
ret = adtfn(set, &e, &ext, &ext, flags);
net/netfilter/ipset/ip_set_hash_netportnet.c
144
hash_netportnet4_init(struct hash_netportnet4_elem *e)
net/netfilter/ipset/ip_set_hash_netportnet.c
146
e->cidr[0] = HOST_MASK;
net/netfilter/ipset/ip_set_hash_netportnet.c
147
e->cidr[1] = HOST_MASK;
net/netfilter/ipset/ip_set_hash_netportnet.c
157
struct hash_netportnet4_elem e = { };
net/netfilter/ipset/ip_set_hash_netportnet.c
160
e.cidr[0] = INIT_CIDR(h->nets[0].cidr[0], HOST_MASK);
net/netfilter/ipset/ip_set_hash_netportnet.c
161
e.cidr[1] = INIT_CIDR(h->nets[0].cidr[1], HOST_MASK);
net/netfilter/ipset/ip_set_hash_netportnet.c
163
e.ccmp = (HOST_MASK << (sizeof(e.cidr[0]) * 8)) | HOST_MASK;
net/netfilter/ipset/ip_set_hash_netportnet.c
166
&e.port, &e.proto))
net/netfilter/ipset/ip_set_hash_netportnet.c
169
ip4addrptr(skb, opt->flags & IPSET_DIM_ONE_SRC, &e.ip[0]);
net/netfilter/ipset/ip_set_hash_netportnet.c
170
ip4addrptr(skb, opt->flags & IPSET_DIM_THREE_SRC, &e.ip[1]);
net/netfilter/ipset/ip_set_hash_netportnet.c
171
e.ip[0] &= ip_set_netmask(e.cidr[0]);
net/netfilter/ipset/ip_set_hash_netportnet.c
172
e.ip[1] &= ip_set_netmask(e.cidr[1]);
net/netfilter/ipset/ip_set_hash_netportnet.c
174
return adtfn(set, &e, &ext, &opt->ext, opt->cmdflags);
net/netfilter/ipset/ip_set_hash_netportnet.c
193
struct hash_netportnet4_elem e = { };
net/netfilter/ipset/ip_set_hash_netportnet.c
203
hash_netportnet4_init(&e);
net/netfilter/ipset/ip_set_hash_netportnet.c
223
e.cidr[0] = nla_get_u8(tb[IPSET_ATTR_CIDR]);
net/netfilter/ipset/ip_set_hash_netportnet.c
224
if (e.cidr[0] > HOST_MASK)
net/netfilter/ipset/ip_set_hash_netportnet.c
229
e.cidr[1] = nla_get_u8(tb[IPSET_ATTR_CIDR2]);
net/netfilter/ipset/ip_set_hash_netportnet.c
230
if (e.cidr[1] > HOST_MASK)
net/netfilter/ipset/ip_set_hash_netportnet.c
234
e.port = nla_get_be16(tb[IPSET_ATTR_PORT]);
net/netfilter/ipset/ip_set_hash_netportnet.c
237
e.proto = nla_get_u8(tb[IPSET_ATTR_PROTO]);
net/netfilter/ipset/ip_set_hash_netportnet.c
238
with_ports = ip_set_proto_with_ports(e.proto);
net/netfilter/ipset/ip_set_hash_netportnet.c
240
if (e.proto == 0)
net/netfilter/ipset/ip_set_hash_netportnet.c
246
if (!(with_ports || e.proto == IPPROTO_ICMP))
net/netfilter/ipset/ip_set_hash_netportnet.c
247
e.port = 0;
net/netfilter/ipset/ip_set_hash_netportnet.c
259
e.ip[0] = htonl(ip & ip_set_hostmask(e.cidr[0]));
net/netfilter/ipset/ip_set_hash_netportnet.c
260
e.ip[1] = htonl(ip2_from & ip_set_hostmask(e.cidr[1]));
net/netfilter/ipset/ip_set_hash_netportnet.c
261
ret = adtfn(set, &e, &ext, &ext, flags);
net/netfilter/ipset/ip_set_hash_netportnet.c
276
ip_set_mask_from_to(ip, ip_to, e.cidr[0]);
net/netfilter/ipset/ip_set_hash_netportnet.c
279
port_to = port = ntohs(e.port);
net/netfilter/ipset/ip_set_hash_netportnet.c
296
ip_set_mask_from_to(ip2_from, ip2_to, e.cidr[1]);
net/netfilter/ipset/ip_set_hash_netportnet.c
309
e.ip[0] = htonl(ip);
net/netfilter/ipset/ip_set_hash_netportnet.c
310
ip = hash_netportnet4_range_to_cidr(ip, ip_to, &e.cidr[0]);
net/netfilter/ipset/ip_set_hash_netportnet.c
312
e.port = htons(p);
net/netfilter/ipset/ip_set_hash_netportnet.c
315
e.ip[1] = htonl(ip2);
net/netfilter/ipset/ip_set_hash_netportnet.c
318
&e);
net/netfilter/ipset/ip_set_hash_netportnet.c
322
ip2_to, &e.cidr[1]);
net/netfilter/ipset/ip_set_hash_netportnet.c
323
ret = adtfn(set, &e, &ext, &ext, flags);
net/netfilter/ipset/ip_set_hash_netportnet.c
439
hash_netportnet6_init(struct hash_netportnet6_elem *e)
net/netfilter/ipset/ip_set_hash_netportnet.c
441
e->cidr[0] = HOST_MASK;
net/netfilter/ipset/ip_set_hash_netportnet.c
442
e->cidr[1] = HOST_MASK;
net/netfilter/ipset/ip_set_hash_netportnet.c
452
struct hash_netportnet6_elem e = { };
net/netfilter/ipset/ip_set_hash_netportnet.c
455
e.cidr[0] = INIT_CIDR(h->nets[0].cidr[0], HOST_MASK);
net/netfilter/ipset/ip_set_hash_netportnet.c
456
e.cidr[1] = INIT_CIDR(h->nets[0].cidr[1], HOST_MASK);
net/netfilter/ipset/ip_set_hash_netportnet.c
458
e.ccmp = (HOST_MASK << (sizeof(u8) * 8)) | HOST_MASK;
net/netfilter/ipset/ip_set_hash_netportnet.c
461
&e.port, &e.proto))
net/netfilter/ipset/ip_set_hash_netportnet.c
464
ip6addrptr(skb, opt->flags & IPSET_DIM_ONE_SRC, &e.ip[0].in6);
net/netfilter/ipset/ip_set_hash_netportnet.c
465
ip6addrptr(skb, opt->flags & IPSET_DIM_THREE_SRC, &e.ip[1].in6);
net/netfilter/ipset/ip_set_hash_netportnet.c
466
ip6_netmask(&e.ip[0], e.cidr[0]);
net/netfilter/ipset/ip_set_hash_netportnet.c
467
ip6_netmask(&e.ip[1], e.cidr[1]);
net/netfilter/ipset/ip_set_hash_netportnet.c
469
return adtfn(set, &e, &ext, &opt->ext, opt->cmdflags);
net/netfilter/ipset/ip_set_hash_netportnet.c
478
struct hash_netportnet6_elem e = { };
net/netfilter/ipset/ip_set_hash_netportnet.c
487
hash_netportnet6_init(&e);
net/netfilter/ipset/ip_set_hash_netportnet.c
496
ret = ip_set_get_ipaddr6(tb[IPSET_ATTR_IP], &e.ip[0]);
net/netfilter/ipset/ip_set_hash_netportnet.c
500
ret = ip_set_get_ipaddr6(tb[IPSET_ATTR_IP2], &e.ip[1]);
net/netfilter/ipset/ip_set_hash_netportnet.c
509
e.cidr[0] = nla_get_u8(tb[IPSET_ATTR_CIDR]);
net/netfilter/ipset/ip_set_hash_netportnet.c
510
if (e.cidr[0] > HOST_MASK)
net/netfilter/ipset/ip_set_hash_netportnet.c
515
e.cidr[1] = nla_get_u8(tb[IPSET_ATTR_CIDR2]);
net/netfilter/ipset/ip_set_hash_netportnet.c
516
if (e.cidr[1] > HOST_MASK)
net/netfilter/ipset/ip_set_hash_netportnet.c
520
ip6_netmask(&e.ip[0], e.cidr[0]);
net/netfilter/ipset/ip_set_hash_netportnet.c
521
ip6_netmask(&e.ip[1], e.cidr[1]);
net/netfilter/ipset/ip_set_hash_netportnet.c
523
e.port = nla_get_be16(tb[IPSET_ATTR_PORT]);
net/netfilter/ipset/ip_set_hash_netportnet.c
526
e.proto = nla_get_u8(tb[IPSET_ATTR_PROTO]);
net/netfilter/ipset/ip_set_hash_netportnet.c
527
with_ports = ip_set_proto_with_ports(e.proto);
net/netfilter/ipset/ip_set_hash_netportnet.c
529
if (e.proto == 0)
net/netfilter/ipset/ip_set_hash_netportnet.c
535
if (!(with_ports || e.proto == IPPROTO_ICMPV6))
net/netfilter/ipset/ip_set_hash_netportnet.c
536
e.port = 0;
net/netfilter/ipset/ip_set_hash_netportnet.c
546
ret = adtfn(set, &e, &ext, &ext, flags);
net/netfilter/ipset/ip_set_hash_netportnet.c
551
port = ntohs(e.port);
net/netfilter/ipset/ip_set_hash_netportnet.c
559
e.port = htons(port);
net/netfilter/ipset/ip_set_hash_netportnet.c
560
ret = adtfn(set, &e, &ext, &ext, flags);
net/netfilter/ipset/ip_set_list_set.c
102
list_for_each_entry_rcu(e, &map->members, list) {
net/netfilter/ipset/ip_set_list_set.c
104
ip_set_timeout_expired(ext_timeout(e, set)))
net/netfilter/ipset/ip_set_list_set.c
106
ret = ip_set_del(e->id, skb, par, opt);
net/netfilter/ipset/ip_set_list_set.c
145
struct set_elem *e = container_of(rcu, struct set_elem, rcu);
net/netfilter/ipset/ip_set_list_set.c
146
struct ip_set *set = e->set;
net/netfilter/ipset/ip_set_list_set.c
148
ip_set_ext_destroy(set, e);
net/netfilter/ipset/ip_set_list_set.c
149
kfree(e);
net/netfilter/ipset/ip_set_list_set.c
153
list_set_del(struct ip_set *set, struct set_elem *e)
net/netfilter/ipset/ip_set_list_set.c
158
list_del_rcu(&e->list);
net/netfilter/ipset/ip_set_list_set.c
159
ip_set_put_byindex(map->net, e->id);
net/netfilter/ipset/ip_set_list_set.c
160
call_rcu(&e->rcu, __list_set_del_rcu);
net/netfilter/ipset/ip_set_list_set.c
164
list_set_replace(struct ip_set *set, struct set_elem *e, struct set_elem *old)
net/netfilter/ipset/ip_set_list_set.c
168
list_replace_rcu(&old->list, &e->list);
net/netfilter/ipset/ip_set_list_set.c
177
struct set_elem *e, *n;
net/netfilter/ipset/ip_set_list_set.c
179
list_for_each_entry_safe(e, n, &map->members, list)
net/netfilter/ipset/ip_set_list_set.c
180
if (ip_set_timeout_expired(ext_timeout(e, set)))
net/netfilter/ipset/ip_set_list_set.c
181
list_set_del(set, e);
net/netfilter/ipset/ip_set_list_set.c
190
struct set_elem *e, *next, *prev = NULL;
net/netfilter/ipset/ip_set_list_set.c
194
list_for_each_entry_rcu(e, &map->members, list) {
net/netfilter/ipset/ip_set_list_set.c
196
ip_set_timeout_expired(ext_timeout(e, set)))
net/netfilter/ipset/ip_set_list_set.c
198
else if (e->id != d->id) {
net/netfilter/ipset/ip_set_list_set.c
199
prev = e;
net/netfilter/ipset/ip_set_list_set.c
207
next = list_next_entry(e, list);
net/netfilter/ipset/ip_set_list_set.c
208
ret = !list_is_last(&e->list, &map->members) &&
net/netfilter/ipset/ip_set_list_set.c
222
struct set_elem *e)
net/netfilter/ipset/ip_set_list_set.c
225
ip_set_init_counter(ext_counter(e, set), ext);
net/netfilter/ipset/ip_set_list_set.c
227
ip_set_init_comment(set, ext_comment(e, set), ext);
net/netfilter/ipset/ip_set_list_set.c
229
ip_set_init_skbinfo(ext_skbinfo(e, set), ext);
net/netfilter/ipset/ip_set_list_set.c
232
ip_set_timeout_set(ext_timeout(e, set), ext->timeout);
net/netfilter/ipset/ip_set_list_set.c
241
struct set_elem *e, *n, *prev, *next;
net/netfilter/ipset/ip_set_list_set.c
246
list_for_each_entry_rcu(e, &map->members, list) {
net/netfilter/ipset/ip_set_list_set.c
248
ip_set_timeout_expired(ext_timeout(e, set)))
net/netfilter/ipset/ip_set_list_set.c
250
else if (d->id == e->id)
net/netfilter/ipset/ip_set_list_set.c
251
n = e;
net/netfilter/ipset/ip_set_list_set.c
252
else if (d->before == 0 || e->id != d->refid)
net/netfilter/ipset/ip_set_list_set.c
255
next = e;
net/netfilter/ipset/ip_set_list_set.c
257
prev = e;
net/netfilter/ipset/ip_set_list_set.c
297
e = kzalloc(set->dsize, GFP_ATOMIC);
net/netfilter/ipset/ip_set_list_set.c
298
if (!e)
net/netfilter/ipset/ip_set_list_set.c
300
e->id = d->id;
net/netfilter/ipset/ip_set_list_set.c
301
e->set = set;
net/netfilter/ipset/ip_set_list_set.c
302
INIT_LIST_HEAD(&e->list);
net/netfilter/ipset/ip_set_list_set.c
303
list_set_init_extensions(set, ext, e);
net/netfilter/ipset/ip_set_list_set.c
305
list_set_replace(set, e, n);
net/netfilter/ipset/ip_set_list_set.c
307
list_add_tail_rcu(&e->list, &next->list);
net/netfilter/ipset/ip_set_list_set.c
309
list_add_rcu(&e->list, &prev->list);
net/netfilter/ipset/ip_set_list_set.c
311
list_add_tail_rcu(&e->list, &map->members);
net/netfilter/ipset/ip_set_list_set.c
323
struct set_elem *e, *n, *next, *prev = NULL;
net/netfilter/ipset/ip_set_list_set.c
325
list_for_each_entry_safe(e, n, &map->members, list) {
net/netfilter/ipset/ip_set_list_set.c
327
ip_set_timeout_expired(ext_timeout(e, set)))
net/netfilter/ipset/ip_set_list_set.c
329
else if (e->id != d->id) {
net/netfilter/ipset/ip_set_list_set.c
330
prev = e;
net/netfilter/ipset/ip_set_list_set.c
335
next = list_next_entry(e, list);
net/netfilter/ipset/ip_set_list_set.c
336
if (list_is_last(&e->list, &map->members) ||
net/netfilter/ipset/ip_set_list_set.c
343
list_set_del(set, e);
net/netfilter/ipset/ip_set_list_set.c
355
struct set_adt_elem e = { .refid = IPSET_INVALID_ID };
net/netfilter/ipset/ip_set_list_set.c
370
e.id = ip_set_get_byname(map->net, tb[IPSET_ATTR_NAME], &s);
net/netfilter/ipset/ip_set_list_set.c
371
if (e.id == IPSET_INVALID_ID)
net/netfilter/ipset/ip_set_list_set.c
382
e.before = f & IPSET_FLAG_BEFORE;
net/netfilter/ipset/ip_set_list_set.c
385
if (e.before && !tb[IPSET_ATTR_NAMEREF]) {
net/netfilter/ipset/ip_set_list_set.c
391
e.refid = ip_set_get_byname(map->net,
net/netfilter/ipset/ip_set_list_set.c
394
if (e.refid == IPSET_INVALID_ID) {
net/netfilter/ipset/ip_set_list_set.c
398
if (!e.before)
net/netfilter/ipset/ip_set_list_set.c
399
e.before = -1;
net/netfilter/ipset/ip_set_list_set.c
404
ret = adtfn(set, &e, &ext, &ext, flags);
net/netfilter/ipset/ip_set_list_set.c
407
if (e.refid != IPSET_INVALID_ID)
net/netfilter/ipset/ip_set_list_set.c
408
ip_set_put_byindex(map->net, e.refid);
net/netfilter/ipset/ip_set_list_set.c
410
ip_set_put_byindex(map->net, e.id);
net/netfilter/ipset/ip_set_list_set.c
419
struct set_elem *e, *n;
net/netfilter/ipset/ip_set_list_set.c
421
list_for_each_entry_safe(e, n, &map->members, list)
net/netfilter/ipset/ip_set_list_set.c
422
list_set_del(set, e);
net/netfilter/ipset/ip_set_list_set.c
442
struct set_elem *e;
net/netfilter/ipset/ip_set_list_set.c
446
list_for_each_entry_rcu(e, &map->members, list)
net/netfilter/ipset/ip_set_list_set.c
485
struct set_elem *e;
net/netfilter/ipset/ip_set_list_set.c
493
list_for_each_entry_rcu(e, &map->members, list) {
net/netfilter/ipset/ip_set_list_set.c
496
ip_set_timeout_expired(ext_timeout(e, set)))) {
net/netfilter/ipset/ip_set_list_set.c
503
ip_set_name_byindex(map->net, e->id, name);
net/netfilter/ipset/ip_set_list_set.c
506
if (ip_set_put_extensions(skb, set, e, true))
net/netfilter/ipset/ip_set_list_set.c
55
struct set_elem *e;
net/netfilter/ipset/ip_set_list_set.c
63
list_for_each_entry_rcu(e, &map->members, list) {
net/netfilter/ipset/ip_set_list_set.c
64
ret = ip_set_test(e->id, skb, par, opt);
net/netfilter/ipset/ip_set_list_set.c
67
if (ip_set_match_extensions(set, ext, mext, flags, e))
net/netfilter/ipset/ip_set_list_set.c
79
struct set_elem *e;
net/netfilter/ipset/ip_set_list_set.c
82
list_for_each_entry_rcu(e, &map->members, list) {
net/netfilter/ipset/ip_set_list_set.c
84
ip_set_timeout_expired(ext_timeout(e, set)))
net/netfilter/ipset/ip_set_list_set.c
86
ret = ip_set_add(e->id, skb, par, opt);
net/netfilter/ipset/ip_set_list_set.c
99
struct set_elem *e;
net/netfilter/ipvs/ip_vs_app.c
545
struct list_head *e;
net/netfilter/ipvs/ip_vs_app.c
556
if ((e = inc->a_list.next) != &app->incs_list)
net/netfilter/ipvs/ip_vs_app.c
557
return list_entry(e, struct ip_vs_app, a_list);
net/netfilter/ipvs/ip_vs_app.c
560
for (e = app->a_list.next; e != &ipvs->app_list; e = e->next) {
net/netfilter/ipvs/ip_vs_app.c
561
app = list_entry(e, struct ip_vs_app, a_list);
net/netfilter/ipvs/ip_vs_conn.c
1099
struct hlist_node *e;
net/netfilter/ipvs/ip_vs_conn.c
1106
e = rcu_dereference(hlist_next_rcu(&cp->c_list));
net/netfilter/ipvs/ip_vs_conn.c
1107
if (e) {
net/netfilter/ipvs/ip_vs_conn.c
1109
return hlist_entry(e, struct ip_vs_conn, c_list);
net/netfilter/ipvs/ip_vs_ctl.c
2342
struct hlist_node *e;
net/netfilter/ipvs/ip_vs_ctl.c
2355
e = rcu_dereference(hlist_next_rcu(&svc->s_list));
net/netfilter/ipvs/ip_vs_ctl.c
2356
if (e)
net/netfilter/ipvs/ip_vs_ctl.c
2357
return hlist_entry(e, struct ip_vs_service, s_list);
net/netfilter/ipvs/ip_vs_ctl.c
2373
e = rcu_dereference(hlist_next_rcu(&svc->f_list));
net/netfilter/ipvs/ip_vs_ctl.c
2374
if (e)
net/netfilter/ipvs/ip_vs_ctl.c
2375
return hlist_entry(e, struct ip_vs_service, f_list);
net/netfilter/ipvs/ip_vs_est.c
122
rate = (s->kstats.conns - e->last_conns) << 9;
net/netfilter/ipvs/ip_vs_est.c
123
e->last_conns = s->kstats.conns;
net/netfilter/ipvs/ip_vs_est.c
124
e->cps += ((s64)rate - (s64)e->cps) >> 2;
net/netfilter/ipvs/ip_vs_est.c
126
rate = (s->kstats.inpkts - e->last_inpkts) << 9;
net/netfilter/ipvs/ip_vs_est.c
127
e->last_inpkts = s->kstats.inpkts;
net/netfilter/ipvs/ip_vs_est.c
128
e->inpps += ((s64)rate - (s64)e->inpps) >> 2;
net/netfilter/ipvs/ip_vs_est.c
130
rate = (s->kstats.outpkts - e->last_outpkts) << 9;
net/netfilter/ipvs/ip_vs_est.c
131
e->last_outpkts = s->kstats.outpkts;
net/netfilter/ipvs/ip_vs_est.c
132
e->outpps += ((s64)rate - (s64)e->outpps) >> 2;
net/netfilter/ipvs/ip_vs_est.c
135
rate = (s->kstats.inbytes - e->last_inbytes) << 4;
net/netfilter/ipvs/ip_vs_est.c
136
e->last_inbytes = s->kstats.inbytes;
net/netfilter/ipvs/ip_vs_est.c
137
e->inbps += ((s64)rate - (s64)e->inbps) >> 2;
net/netfilter/ipvs/ip_vs_est.c
139
rate = (s->kstats.outbytes - e->last_outbytes) << 4;
net/netfilter/ipvs/ip_vs_est.c
140
e->last_outbytes = s->kstats.outbytes;
net/netfilter/ipvs/ip_vs_est.c
141
e->outbps += ((s64)rate - (s64)e->outbps) >> 2;
net/netfilter/ipvs/ip_vs_est.c
80
struct ip_vs_estimator *e;
net/netfilter/ipvs/ip_vs_est.c
85
hlist_for_each_entry_rcu(e, chain, list) {
net/netfilter/ipvs/ip_vs_est.c
920
struct ip_vs_estimator *e = &stats->est;
net/netfilter/ipvs/ip_vs_est.c
922
dst->cps = (e->cps + 0x1FF) >> 10;
net/netfilter/ipvs/ip_vs_est.c
923
dst->inpps = (e->inpps + 0x1FF) >> 10;
net/netfilter/ipvs/ip_vs_est.c
924
dst->outpps = (e->outpps + 0x1FF) >> 10;
net/netfilter/ipvs/ip_vs_est.c
925
dst->inbps = (e->inbps + 0xF) >> 5;
net/netfilter/ipvs/ip_vs_est.c
926
dst->outbps = (e->outbps + 0xF) >> 5;
net/netfilter/ipvs/ip_vs_est.c
95
s = container_of(e, struct ip_vs_stats, est);
net/netfilter/ipvs/ip_vs_lblcr.c
101
struct ip_vs_dest_set_elem *e;
net/netfilter/ipvs/ip_vs_lblcr.c
104
list_for_each_entry(e, &set->list, list) {
net/netfilter/ipvs/ip_vs_lblcr.c
105
if (e->dest == dest)
net/netfilter/ipvs/ip_vs_lblcr.c
110
e = kmalloc_obj(*e, GFP_ATOMIC);
net/netfilter/ipvs/ip_vs_lblcr.c
111
if (e == NULL)
net/netfilter/ipvs/ip_vs_lblcr.c
115
e->dest = dest;
net/netfilter/ipvs/ip_vs_lblcr.c
117
list_add_rcu(&e->list, &set->list);
net/netfilter/ipvs/ip_vs_lblcr.c
125
struct ip_vs_dest_set_elem *e;
net/netfilter/ipvs/ip_vs_lblcr.c
127
e = container_of(head, struct ip_vs_dest_set_elem, rcu_head);
net/netfilter/ipvs/ip_vs_lblcr.c
128
ip_vs_dest_put_and_free(e->dest);
net/netfilter/ipvs/ip_vs_lblcr.c
129
kfree(e);
net/netfilter/ipvs/ip_vs_lblcr.c
135
struct ip_vs_dest_set_elem *e;
net/netfilter/ipvs/ip_vs_lblcr.c
137
list_for_each_entry(e, &set->list, list) {
net/netfilter/ipvs/ip_vs_lblcr.c
138
if (e->dest == dest) {
net/netfilter/ipvs/ip_vs_lblcr.c
142
list_del_rcu(&e->list);
net/netfilter/ipvs/ip_vs_lblcr.c
143
call_rcu(&e->rcu_head, ip_vs_lblcr_elem_rcu_free);
net/netfilter/ipvs/ip_vs_lblcr.c
151
struct ip_vs_dest_set_elem *e, *ep;
net/netfilter/ipvs/ip_vs_lblcr.c
153
list_for_each_entry_safe(e, ep, &set->list, list) {
net/netfilter/ipvs/ip_vs_lblcr.c
154
list_del_rcu(&e->list);
net/netfilter/ipvs/ip_vs_lblcr.c
155
call_rcu(&e->rcu_head, ip_vs_lblcr_elem_rcu_free);
net/netfilter/ipvs/ip_vs_lblcr.c
162
struct ip_vs_dest_set_elem *e;
net/netfilter/ipvs/ip_vs_lblcr.c
167
list_for_each_entry_rcu(e, &set->list, list) {
net/netfilter/ipvs/ip_vs_lblcr.c
168
least = e->dest;
net/netfilter/ipvs/ip_vs_lblcr.c
182
list_for_each_entry_continue_rcu(e, &set->list, list) {
net/netfilter/ipvs/ip_vs_lblcr.c
183
dest = e->dest;
net/netfilter/ipvs/ip_vs_lblcr.c
211
struct ip_vs_dest_set_elem *e;
net/netfilter/ipvs/ip_vs_lblcr.c
219
list_for_each_entry(e, &set->list, list) {
net/netfilter/ipvs/ip_vs_lblcr.c
220
most = e->dest;
net/netfilter/ipvs/ip_vs_lblcr.c
230
list_for_each_entry_continue(e, &set->list, list) {
net/netfilter/ipvs/ip_vs_lblcr.c
231
dest = e->dest;
net/netfilter/nf_conntrack_ecache.c
127
static int __nf_conntrack_eventmask_report(struct nf_conntrack_ecache *e,
net/netfilter/nf_conntrack_ecache.c
137
if (!((events | missed) & e->ctmask))
net/netfilter/nf_conntrack_ecache.c
155
old = READ_ONCE(e->missed);
net/netfilter/nf_conntrack_ecache.c
160
} while (cmpxchg(&e->missed, old, want) != old);
net/netfilter/nf_conntrack_ecache.c
165
static void nf_ct_ecache_tstamp_refresh(struct nf_conntrack_ecache *e)
net/netfilter/nf_conntrack_ecache.c
168
if (local64_read(&e->timestamp))
net/netfilter/nf_conntrack_ecache.c
169
local64_set(&e->timestamp, ktime_get_real_ns());
net/netfilter/nf_conntrack_ecache.c
176
struct nf_conntrack_ecache *e;
net/netfilter/nf_conntrack_ecache.c
184
e = nf_ct_ecache_find(ct);
net/netfilter/nf_conntrack_ecache.c
185
if (!e)
net/netfilter/nf_conntrack_ecache.c
191
item.portid = e->portid ? e->portid : portid;
net/netfilter/nf_conntrack_ecache.c
195
missed = e->portid ? 0 : e->missed;
net/netfilter/nf_conntrack_ecache.c
197
nf_ct_ecache_tstamp_refresh(e);
net/netfilter/nf_conntrack_ecache.c
199
ret = __nf_conntrack_eventmask_report(e, events, missed, &item);
net/netfilter/nf_conntrack_ecache.c
204
if (e->portid == 0 && portid != 0)
net/netfilter/nf_conntrack_ecache.c
205
e->portid = portid;
net/netfilter/nf_conntrack_ecache.c
216
struct nf_conntrack_ecache *e;
net/netfilter/nf_conntrack_ecache.c
223
e = nf_ct_ecache_find(ct);
net/netfilter/nf_conntrack_ecache.c
224
if (e == NULL)
net/netfilter/nf_conntrack_ecache.c
227
events = xchg(&e->cache, 0);
net/netfilter/nf_conntrack_ecache.c
237
__nf_conntrack_eventmask_report(e, events, e->missed, &item);
net/netfilter/nf_conntrack_ecache.c
248
struct nf_conntrack_ecache *e;
net/netfilter/nf_conntrack_ecache.c
257
e = nf_ct_ecache_find(exp->master);
net/netfilter/nf_conntrack_ecache.c
258
if (!e)
net/netfilter/nf_conntrack_ecache.c
261
if (e->expmask & (1 << event)) {
net/netfilter/nf_conntrack_ecache.c
312
static void nf_ct_ecache_tstamp_new(const struct nf_conn *ct, struct nf_conntrack_ecache *e)
net/netfilter/nf_conntrack_ecache.c
320
local64_set(&e->timestamp, ts);
net/netfilter/nf_conntrack_ecache.c
327
struct nf_conntrack_ecache *e;
net/netfilter/nf_conntrack_ecache.c
351
e = nf_ct_ext_add(ct, NF_CT_EXT_ECACHE, gfp);
net/netfilter/nf_conntrack_ecache.c
352
if (e) {
net/netfilter/nf_conntrack_ecache.c
353
nf_ct_ecache_tstamp_new(ct, e);
net/netfilter/nf_conntrack_ecache.c
354
e->ctmask = ctmask;
net/netfilter/nf_conntrack_ecache.c
355
e->expmask = expmask;
net/netfilter/nf_conntrack_ecache.c
358
return e != NULL;
net/netfilter/nf_conntrack_expect.c
535
void nf_ct_expect_iterate_destroy(bool (*iter)(struct nf_conntrack_expect *e, void *data),
net/netfilter/nf_conntrack_expect.c
560
bool (*iter)(struct nf_conntrack_expect *e, void *data),
net/netfilter/nf_conntrack_netlink.c
391
const struct nf_conntrack_ecache *e = nf_ct_ecache_find(ct);
net/netfilter/nf_conntrack_netlink.c
393
if (e) {
net/netfilter/nf_conntrack_netlink.c
394
u64 ts = local64_read(&e->timestamp);
net/netfilter/nf_nat_core.c
934
struct nf_hook_entries *e = rcu_dereference(lpriv->entries);
net/netfilter/nf_nat_core.c
938
if (!e)
net/netfilter/nf_nat_core.c
941
for (i = 0; i < e->num_hook_entries; i++) {
net/netfilter/nf_nat_core.c
942
ret = e->hooks[i].hook(e->hooks[i].priv, skb,
net/netfilter/nf_tables_core.c
208
unsigned long e;
net/netfilter/nf_tables_core.c
213
e = (unsigned long)expr->ops->eval;
net/netfilter/nf_tables_core.c
214
#define X(e, fun) \
net/netfilter/nf_tables_core.c
215
do { if ((e) == (unsigned long)(fun)) \
net/netfilter/nf_tables_core.c
218
X(e, nft_payload_eval);
net/netfilter/nf_tables_core.c
219
X(e, nft_cmp_eval);
net/netfilter/nf_tables_core.c
220
X(e, nft_counter_eval);
net/netfilter/nf_tables_core.c
221
X(e, nft_meta_get_eval);
net/netfilter/nf_tables_core.c
222
X(e, nft_lookup_eval);
net/netfilter/nf_tables_core.c
224
X(e, nft_ct_get_fast_eval);
net/netfilter/nf_tables_core.c
226
X(e, nft_range_eval);
net/netfilter/nf_tables_core.c
227
X(e, nft_immediate_eval);
net/netfilter/nf_tables_core.c
228
X(e, nft_byteorder_eval);
net/netfilter/nf_tables_core.c
229
X(e, nft_dynset_eval);
net/netfilter/nf_tables_core.c
230
X(e, nft_rt_get_eval);
net/netfilter/nf_tables_core.c
231
X(e, nft_bitwise_eval);
net/netfilter/nf_tables_core.c
232
X(e, nft_objref_eval);
net/netfilter/nf_tables_core.c
233
X(e, nft_objref_map_eval);
net/netfilter/nfnetlink_hook.c
348
const struct nf_hook_entries *e;
net/netfilter/nfnetlink_hook.c
353
e = nfnl_hook_entries_head(family, ctx->hook, net, ctx->devname);
net/netfilter/nfnetlink_hook.c
354
if (!e)
net/netfilter/nfnetlink_hook.c
357
if (IS_ERR(e)) {
net/netfilter/nfnetlink_hook.c
362
if ((unsigned long)e != ctx->headv || i >= e->num_hook_entries)
net/netfilter/nfnetlink_hook.c
365
ops = nf_hook_entries_get_hook_ops(e);
net/netfilter/nfnetlink_hook.c
367
for (; i < e->num_hook_entries; i++) {
net/netfilter/nfnetlink_queue.c
1126
nfqnl_mangle(void *data, unsigned int data_len, struct nf_queue_entry *e, int diff)
net/netfilter/nfnetlink_queue.c
1131
unsigned int min_len = skb_transport_offset(e->skb);
net/netfilter/nfnetlink_queue.c
1136
if (pskb_trim(e->skb, data_len))
net/netfilter/nfnetlink_queue.c
1141
if (diff > skb_tailroom(e->skb)) {
net/netfilter/nfnetlink_queue.c
1142
nskb = skb_copy_expand(e->skb, skb_headroom(e->skb),
net/netfilter/nfnetlink_queue.c
1146
kfree_skb(e->skb);
net/netfilter/nfnetlink_queue.c
1147
e->skb = nskb;
net/netfilter/nfnetlink_queue.c
1149
skb_put(e->skb, diff);
net/netfilter/nfnetlink_queue.c
1151
if (skb_ensure_writable(e->skb, data_len))
net/netfilter/nfnetlink_queue.c
1153
skb_copy_to_linear_data(e->skb, data, data_len);
net/netfilter/nfnetlink_queue.c
1154
e->skb->ip_summed = CHECKSUM_NONE;
net/netfilter/nfnetlink_queue.c
971
nf_queue_entry_dup(struct nf_queue_entry *e)
net/netfilter/nfnetlink_queue.c
973
struct nf_queue_entry *entry = kmemdup(e, e->size, GFP_ATOMIC);
net/netfilter/nft_compat.c
253
union nft_entry e = {};
net/netfilter/nft_compat.c
264
nft_target_set_tgchk_param(&par, ctx, target, info, &e, proto, inv);
net/netfilter/nft_compat.c
507
union nft_entry e = {};
net/netfilter/nft_compat.c
518
nft_match_set_mtchk_param(&par, ctx, match, info, &e, proto, inv);
net/netfilter/nft_compat.c
607
const struct nft_expr *e, bool reset)
net/netfilter/nft_compat.c
609
struct nft_xt_match_priv *priv = nft_expr_priv(e);
net/netfilter/nft_compat.c
611
return __nft_match_dump(skb, e, priv->info);
net/netfilter/nft_ct.c
319
struct nf_conntrack_ecache *e = nf_ct_ecache_find(ct);
net/netfilter/nft_ct.c
322
if (e) {
net/netfilter/nft_ct.c
323
if (e->ctmask != ctmask)
net/netfilter/nft_ct.c
324
e->ctmask = ctmask;
net/netfilter/nft_set_pipapo.c
1175
struct nft_pipapo_elem *e)
net/netfilter/nft_set_pipapo.c
1189
f->mt[map[i].to + j].e = e;
net/netfilter/nft_set_pipapo.c
1295
struct nft_pipapo_elem *e, *dup;
net/netfilter/nft_set_pipapo.c
1396
e = nft_elem_priv_cast(elem->priv);
net/netfilter/nft_set_pipapo.c
1397
*elem_priv = &e->priv;
net/netfilter/nft_set_pipapo.c
1399
pipapo_map(m, rulemap, e);
net/netfilter/nft_set_pipapo.c
1529
struct nft_pipapo_elem *e = NULL; /* Keep gcc happy */
net/netfilter/nft_set_pipapo.c
1533
if (r != first && e != f->mt[r].e)
net/netfilter/nft_set_pipapo.c
1536
e = f->mt[r].e;
net/netfilter/nft_set_pipapo.c
1677
struct nft_pipapo_elem *e)
net/netfilter/nft_set_pipapo.c
1680
nft_setelem_data_deactivate(net, set, &e->priv);
net/netfilter/nft_set_pipapo.c
1694
struct nft_pipapo_elem *e;
net/netfilter/nft_set_pipapo.c
1724
e = f->mt[rulemap[i].to].e;
net/netfilter/nft_set_pipapo.c
1729
if (__nft_set_elem_expired(&e->ext, tstamp)) {
net/netfilter/nft_set_pipapo.c
1738
nft_pipapo_gc_deactivate(net, set, e);
net/netfilter/nft_set_pipapo.c
1740
nft_trans_gc_elem_add(gc, e);
net/netfilter/nft_set_pipapo.c
1878
struct nft_pipapo_elem *e = nft_elem_priv_cast(elem_priv);
net/netfilter/nft_set_pipapo.c
1880
nft_clear(net, &e->ext);
net/netfilter/nft_set_pipapo.c
1896
struct nft_pipapo_elem *e;
net/netfilter/nft_set_pipapo.c
1904
e = pipapo_get(m, (const u8 *)elem->key.val.data,
net/netfilter/nft_set_pipapo.c
1906
if (!e)
net/netfilter/nft_set_pipapo.c
1909
nft_set_elem_change_active(net, set, &e->ext);
net/netfilter/nft_set_pipapo.c
1911
return &e->priv;
net/netfilter/nft_set_pipapo.c
1935
struct nft_pipapo_elem *e = nft_elem_priv_cast(elem_priv);
net/netfilter/nft_set_pipapo.c
1937
nft_set_elem_change_active(net, set, &e->ext);
net/netfilter/nft_set_pipapo.c
2073
struct nft_pipapo_elem *e;
net/netfilter/nft_set_pipapo.c
2076
e = nft_elem_priv_cast(elem_priv);
net/netfilter/nft_set_pipapo.c
2077
data = (const u8 *)nft_set_ext_key(&e->ext);
net/netfilter/nft_set_pipapo.c
2087
if (nft_set_ext_exists(&e->ext, NFT_SET_EXT_KEY_END))
net/netfilter/nft_set_pipapo.c
2088
match_end = (const u8 *)nft_set_ext_key_end(&e->ext)->data;
net/netfilter/nft_set_pipapo.c
2111
if (last && f->mt[rulemap[i].to].e == e) {
net/netfilter/nft_set_pipapo.c
2145
struct nft_pipapo_elem *e;
net/netfilter/nft_set_pipapo.c
2147
if (r < f->rules - 1 && f->mt[r + 1].e == f->mt[r].e)
net/netfilter/nft_set_pipapo.c
2153
e = f->mt[r].e;
net/netfilter/nft_set_pipapo.c
2155
iter->err = iter->fn(ctx, set, iter, &e->priv);
net/netfilter/nft_set_pipapo.c
2343
struct nft_pipapo_elem *e;
net/netfilter/nft_set_pipapo.c
2345
if (r < f->rules - 1 && f->mt[r + 1].e == f->mt[r].e)
net/netfilter/nft_set_pipapo.c
2348
e = f->mt[r].e;
net/netfilter/nft_set_pipapo.c
2350
nf_tables_set_elem_destroy(ctx, set, &e->priv);
net/netfilter/nft_set_pipapo.c
476
struct nft_pipapo_elem *e;
net/netfilter/nft_set_pipapo.c
478
e = f->mt[b].e;
net/netfilter/nft_set_pipapo.c
479
if (unlikely(__nft_set_elem_expired(&e->ext, tstamp) ||
net/netfilter/nft_set_pipapo.c
480
!nft_set_elem_active(&e->ext, genmask)))
net/netfilter/nft_set_pipapo.c
491
return e;
net/netfilter/nft_set_pipapo.c
528
struct nft_pipapo_elem *e;
net/netfilter/nft_set_pipapo.c
534
e = pipapo_get_avx2(m, data, genmask, tstamp);
net/netfilter/nft_set_pipapo.c
536
return e;
net/netfilter/nft_set_pipapo.c
539
e = pipapo_get_slow(m, data, genmask, tstamp);
net/netfilter/nft_set_pipapo.c
541
return e;
net/netfilter/nft_set_pipapo.c
577
const struct nft_pipapo_elem *e;
net/netfilter/nft_set_pipapo.c
580
e = pipapo_get_slow(m, (const u8 *)key, 0, get_jiffies_64());
net/netfilter/nft_set_pipapo.c
582
return e ? &e->ext : NULL;
net/netfilter/nft_set_pipapo.c
603
struct nft_pipapo_elem *e;
net/netfilter/nft_set_pipapo.c
605
e = pipapo_get(m, (const u8 *)elem->key.val.data,
net/netfilter/nft_set_pipapo.c
607
if (!e)
net/netfilter/nft_set_pipapo.c
610
return &e->priv;
net/netfilter/nft_set_pipapo.h
102
struct nft_pipapo_elem *e;
net/netfilter/nft_set_pipapo_avx2.c
1237
struct nft_pipapo_elem *e;
net/netfilter/nft_set_pipapo_avx2.c
1239
e = f->mt[ret].e;
net/netfilter/nft_set_pipapo_avx2.c
1240
if (unlikely(__nft_set_elem_expired(&e->ext, tstamp) ||
net/netfilter/nft_set_pipapo_avx2.c
1241
!nft_set_elem_active(&e->ext, genmask))) {
net/netfilter/nft_set_pipapo_avx2.c
1250
return e;
net/netfilter/nft_set_pipapo_avx2.c
1283
const struct nft_pipapo_elem *e;
net/netfilter/nft_set_pipapo_avx2.c
1298
e = pipapo_get_avx2(m, rp, 0, get_jiffies_64());
net/netfilter/nft_set_pipapo_avx2.c
1301
return e ? &e->ext : NULL;
net/netfilter/x_tables.c
837
const char *e = base;
net/netfilter/x_tables.c
845
t = (void *)(e + target_offset);
net/netfilter/x_tables.c
930
const char *e = base;
net/netfilter/x_tables.c
939
t = (void *)(e + target_offset);
net/netfilter/xt_CT.c
58
const struct ipt_entry *e = par->entryinfo;
net/netfilter/xt_CT.c
60
if (e->ip.invflags & IPT_INV_PROTO)
net/netfilter/xt_CT.c
62
return e->ip.proto;
net/netfilter/xt_CT.c
64
const struct ip6t_entry *e = par->entryinfo;
net/netfilter/xt_CT.c
66
if (e->ipv6.invflags & IP6T_INV_PROTO)
net/netfilter/xt_CT.c
68
return e->ipv6.proto;
net/netfilter/xt_TCPMSS.c
266
const struct ipt_entry *e = par->entryinfo;
net/netfilter/xt_TCPMSS.c
279
xt_ematch_foreach(ematch, e)
net/netfilter/xt_TCPMSS.c
290
const struct ip6t_entry *e = par->entryinfo;
net/netfilter/xt_TCPMSS.c
303
xt_ematch_foreach(ematch, e)
net/netfilter/xt_l2tp.c
262
const struct ipt_entry *e = par->entryinfo;
net/netfilter/xt_l2tp.c
263
const struct ipt_ip *ip = &e->ip;
net/netfilter/xt_l2tp.c
289
const struct ip6t_entry *e = par->entryinfo;
net/netfilter/xt_l2tp.c
290
const struct ip6t_ip6 *ip = &e->ipv6;
net/netfilter/xt_multiport.c
32
u_int16_t s, e;
net/netfilter/xt_multiport.c
39
e = minfo->ports[++i];
net/netfilter/xt_multiport.c
40
pr_debug("src or dst matches with %d-%d?\n", s, e);
net/netfilter/xt_multiport.c
44
if (src >= s && src <= e)
net/netfilter/xt_multiport.c
48
if (dst >= s && dst <= e)
net/netfilter/xt_multiport.c
52
if ((dst >= s && dst <= e) ||
net/netfilter/xt_multiport.c
53
(src >= s && src <= e))
net/netfilter/xt_policy.c
100
if (match_xfrm_state(dst->xfrm, e, family)) {
net/netfilter/xt_policy.c
35
match_xfrm_state(const struct xfrm_state *x, const struct xt_policy_elem *e,
net/netfilter/xt_policy.c
38
#define MATCH_ADDR(x,y,z) (!e->match.x || \
net/netfilter/xt_policy.c
39
(xt_addr_cmp(&e->x, &e->y, (const union nf_inet_addr *)(z), family) \
net/netfilter/xt_policy.c
40
^ e->invert.x))
net/netfilter/xt_policy.c
41
#define MATCH(x,y) (!e->match.x || ((e->x == (y)) ^ e->invert.x))
net/netfilter/xt_policy.c
55
const struct xt_policy_elem *e;
net/netfilter/xt_policy.c
69
e = &info->pol[pos];
net/netfilter/xt_policy.c
71
if (match_xfrm_state(sp->xvec[i], e, family)) {
net/netfilter/xt_policy.c
85
const struct xt_policy_elem *e;
net/netfilter/xt_policy.c
98
e = &info->pol[pos];
net/netfilter/xt_recent.c
128
struct recent_entry *e;
net/netfilter/xt_recent.c
136
list_for_each_entry(e, &table->iphash[h], list)
net/netfilter/xt_recent.c
137
if (e->family == family &&
net/netfilter/xt_recent.c
138
memcmp(&e->addr, addrp, sizeof(e->addr)) == 0 &&
net/netfilter/xt_recent.c
139
(ttl == e->ttl || ttl == 0 || e->ttl == 0))
net/netfilter/xt_recent.c
140
return e;
net/netfilter/xt_recent.c
144
static void recent_entry_remove(struct recent_table *t, struct recent_entry *e)
net/netfilter/xt_recent.c
146
list_del(&e->list);
net/netfilter/xt_recent.c
147
list_del(&e->lru_list);
net/netfilter/xt_recent.c
148
kfree(e);
net/netfilter/xt_recent.c
158
struct recent_entry *e;
net/netfilter/xt_recent.c
163
e = list_entry(t->lru_list.next, struct recent_entry, lru_list);
net/netfilter/xt_recent.c
168
if (e == working && update)
net/netfilter/xt_recent.c
174
if (time_after(time, e->stamps[e->index-1]))
net/netfilter/xt_recent.c
175
recent_entry_remove(t, e);
net/netfilter/xt_recent.c
182
struct recent_entry *e;
net/netfilter/xt_recent.c
186
e = list_entry(t->lru_list.next, struct recent_entry, lru_list);
net/netfilter/xt_recent.c
187
recent_entry_remove(t, e);
net/netfilter/xt_recent.c
191
e = kmalloc_flex(*e, stamps, nstamps_max, GFP_ATOMIC);
net/netfilter/xt_recent.c
192
if (e == NULL)
net/netfilter/xt_recent.c
194
memcpy(&e->addr, addr, sizeof(e->addr));
net/netfilter/xt_recent.c
195
e->ttl = ttl;
net/netfilter/xt_recent.c
196
e->stamps[0] = jiffies;
net/netfilter/xt_recent.c
197
e->nstamps = 1;
net/netfilter/xt_recent.c
198
e->index = 1;
net/netfilter/xt_recent.c
199
e->family = family;
net/netfilter/xt_recent.c
201
list_add_tail(&e->list, &t->iphash[recent_entry_hash4(addr)]);
net/netfilter/xt_recent.c
203
list_add_tail(&e->list, &t->iphash[recent_entry_hash6(addr)]);
net/netfilter/xt_recent.c
204
list_add_tail(&e->lru_list, &t->lru_list);
net/netfilter/xt_recent.c
206
return e;
net/netfilter/xt_recent.c
209
static void recent_entry_update(struct recent_table *t, struct recent_entry *e)
net/netfilter/xt_recent.c
211
e->index &= t->nstamps_max_mask;
net/netfilter/xt_recent.c
212
e->stamps[e->index++] = jiffies;
net/netfilter/xt_recent.c
213
if (e->index > e->nstamps)
net/netfilter/xt_recent.c
214
e->nstamps = e->index;
net/netfilter/xt_recent.c
215
list_move_tail(&e->lru_list, &t->lru_list);
net/netfilter/xt_recent.c
231
struct recent_entry *e, *next;
net/netfilter/xt_recent.c
235
list_for_each_entry_safe(e, next, &t->iphash[i], list)
net/netfilter/xt_recent.c
236
recent_entry_remove(t, e);
net/netfilter/xt_recent.c
246
struct recent_entry *e;
net/netfilter/xt_recent.c
281
e = recent_entry_lookup(t, &addr_mask, xt_family(par),
net/netfilter/xt_recent.c
283
if (e == NULL) {
net/netfilter/xt_recent.c
286
e = recent_entry_init(t, &addr_mask, xt_family(par), ttl);
net/netfilter/xt_recent.c
287
if (e == NULL)
net/netfilter/xt_recent.c
296
recent_entry_remove(t, e);
net/netfilter/xt_recent.c
302
for (i = 0; i < e->nstamps; i++) {
net/netfilter/xt_recent.c
303
if (info->seconds && time_after(time, e->stamps[i]))
net/netfilter/xt_recent.c
313
recent_entry_reap(t, time, e,
net/netfilter/xt_recent.c
319
recent_entry_update(t, e);
net/netfilter/xt_recent.c
320
e->ttl = ttl;
net/netfilter/xt_recent.c
484
struct recent_entry *e;
net/netfilter/xt_recent.c
490
list_for_each_entry(e, &t->iphash[st->bucket], list)
net/netfilter/xt_recent.c
492
return e;
net/netfilter/xt_recent.c
500
const struct recent_entry *e = v;
net/netfilter/xt_recent.c
501
const struct list_head *head = e->list.next;
net/netfilter/xt_recent.c
520
const struct recent_entry *e = v;
net/netfilter/xt_recent.c
525
i = (e->index - 1) & t->nstamps_max_mask;
net/netfilter/xt_recent.c
527
if (e->family == NFPROTO_IPV4)
net/netfilter/xt_recent.c
529
&e->addr.ip, e->ttl, e->stamps[i], e->index);
net/netfilter/xt_recent.c
532
&e->addr.in6, e->ttl, e->stamps[i], e->index);
net/netfilter/xt_recent.c
533
for (i = 0; i < e->nstamps; i++)
net/netfilter/xt_recent.c
534
seq_printf(seq, "%s %lu", i ? "," : "", e->stamps[i]);
net/netfilter/xt_recent.c
563
struct recent_entry *e;
net/netfilter/xt_recent.c
611
e = recent_entry_lookup(t, &addr, family, 0);
net/netfilter/xt_recent.c
612
if (e == NULL) {
net/netfilter/xt_recent.c
617
recent_entry_update(t, e);
net/netfilter/xt_recent.c
619
recent_entry_remove(t, e);
net/openvswitch/flow_table.c
818
struct mask_cache_entry *e;
net/openvswitch/flow_table.c
820
e = &entries[index];
net/openvswitch/flow_table.c
821
if (e->skb_hash == skb_hash) {
net/openvswitch/flow_table.c
823
n_cache_hit, &e->mask_index);
net/openvswitch/flow_table.c
825
e->skb_hash = 0;
net/openvswitch/flow_table.c
829
if (!ce || e->skb_hash < ce->skb_hash)
net/openvswitch/flow_table.c
830
ce = e; /* A better replacement cache candidate. */
net/sched/act_gate.c
225
struct tcfg_gate_entry *entry, *e;
net/sched/act_gate.c
227
list_for_each_entry_safe(entry, e, entries, list) {
net/sched/act_ife.c
371
struct tcf_meta_info *e;
net/sched/act_ife.c
384
list_for_each_entry(e, &p->metalist, metalist) {
net/sched/act_ife.c
385
if (!e->ops->get(skb, e))
net/sched/act_ife.c
403
struct tcf_meta_info *e, *n;
net/sched/act_ife.c
405
list_for_each_entry_safe(e, n, &p->metalist, metalist) {
net/sched/act_ife.c
406
list_del(&e->metalist);
net/sched/act_ife.c
407
if (e->metaval) {
net/sched/act_ife.c
408
if (e->ops->release)
net/sched/act_ife.c
409
e->ops->release(e);
net/sched/act_ife.c
411
kfree(e->metaval);
net/sched/act_ife.c
413
module_put(e->ops->owner);
net/sched/act_ife.c
414
kfree(e);
net/sched/act_ife.c
695
struct tcf_meta_info *e;
net/sched/act_ife.c
698
list_for_each_entry_rcu(e, &p->metalist, metalist) {
net/sched/act_ife.c
699
if (metaid == e->metaid) {
net/sched/act_ife.c
700
if (e->ops) {
net/sched/act_ife.c
702
return e->ops->decode(skb, mdata, mlen);
net/sched/act_ife.c
773
struct tcf_meta_info *e;
net/sched/act_ife.c
776
list_for_each_entry_rcu(e, &p->metalist, metalist) {
net/sched/act_ife.c
777
if (e->ops->check_presence) {
net/sched/act_ife.c
778
run_sz = e->ops->check_presence(skb, e);
net/sched/act_ife.c
792
struct tcf_meta_info *e;
net/sched/act_ife.c
839
list_for_each_entry_rcu(e, &p->metalist, metalist) {
net/sched/act_ife.c
840
if (e->ops->encode) {
net/sched/act_ife.c
841
err = e->ops->encode(skb, (void *)(ife_meta + skboff),
net/sched/act_ife.c
842
e);
net/sched/act_meta_mark.c
22
struct tcf_meta_info *e)
net/sched/act_meta_mark.c
26
return ife_encode_meta_u32(ifemark, skbdata, e);
net/sched/act_meta_mark.c
37
static int skbmark_check(struct sk_buff *skb, struct tcf_meta_info *e)
net/sched/act_meta_mark.c
39
return ife_check_meta_u32(skb->mark, e);
net/sched/act_meta_skbprio.c
21
static int skbprio_check(struct sk_buff *skb, struct tcf_meta_info *e)
net/sched/act_meta_skbprio.c
23
return ife_check_meta_u32(skb->priority, e);
net/sched/act_meta_skbprio.c
27
struct tcf_meta_info *e)
net/sched/act_meta_skbprio.c
31
return ife_encode_meta_u32(ifeprio, skbdata, e);
net/sched/act_meta_skbtcindex.c
22
struct tcf_meta_info *e)
net/sched/act_meta_skbtcindex.c
26
return ife_encode_meta_u16(ifetc_index, skbdata, e);
net/sched/act_meta_skbtcindex.c
37
static int skbtcindex_check(struct sk_buff *skb, struct tcf_meta_info *e)
net/sched/act_meta_skbtcindex.c
39
return ife_check_meta_u16(skb->tc_index, e);
net/sched/em_ipt.c
48
} e = {};
net/sched/em_ipt.c
55
mtpar.entryinfo = &e;
net/sched/ematch.c
100
e = NULL;
net/sched/ematch.c
102
return e;
net/sched/ematch.c
125
struct tcf_ematch_ops *e;
net/sched/ematch.c
131
list_for_each_entry(e, &ematch_ops, link)
net/sched/ematch.c
132
if (ops->kind == e->kind)
net/sched/ematch.c
94
struct tcf_ematch_ops *e = NULL;
net/sched/ematch.c
97
list_for_each_entry(e, &ematch_ops, link) {
net/sched/ematch.c
98
if (kind == e->kind) {
net/sched/ematch.c
99
if (!try_module_get(e->owner))
net/sched/sch_api.c
2452
struct proc_dir_entry *e;
net/sched/sch_api.c
2454
e = proc_create_single("psched", 0, net->proc_net, psched_show);
net/sched/sch_api.c
2455
if (e == NULL)
net/sched/sch_taprio.c
1470
struct tc_taprio_sched_entry *e = &offload->entries[i];
net/sched/sch_taprio.c
1472
e->command = entry->command;
net/sched/sch_taprio.c
1473
e->interval = entry->interval;
net/sched/sch_taprio.c
1475
e->gate_mask = tc_map_to_queue_mask(dev,
net/sched/sch_taprio.c
1478
e->gate_mask = entry->gate_mask;
net/tipc/node.c
2211
struct tipc_link_entry *e;
net/tipc/node.c
2220
e = &n->links[bearer_id];
net/tipc/node.c
2221
if (e->link) {
net/tipc/node.c
2223
tipc_link_set_tolerance(e->link, b->tolerance,
net/tipc/node.c
2226
tipc_link_set_mtu(e->link, b->mtu);
net/tipc/node.c
2229
e->mtu = tipc_link_mss(e->link);
net/tipc/node.c
2233
tipc_bearer_xmit(net, bearer_id, &xmitq, &e->maddr, NULL);
net/tipc/topsrv.c
128
struct outqueue_entry *e, *safe;
net/tipc/topsrv.c
138
list_for_each_entry_safe(e, safe, &con->outqueue, list) {
net/tipc/topsrv.c
139
list_del(&e->list);
net/tipc/topsrv.c
140
kfree(e);
net/tipc/topsrv.c
253
struct outqueue_entry *e;
net/tipc/topsrv.c
263
e = list_first_entry(queue, struct outqueue_entry, list);
net/tipc/topsrv.c
264
evt = &e->evt;
net/tipc/topsrv.c
267
if (e->inactive)
net/tipc/topsrv.c
295
list_del(&e->list);
net/tipc/topsrv.c
296
kfree(e);
net/tipc/topsrv.c
318
struct outqueue_entry *e;
net/tipc/topsrv.c
328
e = kmalloc_obj(*e, GFP_ATOMIC);
net/tipc/topsrv.c
329
if (!e)
net/tipc/topsrv.c
331
e->inactive = (event == TIPC_SUBSCR_TIMEOUT);
net/tipc/topsrv.c
332
memcpy(&e->evt, evt, sizeof(*evt));
net/tipc/topsrv.c
334
list_add_tail(&e->list, &con->outqueue);
net/wireless/ibss.c
356
freq->e = 6;
net/wireless/scan.c
3713
iwe.u.freq.e = 0;
net/wireless/scan.c
3722
iwe.u.freq.e = 6;
net/wireless/wext-compat.c
200
range->freq[c].e = 6;
net/wireless/wext-compat.c
231
if (freq->e == 0) {
net/wireless/wext-compat.c
240
for (i = 0; i < freq->e; i++)
net/wireless/wext-compat.c
834
freq->e = 6;
net/wireless/wext-sme.c
130
freq->e = 6;
samples/bpf/trace_output_user.c
27
} *e = data;
samples/bpf/trace_output_user.c
29
if (e->cookie != 0x12345678) {
samples/bpf/trace_output_user.c
31
e->pid, e->cookie, size);
samples/bpf/xdp_sample_user.c
1013
struct map_entry *e, *x = NULL;
samples/bpf/xdp_sample_user.c
1034
hash_for_each_possible(stats_prev->xmit_map, e, node, pair) {
samples/bpf/xdp_sample_user.c
1035
if (e->pair == pair) {
samples/bpf/xdp_sample_user.c
1036
x = e;
samples/bpf/xdp_sample_user.c
329
static int map_entry_init(struct map_entry *e, __u64 pair)
samples/bpf/xdp_sample_user.c
331
e->pair = pair;
samples/bpf/xdp_sample_user.c
332
INIT_HLIST_NODE(&e->node);
samples/bpf/xdp_sample_user.c
333
e->val.timestamp = gettime();
samples/bpf/xdp_sample_user.c
334
e->val.cpu = alloc_record_per_cpu();
samples/bpf/xdp_sample_user.c
335
if (!e->val.cpu)
samples/bpf/xdp_sample_user.c
410
struct map_entry *e, *x = NULL;
samples/bpf/xdp_sample_user.c
415
hash_for_each_possible(rec->xmit_map, e, node, pair) {
samples/bpf/xdp_sample_user.c
416
if (e->pair == pair) {
samples/bpf/xdp_sample_user.c
417
x = e;
samples/bpf/xdp_sample_user.c
550
struct map_entry *e;
samples/bpf/xdp_sample_user.c
555
hash_for_each_safe(r->xmit_map, i, tmp, e, node) {
samples/bpf/xdp_sample_user.c
556
hash_del(&e->node);
samples/bpf/xdp_sample_user.c
557
free(e->val.cpu);
samples/bpf/xdp_sample_user.c
558
free(e);
scripts/asn1_compiler.c
1210
static void dump_element(const struct element *e, int level)
scripts/asn1_compiler.c
1213
const struct type *t = e->type_def;
scripts/asn1_compiler.c
1214
const char *name = e->name ? e->name->content : ".";
scripts/asn1_compiler.c
1218
if (e->class == 0 && e->method == 0 && e->tag == 0)
scripts/asn1_compiler.c
1220
else if (e->class == ASN1_UNIV)
scripts/asn1_compiler.c
1222
asn1_classes[e->class],
scripts/asn1_compiler.c
1223
asn1_methods[e->method],
scripts/asn1_compiler.c
1224
asn1_universal_tags[e->tag]);
scripts/asn1_compiler.c
1227
asn1_classes[e->class],
scripts/asn1_compiler.c
1228
asn1_methods[e->method],
scripts/asn1_compiler.c
1229
e->tag);
scripts/asn1_compiler.c
1232
e->flags & ELEMENT_IMPLICIT ? 'I' : '-',
scripts/asn1_compiler.c
1233
e->flags & ELEMENT_EXPLICIT ? 'E' : '-',
scripts/asn1_compiler.c
1234
e->flags & ELEMENT_TAG_SPECIFIED ? 'T' : '-',
scripts/asn1_compiler.c
1235
e->flags & ELEMENT_SKIPPABLE ? 'S' : '-',
scripts/asn1_compiler.c
1236
e->flags & ELEMENT_CONDITIONAL ? 'C' : '-',
scripts/asn1_compiler.c
1237
"-tTqQcaro"[e->compound],
scripts/asn1_compiler.c
1242
e->action ? e->action->name : "");
scripts/asn1_compiler.c
1243
if (e->compound == TYPE_REF)
scripts/asn1_compiler.c
1244
dump_element(e->type->type->element, level + 3);
scripts/asn1_compiler.c
1246
for (c = e->children; c; c = c->next)
scripts/asn1_compiler.c
1256
static void render_element(FILE *out, struct element *e, struct element *tag);
scripts/asn1_compiler.c
1294
struct element *e;
scripts/asn1_compiler.c
1364
for (e = element_list; e; e = e->list_next)
scripts/asn1_compiler.c
1365
e->flags &= ~ELEMENT_RENDERED;
scripts/asn1_compiler.c
1394
struct element *e, *ce;
scripts/asn1_compiler.c
1398
while ((e = render_list)) {
scripts/asn1_compiler.c
1399
render_list = e->render_next;
scripts/asn1_compiler.c
1404
e->entry_index = entry = nr_entries;
scripts/asn1_compiler.c
1406
for (ce = e->children; ce; ce = ce->next)
scripts/asn1_compiler.c
1410
act = e->action ? "_ACT" : "";
scripts/asn1_compiler.c
1411
switch (e->compound) {
scripts/asn1_compiler.c
1429
if (e->action)
scripts/asn1_compiler.c
1431
e->action->name);
scripts/asn1_compiler.c
1439
static void render_element(FILE *out, struct element *e, struct element *tag)
scripts/asn1_compiler.c
1445
if (e->flags & ELEMENT_SKIPPABLE ||
scripts/asn1_compiler.c
1449
if ((e->type_def && e->type_def->ref_count > 1) ||
scripts/asn1_compiler.c
1453
if (e->type_def && out) {
scripts/asn1_compiler.c
1454
render_more(out, "\t// %s\n", e->type_def->name->content);
scripts/asn1_compiler.c
1458
cond = (e->flags & ELEMENT_CONDITIONAL ||
scripts/asn1_compiler.c
1460
act = e->action ? "_ACT" : "";
scripts/asn1_compiler.c
1461
switch (e->compound) {
scripts/asn1_compiler.c
1465
if (e->name)
scripts/asn1_compiler.c
1466
render_more(out, "\t\t// %s", e->name->content);
scripts/asn1_compiler.c
1471
render_element(out, e->children, e);
scripts/asn1_compiler.c
1488
if (e->class == ASN1_UNIV && e->method == ASN1_PRIM && e->tag == 0)
scripts/asn1_compiler.c
1497
x = tag ?: e;
scripts/asn1_compiler.c
1504
tag = e;
scripts/asn1_compiler.c
1512
asn1_methods[tag->method | e->method],
scripts/asn1_compiler.c
1517
asn1_methods[tag->method | e->method],
scripts/asn1_compiler.c
1523
switch (e->compound) {
scripts/asn1_compiler.c
1525
render_element(out, e->type->type->element, tag);
scripts/asn1_compiler.c
1526
if (e->action)
scripts/asn1_compiler.c
1535
render_opcode(out, "_jump_target(%u),", e->entry_index);
scripts/asn1_compiler.c
1536
if (e->type_def && e->type_def->name)
scripts/asn1_compiler.c
1538
e->type_def->name->content);
scripts/asn1_compiler.c
1540
if (!(e->flags & ELEMENT_RENDERED)) {
scripts/asn1_compiler.c
1541
e->flags |= ELEMENT_RENDERED;
scripts/asn1_compiler.c
1542
*render_list_p = e;
scripts/asn1_compiler.c
1543
render_list_p = &e->render_next;
scripts/asn1_compiler.c
1549
for (ec = e->children; ec; ec = ec->next)
scripts/asn1_compiler.c
1561
render_opcode(out, "_jump_target(%u),", e->entry_index);
scripts/asn1_compiler.c
1562
if (e->type_def && e->type_def->name)
scripts/asn1_compiler.c
1564
e->type_def->name->content);
scripts/asn1_compiler.c
1566
if (!(e->flags & ELEMENT_RENDERED)) {
scripts/asn1_compiler.c
1567
e->flags |= ELEMENT_RENDERED;
scripts/asn1_compiler.c
1568
*render_list_p = e;
scripts/asn1_compiler.c
1569
render_list_p = &e->render_next;
scripts/asn1_compiler.c
1576
render_element(out, e->children, NULL);
scripts/asn1_compiler.c
1578
if (e->compound == SEQUENCE_OF)
scripts/asn1_compiler.c
1597
for (ec = e->children; ec; ec = ec->next)
scripts/asn1_compiler.c
1601
if (e->action)
scripts/asn1_compiler.c
1609
if (e->action)
scripts/asn1_compiler.c
1610
render_opcode(out, "_action(ACT_%s),\n", e->action->name);
scripts/asn1_compiler.c
837
struct element *e = calloc(1, sizeof(*e));
scripts/asn1_compiler.c
838
if (!e) {
scripts/asn1_compiler.c
842
e->list_next = element_list;
scripts/asn1_compiler.c
843
element_list = e;
scripts/asn1_compiler.c
844
return e;
scripts/dtc/flattree.c
123
static void asm_emit_cell(void *e, cell_t val)
scripts/dtc/flattree.c
125
FILE *f = e;
scripts/dtc/flattree.c
133
static void asm_emit_string(void *e, const char *str, int len)
scripts/dtc/flattree.c
135
FILE *f = e;
scripts/dtc/flattree.c
143
static void asm_emit_align(void *e, int a)
scripts/dtc/flattree.c
145
FILE *f = e;
scripts/dtc/flattree.c
150
static void asm_emit_data(void *e, struct data d)
scripts/dtc/flattree.c
152
FILE *f = e;
scripts/dtc/flattree.c
160
asm_emit_cell(e, dtb_ld32(d.val + off));
scripts/dtc/flattree.c
172
static void asm_emit_beginnode(void *e, struct label *labels)
scripts/dtc/flattree.c
174
FILE *f = e;
scripts/dtc/flattree.c
182
asm_emit_cell(e, FDT_BEGIN_NODE);
scripts/dtc/flattree.c
185
static void asm_emit_endnode(void *e, struct label *labels)
scripts/dtc/flattree.c
187
FILE *f = e;
scripts/dtc/flattree.c
191
asm_emit_cell(e, FDT_END_NODE);
scripts/dtc/flattree.c
198
static void asm_emit_property(void *e, struct label *labels)
scripts/dtc/flattree.c
200
FILE *f = e;
scripts/dtc/flattree.c
208
asm_emit_cell(e, FDT_PROP);
scripts/dtc/flattree.c
45
static void bin_emit_cell(void *e, cell_t val)
scripts/dtc/flattree.c
47
struct data *dtbuf = e;
scripts/dtc/flattree.c
52
static void bin_emit_string(void *e, const char *str, int len)
scripts/dtc/flattree.c
54
struct data *dtbuf = e;
scripts/dtc/flattree.c
63
static void bin_emit_align(void *e, int a)
scripts/dtc/flattree.c
65
struct data *dtbuf = e;
scripts/dtc/flattree.c
70
static void bin_emit_data(void *e, struct data d)
scripts/dtc/flattree.c
72
struct data *dtbuf = e;
scripts/dtc/flattree.c
77
static void bin_emit_beginnode(void *e, struct label *labels)
scripts/dtc/flattree.c
79
bin_emit_cell(e, FDT_BEGIN_NODE);
scripts/dtc/flattree.c
82
static void bin_emit_endnode(void *e, struct label *labels)
scripts/dtc/flattree.c
84
bin_emit_cell(e, FDT_END_NODE);
scripts/dtc/flattree.c
87
static void bin_emit_property(void *e, struct label *labels)
scripts/dtc/flattree.c
89
bin_emit_cell(e, FDT_PROP);
scripts/dtc/libfdt/fdt_overlay.c
910
const char *s, *e;
scripts/dtc/libfdt/fdt_overlay.c
946
e = path + path_len;
scripts/dtc/libfdt/fdt_overlay.c
964
if ((e - s) > len && (memcmp(s, "/__overlay__/", len) == 0)) {
scripts/dtc/libfdt/fdt_overlay.c
967
rel_path_len = e - rel_path - 1;
scripts/dtc/libfdt/fdt_overlay.c
968
} else if ((e - s) == len
scripts/gcc-plugins/latent_entropy_plugin.c
408
edge e, last_bb_e;
scripts/gcc-plugins/latent_entropy_plugin.c
414
FOR_EACH_EDGE(e, ei, last_bb_e->src->preds) {
scripts/gcc-plugins/latent_entropy_plugin.c
415
if (ENTRY_BLOCK_PTR_FOR_FN(cfun) == e->src)
scripts/gcc-plugins/latent_entropy_plugin.c
417
if (EXIT_BLOCK_PTR_FOR_FN(cfun) == e->src)
scripts/gcc-plugins/latent_entropy_plugin.c
420
handle_tail_calls(e->src, local_entropy);
scripts/gcc-plugins/randomize_layout_plugin.c
127
u64 e = x->a - rot(x->b, 7);
scripts/gcc-plugins/randomize_layout_plugin.c
130
x->c = x->d + e;
scripts/gcc-plugins/randomize_layout_plugin.c
131
x->d = e + x->a;
scripts/gendwarfksyms/examples/kabi_ex.c
14
enum e e1;
scripts/gendwarfksyms/examples/kabi_ex.h
59
KABI_ENUMERATOR_IGNORE(e, B);
scripts/gendwarfksyms/examples/kabi_ex.h
60
KABI_ENUMERATOR_IGNORE(e, C);
scripts/gendwarfksyms/examples/kabi_ex.h
61
KABI_ENUMERATOR_VALUE(e, D, 123456789);
scripts/gendwarfksyms/types.c
107
struct type_expansion *e;
scripts/gendwarfksyms/types.c
109
hash_for_each_possible(type_map, e, hash, hash_str(name)) {
scripts/gendwarfksyms/types.c
110
if (!strcmp(name, e->name)) {
scripts/gendwarfksyms/types.c
111
*res = e;
scripts/gendwarfksyms/types.c
122
struct type_expansion *e;
scripts/gendwarfksyms/types.c
124
if (__type_map_get(name, &e)) {
scripts/gendwarfksyms/types.c
125
e = xmalloc(sizeof(*e));
scripts/gendwarfksyms/types.c
126
type_expansion_init(e);
scripts/gendwarfksyms/types.c
127
e->name = xstrdup(name);
scripts/gendwarfksyms/types.c
129
hash_add(type_map, &e->hash, hash_str(e->name));
scripts/gendwarfksyms/types.c
132
debug("adding %s", e->name);
scripts/gendwarfksyms/types.c
135
if (type->len <= e->len)
scripts/gendwarfksyms/types.c
136
return e;
scripts/gendwarfksyms/types.c
138
type_list_free(&e->expanded);
scripts/gendwarfksyms/types.c
141
debug("replacing %s", e->name);
scripts/gendwarfksyms/types.c
145
list_replace_init(&type->expanded, &e->expanded);
scripts/gendwarfksyms/types.c
146
e->len = type->len;
scripts/gendwarfksyms/types.c
149
checkp(fputs(e->name, stderr));
scripts/gendwarfksyms/types.c
151
type_list_write(&e->expanded, stderr);
scripts/gendwarfksyms/types.c
155
return e;
scripts/gendwarfksyms/types.c
194
struct type_expansion *e;
scripts/gendwarfksyms/types.c
203
hash_for_each_safe(type_map, e, tmp, hash)
scripts/gendwarfksyms/types.c
206
hash_for_each_safe(type_map, e, tmp, hash)
scripts/gendwarfksyms/types.c
207
es[i++] = e;
scripts/gendwarfksyms/types.c
223
struct type_expansion *e;
scripts/gendwarfksyms/types.c
226
hash_for_each_safe(type_map, e, tmp, hash) {
scripts/gendwarfksyms/types.c
227
type_expansion_free(e);
scripts/gendwarfksyms/types.c
228
free(e);
scripts/gendwarfksyms/types.c
326
struct type_expansion *e;
scripts/gendwarfksyms/types.c
331
if (type_map_get(entry->str, &e))
scripts/gendwarfksyms/types.c
339
if (cache_was_expanded(&expansion_cache, e)) {
scripts/gendwarfksyms/types.c
342
cache_mark_expanded(&expansion_cache, e);
scripts/gendwarfksyms/types.c
343
__calculate_version(version, e);
scripts/genksyms/genksyms.c
333
void free_list(struct string_list *s, struct string_list *e)
scripts/genksyms/genksyms.c
335
while (s != e) {
scripts/genksyms/genksyms.c
507
struct string_list **e, **b;
scripts/genksyms/genksyms.c
520
b = alloca(elem * sizeof(*e));
scripts/genksyms/genksyms.c
521
e = b + elem;
scripts/genksyms/genksyms.c
522
tmp2 = e - 1;
scripts/genksyms/genksyms.c
528
while (b != e) {
scripts/genksyms/genksyms.c
537
struct string_list **e, **b;
scripts/genksyms/genksyms.c
548
b = alloca(elem * sizeof(*e));
scripts/genksyms/genksyms.c
549
e = b + elem;
scripts/genksyms/genksyms.c
550
tmp2 = e - 1;
scripts/genksyms/genksyms.c
556
while (b != e) {
scripts/genksyms/genksyms.h
62
void free_list(struct string_list *s, struct string_list *e);
scripts/genksyms/parse.y
38
struct string_list *b = *pb, *e = *pe;
scripts/genksyms/parse.y
39
*pb = e;
scripts/genksyms/parse.y
40
free_list(b, e);
scripts/genksyms/parse.y
536
yyerror(const char *e)
scripts/genksyms/parse.y
538
error_with_pos("%s", e);
scripts/insert-sys-cert.c
268
static void print_usage(char *e)
scripts/insert-sys-cert.c
270
printf("Usage %s [-s <System.map>] -b <vmlinux> -c <certfile>\n", e);
scripts/kconfig/expr.c
1022
void expr_print(const struct expr *e,
scripts/kconfig/expr.c
1026
if (!e) {
scripts/kconfig/expr.c
1031
if (expr_compare_type(prevtoken, e->type) > 0)
scripts/kconfig/expr.c
1033
switch (e->type) {
scripts/kconfig/expr.c
1035
if (e->left.sym->name)
scripts/kconfig/expr.c
1036
fn(data, e->left.sym, e->left.sym->name);
scripts/kconfig/expr.c
1042
expr_print(e->left.expr, fn, data, E_NOT);
scripts/kconfig/expr.c
1045
if (e->left.sym->name)
scripts/kconfig/expr.c
1046
fn(data, e->left.sym, e->left.sym->name);
scripts/kconfig/expr.c
1050
fn(data, e->right.sym, e->right.sym->name);
scripts/kconfig/expr.c
1054
if (e->left.sym->name)
scripts/kconfig/expr.c
1055
fn(data, e->left.sym, e->left.sym->name);
scripts/kconfig/expr.c
1058
fn(data, NULL, e->type == E_LEQ ? "<=" : "<");
scripts/kconfig/expr.c
1059
fn(data, e->right.sym, e->right.sym->name);
scripts/kconfig/expr.c
1063
if (e->left.sym->name)
scripts/kconfig/expr.c
1064
fn(data, e->left.sym, e->left.sym->name);
scripts/kconfig/expr.c
1067
fn(data, NULL, e->type == E_GEQ ? ">=" : ">");
scripts/kconfig/expr.c
1068
fn(data, e->right.sym, e->right.sym->name);
scripts/kconfig/expr.c
1071
if (e->left.sym->name)
scripts/kconfig/expr.c
1072
fn(data, e->left.sym, e->left.sym->name);
scripts/kconfig/expr.c
1076
fn(data, e->right.sym, e->right.sym->name);
scripts/kconfig/expr.c
1079
expr_print(e->left.expr, fn, data, E_OR);
scripts/kconfig/expr.c
1081
expr_print(e->right.expr, fn, data, E_OR);
scripts/kconfig/expr.c
1084
expr_print(e->left.expr, fn, data, E_AND);
scripts/kconfig/expr.c
1086
expr_print(e->right.expr, fn, data, E_AND);
scripts/kconfig/expr.c
1090
fn(data, e->left.sym, e->left.sym->name);
scripts/kconfig/expr.c
1092
fn(data, e->right.sym, e->right.sym->name);
scripts/kconfig/expr.c
1098
sprintf(buf, "<unknown type %d>", e->type);
scripts/kconfig/expr.c
1103
if (expr_compare_type(prevtoken, e->type) > 0)
scripts/kconfig/expr.c
1112
void expr_fprint(struct expr *e, FILE *out)
scripts/kconfig/expr.c
1114
expr_print(e, expr_print_file_helper, out, E_NONE);
scripts/kconfig/expr.c
1147
void expr_gstr_print(const struct expr *e, struct gstr *gs)
scripts/kconfig/expr.c
1149
expr_print(e, expr_print_gstr_helper, gs, E_NONE);
scripts/kconfig/expr.c
1157
static void expr_print_revdep(struct expr *e,
scripts/kconfig/expr.c
1161
if (e->type == E_OR) {
scripts/kconfig/expr.c
1162
expr_print_revdep(e->left.expr, fn, data, pr_type, title);
scripts/kconfig/expr.c
1163
expr_print_revdep(e->right.expr, fn, data, pr_type, title);
scripts/kconfig/expr.c
1164
} else if (expr_calc_value(e) == pr_type) {
scripts/kconfig/expr.c
1171
expr_print(e, fn, data, E_NONE);
scripts/kconfig/expr.c
1176
void expr_gstr_print_revdep(struct expr *e, struct gstr *gs,
scripts/kconfig/expr.c
1179
expr_print_revdep(e, expr_print_gstr_helper, gs, pr_type, &title);
scripts/kconfig/expr.c
21
static struct expr *expr_eliminate_yn(struct expr *e);
scripts/kconfig/expr.c
267
static struct expr *expr_eliminate_yn(struct expr *e)
scripts/kconfig/expr.c
271
if (e) switch (e->type) {
scripts/kconfig/expr.c
273
l = expr_eliminate_yn(e->left.expr);
scripts/kconfig/expr.c
274
r = expr_eliminate_yn(e->right.expr);
scripts/kconfig/expr.c
289
l = expr_eliminate_yn(e->left.expr);
scripts/kconfig/expr.c
290
r = expr_eliminate_yn(e->right.expr);
scripts/kconfig/expr.c
307
return e;
scripts/kconfig/expr.c
35
struct expr *e;
scripts/kconfig/expr.c
40
hash_for_each_possible(expr_hashtable, e, node, hash) {
scripts/kconfig/expr.c
41
if (e->type == type && e->left._initdata == l &&
scripts/kconfig/expr.c
42
e->right._initdata == r)
scripts/kconfig/expr.c
43
return e;
scripts/kconfig/expr.c
46
e = xmalloc(sizeof(*e));
scripts/kconfig/expr.c
47
e->type = type;
scripts/kconfig/expr.c
48
e->left._initdata = l;
scripts/kconfig/expr.c
49
e->right._initdata = r;
scripts/kconfig/expr.c
50
e->val_is_valid = false;
scripts/kconfig/expr.c
52
hash_add(expr_hashtable, &e->node, hash);
scripts/kconfig/expr.c
537
struct expr *expr_eliminate_dups(struct expr *e)
scripts/kconfig/expr.c
54
return e;
scripts/kconfig/expr.c
540
if (!e)
scripts/kconfig/expr.c
541
return e;
scripts/kconfig/expr.c
548
switch (e->type) {
scripts/kconfig/expr.c
550
l = expr_eliminate_dups(e->left.expr);
scripts/kconfig/expr.c
551
r = expr_eliminate_dups(e->right.expr);
scripts/kconfig/expr.c
552
expr_eliminate_dups1(e->type, &l, &r);
scripts/kconfig/expr.c
553
e = expr_alloc_two(e->type, l, r);
scripts/kconfig/expr.c
557
e = expr_eliminate_yn(e);
scripts/kconfig/expr.c
560
return e;
scripts/kconfig/expr.c
593
struct expr *expr_transform(struct expr *e)
scripts/kconfig/expr.c
595
if (!e)
scripts/kconfig/expr.c
597
switch (e->type) {
scripts/kconfig/expr.c
607
e = expr_alloc_two(e->type,
scripts/kconfig/expr.c
608
expr_transform(e->left.expr),
scripts/kconfig/expr.c
609
expr_transform(e->right.expr));
scripts/kconfig/expr.c
612
switch (e->type) {
scripts/kconfig/expr.c
614
if (e->left.sym->type != S_BOOLEAN)
scripts/kconfig/expr.c
616
if (e->right.sym == &symbol_no) {
scripts/kconfig/expr.c
618
e = expr_alloc_one(E_NOT, expr_alloc_symbol(e->left.sym));
scripts/kconfig/expr.c
621
if (e->right.sym == &symbol_mod) {
scripts/kconfig/expr.c
623
printf("boolean symbol %s tested for 'm'? test forced to 'n'\n", e->left.sym->name);
scripts/kconfig/expr.c
624
e = expr_alloc_symbol(&symbol_no);
scripts/kconfig/expr.c
627
if (e->right.sym == &symbol_yes) {
scripts/kconfig/expr.c
629
e = expr_alloc_symbol(e->left.sym);
scripts/kconfig/expr.c
634
if (e->left.sym->type != S_BOOLEAN)
scripts/kconfig/expr.c
636
if (e->right.sym == &symbol_no) {
scripts/kconfig/expr.c
638
e = expr_alloc_symbol(e->left.sym);
scripts/kconfig/expr.c
641
if (e->right.sym == &symbol_mod) {
scripts/kconfig/expr.c
643
printf("boolean symbol %s tested for 'm'? test forced to 'y'\n", e->left.sym->name);
scripts/kconfig/expr.c
644
e = expr_alloc_symbol(&symbol_yes);
scripts/kconfig/expr.c
647
if (e->right.sym == &symbol_yes) {
scripts/kconfig/expr.c
649
e = expr_alloc_one(E_NOT, e->left.expr);
scripts/kconfig/expr.c
654
switch (e->left.expr->type) {
scripts/kconfig/expr.c
657
e = e->left.expr->left.expr;
scripts/kconfig/expr.c
662
e = expr_alloc_comp(e->left.expr->type == E_EQUAL ? E_UNEQUAL : E_EQUAL,
scripts/kconfig/expr.c
663
e->left.expr->left.sym,
scripts/kconfig/expr.c
664
e->left.expr->right.sym);
scripts/kconfig/expr.c
669
e = expr_alloc_comp(e->left.expr->type == E_LEQ ? E_GTH : E_LTH,
scripts/kconfig/expr.c
670
e->left.expr->left.sym,
scripts/kconfig/expr.c
671
e->left.expr->right.sym);
scripts/kconfig/expr.c
676
e = expr_alloc_comp(e->left.expr->type == E_LTH ? E_GEQ : E_LEQ,
scripts/kconfig/expr.c
677
e->left.expr->left.sym,
scripts/kconfig/expr.c
678
e->left.expr->right.sym);
scripts/kconfig/expr.c
682
e = expr_alloc_and(expr_alloc_one(E_NOT, e->left.expr->left.expr),
scripts/kconfig/expr.c
683
expr_alloc_one(E_NOT, e->left.expr->right.expr));
scripts/kconfig/expr.c
684
e = expr_transform(e);
scripts/kconfig/expr.c
688
e = expr_alloc_or(expr_alloc_one(E_NOT, e->left.expr->left.expr),
scripts/kconfig/expr.c
689
expr_alloc_one(E_NOT, e->left.expr->right.expr));
scripts/kconfig/expr.c
690
e = expr_transform(e);
scripts/kconfig/expr.c
693
if (e->left.expr->left.sym == &symbol_yes)
scripts/kconfig/expr.c
695
e = expr_alloc_symbol(&symbol_no);
scripts/kconfig/expr.c
696
else if (e->left.expr->left.sym == &symbol_mod)
scripts/kconfig/expr.c
698
e = expr_alloc_symbol(&symbol_mod);
scripts/kconfig/expr.c
699
else if (e->left.expr->left.sym == &symbol_no)
scripts/kconfig/expr.c
701
e = expr_alloc_symbol(&symbol_yes);
scripts/kconfig/expr.c
710
return e;
scripts/kconfig/expr.c
784
struct expr *expr_trans_compare(struct expr *e, enum expr_type type, struct symbol *sym)
scripts/kconfig/expr.c
788
if (!e) {
scripts/kconfig/expr.c
789
e = expr_alloc_symbol(sym);
scripts/kconfig/expr.c
791
e = expr_alloc_one(E_NOT, e);
scripts/kconfig/expr.c
792
return e;
scripts/kconfig/expr.c
794
switch (e->type) {
scripts/kconfig/expr.c
796
e1 = expr_trans_compare(e->left.expr, E_EQUAL, sym);
scripts/kconfig/expr.c
797
e2 = expr_trans_compare(e->right.expr, E_EQUAL, sym);
scripts/kconfig/expr.c
799
e = expr_alloc_two(E_AND, e1, e2);
scripts/kconfig/expr.c
801
e = expr_alloc_two(E_OR, e1, e2);
scripts/kconfig/expr.c
803
e = expr_alloc_one(E_NOT, e);
scripts/kconfig/expr.c
804
return e;
scripts/kconfig/expr.c
806
e1 = expr_trans_compare(e->left.expr, E_EQUAL, sym);
scripts/kconfig/expr.c
807
e2 = expr_trans_compare(e->right.expr, E_EQUAL, sym);
scripts/kconfig/expr.c
809
e = expr_alloc_two(E_OR, e1, e2);
scripts/kconfig/expr.c
811
e = expr_alloc_two(E_AND, e1, e2);
scripts/kconfig/expr.c
813
e = expr_alloc_one(E_NOT, e);
scripts/kconfig/expr.c
814
return e;
scripts/kconfig/expr.c
816
return expr_trans_compare(e->left.expr, type == E_EQUAL ? E_UNEQUAL : E_EQUAL, sym);
scripts/kconfig/expr.c
825
return e;
scripts/kconfig/expr.c
829
return expr_alloc_one(E_NOT, e);
scripts/kconfig/expr.c
832
return expr_alloc_one(E_NOT, e);
scripts/kconfig/expr.c
836
return e;
scripts/kconfig/expr.c
840
return expr_alloc_comp(type, e->left.sym, sym);
scripts/kconfig/expr.c
891
static tristate __expr_calc_value(struct expr *e)
scripts/kconfig/expr.c
899
switch (e->type) {
scripts/kconfig/expr.c
901
sym_calc_value(e->left.sym);
scripts/kconfig/expr.c
902
return e->left.sym->curr.tri;
scripts/kconfig/expr.c
904
val1 = expr_calc_value(e->left.expr);
scripts/kconfig/expr.c
905
val2 = expr_calc_value(e->right.expr);
scripts/kconfig/expr.c
908
val1 = expr_calc_value(e->left.expr);
scripts/kconfig/expr.c
909
val2 = expr_calc_value(e->right.expr);
scripts/kconfig/expr.c
912
val1 = expr_calc_value(e->left.expr);
scripts/kconfig/expr.c
922
printf("expr_calc_value: %d?\n", e->type);
scripts/kconfig/expr.c
926
sym_calc_value(e->left.sym);
scripts/kconfig/expr.c
927
sym_calc_value(e->right.sym);
scripts/kconfig/expr.c
928
str1 = sym_get_string_value(e->left.sym);
scripts/kconfig/expr.c
929
str2 = sym_get_string_value(e->right.sym);
scripts/kconfig/expr.c
931
if (e->left.sym->type != S_STRING || e->right.sym->type != S_STRING) {
scripts/kconfig/expr.c
932
k1 = expr_parse_string(str1, e->left.sym->type, &lval);
scripts/kconfig/expr.c
933
k2 = expr_parse_string(str2, e->right.sym->type, &rval);
scripts/kconfig/expr.c
943
switch(e->type) {
scripts/kconfig/expr.c
957
printf("expr_calc_value: relation %d?\n", e->type);
scripts/kconfig/expr.c
967
tristate expr_calc_value(struct expr *e)
scripts/kconfig/expr.c
969
if (!e)
scripts/kconfig/expr.c
972
if (!e->val_is_valid) {
scripts/kconfig/expr.c
973
e->val = __expr_calc_value(e);
scripts/kconfig/expr.c
974
e->val_is_valid = true;
scripts/kconfig/expr.c
977
return e->val;
scripts/kconfig/expr.c
985
struct expr *e;
scripts/kconfig/expr.c
987
hash_for_each(expr_hashtable, e, node)
scripts/kconfig/expr.c
988
e->val_is_valid = false;
scripts/kconfig/expr.h
306
tristate expr_calc_value(struct expr *e);
scripts/kconfig/expr.h
307
struct expr *expr_eliminate_dups(struct expr *e);
scripts/kconfig/expr.h
308
struct expr *expr_transform(struct expr *e);
scripts/kconfig/expr.h
311
struct expr *expr_trans_compare(struct expr *e, enum expr_type type, struct symbol *sym);
scripts/kconfig/expr.h
313
void expr_fprint(struct expr *e, FILE *out);
scripts/kconfig/expr.h
315
void expr_gstr_print(const struct expr *e, struct gstr *gs);
scripts/kconfig/expr.h
316
void expr_gstr_print_revdep(struct expr *e, struct gstr *gs,
scripts/kconfig/expr.h
319
static inline bool expr_is_yes(const struct expr *e)
scripts/kconfig/expr.h
321
return !e || (e->type == E_SYMBOL && e->left.sym == &symbol_yes);
scripts/kconfig/lkc_proto.h
44
void expr_print(const struct expr *e,
scripts/kconfig/menu.c
104
static struct expr *rewrite_m(struct expr *e)
scripts/kconfig/menu.c
106
if (!e)
scripts/kconfig/menu.c
107
return e;
scripts/kconfig/menu.c
109
switch (e->type) {
scripts/kconfig/menu.c
111
e = expr_alloc_one(E_NOT, rewrite_m(e->left.expr));
scripts/kconfig/menu.c
115
e = expr_alloc_two(e->type,
scripts/kconfig/menu.c
116
rewrite_m(e->left.expr),
scripts/kconfig/menu.c
117
rewrite_m(e->right.expr));
scripts/kconfig/menu.c
121
if (e->left.sym == &symbol_mod)
scripts/kconfig/menu.c
122
return expr_alloc_and(e, expr_alloc_symbol(modules_sym));
scripts/kconfig/menu.c
127
return e;
scripts/kconfig/preprocess.c
102
e->name, e->value);
scripts/kconfig/preprocess.c
103
env_del(e);
scripts/kconfig/preprocess.c
48
struct env *e;
scripts/kconfig/preprocess.c
50
e = xmalloc(sizeof(*e));
scripts/kconfig/preprocess.c
51
e->name = xstrdup(name);
scripts/kconfig/preprocess.c
52
e->value = xstrdup(value);
scripts/kconfig/preprocess.c
54
list_add_tail(&e->node, &env_list);
scripts/kconfig/preprocess.c
57
static void env_del(struct env *e)
scripts/kconfig/preprocess.c
59
list_del(&e->node);
scripts/kconfig/preprocess.c
60
free(e->name);
scripts/kconfig/preprocess.c
61
free(e->value);
scripts/kconfig/preprocess.c
62
free(e);
scripts/kconfig/preprocess.c
68
struct env *e;
scripts/kconfig/preprocess.c
74
list_for_each_entry(e, &env_list, node) {
scripts/kconfig/preprocess.c
75
if (!strcmp(name, e->name))
scripts/kconfig/preprocess.c
76
return xstrdup(e->value);
scripts/kconfig/preprocess.c
94
struct env *e, *tmp;
scripts/kconfig/preprocess.c
96
list_for_each_entry_safe(e, tmp, &env_list, node) {
scripts/kconfig/qconf.cc
1687
void ConfigMainWindow::closeEvent(QCloseEvent* e)
scripts/kconfig/qconf.cc
1690
e->accept();
scripts/kconfig/qconf.cc
1711
e->accept();
scripts/kconfig/qconf.cc
1713
e->ignore();
scripts/kconfig/qconf.cc
1716
e->accept();
scripts/kconfig/qconf.cc
1719
e->ignore();
scripts/kconfig/qconf.cc
789
void ConfigList::mouseReleaseEvent(QMouseEvent* e)
scripts/kconfig/qconf.cc
791
QPoint p = e->pos();
scripts/kconfig/qconf.cc
832
Parent::mouseReleaseEvent(e);
scripts/kconfig/qconf.cc
835
void ConfigList::mouseDoubleClickEvent(QMouseEvent* e)
scripts/kconfig/qconf.cc
837
QPoint p = e->pos();
scripts/kconfig/qconf.cc
862
Parent::mouseDoubleClickEvent(e);
scripts/kconfig/qconf.cc
865
void ConfigList::focusInEvent(QFocusEvent *e)
scripts/kconfig/qconf.cc
869
Parent::focusInEvent(e);
scripts/kconfig/qconf.cc
879
void ConfigList::contextMenuEvent(QContextMenuEvent *e)
scripts/kconfig/qconf.cc
895
headerPopup->exec(e->globalPos());
scripts/kconfig/qconf.cc
896
e->accept();
scripts/kconfig/qconf.h
261
void closeEvent(QCloseEvent *e);
scripts/kconfig/qconf.h
58
void keyPressEvent(QKeyEvent *e);
scripts/kconfig/qconf.h
59
void mouseReleaseEvent(QMouseEvent *e);
scripts/kconfig/qconf.h
60
void mouseDoubleClickEvent(QMouseEvent *e);
scripts/kconfig/qconf.h
61
void focusInEvent(QFocusEvent *e);
scripts/kconfig/qconf.h
62
void contextMenuEvent(QContextMenuEvent *e);
scripts/kconfig/symbol.c
1179
static struct symbol *sym_check_expr_deps(const struct expr *e)
scripts/kconfig/symbol.c
1183
if (!e)
scripts/kconfig/symbol.c
1185
switch (e->type) {
scripts/kconfig/symbol.c
1188
sym = sym_check_expr_deps(e->left.expr);
scripts/kconfig/symbol.c
1191
return sym_check_expr_deps(e->right.expr);
scripts/kconfig/symbol.c
1193
return sym_check_expr_deps(e->left.expr);
scripts/kconfig/symbol.c
1200
sym = sym_check_deps(e->left.sym);
scripts/kconfig/symbol.c
1203
return sym_check_deps(e->right.sym);
scripts/kconfig/symbol.c
1205
return sym_check_deps(e->left.sym);
scripts/kconfig/symbol.c
1209
fprintf(stderr, "Oops! How to check %d?\n", e->type);
scripts/sign-file.c
119
ENGINE *e;
scripts/sign-file.c
123
e = ENGINE_by_id("pkcs11");
scripts/sign-file.c
124
ERR(!e, "Load PKCS#11 ENGINE");
scripts/sign-file.c
125
if (ENGINE_init(e))
scripts/sign-file.c
130
ERR(!ENGINE_ctrl_cmd_string(e, "PIN", key_pass, 0), "Set PKCS#11 PIN");
scripts/sign-file.c
131
private_key = ENGINE_load_private_key(e, private_key_name, NULL, NULL);
scripts/ssl-common.h
10
int e, line;
scripts/ssl-common.h
17
while ((e = ERR_peek_error_line(&file, &line))) {
scripts/ssl-common.h
18
ERR_error_string(e, buf);
security/apparmor/file.c
232
int e = 0;
security/apparmor/file.c
240
e = -EACCES;
security/apparmor/file.c
243
cond->uid, NULL, e);
security/apparmor/include/perms.h
158
int e, error = FN1; \
security/apparmor/include/perms.h
159
e = FN2; \
security/apparmor/include/perms.h
160
if (e) \
security/apparmor/include/perms.h
161
error = e; \
security/apparmor/include/policy_unpack.h
184
bool aa_inbounds(struct aa_ext *e, size_t size);
security/apparmor/include/policy_unpack.h
185
size_t aa_unpack_u16_chunk(struct aa_ext *e, char **chunk);
security/apparmor/include/policy_unpack.h
186
bool aa_unpack_X(struct aa_ext *e, enum aa_code code);
security/apparmor/include/policy_unpack.h
187
bool aa_unpack_nameX(struct aa_ext *e, enum aa_code code, const char *name);
security/apparmor/include/policy_unpack.h
188
bool aa_unpack_u32(struct aa_ext *e, u32 *data, const char *name);
security/apparmor/include/policy_unpack.h
189
bool aa_unpack_u64(struct aa_ext *e, u64 *data, const char *name);
security/apparmor/include/policy_unpack.h
190
bool aa_unpack_array(struct aa_ext *e, const char *name, u16 *size);
security/apparmor/include/policy_unpack.h
191
size_t aa_unpack_blob(struct aa_ext *e, char **blob, const char *name);
security/apparmor/include/policy_unpack.h
192
int aa_unpack_str(struct aa_ext *e, const char **string, const char *name);
security/apparmor/include/policy_unpack.h
193
int aa_unpack_strdup(struct aa_ext *e, char **string, const char *name);
security/apparmor/policy_unpack.c
1005
pdb->dfa = unpack_dfa(e, flags);
security/apparmor/policy_unpack.c
1023
if (!aa_unpack_u32(e, &pdb->start[0], "start"))
security/apparmor/policy_unpack.c
1026
if (!aa_unpack_u32(e, &pdb->start[AA_CLASS_FILE], "dfa_start")) {
security/apparmor/policy_unpack.c
1066
error = unpack_strs_table(e, "xtable", false, &pdb->trans);
security/apparmor/policy_unpack.c
1087
e->pos = pos;
security/apparmor/policy_unpack.c
1113
static struct aa_profile *unpack_profile(struct aa_ext *e, char **ns_name)
security/apparmor/policy_unpack.c
1130
if (!aa_unpack_nameX(e, AA_STRUCT, "profile"))
security/apparmor/policy_unpack.c
1132
if (!aa_unpack_str(e, &name, NULL))
security/apparmor/policy_unpack.c
1161
(void) aa_unpack_str(e, &profile->rename, "rename");
security/apparmor/policy_unpack.c
1164
(void) aa_unpack_str(e, &profile->attach.xmatch_str, "attach");
security/apparmor/policy_unpack.c
1167
error = unpack_pdb(e, &profile->attach.xmatch, false, false, &info);
security/apparmor/policy_unpack.c
1175
if (!aa_unpack_u32(e, &tmp, NULL)) {
security/apparmor/policy_unpack.c
1191
(void) aa_unpack_strdup(e, &disconnected, "disconnected");
security/apparmor/policy_unpack.c
1195
(void) aa_unpack_u32(e, &profile->signal, "kill");
security/apparmor/policy_unpack.c
1201
if (!aa_unpack_nameX(e, AA_STRUCT, "flags")) {
security/apparmor/policy_unpack.c
1206
if (!aa_unpack_u32(e, &tmp, NULL))
security/apparmor/policy_unpack.c
1214
if (!aa_unpack_u32(e, &tmp, NULL))
security/apparmor/policy_unpack.c
1216
if (tmp == PACKED_MODE_COMPLAIN || (e->version & FORCE_COMPLAIN_FLAG)) {
security/apparmor/policy_unpack.c
1230
if (!aa_unpack_u32(e, &tmp, NULL))
security/apparmor/policy_unpack.c
1235
if (!aa_unpack_nameX(e, AA_STRUCTEND, NULL))
security/apparmor/policy_unpack.c
1239
if (aa_unpack_u32(e, &profile->path_flags, "path_flags"))
security/apparmor/policy_unpack.c
1247
if (!aa_unpack_cap_low(e, &rules->caps.allow, NULL))
security/apparmor/policy_unpack.c
1249
if (!aa_unpack_cap_low(e, &rules->caps.audit, NULL))
security/apparmor/policy_unpack.c
1251
if (!aa_unpack_cap_low(e, &rules->caps.quiet, NULL))
security/apparmor/policy_unpack.c
1253
if (!aa_unpack_cap_low(e, &tmpcap, NULL))
security/apparmor/policy_unpack.c
1257
if (aa_unpack_nameX(e, AA_STRUCT, "caps64")) {
security/apparmor/policy_unpack.c
1259
if (!aa_unpack_cap_high(e, &rules->caps.allow, NULL))
security/apparmor/policy_unpack.c
1261
if (!aa_unpack_cap_high(e, &rules->caps.audit, NULL))
security/apparmor/policy_unpack.c
1263
if (!aa_unpack_cap_high(e, &rules->caps.quiet, NULL))
security/apparmor/policy_unpack.c
1265
if (!aa_unpack_cap_high(e, &tmpcap, NULL))
security/apparmor/policy_unpack.c
1267
if (!aa_unpack_nameX(e, AA_STRUCTEND, NULL))
security/apparmor/policy_unpack.c
1272
if (aa_unpack_nameX(e, AA_STRUCT, "capsx")) {
security/apparmor/policy_unpack.c
1274
if (!aa_unpack_cap_low(e, &rules->caps.extended, NULL))
security/apparmor/policy_unpack.c
1276
if (!aa_unpack_cap_high(e, &rules->caps.extended, NULL))
security/apparmor/policy_unpack.c
1278
if (!aa_unpack_nameX(e, AA_STRUCTEND, NULL))
security/apparmor/policy_unpack.c
1282
if (!unpack_xattrs(e, profile)) {
security/apparmor/policy_unpack.c
1287
if (!unpack_rlimits(e, rules)) {
security/apparmor/policy_unpack.c
1292
if (!unpack_secmark(e, rules)) {
security/apparmor/policy_unpack.c
1297
if (aa_unpack_nameX(e, AA_STRUCT, "policydb")) {
security/apparmor/policy_unpack.c
1300
error = unpack_pdb(e, &rules->policy, true, false,
security/apparmor/policy_unpack.c
1311
if (!aa_unpack_nameX(e, AA_STRUCTEND, NULL))
security/apparmor/policy_unpack.c
1315
e->version);
security/apparmor/policy_unpack.c
1325
error = unpack_pdb(e, &rules->file, false, true, &info);
security/apparmor/policy_unpack.c
1346
if (aa_unpack_nameX(e, AA_STRUCT, "data")) {
security/apparmor/policy_unpack.c
1365
while (aa_unpack_strdup(e, &key, NULL)) {
security/apparmor/policy_unpack.c
1374
data->size = aa_unpack_blob(e, &data->data, NULL);
security/apparmor/policy_unpack.c
1393
if (!aa_unpack_nameX(e, AA_STRUCTEND, NULL)) {
security/apparmor/policy_unpack.c
1399
if (!aa_unpack_nameX(e, AA_STRUCTEND, NULL)) {
security/apparmor/policy_unpack.c
1420
audit_iface(profile, NULL, name, info, e, error);
security/apparmor/policy_unpack.c
1434
static int verify_header(struct aa_ext *e, int required, const char **ns)
security/apparmor/policy_unpack.c
1440
if (!aa_unpack_u32(e, &e->version, "version")) {
security/apparmor/policy_unpack.c
1443
e, error);
security/apparmor/policy_unpack.c
1452
if (VERSION_LT(e->version, v5) || VERSION_GT(e->version, v9)) {
security/apparmor/policy_unpack.c
1454
e, error);
security/apparmor/policy_unpack.c
1459
if (aa_unpack_str(e, &name, "namespace")) {
security/apparmor/policy_unpack.c
1462
e, error);
security/apparmor/policy_unpack.c
1466
audit_iface(NULL, NULL, NULL, "invalid ns change", e,
security/apparmor/policy_unpack.c
1732
struct aa_ext e = {
security/apparmor/policy_unpack.c
1739
while (e.pos < e.end) {
security/apparmor/policy_unpack.c
1741
error = verify_header(&e, e.pos == e.start, ns);
security/apparmor/policy_unpack.c
1745
start = e.pos;
security/apparmor/policy_unpack.c
1746
profile = unpack_profile(&e, &ns_name);
security/apparmor/policy_unpack.c
1757
error = aa_calc_profile_hash(profile, e.version, start,
security/apparmor/policy_unpack.c
1758
e.pos - start);
security/apparmor/policy_unpack.c
1773
udata->abi = e.version & K_ABI_MASK;
security/apparmor/policy_unpack.c
179
VISIBLE_IF_KUNIT bool aa_inbounds(struct aa_ext *e, size_t size)
security/apparmor/policy_unpack.c
181
return (size <= e->end - e->pos);
security/apparmor/policy_unpack.c
192
VISIBLE_IF_KUNIT size_t aa_unpack_u16_chunk(struct aa_ext *e, char **chunk)
security/apparmor/policy_unpack.c
195
void *pos = e->pos;
security/apparmor/policy_unpack.c
197
if (!aa_inbounds(e, sizeof(u16)))
security/apparmor/policy_unpack.c
199
size = le16_to_cpu(get_unaligned((__le16 *) e->pos));
security/apparmor/policy_unpack.c
200
e->pos += sizeof(__le16);
security/apparmor/policy_unpack.c
201
if (!aa_inbounds(e, size))
security/apparmor/policy_unpack.c
203
*chunk = e->pos;
security/apparmor/policy_unpack.c
204
e->pos += size;
security/apparmor/policy_unpack.c
208
e->pos = pos;
security/apparmor/policy_unpack.c
214
VISIBLE_IF_KUNIT bool aa_unpack_X(struct aa_ext *e, enum aa_code code)
security/apparmor/policy_unpack.c
216
if (!aa_inbounds(e, 1))
security/apparmor/policy_unpack.c
218
if (*(u8 *) e->pos != code)
security/apparmor/policy_unpack.c
220
e->pos++;
security/apparmor/policy_unpack.c
241
VISIBLE_IF_KUNIT bool aa_unpack_nameX(struct aa_ext *e, enum aa_code code, const char *name)
security/apparmor/policy_unpack.c
246
void *pos = e->pos;
security/apparmor/policy_unpack.c
251
if (aa_unpack_X(e, AA_NAME)) {
security/apparmor/policy_unpack.c
253
size_t size = aa_unpack_u16_chunk(e, &tag);
security/apparmor/policy_unpack.c
263
if (aa_unpack_X(e, code))
security/apparmor/policy_unpack.c
267
e->pos = pos;
security/apparmor/policy_unpack.c
272
static bool unpack_u8(struct aa_ext *e, u8 *data, const char *name)
security/apparmor/policy_unpack.c
274
void *pos = e->pos;
security/apparmor/policy_unpack.c
276
if (aa_unpack_nameX(e, AA_U8, name)) {
security/apparmor/policy_unpack.c
277
if (!aa_inbounds(e, sizeof(u8)))
security/apparmor/policy_unpack.c
280
*data = *((u8 *)e->pos);
security/apparmor/policy_unpack.c
281
e->pos += sizeof(u8);
security/apparmor/policy_unpack.c
286
e->pos = pos;
security/apparmor/policy_unpack.c
290
VISIBLE_IF_KUNIT bool aa_unpack_u32(struct aa_ext *e, u32 *data, const char *name)
security/apparmor/policy_unpack.c
292
void *pos = e->pos;
security/apparmor/policy_unpack.c
294
if (aa_unpack_nameX(e, AA_U32, name)) {
security/apparmor/policy_unpack.c
295
if (!aa_inbounds(e, sizeof(u32)))
security/apparmor/policy_unpack.c
298
*data = le32_to_cpu(get_unaligned((__le32 *) e->pos));
security/apparmor/policy_unpack.c
299
e->pos += sizeof(u32);
security/apparmor/policy_unpack.c
304
e->pos = pos;
security/apparmor/policy_unpack.c
309
VISIBLE_IF_KUNIT bool aa_unpack_u64(struct aa_ext *e, u64 *data, const char *name)
security/apparmor/policy_unpack.c
311
void *pos = e->pos;
security/apparmor/policy_unpack.c
313
if (aa_unpack_nameX(e, AA_U64, name)) {
security/apparmor/policy_unpack.c
314
if (!aa_inbounds(e, sizeof(u64)))
security/apparmor/policy_unpack.c
317
*data = le64_to_cpu(get_unaligned((__le64 *) e->pos));
security/apparmor/policy_unpack.c
318
e->pos += sizeof(u64);
security/apparmor/policy_unpack.c
323
e->pos = pos;
security/apparmor/policy_unpack.c
328
static bool aa_unpack_cap_low(struct aa_ext *e, kernel_cap_t *data, const char *name)
security/apparmor/policy_unpack.c
332
if (!aa_unpack_u32(e, &val, name))
security/apparmor/policy_unpack.c
338
static bool aa_unpack_cap_high(struct aa_ext *e, kernel_cap_t *data, const char *name)
security/apparmor/policy_unpack.c
342
if (!aa_unpack_u32(e, &val, name))
security/apparmor/policy_unpack.c
348
VISIBLE_IF_KUNIT bool aa_unpack_array(struct aa_ext *e, const char *name, u16 *size)
security/apparmor/policy_unpack.c
350
void *pos = e->pos;
security/apparmor/policy_unpack.c
352
if (aa_unpack_nameX(e, AA_ARRAY, name)) {
security/apparmor/policy_unpack.c
353
if (!aa_inbounds(e, sizeof(u16)))
security/apparmor/policy_unpack.c
355
*size = le16_to_cpu(get_unaligned((__le16 *) e->pos));
security/apparmor/policy_unpack.c
356
e->pos += sizeof(u16);
security/apparmor/policy_unpack.c
361
e->pos = pos;
security/apparmor/policy_unpack.c
366
VISIBLE_IF_KUNIT size_t aa_unpack_blob(struct aa_ext *e, char **blob, const char *name)
security/apparmor/policy_unpack.c
368
void *pos = e->pos;
security/apparmor/policy_unpack.c
370
if (aa_unpack_nameX(e, AA_BLOB, name)) {
security/apparmor/policy_unpack.c
372
if (!aa_inbounds(e, sizeof(u32)))
security/apparmor/policy_unpack.c
374
size = le32_to_cpu(get_unaligned((__le32 *) e->pos));
security/apparmor/policy_unpack.c
375
e->pos += sizeof(u32);
security/apparmor/policy_unpack.c
376
if (aa_inbounds(e, (size_t) size)) {
security/apparmor/policy_unpack.c
377
*blob = e->pos;
security/apparmor/policy_unpack.c
378
e->pos += size;
security/apparmor/policy_unpack.c
384
e->pos = pos;
security/apparmor/policy_unpack.c
389
VISIBLE_IF_KUNIT int aa_unpack_str(struct aa_ext *e, const char **string, const char *name)
security/apparmor/policy_unpack.c
393
void *pos = e->pos;
security/apparmor/policy_unpack.c
395
if (aa_unpack_nameX(e, AA_STRING, name)) {
security/apparmor/policy_unpack.c
396
size = aa_unpack_u16_chunk(e, &src_str);
security/apparmor/policy_unpack.c
408
e->pos = pos;
security/apparmor/policy_unpack.c
413
VISIBLE_IF_KUNIT int aa_unpack_strdup(struct aa_ext *e, char **string, const char *name)
security/apparmor/policy_unpack.c
416
void *pos = e->pos;
security/apparmor/policy_unpack.c
417
int res = aa_unpack_str(e, &tmp, name);
security/apparmor/policy_unpack.c
425
e->pos = pos;
security/apparmor/policy_unpack.c
441
static struct aa_dfa *unpack_dfa(struct aa_ext *e, int flags)
security/apparmor/policy_unpack.c
447
size = aa_unpack_blob(e, &blob, "aadfa");
security/apparmor/policy_unpack.c
454
size_t sz = blob - (char *) e->start -
security/apparmor/policy_unpack.c
455
((e->pos - e->start) & 7);
security/apparmor/policy_unpack.c
527
static int unpack_strs_table(struct aa_ext *e, const char *name, bool multi,
security/apparmor/policy_unpack.c
530
void *saved_pos = e->pos;
security/apparmor/policy_unpack.c
535
if (aa_unpack_nameX(e, AA_STRUCT, name)) {
security/apparmor/policy_unpack.c
539
if (!aa_unpack_array(e, NULL, &size))
security/apparmor/policy_unpack.c
556
int c, size2 = aa_unpack_strdup(e, &str, NULL);
security/apparmor/policy_unpack.c
564
(unsigned long)(e->pos - saved_pos));
security/apparmor/policy_unpack.c
576
if (!aa_unpack_nameX(e, AA_ARRAYEND, NULL))
security/apparmor/policy_unpack.c
578
if (!aa_unpack_nameX(e, AA_STRUCTEND, NULL))
security/apparmor/policy_unpack.c
585
e->pos = saved_pos;
security/apparmor/policy_unpack.c
589
static bool unpack_xattrs(struct aa_ext *e, struct aa_profile *profile)
security/apparmor/policy_unpack.c
591
void *pos = e->pos;
security/apparmor/policy_unpack.c
593
if (aa_unpack_nameX(e, AA_STRUCT, "xattrs")) {
security/apparmor/policy_unpack.c
597
if (!aa_unpack_array(e, NULL, &size))
security/apparmor/policy_unpack.c
604
if (!aa_unpack_strdup(e, &profile->attach.xattrs[i], NULL))
security/apparmor/policy_unpack.c
607
if (!aa_unpack_nameX(e, AA_ARRAYEND, NULL))
security/apparmor/policy_unpack.c
609
if (!aa_unpack_nameX(e, AA_STRUCTEND, NULL))
security/apparmor/policy_unpack.c
616
e->pos = pos;
security/apparmor/policy_unpack.c
620
static bool unpack_secmark(struct aa_ext *e, struct aa_ruleset *rules)
security/apparmor/policy_unpack.c
622
void *pos = e->pos;
security/apparmor/policy_unpack.c
626
if (aa_unpack_nameX(e, AA_STRUCT, "secmark")) {
security/apparmor/policy_unpack.c
627
if (!aa_unpack_array(e, NULL, &size))
security/apparmor/policy_unpack.c
637
if (!unpack_u8(e, &rules->secmark[i].audit, NULL))
security/apparmor/policy_unpack.c
639
if (!unpack_u8(e, &rules->secmark[i].deny, NULL))
security/apparmor/policy_unpack.c
64
const char *name, const char *info, struct aa_ext *e,
security/apparmor/policy_unpack.c
641
if (!aa_unpack_strdup(e, &rules->secmark[i].label, NULL))
security/apparmor/policy_unpack.c
644
if (!aa_unpack_nameX(e, AA_ARRAYEND, NULL))
security/apparmor/policy_unpack.c
646
if (!aa_unpack_nameX(e, AA_STRUCTEND, NULL))
security/apparmor/policy_unpack.c
661
e->pos = pos;
security/apparmor/policy_unpack.c
665
static bool unpack_rlimits(struct aa_ext *e, struct aa_ruleset *rules)
security/apparmor/policy_unpack.c
667
void *pos = e->pos;
security/apparmor/policy_unpack.c
670
if (aa_unpack_nameX(e, AA_STRUCT, "rlimits")) {
security/apparmor/policy_unpack.c
674
if (!aa_unpack_u32(e, &tmp, NULL))
security/apparmor/policy_unpack.c
678
if (!aa_unpack_array(e, NULL, &size) ||
security/apparmor/policy_unpack.c
684
if (!aa_unpack_u64(e, &tmp2, NULL))
security/apparmor/policy_unpack.c
688
if (!aa_unpack_nameX(e, AA_ARRAYEND, NULL))
security/apparmor/policy_unpack.c
69
if (e)
security/apparmor/policy_unpack.c
690
if (!aa_unpack_nameX(e, AA_STRUCTEND, NULL))
security/apparmor/policy_unpack.c
696
e->pos = pos;
security/apparmor/policy_unpack.c
70
ad.iface.pos = e->pos - e->start;
security/apparmor/policy_unpack.c
777
static int unpack_tagsets(struct aa_ext *e, struct aa_tags_struct *tags)
security/apparmor/policy_unpack.c
782
void *pos = e->pos;
security/apparmor/policy_unpack.c
784
if (!aa_unpack_array(e, "sets", &size))
security/apparmor/policy_unpack.c
792
if (!aa_unpack_u32(e, &sets[i], NULL))
security/apparmor/policy_unpack.c
795
if (!aa_unpack_nameX(e, AA_ARRAYEND, NULL))
security/apparmor/policy_unpack.c
806
e->pos = pos;
security/apparmor/policy_unpack.c
810
static bool unpack_tag_header_ent(struct aa_ext *e, struct aa_tags_header *h)
security/apparmor/policy_unpack.c
812
return aa_unpack_u32(e, &h->mask, NULL) &&
security/apparmor/policy_unpack.c
813
aa_unpack_u32(e, &h->count, NULL) &&
security/apparmor/policy_unpack.c
814
aa_unpack_u32(e, &h->size, NULL) &&
security/apparmor/policy_unpack.c
815
aa_unpack_u32(e, &h->tags, NULL);
security/apparmor/policy_unpack.c
818
static int unpack_tag_headers(struct aa_ext *e, struct aa_tags_struct *tags)
security/apparmor/policy_unpack.c
823
void *pos = e->pos;
security/apparmor/policy_unpack.c
825
if (!aa_unpack_array(e, "hdrs", &size))
security/apparmor/policy_unpack.c
833
if (!unpack_tag_header_ent(e, &hdrs[i]))
security/apparmor/policy_unpack.c
836
if (!aa_unpack_nameX(e, AA_ARRAYEND, NULL))
security/apparmor/policy_unpack.c
847
e->pos = pos;
security/apparmor/policy_unpack.c
852
static int unpack_tags(struct aa_ext *e, struct aa_tags_struct *tags,
security/apparmor/policy_unpack.c
856
void *pos = e->pos;
security/apparmor/policy_unpack.c
860
if (aa_unpack_nameX(e, AA_STRUCT, "tags")) {
security/apparmor/policy_unpack.c
863
if (!aa_unpack_u32(e, &version, "version") || version != 1) {
security/apparmor/policy_unpack.c
867
error = unpack_strs_table(e, "strs", true, &tags->strs);
security/apparmor/policy_unpack.c
872
error = unpack_tag_headers(e, tags);
security/apparmor/policy_unpack.c
877
error = unpack_tagsets(e, tags);
security/apparmor/policy_unpack.c
882
if (!aa_unpack_nameX(e, AA_STRUCTEND, NULL))
security/apparmor/policy_unpack.c
894
e->pos = pos;
security/apparmor/policy_unpack.c
898
static bool unpack_perm(struct aa_ext *e, u32 version, struct aa_perms *perm)
security/apparmor/policy_unpack.c
906
return aa_unpack_u32(e, &reserved, NULL) &&
security/apparmor/policy_unpack.c
907
aa_unpack_u32(e, &perm->allow, NULL) &&
security/apparmor/policy_unpack.c
908
aa_unpack_u32(e, &perm->deny, NULL) &&
security/apparmor/policy_unpack.c
909
aa_unpack_u32(e, &perm->subtree, NULL) &&
security/apparmor/policy_unpack.c
910
aa_unpack_u32(e, &perm->cond, NULL) &&
security/apparmor/policy_unpack.c
911
aa_unpack_u32(e, &perm->kill, NULL) &&
security/apparmor/policy_unpack.c
912
aa_unpack_u32(e, &perm->complain, NULL) &&
security/apparmor/policy_unpack.c
913
aa_unpack_u32(e, &perm->prompt, NULL) &&
security/apparmor/policy_unpack.c
914
aa_unpack_u32(e, &perm->audit, NULL) &&
security/apparmor/policy_unpack.c
915
aa_unpack_u32(e, &perm->quiet, NULL) &&
security/apparmor/policy_unpack.c
916
aa_unpack_u32(e, &perm->hide, NULL) &&
security/apparmor/policy_unpack.c
917
aa_unpack_u32(e, &perm->xindex, NULL) &&
security/apparmor/policy_unpack.c
918
aa_unpack_u32(e, &perm->tag, NULL) &&
security/apparmor/policy_unpack.c
919
aa_unpack_u32(e, &perm->label, NULL);
security/apparmor/policy_unpack.c
922
static ssize_t unpack_perms_table(struct aa_ext *e, struct aa_perms **perms)
security/apparmor/policy_unpack.c
924
void *pos = e->pos;
security/apparmor/policy_unpack.c
932
if (aa_unpack_nameX(e, AA_STRUCT, "perms")) {
security/apparmor/policy_unpack.c
936
if (!aa_unpack_u32(e, &version, "version"))
security/apparmor/policy_unpack.c
938
if (!aa_unpack_array(e, NULL, &size))
security/apparmor/policy_unpack.c
942
e->pos = pos;
security/apparmor/policy_unpack.c
946
if (!unpack_perm(e, version, &(*perms)[i]))
security/apparmor/policy_unpack.c
949
if (!aa_unpack_nameX(e, AA_ARRAYEND, NULL))
security/apparmor/policy_unpack.c
951
if (!aa_unpack_nameX(e, AA_STRUCTEND, NULL))
security/apparmor/policy_unpack.c
961
e->pos = pos;
security/apparmor/policy_unpack.c
965
static int unpack_pdb(struct aa_ext *e, struct aa_policydb **policy,
security/apparmor/policy_unpack.c
970
void *pos = e->pos;
security/apparmor/policy_unpack.c
980
if (unpack_tags(e, &pdb->tags, info) < 0)
security/apparmor/policy_unpack.c
984
size = unpack_perms_table(e, &pdb->perms);
security/apparmor/policy_unpack.c
996
if (aa_unpack_u32(e, &version, "permsv") && version > 2)
security/apparmor/policy_unpack_test.c
103
buf = e->start + TEST_NAMED_ARRAY_BUF_OFFSET;
security/apparmor/policy_unpack_test.c
106
strscpy(buf + 3, TEST_ARRAY_NAME, e->end - (void *)(buf + 3));
security/apparmor/policy_unpack_test.c
110
return e;
security/apparmor/policy_unpack_test.c
122
puf->e = build_aa_ext_struct(puf, test, e_size);
security/apparmor/policy_unpack_test.c
132
KUNIT_EXPECT_TRUE(test, aa_inbounds(puf->e, 0));
security/apparmor/policy_unpack_test.c
133
KUNIT_EXPECT_TRUE(test, aa_inbounds(puf->e, puf->e_size / 2));
security/apparmor/policy_unpack_test.c
134
KUNIT_EXPECT_TRUE(test, aa_inbounds(puf->e, puf->e_size));
security/apparmor/policy_unpack_test.c
141
KUNIT_EXPECT_FALSE(test, aa_inbounds(puf->e, puf->e_size + 1));
security/apparmor/policy_unpack_test.c
149
puf->e->pos += TEST_ARRAY_BUF_OFFSET;
security/apparmor/policy_unpack_test.c
151
KUNIT_EXPECT_TRUE(test, aa_unpack_array(puf->e, NULL, &array_size));
security/apparmor/policy_unpack_test.c
153
KUNIT_EXPECT_PTR_EQ(test, puf->e->pos,
security/apparmor/policy_unpack_test.c
154
puf->e->start + TEST_ARRAY_BUF_OFFSET + sizeof(u16) + 1);
security/apparmor/policy_unpack_test.c
163
puf->e->pos += TEST_NAMED_ARRAY_BUF_OFFSET;
security/apparmor/policy_unpack_test.c
165
KUNIT_EXPECT_TRUE(test, aa_unpack_array(puf->e, name, &array_size));
security/apparmor/policy_unpack_test.c
167
KUNIT_EXPECT_PTR_EQ(test, puf->e->pos,
security/apparmor/policy_unpack_test.c
168
puf->e->start + TEST_ARRAY_BUF_OFFSET + sizeof(u16) + 1);
security/apparmor/policy_unpack_test.c
177
puf->e->pos += TEST_NAMED_ARRAY_BUF_OFFSET;
security/apparmor/policy_unpack_test.c
178
puf->e->end = puf->e->start + TEST_ARRAY_BUF_OFFSET + sizeof(u16);
security/apparmor/policy_unpack_test.c
180
KUNIT_EXPECT_FALSE(test, aa_unpack_array(puf->e, name, &array_size));
security/apparmor/policy_unpack_test.c
181
KUNIT_EXPECT_PTR_EQ(test, puf->e->pos,
security/apparmor/policy_unpack_test.c
182
puf->e->start + TEST_NAMED_ARRAY_BUF_OFFSET);
security/apparmor/policy_unpack_test.c
191
puf->e->pos += TEST_BLOB_BUF_OFFSET;
security/apparmor/policy_unpack_test.c
192
size = aa_unpack_blob(puf->e, &blob, NULL);
security/apparmor/policy_unpack_test.c
205
puf->e->pos += TEST_NAMED_BLOB_BUF_OFFSET;
security/apparmor/policy_unpack_test.c
206
size = aa_unpack_blob(puf->e, &blob, TEST_BLOB_NAME);
security/apparmor/policy_unpack_test.c
220
puf->e->pos += TEST_NAMED_BLOB_BUF_OFFSET;
security/apparmor/policy_unpack_test.c
221
start = puf->e->pos;
security/apparmor/policy_unpack_test.c
222
puf->e->end = puf->e->start + TEST_BLOB_BUF_OFFSET
security/apparmor/policy_unpack_test.c
225
size = aa_unpack_blob(puf->e, &blob, TEST_BLOB_NAME);
security/apparmor/policy_unpack_test.c
228
KUNIT_EXPECT_PTR_EQ(test, puf->e->pos, start);
security/apparmor/policy_unpack_test.c
237
puf->e->pos += TEST_STRING_BUF_OFFSET;
security/apparmor/policy_unpack_test.c
238
size = aa_unpack_str(puf->e, &string, NULL);
security/apparmor/policy_unpack_test.c
250
size = aa_unpack_str(puf->e, &string, TEST_STRING_NAME);
security/apparmor/policy_unpack_test.c
260
void *start = puf->e->pos;
security/apparmor/policy_unpack_test.c
263
puf->e->end = puf->e->pos + TEST_STRING_BUF_OFFSET
security/apparmor/policy_unpack_test.c
266
size = aa_unpack_str(puf->e, &string, TEST_STRING_NAME);
security/apparmor/policy_unpack_test.c
269
KUNIT_EXPECT_PTR_EQ(test, puf->e->pos, start);
security/apparmor/policy_unpack_test.c
278
puf->e->pos += TEST_STRING_BUF_OFFSET;
security/apparmor/policy_unpack_test.c
279
size = aa_unpack_strdup(puf->e, &string, NULL);
security/apparmor/policy_unpack_test.c
283
((uintptr_t)puf->e->start <= (uintptr_t)string)
security/apparmor/policy_unpack_test.c
284
&& ((uintptr_t)string <= (uintptr_t)puf->e->end));
security/apparmor/policy_unpack_test.c
296
size = aa_unpack_strdup(puf->e, &string, TEST_STRING_NAME);
security/apparmor/policy_unpack_test.c
300
((uintptr_t)puf->e->start <= (uintptr_t)string)
security/apparmor/policy_unpack_test.c
301
&& ((uintptr_t)string <= (uintptr_t)puf->e->end));
security/apparmor/policy_unpack_test.c
310
void *start = puf->e->pos;
security/apparmor/policy_unpack_test.c
314
puf->e->end = puf->e->pos + TEST_STRING_BUF_OFFSET
security/apparmor/policy_unpack_test.c
317
size = aa_unpack_strdup(puf->e, &string, TEST_STRING_NAME);
security/apparmor/policy_unpack_test.c
321
KUNIT_EXPECT_PTR_EQ(test, puf->e->pos, start);
security/apparmor/policy_unpack_test.c
331
puf->e->pos += TEST_U32_BUF_OFFSET;
security/apparmor/policy_unpack_test.c
333
success = aa_unpack_nameX(puf->e, AA_U32, NULL);
security/apparmor/policy_unpack_test.c
336
KUNIT_EXPECT_PTR_EQ(test, puf->e->pos,
security/apparmor/policy_unpack_test.c
337
puf->e->start + TEST_U32_BUF_OFFSET + 1);
security/apparmor/policy_unpack_test.c
345
puf->e->pos += TEST_U32_BUF_OFFSET;
security/apparmor/policy_unpack_test.c
347
success = aa_unpack_nameX(puf->e, AA_BLOB, NULL);
security/apparmor/policy_unpack_test.c
350
KUNIT_EXPECT_PTR_EQ(test, puf->e->pos,
security/apparmor/policy_unpack_test.c
351
puf->e->start + TEST_U32_BUF_OFFSET);
security/apparmor/policy_unpack_test.c
360
puf->e->pos += TEST_NAMED_U32_BUF_OFFSET;
security/apparmor/policy_unpack_test.c
362
success = aa_unpack_nameX(puf->e, AA_U32, name);
security/apparmor/policy_unpack_test.c
365
KUNIT_EXPECT_PTR_EQ(test, puf->e->pos,
security/apparmor/policy_unpack_test.c
366
puf->e->start + TEST_U32_BUF_OFFSET + 1);
security/apparmor/policy_unpack_test.c
375
puf->e->pos += TEST_NAMED_U32_BUF_OFFSET;
security/apparmor/policy_unpack_test.c
377
success = aa_unpack_nameX(puf->e, AA_U32, name);
security/apparmor/policy_unpack_test.c
380
KUNIT_EXPECT_PTR_EQ(test, puf->e->pos,
security/apparmor/policy_unpack_test.c
381
puf->e->start + TEST_NAMED_U32_BUF_OFFSET);
security/apparmor/policy_unpack_test.c
390
puf->e->pos += TEST_U16_OFFSET;
security/apparmor/policy_unpack_test.c
396
puf->e->end += TEST_U16_DATA;
security/apparmor/policy_unpack_test.c
398
size = aa_unpack_u16_chunk(puf->e, &chunk);
security/apparmor/policy_unpack_test.c
401
puf->e->start + TEST_U16_OFFSET + 2);
security/apparmor/policy_unpack_test.c
403
KUNIT_EXPECT_PTR_EQ(test, puf->e->pos, (chunk + TEST_U16_DATA));
security/apparmor/policy_unpack_test.c
413
puf->e->pos = puf->e->end - 1;
security/apparmor/policy_unpack_test.c
415
size = aa_unpack_u16_chunk(puf->e, &chunk);
security/apparmor/policy_unpack_test.c
419
KUNIT_EXPECT_PTR_EQ(test, puf->e->pos, puf->e->end - 1);
security/apparmor/policy_unpack_test.c
429
puf->e->pos += TEST_U16_OFFSET;
security/apparmor/policy_unpack_test.c
435
puf->e->end = puf->e->pos + TEST_U16_DATA - 1;
security/apparmor/policy_unpack_test.c
437
size = aa_unpack_u16_chunk(puf->e, &chunk);
security/apparmor/policy_unpack_test.c
441
KUNIT_EXPECT_PTR_EQ(test, puf->e->pos, puf->e->start + TEST_U16_OFFSET);
security/apparmor/policy_unpack_test.c
450
puf->e->pos += TEST_U32_BUF_OFFSET;
security/apparmor/policy_unpack_test.c
452
success = aa_unpack_u32(puf->e, &data, NULL);
security/apparmor/policy_unpack_test.c
456
KUNIT_EXPECT_PTR_EQ(test, puf->e->pos,
security/apparmor/policy_unpack_test.c
457
puf->e->start + TEST_U32_BUF_OFFSET + sizeof(u32) + 1);
security/apparmor/policy_unpack_test.c
467
puf->e->pos += TEST_NAMED_U32_BUF_OFFSET;
security/apparmor/policy_unpack_test.c
469
success = aa_unpack_u32(puf->e, &data, name);
security/apparmor/policy_unpack_test.c
473
KUNIT_EXPECT_PTR_EQ(test, puf->e->pos,
security/apparmor/policy_unpack_test.c
474
puf->e->start + TEST_U32_BUF_OFFSET + sizeof(u32) + 1);
security/apparmor/policy_unpack_test.c
484
puf->e->pos += TEST_NAMED_U32_BUF_OFFSET;
security/apparmor/policy_unpack_test.c
485
puf->e->end = puf->e->start + TEST_U32_BUF_OFFSET + sizeof(u32);
security/apparmor/policy_unpack_test.c
487
success = aa_unpack_u32(puf->e, &data, name);
security/apparmor/policy_unpack_test.c
490
KUNIT_EXPECT_PTR_EQ(test, puf->e->pos,
security/apparmor/policy_unpack_test.c
491
puf->e->start + TEST_NAMED_U32_BUF_OFFSET);
security/apparmor/policy_unpack_test.c
500
puf->e->pos += TEST_U64_BUF_OFFSET;
security/apparmor/policy_unpack_test.c
502
success = aa_unpack_u64(puf->e, &data, NULL);
security/apparmor/policy_unpack_test.c
506
KUNIT_EXPECT_PTR_EQ(test, puf->e->pos,
security/apparmor/policy_unpack_test.c
507
puf->e->start + TEST_U64_BUF_OFFSET + sizeof(u64) + 1);
security/apparmor/policy_unpack_test.c
517
puf->e->pos += TEST_NAMED_U64_BUF_OFFSET;
security/apparmor/policy_unpack_test.c
519
success = aa_unpack_u64(puf->e, &data, name);
security/apparmor/policy_unpack_test.c
52
struct aa_ext *e;
security/apparmor/policy_unpack_test.c
523
KUNIT_EXPECT_PTR_EQ(test, puf->e->pos,
security/apparmor/policy_unpack_test.c
524
puf->e->start + TEST_U64_BUF_OFFSET + sizeof(u64) + 1);
security/apparmor/policy_unpack_test.c
534
puf->e->pos += TEST_NAMED_U64_BUF_OFFSET;
security/apparmor/policy_unpack_test.c
535
puf->e->end = puf->e->start + TEST_U64_BUF_OFFSET + sizeof(u64);
security/apparmor/policy_unpack_test.c
537
success = aa_unpack_u64(puf->e, &data, name);
security/apparmor/policy_unpack_test.c
540
KUNIT_EXPECT_PTR_EQ(test, puf->e->pos,
security/apparmor/policy_unpack_test.c
541
puf->e->start + TEST_NAMED_U64_BUF_OFFSET);
security/apparmor/policy_unpack_test.c
547
bool success = aa_unpack_X(puf->e, AA_NAME);
security/apparmor/policy_unpack_test.c
550
KUNIT_EXPECT_TRUE(test, puf->e->pos == puf->e->start + 1);
security/apparmor/policy_unpack_test.c
556
bool success = aa_unpack_X(puf->e, AA_STRING);
security/apparmor/policy_unpack_test.c
559
KUNIT_EXPECT_TRUE(test, puf->e->pos == puf->e->start);
security/apparmor/policy_unpack_test.c
567
puf->e->pos = puf->e->end;
security/apparmor/policy_unpack_test.c
568
success = aa_unpack_X(puf->e, AA_NAME);
security/apparmor/policy_unpack_test.c
60
struct aa_ext *e;
security/apparmor/policy_unpack_test.c
65
e = kunit_kmalloc(test, sizeof(*e), GFP_USER);
security/apparmor/policy_unpack_test.c
66
KUNIT_EXPECT_NOT_ERR_OR_NULL(test, e);
security/apparmor/policy_unpack_test.c
68
e->start = buf;
security/apparmor/policy_unpack_test.c
69
e->end = e->start + buf_size;
security/apparmor/policy_unpack_test.c
70
e->pos = e->start;
security/apparmor/policy_unpack_test.c
74
strscpy(buf + 3, TEST_STRING_NAME, e->end - (void *)(buf + 3));
security/apparmor/policy_unpack_test.c
76
buf = e->start + TEST_STRING_BUF_OFFSET;
security/apparmor/policy_unpack_test.c
79
strscpy(buf + 3, TEST_STRING_DATA, e->end - (void *)(buf + 3));
security/apparmor/policy_unpack_test.c
80
buf = e->start + TEST_NAMED_U32_BUF_OFFSET;
security/apparmor/policy_unpack_test.c
83
strscpy(buf + 3, TEST_U32_NAME, e->end - (void *)(buf + 3));
security/apparmor/policy_unpack_test.c
87
buf = e->start + TEST_NAMED_U64_BUF_OFFSET;
security/apparmor/policy_unpack_test.c
90
strscpy(buf + 3, TEST_U64_NAME, e->end - (void *)(buf + 3));
security/apparmor/policy_unpack_test.c
94
buf = e->start + TEST_NAMED_BLOB_BUF_OFFSET;
security/apparmor/policy_unpack_test.c
97
strscpy(buf + 3, TEST_BLOB_NAME, e->end - (void *)(buf + 3));
security/apparmor/resource.c
93
int e = 0;
security/apparmor/resource.c
97
e = -EACCES;
security/apparmor/resource.c
99
NULL, NULL, e);
security/integrity/ima/ima_fs.c
131
struct ima_template_entry *e;
security/integrity/ima/ima_fs.c
147
e = qe->entry;
security/integrity/ima/ima_fs.c
148
if (e == NULL)
security/integrity/ima/ima_fs.c
151
template_name = (e->template_desc->name[0] != '\0') ?
security/integrity/ima/ima_fs.c
152
e->template_desc->name : e->template_desc->fmt;
security/integrity/ima/ima_fs.c
159
pcr = !ima_canonical_fmt ? e->pcr : (__force u32)cpu_to_le32(e->pcr);
security/integrity/ima/ima_fs.c
160
ima_putc(m, &pcr, sizeof(e->pcr));
security/integrity/ima/ima_fs.c
163
ima_putc(m, e->digests[algo_idx].digest, hash_digest_size[algo]);
security/integrity/ima/ima_fs.c
178
template_data_len = !ima_canonical_fmt ? e->template_data_len :
security/integrity/ima/ima_fs.c
179
(__force u32)cpu_to_le32(e->template_data_len);
security/integrity/ima/ima_fs.c
180
ima_putc(m, &template_data_len, sizeof(e->template_data_len));
security/integrity/ima/ima_fs.c
184
for (i = 0; i < e->template_desc->num_fields; i++) {
security/integrity/ima/ima_fs.c
187
e->template_desc->fields[i];
security/integrity/ima/ima_fs.c
193
field->field_show(m, show, &e->template_data[i]);
security/integrity/ima/ima_fs.c
230
struct ima_template_entry *e;
security/integrity/ima/ima_fs.c
244
e = qe->entry;
security/integrity/ima/ima_fs.c
245
if (e == NULL)
security/integrity/ima/ima_fs.c
248
template_name = (e->template_desc->name[0] != '\0') ?
security/integrity/ima/ima_fs.c
249
e->template_desc->name : e->template_desc->fmt;
security/integrity/ima/ima_fs.c
252
seq_printf(m, "%2d ", e->pcr);
security/integrity/ima/ima_fs.c
255
ima_print_digest(m, e->digests[algo_idx].digest, hash_digest_size[algo]);
security/integrity/ima/ima_fs.c
261
for (i = 0; i < e->template_desc->num_fields; i++) {
security/integrity/ima/ima_fs.c
263
if (e->template_data[i].len == 0)
security/integrity/ima/ima_fs.c
266
e->template_desc->fields[i]->field_show(m, IMA_SHOW_ASCII,
security/integrity/ima/ima_fs.c
267
&e->template_data[i]);
security/integrity/ima/ima_policy.c
486
struct ima_rule_entry *entry, *e;
security/integrity/ima/ima_policy.c
489
list_for_each_entry_safe(entry, e, &ima_policy_rules, list) {
security/selinux/ss/ebitmap.c
260
int ebitmap_get_bit(const struct ebitmap *e, u32 bit)
security/selinux/ss/ebitmap.c
264
if (e->highbit < bit)
security/selinux/ss/ebitmap.c
267
n = e->node;
security/selinux/ss/ebitmap.c
277
int ebitmap_set_bit(struct ebitmap *e, u32 bit, int value)
security/selinux/ss/ebitmap.c
282
n = e->node;
security/selinux/ss/ebitmap.c
303
e->highbit = prev->startbit +
security/selinux/ss/ebitmap.c
306
e->highbit = 0;
security/selinux/ss/ebitmap.c
311
e->node = n->next;
security/selinux/ss/ebitmap.c
332
e->highbit = new->startbit + EBITMAP_SIZE;
security/selinux/ss/ebitmap.c
338
new->next = e->node;
security/selinux/ss/ebitmap.c
339
e->node = new;
security/selinux/ss/ebitmap.c
345
void ebitmap_destroy(struct ebitmap *e)
security/selinux/ss/ebitmap.c
349
if (!e)
security/selinux/ss/ebitmap.c
352
n = e->node;
security/selinux/ss/ebitmap.c
359
e->highbit = 0;
security/selinux/ss/ebitmap.c
360
e->node = NULL;
security/selinux/ss/ebitmap.c
363
int ebitmap_read(struct ebitmap *e, struct policy_file *fp)
security/selinux/ss/ebitmap.c
373
ebitmap_init(e);
security/selinux/ss/ebitmap.c
380
e->highbit = le32_to_cpu(buf[1]);
security/selinux/ss/ebitmap.c
386
mapunit, BITS_PER_U64, e->highbit);
security/selinux/ss/ebitmap.c
391
e->highbit += EBITMAP_SIZE - 1;
security/selinux/ss/ebitmap.c
392
e->highbit -= (e->highbit % EBITMAP_SIZE);
security/selinux/ss/ebitmap.c
394
if (!e->highbit) {
security/selinux/ss/ebitmap.c
395
e->node = NULL;
security/selinux/ss/ebitmap.c
399
if (e->highbit && !count)
security/selinux/ss/ebitmap.c
416
if (startbit > e->highbit - mapunit) {
security/selinux/ss/ebitmap.c
419
startbit, (e->highbit - mapunit));
security/selinux/ss/ebitmap.c
437
e->node = tmp;
security/selinux/ss/ebitmap.c
464
if (n && n->startbit + EBITMAP_SIZE != e->highbit) {
security/selinux/ss/ebitmap.c
466
e->highbit, n->startbit + EBITMAP_SIZE);
security/selinux/ss/ebitmap.c
477
ebitmap_destroy(e);
security/selinux/ss/ebitmap.c
481
int ebitmap_write(const struct ebitmap *e, struct policy_file *fp)
security/selinux/ss/ebitmap.c
494
ebitmap_for_each_positive_bit(e, n, bit)
security/selinux/ss/ebitmap.c
512
ebitmap_for_each_positive_bit(e, n, bit)
security/selinux/ss/ebitmap.c
560
u32 ebitmap_hash(const struct ebitmap *e, u32 hash)
security/selinux/ss/ebitmap.c
565
hash = jhash_1word(e->highbit, hash);
security/selinux/ss/ebitmap.c
566
for (node = e->node; node; node = node->next) {
security/selinux/ss/ebitmap.h
118
#define ebitmap_for_each_positive_bit(e, n, bit) \
security/selinux/ss/ebitmap.h
119
for ((bit) = ebitmap_start_positive(e, &(n)); \
security/selinux/ss/ebitmap.h
120
(bit) < ebitmap_length(e); \
security/selinux/ss/ebitmap.h
121
(bit) = ebitmap_next_positive(e, &(n), bit))
security/selinux/ss/ebitmap.h
129
int ebitmap_get_bit(const struct ebitmap *e, u32 bit);
security/selinux/ss/ebitmap.h
130
int ebitmap_set_bit(struct ebitmap *e, u32 bit, int value);
security/selinux/ss/ebitmap.h
131
void ebitmap_destroy(struct ebitmap *e);
security/selinux/ss/ebitmap.h
133
int ebitmap_read(struct ebitmap *e, struct policy_file *fp);
security/selinux/ss/ebitmap.h
134
int ebitmap_write(const struct ebitmap *e, struct policy_file *fp);
security/selinux/ss/ebitmap.h
135
u32 ebitmap_hash(const struct ebitmap *e, u32 hash);
security/selinux/ss/ebitmap.h
47
#define ebitmap_length(e) ((e)->highbit)
security/selinux/ss/ebitmap.h
49
static inline u32 ebitmap_start_positive(const struct ebitmap *e,
security/selinux/ss/ebitmap.h
54
for (*n = e->node; *n; *n = (*n)->next) {
security/selinux/ss/ebitmap.h
59
return ebitmap_length(e);
security/selinux/ss/ebitmap.h
62
static inline void ebitmap_init(struct ebitmap *e)
security/selinux/ss/ebitmap.h
64
memset(e, 0, sizeof(*e));
security/selinux/ss/ebitmap.h
67
static inline u32 ebitmap_next_positive(const struct ebitmap *e,
security/selinux/ss/ebitmap.h
81
return ebitmap_length(e);
security/selinux/ss/mls.c
110
e = &context->range.level[l].cat;
security/selinux/ss/mls.c
111
ebitmap_for_each_positive_bit(e, node, i)
security/selinux/ss/mls.c
36
struct ebitmap *e;
security/selinux/ss/mls.c
50
e = &context->range.level[l].cat;
security/selinux/ss/mls.c
51
ebitmap_for_each_positive_bit(e, node, i)
security/selinux/ss/mls.c
91
struct ebitmap *e;
security/selinux/ss/policydb.c
1232
struct constraint_expr *e, *le;
security/selinux/ss/policydb.c
1256
e = kzalloc_obj(*e);
security/selinux/ss/policydb.c
1257
if (!e)
security/selinux/ss/policydb.c
1261
le->next = e;
security/selinux/ss/policydb.c
1263
c->expr = e;
security/selinux/ss/policydb.c
1268
e->expr_type = le32_to_cpu(buf[0]);
security/selinux/ss/policydb.c
1269
e->attr = le32_to_cpu(buf[1]);
security/selinux/ss/policydb.c
1270
e->op = le32_to_cpu(buf[2]);
security/selinux/ss/policydb.c
1272
switch (e->expr_type) {
security/selinux/ss/policydb.c
1289
if (!allowxtarget && (e->attr & CEXPR_XTARGET))
security/selinux/ss/policydb.c
1294
rc = ebitmap_read(&e->names, fp);
security/selinux/ss/policydb.c
1299
e->type_names = kzalloc_obj(*e->type_names);
security/selinux/ss/policydb.c
1300
if (!e->type_names)
security/selinux/ss/policydb.c
1302
type_set_init(e->type_names);
security/selinux/ss/policydb.c
1303
rc = type_set_read(e->type_names, fp);
security/selinux/ss/policydb.c
1311
le = e;
security/selinux/ss/policydb.c
227
struct constraint_expr *e, *etmp;
security/selinux/ss/policydb.c
236
e = constraint->expr;
security/selinux/ss/policydb.c
237
while (e) {
security/selinux/ss/policydb.c
238
etmp = e;
security/selinux/ss/policydb.c
239
e = e->next;
security/selinux/ss/policydb.c
249
e = constraint->expr;
security/selinux/ss/policydb.c
250
while (e) {
security/selinux/ss/policydb.c
251
etmp = e;
security/selinux/ss/policydb.c
252
e = e->next;
security/selinux/ss/policydb.c
2736
struct ebitmap *e = &p->type_attr_map_array[i];
security/selinux/ss/policydb.c
2739
rc = ebitmap_read(e, fp);
security/selinux/ss/policydb.c
2744
rc = ebitmap_set_bit(e, i, 1);
security/selinux/ss/policydb.c
3043
struct constraint_expr *e;
security/selinux/ss/policydb.c
3050
for (e = c->expr; e; e = e->next)
security/selinux/ss/policydb.c
3057
for (e = c->expr; e; e = e->next) {
security/selinux/ss/policydb.c
3058
buf[0] = cpu_to_le32(e->expr_type);
security/selinux/ss/policydb.c
3059
buf[1] = cpu_to_le32(e->attr);
security/selinux/ss/policydb.c
3060
buf[2] = cpu_to_le32(e->op);
security/selinux/ss/policydb.c
3065
switch (e->expr_type) {
security/selinux/ss/policydb.c
3067
rc = ebitmap_write(&e->names, fp);
security/selinux/ss/policydb.c
3072
rc = type_set_write(e->type_names, fp);
security/selinux/ss/policydb.c
3793
struct ebitmap *e = &p->type_attr_map_array[i];
security/selinux/ss/policydb.c
3795
rc = ebitmap_write(e, fp);
security/selinux/ss/services.c
275
struct constraint_expr *e;
security/selinux/ss/services.c
279
for (e = cexpr; e; e = e->next) {
security/selinux/ss/services.c
280
switch (e->expr_type) {
security/selinux/ss/services.c
298
switch (e->attr) {
security/selinux/ss/services.c
312
switch (e->op) {
security/selinux/ss/services.c
356
switch (e->op) {
security/selinux/ss/services.c
382
switch (e->op) {
security/selinux/ss/services.c
398
if (e->attr & CEXPR_TARGET)
security/selinux/ss/services.c
400
else if (e->attr & CEXPR_XTARGET) {
security/selinux/ss/services.c
407
if (e->attr & CEXPR_USER)
security/selinux/ss/services.c
409
else if (e->attr & CEXPR_ROLE)
security/selinux/ss/services.c
411
else if (e->attr & CEXPR_TYPE)
security/selinux/ss/services.c
418
switch (e->op) {
security/selinux/ss/services.c
420
s[++sp] = ebitmap_get_bit(&e->names, val1 - 1);
security/selinux/ss/services.c
423
s[++sp] = !ebitmap_get_bit(&e->names, val1 - 1);
security/tomoyo/common.c
1072
struct tomoyo_task_acl e = {
security/tomoyo/common.c
1077
if (e.domainname)
security/tomoyo/common.c
1078
error = tomoyo_update_domain(&e.head, sizeof(e), param,
security/tomoyo/common.c
1081
tomoyo_put_name(e.domainname);
security/tomoyo/common.c
863
struct tomoyo_manager e = { };
security/tomoyo/common.c
874
e.manager = tomoyo_get_name(manager);
security/tomoyo/common.c
875
if (e.manager) {
security/tomoyo/common.c
876
error = tomoyo_update_policy(&e.head, sizeof(e), &param,
security/tomoyo/common.c
878
tomoyo_put_name(e.manager);
security/tomoyo/condition.c
455
struct tomoyo_condition *e)
security/tomoyo/condition.c
461
e->transit = tomoyo_get_domainname(param);
security/tomoyo/condition.c
477
e->transit = tomoyo_get_name(tomoyo_read_token(param));
security/tomoyo/condition.c
479
if (e->transit)
security/tomoyo/condition.c
503
struct tomoyo_condition e = { };
security/tomoyo/condition.c
505
tomoyo_get_transit_preference(param, &e);
security/tomoyo/condition.c
566
e.argc++;
security/tomoyo/condition.c
567
e.condc++;
security/tomoyo/condition.c
569
e.argc--;
security/tomoyo/condition.c
570
e.condc--;
security/tomoyo/condition.c
581
e.envc++;
security/tomoyo/condition.c
582
e.condc++;
security/tomoyo/condition.c
584
e.envc--;
security/tomoyo/condition.c
585
e.condc--;
security/tomoyo/condition.c
599
e.numbers_count++;
security/tomoyo/condition.c
601
e.numbers_count--;
security/tomoyo/condition.c
611
e.condc++;
security/tomoyo/condition.c
613
e.condc--;
security/tomoyo/condition.c
617
e.names_count++;
security/tomoyo/condition.c
619
e.names_count--;
security/tomoyo/condition.c
631
e.numbers_count++;
security/tomoyo/condition.c
633
e.numbers_count--;
security/tomoyo/condition.c
656
__LINE__, e.condc, e.numbers_count, e.names_count, e.argc,
security/tomoyo/condition.c
657
e.envc);
security/tomoyo/condition.c
659
BUG_ON(e.names_count | e.numbers_count | e.argc | e.envc |
security/tomoyo/condition.c
660
e.condc);
security/tomoyo/condition.c
663
e.size = sizeof(*entry)
security/tomoyo/condition.c
664
+ e.condc * sizeof(struct tomoyo_condition_element)
security/tomoyo/condition.c
665
+ e.numbers_count * sizeof(struct tomoyo_number_union)
security/tomoyo/condition.c
666
+ e.names_count * sizeof(struct tomoyo_name_union)
security/tomoyo/condition.c
667
+ e.argc * sizeof(struct tomoyo_argv)
security/tomoyo/condition.c
668
+ e.envc * sizeof(struct tomoyo_envp);
security/tomoyo/condition.c
669
entry = kzalloc(e.size, GFP_NOFS);
security/tomoyo/condition.c
672
*entry = e;
security/tomoyo/condition.c
673
e.transit = NULL;
security/tomoyo/condition.c
675
numbers_p = (struct tomoyo_number_union *) (condp + e.condc);
security/tomoyo/condition.c
676
names_p = (struct tomoyo_name_union *) (numbers_p + e.numbers_count);
security/tomoyo/condition.c
677
argv = (struct tomoyo_argv *) (names_p + e.names_count);
security/tomoyo/condition.c
678
envp = (struct tomoyo_envp *) (argv + e.argc);
security/tomoyo/condition.c
702
tomoyo_put_name(e.transit);
security/tomoyo/domain.c
245
struct tomoyo_transition_control e = { .type = type };
security/tomoyo/domain.c
261
e.program = tomoyo_get_name(program);
security/tomoyo/domain.c
262
if (!e.program)
security/tomoyo/domain.c
269
e.is_last_name = true;
security/tomoyo/domain.c
271
e.domainname = tomoyo_get_name(domainname);
security/tomoyo/domain.c
272
if (!e.domainname)
security/tomoyo/domain.c
276
error = tomoyo_update_policy(&e.head, sizeof(e), param,
security/tomoyo/domain.c
279
tomoyo_put_name(e.domainname);
security/tomoyo/domain.c
280
tomoyo_put_name(e.program);
security/tomoyo/domain.c
405
struct tomoyo_aggregator e = { };
security/tomoyo/domain.c
413
e.original_name = tomoyo_get_name(original_name);
security/tomoyo/domain.c
414
e.aggregated_name = tomoyo_get_name(aggregated_name);
security/tomoyo/domain.c
415
if (!e.original_name || !e.aggregated_name ||
security/tomoyo/domain.c
416
e.aggregated_name->is_patterned) /* No patterns allowed. */
security/tomoyo/domain.c
419
error = tomoyo_update_policy(&e.head, sizeof(e), param,
security/tomoyo/domain.c
422
tomoyo_put_name(e.original_name);
security/tomoyo/domain.c
423
tomoyo_put_name(e.aggregated_name);
security/tomoyo/domain.c
525
struct tomoyo_domain_info e = { };
security/tomoyo/domain.c
555
e.ns = tomoyo_assign_namespace(domainname);
security/tomoyo/domain.c
556
if (!e.ns)
security/tomoyo/domain.c
566
e.profile = domain->profile;
security/tomoyo/domain.c
567
memcpy(e.group, domain->group, sizeof(e.group));
security/tomoyo/domain.c
569
e.domainname = tomoyo_get_name(domainname);
security/tomoyo/domain.c
570
if (!e.domainname)
security/tomoyo/domain.c
576
entry = tomoyo_commit_ok(&e, sizeof(e));
security/tomoyo/domain.c
585
tomoyo_put_name(e.domainname);
security/tomoyo/environ.c
103
e.env = tomoyo_get_name(data);
security/tomoyo/environ.c
104
if (!e.env)
security/tomoyo/environ.c
106
error = tomoyo_update_domain(&e.head, sizeof(e), param,
security/tomoyo/environ.c
108
tomoyo_put_name(e.env);
security/tomoyo/environ.c
97
struct tomoyo_env_acl e = { .head.type = TOMOYO_TYPE_ENV_ACL };
security/tomoyo/file.c
1001
tomoyo_put_name_union(&e.dev_name);
security/tomoyo/file.c
1002
tomoyo_put_name_union(&e.dir_name);
security/tomoyo/file.c
1003
tomoyo_put_name_union(&e.fs_type);
security/tomoyo/file.c
1004
tomoyo_put_number_union(&e.flags);
security/tomoyo/file.c
393
struct tomoyo_path_acl e = {
security/tomoyo/file.c
399
if (!tomoyo_parse_name_union(param, &e.name))
security/tomoyo/file.c
402
error = tomoyo_update_domain(&e.head, sizeof(e), param,
security/tomoyo/file.c
405
tomoyo_put_name_union(&e.name);
security/tomoyo/file.c
469
struct tomoyo_mkdev_acl e = {
security/tomoyo/file.c
475
if (!tomoyo_parse_name_union(param, &e.name) ||
security/tomoyo/file.c
476
!tomoyo_parse_number_union(param, &e.mode) ||
security/tomoyo/file.c
477
!tomoyo_parse_number_union(param, &e.major) ||
security/tomoyo/file.c
478
!tomoyo_parse_number_union(param, &e.minor))
security/tomoyo/file.c
481
error = tomoyo_update_domain(&e.head, sizeof(e), param,
security/tomoyo/file.c
484
tomoyo_put_name_union(&e.name);
security/tomoyo/file.c
485
tomoyo_put_number_union(&e.mode);
security/tomoyo/file.c
486
tomoyo_put_number_union(&e.major);
security/tomoyo/file.c
487
tomoyo_put_number_union(&e.minor);
security/tomoyo/file.c
548
struct tomoyo_path2_acl e = {
security/tomoyo/file.c
554
if (!tomoyo_parse_name_union(param, &e.name1) ||
security/tomoyo/file.c
555
!tomoyo_parse_name_union(param, &e.name2))
security/tomoyo/file.c
558
error = tomoyo_update_domain(&e.head, sizeof(e), param,
security/tomoyo/file.c
561
tomoyo_put_name_union(&e.name1);
security/tomoyo/file.c
562
tomoyo_put_name_union(&e.name2);
security/tomoyo/file.c
686
struct tomoyo_path_number_acl e = {
security/tomoyo/file.c
692
if (!tomoyo_parse_name_union(param, &e.name) ||
security/tomoyo/file.c
693
!tomoyo_parse_number_union(param, &e.number))
security/tomoyo/file.c
696
error = tomoyo_update_domain(&e.head, sizeof(e), param,
security/tomoyo/file.c
699
tomoyo_put_name_union(&e.name);
security/tomoyo/file.c
700
tomoyo_put_number_union(&e.number);
security/tomoyo/file.c
990
struct tomoyo_mount_acl e = { .head.type = TOMOYO_TYPE_MOUNT_ACL };
security/tomoyo/file.c
993
if (!tomoyo_parse_name_union(param, &e.dev_name) ||
security/tomoyo/file.c
994
!tomoyo_parse_name_union(param, &e.dir_name) ||
security/tomoyo/file.c
995
!tomoyo_parse_name_union(param, &e.fs_type) ||
security/tomoyo/file.c
996
!tomoyo_parse_number_union(param, &e.flags))
security/tomoyo/file.c
999
error = tomoyo_update_domain(&e.head, sizeof(e), param,
security/tomoyo/group.c
106
struct tomoyo_address_group e = { };
security/tomoyo/group.c
109
!tomoyo_parse_ipaddr_union(param, &e.address))
security/tomoyo/group.c
111
error = tomoyo_update_policy(&e.head, sizeof(e), param,
security/tomoyo/group.c
83
struct tomoyo_path_group e = { };
security/tomoyo/group.c
85
e.member_name = tomoyo_get_name(tomoyo_read_token(param));
security/tomoyo/group.c
86
if (!e.member_name) {
security/tomoyo/group.c
90
error = tomoyo_update_policy(&e.head, sizeof(e), param,
security/tomoyo/group.c
92
tomoyo_put_name(e.member_name);
security/tomoyo/group.c
94
struct tomoyo_number_group e = { };
security/tomoyo/group.c
97
!tomoyo_parse_number_union(param, &e.number))
security/tomoyo/group.c
99
error = tomoyo_update_policy(&e.head, sizeof(e), param,
security/tomoyo/memory.c
106
e.group_name = tomoyo_get_name(group_name);
security/tomoyo/memory.c
107
if (!e.group_name)
security/tomoyo/memory.c
113
if (e.group_name != group->group_name ||
security/tomoyo/memory.c
121
struct tomoyo_group *entry = tomoyo_commit_ok(&e, sizeof(e));
security/tomoyo/memory.c
133
tomoyo_put_name(e.group_name);
security/tomoyo/memory.c
98
struct tomoyo_group e = { };
security/tomoyo/network.c
284
struct tomoyo_inet_acl e = { .head.type = TOMOYO_TYPE_INET_ACL };
security/tomoyo/network.c
290
for (e.protocol = 0; e.protocol < TOMOYO_SOCK_MAX; e.protocol++)
security/tomoyo/network.c
291
if (!strcmp(protocol, tomoyo_proto_keyword[e.protocol]))
security/tomoyo/network.c
295
e.perm |= 1 << type;
security/tomoyo/network.c
296
if (e.protocol == TOMOYO_SOCK_MAX || !e.perm)
security/tomoyo/network.c
300
e.address.group =
security/tomoyo/network.c
302
if (!e.address.group)
security/tomoyo/network.c
305
if (!tomoyo_parse_ipaddr_union(param, &e.address))
security/tomoyo/network.c
308
if (!tomoyo_parse_number_union(param, &e.port) ||
security/tomoyo/network.c
309
e.port.values[1] > 65535)
security/tomoyo/network.c
311
error = tomoyo_update_domain(&e.head, sizeof(e), param,
security/tomoyo/network.c
315
tomoyo_put_group(e.address.group);
security/tomoyo/network.c
316
tomoyo_put_number_union(&e.port);
security/tomoyo/network.c
329
struct tomoyo_unix_acl e = { .head.type = TOMOYO_TYPE_UNIX_ACL };
security/tomoyo/network.c
335
for (e.protocol = 0; e.protocol < TOMOYO_SOCK_MAX; e.protocol++)
security/tomoyo/network.c
336
if (!strcmp(protocol, tomoyo_proto_keyword[e.protocol]))
security/tomoyo/network.c
340
e.perm |= 1 << type;
security/tomoyo/network.c
341
if (e.protocol == TOMOYO_SOCK_MAX || !e.perm)
security/tomoyo/network.c
343
if (!tomoyo_parse_name_union(param, &e.name))
security/tomoyo/network.c
345
error = tomoyo_update_domain(&e.head, sizeof(e), param,
security/tomoyo/network.c
348
tomoyo_put_name_union(&e.name);
security/tomoyo/util.c
452
unsigned char e;
security/tomoyo/util.c
457
e = *string++;
security/tomoyo/util.c
458
if (d < '0' || d > '7' || e < '0' || e > '7')
security/tomoyo/util.c
460
c = tomoyo_make_byte(c, d, e);
sound/core/pcm_drm_eld.c
327
int snd_parse_eld(struct device *dev, struct snd_parsed_hdmi_eld *e,
sound/core/pcm_drm_eld.c
333
memset(e, 0, sizeof(*e));
sound/core/pcm_drm_eld.c
334
e->eld_ver = GRAB_BITS(buf, 0, 3, 5);
sound/core/pcm_drm_eld.c
335
if (e->eld_ver != ELD_VER_CEA_861D &&
sound/core/pcm_drm_eld.c
336
e->eld_ver != ELD_VER_PARTIAL) {
sound/core/pcm_drm_eld.c
337
dev_info(dev, "HDMI: Unknown ELD version %d\n", e->eld_ver);
sound/core/pcm_drm_eld.c
341
e->baseline_len = GRAB_BITS(buf, 2, 0, 8);
sound/core/pcm_drm_eld.c
343
e->cea_edid_ver = GRAB_BITS(buf, 4, 5, 3);
sound/core/pcm_drm_eld.c
345
e->support_hdcp = GRAB_BITS(buf, 5, 0, 1);
sound/core/pcm_drm_eld.c
346
e->support_ai = GRAB_BITS(buf, 5, 1, 1);
sound/core/pcm_drm_eld.c
347
e->conn_type = GRAB_BITS(buf, 5, 2, 2);
sound/core/pcm_drm_eld.c
348
e->sad_count = GRAB_BITS(buf, 5, 4, 4);
sound/core/pcm_drm_eld.c
350
e->aud_synch_delay = GRAB_BITS(buf, 6, 0, 8) * 2;
sound/core/pcm_drm_eld.c
351
e->spk_alloc = GRAB_BITS(buf, 7, 0, 7);
sound/core/pcm_drm_eld.c
353
e->port_id = get_unaligned_le64(buf + 8);
sound/core/pcm_drm_eld.c
356
e->manufacture_id = get_unaligned_le16(buf + 16);
sound/core/pcm_drm_eld.c
357
e->product_id = get_unaligned_le16(buf + 18);
sound/core/pcm_drm_eld.c
366
strscpy(e->monitor_name, buf + ELD_FIXED_BYTES, mnl + 1);
sound/core/pcm_drm_eld.c
368
for (i = 0; i < e->sad_count; i++) {
sound/core/pcm_drm_eld.c
373
hdmi_update_short_audio_desc(dev, e->sad + i,
sound/core/pcm_drm_eld.c
382
if (!e->spk_alloc && e->sad_count)
sound/core/pcm_drm_eld.c
383
e->spk_alloc = 0xffff;
sound/core/pcm_drm_eld.c
461
void snd_show_eld(struct device *dev, struct snd_parsed_hdmi_eld *e)
sound/core/pcm_drm_eld.c
466
e->monitor_name,
sound/core/pcm_drm_eld.c
467
eld_connection_type_names[e->conn_type]);
sound/core/pcm_drm_eld.c
469
if (e->spk_alloc) {
sound/core/pcm_drm_eld.c
472
snd_eld_print_channel_allocation(e->spk_alloc, buf, sizeof(buf));
sound/core/pcm_drm_eld.c
476
for (i = 0; i < e->sad_count; i++)
sound/core/pcm_drm_eld.c
477
hdmi_show_short_audio_desc(dev, e->sad + i);
sound/core/pcm_drm_eld.c
508
void snd_print_eld_info(struct snd_parsed_hdmi_eld *e,
sound/core/pcm_drm_eld.c
528
snd_iprintf(buffer, "monitor_name\t\t%s\n", e->monitor_name);
sound/core/pcm_drm_eld.c
530
eld_connection_type_names[e->conn_type]);
sound/core/pcm_drm_eld.c
531
snd_iprintf(buffer, "eld_version\t\t[0x%x] %s\n", e->eld_ver,
sound/core/pcm_drm_eld.c
532
eld_version_names[e->eld_ver]);
sound/core/pcm_drm_eld.c
533
snd_iprintf(buffer, "edid_version\t\t[0x%x] %s\n", e->cea_edid_ver,
sound/core/pcm_drm_eld.c
534
cea_edid_version_names[e->cea_edid_ver]);
sound/core/pcm_drm_eld.c
535
snd_iprintf(buffer, "manufacture_id\t\t0x%x\n", e->manufacture_id);
sound/core/pcm_drm_eld.c
536
snd_iprintf(buffer, "product_id\t\t0x%x\n", e->product_id);
sound/core/pcm_drm_eld.c
537
snd_iprintf(buffer, "port_id\t\t\t0x%llx\n", (long long)e->port_id);
sound/core/pcm_drm_eld.c
538
snd_iprintf(buffer, "support_hdcp\t\t%d\n", e->support_hdcp);
sound/core/pcm_drm_eld.c
539
snd_iprintf(buffer, "support_ai\t\t%d\n", e->support_ai);
sound/core/pcm_drm_eld.c
540
snd_iprintf(buffer, "audio_sync_delay\t%d\n", e->aud_synch_delay);
sound/core/pcm_drm_eld.c
542
snd_eld_print_channel_allocation(e->spk_alloc, buf, sizeof(buf));
sound/core/pcm_drm_eld.c
543
snd_iprintf(buffer, "speakers\t\t[0x%x]%s\n", e->spk_alloc, buf);
sound/core/pcm_drm_eld.c
545
snd_iprintf(buffer, "sad_count\t\t%d\n", e->sad_count);
sound/core/pcm_drm_eld.c
547
for (i = 0; i < e->sad_count; i++)
sound/core/pcm_drm_eld.c
548
hdmi_print_sad_info(i, e->sad + i, buffer);
sound/core/seq/oss/seq_oss_event.c
125
switch (q->e.cmd) {
sound/core/seq/oss/seq_oss_event.c
127
return note_off_event(dp, q->e.dev, q->e.chn, q->e.p1, q->e.p2, ev);
sound/core/seq/oss/seq_oss_event.c
130
return note_on_event(dp, q->e.dev, q->e.chn, q->e.p1, q->e.p2, ev);
sound/core/seq/oss/seq_oss_event.c
133
return set_control_event(dp, q->e.dev, SNDRV_SEQ_EVENT_PGMCHANGE,
sound/core/seq/oss/seq_oss_event.c
134
q->e.chn, 0, q->e.p1, ev);
sound/core/seq/oss/seq_oss_event.c
137
return set_control_event(dp, q->e.dev, SNDRV_SEQ_EVENT_CHANPRESS,
sound/core/seq/oss/seq_oss_event.c
138
q->e.chn, 0, q->e.p1, ev);
sound/core/seq/oss/seq_oss_event.c
142
val = (char)q->e.p1;
sound/core/seq/oss/seq_oss_event.c
144
return set_control_event(dp, q->e.dev, SNDRV_SEQ_EVENT_CONTROLLER,
sound/core/seq/oss/seq_oss_event.c
145
q->e.chn, CTL_PAN, val, ev);
sound/core/seq/oss/seq_oss_event.c
148
val = ((short)q->e.p3 << 8) | (short)q->e.p2;
sound/core/seq/oss/seq_oss_event.c
149
switch (q->e.p1) {
sound/core/seq/oss/seq_oss_event.c
152
return set_control_event(dp, q->e.dev,
sound/core/seq/oss/seq_oss_event.c
154
q->e.chn, 0, val, ev);
sound/core/seq/oss/seq_oss_event.c
157
return set_control_event(dp, q->e.dev,
sound/core/seq/oss/seq_oss_event.c
159
q->e.chn, 0, val*128/100, ev);
sound/core/seq/oss/seq_oss_event.c
161
return set_control_event(dp, q->e.dev,
sound/core/seq/oss/seq_oss_event.c
163
q->e.chn, q->e.p1, val, ev);
sound/core/seq/oss/seq_oss_event.c
167
return snd_seq_oss_synth_raw_event(dp, q->e.dev, q->c, ev);
sound/core/seq/oss/seq_oss_event.h
85
struct evrec_extended e;
sound/firewire/bebob/bebob_proc.c
160
void (*op)(struct snd_info_entry *e, struct snd_info_buffer *b))
sound/firewire/bebob/bebob_stream.c
708
unsigned int i, e, channels, format;
sound/firewire/bebob/bebob_stream.c
729
for (e = 0; e < buf[4]; e++) {
sound/firewire/bebob/bebob_stream.c
730
channels = buf[5 + e * 2];
sound/firewire/bebob/bebob_stream.c
731
format = buf[6 + e * 2];
sound/firewire/fireface/ff-proc.c
37
void (*op)(struct snd_info_entry *e,
sound/firewire/fireworks/fireworks_proc.c
196
void (*op)(struct snd_info_entry *e, struct snd_info_buffer *b))
sound/firewire/motu/motu-proc.c
83
void (*op)(struct snd_info_entry *e,
sound/firewire/oxfw/oxfw-proc.c
79
void (*op)(struct snd_info_entry *e,
sound/firewire/oxfw/oxfw-stream.c
554
unsigned int i, e, channels, type;
sound/firewire/oxfw/oxfw-stream.c
576
for (e = 0; e < format[4]; e++) {
sound/firewire/oxfw/oxfw-stream.c
577
channels = format[5 + e * 2];
sound/firewire/oxfw/oxfw-stream.c
578
type = format[6 + e * 2];
sound/firewire/tascam/tascam-proc.c
54
void (*op)(struct snd_info_entry *e,
sound/hda/codecs/hdmi/eld.c
127
struct snd_parsed_hdmi_eld *e = &eld->info;
sound/hda/codecs/hdmi/eld.c
147
e->conn_type = val;
sound/hda/codecs/hdmi/eld.c
149
e->port_id = val;
sound/hda/codecs/hdmi/eld.c
151
e->support_hdcp = val;
sound/hda/codecs/hdmi/eld.c
153
e->support_ai = val;
sound/hda/codecs/hdmi/eld.c
155
e->aud_synch_delay = val;
sound/hda/codecs/hdmi/eld.c
157
e->spk_alloc = val;
sound/hda/codecs/hdmi/eld.c
159
e->sad_count = val;
sound/hda/codecs/hdmi/eld.c
170
e->sad[n].format = val;
sound/hda/codecs/hdmi/eld.c
172
e->sad[n].channels = val;
sound/hda/codecs/hdmi/eld.c
174
e->sad[n].rates = val;
sound/hda/codecs/hdmi/eld.c
176
e->sad[n].sample_bits = val;
sound/hda/codecs/hdmi/eld.c
178
e->sad[n].max_bitrate = val;
sound/hda/codecs/hdmi/eld.c
180
e->sad[n].profile = val;
sound/hda/codecs/hdmi/eld.c
181
if (n >= e->sad_count)
sound/hda/codecs/hdmi/eld.c
182
e->sad_count = n + 1;
sound/hda/codecs/hdmi/eld.c
189
void snd_hdmi_eld_update_pcm_info(struct snd_parsed_hdmi_eld *e,
sound/hda/codecs/hdmi/eld.c
206
for (i = 0; i < e->sad_count; i++) {
sound/hda/codecs/hdmi/eld.c
207
struct snd_cea_sad *a = &e->sad[i];
sound/hda/common/hda_local.h
698
void snd_hdmi_eld_update_pcm_info(struct snd_parsed_hdmi_eld *e,
sound/isa/gus/gus_volume.c
17
unsigned short e, m, tmp;
sound/isa/gus/gus_volume.c
22
e = 7;
sound/isa/gus/gus_volume.c
24
while (e > 0 && tmp < (1 << e))
sound/isa/gus/gus_volume.c
25
e--;
sound/isa/gus/gus_volume.c
29
e++;
sound/isa/gus/gus_volume.c
32
m = vol - (1 << e);
sound/isa/gus/gus_volume.c
34
if (e > 8)
sound/isa/gus/gus_volume.c
35
m >>= e - 8;
sound/isa/gus/gus_volume.c
36
else if (e < 8)
sound/isa/gus/gus_volume.c
37
m <<= 8 - e;
sound/isa/gus/gus_volume.c
40
return (e << 8) | m;
sound/isa/gus/gus_volume.c
48
unsigned short e, m;
sound/isa/gus/gus_volume.c
52
e = gf1_vol >> 8;
sound/isa/gus/gus_volume.c
54
rvol = 1 << e;
sound/isa/gus/gus_volume.c
55
if (e > 8)
sound/isa/gus/gus_volume.c
56
return rvol | (m << (e - 8));
sound/isa/gus/gus_volume.c
57
return rvol | (m >> (8 - e));
sound/pci/ac97/ac97_codec.c
441
struct ac97_enum *e = (struct ac97_enum *)kcontrol->private_value;
sound/pci/ac97/ac97_codec.c
443
return snd_ctl_enum_info(uinfo, e->shift_l == e->shift_r ? 1 : 2,
sound/pci/ac97/ac97_codec.c
444
e->mask, e->texts);
sound/pci/ac97/ac97_codec.c
451
struct ac97_enum *e = (struct ac97_enum *)kcontrol->private_value;
sound/pci/ac97/ac97_codec.c
454
for (bitmask = 1; bitmask < e->mask; bitmask <<= 1)
sound/pci/ac97/ac97_codec.c
456
val = snd_ac97_read_cache(ac97, e->reg);
sound/pci/ac97/ac97_codec.c
457
ucontrol->value.enumerated.item[0] = (val >> e->shift_l) & (bitmask - 1);
sound/pci/ac97/ac97_codec.c
458
if (e->shift_l != e->shift_r)
sound/pci/ac97/ac97_codec.c
459
ucontrol->value.enumerated.item[1] = (val >> e->shift_r) & (bitmask - 1);
sound/pci/ac97/ac97_codec.c
468
struct ac97_enum *e = (struct ac97_enum *)kcontrol->private_value;
sound/pci/ac97/ac97_codec.c
472
for (bitmask = 1; bitmask < e->mask; bitmask <<= 1)
sound/pci/ac97/ac97_codec.c
474
if (ucontrol->value.enumerated.item[0] > e->mask - 1)
sound/pci/ac97/ac97_codec.c
476
val = ucontrol->value.enumerated.item[0] << e->shift_l;
sound/pci/ac97/ac97_codec.c
477
mask = (bitmask - 1) << e->shift_l;
sound/pci/ac97/ac97_codec.c
478
if (e->shift_l != e->shift_r) {
sound/pci/ac97/ac97_codec.c
479
if (ucontrol->value.enumerated.item[1] > e->mask - 1)
sound/pci/ac97/ac97_codec.c
481
val |= ucontrol->value.enumerated.item[1] << e->shift_r;
sound/pci/ac97/ac97_codec.c
482
mask |= (bitmask - 1) << e->shift_r;
sound/pci/ac97/ac97_codec.c
484
return snd_ac97_update_bits(ac97, e->reg, mask, val);
sound/pci/asihpi/asihpi.c
2176
u16 e;
sound/pci/asihpi/asihpi.c
2180
e = hpi_multiplexer_query_source(h_control,
sound/pci/asihpi/asihpi.c
2183
if (!e)
sound/pci/asihpi/asihpi.c
221
u16 e;
sound/pci/asihpi/asihpi.c
223
e = hpi_outstream_get_info_ex(h_stream, pw_state,
sound/pci/asihpi/asihpi.c
227
e = hpi_instream_get_info_ex(h_stream, pw_state,
sound/pci/asihpi/asihpi.c
230
return e;
sound/pci/asihpi/asihpi.c
2389
u16 e;
sound/pci/asihpi/asihpi.c
2391
e = hpi_sample_clock_get_local_rate(h_control, &rate);
sound/pci/asihpi/asihpi.c
2392
if (!e)
sound/pci/asihpi/asihpi.c
2431
u16 e;
sound/pci/asihpi/asihpi.c
2433
e = hpi_sample_clock_get_sample_rate(h_control, &rate);
sound/pci/asihpi/asihpi.c
2434
if (!e)
sound/pci/asihpi/asihpi.c
558
u16 e;
sound/pci/asihpi/asihpi.c
597
e = hpi_stream_group_add(
sound/pci/asihpi/asihpi.c
600
if (!e) {
sound/pci/asihpi/asihpi.c
603
hpi_handle_error(e);
sound/pci/asihpi/hpifunc.c
2829
u16 e;
sound/pci/asihpi/hpifunc.c
2831
e = hpi_control_query(h_volume, HPI_VOLUME_AUTOFADE, i, 0, &u);
sound/pci/asihpi/hpifunc.c
2833
return e;
sound/pci/au88x0/au88x0_a3d.c
48
short e)
sound/pci/au88x0/au88x0_a3d.c
53
(e << 0x10) | d);
sound/pci/au88x0/au88x0_a3d.c
63
short e)
sound/pci/au88x0/au88x0_a3d.c
68
(e << 0x10) | d);
sound/pci/au88x0/au88x0_a3d.c
89
short *d, short *e)
sound/pci/ctxfi/ctmixer.c
308
unsigned int e;
sound/pci/ctxfi/ctmixer.c
313
e = (x >> 10) & 0x7;
sound/pci/ctxfi/ctmixer.c
316
x >>= (7 - e);
sound/pci/oxygen/xonar_wm87x6.c
981
#define WM8776_FIELD_CTL_VOLUME(a, b, c, d, e, f, g, h, tlv_p) { \
sound/pci/oxygen/xonar_wm87x6.c
982
_WM8776_FIELD_CTL(a " Capture Volume", b, c, d, e, f, g, h), \
sound/pci/riptide/riptide.c
244
#define SEND_SETF(p,b,c,d,e,f,g) sendcmd(p,PARM,SETF|WORD1(b)|BYTE3(c),d|BYTE1(e)|BYTE2(f)|BYTE3(g),RET(0)) /* set sample format at mixer */
sound/pci/riptide/riptide.c
251
#define SEND_LSEL(p,b,c,d,e,f,g,h) sendcmd(p,PARM,LSEL|BYTE1(b)|BYTE2(c)|BYTE3(d),BYTE0(e)|BYTE1(f)|BYTE2(g)|BYTE3(h),RET(0)) /* select paths for internal connections */
sound/pci/riptide/riptide.c
252
#define SEND_SSRC(p,b,c,d,e) sendcmd(p,PARM,SSRC|BYTE1(b)|WORD2(c),WORD0(d)|WORD2(e),RET(0)) /* configure source */
sound/pci/riptide/riptide.c
256
#define SEND_SDGV(p,b,c,d,e) sendcmd(p,PARM,SDGV|BYTE2(b)|BYTE3(c),WORD0(d)|WORD2(e),RET(0)) /* set digital mixer */
sound/pci/riptide/riptide.c
262
#define SEND_TXAC(p,b,c,d,e,f) sendcmd(p,PARM,TXAC|BYTE1(b)|WORD2(c),WORD0(d)|BYTE2(e)|BYTE3(f),RET(0))
sound/soc/atmel/mchp-pdmc.c
149
struct soc_enum *e = (struct soc_enum *)kcontrol->private_value;
sound/soc/atmel/mchp-pdmc.c
152
item = snd_soc_enum_val_to_item(e, dd->sinc_order);
sound/soc/atmel/mchp-pdmc.c
163
struct soc_enum *e = (struct soc_enum *)kcontrol->private_value;
sound/soc/atmel/mchp-pdmc.c
167
if (item[0] >= e->items)
sound/soc/atmel/mchp-pdmc.c
170
val = snd_soc_enum_item_to_val(e, item[0]) << e->shift_l;
sound/soc/atmel/tse850-pcm5142.c
106
struct soc_enum *e = (struct soc_enum *)kctrl->private_value;
sound/soc/atmel/tse850-pcm5142.c
109
if (val >= e->items)
sound/soc/atmel/tse850-pcm5142.c
191
struct soc_enum *e = (struct soc_enum *)kctrl->private_value;
sound/soc/atmel/tse850-pcm5142.c
195
if (uV >= e->items)
sound/soc/atmel/tse850-pcm5142.c
76
struct soc_enum *e = (struct soc_enum *)kctrl->private_value;
sound/soc/atmel/tse850-pcm5142.c
79
if (val >= e->items)
sound/soc/codecs/adau17x1.c
191
struct soc_enum *e = (struct soc_enum *)kcontrol->private_value;
sound/soc/codecs/adau17x1.c
193
unsigned int stream = e->shift_l;
sound/soc/codecs/adau17x1.c
197
if (ucontrol->value.enumerated.item[0] >= e->items)
sound/soc/codecs/adau17x1.c
224
ucontrol->value.enumerated.item[0], e, &update);
sound/soc/codecs/adau17x1.c
235
struct soc_enum *e = (struct soc_enum *)kcontrol->private_value;
sound/soc/codecs/adau17x1.c
236
unsigned int stream = e->shift_l;
sound/soc/codecs/cpcap.c
385
struct soc_enum *e = (struct soc_enum *)kcontrol->private_value;
sound/soc/codecs/cpcap.c
386
unsigned int shift = e->shift_l;
sound/soc/codecs/cpcap.c
429
struct soc_enum *e = (struct soc_enum *)kcontrol->private_value;
sound/soc/codecs/cpcap.c
431
unsigned int mask = BIT(e->shift_l);
sound/soc/codecs/cpcap.c
462
snd_soc_dapm_mux_update_power(dapm, kcontrol, muxval, e, NULL);
sound/soc/codecs/cpcap.c
512
struct soc_enum *e = (struct soc_enum *)kcontrol->private_value;
sound/soc/codecs/cpcap.c
545
snd_soc_dapm_mux_update_power(dapm, kcontrol, muxval, e, NULL);
sound/soc/codecs/cpcap.c
587
struct soc_enum *e = (struct soc_enum *)kcontrol->private_value;
sound/soc/codecs/cpcap.c
612
snd_soc_dapm_mux_update_power(dapm, kcontrol, muxval, e, NULL);
sound/soc/codecs/cs42l43-jack.c
910
struct soc_enum *e = (struct soc_enum *)kcontrol->private_value;
sound/soc/codecs/cs42l43-jack.c
913
if (override >= e->items)
sound/soc/codecs/cs43130.c
1063
struct soc_enum *e = (struct soc_enum *)kcontrol->private_value;
sound/soc/codecs/cs43130.c
1069
if (item[0] >= e->items)
sound/soc/codecs/cs43130.c
1071
val = snd_soc_enum_item_to_val(e, item[0]) << e->shift_l;
sound/soc/codecs/cs47l92.c
120
ret = snd_soc_dapm_mux_update_power(dapm, kcontrol, mux, e, NULL);
sound/soc/codecs/cs47l92.c
60
struct soc_enum *e = (struct soc_enum *)kcontrol->private_value;
sound/soc/codecs/cs47l92.c
65
if (ucontrol->value.enumerated.item[0] > e->items - 1)
sound/soc/codecs/cs47l92.c
72
ep_sel = mux << e->shift_l;
sound/soc/codecs/cs48l32.c
1067
struct soc_enum *e = (struct soc_enum *) kcontrol->private_value;
sound/soc/codecs/cs48l32.c
1069
const unsigned int rate_num = e->mask;
sound/soc/codecs/cs48l32.c
1076
item = snd_soc_enum_val_to_item(e, cached_rate);
sound/soc/codecs/cs48l32.c
1088
struct soc_enum *e = (struct soc_enum *) kcontrol->private_value;
sound/soc/codecs/cs48l32.c
1089
const unsigned int rate_num = e->mask;
sound/soc/codecs/cs48l32.c
1094
if (item >= e->items)
sound/soc/codecs/cs48l32.c
1100
val = snd_soc_enum_item_to_val(e, item);
sound/soc/codecs/cs48l32.c
326
struct soc_enum *e = (struct soc_enum *) kcontrol->private_value;
sound/soc/codecs/cs48l32.c
334
switch (e->reg) {
sound/soc/codecs/cs48l32.c
345
src_val = mux << e->shift_l;
sound/soc/codecs/cs48l32.c
351
e->reg,
sound/soc/codecs/cs48l32.c
355
snd_soc_dapm_mux_update_power(dapm, kcontrol, mux, e, NULL);
sound/soc/codecs/cs48l32.c
390
struct soc_enum *e = (struct soc_enum *) kcontrol->private_value;
sound/soc/codecs/cs48l32.c
418
e->reg,
sound/soc/codecs/cs48l32.c
449
snd_soc_dapm_mux_update_power(dapm, kcontrol, mode, e, NULL);
sound/soc/codecs/cs48l32.c
709
struct soc_enum *e = (struct soc_enum *)kcontrol->private_value;
sound/soc/codecs/cs48l32.c
715
if (cs48l32_is_input_enabled(component, e->reg)) {
sound/soc/codecs/cs48l32.c
952
struct soc_enum *e = (struct soc_enum *) kcontrol->private_value;
sound/soc/codecs/cs48l32.c
955
item = snd_soc_enum_val_to_item(e, cs48l32_codec->eq_mode[e->shift_l]);
sound/soc/codecs/cs48l32.c
967
struct soc_enum *e = (struct soc_enum *) kcontrol->private_value;
sound/soc/codecs/cs48l32.c
972
if (item[0] >= e->items)
sound/soc/codecs/cs48l32.c
975
val = snd_soc_enum_item_to_val(e, item[0]);
sound/soc/codecs/cs48l32.c
978
if (cs48l32_codec->eq_mode[e->shift_l] != val) {
sound/soc/codecs/cs48l32.c
979
cs48l32_codec->eq_mode[e->shift_l] = val;
sound/soc/codecs/es8389.c
72
struct soc_enum *e = (struct soc_enum *)kcontrol->private_value;
sound/soc/codecs/es8389.c
89
return snd_soc_dapm_mux_update_power(dapm, kcontrol, val, e, NULL);
sound/soc/codecs/hdac_hdmi.c
906
struct soc_enum *e = (struct soc_enum *)kcontrol->private_value;
sound/soc/codecs/hdac_hdmi.c
914
const char *cvt_name = e->texts[ucontrol->value.enumerated.item[0]];
sound/soc/codecs/lpass-rx-macro.c
1688
struct soc_enum *e = (struct soc_enum *)kcontrol->private_value;
sound/soc/codecs/lpass-rx-macro.c
1694
if (e->reg == CDC_RX_RXn_RX_PATH_CFG1(rx, 0))
sound/soc/codecs/lpass-rx-macro.c
1696
else if (e->reg == CDC_RX_RXn_RX_PATH_CFG1(rx, 1))
sound/soc/codecs/lpass-rx-macro.c
2471
struct soc_enum *e = (struct soc_enum *)kcontrol->private_value;
sound/soc/codecs/lpass-rx-macro.c
2518
rx_port_value, e, update);
sound/soc/codecs/lpass-tx-macro.c
1075
struct soc_enum *e = (struct soc_enum *)kcontrol->private_value;
sound/soc/codecs/lpass-tx-macro.c
1076
int path = e->shift_l;
sound/soc/codecs/lpass-tx-macro.c
1088
struct soc_enum *e = (struct soc_enum *)kcontrol->private_value;
sound/soc/codecs/lpass-tx-macro.c
1089
int path = e->shift_l;
sound/soc/codecs/lpass-tx-macro.c
805
struct soc_enum *e = (struct soc_enum *)kcontrol->private_value;
sound/soc/codecs/lpass-tx-macro.c
811
if (val >= e->items)
sound/soc/codecs/lpass-tx-macro.c
814
switch (e->reg) {
sound/soc/codecs/lpass-va-macro.c
530
struct soc_enum *e = (struct soc_enum *)kcontrol->private_value;
sound/soc/codecs/lpass-va-macro.c
536
switch (e->reg) {
sound/soc/codecs/lpass-va-macro.c
551
__func__, e->reg);
sound/soc/codecs/lpass-va-macro.c
829
struct soc_enum *e = (struct soc_enum *)kcontrol->private_value;
sound/soc/codecs/lpass-va-macro.c
830
int path = e->shift_l;
sound/soc/codecs/lpass-va-macro.c
842
struct soc_enum *e = (struct soc_enum *)kcontrol->private_value;
sound/soc/codecs/lpass-va-macro.c
843
int path = e->shift_l;
sound/soc/codecs/lpass-wsa-macro.c
2102
struct soc_enum *e = (struct soc_enum *)kcontrol->private_value;
sound/soc/codecs/lpass-wsa-macro.c
2150
rx_port_value, e, update);
sound/soc/codecs/madera.c
1052
struct soc_enum *e = (struct soc_enum *)kcontrol->private_value;
sound/soc/codecs/madera.c
1057
if (item >= e->items)
sound/soc/codecs/madera.c
1066
val = snd_soc_component_read(component, e->reg);
sound/soc/codecs/madera.c
1067
val >>= e->shift_l;
sound/soc/codecs/madera.c
1068
val &= e->mask;
sound/soc/codecs/madera.c
1069
if (snd_soc_enum_item_to_val(e, item) == val) {
sound/soc/codecs/madera.c
1074
if (!madera_can_change_grp_rate(priv, e->reg)) {
sound/soc/codecs/madera.c
2159
struct soc_enum *e = (struct soc_enum *)kcontrol->private_value;
sound/soc/codecs/madera.c
2160
unsigned int reg = e->reg;
sound/soc/codecs/madera.c
548
struct soc_enum *e = (struct soc_enum *)kcontrol->private_value;
sound/soc/codecs/madera.c
553
if (ucontrol->value.enumerated.item[0] > e->items - 1)
sound/soc/codecs/madera.c
620
ret = snd_soc_dapm_mux_update_power(dapm, kcontrol, mux, e, NULL);
sound/soc/codecs/madera.c
653
struct soc_enum *e = (struct soc_enum *)kcontrol->private_value;
sound/soc/codecs/madera.c
663
val = mux << e->shift_l;
sound/soc/codecs/madera.c
664
mask = (e->mask << e->shift_l) | MADERA_IN1L_SRC_SE_MASK;
sound/soc/codecs/madera.c
666
switch (e->reg) {
sound/soc/codecs/madera.c
687
mux, e->reg, inmode, mask, val);
sound/soc/codecs/madera.c
689
ret = regmap_update_bits_check(regmap, e->reg, mask, val, &changed);
sound/soc/codecs/madera.c
695
mux, e, NULL);
sound/soc/codecs/madera.c
873
struct soc_enum *e = (struct soc_enum *)kcontrol->private_value;
sound/soc/codecs/madera.c
875
const int adsp_num = e->shift_l;
sound/soc/codecs/madera.c
882
item = snd_soc_enum_val_to_item(e, cached_rate);
sound/soc/codecs/madera.c
893
struct soc_enum *e = (struct soc_enum *)kcontrol->private_value;
sound/soc/codecs/madera.c
894
const int adsp_num = e->shift_l;
sound/soc/codecs/madera.c
898
if (item >= e->items)
sound/soc/codecs/madera.c
913
} else if (priv->adsp_rate_cache[adsp_num] != e->values[item]) {
sound/soc/codecs/madera.c
915
priv->adsp_rate_cache[adsp_num] = e->values[item];
sound/soc/codecs/max98396.c
1000
snd_soc_dapm_mux_update_power(dapm, kcontrol, item[0], e, NULL);
sound/soc/codecs/max98396.c
978
struct soc_enum *e = (struct soc_enum *)kcontrol->private_value;
sound/soc/codecs/max98396.c
983
if (item[0] >= e->items)
sound/soc/codecs/max98396.c
986
val = snd_soc_enum_item_to_val(e, item[0]) << e->shift_l;
sound/soc/codecs/peb2466.c
382
struct soc_enum *e = (struct soc_enum *)kcontrol->private_value;
sound/soc/codecs/peb2466.c
384
switch (e->reg) {
sound/soc/codecs/peb2466.c
420
struct soc_enum *e = (struct soc_enum *)kcontrol->private_value;
sound/soc/codecs/peb2466.c
431
switch (e->reg) {
sound/soc/codecs/peb2466.c
484
ret = peb2466_write_buf(peb2466, e->reg, peb2466_tone_lookup[index], 4);
sound/soc/codecs/rt700.c
556
struct soc_enum *e = (struct soc_enum *)kcontrol->private_value;
sound/soc/codecs/rt700.c
561
if (item[0] >= e->items)
sound/soc/codecs/rt700.c
574
val = snd_soc_enum_item_to_val(e, item[0]) << e->shift_l;
sound/soc/codecs/rt700.c
592
item[0], e, NULL);
sound/soc/codecs/rt711-sdca.c
815
struct soc_enum *e = (struct soc_enum *)kcontrol->private_value;
sound/soc/codecs/rt711-sdca.c
820
val = (rt711->ge_mode_override >> e->shift_l) & e->mask;
sound/soc/codecs/rt711-sdca.c
821
item = snd_soc_enum_val_to_item(e, val);
sound/soc/codecs/rt711-sdca.c
829
struct soc_enum *e = (struct soc_enum *)kcontrol->private_value;
sound/soc/codecs/rt711-sdca.c
835
if (item[0] >= e->items)
sound/soc/codecs/rt711-sdca.c
838
val = snd_soc_enum_item_to_val(e, item[0]) << e->shift_l;
sound/soc/codecs/rt711-sdca.c
928
struct soc_enum *e = (struct soc_enum *)kcontrol->private_value;
sound/soc/codecs/rt711-sdca.c
932
if (item[0] >= e->items)
sound/soc/codecs/rt711-sdca.c
942
val = snd_soc_enum_item_to_val(e, item[0]) << e->shift_l;
sound/soc/codecs/rt711-sdca.c
959
item[0], e, NULL);
sound/soc/codecs/rt711.c
715
struct soc_enum *e = (struct soc_enum *)kcontrol->private_value;
sound/soc/codecs/rt711.c
720
if (item[0] >= e->items)
sound/soc/codecs/rt711.c
731
val = snd_soc_enum_item_to_val(e, item[0]) << e->shift_l;
sound/soc/codecs/rt711.c
752
item[0], e, NULL);
sound/soc/codecs/rt712-sdca-dmic.c
457
struct soc_enum *e = (struct soc_enum *)kcontrol->private_value;
sound/soc/codecs/rt712-sdca-dmic.c
461
if (item[0] >= e->items)
sound/soc/codecs/rt712-sdca-dmic.c
471
val = snd_soc_enum_item_to_val(e, item[0]) << e->shift_l;
sound/soc/codecs/rt712-sdca-dmic.c
488
item[0], e, NULL);
sound/soc/codecs/rt712-sdca.c
1255
struct soc_enum *e = (struct soc_enum *)kcontrol->private_value;
sound/soc/codecs/rt712-sdca.c
1259
if (item[0] >= e->items)
sound/soc/codecs/rt712-sdca.c
1269
val = snd_soc_enum_item_to_val(e, item[0]) << e->shift_l;
sound/soc/codecs/rt712-sdca.c
1290
snd_soc_dapm_mux_update_power(dapm, kcontrol, item[0], e, NULL);
sound/soc/codecs/rt712-sdca.c
773
struct soc_enum *e = (struct soc_enum *)kcontrol->private_value;
sound/soc/codecs/rt712-sdca.c
778
if (item[0] >= e->items)
sound/soc/codecs/rt712-sdca.c
799
item[0], e, NULL);
sound/soc/codecs/rt715-sdca.c
543
struct soc_enum *e = (struct soc_enum *)kcontrol->private_value;
sound/soc/codecs/rt715-sdca.c
547
if (item[0] >= e->items)
sound/soc/codecs/rt715-sdca.c
562
val = snd_soc_enum_item_to_val(e, item[0]) << e->shift_l;
sound/soc/codecs/rt715-sdca.c
574
snd_soc_dapm_mux_update_power(dapm, kcontrol, item[0], e, NULL);
sound/soc/codecs/rt715.c
548
struct soc_enum *e = (struct soc_enum *)kcontrol->private_value;
sound/soc/codecs/rt715.c
553
reg = RT715_VERB_SET_CONNECT_SEL | e->reg;
sound/soc/codecs/rt715.c
566
if ((e->reg == RT715_MUX_IN3 || e->reg == RT715_MUX_IN4) && (val > 0))
sound/soc/codecs/rt715.c
579
struct soc_enum *e = (struct soc_enum *)kcontrol->private_value;
sound/soc/codecs/rt715.c
584
if (item[0] >= e->items)
sound/soc/codecs/rt715.c
588
val = snd_soc_enum_item_to_val(e, item[0]) << e->shift_l;
sound/soc/codecs/rt715.c
590
reg = RT715_VERB_SET_CONNECT_SEL | e->reg;
sound/soc/codecs/rt715.c
604
reg = RT715_VERB_SET_CONNECT_SEL | e->reg;
sound/soc/codecs/rt715.c
609
item[0], e, NULL);
sound/soc/codecs/rt721-sdca.c
791
struct soc_enum *e = (struct soc_enum *)kcontrol->private_value;
sound/soc/codecs/rt721-sdca.c
796
if (item[0] >= e->items)
sound/soc/codecs/rt721-sdca.c
823
val = snd_soc_enum_item_to_val(e, item[0]) << e->shift_l;
sound/soc/codecs/rt721-sdca.c
846
item[0], e, NULL);
sound/soc/codecs/simple-mux.c
50
struct soc_enum *e = (struct soc_enum *)kcontrol->private_value;
sound/soc/codecs/simple-mux.c
54
if (ucontrol->value.enumerated.item[0] > e->items)
sound/soc/codecs/simple-mux.c
70
e, NULL);
sound/soc/codecs/twl6040.c
327
struct soc_enum *e = (struct soc_enum *)kcontrol->private_value;
sound/soc/codecs/twl6040.c
331
val = twl6040_read(component, e->reg);
sound/soc/codecs/wcd9335.c
1279
struct soc_enum *e = (struct soc_enum *)kc->private_value;
sound/soc/codecs/wcd9335.c
1317
e, update);
sound/soc/codecs/wcd9335.c
1483
struct soc_enum *e = (struct soc_enum *)kc->private_value;
sound/soc/codecs/wcd9335.c
1488
switch (e->reg) {
sound/soc/codecs/wcd9335.c
1532
struct soc_enum *e = (struct soc_enum *)kc->private_value;
sound/soc/codecs/wcd9335.c
1539
if (e->reg == WCD9335_CDC_RX0_RX_PATH_SEC0)
sound/soc/codecs/wcd9335.c
1541
else if (e->reg == WCD9335_CDC_RX1_RX_PATH_SEC0)
sound/soc/codecs/wcd9335.c
1543
else if (e->reg == WCD9335_CDC_RX2_RX_PATH_SEC0)
sound/soc/codecs/wcd934x.c
3294
struct soc_enum *e = (struct soc_enum *)kc->private_value;
sound/soc/codecs/wcd934x.c
3346
e, update);
sound/soc/codecs/wcd934x.c
3356
struct soc_enum *e = (struct soc_enum *)kc->private_value;
sound/soc/codecs/wcd934x.c
3362
if (e->reg == WCD934X_CDC_RX0_RX_PATH_SEC0)
sound/soc/codecs/wcd934x.c
3364
else if (e->reg == WCD934X_CDC_RX1_RX_PATH_SEC0)
sound/soc/codecs/wcd934x.c
3366
else if (e->reg == WCD934X_CDC_RX2_RX_PATH_SEC0)
sound/soc/codecs/wcd934x.c
3388
struct soc_enum *e = (struct soc_enum *)kcontrol->private_value;
sound/soc/codecs/wcd934x.c
3396
if (val > e->items - 1)
sound/soc/codecs/wcd934x.c
3399
switch (e->reg) {
sound/soc/codecs/wcd934x.c
3401
if (e->shift_l == 0)
sound/soc/codecs/wcd934x.c
3403
else if (e->shift_l == 2)
sound/soc/codecs/wcd934x.c
3405
else if (e->shift_l == 4)
sound/soc/codecs/wcd934x.c
3409
if (e->shift_l == 0)
sound/soc/codecs/wcd934x.c
3411
else if (e->shift_l == 2)
sound/soc/codecs/wcd934x.c
3415
if (e->shift_l == 0)
sound/soc/codecs/wcd934x.c
3417
else if (e->shift_l == 2)
sound/soc/codecs/wcd934x.c
3421
if (e->shift_l == 0)
sound/soc/codecs/wcd934x.c
3423
else if (e->shift_l == 2)
sound/soc/codecs/wcd934x.c
3428
__func__, e->reg);
sound/soc/codecs/wcd938x.c
1484
struct soc_enum *e = (struct soc_enum *)kcontrol->private_value;
sound/soc/codecs/wcd938x.c
1485
int path = e->shift_l;
sound/soc/codecs/wcd938x.c
1497
struct soc_enum *e = (struct soc_enum *)kcontrol->private_value;
sound/soc/codecs/wcd938x.c
1498
int path = e->shift_l;
sound/soc/codecs/wcd939x.c
1438
struct soc_enum *e = (struct soc_enum *)kcontrol->private_value;
sound/soc/codecs/wcd939x.c
1439
int path = e->shift_l;
sound/soc/codecs/wcd939x.c
1451
struct soc_enum *e = (struct soc_enum *)kcontrol->private_value;
sound/soc/codecs/wcd939x.c
1452
int path = e->shift_l;
sound/soc/codecs/wm8804.c
162
struct soc_enum *e = (struct soc_enum *)kcontrol->private_value;
sound/soc/codecs/wm8804.c
163
unsigned int val = ucontrol->value.enumerated.item[0] << e->shift_l;
sound/soc/codecs/wm8804.c
164
unsigned int mask = 1 << e->shift_l;
sound/soc/codecs/wm8804.c
172
if (snd_soc_component_test_bits(component, e->reg, mask, val)) {
sound/soc/codecs/wm8804.c
180
snd_soc_component_update_bits(component, e->reg, mask, val);
sound/soc/codecs/wm8998.c
108
struct soc_enum *e = (struct soc_enum *)kcontrol->private_value;
sound/soc/codecs/wm8998.c
117
switch (e->reg) {
sound/soc/codecs/wm8998.c
141
change = snd_soc_component_update_bits(component, e->reg,
sound/soc/codecs/wm8998.c
148
e, NULL);
sound/soc/codecs/wm_adsp.c
323
struct soc_enum *e = (struct soc_enum *)kcontrol->private_value;
sound/soc/codecs/wm_adsp.c
326
ucontrol->value.enumerated.item[0] = dsp[e->shift_l].fw;
sound/soc/codecs/wm_adsp.c
336
struct soc_enum *e = (struct soc_enum *)kcontrol->private_value;
sound/soc/codecs/wm_adsp.c
340
if (ucontrol->value.enumerated.item[0] == dsp[e->shift_l].fw)
sound/soc/codecs/wm_adsp.c
346
mutex_lock(&dsp[e->shift_l].cs_dsp.pwr_lock);
sound/soc/codecs/wm_adsp.c
348
if (dsp[e->shift_l].cs_dsp.booted || !list_empty(&dsp[e->shift_l].compr_list))
sound/soc/codecs/wm_adsp.c
351
dsp[e->shift_l].fw = ucontrol->value.enumerated.item[0];
sound/soc/codecs/wm_adsp.c
353
mutex_unlock(&dsp[e->shift_l].cs_dsp.pwr_lock);
sound/soc/fsl/fsl_audmix.c
116
struct soc_enum *e = (struct soc_enum *)kcontrol->private_value;
sound/soc/fsl/fsl_audmix.c
124
val = snd_soc_enum_item_to_val(e, item[0]);
sound/soc/fsl/fsl_audmix.c
154
struct soc_enum *e = (struct soc_enum *)kcontrol->private_value;
sound/soc/fsl/fsl_audmix.c
170
val = snd_soc_enum_item_to_val(e, item[0]);
sound/soc/fsl/fsl_micfil.c
335
struct soc_enum *e = (struct soc_enum *)kcontrol->private_value;
sound/soc/fsl/fsl_micfil.c
339
int val = snd_soc_enum_item_to_val(e, item[0]);
sound/soc/fsl/fsl_micfil.c
376
struct soc_enum *e = (struct soc_enum *)kcontrol->private_value;
sound/soc/fsl/fsl_micfil.c
379
int val = snd_soc_enum_item_to_val(e, item[0]);
sound/soc/fsl/fsl_micfil.c
401
struct soc_enum *e = (struct soc_enum *)kcontrol->private_value;
sound/soc/fsl/fsl_micfil.c
404
int val = snd_soc_enum_item_to_val(e, item[0]);
sound/soc/fsl/fsl_xcvr.c
116
struct soc_enum *e = (struct soc_enum *)kcontrol->private_value;
sound/soc/fsl/fsl_xcvr.c
119
xcvr->arc_mode = snd_soc_enum_item_to_val(e, item[0]);
sound/soc/fsl/fsl_xcvr.c
219
struct soc_enum *e = (struct soc_enum *)kcontrol->private_value;
sound/soc/fsl/fsl_xcvr.c
224
xcvr->mode = snd_soc_enum_item_to_val(e, item[0]);
sound/soc/intel/atom/sst-atom-controls.c
137
struct sst_enum *e = (struct sst_enum *)kcontrol->private_value;
sound/soc/intel/atom/sst-atom-controls.c
141
uinfo->value.enumerated.items = e->max;
sound/soc/intel/atom/sst-atom-controls.c
143
if (uinfo->value.enumerated.item > e->max - 1)
sound/soc/intel/atom/sst-atom-controls.c
144
uinfo->value.enumerated.item = e->max - 1;
sound/soc/intel/atom/sst-atom-controls.c
146
e->texts[uinfo->value.enumerated.item]);
sound/soc/intel/atom/sst-atom-controls.c
1461
struct sst_enum *e = (void *)kctl->private_value;
sound/soc/intel/atom/sst-atom-controls.c
1463
e->w = w;
sound/soc/intel/atom/sst-atom-controls.c
1466
struct sst_enum *e = (void *)kctl->private_value;
sound/soc/intel/atom/sst-atom-controls.c
1468
e->w = w;
sound/soc/intel/atom/sst-atom-controls.c
162
struct sst_enum *e = (void *)kcontrol->private_value;
sound/soc/intel/atom/sst-atom-controls.c
165
unsigned int ctl_no = e->reg;
sound/soc/intel/atom/sst-atom-controls.c
166
unsigned int is_tx = e->tx;
sound/soc/intel/atom/sst-atom-controls.c
173
for (mux = e->max; mux > 0; mux--)
sound/soc/intel/atom/sst-atom-controls.c
182
e->texts[mux], mux ? map[mux - 1] : -1);
sound/soc/intel/atom/sst-atom-controls.c
193
struct sst_enum *e = (void *)kcontrol->private_value;
sound/soc/intel/atom/sst-atom-controls.c
196
if (e->w && e->w->power)
sound/soc/intel/atom/sst-atom-controls.c
198
else if (!e->w)
sound/soc/intel/atom/sst-atom-controls.c
223
struct sst_enum *e = (void *)kcontrol->private_value;
sound/soc/intel/atom/sst-atom-controls.c
225
unsigned int ctl_no = e->reg;
sound/soc/intel/atom/sst-atom-controls.c
226
unsigned int is_tx = e->tx;
sound/soc/intel/atom/sst-atom-controls.c
235
if (mux > e->max - 1)
sound/soc/intel/atom/sst-atom-controls.c
240
for (i = 0; i < e->max; i++)
sound/soc/intel/atom/sst-atom-controls.c
257
e->texts[mux], map[slot_channel_no]);
sound/soc/intel/catpt/registers.h
66
#define CATPT_HMDC_HDDA(e, ch) BIT(8 * (e) + (ch))
sound/soc/mediatek/common/mtk-btcvsd.c
1061
struct soc_enum *e = (struct soc_enum *)kcontrol->private_value;
sound/soc/mediatek/common/mtk-btcvsd.c
1063
if (ucontrol->value.enumerated.item[0] >= e->items)
sound/soc/mediatek/mt8183/mt8183-dai-adda.c
131
struct soc_enum *e = (struct soc_enum *)kcontrol->private_value;
sound/soc/mediatek/mt8183/mt8183-dai-adda.c
133
if (ucontrol->value.enumerated.item[0] >= e->items)
sound/soc/mediatek/mt8183/mt8183-dai-i2s.c
143
struct soc_enum *e = (struct soc_enum *)kcontrol->private_value;
sound/soc/mediatek/mt8183/mt8183-dai-i2s.c
146
if (ucontrol->value.enumerated.item[0] >= e->items)
sound/soc/mediatek/mt8186/mt8186-dai-i2s.c
132
struct soc_enum *e = (struct soc_enum *)kcontrol->private_value;
sound/soc/mediatek/mt8186/mt8186-dai-i2s.c
135
if (ucontrol->value.enumerated.item[0] >= e->items)
sound/soc/mediatek/mt8186/mt8186-dai-tdm.c
294
struct soc_enum *e = (struct soc_enum *)kcontrol->private_value;
sound/soc/mediatek/mt8186/mt8186-dai-tdm.c
297
if (ucontrol->value.enumerated.item[0] >= e->items)
sound/soc/mediatek/mt8186/mt8186-misc-control.c
146
struct soc_enum *e = (struct soc_enum *)kcontrol->private_value;
sound/soc/mediatek/mt8186/mt8186-misc-control.c
149
if (ucontrol->value.enumerated.item[0] >= e->items)
sound/soc/mediatek/mt8186/mt8186-misc-control.c
189
struct soc_enum *e = (struct soc_enum *)kcontrol->private_value;
sound/soc/mediatek/mt8186/mt8186-misc-control.c
192
if (ucontrol->value.enumerated.item[0] >= e->items)
sound/soc/mediatek/mt8186/mt8186-misc-control.c
92
struct soc_enum *e = (struct soc_enum *)kcontrol->private_value;
sound/soc/mediatek/mt8186/mt8186-misc-control.c
96
if (ucontrol->value.enumerated.item[0] >= e->items)
sound/soc/mediatek/mt8188/mt8188-afe-pcm.c
1674
struct soc_enum *e = (struct soc_enum *)kcontrol->private_value;
sound/soc/mediatek/mt8188/mt8188-afe-pcm.c
1679
snd_soc_enum_val_to_item(e, memif_priv->fs_timing);
sound/soc/mediatek/mt8188/mt8188-afe-pcm.c
1692
struct soc_enum *e = (struct soc_enum *)kcontrol->private_value;
sound/soc/mediatek/mt8188/mt8188-afe-pcm.c
1696
if (item[0] >= e->items)
sound/soc/mediatek/mt8188/mt8188-afe-pcm.c
1701
prev_item = snd_soc_enum_val_to_item(e, memif_priv->fs_timing);
sound/soc/mediatek/mt8188/mt8188-afe-pcm.c
1706
memif_priv->fs_timing = snd_soc_enum_item_to_val(e, item[0]);
sound/soc/mediatek/mt8188/mt8188-dai-dmic.c
542
struct soc_enum *e = (struct soc_enum *)kcontrol->private_value;
sound/soc/mediatek/mt8188/mt8188-dai-dmic.c
549
if (source >= e->items)
sound/soc/mediatek/mt8188/mt8188-dai-etdm.c
1016
struct soc_enum *e = (struct soc_enum *)kcontrol->private_value;
sound/soc/mediatek/mt8188/mt8188-dai-etdm.c
1024
if (source >= e->items)
sound/soc/mediatek/mt8192/mt8192-dai-i2s.c
160
struct soc_enum *e = (struct soc_enum *)kcontrol->private_value;
sound/soc/mediatek/mt8192/mt8192-dai-i2s.c
163
if (ucontrol->value.enumerated.item[0] >= e->items)
sound/soc/mediatek/mt8195/mt8195-dai-etdm.c
699
struct soc_enum *e = (struct soc_enum *)kcontrol->private_value;
sound/soc/mediatek/mt8195/mt8195-dai-etdm.c
706
if (source >= e->items)
sound/soc/meson/aiu-acodec-ctrl.c
36
struct soc_enum *e = (struct soc_enum *)kcontrol->private_value;
sound/soc/meson/aiu-acodec-ctrl.c
39
mux = snd_soc_enum_item_to_val(e, ucontrol->value.enumerated.item[0]);
sound/soc/meson/aiu-acodec-ctrl.c
40
changed = snd_soc_component_test_bits(component, e->reg,
sound/soc/meson/aiu-acodec-ctrl.c
51
snd_soc_component_update_bits(component, e->reg,
sound/soc/meson/aiu-acodec-ctrl.c
57
snd_soc_dapm_mux_update_power(dapm, kcontrol, mux, e, NULL);
sound/soc/meson/aiu-codec-ctrl.c
28
struct soc_enum *e = (struct soc_enum *)kcontrol->private_value;
sound/soc/meson/aiu-codec-ctrl.c
31
mux = snd_soc_enum_item_to_val(e, ucontrol->value.enumerated.item[0]);
sound/soc/meson/aiu-codec-ctrl.c
32
changed = snd_soc_component_test_bits(component, e->reg,
sound/soc/meson/aiu-codec-ctrl.c
43
snd_soc_component_update_bits(component, e->reg,
sound/soc/meson/aiu-codec-ctrl.c
50
snd_soc_component_update_bits(component, e->reg,
sound/soc/meson/aiu-codec-ctrl.c
56
snd_soc_dapm_mux_update_power(dapm, kcontrol, mux, e, NULL);
sound/soc/meson/g12a-toacodec.c
103
snd_soc_dapm_mux_update_power(dapm, kcontrol, mux, e, NULL);
sound/soc/meson/g12a-toacodec.c
69
struct soc_enum *e = (struct soc_enum *)kcontrol->private_value;
sound/soc/meson/g12a-toacodec.c
72
if (ucontrol->value.enumerated.item[0] >= e->items)
sound/soc/meson/g12a-toacodec.c
75
mux = snd_soc_enum_item_to_val(e, ucontrol->value.enumerated.item[0]);
sound/soc/meson/g12a-toacodec.c
99
snd_soc_component_update_bits(component, e->reg,
sound/soc/meson/g12a-tohdmitx.c
115
snd_soc_dapm_mux_update_power(dapm, kcontrol, mux, e, NULL);
sound/soc/meson/g12a-tohdmitx.c
43
struct soc_enum *e = (struct soc_enum *)kcontrol->private_value;
sound/soc/meson/g12a-tohdmitx.c
46
if (ucontrol->value.enumerated.item[0] >= e->items)
sound/soc/meson/g12a-tohdmitx.c
49
mux = snd_soc_enum_item_to_val(e, ucontrol->value.enumerated.item[0]);
sound/soc/meson/g12a-tohdmitx.c
50
changed = snd_soc_component_test_bits(component, e->reg,
sound/soc/meson/g12a-tohdmitx.c
61
snd_soc_component_update_bits(component, e->reg,
sound/soc/meson/g12a-tohdmitx.c
69
snd_soc_dapm_mux_update_power(dapm, kcontrol, mux, e, NULL);
sound/soc/meson/g12a-tohdmitx.c
92
struct soc_enum *e = (struct soc_enum *)kcontrol->private_value;
sound/soc/meson/g12a-tohdmitx.c
95
if (ucontrol->value.enumerated.item[0] >= e->items)
sound/soc/meson/g12a-tohdmitx.c
98
mux = snd_soc_enum_item_to_val(e, ucontrol->value.enumerated.item[0]);
sound/soc/sdca/sdca_asoc.c
125
struct soc_enum *e = (struct soc_enum *)kcontrol->private_value;
sound/soc/sdca/sdca_asoc.c
127
unsigned int reg = e->reg;
sound/soc/sdca/sdca_asoc.c
147
ret = snd_soc_enum_item_to_val(e, item[0]);
sound/soc/soc-dapm.c
2637
int mux, struct soc_enum *e)
sound/soc/soc-dapm.c
2649
if (e && !(strcmp(path->name, e->texts[mux])))
sound/soc/soc-dapm.c
2664
struct snd_kcontrol *kcontrol, int mux, struct soc_enum *e,
sound/soc/soc-dapm.c
2671
ret = dapm_mux_update_power(card, kcontrol, update, mux, e);
sound/soc/soc-dapm.c
3546
struct soc_enum *e = (struct soc_enum *)kcontrol->private_value;
sound/soc/soc-dapm.c
3550
if (e->reg != SND_SOC_NOPM && dapm_kcontrol_is_powered(kcontrol)) {
sound/soc/soc-dapm.c
3551
reg_val = dapm_read(dapm, e->reg);
sound/soc/soc-dapm.c
3557
val = (reg_val >> e->shift_l) & e->mask;
sound/soc/soc-dapm.c
3558
ucontrol->value.enumerated.item[0] = snd_soc_enum_val_to_item(e, val);
sound/soc/soc-dapm.c
3559
if (e->shift_l != e->shift_r) {
sound/soc/soc-dapm.c
3560
val = (reg_val >> e->shift_r) & e->mask;
sound/soc/soc-dapm.c
3561
val = snd_soc_enum_val_to_item(e, val);
sound/soc/soc-dapm.c
3583
struct soc_enum *e = (struct soc_enum *)kcontrol->private_value;
sound/soc/soc-dapm.c
3591
if (item[0] >= e->items)
sound/soc/soc-dapm.c
3594
val = snd_soc_enum_item_to_val(e, item[0]) << e->shift_l;
sound/soc/soc-dapm.c
3595
mask = e->mask << e->shift_l;
sound/soc/soc-dapm.c
3596
if (e->shift_l != e->shift_r) {
sound/soc/soc-dapm.c
3597
if (item[1] > e->items)
sound/soc/soc-dapm.c
3599
val |= snd_soc_enum_item_to_val(e, item[1]) << e->shift_r;
sound/soc/soc-dapm.c
3600
mask |= e->mask << e->shift_r;
sound/soc/soc-dapm.c
3607
if (e->reg != SND_SOC_NOPM)
sound/soc/soc-dapm.c
3608
reg_change = dapm_test_bits(dapm, e->reg, mask, val);
sound/soc/soc-dapm.c
3613
update.reg = e->reg;
sound/soc/soc-dapm.c
3618
ret = dapm_mux_update_power(card, kcontrol, pupdate, item[0], e);
sound/soc/soc-dapm.c
451
struct soc_enum *e = (struct soc_enum *)kcontrol->private_value;
sound/soc/soc-dapm.c
455
if (e->reg != SND_SOC_NOPM) {
sound/soc/soc-dapm.c
458
val = dapm_read(dapm, e->reg);
sound/soc/soc-dapm.c
459
val = (val >> e->shift_l) & e->mask;
sound/soc/soc-dapm.c
460
item = snd_soc_enum_val_to_item(e, val);
sound/soc/soc-dapm.c
471
i = match_string(e->texts, e->items, control_name);
sound/soc/soc-dapm.c
475
path->name = e->texts[i];
sound/soc/soc-dapm.c
712
struct soc_enum *e;
sound/soc/soc-dapm.c
769
e = (struct soc_enum *)kcontrol->private_value;
sound/soc/soc-dapm.c
771
if (e->autodisable) {
sound/soc/soc-dapm.c
782
template.reg = e->reg;
sound/soc/soc-dapm.c
783
template.mask = e->mask;
sound/soc/soc-dapm.c
784
template.shift = e->shift_l;
sound/soc/soc-dapm.c
785
template.off_val = snd_soc_enum_item_to_val(e, 0);
sound/soc/soc-dapm.c
802
} else if (e->reg != SND_SOC_NOPM) {
sound/soc/soc-dapm.c
803
data->value = dapm_read(widget->dapm, e->reg) &
sound/soc/soc-dapm.c
804
(e->mask << e->shift_l);
sound/soc/soc-ops.c
100
val = snd_soc_enum_item_to_val(e, item[0]) << e->shift_l;
sound/soc/soc-ops.c
101
mask = e->mask << e->shift_l;
sound/soc/soc-ops.c
102
if (e->shift_l != e->shift_r) {
sound/soc/soc-ops.c
103
if (item[1] >= e->items)
sound/soc/soc-ops.c
105
val |= snd_soc_enum_item_to_val(e, item[1]) << e->shift_r;
sound/soc/soc-ops.c
106
mask |= e->mask << e->shift_r;
sound/soc/soc-ops.c
109
return snd_soc_component_update_bits(component, e->reg, mask, val);
sound/soc/soc-ops.c
42
struct soc_enum *e = (struct soc_enum *)kcontrol->private_value;
sound/soc/soc-ops.c
44
return snd_ctl_enum_info(uinfo, e->shift_l == e->shift_r ? 1 : 2,
sound/soc/soc-ops.c
45
e->items, e->texts);
sound/soc/soc-ops.c
62
struct soc_enum *e = (struct soc_enum *)kcontrol->private_value;
sound/soc/soc-ops.c
66
reg_val = snd_soc_component_read(component, e->reg);
sound/soc/soc-ops.c
67
val = (reg_val >> e->shift_l) & e->mask;
sound/soc/soc-ops.c
68
item = snd_soc_enum_val_to_item(e, val);
sound/soc/soc-ops.c
70
if (e->shift_l != e->shift_r) {
sound/soc/soc-ops.c
71
val = (reg_val >> e->shift_r) & e->mask;
sound/soc/soc-ops.c
72
item = snd_soc_enum_val_to_item(e, val);
sound/soc/soc-ops.c
93
struct soc_enum *e = (struct soc_enum *)kcontrol->private_value;
sound/soc/soc-ops.c
98
if (item[0] >= e->items)
sound/soc/tegra/tegra210_ahub.c
22
struct soc_enum *e = (struct soc_enum *)kctl->private_value;
sound/soc/tegra/tegra210_ahub.c
32
reg = e->reg + (ahub->soc_data->xbar_part_size * i);
sound/soc/tegra/tegra210_ahub.c
44
for (i = 0; i < e->items; i++) {
sound/soc/tegra/tegra210_ahub.c
45
if (bit_pos == e->values[i]) {
sound/soc/tegra/tegra210_ahub.c
60
struct soc_enum *e = (struct soc_enum *)kctl->private_value;
sound/soc/tegra/tegra210_ahub.c
63
unsigned int value = e->values[item[0]];
sound/soc/tegra/tegra210_ahub.c
67
if (item[0] >= e->items)
sound/soc/tegra/tegra210_ahub.c
83
update[i].reg = e->reg + (ahub->soc_data->xbar_part_size * i);
sound/soc/tegra/tegra210_ahub.c
93
item[0], e,
sound/soc/tegra/tegra210_mbdrc.c
272
struct soc_enum *e = (struct soc_enum *)kcontrol->private_value;
sound/soc/tegra/tegra210_mbdrc.c
275
regmap_read(ope->mbdrc_regmap, e->reg, &val);
sound/soc/tegra/tegra210_mbdrc.c
277
ucontrol->value.enumerated.item[0] = (val >> e->shift_l) & e->mask;
sound/soc/tegra/tegra210_mbdrc.c
287
struct soc_enum *e = (struct soc_enum *)kcontrol->private_value;
sound/soc/tegra/tegra210_mbdrc.c
292
if (ucontrol->value.enumerated.item[0] > e->items - 1)
sound/soc/tegra/tegra210_mbdrc.c
295
val = ucontrol->value.enumerated.item[0] << e->shift_l;
sound/soc/tegra/tegra210_mbdrc.c
296
mask = e->mask << e->shift_l;
sound/soc/tegra/tegra210_mbdrc.c
298
regmap_update_bits_check(ope->mbdrc_regmap, e->reg, mask, val,
sound/usb/mixer_s1810c.c
177
u32 b, u32 c, u32 d, u32 e)
sound/usb/mixer_s1810c.c
189
pkt.e = __cpu_to_le32(e);
sound/usb/mixer_s1810c.c
252
u32 a, b, c, e, n, off, left, right;
sound/usb/mixer_s1810c.c
259
e = MIXER_LEVEL_N99DB;
sound/usb/mixer_s1810c.c
265
snd_s1810c_send_ctl_packet(dev, a, b, c, 0, e);
sound/usb/mixer_s1810c.c
266
snd_s1810c_send_ctl_packet(dev, a, b, c, 1, e);
sound/usb/mixer_s1810c.c
269
snd_s1810c_send_ctl_packet(dev, a, b, 0, 0, e);
sound/usb/mixer_s1810c.c
270
snd_s1810c_send_ctl_packet(dev, a, b, 0, 1, e);
sound/usb/mixer_s1810c.c
276
e = MIXER_LEVEL_N3DB;
sound/usb/mixer_s1810c.c
290
e = MIXER_LEVEL_N99DB;
sound/usb/mixer_s1810c.c
295
snd_s1810c_send_ctl_packet(dev, a, b, c, 0, e);
sound/usb/mixer_s1810c.c
296
snd_s1810c_send_ctl_packet(dev, a, b, c, 1, e);
sound/usb/mixer_s1810c.c
298
e = MIXER_LEVEL_N3DB;
sound/usb/mixer_s1810c.c
344
e = MIXER_LEVEL_N99DB;
sound/usb/mixer_s1810c.c
349
snd_s1810c_send_ctl_packet(dev, a, b, c, 0, e);
sound/usb/mixer_s1810c.c
350
snd_s1810c_send_ctl_packet(dev, a, b, c, 1, e);
sound/usb/mixer_s1810c.c
352
e = MIXER_LEVEL_N3DB;
sound/usb/mixer_s1810c.c
94
__le32 e;
tools/arch/arm64/include/asm/esr.h
274
#define esr_sys64_to_sysreg(e) \
tools/arch/arm64/include/asm/esr.h
275
sys_reg((((e) & ESR_ELx_SYS64_ISS_OP0_MASK) >> \
tools/arch/arm64/include/asm/esr.h
277
(((e) & ESR_ELx_SYS64_ISS_OP1_MASK) >> \
tools/arch/arm64/include/asm/esr.h
279
(((e) & ESR_ELx_SYS64_ISS_CRN_MASK) >> \
tools/arch/arm64/include/asm/esr.h
281
(((e) & ESR_ELx_SYS64_ISS_CRM_MASK) >> \
tools/arch/arm64/include/asm/esr.h
283
(((e) & ESR_ELx_SYS64_ISS_OP2_MASK) >> \
tools/arch/arm64/include/asm/esr.h
286
#define esr_cp15_to_sysreg(e) \
tools/arch/arm64/include/asm/esr.h
288
(((e) & ESR_ELx_SYS64_ISS_OP1_MASK) >> \
tools/arch/arm64/include/asm/esr.h
290
(((e) & ESR_ELx_SYS64_ISS_CRN_MASK) >> \
tools/arch/arm64/include/asm/esr.h
292
(((e) & ESR_ELx_SYS64_ISS_CRM_MASK) >> \
tools/arch/arm64/include/asm/esr.h
294
(((e) & ESR_ELx_SYS64_ISS_OP2_MASK) >> \
tools/arch/powerpc/include/uapi/asm/kvm.h
260
} e;
tools/arch/x86/include/asm/asm.h
30
#define __ASM_REG(reg) __ASM_SEL_RAW(e##reg, r##reg)
tools/bpf/bpftool/cfg.c
259
struct edge_node *e;
tools/bpf/bpftool/cfg.c
261
e = calloc(1, sizeof(*e));
tools/bpf/bpftool/cfg.c
262
if (!e) {
tools/bpf/bpftool/cfg.c
268
e->src = src;
tools/bpf/bpftool/cfg.c
270
e->dst = dst;
tools/bpf/bpftool/cfg.c
272
e->flags |= flags;
tools/bpf/bpftool/cfg.c
274
return e;
tools/bpf/bpftool/cfg.c
280
struct edge_node *e;
tools/bpf/bpftool/cfg.c
284
e = new_edge(bb, bb_next(bb), EDGE_FLAG_FALLTHROUGH);
tools/bpf/bpftool/cfg.c
285
if (!e)
tools/bpf/bpftool/cfg.c
287
list_add_tail(&e->l, &bb->e_succs);
tools/bpf/bpftool/cfg.c
290
e = new_edge(bb_prev(bb), bb, EDGE_FLAG_FALLTHROUGH);
tools/bpf/bpftool/cfg.c
291
if (!e)
tools/bpf/bpftool/cfg.c
293
list_add_tail(&e->l, &bb->e_prevs);
tools/bpf/bpftool/cfg.c
298
e = new_edge(bb, NULL, EDGE_FLAG_EMPTY);
tools/bpf/bpftool/cfg.c
299
if (!e)
tools/bpf/bpftool/cfg.c
301
e->src = bb;
tools/bpf/bpftool/cfg.c
307
e->dst = bb_next(bb);
tools/bpf/bpftool/cfg.c
308
e->flags |= EDGE_FLAG_FALLTHROUGH;
tools/bpf/bpftool/cfg.c
309
list_add_tail(&e->l, &bb->e_succs);
tools/bpf/bpftool/cfg.c
312
e->dst = func_search_bb_with_head(func,
tools/bpf/bpftool/cfg.c
314
e->flags |= EDGE_FLAG_JUMP;
tools/bpf/bpftool/cfg.c
315
list_add_tail(&e->l, &bb->e_succs);
tools/bpf/bpftool/cfg.c
319
e->dst = bb_next(bb);
tools/bpf/bpftool/cfg.c
320
e->flags |= EDGE_FLAG_FALLTHROUGH;
tools/bpf/bpftool/cfg.c
321
list_add_tail(&e->l, &bb->e_succs);
tools/bpf/bpftool/cfg.c
323
e = new_edge(bb, NULL, EDGE_FLAG_JUMP);
tools/bpf/bpftool/cfg.c
324
if (!e)
tools/bpf/bpftool/cfg.c
326
e->src = bb;
tools/bpf/bpftool/cfg.c
327
e->dst = func_search_bb_with_head(func, insn + insn->off + 1);
tools/bpf/bpftool/cfg.c
328
list_add_tail(&e->l, &bb->e_succs);
tools/bpf/bpftool/cfg.c
363
struct edge_node *e, *e2;
tools/bpf/bpftool/cfg.c
365
list_for_each_entry_safe(e, e2, &bb->e_prevs, l) {
tools/bpf/bpftool/cfg.c
366
list_del(&e->l);
tools/bpf/bpftool/cfg.c
367
free(e);
tools/bpf/bpftool/cfg.c
370
list_for_each_entry_safe(e, e2, &bb->e_succs, l) {
tools/bpf/bpftool/cfg.c
371
list_del(&e->l);
tools/bpf/bpftool/cfg.c
372
free(e);
tools/bpf/bpftool/cfg.c
419
struct edge_node *e;
tools/bpf/bpftool/cfg.c
425
list_for_each_entry(e, &bb->e_succs, l) {
tools/bpf/bpftool/cfg.c
427
func_idx, e->src->idx, func_idx, e->dst->idx,
tools/bpf/bpftool/map_perf_ring.c
103
e->header.type, e->header.size);
tools/bpf/bpftool/map_perf_ring.c
60
struct perf_event_sample *e = container_of(event,
tools/bpf/bpftool/map_perf_ring.c
72
jsonw_uint(json_wtr, e->header.type);
tools/bpf/bpftool/map_perf_ring.c
77
if (e->header.type == PERF_RECORD_SAMPLE) {
tools/bpf/bpftool/map_perf_ring.c
79
jsonw_uint(json_wtr, e->time);
tools/bpf/bpftool/map_perf_ring.c
81
print_data_json(e->data, e->size);
tools/bpf/bpftool/map_perf_ring.c
82
} else if (e->header.type == PERF_RECORD_LOST) {
tools/bpf/bpftool/map_perf_ring.c
93
if (e->header.type == PERF_RECORD_SAMPLE) {
tools/bpf/bpftool/map_perf_ring.c
95
e->time / 1000000000ULL, e->time % 1000000000ULL,
tools/bpf/bpftool/map_perf_ring.c
97
fprint_hex(stdout, e->data, e->size, " ");
tools/bpf/bpftool/map_perf_ring.c
99
} else if (e->header.type == PERF_RECORD_LOST) {
tools/bpf/bpftool/pids.c
102
struct pid_iter_entry *e;
tools/bpf/bpftool/pids.c
103
char buf[4096 / sizeof(*e) * sizeof(*e)];
tools/bpf/bpftool/pids.c
164
if (ret % sizeof(*e)) {
tools/bpf/bpftool/pids.c
169
ret /= sizeof(*e);
tools/bpf/bpftool/pids.c
171
e = (void *)buf;
tools/bpf/bpftool/pids.c
172
for (i = 0; i < ret; i++, e++) {
tools/bpf/bpftool/pids.c
173
add_ref(*map, e);
tools/bpf/bpftool/pids.c
31
static void add_ref(struct hashmap *map, struct pid_iter_entry *e)
tools/bpf/bpftool/pids.c
39
hashmap__for_each_key_entry(map, entry, e->id) {
tools/bpf/bpftool/pids.c
43
if (refs->refs[i].pid == e->pid)
tools/bpf/bpftool/pids.c
50
e->id, e->pid, e->comm);
tools/bpf/bpftool/pids.c
55
ref->pid = e->pid;
tools/bpf/bpftool/pids.c
56
memcpy(ref->comm, e->comm, sizeof(ref->comm));
tools/bpf/bpftool/pids.c
67
e->id, e->pid, e->comm);
tools/bpf/bpftool/pids.c
75
e->id, e->pid, e->comm);
tools/bpf/bpftool/pids.c
79
ref->pid = e->pid;
tools/bpf/bpftool/pids.c
80
memcpy(ref->comm, e->comm, sizeof(ref->comm));
tools/bpf/bpftool/pids.c
83
refs->has_bpf_cookie = e->has_bpf_cookie;
tools/bpf/bpftool/pids.c
84
refs->bpf_cookie = e->bpf_cookie;
tools/bpf/bpftool/pids.c
86
err = hashmap__append(map, e->id, refs);
tools/bpf/bpftool/pids.c
89
e->id, strerror(errno));
tools/bpf/bpftool/sign.c
42
unsigned long e;
tools/bpf/bpftool/sign.c
46
while ((e = ERR_get_error_all(&file, &line, NULL, &data, &flags))) {
tools/bpf/bpftool/sign.c
47
ERR_error_string_n(e, buf, sizeof(buf));
tools/bpf/bpftool/skeleton/pid_iter.bpf.c
101
__builtin_memset(&e, 0, sizeof(e));
tools/bpf/bpftool/skeleton/pid_iter.bpf.c
102
e.pid = task->tgid;
tools/bpf/bpftool/skeleton/pid_iter.bpf.c
103
e.id = get_obj_id(file->private_data, obj_type);
tools/bpf/bpftool/skeleton/pid_iter.bpf.c
112
e.has_bpf_cookie = true;
tools/bpf/bpftool/skeleton/pid_iter.bpf.c
113
e.bpf_cookie = get_bpf_cookie(link);
tools/bpf/bpftool/skeleton/pid_iter.bpf.c
117
bpf_probe_read_kernel_str(&e.comm, sizeof(e.comm),
tools/bpf/bpftool/skeleton/pid_iter.bpf.c
119
bpf_seq_write(ctx->meta->seq, &e, sizeof(e));
tools/bpf/bpftool/skeleton/pid_iter.bpf.c
71
struct pid_iter_entry e;
tools/build/tests/ex/ex.c
17
e();
tools/build/tests/ex/ex.c
7
int e(void);
tools/include/linux/bug.h
9
#define BUILD_BUG_ON_ZERO(e) (sizeof(struct { int:-!!(e); }))
tools/include/linux/build_bug.h
16
#define BUILD_BUG_ON_ZERO(e, ...) \
tools/include/linux/build_bug.h
17
__BUILD_BUG_ON_ZERO_MSG(e, ##__VA_ARGS__, #e " is true")
tools/include/linux/build_bug.h
30
#define BUILD_BUG_ON_INVALID(e) ((void)(sizeof((__force long)(e))))
tools/include/linux/compiler.h
253
#define __BUILD_BUG_ON_ZERO_MSG(e, msg, ...) ((int)(sizeof(struct { int:(-!!(e)); })))
tools/include/linux/compiler.h
255
#define __BUILD_BUG_ON_ZERO_MSG(e, msg, ...) ((int)sizeof(struct {_Static_assert(!(e), msg);}))
tools/lib/bpf/bpf_core_read.h
361
#define ___last6(a, b, c, d, e, x) x
tools/lib/bpf/bpf_core_read.h
362
#define ___last7(a, b, c, d, e, f, x) x
tools/lib/bpf/bpf_core_read.h
363
#define ___last8(a, b, c, d, e, f, g, x) x
tools/lib/bpf/bpf_core_read.h
364
#define ___last9(a, b, c, d, e, f, g, h, x) x
tools/lib/bpf/bpf_core_read.h
365
#define ___last10(a, b, c, d, e, f, g, h, i, x) x
tools/lib/bpf/bpf_core_read.h
372
#define ___nolast6(a, b, c, d, e, _) a, b, c, d, e
tools/lib/bpf/bpf_core_read.h
373
#define ___nolast7(a, b, c, d, e, f, _) a, b, c, d, e, f
tools/lib/bpf/bpf_core_read.h
374
#define ___nolast8(a, b, c, d, e, f, g, _) a, b, c, d, e, f, g
tools/lib/bpf/bpf_core_read.h
375
#define ___nolast9(a, b, c, d, e, f, g, h, _) a, b, c, d, e, f, g, h
tools/lib/bpf/bpf_core_read.h
376
#define ___nolast10(a, b, c, d, e, f, g, h, i, _) a, b, c, d, e, f, g, h, i
tools/lib/bpf/bpf_core_read.h
383
#define ___arrow5(a, b, c, d, e) a->b->c->d->e
tools/lib/bpf/bpf_core_read.h
384
#define ___arrow6(a, b, c, d, e, f) a->b->c->d->e->f
tools/lib/bpf/bpf_core_read.h
385
#define ___arrow7(a, b, c, d, e, f, g) a->b->c->d->e->f->g
tools/lib/bpf/bpf_core_read.h
386
#define ___arrow8(a, b, c, d, e, f, g, h) a->b->c->d->e->f->g->h
tools/lib/bpf/bpf_core_read.h
387
#define ___arrow9(a, b, c, d, e, f, g, h, i) a->b->c->d->e->f->g->h->i
tools/lib/bpf/bpf_core_read.h
388
#define ___arrow10(a, b, c, d, e, f, g, h, i, j) a->b->c->d->e->f->g->h->i->j
tools/lib/bpf/btf.c
352
struct btf_enum *e;
tools/lib/bpf/btf.c
371
for (i = 0, e = btf_enum(t); i < vlen; i++, e++) {
tools/lib/bpf/btf.c
372
e->name_off = bswap_32(e->name_off);
tools/lib/bpf/btf.c
373
e->val = bswap_32(e->val);
tools/lib/bpf/btf.h
589
static inline __u64 btf_enum64_value(const struct btf_enum64 *e)
tools/lib/bpf/btf.h
610
const __u32 *e64 = (const __u32 *)e;
tools/lib/bpf/btf_dump.c
2274
const struct btf_enum *e;
tools/lib/bpf/btf_dump.c
2276
for (i = 0, e = btf_enum(t); i < btf_vlen(t); i++, e++) {
tools/lib/bpf/btf_dump.c
2277
if (value != e->val)
tools/lib/bpf/btf_dump.c
2279
btf_dump_type_values(d, "%s", btf_name_of(d, e->name_off));
tools/lib/bpf/btf_dump.c
2285
const struct btf_enum64 *e;
tools/lib/bpf/btf_dump.c
2287
for (i = 0, e = btf_enum64(t); i < btf_vlen(t); i++, e++) {
tools/lib/bpf/btf_dump.c
2288
if (value != btf_enum64_value(e))
tools/lib/bpf/btf_dump.c
2290
btf_dump_type_values(d, "%s", btf_name_of(d, e->name_off));
tools/lib/bpf/libbpf.c
10859
struct bpf_map *s, *e;
tools/lib/bpf/libbpf.c
10865
e = obj->maps + obj->nr_maps;
tools/lib/bpf/libbpf.c
10867
if ((m < s) || (m >= e)) {
tools/lib/bpf/libbpf.c
13967
perf_buffer__process_record(struct perf_event_header *e, void *ctx)
tools/lib/bpf/libbpf.c
13971
void *data = e;
tools/lib/bpf/libbpf.c
13975
return pb->event_cb(pb->ctx, cpu_buf->cpu, e);
tools/lib/bpf/libbpf.c
13977
switch (e->type) {
tools/lib/bpf/libbpf.c
13993
pr_warn("unknown perf sample type %d\n", e->type);
tools/lib/bpf/libbpf.c
2527
const struct btf_enum *e = btf_enum(t);
tools/lib/bpf/libbpf.c
2529
*res = e->val;
tools/lib/bpf/libbpf.c
2531
const struct btf_enum64 *e = btf_enum64(t);
tools/lib/bpf/libbpf.c
2533
*res = btf_enum64_value(e);
tools/lib/bpf/libbpf_utils.c
188
#define SHA256_ROUND(i, a, b, c, d, e, f, g, h) \
tools/lib/bpf/libbpf_utils.c
190
__u32 tmp = h + Sigma_1(e) + Ch(e, f, g) + sha256_K[i] + w[i]; \
tools/lib/bpf/libbpf_utils.c
202
__u32 e = state[4];
tools/lib/bpf/libbpf_utils.c
215
SHA256_ROUND(i + 0, a, b, c, d, e, f, g, h);
tools/lib/bpf/libbpf_utils.c
216
SHA256_ROUND(i + 1, h, a, b, c, d, e, f, g);
tools/lib/bpf/libbpf_utils.c
217
SHA256_ROUND(i + 2, g, h, a, b, c, d, e, f);
tools/lib/bpf/libbpf_utils.c
218
SHA256_ROUND(i + 3, f, g, h, a, b, c, d, e);
tools/lib/bpf/libbpf_utils.c
219
SHA256_ROUND(i + 4, e, f, g, h, a, b, c, d);
tools/lib/bpf/libbpf_utils.c
220
SHA256_ROUND(i + 5, d, e, f, g, h, a, b, c);
tools/lib/bpf/libbpf_utils.c
221
SHA256_ROUND(i + 6, c, d, e, f, g, h, a, b);
tools/lib/bpf/libbpf_utils.c
222
SHA256_ROUND(i + 7, b, c, d, e, f, g, h, a);
tools/lib/bpf/libbpf_utils.c
228
state[4] += e;
tools/lib/bpf/libbpf_utils.c
27
#define ERRNO_OFFSET(e) ((e) - __LIBBPF_ERRNO__START)
tools/lib/bpf/relo_core.c
1203
const struct btf_enum *e;
tools/lib/bpf/relo_core.c
1206
e = btf_enum(t) + spec->raw_spec[0];
tools/lib/bpf/relo_core.c
1207
s = btf__name_by_offset(spec->btf, e->name_off);
tools/lib/bpf/relo_core.c
1209
append_buf(fmt_str, s, e->val);
tools/lib/bpf/relo_core.c
1211
const struct btf_enum64 *e;
tools/lib/bpf/relo_core.c
1214
e = btf_enum64(t) + spec->raw_spec[0];
tools/lib/bpf/relo_core.c
1215
s = btf__name_by_offset(spec->btf, e->name_off);
tools/lib/bpf/relo_core.c
1217
append_buf(fmt_str, s, (unsigned long long)btf_enum64_value(e));
tools/lib/bpf/ringbuf.c
152
e = &rb->events[rb->ring_cnt];
tools/lib/bpf/ringbuf.c
153
memset(e, 0, sizeof(*e));
tools/lib/bpf/ringbuf.c
155
e->events = EPOLLIN;
tools/lib/bpf/ringbuf.c
156
e->data.fd = rb->ring_cnt;
tools/lib/bpf/ringbuf.c
157
if (epoll_ctl(rb->epoll_fd, EPOLL_CTL_ADD, map_fd, e) < 0) {
tools/lib/bpf/ringbuf.c
80
struct epoll_event *e;
tools/lib/perf/evlist.c
332
#define SID(e, x, y) xyarray__entry(e->sample_id, x, y)
tools/lib/perf/evlist.c
521
#define FD(e, x, y) (*(int *) xyarray__entry(e->fd, x, y))
tools/net/ynl/lib/ynl.h
104
ynl_sock_create(const struct ynl_family *yf, struct ynl_error *e);
tools/objtool/klp-diff.c
1014
klp_relocs = elf_create_section(e->out, KLP_RELOCS_SEC, 0,
tools/objtool/klp-diff.c
1024
if (!elf_add_data(e->out, klp_relocs, &klp_reloc, sizeof(klp_reloc)))
tools/objtool/klp-diff.c
1028
if (!sec->sym && !elf_create_section_symbol(e->out, sec))
tools/objtool/klp-diff.c
1031
if (!elf_create_reloc(e->out, klp_relocs,
tools/objtool/klp-diff.c
1037
if (!elf_create_reloc(e->out, klp_relocs,
tools/objtool/klp-diff.c
1057
static int clone_reloc(struct elfs *e, struct reloc *patched_reloc,
tools/objtool/klp-diff.c
1078
if (clone_reloc_klp(e, patched_reloc, sec, offset, export))
tools/objtool/klp-diff.c
1096
out_sym = clone_symbol(e, patched_sym, patched_sym->included || !export);
tools/objtool/klp-diff.c
1110
addend = elf_add_string(e->out, out_sym->sec, str);
tools/objtool/klp-diff.c
1115
if (!elf_create_reloc(e->out, sec, offset, out_sym, addend,
tools/objtool/klp-diff.c
1123
static int clone_sym_relocs(struct elfs *e, struct symbol *patched_sym)
tools/objtool/klp-diff.c
1168
if (convert_reloc_sym(e->patched, patched_reloc)) {
tools/objtool/klp-diff.c
1177
if (clone_reloc(e, patched_reloc, out_sym->sec, offset))
tools/objtool/klp-diff.c
1352
static int validate_special_section_klp_reloc(struct elfs *e, struct symbol *sym)
tools/objtool/klp-diff.c
1364
sym_for_each_reloc(e->patched, sym, reloc) {
tools/objtool/klp-diff.c
1368
if (convert_reloc_sym(e->patched, reloc))
tools/objtool/klp-diff.c
1390
sym_modname = find_modname(e);
tools/objtool/klp-diff.c
1436
static int clone_special_section(struct elfs *e, struct section *patched_sec)
tools/objtool/klp-diff.c
1450
if (!should_keep_special_sym(e->patched, patched_sym))
tools/objtool/klp-diff.c
1453
ret = validate_special_section_klp_reloc(e, patched_sym);
tools/objtool/klp-diff.c
1459
if (!clone_symbol(e, patched_sym, true))
tools/objtool/klp-diff.c
1467
static int clone_special_sections(struct elfs *e)
tools/objtool/klp-diff.c
1471
for_each_sec(e->patched, patched_sec) {
tools/objtool/klp-diff.c
1473
if (clone_special_section(e, patched_sec))
tools/objtool/klp-diff.c
1486
static int create_klp_sections(struct elfs *e)
tools/objtool/klp-diff.c
1498
obj_sec = elf_create_section_pair(e->out, KLP_OBJECTS_SEC, obj_size, 0, 0);
tools/objtool/klp-diff.c
1502
funcs_sec = elf_create_section_pair(e->out, KLP_FUNCS_SEC, func_size, 0, 0);
tools/objtool/klp-diff.c
1506
funcs_sym = elf_create_section_symbol(e->out, funcs_sec);
tools/objtool/klp-diff.c
1510
str_sec = elf_create_section(e->out, KLP_STRINGS_SEC, 0, 0,
tools/objtool/klp-diff.c
1516
if (elf_add_string(e->out, str_sec, "") == -1)
tools/objtool/klp-diff.c
1519
str_sym = elf_create_section_symbol(e->out, str_sec);
tools/objtool/klp-diff.c
1524
obj_data = elf_add_data(e->out, obj_sec, NULL, obj_size);
tools/objtool/klp-diff.c
1528
modname = find_modname(e);
tools/objtool/klp-diff.c
1534
addend = elf_add_string(e->out, str_sec, modname);
tools/objtool/klp-diff.c
1538
if (!elf_create_reloc(e->out, obj_sec,
tools/objtool/klp-diff.c
1545
if (!elf_create_reloc(e->out, obj_sec, offsetof(struct klp_object_ext, funcs),
tools/objtool/klp-diff.c
1549
for_each_sym(e->out, sym) {
tools/objtool/klp-diff.c
1558
func_data = elf_add_data(e->out, funcs_sec, NULL, func_size);
tools/objtool/klp-diff.c
1563
addend = elf_add_string(e->out, str_sec, sym->clone->twin->name);
tools/objtool/klp-diff.c
1567
if (!elf_create_reloc(e->out, funcs_sec,
tools/objtool/klp-diff.c
1573
if (!elf_create_reloc(e->out, funcs_sec,
tools/objtool/klp-diff.c
1580
sympos = find_sympos(e->orig, sym->clone->twin);
tools/objtool/klp-diff.c
1602
sym = find_symbol_by_name(e->out, sym_name);
tools/objtool/klp-diff.c
1606
reloc = find_reloc_by_dest(e->out, sym->sec, sym->offset);
tools/objtool/klp-diff.c
1608
if (!elf_create_reloc(e->out, obj_sec,
tools/objtool/klp-diff.c
1618
sym = find_symbol_by_name(e->out, sym_name);
tools/objtool/klp-diff.c
1622
reloc = find_reloc_by_dest(e->out, sym->sec, sym->offset);
tools/objtool/klp-diff.c
1624
if (!elf_create_reloc(e->out, obj_sec,
tools/objtool/klp-diff.c
1634
sym = find_symbol_by_name(e->out, sym_name);
tools/objtool/klp-diff.c
1638
reloc = find_reloc_by_dest(e->out, sym->sec, sym->offset);
tools/objtool/klp-diff.c
1640
if (!elf_create_reloc(e->out, obj_sec,
tools/objtool/klp-diff.c
1650
sym = find_symbol_by_name(e->out, sym_name);
tools/objtool/klp-diff.c
1654
reloc = find_reloc_by_dest(e->out, sym->sec, sym->offset);
tools/objtool/klp-diff.c
1656
if (!elf_create_reloc(e->out, obj_sec,
tools/objtool/klp-diff.c
1670
static int copy_import_ns(struct elfs *e)
tools/objtool/klp-diff.c
1675
patched_sec = find_section_by_name(e->patched, ".modinfo");
tools/objtool/klp-diff.c
1692
out_sec = find_section_by_name(e->out, ".modinfo");
tools/objtool/klp-diff.c
1694
out_sec = elf_create_section(e->out, ".modinfo", 0,
tools/objtool/klp-diff.c
1704
if (!elf_add_data(e->out, out_sec, import_ns, strlen(import_ns) + 1))
tools/objtool/klp-diff.c
1713
struct elfs e = {0};
tools/objtool/klp-diff.c
1721
e.orig = elf_open_read(argv[0], O_RDONLY);
tools/objtool/klp-diff.c
1722
e.patched = elf_open_read(argv[1], O_RDONLY);
tools/objtool/klp-diff.c
1723
e.out = NULL;
tools/objtool/klp-diff.c
1725
if (!e.orig || !e.patched)
tools/objtool/klp-diff.c
1731
if (read_sym_checksums(e.orig))
tools/objtool/klp-diff.c
1734
if (read_sym_checksums(e.patched))
tools/objtool/klp-diff.c
1737
if (correlate_symbols(&e))
tools/objtool/klp-diff.c
1740
if (mark_changed_functions(&e))
tools/objtool/klp-diff.c
1743
e.out = elf_create_file(&e.orig->ehdr, argv[2]);
tools/objtool/klp-diff.c
1744
if (!e.out)
tools/objtool/klp-diff.c
1755
if (create_fake_symbols(e.patched))
tools/objtool/klp-diff.c
1758
if (clone_included_functions(&e))
tools/objtool/klp-diff.c
1761
if (clone_special_sections(&e))
tools/objtool/klp-diff.c
1764
if (create_klp_sections(&e))
tools/objtool/klp-diff.c
1767
if (copy_import_ns(&e))
tools/objtool/klp-diff.c
1770
if (elf_write(e.out))
tools/objtool/klp-diff.c
1773
return elf_close(e.out);
tools/objtool/klp-diff.c
363
static int correlate_symbols(struct elfs *e)
tools/objtool/klp-diff.c
368
file1_sym = first_file_symbol(e->orig);
tools/objtool/klp-diff.c
369
file2_sym = first_file_symbol(e->patched);
tools/objtool/klp-diff.c
376
for_each_sym(e->orig, sym1) {
tools/objtool/klp-diff.c
383
for_each_sym(e->patched, sym2) {
tools/objtool/klp-diff.c
400
for (; ; file1_sym = next_file_symbol(e->orig, file1_sym),
tools/objtool/klp-diff.c
401
file2_sym = next_file_symbol(e->patched, file2_sym)) {
tools/objtool/klp-diff.c
426
for_each_sym_continue(e->orig, sym1) {
tools/objtool/klp-diff.c
434
for_each_sym_continue(e->patched, sym2) {
tools/objtool/klp-diff.c
452
for_each_sym(e->orig, sym1) {
tools/objtool/klp-diff.c
456
sym2 = find_global_symbol_by_name(e->patched, sym1->name);
tools/objtool/klp-diff.c
464
for_each_sym(e->orig, sym1) {
tools/objtool/klp-diff.c
532
static int clone_sym_relocs(struct elfs *e, struct symbol *patched_sym);
tools/objtool/klp-diff.c
623
static struct symbol *clone_symbol(struct elfs *e, struct symbol *patched_sym,
tools/objtool/klp-diff.c
637
clone_symbol(e, pfx, true);
tools/objtool/klp-diff.c
640
if (!__clone_symbol(e->out, patched_sym, data_too))
tools/objtool/klp-diff.c
643
if (data_too && clone_sym_relocs(e, patched_sym))
tools/objtool/klp-diff.c
671
static int mark_changed_functions(struct elfs *e)
tools/objtool/klp-diff.c
677
for_each_sym(e->orig, sym_orig) {
tools/objtool/klp-diff.c
693
for_each_sym(e->patched, patched_sym) {
tools/objtool/klp-diff.c
705
for_each_sym(e->patched, patched_sym) {
tools/objtool/klp-diff.c
713
static int clone_included_functions(struct elfs *e)
tools/objtool/klp-diff.c
717
for_each_sym(e->patched, patched_sym) {
tools/objtool/klp-diff.c
719
if (!clone_symbol(e, patched_sym, true))
tools/objtool/klp-diff.c
772
static const char *__find_modname(struct elfs *e)
tools/objtool/klp-diff.c
777
sec = find_section_by_name(e->orig, ".modinfo");
tools/objtool/klp-diff.c
787
name = strdup(e->orig->name);
tools/objtool/klp-diff.c
808
static const char *find_modname(struct elfs *e)
tools/objtool/klp-diff.c
812
if (e->modname)
tools/objtool/klp-diff.c
813
return e->modname;
tools/objtool/klp-diff.c
815
modname = __find_modname(e);
tools/objtool/klp-diff.c
816
e->modname = modname;
tools/objtool/klp-diff.c
927
static int clone_reloc_klp(struct elfs *e, struct reloc *patched_reloc,
tools/objtool/klp-diff.c
957
sym = elf_create_symbol(e->out, patched_sym->name, NULL,
tools/objtool/klp-diff.c
966
if (!elf_create_reloc(e->out, sec, offset, sym, addend, reloc_type(patched_reloc)))
tools/objtool/klp-diff.c
978
sym_modname = find_modname(e);
tools/objtool/klp-diff.c
983
sympos = find_sympos(e->orig, patched_sym->twin);
tools/objtool/klp-diff.c
993
klp_sym = find_symbol_by_name(e->out, sym_name);
tools/objtool/klp-diff.c
998
klp_sym = elf_create_symbol(e->out, sym_name, NULL,
tools/perf/arch/x86/tests/hybrid.c
243
static int test_event(const struct evlist_test *e)
tools/perf/arch/x86/tests/hybrid.c
249
if (e->valid && !e->valid()) {
tools/perf/arch/x86/tests/hybrid.c
260
ret = parse_events(evlist, e->name, &err);
tools/perf/arch/x86/tests/hybrid.c
262
pr_debug("failed to parse event '%s', err %d\n", e->name, ret);
tools/perf/arch/x86/tests/hybrid.c
263
parse_events_error__print(&err, e->name);
tools/perf/arch/x86/tests/hybrid.c
268
ret = e->check(evlist);
tools/perf/arch/x86/tests/hybrid.c
290
const struct evlist_test *e = &events[i];
tools/perf/arch/x86/tests/hybrid.c
293
pr_debug("running test %d '%s'\n", i, e->name);
tools/perf/arch/x86/tests/hybrid.c
294
test_ret = test_event(e);
tools/perf/arch/x86/tests/hybrid.c
296
pr_debug("Event test failure: test %d '%s'", i, e->name);
tools/perf/bench/mem-functions.c
155
static union bench_clock clock_diff(union bench_clock *s, union bench_clock *e)
tools/perf/bench/mem-functions.c
160
t.cycles = e->cycles - s->cycles;
tools/perf/bench/mem-functions.c
162
timersub(&e->tv, &s->tv, &t.tv);
tools/perf/builtin-c2c.c
3397
struct perf_mem_event *e;
tools/perf/builtin-c2c.c
3434
e = perf_pmu__mem_events_ptr(pmu, PERF_MEM_EVENTS__LOAD_STORE);
tools/perf/builtin-c2c.c
3439
if (e->tag) {
tools/perf/builtin-diff.c
473
struct evsel *e;
tools/perf/builtin-diff.c
475
evlist__for_each_entry(evlist, e) {
tools/perf/builtin-diff.c
476
if ((evsel->core.attr.type == e->core.attr.type) &&
tools/perf/builtin-diff.c
477
(evsel->core.attr.config == e->core.attr.config))
tools/perf/builtin-diff.c
478
return e;
tools/perf/builtin-mem.c
114
e = perf_pmu__mem_events_ptr(pmu, PERF_MEM_EVENTS__LOAD_STORE);
tools/perf/builtin-mem.c
120
if (e->tag &&
tools/perf/builtin-mem.c
80
struct perf_mem_event *e;
tools/perf/builtin-stat.c
269
#define SID(e, x, y) xyarray__entry(e->core.sample_id, x, y)
tools/perf/builtin-trace.c
3010
const char *e = perf_env__arch_strerrno(env, err);
tools/perf/builtin-trace.c
3012
fprintf(trace->output, "-1 %s (%s)", e, emsg);
tools/perf/builtin-trace.c
4886
int e;
tools/perf/builtin-trace.c
4888
for (e = 0; e < stats->max_errno; ++e) {
tools/perf/builtin-trace.c
4889
if (stats->errnos[e] != 0)
tools/perf/builtin-trace.c
4890
fprintf(fp, "\t\t\t\t%s: %d\n", perf_env__arch_strerrno(trace->host->env, e + 1), stats->errnos[e]);
tools/perf/dlfilters/dlfilter-show-cycles.c
121
struct entry *e = find_entry(tid);
tools/perf/dlfilters/dlfilter-show-cycles.c
123
if (e) {
tools/perf/dlfilters/dlfilter-show-cycles.c
124
print_vals(e->cycles[pos], e->cycles[pos] - e->cycles_rpt[pos]);
tools/perf/dlfilters/dlfilter-show-cycles.c
125
e->cycles_rpt[pos] = e->cycles[pos];
tools/perf/dlfilters/dlfilter-show-cycles.c
50
struct entry *e;
tools/perf/dlfilters/dlfilter-show-cycles.c
52
e = &table[pos];
tools/perf/dlfilters/dlfilter-show-cycles.c
53
while (e->used) {
tools/perf/dlfilters/dlfilter-show-cycles.c
54
if (e->tid == tid)
tools/perf/dlfilters/dlfilter-show-cycles.c
55
return e;
tools/perf/dlfilters/dlfilter-show-cycles.c
58
e = &table[pos];
tools/perf/dlfilters/dlfilter-show-cycles.c
67
e->used = 1;
tools/perf/dlfilters/dlfilter-show-cycles.c
68
e->tid = tid;
tools/perf/dlfilters/dlfilter-show-cycles.c
69
return e;
tools/perf/dlfilters/dlfilter-show-cycles.c
74
struct entry *e = find_entry(tid);
tools/perf/dlfilters/dlfilter-show-cycles.c
76
if (e)
tools/perf/dlfilters/dlfilter-show-cycles.c
77
e->cycles[pos] += cnt;
tools/perf/tests/expr.c
61
static int test(struct expr_parse_ctx *ctx, const char *e, double val2)
tools/perf/tests/expr.c
65
if (expr__parse(&val, ctx, e))
tools/perf/tests/kmod-path.c
47
#define M(path, c, e) \
tools/perf/tests/kmod-path.c
48
TEST_ASSERT_VAL("failed", !test_is_kernel_module(path, c, e))
tools/perf/tests/parse-events.c
2535
static int test_event(const struct evlist_test *e)
tools/perf/tests/parse-events.c
2541
if (e->valid && !e->valid()) {
tools/perf/tests/parse-events.c
2552
ret = __parse_events(evlist, e->name, /*pmu_filter=*/NULL, &err, /*fake_pmu=*/false,
tools/perf/tests/parse-events.c
2555
pr_debug("failed to parse event '%s', err %d\n", e->name, ret);
tools/perf/tests/parse-events.c
2556
parse_events_error__print(&err, e->name);
tools/perf/tests/parse-events.c
2561
ret = e->check(evlist);
tools/perf/tests/parse-events.c
2610
struct evlist_test e = events[i];
tools/perf/tests/parse-events.c
2612
const char *pos = e.name, *end;
tools/perf/tests/parse-events.c
2626
e.name = buf;
tools/perf/tests/parse-events.c
2627
pr_debug("running test %d '%s'\n", i, e.name);
tools/perf/tests/parse-events.c
2628
test_ret = test_event(&e);
tools/perf/tests/parse-events.c
2630
pr_debug("Event test failure: test %d '%s'\n", i, e.name);
tools/perf/tests/parse-events.c
2718
struct evlist_test e = { .name = NULL, };
tools/perf/tests/parse-events.c
2762
e.name = name;
tools/perf/tests/parse-events.c
2763
e.check = test__checkevent_pmu_events;
tools/perf/tests/parse-events.c
2765
test_ret = test_event(&e);
tools/perf/tests/parse-events.c
2789
e.name = name;
tools/perf/tests/parse-events.c
2790
e.check = test__checkevent_pmu_events_mix;
tools/perf/tests/parse-events.c
2791
test_ret = test_event(&e);
tools/perf/tests/parse-events.c
2879
struct evlist_test e = { .name = NULL, };
tools/perf/tests/parse-events.c
2885
e.name = name;
tools/perf/tests/parse-events.c
2886
e.check = test__checkevent_pmu_events_alias;
tools/perf/tests/parse-events.c
2887
return test_event(&e);
tools/perf/tests/time-utils-test.c
174
u64 e = 8000000000000000ULL;
tools/perf/tests/time-utils-test.c
179
.ptime = { {b, b + 1}, {c, c + 123}, {e, e + 5}, },
tools/perf/tests/time-utils-test.c
181
.skip = { b - 1, b + 2, c - 1, c + 124, e - 1, e + 6 },
tools/perf/tests/time-utils-test.c
182
.noskip = { b, b + 1, c, c + 123, e, e + 5 },
tools/perf/util/auxtrace.c
1775
struct perf_record_auxtrace_error *e = &event->auxtrace_error;
tools/perf/util/auxtrace.c
1776
unsigned long long nsecs = e->time;
tools/perf/util/auxtrace.c
1777
const char *msg = e->msg;
tools/perf/util/auxtrace.c
1781
auxtrace_error_name(e->type), e->type);
tools/perf/util/auxtrace.c
1783
if (e->fmt && nsecs) {
tools/perf/util/auxtrace.c
1792
if (!e->fmt)
tools/perf/util/auxtrace.c
1793
msg = (const char *)&e->time;
tools/perf/util/auxtrace.c
1795
if (e->fmt >= 2 && e->machine_pid)
tools/perf/util/auxtrace.c
1796
ret += fprintf(fp, " machine_pid %d vcpu %d", e->machine_pid, e->vcpu);
tools/perf/util/auxtrace.c
1799
e->cpu, e->pid, e->tid, e->ip, e->code, msg);
tools/perf/util/auxtrace.c
1806
struct perf_record_auxtrace_error *e = &event->auxtrace_error;
tools/perf/util/auxtrace.c
1808
if (e->type < PERF_AUXTRACE_ERROR_MAX)
tools/perf/util/auxtrace.c
1809
session->evlist->stats.nr_auxtrace_errors[e->type] += 1;
tools/perf/util/bpf-filter.c
68
#define FD(e, x, y) (*(int *)xyarray__entry(e->core.fd, x, y))
tools/perf/util/branch.c
101
const char *get_branch_type(struct branch_entry *e)
tools/perf/util/branch.c
103
if (e->flags.type == PERF_BR_UNKNOWN)
tools/perf/util/branch.c
106
if (e->flags.type == PERF_BR_EXTEND_ABI)
tools/perf/util/branch.c
107
return branch_new_type_name(e->flags.new_type);
tools/perf/util/branch.c
109
return branch_type_name(e->flags.type);
tools/perf/util/branch.h
90
const char *get_branch_type(struct branch_entry *e);
tools/perf/util/cgroup.c
263
const char *p, *e, *eos = str + strlen(str);
tools/perf/util/cgroup.c
270
e = p ? p : eos;
tools/perf/util/cgroup.c
272
if (e - str) {
tools/perf/util/cgroup.c
275
s = strndup(str, e - str);
tools/perf/util/cgroup.c
304
const char *p, *e, *eos = str + strlen(str);
tools/perf/util/cgroup.c
322
e = p ? p : eos;
tools/perf/util/cgroup.c
325
if (e - str) {
tools/perf/util/cgroup.c
327
s = strndup(str, e - str);
tools/perf/util/cgroup.c
366
const char *p, *e, *eos = str + strlen(str);
tools/perf/util/cgroup.c
377
e = p ? p : eos;
tools/perf/util/cgroup.c
380
if (e - str) {
tools/perf/util/cgroup.c
382
s = strndup(str, e - str);
tools/perf/util/evlist.c
73
#define FD(e, x, y) (*(int *)xyarray__entry(e->core.fd, x, y))
tools/perf/util/evlist.c
74
#define SID(e, x, y) xyarray__entry(e->core.sample_id, x, y)
tools/perf/util/evsel.c
254
#define FD(e, x, y) (*(int *)xyarray__entry(e->core.fd, x, y))
tools/perf/util/evsel.c
3424
struct branch_entry *e;
tools/perf/util/evsel.c
3436
e = &data->branch_stack->entries[0];
tools/perf/util/evsel.c
3443
e = (struct branch_entry *)&data->branch_stack->hw_idx;
tools/perf/util/evsel.c
3457
for (i = 0; i < data->branch_stack->nr; i++, e++)
tools/perf/util/evsel.c
3458
e->flags.value = evsel__bitfield_swap_branch_flags(e->flags.value);
tools/perf/util/genelf.c
121
scn = elf_newscn(e);
tools/perf/util/genelf.c
187
Elf *e;
tools/perf/util/genelf.c
204
e = elf_begin(fd, ELF_C_WRITE, NULL);
tools/perf/util/genelf.c
205
if (!e) {
tools/perf/util/genelf.c
213
ehdr = elf_newehdr(e);
tools/perf/util/genelf.c
230
phdr = elf_newphdr(e, 1);
tools/perf/util/genelf.c
243
scn = elf_newscn(e);
tools/perf/util/genelf.c
282
retval = jit_add_eh_frame_info(e, unwinding,
tools/perf/util/genelf.c
293
scn = elf_newscn(e);
tools/perf/util/genelf.c
329
scn = elf_newscn(e);
tools/perf/util/genelf.c
374
scn = elf_newscn(e);
tools/perf/util/genelf.c
409
scn = elf_newscn(e);
tools/perf/util/genelf.c
452
retval = jit_add_debug_info(e, load_addr, debug, nr_debug_entries);
tools/perf/util/genelf.c
458
if (elf_update(e, ELF_C_WRITE) < 0) {
tools/perf/util/genelf.c
466
(void)elf_end(e);
tools/perf/util/genelf.c
76
jit_add_eh_frame_info(Elf *e, void* unwinding, uint64_t unwinding_header_size,
tools/perf/util/genelf.c
87
scn = elf_newscn(e);
tools/perf/util/genelf.h
13
int jit_add_debug_info(Elf *e, uint64_t code_addr, void *debug, int nr_debug_entries);
tools/perf/util/genelf_debug.c
495
jit_add_debug_info(Elf *e, uint64_t code_addr, void *debug, int nr_debug_entries)
tools/perf/util/genelf_debug.c
513
scn = elf_newscn(e);
tools/perf/util/genelf_debug.c
547
scn = elf_newscn(e);
tools/perf/util/genelf_debug.c
581
scn = elf_newscn(e);
tools/perf/util/genelf_debug.c
615
if (elf_update(e, ELF_C_WRITE) < 0)
tools/perf/util/hwmon_pmu.c
756
#define FD(e, x, y) (*(int *)xyarray__entry(e->core.fd, x, y))
tools/perf/util/intel-pt.c
3653
struct intel_pt_cache_entry *e;
tools/perf/util/intel-pt.c
3676
e = intel_pt_cache_lookup(dso, machine, offset);
tools/perf/util/intel-pt.c
3677
if (!e)
tools/perf/util/intel-pt.c
3680
if (addr + e->byte_cnt + e->length <= event->text_poke.addr) {
tools/perf/util/intel-pt.c
3686
if (e->branch != INTEL_PT_BR_NO_BRANCH)
tools/perf/util/intel-pt.c
621
struct intel_pt_cache_entry *e;
tools/perf/util/intel-pt.c
627
e = auxtrace_cache__alloc_entry(c);
tools/perf/util/intel-pt.c
628
if (!e)
tools/perf/util/intel-pt.c
631
e->insn_cnt = insn_cnt;
tools/perf/util/intel-pt.c
632
e->byte_cnt = byte_cnt;
tools/perf/util/intel-pt.c
633
e->op = intel_pt_insn->op;
tools/perf/util/intel-pt.c
634
e->branch = intel_pt_insn->branch;
tools/perf/util/intel-pt.c
635
e->emulated_ptwrite = intel_pt_insn->emulated_ptwrite;
tools/perf/util/intel-pt.c
636
e->length = intel_pt_insn->length;
tools/perf/util/intel-pt.c
637
e->rel = intel_pt_insn->rel;
tools/perf/util/intel-pt.c
638
memcpy(e->insn, intel_pt_insn->buf, INTEL_PT_INSN_BUF_SZ);
tools/perf/util/intel-pt.c
640
err = auxtrace_cache__add(c, offset, &e->entry);
tools/perf/util/intel-pt.c
642
auxtrace_cache__free_entry(c, e);
tools/perf/util/intel-pt.c
835
struct intel_pt_cache_entry *e;
tools/perf/util/intel-pt.c
837
e = intel_pt_cache_lookup(dso, machine, offset);
tools/perf/util/intel-pt.c
838
if (e &&
tools/perf/util/intel-pt.c
839
(!max_insn_cnt || e->insn_cnt <= max_insn_cnt)) {
tools/perf/util/intel-pt.c
840
*insn_cnt_ptr = e->insn_cnt;
tools/perf/util/intel-pt.c
841
*ip += e->byte_cnt;
tools/perf/util/intel-pt.c
842
intel_pt_insn->op = e->op;
tools/perf/util/intel-pt.c
843
intel_pt_insn->branch = e->branch;
tools/perf/util/intel-pt.c
844
intel_pt_insn->emulated_ptwrite = e->emulated_ptwrite;
tools/perf/util/intel-pt.c
845
intel_pt_insn->length = e->length;
tools/perf/util/intel-pt.c
846
intel_pt_insn->rel = e->rel;
tools/perf/util/intel-pt.c
847
memcpy(intel_pt_insn->buf, e->insn, INTEL_PT_INSN_BUF_SZ);
tools/perf/util/intel-pt.c
926
struct intel_pt_cache_entry *e;
tools/perf/util/intel-pt.c
928
e = intel_pt_cache_lookup(map__dso(al.map), machine, start_offset);
tools/perf/util/intel-pt.c
929
if (e)
tools/perf/util/mem-events.c
103
e->name, pmu->name, pmu->name,
tools/perf/util/mem-events.c
107
if (!e->aux_event) {
tools/perf/util/mem-events.c
110
e->name, pmu->name);
tools/perf/util/mem-events.c
120
e->name, pmu->name);
tools/perf/util/mem-events.c
130
struct perf_mem_event *e;
tools/perf/util/mem-events.c
135
e = &pmu->mem_events[PERF_MEM_EVENTS__LOAD];
tools/perf/util/mem-events.c
136
if (!e->aux_event)
tools/perf/util/mem-events.c
139
return leader->core.attr.config == e->aux_event;
tools/perf/util/mem-events.c
160
struct perf_mem_event *e = perf_pmu__mem_events_ptr(pmu, j);
tools/perf/util/mem-events.c
162
if (!e->tag)
tools/perf/util/mem-events.c
165
if (strstr(e->tag, tok))
tools/perf/util/mem-events.c
182
struct perf_mem_event *e)
tools/perf/util/mem-events.c
187
if (!e->event_name)
tools/perf/util/mem-events.c
190
scnprintf(path, PATH_MAX, "%s/bus/event_source/devices/%s/events/%s", mnt, pmu->name, e->event_name);
tools/perf/util/mem-events.c
205
struct perf_mem_event *e = perf_pmu__mem_events_ptr(pmu, j);
tools/perf/util/mem-events.c
211
if (!e->tag)
tools/perf/util/mem-events.c
214
e->supported |= perf_pmu__mem_events_supported(mnt, pmu, e);
tools/perf/util/mem-events.c
215
if (e->supported)
tools/perf/util/mem-events.c
240
struct perf_mem_event *e = perf_pmu__mem_events_ptr(pmu, j);
tools/perf/util/mem-events.c
243
e->tag ? 13 : 0,
tools/perf/util/mem-events.c
244
e->tag ? : "",
tools/perf/util/mem-events.c
245
e->tag && verbose > 0 ? 25 : 0,
tools/perf/util/mem-events.c
246
e->tag && verbose > 0
tools/perf/util/mem-events.c
249
e->supported ? ": available\n" : "");
tools/perf/util/mem-events.c
272
struct perf_mem_event *e = perf_pmu__mem_events_ptr(pmu, j);
tools/perf/util/mem-events.c
278
if (!e->supported) {
tools/perf/util/mem-events.c
289
if (!s || !perf_pmu__mem_events_supported(mnt, pmu, e))
tools/perf/util/mem-events.c
84
struct perf_mem_event *e;
tools/perf/util/mem-events.c
89
e = &pmu->mem_events[i];
tools/perf/util/mem-events.c
90
if (!e || !e->name)
tools/perf/util/mem-events.c
94
if (e->ldlat) {
tools/perf/util/mem-events.c
95
if (!e->aux_event) {
tools/perf/util/mem-events.c
98
e->name, pmu->name,
tools/perf/util/mem2node.c
22
struct phys_entry *e;
tools/perf/util/mem2node.c
26
e = rb_entry(parent, struct phys_entry, rb_node);
tools/perf/util/mem2node.c
28
if (entry->start < e->start)
tools/perf/util/parse-events.c
440
static void tracepoint_error(struct parse_events_error *e, int err,
tools/perf/util/parse-events.c
446
if (!e)
tools/perf/util/parse-events.c
468
parse_events_error__handle(e, column, strdup(str), strdup(help));
tools/perf/util/python.c
1428
struct metric_expr *e = container_of(pos, struct metric_expr, nd);
tools/perf/util/python.c
1431
if (strcmp(e->metric_name, metric))
tools/perf/util/python.c
1434
if (e->metric_events[0] == NULL)
tools/perf/util/python.c
1438
if (pos2->metric_leader != e->metric_events[0])
tools/perf/util/python.c
1449
mexp = e;
tools/perf/util/python.c
2050
struct metric_expr *e = container_of(mpos, struct metric_expr, nd);
tools/perf/util/python.c
2052
for (int j = 0; e->metric_events[j]; j++) {
tools/perf/util/python.c
2053
idx = evlist__pos(evlist, e->metric_events[j]);
tools/perf/util/python.c
2055
e->metric_events[j] = evlist__at(&pevlist->evlist, idx);
tools/perf/util/scripting-engines/trace-event-python.c
1612
struct perf_record_auxtrace_error *e = &event->auxtrace_error;
tools/perf/util/scripting-engines/trace-event-python.c
1613
u8 cpumode = e->header.misc & PERF_RECORD_MISC_CPUMODE_MASK;
tools/perf/util/scripting-engines/trace-event-python.c
1615
unsigned long long tm = e->time;
tools/perf/util/scripting-engines/trace-event-python.c
1616
const char *msg = e->msg;
tools/perf/util/scripting-engines/trace-event-python.c
1623
if (!e->fmt) {
tools/perf/util/scripting-engines/trace-event-python.c
1625
msg = (const char *)&e->time;
tools/perf/util/scripting-engines/trace-event-python.c
1630
tuple_set_u32(t, 0, e->type);
tools/perf/util/scripting-engines/trace-event-python.c
1631
tuple_set_u32(t, 1, e->code);
tools/perf/util/scripting-engines/trace-event-python.c
1632
tuple_set_s32(t, 2, e->cpu);
tools/perf/util/scripting-engines/trace-event-python.c
1633
tuple_set_s32(t, 3, e->pid);
tools/perf/util/scripting-engines/trace-event-python.c
1634
tuple_set_s32(t, 4, e->tid);
tools/perf/util/scripting-engines/trace-event-python.c
1635
tuple_set_u64(t, 5, e->ip);
tools/perf/util/scripting-engines/trace-event-python.c
1639
tuple_set_s32(t, 9, e->machine_pid);
tools/perf/util/scripting-engines/trace-event-python.c
1640
tuple_set_s32(t, 10, e->vcpu);
tools/perf/util/session.c
2892
struct id_index_entry *e = &ie->entries[i];
tools/perf/util/session.c
2897
fprintf(stdout, " ... id: %"PRI_lu64, e->id);
tools/perf/util/session.c
2898
fprintf(stdout, " idx: %"PRI_lu64, e->idx);
tools/perf/util/session.c
2899
fprintf(stdout, " cpu: %"PRI_ld64, e->cpu);
tools/perf/util/session.c
2900
fprintf(stdout, " tid: %"PRI_ld64, e->tid);
tools/perf/util/session.c
2909
sid = evlist__id2sid(evlist, e->id);
tools/perf/util/session.c
2913
sid->idx = e->idx;
tools/perf/util/session.c
2914
sid->cpu.cpu = e->cpu;
tools/perf/util/session.c
2915
sid->tid = e->tid;
tools/perf/util/session.c
2934
ret = perf_session__set_guest_cpu(session, sid->machine_pid, e->tid, e2->vcpu);
tools/perf/util/session.c
927
struct branch_entry *e = &entries[i];
tools/perf/util/session.c
931
i, e->from, e->to,
tools/perf/util/session.c
932
(unsigned short)e->flags.cycles,
tools/perf/util/session.c
933
e->flags.mispred ? "M" : " ",
tools/perf/util/session.c
934
e->flags.predicted ? "P" : " ",
tools/perf/util/session.c
935
e->flags.abort ? "A" : " ",
tools/perf/util/session.c
936
e->flags.in_tx ? "T" : " ",
tools/perf/util/session.c
937
(unsigned)e->flags.reserved,
tools/perf/util/session.c
938
get_branch_type(e),
tools/perf/util/session.c
939
e->flags.spec ? branch_spec_desc(e->flags.spec) : "");
tools/perf/util/session.c
944
i, e->to, i+1, e->from);
tools/perf/util/session.c
946
printf("..... %2"PRIu64": %016" PRIx64 "\n", i+1, e->from);
tools/perf/util/sort.c
817
static char *hist_entry__get_srcfile(struct hist_entry *e)
tools/perf/util/sort.c
820
struct map *map = e->ms.map;
tools/perf/util/sort.c
825
sf = __get_srcline(map__dso(map), map__rip_2objdump(map, e->ip),
tools/perf/util/sort.c
826
e->ms.sym, false, true, true, e->ip);
tools/perf/util/strfilter.c
130
e = s + 1;
tools/perf/util/strfilter.c
138
cur->r->p = strndup(s, e - s);
tools/perf/util/strfilter.c
142
s = get_token(e, &e);
tools/perf/util/strfilter.c
38
static const char *get_token(const char *s, const char **e)
tools/perf/util/strfilter.c
62
*e = p;
tools/perf/util/strfilter.c
85
const char *e;
tools/perf/util/strfilter.c
93
s = get_token(s, &e);
tools/perf/util/string.c
225
char *e = expr;
tools/perf/util/string.c
237
printed += scnprintf(e + printed, size - printed, " %s ", or_and);
tools/perf/util/string.c
238
printed += scnprintf(e + printed, size - printed,
tools/perf/util/symbol-elf.c
2242
u64 pgoff, u64 s, u64 e)
tools/perf/util/symbol-elf.c
2250
len = e < end ? e - s : end - s;
tools/perf/util/synthetic-events.c
1896
struct id_index_entry *e;
tools/perf/util/synthetic-events.c
1910
e = &ev->id_index.entries[i];
tools/perf/util/synthetic-events.c
1912
e->id = evsel->core.id[j];
tools/perf/util/synthetic-events.c
1914
sid = evlist__id2sid(evlist, e->id);
tools/perf/util/synthetic-events.c
1920
e->idx = sid->idx;
tools/perf/util/synthetic-events.c
1921
e->cpu = sid->cpu.cpu;
tools/perf/util/synthetic-events.c
1922
e->tid = sid->tid;
tools/perf/util/tool_pmu.c
112
static bool read_until_char(struct io *io, char e)
tools/perf/util/tool_pmu.c
120
} while (c != e);
tools/perf/util/tool_pmu.c
220
#define FD(e, x, y) (*(int *)xyarray__entry(e->core.fd, x, y))
tools/perf/util/unwind-libdw.c
154
struct unwind_entry *e = &ui->entries[ui->idx++];
tools/perf/util/unwind-libdw.c
163
e->ip = ip;
tools/perf/util/unwind-libdw.c
164
e->ms.thread = thread__get(al.thread);
tools/perf/util/unwind-libdw.c
165
e->ms.map = map__get(al.map);
tools/perf/util/unwind-libdw.c
166
e->ms.sym = al.sym;
tools/perf/util/unwind-libunwind-local.c
663
struct unwind_entry e;
tools/perf/util/unwind-libunwind-local.c
668
e.ms.sym = thread__find_symbol(thread, PERF_RECORD_MISC_USER, ip, &al);
tools/perf/util/unwind-libunwind-local.c
669
e.ip = ip;
tools/perf/util/unwind-libunwind-local.c
670
e.ms.map = al.map;
tools/perf/util/unwind-libunwind-local.c
671
e.ms.thread = thread__get(al.thread);
tools/perf/util/unwind-libunwind-local.c
678
ret = cb(&e, arg);
tools/power/x86/turbostat/turbostat.c
1991
int pmt_telemdir_filter(const struct dirent *e)
tools/power/x86/turbostat/turbostat.c
1995
return sscanf(e->d_name, "telem%u", &dummy);
tools/sched_ext/include/scx/common.bpf.h
111
#define scx_read_event(e, name) \
tools/sched_ext/include/scx/common.bpf.h
112
(bpf_core_field_exists((e)->name) ? (e)->name : 0)
tools/sched_ext/include/scx/compat.h
42
struct btf_enum *e = btf_enum(t);
tools/sched_ext/include/scx/compat.h
45
n = btf__name_by_offset(__COMPAT_vmlinux_btf, e[i].name_off);
tools/sched_ext/include/scx/compat.h
48
*v = e[i].val;
tools/sched_ext/include/scx/compat.h
53
struct btf_enum64 *e = btf_enum64(t);
tools/sched_ext/include/scx/compat.h
56
n = btf__name_by_offset(__COMPAT_vmlinux_btf, e[i].name_off);
tools/sched_ext/include/scx/compat.h
59
*v = btf_enum64_value(&e[i]);
tools/testing/radix-tree/maple.c
34984
unsigned long e, count = 0, max = 1000;
tools/testing/radix-tree/maple.c
34989
e = 37;
tools/testing/radix-tree/maple.c
34991
e = 74;
tools/testing/radix-tree/maple.c
34998
MT_BUG_ON(mt, count != e);
tools/testing/radix-tree/maple.c
35005
e = 77;
tools/testing/radix-tree/maple.c
35013
MT_BUG_ON(mt, count != e);
tools/testing/radix-tree/maple.c
35025
MT_BUG_ON(mt, count != e);
tools/testing/selftests/alsa/conf.c
145
char *e;
tools/testing/selftests/alsa/conf.c
161
e = strrchr(link, '/');
tools/testing/selftests/alsa/conf.c
162
if (e)
tools/testing/selftests/alsa/conf.c
163
return strdup(e + 1);
tools/testing/selftests/alsa/conf.c
186
e = strdup(path);
tools/testing/selftests/alsa/conf.c
187
if (e == NULL)
tools/testing/selftests/alsa/conf.c
189
return e;
tools/testing/selftests/arm64/fp/vlset.c
107
int t, e;
tools/testing/selftests/arm64/fp/vlset.c
154
e = errno;
tools/testing/selftests/arm64/fp/vlset.c
157
fprintf(stderr, "%s: %s: %s\n", program_name, path, strerror(e));
tools/testing/selftests/bpf/benchs/bench_ringbufs.c
432
struct perf_event_header *e)
tools/testing/selftests/bpf/benchs/bench_ringbufs.c
434
switch (e->type) {
tools/testing/selftests/bpf/prog_tests/core_reloc.c
162
.e = { [2] = 5 }, \
tools/testing/selftests/bpf/prog_tests/core_reloc.c
170
.e = 5, .f = 6, .g = 7, .h = 8, \
tools/testing/selftests/bpf/prog_tests/get_stack_raw_tp.c
30
struct get_stack_trace_t e;
tools/testing/selftests/bpf/prog_tests/get_stack_raw_tp.c
34
memset(&e, 0, sizeof(e));
tools/testing/selftests/bpf/prog_tests/get_stack_raw_tp.c
35
memcpy(&e, data, size <= sizeof(e) ? size : sizeof(e));
tools/testing/selftests/bpf/prog_tests/get_stack_raw_tp.c
63
num_stack = e.kern_stack_size / sizeof(__u64);
tools/testing/selftests/bpf/prog_tests/get_stack_raw_tp.c
68
ks = ksym_search(e.kern_stack[i]);
tools/testing/selftests/bpf/prog_tests/get_stack_raw_tp.c
75
if (e.user_stack_size > 0 && e.user_stack_buildid_size > 0)
tools/testing/selftests/bpf/prog_tests/libbpf_probes.c
11
const struct btf_enum *e;
tools/testing/selftests/bpf/prog_tests/libbpf_probes.c
26
for (e = btf_enum(t), i = 0, n = btf_vlen(t); i < n; e++, i++) {
tools/testing/selftests/bpf/prog_tests/libbpf_probes.c
27
const char *prog_type_name = btf__str_by_offset(btf, e->name_off);
tools/testing/selftests/bpf/prog_tests/libbpf_probes.c
28
enum bpf_prog_type prog_type = (enum bpf_prog_type)e->val;
tools/testing/selftests/bpf/prog_tests/libbpf_probes.c
51
const struct btf_enum *e;
tools/testing/selftests/bpf/prog_tests/libbpf_probes.c
66
for (e = btf_enum(t), i = 0, n = btf_vlen(t); i < n; e++, i++) {
tools/testing/selftests/bpf/prog_tests/libbpf_probes.c
67
const char *map_type_name = btf__str_by_offset(btf, e->name_off);
tools/testing/selftests/bpf/prog_tests/libbpf_probes.c
68
enum bpf_map_type map_type = (enum bpf_map_type)e->val;
tools/testing/selftests/bpf/prog_tests/libbpf_str.c
115
const struct btf_enum *e;
tools/testing/selftests/bpf/prog_tests/libbpf_str.c
127
e = btf_enum(t);
tools/testing/selftests/bpf/prog_tests/libbpf_str.c
129
for (i = 0; i < n; e++, i++) {
tools/testing/selftests/bpf/prog_tests/libbpf_str.c
130
enum bpf_map_type map_type = (enum bpf_map_type)e->val;
tools/testing/selftests/bpf/prog_tests/libbpf_str.c
138
map_type_name = btf__str_by_offset(btf, e->name_off);
tools/testing/selftests/bpf/prog_tests/libbpf_str.c
172
const struct btf_enum *e;
tools/testing/selftests/bpf/prog_tests/libbpf_str.c
184
e = btf_enum(t);
tools/testing/selftests/bpf/prog_tests/libbpf_str.c
186
for (i = 0; i < n; e++, i++) {
tools/testing/selftests/bpf/prog_tests/libbpf_str.c
187
enum bpf_prog_type prog_type = (enum bpf_prog_type)e->val;
tools/testing/selftests/bpf/prog_tests/libbpf_str.c
195
prog_type_name = btf__str_by_offset(btf, e->name_off);
tools/testing/selftests/bpf/prog_tests/libbpf_str.c
25
const struct btf_enum *e;
tools/testing/selftests/bpf/prog_tests/libbpf_str.c
37
e = btf_enum(t);
tools/testing/selftests/bpf/prog_tests/libbpf_str.c
39
for (i = 0; i < n; e++, i++) {
tools/testing/selftests/bpf/prog_tests/libbpf_str.c
40
enum bpf_attach_type attach_type = (enum bpf_attach_type)e->val;
tools/testing/selftests/bpf/prog_tests/libbpf_str.c
48
attach_type_name = btf__str_by_offset(btf, e->name_off);
tools/testing/selftests/bpf/prog_tests/libbpf_str.c
70
const struct btf_enum *e;
tools/testing/selftests/bpf/prog_tests/libbpf_str.c
82
e = btf_enum(t);
tools/testing/selftests/bpf/prog_tests/libbpf_str.c
84
for (i = 0; i < n; e++, i++) {
tools/testing/selftests/bpf/prog_tests/libbpf_str.c
85
enum bpf_link_type link_type = (enum bpf_link_type)e->val;
tools/testing/selftests/bpf/prog_tests/libbpf_str.c
93
link_type_name = btf__str_by_offset(btf, e->name_off);
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
1255
static bool assert_reg_state_eq(struct reg_state *r, struct reg_state *e, const char *ctx)
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
1260
if (r->valid != e->valid) {
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
1263
e->valid ? "<valid>" : "<invalid>");
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
1271
if (!assert_range_eq(t, r->r[t], e->r[t], ctx, t_str(t)))
tools/testing/selftests/bpf/progs/bpf_loop.c
181
int e = map1_lookup_elem(5);
tools/testing/selftests/bpf/progs/bpf_loop.c
190
map1_update_elem(5, e + 1);
tools/testing/selftests/bpf/progs/bpf_loop.c
212
int e = map1_lookup_elem(11);
tools/testing/selftests/bpf/progs/bpf_loop.c
221
map1_update_elem(11, e + 1);
tools/testing/selftests/bpf/progs/btf_dump_test_case_bitfields.c
36
short e: 5;
tools/testing/selftests/bpf/progs/btf_dump_test_case_bitfields.c
64
int e; /* combined with previous bitfield */
tools/testing/selftests/bpf/progs/btf_dump_test_case_packing.c
36
} __attribute__((packed)) e;
tools/testing/selftests/bpf/progs/btf_dump_test_case_syntax.c
117
int e[5];
tools/testing/selftests/bpf/progs/btf_dump_test_case_syntax.c
137
ptr_arr_t e;
tools/testing/selftests/bpf/progs/btf_dump_test_case_syntax.c
172
enum e2 e;
tools/testing/selftests/bpf/progs/btf_dump_test_case_syntax.c
202
int e;
tools/testing/selftests/bpf/progs/btf_dump_test_case_syntax.c
223
} e;
tools/testing/selftests/bpf/progs/core_reloc_types.h
539
int a, b, c, d, e, f, g, h;
tools/testing/selftests/bpf/progs/core_reloc_types.h
561
int e[3] __bpf_aligned;
tools/testing/selftests/bpf/progs/core_reloc_types.h
574
arr_t e;
tools/testing/selftests/bpf/progs/core_reloc_types.h
600
arr4_t e;
tools/testing/selftests/bpf/progs/fentry_many_args.c
11
int BPF_PROG(test1, __u64 a, void *b, short c, int d, void *e, char f,
tools/testing/selftests/bpf/progs/fentry_many_args.c
15
e == (void *)20 && f == 21 && g == 22;
tools/testing/selftests/bpf/progs/fentry_many_args.c
21
int BPF_PROG(test2, __u64 a, void *b, short c, int d, void *e, char f,
tools/testing/selftests/bpf/progs/fentry_many_args.c
25
e == (void *)20 && f == 21 && g == 22 && h == 23 &&
tools/testing/selftests/bpf/progs/fentry_many_args.c
32
int BPF_PROG(test3, __u64 a, __u64 b, __u64 c, __u64 d, __u64 e, __u64 f,
tools/testing/selftests/bpf/progs/fentry_many_args.c
36
e == 20 && f == 21 && g == 22 && h == 23 &&
tools/testing/selftests/bpf/progs/fentry_test.c
43
int BPF_PROG(test5, __u64 a, void *b, short c, int d, __u64 e)
tools/testing/selftests/bpf/progs/fentry_test.c
46
e == 15;
tools/testing/selftests/bpf/progs/fentry_test.c
52
int BPF_PROG(test6, __u64 a, void *b, short c, int d, void * e, __u64 f)
tools/testing/selftests/bpf/progs/fentry_test.c
55
e == (void *)20 && f == 21;
tools/testing/selftests/bpf/progs/fexit_many_args.c
11
int BPF_PROG(test1, __u64 a, void *b, short c, int d, void *e, char f,
tools/testing/selftests/bpf/progs/fexit_many_args.c
15
e == (void *)20 && f == 21 && g == 22 && ret == 133;
tools/testing/selftests/bpf/progs/fexit_many_args.c
21
int BPF_PROG(test2, __u64 a, void *b, short c, int d, void *e, char f,
tools/testing/selftests/bpf/progs/fexit_many_args.c
26
e == (void *)20 && f == 21 && g == 22 && h == 23 &&
tools/testing/selftests/bpf/progs/fexit_many_args.c
33
int BPF_PROG(test3, __u64 a, __u64 b, __u64 c, __u64 d, __u64 e, __u64 f,
tools/testing/selftests/bpf/progs/fexit_many_args.c
37
e == 20 && f == 21 && g == 22 && h == 23 &&
tools/testing/selftests/bpf/progs/fexit_test.c
44
int BPF_PROG(test5, __u64 a, void *b, short c, int d, __u64 e, int ret)
tools/testing/selftests/bpf/progs/fexit_test.c
47
e == 15 && ret == 65;
tools/testing/selftests/bpf/progs/fexit_test.c
53
int BPF_PROG(test6, __u64 a, void *b, short c, int d, void *e, __u64 f, int ret)
tools/testing/selftests/bpf/progs/fexit_test.c
56
e == (void *)20 && f == 21 && ret == 111;
tools/testing/selftests/bpf/progs/fsession_test.c
64
int BPF_PROG(test4, __u64 a, void *b, short c, int d, __u64 e, int ret)
tools/testing/selftests/bpf/progs/fsession_test.c
70
e == 15 && ret == 0;
tools/testing/selftests/bpf/progs/fsession_test.c
75
e == 15 && ret == 65;
tools/testing/selftests/bpf/progs/kfunc_call_test.c
85
int e = 0;
tools/testing/selftests/bpf/progs/kfunc_call_test.c
95
bpf_kfunc_call_test_mem_len_pass1(&e, sizeof(e));
tools/testing/selftests/bpf/progs/modify_return.c
55
int BPF_PROG(fentry_test2, int a, int *b, short c, int d, void *e, char f,
tools/testing/selftests/bpf/progs/modify_return.c
65
int BPF_PROG(fmod_ret_test2, int a, int *b, short c, int d, void *e, char f,
tools/testing/selftests/bpf/progs/modify_return.c
76
int BPF_PROG(fexit_test2, int a, int *b, short c, int d, void *e, char f,
tools/testing/selftests/bpf/progs/percpu_alloc_array.c
113
struct elem *e;
tools/testing/selftests/bpf/progs/percpu_alloc_array.c
116
e = bpf_map_lookup_elem(&array, &index);
tools/testing/selftests/bpf/progs/percpu_alloc_array.c
117
if (!e)
tools/testing/selftests/bpf/progs/percpu_alloc_array.c
121
p = bpf_kptr_xchg(&e->pc, NULL);
tools/testing/selftests/bpf/progs/percpu_alloc_array.c
135
struct elem *e;
tools/testing/selftests/bpf/progs/percpu_alloc_array.c
140
e = bpf_map_lookup_elem(&array, &index);
tools/testing/selftests/bpf/progs/percpu_alloc_array.c
141
if (!e)
tools/testing/selftests/bpf/progs/percpu_alloc_array.c
145
p = e->pc;
tools/testing/selftests/bpf/progs/percpu_alloc_array.c
151
p1 = bpf_kptr_xchg(&e->pc, p);
tools/testing/selftests/bpf/progs/percpu_alloc_array.c
170
p1 = bpf_kptr_xchg(&e->pc, NULL);
tools/testing/selftests/bpf/progs/percpu_alloc_array.c
29
struct elem *e;
tools/testing/selftests/bpf/progs/percpu_alloc_array.c
32
e = bpf_map_lookup_elem(&array, &index);
tools/testing/selftests/bpf/progs/percpu_alloc_array.c
33
if (!e)
tools/testing/selftests/bpf/progs/percpu_alloc_array.c
40
p = bpf_kptr_xchg(&e->pc, p);
tools/testing/selftests/bpf/progs/percpu_alloc_array.c
53
struct elem *e;
tools/testing/selftests/bpf/progs/percpu_alloc_array.c
56
e = bpf_map_lookup_elem(&array, &index);
tools/testing/selftests/bpf/progs/percpu_alloc_array.c
57
if (!e)
tools/testing/selftests/bpf/progs/percpu_alloc_array.c
60
p = e->pc;
tools/testing/selftests/bpf/progs/percpu_alloc_array.c
83
struct elem *e;
tools/testing/selftests/bpf/progs/percpu_alloc_array.c
88
e = bpf_map_lookup_elem(&array, &index);
tools/testing/selftests/bpf/progs/percpu_alloc_array.c
89
if (!e)
tools/testing/selftests/bpf/progs/percpu_alloc_array.c
92
p = e->pc;
tools/testing/selftests/bpf/progs/percpu_alloc_cgrp_local_storage.c
27
struct elem *e;
tools/testing/selftests/bpf/progs/percpu_alloc_cgrp_local_storage.c
30
e = bpf_cgrp_storage_get(&cgrp, task->cgroups->dfl_cgrp, 0,
tools/testing/selftests/bpf/progs/percpu_alloc_cgrp_local_storage.c
32
if (!e)
tools/testing/selftests/bpf/progs/percpu_alloc_cgrp_local_storage.c
39
p = bpf_kptr_xchg(&e->pc, p);
tools/testing/selftests/bpf/progs/percpu_alloc_cgrp_local_storage.c
53
struct elem *e;
tools/testing/selftests/bpf/progs/percpu_alloc_cgrp_local_storage.c
56
e = bpf_cgrp_storage_get(&cgrp, task->cgroups->dfl_cgrp, 0, 0);
tools/testing/selftests/bpf/progs/percpu_alloc_cgrp_local_storage.c
57
if (!e)
tools/testing/selftests/bpf/progs/percpu_alloc_cgrp_local_storage.c
60
p = e->pc;
tools/testing/selftests/bpf/progs/percpu_alloc_cgrp_local_storage.c
82
struct elem *e;
tools/testing/selftests/bpf/progs/percpu_alloc_cgrp_local_storage.c
89
e = bpf_cgrp_storage_get(&cgrp, task->cgroups->dfl_cgrp, 0, 0);
tools/testing/selftests/bpf/progs/percpu_alloc_cgrp_local_storage.c
90
if (!e)
tools/testing/selftests/bpf/progs/percpu_alloc_cgrp_local_storage.c
93
p = e->pc;
tools/testing/selftests/bpf/progs/percpu_alloc_fail.c
103
p1 = bpf_kptr_xchg(&e->pc, p);
tools/testing/selftests/bpf/progs/percpu_alloc_fail.c
43
struct elem *e;
tools/testing/selftests/bpf/progs/percpu_alloc_fail.c
46
e = bpf_map_lookup_elem(&array, &index);
tools/testing/selftests/bpf/progs/percpu_alloc_fail.c
47
if (!e)
tools/testing/selftests/bpf/progs/percpu_alloc_fail.c
54
p = bpf_kptr_xchg(&e->pc, p);
tools/testing/selftests/bpf/progs/percpu_alloc_fail.c
58
e->pc = (struct val_t __percpu_kptr *)ret;
tools/testing/selftests/bpf/progs/percpu_alloc_fail.c
68
struct elem *e;
tools/testing/selftests/bpf/progs/percpu_alloc_fail.c
71
e = bpf_map_lookup_elem(&array, &index);
tools/testing/selftests/bpf/progs/percpu_alloc_fail.c
72
if (!e)
tools/testing/selftests/bpf/progs/percpu_alloc_fail.c
79
p = bpf_kptr_xchg(&e->pc, p2);
tools/testing/selftests/bpf/progs/percpu_alloc_fail.c
92
struct elem *e;
tools/testing/selftests/bpf/progs/percpu_alloc_fail.c
95
e = bpf_map_lookup_elem(&array, &index);
tools/testing/selftests/bpf/progs/percpu_alloc_fail.c
96
if (!e)
tools/testing/selftests/bpf/progs/res_spin_lock.c
101
e = bpf_map_lookup_elem(&arrmap, &key);
tools/testing/selftests/bpf/progs/res_spin_lock.c
102
if (!e)
tools/testing/selftests/bpf/progs/res_spin_lock.c
104
locks[i] = &e->lock;
tools/testing/selftests/bpf/progs/res_spin_lock.c
113
e = bpf_map_lookup_elem(&arrmap, &key);
tools/testing/selftests/bpf/progs/res_spin_lock.c
114
if (!e)
tools/testing/selftests/bpf/progs/res_spin_lock.c
116
locks[i] = &e->lock;
tools/testing/selftests/bpf/progs/res_spin_lock.c
88
struct arr_elem *e;
tools/testing/selftests/bpf/progs/test_core_reloc_mods.c
17
int a, b, c, d, e, f, g, h;
tools/testing/selftests/bpf/progs/test_core_reloc_mods.c
39
int e[3];
tools/testing/selftests/bpf/progs/test_core_reloc_mods.c
66
CORE_READ(&out->e, &in->e[2]) ||
tools/testing/selftests/bpf/progs/test_global_func9.c
116
enum E e = E_ITEM;
tools/testing/selftests/bpf/progs/test_global_func9.c
118
result |= qux(&e);
tools/testing/selftests/bpf/progs/test_global_func9.c
53
__noinline int qux(enum E *e)
tools/testing/selftests/bpf/progs/test_global_func9.c
55
if (e)
tools/testing/selftests/bpf/progs/test_global_func9.c
56
return *e;
tools/testing/selftests/bpf/progs/timer_crash.c
32
struct map_elem *e, value = {};
tools/testing/selftests/bpf/progs/timer_crash.c
46
e = bpf_map_lookup_elem(map, &(int){0});
tools/testing/selftests/bpf/progs/timer_crash.c
47
if (!e)
tools/testing/selftests/bpf/progs/timer_crash.c
49
bpf_timer_cancel(&e->timer);
tools/testing/selftests/bpf/progs/tracing_struct.c
109
int, c, int, d, struct bpf_testmod_struct_arg_2, e)
tools/testing/selftests/bpf/progs/tracing_struct.c
115
t4_e_a = e.a;
tools/testing/selftests/bpf/progs/tracing_struct.c
116
t4_e_b = e.b;
tools/testing/selftests/bpf/progs/tracing_struct.c
122
int, c, int, d, struct bpf_testmod_struct_arg_2, e, int, ret)
tools/testing/selftests/bpf/progs/tracing_struct_many_args.c
24
void *, e, struct bpf_testmod_struct_arg_4, f)
tools/testing/selftests/bpf/progs/tracing_struct_many_args.c
30
t7_e = (long)e;
tools/testing/selftests/bpf/progs/tracing_struct_many_args.c
38
void *, e, struct bpf_testmod_struct_arg_4, f, int, ret)
tools/testing/selftests/bpf/progs/tracing_struct_many_args.c
46
void *, e, struct bpf_testmod_struct_arg_4, f, int, g)
tools/testing/selftests/bpf/progs/tracing_struct_many_args.c
52
t8_e = (long)e;
tools/testing/selftests/bpf/progs/tracing_struct_many_args.c
61
void *, e, struct bpf_testmod_struct_arg_4, f, int, g,
tools/testing/selftests/bpf/progs/tracing_struct_many_args.c
69
int BPF_PROG2(test_struct_many_args_5, __u64, a, void *, b, short, c, int, d, void *, e,
tools/testing/selftests/bpf/progs/tracing_struct_many_args.c
76
t9_e = (long)e;
tools/testing/selftests/bpf/progs/tracing_struct_many_args.c
88
int BPF_PROG2(test_struct_many_args_6, __u64, a, void *, b, short, c, int, d, void *, e,
tools/testing/selftests/bpf/sdt.h
51
# define _SDT_ASM_5(a, b, c, d, e) a,b,c,d,e;
tools/testing/selftests/bpf/sdt.h
59
# define _SDT_DEPAREN_5(a,b,c,d,e) a b c d e
tools/testing/selftests/bpf/sdt.h
60
# define _SDT_DEPAREN_6(a,b,c,d,e,f) a b c d e f
tools/testing/selftests/bpf/sdt.h
61
# define _SDT_DEPAREN_7(a,b,c,d,e,f,g) a b c d e f g
tools/testing/selftests/bpf/sdt.h
62
# define _SDT_DEPAREN_8(a,b,c,d,e,f,g,h) a b c d e f g h
tools/testing/selftests/bpf/sdt.h
63
# define _SDT_DEPAREN_9(a,b,c,d,e,f,g,h,i) a b c d e f g h i
tools/testing/selftests/bpf/sdt.h
64
# define _SDT_DEPAREN_10(a,b,c,d,e,f,g,h,i,j) a b c d e f g h i j
tools/testing/selftests/bpf/sdt.h
65
# define _SDT_DEPAREN_11(a,b,c,d,e,f,g,h,i,j,k) a b c d e f g h i j k
tools/testing/selftests/bpf/sdt.h
66
# define _SDT_DEPAREN_12(a,b,c,d,e,f,g,h,i,j,k,l) a b c d e f g h i j k l
tools/testing/selftests/bpf/sdt.h
87
# define _SDT_ASM_5(a, b, c, d, e) _SDT_S(a) "," _SDT_S(b) "," \
tools/testing/selftests/bpf/sdt.h
89
_SDT_S(e) "\n"
tools/testing/selftests/bpf/test_kmods/bpf_test_rqspinlock.c
265
struct perf_event *e;
tools/testing/selftests/bpf/test_kmods/bpf_test_rqspinlock.c
267
e = perf_event_create_kernel_counter(&hw_attr, i, NULL, nmi_cb, NULL);
tools/testing/selftests/bpf/test_kmods/bpf_test_rqspinlock.c
268
if (IS_ERR(e)) {
tools/testing/selftests/bpf/test_kmods/bpf_test_rqspinlock.c
269
ret = PTR_ERR(e);
tools/testing/selftests/bpf/test_kmods/bpf_test_rqspinlock.c
272
rqsl_evts[i - 1] = e;
tools/testing/selftests/bpf/test_kmods/bpf_testmod.c
100
int c, int d, struct bpf_testmod_struct_arg_2 e) {
tools/testing/selftests/bpf/test_kmods/bpf_testmod.c
101
bpf_testmod_test_struct_arg_result = a.a + b + c + d + e.a + e.b;
tools/testing/selftests/bpf/test_kmods/bpf_testmod.c
118
bpf_testmod_test_struct_arg_7(u64 a, void *b, short c, int d, void *e,
tools/testing/selftests/bpf/test_kmods/bpf_testmod.c
122
(long)e + f.a + f.b;
tools/testing/selftests/bpf/test_kmods/bpf_testmod.c
127
bpf_testmod_test_struct_arg_8(u64 a, void *b, short c, int d, void *e,
tools/testing/selftests/bpf/test_kmods/bpf_testmod.c
131
(long)e + f.a + f.b + g;
tools/testing/selftests/bpf/test_kmods/bpf_testmod.c
136
bpf_testmod_test_struct_arg_9(u64 a, void *b, short c, int d, void *e, char f,
tools/testing/selftests/bpf/test_kmods/bpf_testmod.c
139
bpf_testmod_test_struct_arg_result = a + (long)b + c + d + (long)e +
tools/testing/selftests/bpf/test_kmods/bpf_testmod.c
434
void *e, char f, int g)
tools/testing/selftests/bpf/test_kmods/bpf_testmod.c
436
return a + (long)b + c + d + (long)e + f + g;
tools/testing/selftests/bpf/test_kmods/bpf_testmod.c
440
void *e, char f, int g,
tools/testing/selftests/bpf/test_kmods/bpf_testmod.c
444
return a + (long)b + c + d + (long)e + f + g + h + i + j + k;
tools/testing/selftests/bpf/test_progs.c
1229
const struct btf_enum *e;
tools/testing/selftests/bpf/test_progs.c
1239
e = btf_enum(t);
tools/testing/selftests/bpf/test_progs.c
1240
for (j = 0, vlen = btf_vlen(t); j < vlen; j++, e++) {
tools/testing/selftests/bpf/test_progs.c
1241
name = btf__str_by_offset(btf, e->name_off);
tools/testing/selftests/bpf/test_progs.c
1243
return e->val;
tools/testing/selftests/bpf/usdt.h
536
#define __usdt_asm_ops6(a,b,c,d,e,x) __usdt_asm_ops5(a,b,c,d,e), __usdt_asm_op(6, x)
tools/testing/selftests/bpf/usdt.h
537
#define __usdt_asm_ops7(a,b,c,d,e,f,x) __usdt_asm_ops6(a,b,c,d,e,f), __usdt_asm_op(7, x)
tools/testing/selftests/bpf/usdt.h
538
#define __usdt_asm_ops8(a,b,c,d,e,f,g,x) __usdt_asm_ops7(a,b,c,d,e,f,g), __usdt_asm_op(8, x)
tools/testing/selftests/bpf/usdt.h
539
#define __usdt_asm_ops9(a,b,c,d,e,f,g,h,x) __usdt_asm_ops8(a,b,c,d,e,f,g,h), __usdt_asm_op(9, x)
tools/testing/selftests/bpf/usdt.h
540
#define __usdt_asm_ops10(a,b,c,d,e,f,g,h,i,x) __usdt_asm_ops9(a,b,c,d,e,f,g,h,i), __usdt_asm_op(10, x)
tools/testing/selftests/bpf/usdt.h
541
#define __usdt_asm_ops11(a,b,c,d,e,f,g,h,i,j,x) __usdt_asm_ops10(a,b,c,d,e,f,g,h,i,j), __usdt_asm_op(11, x)
tools/testing/selftests/bpf/usdt.h
542
#define __usdt_asm_ops12(a,b,c,d,e,f,g,h,i,j,k,x) __usdt_asm_ops11(a,b,c,d,e,f,g,h,i,j,k), __usdt_asm_op(12, x)
tools/testing/selftests/bpf/veristat.c
1843
struct btf_enum *e = btf_enum(t);
tools/testing/selftests/bpf/veristat.c
1846
for (i = 0; i < n; ++i, ++e) {
tools/testing/selftests/bpf/veristat.c
1847
const char *cur_name = btf__name_by_offset(btf, e->name_off);
tools/testing/selftests/bpf/veristat.c
1850
*retval = e->val;
tools/testing/selftests/bpf/veristat.c
1855
struct btf_enum64 *e = btf_enum64(t);
tools/testing/selftests/bpf/veristat.c
1858
for (i = 0; i < n; ++i, ++e) {
tools/testing/selftests/bpf/veristat.c
1859
const char *cur_name = btf__name_by_offset(btf, e->name_off);
tools/testing/selftests/bpf/veristat.c
1860
__u64 value = btf_enum64_value(e);
tools/testing/selftests/connector/proc_filter.c
172
event->event_data.id.e.euid);
tools/testing/selftests/connector/proc_filter.c
179
event->event_data.id.e.egid);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1007
struct epoll_event e;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1017
e.events = EPOLLIN;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1018
ASSERT_EQ(epoll_ctl(efd[1], EPOLL_CTL_ADD, sfd[0], &e), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1020
e.events = EPOLLIN;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1021
ASSERT_EQ(epoll_ctl(efd[0], EPOLL_CTL_ADD, efd[1], &e), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1028
EXPECT_EQ(epoll_wait(efd[0], &e, 1, 0), 1);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1033
EXPECT_EQ(epoll_wait(efd[0], &e, 1, 0), 1);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1055
struct epoll_event e;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1065
e.events = EPOLLIN | EPOLLET;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1066
ASSERT_EQ(epoll_ctl(efd[1], EPOLL_CTL_ADD, sfd[0], &e), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1068
e.events = EPOLLIN;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1069
ASSERT_EQ(epoll_ctl(efd[0], EPOLL_CTL_ADD, efd[1], &e), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1076
EXPECT_EQ(epoll_wait(efd[0], &e, 1, 0), 1);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1081
EXPECT_EQ(epoll_wait(efd[0], &e, 1, 0), 1);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1103
struct epoll_event e;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1113
e.events = EPOLLIN;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1114
ASSERT_EQ(epoll_ctl(efd[1], EPOLL_CTL_ADD, sfd[0], &e), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1116
e.events = EPOLLIN | EPOLLET;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1117
ASSERT_EQ(epoll_ctl(efd[0], EPOLL_CTL_ADD, efd[1], &e), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1124
EXPECT_EQ(epoll_wait(efd[0], &e, 1, 0), 1);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1129
EXPECT_EQ(epoll_wait(efd[0], &e, 1, 0), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1151
struct epoll_event e;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1161
e.events = EPOLLIN | EPOLLET;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1162
ASSERT_EQ(epoll_ctl(efd[1], EPOLL_CTL_ADD, sfd[0], &e), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1164
e.events = EPOLLIN | EPOLLET;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1165
ASSERT_EQ(epoll_ctl(efd[0], EPOLL_CTL_ADD, efd[1], &e), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1172
EXPECT_EQ(epoll_wait(efd[0], &e, 1, 0), 1);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1177
EXPECT_EQ(epoll_wait(efd[0], &e, 1, 0), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1197
struct epoll_event e;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1210
e.events = EPOLLIN;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1211
ASSERT_EQ(epoll_ctl(ctx.efd[1], EPOLL_CTL_ADD, ctx.sfd[0], &e), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1213
e.events = EPOLLIN;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1214
ASSERT_EQ(epoll_ctl(ctx.efd[0], EPOLL_CTL_ADD, ctx.efd[1], &e), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1220
if (epoll_wait(ctx.efd[0], &e, 1, -1) > 0)
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1249
struct epoll_event e;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1262
e.events = EPOLLIN | EPOLLET;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1263
ASSERT_EQ(epoll_ctl(ctx.efd[1], EPOLL_CTL_ADD, ctx.sfd[0], &e), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1265
e.events = EPOLLIN;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1266
ASSERT_EQ(epoll_ctl(ctx.efd[0], EPOLL_CTL_ADD, ctx.efd[1], &e), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1272
if (epoll_wait(ctx.efd[0], &e, 1, -1) > 0)
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1301
struct epoll_event e;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1314
e.events = EPOLLIN;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1315
ASSERT_EQ(epoll_ctl(ctx.efd[1], EPOLL_CTL_ADD, ctx.sfd[0], &e), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1317
e.events = EPOLLIN | EPOLLET;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1318
ASSERT_EQ(epoll_ctl(ctx.efd[0], EPOLL_CTL_ADD, ctx.efd[1], &e), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1324
if (epoll_wait(ctx.efd[0], &e, 1, -1) > 0)
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1353
struct epoll_event e;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1366
e.events = EPOLLIN | EPOLLET;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1367
ASSERT_EQ(epoll_ctl(ctx.efd[1], EPOLL_CTL_ADD, ctx.sfd[0], &e), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1369
e.events = EPOLLIN | EPOLLET;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1370
ASSERT_EQ(epoll_ctl(ctx.efd[0], EPOLL_CTL_ADD, ctx.efd[1], &e), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1376
if (epoll_wait(ctx.efd[0], &e, 1, -1) > 0)
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1405
struct epoll_event e;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1418
e.events = EPOLLIN;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1419
ASSERT_EQ(epoll_ctl(ctx.efd[1], EPOLL_CTL_ADD, ctx.sfd[0], &e), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1421
e.events = EPOLLIN;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1422
ASSERT_EQ(epoll_ctl(ctx.efd[0], EPOLL_CTL_ADD, ctx.efd[1], &e), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1428
if (epoll_wait(ctx.efd[0], &e, 1, -1) > 0)
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1456
struct epoll_event e;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1469
e.events = EPOLLIN | EPOLLET;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1470
ASSERT_EQ(epoll_ctl(ctx.efd[1], EPOLL_CTL_ADD, ctx.sfd[0], &e), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1472
e.events = EPOLLIN;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1473
ASSERT_EQ(epoll_ctl(ctx.efd[0], EPOLL_CTL_ADD, ctx.efd[1], &e), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1479
if (epoll_wait(ctx.efd[0], &e, 1, -1) > 0)
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1507
struct epoll_event e;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1520
e.events = EPOLLIN;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1521
ASSERT_EQ(epoll_ctl(ctx.efd[1], EPOLL_CTL_ADD, ctx.sfd[0], &e), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1523
e.events = EPOLLIN | EPOLLET;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1524
ASSERT_EQ(epoll_ctl(ctx.efd[0], EPOLL_CTL_ADD, ctx.efd[1], &e), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1530
if (epoll_wait(ctx.efd[0], &e, 1, -1) > 0)
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1558
struct epoll_event e;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1571
e.events = EPOLLIN | EPOLLET;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1572
ASSERT_EQ(epoll_ctl(ctx.efd[1], EPOLL_CTL_ADD, ctx.sfd[0], &e), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1574
e.events = EPOLLIN | EPOLLET;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1575
ASSERT_EQ(epoll_ctl(ctx.efd[0], EPOLL_CTL_ADD, ctx.efd[1], &e), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1581
if (epoll_wait(ctx.efd[0], &e, 1, -1) > 0)
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1609
struct epoll_event e;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1622
e.events = EPOLLIN;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1623
ASSERT_EQ(epoll_ctl(ctx.efd[1], EPOLL_CTL_ADD, ctx.sfd[0], &e), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1625
e.events = EPOLLIN;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1626
ASSERT_EQ(epoll_ctl(ctx.efd[0], EPOLL_CTL_ADD, ctx.efd[1], &e), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1632
if (epoll_wait(ctx.efd[1], &e, 1, -1) > 0)
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1661
struct epoll_event e;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
167
struct epoll_event e;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1674
e.events = EPOLLIN | EPOLLET;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1675
ASSERT_EQ(epoll_ctl(ctx.efd[1], EPOLL_CTL_ADD, ctx.sfd[0], &e), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1677
e.events = EPOLLIN;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1678
ASSERT_EQ(epoll_ctl(ctx.efd[0], EPOLL_CTL_ADD, ctx.efd[1], &e), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1684
if (epoll_wait(ctx.efd[1], &e, 1, -1) > 0)
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1713
struct epoll_event e;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1726
e.events = EPOLLIN;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1727
ASSERT_EQ(epoll_ctl(ctx.efd[1], EPOLL_CTL_ADD, ctx.sfd[0], &e), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1729
e.events = EPOLLIN | EPOLLET;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1730
ASSERT_EQ(epoll_ctl(ctx.efd[0], EPOLL_CTL_ADD, ctx.efd[1], &e), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1736
if (epoll_wait(ctx.efd[1], &e, 1, -1) > 0)
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
174
e.events = EPOLLIN;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
175
ASSERT_EQ(epoll_ctl(efd, EPOLL_CTL_ADD, sfd[0], &e), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1765
struct epoll_event e;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1778
e.events = EPOLLIN | EPOLLET;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1779
ASSERT_EQ(epoll_ctl(ctx.efd[1], EPOLL_CTL_ADD, ctx.sfd[0], &e), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1781
e.events = EPOLLIN | EPOLLET;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1782
ASSERT_EQ(epoll_ctl(ctx.efd[0], EPOLL_CTL_ADD, ctx.efd[1], &e), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1788
if (epoll_wait(ctx.efd[1], &e, 1, -1) > 0)
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
179
EXPECT_EQ(epoll_wait(efd, &e, 1, 0), 1);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
180
EXPECT_EQ(epoll_wait(efd, &e, 1, 0), 1);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1818
struct epoll_event e;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1831
e.events = EPOLLIN;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1832
ASSERT_EQ(epoll_ctl(ctx.efd[1], EPOLL_CTL_ADD, ctx.sfd[0], &e), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1834
e.events = EPOLLIN;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1835
ASSERT_EQ(epoll_ctl(ctx.efd[0], EPOLL_CTL_ADD, ctx.efd[1], &e), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1844
if (epoll_wait(ctx.efd[1], &e, 1, 0) > 0)
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1875
struct epoll_event e;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1888
e.events = EPOLLIN | EPOLLET;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1889
ASSERT_EQ(epoll_ctl(ctx.efd[1], EPOLL_CTL_ADD, ctx.sfd[0], &e), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1891
e.events = EPOLLIN;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1892
ASSERT_EQ(epoll_ctl(ctx.efd[0], EPOLL_CTL_ADD, ctx.efd[1], &e), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1901
if (epoll_wait(ctx.efd[1], &e, 1, 0) > 0)
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1932
struct epoll_event e;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1945
e.events = EPOLLIN;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1946
ASSERT_EQ(epoll_ctl(ctx.efd[1], EPOLL_CTL_ADD, ctx.sfd[0], &e), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1948
e.events = EPOLLIN | EPOLLET;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1949
ASSERT_EQ(epoll_ctl(ctx.efd[0], EPOLL_CTL_ADD, ctx.efd[1], &e), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1958
if (epoll_wait(ctx.efd[1], &e, 1, 0) > 0)
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
198
struct epoll_event e;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
1989
struct epoll_event e;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
2002
e.events = EPOLLIN | EPOLLET;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
2003
ASSERT_EQ(epoll_ctl(ctx.efd[1], EPOLL_CTL_ADD, ctx.sfd[0], &e), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
2005
e.events = EPOLLIN | EPOLLET;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
2006
ASSERT_EQ(epoll_ctl(ctx.efd[0], EPOLL_CTL_ADD, ctx.efd[1], &e), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
2015
if (epoll_wait(ctx.efd[1], &e, 1, 0) > 0)
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
2045
struct epoll_event e;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
205
e.events = EPOLLIN | EPOLLET;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
2058
e.events = EPOLLIN;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
2059
ASSERT_EQ(epoll_ctl(ctx.efd[1], EPOLL_CTL_ADD, ctx.sfd[0], &e), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
206
ASSERT_EQ(epoll_ctl(efd, EPOLL_CTL_ADD, sfd[0], &e), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
2061
e.events = EPOLLIN;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
2062
ASSERT_EQ(epoll_ctl(ctx.efd[0], EPOLL_CTL_ADD, ctx.efd[1], &e), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
2068
if (epoll_wait(ctx.efd[1], &e, 1, -1) > 0)
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
2097
struct epoll_event e;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
210
EXPECT_EQ(epoll_wait(efd, &e, 1, 0), 1);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
211
EXPECT_EQ(epoll_wait(efd, &e, 1, 0), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
2110
e.events = EPOLLIN | EPOLLET;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
2111
ASSERT_EQ(epoll_ctl(ctx.efd[1], EPOLL_CTL_ADD, ctx.sfd[0], &e), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
2113
e.events = EPOLLIN;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
2114
ASSERT_EQ(epoll_ctl(ctx.efd[0], EPOLL_CTL_ADD, ctx.efd[1], &e), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
2120
if (epoll_wait(ctx.efd[1], &e, 1, -1) > 0)
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
2149
struct epoll_event e;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
2162
e.events = EPOLLIN;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
2163
ASSERT_EQ(epoll_ctl(ctx.efd[1], EPOLL_CTL_ADD, ctx.sfd[0], &e), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
2165
e.events = EPOLLIN | EPOLLET;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
2166
ASSERT_EQ(epoll_ctl(ctx.efd[0], EPOLL_CTL_ADD, ctx.efd[1], &e), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
2172
if (epoll_wait(ctx.efd[1], &e, 1, -1) > 0)
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
2201
struct epoll_event e;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
2214
e.events = EPOLLIN | EPOLLET;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
2215
ASSERT_EQ(epoll_ctl(ctx.efd[1], EPOLL_CTL_ADD, ctx.sfd[0], &e), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
2217
e.events = EPOLLIN | EPOLLET;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
2218
ASSERT_EQ(epoll_ctl(ctx.efd[0], EPOLL_CTL_ADD, ctx.efd[1], &e), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
2224
if (epoll_wait(ctx.efd[1], &e, 1, -1) > 0)
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
2254
struct epoll_event e;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
2267
e.events = EPOLLIN;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
2268
ASSERT_EQ(epoll_ctl(ctx.efd[1], EPOLL_CTL_ADD, ctx.sfd[0], &e), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
2270
e.events = EPOLLIN;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
2271
ASSERT_EQ(epoll_ctl(ctx.efd[0], EPOLL_CTL_ADD, ctx.efd[1], &e), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
2280
if (epoll_wait(ctx.efd[1], &e, 1, 0) > 0)
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
2310
struct epoll_event e;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
2323
e.events = EPOLLIN | EPOLLET;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
2324
ASSERT_EQ(epoll_ctl(ctx.efd[1], EPOLL_CTL_ADD, ctx.sfd[0], &e), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
2326
e.events = EPOLLIN;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
2327
ASSERT_EQ(epoll_ctl(ctx.efd[0], EPOLL_CTL_ADD, ctx.efd[1], &e), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
2333
if (epoll_wait(ctx.efd[1], &e, 1, -1) > 0)
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
2363
struct epoll_event e;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
2376
e.events = EPOLLIN;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
2377
ASSERT_EQ(epoll_ctl(ctx.efd[1], EPOLL_CTL_ADD, ctx.sfd[0], &e), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
2379
e.events = EPOLLIN | EPOLLET;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
2380
ASSERT_EQ(epoll_ctl(ctx.efd[0], EPOLL_CTL_ADD, ctx.efd[1], &e), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
2389
if (epoll_wait(ctx.efd[1], &e, 1, 0) > 0)
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
2419
struct epoll_event e;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
2432
e.events = EPOLLIN | EPOLLET;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
2433
ASSERT_EQ(epoll_ctl(ctx.efd[1], EPOLL_CTL_ADD, ctx.sfd[0], &e), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
2435
e.events = EPOLLIN | EPOLLET;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
2436
ASSERT_EQ(epoll_ctl(ctx.efd[0], EPOLL_CTL_ADD, ctx.efd[1], &e), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
2442
if (epoll_wait(ctx.efd[1], &e, 1, -1) > 0)
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
2703
struct epoll_event e;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
2720
e.events = EPOLLIN;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
2721
ASSERT_EQ(epoll_ctl(ctx.efd[1], EPOLL_CTL_ADD, ctx.sfd[0], &e), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
2723
e.events = EPOLLIN;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
2724
ASSERT_EQ(epoll_ctl(ctx.efd[2], EPOLL_CTL_ADD, ctx.sfd[2], &e), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
2726
e.events = EPOLLIN;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
2727
ASSERT_EQ(epoll_ctl(ctx.efd[0], EPOLL_CTL_ADD, ctx.efd[1], &e), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
2729
e.events = EPOLLIN;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
2730
ASSERT_EQ(epoll_ctl(ctx.efd[0], EPOLL_CTL_ADD, ctx.efd[2], &e), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
2736
if (epoll_wait(ctx.efd[0], &e, 1, -1) > 0)
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
2768
struct epoll_event e;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
2785
e.events = EPOLLIN;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
2786
ASSERT_EQ(epoll_ctl(ctx.efd[1], EPOLL_CTL_ADD, ctx.sfd[0], &e), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
2788
e.events = EPOLLIN;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
2789
ASSERT_EQ(epoll_ctl(ctx.efd[2], EPOLL_CTL_ADD, ctx.sfd[2], &e), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
2791
e.events = EPOLLIN | EPOLLET;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
2792
ASSERT_EQ(epoll_ctl(ctx.efd[0], EPOLL_CTL_ADD, ctx.efd[1], &e), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
2794
e.events = EPOLLIN | EPOLLET;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
2795
ASSERT_EQ(epoll_ctl(ctx.efd[0], EPOLL_CTL_ADD, ctx.efd[2], &e), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
2801
if (epoll_wait(ctx.efd[0], &e, 1, -1) > 0)
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
2833
struct epoll_event e;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
2850
e.events = EPOLLIN;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
2851
ASSERT_EQ(epoll_ctl(ctx.efd[1], EPOLL_CTL_ADD, ctx.sfd[0], &e), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
2853
e.events = EPOLLIN;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
2854
ASSERT_EQ(epoll_ctl(ctx.efd[2], EPOLL_CTL_ADD, ctx.sfd[2], &e), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
2856
e.events = EPOLLIN;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
2857
ASSERT_EQ(epoll_ctl(ctx.efd[0], EPOLL_CTL_ADD, ctx.efd[1], &e), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
2859
e.events = EPOLLIN;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
2860
ASSERT_EQ(epoll_ctl(ctx.efd[0], EPOLL_CTL_ADD, ctx.efd[2], &e), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
2866
if (epoll_wait(ctx.efd[0], &e, 1, -1) > 0)
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
2898
struct epoll_event e;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
2915
e.events = EPOLLIN;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
2916
ASSERT_EQ(epoll_ctl(ctx.efd[1], EPOLL_CTL_ADD, ctx.sfd[0], &e), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
2918
e.events = EPOLLIN;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
2919
ASSERT_EQ(epoll_ctl(ctx.efd[2], EPOLL_CTL_ADD, ctx.sfd[2], &e), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
2921
e.events = EPOLLIN | EPOLLET;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
2922
ASSERT_EQ(epoll_ctl(ctx.efd[0], EPOLL_CTL_ADD, ctx.efd[1], &e), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
2924
e.events = EPOLLIN | EPOLLET;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
2925
ASSERT_EQ(epoll_ctl(ctx.efd[0], EPOLL_CTL_ADD, ctx.efd[2], &e), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
2931
if (epoll_wait(ctx.efd[0], &e, 1, -1) > 0)
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
2964
struct epoll_event e;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
2981
e.events = EPOLLIN;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
2982
ASSERT_EQ(epoll_ctl(ctx.efd[1], EPOLL_CTL_ADD, ctx.sfd[0], &e), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
2984
e.events = EPOLLIN;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
2985
ASSERT_EQ(epoll_ctl(ctx.efd[2], EPOLL_CTL_ADD, ctx.sfd[2], &e), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
2987
e.events = EPOLLIN;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
2988
ASSERT_EQ(epoll_ctl(ctx.efd[0], EPOLL_CTL_ADD, ctx.efd[1], &e), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
2990
e.events = EPOLLIN;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
2991
ASSERT_EQ(epoll_ctl(ctx.efd[0], EPOLL_CTL_ADD, ctx.efd[2], &e), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
3000
if (epoll_wait(ctx.efd[0], &e, 1, 0) > 0)
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
3034
struct epoll_event e;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
3051
e.events = EPOLLIN;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
3052
ASSERT_EQ(epoll_ctl(ctx.efd[1], EPOLL_CTL_ADD, ctx.sfd[0], &e), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
3054
e.events = EPOLLIN;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
3055
ASSERT_EQ(epoll_ctl(ctx.efd[2], EPOLL_CTL_ADD, ctx.sfd[2], &e), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
3057
e.events = EPOLLIN | EPOLLET;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
3058
ASSERT_EQ(epoll_ctl(ctx.efd[0], EPOLL_CTL_ADD, ctx.efd[1], &e), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
306
struct epoll_event e;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
3060
e.events = EPOLLIN | EPOLLET;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
3061
ASSERT_EQ(epoll_ctl(ctx.efd[0], EPOLL_CTL_ADD, ctx.efd[2], &e), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
3070
if (epoll_wait(ctx.efd[0], &e, 1, 0) > 0)
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
3094
struct epoll_event e;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
3101
e.events = EPOLLIN | EPOLLERR | EPOLLET;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
3102
epoll_ctl(ctx->efd[0], EPOLL_CTL_MOD, ctx->sfd[0], &e);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
3122
struct epoll_event e;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
313
e.events = EPOLLIN;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
3134
e.events = EPOLLIN | EPOLLERR | EPOLLET;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
3135
ASSERT_EQ(epoll_ctl(ctx.efd[0], EPOLL_CTL_ADD, ctx.sfd[0], &e), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
314
ASSERT_EQ(epoll_ctl(efd, EPOLL_CTL_ADD, sfd[0], &e), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
3140
ret = epoll_wait(ctx.efd[0], &e, 1, 1000);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
3170
struct epoll_event e;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
3194
ret = epoll_pwait(ctx->epfd, &e, 1, 2000, &sigmask);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
3202
ret = read(e.data.fd, &v, sizeof(v));
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
321
ASSERT_EQ(epoll_wait(efd, &e, 1, 0), 1);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
3233
struct epoll_event e;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
3246
e.events = EPOLLIN | EPOLLET;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
3247
e.data.fd = ctx.evfd[i];
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
3248
ASSERT_EQ(epoll_ctl(ctx.epfd, EPOLL_CTL_ADD, ctx.evfd[i], &e), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
326
ASSERT_EQ(epoll_wait(efd, &e, 1, 0), 1);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
3400
struct epoll_event e;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
3407
e.events = EPOLLIN;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
3408
ASSERT_EQ(epoll_ctl(efd, EPOLL_CTL_ADD, sfd[0], &e), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
3412
EXPECT_EQ(sys_epoll_pwait2(efd, &e, 1, NULL, NULL, 0), 1);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
3413
EXPECT_EQ(sys_epoll_pwait2(efd, &e, 1, NULL, NULL, 0), 1);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
3428
struct epoll_event e;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
3435
e.events = EPOLLIN;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
3436
ASSERT_EQ(epoll_ctl(efd, EPOLL_CTL_ADD, sfd[0], &e), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
3442
EXPECT_EQ(sys_epoll_pwait2(efd, &e, 1, &ts, NULL, 0), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
345
struct epoll_event e;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
3462
struct epoll_event e;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
3472
e.events = EPOLLIN;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
3473
ASSERT_EQ(epoll_ctl(ctx.efd[0], EPOLL_CTL_ADD, ctx.sfd[0], &e), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
352
e.events = EPOLLIN | EPOLLET;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
353
ASSERT_EQ(epoll_ctl(efd, EPOLL_CTL_ADD, sfd[0], &e), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
360
ASSERT_EQ(epoll_wait(efd, &e, 1, 0), 1);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
365
ASSERT_EQ(epoll_wait(efd, &e, 1, 0), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
475
struct epoll_event e;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
485
e.events = EPOLLIN;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
486
ASSERT_EQ(epoll_ctl(ctx.efd[0], EPOLL_CTL_ADD, ctx.sfd[0], &e), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
492
if (epoll_wait(ctx.efd[0], &e, 1, -1) > 0)
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
518
struct epoll_event e;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
52
struct epoll_event e;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
528
e.events = EPOLLIN | EPOLLET;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
529
ASSERT_EQ(epoll_ctl(ctx.efd[0], EPOLL_CTL_ADD, ctx.sfd[0], &e), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
535
if (epoll_wait(ctx.efd[0], &e, 1, -1) > 0)
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
55
if (epoll_wait(ctx->efd[0], &e, 1, -1) > 0)
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
64
struct epoll_event e;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
659
struct epoll_event e;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
669
e.events = EPOLLIN;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
670
ASSERT_EQ(epoll_ctl(ctx.efd[0], EPOLL_CTL_ADD, ctx.sfd[0], &e), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
676
if (epoll_wait(ctx.efd[0], &e, 1, -1) > 0)
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
70
if (epoll_wait(ctx->efd[0], &e, 1, 0) > 0)
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
702
struct epoll_event e;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
712
e.events = EPOLLIN | EPOLLET;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
713
ASSERT_EQ(epoll_ctl(ctx.efd[0], EPOLL_CTL_ADD, ctx.sfd[0], &e), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
719
if (epoll_wait(ctx.efd[0], &e, 1, -1) > 0)
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
79
struct epoll_event e;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
82
if (epoll_wait(ctx->efd[0], &e, 1, -1) > 0)
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
846
struct epoll_event e;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
856
e.events = EPOLLIN;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
857
ASSERT_EQ(epoll_ctl(efd[1], EPOLL_CTL_ADD, sfd[0], &e), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
859
e.events = EPOLLIN;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
860
ASSERT_EQ(epoll_ctl(efd[0], EPOLL_CTL_ADD, efd[1], &e), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
864
EXPECT_EQ(epoll_wait(efd[0], &e, 1, 0), 1);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
865
EXPECT_EQ(epoll_wait(efd[0], &e, 1, 0), 1);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
886
struct epoll_event e;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
896
e.events = EPOLLIN | EPOLLET;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
897
ASSERT_EQ(epoll_ctl(efd[1], EPOLL_CTL_ADD, sfd[0], &e), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
899
e.events = EPOLLIN;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
900
ASSERT_EQ(epoll_ctl(efd[0], EPOLL_CTL_ADD, efd[1], &e), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
904
EXPECT_EQ(epoll_wait(efd[0], &e, 1, 0), 1);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
905
EXPECT_EQ(epoll_wait(efd[0], &e, 1, 0), 1);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
91
struct epoll_event e;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
926
struct epoll_event e;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
936
e.events = EPOLLIN;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
937
ASSERT_EQ(epoll_ctl(efd[1], EPOLL_CTL_ADD, sfd[0], &e), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
939
e.events = EPOLLIN | EPOLLET;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
940
ASSERT_EQ(epoll_ctl(efd[0], EPOLL_CTL_ADD, efd[1], &e), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
944
EXPECT_EQ(epoll_wait(efd[0], &e, 1, 0), 1);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
945
EXPECT_EQ(epoll_wait(efd[0], &e, 1, 0), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
966
struct epoll_event e;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
97
if (epoll_wait(ctx->efd[0], &e, 1, 0) > 0)
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
976
e.events = EPOLLIN | EPOLLET;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
977
ASSERT_EQ(epoll_ctl(efd[1], EPOLL_CTL_ADD, sfd[0], &e), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
979
e.events = EPOLLIN | EPOLLET;
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
980
ASSERT_EQ(epoll_ctl(efd[0], EPOLL_CTL_ADD, efd[1], &e), 0);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
984
EXPECT_EQ(epoll_wait(efd[0], &e, 1, 0), 1);
tools/testing/selftests/filesystems/epoll/epoll_wakeup_test.c
985
EXPECT_EQ(epoll_wait(efd[0], &e, 1, 0), 0);
tools/testing/selftests/kvm/include/test_util.h
57
#define TEST_ASSERT(e, fmt, ...) \
tools/testing/selftests/kvm/include/test_util.h
58
test_assert((e), #e, __FILE__, __LINE__, fmt, ##__VA_ARGS__)
tools/testing/selftests/kvm/lib/x86/processor.c
697
struct idt_entry *e = &base[vector];
tools/testing/selftests/kvm/lib/x86/processor.c
699
memset(e, 0, sizeof(*e));
tools/testing/selftests/kvm/lib/x86/processor.c
700
e->offset0 = addr;
tools/testing/selftests/kvm/lib/x86/processor.c
701
e->selector = selector;
tools/testing/selftests/kvm/lib/x86/processor.c
702
e->ist = 0;
tools/testing/selftests/kvm/lib/x86/processor.c
703
e->type = 14;
tools/testing/selftests/kvm/lib/x86/processor.c
704
e->dpl = dpl;
tools/testing/selftests/kvm/lib/x86/processor.c
705
e->p = 1;
tools/testing/selftests/kvm/lib/x86/processor.c
706
e->offset1 = addr >> 16;
tools/testing/selftests/kvm/lib/x86/processor.c
707
e->offset2 = addr >> 32;
tools/testing/selftests/kvm/x86/pmu_event_filter_test.c
689
uint64_t e = ~0ul;
tools/testing/selftests/kvm/x86/pmu_event_filter_test.c
696
r = set_pmu_single_event_filter(vcpu, e, 0, KVM_PMU_EVENT_ALLOW);
tools/testing/selftests/kvm/x86/pmu_event_filter_test.c
699
r = set_pmu_single_event_filter(vcpu, e,
tools/testing/selftests/kvm/x86/pmu_event_filter_test.c
704
e = KVM_PMU_ENCODE_MASKED_ENTRY(0xff, 0xff, 0xff, 0xf);
tools/testing/selftests/kvm/x86/pmu_event_filter_test.c
705
r = set_pmu_single_event_filter(vcpu, e,
tools/testing/selftests/kvm/x86/xen_shinfo_test.c
737
struct kvm_irq_routing_xen_evtchn e;
tools/testing/selftests/kvm/x86/xen_shinfo_test.c
738
e.port = EVTCHN_TEST2;
tools/testing/selftests/kvm/x86/xen_shinfo_test.c
739
e.vcpu = vcpu->id;
tools/testing/selftests/kvm/x86/xen_shinfo_test.c
740
e.priority = KVM_IRQ_ROUTING_XEN_EVTCHN_PRIO_2LEVEL;
tools/testing/selftests/kvm/x86/xen_shinfo_test.c
742
vm_ioctl(vm, KVM_XEN_HVM_EVTCHN_SEND, &e);
tools/testing/selftests/net/fin_ack_lat.c
17
static unsigned long timediff(struct timeval s, struct timeval e)
tools/testing/selftests/net/fin_ack_lat.c
22
e_us = e.tv_sec * 1000000 + e.tv_usec;
tools/testing/selftests/net/ipsec.c
2042
char *a, char *e, char *c, char *ae)
tools/testing/selftests/net/ipsec.c
2051
if (e)
tools/testing/selftests/net/ipsec.c
2052
strncpy(desc.e_algo, e, ALGO_LEN - 1);
tools/testing/selftests/net/netfilter/conntrack_reverse_clash.c
30
static void die(const char *e)
tools/testing/selftests/net/netfilter/conntrack_reverse_clash.c
32
perror(e);
tools/testing/selftests/net/tcp_ao/lib/ftrace-tcp.c
138
static bool lookup_expected_event(int event_type, struct trace_point *e)
tools/testing/selftests/net/tcp_ao/lib/ftrace-tcp.c
149
if (p->family != e->family)
tools/testing/selftests/net/tcp_ao/lib/ftrace-tcp.c
155
if (memcmp(&p->src, &e->src, sk_size))
tools/testing/selftests/net/tcp_ao/lib/ftrace-tcp.c
157
if (memcmp(&p->dst, &e->dst, sk_size))
tools/testing/selftests/net/tcp_ao/lib/ftrace-tcp.c
159
if (p->src_port >= 0 && p->src_port != e->src_port)
tools/testing/selftests/net/tcp_ao/lib/ftrace-tcp.c
161
if (p->dst_port >= 0 && p->dst_port != e->dst_port)
tools/testing/selftests/net/tcp_ao/lib/ftrace-tcp.c
163
if (p->L3index >= 0 && p->L3index != e->L3index)
tools/testing/selftests/net/tcp_ao/lib/ftrace-tcp.c
166
if (p->fin >= 0 && p->fin != e->fin)
tools/testing/selftests/net/tcp_ao/lib/ftrace-tcp.c
168
if (p->syn >= 0 && p->syn != e->syn)
tools/testing/selftests/net/tcp_ao/lib/ftrace-tcp.c
170
if (p->rst >= 0 && p->rst != e->rst)
tools/testing/selftests/net/tcp_ao/lib/ftrace-tcp.c
172
if (p->psh >= 0 && p->psh != e->psh)
tools/testing/selftests/net/tcp_ao/lib/ftrace-tcp.c
174
if (p->ack >= 0 && p->ack != e->ack)
tools/testing/selftests/net/tcp_ao/lib/ftrace-tcp.c
177
if (p->keyid >= 0 && p->keyid != e->keyid)
tools/testing/selftests/net/tcp_ao/lib/ftrace-tcp.c
179
if (p->rnext >= 0 && p->rnext != e->rnext)
tools/testing/selftests/net/tcp_ao/lib/ftrace-tcp.c
181
if (p->maclen >= 0 && p->maclen != e->maclen)
tools/testing/selftests/net/tcp_ao/lib/ftrace-tcp.c
183
if (p->sne >= 0 && p->sne != e->sne)
tools/testing/selftests/net/tcp_ao/lib/ftrace-tcp.c
419
static void dump_trace_event(struct expected_trace_point *e)
tools/testing/selftests/net/tcp_ao/lib/ftrace-tcp.c
423
if (!inet_ntop(e->family, &e->src, src, INET6_ADDRSTRLEN))
tools/testing/selftests/net/tcp_ao/lib/ftrace-tcp.c
425
if (!inet_ntop(e->family, &e->dst, dst, INET6_ADDRSTRLEN))
tools/testing/selftests/net/tcp_ao/lib/ftrace-tcp.c
428
trace_event_names[e->type],
tools/testing/selftests/net/tcp_ao/lib/ftrace-tcp.c
429
src, e->src_port, dst, e->dst_port, e->L3index,
tools/testing/selftests/net/tcp_ao/lib/ftrace-tcp.c
430
e->fin ? "F" : "", e->syn ? "S" : "", e->rst ? "R" : "",
tools/testing/selftests/net/tcp_ao/lib/ftrace-tcp.c
431
e->psh ? "P" : "", e->ack ? "." : "",
tools/testing/selftests/net/tcp_ao/lib/ftrace-tcp.c
432
e->keyid, e->rnext, e->maclen, e->sne, e->matched);
tools/testing/selftests/net/tcp_ao/lib/ftrace-tcp.c
443
struct expected_trace_point *e = &exp_tps[i];
tools/testing/selftests/net/tcp_ao/lib/ftrace-tcp.c
445
total_matched += e->matched;
tools/testing/selftests/net/tcp_ao/lib/ftrace-tcp.c
446
matches_per_type[e->type] += e->matched;
tools/testing/selftests/net/tcp_ao/lib/ftrace-tcp.c
447
if (!e->matched)
tools/testing/selftests/openat2/helpers.h
18
#define BUILD_BUG_ON(e) ((void)(sizeof(struct { int:(-!!(e)); })))
tools/testing/selftests/powerpc/nx-gzip/include/nxu.h
584
char *nx_strerror(int e);
tools/testing/selftests/powerpc/pmu/count_instructions.c
19
static void setup_event(struct event *e, u64 config, char *name)
tools/testing/selftests/powerpc/pmu/count_instructions.c
21
event_init_opts(e, config, PERF_TYPE_HARDWARE, name);
tools/testing/selftests/powerpc/pmu/count_instructions.c
23
e->attr.disabled = 1;
tools/testing/selftests/powerpc/pmu/count_instructions.c
24
e->attr.exclude_kernel = 1;
tools/testing/selftests/powerpc/pmu/count_instructions.c
25
e->attr.exclude_hv = 1;
tools/testing/selftests/powerpc/pmu/count_instructions.c
26
e->attr.exclude_idle = 1;
tools/testing/selftests/powerpc/pmu/count_stcx_fail.c
19
static void setup_event(struct event *e, u64 config, int type, char *name)
tools/testing/selftests/powerpc/pmu/count_stcx_fail.c
21
event_init_opts(e, config, type, name);
tools/testing/selftests/powerpc/pmu/count_stcx_fail.c
23
e->attr.disabled = 1;
tools/testing/selftests/powerpc/pmu/count_stcx_fail.c
24
e->attr.exclude_kernel = 1;
tools/testing/selftests/powerpc/pmu/count_stcx_fail.c
25
e->attr.exclude_hv = 1;
tools/testing/selftests/powerpc/pmu/count_stcx_fail.c
26
e->attr.exclude_idle = 1;
tools/testing/selftests/powerpc/pmu/ebb/ebb.c
275
int ebb_event_enable(struct event *e)
tools/testing/selftests/powerpc/pmu/ebb/ebb.c
282
rc = ioctl(e->fd, PERF_EVENT_IOC_ENABLE);
tools/testing/selftests/powerpc/pmu/ebb/ebb.c
286
rc = event_read(e);
tools/testing/selftests/powerpc/pmu/ebb/ebb.c
331
void event_ebb_init(struct event *e)
tools/testing/selftests/powerpc/pmu/ebb/ebb.c
333
e->attr.config |= (1ull << 63);
tools/testing/selftests/powerpc/pmu/ebb/ebb.c
336
void event_bhrb_init(struct event *e, unsigned ifm)
tools/testing/selftests/powerpc/pmu/ebb/ebb.c
338
e->attr.config |= (1ull << 62) | ((u64)ifm << 60);
tools/testing/selftests/powerpc/pmu/ebb/ebb.c
341
void event_leader_ebb_init(struct event *e)
tools/testing/selftests/powerpc/pmu/ebb/ebb.c
343
event_ebb_init(e);
tools/testing/selftests/powerpc/pmu/ebb/ebb.c
345
e->attr.exclusive = 1;
tools/testing/selftests/powerpc/pmu/ebb/ebb.c
346
e->attr.pinned = 1;
tools/testing/selftests/powerpc/pmu/ebb/ebb.h
47
void event_leader_ebb_init(struct event *e);
tools/testing/selftests/powerpc/pmu/ebb/ebb.h
48
void event_ebb_init(struct event *e);
tools/testing/selftests/powerpc/pmu/ebb/ebb.h
49
void event_bhrb_init(struct event *e, unsigned ifm);
tools/testing/selftests/powerpc/pmu/ebb/ebb.h
52
int ebb_event_enable(struct event *e);
tools/testing/selftests/powerpc/pmu/ebb/trace.c
106
struct trace_entry *e;
tools/testing/selftests/powerpc/pmu/ebb/trace.c
109
e = trace_alloc_entry(tb, sizeof(value));
tools/testing/selftests/powerpc/pmu/ebb/trace.c
110
if (!e)
tools/testing/selftests/powerpc/pmu/ebb/trace.c
113
e->type = TRACE_TYPE_COUNTER;
tools/testing/selftests/powerpc/pmu/ebb/trace.c
114
p = (u64 *)e->data;
tools/testing/selftests/powerpc/pmu/ebb/trace.c
122
struct trace_entry *e;
tools/testing/selftests/powerpc/pmu/ebb/trace.c
129
e = trace_alloc_entry(tb, len + 1);
tools/testing/selftests/powerpc/pmu/ebb/trace.c
130
if (!e)
tools/testing/selftests/powerpc/pmu/ebb/trace.c
133
e->type = TRACE_TYPE_STRING;
tools/testing/selftests/powerpc/pmu/ebb/trace.c
134
p = (char *)e->data;
tools/testing/selftests/powerpc/pmu/ebb/trace.c
144
struct trace_entry *e;
tools/testing/selftests/powerpc/pmu/ebb/trace.c
146
e = trace_alloc_entry(tb, 0);
tools/testing/selftests/powerpc/pmu/ebb/trace.c
147
if (!e)
tools/testing/selftests/powerpc/pmu/ebb/trace.c
150
e->type = TRACE_TYPE_INDENT;
tools/testing/selftests/powerpc/pmu/ebb/trace.c
157
struct trace_entry *e;
tools/testing/selftests/powerpc/pmu/ebb/trace.c
159
e = trace_alloc_entry(tb, 0);
tools/testing/selftests/powerpc/pmu/ebb/trace.c
160
if (!e)
tools/testing/selftests/powerpc/pmu/ebb/trace.c
163
e->type = TRACE_TYPE_OUTDENT;
tools/testing/selftests/powerpc/pmu/ebb/trace.c
200
static void trace_print_reg(struct trace_entry *e)
tools/testing/selftests/powerpc/pmu/ebb/trace.c
205
p = (u64 *)e->data;
tools/testing/selftests/powerpc/pmu/ebb/trace.c
216
static void trace_print_counter(struct trace_entry *e)
tools/testing/selftests/powerpc/pmu/ebb/trace.c
220
value = (u64 *)e->data;
tools/testing/selftests/powerpc/pmu/ebb/trace.c
224
static void trace_print_string(struct trace_entry *e)
tools/testing/selftests/powerpc/pmu/ebb/trace.c
228
str = (char *)e->data;
tools/testing/selftests/powerpc/pmu/ebb/trace.c
235
static void trace_print_entry(struct trace_entry *e, int seq, int *prefix)
tools/testing/selftests/powerpc/pmu/ebb/trace.c
237
switch (e->type) {
tools/testing/selftests/powerpc/pmu/ebb/trace.c
240
trace_print_reg(e);
tools/testing/selftests/powerpc/pmu/ebb/trace.c
244
trace_print_counter(e);
tools/testing/selftests/powerpc/pmu/ebb/trace.c
248
trace_print_string(e);
tools/testing/selftests/powerpc/pmu/ebb/trace.c
264
printf("entry @ %p type %d\n", e, e->type);
tools/testing/selftests/powerpc/pmu/ebb/trace.c
271
struct trace_entry *e;
tools/testing/selftests/powerpc/pmu/ebb/trace.c
288
e = p;
tools/testing/selftests/powerpc/pmu/ebb/trace.c
290
trace_print_entry(e, i, &prefix);
tools/testing/selftests/powerpc/pmu/ebb/trace.c
293
p = (void *)e + sizeof(*e) + e->length;
tools/testing/selftests/powerpc/pmu/ebb/trace.c
78
struct trace_entry *e;
tools/testing/selftests/powerpc/pmu/ebb/trace.c
80
e = trace_alloc(tb, sizeof(*e) + payload_size);
tools/testing/selftests/powerpc/pmu/ebb/trace.c
81
if (e)
tools/testing/selftests/powerpc/pmu/ebb/trace.c
82
e->length = payload_size;
tools/testing/selftests/powerpc/pmu/ebb/trace.c
84
return e;
tools/testing/selftests/powerpc/pmu/ebb/trace.c
89
struct trace_entry *e;
tools/testing/selftests/powerpc/pmu/ebb/trace.c
92
e = trace_alloc_entry(tb, sizeof(reg) + sizeof(value));
tools/testing/selftests/powerpc/pmu/ebb/trace.c
93
if (!e)
tools/testing/selftests/powerpc/pmu/ebb/trace.c
96
e->type = TRACE_TYPE_REG;
tools/testing/selftests/powerpc/pmu/ebb/trace.c
97
p = (u64 *)e->data;
tools/testing/selftests/powerpc/pmu/event.c
100
void event_close(struct event *e)
tools/testing/selftests/powerpc/pmu/event.c
102
close(e->fd);
tools/testing/selftests/powerpc/pmu/event.c
105
int event_enable(struct event *e)
tools/testing/selftests/powerpc/pmu/event.c
107
return ioctl(e->fd, PERF_EVENT_IOC_ENABLE);
tools/testing/selftests/powerpc/pmu/event.c
110
int event_disable(struct event *e)
tools/testing/selftests/powerpc/pmu/event.c
112
return ioctl(e->fd, PERF_EVENT_IOC_DISABLE);
tools/testing/selftests/powerpc/pmu/event.c
115
int event_reset(struct event *e)
tools/testing/selftests/powerpc/pmu/event.c
117
return ioctl(e->fd, PERF_EVENT_IOC_RESET);
tools/testing/selftests/powerpc/pmu/event.c
120
int event_read(struct event *e)
tools/testing/selftests/powerpc/pmu/event.c
124
rc = read(e->fd, &e->result, sizeof(e->result));
tools/testing/selftests/powerpc/pmu/event.c
125
if (rc != sizeof(e->result)) {
tools/testing/selftests/powerpc/pmu/event.c
126
fprintf(stderr, "read error on event %p!\n", e);
tools/testing/selftests/powerpc/pmu/event.c
133
void event_report_justified(struct event *e, int name_width, int result_width)
tools/testing/selftests/powerpc/pmu/event.c
135
printf("%*s: result %*llu ", name_width, e->name, result_width,
tools/testing/selftests/powerpc/pmu/event.c
136
e->result.value);
tools/testing/selftests/powerpc/pmu/event.c
138
if (e->result.running == e->result.enabled)
tools/testing/selftests/powerpc/pmu/event.c
139
printf("running/enabled %llu\n", e->result.running);
tools/testing/selftests/powerpc/pmu/event.c
141
printf("running %llu enabled %llu\n", e->result.running,
tools/testing/selftests/powerpc/pmu/event.c
142
e->result.enabled);
tools/testing/selftests/powerpc/pmu/event.c
145
void event_report(struct event *e)
tools/testing/selftests/powerpc/pmu/event.c
147
event_report_justified(e, 0, 0);
tools/testing/selftests/powerpc/pmu/event.c
24
static void __event_init_opts(struct event *e, u64 config,
tools/testing/selftests/powerpc/pmu/event.c
27
memset(e, 0, sizeof(*e));
tools/testing/selftests/powerpc/pmu/event.c
29
e->name = name;
tools/testing/selftests/powerpc/pmu/event.c
31
e->attr.type = type;
tools/testing/selftests/powerpc/pmu/event.c
32
e->attr.config = config;
tools/testing/selftests/powerpc/pmu/event.c
33
e->attr.size = sizeof(e->attr);
tools/testing/selftests/powerpc/pmu/event.c
35
e->attr.read_format = PERF_FORMAT_TOTAL_TIME_ENABLED | \
tools/testing/selftests/powerpc/pmu/event.c
38
e->attr.sample_period = 1000;
tools/testing/selftests/powerpc/pmu/event.c
39
e->attr.sample_type = PERF_SAMPLE_REGS_INTR;
tools/testing/selftests/powerpc/pmu/event.c
40
e->attr.disabled = 1;
tools/testing/selftests/powerpc/pmu/event.c
44
void event_init_opts(struct event *e, u64 config, int type, char *name)
tools/testing/selftests/powerpc/pmu/event.c
46
__event_init_opts(e, config, type, name, false);
tools/testing/selftests/powerpc/pmu/event.c
49
void event_init_named(struct event *e, u64 config, char *name)
tools/testing/selftests/powerpc/pmu/event.c
51
event_init_opts(e, config, PERF_TYPE_RAW, name);
tools/testing/selftests/powerpc/pmu/event.c
54
void event_init(struct event *e, u64 config)
tools/testing/selftests/powerpc/pmu/event.c
56
event_init_opts(e, config, PERF_TYPE_RAW, "event");
tools/testing/selftests/powerpc/pmu/event.c
59
void event_init_sampling(struct event *e, u64 config)
tools/testing/selftests/powerpc/pmu/event.c
61
__event_init_opts(e, config, PERF_TYPE_RAW, "event", true);
tools/testing/selftests/powerpc/pmu/event.c
69
int event_open_with_options(struct event *e, pid_t pid, int cpu, int group_fd)
tools/testing/selftests/powerpc/pmu/event.c
71
e->fd = perf_event_open(&e->attr, pid, cpu, group_fd, 0);
tools/testing/selftests/powerpc/pmu/event.c
72
if (e->fd == -1) {
tools/testing/selftests/powerpc/pmu/event.c
80
int event_open_with_group(struct event *e, int group_fd)
tools/testing/selftests/powerpc/pmu/event.c
82
return event_open_with_options(e, PERF_CURRENT_PID, PERF_NO_CPU, group_fd);
tools/testing/selftests/powerpc/pmu/event.c
85
int event_open_with_pid(struct event *e, pid_t pid)
tools/testing/selftests/powerpc/pmu/event.c
87
return event_open_with_options(e, pid, PERF_NO_CPU, PERF_NO_GROUP);
tools/testing/selftests/powerpc/pmu/event.c
90
int event_open_with_cpu(struct event *e, int cpu)
tools/testing/selftests/powerpc/pmu/event.c
92
return event_open_with_options(e, PERF_NO_PID, cpu, PERF_NO_GROUP);
tools/testing/selftests/powerpc/pmu/event.c
95
int event_open(struct event *e)
tools/testing/selftests/powerpc/pmu/event.c
97
return event_open_with_options(e, PERF_CURRENT_PID, PERF_NO_CPU, PERF_NO_GROUP);
tools/testing/selftests/powerpc/pmu/event.h
32
void event_init(struct event *e, u64 config);
tools/testing/selftests/powerpc/pmu/event.h
33
void event_init_named(struct event *e, u64 config, char *name);
tools/testing/selftests/powerpc/pmu/event.h
34
void event_init_opts(struct event *e, u64 config, int type, char *name);
tools/testing/selftests/powerpc/pmu/event.h
35
void event_init_sampling(struct event *e, u64 config);
tools/testing/selftests/powerpc/pmu/event.h
36
int event_open_with_options(struct event *e, pid_t pid, int cpu, int group_fd);
tools/testing/selftests/powerpc/pmu/event.h
37
int event_open_with_group(struct event *e, int group_fd);
tools/testing/selftests/powerpc/pmu/event.h
38
int event_open_with_pid(struct event *e, pid_t pid);
tools/testing/selftests/powerpc/pmu/event.h
39
int event_open_with_cpu(struct event *e, int cpu);
tools/testing/selftests/powerpc/pmu/event.h
40
int event_open(struct event *e);
tools/testing/selftests/powerpc/pmu/event.h
41
void event_close(struct event *e);
tools/testing/selftests/powerpc/pmu/event.h
42
int event_enable(struct event *e);
tools/testing/selftests/powerpc/pmu/event.h
43
int event_disable(struct event *e);
tools/testing/selftests/powerpc/pmu/event.h
44
int event_reset(struct event *e);
tools/testing/selftests/powerpc/pmu/event.h
45
int event_read(struct event *e);
tools/testing/selftests/powerpc/pmu/event.h
46
void event_report_justified(struct event *e, int name_width, int result_width);
tools/testing/selftests/powerpc/pmu/event.h
47
void event_report(struct event *e);
tools/testing/selftests/powerpc/pmu/event_code_tests/event_alternatives_tests_p10.c
27
struct event *e, events[5];
tools/testing/selftests/powerpc/pmu/event_code_tests/event_alternatives_tests_p10.c
48
e = &events[0];
tools/testing/selftests/powerpc/pmu/event_code_tests/event_alternatives_tests_p10.c
49
event_init(e, 0x0001e);
tools/testing/selftests/powerpc/pmu/event_code_tests/event_alternatives_tests_p10.c
51
e = &events[1];
tools/testing/selftests/powerpc/pmu/event_code_tests/event_alternatives_tests_p10.c
52
event_init(e, EventCode_1);
tools/testing/selftests/powerpc/pmu/event_code_tests/event_alternatives_tests_p10.c
54
e = &events[2];
tools/testing/selftests/powerpc/pmu/event_code_tests/event_alternatives_tests_p10.c
55
event_init(e, EventCode_2);
tools/testing/selftests/powerpc/pmu/event_code_tests/event_alternatives_tests_p10.c
57
e = &events[3];
tools/testing/selftests/powerpc/pmu/event_code_tests/event_alternatives_tests_p10.c
58
event_init(e, EventCode_3);
tools/testing/selftests/powerpc/pmu/event_code_tests/event_alternatives_tests_p10.c
60
e = &events[4];
tools/testing/selftests/powerpc/pmu/event_code_tests/event_alternatives_tests_p10.c
61
event_init(e, EventCode_4);
tools/testing/selftests/powerpc/pmu/event_code_tests/event_alternatives_tests_p10.c
76
e = &events[0];
tools/testing/selftests/powerpc/pmu/event_code_tests/event_alternatives_tests_p10.c
77
event_init(e, 0x00002);
tools/testing/selftests/powerpc/pmu/event_code_tests/event_alternatives_tests_p10.c
79
e = &events[1];
tools/testing/selftests/powerpc/pmu/event_code_tests/event_alternatives_tests_p10.c
80
event_init(e, EventCode_1);
tools/testing/selftests/powerpc/pmu/event_code_tests/event_alternatives_tests_p10.c
82
e = &events[2];
tools/testing/selftests/powerpc/pmu/event_code_tests/event_alternatives_tests_p10.c
83
event_init(e, EventCode_2);
tools/testing/selftests/powerpc/pmu/event_code_tests/event_alternatives_tests_p10.c
85
e = &events[3];
tools/testing/selftests/powerpc/pmu/event_code_tests/event_alternatives_tests_p10.c
86
event_init(e, EventCode_3);
tools/testing/selftests/powerpc/pmu/event_code_tests/event_alternatives_tests_p10.c
88
e = &events[4];
tools/testing/selftests/powerpc/pmu/event_code_tests/event_alternatives_tests_p10.c
89
event_init(e, EventCode_4);
tools/testing/selftests/powerpc/pmu/event_code_tests/group_constraint_pmc_count_test.c
21
struct event *e, events[5];
tools/testing/selftests/powerpc/pmu/event_code_tests/group_constraint_pmc_count_test.c
33
e = &events[0];
tools/testing/selftests/powerpc/pmu/event_code_tests/group_constraint_pmc_count_test.c
34
event_init(e, 0x1001a);
tools/testing/selftests/powerpc/pmu/event_code_tests/group_constraint_pmc_count_test.c
36
e = &events[1];
tools/testing/selftests/powerpc/pmu/event_code_tests/group_constraint_pmc_count_test.c
37
event_init(e, 0x200fc);
tools/testing/selftests/powerpc/pmu/event_code_tests/group_constraint_pmc_count_test.c
39
e = &events[2];
tools/testing/selftests/powerpc/pmu/event_code_tests/group_constraint_pmc_count_test.c
40
event_init(e, 0x30080);
tools/testing/selftests/powerpc/pmu/event_code_tests/group_constraint_pmc_count_test.c
42
e = &events[3];
tools/testing/selftests/powerpc/pmu/event_code_tests/group_constraint_pmc_count_test.c
43
event_init(e, 0x40054);
tools/testing/selftests/powerpc/pmu/event_code_tests/group_constraint_pmc_count_test.c
45
e = &events[4];
tools/testing/selftests/powerpc/pmu/event_code_tests/group_constraint_pmc_count_test.c
46
event_init(e, 0x0002c);
tools/testing/selftests/powerpc/pmu/event_code_tests/group_constraint_unit_test.c
29
struct event *e, events[3];
tools/testing/selftests/powerpc/pmu/event_code_tests/group_constraint_unit_test.c
41
e = &events[0];
tools/testing/selftests/powerpc/pmu/event_code_tests/group_constraint_unit_test.c
42
event_init(e, EventCode_1);
tools/testing/selftests/powerpc/pmu/event_code_tests/group_constraint_unit_test.c
48
e = &events[1];
tools/testing/selftests/powerpc/pmu/event_code_tests/group_constraint_unit_test.c
49
event_init(e, EventCode_2);
tools/testing/selftests/powerpc/pmu/event_code_tests/group_constraint_unit_test.c
55
e = &events[2];
tools/testing/selftests/powerpc/pmu/event_code_tests/group_constraint_unit_test.c
56
event_init(e, EventCode_3);
tools/testing/selftests/powerpc/pmu/event_code_tests/group_pmc56_exclude_constraints_test.c
22
struct event *e, events[3];
tools/testing/selftests/powerpc/pmu/event_code_tests/group_pmc56_exclude_constraints_test.c
35
e = &events[0];
tools/testing/selftests/powerpc/pmu/event_code_tests/group_pmc56_exclude_constraints_test.c
36
event_init(e, 0x500fa);
tools/testing/selftests/powerpc/pmu/event_code_tests/group_pmc56_exclude_constraints_test.c
38
e = &events[1];
tools/testing/selftests/powerpc/pmu/event_code_tests/group_pmc56_exclude_constraints_test.c
39
event_init(e, 0x600f4);
tools/testing/selftests/powerpc/pmu/event_code_tests/group_pmc56_exclude_constraints_test.c
41
e = &events[2];
tools/testing/selftests/powerpc/pmu/event_code_tests/group_pmc56_exclude_constraints_test.c
42
event_init(e, 0x22C040);
tools/testing/selftests/powerpc/pmu/per_event_excludes.c
25
struct event *e, events[4];
tools/testing/selftests/powerpc/pmu/per_event_excludes.c
34
e = &events[0];
tools/testing/selftests/powerpc/pmu/per_event_excludes.c
35
event_init_opts(e, PERF_COUNT_HW_INSTRUCTIONS,
tools/testing/selftests/powerpc/pmu/per_event_excludes.c
37
e->attr.disabled = 1;
tools/testing/selftests/powerpc/pmu/per_event_excludes.c
39
e = &events[1];
tools/testing/selftests/powerpc/pmu/per_event_excludes.c
40
event_init_opts(e, PERF_COUNT_HW_INSTRUCTIONS,
tools/testing/selftests/powerpc/pmu/per_event_excludes.c
42
e->attr.disabled = 1;
tools/testing/selftests/powerpc/pmu/per_event_excludes.c
43
e->attr.exclude_user = 1;
tools/testing/selftests/powerpc/pmu/per_event_excludes.c
44
e->attr.exclude_hv = 1;
tools/testing/selftests/powerpc/pmu/per_event_excludes.c
46
e = &events[2];
tools/testing/selftests/powerpc/pmu/per_event_excludes.c
47
event_init_opts(e, PERF_COUNT_HW_INSTRUCTIONS,
tools/testing/selftests/powerpc/pmu/per_event_excludes.c
49
e->attr.disabled = 1;
tools/testing/selftests/powerpc/pmu/per_event_excludes.c
50
e->attr.exclude_user = 1;
tools/testing/selftests/powerpc/pmu/per_event_excludes.c
51
e->attr.exclude_kernel = 1;
tools/testing/selftests/powerpc/pmu/per_event_excludes.c
53
e = &events[3];
tools/testing/selftests/powerpc/pmu/per_event_excludes.c
54
event_init_opts(e, PERF_COUNT_HW_INSTRUCTIONS,
tools/testing/selftests/powerpc/pmu/per_event_excludes.c
56
e->attr.disabled = 1;
tools/testing/selftests/powerpc/pmu/per_event_excludes.c
57
e->attr.exclude_hv = 1;
tools/testing/selftests/powerpc/pmu/per_event_excludes.c
58
e->attr.exclude_kernel = 1;
tools/testing/selftests/powerpc/security/spectre_v2.c
64
static void setup_event(struct event *e, u64 config, char *name)
tools/testing/selftests/powerpc/security/spectre_v2.c
66
event_init_named(e, config, name);
tools/testing/selftests/powerpc/security/spectre_v2.c
68
e->attr.disabled = 1;
tools/testing/selftests/powerpc/security/spectre_v2.c
69
e->attr.exclude_kernel = 1;
tools/testing/selftests/powerpc/security/spectre_v2.c
70
e->attr.exclude_hv = 1;
tools/testing/selftests/powerpc/security/spectre_v2.c
71
e->attr.exclude_idle = 1;
tools/testing/selftests/vfio/lib/drivers/dsa/registers.h
657
struct __evl_entry e;
tools/testing/selftests/vfio/lib/drivers/dsa/registers.h
662
struct __evl_entry e;
tools/tracing/latency/latency-collector.c
824
const struct entry *e)
tools/tracing/latency/latency-collector.c
826
q->entries[q->next_prod_idx] = *e;
tools/tracing/latency/latency-collector.c
832
struct entry e = q->entries[q->next_cons_idx];
tools/tracing/latency/latency-collector.c
835
return e;
tools/tracing/latency/latency-collector.c
849
const struct entry *e)
tools/tracing/latency/latency-collector.c
855
queue_push_to_back(q, e);
tools/tracing/latency/latency-collector.c
865
struct entry e;
tools/tracing/latency/latency-collector.c
870
e = queue_pop_from_front(q);
tools/tracing/latency/latency-collector.c
877
return e;
tools/tracing/rtla/src/timerlat_top.c
370
int e = 0;
tools/tracing/rtla/src/timerlat_top.c
379
e++;
tools/tracing/rtla/src/timerlat_top.c
388
trace_seq_printf(s, "ALL #%-6llu e%d |", count, e);
tools/usb/ffs-aio-example/multibuff/device_app/aio_multibuff.c
359
struct io_event e[BUFS_MAX];
tools/usb/ffs-aio-example/multibuff/device_app/aio_multibuff.c
361
ret = io_getevents(ctx, 1, BUFS_MAX, e, NULL);
tools/usb/ffs-aio-example/simple/device_app/aio_simple.c
327
struct io_event e[2];
tools/usb/ffs-aio-example/simple/device_app/aio_simple.c
329
ret = io_getevents(ctx, 1, 2, e, NULL);
tools/usb/ffs-aio-example/simple/device_app/aio_simple.c
332
if (e[i].obj->aio_fildes == ep[0]) {
tools/usb/ffs-aio-example/simple/device_app/aio_simple.c
333
printf("ev=in; ret=%lu\n", e[i].res);
tools/usb/ffs-aio-example/simple/device_app/aio_simple.c
335
} else if (e[i].obj->aio_fildes == ep[1]) {
tools/usb/ffs-aio-example/simple/device_app/aio_simple.c
336
printf("ev=out; ret=%lu\n", e[i].res);
tools/virtio/vhost_net_test.c
64
int fd, e;
tools/virtio/vhost_net_test.c
77
e = ioctl(fd, TUNSETIFF, &ifr);
tools/virtio/vhost_net_test.c
78
if (e < 0) {
tools/virtio/vhost_net_test.c
81
return e;
tools/virtio/vhost_net_test.c
84
e = ioctl(fd, TUNSETVNETHDRSZ, &len);
tools/virtio/vhost_net_test.c
85
if (e < 0) {
tools/virtio/vhost_net_test.c
88
return e;
tools/virtio/vhost_net_test.c
91
e = ioctl(fd, SIOCGIFHWADDR, &ifr);
tools/virtio/vhost_net_test.c
92
if (e < 0) {
tools/virtio/vhost_net_test.c
95
return e;
virt/kvm/eventfd.c
274
struct kvm_kernel_irq_routing_entry *e;
virt/kvm/eventfd.c
284
e = entries;
virt/kvm/eventfd.c
286
irqfd->irq_entry = *e;
virt/kvm/irqchip.c
107
struct kvm_kernel_irq_routing_entry *e;
virt/kvm/irqchip.c
110
hlist_for_each_entry_safe(e, n, &rt->map[i], link) {
virt/kvm/irqchip.c
111
hlist_del(&e->link);
virt/kvm/irqchip.c
112
kfree(e);
virt/kvm/irqchip.c
129
struct kvm_kernel_irq_routing_entry *e,
virt/kvm/irqchip.c
146
e->gsi = gsi;
virt/kvm/irqchip.c
147
e->type = ue->type;
virt/kvm/irqchip.c
148
r = kvm_set_routing_entry(kvm, e, ue);
virt/kvm/irqchip.c
151
if (e->type == KVM_IRQ_ROUTING_IRQCHIP)
virt/kvm/irqchip.c
152
rt->chip[e->irqchip.irqchip][e->irqchip.pin] = e->gsi;
virt/kvm/irqchip.c
154
hlist_add_head(&e->link, &rt->map[e->gsi]);
virt/kvm/irqchip.c
174
struct kvm_kernel_irq_routing_entry *e;
virt/kvm/irqchip.c
197
e = kzalloc_obj(*e, GFP_KERNEL_ACCOUNT);
virt/kvm/irqchip.c
198
if (!e)
virt/kvm/irqchip.c
212
r = setup_routing_entry(kvm, new, e, ue);
virt/kvm/irqchip.c
232
kfree(e);
virt/kvm/irqchip.c
25
struct kvm_kernel_irq_routing_entry *e;
virt/kvm/irqchip.c
31
hlist_for_each_entry(e, &irq_rt->map[gsi], link) {
virt/kvm/irqchip.c
32
entries[n] = *e;