arch/alpha/boot/stdio.c
228
long * ip = va_arg(args, long *);
arch/alpha/boot/stdio.c
229
*ip = (str - buf);
arch/alpha/boot/stdio.c
231
size_t * ip = va_arg(args, size_t *);
arch/alpha/boot/stdio.c
232
*ip = (str - buf);
arch/alpha/boot/stdio.c
234
int * ip = va_arg(args, int *);
arch/alpha/boot/stdio.c
235
*ip = (str - buf);
arch/alpha/kernel/smc37c669.c
2332
struct INODE *ip;
arch/alpha/kernel/smc37c669.c
2334
allocinode( smc_ddb.name, 1, &ip );
arch/alpha/kernel/smc37c669.c
2335
ip->dva = &smc_ddb;
arch/alpha/kernel/smc37c669.c
2336
ip->attr = ATTR$M_WRITE | ATTR$M_READ;
arch/alpha/kernel/smc37c669.c
2337
ip->len[0] = 0x30;
arch/alpha/kernel/smc37c669.c
2338
ip->misc = 0;
arch/alpha/kernel/smc37c669.c
2339
INODE_UNLOCK( ip );
arch/alpha/kernel/smc37c669.c
2346
struct INODE *ip;
arch/alpha/kernel/smc37c669.c
2351
ip = fp->ip;
arch/alpha/kernel/smc37c669.c
2352
INODE_LOCK( ip );
arch/alpha/kernel/smc37c669.c
2354
if ( ip->misc ) {
arch/alpha/kernel/smc37c669.c
2355
INODE_UNLOCK( ip );
arch/alpha/kernel/smc37c669.c
2358
ip->misc++;
arch/alpha/kernel/smc37c669.c
2364
INODE_UNLOCK( ip );
arch/alpha/kernel/smc37c669.c
2371
struct INODE *ip;
arch/alpha/kernel/smc37c669.c
2373
ip = fp->ip;
arch/alpha/kernel/smc37c669.c
2375
INODE_LOCK( ip );
arch/alpha/kernel/smc37c669.c
2376
ip->misc--;
arch/alpha/kernel/smc37c669.c
2377
INODE_UNLOCK( ip );
arch/alpha/kernel/smc37c669.c
2387
struct INODE *ip;
arch/alpha/kernel/smc37c669.c
2392
ip = fp->ip;
arch/alpha/kernel/smc37c669.c
2398
if ( !inrange( *fp->offset, 0, ip->len[0] ) )
arch/alpha/kernel/smc37c669.c
2413
struct INODE *ip;
arch/alpha/kernel/smc37c669.c
2417
ip = fp->ip;
arch/alpha/kernel/smc37c669.c
2423
if ( !inrange( *fp->offset, 0, ip->len[0] ) )
arch/arc/kernel/kgdb.c
188
void kgdb_arch_set_pc(struct pt_regs *regs, unsigned long ip)
arch/arc/kernel/kgdb.c
190
instruction_pointer(regs) = ip;
arch/arm/common/sa1111.c
289
u32 ip, mask = sa1111_irqmask(d);
arch/arm/common/sa1111.c
292
ip = readl_relaxed(mapbase + SA1111_INTPOL0);
arch/arm/common/sa1111.c
294
writel_relaxed(ip ^ mask, mapbase + SA1111_INTPOL0);
arch/arm/common/sa1111.c
295
writel_relaxed(ip, mapbase + SA1111_INTPOL0);
arch/arm/common/sa1111.c
313
u32 ip, mask = sa1111_irqmask(d);
arch/arm/common/sa1111.c
321
ip = readl_relaxed(mapbase + SA1111_INTPOL0);
arch/arm/common/sa1111.c
323
ip &= ~mask;
arch/arm/common/sa1111.c
325
ip |= mask;
arch/arm/common/sa1111.c
326
writel_relaxed(ip, mapbase + SA1111_INTPOL0);
arch/arm/common/sa1111.c
327
writel_relaxed(ip, mapbase + SA1111_WAKEPOL0);
arch/arm/include/asm/assembler.h
128
stmdb sp!, {r0-r3, ip, lr}
arch/arm/include/asm/assembler.h
132
ldmia sp!, {r0-r3, ip, lr}
arch/arm/include/asm/assembler.h
144
stmdb sp!, {r0-r3, ip, lr}
arch/arm/include/asm/assembler.h
148
ldmia sp!, {r0-r3, ip, lr}
arch/arm/kernel/ftrace.c
148
unsigned long ip = rec->ip;
arch/arm/kernel/ftrace.c
158
new = ftrace_call_replace(ip, aaddr, !mod);
arch/arm/kernel/ftrace.c
161
aaddr = get_module_plt(mod, ip, aaddr);
arch/arm/kernel/ftrace.c
162
new = ftrace_call_replace(ip, aaddr, true);
arch/arm/kernel/ftrace.c
166
return ftrace_modify_code(rec->ip, old, new, true);
arch/arm/kernel/ftrace.c
175
unsigned long ip = rec->ip;
arch/arm/kernel/ftrace.c
177
old = ftrace_call_replace(ip, adjust_address(rec, old_addr), true);
arch/arm/kernel/ftrace.c
179
new = ftrace_call_replace(ip, adjust_address(rec, addr), true);
arch/arm/kernel/ftrace.c
181
return ftrace_modify_code(rec->ip, old, new, true);
arch/arm/kernel/ftrace.c
190
unsigned long ip = rec->ip;
arch/arm/kernel/ftrace.c
203
old = ftrace_call_replace(ip, aaddr,
arch/arm/kernel/ftrace.c
207
aaddr = get_module_plt(mod, ip, aaddr);
arch/arm/kernel/ftrace.c
208
old = ftrace_call_replace(ip, aaddr, true);
arch/arm/kernel/ftrace.c
219
ret = ftrace_modify_code(ip, old, new, !is_kernel_inittext(ip));
arch/arm/kernel/ftrace.c
74
likely(!is_kernel_inittext(rec->ip)))
arch/arm/kernel/ptrace.c
81
REG_OFFSET_NAME(ip),
arch/arm/kernel/ptrace.c
832
unsigned long ip;
arch/arm/kernel/ptrace.c
838
ip = regs->ARM_ip;
arch/arm/kernel/ptrace.c
846
regs->ARM_ip = ip;
arch/arm/lib/bitops.h
10
strbne r1, [ip] @ assert word-aligned
arch/arm/lib/bitops.h
100
strbne r1, [ip] @ assert word-aligned
arch/arm/lib/bitops.h
103
save_and_disable_irqs ip
arch/arm/lib/bitops.h
110
restore_irqs ip
arch/arm/lib/bitops.h
34
ands ip, r1, #3
arch/arm/lib/bitops.h
35
strbne r1, [ip] @ assert word-aligned
arch/arm/lib/bitops.h
50
strex ip, r2, [r1]
arch/arm/lib/bitops.h
51
cmp ip, #0
arch/arm/lib/bitops.h
72
ands ip, r1, #3
arch/arm/lib/bitops.h
73
strbne r1, [ip] @ assert word-aligned
arch/arm/lib/bitops.h
78
save_and_disable_irqs ip
arch/arm/lib/bitops.h
82
restore_irqs ip
arch/arm/lib/bitops.h
9
ands ip, r1, #3
arch/arm/lib/bitops.h
99
ands ip, r1, #3
arch/arm/mach-bcm/bcm_kona_smc.c
108
: "=r" (ip), "=r" (r0)
arch/arm/mach-bcm/bcm_kona_smc.c
112
BUG_ON(ip != SEC_EXIT_NORMAL);
arch/arm/mach-bcm/bcm_kona_smc.c
89
register u32 ip asm("ip"); /* Also called r12 */
arch/arm/mm/alignment.c
770
static int alignment_get_arm(struct pt_regs *regs, u32 *ip, u32 *inst)
arch/arm/mm/alignment.c
776
fault = get_user(instr, ip);
arch/arm/mm/alignment.c
778
fault = get_kernel_nofault(instr, ip);
arch/arm/mm/alignment.c
785
static int alignment_get_thumb(struct pt_regs *regs, u16 *ip, u16 *inst)
arch/arm/mm/alignment.c
791
fault = get_user(instr, ip);
arch/arm/mm/alignment.c
793
fault = get_kernel_nofault(instr, ip);
arch/arm/probes/decode.c
227
u16 *ip = (u16 *)asi->insn;
arch/arm/probes/decode.c
229
*ip++ = __opcode_to_mem_thumb16(insn >> 16);
arch/arm/probes/decode.c
230
*ip++ = __opcode_to_mem_thumb16(insn);
arch/arm64/include/asm/ftrace.h
172
void ftrace_graph_func(unsigned long ip, unsigned long parent_ip,
arch/arm64/kernel/compat_alignment.c
284
static int alignment_get_arm(struct pt_regs *regs, __le32 __user *ip, u32 *inst)
arch/arm64/kernel/compat_alignment.c
289
fault = get_user(instr, ip);
arch/arm64/kernel/compat_alignment.c
297
static int alignment_get_thumb(struct pt_regs *regs, __le16 __user *ip, u16 *inst)
arch/arm64/kernel/compat_alignment.c
302
fault = get_user(instr, ip);
arch/arm64/kernel/ftrace.c
299
unsigned long pc = rec->ip;
arch/arm64/kernel/ftrace.c
371
unsigned long literal = ALIGN_DOWN(rec->ip - 12, 8);
arch/arm64/kernel/ftrace.c
395
unsigned long pc = rec->ip;
arch/arm64/kernel/ftrace.c
416
unsigned long pc = rec->ip;
arch/arm64/kernel/ftrace.c
462
unsigned long pc = rec->ip - AARCH64_INSN_SIZE;
arch/arm64/kernel/ftrace.c
484
unsigned long pc = rec->ip;
arch/arm64/kernel/ftrace.c
550
void ftrace_graph_func(unsigned long ip, unsigned long parent_ip,
arch/arm64/kernel/ftrace.c
563
if (!function_graph_enter_regs(old, ip, frame_pointer,
arch/arm64/kernel/stacktrace.c
424
bool (*consume_entry)(void *cookie, u64 ip, u64 sp, u64 fp);
arch/arm64/kernel/stacktrace.c
437
noinline noinstr void arch_bpf_stack_walk(bool (*consume_entry)(void *cookie, u64 ip, u64 sp,
arch/arm64/kvm/at.c
1062
#define set_perms(w, wr, ip) \
arch/arm64/kvm/at.c
1065
switch ((ip)) { \
arch/arm64/kvm/at.c
1117
set_ ## w ## _wxn((wr), ((ip) == 0b0110)); \
arch/arm64/net/bpf_jit_comp.c
2889
static bool is_long_jump(void *ip, void *target)
arch/arm64/net/bpf_jit_comp.c
2897
offset = (long)target - (long)ip;
arch/arm64/net/bpf_jit_comp.c
2901
static int gen_branch_or_nop(enum aarch64_insn_branch_type type, void *ip,
arch/arm64/net/bpf_jit_comp.c
2911
if (is_long_jump(ip, addr))
arch/arm64/net/bpf_jit_comp.c
2916
*insn = aarch64_insn_gen_branch_imm((unsigned long)ip,
arch/arm64/net/bpf_jit_comp.c
2983
int bpf_arch_text_poke(void *ip, enum bpf_text_poke_type old_t,
arch/arm64/net/bpf_jit_comp.c
3000
if (!bpf_address_lookup((unsigned long)ip, &size, &offset, namebuf))
arch/arm64/net/bpf_jit_comp.c
3007
image = ip - offset;
arch/arm64/net/bpf_jit_comp.c
3021
ip = image + POKE_OFFSET * AARCH64_INSN_SIZE;
arch/arm64/net/bpf_jit_comp.c
3025
if (WARN_ON((is_long_jump(ip, new_addr) || is_long_jump(ip, old_addr)) &&
arch/arm64/net/bpf_jit_comp.c
3031
if (gen_branch_or_nop(branch_type, ip, old_addr, plt, &old_insn) < 0)
arch/arm64/net/bpf_jit_comp.c
3036
if (gen_branch_or_nop(branch_type, ip, new_addr, plt, &new_insn) < 0)
arch/arm64/net/bpf_jit_comp.c
3039
if (is_long_jump(ip, new_addr))
arch/arm64/net/bpf_jit_comp.c
3041
else if (is_long_jump(ip, old_addr))
arch/arm64/net/bpf_jit_comp.c
3072
if (aarch64_insn_read(ip, &replaced)) {
arch/arm64/net/bpf_jit_comp.c
3097
ret = aarch64_insn_patch_text_nosync(ip, new_insn);
arch/csky/kernel/ftrace.c
113
int ret = ftrace_check_current_nop(rec->ip);
arch/csky/kernel/ftrace.c
118
return ftrace_modify_code(rec->ip, addr, true, false);
arch/csky/kernel/ftrace.c
124
return ftrace_modify_code(rec->ip, addr, false, false);
arch/csky/kernel/ftrace.c
142
return ftrace_modify_code(rec->ip, addr, true, true);
arch/csky/kernel/probes/ftrace.c
18
bit = ftrace_test_recursion_trylock(ip, parent_ip);
arch/csky/kernel/probes/ftrace.c
23
p = get_kprobe((kprobe_opcode_t *)ip);
arch/csky/kernel/probes/ftrace.c
25
p = get_kprobe((kprobe_opcode_t *)(ip - MCOUNT_INSN_SIZE));
arch/csky/kernel/probes/ftrace.c
38
ip -= MCOUNT_INSN_SIZE;
arch/csky/kernel/probes/ftrace.c
39
instruction_pointer_set(regs, ip);
arch/csky/kernel/probes/ftrace.c
6
void kprobe_ftrace_handler(unsigned long ip, unsigned long parent_ip,
arch/hexagon/kernel/traps.c
111
if (!ip)
arch/hexagon/kernel/traps.c
112
ip = *(fp+1);
arch/hexagon/kernel/traps.c
121
name = kallsyms_lookup(ip, &size, &offset, &modname, tmpstr);
arch/hexagon/kernel/traps.c
123
printk("%s[%p] 0x%lx: %s + 0x%lx", loglvl, fp, ip, name, offset);
arch/hexagon/kernel/traps.c
157
ip = pt_elr(regs);
arch/hexagon/kernel/traps.c
159
ip = *(newfp + 1);
arch/hexagon/kernel/traps.c
163
if (ip == 0x0)
arch/hexagon/kernel/traps.c
78
unsigned long ip, const char *loglvl)
arch/loongarch/include/asm/ftrace.h
55
ftrace_regs_set_instruction_pointer(struct ftrace_regs *fregs, unsigned long ip)
arch/loongarch/include/asm/ftrace.h
57
instruction_pointer_set(&arch_ftrace_regs(fregs)->regs, ip);
arch/loongarch/include/asm/ftrace.h
71
void ftrace_graph_func(unsigned long ip, unsigned long parent_ip,
arch/loongarch/include/asm/inst.h
422
static inline bool is_break_ins(union loongarch_instruction *ip)
arch/loongarch/include/asm/inst.h
424
return ip->reg0i15_format.opcode == break_op;
arch/loongarch/include/asm/inst.h
427
static inline bool is_pc_ins(union loongarch_instruction *ip)
arch/loongarch/include/asm/inst.h
429
return ip->reg1i20_format.opcode >= pcaddi_op &&
arch/loongarch/include/asm/inst.h
430
ip->reg1i20_format.opcode <= pcaddu18i_op;
arch/loongarch/include/asm/inst.h
433
static inline bool is_branch_ins(union loongarch_instruction *ip)
arch/loongarch/include/asm/inst.h
435
return ip->reg1i21_format.opcode >= beqz_op &&
arch/loongarch/include/asm/inst.h
436
ip->reg1i21_format.opcode <= bgeu_op;
arch/loongarch/include/asm/inst.h
439
static inline bool is_ra_save_ins(union loongarch_instruction *ip)
arch/loongarch/include/asm/inst.h
444
return ip->reg2i12_format.opcode == opcode &&
arch/loongarch/include/asm/inst.h
445
ip->reg2i12_format.rj == LOONGARCH_GPR_SP &&
arch/loongarch/include/asm/inst.h
446
ip->reg2i12_format.rd == LOONGARCH_GPR_RA &&
arch/loongarch/include/asm/inst.h
447
!is_imm12_negative(ip->reg2i12_format.immediate);
arch/loongarch/include/asm/inst.h
450
static inline bool is_stack_alloc_ins(union loongarch_instruction *ip)
arch/loongarch/include/asm/inst.h
455
return ip->reg2i12_format.opcode == opcode &&
arch/loongarch/include/asm/inst.h
456
ip->reg2i12_format.rj == LOONGARCH_GPR_SP &&
arch/loongarch/include/asm/inst.h
457
ip->reg2i12_format.rd == LOONGARCH_GPR_SP &&
arch/loongarch/include/asm/inst.h
458
is_imm12_negative(ip->reg2i12_format.immediate);
arch/loongarch/include/asm/inst.h
461
static inline bool is_self_loop_ins(union loongarch_instruction *ip, struct pt_regs *regs)
arch/loongarch/include/asm/inst.h
463
switch (ip->reg0i26_format.opcode) {
arch/loongarch/include/asm/inst.h
466
if (ip->reg0i26_format.immediate_l == 0
arch/loongarch/include/asm/inst.h
467
&& ip->reg0i26_format.immediate_h == 0)
arch/loongarch/include/asm/inst.h
471
switch (ip->reg1i21_format.opcode) {
arch/loongarch/include/asm/inst.h
475
if (ip->reg1i21_format.immediate_l == 0
arch/loongarch/include/asm/inst.h
476
&& ip->reg1i21_format.immediate_h == 0)
arch/loongarch/include/asm/inst.h
480
switch (ip->reg2i16_format.opcode) {
arch/loongarch/include/asm/inst.h
487
if (ip->reg2i16_format.immediate == 0)
arch/loongarch/include/asm/inst.h
491
if (regs->regs[ip->reg2i16_format.rj] +
arch/loongarch/include/asm/inst.h
492
((unsigned long)ip->reg2i16_format.immediate << 2) == (unsigned long)ip)
arch/loongarch/kernel/ftrace_dyn.c
122
pc = rec->ip + LOONGARCH_INSN_SIZE;
arch/loongarch/kernel/ftrace_dyn.c
174
pc = rec->ip;
arch/loongarch/kernel/ftrace_dyn.c
186
pc = rec->ip + LOONGARCH_INSN_SIZE;
arch/loongarch/kernel/ftrace_dyn.c
202
pc = rec->ip + LOONGARCH_INSN_SIZE;
arch/loongarch/kernel/ftrace_dyn.c
240
void ftrace_graph_func(unsigned long ip, unsigned long parent_ip,
arch/loongarch/kernel/ftrace_dyn.c
253
if (!function_graph_enter_regs(old, ip, 0, parent, fregs))
arch/loongarch/kernel/ftrace_dyn.c
289
void kprobe_ftrace_handler(unsigned long ip, unsigned long parent_ip,
arch/loongarch/kernel/ftrace_dyn.c
300
bit = ftrace_test_recursion_trylock(ip, parent_ip);
arch/loongarch/kernel/ftrace_dyn.c
304
p = get_kprobe((kprobe_opcode_t *)ip);
arch/loongarch/kernel/ftrace_dyn.c
318
instruction_pointer_set(regs, ip);
arch/loongarch/kernel/ftrace_dyn.c
65
unsigned long pc = rec->ip + LOONGARCH_INSN_SIZE;
arch/loongarch/kernel/kgdb.c
300
union loongarch_instruction *ip = (union loongarch_instruction *)pc;
arch/loongarch/kernel/kgdb.c
309
si_h = ip->reg0i26_format.immediate_h;
arch/loongarch/kernel/kgdb.c
310
si_l = ip->reg0i26_format.immediate_l;
arch/loongarch/kernel/kgdb.c
311
switch (ip->reg0i26_format.opcode) {
arch/loongarch/kernel/kgdb.c
321
rj = ip->reg1i21_format.rj;
arch/loongarch/kernel/kgdb.c
323
si_l = ip->reg1i21_format.immediate_l;
arch/loongarch/kernel/kgdb.c
324
si_h = ip->reg1i21_format.immediate_h;
arch/loongarch/kernel/kgdb.c
326
switch (ip->reg1i21_format.opcode) {
arch/loongarch/kernel/kgdb.c
343
rj = ip->reg2i16_format.rj;
arch/loongarch/kernel/kgdb.c
344
rd = ip->reg2i16_format.rd;
arch/loongarch/kernel/kgdb.c
345
si = ip->reg2i16_format.immediate;
arch/loongarch/kernel/kgdb.c
346
switch (ip->reg2i16_format.opcode) {
arch/loongarch/kernel/traps.c
816
union loongarch_instruction *ip = (union loongarch_instruction *)pc;
arch/loongarch/kernel/traps.c
835
if (!is_self_loop_ins(ip, regs)) {
arch/loongarch/kernel/unwind_orc.c
100
static struct orc_entry *orc_module_find(unsigned long ip)
arch/loongarch/kernel/unwind_orc.c
107
static struct orc_entry *orc_find(unsigned long ip);
arch/loongarch/kernel/unwind_orc.c
119
static struct orc_entry *orc_ftrace_find(unsigned long ip)
arch/loongarch/kernel/unwind_orc.c
124
ops = ftrace_ops_trampoline(ip);
arch/loongarch/kernel/unwind_orc.c
135
offset = ip - ops->trampoline;
arch/loongarch/kernel/unwind_orc.c
139
if (ip == tramp_addr)
arch/loongarch/kernel/unwind_orc.c
145
static struct orc_entry *orc_ftrace_find(unsigned long ip)
arch/loongarch/kernel/unwind_orc.c
151
static struct orc_entry *orc_find(unsigned long ip)
arch/loongarch/kernel/unwind_orc.c
155
if (ip == 0)
arch/loongarch/kernel/unwind_orc.c
159
if (ip >= LOOKUP_START_IP && ip < LOOKUP_STOP_IP) {
arch/loongarch/kernel/unwind_orc.c
162
idx = (ip - LOOKUP_START_IP) / LOOKUP_BLOCK_SIZE;
arch/loongarch/kernel/unwind_orc.c
166
idx, lookup_num_blocks, (void *)ip);
arch/loongarch/kernel/unwind_orc.c
176
idx, lookup_num_blocks, start, stop, (void *)ip);
arch/loongarch/kernel/unwind_orc.c
181
__start_orc_unwind + start, stop - start, ip);
arch/loongarch/kernel/unwind_orc.c
185
if (is_kernel_inittext(ip))
arch/loongarch/kernel/unwind_orc.c
187
__stop_orc_unwind_ip - __start_orc_unwind_ip, ip);
arch/loongarch/kernel/unwind_orc.c
190
orc = orc_module_find(ip);
arch/loongarch/kernel/unwind_orc.c
194
return orc_ftrace_find(ip);
arch/loongarch/kernel/unwind_orc.c
54
static inline unsigned long orc_ip(const int *ip)
arch/loongarch/kernel/unwind_orc.c
56
return (unsigned long)ip + *ip;
arch/loongarch/kernel/unwind_orc.c
60
unsigned int num_entries, unsigned long ip)
arch/loongarch/kernel/unwind_orc.c
78
if (orc_ip(mid) <= ip) {
arch/loongarch/kernel/unwind_orc.c
89
static struct orc_entry *orc_module_find(unsigned long ip)
arch/loongarch/kernel/unwind_orc.c
93
mod = __module_address(ip);
arch/loongarch/kernel/unwind_orc.c
97
return __orc_find(mod->arch.orc_unwind_ip, mod->arch.orc_unwind, mod->arch.num_orcs, ip);
arch/loongarch/kernel/unwind_prologue.c
114
union loongarch_instruction *ip, *ip_end;
arch/loongarch/kernel/unwind_prologue.c
137
ip = (union loongarch_instruction *)(pc - offset);
arch/loongarch/kernel/unwind_prologue.c
140
while (ip < ip_end) {
arch/loongarch/kernel/unwind_prologue.c
141
if (is_stack_alloc_ins(ip)) {
arch/loongarch/kernel/unwind_prologue.c
142
frame_size = (1 << 12) - ip->reg2i12_format.immediate;
arch/loongarch/kernel/unwind_prologue.c
143
ip++;
arch/loongarch/kernel/unwind_prologue.c
146
ip++;
arch/loongarch/kernel/unwind_prologue.c
160
while (ip < ip_end) {
arch/loongarch/kernel/unwind_prologue.c
161
if (is_ra_save_ins(ip)) {
arch/loongarch/kernel/unwind_prologue.c
162
frame_ra = ip->reg2i12_format.immediate;
arch/loongarch/kernel/unwind_prologue.c
165
if (is_branch_ins(ip))
arch/loongarch/kernel/unwind_prologue.c
167
ip++;
arch/loongarch/net/bpf_jit.c
1356
static int emit_jump_or_nops(void *target, void *ip, u32 *insns, bool is_call)
arch/loongarch/net/bpf_jit.c
1391
int bpf_arch_text_poke(void *ip, enum bpf_text_poke_type old_t,
arch/loongarch/net/bpf_jit.c
1407
if (!bpf_address_lookup((unsigned long)ip, &size, &offset, namebuf))
arch/loongarch/net/bpf_jit.c
1410
image = ip - offset;
arch/loongarch/net/bpf_jit.c
1418
ip = image + LOONGARCH_INSN_SIZE;
arch/loongarch/net/bpf_jit.c
1422
ret = emit_jump_or_nops(old_addr, ip, old_insns, is_call);
arch/loongarch/net/bpf_jit.c
1426
if (memcmp(ip, old_insns, LOONGARCH_LONG_JUMP_NBYTES))
arch/loongarch/net/bpf_jit.c
1430
ret = emit_jump_or_nops(new_addr, ip, new_insns, is_call);
arch/loongarch/net/bpf_jit.c
1436
if (memcmp(ip, new_insns, LOONGARCH_LONG_JUMP_NBYTES))
arch/loongarch/net/bpf_jit.c
1437
ret = larch_insn_text_copy(ip, new_insns, LOONGARCH_LONG_JUMP_NBYTES);
arch/m68k/mac/misc.c
560
const unsigned short int *ip;
arch/m68k/mac/misc.c
597
ip = __mon_yday[__isleap(y)];
arch/m68k/mac/misc.c
598
for (y = 11; days < (long int) ip[y]; --y)
arch/m68k/mac/misc.c
600
days -= ip[y];
arch/microblaze/kernel/ftrace.c
135
imm = *(unsigned int *)rec->ip;
arch/microblaze/kernel/ftrace.c
138
bralid = *(unsigned int *)(rec->ip + 4);
arch/microblaze/kernel/ftrace.c
144
ret = ftrace_modify_code(rec->ip, MICROBLAZE_NOP);
arch/microblaze/kernel/ftrace.c
145
ret += ftrace_modify_code(rec->ip + 4, MICROBLAZE_NOP);
arch/microblaze/kernel/ftrace.c
147
ret = ftrace_modify_code(rec->ip, MICROBLAZE_BRI);
arch/microblaze/kernel/ftrace.c
157
__func__, (unsigned int)addr, (unsigned int)rec->ip, imm);
arch/microblaze/kernel/ftrace.c
158
ret = ftrace_modify_code(rec->ip, imm);
arch/microblaze/kernel/ftrace.c
161
ret += ftrace_modify_code(rec->ip + 4, bralid);
arch/microblaze/kernel/ftrace.c
168
unsigned long ip = (unsigned long)(&ftrace_call);
arch/microblaze/kernel/ftrace.c
178
__func__, (unsigned int)func, (unsigned int)ip, upper, lower);
arch/microblaze/kernel/ftrace.c
181
ret = ftrace_modify_code(ip, upper);
arch/microblaze/kernel/ftrace.c
182
ret += ftrace_modify_code(ip + 4, lower);
arch/microblaze/kernel/ftrace.c
197
unsigned long ip = (unsigned long)(&ftrace_call_graph);
arch/microblaze/kernel/ftrace.c
199
old_jump = *(unsigned int *)ip; /* save jump over instruction */
arch/microblaze/kernel/ftrace.c
200
ret = ftrace_modify_code(ip, MICROBLAZE_NOP);
arch/microblaze/kernel/ftrace.c
209
unsigned long ip = (unsigned long)(&ftrace_call_graph);
arch/microblaze/kernel/ftrace.c
211
ret = ftrace_modify_code(ip, old_jump);
arch/microblaze/kernel/kgdb.c
108
void kgdb_arch_set_pc(struct pt_regs *regs, unsigned long ip)
arch/microblaze/kernel/kgdb.c
110
regs->pc = ip;
arch/mips/kernel/ftrace.c
104
safe_store_code(new_code1, ip, faulted);
arch/mips/kernel/ftrace.c
108
ip += 4;
arch/mips/kernel/ftrace.c
109
safe_store_code(new_code2, ip, faulted);
arch/mips/kernel/ftrace.c
113
ip -= 4;
arch/mips/kernel/ftrace.c
114
flush_icache_range(ip, ip + 8);
arch/mips/kernel/ftrace.c
119
static int ftrace_modify_code_2r(unsigned long ip, unsigned int new_code1,
arch/mips/kernel/ftrace.c
124
ip += 4;
arch/mips/kernel/ftrace.c
125
safe_store_code(new_code2, ip, faulted);
arch/mips/kernel/ftrace.c
129
ip -= 4;
arch/mips/kernel/ftrace.c
130
safe_store_code(new_code1, ip, faulted);
arch/mips/kernel/ftrace.c
134
flush_icache_range(ip, ip + 8);
arch/mips/kernel/ftrace.c
176
unsigned long ip = rec->ip;
arch/mips/kernel/ftrace.c
182
new = core_kernel_text(ip) ? INSN_NOP : INSN_B_1F;
arch/mips/kernel/ftrace.c
184
return ftrace_modify_code(ip, new);
arch/mips/kernel/ftrace.c
193
return ftrace_modify_code_2(ip, new, INSN_NOP);
arch/mips/kernel/ftrace.c
200
unsigned long ip = rec->ip;
arch/mips/kernel/ftrace.c
206
if (!core_kernel_text(ip) && !uasm_in_compat_space_p(MCOUNT_ADDR))
arch/mips/kernel/ftrace.c
209
new = core_kernel_text(ip) ? insn_jal_ftrace_caller : insn_la_mcount[0];
arch/mips/kernel/ftrace.c
212
return ftrace_modify_code(ip, new);
arch/mips/kernel/ftrace.c
214
return ftrace_modify_code_2r(ip, new, core_kernel_text(ip) ?
arch/mips/kernel/ftrace.c
271
unsigned long sp, ip, tmp;
arch/mips/kernel/ftrace.c
280
ip = self_ra - (core_kernel_text(self_ra) ? 16 : 24);
arch/mips/kernel/ftrace.c
288
safe_load_code(code, ip, faulted);
arch/mips/kernel/ftrace.c
301
ip -= 4;
arch/mips/kernel/ftrace.c
83
static int ftrace_modify_code(unsigned long ip, unsigned int new_code)
arch/mips/kernel/ftrace.c
88
safe_store_code(new_code, ip, faulted);
arch/mips/kernel/ftrace.c
93
flush_icache_range(ip, ip + 8);
arch/mips/kernel/ftrace.c
99
static int ftrace_modify_code_2(unsigned long ip, unsigned int new_code1,
arch/mips/kernel/process.c
193
static inline int is_jr_ra_ins(union mips_instruction *ip)
arch/mips/kernel/process.c
200
if (mm_insn_16bit(ip->word >> 16)) {
arch/mips/kernel/process.c
201
if (ip->mm16_r5_format.opcode == mm_pool16c_op &&
arch/mips/kernel/process.c
202
ip->mm16_r5_format.rt == mm_jr16_op &&
arch/mips/kernel/process.c
203
ip->mm16_r5_format.imm == 31)
arch/mips/kernel/process.c
208
if (ip->r_format.opcode == mm_pool32a_op &&
arch/mips/kernel/process.c
209
ip->r_format.func == mm_pool32axf_op &&
arch/mips/kernel/process.c
210
((ip->u_format.uimmediate >> 6) & GENMASK(9, 0)) == mm_jalr_op &&
arch/mips/kernel/process.c
211
ip->r_format.rt == 31)
arch/mips/kernel/process.c
215
if (ip->r_format.opcode == spec_op &&
arch/mips/kernel/process.c
216
ip->r_format.func == jr_op &&
arch/mips/kernel/process.c
217
ip->r_format.rs == 31)
arch/mips/kernel/process.c
223
static inline int is_ra_save_ins(union mips_instruction *ip, int *poff)
arch/mips/kernel/process.c
235
if (mm_insn_16bit(ip->word >> 16)) {
arch/mips/kernel/process.c
236
switch (ip->mm16_r5_format.opcode) {
arch/mips/kernel/process.c
238
if (ip->mm16_r5_format.rt != 31)
arch/mips/kernel/process.c
241
*poff = ip->mm16_r5_format.imm;
arch/mips/kernel/process.c
246
switch (ip->mm16_m_format.func) {
arch/mips/kernel/process.c
248
*poff = ip->mm16_m_format.imm;
arch/mips/kernel/process.c
249
*poff += 1 + ip->mm16_m_format.rlist;
arch/mips/kernel/process.c
262
switch (ip->i_format.opcode) {
arch/mips/kernel/process.c
264
if (ip->i_format.rs != 29)
arch/mips/kernel/process.c
266
if (ip->i_format.rt != 31)
arch/mips/kernel/process.c
269
*poff = ip->i_format.simmediate / sizeof(ulong);
arch/mips/kernel/process.c
273
switch (ip->mm_m_format.func) {
arch/mips/kernel/process.c
275
if (ip->mm_m_format.rd < 0x10)
arch/mips/kernel/process.c
277
if (ip->mm_m_format.base != 29)
arch/mips/kernel/process.c
280
*poff = ip->mm_m_format.simmediate;
arch/mips/kernel/process.c
281
*poff += (ip->mm_m_format.rd & 0xf) * sizeof(u32);
arch/mips/kernel/process.c
293
if ((ip->i_format.opcode == sw_op || ip->i_format.opcode == sd_op) &&
arch/mips/kernel/process.c
294
ip->i_format.rs == 29 && ip->i_format.rt == 31) {
arch/mips/kernel/process.c
295
*poff = ip->i_format.simmediate / sizeof(ulong);
arch/mips/kernel/process.c
299
if ((ip->loongson3_lswc2_format.opcode == swc2_op) &&
arch/mips/kernel/process.c
300
(ip->loongson3_lswc2_format.ls == 1) &&
arch/mips/kernel/process.c
301
(ip->loongson3_lswc2_format.fr == 0) &&
arch/mips/kernel/process.c
302
(ip->loongson3_lswc2_format.base == 29)) {
arch/mips/kernel/process.c
303
if (ip->loongson3_lswc2_format.rt == 31) {
arch/mips/kernel/process.c
304
*poff = ip->loongson3_lswc2_format.offset << 1;
arch/mips/kernel/process.c
307
if (ip->loongson3_lswc2_format.rq == 31) {
arch/mips/kernel/process.c
308
*poff = (ip->loongson3_lswc2_format.offset << 1) + 1;
arch/mips/kernel/process.c
317
static inline int is_jump_ins(union mips_instruction *ip)
arch/mips/kernel/process.c
328
if (mm_insn_16bit(ip->word >> 16)) {
arch/mips/kernel/process.c
329
if ((ip->mm16_r5_format.opcode == mm_pool16c_op &&
arch/mips/kernel/process.c
330
(ip->mm16_r5_format.rt & mm_jr16_op) == mm_jr16_op))
arch/mips/kernel/process.c
335
if (ip->j_format.opcode == mm_j32_op)
arch/mips/kernel/process.c
337
if (ip->j_format.opcode == mm_jal32_op)
arch/mips/kernel/process.c
339
if (ip->r_format.opcode != mm_pool32a_op ||
arch/mips/kernel/process.c
340
ip->r_format.func != mm_pool32axf_op)
arch/mips/kernel/process.c
342
return ((ip->u_format.uimmediate >> 6) & mm_jalr_op) == mm_jalr_op;
arch/mips/kernel/process.c
344
if (ip->j_format.opcode == j_op)
arch/mips/kernel/process.c
346
if (ip->j_format.opcode == jal_op)
arch/mips/kernel/process.c
348
if (ip->r_format.opcode != spec_op)
arch/mips/kernel/process.c
350
return ip->r_format.func == jalr_op || ip->r_format.func == jr_op;
arch/mips/kernel/process.c
354
static inline int is_sp_move_ins(union mips_instruction *ip, int *frame_size)
arch/mips/kernel/process.c
367
if (mm_insn_16bit(ip->word >> 16)) {
arch/mips/kernel/process.c
368
if (ip->mm16_r3_format.opcode == mm_pool16d_op &&
arch/mips/kernel/process.c
369
ip->mm16_r3_format.simmediate & mm_addiusp_func) {
arch/mips/kernel/process.c
370
tmp = ip->mm_b0_format.simmediate >> 1;
arch/mips/kernel/process.c
377
if (ip->mm16_r5_format.opcode == mm_pool16d_op &&
arch/mips/kernel/process.c
378
ip->mm16_r5_format.rt == 29) {
arch/mips/kernel/process.c
379
tmp = ip->mm16_r5_format.imm >> 1;
arch/mips/kernel/process.c
386
if (ip->mm_i_format.opcode == mm_addiu32_op &&
arch/mips/kernel/process.c
387
ip->mm_i_format.rt == 29 && ip->mm_i_format.rs == 29) {
arch/mips/kernel/process.c
388
*frame_size = -ip->i_format.simmediate;
arch/mips/kernel/process.c
393
if (ip->i_format.rs != 29 || ip->i_format.rt != 29)
arch/mips/kernel/process.c
396
if (ip->i_format.opcode == addiu_op ||
arch/mips/kernel/process.c
397
ip->i_format.opcode == daddiu_op) {
arch/mips/kernel/process.c
398
*frame_size = -ip->i_format.simmediate;
arch/mips/kernel/process.c
408
union mips_instruction insn, *ip, *ip_end;
arch/mips/kernel/process.c
415
ip = (void *)msk_isa16_mode((ulong)info->func);
arch/mips/kernel/process.c
416
if (!ip)
arch/mips/kernel/process.c
419
ip_end = (void *)ip + (info->func_size ? info->func_size : 512);
arch/mips/kernel/process.c
421
while (ip < ip_end) {
arch/mips/kernel/process.c
422
ip = (void *)ip + last_insn_size;
arch/mips/kernel/process.c
424
if (is_mmips && mm_insn_16bit(ip->halfword[0])) {
arch/mips/kernel/process.c
425
insn.word = ip->halfword[0] << 16;
arch/mips/kernel/process.c
428
insn.word = ip->halfword[0] << 16 | ip->halfword[1];
arch/mips/kernel/process.c
431
insn.word = ip->word;
arch/mips/kernel/process.c
435
if (is_jr_ra_ins(ip)) {
arch/mips/kernel/process.c
440
} else if (!saw_jump && is_jump_ins(ip)) {
arch/mips/kernel/process.c
482
union mips_instruction *ip = (void *)schedule;
arch/mips/kernel/process.c
486
for (i = 0; i < max_insns; i++, ip++) {
arch/mips/kernel/process.c
487
if (ip->j_format.opcode == j_op)
arch/mips/kernel/process.c
488
return J_TARGET(ip, ip->j_format.target);
arch/mips/mm/uasm-micromips.c
158
const struct insn *ip;
arch/mips/mm/uasm-micromips.c
167
ip = &insn_table_MM[opc];
arch/mips/mm/uasm-micromips.c
169
op = ip->match;
arch/mips/mm/uasm-micromips.c
171
if (ip->fields & RS) {
arch/mips/mm/uasm-micromips.c
178
if (ip->fields & RT) {
arch/mips/mm/uasm-micromips.c
185
if (ip->fields & RD)
arch/mips/mm/uasm-micromips.c
187
if (ip->fields & RE)
arch/mips/mm/uasm-micromips.c
189
if (ip->fields & SIMM)
arch/mips/mm/uasm-micromips.c
191
if (ip->fields & UIMM)
arch/mips/mm/uasm-micromips.c
193
if (ip->fields & BIMM)
arch/mips/mm/uasm-micromips.c
195
if (ip->fields & JIMM)
arch/mips/mm/uasm-micromips.c
197
if (ip->fields & FUNC)
arch/mips/mm/uasm-micromips.c
199
if (ip->fields & SET)
arch/mips/mm/uasm-micromips.c
201
if (ip->fields & SCIMM)
arch/mips/mm/uasm-mips.c
234
const struct insn *ip;
arch/mips/mm/uasm-mips.c
243
ip = &insn_table[opc];
arch/mips/mm/uasm-mips.c
245
op = ip->match;
arch/mips/mm/uasm-mips.c
247
if (ip->fields & RS)
arch/mips/mm/uasm-mips.c
249
if (ip->fields & RT)
arch/mips/mm/uasm-mips.c
251
if (ip->fields & RD)
arch/mips/mm/uasm-mips.c
253
if (ip->fields & RE)
arch/mips/mm/uasm-mips.c
255
if (ip->fields & SIMM)
arch/mips/mm/uasm-mips.c
257
if (ip->fields & UIMM)
arch/mips/mm/uasm-mips.c
259
if (ip->fields & BIMM)
arch/mips/mm/uasm-mips.c
261
if (ip->fields & JIMM)
arch/mips/mm/uasm-mips.c
263
if (ip->fields & FUNC)
arch/mips/mm/uasm-mips.c
265
if (ip->fields & SET)
arch/mips/mm/uasm-mips.c
267
if (ip->fields & SCIMM)
arch/mips/mm/uasm-mips.c
269
if (ip->fields & SIMM9)
arch/mips/rb532/irq.c
100
int ipnum = 0x100 << ip;
arch/mips/rb532/irq.c
105
static inline void ack_local_irq(unsigned int ip)
arch/mips/rb532/irq.c
107
int ipnum = 0x100 << ip;
arch/mips/rb532/irq.c
115
int ip = irq_nr - GROUP0_IRQ_BASE;
arch/mips/rb532/irq.c
118
if (ip < 0)
arch/mips/rb532/irq.c
121
group = ip >> 5;
arch/mips/rb532/irq.c
123
ip &= (1 << 5) - 1;
arch/mips/rb532/irq.c
124
intr_bit = 1 << ip;
arch/mips/rb532/irq.c
136
int ip = irq_nr - GROUP0_IRQ_BASE;
arch/mips/rb532/irq.c
139
if (ip < 0) {
arch/mips/rb532/irq.c
142
group = ip >> 5;
arch/mips/rb532/irq.c
144
ip &= (1 << 5) - 1;
arch/mips/rb532/irq.c
145
intr_bit = 1 << ip;
arch/mips/rb532/irq.c
215
unsigned int ip, pend, group;
arch/mips/rb532/irq.c
222
ip = (cp0_cause & 0x7c00);
arch/mips/rb532/irq.c
223
if (ip) {
arch/mips/rb532/irq.c
224
group = 21 + (fls(ip) - 32);
arch/mips/rb532/irq.c
91
static inline void enable_local_irq(unsigned int ip)
arch/mips/rb532/irq.c
93
int ipnum = 0x100 << ip;
arch/mips/rb532/irq.c
98
static inline void disable_local_irq(unsigned int ip)
arch/parisc/include/asm/unwind.h
64
unsigned long sp, ip, rp, r31;
arch/parisc/kernel/ftrace.c
118
void *ip;
arch/parisc/kernel/ftrace.c
163
if (rec->ip & 0x4) {
arch/parisc/kernel/ftrace.c
169
ip = (void *)(rec->ip + 4 - size);
arch/parisc/kernel/ftrace.c
171
ret = copy_from_kernel_nofault(insn, ip, size);
arch/parisc/kernel/ftrace.c
180
__patch_text_multiple(ip, tramp, size);
arch/parisc/kernel/ftrace.c
193
__patch_text((void *)rec->ip, INSN_NOP);
arch/parisc/kernel/ftrace.c
194
__patch_text_multiple((void *)rec->ip + 4 - sizeof(insn),
arch/parisc/kernel/ftrace.c
201
void kprobe_ftrace_handler(unsigned long ip, unsigned long parent_ip,
arch/parisc/kernel/ftrace.c
212
bit = ftrace_test_recursion_trylock(ip, parent_ip);
arch/parisc/kernel/ftrace.c
217
p = get_kprobe((kprobe_opcode_t *)ip);
arch/parisc/kernel/ftrace.c
231
regs->iaoq[0] = ip;
arch/parisc/kernel/ftrace.c
232
regs->iaoq[1] = ip + 4;
arch/parisc/kernel/ftrace.c
235
regs->iaoq[0] = ip + 4;
arch/parisc/kernel/ftrace.c
236
regs->iaoq[1] = ip + 8;
arch/parisc/kernel/kgdb.c
150
void kgdb_arch_set_pc(struct pt_regs *regs, unsigned long ip)
arch/parisc/kernel/kgdb.c
152
regs->iaoq[0] = ip;
arch/parisc/kernel/kgdb.c
153
regs->iaoq[1] = ip + 4;
arch/parisc/kernel/perf_event.c
20
if (unwind_once(&info) < 0 || info.ip == 0)
arch/parisc/kernel/perf_event.c
23
if (!__kernel_text_address(info.ip) ||
arch/parisc/kernel/perf_event.c
24
perf_callchain_store(entry, info.ip))
arch/parisc/kernel/process.c
265
unsigned long ip;
arch/parisc/kernel/process.c
278
ip = info.ip;
arch/parisc/kernel/process.c
279
if (!in_sched_functions(ip))
arch/parisc/kernel/process.c
280
return ip;
arch/parisc/kernel/stacktrace.c
23
if (unwind_once(&info) < 0 || info.ip == 0)
arch/parisc/kernel/stacktrace.c
26
if (__kernel_text_address(info.ip))
arch/parisc/kernel/stacktrace.c
27
if (!fn(cookie, info.ip))
arch/parisc/kernel/traps.c
187
if (unwind_once(info) < 0 || info->ip == 0)
arch/parisc/kernel/traps.c
190
if (__kernel_text_address(info->ip)) {
arch/parisc/kernel/traps.c
192
loglvl, info->ip, (void *) info->ip);
arch/parisc/kernel/unwind.c
287
e = find_unwind_entry(info->ip);
arch/parisc/kernel/unwind.c
292
(void *) info->ip);
arch/parisc/kernel/unwind.c
325
"prev_sp=%lx prev_ip=%lx\n", info->ip,
arch/parisc/kernel/unwind.c
338
npc < info->ip;
arch/parisc/kernel/unwind.c
348
"%lx, frame_size = %ld\n", info->ip,
arch/parisc/kernel/unwind.c
354
"%lx, frame_size = %ld\n", info->ip,
arch/parisc/kernel/unwind.c
361
"-20(sp) @ %lx\n", info->ip, npc);
arch/parisc/kernel/unwind.c
367
"-16(sp) @ %lx\n", info->ip, npc);
arch/parisc/kernel/unwind.c
387
"prev_ip=%lx npc=%lx\n", info->ip, info->prev_sp,
arch/parisc/kernel/unwind.c
398
info->ip = regs->iaoq[0];
arch/parisc/kernel/unwind.c
403
t ? (int)t->pid : -1, info->sp, info->ip);
arch/parisc/kernel/unwind.c
457
next_frame->ip = next_frame->prev_ip;
arch/parisc/kernel/unwind.c
463
next_frame->sp, next_frame->ip);
arch/parisc/kernel/unwind.c
474
} while (!ret && !(info->ip & 3));
arch/parisc/kernel/unwind.c
489
if (unwind_once(&info) < 0 || info.ip == 0)
arch/parisc/kernel/unwind.c
491
if (!kernel_text_address(info.ip))
arch/parisc/kernel/unwind.c
493
} while (info.ip && level--);
arch/parisc/kernel/unwind.c
495
return info.ip;
arch/powerpc/boot/stdio.c
264
long * ip = va_arg(args, long *);
arch/powerpc/boot/stdio.c
265
*ip = (str - buf);
arch/powerpc/boot/stdio.c
267
size_t * ip = va_arg(args, size_t *);
arch/powerpc/boot/stdio.c
268
*ip = (str - buf);
arch/powerpc/boot/stdio.c
270
int * ip = va_arg(args, int *);
arch/powerpc/boot/stdio.c
271
*ip = (str - buf);
arch/powerpc/include/asm/ftrace.h
21
unsigned long prepare_ftrace_return(unsigned long parent, unsigned long ip,
arch/powerpc/include/asm/ftrace.h
70
unsigned long ip)
arch/powerpc/include/asm/ftrace.h
72
regs_set_return_ip(&arch_ftrace_regs(fregs)->regs, ip);
arch/powerpc/include/asm/ftrace.h
84
void ftrace_graph_func(unsigned long ip, unsigned long parent_ip,
arch/powerpc/include/asm/imc-pmu.h
78
__be64 ip;
arch/powerpc/include/asm/mpc52xx_psc.h
201
u8 ip; /* PSC + 0x34 */
arch/powerpc/include/asm/mpc52xx_psc.h
342
u8 ip; /* PSC + 0x44 */
arch/powerpc/include/asm/ptrace.h
188
static inline void regs_set_return_ip(struct pt_regs *regs, unsigned long ip)
arch/powerpc/include/asm/ptrace.h
190
regs->nip = ip;
arch/powerpc/kernel/kprobes.c
110
unsigned long ip = ftrace_location(addr);
arch/powerpc/kernel/kprobes.c
112
if (ip)
arch/powerpc/kernel/kprobes.c
113
return offset <= (ip - addr);
arch/powerpc/kernel/process.c
2228
unsigned long ip, sp;
arch/powerpc/kernel/process.c
2240
ip = READ_ONCE_NOCHECK(((unsigned long *)sp)[STACK_FRAME_LR_SAVE]);
arch/powerpc/kernel/process.c
2241
if (!in_sched_functions(ip))
arch/powerpc/kernel/process.c
2242
return ip;
arch/powerpc/kernel/process.c
2284
unsigned long sp, ip, lr, newsp;
arch/powerpc/kernel/process.c
2312
ip = stack[STACK_FRAME_LR_SAVE];
arch/powerpc/kernel/process.c
2313
if (!firstframe || ip != lr) {
arch/powerpc/kernel/process.c
2315
loglvl, sp, ip, (void *)ip);
arch/powerpc/kernel/process.c
2317
&ftrace_idx, ip, stack);
arch/powerpc/kernel/process.c
2318
if (ret_addr != ip)
arch/powerpc/kernel/stacktrace.c
128
ip = stack[STACK_FRAME_LR_SAVE];
arch/powerpc/kernel/stacktrace.c
129
if (!__kernel_text_address(ip))
arch/powerpc/kernel/stacktrace.c
136
ip = ftrace_graph_ret_addr(task, &graph_idx, ip, stack);
arch/powerpc/kernel/stacktrace.c
143
if (ip == (unsigned long)arch_rethook_trampoline)
arch/powerpc/kernel/stacktrace.c
147
if (!consume_entry(cookie, ip))
arch/powerpc/kernel/stacktrace.c
45
unsigned long newsp, ip;
arch/powerpc/kernel/stacktrace.c
51
ip = stack[STACK_FRAME_LR_SAVE];
arch/powerpc/kernel/stacktrace.c
53
if (!consume_entry(cookie, ip))
arch/powerpc/kernel/stacktrace.c
97
unsigned long ip;
arch/powerpc/kernel/trace/ftrace.c
103
static inline int ftrace_modify_code(unsigned long ip, ppc_inst_t old, ppc_inst_t new)
arch/powerpc/kernel/trace/ftrace.c
105
int ret = ftrace_validate_inst(ip, old);
arch/powerpc/kernel/trace/ftrace.c
108
ret = patch_instruction((u32 *)ip, new);
arch/powerpc/kernel/trace/ftrace.c
118
static unsigned long find_ftrace_tramp(unsigned long ip)
arch/powerpc/kernel/trace/ftrace.c
125
else if (is_offset_in_branch_range(ftrace_tramps[i] - ip))
arch/powerpc/kernel/trace/ftrace.c
132
static unsigned long ftrace_lookup_module_stub(unsigned long ip, unsigned long addr)
arch/powerpc/kernel/trace/ftrace.c
137
mod = __module_text_address(ip);
arch/powerpc/kernel/trace/ftrace.c
139
pr_err("No module loaded at addr=%lx\n", ip);
arch/powerpc/kernel/trace/ftrace.c
144
static unsigned long ftrace_lookup_module_stub(unsigned long ip, unsigned long addr)
arch/powerpc/kernel/trace/ftrace.c
161
unsigned long ip;
arch/powerpc/kernel/trace/ftrace.c
165
ip = ftrace_get_ool_stub(rec) + MCOUNT_INSN_SIZE; /* second instruction in stub */
arch/powerpc/kernel/trace/ftrace.c
167
ip = rec->ip;
arch/powerpc/kernel/trace/ftrace.c
169
if (!is_offset_in_branch_range(addr - ip) && addr != FTRACE_ADDR &&
arch/powerpc/kernel/trace/ftrace.c
174
ip, rec->ip, addr);
arch/powerpc/kernel/trace/ftrace.c
180
if (is_offset_in_branch_range(addr - ip))
arch/powerpc/kernel/trace/ftrace.c
183
else if (core_kernel_text(ip))
arch/powerpc/kernel/trace/ftrace.c
185
stub = find_ftrace_tramp(ip);
arch/powerpc/kernel/trace/ftrace.c
187
stub = ftrace_lookup_module_stub(ip, addr);
arch/powerpc/kernel/trace/ftrace.c
190
pr_err("0x%lx (0x%lx): No ftrace stubs reachable\n", ip, rec->ip);
arch/powerpc/kernel/trace/ftrace.c
194
*call_inst = ftrace_create_branch_inst(ip, stub, 1);
arch/powerpc/kernel/trace/ftrace.c
219
if (is_kernel_inittext(rec->ip)) {
arch/powerpc/kernel/trace/ftrace.c
223
} else if (is_kernel_text(rec->ip)) {
arch/powerpc/kernel/trace/ftrace.c
230
!is_offset_in_branch_range((long)rec->ip -
arch/powerpc/kernel/trace/ftrace.c
255
if (!is_offset_in_branch_range((long)rec->ip - (long)&ool_stub->insn[0]) ||
arch/powerpc/kernel/trace/ftrace.c
256
!is_offset_in_branch_range((long)(rec->ip + MCOUNT_INSN_SIZE) -
arch/powerpc/kernel/trace/ftrace.c
259
__func__, (void *)rec->ip, (void *)&ool_stub->insn[0]);
arch/powerpc/kernel/trace/ftrace.c
283
if (!ret && create_branch(&inst, &ool_stub->insn[3], rec->ip + MCOUNT_INSN_SIZE, 0))
arch/powerpc/kernel/trace/ftrace.c
319
return patch_ulong((void *)(rec->ip - MCOUNT_INSN_SIZE - sizeof(unsigned long)),
arch/powerpc/kernel/trace/ftrace.c
349
unsigned long ip = rec->ip;
arch/powerpc/kernel/trace/ftrace.c
353
if (WARN_ON(!IS_ENABLED(CONFIG_MODULES) || core_kernel_text(ip)))
arch/powerpc/kernel/trace/ftrace.c
358
ip = ftrace_get_ool_stub(rec) + MCOUNT_INSN_SIZE; /* second instruction in stub */
arch/powerpc/kernel/trace/ftrace.c
365
ret = ftrace_modify_code(ip, old, new);
arch/powerpc/kernel/trace/ftrace.c
372
ret = ftrace_modify_code(rec->ip, ppc_inst(PPC_RAW_NOP()),
arch/powerpc/kernel/trace/ftrace.c
373
ppc_inst(PPC_RAW_BRANCH((long)ftrace_get_ool_stub(rec) - (long)rec->ip)));
arch/powerpc/kernel/trace/ftrace.c
392
unsigned long ip, new_addr, addr;
arch/powerpc/kernel/trace/ftrace.c
399
ip = rec->ip;
arch/powerpc/kernel/trace/ftrace.c
409
ip = ftrace_get_ool_stub(rec) + MCOUNT_INSN_SIZE;
arch/powerpc/kernel/trace/ftrace.c
441
ret = ftrace_modify_code(ip, old, new);
arch/powerpc/kernel/trace/ftrace.c
447
(long)rec->ip));
arch/powerpc/kernel/trace/ftrace.c
449
ip = rec->ip;
arch/powerpc/kernel/trace/ftrace.c
452
ret = ftrace_modify_code(ip, call_inst, nop_inst);
arch/powerpc/kernel/trace/ftrace.c
454
ret = ftrace_modify_code(ip, nop_inst, call_inst);
arch/powerpc/kernel/trace/ftrace.c
472
unsigned long addr, ip = rec->ip;
arch/powerpc/kernel/trace/ftrace.c
480
ret = ftrace_validate_inst(ip - 4, ppc_inst(PPC_RAW_NOP()));
arch/powerpc/kernel/trace/ftrace.c
482
ret = ftrace_validate_inst(ip, ppc_inst(PPC_RAW_NOP()));
arch/powerpc/kernel/trace/ftrace.c
485
ret = ftrace_validate_inst(ip - 8, ppc_inst(PPC_RAW_MFLR(_R0)));
arch/powerpc/kernel/trace/ftrace.c
488
ret = ftrace_modify_code(ip - 4, ppc_inst(PPC_RAW_STW(_R0, _R1, 4)),
arch/powerpc/kernel/trace/ftrace.c
492
ret = ftrace_read_inst(ip - 4, &old);
arch/powerpc/kernel/trace/ftrace.c
495
ret = ftrace_validate_inst(ip - 8, ppc_inst(PPC_RAW_MFLR(_R0)));
arch/powerpc/kernel/trace/ftrace.c
498
ret = ftrace_modify_code(ip - 4, ppc_inst(PPC_RAW_STD(_R0, _R1, 16)),
arch/powerpc/kernel/trace/ftrace.c
521
ret = ftrace_modify_code(ip - 4, old, new);
arch/powerpc/kernel/trace/ftrace.c
522
} else if (is_offset_in_branch_range(addr - ip)) {
arch/powerpc/kernel/trace/ftrace.c
524
old = ftrace_create_branch_inst(ip, addr, 1);
arch/powerpc/kernel/trace/ftrace.c
525
ret = ftrace_modify_code(ip, old, new);
arch/powerpc/kernel/trace/ftrace.c
526
} else if (core_kernel_text(ip) || (IS_ENABLED(CONFIG_MODULES) && mod)) {
arch/powerpc/kernel/trace/ftrace.c
531
ret = ftrace_read_inst(ip, &old);
arch/powerpc/kernel/trace/ftrace.c
535
pr_err("0x%lx: expected (bl) != found (%08lx)\n", ip, ppc_inst_as_ulong(old));
arch/powerpc/kernel/trace/ftrace.c
538
ret = patch_instruction((u32 *)ip, new);
arch/powerpc/kernel/trace/ftrace.c
552
unsigned long ip = (unsigned long)(&ftrace_call);
arch/powerpc/kernel/trace/ftrace.c
564
new = ftrace_create_branch_inst(ip, ppc_function_entry(func), 1);
arch/powerpc/kernel/trace/ftrace.c
565
ret = ftrace_modify_code(ip, old, new);
arch/powerpc/kernel/trace/ftrace.c
569
ip = (unsigned long)(&ftrace_regs_call);
arch/powerpc/kernel/trace/ftrace.c
571
new = ftrace_create_branch_inst(ip, ppc_function_entry(func), 1);
arch/powerpc/kernel/trace/ftrace.c
572
ret = ftrace_modify_code(ip, old, new);
arch/powerpc/kernel/trace/ftrace.c
679
void ftrace_graph_func(unsigned long ip, unsigned long parent_ip,
arch/powerpc/kernel/trace/ftrace.c
68
static ppc_inst_t ftrace_create_branch_inst(unsigned long ip, unsigned long addr, int link)
arch/powerpc/kernel/trace/ftrace.c
690
if (!function_graph_enter_regs(parent_ip, ip, 0, (unsigned long *)sp, fregs))
arch/powerpc/kernel/trace/ftrace.c
72
WARN_ON(!is_offset_in_branch_range(addr - ip));
arch/powerpc/kernel/trace/ftrace.c
73
create_branch(&op, (u32 *)ip, addr, link ? BRANCH_SET_LINK : 0);
arch/powerpc/kernel/trace/ftrace.c
78
static inline int ftrace_read_inst(unsigned long ip, ppc_inst_t *op)
arch/powerpc/kernel/trace/ftrace.c
80
if (copy_inst_from_kernel_nofault(op, (void *)ip)) {
arch/powerpc/kernel/trace/ftrace.c
81
pr_err("0x%lx: fetching instruction failed\n", ip);
arch/powerpc/kernel/trace/ftrace.c
88
static inline int ftrace_validate_inst(unsigned long ip, ppc_inst_t inst)
arch/powerpc/kernel/trace/ftrace.c
93
ret = ftrace_read_inst(ip, &op);
arch/powerpc/kernel/trace/ftrace.c
96
ip, ppc_inst_as_ulong(inst), ppc_inst_as_ulong(op));
arch/powerpc/kernel/trace/ftrace_64_pg.c
106
static unsigned long find_bl_target(unsigned long ip, ppc_inst_t op)
arch/powerpc/kernel/trace/ftrace_64_pg.c
115
return ip + (long)offset;
arch/powerpc/kernel/trace/ftrace_64_pg.c
124
mod = __module_text_address(rec->ip);
arch/powerpc/kernel/trace/ftrace_64_pg.c
126
pr_err("No module loaded at addr=%lx\n", rec->ip);
arch/powerpc/kernel/trace/ftrace_64_pg.c
136
unsigned long ip = rec->ip;
arch/powerpc/kernel/trace/ftrace_64_pg.c
146
if (copy_inst_from_kernel_nofault(&op, (void *)ip)) {
arch/powerpc/kernel/trace/ftrace_64_pg.c
158
tramp = find_bl_target(ip, op);
arch/powerpc/kernel/trace/ftrace_64_pg.c
160
pr_devel("ip:%lx jumps to %lx", ip, tramp);
arch/powerpc/kernel/trace/ftrace_64_pg.c
177
if (copy_inst_from_kernel_nofault(&op, (void *)(ip - 4))) {
arch/powerpc/kernel/trace/ftrace_64_pg.c
178
pr_err("Fetching instruction at %lx failed.\n", ip - 4);
arch/powerpc/kernel/trace/ftrace_64_pg.c
194
if (copy_inst_from_kernel_nofault(&op, (void *)(ip + 4))) {
arch/powerpc/kernel/trace/ftrace_64_pg.c
226
if (patch_instruction((u32 *)ip, pop)) {
arch/powerpc/kernel/trace/ftrace_64_pg.c
240
static unsigned long find_ftrace_tramp(unsigned long ip)
arch/powerpc/kernel/trace/ftrace_64_pg.c
251
else if (is_offset_in_branch_range(ftrace_tramps[i] - ip))
arch/powerpc/kernel/trace/ftrace_64_pg.c
328
unsigned long tramp, ip = rec->ip;
arch/powerpc/kernel/trace/ftrace_64_pg.c
332
if (copy_inst_from_kernel_nofault(&op, (void *)ip)) {
arch/powerpc/kernel/trace/ftrace_64_pg.c
344
tramp = find_bl_target(ip, op);
arch/powerpc/kernel/trace/ftrace_64_pg.c
346
pr_devel("ip:%lx jumps to %lx", ip, tramp);
arch/powerpc/kernel/trace/ftrace_64_pg.c
350
if (!find_ftrace_tramp(ip)) {
arch/powerpc/kernel/trace/ftrace_64_pg.c
352
(void *)ip);
arch/powerpc/kernel/trace/ftrace_64_pg.c
357
if (patch_instruction((u32 *)ip, ppc_inst(PPC_RAW_NOP()))) {
arch/powerpc/kernel/trace/ftrace_64_pg.c
368
unsigned long ip = rec->ip;
arch/powerpc/kernel/trace/ftrace_64_pg.c
376
if (test_24bit_addr(ip, addr)) {
arch/powerpc/kernel/trace/ftrace_64_pg.c
378
old = ftrace_call_replace(ip, addr, 1);
arch/powerpc/kernel/trace/ftrace_64_pg.c
380
return ftrace_modify_code(ip, old, new);
arch/powerpc/kernel/trace/ftrace_64_pg.c
381
} else if (core_kernel_text(ip)) {
arch/powerpc/kernel/trace/ftrace_64_pg.c
396
static bool expected_nop_sequence(void *ip, ppc_inst_t op0, ppc_inst_t op1)
arch/powerpc/kernel/trace/ftrace_64_pg.c
409
void *ip = (void *)rec->ip;
arch/powerpc/kernel/trace/ftrace_64_pg.c
417
if (copy_inst_from_kernel_nofault(op, ip))
arch/powerpc/kernel/trace/ftrace_64_pg.c
421
copy_inst_from_kernel_nofault(op + 1, ip + 4))
arch/powerpc/kernel/trace/ftrace_64_pg.c
424
if (!expected_nop_sequence(ip, op[0], op[1])) {
arch/powerpc/kernel/trace/ftrace_64_pg.c
425
pr_err("Unexpected call sequence at %p: %08lx %08lx\n", ip,
arch/powerpc/kernel/trace/ftrace_64_pg.c
456
if (patch_branch(ip, tramp, BRANCH_SET_LINK)) {
arch/powerpc/kernel/trace/ftrace_64_pg.c
46
ftrace_call_replace(unsigned long ip, unsigned long addr, int link)
arch/powerpc/kernel/trace/ftrace_64_pg.c
473
void *ip = (void *)rec->ip;
arch/powerpc/kernel/trace/ftrace_64_pg.c
489
if (copy_inst_from_kernel_nofault(&op, ip)) {
arch/powerpc/kernel/trace/ftrace_64_pg.c
490
pr_err("Unable to read ftrace location %p\n", ip);
arch/powerpc/kernel/trace/ftrace_64_pg.c
496
ip, ppc_inst_as_ulong(op));
arch/powerpc/kernel/trace/ftrace_64_pg.c
500
tramp = find_ftrace_tramp((unsigned long)ip);
arch/powerpc/kernel/trace/ftrace_64_pg.c
502
pr_err("No ftrace trampolines reachable from %ps\n", ip);
arch/powerpc/kernel/trace/ftrace_64_pg.c
506
if (patch_branch(ip, tramp, BRANCH_SET_LINK)) {
arch/powerpc/kernel/trace/ftrace_64_pg.c
516
unsigned long ip = rec->ip;
arch/powerpc/kernel/trace/ftrace_64_pg.c
524
if (test_24bit_addr(ip, addr)) {
arch/powerpc/kernel/trace/ftrace_64_pg.c
527
new = ftrace_call_replace(ip, addr, 1);
arch/powerpc/kernel/trace/ftrace_64_pg.c
528
return ftrace_modify_code(ip, old, new);
arch/powerpc/kernel/trace/ftrace_64_pg.c
529
} else if (core_kernel_text(ip)) {
arch/powerpc/kernel/trace/ftrace_64_pg.c
53
create_branch(&op, (u32 *)ip, addr, link ? BRANCH_SET_LINK : 0);
arch/powerpc/kernel/trace/ftrace_64_pg.c
546
unsigned long ip = rec->ip;
arch/powerpc/kernel/trace/ftrace_64_pg.c
560
if (copy_inst_from_kernel_nofault(&op, (void *)ip)) {
arch/powerpc/kernel/trace/ftrace_64_pg.c
572
tramp = find_bl_target(ip, op);
arch/powerpc/kernel/trace/ftrace_64_pg.c
575
pr_devel("ip:%lx jumps to %lx", ip, tramp);
arch/powerpc/kernel/trace/ftrace_64_pg.c
59
ftrace_modify_code(unsigned long ip, ppc_inst_t old, ppc_inst_t new)
arch/powerpc/kernel/trace/ftrace_64_pg.c
594
if (test_24bit_addr(ip, addr)) {
arch/powerpc/kernel/trace/ftrace_64_pg.c
596
if (patch_branch((u32 *)ip, addr, BRANCH_SET_LINK)) {
arch/powerpc/kernel/trace/ftrace_64_pg.c
623
if (patch_branch((u32 *)ip, tramp, BRANCH_SET_LINK)) {
arch/powerpc/kernel/trace/ftrace_64_pg.c
640
unsigned long ip = rec->ip;
arch/powerpc/kernel/trace/ftrace_64_pg.c
648
if (test_24bit_addr(ip, addr) && test_24bit_addr(ip, old_addr)) {
arch/powerpc/kernel/trace/ftrace_64_pg.c
650
old = ftrace_call_replace(ip, old_addr, 1);
arch/powerpc/kernel/trace/ftrace_64_pg.c
651
new = ftrace_call_replace(ip, addr, 1);
arch/powerpc/kernel/trace/ftrace_64_pg.c
652
return ftrace_modify_code(ip, old, new);
arch/powerpc/kernel/trace/ftrace_64_pg.c
653
} else if (core_kernel_text(ip)) {
arch/powerpc/kernel/trace/ftrace_64_pg.c
670
unsigned long ip = (unsigned long)(&ftrace_call);
arch/powerpc/kernel/trace/ftrace_64_pg.c
675
new = ftrace_call_replace(ip, (unsigned long)func, 1);
arch/powerpc/kernel/trace/ftrace_64_pg.c
676
ret = ftrace_modify_code(ip, old, new);
arch/powerpc/kernel/trace/ftrace_64_pg.c
680
ip = (unsigned long)(&ftrace_regs_call);
arch/powerpc/kernel/trace/ftrace_64_pg.c
682
new = ftrace_call_replace(ip, (unsigned long)func, 1);
arch/powerpc/kernel/trace/ftrace_64_pg.c
683
ret = ftrace_modify_code(ip, old, new);
arch/powerpc/kernel/trace/ftrace_64_pg.c
72
if (copy_inst_from_kernel_nofault(&replaced, (void *)ip))
arch/powerpc/kernel/trace/ftrace_64_pg.c
759
unsigned long ip = (unsigned long)(&ftrace_graph_call);
arch/powerpc/kernel/trace/ftrace_64_pg.c
767
old = ftrace_call_replace(ip, enable ? stub : addr, 0);
arch/powerpc/kernel/trace/ftrace_64_pg.c
768
new = ftrace_call_replace(ip, enable ? addr : stub, 0);
arch/powerpc/kernel/trace/ftrace_64_pg.c
77
pr_err("%p: replaced (%08lx) != old (%08lx)", (void *)ip,
arch/powerpc/kernel/trace/ftrace_64_pg.c
770
return ftrace_modify_code(ip, old, new);
arch/powerpc/kernel/trace/ftrace_64_pg.c
788
__prepare_ftrace_return(unsigned long parent, unsigned long ip, unsigned long sp,
arch/powerpc/kernel/trace/ftrace_64_pg.c
801
if (!function_graph_enter_regs(parent, ip, 0, (unsigned long *)sp, fregs))
arch/powerpc/kernel/trace/ftrace_64_pg.c
809
void ftrace_graph_func(unsigned long ip, unsigned long parent_ip,
arch/powerpc/kernel/trace/ftrace_64_pg.c
812
arch_ftrace_regs(fregs)->regs.link = __prepare_ftrace_return(parent_ip, ip,
arch/powerpc/kernel/trace/ftrace_64_pg.c
816
unsigned long prepare_ftrace_return(unsigned long parent, unsigned long ip,
arch/powerpc/kernel/trace/ftrace_64_pg.c
819
return __prepare_ftrace_return(parent, ip, sp, NULL);
arch/powerpc/kernel/trace/ftrace_64_pg.c
83
return patch_instruction((u32 *)ip, new);
arch/powerpc/kernel/trace/ftrace_64_pg.c
89
static int test_24bit_addr(unsigned long ip, unsigned long addr)
arch/powerpc/kernel/trace/ftrace_64_pg.c
93
return is_offset_in_branch_range(addr - ip);
arch/powerpc/kvm/book3s_hv.c
164
int *ip)
arch/powerpc/kvm/book3s_hv.c
166
int i = *ip;
arch/powerpc/kvm/book3s_hv.c
172
*ip = i;
arch/powerpc/net/bpf_jit_comp.c
1102
static int bpf_modify_inst(void *ip, ppc_inst_t old_inst, ppc_inst_t new_inst)
arch/powerpc/net/bpf_jit_comp.c
1106
if (copy_inst_from_kernel_nofault(&org_inst, ip)) {
arch/powerpc/net/bpf_jit_comp.c
1107
pr_err("0x%lx: fetching instruction failed\n", (unsigned long)ip);
arch/powerpc/net/bpf_jit_comp.c
1113
(unsigned long)ip, ppc_inst_as_ulong(old_inst), ppc_inst_as_ulong(org_inst));
arch/powerpc/net/bpf_jit_comp.c
1120
return patch_instruction(ip, new_inst);
arch/powerpc/net/bpf_jit_comp.c
1167
int bpf_arch_text_poke(void *ip, enum bpf_text_poke_type old_t,
arch/powerpc/net/bpf_jit_comp.c
1179
bpf_func = (unsigned long)ip;
arch/powerpc/net/bpf_jit_comp.c
1199
if (create_branch(&old_inst, ip, (unsigned long)old_addr, 0))
arch/powerpc/net/bpf_jit_comp.c
1203
if (create_branch(&new_inst, ip, (unsigned long)new_addr, 0))
arch/powerpc/net/bpf_jit_comp.c
1206
ret = bpf_modify_inst(ip, old_inst, new_inst);
arch/powerpc/net/bpf_jit_comp.c
1217
ip = (void *)(bpf_func_end - bpf_jit_ool_stub + 4);
arch/powerpc/net/bpf_jit_comp.c
1219
if (!is_offset_in_branch_range((long)ip - 4 - bpf_func)) {
arch/powerpc/net/bpf_jit_comp.c
1228
if (is_offset_in_branch_range(ip - old_addr))
arch/powerpc/net/bpf_jit_comp.c
1229
create_branch(&old_inst, ip, (unsigned long)old_addr, branch_flags);
arch/powerpc/net/bpf_jit_comp.c
1231
create_branch(&old_inst, ip, bpf_func_end - bpf_jit_long_branch_stub,
arch/powerpc/net/bpf_jit_comp.c
1237
if (is_offset_in_branch_range(ip - new_addr))
arch/powerpc/net/bpf_jit_comp.c
1238
create_branch(&new_inst, ip, (unsigned long)new_addr, branch_flags);
arch/powerpc/net/bpf_jit_comp.c
1240
create_branch(&new_inst, ip, bpf_func_end - bpf_jit_long_branch_stub,
arch/powerpc/net/bpf_jit_comp.c
1251
if ((new_addr && !is_offset_in_branch_range(new_addr - ip)) ||
arch/powerpc/net/bpf_jit_comp.c
1252
(old_addr && !is_offset_in_branch_range(old_addr - ip)))
arch/powerpc/net/bpf_jit_comp.c
1254
(new_addr && !is_offset_in_branch_range(new_addr - ip)) ?
arch/powerpc/net/bpf_jit_comp.c
1260
ret = bpf_modify_inst(ip, old_inst, new_inst);
arch/powerpc/net/bpf_jit_comp.c
1265
ip = (void *)bpf_func;
arch/powerpc/net/bpf_jit_comp.c
1269
create_branch(&new_inst, ip, bpf_func_end - bpf_jit_ool_stub, 0);
arch/powerpc/net/bpf_jit_comp.c
1272
create_branch(&old_inst, ip, bpf_func_end - bpf_jit_ool_stub, 0);
arch/powerpc/net/bpf_jit_comp.c
1274
ret = bpf_modify_inst(ip, old_inst, new_inst);
arch/powerpc/net/bpf_jit_comp64.c
383
unsigned long ip;
arch/powerpc/net/bpf_jit_comp64.c
388
ip = stack[STACK_FRAME_LR_SAVE];
arch/powerpc/net/bpf_jit_comp64.c
389
if (!ip)
arch/powerpc/net/bpf_jit_comp64.c
397
if (ip && !consume_fn(cookie, ip, sp, sp))
arch/powerpc/perf/imc-pmu.c
1294
data->ip = be64_to_cpu(READ_ONCE(mem->ip));
arch/powerpc/perf/imc-pmu.c
1320
if (is_kernel_addr(data->ip))
arch/powerpc/platforms/pasemi/iommu.c
109
u32 *ip;
arch/powerpc/platforms/pasemi/iommu.c
116
ip = ((u32 *)tbl->it_base) + index;
arch/powerpc/platforms/pasemi/iommu.c
119
*(ip++) = iob_l2_emptyval;
arch/powerpc/platforms/pasemi/iommu.c
82
u32 *ip;
arch/powerpc/platforms/pasemi/iommu.c
90
ip = ((u32 *)tbl->it_base) + index;
arch/powerpc/platforms/pasemi/iommu.c
95
*(ip++) = IOBMAP_L2E_V | rpn;
arch/powerpc/xmon/xmon.c
1709
unsigned long ip;
arch/powerpc/xmon/xmon.c
1721
if (!mread(sp + LRSAVE_OFFSET, &ip, sizeof(unsigned long))
arch/powerpc/xmon/xmon.c
1742
if (lr == ip) {
arch/powerpc/xmon/xmon.c
1755
xmon_print_symbol(ip, " ", " (unreliable)\n");
arch/powerpc/xmon/xmon.c
1761
xmon_print_symbol(ip, " ", "\n");
arch/riscv/include/asm/ftrace.h
225
void ftrace_graph_func(unsigned long ip, unsigned long parent_ip,
arch/riscv/kernel/ftrace.c
119
unsigned long distance, orig_addr, pc = rec->ip - MCOUNT_AUIPC_SIZE;
arch/riscv/kernel/ftrace.c
143
if (patch_insn_write((void *)rec->ip, &nop4, MCOUNT_NOP4_SIZE))
arch/riscv/kernel/ftrace.c
158
unsigned long pc = rec->ip - MCOUNT_AUIPC_SIZE;
arch/riscv/kernel/ftrace.c
215
unsigned long caller = rec->ip - MCOUNT_AUIPC_SIZE;
arch/riscv/kernel/ftrace.c
250
void ftrace_graph_func(unsigned long ip, unsigned long parent_ip,
arch/riscv/kernel/ftrace.c
267
if (!function_graph_enter_regs(old, ip, frame_pointer, parent, fregs))
arch/riscv/kernel/ftrace.c
98
unsigned long literal = ALIGN_DOWN(rec->ip - 12, 8);
arch/riscv/net/bpf_jit_comp64.c
172
u64 ip = (u64)(ctx->ro_insns + ctx->ninsns);
arch/riscv/net/bpf_jit_comp64.c
173
s64 off = addr - ip;
arch/riscv/net/bpf_jit_comp64.c
456
u64 ip;
arch/riscv/net/bpf_jit_comp64.c
463
ip = (u64)(long)(ctx->ro_insns + ctx->ninsns);
arch/riscv/net/bpf_jit_comp64.c
464
off = addr - ip;
arch/riscv/net/bpf_jit_comp64.c
837
static int gen_jump_or_nops(void *target, void *ip, u32 *insns, bool is_call)
arch/riscv/net/bpf_jit_comp64.c
851
rvoff = (s64)(target - ip);
arch/riscv/net/bpf_jit_comp64.c
855
int bpf_arch_text_poke(void *ip, enum bpf_text_poke_type old_t,
arch/riscv/net/bpf_jit_comp64.c
863
if (!is_kernel_text((unsigned long)ip) &&
arch/riscv/net/bpf_jit_comp64.c
864
!is_bpf_text_address((unsigned long)ip))
arch/riscv/net/bpf_jit_comp64.c
868
ret = gen_jump_or_nops(old_addr, ip, old_insns, is_call);
arch/riscv/net/bpf_jit_comp64.c
872
if (memcmp(ip, old_insns, RV_FENTRY_NBYTES))
arch/riscv/net/bpf_jit_comp64.c
876
ret = gen_jump_or_nops(new_addr, ip, new_insns, is_call);
arch/riscv/net/bpf_jit_comp64.c
882
if (memcmp(ip, new_insns, RV_FENTRY_NBYTES))
arch/riscv/net/bpf_jit_comp64.c
883
ret = patch_text(ip, new_insns, RV_FENTRY_NBYTES);
arch/s390/boot/printk.c
124
static noinline char *findsym(unsigned long ip, unsigned short *off, unsigned short *len)
arch/s390/boot/printk.c
138
if (ip < start) {
arch/s390/boot/printk.c
142
if (ip > start + size) {
arch/s390/boot/printk.c
146
*off = ip - start;
arch/s390/boot/printk.c
154
static noinline char *strsym(char *buf, void *ip)
arch/s390/boot/printk.c
160
p = findsym((unsigned long)ip, &off, &len);
arch/s390/boot/printk.c
170
as_hex(buf, (unsigned long)ip, 16);
arch/s390/include/asm/ftrace.h
116
void ftrace_graph_func(unsigned long ip, unsigned long parent_ip,
arch/s390/include/asm/ftrace.h
69
unsigned long ip)
arch/s390/include/asm/ftrace.h
71
arch_ftrace_regs(fregs)->regs.psw.addr = ip;
arch/s390/include/asm/unwind.h
39
unsigned long sp, ip;
arch/s390/include/asm/unwind.h
48
unsigned long ip)
arch/s390/include/asm/unwind.h
50
ip = ftrace_graph_ret_addr(state->task, &state->graph_idx, ip, (void *)state->sp);
arch/s390/include/asm/unwind.h
52
if (is_rethook_trampoline(ip))
arch/s390/include/asm/unwind.h
53
ip = rethook_find_ret_addr(state->task, state->sp, &state->kr_cur);
arch/s390/include/asm/unwind.h
55
return ip;
arch/s390/include/uapi/asm/qeth.h
37
__u8 ip;
arch/s390/kernel/dumpstack.c
139
loglvl, state.ip, (void *) state.ip);
arch/s390/kernel/early.c
156
unsigned long ip;
arch/s390/kernel/early.c
161
ip = __rewind_psw(regs->psw, regs->int_code >> 16);
arch/s390/kernel/early.c
165
if (report_bug(ip, regs) == BUG_TRAP_TYPE_WARN)
arch/s390/kernel/ftrace.c
107
if (copy_from_kernel_nofault(&old, (void *)rec->ip, sizeof(old)))
arch/s390/kernel/ftrace.c
117
tmp.rest_of_intercepted_function = rec->ip + sizeof(struct ftrace_insn);
arch/s390/kernel/ftrace.c
121
disp = ((char *)trampoline - (char *)rec->ip) / 2;
arch/s390/kernel/ftrace.c
122
insn = (struct ftrace_insn *)rec->ip;
arch/s390/kernel/ftrace.c
135
if (copy_from_kernel_nofault(&insn, (void *)rec->ip, sizeof(insn)))
arch/s390/kernel/ftrace.c
138
trampoline = (void *)(rec->ip + disp);
arch/s390/kernel/ftrace.c
147
ftrace_generate_branch_insn(unsigned long ip, unsigned long target)
arch/s390/kernel/ftrace.c
151
.disp = target ? (target - ip) / 2 : 0 };
arch/s390/kernel/ftrace.c
154
static int ftrace_patch_branch_insn(unsigned long ip, unsigned long old_target,
arch/s390/kernel/ftrace.c
157
struct ftrace_insn orig = ftrace_generate_branch_insn(ip, old_target);
arch/s390/kernel/ftrace.c
158
struct ftrace_insn new = ftrace_generate_branch_insn(ip, target);
arch/s390/kernel/ftrace.c
161
if (!IS_ALIGNED(ip, 8))
arch/s390/kernel/ftrace.c
163
if (copy_from_kernel_nofault(&old, (void *)ip, sizeof(old)))
arch/s390/kernel/ftrace.c
168
s390_kernel_write((void *)ip, &new, sizeof(new));
arch/s390/kernel/ftrace.c
194
return ftrace_patch_branch_insn(rec->ip, old_addr, addr);
arch/s390/kernel/ftrace.c
219
return ftrace_patch_branch_insn(rec->ip, addr, 0);
arch/s390/kernel/ftrace.c
221
return ftrace_patch_branch_mask((void *)rec->ip, 0xc0f4, false);
arch/s390/kernel/ftrace.c
233
return ftrace_patch_branch_mask((void *)rec->ip, 0xc004, true);
arch/s390/kernel/ftrace.c
239
return ftrace_patch_branch_insn(rec->ip, 0, addr);
arch/s390/kernel/ftrace.c
266
void ftrace_graph_func(unsigned long ip, unsigned long parent_ip,
arch/s390/kernel/ftrace.c
276
if (!function_graph_enter_regs(*parent, ip, 0, (unsigned long *)sp, fregs))
arch/s390/kernel/ftrace.c
283
void kprobe_ftrace_handler(unsigned long ip, unsigned long parent_ip,
arch/s390/kernel/ftrace.c
294
bit = ftrace_test_recursion_trylock(ip, parent_ip);
arch/s390/kernel/ftrace.c
300
p = get_kprobe((kprobe_opcode_t *)ip);
arch/s390/kernel/ftrace.c
314
instruction_pointer_set(regs, ip);
arch/s390/kernel/ftrace.c
318
instruction_pointer_set(regs, ip + MCOUNT_INSN_SIZE);
arch/s390/kernel/kprobes.c
190
unsigned long ip)
arch/s390/kernel/kprobes.c
203
per_kprobe.start.val = ip;
arch/s390/kernel/kprobes.c
204
per_kprobe.end.val = ip;
arch/s390/kernel/kprobes.c
215
regs->psw.addr = ip;
arch/s390/kernel/kprobes.c
221
unsigned long ip)
arch/s390/kernel/kprobes.c
227
regs->psw.addr = ip;
arch/s390/kernel/kprobes.c
348
unsigned long ip = regs->psw.addr;
arch/s390/kernel/kprobes.c
352
ip += (unsigned long) p->addr - (unsigned long) p->ainsn.insn;
arch/s390/kernel/kprobes.c
356
if (ip - (unsigned long) p->ainsn.insn == ilen)
arch/s390/kernel/kprobes.c
357
ip = (unsigned long) p->addr + ilen;
arch/s390/kernel/kprobes.c
366
disable_singlestep(kcb, regs, ip);
arch/s390/kernel/module.c
336
char *ip;
arch/s390/kernel/module.c
339
ip = plt_base + info->plt_offset;
arch/s390/kernel/module.c
349
*(int *)&insn[10] = (jump_r1 - (ip + 8)) / 2;
arch/s390/kernel/module.c
355
write(ip, insn, sizeof(insn));
arch/s390/kernel/process.c
204
unsigned long ip = 0;
arch/s390/kernel/process.c
214
ip = 0;
arch/s390/kernel/process.c
218
ip = unwind_get_return_address(&state);
arch/s390/kernel/process.c
219
if (!ip)
arch/s390/kernel/process.c
222
if (!in_sched_functions(ip))
arch/s390/kernel/process.c
227
return ip;
arch/s390/kernel/stacktrace.c
106
unsigned long ip, sp;
arch/s390/kernel/stacktrace.c
110
ip = instruction_pointer(regs);
arch/s390/kernel/stacktrace.c
111
if (!store_ip(consume_entry, cookie, entry, perf, ip))
arch/s390/kernel/stacktrace.c
122
if (!sp && ip_within_vdso(ip)) {
arch/s390/kernel/stacktrace.c
124
if (__get_user(ip, &sf_vdso->return_address))
arch/s390/kernel/stacktrace.c
132
if (__get_user(ip, &sf->gprs[8]))
arch/s390/kernel/stacktrace.c
136
if (sp & 0x7 || ip_invalid(ip))
arch/s390/kernel/stacktrace.c
138
if (!store_ip(consume_entry, cookie, entry, perf, ip))
arch/s390/kernel/stacktrace.c
52
if (state.ip == (unsigned long)arch_rethook_trampoline)
arch/s390/kernel/stacktrace.c
68
unsigned long ip)
arch/s390/kernel/stacktrace.c
72
if (perf_callchain_store(entry, ip))
arch/s390/kernel/stacktrace.c
77
return consume_entry(cookie, ip);
arch/s390/kernel/stacktrace.c
80
static inline bool ip_invalid(unsigned long ip)
arch/s390/kernel/stacktrace.c
86
if (ip & 1)
arch/s390/kernel/stacktrace.c
88
if (ip < mmap_min_addr)
arch/s390/kernel/stacktrace.c
90
if (ip >= current->mm->context.asce_limit)
arch/s390/kernel/stacktrace.c
95
static inline bool ip_within_vdso(unsigned long ip)
arch/s390/kernel/stacktrace.c
97
return in_range(ip, current->mm->context.vdso_base, vdso_text_size());
arch/s390/kernel/unwind_bc.c
114
state->ip = unwind_recover_ret_addr(state, ip);
arch/s390/kernel/unwind_bc.c
132
unsigned long ip, sp;
arch/s390/kernel/unwind_bc.c
146
ip = regs->psw.addr;
arch/s390/kernel/unwind_bc.c
165
ip = READ_ONCE_NOCHECK(sf->gprs[8]);
arch/s390/kernel/unwind_bc.c
171
state->ip = unwind_recover_ret_addr(state, ip);
arch/s390/kernel/unwind_bc.c
18
return __kernel_text_address(state->ip) ? state->ip : 0;
arch/s390/kernel/unwind_bc.c
61
unsigned long sp, ip;
arch/s390/kernel/unwind_bc.c
68
ip = READ_ONCE_NOCHECK(sf->gprs[8]);
arch/s390/kernel/unwind_bc.c
72
if (!__kernel_text_address(ip) || state->ip == unwind_recover_ret_addr(state, ip)) {
arch/s390/kernel/unwind_bc.c
86
ip = READ_ONCE_NOCHECK(sf->gprs[8]);
arch/s390/kernel/unwind_bc.c
96
ip = READ_ONCE_NOCHECK(regs->psw.addr);
arch/s390/lib/test_unwind.c
268
static void notrace __used test_unwind_ftrace_handler(unsigned long ip,
arch/s390/lib/test_unwind.c
84
(void *)state.sp, (void *)state.ip);
arch/s390/net/bpf_jit_comp.c
2415
int bpf_arch_text_poke(void *ip, enum bpf_text_poke_type old_t,
arch/s390/net/bpf_jit_comp.c
2428
err = copy_from_kernel_nofault(&insn, ip, sizeof(insn));
arch/s390/net/bpf_jit_comp.c
2435
insn.disp == ((char *)new_addr - (char *)ip) >> 1) {
arch/s390/net/bpf_jit_comp.c
2442
plt = ip + (insn.disp << 1);
arch/s390/net/bpf_jit_comp.c
2447
ret = (char *)ip + 6;
arch/s390/net/bpf_jit_comp.c
2459
s390_kernel_write((char *)ip + 1, (char *)&insn.opc + 1, 1);
arch/sh/kernel/cpu/irq/imask.c
31
static inline void set_interrupt_registers(int ip)
arch/sh/kernel/cpu/irq/imask.c
50
: "r" (~0xf0), "r" (ip << 4)
arch/sh/kernel/ftrace.c
168
do_ftrace_mod_code(unsigned long ip, void *new_code)
arch/sh/kernel/ftrace.c
170
mod_code_ip = (void *)ip;
arch/sh/kernel/ftrace.c
192
static int ftrace_modify_code(unsigned long ip, unsigned char *old_code,
arch/sh/kernel/ftrace.c
206
if (copy_from_kernel_nofault(replaced, (void *)ip, MCOUNT_INSN_SIZE))
arch/sh/kernel/ftrace.c
214
if (do_ftrace_mod_code(ip, new_code))
arch/sh/kernel/ftrace.c
217
flush_icache_range(ip, ip + MCOUNT_INSN_SIZE);
arch/sh/kernel/ftrace.c
224
unsigned long ip = (unsigned long)(&ftrace_call) + MCOUNT_INSN_OFFSET;
arch/sh/kernel/ftrace.c
227
memcpy(old, (unsigned char *)ip, MCOUNT_INSN_SIZE);
arch/sh/kernel/ftrace.c
228
new = ftrace_call_replace(ip, (unsigned long)func);
arch/sh/kernel/ftrace.c
230
return ftrace_modify_code(ip, old, new);
arch/sh/kernel/ftrace.c
237
unsigned long ip = rec->ip;
arch/sh/kernel/ftrace.c
239
old = ftrace_call_replace(ip, addr);
arch/sh/kernel/ftrace.c
240
new = ftrace_nop_replace(ip);
arch/sh/kernel/ftrace.c
242
return ftrace_modify_code(rec->ip, old, new);
arch/sh/kernel/ftrace.c
248
unsigned long ip = rec->ip;
arch/sh/kernel/ftrace.c
250
old = ftrace_nop_replace(ip);
arch/sh/kernel/ftrace.c
251
new = ftrace_call_replace(ip, addr);
arch/sh/kernel/ftrace.c
253
return ftrace_modify_code(rec->ip, old, new);
arch/sh/kernel/ftrace.c
261
static int ftrace_mod(unsigned long ip, unsigned long old_addr,
arch/sh/kernel/ftrace.c
266
if (copy_from_kernel_nofault(code, (void *)ip, MCOUNT_INSN_SIZE))
arch/sh/kernel/ftrace.c
272
__raw_writel(new_addr, ip);
arch/sh/kernel/ftrace.c
278
unsigned long ip, old_addr, new_addr;
arch/sh/kernel/ftrace.c
280
ip = (unsigned long)(&ftrace_graph_call) + GRAPH_INSN_OFFSET;
arch/sh/kernel/ftrace.c
284
return ftrace_mod(ip, old_addr, new_addr);
arch/sh/kernel/ftrace.c
289
unsigned long ip, old_addr, new_addr;
arch/sh/kernel/ftrace.c
291
ip = (unsigned long)(&ftrace_graph_call) + GRAPH_INSN_OFFSET;
arch/sh/kernel/ftrace.c
295
return ftrace_mod(ip, old_addr, new_addr);
arch/sh/kernel/ftrace.c
48
static unsigned char *ftrace_nop_replace(unsigned long ip)
arch/sh/kernel/ftrace.c
50
__raw_writel(ip + MCOUNT_INSN_SIZE, ftrace_nop);
arch/sh/kernel/ftrace.c
54
static unsigned char *ftrace_call_replace(unsigned long ip, unsigned long addr)
arch/sh/kernel/kgdb.c
296
void kgdb_arch_set_pc(struct pt_regs *regs, unsigned long ip)
arch/sh/kernel/kgdb.c
298
regs->pc = ip;
arch/sparc/include/asm/perf_event.h
23
(regs)->tpc = (ip); \
arch/sparc/include/asm/perf_event.h
8
#define perf_arch_fetch_caller_regs(regs, ip) \
arch/sparc/kernel/ftrace.c
104
unsigned long ip = (unsigned long)(&ftrace_graph_call);
arch/sparc/kernel/ftrace.c
108
new = ftrace_call_replace(ip, (unsigned long) &ftrace_stub);
arch/sparc/kernel/ftrace.c
110
return ftrace_modify_code(ip, old, new);
arch/sparc/kernel/ftrace.c
15
static u32 ftrace_call_replace(unsigned long ip, unsigned long addr)
arch/sparc/kernel/ftrace.c
20
off = ((s32)addr - (s32)ip);
arch/sparc/kernel/ftrace.c
26
static int ftrace_modify_code(unsigned long ip, u32 old, u32 new)
arch/sparc/kernel/ftrace.c
47
: [new] "0" (new), [old] "r" (old), [ip] "r" (ip)
arch/sparc/kernel/ftrace.c
58
unsigned long ip = rec->ip;
arch/sparc/kernel/ftrace.c
61
old = ftrace_call_replace(ip, addr);
arch/sparc/kernel/ftrace.c
63
return ftrace_modify_code(ip, old, new);
arch/sparc/kernel/ftrace.c
68
unsigned long ip = rec->ip;
arch/sparc/kernel/ftrace.c
72
new = ftrace_call_replace(ip, addr);
arch/sparc/kernel/ftrace.c
73
return ftrace_modify_code(ip, old, new);
arch/sparc/kernel/ftrace.c
78
unsigned long ip = (unsigned long)(&ftrace_call);
arch/sparc/kernel/ftrace.c
82
new = ftrace_call_replace(ip, (unsigned long)func);
arch/sparc/kernel/ftrace.c
83
return ftrace_modify_code(ip, old, new);
arch/sparc/kernel/ftrace.c
94
unsigned long ip = (unsigned long)(&ftrace_graph_call);
arch/sparc/kernel/ftrace.c
98
new = ftrace_call_replace(ip, (unsigned long) &ftrace_graph_caller);
arch/sparc/kernel/ftrace.c
99
return ftrace_modify_code(ip, old, new);
arch/sparc/kernel/kgdb_32.c
163
void kgdb_arch_set_pc(struct pt_regs *regs, unsigned long ip)
arch/sparc/kernel/kgdb_32.c
165
regs->pc = ip;
arch/sparc/kernel/kgdb_64.c
192
void kgdb_arch_set_pc(struct pt_regs *regs, unsigned long ip)
arch/sparc/kernel/kgdb_64.c
194
regs->tpc = ip;
arch/sparc/kernel/of_device_64.c
539
struct device_node *pp, *ip;
arch/sparc/kernel/of_device_64.c
566
ip = NULL;
arch/sparc/kernel/of_device_64.c
590
ip = iret;
arch/sparc/kernel/of_device_64.c
608
ip = pp;
arch/sparc/kernel/of_device_64.c
615
if (!ip)
arch/sparc/kernel/of_device_64.c
618
irq = ip->irq_trans->irq_build(op->dev.of_node, irq,
arch/sparc/kernel/of_device_64.c
619
ip->irq_trans->data);
arch/sparc/kernel/of_device_64.c
622
op->dev.of_node, ip, orig_irq, irq);
arch/um/include/shared/kern_util.h
27
extern unsigned long segv(struct faultinfo fi, unsigned long ip,
arch/um/include/shared/kern_util.h
30
extern int handle_page_fault(unsigned long address, unsigned long ip,
arch/um/kernel/process.c
281
unsigned long stack_page, sp, ip;
arch/um/kernel/process.c
298
ip = *((unsigned long *) sp);
arch/um/kernel/process.c
299
if (in_sched_functions(ip))
arch/um/kernel/process.c
302
else if (kernel_text_address(ip) && seen_sched)
arch/um/kernel/process.c
303
return ip;
arch/um/kernel/trap.c
140
int handle_page_fault(unsigned long address, unsigned long ip,
arch/um/kernel/trap.c
260
static void bad_segv(struct faultinfo fi, unsigned long ip)
arch/um/kernel/trap.c
308
unsigned long segv(struct faultinfo fi, unsigned long ip, int is_user,
arch/um/kernel/trap.c
352
address, ip);
arch/um/kernel/trap.c
356
err = handle_page_fault(address, ip, is_write, is_user,
arch/um/kernel/trap.c
370
else if (!is_user && arch_fixup(ip, regs))
arch/um/kernel/trap.c
376
address, ip);
arch/x86/boot/compressed/ident_map_64.c
346
unsigned long address, unsigned long ip)
arch/x86/boot/compressed/ident_map_64.c
355
error_puthex(ip - (unsigned long)_head);
arch/x86/boot/compressed/ident_map_64.c
379
do_pf_error("Unexpected page-fault:", error_code, address, regs->ip);
arch/x86/boot/compressed/ident_map_64.c
381
do_pf_error("Page-fault on GHCB page:", error_code, address, regs->ip);
arch/x86/boot/compressed/sev-handle-vc.c
48
memcpy(buffer, (unsigned char *)ctxt->regs->ip, MAX_INSN_SIZE);
arch/x86/boot/printf.c
230
long *ip = va_arg(args, long *);
arch/x86/boot/printf.c
231
*ip = (str - buf);
arch/x86/boot/printf.c
233
int *ip = va_arg(args, int *);
arch/x86/boot/printf.c
234
*ip = (str - buf);
arch/x86/boot/startup/sev-shared.c
446
u16 opcode = *(unsigned short *)regs->ip;
arch/x86/boot/startup/sev-shared.c
511
regs->ip += 2;
arch/x86/coco/sev/vc-handle.c
1054
exit_code, regs->ip);
arch/x86/coco/sev/vc-handle.c
1058
exit_code, regs->ip);
arch/x86/coco/sev/vc-handle.c
1062
exit_code, regs->ip);
arch/x86/coco/sev/vc-handle.c
131
return copy_from_kernel_nofault(buffer, (unsigned char *)ctxt->regs->ip, MAX_INSN_SIZE);
arch/x86/coco/sev/vc-handle.c
144
ctxt->fi.cr2 = ctxt->regs->ip;
arch/x86/coco/sev/vc-handle.c
172
ctxt->fi.cr2 = ctxt->regs->ip;
arch/x86/coco/sev/vc-handle.c
917
error_code, regs->ip);
arch/x86/coco/sev/vc-handle.c
922
error_code, regs->ip);
arch/x86/coco/sev/vc-handle.c
927
error_code, regs->ip);
arch/x86/coco/sev/vc-shared.c
122
ctxt->regs->ip += ctxt->insn.length;
arch/x86/coco/sev/vc-shared.c
93
opcode, exit_code, ctxt->regs->ip);
arch/x86/coco/tdx/tdx.c
590
if (copy_from_kernel_nofault(buffer, (void *)regs->ip, MAX_INSN_SIZE))
arch/x86/coco/tdx/tdx.c
777
regs->ip += insn_len;
arch/x86/coco/tdx/tdx.c
874
regs->ip += insn_len;
arch/x86/entry/entry_fred.c
32
fred_event_data(regs), regs->cs, regs->ip);
arch/x86/entry/entry_fred.c
42
fred_event_data(regs), regs->cs, regs->ip);
arch/x86/entry/syscall_32.c
329
regs->ip = landing_pad;
arch/x86/entry/syscall_32.c
346
if (unlikely(regs->ip != landing_pad))
arch/x86/entry/syscall_64.c
113
if (unlikely(regs->cx != regs->ip || regs->r11 != regs->flags))
arch/x86/entry/syscall_64.c
128
if (unlikely(regs->ip >= TASK_SIZE_MAX))
arch/x86/entry/vdso/extable.c
36
if (regs->ip == base + extable[i].insn) {
arch/x86/entry/vdso/extable.c
37
regs->ip = base + extable[i].fixup;
arch/x86/entry/vdso/vma.c
267
if (regs->ip == vdso + image->sym_vdso32_sigreturn_landing_pad ||
arch/x86/entry/vdso/vma.c
268
regs->ip == vdso + image->sym_vdso32_rt_sigreturn_landing_pad)
arch/x86/entry/vdso/vma.c
69
unsigned long ipoffset = regs->ip -
arch/x86/entry/vdso/vma.c
73
regs->ip = new_vma->vm_start + ipoffset;
arch/x86/entry/vsyscall/vsyscall_64.c
131
if (address != regs->ip) {
arch/x86/entry/vsyscall/vsyscall_64.c
227
if ((!tmp && regs->orig_ax != syscall_nr) || regs->ip != address) {
arch/x86/entry/vsyscall/vsyscall_64.c
274
regs->ip = caller;
arch/x86/entry/vsyscall/vsyscall_64.c
80
message, regs->ip, regs->cs,
arch/x86/events/amd/lbr.c
186
br[out].from = sign_ext_branch_ip(entry.from.split.ip);
arch/x86/events/amd/lbr.c
187
br[out].to = sign_ext_branch_ip(entry.to.split.ip);
arch/x86/events/amd/lbr.c
44
u64 ip:58;
arch/x86/events/amd/lbr.c
53
u64 ip:58;
arch/x86/events/amd/lbr.c
91
static __always_inline u64 sign_ext_branch_ip(u64 ip)
arch/x86/events/amd/lbr.c
95
return (u64)(((s64)ip << shift) >> shift);
arch/x86/events/core.c
2864
if (perf_callchain_store(entry, regs->ip))
arch/x86/events/core.c
2989
perf_callchain_store(entry, regs->ip);
arch/x86/events/core.c
3066
return regs->ip + code_segment_base(regs);
arch/x86/events/intel/ds.c
1230
data.ip = at->from;
arch/x86/events/intel/ds.c
2005
unsigned long ip = regs->ip;
arch/x86/events/intel/ds.c
2025
if (kernel_ip(ip) != kernel_ip(to))
arch/x86/events/intel/ds.c
2032
if ((ip - to) > PEBS_FIXUP_SIZE)
arch/x86/events/intel/ds.c
2038
if (ip == to) {
arch/x86/events/intel/ds.c
2043
size = ip - to;
arch/x86/events/intel/ds.c
2044
if (!kernel_ip(ip)) {
arch/x86/events/intel/ds.c
2079
} while (to < ip);
arch/x86/events/intel/ds.c
2081
if (to == ip) {
arch/x86/events/intel/ds.c
2264
set_linear_ip(regs, pebs->ip);
arch/x86/events/intel/ds.c
2279
set_linear_ip(regs, pebs->ip);
arch/x86/events/intel/ds.c
2435
u64 sample_type, u64 ip,
arch/x86/events/intel/ds.c
2439
set_linear_ip(regs, ip);
arch/x86/events/intel/ds.c
2453
set_linear_ip(regs, gprs->ip);
arch/x86/events/intel/ds.c
2535
basic->ip, basic->tsc, retire);
arch/x86/events/intel/ds.c
2658
basic->ip, basic->tsc, retire);
arch/x86/events/intel/ds.c
717
u64 flags, ip;
arch/x86/events/intel/ds.c
725
u64 flags, ip;
arch/x86/events/intel/ds.c
737
u64 flags, ip;
arch/x86/events/intel/ds.c
766
u64 flags, ip;
arch/x86/events/perf_event.h
1346
static inline bool kernel_ip(unsigned long ip)
arch/x86/events/perf_event.h
1349
return ip > PAGE_OFFSET;
arch/x86/events/perf_event.h
1351
return (long)ip < 0;
arch/x86/events/perf_event.h
1367
static inline void set_linear_ip(struct pt_regs *regs, unsigned long ip)
arch/x86/events/perf_event.h
1369
regs->cs = kernel_ip(ip) ? __KERNEL_CS : __USER_CS;
arch/x86/events/perf_event.h
1372
regs->ip = ip;
arch/x86/hyperv/hv_init.c
678
wrmsrq(HV_X64_MSR_CRASH_P2, regs->ip);
arch/x86/include/asm/alternative.h
117
extern int x86_call_depth_emit_accounting(u8 **pprog, void *func, void *ip);
arch/x86/include/asm/alternative.h
128
void *func, void *ip)
arch/x86/include/asm/cpu.h
31
extern bool handle_guest_split_lock(unsigned long ip);
arch/x86/include/asm/cpu.h
42
static inline bool handle_guest_split_lock(unsigned long ip)
arch/x86/include/asm/elf.h
131
pr_reg[12] = regs->ip; \
arch/x86/include/asm/elf.h
207
(pr_reg)[16] = (regs)->ip; \
arch/x86/include/asm/extable.h
40
extern int ex_get_fixup_type(unsigned long ip);
arch/x86/include/asm/ftrace.h
120
void prepare_ftrace_return(unsigned long ip, unsigned long *parent,
arch/x86/include/asm/ftrace.h
65
(_regs)->ip = arch_ftrace_regs(fregs)->regs.ip; \
arch/x86/include/asm/ftrace.h
72
do { arch_ftrace_regs(fregs)->regs.ip = (_ip); } while (0)
arch/x86/include/asm/ftrace.h
83
void ftrace_graph_func(unsigned long ip, unsigned long parent_ip,
arch/x86/include/asm/insn-eval.h
27
int insn_get_effective_ip(struct pt_regs *regs, unsigned long *ip);
arch/x86/include/asm/kexec.h
117
newregs->ip = _THIS_IP_;
arch/x86/include/asm/perf_event.h
477
u64 ip;
arch/x86/include/asm/perf_event.h
500
u64 flags, ip, ax, cx, dx, bx, sp, bp, si, di;
arch/x86/include/asm/perf_event.h
562
u64 ip;
arch/x86/include/asm/perf_event.h
587
u64 flags, ip, ax, cx, dx, bx, sp, bp, si, di;
arch/x86/include/asm/perf_event.h
728
(regs)->ip = (__ip); \
arch/x86/include/asm/processor.h
286
unsigned long ip;
arch/x86/include/asm/processor.h
680
#define KSTK_EIP(task) (task_pt_regs(task)->ip)
arch/x86/include/asm/ptrace.h
142
unsigned long ip;
arch/x86/include/asm/ptrace.h
264
bool ret = (regs->ip >= (unsigned long)entry_SYSCALL_64 &&
arch/x86/include/asm/ptrace.h
265
regs->ip < (unsigned long)entry_SYSCALL_64_safe_stack);
arch/x86/include/asm/ptrace.h
267
ret = ret || (regs->ip >= (unsigned long)entry_SYSRETQ_unsafe_stack &&
arch/x86/include/asm/ptrace.h
268
regs->ip < (unsigned long)entry_SYSRETQ_end);
arch/x86/include/asm/ptrace.h
270
ret = ret || (regs->ip >= (unsigned long)entry_SYSCALL_compat &&
arch/x86/include/asm/ptrace.h
271
regs->ip < (unsigned long)entry_SYSCALL_compat_safe_stack);
arch/x86/include/asm/ptrace.h
272
ret = ret || (regs->ip >= (unsigned long)entry_SYSRETL_compat_unsafe_stack &&
arch/x86/include/asm/ptrace.h
273
regs->ip < (unsigned long)entry_SYSRETL_compat_end);
arch/x86/include/asm/ptrace.h
287
return regs->ip;
arch/x86/include/asm/ptrace.h
293
regs->ip = val;
arch/x86/include/asm/ptrace.h
48
unsigned long ip;
arch/x86/include/asm/text-patching.h
136
void int3_emulate_jmp(struct pt_regs *regs, unsigned long ip)
arch/x86/include/asm/text-patching.h
138
regs->ip = ip;
arch/x86/include/asm/text-patching.h
169
int3_emulate_push(regs, regs->ip - INT3_INSN_SIZE + CALL_INSN_SIZE);
arch/x86/include/asm/text-patching.h
176
unsigned long ip = int3_emulate_pop(regs);
arch/x86/include/asm/text-patching.h
177
int3_emulate_jmp(regs, ip);
arch/x86/include/asm/text-patching.h
208
void int3_emulate_jcc(struct pt_regs *regs, u8 cc, unsigned long ip, unsigned long disp)
arch/x86/include/asm/text-patching.h
211
ip += disp;
arch/x86/include/asm/text-patching.h
213
int3_emulate_jmp(regs, ip);
arch/x86/include/asm/unwind.h
11
#define IRET_FRAME_OFFSET (offsetof(struct pt_regs, ip))
arch/x86/include/asm/unwind.h
25
unsigned long sp, bp, ip;
arch/x86/include/asm/unwind.h
29
unsigned long *bp, *orig_sp, ip;
arch/x86/include/asm/user_32.h
91
unsigned long ip;
arch/x86/include/asm/user_64.h
86
unsigned long ip;
arch/x86/include/asm/xen/interface.h
308
uint64_t ip;
arch/x86/include/asm/xen/interface_64.h
117
__DECL_REG(ip);
arch/x86/include/uapi/asm/mce.h
19
__u64 ip; /* Instruction Pointer when the error happened */
arch/x86/include/uapi/asm/sigcontext.h
217
__u32 ip;
arch/x86/include/uapi/asm/sigcontext.h
255
__u64 ip;
arch/x86/kernel/alternative.c
1952
unsigned long addr = regs->ip - fineibt_preamble_ud;
arch/x86/kernel/alternative.c
1967
regs->ip = *target;
arch/x86/kernel/alternative.c
1986
if (regs->ip < (unsigned long)__bhi_args ||
arch/x86/kernel/alternative.c
1987
regs->ip >= (unsigned long)__bhi_args_end)
arch/x86/kernel/alternative.c
2033
unsigned long addr = regs->ip - fineibt_paranoid_ud;
arch/x86/kernel/alternative.c
2067
if (is_paranoid_thunk(regs->ip)) {
arch/x86/kernel/alternative.c
2071
regs->ip = *target;
arch/x86/kernel/alternative.c
2324
if (regs->ip - INT3_INSN_SIZE != selftest)
arch/x86/kernel/alternative.c
2830
void *ip;
arch/x86/kernel/alternative.c
2851
ip = (void *) regs->ip - INT3_INSN_SIZE;
arch/x86/kernel/alternative.c
2857
tpl = __inline_bsearch(ip, text_poke_array.vec, text_poke_array.nr_entries,
arch/x86/kernel/alternative.c
2864
if (text_poke_addr(tpl) != ip)
arch/x86/kernel/alternative.c
2868
ip += tpl->len;
arch/x86/kernel/alternative.c
2883
int3_emulate_call(regs, (long)ip + tpl->disp);
arch/x86/kernel/alternative.c
2888
int3_emulate_jmp(regs, (long)ip + tpl->disp);
arch/x86/kernel/alternative.c
2892
int3_emulate_jcc(regs, tpl->opcode & 0xf, (long)ip, tpl->disp);
arch/x86/kernel/apic/io_apic.c
2217
struct ioapic *ip = &ioapics[ioapic];
arch/x86/kernel/apic/io_apic.c
2218
struct ioapic_domain_cfg *cfg = &ip->irqdomain_cfg;
arch/x86/kernel/apic/io_apic.c
2246
ip->irqdomain = irq_domain_create_hierarchy(parent, 0, hwirqs, fn, cfg->ops,
arch/x86/kernel/apic/io_apic.c
2248
if (!ip->irqdomain) {
arch/x86/kernel/asm-offsets.c
65
OFFSET(IA32_SIGCONTEXT_ip, sigcontext_32, ip);
arch/x86/kernel/asm-offsets_32.c
27
OFFSET(PT_EIP, pt_regs, ip);
arch/x86/kernel/callthunks.c
301
int x86_call_depth_emit_accounting(u8 **pprog, void *func, void *ip)
arch/x86/kernel/callthunks.c
314
text_poke_apply_relocation(insn_buff, ip, tmpl_size, skl_call_thunk_template, tmpl_size);
arch/x86/kernel/cet.c
120
if (unlikely(regs->ip == (unsigned long)&ibt_selftest_noendbr)) {
arch/x86/kernel/cet.c
129
__warn(__FILE__, __LINE__, (void *)regs->ip, TAINT_WARN, regs, NULL);
arch/x86/kernel/cet.c
72
regs->ip, regs->sp, ssp, error_code,
arch/x86/kernel/cet.c
75
print_vma_addr(KERN_CONT " in ", regs->ip);
arch/x86/kernel/cfi.c
38
if (copy_from_kernel_nofault(buffer, (void *)regs->ip - 12, MAX_INSN_SIZE))
arch/x86/kernel/cfi.c
47
if (copy_from_kernel_nofault(buffer, (void *)regs->ip - 6, MAX_INSN_SIZE))
arch/x86/kernel/cfi.c
70
unsigned long target, addr = regs->ip;
arch/x86/kernel/cpu/bus_lock.c
242
static void split_lock_warn(unsigned long ip)
arch/x86/kernel/cpu/bus_lock.c
250
current->comm, current->pid, ip);
arch/x86/kernel/cpu/bus_lock.c
277
bool handle_guest_split_lock(unsigned long ip)
arch/x86/kernel/cpu/bus_lock.c
280
split_lock_warn(ip);
arch/x86/kernel/cpu/bus_lock.c
286
sld_state == sld_fatal ? "fatal" : "bogus", ip);
arch/x86/kernel/cpu/bus_lock.c
323
split_lock_warn(regs->ip);
arch/x86/kernel/cpu/bus_lock.c
340
current->comm, current->pid, regs->ip);
arch/x86/kernel/cpu/mce/core.c
185
if (m->ip) {
arch/x86/kernel/cpu/mce/core.c
188
m->cs, m->ip);
arch/x86/kernel/cpu/mce/core.c
191
pr_cont("{%pS}", (void *)(unsigned long)m->ip);
arch/x86/kernel/cpu/mce/core.c
360
return offsetof(struct mce, ip);
arch/x86/kernel/cpu/mce/core.c
377
regs->ip, (void *)regs->ip);
arch/x86/kernel/cpu/mce/core.c
380
(unsigned int)regs->cx, regs->ip, (void *)regs->ip);
arch/x86/kernel/cpu/mce/core.c
479
m->ip = regs->ip;
arch/x86/kernel/cpu/mce/core.c
492
m->ip = mce_rdmsrq(mca_cfg.rip_msr);
arch/x86/kernel/cpu/mce/core.c
909
m->ip = regs->ip;
arch/x86/kernel/cpu/mce/inject.c
164
regs.ip = m->ip;
arch/x86/kernel/cpu/mce/severity.c
245
if (copy_from_kernel_nofault(insn_buf, (void *)regs->ip, MAX_INSN_SIZE))
arch/x86/kernel/cpu/mce/severity.c
299
fixup_type = ex_get_fixup_type(m->ip);
arch/x86/kernel/cpu/vmware.c
554
ghcb_set_rip(ghcb, regs->ip);
arch/x86/kernel/doublefault_32.c
55
regs.ip = TSS(ip);
arch/x86/kernel/doublefault_32.c
94
.ip = (unsigned long) asm_exc_double_fault,
arch/x86/kernel/dumpstack.c
117
unsigned long prologue = regs->ip - PROLOGUE_SIZE;
arch/x86/kernel/dumpstack.c
138
printk("%sEIP: %pS\n", loglvl, (void *)regs->ip);
arch/x86/kernel/dumpstack.c
140
printk("%sRIP: %04x:%pS\n", loglvl, (int)regs->cs, (void *)regs->ip);
arch/x86/kernel/dumpstack.c
263
if (regs && stack == ®s->ip)
arch/x86/kernel/ftrace.c
118
ftrace_modify_code_direct(unsigned long ip, const char *old_code,
arch/x86/kernel/ftrace.c
121
int ret = ftrace_verify_code(ip, old_code);
arch/x86/kernel/ftrace.c
127
smp_text_poke_batch_add((void *)ip, new_code, MCOUNT_INSN_SIZE, NULL);
arch/x86/kernel/ftrace.c
129
text_poke_early((void *)ip, new_code, MCOUNT_INSN_SIZE);
arch/x86/kernel/ftrace.c
135
unsigned long ip = rec->ip;
arch/x86/kernel/ftrace.c
138
old = ftrace_call_replace(ip, addr);
arch/x86/kernel/ftrace.c
150
return ftrace_modify_code_direct(ip, old, new);
arch/x86/kernel/ftrace.c
162
unsigned long ip = rec->ip;
arch/x86/kernel/ftrace.c
166
new = ftrace_call_replace(ip, addr);
arch/x86/kernel/ftrace.c
169
return ftrace_modify_code_direct(rec->ip, old, new);
arch/x86/kernel/ftrace.c
189
unsigned long ip;
arch/x86/kernel/ftrace.c
192
ip = (unsigned long)(&ftrace_call);
arch/x86/kernel/ftrace.c
193
new = ftrace_call_replace(ip, (unsigned long)func);
arch/x86/kernel/ftrace.c
194
smp_text_poke_single((void *)ip, new, MCOUNT_INSN_SIZE, NULL);
arch/x86/kernel/ftrace.c
196
ip = (unsigned long)(&ftrace_regs_call);
arch/x86/kernel/ftrace.c
197
new = ftrace_call_replace(ip, (unsigned long)func);
arch/x86/kernel/ftrace.c
198
smp_text_poke_single((void *)ip, new, MCOUNT_INSN_SIZE, NULL);
arch/x86/kernel/ftrace.c
224
old = ftrace_call_replace(rec->ip, ftrace_get_addr_curr(rec));
arch/x86/kernel/ftrace.c
228
ret = ftrace_verify_code(rec->ip, old);
arch/x86/kernel/ftrace.c
247
new = ftrace_call_replace(rec->ip, ftrace_get_addr_new(rec));
arch/x86/kernel/ftrace.c
255
smp_text_poke_batch_add((void *)rec->ip, new, MCOUNT_INSN_SIZE, NULL);
arch/x86/kernel/ftrace.c
321
void *ip, *dest;
arch/x86/kernel/ftrace.c
361
ip = trampoline + size;
arch/x86/kernel/ftrace.c
362
if (cpu_wants_rethunk_at(ip))
arch/x86/kernel/ftrace.c
363
__text_gen_insn(ip, JMP32_INSN_OPCODE, ip, x86_return_thunk, JMP32_INSN_SIZE);
arch/x86/kernel/ftrace.c
365
memcpy(ip, retq, sizeof(retq));
arch/x86/kernel/ftrace.c
370
ip = trampoline + (jmp_offset - start_offset);
arch/x86/kernel/ftrace.c
371
if (WARN_ON(*(char *)ip != 0x75))
arch/x86/kernel/ftrace.c
373
ret = copy_from_kernel_nofault(ip, x86_nops[2], 2);
arch/x86/kernel/ftrace.c
474
unsigned long ip;
arch/x86/kernel/ftrace.c
494
ip = ops->trampoline + offset;
arch/x86/kernel/ftrace.c
499
new = ftrace_call_replace(ip, (unsigned long)func);
arch/x86/kernel/ftrace.c
500
smp_text_poke_single((void *)ip, new, MCOUNT_INSN_SIZE, NULL);
arch/x86/kernel/ftrace.c
584
static const char *ftrace_jmp_replace(unsigned long ip, unsigned long addr)
arch/x86/kernel/ftrace.c
586
return text_gen_insn(JMP32_INSN_OPCODE, (void *)ip, (void *)addr);
arch/x86/kernel/ftrace.c
589
static int ftrace_mod_jmp(unsigned long ip, void *func)
arch/x86/kernel/ftrace.c
593
new = ftrace_jmp_replace(ip, (unsigned long)func);
arch/x86/kernel/ftrace.c
594
smp_text_poke_single((void *)ip, new, MCOUNT_INSN_SIZE, NULL);
arch/x86/kernel/ftrace.c
600
unsigned long ip = (unsigned long)(&ftrace_graph_call);
arch/x86/kernel/ftrace.c
602
return ftrace_mod_jmp(ip, &ftrace_graph_caller);
arch/x86/kernel/ftrace.c
607
unsigned long ip = (unsigned long)(&ftrace_graph_call);
arch/x86/kernel/ftrace.c
609
return ftrace_mod_jmp(ip, &ftrace_stub);
arch/x86/kernel/ftrace.c
639
void prepare_ftrace_return(unsigned long ip, unsigned long *parent,
arch/x86/kernel/ftrace.c
647
if (!function_graph_enter(*parent, ip, frame_pointer, parent))
arch/x86/kernel/ftrace.c
652
void ftrace_graph_func(unsigned long ip, unsigned long parent_ip,
arch/x86/kernel/ftrace.c
664
if (!function_graph_enter_regs(*parent, ip, 0, parent, fregs))
arch/x86/kernel/ftrace.c
71
static const char *ftrace_call_replace(unsigned long ip, unsigned long addr)
arch/x86/kernel/ftrace.c
79
return text_gen_insn(JMP32_INSN_OPCODE, (void *)ip, (void *)addr);
arch/x86/kernel/ftrace.c
81
return text_gen_insn(CALL_INSN_OPCODE, (void *)ip, (void *)addr);
arch/x86/kernel/ftrace.c
85
static int ftrace_verify_code(unsigned long ip, const char *old_code)
arch/x86/kernel/ftrace.c
97
if (copy_from_kernel_nofault(cur_code, (void *)ip, MCOUNT_INSN_SIZE)) {
arch/x86/kernel/kgdb.c
452
linux_regs->ip = addr;
arch/x86/kernel/kgdb.c
57
{ "ip", 4, offsetof(struct pt_regs, ip) },
arch/x86/kernel/kgdb.c
711
if (exception == 3 && kgdb_isremovedbreak(regs->ip - 1)) {
arch/x86/kernel/kgdb.c
712
regs->ip -= 1;
arch/x86/kernel/kgdb.c
725
void kgdb_arch_set_pc(struct pt_regs *regs, unsigned long ip)
arch/x86/kernel/kgdb.c
727
regs->ip = ip;
arch/x86/kernel/kgdb.c
80
{ "ip", 8, offsetof(struct pt_regs, ip) },
arch/x86/kernel/kprobes/core.c
1020
if ((unsigned long)p->ainsn.insn < regs->ip &&
arch/x86/kernel/kprobes/core.c
1021
(unsigned long)p->ainsn.insn + MAX_INSN_SIZE > regs->ip) {
arch/x86/kernel/kprobes/core.c
1038
if (unlikely(regs->ip == (unsigned long)cur->ainsn.insn)) {
arch/x86/kernel/kprobes/core.c
1049
regs->ip = (unsigned long)cur->addr;
arch/x86/kernel/kprobes/core.c
501
regs->ip = regs->ip - INT3_INSN_SIZE + p->ainsn.size;
arch/x86/kernel/kprobes/core.c
513
unsigned long func = regs->ip - INT3_INSN_SIZE + p->ainsn.size;
arch/x86/kernel/kprobes/core.c
522
unsigned long ip = regs->ip - INT3_INSN_SIZE + p->ainsn.size;
arch/x86/kernel/kprobes/core.c
524
ip += p->ainsn.rel32;
arch/x86/kernel/kprobes/core.c
525
int3_emulate_jmp(regs, ip);
arch/x86/kernel/kprobes/core.c
531
unsigned long ip = regs->ip - INT3_INSN_SIZE + p->ainsn.size;
arch/x86/kernel/kprobes/core.c
533
int3_emulate_jcc(regs, p->ainsn.jcc.type, ip, p->ainsn.rel32);
arch/x86/kernel/kprobes/core.c
539
unsigned long ip = regs->ip - INT3_INSN_SIZE + p->ainsn.size;
arch/x86/kernel/kprobes/core.c
568
ip += p->ainsn.rel32;
arch/x86/kernel/kprobes/core.c
569
int3_emulate_jmp(regs, ip);
arch/x86/kernel/kprobes/core.c
598
int3_emulate_push(regs, regs->ip - INT3_INSN_SIZE + p->ainsn.size);
arch/x86/kernel/kprobes/core.c
879
regs->ip = (unsigned long)p->ainsn.insn;
arch/x86/kernel/kprobes/core.c
898
regs->ip = (unsigned long)p->ainsn.insn;
arch/x86/kernel/kprobes/core.c
929
regs->ip += (orig_ip - copy_ip) - INT3_INSN_SIZE;
arch/x86/kernel/kprobes/core.c
987
addr = (kprobe_opcode_t *)(regs->ip - sizeof(kprobe_opcode_t));
arch/x86/kernel/kprobes/ftrace.c
17
void kprobe_ftrace_handler(unsigned long ip, unsigned long parent_ip,
arch/x86/kernel/kprobes/ftrace.c
28
bit = ftrace_test_recursion_trylock(ip, parent_ip);
arch/x86/kernel/kprobes/ftrace.c
32
p = get_kprobe((kprobe_opcode_t *)ip);
arch/x86/kernel/kprobes/ftrace.c
43
instruction_pointer_set(regs, ip + INT3_INSN_SIZE);
arch/x86/kernel/kprobes/ftrace.c
53
instruction_pointer_set(regs, ip + MCOUNT_INSN_SIZE);
arch/x86/kernel/kprobes/opt.c
191
regs->ip = (unsigned long)op->kp.addr + INT3_INSN_SIZE;
arch/x86/kernel/kprobes/opt.c
544
regs->ip = (unsigned long)op->optinsn.insn + TMPL_END_IDX;
arch/x86/kernel/nmi.c
376
if (regs->ip == __this_cpu_read(last_nmi_rip))
arch/x86/kernel/nmi.c
381
__this_cpu_write(last_nmi_rip, regs->ip);
arch/x86/kernel/perf_regs.c
162
regs_user_copy->ip = user_regs->ip;
arch/x86/kernel/perf_regs.c
29
PT_REGS_OFFSET(PERF_REG_X86_IP, ip),
arch/x86/kernel/process.c
260
childregs->ip = 0;
arch/x86/kernel/process_32.c
120
regs->ip = new_ip;
arch/x86/kernel/process_64.c
549
regs->ip = new_ip;
arch/x86/kernel/ptrace.c
110
REG_OFFSET_NAME(ip),
arch/x86/kernel/ptrace.c
1417
user_mode(regs) ? (void __user *)regs->ip : NULL);
arch/x86/kernel/ptrace.c
918
R32(eip, ip);
arch/x86/kernel/ptrace.c
992
R32(eip, ip);
arch/x86/kernel/rethook.c
76
regs->ip = (unsigned long)&arch_rethook_trampoline;
arch/x86/kernel/signal.c
280
regs->ip -= 2;
arch/x86/kernel/signal.c
351
regs->ip -= 2;
arch/x86/kernel/signal.c
356
regs->ip -= 2;
arch/x86/kernel/signal.c
377
regs->ip, regs->sp, regs->orig_ax);
arch/x86/kernel/signal.c
378
print_vma_addr(KERN_CONT " in ", regs->ip);
arch/x86/kernel/signal_32.c
127
regs->ip = sc.ip;
arch/x86/kernel/signal_32.c
236
unsafe_put_user(regs->ip, &sc->ip, Efault);
arch/x86/kernel/signal_32.c
306
regs->ip = (unsigned long) ksig->ka.sa.sa_handler;
arch/x86/kernel/signal_32.c
388
regs->ip = (unsigned long) ksig->ka.sa.sa_handler;
arch/x86/kernel/signal_64.c
121
unsafe_put_user(regs->ip, &sc->ip, Efault);
arch/x86/kernel/signal_64.c
210
regs->ip = (unsigned long) ksig->ka.sa.sa_handler;
arch/x86/kernel/signal_64.c
344
regs->ip = (unsigned long) ksig->ka.sa.sa_handler;
arch/x86/kernel/signal_64.c
70
regs->ip = sc.ip;
arch/x86/kernel/stacktrace.c
111
if (!consume_entry(cookie, regs->ip))
arch/x86/kernel/stacktrace.c
21
if (regs && !consume_entry(cookie, regs->ip))
arch/x86/kernel/step.c
18
addr = regs->ip;
arch/x86/kernel/traps.c
1108
__memcpy(&tmp.ip, (void *)bad_regs->sp, 5*8);
arch/x86/kernel/traps.c
1111
__memcpy(&tmp, bad_regs, offsetof(struct pt_regs, ip));
arch/x86/kernel/traps.c
1132
return (regs->ip - (unsigned long)__begin_SYSENTER_singlestep_region) <
arch/x86/kernel/traps.c
1136
return (regs->ip - (unsigned long)entry_SYSENTER_compat) <
arch/x86/kernel/traps.c
271
if ((void *)regs->ip == &__WARN_trap)
arch/x86/kernel/traps.c
327
regs->ip, regs->sp, error_code);
arch/x86/kernel/traps.c
328
print_vma_addr(KERN_CONT " in ", regs->ip);
arch/x86/kernel/traps.c
402
unsigned long addr = regs->ip;
arch/x86/kernel/traps.c
435
if (report_bug(regs->ip, regs) == BUG_TRAP_TYPE_WARN) {
arch/x86/kernel/traps.c
453
(void *)regs->ip);
arch/x86/kernel/traps.c
467
if (regs->ip == addr)
arch/x86/kernel/traps.c
468
regs->ip += ud_len;
arch/x86/kernel/traps.c
470
regs->ip = addr;
arch/x86/kernel/traps.c
625
regs->ip == (unsigned long)native_irq_return_iret)
arch/x86/kernel/traps.c
636
gpregs->ip = p[0];
arch/x86/kernel/traps.c
654
regs->ip = (unsigned long)asm_exc_general_protection;
arch/x86/kernel/traps.c
760
if (copy_from_kernel_nofault(insn_buf, (void *)regs->ip,
arch/x86/kernel/traps.c
806
unsigned long ip;
arch/x86/kernel/traps.c
811
if (insn_get_effective_ip(regs, &ip))
arch/x86/kernel/traps.c
814
if (get_user(byte, (const char __user *)ip))
arch/x86/kernel/traps.c
822
current->comm, task_pid_nr(current), ip);
arch/x86/kernel/traps.c
823
print_vma_addr(KERN_CONT " in ", ip);
arch/x86/kernel/traps.c
828
regs->ip += 1;
arch/x86/kernel/umip.c
132
task_pid_nr(tsk), regs->ip, regs->sp, &vaf);
arch/x86/kernel/umip.c
420
regs->ip += insn.length;
arch/x86/kernel/unwind_frame.c
19
return __kernel_text_address(state->ip) ? state->ip : 0;
arch/x86/kernel/unwind_frame.c
249
state->ip = 0;
arch/x86/kernel/unwind_frame.c
253
state->ip = unwind_recover_ret_addr(state, addr, addr_p);
arch/x86/kernel/unwind_frame.c
28
return state->regs ? &state->regs->ip : state->bp + 1;
arch/x86/kernel/unwind_frame.c
300
state->ip = 0;
arch/x86/kernel/unwind_frame.c
337
if (state->got_irq && in_entry_code(state->ip))
arch/x86/kernel/unwind_frame.c
399
if (regs && regs->ip == 0 && (unsigned long *)regs->sp >= first_frame) {
arch/x86/kernel/unwind_frame.c
73
static bool in_entry_code(unsigned long ip)
arch/x86/kernel/unwind_frame.c
75
char *addr = (char *)ip;
arch/x86/kernel/unwind_orc.c
103
if (orc_ip(mid) <= ip) {
arch/x86/kernel/unwind_orc.c
114
static struct orc_entry *orc_module_find(unsigned long ip)
arch/x86/kernel/unwind_orc.c
118
mod = __module_address(ip);
arch/x86/kernel/unwind_orc.c
122
mod->arch.num_orcs, ip);
arch/x86/kernel/unwind_orc.c
125
static struct orc_entry *orc_module_find(unsigned long ip)
arch/x86/kernel/unwind_orc.c
132
static struct orc_entry *orc_find(unsigned long ip);
arch/x86/kernel/unwind_orc.c
144
static struct orc_entry *orc_ftrace_find(unsigned long ip)
arch/x86/kernel/unwind_orc.c
149
ops = ftrace_ops_trampoline(ip);
arch/x86/kernel/unwind_orc.c
160
offset = ip - ops->trampoline;
arch/x86/kernel/unwind_orc.c
164
if (ip == tramp_addr)
arch/x86/kernel/unwind_orc.c
170
static struct orc_entry *orc_ftrace_find(unsigned long ip)
arch/x86/kernel/unwind_orc.c
185
static struct orc_entry *orc_bpf_find(unsigned long ip)
arch/x86/kernel/unwind_orc.c
188
if (bpf_has_frame_pointer(ip))
arch/x86/kernel/unwind_orc.c
209
static struct orc_entry *orc_find(unsigned long ip)
arch/x86/kernel/unwind_orc.c
213
if (ip == 0)
arch/x86/kernel/unwind_orc.c
217
if (ip >= LOOKUP_START_IP && ip < LOOKUP_STOP_IP) {
arch/x86/kernel/unwind_orc.c
220
idx = (ip - LOOKUP_START_IP) / LOOKUP_BLOCK_SIZE;
arch/x86/kernel/unwind_orc.c
224
idx, lookup_num_blocks, (void *)ip);
arch/x86/kernel/unwind_orc.c
234
idx, lookup_num_blocks, start, stop, (void *)ip);
arch/x86/kernel/unwind_orc.c
239
__start_orc_unwind + start, stop - start, ip);
arch/x86/kernel/unwind_orc.c
243
if (is_kernel_inittext(ip))
arch/x86/kernel/unwind_orc.c
245
__stop_orc_unwind_ip - __start_orc_unwind_ip, ip);
arch/x86/kernel/unwind_orc.c
248
orc = orc_module_find(ip);
arch/x86/kernel/unwind_orc.c
253
orc = orc_bpf_find(ip);
arch/x86/kernel/unwind_orc.c
257
return orc_ftrace_find(ip);
arch/x86/kernel/unwind_orc.c
385
return __kernel_text_address(state->ip) ? state->ip : 0;
arch/x86/kernel/unwind_orc.c
395
return &state->regs->ip;
arch/x86/kernel/unwind_orc.c
427
unsigned long *ip, unsigned long *sp)
arch/x86/kernel/unwind_orc.c
437
*ip = READ_ONCE_NOCHECK(regs->ip);
arch/x86/kernel/unwind_orc.c
443
unsigned long *ip, unsigned long *sp)
arch/x86/kernel/unwind_orc.c
450
*ip = READ_ONCE_NOCHECK(regs->ip);
arch/x86/kernel/unwind_orc.c
486
unsigned long ip_p, sp, tmp, orig_ip = state->ip, prev_sp = state->sp;
arch/x86/kernel/unwind_orc.c
510
orc = orc_find(state->signal ? state->ip : state->ip - 1);
arch/x86/kernel/unwind_orc.c
552
(void *)state->ip);
arch/x86/kernel/unwind_orc.c
560
(void *)state->ip);
arch/x86/kernel/unwind_orc.c
568
(void *)state->ip);
arch/x86/kernel/unwind_orc.c
576
(void *)state->ip);
arch/x86/kernel/unwind_orc.c
583
orc->sp_reg, (void *)state->ip);
arch/x86/kernel/unwind_orc.c
600
if (!deref_stack_reg(state, ip_p, &state->ip))
arch/x86/kernel/unwind_orc.c
603
state->ip = unwind_recover_ret_addr(state, state->ip,
arch/x86/kernel/unwind_orc.c
611
if (!deref_stack_regs(state, sp, &state->ip, &state->sp)) {
arch/x86/kernel/unwind_orc.c
626
state->ip = unwind_recover_rethook(state, state->ip,
arch/x86/kernel/unwind_orc.c
634
if (!deref_stack_iret_regs(state, sp, &state->ip, &state->sp)) {
arch/x86/kernel/unwind_orc.c
640
state->ip = unwind_recover_rethook(state, state->ip,
arch/x86/kernel/unwind_orc.c
719
state->ip = regs->ip;
arch/x86/kernel/unwind_orc.c
730
: "=r" (state->ip), "=r" (state->sp),
arch/x86/kernel/unwind_orc.c
738
state->ip = READ_ONCE_NOCHECK(frame->ret_addr);
arch/x86/kernel/unwind_orc.c
739
state->signal = (void *)state->ip == ret_from_fork_asm;
arch/x86/kernel/unwind_orc.c
79
static inline unsigned long orc_ip(const int *ip)
arch/x86/kernel/unwind_orc.c
81
return (unsigned long)ip + *ip;
arch/x86/kernel/unwind_orc.c
85
unsigned int num_entries, unsigned long ip)
arch/x86/kernel/uprobes.c
1247
regs->ip += correction;
arch/x86/kernel/uprobes.c
1326
unsigned long new_ip = regs->ip += auprobe->branch.ilen;
arch/x86/kernel/uprobes.c
1345
regs->ip = new_ip + offs;
arch/x86/kernel/uprobes.c
1355
regs->ip += auprobe->push.ilen;
arch/x86/kernel/uprobes.c
1615
regs->ip = utask->xol_vaddr;
arch/x86/kernel/uprobes.c
1669
regs->ip = utask->vaddr;
arch/x86/kernel/uprobes.c
1733
regs->ip = utask->vaddr;
arch/x86/kernel/uprobes.c
1778
current->pid, regs->sp, regs->ip);
arch/x86/kernel/uprobes.c
377
unsigned long err, ip, sp, tramp;
arch/x86/kernel/uprobes.c
385
if (unlikely(regs->ip != trampoline_check_ip(tramp)))
arch/x86/kernel/uprobes.c
399
ip = regs->ip;
arch/x86/kernel/uprobes.c
423
args.ax = regs->ip;
arch/x86/kernel/uprobes.c
424
regs->ip = ip;
arch/x86/kernel/uprobes.c
432
regs->cx = regs->ip;
arch/x86/kernel/uprobes.c
764
static bool __in_uprobe_trampoline(unsigned long ip)
arch/x86/kernel/uprobes.c
766
struct vm_area_struct *vma = vma_lookup(current->mm, ip);
arch/x86/kernel/uprobes.c
771
static bool in_uprobe_trampoline(unsigned long ip)
arch/x86/kernel/uprobes.c
779
found = __in_uprobe_trampoline(ip);
arch/x86/kernel/uprobes.c
786
found = __in_uprobe_trampoline(ip);
arch/x86/kernel/uprobes.c
808
unsigned long ip, sp, sret;
arch/x86/kernel/uprobes.c
812
if (!in_uprobe_trampoline(regs->ip))
arch/x86/kernel/uprobes.c
819
ip = regs->ip;
arch/x86/kernel/uprobes.c
829
regs->ip = args.retaddr - 5;
arch/x86/kernel/uprobes.c
839
handle_syscall_uprobe(regs, regs->ip);
arch/x86/kernel/uprobes.c
847
if (args.retaddr - 5 == regs->ip)
arch/x86/kernel/uprobes.c
848
regs->ip += 5;
arch/x86/kernel/uprobes.c
860
if (args.retaddr - 5 != regs->ip)
arch/x86/kernel/uprobes.c
861
args.retaddr = regs->ip;
arch/x86/kernel/uprobes.c
866
regs->ip = ip;
arch/x86/kernel/uprobes.c
874
regs->cx = regs->ip;
arch/x86/kernel/vm86_32.c
126
unsafe_put_user(regs->pt.ip, &user->regs.eip, Efault_end);
arch/x86/kernel/vm86_32.c
263
vm86regs.pt.ip = v.regs.eip;
arch/x86/kernel/vm86_32.c
552
unsigned short ip, sp, orig_flags;
arch/x86/kernel/vm86_32.c
565
ip = IP(regs);
arch/x86/kernel/vm86_32.c
570
switch (opcode = popb(csp, ip, simulate_sigsegv)) {
arch/x86/kernel/vm86_32.c
596
IP(regs) = ip;
arch/x86/kernel/vm86_32.c
610
IP(regs) = ip;
arch/x86/kernel/vm86_32.c
622
int intno = popb(csp, ip, simulate_sigsegv);
arch/x86/kernel/vm86_32.c
623
IP(regs) = ip;
arch/x86/kernel/vm86_32.c
664
IP(regs) = ip;
arch/x86/kernel/vm86_32.c
676
IP(regs) = ip;
arch/x86/kernel/vm86_32.c
81
#define IP(regs) (*(unsigned short *)&((regs)->pt.ip))
arch/x86/kvm/emulate.c
2623
tss->ip = ctxt->_eip;
arch/x86/kvm/emulate.c
2647
ctxt->_eip = tss->ip;
arch/x86/kvm/tss.h
43
u16 ip;
arch/x86/lib/error-inject.c
23
regs->ip = (unsigned long)&just_return_func;
arch/x86/lib/insn-eval.c
1047
tmp = regs->ip + insn->length;
arch/x86/lib/insn-eval.c
1472
int insn_get_effective_ip(struct pt_regs *regs, unsigned long *ip)
arch/x86/lib/insn-eval.c
1488
*ip = seg_base + regs->ip;
arch/x86/lib/insn-eval.c
1509
unsigned long ip;
arch/x86/lib/insn-eval.c
1512
if (insn_get_effective_ip(regs, &ip))
arch/x86/lib/insn-eval.c
1515
not_copied = copy_from_user(buf, (void __user *)ip, MAX_INSN_SIZE);
arch/x86/lib/insn-eval.c
1537
unsigned long ip;
arch/x86/lib/insn-eval.c
1540
if (insn_get_effective_ip(regs, &ip))
arch/x86/lib/insn-eval.c
1543
not_copied = __copy_from_user_inatomic(buf, (void __user *)ip, MAX_INSN_SIZE);
arch/x86/lib/insn-eval.c
221
case offsetof(struct pt_regs, ip):
arch/x86/lib/insn-eval.c
291
if (regoff == offsetof(struct pt_regs, ip)) {
arch/x86/math-emu/fpu_entry.c
550
u_char __user *ip = *fpu_eip;
arch/x86/math-emu/fpu_entry.c
557
FPU_get_user(byte, ip);
arch/x86/math-emu/fpu_entry.c
598
ip++;
arch/x86/math-emu/fpu_entry.c
601
FPU_get_user(byte, ip);
arch/x86/math-emu/fpu_entry.c
610
*fpu_eip = ip;
arch/x86/math-emu/fpu_system.h
84
#define FPU_EIP (FPU_info->regs->ip)
arch/x86/mm/extable.c
172
(unsigned int)regs->ax, regs->ip, (void *)regs->ip);
arch/x86/mm/extable.c
178
(unsigned int)regs->cx, regs->ip, (void *)regs->ip);
arch/x86/mm/extable.c
292
int ex_get_fixup_type(unsigned long ip)
arch/x86/mm/extable.c
294
const struct exception_table_entry *e = search_exception_tables(ip);
arch/x86/mm/extable.c
319
e = search_exception_tables(regs->ip);
arch/x86/mm/extable.c
41
regs->ip = ex_fixup_addr(e);
arch/x86/mm/extable.c
426
(unsigned)trapnr, (unsigned long)regs->cs, regs->ip,
arch/x86/mm/extable.c
67
len = next_ip - regs->ip;
arch/x86/mm/extable.c
71
if (insn_decode(&insn, (void *) regs->ip, len, INSN_MODE_KERN))
arch/x86/mm/fault.c
429
if (address != regs->ip)
arch/x86/mm/fault.c
439
regs->ip = address;
arch/x86/mm/fault.c
759
(void *)regs->ip, (void *)regs->sp, error_code);
arch/x86/mm/fault.c
761
print_vma_addr(KERN_CONT " in ", regs->ip);
arch/x86/mm/mmio-mod.c
111
pr_emerg("faulting IP is at %pS\n", (void *)regs->ip);
arch/x86/mm/mmio-mod.c
112
pr_emerg("last faulting IP was at %pS\n", (void *)my_reason->ip);
arch/x86/mm/mmio-mod.c
145
my_reason->ip = instptr;
arch/x86/mm/mmio-mod.c
182
unsigned char *ip = (unsigned char *)instptr;
arch/x86/mm/mmio-mod.c
185
my_trace->value = (*ip) << 16 | *(ip + 1) << 8 |
arch/x86/mm/mmio-mod.c
186
*(ip + 2);
arch/x86/mm/mmio-mod.c
208
my_trace->value = get_ins_reg_val(my_reason->ip, regs);
arch/x86/mm/mmio-mod.c
29
unsigned long ip;
arch/x86/net/bpf_jit_comp.c
1493
bpf_prog_report_arena_violation(is_write, addr, regs->ip);
arch/x86/net/bpf_jit_comp.c
1499
regs->ip += insn_len;
arch/x86/net/bpf_jit_comp.c
1610
static int emit_spectre_bhb_barrier(u8 **pprog, u8 *ip,
arch/x86/net/bpf_jit_comp.c
1620
ip += 2;
arch/x86/net/bpf_jit_comp.c
1623
ip += x86_call_depth_emit_accounting(&prog, func, ip);
arch/x86/net/bpf_jit_comp.c
1625
if (emit_call(&prog, func, ip))
arch/x86/net/bpf_jit_comp.c
2443
u8 *ip = image + addrs[i - 1];
arch/x86/net/bpf_jit_comp.c
2448
ip += 7;
arch/x86/net/bpf_jit_comp.c
2454
ip += 2;
arch/x86/net/bpf_jit_comp.c
2456
ip += x86_call_depth_emit_accounting(&prog, func, ip);
arch/x86/net/bpf_jit_comp.c
2457
if (emit_call(&prog, func, ip))
arch/x86/net/bpf_jit_comp.c
2732
u8 *ip = image + addrs[i - 1];
arch/x86/net/bpf_jit_comp.c
2734
if (emit_spectre_bhb_barrier(&prog, ip, bpf_prog))
arch/x86/net/bpf_jit_comp.c
3071
static int emit_cond_near_jump(u8 **pprog, void *func, void *ip, u8 jmp_cond)
arch/x86/net/bpf_jit_comp.c
3076
offset = func - (ip + 2 + 4);
arch/x86/net/bpf_jit_comp.c
4007
void arch_bpf_stack_walk(bool (*consume_fn)(void *cookie, u64 ip, u64 sp, u64 bp), void *cookie)
arch/x86/net/bpf_jit_comp.c
416
static int emit_call(u8 **prog, void *func, void *ip);
arch/x86/net/bpf_jit_comp.c
418
static void emit_fineibt(u8 **pprog, u8 *ip, u32 hash, int arity)
arch/x86/net/bpf_jit_comp.c
426
emit_call(&prog, __bhi_args[arity], ip + 11);
arch/x86/net/bpf_jit_comp.c
449
static void emit_cfi(u8 **pprog, u8 *ip, u32 hash, int arity)
arch/x86/net/bpf_jit_comp.c
455
emit_fineibt(&prog, ip, hash, arity);
arch/x86/net/bpf_jit_comp.c
506
static void emit_prologue(u8 **pprog, u8 *ip, u32 stack_depth, bool ebpf_from_cbpf,
arch/x86/net/bpf_jit_comp.c
513
emit_cfi(&prog, ip, cfi_bpf_subprog_hash, 5);
arch/x86/net/bpf_jit_comp.c
515
emit_cfi(&prog, ip, cfi_bpf_hash, 1);
arch/x86/net/bpf_jit_comp.c
559
static int emit_patch(u8 **pprog, void *func, void *ip, u8 opcode)
arch/x86/net/bpf_jit_comp.c
564
offset = func - (ip + X86_PATCH_SIZE);
arch/x86/net/bpf_jit_comp.c
574
static int emit_call(u8 **pprog, void *func, void *ip)
arch/x86/net/bpf_jit_comp.c
576
return emit_patch(pprog, func, ip, 0xE8);
arch/x86/net/bpf_jit_comp.c
579
static int emit_rsb_call(u8 **pprog, void *func, void *ip)
arch/x86/net/bpf_jit_comp.c
582
ip += x86_call_depth_emit_accounting(pprog, func, ip);
arch/x86/net/bpf_jit_comp.c
583
return emit_patch(pprog, func, ip, 0xE8);
arch/x86/net/bpf_jit_comp.c
586
static int emit_jump(u8 **pprog, void *func, void *ip)
arch/x86/net/bpf_jit_comp.c
588
return emit_patch(pprog, func, ip, 0xE9);
arch/x86/net/bpf_jit_comp.c
591
static int __bpf_arch_text_poke(void *ip, enum bpf_text_poke_type old_t,
arch/x86/net/bpf_jit_comp.c
605
emit_call(&prog, old_addr, ip) :
arch/x86/net/bpf_jit_comp.c
606
emit_jump(&prog, old_addr, ip);
arch/x86/net/bpf_jit_comp.c
615
emit_call(&prog, new_addr, ip) :
arch/x86/net/bpf_jit_comp.c
616
emit_jump(&prog, new_addr, ip);
arch/x86/net/bpf_jit_comp.c
623
if (memcmp(ip, old_insn, X86_PATCH_SIZE))
arch/x86/net/bpf_jit_comp.c
626
if (memcmp(ip, new_insn, X86_PATCH_SIZE)) {
arch/x86/net/bpf_jit_comp.c
627
smp_text_poke_single(ip, new_insn, X86_PATCH_SIZE, NULL);
arch/x86/net/bpf_jit_comp.c
635
int bpf_arch_text_poke(void *ip, enum bpf_text_poke_type old_t,
arch/x86/net/bpf_jit_comp.c
639
if (!is_kernel_text((long)ip) &&
arch/x86/net/bpf_jit_comp.c
640
!is_bpf_text_address((long)ip))
arch/x86/net/bpf_jit_comp.c
648
if (is_endbr(ip))
arch/x86/net/bpf_jit_comp.c
649
ip += ENDBR_INSN_SIZE;
arch/x86/net/bpf_jit_comp.c
651
return __bpf_arch_text_poke(ip, old_t, new_t, old_addr, new_addr);
arch/x86/net/bpf_jit_comp.c
668
static void emit_indirect_jump(u8 **pprog, int bpf_reg, u8 *ip)
arch/x86/net/bpf_jit_comp.c
676
emit_jump(&prog, its_static_thunk(reg + 8*ereg), ip);
arch/x86/net/bpf_jit_comp.c
683
emit_jump(&prog, &__x86_indirect_jump_thunk_array[reg + 8*ereg], ip);
arch/x86/net/bpf_jit_comp.c
685
emit_jump(&prog, &__x86_indirect_thunk_array[reg + 8*ereg], ip);
arch/x86/net/bpf_jit_comp.c
695
static void emit_return(u8 **pprog, u8 *ip)
arch/x86/net/bpf_jit_comp.c
700
emit_jump(&prog, x86_return_thunk, ip);
arch/x86/net/bpf_jit_comp.c
728
u32 stack_depth, u8 *ip,
arch/x86/net/bpf_jit_comp.c
807
emit_indirect_jump(&prog, BPF_REG_4 /* R4 -> rcx */, ip + (prog - start));
arch/x86/net/bpf_jit_comp.c
816
u8 **pprog, u8 *ip,
arch/x86/net/bpf_jit_comp.c
834
poke->tailcall_bypass = ip + (prog - start);
arch/x86/net/bpf_jit_comp.c
836
poke->tailcall_target = ip + ctx->tail_call_direct_label - X86_PATCH_SIZE;
arch/x86/net/bpf_jit_comp32.c
1271
static int emit_jmp_edx(u8 **pprog, u8 *ip)
arch/x86/net/bpf_jit_comp32.c
1277
EMIT1_off32(0xE9, (u8 *)__x86_indirect_thunk_edx - (ip + 5));
arch/x86/net/bpf_jit_comp32.c
1299
static void emit_bpf_tail_call(u8 **pprog, u8 *ip)
arch/x86/net/bpf_jit_comp32.c
1381
cnt += emit_jmp_edx(&prog, ip + cnt);
arch/x86/platform/uv/uv_nmi.c
716
cpu, current->pid, current->comm, (void *)regs->ip);
arch/x86/um/signal.c
149
PUTREG(IP, ip);
arch/x86/um/signal.c
63
GETREG(IP, ip);
arch/x86/xen/pmu.c
429
return xenpmu_data->pmu.r.regs.ip;
arch/x86/xen/pmu.c
441
regs->ip = xen_regs->ip;
drivers/accessibility/speakup/devsynth.c
103
static int speakup_file_open(struct inode *ip, struct file *fp)
drivers/accessibility/speakup/devsynth.c
112
static int speakup_file_release(struct inode *ip, struct file *fp)
drivers/acpi/apei/apei-base.c
149
u32 i, ip;
drivers/acpi/apei/apei-base.c
153
ctx->ip = 0;
drivers/acpi/apei/apei-base.c
162
ip = 0;
drivers/acpi/apei/apei-base.c
167
if (ip == ctx->ip) {
drivers/acpi/apei/apei-base.c
180
ctx->ip++;
drivers/acpi/apei/apei-base.c
182
ip++;
drivers/acpi/apei/apei-base.c
183
if (ctx->ip < ip)
drivers/acpi/apei/apei-internal.h
25
u32 ip;
drivers/acpi/apei/erst.c
249
ctx->ip += 2;
drivers/acpi/apei/erst.c
259
ctx->ip = ctx->value;
drivers/block/aoe/aoe.h
186
} ip;
drivers/block/aoe/aoecmd.c
1049
if (rq == d->ip.rq)
drivers/block/aoe/aoecmd.c
1050
d->ip.rq = NULL;
drivers/block/aoe/aoecmd.c
1071
if (buf == d->ip.buf)
drivers/block/aoe/aoecmd.c
1072
d->ip.buf = NULL;
drivers/block/aoe/aoecmd.c
392
d->ip.buf = NULL;
drivers/block/aoe/aoecmd.c
854
if (d->ip.buf)
drivers/block/aoe/aoecmd.c
855
return d->ip.buf;
drivers/block/aoe/aoecmd.c
856
rq = d->ip.rq;
drivers/block/aoe/aoecmd.c
864
d->ip.rq = rq;
drivers/block/aoe/aoecmd.c
865
d->ip.nxbio = rq->bio;
drivers/block/aoe/aoecmd.c
877
bio = d->ip.nxbio;
drivers/block/aoe/aoecmd.c
880
d->ip.nxbio = bio;
drivers/block/aoe/aoecmd.c
882
d->ip.rq = NULL;
drivers/block/aoe/aoecmd.c
883
return d->ip.buf = buf;
drivers/block/aoe/aoedev.c
166
aoe_failbuf(d, d->ip.buf);
drivers/block/aoe/aoedev.c
167
rq = d->ip.rq;
drivers/block/aoe/aoedev.c
172
while ((bio = d->ip.nxbio)) {
drivers/block/aoe/aoedev.c
174
d->ip.nxbio = bio->bi_next;
drivers/bus/intel-ixp4xx-eb.c
211
const struct ixp4xx_exp_tim_prop *ip = &ixp4xx_exp_tim_props[i];
drivers/bus/intel-ixp4xx-eb.c
214
ret = of_property_read_u32(np, ip->prop, &val);
drivers/bus/intel-ixp4xx-eb.c
219
if (ip->max == 1) {
drivers/bus/intel-ixp4xx-eb.c
221
cs_cfg |= ip->mask;
drivers/bus/intel-ixp4xx-eb.c
223
cs_cfg &= ~ip->mask;
drivers/bus/intel-ixp4xx-eb.c
226
ip->prop);
drivers/bus/intel-ixp4xx-eb.c
230
if (val > ip->max) {
drivers/bus/intel-ixp4xx-eb.c
233
cs_index, ip->prop, val, ip->max);
drivers/bus/intel-ixp4xx-eb.c
234
val = ip->max;
drivers/bus/intel-ixp4xx-eb.c
237
cs_cfg &= ~ip->mask;
drivers/bus/intel-ixp4xx-eb.c
238
cs_cfg |= (val << ip->shift);
drivers/bus/intel-ixp4xx-eb.c
239
dev_info(eb->dev, "CS%d set %s to %u\n", cs_index, ip->prop, val);
drivers/cxl/core/region.c
985
struct cxl_region_params *ip = &iter->region->params;
drivers/cxl/core/region.c
987
if (!ip->res || ip->res->start < p->res->start)
drivers/cxl/core/region.c
996
dev_name(&iter->region->dev), ip->res, p->res);
drivers/firmware/efi/cper-x86.c
320
newpfx, err_info->ip);
drivers/firmware/efi/cper.c
246
printk("%s""IP: 0x%016llx\n", pfx, proc->ip);
drivers/firmware/iscsi_ibft.c
192
static int address_not_null(u8 *ip)
drivers/firmware/iscsi_ibft.c
194
return (memcmp(ip, nulls, 16) && memcmp(ip, mapped_nulls, 16));
drivers/firmware/iscsi_ibft.c
200
static ssize_t sprintf_ipaddr(char *buf, u8 *ip)
drivers/firmware/iscsi_ibft.c
204
if (ip[0] == 0 && ip[1] == 0 && ip[2] == 0 && ip[3] == 0 &&
drivers/firmware/iscsi_ibft.c
205
ip[4] == 0 && ip[5] == 0 && ip[6] == 0 && ip[7] == 0 &&
drivers/firmware/iscsi_ibft.c
206
ip[8] == 0 && ip[9] == 0 && ip[10] == 0xff && ip[11] == 0xff) {
drivers/firmware/iscsi_ibft.c
210
str += sprintf(buf, "%pI4", ip + 12);
drivers/firmware/iscsi_ibft.c
215
str += sprintf(str, "%pI6", ip);
drivers/gpio/gpiolib-cdev.c
1249
static long linereq_get_values(struct linereq *lr, void __user *ip)
drivers/gpio/gpiolib-cdev.c
1259
if (copy_from_user(&lv, ip, sizeof(lv)))
drivers/gpio/gpiolib-cdev.c
1315
if (copy_to_user(ip, &lv, sizeof(lv)))
drivers/gpio/gpiolib-cdev.c
1321
static long linereq_set_values(struct linereq *lr, void __user *ip)
drivers/gpio/gpiolib-cdev.c
1329
if (copy_from_user(&lv, ip, sizeof(lv)))
drivers/gpio/gpiolib-cdev.c
1376
static long linereq_set_config(struct linereq *lr, void __user *ip)
drivers/gpio/gpiolib-cdev.c
1385
if (copy_from_user(&lc, ip, sizeof(lc)))
drivers/gpio/gpiolib-cdev.c
1434
void __user *ip = (void __user *)arg;
drivers/gpio/gpiolib-cdev.c
1443
return linereq_get_values(lr, ip);
drivers/gpio/gpiolib-cdev.c
1445
return linereq_set_values(lr, ip);
drivers/gpio/gpiolib-cdev.c
1447
return linereq_set_config(lr, ip);
drivers/gpio/gpiolib-cdev.c
1589
static int linereq_create(struct gpio_device *gdev, void __user *ip)
drivers/gpio/gpiolib-cdev.c
1599
if (copy_from_user(&ulr, ip, sizeof(ulr)))
drivers/gpio/gpiolib-cdev.c
164
void __user *ip)
drivers/gpio/gpiolib-cdev.c
171
if (copy_from_user(&gcnf, ip, sizeof(gcnf)))
drivers/gpio/gpiolib-cdev.c
1719
if (copy_to_user(ip, &ulr, sizeof(ulr))) {
drivers/gpio/gpiolib-cdev.c
1902
void __user *ip = (void __user *)arg;
drivers/gpio/gpiolib-cdev.c
1924
if (copy_to_user(ip, &ghd, sizeof(ghd)))
drivers/gpio/gpiolib-cdev.c
2013
static int lineevent_create(struct gpio_device *gdev, void __user *ip)
drivers/gpio/gpiolib-cdev.c
2027
if (copy_from_user(&eventreq, ip, sizeof(eventreq)))
drivers/gpio/gpiolib-cdev.c
208
void __user *ip = (void __user *)arg;
drivers/gpio/gpiolib-cdev.c
2146
if (copy_to_user(ip, &eventreq, sizeof(eventreq))) {
drivers/gpio/gpiolib-cdev.c
2312
static int chipinfo_get(struct gpio_chardev_data *cdev, void __user *ip)
drivers/gpio/gpiolib-cdev.c
232
if (copy_to_user(ip, &ghd, sizeof(ghd)))
drivers/gpio/gpiolib-cdev.c
2322
if (copy_to_user(ip, &chipinfo, sizeof(chipinfo)))
drivers/gpio/gpiolib-cdev.c
2342
static int lineinfo_get_v1(struct gpio_chardev_data *cdev, void __user *ip,
drivers/gpio/gpiolib-cdev.c
2349
if (copy_from_user(&lineinfo, ip, sizeof(lineinfo)))
drivers/gpio/gpiolib-cdev.c
2368
if (copy_to_user(ip, &lineinfo, sizeof(lineinfo))) {
drivers/gpio/gpiolib-cdev.c
2378
static int lineinfo_get(struct gpio_chardev_data *cdev, void __user *ip,
drivers/gpio/gpiolib-cdev.c
2384
if (copy_from_user(&lineinfo, ip, sizeof(lineinfo)))
drivers/gpio/gpiolib-cdev.c
2404
if (copy_to_user(ip, &lineinfo, sizeof(lineinfo))) {
drivers/gpio/gpiolib-cdev.c
2413
static int lineinfo_unwatch(struct gpio_chardev_data *cdev, void __user *ip)
drivers/gpio/gpiolib-cdev.c
2417
if (copy_from_user(&offset, ip, sizeof(offset)))
drivers/gpio/gpiolib-cdev.c
2436
void __user *ip = (void __user *)arg;
drivers/gpio/gpiolib-cdev.c
244
if (copy_from_user(&ghd, ip, sizeof(ghd)))
drivers/gpio/gpiolib-cdev.c
2447
return chipinfo_get(cdev, ip);
drivers/gpio/gpiolib-cdev.c
2450
return linehandle_create(gdev, ip);
drivers/gpio/gpiolib-cdev.c
2452
return lineevent_create(gdev, ip);
drivers/gpio/gpiolib-cdev.c
2454
return lineinfo_get_v1(cdev, ip, false);
drivers/gpio/gpiolib-cdev.c
2456
return lineinfo_get_v1(cdev, ip, true);
drivers/gpio/gpiolib-cdev.c
2459
return lineinfo_get(cdev, ip, false);
drivers/gpio/gpiolib-cdev.c
2461
return lineinfo_get(cdev, ip, true);
drivers/gpio/gpiolib-cdev.c
2463
return linereq_create(gdev, ip);
drivers/gpio/gpiolib-cdev.c
2465
return lineinfo_unwatch(cdev, ip);
drivers/gpio/gpiolib-cdev.c
259
return linehandle_set_config(lh, ip);
drivers/gpio/gpiolib-cdev.c
303
static int linehandle_create(struct gpio_device *gdev, void __user *ip)
drivers/gpio/gpiolib-cdev.c
310
if (copy_from_user(&handlereq, ip, sizeof(handlereq)))
drivers/gpio/gpiolib-cdev.c
385
if (copy_to_user(ip, &handlereq, sizeof(handlereq)))
drivers/gpu/drm/amd/amdgpu/amdgpu.h
1251
uint8_t ip, uint8_t inst)
drivers/gpu/drm/amd/amdgpu/amdgpu.h
1256
return adev->ip_versions[ip][inst] & ~0xFFU;
drivers/gpu/drm/amd/amdgpu/amdgpu.h
1260
uint8_t ip, uint8_t inst)
drivers/gpu/drm/amd/amdgpu/amdgpu.h
1263
return adev->ip_versions[ip][inst];
drivers/gpu/drm/amd/amdgpu/amdgpu_discovery.c
1136
struct ip_v4 *ip;
drivers/gpu/drm/amd/amdgpu/amdgpu_discovery.c
1139
ip = (struct ip_v4 *)(discovery_bin + ip_offset);
drivers/gpu/drm/amd/amdgpu/amdgpu_discovery.c
1140
inst = ip->instance_number;
drivers/gpu/drm/amd/amdgpu/amdgpu_discovery.c
1141
hw_id = le16_to_cpu(ip->hw_id);
drivers/gpu/drm/amd/amdgpu/amdgpu_discovery.c
1182
ip->num_base_address);
drivers/gpu/drm/amd/amdgpu/amdgpu_discovery.c
1187
ip_hw_instance->hw_id = le16_to_cpu(ip->hw_id); /* == ii */
drivers/gpu/drm/amd/amdgpu/amdgpu_discovery.c
1188
ip_hw_instance->num_instance = ip->instance_number;
drivers/gpu/drm/amd/amdgpu/amdgpu_discovery.c
1189
ip_hw_instance->major = ip->major;
drivers/gpu/drm/amd/amdgpu/amdgpu_discovery.c
1190
ip_hw_instance->minor = ip->minor;
drivers/gpu/drm/amd/amdgpu/amdgpu_discovery.c
1191
ip_hw_instance->revision = ip->revision;
drivers/gpu/drm/amd/amdgpu/amdgpu_discovery.c
1196
ip_hw_instance->num_base_addresses = ip->num_base_address;
drivers/gpu/drm/amd/amdgpu/amdgpu_discovery.c
1201
lower_32_bits(le64_to_cpu(ip->base_address_64[kk])) & 0x3FFFFFFF;
drivers/gpu/drm/amd/amdgpu/amdgpu_discovery.c
1203
ip_hw_instance->base_addr[kk] = ip->base_address[kk];
drivers/gpu/drm/amd/amdgpu/amdgpu_discovery.c
1212
ip_offset += struct_size(ip, base_address_64,
drivers/gpu/drm/amd/amdgpu/amdgpu_discovery.c
1213
ip->num_base_address);
drivers/gpu/drm/amd/amdgpu/amdgpu_discovery.c
1215
ip_offset += struct_size(ip, base_address,
drivers/gpu/drm/amd/amdgpu/amdgpu_discovery.c
1216
ip->num_base_address);
drivers/gpu/drm/amd/amdgpu/amdgpu_discovery.c
1391
struct ip_v4 *ip;
drivers/gpu/drm/amd/amdgpu/amdgpu_discovery.c
1436
ip = (struct ip_v4 *)(discovery_bin + ip_offset);
drivers/gpu/drm/amd/amdgpu/amdgpu_discovery.c
1438
inst = ip->instance_number;
drivers/gpu/drm/amd/amdgpu/amdgpu_discovery.c
1439
hw_id = le16_to_cpu(ip->hw_id);
drivers/gpu/drm/amd/amdgpu/amdgpu_discovery.c
1443
num_base_address = ip->num_base_address;
drivers/gpu/drm/amd/amdgpu/amdgpu_discovery.c
1446
hw_id_names[le16_to_cpu(ip->hw_id)],
drivers/gpu/drm/amd/amdgpu/amdgpu_discovery.c
1447
le16_to_cpu(ip->hw_id),
drivers/gpu/drm/amd/amdgpu/amdgpu_discovery.c
1448
ip->instance_number,
drivers/gpu/drm/amd/amdgpu/amdgpu_discovery.c
1449
ip->major, ip->minor,
drivers/gpu/drm/amd/amdgpu/amdgpu_discovery.c
1450
ip->revision);
drivers/gpu/drm/amd/amdgpu/amdgpu_discovery.c
1452
if (le16_to_cpu(ip->hw_id) == VCN_HWID) {
drivers/gpu/drm/amd/amdgpu/amdgpu_discovery.c
1462
ip->revision & 0xc0;
drivers/gpu/drm/amd/amdgpu/amdgpu_discovery.c
1465
(1U << ip->instance_number);
drivers/gpu/drm/amd/amdgpu/amdgpu_discovery.c
1467
(1U << ip->instance_number);
drivers/gpu/drm/amd/amdgpu/amdgpu_discovery.c
1473
ip->revision &= ~0xc0;
drivers/gpu/drm/amd/amdgpu/amdgpu_discovery.c
1475
if (le16_to_cpu(ip->hw_id) == SDMA0_HWID ||
drivers/gpu/drm/amd/amdgpu/amdgpu_discovery.c
1476
le16_to_cpu(ip->hw_id) == SDMA1_HWID ||
drivers/gpu/drm/amd/amdgpu/amdgpu_discovery.c
1477
le16_to_cpu(ip->hw_id) == SDMA2_HWID ||
drivers/gpu/drm/amd/amdgpu/amdgpu_discovery.c
1478
le16_to_cpu(ip->hw_id) == SDMA3_HWID) {
drivers/gpu/drm/amd/amdgpu/amdgpu_discovery.c
1483
(1U << ip->instance_number);
drivers/gpu/drm/amd/amdgpu/amdgpu_discovery.c
1491
if (le16_to_cpu(ip->hw_id) == VPE_HWID) {
drivers/gpu/drm/amd/amdgpu/amdgpu_discovery.c
1500
if (le16_to_cpu(ip->hw_id) == UMC_HWID) {
drivers/gpu/drm/amd/amdgpu/amdgpu_discovery.c
1505
if (le16_to_cpu(ip->hw_id) == GC_HWID)
drivers/gpu/drm/amd/amdgpu/amdgpu_discovery.c
1507
(1U << ip->instance_number);
drivers/gpu/drm/amd/amdgpu/amdgpu_discovery.c
1509
if (!wafl_ver && le16_to_cpu(ip->hw_id) == WAFLC_HWID)
drivers/gpu/drm/amd/amdgpu/amdgpu_discovery.c
1510
wafl_ver = IP_VERSION_FULL(ip->major, ip->minor,
drivers/gpu/drm/amd/amdgpu/amdgpu_discovery.c
1511
ip->revision, 0, 0);
drivers/gpu/drm/amd/amdgpu/amdgpu_discovery.c
1528
ip->base_address[k] =
drivers/gpu/drm/amd/amdgpu/amdgpu_discovery.c
1529
lower_32_bits(le64_to_cpu(ip->base_address_64[k])) & 0x3FFFFFFF;
drivers/gpu/drm/amd/amdgpu/amdgpu_discovery.c
1531
ip->base_address[k] = le32_to_cpu(ip->base_address[k]);
drivers/gpu/drm/amd/amdgpu/amdgpu_discovery.c
1532
DRM_DEBUG("\t0x%08x\n", ip->base_address[k]);
drivers/gpu/drm/amd/amdgpu/amdgpu_discovery.c
1536
if (hw_id_map[hw_ip] == le16_to_cpu(ip->hw_id) &&
drivers/gpu/drm/amd/amdgpu/amdgpu_discovery.c
1539
hw_id_names[le16_to_cpu(ip->hw_id)]);
drivers/gpu/drm/amd/amdgpu/amdgpu_discovery.c
1540
adev->reg_offset[hw_ip][ip->instance_number] =
drivers/gpu/drm/amd/amdgpu/amdgpu_discovery.c
1541
ip->base_address;
drivers/gpu/drm/amd/amdgpu/amdgpu_discovery.c
1556
subrev = ip->sub_revision;
drivers/gpu/drm/amd/amdgpu/amdgpu_discovery.c
1557
variant = ip->variant;
drivers/gpu/drm/amd/amdgpu/amdgpu_discovery.c
1561
[ip->instance_number] =
drivers/gpu/drm/amd/amdgpu/amdgpu_discovery.c
1562
IP_VERSION_FULL(ip->major,
drivers/gpu/drm/amd/amdgpu/amdgpu_discovery.c
1563
ip->minor,
drivers/gpu/drm/amd/amdgpu/amdgpu_discovery.c
1564
ip->revision,
drivers/gpu/drm/amd/amdgpu/amdgpu_discovery.c
1572
ip_offset += struct_size(ip, base_address_64, ip->num_base_address);
drivers/gpu/drm/amd/amdgpu/amdgpu_discovery.c
1574
ip_offset += struct_size(ip, base_address, ip->num_base_address);
drivers/gpu/drm/amd/amdgpu/amdgpu_discovery.c
710
struct ip *ip;
drivers/gpu/drm/amd/amdgpu/amdgpu_discovery.c
730
ip = (struct ip *)(discovery_bin + ip_offset);
drivers/gpu/drm/amd/amdgpu/amdgpu_discovery.c
731
inst = ip->number_instance;
drivers/gpu/drm/amd/amdgpu/amdgpu_discovery.c
732
hw_id = le16_to_cpu(ip->hw_id);
drivers/gpu/drm/amd/amdgpu/amdgpu_discovery.c
736
if (ip->harvest == 1) {
drivers/gpu/drm/amd/amdgpu/amdgpu_discovery.c
762
ip_offset += struct_size(ip, base_address,
drivers/gpu/drm/amd/amdgpu/amdgpu_discovery.c
763
ip->num_base_address);
drivers/gpu/drm/amd/amdgpu/amdgpu_imu.h
54
#define IMU_RLC_RAM_GOLDEN_VALUE(ip, inst, reg, data, addr_mask) \
drivers/gpu/drm/amd/amdgpu/amdgpu_imu.h
55
{ ip##_HWIP, inst, reg##_BASE_IDX, reg, data, addr_mask }
drivers/gpu/drm/amd/amdgpu/amdgpu_kms.c
175
amdgpu_ip_get_block_type(struct amdgpu_device *adev, uint32_t ip)
drivers/gpu/drm/amd/amdgpu/amdgpu_kms.c
179
switch (ip) {
drivers/gpu/drm/amd/amdgpu/amdgpu_kms.c
694
struct drm_amdgpu_info_hw_ip ip = {};
drivers/gpu/drm/amd/amdgpu/amdgpu_kms.c
696
ret = amdgpu_hw_ip_info(adev, info, &ip);
drivers/gpu/drm/amd/amdgpu/amdgpu_kms.c
700
ret = copy_to_user(out, &ip, min_t(size_t, size, sizeof(ip)));
drivers/gpu/drm/amd/amdgpu/amdgpu_mca.c
539
seq_printf(m, "mca entry[%d].ip: %d\n", idx, entry->ip);
drivers/gpu/drm/amd/amdgpu/amdgpu_mca.h
118
enum amdgpu_mca_ip ip;
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.h
383
#define AMDGPU_RAS_REG_ENTRY(ip, inst, reg_lo, reg_hi) \
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.h
384
ip##_HWIP, inst, reg_lo##_BASE_IDX, reg_lo, reg_hi##_BASE_IDX, reg_hi
drivers/gpu/drm/amd/amdgpu/amdgpu_vcn.h
101
#define SOC15_DPG_MODE_OFFSET(ip, inst_idx, reg) \
drivers/gpu/drm/amd/amdgpu/amdgpu_vcn.h
107
addr = (adev->reg_offset[ip##_HWIP][inst_idx][reg##_BASE_IDX] + reg); \
drivers/gpu/drm/amd/amdgpu/amdgpu_vcn.h
163
#define SOC24_DPG_MODE_OFFSET(ip, inst_idx, reg) \
drivers/gpu/drm/amd/amdgpu/amdgpu_vcn.h
169
addr = (adev->reg_offset[ip##_HWIP][inst_idx][reg##_BASE_IDX] + reg); \
drivers/gpu/drm/amd/amdgpu/amdgpu_vcn.h
80
#define RREG32_SOC15_DPG_MODE_1_0(ip, inst_idx, reg, mask, sram_sel) \
drivers/gpu/drm/amd/amdgpu/amdgpu_vcn.h
81
({ WREG32_SOC15(ip, inst_idx, mmUVD_DPG_LMA_MASK, mask); \
drivers/gpu/drm/amd/amdgpu/amdgpu_vcn.h
82
WREG32_SOC15(ip, inst_idx, mmUVD_DPG_LMA_CTL, \
drivers/gpu/drm/amd/amdgpu/amdgpu_vcn.h
84
((adev->reg_offset[ip##_HWIP][inst_idx][reg##_BASE_IDX] + reg) \
drivers/gpu/drm/amd/amdgpu/amdgpu_vcn.h
87
RREG32_SOC15(ip, inst_idx, mmUVD_DPG_LMA_DATA); \
drivers/gpu/drm/amd/amdgpu/amdgpu_vcn.h
90
#define WREG32_SOC15_DPG_MODE_1_0(ip, inst_idx, reg, value, mask, sram_sel) \
drivers/gpu/drm/amd/amdgpu/amdgpu_vcn.h
92
WREG32_SOC15(ip, inst_idx, mmUVD_DPG_LMA_DATA, value); \
drivers/gpu/drm/amd/amdgpu/amdgpu_vcn.h
93
WREG32_SOC15(ip, inst_idx, mmUVD_DPG_LMA_MASK, mask); \
drivers/gpu/drm/amd/amdgpu/amdgpu_vcn.h
94
WREG32_SOC15(ip, inst_idx, mmUVD_DPG_LMA_CTL, \
drivers/gpu/drm/amd/amdgpu/amdgpu_vcn.h
96
((adev->reg_offset[ip##_HWIP][inst_idx][reg##_BASE_IDX] + reg) \
drivers/gpu/drm/amd/amdgpu/amdgpu_xcp.c
109
struct amdgpu_xcp_ip *ip)
drivers/gpu/drm/amd/amdgpu/amdgpu_xcp.c
113
if (!ip)
drivers/gpu/drm/amd/amdgpu/amdgpu_xcp.c
117
xcp->ip[ip->ip_id] = *ip;
drivers/gpu/drm/amd/amdgpu/amdgpu_xcp.c
118
xcp->ip[ip->ip_id].valid = true;
drivers/gpu/drm/amd/amdgpu/amdgpu_xcp.c
145
struct amdgpu_xcp_ip ip;
drivers/gpu/drm/amd/amdgpu/amdgpu_xcp.c
163
&ip);
drivers/gpu/drm/amd/amdgpu/amdgpu_xcp.c
167
__amdgpu_xcp_add_block(xcp_mgr, i, &ip);
drivers/gpu/drm/amd/amdgpu/amdgpu_xcp.c
358
enum AMDGPU_XCP_IP_BLOCK ip, int instance)
drivers/gpu/drm/amd/amdgpu/amdgpu_xcp.c
363
if (ip >= AMDGPU_XCP_MAX_BLOCKS)
drivers/gpu/drm/amd/amdgpu/amdgpu_xcp.c
368
if ((xcp->valid) && (xcp->ip[ip].valid) &&
drivers/gpu/drm/amd/amdgpu/amdgpu_xcp.c
369
(xcp->ip[ip].inst_mask & BIT(instance)))
drivers/gpu/drm/amd/amdgpu/amdgpu_xcp.c
380
enum AMDGPU_XCP_IP_BLOCK ip,
drivers/gpu/drm/amd/amdgpu/amdgpu_xcp.c
383
if (!xcp->valid || !inst_mask || !(xcp->ip[ip].valid))
drivers/gpu/drm/amd/amdgpu/amdgpu_xcp.c
386
*inst_mask = xcp->ip[ip].inst_mask;
drivers/gpu/drm/amd/amdgpu/amdgpu_xcp.c
559
if (adev->xcp_mgr->xcp[xcp_id].ip[ip_blk].inst_mask & inst_mask) {
drivers/gpu/drm/amd/amdgpu/amdgpu_xcp.c
77
xcp_ip = &xcp->ip[i];
drivers/gpu/drm/amd/amdgpu/amdgpu_xcp.h
104
struct amdgpu_xcp_ip ip[AMDGPU_XCP_MAX_BLOCKS];
drivers/gpu/drm/amd/amdgpu/amdgpu_xcp.h
143
struct amdgpu_xcp_ip *ip);
drivers/gpu/drm/amd/amdgpu/amdgpu_xcp.h
167
enum AMDGPU_XCP_IP_BLOCK ip, int instance);
drivers/gpu/drm/amd/amdgpu/amdgpu_xcp.h
170
enum AMDGPU_XCP_IP_BLOCK ip,
drivers/gpu/drm/amd/amdgpu/aqua_vanjaram.c
166
struct amdgpu_xcp_ip *ip)
drivers/gpu/drm/amd/amdgpu/aqua_vanjaram.c
197
ip->inst_mask = XCP_INST_MASK(num_xcc_xcp, xcp_id);
drivers/gpu/drm/amd/amdgpu/aqua_vanjaram.c
198
ip->ip_funcs = &gfxhub_v1_2_xcp_funcs;
drivers/gpu/drm/amd/amdgpu/aqua_vanjaram.c
201
ip->inst_mask = XCP_INST_MASK(num_xcc_xcp, xcp_id);
drivers/gpu/drm/amd/amdgpu/aqua_vanjaram.c
202
ip->ip_funcs = &gfx_v9_4_3_xcp_funcs;
drivers/gpu/drm/amd/amdgpu/aqua_vanjaram.c
205
ip->inst_mask = XCP_INST_MASK(num_sdma_xcp, xcp_id);
drivers/gpu/drm/amd/amdgpu/aqua_vanjaram.c
206
ip->ip_funcs = &sdma_v4_4_2_xcp_funcs;
drivers/gpu/drm/amd/amdgpu/aqua_vanjaram.c
209
ip->inst_mask =
drivers/gpu/drm/amd/amdgpu/aqua_vanjaram.c
217
ip->ip_id = ip_id;
drivers/gpu/drm/amd/amdgpu/aqua_vanjaram.c
499
struct amdgpu_xcp_ip *ip)
drivers/gpu/drm/amd/amdgpu/aqua_vanjaram.c
501
if (!ip)
drivers/gpu/drm/amd/amdgpu/aqua_vanjaram.c
504
return __aqua_vanjaram_get_xcp_ip_info(xcp_mgr, xcp_id, ip_id, ip);
drivers/gpu/drm/amd/amdgpu/soc15.h
100
#define SOC15_REG_GOLDEN_VALUE(ip, inst, reg, and_mask, or_mask) \
drivers/gpu/drm/amd/amdgpu/soc15.h
101
{ ip##_HWIP, inst, reg##_BASE_IDX, reg, and_mask, or_mask }
drivers/gpu/drm/amd/amdgpu/soc15.h
90
#define SOC15_REG_ENTRY(ip, inst, reg) ip##_HWIP, inst, reg##_BASE_IDX, reg
drivers/gpu/drm/amd/amdgpu/soc15.h
91
#define SOC15_REG_ENTRY_STR(ip, inst, reg) \
drivers/gpu/drm/amd/amdgpu/soc15.h
92
{ ip##_HWIP, inst, reg##_BASE_IDX, reg, #reg }
drivers/gpu/drm/amd/amdgpu/soc15_common.h
100
#define SOC15_WAIT_ON_RREG(ip, inst, reg, expected_value, mask) \
drivers/gpu/drm/amd/amdgpu/soc15_common.h
102
(adev->reg_offset[ip##_HWIP][inst][reg##_BASE_IDX] + (reg)), \
drivers/gpu/drm/amd/amdgpu/soc15_common.h
105
#define SOC15_WAIT_ON_RREG_OFFSET(ip, inst, reg, offset, expected_value, mask) \
drivers/gpu/drm/amd/amdgpu/soc15_common.h
107
(adev->reg_offset[ip##_HWIP][inst][reg##_BASE_IDX] + (reg) + (offset)), \
drivers/gpu/drm/amd/amdgpu/soc15_common.h
138
#define WREG32_SOC15_RLC_SHADOW(ip, inst, reg, value) \
drivers/gpu/drm/amd/amdgpu/soc15_common.h
139
__WREG32_SOC15_RLC__((adev->reg_offset[ip##_HWIP][inst][reg##_BASE_IDX] + reg), value, AMDGPU_REGS_RLC, GC_HWIP, inst)
drivers/gpu/drm/amd/amdgpu/soc15_common.h
151
#define WREG32_SOC15_RLC_SHADOW_EX(prefix, ip, inst, reg, value) \
drivers/gpu/drm/amd/amdgpu/soc15_common.h
153
uint32_t target_reg = adev->reg_offset[ip##_HWIP][inst][reg##_BASE_IDX] + reg;\
drivers/gpu/drm/amd/amdgpu/soc15_common.h
169
#define RREG32_SOC15_RLC(ip, inst, reg) \
drivers/gpu/drm/amd/amdgpu/soc15_common.h
170
__RREG32_SOC15_RLC__(adev->reg_offset[ip##_HWIP][inst][reg##_BASE_IDX] + reg, AMDGPU_REGS_RLC, ip##_HWIP, inst)
drivers/gpu/drm/amd/amdgpu/soc15_common.h
172
#define WREG32_SOC15_RLC(ip, inst, reg, value) \
drivers/gpu/drm/amd/amdgpu/soc15_common.h
174
uint32_t target_reg = adev->reg_offset[ip##_HWIP][inst][reg##_BASE_IDX] + reg;\
drivers/gpu/drm/amd/amdgpu/soc15_common.h
175
__WREG32_SOC15_RLC__(target_reg, value, AMDGPU_REGS_RLC, ip##_HWIP, inst); \
drivers/gpu/drm/amd/amdgpu/soc15_common.h
178
#define WREG32_SOC15_RLC_EX(prefix, ip, inst, reg, value) \
drivers/gpu/drm/amd/amdgpu/soc15_common.h
184
#define WREG32_FIELD15_RLC(ip, idx, reg, field, val) \
drivers/gpu/drm/amd/amdgpu/soc15_common.h
185
__WREG32_SOC15_RLC__((adev->reg_offset[ip##_HWIP][idx][mm##reg##_BASE_IDX] + mm##reg), \
drivers/gpu/drm/amd/amdgpu/soc15_common.h
186
(__RREG32_SOC15_RLC__(adev->reg_offset[ip##_HWIP][idx][mm##reg##_BASE_IDX] + mm##reg, \
drivers/gpu/drm/amd/amdgpu/soc15_common.h
187
AMDGPU_REGS_RLC, ip##_HWIP, idx) & \
drivers/gpu/drm/amd/amdgpu/soc15_common.h
189
AMDGPU_REGS_RLC, ip##_HWIP, idx)
drivers/gpu/drm/amd/amdgpu/soc15_common.h
191
#define WREG32_SOC15_OFFSET_RLC(ip, inst, reg, offset, value) \
drivers/gpu/drm/amd/amdgpu/soc15_common.h
192
__WREG32_SOC15_RLC__((adev->reg_offset[ip##_HWIP][inst][reg##_BASE_IDX] + reg) + offset, value, AMDGPU_REGS_RLC, ip##_HWIP, inst)
drivers/gpu/drm/amd/amdgpu/soc15_common.h
194
#define RREG32_SOC15_OFFSET_RLC(ip, inst, reg, offset) \
drivers/gpu/drm/amd/amdgpu/soc15_common.h
195
__RREG32_SOC15_RLC__((adev->reg_offset[ip##_HWIP][inst][reg##_BASE_IDX] + reg) + offset, AMDGPU_REGS_RLC, ip##_HWIP, inst)
drivers/gpu/drm/amd/amdgpu/soc15_common.h
198
#define RREG32_SOC15_EXT(ip, inst, reg, ext) \
drivers/gpu/drm/amd/amdgpu/soc15_common.h
199
RREG32_PCIE_EXT((adev->reg_offset[ip##_HWIP][inst][reg##_BASE_IDX] + reg) * 4 \
drivers/gpu/drm/amd/amdgpu/soc15_common.h
202
#define WREG32_SOC15_EXT(ip, inst, reg, ext, value) \
drivers/gpu/drm/amd/amdgpu/soc15_common.h
203
WREG32_PCIE_EXT((adev->reg_offset[ip##_HWIP][inst][reg##_BASE_IDX] + reg) * 4 \
drivers/gpu/drm/amd/amdgpu/soc15_common.h
28
#define GET_INST(ip, inst) \
drivers/gpu/drm/amd/amdgpu/soc15_common.h
30
adev->ip_map.logical_to_dev_inst(adev, ip##_HWIP, inst) : inst)
drivers/gpu/drm/amd/amdgpu/soc15_common.h
31
#define GET_MASK(ip, mask) \
drivers/gpu/drm/amd/amdgpu/soc15_common.h
33
adev->ip_map.logical_to_dev_mask(adev, ip##_HWIP, mask) : mask)
drivers/gpu/drm/amd/amdgpu/soc15_common.h
36
#define SOC15_REG_OFFSET(ip, inst, reg) (adev->reg_offset[ip##_HWIP][inst][reg##_BASE_IDX] + reg)
drivers/gpu/drm/amd/amdgpu/soc15_common.h
37
#define SOC15_REG_OFFSET1(ip, inst, reg, offset) \
drivers/gpu/drm/amd/amdgpu/soc15_common.h
38
(adev->reg_offset[ip##_HWIP][inst][reg##_BASE_IDX] + (reg)+(offset))
drivers/gpu/drm/amd/amdgpu/soc15_common.h
50
#define WREG32_FIELD15(ip, idx, reg, field, val) \
drivers/gpu/drm/amd/amdgpu/soc15_common.h
51
__WREG32_SOC15_RLC__(adev->reg_offset[ip##_HWIP][idx][mm##reg##_BASE_IDX] + mm##reg, \
drivers/gpu/drm/amd/amdgpu/soc15_common.h
53
adev->reg_offset[ip##_HWIP][idx][mm##reg##_BASE_IDX] + mm##reg, \
drivers/gpu/drm/amd/amdgpu/soc15_common.h
54
0, ip##_HWIP, idx) & \
drivers/gpu/drm/amd/amdgpu/soc15_common.h
56
0, ip##_HWIP, idx)
drivers/gpu/drm/amd/amdgpu/soc15_common.h
58
#define WREG32_FIELD15_PREREG(ip, idx, reg_name, field, val) \
drivers/gpu/drm/amd/amdgpu/soc15_common.h
59
__WREG32_SOC15_RLC__(adev->reg_offset[ip##_HWIP][idx][reg##reg_name##_BASE_IDX] + reg##reg_name, \
drivers/gpu/drm/amd/amdgpu/soc15_common.h
61
adev->reg_offset[ip##_HWIP][idx][reg##reg_name##_BASE_IDX] + reg##reg_name, \
drivers/gpu/drm/amd/amdgpu/soc15_common.h
62
0, ip##_HWIP, idx) & \
drivers/gpu/drm/amd/amdgpu/soc15_common.h
64
0, ip##_HWIP, idx)
drivers/gpu/drm/amd/amdgpu/soc15_common.h
66
#define RREG32_SOC15(ip, inst, reg) \
drivers/gpu/drm/amd/amdgpu/soc15_common.h
67
__RREG32_SOC15_RLC__(adev->reg_offset[ip##_HWIP][inst][reg##_BASE_IDX] + reg, \
drivers/gpu/drm/amd/amdgpu/soc15_common.h
68
0, ip##_HWIP, inst)
drivers/gpu/drm/amd/amdgpu/soc15_common.h
70
#define RREG32_SOC15_IP(ip, reg) __RREG32_SOC15_RLC__(reg, 0, ip##_HWIP, 0)
drivers/gpu/drm/amd/amdgpu/soc15_common.h
72
#define RREG32_SOC15_IP_NO_KIQ(ip, reg, inst) __RREG32_SOC15_RLC__(reg, AMDGPU_REGS_NO_KIQ, ip##_HWIP, inst)
drivers/gpu/drm/amd/amdgpu/soc15_common.h
74
#define RREG32_SOC15_NO_KIQ(ip, inst, reg) \
drivers/gpu/drm/amd/amdgpu/soc15_common.h
75
__RREG32_SOC15_RLC__(adev->reg_offset[ip##_HWIP][inst][reg##_BASE_IDX] + reg, \
drivers/gpu/drm/amd/amdgpu/soc15_common.h
76
AMDGPU_REGS_NO_KIQ, ip##_HWIP, inst)
drivers/gpu/drm/amd/amdgpu/soc15_common.h
78
#define RREG32_SOC15_OFFSET(ip, inst, reg, offset) \
drivers/gpu/drm/amd/amdgpu/soc15_common.h
79
__RREG32_SOC15_RLC__((adev->reg_offset[ip##_HWIP][inst][reg##_BASE_IDX] + (reg)) + \
drivers/gpu/drm/amd/amdgpu/soc15_common.h
80
(offset), 0, ip##_HWIP, inst)
drivers/gpu/drm/amd/amdgpu/soc15_common.h
82
#define WREG32_SOC15(ip, inst, reg, value) \
drivers/gpu/drm/amd/amdgpu/soc15_common.h
83
__WREG32_SOC15_RLC__((adev->reg_offset[ip##_HWIP][inst][reg##_BASE_IDX] + reg), \
drivers/gpu/drm/amd/amdgpu/soc15_common.h
84
value, 0, ip##_HWIP, inst)
drivers/gpu/drm/amd/amdgpu/soc15_common.h
86
#define WREG32_SOC15_IP(ip, reg, value) \
drivers/gpu/drm/amd/amdgpu/soc15_common.h
87
__WREG32_SOC15_RLC__(reg, value, 0, ip##_HWIP, 0)
drivers/gpu/drm/amd/amdgpu/soc15_common.h
89
#define WREG32_SOC15_IP_NO_KIQ(ip, reg, value, inst) \
drivers/gpu/drm/amd/amdgpu/soc15_common.h
90
__WREG32_SOC15_RLC__(reg, value, AMDGPU_REGS_NO_KIQ, ip##_HWIP, inst)
drivers/gpu/drm/amd/amdgpu/soc15_common.h
92
#define WREG32_SOC15_NO_KIQ(ip, inst, reg, value) \
drivers/gpu/drm/amd/amdgpu/soc15_common.h
93
__WREG32_SOC15_RLC__(adev->reg_offset[ip##_HWIP][inst][reg##_BASE_IDX] + reg, \
drivers/gpu/drm/amd/amdgpu/soc15_common.h
94
value, AMDGPU_REGS_NO_KIQ, ip##_HWIP, inst)
drivers/gpu/drm/amd/amdgpu/soc15_common.h
96
#define WREG32_SOC15_OFFSET(ip, inst, reg, offset, value) \
drivers/gpu/drm/amd/amdgpu/soc15_common.h
97
__WREG32_SOC15_RLC__((adev->reg_offset[ip##_HWIP][inst][reg##_BASE_IDX] + reg) + offset, \
drivers/gpu/drm/amd/amdgpu/soc15_common.h
98
value, 0, ip##_HWIP, inst)
drivers/gpu/drm/amd/amdgpu/soc_v1_0.c
465
struct amdgpu_xcp_ip *ip)
drivers/gpu/drm/amd/amdgpu/soc_v1_0.c
496
ip->inst_mask = XCP_INST_MASK(num_xcc_xcp, xcp_id);
drivers/gpu/drm/amd/amdgpu/soc_v1_0.c
497
ip->ip_funcs = &gfxhub_v12_1_xcp_funcs;
drivers/gpu/drm/amd/amdgpu/soc_v1_0.c
500
ip->inst_mask = XCP_INST_MASK(num_xcc_xcp, xcp_id);
drivers/gpu/drm/amd/amdgpu/soc_v1_0.c
501
ip->ip_funcs = &gfx_v12_1_xcp_funcs;
drivers/gpu/drm/amd/amdgpu/soc_v1_0.c
504
ip->inst_mask = XCP_INST_MASK(num_sdma_xcp, xcp_id);
drivers/gpu/drm/amd/amdgpu/soc_v1_0.c
505
ip->ip_funcs = &sdma_v7_1_xcp_funcs;
drivers/gpu/drm/amd/amdgpu/soc_v1_0.c
508
ip->inst_mask =
drivers/gpu/drm/amd/amdgpu/soc_v1_0.c
516
ip->ip_id = ip_id;
drivers/gpu/drm/amd/amdgpu/soc_v1_0.c
778
struct amdgpu_xcp_ip *ip)
drivers/gpu/drm/amd/amdgpu/soc_v1_0.c
780
if (!ip)
drivers/gpu/drm/amd/amdgpu/soc_v1_0.c
783
return __soc_v1_0_get_xcp_ip_info(xcp_mgr, xcp_id, ip_id, ip);
drivers/gpu/drm/amd/display/dc/dml/calcs/dcn_calcs.c
1585
dc->dml.ip.rob_buffer_size_kbytes = dc->dcn_ip->rob_buffer_size_in_kbyte;
drivers/gpu/drm/amd/display/dc/dml/calcs/dcn_calcs.c
1586
dc->dml.ip.det_buffer_size_kbytes = dc->dcn_ip->det_buffer_size_in_kbyte;
drivers/gpu/drm/amd/display/dc/dml/calcs/dcn_calcs.c
1587
dc->dml.ip.dpp_output_buffer_pixels = dc->dcn_ip->dpp_output_buffer_pixels;
drivers/gpu/drm/amd/display/dc/dml/calcs/dcn_calcs.c
1588
dc->dml.ip.opp_output_buffer_lines = dc->dcn_ip->opp_output_buffer_lines;
drivers/gpu/drm/amd/display/dc/dml/calcs/dcn_calcs.c
1589
dc->dml.ip.pixel_chunk_size_kbytes = dc->dcn_ip->pixel_chunk_size_in_kbyte;
drivers/gpu/drm/amd/display/dc/dml/calcs/dcn_calcs.c
1590
dc->dml.ip.pte_enable = dc->dcn_ip->pte_enable == dcn_bw_yes;
drivers/gpu/drm/amd/display/dc/dml/calcs/dcn_calcs.c
1591
dc->dml.ip.pte_chunk_size_kbytes = dc->dcn_ip->pte_chunk_size;
drivers/gpu/drm/amd/display/dc/dml/calcs/dcn_calcs.c
1592
dc->dml.ip.meta_chunk_size_kbytes = dc->dcn_ip->meta_chunk_size;
drivers/gpu/drm/amd/display/dc/dml/calcs/dcn_calcs.c
1593
dc->dml.ip.writeback_chunk_size_kbytes = dc->dcn_ip->writeback_chunk_size;
drivers/gpu/drm/amd/display/dc/dml/calcs/dcn_calcs.c
1594
dc->dml.ip.line_buffer_size_bits = dc->dcn_ip->line_buffer_size;
drivers/gpu/drm/amd/display/dc/dml/calcs/dcn_calcs.c
1595
dc->dml.ip.max_line_buffer_lines = dc->dcn_ip->max_line_buffer_lines;
drivers/gpu/drm/amd/display/dc/dml/calcs/dcn_calcs.c
1596
dc->dml.ip.IsLineBufferBppFixed = dc->dcn_ip->is_line_buffer_bpp_fixed == dcn_bw_yes;
drivers/gpu/drm/amd/display/dc/dml/calcs/dcn_calcs.c
1597
dc->dml.ip.LineBufferFixedBpp = dc->dcn_ip->line_buffer_fixed_bpp;
drivers/gpu/drm/amd/display/dc/dml/calcs/dcn_calcs.c
1598
dc->dml.ip.writeback_luma_buffer_size_kbytes = dc->dcn_ip->writeback_luma_buffer_size;
drivers/gpu/drm/amd/display/dc/dml/calcs/dcn_calcs.c
1599
dc->dml.ip.writeback_chroma_buffer_size_kbytes = dc->dcn_ip->writeback_chroma_buffer_size;
drivers/gpu/drm/amd/display/dc/dml/calcs/dcn_calcs.c
1600
dc->dml.ip.max_num_dpp = dc->dcn_ip->max_num_dpp;
drivers/gpu/drm/amd/display/dc/dml/calcs/dcn_calcs.c
1601
dc->dml.ip.max_num_wb = dc->dcn_ip->max_num_writeback;
drivers/gpu/drm/amd/display/dc/dml/calcs/dcn_calcs.c
1602
dc->dml.ip.max_dchub_pscl_bw_pix_per_clk = dc->dcn_ip->max_dchub_topscl_throughput;
drivers/gpu/drm/amd/display/dc/dml/calcs/dcn_calcs.c
1603
dc->dml.ip.max_pscl_lb_bw_pix_per_clk = dc->dcn_ip->max_pscl_tolb_throughput;
drivers/gpu/drm/amd/display/dc/dml/calcs/dcn_calcs.c
1604
dc->dml.ip.max_lb_vscl_bw_pix_per_clk = dc->dcn_ip->max_lb_tovscl_throughput;
drivers/gpu/drm/amd/display/dc/dml/calcs/dcn_calcs.c
1605
dc->dml.ip.max_vscl_hscl_bw_pix_per_clk = dc->dcn_ip->max_vscl_tohscl_throughput;
drivers/gpu/drm/amd/display/dc/dml/calcs/dcn_calcs.c
1606
dc->dml.ip.max_hscl_ratio = dc->dcn_ip->max_hscl_ratio;
drivers/gpu/drm/amd/display/dc/dml/calcs/dcn_calcs.c
1607
dc->dml.ip.max_vscl_ratio = dc->dcn_ip->max_vscl_ratio;
drivers/gpu/drm/amd/display/dc/dml/calcs/dcn_calcs.c
1608
dc->dml.ip.max_hscl_taps = dc->dcn_ip->max_hscl_taps;
drivers/gpu/drm/amd/display/dc/dml/calcs/dcn_calcs.c
1609
dc->dml.ip.max_vscl_taps = dc->dcn_ip->max_vscl_taps;
drivers/gpu/drm/amd/display/dc/dml/calcs/dcn_calcs.c
1611
dc->dml.ip.dispclk_ramp_margin_percent = dc->dcn_ip->dispclk_ramping_margin;
drivers/gpu/drm/amd/display/dc/dml/calcs/dcn_calcs.c
1612
dc->dml.ip.underscan_factor = dc->dcn_ip->under_scan_factor;
drivers/gpu/drm/amd/display/dc/dml/calcs/dcn_calcs.c
1613
dc->dml.ip.max_inter_dcn_tile_repeaters = dc->dcn_ip->max_inter_dcn_tile_repeaters;
drivers/gpu/drm/amd/display/dc/dml/calcs/dcn_calcs.c
1614
dc->dml.ip.can_vstartup_lines_exceed_vsync_plus_back_porch_lines_minus_one =
drivers/gpu/drm/amd/display/dc/dml/calcs/dcn_calcs.c
1616
dc->dml.ip.bug_forcing_LC_req_same_size_fixed =
drivers/gpu/drm/amd/display/dc/dml/calcs/dcn_calcs.c
1618
dc->dml.ip.dcfclk_cstate_latency = dc->dcn_ip->dcfclk_cstate_latency;
drivers/gpu/drm/amd/display/dc/dml/dcn10/dcn10_fpu.c
135
dml->ip.max_num_dpp = 3;
drivers/gpu/drm/amd/display/dc/dml/dcn20/dcn20_fpu.c
1194
context->res_ctx.pipe_ctx[i].det_buffer_size_kb = context->bw_ctx.dml.ip.det_buffer_size_kbytes;
drivers/gpu/drm/amd/display/dc/dml/dcn20/dcn20_fpu.c
1221
context->bw_ctx.bw.dcn.compbuf_size_kb = context->bw_ctx.dml.ip.config_return_buffer_size_in_kbytes
drivers/gpu/drm/amd/display/dc/dml/dcn20/dcn20_fpu.c
1222
- context->bw_ctx.dml.ip.det_buffer_size_kbytes * pipe_idx;
drivers/gpu/drm/amd/display/dc/dml/dcn20/dcn20_fpu.c
1542
pipes[pipe_cnt].pipe.src.num_cursors = dc->dml.ip.number_of_cursors;
drivers/gpu/drm/amd/display/dc/dml/dcn20/display_rq_dlg_calc_20.c
1006
dppclk_delay_subtotal = mode_lib->ip.dppclk_delay_subtotal;
drivers/gpu/drm/amd/display/dc/dml/dcn20/display_rq_dlg_calc_20.c
1007
dispclk_delay_subtotal = mode_lib->ip.dispclk_delay_subtotal;
drivers/gpu/drm/amd/display/dc/dml/dcn20/display_rq_dlg_calc_20.c
1010
dppclk_delay_subtotal += mode_lib->ip.dppclk_delay_scl;
drivers/gpu/drm/amd/display/dc/dml/dcn20/display_rq_dlg_calc_20.c
1012
dppclk_delay_subtotal += mode_lib->ip.dppclk_delay_scl_lb_only;
drivers/gpu/drm/amd/display/dc/dml/dcn20/display_rq_dlg_calc_20.c
1014
dppclk_delay_subtotal += mode_lib->ip.dppclk_delay_cnvc_formatter
drivers/gpu/drm/amd/display/dc/dml/dcn20/display_rq_dlg_calc_20.c
1015
+ src->num_cursors * mode_lib->ip.dppclk_delay_cnvc_cursor;
drivers/gpu/drm/amd/display/dc/dml/dcn20/display_rq_dlg_calc_20.c
193
unsigned int detile_buf_size_in_bytes = mode_lib->ip.det_buffer_size_kbytes * 1024;
drivers/gpu/drm/amd/display/dc/dml/dcn20/display_rq_dlg_calc_20.c
249
unsigned int detile_buf_size_in_bytes = mode_lib->ip.det_buffer_size_kbytes * 1024;
drivers/gpu/drm/amd/display/dc/dml/dcn20/display_rq_dlg_calc_20.c
394
const unsigned int dpte_buf_in_pte_reqs = mode_lib->ip.dpte_buffer_size_in_pte_reqs_luma;
drivers/gpu/drm/amd/display/dc/dml/dcn20/display_rq_dlg_calc_20.c
396
mode_lib->ip.pde_proc_buffer_size_64k_reqs;
drivers/gpu/drm/amd/display/dc/dml/dcn20/display_rq_dlg_calc_20.c
803
unsigned int min_vblank = mode_lib->ip.min_vblank_lines;
drivers/gpu/drm/amd/display/dc/dml/dcn20/display_rq_dlg_calc_20v2.c
1007
dppclk_delay_subtotal = mode_lib->ip.dppclk_delay_subtotal;
drivers/gpu/drm/amd/display/dc/dml/dcn20/display_rq_dlg_calc_20v2.c
1008
dispclk_delay_subtotal = mode_lib->ip.dispclk_delay_subtotal;
drivers/gpu/drm/amd/display/dc/dml/dcn20/display_rq_dlg_calc_20v2.c
1011
dppclk_delay_subtotal += mode_lib->ip.dppclk_delay_scl;
drivers/gpu/drm/amd/display/dc/dml/dcn20/display_rq_dlg_calc_20v2.c
1013
dppclk_delay_subtotal += mode_lib->ip.dppclk_delay_scl_lb_only;
drivers/gpu/drm/amd/display/dc/dml/dcn20/display_rq_dlg_calc_20v2.c
1015
dppclk_delay_subtotal += mode_lib->ip.dppclk_delay_cnvc_formatter
drivers/gpu/drm/amd/display/dc/dml/dcn20/display_rq_dlg_calc_20v2.c
1016
+ src->num_cursors * mode_lib->ip.dppclk_delay_cnvc_cursor;
drivers/gpu/drm/amd/display/dc/dml/dcn20/display_rq_dlg_calc_20v2.c
193
unsigned int detile_buf_size_in_bytes = mode_lib->ip.det_buffer_size_kbytes * 1024;
drivers/gpu/drm/amd/display/dc/dml/dcn20/display_rq_dlg_calc_20v2.c
249
unsigned int detile_buf_size_in_bytes = mode_lib->ip.det_buffer_size_kbytes * 1024;
drivers/gpu/drm/amd/display/dc/dml/dcn20/display_rq_dlg_calc_20v2.c
394
const unsigned int dpte_buf_in_pte_reqs = mode_lib->ip.dpte_buffer_size_in_pte_reqs_luma;
drivers/gpu/drm/amd/display/dc/dml/dcn20/display_rq_dlg_calc_20v2.c
396
mode_lib->ip.pde_proc_buffer_size_64k_reqs;
drivers/gpu/drm/amd/display/dc/dml/dcn20/display_rq_dlg_calc_20v2.c
803
unsigned int min_vblank = mode_lib->ip.min_vblank_lines;
drivers/gpu/drm/amd/display/dc/dml/dcn21/display_rq_dlg_calc_21.c
1046
dppclk_delay_subtotal = mode_lib->ip.dppclk_delay_subtotal;
drivers/gpu/drm/amd/display/dc/dml/dcn21/display_rq_dlg_calc_21.c
1047
dispclk_delay_subtotal = mode_lib->ip.dispclk_delay_subtotal;
drivers/gpu/drm/amd/display/dc/dml/dcn21/display_rq_dlg_calc_21.c
1050
dppclk_delay_subtotal += mode_lib->ip.dppclk_delay_scl;
drivers/gpu/drm/amd/display/dc/dml/dcn21/display_rq_dlg_calc_21.c
1052
dppclk_delay_subtotal += mode_lib->ip.dppclk_delay_scl_lb_only;
drivers/gpu/drm/amd/display/dc/dml/dcn21/display_rq_dlg_calc_21.c
1054
dppclk_delay_subtotal += mode_lib->ip.dppclk_delay_cnvc_formatter
drivers/gpu/drm/amd/display/dc/dml/dcn21/display_rq_dlg_calc_21.c
1055
+ src->num_cursors * mode_lib->ip.dppclk_delay_cnvc_cursor;
drivers/gpu/drm/amd/display/dc/dml/dcn21/display_rq_dlg_calc_21.c
171
unsigned int detile_buf_size_in_bytes = mode_lib->ip.det_buffer_size_kbytes * 1024;
drivers/gpu/drm/amd/display/dc/dml/dcn21/display_rq_dlg_calc_21.c
231
unsigned int detile_buf_size_in_bytes = mode_lib->ip.det_buffer_size_kbytes * 1024;
drivers/gpu/drm/amd/display/dc/dml/dcn21/display_rq_dlg_calc_21.c
385
mode_lib->ip.dpte_buffer_size_in_pte_reqs_luma + mode_lib->ip.dpte_buffer_size_in_pte_reqs_chroma;
drivers/gpu/drm/amd/display/dc/dml/dcn21/display_rq_dlg_calc_21.c
387
mode_lib->ip.pde_proc_buffer_size_64k_reqs;
drivers/gpu/drm/amd/display/dc/dml/dcn21/display_rq_dlg_calc_21.c
849
unsigned int min_vblank = mode_lib->ip.min_vblank_lines;
drivers/gpu/drm/amd/display/dc/dml/dcn30/dcn30_fpu.c
218
if (dc->dml.ip.writeback_max_hscl_taps > 1) {
drivers/gpu/drm/amd/display/dc/dml/dcn30/dcn30_fpu.c
258
dc->current_state->bw_ctx.dml.ip.writeback_line_buffer_buffer_size);
drivers/gpu/drm/amd/display/dc/dml/dcn30/display_rq_dlg_calc_30.c
1107
dppclk_delay_subtotal = mode_lib->ip.dppclk_delay_subtotal;
drivers/gpu/drm/amd/display/dc/dml/dcn30/display_rq_dlg_calc_30.c
1108
dispclk_delay_subtotal = mode_lib->ip.dispclk_delay_subtotal;
drivers/gpu/drm/amd/display/dc/dml/dcn30/display_rq_dlg_calc_30.c
1111
dppclk_delay_subtotal += mode_lib->ip.dppclk_delay_scl;
drivers/gpu/drm/amd/display/dc/dml/dcn30/display_rq_dlg_calc_30.c
1113
dppclk_delay_subtotal += mode_lib->ip.dppclk_delay_scl_lb_only;
drivers/gpu/drm/amd/display/dc/dml/dcn30/display_rq_dlg_calc_30.c
1115
dppclk_delay_subtotal += mode_lib->ip.dppclk_delay_cnvc_formatter
drivers/gpu/drm/amd/display/dc/dml/dcn30/display_rq_dlg_calc_30.c
1116
+ src->num_cursors * mode_lib->ip.dppclk_delay_cnvc_cursor;
drivers/gpu/drm/amd/display/dc/dml/dcn30/display_rq_dlg_calc_30.c
117
unsigned int detile_buf_size_in_bytes = mode_lib->ip.det_buffer_size_kbytes * 1024;
drivers/gpu/drm/amd/display/dc/dml/dcn30/display_rq_dlg_calc_30.c
1425
unsigned int levels = mode_lib->ip.gpuvm_max_page_table_levels;
drivers/gpu/drm/amd/display/dc/dml/dcn30/display_rq_dlg_calc_30.c
1429
levels = levels * (mode_lib->ip.hostvm_max_page_table_levels+1);
drivers/gpu/drm/amd/display/dc/dml/dcn30/display_rq_dlg_calc_30.c
173
unsigned int detile_buf_size_in_bytes = mode_lib->ip.det_buffer_size_kbytes * 1024;
drivers/gpu/drm/amd/display/dc/dml/dcn30/display_rq_dlg_calc_30.c
342
(is_chroma ? mode_lib->ip.dpte_buffer_size_in_pte_reqs_chroma : mode_lib->ip.dpte_buffer_size_in_pte_reqs_luma)
drivers/gpu/drm/amd/display/dc/dml/dcn30/display_rq_dlg_calc_30.c
343
: (mode_lib->ip.dpte_buffer_size_in_pte_reqs_luma + mode_lib->ip.dpte_buffer_size_in_pte_reqs_chroma);
drivers/gpu/drm/amd/display/dc/dml/dcn30/display_rq_dlg_calc_30.c
915
unsigned int min_vblank = mode_lib->ip.min_vblank_lines;
drivers/gpu/drm/amd/display/dc/dml/dcn31/dcn31_fpu.c
587
context->bw_ctx.bw.dcn.compbuf_size_kb = context->bw_ctx.dml.ip.config_return_buffer_size_in_kbytes - total_det;
drivers/gpu/drm/amd/display/dc/dml/dcn31/display_mode_vba_31.c
4645
PatchDETBufferSizeInKByte(v->NumberOfActivePlanes, v->NoOfDPPThisState, v->ip.config_return_buffer_size_in_kbytes, v->DETBufferSizeInKByte);
drivers/gpu/drm/amd/display/dc/dml/dcn31/display_rq_dlg_calc_31.c
114
unsigned int detile_buf_size_in_bytes = mode_lib->ip.det_buffer_size_kbytes * 1024;
drivers/gpu/drm/amd/display/dc/dml/dcn31/display_rq_dlg_calc_31.c
180
unsigned int detile_buf_size_in_bytes = mode_lib->ip.det_buffer_size_kbytes * 1024;
drivers/gpu/drm/amd/display/dc/dml/dcn31/display_rq_dlg_calc_31.c
363
dual_plane_en ? (is_chroma ? mode_lib->ip.dpte_buffer_size_in_pte_reqs_chroma : mode_lib->ip.dpte_buffer_size_in_pte_reqs_luma) : (mode_lib->ip.dpte_buffer_size_in_pte_reqs_luma
drivers/gpu/drm/amd/display/dc/dml/dcn31/display_rq_dlg_calc_31.c
364
+ mode_lib->ip.dpte_buffer_size_in_pte_reqs_chroma);
drivers/gpu/drm/amd/display/dc/dml/dcn314/dcn314_fpu.c
392
context->bw_ctx.dml.ip.det_buffer_size_kbytes = DCN3_14_DEFAULT_DET_SIZE;
drivers/gpu/drm/amd/display/dc/dml/dcn314/dcn314_fpu.c
402
context->bw_ctx.dml.ip.det_buffer_size_kbytes = 192;
drivers/gpu/drm/amd/display/dc/dml/dcn314/dcn314_fpu.c
407
context->bw_ctx.dml.ip.det_buffer_size_kbytes = dc->debug.crb_alloc_policy * 64;
drivers/gpu/drm/amd/display/dc/dml/dcn314/dcn314_fpu.c
409
context->bw_ctx.dml.ip.det_buffer_size_kbytes = 192;
drivers/gpu/drm/amd/display/dc/dml/dcn314/dcn314_fpu.c
413
context->bw_ctx.dml.ip.odm_combine_4to1_supported = true;
drivers/gpu/drm/amd/display/dc/dml/dcn314/display_mode_vba_314.c
2569
v->ip.VBlankNomDefaultUS,
drivers/gpu/drm/amd/display/dc/dml/dcn314/display_mode_vba_314.c
5098
v->ip.VBlankNomDefaultUS,
drivers/gpu/drm/amd/display/dc/dml/dcn314/display_rq_dlg_calc_314.c
202
unsigned int detile_buf_size_in_bytes = mode_lib->ip.det_buffer_size_kbytes * 1024;
drivers/gpu/drm/amd/display/dc/dml/dcn314/display_rq_dlg_calc_314.c
268
unsigned int detile_buf_size_in_bytes = mode_lib->ip.det_buffer_size_kbytes * 1024;
drivers/gpu/drm/amd/display/dc/dml/dcn314/display_rq_dlg_calc_314.c
451
dual_plane_en ? (is_chroma ? mode_lib->ip.dpte_buffer_size_in_pte_reqs_chroma : mode_lib->ip.dpte_buffer_size_in_pte_reqs_luma) : (mode_lib->ip.dpte_buffer_size_in_pte_reqs_luma
drivers/gpu/drm/amd/display/dc/dml/dcn314/display_rq_dlg_calc_314.c
452
+ mode_lib->ip.dpte_buffer_size_in_pte_reqs_chroma);
drivers/gpu/drm/amd/display/dc/dml/dcn32/dcn32_fpu.c
1787
context->bw_ctx.bw.dcn.compbuf_size_kb = context->bw_ctx.dml.ip.config_return_buffer_size_in_kbytes;
drivers/gpu/drm/amd/display/dc/dml/dcn32/display_mode_vba_32.c
1967
mode_lib->vba.ip.pixel_chunk_size_kbytes,
drivers/gpu/drm/amd/display/dc/dml/dcn32/display_mode_vba_32.c
1968
mode_lib->vba.ip.rob_buffer_size_kbytes,
drivers/gpu/drm/amd/display/dc/dml/dcn32/display_mode_vba_32.c
233
mode_lib->vba.ip.pixel_chunk_size_kbytes,
drivers/gpu/drm/amd/display/dc/dml/dcn32/display_mode_vba_32.c
234
mode_lib->vba.ip.rob_buffer_size_kbytes,
drivers/gpu/drm/amd/display/dc/dml/dcn32/display_mode_vba_32.c
2529
mode_lib->vba.ip.dsc_delay_factor_wa);
drivers/gpu/drm/amd/display/dc/dml/dcn32/display_mode_vba_32.c
2567
mode_lib->vba.ip.pixel_chunk_size_kbytes,
drivers/gpu/drm/amd/display/dc/dml/dcn32/display_mode_vba_32.c
2568
mode_lib->vba.ip.rob_buffer_size_kbytes,
drivers/gpu/drm/amd/display/dc/dml/dcn32/display_mode_vba_32.c
3321
mode_lib->vba.ip.min_prefetch_in_strobe_us : 0,
drivers/gpu/drm/amd/display/dc/dml/dcn32/display_mode_vba_32.c
370
mode_lib->vba.PixelClockBackEnd[k], mode_lib->vba.ip.dsc_delay_factor_wa);
drivers/gpu/drm/amd/display/dc/dml/dcn32/display_mode_vba_32.c
815
mode_lib->vba.ip.min_prefetch_in_strobe_us : 0,
drivers/gpu/drm/amd/display/dc/dml/dcn35/dcn35_fpu.c
530
context->bw_ctx.dml.ip.det_buffer_size_kbytes = 384;/*per guide*/
drivers/gpu/drm/amd/display/dc/dml/dcn35/dcn35_fpu.c
544
context->bw_ctx.dml.ip.det_buffer_size_kbytes = 192;
drivers/gpu/drm/amd/display/dc/dml/dcn35/dcn35_fpu.c
550
context->bw_ctx.dml.ip.det_buffer_size_kbytes =
drivers/gpu/drm/amd/display/dc/dml/dcn35/dcn35_fpu.c
553
context->bw_ctx.dml.ip.det_buffer_size_kbytes = 192;
drivers/gpu/drm/amd/display/dc/dml/dcn351/dcn351_fpu.c
563
context->bw_ctx.dml.ip.det_buffer_size_kbytes = 384;/*per guide*/
drivers/gpu/drm/amd/display/dc/dml/dcn351/dcn351_fpu.c
577
context->bw_ctx.dml.ip.det_buffer_size_kbytes = 192;
drivers/gpu/drm/amd/display/dc/dml/dcn351/dcn351_fpu.c
583
context->bw_ctx.dml.ip.det_buffer_size_kbytes =
drivers/gpu/drm/amd/display/dc/dml/dcn351/dcn351_fpu.c
586
context->bw_ctx.dml.ip.det_buffer_size_kbytes = 192;
drivers/gpu/drm/amd/display/dc/dml/display_mode_lib.c
99
lib->ip = *ip_params;
drivers/gpu/drm/amd/display/dc/dml/display_mode_lib.h
86
struct _vcs_dpi_ip_params_st ip;
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
420
ip_params_st *ip = &mode_lib->vba.ip;
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
423
mode_lib->vba.UseMinimumRequiredDCFCLK = ip->use_min_dcfclk;
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
424
mode_lib->vba.ClampMinDCFCLK = ip->clamp_min_dcfclk;
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
425
mode_lib->vba.MaxNumDPP = ip->max_num_dpp;
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
426
mode_lib->vba.MaxNumOTG = ip->max_num_otg;
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
427
mode_lib->vba.MaxNumHDMIFRLOutputs = ip->max_num_hdmi_frl_outputs;
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
428
mode_lib->vba.MaxNumWriteback = ip->max_num_wb;
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
429
mode_lib->vba.CursorChunkSize = ip->cursor_chunk_size;
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
430
mode_lib->vba.CursorBufferSize = ip->cursor_buffer_size;
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
432
mode_lib->vba.MaxDCHUBToPSCLThroughput = ip->max_dchub_pscl_bw_pix_per_clk;
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
433
mode_lib->vba.MaxPSCLToLBThroughput = ip->max_pscl_lb_bw_pix_per_clk;
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
434
mode_lib->vba.ROBBufferSizeInKByte = ip->rob_buffer_size_kbytes;
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
435
mode_lib->vba.DETBufferSizeInKByte[0] = ip->det_buffer_size_kbytes;
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
436
mode_lib->vba.ConfigReturnBufferSizeInKByte = ip->config_return_buffer_size_in_kbytes;
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
437
mode_lib->vba.CompressedBufferSegmentSizeInkByte = ip->compressed_buffer_segment_size_in_kbytes;
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
438
mode_lib->vba.MetaFIFOSizeInKEntries = ip->meta_fifo_size_in_kentries;
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
439
mode_lib->vba.ZeroSizeBufferEntries = ip->zero_size_buffer_entries;
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
440
mode_lib->vba.COMPBUF_RESERVED_SPACE_64B = ip->compbuf_reserved_space_64b;
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
441
mode_lib->vba.COMPBUF_RESERVED_SPACE_ZS = ip->compbuf_reserved_space_zs;
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
442
mode_lib->vba.MaximumDSCBitsPerComponent = ip->maximum_dsc_bits_per_component;
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
443
mode_lib->vba.DSC422NativeSupport = ip->dsc422_native_support;
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
445
mode_lib->vba.nomDETInKByte = ip->det_buffer_size_kbytes;
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
446
mode_lib->vba.CompbufReservedSpace64B = ip->compbuf_reserved_space_64b;
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
447
mode_lib->vba.CompbufReservedSpaceZs = ip->compbuf_reserved_space_zs;
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
448
mode_lib->vba.CompressedBufferSegmentSizeInkByteFinal = ip->compressed_buffer_segment_size_in_kbytes;
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
449
mode_lib->vba.LineBufferSizeFinal = ip->line_buffer_size_bits;
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
450
mode_lib->vba.AlphaPixelChunkSizeInKByte = ip->alpha_pixel_chunk_size_kbytes; // not ysed
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
451
mode_lib->vba.MinPixelChunkSizeBytes = ip->min_pixel_chunk_size_bytes; // not used
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
452
mode_lib->vba.MaximumPixelsPerLinePerDSCUnit = ip->maximum_pixels_per_line_per_dsc_unit;
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
453
mode_lib->vba.MaxNumDP2p0Outputs = ip->max_num_dp2p0_outputs;
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
454
mode_lib->vba.MaxNumDP2p0Streams = ip->max_num_dp2p0_streams;
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
455
mode_lib->vba.DCCMetaBufferSizeBytes = ip->dcc_meta_buffer_size_bytes;
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
457
mode_lib->vba.PixelChunkSizeInKByte = ip->pixel_chunk_size_kbytes;
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
458
mode_lib->vba.MetaChunkSize = ip->meta_chunk_size_kbytes;
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
459
mode_lib->vba.MinMetaChunkSizeBytes = ip->min_meta_chunk_size_bytes;
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
460
mode_lib->vba.WritebackChunkSize = ip->writeback_chunk_size_kbytes;
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
461
mode_lib->vba.LineBufferSize = ip->line_buffer_size_bits;
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
462
mode_lib->vba.MaxLineBufferLines = ip->max_line_buffer_lines;
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
463
mode_lib->vba.PTEBufferSizeInRequestsLuma = ip->dpte_buffer_size_in_pte_reqs_luma;
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
464
mode_lib->vba.PTEBufferSizeInRequestsChroma = ip->dpte_buffer_size_in_pte_reqs_chroma;
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
465
mode_lib->vba.DPPOutputBufferPixels = ip->dpp_output_buffer_pixels;
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
466
mode_lib->vba.OPPOutputBufferLines = ip->opp_output_buffer_lines;
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
467
mode_lib->vba.MaxHSCLRatio = ip->max_hscl_ratio;
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
468
mode_lib->vba.MaxVSCLRatio = ip->max_vscl_ratio;
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
469
mode_lib->vba.WritebackInterfaceLumaBufferSize = ip->writeback_luma_buffer_size_kbytes * 1024;
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
470
mode_lib->vba.WritebackInterfaceChromaBufferSize = ip->writeback_chroma_buffer_size_kbytes * 1024;
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
472
mode_lib->vba.WritebackInterfaceBufferSize = ip->writeback_interface_buffer_size_kbytes;
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
473
mode_lib->vba.WritebackLineBufferSize = ip->writeback_line_buffer_buffer_size;
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
476
ip->writeback_chroma_line_buffer_width_pixels;
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
478
ip->writeback_line_buffer_luma_buffer_size;
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
480
ip->writeback_line_buffer_chroma_buffer_size;
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
481
mode_lib->vba.Writeback10bpc420Supported = ip->writeback_10bpc420_supported;
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
482
mode_lib->vba.WritebackMaxHSCLRatio = ip->writeback_max_hscl_ratio;
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
483
mode_lib->vba.WritebackMaxVSCLRatio = ip->writeback_max_vscl_ratio;
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
484
mode_lib->vba.WritebackMinHSCLRatio = ip->writeback_min_hscl_ratio;
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
485
mode_lib->vba.WritebackMinVSCLRatio = ip->writeback_min_vscl_ratio;
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
486
mode_lib->vba.WritebackMaxHSCLTaps = ip->writeback_max_hscl_taps;
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
487
mode_lib->vba.WritebackMaxVSCLTaps = ip->writeback_max_vscl_taps;
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
489
mode_lib->vba.GPUVMMaxPageTableLevels = ip->gpuvm_max_page_table_levels;
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
490
mode_lib->vba.HostVMMaxNonCachedPageTableLevels = ip->hostvm_max_page_table_levels;
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
491
mode_lib->vba.HostVMMaxPageTableLevels = ip->hostvm_max_page_table_levels;
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
492
mode_lib->vba.HostVMCachedPageTableLevels = ip->hostvm_cached_page_table_levels;
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
493
mode_lib->vba.MaxInterDCNTileRepeaters = ip->max_inter_dcn_tile_repeaters;
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
494
mode_lib->vba.NumberOfDSC = ip->num_dsc;
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
495
mode_lib->vba.ODMCapability = ip->odm_capable;
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
496
mode_lib->vba.DISPCLKRampingMargin = ip->dispclk_ramp_margin_percent;
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
498
mode_lib->vba.XFCSupported = ip->xfc_supported;
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
499
mode_lib->vba.XFCFillBWOverhead = ip->xfc_fill_bw_overhead_percent;
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
500
mode_lib->vba.XFCFillConstant = ip->xfc_fill_constant_bytes;
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
501
mode_lib->vba.DPPCLKDelaySubtotal = ip->dppclk_delay_subtotal;
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
502
mode_lib->vba.DPPCLKDelaySCL = ip->dppclk_delay_scl;
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
503
mode_lib->vba.DPPCLKDelaySCLLBOnly = ip->dppclk_delay_scl_lb_only;
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
504
mode_lib->vba.DPPCLKDelayCNVCFormater = ip->dppclk_delay_cnvc_formatter;
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
505
mode_lib->vba.DPPCLKDelayCNVCCursor = ip->dppclk_delay_cnvc_cursor;
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
506
mode_lib->vba.DISPCLKDelaySubtotal = ip->dispclk_delay_subtotal;
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
507
mode_lib->vba.DynamicMetadataVMEnabled = ip->dynamic_metadata_vm_enabled;
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
508
mode_lib->vba.ODMCombine4To1Supported = ip->odm_combine_4to1_supported;
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
509
mode_lib->vba.ProgressiveToInterlaceUnitInOPP = ip->ptoi_supported;
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
510
mode_lib->vba.PDEProcessingBufIn64KBReqs = ip->pde_proc_buffer_size_64k_reqs;
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
511
mode_lib->vba.PTEGroupSize = ip->pte_group_size_bytes;
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
512
mode_lib->vba.SupportGFX7CompatibleTilingIn32bppAnd64bpp = ip->gfx7_compat_tiling_supported;
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
518
ip_params_st *ip = &mode_lib->vba.ip;
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
577
mode_lib->vba.DETBufferSizeInKByte[mode_lib->vba.NumberOfActivePlanes] = ip->det_buffer_size_kbytes;
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
58
|| memcmp(&mode_lib->ip, &mode_lib->vba.ip, sizeof(mode_lib->vba.ip)) != 0
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
595
if (dst->interlaced && !ip->ptoi_supported) {
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
611
ip->dcc_supported : src->dcc && ip->dcc_supported;
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
64
mode_lib->vba.ip = mode_lib->ip;
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
653
ip->maximum_dsc_bits_per_component;
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
700
&& ip->xfc_supported;
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
709
if (ip->is_line_buffer_bpp_fixed)
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
711
ip->line_buffer_fixed_bpp;
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
905
mode_lib->vba.GPUVMEnable = mode_lib->vba.GPUVMEnable && !!ip->gpuvm_enable;
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
906
mode_lib->vba.HostVMEnable = mode_lib->vba.HostVMEnable && !!ip->hostvm_enable;
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
964
|| memcmp(&mode_lib->ip, &mode_lib->vba.ip, sizeof(mode_lib->vba.ip)) != 0
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
971
mode_lib->vba.ip = mode_lib->ip;
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.h
260
ip_params_st ip;
drivers/gpu/drm/amd/display/dc/dml/dml1_display_rq_dlg_calc.c
1014
unsigned int min_vblank = mode_lib->ip.min_vblank_lines;
drivers/gpu/drm/amd/display/dc/dml/dml1_display_rq_dlg_calc.c
1246
dppclk_delay_subtotal = mode_lib->ip.dppclk_delay_subtotal;
drivers/gpu/drm/amd/display/dc/dml/dml1_display_rq_dlg_calc.c
1247
dispclk_delay_subtotal = mode_lib->ip.dispclk_delay_subtotal;
drivers/gpu/drm/amd/display/dc/dml/dml1_display_rq_dlg_calc.c
236
unsigned int detile_buf_size_in_bytes = mode_lib->ip.det_buffer_size_kbytes * 1024;
drivers/gpu/drm/amd/display/dc/dml/dml1_display_rq_dlg_calc.c
288
unsigned int detile_buf_size_in_bytes = mode_lib->ip.det_buffer_size_kbytes * 1024;
drivers/gpu/drm/amd/display/dc/dml/dml1_display_rq_dlg_calc.c
480
dpte_buf_in_pte_reqs = mode_lib->ip.dpte_buffer_size_in_pte_reqs_luma;
drivers/gpu/drm/amd/display/dc/dml/dml1_display_rq_dlg_calc.c
796
dpte_buf_in_pte_reqs = mode_lib->ip.dpte_buffer_size_in_pte_reqs_luma;
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10066
mode_lib->ms.ip = mode_lib->ip;
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10260
dml_get_var_func(pixel_chunk_size_in_kbyte, dml_uint_t, mode_lib->ms.ip.pixel_chunk_size_kbytes);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10261
dml_get_var_func(alpha_pixel_chunk_size_in_kbyte, dml_uint_t, mode_lib->ms.ip.alpha_pixel_chunk_size_kbytes);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10262
dml_get_var_func(meta_chunk_size_in_kbyte, dml_uint_t, mode_lib->ms.ip.meta_chunk_size_kbytes);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10263
dml_get_var_func(min_pixel_chunk_size_in_byte, dml_uint_t, mode_lib->ms.ip.min_pixel_chunk_size_bytes);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10264
dml_get_var_func(min_meta_chunk_size_in_byte, dml_uint_t, mode_lib->ms.ip.min_meta_chunk_size_bytes);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
6225
CalculatePrefetchSchedule_params->DPPCLKDelaySubtotalPlusCNVCFormater = mode_lib->ms.ip.dppclk_delay_subtotal + mode_lib->ms.ip.dppclk_delay_cnvc_formatter;
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
6226
CalculatePrefetchSchedule_params->DPPCLKDelaySCL = mode_lib->ms.ip.dppclk_delay_scl;
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
6227
CalculatePrefetchSchedule_params->DPPCLKDelaySCLLBOnly = mode_lib->ms.ip.dppclk_delay_scl_lb_only;
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
6228
CalculatePrefetchSchedule_params->DPPCLKDelayCNVCCursor = mode_lib->ms.ip.dppclk_delay_cnvc_cursor;
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
6229
CalculatePrefetchSchedule_params->DISPCLKDelaySubtotal = mode_lib->ms.ip.dispclk_delay_subtotal;
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
6232
CalculatePrefetchSchedule_params->MaxInterDCNTileRepeaters = mode_lib->ms.ip.max_inter_dcn_tile_repeaters;
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
6239
CalculatePrefetchSchedule_params->DynamicMetadataVMEnabled = mode_lib->ms.ip.dynamic_metadata_vm_enabled;
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
6327
mode_lib->ms.ip.pixel_chunk_size_kbytes,
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
6329
mode_lib->ms.ip.meta_chunk_size_kbytes,
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
6386
myPipe->ProgressiveToInterlaceUnitInOPP = mode_lib->ms.ip.ptoi_supported;
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
6427
mode_lib->ms.ip.cursor_buffer_size,
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
6652
CalculateWatermarks_params->MaxLineBufferLines = mode_lib->ms.ip.max_line_buffer_lines;
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
6653
CalculateWatermarks_params->LineBufferSize = mode_lib->ms.ip.line_buffer_size_bits;
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
6654
CalculateWatermarks_params->WritebackInterfaceBufferSize = mode_lib->ms.ip.writeback_interface_buffer_size_kbytes;
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
6664
CalculateWatermarks_params->WritebackChunkSize = mode_lib->ms.ip.writeback_chunk_size_kbytes;
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
6722
CalculateVMRowAndSwath_params->PTEBufferSizeInRequestsLuma = mode_lib->ms.ip.dpte_buffer_size_in_pte_reqs_luma;
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
6723
CalculateVMRowAndSwath_params->PTEBufferSizeInRequestsChroma = mode_lib->ms.ip.dpte_buffer_size_in_pte_reqs_chroma;
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
6724
CalculateVMRowAndSwath_params->DCCMetaBufferSizeBytes = mode_lib->ms.ip.dcc_meta_buffer_size_bytes;
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
6798
mode_lib->ms.ip.config_return_buffer_size_in_kbytes,
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
6799
mode_lib->ms.ip.config_return_buffer_segment_size_in_kbytes,
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
6800
mode_lib->ms.ip.rob_buffer_size_kbytes,
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
6801
mode_lib->ms.ip.max_num_dpp,
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
6810
PixelClockAdjustmentForProgressiveToInterlaceUnit(&mode_lib->ms.cache_display_cfg, mode_lib->ms.ip.ptoi_supported);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
6834
|| mode_lib->ms.cache_display_cfg.plane.HRatio[k] > mode_lib->ms.ip.max_hscl_ratio
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
6835
|| mode_lib->ms.cache_display_cfg.plane.VRatio[k] > mode_lib->ms.ip.max_vscl_ratio
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
6846
mode_lib->ms.cache_display_cfg.plane.HRatioChroma[k] > mode_lib->ms.ip.max_hscl_ratio ||
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
6847
mode_lib->ms.cache_display_cfg.plane.VRatioChroma[k] > mode_lib->ms.ip.max_vscl_ratio ||
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
6919
(mode_lib->ms.WriteBandwidth[k] > mode_lib->ms.ip.writeback_interface_buffer_size_kbytes * 1024 / mode_lib->ms.state.writeback_latency_us)) {
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
6933
if (s->TotalNumberOfActiveWriteback > (dml_uint_t) mode_lib->ms.ip.max_num_wb) {
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
6941
if (mode_lib->ms.cache_display_cfg.writeback.WritebackHRatio[k] > mode_lib->ms.ip.writeback_max_hscl_ratio
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
6942
|| mode_lib->ms.cache_display_cfg.writeback.WritebackVRatio[k] > mode_lib->ms.ip.writeback_max_vscl_ratio
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
6943
|| mode_lib->ms.cache_display_cfg.writeback.WritebackHRatio[k] < mode_lib->ms.ip.writeback_min_hscl_ratio
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
6944
|| mode_lib->ms.cache_display_cfg.writeback.WritebackVRatio[k] < mode_lib->ms.ip.writeback_min_vscl_ratio
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
6945
|| mode_lib->ms.cache_display_cfg.writeback.WritebackHTaps[k] > (dml_uint_t) mode_lib->ms.ip.writeback_max_hscl_taps
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
6946
|| mode_lib->ms.cache_display_cfg.writeback.WritebackVTaps[k] > (dml_uint_t) mode_lib->ms.ip.writeback_max_vscl_taps
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
6952
if (2.0 * mode_lib->ms.cache_display_cfg.writeback.WritebackDestinationWidth[k] * (mode_lib->ms.cache_display_cfg.writeback.WritebackVTaps[k] - 1) * 57 > mode_lib->ms.ip.writeback_line_buffer_buffer_size) {
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
6964
mode_lib->ms.ip.max_dchub_pscl_bw_pix_per_clk,
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
6965
mode_lib->ms.ip.max_pscl_lb_bw_pix_per_clk,
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
6998
mode_lib->ms.MaximumSwathWidthInLineBufferLuma = mode_lib->ms.ip.line_buffer_size_bits * dml_max(mode_lib->ms.cache_display_cfg.plane.HRatio[k], 1.0) / mode_lib->ms.cache_display_cfg.plane.LBBitPerPixel[k] /
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
7004
mode_lib->ms.ip.line_buffer_size_bits
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
7040
CalculateSwathAndDETConfiguration_params->ConfigReturnBufferSizeInKByte = mode_lib->ms.ip.config_return_buffer_size_in_kbytes;
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
7041
CalculateSwathAndDETConfiguration_params->ROBBufferSizeInKByte = mode_lib->ms.ip.rob_buffer_size_kbytes;
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
7044
CalculateSwathAndDETConfiguration_params->PixelChunkSizeInKByte = mode_lib->ms.ip.pixel_chunk_size_kbytes;
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
7049
CalculateSwathAndDETConfiguration_params->ConfigReturnBufferSegmentSizeInkByte = mode_lib->ms.ip.config_return_buffer_segment_size_in_kbytes;
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
7050
CalculateSwathAndDETConfiguration_params->CompressedBufferSegmentSizeInkByteFinal = mode_lib->ms.ip.compressed_buffer_segment_size_in_kbytes;
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
7119
mode_lib->ms.ip.maximum_pixels_per_line_per_dsc_unit,
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
7128
mode_lib->ms.ip.max_num_dpp,
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
7131
mode_lib->ms.ip.dispclk_ramp_margin_percent,
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
7142
mode_lib->ms.ip.maximum_pixels_per_line_per_dsc_unit,
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
7151
mode_lib->ms.ip.max_num_dpp,
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
7154
mode_lib->ms.ip.dispclk_ramp_margin_percent,
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
7224
} else if (mode_lib->ms.TotalNumberOfActiveDPP[j] < (dml_uint_t) mode_lib->ms.ip.max_num_dpp) {
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
7252
while (!(mode_lib->ms.TotalNumberOfActiveDPP[j] >= (dml_uint_t) mode_lib->ms.ip.max_num_dpp || mode_lib->ms.TotalNumberOfSingleDPPSurfaces[j] == 0)) {
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
7285
mode_lib->ms.ip.writeback_line_buffer_buffer_size,
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
7314
if (mode_lib->ms.TotalNumberOfActiveDPP[j] > (dml_uint_t) mode_lib->ms.ip.max_num_dpp) {
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
7339
mode_lib->ms.support.NumberOfOTGSupport = (s->TotalNumberOfActiveOTG <= (dml_uint_t) mode_lib->ms.ip.max_num_otg);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
7340
mode_lib->ms.support.NumberOfHDMIFRLSupport = (s->TotalNumberOfActiveHDMIFRL <= (dml_uint_t) mode_lib->ms.ip.max_num_hdmi_frl_outputs);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
7341
mode_lib->ms.support.NumberOfDP2p0Support = (s->TotalNumberOfActiveDP2p0 <= (dml_uint_t) mode_lib->ms.ip.max_num_dp2p0_streams && s->TotalNumberOfActiveDP2p0Outputs <= (dml_uint_t) mode_lib->ms.ip.max_num_dp2p0_outputs);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
7350
|| mode_lib->ms.cache_display_cfg.output.DSCInputBitPerComponent[k] > (dml_uint_t) mode_lib->ms.ip.maximum_dsc_bits_per_component
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
7392
if (mode_lib->ms.cache_display_cfg.output.OutputFormat[k] == dml_420 && mode_lib->ms.cache_display_cfg.timing.Interlace[k] == 1 && mode_lib->ms.ip.ptoi_supported == true)
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
7397
if ((mode_lib->ms.cache_display_cfg.output.DSCEnable[k] == dml_dsc_enable || mode_lib->ms.cache_display_cfg.output.DSCEnable[k] == dml_dsc_enable_if_necessary) && mode_lib->ms.cache_display_cfg.output.OutputFormat[k] == dml_n422 && !mode_lib->ms.ip.dsc422_native_support)
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
7519
if (mode_lib->ms.cache_display_cfg.timing.HActive[k] > 4 * (dml_uint_t) mode_lib->ms.ip.maximum_pixels_per_line_per_dsc_unit)
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
7525
if (mode_lib->ms.cache_display_cfg.timing.HActive[k] > 2 * (dml_uint_t) mode_lib->ms.ip.maximum_pixels_per_line_per_dsc_unit)
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
7531
if (mode_lib->ms.cache_display_cfg.timing.HActive[k] > (dml_uint_t) mode_lib->ms.ip.maximum_pixels_per_line_per_dsc_unit)
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
7539
if (s->TotalDSCUnitsRequired > (dml_uint_t) mode_lib->ms.ip.num_dsc) {
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
7579
CalculateSwathAndDETConfiguration_params->ConfigReturnBufferSizeInKByte = mode_lib->ms.ip.config_return_buffer_size_in_kbytes;
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
7580
CalculateSwathAndDETConfiguration_params->ROBBufferSizeInKByte = mode_lib->ms.ip.rob_buffer_size_kbytes;
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
7583
CalculateSwathAndDETConfiguration_params->PixelChunkSizeInKByte = mode_lib->ms.ip.pixel_chunk_size_kbytes;
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
7588
CalculateSwathAndDETConfiguration_params->ConfigReturnBufferSegmentSizeInkByte = mode_lib->ms.ip.config_return_buffer_segment_size_in_kbytes;
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
7589
CalculateSwathAndDETConfiguration_params->CompressedBufferSegmentSizeInkByteFinal = mode_lib->ms.ip.compressed_buffer_segment_size_in_kbytes;
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
7737
s->SurfParameters[k].ProgressiveToInterlaceUnitInOPP = mode_lib->ms.ip.ptoi_supported;
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
7812
mode_lib->ms.ip.cursor_buffer_size,
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
7895
mode_lib->ms.ip.ptoi_supported,
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
7896
mode_lib->ms.ip.vblank_nom_default_us,
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
7966
UseMinimumDCFCLK_params->MaxInterDCNTileRepeaters = mode_lib->ms.ip.max_inter_dcn_tile_repeaters;
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
7974
UseMinimumDCFCLK_params->PixelChunkSizeInKByte = mode_lib->ms.ip.pixel_chunk_size_kbytes;
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
7975
UseMinimumDCFCLK_params->MetaChunkSize = mode_lib->ms.ip.meta_chunk_size_kbytes;
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
7982
UseMinimumDCFCLK_params->DynamicMetadataVMEnabled = mode_lib->ms.ip.dynamic_metadata_vm_enabled;
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
7984
UseMinimumDCFCLK_params->ProgressiveToInterlaceUnitInOPP = mode_lib->ms.ip.ptoi_supported;
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
8034
if ((mode_lib->ms.ip.rob_buffer_size_kbytes - mode_lib->ms.ip.pixel_chunk_size_kbytes) * 1024 / mode_lib->ms.ReturnBWPerState[j] >
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
8073
if (mode_lib->ms.cache_display_cfg.plane.CursorBPP[k] == 64 && mode_lib->ms.ip.cursor_64bpp_support == false) {
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
8274
mode_lib->ms.ip.config_return_buffer_size_in_kbytes,
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
8275
mode_lib->ms.ip.config_return_buffer_segment_size_in_kbytes,
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
8276
mode_lib->ms.ip.rob_buffer_size_kbytes,
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
8277
mode_lib->ms.ip.max_num_dpp,
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
8286
PixelClockAdjustmentForProgressiveToInterlaceUnit(&mode_lib->ms.cache_display_cfg, mode_lib->ms.ip.ptoi_supported);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
8363
mode_lib->ms.ip.writeback_line_buffer_buffer_size,
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
8376
mode_lib->ms.ip.dispclk_ramp_margin_percent,
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
8397
mode_lib->ms.ip.max_dchub_pscl_bw_pix_per_clk,
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
8398
mode_lib->ms.ip.max_pscl_lb_bw_pix_per_clk,
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
8502
CalculateSwathAndDETConfiguration_params->ConfigReturnBufferSizeInKByte = mode_lib->ms.ip.config_return_buffer_size_in_kbytes;
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
8503
CalculateSwathAndDETConfiguration_params->ROBBufferSizeInKByte = mode_lib->ms.ip.rob_buffer_size_kbytes;
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
8506
CalculateSwathAndDETConfiguration_params->PixelChunkSizeInKByte = mode_lib->ms.ip.pixel_chunk_size_kbytes;
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
8511
CalculateSwathAndDETConfiguration_params->ConfigReturnBufferSegmentSizeInkByte = mode_lib->ms.ip.config_return_buffer_segment_size_in_kbytes;
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
8512
CalculateSwathAndDETConfiguration_params->CompressedBufferSegmentSizeInkByteFinal = mode_lib->ms.ip.compressed_buffer_segment_size_in_kbytes;
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
8687
s->SurfaceParameters[k].ProgressiveToInterlaceUnitInOPP = mode_lib->ms.ip.ptoi_supported;
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
8709
CalculateVMRowAndSwath_params->PTEBufferSizeInRequestsLuma = mode_lib->ms.ip.dpte_buffer_size_in_pte_reqs_luma;
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
8710
CalculateVMRowAndSwath_params->PTEBufferSizeInRequestsChroma = mode_lib->ms.ip.dpte_buffer_size_in_pte_reqs_chroma;
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
8711
CalculateVMRowAndSwath_params->DCCMetaBufferSizeBytes = mode_lib->ms.ip.dcc_meta_buffer_size_bytes;
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
8818
mode_lib->ms.ip.pixel_chunk_size_kbytes,
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
8820
mode_lib->ms.ip.meta_chunk_size_kbytes,
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
8891
mode_lib->ms.ip.cursor_buffer_size,
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
8916
mode_lib->ms.ip.ptoi_supported,
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
8917
mode_lib->ms.ip.vblank_nom_default_us,
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
8992
myPipe->ProgressiveToInterlaceUnitInOPP = mode_lib->ms.ip.ptoi_supported;
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
9004
CalculatePrefetchSchedule_params->DPPCLKDelaySubtotalPlusCNVCFormater = mode_lib->ms.ip.dppclk_delay_subtotal + mode_lib->ms.ip.dppclk_delay_cnvc_formatter;
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
9005
CalculatePrefetchSchedule_params->DPPCLKDelaySCL = mode_lib->ms.ip.dppclk_delay_scl;
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
9006
CalculatePrefetchSchedule_params->DPPCLKDelaySCLLBOnly = mode_lib->ms.ip.dppclk_delay_scl_lb_only;
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
9007
CalculatePrefetchSchedule_params->DPPCLKDelayCNVCCursor = mode_lib->ms.ip.dppclk_delay_cnvc_cursor;
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
9008
CalculatePrefetchSchedule_params->DISPCLKDelaySubtotal = mode_lib->ms.ip.dispclk_delay_subtotal;
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
9011
CalculatePrefetchSchedule_params->MaxInterDCNTileRepeaters = mode_lib->ms.ip.max_inter_dcn_tile_repeaters;
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
9020
CalculatePrefetchSchedule_params->DynamicMetadataVMEnabled = mode_lib->ms.ip.dynamic_metadata_vm_enabled;
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
9079
mode_lib->ms.ip.cursor_buffer_size,
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
9425
CalculateWatermarks_params->MaxLineBufferLines = mode_lib->ms.ip.max_line_buffer_lines;
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
9426
CalculateWatermarks_params->LineBufferSize = mode_lib->ms.ip.line_buffer_size_bits;
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
9427
CalculateWatermarks_params->WritebackInterfaceBufferSize = mode_lib->ms.ip.writeback_interface_buffer_size_kbytes;
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
9437
CalculateWatermarks_params->WritebackChunkSize = mode_lib->ms.ip.writeback_chunk_size_kbytes;
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
9546
mode_lib->ms.ip.meta_chunk_size_kbytes,
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
9547
mode_lib->ms.ip.min_meta_chunk_size_bytes,
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
9691
if (mode_lib->ms.cache_display_cfg.plane.DynamicMetadataEnable[k] && mode_lib->ms.ip.dynamic_metadata_vm_enabled) {
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
9695
isInterlaceTiming = (mode_lib->ms.cache_display_cfg.timing.Interlace[k] && !mode_lib->ms.ip.ptoi_supported);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
9779
CalculateStutterEfficiency_params->MetaFIFOSizeInKEntries = mode_lib->ms.ip.meta_fifo_size_in_kentries;
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
9780
CalculateStutterEfficiency_params->ZeroSizeBufferEntries = mode_lib->ms.ip.zero_size_buffer_entries;
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
9781
CalculateStutterEfficiency_params->PixelChunkSizeInKByte = mode_lib->ms.ip.pixel_chunk_size_kbytes;
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
9783
CalculateStutterEfficiency_params->ROBBufferSizeInKByte = mode_lib->ms.ip.rob_buffer_size_kbytes;
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
9795
CalculateStutterEfficiency_params->ProgressiveToInterlaceUnitInOPP = mode_lib->ms.ip.ptoi_supported;
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
9851
mode_lib->ms.ip.meta_fifo_size_in_kentries,
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
9852
mode_lib->ms.ip.zero_size_buffer_entries,
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
9853
mode_lib->ms.ip.pixel_chunk_size_kbytes,
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
9855
mode_lib->ms.ip.rob_buffer_size_kbytes,
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
9867
mode_lib->ms.ip.ptoi_supported,
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
9980
PTEBufferSizeInRequests = is_plane1 ? mode_lib->ip.dpte_buffer_size_in_pte_reqs_chroma : mode_lib->ip.dpte_buffer_size_in_pte_reqs_luma;
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
9989
dml_print("DML_DLG: %s: dpte_buffer_size_in_pte_reqs_luma = %u\n", __func__, mode_lib->ip.dpte_buffer_size_in_pte_reqs_luma);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
9990
dml_print("DML_DLG: %s: dpte_buffer_size_in_pte_reqs_chroma = %u\n", __func__, mode_lib->ip.dpte_buffer_size_in_pte_reqs_chroma);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core_structs.h
1904
struct ip_params_st ip;
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core_structs.h
811
struct ip_params_st ip;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4.c
135
memcpy(&core->clean_me_up.mode_lib.ip, in_out->explicit_ip_bb, in_out->explicit_ip_bb_size);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4.c
141
core->clean_me_up.mode_lib.ip.subvp_pstate_allow_width_us = core_dcn4_ip_caps_base.subvp_pstate_allow_width_us;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4.c
142
core->clean_me_up.mode_lib.ip.subvp_fw_processing_delay_us = core_dcn4_ip_caps_base.subvp_pstate_allow_width_us;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4.c
143
core->clean_me_up.mode_lib.ip.subvp_swath_height_margin_lines = core_dcn4_ip_caps_base.subvp_swath_height_margin_lines;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4.c
145
memcpy(&core->clean_me_up.mode_lib.ip, &core_dcn4_ip_caps_base, sizeof(struct dml2_core_ip_params));
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4.c
146
patch_ip_params_with_ip_caps(&core->clean_me_up.mode_lib.ip, in_out->ip_caps);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4.c
147
core->clean_me_up.mode_lib.ip.imall_supported = false;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10515
mode_lib->ip.config_return_buffer_size_in_kbytes,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10516
mode_lib->ip.config_return_buffer_segment_size_in_kbytes,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10517
mode_lib->ip.rob_buffer_size_kbytes,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10518
mode_lib->ip.max_num_dpp,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10521
mode_lib->ip.dcn_mrq_present,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10529
PixelClockAdjustmentForProgressiveToInterlaceUnit(display_cfg, mode_lib->ip.ptoi_supported, s->PixelClockBackEnd);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10537
mode_lib->ip.max_dchub_pscl_bw_pix_per_clk,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10538
mode_lib->ip.max_pscl_lb_bw_pix_per_clk,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10613
CalculateSwathAndDETConfiguration_params->ConfigReturnBufferSizeInKByte = mode_lib->ip.config_return_buffer_size_in_kbytes;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10616
CalculateSwathAndDETConfiguration_params->rob_buffer_size_kbytes = mode_lib->ip.rob_buffer_size_kbytes;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10617
CalculateSwathAndDETConfiguration_params->pixel_chunk_size_kbytes = mode_lib->ip.pixel_chunk_size_kbytes;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10618
CalculateSwathAndDETConfiguration_params->rob_buffer_size_kbytes = mode_lib->ip.rob_buffer_size_kbytes;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10619
CalculateSwathAndDETConfiguration_params->pixel_chunk_size_kbytes = mode_lib->ip.pixel_chunk_size_kbytes;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10623
CalculateSwathAndDETConfiguration_params->ConfigReturnBufferSegmentSizeInkByte = mode_lib->ip.config_return_buffer_segment_size_in_kbytes;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10624
CalculateSwathAndDETConfiguration_params->CompressedBufferSegmentSizeInkByte = mode_lib->ip.compressed_buffer_segment_size_in_kbytes;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10641
CalculateSwathAndDETConfiguration_params->mrq_present = mode_lib->ip.dcn_mrq_present;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10673
mode_lib->ip.maximum_dsc_bits_per_component,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10730
s->SurfaceParameters[k].ProgressiveToInterlaceUnitInOPP = mode_lib->ip.ptoi_supported;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10753
CalculateVMRowAndSwath_params->PTEBufferSizeInRequestsLuma = mode_lib->ip.dpte_buffer_size_in_pte_reqs_luma;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10754
CalculateVMRowAndSwath_params->PTEBufferSizeInRequestsChroma = mode_lib->ip.dpte_buffer_size_in_pte_reqs_chroma;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10759
CalculateVMRowAndSwath_params->DCCMetaBufferSizeBytes = mode_lib->ip.dcc_meta_buffer_size_bytes;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10760
CalculateVMRowAndSwath_params->mrq_present = mode_lib->ip.dcn_mrq_present;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10819
if (mode_lib->soc.mall_allocated_for_dcn_mbytes == 0 || mode_lib->ip.dcn_mrq_present) {
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10842
calculate_mcache_setting_params->imall_enable = mode_lib->ip.imall_supported && display_cfg->plane_descriptors[k].overrides.legacy_svp_config == dml2_svp_mode_override_imall;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10921
mode_lib->ip.remote_iommu_outstanding_translations,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10939
calculate_tdlut_setting_params->cursor_buffer_size = mode_lib->ip.cursor_buffer_size;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10961
mode_lib->ip.rob_buffer_size_kbytes,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10966
mode_lib->ip.pixel_chunk_size_kbytes,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10980
mode_lib->ip.meta_chunk_size_kbytes,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10981
mode_lib->ip.dchub_arb_to_ret_delay,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10983
mode_lib->ip.hostvm_mode,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11011
calculate_bytes_to_fetch_required_to_hide_latency_params->mrq_present = mode_lib->ip.dcn_mrq_present;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11094
s->lb_source_lines_l[k] = get_num_lb_source_lines(mode_lib->ip.max_line_buffer_lines, mode_lib->ip.line_buffer_size_bits,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11101
s->lb_source_lines_c[k] = get_num_lb_source_lines(mode_lib->ip.max_line_buffer_lines, mode_lib->ip.line_buffer_size_bits,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11120
mode_lib->ip.cursor_buffer_size,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11158
mode_lib->ip.ptoi_supported,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11159
mode_lib->ip.vblank_nom_default_us,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11180
CheckGlobalPrefetchAdmissibility_params->chunk_bytes_l = mode_lib->ip.pixel_chunk_size_kbytes * 1024;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11181
CheckGlobalPrefetchAdmissibility_params->chunk_bytes_c = mode_lib->ip.pixel_chunk_size_kbytes * 1024;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11186
CheckGlobalPrefetchAdmissibility_params->rob_buffer_size_kbytes = mode_lib->ip.rob_buffer_size_kbytes;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11249
myPipe->ProgressiveToInterlaceUnitInOPP = mode_lib->ip.ptoi_supported;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11258
CalculatePrefetchSchedule_params->DPPCLKDelaySubtotalPlusCNVCFormater = mode_lib->ip.dppclk_delay_subtotal + mode_lib->ip.dppclk_delay_cnvc_formatter;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11259
CalculatePrefetchSchedule_params->DPPCLKDelaySCL = mode_lib->ip.dppclk_delay_scl;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11260
CalculatePrefetchSchedule_params->DPPCLKDelaySCLLBOnly = mode_lib->ip.dppclk_delay_scl_lb_only;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11261
CalculatePrefetchSchedule_params->DPPCLKDelayCNVCCursor = mode_lib->ip.dppclk_delay_cnvc_cursor;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11262
CalculatePrefetchSchedule_params->DISPCLKDelaySubtotal = mode_lib->ip.dispclk_delay_subtotal;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11265
CalculatePrefetchSchedule_params->MaxInterDCNTileRepeaters = mode_lib->ip.max_inter_dcn_tile_repeaters;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11269
CalculatePrefetchSchedule_params->DynamicMetadataVMEnabled = mode_lib->ip.dynamic_metadata_vm_enabled;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11299
CalculatePrefetchSchedule_params->mrq_present = mode_lib->ip.dcn_mrq_present;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11555
mode_lib->ip.max_flip_time_us,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11556
mode_lib->ip.max_flip_time_lines,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11561
mode_lib->ip.dcn_mrq_present && display_cfg->plane_descriptors[k].surface.dcc.enable,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11714
CalculateWatermarks_params->MaxLineBufferLines = mode_lib->ip.max_line_buffer_lines;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11715
CalculateWatermarks_params->LineBufferSize = mode_lib->ip.line_buffer_size_bits;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11716
CalculateWatermarks_params->WritebackInterfaceBufferSize = mode_lib->ip.writeback_interface_buffer_size_kbytes;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11722
CalculateWatermarks_params->WritebackChunkSize = mode_lib->ip.writeback_chunk_size_kbytes;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11820
CalculateMetaAndPTETimes_params->mrq_present = mode_lib->ip.dcn_mrq_present;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11822
CalculateMetaAndPTETimes_params->MetaChunkSize = mode_lib->ip.meta_chunk_size_kbytes;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11823
CalculateMetaAndPTETimes_params->MinMetaChunkSizeBytes = mode_lib->ip.min_meta_chunk_size_bytes;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11867
mode_lib->ip.dcn_mrq_present,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11896
if (display_cfg->plane_descriptors[k].dynamic_meta_data.enable && mode_lib->ip.dynamic_metadata_vm_enabled) {
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11900
isInterlaceTiming = (display_cfg->stream_descriptors[display_cfg->plane_descriptors[k].stream_index].timing.interlaced && !mode_lib->ip.ptoi_supported);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11969
CalculateStutterEfficiency_params->MetaFIFOSizeInKEntries = mode_lib->ip.meta_fifo_size_in_kentries;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11970
CalculateStutterEfficiency_params->ZeroSizeBufferEntries = mode_lib->ip.zero_size_buffer_entries;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11971
CalculateStutterEfficiency_params->PixelChunkSizeInKByte = mode_lib->ip.pixel_chunk_size_kbytes;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11973
CalculateStutterEfficiency_params->ROBBufferSizeInKByte = mode_lib->ip.rob_buffer_size_kbytes;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11978
CalculateStutterEfficiency_params->CompbufReservedSpaceZs = mode_lib->ip.compbuf_reserved_space_zs;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11984
CalculateStutterEfficiency_params->ProgressiveToInterlaceUnitInOPP = mode_lib->ip.ptoi_supported;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
12003
CalculateStutterEfficiency_params->rob_alloc_compressed = mode_lib->ip.dcn_mrq_present;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
12041
mode_lib->mp.dcfclk_deep_sleep_hysteresis = (unsigned int)math_max2(32, (double)mode_lib->ip.pixel_chunk_size_kbytes * 1024 * 3 / 4 / 64 - min_return_latency_in_DCFCLK_cycles);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
12119
unsigned int PTEBufferSizeInRequests = is_plane1 ? mode_lib->ip.dpte_buffer_size_in_pte_reqs_chroma : mode_lib->ip.dpte_buffer_size_in_pte_reqs_luma;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
12128
DML_LOG_VERBOSE("DML: %s: dpte_buffer_size_in_pte_reqs_luma = %u\n", __func__, mode_lib->ip.dpte_buffer_size_in_pte_reqs_luma);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
12129
DML_LOG_VERBOSE("DML: %s: dpte_buffer_size_in_pte_reqs_chroma = %u\n", __func__, mode_lib->ip.dpte_buffer_size_in_pte_reqs_chroma);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
12281
pixel_chunk_bytes = (unsigned int)(mode_lib->ip.pixel_chunk_size_kbytes * 1024);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
12282
min_pixel_chunk_bytes = (unsigned int)(mode_lib->ip.min_pixel_chunk_size_bytes);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
12296
p1_pixel_chunk_bytes = (unsigned int)(mode_lib->ip.alpha_pixel_chunk_size_kbytes * 1024);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
12717
arb_param->sdpif_request_rate_limit = (3 * mode_lib->ip.words_per_channel * mode_lib->soc.clk_table.dram_config.channel_count) / 4;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
12723
arb_param->compbuf_size = mode_lib->mp.CompressedBufferSizeInkByte / mode_lib->ip.compressed_buffer_segment_size_in_kbytes;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
12754
out->det_size = dml_get_det_buffer_size_kbytes(mode_lib, pipe_index) / mode_lib->ip.config_return_buffer_segment_size_in_kbytes;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
12978
phantom_processing_delay_pix = (double)((mode_lib->ip.subvp_fw_processing_delay_us + mode_lib->ip.subvp_pstate_allow_width_us) *
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
12990
phantom_v_active_lines = phantom_processing_delay_lines + phantom_min_v_active_lines + mode_lib->ip.subvp_swath_height_margin_lines;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
13007
DML_LOG_VERBOSE("DML::%s: subvp_fw_processing_delay_us = %d\n", __func__, mode_lib->ip.subvp_fw_processing_delay_us);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
13008
DML_LOG_VERBOSE("DML::%s: subvp_pstate_allow_width_us = %d\n", __func__, mode_lib->ip.subvp_pstate_allow_width_us);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
13009
DML_LOG_VERBOSE("DML::%s: subvp_swath_height_margin_lines = %d\n", __func__, mode_lib->ip.subvp_swath_height_margin_lines);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
13350
if (((mode_lib->ip.rob_buffer_size_kbytes - mode_lib->ip.pixel_chunk_size_kbytes) * 1024
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
7329
mode_lib->ip.remote_iommu_outstanding_translations,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
7344
calculate_tdlut_setting_params->cursor_buffer_size = mode_lib->ip.cursor_buffer_size;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
7371
mode_lib->ip.rob_buffer_size_kbytes,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
7376
mode_lib->ip.pixel_chunk_size_kbytes,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
7390
mode_lib->ip.meta_chunk_size_kbytes,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
7391
mode_lib->ip.dchub_arb_to_ret_delay,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
7393
mode_lib->ip.hostvm_mode,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
7412
s->lb_source_lines_l[k] = get_num_lb_source_lines(mode_lib->ip.max_line_buffer_lines, mode_lib->ip.line_buffer_size_bits,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
7419
s->lb_source_lines_c[k] = get_num_lb_source_lines(mode_lib->ip.max_line_buffer_lines, mode_lib->ip.line_buffer_size_bits,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
7461
myPipe->ProgressiveToInterlaceUnitInOPP = mode_lib->ip.ptoi_supported;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
7471
CalculatePrefetchSchedule_params->DPPCLKDelaySubtotalPlusCNVCFormater = mode_lib->ip.dppclk_delay_subtotal + mode_lib->ip.dppclk_delay_cnvc_formatter;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
7472
CalculatePrefetchSchedule_params->DPPCLKDelaySCL = mode_lib->ip.dppclk_delay_scl;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
7473
CalculatePrefetchSchedule_params->DPPCLKDelaySCLLBOnly = mode_lib->ip.dppclk_delay_scl_lb_only;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
7474
CalculatePrefetchSchedule_params->DPPCLKDelayCNVCCursor = mode_lib->ip.dppclk_delay_cnvc_cursor;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
7475
CalculatePrefetchSchedule_params->DISPCLKDelaySubtotal = mode_lib->ip.dispclk_delay_subtotal;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
7478
CalculatePrefetchSchedule_params->MaxInterDCNTileRepeaters = mode_lib->ip.max_inter_dcn_tile_repeaters;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
7482
CalculatePrefetchSchedule_params->DynamicMetadataVMEnabled = mode_lib->ip.dynamic_metadata_vm_enabled;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
7512
CalculatePrefetchSchedule_params->mrq_present = mode_lib->ip.dcn_mrq_present;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
7722
CheckGlobalPrefetchAdmissibility_params->chunk_bytes_l = mode_lib->ip.pixel_chunk_size_kbytes * 1024;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
7723
CheckGlobalPrefetchAdmissibility_params->chunk_bytes_c = mode_lib->ip.pixel_chunk_size_kbytes * 1024;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
7728
CheckGlobalPrefetchAdmissibility_params->rob_buffer_size_kbytes = mode_lib->ip.rob_buffer_size_kbytes;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
7804
mode_lib->ip.max_flip_time_us,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
7805
mode_lib->ip.max_flip_time_lines,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
7810
mode_lib->ip.dcn_mrq_present && display_cfg->plane_descriptors[k].surface.dcc.enable,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
7899
CalculateWatermarks_params->MaxLineBufferLines = mode_lib->ip.max_line_buffer_lines;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
7900
CalculateWatermarks_params->LineBufferSize = mode_lib->ip.line_buffer_size_bits;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
7901
CalculateWatermarks_params->WritebackInterfaceBufferSize = mode_lib->ip.writeback_interface_buffer_size_kbytes;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
7907
CalculateWatermarks_params->WritebackChunkSize = mode_lib->ip.writeback_chunk_size_kbytes;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
8001
DML_LOG_VERBOSE("DML::%s: ip.compressed_buffer_segment_size_in_kbytes = %u\n", __func__, mode_lib->ip.compressed_buffer_segment_size_in_kbytes);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
8002
DML_LOG_VERBOSE("DML::%s: ip.dcn_mrq_present = %u\n", __func__, mode_lib->ip.dcn_mrq_present);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
8009
mode_lib->ip.config_return_buffer_size_in_kbytes,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
8010
mode_lib->ip.config_return_buffer_segment_size_in_kbytes,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
8011
mode_lib->ip.rob_buffer_size_kbytes,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
8012
mode_lib->ip.max_num_dpp,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
8015
mode_lib->ip.dcn_mrq_present,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
8022
PixelClockAdjustmentForProgressiveToInterlaceUnit(display_cfg, mode_lib->ip.ptoi_supported, s->PixelClockBackEnd);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
8040
|| display_cfg->plane_descriptors[k].composition.scaler_info.plane0.h_ratio > mode_lib->ip.max_hscl_ratio
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
8041
|| display_cfg->plane_descriptors[k].composition.scaler_info.plane0.v_ratio > mode_lib->ip.max_vscl_ratio
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
8048
display_cfg->plane_descriptors[k].composition.scaler_info.plane1.h_ratio > mode_lib->ip.max_hscl_ratio ||
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
8049
display_cfg->plane_descriptors[k].composition.scaler_info.plane1.v_ratio > mode_lib->ip.max_vscl_ratio ||
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
8137
(mode_lib->ms.WriteBandwidth[k][0] > mode_lib->ip.writeback_interface_buffer_size_kbytes * 1024 / ((double)mode_lib->soc.qos_parameters.writeback.base_latency_us))) {
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
8147
if (display_cfg->stream_descriptors[display_cfg->plane_descriptors[k].stream_index].writeback.writeback_stream[0].h_ratio > mode_lib->ip.writeback_max_hscl_ratio
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
8148
|| display_cfg->stream_descriptors[display_cfg->plane_descriptors[k].stream_index].writeback.writeback_stream[0].v_ratio > mode_lib->ip.writeback_max_vscl_ratio
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
8149
|| display_cfg->stream_descriptors[display_cfg->plane_descriptors[k].stream_index].writeback.writeback_stream[0].h_ratio < mode_lib->ip.writeback_min_hscl_ratio
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
8150
|| display_cfg->stream_descriptors[display_cfg->plane_descriptors[k].stream_index].writeback.writeback_stream[0].v_ratio < mode_lib->ip.writeback_min_vscl_ratio
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
8151
|| display_cfg->stream_descriptors[display_cfg->plane_descriptors[k].stream_index].writeback.writeback_stream[0].h_taps > (unsigned int) mode_lib->ip.writeback_max_hscl_taps
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
8152
|| display_cfg->stream_descriptors[display_cfg->plane_descriptors[k].stream_index].writeback.writeback_stream[0].v_taps > (unsigned int) mode_lib->ip.writeback_max_vscl_taps
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
8158
if (2.0 * display_cfg->stream_descriptors[display_cfg->plane_descriptors[k].stream_index].writeback.writeback_stream[0].output_height * (display_cfg->stream_descriptors[display_cfg->plane_descriptors[k].stream_index].writeback.writeback_stream[0].v_taps - 1) * 57 > mode_lib->ip.writeback_line_buffer_buffer_size) {
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
8170
mode_lib->ip.max_dchub_pscl_bw_pix_per_clk,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
8171
mode_lib->ip.max_pscl_lb_bw_pix_per_clk,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
8206
unsigned lb_buffer_size_bits_luma = mode_lib->ip.line_buffer_size_bits;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
8207
unsigned lb_buffer_size_bits_chroma = mode_lib->ip.line_buffer_size_bits;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
8243
if (display_cfg->plane_descriptors[k].cursor.cursor_bpp == 64 && mode_lib->ip.cursor_64bpp_support == false)
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
8282
if (mode_lib->ip.dcn_mrq_present && display_cfg->plane_descriptors[k].surface.dcc.enable) {
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
8326
CalculateSwathAndDETConfiguration_params->ConfigReturnBufferSizeInKByte = mode_lib->ip.config_return_buffer_size_in_kbytes;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
8329
CalculateSwathAndDETConfiguration_params->rob_buffer_size_kbytes = mode_lib->ip.rob_buffer_size_kbytes;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
8330
CalculateSwathAndDETConfiguration_params->pixel_chunk_size_kbytes = mode_lib->ip.pixel_chunk_size_kbytes;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
8331
CalculateSwathAndDETConfiguration_params->rob_buffer_size_kbytes = mode_lib->ip.rob_buffer_size_kbytes;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
8332
CalculateSwathAndDETConfiguration_params->pixel_chunk_size_kbytes = mode_lib->ip.pixel_chunk_size_kbytes;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
8336
CalculateSwathAndDETConfiguration_params->ConfigReturnBufferSegmentSizeInkByte = mode_lib->ip.config_return_buffer_segment_size_in_kbytes;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
8337
CalculateSwathAndDETConfiguration_params->CompressedBufferSegmentSizeInkByte = mode_lib->ip.compressed_buffer_segment_size_in_kbytes;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
8354
CalculateSwathAndDETConfiguration_params->mrq_present = mode_lib->ip.dcn_mrq_present;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
8411
mode_lib->ip.maximum_pixels_per_line_per_dsc_unit,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
8420
mode_lib->ip.max_num_dpp,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
8421
mode_lib->ip.max_num_opp,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
8432
mode_lib->ip.maximum_pixels_per_line_per_dsc_unit,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
8441
mode_lib->ip.max_num_dpp,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
8442
mode_lib->ip.max_num_opp,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
8464
mode_lib->ip.maximum_dsc_bits_per_component,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
8539
if (mode_lib->ip.imall_supported && display_cfg->plane_descriptors[k].overrides.legacy_svp_config == dml2_svp_mode_override_imall)
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
8586
if (mode_lib->ms.TotalNumberOfActiveDPP > (unsigned int)mode_lib->ip.max_num_dpp)
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
8588
if (mode_lib->ms.TotalNumberOfActiveOPP > (unsigned int)mode_lib->ip.max_num_opp)
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
8612
mode_lib->ip.writeback_line_buffer_buffer_size));
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
8661
if (s->TotalNumberOfActiveWriteback > (unsigned int)mode_lib->ip.max_num_wb) {
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
8664
mode_lib->ms.support.NumberOfOTGSupport = (s->TotalNumberOfActiveOTG <= (unsigned int)mode_lib->ip.max_num_otg);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
8665
mode_lib->ms.support.NumberOfHDMIFRLSupport = (s->TotalNumberOfActiveHDMIFRL <= (unsigned int)mode_lib->ip.max_num_hdmi_frl_outputs);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
8666
mode_lib->ms.support.NumberOfDP2p0Support = (s->TotalNumberOfActiveDP2p0 <= (unsigned int)mode_lib->ip.max_num_dp2p0_streams && s->TotalNumberOfActiveDP2p0Outputs <= (unsigned int)mode_lib->ip.max_num_dp2p0_outputs);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
8692
if (display_cfg->stream_descriptors[display_cfg->plane_descriptors[k].stream_index].output.output_format == dml2_420 && display_cfg->stream_descriptors[display_cfg->plane_descriptors[k].stream_index].timing.interlaced == 1 && mode_lib->ip.ptoi_supported == true)
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
8695
if ((display_cfg->stream_descriptors[display_cfg->plane_descriptors[k].stream_index].timing.dsc.enable == dml2_dsc_enable || display_cfg->stream_descriptors[display_cfg->plane_descriptors[k].stream_index].timing.dsc.enable == dml2_dsc_enable_if_necessary) && display_cfg->stream_descriptors[display_cfg->plane_descriptors[k].stream_index].output.output_format == dml2_n422 && !mode_lib->ip.dsc422_native_support)
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
8827
if (display_cfg->stream_descriptors[display_cfg->plane_descriptors[k].stream_index].timing.h_active > s->NumDSCUnitRequired * (unsigned int)mode_lib->ip.maximum_pixels_per_line_per_dsc_unit)
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
8838
if (s->TotalDSCUnitsRequired > (unsigned int)mode_lib->ip.num_dsc) {
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
8846
mode_lib->ip.maximum_dsc_bits_per_component,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
8939
s->SurfParameters[k].ProgressiveToInterlaceUnitInOPP = mode_lib->ip.ptoi_supported;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
8963
CalculateVMRowAndSwath_params->PTEBufferSizeInRequestsLuma = mode_lib->ip.dpte_buffer_size_in_pte_reqs_luma;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
8964
CalculateVMRowAndSwath_params->PTEBufferSizeInRequestsChroma = mode_lib->ip.dpte_buffer_size_in_pte_reqs_chroma;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
8969
CalculateVMRowAndSwath_params->DCCMetaBufferSizeBytes = mode_lib->ip.dcc_meta_buffer_size_bytes;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
8970
CalculateVMRowAndSwath_params->mrq_present = mode_lib->ip.dcn_mrq_present;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
9050
calculate_bytes_to_fetch_required_to_hide_latency_params->mrq_present = mode_lib->ip.dcn_mrq_present;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
9132
mode_lib->ip.cursor_buffer_size,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
9211
bool isInterlaceTiming = (display_cfg->stream_descriptors[display_cfg->plane_descriptors[k].stream_index].timing.interlaced && !mode_lib->ip.ptoi_supported);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
9213
mode_lib->ip.ptoi_supported,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
9214
mode_lib->ip.vblank_nom_default_us,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
9344
if (mode_lib->soc.mcache_size_bytes == 0 || mode_lib->ip.dcn_mrq_present) {
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
9367
calculate_mcache_setting_params->imall_enable = mode_lib->ip.imall_supported && display_cfg->plane_descriptors[k].overrides.legacy_svp_config == dml2_svp_mode_override_imall;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
9505
if (((mode_lib->ip.rob_buffer_size_kbytes - mode_lib->ip.pixel_chunk_size_kbytes) * 1024
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
9512
if (mode_lib->ip.rob_buffer_size_kbytes * 1024 >= mode_lib->soc.qos_parameters.qos_params.dcn32x.loaded_round_trip_latency_fclk_cycles * mode_lib->soc.fabric_datapath_to_dcn_data_return_bytes) {
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_shared_types.h
2305
struct dml2_core_ip_params ip;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_dpmm/dml2_dpmm_dcn4.c
631
dispclk_khz = mode_support_result->global.dispclk_khz * (1 + in_out->soc_bb->dcn_downspread_percent / 100.0) * (1.0 + in_out->ip->dispclk_ramp_margin_percent / 100.0);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_top/dml2_top_soc15.c
814
l->dppm_map_mode_params.ip = &dml->core_instance.clean_me_up.mode_lib.ip;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_top/dml2_top_soc15.c
986
l->dppm_map_mode_params.ip = &dml->core_instance.clean_me_up.mode_lib.ip;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/inc/dml2_internal_shared_types.h
84
struct dml2_core_ip_params *ip;
drivers/gpu/drm/amd/display/dc/dml2_0/dml2_translation_helper.c
1308
dml_dispcfg->plane.GPUVMEnable = dml2->v20.dml_core_ctx.ip.gpuvm_enable;
drivers/gpu/drm/amd/display/dc/dml2_0/dml2_translation_helper.c
1309
dml_dispcfg->plane.GPUVMMaxPageTableLevels = dml2->v20.dml_core_ctx.ip.gpuvm_max_page_table_levels;
drivers/gpu/drm/amd/display/dc/dml2_0/dml2_translation_helper.c
1310
dml_dispcfg->plane.HostVMEnable = dml2->v20.dml_core_ctx.ip.hostvm_enable;
drivers/gpu/drm/amd/display/dc/dml2_0/dml2_translation_helper.c
1311
dml_dispcfg->plane.HostVMMaxPageTableLevels = dml2->v20.dml_core_ctx.ip.hostvm_max_page_table_levels;
drivers/gpu/drm/amd/display/dc/dml2_0/dml2_translation_helper.c
1312
if (dml2->v20.dml_core_ctx.ip.hostvm_enable)
drivers/gpu/drm/amd/display/dc/dml2_0/dml2_translation_helper.c
615
const struct _vcs_dpi_ip_params_st *in_ip_params = &in->dml.ip;
drivers/gpu/drm/amd/display/dc/dml2_0/dml2_utils.c
296
context->bw_ctx.bw.dcn.compbuf_size_kb = in_ctx->v20.dml_core_ctx.ip.config_return_buffer_size_in_kbytes;
drivers/gpu/drm/amd/display/dc/dml2_0/dml2_wrapper.c
75
initialize_dml2_ip_params(*dml2, in_dc, &(*dml2)->v20.dml_core_ctx.ip);
drivers/gpu/drm/amd/display/dc/dml2_0/dml2_wrapper_fpu.c
315
s->optimize_configuration_params.ip_params = &dml2->v20.dml_core_ctx.ip;
drivers/gpu/drm/amd/display/dc/hwss/dcn20/dcn20_hwseq.c
2401
if (context->bw_ctx.dml.ip.min_comp_buffer_size_kbytes) {
drivers/gpu/drm/amd/display/dc/hwss/dcn20/dcn20_hwseq.c
2402
compbuf_size_kb = context->bw_ctx.dml.ip.min_comp_buffer_size_kbytes;
drivers/gpu/drm/amd/display/dc/hwss/dcn20/dcn20_hwseq.c
2403
dc->optimized_required |= (compbuf_size_kb != dc->current_state->bw_ctx.dml.ip.min_comp_buffer_size_kbytes);
drivers/gpu/drm/amd/display/dc/resource/dcn10/dcn10_resource.c
1672
dc->dml.ip.max_num_dpp = pool->base.pipe_count;
drivers/gpu/drm/amd/display/dc/resource/dcn31/dcn31_resource.c
1733
context->bw_ctx.dml.ip.det_buffer_size_kbytes = DCN3_1_DEFAULT_DET_SIZE;
drivers/gpu/drm/amd/display/dc/resource/dcn31/dcn31_resource.c
1741
context->bw_ctx.dml.ip.det_buffer_size_kbytes = 192;
drivers/gpu/drm/amd/display/dc/resource/dcn31/dcn31_resource.c
1746
context->bw_ctx.dml.ip.det_buffer_size_kbytes = dc->debug.crb_alloc_policy * 64;
drivers/gpu/drm/amd/display/dc/resource/dcn31/dcn31_resource.c
1748
context->bw_ctx.dml.ip.det_buffer_size_kbytes = 192;
drivers/gpu/drm/amd/display/dc/resource/dcn31/dcn31_resource.c
1757
return context->bw_ctx.dml.ip.det_buffer_size_kbytes;
drivers/gpu/drm/amd/display/dc/resource/dcn315/dcn315_resource.c
1704
const int max_usable_det = context->bw_ctx.dml.ip.config_return_buffer_size_in_kbytes - DCN3_15_MIN_COMPBUF_SIZE_KB;
drivers/gpu/drm/amd/display/dc/resource/dcn315/dcn315_resource.c
1815
context->bw_ctx.dml.ip.det_buffer_size_kbytes =
drivers/gpu/drm/amd/display/dc/resource/dcn315/dcn315_resource.c
1817
if (context->bw_ctx.dml.ip.det_buffer_size_kbytes > DCN3_15_MAX_DET_SIZE)
drivers/gpu/drm/amd/display/dc/resource/dcn315/dcn315_resource.c
1818
context->bw_ctx.dml.ip.det_buffer_size_kbytes = DCN3_15_MAX_DET_SIZE;
drivers/gpu/drm/amd/display/dc/resource/dcn315/dcn315_resource.c
1825
context->bw_ctx.dml.ip.det_buffer_size_kbytes =
drivers/gpu/drm/amd/display/dc/resource/dcn315/dcn315_resource.c
1831
context->bw_ctx.dml.ip.det_buffer_size_kbytes = 192;
drivers/gpu/drm/amd/display/dc/resource/dcn316/dcn316_resource.c
1649
const int max_usable_det = context->bw_ctx.dml.ip.config_return_buffer_size_in_kbytes - DCN3_16_MIN_COMPBUF_SIZE_KB;
drivers/gpu/drm/amd/display/dc/resource/dcn316/dcn316_resource.c
1699
context->bw_ctx.dml.ip.det_buffer_size_kbytes =
drivers/gpu/drm/amd/display/dc/resource/dcn316/dcn316_resource.c
1701
if (context->bw_ctx.dml.ip.det_buffer_size_kbytes > DCN3_16_MAX_DET_SIZE)
drivers/gpu/drm/amd/display/dc/resource/dcn316/dcn316_resource.c
1702
context->bw_ctx.dml.ip.det_buffer_size_kbytes = DCN3_16_MAX_DET_SIZE;
drivers/gpu/drm/amd/display/dc/resource/dcn316/dcn316_resource.c
1703
ASSERT(context->bw_ctx.dml.ip.det_buffer_size_kbytes >= DCN3_16_DEFAULT_DET_SIZE);
drivers/gpu/drm/amd/display/dc/resource/dcn316/dcn316_resource.c
1709
context->bw_ctx.dml.ip.det_buffer_size_kbytes =
drivers/gpu/drm/amd/display/dc/resource/dcn316/dcn316_resource.c
1712
context->bw_ctx.dml.ip.det_buffer_size_kbytes = 192;
drivers/gpu/drm/amd/display/dc/resource/dcn32/dcn32_resource.c
902
ctx->dc->dml.ip.det_buffer_size_kbytes,
drivers/gpu/drm/amd/display/dc/resource/dcn32/dcn32_resource.c
903
ctx->dc->dml.ip.pixel_chunk_size_kbytes,
drivers/gpu/drm/amd/display/dc/resource/dcn32/dcn32_resource.c
904
ctx->dc->dml.ip.config_return_buffer_size_in_kbytes);
drivers/gpu/drm/amd/display/dc/resource/dcn321/dcn321_resource.c
896
ctx->dc->dml.ip.det_buffer_size_kbytes,
drivers/gpu/drm/amd/display/dc/resource/dcn321/dcn321_resource.c
897
ctx->dc->dml.ip.pixel_chunk_size_kbytes,
drivers/gpu/drm/amd/display/dc/resource/dcn321/dcn321_resource.c
898
ctx->dc->dml.ip.config_return_buffer_size_in_kbytes);
drivers/gpu/drm/amd/include/discovery.h
162
ip *ip_list;
drivers/gpu/drm/amd/pm/swsmu/smu13/smu_v13_0_6_ppt.c
106
enum amdgpu_mca_ip ip;
drivers/gpu/drm/amd/pm/swsmu/smu13/smu_v13_0_6_ppt.c
113
enum amdgpu_mca_ip ip;
drivers/gpu/drm/amd/pm/swsmu/smu13/smu_v13_0_6_ppt.c
3537
.ip = AMDGPU_MCA_IP_UMC,
drivers/gpu/drm/amd/pm/swsmu/smu13/smu_v13_0_6_ppt.c
3541
.ip = AMDGPU_MCA_IP_SMU,
drivers/gpu/drm/amd/pm/swsmu/smu13/smu_v13_0_6_ppt.c
3546
.ip = AMDGPU_MCA_IP_SMU,
drivers/gpu/drm/amd/pm/swsmu/smu13/smu_v13_0_6_ppt.c
3553
.ip = AMDGPU_MCA_IP_SMU,
drivers/gpu/drm/amd/pm/swsmu/smu13/smu_v13_0_6_ppt.c
3560
.ip = AMDGPU_MCA_IP_PCS_XGMI,
drivers/gpu/drm/amd/pm/swsmu/smu13/smu_v13_0_6_ppt.c
3564
.ip = AMDGPU_MCA_IP_SMU,
drivers/gpu/drm/amd/pm/swsmu/smu13/smu_v13_0_6_ppt.c
3571
.ip = AMDGPU_MCA_IP_SMU,
drivers/gpu/drm/amd/pm/swsmu/smu13/smu_v13_0_6_ppt.c
3612
if (mca_decode_ipid_to_hwip(entry->regs[MCA_REG_IDX_IPID]) != mca_ras->ip)
drivers/gpu/drm/amd/ras/ras_mgr/ras_sys.h
67
#define RAS_DEV_RREG32_SOC15(dev, ip, inst, reg) \
drivers/gpu/drm/amd/ras/ras_mgr/ras_sys.h
70
__RREG32_SOC15_RLC__(adev->reg_offset[ip##_HWIP][inst][reg##_BASE_IDX] + reg, \
drivers/gpu/drm/amd/ras/ras_mgr/ras_sys.h
71
0, ip##_HWIP, inst); \
drivers/gpu/drm/amd/ras/ras_mgr/ras_sys.h
74
#define RAS_DEV_WREG32_SOC15(dev, ip, inst, reg, value) \
drivers/gpu/drm/amd/ras/ras_mgr/ras_sys.h
77
__WREG32_SOC15_RLC__((adev->reg_offset[ip##_HWIP][inst][reg##_BASE_IDX] + reg), \
drivers/gpu/drm/amd/ras/ras_mgr/ras_sys.h
78
value, 0, ip##_HWIP, inst); \
drivers/gpu/drm/amd/ras/ras_mgr/ras_sys.h
82
#define RAS_GET_INST(dev, ip, inst) \
drivers/gpu/drm/amd/ras/ras_mgr/ras_sys.h
86
adev->ip_map.logical_to_dev_inst(adev, ip##_HWIP, inst) : inst; \
drivers/gpu/drm/amd/ras/ras_mgr/ras_sys.h
89
#define RAS_GET_MASK(dev, ip, mask) \
drivers/gpu/drm/amd/ras/ras_mgr/ras_sys.h
93
adev->ip_map.logical_to_dev_mask(adev, ip##_HWIP, mask) : mask); \
drivers/gpu/drm/i915/display/intel_display_device.c
1156
.__runtime_defaults.ip.ver = 13, \
drivers/gpu/drm/i915/display/intel_display_device.c
1722
DISPLAY_RUNTIME_INFO(display)->ip = ip_ver;
drivers/gpu/drm/i915/display/intel_display_device.c
1738
DISPLAY_RUNTIME_INFO(display)->ip.ver,
drivers/gpu/drm/i915/display/intel_display_device.c
1739
DISPLAY_RUNTIME_INFO(display)->ip.rel,
drivers/gpu/drm/i915/display/intel_display_device.c
1948
if (runtime->ip.rel)
drivers/gpu/drm/i915/display/intel_display_device.c
1950
runtime->ip.ver,
drivers/gpu/drm/i915/display/intel_display_device.c
1951
runtime->ip.rel);
drivers/gpu/drm/i915/display/intel_display_device.c
1954
runtime->ip.ver);
drivers/gpu/drm/i915/display/intel_display_device.c
240
.__runtime_defaults.ip.ver = 2, \
drivers/gpu/drm/i915/display/intel_display_device.c
253
.__runtime_defaults.ip.ver = 2, \
drivers/gpu/drm/i915/display/intel_display_device.c
303
.__runtime_defaults.ip.ver = 3, \
drivers/gpu/drm/i915/display/intel_display_device.c
392
.__runtime_defaults.ip.ver = 4, \
drivers/gpu/drm/i915/display/intel_display_device.c
449
.__runtime_defaults.ip.ver = 5, \
drivers/gpu/drm/i915/display/intel_display_device.c
478
.__runtime_defaults.ip.ver = 6,
drivers/gpu/drm/i915/display/intel_display_device.c
503
.__runtime_defaults.ip.ver = 7,
drivers/gpu/drm/i915/display/intel_display_device.c
532
.__runtime_defaults.ip.ver = 7,
drivers/gpu/drm/i915/display/intel_display_device.c
580
.__runtime_defaults.ip.ver = 7,
drivers/gpu/drm/i915/display/intel_display_device.c
633
.__runtime_defaults.ip.ver = 8,
drivers/gpu/drm/i915/display/intel_display_device.c
653
.__runtime_defaults.ip.ver = 8,
drivers/gpu/drm/i915/display/intel_display_device.c
675
.__runtime_defaults.ip.ver = 9,
drivers/gpu/drm/i915/display/intel_display_device.c
848
.__runtime_defaults.ip.ver = 9,
drivers/gpu/drm/i915/display/intel_display_device.c
864
.__runtime_defaults.ip.ver = 10,
drivers/gpu/drm/i915/display/intel_display_device.c
899
.__runtime_defaults.ip.ver = 11, \
drivers/gpu/drm/i915/display/intel_display_device.c
990
.__runtime_defaults.ip.ver = 12, \
drivers/gpu/drm/i915/display/intel_display_device.h
242
#define DISPLAY_VER(__display) (DISPLAY_RUNTIME_INFO(__display)->ip.ver)
drivers/gpu/drm/i915/display/intel_display_device.h
243
#define DISPLAY_VERx100(__display) (DISPLAY_RUNTIME_INFO(__display)->ip.ver * 100 + \
drivers/gpu/drm/i915/display/intel_display_device.h
244
DISPLAY_RUNTIME_INFO(__display)->ip.rel)
drivers/gpu/drm/i915/display/intel_display_device.h
287
} ip;
drivers/gpu/drm/i915/i915_drv.h
355
#define GRAPHICS_VER(i915) (RUNTIME_INFO(i915)->graphics.ip.ver)
drivers/gpu/drm/i915/i915_drv.h
356
#define GRAPHICS_VER_FULL(i915) IP_VER(RUNTIME_INFO(i915)->graphics.ip.ver, \
drivers/gpu/drm/i915/i915_drv.h
357
RUNTIME_INFO(i915)->graphics.ip.rel)
drivers/gpu/drm/i915/i915_drv.h
361
#define MEDIA_VER(i915) (RUNTIME_INFO(i915)->media.ip.ver)
drivers/gpu/drm/i915/i915_drv.h
362
#define MEDIA_VER_FULL(i915) IP_VER(RUNTIME_INFO(i915)->media.ip.ver, \
drivers/gpu/drm/i915/i915_drv.h
363
RUNTIME_INFO(i915)->media.ip.rel)
drivers/gpu/drm/i915/i915_pci.c
45
.__runtime.graphics.ip.ver = (x), \
drivers/gpu/drm/i915/i915_pci.c
46
.__runtime.media.ip.ver = (x)
drivers/gpu/drm/i915/i915_pci.c
664
.__runtime.graphics.ip.rel = 10,
drivers/gpu/drm/i915/i915_pci.c
724
.__runtime.graphics.ip.ver = 12, \
drivers/gpu/drm/i915/i915_pci.c
725
.__runtime.graphics.ip.rel = 55, \
drivers/gpu/drm/i915/i915_pci.c
726
.__runtime.media.ip.ver = 12, \
drivers/gpu/drm/i915/i915_pci.c
727
.__runtime.media.ip.rel = 55, \
drivers/gpu/drm/i915/i915_pci.c
765
.__runtime.graphics.ip.ver = 12,
drivers/gpu/drm/i915/i915_pci.c
766
.__runtime.graphics.ip.rel = 70,
drivers/gpu/drm/i915/i915_pci.c
767
.__runtime.media.ip.ver = 13,
drivers/gpu/drm/i915/i915_pci.c
952
return i915_pci_resource_valid(pdev, intel_mmio_bar(intel_info->__runtime.graphics.ip.ver));
drivers/gpu/drm/i915/intel_device_info.c
101
if (runtime->media.ip.rel)
drivers/gpu/drm/i915/intel_device_info.c
103
runtime->media.ip.ver,
drivers/gpu/drm/i915/intel_device_info.c
104
runtime->media.ip.rel);
drivers/gpu/drm/i915/intel_device_info.c
107
runtime->media.ip.ver);
drivers/gpu/drm/i915/intel_device_info.c
297
static void ip_ver_read(struct drm_i915_private *i915, u32 offset, struct intel_ip_version *ip)
drivers/gpu/drm/i915/intel_device_info.c
302
u8 expected_ver = ip->ver;
drivers/gpu/drm/i915/intel_device_info.c
303
u8 expected_rel = ip->rel;
drivers/gpu/drm/i915/intel_device_info.c
312
ip->ver = REG_FIELD_GET(GMD_ID_ARCH_MASK, val);
drivers/gpu/drm/i915/intel_device_info.c
313
ip->rel = REG_FIELD_GET(GMD_ID_RELEASE_MASK, val);
drivers/gpu/drm/i915/intel_device_info.c
314
ip->step = REG_FIELD_GET(GMD_ID_STEP, val);
drivers/gpu/drm/i915/intel_device_info.c
317
if (IP_VER(ip->ver, ip->rel) < IP_VER(expected_ver, expected_rel))
drivers/gpu/drm/i915/intel_device_info.c
320
ip->ver, ip->rel, offset, val, expected_ver, expected_rel);
drivers/gpu/drm/i915/intel_device_info.c
338
drm_WARN_ON(&i915->drm, RUNTIME_INFO(i915)->graphics.ip.ver > 12);
drivers/gpu/drm/i915/intel_device_info.c
343
RUNTIME_INFO(i915)->media.ip =
drivers/gpu/drm/i915/intel_device_info.c
344
RUNTIME_INFO(i915)->graphics.ip;
drivers/gpu/drm/i915/intel_device_info.c
349
&runtime->graphics.ip);
drivers/gpu/drm/i915/intel_device_info.c
351
if (runtime->graphics.ip.ver == 0x0 &&
drivers/gpu/drm/i915/intel_device_info.c
353
RUNTIME_INFO(i915)->graphics.ip.ver = 12;
drivers/gpu/drm/i915/intel_device_info.c
354
RUNTIME_INFO(i915)->graphics.ip.rel = 70;
drivers/gpu/drm/i915/intel_device_info.c
357
&runtime->media.ip);
drivers/gpu/drm/i915/intel_device_info.c
93
if (runtime->graphics.ip.rel)
drivers/gpu/drm/i915/intel_device_info.c
95
runtime->graphics.ip.ver,
drivers/gpu/drm/i915/intel_device_info.c
96
runtime->graphics.ip.rel);
drivers/gpu/drm/i915/intel_device_info.c
99
runtime->graphics.ip.ver);
drivers/gpu/drm/i915/intel_device_info.h
196
struct intel_ip_version ip;
drivers/gpu/drm/i915/intel_device_info.h
199
struct intel_ip_version ip;
drivers/gpu/drm/i915/intel_step.c
159
&RUNTIME_INFO(i915)->graphics.ip);
drivers/gpu/drm/i915/intel_step.c
161
&RUNTIME_INFO(i915)->media.ip);
drivers/gpu/drm/i915/selftests/mock_gem_device.c
124
.__runtime.graphics.ip.ver = -1,
drivers/gpu/drm/lima/lima_bcast.c
11
#define bcast_write(reg, data) writel(data, ip->iomem + reg)
drivers/gpu/drm/lima/lima_bcast.c
12
#define bcast_read(reg) readl(ip->iomem + reg)
drivers/gpu/drm/lima/lima_bcast.c
17
struct lima_ip *ip = dev->ip + lima_ip_bcast;
drivers/gpu/drm/lima/lima_bcast.c
29
static int lima_bcast_hw_init(struct lima_ip *ip)
drivers/gpu/drm/lima/lima_bcast.c
31
bcast_write(LIMA_BCAST_BROADCAST_MASK, ip->data.mask << 16);
drivers/gpu/drm/lima/lima_bcast.c
32
bcast_write(LIMA_BCAST_INTERRUPT_MASK, ip->data.mask);
drivers/gpu/drm/lima/lima_bcast.c
36
int lima_bcast_resume(struct lima_ip *ip)
drivers/gpu/drm/lima/lima_bcast.c
38
return lima_bcast_hw_init(ip);
drivers/gpu/drm/lima/lima_bcast.c
41
void lima_bcast_suspend(struct lima_ip *ip)
drivers/gpu/drm/lima/lima_bcast.c
46
int lima_bcast_mask_irq(struct lima_ip *ip)
drivers/gpu/drm/lima/lima_bcast.c
53
int lima_bcast_reset(struct lima_ip *ip)
drivers/gpu/drm/lima/lima_bcast.c
55
return lima_bcast_hw_init(ip);
drivers/gpu/drm/lima/lima_bcast.c
58
int lima_bcast_init(struct lima_ip *ip)
drivers/gpu/drm/lima/lima_bcast.c
63
if (ip->dev->ip[i].present)
drivers/gpu/drm/lima/lima_bcast.c
64
ip->data.mask |= 1 << (i - lima_ip_pp0);
drivers/gpu/drm/lima/lima_bcast.c
67
return lima_bcast_hw_init(ip);
drivers/gpu/drm/lima/lima_bcast.c
70
void lima_bcast_fini(struct lima_ip *ip)
drivers/gpu/drm/lima/lima_bcast.h
10
void lima_bcast_suspend(struct lima_ip *ip);
drivers/gpu/drm/lima/lima_bcast.h
11
int lima_bcast_init(struct lima_ip *ip);
drivers/gpu/drm/lima/lima_bcast.h
12
void lima_bcast_fini(struct lima_ip *ip);
drivers/gpu/drm/lima/lima_bcast.h
16
int lima_bcast_mask_irq(struct lima_ip *ip);
drivers/gpu/drm/lima/lima_bcast.h
17
int lima_bcast_reset(struct lima_ip *ip);
drivers/gpu/drm/lima/lima_bcast.h
9
int lima_bcast_resume(struct lima_ip *ip);
drivers/gpu/drm/lima/lima_device.c
212
struct lima_ip *ip = dev->ip + index;
drivers/gpu/drm/lima/lima_device.c
221
ip->dev = dev;
drivers/gpu/drm/lima/lima_device.c
222
ip->id = index;
drivers/gpu/drm/lima/lima_device.c
223
ip->iomem = dev->iomem + offset;
drivers/gpu/drm/lima/lima_device.c
229
ip->irq = err;
drivers/gpu/drm/lima/lima_device.c
232
err = desc->init(ip);
drivers/gpu/drm/lima/lima_device.c
234
ip->present = true;
drivers/gpu/drm/lima/lima_device.c
245
struct lima_ip *ip = ldev->ip + index;
drivers/gpu/drm/lima/lima_device.c
247
if (ip->present)
drivers/gpu/drm/lima/lima_device.c
248
desc->fini(ip);
drivers/gpu/drm/lima/lima_device.c
254
struct lima_ip *ip = ldev->ip + index;
drivers/gpu/drm/lima/lima_device.c
257
if (ip->present)
drivers/gpu/drm/lima/lima_device.c
258
ret = desc->resume(ip);
drivers/gpu/drm/lima/lima_device.c
266
struct lima_ip *ip = ldev->ip + index;
drivers/gpu/drm/lima/lima_device.c
268
if (ip->present)
drivers/gpu/drm/lima/lima_device.c
269
desc->suspend(ip);
drivers/gpu/drm/lima/lima_device.c
27
int (*init)(struct lima_ip *ip);
drivers/gpu/drm/lima/lima_device.c
28
void (*fini)(struct lima_ip *ip);
drivers/gpu/drm/lima/lima_device.c
283
pipe->l2_cache[pipe->num_l2_cache++] = dev->ip + lima_ip_l2_cache0;
drivers/gpu/drm/lima/lima_device.c
284
pipe->mmu[pipe->num_mmu++] = dev->ip + lima_ip_gpmmu;
drivers/gpu/drm/lima/lima_device.c
285
pipe->processor[pipe->num_processor++] = dev->ip + lima_ip_gp;
drivers/gpu/drm/lima/lima_device.c
29
int (*resume)(struct lima_ip *ip);
drivers/gpu/drm/lima/lima_device.c
30
void (*suspend)(struct lima_ip *ip);
drivers/gpu/drm/lima/lima_device.c
316
struct lima_ip *pp = dev->ip + lima_ip_pp0 + i;
drivers/gpu/drm/lima/lima_device.c
317
struct lima_ip *ppmmu = dev->ip + lima_ip_ppmmu0 + i;
drivers/gpu/drm/lima/lima_device.c
321
l2_cache = dev->ip + lima_ip_l2_cache0;
drivers/gpu/drm/lima/lima_device.c
323
l2_cache = dev->ip + lima_ip_l2_cache1 + (i >> 2);
drivers/gpu/drm/lima/lima_device.c
333
if (dev->ip[lima_ip_bcast].present) {
drivers/gpu/drm/lima/lima_device.c
334
pipe->bcast_processor = dev->ip + lima_ip_pp_bcast;
drivers/gpu/drm/lima/lima_device.c
335
pipe->bcast_mmu = dev->ip + lima_ip_ppmmu_bcast;
drivers/gpu/drm/lima/lima_device.c
80
const char *lima_ip_name(struct lima_ip *ip)
drivers/gpu/drm/lima/lima_device.c
82
return lima_ip_desc[ip->id].name;
drivers/gpu/drm/lima/lima_device.h
120
const char *lima_ip_name(struct lima_ip *ip);
drivers/gpu/drm/lima/lima_device.h
124
static inline int lima_poll_timeout(struct lima_ip *ip, lima_poll_func_t func,
drivers/gpu/drm/lima/lima_device.h
131
if (func(ip))
drivers/gpu/drm/lima/lima_device.h
93
struct lima_ip ip[lima_ip_num];
drivers/gpu/drm/lima/lima_dlbu.c
12
#define dlbu_write(reg, data) writel(data, ip->iomem + reg)
drivers/gpu/drm/lima/lima_dlbu.c
13
#define dlbu_read(reg) readl(ip->iomem + reg)
drivers/gpu/drm/lima/lima_dlbu.c
18
struct lima_ip *ip = dev->ip + lima_ip_dlbu;
drivers/gpu/drm/lima/lima_dlbu.c
32
struct lima_ip *ip = dev->ip + lima_ip_dlbu;
drivers/gpu/drm/lima/lima_dlbu.c
37
void lima_dlbu_set_reg(struct lima_ip *ip, u32 *reg)
drivers/gpu/drm/lima/lima_dlbu.c
45
static int lima_dlbu_hw_init(struct lima_ip *ip)
drivers/gpu/drm/lima/lima_dlbu.c
47
struct lima_device *dev = ip->dev;
drivers/gpu/drm/lima/lima_dlbu.c
55
int lima_dlbu_resume(struct lima_ip *ip)
drivers/gpu/drm/lima/lima_dlbu.c
57
return lima_dlbu_hw_init(ip);
drivers/gpu/drm/lima/lima_dlbu.c
60
void lima_dlbu_suspend(struct lima_ip *ip)
drivers/gpu/drm/lima/lima_dlbu.c
65
int lima_dlbu_init(struct lima_ip *ip)
drivers/gpu/drm/lima/lima_dlbu.c
67
return lima_dlbu_hw_init(ip);
drivers/gpu/drm/lima/lima_dlbu.c
70
void lima_dlbu_fini(struct lima_ip *ip)
drivers/gpu/drm/lima/lima_dlbu.h
13
void lima_dlbu_set_reg(struct lima_ip *ip, u32 *reg);
drivers/gpu/drm/lima/lima_dlbu.h
15
int lima_dlbu_resume(struct lima_ip *ip);
drivers/gpu/drm/lima/lima_dlbu.h
16
void lima_dlbu_suspend(struct lima_ip *ip);
drivers/gpu/drm/lima/lima_dlbu.h
17
int lima_dlbu_init(struct lima_ip *ip);
drivers/gpu/drm/lima/lima_dlbu.h
18
void lima_dlbu_fini(struct lima_ip *ip);
drivers/gpu/drm/lima/lima_gp.c
100
ip->data.async_reset = false;
drivers/gpu/drm/lima/lima_gp.c
131
struct lima_ip *ip = pipe->processor[0];
drivers/gpu/drm/lima/lima_gp.c
161
lima_gp_soft_reset_async_wait(ip);
drivers/gpu/drm/lima/lima_gp.c
164
writel(f[i], ip->iomem + LIMA_GP_VSCL_START_ADDR + i * 4);
drivers/gpu/drm/lima/lima_gp.c
17
#define gp_write(reg, data) writel(data, ip->iomem + reg)
drivers/gpu/drm/lima/lima_gp.c
170
static int lima_gp_bus_stop_poll(struct lima_ip *ip)
drivers/gpu/drm/lima/lima_gp.c
175
static int lima_gp_hard_reset_poll(struct lima_ip *ip)
drivers/gpu/drm/lima/lima_gp.c
18
#define gp_read(reg) readl(ip->iomem + reg)
drivers/gpu/drm/lima/lima_gp.c
181
static int lima_gp_hard_reset(struct lima_ip *ip)
drivers/gpu/drm/lima/lima_gp.c
183
struct lima_device *dev = ip->dev;
drivers/gpu/drm/lima/lima_gp.c
190
ret = lima_poll_timeout(ip, lima_gp_bus_stop_poll, 10, 100);
drivers/gpu/drm/lima/lima_gp.c
192
dev_err(dev->dev, "%s bus stop timeout\n", lima_ip_name(ip));
drivers/gpu/drm/lima/lima_gp.c
196
ret = lima_poll_timeout(ip, lima_gp_hard_reset_poll, 10, 100);
drivers/gpu/drm/lima/lima_gp.c
198
dev_err(dev->dev, "%s hard reset timeout\n", lima_ip_name(ip));
drivers/gpu/drm/lima/lima_gp.c
210
ip->data.async_reset = false;
drivers/gpu/drm/lima/lima_gp.c
22
struct lima_ip *ip = data;
drivers/gpu/drm/lima/lima_gp.c
222
struct lima_ip *ip = pipe->processor[0];
drivers/gpu/drm/lima/lima_gp.c
224
dev_err(ip->dev->dev, "%s task error int_state=%x status=%x\n",
drivers/gpu/drm/lima/lima_gp.c
225
lima_ip_name(ip), gp_read(LIMA_GP_INT_STAT),
drivers/gpu/drm/lima/lima_gp.c
228
lima_gp_hard_reset(ip);
drivers/gpu/drm/lima/lima_gp.c
23
struct lima_device *dev = ip->dev;
drivers/gpu/drm/lima/lima_gp.c
238
struct lima_ip *ip = pipe->processor[0];
drivers/gpu/drm/lima/lima_gp.c
245
struct lima_ip *ip = pipe->processor[0];
drivers/gpu/drm/lima/lima_gp.c
273
static void lima_gp_print_version(struct lima_ip *ip)
drivers/gpu/drm/lima/lima_gp.c
298
dev_info(ip->dev->dev, "%s - %s version major %d minor %d\n",
drivers/gpu/drm/lima/lima_gp.c
299
lima_ip_name(ip), name, major, minor);
drivers/gpu/drm/lima/lima_gp.c
305
static int lima_gp_hw_init(struct lima_ip *ip)
drivers/gpu/drm/lima/lima_gp.c
307
ip->data.async_reset = false;
drivers/gpu/drm/lima/lima_gp.c
308
lima_gp_soft_reset_async(ip);
drivers/gpu/drm/lima/lima_gp.c
309
return lima_gp_soft_reset_async_wait(ip);
drivers/gpu/drm/lima/lima_gp.c
312
int lima_gp_resume(struct lima_ip *ip)
drivers/gpu/drm/lima/lima_gp.c
314
return lima_gp_hw_init(ip);
drivers/gpu/drm/lima/lima_gp.c
317
void lima_gp_suspend(struct lima_ip *ip)
drivers/gpu/drm/lima/lima_gp.c
322
int lima_gp_init(struct lima_ip *ip)
drivers/gpu/drm/lima/lima_gp.c
324
struct lima_device *dev = ip->dev;
drivers/gpu/drm/lima/lima_gp.c
327
lima_gp_print_version(ip);
drivers/gpu/drm/lima/lima_gp.c
329
err = lima_gp_hw_init(ip);
drivers/gpu/drm/lima/lima_gp.c
333
err = devm_request_irq(dev->dev, ip->irq, lima_gp_irq_handler,
drivers/gpu/drm/lima/lima_gp.c
334
IRQF_SHARED, lima_ip_name(ip), ip);
drivers/gpu/drm/lima/lima_gp.c
337
lima_ip_name(ip));
drivers/gpu/drm/lima/lima_gp.c
346
void lima_gp_fini(struct lima_ip *ip)
drivers/gpu/drm/lima/lima_gp.c
348
struct lima_device *dev = ip->dev;
drivers/gpu/drm/lima/lima_gp.c
350
devm_free_irq(dev->dev, ip->irq, ip);
drivers/gpu/drm/lima/lima_gp.c
38
lima_ip_name(ip), status);
drivers/gpu/drm/lima/lima_gp.c
41
lima_ip_name(ip), state, status);
drivers/gpu/drm/lima/lima_gp.c
68
static void lima_gp_soft_reset_async(struct lima_ip *ip)
drivers/gpu/drm/lima/lima_gp.c
70
if (ip->data.async_reset)
drivers/gpu/drm/lima/lima_gp.c
76
ip->data.async_reset = true;
drivers/gpu/drm/lima/lima_gp.c
79
static int lima_gp_soft_reset_async_wait(struct lima_ip *ip)
drivers/gpu/drm/lima/lima_gp.c
81
struct lima_device *dev = ip->dev;
drivers/gpu/drm/lima/lima_gp.c
85
if (!ip->data.async_reset)
drivers/gpu/drm/lima/lima_gp.c
88
err = readl_poll_timeout(ip->iomem + LIMA_GP_INT_RAWSTAT, v,
drivers/gpu/drm/lima/lima_gp.c
93
lima_ip_name(ip));
drivers/gpu/drm/lima/lima_gp.h
10
int lima_gp_resume(struct lima_ip *ip);
drivers/gpu/drm/lima/lima_gp.h
11
void lima_gp_suspend(struct lima_ip *ip);
drivers/gpu/drm/lima/lima_gp.h
12
int lima_gp_init(struct lima_ip *ip);
drivers/gpu/drm/lima/lima_gp.h
13
void lima_gp_fini(struct lima_ip *ip);
drivers/gpu/drm/lima/lima_l2_cache.c
11
#define l2_cache_write(reg, data) writel(data, ip->iomem + reg)
drivers/gpu/drm/lima/lima_l2_cache.c
12
#define l2_cache_read(reg) readl(ip->iomem + reg)
drivers/gpu/drm/lima/lima_l2_cache.c
14
static int lima_l2_cache_wait_idle(struct lima_ip *ip)
drivers/gpu/drm/lima/lima_l2_cache.c
16
struct lima_device *dev = ip->dev;
drivers/gpu/drm/lima/lima_l2_cache.c
20
err = readl_poll_timeout(ip->iomem + LIMA_L2_CACHE_STATUS, v,
drivers/gpu/drm/lima/lima_l2_cache.c
25
lima_ip_name(ip));
drivers/gpu/drm/lima/lima_l2_cache.c
31
int lima_l2_cache_flush(struct lima_ip *ip)
drivers/gpu/drm/lima/lima_l2_cache.c
35
spin_lock(&ip->data.lock);
drivers/gpu/drm/lima/lima_l2_cache.c
37
ret = lima_l2_cache_wait_idle(ip);
drivers/gpu/drm/lima/lima_l2_cache.c
38
spin_unlock(&ip->data.lock);
drivers/gpu/drm/lima/lima_l2_cache.c
42
static int lima_l2_cache_hw_init(struct lima_ip *ip)
drivers/gpu/drm/lima/lima_l2_cache.c
46
err = lima_l2_cache_flush(ip);
drivers/gpu/drm/lima/lima_l2_cache.c
58
int lima_l2_cache_resume(struct lima_ip *ip)
drivers/gpu/drm/lima/lima_l2_cache.c
60
return lima_l2_cache_hw_init(ip);
drivers/gpu/drm/lima/lima_l2_cache.c
63
void lima_l2_cache_suspend(struct lima_ip *ip)
drivers/gpu/drm/lima/lima_l2_cache.c
68
int lima_l2_cache_init(struct lima_ip *ip)
drivers/gpu/drm/lima/lima_l2_cache.c
72
struct lima_device *dev = ip->dev;
drivers/gpu/drm/lima/lima_l2_cache.c
75
if (ip->id == lima_ip_l2_cache2) {
drivers/gpu/drm/lima/lima_l2_cache.c
77
if (dev->ip[i].present)
drivers/gpu/drm/lima/lima_l2_cache.c
84
spin_lock_init(&ip->data.lock);
drivers/gpu/drm/lima/lima_l2_cache.c
88
lima_ip_name(ip),
drivers/gpu/drm/lima/lima_l2_cache.c
94
return lima_l2_cache_hw_init(ip);
drivers/gpu/drm/lima/lima_l2_cache.c
97
void lima_l2_cache_fini(struct lima_ip *ip)
drivers/gpu/drm/lima/lima_l2_cache.h
10
void lima_l2_cache_suspend(struct lima_ip *ip);
drivers/gpu/drm/lima/lima_l2_cache.h
11
int lima_l2_cache_init(struct lima_ip *ip);
drivers/gpu/drm/lima/lima_l2_cache.h
12
void lima_l2_cache_fini(struct lima_ip *ip);
drivers/gpu/drm/lima/lima_l2_cache.h
14
int lima_l2_cache_flush(struct lima_ip *ip);
drivers/gpu/drm/lima/lima_l2_cache.h
9
int lima_l2_cache_resume(struct lima_ip *ip);
drivers/gpu/drm/lima/lima_mmu.c
100
if (ip->id == lima_ip_ppmmu_bcast)
drivers/gpu/drm/lima/lima_mmu.c
105
dev_err(dev->dev, "%s dte write test fail\n", lima_ip_name(ip));
drivers/gpu/drm/lima/lima_mmu.c
109
err = devm_request_irq(dev->dev, ip->irq, lima_mmu_irq_handler,
drivers/gpu/drm/lima/lima_mmu.c
110
IRQF_SHARED, lima_ip_name(ip), ip);
drivers/gpu/drm/lima/lima_mmu.c
112
dev_err(dev->dev, "%s fail to request irq\n", lima_ip_name(ip));
drivers/gpu/drm/lima/lima_mmu.c
116
return lima_mmu_hw_init(ip);
drivers/gpu/drm/lima/lima_mmu.c
119
void lima_mmu_fini(struct lima_ip *ip)
drivers/gpu/drm/lima/lima_mmu.c
121
struct lima_device *dev = ip->dev;
drivers/gpu/drm/lima/lima_mmu.c
123
if (ip->id == lima_ip_ppmmu_bcast)
drivers/gpu/drm/lima/lima_mmu.c
126
devm_free_irq(dev->dev, ip->irq, ip);
drivers/gpu/drm/lima/lima_mmu.c
129
void lima_mmu_flush_tlb(struct lima_ip *ip)
drivers/gpu/drm/lima/lima_mmu.c
13
#define mmu_write(reg, data) writel(data, ip->iomem + reg)
drivers/gpu/drm/lima/lima_mmu.c
134
void lima_mmu_switch_vm(struct lima_ip *ip, struct lima_vm *vm)
drivers/gpu/drm/lima/lima_mmu.c
136
struct lima_device *dev = ip->dev;
drivers/gpu/drm/lima/lima_mmu.c
14
#define mmu_read(reg) readl(ip->iomem + reg)
drivers/gpu/drm/lima/lima_mmu.c
153
void lima_mmu_page_fault_resume(struct lima_ip *ip)
drivers/gpu/drm/lima/lima_mmu.c
155
struct lima_device *dev = ip->dev;
drivers/gpu/drm/lima/lima_mmu.c
160
dev_info(dev->dev, "%s resume\n", lima_ip_name(ip));
drivers/gpu/drm/lima/lima_mmu.c
21
__ret = readl_poll_timeout(ip->iomem + (addr), val, \
drivers/gpu/drm/lima/lima_mmu.c
26
lima_ip_name(ip), cmd); \
drivers/gpu/drm/lima/lima_mmu.c
32
struct lima_ip *ip = data;
drivers/gpu/drm/lima/lima_mmu.c
33
struct lima_device *dev = ip->dev;
drivers/gpu/drm/lima/lima_mmu.c
45
lima_ip_name(ip), fault, LIMA_MMU_STATUS_BUS_ID(status),
drivers/gpu/drm/lima/lima_mmu.c
50
dev_err(dev->dev, "%s irq bus error\n", lima_ip_name(ip));
drivers/gpu/drm/lima/lima_mmu.c
56
pipe = dev->pipe + (ip->id == lima_ip_gpmmu ? lima_pipe_gp : lima_pipe_pp);
drivers/gpu/drm/lima/lima_mmu.c
62
static int lima_mmu_hw_init(struct lima_ip *ip)
drivers/gpu/drm/lima/lima_mmu.c
64
struct lima_device *dev = ip->dev;
drivers/gpu/drm/lima/lima_mmu.c
82
int lima_mmu_resume(struct lima_ip *ip)
drivers/gpu/drm/lima/lima_mmu.c
84
if (ip->id == lima_ip_ppmmu_bcast)
drivers/gpu/drm/lima/lima_mmu.c
87
return lima_mmu_hw_init(ip);
drivers/gpu/drm/lima/lima_mmu.c
90
void lima_mmu_suspend(struct lima_ip *ip)
drivers/gpu/drm/lima/lima_mmu.c
95
int lima_mmu_init(struct lima_ip *ip)
drivers/gpu/drm/lima/lima_mmu.c
97
struct lima_device *dev = ip->dev;
drivers/gpu/drm/lima/lima_mmu.h
10
int lima_mmu_resume(struct lima_ip *ip);
drivers/gpu/drm/lima/lima_mmu.h
11
void lima_mmu_suspend(struct lima_ip *ip);
drivers/gpu/drm/lima/lima_mmu.h
12
int lima_mmu_init(struct lima_ip *ip);
drivers/gpu/drm/lima/lima_mmu.h
13
void lima_mmu_fini(struct lima_ip *ip);
drivers/gpu/drm/lima/lima_mmu.h
15
void lima_mmu_flush_tlb(struct lima_ip *ip);
drivers/gpu/drm/lima/lima_mmu.h
16
void lima_mmu_switch_vm(struct lima_ip *ip, struct lima_vm *vm);
drivers/gpu/drm/lima/lima_mmu.h
17
void lima_mmu_page_fault_resume(struct lima_ip *ip);
drivers/gpu/drm/lima/lima_pmu.c
107
if (ip->dev->id == lima_gpu_mali400)
drivers/gpu/drm/lima/lima_pmu.c
11
#define pmu_write(reg, data) writel(data, ip->iomem + reg)
drivers/gpu/drm/lima/lima_pmu.c
110
lima_pmu_wait_cmd(ip);
drivers/gpu/drm/lima/lima_pmu.c
114
int lima_pmu_resume(struct lima_ip *ip)
drivers/gpu/drm/lima/lima_pmu.c
116
return lima_pmu_hw_init(ip);
drivers/gpu/drm/lima/lima_pmu.c
119
void lima_pmu_suspend(struct lima_ip *ip)
drivers/gpu/drm/lima/lima_pmu.c
12
#define pmu_read(reg) readl(ip->iomem + reg)
drivers/gpu/drm/lima/lima_pmu.c
121
lima_pmu_hw_fini(ip);
drivers/gpu/drm/lima/lima_pmu.c
124
int lima_pmu_init(struct lima_ip *ip)
drivers/gpu/drm/lima/lima_pmu.c
126
return lima_pmu_hw_init(ip);
drivers/gpu/drm/lima/lima_pmu.c
129
void lima_pmu_fini(struct lima_ip *ip)
drivers/gpu/drm/lima/lima_pmu.c
131
lima_pmu_hw_fini(ip);
drivers/gpu/drm/lima/lima_pmu.c
14
static int lima_pmu_wait_cmd(struct lima_ip *ip)
drivers/gpu/drm/lima/lima_pmu.c
16
struct lima_device *dev = ip->dev;
drivers/gpu/drm/lima/lima_pmu.c
20
err = readl_poll_timeout(ip->iomem + LIMA_PMU_INT_RAWSTAT,
drivers/gpu/drm/lima/lima_pmu.c
25
lima_ip_name(ip));
drivers/gpu/drm/lima/lima_pmu.c
33
static u32 lima_pmu_get_ip_mask(struct lima_ip *ip)
drivers/gpu/drm/lima/lima_pmu.c
35
struct lima_device *dev = ip->dev;
drivers/gpu/drm/lima/lima_pmu.c
44
if (dev->ip[lima_ip_pp0 + i].present)
drivers/gpu/drm/lima/lima_pmu.c
48
if (dev->ip[lima_ip_pp0].present)
drivers/gpu/drm/lima/lima_pmu.c
51
if (dev->ip[i].present) {
drivers/gpu/drm/lima/lima_pmu.c
57
if (dev->ip[i].present) {
drivers/gpu/drm/lima/lima_pmu.c
67
static int lima_pmu_hw_init(struct lima_ip *ip)
drivers/gpu/drm/lima/lima_pmu.c
85
err = lima_pmu_wait_cmd(ip);
drivers/gpu/drm/lima/lima_pmu.c
92
static void lima_pmu_hw_fini(struct lima_ip *ip)
drivers/gpu/drm/lima/lima_pmu.c
96
if (!ip->data.mask)
drivers/gpu/drm/lima/lima_pmu.c
97
ip->data.mask = lima_pmu_get_ip_mask(ip);
drivers/gpu/drm/lima/lima_pmu.c
99
stat = ~pmu_read(LIMA_PMU_STATUS) & ip->data.mask;
drivers/gpu/drm/lima/lima_pmu.h
10
void lima_pmu_suspend(struct lima_ip *ip);
drivers/gpu/drm/lima/lima_pmu.h
11
int lima_pmu_init(struct lima_ip *ip);
drivers/gpu/drm/lima/lima_pmu.h
12
void lima_pmu_fini(struct lima_ip *ip);
drivers/gpu/drm/lima/lima_pmu.h
9
int lima_pmu_resume(struct lima_ip *ip);
drivers/gpu/drm/lima/lima_pp.c
104
static void lima_pp_soft_reset_async(struct lima_ip *ip)
drivers/gpu/drm/lima/lima_pp.c
106
if (ip->data.async_reset)
drivers/gpu/drm/lima/lima_pp.c
112
ip->data.async_reset = true;
drivers/gpu/drm/lima/lima_pp.c
115
static int lima_pp_soft_reset_poll(struct lima_ip *ip)
drivers/gpu/drm/lima/lima_pp.c
121
static int lima_pp_soft_reset_async_wait_one(struct lima_ip *ip)
drivers/gpu/drm/lima/lima_pp.c
123
struct lima_device *dev = ip->dev;
drivers/gpu/drm/lima/lima_pp.c
126
ret = lima_poll_timeout(ip, lima_pp_soft_reset_poll, 0, 100);
drivers/gpu/drm/lima/lima_pp.c
128
dev_err(dev->dev, "%s reset time out\n", lima_ip_name(ip));
drivers/gpu/drm/lima/lima_pp.c
137
static int lima_pp_soft_reset_async_wait(struct lima_ip *ip)
drivers/gpu/drm/lima/lima_pp.c
141
if (!ip->data.async_reset)
drivers/gpu/drm/lima/lima_pp.c
144
if (ip->id == lima_ip_pp_bcast) {
drivers/gpu/drm/lima/lima_pp.c
145
struct lima_device *dev = ip->dev;
drivers/gpu/drm/lima/lima_pp.c
152
err = lima_pp_soft_reset_async_wait_one(ip);
drivers/gpu/drm/lima/lima_pp.c
154
ip->data.async_reset = false;
drivers/gpu/drm/lima/lima_pp.c
158
static void lima_pp_write_frame(struct lima_ip *ip, u32 *frame, u32 *wb)
drivers/gpu/drm/lima/lima_pp.c
163
writel(frame[i], ip->iomem + LIMA_PP_FRAME + i * 4);
drivers/gpu/drm/lima/lima_pp.c
167
writel(wb[n++], ip->iomem + LIMA_PP_WB(i) + j * 4);
drivers/gpu/drm/lima/lima_pp.c
171
static int lima_pp_bus_stop_poll(struct lima_ip *ip)
drivers/gpu/drm/lima/lima_pp.c
176
static int lima_pp_hard_reset_poll(struct lima_ip *ip)
drivers/gpu/drm/lima/lima_pp.c
18
#define pp_write(reg, data) writel(data, ip->iomem + reg)
drivers/gpu/drm/lima/lima_pp.c
182
static int lima_pp_hard_reset(struct lima_ip *ip)
drivers/gpu/drm/lima/lima_pp.c
184
struct lima_device *dev = ip->dev;
drivers/gpu/drm/lima/lima_pp.c
19
#define pp_read(reg) readl(ip->iomem + reg)
drivers/gpu/drm/lima/lima_pp.c
191
ret = lima_poll_timeout(ip, lima_pp_bus_stop_poll, 10, 100);
drivers/gpu/drm/lima/lima_pp.c
193
dev_err(dev->dev, "%s bus stop timeout\n", lima_ip_name(ip));
drivers/gpu/drm/lima/lima_pp.c
198
ret = lima_poll_timeout(ip, lima_pp_hard_reset_poll, 10, 100);
drivers/gpu/drm/lima/lima_pp.c
200
dev_err(dev->dev, "%s hard reset timeout\n", lima_ip_name(ip));
drivers/gpu/drm/lima/lima_pp.c
21
static void lima_pp_handle_irq(struct lima_ip *ip, u32 state)
drivers/gpu/drm/lima/lima_pp.c
212
ip->data.async_reset = false;
drivers/gpu/drm/lima/lima_pp.c
217
static void lima_pp_print_version(struct lima_ip *ip)
drivers/gpu/drm/lima/lima_pp.c
23
struct lima_device *dev = ip->dev;
drivers/gpu/drm/lima/lima_pp.c
242
dev_info(ip->dev->dev, "%s - %s version major %d minor %d\n",
drivers/gpu/drm/lima/lima_pp.c
243
lima_ip_name(ip), name, major, minor);
drivers/gpu/drm/lima/lima_pp.c
246
static int lima_pp_hw_init(struct lima_ip *ip)
drivers/gpu/drm/lima/lima_pp.c
248
ip->data.async_reset = false;
drivers/gpu/drm/lima/lima_pp.c
249
lima_pp_soft_reset_async(ip);
drivers/gpu/drm/lima/lima_pp.c
250
return lima_pp_soft_reset_async_wait(ip);
drivers/gpu/drm/lima/lima_pp.c
253
int lima_pp_resume(struct lima_ip *ip)
drivers/gpu/drm/lima/lima_pp.c
255
return lima_pp_hw_init(ip);
drivers/gpu/drm/lima/lima_pp.c
258
void lima_pp_suspend(struct lima_ip *ip)
drivers/gpu/drm/lima/lima_pp.c
263
int lima_pp_init(struct lima_ip *ip)
drivers/gpu/drm/lima/lima_pp.c
265
struct lima_device *dev = ip->dev;
drivers/gpu/drm/lima/lima_pp.c
268
lima_pp_print_version(ip);
drivers/gpu/drm/lima/lima_pp.c
270
err = lima_pp_hw_init(ip);
drivers/gpu/drm/lima/lima_pp.c
274
err = devm_request_irq(dev->dev, ip->irq, lima_pp_irq_handler,
drivers/gpu/drm/lima/lima_pp.c
275
IRQF_SHARED, lima_ip_name(ip), ip);
drivers/gpu/drm/lima/lima_pp.c
278
lima_ip_name(ip));
drivers/gpu/drm/lima/lima_pp.c
287
void lima_pp_fini(struct lima_ip *ip)
drivers/gpu/drm/lima/lima_pp.c
289
struct lima_device *dev = ip->dev;
drivers/gpu/drm/lima/lima_pp.c
291
devm_free_irq(dev->dev, ip->irq, ip);
drivers/gpu/drm/lima/lima_pp.c
294
int lima_pp_bcast_resume(struct lima_ip *ip)
drivers/gpu/drm/lima/lima_pp.c
297
ip->data.async_reset = false;
drivers/gpu/drm/lima/lima_pp.c
30
lima_ip_name(ip), state, status);
drivers/gpu/drm/lima/lima_pp.c
301
void lima_pp_bcast_suspend(struct lima_ip *ip)
drivers/gpu/drm/lima/lima_pp.c
306
int lima_pp_bcast_init(struct lima_ip *ip)
drivers/gpu/drm/lima/lima_pp.c
308
struct lima_device *dev = ip->dev;
drivers/gpu/drm/lima/lima_pp.c
311
err = devm_request_irq(dev->dev, ip->irq, lima_pp_bcast_irq_handler,
drivers/gpu/drm/lima/lima_pp.c
312
IRQF_SHARED, lima_ip_name(ip), ip);
drivers/gpu/drm/lima/lima_pp.c
315
lima_ip_name(ip));
drivers/gpu/drm/lima/lima_pp.c
322
void lima_pp_bcast_fini(struct lima_ip *ip)
drivers/gpu/drm/lima/lima_pp.c
324
struct lima_device *dev = ip->dev;
drivers/gpu/drm/lima/lima_pp.c
326
devm_free_irq(dev->dev, ip->irq, ip);
drivers/gpu/drm/lima/lima_pp.c
359
struct lima_ip *ip = pipe->bcast_processor;
drivers/gpu/drm/lima/lima_pp.c
369
lima_dlbu_set_reg(dev->ip + lima_ip_dlbu, frame->dlbu_regs);
drivers/gpu/drm/lima/lima_pp.c
375
lima_pp_soft_reset_async_wait(ip);
drivers/gpu/drm/lima/lima_pp.c
377
lima_pp_write_frame(ip, frame->frame, frame->wb);
drivers/gpu/drm/lima/lima_pp.c
380
struct lima_ip *ip = pipe->processor[i];
drivers/gpu/drm/lima/lima_pp.c
395
struct lima_ip *ip = pipe->processor[i];
drivers/gpu/drm/lima/lima_pp.c
402
lima_pp_soft_reset_async_wait(ip);
drivers/gpu/drm/lima/lima_pp.c
404
lima_pp_write_frame(ip, frame->frame, frame->wb);
drivers/gpu/drm/lima/lima_pp.c
428
struct lima_ip *ip = pipe->processor[i];
drivers/gpu/drm/lima/lima_pp.c
43
struct lima_ip *ip = data;
drivers/gpu/drm/lima/lima_pp.c
430
dev_err(ip->dev->dev, "%s task error %d int_state=%x status=%x\n",
drivers/gpu/drm/lima/lima_pp.c
431
lima_ip_name(ip), i, pp_read(LIMA_PP_INT_STATUS),
drivers/gpu/drm/lima/lima_pp.c
434
lima_pp_hard_reset(ip);
drivers/gpu/drm/lima/lima_pp.c
44
struct lima_device *dev = ip->dev;
drivers/gpu/drm/lima/lima_pp.c
452
struct lima_ip *ip = pipe->processor[i];
drivers/gpu/drm/lima/lima_pp.c
52
lima_pp_handle_irq(ip, state);
drivers/gpu/drm/lima/lima_pp.c
76
struct lima_ip *ip = pipe->processor[i];
drivers/gpu/drm/lima/lima_pp.c
89
lima_pp_handle_irq(ip, state);
drivers/gpu/drm/lima/lima_pp.h
10
int lima_pp_resume(struct lima_ip *ip);
drivers/gpu/drm/lima/lima_pp.h
11
void lima_pp_suspend(struct lima_ip *ip);
drivers/gpu/drm/lima/lima_pp.h
12
int lima_pp_init(struct lima_ip *ip);
drivers/gpu/drm/lima/lima_pp.h
13
void lima_pp_fini(struct lima_ip *ip);
drivers/gpu/drm/lima/lima_pp.h
15
int lima_pp_bcast_resume(struct lima_ip *ip);
drivers/gpu/drm/lima/lima_pp.h
16
void lima_pp_bcast_suspend(struct lima_ip *ip);
drivers/gpu/drm/lima/lima_pp.h
17
int lima_pp_bcast_init(struct lima_ip *ip);
drivers/gpu/drm/lima/lima_pp.h
18
void lima_pp_bcast_fini(struct lima_ip *ip);
drivers/gpu/drm/lima/lima_sched.c
277
struct lima_ip *ip = pipe->processor[0];
drivers/gpu/drm/lima/lima_sched.c
278
int pipe_id = ip->id == lima_ip_gp ? lima_pipe_gp : lima_pipe_pp;
drivers/gpu/drm/lima/lima_sched.c
279
struct lima_device *dev = ip->dev;
drivers/gpu/drm/lima/lima_sched.c
408
struct lima_ip *ip = pipe->processor[0];
drivers/gpu/drm/lima/lima_sched.c
416
DRM_WARN("%s spurious timeout\n", lima_ip_name(ip));
drivers/gpu/drm/lima/lima_sched.c
433
DRM_WARN("%s unexpectedly high interrupt latency\n", lima_ip_name(ip));
drivers/gpu/drm/lima/lima_sched.c
445
DRM_ERROR("%s job timeout\n", lima_ip_name(ip));
drivers/gpu/drm/nouveau/nvkm/engine/gr/ctxnv40.h
81
int ip = 0;
drivers/gpu/drm/nouveau/nvkm/engine/gr/ctxnv40.h
84
ip = ctx->ctxprog_label[name] << CP_BRA_IP_SHIFT;
drivers/gpu/drm/nouveau/nvkm/engine/gr/ctxnv40.h
85
if (ip == 0)
drivers/gpu/drm/nouveau/nvkm/engine/gr/ctxnv40.h
86
ip = 0xff000000 | (name << CP_BRA_IP_SHIFT);
drivers/gpu/drm/nouveau/nvkm/engine/gr/ctxnv40.h
89
cp_out(ctx, CP_BRA | (mod << 18) | ip | flag |
drivers/gpu/drm/nouveau/nvkm/falcon/gm200.c
33
int nr = (tidx & 0x00ff0000) >> 16, sp, ip;
drivers/gpu/drm/nouveau/nvkm/falcon/gm200.c
38
ip = nvkm_falcon_rd32(falcon, 0x14c);
drivers/gpu/drm/nouveau/nvkm/falcon/gm200.c
39
FLCN_ERR(falcon, "TRACEPC: %08x", ip);
drivers/gpu/drm/vc4/vc4_validate_shaders.c
189
uint64_t inst = validation_state->shader[validation_state->ip];
drivers/gpu/drm/vc4/vc4_validate_shaders.c
313
uint64_t inst = validation_state->shader[validation_state->ip];
drivers/gpu/drm/vc4/vc4_validate_shaders.c
392
uint64_t inst = validation_state->shader[validation_state->ip];
drivers/gpu/drm/vc4/vc4_validate_shaders.c
483
uint64_t inst = validation_state->shader[validation_state->ip];
drivers/gpu/drm/vc4/vc4_validate_shaders.c
53
uint32_t ip;
drivers/gpu/drm/vc4/vc4_validate_shaders.c
558
uint64_t inst = validation_state->shader[validation_state->ip];
drivers/gpu/drm/vc4/vc4_validate_shaders.c
580
int ip)
drivers/gpu/drm/vc4/vc4_validate_shaders.c
594
validation_state->ip);
drivers/gpu/drm/vc4/vc4_validate_shaders.c
605
uint64_t inst = validation_state->shader[validation_state->ip];
drivers/gpu/drm/vc4/vc4_validate_shaders.c
640
int ip;
drivers/gpu/drm/vc4/vc4_validate_shaders.c
643
for (ip = 0; ip < validation_state->max_ip; ip++) {
drivers/gpu/drm/vc4/vc4_validate_shaders.c
644
uint64_t inst = validation_state->shader[ip];
drivers/gpu/drm/vc4/vc4_validate_shaders.c
647
uint32_t after_delay_ip = ip + 4;
drivers/gpu/drm/vc4/vc4_validate_shaders.c
657
validation_state->max_ip = ip + 3;
drivers/gpu/drm/vc4/vc4_validate_shaders.c
664
if (ip - last_branch < 4) {
drivers/gpu/drm/vc4/vc4_validate_shaders.c
665
DRM_DEBUG("Branch at %d during delay slots\n", ip);
drivers/gpu/drm/vc4/vc4_validate_shaders.c
668
last_branch = ip;
drivers/gpu/drm/vc4/vc4_validate_shaders.c
694
ip, branch_target_ip,
drivers/gpu/drm/vc4/vc4_validate_shaders.c
706
ip, after_delay_ip, validation_state->max_ip);
drivers/gpu/drm/vc4/vc4_validate_shaders.c
749
uint32_t ip = validation_state->ip;
drivers/gpu/drm/vc4/vc4_validate_shaders.c
751
if (!test_bit(ip, validation_state->branch_targets))
drivers/gpu/drm/vc4/vc4_validate_shaders.c
787
uint32_t ip;
drivers/gpu/drm/vc4/vc4_validate_shaders.c
813
for (ip = 0; ip < validation_state.max_ip; ip++) {
drivers/gpu/drm/vc4/vc4_validate_shaders.c
814
uint64_t inst = validation_state.shader[ip];
drivers/gpu/drm/vc4/vc4_validate_shaders.c
817
validation_state.ip = ip;
drivers/gpu/drm/vc4/vc4_validate_shaders.c
822
if (ip == last_thread_switch_ip + 3) {
drivers/gpu/drm/vc4/vc4_validate_shaders.c
846
DRM_DEBUG("Bad write at ip %d\n", ip);
drivers/gpu/drm/vc4/vc4_validate_shaders.c
856
shader_end_ip = ip;
drivers/gpu/drm/vc4/vc4_validate_shaders.c
863
if (ip < last_thread_switch_ip + 3) {
drivers/gpu/drm/vc4/vc4_validate_shaders.c
865
"last switch at ip %d\n", ip);
drivers/gpu/drm/vc4/vc4_validate_shaders.c
868
last_thread_switch_ip = ip;
drivers/gpu/drm/vc4/vc4_validate_shaders.c
876
DRM_DEBUG("Bad LOAD_IMM write at ip %d\n", ip);
drivers/gpu/drm/vc4/vc4_validate_shaders.c
883
&validation_state, ip))
drivers/gpu/drm/vc4/vc4_validate_shaders.c
886
if (ip < last_thread_switch_ip + 3) {
drivers/gpu/drm/vc4/vc4_validate_shaders.c
888
ip);
drivers/gpu/drm/vc4/vc4_validate_shaders.c
895
"instruction %d\n", sig, ip);
drivers/gpu/drm/vc4/vc4_validate_shaders.c
902
if (found_shader_end && ip == shader_end_ip + 2)
drivers/gpu/drm/vc4/vc4_validate_shaders.c
906
if (ip == validation_state.max_ip) {
drivers/gpu/drm/xe/tests/xe_pci.c
156
const struct xe_ip *ip = find_graphics_ip(verx100);
drivers/gpu/drm/xe/tests/xe_pci.c
158
return ip ? ip->name : "";
drivers/gpu/drm/xe/tests/xe_pci.c
163
const struct xe_ip *ip = find_media_ip(verx100);
drivers/gpu/drm/xe/tests/xe_pci.c
165
return ip ? ip->name : "";
drivers/gpu/drm/xe/xe_guc_ct.c
2178
struct drm_printer ip = xe_gt_info_printer(gt);
drivers/gpu/drm/xe/xe_guc_ct.c
2179
struct drm_printer lp = drm_line_printer(&ip, "Capture", ++g_count);
drivers/gpu/drm/xe/xe_guc_log.c
583
struct drm_printer ip = xe_gt_info_printer(gt);
drivers/gpu/drm/xe/xe_guc_log.c
584
struct drm_printer lp = drm_line_printer(&ip, "Capture", ++g_count);
drivers/iio/industrialio-event.c
558
int __user *ip = (int __user *)arg;
drivers/iio/industrialio-event.c
565
if (copy_to_user(ip, &fd, sizeof(fd)))
drivers/infiniband/core/roce_gid_mgmt.c
338
struct sockaddr_in ip;
drivers/infiniband/core/roce_gid_mgmt.c
360
entry->ip.sin_family = AF_INET;
drivers/infiniband/core/roce_gid_mgmt.c
361
entry->ip.sin_addr.s_addr = ifa->ifa_address;
drivers/infiniband/core/roce_gid_mgmt.c
369
(struct sockaddr *)&sin_iter->ip);
drivers/infiniband/hw/irdma/cm.h
399
int irdma_add_arp(struct irdma_pci_f *rf, u32 *ip, bool ipv4, const u8 *mac);
drivers/infiniband/hw/irdma/utils.c
18
u32 ip[4] = {};
drivers/infiniband/hw/irdma/utils.c
21
ip[0] = *ip_addr;
drivers/infiniband/hw/irdma/utils.c
23
memcpy(ip, ip_addr, sizeof(ip));
drivers/infiniband/hw/irdma/utils.c
27
if (!memcmp(rf->arp_table[arp_index].ip_addr, ip, sizeof(ip)))
drivers/infiniband/hw/irdma/utils.c
45
memcpy(rf->arp_table[arp_index].ip_addr, ip,
drivers/infiniband/hw/irdma/utils.c
80
int irdma_add_arp(struct irdma_pci_f *rf, u32 *ip, bool ipv4, const u8 *mac)
drivers/infiniband/hw/irdma/utils.c
84
arpidx = irdma_arp_table(rf, &ip[0], ipv4, NULL, IRDMA_ARP_RESOLVE);
drivers/infiniband/hw/irdma/utils.c
89
irdma_manage_arp_cache(rf, rf->arp_table[arpidx].mac_addr, ip,
drivers/infiniband/hw/irdma/utils.c
93
irdma_manage_arp_cache(rf, mac, ip, ipv4, IRDMA_ARP_ADD);
drivers/infiniband/hw/irdma/utils.c
95
return irdma_arp_table(rf, ip, ipv4, NULL, IRDMA_ARP_RESOLVE);
drivers/infiniband/hw/irdma/verbs.c
4780
u8 *ip = (u8 *)ip_addr;
drivers/infiniband/hw/irdma/verbs.c
4786
mac4[3] = ip[2] & 0x7F;
drivers/infiniband/hw/irdma/verbs.c
4787
mac4[4] = ip[1];
drivers/infiniband/hw/irdma/verbs.c
4788
mac4[5] = ip[0];
drivers/infiniband/hw/irdma/verbs.c
4794
mac6[2] = ip[3];
drivers/infiniband/hw/irdma/verbs.c
4795
mac6[3] = ip[2];
drivers/infiniband/hw/irdma/verbs.c
4796
mac6[4] = ip[1];
drivers/infiniband/hw/irdma/verbs.c
4797
mac6[5] = ip[0];
drivers/infiniband/sw/rdmavt/cq.c
210
cq->ip = rvt_create_mmap_info(rdi, sz, udata, u_wc);
drivers/infiniband/sw/rdmavt/cq.c
211
if (IS_ERR(cq->ip)) {
drivers/infiniband/sw/rdmavt/cq.c
212
err = PTR_ERR(cq->ip);
drivers/infiniband/sw/rdmavt/cq.c
216
err = ib_copy_to_udata(udata, &cq->ip->offset,
drivers/infiniband/sw/rdmavt/cq.c
217
sizeof(cq->ip->offset));
drivers/infiniband/sw/rdmavt/cq.c
232
if (cq->ip) {
drivers/infiniband/sw/rdmavt/cq.c
234
list_add(&cq->ip->pending_mmaps, &rdi->pending_mmaps);
drivers/infiniband/sw/rdmavt/cq.c
264
kfree(cq->ip);
drivers/infiniband/sw/rdmavt/cq.c
287
if (cq->ip)
drivers/infiniband/sw/rdmavt/cq.c
288
kref_put(&cq->ip->ref, rvt_release_mmap_info);
drivers/infiniband/sw/rdmavt/cq.c
39
if (cq->ip) {
drivers/infiniband/sw/rdmavt/cq.c
434
if (cq->ip) {
drivers/infiniband/sw/rdmavt/cq.c
435
struct rvt_mmap_info *ip = cq->ip;
drivers/infiniband/sw/rdmavt/cq.c
437
rvt_update_mmap_info(rdi, ip, sz, u_wc);
drivers/infiniband/sw/rdmavt/cq.c
444
ret = ib_copy_to_udata(udata, &ip->offset,
drivers/infiniband/sw/rdmavt/cq.c
445
sizeof(ip->offset));
drivers/infiniband/sw/rdmavt/cq.c
451
if (list_empty(&ip->pending_mmaps))
drivers/infiniband/sw/rdmavt/cq.c
452
list_add(&ip->pending_mmaps, &rdi->pending_mmaps);
drivers/infiniband/sw/rdmavt/cq.c
487
if (cq->ip)
drivers/infiniband/sw/rdmavt/mmap.c
119
struct rvt_mmap_info *ip;
drivers/infiniband/sw/rdmavt/mmap.c
124
ip = kmalloc_node(sizeof(*ip), GFP_KERNEL, rdi->dparms.node);
drivers/infiniband/sw/rdmavt/mmap.c
125
if (!ip)
drivers/infiniband/sw/rdmavt/mmap.c
133
ip->offset = rdi->mmap_offset;
drivers/infiniband/sw/rdmavt/mmap.c
137
INIT_LIST_HEAD(&ip->pending_mmaps);
drivers/infiniband/sw/rdmavt/mmap.c
138
ip->size = size;
drivers/infiniband/sw/rdmavt/mmap.c
139
ip->context =
drivers/infiniband/sw/rdmavt/mmap.c
142
ip->obj = obj;
drivers/infiniband/sw/rdmavt/mmap.c
143
kref_init(&ip->ref);
drivers/infiniband/sw/rdmavt/mmap.c
145
return ip;
drivers/infiniband/sw/rdmavt/mmap.c
155
void rvt_update_mmap_info(struct rvt_dev_info *rdi, struct rvt_mmap_info *ip,
drivers/infiniband/sw/rdmavt/mmap.c
163
ip->offset = rdi->mmap_offset;
drivers/infiniband/sw/rdmavt/mmap.c
167
ip->size = size;
drivers/infiniband/sw/rdmavt/mmap.c
168
ip->obj = obj;
drivers/infiniband/sw/rdmavt/mmap.c
30
struct rvt_mmap_info *ip =
drivers/infiniband/sw/rdmavt/mmap.c
32
struct rvt_dev_info *rdi = ib_to_rvt(ip->context->device);
drivers/infiniband/sw/rdmavt/mmap.c
35
list_del(&ip->pending_mmaps);
drivers/infiniband/sw/rdmavt/mmap.c
38
vfree(ip->obj);
drivers/infiniband/sw/rdmavt/mmap.c
39
kfree(ip);
drivers/infiniband/sw/rdmavt/mmap.c
44
struct rvt_mmap_info *ip = vma->vm_private_data;
drivers/infiniband/sw/rdmavt/mmap.c
46
kref_get(&ip->ref);
drivers/infiniband/sw/rdmavt/mmap.c
51
struct rvt_mmap_info *ip = vma->vm_private_data;
drivers/infiniband/sw/rdmavt/mmap.c
53
kref_put(&ip->ref, rvt_release_mmap_info);
drivers/infiniband/sw/rdmavt/mmap.c
73
struct rvt_mmap_info *ip, *pp;
drivers/infiniband/sw/rdmavt/mmap.c
82
list_for_each_entry_safe(ip, pp, &rdi->pending_mmaps,
drivers/infiniband/sw/rdmavt/mmap.c
85
if (context != ip->context || (__u64)offset != ip->offset)
drivers/infiniband/sw/rdmavt/mmap.c
88
if (size > ip->size)
drivers/infiniband/sw/rdmavt/mmap.c
91
list_del_init(&ip->pending_mmaps);
drivers/infiniband/sw/rdmavt/mmap.c
94
ret = remap_vmalloc_range(vma, ip->obj, 0);
drivers/infiniband/sw/rdmavt/mmap.c
98
vma->vm_private_data = ip;
drivers/infiniband/sw/rdmavt/mmap.h
16
void rvt_update_mmap_info(struct rvt_dev_info *rdi, struct rvt_mmap_info *ip,
drivers/infiniband/sw/rdmavt/qp.c
1204
qp->ip = rvt_create_mmap_info(rdi, s, udata,
drivers/infiniband/sw/rdmavt/qp.c
1206
if (IS_ERR(qp->ip)) {
drivers/infiniband/sw/rdmavt/qp.c
1207
ret = PTR_ERR(qp->ip);
drivers/infiniband/sw/rdmavt/qp.c
1211
ret = ib_copy_to_udata(udata, &qp->ip->offset,
drivers/infiniband/sw/rdmavt/qp.c
1212
sizeof(qp->ip->offset));
drivers/infiniband/sw/rdmavt/qp.c
1242
if (qp->ip) {
drivers/infiniband/sw/rdmavt/qp.c
1244
list_add(&qp->ip->pending_mmaps, &rdi->pending_mmaps);
drivers/infiniband/sw/rdmavt/qp.c
1251
if (qp->ip)
drivers/infiniband/sw/rdmavt/qp.c
1252
kref_put(&qp->ip->ref, rvt_release_mmap_info);
drivers/infiniband/sw/rdmavt/qp.c
1332
if (qp->ip) {
drivers/infiniband/sw/rdmavt/qp.c
1352
if (qp->ip)
drivers/infiniband/sw/rdmavt/qp.c
1694
if (qp->ip)
drivers/infiniband/sw/rdmavt/qp.c
1695
kref_put(&qp->ip->ref, rvt_release_mmap_info);
drivers/infiniband/sw/rdmavt/qp.c
2319
static inline u32 get_rvt_head(struct rvt_rq *rq, void *ip)
drivers/infiniband/sw/rdmavt/qp.c
2323
if (ip)
drivers/infiniband/sw/rdmavt/qp.c
2353
void *ip = NULL;
drivers/infiniband/sw/rdmavt/qp.c
2359
ip = srq->ip;
drivers/infiniband/sw/rdmavt/qp.c
2364
ip = qp->ip;
drivers/infiniband/sw/rdmavt/qp.c
2373
if (ip) {
drivers/infiniband/sw/rdmavt/qp.c
2385
head = get_rvt_head(rq, ip);
drivers/infiniband/sw/rdmavt/qp.c
2402
if (ip)
drivers/infiniband/sw/rdmavt/qp.c
2423
get_rvt_head(rq, ip), tail);
drivers/infiniband/sw/rdmavt/rc.c
71
if (qp->ip) {
drivers/infiniband/sw/rdmavt/srq.c
101
list_add(&srq->ip->pending_mmaps, &dev->pending_mmaps);
drivers/infiniband/sw/rdmavt/srq.c
108
kfree(srq->ip);
drivers/infiniband/sw/rdmavt/srq.c
227
if (srq->ip) {
drivers/infiniband/sw/rdmavt/srq.c
228
struct rvt_mmap_info *ip = srq->ip;
drivers/infiniband/sw/rdmavt/srq.c
232
rvt_update_mmap_info(dev, ip, s, tmp_rq.wq);
drivers/infiniband/sw/rdmavt/srq.c
239
ret = ib_copy_to_udata(udata, &ip->offset,
drivers/infiniband/sw/rdmavt/srq.c
240
sizeof(ip->offset));
drivers/infiniband/sw/rdmavt/srq.c
250
if (list_empty(&ip->pending_mmaps))
drivers/infiniband/sw/rdmavt/srq.c
251
list_add(&ip->pending_mmaps,
drivers/infiniband/sw/rdmavt/srq.c
302
if (srq->ip)
drivers/infiniband/sw/rdmavt/srq.c
303
kref_put(&srq->ip->ref, rvt_release_mmap_info);
drivers/infiniband/sw/rdmavt/srq.c
71
srq->ip = rvt_create_mmap_info(dev, s, udata, srq->rq.wq);
drivers/infiniband/sw/rdmavt/srq.c
72
if (IS_ERR(srq->ip)) {
drivers/infiniband/sw/rdmavt/srq.c
73
ret = PTR_ERR(srq->ip);
drivers/infiniband/sw/rdmavt/srq.c
77
ret = ib_copy_to_udata(udata, &srq->ip->offset,
drivers/infiniband/sw/rdmavt/srq.c
78
sizeof(srq->ip->offset));
drivers/infiniband/sw/rdmavt/srq.c
99
if (srq->ip) {
drivers/infiniband/sw/rdmavt/trace_cq.h
40
__field(struct rvt_mmap_info *, ip)
drivers/infiniband/sw/rdmavt/trace_cq.h
47
__entry->ip = cq->ip;
drivers/infiniband/sw/rdmavt/trace_cq.h
55
__entry->ip ? "true" : "false", __entry->cqe,
drivers/infiniband/sw/rxe/rxe_cq.c
58
cq->queue->buf, cq->queue->buf_size, &cq->queue->ip);
drivers/infiniband/sw/rxe/rxe_mmap.c
106
vma->vm_private_data = ip;
drivers/infiniband/sw/rxe/rxe_mmap.c
118
struct rxe_mmap_info *ip;
drivers/infiniband/sw/rxe/rxe_mmap.c
123
ip = kmalloc_obj(*ip);
drivers/infiniband/sw/rxe/rxe_mmap.c
124
if (!ip)
drivers/infiniband/sw/rxe/rxe_mmap.c
134
ip->info.offset = rxe->mmap_offset;
drivers/infiniband/sw/rxe/rxe_mmap.c
139
INIT_LIST_HEAD(&ip->pending_mmaps);
drivers/infiniband/sw/rxe/rxe_mmap.c
140
ip->info.size = size;
drivers/infiniband/sw/rxe/rxe_mmap.c
141
ip->context =
drivers/infiniband/sw/rxe/rxe_mmap.c
144
ip->obj = obj;
drivers/infiniband/sw/rxe/rxe_mmap.c
145
kref_init(&ip->ref);
drivers/infiniband/sw/rxe/rxe_mmap.c
147
return ip;
drivers/infiniband/sw/rxe/rxe_mmap.c
18
struct rxe_mmap_info *ip = container_of(ref,
drivers/infiniband/sw/rxe/rxe_mmap.c
20
struct rxe_dev *rxe = to_rdev(ip->context->device);
drivers/infiniband/sw/rxe/rxe_mmap.c
24
if (!list_empty(&ip->pending_mmaps))
drivers/infiniband/sw/rxe/rxe_mmap.c
25
list_del(&ip->pending_mmaps);
drivers/infiniband/sw/rxe/rxe_mmap.c
29
vfree(ip->obj); /* buf */
drivers/infiniband/sw/rxe/rxe_mmap.c
30
kfree(ip);
drivers/infiniband/sw/rxe/rxe_mmap.c
39
struct rxe_mmap_info *ip = vma->vm_private_data;
drivers/infiniband/sw/rxe/rxe_mmap.c
41
kref_get(&ip->ref);
drivers/infiniband/sw/rxe/rxe_mmap.c
46
struct rxe_mmap_info *ip = vma->vm_private_data;
drivers/infiniband/sw/rxe/rxe_mmap.c
48
kref_put(&ip->ref, rxe_mmap_release);
drivers/infiniband/sw/rxe/rxe_mmap.c
67
struct rxe_mmap_info *ip, *pp;
drivers/infiniband/sw/rxe/rxe_mmap.c
76
list_for_each_entry_safe(ip, pp, &rxe->pending_mmaps, pending_mmaps) {
drivers/infiniband/sw/rxe/rxe_mmap.c
77
if (context != ip->context || (__u64)offset != ip->info.offset)
drivers/infiniband/sw/rxe/rxe_mmap.c
81
if (size > ip->info.size) {
drivers/infiniband/sw/rxe/rxe_mmap.c
96
list_del_init(&ip->pending_mmaps);
drivers/infiniband/sw/rxe/rxe_mmap.c
99
ret = remap_vmalloc_range(vma, ip->obj, 0);
drivers/infiniband/sw/rxe/rxe_qp.c
260
&qp->sq.queue->ip);
drivers/infiniband/sw/rxe/rxe_qp.c
351
&qp->rq.queue->ip);
drivers/infiniband/sw/rxe/rxe_queue.c
163
new_q->buf_size, &new_q->ip);
drivers/infiniband/sw/rxe/rxe_queue.c
17
struct rxe_mmap_info *ip = NULL;
drivers/infiniband/sw/rxe/rxe_queue.c
195
if (q->ip)
drivers/infiniband/sw/rxe/rxe_queue.c
196
kref_put(&q->ip->ref, rxe_mmap_release);
drivers/infiniband/sw/rxe/rxe_queue.c
20
ip = rxe_create_mmap_info(rxe, buf_size, udata, buf);
drivers/infiniband/sw/rxe/rxe_queue.c
21
if (IS_ERR(ip)) {
drivers/infiniband/sw/rxe/rxe_queue.c
22
err = PTR_ERR(ip);
drivers/infiniband/sw/rxe/rxe_queue.c
26
if (copy_to_user(outbuf, &ip->info, sizeof(ip->info))) {
drivers/infiniband/sw/rxe/rxe_queue.c
32
list_add(&ip->pending_mmaps, &rxe->pending_mmaps);
drivers/infiniband/sw/rxe/rxe_queue.c
36
*ip_p = ip;
drivers/infiniband/sw/rxe/rxe_queue.c
41
kfree(ip);
drivers/infiniband/sw/rxe/rxe_queue.h
65
struct rxe_mmap_info *ip;
drivers/infiniband/sw/rxe/rxe_srq.c
74
q->buf_size, &q->ip);
drivers/input/evdev.c
1036
int __user *ip = (int __user *)p;
drivers/input/evdev.c
1045
return put_user(EV_VERSION, ip);
drivers/input/evdev.c
1055
if (put_user(dev->rep[REP_DELAY], ip))
drivers/input/evdev.c
1057
if (put_user(dev->rep[REP_PERIOD], ip + 1))
drivers/input/evdev.c
1064
if (get_user(u, ip))
drivers/input/evdev.c
1066
if (get_user(v, ip + 1))
drivers/input/evdev.c
1080
if (put_user(i, ip))
drivers/input/evdev.c
1150
return evdev_handle_mt_request(dev, size, ip);
drivers/input/evdev.c
806
int __user *ip = (int __user *)p;
drivers/input/evdev.c
817
if (put_user(ke.keycode, ip + 1))
drivers/input/evdev.c
847
int __user *ip = (int __user *)p;
drivers/input/evdev.c
852
if (get_user(ke.keycode, ip + 1))
drivers/input/evdev.c
919
int __user *ip)
drivers/input/evdev.c
926
if (get_user(code, &ip[0]))
drivers/input/evdev.c
934
if (put_user(value, &ip[1 + i]))
drivers/ipack/devices/scc2698.h
61
u8 dd, ip; /* Input port register of block */
drivers/isdn/hardware/mISDN/ipac.h
48
struct ipac_hw *ip;
drivers/isdn/hardware/mISDN/mISDNipac.c
1002
hscx->bch.nr, hscx->ip->name, count);
drivers/isdn/hardware/mISDN/mISDNipac.c
1029
if (hx->ip->type & IPAC_TYPE_IPACX)
drivers/isdn/hardware/mISDN/mISDNipac.c
1033
pr_debug("%s: B%1d RSTAB %02x\n", hx->ip->name, hx->bch.nr, rstab);
drivers/isdn/hardware/mISDN/mISDNipac.c
1039
hx->ip->name, hx->bch.nr);
drivers/isdn/hardware/mISDN/mISDNipac.c
1044
hx->ip->name, hx->bch.nr,
drivers/isdn/hardware/mISDN/mISDNipac.c
1050
hx->ip->name, hx->bch.nr);
drivers/isdn/hardware/mISDN/mISDNipac.c
1055
if (hx->ip->type & IPAC_TYPE_IPACX)
drivers/isdn/hardware/mISDN/mISDNipac.c
1067
hx->ip->name, hx->bch.nr, hx->bch.rx_skb->len);
drivers/isdn/hardware/mISDN/mISDNipac.c
1080
if (hx->ip->type & IPAC_TYPE_IPACX)
drivers/isdn/hardware/mISDN/mISDNipac.c
1082
else if (hx->ip->type & IPAC_TYPE_IPAC) {
drivers/isdn/hardware/mISDN/mISDNipac.c
1087
pr_debug("%s: B%1d EXIRB %02x\n", hx->ip->name,
drivers/isdn/hardware/mISDN/mISDNipac.c
1092
ipac_irq(&hx->ip->hscx[0], ista);
drivers/isdn/hardware/mISDN/mISDNipac.c
1095
pr_debug("%s: B%1d EXIRB %02x\n", hx->ip->name,
drivers/isdn/hardware/mISDN/mISDNipac.c
1103
pr_debug("%s: B%1d EXIRB %02x\n", hx->ip->name,
drivers/isdn/hardware/mISDN/mISDNipac.c
1112
pr_debug("%s: B%1d ISTAB %02x\n", hx->ip->name, hx->bch.nr, istab);
drivers/isdn/hardware/mISDN/mISDNipac.c
1127
pr_debug("%s: B%1d RFO error\n", hx->ip->name, hx->bch.nr);
drivers/isdn/hardware/mISDN/mISDNipac.c
1141
pr_debug("%s: B%1d XDU error at len %d\n", hx->ip->name,
drivers/isdn/hardware/mISDN/mISDNipac.c
1215
pr_debug("%s: HSCX %c protocol %x-->%x ch %d\n", hscx->ip->name,
drivers/isdn/hardware/mISDN/mISDNipac.c
1217
if (hscx->ip->type & IPAC_TYPE_IPACX) {
drivers/isdn/hardware/mISDN/mISDNipac.c
1219
WriteIPAC(hscx->ip, ISACX_BCHA_TSDP_BC1, 0x80);
drivers/isdn/hardware/mISDN/mISDNipac.c
1220
WriteIPAC(hscx->ip, ISACX_BCHA_CR, 0x88);
drivers/isdn/hardware/mISDN/mISDNipac.c
1222
WriteIPAC(hscx->ip, ISACX_BCHB_TSDP_BC1, 0x81);
drivers/isdn/hardware/mISDN/mISDNipac.c
1223
WriteIPAC(hscx->ip, ISACX_BCHB_CR, 0x88);
drivers/isdn/hardware/mISDN/mISDNipac.c
1249
pr_info("%s: protocol not known %x\n", hscx->ip->name,
drivers/isdn/hardware/mISDN/mISDNipac.c
1253
} else if (hscx->ip->type & IPAC_TYPE_IPAC) { /* IPAC */
drivers/isdn/hardware/mISDN/mISDNipac.c
1285
pr_info("%s: protocol not known %x\n", hscx->ip->name,
drivers/isdn/hardware/mISDN/mISDNipac.c
1289
} else if (hscx->ip->type & IPAC_TYPE_HSCX) { /* HSCX */
drivers/isdn/hardware/mISDN/mISDNipac.c
1321
pr_info("%s: protocol not known %x\n", hscx->ip->name,
drivers/isdn/hardware/mISDN/mISDNipac.c
1342
spin_lock_irqsave(hx->ip->hwlock, flags);
drivers/isdn/hardware/mISDN/mISDNipac.c
1348
spin_unlock_irqrestore(hx->ip->hwlock, flags);
drivers/isdn/hardware/mISDN/mISDNipac.c
1351
spin_lock_irqsave(hx->ip->hwlock, flags);
drivers/isdn/hardware/mISDN/mISDNipac.c
1356
spin_unlock_irqrestore(hx->ip->hwlock, flags);
drivers/isdn/hardware/mISDN/mISDNipac.c
1362
spin_lock_irqsave(hx->ip->hwlock, flags);
drivers/isdn/hardware/mISDN/mISDNipac.c
1365
spin_unlock_irqrestore(hx->ip->hwlock, flags);
drivers/isdn/hardware/mISDN/mISDNipac.c
1372
hx->ip->name, __func__, hh->prim, hh->id);
drivers/isdn/hardware/mISDN/mISDNipac.c
1394
pr_debug("%s: %s cmd:%x %p\n", hx->ip->name, __func__, cmd, arg);
drivers/isdn/hardware/mISDN/mISDNipac.c
1399
spin_lock_irqsave(hx->ip->hwlock, flags);
drivers/isdn/hardware/mISDN/mISDNipac.c
1402
spin_unlock_irqrestore(hx->ip->hwlock, flags);
drivers/isdn/hardware/mISDN/mISDNipac.c
1405
module_put(hx->ip->owner);
drivers/isdn/hardware/mISDN/mISDNipac.c
1413
hx->ip->name, __func__, cmd);
drivers/isdn/hardware/mISDN/mISDNipac.c
1439
if (hx->ip->type & IPAC_TYPE_HSCX) {
drivers/isdn/hardware/mISDN/mISDNipac.c
1442
pr_debug("%s: HSCX VSTR %02x\n", hx->ip->name, val);
drivers/isdn/hardware/mISDN/mISDNipac.c
1444
pr_notice("%s: HSCX version %s\n", hx->ip->name,
drivers/isdn/hardware/mISDN/mISDNipac.c
1608
ipac->hscx[i].ip = ipac;
drivers/isdn/hardware/mISDN/mISDNipac.c
29
#define ReadHSCX(h, o) (h->ip->read_reg(h->ip->hw, h->off + o))
drivers/isdn/hardware/mISDN/mISDNipac.c
30
#define WriteHSCX(h, o, v) (h->ip->write_reg(h->ip->hw, h->off + o, v))
drivers/isdn/hardware/mISDN/mISDNipac.c
31
#define ReadIPAC(ip, o) (ip->read_reg(ip->hw, o))
drivers/isdn/hardware/mISDN/mISDNipac.c
32
#define WriteIPAC(ip, o, v) (ip->write_reg(ip->hw, o, v))
drivers/isdn/hardware/mISDN/mISDNipac.c
886
pr_debug("%s: B%1d CEC %d us\n", hx->ip->name, hx->bch.nr,
drivers/isdn/hardware/mISDN/mISDNipac.c
889
pr_info("%s: B%1d CEC timeout\n", hx->ip->name, hx->bch.nr);
drivers/isdn/hardware/mISDN/mISDNipac.c
906
pr_debug("%s: B%1d XFW %d us\n", hx->ip->name, hx->bch.nr,
drivers/isdn/hardware/mISDN/mISDNipac.c
909
pr_info("%s: B%1d XFW timeout\n", hx->ip->name, hx->bch.nr);
drivers/isdn/hardware/mISDN/mISDNipac.c
915
if (hx->ip->type & IPAC_TYPE_IPACX)
drivers/isdn/hardware/mISDN/mISDNipac.c
929
pr_debug("%s: B%1d %d\n", hscx->ip->name, hscx->bch.nr, count);
drivers/isdn/hardware/mISDN/mISDNipac.c
941
hscx->ip->name, hscx->bch.nr, count);
drivers/isdn/hardware/mISDN/mISDNipac.c
946
if (hscx->ip->type & IPAC_TYPE_IPACX)
drivers/isdn/hardware/mISDN/mISDNipac.c
947
hscx->ip->read_fifo(hscx->ip->hw,
drivers/isdn/hardware/mISDN/mISDNipac.c
950
hscx->ip->read_fifo(hscx->ip->hw,
drivers/isdn/hardware/mISDN/mISDNipac.c
957
hscx->bch.nr, hscx->ip->name, count);
drivers/isdn/hardware/mISDN/mISDNipac.c
986
pr_debug("%s: B%1d %d/%d/%d\n", hscx->ip->name, hscx->bch.nr,
drivers/isdn/hardware/mISDN/mISDNipac.c
990
if (hscx->ip->type & IPAC_TYPE_IPACX)
drivers/isdn/hardware/mISDN/mISDNipac.c
991
hscx->ip->write_fifo(hscx->ip->hw,
drivers/isdn/hardware/mISDN/mISDNipac.c
995
hscx->ip->write_fifo(hscx->ip->hw,
drivers/isdn/mISDN/l1oip_core.c
1356
hc->remoteip = ip[l1oip_cnt << 2] << 24
drivers/isdn/mISDN/l1oip_core.c
1357
| ip[(l1oip_cnt << 2) + 1] << 16
drivers/isdn/mISDN/l1oip_core.c
1358
| ip[(l1oip_cnt << 2) + 2] << 8
drivers/isdn/mISDN/l1oip_core.c
1359
| ip[(l1oip_cnt << 2) + 3];
drivers/isdn/mISDN/l1oip_core.c
238
static u_int ip[MAX_CARDS * 4];
drivers/isdn/mISDN/l1oip_core.c
252
module_param_array(ip, uint, NULL, S_IRUGO | S_IWUSR);
drivers/media/platform/renesas/rzg2l-cru/rzg2l-core.c
67
sink = &cru->ip.subdev.entity;
drivers/media/platform/renesas/rzg2l-cru/rzg2l-core.c
76
cru->ip.remote = cru->csi.subdev;
drivers/media/platform/renesas/rzg2l-cru/rzg2l-core.c
79
source = &cru->ip.subdev.entity;
drivers/media/platform/renesas/rzg2l-cru/rzg2l-cru.h
154
struct rzg2l_cru_ip ip;
drivers/media/platform/renesas/rzg2l-cru/rzg2l-ip.c
148
state = v4l2_subdev_lock_and_get_active_state(&cru->ip.subdev);
drivers/media/platform/renesas/rzg2l-cru/rzg2l-ip.c
164
ret = v4l2_subdev_call(cru->ip.remote, video, s_stream, enable);
drivers/media/platform/renesas/rzg2l-cru/rzg2l-ip.c
168
ret = v4l2_subdev_call(cru->ip.remote, video, post_streamoff);
drivers/media/platform/renesas/rzg2l-cru/rzg2l-ip.c
175
ret = v4l2_subdev_call(cru->ip.remote, video, pre_streamon, 0);
drivers/media/platform/renesas/rzg2l-cru/rzg2l-ip.c
185
v4l2_subdev_call(cru->ip.remote, video, post_streamoff);
drivers/media/platform/renesas/rzg2l-cru/rzg2l-ip.c
189
ret = v4l2_subdev_call(cru->ip.remote, video, s_stream, enable);
drivers/media/platform/renesas/rzg2l-cru/rzg2l-ip.c
195
v4l2_subdev_call(cru->ip.remote, video, post_streamoff);
drivers/media/platform/renesas/rzg2l-cru/rzg2l-ip.c
332
struct rzg2l_cru_ip *ip = &cru->ip;
drivers/media/platform/renesas/rzg2l-cru/rzg2l-ip.c
335
ip->subdev.dev = cru->dev;
drivers/media/platform/renesas/rzg2l-cru/rzg2l-ip.c
336
v4l2_subdev_init(&ip->subdev, &rzg2l_cru_ip_subdev_ops);
drivers/media/platform/renesas/rzg2l-cru/rzg2l-ip.c
337
ip->subdev.internal_ops = &rzg2l_cru_ip_internal_ops;
drivers/media/platform/renesas/rzg2l-cru/rzg2l-ip.c
338
v4l2_set_subdevdata(&ip->subdev, cru);
drivers/media/platform/renesas/rzg2l-cru/rzg2l-ip.c
339
snprintf(ip->subdev.name, sizeof(ip->subdev.name),
drivers/media/platform/renesas/rzg2l-cru/rzg2l-ip.c
341
ip->subdev.flags = V4L2_SUBDEV_FL_HAS_DEVNODE;
drivers/media/platform/renesas/rzg2l-cru/rzg2l-ip.c
343
ip->subdev.entity.function = MEDIA_ENT_F_PROC_VIDEO_PIXEL_FORMATTER;
drivers/media/platform/renesas/rzg2l-cru/rzg2l-ip.c
344
ip->subdev.entity.ops = &rzg2l_cru_ip_entity_ops;
drivers/media/platform/renesas/rzg2l-cru/rzg2l-ip.c
346
ip->pads[RZG2L_CRU_IP_SINK].flags = MEDIA_PAD_FL_SINK |
drivers/media/platform/renesas/rzg2l-cru/rzg2l-ip.c
348
ip->pads[RZG2L_CRU_IP_SOURCE].flags = MEDIA_PAD_FL_SOURCE |
drivers/media/platform/renesas/rzg2l-cru/rzg2l-ip.c
351
ret = media_entity_pads_init(&ip->subdev.entity, 2, ip->pads);
drivers/media/platform/renesas/rzg2l-cru/rzg2l-ip.c
355
ret = v4l2_subdev_init_finalize(&ip->subdev);
drivers/media/platform/renesas/rzg2l-cru/rzg2l-ip.c
359
ret = v4l2_device_register_subdev(&cru->v4l2_dev, &ip->subdev);
drivers/media/platform/renesas/rzg2l-cru/rzg2l-ip.c
365
v4l2_subdev_cleanup(&ip->subdev);
drivers/media/platform/renesas/rzg2l-cru/rzg2l-ip.c
367
media_entity_cleanup(&ip->subdev.entity);
drivers/media/platform/renesas/rzg2l-cru/rzg2l-ip.c
374
struct rzg2l_cru_ip *ip = &cru->ip;
drivers/media/platform/renesas/rzg2l-cru/rzg2l-ip.c
376
media_entity_cleanup(&ip->subdev.entity);
drivers/media/platform/renesas/rzg2l-cru/rzg2l-ip.c
377
v4l2_subdev_cleanup(&ip->subdev);
drivers/media/platform/renesas/rzg2l-cru/rzg2l-ip.c
378
v4l2_device_unregister_subdev(&ip->subdev);
drivers/media/platform/renesas/rzg2l-cru/rzg2l-video.c
413
remote_pad = media_pad_remote_pad_unique(&cru->ip.pads[RZG2L_CRU_IP_SINK]);
drivers/media/platform/renesas/rzg2l-cru/rzg2l-video.c
414
ret = v4l2_subdev_call(cru->ip.remote, pad, get_frame_desc, remote_pad->index, &fd);
drivers/media/usb/pvrusb2/pvrusb2-std.c
176
const struct std_name *ip,*gp;
drivers/media/usb/pvrusb2/pvrusb2-std.c
185
ip = std_items + idx2;
drivers/media/usb/pvrusb2/pvrusb2-std.c
186
if (!(gp->id & ip->id & id)) continue;
drivers/media/usb/pvrusb2/pvrusb2-std.c
205
ip->name);
drivers/media/usb/usbtv/usbtv-video.c
463
static void usbtv_iso_cb(struct urb *ip)
drivers/media/usb/usbtv/usbtv-video.c
467
struct usbtv *usbtv = (struct usbtv *)ip->context;
drivers/media/usb/usbtv/usbtv-video.c
469
switch (ip->status) {
drivers/media/usb/usbtv/usbtv-video.c
485
for (i = 0; i < ip->number_of_packets; i++) {
drivers/media/usb/usbtv/usbtv-video.c
486
int size = ip->iso_frame_desc[i].actual_length;
drivers/media/usb/usbtv/usbtv-video.c
487
unsigned char *data = ip->transfer_buffer +
drivers/media/usb/usbtv/usbtv-video.c
488
ip->iso_frame_desc[i].offset;
drivers/media/usb/usbtv/usbtv-video.c
497
ret = usb_submit_urb(ip, GFP_ATOMIC);
drivers/media/usb/usbtv/usbtv-video.c
504
struct urb *ip;
drivers/media/usb/usbtv/usbtv-video.c
508
ip = usb_alloc_urb(USBTV_ISOC_PACKETS, GFP_KERNEL);
drivers/media/usb/usbtv/usbtv-video.c
509
if (ip == NULL)
drivers/media/usb/usbtv/usbtv-video.c
512
ip->dev = usbtv->udev;
drivers/media/usb/usbtv/usbtv-video.c
513
ip->context = usbtv;
drivers/media/usb/usbtv/usbtv-video.c
514
ip->pipe = usb_rcvisocpipe(usbtv->udev, USBTV_VIDEO_ENDP);
drivers/media/usb/usbtv/usbtv-video.c
515
ip->interval = 1;
drivers/media/usb/usbtv/usbtv-video.c
516
ip->transfer_flags = URB_ISO_ASAP;
drivers/media/usb/usbtv/usbtv-video.c
517
ip->transfer_buffer = kcalloc(USBTV_ISOC_PACKETS, size,
drivers/media/usb/usbtv/usbtv-video.c
519
if (!ip->transfer_buffer) {
drivers/media/usb/usbtv/usbtv-video.c
520
usb_free_urb(ip);
drivers/media/usb/usbtv/usbtv-video.c
523
ip->complete = usbtv_iso_cb;
drivers/media/usb/usbtv/usbtv-video.c
524
ip->number_of_packets = USBTV_ISOC_PACKETS;
drivers/media/usb/usbtv/usbtv-video.c
525
ip->transfer_buffer_length = size * USBTV_ISOC_PACKETS;
drivers/media/usb/usbtv/usbtv-video.c
527
ip->iso_frame_desc[i].offset = size * i;
drivers/media/usb/usbtv/usbtv-video.c
528
ip->iso_frame_desc[i].length = size;
drivers/media/usb/usbtv/usbtv-video.c
531
return ip;
drivers/media/usb/usbtv/usbtv-video.c
541
struct urb *ip = usbtv->isoc_urbs[i];
drivers/media/usb/usbtv/usbtv-video.c
543
if (ip == NULL)
drivers/media/usb/usbtv/usbtv-video.c
545
usb_kill_urb(ip);
drivers/media/usb/usbtv/usbtv-video.c
546
kfree(ip->transfer_buffer);
drivers/media/usb/usbtv/usbtv-video.c
547
usb_free_urb(ip);
drivers/media/usb/usbtv/usbtv-video.c
584
struct urb *ip;
drivers/media/usb/usbtv/usbtv-video.c
586
ip = usbtv_setup_iso_transfer(usbtv);
drivers/media/usb/usbtv/usbtv-video.c
587
if (ip == NULL) {
drivers/media/usb/usbtv/usbtv-video.c
591
usbtv->isoc_urbs[i] = ip;
drivers/media/usb/usbtv/usbtv-video.c
593
ret = usb_submit_urb(ip, GFP_KERNEL);
drivers/mfd/qcom-pm8xxx.c
83
unsigned int *ip)
drivers/mfd/qcom-pm8xxx.c
94
rc = regmap_read(chip->regmap, SSBI_REG_ADDR_IRQ_IT_STATUS, ip);
drivers/misc/hpilo.c
531
static int ilo_close(struct inode *ip, struct file *fp)
drivers/misc/hpilo.c
538
slot = iminor(ip) % max_ccb;
drivers/misc/hpilo.c
539
hw = container_of(ip->i_cdev, struct ilo_hwinfo, cdev);
drivers/misc/hpilo.c
562
static int ilo_open(struct inode *ip, struct file *fp)
drivers/misc/hpilo.c
569
slot = iminor(ip) % max_ccb;
drivers/misc/hpilo.c
570
hw = container_of(ip->i_cdev, struct ilo_hwinfo, cdev);
drivers/misc/kgdbts.c
309
unsigned long ip;
drivers/misc/kgdbts.c
315
ip = instruction_pointer(&kgdbts_regs);
drivers/misc/kgdbts.c
316
v2printk("Stopped at IP: %lx\n", ip);
drivers/misc/kgdbts.c
319
if (addr + BREAK_INSTR_SIZE == ip)
drivers/misc/kgdbts.c
324
ip + offset == sstep_addr &&
drivers/misc/kgdbts.c
329
} else if (strcmp(arg, "silent") && ip + offset != addr) {
drivers/misc/kgdbts.c
331
ip + offset, addr);
drivers/misc/kgdbts.c
335
ip += offset;
drivers/misc/kgdbts.c
336
cont_addr = ip;
drivers/misc/kgdbts.c
338
instruction_pointer_set(&kgdbts_regs, ip);
drivers/mtd/inftlmount.c
177
ip = &mh->Partitions[i];
drivers/mtd/inftlmount.c
178
ip->virtualUnits = le32_to_cpu(ip->virtualUnits);
drivers/mtd/inftlmount.c
179
ip->firstUnit = le32_to_cpu(ip->firstUnit);
drivers/mtd/inftlmount.c
180
ip->lastUnit = le32_to_cpu(ip->lastUnit);
drivers/mtd/inftlmount.c
181
ip->flags = le32_to_cpu(ip->flags);
drivers/mtd/inftlmount.c
182
ip->spareUnits = le32_to_cpu(ip->spareUnits);
drivers/mtd/inftlmount.c
183
ip->Reserved0 = le32_to_cpu(ip->Reserved0);
drivers/mtd/inftlmount.c
191
i, ip->virtualUnits, ip->firstUnit,
drivers/mtd/inftlmount.c
192
ip->lastUnit, ip->flags,
drivers/mtd/inftlmount.c
193
ip->spareUnits);
drivers/mtd/inftlmount.c
195
if (ip->Reserved0 != ip->firstUnit) {
drivers/mtd/inftlmount.c
205
instr->addr = ip->Reserved0 * inftl->EraseSize;
drivers/mtd/inftlmount.c
209
if ((ip->lastUnit - ip->firstUnit + 1) < ip->virtualUnits) {
drivers/mtd/inftlmount.c
213
"virtualUnits %d\n", i, ip->lastUnit,
drivers/mtd/inftlmount.c
214
ip->firstUnit, ip->Reserved0);
drivers/mtd/inftlmount.c
217
if (ip->Reserved1 != 0) {
drivers/mtd/inftlmount.c
221
i, ip->Reserved1);
drivers/mtd/inftlmount.c
225
if (ip->flags & INFTL_BDTL)
drivers/mtd/inftlmount.c
236
inftl->nb_boot_blocks = ip->firstUnit;
drivers/mtd/inftlmount.c
237
inftl->numvunits = ip->virtualUnits;
drivers/mtd/inftlmount.c
255
inftl->firstEUN = ip->firstUnit;
drivers/mtd/inftlmount.c
256
inftl->lastEUN = ip->lastUnit;
drivers/mtd/inftlmount.c
257
inftl->nb_blocks = ip->lastUnit + 1;
drivers/mtd/inftlmount.c
39
struct INFTLPartition *ip;
drivers/mtd/nand/raw/diskonchip.c
1018
struct INFTLPartition *ip;
drivers/mtd/nand/raw/diskonchip.c
1079
ip = &(mh->Partitions[i]);
drivers/mtd/nand/raw/diskonchip.c
1080
le32_to_cpus(&ip->virtualUnits);
drivers/mtd/nand/raw/diskonchip.c
1081
le32_to_cpus(&ip->firstUnit);
drivers/mtd/nand/raw/diskonchip.c
1082
le32_to_cpus(&ip->lastUnit);
drivers/mtd/nand/raw/diskonchip.c
1083
le32_to_cpus(&ip->flags);
drivers/mtd/nand/raw/diskonchip.c
1084
le32_to_cpus(&ip->spareUnits);
drivers/mtd/nand/raw/diskonchip.c
1085
le32_to_cpus(&ip->Reserved0);
drivers/mtd/nand/raw/diskonchip.c
1093
i, ip->virtualUnits, ip->firstUnit,
drivers/mtd/nand/raw/diskonchip.c
1094
ip->lastUnit, ip->flags,
drivers/mtd/nand/raw/diskonchip.c
1095
ip->spareUnits);
drivers/mtd/nand/raw/diskonchip.c
1098
(i == 0) && (ip->firstUnit > 0)) {
drivers/mtd/nand/raw/diskonchip.c
1101
parts[0].size = (uint64_t)mtd->erasesize * ip->firstUnit;
drivers/mtd/nand/raw/diskonchip.c
1105
if (ip->flags & INFTL_BINARY)
drivers/mtd/nand/raw/diskonchip.c
1109
parts[numparts].offset = ip->firstUnit << vshift;
drivers/mtd/nand/raw/diskonchip.c
1110
parts[numparts].size = (1 + ip->lastUnit - ip->firstUnit) << vshift;
drivers/mtd/nand/raw/diskonchip.c
1112
if (ip->lastUnit > lastvunit)
drivers/mtd/nand/raw/diskonchip.c
1113
lastvunit = ip->lastUnit;
drivers/mtd/nand/raw/diskonchip.c
1114
if (ip->flags & INFTL_LAST)
drivers/net/bonding/bond_main.c
2931
__be32 ip = *(__be32 *)priv->data;
drivers/net/bonding/bond_main.c
2933
return ip == bond_confirm_addr(upper, 0, ip);
drivers/net/bonding/bond_main.c
2936
static bool bond_has_this_ip(struct bonding *bond, __be32 ip)
drivers/net/bonding/bond_main.c
2939
.data = (void *)&ip,
drivers/net/bonding/bond_main.c
2943
if (ip == bond_confirm_addr(bond->dev, 0, ip))
drivers/net/bonding/bond_main.c
6267
__be32 ip;
drivers/net/bonding/bond_main.c
6270
if (!in4_pton(arp_ip_target[i], -1, (u8 *)&ip, -1, NULL) ||
drivers/net/bonding/bond_main.c
6271
!bond_is_ip_target_ok(ip)) {
drivers/net/bonding/bond_main.c
6276
if (bond_get_targets_ip(arp_target, ip) == -1)
drivers/net/bonding/bond_main.c
6277
arp_target[arp_ip_count++] = ip;
drivers/net/bonding/bond_main.c
6280
&ip);
drivers/net/dsa/bcm_sf2_cfp.c
358
struct flow_match_ip ip;
drivers/net/dsa/bcm_sf2_cfp.c
400
flow_rule_match_ip(flow->rule, &ip);
drivers/net/dsa/bcm_sf2_cfp.c
435
core_writel(priv, ip.key->tos << IPTOS_SHIFT |
drivers/net/ethernet/apple/bmac.c
1485
int i,*ip;
drivers/net/ethernet/apple/bmac.c
1488
ip = (int*)(cp+i);
drivers/net/ethernet/apple/bmac.c
1491
le32_to_cpup(ip+0),
drivers/net/ethernet/apple/bmac.c
1492
le32_to_cpup(ip+1),
drivers/net/ethernet/apple/bmac.c
1493
le32_to_cpup(ip+2),
drivers/net/ethernet/apple/bmac.c
1494
le32_to_cpup(ip+3));
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
1338
} ip;
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
1340
ip.hdr = skb_network_header(skb);
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
1361
if (ip.v4->version == 4)
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
1366
proto = (ip.v4->version == 4) ? ip.v4->protocol :
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
1367
ip.v6->nexthdr;
drivers/net/ethernet/chelsio/cxgb4/clip_tbl.c
255
char ip[60];
drivers/net/ethernet/chelsio/cxgb4/clip_tbl.c
263
ip[0] = '\0';
drivers/net/ethernet/chelsio/cxgb4/clip_tbl.c
264
sprintf(ip, "%pISc", &ce->addr);
drivers/net/ethernet/chelsio/cxgb4/clip_tbl.c
265
seq_printf(seq, "%-25s %u\n", ip,
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
1129
static bool is_inaddr_any(u8 *ip, int family)
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
1136
addr = (struct in_addr *)ip;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
1142
addr6 = (struct in6_addr *)ip;
drivers/net/ethernet/chelsio/cxgb4/l2t.c
698
char ip[60];
drivers/net/ethernet/chelsio/cxgb4/l2t.c
704
ip[0] = '\0';
drivers/net/ethernet/chelsio/cxgb4/l2t.c
706
sprintf(ip, e->v6 ? "%pI6c" : "%pI4", e->addr);
drivers/net/ethernet/chelsio/cxgb4/l2t.c
708
e->idx + d->l2t_start, ip, e->dmac,
drivers/net/ethernet/chelsio/inline_crypto/ch_ktls/chcr_ktls.c
1050
ip = (struct iphdr *)(buf + maclen);
drivers/net/ethernet/chelsio/inline_crypto/ch_ktls/chcr_ktls.c
1051
ip->tot_len = htons(pktlen - maclen);
drivers/net/ethernet/chelsio/inline_crypto/ch_ktls/chcr_ktls.c
998
struct iphdr *ip;
drivers/net/ethernet/chelsio/libcxgb/libcxgb_cm.c
52
struct iphdr *ip = (struct iphdr *)((u8 *)(req + 1) + eth_len);
drivers/net/ethernet/chelsio/libcxgb/libcxgb_cm.c
57
if (ip->version == 4) {
drivers/net/ethernet/chelsio/libcxgb/libcxgb_cm.c
59
__func__, ntohl(ip->saddr), ntohl(ip->daddr),
drivers/net/ethernet/chelsio/libcxgb/libcxgb_cm.c
62
memcpy(peer_ip, &ip->saddr, 4);
drivers/net/ethernet/chelsio/libcxgb/libcxgb_cm.c
63
memcpy(local_ip, &ip->daddr, 4);
drivers/net/ethernet/emulex/benet/be_main.c
1129
struct iphdr *ip;
drivers/net/ethernet/emulex/benet/be_main.c
1140
ip = (struct iphdr *)ip_hdr(skb);
drivers/net/ethernet/emulex/benet/be_main.c
1141
if (unlikely(pskb_trim(skb, eth_hdr_len + ntohs(ip->tot_len))))
drivers/net/ethernet/huawei/hinic/hinic_tx.c
199
static void get_inner_l3_l4_type(struct sk_buff *skb, union hinic_l3 *ip,
drivers/net/ethernet/huawei/hinic/hinic_tx.c
207
if (ip->v4->version == 4) {
drivers/net/ethernet/huawei/hinic/hinic_tx.c
211
*l4_proto = ip->v4->protocol;
drivers/net/ethernet/huawei/hinic/hinic_tx.c
212
} else if (ip->v4->version == 6) {
drivers/net/ethernet/huawei/hinic/hinic_tx.c
214
exthdr = ip->hdr + sizeof(*ip->v6);
drivers/net/ethernet/huawei/hinic/hinic_tx.c
215
*l4_proto = ip->v6->nexthdr;
drivers/net/ethernet/huawei/hinic/hinic_tx.c
266
static __sum16 csum_magic(union hinic_l3 *ip, unsigned short proto)
drivers/net/ethernet/huawei/hinic/hinic_tx.c
268
return (ip->v4->version == 4) ?
drivers/net/ethernet/huawei/hinic/hinic_tx.c
269
csum_tcpudp_magic(ip->v4->saddr, ip->v4->daddr, 0, proto, 0) :
drivers/net/ethernet/huawei/hinic/hinic_tx.c
270
csum_ipv6_magic(&ip->v6->saddr, &ip->v6->daddr, 0, proto, 0);
drivers/net/ethernet/huawei/hinic/hinic_tx.c
279
union hinic_l3 ip;
drivers/net/ethernet/huawei/hinic/hinic_tx.c
294
ip.hdr = skb_network_header(skb);
drivers/net/ethernet/huawei/hinic/hinic_tx.c
298
if (ip.v4->version == 4) {
drivers/net/ethernet/huawei/hinic/hinic_tx.c
299
ip.v4->tot_len = 0;
drivers/net/ethernet/huawei/hinic/hinic_tx.c
301
} else if (ip.v4->version == 6) {
drivers/net/ethernet/huawei/hinic/hinic_tx.c
311
l4.udp->check = ~csum_magic(&ip, IPPROTO_UDP);
drivers/net/ethernet/huawei/hinic/hinic_tx.c
321
ip.hdr = skb_inner_network_header(skb);
drivers/net/ethernet/huawei/hinic/hinic_tx.c
324
ip.hdr = skb_network_header(skb);
drivers/net/ethernet/huawei/hinic/hinic_tx.c
330
if (ip.v4->version == 4)
drivers/net/ethernet/huawei/hinic/hinic_tx.c
331
ip.v4->tot_len = 0;
drivers/net/ethernet/huawei/hinic/hinic_tx.c
333
ip.v6->payload_len = 0;
drivers/net/ethernet/huawei/hinic/hinic_tx.c
335
get_inner_l3_l4_type(skb, &ip, &l4, TX_OFFLOAD_TSO, &l3_offload,
drivers/net/ethernet/huawei/hinic/hinic_tx.c
342
l4.tcp->check = ~csum_magic(&ip, IPPROTO_TCP);
drivers/net/ethernet/huawei/hinic/hinic_tx.c
360
union hinic_l3 ip;
drivers/net/ethernet/huawei/hinic/hinic_tx.c
371
ip.hdr = skb_network_header(skb);
drivers/net/ethernet/huawei/hinic/hinic_tx.c
373
if (ip.v4->version == 4) {
drivers/net/ethernet/huawei/hinic/hinic_tx.c
375
l4_proto = ip.v4->protocol;
drivers/net/ethernet/huawei/hinic/hinic_tx.c
376
} else if (ip.v4->version == 6) {
drivers/net/ethernet/huawei/hinic/hinic_tx.c
382
exthdr = ip.hdr + sizeof(*ip.v6);
drivers/net/ethernet/huawei/hinic/hinic_tx.c
383
l4_proto = ip.v6->nexthdr;
drivers/net/ethernet/huawei/hinic/hinic_tx.c
400
ip.hdr = skb_inner_network_header(skb);
drivers/net/ethernet/huawei/hinic/hinic_tx.c
409
ip.hdr = skb_inner_network_header(skb);
drivers/net/ethernet/huawei/hinic/hinic_tx.c
421
ip.hdr = skb_network_header(skb);
drivers/net/ethernet/huawei/hinic/hinic_tx.c
426
get_inner_l3_l4_type(skb, &ip, &l4, TX_OFFLOAD_CSUM, &l3_type,
drivers/net/ethernet/huawei/hinic3/hinic3_tx.c
236
union hinic3_ip ip;
drivers/net/ethernet/huawei/hinic3/hinic3_tx.c
242
ip.hdr = skb_network_header(skb);
drivers/net/ethernet/huawei/hinic3/hinic3_tx.c
243
if (ip.v4->version == 4) {
drivers/net/ethernet/huawei/hinic3/hinic3_tx.c
244
l4_proto = ip.v4->protocol;
drivers/net/ethernet/huawei/hinic3/hinic3_tx.c
245
} else if (ip.v4->version == 6) {
drivers/net/ethernet/huawei/hinic3/hinic3_tx.c
250
exthdr = ip.hdr + sizeof(*ip.v6);
drivers/net/ethernet/huawei/hinic3/hinic3_tx.c
251
l4_proto = ip.v6->nexthdr;
drivers/net/ethernet/huawei/hinic3/hinic3_tx.c
274
static void get_inner_l3_l4_type(struct sk_buff *skb, union hinic3_ip *ip,
drivers/net/ethernet/huawei/hinic3/hinic3_tx.c
281
if (ip->v4->version == 4) {
drivers/net/ethernet/huawei/hinic3/hinic3_tx.c
283
*l4_proto = ip->v4->protocol;
drivers/net/ethernet/huawei/hinic3/hinic3_tx.c
284
} else if (ip->v4->version == 6) {
drivers/net/ethernet/huawei/hinic3/hinic3_tx.c
286
exthdr = ip->hdr + sizeof(*ip->v6);
drivers/net/ethernet/huawei/hinic3/hinic3_tx.c
287
*l4_proto = ip->v6->nexthdr;
drivers/net/ethernet/huawei/hinic3/hinic3_tx.c
322
static __sum16 csum_magic(union hinic3_ip *ip, unsigned short proto)
drivers/net/ethernet/huawei/hinic3/hinic3_tx.c
324
return (ip->v4->version == 4) ?
drivers/net/ethernet/huawei/hinic3/hinic3_tx.c
325
csum_tcpudp_magic(ip->v4->saddr, ip->v4->daddr, 0, proto, 0) :
drivers/net/ethernet/huawei/hinic3/hinic3_tx.c
326
csum_ipv6_magic(&ip->v6->saddr, &ip->v6->daddr, 0, proto, 0);
drivers/net/ethernet/huawei/hinic3/hinic3_tx.c
334
union hinic3_ip ip;
drivers/net/ethernet/huawei/hinic3/hinic3_tx.c
355
ip.hdr = skb_network_header(skb);
drivers/net/ethernet/huawei/hinic3/hinic3_tx.c
358
l4.udp->check = ~csum_magic(&ip, IPPROTO_UDP);
drivers/net/ethernet/huawei/hinic3/hinic3_tx.c
363
ip.hdr = skb_inner_network_header(skb);
drivers/net/ethernet/huawei/hinic3/hinic3_tx.c
366
ip.hdr = skb_network_header(skb);
drivers/net/ethernet/huawei/hinic3/hinic3_tx.c
370
get_inner_l3_l4_type(skb, &ip, &l4, &l3_type, &l4_proto);
drivers/net/ethernet/huawei/hinic3/hinic3_tx.c
373
l4.tcp->check = ~csum_magic(&ip, IPPROTO_TCP);
drivers/net/ethernet/intel/e1000e/netdev.c
5753
const struct iphdr *ip = (struct iphdr *)((u8 *)skb->data + 14);
drivers/net/ethernet/intel/e1000e/netdev.c
5756
if (ip->protocol != IPPROTO_UDP)
drivers/net/ethernet/intel/e1000e/netdev.c
5759
udp = (struct udphdr *)((u8 *)ip + (ip->ihl << 2));
drivers/net/ethernet/intel/i40e/i40e.h
317
} ip;
drivers/net/ethernet/intel/i40e/i40e.h
318
#define dst_ipv6 ip.v6.dst_ip6.s6_addr32
drivers/net/ethernet/intel/i40e/i40e.h
319
#define src_ipv6 ip.v6.src_ip6.s6_addr32
drivers/net/ethernet/intel/i40e/i40e.h
320
#define dst_ipv4 ip.v4.dst_ip.s_addr
drivers/net/ethernet/intel/i40e/i40e.h
321
#define src_ipv4 ip.v4.src_ip.s_addr
drivers/net/ethernet/intel/i40e/i40e_main.c
8480
!ipv6_addr_any(&filter->ip.v6.src_ip6))
drivers/net/ethernet/intel/i40e/i40e_main.c
8509
!ipv6_addr_any(&filter->ip.v6.dst_ip6)) {
drivers/net/ethernet/intel/i40e/i40e_txrx.c
173
struct iphdr ip = {};
drivers/net/ethernet/intel/i40e/i40e_txrx.c
178
ip.protocol = l4proto;
drivers/net/ethernet/intel/i40e/i40e_txrx.c
179
ip.version = 0x4;
drivers/net/ethernet/intel/i40e/i40e_txrx.c
180
ip.ihl = 0x5;
drivers/net/ethernet/intel/i40e/i40e_txrx.c
182
ip.daddr = data->dst_ip;
drivers/net/ethernet/intel/i40e/i40e_txrx.c
183
ip.saddr = data->src_ip;
drivers/net/ethernet/intel/i40e/i40e_txrx.c
211
memcpy(tmp, &ip, sizeof(ip));
drivers/net/ethernet/intel/i40e/i40e_txrx.c
212
tmp += sizeof(ip);
drivers/net/ethernet/intel/i40e/i40e_txrx.c
3085
} ip;
drivers/net/ethernet/intel/i40e/i40e_txrx.c
3108
ip.hdr = skb_inner_network_header(skb);
drivers/net/ethernet/intel/i40e/i40e_txrx.c
3110
ip.hdr = skb_network_header(skb);
drivers/net/ethernet/intel/i40e/i40e_txrx.c
3114
if (ip.v4->version == 4) {
drivers/net/ethernet/intel/i40e/i40e_txrx.c
3115
ip.v4->tot_len = 0;
drivers/net/ethernet/intel/i40e/i40e_txrx.c
3116
ip.v4->check = 0;
drivers/net/ethernet/intel/i40e/i40e_txrx.c
3120
ip.v6->payload_len = 0;
drivers/net/ethernet/intel/i40e/i40e_txrx.c
3144
ip.hdr = skb_inner_network_header(skb);
drivers/net/ethernet/intel/i40e/i40e_txrx.c
3148
if (ip.v4->version == 4) {
drivers/net/ethernet/intel/i40e/i40e_txrx.c
3149
ip.v4->tot_len = 0;
drivers/net/ethernet/intel/i40e/i40e_txrx.c
3150
ip.v4->check = 0;
drivers/net/ethernet/intel/i40e/i40e_txrx.c
3152
ip.v6->payload_len = 0;
drivers/net/ethernet/intel/i40e/i40e_txrx.c
3251
} ip;
drivers/net/ethernet/intel/i40e/i40e_txrx.c
3269
ip.hdr = skb_inner_network_header(skb);
drivers/net/ethernet/intel/i40e/i40e_txrx.c
3272
ip.hdr = skb_network_header(skb);
drivers/net/ethernet/intel/i40e/i40e_txrx.c
3279
if (ip.v4->version == 4)
drivers/net/ethernet/intel/i40e/i40e_txrx.c
3285
offset = ((ip.hdr - skb->data) / 2) << I40E_TX_DESC_LENGTH_MACLEN_SHIFT;
drivers/net/ethernet/intel/i40e/i40e_txrx.c
3295
l4_proto = ip.v4->protocol;
drivers/net/ethernet/intel/i40e/i40e_txrx.c
3301
exthdr = ip.hdr + sizeof(*ip.v6);
drivers/net/ethernet/intel/i40e/i40e_txrx.c
3302
l4_proto = ip.v6->nexthdr;
drivers/net/ethernet/intel/i40e/i40e_txrx.c
3333
tunnel |= ((l4.hdr - ip.hdr) / 4) <<
drivers/net/ethernet/intel/i40e/i40e_txrx.c
3337
ip.hdr = skb_inner_network_header(skb);
drivers/net/ethernet/intel/i40e/i40e_txrx.c
3340
tunnel |= ((ip.hdr - l4.hdr) / 2) <<
drivers/net/ethernet/intel/i40e/i40e_txrx.c
3358
if (ip.v4->version == 4)
drivers/net/ethernet/intel/i40e/i40e_txrx.c
3360
if (ip.v6->version == 6)
drivers/net/ethernet/intel/i40e/i40e_txrx.c
3366
l4_proto = ip.v4->protocol;
drivers/net/ethernet/intel/i40e/i40e_txrx.c
3376
exthdr = ip.hdr + sizeof(*ip.v6);
drivers/net/ethernet/intel/i40e/i40e_txrx.c
3377
l4_proto = ip.v6->nexthdr;
drivers/net/ethernet/intel/i40e/i40e_txrx.c
3384
offset |= ((l4.hdr - ip.hdr) / 4) << I40E_TX_DESC_LENGTH_IPLEN_SHIFT;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
3835
memcpy(&cfilter.ip.v4.dst_ip, tcf.dst_ip,
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
3836
sizeof(cfilter.ip.v4.dst_ip));
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
3838
memcpy(&cfilter.ip.v4.src_ip, tcf.src_ip,
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
3839
sizeof(cfilter.ip.v4.src_ip));
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
3844
memcpy(&cfilter.ip.v6.dst_ip6, tcf.dst_ip,
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
3845
sizeof(cfilter.ip.v6.dst_ip6));
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
3847
memcpy(&cfilter.ip.v6.src_ip6, tcf.src_ip,
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
3848
sizeof(cfilter.ip.v6.src_ip6));
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
3888
if (memcmp(&cfilter.ip.v4.dst_ip, &cf->ip.v4.dst_ip,
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
3893
if (memcmp(&cfilter.ip.v6.dst_ip6, &cf->ip.v6.dst_ip6,
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
3894
sizeof(cfilter.ip.v6.dst_ip6)))
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
3981
memcpy(&cfilter->ip.v4.dst_ip, tcf.dst_ip,
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
3982
sizeof(cfilter->ip.v4.dst_ip));
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
3984
memcpy(&cfilter->ip.v4.src_ip, tcf.src_ip,
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
3985
sizeof(cfilter->ip.v4.src_ip));
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
3990
memcpy(&cfilter->ip.v6.dst_ip6, tcf.dst_ip,
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
3991
sizeof(cfilter->ip.v6.dst_ip6));
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
3993
memcpy(&cfilter->ip.v6.src_ip6, tcf.src_ip,
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
3994
sizeof(cfilter->ip.v6.src_ip6));
drivers/net/ethernet/intel/iavf/iavf_txrx.c
1738
} ip;
drivers/net/ethernet/intel/iavf/iavf_txrx.c
1758
ip.hdr = skb_network_header(skb);
drivers/net/ethernet/intel/iavf/iavf_txrx.c
1762
if (ip.v4->version == 4) {
drivers/net/ethernet/intel/iavf/iavf_txrx.c
1763
ip.v4->tot_len = 0;
drivers/net/ethernet/intel/iavf/iavf_txrx.c
1764
ip.v4->check = 0;
drivers/net/ethernet/intel/iavf/iavf_txrx.c
1766
ip.v6->payload_len = 0;
drivers/net/ethernet/intel/iavf/iavf_txrx.c
1789
ip.hdr = skb_inner_network_header(skb);
drivers/net/ethernet/intel/iavf/iavf_txrx.c
1793
if (ip.v4->version == 4) {
drivers/net/ethernet/intel/iavf/iavf_txrx.c
1794
ip.v4->tot_len = 0;
drivers/net/ethernet/intel/iavf/iavf_txrx.c
1795
ip.v4->check = 0;
drivers/net/ethernet/intel/iavf/iavf_txrx.c
1797
ip.v6->payload_len = 0;
drivers/net/ethernet/intel/iavf/iavf_txrx.c
1854
} ip;
drivers/net/ethernet/intel/iavf/iavf_txrx.c
1868
ip.hdr = skb_network_header(skb);
drivers/net/ethernet/intel/iavf/iavf_txrx.c
1872
offset = ((ip.hdr - skb->data) / 2) << IAVF_TX_DESC_LENGTH_MACLEN_SHIFT;
drivers/net/ethernet/intel/iavf/iavf_txrx.c
1882
l4_proto = ip.v4->protocol;
drivers/net/ethernet/intel/iavf/iavf_txrx.c
1886
exthdr = ip.hdr + sizeof(*ip.v6);
drivers/net/ethernet/intel/iavf/iavf_txrx.c
1887
l4_proto = ip.v6->nexthdr;
drivers/net/ethernet/intel/iavf/iavf_txrx.c
1917
tunnel |= ((l4.hdr - ip.hdr) / 4) <<
drivers/net/ethernet/intel/iavf/iavf_txrx.c
1921
ip.hdr = skb_inner_network_header(skb);
drivers/net/ethernet/intel/iavf/iavf_txrx.c
1924
tunnel |= ((ip.hdr - l4.hdr) / 2) <<
drivers/net/ethernet/intel/iavf/iavf_txrx.c
1942
if (ip.v4->version == 4)
drivers/net/ethernet/intel/iavf/iavf_txrx.c
1944
if (ip.v6->version == 6)
drivers/net/ethernet/intel/iavf/iavf_txrx.c
1950
l4_proto = ip.v4->protocol;
drivers/net/ethernet/intel/iavf/iavf_txrx.c
1960
exthdr = ip.hdr + sizeof(*ip.v6);
drivers/net/ethernet/intel/iavf/iavf_txrx.c
1961
l4_proto = ip.v6->nexthdr;
drivers/net/ethernet/intel/iavf/iavf_txrx.c
1968
offset |= ((l4.hdr - ip.hdr) / 4) << IAVF_TX_DESC_LENGTH_IPLEN_SHIFT;
drivers/net/ethernet/intel/ice/ice_arfs.c
321
fltr_info->ip.v4.proto = ip_proto;
drivers/net/ethernet/intel/ice/ice_arfs.c
325
fltr_info->ip.v4.src_ip = fk->addrs.v4addrs.src;
drivers/net/ethernet/intel/ice/ice_arfs.c
326
fltr_info->ip.v4.dst_ip = fk->addrs.v4addrs.dst;
drivers/net/ethernet/intel/ice/ice_arfs.c
327
fltr_info->ip.v4.src_port = fk->ports.src;
drivers/net/ethernet/intel/ice/ice_arfs.c
328
fltr_info->ip.v4.dst_port = fk->ports.dst;
drivers/net/ethernet/intel/ice/ice_arfs.c
330
fltr_info->ip.v6.proto = ip_proto;
drivers/net/ethernet/intel/ice/ice_arfs.c
334
memcpy(&fltr_info->ip.v6.src_ip, &fk->addrs.v6addrs.src,
drivers/net/ethernet/intel/ice/ice_arfs.c
336
memcpy(&fltr_info->ip.v6.dst_ip, &fk->addrs.v6addrs.dst,
drivers/net/ethernet/intel/ice/ice_arfs.c
338
fltr_info->ip.v6.src_port = fk->ports.src;
drivers/net/ethernet/intel/ice/ice_arfs.c
339
fltr_info->ip.v6.dst_port = fk->ports.dst;
drivers/net/ethernet/intel/ice/ice_arfs.c
409
fltr_info->ip.v4.src_port == fk->ports.src &&
drivers/net/ethernet/intel/ice/ice_arfs.c
410
fltr_info->ip.v4.dst_port == fk->ports.dst &&
drivers/net/ethernet/intel/ice/ice_arfs.c
411
fltr_info->ip.v4.src_ip == fk->addrs.v4addrs.src &&
drivers/net/ethernet/intel/ice/ice_arfs.c
412
fltr_info->ip.v4.dst_ip == fk->addrs.v4addrs.dst &&
drivers/net/ethernet/intel/ice/ice_arfs.c
413
fltr_info->ip.v4.proto == ip_proto;
drivers/net/ethernet/intel/ice/ice_arfs.c
415
return fltr_info->ip.v6.src_port == fk->ports.src &&
drivers/net/ethernet/intel/ice/ice_arfs.c
416
fltr_info->ip.v6.dst_port == fk->ports.dst &&
drivers/net/ethernet/intel/ice/ice_arfs.c
417
fltr_info->ip.v6.proto == ip_proto &&
drivers/net/ethernet/intel/ice/ice_arfs.c
418
!memcmp(&fltr_info->ip.v6.src_ip, &fk->addrs.v6addrs.src,
drivers/net/ethernet/intel/ice/ice_arfs.c
420
!memcmp(&fltr_info->ip.v6.dst_ip, &fk->addrs.v6addrs.dst,
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
151
fsp->h_u.usr_ip4_spec.l4_4_bytes = rule->ip.v4.l4_header;
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
152
fsp->h_u.usr_ip4_spec.tos = rule->ip.v4.tos;
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
153
fsp->h_u.usr_ip4_spec.ip4src = rule->ip.v4.src_ip;
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
154
fsp->h_u.usr_ip4_spec.ip4dst = rule->ip.v4.dst_ip;
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
165
fsp->h_u.tcp_ip4_spec.psrc = rule->ip.v4.src_port;
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
166
fsp->h_u.tcp_ip4_spec.pdst = rule->ip.v4.dst_port;
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
167
fsp->h_u.tcp_ip4_spec.ip4src = rule->ip.v4.src_ip;
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
168
fsp->h_u.tcp_ip4_spec.ip4dst = rule->ip.v4.dst_ip;
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
175
fsp->h_u.usr_ip6_spec.l4_4_bytes = rule->ip.v6.l4_header;
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
176
fsp->h_u.usr_ip6_spec.tclass = rule->ip.v6.tc;
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
177
fsp->h_u.usr_ip6_spec.l4_proto = rule->ip.v6.proto;
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
178
memcpy(fsp->h_u.tcp_ip6_spec.ip6src, rule->ip.v6.src_ip,
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
180
memcpy(fsp->h_u.tcp_ip6_spec.ip6dst, rule->ip.v6.dst_ip,
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1895
input->ip.v4.dst_port = fsp->h_u.tcp_ip4_spec.pdst;
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1896
input->ip.v4.src_port = fsp->h_u.tcp_ip4_spec.psrc;
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1897
input->ip.v4.dst_ip = fsp->h_u.tcp_ip4_spec.ip4dst;
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1898
input->ip.v4.src_ip = fsp->h_u.tcp_ip4_spec.ip4src;
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1905
input->ip.v4.dst_ip = fsp->h_u.usr_ip4_spec.ip4dst;
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1906
input->ip.v4.src_ip = fsp->h_u.usr_ip4_spec.ip4src;
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1907
input->ip.v4.l4_header = fsp->h_u.usr_ip4_spec.l4_4_bytes;
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1908
input->ip.v4.proto = fsp->h_u.usr_ip4_spec.proto;
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1909
input->ip.v4.ip_ver = fsp->h_u.usr_ip4_spec.ip_ver;
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1910
input->ip.v4.tos = fsp->h_u.usr_ip4_spec.tos;
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1921
memcpy(input->ip.v6.dst_ip, fsp->h_u.usr_ip6_spec.ip6dst,
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1923
memcpy(input->ip.v6.src_ip, fsp->h_u.usr_ip6_spec.ip6src,
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1925
input->ip.v6.dst_port = fsp->h_u.tcp_ip6_spec.pdst;
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1926
input->ip.v6.src_port = fsp->h_u.tcp_ip6_spec.psrc;
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1927
input->ip.v6.tc = fsp->h_u.tcp_ip6_spec.tclass;
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
193
memcpy(fsp->h_u.tcp_ip6_spec.ip6src, rule->ip.v6.src_ip,
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1937
memcpy(input->ip.v6.dst_ip, fsp->h_u.usr_ip6_spec.ip6dst,
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1939
memcpy(input->ip.v6.src_ip, fsp->h_u.usr_ip6_spec.ip6src,
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1941
input->ip.v6.l4_header = fsp->h_u.usr_ip6_spec.l4_4_bytes;
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1942
input->ip.v6.tc = fsp->h_u.usr_ip6_spec.tclass;
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1946
input->ip.v6.proto = IPPROTO_NONE;
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1948
input->ip.v6.proto = fsp->h_u.usr_ip6_spec.l4_proto;
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
195
memcpy(fsp->h_u.tcp_ip6_spec.ip6dst, rule->ip.v6.dst_ip,
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
197
fsp->h_u.tcp_ip6_spec.psrc = rule->ip.v6.src_port;
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
198
fsp->h_u.tcp_ip6_spec.pdst = rule->ip.v6.dst_port;
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
207
fsp->h_u.tcp_ip6_spec.tclass = rule->ip.v6.tc;
drivers/net/ethernet/intel/ice/ice_fdir.c
1000
input->ip.v4.dst_ip);
drivers/net/ethernet/intel/ice/ice_fdir.c
1016
input->ip.v4.sec_parm_idx);
drivers/net/ethernet/intel/ice/ice_fdir.c
1020
input->ip.v6.sec_parm_idx);
drivers/net/ethernet/intel/ice/ice_fdir.c
1024
input->ip.v4.sec_parm_idx);
drivers/net/ethernet/intel/ice/ice_fdir.c
1028
input->ip.v6.sec_parm_idx);
drivers/net/ethernet/intel/ice/ice_fdir.c
1032
input->ip.v4.src_ip);
drivers/net/ethernet/intel/ice/ice_fdir.c
1034
input->ip.v4.dst_ip);
drivers/net/ethernet/intel/ice/ice_fdir.c
1036
input->ip.v4.sec_parm_idx);
drivers/net/ethernet/intel/ice/ice_fdir.c
1040
input->ip.v6.src_ip);
drivers/net/ethernet/intel/ice/ice_fdir.c
1042
input->ip.v6.dst_ip);
drivers/net/ethernet/intel/ice/ice_fdir.c
1044
input->ip.v6.sec_parm_idx);
drivers/net/ethernet/intel/ice/ice_fdir.c
1049
input->ip.v4.dst_port);
drivers/net/ethernet/intel/ice/ice_fdir.c
1054
input->ip.v6.dst_port);
drivers/net/ethernet/intel/ice/ice_fdir.c
1062
input->ip.v6.src_ip);
drivers/net/ethernet/intel/ice/ice_fdir.c
1064
input->ip.v6.dst_ip);
drivers/net/ethernet/intel/ice/ice_fdir.c
1066
input->ip.v6.src_port);
drivers/net/ethernet/intel/ice/ice_fdir.c
1068
input->ip.v6.dst_port);
drivers/net/ethernet/intel/ice/ice_fdir.c
1069
ice_pkt_insert_u8_tc(loc, ICE_IPV6_TC_OFFSET, input->ip.v6.tc);
drivers/net/ethernet/intel/ice/ice_fdir.c
1070
ice_pkt_insert_u8(loc, ICE_IPV6_HLIM_OFFSET, input->ip.v6.hlim);
drivers/net/ethernet/intel/ice/ice_fdir.c
1075
input->ip.v6.src_ip);
drivers/net/ethernet/intel/ice/ice_fdir.c
1077
input->ip.v6.dst_ip);
drivers/net/ethernet/intel/ice/ice_fdir.c
1079
input->ip.v6.src_port);
drivers/net/ethernet/intel/ice/ice_fdir.c
1081
input->ip.v6.dst_port);
drivers/net/ethernet/intel/ice/ice_fdir.c
1082
ice_pkt_insert_u8_tc(loc, ICE_IPV6_TC_OFFSET, input->ip.v6.tc);
drivers/net/ethernet/intel/ice/ice_fdir.c
1083
ice_pkt_insert_u8(loc, ICE_IPV6_HLIM_OFFSET, input->ip.v6.hlim);
drivers/net/ethernet/intel/ice/ice_fdir.c
1088
input->ip.v6.src_ip);
drivers/net/ethernet/intel/ice/ice_fdir.c
1090
input->ip.v6.dst_ip);
drivers/net/ethernet/intel/ice/ice_fdir.c
1092
input->ip.v6.src_port);
drivers/net/ethernet/intel/ice/ice_fdir.c
1094
input->ip.v6.dst_port);
drivers/net/ethernet/intel/ice/ice_fdir.c
1095
ice_pkt_insert_u8_tc(loc, ICE_IPV6_TC_OFFSET, input->ip.v6.tc);
drivers/net/ethernet/intel/ice/ice_fdir.c
1096
ice_pkt_insert_u8(loc, ICE_IPV6_HLIM_OFFSET, input->ip.v6.hlim);
drivers/net/ethernet/intel/ice/ice_fdir.c
1101
input->ip.v6.src_ip);
drivers/net/ethernet/intel/ice/ice_fdir.c
1103
input->ip.v6.dst_ip);
drivers/net/ethernet/intel/ice/ice_fdir.c
1104
ice_pkt_insert_u8_tc(loc, ICE_IPV6_TC_OFFSET, input->ip.v6.tc);
drivers/net/ethernet/intel/ice/ice_fdir.c
1105
ice_pkt_insert_u8(loc, ICE_IPV6_HLIM_OFFSET, input->ip.v6.hlim);
drivers/net/ethernet/intel/ice/ice_fdir.c
1107
input->ip.v6.proto);
drivers/net/ethernet/intel/ice/ice_fdir.c
1233
if (a->ip.v4.dst_ip == b->ip.v4.dst_ip &&
drivers/net/ethernet/intel/ice/ice_fdir.c
1234
a->ip.v4.src_ip == b->ip.v4.src_ip &&
drivers/net/ethernet/intel/ice/ice_fdir.c
1235
a->ip.v4.dst_port == b->ip.v4.dst_port &&
drivers/net/ethernet/intel/ice/ice_fdir.c
1236
a->ip.v4.src_port == b->ip.v4.src_port)
drivers/net/ethernet/intel/ice/ice_fdir.c
1240
if (a->ip.v4.dst_ip == b->ip.v4.dst_ip &&
drivers/net/ethernet/intel/ice/ice_fdir.c
1241
a->ip.v4.src_ip == b->ip.v4.src_ip &&
drivers/net/ethernet/intel/ice/ice_fdir.c
1242
a->ip.v4.l4_header == b->ip.v4.l4_header &&
drivers/net/ethernet/intel/ice/ice_fdir.c
1243
a->ip.v4.proto == b->ip.v4.proto &&
drivers/net/ethernet/intel/ice/ice_fdir.c
1244
a->ip.v4.ip_ver == b->ip.v4.ip_ver &&
drivers/net/ethernet/intel/ice/ice_fdir.c
1245
a->ip.v4.tos == b->ip.v4.tos)
drivers/net/ethernet/intel/ice/ice_fdir.c
1251
if (a->ip.v6.dst_port == b->ip.v6.dst_port &&
drivers/net/ethernet/intel/ice/ice_fdir.c
1252
a->ip.v6.src_port == b->ip.v6.src_port &&
drivers/net/ethernet/intel/ice/ice_fdir.c
1253
!ice_cmp_ipv6_addr(a->ip.v6.dst_ip,
drivers/net/ethernet/intel/ice/ice_fdir.c
1254
b->ip.v6.dst_ip) &&
drivers/net/ethernet/intel/ice/ice_fdir.c
1255
!ice_cmp_ipv6_addr(a->ip.v6.src_ip,
drivers/net/ethernet/intel/ice/ice_fdir.c
1256
b->ip.v6.src_ip))
drivers/net/ethernet/intel/ice/ice_fdir.c
1260
if (a->ip.v6.dst_port == b->ip.v6.dst_port &&
drivers/net/ethernet/intel/ice/ice_fdir.c
1261
a->ip.v6.src_port == b->ip.v6.src_port)
drivers/net/ethernet/intel/ice/ice_fdir.c
867
switch (input->ip.v4.proto) {
drivers/net/ethernet/intel/ice/ice_fdir.c
882
switch (input->ip.v6.proto) {
drivers/net/ethernet/intel/ice/ice_fdir.c
941
input->ip.v4.src_ip);
drivers/net/ethernet/intel/ice/ice_fdir.c
943
input->ip.v4.src_port);
drivers/net/ethernet/intel/ice/ice_fdir.c
945
input->ip.v4.dst_ip);
drivers/net/ethernet/intel/ice/ice_fdir.c
947
input->ip.v4.dst_port);
drivers/net/ethernet/intel/ice/ice_fdir.c
948
ice_pkt_insert_u8(loc, ICE_IPV4_TOS_OFFSET, input->ip.v4.tos);
drivers/net/ethernet/intel/ice/ice_fdir.c
949
ice_pkt_insert_u8(loc, ICE_IPV4_TTL_OFFSET, input->ip.v4.ttl);
drivers/net/ethernet/intel/ice/ice_fdir.c
956
input->ip.v4.src_ip);
drivers/net/ethernet/intel/ice/ice_fdir.c
958
input->ip.v4.src_port);
drivers/net/ethernet/intel/ice/ice_fdir.c
960
input->ip.v4.dst_ip);
drivers/net/ethernet/intel/ice/ice_fdir.c
962
input->ip.v4.dst_port);
drivers/net/ethernet/intel/ice/ice_fdir.c
963
ice_pkt_insert_u8(loc, ICE_IPV4_TOS_OFFSET, input->ip.v4.tos);
drivers/net/ethernet/intel/ice/ice_fdir.c
964
ice_pkt_insert_u8(loc, ICE_IPV4_TTL_OFFSET, input->ip.v4.ttl);
drivers/net/ethernet/intel/ice/ice_fdir.c
971
input->ip.v4.src_ip);
drivers/net/ethernet/intel/ice/ice_fdir.c
973
input->ip.v4.src_port);
drivers/net/ethernet/intel/ice/ice_fdir.c
975
input->ip.v4.dst_ip);
drivers/net/ethernet/intel/ice/ice_fdir.c
977
input->ip.v4.dst_port);
drivers/net/ethernet/intel/ice/ice_fdir.c
978
ice_pkt_insert_u8(loc, ICE_IPV4_TOS_OFFSET, input->ip.v4.tos);
drivers/net/ethernet/intel/ice/ice_fdir.c
979
ice_pkt_insert_u8(loc, ICE_IPV4_TTL_OFFSET, input->ip.v4.ttl);
drivers/net/ethernet/intel/ice/ice_fdir.c
984
input->ip.v4.src_ip);
drivers/net/ethernet/intel/ice/ice_fdir.c
986
input->ip.v4.dst_ip);
drivers/net/ethernet/intel/ice/ice_fdir.c
987
ice_pkt_insert_u8(loc, ICE_IPV4_TOS_OFFSET, input->ip.v4.tos);
drivers/net/ethernet/intel/ice/ice_fdir.c
988
ice_pkt_insert_u8(loc, ICE_IPV4_TTL_OFFSET, input->ip.v4.ttl);
drivers/net/ethernet/intel/ice/ice_fdir.c
990
input->ip.v4.proto);
drivers/net/ethernet/intel/ice/ice_fdir.c
998
input->ip.v4.src_ip);
drivers/net/ethernet/intel/ice/ice_fdir.h
172
} ip, mask;
drivers/net/ethernet/intel/ice/ice_tc_lib.h
100
#define dst_ipv6_addr ip.v6.dst_ip6.s6_addr
drivers/net/ethernet/intel/ice/ice_tc_lib.h
101
#define src_ipv6 ip.v6.src_ip6.s6_addr32
drivers/net/ethernet/intel/ice/ice_tc_lib.h
102
#define src_ipv6_addr ip.v6.src_ip6.s6_addr
drivers/net/ethernet/intel/ice/ice_tc_lib.h
103
#define dst_ipv4 ip.v4.dst_ip.s_addr
drivers/net/ethernet/intel/ice/ice_tc_lib.h
104
#define src_ipv4 ip.v4.src_ip.s_addr
drivers/net/ethernet/intel/ice/ice_tc_lib.h
98
} ip;
drivers/net/ethernet/intel/ice/ice_tc_lib.h
99
#define dst_ipv6 ip.v6.dst_ip6.s6_addr32
drivers/net/ethernet/intel/ice/ice_txrx.c
1594
} ip;
drivers/net/ethernet/intel/ice/ice_txrx.c
1610
ip.hdr = skb_inner_network_header(skb);
drivers/net/ethernet/intel/ice/ice_txrx.c
1613
ip.hdr = skb_network_header(skb);
drivers/net/ethernet/intel/ice/ice_txrx.c
1618
l2_len = ip.hdr - skb->data;
drivers/net/ethernet/intel/ice/ice_txrx.c
1624
if (ip.v4->version == 4)
drivers/net/ethernet/intel/ice/ice_txrx.c
1626
else if (ip.v6->version == 6)
drivers/net/ethernet/intel/ice/ice_txrx.c
1638
l4_proto = ip.v4->protocol;
drivers/net/ethernet/intel/ice/ice_txrx.c
1643
exthdr = ip.hdr + sizeof(*ip.v6);
drivers/net/ethernet/intel/ice/ice_txrx.c
1644
l4_proto = ip.v6->nexthdr;
drivers/net/ethernet/intel/ice/ice_txrx.c
1675
tunnel |= ((l4.hdr - ip.hdr) / 4) <<
drivers/net/ethernet/intel/ice/ice_txrx.c
1679
ip.hdr = skb_inner_network_header(skb);
drivers/net/ethernet/intel/ice/ice_txrx.c
1682
tunnel |= ((ip.hdr - l4.hdr) / 2) <<
drivers/net/ethernet/intel/ice/ice_txrx.c
1705
if (ip.v4->version == 4)
drivers/net/ethernet/intel/ice/ice_txrx.c
1707
if (ip.v6->version == 6)
drivers/net/ethernet/intel/ice/ice_txrx.c
1713
l4_proto = ip.v4->protocol;
drivers/net/ethernet/intel/ice/ice_txrx.c
1724
exthdr = ip.hdr + sizeof(*ip.v6);
drivers/net/ethernet/intel/ice/ice_txrx.c
1725
l4_proto = ip.v6->nexthdr;
drivers/net/ethernet/intel/ice/ice_txrx.c
1734
l3_len = l4.hdr - ip.hdr;
drivers/net/ethernet/intel/ice/ice_txrx.c
1841
} ip;
drivers/net/ethernet/intel/ice/ice_txrx.c
1866
ip.hdr = skb_inner_network_header(skb);
drivers/net/ethernet/intel/ice/ice_txrx.c
1868
ip.hdr = skb_network_header(skb);
drivers/net/ethernet/intel/ice/ice_txrx.c
1872
if (ip.v4->version == 4) {
drivers/net/ethernet/intel/ice/ice_txrx.c
1873
ip.v4->tot_len = 0;
drivers/net/ethernet/intel/ice/ice_txrx.c
1874
ip.v4->check = 0;
drivers/net/ethernet/intel/ice/ice_txrx.c
1876
ip.v6->payload_len = 0;
drivers/net/ethernet/intel/ice/ice_txrx.c
1899
ip.hdr = skb_inner_network_header(skb);
drivers/net/ethernet/intel/ice/ice_txrx.c
1903
if (ip.v4->version == 4) {
drivers/net/ethernet/intel/ice/ice_txrx.c
1904
ip.v4->tot_len = 0;
drivers/net/ethernet/intel/ice/ice_txrx.c
1905
ip.v4->check = 0;
drivers/net/ethernet/intel/ice/ice_txrx.c
1907
ip.v6->payload_len = 0;
drivers/net/ethernet/intel/ice/virt/fdir.c
1000
input->ip.v4.src_port = tcph->source;
drivers/net/ethernet/intel/ice/virt/fdir.c
1001
input->ip.v4.dst_port = tcph->dest;
drivers/net/ethernet/intel/ice/virt/fdir.c
1003
input->ip.v6.src_port = tcph->source;
drivers/net/ethernet/intel/ice/virt/fdir.c
1004
input->ip.v6.dst_port = tcph->dest;
drivers/net/ethernet/intel/ice/virt/fdir.c
1017
input->ip.v4.src_port = udph->source;
drivers/net/ethernet/intel/ice/virt/fdir.c
1018
input->ip.v4.dst_port = udph->dest;
drivers/net/ethernet/intel/ice/virt/fdir.c
1020
input->ip.v6.src_port = udph->source;
drivers/net/ethernet/intel/ice/virt/fdir.c
1021
input->ip.v6.dst_port = udph->dest;
drivers/net/ethernet/intel/ice/virt/fdir.c
1036
input->ip.v4.src_port = sctph->source;
drivers/net/ethernet/intel/ice/virt/fdir.c
1037
input->ip.v4.dst_port = sctph->dest;
drivers/net/ethernet/intel/ice/virt/fdir.c
1039
input->ip.v6.src_port = sctph->source;
drivers/net/ethernet/intel/ice/virt/fdir.c
1040
input->ip.v6.dst_port = sctph->dest;
drivers/net/ethernet/intel/ice/virt/fdir.c
1075
input->ip.v4.sec_parm_idx = esph->spi;
drivers/net/ethernet/intel/ice/virt/fdir.c
1077
input->ip.v6.sec_parm_idx = esph->spi;
drivers/net/ethernet/intel/ice/virt/fdir.c
1089
input->ip.v4.sec_parm_idx = ah->spi;
drivers/net/ethernet/intel/ice/virt/fdir.c
1091
input->ip.v6.sec_parm_idx = ah->spi;
drivers/net/ethernet/intel/ice/virt/fdir.c
1108
input->ip.v4.dst_port = cpu_to_be16(PFCP_PORT_NR);
drivers/net/ethernet/intel/ice/virt/fdir.c
1110
input->ip.v6.dst_port = cpu_to_be16(PFCP_PORT_NR);
drivers/net/ethernet/intel/ice/virt/fdir.c
1259
if (memcmp(&a->ip, &b->ip, sizeof(a->ip)))
drivers/net/ethernet/intel/ice/virt/fdir.c
968
input->ip.v4.src_ip = iph->saddr;
drivers/net/ethernet/intel/ice/virt/fdir.c
969
input->ip.v4.dst_ip = iph->daddr;
drivers/net/ethernet/intel/ice/virt/fdir.c
970
input->ip.v4.tos = iph->tos;
drivers/net/ethernet/intel/ice/virt/fdir.c
971
input->ip.v4.proto = iph->protocol;
drivers/net/ethernet/intel/ice/virt/fdir.c
980
memcpy(input->ip.v6.src_ip,
drivers/net/ethernet/intel/ice/virt/fdir.c
983
memcpy(input->ip.v6.dst_ip,
drivers/net/ethernet/intel/ice/virt/fdir.c
986
input->ip.v6.tc = ((u8)(ip6h->priority) << 4) |
drivers/net/ethernet/intel/ice/virt/fdir.c
988
input->ip.v6.proto = ip6h->nexthdr;
drivers/net/ethernet/intel/idpf/idpf_singleq_txrx.c
116
if (ip.v4->version == 4)
drivers/net/ethernet/intel/idpf/idpf_singleq_txrx.c
118
if (ip.v6->version == 6)
drivers/net/ethernet/intel/idpf/idpf_singleq_txrx.c
124
l4_proto = ip.v4->protocol;
drivers/net/ethernet/intel/idpf/idpf_singleq_txrx.c
135
l4_proto = ip.v6->nexthdr;
drivers/net/ethernet/intel/idpf/idpf_singleq_txrx.c
138
sizeof(*ip.v6), &l4_proto,
drivers/net/ethernet/intel/idpf/idpf_singleq_txrx.c
145
l3_len = l4.hdr - ip.hdr;
drivers/net/ethernet/intel/idpf/idpf_singleq_txrx.c
24
} ip;
drivers/net/ethernet/intel/idpf/idpf_singleq_txrx.c
37
ip.hdr = skb_network_header(skb);
drivers/net/ethernet/intel/idpf/idpf_singleq_txrx.c
41
l2_len = ip.hdr - skb->data;
drivers/net/ethernet/intel/idpf/idpf_singleq_txrx.c
57
l4_proto = ip.v4->protocol;
drivers/net/ethernet/intel/idpf/idpf_singleq_txrx.c
61
l4_proto = ip.v6->nexthdr;
drivers/net/ethernet/intel/idpf/idpf_singleq_txrx.c
64
sizeof(*ip.v6),
drivers/net/ethernet/intel/idpf/idpf_singleq_txrx.c
92
(l4.hdr - ip.hdr) / 4);
drivers/net/ethernet/intel/idpf/idpf_singleq_txrx.c
95
ip.hdr = skb_inner_network_header(skb);
drivers/net/ethernet/intel/idpf/idpf_singleq_txrx.c
99
(ip.hdr - l4.hdr) / 2);
drivers/net/ethernet/intel/idpf/idpf_txrx.c
2828
} ip;
drivers/net/ethernet/intel/idpf/idpf_txrx.c
2846
ip.hdr = skb_network_header(skb);
drivers/net/ethernet/intel/idpf/idpf_txrx.c
2850
if (ip.v4->version == 4) {
drivers/net/ethernet/intel/idpf/idpf_txrx.c
2851
ip.v4->tot_len = 0;
drivers/net/ethernet/intel/idpf/idpf_txrx.c
2852
ip.v4->check = 0;
drivers/net/ethernet/intel/idpf/idpf_txrx.c
2853
} else if (ip.v6->version == 6) {
drivers/net/ethernet/intel/idpf/idpf_txrx.c
2854
ip.v6->payload_len = 0;
drivers/net/ethernet/intel/igb/igb_main.c
6058
} ip;
drivers/net/ethernet/intel/igb/igb_main.c
6077
ip.hdr = skb_network_header(skb);
drivers/net/ethernet/intel/igb/igb_main.c
6085
if (ip.v4->version == 4) {
drivers/net/ethernet/intel/igb/igb_main.c
6087
unsigned char *trans_start = ip.hdr + (ip.v4->ihl * 4);
drivers/net/ethernet/intel/igb/igb_main.c
6092
ip.v4->check = csum_fold(csum_partial(trans_start,
drivers/net/ethernet/intel/igb/igb_main.c
6097
ip.v4->tot_len = 0;
drivers/net/ethernet/intel/igb/igb_main.c
6102
ip.v6->payload_len = 0;
drivers/net/ethernet/intel/igb/igb_main.c
6133
vlan_macip_lens = l4.hdr - ip.hdr;
drivers/net/ethernet/intel/igb/igb_main.c
6134
vlan_macip_lens |= (ip.hdr - skb->data) << E1000_ADVTXD_MACLEN_SHIFT;
drivers/net/ethernet/intel/igbvf/netdev.c
2000
} ip;
drivers/net/ethernet/intel/igbvf/netdev.c
2018
ip.hdr = skb_network_header(skb);
drivers/net/ethernet/intel/igbvf/netdev.c
2025
if (ip.v4->version == 4) {
drivers/net/ethernet/intel/igbvf/netdev.c
2027
unsigned char *trans_start = ip.hdr + (ip.v4->ihl * 4);
drivers/net/ethernet/intel/igbvf/netdev.c
2032
ip.v4->check = csum_fold(csum_partial(trans_start,
drivers/net/ethernet/intel/igbvf/netdev.c
2037
ip.v4->tot_len = 0;
drivers/net/ethernet/intel/igbvf/netdev.c
2039
ip.v6->payload_len = 0;
drivers/net/ethernet/intel/igbvf/netdev.c
2057
vlan_macip_lens = l4.hdr - ip.hdr;
drivers/net/ethernet/intel/igbvf/netdev.c
2058
vlan_macip_lens |= (ip.hdr - skb->data) << E1000_ADVTXD_MACLEN_SHIFT;
drivers/net/ethernet/intel/igc/igc_main.c
1462
} ip;
drivers/net/ethernet/intel/igc/igc_main.c
1481
ip.hdr = skb_network_header(skb);
drivers/net/ethernet/intel/igc/igc_main.c
1488
if (ip.v4->version == 4) {
drivers/net/ethernet/intel/igc/igc_main.c
1490
unsigned char *trans_start = ip.hdr + (ip.v4->ihl * 4);
drivers/net/ethernet/intel/igc/igc_main.c
1495
ip.v4->check = csum_fold(csum_partial(trans_start,
drivers/net/ethernet/intel/igc/igc_main.c
1500
ip.v4->tot_len = 0;
drivers/net/ethernet/intel/igc/igc_main.c
1505
ip.v6->payload_len = 0;
drivers/net/ethernet/intel/igc/igc_main.c
1536
vlan_macip_lens = l4.hdr - ip.hdr;
drivers/net/ethernet/intel/igc/igc_main.c
1537
vlan_macip_lens |= (ip.hdr - skb->data) << IGC_ADVTXD_MACLEN_SHIFT;
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
8710
} ip;
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
8731
ip.hdr = skb_inner_network_header(skb);
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
8733
ip.hdr = skb_network_header(skb);
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
8741
if (ip.v4->version == 4) {
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
8743
unsigned char *trans_start = ip.hdr + (ip.v4->ihl * 4);
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
8750
ip.v4->check = (skb_shinfo(skb)->gso_type & SKB_GSO_PARTIAL) ?
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
8755
ip.v4->tot_len = 0;
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
8760
ip.v6->payload_len = 0;
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
8795
vlan_macip_lens = l4.hdr - ip.hdr;
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
8796
vlan_macip_lens |= (ip.hdr - skb->data) << IXGBE_ADVTXD_MACLEN_SHIFT;
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
9210
common.ip ^= hdr.ipv4->saddr ^ hdr.ipv4->daddr;
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
9214
common.ip ^= hdr.ipv6->saddr.s6_addr32[0] ^
drivers/net/ethernet/intel/ixgbe/ixgbe_type.h
3167
__be32 ip;
drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c
3812
} ip;
drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c
3832
ip.hdr = skb_inner_network_header(skb);
drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c
3834
ip.hdr = skb_network_header(skb);
drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c
3841
if (ip.v4->version == 4) {
drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c
3843
unsigned char *trans_start = ip.hdr + (ip.v4->ihl * 4);
drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c
3850
ip.v4->check = (skb_shinfo(skb)->gso_type & SKB_GSO_PARTIAL) ?
drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c
3855
ip.v4->tot_len = 0;
drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c
3860
ip.v6->payload_len = 0;
drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c
3888
vlan_macip_lens = l4.hdr - ip.hdr;
drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c
3889
vlan_macip_lens |= (ip.hdr - skb->data) << IXGBE_ADVTXD_MACLEN_SHIFT;
drivers/net/ethernet/mediatek/mtk_eth_soc.c
3028
static void mtk_hwlro_val_ipaddr(struct mtk_eth *eth, int idx, __be32 ip)
drivers/net/ethernet/mediatek/mtk_eth_soc.c
3037
mtk_w32(eth, ip, MTK_LRO_DIP_DW0_CFG(idx));
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
368
const struct iphdr *ip;
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
384
ip = (const struct iphdr *)(skb->data + nhoff);
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
385
if (ip_is_fragment(ip))
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
388
if ((ip->protocol != IPPROTO_TCP) && (ip->protocol != IPPROTO_UDP))
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
390
ports = (const __be16 *)(skb->data + nhoff + 4 * ip->ihl);
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
392
ip_proto = ip->protocol;
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
393
src_ip = ip->saddr;
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
394
dst_ip = ip->daddr;
drivers/net/ethernet/mellanox/mlx4/en_tx.c
1046
} ip;
drivers/net/ethernet/mellanox/mlx4/en_tx.c
1049
ip.hdr = skb_inner_network_header(skb);
drivers/net/ethernet/mellanox/mlx4/en_tx.c
1050
proto = (ip.v4->version == 4) ? ip.v4->protocol :
drivers/net/ethernet/mellanox/mlx4/en_tx.c
1051
ip.v6->nexthdr;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_ct.c
1052
__be32 tmp_addr = rev_tuple.ip.src_v4;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_ct.c
1054
rev_tuple.ip.src_v4 = rev_tuple.ip.dst_v4;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_ct.c
1055
rev_tuple.ip.dst_v4 = tmp_addr;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_ct.c
1057
struct in6_addr tmp_addr = rev_tuple.ip.src_v6;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_ct.c
1059
rev_tuple.ip.src_v6 = rev_tuple.ip.dst_v6;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_ct.c
1060
rev_tuple.ip.dst_v6 = tmp_addr;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_ct.c
129
} ip;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_ct.c
1435
tuple->ip.src_v4 = flow_keys.addrs.v4addrs.src;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_ct.c
1436
tuple->ip.dst_v4 = flow_keys.addrs.v4addrs.dst;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_ct.c
1441
tuple->ip.src_v6 = flow_keys.addrs.v6addrs.src;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_ct.c
1442
tuple->ip.dst_v6 = flow_keys.addrs.v6addrs.dst;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_ct.c
255
tuple->ip.src_v4 = match.key->src;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_ct.c
256
tuple->ip.dst_v4 = match.key->dst;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_ct.c
261
tuple->ip.src_v6 = match.key->src;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_ct.c
262
tuple->ip.dst_v6 = match.key->dst;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_ct.c
306
tuple->ip.src_v4 = cpu_to_be32(val);
drivers/net/ethernet/mellanox/mlx5/core/en/tc_ct.c
308
tuple->ip.dst_v4 = cpu_to_be32(val);
drivers/net/ethernet/mellanox/mlx5/core/en/tc_ct.c
317
tuple->ip.src_v6.s6_addr32[ip6_offset] = cpu_to_be32(val);
drivers/net/ethernet/mellanox/mlx5/core/en/tc_ct.c
319
tuple->ip.dst_v6.s6_addr32[ip6_offset - 4] = cpu_to_be32(val);
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
202
char *ip;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
209
ip = (char *)vlan + VLAN_HLEN;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
215
ip = (char *)eth + ETH_HLEN;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
218
return ip;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
232
struct iphdr *ip;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
284
ip = (struct iphdr *)gen_eth_tnl_hdr(encap_header, attr.route_dev, e,
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
288
ip->tos = tun_key->tos;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
289
ip->version = 0x4;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
290
ip->ihl = 0x5;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
291
ip->ttl = attr.ttl;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
292
ip->daddr = attr.fl.fl4.daddr;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
293
ip->saddr = attr.fl.fl4.saddr;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
296
err = mlx5e_gen_ip_tunnel_header((char *)ip + sizeof(struct iphdr),
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
297
&ip->protocol, e);
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
348
struct iphdr *ip;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
389
ip = (struct iphdr *)gen_eth_tnl_hdr(encap_header, attr.route_dev, e,
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
393
ip->tos = tun_key->tos;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
394
ip->version = 0x4;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
395
ip->ihl = 0x5;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
396
ip->ttl = attr.ttl;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
397
ip->daddr = attr.fl.fl4.daddr;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
398
ip->saddr = attr.fl.fl4.saddr;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
401
err = mlx5e_gen_ip_tunnel_header((char *)ip + sizeof(struct iphdr),
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
402
&ip->protocol, e);
drivers/net/ethernet/mellanox/mlx5/core/en/txrx.h
507
const struct iphdr *ip = tunnel ? inner_ip_hdr(skb) : ip_hdr(skb);
drivers/net/ethernet/mellanox/mlx5/core/en/txrx.h
520
if (ip->version == 4) {
drivers/net/ethernet/mellanox/mlx5/core/en/txrx.h
521
th->check = ~tcp_v4_check(len, ip->saddr, ip->daddr, 0);
drivers/net/ethernet/mellanox/mlx5/core/en/txrx.h
530
if (ip->version == 4) {
drivers/net/ethernet/mellanox/mlx5/core/en/txrx.h
531
uh->check = ~udp_v4_check(len, ip->saddr, ip->daddr, 0);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/psp_rxtx.c
195
const struct iphdr *ip = ip_hdr(skb);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/psp_rxtx.c
197
th->check = ~tcp_v4_check(len, ip->saddr, ip->daddr, 0);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1226
void *ip;
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1232
ip = skb->data + network_depth;
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1233
rc = ((proto == htons(ETH_P_IP)) ? IP_ECN_set_ce((struct iphdr *)ip) :
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
1234
IP6_ECN_set_ce(skb, (struct ipv6hdr *)ip));
drivers/net/ethernet/mellanox/mlx5/core/lib/macsec_fs.c
121
struct mlx5_flow_handle *ip;
drivers/net/ethernet/mellanox/mlx5/core/lib/macsec_fs.c
1989
mlx5_del_flow_rules(rx_rule->ip);
drivers/net/ethernet/mellanox/mlx5/core/lib/macsec_fs.c
2156
rx_rule->ip = new_rule;
drivers/net/ethernet/mellanox/mlx5/core/lib/macsec_fs.c
2180
mlx5_del_flow_rules(rx_rule->ip);
drivers/net/ethernet/mellanox/mlx5/core/lib/macsec_fs.c
2181
rx_rule->ip = NULL;
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
2022
MLXSW_ITEM32(afa, ip, s_d, 0x00, 31, 1);
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
2034
MLXSW_ITEM32(afa, ip, m_l, 0x00, 30, 1);
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
2039
MLXSW_ITEM32(afa, ip, ip_63_32, 0x08, 0, 32);
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
2044
MLXSW_ITEM32(afa, ip, ip_31_0, 0x0C, 0, 32);
drivers/net/ethernet/microsoft/mana/mana_en.c
152
struct iphdr *ip = ip_hdr(skb);
drivers/net/ethernet/microsoft/mana/mana_en.c
154
if (ip->protocol == IPPROTO_TCP)
drivers/net/ethernet/microsoft/mana/mana_en.c
157
if (ip->protocol == IPPROTO_UDP)
drivers/net/ethernet/mscc/ocelot_vcap.c
451
arp->ip);
drivers/net/ethernet/qlogic/netxen/netxen_nic.h
1771
int netxen_config_ipaddr(struct netxen_adapter *adapter, __be32 ip, int cmd);
drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c
875
int netxen_config_ipaddr(struct netxen_adapter *adapter, __be32 ip, int cmd)
drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c
888
memcpy(&req.words[1], &ip, sizeof(u32));
drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c
894
(cmd == NX_IP_UP) ? "Add" : "Remove", ip);
drivers/net/ethernet/qlogic/qed/qed_iscsi.c
1224
memcpy(con->local_ip, conn_info->src.ip, sizeof(con->local_ip));
drivers/net/ethernet/qlogic/qed/qed_iscsi.c
1225
memcpy(con->remote_ip, conn_info->dst.ip, sizeof(con->remote_ip));
drivers/net/ethernet/qlogic/qed/qed_nvmetcp.c
688
memcpy(con->local_ip, conn_info->src.ip, sizeof(con->local_ip));
drivers/net/ethernet/qlogic/qed/qed_nvmetcp.c
689
memcpy(con->remote_ip, conn_info->dst.ip, sizeof(con->remote_ip));
drivers/net/ethernet/qlogic/qede/qede_filter.c
1428
struct iphdr *ip = (struct iphdr *)(header + ETH_HLEN);
drivers/net/ethernet/qlogic/qede/qede_filter.c
1432
ip->saddr = t->src_ipv4;
drivers/net/ethernet/qlogic/qede/qede_filter.c
1433
ip->daddr = t->dst_ipv4;
drivers/net/ethernet/qlogic/qede/qede_filter.c
1434
ip->version = 0x4;
drivers/net/ethernet/qlogic/qede/qede_filter.c
1435
ip->ihl = 0x5;
drivers/net/ethernet/qlogic/qede/qede_filter.c
1436
ip->protocol = t->ip_proto;
drivers/net/ethernet/qlogic/qede/qede_filter.c
1437
ip->tot_len = cpu_to_be16(qede_flow_get_min_header_size(t) - ETH_HLEN);
drivers/net/ethernet/qlogic/qla3xxx.c
2281
const struct iphdr *ip = ip_hdr(skb);
drivers/net/ethernet/qlogic/qla3xxx.c
2284
mac_iocb_ptr->ip_hdr_len = ip->ihl;
drivers/net/ethernet/qlogic/qla3xxx.c
2286
if (ip->protocol == IPPROTO_TCP) {
drivers/net/ethernet/qlogic/qlcnic/qlcnic.h
2152
__be32 ip, int cmd)
drivers/net/ethernet/qlogic/qlcnic/qlcnic.h
2154
adapter->nic_ops->config_ipaddr(adapter, ip, cmd);
drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_hw.c
1977
void qlcnic_83xx_config_ipaddr(struct qlcnic_adapter *adapter, __be32 ip,
drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_hw.c
2004
temp_ip = swab32(ntohl(ip));
drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_hw.c
2010
(mode == QLCNIC_IP_UP) ? "Add" : "Remove", ip);
drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c
936
__be32 ip, int cmd)
drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c
951
ipa->ipv4 = ip;
drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c
957
(cmd == QLCNIC_IP_UP) ? "Add" : "Remove", ip);
drivers/net/ethernet/realtek/8139cp.c
766
const struct iphdr *ip = ip_hdr(skb);
drivers/net/ethernet/realtek/8139cp.c
767
if (ip->protocol == IPPROTO_TCP)
drivers/net/ethernet/realtek/8139cp.c
769
else if (ip->protocol == IPPROTO_UDP)
drivers/net/ethernet/realtek/r8169_main.c
4458
const struct iphdr *ip = ip_hdr(skb);
drivers/net/ethernet/realtek/r8169_main.c
4460
if (ip->protocol == IPPROTO_TCP)
drivers/net/ethernet/realtek/r8169_main.c
4462
else if (ip->protocol == IPPROTO_UDP)
drivers/net/ethernet/sfc/ef10.c
2241
struct iphdr *ip;
drivers/net/ethernet/sfc/ef10.c
2258
ip = ip_hdr(skb);
drivers/net/ethernet/sfc/ef10.c
2259
if (ip->version == 4)
drivers/net/ethernet/sfc/ef10.c
2260
outer_ipv4_id = ntohs(ip->id);
drivers/net/ethernet/sfc/ef10.c
2262
ip = inner_ip_hdr(skb);
drivers/net/ethernet/sfc/ef10.c
2265
ip = ip_hdr(skb);
drivers/net/ethernet/sfc/ef10.c
2284
if (ip->version == 4) {
drivers/net/ethernet/sfc/ef10.c
2285
ip->tot_len = htons(ip_tot_len);
drivers/net/ethernet/sfc/ef10.c
2286
ip->check = 0;
drivers/net/ethernet/sfc/ef10.c
2287
inner_ipv4_id = ntohs(ip->id);
drivers/net/ethernet/sfc/ef10.c
2289
((struct ipv6hdr *)ip)->payload_len = htons(ip_tot_len);
drivers/net/ethernet/sfc/falcon/selftest.c
305
received.ip.saddr = payload->ip.saddr;
drivers/net/ethernet/sfc/falcon/selftest.c
307
received.ip.check = payload->ip.check;
drivers/net/ethernet/sfc/falcon/selftest.c
335
if (memcmp(&received.ip, &payload->ip, sizeof(payload->ip)) != 0) {
drivers/net/ethernet/sfc/falcon/selftest.c
394
payload->ip.daddr = htonl(INADDR_LOOPBACK);
drivers/net/ethernet/sfc/falcon/selftest.c
395
payload->ip.ihl = 5;
drivers/net/ethernet/sfc/falcon/selftest.c
396
payload->ip.check = (__force __sum16) htons(0xdead);
drivers/net/ethernet/sfc/falcon/selftest.c
397
payload->ip.tot_len = htons(sizeof(*payload) -
drivers/net/ethernet/sfc/falcon/selftest.c
398
offsetof(struct ef4_loopback_payload, ip));
drivers/net/ethernet/sfc/falcon/selftest.c
399
payload->ip.version = IPVERSION;
drivers/net/ethernet/sfc/falcon/selftest.c
400
payload->ip.protocol = IPPROTO_UDP;
drivers/net/ethernet/sfc/falcon/selftest.c
441
payload->ip.saddr = htonl(INADDR_LOOPBACK | (i << 2));
drivers/net/ethernet/sfc/falcon/selftest.c
45
struct iphdr ip;
drivers/net/ethernet/sfc/selftest.c
302
received.ip.saddr = payload->ip.saddr;
drivers/net/ethernet/sfc/selftest.c
304
received.ip.check = payload->ip.check;
drivers/net/ethernet/sfc/selftest.c
332
if (memcmp(&received.ip, &payload->ip, sizeof(payload->ip)) != 0) {
drivers/net/ethernet/sfc/selftest.c
391
payload->ip.daddr = htonl(INADDR_LOOPBACK);
drivers/net/ethernet/sfc/selftest.c
392
payload->ip.ihl = 5;
drivers/net/ethernet/sfc/selftest.c
393
payload->ip.check = (__force __sum16) htons(0xdead);
drivers/net/ethernet/sfc/selftest.c
394
payload->ip.tot_len = htons(sizeof(*payload) -
drivers/net/ethernet/sfc/selftest.c
395
offsetof(struct efx_loopback_payload, ip));
drivers/net/ethernet/sfc/selftest.c
396
payload->ip.version = IPVERSION;
drivers/net/ethernet/sfc/selftest.c
397
payload->ip.protocol = IPPROTO_UDP;
drivers/net/ethernet/sfc/selftest.c
438
payload->ip.saddr = htonl(INADDR_LOOPBACK | (i << 2));
drivers/net/ethernet/sfc/selftest.c
47
struct iphdr ip;
drivers/net/ethernet/sfc/siena/selftest.c
303
received.ip.saddr = payload->ip.saddr;
drivers/net/ethernet/sfc/siena/selftest.c
305
received.ip.check = payload->ip.check;
drivers/net/ethernet/sfc/siena/selftest.c
333
if (memcmp(&received.ip, &payload->ip, sizeof(payload->ip)) != 0) {
drivers/net/ethernet/sfc/siena/selftest.c
392
payload->ip.daddr = htonl(INADDR_LOOPBACK);
drivers/net/ethernet/sfc/siena/selftest.c
393
payload->ip.ihl = 5;
drivers/net/ethernet/sfc/siena/selftest.c
394
payload->ip.check = (__force __sum16) htons(0xdead);
drivers/net/ethernet/sfc/siena/selftest.c
395
payload->ip.tot_len = htons(sizeof(*payload) -
drivers/net/ethernet/sfc/siena/selftest.c
396
offsetof(struct efx_loopback_payload, ip));
drivers/net/ethernet/sfc/siena/selftest.c
397
payload->ip.version = IPVERSION;
drivers/net/ethernet/sfc/siena/selftest.c
398
payload->ip.protocol = IPPROTO_UDP;
drivers/net/ethernet/sfc/siena/selftest.c
439
payload->ip.saddr = htonl(INADDR_LOOPBACK | (i << 2));
drivers/net/ethernet/sfc/siena/selftest.c
48
struct iphdr ip;
drivers/net/ethernet/sfc/tc.c
371
MAP_KEY_AND_MASK(IP, ip, tos, ip_tos);
drivers/net/ethernet/sfc/tc.c
372
MAP_KEY_AND_MASK(IP, ip, ttl, ip_ttl);
drivers/net/ethernet/sfc/tc.c
419
MAP_ENC_KEY_AND_MASK(IP, ip, enc_ip, tos, enc_ip_tos);
drivers/net/ethernet/sfc/tc.c
420
MAP_ENC_KEY_AND_MASK(IP, ip, enc_ip, ttl, enc_ip_ttl);
drivers/net/ethernet/sfc/tc_encap_actions.c
271
struct iphdr *ip;
drivers/net/ethernet/sfc/tc_encap_actions.c
273
ip = (struct iphdr *)(encap->encap_hdr + encap->encap_hdr_len);
drivers/net/ethernet/sfc/tc_encap_actions.c
274
encap->encap_hdr_len += sizeof(*ip);
drivers/net/ethernet/sfc/tc_encap_actions.c
276
ip->daddr = key->u.ipv4.dst;
drivers/net/ethernet/sfc/tc_encap_actions.c
277
ip->saddr = key->u.ipv4.src;
drivers/net/ethernet/sfc/tc_encap_actions.c
278
ip->ttl = neigh->ttl;
drivers/net/ethernet/sfc/tc_encap_actions.c
279
ip->protocol = ipproto;
drivers/net/ethernet/sfc/tc_encap_actions.c
280
ip->version = 0x4;
drivers/net/ethernet/sfc/tc_encap_actions.c
281
ip->ihl = 0x5;
drivers/net/ethernet/sfc/tc_encap_actions.c
282
ip->tot_len = cpu_to_be16(ip->ihl * 4 + len);
drivers/net/ethernet/sfc/tc_encap_actions.c
283
ip_send_check(ip);
drivers/net/ethernet/sfc/tc_encap_actions.c
291
struct ipv6hdr *ip;
drivers/net/ethernet/sfc/tc_encap_actions.c
293
ip = (struct ipv6hdr *)(encap->encap_hdr + encap->encap_hdr_len);
drivers/net/ethernet/sfc/tc_encap_actions.c
294
encap->encap_hdr_len += sizeof(*ip);
drivers/net/ethernet/sfc/tc_encap_actions.c
296
ip6_flow_hdr(ip, key->tos, key->label);
drivers/net/ethernet/sfc/tc_encap_actions.c
297
ip->daddr = key->u.ipv6.dst;
drivers/net/ethernet/sfc/tc_encap_actions.c
298
ip->saddr = key->u.ipv6.src;
drivers/net/ethernet/sfc/tc_encap_actions.c
299
ip->hop_limit = neigh->ttl;
drivers/net/ethernet/sfc/tc_encap_actions.c
300
ip->nexthdr = ipproto;
drivers/net/ethernet/sfc/tc_encap_actions.c
301
ip->version = 0x6;
drivers/net/ethernet/sfc/tc_encap_actions.c
302
ip->payload_len = cpu_to_be16(len);
drivers/net/ethernet/sgi/ioc3-eth.c
1033
spin_lock_irq(&ip->ioc3_lock);
drivers/net/ethernet/sgi/ioc3-eth.c
1038
produce = ip->tx_pi;
drivers/net/ethernet/sgi/ioc3-eth.c
1039
desc = &ip->txr[produce];
drivers/net/ethernet/sgi/ioc3-eth.c
1061
d1 = dma_map_single(ip->dma_dev, skb->data, s1, DMA_TO_DEVICE);
drivers/net/ethernet/sgi/ioc3-eth.c
1062
if (dma_mapping_error(ip->dma_dev, d1))
drivers/net/ethernet/sgi/ioc3-eth.c
1064
d2 = dma_map_single(ip->dma_dev, (void *)b2, s1, DMA_TO_DEVICE);
drivers/net/ethernet/sgi/ioc3-eth.c
1065
if (dma_mapping_error(ip->dma_dev, d2)) {
drivers/net/ethernet/sgi/ioc3-eth.c
1066
dma_unmap_single(ip->dma_dev, d1, len, DMA_TO_DEVICE);
drivers/net/ethernet/sgi/ioc3-eth.c
1077
d = dma_map_single(ip->dma_dev, skb->data, len, DMA_TO_DEVICE);
drivers/net/ethernet/sgi/ioc3-eth.c
1078
if (dma_mapping_error(ip->dma_dev, d))
drivers/net/ethernet/sgi/ioc3-eth.c
1085
ip->tx_skbs[produce] = skb; /* Remember skb */
drivers/net/ethernet/sgi/ioc3-eth.c
1087
ip->tx_pi = produce;
drivers/net/ethernet/sgi/ioc3-eth.c
1088
writel(produce << 7, &ip->regs->etpir); /* Fire ... */
drivers/net/ethernet/sgi/ioc3-eth.c
109
static void ioc3_start(struct ioc3_private *ip);
drivers/net/ethernet/sgi/ioc3-eth.c
1090
ip->txqlen++;
drivers/net/ethernet/sgi/ioc3-eth.c
1092
if (ip->txqlen >= (TX_RING_ENTRIES - 1))
drivers/net/ethernet/sgi/ioc3-eth.c
1095
spin_unlock_irq(&ip->ioc3_lock);
drivers/net/ethernet/sgi/ioc3-eth.c
110
static inline void ioc3_stop(struct ioc3_private *ip);
drivers/net/ethernet/sgi/ioc3-eth.c
1103
spin_unlock_irq(&ip->ioc3_lock);
drivers/net/ethernet/sgi/ioc3-eth.c
1110
struct ioc3_private *ip = netdev_priv(dev);
drivers/net/ethernet/sgi/ioc3-eth.c
1114
spin_lock_irq(&ip->ioc3_lock);
drivers/net/ethernet/sgi/ioc3-eth.c
1116
ioc3_stop(ip);
drivers/net/ethernet/sgi/ioc3-eth.c
1117
ioc3_free_rx_bufs(ip);
drivers/net/ethernet/sgi/ioc3-eth.c
1118
ioc3_clean_tx_ring(ip);
drivers/net/ethernet/sgi/ioc3-eth.c
1123
spin_unlock_irq(&ip->ioc3_lock);
drivers/net/ethernet/sgi/ioc3-eth.c
1126
ioc3_start(ip);
drivers/net/ethernet/sgi/ioc3-eth.c
1127
ioc3_mii_init(ip);
drivers/net/ethernet/sgi/ioc3-eth.c
1128
ioc3_mii_start(ip);
drivers/net/ethernet/sgi/ioc3-eth.c
113
static void ioc3_free_rx_bufs(struct ioc3_private *ip);
drivers/net/ethernet/sgi/ioc3-eth.c
1130
spin_unlock_irq(&ip->ioc3_lock);
drivers/net/ethernet/sgi/ioc3-eth.c
114
static inline void ioc3_clean_tx_ring(struct ioc3_private *ip);
drivers/net/ethernet/sgi/ioc3-eth.c
1168
struct ioc3_private *ip = netdev_priv(dev);
drivers/net/ethernet/sgi/ioc3-eth.c
1170
spin_lock_irq(&ip->ioc3_lock);
drivers/net/ethernet/sgi/ioc3-eth.c
1171
mii_ethtool_get_link_ksettings(&ip->mii, cmd);
drivers/net/ethernet/sgi/ioc3-eth.c
1172
spin_unlock_irq(&ip->ioc3_lock);
drivers/net/ethernet/sgi/ioc3-eth.c
1180
struct ioc3_private *ip = netdev_priv(dev);
drivers/net/ethernet/sgi/ioc3-eth.c
1183
spin_lock_irq(&ip->ioc3_lock);
drivers/net/ethernet/sgi/ioc3-eth.c
1184
rc = mii_ethtool_set_link_ksettings(&ip->mii, cmd);
drivers/net/ethernet/sgi/ioc3-eth.c
1185
spin_unlock_irq(&ip->ioc3_lock);
drivers/net/ethernet/sgi/ioc3-eth.c
1192
struct ioc3_private *ip = netdev_priv(dev);
drivers/net/ethernet/sgi/ioc3-eth.c
1195
spin_lock_irq(&ip->ioc3_lock);
drivers/net/ethernet/sgi/ioc3-eth.c
1196
rc = mii_nway_restart(&ip->mii);
drivers/net/ethernet/sgi/ioc3-eth.c
1197
spin_unlock_irq(&ip->ioc3_lock);
drivers/net/ethernet/sgi/ioc3-eth.c
1204
struct ioc3_private *ip = netdev_priv(dev);
drivers/net/ethernet/sgi/ioc3-eth.c
1207
spin_lock_irq(&ip->ioc3_lock);
drivers/net/ethernet/sgi/ioc3-eth.c
1208
rc = mii_link_ok(&ip->mii);
drivers/net/ethernet/sgi/ioc3-eth.c
1209
spin_unlock_irq(&ip->ioc3_lock);
drivers/net/ethernet/sgi/ioc3-eth.c
1224
struct ioc3_private *ip = netdev_priv(dev);
drivers/net/ethernet/sgi/ioc3-eth.c
1227
spin_lock_irq(&ip->ioc3_lock);
drivers/net/ethernet/sgi/ioc3-eth.c
1228
rc = generic_mii_ioctl(&ip->mii, if_mii(rq), cmd, NULL);
drivers/net/ethernet/sgi/ioc3-eth.c
1229
spin_unlock_irq(&ip->ioc3_lock);
drivers/net/ethernet/sgi/ioc3-eth.c
123
static inline int ioc3_alloc_skb(struct ioc3_private *ip, struct sk_buff **skb,
drivers/net/ethernet/sgi/ioc3-eth.c
1236
struct ioc3_private *ip = netdev_priv(dev);
drivers/net/ethernet/sgi/ioc3-eth.c
1237
struct ioc3_ethregs *regs = ip->regs;
drivers/net/ethernet/sgi/ioc3-eth.c
1241
spin_lock_irq(&ip->ioc3_lock);
drivers/net/ethernet/sgi/ioc3-eth.c
1244
ip->emcr |= EMCR_PROMISC;
drivers/net/ethernet/sgi/ioc3-eth.c
1245
writel(ip->emcr, ®s->emcr);
drivers/net/ethernet/sgi/ioc3-eth.c
1248
ip->emcr &= ~EMCR_PROMISC;
drivers/net/ethernet/sgi/ioc3-eth.c
1249
writel(ip->emcr, ®s->emcr); /* Clear promiscuous. */
drivers/net/ethernet/sgi/ioc3-eth.c
1258
ip->ehar_h = 0xffffffff;
drivers/net/ethernet/sgi/ioc3-eth.c
1259
ip->ehar_l = 0xffffffff;
drivers/net/ethernet/sgi/ioc3-eth.c
1264
ip->ehar_h = ehar >> 32;
drivers/net/ethernet/sgi/ioc3-eth.c
1265
ip->ehar_l = ehar & 0xffffffff;
drivers/net/ethernet/sgi/ioc3-eth.c
1267
writel(ip->ehar_h, ®s->ehar_h);
drivers/net/ethernet/sgi/ioc3-eth.c
1268
writel(ip->ehar_l, ®s->ehar_l);
drivers/net/ethernet/sgi/ioc3-eth.c
1271
spin_unlock_irq(&ip->ioc3_lock);
drivers/net/ethernet/sgi/ioc3-eth.c
139
d = dma_map_single(ip->dma_dev, new_skb->data,
drivers/net/ethernet/sgi/ioc3-eth.c
142
if (dma_mapping_error(ip->dma_dev, d)) {
drivers/net/ethernet/sgi/ioc3-eth.c
229
struct ioc3_private *ip = netdev_priv(dev);
drivers/net/ethernet/sgi/ioc3-eth.c
233
&ip->regs->emar_h);
drivers/net/ethernet/sgi/ioc3-eth.c
238
&ip->regs->emar_l);
drivers/net/ethernet/sgi/ioc3-eth.c
243
struct ioc3_private *ip = netdev_priv(dev);
drivers/net/ethernet/sgi/ioc3-eth.c
248
spin_lock_irq(&ip->ioc3_lock);
drivers/net/ethernet/sgi/ioc3-eth.c
250
spin_unlock_irq(&ip->ioc3_lock);
drivers/net/ethernet/sgi/ioc3-eth.c
260
struct ioc3_private *ip = netdev_priv(dev);
drivers/net/ethernet/sgi/ioc3-eth.c
261
struct ioc3_ethregs *regs = ip->regs;
drivers/net/ethernet/sgi/ioc3-eth.c
275
struct ioc3_private *ip = netdev_priv(dev);
drivers/net/ethernet/sgi/ioc3-eth.c
276
struct ioc3_ethregs *regs = ip->regs;
drivers/net/ethernet/sgi/ioc3-eth.c
286
static int ioc3_mii_init(struct ioc3_private *ip);
drivers/net/ethernet/sgi/ioc3-eth.c
290
struct ioc3_private *ip = netdev_priv(dev);
drivers/net/ethernet/sgi/ioc3-eth.c
291
struct ioc3_ethregs *regs = ip->regs;
drivers/net/ethernet/sgi/ioc3-eth.c
367
struct ioc3_private *ip = netdev_priv(dev);
drivers/net/ethernet/sgi/ioc3-eth.c
375
rxr = ip->rxr; /* Ring base */
drivers/net/ethernet/sgi/ioc3-eth.c
376
rx_entry = ip->rx_ci; /* RX consume index */
drivers/net/ethernet/sgi/ioc3-eth.c
377
n_entry = ip->rx_pi;
drivers/net/ethernet/sgi/ioc3-eth.c
379
skb = ip->rx_skbs[rx_entry];
drivers/net/ethernet/sgi/ioc3-eth.c
390
if (ioc3_alloc_skb(ip, &new_skb, &rxb, &d)) {
drivers/net/ethernet/sgi/ioc3-eth.c
405
dma_unmap_single(ip->dma_dev, rxr[rx_entry],
drivers/net/ethernet/sgi/ioc3-eth.c
410
ip->rx_skbs[rx_entry] = NULL; /* Poison */
drivers/net/ethernet/sgi/ioc3-eth.c
429
ip->rx_skbs[n_entry] = new_skb;
drivers/net/ethernet/sgi/ioc3-eth.c
436
skb = ip->rx_skbs[rx_entry];
drivers/net/ethernet/sgi/ioc3-eth.c
440
writel((n_entry << 3) | ERPIR_ARM, &ip->regs->erpir);
drivers/net/ethernet/sgi/ioc3-eth.c
441
ip->rx_pi = n_entry;
drivers/net/ethernet/sgi/ioc3-eth.c
442
ip->rx_ci = rx_entry;
drivers/net/ethernet/sgi/ioc3-eth.c
447
struct ioc3_private *ip = netdev_priv(dev);
drivers/net/ethernet/sgi/ioc3-eth.c
448
struct ioc3_ethregs *regs = ip->regs;
drivers/net/ethernet/sgi/ioc3-eth.c
454
spin_lock(&ip->ioc3_lock);
drivers/net/ethernet/sgi/ioc3-eth.c
458
o_entry = ip->tx_ci;
drivers/net/ethernet/sgi/ioc3-eth.c
464
skb = ip->tx_skbs[o_entry];
drivers/net/ethernet/sgi/ioc3-eth.c
467
ip->tx_skbs[o_entry] = NULL;
drivers/net/ethernet/sgi/ioc3-eth.c
477
ip->txqlen -= packets;
drivers/net/ethernet/sgi/ioc3-eth.c
479
if (netif_queue_stopped(dev) && ip->txqlen < TX_RING_ENTRIES)
drivers/net/ethernet/sgi/ioc3-eth.c
482
ip->tx_ci = o_entry;
drivers/net/ethernet/sgi/ioc3-eth.c
483
spin_unlock(&ip->ioc3_lock);
drivers/net/ethernet/sgi/ioc3-eth.c
494
struct ioc3_private *ip = netdev_priv(dev);
drivers/net/ethernet/sgi/ioc3-eth.c
496
spin_lock(&ip->ioc3_lock);
drivers/net/ethernet/sgi/ioc3-eth.c
511
ioc3_stop(ip);
drivers/net/ethernet/sgi/ioc3-eth.c
512
ioc3_free_rx_bufs(ip);
drivers/net/ethernet/sgi/ioc3-eth.c
513
ioc3_clean_tx_ring(ip);
drivers/net/ethernet/sgi/ioc3-eth.c
518
spin_unlock(&ip->ioc3_lock);
drivers/net/ethernet/sgi/ioc3-eth.c
521
ioc3_start(ip);
drivers/net/ethernet/sgi/ioc3-eth.c
522
ioc3_mii_init(ip);
drivers/net/ethernet/sgi/ioc3-eth.c
526
spin_unlock(&ip->ioc3_lock);
drivers/net/ethernet/sgi/ioc3-eth.c
534
struct ioc3_private *ip = netdev_priv(dev_id);
drivers/net/ethernet/sgi/ioc3-eth.c
535
struct ioc3_ethregs *regs = ip->regs;
drivers/net/ethernet/sgi/ioc3-eth.c
553
static inline void ioc3_setup_duplex(struct ioc3_private *ip)
drivers/net/ethernet/sgi/ioc3-eth.c
555
struct ioc3_ethregs *regs = ip->regs;
drivers/net/ethernet/sgi/ioc3-eth.c
557
spin_lock_irq(&ip->ioc3_lock);
drivers/net/ethernet/sgi/ioc3-eth.c
559
if (ip->mii.full_duplex) {
drivers/net/ethernet/sgi/ioc3-eth.c
561
ip->emcr |= EMCR_DUPLEX;
drivers/net/ethernet/sgi/ioc3-eth.c
564
ip->emcr &= ~EMCR_DUPLEX;
drivers/net/ethernet/sgi/ioc3-eth.c
566
writel(ip->emcr, ®s->emcr);
drivers/net/ethernet/sgi/ioc3-eth.c
568
spin_unlock_irq(&ip->ioc3_lock);
drivers/net/ethernet/sgi/ioc3-eth.c
573
struct ioc3_private *ip = timer_container_of(ip, t, ioc3_timer);
drivers/net/ethernet/sgi/ioc3-eth.c
576
mii_check_media(&ip->mii, 1, 0);
drivers/net/ethernet/sgi/ioc3-eth.c
577
ioc3_setup_duplex(ip);
drivers/net/ethernet/sgi/ioc3-eth.c
579
ip->ioc3_timer.expires = jiffies + ((12 * HZ) / 10); /* 1.2s */
drivers/net/ethernet/sgi/ioc3-eth.c
580
add_timer(&ip->ioc3_timer);
drivers/net/ethernet/sgi/ioc3-eth.c
587
static int ioc3_mii_init(struct ioc3_private *ip)
drivers/net/ethernet/sgi/ioc3-eth.c
593
word = ioc3_mdio_read(ip->mii.dev, i, MII_PHYSID1);
drivers/net/ethernet/sgi/ioc3-eth.c
596
ip->mii.phy_id = i;
drivers/net/ethernet/sgi/ioc3-eth.c
600
ip->mii.phy_id = -1;
drivers/net/ethernet/sgi/ioc3-eth.c
604
static void ioc3_mii_start(struct ioc3_private *ip)
drivers/net/ethernet/sgi/ioc3-eth.c
606
ip->ioc3_timer.expires = jiffies + (12 * HZ) / 10; /* 1.2 sec. */
drivers/net/ethernet/sgi/ioc3-eth.c
607
add_timer(&ip->ioc3_timer);
drivers/net/ethernet/sgi/ioc3-eth.c
610
static inline void ioc3_tx_unmap(struct ioc3_private *ip, int entry)
drivers/net/ethernet/sgi/ioc3-eth.c
615
desc = &ip->txr[entry];
drivers/net/ethernet/sgi/ioc3-eth.c
620
dma_unmap_single(ip->dma_dev, be64_to_cpu(desc->p1),
drivers/net/ethernet/sgi/ioc3-eth.c
625
dma_unmap_single(ip->dma_dev, be64_to_cpu(desc->p2),
drivers/net/ethernet/sgi/ioc3-eth.c
630
static inline void ioc3_clean_tx_ring(struct ioc3_private *ip)
drivers/net/ethernet/sgi/ioc3-eth.c
636
skb = ip->tx_skbs[i];
drivers/net/ethernet/sgi/ioc3-eth.c
638
ioc3_tx_unmap(ip, i);
drivers/net/ethernet/sgi/ioc3-eth.c
639
ip->tx_skbs[i] = NULL;
drivers/net/ethernet/sgi/ioc3-eth.c
642
ip->txr[i].cmd = 0;
drivers/net/ethernet/sgi/ioc3-eth.c
644
ip->tx_pi = 0;
drivers/net/ethernet/sgi/ioc3-eth.c
645
ip->tx_ci = 0;
drivers/net/ethernet/sgi/ioc3-eth.c
648
static void ioc3_free_rx_bufs(struct ioc3_private *ip)
drivers/net/ethernet/sgi/ioc3-eth.c
653
n_entry = ip->rx_ci;
drivers/net/ethernet/sgi/ioc3-eth.c
654
rx_entry = ip->rx_pi;
drivers/net/ethernet/sgi/ioc3-eth.c
657
skb = ip->rx_skbs[n_entry];
drivers/net/ethernet/sgi/ioc3-eth.c
659
dma_unmap_single(ip->dma_dev,
drivers/net/ethernet/sgi/ioc3-eth.c
660
be64_to_cpu(ip->rxr[n_entry]),
drivers/net/ethernet/sgi/ioc3-eth.c
670
struct ioc3_private *ip = netdev_priv(dev);
drivers/net/ethernet/sgi/ioc3-eth.c
680
if (ioc3_alloc_skb(ip, &ip->rx_skbs[i], &rxb, &d))
drivers/net/ethernet/sgi/ioc3-eth.c
684
ip->rxr[i] = cpu_to_be64(ioc3_map(d, PCI64_ATTR_BAR));
drivers/net/ethernet/sgi/ioc3-eth.c
686
ip->rx_ci = 0;
drivers/net/ethernet/sgi/ioc3-eth.c
687
ip->rx_pi = RX_BUFFS;
drivers/net/ethernet/sgi/ioc3-eth.c
692
static inline void ioc3_ssram_disc(struct ioc3_private *ip)
drivers/net/ethernet/sgi/ioc3-eth.c
694
struct ioc3_ethregs *regs = ip->regs;
drivers/net/ethernet/sgi/ioc3-eth.c
695
u32 *ssram0 = &ip->ssram[0x0000];
drivers/net/ethernet/sgi/ioc3-eth.c
696
u32 *ssram1 = &ip->ssram[0x4000];
drivers/net/ethernet/sgi/ioc3-eth.c
709
ip->emcr |= EMCR_RAMPAR;
drivers/net/ethernet/sgi/ioc3-eth.c
712
ip->emcr |= EMCR_BUFSIZ | EMCR_RAMPAR;
drivers/net/ethernet/sgi/ioc3-eth.c
718
struct ioc3_private *ip = netdev_priv(dev);
drivers/net/ethernet/sgi/ioc3-eth.c
719
struct ioc3_ethregs *regs = ip->regs;
drivers/net/ethernet/sgi/ioc3-eth.c
721
timer_delete_sync(&ip->ioc3_timer); /* Kill if running */
drivers/net/ethernet/sgi/ioc3-eth.c
735
writel(ip->ehar_h, ®s->ehar_h);
drivers/net/ethernet/sgi/ioc3-eth.c
736
writel(ip->ehar_l, ®s->ehar_l);
drivers/net/ethernet/sgi/ioc3-eth.c
740
static void ioc3_start(struct ioc3_private *ip)
drivers/net/ethernet/sgi/ioc3-eth.c
742
struct ioc3_ethregs *regs = ip->regs;
drivers/net/ethernet/sgi/ioc3-eth.c
746
ring = ioc3_map(ip->rxr_dma, PCI64_ATTR_PREC);
drivers/net/ethernet/sgi/ioc3-eth.c
749
writel(ip->rx_ci << 3, ®s->ercir);
drivers/net/ethernet/sgi/ioc3-eth.c
750
writel((ip->rx_pi << 3) | ERPIR_ARM, ®s->erpir);
drivers/net/ethernet/sgi/ioc3-eth.c
752
ring = ioc3_map(ip->txr_dma, PCI64_ATTR_PREC);
drivers/net/ethernet/sgi/ioc3-eth.c
754
ip->txqlen = 0; /* nothing queued */
drivers/net/ethernet/sgi/ioc3-eth.c
759
writel(ip->tx_pi << 7, ®s->etpir);
drivers/net/ethernet/sgi/ioc3-eth.c
760
writel(ip->tx_ci << 7, ®s->etcir);
drivers/net/ethernet/sgi/ioc3-eth.c
763
ip->emcr |= ((RX_OFFSET / 2) << EMCR_RXOFF_SHIFT) | EMCR_TXDMAEN |
drivers/net/ethernet/sgi/ioc3-eth.c
765
writel(ip->emcr, ®s->emcr);
drivers/net/ethernet/sgi/ioc3-eth.c
772
static inline void ioc3_stop(struct ioc3_private *ip)
drivers/net/ethernet/sgi/ioc3-eth.c
774
struct ioc3_ethregs *regs = ip->regs;
drivers/net/ethernet/sgi/ioc3-eth.c
783
struct ioc3_private *ip = netdev_priv(dev);
drivers/net/ethernet/sgi/ioc3-eth.c
785
ip->ehar_h = 0;
drivers/net/ethernet/sgi/ioc3-eth.c
786
ip->ehar_l = 0;
drivers/net/ethernet/sgi/ioc3-eth.c
793
ioc3_start(ip);
drivers/net/ethernet/sgi/ioc3-eth.c
794
ioc3_mii_start(ip);
drivers/net/ethernet/sgi/ioc3-eth.c
802
struct ioc3_private *ip = netdev_priv(dev);
drivers/net/ethernet/sgi/ioc3-eth.c
804
timer_delete_sync(&ip->ioc3_timer);
drivers/net/ethernet/sgi/ioc3-eth.c
808
ioc3_stop(ip);
drivers/net/ethernet/sgi/ioc3-eth.c
810
ioc3_free_rx_bufs(ip);
drivers/net/ethernet/sgi/ioc3-eth.c
811
ioc3_clean_tx_ring(ip);
drivers/net/ethernet/sgi/ioc3-eth.c
831
struct ioc3_private *ip;
drivers/net/ethernet/sgi/ioc3-eth.c
852
ip = netdev_priv(dev);
drivers/net/ethernet/sgi/ioc3-eth.c
853
ip->dma_dev = pdev->dev.parent;
drivers/net/ethernet/sgi/ioc3-eth.c
854
ip->regs = devm_platform_ioremap_resource(pdev, 0);
drivers/net/ethernet/sgi/ioc3-eth.c
855
if (IS_ERR(ip->regs)) {
drivers/net/ethernet/sgi/ioc3-eth.c
856
err = PTR_ERR(ip->regs);
drivers/net/ethernet/sgi/ioc3-eth.c
860
ip->ssram = devm_platform_ioremap_resource(pdev, 1);
drivers/net/ethernet/sgi/ioc3-eth.c
861
if (IS_ERR(ip->ssram)) {
drivers/net/ethernet/sgi/ioc3-eth.c
862
err = PTR_ERR(ip->ssram);
drivers/net/ethernet/sgi/ioc3-eth.c
879
spin_lock_init(&ip->ioc3_lock);
drivers/net/ethernet/sgi/ioc3-eth.c
880
timer_setup(&ip->ioc3_timer, ioc3_timer, 0);
drivers/net/ethernet/sgi/ioc3-eth.c
882
ioc3_stop(ip);
drivers/net/ethernet/sgi/ioc3-eth.c
885
ip->rxr = dma_alloc_coherent(ip->dma_dev, RX_RING_SIZE, &ip->rxr_dma,
drivers/net/ethernet/sgi/ioc3-eth.c
887
if (!ip->rxr) {
drivers/net/ethernet/sgi/ioc3-eth.c
894
ip->tx_ring = dma_alloc_coherent(ip->dma_dev, TX_RING_SIZE + SZ_16K - 1,
drivers/net/ethernet/sgi/ioc3-eth.c
895
&ip->txr_dma, GFP_KERNEL);
drivers/net/ethernet/sgi/ioc3-eth.c
896
if (!ip->tx_ring) {
drivers/net/ethernet/sgi/ioc3-eth.c
902
ip->txr = PTR_ALIGN(ip->tx_ring, SZ_16K);
drivers/net/ethernet/sgi/ioc3-eth.c
903
ip->txr_dma = ALIGN(ip->txr_dma, SZ_16K);
drivers/net/ethernet/sgi/ioc3-eth.c
907
ip->mii.phy_id_mask = 0x1f;
drivers/net/ethernet/sgi/ioc3-eth.c
908
ip->mii.reg_num_mask = 0x1f;
drivers/net/ethernet/sgi/ioc3-eth.c
909
ip->mii.dev = dev;
drivers/net/ethernet/sgi/ioc3-eth.c
910
ip->mii.mdio_read = ioc3_mdio_read;
drivers/net/ethernet/sgi/ioc3-eth.c
911
ip->mii.mdio_write = ioc3_mdio_write;
drivers/net/ethernet/sgi/ioc3-eth.c
913
ioc3_mii_init(ip);
drivers/net/ethernet/sgi/ioc3-eth.c
915
if (ip->mii.phy_id == -1) {
drivers/net/ethernet/sgi/ioc3-eth.c
921
ioc3_mii_start(ip);
drivers/net/ethernet/sgi/ioc3-eth.c
922
ioc3_ssram_disc(ip);
drivers/net/ethernet/sgi/ioc3-eth.c
932
sw_physid1 = ioc3_mdio_read(dev, ip->mii.phy_id, MII_PHYSID1);
drivers/net/ethernet/sgi/ioc3-eth.c
933
sw_physid2 = ioc3_mdio_read(dev, ip->mii.phy_id, MII_PHYSID2);
drivers/net/ethernet/sgi/ioc3-eth.c
939
mii_check_media(&ip->mii, 1, 1);
drivers/net/ethernet/sgi/ioc3-eth.c
940
ioc3_setup_duplex(ip);
drivers/net/ethernet/sgi/ioc3-eth.c
946
ip->mii.phy_id, vendor, model, rev);
drivers/net/ethernet/sgi/ioc3-eth.c
948
ip->emcr & EMCR_BUFSIZ ? 128 : 64);
drivers/net/ethernet/sgi/ioc3-eth.c
953
timer_delete_sync(&ip->ioc3_timer);
drivers/net/ethernet/sgi/ioc3-eth.c
954
if (ip->rxr)
drivers/net/ethernet/sgi/ioc3-eth.c
955
dma_free_coherent(ip->dma_dev, RX_RING_SIZE, ip->rxr,
drivers/net/ethernet/sgi/ioc3-eth.c
956
ip->rxr_dma);
drivers/net/ethernet/sgi/ioc3-eth.c
957
if (ip->tx_ring)
drivers/net/ethernet/sgi/ioc3-eth.c
958
dma_free_coherent(ip->dma_dev, TX_RING_SIZE + SZ_16K - 1, ip->tx_ring,
drivers/net/ethernet/sgi/ioc3-eth.c
959
ip->txr_dma);
drivers/net/ethernet/sgi/ioc3-eth.c
968
struct ioc3_private *ip = netdev_priv(dev);
drivers/net/ethernet/sgi/ioc3-eth.c
970
dma_free_coherent(ip->dma_dev, RX_RING_SIZE, ip->rxr, ip->rxr_dma);
drivers/net/ethernet/sgi/ioc3-eth.c
971
dma_free_coherent(ip->dma_dev, TX_RING_SIZE + SZ_16K - 1, ip->tx_ring, ip->txr_dma);
drivers/net/ethernet/sgi/ioc3-eth.c
974
timer_delete_sync(&ip->ioc3_timer);
drivers/net/ethernet/sgi/ioc3-eth.c
981
struct ioc3_private *ip = netdev_priv(dev);
drivers/net/ethernet/via/via-velocity.c
2605
const struct iphdr *ip = ip_hdr(skb);
drivers/net/ethernet/via/via-velocity.c
2606
if (ip->protocol == IPPROTO_TCP)
drivers/net/ethernet/via/via-velocity.c
2608
else if (ip->protocol == IPPROTO_UDP)
drivers/net/ethernet/wangxun/libwx/wx_type.h
760
u32 ip:3; /* outer ip*/
drivers/net/ethernet/wangxun/libwx/wx_type.h
768
#define WX_PTT(mac, ip, etype, eip, proto, layer)\
drivers/net/ethernet/wangxun/libwx/wx_type.h
771
WX_DEC_PTYPE_IP_##ip, /* ip */ \
drivers/net/ethernet/wangxun/txgbe/txgbe_fdir.c
289
common.ip ^= hdr.ipv6->saddr.s6_addr32[0] ^
drivers/net/ethernet/wangxun/txgbe/txgbe_fdir.c
299
common.ip ^= hdr.ipv4->saddr ^ hdr.ipv4->daddr;
drivers/net/ethernet/wangxun/txgbe/txgbe_type.h
271
__be32 ip;
drivers/net/gtp.c
135
static inline u32 ipv4_hashfn(__be32 ip)
drivers/net/gtp.c
137
return jhash_1word((__force u32)ip, gtp_h_initval);
drivers/net/gtp.c
2081
__be32 ip = nla_get_be32(nla[GTPA_MS_ADDRESS]);
drivers/net/gtp.c
2086
return ipv4_pdp_find(gtp, ip);
drivers/net/hyperv/netvsc_drv.c
397
struct iphdr *ip = ip_hdr(skb);
drivers/net/hyperv/netvsc_drv.c
399
if (ip->protocol == IPPROTO_TCP)
drivers/net/hyperv/netvsc_drv.c
401
else if (ip->protocol == IPPROTO_UDP)
drivers/net/netconsole.c
406
np_info(np, "local IPv4 address %pI4\n", &np->local_ip.ip);
drivers/net/netconsole.c
413
np_info(np, "remote IPv4 address %pI4\n", &np->remote_ip.ip);
drivers/net/slip/slhc.c
238
struct iphdr *ip;
drivers/net/slip/slhc.c
250
ip = (struct iphdr *) icp;
drivers/net/slip/slhc.c
251
if (ip->version != 4 || ip->ihl < 5)
drivers/net/slip/slhc.c
255
if (ip->protocol != IPPROTO_TCP || (ntohs(ip->frag_off) & 0x3fff)) {
drivers/net/slip/slhc.c
257
if(ip->protocol != IPPROTO_TCP)
drivers/net/slip/slhc.c
263
nlen = ip->ihl * 4;
drivers/net/slip/slhc.c
297
if( ip->saddr == cs->cs_ip.saddr
drivers/net/slip/slhc.c
298
&& ip->daddr == cs->cs_ip.daddr
drivers/net/slip/slhc.c
353
if(ip->version != cs->cs_ip.version || ip->ihl != cs->cs_ip.ihl
drivers/net/slip/slhc.c
354
|| ip->tos != cs->cs_ip.tos
drivers/net/slip/slhc.c
355
|| (ip->frag_off & htons(0x4000)) != (cs->cs_ip.frag_off & htons(0x4000))
drivers/net/slip/slhc.c
356
|| ip->ttl != cs->cs_ip.ttl
drivers/net/slip/slhc.c
358
|| (ip->ihl > 5 && memcmp(ip+1,cs->cs_ipopt,((ip->ihl)-5)*4) != 0)
drivers/net/slip/slhc.c
405
if(ip->tot_len != cs->cs_ip.tot_len &&
drivers/net/slip/slhc.c
431
deltaS = ntohs(ip->id) - ntohs(cs->cs_ip.id);
drivers/net/slip/slhc.c
442
memcpy(&cs->cs_ip,ip,20);
drivers/net/slip/slhc.c
476
memcpy(&cs->cs_ip,ip,20);
drivers/net/slip/slhc.c
478
if (ip->ihl > 5)
drivers/net/slip/slhc.c
479
memcpy(cs->cs_ipopt, ip+1, ((ip->ihl) - 5) * 4);
drivers/net/slip/slhc.c
498
struct iphdr *ip;
drivers/net/slip/slhc.c
535
ip = &cs->cs_ip;
drivers/net/slip/slhc.c
547
hdrlen = ip->ihl * 4 + thp->doff * 4;
drivers/net/slip/slhc.c
553
i = ntohs(ip->tot_len) - hdrlen;
drivers/net/slip/slhc.c
561
ntohs(ip->tot_len) - hdrlen);
drivers/net/slip/slhc.c
597
ip->id = htons (ntohs (ip->id) + x);
drivers/net/slip/slhc.c
599
ip->id = htons (ntohs (ip->id) + 1);
drivers/net/slip/slhc.c
611
ip->tot_len = htons(len);
drivers/net/slip/slhc.c
612
ip->check = 0;
drivers/net/slip/slhc.c
617
memcpy(cp, ip, 20);
drivers/net/slip/slhc.c
620
if (ip->ihl > 5) {
drivers/net/slip/slhc.c
621
memcpy(cp, cs->cs_ipopt, (ip->ihl - 5) * 4);
drivers/net/slip/slhc.c
622
cp += (ip->ihl - 5) * 4;
drivers/net/slip/slhc.c
625
put_unaligned(ip_fast_csum(icp, ip->ihl),
drivers/net/vxlan/vxlan_core.c
1020
rc = vxlan_fdb_replace(f, ip, port, vni,
drivers/net/vxlan/vxlan_core.c
1032
rc = vxlan_fdb_append(f, ip, port, vni, ifindex, &rd);
drivers/net/vxlan/vxlan_core.c
1068
const u8 *mac, union vxlan_addr *ip,
drivers/net/vxlan/vxlan_core.c
1084
netdev_dbg(vxlan->dev, "add %pM -> %pIS\n", mac, ip);
drivers/net/vxlan/vxlan_core.c
1085
rc = vxlan_fdb_create(vxlan, mac, ip, state, port, src_vni,
drivers/net/vxlan/vxlan_core.c
1104
const u8 *mac, union vxlan_addr *ip,
drivers/net/vxlan/vxlan_core.c
1121
return vxlan_fdb_update_existing(vxlan, ip, state, flags, port,
drivers/net/vxlan/vxlan_core.c
1128
return vxlan_fdb_update_create(vxlan, mac, ip, state, flags,
drivers/net/vxlan/vxlan_core.c
1144
union vxlan_addr *ip, __be16 *port, __be32 *src_vni,
drivers/net/vxlan/vxlan_core.c
1158
err = vxlan_nla_get_addr(ip, tb[NDA_DST]);
drivers/net/vxlan/vxlan_core.c
1167
ip->sin.sin_addr.s_addr = htonl(INADDR_ANY);
drivers/net/vxlan/vxlan_core.c
1168
ip->sa.sa_family = AF_INET;
drivers/net/vxlan/vxlan_core.c
1171
ip->sin6.sin6_addr = in6addr_any;
drivers/net/vxlan/vxlan_core.c
1172
ip->sa.sa_family = AF_INET6;
drivers/net/vxlan/vxlan_core.c
1237
union vxlan_addr ip;
drivers/net/vxlan/vxlan_core.c
1252
err = vxlan_fdb_parse(tb, vxlan, &ip, &port, &src_vni, &vni, &ifindex,
drivers/net/vxlan/vxlan_core.c
1257
if (vxlan->default_dst.remote_ip.sa.sa_family != ip.sa.sa_family)
drivers/net/vxlan/vxlan_core.c
1261
err = vxlan_fdb_update(vxlan, addr, &ip, ndm->ndm_state, flags,
drivers/net/vxlan/vxlan_core.c
1274
const unsigned char *addr, union vxlan_addr ip,
drivers/net/vxlan/vxlan_core.c
1286
if (!vxlan_addr_any(&ip)) {
drivers/net/vxlan/vxlan_core.c
1287
rd = vxlan_fdb_find_rdst(f, &ip, port, vni, ifindex);
drivers/net/vxlan/vxlan_core.c
1313
union vxlan_addr ip;
drivers/net/vxlan/vxlan_core.c
1319
err = vxlan_fdb_parse(tb, vxlan, &ip, &port, &src_vni, &vni, &ifindex,
drivers/net/vxlan/vxlan_core.c
1325
err = __vxlan_fdb_delete(vxlan, addr, ip, port, src_vni, vni, ifindex,
drivers/net/vxlan/vxlan_core.c
3165
union vxlan_addr ip;
drivers/net/vxlan/vxlan_core.c
3167
err = vxlan_nla_get_addr(&ip, tb[NDA_DST]);
drivers/net/vxlan/vxlan_core.c
3173
desc.dst_ip = ip;
drivers/net/vxlan/vxlan_core.c
428
union vxlan_addr *ip, __be16 port,
drivers/net/vxlan/vxlan_core.c
434
if (vxlan_addr_equal(&rd->remote_ip, ip) &&
drivers/net/vxlan/vxlan_core.c
549
union vxlan_addr *ip, __be16 port, __be32 vni,
drivers/net/vxlan/vxlan_core.c
554
rd = vxlan_fdb_find_rdst(f, ip, port, vni, ifindex);
drivers/net/vxlan/vxlan_core.c
564
rd->remote_ip = *ip;
drivers/net/vxlan/vxlan_core.c
574
union vxlan_addr *ip, __be16 port, __be32 vni,
drivers/net/vxlan/vxlan_core.c
579
rd = vxlan_fdb_find_rdst(f, ip, port, vni, ifindex);
drivers/net/vxlan/vxlan_core.c
592
rd->remote_ip = *ip;
drivers/net/vxlan/vxlan_core.c
857
const u8 *mac, union vxlan_addr *ip,
drivers/net/vxlan/vxlan_core.c
871
netdev_dbg(vxlan->dev, "add %pM -> %pIS\n", mac, ip);
drivers/net/vxlan/vxlan_core.c
879
rc = vxlan_fdb_append(f, ip, port, vni, ifindex, &rd);
drivers/net/vxlan/vxlan_core.c
969
union vxlan_addr *ip,
drivers/net/vxlan/vxlan_multicast.c
101
union vxlan_addr *ip, int ifindex)
drivers/net/vxlan/vxlan_multicast.c
106
if (vxlan_group_used_match(ip, ifindex,
drivers/net/vxlan/vxlan_multicast.c
115
if (vxlan_group_used_match(ip, ifindex,
drivers/net/vxlan/vxlan_multicast.c
128
union vxlan_addr *ip = (rip ? : &dev->default_dst.remote_ip);
drivers/net/vxlan/vxlan_multicast.c
164
if (!vxlan_group_used_by_vnifilter(vxlan, ip, ifindex))
drivers/net/vxlan/vxlan_multicast.c
167
if (!vxlan_group_used_match(ip, ifindex,
drivers/net/vxlan/vxlan_multicast.c
20
union vxlan_addr *ip = (rip ? : &vxlan->default_dst.remote_ip);
drivers/net/vxlan/vxlan_multicast.c
25
if (ip->sa.sa_family == AF_INET) {
drivers/net/vxlan/vxlan_multicast.c
28
.imr_multiaddr.s_addr = ip->sin.sin_addr.s_addr,
drivers/net/vxlan/vxlan_multicast.c
43
&ip->sin6.sin6_addr);
drivers/net/vxlan/vxlan_multicast.c
54
union vxlan_addr *ip = (rip ? : &vxlan->default_dst.remote_ip);
drivers/net/vxlan/vxlan_multicast.c
59
if (ip->sa.sa_family == AF_INET) {
drivers/net/vxlan/vxlan_multicast.c
62
.imr_multiaddr.s_addr = ip->sin.sin_addr.s_addr,
drivers/net/vxlan/vxlan_multicast.c
77
&ip->sin6.sin6_addr);
drivers/net/vxlan/vxlan_multicast.c
85
static bool vxlan_group_used_match(union vxlan_addr *ip, int ifindex,
drivers/net/vxlan/vxlan_multicast.c
91
if (!vxlan_addr_equal(rip, ip))
drivers/net/vxlan/vxlan_private.h
100
ip->sa.sa_family = AF_INET;
drivers/net/vxlan/vxlan_private.h
108
const union vxlan_addr *ip)
drivers/net/vxlan/vxlan_private.h
110
if (ip->sa.sa_family == AF_INET6)
drivers/net/vxlan/vxlan_private.h
111
return nla_put_in6_addr(skb, attr, &ip->sin6.sin6_addr);
drivers/net/vxlan/vxlan_private.h
113
return nla_put_in_addr(skb, attr, ip->sin.sin_addr.s_addr);
drivers/net/vxlan/vxlan_private.h
116
static inline bool vxlan_addr_is_multicast(const union vxlan_addr *ip)
drivers/net/vxlan/vxlan_private.h
118
if (ip->sa.sa_family == AF_INET6)
drivers/net/vxlan/vxlan_private.h
119
return ipv6_addr_is_multicast(&ip->sin6.sin6_addr);
drivers/net/vxlan/vxlan_private.h
121
return ipv4_is_multicast(ip->sin.sin_addr.s_addr);
drivers/net/vxlan/vxlan_private.h
132
static inline int vxlan_nla_get_addr(union vxlan_addr *ip,
drivers/net/vxlan/vxlan_private.h
138
ip->sin.sin_addr.s_addr = nla_get_in_addr(nla);
drivers/net/vxlan/vxlan_private.h
139
ip->sa.sa_family = AF_INET;
drivers/net/vxlan/vxlan_private.h
147
const union vxlan_addr *ip)
drivers/net/vxlan/vxlan_private.h
149
return nla_put_in_addr(skb, attr, ip->sin.sin_addr.s_addr);
drivers/net/vxlan/vxlan_private.h
152
static inline bool vxlan_addr_is_multicast(const union vxlan_addr *ip)
drivers/net/vxlan/vxlan_private.h
154
return ipv4_is_multicast(ip->sin.sin_addr.s_addr);
drivers/net/vxlan/vxlan_private.h
159
static inline size_t vxlan_addr_size(const union vxlan_addr *ip)
drivers/net/vxlan/vxlan_private.h
161
if (ip->sa.sa_family == AF_INET6)
drivers/net/vxlan/vxlan_private.h
182
const u8 *mac, union vxlan_addr *ip,
drivers/net/vxlan/vxlan_private.h
188
const unsigned char *addr, union vxlan_addr ip,
drivers/net/vxlan/vxlan_private.h
192
const u8 *mac, union vxlan_addr *ip,
drivers/net/vxlan/vxlan_private.h
91
static inline int vxlan_nla_get_addr(union vxlan_addr *ip,
drivers/net/vxlan/vxlan_private.h
95
ip->sin6.sin6_addr = nla_get_in6_addr(nla);
drivers/net/vxlan/vxlan_private.h
96
ip->sa.sa_family = AF_INET6;
drivers/net/vxlan/vxlan_private.h
99
ip->sin.sin_addr.s_addr = nla_get_in_addr(nla);
drivers/net/wireguard/allowedips.c
131
u8 ip[16] __aligned(__alignof(u64));
drivers/net/wireguard/allowedips.c
135
swap_endian(ip, be_ip, bits);
drivers/net/wireguard/allowedips.c
139
node = find_node(rcu_dereference_bh(root), bits, ip);
drivers/net/wireguard/allowedips.c
327
int wg_allowedips_insert_v4(struct allowedips *table, const struct in_addr *ip,
drivers/net/wireguard/allowedips.c
334
swap_endian(key, (const u8 *)ip, 32);
drivers/net/wireguard/allowedips.c
338
int wg_allowedips_insert_v6(struct allowedips *table, const struct in6_addr *ip,
drivers/net/wireguard/allowedips.c
345
swap_endian(key, (const u8 *)ip, 128);
drivers/net/wireguard/allowedips.c
349
int wg_allowedips_remove_v4(struct allowedips *table, const struct in_addr *ip,
drivers/net/wireguard/allowedips.c
356
swap_endian(key, (const u8 *)ip, 32);
drivers/net/wireguard/allowedips.c
360
int wg_allowedips_remove_v6(struct allowedips *table, const struct in6_addr *ip,
drivers/net/wireguard/allowedips.c
367
swap_endian(key, (const u8 *)ip, 128);
drivers/net/wireguard/allowedips.c
383
int wg_allowedips_read_node(struct allowedips_node *node, u8 ip[16], u8 *cidr)
drivers/net/wireguard/allowedips.c
386
swap_endian(ip, node->bits, node->bitlen);
drivers/net/wireguard/allowedips.c
387
memset(ip + cidr_bytes, 0, node->bitlen / 8U - cidr_bytes);
drivers/net/wireguard/allowedips.c
389
ip[cidr_bytes - 1U] &= ~0U << (-node->cidr % 8U);
drivers/net/wireguard/allowedips.h
37
int wg_allowedips_insert_v4(struct allowedips *table, const struct in_addr *ip,
drivers/net/wireguard/allowedips.h
39
int wg_allowedips_insert_v6(struct allowedips *table, const struct in6_addr *ip,
drivers/net/wireguard/allowedips.h
41
int wg_allowedips_remove_v4(struct allowedips *table, const struct in_addr *ip,
drivers/net/wireguard/allowedips.h
43
int wg_allowedips_remove_v6(struct allowedips *table, const struct in6_addr *ip,
drivers/net/wireguard/allowedips.h
48
int wg_allowedips_read_node(struct allowedips_node *node, u8 ip[16], u8 *cidr);
drivers/net/wireguard/netlink.c
148
u8 cidr, ip[16] __aligned(__alignof(u64));
drivers/net/wireguard/netlink.c
151
family = wg_allowedips_read_node(allowedips_node, ip, &cidr);
drivers/net/wireguard/netlink.c
152
if (get_allowedips(skb, ip, cidr, family)) {
drivers/net/wireguard/netlink.c
46
static int get_allowedips(struct sk_buff *skb, const u8 *ip, u8 cidr,
drivers/net/wireguard/netlink.c
58
sizeof(struct in6_addr) : sizeof(struct in_addr), ip)) {
drivers/net/wireguard/ratelimiter.c
103
memcpy(&ip, &ipv6_hdr(skb)->saddr, sizeof(ip));
drivers/net/wireguard/ratelimiter.c
104
bucket = &table_v6[hsiphash_3u32(net_word, ip >> 32, ip, &key) &
drivers/net/wireguard/ratelimiter.c
112
if (entry->net == net && entry->ip == ip) {
drivers/net/wireguard/ratelimiter.c
143
entry->ip = ip;
drivers/net/wireguard/ratelimiter.c
27
u64 last_time_ns, tokens, ip;
drivers/net/wireguard/ratelimiter.c
93
u64 ip;
drivers/net/wireguard/ratelimiter.c
96
ip = (u64 __force)ip_hdr(skb)->saddr;
drivers/net/wireguard/ratelimiter.c
97
bucket = &table_v4[hsiphash_2u32(net_word, ip, &key) &
drivers/net/wireguard/selftest/allowedips.c
128
node->ip.ip &= node->mask.ip;
drivers/net/wireguard/selftest/allowedips.c
130
node->ip.ip6[0] &= node->mask.ip6[0];
drivers/net/wireguard/selftest/allowedips.c
131
node->ip.ip6[1] &= node->mask.ip6[1];
drivers/net/wireguard/selftest/allowedips.c
132
node->ip.ip6[2] &= node->mask.ip6[2];
drivers/net/wireguard/selftest/allowedips.c
133
node->ip.ip6[3] &= node->mask.ip6[3];
drivers/net/wireguard/selftest/allowedips.c
138
horrible_match_v4(const struct horrible_allowedips_node *node, struct in_addr *ip)
drivers/net/wireguard/selftest/allowedips.c
140
return (ip->s_addr & node->mask.ip) == node->ip.ip;
drivers/net/wireguard/selftest/allowedips.c
144
horrible_match_v6(const struct horrible_allowedips_node *node, struct in6_addr *ip)
drivers/net/wireguard/selftest/allowedips.c
146
return (ip->in6_u.u6_addr32[0] & node->mask.ip6[0]) == node->ip.ip6[0] &&
drivers/net/wireguard/selftest/allowedips.c
147
(ip->in6_u.u6_addr32[1] & node->mask.ip6[1]) == node->ip.ip6[1] &&
drivers/net/wireguard/selftest/allowedips.c
148
(ip->in6_u.u6_addr32[2] & node->mask.ip6[2]) == node->ip.ip6[2] &&
drivers/net/wireguard/selftest/allowedips.c
149
(ip->in6_u.u6_addr32[3] & node->mask.ip6[3]) == node->ip.ip6[3];
drivers/net/wireguard/selftest/allowedips.c
161
!memcmp(&other->ip, &node->ip, sizeof(union nf_inet_addr))) {
drivers/net/wireguard/selftest/allowedips.c
182
struct in_addr *ip, u8 cidr, void *value)
drivers/net/wireguard/selftest/allowedips.c
188
node->ip.in = *ip;
drivers/net/wireguard/selftest/allowedips.c
199
struct in6_addr *ip, u8 cidr, void *value)
drivers/net/wireguard/selftest/allowedips.c
205
node->ip.in6 = *ip;
drivers/net/wireguard/selftest/allowedips.c
215
horrible_allowedips_lookup_v4(struct horrible_allowedips *table, struct in_addr *ip)
drivers/net/wireguard/selftest/allowedips.c
220
if (node->ip_version == 4 && horrible_match_v4(node, ip))
drivers/net/wireguard/selftest/allowedips.c
227
horrible_allowedips_lookup_v6(struct horrible_allowedips *table, struct in6_addr *ip)
drivers/net/wireguard/selftest/allowedips.c
232
if (node->ip_version == 6 && horrible_match_v6(node, ip))
drivers/net/wireguard/selftest/allowedips.c
257
u8 ip[16], mutate_mask[16], mutated[16];
drivers/net/wireguard/selftest/allowedips.c
287
get_random_bytes(ip, 4);
drivers/net/wireguard/selftest/allowedips.c
290
if (wg_allowedips_insert_v4(&t, (struct in_addr *)ip, cidr,
drivers/net/wireguard/selftest/allowedips.c
295
if (horrible_allowedips_insert_v4(&h, (struct in_addr *)ip,
drivers/net/wireguard/selftest/allowedips.c
301
memcpy(mutated, ip, 4);
drivers/net/wireguard/selftest/allowedips.c
331
get_random_bytes(ip, 16);
drivers/net/wireguard/selftest/allowedips.c
334
if (wg_allowedips_insert_v6(&t, (struct in6_addr *)ip, cidr,
drivers/net/wireguard/selftest/allowedips.c
339
if (horrible_allowedips_insert_v6(&h, (struct in6_addr *)ip,
drivers/net/wireguard/selftest/allowedips.c
345
memcpy(mutated, ip, 16);
drivers/net/wireguard/selftest/allowedips.c
384
get_random_bytes(ip, 4);
drivers/net/wireguard/selftest/allowedips.c
385
if (lookup(t.root4, 32, ip) != horrible_allowedips_lookup_v4(&h, (struct in_addr *)ip)) {
drivers/net/wireguard/selftest/allowedips.c
389
get_random_bytes(ip, 16);
drivers/net/wireguard/selftest/allowedips.c
390
if (lookup(t.root6, 128, ip) != horrible_allowedips_lookup_v6(&h, (struct in6_addr *)ip)) {
drivers/net/wireguard/selftest/allowedips.c
426
static struct in_addr ip;
drivers/net/wireguard/selftest/allowedips.c
427
u8 *split = (u8 *)&ip;
drivers/net/wireguard/selftest/allowedips.c
433
return &ip;
drivers/net/wireguard/selftest/allowedips.c
438
static struct in6_addr ip;
drivers/net/wireguard/selftest/allowedips.c
439
__be32 *split = (__be32 *)&ip;
drivers/net/wireguard/selftest/allowedips.c
445
return &ip;
drivers/net/wireguard/selftest/allowedips.c
460
wg_allowedips_insert_v##version(&t, ip##version(ipa, ipb, ipc, ipd), \
drivers/net/wireguard/selftest/allowedips.c
464
wg_allowedips_remove_v##version(&t, ip##version(ipa, ipb, ipc, ipd), \
drivers/net/wireguard/selftest/allowedips.c
477
ip##version(ipa, ipb, ipc, ipd)) == (mem); \
drivers/net/wireguard/selftest/allowedips.c
483
ip##version(ipa, ipb, ipc, ipd)) != (mem); \
drivers/net/wireguard/selftest/allowedips.c
503
struct in6_addr ip;
drivers/net/wireguard/selftest/allowedips.c
648
memset(&ip, 0xff, 8);
drivers/net/wireguard/selftest/allowedips.c
649
memcpy((u8 *)&ip + 8, &part, 8);
drivers/net/wireguard/selftest/allowedips.c
650
wg_allowedips_insert_v6(&t, &ip, 128, a, &mutex);
drivers/net/wireguard/selftest/allowedips.c
651
memcpy(&ip, &part, 8);
drivers/net/wireguard/selftest/allowedips.c
652
memset((u8 *)&ip + 8, 0, 8);
drivers/net/wireguard/selftest/allowedips.c
653
wg_allowedips_insert_v6(&t, &ip, 128, a, &mutex);
drivers/net/wireguard/selftest/allowedips.c
655
memset(&ip, 0, 16);
drivers/net/wireguard/selftest/allowedips.c
656
wg_allowedips_insert_v6(&t, &ip, 128, a, &mutex);
drivers/net/wireguard/selftest/allowedips.c
666
u8 cidr, ip[16] __aligned(__alignof(u64));
drivers/net/wireguard/selftest/allowedips.c
667
int family = wg_allowedips_read_node(iter_node, ip, &cidr);
drivers/net/wireguard/selftest/allowedips.c
672
!memcmp(ip, ip4(192, 95, 5, 64), sizeof(struct in_addr)))
drivers/net/wireguard/selftest/allowedips.c
675
!memcmp(ip, ip6(0x26075300, 0x60006b00, 0, 0xc05f0543),
drivers/net/wireguard/selftest/allowedips.c
679
!memcmp(ip, ip4(10, 1, 0, 16), sizeof(struct in_addr)))
drivers/net/wireguard/selftest/allowedips.c
682
!memcmp(ip, ip6(0x26075300, 0x6d8a6bf8, 0xdab1e000, 0),
drivers/net/wireguard/selftest/allowedips.c
686
!memcmp(ip, ip6(0x26075000, 0, 0, 0),
drivers/net/wireguard/selftest/allowedips.c
84
union nf_inet_addr ip;
drivers/net/wireless/ath/ath5k/eeprom.c
654
const u16 *ip;
drivers/net/wireless/ath/ath5k/eeprom.c
658
ip = intercepts3_2;
drivers/net/wireless/ath/ath5k/eeprom.c
660
ip = intercepts3;
drivers/net/wireless/ath/ath5k/eeprom.c
663
vp[i] = (ip[i] * max + (100 - ip[i]) * min) / 100;
drivers/net/wireless/ath/ath9k/ar9003_eeprom.c
4272
int ip = 0;
drivers/net/wireless/ath/ath9k/ar9003_eeprom.c
4282
for (ip = 0; ip < np; ip++) {
drivers/net/wireless/ath/ath9k/ar9003_eeprom.c
4283
dx = x - px[ip];
drivers/net/wireless/ath/ath9k/ar9003_eeprom.c
4289
hx = px[ip];
drivers/net/wireless/ath/ath9k/ar9003_eeprom.c
4290
hy = py[ip];
drivers/net/wireless/ath/ath9k/ar9003_eeprom.c
4298
lx = px[ip];
drivers/net/wireless/ath/ath9k/ar9003_eeprom.c
4299
ly = py[ip];
drivers/net/wireless/intel/ipw2x00/libipw_tx.c
211
struct iphdr *ip;
drivers/net/wireless/intel/ipw2x00/libipw_tx.c
217
ip = ip_hdr(skb);
drivers/net/wireless/intel/ipw2x00/libipw_tx.c
218
switch (ip->tos & 0xfc) {
drivers/net/wireless/intel/iwlegacy/3945.c
44
#define IL_DECLARE_RATE_INFO(r, ip, in, rp, rn, pp, np) \
drivers/net/wireless/intel/iwlegacy/3945.c
47
RATE_##ip##M_IDX, \
drivers/net/wireless/intel/iwlegacy/3945.c
54
RATE_##ip##M_IDX_TBL }
drivers/net/wireless/intel/iwlegacy/4965-rs.c
61
#define IL_DECLARE_RATE_INFO(r, s, ip, in, rp, rn, pp, np) \
drivers/net/wireless/intel/iwlegacy/4965-rs.c
66
RATE_##ip##M_IDX, \
drivers/net/wireless/intel/iwlwifi/dvm/rs.c
57
#define IWL_DECLARE_RATE_INFO(r, s, ip, in, rp, rn, pp, np) \
drivers/net/wireless/intel/iwlwifi/dvm/rs.c
63
IWL_RATE_##ip##M_INDEX, \
drivers/net/wireless/realtek/rtlwifi/base.c
1497
const struct iphdr *ip;
drivers/net/wireless/realtek/rtlwifi/base.c
1506
ip = (struct iphdr *)((u8 *)ether_type_ptr +
drivers/net/wireless/realtek/rtlwifi/base.c
1508
if (IPPROTO_UDP == ip->protocol) {
drivers/net/wireless/realtek/rtlwifi/base.c
1509
struct udphdr *udp = (struct udphdr *)((u8 *)ip +
drivers/net/wireless/realtek/rtlwifi/base.c
1510
(ip->ihl << 2));
drivers/pinctrl/aspeed/pinctrl-aspeed-g4.c
2554
if (!ctx->maps[desc->ip])
drivers/pinctrl/aspeed/pinctrl-aspeed-g4.c
2576
if (desc->ip == ASPEED_IP_SCU && desc->reg == HW_STRAP1 &&
drivers/pinctrl/aspeed/pinctrl-aspeed-g4.c
2580
if (desc->ip == ASPEED_IP_SCU && desc->reg == HW_STRAP2)
drivers/pinctrl/aspeed/pinctrl-aspeed-g4.c
2583
ret = regmap_update_bits(ctx->maps[desc->ip], desc->reg,
drivers/pinctrl/aspeed/pinctrl-aspeed-g5.c
2617
int ip)
drivers/pinctrl/aspeed/pinctrl-aspeed-g5.c
2619
if (ip == ASPEED_IP_SCU) {
drivers/pinctrl/aspeed/pinctrl-aspeed-g5.c
2620
WARN(!ctx->maps[ip], "Missing SCU syscon!");
drivers/pinctrl/aspeed/pinctrl-aspeed-g5.c
2621
return ctx->maps[ip];
drivers/pinctrl/aspeed/pinctrl-aspeed-g5.c
2624
if (ip >= ASPEED_NR_PINMUX_IPS)
drivers/pinctrl/aspeed/pinctrl-aspeed-g5.c
2627
if (likely(ctx->maps[ip]))
drivers/pinctrl/aspeed/pinctrl-aspeed-g5.c
2628
return ctx->maps[ip];
drivers/pinctrl/aspeed/pinctrl-aspeed-g5.c
2630
if (ip == ASPEED_IP_GFX) {
drivers/pinctrl/aspeed/pinctrl-aspeed-g5.c
2649
if (ip == ASPEED_IP_LPC) {
drivers/pinctrl/aspeed/pinctrl-aspeed-g5.c
2685
map = aspeed_g5_acquire_regmap(ctx, desc->ip);
drivers/pinctrl/aspeed/pinctrl-aspeed-g5.c
2689
desc->ip);
drivers/pinctrl/aspeed/pinctrl-aspeed-g5.c
2693
ret = aspeed_sig_desc_eval(desc, enabled, ctx->maps[desc->ip]);
drivers/pinctrl/aspeed/pinctrl-aspeed-g5.c
2727
map = aspeed_g5_acquire_regmap(ctx, desc->ip);
drivers/pinctrl/aspeed/pinctrl-aspeed-g5.c
2731
desc->ip);
drivers/pinctrl/aspeed/pinctrl-aspeed-g5.c
2754
if (desc->ip == ASPEED_IP_SCU && desc->reg == HW_STRAP1 &&
drivers/pinctrl/aspeed/pinctrl-aspeed-g5.c
2758
if (desc->ip == ASPEED_IP_SCU && desc->reg == HW_STRAP2)
drivers/pinctrl/aspeed/pinctrl-aspeed-g5.c
2762
if (desc->ip == ASPEED_IP_SCU && desc->reg == HW_STRAP1) {
drivers/pinctrl/aspeed/pinctrl-aspeed-g5.c
2766
ret = regmap_write(ctx->maps[desc->ip],
drivers/pinctrl/aspeed/pinctrl-aspeed-g5.c
2773
ret = regmap_update_bits(ctx->maps[desc->ip], desc->reg,
drivers/pinctrl/aspeed/pinctrl-aspeed-g6.c
2672
if (!ctx->maps[desc->ip])
drivers/pinctrl/aspeed/pinctrl-aspeed-g6.c
2675
WARN_ON(desc->ip != ASPEED_IP_SCU);
drivers/pinctrl/aspeed/pinctrl-aspeed-g6.c
2693
ret = regmap_update_bits(ctx->maps[desc->ip],
drivers/pinctrl/aspeed/pinctrl-aspeed-g6.c
2698
ret = regmap_update_bits(ctx->maps[desc->ip], desc->reg,
drivers/pinctrl/aspeed/pinmux-aspeed.c
18
aspeed_pinmux_ips[desc->ip], desc->reg,
drivers/pinctrl/aspeed/pinmux-aspeed.c
95
ret = aspeed_sig_desc_eval(desc, enabled, ctx->maps[desc->ip]);
drivers/pinctrl/aspeed/pinmux-aspeed.h
442
unsigned int ip;
drivers/pinctrl/aspeed/pinmux-aspeed.h
487
#define SIG_DESC_IP_BIT(ip, reg, idx, val) \
drivers/pinctrl/aspeed/pinmux-aspeed.h
488
{ ip, reg, BIT_MASK(idx), val, (((val) + 1) & 1) }
drivers/pinctrl/aspeed/pinmux-aspeed.h
501
#define SIG_DESC_IP_SET(ip, reg, idx) SIG_DESC_IP_BIT(ip, reg, idx, 1)
drivers/platform/x86/intel/pmc/core.c
314
static void pmc_core_display_map(struct seq_file *s, int index, int idx, int ip,
drivers/platform/x86/intel/pmc/core.c
318
pmc_idx, ip, pf_map[idx][index].name,
drivers/platform/x86/intel/pmc/core.c
331
unsigned int index, iter, idx, ip = 0;
drivers/platform/x86/intel/pmc/core.c
345
index < pmc->map->ppfear_buckets * 8; ip++, index++)
drivers/platform/x86/intel/pmc/core.c
346
pmc_core_display_map(s, index, idx, ip, pmc_idx,
drivers/pwm/pwm-bcm-iproc.c
102
value = readl(ip->base + IPROC_PWM_PERIOD_OFFSET(pwm->hwpwm));
drivers/pwm/pwm-bcm-iproc.c
106
value = readl(ip->base + IPROC_PWM_DUTY_CYCLE_OFFSET(pwm->hwpwm));
drivers/pwm/pwm-bcm-iproc.c
117
struct iproc_pwmc *ip = to_iproc_pwmc(chip);
drivers/pwm/pwm-bcm-iproc.c
121
rate = clk_get_rate(ip->clk);
drivers/pwm/pwm-bcm-iproc.c
154
iproc_pwmc_disable(ip, pwm->hwpwm);
drivers/pwm/pwm-bcm-iproc.c
157
value = readl(ip->base + IPROC_PWM_PRESCALE_OFFSET);
drivers/pwm/pwm-bcm-iproc.c
160
writel(value, ip->base + IPROC_PWM_PRESCALE_OFFSET);
drivers/pwm/pwm-bcm-iproc.c
163
writel(period, ip->base + IPROC_PWM_PERIOD_OFFSET(pwm->hwpwm));
drivers/pwm/pwm-bcm-iproc.c
164
writel(duty, ip->base + IPROC_PWM_DUTY_CYCLE_OFFSET(pwm->hwpwm));
drivers/pwm/pwm-bcm-iproc.c
167
value = readl(ip->base + IPROC_PWM_CTRL_OFFSET);
drivers/pwm/pwm-bcm-iproc.c
174
writel(value, ip->base + IPROC_PWM_CTRL_OFFSET);
drivers/pwm/pwm-bcm-iproc.c
177
iproc_pwmc_enable(ip, pwm->hwpwm);
drivers/pwm/pwm-bcm-iproc.c
190
struct iproc_pwmc *ip;
drivers/pwm/pwm-bcm-iproc.c
195
chip = devm_pwmchip_alloc(&pdev->dev, 4, sizeof(*ip));
drivers/pwm/pwm-bcm-iproc.c
198
ip = to_iproc_pwmc(chip);
drivers/pwm/pwm-bcm-iproc.c
200
platform_set_drvdata(pdev, ip);
drivers/pwm/pwm-bcm-iproc.c
204
ip->base = devm_platform_ioremap_resource(pdev, 0);
drivers/pwm/pwm-bcm-iproc.c
205
if (IS_ERR(ip->base))
drivers/pwm/pwm-bcm-iproc.c
206
return PTR_ERR(ip->base);
drivers/pwm/pwm-bcm-iproc.c
208
ip->clk = devm_clk_get_enabled(&pdev->dev, NULL);
drivers/pwm/pwm-bcm-iproc.c
209
if (IS_ERR(ip->clk))
drivers/pwm/pwm-bcm-iproc.c
210
return dev_err_probe(&pdev->dev, PTR_ERR(ip->clk),
drivers/pwm/pwm-bcm-iproc.c
214
value = readl(ip->base + IPROC_PWM_CTRL_OFFSET);
drivers/pwm/pwm-bcm-iproc.c
221
writel(value, ip->base + IPROC_PWM_CTRL_OFFSET);
drivers/pwm/pwm-bcm-iproc.c
46
static void iproc_pwmc_enable(struct iproc_pwmc *ip, unsigned int channel)
drivers/pwm/pwm-bcm-iproc.c
50
value = readl(ip->base + IPROC_PWM_CTRL_OFFSET);
drivers/pwm/pwm-bcm-iproc.c
52
writel(value, ip->base + IPROC_PWM_CTRL_OFFSET);
drivers/pwm/pwm-bcm-iproc.c
58
static void iproc_pwmc_disable(struct iproc_pwmc *ip, unsigned int channel)
drivers/pwm/pwm-bcm-iproc.c
62
value = readl(ip->base + IPROC_PWM_CTRL_OFFSET);
drivers/pwm/pwm-bcm-iproc.c
64
writel(value, ip->base + IPROC_PWM_CTRL_OFFSET);
drivers/pwm/pwm-bcm-iproc.c
73
struct iproc_pwmc *ip = to_iproc_pwmc(chip);
drivers/pwm/pwm-bcm-iproc.c
77
value = readl(ip->base + IPROC_PWM_CTRL_OFFSET);
drivers/pwm/pwm-bcm-iproc.c
89
rate = clk_get_rate(ip->clk);
drivers/pwm/pwm-bcm-iproc.c
96
value = readl(ip->base + IPROC_PWM_PRESCALE_OFFSET);
drivers/regulator/qcom_rpm-regulator.c
112
.ip = { 1, 0x000003FF, 0 },
drivers/regulator/qcom_rpm-regulator.c
123
.ip = { 1, 0x000003FF, 0 },
drivers/regulator/qcom_rpm-regulator.c
29
struct request_member ip; /* peak current in mA */
drivers/regulator/qcom_rpm-regulator.c
68
.ip = { 0, 0x00FFF000, 12 },
drivers/regulator/qcom_rpm-regulator.c
79
.ip = { 0, 0x00FFF000, 12 },
drivers/s390/block/dasd_diag.c
240
addr_t ip;
drivers/s390/block/dasd_diag.c
245
ip = (addr_t) param32;
drivers/s390/block/dasd_diag.c
248
ip = (addr_t) param64;
drivers/s390/block/dasd_diag.c
254
if (!ip) { /* no intparm: unsolicited interrupt */
drivers/s390/block/dasd_diag.c
259
cqr = (struct dasd_ccw_req *) ip;
drivers/s390/char/sclp_tty.c
359
unsigned char *ip, *op;
drivers/s390/char/sclp_tty.c
364
ip = op = buf;
drivers/s390/char/sclp_tty.c
367
if (*ip == CASE_DELIMITER) {
drivers/s390/char/sclp_tty.c
369
if (count && ip[1] == CASE_DELIMITER) {
drivers/s390/char/sclp_tty.c
374
*op++ = *ip++;
drivers/s390/char/sclp_tty.c
383
ip++;
drivers/s390/char/sclp_tty.c
390
*op++ = _ebc_toupper[(int) *ip++];
drivers/s390/char/sclp_tty.c
393
*op++ = _ebc_tolower[(int) *ip++];
drivers/s390/char/sclp_tty.c
396
*op++ = *ip++;
drivers/s390/net/qeth_core_mpc.h
354
struct in6_addr ip;
drivers/s390/net/qeth_l3.h
114
int qeth_l3_modify_rxip_vipa(struct qeth_card *card, bool add, const u8 *ip,
drivers/s390/net/qeth_l3_main.c
1263
if (type->ip == QETHARP_IP_ADDR_V4) {
drivers/s390/net/qeth_l3_main.c
1272
} else if (type->ip == QETHARP_IP_ADDR_V6) {
drivers/s390/net/qeth_l3_main.c
1293
return (type->ip == QETHARP_IP_ADDR_V4 && prot == QETH_PROT_IPV4) ||
drivers/s390/net/qeth_l3_main.c
1294
(type->ip == QETHARP_IP_ADDR_V6 && prot == QETH_PROT_IPV6);
drivers/s390/net/qeth_l3_main.c
1339
QETH_CARD_TEXT_(card, 4, "%i", etype->ip);
drivers/s390/net/qeth_l3_main.c
330
cmd->data.setdelipm.ip = addr->u.a6.addr;
drivers/s390/net/qeth_l3_main.c
333
cmd->data.setdelipm.ip.s6_addr32[3] = addr->u.a4.addr;
drivers/s390/net/qeth_l3_main.c
591
int qeth_l3_modify_rxip_vipa(struct qeth_card *card, bool add, const u8 *ip,
drivers/s390/net/qeth_l3_main.c
599
memcpy(&addr.u.a4.addr, ip, 4);
drivers/s390/net/qeth_l3_main.c
601
memcpy(&addr.u.a6.addr, ip, 16);
drivers/scsi/advansys.c
7099
sector_t capacity, int ip[])
drivers/scsi/advansys.c
7108
ip[0] = 255;
drivers/scsi/advansys.c
7109
ip[1] = 63;
drivers/scsi/advansys.c
7111
ip[0] = 64;
drivers/scsi/advansys.c
7112
ip[1] = 32;
drivers/scsi/advansys.c
7117
ip[0] = 255;
drivers/scsi/advansys.c
7118
ip[1] = 63;
drivers/scsi/advansys.c
7120
ip[0] = 64;
drivers/scsi/advansys.c
7121
ip[1] = 32;
drivers/scsi/advansys.c
7124
ip[2] = (unsigned long)capacity / (ip[0] * ip[1]);
drivers/scsi/aha1740.c
514
sector_t capacity, int* ip)
drivers/scsi/aha1740.c
520
if (extended && (ip[2] > 1024)) {
drivers/scsi/aha1740.c
521
ip[0] = 255;
drivers/scsi/aha1740.c
522
ip[1] = 63;
drivers/scsi/aha1740.c
523
ip[2] = size / (255 * 63);
drivers/scsi/aha1740.c
525
ip[0] = 64;
drivers/scsi/aha1740.c
526
ip[1] = 32;
drivers/scsi/aha1740.c
527
ip[2] = size >> 11;
drivers/scsi/atp870u.c
1696
sector_t capacity, int *ip)
drivers/scsi/atp870u.c
1708
ip[0] = heads;
drivers/scsi/atp870u.c
1709
ip[1] = sectors;
drivers/scsi/atp870u.c
1710
ip[2] = cylinders;
drivers/scsi/be2iscsi/be_iscsi.c
344
u8 *ip = NULL, *subnet = NULL, *gw;
drivers/scsi/be2iscsi/be_iscsi.c
375
ip = info->value;
drivers/scsi/be2iscsi/be_iscsi.c
382
ip, subnet);
drivers/scsi/be2iscsi/be_iscsi.c
393
ip = info->value;
drivers/scsi/be2iscsi/be_iscsi.c
396
ip, subnet);
drivers/scsi/be2iscsi/be_mgmt.c
427
static inline bool beiscsi_if_zero_ip(u8 *ip, u32 ip_type)
drivers/scsi/be2iscsi/be_mgmt.c
432
while (len && !ip[len - 1])
drivers/scsi/be2iscsi/be_mgmt.c
552
beiscsi_if_set_ip(struct beiscsi_hba *phba, u8 *ip,
drivers/scsi/be2iscsi/be_mgmt.c
575
memcpy(req->ip_params.ip_record.ip_addr.addr, ip, ip_len);
drivers/scsi/be2iscsi/be_mgmt.c
597
u8 *ip, u8 *subnet)
drivers/scsi/be2iscsi/be_mgmt.c
637
if (ip)
drivers/scsi/be2iscsi/be_mgmt.c
638
rc = beiscsi_if_set_ip(phba, ip, subnet, ip_type);
drivers/scsi/be2iscsi/be_mgmt.h
175
u8 *ip, u8 *subnet);
drivers/scsi/bfa/bfa_fc.h
1410
u8 ip[16]; /* IP Address */
drivers/scsi/imm.c
958
sector_t capacity, int ip[])
drivers/scsi/imm.c
960
ip[0] = 0x40;
drivers/scsi/imm.c
961
ip[1] = 0x20;
drivers/scsi/imm.c
962
ip[2] = ((unsigned long) capacity + 1) / (ip[0] * ip[1]);
drivers/scsi/imm.c
963
if (ip[2] > 1024) {
drivers/scsi/imm.c
964
ip[0] = 0xff;
drivers/scsi/imm.c
965
ip[1] = 0x3f;
drivers/scsi/imm.c
966
ip[2] = ((unsigned long) capacity + 1) / (ip[0] * ip[1]);
drivers/scsi/iscsi_boot_sysfs.c
160
iscsi_boot_rd_attr(eth_ip, ip-addr, ISCSI_BOOT_ETH_IP_ADDR);
drivers/scsi/iscsi_boot_sysfs.c
78
iscsi_boot_rd_attr(tgt_ip, ip-addr, ISCSI_BOOT_TGT_IP_ADDR);
drivers/scsi/ppa.c
849
sector_t capacity, int ip[])
drivers/scsi/ppa.c
851
ip[0] = 0x40;
drivers/scsi/ppa.c
852
ip[1] = 0x20;
drivers/scsi/ppa.c
853
ip[2] = ((unsigned long) capacity + 1) / (ip[0] * ip[1]);
drivers/scsi/ppa.c
854
if (ip[2] > 1024) {
drivers/scsi/ppa.c
855
ip[0] = 0xff;
drivers/scsi/ppa.c
856
ip[1] = 0x3f;
drivers/scsi/ppa.c
857
ip[2] = ((unsigned long) capacity + 1) / (ip[0] * ip[1]);
drivers/scsi/ppa.c
858
if (ip[2] > 1023)
drivers/scsi/ppa.c
859
ip[2] = 1023;
drivers/scsi/qedi/qedi_iscsi.c
518
conn_info->src.ip[0] = ntohl(qedi_ep->src_addr[0]);
drivers/scsi/qedi/qedi_iscsi.c
519
conn_info->dst.ip[0] = ntohl(qedi_ep->dst_addr[0]);
drivers/scsi/qedi/qedi_iscsi.c
528
conn_info->src.ip[i] = ntohl(qedi_ep->src_addr[i]);
drivers/scsi/qedi/qedi_iscsi.c
529
conn_info->dst.ip[i] = ntohl(qedi_ep->dst_addr[i]);
drivers/scsi/qedi/qedi_main.c
2041
char *fmt, *ip, *sub, *gw;
drivers/scsi/qedi/qedi_main.c
2054
ip = ipv6_en ? initiator->ipv6.addr.byte : initiator->ipv4.addr.byte;
drivers/scsi/qedi/qedi_main.c
2063
ip = ipv6_en ? "0::0" : "0.0.0.0";
drivers/scsi/qedi/qedi_main.c
2064
sub = ip;
drivers/scsi/qedi/qedi_main.c
2065
gw = ip;
drivers/scsi/qedi/qedi_main.c
2071
rc = snprintf(buf, ip_len, fmt, ip);
drivers/scsi/qla4xxx/ql4_mbx.c
2000
char *ip;
drivers/scsi/qla4xxx/ql4_mbx.c
2052
ip = (char *)&addr->sin_addr;
drivers/scsi/qla4xxx/ql4_mbx.c
2053
memcpy(fw_ddb_entry->ip_addr, ip, IP_ADDR_LEN);
drivers/scsi/qla4xxx/ql4_mbx.c
2061
ip = (char *)&addr6->sin6_addr;
drivers/scsi/qla4xxx/ql4_mbx.c
2062
memcpy(fw_ddb_entry->ip_addr, ip, IPv6_ADDR_LEN);
drivers/scsi/qla4xxx/ql4_os.c
6585
char *ip;
drivers/scsi/qla4xxx/ql4_os.c
6596
ip = (char *)&addr6->sin6_addr;
drivers/scsi/qla4xxx/ql4_os.c
6597
memcpy(ip, fw_ddb_entry->ip_addr, IPv6_ADDR_LEN);
drivers/scsi/qla4xxx/ql4_os.c
6604
ip = (char *)&addr->sin_addr;
drivers/scsi/qla4xxx/ql4_os.c
6605
memcpy(ip, fw_ddb_entry->ip_addr, IP_ADDR_LEN);
drivers/scsi/qlogicfas408.c
496
sector_t capacity, int ip[])
drivers/scsi/qlogicfas408.c
499
ip[0] = 0x40;
drivers/scsi/qlogicfas408.c
500
ip[1] = 0x20;
drivers/scsi/qlogicfas408.c
501
ip[2] = (unsigned long) capacity / (ip[0] * ip[1]);
drivers/scsi/qlogicfas408.c
502
if (ip[2] > 1024) {
drivers/scsi/qlogicfas408.c
503
ip[0] = 0xff;
drivers/scsi/qlogicfas408.c
504
ip[1] = 0x3f;
drivers/scsi/qlogicfas408.c
505
ip[2] = (unsigned long) capacity / (ip[0] * ip[1]);
drivers/scsi/qlogicfas408.c
507
if (ip[2] > 1023)
drivers/scsi/qlogicfas408.c
508
ip[2] = 1023;
drivers/scsi/qlogicfas408.h
111
sector_t capacity, int ip[]);
drivers/scsi/scsicam.c
218
int scsicam_bios_param(struct gendisk *disk, sector_t capacity, int *ip)
drivers/scsi/scsicam.c
224
if (scsi_partsize(disk, capacity, ip))
drivers/scsi/scsicam.c
232
ret = setsize((unsigned long)capacity, (unsigned int *)ip + 2,
drivers/scsi/scsicam.c
233
(unsigned int *)ip + 0, (unsigned int *)ip + 1);
drivers/scsi/scsicam.c
240
if (ret || ip[0] > 255 || ip[1] > 63) {
drivers/scsi/scsicam.c
242
ip[0] = 255;
drivers/scsi/scsicam.c
243
ip[1] = 63;
drivers/scsi/scsicam.c
245
ip[0] = 64;
drivers/scsi/scsicam.c
246
ip[1] = 32;
drivers/scsi/scsicam.c
250
ip[2] = 65535;
drivers/scsi/scsicam.c
252
ip[2] = (unsigned long)capacity / (ip[0] * ip[1]);
drivers/scsi/sg.c
1004
return put_user(srp->header.pack_id, ip);
drivers/scsi/sg.c
1008
return put_user(-1, ip);
drivers/scsi/sg.c
1017
return put_user(val, ip);
drivers/scsi/sg.c
1019
return put_user(sdp->sg_tablesize, ip);
drivers/scsi/sg.c
1021
result = get_user(val, ip);
drivers/scsi/sg.c
1044
return put_user(val, ip);
drivers/scsi/sg.c
1046
result = get_user(val, ip);
drivers/scsi/sg.c
1052
return put_user((int) sfp->cmd_q, ip);
drivers/scsi/sg.c
1054
result = get_user(val, ip);
drivers/scsi/sg.c
1060
return put_user((int) sfp->keep_orphan, ip);
drivers/scsi/sg.c
1062
result = get_user(val, ip);
drivers/scsi/sg.c
1070
return put_user(sg_version_num, ip);
drivers/scsi/sg.c
1074
return put_user(val, ip);
drivers/scsi/sg.c
1100
return put_user(sdp->device->host->hostt->emulated, ip);
drivers/scsi/sg.c
1107
result = get_user(val, ip);
drivers/scsi/sg.c
1114
ip);
drivers/scsi/sg.c
918
int __user *ip = p;
drivers/scsi/sg.c
950
result = get_user(val, ip);
drivers/scsi/sg.c
973
return put_user(0, ip);
drivers/scsi/sg.c
993
result = get_user(val, ip);
drivers/spi/spi-dw-mmio.c
222
dwsmmio->dws.ip = DW_HSSI_ID;
drivers/spi/spi-dw-mmio.c
232
dwsmmio->dws.ip = DW_HSSI_ID;
drivers/spi/spi-dw.h
154
u32 ip; /* Synopsys DW SSI IP-core ID */
drivers/spi/spi-dw.h
23
((_dws)->ip == DW_ ## _ip ## _ID)
drivers/spi/spi-sifive.c
211
u32 ip = sifive_spi_read(spi, SIFIVE_SPI_REG_IP);
drivers/spi/spi-sifive.c
213
if (ip & (SIFIVE_SPI_IP_TXWM | SIFIVE_SPI_IP_RXWM)) {
drivers/tee/tee_core.c
364
struct tee_ioctl_param *ip)
drivers/tee/tee_core.c
374
if (ip->c != TEE_MEMREF_NULL) {
drivers/tee/tee_core.c
384
shm = tee_shm_get_from_id(ctx, ip->c);
drivers/tee/tee_core.c
393
if ((ip->a + ip->b) < ip->a ||
drivers/tee/tee_core.c
394
(ip->a + ip->b) > shm->size) {
drivers/tee/tee_core.c
425
memref->shm_offs = ip->a + offs;
drivers/tee/tee_core.c
426
memref->size = ip->b;
drivers/tee/tee_core.c
439
struct tee_ioctl_param ip;
drivers/tee/tee_core.c
442
if (copy_from_user(&ip, uparams + n, sizeof(ip)))
drivers/tee/tee_core.c
446
if (ip.attr & ~TEE_IOCTL_PARAM_ATTR_MASK)
drivers/tee/tee_core.c
449
params[n].attr = ip.attr;
drivers/tee/tee_core.c
450
switch (ip.attr & TEE_IOCTL_PARAM_ATTR_TYPE_MASK) {
drivers/tee/tee_core.c
457
params[n].u.value.a = ip.a;
drivers/tee/tee_core.c
458
params[n].u.value.b = ip.b;
drivers/tee/tee_core.c
459
params[n].u.value.c = ip.c;
drivers/tee/tee_core.c
464
params[n].u.ubuf.uaddr = u64_to_user_ptr(ip.a);
drivers/tee/tee_core.c
465
params[n].u.ubuf.size = ip.b;
drivers/tee/tee_core.c
474
params[n].u.objref.id = ip.a;
drivers/tee/tee_core.c
475
params[n].u.objref.flags = ip.b;
drivers/tee/tee_core.c
481
&ip);
drivers/tee/tee_core.c
767
struct tee_ioctl_param ip;
drivers/tee/tee_core.c
770
ip.attr = p->attr;
drivers/tee/tee_core.c
774
ip.a = p->u.value.a;
drivers/tee/tee_core.c
775
ip.b = p->u.value.b;
drivers/tee/tee_core.c
776
ip.c = p->u.value.c;
drivers/tee/tee_core.c
781
ip.a = (__force unsigned long)p->u.ubuf.uaddr;
drivers/tee/tee_core.c
782
ip.b = p->u.ubuf.size;
drivers/tee/tee_core.c
783
ip.c = 0;
drivers/tee/tee_core.c
787
ip.a = p->u.objref.id;
drivers/tee/tee_core.c
788
ip.b = p->u.objref.flags;
drivers/tee/tee_core.c
789
ip.c = 0;
drivers/tee/tee_core.c
794
ip.b = p->u.memref.size;
drivers/tee/tee_core.c
796
ip.a = 0;
drivers/tee/tee_core.c
797
ip.c = (u64)-1; /* invalid shm id */
drivers/tee/tee_core.c
800
ip.a = p->u.memref.shm_offs;
drivers/tee/tee_core.c
801
ip.c = p->u.memref.shm->id;
drivers/tee/tee_core.c
804
ip.a = 0;
drivers/tee/tee_core.c
805
ip.b = 0;
drivers/tee/tee_core.c
806
ip.c = 0;
drivers/tee/tee_core.c
810
if (copy_to_user(uparams + n, &ip, sizeof(ip)))
drivers/tee/tee_core.c
875
struct tee_ioctl_param ip;
drivers/tee/tee_core.c
877
if (copy_from_user(&ip, uparams + n, sizeof(ip)))
drivers/tee/tee_core.c
881
if (ip.attr & ~TEE_IOCTL_PARAM_ATTR_MASK)
drivers/tee/tee_core.c
884
p->attr = ip.attr;
drivers/tee/tee_core.c
885
switch (ip.attr & TEE_IOCTL_PARAM_ATTR_TYPE_MASK) {
drivers/tee/tee_core.c
889
p->u.value.a = ip.a;
drivers/tee/tee_core.c
890
p->u.value.b = ip.b;
drivers/tee/tee_core.c
891
p->u.value.c = ip.c;
drivers/tee/tee_core.c
895
p->u.ubuf.uaddr = u64_to_user_ptr(ip.a);
drivers/tee/tee_core.c
896
p->u.ubuf.size = ip.b;
drivers/tee/tee_core.c
905
p->u.objref.id = ip.a;
drivers/tee/tee_core.c
906
p->u.objref.flags = ip.b;
drivers/tee/tee_core.c
919
p->u.memref.size = ip.b;
drivers/tty/moxa.c
1414
u16 __iomem *ip)
drivers/tty/moxa.c
1446
intr = readw(ip); /* port irq status */
drivers/tty/moxa.c
1450
writew(0, ip); /* ACK port */
drivers/tty/moxa.c
1475
u16 __iomem *ip;
drivers/tty/moxa.c
1486
ip = NULL;
drivers/tty/moxa.c
1488
ip = brd->intTable + readb(brd->intNdx);
drivers/tty/moxa.c
1491
moxa_poll_port(&brd->ports[port], !!ip, ip + port);
drivers/tty/moxa.c
1493
if (ip)
drivers/tty/serial/sifive.c
525
u32 ip;
drivers/tty/serial/sifive.c
529
ip = __ssp_readl(ssp, SIFIVE_SERIAL_IP_OFFS);
drivers/tty/serial/sifive.c
530
if (!ip) {
drivers/tty/serial/sifive.c
535
if (ip & SIFIVE_SERIAL_IP_RXWM_MASK)
drivers/tty/serial/sifive.c
537
if (ip & SIFIVE_SERIAL_IP_TXWM_MASK)
drivers/usb/dwc3/core.c
995
dwc->ip = DWC3_GSNPS_ID(reg);
drivers/usb/dwc3/core.c
996
if (dwc->ip == DWC4_IP)
drivers/usb/dwc3/core.c
997
dwc->ip = DWC32_IP;
drivers/usb/dwc3/core.h
1265
u32 ip;
drivers/usb/dwc3/core.h
1589
(dwc->ip == _ip##_IP)
drivers/video/fbdev/imsttfb.c
1355
__u32 i, tmp, *ip, *end;
drivers/video/fbdev/imsttfb.c
1363
ip = (__u32 *)info->screen_base;
drivers/video/fbdev/imsttfb.c
1365
while (ip < end)
drivers/video/fbdev/imsttfb.c
1366
*ip++ = 0;
fs/fat/dir.c
147
const wchar_t *ip;
fs/fat/dir.c
152
ip = uni;
fs/fat/dir.c
155
while (*ip && ((len - NLS_MAX_CHARSET_SIZE) > 0)) {
fs/fat/dir.c
156
ec = *ip++;
fs/fat/dir.c
174
if (unlikely(*ip)) {
fs/fat/namei_vfat.c
330
wchar_t *ip, *ext_start, *end, *name_start;
fs/fat/namei_vfat.c
380
for (baselen = i = 0, p = base, ip = uname; i < sz; i++, ip++) {
fs/fat/namei_vfat.c
382
ip, &base_info);
fs/fat/namei_vfat.c
397
if ((chi < chl - 1) || (ip + 1) - uname < sz)
fs/fat/namei_vfat.c
408
for (p = ext, ip = ext_start; extlen < 3 && ip < end; ip++) {
fs/fat/namei_vfat.c
410
ip, &ext_info);
fs/fat/namei_vfat.c
423
if (ip + 1 != end)
fs/fat/namei_vfat.c
513
const unsigned char *ip;
fs/fat/namei_vfat.c
528
for (i = 0, ip = name, op = outname, *outlen = 0;
fs/fat/namei_vfat.c
531
if (escape && (*ip == ':')) {
fs/fat/namei_vfat.c
537
if (hex2bin(uc, ip + 1, 2) < 0)
fs/fat/namei_vfat.c
543
ip += 5;
fs/fat/namei_vfat.c
546
charlen = nls->char2uni(ip, len - i,
fs/fat/namei_vfat.c
550
ip += charlen;
fs/freevxfs/vxfs.h
221
#define VXFS_IS_TYPE(ip,type) (((ip)->vii_mode & VXFS_TYPE_MASK) == (type))
fs/freevxfs/vxfs.h
245
#define VXFS_IS_ORG(ip,org) ((ip)->vii_orgtype == (org))
fs/freevxfs/vxfs.h
246
#define VXFS_ISNONE(ip) VXFS_IS_ORG((ip), VXFS_ORG_NONE)
fs/freevxfs/vxfs.h
247
#define VXFS_ISEXT4(ip) VXFS_IS_ORG((ip), VXFS_ORG_EXT4)
fs/freevxfs/vxfs.h
248
#define VXFS_ISIMMED(ip) VXFS_IS_ORG((ip), VXFS_ORG_IMMED)
fs/freevxfs/vxfs.h
249
#define VXFS_ISTYPED(ip) VXFS_IS_ORG((ip), VXFS_ORG_TYPED)
fs/freevxfs/vxfs_bmap.c
107
vxfs_bmap_indir(struct inode *ip, long indir, int size, long block)
fs/freevxfs/vxfs_bmap.c
109
struct vxfs_sb_info *sbi = VXFS_SBI(ip->i_sb);
fs/freevxfs/vxfs_bmap.c
114
for (i = 0; i < size * VXFS_TYPED_PER_BLOCK(ip->i_sb); i++) {
fs/freevxfs/vxfs_bmap.c
118
bp = sb_bread(ip->i_sb,
fs/freevxfs/vxfs_bmap.c
119
indir + (i / VXFS_TYPED_PER_BLOCK(ip->i_sb)));
fs/freevxfs/vxfs_bmap.c
124
(i % VXFS_TYPED_PER_BLOCK(ip->i_sb));
fs/freevxfs/vxfs_bmap.c
135
pblock = vxfs_bmap_indir(ip,
fs/freevxfs/vxfs_bmap.c
186
vxfs_bmap_typed(struct inode *ip, long iblock)
fs/freevxfs/vxfs_bmap.c
188
struct vxfs_inode_info *vip = VXFS_INO(ip);
fs/freevxfs/vxfs_bmap.c
189
struct vxfs_sb_info *sbi = VXFS_SBI(ip->i_sb);
fs/freevxfs/vxfs_bmap.c
205
pblock = vxfs_bmap_indir(ip,
fs/freevxfs/vxfs_bmap.c
250
vxfs_bmap1(struct inode *ip, long iblock)
fs/freevxfs/vxfs_bmap.c
252
struct vxfs_inode_info *vip = VXFS_INO(ip);
fs/freevxfs/vxfs_bmap.c
255
return vxfs_bmap_ext4(ip, iblock);
fs/freevxfs/vxfs_bmap.c
257
return vxfs_bmap_typed(ip, iblock);
fs/freevxfs/vxfs_bmap.c
264
ip->i_ino, vip->vii_orgtype);
fs/freevxfs/vxfs_bmap.c
269
ip->i_ino, vip->vii_orgtype);
fs/freevxfs/vxfs_bmap.c
43
vxfs_bmap_ext4(struct inode *ip, long bn)
fs/freevxfs/vxfs_bmap.c
45
struct super_block *sb = ip->i_sb;
fs/freevxfs/vxfs_bmap.c
46
struct vxfs_inode_info *vip = VXFS_INO(ip);
fs/freevxfs/vxfs_fshead.c
53
vxfs_getfsh(struct inode *ip, int which)
fs/freevxfs/vxfs_fshead.c
57
bp = vxfs_bread(ip, which);
fs/freevxfs/vxfs_inode.c
255
struct inode *ip;
fs/freevxfs/vxfs_inode.c
258
ip = iget_locked(sbp, ino);
fs/freevxfs/vxfs_inode.c
259
if (!ip)
fs/freevxfs/vxfs_inode.c
261
if (!(inode_state_read_once(ip) & I_NEW))
fs/freevxfs/vxfs_inode.c
262
return ip;
fs/freevxfs/vxfs_inode.c
264
vip = VXFS_INO(ip);
fs/freevxfs/vxfs_inode.c
267
iget_failed(ip);
fs/freevxfs/vxfs_inode.c
276
if (S_ISREG(ip->i_mode)) {
fs/freevxfs/vxfs_inode.c
277
ip->i_fop = &generic_ro_fops;
fs/freevxfs/vxfs_inode.c
278
ip->i_mapping->a_ops = aops;
fs/freevxfs/vxfs_inode.c
279
} else if (S_ISDIR(ip->i_mode)) {
fs/freevxfs/vxfs_inode.c
280
ip->i_op = &vxfs_dir_inode_ops;
fs/freevxfs/vxfs_inode.c
281
ip->i_fop = &vxfs_dir_operations;
fs/freevxfs/vxfs_inode.c
282
ip->i_mapping->a_ops = aops;
fs/freevxfs/vxfs_inode.c
283
} else if (S_ISLNK(ip->i_mode)) {
fs/freevxfs/vxfs_inode.c
285
ip->i_op = &page_symlink_inode_operations;
fs/freevxfs/vxfs_inode.c
286
inode_nohighmem(ip);
fs/freevxfs/vxfs_inode.c
287
ip->i_mapping->a_ops = &vxfs_aops;
fs/freevxfs/vxfs_inode.c
289
ip->i_op = &simple_symlink_inode_operations;
fs/freevxfs/vxfs_inode.c
290
ip->i_link = vip->vii_immed.vi_immed;
fs/freevxfs/vxfs_inode.c
291
nd_terminate_link(ip->i_link, ip->i_size,
fs/freevxfs/vxfs_inode.c
295
init_special_inode(ip, ip->i_mode, old_decode_dev(vip->vii_rdev));
fs/freevxfs/vxfs_inode.c
297
unlock_new_inode(ip);
fs/freevxfs/vxfs_inode.c
298
return ip;
fs/freevxfs/vxfs_inode.c
310
vxfs_evict_inode(struct inode *ip)
fs/freevxfs/vxfs_inode.c
312
truncate_inode_pages_final(&ip->i_data);
fs/freevxfs/vxfs_inode.c
313
clear_inode(ip);
fs/freevxfs/vxfs_lookup.c
167
struct inode *ip = NULL;
fs/freevxfs/vxfs_lookup.c
175
ip = vxfs_iget(dip->i_sb, ino);
fs/freevxfs/vxfs_lookup.c
176
return d_splice_alias(ip, dp);
fs/freevxfs/vxfs_lookup.c
194
struct inode *ip = file_inode(fp);
fs/freevxfs/vxfs_lookup.c
195
struct super_block *sbp = ip->i_sb;
fs/freevxfs/vxfs_lookup.c
206
if (!dir_emit(ctx, "..", 2, VXFS_INO(ip)->vii_dotdot, DT_DIR))
fs/freevxfs/vxfs_lookup.c
211
limit = VXFS_DIRROUND(ip->i_size);
fs/freevxfs/vxfs_lookup.c
223
pp = vxfs_get_page(ip->i_mapping, pos >> PAGE_SHIFT);
fs/freevxfs/vxfs_lookup.c
59
vxfs_find_entry(struct inode *ip, struct dentry *dp, struct page **ppp)
fs/freevxfs/vxfs_lookup.c
61
u_long bsize = ip->i_sb->s_blocksize;
fs/freevxfs/vxfs_lookup.c
64
loff_t limit = VXFS_DIRROUND(ip->i_size);
fs/freevxfs/vxfs_lookup.c
67
struct vxfs_sb_info *sbi = VXFS_SBI(ip->i_sb);
fs/freevxfs/vxfs_lookup.c
74
pp = vxfs_get_page(ip->i_mapping, pos >> PAGE_SHIFT);
fs/freevxfs/vxfs_subr.c
104
pblock = vxfs_bmap1(ip, iblock);
fs/freevxfs/vxfs_subr.c
106
map_bh(bp, ip->i_sb, pblock);
fs/freevxfs/vxfs_subr.c
72
vxfs_bread(struct inode *ip, int block)
fs/freevxfs/vxfs_subr.c
77
pblock = vxfs_bmap1(ip, block);
fs/freevxfs/vxfs_subr.c
78
bp = sb_bread(ip->i_sb, pblock);
fs/freevxfs/vxfs_subr.c
99
vxfs_getblk(struct inode *ip, sector_t iblock,
fs/gfs2/acl.c
109
struct gfs2_inode *ip = GFS2_I(inode);
fs/gfs2/acl.c
118
ret = gfs2_qa_get(ip);
fs/gfs2/acl.c
122
if (!gfs2_glock_is_locked_by_me(ip->i_gl)) {
fs/gfs2/acl.c
123
ret = gfs2_glock_nq_init(ip->i_gl, LM_ST_EXCLUSIVE, 0, &gh);
fs/gfs2/acl.c
146
gfs2_qa_put(ip);
fs/gfs2/acl.c
42
struct gfs2_inode *ip = GFS2_I(inode);
fs/gfs2/acl.c
48
if (!ip->i_eattr)
fs/gfs2/acl.c
52
len = gfs2_xattr_acl_get(ip, name, &data);
fs/gfs2/acl.c
62
struct gfs2_inode *ip = GFS2_I(inode);
fs/gfs2/acl.c
70
if (!gfs2_glock_is_locked_by_me(ip->i_gl)) {
fs/gfs2/acl.c
71
int ret = gfs2_glock_nq_init(ip->i_gl, LM_ST_SHARED,
fs/gfs2/aops.c
105
struct gfs2_inode *ip = GFS2_I(inode);
fs/gfs2/aops.c
114
gfs2_trans_add_databufs(ip->i_gl, folio, 0, folio_size(folio));
fs/gfs2/aops.c
129
struct gfs2_inode *ip = GFS2_I(inode);
fs/gfs2/aops.c
135
if (gfs2_assert_withdraw(sdp, ip->i_gl->gl_state == LM_ST_EXCLUSIVE))
fs/gfs2/aops.c
364
struct gfs2_inode *ip = GFS2_I(mapping->host);
fs/gfs2/aops.c
370
gfs2_log_flush(sdp, ip->i_gl, GFS2_LOG_HEAD_FLUSH_NORMAL |
fs/gfs2/aops.c
384
static int stuffed_read_folio(struct gfs2_inode *ip, struct folio *folio)
fs/gfs2/aops.c
387
size_t dsize = i_size_read(&ip->i_inode);
fs/gfs2/aops.c
399
error = gfs2_meta_inode_buffer(ip, &dibh);
fs/gfs2/aops.c
421
struct gfs2_inode *ip = GFS2_I(inode);
fs/gfs2/aops.c
425
if (!gfs2_is_jdata(ip) ||
fs/gfs2/aops.c
428
} else if (gfs2_is_stuffed(ip)) {
fs/gfs2/aops.c
429
error = stuffed_read_folio(ip, folio);
fs/gfs2/aops.c
449
ssize_t gfs2_internal_read(struct gfs2_inode *ip, char *buf, loff_t *pos,
fs/gfs2/aops.c
452
struct address_space *mapping = ip->i_inode.i_mapping;
fs/gfs2/aops.c
495
struct gfs2_inode *ip = GFS2_I(inode);
fs/gfs2/aops.c
497
if (gfs2_is_stuffed(ip))
fs/gfs2/aops.c
499
else if (gfs2_is_jdata(ip))
fs/gfs2/aops.c
563
struct gfs2_inode *ip = GFS2_I(mapping->host);
fs/gfs2/aops.c
568
error = gfs2_glock_nq_init(ip->i_gl, LM_ST_SHARED, LM_FLAG_ANY, &i_gh);
fs/gfs2/aops.c
572
if (!gfs2_is_stuffed(ip))
fs/gfs2/bmap.c
1002
if (gfs2_is_jdata(ip) || gfs2_is_stuffed(ip)) {
fs/gfs2/bmap.c
1019
struct gfs2_inode *ip = GFS2_I(inode);
fs/gfs2/bmap.c
102
if (i_size_read(&ip->i_inode)) {
fs/gfs2/bmap.c
1024
unstuff = gfs2_is_stuffed(ip) &&
fs/gfs2/bmap.c
1025
pos + length > gfs2_max_stuffed_size(ip);
fs/gfs2/bmap.c
1033
gfs2_write_calc_reserv(ip, iomap->length, &data_blocks,
fs/gfs2/bmap.c
1036
ret = gfs2_quota_lock_check(ip, &ap);
fs/gfs2/bmap.c
1040
ret = gfs2_inplace_reserve(ip, &ap);
fs/gfs2/bmap.c
1045
if (gfs2_is_jdata(ip))
fs/gfs2/bmap.c
1051
rblocks += gfs2_rg_blocks(ip, data_blocks + ind_blocks);
fs/gfs2/bmap.c
1059
ret = gfs2_unstuff_dinode(ip);
fs/gfs2/bmap.c
107
error = gfs2_alloc_blocks(ip, &block, &n, 0);
fs/gfs2/bmap.c
1073
gfs2_inplace_release(ip);
fs/gfs2/bmap.c
1074
punch_hole(ip, iomap->offset, iomap->length);
fs/gfs2/bmap.c
1091
gfs2_inplace_release(ip);
fs/gfs2/bmap.c
1093
gfs2_quota_unlock(ip);
fs/gfs2/bmap.c
1101
struct gfs2_inode *ip = GFS2_I(inode);
fs/gfs2/bmap.c
1105
if (gfs2_is_jdata(ip))
fs/gfs2/bmap.c
1108
trace_gfs2_iomap_start(ip, pos, length, flags);
fs/gfs2/bmap.c
111
gfs2_trans_remove_revoke(GFS2_SB(&ip->i_inode), block, 1);
fs/gfs2/bmap.c
112
error = gfs2_dir_get_new_buffer(ip, block, &bh);
fs/gfs2/bmap.c
1145
trace_gfs2_iomap_end(ip, iomap, ret);
fs/gfs2/bmap.c
1152
struct gfs2_inode *ip = GFS2_I(inode);
fs/gfs2/bmap.c
1171
if (!gfs2_is_stuffed(ip))
fs/gfs2/bmap.c
1172
gfs2_ordered_add_inode(ip);
fs/gfs2/bmap.c
1177
gfs2_inplace_release(ip);
fs/gfs2/bmap.c
1179
if (ip->i_qadata && ip->i_qadata->qa_qd_num)
fs/gfs2/bmap.c
1180
gfs2_quota_unlock(ip);
fs/gfs2/bmap.c
1189
punch_hole(ip, hstart, hend - hstart);
fs/gfs2/bmap.c
119
error = gfs2_unstuffer_folio(ip, dibh, block, folio);
fs/gfs2/bmap.c
1198
set_bit(GLF_DIRTY, &ip->i_gl->gl_flags);
fs/gfs2/bmap.c
1231
struct gfs2_inode *ip = GFS2_I(inode);
fs/gfs2/bmap.c
1240
trace_gfs2_bmap(ip, bh_map, lblock, create, 1);
fs/gfs2/bmap.c
1262
trace_gfs2_bmap(ip, bh_map, lblock, create, ret);
fs/gfs2/bmap.c
127
gfs2_trans_add_meta(ip->i_gl, dibh);
fs/gfs2/bmap.c
131
if (i_size_read(&ip->i_inode)) {
fs/gfs2/bmap.c
133
gfs2_add_inode_blocks(&ip->i_inode, 1);
fs/gfs2/bmap.c
134
di->di_blocks = cpu_to_be64(gfs2_get_inode_blocks(&ip->i_inode));
fs/gfs2/bmap.c
137
ip->i_height = 1;
fs/gfs2/bmap.c
1374
struct gfs2_inode *ip = GFS2_I(inode);
fs/gfs2/bmap.c
1377
int journaled = gfs2_is_jdata(ip);
fs/gfs2/bmap.c
1381
if (!gfs2_is_stuffed(ip)) {
fs/gfs2/bmap.c
1398
error = gfs2_meta_inode_buffer(ip, &dibh);
fs/gfs2/bmap.c
1402
gfs2_trans_add_meta(ip->i_gl, dibh);
fs/gfs2/bmap.c
1404
if (gfs2_is_stuffed(ip))
fs/gfs2/bmap.c
1407
ip->i_diskflags |= GFS2_DIF_TRUNC_IN_PROG;
fs/gfs2/bmap.c
1410
inode_set_mtime_to_ts(&ip->i_inode, inode_set_ctime_current(&ip->i_inode));
fs/gfs2/bmap.c
1411
gfs2_dinode_out(ip, dibh->b_data);
fs/gfs2/bmap.c
1471
static int sweep_bh_for_rgrps(struct gfs2_inode *ip, struct gfs2_holder *rd_gh,
fs/gfs2/bmap.c
1475
struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode);
fs/gfs2/bmap.c
1518
if (gfs2_rs_active(&ip->i_res) &&
fs/gfs2/bmap.c
1519
rgd == ip->i_res.rs_rgd)
fs/gfs2/bmap.c
1520
gfs2_rs_deltree(&ip->i_res);
fs/gfs2/bmap.c
1533
isize_blks = gfs2_get_inode_blocks(&ip->i_inode);
fs/gfs2/bmap.c
1542
else if (ip->i_depth)
fs/gfs2/bmap.c
1547
down_write(&ip->i_rw_mutex);
fs/gfs2/bmap.c
155
int gfs2_unstuff_dinode(struct gfs2_inode *ip)
fs/gfs2/bmap.c
157
struct inode *inode = &ip->i_inode;
fs/gfs2/bmap.c
1574
gfs2_trans_add_meta(ip->i_gl, bh);
fs/gfs2/bmap.c
1582
__gfs2_free_blocks(ip, rgd, bstart, (u32)blen, meta);
fs/gfs2/bmap.c
1584
gfs2_add_inode_blocks(&ip->i_inode, -blen);
fs/gfs2/bmap.c
1590
__gfs2_free_blocks(ip, rgd, bstart, (u32)blen, meta);
fs/gfs2/bmap.c
1592
gfs2_add_inode_blocks(&ip->i_inode, -blen);
fs/gfs2/bmap.c
1601
ret = gfs2_meta_inode_buffer(ip, &dibh);
fs/gfs2/bmap.c
1607
inode_set_mtime_to_ts(&ip->i_inode, inode_set_ctime_current(&ip->i_inode));
fs/gfs2/bmap.c
1608
gfs2_trans_add_meta(ip->i_gl, dibh);
fs/gfs2/bmap.c
1609
gfs2_dinode_out(ip, dibh->b_data);
fs/gfs2/bmap.c
161
down_write(&ip->i_rw_mutex);
fs/gfs2/bmap.c
1611
up_write(&ip->i_rw_mutex);
fs/gfs2/bmap.c
166
error = __gfs2_unstuff_inode(ip, folio);
fs/gfs2/bmap.c
170
up_write(&ip->i_rw_mutex);
fs/gfs2/bmap.c
1731
static int punch_hole(struct gfs2_inode *ip, u64 offset, u64 length)
fs/gfs2/bmap.c
1733
struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode);
fs/gfs2/bmap.c
1734
u64 maxsize = sdp->sd_heightsize[ip->i_height];
fs/gfs2/bmap.c
1744
unsigned int strip_h = ip->i_height - 1;
fs/gfs2/bmap.c
1787
find_metapath(sdp, lend, &mp, ip->i_height);
fs/gfs2/bmap.c
1791
for (mp_h = ip->i_height - 1; mp_h > 0; mp_h--) {
fs/gfs2/bmap.c
1798
find_metapath(sdp, lblock, &mp, ip->i_height);
fs/gfs2/bmap.c
1801
for (mp_h = ip->i_height - 1; mp_h > 0; mp_h--) {
fs/gfs2/bmap.c
1807
ret = gfs2_meta_inode_buffer(ip, &dibh);
fs/gfs2/bmap.c
1812
ret = lookup_metapath(ip, &mp);
fs/gfs2/bmap.c
1820
gfs2_metapath_ra(ip->i_gl, start, end);
fs/gfs2/bmap.c
1823
if (mp.mp_aheight == ip->i_height)
fs/gfs2/bmap.c
1832
ret = gfs2_quota_hold(ip, NO_UID_QUOTA_CHANGE, NO_GID_QUOTA_CHANGE);
fs/gfs2/bmap.c
1850
(unsigned long long)ip->i_no_addr,
fs/gfs2/bmap.c
1851
prev_bnr, ip->i_height, strip_h, mp_h);
fs/gfs2/bmap.c
1871
ret = sweep_bh_for_rgrps(ip, &rd_gh, mp.mp_bh[mp_h],
fs/gfs2/bmap.c
1873
mp_h != ip->i_height - 1,
fs/gfs2/bmap.c
1923
ret = fillup_metapath(ip, &mp, mp_h);
fs/gfs2/bmap.c
1928
if (mp.mp_aheight > 1 && strip_h == ip->i_height - 1) {
fs/gfs2/bmap.c
1940
gfs2_metapath_ra(ip->i_gl, start, end);
fs/gfs2/bmap.c
1949
if (mp.mp_aheight < ip->i_height) /* We have a partial height */
fs/gfs2/bmap.c
1969
down_write(&ip->i_rw_mutex);
fs/gfs2/bmap.c
1972
gfs2_quota_change(ip, -(s64)btotal, ip->i_inode.i_uid,
fs/gfs2/bmap.c
1973
ip->i_inode.i_gid);
fs/gfs2/bmap.c
1974
inode_set_mtime_to_ts(&ip->i_inode, inode_set_ctime_current(&ip->i_inode));
fs/gfs2/bmap.c
1975
gfs2_trans_add_meta(ip->i_gl, dibh);
fs/gfs2/bmap.c
1976
gfs2_dinode_out(ip, dibh->b_data);
fs/gfs2/bmap.c
1977
up_write(&ip->i_rw_mutex);
fs/gfs2/bmap.c
1985
up_write(&ip->i_rw_mutex);
fs/gfs2/bmap.c
1989
gfs2_quota_unhold(ip);
fs/gfs2/bmap.c
1995
static int trunc_end(struct gfs2_inode *ip)
fs/gfs2/bmap.c
1997
struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode);
fs/gfs2/bmap.c
2005
down_write(&ip->i_rw_mutex);
fs/gfs2/bmap.c
2007
error = gfs2_meta_inode_buffer(ip, &dibh);
fs/gfs2/bmap.c
2011
if (!i_size_read(&ip->i_inode)) {
fs/gfs2/bmap.c
2012
ip->i_height = 0;
fs/gfs2/bmap.c
2013
ip->i_goal = ip->i_no_addr;
fs/gfs2/bmap.c
2015
gfs2_ordered_del_inode(ip);
fs/gfs2/bmap.c
2017
inode_set_mtime_to_ts(&ip->i_inode, inode_set_ctime_current(&ip->i_inode));
fs/gfs2/bmap.c
2018
ip->i_diskflags &= ~GFS2_DIF_TRUNC_IN_PROG;
fs/gfs2/bmap.c
2020
gfs2_trans_add_meta(ip->i_gl, dibh);
fs/gfs2/bmap.c
2021
gfs2_dinode_out(ip, dibh->b_data);
fs/gfs2/bmap.c
2025
up_write(&ip->i_rw_mutex);
fs/gfs2/bmap.c
2043
struct gfs2_inode *ip = GFS2_I(inode);
fs/gfs2/bmap.c
2049
if (gfs2_is_stuffed(ip))
fs/gfs2/bmap.c
2052
error = punch_hole(ip, newsize, 0);
fs/gfs2/bmap.c
2054
error = trunc_end(ip);
fs/gfs2/bmap.c
2081
struct gfs2_inode *ip = GFS2_I(inode);
fs/gfs2/bmap.c
2088
if (gfs2_is_stuffed(ip) && size > gfs2_max_stuffed_size(ip)) {
fs/gfs2/bmap.c
2089
error = gfs2_quota_lock_check(ip, &ap);
fs/gfs2/bmap.c
2093
error = gfs2_inplace_reserve(ip, &ap);
fs/gfs2/bmap.c
2101
gfs2_is_jdata(ip) ? RES_JDATA : 0) +
fs/gfs2/bmap.c
2108
error = gfs2_unstuff_dinode(ip);
fs/gfs2/bmap.c
2113
error = gfs2_meta_inode_buffer(ip, &dibh);
fs/gfs2/bmap.c
2118
inode_set_mtime_to_ts(&ip->i_inode, inode_set_ctime_current(&ip->i_inode));
fs/gfs2/bmap.c
2119
gfs2_trans_add_meta(ip->i_gl, dibh);
fs/gfs2/bmap.c
2120
gfs2_dinode_out(ip, dibh->b_data);
fs/gfs2/bmap.c
2127
gfs2_inplace_release(ip);
fs/gfs2/bmap.c
2129
gfs2_quota_unlock(ip);
fs/gfs2/bmap.c
2148
struct gfs2_inode *ip = GFS2_I(inode);
fs/gfs2/bmap.c
2159
ret = gfs2_qa_get(ip);
fs/gfs2/bmap.c
2170
gfs2_rs_delete(ip);
fs/gfs2/bmap.c
2171
gfs2_qa_put(ip);
fs/gfs2/bmap.c
2175
int gfs2_truncatei_resume(struct gfs2_inode *ip)
fs/gfs2/bmap.c
2178
error = punch_hole(ip, i_size_read(&ip->i_inode), 0);
fs/gfs2/bmap.c
2180
error = trunc_end(ip);
fs/gfs2/bmap.c
2184
int gfs2_file_dealloc(struct gfs2_inode *ip)
fs/gfs2/bmap.c
2186
return punch_hole(ip, 0, 0);
fs/gfs2/bmap.c
2263
struct gfs2_inode *ip = GFS2_I(jd->jd_inode);
fs/gfs2/bmap.c
2287
lblock += (bh.b_size >> ip->i_inode.i_blkbits);
fs/gfs2/bmap.c
2316
int gfs2_write_alloc_required(struct gfs2_inode *ip, u64 offset,
fs/gfs2/bmap.c
2319
struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode);
fs/gfs2/bmap.c
2328
if (gfs2_is_stuffed(ip)) {
fs/gfs2/bmap.c
2329
if (offset + len > gfs2_max_stuffed_size(ip))
fs/gfs2/bmap.c
2335
BUG_ON(gfs2_is_dir(ip));
fs/gfs2/bmap.c
2336
end_of_file = (i_size_read(&ip->i_inode) + sdp->sd_sb.sb_bsize - 1) >> shift;
fs/gfs2/bmap.c
2339
if (lblock_stop > end_of_file && ip != GFS2_I(sdp->sd_rindex))
fs/gfs2/bmap.c
2346
gfs2_block_map(&ip->i_inode, lblock, &bh, 0);
fs/gfs2/bmap.c
2350
lblock += (bh.b_size >> ip->i_inode.i_blkbits);
fs/gfs2/bmap.c
2358
struct gfs2_inode *ip = GFS2_I(inode);
fs/gfs2/bmap.c
2367
error = gfs2_meta_inode_buffer(ip, &dibh);
fs/gfs2/bmap.c
2370
gfs2_trans_add_meta(ip->i_gl, dibh);
fs/gfs2/bmap.c
2416
struct gfs2_inode *ip = GFS2_I(inode);
fs/gfs2/bmap.c
2422
if (!gfs2_is_stuffed(ip)) {
fs/gfs2/bmap.c
2451
if (gfs2_is_jdata(ip))
fs/gfs2/bmap.c
2459
if (gfs2_is_stuffed(ip)) {
fs/gfs2/bmap.c
2465
if (gfs2_is_jdata(ip)) {
fs/gfs2/bmap.c
2477
if (!gfs2_is_stuffed(ip))
fs/gfs2/bmap.c
2478
error = punch_hole(ip, offset, length);
fs/gfs2/bmap.c
324
static int __fillup_metapath(struct gfs2_inode *ip, struct metapath *mp,
fs/gfs2/bmap.c
334
ret = gfs2_meta_buffer(ip, GFS2_METATYPE_IN, dblock, &mp->mp_bh[x + 1]);
fs/gfs2/bmap.c
359
static int lookup_metapath(struct gfs2_inode *ip, struct metapath *mp)
fs/gfs2/bmap.c
361
return __fillup_metapath(ip, mp, 0, ip->i_height - 1);
fs/gfs2/bmap.c
375
static int fillup_metapath(struct gfs2_inode *ip, struct metapath *mp, int h)
fs/gfs2/bmap.c
387
ret = __fillup_metapath(ip, mp, x, h);
fs/gfs2/bmap.c
43
static int punch_hole(struct gfs2_inode *ip, u64 offset, u64 length);
fs/gfs2/bmap.c
472
struct gfs2_inode *ip = GFS2_I(inode);
fs/gfs2/bmap.c
54
static int gfs2_unstuffer_folio(struct gfs2_inode *ip, struct buffer_head *dibh,
fs/gfs2/bmap.c
540
ret = fillup_metapath(ip, mp, ip->i_height - 1);
fs/gfs2/bmap.c
57
struct inode *inode = &ip->i_inode;
fs/gfs2/bmap.c
664
struct gfs2_inode *ip = GFS2_I(inode);
fs/gfs2/bmap.c
680
gfs2_trans_add_meta(ip->i_gl, dibh);
fs/gfs2/bmap.c
682
down_write(&ip->i_rw_mutex);
fs/gfs2/bmap.c
689
if (mp->mp_fheight == ip->i_height) {
fs/gfs2/bmap.c
696
iblks = mp->mp_fheight - ip->i_height;
fs/gfs2/bmap.c
70
if (gfs2_is_jdata(ip)) {
fs/gfs2/bmap.c
708
ret = gfs2_alloc_blocks(ip, &bn, &n, 0);
fs/gfs2/bmap.c
712
if (state != ALLOC_DATA || gfs2_is_jdata(ip))
fs/gfs2/bmap.c
722
for (; i - 1 < mp->mp_fheight - ip->i_height && n > 0;
fs/gfs2/bmap.c
724
gfs2_indirect_init(mp, ip->i_gl, i, 0, bn++);
fs/gfs2/bmap.c
725
if (i - 1 == mp->mp_fheight - ip->i_height) {
fs/gfs2/bmap.c
750
gfs2_trans_add_meta(ip->i_gl, mp->mp_bh[i-1]);
fs/gfs2/bmap.c
752
gfs2_indirect_init(mp, ip->i_gl, i,
fs/gfs2/bmap.c
762
gfs2_trans_add_meta(ip->i_gl, mp->mp_bh[end_of_metadata]);
fs/gfs2/bmap.c
775
ip->i_height = mp->mp_fheight;
fs/gfs2/bmap.c
776
gfs2_add_inode_blocks(&ip->i_inode, alloced);
fs/gfs2/bmap.c
777
gfs2_dinode_out(ip, dibh->b_data);
fs/gfs2/bmap.c
779
up_write(&ip->i_rw_mutex);
fs/gfs2/bmap.c
797
struct gfs2_inode *ip = GFS2_I(inode);
fs/gfs2/bmap.c
809
if (gfs2_is_stuffed(ip) || mp->mp_fheight != mp->mp_aheight) {
fs/gfs2/bmap.c
81
gfs2_trans_add_data(ip->i_gl, bh);
fs/gfs2/bmap.c
818
first = metapointer(ip->i_height - 1, mp);
fs/gfs2/bmap.c
819
end = metaend(ip->i_height - 1, mp);
fs/gfs2/bmap.c
84
gfs2_ordered_add_inode(ip);
fs/gfs2/bmap.c
844
struct gfs2_inode *ip = GFS2_I(inode);
fs/gfs2/bmap.c
859
down_read(&ip->i_rw_mutex);
fs/gfs2/bmap.c
861
ret = gfs2_meta_inode_buffer(ip, &dibh);
fs/gfs2/bmap.c
866
if (gfs2_is_stuffed(ip)) {
fs/gfs2/bmap.c
868
loff_t max_size = gfs2_max_stuffed_size(ip);
fs/gfs2/bmap.c
886
iomap->addr = (ip->i_no_addr << inode->i_blkbits) +
fs/gfs2/bmap.c
90
static int __gfs2_unstuff_inode(struct gfs2_inode *ip, struct folio *folio)
fs/gfs2/bmap.c
900
height = ip->i_height;
fs/gfs2/bmap.c
904
if (height > ip->i_height || gfs2_is_stuffed(ip))
fs/gfs2/bmap.c
907
ret = lookup_metapath(ip, mp);
fs/gfs2/bmap.c
911
if (mp->mp_aheight != ip->i_height)
fs/gfs2/bmap.c
914
ptr = metapointer(ip->i_height - 1, mp);
fs/gfs2/bmap.c
918
bh = mp->mp_bh[ip->i_height - 1];
fs/gfs2/bmap.c
931
up_read(&ip->i_rw_mutex);
fs/gfs2/bmap.c
938
else if (height == ip->i_height)
fs/gfs2/bmap.c
95
int isdir = gfs2_is_dir(ip);
fs/gfs2/bmap.c
953
if (pos < size && height == ip->i_height)
fs/gfs2/bmap.c
966
struct gfs2_inode *ip = GFS2_I(inode);
fs/gfs2/bmap.c
973
if (!gfs2_is_jdata(ip) && !gfs2_is_stuffed(ip))
fs/gfs2/bmap.c
98
error = gfs2_meta_inode_buffer(ip, &dibh);
fs/gfs2/bmap.c
991
struct gfs2_inode *ip = GFS2_I(inode);
fs/gfs2/bmap.c
994
if (gfs2_is_jdata(ip) && !gfs2_is_stuffed(ip))
fs/gfs2/bmap.c
995
gfs2_trans_add_databufs(ip->i_gl, folio,
fs/gfs2/bmap.h
28
static inline void gfs2_write_calc_reserv(const struct gfs2_inode *ip,
fs/gfs2/bmap.h
33
const struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode);
fs/gfs2/bmap.h
36
BUG_ON(gfs2_is_dir(ip));
fs/gfs2/bmap.h
50
int gfs2_unstuff_dinode(struct gfs2_inode *ip);
fs/gfs2/bmap.h
62
int gfs2_truncatei_resume(struct gfs2_inode *ip);
fs/gfs2/bmap.h
63
int gfs2_file_dealloc(struct gfs2_inode *ip);
fs/gfs2/bmap.h
64
int gfs2_write_alloc_required(struct gfs2_inode *ip, u64 offset,
fs/gfs2/dentry.c
42
struct gfs2_inode *ip = NULL;
fs/gfs2/dentry.c
54
ip = GFS2_I(inode);
fs/gfs2/dentry.c
67
error = gfs2_dir_check(dir, name, ip);
fs/gfs2/dir.c
103
static int gfs2_dir_get_existing_buffer(struct gfs2_inode *ip, u64 block,
fs/gfs2/dir.c
109
error = gfs2_meta_read(ip->i_gl, block, DIO_WAIT, 0, &bh);
fs/gfs2/dir.c
112
if (gfs2_metatype_check(GFS2_SB(&ip->i_inode), bh, GFS2_METATYPE_JD)) {
fs/gfs2/dir.c
120
static int gfs2_dir_write_stuffed(struct gfs2_inode *ip, const char *buf,
fs/gfs2/dir.c
126
error = gfs2_meta_inode_buffer(ip, &dibh);
fs/gfs2/dir.c
130
gfs2_trans_add_meta(ip->i_gl, dibh);
fs/gfs2/dir.c
132
if (ip->i_inode.i_size < offset + size)
fs/gfs2/dir.c
133
i_size_write(&ip->i_inode, offset + size);
fs/gfs2/dir.c
134
inode_set_mtime_to_ts(&ip->i_inode, inode_set_ctime_current(&ip->i_inode));
fs/gfs2/dir.c
135
gfs2_dinode_out(ip, dibh->b_data);
fs/gfs2/dir.c
1371
struct gfs2_inode *ip = GFS2_I(inode);
fs/gfs2/dir.c
1384
error = get_leaf(ip, lfn, &bh);
fs/gfs2/dir.c
1420
error = get_leaf(ip, lfn, &bh);
fs/gfs2/dir.c
1439
gfs2_consist_inode(ip);
fs/gfs2/dir.c
1458
error = do_filldir_main(ip, ctx, darr, entries, need_sort ?
fs/gfs2/dir.c
1483
struct gfs2_inode *ip = GFS2_I(inode);
fs/gfs2/dir.c
1484
struct gfs2_glock *gl = ip->i_gl;
fs/gfs2/dir.c
1499
blocknr = be64_to_cpu(ip->i_hash_cache[f_ra->start]);
fs/gfs2/dir.c
153
static int gfs2_dir_write_data(struct gfs2_inode *ip, const char *buf,
fs/gfs2/dir.c
156
struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode);
fs/gfs2/dir.c
1675
const struct gfs2_inode *ip)
fs/gfs2/dir.c
168
if (gfs2_is_stuffed(ip) && offset + size <= gfs2_max_stuffed_size(ip))
fs/gfs2/dir.c
1685
if (ip) {
fs/gfs2/dir.c
1686
if (be64_to_cpu(dent->de_inum.no_addr) != ip->i_no_addr)
fs/gfs2/dir.c
1689
ip->i_no_formal_ino)
fs/gfs2/dir.c
169
return gfs2_dir_write_stuffed(ip, buf, (unsigned int)offset,
fs/gfs2/dir.c
1691
if (unlikely(IF2DT(ip->i_inode.i_mode) !=
fs/gfs2/dir.c
172
if (gfs2_assert_warn(sdp, gfs2_is_jdata(ip)))
fs/gfs2/dir.c
1727
struct gfs2_inode *ip = GFS2_I(inode);
fs/gfs2/dir.c
1734
index = name->hash >> (32 - ip->i_depth);
fs/gfs2/dir.c
1735
error = get_first_leaf(ip, index, &obh);
fs/gfs2/dir.c
1745
error = get_leaf(ip, bn, &obh);
fs/gfs2/dir.c
175
if (gfs2_is_stuffed(ip)) {
fs/gfs2/dir.c
1750
gfs2_trans_add_meta(ip->i_gl, obh);
fs/gfs2/dir.c
176
error = gfs2_unstuff_dinode(ip);
fs/gfs2/dir.c
1762
error = gfs2_meta_inode_buffer(ip, &bh);
fs/gfs2/dir.c
1765
gfs2_trans_add_meta(ip->i_gl, bh);
fs/gfs2/dir.c
1766
gfs2_add_inode_blocks(&ip->i_inode, 1);
fs/gfs2/dir.c
1767
gfs2_dinode_out(ip, bh->b_data);
fs/gfs2/dir.c
1772
static u16 gfs2_inode_ra_len(const struct gfs2_inode *ip)
fs/gfs2/dir.c
1774
u64 where = ip->i_no_addr + 1;
fs/gfs2/dir.c
1775
if (ip->i_eattr == where)
fs/gfs2/dir.c
1799
struct gfs2_inode *ip = GFS2_I(inode);
fs/gfs2/dir.c
1818
tv = inode_set_ctime_current(&ip->i_inode);
fs/gfs2/dir.c
1819
if (ip->i_diskflags & GFS2_DIF_EXHASH) {
fs/gfs2/dir.c
1828
ip->i_entries++;
fs/gfs2/dir.c
1829
inode_set_mtime_to_ts(&ip->i_inode, tv);
fs/gfs2/dir.c
1831
inc_nlink(&ip->i_inode);
fs/gfs2/dir.c
1836
if (!(ip->i_diskflags & GFS2_DIF_EXHASH)) {
fs/gfs2/dir.c
1847
if (ip->i_depth < GFS2_DIR_MAX_DEPTH) {
fs/gfs2/dir.c
1848
error = dir_double_exhash(ip);
fs/gfs2/dir.c
194
error = gfs2_alloc_extent(&ip->i_inode, lblock, &dblock,
fs/gfs2/dir.c
204
error = gfs2_dir_get_new_buffer(ip, dblock, &bh);
fs/gfs2/dir.c
206
error = gfs2_dir_get_existing_buffer(ip, dblock, &bh);
fs/gfs2/dir.c
211
gfs2_trans_add_meta(ip->i_gl, bh);
fs/gfs2/dir.c
2160
struct gfs2_inode *ip = GFS2_I(inode);
fs/gfs2/dir.c
2173
if (!(ip->i_diskflags & GFS2_DIF_EXHASH) &&
fs/gfs2/dir.c
225
error = gfs2_meta_inode_buffer(ip, &dibh);
fs/gfs2/dir.c
229
if (ip->i_inode.i_size < offset + copied)
fs/gfs2/dir.c
230
i_size_write(&ip->i_inode, offset + copied);
fs/gfs2/dir.c
231
inode_set_mtime_to_ts(&ip->i_inode, inode_set_ctime_current(&ip->i_inode));
fs/gfs2/dir.c
233
gfs2_trans_add_meta(ip->i_gl, dibh);
fs/gfs2/dir.c
234
gfs2_dinode_out(ip, dibh->b_data);
fs/gfs2/dir.c
244
static int gfs2_dir_read_stuffed(struct gfs2_inode *ip, __be64 *buf,
fs/gfs2/dir.c
250
error = gfs2_meta_inode_buffer(ip, &dibh);
fs/gfs2/dir.c
268
static int gfs2_dir_read_data(struct gfs2_inode *ip, __be64 *buf,
fs/gfs2/dir.c
271
struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode);
fs/gfs2/dir.c
278
if (gfs2_is_stuffed(ip))
fs/gfs2/dir.c
279
return gfs2_dir_read_stuffed(ip, buf, size);
fs/gfs2/dir.c
281
if (gfs2_assert_warn(sdp, gfs2_is_jdata(ip)))
fs/gfs2/dir.c
297
error = gfs2_get_extent(&ip->i_inode, lblock,
fs/gfs2/dir.c
302
bh = gfs2_meta_ra(ip->i_gl, dblock, extlen);
fs/gfs2/dir.c
304
error = gfs2_meta_read(ip->i_gl, dblock, DIO_WAIT, 0, &bh);
fs/gfs2/dir.c
335
static __be64 *gfs2_dir_get_hash_table(struct gfs2_inode *ip)
fs/gfs2/dir.c
337
struct inode *inode = &ip->i_inode;
fs/gfs2/dir.c
342
BUG_ON(!(ip->i_diskflags & GFS2_DIF_EXHASH));
fs/gfs2/dir.c
344
hc = ip->i_hash_cache;
fs/gfs2/dir.c
348
hsize = BIT(ip->i_depth);
fs/gfs2/dir.c
350
if (hsize != i_size_read(&ip->i_inode)) {
fs/gfs2/dir.c
351
gfs2_consist_inode(ip);
fs/gfs2/dir.c
362
ret = gfs2_dir_read_data(ip, hc, hsize);
fs/gfs2/dir.c
369
if (likely(!ip->i_hash_cache)) {
fs/gfs2/dir.c
370
ip->i_hash_cache = hc;
fs/gfs2/dir.c
376
return ip->i_hash_cache;
fs/gfs2/dir.c
385
void gfs2_dir_hash_inval(struct gfs2_inode *ip)
fs/gfs2/dir.c
389
spin_lock(&ip->i_inode.i_lock);
fs/gfs2/dir.c
390
hc = ip->i_hash_cache;
fs/gfs2/dir.c
391
ip->i_hash_cache = NULL;
fs/gfs2/dir.c
392
spin_unlock(&ip->i_inode.i_lock);
fs/gfs2/dir.c
715
struct gfs2_inode *ip = GFS2_I(inode);
fs/gfs2/dir.c
721
gfs2_trans_add_meta(ip->i_gl, bh);
fs/gfs2/dir.c
816
struct gfs2_inode *ip = GFS2_I(inode);
fs/gfs2/dir.c
819
if (ip->i_diskflags & GFS2_DIF_EXHASH) {
fs/gfs2/dir.c
821
unsigned int hsize = BIT(ip->i_depth);
fs/gfs2/dir.c
825
gfs2_consist_inode(ip);
fs/gfs2/dir.c
829
index = name->hash >> (32 - ip->i_depth);
fs/gfs2/dir.c
830
error = get_first_leaf(ip, index, &bh);
fs/gfs2/dir.c
844
error = get_leaf(ip, ln, &bh);
fs/gfs2/dir.c
851
error = gfs2_meta_inode_buffer(ip, &bh);
fs/gfs2/dir.c
866
struct gfs2_inode *ip = GFS2_I(inode);
fs/gfs2/dir.c
875
error = gfs2_alloc_blocks(ip, &bn, &n, 0);
fs/gfs2/dir.c
878
bh = gfs2_meta_new(ip->i_gl, bn);
fs/gfs2/dir.c
883
gfs2_trans_add_meta(ip->i_gl, bh);
fs/gfs2/dir.c
890
leaf->lf_inode = cpu_to_be64(ip->i_no_addr);
fs/gfs2/dir.c
90
int gfs2_dir_get_new_buffer(struct gfs2_inode *ip, u64 block,
fs/gfs2/dir.c
95
bh = gfs2_meta_new(ip->i_gl, block);
fs/gfs2/dir.c
96
gfs2_trans_add_meta(ip->i_gl, bh);
fs/gfs2/dir.h
30
const struct gfs2_inode *ip);
fs/gfs2/dir.h
32
const struct gfs2_inode *ip, struct gfs2_diradd *da);
fs/gfs2/dir.h
49
int gfs2_dir_get_new_buffer(struct gfs2_inode *ip, u64 block,
fs/gfs2/dir.h
51
void gfs2_dir_hash_inval(struct gfs2_inode *ip);
fs/gfs2/export.c
106
ip = GFS2_I(inode);
fs/gfs2/export.c
109
gnfd.inum.no_addr = ip->i_no_addr;
fs/gfs2/export.c
110
gnfd.inum.no_formal_ino = ip->i_no_formal_ino;
fs/gfs2/export.c
33
struct gfs2_inode *ip = GFS2_I(inode);
fs/gfs2/export.c
43
fh[0] = cpu_to_be32(ip->i_no_formal_ino >> 32);
fs/gfs2/export.c
44
fh[1] = cpu_to_be32(ip->i_no_formal_ino & 0xFFFFFFFF);
fs/gfs2/export.c
45
fh[2] = cpu_to_be32(ip->i_no_addr >> 32);
fs/gfs2/export.c
46
fh[3] = cpu_to_be32(ip->i_no_addr & 0xFFFFFFFF);
fs/gfs2/export.c
52
ip = GFS2_I(parent);
fs/gfs2/export.c
54
fh[4] = cpu_to_be32(ip->i_no_formal_ino >> 32);
fs/gfs2/export.c
55
fh[5] = cpu_to_be32(ip->i_no_formal_ino & 0xFFFFFFFF);
fs/gfs2/export.c
56
fh[6] = cpu_to_be32(ip->i_no_addr >> 32);
fs/gfs2/export.c
57
fh[7] = cpu_to_be32(ip->i_no_addr & 0xFFFFFFFF);
fs/gfs2/export.c
90
struct gfs2_inode *dip, *ip;
fs/gfs2/file.c
1016
struct gfs2_inode *ip = GFS2_I(inode);
fs/gfs2/file.c
1037
gfs2_holder_init(ip->i_gl, LM_ST_EXCLUSIVE, 0, gh);
fs/gfs2/file.c
1104
struct gfs2_inode *ip = GFS2_I(inode);
fs/gfs2/file.c
1111
ret = gfs2_glock_nq_init(ip->i_gl, LM_ST_SHARED, 0, &gh);
fs/gfs2/file.c
1180
struct gfs2_inode *ip = GFS2_I(inode);
fs/gfs2/file.c
1185
error = gfs2_meta_inode_buffer(ip, &dibh);
fs/gfs2/file.c
1189
gfs2_trans_add_meta(ip->i_gl, dibh);
fs/gfs2/file.c
1191
if (gfs2_is_stuffed(ip)) {
fs/gfs2/file.c
1192
error = gfs2_unstuff_dinode(ip);
fs/gfs2/file.c
1230
static void calc_max_reserv(struct gfs2_inode *ip, loff_t *len,
fs/gfs2/file.c
1235
const struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode);
fs/gfs2/file.c
1248
gfs2_write_calc_reserv(ip, max, data_blocks, ind_blocks);
fs/gfs2/file.c
1256
struct gfs2_inode *ip = GFS2_I(inode);
fs/gfs2/file.c
1281
gfs2_write_calc_reserv(ip, PAGE_SIZE, &data_blocks, &ind_blocks);
fs/gfs2/file.c
1287
if (!gfs2_write_alloc_required(ip, offset, bytes)) {
fs/gfs2/file.c
1303
gfs2_write_calc_reserv(ip, bytes, &data_blocks, &ind_blocks);
fs/gfs2/file.c
1306
error = gfs2_quota_lock_check(ip, &ap);
fs/gfs2/file.c
1315
error = gfs2_inplace_reserve(ip, &ap);
fs/gfs2/file.c
1320
if (ip->i_res.rs_reserved < max_blks)
fs/gfs2/file.c
1321
max_blks = ip->i_res.rs_reserved;
fs/gfs2/file.c
1326
calc_max_reserv(ip, &max_bytes, &data_blocks,
fs/gfs2/file.c
1330
RES_RG_HDR + gfs2_rg_blocks(ip, data_blocks + ind_blocks);
fs/gfs2/file.c
1331
if (gfs2_is_jdata(ip))
fs/gfs2/file.c
1347
gfs2_inplace_release(ip);
fs/gfs2/file.c
1348
gfs2_quota_unlock(ip);
fs/gfs2/file.c
1362
gfs2_inplace_release(ip);
fs/gfs2/file.c
1364
gfs2_quota_unlock(ip);
fs/gfs2/file.c
1372
struct gfs2_inode *ip = GFS2_I(inode);
fs/gfs2/file.c
1379
if (gfs2_is_jdata(ip) && inode != sdp->sd_rindex)
fs/gfs2/file.c
1384
gfs2_holder_init(ip->i_gl, LM_ST_EXCLUSIVE, 0, &gh);
fs/gfs2/file.c
1405
gfs2_rs_deltree(&ip->i_res);
fs/gfs2/file.c
1442
struct gfs2_inode *ip = GFS2_I(file->f_mapping->host);
fs/gfs2/file.c
1458
ret = dlm_posix_cancel(ls->ls_dlm, ip->i_no_addr, file, fl);
fs/gfs2/file.c
1460
ret = dlm_posix_get(ls->ls_dlm, ip->i_no_addr, file, fl);
fs/gfs2/file.c
1462
ret = dlm_posix_unlock(ls->ls_dlm, ip->i_no_addr, file, fl);
fs/gfs2/file.c
1464
ret = dlm_posix_lock(ls->ls_dlm, ip->i_no_addr, file, cmd, fl);
fs/gfs2/file.c
1489
struct gfs2_inode *ip = GFS2_I(file_inode(file));
fs/gfs2/file.c
1514
error = gfs2_glock_get(GFS2_SB(&ip->i_inode), ip->i_no_addr,
fs/gfs2/file.c
1537
gfs2_assert_warn(GFS2_SB(&ip->i_inode), !error);
fs/gfs2/file.c
161
struct gfs2_inode *ip = GFS2_I(inode);
fs/gfs2/file.c
169
gfs2_holder_init(ip->i_gl, LM_ST_SHARED, 0, &gh);
fs/gfs2/file.c
174
fsflags = gfs2_gfsflags_to_fsflags(inode, ip->i_diskflags);
fs/gfs2/file.c
186
struct gfs2_inode *ip = GFS2_I(inode);
fs/gfs2/file.c
190
if ((ip->i_eattr == 0) && !is_sxid(inode->i_mode))
fs/gfs2/file.c
192
if (ip->i_diskflags & GFS2_DIF_IMMUTABLE)
fs/gfs2/file.c
194
if (ip->i_diskflags & GFS2_DIF_APPENDONLY)
fs/gfs2/file.c
196
if (ip->i_diskflags & GFS2_DIF_NOATIME)
fs/gfs2/file.c
198
if (ip->i_diskflags & GFS2_DIF_SYNC)
fs/gfs2/file.c
221
struct gfs2_inode *ip = GFS2_I(inode);
fs/gfs2/file.c
228
error = gfs2_glock_nq_init(ip->i_gl, LM_ST_EXCLUSIVE, 0, &gh);
fs/gfs2/file.c
233
flags = ip->i_diskflags;
fs/gfs2/file.c
245
gfs2_log_flush(sdp, ip->i_gl,
fs/gfs2/file.c
256
gfs2_ordered_del_inode(ip);
fs/gfs2/file.c
261
error = gfs2_meta_inode_buffer(ip, &bh);
fs/gfs2/file.c
265
gfs2_trans_add_meta(ip->i_gl, bh);
fs/gfs2/file.c
266
ip->i_diskflags = new_flags;
fs/gfs2/file.c
267
gfs2_dinode_out(ip, bh->b_data);
fs/gfs2/file.c
371
struct gfs2_inode *ip = GFS2_I(inode);
fs/gfs2/file.c
375
if (hint > atomic_read(&ip->i_sizehint))
fs/gfs2/file.c
376
atomic_set(&ip->i_sizehint, hint);
fs/gfs2/file.c
420
struct gfs2_inode *ip = GFS2_I(inode);
fs/gfs2/file.c
433
gfs2_holder_init(ip->i_gl, LM_ST_EXCLUSIVE, 0, &gh);
fs/gfs2/file.c
458
set_bit(GLF_DIRTY, &ip->i_gl->gl_flags);
fs/gfs2/file.c
459
set_bit(GIF_SW_PAGED, &ip->i_flags);
fs/gfs2/file.c
466
if (!gfs2_is_stuffed(ip) &&
fs/gfs2/file.c
467
!gfs2_write_alloc_required(ip, pos, length)) {
fs/gfs2/file.c
483
gfs2_write_calc_reserv(ip, length, &data_blocks, &ind_blocks);
fs/gfs2/file.c
485
err = gfs2_quota_lock_check(ip, &ap);
fs/gfs2/file.c
490
err = gfs2_inplace_reserve(ip, &ap);
fs/gfs2/file.c
497
if (gfs2_is_jdata(ip))
fs/gfs2/file.c
501
rblocks += gfs2_rg_blocks(ip, data_blocks + ind_blocks);
fs/gfs2/file.c
510
if (gfs2_is_stuffed(ip)) {
fs/gfs2/file.c
511
err = gfs2_unstuff_dinode(ip);
fs/gfs2/file.c
537
gfs2_inplace_release(ip);
fs/gfs2/file.c
539
gfs2_quota_unlock(ip);
fs/gfs2/file.c
555
struct gfs2_inode *ip = GFS2_I(inode);
fs/gfs2/file.c
560
gfs2_holder_init(ip->i_gl, LM_ST_SHARED, 0, &gh);
fs/gfs2/file.c
593
struct gfs2_inode *ip = GFS2_I(file->f_mapping->host);
fs/gfs2/file.c
596
!IS_NOATIME(&ip->i_inode)) {
fs/gfs2/file.c
60
struct gfs2_inode *ip = GFS2_I(file->f_mapping->host);
fs/gfs2/file.c
600
error = gfs2_glock_nq_init(ip->i_gl, LM_ST_SHARED, LM_FLAG_ANY,
fs/gfs2/file.c
66
error = gfs2_glock_nq_init(ip->i_gl, LM_ST_SHARED, LM_FLAG_ANY,
fs/gfs2/file.c
677
struct gfs2_inode *ip = GFS2_I(inode);
fs/gfs2/file.c
682
if (S_ISREG(ip->i_inode.i_mode)) {
fs/gfs2/file.c
683
error = gfs2_glock_nq_init(ip->i_gl, LM_ST_SHARED, LM_FLAG_ANY,
fs/gfs2/file.c
708
struct gfs2_inode *ip = GFS2_I(inode);
fs/gfs2/file.c
714
if (gfs2_rs_active(&ip->i_res))
fs/gfs2/file.c
715
gfs2_rs_delete(ip);
fs/gfs2/file.c
716
gfs2_qa_put(ip);
fs/gfs2/file.c
748
struct gfs2_inode *ip = GFS2_I(inode);
fs/gfs2/file.c
757
if (!gfs2_is_jdata(ip))
fs/gfs2/file.c
766
if (gfs2_is_jdata(ip))
fs/gfs2/file.c
770
gfs2_ail_flush(ip->i_gl, 1);
fs/gfs2/file.c
815
struct gfs2_inode *ip = GFS2_I(file->f_mapping->host);
fs/gfs2/file.c
840
gfs2_holder_init(ip->i_gl, LM_ST_DEFERRED, 0, gh);
fs/gfs2/file.c
879
struct gfs2_inode *ip = GFS2_I(inode);
fs/gfs2/file.c
903
gfs2_holder_init(ip->i_gl, LM_ST_DEFERRED, 0, gh);
fs/gfs2/file.c
909
if (iocb->ki_pos + iov_iter_count(from) > i_size_read(&ip->i_inode))
fs/gfs2/file.c
951
struct gfs2_inode *ip;
fs/gfs2/file.c
982
ip = GFS2_I(iocb->ki_filp->f_mapping->host);
fs/gfs2/file.c
983
gfs2_holder_init(ip->i_gl, LM_ST_SHARED, 0, &gh);
fs/gfs2/glock.c
1182
struct gfs2_holder *gh, unsigned long ip)
fs/gfs2/glock.c
1186
gh->gh_ip = ip;
fs/gfs2/glock.c
877
struct gfs2_inode *ip;
fs/gfs2/glock.c
880
ip = gl->gl_object;
fs/gfs2/glock.c
881
if (ip && !igrab(&ip->i_inode))
fs/gfs2/glock.c
882
ip = NULL;
fs/gfs2/glock.c
884
if (ip) {
fs/gfs2/glock.c
885
wait_on_new_inode(&ip->i_inode);
fs/gfs2/glock.c
886
if (is_bad_inode(&ip->i_inode)) {
fs/gfs2/glock.c
887
iput(&ip->i_inode);
fs/gfs2/glock.c
888
ip = NULL;
fs/gfs2/glock.c
891
return ip;
fs/gfs2/glock.c
896
struct gfs2_inode *ip;
fs/gfs2/glock.c
911
ip = gfs2_grab_existing_inode(gl);
fs/gfs2/glock.c
912
if (ip) {
fs/gfs2/glock.c
914
d_prune_aliases(&ip->i_inode);
fs/gfs2/glock.c
915
iput(&ip->i_inode);
fs/gfs2/glock.c
919
ip = gfs2_grab_existing_inode(gl);
fs/gfs2/glock.c
920
if (ip) {
fs/gfs2/glock.c
921
gfs2_glock_poke(ip->i_gl);
fs/gfs2/glock.c
922
iput(&ip->i_inode);
fs/gfs2/glock.h
193
unsigned long ip);
fs/gfs2/glops.c
250
struct gfs2_inode *ip;
fs/gfs2/glops.c
253
ip = gl->gl_object;
fs/gfs2/glops.c
254
if (ip)
fs/gfs2/glops.c
255
set_bit(GIF_GLOP_PENDING, &ip->i_flags);
fs/gfs2/glops.c
257
return ip;
fs/gfs2/glops.c
271
static void gfs2_clear_glop_pending(struct gfs2_inode *ip)
fs/gfs2/glops.c
273
if (!ip)
fs/gfs2/glops.c
276
clear_bit_unlock(GIF_GLOP_PENDING, &ip->i_flags);
fs/gfs2/glops.c
277
wake_up_bit(&ip->i_flags, GIF_GLOP_PENDING);
fs/gfs2/glops.c
305
struct gfs2_inode *ip = gfs2_glock2inode(gl);
fs/gfs2/glops.c
306
int isreg = ip && S_ISREG(ip->i_inode.i_mode);
fs/gfs2/glops.c
311
if (test_and_clear_bit(GIF_SW_PAGED, &ip->i_flags))
fs/gfs2/glops.c
312
unmap_shared_mapping_range(ip->i_inode.i_mapping, 0, 0);
fs/gfs2/glops.c
313
inode_dio_wait(&ip->i_inode);
fs/gfs2/glops.c
324
struct address_space *mapping = ip->i_inode.i_mapping;
fs/gfs2/glops.c
343
gfs2_clear_glop_pending(ip);
fs/gfs2/glops.c
360
struct gfs2_inode *ip = gfs2_glock2inode(gl);
fs/gfs2/glops.c
367
if (ip) {
fs/gfs2/glops.c
369
forget_all_cached_acls(&ip->i_inode);
fs/gfs2/glops.c
370
security_inode_invalidate_secctx(&ip->i_inode);
fs/gfs2/glops.c
371
gfs2_dir_hash_inval(ip);
fs/gfs2/glops.c
375
if (ip == GFS2_I(glock_sbd(gl)->sd_rindex)) {
fs/gfs2/glops.c
381
if (ip && S_ISREG(ip->i_inode.i_mode))
fs/gfs2/glops.c
382
truncate_inode_pages(ip->i_inode.i_mapping, 0);
fs/gfs2/glops.c
384
gfs2_clear_glop_pending(ip);
fs/gfs2/glops.c
387
static int gfs2_dinode_in(struct gfs2_inode *ip, const void *buf)
fs/gfs2/glops.c
389
struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode);
fs/gfs2/glops.c
394
struct inode *inode = &ip->i_inode;
fs/gfs2/glops.c
397
if (unlikely(ip->i_no_addr != be64_to_cpu(str->di_num.no_addr))) {
fs/gfs2/glops.c
398
gfs2_consist_inode(ip);
fs/gfs2/glops.c
402
gfs2_consist_inode(ip);
fs/gfs2/glops.c
405
ip->i_no_formal_ino = be64_to_cpu(str->di_num.no_formal_ino);
fs/gfs2/glops.c
433
ip->i_goal = be64_to_cpu(str->di_goal_meta);
fs/gfs2/glops.c
434
ip->i_generation = be64_to_cpu(str->di_generation);
fs/gfs2/glops.c
436
ip->i_diskflags = be32_to_cpu(str->di_flags);
fs/gfs2/glops.c
437
ip->i_eattr = be64_to_cpu(str->di_eattr);
fs/gfs2/glops.c
442
gfs2_consist_inode(ip);
fs/gfs2/glops.c
445
ip->i_height = (u8)height;
fs/gfs2/glops.c
449
gfs2_consist_inode(ip);
fs/gfs2/glops.c
452
if ((ip->i_diskflags & GFS2_DIF_EXHASH) &&
fs/gfs2/glops.c
454
gfs2_consist_inode(ip);
fs/gfs2/glops.c
457
ip->i_depth = (u8)depth;
fs/gfs2/glops.c
458
ip->i_entries = be32_to_cpu(str->di_entries);
fs/gfs2/glops.c
460
if (gfs2_is_stuffed(ip) && inode->i_size > gfs2_max_stuffed_size(ip)) {
fs/gfs2/glops.c
461
gfs2_consist_inode(ip);
fs/gfs2/glops.c
477
static int gfs2_inode_refresh(struct gfs2_inode *ip)
fs/gfs2/glops.c
482
error = gfs2_meta_inode_buffer(ip, &dibh);
fs/gfs2/glops.c
486
error = gfs2_dinode_in(ip, dibh->b_data);
fs/gfs2/glops.c
500
struct gfs2_inode *ip = gl->gl_object;
fs/gfs2/glops.c
504
if (!ip) /* no inode to populate - read it in later */
fs/gfs2/glops.c
507
error = gfs2_inode_refresh(ip);
fs/gfs2/glops.c
510
io_gl = ip->i_iopen_gh.gh_gl;
fs/gfs2/glops.c
511
io_gl->gl_no_formal_ino = ip->i_no_formal_ino;
fs/gfs2/glops.c
518
struct gfs2_inode *ip = gl->gl_object;
fs/gfs2/glops.c
521
if (!ip) /* no inode to populate - read it in later */
fs/gfs2/glops.c
525
inode_dio_wait(&ip->i_inode);
fs/gfs2/glops.c
527
if ((ip->i_diskflags & GFS2_DIF_TRUNC_IN_PROG) &&
fs/gfs2/glops.c
530
error = gfs2_truncatei_resume(ip);
fs/gfs2/glops.c
546
struct gfs2_inode *ip = gl->gl_object;
fs/gfs2/glops.c
547
const struct inode *inode = &ip->i_inode;
fs/gfs2/glops.c
549
if (ip == NULL)
fs/gfs2/glops.c
554
(unsigned long long)ip->i_no_formal_ino,
fs/gfs2/glops.c
555
(unsigned long long)ip->i_no_addr,
fs/gfs2/glops.c
556
IF2DT(inode->i_mode), ip->i_flags,
fs/gfs2/glops.c
557
(unsigned int)ip->i_diskflags,
fs/gfs2/glops.c
600
struct gfs2_inode *ip = GFS2_I(sdp->sd_jdesc->jd_inode);
fs/gfs2/glops.c
601
struct gfs2_glock *j_gl = ip->i_gl;
fs/gfs2/glops.c
628
struct gfs2_inode *ip = gl->gl_object;
fs/gfs2/glops.c
635
gl->gl_state == LM_ST_SHARED && ip) {
fs/gfs2/incore.h
876
static inline unsigned gfs2_max_stuffed_size(const struct gfs2_inode *ip)
fs/gfs2/incore.h
878
return GFS2_SB(&ip->i_inode)->sd_sb.sb_bsize - sizeof(struct gfs2_dinode);
fs/gfs2/inode.c
1045
struct gfs2_inode *ip = GFS2_I(inode);
fs/gfs2/inode.c
1059
gfs2_holder_init(ip->i_gl, LM_ST_EXCLUSIVE, 0, &gh);
fs/gfs2/inode.c
1098
if (ip->i_inode.i_nlink == (u32)-1)
fs/gfs2/inode.c
1124
error = gfs2_meta_inode_buffer(ip, &dibh);
fs/gfs2/inode.c
1128
error = gfs2_dir_add(dir, &dentry->d_name, ip, &da);
fs/gfs2/inode.c
1132
gfs2_trans_add_meta(ip->i_gl, dibh);
fs/gfs2/inode.c
1133
inc_nlink(&ip->i_inode);
fs/gfs2/inode.c
1134
inode_set_ctime_current(&ip->i_inode);
fs/gfs2/inode.c
1173
const struct gfs2_inode *ip)
fs/gfs2/inode.c
1177
if (IS_IMMUTABLE(&ip->i_inode) || IS_APPEND(&ip->i_inode))
fs/gfs2/inode.c
1182
!uid_eq(ip->i_inode.i_uid, current_fsuid()) && !capable(CAP_FOWNER))
fs/gfs2/inode.c
1193
return gfs2_dir_check(&dip->i_inode, name, ip);
fs/gfs2/inode.c
1211
struct gfs2_inode *ip = GFS2_I(inode);
fs/gfs2/inode.c
1218
ip->i_entries = 0;
fs/gfs2/inode.c
1247
struct gfs2_inode *ip = GFS2_I(inode);
fs/gfs2/inode.c
1259
gfs2_holder_init(ip->i_gl, LM_ST_EXCLUSIVE, 0, &gh);
fs/gfs2/inode.c
1261
rgd = gfs2_blk2rgrpd(sdp, ip->i_no_addr, 1);
fs/gfs2/inode.c
1282
if (ip->i_entries > 2 || inode->i_nlink > 2)
fs/gfs2/inode.c
1290
error = gfs2_unlink_ok(dip, &dentry->d_name, ip);
fs/gfs2/inode.c
132
struct gfs2_inode *ip;
fs/gfs2/inode.c
141
ip = GFS2_I(inode);
fs/gfs2/inode.c
1463
static int update_moved_ino(struct gfs2_inode *ip, struct gfs2_inode *ndip,
fs/gfs2/inode.c
1467
return gfs2_dir_mvino(ip, &gfs2_qdotdot, ndip, DT_DIR);
fs/gfs2/inode.c
1469
inode_set_ctime_current(&ip->i_inode);
fs/gfs2/inode.c
1470
mark_inode_dirty_sync(&ip->i_inode);
fs/gfs2/inode.c
1490
struct gfs2_inode *ip = GFS2_I(d_inode(odentry));
fs/gfs2/inode.c
150
&ip->i_gl);
fs/gfs2/inode.c
1505
if (ip == nip)
fs/gfs2/inode.c
1523
if (S_ISDIR(ip->i_inode.i_mode)) {
fs/gfs2/inode.c
1526
error = gfs2_ok_to_move(ip, ndip);
fs/gfs2/inode.c
1539
gfs2_holder_init(ip->i_gl, LM_ST_EXCLUSIVE, GL_ASYNC, ghs + num_gh);
fs/gfs2/inode.c
1579
if (ip->i_inode.i_nlink == 0)
fs/gfs2/inode.c
1584
error = gfs2_unlink_ok(odip, &odentry->d_name, ip);
fs/gfs2/inode.c
1638
if (S_ISDIR(ip->i_inode.i_mode) &&
fs/gfs2/inode.c
1687
error = update_moved_ino(ip, ndip, dir_rename);
fs/gfs2/inode.c
1695
error = gfs2_dir_add(ndir, &ndentry->d_name, ip, &da);
fs/gfs2/inode.c
170
&ip->i_iopen_gh);
fs/gfs2/inode.c
181
error = gfs2_glock_nq_init(ip->i_gl, LM_ST_EXCLUSIVE,
fs/gfs2/inode.c
188
gfs2_inode_already_deleted(ip->i_gl, no_formal_ino))
fs/gfs2/inode.c
1909
struct gfs2_inode *ip = GFS2_I(inode);
fs/gfs2/inode.c
1919
gfs2_holder_init(ip->i_gl, LM_ST_SHARED, 0, &i_gh);
fs/gfs2/inode.c
1926
size = (unsigned int)i_size_read(&ip->i_inode);
fs/gfs2/inode.c
1928
gfs2_consist_inode(ip);
fs/gfs2/inode.c
1933
error = gfs2_meta_inode_buffer(ip, &dibh);
fs/gfs2/inode.c
1969
struct gfs2_inode *ip;
fs/gfs2/inode.c
1975
ip = GFS2_I(inode);
fs/gfs2/inode.c
1976
gl = rcu_dereference_check(ip->i_gl, !may_not_block);
fs/gfs2/inode.c
199
set_bit(GLF_INSTANTIATE_NEEDED, &ip->i_gl->gl_flags);
fs/gfs2/inode.c
2025
struct gfs2_inode *ip = GFS2_I(inode);
fs/gfs2/inode.c
2041
error = gfs2_qa_get(ip);
fs/gfs2/inode.c
2049
error = gfs2_quota_lock(ip, nuid, ngid);
fs/gfs2/inode.c
2053
ap.target = gfs2_get_inode_blocks(&ip->i_inode);
fs/gfs2/inode.c
2057
error = gfs2_quota_check(ip, nuid, ngid, &ap);
fs/gfs2/inode.c
206
glock_set_object(ip->i_gl, ip);
fs/gfs2/inode.c
2072
gfs2_quota_change(ip, -(s64)ap.target, ouid, ogid);
fs/gfs2/inode.c
2073
gfs2_quota_change(ip, ap.target, nuid, ngid);
fs/gfs2/inode.c
2079
gfs2_quota_unlock(ip);
fs/gfs2/inode.c
2081
gfs2_qa_put(ip);
fs/gfs2/inode.c
2101
struct gfs2_inode *ip = GFS2_I(inode);
fs/gfs2/inode.c
2105
error = gfs2_qa_get(ip);
fs/gfs2/inode.c
2109
error = gfs2_glock_nq_init(ip->i_gl, LM_ST_EXCLUSIVE, 0, &i_gh);
fs/gfs2/inode.c
212
glock_clear_object(ip->i_gl, ip);
fs/gfs2/inode.c
2137
gfs2_qa_put(ip);
fs/gfs2/inode.c
216
ip->i_no_formal_ino = no_formal_ino;
fs/gfs2/inode.c
2163
struct gfs2_inode *ip = GFS2_I(inode);
fs/gfs2/inode.c
2169
if (gfs2_glock_is_locked_by_me(ip->i_gl) == NULL) {
fs/gfs2/inode.c
2170
error = gfs2_glock_nq_init(ip->i_gl, LM_ST_SHARED, LM_FLAG_ANY, &gh);
fs/gfs2/inode.c
2175
gfsflags = ip->i_diskflags;
fs/gfs2/inode.c
2206
struct gfs2_inode *ip = GFS2_I(inode);
fs/gfs2/inode.c
2213
ret = gfs2_glock_nq_init(ip->i_gl, LM_ST_SHARED, 0, &gh);
fs/gfs2/inode.c
222
glock_set_object(ip->i_iopen_gh.gh_gl, ip);
fs/gfs2/inode.c
2236
struct gfs2_inode *ip = GFS2_I(inode);
fs/gfs2/inode.c
2241
ret = gfs2_glock_nq_init(ip->i_gl, LM_ST_SHARED, 0, &gh);
fs/gfs2/inode.c
2255
struct gfs2_inode *ip = GFS2_I(inode);
fs/gfs2/inode.c
2260
ret = gfs2_glock_nq_init(ip->i_gl, LM_ST_SHARED, 0, &gh);
fs/gfs2/inode.c
2274
struct gfs2_inode *ip = GFS2_I(inode);
fs/gfs2/inode.c
2275
struct gfs2_glock *gl = ip->i_gl;
fs/gfs2/inode.c
228
if (no_formal_ino && ip->i_no_formal_ino &&
fs/gfs2/inode.c
229
no_formal_ino != ip->i_no_formal_ino) {
fs/gfs2/inode.c
239
if (gfs2_holder_initialized(&ip->i_iopen_gh))
fs/gfs2/inode.c
240
gfs2_glock_dq_uninit(&ip->i_iopen_gh);
fs/gfs2/inode.c
243
if (ip->i_gl) {
fs/gfs2/inode.c
244
gfs2_glock_put(ip->i_gl);
fs/gfs2/inode.c
245
ip->i_gl = NULL;
fs/gfs2/inode.c
418
static int alloc_dinode(struct gfs2_inode *ip, u32 flags, unsigned *dblocks)
fs/gfs2/inode.c
420
struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode);
fs/gfs2/inode.c
424
error = gfs2_quota_lock_check(ip, &ap);
fs/gfs2/inode.c
428
error = gfs2_inplace_reserve(ip, &ap);
fs/gfs2/inode.c
436
error = gfs2_alloc_blocks(ip, &ip->i_no_addr, dblocks, 1);
fs/gfs2/inode.c
440
ip->i_no_formal_ino = ip->i_generation;
fs/gfs2/inode.c
441
ip->i_inode.i_ino = ip->i_no_addr;
fs/gfs2/inode.c
442
ip->i_goal = ip->i_no_addr;
fs/gfs2/inode.c
444
ip->i_eattr = ip->i_no_addr + 1;
fs/gfs2/inode.c
449
gfs2_inplace_release(ip);
fs/gfs2/inode.c
451
gfs2_quota_unlock(ip);
fs/gfs2/inode.c
456
static void gfs2_final_release_pages(struct gfs2_inode *ip)
fs/gfs2/inode.c
458
struct inode *inode = &ip->i_inode;
fs/gfs2/inode.c
459
struct gfs2_glock *gl = ip->i_gl;
fs/gfs2/inode.c
474
int gfs2_dinode_dealloc(struct gfs2_inode *ip)
fs/gfs2/inode.c
476
struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode);
fs/gfs2/inode.c
481
if (gfs2_get_inode_blocks(&ip->i_inode) != 1) {
fs/gfs2/inode.c
482
gfs2_consist_inode(ip);
fs/gfs2/inode.c
488
error = gfs2_quota_hold(ip, NO_UID_QUOTA_CHANGE, NO_GID_QUOTA_CHANGE);
fs/gfs2/inode.c
492
rgd = gfs2_blk2rgrpd(sdp, ip->i_no_addr, 1);
fs/gfs2/inode.c
494
gfs2_consist_inode(ip);
fs/gfs2/inode.c
509
gfs2_free_di(rgd, ip);
fs/gfs2/inode.c
511
gfs2_final_release_pages(ip);
fs/gfs2/inode.c
518
gfs2_quota_unhold(ip);
fs/gfs2/inode.c
547
static void gfs2_init_xattr(struct gfs2_inode *ip)
fs/gfs2/inode.c
549
struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode);
fs/gfs2/inode.c
553
bh = gfs2_meta_new(ip->i_gl, ip->i_eattr);
fs/gfs2/inode.c
554
gfs2_trans_add_meta(ip->i_gl, bh);
fs/gfs2/inode.c
574
static void init_dinode(struct gfs2_inode *dip, struct gfs2_inode *ip,
fs/gfs2/inode.c
580
dibh = gfs2_meta_new(ip->i_gl, ip->i_no_addr);
fs/gfs2/inode.c
581
gfs2_trans_add_meta(ip->i_gl, dibh);
fs/gfs2/inode.c
583
gfs2_dinode_out(ip, di);
fs/gfs2/inode.c
585
di->di_major = cpu_to_be32(imajor(&ip->i_inode));
fs/gfs2/inode.c
586
di->di_minor = cpu_to_be32(iminor(&ip->i_inode));
fs/gfs2/inode.c
594
switch(ip->i_inode.i_mode & S_IFMT) {
fs/gfs2/inode.c
599
memcpy(dibh->b_data + sizeof(struct gfs2_dinode), symname, ip->i_inode.i_size);
fs/gfs2/inode.c
630
struct gfs2_inode *ip, struct gfs2_diradd *da)
fs/gfs2/inode.c
654
error = gfs2_dir_add(&dip->i_inode, name, ip, da);
fs/gfs2/inode.c
709
struct gfs2_inode *dip = GFS2_I(dir), *ip;
fs/gfs2/inode.c
770
ip = GFS2_I(inode);
fs/gfs2/inode.c
776
error = gfs2_qa_get(ip);
fs/gfs2/inode.c
787
ip->i_goal = dip->i_goal;
fs/gfs2/inode.c
788
ip->i_diskflags = 0;
fs/gfs2/inode.c
789
ip->i_eattr = 0;
fs/gfs2/inode.c
790
ip->i_height = 0;
fs/gfs2/inode.c
791
ip->i_depth = 0;
fs/gfs2/inode.c
792
ip->i_entries = 0;
fs/gfs2/inode.c
793
ip->i_no_addr = 0; /* Temporarily zero until real addr is assigned */
fs/gfs2/inode.c
799
ip->i_diskflags |= GFS2_DIF_JDATA;
fs/gfs2/inode.c
803
ip->i_diskflags |= (dip->i_diskflags & GFS2_DIF_INHERIT_JDATA);
fs/gfs2/inode.c
804
ip->i_diskflags |= GFS2_DIF_JDATA;
fs/gfs2/inode.c
805
ip->i_entries = 2;
fs/gfs2/inode.c
811
ip->i_diskflags |= GFS2_DIF_SYSTEM;
fs/gfs2/inode.c
822
error = alloc_dinode(ip, aflags, &blocks);
fs/gfs2/inode.c
828
error = gfs2_glock_get(sdp, ip->i_no_addr, &gfs2_inode_glops, CREATE, &ip->i_gl);
fs/gfs2/inode.c
832
error = gfs2_glock_get(sdp, ip->i_no_addr, &gfs2_iopen_glops, CREATE, &io_gl);
fs/gfs2/inode.c
836
io_gl->gl_no_formal_ino = ip->i_no_formal_ino;
fs/gfs2/inode.c
839
error = insert_inode_locked4(inode, ip->i_no_addr, iget_test, &ip->i_no_addr);
fs/gfs2/inode.c
846
&ip->i_iopen_gh);
fs/gfs2/inode.c
850
error = gfs2_glock_nq_init(ip->i_gl, LM_ST_EXCLUSIVE, GL_SKIP, &gh);
fs/gfs2/inode.c
853
clear_bit(GLF_INSTANTIATE_NEEDED, &ip->i_gl->gl_flags);
fs/gfs2/inode.c
860
gfs2_init_xattr(ip);
fs/gfs2/inode.c
863
init_dinode(dip, ip, symname);
fs/gfs2/inode.c
866
glock_set_object(ip->i_gl, ip);
fs/gfs2/inode.c
867
glock_set_object(io_gl, ip);
fs/gfs2/inode.c
885
error = security_inode_init_security(&ip->i_inode, &dip->i_inode, name,
fs/gfs2/inode.c
890
error = link_dinode(dip, name, ip, &da);
fs/gfs2/inode.c
903
gfs2_qa_put(ip);
fs/gfs2/inode.c
911
glock_clear_object(ip->i_gl, ip);
fs/gfs2/inode.c
912
glock_clear_object(io_gl, ip);
fs/gfs2/inode.c
914
gfs2_glock_dq_uninit(&ip->i_iopen_gh);
fs/gfs2/inode.c
919
if (ip->i_eattr)
fs/gfs2/inode.c
920
dealloc_error = gfs2_ea_dealloc(ip, xattr_initialized);
fs/gfs2/inode.c
924
dealloc_error = gfs2_dinode_dealloc(ip);
fs/gfs2/inode.c
927
ip->i_no_addr = 0;
fs/gfs2/inode.c
929
if (ip->i_gl) {
fs/gfs2/inode.c
930
gfs2_glock_put(ip->i_gl);
fs/gfs2/inode.c
931
ip->i_gl = NULL;
fs/gfs2/inode.c
933
gfs2_rs_deltree(&ip->i_res);
fs/gfs2/inode.c
934
gfs2_qa_put(ip);
fs/gfs2/inode.h
103
void gfs2_dinode_out(const struct gfs2_inode *ip, void *buf);
fs/gfs2/inode.h
16
ssize_t gfs2_internal_read(struct gfs2_inode *ip,
fs/gfs2/inode.h
20
static inline int gfs2_is_stuffed(const struct gfs2_inode *ip)
fs/gfs2/inode.h
22
return !ip->i_height;
fs/gfs2/inode.h
25
static inline int gfs2_is_jdata(const struct gfs2_inode *ip)
fs/gfs2/inode.h
27
return ip->i_diskflags & GFS2_DIF_JDATA;
fs/gfs2/inode.h
40
static inline int gfs2_is_dir(const struct gfs2_inode *ip)
fs/gfs2/inode.h
42
return S_ISDIR(ip->i_inode.i_mode);
fs/gfs2/inode.h
62
static inline int gfs2_check_inum(const struct gfs2_inode *ip, u64 no_addr,
fs/gfs2/inode.h
65
return ip->i_no_addr == no_addr && ip->i_no_formal_ino == no_formal_ino;
fs/gfs2/inode.h
68
static inline void gfs2_inum_out(const struct gfs2_inode *ip,
fs/gfs2/inode.h
71
dent->de_inum.no_formal_ino = cpu_to_be64(ip->i_no_formal_ino);
fs/gfs2/inode.h
72
dent->de_inum.no_addr = cpu_to_be64(ip->i_no_addr);
fs/gfs2/inode.h
96
int gfs2_dinode_dealloc(struct gfs2_inode *ip);
fs/gfs2/log.c
708
static void __ordered_del_inode(struct gfs2_inode *ip)
fs/gfs2/log.c
710
if (!list_empty(&ip->i_ordered))
fs/gfs2/log.c
711
list_del_init(&ip->i_ordered);
fs/gfs2/log.c
716
struct gfs2_inode *ip;
fs/gfs2/log.c
722
ip = list_first_entry(&sdp->sd_log_ordered, struct gfs2_inode, i_ordered);
fs/gfs2/log.c
723
if (ip->i_inode.i_mapping->nrpages == 0) {
fs/gfs2/log.c
724
__ordered_del_inode(ip);
fs/gfs2/log.c
727
list_move(&ip->i_ordered, &written);
fs/gfs2/log.c
729
filemap_fdatawrite(ip->i_inode.i_mapping);
fs/gfs2/log.c
738
struct gfs2_inode *ip;
fs/gfs2/log.c
742
ip = list_first_entry(&sdp->sd_log_ordered, struct gfs2_inode, i_ordered);
fs/gfs2/log.c
743
__ordered_del_inode(ip);
fs/gfs2/log.c
744
if (ip->i_inode.i_mapping->nrpages == 0)
fs/gfs2/log.c
747
filemap_fdatawait(ip->i_inode.i_mapping);
fs/gfs2/log.c
753
void gfs2_ordered_del_inode(struct gfs2_inode *ip)
fs/gfs2/log.c
755
struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode);
fs/gfs2/log.c
758
__ordered_del_inode(ip);
fs/gfs2/log.h
47
static inline void gfs2_ordered_add_inode(struct gfs2_inode *ip)
fs/gfs2/log.h
49
struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode);
fs/gfs2/log.h
51
if (gfs2_is_jdata(ip) || !gfs2_is_ordered(sdp))
fs/gfs2/log.h
54
if (list_empty(&ip->i_ordered)) {
fs/gfs2/log.h
56
if (list_empty(&ip->i_ordered))
fs/gfs2/log.h
57
list_add(&ip->i_ordered, &sdp->sd_log_ordered);
fs/gfs2/log.h
62
void gfs2_ordered_del_inode(struct gfs2_inode *ip);
fs/gfs2/lops.c
1003
struct gfs2_inode *ip = GFS2_I(jd->jd_inode);
fs/gfs2/lops.c
1004
struct gfs2_glock *gl = ip->i_gl;
fs/gfs2/lops.c
1051
struct gfs2_inode *ip = GFS2_I(jd->jd_inode);
fs/gfs2/lops.c
1055
gfs2_inode_metasync(ip->i_gl);
fs/gfs2/lops.c
1062
gfs2_inode_metasync(ip->i_gl);
fs/gfs2/lops.c
778
struct gfs2_inode *ip = GFS2_I(jd->jd_inode);
fs/gfs2/lops.c
780
struct gfs2_glock *gl = ip->i_gl;
fs/gfs2/lops.c
831
struct gfs2_inode *ip = GFS2_I(jd->jd_inode);
fs/gfs2/lops.c
835
gfs2_inode_metasync(ip->i_gl);
fs/gfs2/lops.c
841
gfs2_inode_metasync(ip->i_gl);
fs/gfs2/main.c
36
struct gfs2_inode *ip = foo;
fs/gfs2/main.c
38
inode_init_once(&ip->i_inode);
fs/gfs2/main.c
39
atomic_set(&ip->i_sizehint, 0);
fs/gfs2/main.c
40
init_rwsem(&ip->i_rw_mutex);
fs/gfs2/main.c
41
INIT_LIST_HEAD(&ip->i_ordered);
fs/gfs2/main.c
42
ip->i_qadata = NULL;
fs/gfs2/main.c
43
gfs2_holder_mark_uninitialized(&ip->i_rgd_gh);
fs/gfs2/main.c
44
memset(&ip->i_res, 0, sizeof(ip->i_res));
fs/gfs2/main.c
45
RB_CLEAR_NODE(&ip->i_res.rs_node);
fs/gfs2/main.c
46
ip->i_hash_cache = NULL;
fs/gfs2/main.c
47
gfs2_holder_mark_uninitialized(&ip->i_iopen_gh);
fs/gfs2/meta_io.c
410
static struct buffer_head *gfs2_getjdatabuf(struct gfs2_inode *ip, u64 blkno)
fs/gfs2/meta_io.c
412
struct address_space *mapping = ip->i_inode.i_mapping;
fs/gfs2/meta_io.c
413
struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode);
fs/gfs2/meta_io.c
439
void gfs2_journal_wipe(struct gfs2_inode *ip, u64 bstart, u32 blen)
fs/gfs2/meta_io.c
441
struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode);
fs/gfs2/meta_io.c
446
if (!ip->i_gl)
fs/gfs2/meta_io.c
452
bh = gfs2_getbuf(ip->i_gl, bstart, NO_CREATE);
fs/gfs2/meta_io.c
453
if (!bh && gfs2_is_jdata(ip)) {
fs/gfs2/meta_io.c
454
bh = gfs2_getjdatabuf(ip, bstart);
fs/gfs2/meta_io.c
483
int gfs2_meta_buffer(struct gfs2_inode *ip, u32 mtype, u64 num,
fs/gfs2/meta_io.c
486
struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode);
fs/gfs2/meta_io.c
487
struct gfs2_glock *gl = ip->i_gl;
fs/gfs2/meta_io.c
492
if (num == ip->i_no_addr)
fs/gfs2/meta_io.c
493
rahead = ip->i_rahead;
fs/gfs2/meta_io.h
63
void gfs2_journal_wipe(struct gfs2_inode *ip, u64 bstart, u32 blen);
fs/gfs2/meta_io.h
64
int gfs2_meta_buffer(struct gfs2_inode *ip, u32 mtype, u64 num,
fs/gfs2/meta_io.h
67
static inline int gfs2_meta_inode_buffer(struct gfs2_inode *ip,
fs/gfs2/meta_io.h
70
return gfs2_meta_buffer(ip, GFS2_METATYPE_DI, ip->i_no_addr, bhp);
fs/gfs2/ops_fstype.c
612
struct gfs2_inode *ip;
fs/gfs2/ops_fstype.c
657
ip = GFS2_I(sdp->sd_sc_inode);
fs/gfs2/ops_fstype.c
658
error = gfs2_glock_nq_init(ip->i_gl, LM_ST_EXCLUSIVE, GL_NOPID,
fs/gfs2/ops_fstype.c
665
error = gfs2_meta_inode_buffer(ip, &sdp->sd_sc_bh);
fs/gfs2/ops_fstype.c
698
struct gfs2_inode *ip;
fs/gfs2/ops_fstype.c
752
ip = GFS2_I(sdp->sd_jdesc->jd_inode);
fs/gfs2/ops_fstype.c
753
error = gfs2_glock_nq_init(ip->i_gl, LM_ST_SHARED,
fs/gfs2/ops_fstype.c
894
struct gfs2_inode *ip;
fs/gfs2/ops_fstype.c
921
ip = GFS2_I(sdp->sd_qc_inode);
fs/gfs2/ops_fstype.c
922
error = gfs2_glock_nq_init(ip->i_gl, LM_ST_EXCLUSIVE, GL_NOPID,
fs/gfs2/quota.c
1006
error = gfs2_internal_read(ip, (char *)&q, &pos, sizeof(q));
fs/gfs2/quota.c
1027
struct gfs2_inode *ip = GFS2_I(sdp->sd_quota_inode);
fs/gfs2/quota.c
1051
error = gfs2_glock_nq_init(ip->i_gl, LM_ST_SHARED, 0, &i_gh);
fs/gfs2/quota.c
1074
int gfs2_quota_lock(struct gfs2_inode *ip, kuid_t uid, kgid_t gid)
fs/gfs2/quota.c
1076
struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode);
fs/gfs2/quota.c
1084
error = gfs2_quota_hold(ip, uid, gid);
fs/gfs2/quota.c
1088
sort(ip->i_qadata->qa_qd, ip->i_qadata->qa_qd_num,
fs/gfs2/quota.c
1091
for (x = 0; x < ip->i_qadata->qa_qd_num; x++) {
fs/gfs2/quota.c
1092
qd = ip->i_qadata->qa_qd[x];
fs/gfs2/quota.c
1093
error = do_glock(qd, NO_FORCE, &ip->i_qadata->qa_qd_ghs[x]);
fs/gfs2/quota.c
1099
set_bit(GIF_QD_LOCKED, &ip->i_flags);
fs/gfs2/quota.c
1102
gfs2_glock_dq_uninit(&ip->i_qadata->qa_qd_ghs[x]);
fs/gfs2/quota.c
1103
gfs2_quota_unhold(ip);
fs/gfs2/quota.c
1145
void gfs2_quota_unlock(struct gfs2_inode *ip)
fs/gfs2/quota.c
1147
struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode);
fs/gfs2/quota.c
1152
if (!test_and_clear_bit(GIF_QD_LOCKED, &ip->i_flags))
fs/gfs2/quota.c
1155
for (x = 0; x < ip->i_qadata->qa_qd_num; x++) {
fs/gfs2/quota.c
1160
qd = ip->i_qadata->qa_qd[x];
fs/gfs2/quota.c
1163
gfs2_glock_dq_uninit(&ip->i_qadata->qa_qd_ghs[x]);
fs/gfs2/quota.c
1192
gfs2_quota_unhold(ip);
fs/gfs2/quota.c
1226
int gfs2_quota_check(struct gfs2_inode *ip, kuid_t uid, kgid_t gid,
fs/gfs2/quota.c
1229
struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode);
fs/gfs2/quota.c
1236
if (!test_bit(GIF_QD_LOCKED, &ip->i_flags))
fs/gfs2/quota.c
1239
for (x = 0; x < ip->i_qadata->qa_qd_num; x++) {
fs/gfs2/quota.c
1240
qd = ip->i_qadata->qa_qd[x];
fs/gfs2/quota.c
1284
void gfs2_quota_change(struct gfs2_inode *ip, s64 change,
fs/gfs2/quota.c
1289
struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode);
fs/gfs2/quota.c
1294
if (ip->i_diskflags & GFS2_DIF_SYSTEM)
fs/gfs2/quota.c
1297
if (gfs2_assert_withdraw(sdp, ip->i_qadata &&
fs/gfs2/quota.c
1298
ip->i_qadata->qa_ref > 0))
fs/gfs2/quota.c
1300
for (x = 0; x < ip->i_qadata->qa_qd_num; x++) {
fs/gfs2/quota.c
1301
qd = ip->i_qadata->qa_qd[x];
fs/gfs2/quota.c
1391
struct gfs2_inode *ip = GFS2_I(sdp->sd_qc_inode);
fs/gfs2/quota.c
1423
error = gfs2_get_extent(&ip->i_inode, x, &dblock, &extlen);
fs/gfs2/quota.c
1428
bh = gfs2_meta_ra(ip->i_gl, dblock, extlen);
fs/gfs2/quota.c
1705
struct gfs2_inode *ip = GFS2_I(sdp->sd_quota_inode);
fs/gfs2/quota.c
1728
error = gfs2_qa_get(ip);
fs/gfs2/quota.c
1732
inode_lock(&ip->i_inode);
fs/gfs2/quota.c
1736
error = gfs2_glock_nq_init(ip->i_gl, LM_ST_EXCLUSIVE, 0, &i_gh);
fs/gfs2/quota.c
1762
alloc_required = gfs2_write_alloc_required(ip, offset, sizeof(struct gfs2_quota));
fs/gfs2/quota.c
1763
if (gfs2_is_stuffed(ip))
fs/gfs2/quota.c
1767
gfs2_write_calc_reserv(ip, sizeof(struct gfs2_quota),
fs/gfs2/quota.c
1771
error = gfs2_inplace_reserve(ip, &ap);
fs/gfs2/quota.c
1774
blocks += gfs2_rg_blocks(ip, blocks);
fs/gfs2/quota.c
1791
gfs2_inplace_release(ip);
fs/gfs2/quota.c
1797
gfs2_qa_put(ip);
fs/gfs2/quota.c
1798
inode_unlock(&ip->i_inode);
fs/gfs2/quota.c
397
struct gfs2_inode *ip = GFS2_I(inode);
fs/gfs2/quota.c
423
error = gfs2_meta_read(ip->i_gl, iomap.addr >> inode->i_blkbits,
fs/gfs2/quota.c
544
int gfs2_qa_get(struct gfs2_inode *ip)
fs/gfs2/quota.c
546
struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode);
fs/gfs2/quota.c
547
struct inode *inode = &ip->i_inode;
fs/gfs2/quota.c
553
if (ip->i_qadata == NULL) {
fs/gfs2/quota.c
562
if (ip->i_qadata == NULL)
fs/gfs2/quota.c
563
ip->i_qadata = tmp;
fs/gfs2/quota.c
567
ip->i_qadata->qa_ref++;
fs/gfs2/quota.c
572
void gfs2_qa_put(struct gfs2_inode *ip)
fs/gfs2/quota.c
574
struct inode *inode = &ip->i_inode;
fs/gfs2/quota.c
577
if (ip->i_qadata && --ip->i_qadata->qa_ref == 0) {
fs/gfs2/quota.c
578
kmem_cache_free(gfs2_qadata_cachep, ip->i_qadata);
fs/gfs2/quota.c
579
ip->i_qadata = NULL;
fs/gfs2/quota.c
584
int gfs2_quota_hold(struct gfs2_inode *ip, kuid_t uid, kgid_t gid)
fs/gfs2/quota.c
586
struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode);
fs/gfs2/quota.c
593
error = gfs2_qa_get(ip);
fs/gfs2/quota.c
597
qd = ip->i_qadata->qa_qd;
fs/gfs2/quota.c
599
if (gfs2_assert_warn(sdp, !ip->i_qadata->qa_qd_num) ||
fs/gfs2/quota.c
600
gfs2_assert_warn(sdp, !test_bit(GIF_QD_LOCKED, &ip->i_flags))) {
fs/gfs2/quota.c
602
gfs2_qa_put(ip);
fs/gfs2/quota.c
606
error = qdsb_get(sdp, make_kqid_uid(ip->i_inode.i_uid), qd);
fs/gfs2/quota.c
609
ip->i_qadata->qa_qd_num++;
fs/gfs2/quota.c
612
error = qdsb_get(sdp, make_kqid_gid(ip->i_inode.i_gid), qd);
fs/gfs2/quota.c
615
ip->i_qadata->qa_qd_num++;
fs/gfs2/quota.c
619
!uid_eq(uid, ip->i_inode.i_uid)) {
fs/gfs2/quota.c
623
ip->i_qadata->qa_qd_num++;
fs/gfs2/quota.c
628
!gid_eq(gid, ip->i_inode.i_gid)) {
fs/gfs2/quota.c
632
ip->i_qadata->qa_qd_num++;
fs/gfs2/quota.c
638
gfs2_quota_unhold(ip);
fs/gfs2/quota.c
643
void gfs2_quota_unhold(struct gfs2_inode *ip)
fs/gfs2/quota.c
645
struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode);
fs/gfs2/quota.c
648
if (ip->i_qadata == NULL)
fs/gfs2/quota.c
651
gfs2_assert_warn(sdp, !test_bit(GIF_QD_LOCKED, &ip->i_flags));
fs/gfs2/quota.c
653
for (x = 0; x < ip->i_qadata->qa_qd_num; x++) {
fs/gfs2/quota.c
654
qdsb_put(ip->i_qadata->qa_qd[x]);
fs/gfs2/quota.c
655
ip->i_qadata->qa_qd[x] = NULL;
fs/gfs2/quota.c
657
ip->i_qadata->qa_qd_num = 0;
fs/gfs2/quota.c
658
gfs2_qa_put(ip);
fs/gfs2/quota.c
676
struct gfs2_inode *ip = GFS2_I(sdp->sd_qc_inode);
fs/gfs2/quota.c
681
gfs2_trans_add_meta(ip->i_gl, qd->qd_bh);
fs/gfs2/quota.c
729
struct gfs2_inode *ip = GFS2_I(sdp->sd_quota_inode);
fs/gfs2/quota.c
730
struct inode *inode = &ip->i_inode;
fs/gfs2/quota.c
769
gfs2_trans_add_data(ip->i_gl, bh);
fs/gfs2/quota.c
841
struct gfs2_inode *ip = GFS2_I(sdp->sd_quota_inode);
fs/gfs2/quota.c
842
struct inode *inode = &ip->i_inode;
fs/gfs2/quota.c
847
if (gfs2_is_stuffed(ip)) {
fs/gfs2/quota.c
848
err = gfs2_unstuff_dinode(ip);
fs/gfs2/quota.c
854
err = gfs2_internal_read(ip, (char *)&q, &loc, sizeof(q));
fs/gfs2/quota.c
897
struct gfs2_inode *ip = GFS2_I(sdp->sd_quota_inode);
fs/gfs2/quota.c
908
gfs2_write_calc_reserv(ip, sizeof(struct gfs2_quota),
fs/gfs2/quota.c
916
inode_lock(&ip->i_inode);
fs/gfs2/quota.c
924
error = gfs2_glock_nq_init(ip->i_gl, LM_ST_EXCLUSIVE, 0, &i_gh);
fs/gfs2/quota.c
930
if (gfs2_write_alloc_required(ip, offset,
fs/gfs2/quota.c
948
error = gfs2_inplace_reserve(ip, &ap);
fs/gfs2/quota.c
953
blocks += gfs2_rg_blocks(ip, reserved) + nalloc * ind_blocks + RES_STATFS;
fs/gfs2/quota.c
974
gfs2_inplace_release(ip);
fs/gfs2/quota.c
980
inode_unlock(&ip->i_inode);
fs/gfs2/quota.c
982
gfs2_log_flush(glock_sbd(ip->i_gl), ip->i_gl,
fs/gfs2/quota.c
998
struct gfs2_inode *ip = GFS2_I(sdp->sd_quota_inode);
fs/gfs2/quota.h
18
int gfs2_qa_get(struct gfs2_inode *ip);
fs/gfs2/quota.h
19
void gfs2_qa_put(struct gfs2_inode *ip);
fs/gfs2/quota.h
20
int gfs2_quota_hold(struct gfs2_inode *ip, kuid_t uid, kgid_t gid);
fs/gfs2/quota.h
21
void gfs2_quota_unhold(struct gfs2_inode *ip);
fs/gfs2/quota.h
23
int gfs2_quota_lock(struct gfs2_inode *ip, kuid_t uid, kgid_t gid);
fs/gfs2/quota.h
24
void gfs2_quota_unlock(struct gfs2_inode *ip);
fs/gfs2/quota.h
26
int gfs2_quota_check(struct gfs2_inode *ip, kuid_t uid, kgid_t gid,
fs/gfs2/quota.h
28
void gfs2_quota_change(struct gfs2_inode *ip, s64 change,
fs/gfs2/quota.h
40
static inline int gfs2_quota_lock_check(struct gfs2_inode *ip,
fs/gfs2/quota.h
43
struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode);
fs/gfs2/quota.h
50
ret = gfs2_quota_lock(ip, NO_UID_QUOTA_CHANGE, NO_GID_QUOTA_CHANGE);
fs/gfs2/quota.h
55
ret = gfs2_quota_check(ip, ip->i_inode.i_uid, ip->i_inode.i_gid, ap);
fs/gfs2/quota.h
57
gfs2_quota_unlock(ip);
fs/gfs2/recovery.c
309
struct gfs2_inode *ip;
fs/gfs2/recovery.c
315
ip = GFS2_I(inode);
fs/gfs2/recovery.c
317
error = gfs2_meta_inode_buffer(ip, &bh);
fs/gfs2/recovery.c
348
gfs2_inode_metasync(ip->i_gl);
fs/gfs2/recovery.c
35
struct gfs2_inode *ip = GFS2_I(jd->jd_inode);
fs/gfs2/recovery.c
36
struct gfs2_glock *gl = ip->i_gl;
fs/gfs2/recovery.c
401
struct gfs2_inode *ip = GFS2_I(jd->jd_inode);
fs/gfs2/recovery.c
42
error = gfs2_get_extent(&ip->i_inode, blk, &dblock, &extlen);
fs/gfs2/recovery.c
443
error = gfs2_glock_nq_init(ip->i_gl, LM_ST_SHARED,
fs/gfs2/recovery.c
46
gfs2_consist_inode(ip);
fs/gfs2/rgrp.c
1001
error = read_rindex_entry(ip);
fs/gfs2/rgrp.c
1036
struct gfs2_inode *ip = GFS2_I(sdp->sd_rindex);
fs/gfs2/rgrp.c
1037
struct gfs2_glock *gl = ip->i_gl;
fs/gfs2/rgrp.c
1051
error = gfs2_ri_update(ip);
fs/gfs2/rgrp.c
1488
static void rs_insert(struct gfs2_inode *ip)
fs/gfs2/rgrp.c
1492
struct gfs2_blkreserv *rs = &ip->i_res;
fs/gfs2/rgrp.c
1562
static void rg_mblk_search(struct gfs2_rgrpd *rgd, struct gfs2_inode *ip,
fs/gfs2/rgrp.c
1567
struct gfs2_blkreserv *rs = &ip->i_res;
fs/gfs2/rgrp.c
1571
struct inode *inode = &ip->i_inode;
fs/gfs2/rgrp.c
1585
extlen = max_t(u32, atomic_read(&ip->i_sizehint), ap->target);
fs/gfs2/rgrp.c
1592
if (rgrp_contains_block(rgd, ip->i_goal))
fs/gfs2/rgrp.c
1593
goal = ip->i_goal;
fs/gfs2/rgrp.c
1600
ret = gfs2_rbm_find(&rbm, GFS2_BLKST_FREE, &extlen, &ip->i_res, true);
fs/gfs2/rgrp.c
1604
rs_insert(ip);
fs/gfs2/rgrp.c
1847
struct gfs2_inode *ip;
fs/gfs2/rgrp.c
1880
ip = gl->gl_object;
fs/gfs2/rgrp.c
1882
if (ip || !gfs2_queue_verify_delete(gl, false))
fs/gfs2/rgrp.c
1987
static u32 gfs2_orlov_skip(const struct gfs2_inode *ip)
fs/gfs2/rgrp.c
1989
const struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode);
fs/gfs2/rgrp.c
2043
int gfs2_inplace_reserve(struct gfs2_inode *ip, struct gfs2_alloc_parms *ap)
fs/gfs2/rgrp.c
2045
struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode);
fs/gfs2/rgrp.c
2047
struct gfs2_blkreserv *rs = &ip->i_res;
fs/gfs2/rgrp.c
2064
rgrp_contains_block(rs->rs_rgd, ip->i_goal)) {
fs/gfs2/rgrp.c
2067
check_and_update_goal(ip);
fs/gfs2/rgrp.c
2068
rs->rs_rgd = begin = gfs2_blk2rgrpd(sdp, ip->i_goal, 1);
fs/gfs2/rgrp.c
2070
if (S_ISDIR(ip->i_inode.i_mode) && (ap->aflags & GFS2_AF_ORLOV))
fs/gfs2/rgrp.c
2071
skip = gfs2_orlov_skip(ip);
fs/gfs2/rgrp.c
2095
&ip->i_rgd_gh);
fs/gfs2/rgrp.c
2104
&ip->i_rgd_gh);
fs/gfs2/rgrp.c
2107
gfs2_glock_dq_uninit(&ip->i_rgd_gh);
fs/gfs2/rgrp.c
2120
error = gfs2_instantiate(&ip->i_rgd_gh);
fs/gfs2/rgrp.c
2127
rg_mblk_search(rs->rs_rgd, ip, ap);
fs/gfs2/rgrp.c
2153
ip->i_no_addr);
fs/gfs2/rgrp.c
2163
gfs2_glock_dq_uninit(&ip->i_rgd_gh);
fs/gfs2/rgrp.c
2177
if (ip == GFS2_I(sdp->sd_rindex) && !sdp->sd_rindex_uptodate) {
fs/gfs2/rgrp.c
2178
error = gfs2_ri_update(ip);
fs/gfs2/rgrp.c
2201
void gfs2_inplace_release(struct gfs2_inode *ip)
fs/gfs2/rgrp.c
2203
struct gfs2_blkreserv *rs = &ip->i_res;
fs/gfs2/rgrp.c
2214
if (gfs2_holder_initialized(&ip->i_rgd_gh))
fs/gfs2/rgrp.c
2215
gfs2_glock_dq_uninit(&ip->i_rgd_gh);
fs/gfs2/rgrp.c
2346
static void gfs2_adjust_reservation(struct gfs2_inode *ip,
fs/gfs2/rgrp.c
2349
struct gfs2_blkreserv *rs = &ip->i_res;
fs/gfs2/rgrp.c
2370
atomic_add(RGRP_RSRV_ADDBLKS, &ip->i_sizehint);
fs/gfs2/rgrp.c
2388
const struct gfs2_inode *ip, bool dinode)
fs/gfs2/rgrp.c
2392
if (gfs2_rs_active(&ip->i_res)) {
fs/gfs2/rgrp.c
2393
goal = ip->i_res.rs_start;
fs/gfs2/rgrp.c
2395
if (!dinode && rgrp_contains_block(rbm->rgd, ip->i_goal))
fs/gfs2/rgrp.c
2396
goal = ip->i_goal;
fs/gfs2/rgrp.c
2416
int gfs2_alloc_blocks(struct gfs2_inode *ip, u64 *bn, unsigned int *nblocks,
fs/gfs2/rgrp.c
2419
struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode);
fs/gfs2/rgrp.c
2421
struct gfs2_rbm rbm = { .rgd = ip->i_res.rs_rgd, };
fs/gfs2/rgrp.c
2426
BUG_ON(ip->i_res.rs_reserved < *nblocks);
fs/gfs2/rgrp.c
2429
if (gfs2_rs_active(&ip->i_res)) {
fs/gfs2/rgrp.c
2430
gfs2_set_alloc_start(&rbm, ip, dinode);
fs/gfs2/rgrp.c
2431
error = gfs2_rbm_find(&rbm, GFS2_BLKST_FREE, &minext, &ip->i_res, false);
fs/gfs2/rgrp.c
2434
gfs2_set_alloc_start(&rbm, ip, dinode);
fs/gfs2/rgrp.c
2441
(unsigned long long)ip->i_no_addr, error, *nblocks,
fs/gfs2/rgrp.c
2451
ip->i_goal = block + *nblocks - 1;
fs/gfs2/rgrp.c
2452
error = gfs2_meta_inode_buffer(ip, &dibh);
fs/gfs2/rgrp.c
2456
gfs2_trans_add_meta(ip->i_gl, dibh);
fs/gfs2/rgrp.c
2458
cpu_to_be64(ip->i_goal);
fs/gfs2/rgrp.c
2463
gfs2_adjust_reservation(ip, &rbm, *nblocks);
fs/gfs2/rgrp.c
2483
ip->i_generation = generation;
fs/gfs2/rgrp.c
2494
gfs2_quota_change(ip, *nblocks, ip->i_inode.i_uid, ip->i_inode.i_gid);
fs/gfs2/rgrp.c
2496
trace_gfs2_block_alloc(ip, rbm.rgd, block, *nblocks,
fs/gfs2/rgrp.c
2517
void __gfs2_free_blocks(struct gfs2_inode *ip, struct gfs2_rgrpd *rgd,
fs/gfs2/rgrp.c
2520
struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode);
fs/gfs2/rgrp.c
2524
trace_gfs2_block_alloc(ip, rgd, bstart, blen, GFS2_BLKST_FREE);
fs/gfs2/rgrp.c
2532
if (meta || ip->i_depth || gfs2_is_jdata(ip))
fs/gfs2/rgrp.c
2533
gfs2_journal_wipe(ip, bstart, blen);
fs/gfs2/rgrp.c
2545
void gfs2_free_meta(struct gfs2_inode *ip, struct gfs2_rgrpd *rgd,
fs/gfs2/rgrp.c
2548
struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode);
fs/gfs2/rgrp.c
2550
__gfs2_free_blocks(ip, rgd, bstart, blen, 1);
fs/gfs2/rgrp.c
2552
gfs2_quota_change(ip, -(s64)blen, ip->i_inode.i_uid, ip->i_inode.i_gid);
fs/gfs2/rgrp.c
2557
struct gfs2_inode *ip = GFS2_I(inode);
fs/gfs2/rgrp.c
2560
u64 blkno = ip->i_no_addr;
fs/gfs2/rgrp.c
2567
trace_gfs2_block_alloc(ip, rgd, blkno, 1, GFS2_BLKST_UNLINKED);
fs/gfs2/rgrp.c
2574
void gfs2_free_di(struct gfs2_rgrpd *rgd, struct gfs2_inode *ip)
fs/gfs2/rgrp.c
2579
rgblk_free(sdp, rgd, ip->i_no_addr, 1, GFS2_BLKST_FREE);
fs/gfs2/rgrp.c
2591
trace_gfs2_block_alloc(ip, rgd, ip->i_no_addr, 1, GFS2_BLKST_FREE);
fs/gfs2/rgrp.c
2592
gfs2_quota_change(ip, -1, ip->i_inode.i_uid, ip->i_inode.i_gid);
fs/gfs2/rgrp.c
2593
gfs2_journal_wipe(ip, ip->i_no_addr, 1);
fs/gfs2/rgrp.c
2658
void gfs2_rlist_add(struct gfs2_inode *ip, struct gfs2_rgrp_list *rlist,
fs/gfs2/rgrp.c
2661
struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode);
fs/gfs2/rgrp.c
2680
rgd = ip->i_res.rs_rgd;
fs/gfs2/rgrp.c
601
void check_and_update_goal(struct gfs2_inode *ip)
fs/gfs2/rgrp.c
603
struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode);
fs/gfs2/rgrp.c
604
if (!ip->i_goal || gfs2_blk2rgrpd(sdp, ip->i_goal, 1) == NULL)
fs/gfs2/rgrp.c
605
ip->i_goal = ip->i_no_addr;
fs/gfs2/rgrp.c
622
struct gfs2_inode *ip = container_of(rs, struct gfs2_inode, i_res);
fs/gfs2/rgrp.c
626
(unsigned long long)ip->i_no_addr,
fs/gfs2/rgrp.c
685
void gfs2_rs_delete(struct gfs2_inode *ip)
fs/gfs2/rgrp.c
687
struct inode *inode = &ip->i_inode;
fs/gfs2/rgrp.c
689
down_write(&ip->i_rw_mutex);
fs/gfs2/rgrp.c
691
gfs2_rs_deltree(&ip->i_res);
fs/gfs2/rgrp.c
692
up_write(&ip->i_rw_mutex);
fs/gfs2/rgrp.c
845
struct gfs2_inode *ip = GFS2_I(inode);
fs/gfs2/rgrp.c
854
error = gfs2_internal_read(ip, buf, &pos,
fs/gfs2/rgrp.c
895
static int read_rindex_entry(struct gfs2_inode *ip)
fs/gfs2/rgrp.c
897
struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode);
fs/gfs2/rgrp.c
903
if (pos >= i_size_read(&ip->i_inode))
fs/gfs2/rgrp.c
906
error = gfs2_internal_read(ip, (char *)&buf, &pos,
fs/gfs2/rgrp.c
995
static int gfs2_ri_update(struct gfs2_inode *ip)
fs/gfs2/rgrp.c
997
struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode);
fs/gfs2/rgrp.h
37
struct gfs2_alloc *gfs2_alloc_get(struct gfs2_inode *ip);
fs/gfs2/rgrp.h
40
int gfs2_inplace_reserve(struct gfs2_inode *ip,
fs/gfs2/rgrp.h
42
void gfs2_inplace_release(struct gfs2_inode *ip);
fs/gfs2/rgrp.h
44
int gfs2_alloc_blocks(struct gfs2_inode *ip, u64 *bn, unsigned int *n,
fs/gfs2/rgrp.h
48
void gfs2_rs_delete(struct gfs2_inode *ip);
fs/gfs2/rgrp.h
49
void __gfs2_free_blocks(struct gfs2_inode *ip, struct gfs2_rgrpd *rgd,
fs/gfs2/rgrp.h
51
void gfs2_free_meta(struct gfs2_inode *ip, struct gfs2_rgrpd *rgd,
fs/gfs2/rgrp.h
53
void gfs2_free_di(struct gfs2_rgrpd *rgd, struct gfs2_inode *ip);
fs/gfs2/rgrp.h
65
void gfs2_rlist_add(struct gfs2_inode *ip, struct gfs2_rgrp_list *rlist,
fs/gfs2/rgrp.h
92
void check_and_update_goal(struct gfs2_inode *ip);
fs/gfs2/super.c
1007
struct gfs2_inode *ip = GFS2_I(inode);
fs/gfs2/super.c
1011
gfs2_holder_initialized(&ip->i_iopen_gh)) {
fs/gfs2/super.c
1012
struct gfs2_glock *gl = ip->i_iopen_gh.gh_gl;
fs/gfs2/super.c
1024
gfs2_holder_initialized(&ip->i_iopen_gh)) {
fs/gfs2/super.c
1025
struct gfs2_glock *gl = ip->i_iopen_gh.gh_gl;
fs/gfs2/super.c
109
struct gfs2_inode *ip = GFS2_I(jd->jd_inode);
fs/gfs2/super.c
1177
struct gfs2_inode *ip = GFS2_I(inode);
fs/gfs2/super.c
1179
struct gfs2_holder *gh = &ip->i_iopen_gh;
fs/gfs2/super.c
118
if (gfs2_write_alloc_required(ip, 0, size)) {
fs/gfs2/super.c
119
gfs2_consist_inode(ip);
fs/gfs2/super.c
1212
glock_needs_demote(ip->i_gl),
fs/gfs2/super.c
1216
if (glock_needs_demote(ip->i_gl))
fs/gfs2/super.c
1239
struct gfs2_inode *ip = GFS2_I(inode);
fs/gfs2/super.c
1244
if (gfs2_holder_initialized(&ip->i_iopen_gh) &&
fs/gfs2/super.c
1245
test_bit(GLF_DEFER_DELETE, &ip->i_iopen_gh.gh_gl->gl_flags))
fs/gfs2/super.c
1253
ret = gfs2_glock_nq_init(ip->i_gl, LM_ST_EXCLUSIVE, GL_SKIP, gh);
fs/gfs2/super.c
1257
if (gfs2_inode_already_deleted(ip->i_gl, ip->i_no_formal_ino))
fs/gfs2/super.c
1259
ret = gfs2_check_blk_type(sdp, ip->i_no_addr, GFS2_BLKST_UNLINKED);
fs/gfs2/super.c
1273
if (gfs2_holder_initialized(&ip->i_iopen_gh) &&
fs/gfs2/super.c
1274
test_bit(HIF_HOLDER, &ip->i_iopen_gh.gh_iflags))
fs/gfs2/super.c
1285
struct gfs2_inode *ip = GFS2_I(inode);
fs/gfs2/super.c
1289
(ip->i_diskflags & GFS2_DIF_EXHASH)) {
fs/gfs2/super.c
1290
ret = gfs2_dir_exhash_dealloc(ip);
fs/gfs2/super.c
1295
if (ip->i_eattr) {
fs/gfs2/super.c
1296
ret = gfs2_ea_dealloc(ip, true);
fs/gfs2/super.c
1301
if (!gfs2_is_stuffed(ip)) {
fs/gfs2/super.c
1302
ret = gfs2_file_dealloc(ip);
fs/gfs2/super.c
1319
ret = gfs2_dinode_dealloc(ip);
fs/gfs2/super.c
1320
if (!ret && ip->i_gl)
fs/gfs2/super.c
1321
gfs2_inode_remember_delete(ip->i_gl, ip->i_no_formal_ino);
fs/gfs2/super.c
1335
struct gfs2_inode *ip = GFS2_I(inode);
fs/gfs2/super.c
1339
gfs2_log_flush(sdp, ip->i_gl, GFS2_LOG_HEAD_FLUSH_NORMAL |
fs/gfs2/super.c
1341
metamapping = gfs2_glock2aspace(ip->i_gl);
fs/gfs2/super.c
1342
if (test_bit(GLF_DIRTY, &ip->i_gl->gl_flags)) {
fs/gfs2/super.c
1347
gfs2_ail_flush(ip->i_gl, 0);
fs/gfs2/super.c
135
struct gfs2_inode *ip = GFS2_I(sdp->sd_jdesc->jd_inode);
fs/gfs2/super.c
136
struct gfs2_glock *j_gl = ip->i_gl;
fs/gfs2/super.c
1385
struct gfs2_inode *ip = GFS2_I(inode);
fs/gfs2/super.c
1391
if (inode->i_nlink || sb_rdonly(sb) || !ip->i_no_addr)
fs/gfs2/super.c
1405
struct gfs2_glock *io_gl = ip->i_iopen_gh.gh_gl;
fs/gfs2/super.c
1420
if (gfs2_rs_active(&ip->i_res))
fs/gfs2/super.c
1421
gfs2_rs_deltree(&ip->i_res);
fs/gfs2/super.c
1429
if (ip->i_qadata)
fs/gfs2/super.c
1430
gfs2_assert_warn(sdp, ip->i_qadata->qa_ref == 0);
fs/gfs2/super.c
1431
gfs2_rs_deltree(&ip->i_res);
fs/gfs2/super.c
1432
gfs2_ordered_del_inode(ip);
fs/gfs2/super.c
1434
gfs2_dir_hash_inval(ip);
fs/gfs2/super.c
1435
if (gfs2_holder_initialized(&ip->i_iopen_gh)) {
fs/gfs2/super.c
1436
struct gfs2_glock *gl = ip->i_iopen_gh.gh_gl;
fs/gfs2/super.c
1438
glock_clear_object(gl, ip);
fs/gfs2/super.c
1440
ip->i_iopen_gh.gh_flags |= GL_NOCACHE;
fs/gfs2/super.c
1441
gfs2_glock_dq_uninit(&ip->i_iopen_gh);
fs/gfs2/super.c
1444
if (ip->i_gl) {
fs/gfs2/super.c
1445
glock_clear_object(ip->i_gl, ip);
fs/gfs2/super.c
1446
wait_on_bit_io(&ip->i_flags, GIF_GLOP_PENDING, TASK_UNINTERRUPTIBLE);
fs/gfs2/super.c
1447
gfs2_glock_put_eventually(ip->i_gl);
fs/gfs2/super.c
1448
rcu_assign_pointer(ip->i_gl, NULL);
fs/gfs2/super.c
1454
struct gfs2_inode *ip;
fs/gfs2/super.c
1456
ip = alloc_inode_sb(sb, gfs2_inode_cachep, GFP_KERNEL);
fs/gfs2/super.c
1457
if (!ip)
fs/gfs2/super.c
1459
ip->i_no_addr = 0;
fs/gfs2/super.c
1460
ip->i_no_formal_ino = 0;
fs/gfs2/super.c
1461
ip->i_flags = 0;
fs/gfs2/super.c
1462
ip->i_gl = NULL;
fs/gfs2/super.c
1463
gfs2_holder_mark_uninitialized(&ip->i_iopen_gh);
fs/gfs2/super.c
1464
memset(&ip->i_res, 0, sizeof(ip->i_res));
fs/gfs2/super.c
1465
RB_CLEAR_NODE(&ip->i_res.rs_node);
fs/gfs2/super.c
1466
ip->i_diskflags = 0;
fs/gfs2/super.c
1467
ip->i_rahead = 0;
fs/gfs2/super.c
1468
return &ip->i_inode;
fs/gfs2/super.c
326
struct gfs2_inode *ip;
fs/gfs2/super.c
344
ip = GFS2_I(jd->jd_inode);
fs/gfs2/super.c
345
error = gfs2_glock_nq_init(ip->i_gl, LM_ST_SHARED, 0, &lfcc->gh);
fs/gfs2/super.c
393
void gfs2_dinode_out(const struct gfs2_inode *ip, void *buf)
fs/gfs2/super.c
395
const struct inode *inode = &ip->i_inode;
fs/gfs2/super.c
401
str->di_num.no_addr = cpu_to_be64(ip->i_no_addr);
fs/gfs2/super.c
402
str->di_num.no_formal_ino = cpu_to_be64(ip->i_no_formal_ino);
fs/gfs2/super.c
413
str->di_goal_meta = cpu_to_be64(ip->i_goal);
fs/gfs2/super.c
414
str->di_goal_data = cpu_to_be64(ip->i_goal);
fs/gfs2/super.c
415
str->di_generation = cpu_to_be64(ip->i_generation);
fs/gfs2/super.c
417
str->di_flags = cpu_to_be32(ip->i_diskflags);
fs/gfs2/super.c
418
str->di_height = cpu_to_be16(ip->i_height);
fs/gfs2/super.c
420
!(ip->i_diskflags & GFS2_DIF_EXHASH) ?
fs/gfs2/super.c
422
str->di_depth = cpu_to_be16(ip->i_depth);
fs/gfs2/super.c
423
str->di_entries = cpu_to_be32(ip->i_entries);
fs/gfs2/super.c
425
str->di_eattr = cpu_to_be64(ip->i_eattr);
fs/gfs2/super.c
441
struct gfs2_inode *ip = GFS2_I(inode);
fs/gfs2/super.c
443
struct address_space *metamapping = gfs2_glock2aspace(ip->i_gl);
fs/gfs2/super.c
446
bool flush_all = (wbc->sync_mode == WB_SYNC_ALL || gfs2_is_jdata(ip));
fs/gfs2/super.c
449
gfs2_log_flush(GFS2_SB(inode), ip->i_gl,
fs/gfs2/super.c
463
gfs2_ordered_del_inode(ip);
fs/gfs2/super.c
484
struct gfs2_inode *ip = GFS2_I(inode);
fs/gfs2/super.c
493
if (unlikely(!ip->i_gl))
fs/gfs2/super.c
498
if (!gfs2_glock_is_locked_by_me(ip->i_gl)) {
fs/gfs2/super.c
499
ret = gfs2_glock_nq_init(ip->i_gl, LM_ST_EXCLUSIVE, 0, &gh);
fs/gfs2/super.c
502
gfs2_dump_glock(NULL, ip->i_gl, true);
fs/gfs2/super.c
506
} else if (WARN_ON_ONCE(ip->i_gl->gl_state != LM_ST_EXCLUSIVE))
fs/gfs2/super.c
518
ret = gfs2_meta_inode_buffer(ip, &bh);
fs/gfs2/super.c
520
gfs2_trans_add_meta(ip->i_gl, bh);
fs/gfs2/super.c
521
gfs2_dinode_out(ip, bh->b_data);
fs/gfs2/trace_gfs2.h
444
TP_PROTO(const struct gfs2_inode *ip, const struct buffer_head *bh,
fs/gfs2/trace_gfs2.h
447
TP_ARGS(ip, bh, lblock, create, errno),
fs/gfs2/trace_gfs2.h
461
__entry->dev = glock_sbd(ip->i_gl)->sd_vfs->s_dev;
fs/gfs2/trace_gfs2.h
464
__entry->inum = ip->i_no_addr;
fs/gfs2/trace_gfs2.h
483
TP_PROTO(const struct gfs2_inode *ip, loff_t pos, ssize_t length,
fs/gfs2/trace_gfs2.h
486
TP_ARGS(ip, pos, length, flags),
fs/gfs2/trace_gfs2.h
497
__entry->dev = glock_sbd(ip->i_gl)->sd_vfs->s_dev;
fs/gfs2/trace_gfs2.h
498
__entry->inum = ip->i_no_addr;
fs/gfs2/trace_gfs2.h
513
TP_PROTO(const struct gfs2_inode *ip, struct iomap *iomap, int ret),
fs/gfs2/trace_gfs2.h
515
TP_ARGS(ip, iomap, ret),
fs/gfs2/trace_gfs2.h
529
__entry->dev = glock_sbd(ip->i_gl)->sd_vfs->s_dev;
fs/gfs2/trace_gfs2.h
530
__entry->inum = ip->i_no_addr;
fs/gfs2/trace_gfs2.h
534
(iomap->addr >> ip->i_inode.i_blkbits);
fs/gfs2/trace_gfs2.h
553
TP_PROTO(const struct gfs2_inode *ip, struct gfs2_rgrpd *rgd,
fs/gfs2/trace_gfs2.h
556
TP_ARGS(ip, rgd, block, len, block_state),
fs/gfs2/trace_gfs2.h
573
__entry->inum = ip->i_no_addr;
fs/gfs2/trans.c
42
unsigned long ip)
fs/gfs2/trans.c
55
tr->tr_ip = ip;
fs/gfs2/trans.h
28
static inline unsigned int gfs2_rg_blocks(const struct gfs2_inode *ip, unsigned requested)
fs/gfs2/trans.h
30
struct gfs2_rgrpd *rgd = ip->i_res.rs_rgd;
fs/gfs2/trans.h
39
unsigned long ip);
fs/gfs2/util.c
334
void gfs2_consist_inode_i(struct gfs2_inode *ip,
fs/gfs2/util.c
337
struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode);
fs/gfs2/util.c
343
(unsigned long long)ip->i_no_formal_ino,
fs/gfs2/util.c
344
(unsigned long long)ip->i_no_addr,
fs/gfs2/util.c
346
gfs2_dump_glock(NULL, ip->i_gl, 1);
fs/gfs2/util.c
58
struct gfs2_inode *ip;
fs/gfs2/util.c
60
ip = GFS2_I(jd->jd_inode);
fs/gfs2/util.c
61
error = gfs2_glock_nq_init(ip->i_gl, LM_ST_SHARED, LM_FLAG_RECOVER |
fs/gfs2/util.h
70
void gfs2_consist_inode_i(struct gfs2_inode *ip,
fs/gfs2/util.h
73
#define gfs2_consist_inode(ip) \
fs/gfs2/util.h
74
gfs2_consist_inode_i((ip), __func__, __FILE__, __LINE__)
fs/gfs2/xattr.c
100
gfs2_consist_inode(ip);
fs/gfs2/xattr.c
1005
gfs2_trans_add_meta(ip->i_gl, indbh);
fs/gfs2/xattr.c
1009
error = gfs2_alloc_blocks(ip, &blk, &n, 0);
fs/gfs2/xattr.c
1013
indbh = gfs2_meta_new(ip->i_gl, blk);
fs/gfs2/xattr.c
1014
gfs2_trans_add_meta(ip->i_gl, indbh);
fs/gfs2/xattr.c
1019
*eablk = cpu_to_be64(ip->i_eattr);
fs/gfs2/xattr.c
1020
ip->i_eattr = blk;
fs/gfs2/xattr.c
1021
ip->i_diskflags |= GFS2_DIF_EA_INDIRECT;
fs/gfs2/xattr.c
1022
gfs2_add_inode_blocks(&ip->i_inode, 1);
fs/gfs2/xattr.c
1027
error = ea_alloc_blk(ip, &newbh);
fs/gfs2/xattr.c
1032
error = ea_write(ip, GFS2_EA_BH2FIRST(newbh), er);
fs/gfs2/xattr.c
1038
ea_set_remove_stuffed(ip, private);
fs/gfs2/xattr.c
1045
static int ea_set_i(struct gfs2_inode *ip, int type, const char *name,
fs/gfs2/xattr.c
105
gfs2_consist_inode(ip);
fs/gfs2/xattr.c
1063
error = ea_foreach(ip, ea_set_simple, &es);
fs/gfs2/xattr.c
1069
if (!(ip->i_diskflags & GFS2_DIF_EA_INDIRECT))
fs/gfs2/xattr.c
1071
if (GFS2_EAREQ_SIZE_STUFFED(&er) > GFS2_SB(&ip->i_inode)->sd_jbsize)
fs/gfs2/xattr.c
1072
blks += DIV_ROUND_UP(er.er_data_len, GFS2_SB(&ip->i_inode)->sd_jbsize);
fs/gfs2/xattr.c
1074
return ea_alloc_skeleton(ip, &er, blks, ea_set_block, el);
fs/gfs2/xattr.c
1077
static int ea_set_remove_unstuffed(struct gfs2_inode *ip,
fs/gfs2/xattr.c
1082
gfs2_assert_withdraw(GFS2_SB(&ip->i_inode),
fs/gfs2/xattr.c
1086
return ea_remove_unstuffed(ip, el->el_bh, el->el_ea, el->el_prev, 0);
fs/gfs2/xattr.c
1089
static int ea_remove_stuffed(struct gfs2_inode *ip, struct gfs2_ea_location *el)
fs/gfs2/xattr.c
109
gfs2_consist_inode(ip);
fs/gfs2/xattr.c
1095
error = gfs2_trans_begin(GFS2_SB(&ip->i_inode), RES_DINODE + RES_EATTR, 0);
fs/gfs2/xattr.c
1099
gfs2_trans_add_meta(ip->i_gl, el->el_bh);
fs/gfs2/xattr.c
1113
inode_set_ctime_current(&ip->i_inode);
fs/gfs2/xattr.c
1114
__mark_inode_dirty(&ip->i_inode, I_DIRTY_DATASYNC);
fs/gfs2/xattr.c
1116
gfs2_trans_end(GFS2_SB(&ip->i_inode));
fs/gfs2/xattr.c
112
error = ea_call(ip, bh, ea, prev, data);
fs/gfs2/xattr.c
1134
static int gfs2_xattr_remove(struct gfs2_inode *ip, int type, const char *name)
fs/gfs2/xattr.c
1139
if (!ip->i_eattr)
fs/gfs2/xattr.c
1142
error = gfs2_ea_find(ip, type, name, &el);
fs/gfs2/xattr.c
1149
error = ea_remove_stuffed(ip, &el);
fs/gfs2/xattr.c
1151
error = ea_remove_unstuffed(ip, el.el_bh, el.el_ea, el.el_prev, 0);
fs/gfs2/xattr.c
1175
struct gfs2_inode *ip = GFS2_I(inode);
fs/gfs2/xattr.c
1187
error = gfs2_xattr_remove(ip, type, name);
fs/gfs2/xattr.c
119
gfs2_consist_inode(ip);
fs/gfs2/xattr.c
1196
if (!ip->i_eattr) {
fs/gfs2/xattr.c
1199
return ea_init(ip, type, name, value, size);
fs/gfs2/xattr.c
1202
error = gfs2_ea_find(ip, type, name, &el);
fs/gfs2/xattr.c
1207
if (ip->i_diskflags & GFS2_DIF_APPENDONLY) {
fs/gfs2/xattr.c
1215
error = ea_set_i(ip, type, name, value, size, &el);
fs/gfs2/xattr.c
1217
ea_set_remove_unstuffed(ip, &el);
fs/gfs2/xattr.c
1226
error = ea_set_i(ip, type, name, value, size, NULL);
fs/gfs2/xattr.c
1237
struct gfs2_inode *ip = GFS2_I(inode);
fs/gfs2/xattr.c
1241
ret = gfs2_qa_get(ip);
fs/gfs2/xattr.c
1247
if (!gfs2_glock_is_locked_by_me(ip->i_gl)) {
fs/gfs2/xattr.c
1248
ret = gfs2_glock_nq_init(ip->i_gl, LM_ST_EXCLUSIVE, 0, &gh);
fs/gfs2/xattr.c
1252
if (WARN_ON_ONCE(ip->i_gl->gl_state != LM_ST_EXCLUSIVE)) {
fs/gfs2/xattr.c
1262
gfs2_qa_put(ip);
fs/gfs2/xattr.c
1266
static int ea_dealloc_indirect(struct gfs2_inode *ip)
fs/gfs2/xattr.c
1268
struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode);
fs/gfs2/xattr.c
1286
error = gfs2_meta_read(ip->i_gl, ip->i_eattr, DIO_WAIT, 0, &indbh);
fs/gfs2/xattr.c
129
static int ea_foreach(struct gfs2_inode *ip, ea_call_t ea_call, void *data)
fs/gfs2/xattr.c
1309
gfs2_rlist_add(ip, &rlist, bstart);
fs/gfs2/xattr.c
1316
gfs2_rlist_add(ip, &rlist, bstart);
fs/gfs2/xattr.c
1336
gfs2_trans_add_meta(ip->i_gl, indbh);
fs/gfs2/xattr.c
135
error = gfs2_meta_read(ip->i_gl, ip->i_eattr, DIO_WAIT, 0, &bh);
fs/gfs2/xattr.c
1354
gfs2_free_meta(ip, rgd, bstart, blen);
fs/gfs2/xattr.c
1361
gfs2_add_inode_blocks(&ip->i_inode, -1);
fs/gfs2/xattr.c
1364
gfs2_free_meta(ip, rgd, bstart, blen);
fs/gfs2/xattr.c
1366
ip->i_diskflags &= ~GFS2_DIF_EA_INDIRECT;
fs/gfs2/xattr.c
1368
error = gfs2_meta_inode_buffer(ip, &dibh);
fs/gfs2/xattr.c
1370
gfs2_trans_add_meta(ip->i_gl, dibh);
fs/gfs2/xattr.c
1371
gfs2_dinode_out(ip, dibh->b_data);
fs/gfs2/xattr.c
1386
static int ea_dealloc_block(struct gfs2_inode *ip, bool initialized)
fs/gfs2/xattr.c
1388
struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode);
fs/gfs2/xattr.c
139
if (!(ip->i_diskflags & GFS2_DIF_EA_INDIRECT)) {
fs/gfs2/xattr.c
1398
rgd = gfs2_blk2rgrpd(sdp, ip->i_eattr, 1);
fs/gfs2/xattr.c
140
error = ea_foreach_i(ip, bh, ea_call, data);
fs/gfs2/xattr.c
1400
gfs2_consist_inode(ip);
fs/gfs2/xattr.c
1414
gfs2_free_meta(ip, rgd, ip->i_eattr, 1);
fs/gfs2/xattr.c
1416
ip->i_eattr = 0;
fs/gfs2/xattr.c
1417
gfs2_add_inode_blocks(&ip->i_inode, -1);
fs/gfs2/xattr.c
1420
error = gfs2_meta_inode_buffer(ip, &dibh);
fs/gfs2/xattr.c
1422
gfs2_trans_add_meta(ip->i_gl, dibh);
fs/gfs2/xattr.c
1423
gfs2_dinode_out(ip, dibh->b_data);
fs/gfs2/xattr.c
144
if (gfs2_metatype_check(GFS2_SB(&ip->i_inode), bh, GFS2_METATYPE_IN)) {
fs/gfs2/xattr.c
1443
int gfs2_ea_dealloc(struct gfs2_inode *ip, bool initialized)
fs/gfs2/xattr.c
1447
error = gfs2_rindex_update(GFS2_SB(&ip->i_inode));
fs/gfs2/xattr.c
1451
error = gfs2_quota_hold(ip, NO_UID_QUOTA_CHANGE, NO_GID_QUOTA_CHANGE);
fs/gfs2/xattr.c
1456
error = ea_foreach(ip, ea_dealloc_unstuffed, NULL);
fs/gfs2/xattr.c
1460
if (ip->i_diskflags & GFS2_DIF_EA_INDIRECT) {
fs/gfs2/xattr.c
1461
error = ea_dealloc_indirect(ip);
fs/gfs2/xattr.c
1467
error = ea_dealloc_block(ip, initialized);
fs/gfs2/xattr.c
1470
gfs2_quota_unhold(ip);
fs/gfs2/xattr.c
150
end = eablk + GFS2_SB(&ip->i_inode)->sd_inptrs;
fs/gfs2/xattr.c
159
error = gfs2_meta_read(ip->i_gl, bn, DIO_WAIT, 0, &eabh);
fs/gfs2/xattr.c
162
error = ea_foreach_i(ip, eabh, ea_call, data);
fs/gfs2/xattr.c
179
static int ea_find_i(struct gfs2_inode *ip, struct buffer_head *bh,
fs/gfs2/xattr.c
203
static int gfs2_ea_find(struct gfs2_inode *ip, int type, const char *name,
fs/gfs2/xattr.c
216
error = ea_foreach(ip, ea_find_i, &ef);
fs/gfs2/xattr.c
233
static int ea_dealloc_unstuffed(struct gfs2_inode *ip, struct buffer_head *bh,
fs/gfs2/xattr.c
238
struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode);
fs/gfs2/xattr.c
268
gfs2_consist_inode(ip);
fs/gfs2/xattr.c
282
gfs2_trans_add_meta(ip->i_gl, bh);
fs/gfs2/xattr.c
294
gfs2_free_meta(ip, rgd, bstart, blen);
fs/gfs2/xattr.c
300
gfs2_add_inode_blocks(&ip->i_inode, -1);
fs/gfs2/xattr.c
303
gfs2_free_meta(ip, rgd, bstart, blen);
fs/gfs2/xattr.c
318
inode_set_ctime_current(&ip->i_inode);
fs/gfs2/xattr.c
319
__mark_inode_dirty(&ip->i_inode, I_DIRTY_DATASYNC);
fs/gfs2/xattr.c
328
static int ea_remove_unstuffed(struct gfs2_inode *ip, struct buffer_head *bh,
fs/gfs2/xattr.c
334
error = gfs2_rindex_update(GFS2_SB(&ip->i_inode));
fs/gfs2/xattr.c
338
error = gfs2_quota_hold(ip, NO_UID_QUOTA_CHANGE, NO_GID_QUOTA_CHANGE);
fs/gfs2/xattr.c
342
error = ea_dealloc_unstuffed(ip, bh, ea, prev, (leave) ? &error : NULL);
fs/gfs2/xattr.c
344
gfs2_quota_unhold(ip);
fs/gfs2/xattr.c
354
static int ea_list_i(struct gfs2_inode *ip, struct buffer_head *bh,
fs/gfs2/xattr.c
358
struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode);
fs/gfs2/xattr.c
418
struct gfs2_inode *ip = GFS2_I(d_inode(dentry));
fs/gfs2/xattr.c
429
error = gfs2_glock_nq_init(ip->i_gl, LM_ST_SHARED, LM_FLAG_ANY, &i_gh);
fs/gfs2/xattr.c
433
if (ip->i_eattr) {
fs/gfs2/xattr.c
436
error = ea_foreach(ip, ea_list_i, &ei);
fs/gfs2/xattr.c
457
static int gfs2_iter_unstuffed(struct gfs2_inode *ip, struct gfs2_ea_header *ea,
fs/gfs2/xattr.c
460
struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode);
fs/gfs2/xattr.c
475
error = gfs2_meta_read(ip->i_gl, be64_to_cpu(*dataptrs), 0, 0,
fs/gfs2/xattr.c
508
gfs2_trans_add_meta(ip->i_gl, bh[x]);
fs/gfs2/xattr.c
522
static int gfs2_ea_get_copy(struct gfs2_inode *ip, struct gfs2_ea_location *el,
fs/gfs2/xattr.c
534
ret = gfs2_iter_unstuffed(ip, el->el_ea, NULL, data);
fs/gfs2/xattr.c
540
int gfs2_xattr_acl_get(struct gfs2_inode *ip, const char *name, char **ppdata)
fs/gfs2/xattr.c
547
error = gfs2_ea_find(ip, GFS2_EATYPE_SYS, name, &el);
fs/gfs2/xattr.c
561
error = gfs2_ea_get_copy(ip, &el, data, len);
fs/gfs2/xattr.c
584
struct gfs2_inode *ip = GFS2_I(inode);
fs/gfs2/xattr.c
588
if (!ip->i_eattr)
fs/gfs2/xattr.c
593
error = gfs2_ea_find(ip, type, name, &el);
fs/gfs2/xattr.c
599
error = gfs2_ea_get_copy(ip, &el, buffer, size);
fs/gfs2/xattr.c
611
struct gfs2_inode *ip = GFS2_I(inode);
fs/gfs2/xattr.c
617
if (!gfs2_glock_is_locked_by_me(ip->i_gl)) {
fs/gfs2/xattr.c
618
ret = gfs2_glock_nq_init(ip->i_gl, LM_ST_SHARED, LM_FLAG_ANY, &gh);
fs/gfs2/xattr.c
638
static int ea_alloc_blk(struct gfs2_inode *ip, struct buffer_head **bhp)
fs/gfs2/xattr.c
640
struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode);
fs/gfs2/xattr.c
646
error = gfs2_alloc_blocks(ip, &block, &n, 0);
fs/gfs2/xattr.c
650
*bhp = gfs2_meta_new(ip->i_gl, block);
fs/gfs2/xattr.c
651
gfs2_trans_add_meta(ip->i_gl, *bhp);
fs/gfs2/xattr.c
661
gfs2_add_inode_blocks(&ip->i_inode, 1);
fs/gfs2/xattr.c
678
static int ea_write(struct gfs2_inode *ip, struct gfs2_ea_header *ea,
fs/gfs2/xattr.c
681
struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode);
fs/gfs2/xattr.c
708
error = gfs2_alloc_blocks(ip, &block, &n, 0);
fs/gfs2/xattr.c
712
bh = gfs2_meta_new(ip->i_gl, block);
fs/gfs2/xattr.c
713
gfs2_trans_add_meta(ip->i_gl, bh);
fs/gfs2/xattr.c
716
gfs2_add_inode_blocks(&ip->i_inode, 1);
fs/gfs2/xattr.c
738
typedef int (*ea_skeleton_call_t) (struct gfs2_inode *ip,
fs/gfs2/xattr.c
741
static int ea_alloc_skeleton(struct gfs2_inode *ip, struct gfs2_ea_request *er,
fs/gfs2/xattr.c
748
error = gfs2_rindex_update(GFS2_SB(&ip->i_inode));
fs/gfs2/xattr.c
752
error = gfs2_quota_lock_check(ip, &ap);
fs/gfs2/xattr.c
756
error = gfs2_inplace_reserve(ip, &ap);
fs/gfs2/xattr.c
760
error = gfs2_trans_begin(GFS2_SB(&ip->i_inode),
fs/gfs2/xattr.c
761
blks + gfs2_rg_blocks(ip, blks) +
fs/gfs2/xattr.c
766
error = skeleton_call(ip, er, private);
fs/gfs2/xattr.c
770
inode_set_ctime_current(&ip->i_inode);
fs/gfs2/xattr.c
771
__mark_inode_dirty(&ip->i_inode, I_DIRTY_DATASYNC);
fs/gfs2/xattr.c
774
gfs2_trans_end(GFS2_SB(&ip->i_inode));
fs/gfs2/xattr.c
776
gfs2_inplace_release(ip);
fs/gfs2/xattr.c
778
gfs2_quota_unlock(ip);
fs/gfs2/xattr.c
782
static int ea_init_i(struct gfs2_inode *ip, struct gfs2_ea_request *er,
fs/gfs2/xattr.c
788
error = ea_alloc_blk(ip, &bh);
fs/gfs2/xattr.c
792
ip->i_eattr = bh->b_blocknr;
fs/gfs2/xattr.c
793
error = ea_write(ip, GFS2_EA_BH2FIRST(bh), er);
fs/gfs2/xattr.c
805
static int ea_init(struct gfs2_inode *ip, int type, const char *name,
fs/gfs2/xattr.c
809
unsigned int jbsize = GFS2_SB(&ip->i_inode)->sd_jbsize;
fs/gfs2/xattr.c
821
return ea_alloc_skeleton(ip, &er, blks, ea_init_i, NULL);
fs/gfs2/xattr.c
84
typedef int (*ea_call_t) (struct gfs2_inode *ip, struct buffer_head *bh,
fs/gfs2/xattr.c
841
static void ea_set_remove_stuffed(struct gfs2_inode *ip,
fs/gfs2/xattr.c
848
gfs2_trans_add_meta(ip->i_gl, el->el_bh);
fs/gfs2/xattr.c
855
gfs2_assert_withdraw(GFS2_SB(&ip->i_inode), GFS2_EA2NEXT(prev) == ea);
fs/gfs2/xattr.c
875
static int ea_set_simple_noalloc(struct gfs2_inode *ip, struct buffer_head *bh,
fs/gfs2/xattr.c
88
static int ea_foreach_i(struct gfs2_inode *ip, struct buffer_head *bh,
fs/gfs2/xattr.c
881
error = gfs2_trans_begin(GFS2_SB(&ip->i_inode), RES_DINODE + 2 * RES_EATTR, 0);
fs/gfs2/xattr.c
885
gfs2_trans_add_meta(ip->i_gl, bh);
fs/gfs2/xattr.c
890
ea_write(ip, ea, er);
fs/gfs2/xattr.c
893
ea_set_remove_stuffed(ip, es->es_el);
fs/gfs2/xattr.c
895
inode_set_ctime_current(&ip->i_inode);
fs/gfs2/xattr.c
896
__mark_inode_dirty(&ip->i_inode, I_DIRTY_DATASYNC);
fs/gfs2/xattr.c
898
gfs2_trans_end(GFS2_SB(&ip->i_inode));
fs/gfs2/xattr.c
902
static int ea_set_simple_alloc(struct gfs2_inode *ip,
fs/gfs2/xattr.c
909
gfs2_trans_add_meta(ip->i_gl, es->es_bh);
fs/gfs2/xattr.c
91
struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode);
fs/gfs2/xattr.c
914
error = ea_write(ip, ea, er);
fs/gfs2/xattr.c
919
ea_set_remove_stuffed(ip, es->es_el);
fs/gfs2/xattr.c
924
static int ea_set_simple(struct gfs2_inode *ip, struct buffer_head *bh,
fs/gfs2/xattr.c
933
stuffed = ea_calc_size(GFS2_SB(&ip->i_inode), es->es_er->er_name_len,
fs/gfs2/xattr.c
940
error = ea_remove_unstuffed(ip, bh, ea, prev, 1);
fs/gfs2/xattr.c
95
if (gfs2_metatype_check(GFS2_SB(&ip->i_inode), bh, GFS2_METATYPE_EA))
fs/gfs2/xattr.c
951
error = ea_set_simple_noalloc(ip, bh, ea, es);
fs/gfs2/xattr.c
960
GFS2_SB(&ip->i_inode)->sd_jbsize);
fs/gfs2/xattr.c
962
error = ea_alloc_skeleton(ip, es->es_er, blks,
fs/gfs2/xattr.c
971
static int ea_set_block(struct gfs2_inode *ip, struct gfs2_ea_request *er,
fs/gfs2/xattr.c
974
struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode);
fs/gfs2/xattr.c
980
if (ip->i_diskflags & GFS2_DIF_EA_INDIRECT) {
fs/gfs2/xattr.c
983
error = gfs2_meta_read(ip->i_gl, ip->i_eattr, DIO_WAIT, 0,
fs/gfs2/xattr.h
57
int gfs2_ea_dealloc(struct gfs2_inode *ip, bool initialized);
fs/gfs2/xattr.h
61
int gfs2_xattr_acl_get(struct gfs2_inode *ip, const char *name, char **data);
fs/hfsplus/unicode.c
154
const hfsplus_unichr *ip;
fs/hfsplus/unicode.c
162
ip = ustr->unicode;
fs/hfsplus/unicode.c
176
c0 = be16_to_cpu(*ip++);
fs/hfsplus/unicode.c
191
c1 = be16_to_cpu(*ip) - Hangul_VBase;
fs/hfsplus/unicode.c
197
ip++;
fs/hfsplus/unicode.c
201
c1 = be16_to_cpu(*ip) - Hangul_TBase;
fs/hfsplus/unicode.c
204
ip++;
fs/hfsplus/unicode.c
214
c1 = be16_to_cpu(*ip);
fs/hfsplus/unicode.c
238
ip++;
fs/hfsplus/unicode.c
246
be16_to_cpu(ip[i]));
fs/hfsplus/unicode.c
254
ip += i;
fs/isofs/joliet.c
20
__be16 *ip, ch;
fs/isofs/joliet.c
23
ip = uni;
fs/isofs/joliet.c
26
while ((ch = get_unaligned(ip)) && len) {
fs/isofs/joliet.c
33
ip++;
fs/jfs/inode.c
205
int jfs_get_block(struct inode *ip, sector_t lblock,
fs/jfs/inode.c
213
s32 xlen = bh_result->b_size >> ip->i_blkbits;
fs/jfs/inode.c
219
IWRITE_LOCK(ip, RDWRLOCK_NORMAL);
fs/jfs/inode.c
221
IREAD_LOCK(ip, RDWRLOCK_NORMAL);
fs/jfs/inode.c
223
if (((lblock64 << ip->i_sb->s_blocksize_bits) < ip->i_size) &&
fs/jfs/inode.c
224
(!xtLookup(ip, lblock64, xlen, &xflag, &xaddr, &xlen, 0)) &&
fs/jfs/inode.c
236
rc = extRecord(ip, &xad);
fs/jfs/inode.c
242
map_bh(bh_result, ip->i_sb, xaddr);
fs/jfs/inode.c
243
bh_result->b_size = xlen << ip->i_blkbits;
fs/jfs/inode.c
252
if ((rc = extHint(ip, lblock64 << ip->i_sb->s_blocksize_bits, &xad)))
fs/jfs/inode.c
254
rc = extAlloc(ip, xlen, lblock64, &xad, false);
fs/jfs/inode.c
259
map_bh(bh_result, ip->i_sb, addressXAD(&xad));
fs/jfs/inode.c
260
bh_result->b_size = lengthXAD(&xad) << ip->i_blkbits;
fs/jfs/inode.c
267
IWRITE_UNLOCK(ip);
fs/jfs/inode.c
269
IREAD_UNLOCK(ip);
fs/jfs/inode.c
373
void jfs_truncate_nolock(struct inode *ip, loff_t length)
fs/jfs/inode.c
380
if (test_cflag(COMMIT_Nolink, ip) || isReadOnly(ip)) {
fs/jfs/inode.c
381
xtTruncate(0, ip, length, COMMIT_WMAP);
fs/jfs/inode.c
386
tid = txBegin(ip->i_sb, 0);
fs/jfs/inode.c
394
mutex_lock(&JFS_IP(ip)->commit_mutex);
fs/jfs/inode.c
396
newsize = xtTruncate(tid, ip, length,
fs/jfs/inode.c
400
mutex_unlock(&JFS_IP(ip)->commit_mutex);
fs/jfs/inode.c
404
inode_set_mtime_to_ts(ip, inode_set_ctime_current(ip));
fs/jfs/inode.c
405
mark_inode_dirty(ip);
fs/jfs/inode.c
407
txCommit(tid, 1, &ip, 0);
fs/jfs/inode.c
409
mutex_unlock(&JFS_IP(ip)->commit_mutex);
fs/jfs/inode.c
413
void jfs_truncate(struct inode *ip)
fs/jfs/inode.c
415
jfs_info("jfs_truncate: size = 0x%lx", (ulong) ip->i_size);
fs/jfs/inode.c
417
block_truncate_page(ip->i_mapping, ip->i_size, jfs_get_block);
fs/jfs/inode.c
419
IWRITE_LOCK(ip, RDWRLOCK_NORMAL);
fs/jfs/inode.c
420
jfs_truncate_nolock(ip, ip->i_size);
fs/jfs/inode.c
421
IWRITE_UNLOCK(ip);
fs/jfs/jfs_discard.c
103
agno = BLKTOAG(start, JFS_SBI(ip->i_sb));
fs/jfs/jfs_discard.c
104
agno_end = BLKTOAG(end, JFS_SBI(ip->i_sb));
fs/jfs/jfs_discard.c
106
trimmed += dbDiscardAG(ip, agno, minlen);
fs/jfs/jfs_discard.c
32
void jfs_issue_discard(struct inode *ip, u64 blkno, u64 nblocks)
fs/jfs/jfs_discard.c
34
struct super_block *sb = ip->i_sb;
fs/jfs/jfs_discard.c
65
int jfs_ioc_trim(struct inode *ip, struct fstrim_range *range)
fs/jfs/jfs_discard.c
67
struct inode *ipbmap = JFS_SBI(ip->i_sb)->ipbmap;
fs/jfs/jfs_discard.c
87
bmp = JFS_SBI(ip->i_sb)->bmap;
fs/jfs/jfs_discard.h
10
extern void jfs_issue_discard(struct inode *ip, u64 blkno, u64 nblocks);
fs/jfs/jfs_discard.h
11
extern int jfs_ioc_trim(struct inode *ip, struct fstrim_range *range);
fs/jfs/jfs_dmap.c
1004
jfs_error(ip->i_sb, "the block is outside the filesystem\n");
fs/jfs/jfs_dmap.c
1571
s64 dbDiscardAG(struct inode *ip, int agno, s64 minlen)
fs/jfs/jfs_dmap.c
1573
struct inode *ipbmap = JFS_SBI(ip->i_sb)->ipbmap;
fs/jfs/jfs_dmap.c
1574
struct bmap *bmp = JFS_SBI(ip->i_sb)->bmap;
fs/jfs/jfs_dmap.c
1644
jfs_issue_discard(ip, tt->blkno, tt->nblocks);
fs/jfs/jfs_dmap.c
1645
dbFree(ip, tt->blkno, tt->nblocks);
fs/jfs/jfs_dmap.c
3191
int dbAllocBottomUp(struct inode *ip, s64 blkno, s64 nblocks)
fs/jfs/jfs_dmap.c
3197
struct inode *ipbmap = JFS_SBI(ip->i_sb)->ipbmap;
fs/jfs/jfs_dmap.c
3198
struct bmap *bmp = JFS_SBI(ip->i_sb)->bmap;
fs/jfs/jfs_dmap.c
344
int dbFree(struct inode *ip, s64 blkno, s64 nblocks)
fs/jfs/jfs_dmap.c
350
struct inode *ipbmap = JFS_SBI(ip->i_sb)->ipbmap;
fs/jfs/jfs_dmap.c
351
struct bmap *bmp = JFS_SBI(ip->i_sb)->bmap;
fs/jfs/jfs_dmap.c
362
jfs_error(ip->i_sb, "block to be freed is outside the map\n");
fs/jfs/jfs_dmap.c
399
jfs_error(ip->i_sb, "error in block map\n");
fs/jfs/jfs_dmap.c
714
int dbAlloc(struct inode *ip, s64 hint, s64 nblocks, s64 * results)
fs/jfs/jfs_dmap.c
717
struct inode *ipbmap = JFS_SBI(ip->i_sb)->ipbmap;
fs/jfs/jfs_dmap.c
735
bmp = JFS_SBI(ip->i_sb)->bmap;
fs/jfs/jfs_dmap.c
741
jfs_error(ip->i_sb, "the hint is outside the map\n");
fs/jfs/jfs_dmap.c
818
((writers == 1) && (JFS_IP(ip)->active_ag != agno))) {
fs/jfs/jfs_dmap.c
87
static int dbExtend(struct inode *ip, s64 blkno, s64 nblocks, s64 addnblocks);
fs/jfs/jfs_dmap.c
924
dbReAlloc(struct inode *ip,
fs/jfs/jfs_dmap.c
931
if ((rc = dbExtend(ip, blkno, nblocks, addnblocks)) == 0) {
fs/jfs/jfs_dmap.c
945
(ip, blkno + nblocks - 1, addnblocks + nblocks, results));
fs/jfs/jfs_dmap.c
972
static int dbExtend(struct inode *ip, s64 blkno, s64 nblocks, s64 addnblocks)
fs/jfs/jfs_dmap.c
974
struct jfs_sb_info *sbi = JFS_SBI(ip->i_sb);
fs/jfs/jfs_dmap.h
126
#define AGTOBLK(a,ip) \
fs/jfs/jfs_dmap.h
127
((s64)(a) << (JFS_SBI((ip)->i_sb)->bmap->db_agl2size))
fs/jfs/jfs_dmap.h
297
extern int dbAllocBottomUp(struct inode *ip, s64 blkno, s64 nblocks);
fs/jfs/jfs_dmap.h
301
extern s64 dbDiscardAG(struct inode *ip, int agno, s64 minlen);
fs/jfs/jfs_dtree.c
1002
if (!DO_INDEX(ip))
fs/jfs/jfs_dtree.c
1003
ip->i_size = xlen << sbi->l2bsize;
fs/jfs/jfs_dtree.c
1027
rc = dquot_alloc_block(ip, n);
fs/jfs/jfs_dtree.c
1042
if ((rc = dtExtendPage(tid, ip, split, btstack))) {
fs/jfs/jfs_dtree.c
1047
dbFree(ip, nxaddr, (s64) xlen);
fs/jfs/jfs_dtree.c
1052
dbFree(ip, xaddr, (s64) n);
fs/jfs/jfs_dtree.c
1054
} else if (!DO_INDEX(ip))
fs/jfs/jfs_dtree.c
1055
ip->i_size = lengthPXD(pxd) << sbi->l2bsize;
fs/jfs/jfs_dtree.c
1078
if ((rc = dbAlloc(ip, 0, (s64) xlen, &xaddr)) == 0) {
fs/jfs/jfs_dtree.c
1092
if ((rc = dtSplitPage(tid, ip, split, &rmp, &rp, &rpxd))) {
fs/jfs/jfs_dtree.c
1099
if (!DO_INDEX(ip))
fs/jfs/jfs_dtree.c
1100
ip->i_size += PSIZE;
fs/jfs/jfs_dtree.c
1134
DT_GETPAGE(ip, parent->bn, smp, PSIZE, sp, rc);
fs/jfs/jfs_dtree.c
1244
dtSplitRoot(tid, ip, split, &rmp) :
fs/jfs/jfs_dtree.c
1245
dtSplitPage(tid, ip, split, &rmp, &rp, &rpxd);
fs/jfs/jfs_dtree.c
1257
BT_MARK_DIRTY(smp, ip);
fs/jfs/jfs_dtree.c
1261
tlck = txLock(tid, ip, smp, tlckDTREE | tlckENTRY);
fs/jfs/jfs_dtree.c
1300
dbFree(ip, addressPXD(pxd), (s64) lengthPXD(pxd));
fs/jfs/jfs_dtree.c
1307
dquot_free_block(ip, quota_allocation);
fs/jfs/jfs_dtree.c
1326
static int dtSplitPage(tid_t tid, struct inode *ip, struct dtsplit * split,
fs/jfs/jfs_dtree.c
1354
sp = DT_PAGE(ip, smp);
fs/jfs/jfs_dtree.c
1363
rmp = get_metapage(ip, rbn, PSIZE, 1);
fs/jfs/jfs_dtree.c
1368
rc = dquot_alloc_block(ip, lengthPXD(pxd));
fs/jfs/jfs_dtree.c
1374
jfs_info("dtSplitPage: ip:0x%p smp:0x%p rmp:0x%p", ip, smp, rmp);
fs/jfs/jfs_dtree.c
1376
BT_MARK_DIRTY(rmp, ip);
fs/jfs/jfs_dtree.c
1380
tlck = txLock(tid, ip, rmp, tlckDTREE | tlckNEW);
fs/jfs/jfs_dtree.c
1387
BT_MARK_DIRTY(smp, ip);
fs/jfs/jfs_dtree.c
1393
tlck = txLock(tid, ip, smp, tlckDTREE | tlckENTRY);
fs/jfs/jfs_dtree.c
140
static int dtSplitUp(tid_t tid, struct inode *ip,
fs/jfs/jfs_dtree.c
143
static int dtSplitPage(tid_t tid, struct inode *ip, struct dtsplit * split,
fs/jfs/jfs_dtree.c
146
static int dtExtendPage(tid_t tid, struct inode *ip,
fs/jfs/jfs_dtree.c
1471
DT_GETPAGE(ip, nextbn, mp, PSIZE, p, rc);
fs/jfs/jfs_dtree.c
1477
BT_MARK_DIRTY(mp, ip);
fs/jfs/jfs_dtree.c
1481
tlck = txLock(tid, ip, mp, tlckDTREE | tlckRELINK);
fs/jfs/jfs_dtree.c
1483
tlck, ip, mp);
fs/jfs/jfs_dtree.c
149
static int dtSplitRoot(tid_t tid, struct inode *ip,
fs/jfs/jfs_dtree.c
152
static int dtDeleteUp(tid_t tid, struct inode *ip, struct metapage * fmp,
fs/jfs/jfs_dtree.c
1521
if (DO_INDEX(ip))
fs/jfs/jfs_dtree.c
155
static int dtRelink(tid_t tid, struct inode *ip, dtpage_t * p);
fs/jfs/jfs_dtree.c
1561
dtMoveEntry(sp, nxt, rp, &sdtlck, &rdtlck, DO_INDEX(ip));
fs/jfs/jfs_dtree.c
157
static int dtReadFirst(struct inode *ip, struct btstack * btstack);
fs/jfs/jfs_dtree.c
1577
if ((rp->header.flag & BT_LEAF) && DO_INDEX(ip)) {
fs/jfs/jfs_dtree.c
1584
modify_index(tid, ip, le32_to_cpu(ldtentry->index),
fs/jfs/jfs_dtree.c
159
static int dtReadNext(struct inode *ip,
fs/jfs/jfs_dtree.c
1639
struct inode *ip, struct dtsplit * split, struct btstack * btstack)
fs/jfs/jfs_dtree.c
1641
struct super_block *sb = ip->i_sb;
fs/jfs/jfs_dtree.c
1665
sp = DT_PAGE(ip, smp);
fs/jfs/jfs_dtree.c
1669
DT_GETPAGE(ip, parent->bn, pmp, PSIZE, pp, rc);
fs/jfs/jfs_dtree.c
1692
tlck = txMaplock(tid, ip, tlckDTREE | tlckRELOCATE);
fs/jfs/jfs_dtree.c
1701
if (DO_INDEX(ip)) {
fs/jfs/jfs_dtree.c
1709
modify_index(tid, ip,
fs/jfs/jfs_dtree.c
1723
jfs_info("dtExtendPage: ip:0x%p smp:0x%p sp:0x%p", ip, smp, sp);
fs/jfs/jfs_dtree.c
1725
BT_MARK_DIRTY(smp, ip);
fs/jfs/jfs_dtree.c
1729
tlck = txLock(tid, ip, smp, tlckDTREE | type);
fs/jfs/jfs_dtree.c
1819
BT_MARK_DIRTY(pmp, ip);
fs/jfs/jfs_dtree.c
1835
tlck = txLock(tid, ip, pmp, tlckDTREE | tlckENTRY);
fs/jfs/jfs_dtree.c
1872
struct inode *ip, struct dtsplit * split, struct metapage ** rmpp)
fs/jfs/jfs_dtree.c
1874
struct super_block *sb = ip->i_sb;
fs/jfs/jfs_dtree.c
1896
sp = &JFS_IP(ip)->i_dtroot;
fs/jfs/jfs_dtree.c
1910
rmp = get_metapage(ip, rbn, xsize, 1);
fs/jfs/jfs_dtree.c
1917
rc = dquot_alloc_block(ip, lengthPXD(pxd));
fs/jfs/jfs_dtree.c
1923
BT_MARK_DIRTY(rmp, ip);
fs/jfs/jfs_dtree.c
1927
tlck = txLock(tid, ip, rmp, tlckDTREE | tlckNEW);
fs/jfs/jfs_dtree.c
1991
if ((rp->header.flag & BT_LEAF) && DO_INDEX(ip)) {
fs/jfs/jfs_dtree.c
1999
modify_index(tid, ip, le32_to_cpu(ldtentry->index),
fs/jfs/jfs_dtree.c
2020
BT_MARK_DIRTY(smp, ip);
fs/jfs/jfs_dtree.c
2024
tlck = txLock(tid, ip, smp, tlckDTREE | tlckNEW | tlckBTROOT);
fs/jfs/jfs_dtree.c
2079
struct inode *ip, struct component_name * key, ino_t * ino, int flag)
fs/jfs/jfs_dtree.c
2102
if ((rc = dtSearch(ip, key, ino, &btstack, flag)))
fs/jfs/jfs_dtree.c
2106
DT_GETSEARCH(ip, btstack.top, bn, mp, p, index);
fs/jfs/jfs_dtree.c
2113
if (DO_INDEX(ip)) {
fs/jfs/jfs_dtree.c
2126
DT_GETPAGE(ip, le64_to_cpu(p->header.next),
fs/jfs/jfs_dtree.c
2145
free_index(tid, ip, table_index, next_index);
fs/jfs/jfs_dtree.c
2152
rc = dtDeleteUp(tid, ip, mp, p, &btstack);
fs/jfs/jfs_dtree.c
2160
BT_MARK_DIRTY(mp, ip);
fs/jfs/jfs_dtree.c
2164
tlck = txLock(tid, ip, mp, tlckDTREE | tlckENTRY);
fs/jfs/jfs_dtree.c
2200
if (DO_INDEX(ip) && index < p->header.nextindex) {
fs/jfs/jfs_dtree.c
2208
modify_index(tid, ip,
fs/jfs/jfs_dtree.c
2233
static int dtDeleteUp(tid_t tid, struct inode *ip,
fs/jfs/jfs_dtree.c
2257
dtInitRoot(tid, ip, PARENT(ip));
fs/jfs/jfs_dtree.c
2274
tlck = txMaplock(tid, ip, tlckDTREE | tlckFREE);
fs/jfs/jfs_dtree.c
2281
if ((rc = dtRelink(tid, ip, fp))) {
fs/jfs/jfs_dtree.c
2289
dquot_free_block(ip, xlen);
fs/jfs/jfs_dtree.c
2304
DT_GETPAGE(ip, parent->bn, mp, PSIZE, p, rc);
fs/jfs/jfs_dtree.c
2333
dtInitRoot(tid, ip, PARENT(ip));
fs/jfs/jfs_dtree.c
2349
txMaplock(tid, ip,
fs/jfs/jfs_dtree.c
2357
if ((rc = dtRelink(tid, ip, p))) {
fs/jfs/jfs_dtree.c
2365
dquot_free_block(ip, xlen);
fs/jfs/jfs_dtree.c
237
static struct dir_table_slot *find_index(struct inode *ip, u32 index,
fs/jfs/jfs_dtree.c
2380
BT_MARK_DIRTY(mp, ip);
fs/jfs/jfs_dtree.c
2386
tlck = txLock(tid, ip, mp, tlckDTREE | tlckENTRY);
fs/jfs/jfs_dtree.c
240
struct jfs_inode_info *jfs_ip = JFS_IP(ip);
fs/jfs/jfs_dtree.c
2428
if (!DO_INDEX(ip))
fs/jfs/jfs_dtree.c
2429
ip->i_size -= PSIZE;
fs/jfs/jfs_dtree.c
2445
static int dtRelink(tid_t tid, struct inode *ip, dtpage_t * p)
fs/jfs/jfs_dtree.c
2459
DT_GETPAGE(ip, nextbn, mp, PSIZE, p, rc);
fs/jfs/jfs_dtree.c
2463
BT_MARK_DIRTY(mp, ip);
fs/jfs/jfs_dtree.c
2469
tlck = txLock(tid, ip, mp, tlckDTREE | tlckRELINK);
fs/jfs/jfs_dtree.c
2471
tlck, ip, mp);
fs/jfs/jfs_dtree.c
2488
DT_GETPAGE(ip, prevbn, mp, PSIZE, p, rc);
fs/jfs/jfs_dtree.c
2492
BT_MARK_DIRTY(mp, ip);
fs/jfs/jfs_dtree.c
2498
tlck = txLock(tid, ip, mp, tlckDTREE | tlckRELINK);
fs/jfs/jfs_dtree.c
2500
tlck, ip, mp);
fs/jfs/jfs_dtree.c
2524
void dtInitRoot(tid_t tid, struct inode *ip, u32 idotdot)
fs/jfs/jfs_dtree.c
2526
struct jfs_inode_info *jfs_ip = JFS_IP(ip);
fs/jfs/jfs_dtree.c
2539
if (DO_INDEX(ip)) {
fs/jfs/jfs_dtree.c
2540
if (!jfs_dirtable_inline(ip)) {
fs/jfs/jfs_dtree.c
2558
xtTruncate(tid, ip, 0, COMMIT_PWMAP);
fs/jfs/jfs_dtree.c
2559
set_cflag(COMMIT_Stale, ip);
fs/jfs/jfs_dtree.c
2563
ip->i_size = 1;
fs/jfs/jfs_dtree.c
2567
ip->i_size = IDATASIZE;
fs/jfs/jfs_dtree.c
2574
tlck = txLock(tid, ip, (struct metapage *) & jfs_ip->bxflag,
fs/jfs/jfs_dtree.c
260
if (jfs_dirtable_inline(ip)) {
fs/jfs/jfs_dtree.c
270
JFS_SBI(ip->i_sb)->l2nbperpage;
fs/jfs/jfs_dtree.c
2716
struct inode *ip = file_inode(file);
fs/jfs/jfs_dtree.c
2717
struct nls_table *codepage = JFS_SBI(ip->i_sb)->nls_tab;
fs/jfs/jfs_dtree.c
2748
if (DO_INDEX(ip)) {
fs/jfs/jfs_dtree.c
2770
if (dtEmpty(ip) ||
fs/jfs/jfs_dtree.c
2771
(dir_index >= JFS_IP(ip)->next_index)) {
fs/jfs/jfs_dtree.c
2777
rc = read_index(ip, dir_index, &dirtab_slot);
fs/jfs/jfs_dtree.c
278
*mp = read_index_page(ip, blkno);
fs/jfs/jfs_dtree.c
2783
if (loop_count++ > JFS_IP(ip)->next_index) {
fs/jfs/jfs_dtree.c
2797
DT_GETPAGE(ip, bn, mp, PSIZE, p, rc);
fs/jfs/jfs_dtree.c
2814
if (!dir_emit(ctx, ".", 1, ip->i_ino, DT_DIR))
fs/jfs/jfs_dtree.c
2821
if (!dir_emit(ctx, "..", 2, PARENT(ip), DT_DIR))
fs/jfs/jfs_dtree.c
2827
if (dtEmpty(ip)) {
fs/jfs/jfs_dtree.c
2832
if ((rc = dtReadFirst(ip, &btstack)))
fs/jfs/jfs_dtree.c
2835
DT_GETSEARCH(ip, btstack.top, bn, mp, p, index);
fs/jfs/jfs_dtree.c
2850
if (!dir_emit(ctx, ".", 1, ip->i_ino, DT_DIR))
fs/jfs/jfs_dtree.c
2859
if (!dir_emit(ctx, "..", 2, PARENT(ip), DT_DIR))
fs/jfs/jfs_dtree.c
2869
if (dtEmpty(ip)) {
fs/jfs/jfs_dtree.c
2874
if ((rc = dtReadNext(ip, &ctx->pos, &btstack))) {
fs/jfs/jfs_dtree.c
2881
DT_GETSEARCH(ip, btstack.top, bn, mp, p, index);
fs/jfs/jfs_dtree.c
2908
i, stbl[i], (long)ip->i_ino, (long long)bn);
fs/jfs/jfs_dtree.c
292
static inline void lock_index(tid_t tid, struct inode *ip, struct metapage * mp,
fs/jfs/jfs_dtree.c
2940
JFS_IP(ip)->next_index)) {
fs/jfs/jfs_dtree.c
2941
if (!page_fixed && !isReadOnly(ip)) {
fs/jfs/jfs_dtree.c
2979
jfs_error(ip->i_sb,
fs/jfs/jfs_dtree.c
2981
(long)ip->i_ino,
fs/jfs/jfs_dtree.c
299
tlck = txLock(tid, ip, mp, tlckDATA);
fs/jfs/jfs_dtree.c
3031
if ((rc = add_missing_indices(ip, bn)))
fs/jfs/jfs_dtree.c
3041
DT_GETPAGE(ip, bn, mp, PSIZE, p, rc);
fs/jfs/jfs_dtree.c
3060
static int dtReadFirst(struct inode *ip, struct btstack * btstack)
fs/jfs/jfs_dtree.c
3079
DT_GETPAGE(ip, bn, mp, psize, p, rc);
fs/jfs/jfs_dtree.c
3101
jfs_error(ip->i_sb, "btstack overrun\n");
fs/jfs/jfs_dtree.c
3113
jfs_error(ip->i_sb, "stbl[0] out of bound\n");
fs/jfs/jfs_dtree.c
3121
psize = lengthPXD(xd) << JFS_SBI(ip->i_sb)->l2bsize;
fs/jfs/jfs_dtree.c
3139
static int dtReadNext(struct inode *ip, loff_t * offset,
fs/jfs/jfs_dtree.c
3160
if ((rc = dtReadFirst(ip, btstack)))
fs/jfs/jfs_dtree.c
3164
DT_GETSEARCH(ip, btstack->top, bn, mp, p, index);
fs/jfs/jfs_dtree.c
322
static u32 add_index(tid_t tid, struct inode *ip, s64 bn, int slot)
fs/jfs/jfs_dtree.c
3223
DT_GETPAGE(ip, bn, mp, PSIZE, p, rc);
fs/jfs/jfs_dtree.c
324
struct super_block *sb = ip->i_sb;
fs/jfs/jfs_dtree.c
3244
DT_GETPAGE(ip, bn, mp, PSIZE, p, rc);
fs/jfs/jfs_dtree.c
326
struct jfs_inode_info *jfs_ip = JFS_IP(ip);
fs/jfs/jfs_dtree.c
3264
DT_GETPAGE(ip, bn, mp, PSIZE, p, rc);
fs/jfs/jfs_dtree.c
3285
DT_GETPAGE(ip, bn, mp, PSIZE, p, rc);
fs/jfs/jfs_dtree.c
338
ASSERT(DO_INDEX(ip));
fs/jfs/jfs_dtree.c
352
ip->i_size = (loff_t) (index - 1) << 3;
fs/jfs/jfs_dtree.c
362
set_cflag(COMMIT_Dirtable, ip);
fs/jfs/jfs_dtree.c
3670
if (data->leaf.ip) {
fs/jfs/jfs_dtree.c
3675
data->leaf.ip,
fs/jfs/jfs_dtree.c
373
if (dquot_alloc_block(ip, sbi->nbperpage))
fs/jfs/jfs_dtree.c
375
if (dbAlloc(ip, 0, sbi->nbperpage, &xaddr)) {
fs/jfs/jfs_dtree.c
3754
if ((p->header.flag & BT_LEAF) && data->leaf.ip) {
fs/jfs/jfs_dtree.c
376
dquot_free_block(ip, sbi->nbperpage);
fs/jfs/jfs_dtree.c
3764
modify_index(data->leaf.tid, data->leaf.ip,
fs/jfs/jfs_dtree.c
389
xtInitRoot(tid, ip);
fs/jfs/jfs_dtree.c
394
if (xtInsert(tid, ip, 0, 0, sbi->nbperpage, &xaddr, 0)) {
fs/jfs/jfs_dtree.c
399
dbFree(ip, xaddr, sbi->nbperpage);
fs/jfs/jfs_dtree.c
400
dquot_free_block(ip, sbi->nbperpage);
fs/jfs/jfs_dtree.c
403
ip->i_size = PSIZE;
fs/jfs/jfs_dtree.c
405
mp = get_index_page(ip, 0);
fs/jfs/jfs_dtree.c
408
xtTruncate(tid, ip, 0, COMMIT_PWMAP);
fs/jfs/jfs_dtree.c
413
tlck = txLock(tid, ip, mp, tlckDATA);
fs/jfs/jfs_dtree.c
4244
int dtModify(tid_t tid, struct inode *ip,
fs/jfs/jfs_dtree.c
4265
if ((rc = dtSearch(ip, key, orig_ino, &btstack, flag)))
fs/jfs/jfs_dtree.c
4269
DT_GETSEARCH(ip, btstack.top, bn, mp, p, index);
fs/jfs/jfs_dtree.c
4271
BT_MARK_DIRTY(mp, ip);
fs/jfs/jfs_dtree.c
4275
tlck = txLock(tid, ip, mp, tlckDTREE | tlckENTRY);
fs/jfs/jfs_dtree.c
430
clear_cflag(COMMIT_Dirtable, ip);
fs/jfs/jfs_dtree.c
441
if (xtInsert(tid, ip, 0, blkno, sbi->nbperpage, &xaddr, 0)) {
fs/jfs/jfs_dtree.c
445
ip->i_size += PSIZE;
fs/jfs/jfs_dtree.c
447
if ((mp = get_index_page(ip, blkno)))
fs/jfs/jfs_dtree.c
450
xtTruncate(tid, ip, offset, COMMIT_PWMAP);
fs/jfs/jfs_dtree.c
452
mp = read_index_page(ip, blkno);
fs/jfs/jfs_dtree.c
459
lock_index(tid, ip, mp, index);
fs/jfs/jfs_dtree.c
484
static void free_index(tid_t tid, struct inode *ip, u32 index, u32 next)
fs/jfs/jfs_dtree.c
490
dirtab_slot = find_index(ip, index, &mp, &lblock);
fs/jfs/jfs_dtree.c
500
lock_index(tid, ip, mp, index);
fs/jfs/jfs_dtree.c
504
set_cflag(COMMIT_Dirtable, ip);
fs/jfs/jfs_dtree.c
512
static void modify_index(tid_t tid, struct inode *ip, u32 index, s64 bn,
fs/jfs/jfs_dtree.c
517
dirtab_slot = find_index(ip, index, mp, lblock);
fs/jfs/jfs_dtree.c
526
lock_index(tid, ip, *mp, index);
fs/jfs/jfs_dtree.c
529
set_cflag(COMMIT_Dirtable, ip);
fs/jfs/jfs_dtree.c
537
static int read_index(struct inode *ip, u32 index,
fs/jfs/jfs_dtree.c
544
slot = find_index(ip, index, &mp, &lblock);
fs/jfs/jfs_dtree.c
568
int dtSearch(struct inode *ip, struct component_name * key, ino_t * data,
fs/jfs/jfs_dtree.c
583
struct super_block *sb = ip->i_sb;
fs/jfs/jfs_dtree.c
624
DT_GETPAGE(ip, bn, mp, psize, p, rc);
fs/jfs/jfs_dtree.c
786
psize = lengthPXD(pxd) << JFS_SBI(ip->i_sb)->l2bsize;
fs/jfs/jfs_dtree.c
815
int dtInsert(tid_t tid, struct inode *ip,
fs/jfs/jfs_dtree.c
837
DT_GETSEARCH(ip, btstack->top, bn, mp, p, index);
fs/jfs/jfs_dtree.c
844
if (DO_INDEX(ip)) {
fs/jfs/jfs_dtree.c
845
if (JFS_IP(ip)->next_index == DIREND) {
fs/jfs/jfs_dtree.c
851
data.leaf.ip = ip;
fs/jfs/jfs_dtree.c
854
data.leaf.ip = NULL; /* signifies legacy directory format */
fs/jfs/jfs_dtree.c
871
rc = dtSplitUp(tid, ip, &split, btstack);
fs/jfs/jfs_dtree.c
880
BT_MARK_DIRTY(mp, ip);
fs/jfs/jfs_dtree.c
884
tlck = txLock(tid, ip, mp, tlckDTREE | tlckENTRY);
fs/jfs/jfs_dtree.c
927
struct inode *ip, struct dtsplit * split, struct btstack * btstack)
fs/jfs/jfs_dtree.c
929
struct jfs_sb_info *sbi = JFS_SBI(ip->i_sb);
fs/jfs/jfs_dtree.c
954
sp = DT_PAGE(ip, smp);
fs/jfs/jfs_dtree.c
982
if ((rc = dbAlloc(ip, 0, (s64) xlen, &xaddr))) {
fs/jfs/jfs_dtree.c
993
rc = dtSplitRoot(tid, ip, split, &rmp);
fs/jfs/jfs_dtree.c
996
dbFree(ip, xaddr, xlen);
fs/jfs/jfs_dtree.h
17
struct inode *ip;
fs/jfs/jfs_dtree.h
241
extern void dtInitRoot(tid_t tid, struct inode *ip, u32 idotdot);
fs/jfs/jfs_dtree.h
243
extern int dtSearch(struct inode *ip, struct component_name * key,
fs/jfs/jfs_dtree.h
246
extern int dtInsert(tid_t tid, struct inode *ip, struct component_name * key,
fs/jfs/jfs_dtree.h
249
extern int dtDelete(tid_t tid, struct inode *ip, struct component_name * key,
fs/jfs/jfs_dtree.h
252
extern int dtModify(tid_t tid, struct inode *ip, struct component_name * key,
fs/jfs/jfs_extent.c
127
if ((rc = extBalloc(ip, hint ? hint : INOHINT(ip), &nxlen, &nxaddr))) {
fs/jfs/jfs_extent.c
128
mutex_unlock(&JFS_IP(ip)->commit_mutex);
fs/jfs/jfs_extent.c
133
rc = dquot_alloc_block(ip, nxlen);
fs/jfs/jfs_extent.c
135
dbFree(ip, nxaddr, (s64) nxlen);
fs/jfs/jfs_extent.c
136
mutex_unlock(&JFS_IP(ip)->commit_mutex);
fs/jfs/jfs_extent.c
148
rc = xtExtend(0, ip, xoff, (int) nxlen, 0);
fs/jfs/jfs_extent.c
150
rc = xtInsert(0, ip, xflag, xoff, (int) nxlen, &nxaddr, 0);
fs/jfs/jfs_extent.c
156
dbFree(ip, nxaddr, nxlen);
fs/jfs/jfs_extent.c
157
dquot_free_block(ip, nxlen);
fs/jfs/jfs_extent.c
158
mutex_unlock(&JFS_IP(ip)->commit_mutex);
fs/jfs/jfs_extent.c
168
mark_inode_dirty(ip);
fs/jfs/jfs_extent.c
170
mutex_unlock(&JFS_IP(ip)->commit_mutex);
fs/jfs/jfs_extent.c
176
if (test_and_clear_cflag(COMMIT_Synclist,ip))
fs/jfs/jfs_extent.c
177
jfs_commit_inode(ip, 0);
fs/jfs/jfs_extent.c
197
int extHint(struct inode *ip, s64 offset, xad_t * xp)
fs/jfs/jfs_extent.c
199
struct super_block *sb = ip->i_sb;
fs/jfs/jfs_extent.c
220
rc = xtLookup(ip, prev, nbperpage, &xflag, &xaddr, &xlen, 0);
fs/jfs/jfs_extent.c
224
jfs_error(ip->i_sb, "corrupt xtree\n");
fs/jfs/jfs_extent.c
257
int extRecord(struct inode *ip, xad_t * xp)
fs/jfs/jfs_extent.c
261
if (isReadOnly(ip)) {
fs/jfs/jfs_extent.c
262
jfs_error(ip->i_sb, "read-only filesystem\n");
fs/jfs/jfs_extent.c
266
txBeginAnon(ip->i_sb);
fs/jfs/jfs_extent.c
268
mutex_lock(&JFS_IP(ip)->commit_mutex);
fs/jfs/jfs_extent.c
271
rc = xtUpdate(0, ip, xp);
fs/jfs/jfs_extent.c
273
mutex_unlock(&JFS_IP(ip)->commit_mutex);
fs/jfs/jfs_extent.c
309
extBalloc(struct inode *ip, s64 hint, s64 * nblocks, s64 * blkno)
fs/jfs/jfs_extent.c
311
struct jfs_inode_info *ji = JFS_IP(ip);
fs/jfs/jfs_extent.c
312
struct jfs_sb_info *sbi = JFS_SBI(ip->i_sb);
fs/jfs/jfs_extent.c
336
while ((rc = dbAlloc(ip, hint, nb, &daddr)) != 0) {
fs/jfs/jfs_extent.c
354
if (S_ISREG(ip->i_mode) && (ji->fileset == FILESYSTEM_I)) {
fs/jfs/jfs_extent.c
70
extAlloc(struct inode *ip, s64 xlen, s64 pno, xad_t * xp, bool abnr)
fs/jfs/jfs_extent.c
72
struct jfs_sb_info *sbi = JFS_SBI(ip->i_sb);
fs/jfs/jfs_extent.c
77
if (isReadOnly(ip)) {
fs/jfs/jfs_extent.c
78
jfs_error(ip->i_sb, "read-only filesystem\n");
fs/jfs/jfs_extent.c
83
txBeginAnon(ip->i_sb);
fs/jfs/jfs_extent.c
86
mutex_lock(&JFS_IP(ip)->commit_mutex);
fs/jfs/jfs_extent.h
10
(addressPXD(&(JFS_IP(ip)->ixpxd)) + lengthPXD(&(JFS_IP(ip)->ixpxd)) - 1)
fs/jfs/jfs_extent.h
9
#define INOHINT(ip) \
fs/jfs/jfs_imap.c
1121
invalidate_pxd_metapages(ip, freepxd);
fs/jfs/jfs_imap.c
1182
jfs_error(ip->i_sb, "the pmap does not show inode free\n");
fs/jfs/jfs_imap.c
1296
diInitInode(struct inode *ip, int iagno, int ino, int extno, struct iag * iagp)
fs/jfs/jfs_imap.c
1298
struct jfs_inode_info *jfs_ip = JFS_IP(ip);
fs/jfs/jfs_imap.c
1300
ip->i_ino = (iagno << L2INOSPERIAG) + ino;
fs/jfs/jfs_imap.c
1323
int diAlloc(struct inode *pip, bool dir, struct inode *ip)
fs/jfs/jfs_imap.c
1339
JFS_IP(ip)->ipimap = ipimap;
fs/jfs/jfs_imap.c
1340
JFS_IP(ip)->fileset = FILESYSTEM_I;
fs/jfs/jfs_imap.c
1438
diInitInode(ip, iagno, ino, extno,
fs/jfs/jfs_imap.c
1503
jfs_error(ip->i_sb,
fs/jfs/jfs_imap.c
1521
diInitInode(ip, iagno, ino, extno,
fs/jfs/jfs_imap.c
1560
diInitInode(ip, iagno,
fs/jfs/jfs_imap.c
1590
rc = diAllocAG(imap, agno, dir, ip);
fs/jfs/jfs_imap.c
1600
return (diAllocAny(imap, agno, dir, ip));
fs/jfs/jfs_imap.c
1634
diAllocAG(struct inomap * imap, int agno, bool dir, struct inode *ip)
fs/jfs/jfs_imap.c
1645
jfs_error(ip->i_sb, "numfree > numinos\n");
fs/jfs/jfs_imap.c
1669
if ((rc = diAllocExt(imap, agno, ip)) != -ENOSPC)
fs/jfs/jfs_imap.c
1676
return (diAllocIno(imap, agno, ip));
fs/jfs/jfs_imap.c
1704
diAllocAny(struct inomap * imap, int agno, bool dir, struct inode *ip)
fs/jfs/jfs_imap.c
1716
rc = diAllocAG(imap, ag, dir, ip);
fs/jfs/jfs_imap.c
1729
rc = diAllocAG(imap, ag, dir, ip);
fs/jfs/jfs_imap.c
1768
static int diAllocIno(struct inomap * imap, int agno, struct inode *ip)
fs/jfs/jfs_imap.c
1796
jfs_error(ip->i_sb, "nfreeinos = 0, but iag on freelist\n");
fs/jfs/jfs_imap.c
1807
jfs_error(ip->i_sb,
fs/jfs/jfs_imap.c
1823
jfs_error(ip->i_sb, "no free extent found\n");
fs/jfs/jfs_imap.c
1834
jfs_error(ip->i_sb, "free inode not found\n");
fs/jfs/jfs_imap.c
1853
diInitInode(ip, iagno, ino, extno, iagp);
fs/jfs/jfs_imap.c
1892
static int diAllocExt(struct inomap * imap, int agno, struct inode *ip)
fs/jfs/jfs_imap.c
1920
jfs_error(ip->i_sb, "error reading iag\n");
fs/jfs/jfs_imap.c
1932
jfs_error(ip->i_sb, "free ext summary map not found\n");
fs/jfs/jfs_imap.c
1945
jfs_error(ip->i_sb, "free extent not found\n");
fs/jfs/jfs_imap.c
1971
diInitInode(ip, iagno, extno << L2INOSPEREXT, extno, iagp);
fs/jfs/jfs_imap.c
290
int diRead(struct inode *ip)
fs/jfs/jfs_imap.c
292
struct jfs_sb_info *sbi = JFS_SBI(ip->i_sb);
fs/jfs/jfs_imap.c
2987
struct inode *ip;
fs/jfs/jfs_imap.c
2993
ip = diReadSpecial(sb, FILESYSTEM_I, 1);
fs/jfs/jfs_imap.c
2994
if (ip == NULL) {
fs/jfs/jfs_imap.c
3010
if (xtInsert(tid, ip, 0, blkno, xlen, xaddr, 0)) {
fs/jfs/jfs_imap.c
3017
ip->i_size += PSIZE;
fs/jfs/jfs_imap.c
3018
inode_add_bytes(ip, PSIZE);
fs/jfs/jfs_imap.c
3019
txCommit(tid, 1, &ip, COMMIT_FORCE);
fs/jfs/jfs_imap.c
3022
diFreeSpecial(ip);
fs/jfs/jfs_imap.c
3034
static int copy_from_dinode(struct dinode * dip, struct inode *ip)
fs/jfs/jfs_imap.c
3036
struct jfs_inode_info *jfs_ip = JFS_IP(ip);
fs/jfs/jfs_imap.c
3037
struct jfs_sb_info *sbi = JFS_SBI(ip->i_sb);
fs/jfs/jfs_imap.c
305
jfs_info("diRead: ino = %ld", ip->i_ino);
fs/jfs/jfs_imap.c
3050
jfs_set_inode_flags(ip);
fs/jfs/jfs_imap.c
3052
ip->i_mode = le32_to_cpu(dip->di_mode) & 0xffff;
fs/jfs/jfs_imap.c
3054
ip->i_mode = (ip->i_mode & ~0777) | (0777 & ~sbi->umask);
fs/jfs/jfs_imap.c
3056
if (S_ISDIR(ip->i_mode)) {
fs/jfs/jfs_imap.c
3057
if (ip->i_mode & 0400)
fs/jfs/jfs_imap.c
3058
ip->i_mode |= 0100;
fs/jfs/jfs_imap.c
3059
if (ip->i_mode & 0040)
fs/jfs/jfs_imap.c
3060
ip->i_mode |= 0010;
fs/jfs/jfs_imap.c
3061
if (ip->i_mode & 0004)
fs/jfs/jfs_imap.c
3062
ip->i_mode |= 0001;
fs/jfs/jfs_imap.c
3065
set_nlink(ip, le32_to_cpu(dip->di_nlink));
fs/jfs/jfs_imap.c
3069
ip->i_uid = jfs_ip->saved_uid;
fs/jfs/jfs_imap.c
3071
ip->i_uid = sbi->uid;
fs/jfs/jfs_imap.c
3076
ip->i_gid = jfs_ip->saved_gid;
fs/jfs/jfs_imap.c
3078
ip->i_gid = sbi->gid;
fs/jfs/jfs_imap.c
308
JFS_IP(ip)->ipimap = ipimap;
fs/jfs/jfs_imap.c
3081
ip->i_size = le64_to_cpu(dip->di_size);
fs/jfs/jfs_imap.c
3082
inode_set_atime(ip, le32_to_cpu(dip->di_atime.tv_sec),
fs/jfs/jfs_imap.c
3084
inode_set_mtime(ip, le32_to_cpu(dip->di_mtime.tv_sec),
fs/jfs/jfs_imap.c
3086
inode_set_ctime(ip, le32_to_cpu(dip->di_ctime.tv_sec),
fs/jfs/jfs_imap.c
3088
ip->i_blocks = LBLK2PBLK(ip->i_sb, le64_to_cpu(dip->di_nblocks));
fs/jfs/jfs_imap.c
3089
ip->i_generation = le32_to_cpu(dip->di_gen);
fs/jfs/jfs_imap.c
3098
if (S_ISCHR(ip->i_mode) || S_ISBLK(ip->i_mode)) {
fs/jfs/jfs_imap.c
3100
ip->i_rdev = new_decode_dev(jfs_ip->dev);
fs/jfs/jfs_imap.c
3103
if (S_ISDIR(ip->i_mode)) {
fs/jfs/jfs_imap.c
3105
} else if (S_ISREG(ip->i_mode) || S_ISLNK(ip->i_mode)) {
fs/jfs/jfs_imap.c
311
iagno = INOTOIAG(ip->i_ino);
fs/jfs/jfs_imap.c
3127
static void copy_to_dinode(struct dinode * dip, struct inode *ip)
fs/jfs/jfs_imap.c
3129
struct jfs_inode_info *jfs_ip = JFS_IP(ip);
fs/jfs/jfs_imap.c
3130
struct jfs_sb_info *sbi = JFS_SBI(ip->i_sb);
fs/jfs/jfs_imap.c
3134
dip->di_number = cpu_to_le32(ip->i_ino);
fs/jfs/jfs_imap.c
3135
dip->di_gen = cpu_to_le32(ip->i_generation);
fs/jfs/jfs_imap.c
3136
dip->di_size = cpu_to_le64(ip->i_size);
fs/jfs/jfs_imap.c
3137
dip->di_nblocks = cpu_to_le64(PBLK2LBLK(ip->i_sb, ip->i_blocks));
fs/jfs/jfs_imap.c
3138
dip->di_nlink = cpu_to_le32(ip->i_nlink);
fs/jfs/jfs_imap.c
3140
dip->di_uid = cpu_to_le32(i_uid_read(ip));
fs/jfs/jfs_imap.c
3145
dip->di_gid = cpu_to_le32(i_gid_read(ip));
fs/jfs/jfs_imap.c
3155
ip->i_mode);
fs/jfs/jfs_imap.c
3159
dip->di_atime.tv_sec = cpu_to_le32(inode_get_atime_sec(ip));
fs/jfs/jfs_imap.c
3160
dip->di_atime.tv_nsec = cpu_to_le32(inode_get_atime_nsec(ip));
fs/jfs/jfs_imap.c
3161
dip->di_ctime.tv_sec = cpu_to_le32(inode_get_ctime_sec(ip));
fs/jfs/jfs_imap.c
3162
dip->di_ctime.tv_nsec = cpu_to_le32(inode_get_ctime_nsec(ip));
fs/jfs/jfs_imap.c
3163
dip->di_mtime.tv_sec = cpu_to_le32(inode_get_mtime_sec(ip));
fs/jfs/jfs_imap.c
3164
dip->di_mtime.tv_nsec = cpu_to_le32(inode_get_mtime_nsec(ip));
fs/jfs/jfs_imap.c
3172
if (S_ISCHR(ip->i_mode) || S_ISBLK(ip->i_mode))
fs/jfs/jfs_imap.c
326
ino = ip->i_ino & (INOSPERIAG - 1);
fs/jfs/jfs_imap.c
342
agno = BLKTOAG(agstart, JFS_SBI(ip->i_sb));
fs/jfs/jfs_imap.c
377
if (ip->i_ino != le32_to_cpu(dp->di_number)) {
fs/jfs/jfs_imap.c
378
jfs_error(ip->i_sb, "i_ino != di_number\n");
fs/jfs/jfs_imap.c
384
rc = copy_from_dinode(dp, ip);
fs/jfs/jfs_imap.c
389
JFS_IP(ip)->agstart = agstart;
fs/jfs/jfs_imap.c
390
JFS_IP(ip)->active_ag = -1;
fs/jfs/jfs_imap.c
423
struct inode *ip;
fs/jfs/jfs_imap.c
426
ip = new_inode(sb);
fs/jfs/jfs_imap.c
427
if (ip == NULL) {
fs/jfs/jfs_imap.c
429
return ip;
fs/jfs/jfs_imap.c
434
JFS_IP(ip)->ipimap = sbi->ipaimap2;
fs/jfs/jfs_imap.c
437
JFS_IP(ip)->ipimap = sbi->ipaimap;
fs/jfs/jfs_imap.c
442
ip->i_ino = inum;
fs/jfs/jfs_imap.c
447
mp = read_metapage(ip, address << sbi->l2nbperpage, PSIZE, 1);
fs/jfs/jfs_imap.c
449
set_nlink(ip, 1); /* Don't want iput() deleting it */
fs/jfs/jfs_imap.c
450
iput(ip);
fs/jfs/jfs_imap.c
459
if ((copy_from_dinode(dp, ip) != 0) || (ip->i_nlink == 0)) {
fs/jfs/jfs_imap.c
461
set_nlink(ip, 1); /* Don't want iput() deleting it */
fs/jfs/jfs_imap.c
462
iput(ip);
fs/jfs/jfs_imap.c
469
ip->i_mapping->a_ops = &jfs_metapage_aops;
fs/jfs/jfs_imap.c
470
mapping_set_gfp_mask(ip->i_mapping, GFP_NOFS);
fs/jfs/jfs_imap.c
473
ip->i_flags |= S_NOQUOTA;
fs/jfs/jfs_imap.c
475
if ((inum == FILESYSTEM_I) && (JFS_IP(ip)->ipimap == sbi->ipaimap)) {
fs/jfs/jfs_imap.c
483
inode_fake_hash(ip);
fs/jfs/jfs_imap.c
485
return (ip);
fs/jfs/jfs_imap.c
500
void diWriteSpecial(struct inode *ip, int secondary)
fs/jfs/jfs_imap.c
502
struct jfs_sb_info *sbi = JFS_SBI(ip->i_sb);
fs/jfs/jfs_imap.c
505
ino_t inum = ip->i_ino;
fs/jfs/jfs_imap.c
518
mp = read_metapage(ip, address << sbi->l2nbperpage, PSIZE, 1);
fs/jfs/jfs_imap.c
529
copy_to_dinode(dp, ip);
fs/jfs/jfs_imap.c
530
memcpy(&dp->di_xtroot, &JFS_IP(ip)->i_xtroot, 288);
fs/jfs/jfs_imap.c
544
void diFreeSpecial(struct inode *ip)
fs/jfs/jfs_imap.c
546
if (ip == NULL) {
fs/jfs/jfs_imap.c
550
filemap_write_and_wait(ip->i_mapping);
fs/jfs/jfs_imap.c
551
truncate_inode_pages(ip->i_mapping, 0);
fs/jfs/jfs_imap.c
552
iput(ip);
fs/jfs/jfs_imap.c
581
int diWrite(tid_t tid, struct inode *ip)
fs/jfs/jfs_imap.c
583
struct jfs_sb_info *sbi = JFS_SBI(ip->i_sb);
fs/jfs/jfs_imap.c
584
struct jfs_inode_info *jfs_ip = JFS_IP(ip);
fs/jfs/jfs_imap.c
605
ino = ip->i_ino & (INOSPERIAG - 1);
fs/jfs/jfs_imap.c
610
jfs_error(ip->i_sb, "ixpxd invalid\n");
fs/jfs/jfs_imap.c
671
if (S_ISDIR(ip->i_mode) && (lid = jfs_ip->xtlid)) {
fs/jfs/jfs_imap.c
764
if (S_ISLNK(ip->i_mode) && ip->i_size < IDATASIZE) {
fs/jfs/jfs_imap.c
775
if (test_cflag(COMMIT_Inlineea, ip)) {
fs/jfs/jfs_imap.c
782
clear_cflag(COMMIT_Inlineea, ip);
fs/jfs/jfs_imap.c
790
copy_to_dinode(dp, ip);
fs/jfs/jfs_imap.c
791
if (test_and_clear_cflag(COMMIT_Dirtable, ip)) {
fs/jfs/jfs_imap.c
845
int diFree(struct inode *ip)
fs/jfs/jfs_imap.c
848
ino_t inum = ip->i_ino;
fs/jfs/jfs_imap.c
854
struct inode *ipimap = JFS_SBI(ip->i_sb)->ipimap;
fs/jfs/jfs_imap.c
878
jfs_error(ip->i_sb, "inum = %d, iagno = %d, nextiag = %d\n",
fs/jfs/jfs_imap.c
885
agno = BLKTOAG(JFS_IP(ip)->agstart, JFS_SBI(ip->i_sb));
fs/jfs/jfs_imap.c
914
jfs_error(ip->i_sb, "wmap shows inode already free\n");
fs/jfs/jfs_imap.c
921
jfs_error(ip->i_sb, "invalid inoext\n");
fs/jfs/jfs_imap.c
933
jfs_error(ip->i_sb, "numfree > numinos\n");
fs/jfs/jfs_incore.h
108
#define IREAD_LOCK(ip, subclass) \
fs/jfs/jfs_incore.h
109
down_read_nested(&JFS_IP(ip)->rdwrlock, subclass)
fs/jfs/jfs_incore.h
110
#define IREAD_UNLOCK(ip) up_read(&JFS_IP(ip)->rdwrlock)
fs/jfs/jfs_incore.h
111
#define IWRITE_LOCK(ip, subclass) \
fs/jfs/jfs_incore.h
112
down_write_nested(&JFS_IP(ip)->rdwrlock, subclass)
fs/jfs/jfs_incore.h
113
#define IWRITE_UNLOCK(ip) up_write(&JFS_IP(ip)->rdwrlock)
fs/jfs/jfs_incore.h
151
#define set_cflag(flag, ip) set_bit(flag, &(JFS_IP(ip)->cflag))
fs/jfs/jfs_incore.h
152
#define clear_cflag(flag, ip) clear_bit(flag, &(JFS_IP(ip)->cflag))
fs/jfs/jfs_incore.h
153
#define test_cflag(flag, ip) test_bit(flag, &(JFS_IP(ip)->cflag))
fs/jfs/jfs_incore.h
154
#define test_and_clear_cflag(flag, ip) \
fs/jfs/jfs_incore.h
155
test_and_clear_bit(flag, &(JFS_IP(ip)->cflag))
fs/jfs/jfs_logmgr.c
386
p = (caddr_t) &JFS_IP(tlck->ip)->i_dtroot;
fs/jfs/jfs_logmgr.c
388
p = (caddr_t) &JFS_IP(tlck->ip)->i_xtroot;
fs/jfs/jfs_metapage.c
898
void __invalidate_metapages(struct inode *ip, s64 addr, int len)
fs/jfs/jfs_metapage.c
901
int l2BlocksPerPage = PAGE_SHIFT - ip->i_blkbits;
fs/jfs/jfs_metapage.c
905
JFS_SBI(ip->i_sb)->direct_inode->i_mapping;
fs/jfs/jfs_metapage.h
135
#define invalidate_pxd_metapages(ip, pxd) \
fs/jfs/jfs_metapage.h
136
__invalidate_metapages((ip), addressPXD(&(pxd)), lengthPXD(&(pxd)))
fs/jfs/jfs_metapage.h
137
#define invalidate_dxd_metapages(ip, dxd) \
fs/jfs/jfs_metapage.h
138
__invalidate_metapages((ip), addressDXD(&(dxd)), lengthDXD(&(dxd)))
fs/jfs/jfs_metapage.h
139
#define invalidate_xad_metapages(ip, xad) \
fs/jfs/jfs_metapage.h
140
__invalidate_metapages((ip), addressXAD(&(xad)), lengthXAD(&(xad)))
fs/jfs/jfs_txnmgr.c
1011
if (S_ISDIR(ip->i_mode))
fs/jfs/jfs_txnmgr.c
1013
tlck->ip = ip;
fs/jfs/jfs_txnmgr.c
1138
struct inode *ip;
fs/jfs/jfs_txnmgr.c
1199
ip = cd.iplist[n];
fs/jfs/jfs_txnmgr.c
1200
if (ip->i_ino > top) {
fs/jfs/jfs_txnmgr.c
1201
top = ip->i_ino;
fs/jfs/jfs_txnmgr.c
1203
cd.iplist[k] = ip;
fs/jfs/jfs_txnmgr.c
1207
ip = cd.iplist[k];
fs/jfs/jfs_txnmgr.c
1208
jfs_ip = JFS_IP(ip);
fs/jfs/jfs_txnmgr.c
1238
clear_cflag(COMMIT_Dirty, ip);
fs/jfs/jfs_txnmgr.c
1256
if (((rc = diWrite(tid, ip))))
fs/jfs/jfs_txnmgr.c
1272
ihold(tblk->u.ip);
fs/jfs/jfs_txnmgr.c
1290
if (inode_state_read_once(tblk->u.ip) & I_SYNC)
fs/jfs/jfs_txnmgr.c
1295
((tblk->u.ip->i_nlink == 0) &&
fs/jfs/jfs_txnmgr.c
1296
!test_cflag(COMMIT_Nolink, tblk->u.ip)));
fs/jfs/jfs_txnmgr.c
1343
ip = cd.iplist[k];
fs/jfs/jfs_txnmgr.c
1344
jfs_ip = JFS_IP(ip);
fs/jfs/jfs_txnmgr.c
1375
struct inode *ip;
fs/jfs/jfs_txnmgr.c
1389
ip = tlck->ip;
fs/jfs/jfs_txnmgr.c
1390
lrd->aggregate = cpu_to_le32(JFS_SBI(ip->i_sb)->aggregate);
fs/jfs/jfs_txnmgr.c
1391
lrd->log.redopage.fileset = cpu_to_le32(JFS_IP(ip)->fileset);
fs/jfs/jfs_txnmgr.c
1392
lrd->log.redopage.inode = cpu_to_le32(ip->i_ino);
fs/jfs/jfs_txnmgr.c
1524
if (jfs_dirtable_inline(tlck->ip)) {
fs/jfs/jfs_txnmgr.c
159
static void txAllocPMap(struct inode *ip, struct maplock * maplock,
fs/jfs/jfs_txnmgr.c
1659
struct inode *ip;
fs/jfs/jfs_txnmgr.c
1669
ip = tlck->ip;
fs/jfs/jfs_txnmgr.c
1680
p = (xtpage_t *) &JFS_IP(ip)->i_xtroot;
fs/jfs/jfs_txnmgr.c
1681
if (S_ISDIR(ip->i_mode))
fs/jfs/jfs_txnmgr.c
1758
tlck->ip, mp, tlck, lwm, xadlock->count);
fs/jfs/jfs_txnmgr.c
1861
tlck->ip, mp, xadlock->count);
fs/jfs/jfs_txnmgr.c
1866
if (((tblk->xflag & COMMIT_PWMAP) || S_ISDIR(ip->i_mode))
fs/jfs/jfs_txnmgr.c
1984
tlck->ip, mp, xadlock->count, lwm, next);
fs/jfs/jfs_txnmgr.c
2005
ip, mp, pxdlock->count, hwm);
fs/jfs/jfs_txnmgr.c
2024
tlck->ip, mp, xadlock->count, next, hwm);
fs/jfs/jfs_txnmgr.c
2126
void txEA(tid_t tid, struct inode *ip, dxd_t * oldea, dxd_t * newea)
fs/jfs/jfs_txnmgr.c
2140
tlck = txMaplock(tid, ip, tlckMAP);
fs/jfs/jfs_txnmgr.c
2151
set_cflag(COMMIT_Inlineea, ip);
fs/jfs/jfs_txnmgr.c
2158
if (!test_cflag(COMMIT_Nolink, ip) && oldea->flag & DXD_EXTENT) {
fs/jfs/jfs_txnmgr.c
2160
tlck = txMaplock(tid, ip, tlckMAP);
fs/jfs/jfs_txnmgr.c
2248
struct inode *ip;
fs/jfs/jfs_txnmgr.c
2366
ip = tblk->u.ip;
fs/jfs/jfs_txnmgr.c
2367
diUpdatePMap(ipimap, ip->i_ino, true, tblk);
fs/jfs/jfs_txnmgr.c
2368
iput(ip);
fs/jfs/jfs_txnmgr.c
2392
static void txAllocPMap(struct inode *ip, struct maplock * maplock,
fs/jfs/jfs_txnmgr.c
2395
struct inode *ipbmap = JFS_SBI(ip->i_sb)->ipbmap;
fs/jfs/jfs_txnmgr.c
2450
void txFreeMap(struct inode *ip,
fs/jfs/jfs_txnmgr.c
2453
struct inode *ipbmap = JFS_SBI(ip->i_sb)->ipbmap;
fs/jfs/jfs_txnmgr.c
2516
dbFree(ip, xaddr, (s64) xlen);
fs/jfs/jfs_txnmgr.c
2525
dbFree(ip, xaddr, (s64) xlen);
fs/jfs/jfs_txnmgr.c
2535
dbFree(ip, xaddr, (s64) xlen);
fs/jfs/jfs_txnmgr.c
2548
void txFreelock(struct inode *ip)
fs/jfs/jfs_txnmgr.c
2550
struct jfs_inode_info *jfs_ip = JFS_IP(ip);
fs/jfs/jfs_txnmgr.c
2608
JFS_IP(tlck->ip)->xtlid = 0;
fs/jfs/jfs_txnmgr.c
2814
struct inode *ip;
fs/jfs/jfs_txnmgr.c
2827
ip = &jfs_ip->vfs_inode;
fs/jfs/jfs_txnmgr.c
2834
tid = txBegin(ip->i_sb, COMMIT_INODE | COMMIT_FORCE);
fs/jfs/jfs_txnmgr.c
2836
txCommit(tid, 1, &ip, 0);
fs/jfs/jfs_txnmgr.c
2885
struct inode *ip;
fs/jfs/jfs_txnmgr.c
2899
ip = &jfs_ip->vfs_inode;
fs/jfs/jfs_txnmgr.c
2901
if (! igrab(ip)) {
fs/jfs/jfs_txnmgr.c
2912
tid = txBegin(ip->i_sb, COMMIT_INODE);
fs/jfs/jfs_txnmgr.c
2913
txCommit(tid, 1, &ip, 0);
fs/jfs/jfs_txnmgr.c
2917
iput(ip);
fs/jfs/jfs_txnmgr.c
2936
iput(ip);
fs/jfs/jfs_txnmgr.c
583
struct tlock *txLock(tid_t tid, struct inode *ip, struct metapage * mp,
fs/jfs/jfs_txnmgr.c
586
struct jfs_inode_info *jfs_ip = JFS_IP(ip);
fs/jfs/jfs_txnmgr.c
598
if (S_ISDIR(ip->i_mode) && (type & tlckXTREE) &&
fs/jfs/jfs_txnmgr.c
613
jfs_info("txLock: tid:%d ip:0x%p mp:0x%p lid:%d", tid, ip, mp, lid);
fs/jfs/jfs_txnmgr.c
714
set_cflag(COMMIT_Synclist, ip);
fs/jfs/jfs_txnmgr.c
720
if (S_ISDIR(ip->i_mode))
fs/jfs/jfs_txnmgr.c
726
tlck->ip = ip;
fs/jfs/jfs_txnmgr.c
827
ip, sizeof(*ip), 0);
fs/jfs/jfs_txnmgr.c
988
struct tlock *txMaplock(tid_t tid, struct inode *ip, int type)
fs/jfs/jfs_txnmgr.c
990
struct jfs_inode_info *jfs_ip = JFS_IP(ip);
fs/jfs/jfs_txnmgr.h
53
struct inode *ip; /* inode being deleted */
fs/jfs/jfs_txnmgr.h
92
struct inode *ip; /* 4/8: object */
fs/jfs/jfs_xtree.c
1008
BT_MARK_DIRTY(smp, ip);
fs/jfs/jfs_xtree.c
1010
if (!test_cflag(COMMIT_Nolink, ip)) {
fs/jfs/jfs_xtree.c
1014
tlck = txLock(tid, ip, rmp, tlckXTREE | tlckNEW);
fs/jfs/jfs_xtree.c
1020
tlck = txLock(tid, ip, smp, tlckXTREE | tlckGROW);
fs/jfs/jfs_xtree.c
1060
if (!test_cflag(COMMIT_Nolink, ip)) {
fs/jfs/jfs_xtree.c
1080
p = xt_getpage(ip, nextbn, &mp);
fs/jfs/jfs_xtree.c
1086
BT_MARK_DIRTY(mp, ip);
fs/jfs/jfs_xtree.c
1092
if (!test_cflag(COMMIT_Nolink, ip))
fs/jfs/jfs_xtree.c
1093
tlck = txLock(tid, ip, mp, tlckXTREE | tlckRELINK);
fs/jfs/jfs_xtree.c
111
static inline xtpage_t *xt_getpage(struct inode *ip, s64 bn, struct metapage **mp)
fs/jfs/jfs_xtree.c
1131
if (!test_cflag(COMMIT_Nolink, ip)) {
fs/jfs/jfs_xtree.c
116
BT_GETPAGE(ip, bn, *mp, xtpage_t, PSIZE, p, rc, i_xtroot);
fs/jfs/jfs_xtree.c
1161
if (!test_cflag(COMMIT_Nolink, ip)) {
fs/jfs/jfs_xtree.c
1170
if (!test_cflag(COMMIT_Nolink, ip)) {
fs/jfs/jfs_xtree.c
1189
dquot_free_block(ip, quota_allocation);
fs/jfs/jfs_xtree.c
1217
struct inode *ip, struct xtsplit * split, struct metapage ** rmpp)
fs/jfs/jfs_xtree.c
1231
sp = (xtpage_t *) &JFS_IP(ip)->i_xtroot;
fs/jfs/jfs_xtree.c
1242
rmp = get_metapage(ip, rbn, PSIZE, 1);
fs/jfs/jfs_xtree.c
1247
rc = dquot_alloc_block(ip, lengthPXD(pxd));
fs/jfs/jfs_xtree.c
125
jfs_error(ip->i_sb, "xt_getpage: xtree page corrupt\n");
fs/jfs/jfs_xtree.c
1253
jfs_info("xtSplitRoot: ip:0x%p rmp:0x%p", ip, rmp);
fs/jfs/jfs_xtree.c
1260
BT_MARK_DIRTY(rmp, ip);
fs/jfs/jfs_xtree.c
1296
if (!test_cflag(COMMIT_Nolink, ip)) {
fs/jfs/jfs_xtree.c
1297
tlck = txLock(tid, ip, rmp, tlckXTREE | tlckNEW);
fs/jfs/jfs_xtree.c
1316
BT_MARK_DIRTY(split->mp, ip);
fs/jfs/jfs_xtree.c
1319
XT_PUTENTRY(xad, XAD_NEW, 0, JFS_SBI(ip->i_sb)->nbperpage, rbn);
fs/jfs/jfs_xtree.c
1327
if (!test_cflag(COMMIT_Nolink, ip)) {
fs/jfs/jfs_xtree.c
1328
tlck = txLock(tid, ip, split->mp, tlckXTREE | tlckGROW);
fs/jfs/jfs_xtree.c
1352
struct inode *ip, s64 xoff, /* delta extent offset */
fs/jfs/jfs_xtree.c
1372
if ((rc = xtSearch(ip, xoff - 1, NULL, &cmp, &btstack, XT_INSERT)))
fs/jfs/jfs_xtree.c
1376
XT_GETSEARCH(ip, btstack.top, bn, mp, p, index);
fs/jfs/jfs_xtree.c
138
int xtLookup(struct inode *ip, s64 lstart,
fs/jfs/jfs_xtree.c
1380
jfs_error(ip->i_sb, "xtSearch did not find extent\n");
fs/jfs/jfs_xtree.c
1388
jfs_error(ip->i_sb, "extension is not contiguous\n");
fs/jfs/jfs_xtree.c
1397
BT_MARK_DIRTY(mp, ip);
fs/jfs/jfs_xtree.c
1398
if (!test_cflag(COMMIT_Nolink, ip)) {
fs/jfs/jfs_xtree.c
1399
tlck = txLock(tid, ip, mp, tlckXTREE | tlckGROW);
fs/jfs/jfs_xtree.c
1431
if ((rc = xtSplitUp(tid, ip, &split, &btstack)))
fs/jfs/jfs_xtree.c
1435
p = xt_getpage(ip, bn, &mp);
fs/jfs/jfs_xtree.c
1451
p = xt_getpage(ip, bn, &mp);
fs/jfs/jfs_xtree.c
1455
BT_MARK_DIRTY(mp, ip);
fs/jfs/jfs_xtree.c
1456
if (!test_cflag(COMMIT_Nolink, ip)) {
fs/jfs/jfs_xtree.c
1457
tlck = txLock(tid, ip, mp, tlckXTREE|tlckGROW);
fs/jfs/jfs_xtree.c
1486
if (!test_cflag(COMMIT_Nolink, ip)) {
fs/jfs/jfs_xtree.c
1513
int xtUpdate(tid_t tid, struct inode *ip, xad_t * nxad)
fs/jfs/jfs_xtree.c
1537
if ((rc = xtSearch(ip, nxoff, NULL, &cmp, &btstack, XT_INSERT)))
fs/jfs/jfs_xtree.c
1541
XT_GETSEARCH(ip, btstack.top, bn, mp, p, index0);
fs/jfs/jfs_xtree.c
1545
jfs_error(ip->i_sb, "Could not find extent\n");
fs/jfs/jfs_xtree.c
1549
BT_MARK_DIRTY(mp, ip);
fs/jfs/jfs_xtree.c
1553
if (!test_cflag(COMMIT_Nolink, ip)) {
fs/jfs/jfs_xtree.c
1554
tlck = txLock(tid, ip, mp, tlckXTREE | tlckGROW);
fs/jfs/jfs_xtree.c
1568
jfs_error(ip->i_sb,
fs/jfs/jfs_xtree.c
158
size = ((u64) ip->i_size + (JFS_SBI(ip->i_sb)->bsize - 1)) >>
fs/jfs/jfs_xtree.c
159
JFS_SBI(ip->i_sb)->l2bsize;
fs/jfs/jfs_xtree.c
168
if ((rc = xtSearch(ip, lstart, &next, &cmp, &btstack, 0))) {
fs/jfs/jfs_xtree.c
1698
jfs_error(ip->i_sb, "xoff >= nxoff\n");
fs/jfs/jfs_xtree.c
1725
if ((rc = xtSplitUp(tid, ip, &split, &btstack)))
fs/jfs/jfs_xtree.c
1729
p = xt_getpage(ip, bn, &mp);
fs/jfs/jfs_xtree.c
1745
p = xt_getpage(ip, bn, &mp);
fs/jfs/jfs_xtree.c
1749
BT_MARK_DIRTY(mp, ip);
fs/jfs/jfs_xtree.c
1750
if (!test_cflag(COMMIT_Nolink, ip)) {
fs/jfs/jfs_xtree.c
1751
tlck = txLock(tid, ip, mp, tlckXTREE|tlckGROW);
fs/jfs/jfs_xtree.c
1794
if (!test_cflag(COMMIT_Nolink, ip)) {
fs/jfs/jfs_xtree.c
180
XT_GETSEARCH(ip, btstack.top, bn, mp, p, index);
fs/jfs/jfs_xtree.c
1806
p = xt_getpage(ip, bn, &mp);
fs/jfs/jfs_xtree.c
1810
BT_MARK_DIRTY(mp, ip);
fs/jfs/jfs_xtree.c
1811
if (!test_cflag(COMMIT_Nolink, ip)) {
fs/jfs/jfs_xtree.c
1812
tlck = txLock(tid, ip, mp, tlckXTREE | tlckGROW);
fs/jfs/jfs_xtree.c
1830
if ((rc = xtSearch(ip, nxoff, NULL, &cmp, &btstack, XT_INSERT)))
fs/jfs/jfs_xtree.c
1834
XT_GETSEARCH(ip, btstack.top, bn, mp, p, index0);
fs/jfs/jfs_xtree.c
1838
jfs_error(ip->i_sb, "xtSearch failed\n");
fs/jfs/jfs_xtree.c
1844
jfs_error(ip->i_sb, "unexpected value of index\n");
fs/jfs/jfs_xtree.c
1878
if ((rc = xtSplitUp(tid, ip, &split, &btstack)))
fs/jfs/jfs_xtree.c
1882
p = xt_getpage(ip, bn, &mp);
fs/jfs/jfs_xtree.c
1899
p = xt_getpage(ip, bn, &mp);
fs/jfs/jfs_xtree.c
1903
BT_MARK_DIRTY(mp, ip);
fs/jfs/jfs_xtree.c
1904
if (!test_cflag(COMMIT_Nolink, ip)) {
fs/jfs/jfs_xtree.c
1905
tlck = txLock(tid, ip, mp, tlckXTREE|tlckGROW);
fs/jfs/jfs_xtree.c
1925
if (!test_cflag(COMMIT_Nolink, ip)) {
fs/jfs/jfs_xtree.c
1957
struct inode *ip, int xflag, s64 xoff, s32 maxblocks,
fs/jfs/jfs_xtree.c
1991
if ((rc = xtSearch(ip, xoff, &next, &cmp, &btstack, XT_INSERT)))
fs/jfs/jfs_xtree.c
1995
XT_GETSEARCH(ip, btstack.top, bn, mp, p, index);
fs/jfs/jfs_xtree.c
2027
nblocks = JFS_SBI(ip->i_sb)->nbperpage;
fs/jfs/jfs_xtree.c
2029
if ((rc = dbAllocBottomUp(ip, xaddr, (s64) nblocks)) == 0) {
fs/jfs/jfs_xtree.c
2048
if ((rc = dbAllocBottomUp(ip, xaddr, (s64) xlen)))
fs/jfs/jfs_xtree.c
2057
if ((rc = xtSplitUp(tid, ip, &split, &btstack))) {
fs/jfs/jfs_xtree.c
2059
dbFree(ip, *xaddrp, (s64) * xlenp);
fs/jfs/jfs_xtree.c
2075
if ((rc = dbAllocBottomUp(ip, xaddr, (s64) xlen)))
fs/jfs/jfs_xtree.c
2078
BT_MARK_DIRTY(mp, ip);
fs/jfs/jfs_xtree.c
2084
tlck = txLock(tid, ip, mp, tlckXTREE | tlckGROW);
fs/jfs/jfs_xtree.c
2114
void xtInitRoot(tid_t tid, struct inode *ip)
fs/jfs/jfs_xtree.c
2123
txLock(tid, ip, (struct metapage *) &JFS_IP(ip)->bxflag,
fs/jfs/jfs_xtree.c
2125
p = &JFS_IP(ip)->i_xtroot;
fs/jfs/jfs_xtree.c
2130
if (S_ISDIR(ip->i_mode))
fs/jfs/jfs_xtree.c
2134
ip->i_size = 0;
fs/jfs/jfs_xtree.c
2203
s64 xtTruncate(tid_t tid, struct inode *ip, s64 newsize, int flag)
fs/jfs/jfs_xtree.c
2268
teof = (newsize + (JFS_SBI(ip->i_sb)->bsize - 1)) >>
fs/jfs/jfs_xtree.c
2269
JFS_SBI(ip->i_sb)->l2bsize;
fs/jfs/jfs_xtree.c
2285
p = xt_getpage(ip, bn, &mp);
fs/jfs/jfs_xtree.c
2304
tlck = txLock(tid, ip, mp, tlckXTREE|tlckGROW);
fs/jfs/jfs_xtree.c
2305
BT_MARK_DIRTY(mp, ip);
fs/jfs/jfs_xtree.c
231
static int xtSearch(struct inode *ip, s64 xoff, s64 *nextp,
fs/jfs/jfs_xtree.c
2334
newsize = (xoff + xlen) << JFS_SBI(ip->i_sb)->l2bsize;
fs/jfs/jfs_xtree.c
2337
tlck = txLock(tid, ip, mp, tlckXTREE);
fs/jfs/jfs_xtree.c
234
struct jfs_inode_info *jfs_ip = JFS_IP(ip);
fs/jfs/jfs_xtree.c
2342
BT_MARK_DIRTY(mp, ip);
fs/jfs/jfs_xtree.c
2358
if (S_ISDIR(ip->i_mode) && (teof == 0))
fs/jfs/jfs_xtree.c
2359
invalidate_xad_metapages(ip, *xad);
fs/jfs/jfs_xtree.c
2423
txFreeMap(ip, pxdlock, NULL, COMMIT_WMAP);
fs/jfs/jfs_xtree.c
2451
txFreeMap(ip, (struct maplock *) & xadlock,
fs/jfs/jfs_xtree.c
2481
txFreeMap(ip, (struct maplock *) & xadlock, NULL, COMMIT_WMAP);
fs/jfs/jfs_xtree.c
2523
p = xt_getpage(ip, bn, &mp);
fs/jfs/jfs_xtree.c
2540
tlck = txLock(tid, ip, mp, tlckXTREE);
fs/jfs/jfs_xtree.c
2556
txFreeMap(ip, (struct maplock *) & xadlock,
fs/jfs/jfs_xtree.c
2559
BT_MARK_DIRTY(mp, ip);
fs/jfs/jfs_xtree.c
2616
tlck = txLock(tid, ip, mp, tlckXTREE);
fs/jfs/jfs_xtree.c
2628
txFreeMap(ip, (struct maplock *) & xadlock, NULL,
fs/jfs/jfs_xtree.c
2631
BT_MARK_DIRTY(mp, ip);
fs/jfs/jfs_xtree.c
2644
JFS_IP(ip)->mode2 |= INLINEEA;
fs/jfs/jfs_xtree.c
2692
jfs_error(ip->i_sb, "stack overrun!\n");
fs/jfs/jfs_xtree.c
270
p = xt_getpage(ip, bn, &mp);
fs/jfs/jfs_xtree.c
2717
if (S_ISDIR(ip->i_mode) && !newsize)
fs/jfs/jfs_xtree.c
2718
ip->i_size = 1; /* fsck hates zero-length directories */
fs/jfs/jfs_xtree.c
2720
ip->i_size = newsize;
fs/jfs/jfs_xtree.c
2723
dquot_free_block(ip, nfreed);
fs/jfs/jfs_xtree.c
2729
txFreelock(ip);
fs/jfs/jfs_xtree.c
2758
s64 xtTruncate_pmap(tid_t tid, struct inode *ip, s64 committed_size)
fs/jfs/jfs_xtree.c
2784
xoff = (committed_size >> JFS_SBI(ip->i_sb)->l2bsize) - 1;
fs/jfs/jfs_xtree.c
2785
rc = xtSearch(ip, xoff, NULL, &cmp, &btstack, 0);
fs/jfs/jfs_xtree.c
2789
XT_GETSEARCH(ip, btstack.top, bn, mp, p, index);
fs/jfs/jfs_xtree.c
2793
jfs_error(ip->i_sb, "did not find extent\n");
fs/jfs/jfs_xtree.c
2808
p = xt_getpage(ip, bn, &mp);
fs/jfs/jfs_xtree.c
2832
return (xoff + xlen) << JFS_SBI(ip->i_sb)->l2bsize;
fs/jfs/jfs_xtree.c
2834
tlck = txLock(tid, ip, mp, tlckXTREE);
fs/jfs/jfs_xtree.c
2853
p = xt_getpage(ip, bn, &mp);
fs/jfs/jfs_xtree.c
2867
tlck = txLock(tid, ip, mp, tlckXTREE);
fs/jfs/jfs_xtree.c
2892
jfs_error(ip->i_sb, "stack overrun!\n");
fs/jfs/jfs_xtree.c
494
jfs_error(ip->i_sb, "stack overrun!\n");
fs/jfs/jfs_xtree.c
529
struct inode *ip, int xflag, s64 xoff, s32 xlen, s64 * xaddrp,
fs/jfs/jfs_xtree.c
556
if ((rc = xtSearch(ip, xoff, &next, &cmp, &btstack, XT_INSERT)))
fs/jfs/jfs_xtree.c
560
XT_GETSEARCH(ip, btstack.top, bn, mp, p, index);
fs/jfs/jfs_xtree.c
580
if ((rc = dquot_alloc_block(ip, xlen)))
fs/jfs/jfs_xtree.c
582
if ((rc = dbAlloc(ip, hint, (s64) xlen, &xaddr))) {
fs/jfs/jfs_xtree.c
583
dquot_free_block(ip, xlen);
fs/jfs/jfs_xtree.c
608
if ((rc = xtSplitUp(tid, ip, &split, &btstack))) {
fs/jfs/jfs_xtree.c
611
dbFree(ip, xaddr, (s64) xlen);
fs/jfs/jfs_xtree.c
612
dquot_free_block(ip, xlen);
fs/jfs/jfs_xtree.c
629
BT_MARK_DIRTY(mp, ip);
fs/jfs/jfs_xtree.c
644
if (!test_cflag(COMMIT_Nolink, ip)) {
fs/jfs/jfs_xtree.c
645
tlck = txLock(tid, ip, mp, tlckXTREE | tlckGROW);
fs/jfs/jfs_xtree.c
680
struct inode *ip, struct xtsplit * split, struct btstack * btstack)
fs/jfs/jfs_xtree.c
703
sp = XT_PAGE(ip, smp);
fs/jfs/jfs_xtree.c
706
if ((sp->header.flag & BT_ROOT) && (!S_ISDIR(ip->i_mode)) &&
fs/jfs/jfs_xtree.c
708
(JFS_IP(ip)->mode2 & INLINEEA)) {
fs/jfs/jfs_xtree.c
710
JFS_IP(ip)->mode2 &= ~INLINEEA;
fs/jfs/jfs_xtree.c
712
BT_MARK_DIRTY(smp, ip);
fs/jfs/jfs_xtree.c
735
if (!test_cflag(COMMIT_Nolink, ip)) {
fs/jfs/jfs_xtree.c
736
tlck = txLock(tid, ip, smp, tlckXTREE | tlckGROW);
fs/jfs/jfs_xtree.c
758
xlen = JFS_SBI(ip->i_sb)->nbperpage;
fs/jfs/jfs_xtree.c
760
if ((rc = dbAlloc(ip, (s64) 0, (s64) xlen, &xaddr))
fs/jfs/jfs_xtree.c
785
xtSplitRoot(tid, ip, split, &rmp) :
fs/jfs/jfs_xtree.c
786
xtSplitPage(tid, ip, split, &rmp, &rbn);
fs/jfs/jfs_xtree.c
819
rcp = XT_PAGE(ip, rcmp);
fs/jfs/jfs_xtree.c
825
sp = xt_getpage(ip, parent->bn, &smp);
fs/jfs/jfs_xtree.c
84
static int xtSearch(struct inode *ip, s64 xoff, s64 *next, int *cmpp,
fs/jfs/jfs_xtree.c
850
split->len = JFS_SBI(ip->i_sb)->nbperpage;
fs/jfs/jfs_xtree.c
861
xtSplitRoot(tid, ip, split, &rmp) :
fs/jfs/jfs_xtree.c
862
xtSplitPage(tid, ip, split, &rmp, &rbn);
fs/jfs/jfs_xtree.c
88
struct inode *ip,
fs/jfs/jfs_xtree.c
884
BT_MARK_DIRTY(smp, ip);
fs/jfs/jfs_xtree.c
898
JFS_SBI(ip->i_sb)->nbperpage, rcbn);
fs/jfs/jfs_xtree.c
904
if (!test_cflag(COMMIT_Nolink, ip)) {
fs/jfs/jfs_xtree.c
905
tlck = txLock(tid, ip, smp,
fs/jfs/jfs_xtree.c
91
static int xtSplitPage(tid_t tid, struct inode *ip, struct xtsplit * split,
fs/jfs/jfs_xtree.c
94
static int xtSplitRoot(tid_t tid, struct inode *ip,
fs/jfs/jfs_xtree.c
949
xtSplitPage(tid_t tid, struct inode *ip,
fs/jfs/jfs_xtree.c
970
sp = XT_PAGE(ip, smp);
fs/jfs/jfs_xtree.c
980
rc = dquot_alloc_block(ip, lengthPXD(pxd));
fs/jfs/jfs_xtree.c
989
rmp = get_metapage(ip, rbn, PSIZE, 1);
fs/jfs/jfs_xtree.c
995
jfs_info("xtSplitPage: ip:0x%p smp:0x%p rmp:0x%p", ip, smp, rmp);
fs/jfs/jfs_xtree.c
997
BT_MARK_DIRTY(rmp, ip);
fs/jfs/jfs_xtree.h
100
extern int xtLookup(struct inode *ip, s64 lstart, s64 llen,
fs/jfs/jfs_xtree.h
102
extern void xtInitRoot(tid_t tid, struct inode *ip);
fs/jfs/jfs_xtree.h
103
extern int xtInsert(tid_t tid, struct inode *ip,
fs/jfs/jfs_xtree.h
105
extern int xtExtend(tid_t tid, struct inode *ip, s64 xoff, int xlen,
fs/jfs/jfs_xtree.h
107
extern int xtUpdate(tid_t tid, struct inode *ip, struct xad *nxad);
fs/jfs/jfs_xtree.h
108
extern s64 xtTruncate(tid_t tid, struct inode *ip, s64 newsize, int type);
fs/jfs/jfs_xtree.h
109
extern s64 xtTruncate_pmap(tid_t tid, struct inode *ip, s64 committed_size);
fs/jfs/jfs_xtree.h
111
struct inode *ip, int xflag, s64 xoff, int maxblocks,
fs/jfs/namei.c
1002
xtTruncate(tid, ip, 0, COMMIT_PWMAP);
fs/jfs/namei.c
102
mutex_lock_nested(&JFS_IP(ip)->commit_mutex, COMMIT_MUTEX_CHILD);
fs/jfs/namei.c
1020
ino = ip->i_ino;
fs/jfs/namei.c
1025
xtTruncate(tid, ip, 0, COMMIT_PWMAP);
fs/jfs/namei.c
1031
mark_inode_dirty(ip);
fs/jfs/namei.c
104
rc = jfs_init_acl(tid, ip, dip);
fs/jfs/namei.c
1040
iplist[1] = ip;
fs/jfs/namei.c
1045
mutex_unlock(&JFS_IP(ip)->commit_mutex);
fs/jfs/namei.c
1048
free_ea_wmap(ip);
fs/jfs/namei.c
1049
clear_nlink(ip);
fs/jfs/namei.c
1050
discard_new_inode(ip);
fs/jfs/namei.c
1052
d_instantiate_new(dentry, ip);
fs/jfs/namei.c
108
rc = jfs_init_security(tid, ip, dip, &dentry->d_name);
fs/jfs/namei.c
1197
tblk->u.ip = new_ip;
fs/jfs/namei.c
1208
tblk->u.ip = new_ip;
fs/jfs/namei.c
122
tblk->ino = ip->i_ino;
fs/jfs/namei.c
123
tblk->u.ixpxd = JFS_IP(ip)->ixpxd;
fs/jfs/namei.c
126
iplist[1] = ip;
fs/jfs/namei.c
131
xtInitRoot(tid, ip);
fs/jfs/namei.c
1365
struct inode *ip;
fs/jfs/namei.c
137
ino = ip->i_ino;
fs/jfs/namei.c
1380
ip = ialloc(dir, mode);
fs/jfs/namei.c
1381
if (IS_ERR(ip)) {
fs/jfs/namei.c
1382
rc = PTR_ERR(ip);
fs/jfs/namei.c
1385
jfs_ip = JFS_IP(ip);
fs/jfs/namei.c
1390
mutex_lock_nested(&JFS_IP(ip)->commit_mutex, COMMIT_MUTEX_CHILD);
fs/jfs/namei.c
1392
rc = jfs_init_acl(tid, ip, dir);
fs/jfs/namei.c
1396
rc = jfs_init_security(tid, ip, dir, &dentry->d_name);
fs/jfs/namei.c
1409
tblk->ino = ip->i_ino;
fs/jfs/namei.c
1410
tblk->u.ixpxd = JFS_IP(ip)->ixpxd;
fs/jfs/namei.c
1412
ino = ip->i_ino;
fs/jfs/namei.c
1418
ip->i_op = &jfs_file_inode_operations;
fs/jfs/namei.c
1420
init_special_inode(ip, ip->i_mode, rdev);
fs/jfs/namei.c
1422
mark_inode_dirty(ip);
fs/jfs/namei.c
1429
iplist[1] = ip;
fs/jfs/namei.c
1434
mutex_unlock(&JFS_IP(ip)->commit_mutex);
fs/jfs/namei.c
1437
free_ea_wmap(ip);
fs/jfs/namei.c
1438
clear_nlink(ip);
fs/jfs/namei.c
1439
discard_new_inode(ip);
fs/jfs/namei.c
1441
d_instantiate_new(dentry, ip);
fs/jfs/namei.c
1456
struct inode *ip;
fs/jfs/namei.c
1467
ip = NULL;
fs/jfs/namei.c
147
ip->i_op = &jfs_file_inode_operations;
fs/jfs/namei.c
1470
ip = ERR_PTR(rc);
fs/jfs/namei.c
1472
ip = jfs_iget(dip->i_sb, inum);
fs/jfs/namei.c
1473
if (IS_ERR(ip))
fs/jfs/namei.c
1477
return d_splice_alias(ip, dentry);
fs/jfs/namei.c
148
ip->i_fop = &jfs_file_operations;
fs/jfs/namei.c
149
ip->i_mapping->a_ops = &jfs_aops;
fs/jfs/namei.c
151
mark_inode_dirty(ip);
fs/jfs/namei.c
161
mutex_unlock(&JFS_IP(ip)->commit_mutex);
fs/jfs/namei.c
164
free_ea_wmap(ip);
fs/jfs/namei.c
165
clear_nlink(ip);
fs/jfs/namei.c
166
discard_new_inode(ip);
fs/jfs/namei.c
168
d_instantiate_new(dentry, ip);
fs/jfs/namei.c
201
struct inode *ip = NULL; /* child directory inode */
fs/jfs/namei.c
226
ip = ialloc(dip, S_IFDIR | mode);
fs/jfs/namei.c
227
if (IS_ERR(ip)) {
fs/jfs/namei.c
228
rc = PTR_ERR(ip);
fs/jfs/namei.c
235
mutex_lock_nested(&JFS_IP(ip)->commit_mutex, COMMIT_MUTEX_CHILD);
fs/jfs/namei.c
237
rc = jfs_init_acl(tid, ip, dip);
fs/jfs/namei.c
241
rc = jfs_init_security(tid, ip, dip, &dentry->d_name);
fs/jfs/namei.c
255
tblk->ino = ip->i_ino;
fs/jfs/namei.c
256
tblk->u.ixpxd = JFS_IP(ip)->ixpxd;
fs/jfs/namei.c
259
iplist[1] = ip;
fs/jfs/namei.c
264
dtInitRoot(tid, ip, dip->i_ino);
fs/jfs/namei.c
270
ino = ip->i_ino;
fs/jfs/namei.c
280
set_nlink(ip, 2); /* for '.' */
fs/jfs/namei.c
281
ip->i_op = &jfs_dir_inode_operations;
fs/jfs/namei.c
282
ip->i_fop = &jfs_dir_operations;
fs/jfs/namei.c
284
mark_inode_dirty(ip);
fs/jfs/namei.c
295
mutex_unlock(&JFS_IP(ip)->commit_mutex);
fs/jfs/namei.c
298
free_ea_wmap(ip);
fs/jfs/namei.c
299
clear_nlink(ip);
fs/jfs/namei.c
300
discard_new_inode(ip);
fs/jfs/namei.c
302
d_instantiate_new(dentry, ip);
fs/jfs/namei.c
338
struct inode *ip = d_inode(dentry);
fs/jfs/namei.c
350
rc = dquot_initialize(ip);
fs/jfs/namei.c
355
if (!dtEmpty(ip)) {
fs/jfs/namei.c
367
mutex_lock_nested(&JFS_IP(ip)->commit_mutex, COMMIT_MUTEX_CHILD);
fs/jfs/namei.c
370
iplist[1] = ip;
fs/jfs/namei.c
374
tblk->u.ip = ip;
fs/jfs/namei.c
379
ino = ip->i_ino;
fs/jfs/namei.c
385
mutex_unlock(&JFS_IP(ip)->commit_mutex);
fs/jfs/namei.c
401
if (JFS_IP(ip)->ea.flag & DXD_EXTENT) {
fs/jfs/namei.c
403
txEA(tid, ip, &JFS_IP(ip)->ea, NULL);
fs/jfs/namei.c
405
JFS_IP(ip)->ea.flag = 0;
fs/jfs/namei.c
408
if (JFS_IP(ip)->acl.flag & DXD_EXTENT) {
fs/jfs/namei.c
410
txEA(tid, ip, &JFS_IP(ip)->acl, NULL);
fs/jfs/namei.c
412
JFS_IP(ip)->acl.flag = 0;
fs/jfs/namei.c
415
clear_nlink(ip);
fs/jfs/namei.c
416
mark_inode_dirty(ip);
fs/jfs/namei.c
422
mutex_unlock(&JFS_IP(ip)->commit_mutex);
fs/jfs/namei.c
468
struct inode *ip = d_inode(dentry);
fs/jfs/namei.c
482
rc = dquot_initialize(ip);
fs/jfs/namei.c
489
IWRITE_LOCK(ip, RDWRLOCK_NORMAL);
fs/jfs/namei.c
494
mutex_lock_nested(&JFS_IP(ip)->commit_mutex, COMMIT_MUTEX_CHILD);
fs/jfs/namei.c
497
iplist[1] = ip;
fs/jfs/namei.c
502
ino = ip->i_ino;
fs/jfs/namei.c
508
mutex_unlock(&JFS_IP(ip)->commit_mutex);
fs/jfs/namei.c
510
IWRITE_UNLOCK(ip);
fs/jfs/namei.c
514
ASSERT(ip->i_nlink);
fs/jfs/namei.c
517
inode_set_ctime_to_ts(dip, inode_set_ctime_current(ip)));
fs/jfs/namei.c
521
inode_dec_link_count(ip);
fs/jfs/namei.c
526
if (ip->i_nlink == 0) {
fs/jfs/namei.c
527
assert(!test_cflag(COMMIT_Nolink, ip));
fs/jfs/namei.c
529
if ((new_size = commitZeroLink(tid, ip)) < 0) {
fs/jfs/namei.c
532
mutex_unlock(&JFS_IP(ip)->commit_mutex);
fs/jfs/namei.c
534
IWRITE_UNLOCK(ip);
fs/jfs/namei.c
540
tblk->u.ip = ip;
fs/jfs/namei.c
561
mutex_unlock(&JFS_IP(ip)->commit_mutex);
fs/jfs/namei.c
566
mutex_lock(&JFS_IP(ip)->commit_mutex);
fs/jfs/namei.c
567
new_size = xtTruncate_pmap(tid, ip, new_size);
fs/jfs/namei.c
574
mutex_unlock(&JFS_IP(ip)->commit_mutex);
fs/jfs/namei.c
577
if (ip->i_nlink == 0)
fs/jfs/namei.c
578
set_cflag(COMMIT_Nolink, ip);
fs/jfs/namei.c
580
IWRITE_UNLOCK(ip);
fs/jfs/namei.c
624
static s64 commitZeroLink(tid_t tid, struct inode *ip)
fs/jfs/namei.c
629
jfs_info("commitZeroLink: tid = %d, ip = 0x%p", tid, ip);
fs/jfs/namei.c
631
filetype = ip->i_mode & S_IFMT;
fs/jfs/namei.c
637
if (ip->i_size < IDATASIZE) {
fs/jfs/namei.c
638
ip->i_size = 0;
fs/jfs/namei.c
647
set_cflag(COMMIT_Freewmap, ip);
fs/jfs/namei.c
656
if (JFS_IP(ip)->ea.flag & DXD_EXTENT)
fs/jfs/namei.c
658
txEA(tid, ip, &JFS_IP(ip)->ea, NULL);
fs/jfs/namei.c
663
if (JFS_IP(ip)->acl.flag & DXD_EXTENT)
fs/jfs/namei.c
665
txEA(tid, ip, &JFS_IP(ip)->acl, NULL);
fs/jfs/namei.c
673
if (ip->i_size)
fs/jfs/namei.c
674
return xtTruncate_pmap(tid, ip, 0);
fs/jfs/namei.c
68
struct inode *ip = NULL; /* child directory inode */
fs/jfs/namei.c
690
void jfs_free_zero_link(struct inode *ip)
fs/jfs/namei.c
694
jfs_info("jfs_free_zero_link: ip = 0x%p", ip);
fs/jfs/namei.c
699
type = ip->i_mode & S_IFMT;
fs/jfs/namei.c
706
if (ip->i_size < IDATASIZE)
fs/jfs/namei.c
716
if (JFS_IP(ip)->ea.flag & DXD_EXTENT) {
fs/jfs/namei.c
717
s64 xaddr = addressDXD(&JFS_IP(ip)->ea);
fs/jfs/namei.c
718
int xlen = lengthDXD(&JFS_IP(ip)->ea);
fs/jfs/namei.c
723
invalidate_dxd_metapages(ip, JFS_IP(ip)->ea);
fs/jfs/namei.c
731
txFreeMap(ip, pxdlock, NULL, COMMIT_WMAP);
fs/jfs/namei.c
737
if (JFS_IP(ip)->acl.flag & DXD_EXTENT) {
fs/jfs/namei.c
738
s64 xaddr = addressDXD(&JFS_IP(ip)->acl);
fs/jfs/namei.c
739
int xlen = lengthDXD(&JFS_IP(ip)->acl);
fs/jfs/namei.c
743
invalidate_dxd_metapages(ip, JFS_IP(ip)->acl);
fs/jfs/namei.c
751
txFreeMap(ip, pxdlock, NULL, COMMIT_WMAP);
fs/jfs/namei.c
759
if (ip->i_size)
fs/jfs/namei.c
760
xtTruncate(0, ip, 0, COMMIT_WMAP);
fs/jfs/namei.c
792
struct inode *ip = d_inode(old_dentry);
fs/jfs/namei.c
804
if (isReadOnly(ip)) {
fs/jfs/namei.c
805
jfs_error(ip->i_sb, "read-only filesystem\n");
fs/jfs/namei.c
809
tid = txBegin(ip->i_sb, 0);
fs/jfs/namei.c
812
mutex_lock_nested(&JFS_IP(ip)->commit_mutex, COMMIT_MUTEX_CHILD);
fs/jfs/namei.c
826
ino = ip->i_ino;
fs/jfs/namei.c
831
inc_nlink(ip); /* for new link */
fs/jfs/namei.c
832
inode_set_ctime_current(ip);
fs/jfs/namei.c
835
ihold(ip);
fs/jfs/namei.c
837
iplist[0] = ip;
fs/jfs/namei.c
842
drop_nlink(ip); /* never instantiated */
fs/jfs/namei.c
843
iput(ip);
fs/jfs/namei.c
845
d_instantiate(dentry, ip);
fs/jfs/namei.c
853
mutex_unlock(&JFS_IP(ip)->commit_mutex);
fs/jfs/namei.c
888
struct inode *ip;
fs/jfs/namei.c
918
ip = ialloc(dip, S_IFLNK | 0777);
fs/jfs/namei.c
919
if (IS_ERR(ip)) {
fs/jfs/namei.c
920
rc = PTR_ERR(ip);
fs/jfs/namei.c
927
mutex_lock_nested(&JFS_IP(ip)->commit_mutex, COMMIT_MUTEX_CHILD);
fs/jfs/namei.c
929
rc = jfs_init_security(tid, ip, dip, &dentry->d_name);
fs/jfs/namei.c
93
ip = ialloc(dip, mode);
fs/jfs/namei.c
935
tblk->ino = ip->i_ino;
fs/jfs/namei.c
936
tblk->u.ixpxd = JFS_IP(ip)->ixpxd;
fs/jfs/namei.c
94
if (IS_ERR(ip)) {
fs/jfs/namei.c
942
ip->i_mode |= 0777;
fs/jfs/namei.c
947
xtInitRoot(tid, ip);
fs/jfs/namei.c
95
rc = PTR_ERR(ip);
fs/jfs/namei.c
954
ip->i_op = &jfs_fast_symlink_inode_operations;
fs/jfs/namei.c
956
ip->i_link = JFS_IP(ip)->i_inline_all;
fs/jfs/namei.c
957
memcpy(ip->i_link, name, ssize);
fs/jfs/namei.c
958
ip->i_size = ssize - 1;
fs/jfs/namei.c
964
if (ssize > sizeof (JFS_IP(ip)->i_inline))
fs/jfs/namei.c
965
JFS_IP(ip)->mode2 &= ~INLINEEA;
fs/jfs/namei.c
974
jfs_info("jfs_symlink: allocate extent ip:0x%p", ip);
fs/jfs/namei.c
976
ip->i_op = &jfs_symlink_inode_operations;
fs/jfs/namei.c
977
inode_nohighmem(ip);
fs/jfs/namei.c
978
ip->i_mapping->a_ops = &jfs_aops;
fs/jfs/namei.c
985
sb = ip->i_sb;
fs/jfs/namei.c
990
if ((rc = xtInsert(tid, ip, 0, 0, xlen, &xaddr, 0))) {
fs/jfs/namei.c
994
ip->i_size = ssize - 1;
fs/jfs/namei.c
999
mp = get_metapage(ip, xaddr, PSIZE, 1);
fs/jfs/xattr.c
143
static int ea_write_inline(struct inode *ip, struct jfs_ea_list *ealist,
fs/jfs/xattr.c
146
struct jfs_inode_info *ji = JFS_IP(ip);
fs/jfs/xattr.c
201
static int ea_write(struct inode *ip, struct jfs_ea_list *ealist, int size,
fs/jfs/xattr.c
204
struct super_block *sb = ip->i_sb;
fs/jfs/xattr.c
205
struct jfs_inode_info *ji = JFS_IP(ip);
fs/jfs/xattr.c
220
if (!ea_write_inline(ip, ealist, size, ea))
fs/jfs/xattr.c
228
rc = dquot_alloc_block(ip, nblocks);
fs/jfs/xattr.c
232
rc = dbAlloc(ip, INOHINT(ip), nblocks, &blkno);
fs/jfs/xattr.c
235
dquot_free_block(ip, nblocks);
fs/jfs/xattr.c
256
if (!(mp = get_metapage(ip, blkno + i, bytes_to_write, 1))) {
fs/jfs/xattr.c
300
dquot_free_block(ip, nblocks);
fs/jfs/xattr.c
302
dbFree(ip, blkno, nblocks);
fs/jfs/xattr.c
317
static int ea_read_inline(struct inode *ip, struct jfs_ea_list *ealist)
fs/jfs/xattr.c
319
struct jfs_inode_info *ji = JFS_IP(ip);
fs/jfs/xattr.c
351
static int ea_read(struct inode *ip, struct jfs_ea_list *ealist)
fs/jfs/xattr.c
353
struct super_block *sb = ip->i_sb;
fs/jfs/xattr.c
354
struct jfs_inode_info *ji = JFS_IP(ip);
fs/jfs/xattr.c
366
return ea_read_inline(ip, ealist);
fs/jfs/xattr.c
396
if (!(mp = read_metapage(ip, blkno + i, bytes_to_read, 1)))
fs/ntfs3/file.c
1051
for (ip = 0; ip < pages_per_frame; ip++) {
fs/ntfs3/file.c
1052
folio = page_folio(pages[ip]);
fs/ntfs3/file.c
1060
ip = off >> PAGE_SHIFT;
fs/ntfs3/file.c
1062
for (; ip < pages_per_frame; ip++, off = 0) {
fs/ntfs3/file.c
1063
folio = page_folio(pages[ip]);
fs/ntfs3/file.c
1073
for (ip = 0; ip < pages_per_frame; ip++) {
fs/ntfs3/file.c
1074
folio = page_folio(pages[ip]);
fs/ntfs3/file.c
1116
for (ip = 0; ip < pages_per_frame;
fs/ntfs3/file.c
1117
ip++) {
fs/ntfs3/file.c
1118
folio = page_folio(pages[ip]);
fs/ntfs3/file.c
1129
ip = off >> PAGE_SHIFT;
fs/ntfs3/file.c
1136
folio = page_folio(pages[ip]);
fs/ntfs3/file.c
1149
ip++;
fs/ntfs3/file.c
1158
for (ip = 0; ip < pages_per_frame; ip++) {
fs/ntfs3/file.c
1159
folio = page_folio(pages[ip]);
fs/ntfs3/file.c
994
u32 ip, off;
fs/ocfs2/dlmfs/dlmfs.c
121
struct dlmfs_inode_private *ip;
fs/ocfs2/dlmfs/dlmfs.c
144
ip = DLMFS_I(inode);
fs/ocfs2/dlmfs/dlmfs.c
146
status = user_dlm_cluster_lock(&ip->ip_lockres, level, flags);
fs/ocfs2/dlmfs/dlmfs.c
167
struct dlmfs_inode_private *ip = DLMFS_I(inode);
fs/ocfs2/dlmfs/dlmfs.c
178
user_dlm_cluster_unlock(&ip->ip_lockres, level);
fs/ocfs2/dlmfs/dlmfs.c
211
struct dlmfs_inode_private *ip = DLMFS_I(inode);
fs/ocfs2/dlmfs/dlmfs.c
213
poll_wait(file, &ip->ip_lockres.l_event, wait);
fs/ocfs2/dlmfs/dlmfs.c
215
spin_lock(&ip->ip_lockres.l_lock);
fs/ocfs2/dlmfs/dlmfs.c
216
if (ip->ip_lockres.l_flags & USER_LOCK_BLOCKED)
fs/ocfs2/dlmfs/dlmfs.c
218
spin_unlock(&ip->ip_lockres.l_lock);
fs/ocfs2/dlmfs/dlmfs.c
270
struct dlmfs_inode_private *ip =
fs/ocfs2/dlmfs/dlmfs.c
273
ip->ip_conn = NULL;
fs/ocfs2/dlmfs/dlmfs.c
274
ip->ip_parent = NULL;
fs/ocfs2/dlmfs/dlmfs.c
276
inode_init_once(&ip->ip_vfs_inode);
fs/ocfs2/dlmfs/dlmfs.c
281
struct dlmfs_inode_private *ip;
fs/ocfs2/dlmfs/dlmfs.c
283
ip = alloc_inode_sb(sb, dlmfs_inode_cache, GFP_NOFS);
fs/ocfs2/dlmfs/dlmfs.c
284
if (!ip)
fs/ocfs2/dlmfs/dlmfs.c
287
return &ip->ip_vfs_inode;
fs/ocfs2/dlmfs/dlmfs.c
298
struct dlmfs_inode_private *ip;
fs/ocfs2/dlmfs/dlmfs.c
306
ip = DLMFS_I(inode);
fs/ocfs2/dlmfs/dlmfs.c
307
lockres = &ip->ip_lockres;
fs/ocfs2/dlmfs/dlmfs.c
318
iput(ip->ip_parent);
fs/ocfs2/dlmfs/dlmfs.c
322
mlog(0, "we're a directory, ip->ip_conn = 0x%p\n", ip->ip_conn);
fs/ocfs2/dlmfs/dlmfs.c
325
if (ip->ip_conn)
fs/ocfs2/dlmfs/dlmfs.c
326
user_dlm_unregister(ip->ip_conn);
fs/ocfs2/dlmfs/dlmfs.c
328
ip->ip_parent = NULL;
fs/ocfs2/dlmfs/dlmfs.c
329
ip->ip_conn = NULL;
fs/ocfs2/dlmfs/dlmfs.c
356
struct dlmfs_inode_private *ip;
fs/ocfs2/dlmfs/dlmfs.c
365
ip = DLMFS_I(inode);
fs/ocfs2/dlmfs/dlmfs.c
366
ip->ip_conn = DLMFS_I(parent)->ip_conn;
fs/ocfs2/dlmfs/dlmfs.c
380
user_dlm_lock_res_init(&ip->ip_lockres, dentry);
fs/ocfs2/dlmfs/dlmfs.c
386
ip->ip_parent = igrab(parent);
fs/ocfs2/dlmfs/dlmfs.c
387
BUG_ON(!ip->ip_parent);
fs/ocfs2/dlmfs/dlmfs.c
413
struct dlmfs_inode_private *ip;
fs/ocfs2/dlmfs/dlmfs.c
432
ip = DLMFS_I(inode);
fs/ocfs2/dlmfs/dlmfs.c
441
ip->ip_conn = conn;
fs/ocfs2/dlmfs/userdlm.c
65
struct dlmfs_inode_private *ip;
fs/ocfs2/dlmfs/userdlm.c
67
ip = container_of(lockres,
fs/ocfs2/dlmfs/userdlm.c
70
return ip->ip_conn;
fs/ocfs2/dlmfs/userdlm.c
76
struct dlmfs_inode_private *ip;
fs/ocfs2/dlmfs/userdlm.c
78
ip = container_of(lockres,
fs/ocfs2/dlmfs/userdlm.c
81
return &ip->ip_vfs_inode;
fs/pstore/ftrace.c
26
static void notrace pstore_ftrace_call(unsigned long ip,
fs/pstore/ftrace.c
44
bit = ftrace_test_recursion_trylock(ip, parent_ip);
fs/pstore/ftrace.c
50
rec.ip = ip;
fs/pstore/inode.c
117
rec->ip, rec->parent_ip, (void *)rec->ip,
fs/smb/client/dns_resolve.c
25
char *ip;
fs/smb/client/dns_resolve.c
29
namelen, NULL, &ip, NULL, false);
fs/smb/client/dns_resolve.c
35
__func__, (int)namelen, (int)namelen, name, ip);
fs/smb/client/dns_resolve.c
37
rc = cifs_convert_address(addr, ip, strlen(ip));
fs/smb/client/dns_resolve.c
38
kfree(ip);
fs/xfs/libxfs/xfs_attr.c
1004
if (!xfs_is_metadir_inode(ip))
fs/xfs/libxfs/xfs_attr.c
1005
ASSERT(!XFS_NOT_DQATTACHED(mp, ip));
fs/xfs/libxfs/xfs_attr.c
1009
error = xfs_trans_alloc_inode(ip, &M_RES(mp)->tr_addafork, blks, 0,
fs/xfs/libxfs/xfs_attr.c
1014
if (xfs_inode_has_attr_fork(ip))
fs/xfs/libxfs/xfs_attr.c
1017
error = xfs_bmap_add_attrfork(tp, ip, size, rsvd);
fs/xfs/libxfs/xfs_attr.c
1022
xfs_iunlock(ip, XFS_ILOCK_EXCL);
fs/xfs/libxfs/xfs_attr.c
1027
xfs_iunlock(ip, XFS_ILOCK_EXCL);
fs/xfs/libxfs/xfs_attr.c
1038
const struct xfs_inode *ip)
fs/xfs/libxfs/xfs_attr.c
1040
return xfs_inode_has_attr_fork(ip) && xfs_attr_is_shortform(ip);
fs/xfs/libxfs/xfs_attr.c
67
struct xfs_inode *ip)
fs/xfs/libxfs/xfs_attr.c
69
if (!xfs_inode_has_attr_fork(ip))
fs/xfs/libxfs/xfs_attr.c
71
if (ip->i_af.if_format == XFS_DINODE_FMT_EXTENTS &&
fs/xfs/libxfs/xfs_attr.c
72
ip->i_af.if_nextents == 0)
fs/xfs/libxfs/xfs_attr.c
83
struct xfs_inode *ip)
fs/xfs/libxfs/xfs_attr.c
85
struct xfs_ifork *ifp = &ip->i_af;
fs/xfs/libxfs/xfs_attr.c
995
struct xfs_inode *ip, /* incore inode pointer */
fs/xfs/libxfs/xfs_attr.c
999
struct xfs_mount *mp = ip->i_mount;
fs/xfs/libxfs/xfs_attr.h
550
int xfs_inode_hasattr(struct xfs_inode *ip);
fs/xfs/libxfs/xfs_attr.h
551
bool xfs_attr_is_leaf(struct xfs_inode *ip);
fs/xfs/libxfs/xfs_attr.h
577
const struct xfs_inode *ip)
fs/xfs/libxfs/xfs_attr.h
579
return ip->i_af.if_format == XFS_DINODE_FMT_LOCAL ||
fs/xfs/libxfs/xfs_attr.h
580
(ip->i_af.if_format == XFS_DINODE_FMT_EXTENTS &&
fs/xfs/libxfs/xfs_attr.h
581
ip->i_af.if_nextents == 0);
fs/xfs/libxfs/xfs_attr.h
651
int xfs_attr_add_fork(struct xfs_inode *ip, int size, int rsvd);
fs/xfs/libxfs/xfs_attr_leaf.c
928
struct xfs_inode *ip,
fs/xfs/libxfs/xfs_attr_leaf.c
931
ASSERT(ip->i_af.if_nextents == 0);
fs/xfs/libxfs/xfs_attr_leaf.c
933
xfs_ifork_zap_attr(ip);
fs/xfs/libxfs/xfs_attr_leaf.c
934
ip->i_forkoff = 0;
fs/xfs/libxfs/xfs_attr_leaf.c
935
xfs_trans_log_inode(tp, ip, XFS_ILOG_CORE);
fs/xfs/libxfs/xfs_attr_leaf.h
59
void xfs_attr_fork_remove(struct xfs_inode *ip, struct xfs_trans *tp);
fs/xfs/libxfs/xfs_attr_remote.c
563
struct xfs_inode *ip,
fs/xfs/libxfs/xfs_attr_remote.c
567
struct xfs_mount *mp = ip->i_mount;
fs/xfs/libxfs/xfs_attr_remote.c
571
xfs_assert_ilocked(ip, XFS_ILOCK_EXCL);
fs/xfs/libxfs/xfs_attr_remote.c
575
xfs_bmap_mark_sick(ip, XFS_ATTR_FORK);
fs/xfs/libxfs/xfs_attr_remote.h
18
int xfs_attr_rmtval_stale(struct xfs_inode *ip, struct xfs_bmbt_irec *map,
fs/xfs/libxfs/xfs_bmap.c
1002
int default_size = xfs_default_attroffset(ip) >> 3;
fs/xfs/libxfs/xfs_bmap.c
1004
switch (ip->i_df.if_format) {
fs/xfs/libxfs/xfs_bmap.c
1006
ip->i_forkoff = default_size;
fs/xfs/libxfs/xfs_bmap.c
1011
ip->i_forkoff = xfs_attr_shortform_bytesfit(ip, size);
fs/xfs/libxfs/xfs_bmap.c
1012
if (!ip->i_forkoff)
fs/xfs/libxfs/xfs_bmap.c
1013
ip->i_forkoff = default_size;
fs/xfs/libxfs/xfs_bmap.c
1030
struct xfs_inode *ip, /* incore inode pointer */
fs/xfs/libxfs/xfs_bmap.c
1038
xfs_assert_ilocked(ip, XFS_ILOCK_EXCL);
fs/xfs/libxfs/xfs_bmap.c
1039
if (!xfs_is_metadir_inode(ip))
fs/xfs/libxfs/xfs_bmap.c
1040
ASSERT(!XFS_NOT_DQATTACHED(mp, ip));
fs/xfs/libxfs/xfs_bmap.c
1041
ASSERT(!xfs_inode_has_attr_fork(ip));
fs/xfs/libxfs/xfs_bmap.c
1043
xfs_trans_log_inode(tp, ip, XFS_ILOG_CORE);
fs/xfs/libxfs/xfs_bmap.c
1044
error = xfs_bmap_set_attrforkoff(ip, size);
fs/xfs/libxfs/xfs_bmap.c
1048
xfs_ifork_init_attr(ip, XFS_DINODE_FMT_EXTENTS, 0);
fs/xfs/libxfs/xfs_bmap.c
1050
switch (ip->i_df.if_format) {
fs/xfs/libxfs/xfs_bmap.c
1052
error = xfs_bmap_add_attrfork_local(tp, ip, &logflags);
fs/xfs/libxfs/xfs_bmap.c
1055
error = xfs_bmap_add_attrfork_extents(tp, ip, &logflags);
fs/xfs/libxfs/xfs_bmap.c
1058
error = xfs_bmap_add_attrfork_btree(tp, ip, &logflags);
fs/xfs/libxfs/xfs_bmap.c
1065
xfs_trans_log_inode(tp, ip, logflags);
fs/xfs/libxfs/xfs_bmap.c
1096
struct xfs_inode *ip,
fs/xfs/libxfs/xfs_bmap.c
1101
struct xfs_mount *mp = ip->i_mount;
fs/xfs/libxfs/xfs_bmap.c
1113
ip->i_ino, forkname, fa);
fs/xfs/libxfs/xfs_bmap.c
1131
struct xfs_inode *ip = cur->bc_ino.ip;
fs/xfs/libxfs/xfs_bmap.c
1138
struct xfs_ifork *ifp = xfs_ifork_ptr(ip, whichfork);
fs/xfs/libxfs/xfs_bmap.c
1145
xfs_warn(ip->i_mount, "corrupt dinode %llu, (btree extents).",
fs/xfs/libxfs/xfs_bmap.c
1146
(unsigned long long)ip->i_ino);
fs/xfs/libxfs/xfs_bmap.c
1147
xfs_inode_verifier_error(ip, -EFSCORRUPTED, __func__, block,
fs/xfs/libxfs/xfs_bmap.c
1149
xfs_bmap_mark_sick(ip, whichfork);
fs/xfs/libxfs/xfs_bmap.c
1160
fa = xfs_bmap_validate_extent(ip, whichfork, &new);
fs/xfs/libxfs/xfs_bmap.c
1162
xfs_inode_verifier_error(ip, -EFSCORRUPTED,
fs/xfs/libxfs/xfs_bmap.c
1165
xfs_bmap_mark_sick(ip, whichfork);
fs/xfs/libxfs/xfs_bmap.c
1166
return xfs_bmap_complain_bad_rec(ip, whichfork, fa,
fs/xfs/libxfs/xfs_bmap.c
1169
xfs_iext_insert(ip, &ir->icur, &new,
fs/xfs/libxfs/xfs_bmap.c
1171
trace_xfs_read_extent(ip, &ir->icur,
fs/xfs/libxfs/xfs_bmap.c
1185
struct xfs_inode *ip,
fs/xfs/libxfs/xfs_bmap.c
1189
struct xfs_ifork *ifp = xfs_ifork_ptr(ip, whichfork);
fs/xfs/libxfs/xfs_bmap.c
1190
struct xfs_mount *mp = ip->i_mount;
fs/xfs/libxfs/xfs_bmap.c
1197
xfs_assert_ilocked(ip, XFS_ILOCK_EXCL);
fs/xfs/libxfs/xfs_bmap.c
1201
cur = xfs_bmbt_init_cursor(mp, tp, ip, whichfork);
fs/xfs/libxfs/xfs_bmap.c
1209
xfs_bmap_mark_sick(ip, whichfork);
fs/xfs/libxfs/xfs_bmap.c
1223
xfs_bmap_mark_sick(ip, whichfork);
fs/xfs/libxfs/xfs_bmap.c
1237
struct xfs_inode *ip, /* incore inode */
fs/xfs/libxfs/xfs_bmap.c
1242
struct xfs_ifork *ifp = xfs_ifork_ptr(ip, whichfork);
fs/xfs/libxfs/xfs_bmap.c
1256
error = xfs_iread_extents(tp, ip, whichfork);
fs/xfs/libxfs/xfs_bmap.c
1285
struct xfs_inode *ip, /* incore inode */
fs/xfs/libxfs/xfs_bmap.c
1289
struct xfs_ifork *ifp = xfs_ifork_ptr(ip, whichfork);
fs/xfs/libxfs/xfs_bmap.c
1303
xfs_bmap_mark_sick(ip, whichfork);
fs/xfs/libxfs/xfs_bmap.c
1307
error = xfs_iread_extents(tp, ip, whichfork);
fs/xfs/libxfs/xfs_bmap.c
1311
if (!xfs_iext_lookup_extent_before(ip, ifp, last_block, &icur, &got))
fs/xfs/libxfs/xfs_bmap.c
1319
struct xfs_inode *ip,
fs/xfs/libxfs/xfs_bmap.c
1324
struct xfs_ifork *ifp = xfs_ifork_ptr(ip, whichfork);
fs/xfs/libxfs/xfs_bmap.c
1328
error = xfs_iread_extents(tp, ip, whichfork);
fs/xfs/libxfs/xfs_bmap.c
134
static inline bool xfs_bmap_needs_btree(struct xfs_inode *ip, int whichfork)
fs/xfs/libxfs/xfs_bmap.c
1359
error = xfs_bmap_last_extent(NULL, bma->ip, whichfork, &rec,
fs/xfs/libxfs/xfs_bmap.c
136
struct xfs_ifork *ifp = xfs_ifork_ptr(ip, whichfork);
fs/xfs/libxfs/xfs_bmap.c
1386
struct xfs_inode *ip,
fs/xfs/libxfs/xfs_bmap.c
1390
struct xfs_ifork *ifp = xfs_ifork_ptr(ip, whichfork);
fs/xfs/libxfs/xfs_bmap.c
140
ifp->if_nextents > XFS_IFORK_MAXEXT(ip, whichfork);
fs/xfs/libxfs/xfs_bmap.c
1400
if (XFS_IS_CORRUPT(ip->i_mount, !xfs_ifork_has_extents(ifp))) {
fs/xfs/libxfs/xfs_bmap.c
1401
xfs_bmap_mark_sick(ip, whichfork);
fs/xfs/libxfs/xfs_bmap.c
1405
error = xfs_bmap_last_extent(NULL, ip, whichfork, &rec, &is_empty);
fs/xfs/libxfs/xfs_bmap.c
1419
struct xfs_inode *ip,
fs/xfs/libxfs/xfs_bmap.c
1424
struct xfs_mount *mp = ip->i_mount;
fs/xfs/libxfs/xfs_bmap.c
1426
if (xfs_ifork_is_realtime(ip, whichfork) && xfs_has_rtgroups(mp)) {
fs/xfs/libxfs/xfs_bmap.c
1443
struct xfs_mount *mp = bma->ip->i_mount;
fs/xfs/libxfs/xfs_bmap.c
1444
struct xfs_ifork *ifp = xfs_ifork_ptr(bma->ip, whichfork);
fs/xfs/libxfs/xfs_bmap.c
146
static inline bool xfs_bmap_wants_extents(struct xfs_inode *ip, int whichfork)
fs/xfs/libxfs/xfs_bmap.c
148
struct xfs_ifork *ifp = xfs_ifork_ptr(ip, whichfork);
fs/xfs/libxfs/xfs_bmap.c
1505
xfs_bmap_same_rtgroup(bma->ip, whichfork, &LEFT, new))
fs/xfs/libxfs/xfs_bmap.c
152
ifp->if_nextents <= XFS_IFORK_MAXEXT(ip, whichfork);
fs/xfs/libxfs/xfs_bmap.c
1530
xfs_bmap_same_rtgroup(bma->ip, whichfork, new, &RIGHT))
fs/xfs/libxfs/xfs_bmap.c
1547
xfs_iext_remove(bma->ip, &bma->icur, state);
fs/xfs/libxfs/xfs_bmap.c
1548
xfs_iext_remove(bma->ip, &bma->icur, state);
fs/xfs/libxfs/xfs_bmap.c
1550
xfs_iext_update_extent(bma->ip, state, &bma->icur, &LEFT);
fs/xfs/libxfs/xfs_bmap.c
1596
xfs_iext_remove(bma->ip, &bma->icur, state);
fs/xfs/libxfs/xfs_bmap.c
1598
xfs_iext_update_extent(bma->ip, state, &bma->icur, &LEFT);
fs/xfs/libxfs/xfs_bmap.c
1631
xfs_iext_remove(bma->ip, &bma->icur, state);
fs/xfs/libxfs/xfs_bmap.c
1633
xfs_iext_update_extent(bma->ip, state, &bma->icur, &PREV);
fs/xfs/libxfs/xfs_bmap.c
1662
xfs_iext_update_extent(bma->ip, state, &bma->icur, &PREV);
fs/xfs/libxfs/xfs_bmap.c
1696
da_new = XFS_FILBLKS_MIN(xfs_bmap_worst_indlen(bma->ip, temp),
fs/xfs/libxfs/xfs_bmap.c
1705
xfs_iext_update_extent(bma->ip, state, &bma->icur, &PREV);
fs/xfs/libxfs/xfs_bmap.c
1707
xfs_iext_update_extent(bma->ip, state, &bma->icur, &LEFT);
fs/xfs/libxfs/xfs_bmap.c
1733
xfs_iext_update_extent(bma->ip, state, &bma->icur, new);
fs/xfs/libxfs/xfs_bmap.c
1758
if (xfs_bmap_needs_btree(bma->ip, whichfork)) {
fs/xfs/libxfs/xfs_bmap.c
1759
error = xfs_bmap_extents_to_btree(bma->tp, bma->ip,
fs/xfs/libxfs/xfs_bmap.c
176
struct xfs_inode *ip, /* incore inode pointer */
fs/xfs/libxfs/xfs_bmap.c
1767
da_new = XFS_FILBLKS_MIN(xfs_bmap_worst_indlen(bma->ip, temp),
fs/xfs/libxfs/xfs_bmap.c
1775
xfs_iext_insert(bma->ip, &bma->icur, &PREV, state);
fs/xfs/libxfs/xfs_bmap.c
179
struct xfs_mount *mp = ip->i_mount;
fs/xfs/libxfs/xfs_bmap.c
1807
da_new = XFS_FILBLKS_MIN(xfs_bmap_worst_indlen(bma->ip, temp),
fs/xfs/libxfs/xfs_bmap.c
1813
xfs_iext_update_extent(bma->ip, state, &bma->icur, &PREV);
fs/xfs/libxfs/xfs_bmap.c
1815
xfs_iext_update_extent(bma->ip, state, &bma->icur, &RIGHT);
fs/xfs/libxfs/xfs_bmap.c
1824
xfs_iext_update_extent(bma->ip, state, &bma->icur, new);
fs/xfs/libxfs/xfs_bmap.c
1849
if (xfs_bmap_needs_btree(bma->ip, whichfork)) {
fs/xfs/libxfs/xfs_bmap.c
1850
error = xfs_bmap_extents_to_btree(bma->tp, bma->ip,
fs/xfs/libxfs/xfs_bmap.c
1858
da_new = XFS_FILBLKS_MIN(xfs_bmap_worst_indlen(bma->ip, temp),
fs/xfs/libxfs/xfs_bmap.c
1864
xfs_iext_insert(bma->ip, &bma->icur, &PREV, state);
fs/xfs/libxfs/xfs_bmap.c
1901
nullstartblock(xfs_bmap_worst_indlen(bma->ip,
fs/xfs/libxfs/xfs_bmap.c
1907
nullstartblock(xfs_bmap_worst_indlen(bma->ip,
fs/xfs/libxfs/xfs_bmap.c
1909
xfs_iext_update_extent(bma->ip, state, &bma->icur, &PREV);
fs/xfs/libxfs/xfs_bmap.c
1912
xfs_iext_insert(bma->ip, &bma->icur, &RIGHT, state);
fs/xfs/libxfs/xfs_bmap.c
1913
xfs_iext_insert(bma->ip, &bma->icur, &LEFT, state);
fs/xfs/libxfs/xfs_bmap.c
1938
if (xfs_bmap_needs_btree(bma->ip, whichfork)) {
fs/xfs/libxfs/xfs_bmap.c
1939
error = xfs_bmap_extents_to_btree(bma->tp, bma->ip,
fs/xfs/libxfs/xfs_bmap.c
1965
xfs_rmap_map_extent(bma->tp, bma->ip, whichfork, new);
fs/xfs/libxfs/xfs_bmap.c
1968
if (xfs_bmap_needs_btree(bma->ip, whichfork)) {
fs/xfs/libxfs/xfs_bmap.c
1972
error = xfs_bmap_extents_to_btree(bma->tp, bma->ip,
fs/xfs/libxfs/xfs_bmap.c
1981
xfs_mod_delalloc(bma->ip, 0, (int64_t)da_new - da_old);
fs/xfs/libxfs/xfs_bmap.c
1994
xfs_bmap_check_leaf_extents(bma->cur, bma->ip, whichfork);
fs/xfs/libxfs/xfs_bmap.c
2010
xfs_inode_t *ip, /* incore inode pointer */
fs/xfs/libxfs/xfs_bmap.c
2026
struct xfs_mount *mp = ip->i_mount;
fs/xfs/libxfs/xfs_bmap.c
2032
ifp = xfs_ifork_ptr(ip, whichfork);
fs/xfs/libxfs/xfs_bmap.c
204
struct xfs_inode *ip)
fs/xfs/libxfs/xfs_bmap.c
206
if (ip->i_df.if_format == XFS_DINODE_FMT_DEV)
fs/xfs/libxfs/xfs_bmap.c
2076
xfs_bmap_same_rtgroup(ip, whichfork, &LEFT, new))
fs/xfs/libxfs/xfs_bmap.c
208
return M_IGEO(ip->i_mount)->attr_fork_offset;
fs/xfs/libxfs/xfs_bmap.c
2101
xfs_bmap_same_rtgroup(ip, whichfork, new, &RIGHT))
fs/xfs/libxfs/xfs_bmap.c
2117
xfs_iext_remove(ip, icur, state);
fs/xfs/libxfs/xfs_bmap.c
2118
xfs_iext_remove(ip, icur, state);
fs/xfs/libxfs/xfs_bmap.c
2120
xfs_iext_update_extent(ip, state, icur, &LEFT);
fs/xfs/libxfs/xfs_bmap.c
2175
xfs_iext_remove(ip, icur, state);
fs/xfs/libxfs/xfs_bmap.c
2177
xfs_iext_update_extent(ip, state, icur, &LEFT);
fs/xfs/libxfs/xfs_bmap.c
218
xfs_inode_t *ip,
fs/xfs/libxfs/xfs_bmap.c
222
ip->i_df.if_format != XFS_DINODE_FMT_DEV &&
fs/xfs/libxfs/xfs_bmap.c
2220
xfs_iext_remove(ip, icur, state);
fs/xfs/libxfs/xfs_bmap.c
2222
xfs_iext_update_extent(ip, state, icur, &PREV);
fs/xfs/libxfs/xfs_bmap.c
223
ip->i_df.if_format != XFS_DINODE_FMT_BTREE) {
fs/xfs/libxfs/xfs_bmap.c
224
uint dfl_forkoff = xfs_default_attroffset(ip) >> 3;
fs/xfs/libxfs/xfs_bmap.c
226
if (dfl_forkoff > ip->i_forkoff)
fs/xfs/libxfs/xfs_bmap.c
2264
xfs_iext_update_extent(ip, state, icur, &PREV);
fs/xfs/libxfs/xfs_bmap.c
227
ip->i_forkoff = dfl_forkoff;
fs/xfs/libxfs/xfs_bmap.c
2296
xfs_iext_update_extent(ip, state, icur, &PREV);
fs/xfs/libxfs/xfs_bmap.c
2298
xfs_iext_update_extent(ip, state, icur, &LEFT);
fs/xfs/libxfs/xfs_bmap.c
2334
xfs_iext_update_extent(ip, state, icur, &PREV);
fs/xfs/libxfs/xfs_bmap.c
2335
xfs_iext_insert(ip, icur, new, state);
fs/xfs/libxfs/xfs_bmap.c
2376
xfs_iext_update_extent(ip, state, icur, &PREV);
fs/xfs/libxfs/xfs_bmap.c
2378
xfs_iext_update_extent(ip, state, icur, &RIGHT);
fs/xfs/libxfs/xfs_bmap.c
2412
xfs_iext_update_extent(ip, state, icur, &PREV);
fs/xfs/libxfs/xfs_bmap.c
2414
xfs_iext_insert(ip, icur, new, state);
fs/xfs/libxfs/xfs_bmap.c
2466
xfs_iext_update_extent(ip, state, icur, &PREV);
fs/xfs/libxfs/xfs_bmap.c
2468
xfs_iext_insert(ip, icur, &r[1], state);
fs/xfs/libxfs/xfs_bmap.c
2469
xfs_iext_insert(ip, icur, &r[0], state);
fs/xfs/libxfs/xfs_bmap.c
2535
xfs_rmap_convert_extent(mp, tp, ip, whichfork, new);
fs/xfs/libxfs/xfs_bmap.c
2538
if (xfs_bmap_needs_btree(ip, whichfork)) {
fs/xfs/libxfs/xfs_bmap.c
2542
error = xfs_bmap_extents_to_btree(tp, ip, &cur, 0,
fs/xfs/libxfs/xfs_bmap.c
2555
xfs_bmap_check_leaf_extents(*curp, ip, whichfork);
fs/xfs/libxfs/xfs_bmap.c
2570
struct xfs_inode *ip,
fs/xfs/libxfs/xfs_bmap.c
2578
struct xfs_ifork *ifp = xfs_ifork_ptr(ip, whichfork);
fs/xfs/libxfs/xfs_bmap.c
2579
struct xfs_mount *mp = ip->i_mount;
fs/xfs/libxfs/xfs_bmap.c
2622
xfs_bmap_same_rtgroup(ip, whichfork, &left, new))
fs/xfs/libxfs/xfs_bmap.c
2633
xfs_bmap_same_rtgroup(ip, whichfork, new, &right))
fs/xfs/libxfs/xfs_bmap.c
2649
xfs_iext_remove(ip, icur, state);
fs/xfs/libxfs/xfs_bmap.c
2651
xfs_iext_update_extent(ip, state, icur, &left);
fs/xfs/libxfs/xfs_bmap.c
2698
xfs_iext_update_extent(ip, state, icur, &left);
fs/xfs/libxfs/xfs_bmap.c
2729
xfs_iext_update_extent(ip, state, icur, &right);
fs/xfs/libxfs/xfs_bmap.c
2755
xfs_iext_insert(ip, icur, new, state);
fs/xfs/libxfs/xfs_bmap.c
2784
xfs_rmap_map_extent(tp, ip, whichfork, new);
fs/xfs/libxfs/xfs_bmap.c
2787
if (xfs_bmap_needs_btree(ip, whichfork)) {
fs/xfs/libxfs/xfs_bmap.c
2791
error = xfs_bmap_extents_to_btree(tp, ip, curp, 0,
fs/xfs/libxfs/xfs_bmap.c
2803
xfs_bmap_check_leaf_extents(cur, ip, whichfork);
fs/xfs/libxfs/xfs_bmap.c
2996
struct xfs_mount *mp = ap->ip->i_mount;
fs/xfs/libxfs/xfs_bmap.c
2998
if (XFS_IS_REALTIME_INODE(ap->ip) &&
fs/xfs/libxfs/xfs_bmap.c
3240
bool isrt = XFS_IS_REALTIME_INODE(ap->ip) &&
fs/xfs/libxfs/xfs_bmap.c
3255
xfs_mod_delalloc(ap->ip, -(int64_t)ap->length, 0);
fs/xfs/libxfs/xfs_bmap.c
3267
ap->ip->i_delayed_blks += ap->length;
fs/xfs/libxfs/xfs_bmap.c
3268
xfs_trans_mod_dquot_byino(ap->tp, ap->ip, isrt ?
fs/xfs/libxfs/xfs_bmap.c
3275
ap->ip->i_nblocks += ap->length;
fs/xfs/libxfs/xfs_bmap.c
3276
xfs_trans_log_inode(ap->tp, ap->ip, XFS_ILOG_CORE);
fs/xfs/libxfs/xfs_bmap.c
3278
ap->ip->i_delayed_blks -= ap->length;
fs/xfs/libxfs/xfs_bmap.c
3279
xfs_mod_delalloc(ap->ip, -(int64_t)ap->length, 0);
fs/xfs/libxfs/xfs_bmap.c
3285
xfs_trans_mod_dquot_byino(ap->tp, ap->ip, fld, ap->length);
fs/xfs/libxfs/xfs_bmap.c
3304
align = xfs_get_cowextsz_hint(ap->ip);
fs/xfs/libxfs/xfs_bmap.c
3306
align = xfs_get_extsz_hint(ap->ip);
fs/xfs/libxfs/xfs_bmap.c
3389
ap->blkno = XFS_AGB_TO_FSB(ap->ip->i_mount, 0, 0);
fs/xfs/libxfs/xfs_bmap.c
343
xfs_inode_t *ip, /* incore inode pointer */
fs/xfs/libxfs/xfs_bmap.c
346
struct xfs_mount *mp = ip->i_mount;
fs/xfs/libxfs/xfs_bmap.c
347
struct xfs_ifork *ifp = xfs_ifork_ptr(ip, whichfork);
fs/xfs/libxfs/xfs_bmap.c
3593
ap->blkno = XFS_INO_TO_FSB(args->mp, ap->ip->i_ino);
fs/xfs/libxfs/xfs_bmap.c
3630
struct xfs_mount *mp = ap->ip->i_mount;
fs/xfs/libxfs/xfs_bmap.c
364
if (ip->i_df.if_nextents > 10000)
fs/xfs/libxfs/xfs_bmap.c
3660
xfs_inode_is_filestream(ap->ip))
fs/xfs/libxfs/xfs_bmap.c
3815
struct xfs_inode *ip,
fs/xfs/libxfs/xfs_bmap.c
3822
struct xfs_mount *mp = ip->i_mount;
fs/xfs/libxfs/xfs_bmap.c
3824
struct xfs_ifork *ifp = xfs_ifork_ptr(ip, whichfork);
fs/xfs/libxfs/xfs_bmap.c
3835
xfs_assert_ilocked(ip, XFS_ILOCK_SHARED | XFS_ILOCK_EXCL);
fs/xfs/libxfs/xfs_bmap.c
3838
xfs_bmap_mark_sick(ip, whichfork);
fs/xfs/libxfs/xfs_bmap.c
3844
xfs_bmap_mark_sick(ip, whichfork);
fs/xfs/libxfs/xfs_bmap.c
3853
error = xfs_iread_extents(NULL, ip, whichfork);
fs/xfs/libxfs/xfs_bmap.c
3857
if (!xfs_iext_lookup_extent(ip, ifp, bno, &icur, &got))
fs/xfs/libxfs/xfs_bmap.c
3900
struct xfs_mount *mp = bma->ip->i_mount;
fs/xfs/libxfs/xfs_bmap.c
3902
struct xfs_ifork *ifp = xfs_ifork_ptr(bma->ip, whichfork);
fs/xfs/libxfs/xfs_bmap.c
3935
XFS_IS_REALTIME_INODE(bma->ip))
fs/xfs/libxfs/xfs_bmap.c
3944
if (WARN_ON_ONCE(!xfs_valid_startblock(bma->ip, bma->blkno))) {
fs/xfs/libxfs/xfs_bmap.c
3945
xfs_bmap_mark_sick(bma->ip, whichfork);
fs/xfs/libxfs/xfs_bmap.c
3950
error = xfs_zero_extent(bma->ip, bma->blkno, bma->length);
fs/xfs/libxfs/xfs_bmap.c
3956
bma->cur = xfs_bmbt_init_cursor(mp, bma->tp, bma->ip, whichfork);
fs/xfs/libxfs/xfs_bmap.c
3977
error = xfs_bmap_add_extent_hole_real(bma->tp, bma->ip,
fs/xfs/libxfs/xfs_bmap.c
4006
struct xfs_ifork *ifp = xfs_ifork_ptr(bma->ip, whichfork);
fs/xfs/libxfs/xfs_bmap.c
4026
bma->cur = xfs_bmbt_init_cursor(bma->ip->i_mount, bma->tp,
fs/xfs/libxfs/xfs_bmap.c
4027
bma->ip, whichfork);
fs/xfs/libxfs/xfs_bmap.c
4037
error = xfs_zero_extent(bma->ip, mval->br_startblock,
fs/xfs/libxfs/xfs_bmap.c
4043
error = xfs_bmap_add_extent_unwritten_real(bma->tp, bma->ip, whichfork,
fs/xfs/libxfs/xfs_bmap.c
4080
struct xfs_inode *ip,
fs/xfs/libxfs/xfs_bmap.c
4083
struct xfs_ifork *ifp = xfs_ifork_ptr(ip, fork);
fs/xfs/libxfs/xfs_bmap.c
4104
struct xfs_ifork *ifp = xfs_ifork_ptr(bma->ip, whichfork);
fs/xfs/libxfs/xfs_bmap.c
4114
xfs_trans_log_inode(bma->tp, bma->ip, bma->logflags);
fs/xfs/libxfs/xfs_bmap.c
4137
struct xfs_inode *ip, /* incore inode */
fs/xfs/libxfs/xfs_bmap.c
4147
.ip = ip,
fs/xfs/libxfs/xfs_bmap.c
4150
struct xfs_mount *mp = ip->i_mount;
fs/xfs/libxfs/xfs_bmap.c
4152
struct xfs_ifork *ifp = xfs_ifork_ptr(ip, whichfork);
fs/xfs/libxfs/xfs_bmap.c
4178
xfs_assert_ilocked(ip, XFS_ILOCK_EXCL);
fs/xfs/libxfs/xfs_bmap.c
4195
xfs_bmap_mark_sick(ip, whichfork);
fs/xfs/libxfs/xfs_bmap.c
4204
error = xfs_iread_extents(tp, ip, whichfork);
fs/xfs/libxfs/xfs_bmap.c
4208
if (!xfs_iext_lookup_extent(ip, ifp, bno, &bma.icur, &bma.got))
fs/xfs/libxfs/xfs_bmap.c
4212
bma.minleft = xfs_bmapi_minleft(tp, ip, whichfork);
fs/xfs/libxfs/xfs_bmap.c
4285
XFS_IS_REALTIME_INODE(ip),
fs/xfs/libxfs/xfs_bmap.c
4317
error = xfs_bmap_btree_to_extents(tp, ip, bma.cur, &bma.logflags,
fs/xfs/libxfs/xfs_bmap.c
4323
ifp->if_nextents > XFS_IFORK_MAXEXT(ip, whichfork));
fs/xfs/libxfs/xfs_bmap.c
4356
struct xfs_inode *ip,
fs/xfs/libxfs/xfs_bmap.c
4362
struct xfs_ifork *ifp = xfs_ifork_ptr(ip, whichfork);
fs/xfs/libxfs/xfs_bmap.c
4363
struct xfs_mount *mp = ip->i_mount;
fs/xfs/libxfs/xfs_bmap.c
4382
xfs_ilock(ip, XFS_ILOCK_EXCL);
fs/xfs/libxfs/xfs_bmap.c
4383
xfs_trans_ijoin(tp, ip, 0);
fs/xfs/libxfs/xfs_bmap.c
4385
error = xfs_iext_count_extend(tp, ip, whichfork,
fs/xfs/libxfs/xfs_bmap.c
4390
if (!xfs_iext_lookup_extent(ip, ifp, offset_fsb, &bma.icur, &bma.got) ||
fs/xfs/libxfs/xfs_bmap.c
4407
xfs_bmbt_to_iomap(ip, iomap, &bma.got, 0, flags,
fs/xfs/libxfs/xfs_bmap.c
4408
xfs_iomap_inode_sequence(ip, flags));
fs/xfs/libxfs/xfs_bmap.c
4415
bma.ip = ip;
fs/xfs/libxfs/xfs_bmap.c
4417
bma.minleft = xfs_bmapi_minleft(tp, ip, whichfork);
fs/xfs/libxfs/xfs_bmap.c
4455
xfs_bmbt_to_iomap(ip, iomap, &bma.got, 0, flags,
fs/xfs/libxfs/xfs_bmap.c
4456
xfs_iomap_inode_sequence(ip, flags));
fs/xfs/libxfs/xfs_bmap.c
4461
xfs_refcount_alloc_cow_extent(tp, XFS_IS_REALTIME_INODE(ip),
fs/xfs/libxfs/xfs_bmap.c
4464
error = xfs_bmap_btree_to_extents(tp, ip, bma.cur, &bma.logflags,
fs/xfs/libxfs/xfs_bmap.c
4471
xfs_iunlock(ip, XFS_ILOCK_EXCL);
fs/xfs/libxfs/xfs_bmap.c
4478
xfs_iunlock(ip, XFS_ILOCK_EXCL);
fs/xfs/libxfs/xfs_bmap.c
4488
struct xfs_inode *ip,
fs/xfs/libxfs/xfs_bmap.c
4503
error = xfs_bmapi_convert_one_delalloc(ip, whichfork, offset,
fs/xfs/libxfs/xfs_bmap.c
4515
struct xfs_inode *ip,
fs/xfs/libxfs/xfs_bmap.c
4521
struct xfs_mount *mp = ip->i_mount;
fs/xfs/libxfs/xfs_bmap.c
4529
ifp = xfs_ifork_ptr(ip, whichfork);
fs/xfs/libxfs/xfs_bmap.c
4532
xfs_assert_ilocked(ip, XFS_ILOCK_EXCL);
fs/xfs/libxfs/xfs_bmap.c
4540
xfs_bmap_mark_sick(ip, whichfork);
fs/xfs/libxfs/xfs_bmap.c
4547
error = xfs_iread_extents(tp, ip, whichfork);
fs/xfs/libxfs/xfs_bmap.c
4551
if (xfs_iext_lookup_extent(ip, ifp, bno, &icur, &got)) {
fs/xfs/libxfs/xfs_bmap.c
4557
ip->i_nblocks += len;
fs/xfs/libxfs/xfs_bmap.c
4558
ip->i_delayed_blks -= len; /* see xfs_bmap_defer_add */
fs/xfs/libxfs/xfs_bmap.c
4559
xfs_trans_log_inode(tp, ip, XFS_ILOG_CORE);
fs/xfs/libxfs/xfs_bmap.c
4562
cur = xfs_bmbt_init_cursor(mp, tp, ip, whichfork);
fs/xfs/libxfs/xfs_bmap.c
4572
error = xfs_bmap_add_extent_hole_real(tp, ip, whichfork, &icur,
fs/xfs/libxfs/xfs_bmap.c
4577
error = xfs_bmap_btree_to_extents(tp, ip, cur, &logflags, whichfork);
fs/xfs/libxfs/xfs_bmap.c
4580
if (ip->i_df.if_format != XFS_DINODE_FMT_EXTENTS)
fs/xfs/libxfs/xfs_bmap.c
4582
else if (ip->i_df.if_format != XFS_DINODE_FMT_BTREE)
fs/xfs/libxfs/xfs_bmap.c
4586
xfs_trans_log_inode(tp, ip, logflags);
fs/xfs/libxfs/xfs_bmap.c
4663
struct xfs_inode *ip,
fs/xfs/libxfs/xfs_bmap.c
4670
struct xfs_mount *mp = ip->i_mount;
fs/xfs/libxfs/xfs_bmap.c
4671
struct xfs_ifork *ifp = xfs_ifork_ptr(ip, whichfork);
fs/xfs/libxfs/xfs_bmap.c
4682
isrt = xfs_ifork_is_realtime(ip, whichfork);
fs/xfs/libxfs/xfs_bmap.c
4697
xfs_quota_unreserve_blkres(ip, del->br_blockcount);
fs/xfs/libxfs/xfs_bmap.c
4698
ip->i_delayed_blks -= del->br_blockcount;
fs/xfs/libxfs/xfs_bmap.c
4710
xfs_iext_remove(ip, icur, state);
fs/xfs/libxfs/xfs_bmap.c
4719
da_new = XFS_FILBLKS_MIN(xfs_bmap_worst_indlen(ip,
fs/xfs/libxfs/xfs_bmap.c
4722
xfs_iext_update_extent(ip, state, icur, got);
fs/xfs/libxfs/xfs_bmap.c
4729
da_new = XFS_FILBLKS_MIN(xfs_bmap_worst_indlen(ip,
fs/xfs/libxfs/xfs_bmap.c
4732
xfs_iext_update_extent(ip, state, icur, got);
fs/xfs/libxfs/xfs_bmap.c
4745
got_indlen = xfs_bmap_worst_indlen(ip, got->br_blockcount);
fs/xfs/libxfs/xfs_bmap.c
4748
new_indlen = xfs_bmap_worst_indlen(ip, new.br_blockcount);
fs/xfs/libxfs/xfs_bmap.c
4776
xfs_iext_update_extent(ip, state, icur, got);
fs/xfs/libxfs/xfs_bmap.c
4778
xfs_iext_insert(ip, icur, &new, state);
fs/xfs/libxfs/xfs_bmap.c
4794
if (xfs_is_zoned_inode(ip))
fs/xfs/libxfs/xfs_bmap.c
4802
xfs_mod_delalloc(ip, -(int64_t)del->br_blockcount, -da_diff);
fs/xfs/libxfs/xfs_bmap.c
4807
struct xfs_inode *ip,
fs/xfs/libxfs/xfs_bmap.c
4812
struct xfs_mount *mp = ip->i_mount;
fs/xfs/libxfs/xfs_bmap.c
4813
struct xfs_ifork *ifp = xfs_ifork_ptr(ip, XFS_COW_FORK);
fs/xfs/libxfs/xfs_bmap.c
4838
xfs_iext_remove(ip, icur, state);
fs/xfs/libxfs/xfs_bmap.c
4848
xfs_iext_update_extent(ip, state, icur, got);
fs/xfs/libxfs/xfs_bmap.c
4855
xfs_iext_update_extent(ip, state, icur, got);
fs/xfs/libxfs/xfs_bmap.c
4868
xfs_iext_update_extent(ip, state, icur, got);
fs/xfs/libxfs/xfs_bmap.c
4870
xfs_iext_insert(ip, icur, &new, state);
fs/xfs/libxfs/xfs_bmap.c
4873
ip->i_delayed_blks -= del->br_blockcount;
fs/xfs/libxfs/xfs_bmap.c
4910
xfs_inode_t *ip, /* incore inode pointer */
fs/xfs/libxfs/xfs_bmap.c
4936
mp = ip->i_mount;
fs/xfs/libxfs/xfs_bmap.c
4939
ifp = xfs_ifork_ptr(ip, whichfork);
fs/xfs/libxfs/xfs_bmap.c
4959
ifp->if_nextents >= XFS_IFORK_MAXEXT(ip, whichfork) &&
fs/xfs/libxfs/xfs_bmap.c
4964
if (xfs_ifork_is_realtime(ip, whichfork))
fs/xfs/libxfs/xfs_bmap.c
4991
xfs_iext_remove(ip, icur, state);
fs/xfs/libxfs/xfs_bmap.c
5014
xfs_iext_update_extent(ip, state, icur, &got);
fs/xfs/libxfs/xfs_bmap.c
5028
xfs_iext_update_extent(ip, state, icur, &got);
fs/xfs/libxfs/xfs_bmap.c
5045
xfs_iext_update_extent(ip, state, icur, &got);
fs/xfs/libxfs/xfs_bmap.c
5092
xfs_iext_update_extent(ip, state, icur, &old);
fs/xfs/libxfs/xfs_bmap.c
5105
xfs_iext_insert(ip, icur, &new, state);
fs/xfs/libxfs/xfs_bmap.c
5110
xfs_rmap_unmap_extent(tp, ip, whichfork, del);
fs/xfs/libxfs/xfs_bmap.c
5116
bool isrt = xfs_ifork_is_realtime(ip, whichfork);
fs/xfs/libxfs/xfs_bmap.c
5118
if (xfs_is_reflink_inode(ip) && whichfork == XFS_DATA_FORK) {
fs/xfs/libxfs/xfs_bmap.c
5154
ip->i_nblocks -= nblks;
fs/xfs/libxfs/xfs_bmap.c
5159
xfs_trans_mod_dquot_byino(tp, ip, qfield, (long)-nblks);
fs/xfs/libxfs/xfs_bmap.c
5173
struct xfs_inode *ip, /* incore inode */
fs/xfs/libxfs/xfs_bmap.c
5188
struct xfs_mount *mp = ip->i_mount;
fs/xfs/libxfs/xfs_bmap.c
5197
trace_xfs_bunmap(ip, start, len, flags, _RET_IP_);
fs/xfs/libxfs/xfs_bmap.c
5201
ifp = xfs_ifork_ptr(ip, whichfork);
fs/xfs/libxfs/xfs_bmap.c
5203
xfs_bmap_mark_sick(ip, whichfork);
fs/xfs/libxfs/xfs_bmap.c
5209
xfs_assert_ilocked(ip, XFS_ILOCK_EXCL);
fs/xfs/libxfs/xfs_bmap.c
5213
error = xfs_iread_extents(tp, ip, whichfork);
fs/xfs/libxfs/xfs_bmap.c
5222
isrt = xfs_ifork_is_realtime(ip, whichfork);
fs/xfs/libxfs/xfs_bmap.c
5225
if (!xfs_iext_lookup_extent_before(ip, ifp, &end, &icur, &got)) {
fs/xfs/libxfs/xfs_bmap.c
5234
cur = xfs_bmbt_init_cursor(mp, tp, ip, whichfork);
fs/xfs/libxfs/xfs_bmap.c
5318
error = xfs_bmap_add_extent_unwritten_real(tp, ip,
fs/xfs/libxfs/xfs_bmap.c
5379
ip, whichfork, &icur, &cur,
fs/xfs/libxfs/xfs_bmap.c
5388
ip, whichfork, &icur, &cur,
fs/xfs/libxfs/xfs_bmap.c
5398
xfs_bmap_del_extent_delay(ip, whichfork, &icur, &got,
fs/xfs/libxfs/xfs_bmap.c
5401
error = xfs_bmap_del_extent_real(ip, tp, &icur, cur,
fs/xfs/libxfs/xfs_bmap.c
5432
if (xfs_bmap_needs_btree(ip, whichfork)) {
fs/xfs/libxfs/xfs_bmap.c
5434
error = xfs_bmap_extents_to_btree(tp, ip, &cur, 0,
fs/xfs/libxfs/xfs_bmap.c
5438
error = xfs_bmap_btree_to_extents(tp, ip, cur, &logflags,
fs/xfs/libxfs/xfs_bmap.c
544
#define xfs_bmap_check_leaf_extents(cur, ip, whichfork) do { } while (0)
fs/xfs/libxfs/xfs_bmap.c
5458
xfs_trans_log_inode(tp, ip, logflags);
fs/xfs/libxfs/xfs_bmap.c
5471
struct xfs_inode *ip,
fs/xfs/libxfs/xfs_bmap.c
5480
error = __xfs_bunmapi(tp, ip, bno, &len, flags, nexts);
fs/xfs/libxfs/xfs_bmap.c
5491
struct xfs_inode *ip,
fs/xfs/libxfs/xfs_bmap.c
5509
!xfs_bmap_same_rtgroup(ip, whichfork, left, got))
fs/xfs/libxfs/xfs_bmap.c
5527
struct xfs_inode *ip,
fs/xfs/libxfs/xfs_bmap.c
5536
struct xfs_ifork *ifp = xfs_ifork_ptr(ip, whichfork);
fs/xfs/libxfs/xfs_bmap.c
5540
struct xfs_mount *mp = ip->i_mount;
fs/xfs/libxfs/xfs_bmap.c
5544
xfs_assert_ilocked(ip, XFS_IOLOCK_EXCL | XFS_ILOCK_EXCL);
fs/xfs/libxfs/xfs_bmap.c
5545
ASSERT(xfs_bmse_can_merge(ip, whichfork, left, got, shift));
fs/xfs/libxfs/xfs_bmap.c
5592
error = xfs_bmap_btree_to_extents(tp, ip, cur, logflags, whichfork);
fs/xfs/libxfs/xfs_bmap.c
5597
xfs_iext_remove(ip, icur, 0);
fs/xfs/libxfs/xfs_bmap.c
5599
xfs_iext_update_extent(ip, xfs_bmap_fork_to_state(whichfork), icur,
fs/xfs/libxfs/xfs_bmap.c
5603
xfs_rmap_unmap_extent(tp, ip, whichfork, got);
fs/xfs/libxfs/xfs_bmap.c
5606
xfs_rmap_map_extent(tp, ip, whichfork, &new);
fs/xfs/libxfs/xfs_bmap.c
5613
struct xfs_inode *ip,
fs/xfs/libxfs/xfs_bmap.c
562
struct xfs_inode *ip, /* incore inode pointer */
fs/xfs/libxfs/xfs_bmap.c
5621
struct xfs_mount *mp = ip->i_mount;
fs/xfs/libxfs/xfs_bmap.c
5645
xfs_iext_update_extent(ip, xfs_bmap_fork_to_state(whichfork), icur,
fs/xfs/libxfs/xfs_bmap.c
5649
xfs_rmap_unmap_extent(tp, ip, whichfork, &prev);
fs/xfs/libxfs/xfs_bmap.c
5650
xfs_rmap_map_extent(tp, ip, whichfork, got);
fs/xfs/libxfs/xfs_bmap.c
5657
struct xfs_inode *ip,
fs/xfs/libxfs/xfs_bmap.c
5663
struct xfs_mount *mp = ip->i_mount;
fs/xfs/libxfs/xfs_bmap.c
5664
struct xfs_ifork *ifp = xfs_ifork_ptr(ip, whichfork);
fs/xfs/libxfs/xfs_bmap.c
567
struct xfs_ifork *ifp = xfs_ifork_ptr(ip, whichfork);
fs/xfs/libxfs/xfs_bmap.c
5674
xfs_bmap_mark_sick(ip, whichfork);
fs/xfs/libxfs/xfs_bmap.c
568
struct xfs_mount *mp = ip->i_mount;
fs/xfs/libxfs/xfs_bmap.c
5681
xfs_assert_ilocked(ip, XFS_IOLOCK_EXCL | XFS_ILOCK_EXCL);
fs/xfs/libxfs/xfs_bmap.c
5683
error = xfs_iread_extents(tp, ip, whichfork);
fs/xfs/libxfs/xfs_bmap.c
5688
cur = xfs_bmbt_init_cursor(mp, tp, ip, whichfork);
fs/xfs/libxfs/xfs_bmap.c
5690
if (!xfs_iext_lookup_extent(ip, ifp, *next_fsb, &icur, &got)) {
fs/xfs/libxfs/xfs_bmap.c
5695
xfs_bmap_mark_sick(ip, whichfork);
fs/xfs/libxfs/xfs_bmap.c
5707
if (xfs_bmse_can_merge(ip, whichfork, &prev, &got,
fs/xfs/libxfs/xfs_bmap.c
5709
error = xfs_bmse_merge(tp, ip, whichfork,
fs/xfs/libxfs/xfs_bmap.c
5723
error = xfs_bmap_shift_update_extent(tp, ip, whichfork, &icur, &got,
fs/xfs/libxfs/xfs_bmap.c
5739
xfs_trans_log_inode(tp, ip, logflags);
fs/xfs/libxfs/xfs_bmap.c
5746
struct xfs_inode *ip,
fs/xfs/libxfs/xfs_bmap.c
5754
xfs_assert_ilocked(ip, XFS_IOLOCK_EXCL);
fs/xfs/libxfs/xfs_bmap.c
5756
if (xfs_is_shutdown(ip->i_mount))
fs/xfs/libxfs/xfs_bmap.c
5759
xfs_ilock(ip, XFS_ILOCK_EXCL);
fs/xfs/libxfs/xfs_bmap.c
5760
error = xfs_bmap_last_extent(NULL, ip, XFS_DATA_FORK, &got, &is_empty);
fs/xfs/libxfs/xfs_bmap.c
5764
xfs_iunlock(ip, XFS_ILOCK_EXCL);
fs/xfs/libxfs/xfs_bmap.c
5772
struct xfs_inode *ip,
fs/xfs/libxfs/xfs_bmap.c
5779
struct xfs_mount *mp = ip->i_mount;
fs/xfs/libxfs/xfs_bmap.c
578
if (!xfs_bmap_wants_extents(ip, whichfork))
fs/xfs/libxfs/xfs_bmap.c
5780
struct xfs_ifork *ifp = xfs_ifork_ptr(ip, whichfork);
fs/xfs/libxfs/xfs_bmap.c
5790
xfs_bmap_mark_sick(ip, whichfork);
fs/xfs/libxfs/xfs_bmap.c
5797
xfs_assert_ilocked(ip, XFS_IOLOCK_EXCL | XFS_ILOCK_EXCL);
fs/xfs/libxfs/xfs_bmap.c
5799
error = xfs_iread_extents(tp, ip, whichfork);
fs/xfs/libxfs/xfs_bmap.c
5804
cur = xfs_bmbt_init_cursor(mp, tp, ip, whichfork);
fs/xfs/libxfs/xfs_bmap.c
5814
if (!xfs_iext_lookup_extent(ip, ifp, *next_fsb, &icur, &got)) {
fs/xfs/libxfs/xfs_bmap.c
5820
xfs_bmap_mark_sick(ip, whichfork);
fs/xfs/libxfs/xfs_bmap.c
5826
xfs_bmap_mark_sick(ip, whichfork);
fs/xfs/libxfs/xfs_bmap.c
5844
if (xfs_bmse_can_merge(ip, whichfork, &got, &next,
fs/xfs/libxfs/xfs_bmap.c
5849
error = xfs_bmap_shift_update_extent(tp, ip, whichfork, &icur, &got,
fs/xfs/libxfs/xfs_bmap.c
5865
xfs_trans_log_inode(tp, ip, logflags);
fs/xfs/libxfs/xfs_bmap.c
5878
struct xfs_inode *ip,
fs/xfs/libxfs/xfs_bmap.c
5882
struct xfs_ifork *ifp = xfs_ifork_ptr(ip, whichfork);
fs/xfs/libxfs/xfs_bmap.c
5886
struct xfs_mount *mp = ip->i_mount;
fs/xfs/libxfs/xfs_bmap.c
5895
xfs_bmap_mark_sick(ip, whichfork);
fs/xfs/libxfs/xfs_bmap.c
5903
error = xfs_iread_extents(tp, ip, whichfork);
fs/xfs/libxfs/xfs_bmap.c
5910
if (!xfs_iext_lookup_extent(ip, ifp, split_fsb, &icur, &got) ||
fs/xfs/libxfs/xfs_bmap.c
5921
cur = xfs_bmbt_init_cursor(mp, tp, ip, whichfork);
fs/xfs/libxfs/xfs_bmap.c
5933
xfs_iext_update_extent(ip, xfs_bmap_fork_to_state(whichfork), &icur,
fs/xfs/libxfs/xfs_bmap.c
5946
xfs_iext_insert(ip, &icur, &new, 0);
fs/xfs/libxfs/xfs_bmap.c
5971
if (xfs_bmap_needs_btree(ip, whichfork)) {
fs/xfs/libxfs/xfs_bmap.c
5975
error = xfs_bmap_extents_to_btree(tp, ip, &cur, 0,
fs/xfs/libxfs/xfs_bmap.c
5987
xfs_trans_log_inode(tp, ip, logflags);
fs/xfs/libxfs/xfs_bmap.c
5996
struct xfs_inode *ip,
fs/xfs/libxfs/xfs_bmap.c
6010
bi->bi_owner = ip;
fs/xfs/libxfs/xfs_bmap.c
6021
struct xfs_inode *ip,
fs/xfs/libxfs/xfs_bmap.c
6025
__xfs_bmap_add(tp, XFS_BMAP_MAP, ip, whichfork, PREV);
fs/xfs/libxfs/xfs_bmap.c
6032
struct xfs_inode *ip,
fs/xfs/libxfs/xfs_bmap.c
6036
__xfs_bmap_add(tp, XFS_BMAP_UNMAP, ip, whichfork, PREV);
fs/xfs/libxfs/xfs_bmap.c
605
xfs_rmap_ino_bmbt_owner(&oinfo, ip->i_ino, whichfork);
fs/xfs/libxfs/xfs_bmap.c
611
ip->i_nblocks--;
fs/xfs/libxfs/xfs_bmap.c
612
xfs_trans_mod_dquot_byino(tp, ip, XFS_TRANS_DQ_BCOUNT, -1L);
fs/xfs/libxfs/xfs_bmap.c
6130
struct xfs_inode *ip,
fs/xfs/libxfs/xfs_bmap.c
6134
return xfs_bmap_validate_extent_raw(ip->i_mount,
fs/xfs/libxfs/xfs_bmap.c
6135
XFS_IS_REALTIME_INODE(ip), whichfork, irec);
fs/xfs/libxfs/xfs_bmap.c
6151
struct xfs_inode *ip,
fs/xfs/libxfs/xfs_bmap.c
6159
xfs_assert_ilocked(ip, XFS_ILOCK_EXCL);
fs/xfs/libxfs/xfs_bmap.c
616
xfs_bmap_broot_realloc(ip, whichfork, 0);
fs/xfs/libxfs/xfs_bmap.c
6163
error = __xfs_bunmapi(*tpp, ip, startoff, &unmap_len, flags,
fs/xfs/libxfs/xfs_bmap.c
6195
fa = xfs_bmap_validate_extent(cur->bc_ino.ip, cur->bc_ino.whichfork,
fs/xfs/libxfs/xfs_bmap.c
6199
return xfs_bmap_complain_bad_rec(cur->bc_ino.ip,
fs/xfs/libxfs/xfs_bmap.c
6224
struct xfs_inode *ip)
fs/xfs/libxfs/xfs_bmap.c
6230
if (!xfs_is_always_cow_inode(ip) &&
fs/xfs/libxfs/xfs_bmap.c
6231
(ip->i_diflags & XFS_DIFLAG_EXTSIZE) && ip->i_extsize)
fs/xfs/libxfs/xfs_bmap.c
6232
return ip->i_extsize;
fs/xfs/libxfs/xfs_bmap.c
6233
if (XFS_IS_REALTIME_INODE(ip) &&
fs/xfs/libxfs/xfs_bmap.c
6234
ip->i_mount->m_sb.sb_rextsize > 1)
fs/xfs/libxfs/xfs_bmap.c
6235
return ip->i_mount->m_sb.sb_rextsize;
fs/xfs/libxfs/xfs_bmap.c
6247
struct xfs_inode *ip)
fs/xfs/libxfs/xfs_bmap.c
6252
if (ip->i_diflags2 & XFS_DIFLAG2_COWEXTSIZE)
fs/xfs/libxfs/xfs_bmap.c
6253
a = ip->i_cowextsize;
fs/xfs/libxfs/xfs_bmap.c
6254
if (XFS_IS_REALTIME_INODE(ip)) {
fs/xfs/libxfs/xfs_bmap.c
6256
if (ip->i_diflags & XFS_DIFLAG_EXTSIZE)
fs/xfs/libxfs/xfs_bmap.c
6257
b = ip->i_extsize;
fs/xfs/libxfs/xfs_bmap.c
6259
b = xfs_get_extsz_hint(ip);
fs/xfs/libxfs/xfs_bmap.c
630
struct xfs_inode *ip, /* incore inode pointer */
fs/xfs/libxfs/xfs_bmap.c
651
mp = ip->i_mount;
fs/xfs/libxfs/xfs_bmap.c
653
ifp = xfs_ifork_ptr(ip, whichfork);
fs/xfs/libxfs/xfs_bmap.c
660
block = xfs_bmap_broot_realloc(ip, whichfork, 1);
fs/xfs/libxfs/xfs_bmap.c
665
xfs_bmbt_init_block(ip, block, NULL, 1, 1);
fs/xfs/libxfs/xfs_bmap.c
669
cur = xfs_bmbt_init_cursor(mp, tp, ip, whichfork);
fs/xfs/libxfs/xfs_bmap.c
679
xfs_rmap_ino_bmbt_owner(&args.oinfo, ip->i_ino, whichfork);
fs/xfs/libxfs/xfs_bmap.c
685
XFS_INO_TO_FSB(mp, ip->i_ino));
fs/xfs/libxfs/xfs_bmap.c
698
ip->i_nblocks++;
fs/xfs/libxfs/xfs_bmap.c
699
xfs_trans_mod_dquot_byino(tp, ip, XFS_TRANS_DQ_BCOUNT, 1L);
fs/xfs/libxfs/xfs_bmap.c
710
xfs_bmbt_init_block(ip, ablock, abp, 0, 0);
fs/xfs/libxfs/xfs_bmap.c
744
xfs_trans_mod_dquot_byino(tp, ip, XFS_TRANS_DQ_BCOUNT, -1L);
fs/xfs/libxfs/xfs_bmap.c
746
xfs_bmap_broot_realloc(ip, whichfork, 0);
fs/xfs/libxfs/xfs_bmap.c
763
struct xfs_inode *ip,
fs/xfs/libxfs/xfs_bmap.c
766
struct xfs_ifork *ifp = xfs_ifork_ptr(ip, whichfork);
fs/xfs/libxfs/xfs_bmap.c
773
xfs_bmap_forkoff_reset(ip, whichfork);
fs/xfs/libxfs/xfs_bmap.c
777
xfs_trans_log_inode(tp, ip, XFS_ILOG_CORE);
fs/xfs/libxfs/xfs_bmap.c
784
xfs_inode_t *ip, /* incore inode pointer */
fs/xfs/libxfs/xfs_bmap.c
790
struct xfs_inode *ip,
fs/xfs/libxfs/xfs_bmap.c
806
ASSERT(!(S_ISREG(VFS_I(ip)->i_mode) && whichfork == XFS_DATA_FORK));
fs/xfs/libxfs/xfs_bmap.c
807
ifp = xfs_ifork_ptr(ip, whichfork);
fs/xfs/libxfs/xfs_bmap.c
811
xfs_bmap_local_to_extents_empty(tp, ip, whichfork);
fs/xfs/libxfs/xfs_bmap.c
820
args.mp = ip->i_mount;
fs/xfs/libxfs/xfs_bmap.c
823
xfs_rmap_ino_owner(&args.oinfo, ip->i_ino, whichfork, 0);
fs/xfs/libxfs/xfs_bmap.c
832
XFS_INO_TO_FSB(args.mp, ip->i_ino));
fs/xfs/libxfs/xfs_bmap.c
853
init_fn(tp, bp, ip, ifp, priv);
fs/xfs/libxfs/xfs_bmap.c
856
xfs_idata_realloc(ip, -ifp->if_bytes, whichfork);
fs/xfs/libxfs/xfs_bmap.c
857
xfs_bmap_local_to_extents_empty(tp, ip, whichfork);
fs/xfs/libxfs/xfs_bmap.c
868
xfs_iext_insert(ip, &icur, &rec, 0);
fs/xfs/libxfs/xfs_bmap.c
871
ip->i_nblocks = 1;
fs/xfs/libxfs/xfs_bmap.c
872
xfs_trans_mod_dquot_byino(tp, ip, XFS_TRANS_DQ_BCOUNT, 1L);
fs/xfs/libxfs/xfs_bmap.c
886
xfs_inode_t *ip, /* incore inode pointer */
fs/xfs/libxfs/xfs_bmap.c
889
struct xfs_btree_block *block = ip->i_df.if_broot;
fs/xfs/libxfs/xfs_bmap.c
895
mp = ip->i_mount;
fs/xfs/libxfs/xfs_bmap.c
897
if (xfs_bmap_bmdr_space(block) <= xfs_inode_data_fork_size(ip))
fs/xfs/libxfs/xfs_bmap.c
900
cur = xfs_bmbt_init_cursor(mp, tp, ip, XFS_DATA_FORK);
fs/xfs/libxfs/xfs_bmap.c
931
struct xfs_inode *ip, /* incore inode pointer */
fs/xfs/libxfs/xfs_bmap.c
937
if (ip->i_df.if_nextents * sizeof(struct xfs_bmbt_rec) <=
fs/xfs/libxfs/xfs_bmap.c
938
xfs_inode_data_fork_size(ip))
fs/xfs/libxfs/xfs_bmap.c
941
error = xfs_bmap_extents_to_btree(tp, ip, &cur, 0, flags,
fs/xfs/libxfs/xfs_bmap.c
964
struct xfs_inode *ip, /* incore inode pointer */
fs/xfs/libxfs/xfs_bmap.c
969
if (ip->i_df.if_bytes <= xfs_inode_data_fork_size(ip))
fs/xfs/libxfs/xfs_bmap.c
972
if (S_ISDIR(VFS_I(ip)->i_mode)) {
fs/xfs/libxfs/xfs_bmap.c
974
dargs.geo = ip->i_mount->m_dir_geo;
fs/xfs/libxfs/xfs_bmap.c
975
dargs.dp = ip;
fs/xfs/libxfs/xfs_bmap.c
979
dargs.owner = ip->i_ino;
fs/xfs/libxfs/xfs_bmap.c
983
if (S_ISLNK(VFS_I(ip)->i_mode))
fs/xfs/libxfs/xfs_bmap.c
984
return xfs_bmap_local_to_extents(tp, ip, 1, flags,
fs/xfs/libxfs/xfs_bmap.c
990
xfs_bmap_mark_sick(ip, XFS_ATTR_FORK);
fs/xfs/libxfs/xfs_bmap.c
999
struct xfs_inode *ip,
fs/xfs/libxfs/xfs_bmap.h
175
#define xfs_valid_startblock(ip, startblock) \
fs/xfs/libxfs/xfs_bmap.h
176
((startblock) != 0 || XFS_IS_REALTIME_INODE(ip))
fs/xfs/libxfs/xfs_bmap.h
183
int xfs_bmap_add_attrfork(struct xfs_trans *tp, struct xfs_inode *ip,
fs/xfs/libxfs/xfs_bmap.h
186
struct xfs_inode *ip, int whichfork);
fs/xfs/libxfs/xfs_bmap.h
187
int xfs_bmap_local_to_extents(struct xfs_trans *tp, struct xfs_inode *ip,
fs/xfs/libxfs/xfs_bmap.h
190
struct xfs_inode *ip, struct xfs_ifork *ifp,
fs/xfs/libxfs/xfs_bmap.h
194
int xfs_bmap_first_unused(struct xfs_trans *tp, struct xfs_inode *ip,
fs/xfs/libxfs/xfs_bmap.h
196
int xfs_bmap_last_before(struct xfs_trans *tp, struct xfs_inode *ip,
fs/xfs/libxfs/xfs_bmap.h
198
int xfs_bmap_last_offset(struct xfs_inode *ip, xfs_fileoff_t *unused,
fs/xfs/libxfs/xfs_bmap.h
200
int xfs_bmapi_read(struct xfs_inode *ip, xfs_fileoff_t bno,
fs/xfs/libxfs/xfs_bmap.h
203
int xfs_bmapi_write(struct xfs_trans *tp, struct xfs_inode *ip,
fs/xfs/libxfs/xfs_bmap.h
206
int xfs_bunmapi(struct xfs_trans *tp, struct xfs_inode *ip,
fs/xfs/libxfs/xfs_bmap.h
209
void xfs_bmap_del_extent_delay(struct xfs_inode *ip, int whichfork,
fs/xfs/libxfs/xfs_bmap.h
212
void xfs_bmap_del_extent_cow(struct xfs_inode *ip,
fs/xfs/libxfs/xfs_bmap.h
215
uint xfs_default_attroffset(struct xfs_inode *ip);
fs/xfs/libxfs/xfs_bmap.h
216
int xfs_bmap_collapse_extents(struct xfs_trans *tp, struct xfs_inode *ip,
fs/xfs/libxfs/xfs_bmap.h
219
int xfs_bmap_can_insert_extents(struct xfs_inode *ip, xfs_fileoff_t off,
fs/xfs/libxfs/xfs_bmap.h
22
struct xfs_inode *ip; /* incore inode pointer */
fs/xfs/libxfs/xfs_bmap.h
221
int xfs_bmap_insert_extents(struct xfs_trans *tp, struct xfs_inode *ip,
fs/xfs/libxfs/xfs_bmap.h
224
int xfs_bmap_split_extent(struct xfs_trans *tp, struct xfs_inode *ip,
fs/xfs/libxfs/xfs_bmap.h
226
int xfs_bmapi_convert_delalloc(struct xfs_inode *ip, int whichfork,
fs/xfs/libxfs/xfs_bmap.h
229
struct xfs_inode *ip, int whichfork,
fs/xfs/libxfs/xfs_bmap.h
232
xfs_extlen_t xfs_bmapi_minleft(struct xfs_trans *tp, struct xfs_inode *ip,
fs/xfs/libxfs/xfs_bmap.h
236
xfs_filblks_t xfs_bmap_worst_indlen(struct xfs_inode *ip, xfs_filblks_t len);
fs/xfs/libxfs/xfs_bmap.h
257
void xfs_bmap_map_extent(struct xfs_trans *tp, struct xfs_inode *ip,
fs/xfs/libxfs/xfs_bmap.h
259
void xfs_bmap_unmap_extent(struct xfs_trans *tp, struct xfs_inode *ip,
fs/xfs/libxfs/xfs_bmap.h
276
xfs_failaddr_t xfs_bmap_validate_extent(struct xfs_inode *ip, int whichfork,
fs/xfs/libxfs/xfs_bmap.h
278
int xfs_bmap_complain_bad_rec(struct xfs_inode *ip, int whichfork,
fs/xfs/libxfs/xfs_bmap.h
281
int xfs_bmapi_remap(struct xfs_trans *tp, struct xfs_inode *ip,
fs/xfs/libxfs/xfs_bmap.h
284
int xfs_bunmapi_range(struct xfs_trans **tpp, struct xfs_inode *ip,
fs/xfs/libxfs/xfs_bmap.h
300
xfs_extlen_t xfs_get_extsz_hint(struct xfs_inode *ip);
fs/xfs/libxfs/xfs_bmap.h
301
xfs_extlen_t xfs_get_cowextsz_hint(struct xfs_inode *ip);
fs/xfs/libxfs/xfs_bmap_btree.c
187
cur->bc_ino.ip, cur->bc_ino.whichfork);
fs/xfs/libxfs/xfs_bmap_btree.c
199
(dst->bc_ino.ip->i_diflags & XFS_DIFLAG_REALTIME));
fs/xfs/libxfs/xfs_bmap_btree.c
220
xfs_rmap_ino_bmbt_owner(&args.oinfo, cur->bc_ino.ip->i_ino,
fs/xfs/libxfs/xfs_bmap_btree.c
234
args.minleft = xfs_bmapi_minleft(cur->bc_tp, cur->bc_ino.ip,
fs/xfs/libxfs/xfs_bmap_btree.c
260
cur->bc_ino.ip->i_nblocks++;
fs/xfs/libxfs/xfs_bmap_btree.c
261
xfs_trans_log_inode(args.tp, cur->bc_ino.ip, XFS_ILOG_CORE);
fs/xfs/libxfs/xfs_bmap_btree.c
262
xfs_trans_mod_dquot_byino(args.tp, cur->bc_ino.ip,
fs/xfs/libxfs/xfs_bmap_btree.c
277
struct xfs_inode *ip = cur->bc_ino.ip;
fs/xfs/libxfs/xfs_bmap_btree.c
283
xfs_rmap_ino_bmbt_owner(&oinfo, ip->i_ino, cur->bc_ino.whichfork);
fs/xfs/libxfs/xfs_bmap_btree.c
289
ip->i_nblocks--;
fs/xfs/libxfs/xfs_bmap_btree.c
290
xfs_trans_log_inode(tp, ip, XFS_ILOG_CORE);
fs/xfs/libxfs/xfs_bmap_btree.c
291
xfs_trans_mod_dquot_byino(tp, ip, XFS_TRANS_DQ_BCOUNT, -1L);
fs/xfs/libxfs/xfs_bmap_btree.c
31
struct xfs_inode *ip,
fs/xfs/libxfs/xfs_bmap_btree.c
38
xfs_btree_init_buf(ip->i_mount, bp, &xfs_bmbt_ops, level,
fs/xfs/libxfs/xfs_bmap_btree.c
39
numrecs, ip->i_ino);
fs/xfs/libxfs/xfs_bmap_btree.c
41
xfs_btree_init_block(ip->i_mount, buf, &xfs_bmbt_ops, level,
fs/xfs/libxfs/xfs_bmap_btree.c
42
numrecs, ip->i_ino);
fs/xfs/libxfs/xfs_bmap_btree.c
50
struct xfs_inode *ip,
fs/xfs/libxfs/xfs_bmap_btree.c
542
struct xfs_inode *ip,
fs/xfs/libxfs/xfs_bmap_btree.c
546
struct xfs_mount *mp = ip->i_mount;
fs/xfs/libxfs/xfs_bmap_btree.c
547
struct xfs_ifork *ifp = xfs_ifork_ptr(ip, whichfork);
fs/xfs/libxfs/xfs_bmap_btree.c
56
struct xfs_mount *mp = ip->i_mount;
fs/xfs/libxfs/xfs_bmap_btree.c
585
xfs_inode_fork_size(ip, whichfork));
fs/xfs/libxfs/xfs_bmap_btree.c
604
xfs_inode_fork_size(ip, whichfork));
fs/xfs/libxfs/xfs_bmap_btree.c
613
return xfs_bmap_broot_realloc(cur->bc_ino.ip, cur->bc_ino.whichfork,
fs/xfs/libxfs/xfs_bmap_btree.c
63
xfs_bmbt_init_block(ip, rblock, NULL, 0, 0);
fs/xfs/libxfs/xfs_bmap_btree.c
656
struct xfs_inode *ip,
fs/xfs/libxfs/xfs_bmap_btree.c
678
cur->bc_ino.ip = ip;
fs/xfs/libxfs/xfs_bmap_btree.c
682
struct xfs_ifork *ifp = xfs_ifork_ptr(ip, whichfork);
fs/xfs/libxfs/xfs_bmap_btree.c
685
cur->bc_ino.forksize = xfs_inode_fork_size(ip, whichfork);
fs/xfs/libxfs/xfs_bmap_btree.c
725
ifp = xfs_ifork_ptr(cur->bc_ino.ip, whichfork);
fs/xfs/libxfs/xfs_bmap_btree.c
740
xfs_trans_log_inode(tp, cur->bc_ino.ip, flags);
fs/xfs/libxfs/xfs_bmap_btree.c
814
struct xfs_inode *ip,
fs/xfs/libxfs/xfs_bmap_btree.c
824
ASSERT(xfs_ifork_ptr(ip, whichfork)->if_format == XFS_DINODE_FMT_BTREE);
fs/xfs/libxfs/xfs_bmap_btree.c
826
cur = xfs_bmbt_init_cursor(ip->i_mount, tp, ip, whichfork);
fs/xfs/libxfs/xfs_bmap_btree.h
201
struct xfs_btree_block *xfs_bmap_broot_realloc(struct xfs_inode *ip,
fs/xfs/libxfs/xfs_bmap_btree.h
41
extern int xfs_bmbt_change_owner(struct xfs_trans *tp, struct xfs_inode *ip,
fs/xfs/libxfs/xfs_bmap_btree.h
58
void xfs_bmbt_init_block(struct xfs_inode *ip, struct xfs_btree_block *buf,
fs/xfs/libxfs/xfs_btree.c
1308
return cur->bc_ino.ip->i_ino;
fs/xfs/libxfs/xfs_btree.c
1531
xfs_trans_log_inode(cur->bc_tp, cur->bc_ino.ip,
fs/xfs/libxfs/xfs_btree.c
1547
xfs_trans_log_inode(cur->bc_tp, cur->bc_ino.ip,
fs/xfs/libxfs/xfs_btree.c
1578
xfs_trans_log_inode(cur->bc_tp, cur->bc_ino.ip,
fs/xfs/libxfs/xfs_btree.c
1647
xfs_trans_log_inode(cur->bc_tp, cur->bc_ino.ip,
fs/xfs/libxfs/xfs_btree.c
3132
cur->bc_nlevels - 1, 1, cur->bc_ino.ip->i_ino);
fs/xfs/libxfs/xfs_btree.c
3247
cur->bc_ino.ip->i_ino));
fs/xfs/libxfs/xfs_btree.c
3452
struct xfs_inode *ip = cur->bc_ino.ip;
fs/xfs/libxfs/xfs_btree.c
3466
xfs_trans_log_inode(cur->bc_tp, ip, logflags);
fs/xfs/libxfs/xfs_btree.c
374
cur->bc_ino.ip->i_ino,
fs/xfs/libxfs/xfs_btree.c
3830
cur->bc_ino.ip->i_ino);
fs/xfs/libxfs/xfs_btree.c
3905
struct xfs_inode *ip = cur->bc_ino.ip;
fs/xfs/libxfs/xfs_btree.c
3971
xfs_trans_log_inode(cur->bc_tp, ip,
fs/xfs/libxfs/xfs_btree.c
5606
struct xfs_inode *ip = cur->bc_ino.ip;
fs/xfs/libxfs/xfs_btree.c
5609
ASSERT(xfs_is_metadir_inode(ip));
fs/xfs/libxfs/xfs_btree.c
5611
xfs_rmap_ino_bmbt_owner(&args.oinfo, ip->i_ino, cur->bc_ino.whichfork);
fs/xfs/libxfs/xfs_btree.c
5613
XFS_INO_TO_FSB(cur->bc_mp, ip->i_ino));
fs/xfs/libxfs/xfs_btree.c
5622
xfs_metafile_resv_alloc_space(ip, &args);
fs/xfs/libxfs/xfs_btree.c
5637
struct xfs_inode *ip = cur->bc_ino.ip;
fs/xfs/libxfs/xfs_btree.c
5642
ASSERT(xfs_is_metadir_inode(ip));
fs/xfs/libxfs/xfs_btree.c
5644
xfs_rmap_ino_bmbt_owner(&oinfo, ip->i_ino, cur->bc_ino.whichfork);
fs/xfs/libxfs/xfs_btree.c
5650
xfs_metafile_resv_free_space(ip, tp, 1);
fs/xfs/libxfs/xfs_btree.c
822
return xfs_ifork_ptr(cur->bc_ino.ip, cur->bc_ino.whichfork);
fs/xfs/libxfs/xfs_btree.h
281
struct xfs_inode *ip;
fs/xfs/libxfs/xfs_btree_staging.c
312
level, nr_this_block, cur->bc_ino.ip->i_ino);
fs/xfs/libxfs/xfs_dir2.c
1015
error = xfs_droplink(tp, ip);
fs/xfs/libxfs/xfs_dir2.c
1019
error = xfs_dir_removename(tp, dp, name, ip->i_ino, resblks);
fs/xfs/libxfs/xfs_dir2.c
1027
error = xfs_parent_removename(tp, du->ppargs, dp, name, ip);
fs/xfs/libxfs/xfs_dir2.c
1032
xfs_dir_update_hook(dp, ip, -1, name);
fs/xfs/libxfs/xfs_dir2.c
1052
struct xfs_inode *ip1 = du1->ip;
fs/xfs/libxfs/xfs_dir2.c
1055
struct xfs_inode *ip2 = du2->ip;
fs/xfs/libxfs/xfs_dir2.c
1197
struct xfs_inode *src_ip = du_src->ip;
fs/xfs/libxfs/xfs_dir2.c
1200
struct xfs_inode *target_ip = du_tgt->ip;
fs/xfs/libxfs/xfs_dir2.c
1244
if (du_wip->ip) {
fs/xfs/libxfs/xfs_dir2.c
1247
ASSERT(VFS_I(du_wip->ip)->i_nlink == 0);
fs/xfs/libxfs/xfs_dir2.c
1249
pag = xfs_perag_get(mp, XFS_INO_TO_AGNO(mp, du_wip->ip->i_ino));
fs/xfs/libxfs/xfs_dir2.c
1250
error = xfs_iunlink_remove(tp, pag, du_wip->ip);
fs/xfs/libxfs/xfs_dir2.c
1255
xfs_bumplink(tp, du_wip->ip);
fs/xfs/libxfs/xfs_dir2.c
1360
if (du_wip->ip)
fs/xfs/libxfs/xfs_dir2.c
1361
error = xfs_dir_replace(tp, src_dp, src_name, du_wip->ip->i_ino,
fs/xfs/libxfs/xfs_dir2.c
1377
src_name, du_wip->ip);
fs/xfs/libxfs/xfs_dir2.c
1408
if (du_wip->ip)
fs/xfs/libxfs/xfs_dir2.c
1409
xfs_dir_update_hook(src_dp, du_wip->ip, 1, src_name);
fs/xfs/libxfs/xfs_dir2.c
793
struct xfs_inode *ip,
fs/xfs/libxfs/xfs_dir2.c
800
.ip = ip,
fs/xfs/libxfs/xfs_dir2.c
804
struct xfs_mount *mp = ip->i_mount;
fs/xfs/libxfs/xfs_dir2.c
852
struct xfs_inode *ip = du->ip;
fs/xfs/libxfs/xfs_dir2.c
855
xfs_assert_ilocked(ip, XFS_ILOCK_EXCL);
fs/xfs/libxfs/xfs_dir2.c
858
error = xfs_dir_createname(tp, dp, name, ip->i_ino, resblks);
fs/xfs/libxfs/xfs_dir2.c
867
if (S_ISDIR(VFS_I(ip)->i_mode)) {
fs/xfs/libxfs/xfs_dir2.c
868
error = xfs_dir_init(tp, ip, dp);
fs/xfs/libxfs/xfs_dir2.c
880
error = xfs_parent_addname(tp, du->ppargs, dp, name, ip);
fs/xfs/libxfs/xfs_dir2.c
885
xfs_dir_update_hook(dp, ip, 1, name);
fs/xfs/libxfs/xfs_dir2.c
902
struct xfs_inode *ip = du->ip;
fs/xfs/libxfs/xfs_dir2.c
906
xfs_assert_ilocked(ip, XFS_ILOCK_EXCL);
fs/xfs/libxfs/xfs_dir2.c
908
ASSERT(!S_ISDIR(VFS_I(ip)->i_mode));
fs/xfs/libxfs/xfs_dir2.c
919
if (VFS_I(ip)->i_nlink == 0) {
fs/xfs/libxfs/xfs_dir2.c
922
pag = xfs_perag_get(mp, XFS_INO_TO_AGNO(mp, ip->i_ino));
fs/xfs/libxfs/xfs_dir2.c
923
error = xfs_iunlink_remove(tp, pag, ip);
fs/xfs/libxfs/xfs_dir2.c
929
error = xfs_dir_createname(tp, dp, name, ip->i_ino, resblks);
fs/xfs/libxfs/xfs_dir2.c
936
xfs_bumplink(tp, ip);
fs/xfs/libxfs/xfs_dir2.c
945
error = xfs_parent_addname(tp, du->ppargs, dp, name, ip);
fs/xfs/libxfs/xfs_dir2.c
950
xfs_dir_update_hook(dp, ip, 1, name);
fs/xfs/libxfs/xfs_dir2.c
966
struct xfs_inode *ip = du->ip;
fs/xfs/libxfs/xfs_dir2.c
969
xfs_assert_ilocked(ip, XFS_ILOCK_EXCL);
fs/xfs/libxfs/xfs_dir2.c
975
if (S_ISDIR(VFS_I(ip)->i_mode)) {
fs/xfs/libxfs/xfs_dir2.c
976
ASSERT(VFS_I(ip)->i_nlink >= 2);
fs/xfs/libxfs/xfs_dir2.c
977
if (VFS_I(ip)->i_nlink != 2)
fs/xfs/libxfs/xfs_dir2.c
979
if (!xfs_dir_isempty(ip))
fs/xfs/libxfs/xfs_dir2.c
988
error = xfs_droplink(tp, ip);
fs/xfs/libxfs/xfs_dir2.c
999
error = xfs_dir_replace(tp, ip, &xfs_name_dotdot,
fs/xfs/libxfs/xfs_dir2.h
313
const struct xfs_inode *ip;
fs/xfs/libxfs/xfs_dir2.h
319
void xfs_dir_update_hook(struct xfs_inode *dp, struct xfs_inode *ip,
fs/xfs/libxfs/xfs_dir2.h
333
# define xfs_dir_update_hook(dp, ip, delta, name) ((void)0)
fs/xfs/libxfs/xfs_dir2.h
341
struct xfs_inode *ip;
fs/xfs/libxfs/xfs_dquot_buf.c
363
struct xfs_inode *ip;
fs/xfs/libxfs/xfs_dquot_buf.c
389
error = xfs_trans_metafile_iget(tp, ino, metafile_type, &ip);
fs/xfs/libxfs/xfs_dquot_buf.c
392
metafile_type, &ip);
fs/xfs/libxfs/xfs_dquot_buf.c
402
if (XFS_IS_CORRUPT(mp, ip->i_df.if_format != XFS_DINODE_FMT_EXTENTS &&
fs/xfs/libxfs/xfs_dquot_buf.c
403
ip->i_df.if_format != XFS_DINODE_FMT_BTREE)) {
fs/xfs/libxfs/xfs_dquot_buf.c
404
xfs_irele(ip);
fs/xfs/libxfs/xfs_dquot_buf.c
409
if (XFS_IS_CORRUPT(mp, ip->i_projid != 0)) {
fs/xfs/libxfs/xfs_dquot_buf.c
410
xfs_irele(ip);
fs/xfs/libxfs/xfs_dquot_buf.c
415
*ipp = ip;
fs/xfs/libxfs/xfs_dquot_buf.c
441
xfs_trans_log_inode(upd.tp, upd.ip, XFS_ILOG_CORE);
fs/xfs/libxfs/xfs_dquot_buf.c
447
xfs_finish_inode_setup(upd.ip);
fs/xfs/libxfs/xfs_dquot_buf.c
448
*ipp = upd.ip;
fs/xfs/libxfs/xfs_dquot_buf.c
458
struct xfs_inode *ip)
fs/xfs/libxfs/xfs_dquot_buf.c
464
.ip = ip,
fs/xfs/libxfs/xfs_dquot_buf.c
476
xfs_trans_log_inode(upd.tp, upd.ip, XFS_ILOG_CORE);
fs/xfs/libxfs/xfs_exchmaps.c
102
trace_xfs_exchmaps_update_inode_size(ip, len);
fs/xfs/libxfs/xfs_exchmaps.c
104
ip->i_disk_size = len;
fs/xfs/libxfs/xfs_exchmaps.c
105
xfs_trans_log_inode(tp, ip, XFS_ILOG_CORE);
fs/xfs/libxfs/xfs_exchmaps.c
1146
struct xfs_inode *ip)
fs/xfs/libxfs/xfs_exchmaps.c
1148
trace_xfs_reflink_set_inode_flag(ip);
fs/xfs/libxfs/xfs_exchmaps.c
1150
ip->i_diflags2 |= XFS_DIFLAG2_REFLINK;
fs/xfs/libxfs/xfs_exchmaps.c
1151
xfs_trans_log_inode(tp, ip, XFS_ILOG_CORE);
fs/xfs/libxfs/xfs_exchmaps.c
1182
struct xfs_inode *ip)
fs/xfs/libxfs/xfs_exchmaps.c
1184
if (xfs_inode_has_large_extent_counts(ip))
fs/xfs/libxfs/xfs_exchmaps.c
1187
ip->i_diflags2 |= XFS_DIFLAG2_NREXT64;
fs/xfs/libxfs/xfs_exchmaps.c
1188
xfs_trans_log_inode(tp, ip, XFS_ILOG_CORE);
fs/xfs/libxfs/xfs_exchmaps.c
491
struct xfs_inode *ip = xmi->xmi_ip2;
fs/xfs/libxfs/xfs_exchmaps.c
492
struct xfs_ifork *ifp = xfs_ifork_ptr(ip, XFS_DATA_FORK);
fs/xfs/libxfs/xfs_exchmaps.c
497
ip->i_disk_size > xfs_inode_data_fork_size(ip))
fs/xfs/libxfs/xfs_exchmaps.c
501
buf = kmalloc(ip->i_disk_size + 1,
fs/xfs/libxfs/xfs_exchmaps.c
508
error = xfs_symlink_remote_read(ip, buf);
fs/xfs/libxfs/xfs_exchmaps.c
513
error = xfs_symlink_remote_truncate(tp, ip);
fs/xfs/libxfs/xfs_exchmaps.c
521
xfs_init_local_fork(ip, XFS_DATA_FORK, buf, ip->i_disk_size);
fs/xfs/libxfs/xfs_exchmaps.c
522
xfs_trans_log_inode(tp, ip, XFS_ILOG_DDATA | XFS_ILOG_CORE);
fs/xfs/libxfs/xfs_exchmaps.c
532
struct xfs_inode *ip)
fs/xfs/libxfs/xfs_exchmaps.c
534
trace_xfs_reflink_unset_inode_flag(ip);
fs/xfs/libxfs/xfs_exchmaps.c
536
ip->i_diflags2 &= ~XFS_DIFLAG2_REFLINK;
fs/xfs/libxfs/xfs_exchmaps.c
537
xfs_trans_log_inode(tp, ip, XFS_ILOG_CORE);
fs/xfs/libxfs/xfs_exchmaps.c
62
struct xfs_inode *ip)
fs/xfs/libxfs/xfs_exchmaps.c
66
if (xfs_is_reflink_inode(ip))
fs/xfs/libxfs/xfs_exchmaps.c
67
xfs_ifork_init_cow(ip);
fs/xfs/libxfs/xfs_exchmaps.c
69
cfork = xfs_ifork_ptr(ip, XFS_COW_FORK);
fs/xfs/libxfs/xfs_exchmaps.c
73
xfs_inode_set_cowblocks_tag(ip);
fs/xfs/libxfs/xfs_exchmaps.c
75
xfs_inode_clear_cowblocks_tag(ip);
fs/xfs/libxfs/xfs_exchmaps.c
86
struct xfs_inode *ip,
fs/xfs/libxfs/xfs_exchmaps.c
864
struct xfs_inode *ip,
fs/xfs/libxfs/xfs_exchmaps.c
867
struct xfs_mount *mp = ip->i_mount;
fs/xfs/libxfs/xfs_exchmaps.c
869
struct xfs_ifork *ifp = xfs_ifork_ptr(ip, whichfork);
fs/xfs/libxfs/xfs_exchmaps.c
905
struct xfs_inode *ip,
fs/xfs/libxfs/xfs_exchmaps.c
917
error = xfs_bmapi_read(ip, off, blockcount, nrec, &nimaps, bmap_flags);
fs/xfs/libxfs/xfs_exchmaps.c
99
if (len <= ip->i_disk_size)
fs/xfs/libxfs/xfs_health.h
218
void xfs_inode_mark_sick(struct xfs_inode *ip, unsigned int mask);
fs/xfs/libxfs/xfs_health.h
219
void xfs_inode_mark_corrupt(struct xfs_inode *ip, unsigned int mask);
fs/xfs/libxfs/xfs_health.h
220
void xfs_inode_mark_healthy(struct xfs_inode *ip, unsigned int mask);
fs/xfs/libxfs/xfs_health.h
221
void xfs_inode_measure_sickness(struct xfs_inode *ip, unsigned int *sick,
fs/xfs/libxfs/xfs_health.h
225
void xfs_bmap_mark_sick(struct xfs_inode *ip, int whichfork);
fs/xfs/libxfs/xfs_health.h
227
void xfs_dirattr_mark_sick(struct xfs_inode *ip, int whichfork);
fs/xfs/libxfs/xfs_health.h
263
xfs_inode_has_sickness(struct xfs_inode *ip, unsigned int mask)
fs/xfs/libxfs/xfs_health.h
267
xfs_inode_measure_sickness(ip, &sick, &checked);
fs/xfs/libxfs/xfs_health.h
278
xfs_inode_is_healthy(struct xfs_inode *ip)
fs/xfs/libxfs/xfs_health.h
280
return !xfs_inode_has_sickness(ip, -1U);
fs/xfs/libxfs/xfs_health.h
287
void xfs_bulkstat_health(struct xfs_inode *ip, struct xfs_bulkstat *bs);
fs/xfs/libxfs/xfs_iext_tree.c
1001
struct xfs_ifork *ifp = xfs_iext_state_to_fork(ip, state);
fs/xfs/libxfs/xfs_iext_tree.c
1015
trace_xfs_bmap_pre_update(ip, cur, state, _RET_IP_);
fs/xfs/libxfs/xfs_iext_tree.c
1017
trace_xfs_bmap_post_update(ip, cur, state, _RET_IP_);
fs/xfs/libxfs/xfs_iext_tree.c
677
struct xfs_inode *ip,
fs/xfs/libxfs/xfs_iext_tree.c
682
struct xfs_ifork *ifp = xfs_iext_state_to_fork(ip, state);
fs/xfs/libxfs/xfs_iext_tree.c
685
trace_xfs_iext_insert(ip, cur, state, _RET_IP_);
fs/xfs/libxfs/xfs_iext_tree.c
880
struct xfs_inode *ip,
fs/xfs/libxfs/xfs_iext_tree.c
884
struct xfs_ifork *ifp = xfs_iext_state_to_fork(ip, state);
fs/xfs/libxfs/xfs_iext_tree.c
889
trace_xfs_iext_remove(ip, cur, state, _RET_IP_);
fs/xfs/libxfs/xfs_iext_tree.c
937
struct xfs_inode *ip,
fs/xfs/libxfs/xfs_iext_tree.c
943
XFS_STATS_INC(ip->i_mount, xs_look_exlist);
fs/xfs/libxfs/xfs_iext_tree.c
978
struct xfs_inode *ip,
fs/xfs/libxfs/xfs_iext_tree.c
985
if (xfs_iext_lookup_extent(ip, ifp, *end - 1, cur, gotp) &&
fs/xfs/libxfs/xfs_iext_tree.c
996
struct xfs_inode *ip,
fs/xfs/libxfs/xfs_inode_buf.c
179
struct xfs_inode *ip,
fs/xfs/libxfs/xfs_inode_buf.c
182
struct inode *inode = VFS_I(ip);
fs/xfs/libxfs/xfs_inode_buf.c
186
ASSERT(ip->i_cowfp == NULL);
fs/xfs/libxfs/xfs_inode_buf.c
188
fa = xfs_dinode_verify(ip->i_mount, ip->i_ino, from);
fs/xfs/libxfs/xfs_inode_buf.c
190
xfs_inode_verifier_error(ip, -EFSCORRUPTED, "dinode", from,
fs/xfs/libxfs/xfs_inode_buf.c
200
if (!xfs_has_v3inodes(ip->i_mount))
fs/xfs/libxfs/xfs_inode_buf.c
201
ip->i_flushiter = be16_to_cpu(from->di_flushiter);
fs/xfs/libxfs/xfs_inode_buf.c
215
ip->i_projid = 0;
fs/xfs/libxfs/xfs_inode_buf.c
218
ip->i_projid = (prid_t)be16_to_cpu(from->di_projid_hi) << 16 |
fs/xfs/libxfs/xfs_inode_buf.c
221
ip->i_metatype = be16_to_cpu(from->di_metatype);
fs/xfs/libxfs/xfs_inode_buf.c
240
ip->i_disk_size = be64_to_cpu(from->di_size);
fs/xfs/libxfs/xfs_inode_buf.c
241
ip->i_nblocks = be64_to_cpu(from->di_nblocks);
fs/xfs/libxfs/xfs_inode_buf.c
242
ip->i_extsize = be32_to_cpu(from->di_extsize);
fs/xfs/libxfs/xfs_inode_buf.c
243
ip->i_forkoff = from->di_forkoff;
fs/xfs/libxfs/xfs_inode_buf.c
244
ip->i_diflags = be16_to_cpu(from->di_flags);
fs/xfs/libxfs/xfs_inode_buf.c
245
ip->i_next_unlinked = be32_to_cpu(from->di_next_unlinked);
fs/xfs/libxfs/xfs_inode_buf.c
248
xfs_iflags_set(ip, XFS_IPRESERVE_DM_FIELDS);
fs/xfs/libxfs/xfs_inode_buf.c
250
if (xfs_has_v3inodes(ip->i_mount)) {
fs/xfs/libxfs/xfs_inode_buf.c
253
ip->i_crtime = xfs_inode_from_disk_ts(from, from->di_crtime);
fs/xfs/libxfs/xfs_inode_buf.c
254
ip->i_diflags2 = be64_to_cpu(from->di_flags2);
fs/xfs/libxfs/xfs_inode_buf.c
256
ip->i_cowextsize = be32_to_cpu(from->di_cowextsize);
fs/xfs/libxfs/xfs_inode_buf.c
261
error = xfs_iformat_data_fork(ip, from);
fs/xfs/libxfs/xfs_inode_buf.c
265
error = xfs_iformat_attr_fork(ip, from);
fs/xfs/libxfs/xfs_inode_buf.c
269
if (xfs_is_reflink_inode(ip))
fs/xfs/libxfs/xfs_inode_buf.c
270
xfs_ifork_init_cow(ip);
fs/xfs/libxfs/xfs_inode_buf.c
271
if (xfs_is_metadir_inode(ip)) {
fs/xfs/libxfs/xfs_inode_buf.c
272
XFS_STATS_DEC(ip->i_mount, xs_inodes_active);
fs/xfs/libxfs/xfs_inode_buf.c
273
XFS_STATS_INC(ip->i_mount, xs_inodes_meta);
fs/xfs/libxfs/xfs_inode_buf.c
278
xfs_idestroy_fork(&ip->i_df);
fs/xfs/libxfs/xfs_inode_buf.c
285
struct xfs_inode *ip,
fs/xfs/libxfs/xfs_inode_buf.c
291
if (xfs_inode_has_bigtime(ip))
fs/xfs/libxfs/xfs_inode_buf.c
303
struct xfs_inode *ip,
fs/xfs/libxfs/xfs_inode_buf.c
306
if (xfs_inode_has_large_extent_counts(ip)) {
fs/xfs/libxfs/xfs_inode_buf.c
307
to->di_big_nextents = cpu_to_be64(xfs_ifork_nextents(&ip->i_df));
fs/xfs/libxfs/xfs_inode_buf.c
308
to->di_big_anextents = cpu_to_be32(xfs_ifork_nextents(&ip->i_af));
fs/xfs/libxfs/xfs_inode_buf.c
315
to->di_nextents = cpu_to_be32(xfs_ifork_nextents(&ip->i_df));
fs/xfs/libxfs/xfs_inode_buf.c
316
to->di_anextents = cpu_to_be16(xfs_ifork_nextents(&ip->i_af));
fs/xfs/libxfs/xfs_inode_buf.c
322
struct xfs_inode *ip,
fs/xfs/libxfs/xfs_inode_buf.c
326
struct inode *inode = VFS_I(ip);
fs/xfs/libxfs/xfs_inode_buf.c
329
if (xfs_is_metadir_inode(ip))
fs/xfs/libxfs/xfs_inode_buf.c
330
to->di_metatype = cpu_to_be16(ip->i_metatype);
fs/xfs/libxfs/xfs_inode_buf.c
334
to->di_format = xfs_ifork_format(&ip->i_df);
fs/xfs/libxfs/xfs_inode_buf.c
337
to->di_projid_lo = cpu_to_be16(ip->i_projid & 0xffff);
fs/xfs/libxfs/xfs_inode_buf.c
338
to->di_projid_hi = cpu_to_be16(ip->i_projid >> 16);
fs/xfs/libxfs/xfs_inode_buf.c
340
to->di_atime = xfs_inode_to_disk_ts(ip, inode_get_atime(inode));
fs/xfs/libxfs/xfs_inode_buf.c
341
to->di_mtime = xfs_inode_to_disk_ts(ip, inode_get_mtime(inode));
fs/xfs/libxfs/xfs_inode_buf.c
342
to->di_ctime = xfs_inode_to_disk_ts(ip, inode_get_ctime(inode));
fs/xfs/libxfs/xfs_inode_buf.c
347
to->di_size = cpu_to_be64(ip->i_disk_size);
fs/xfs/libxfs/xfs_inode_buf.c
348
to->di_nblocks = cpu_to_be64(ip->i_nblocks);
fs/xfs/libxfs/xfs_inode_buf.c
349
to->di_extsize = cpu_to_be32(ip->i_extsize);
fs/xfs/libxfs/xfs_inode_buf.c
350
to->di_forkoff = ip->i_forkoff;
fs/xfs/libxfs/xfs_inode_buf.c
351
to->di_aformat = xfs_ifork_format(&ip->i_af);
fs/xfs/libxfs/xfs_inode_buf.c
352
to->di_flags = cpu_to_be16(ip->i_diflags);
fs/xfs/libxfs/xfs_inode_buf.c
354
if (xfs_has_v3inodes(ip->i_mount)) {
fs/xfs/libxfs/xfs_inode_buf.c
357
to->di_crtime = xfs_inode_to_disk_ts(ip, ip->i_crtime);
fs/xfs/libxfs/xfs_inode_buf.c
358
to->di_flags2 = cpu_to_be64(ip->i_diflags2);
fs/xfs/libxfs/xfs_inode_buf.c
360
to->di_cowextsize = cpu_to_be32(ip->i_cowextsize);
fs/xfs/libxfs/xfs_inode_buf.c
361
to->di_ino = cpu_to_be64(ip->i_ino);
fs/xfs/libxfs/xfs_inode_buf.c
364
uuid_copy(&to->di_uuid, &ip->i_mount->m_sb.sb_meta_uuid);
fs/xfs/libxfs/xfs_inode_buf.c
368
to->di_flushiter = cpu_to_be16(ip->i_flushiter);
fs/xfs/libxfs/xfs_inode_buf.c
372
xfs_inode_to_disk_iext_counters(ip, to);
fs/xfs/libxfs/xfs_inode_buf.h
25
void xfs_inode_to_disk(struct xfs_inode *ip, struct xfs_dinode *to,
fs/xfs/libxfs/xfs_inode_buf.h
27
int xfs_inode_from_disk(struct xfs_inode *ip, struct xfs_dinode *from);
fs/xfs/libxfs/xfs_inode_fork.c
109
struct xfs_inode *ip,
fs/xfs/libxfs/xfs_inode_fork.c
113
struct xfs_mount *mp = ip->i_mount;
fs/xfs/libxfs/xfs_inode_fork.c
114
struct xfs_ifork *ifp = xfs_ifork_ptr(ip, whichfork);
fs/xfs/libxfs/xfs_inode_fork.c
128
xfs_warn(ip->i_mount, "corrupt inode %llu ((a)extents = %llu).",
fs/xfs/libxfs/xfs_inode_fork.c
129
ip->i_ino, nex);
fs/xfs/libxfs/xfs_inode_fork.c
130
xfs_inode_verifier_error(ip, -EFSCORRUPTED,
fs/xfs/libxfs/xfs_inode_fork.c
133
xfs_inode_mark_sick(ip, XFS_SICK_INO_CORE);
fs/xfs/libxfs/xfs_inode_fork.c
148
fa = xfs_bmap_validate_extent(ip, whichfork, &new);
fs/xfs/libxfs/xfs_inode_fork.c
150
xfs_inode_verifier_error(ip, -EFSCORRUPTED,
fs/xfs/libxfs/xfs_inode_fork.c
153
xfs_inode_mark_sick(ip, XFS_SICK_INO_CORE);
fs/xfs/libxfs/xfs_inode_fork.c
154
return xfs_bmap_complain_bad_rec(ip, whichfork,
fs/xfs/libxfs/xfs_inode_fork.c
158
xfs_iext_insert(ip, &icur, &new, state);
fs/xfs/libxfs/xfs_inode_fork.c
159
trace_xfs_read_extent(ip, &icur, state, _THIS_IP_);
fs/xfs/libxfs/xfs_inode_fork.c
176
struct xfs_inode *ip,
fs/xfs/libxfs/xfs_inode_fork.c
180
struct xfs_mount *mp = ip->i_mount;
fs/xfs/libxfs/xfs_inode_fork.c
188
ifp = xfs_ifork_ptr(ip, whichfork);
fs/xfs/libxfs/xfs_inode_fork.c
201
if (unlikely(ifp->if_nextents <= XFS_IFORK_MAXEXT(ip, whichfork) ||
fs/xfs/libxfs/xfs_inode_fork.c
205
ifp->if_nextents > ip->i_nblocks) ||
fs/xfs/libxfs/xfs_inode_fork.c
208
(unsigned long long) ip->i_ino);
fs/xfs/libxfs/xfs_inode_fork.c
209
xfs_inode_verifier_error(ip, -EFSCORRUPTED,
fs/xfs/libxfs/xfs_inode_fork.c
212
xfs_inode_mark_sick(ip, XFS_SICK_INO_CORE);
fs/xfs/libxfs/xfs_inode_fork.c
221
xfs_bmdr_to_bmbt(ip, dfp, XFS_DFORK_SIZE(dip, ip->i_mount, whichfork),
fs/xfs/libxfs/xfs_inode_fork.c
232
struct xfs_inode *ip,
fs/xfs/libxfs/xfs_inode_fork.c
235
struct inode *inode = VFS_I(ip);
fs/xfs/libxfs/xfs_inode_fork.c
245
ip->i_df.if_format = dip->di_format;
fs/xfs/libxfs/xfs_inode_fork.c
246
ip->i_df.if_nextents = xfs_dfork_data_extents(dip);
fs/xfs/libxfs/xfs_inode_fork.c
247
smp_store_release(&ip->i_df.if_needextents,
fs/xfs/libxfs/xfs_inode_fork.c
248
ip->i_df.if_format == XFS_DINODE_FMT_BTREE ? 1 : 0);
fs/xfs/libxfs/xfs_inode_fork.c
255
ip->i_disk_size = 0;
fs/xfs/libxfs/xfs_inode_fork.c
261
switch (ip->i_df.if_format) {
fs/xfs/libxfs/xfs_inode_fork.c
263
error = xfs_iformat_local(ip, dip, XFS_DATA_FORK,
fs/xfs/libxfs/xfs_inode_fork.c
266
error = xfs_ifork_verify_local_data(ip);
fs/xfs/libxfs/xfs_inode_fork.c
269
return xfs_iformat_extents(ip, dip, XFS_DATA_FORK);
fs/xfs/libxfs/xfs_inode_fork.c
271
return xfs_iformat_btree(ip, dip, XFS_DATA_FORK);
fs/xfs/libxfs/xfs_inode_fork.c
273
switch (ip->i_metatype) {
fs/xfs/libxfs/xfs_inode_fork.c
275
return xfs_iformat_rtrmap(ip, dip);
fs/xfs/libxfs/xfs_inode_fork.c
277
return xfs_iformat_rtrefcount(ip, dip);
fs/xfs/libxfs/xfs_inode_fork.c
283
xfs_inode_verifier_error(ip, -EFSCORRUPTED, __func__,
fs/xfs/libxfs/xfs_inode_fork.c
285
xfs_inode_mark_sick(ip, XFS_SICK_INO_CORE);
fs/xfs/libxfs/xfs_inode_fork.c
290
xfs_inode_verifier_error(ip, -EFSCORRUPTED, __func__, dip,
fs/xfs/libxfs/xfs_inode_fork.c
292
xfs_inode_mark_sick(ip, XFS_SICK_INO_CORE);
fs/xfs/libxfs/xfs_inode_fork.c
308
struct xfs_inode *ip,
fs/xfs/libxfs/xfs_inode_fork.c
319
ip->i_af.if_format = format;
fs/xfs/libxfs/xfs_inode_fork.c
320
ip->i_af.if_nextents = nextents;
fs/xfs/libxfs/xfs_inode_fork.c
321
smp_store_release(&ip->i_af.if_needextents,
fs/xfs/libxfs/xfs_inode_fork.c
322
ip->i_af.if_format == XFS_DINODE_FMT_BTREE ? 1 : 0);
fs/xfs/libxfs/xfs_inode_fork.c
327
struct xfs_inode *ip)
fs/xfs/libxfs/xfs_inode_fork.c
329
xfs_idestroy_fork(&ip->i_af);
fs/xfs/libxfs/xfs_inode_fork.c
330
memset(&ip->i_af, 0, sizeof(struct xfs_ifork));
fs/xfs/libxfs/xfs_inode_fork.c
331
ip->i_af.if_format = XFS_DINODE_FMT_EXTENTS;
fs/xfs/libxfs/xfs_inode_fork.c
336
struct xfs_inode *ip,
fs/xfs/libxfs/xfs_inode_fork.c
346
xfs_ifork_init_attr(ip, dip->di_aformat, naextents);
fs/xfs/libxfs/xfs_inode_fork.c
348
switch (ip->i_af.if_format) {
fs/xfs/libxfs/xfs_inode_fork.c
350
error = xfs_iformat_local(ip, dip, XFS_ATTR_FORK,
fs/xfs/libxfs/xfs_inode_fork.c
353
error = xfs_ifork_verify_local_attr(ip);
fs/xfs/libxfs/xfs_inode_fork.c
356
error = xfs_iformat_extents(ip, dip, XFS_ATTR_FORK);
fs/xfs/libxfs/xfs_inode_fork.c
359
error = xfs_iformat_btree(ip, dip, XFS_ATTR_FORK);
fs/xfs/libxfs/xfs_inode_fork.c
362
xfs_inode_verifier_error(ip, error, __func__, dip,
fs/xfs/libxfs/xfs_inode_fork.c
364
xfs_inode_mark_sick(ip, XFS_SICK_INO_CORE);
fs/xfs/libxfs/xfs_inode_fork.c
37
struct xfs_inode *ip,
fs/xfs/libxfs/xfs_inode_fork.c
370
xfs_ifork_zap_attr(ip);
fs/xfs/libxfs/xfs_inode_fork.c
42
struct xfs_ifork *ifp = xfs_ifork_ptr(ip, whichfork);
fs/xfs/libxfs/xfs_inode_fork.c
454
struct xfs_inode *ip,
fs/xfs/libxfs/xfs_inode_fork.c
458
struct xfs_ifork *ifp = xfs_ifork_ptr(ip, whichfork);
fs/xfs/libxfs/xfs_inode_fork.c
462
ASSERT(new_size <= xfs_inode_fork_size(ip, whichfork));
fs/xfs/libxfs/xfs_inode_fork.c
509
struct xfs_inode *ip,
fs/xfs/libxfs/xfs_inode_fork.c
514
struct xfs_ifork *ifp = xfs_ifork_ptr(ip, whichfork);
fs/xfs/libxfs/xfs_inode_fork.c
519
xfs_assert_ilocked(ip, XFS_ILOCK_EXCL | XFS_ILOCK_SHARED);
fs/xfs/libxfs/xfs_inode_fork.c
52
zero_terminate = S_ISLNK(VFS_I(ip)->i_mode);
fs/xfs/libxfs/xfs_inode_fork.c
525
ASSERT(xfs_bmap_validate_extent(ip, whichfork, &rec) == NULL);
fs/xfs/libxfs/xfs_inode_fork.c
527
trace_xfs_write_extent(ip, &icur, state, _RET_IP_);
fs/xfs/libxfs/xfs_inode_fork.c
549
struct xfs_inode *ip,
fs/xfs/libxfs/xfs_inode_fork.c
566
ifp = xfs_ifork_ptr(ip, whichfork);
fs/xfs/libxfs/xfs_inode_fork.c
576
mp = ip->i_mount;
fs/xfs/libxfs/xfs_inode_fork.c
582
ASSERT(ifp->if_bytes <= xfs_inode_fork_size(ip, whichfork));
fs/xfs/libxfs/xfs_inode_fork.c
591
(void)xfs_iextents_copy(ip, (xfs_bmbt_rec_t *)cp,
fs/xfs/libxfs/xfs_inode_fork.c
601
xfs_inode_fork_size(ip, whichfork));
fs/xfs/libxfs/xfs_inode_fork.c
612
linux_to_xfs_dev_t(VFS_I(ip)->i_rdev));
fs/xfs/libxfs/xfs_inode_fork.c
622
switch (ip->i_metatype) {
fs/xfs/libxfs/xfs_inode_fork.c
624
xfs_iflush_rtrmap(ip, dip);
fs/xfs/libxfs/xfs_inode_fork.c
627
xfs_iflush_rtrefcount(ip, dip);
fs/xfs/libxfs/xfs_inode_fork.c
644
struct xfs_inode *ip,
fs/xfs/libxfs/xfs_inode_fork.c
648
return ip->i_cowfp;
fs/xfs/libxfs/xfs_inode_fork.c
650
return &ip->i_af;
fs/xfs/libxfs/xfs_inode_fork.c
651
return &ip->i_df;
fs/xfs/libxfs/xfs_inode_fork.c
659
struct xfs_inode *ip)
fs/xfs/libxfs/xfs_inode_fork.c
661
if (ip->i_cowfp)
fs/xfs/libxfs/xfs_inode_fork.c
664
ip->i_cowfp = kmem_cache_zalloc(xfs_ifork_cache,
fs/xfs/libxfs/xfs_inode_fork.c
666
ip->i_cowfp->if_format = XFS_DINODE_FMT_EXTENTS;
fs/xfs/libxfs/xfs_inode_fork.c
672
struct xfs_inode *ip)
fs/xfs/libxfs/xfs_inode_fork.c
676
switch (VFS_I(ip)->i_mode & S_IFMT) {
fs/xfs/libxfs/xfs_inode_fork.c
678
struct xfs_mount *mp = ip->i_mount;
fs/xfs/libxfs/xfs_inode_fork.c
679
struct xfs_ifork *ifp = xfs_ifork_ptr(ip, XFS_DATA_FORK);
fs/xfs/libxfs/xfs_inode_fork.c
686
struct xfs_ifork *ifp = xfs_ifork_ptr(ip, XFS_DATA_FORK);
fs/xfs/libxfs/xfs_inode_fork.c
696
xfs_inode_verifier_error(ip, -EFSCORRUPTED, "data fork",
fs/xfs/libxfs/xfs_inode_fork.c
697
ip->i_df.if_data, ip->i_df.if_bytes, fa);
fs/xfs/libxfs/xfs_inode_fork.c
707
struct xfs_inode *ip)
fs/xfs/libxfs/xfs_inode_fork.c
709
struct xfs_ifork *ifp = &ip->i_af;
fs/xfs/libxfs/xfs_inode_fork.c
712
if (!xfs_inode_has_attr_fork(ip)) {
fs/xfs/libxfs/xfs_inode_fork.c
715
struct xfs_ifork *ifp = &ip->i_af;
fs/xfs/libxfs/xfs_inode_fork.c
721
xfs_inode_verifier_error(ip, -EFSCORRUPTED, "attr fork",
fs/xfs/libxfs/xfs_inode_fork.c
739
struct xfs_inode *ip,
fs/xfs/libxfs/xfs_inode_fork.c
743
struct xfs_mount *mp = ip->i_mount;
fs/xfs/libxfs/xfs_inode_fork.c
745
xfs_inode_has_large_extent_counts(ip);
fs/xfs/libxfs/xfs_inode_fork.c
746
struct xfs_ifork *ifp = xfs_ifork_ptr(ip, whichfork);
fs/xfs/libxfs/xfs_inode_fork.c
765
ip->i_diflags2 |= XFS_DIFLAG2_NREXT64;
fs/xfs/libxfs/xfs_inode_fork.c
766
xfs_trans_log_inode(tp, ip, XFS_ILOG_CORE);
fs/xfs/libxfs/xfs_inode_fork.c
77
struct xfs_inode *ip,
fs/xfs/libxfs/xfs_inode_fork.c
774
struct xfs_inode *ip,
fs/xfs/libxfs/xfs_inode_fork.c
777
return XFS_IS_REALTIME_INODE(ip) && whichfork != XFS_ATTR_FORK;
fs/xfs/libxfs/xfs_inode_fork.c
87
if (unlikely(size > XFS_DFORK_SIZE(dip, ip->i_mount, whichfork))) {
fs/xfs/libxfs/xfs_inode_fork.c
88
xfs_warn(ip->i_mount,
fs/xfs/libxfs/xfs_inode_fork.c
90
(unsigned long long) ip->i_ino, size,
fs/xfs/libxfs/xfs_inode_fork.c
91
XFS_DFORK_SIZE(dip, ip->i_mount, whichfork));
fs/xfs/libxfs/xfs_inode_fork.c
92
xfs_inode_verifier_error(ip, -EFSCORRUPTED,
fs/xfs/libxfs/xfs_inode_fork.c
95
xfs_inode_mark_sick(ip, XFS_SICK_INO_CORE);
fs/xfs/libxfs/xfs_inode_fork.c
99
xfs_init_local_fork(ip, whichfork, XFS_DFORK_PTR(dip, whichfork), size);
fs/xfs/libxfs/xfs_inode_fork.h
161
void xfs_ifork_zap_attr(struct xfs_inode *ip);
fs/xfs/libxfs/xfs_inode_fork.h
162
void xfs_ifork_init_attr(struct xfs_inode *ip, enum xfs_dinode_fmt format,
fs/xfs/libxfs/xfs_inode_fork.h
164
struct xfs_ifork *xfs_iext_state_to_fork(struct xfs_inode *ip, int state);
fs/xfs/libxfs/xfs_inode_fork.h
171
void * xfs_idata_realloc(struct xfs_inode *ip, int64_t byte_diff,
fs/xfs/libxfs/xfs_inode_fork.h
181
void xfs_init_local_fork(struct xfs_inode *ip, int whichfork,
fs/xfs/libxfs/xfs_inode_fork.h
194
bool xfs_iext_lookup_extent(struct xfs_inode *ip,
fs/xfs/libxfs/xfs_inode_fork.h
198
bool xfs_iext_lookup_extent_before(struct xfs_inode *ip,
fs/xfs/libxfs/xfs_inode_fork.h
205
void xfs_iext_update_extent(struct xfs_inode *ip, int state,
fs/xfs/libxfs/xfs_inode_fork.h
259
extern void xfs_ifork_init_cow(struct xfs_inode *ip);
fs/xfs/libxfs/xfs_inode_fork.h
261
int xfs_ifork_verify_local_data(struct xfs_inode *ip);
fs/xfs/libxfs/xfs_inode_fork.h
262
int xfs_ifork_verify_local_attr(struct xfs_inode *ip);
fs/xfs/libxfs/xfs_inode_fork.h
263
int xfs_iext_count_extend(struct xfs_trans *tp, struct xfs_inode *ip,
fs/xfs/libxfs/xfs_inode_fork.h
265
bool xfs_ifork_is_realtime(struct xfs_inode *ip, int whichfork);
fs/xfs/libxfs/xfs_inode_fork.h
79
#define XFS_IFORK_MAXEXT(ip, w) \
fs/xfs/libxfs/xfs_inode_fork.h
80
(xfs_inode_fork_size(ip, w) / sizeof(xfs_bmbt_rec_t))
fs/xfs/libxfs/xfs_inode_util.c
100
if (ip->i_diflags & XFS_DIFLAG_APPEND)
fs/xfs/libxfs/xfs_inode_util.c
102
if (ip->i_diflags & XFS_DIFLAG_SYNC)
fs/xfs/libxfs/xfs_inode_util.c
104
if (ip->i_diflags & XFS_DIFLAG_NOATIME)
fs/xfs/libxfs/xfs_inode_util.c
106
if (ip->i_diflags & XFS_DIFLAG_NODUMP)
fs/xfs/libxfs/xfs_inode_util.c
108
if (ip->i_diflags & XFS_DIFLAG_RTINHERIT)
fs/xfs/libxfs/xfs_inode_util.c
110
if (ip->i_diflags & XFS_DIFLAG_PROJINHERIT)
fs/xfs/libxfs/xfs_inode_util.c
112
if (ip->i_diflags & XFS_DIFLAG_NOSYMLINKS)
fs/xfs/libxfs/xfs_inode_util.c
114
if (ip->i_diflags & XFS_DIFLAG_EXTSIZE)
fs/xfs/libxfs/xfs_inode_util.c
116
if (ip->i_diflags & XFS_DIFLAG_EXTSZINHERIT)
fs/xfs/libxfs/xfs_inode_util.c
118
if (ip->i_diflags & XFS_DIFLAG_NODEFRAG)
fs/xfs/libxfs/xfs_inode_util.c
120
if (ip->i_diflags & XFS_DIFLAG_FILESTREAM)
fs/xfs/libxfs/xfs_inode_util.c
124
if (ip->i_diflags2 & XFS_DIFLAG2_ANY) {
fs/xfs/libxfs/xfs_inode_util.c
125
if (ip->i_diflags2 & XFS_DIFLAG2_DAX)
fs/xfs/libxfs/xfs_inode_util.c
127
if (ip->i_diflags2 & XFS_DIFLAG2_COWEXTSIZE)
fs/xfs/libxfs/xfs_inode_util.c
131
if (xfs_inode_has_attr_fork(ip))
fs/xfs/libxfs/xfs_inode_util.c
149
struct xfs_inode *ip,
fs/xfs/libxfs/xfs_inode_util.c
154
umode_t mode = VFS_I(ip)->i_mode;
fs/xfs/libxfs/xfs_inode_util.c
161
ip->i_extsize = pip->i_extsize;
fs/xfs/libxfs/xfs_inode_util.c
167
xfs_has_realtime(ip->i_mount))
fs/xfs/libxfs/xfs_inode_util.c
171
ip->i_extsize = pip->i_extsize;
fs/xfs/libxfs/xfs_inode_util.c
192
ip->i_diflags |= di_flags;
fs/xfs/libxfs/xfs_inode_util.c
204
failaddr = xfs_inode_validate_extsize(ip->i_mount, ip->i_extsize,
fs/xfs/libxfs/xfs_inode_util.c
205
VFS_I(ip)->i_mode, ip->i_diflags);
fs/xfs/libxfs/xfs_inode_util.c
207
ip->i_diflags &= ~(XFS_DIFLAG_EXTSIZE |
fs/xfs/libxfs/xfs_inode_util.c
209
ip->i_extsize = 0;
fs/xfs/libxfs/xfs_inode_util.c
216
struct xfs_inode *ip,
fs/xfs/libxfs/xfs_inode_util.c
222
ip->i_diflags2 |= XFS_DIFLAG2_COWEXTSIZE;
fs/xfs/libxfs/xfs_inode_util.c
223
ip->i_cowextsize = pip->i_cowextsize;
fs/xfs/libxfs/xfs_inode_util.c
226
ip->i_diflags2 |= XFS_DIFLAG2_DAX;
fs/xfs/libxfs/xfs_inode_util.c
228
ip->i_diflags2 |= XFS_DIFLAG2_METADATA;
fs/xfs/libxfs/xfs_inode_util.c
231
failaddr = xfs_inode_validate_cowextsize(ip->i_mount, ip->i_cowextsize,
fs/xfs/libxfs/xfs_inode_util.c
232
VFS_I(ip)->i_mode, ip->i_diflags, ip->i_diflags2);
fs/xfs/libxfs/xfs_inode_util.c
234
ip->i_diflags2 &= ~XFS_DIFLAG2_COWEXTSIZE;
fs/xfs/libxfs/xfs_inode_util.c
235
ip->i_cowextsize = 0;
fs/xfs/libxfs/xfs_inode_util.c
269
struct xfs_inode *ip)
fs/xfs/libxfs/xfs_inode_util.c
274
struct inode *inode = VFS_I(ip);
fs/xfs/libxfs/xfs_inode_util.c
29
struct xfs_inode *ip,
fs/xfs/libxfs/xfs_inode_util.c
291
ip->i_projid = 0;
fs/xfs/libxfs/xfs_inode_util.c
302
ip->i_projid = xfs_get_initial_prid(pip);
fs/xfs/libxfs/xfs_inode_util.c
305
ip->i_disk_size = 0;
fs/xfs/libxfs/xfs_inode_util.c
306
ip->i_df.if_nextents = 0;
fs/xfs/libxfs/xfs_inode_util.c
307
ASSERT(ip->i_nblocks == 0);
fs/xfs/libxfs/xfs_inode_util.c
309
ip->i_extsize = 0;
fs/xfs/libxfs/xfs_inode_util.c
310
ip->i_diflags = 0;
fs/xfs/libxfs/xfs_inode_util.c
315
ip->i_cowextsize = 0;
fs/xfs/libxfs/xfs_inode_util.c
319
xfs_trans_ichgtime(tp, ip, times);
fs/xfs/libxfs/xfs_inode_util.c
327
ip->i_df.if_format = XFS_DINODE_FMT_DEV;
fs/xfs/libxfs/xfs_inode_util.c
333
xfs_inode_inherit_flags(ip, pip);
fs/xfs/libxfs/xfs_inode_util.c
335
xfs_inode_inherit_flags2(ip, pip);
fs/xfs/libxfs/xfs_inode_util.c
338
ip->i_df.if_format = XFS_DINODE_FMT_EXTENTS;
fs/xfs/libxfs/xfs_inode_util.c
339
ip->i_df.if_bytes = 0;
fs/xfs/libxfs/xfs_inode_util.c
34
(ip->i_diflags & XFS_DIFLAG_PREALLOC);
fs/xfs/libxfs/xfs_inode_util.c
340
ip->i_df.if_data = NULL;
fs/xfs/libxfs/xfs_inode_util.c
347
ip->i_forkoff = xfs_default_attroffset(ip) >> 3;
fs/xfs/libxfs/xfs_inode_util.c
348
xfs_ifork_init_attr(ip, XFS_DINODE_FMT_EXTENTS, 0);
fs/xfs/libxfs/xfs_inode_util.c
358
xfs_trans_log_inode(tp, ip, flags);
fs/xfs/libxfs/xfs_inode_util.c
404
struct xfs_inode *ip;
fs/xfs/libxfs/xfs_inode_util.c
410
ip = xfs_iunlink_lookup(pag, next_agino);
fs/xfs/libxfs/xfs_inode_util.c
411
if (!ip)
fs/xfs/libxfs/xfs_inode_util.c
414
ip->i_prev_unlinked = prev_agino;
fs/xfs/libxfs/xfs_inode_util.c
463
struct xfs_inode *ip)
fs/xfs/libxfs/xfs_inode_util.c
468
xfs_agino_t agino = XFS_INO_TO_AGINO(mp, ip->i_ino);
fs/xfs/libxfs/xfs_inode_util.c
50
if (S_ISDIR(VFS_I(ip)->i_mode)) {
fs/xfs/libxfs/xfs_inode_util.c
500
error = xfs_iunlink_log_inode(tp, ip, pag, next_agino);
fs/xfs/libxfs/xfs_inode_util.c
503
ip->i_next_unlinked = next_agino;
fs/xfs/libxfs/xfs_inode_util.c
507
ip->i_prev_unlinked = NULLAGINO;
fs/xfs/libxfs/xfs_inode_util.c
521
struct xfs_inode *ip)
fs/xfs/libxfs/xfs_inode_util.c
528
ASSERT(VFS_I(ip)->i_nlink == 0);
fs/xfs/libxfs/xfs_inode_util.c
529
ASSERT(VFS_I(ip)->i_mode != 0);
fs/xfs/libxfs/xfs_inode_util.c
530
trace_xfs_iunlink(ip);
fs/xfs/libxfs/xfs_inode_util.c
532
pag = xfs_perag_get(mp, XFS_INO_TO_AGNO(mp, ip->i_ino));
fs/xfs/libxfs/xfs_inode_util.c
539
error = xfs_iunlink_insert_inode(tp, pag, agibp, ip);
fs/xfs/libxfs/xfs_inode_util.c
550
struct xfs_inode *ip)
fs/xfs/libxfs/xfs_inode_util.c
554
xfs_agino_t agino = XFS_INO_TO_AGINO(mp, ip->i_ino);
fs/xfs/libxfs/xfs_inode_util.c
559
trace_xfs_iunlink_remove(ip);
fs/xfs/libxfs/xfs_inode_util.c
578
error = xfs_iunlink_log_inode(tp, ip, pag, NULLAGINO);
fs/xfs/libxfs/xfs_inode_util.c
586
error = xfs_iunlink_update_backref(pag, ip->i_prev_unlinked,
fs/xfs/libxfs/xfs_inode_util.c
587
ip->i_next_unlinked);
fs/xfs/libxfs/xfs_inode_util.c
589
error = xfs_iunlink_reload_next(tp, agibp, ip->i_prev_unlinked,
fs/xfs/libxfs/xfs_inode_util.c
59
} else if (S_ISREG(VFS_I(ip)->i_mode)) {
fs/xfs/libxfs/xfs_inode_util.c
590
ip->i_next_unlinked);
fs/xfs/libxfs/xfs_inode_util.c
597
prev_ip = xfs_iunlink_lookup(pag, ip->i_prev_unlinked);
fs/xfs/libxfs/xfs_inode_util.c
599
xfs_inode_mark_sick(ip, XFS_SICK_INO_CORE);
fs/xfs/libxfs/xfs_inode_util.c
604
ip->i_next_unlinked);
fs/xfs/libxfs/xfs_inode_util.c
605
prev_ip->i_next_unlinked = ip->i_next_unlinked;
fs/xfs/libxfs/xfs_inode_util.c
609
ip->i_next_unlinked);
fs/xfs/libxfs/xfs_inode_util.c
612
ip->i_next_unlinked = NULLAGINO;
fs/xfs/libxfs/xfs_inode_util.c
613
ip->i_prev_unlinked = 0;
fs/xfs/libxfs/xfs_inode_util.c
624
struct xfs_inode *ip)
fs/xfs/libxfs/xfs_inode_util.c
629
trace_xfs_iunlink_remove(ip);
fs/xfs/libxfs/xfs_inode_util.c
636
return xfs_iunlink_remove_inode(tp, pag, agibp, ip);
fs/xfs/libxfs/xfs_inode_util.c
647
struct xfs_inode *ip)
fs/xfs/libxfs/xfs_inode_util.c
649
struct inode *inode = VFS_I(ip);
fs/xfs/libxfs/xfs_inode_util.c
651
xfs_trans_ichgtime(tp, ip, XFS_ICHGTIME_CHG);
fs/xfs/libxfs/xfs_inode_util.c
656
ip->i_ino);
fs/xfs/libxfs/xfs_inode_util.c
662
xfs_trans_log_inode(tp, ip, XFS_ILOG_CORE);
fs/xfs/libxfs/xfs_inode_util.c
667
return xfs_iunlink(tp, ip);
fs/xfs/libxfs/xfs_inode_util.c
676
struct xfs_inode *ip)
fs/xfs/libxfs/xfs_inode_util.c
678
struct inode *inode = VFS_I(ip);
fs/xfs/libxfs/xfs_inode_util.c
680
xfs_trans_ichgtime(tp, ip, XFS_ICHGTIME_CHG);
fs/xfs/libxfs/xfs_inode_util.c
685
ip->i_ino);
fs/xfs/libxfs/xfs_inode_util.c
689
xfs_trans_log_inode(tp, ip, XFS_ILOG_CORE);
fs/xfs/libxfs/xfs_inode_util.c
697
struct xfs_inode *ip,
fs/xfs/libxfs/xfs_inode_util.c
700
struct xfs_mount *mp = ip->i_mount;
fs/xfs/libxfs/xfs_inode_util.c
709
error = xfs_difree(tp, pag, ip->i_ino, xic);
fs/xfs/libxfs/xfs_inode_util.c
71
struct xfs_inode *ip,
fs/xfs/libxfs/xfs_inode_util.c
713
error = xfs_iunlink_remove(tp, pag, ip);
fs/xfs/libxfs/xfs_inode_util.c
722
if (ip->i_df.if_format == XFS_DINODE_FMT_LOCAL) {
fs/xfs/libxfs/xfs_inode_util.c
723
kfree(ip->i_df.if_data);
fs/xfs/libxfs/xfs_inode_util.c
724
ip->i_df.if_data = NULL;
fs/xfs/libxfs/xfs_inode_util.c
725
ip->i_df.if_bytes = 0;
fs/xfs/libxfs/xfs_inode_util.c
728
VFS_I(ip)->i_mode = 0; /* mark incore inode as free */
fs/xfs/libxfs/xfs_inode_util.c
729
ip->i_diflags = 0;
fs/xfs/libxfs/xfs_inode_util.c
730
ip->i_diflags2 = mp->m_ino_geo.new_diflags2;
fs/xfs/libxfs/xfs_inode_util.c
731
ip->i_forkoff = 0; /* mark the attr fork not in use */
fs/xfs/libxfs/xfs_inode_util.c
732
ip->i_df.if_format = XFS_DINODE_FMT_EXTENTS;
fs/xfs/libxfs/xfs_inode_util.c
738
VFS_I(ip)->i_generation++;
fs/xfs/libxfs/xfs_inode_util.c
739
xfs_trans_log_inode(tp, ip, XFS_ILOG_CORE);
fs/xfs/libxfs/xfs_inode_util.c
75
(ip->i_diflags2 & (XFS_DIFLAG2_REFLINK |
fs/xfs/libxfs/xfs_inode_util.c
89
struct xfs_inode *ip)
fs/xfs/libxfs/xfs_inode_util.c
93
if (ip->i_diflags & XFS_DIFLAG_ANY) {
fs/xfs/libxfs/xfs_inode_util.c
94
if (ip->i_diflags & XFS_DIFLAG_REALTIME)
fs/xfs/libxfs/xfs_inode_util.c
96
if (ip->i_diflags & XFS_DIFLAG_PREALLOC)
fs/xfs/libxfs/xfs_inode_util.c
98
if (ip->i_diflags & XFS_DIFLAG_IMMUTABLE)
fs/xfs/libxfs/xfs_inode_util.h
11
uint16_t xfs_flags2diflags(struct xfs_inode *ip, unsigned int xflags);
fs/xfs/libxfs/xfs_inode_util.h
12
uint64_t xfs_flags2diflags2(struct xfs_inode *ip, unsigned int xflags);
fs/xfs/libxfs/xfs_inode_util.h
13
uint32_t xfs_dic2xflags(struct xfs_inode *ip);
fs/xfs/libxfs/xfs_inode_util.h
14
uint32_t xfs_ip2xflags(struct xfs_inode *ip);
fs/xfs/libxfs/xfs_inode_util.h
48
void xfs_trans_ichgtime(struct xfs_trans *tp, struct xfs_inode *ip, int flags);
fs/xfs/libxfs/xfs_inode_util.h
51
struct xfs_inode *ip);
fs/xfs/libxfs/xfs_inode_util.h
54
struct xfs_inode *ip, struct xfs_icluster *xic);
fs/xfs/libxfs/xfs_inode_util.h
56
int xfs_iunlink(struct xfs_trans *tp, struct xfs_inode *ip);
fs/xfs/libxfs/xfs_inode_util.h
58
struct xfs_inode *ip);
fs/xfs/libxfs/xfs_inode_util.h
59
int xfs_droplink(struct xfs_trans *tp, struct xfs_inode *ip);
fs/xfs/libxfs/xfs_inode_util.h
60
void xfs_bumplink(struct xfs_trans *tp, struct xfs_inode *ip);
fs/xfs/libxfs/xfs_metadir.c
170
if (upd->ip) {
fs/xfs/libxfs/xfs_metadir.c
172
xfs_iunlock(upd->ip, XFS_ILOCK_EXCL);
fs/xfs/libxfs/xfs_metadir.c
193
ASSERT(upd->ip == NULL);
fs/xfs/libxfs/xfs_metadir.c
282
error = xfs_icreate(upd->tp, ino, &args, &upd->ip);
fs/xfs/libxfs/xfs_metadir.c
285
du.ip = upd->ip;
fs/xfs/libxfs/xfs_metadir.c
286
xfs_metafile_set_iflag(upd->tp, upd->ip, upd->metafile_type);
fs/xfs/libxfs/xfs_metadir.c
330
ASSERT(upd->ip != NULL);
fs/xfs/libxfs/xfs_metadir.c
338
error = xfs_trans_alloc_dir(upd->dp, &M_RES(mp)->tr_link, upd->ip,
fs/xfs/libxfs/xfs_metadir.c
347
xfs_iunlock(upd->ip, XFS_ILOCK_EXCL);
fs/xfs/libxfs/xfs_metadir.c
375
.ip = upd->ip,
fs/xfs/libxfs/xfs_metadir.c
384
xfs_assert_ilocked(upd->ip, XFS_ILOCK_EXCL);
fs/xfs/libxfs/xfs_metadir.c
388
xfs_mode_to_ftype(VFS_I(upd->ip)->i_mode));
fs/xfs/libxfs/xfs_metadir.c
472
xfs_finish_inode_setup(upd.ip);
fs/xfs/libxfs/xfs_metadir.c
473
*ipp = upd.ip;
fs/xfs/libxfs/xfs_metadir.c
480
if (upd.ip) {
fs/xfs/libxfs/xfs_metadir.c
481
xfs_finish_inode_setup(upd.ip);
fs/xfs/libxfs/xfs_metadir.c
482
xfs_irele(upd.ip);
fs/xfs/libxfs/xfs_metadir.h
21
struct xfs_inode *ip;
fs/xfs/libxfs/xfs_metafile.c
135
struct xfs_inode *ip,
fs/xfs/libxfs/xfs_metafile.c
138
struct xfs_mount *mp = ip->i_mount;
fs/xfs/libxfs/xfs_metafile.c
141
ASSERT(xfs_is_metadir_inode(ip));
fs/xfs/libxfs/xfs_metafile.c
156
xfs_mod_delalloc(ip, 0, -from_resv);
fs/xfs/libxfs/xfs_metafile.c
173
error = xfs_dec_fdblocks(ip->i_mount, len, true);
fs/xfs/libxfs/xfs_metafile.c
185
ip->i_nblocks += args->len;
fs/xfs/libxfs/xfs_metafile.c
186
xfs_trans_log_inode(args->tp, ip, XFS_ILOG_CORE);
fs/xfs/libxfs/xfs_metafile.c
192
struct xfs_inode *ip,
fs/xfs/libxfs/xfs_metafile.c
196
struct xfs_mount *mp = ip->i_mount;
fs/xfs/libxfs/xfs_metafile.c
199
ASSERT(xfs_is_metadir_inode(ip));
fs/xfs/libxfs/xfs_metafile.c
203
ip->i_nblocks -= len;
fs/xfs/libxfs/xfs_metafile.c
204
xfs_trans_log_inode(tp, ip, XFS_ILOG_CORE);
fs/xfs/libxfs/xfs_metafile.c
218
xfs_mod_delalloc(ip, 0, to_resv);
fs/xfs/libxfs/xfs_metafile.c
50
struct xfs_inode *ip,
fs/xfs/libxfs/xfs_metafile.c
53
VFS_I(ip)->i_mode &= ~0777;
fs/xfs/libxfs/xfs_metafile.c
54
VFS_I(ip)->i_uid = GLOBAL_ROOT_UID;
fs/xfs/libxfs/xfs_metafile.c
55
VFS_I(ip)->i_gid = GLOBAL_ROOT_GID;
fs/xfs/libxfs/xfs_metafile.c
56
if (S_ISDIR(VFS_I(ip)->i_mode))
fs/xfs/libxfs/xfs_metafile.c
57
ip->i_diflags |= XFS_METADIR_DIFLAGS;
fs/xfs/libxfs/xfs_metafile.c
59
ip->i_diflags |= XFS_METAFILE_DIFLAGS;
fs/xfs/libxfs/xfs_metafile.c
60
ip->i_diflags2 &= ~XFS_DIFLAG2_DAX;
fs/xfs/libxfs/xfs_metafile.c
61
ip->i_diflags2 |= XFS_DIFLAG2_METADATA;
fs/xfs/libxfs/xfs_metafile.c
62
ip->i_metatype = metafile_type;
fs/xfs/libxfs/xfs_metafile.c
63
xfs_trans_log_inode(tp, ip, XFS_ILOG_CORE);
fs/xfs/libxfs/xfs_metafile.c
65
XFS_STATS_DEC(ip->i_mount, xs_inodes_active);
fs/xfs/libxfs/xfs_metafile.c
66
XFS_STATS_INC(ip->i_mount, xs_inodes_meta);
fs/xfs/libxfs/xfs_metafile.c
73
struct xfs_inode *ip)
fs/xfs/libxfs/xfs_metafile.c
75
ASSERT(xfs_is_metadir_inode(ip));
fs/xfs/libxfs/xfs_metafile.c
76
ASSERT(VFS_I(ip)->i_nlink == 0);
fs/xfs/libxfs/xfs_metafile.c
78
ip->i_diflags2 &= ~XFS_DIFLAG2_METADATA;
fs/xfs/libxfs/xfs_metafile.c
79
xfs_trans_log_inode(tp, ip, XFS_ILOG_CORE);
fs/xfs/libxfs/xfs_metafile.c
80
XFS_STATS_INC(ip->i_mount, xs_inodes_active);
fs/xfs/libxfs/xfs_metafile.c
81
XFS_STATS_DEC(ip->i_mount, xs_inodes_meta);
fs/xfs/libxfs/xfs_metafile.h
22
void xfs_metafile_set_iflag(struct xfs_trans *tp, struct xfs_inode *ip,
fs/xfs/libxfs/xfs_metafile.h
24
void xfs_metafile_clear_iflag(struct xfs_trans *tp, struct xfs_inode *ip);
fs/xfs/libxfs/xfs_metafile.h
30
void xfs_metafile_resv_alloc_space(struct xfs_inode *ip,
fs/xfs/libxfs/xfs_metafile.h
32
void xfs_metafile_resv_free_space(struct xfs_inode *ip, struct xfs_trans *tp,
fs/xfs/libxfs/xfs_parent.c
309
struct xfs_inode *ip,
fs/xfs/libxfs/xfs_parent.c
315
xfs_parent_da_args_init(scratch, tp, pptr, ip, ip->i_ino, parent_name);
fs/xfs/libxfs/xfs_parent.c
344
struct xfs_inode *ip,
fs/xfs/libxfs/xfs_parent.c
350
if (!xfs_parent_sanity_check(ip->i_mount, parent_name, pptr)) {
fs/xfs/libxfs/xfs_parent.c
356
xfs_parent_da_args_init(scratch, NULL, pptr, ip, owner, parent_name);
fs/xfs/libxfs/xfs_parent.c
367
struct xfs_inode *ip,
fs/xfs/libxfs/xfs_parent.c
373
if (!xfs_parent_sanity_check(ip->i_mount, parent_name, pptr)) {
fs/xfs/libxfs/xfs_parent.c
379
xfs_parent_da_args_init(scratch, NULL, pptr, ip, owner, parent_name);
fs/xfs/libxfs/xfs_parent.h
100
int xfs_parent_lookup(struct xfs_trans *tp, struct xfs_inode *ip,
fs/xfs/libxfs/xfs_parent.h
103
int xfs_parent_set(struct xfs_inode *ip, xfs_ino_t owner,
fs/xfs/libxfs/xfs_parent.h
106
int xfs_parent_unset(struct xfs_inode *ip, xfs_ino_t owner,
fs/xfs/libxfs/xfs_quota_defs.h
183
struct xfs_inode *ip);
fs/xfs/libxfs/xfs_rmap.c
2770
struct xfs_inode *ip,
fs/xfs/libxfs/xfs_rmap.c
2775
bool isrt = xfs_ifork_is_realtime(ip, whichfork);
fs/xfs/libxfs/xfs_rmap.c
2780
if (whichfork != XFS_ATTR_FORK && xfs_is_reflink_inode(ip))
fs/xfs/libxfs/xfs_rmap.c
2783
__xfs_rmap_add(tp, type, ip->i_ino, isrt, whichfork, PREV);
fs/xfs/libxfs/xfs_rmap.c
2790
struct xfs_inode *ip,
fs/xfs/libxfs/xfs_rmap.c
2795
bool isrt = xfs_ifork_is_realtime(ip, whichfork);
fs/xfs/libxfs/xfs_rmap.c
2800
if (whichfork != XFS_ATTR_FORK && xfs_is_reflink_inode(ip))
fs/xfs/libxfs/xfs_rmap.c
2803
__xfs_rmap_add(tp, type, ip->i_ino, isrt, whichfork, PREV);
fs/xfs/libxfs/xfs_rmap.c
2816
struct xfs_inode *ip,
fs/xfs/libxfs/xfs_rmap.c
2821
bool isrt = xfs_ifork_is_realtime(ip, whichfork);
fs/xfs/libxfs/xfs_rmap.c
2826
if (whichfork != XFS_ATTR_FORK && xfs_is_reflink_inode(ip))
fs/xfs/libxfs/xfs_rmap.c
2829
__xfs_rmap_add(tp, type, ip->i_ino, isrt, whichfork, PREV);
fs/xfs/libxfs/xfs_rmap.h
182
void xfs_rmap_map_extent(struct xfs_trans *tp, struct xfs_inode *ip,
fs/xfs/libxfs/xfs_rmap.h
184
void xfs_rmap_unmap_extent(struct xfs_trans *tp, struct xfs_inode *ip,
fs/xfs/libxfs/xfs_rmap.h
187
struct xfs_inode *ip, int whichfork,
fs/xfs/libxfs/xfs_rtbitmap.c
1329
struct xfs_inode *ip,
fs/xfs/libxfs/xfs_rtbitmap.c
1334
struct xfs_mount *mp = ip->i_mount;
fs/xfs/libxfs/xfs_rtbitmap.c
1344
xfs_ilock(ip, XFS_ILOCK_EXCL);
fs/xfs/libxfs/xfs_rtbitmap.c
1345
xfs_trans_ijoin(tp, ip, XFS_ILOCK_EXCL);
fs/xfs/libxfs/xfs_rtbitmap.c
1347
error = xfs_iext_count_extend(tp, ip, XFS_DATA_FORK,
fs/xfs/libxfs/xfs_rtbitmap.c
1352
error = xfs_bmapi_write(tp, ip, offset_fsb, count_fsb,
fs/xfs/libxfs/xfs_rtbitmap.c
1373
struct xfs_inode *ip = rtg->rtg_inodes[type];
fs/xfs/libxfs/xfs_rtbitmap.c
1391
xfs_ilock(ip, XFS_ILOCK_EXCL);
fs/xfs/libxfs/xfs_rtbitmap.c
1392
xfs_trans_ijoin(tp, ip, XFS_ILOCK_EXCL);
fs/xfs/libxfs/xfs_rtbitmap.c
1412
hdr->rt_owner = cpu_to_be64(ip->i_ino);
fs/xfs/libxfs/xfs_rtbitmap.c
1477
struct xfs_inode *ip,
fs/xfs/libxfs/xfs_rtbitmap.c
1483
ip->i_disk_size = mp->m_sb.sb_rbmblocks * mp->m_sb.sb_blocksize;
fs/xfs/libxfs/xfs_rtbitmap.c
1485
ip->i_diflags |= XFS_DIFLAG_NEWRTBM;
fs/xfs/libxfs/xfs_rtbitmap.c
1486
inode_set_atime(VFS_I(ip), 0, 0);
fs/xfs/libxfs/xfs_rtbitmap.c
1488
xfs_trans_log_inode(tp, ip, XFS_ILOG_CORE);
fs/xfs/libxfs/xfs_rtbitmap.c
1495
struct xfs_inode *ip,
fs/xfs/libxfs/xfs_rtbitmap.c
1501
ip->i_disk_size = mp->m_rsumblocks * mp->m_sb.sb_blocksize;
fs/xfs/libxfs/xfs_rtbitmap.c
1502
xfs_trans_log_inode(tp, ip, XFS_ILOG_CORE);
fs/xfs/libxfs/xfs_rtbitmap.c
155
struct xfs_inode *ip = args->rtg->rtg_inodes[type];
fs/xfs/libxfs/xfs_rtbitmap.c
195
error = xfs_bmapi_read(ip, block, 1, &map, &nmap, 0);
fs/xfs/libxfs/xfs_rtbitmap.c
217
if (hdr->rt_owner != cpu_to_be64(ip->i_ino)) {
fs/xfs/libxfs/xfs_rtbitmap.h
425
int xfs_rtbitmap_create(struct xfs_rtgroup *rtg, struct xfs_inode *ip,
fs/xfs/libxfs/xfs_rtbitmap.h
427
int xfs_rtsummary_create(struct xfs_rtgroup *rtg, struct xfs_inode *ip,
fs/xfs/libxfs/xfs_rtgroup.c
308
const struct xfs_inode *ip =
fs/xfs/libxfs/xfs_rtgroup.c
311
printk(KERN_CONT " rgno=%u metatype=%s", ip->i_projid,
fs/xfs/libxfs/xfs_rtgroup.c
312
xfs_metafile_type_str(ip->i_metatype));
fs/xfs/libxfs/xfs_rtgroup.c
324
struct xfs_inode *ip,
fs/xfs/libxfs/xfs_rtgroup.c
328
lockdep_set_class_and_subclass(&ip->i_lock, &xfs_rtginode_lock_class,
fs/xfs/libxfs/xfs_rtgroup.c
330
lock_set_cmp_fn(&ip->i_lock, xfs_rtginode_ilock_cmp_fn,
fs/xfs/libxfs/xfs_rtgroup.c
334
#define xfs_rtginode_lockdep_setup(ip, rgno, type) do { } while (0)
fs/xfs/libxfs/xfs_rtgroup.c
351
struct xfs_inode *ip,
fs/xfs/libxfs/xfs_rtgroup.c
447
struct xfs_inode *ip;
fs/xfs/libxfs/xfs_rtgroup.c
470
&ip);
fs/xfs/libxfs/xfs_rtgroup.c
483
ops->metafile_type, &ip);
fs/xfs/libxfs/xfs_rtgroup.c
493
if (XFS_IS_CORRUPT(mp, !((1U << ip->i_df.if_format) & ops->fmt_mask))) {
fs/xfs/libxfs/xfs_rtgroup.c
494
xfs_irele(ip);
fs/xfs/libxfs/xfs_rtgroup.c
499
if (XFS_IS_CORRUPT(mp, ip->i_projid != rtg_rgno(rtg))) {
fs/xfs/libxfs/xfs_rtgroup.c
500
xfs_irele(ip);
fs/xfs/libxfs/xfs_rtgroup.c
505
xfs_rtginode_lockdep_setup(ip, rtg_rgno(rtg), type);
fs/xfs/libxfs/xfs_rtgroup.c
506
rtg->rtg_inodes[type] = ip;
fs/xfs/libxfs/xfs_rtgroup.c
555
xfs_rtginode_lockdep_setup(upd.ip, rtg_rgno(rtg), type);
fs/xfs/libxfs/xfs_rtgroup.c
557
upd.ip->i_projid = rtg_rgno(rtg);
fs/xfs/libxfs/xfs_rtgroup.c
558
error = ops->create(rtg, upd.ip, upd.tp, init);
fs/xfs/libxfs/xfs_rtgroup.c
567
xfs_finish_inode_setup(upd.ip);
fs/xfs/libxfs/xfs_rtgroup.c
568
rtg->rtg_inodes[type] = upd.ip;
fs/xfs/libxfs/xfs_rtgroup.c
574
if (upd.ip) {
fs/xfs/libxfs/xfs_rtgroup.c
575
xfs_finish_inode_setup(upd.ip);
fs/xfs/libxfs/xfs_rtgroup.c
576
xfs_irele(upd.ip);
fs/xfs/libxfs/xfs_rtrefcount_btree.c
363
xfs_inode_fork_size(cur->bc_ino.ip, cur->bc_ino.whichfork));
fs/xfs/libxfs/xfs_rtrefcount_btree.c
405
struct xfs_inode *ip = rtg_refcount(rtg);
fs/xfs/libxfs/xfs_rtrefcount_btree.c
409
xfs_assert_ilocked(ip, XFS_ILOCK_SHARED | XFS_ILOCK_EXCL);
fs/xfs/libxfs/xfs_rtrefcount_btree.c
414
cur->bc_ino.ip = ip;
fs/xfs/libxfs/xfs_rtrefcount_btree.c
418
cur->bc_nlevels = be16_to_cpu(ip->i_df.if_broot->bb_level) + 1;
fs/xfs/libxfs/xfs_rtrefcount_btree.c
419
cur->bc_ino.forksize = xfs_inode_fork_size(ip, XFS_DATA_FORK);
fs/xfs/libxfs/xfs_rtrefcount_btree.c
445
ifp = xfs_ifork_ptr(cur->bc_ino.ip, XFS_DATA_FORK);
fs/xfs/libxfs/xfs_rtrefcount_btree.c
449
cur->bc_ino.ip->i_projid = cur->bc_group->xg_gno;
fs/xfs/libxfs/xfs_rtrefcount_btree.c
450
xfs_trans_log_inode(tp, cur->bc_ino.ip, flags);
fs/xfs/libxfs/xfs_rtrefcount_btree.c
586
struct xfs_inode *ip,
fs/xfs/libxfs/xfs_rtrefcount_btree.c
591
struct xfs_mount *mp = ip->i_mount;
fs/xfs/libxfs/xfs_rtrefcount_btree.c
605
ip->i_ino);
fs/xfs/libxfs/xfs_rtrefcount_btree.c
630
struct xfs_inode *ip,
fs/xfs/libxfs/xfs_rtrefcount_btree.c
633
struct xfs_mount *mp = ip->i_mount;
fs/xfs/libxfs/xfs_rtrefcount_btree.c
645
if (!xfs_has_reflink(ip->i_mount)) {
fs/xfs/libxfs/xfs_rtrefcount_btree.c
646
xfs_inode_mark_sick(ip, XFS_SICK_INO_CORE);
fs/xfs/libxfs/xfs_rtrefcount_btree.c
656
xfs_inode_mark_sick(ip, XFS_SICK_INO_CORE);
fs/xfs/libxfs/xfs_rtrefcount_btree.c
660
broot = xfs_broot_alloc(xfs_ifork_ptr(ip, XFS_DATA_FORK),
fs/xfs/libxfs/xfs_rtrefcount_btree.c
663
xfs_rtrefcountbt_from_disk(ip, dfp, dsize, broot);
fs/xfs/libxfs/xfs_rtrefcount_btree.c
716
struct xfs_inode *ip,
fs/xfs/libxfs/xfs_rtrefcount_btree.c
719
struct xfs_ifork *ifp = xfs_ifork_ptr(ip, XFS_DATA_FORK);
fs/xfs/libxfs/xfs_rtrefcount_btree.c
725
xfs_inode_fork_size(ip, XFS_DATA_FORK));
fs/xfs/libxfs/xfs_rtrefcount_btree.c
726
xfs_rtrefcountbt_to_disk(ip->i_mount, ifp->if_broot,
fs/xfs/libxfs/xfs_rtrefcount_btree.c
728
XFS_DFORK_SIZE(dip, ip->i_mount, XFS_DATA_FORK));
fs/xfs/libxfs/xfs_rtrefcount_btree.c
737
struct xfs_inode *ip,
fs/xfs/libxfs/xfs_rtrefcount_btree.c
741
struct xfs_ifork *ifp = xfs_ifork_ptr(ip, XFS_DATA_FORK);
fs/xfs/libxfs/xfs_rtrefcount_btree.c
742
struct xfs_mount *mp = ip->i_mount;
fs/xfs/libxfs/xfs_rtrefcount_btree.c
754
ip->i_ino);
fs/xfs/libxfs/xfs_rtrefcount_btree.c
755
xfs_trans_log_inode(tp, ip, XFS_ILOG_CORE | XFS_ILOG_DBROOT);
fs/xfs/libxfs/xfs_rtrefcount_btree.h
180
int xfs_iformat_rtrefcount(struct xfs_inode *ip, struct xfs_dinode *dip);
fs/xfs/libxfs/xfs_rtrefcount_btree.h
184
void xfs_iflush_rtrefcount(struct xfs_inode *ip, struct xfs_dinode *dip);
fs/xfs/libxfs/xfs_rtrefcount_btree.h
186
int xfs_rtrefcountbt_create(struct xfs_rtgroup *rtg, struct xfs_inode *ip,
fs/xfs/libxfs/xfs_rtrefcount_btree.h
21
struct xfs_rtgroup *rtg, struct xfs_inode *ip,
fs/xfs/libxfs/xfs_rtrmap_btree.c
464
xfs_inode_fork_size(cur->bc_ino.ip, cur->bc_ino.whichfork));
fs/xfs/libxfs/xfs_rtrmap_btree.c
508
struct xfs_inode *ip = rtg_rmap(rtg);
fs/xfs/libxfs/xfs_rtrmap_btree.c
512
xfs_assert_ilocked(ip, XFS_ILOCK_SHARED | XFS_ILOCK_EXCL);
fs/xfs/libxfs/xfs_rtrmap_btree.c
517
cur->bc_ino.ip = ip;
fs/xfs/libxfs/xfs_rtrmap_btree.c
520
cur->bc_nlevels = be16_to_cpu(ip->i_df.if_broot->bb_level) + 1;
fs/xfs/libxfs/xfs_rtrmap_btree.c
521
cur->bc_ino.forksize = xfs_inode_fork_size(ip, XFS_DATA_FORK);
fs/xfs/libxfs/xfs_rtrmap_btree.c
662
ifp = xfs_ifork_ptr(cur->bc_ino.ip, XFS_DATA_FORK);
fs/xfs/libxfs/xfs_rtrmap_btree.c
666
cur->bc_ino.ip->i_projid = cur->bc_group->xg_gno;
fs/xfs/libxfs/xfs_rtrmap_btree.c
667
xfs_trans_log_inode(tp, cur->bc_ino.ip, flags);
fs/xfs/libxfs/xfs_rtrmap_btree.c
826
struct xfs_inode *ip,
fs/xfs/libxfs/xfs_rtrmap_btree.c
831
struct xfs_mount *mp = ip->i_mount;
fs/xfs/libxfs/xfs_rtrmap_btree.c
842
xfs_btree_init_block(mp, rblock, &xfs_rtrmapbt_ops, 0, 0, ip->i_ino);
fs/xfs/libxfs/xfs_rtrmap_btree.c
866
struct xfs_inode *ip,
fs/xfs/libxfs/xfs_rtrmap_btree.c
869
struct xfs_mount *mp = ip->i_mount;
fs/xfs/libxfs/xfs_rtrmap_btree.c
880
if (!xfs_has_rmapbt(ip->i_mount)) {
fs/xfs/libxfs/xfs_rtrmap_btree.c
881
xfs_inode_mark_sick(ip, XFS_SICK_INO_CORE);
fs/xfs/libxfs/xfs_rtrmap_btree.c
891
xfs_inode_mark_sick(ip, XFS_SICK_INO_CORE);
fs/xfs/libxfs/xfs_rtrmap_btree.c
895
broot = xfs_broot_alloc(xfs_ifork_ptr(ip, XFS_DATA_FORK),
fs/xfs/libxfs/xfs_rtrmap_btree.c
898
xfs_rtrmapbt_from_disk(ip, dfp, dsize, broot);
fs/xfs/libxfs/xfs_rtrmap_btree.c
948
struct xfs_inode *ip,
fs/xfs/libxfs/xfs_rtrmap_btree.c
951
struct xfs_ifork *ifp = xfs_ifork_ptr(ip, XFS_DATA_FORK);
fs/xfs/libxfs/xfs_rtrmap_btree.c
957
xfs_inode_fork_size(ip, XFS_DATA_FORK));
fs/xfs/libxfs/xfs_rtrmap_btree.c
958
xfs_rtrmapbt_to_disk(ip->i_mount, ifp->if_broot, ifp->if_broot_bytes,
fs/xfs/libxfs/xfs_rtrmap_btree.c
959
dfp, XFS_DFORK_SIZE(dip, ip->i_mount, XFS_DATA_FORK));
fs/xfs/libxfs/xfs_rtrmap_btree.c
968
struct xfs_inode *ip,
fs/xfs/libxfs/xfs_rtrmap_btree.c
972
struct xfs_ifork *ifp = xfs_ifork_ptr(ip, XFS_DATA_FORK);
fs/xfs/libxfs/xfs_rtrmap_btree.c
973
struct xfs_mount *mp = ip->i_mount;
fs/xfs/libxfs/xfs_rtrmap_btree.c
984
ip->i_ino);
fs/xfs/libxfs/xfs_rtrmap_btree.c
985
xfs_trans_log_inode(tp, ip, XFS_ILOG_CORE | XFS_ILOG_DBROOT);
fs/xfs/libxfs/xfs_rtrmap_btree.h
191
int xfs_iformat_rtrmap(struct xfs_inode *ip, struct xfs_dinode *dip);
fs/xfs/libxfs/xfs_rtrmap_btree.h
195
void xfs_iflush_rtrmap(struct xfs_inode *ip, struct xfs_dinode *dip);
fs/xfs/libxfs/xfs_rtrmap_btree.h
197
int xfs_rtrmapbt_create(struct xfs_rtgroup *rtg, struct xfs_inode *ip,
fs/xfs/libxfs/xfs_rtrmap_btree.h
22
struct xfs_rtgroup *rtg, struct xfs_inode *ip,
fs/xfs/libxfs/xfs_symlink_remote.c
173
struct xfs_inode *ip,
fs/xfs/libxfs/xfs_symlink_remote.c
177
struct xfs_mount *mp = ip->i_mount;
fs/xfs/libxfs/xfs_symlink_remote.c
199
buf += xfs_symlink_hdr_set(mp, ip->i_ino, 0, ifp->if_bytes, bp);
fs/xfs/libxfs/xfs_symlink_remote.c
240
struct xfs_inode *ip,
fs/xfs/libxfs/xfs_symlink_remote.c
243
struct xfs_mount *mp = ip->i_mount;
fs/xfs/libxfs/xfs_symlink_remote.c
248
int pathlen = ip->i_disk_size;
fs/xfs/libxfs/xfs_symlink_remote.c
256
xfs_assert_ilocked(ip, XFS_ILOCK_SHARED | XFS_ILOCK_EXCL);
fs/xfs/libxfs/xfs_symlink_remote.c
259
error = xfs_bmapi_read(ip, 0, fsblocks, mval, &nmaps, 0);
fs/xfs/libxfs/xfs_symlink_remote.c
271
xfs_inode_mark_sick(ip, XFS_SICK_INO_SYMLINK);
fs/xfs/libxfs/xfs_symlink_remote.c
280
if (!xfs_symlink_hdr_ok(ip->i_ino, offset,
fs/xfs/libxfs/xfs_symlink_remote.c
282
xfs_inode_mark_sick(ip, XFS_SICK_INO_SYMLINK);
fs/xfs/libxfs/xfs_symlink_remote.c
286
offset, byte_cnt, ip->i_ino);
fs/xfs/libxfs/xfs_symlink_remote.c
304
link[ip->i_disk_size] = '\0';
fs/xfs/libxfs/xfs_symlink_remote.c
315
struct xfs_inode *ip,
fs/xfs/libxfs/xfs_symlink_remote.c
336
if (pathlen <= xfs_inode_data_fork_size(ip)) {
fs/xfs/libxfs/xfs_symlink_remote.c
337
xfs_init_local_fork(ip, XFS_DATA_FORK, target_path, pathlen);
fs/xfs/libxfs/xfs_symlink_remote.c
339
ip->i_disk_size = pathlen;
fs/xfs/libxfs/xfs_symlink_remote.c
340
ip->i_df.if_format = XFS_DINODE_FMT_LOCAL;
fs/xfs/libxfs/xfs_symlink_remote.c
341
xfs_trans_log_inode(tp, ip, XFS_ILOG_DDATA | XFS_ILOG_CORE);
fs/xfs/libxfs/xfs_symlink_remote.c
346
error = xfs_bmapi_write(tp, ip, 0, fs_blocks, XFS_BMAPI_METADATA,
fs/xfs/libxfs/xfs_symlink_remote.c
351
ip->i_disk_size = pathlen;
fs/xfs/libxfs/xfs_symlink_remote.c
352
xfs_trans_log_inode(tp, ip, XFS_ILOG_CORE);
fs/xfs/libxfs/xfs_symlink_remote.c
391
struct xfs_inode *ip)
fs/xfs/libxfs/xfs_symlink_remote.c
402
error = xfs_bmapi_read(ip, 0, XFS_MAX_FILEOFF, mval, &nmaps, 0);
fs/xfs/libxfs/xfs_symlink_remote.c
421
error = xfs_bunmapi(tp, ip, 0, XFS_MAX_FILEOFF, 0, nmaps, &done);
fs/xfs/libxfs/xfs_symlink_remote.c
426
xfs_inode_mark_sick(ip, XFS_SICK_INO_SYMLINK);
fs/xfs/libxfs/xfs_symlink_remote.c
430
xfs_trans_log_inode(tp, ip, XFS_ILOG_CORE);
fs/xfs/libxfs/xfs_symlink_remote.h
19
struct xfs_inode *ip, struct xfs_ifork *ifp,
fs/xfs/libxfs/xfs_symlink_remote.h
22
int xfs_symlink_remote_read(struct xfs_inode *ip, char *link);
fs/xfs/libxfs/xfs_symlink_remote.h
23
int xfs_symlink_write_target(struct xfs_trans *tp, struct xfs_inode *ip,
fs/xfs/libxfs/xfs_symlink_remote.h
26
int xfs_symlink_remote_truncate(struct xfs_trans *tp, struct xfs_inode *ip);
fs/xfs/libxfs/xfs_trans_inode.c
121
struct xfs_inode *ip)
fs/xfs/libxfs/xfs_trans_inode.c
125
xfs_trans_log_inode(*tpp, ip, XFS_ILOG_CORE);
fs/xfs/libxfs/xfs_trans_inode.c
128
xfs_trans_ijoin(*tpp, ip, 0);
fs/xfs/libxfs/xfs_trans_inode.c
29
struct xfs_inode *ip,
fs/xfs/libxfs/xfs_trans_inode.c
34
xfs_assert_ilocked(ip, XFS_ILOCK_EXCL);
fs/xfs/libxfs/xfs_trans_inode.c
35
if (ip->i_itemp == NULL)
fs/xfs/libxfs/xfs_trans_inode.c
36
xfs_inode_item_init(ip, ip->i_mount);
fs/xfs/libxfs/xfs_trans_inode.c
37
iip = ip->i_itemp;
fs/xfs/libxfs/xfs_trans_inode.c
41
ASSERT(!xfs_iflags_test(ip, XFS_ISTALE));
fs/xfs/libxfs/xfs_trans_inode.c
56
struct xfs_inode *ip,
fs/xfs/libxfs/xfs_trans_inode.c
59
struct inode *inode = VFS_I(ip);
fs/xfs/libxfs/xfs_trans_inode.c
63
xfs_assert_ilocked(ip, XFS_ILOCK_EXCL);
fs/xfs/libxfs/xfs_trans_inode.c
74
ip->i_crtime = tv;
fs/xfs/libxfs/xfs_trans_inode.c
88
struct xfs_inode *ip,
fs/xfs/libxfs/xfs_trans_inode.c
91
struct xfs_inode_log_item *iip = ip->i_itemp;
fs/xfs/libxfs/xfs_trans_inode.c
92
struct inode *inode = VFS_I(ip);
fs/xfs/libxfs/xfs_trans_inode.c
95
xfs_assert_ilocked(ip, XFS_ILOCK_EXCL);
fs/xfs/libxfs/xfs_trans_inode.c
96
ASSERT(!xfs_iflags_test(ip, XFS_ISTALE));
fs/xfs/scrub/agheader.c
945
struct xfs_inode *ip;
fs/xfs/scrub/agheader.c
956
ip = xfs_iunlink_lookup(sc->sa.pag, agino);
fs/xfs/scrub/agheader.c
957
if (!ip) {
fs/xfs/scrub/agheader.c
962
if (!xfs_inode_on_unlinked_list(ip)) {
fs/xfs/scrub/agheader.c
967
agino = ip->i_next_unlinked;
fs/xfs/scrub/agheader_repair.c
1022
struct xfs_inode *ip;
fs/xfs/scrub/agheader_repair.c
1024
ip = xfs_iunlink_lookup(sc->sa.pag, agino);
fs/xfs/scrub/agheader_repair.c
1025
if (!ip)
fs/xfs/scrub/agheader_repair.c
1028
return ip->i_next_unlinked;
fs/xfs/scrub/agheader_repair.c
1044
struct xfs_inode *ip;
fs/xfs/scrub/agheader_repair.c
1048
error = xchk_iget(ragi->sc, xfs_agino_to_ino(sc->sa.pag, agino), &ip);
fs/xfs/scrub/agheader_repair.c
1052
trace_xrep_iunlink_reload_next(ip, prev_agino);
fs/xfs/scrub/agheader_repair.c
1055
if (VFS_I(ip)->i_nlink != 0) {
fs/xfs/scrub/agheader_repair.c
1060
ip->i_prev_unlinked = prev_agino;
fs/xfs/scrub/agheader_repair.c
1061
ret = ip->i_next_unlinked;
fs/xfs/scrub/agheader_repair.c
1069
xchk_irele(sc, ip);
fs/xfs/scrub/agheader_repair.c
1117
struct xfs_inode *ip)
fs/xfs/scrub/agheader_repair.c
1121
if (XFS_INO_TO_AGNO(mp, ip->i_ino) != pag_agno(pag))
fs/xfs/scrub/agheader_repair.c
1124
if (!xfs_inode_on_unlinked_list(ip))
fs/xfs/scrub/agheader_repair.c
1140
struct xfs_inode *ip = ragi->lookup_batch[batch_idx];
fs/xfs/scrub/agheader_repair.c
1145
ASSERT(XFS_INO_TO_AGNO(mp, ip->i_ino) == pag_agno(ragi->sc->sa.pag));
fs/xfs/scrub/agheader_repair.c
1146
ASSERT(xfs_inode_on_unlinked_list(ip));
fs/xfs/scrub/agheader_repair.c
1148
agino = XFS_INO_TO_AGINO(mp, ip->i_ino);
fs/xfs/scrub/agheader_repair.c
1152
ragi->iunlink_heads[bucket], ip);
fs/xfs/scrub/agheader_repair.c
1158
if (ip->i_prev_unlinked == NULLAGINO) {
fs/xfs/scrub/agheader_repair.c
1199
struct xfs_inode *ip = ragi->lookup_batch[i];
fs/xfs/scrub/agheader_repair.c
1201
if (done || !xrep_iunlink_igrab(pag, ip))
fs/xfs/scrub/agheader_repair.c
1216
if (XFS_INO_TO_AGNO(mp, ip->i_ino) != pag_agno(pag))
fs/xfs/scrub/agheader_repair.c
1218
first_index = XFS_INO_TO_AGINO(mp, ip->i_ino + 1);
fs/xfs/scrub/agheader_repair.c
1219
if (first_index < XFS_INO_TO_AGINO(mp, ip->i_ino))
fs/xfs/scrub/agheader_repair.c
1258
struct xfs_inode *ip;
fs/xfs/scrub/agheader_repair.c
1273
ip = radix_tree_lookup(&sc->sa.pag->pag_ici_root, agino);
fs/xfs/scrub/agheader_repair.c
1275
if (ip)
fs/xfs/scrub/agheader_repair.c
1284
&ip);
fs/xfs/scrub/agheader_repair.c
1288
trace_xrep_iunlink_reload_ondisk(ip);
fs/xfs/scrub/agheader_repair.c
1290
if (VFS_I(ip)->i_nlink == 0)
fs/xfs/scrub/agheader_repair.c
1292
xchk_irele(sc, ip);
fs/xfs/scrub/agheader_repair.c
1330
struct xfs_inode *ip;
fs/xfs/scrub/agheader_repair.c
1340
ip = xfs_iunlink_lookup(sc->sa.pag, next_agino);
fs/xfs/scrub/agheader_repair.c
1341
if (!ip) {
fs/xfs/scrub/agheader_repair.c
1358
next_agino = ip->i_next_unlinked;
fs/xfs/scrub/agheader_repair.c
1362
if (!xfs_inode_on_unlinked_list(ip)) {
fs/xfs/scrub/agheader_repair.c
1371
next_agino = ip->i_next_unlinked;
fs/xfs/scrub/agheader_repair.c
1402
next_agino = ip->i_next_unlinked;
fs/xfs/scrub/agheader_repair.c
1535
struct xfs_inode *ip;
fs/xfs/scrub/agheader_repair.c
1540
ip = xfs_iunlink_lookup(pag, agino);
fs/xfs/scrub/agheader_repair.c
1541
if (!ip) {
fs/xfs/scrub/agheader_repair.c
1548
error = xchk_iget(sc, xfs_agino_to_ino(pag, agino), &ip);
fs/xfs/scrub/agheader_repair.c
1559
trace_xrep_iunlink_relink_prev(ip, prev_agino);
fs/xfs/scrub/agheader_repair.c
1560
ip->i_prev_unlinked = prev_agino;
fs/xfs/scrub/agheader_repair.c
1564
if (ip->i_next_unlinked != next_agino) {
fs/xfs/scrub/agheader_repair.c
1565
error = xfs_iunlink_log_inode(sc->tp, ip, pag, next_agino);
fs/xfs/scrub/agheader_repair.c
1569
trace_xrep_iunlink_relink_next(ip, next_agino);
fs/xfs/scrub/agheader_repair.c
1570
ip->i_next_unlinked = next_agino;
fs/xfs/scrub/agheader_repair.c
1580
xchk_irele(sc, ip);
fs/xfs/scrub/agheader_repair.c
1593
struct xfs_inode *ip;
fs/xfs/scrub/agheader_repair.c
1600
ip = xfs_iunlink_lookup(pag, agino);
fs/xfs/scrub/agheader_repair.c
1601
if (!ip) {
fs/xfs/scrub/agheader_repair.c
1608
error = xchk_iget(sc, xfs_agino_to_ino(pag, agino), &ip);
fs/xfs/scrub/agheader_repair.c
1619
error = xfs_iunlink_log_inode(sc->tp, ip, pag, next_agino);
fs/xfs/scrub/agheader_repair.c
1623
trace_xrep_iunlink_relink_next(ip, next_agino);
fs/xfs/scrub/agheader_repair.c
1624
ip->i_next_unlinked = next_agino;
fs/xfs/scrub/agheader_repair.c
1628
if (ip->i_prev_unlinked != prev_agino) {
fs/xfs/scrub/agheader_repair.c
1629
trace_xrep_iunlink_relink_prev(ip, prev_agino);
fs/xfs/scrub/agheader_repair.c
1630
ip->i_prev_unlinked = prev_agino;
fs/xfs/scrub/agheader_repair.c
1640
xchk_irele(sc, ip);
fs/xfs/scrub/attr.c
168
struct xfs_inode *ip,
fs/xfs/scrub/attr.c
180
.dp = ip,
fs/xfs/scrub/attr.c
185
.owner = ip->i_ino,
fs/xfs/scrub/attr.c
202
xchk_ino_set_preen(sc, ip->i_ino);
fs/xfs/scrub/attr.c
576
struct xfs_ifork *ifp = &sc->ip->i_af;
fs/xfs/scrub/attr.c
59
if (!sc->ip)
fs/xfs/scrub/attr.c
62
ifp = xfs_ifork_ptr(sc->ip, XFS_ATTR_FORK);
fs/xfs/scrub/attr.c
649
if (!xfs_inode_hasattr(sc->ip))
fs/xfs/scrub/attr.c
660
if (sc->ip->i_af.if_format == XFS_DINODE_FMT_LOCAL)
fs/xfs/scrub/attr.c
681
error = xchk_xattr_walk(sc, sc->ip, xchk_xattr_actor, NULL, NULL);
fs/xfs/scrub/attr_repair.c
1012
trace_xrep_xattr_reset_fork(sc->ip, sc->ip);
fs/xfs/scrub/attr_repair.c
1015
if (xfs_ifork_has_extents(&sc->ip->i_af)) {
fs/xfs/scrub/attr_repair.c
1016
error = xrep_reap_ifork(sc, sc->ip, XFS_ATTR_FORK);
fs/xfs/scrub/attr_repair.c
1021
error = xrep_xattr_fork_remove(sc, sc->ip);
fs/xfs/scrub/attr_repair.c
1025
return xfs_trans_roll_inode(&sc->tp, sc->ip);
fs/xfs/scrub/attr_repair.c
1039
trace_xrep_xattr_reset_fork(sc->ip, sc->tempip);
fs/xfs/scrub/attr_repair.c
1064
struct xfs_inode *ip = rx->sc->ip;
fs/xfs/scrub/attr_repair.c
1068
if (rx->sc->ip->i_af.if_format == XFS_DINODE_FMT_LOCAL) {
fs/xfs/scrub/attr_repair.c
1088
error = xfs_iread_extents(rx->sc->tp, ip, XFS_ATTR_FORK);
fs/xfs/scrub/attr_repair.c
1119
error = xfs_parent_set(sc->tempip, sc->ip->i_ino, xname,
fs/xfs/scrub/attr_repair.c
1128
error = xfs_parent_unset(sc->tempip, sc->ip->i_ino, xname,
fs/xfs/scrub/attr_repair.c
1259
if (p->ip->i_ino != sc->ip->i_ino)
fs/xfs/scrub/attr_repair.c
1298
.owner = sc->ip->i_ino,
fs/xfs/scrub/attr_repair.c
1322
ifp = xfs_ifork_ptr(sc->ip, XFS_ATTR_FORK);
fs/xfs/scrub/attr_repair.c
1331
xfs_trans_log_inode(sc->tp, sc->ip,
fs/xfs/scrub/attr_repair.c
1347
ip_local = sc->ip->i_af.if_format == XFS_DINODE_FMT_LOCAL;
fs/xfs/scrub/attr_repair.c
1360
forkoff = xfs_attr_shortform_bytesfit(sc->ip, newsize);
fs/xfs/scrub/attr_repair.c
1362
sc->ip->i_forkoff = forkoff;
fs/xfs/scrub/attr_repair.c
1431
xfs_trans_ijoin(sc->tp, sc->ip, 0);
fs/xfs/scrub/attr_repair.c
1439
trace_xrep_xattr_rebuild_tree(sc->ip, sc->tempip);
fs/xfs/scrub/attr_repair.c
1503
xfs_forget_acl(VFS_I(sc->ip), SGI_ACL_FILE);
fs/xfs/scrub/attr_repair.c
1504
xfs_forget_acl(VFS_I(sc->ip), SGI_ACL_DEFAULT);
fs/xfs/scrub/attr_repair.c
1619
if (!xfs_inode_hasattr(sc->ip))
fs/xfs/scrub/attr_repair.c
213
trace_xrep_xattr_salvage_pptr(rx->sc->ip, flags, name,
fs/xfs/scrub/attr_repair.c
222
trace_xrep_xattr_salvage_rec(rx->sc->ip, flags, name,
fs/xfs/scrub/attr_repair.c
329
.dp = rx->sc->ip,
fs/xfs/scrub/attr_repair.c
332
.owner = rx->sc->ip->i_ino,
fs/xfs/scrub/attr_repair.c
465
ifp = xfs_ifork_ptr(rx->sc->ip, XFS_ATTR_FORK);
fs/xfs/scrub/attr_repair.c
584
trace_xrep_xattr_recover_leafblock(rx->sc->ip, dabno,
fs/xfs/scrub/attr_repair.c
593
xfs_attr3_leaf_header_check(bp, rx->sc->ip->i_ino) == NULL)
fs/xfs/scrub/attr_repair.c
620
.owner = rx->sc->ip->i_ino,
fs/xfs/scrub/attr_repair.c
786
xfs_assert_ilocked(rx->sc->ip, XFS_ILOCK_EXCL);
fs/xfs/scrub/attr_repair.c
810
trace_xrep_xattr_full_reset(sc->ip, sc->tempip);
fs/xfs/scrub/attr_repair.c
826
xfs_trans_ijoin(sc->tp, sc->ip, 0);
fs/xfs/scrub/attr_repair.c
902
error = xfs_bmapi_read(sc->ip, offset, XFS_MAX_FILEOFF - offset,
fs/xfs/scrub/attr_repair.c
954
struct xfs_inode *ip)
fs/xfs/scrub/attr_repair.c
957
struct xfs_ifork *ifp = xfs_ifork_ptr(ip, XFS_ATTR_FORK);
fs/xfs/scrub/attr_repair.c
966
if (ip->i_df.if_format == XFS_DINODE_FMT_BTREE) {
fs/xfs/scrub/attr_repair.c
968
hdr = xfs_idata_realloc(ip, (int)sizeof(*hdr) - ifp->if_bytes,
fs/xfs/scrub/attr_repair.c
972
xfs_trans_log_inode(sc->tp, ip,
fs/xfs/scrub/attr_repair.c
985
ip->i_ino, ifp->if_nextents, ifp->if_format);
fs/xfs/scrub/attr_repair.c
997
xfs_attr_fork_remove(ip, sc->tp);
fs/xfs/scrub/bmap.c
1006
xchk_ino_set_preen(info->sc, info->sc->ip->i_ino);
fs/xfs/scrub/bmap.c
1025
struct xfs_inode *ip = sc->ip;
fs/xfs/scrub/bmap.c
1026
struct xfs_ifork *ifp = xfs_ifork_ptr(ip, whichfork);
fs/xfs/scrub/bmap.c
1034
info.is_rt = xfs_ifork_is_realtime(ip, whichfork);
fs/xfs/scrub/bmap.c
1036
info.is_shared = whichfork == XFS_DATA_FORK && xfs_is_reflink_inode(ip);
fs/xfs/scrub/bmap.c
1043
xchk_ino_set_corrupt(sc, sc->ip->i_ino);
fs/xfs/scrub/bmap.c
1055
xchk_ino_set_corrupt(sc, sc->ip->i_ino);
fs/xfs/scrub/bmap.c
1093
error = xfs_bmap_last_offset(ip, &endoff, whichfork);
fs/xfs/scrub/bmap.c
1117
xchk_bmap_iextent_delalloc(ip, &info, &irec);
fs/xfs/scrub/bmap.c
1119
xchk_bmap_iextent(ip, &info, &irec);
fs/xfs/scrub/bmap.c
1140
xchk_ino_set_corrupt(sc, sc->ip->i_ino);
fs/xfs/scrub/bmap.c
1168
xchk_ino_set_corrupt(sc, sc->ip->i_ino);
fs/xfs/scrub/bmap.c
158
if (xfs_ifork_is_realtime(info->sc->ip, info->whichfork))
fs/xfs/scrub/bmap.c
208
uint64_t owner = info->sc->ip->i_ino;
fs/xfs/scrub/bmap.c
323
struct xfs_inode *ip,
fs/xfs/scrub/bmap.c
332
xfs_rtb_to_rgno(ip->i_mount, irec->br_startblock),
fs/xfs/scrub/bmap.c
354
if (!xfs_is_reflink_inode(info->sc->ip)) {
fs/xfs/scrub/bmap.c
355
xfs_rmap_ino_owner(&oinfo, info->sc->ip->i_ino,
fs/xfs/scrub/bmap.c
384
struct xfs_inode *ip,
fs/xfs/scrub/bmap.c
409
if (!xfs_is_reflink_inode(info->sc->ip)) {
fs/xfs/scrub/bmap.c
410
xfs_rmap_ino_owner(&oinfo, info->sc->ip->i_ino,
fs/xfs/scrub/bmap.c
422
xfs_rmap_ino_owner(&oinfo, info->sc->ip->i_ino,
fs/xfs/scrub/bmap.c
452
struct xfs_inode *ip,
fs/xfs/scrub/bmap.c
456
struct xfs_mount *mp = ip->i_mount;
fs/xfs/scrub/bmap.c
459
if (!S_ISDIR(VFS_I(ip)->i_mode) && info->whichfork != XFS_ATTR_FORK)
fs/xfs/scrub/bmap.c
474
struct xfs_inode *ip,
fs/xfs/scrub/bmap.c
493
xchk_bmap_dirattr_extent(ip, info, irec);
fs/xfs/scrub/bmap.c
515
xchk_bmap_rt_iextent_xref(ip, info, irec);
fs/xfs/scrub/bmap.c
517
xchk_bmap_iextent_xref(ip, info, irec);
fs/xfs/scrub/bmap.c
53
if (S_ISREG(VFS_I(sc->ip)->i_mode) &&
fs/xfs/scrub/bmap.c
530
struct xfs_inode *ip = bs->cur->bc_ino.ip;
fs/xfs/scrub/bmap.c
533
struct xfs_ifork *ifp = xfs_ifork_ptr(ip, info->whichfork);
fs/xfs/scrub/bmap.c
546
if (owner != ip->i_ino)
fs/xfs/scrub/bmap.c
55
struct address_space *mapping = VFS_I(sc->ip)->i_mapping;
fs/xfs/scrub/bmap.c
563
if (xfs_bmap_validate_extent(ip, info->whichfork, &irec) != NULL) {
fs/xfs/scrub/bmap.c
569
if (!xfs_iext_lookup_extent(ip, ifp, irec.br_startoff, &icur,
fs/xfs/scrub/bmap.c
588
struct xfs_ifork *ifp = xfs_ifork_ptr(sc->ip, whichfork);
fs/xfs/scrub/bmap.c
590
struct xfs_inode *ip = sc->ip;
fs/xfs/scrub/bmap.c
597
error = xfs_iread_extents(sc->tp, ip, whichfork);
fs/xfs/scrub/bmap.c
602
cur = xfs_bmbt_init_cursor(mp, sc->tp, ip, whichfork);
fs/xfs/scrub/bmap.c
603
xfs_rmap_ino_bmbt_owner(&oinfo, ip->i_ino, whichfork);
fs/xfs/scrub/bmap.c
62
error = xfs_break_layouts(VFS_I(sc->ip),
fs/xfs/scrub/bmap.c
631
if (rec->rm_owner != sc->ip->i_ino)
fs/xfs/scrub/bmap.c
640
ifp = xfs_ifork_ptr(sc->ip, sbcri->whichfork);
fs/xfs/scrub/bmap.c
646
have_map = xfs_iext_lookup_extent(sc->ip, ifp, rec->rm_offset,
fs/xfs/scrub/bmap.c
68
inode_dio_wait(VFS_I(sc->ip));
fs/xfs/scrub/bmap.c
748
struct xfs_inode *ip)
fs/xfs/scrub/bmap.c
750
struct xfs_ifork *ifp = &ip->i_af;
fs/xfs/scrub/bmap.c
776
if ((VFS_I(ip)->i_mode & 0777) != 0)
fs/xfs/scrub/bmap.c
778
if (!uid_eq(VFS_I(ip)->i_uid, GLOBAL_ROOT_UID))
fs/xfs/scrub/bmap.c
780
if (!gid_eq(VFS_I(ip)->i_gid, GLOBAL_ROOT_GID))
fs/xfs/scrub/bmap.c
792
struct xfs_inode *ip)
fs/xfs/scrub/bmap.c
794
struct xfs_ifork *ifp = &ip->i_df;
fs/xfs/scrub/bmap.c
811
return i_size_read(VFS_I(ip)) != 0;
fs/xfs/scrub/bmap.c
832
return xchk_bmap_check_empty_attrfork(sc->ip);
fs/xfs/scrub/bmap.c
834
return xchk_bmap_check_empty_datafork(sc->ip);
fs/xfs/scrub/bmap.c
846
if (xfs_ifork_is_realtime(sc->ip, whichfork)) {
fs/xfs/scrub/bmap.c
876
struct xfs_inode *ip,
fs/xfs/scrub/bmap.c
92
VFS_I(sc->ip)->i_mapping);
fs/xfs/scrub/bmap.c
966
ifp = xfs_ifork_ptr(info->sc->ip, info->whichfork);
fs/xfs/scrub/bmap_repair.c
110
if (XFS_IS_REALTIME_INODE(sc->ip)) {
fs/xfs/scrub/bmap_repair.c
163
fa = xfs_bmap_validate_extent(sc->ip, rb->whichfork, &irec);
fs/xfs/scrub/bmap_repair.c
169
trace_xrep_bmap_found(sc->ip, rb->whichfork, &irec);
fs/xfs/scrub/bmap_repair.c
203
if (XFS_IS_REALTIME_INODE(sc->ip) &&
fs/xfs/scrub/bmap_repair.c
255
if (rec->rm_owner != rb->sc->ip->i_ino)
fs/xfs/scrub/bmap_repair.c
388
if (!XFS_IS_REALTIME_INODE(sc->ip))
fs/xfs/scrub/bmap_repair.c
419
if (rec->rm_owner != rb->sc->ip->i_ino)
fs/xfs/scrub/bmap_repair.c
487
struct xfs_inode *ip = rb->sc->ip;
fs/xfs/scrub/bmap_repair.c
488
struct xfs_ifork *ifp = xfs_ifork_ptr(ip, rb->whichfork);
fs/xfs/scrub/bmap_repair.c
495
if (rb->whichfork == XFS_ATTR_FORK || ip->i_delayed_blks == 0)
fs/xfs/scrub/bmap_repair.c
504
trace_xrep_bmap_found(ip, rb->whichfork, &irec);
fs/xfs/scrub/bmap_repair.c
534
if (!xfs_is_metadir_inode(sc->ip)) {
fs/xfs/scrub/bmap_repair.c
626
sc->ip->i_diflags2 |= XFS_DIFLAG2_REFLINK;
fs/xfs/scrub/bmap_repair.c
633
sc->ip->i_nblocks = rb->nblocks + delta;
fs/xfs/scrub/bmap_repair.c
634
xfs_trans_log_inode(sc->tp, sc->ip, XFS_ILOG_CORE);
fs/xfs/scrub/bmap_repair.c
640
xfs_trans_mod_dquot_byino(sc->tp, sc->ip, XFS_TRANS_DQ_BCOUNT, delta);
fs/xfs/scrub/bmap_repair.c
711
error = xfs_trans_reserve_more_inode(sc->tp, sc->ip,
fs/xfs/scrub/bmap_repair.c
763
xfs_rmap_ino_bmbt_owner(&oinfo, sc->ip->i_ino, rb->whichfork);
fs/xfs/scrub/bmap_repair.c
777
bmap_cur = xfs_bmbt_init_cursor(sc->mp, NULL, sc->ip, XFS_STAGING_FORK);
fs/xfs/scrub/bmap_repair.c
785
if (rb->real_mappings <= XFS_IFORK_MAXEXT(sc->ip, rb->whichfork)) {
fs/xfs/scrub/bmap_repair.c
836
xfs_rmap_ino_bmbt_owner(&oinfo, sc->ip->i_ino, rb->whichfork);
fs/xfs/scrub/bmap_repair.c
846
struct xfs_ifork *ifp = xfs_ifork_ptr(sc->ip, whichfork);
fs/xfs/scrub/bmap_repair.c
880
switch (VFS_I(sc->ip)->i_mode & S_IFMT) {
fs/xfs/scrub/bmap_repair.c
904
if (xfs_is_reflink_inode(sc->ip))
fs/xfs/scrub/bmap_repair.c
908
if (!S_ISREG(VFS_I(sc->ip)->i_mode))
fs/xfs/scrub/bmap_repair.c
958
xfs_trans_ijoin(sc->tp, sc->ip, 0);
fs/xfs/scrub/btree.c
41
trace_xchk_deadlock_retry(sc->ip, sc->sm, *error);
fs/xfs/scrub/btree.c
484
xfs_inode_has_attr_fork(bs->sc->ip))
fs/xfs/scrub/common.c
1067
ASSERT(sc->ip != NULL);
fs/xfs/scrub/common.c
1080
struct xfs_inode *ip)
fs/xfs/scrub/common.c
1082
if (VFS_I(ip)->i_generation != sc->sm->sm_gen) {
fs/xfs/scrub/common.c
1083
xchk_irele(sc, ip);
fs/xfs/scrub/common.c
1087
sc->ip = ip;
fs/xfs/scrub/common.c
1099
struct xfs_inode *ip)
fs/xfs/scrub/common.c
1101
if (!igrab(VFS_I(ip))) {
fs/xfs/scrub/common.c
1102
xchk_ino_set_corrupt(sc, ip->i_ino);
fs/xfs/scrub/common.c
1106
sc->ip = ip;
fs/xfs/scrub/common.c
1126
struct xfs_inode *ip = NULL;
fs/xfs/scrub/common.c
1149
error = xchk_iget_safe(sc, sc->sm->sm_ino, &ip);
fs/xfs/scrub/common.c
1151
return xchk_install_handle_inode(sc, ip);
fs/xfs/scrub/common.c
1176
error = xchk_iget_agi(sc, sc->sm->sm_ino, &agi_bp, &ip);
fs/xfs/scrub/common.c
1180
return xchk_install_handle_inode(sc, ip);
fs/xfs/scrub/common.c
1240
struct xfs_inode *ip)
fs/xfs/scrub/common.c
1255
spin_lock(&VFS_I(ip)->i_lock);
fs/xfs/scrub/common.c
1256
inode_state_clear(VFS_I(ip), I_DONTCACHE);
fs/xfs/scrub/common.c
1257
spin_unlock(&VFS_I(ip)->i_lock);
fs/xfs/scrub/common.c
1260
xfs_irele(ip);
fs/xfs/scrub/common.c
1305
xfs_ilock(sc->ip, ilock_flags);
fs/xfs/scrub/common.c
1314
if (xfs_ilock_nowait(sc->ip, ilock_flags)) {
fs/xfs/scrub/common.c
1328
xfs_iunlock(sc->ip, ilock_flags);
fs/xfs/scrub/common.c
1430
if (sc->ip->i_diflags & XFS_DIFLAG_REALTIME) {
fs/xfs/scrub/common.c
1431
xchk_ino_set_corrupt(sc, sc->ip->i_ino);
fs/xfs/scrub/common.c
1436
if (xfs_is_reflink_inode(sc->ip)) {
fs/xfs/scrub/common.c
1437
xchk_ino_set_corrupt(sc, sc->ip->i_ino);
fs/xfs/scrub/common.c
1448
error = xfs_reflink_inode_has_shared_extents(sc->tp, sc->ip,
fs/xfs/scrub/common.c
1454
xchk_ino_set_corrupt(sc, sc->ip->i_ino);
fs/xfs/scrub/common.c
1461
if (xfs_inode_hasattr(sc->ip)) {
fs/xfs/scrub/common.c
1463
xchk_ino_set_corrupt(sc, sc->ip->i_ino);
fs/xfs/scrub/common.c
1531
struct xfs_inode *ip;
fs/xfs/scrub/common.c
1553
ip = radix_tree_lookup(&pag->pag_ici_root, agino);
fs/xfs/scrub/common.c
1554
if (!ip) {
fs/xfs/scrub/common.c
1564
spin_lock(&ip->i_flags_lock);
fs/xfs/scrub/common.c
1565
if (ip->i_ino != ino)
fs/xfs/scrub/common.c
1568
trace_xchk_inode_is_allocated(ip);
fs/xfs/scrub/common.c
1583
if (!(ip->i_flags & (XFS_NEED_INACTIVE | XFS_INEW | XFS_IRECLAIMABLE |
fs/xfs/scrub/common.c
1586
ASSERT(VFS_I(ip)->i_mode != 0);
fs/xfs/scrub/common.c
1600
if (ip->i_flags & XFS_INEW) {
fs/xfs/scrub/common.c
1602
ASSERT(VFS_I(ip)->i_mode != 0);
fs/xfs/scrub/common.c
1609
if ((ip->i_flags & XFS_NEED_INACTIVE) &&
fs/xfs/scrub/common.c
1610
!(ip->i_flags & XFS_INACTIVATING)) {
fs/xfs/scrub/common.c
1612
ASSERT(VFS_I(ip)->i_mode != 0);
fs/xfs/scrub/common.c
1648
*inuse = VFS_I(ip)->i_mode != 0;
fs/xfs/scrub/common.c
1652
spin_unlock(&ip->i_flags_lock);
fs/xfs/scrub/common.c
1660
xchk_inode_is_dirtree_root(const struct xfs_inode *ip)
fs/xfs/scrub/common.c
1662
struct xfs_mount *mp = ip->i_mount;
fs/xfs/scrub/common.c
1664
return ip == mp->m_rootip ||
fs/xfs/scrub/common.c
1665
(xfs_has_metadir(mp) && ip == mp->m_metadirip);
fs/xfs/scrub/common.c
1670
xchk_inode_is_sb_rooted(const struct xfs_inode *ip)
fs/xfs/scrub/common.c
1672
return xchk_inode_is_dirtree_root(ip) ||
fs/xfs/scrub/common.c
1673
xfs_is_sb_inum(ip->i_mount, ip->i_ino);
fs/xfs/scrub/common.c
1678
xchk_inode_rootdir_inum(const struct xfs_inode *ip)
fs/xfs/scrub/common.c
168
trace_xchk_deadlock_retry(sc->ip, sc->sm, *error);
fs/xfs/scrub/common.c
1680
struct xfs_mount *mp = ip->i_mount;
fs/xfs/scrub/common.c
1682
if (xfs_is_metadir_inode(ip))
fs/xfs/scrub/common.c
1701
switch (sc->ip->i_metatype) {
fs/xfs/scrub/common.c
1730
struct xfs_ifork *ifp = xfs_ifork_ptr(sc->ip, whichfork);
fs/xfs/scrub/common.c
1743
return xfs_bmap_count_blocks(sc->tp, sc->ip, whichfork, nextents,
fs/xfs/scrub/common.c
510
if (sc->ip)
fs/xfs/scrub/common.c
763
if (sc->ip && !xfs_is_internal_inode(sc->ip))
fs/xfs/scrub/common.c
92
sc->ip ? sc->ip : XFS_I(file_inode(sc->file)),
fs/xfs/scrub/common.h
175
int xchk_install_live_inode(struct xfs_scrub *sc, struct xfs_inode *ip);
fs/xfs/scrub/common.h
191
void xchk_irele(struct xfs_scrub *sc, struct xfs_inode *ip);
fs/xfs/scrub/common.h
192
int xchk_install_handle_inode(struct xfs_scrub *sc, struct xfs_inode *ip);
fs/xfs/scrub/common.h
225
bool xchk_pptr_looks_zapped(struct xfs_inode *ip);
fs/xfs/scrub/common.h
268
bool xchk_inode_is_dirtree_root(const struct xfs_inode *ip);
fs/xfs/scrub/common.h
269
bool xchk_inode_is_sb_rooted(const struct xfs_inode *ip);
fs/xfs/scrub/common.h
270
xfs_ino_t xchk_inode_rootdir_inum(const struct xfs_inode *ip);
fs/xfs/scrub/cow_repair.c
104
trace_xrep_cow_mark_file_range(xc->sc->ip, startblock, startoff,
fs/xfs/scrub/cow_repair.c
430
XFS_INO_TO_FSB(sc->mp, sc->ip->i_ino));
fs/xfs/scrub/cow_repair.c
482
struct xfs_inode *ip = xc->sc->ip;
fs/xfs/scrub/cow_repair.c
483
struct xfs_ifork *ifp = xfs_ifork_ptr(ip, XFS_COW_FORK);
fs/xfs/scrub/cow_repair.c
485
if (!xfs_iext_lookup_extent(ip, ifp, startoff, icur, got))
fs/xfs/scrub/cow_repair.c
515
struct xfs_inode *ip,
fs/xfs/scrub/cow_repair.c
525
trace_xrep_cow_replace_mapping(ip, got, repl->fsbno, repl->len);
fs/xfs/scrub/cow_repair.c
533
xfs_iext_update_extent(ip, BMAP_COWFORK, icur, &new);
fs/xfs/scrub/cow_repair.c
544
xfs_iext_update_extent(ip, BMAP_COWFORK, icur, &new);
fs/xfs/scrub/cow_repair.c
549
xfs_iext_insert(ip, icur, &new, BMAP_COWFORK);
fs/xfs/scrub/cow_repair.c
586
if (XFS_IS_REALTIME_INODE(sc->ip))
fs/xfs/scrub/cow_repair.c
597
xrep_cow_replace_mapping(sc->ip, &icur, &got, &repl);
fs/xfs/scrub/cow_repair.c
599
xfs_inode_set_cowblocks_tag(sc->ip);
fs/xfs/scrub/cow_repair.c
605
if (XFS_IS_REALTIME_INODE(sc->ip))
fs/xfs/scrub/cow_repair.c
657
struct xfs_ifork *ifp = xfs_ifork_ptr(sc->ip, XFS_COW_FORK);
fs/xfs/scrub/cow_repair.c
671
if (xfs_inode_has_bigrtalloc(sc->ip))
fs/xfs/scrub/cow_repair.c
675
if (xfs_is_metadir_inode(sc->ip) && XFS_IS_REALTIME_INODE(sc->ip))
fs/xfs/scrub/cow_repair.c
692
xfs_trans_ijoin(sc->tp, sc->ip, 0);
fs/xfs/scrub/cow_repair.c
696
if (XFS_IS_REALTIME_INODE(sc->ip))
fs/xfs/scrub/cow_repair.c
722
if (XFS_IS_REALTIME_INODE(sc->ip))
fs/xfs/scrub/cow_repair.c
740
if (XFS_IS_REALTIME_INODE(sc->ip))
fs/xfs/scrub/cow_repair.c
750
if (XFS_IS_REALTIME_INODE(sc->ip))
fs/xfs/scrub/dabtree.c
340
struct xfs_inode *ip = ds->dargs.dp;
fs/xfs/scrub/dabtree.c
394
if (xfs_has_crc(ip->i_mount)) {
fs/xfs/scrub/dabtree.c
396
if (owner != ip->i_ino)
fs/xfs/scrub/dabtree.c
421
blk->hashval = xfs_dir2_leaf_lasthash(ip, blk->bp, pmaxrecs);
fs/xfs/scrub/dabtree.c
430
blk->hashval = xfs_dir2_leaf_lasthash(ip, blk->bp, pmaxrecs);
fs/xfs/scrub/dabtree.c
44
trace_xchk_deadlock_retry(sc->ip, sc->sm, *error);
fs/xfs/scrub/dabtree.c
440
xfs_da3_node_hdr_from_disk(ip->i_mount, &nodehdr, node);
fs/xfs/scrub/dabtree.c
511
if (!xfs_ifork_has_extents(xfs_ifork_ptr(sc->ip, whichfork)))
fs/xfs/scrub/dabtree.c
518
ds->dargs.dp = sc->ip;
fs/xfs/scrub/dabtree.c
522
ds->dargs.owner = sc->ip->i_ino;
fs/xfs/scrub/dir.c
1005
trace_xchk_dir_ultraslowpath(sc->ip, xname, ip->i_ino);
fs/xfs/scrub/dir.c
1007
error = xchk_dir_trylock_for_pptrs(sc, ip, &lockmode);
fs/xfs/scrub/dir.c
101
if (xfs_mode_to_ftype(VFS_I(ip)->i_mode) != ftype)
fs/xfs/scrub/dir.c
1021
error = xchk_dir_parent_pointer(sd, xname, ip);
fs/xfs/scrub/dir.c
1023
xfs_iunlock(ip, lockmode);
fs/xfs/scrub/dir.c
1025
xchk_irele(sc, ip);
fs/xfs/scrub/dir.c
1068
if (!S_ISDIR(VFS_I(sc->ip)->i_mode))
fs/xfs/scrub/dir.c
1077
if (sc->ip->i_disk_size < xfs_dir2_sf_hdr_size(0)) {
fs/xfs/scrub/dir.c
1078
xchk_ino_set_corrupt(sc, sc->ip->i_ino);
fs/xfs/scrub/dir.c
109
if (xfs_is_metadir_inode(ip) != xfs_is_metadir_inode(sc->ip))
fs/xfs/scrub/dir.c
1121
error = xchk_dir_walk(sc, sc->ip, xchk_dir_actor, sd);
fs/xfs/scrub/dir.c
120
struct xfs_inode *ip)
fs/xfs/scrub/dir.c
122
if (!xfs_ilock_nowait(ip, XFS_IOLOCK_SHARED))
fs/xfs/scrub/dir.c
125
if (!xfs_ilock_nowait(ip, XFS_ILOCK_SHARED)) {
fs/xfs/scrub/dir.c
126
xfs_iunlock(ip, XFS_IOLOCK_SHARED);
fs/xfs/scrub/dir.c
130
if (!xfs_inode_has_attr_fork(ip) || !xfs_need_iread_extents(&ip->i_af))
fs/xfs/scrub/dir.c
133
xfs_iunlock(ip, XFS_ILOCK_SHARED);
fs/xfs/scrub/dir.c
135
if (!xfs_ilock_nowait(ip, XFS_ILOCK_EXCL)) {
fs/xfs/scrub/dir.c
136
xfs_iunlock(ip, XFS_IOLOCK_SHARED);
fs/xfs/scrub/dir.c
148
struct xfs_inode *ip)
fs/xfs/scrub/dir.c
153
xfs_inode_to_parent_rec(&sd->pptr_rec, sc->ip);
fs/xfs/scrub/dir.c
154
error = xfs_parent_lookup(sc->tp, ip, name, &sd->pptr_rec,
fs/xfs/scrub/dir.c
168
struct xfs_inode *ip)
fs/xfs/scrub/dir.c
180
if (ip == sc->ip) {
fs/xfs/scrub/dir.c
186
lockmode = xchk_dir_lock_child(sc, ip);
fs/xfs/scrub/dir.c
190
.ino = ip->i_ino,
fs/xfs/scrub/dir.c
194
trace_xchk_dir_defer(sc->ip, name, ip->i_ino);
fs/xfs/scrub/dir.c
210
error = xchk_dir_parent_pointer(sd, name, ip);
fs/xfs/scrub/dir.c
211
xfs_iunlock(ip, lockmode);
fs/xfs/scrub/dir.c
231
struct xfs_inode *ip;
fs/xfs/scrub/dir.c
289
error = xchk_iget(sc, ino, &ip);
fs/xfs/scrub/dir.c
298
xchk_dir_check_ftype(sc, offset, ip, name->type);
fs/xfs/scrub/dir.c
301
error = xchk_dir_check_pptr_fast(sd, dapos, name, ip);
fs/xfs/scrub/dir.c
307
xchk_irele(sc, ip);
fs/xfs/scrub/dir.c
485
error = xfs_dir3_block_read(sc->tp, sc->ip, sc->ip->i_ino, &bp);
fs/xfs/scrub/dir.c
488
error = xfs_dir3_data_read(sc->tp, sc->ip, sc->ip->i_ino, lblk,
fs/xfs/scrub/dir.c
641
error = xfs_dir3_leaf_read(sc->tp, sc->ip, sc->ip->i_ino, lblk, &bp);
fs/xfs/scrub/dir.c
647
xfs_dir2_leaf_hdr_from_disk(sc->ip->i_mount, &leafhdr, leaf);
fs/xfs/scrub/dir.c
702
error = xfs_dir3_data_read(sc->tp, sc->ip, args->owner,
fs/xfs/scrub/dir.c
747
error = xfs_dir2_free_read(sc->tp, sc->ip, sc->ip->i_ino, lblk, &bp);
fs/xfs/scrub/dir.c
760
xfs_dir2_free_hdr_from_disk(sc->ip->i_mount, &freehdr, bp->b_addr);
fs/xfs/scrub/dir.c
767
error = xfs_dir3_data_read(sc->tp, sc->ip, args->owner,
fs/xfs/scrub/dir.c
791
.dp = sc->ip,
fs/xfs/scrub/dir.c
795
.owner = sc->ip->i_ino,
fs/xfs/scrub/dir.c
797
struct xfs_ifork *ifp = xfs_ifork_ptr(sc->ip, XFS_DATA_FORK);
fs/xfs/scrub/dir.c
825
found = xfs_iext_lookup_extent(sc->ip, ifp, lblk, &icur, &got);
fs/xfs/scrub/dir.c
854
found = xfs_iext_lookup_extent(sc->ip, ifp, lblk, &icur, &got);
fs/xfs/scrub/dir.c
861
if (xfs_iext_lookup_extent(sc->ip, ifp, leaf_lblk, &icur, &got) &&
fs/xfs/scrub/dir.c
880
found = xfs_iext_lookup_extent(sc->ip, ifp, lblk, &icur, &got);
fs/xfs/scrub/dir.c
90
struct xfs_inode *ip,
fs/xfs/scrub/dir.c
918
found = xfs_iext_lookup_extent(sc->ip, ifp, lblk, &icur, &got);
fs/xfs/scrub/dir.c
944
error = xchk_dir_lookup(sc, sc->ip, xname, &child_ino);
fs/xfs/scrub/dir.c
966
struct xfs_inode *ip;
fs/xfs/scrub/dir.c
980
error = xchk_iget(sc, dirent->ino, &ip);
fs/xfs/scrub/dir.c
992
lockmode = xchk_dir_lock_child(sc, ip);
fs/xfs/scrub/dir.c
994
trace_xchk_dir_slowpath(sc->ip, xname, ip->i_ino);
fs/xfs/scrub/dir_repair.c
1015
struct xfs_inode *dp = rd->sc->ip;
fs/xfs/scrub/dir_repair.c
1098
struct xfs_inode *ip,
fs/xfs/scrub/dir_repair.c
1109
.type = xfs_mode_to_ftype(VFS_I(ip)->i_mode),
fs/xfs/scrub/dir_repair.c
1128
if (parent_ino != sc->ip->i_ino ||
fs/xfs/scrub/dir_repair.c
1129
parent_gen != VFS_I(sc->ip)->i_generation)
fs/xfs/scrub/dir_repair.c
1133
error = xrep_dir_stash_createname(rd, &xname, ip->i_ino);
fs/xfs/scrub/dir_repair.c
1154
if (ino != rd->sc->ip->i_ino)
fs/xfs/scrub/dir_repair.c
1184
const struct xfs_inode *ip)
fs/xfs/scrub/dir_repair.c
1186
return ip != rd->sc->ip && !xrep_is_tempfile(ip);
fs/xfs/scrub/dir_repair.c
1198
struct xfs_inode *ip)
fs/xfs/scrub/dir_repair.c
1203
if (!xrep_dir_want_scan(rd, ip))
fs/xfs/scrub/dir_repair.c
1206
if (S_ISDIR(VFS_I(ip)->i_mode) && xfs_need_iread_extents(&ip->i_df)) {
fs/xfs/scrub/dir_repair.c
1211
if (xfs_inode_has_attr_fork(ip) && xfs_need_iread_extents(&ip->i_af))
fs/xfs/scrub/dir_repair.c
1215
xfs_ilock(ip, lock_mode);
fs/xfs/scrub/dir_repair.c
1226
struct xfs_inode *ip)
fs/xfs/scrub/dir_repair.c
1231
lock_mode = xrep_dir_scan_ilock(rd, ip);
fs/xfs/scrub/dir_repair.c
1233
if (!xrep_dir_want_scan(rd, ip))
fs/xfs/scrub/dir_repair.c
1240
if (xchk_pptr_looks_zapped(ip)) {
fs/xfs/scrub/dir_repair.c
1245
error = xchk_xattr_walk(rd->sc, ip, xrep_dir_scan_pptr, NULL, rd);
fs/xfs/scrub/dir_repair.c
1249
if (S_ISDIR(VFS_I(ip)->i_mode)) {
fs/xfs/scrub/dir_repair.c
1254
if (xchk_dir_looks_zapped(ip)) {
fs/xfs/scrub/dir_repair.c
1259
error = xchk_dir_walk(rd->sc, ip, xrep_dir_scan_dirent, rd);
fs/xfs/scrub/dir_repair.c
1265
xchk_iscan_mark_visited(&rd->pscan.iscan, ip);
fs/xfs/scrub/dir_repair.c
1266
xfs_iunlock(ip, lock_mode);
fs/xfs/scrub/dir_repair.c
1279
struct xfs_inode *ip;
fs/xfs/scrub/dir_repair.c
1283
if (xchk_inode_is_dirtree_root(sc->ip))
fs/xfs/scrub/dir_repair.c
1284
xrep_findparent_scan_found(&rd->pscan, sc->ip->i_ino);
fs/xfs/scrub/dir_repair.c
1298
while ((error = xchk_iscan_iter(&rd->pscan.iscan, &ip)) == 1) {
fs/xfs/scrub/dir_repair.c
1301
error = xrep_dir_scan_file(rd, ip);
fs/xfs/scrub/dir_repair.c
1302
xchk_irele(sc, ip);
fs/xfs/scrub/dir_repair.c
1372
if (p->dp->i_ino == sc->ip->i_ino &&
fs/xfs/scrub/dir_repair.c
1373
xchk_iscan_want_live_update(&rd->pscan.iscan, p->ip->i_ino)) {
fs/xfs/scrub/dir_repair.c
1377
p->ip->i_ino);
fs/xfs/scrub/dir_repair.c
1380
p->ip->i_ino);
fs/xfs/scrub/dir_repair.c
1391
if (p->ip->i_ino == sc->ip->i_ino &&
fs/xfs/scrub/dir_repair.c
1471
.owner = sc->ip->i_ino,
fs/xfs/scrub/dir_repair.c
1495
ifp = xfs_ifork_ptr(sc->ip, XFS_DATA_FORK);
fs/xfs/scrub/dir_repair.c
1503
xfs_trans_log_inode(sc->tp, sc->ip,
fs/xfs/scrub/dir_repair.c
1545
struct xfs_inode *dp = sc->ip;
fs/xfs/scrub/dir_repair.c
1680
ip_local = sc->ip->i_df.if_format == XFS_DINODE_FMT_LOCAL;
fs/xfs/scrub/dir_repair.c
1690
sc->tempip->i_disk_size <= xfs_inode_data_fork_size(sc->ip)) {
fs/xfs/scrub/dir_repair.c
1730
trace_xrep_dir_rebuild_tree(sc->ip, rd->pscan.parent_ino);
fs/xfs/scrub/dir_repair.c
1843
error = xchk_dir_lookup(sc, sc->ip, &xfs_name_dotdot, &orig_parent);
fs/xfs/scrub/dir_repair.c
1880
error = xchk_dir_lookup(sc, sc->ip, &xfs_name_dotdot, &new_parent);
fs/xfs/scrub/dir_repair.c
1888
if (orig_parent == new_parent && VFS_I(sc->ip)->i_nlink > 0) {
fs/xfs/scrub/dir_repair.c
223
error = xfs_dir_lookup(sc->tp, sc->ip, &xfs_name_dotdot, &ino, NULL);
fs/xfs/scrub/dir_repair.c
307
if (ino == rd->sc->ip->i_ino)
fs/xfs/scrub/dir_repair.c
389
struct xfs_inode *ip;
fs/xfs/scrub/dir_repair.c
408
trace_xrep_dir_salvaged_parent(sc->ip, ino);
fs/xfs/scrub/dir_repair.c
412
trace_xrep_dir_salvage_entry(sc->ip, &xname, ino);
fs/xfs/scrub/dir_repair.c
418
error = xchk_iget(sc, ino, &ip);
fs/xfs/scrub/dir_repair.c
423
if (xfs_is_metadir_inode(ip) != xfs_is_metadir_inode(rd->sc->ip)) {
fs/xfs/scrub/dir_repair.c
424
xchk_irele(sc, ip);
fs/xfs/scrub/dir_repair.c
428
xname.type = xfs_mode_to_ftype(VFS_I(ip)->i_mode);
fs/xfs/scrub/dir_repair.c
429
xchk_irele(sc, ip);
fs/xfs/scrub/dir_repair.c
525
ifp = xfs_ifork_ptr(rd->sc->ip, XFS_DATA_FORK);
fs/xfs/scrub/dir_repair.c
530
trace_xrep_dir_salvaged_parent(rd->sc->ip, ino);
fs/xfs/scrub/dir_repair.c
565
struct xfs_inode *dp = rd->sc->ip;
fs/xfs/scrub/dir_repair.c
617
error = xfs_da_read_buf(rd->sc->tp, rd->sc->ip, dabno,
fs/xfs/scrub/dir_repair.c
625
trace_xrep_dir_recover_dirblock(rd->sc->ip, dabno,
fs/xfs/scrub/dir_repair.c
647
if (xfs_dir3_block_header_check(bp, rd->sc->ip->i_ino) != NULL)
fs/xfs/scrub/dir_repair.c
654
if (xfs_dir3_data_header_check(bp, rd->sc->ip->i_ino) != NULL)
fs/xfs/scrub/dir_repair.c
679
rd->args.owner = rd->sc->ip->i_ino;
fs/xfs/scrub/dir_repair.c
975
error = xfs_bmapi_read(sc->ip, offset, geo->leafblk - offset,
fs/xfs/scrub/dirtree.c
1003
if (xchk_inode_is_dirtree_root(sc->ip))
fs/xfs/scrub/dirtree.c
1005
if (VFS_I(sc->ip)->i_nlink == 0)
fs/xfs/scrub/dirtree.c
156
struct xfs_inode *ip,
fs/xfs/scrub/dirtree.c
178
error = xino_bitmap_set(&path->seen_inodes, ip->i_ino);
fs/xfs/scrub/dirtree.c
203
struct xfs_inode *ip,
fs/xfs/scrub/dirtree.c
235
trace_xchk_dirtree_create_path(sc, ip, dl->nr_paths, &xname, rec);
fs/xfs/scrub/dirtree.c
250
error = xchk_dirpath_append(dl, sc->ip, path, &xname, rec);
fs/xfs/scrub/dirtree.c
289
error = xfs_parent_lookup(sc->tp, sc->ip, &dl->xname, &dl->pptr_rec,
fs/xfs/scrub/dirtree.c
292
trace_xchk_dirpath_disappeared(dl->sc, sc->ip, path->path_nr,
fs/xfs/scrub/dirtree.c
308
struct xfs_inode *ip,
fs/xfs/scrub/dirtree.c
397
if (parent_ino == sc->ip->i_ino) {
fs/xfs/scrub/dirtree.c
529
trace_xchk_dirpath_walk_upwards(sc, sc->ip, path->path_nr, &dl->xname,
fs/xfs/scrub/dirtree.c
536
if (be64_to_cpu(dl->pptr_rec.p_ino) == sc->ip->i_ino) {
fs/xfs/scrub/dirtree.c
550
is_metadir = xfs_is_metadir_inode(sc->ip);
fs/xfs/scrub/dirtree.c
613
if (p->ip->i_ino != child_ino || p->dp->i_ino != *cursor)
fs/xfs/scrub/dirtree.c
635
if (p->ip->i_ino == dl->scan_ino &&
fs/xfs/scrub/dirtree.c
641
if (p->ip->i_ino == dl->scan_ino &&
fs/xfs/scrub/dirtree.c
649
p->ip, p->name);
fs/xfs/scrub/dirtree.c
672
if (!xino_bitmap_test(&path->seen_inodes, p->ip->i_ino))
fs/xfs/scrub/dirtree.c
705
trace_xchk_dirtree_live_update(dl->sc, p->dp, action, p->ip, p->delta,
fs/xfs/scrub/dirtree.c
804
if (xchk_pptr_looks_zapped(sc->ip)) {
fs/xfs/scrub/dirtree.c
814
error = xchk_xattr_walk(sc, sc->ip, xchk_dirtree_create_path,
fs/xfs/scrub/dirtree.c
920
if (!S_ISDIR(VFS_I(sc->ip)->i_mode))
fs/xfs/scrub/dirtree.c
930
dl->root_ino = xchk_inode_rootdir_inum(sc->ip);
fs/xfs/scrub/dirtree.c
931
dl->scan_ino = sc->ip->i_ino;
fs/xfs/scrub/dirtree.c
933
trace_xchk_dirtree_start(sc->ip, sc->sm, 0);
fs/xfs/scrub/dirtree.c
957
xchk_ino_xref_set_corrupt(sc, sc->ip->i_ino);
fs/xfs/scrub/dirtree.c
982
xchk_ino_set_corrupt(sc, sc->ip->i_ino);
fs/xfs/scrub/dirtree.c
985
xchk_ino_set_corrupt(sc, sc->ip->i_ino);
fs/xfs/scrub/dirtree.c
987
xchk_ino_xref_set_corrupt(sc, sc->ip->i_ino);
fs/xfs/scrub/dirtree.c
993
trace_xchk_dirtree_done(sc->ip, sc->sm, error);
fs/xfs/scrub/dirtree.h
162
int xchk_dirpath_append(struct xchk_dirtree *dl, struct xfs_inode *ip,
fs/xfs/scrub/dirtree_repair.c
403
error = xfs_trans_alloc_dir(dp, &M_RES(mp)->tr_remove, sc->ip,
fs/xfs/scrub/dirtree_repair.c
421
trace_xrep_dirtree_delete_path(dl->sc, sc->ip, path->path_nr,
fs/xfs/scrub/dirtree_repair.c
436
error = xchk_dir_lookup(sc, sc->ip, &xfs_name_dotdot, &dotdot_ino);
fs/xfs/scrub/dirtree_repair.c
451
error = xfs_dir_replace(sc->tp, sc->ip, &xfs_name_dotdot,
fs/xfs/scrub/dirtree_repair.c
458
error = xfs_droplink(sc->tp, sc->ip);
fs/xfs/scrub/dirtree_repair.c
462
error = xfs_dir_removename(sc->tp, dp, &dl->xname, sc->ip->i_ino,
fs/xfs/scrub/dirtree_repair.c
471
&dl->xname, sc->ip);
fs/xfs/scrub/dirtree_repair.c
480
xfs_dir_update_hook(dp, sc->ip, -1, &dl->xname);
fs/xfs/scrub/dirtree_repair.c
491
xfs_iunlock(sc->ip, XFS_ILOCK_EXCL);
fs/xfs/scrub/dirtree_repair.c
590
trace_xrep_dirtree_create_adoption(sc, sc->ip, dl->nr_paths,
fs/xfs/scrub/dirtree_repair.c
593
error = xchk_dirpath_append(dl, sc->ip, path, &dl->xname,
fs/xfs/scrub/findparent.c
109
if (ino != fpi->sc->ip->i_ino)
fs/xfs/scrub/findparent.c
127
trace_xrep_findparent_dirent(fpi->sc->ip, 0);
fs/xfs/scrub/findparent.c
132
trace_xrep_findparent_dirent(fpi->sc->ip, fpi->dp->i_ino);
fs/xfs/scrub/findparent.c
159
if (dp == sc->ip || dp == sc->tempip)
fs/xfs/scrub/findparent.c
176
if (xfs_is_metadir_inode(dp) != xfs_is_metadir_inode(sc->ip))
fs/xfs/scrub/findparent.c
230
if (p->ip->i_ino == sc->ip->i_ino &&
fs/xfs/scrub/findparent.c
309
ASSERT(S_ISDIR(VFS_IC(sc->ip)->i_mode));
fs/xfs/scrub/findparent.c
370
if (sc->ip == sc->mp->m_rootip) {
fs/xfs/scrub/findparent.c
376
if (sc->ip == sc->mp->m_metadirip) {
fs/xfs/scrub/findparent.c
382
if (VFS_I(sc->ip)->i_nlink == 0) {
fs/xfs/scrub/findparent.c
383
*parent_ino = xchk_inode_rootdir_inum(sc->ip);
fs/xfs/scrub/findparent.c
391
*parent_ino == sc->ip->i_ino) {
fs/xfs/scrub/findparent.c
425
if (sc->ip->i_ino == sc->mp->m_sb.sb_rootino)
fs/xfs/scrub/findparent.c
428
if (sc->ip->i_ino == sc->mp->m_sb.sb_metadirino)
fs/xfs/scrub/findparent.c
431
if (VFS_I(sc->ip)->i_nlink == 0)
fs/xfs/scrub/findparent.c
432
return xchk_inode_rootdir_inum(sc->ip);
fs/xfs/scrub/findparent.c
446
dentry = d_find_alias(VFS_I(sc->ip));
fs/xfs/scrub/findparent.c
454
ASSERT(parent->d_sb == sc->ip->i_mount->m_super);
fs/xfs/scrub/findparent.c
460
trace_xrep_findparent_from_dcache(sc->ip, XFS_I(pip)->i_ino);
fs/xfs/scrub/health.c
158
return xfs_inode_has_sickness(sc->ip, mask);
fs/xfs/scrub/health.c
229
if (!sc->ip)
fs/xfs/scrub/health.c
241
xfs_inode_mark_corrupt(sc->ip, mask);
fs/xfs/scrub/health.c
243
xfs_inode_mark_healthy(sc->ip, mask);
fs/xfs/scrub/inode.c
126
error = xchk_iget_safe(sc, sc->sm->sm_ino, &ip);
fs/xfs/scrub/inode.c
128
return xchk_install_handle_iscrub(sc, ip);
fs/xfs/scrub/inode.c
160
error = xchk_iget_agi(sc, sc->sm->sm_ino, &agi_bp, &ip);
fs/xfs/scrub/inode.c
164
return xchk_install_handle_iscrub(sc, ip);
fs/xfs/scrub/inode.c
470
if (dip->di_mode == 0 && sc->ip)
fs/xfs/scrub/inode.c
55
struct xfs_inode *ip)
fs/xfs/scrub/inode.c
59
error = xchk_install_handle_inode(sc, ip);
fs/xfs/scrub/inode.c
713
xchk_ino_xref_set_corrupt(sc, sc->ip->i_ino);
fs/xfs/scrub/inode.c
719
xchk_ino_xref_set_corrupt(sc, sc->ip->i_ino);
fs/xfs/scrub/inode.c
723
xchk_ino_xref_set_corrupt(sc, sc->ip->i_ino);
fs/xfs/scrub/inode.c
73
if (xfs_is_metadir_inode(ip) && !S_ISDIR(VFS_I(ip)->i_mode)) {
fs/xfs/scrub/inode.c
74
xchk_irele(sc, ip);
fs/xfs/scrub/inode.c
75
sc->ip = NULL;
fs/xfs/scrub/inode.c
776
error = xfs_reflink_inode_has_shared_extents(sc->tp, sc->ip,
fs/xfs/scrub/inode.c
781
if (xfs_is_reflink_inode(sc->ip) && !has_shared)
fs/xfs/scrub/inode.c
783
else if (!xfs_is_reflink_inode(sc->ip) && has_shared)
fs/xfs/scrub/inode.c
795
if (VFS_I(sc->ip)->i_nlink == 0) {
fs/xfs/scrub/inode.c
796
if (!xfs_inode_on_unlinked_list(sc->ip))
fs/xfs/scrub/inode.c
797
xchk_ino_set_corrupt(sc, sc->ip->i_ino);
fs/xfs/scrub/inode.c
799
if (xfs_inode_on_unlinked_list(sc->ip))
fs/xfs/scrub/inode.c
800
xchk_ino_set_corrupt(sc, sc->ip->i_ino);
fs/xfs/scrub/inode.c
817
if (!sc->ip) {
fs/xfs/scrub/inode.c
823
xfs_inode_to_disk(sc->ip, &di, 0);
fs/xfs/scrub/inode.c
824
xchk_dinode(sc, &di, sc->ip->i_ino);
fs/xfs/scrub/inode.c
833
if (S_ISREG(VFS_I(sc->ip)->i_mode))
fs/xfs/scrub/inode.c
834
xchk_inode_check_reflink_iflag(sc, sc->ip->i_ino);
fs/xfs/scrub/inode.c
838
xchk_inode_xref(sc, sc->ip->i_ino, &di);
fs/xfs/scrub/inode.c
93
struct xfs_inode *ip;
fs/xfs/scrub/inode_repair.c
1600
iget_error = xchk_iget(sc, ino, &sc->ip);
fs/xfs/scrub/inode_repair.c
1626
xfs_inode_mark_sick(sc->ip, ri->ino_sick_mask);
fs/xfs/scrub/inode_repair.c
1676
if (xfs_is_reflink_inode(sc->ip)) {
fs/xfs/scrub/inode_repair.c
1682
} else if (XFS_IS_REALTIME_INODE(sc->ip)) {
fs/xfs/scrub/inode_repair.c
1692
sc->ip->i_df.if_nextents = nextents;
fs/xfs/scrub/inode_repair.c
1695
ifp = xfs_ifork_ptr(sc->ip, XFS_ATTR_FORK);
fs/xfs/scrub/inode_repair.c
1712
sc->ip->i_nblocks = count + acount;
fs/xfs/scrub/inode_repair.c
1725
if (!uid_valid(VFS_I(sc->ip)->i_uid)) {
fs/xfs/scrub/inode_repair.c
1726
i_uid_write(VFS_I(sc->ip), 0);
fs/xfs/scrub/inode_repair.c
1732
if (!gid_valid(VFS_I(sc->ip)->i_gid)) {
fs/xfs/scrub/inode_repair.c
1733
i_gid_write(VFS_I(sc->ip), 0);
fs/xfs/scrub/inode_repair.c
1739
if (sc->ip->i_projid == -1U) {
fs/xfs/scrub/inode_repair.c
1740
sc->ip->i_projid = 0;
fs/xfs/scrub/inode_repair.c
1748
VFS_I(sc->ip)->i_mode &= ~(S_ISUID | S_ISGID);
fs/xfs/scrub/inode_repair.c
1753
struct xfs_inode *ip,
fs/xfs/scrub/inode_repair.c
1757
*ts = timestamp_truncate(*ts, VFS_I(ip));
fs/xfs/scrub/inode_repair.c
1763
struct xfs_inode *ip)
fs/xfs/scrub/inode_repair.c
1766
struct inode *inode = VFS_I(ip);
fs/xfs/scrub/inode_repair.c
1769
xrep_clamp_timestamp(ip, &tstamp);
fs/xfs/scrub/inode_repair.c
1773
xrep_clamp_timestamp(ip, &tstamp);
fs/xfs/scrub/inode_repair.c
1777
xrep_clamp_timestamp(ip, &tstamp);
fs/xfs/scrub/inode_repair.c
1780
xrep_clamp_timestamp(ip, &ip->i_crtime);
fs/xfs/scrub/inode_repair.c
1792
mode = VFS_I(sc->ip)->i_mode;
fs/xfs/scrub/inode_repair.c
1795
if (sc->ip->i_diflags & ~XFS_DIFLAG_ANY)
fs/xfs/scrub/inode_repair.c
1796
sc->ip->i_diflags &= ~XFS_DIFLAG_ANY;
fs/xfs/scrub/inode_repair.c
1799
if (sc->ip->i_ino == sc->mp->m_sb.sb_rbmino)
fs/xfs/scrub/inode_repair.c
1800
sc->ip->i_diflags |= XFS_DIFLAG_NEWRTBM;
fs/xfs/scrub/inode_repair.c
1802
sc->ip->i_diflags &= ~XFS_DIFLAG_NEWRTBM;
fs/xfs/scrub/inode_repair.c
1806
sc->ip->i_diflags &= ~(XFS_DIFLAG_RTINHERIT |
fs/xfs/scrub/inode_repair.c
1813
sc->ip->i_diflags &= ~(XFS_DIFLAG_REALTIME |
fs/xfs/scrub/inode_repair.c
1817
if (sc->ip->i_diflags & XFS_DIFLAG_REALTIME)
fs/xfs/scrub/inode_repair.c
1818
sc->ip->i_diflags &= ~XFS_DIFLAG_FILESTREAM;
fs/xfs/scrub/inode_repair.c
1821
if ((sc->ip->i_diflags & XFS_DIFLAG_IMMUTABLE) &&
fs/xfs/scrub/inode_repair.c
1822
(sc->ip->i_diflags & XFS_DIFLAG_APPEND))
fs/xfs/scrub/inode_repair.c
1823
sc->ip->i_diflags &= ~XFS_DIFLAG_APPEND;
fs/xfs/scrub/inode_repair.c
1826
if (sc->ip->i_diflags2 & ~XFS_DIFLAG2_ANY)
fs/xfs/scrub/inode_repair.c
1827
sc->ip->i_diflags2 &= ~XFS_DIFLAG2_ANY;
fs/xfs/scrub/inode_repair.c
1831
sc->ip->i_diflags2 &= ~XFS_DIFLAG2_REFLINK;
fs/xfs/scrub/inode_repair.c
1835
sc->ip->i_diflags2 &= ~XFS_DIFLAG2_DAX;
fs/xfs/scrub/inode_repair.c
1855
error = xfs_iread_extents(sc->tp, sc->ip, XFS_DATA_FORK);
fs/xfs/scrub/inode_repair.c
1860
ifp = xfs_ifork_ptr(sc->ip, XFS_DATA_FORK);
fs/xfs/scrub/inode_repair.c
1862
if (!xfs_iext_lookup_extent_before(sc->ip, ifp, &off, &icur, &got)) {
fs/xfs/scrub/inode_repair.c
1868
sc->ip->i_disk_size = min_t(loff_t, XFS_DIR2_SPACE_SIZE,
fs/xfs/scrub/inode_repair.c
1881
ifp = xfs_ifork_ptr(sc->ip, XFS_DATA_FORK);
fs/xfs/scrub/inode_repair.c
1882
sc->ip->i_disk_size = ifp->if_bytes;
fs/xfs/scrub/inode_repair.c
1895
switch (sc->ip->i_df.if_format) {
fs/xfs/scrub/inode_repair.c
1912
if ((sc->ip->i_diflags & XFS_DIFLAG_RTINHERIT) &&
fs/xfs/scrub/inode_repair.c
1913
(sc->ip->i_diflags & XFS_DIFLAG_EXTSZINHERIT) &&
fs/xfs/scrub/inode_repair.c
1914
xfs_extlen_to_rtxmod(sc->mp, sc->ip->i_extsize) > 0) {
fs/xfs/scrub/inode_repair.c
1915
sc->ip->i_extsize = 0;
fs/xfs/scrub/inode_repair.c
1916
sc->ip->i_diflags &= ~XFS_DIFLAG_EXTSZINHERIT;
fs/xfs/scrub/inode_repair.c
1926
struct xfs_inode *ip = sc->ip;
fs/xfs/scrub/inode_repair.c
1927
struct inode *inode = VFS_I(ip);
fs/xfs/scrub/inode_repair.c
1940
if (xchk_inode_is_sb_rooted(ip))
fs/xfs/scrub/inode_repair.c
1944
if (xfs_inode_has_attr_fork(ip))
fs/xfs/scrub/inode_repair.c
1947
return xfs_bmap_add_attrfork(sc->tp, ip,
fs/xfs/scrub/inode_repair.c
1957
if ((sc->ip->i_diflags & XFS_DIFLAG_RTINHERIT) &&
fs/xfs/scrub/inode_repair.c
1958
(sc->ip->i_diflags2 & XFS_DIFLAG2_COWEXTSIZE) &&
fs/xfs/scrub/inode_repair.c
1959
sc->ip->i_extsize % sc->mp->m_sb.sb_rextsize > 0) {
fs/xfs/scrub/inode_repair.c
1960
sc->ip->i_cowextsize = 0;
fs/xfs/scrub/inode_repair.c
1961
sc->ip->i_diflags2 &= ~XFS_DIFLAG2_COWEXTSIZE;
fs/xfs/scrub/inode_repair.c
1978
xrep_inode_timestamps(sc->ip);
fs/xfs/scrub/inode_repair.c
1985
if (S_ISDIR(VFS_I(sc->ip)->i_mode))
fs/xfs/scrub/inode_repair.c
1991
xfs_trans_log_inode(sc->tp, sc->ip, XFS_ILOG_CORE);
fs/xfs/scrub/inode_repair.c
2003
unsigned int nlink = VFS_I(sc->ip)->i_nlink;
fs/xfs/scrub/inode_repair.c
2010
if (nlink > 0 && xfs_inode_on_unlinked_list(sc->ip)) {
fs/xfs/scrub/inode_repair.c
2015
XFS_INO_TO_AGNO(sc->mp, sc->ip->i_ino));
fs/xfs/scrub/inode_repair.c
2016
error = xfs_iunlink_remove(sc->tp, pag, sc->ip);
fs/xfs/scrub/inode_repair.c
2026
if (nlink == 0 && !xfs_inode_on_unlinked_list(sc->ip)) {
fs/xfs/scrub/inode_repair.c
2027
error = xfs_iunlink(sc->tp, sc->ip);
fs/xfs/scrub/inode_repair.c
2046
if (!sc->ip) {
fs/xfs/scrub/inode_repair.c
2063
if (!sc->ip)
fs/xfs/scrub/inode_repair.c
2067
xfs_trans_ijoin(sc->tp, sc->ip, 0);
fs/xfs/scrub/inode_repair.c
2078
if (xfs_is_reflink_inode(sc->ip)) {
fs/xfs/scrub/inode_repair.c
2079
error = xfs_reflink_clear_inode_flag(sc->ip, &sc->tp);
fs/xfs/scrub/iscan.c
728
struct xfs_inode *ip)
fs/xfs/scrub/iscan.c
731
iscan->__visited_ino = ip->i_ino;
fs/xfs/scrub/iscan.h
97
void xchk_iscan_mark_visited(struct xchk_iscan *iscan, struct xfs_inode *ip);
fs/xfs/scrub/listxattr.c
111
struct xfs_inode *ip,
fs/xfs/scrub/listxattr.c
118
error = xfs_attr3_leaf_read(sc->tp, ip, ip->i_ino, 0, &leaf_bp);
fs/xfs/scrub/listxattr.c
122
error = xchk_xattr_walk_leaf_entries(sc, ip, attr_fn, leaf_bp, priv);
fs/xfs/scrub/listxattr.c
131
struct xfs_inode *ip,
fs/xfs/scrub/listxattr.c
154
error = xfs_da3_node_read(tp, ip, blkno, &bp, XFS_ATTR_FORK);
fs/xfs/scrub/listxattr.c
169
fa = xfs_da3_node_header_check(bp, ip->i_ino);
fs/xfs/scrub/listxattr.c
198
fa = xfs_attr3_leaf_header_check(bp, ip->i_ino);
fs/xfs/scrub/listxattr.c
222
struct xfs_inode *ip,
fs/xfs/scrub/listxattr.c
236
error = xchk_xattr_find_leftmost_leaf(sc, ip, &seen_dablks, &leaf_bp);
fs/xfs/scrub/listxattr.c
243
error = xchk_xattr_walk_leaf_entries(sc, ip, attr_fn, leaf_bp,
fs/xfs/scrub/listxattr.c
269
error = xfs_attr3_leaf_read(sc->tp, ip, ip->i_ino,
fs/xfs/scrub/listxattr.c
29
struct xfs_inode *ip,
fs/xfs/scrub/listxattr.c
296
struct xfs_inode *ip,
fs/xfs/scrub/listxattr.c
303
xfs_assert_ilocked(ip, XFS_ILOCK_SHARED | XFS_ILOCK_EXCL);
fs/xfs/scrub/listxattr.c
305
if (!xfs_inode_hasattr(ip))
fs/xfs/scrub/listxattr.c
308
if (ip->i_af.if_format == XFS_DINODE_FMT_LOCAL)
fs/xfs/scrub/listxattr.c
309
return xchk_xattr_walk_sf(sc, ip, attr_fn, priv);
fs/xfs/scrub/listxattr.c
312
error = xfs_iread_extents(sc->tp, ip, XFS_ATTR_FORK);
fs/xfs/scrub/listxattr.c
316
if (xfs_attr_is_leaf(ip))
fs/xfs/scrub/listxattr.c
317
return xchk_xattr_walk_leaf(sc, ip, attr_fn, priv);
fs/xfs/scrub/listxattr.c
319
return xchk_xattr_walk_node(sc, ip, attr_fn, leaf_fn, priv);
fs/xfs/scrub/listxattr.c
33
struct xfs_attr_sf_hdr *hdr = ip->i_af.if_data;
fs/xfs/scrub/listxattr.c
40
error = attr_fn(sc, ip, sfe->flags, sfe->nameval, sfe->namelen,
fs/xfs/scrub/listxattr.c
56
struct xfs_inode *ip,
fs/xfs/scrub/listxattr.c
94
error = attr_fn(sc, ip, entry->flags, name, namelen, value,
fs/xfs/scrub/listxattr.h
16
int xchk_xattr_walk(struct xfs_scrub *sc, struct xfs_inode *ip,
fs/xfs/scrub/listxattr.h
9
typedef int (*xchk_xattr_fn)(struct xfs_scrub *sc, struct xfs_inode *ip,
fs/xfs/scrub/metapath.c
120
mpath->xname.type = xfs_mode_to_ftype(VFS_I(ip)->i_mode);
fs/xfs/scrub/metapath.c
145
struct xfs_inode *ip;
fs/xfs/scrub/metapath.c
152
ip = rtg->rtg_inodes[type];
fs/xfs/scrub/metapath.c
153
if (!ip) {
fs/xfs/scrub/metapath.c
159
xfs_rtginode_path(rtg_rgno(rtg), type), ip);
fs/xfs/scrub/metapath.c
192
struct xfs_inode *ip = NULL;
fs/xfs/scrub/metapath.c
199
ip = qi->qi_uquotaip;
fs/xfs/scrub/metapath.c
202
ip = qi->qi_gquotaip;
fs/xfs/scrub/metapath.c
205
ip = qi->qi_pquotaip;
fs/xfs/scrub/metapath.c
211
if (!ip)
fs/xfs/scrub/metapath.c
215
kstrdup_const(xfs_dqinode_path(type), GFP_KERNEL), ip);
fs/xfs/scrub/metapath.c
317
xchk_ino_set_corrupt(sc, sc->ip->i_ino);
fs/xfs/scrub/metapath.c
332
xchk_ino_set_corrupt(sc, sc->ip->i_ino);
fs/xfs/scrub/metapath.c
338
if (ino != sc->ip->i_ino) {
fs/xfs/scrub/metapath.c
340
xchk_ino_set_corrupt(sc, sc->ip->i_ino);
fs/xfs/scrub/metapath.c
360
mpath->du.ip = sc->ip;
fs/xfs/scrub/metapath.c
367
trace_xrep_metapath_link(sc, mpath->path, mpath->dp, sc->ip->i_ino);
fs/xfs/scrub/metapath.c
377
struct xfs_inode *ip)
fs/xfs/scrub/metapath.c
386
if (!ip) {
fs/xfs/scrub/metapath.c
395
mpath->du.ip = ip;
fs/xfs/scrub/metapath.c
400
xfs_inode_to_parent_rec(&rec, ip);
fs/xfs/scrub/metapath.c
401
error = xfs_parent_lookup(sc->tp, ip, &mpath->xname, &rec,
fs/xfs/scrub/metapath.c
445
xfs_trans_ijoin(sc->tp, sc->ip, 0);
fs/xfs/scrub/metapath.c
465
if (ino == sc->ip->i_ino) {
fs/xfs/scrub/metapath.c
493
struct xfs_inode *ip)
fs/xfs/scrub/metapath.c
500
if (!ip || xfs_ilock_nowait(ip, XFS_ILOCK_EXCL))
fs/xfs/scrub/metapath.c
529
struct xfs_inode *ip = NULL;
fs/xfs/scrub/metapath.c
533
ASSERT(*alleged_child != sc->ip->i_ino);
fs/xfs/scrub/metapath.c
546
error = xchk_iget(sc, *alleged_child, &ip);
fs/xfs/scrub/metapath.c
556
error = xchk_metapath_ilock_parent_and_child(mpath, ip);
fs/xfs/scrub/metapath.c
562
if (ip)
fs/xfs/scrub/metapath.c
563
xfs_trans_ijoin(sc->tp, ip, 0);
fs/xfs/scrub/metapath.c
578
if (ino == sc->ip->i_ino) {
fs/xfs/scrub/metapath.c
595
error = xrep_metapath_unlink(mpath, ino, ip);
fs/xfs/scrub/metapath.c
606
if (ip) {
fs/xfs/scrub/metapath.c
607
xfs_iunlock(ip, XFS_ILOCK_EXCL);
fs/xfs/scrub/metapath.c
608
xchk_irele(sc, ip);
fs/xfs/scrub/metapath.c
641
error = xfs_attr_add_fork(sc->ip,
fs/xfs/scrub/metapath.c
91
struct xfs_inode *ip)
fs/xfs/scrub/metapath.c
99
error = xchk_install_live_inode(sc, ip);
fs/xfs/scrub/newbt.c
127
XFS_INO_TO_FSB(sc->mp, sc->ip->i_ino),
fs/xfs/scrub/newbt.c
130
xnr->ifake.if_fork_size = xfs_inode_fork_size(sc->ip, whichfork);
fs/xfs/scrub/newbt.c
147
ASSERT(xfs_is_metadir_inode(sc->ip));
fs/xfs/scrub/newbt.c
149
xfs_rmap_ino_bmbt_owner(&oinfo, sc->ip->i_ino, XFS_DATA_FORK);
fs/xfs/scrub/newbt.c
164
XFS_INO_TO_FSB(sc->mp, sc->ip->i_ino),
fs/xfs/scrub/newbt.c
167
xnr->ifake.if_fork_size = xfs_inode_fork_size(sc->ip, XFS_DATA_FORK);
fs/xfs/scrub/newbt.c
410
if (xnr->sc->ip)
fs/xfs/scrub/newbt.c
528
if (sc->ip) {
fs/xfs/scrub/nlinks.c
178
trace_xchk_nlinks_live_update(xnc->sc->mp, p->dp, action, p->ip->i_ino,
fs/xfs/scrub/nlinks.c
188
error = xchk_nlinks_update_incore(xnc, p->ip->i_ino, p->delta,
fs/xfs/scrub/nlinks.c
190
if (!error && S_ISDIR(VFS_IC(p->ip)->i_mode))
fs/xfs/scrub/nlinks.c
202
if (S_ISDIR(VFS_IC(p->ip)->i_mode) &&
fs/xfs/scrub/nlinks.c
203
xchk_iscan_want_live_update(&xnc->collect_iscan, p->ip->i_ino)) {
fs/xfs/scrub/nlinks.c
328
struct xfs_inode *ip,
fs/xfs/scrub/nlinks.c
360
trace_xchk_nlinks_collect_pptr(sc->mp, ip, &xname, pptr_rec);
fs/xfs/scrub/nlinks.c
381
struct xfs_inode *ip)
fs/xfs/scrub/nlinks.c
389
if (xfs_need_iread_extents(&ip->i_df))
fs/xfs/scrub/nlinks.c
396
if (xfs_has_parent(ip->i_mount) && xfs_inode_has_attr_fork(ip) &&
fs/xfs/scrub/nlinks.c
397
xfs_need_iread_extents(&ip->i_af))
fs/xfs/scrub/nlinks.c
405
xfs_ilock(ip, lock_mode);
fs/xfs/scrub/nlinks.c
548
struct xfs_inode *ip)
fs/xfs/scrub/nlinks.c
550
xfs_ilock(ip, XFS_IOLOCK_SHARED);
fs/xfs/scrub/nlinks.c
551
xchk_iscan_mark_visited(&xnc->collect_iscan, ip);
fs/xfs/scrub/nlinks.c
552
xfs_iunlock(ip, XFS_IOLOCK_SHARED);
fs/xfs/scrub/nlinks.c
562
struct xfs_inode *ip;
fs/xfs/scrub/nlinks.c
588
while ((error = xchk_iscan_iter(&xnc->collect_iscan, &ip)) == 1) {
fs/xfs/scrub/nlinks.c
589
if (S_ISDIR(VFS_I(ip)->i_mode))
fs/xfs/scrub/nlinks.c
590
error = xchk_nlinks_collect_dir(xnc, ip);
fs/xfs/scrub/nlinks.c
592
error = xchk_nlinks_collect_file(xnc, ip);
fs/xfs/scrub/nlinks.c
593
xchk_irele(sc, ip);
fs/xfs/scrub/nlinks.c
670
struct xfs_inode *ip)
fs/xfs/scrub/nlinks.c
683
if (xrep_is_tempfile(ip))
fs/xfs/scrub/nlinks.c
686
xfs_ilock(ip, XFS_ILOCK_SHARED);
fs/xfs/scrub/nlinks.c
695
error = xchk_nlinks_comparison_read(xnc, ip->i_ino, &obs);
fs/xfs/scrub/nlinks.c
707
if (!xfs_has_ftype(sc->mp) && S_ISDIR(VFS_I(ip)->i_mode))
fs/xfs/scrub/nlinks.c
710
total_links = xchk_nlink_total(ip, &obs);
fs/xfs/scrub/nlinks.c
711
actual_nlink = VFS_I(ip)->i_nlink;
fs/xfs/scrub/nlinks.c
713
trace_xchk_nlinks_compare_inode(sc->mp, ip, &obs);
fs/xfs/scrub/nlinks.c
721
xchk_ino_set_corrupt(sc, ip->i_ino);
fs/xfs/scrub/nlinks.c
724
xchk_ino_set_warning(sc, ip->i_ino);
fs/xfs/scrub/nlinks.c
729
xchk_ino_set_corrupt(sc, ip->i_ino);
fs/xfs/scrub/nlinks.c
733
if (S_ISDIR(VFS_I(ip)->i_mode) && actual_nlink > 0) {
fs/xfs/scrub/nlinks.c
743
xchk_ino_xref_set_corrupt(sc, ip->i_ino);
fs/xfs/scrub/nlinks.c
750
xchk_ino_set_corrupt(sc, ip->i_ino);
fs/xfs/scrub/nlinks.c
759
xchk_ino_set_corrupt(sc, ip->i_ino);
fs/xfs/scrub/nlinks.c
764
if (xchk_inode_is_dirtree_root(ip)) {
fs/xfs/scrub/nlinks.c
772
xchk_ino_set_corrupt(sc, ip->i_ino);
fs/xfs/scrub/nlinks.c
781
xchk_ino_set_corrupt(sc, ip->i_ino);
fs/xfs/scrub/nlinks.c
791
xfs_iunlock(ip, XFS_ILOCK_SHARED);
fs/xfs/scrub/nlinks.c
809
struct xfs_inode *ip;
fs/xfs/scrub/nlinks.c
819
error = xchk_iget_agi(xnc->sc, ino, &agi_bp, &ip);
fs/xfs/scrub/nlinks.c
822
error = xchk_nlinks_compare_inode(xnc, ip);
fs/xfs/scrub/nlinks.c
823
xchk_irele(xnc->sc, ip);
fs/xfs/scrub/nlinks.c
896
struct xfs_inode *ip;
fs/xfs/scrub/nlinks.c
917
while ((error = xchk_nlinks_compare_iter(xnc, &ip)) == 1) {
fs/xfs/scrub/nlinks.c
918
error = xchk_nlinks_compare_inode(xnc, ip);
fs/xfs/scrub/nlinks.c
919
xchk_iscan_mark_visited(&xnc->compare_iscan, ip);
fs/xfs/scrub/nlinks.c
920
xchk_irele(sc, ip);
fs/xfs/scrub/nlinks.h
104
if (ip && S_ISDIR(VFS_I(ip)->i_mode) && VFS_I(ip)->i_nlink)
fs/xfs/scrub/nlinks.h
99
xchk_nlink_total(struct xfs_inode *ip, const struct xchk_nlink *live)
fs/xfs/scrub/nlinks_repair.c
107
if (xrep_is_tempfile(ip))
fs/xfs/scrub/nlinks_repair.c
145
xfs_trans_ijoin(sc->tp, ip, 0);
fs/xfs/scrub/nlinks_repair.c
155
error = xfarray_load_sparse(xnc->nlinks, ip->i_ino, &obs);
fs/xfs/scrub/nlinks_repair.c
165
total_links = xchk_nlink_total(ip, &obs);
fs/xfs/scrub/nlinks_repair.c
166
actual_nlink = VFS_I(ip)->i_nlink;
fs/xfs/scrub/nlinks_repair.c
178
if (!S_ISDIR(VFS_I(ip)->i_mode) && obs.children != 0) {
fs/xfs/scrub/nlinks_repair.c
179
trace_xrep_nlinks_unfixable_inode(mp, ip, &obs);
fs/xfs/scrub/nlinks_repair.c
189
xrep_nlinks_is_orphaned(sc, ip, actual_nlink, &obs)) {
fs/xfs/scrub/nlinks_repair.c
209
error = xfarray_load_sparse(xnc->nlinks, ip->i_ino, &obs);
fs/xfs/scrub/nlinks_repair.c
214
total_links = xchk_nlink_total(ip, &obs);
fs/xfs/scrub/nlinks_repair.c
215
actual_nlink = VFS_I(ip)->i_nlink;
fs/xfs/scrub/nlinks_repair.c
223
if (total_links > 0 && xfs_inode_on_unlinked_list(ip)) {
fs/xfs/scrub/nlinks_repair.c
234
if (total_links == 0 && !xfs_inode_on_unlinked_list(ip)) {
fs/xfs/scrub/nlinks_repair.c
235
error = xfs_iunlink(sc->tp, ip);
fs/xfs/scrub/nlinks_repair.c
243
trace_xrep_nlinks_update_inode(mp, ip, &obs);
fs/xfs/scrub/nlinks_repair.c
245
set_nlink(VFS_I(ip), min_t(unsigned long long, total_links,
fs/xfs/scrub/nlinks_repair.c
255
xfs_trans_log_inode(sc->tp, ip, XFS_ILOG_CORE);
fs/xfs/scrub/nlinks_repair.c
318
ASSERT(sc->ip == NULL);
fs/xfs/scrub/nlinks_repair.c
320
while ((error = xrep_nlinks_iter(xnc, &sc->ip)) == 1) {
fs/xfs/scrub/nlinks_repair.c
328
xchk_iscan_mark_visited(&xnc->compare_iscan, sc->ip);
fs/xfs/scrub/nlinks_repair.c
329
xchk_irele(sc, sc->ip);
fs/xfs/scrub/nlinks_repair.c
330
sc->ip = NULL;
fs/xfs/scrub/nlinks_repair.c
59
struct xfs_inode *ip,
fs/xfs/scrub/nlinks_repair.c
65
if (xchk_inode_is_dirtree_root(ip) || ip == sc->orphanage)
fs/xfs/scrub/nlinks_repair.c
78
pag = xfs_perag_get(sc->mp, XFS_INO_TO_AGNO(sc->mp, sc->ip->i_ino));
fs/xfs/scrub/nlinks_repair.c
79
error = xfs_iunlink_remove(sc->tp, pag, sc->ip);
fs/xfs/scrub/nlinks_repair.c
95
struct xfs_inode *ip = sc->ip;
fs/xfs/scrub/orphanage.c
289
ASSERT(sc->ip != NULL);
fs/xfs/scrub/orphanage.c
293
if (sc->ip == sc->orphanage)
fs/xfs/scrub/orphanage.c
295
if (xchk_inode_is_sb_rooted(sc->ip))
fs/xfs/scrub/orphanage.c
297
if (xfs_is_internal_inode(sc->ip))
fs/xfs/scrub/orphanage.c
320
ASSERT(sc->ip != NULL);
fs/xfs/scrub/orphanage.c
331
if (S_ISDIR(VFS_I(sc->ip)->i_mode))
fs/xfs/scrub/orphanage.c
350
sc->ip, XFS_ILOCK_EXCL);
fs/xfs/scrub/orphanage.c
355
xfs_trans_ijoin(sc->tp, sc->ip, 0);
fs/xfs/scrub/orphanage.c
374
error = xfs_trans_reserve_quota_nblks(sc->tp, sc->ip,
fs/xfs/scrub/orphanage.c
405
xname->len = snprintf(namebuf, MAXNAMELEN, "%llu", sc->ip->i_ino);
fs/xfs/scrub/orphanage.c
406
xname->type = xfs_mode_to_ftype(VFS_I(sc->ip)->i_mode);
fs/xfs/scrub/orphanage.c
412
sc->ip->i_ino, ++incr);
fs/xfs/scrub/orphanage.c
499
while ((d_child = d_find_alias(VFS_I(sc->ip))) != NULL) {
fs/xfs/scrub/orphanage.c
531
bool isdir = S_ISDIR(VFS_I(sc->ip)->i_mode);
fs/xfs/scrub/orphanage.c
535
sc->ip->i_ino);
fs/xfs/scrub/orphanage.c
546
if (!xfs_inode_has_attr_fork(sc->ip) && xfs_has_parent(sc->mp)) {
fs/xfs/scrub/orphanage.c
549
error = xfs_bmap_add_attrfork(sc->tp, sc->ip, sf_size, true);
fs/xfs/scrub/orphanage.c
556
sc->ip->i_ino, adopt->orphanage_blkres);
fs/xfs/scrub/orphanage.c
572
xfs_bumplink(sc->tp, sc->ip);
fs/xfs/scrub/orphanage.c
573
xfs_trans_log_inode(sc->tp, sc->ip, XFS_ILOG_CORE);
fs/xfs/scrub/orphanage.c
578
error = xfs_dir_replace(sc->tp, sc->ip, &xfs_name_dotdot,
fs/xfs/scrub/orphanage.c
587
sc->orphanage, adopt->xname, sc->ip);
fs/xfs/scrub/orphanage.c
597
xfs_dir_update_hook(sc->orphanage, sc->ip, 1, adopt->xname);
fs/xfs/scrub/orphanage.c
621
trace_xrep_adoption_trans_roll(sc->orphanage, sc->ip,
fs/xfs/scrub/parent.c
128
if (sc->ip == mp->m_rootip) {
fs/xfs/scrub/parent.c
129
if (sc->ip->i_ino != mp->m_sb.sb_rootino ||
fs/xfs/scrub/parent.c
130
sc->ip->i_ino != parent_ino)
fs/xfs/scrub/parent.c
136
if (sc->ip == mp->m_metadirip) {
fs/xfs/scrub/parent.c
137
if (sc->ip->i_ino != mp->m_sb.sb_metadirino ||
fs/xfs/scrub/parent.c
138
sc->ip->i_ino != parent_ino)
fs/xfs/scrub/parent.c
144
if (sc->ip->i_ino == parent_ino) {
fs/xfs/scrub/parent.c
153
expected_nlink = VFS_I(sc->ip)->i_nlink == 0 ? 0 : 1;
fs/xfs/scrub/parent.c
172
if (dp == sc->ip || xrep_is_tempfile(dp) ||
fs/xfs/scrub/parent.c
197
if (xfs_is_metadir_inode(dp) != xfs_is_metadir_inode(sc->ip)) {
fs/xfs/scrub/parent.c
273
struct xfs_inode *ip,
fs/xfs/scrub/parent.c
308
error = xchk_dir_lookup(sc, sc->ip, &xfs_name_dotdot, &pp->parent_ino);
fs/xfs/scrub/parent.c
317
if (xchk_inode_is_dirtree_root(sc->ip)) {
fs/xfs/scrub/parent.c
318
if (sc->ip->i_ino != pp->parent_ino)
fs/xfs/scrub/parent.c
327
if (VFS_I(sc->ip)->i_nlink == 0)
fs/xfs/scrub/parent.c
334
error = xchk_xattr_walk(sc, sc->ip, xchk_parent_scan_dotdot, NULL, pp);
fs/xfs/scrub/parent.c
401
if (child_ino != sc->ip->i_ino) {
fs/xfs/scrub/parent.c
417
struct xfs_inode *ip;
fs/xfs/scrub/parent.c
428
error = xchk_iget(sc, parent_ino, &ip);
fs/xfs/scrub/parent.c
437
if (!S_ISDIR(VFS_I(ip)->i_mode)) {
fs/xfs/scrub/parent.c
443
if (VFS_I(ip)->i_generation != be32_to_cpu(pptr->p_gen)) {
fs/xfs/scrub/parent.c
448
*dpp = ip;
fs/xfs/scrub/parent.c
451
xchk_irele(sc, ip);
fs/xfs/scrub/parent.c
463
struct xfs_inode *ip,
fs/xfs/scrub/parent.c
493
if (parent_ino == sc->ip->i_ino) {
fs/xfs/scrub/parent.c
515
trace_xchk_parent_defer(sc->ip, &xname, dp->i_ino);
fs/xfs/scrub/parent.c
556
error = xfs_parent_lookup(sc->tp, sc->ip, xname, pptr, &pp->pptr_args);
fs/xfs/scrub/parent.c
602
trace_xchk_parent_slowpath(sc->ip, xname, dp->i_ino);
fs/xfs/scrub/parent.c
613
trace_xchk_parent_ultraslowpath(sc->ip, xname, dp->i_ino);
fs/xfs/scrub/parent.c
675
struct xfs_inode *ip,
fs/xfs/scrub/parent.c
716
error = xchk_xattr_walk(sc, sc->ip, xchk_parent_count_pptr,
fs/xfs/scrub/parent.c
727
if (S_ISDIR(VFS_I(sc->ip)->i_mode)) {
fs/xfs/scrub/parent.c
728
if (xchk_inode_is_dirtree_root(sc->ip))
fs/xfs/scrub/parent.c
731
if (VFS_I(sc->ip)->i_nlink == 0 && pp->pptrs_found > 0)
fs/xfs/scrub/parent.c
732
xchk_ino_set_corrupt(sc, sc->ip->i_ino);
fs/xfs/scrub/parent.c
733
else if (VFS_I(sc->ip)->i_nlink > 0 &&
fs/xfs/scrub/parent.c
735
xchk_ino_set_corrupt(sc, sc->ip->i_ino);
fs/xfs/scrub/parent.c
742
if (xfs_has_metadir(sc->mp) && xchk_inode_is_sb_rooted(sc->ip))
fs/xfs/scrub/parent.c
745
if (VFS_I(sc->ip)->i_nlink != pp->pptrs_found)
fs/xfs/scrub/parent.c
746
xchk_ino_set_corrupt(sc, sc->ip->i_ino);
fs/xfs/scrub/parent.c
75
if (sc->ip->i_ino == ino)
fs/xfs/scrub/parent.c
779
error = xchk_xattr_walk(sc, sc->ip, xchk_parent_scan_attr, NULL, pp);
fs/xfs/scrub/parent.c
812
if (S_ISDIR(VFS_I(sc->ip)->i_mode)) {
fs/xfs/scrub/parent.c
855
if (!S_ISDIR(VFS_I(sc->ip)->i_mode))
fs/xfs/scrub/parent.c
859
if (!xfs_verify_dir_ino(mp, sc->ip->i_ino)) {
fs/xfs/scrub/parent.c
869
error = xchk_dir_lookup(sc, sc->ip, &xfs_name_dotdot,
fs/xfs/scrub/parent.c
903
struct xfs_inode *ip)
fs/xfs/scrub/parent.c
905
struct inode *inode = VFS_I(ip);
fs/xfs/scrub/parent.c
907
ASSERT(xfs_has_parent(ip->i_mount));
fs/xfs/scrub/parent.c
921
if (xchk_inode_is_dirtree_root(ip))
fs/xfs/scrub/parent.c
929
if (xchk_inode_is_sb_rooted(ip))
fs/xfs/scrub/parent.c
937
if (!xfs_inode_has_attr_fork(ip))
fs/xfs/scrub/parent.c
941
if (xfs_ifork_zapped(ip, XFS_ATTR_FORK))
fs/xfs/scrub/parent.c
951
return ip->i_af.if_format == XFS_DINODE_FMT_EXTENTS &&
fs/xfs/scrub/parent.c
952
ip->i_af.if_nextents == 0;
fs/xfs/scrub/parent_repair.c
1163
error = xchk_xattr_walk(sc, sc->ip, xrep_parent_stash_xattr,
fs/xfs/scrub/parent_repair.c
1174
error = xchk_xattr_walk(sc, sc->ip, xrep_parent_stash_xattr,
fs/xfs/scrub/parent_repair.c
1208
return xfs_attr_add_fork(sc->ip, sizeof(struct xfs_attr_sf_hdr), 1);
fs/xfs/scrub/parent_repair.c
1331
if (xchk_inode_is_dirtree_root(sc->ip)) {
fs/xfs/scrub/parent_repair.c
1332
xrep_findparent_scan_found(&rp->pscan, sc->ip->i_ino);
fs/xfs/scrub/parent_repair.c
1374
xchk_inode_is_sb_rooted(sc->ip))
fs/xfs/scrub/parent_repair.c
1384
if (S_ISDIR(VFS_I(sc->ip)->i_mode))
fs/xfs/scrub/parent_repair.c
1394
struct xfs_inode *ip,
fs/xfs/scrub/parent_repair.c
1427
struct xfs_inode *ip = sc->ip;
fs/xfs/scrub/parent_repair.c
1434
error = xchk_xattr_walk(sc, ip, xrep_parent_count_pptr, NULL, rp);
fs/xfs/scrub/parent_repair.c
1443
if (xfs_has_metadir(sc->mp) && xchk_inode_is_sb_rooted(sc->ip))
fs/xfs/scrub/parent_repair.c
1446
if (rp->parents > 0 && xfs_inode_on_unlinked_list(ip)) {
fs/xfs/scrub/parent_repair.c
1447
xfs_trans_ijoin(sc->tp, sc->ip, 0);
fs/xfs/scrub/parent_repair.c
1454
pag = xfs_perag_get(sc->mp, XFS_INO_TO_AGNO(sc->mp, ip->i_ino));
fs/xfs/scrub/parent_repair.c
1460
error = xfs_iunlink_remove(sc->tp, pag, ip);
fs/xfs/scrub/parent_repair.c
1464
} else if (rp->parents == 0 && !xfs_inode_on_unlinked_list(ip)) {
fs/xfs/scrub/parent_repair.c
1465
xfs_trans_ijoin(sc->tp, sc->ip, 0);
fs/xfs/scrub/parent_repair.c
1472
error = xfs_iunlink(sc->tp, ip);
fs/xfs/scrub/parent_repair.c
1478
if (VFS_I(ip)->i_nlink != rp->parents) {
fs/xfs/scrub/parent_repair.c
1480
xfs_trans_ijoin(sc->tp, sc->ip, 0);
fs/xfs/scrub/parent_repair.c
1484
set_nlink(VFS_I(ip), min_t(unsigned long long, rp->parents,
fs/xfs/scrub/parent_repair.c
1490
xfs_trans_log_inode(sc->tp, ip, XFS_ILOG_CORE);
fs/xfs/scrub/parent_repair.c
1611
if (xfs_has_parent(sc->mp) && !S_ISDIR(VFS_I(sc->ip)->i_mode)) {
fs/xfs/scrub/parent_repair.c
251
xfs_inode_measure_sickness(sc->ip, &sick, &checked);
fs/xfs/scrub/parent_repair.c
305
return xfs_parent_set(sc->tempip, sc->ip->i_ino, xname,
fs/xfs/scrub/parent_repair.c
312
return xfs_parent_unset(sc->tempip, sc->ip->i_ino, xname,
fs/xfs/scrub/parent_repair.c
437
if (ino != rp->sc->ip->i_ino)
fs/xfs/scrub/parent_repair.c
445
if (name->type != xfs_mode_to_ftype(VFS_I(sc->ip)->i_mode))
fs/xfs/scrub/parent_repair.c
470
const struct xfs_inode *ip)
fs/xfs/scrub/parent_repair.c
472
return ip != rp->sc->ip && !xrep_is_tempfile(ip);
fs/xfs/scrub/parent_repair.c
484
struct xfs_inode *ip)
fs/xfs/scrub/parent_repair.c
489
if (!xrep_parent_want_scan(rp, ip))
fs/xfs/scrub/parent_repair.c
492
if (S_ISDIR(VFS_I(ip)->i_mode) && xfs_need_iread_extents(&ip->i_df)) {
fs/xfs/scrub/parent_repair.c
498
xfs_ilock(ip, lock_mode);
fs/xfs/scrub/parent_repair.c
509
struct xfs_inode *ip)
fs/xfs/scrub/parent_repair.c
514
lock_mode = xrep_parent_scan_ilock(rp, ip);
fs/xfs/scrub/parent_repair.c
516
if (!xrep_parent_want_scan(rp, ip))
fs/xfs/scrub/parent_repair.c
519
if (S_ISDIR(VFS_I(ip)->i_mode)) {
fs/xfs/scrub/parent_repair.c
524
if (xchk_dir_looks_zapped(ip)) {
fs/xfs/scrub/parent_repair.c
529
error = xchk_dir_walk(rp->sc, ip, xrep_parent_scan_dirent, rp);
fs/xfs/scrub/parent_repair.c
535
xchk_iscan_mark_visited(&rp->pscan.iscan, ip);
fs/xfs/scrub/parent_repair.c
536
xfs_iunlock(ip, lock_mode);
fs/xfs/scrub/parent_repair.c
560
struct xfs_inode *ip;
fs/xfs/scrub/parent_repair.c
574
while ((error = xchk_iscan_iter(&rp->pscan.iscan, &ip)) == 1) {
fs/xfs/scrub/parent_repair.c
577
error = xrep_parent_scan_file(rp, ip);
fs/xfs/scrub/parent_repair.c
578
xchk_irele(sc, ip);
fs/xfs/scrub/parent_repair.c
648
if (p->ip->i_ino == sc->ip->i_ino &&
fs/xfs/scrub/parent_repair.c
681
error = xchk_dir_lookup(sc, sc->ip, &xfs_name_dotdot, &ino);
fs/xfs/scrub/parent_repair.c
685
xfs_trans_ijoin(sc->tp, sc->ip, 0);
fs/xfs/scrub/parent_repair.c
687
trace_xrep_parent_reset_dotdot(sc->ip, rp->pscan.parent_ino);
fs/xfs/scrub/parent_repair.c
695
error = xfs_trans_reserve_more_inode(sc->tp, sc->ip, spaceres, 0,
fs/xfs/scrub/parent_repair.c
700
error = xfs_dir_replace(sc->tp, sc->ip, &xfs_name_dotdot,
fs/xfs/scrub/parent_repair.c
716
struct xfs_inode *ip,
fs/xfs/scrub/parent_repair.c
754
error = xchk_xattr_walk(sc, sc->ip, xrep_parent_lookup_pptr, NULL,
fs/xfs/scrub/parent_repair.c
777
if (S_ISDIR(VFS_I(sc->ip)->i_mode)) {
fs/xfs/scrub/parent_repair.c
784
error = xchk_dir_lookup(sc, sc->ip, &xfs_name_dotdot,
fs/xfs/scrub/parent_repair.c
832
if (S_ISDIR(VFS_I(sc->ip)->i_mode))
fs/xfs/scrub/parent_repair.c
833
error = xchk_dir_lookup(sc, sc->ip, &xfs_name_dotdot,
fs/xfs/scrub/parent_repair.c
844
if (orig_parent == new_parent && VFS_I(sc->ip)->i_nlink > 0) {
fs/xfs/scrub/parent_repair.c
893
struct xfs_inode *ip,
fs/xfs/scrub/parent_repair.c
904
.dp = ip,
fs/xfs/scrub/parent_repair.c
909
.owner = ip->i_ino,
fs/xfs/scrub/parent_repair.c
932
struct xfs_inode *ip,
fs/xfs/scrub/parent_repair.c
952
error = xrep_parent_fetch_xattr_remote(rp, ip, attr_flags,
fs/xfs/scrub/parent_repair.c
987
.owner = rp->sc->ip->i_ino,
fs/xfs/scrub/quota.c
102
error = xfs_bmapi_read(sc->ip, offset, 1, &irec, &nmaps, 0);
fs/xfs/scrub/quota.c
279
ifp = xfs_ifork_ptr(sc->ip, XFS_DATA_FORK);
fs/xfs/scrub/quota_repair.c
125
error = xfs_bmapi_read(sc->ip, offset, 1, &irec, &nmaps, 0);
fs/xfs/scrub/quota_repair.c
414
ifp = xfs_ifork_ptr(sc->ip, XFS_DATA_FORK);
fs/xfs/scrub/quota_repair.c
433
xfs_trans_ijoin(sc->tp, sc->ip, 0);
fs/xfs/scrub/quota_repair.c
437
error = xfs_bmapi_write(sc->tp, sc->ip,
fs/xfs/scrub/quota_repair.c
452
xfs_trans_ijoin(sc->tp, sc->ip, 0);
fs/xfs/scrub/quota_repair.c
458
error = xfs_bunmapi_range(&sc->tp, sc->ip, 0,
fs/xfs/scrub/quota_repair.c
465
error = xfs_reflink_cancel_cow_blocks(sc->ip, &sc->tp, 0,
fs/xfs/scrub/quota_repair.c
469
sc->ip->i_diflags2 &= ~XFS_DIFLAG2_REFLINK;
fs/xfs/scrub/quota_repair.c
475
xfs_trans_log_inode(sc->tp, sc->ip, XFS_ILOG_CORE);
fs/xfs/scrub/quota_repair.c
66
xfs_trans_ijoin(sc->tp, sc->ip, 0);
fs/xfs/scrub/quota_repair.c
74
error = xfs_bmapi_write(sc->tp, sc->ip, dq->q_fileoffset,
fs/xfs/scrub/quotacheck.c
392
struct xfs_inode *ip)
fs/xfs/scrub/quotacheck.c
398
bool isreg = S_ISREG(VFS_I(ip)->i_mode);
fs/xfs/scrub/quotacheck.c
401
if (xfs_is_metadir_inode(ip) ||
fs/xfs/scrub/quotacheck.c
402
xfs_is_quota_inode(&tp->t_mountp->m_sb, ip->i_ino)) {
fs/xfs/scrub/quotacheck.c
409
xchk_iscan_mark_visited(&xqc->iscan, ip);
fs/xfs/scrub/quotacheck.c
414
xfs_ilock(ip, XFS_IOLOCK_SHARED);
fs/xfs/scrub/quotacheck.c
416
xfs_ilock(ip, XFS_MMAPLOCK_SHARED);
fs/xfs/scrub/quotacheck.c
417
if (XFS_IS_REALTIME_INODE(ip)) {
fs/xfs/scrub/quotacheck.c
423
ilock_flags = xfs_ilock_data_map_shared(ip);
fs/xfs/scrub/quotacheck.c
424
error = xfs_iread_extents(tp, ip, XFS_DATA_FORK);
fs/xfs/scrub/quotacheck.c
429
xfs_ilock(ip, XFS_ILOCK_SHARED);
fs/xfs/scrub/quotacheck.c
431
xfs_inode_count_blocks(tp, ip, &nblks, &rtblks);
fs/xfs/scrub/quotacheck.c
441
id = xfs_qm_id_for_quotatype(ip, XFS_DQTYPE_USER);
fs/xfs/scrub/quotacheck.c
449
id = xfs_qm_id_for_quotatype(ip, XFS_DQTYPE_GROUP);
fs/xfs/scrub/quotacheck.c
457
id = xfs_qm_id_for_quotatype(ip, XFS_DQTYPE_PROJ);
fs/xfs/scrub/quotacheck.c
465
xchk_iscan_mark_visited(&xqc->iscan, ip);
fs/xfs/scrub/quotacheck.c
475
xfs_iunlock(ip, ilock_flags);
fs/xfs/scrub/quotacheck.c
477
xfs_iunlock(ip, XFS_MMAPLOCK_SHARED);
fs/xfs/scrub/quotacheck.c
478
xfs_iunlock(ip, XFS_IOLOCK_SHARED);
fs/xfs/scrub/quotacheck.c
488
struct xfs_inode *ip;
fs/xfs/scrub/quotacheck.c
510
while ((error = xchk_iscan_iter(&xqc->iscan, &ip)) == 1) {
fs/xfs/scrub/quotacheck.c
511
error = xqcheck_collect_inode(xqc, ip);
fs/xfs/scrub/quotacheck.c
512
xchk_irele(sc, ip);
fs/xfs/scrub/readdir.c
335
args.owner = sc->ip->i_ino;
fs/xfs/scrub/readdir.c
354
struct xfs_inode *ip)
fs/xfs/scrub/readdir.c
359
if (!xfs_ilock_nowait(ip, XFS_IOLOCK_SHARED))
fs/xfs/scrub/readdir.c
363
if (!xfs_ilock_nowait(ip, XFS_ILOCK_EXCL))
fs/xfs/scrub/readdir.c
370
xfs_iunlock(ip, XFS_IOLOCK_SHARED);
fs/xfs/scrub/readdir.c
396
struct xfs_inode *ip,
fs/xfs/scrub/readdir.c
405
*lockmode = xchk_dir_trylock_both(sc, ip);
fs/xfs/scrub/readdir.h
19
int xchk_dir_trylock_for_pptrs(struct xfs_scrub *sc, struct xfs_inode *ip,
fs/xfs/scrub/reap.c
1100
ASSERT(sc->ip != NULL);
fs/xfs/scrub/reap.c
111
struct xfs_inode *ip;
fs/xfs/scrub/reap.c
1132
error = xfs_trans_roll_inode(&sc->tp, sc->ip);
fs/xfs/scrub/reap.c
1166
ASSERT(sc->ip != NULL);
fs/xfs/scrub/reap.c
1204
ASSERT(sc->ip != NULL);
fs/xfs/scrub/reap.c
1205
ASSERT(xfs_is_metadir_inode(sc->ip));
fs/xfs/scrub/reap.c
1208
xfs_rmap_ino_bmbt_owner(&oinfo, sc->ip->i_ino, XFS_DATA_FORK);
fs/xfs/scrub/reap.c
1252
xfs_rmap_ino_owner(&oinfo, rs->ip->i_ino, rs->whichfork,
fs/xfs/scrub/reap.c
1355
error = xfs_bmapi_read(rs->ip, off, max_off - off, &hmap,
fs/xfs/scrub/reap.c
1449
xfs_bmap_unmap_extent(sc->tp, rs->ip, rs->whichfork, imap);
fs/xfs/scrub/reap.c
1450
xfs_trans_mod_dquot_byino(sc->tp, rs->ip, XFS_TRANS_DQ_BCOUNT,
fs/xfs/scrub/reap.c
1452
xfs_rmap_unmap_extent(sc->tp, rs->ip, rs->whichfork, imap);
fs/xfs/scrub/reap.c
1482
xfs_bmap_unmap_extent(sc->tp, rs->ip, rs->whichfork, imap);
fs/xfs/scrub/reap.c
1483
xfs_trans_mod_dquot_byino(sc->tp, rs->ip, XFS_TRANS_DQ_BCOUNT,
fs/xfs/scrub/reap.c
1605
trace_xreap_ifork_extent(sc, rs->ip, rs->whichfork, imap);
fs/xfs/scrub/reap.c
1645
struct xfs_inode *ip,
fs/xfs/scrub/reap.c
1650
.ip = ip,
fs/xfs/scrub/reap.c
1658
ASSERT(ip == sc->ip || ip == sc->tempip);
fs/xfs/scrub/reap.c
1659
ASSERT(whichfork == XFS_ATTR_FORK || !XFS_IS_REALTIME_INODE(ip));
fs/xfs/scrub/reap.c
1667
error = xfs_bmapi_read(ip, off, XFS_MAX_FILEOFF - off, &imap,
fs/xfs/scrub/reap.c
719
ASSERT(sc->ip == NULL);
fs/xfs/scrub/reap.c
768
ASSERT(sc->ip == NULL);
fs/xfs/scrub/reap.c
800
ASSERT(sc->ip != NULL);
fs/xfs/scrub/reap.c
844
error = xfs_trans_roll_inode(&sc->tp, sc->ip);
fs/xfs/scrub/reap.c
881
ASSERT(sc->ip != NULL);
fs/xfs/scrub/reap.h
16
int xrep_reap_ifork(struct xfs_scrub *sc, struct xfs_inode *ip, int whichfork);
fs/xfs/scrub/repair.c
1215
if (xfs_inode_hasattr(sc->ip)) {
fs/xfs/scrub/repair.c
1222
if (xfs_is_reflink_inode(sc->ip)) {
fs/xfs/scrub/repair.c
1224
xfs_trans_ijoin(sc->tp, sc->ip, 0);
fs/xfs/scrub/repair.c
1225
error = xfs_reflink_clear_inode_flag(sc->ip, &sc->tp);
fs/xfs/scrub/repair.c
1234
if (xfs_inode_hasattr(sc->ip) && !xfs_has_metadir(sc->mp)) {
fs/xfs/scrub/repair.c
1237
xfs_trans_ijoin(sc->tp, sc->ip, 0);
fs/xfs/scrub/repair.c
1348
new_blocks - sc->ip->i_nblocks;
fs/xfs/scrub/repair.c
1350
sc->ip->i_nblocks = new_blocks;
fs/xfs/scrub/repair.c
1352
xfs_trans_log_inode(sc->tp, sc->ip, XFS_ILOG_CORE);
fs/xfs/scrub/repair.c
1354
xfs_trans_mod_dquot_byino(sc->tp, sc->ip, XFS_TRANS_DQ_BCOUNT,
fs/xfs/scrub/repair.c
204
if (!sc->ip)
fs/xfs/scrub/repair.c
206
return xfs_trans_roll_inode(&sc->tp, sc->ip);
fs/xfs/scrub/repair.c
810
ASSERT(sc->ip != NULL);
fs/xfs/scrub/repair.c
812
error = xfs_qm_dqattach(sc->ip);
fs/xfs/scrub/repair.c
819
(unsigned long long)sc->ip->i_ino, error);
fs/xfs/scrub/repair.c
820
if (XFS_IS_UQUOTA_ON(sc->mp) && !sc->ip->i_udquot)
fs/xfs/scrub/repair.c
822
if (XFS_IS_GQUOTA_ON(sc->mp) && !sc->ip->i_gdquot)
fs/xfs/scrub/repair.c
824
if (XFS_IS_PQUOTA_ON(sc->mp) && !sc->ip->i_pdquot)
fs/xfs/scrub/repair.c
852
inode_has_nrext64 = xfs_inode_has_large_extent_counts(sc->ip);
fs/xfs/scrub/repair.c
865
sc->ip->i_diflags2 |= XFS_DIFLAG2_NREXT64;
fs/xfs/scrub/repair.c
866
xfs_trans_log_inode(sc->tp, sc->ip, XFS_ILOG_CORE);
fs/xfs/scrub/rmap_repair.c
443
struct xfs_inode *ip,
fs/xfs/scrub/rmap_repair.c
452
ifp = xfs_ifork_ptr(ip, rf->whichfork);
fs/xfs/scrub/rmap_repair.c
453
cur = xfs_bmbt_init_cursor(rr->sc->mp, rr->sc->tp, ip, rf->whichfork);
fs/xfs/scrub/rmap_repair.c
455
if (!xfs_ifork_is_realtime(ip, rf->whichfork) &&
fs/xfs/scrub/rmap_repair.c
505
struct xfs_inode *ip)
fs/xfs/scrub/rmap_repair.c
516
switch (ip->i_metatype) {
fs/xfs/scrub/rmap_repair.c
529
if (ip == rtg->rtg_inodes[type])
fs/xfs/scrub/rmap_repair.c
537
if (ip->i_nblocks) {
fs/xfs/scrub/rmap_repair.c
545
switch (ip->i_metatype) {
fs/xfs/scrub/rmap_repair.c
569
struct xfs_inode *ip,
fs/xfs/scrub/rmap_repair.c
573
.accum = { .rm_owner = ip->i_ino, },
fs/xfs/scrub/rmap_repair.c
577
struct xfs_ifork *ifp = xfs_ifork_ptr(ip, whichfork);
fs/xfs/scrub/rmap_repair.c
591
error = xrep_rmap_scan_bmbt(&rf, ip, &mappings_done);
fs/xfs/scrub/rmap_repair.c
597
if (xfs_ifork_is_realtime(ip, whichfork))
fs/xfs/scrub/rmap_repair.c
602
return xrep_rmap_scan_meta_btree(&rf, ip);
fs/xfs/scrub/rmap_repair.c
616
struct xfs_inode *ip)
fs/xfs/scrub/rmap_repair.c
620
if (xfs_need_iread_extents(&ip->i_df)) {
fs/xfs/scrub/rmap_repair.c
625
if (xfs_inode_has_attr_fork(ip) && xfs_need_iread_extents(&ip->i_af))
fs/xfs/scrub/rmap_repair.c
629
xfs_ilock(ip, lock_mode);
fs/xfs/scrub/rmap_repair.c
637
struct xfs_inode *ip)
fs/xfs/scrub/rmap_repair.c
639
unsigned int lock_mode = xrep_rmap_scan_ilock(ip);
fs/xfs/scrub/rmap_repair.c
643
error = xrep_rmap_scan_ifork(rr, ip, XFS_DATA_FORK);
fs/xfs/scrub/rmap_repair.c
648
error = xrep_rmap_scan_ifork(rr, ip, XFS_ATTR_FORK);
fs/xfs/scrub/rmap_repair.c
654
xchk_iscan_mark_visited(&rr->iscan, ip);
fs/xfs/scrub/rmap_repair.c
656
xfs_iunlock(ip, lock_mode);
fs/xfs/scrub/rmap_repair.c
908
struct xfs_inode *ip;
fs/xfs/scrub/rmap_repair.c
954
while ((error = xchk_iscan_iter(&rr->iscan, &ip)) == 1) {
fs/xfs/scrub/rmap_repair.c
955
error = xrep_rmap_scan_inode(rr, ip);
fs/xfs/scrub/rmap_repair.c
956
xchk_irele(sc, ip);
fs/xfs/scrub/rtbitmap.c
153
struct xfs_inode *ip = sc->ip;
fs/xfs/scrub/rtbitmap.c
159
endoff = XFS_B_TO_FSB(mp, ip->i_disk_size);
fs/xfs/scrub/rtbitmap.c
160
if (xfs_iext_lookup_extent(ip, &ip->i_df, endoff, &icur, &map)) {
fs/xfs/scrub/rtbitmap.c
173
error = xfs_bmapi_read(ip, off, endoff - off, &map, &nmap,
fs/xfs/scrub/rtbitmap_repair.c
385
hdr->rt_owner = cpu_to_be64(sc->ip->i_ino);
fs/xfs/scrub/rtbitmap_repair.c
413
ASSERT(sc->ip != NULL);
fs/xfs/scrub/rtbitmap_repair.c
422
error = xfs_bmapi_read(sc->ip, off, len - off, &map, &nmaps,
fs/xfs/scrub/rtbitmap_repair.c
456
error = xfs_bmapi_write(sc->tp, sc->ip, map.br_startoff,
fs/xfs/scrub/rtbitmap_repair.c
497
sc->ip->i_disk_size = roundup_64(sc->ip->i_disk_size,
fs/xfs/scrub/rtbitmap_repair.c
500
if (sc->ip->i_disk_size < XFS_FSB_TO_B(mp, rtb->rbmblocks))
fs/xfs/scrub/rtbitmap_repair.c
501
sc->ip->i_disk_size = XFS_FSB_TO_B(mp, rtb->rbmblocks);
fs/xfs/scrub/rtbitmap_repair.c
503
xfs_trans_log_inode(sc->tp, sc->ip, XFS_ILOG_CORE);
fs/xfs/scrub/rtbitmap_repair.c
551
xfs_trans_ijoin(sc->tp, sc->ip, 0);
fs/xfs/scrub/rtrefcount_repair.c
363
if (rec->rm_owner != rr->sc->ip->i_ino)
fs/xfs/scrub/rtrefcount_repair.c
730
xfs_trans_ijoin(sc->tp, sc->ip, 0);
fs/xfs/scrub/rtrmap.c
247
struct xfs_inode *ip = rtg_rmap(sc->sr.rtg);
fs/xfs/scrub/rtrmap.c
256
xfs_rmap_ino_bmbt_owner(&oinfo, ip->i_ino, XFS_DATA_FORK);
fs/xfs/scrub/rtrmap_repair.c
266
struct xfs_inode *ip,
fs/xfs/scrub/rtrmap_repair.c
271
struct xfs_ifork *ifp = xfs_ifork_ptr(ip, XFS_DATA_FORK);
fs/xfs/scrub/rtrmap_repair.c
285
cur = xfs_bmbt_init_cursor(rr->sc->mp, rr->sc->tp, ip, XFS_DATA_FORK);
fs/xfs/scrub/rtrmap_repair.c
324
struct xfs_inode *ip)
fs/xfs/scrub/rtrmap_repair.c
327
.accum = { .rm_owner = ip->i_ino, },
fs/xfs/scrub/rtrmap_repair.c
330
struct xfs_ifork *ifp = xfs_ifork_ptr(ip, XFS_DATA_FORK);
fs/xfs/scrub/rtrmap_repair.c
341
error = xrep_rtrmap_scan_bmbt(&rf, ip, &mappings_done);
fs/xfs/scrub/rtrmap_repair.c
357
struct xfs_inode *ip)
fs/xfs/scrub/rtrmap_repair.c
363
if (rr->sc->ip == ip)
fs/xfs/scrub/rtrmap_repair.c
366
lock_mode = xfs_ilock_data_map_shared(ip);
fs/xfs/scrub/rtrmap_repair.c
369
if (XFS_IS_REALTIME_INODE(ip)) {
fs/xfs/scrub/rtrmap_repair.c
370
error = xrep_rtrmap_scan_dfork(rr, ip);
fs/xfs/scrub/rtrmap_repair.c
375
xchk_iscan_mark_visited(&rr->iscan, ip);
fs/xfs/scrub/rtrmap_repair.c
377
xfs_iunlock(ip, lock_mode);
fs/xfs/scrub/rtrmap_repair.c
395
if (rec->rm_owner != rr->sc->ip->i_ino)
fs/xfs/scrub/rtrmap_repair.c
544
struct xfs_inode *ip;
fs/xfs/scrub/rtrmap_repair.c
582
while ((error = xchk_iscan_iter(&rr->iscan, &ip)) == 1) {
fs/xfs/scrub/rtrmap_repair.c
583
error = xrep_rtrmap_scan_inode(rr, ip);
fs/xfs/scrub/rtrmap_repair.c
584
xchk_irele(sc, ip);
fs/xfs/scrub/rtrmap_repair.c
960
xfs_trans_ijoin(sc->tp, sc->ip, 0);
fs/xfs/scrub/rtsummary.c
233
struct xfs_inode *ip = sc->ip;
fs/xfs/scrub/rtsummary.c
245
endoff = XFS_B_TO_FSB(mp, ip->i_disk_size);
fs/xfs/scrub/rtsummary.c
246
if (xfs_iext_lookup_extent(ip, &ip->i_df, endoff, &icur, &map)) {
fs/xfs/scrub/rtsummary.c
260
error = xfs_bmapi_read(ip, off, endoff - off, &map, &nmap,
fs/xfs/scrub/rtsummary_repair.c
145
xfs_trans_ijoin(sc->tp, sc->ip, 0);
fs/xfs/scrub/rtsummary_repair.c
94
hdr->rt_owner = cpu_to_be64(sc->ip->i_ino);
fs/xfs/scrub/scrub.c
200
ASSERT(sc->ip == sub->sc.ip);
fs/xfs/scrub/scrub.c
244
if (sc->ip) {
fs/xfs/scrub/scrub.c
247
xchk_irele(sc, sc->ip);
fs/xfs/scrub/scrub.c
248
sc->ip = NULL;
fs/xfs/scrub/scrub.c
876
struct xfs_inode *ip;
fs/xfs/scrub/scrub.c
880
error = xfs_iget(mp, tp, head->svh_ino, XCHK_IGET_FLAGS, 0, &ip);
fs/xfs/scrub/scrub.c
885
if (VFS_I(ip)->i_generation != head->svh_gen) {
fs/xfs/scrub/scrub.c
886
xfs_irele(ip);
fs/xfs/scrub/scrub.c
890
return ip;
fs/xfs/scrub/scrub.h
151
struct xfs_inode *ip;
fs/xfs/scrub/symlink.c
51
struct xfs_inode *ip = sc->ip;
fs/xfs/scrub/symlink.c
56
if (!S_ISLNK(VFS_I(ip)->i_mode))
fs/xfs/scrub/symlink.c
64
ifp = xfs_ifork_ptr(ip, XFS_DATA_FORK);
fs/xfs/scrub/symlink.c
65
len = ip->i_disk_size;
fs/xfs/scrub/symlink.c
75
if (len > xfs_inode_data_fork_size(ip) ||
fs/xfs/scrub/symlink.c
76
len > strnlen(ifp->if_data, xfs_inode_data_fork_size(ip)))
fs/xfs/scrub/symlink.c
82
error = xfs_symlink_remote_read(sc->ip, sc->buf);
fs/xfs/scrub/symlink_repair.c
114
len = min_t(loff_t, ip->i_disk_size, XFS_SYMLINK_MAXLEN);
fs/xfs/scrub/symlink_repair.c
116
error = xfs_bmapi_read(ip, 0, fsblocks, mval, &nmaps, 0);
fs/xfs/scrub/symlink_repair.c
147
hdr_ok = xfs_symlink_hdr_ok(ip->i_ino, offset, byte_cnt, bp);
fs/xfs/scrub/symlink_repair.c
167
struct xfs_inode *ip = sc->ip;
fs/xfs/scrub/symlink_repair.c
173
ifp = xfs_ifork_ptr(ip, XFS_DATA_FORK);
fs/xfs/scrub/symlink_repair.c
183
if (xfs_inode_has_sickness(sc->ip, XFS_SICK_INO_SYMLINK_ZAPPED) &&
fs/xfs/scrub/symlink_repair.c
184
sc->ip->i_disk_size == 1 && old_target[0] == '?')
fs/xfs/scrub/symlink_repair.c
215
if (sc->ip->i_df.if_format == XFS_DINODE_FMT_LOCAL)
fs/xfs/scrub/symlink_repair.c
228
if (strlen(target_buf) != sc->ip->i_disk_size)
fs/xfs/scrub/symlink_repair.c
241
trace_xrep_symlink_salvage_target(sc->ip, target_buf,
fs/xfs/scrub/symlink_repair.c
250
struct xfs_inode *ip,
fs/xfs/scrub/symlink_repair.c
257
xfs_symlink_local_to_remote(tp, bp, ip, ifp, NULL);
fs/xfs/scrub/symlink_repair.c
262
dsl->sl_owner = cpu_to_be64(sc->ip->i_ino);
fs/xfs/scrub/symlink_repair.c
294
xfs_trans_log_inode(sc->tp, sc->ip, 0);
fs/xfs/scrub/symlink_repair.c
309
ifp = xfs_ifork_ptr(sc->ip, XFS_DATA_FORK);
fs/xfs/scrub/symlink_repair.c
317
xfs_trans_log_inode(sc->tp, sc->ip,
fs/xfs/scrub/symlink_repair.c
333
ip_local = sc->ip->i_df.if_format == XFS_DINODE_FMT_LOCAL;
fs/xfs/scrub/symlink_repair.c
342
sc->tempip->i_disk_size <= xfs_inode_data_fork_size(sc->ip)) {
fs/xfs/scrub/symlink_repair.c
403
trace_xrep_symlink_rebuild(sc->ip);
fs/xfs/scrub/symlink_repair.c
437
error = xfs_symlink_write_target(sc->tp, sc->tempip, sc->ip->i_ino,
fs/xfs/scrub/symlink_repair.c
98
struct xfs_inode *ip = sc->ip;
fs/xfs/scrub/tempfile.c
212
if (!sc->ip || !xfs_is_metadir_inode(sc->ip))
fs/xfs/scrub/tempfile.c
375
xfs_lock_two_inodes(sc->ip, XFS_ILOCK_EXCL, sc->tempip, XFS_ILOCK_EXCL);
fs/xfs/scrub/tempfile.c
624
if (!xfs_ifork_ptr(sc->ip, whichfork) ||
fs/xfs/scrub/tempfile.c
626
ASSERT(xfs_ifork_ptr(sc->ip, whichfork) != NULL);
fs/xfs/scrub/tempfile.c
633
req->ip2 = sc->ip;
fs/xfs/scrub/tempfile.c
672
ifp = xfs_ifork_ptr(sc->ip, whichfork);
fs/xfs/scrub/tempfile.c
710
req->ip2_bcount = sc->ip->i_nblocks;
fs/xfs/scrub/tempfile.c
809
xfs_assert_ilocked(sc->ip, XFS_ILOCK_EXCL);
fs/xfs/scrub/tempfile.c
869
xfs_exchrange_ilock(sc->tp, sc->ip, sc->tempip);
fs/xfs/scrub/tempfile.c
900
temp = i_size_read(VFS_I(sc->ip));
fs/xfs/scrub/tempfile.c
901
i_size_write(VFS_I(sc->ip), i_size_read(VFS_I(sc->tempip)));
fs/xfs/scrub/tempfile.c
924
ifp = xfs_ifork_ptr(sc->ip, whichfork);
fs/xfs/scrub/tempfile.c
934
xfs_inode_data_fork_size(sc->ip));
fs/xfs/scrub/tempfile.c
937
ASSERT(sc->tempip->i_forkoff >= sc->ip->i_forkoff);
fs/xfs/scrub/tempfile.c
946
xfs_init_local_fork(sc->ip, whichfork, temp_ifp->if_data,
fs/xfs/scrub/tempfile.c
950
i_size_write(VFS_I(sc->ip), i_size_read(VFS_I(sc->tempip)));
fs/xfs/scrub/tempfile.c
951
sc->ip->i_disk_size = sc->tempip->i_disk_size;
fs/xfs/scrub/tempfile.c
955
xfs_trans_log_inode(sc->tp, sc->ip, ilog_flags);
fs/xfs/scrub/tempfile.c
961
const struct xfs_inode *ip)
fs/xfs/scrub/tempfile.c
963
const struct inode *inode = &ip->i_vnode;
fs/xfs/scrub/tempfile.c
964
struct xfs_mount *mp = ip->i_mount;
fs/xfs/scrub/tempfile.c
973
if (xfs_has_metadir(mp) && (ip->i_diflags2 & XFS_DIFLAG2_METADATA))
fs/xfs/scrub/tempfile.c
974
return __xfs_iflags_test(ip, XFS_IRECOVERY);
fs/xfs/scrub/tempfile.h
40
bool xrep_is_tempfile(const struct xfs_inode *ip);
fs/xfs/scrub/tempfile.h
46
# define xrep_is_tempfile(ip) (false)
fs/xfs/scrub/trace.c
51
return XFS_INO_TO_FSB(cur->bc_mp, cur->bc_ino.ip->i_ino);
fs/xfs/scrub/trace.h
150
TP_PROTO(struct xfs_inode *ip, struct xfs_scrub_metadata *sm,
fs/xfs/scrub/trace.h
152
TP_ARGS(ip, sm, error),
fs/xfs/scrub/trace.h
1591
TP_PROTO(struct xfs_mount *mp, struct xfs_inode *ip,
fs/xfs/scrub/trace.h
1593
TP_ARGS(mp, ip, live),
fs/xfs/scrub/trace.h
1605
__entry->ino = ip->i_ino;
fs/xfs/scrub/trace.h
1606
__entry->ftype = xfs_mode_to_ftype(VFS_I(ip)->i_mode);
fs/xfs/scrub/trace.h
1607
__entry->nlink = VFS_I(ip)->i_nlink;
fs/xfs/scrub/trace.h
1623
TP_PROTO(struct xfs_mount *mp, struct xfs_inode *ip, \
fs/xfs/scrub/trace.h
1625
TP_ARGS(mp, ip, live))
fs/xfs/scrub/trace.h
1629
TP_PROTO(struct xfs_inode *ip, const struct xfs_name *name,
fs/xfs/scrub/trace.h
1631
TP_ARGS(ip, name, far_ino),
fs/xfs/scrub/trace.h
164
__entry->dev = ip->i_mount->m_super->s_dev;
fs/xfs/scrub/trace.h
1640
__entry->dev = ip->i_mount->m_super->s_dev;
fs/xfs/scrub/trace.h
1641
__entry->ino = ip->i_ino;
fs/xfs/scrub/trace.h
165
__entry->ino = ip->i_ino;
fs/xfs/scrub/trace.h
1655
TP_PROTO(struct xfs_inode *ip, const struct xfs_name *name, \
fs/xfs/scrub/trace.h
1657
TP_ARGS(ip, name, far_ino))
fs/xfs/scrub/trace.h
1666
TP_PROTO(struct xfs_scrub *sc, struct xfs_inode *ip,
fs/xfs/scrub/trace.h
1669
TP_ARGS(sc, ip, path_nr, name, pptr),
fs/xfs/scrub/trace.h
1683
__entry->child_ino = ip->i_ino;
fs/xfs/scrub/trace.h
1684
__entry->child_gen = VFS_I(ip)->i_generation;
fs/xfs/scrub/trace.h
1702
TP_PROTO(struct xfs_scrub *sc, struct xfs_inode *ip, \
fs/xfs/scrub/trace.h
1705
TP_ARGS(sc, ip, path_nr, name, pptr))
fs/xfs/scrub/trace.h
1710
TP_PROTO(struct xfs_scrub *sc, struct xfs_inode *ip,
fs/xfs/scrub/trace.h
1714
TP_ARGS(sc, ip, path_nr, step_nr, name, pptr),
fs/xfs/scrub/trace.h
1730
__entry->child_ino = ip->i_ino;
fs/xfs/scrub/trace.h
1731
__entry->child_gen = VFS_I(ip)->i_generation;
fs/xfs/scrub/trace.h
1750
TP_PROTO(struct xfs_scrub *sc, struct xfs_inode *ip, \
fs/xfs/scrub/trace.h
1754
TP_ARGS(sc, ip, path_nr, step_nr, name, pptr))
fs/xfs/scrub/trace.h
1833
__entry->ino = dl->sc->ip->i_ino;
fs/xfs/scrub/trace.h
185
TP_PROTO(struct xfs_inode *ip, struct xfs_scrub_metadata *sm, \
fs/xfs/scrub/trace.h
1861
const struct xfs_inode *ip, const struct xfs_name *xname),
fs/xfs/scrub/trace.h
1862
TP_ARGS(sc, path_nr, step_nr, dp, ip, xname),
fs/xfs/scrub/trace.h
187
TP_ARGS(ip, sm, error))
fs/xfs/scrub/trace.h
1876
__entry->child_ino = ip->i_ino;
fs/xfs/scrub/trace.h
1893
int action, const struct xfs_inode *ip, int delta,
fs/xfs/scrub/trace.h
1895
TP_ARGS(sc, dp, action, ip, delta, xname),
fs/xfs/scrub/trace.h
1909
__entry->child_ino = ip->i_ino;
fs/xfs/scrub/trace.h
1936
__entry->scrub_ino = sc->ip ? sc->ip->i_ino : NULLFSINO;
fs/xfs/scrub/trace.h
2197
TP_PROTO(struct xfs_inode *ip, int whichfork,
fs/xfs/scrub/trace.h
2199
TP_ARGS(ip, whichfork, irec),
fs/xfs/scrub/trace.h
2210
__entry->dev = VFS_I(ip)->i_sb->s_dev;
fs/xfs/scrub/trace.h
2211
__entry->ino = ip->i_ino;
fs/xfs/scrub/trace.h
225
TP_PROTO(struct xfs_inode *ip, struct xfs_scrub_vec_head *vhead),
fs/xfs/scrub/trace.h
226
TP_ARGS(ip, vhead),
fs/xfs/scrub/trace.h
238
__entry->dev = ip->i_mount->m_super->s_dev;
fs/xfs/scrub/trace.h
239
__entry->ino = ip->i_ino;
fs/xfs/scrub/trace.h
2527
__entry->size = sc->ip->i_disk_size;
fs/xfs/scrub/trace.h
2528
__entry->nblocks = sc->ip->i_nblocks;
fs/xfs/scrub/trace.h
2529
__entry->flags = sc->ip->i_diflags;
fs/xfs/scrub/trace.h
2530
__entry->flags2 = sc->ip->i_diflags2;
fs/xfs/scrub/trace.h
2531
__entry->nextents = sc->ip->i_df.if_nextents;
fs/xfs/scrub/trace.h
2532
__entry->format = sc->ip->i_df.if_format;
fs/xfs/scrub/trace.h
2533
__entry->anextents = sc->ip->i_af.if_nextents;
fs/xfs/scrub/trace.h
2534
__entry->aformat = sc->ip->i_af.if_format;
fs/xfs/scrub/trace.h
259
TP_PROTO(struct xfs_inode *ip, struct xfs_scrub_vec_head *vhead), \
fs/xfs/scrub/trace.h
260
TP_ARGS(ip, vhead))
fs/xfs/scrub/trace.h
2649
TP_PROTO(struct xfs_inode *ip, xfs_fsblock_t startblock,
fs/xfs/scrub/trace.h
2651
TP_ARGS(ip, startblock, startoff, blockcount),
fs/xfs/scrub/trace.h
2660
__entry->dev = ip->i_mount->m_super->s_dev;
fs/xfs/scrub/trace.h
2661
__entry->ino = ip->i_ino;
fs/xfs/scrub/trace.h
2675
TP_PROTO(struct xfs_inode *ip, const struct xfs_bmbt_irec *irec,
fs/xfs/scrub/trace.h
2677
TP_ARGS(ip, irec, new_startblock, new_blockcount),
fs/xfs/scrub/trace.h
2689
__entry->dev = ip->i_mount->m_super->s_dev;
fs/xfs/scrub/trace.h
2690
__entry->ino = ip->i_ino;
fs/xfs/scrub/trace.h
2878
TP_PROTO(struct xfs_scrub *sc, struct xfs_inode *ip, int whichfork,
fs/xfs/scrub/trace.h
2880
TP_ARGS(sc, ip, whichfork, irec),
fs/xfs/scrub/trace.h
2893
__entry->ino = ip->i_ino;
fs/xfs/scrub/trace.h
2939
TP_PROTO(struct xfs_inode *ip, xfs_dablk_t dabno, uint16_t magic),
fs/xfs/scrub/trace.h
2940
TP_ARGS(ip, dabno, magic),
fs/xfs/scrub/trace.h
2948
__entry->dev = ip->i_mount->m_super->s_dev;
fs/xfs/scrub/trace.h
2949
__entry->ino = ip->i_ino;
fs/xfs/scrub/trace.h
2961
TP_PROTO(struct xfs_inode *ip, unsigned int flags, char *name,
fs/xfs/scrub/trace.h
2963
TP_ARGS(ip, flags, name, namelen, valuelen),
fs/xfs/scrub/trace.h
2973
__entry->dev = ip->i_mount->m_super->s_dev;
fs/xfs/scrub/trace.h
2974
__entry->ino = ip->i_ino;
fs/xfs/scrub/trace.h
2990
TP_PROTO(struct xfs_inode *ip, unsigned int flags, char *name, \
fs/xfs/scrub/trace.h
2992
TP_ARGS(ip, flags, name, namelen, valuelen))
fs/xfs/scrub/trace.h
2999
TP_PROTO(struct xfs_inode *ip, unsigned int flags, const void *name,
fs/xfs/scrub/trace.h
3001
TP_ARGS(ip, flags, name, namelen, value, valuelen),
fs/xfs/scrub/trace.h
3013
__entry->dev = ip->i_mount->m_super->s_dev;
fs/xfs/scrub/trace.h
3014
__entry->ino = ip->i_ino;
fs/xfs/scrub/trace.h
3030
TP_PROTO(struct xfs_inode *ip, unsigned int flags, const void *name, \
fs/xfs/scrub/trace.h
3032
TP_ARGS(ip, flags, name, namelen, value, valuelen))
fs/xfs/scrub/trace.h
3037
TP_PROTO(struct xfs_inode *ip, struct xfs_inode *arg_ip),
fs/xfs/scrub/trace.h
3038
TP_ARGS(ip, arg_ip),
fs/xfs/scrub/trace.h
3045
__entry->dev = ip->i_mount->m_super->s_dev;
fs/xfs/scrub/trace.h
3046
__entry->ino = ip->i_ino;
fs/xfs/scrub/trace.h
3056
TP_PROTO(struct xfs_inode *ip, struct xfs_inode *arg_ip), \
fs/xfs/scrub/trace.h
3057
TP_ARGS(ip, arg_ip))
fs/xfs/scrub/trace.h
3063
TP_PROTO(struct xfs_inode *ip, const struct xfs_inode *dp,
fs/xfs/scrub/trace.h
3065
TP_ARGS(ip, dp, name),
fs/xfs/scrub/trace.h
3075
__entry->dev = ip->i_mount->m_super->s_dev;
fs/xfs/scrub/trace.h
3076
__entry->ino = ip->i_ino;
fs/xfs/scrub/trace.h
3092
TP_PROTO(struct xfs_inode *ip, const struct xfs_inode *dp, \
fs/xfs/scrub/trace.h
3094
TP_ARGS(ip, dp, name))
fs/xfs/scrub/trace.h
3191
TP_PROTO(struct xfs_inode *dp, struct xfs_inode *ip, bool moved),
fs/xfs/scrub/trace.h
3192
TP_ARGS(dp, ip, moved),
fs/xfs/scrub/trace.h
3202
__entry->child_ino = ip->i_ino;
fs/xfs/scrub/trace.h
3213
TP_PROTO(struct xfs_inode *dp, struct xfs_inode *ip, bool moved), \
fs/xfs/scrub/trace.h
3214
TP_ARGS(dp, ip, moved))
fs/xfs/scrub/trace.h
3244
TP_PROTO(struct xfs_inode *ip, const struct xfs_name *name,
fs/xfs/scrub/trace.h
3246
TP_ARGS(ip, name, pptr),
fs/xfs/scrub/trace.h
3256
__entry->dev = ip->i_mount->m_super->s_dev;
fs/xfs/scrub/trace.h
3257
__entry->ino = ip->i_ino;
fs/xfs/scrub/trace.h
3273
TP_PROTO(struct xfs_inode *ip, const struct xfs_name *name, \
fs/xfs/scrub/trace.h
3275
TP_ARGS(ip, name, pptr))
fs/xfs/scrub/trace.h
3282
TP_PROTO(struct xfs_inode *ip, const struct xfs_inode *dp,
fs/xfs/scrub/trace.h
3284
TP_ARGS(ip, dp, name),
fs/xfs/scrub/trace.h
3294
__entry->dev = ip->i_mount->m_super->s_dev;
fs/xfs/scrub/trace.h
3295
__entry->ino = ip->i_ino;
fs/xfs/scrub/trace.h
3311
TP_PROTO(struct xfs_inode *ip, const struct xfs_inode *dp, \
fs/xfs/scrub/trace.h
3313
TP_ARGS(ip, dp, name))
fs/xfs/scrub/trace.h
3385
TP_PROTO(struct xfs_inode *ip, char *target, unsigned int targetlen),
fs/xfs/scrub/trace.h
3386
TP_ARGS(ip, target, targetlen),
fs/xfs/scrub/trace.h
3394
__entry->dev = ip->i_mount->m_super->s_dev;
fs/xfs/scrub/trace.h
3395
__entry->ino = ip->i_ino;
fs/xfs/scrub/trace.h
3408
TP_PROTO(struct xfs_inode *ip),
fs/xfs/scrub/trace.h
3409
TP_ARGS(ip),
fs/xfs/scrub/trace.h
3415
__entry->dev = ip->i_mount->m_super->s_dev;
fs/xfs/scrub/trace.h
3416
__entry->ino = ip->i_ino;
fs/xfs/scrub/trace.h
342
__entry->dev = sc->ip->i_mount->m_super->s_dev;
fs/xfs/scrub/trace.h
3425
TP_PROTO(struct xfs_inode *ip), \
fs/xfs/scrub/trace.h
3426
TP_ARGS(ip))
fs/xfs/scrub/trace.h
343
__entry->ino = sc->ip->i_ino;
fs/xfs/scrub/trace.h
3432
xfs_agino_t bucket_agino, struct xfs_inode *ip),
fs/xfs/scrub/trace.h
3433
TP_ARGS(pag, bucket, bucket_agino, ip),
fs/xfs/scrub/trace.h
3446
__entry->agino = XFS_INO_TO_AGINO(pag_mount(pag), ip->i_ino);
fs/xfs/scrub/trace.h
3449
__entry->prev_agino = ip->i_prev_unlinked;
fs/xfs/scrub/trace.h
3450
__entry->next_agino = ip->i_next_unlinked;
fs/xfs/scrub/trace.h
3463
TP_PROTO(struct xfs_inode *ip, xfs_agino_t prev_agino),
fs/xfs/scrub/trace.h
3464
TP_ARGS(ip, prev_agino),
fs/xfs/scrub/trace.h
3475
__entry->dev = ip->i_mount->m_super->s_dev;
fs/xfs/scrub/trace.h
3476
__entry->agno = XFS_INO_TO_AGNO(ip->i_mount, ip->i_ino);
fs/xfs/scrub/trace.h
3477
__entry->agino = XFS_INO_TO_AGINO(ip->i_mount, ip->i_ino);
fs/xfs/scrub/trace.h
3478
__entry->old_prev_agino = ip->i_prev_unlinked;
fs/xfs/scrub/trace.h
3480
__entry->next_agino = ip->i_next_unlinked;
fs/xfs/scrub/trace.h
3481
__entry->nlink = VFS_I(ip)->i_nlink;
fs/xfs/scrub/trace.h
3495
TP_PROTO(struct xfs_inode *ip),
fs/xfs/scrub/trace.h
3496
TP_ARGS(ip),
fs/xfs/scrub/trace.h
3505
__entry->dev = ip->i_mount->m_super->s_dev;
fs/xfs/scrub/trace.h
3506
__entry->agno = XFS_INO_TO_AGNO(ip->i_mount, ip->i_ino);
fs/xfs/scrub/trace.h
3507
__entry->agino = XFS_INO_TO_AGINO(ip->i_mount, ip->i_ino);
fs/xfs/scrub/trace.h
3508
__entry->nlink = VFS_I(ip)->i_nlink;
fs/xfs/scrub/trace.h
3509
__entry->next_agino = ip->i_next_unlinked;
fs/xfs/scrub/trace.h
3582
TP_PROTO(struct xfs_inode *ip, xfs_agino_t next_agino),
fs/xfs/scrub/trace.h
3583
TP_ARGS(ip, next_agino),
fs/xfs/scrub/trace.h
3592
__entry->dev = ip->i_mount->m_super->s_dev;
fs/xfs/scrub/trace.h
3593
__entry->agno = XFS_INO_TO_AGNO(ip->i_mount, ip->i_ino);
fs/xfs/scrub/trace.h
3594
__entry->agino = XFS_INO_TO_AGINO(ip->i_mount, ip->i_ino);
fs/xfs/scrub/trace.h
3595
__entry->next_agino = ip->i_next_unlinked;
fs/xfs/scrub/trace.h
3608
TP_PROTO(struct xfs_inode *ip, xfs_agino_t prev_agino),
fs/xfs/scrub/trace.h
3609
TP_ARGS(ip, prev_agino),
fs/xfs/scrub/trace.h
3618
__entry->dev = ip->i_mount->m_super->s_dev;
fs/xfs/scrub/trace.h
3619
__entry->agno = XFS_INO_TO_AGNO(ip->i_mount, ip->i_ino);
fs/xfs/scrub/trace.h
3620
__entry->agino = XFS_INO_TO_AGINO(ip->i_mount, ip->i_ino);
fs/xfs/scrub/trace.h
3621
__entry->prev_agino = ip->i_prev_unlinked;
fs/xfs/scrub/trace.h
440
__entry->dev = sc->ip->i_mount->m_super->s_dev;
fs/xfs/scrub/trace.h
441
__entry->ino = sc->ip->i_ino;
fs/xfs/scrub/trace.h
616
__entry->ino = cur->bc_ino.ip->i_ino;
fs/xfs/scrub/trace.h
696
__entry->ino = sc->ip->i_ino;
fs/xfs/scrub/trace.h
832
TP_PROTO(struct xfs_inode *ip),
fs/xfs/scrub/trace.h
833
TP_ARGS(ip),
fs/xfs/scrub/trace.h
841
__entry->dev = VFS_I(ip)->i_sb->s_dev;
fs/xfs/scrub/trace.h
842
__entry->ino = ip->i_ino;
fs/xfs/scrub/trace.h
843
__entry->iflags = ip->i_flags;
fs/xfs/scrub/trace.h
844
__entry->mode = VFS_I(ip)->i_mode;
fs/xfs/xfs_acl.c
131
struct xfs_inode *ip = XFS_I(inode);
fs/xfs/xfs_acl.c
132
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_acl.c
135
.dp = ip,
fs/xfs/xfs_acl.c
144
trace_xfs_get_acl(ip);
fs/xfs/xfs_acl.c
177
struct xfs_inode *ip = XFS_I(inode);
fs/xfs/xfs_acl.c
179
.dp = ip,
fs/xfs/xfs_acl.c
225
struct xfs_inode *ip = XFS_I(inode);
fs/xfs/xfs_acl.c
226
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_acl.c
234
xfs_ilock(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_acl.c
235
xfs_trans_ijoin(tp, ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_acl.c
238
xfs_trans_log_inode(tp, ip, XFS_ILOG_CORE);
fs/xfs/xfs_aops.c
109
struct xfs_inode *ip = XFS_I(ioend->io_inode);
fs/xfs/xfs_aops.c
110
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_aops.c
111
bool is_zoned = xfs_is_zoned_inode(ip);
fs/xfs/xfs_aops.c
143
xfs_reflink_cancel_cow_range(ip, offset, size, true);
fs/xfs/xfs_aops.c
144
xfs_bmap_punch_delalloc_range(ip, XFS_DATA_FORK, offset,
fs/xfs/xfs_aops.c
154
error = xfs_zoned_end_io(ip, offset, size, ioend->io_sector,
fs/xfs/xfs_aops.c
157
error = xfs_reflink_end_cow(ip, offset, size);
fs/xfs/xfs_aops.c
159
error = xfs_iomap_write_unwritten(ip, offset, size, false);
fs/xfs/xfs_aops.c
164
error = xfs_setfilesize(ip, offset, size);
fs/xfs/xfs_aops.c
190
struct xfs_inode *ip =
fs/xfs/xfs_aops.c
196
spin_lock_irqsave(&ip->i_ioend_lock, flags);
fs/xfs/xfs_aops.c
197
list_replace_init(&ip->i_ioend_list, &tmp);
fs/xfs/xfs_aops.c
198
spin_unlock_irqrestore(&ip->i_ioend_lock, flags);
fs/xfs/xfs_aops.c
219
struct xfs_inode *ip = XFS_I(ioend->io_inode);
fs/xfs/xfs_aops.c
220
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_aops.c
232
spin_lock_irqsave(&ip->i_ioend_lock, flags);
fs/xfs/xfs_aops.c
233
if (list_empty(&ip->i_ioend_list))
fs/xfs/xfs_aops.c
235
&ip->i_ioend_work));
fs/xfs/xfs_aops.c
236
list_add_tail(&ioend->io_list, &ip->i_ioend_list);
fs/xfs/xfs_aops.c
237
spin_unlock_irqrestore(&ip->i_ioend_lock, flags);
fs/xfs/xfs_aops.c
262
struct xfs_inode *ip = XFS_I(folio->mapping->host);
fs/xfs/xfs_aops.c
263
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_aops.c
270
folio, ip->i_ino, pos);
fs/xfs/xfs_aops.c
277
xfs_bmap_punch_delalloc_range(ip, XFS_DATA_FORK, pos,
fs/xfs/xfs_aops.c
288
struct xfs_inode *ip,
fs/xfs/xfs_aops.c
309
if (XFS_WPC(wpc)->data_seq != READ_ONCE(ip->i_df.if_seq)) {
fs/xfs/xfs_aops.c
310
trace_xfs_wb_data_iomap_invalid(ip, &wpc->iomap,
fs/xfs/xfs_aops.c
314
if (xfs_inode_has_cow_data(ip) &&
fs/xfs/xfs_aops.c
315
XFS_WPC(wpc)->cow_seq != READ_ONCE(ip->i_cowfp->if_seq)) {
fs/xfs/xfs_aops.c
316
trace_xfs_wb_cow_iomap_invalid(ip, &wpc->iomap,
fs/xfs/xfs_aops.c
329
struct xfs_inode *ip = XFS_I(wpc->inode);
fs/xfs/xfs_aops.c
330
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_aops.c
362
if (xfs_imap_valid(wpc, ip, offset))
fs/xfs/xfs_aops.c
374
xfs_ilock(ip, XFS_ILOCK_SHARED);
fs/xfs/xfs_aops.c
375
ASSERT(!xfs_need_iread_extents(&ip->i_df));
fs/xfs/xfs_aops.c
381
if (xfs_inode_has_cow_data(ip) &&
fs/xfs/xfs_aops.c
382
xfs_iext_lookup_extent(ip, ip->i_cowfp, offset_fsb, &icur, &imap))
fs/xfs/xfs_aops.c
385
XFS_WPC(wpc)->cow_seq = READ_ONCE(ip->i_cowfp->if_seq);
fs/xfs/xfs_aops.c
386
xfs_iunlock(ip, XFS_ILOCK_SHARED);
fs/xfs/xfs_aops.c
396
if (xfs_imap_valid(wpc, ip, offset)) {
fs/xfs/xfs_aops.c
397
xfs_iunlock(ip, XFS_ILOCK_SHARED);
fs/xfs/xfs_aops.c
406
if (!xfs_iext_lookup_extent(ip, &ip->i_df, offset_fsb, &icur, &imap))
fs/xfs/xfs_aops.c
408
XFS_WPC(wpc)->data_seq = READ_ONCE(ip->i_df.if_seq);
fs/xfs/xfs_aops.c
409
xfs_iunlock(ip, XFS_ILOCK_SHARED);
fs/xfs/xfs_aops.c
434
xfs_bmbt_to_iomap(ip, &wpc->iomap, &imap, 0, 0, XFS_WPC(wpc)->data_seq);
fs/xfs/xfs_aops.c
435
trace_xfs_map_blocks_found(ip, offset, count, whichfork, &imap);
fs/xfs/xfs_aops.c
449
error = xfs_bmapi_convert_delalloc(ip, whichfork, offset,
fs/xfs/xfs_aops.c
479
trace_xfs_map_blocks_alloc(ip, offset, count, whichfork, &imap);
fs/xfs/xfs_aops.c
52
struct xfs_inode *ip,
fs/xfs/xfs_aops.c
56
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_aops.c
574
struct xfs_inode *ip = XFS_I(wpc->inode);
fs/xfs/xfs_aops.c
575
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_aops.c
601
xfs_ilock(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_aops.c
602
if (!xfs_iext_lookup_extent(ip, ip->i_cowfp, offset_fsb, &icur, &imap))
fs/xfs/xfs_aops.c
609
xfs_iunlock(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_aops.c
610
xfs_bmbt_to_iomap(ip, &wpc->iomap, &imap, 0, 0, 0);
fs/xfs/xfs_aops.c
618
xfs_bmap_del_extent_delay(ip, XFS_COW_FORK, &icur, &imap, &del,
fs/xfs/xfs_aops.c
620
xfs_iunlock(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_aops.c
629
trace_xfs_zoned_map_blocks(ip, offset, wpc->iomap.length);
fs/xfs/xfs_aops.c
65
xfs_ilock(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_aops.c
66
isize = xfs_new_eof(ip, offset + size);
fs/xfs/xfs_aops.c
678
struct xfs_inode *ip = XFS_I(mapping->host);
fs/xfs/xfs_aops.c
68
xfs_iunlock(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_aops.c
680
xfs_iflags_clear(ip, XFS_ITRUNCATED);
fs/xfs/xfs_aops.c
682
if (xfs_is_zoned_inode(ip)) {
fs/xfs/xfs_aops.c
714
struct xfs_inode *ip = XFS_I(mapping->host);
fs/xfs/xfs_aops.c
716
xfs_iflags_clear(ip, XFS_ITRUNCATED);
fs/xfs/xfs_aops.c
718
xfs_inode_buftarg(ip)->bt_daxdev, wbc);
fs/xfs/xfs_aops.c
726
struct xfs_inode *ip = XFS_I(mapping->host);
fs/xfs/xfs_aops.c
728
trace_xfs_vm_bmap(ip);
fs/xfs/xfs_aops.c
73
trace_xfs_setfilesize(ip, offset, size);
fs/xfs/xfs_aops.c
739
if (xfs_is_cow_inode(ip) || XFS_IS_REALTIME_INODE(ip))
fs/xfs/xfs_aops.c
75
ip->i_disk_size = isize;
fs/xfs/xfs_aops.c
76
xfs_trans_ijoin(tp, ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_aops.c
766
struct xfs_inode *ip = XFS_I(file_inode(swap_file));
fs/xfs/xfs_aops.c
768
if (xfs_is_zoned_inode(ip))
fs/xfs/xfs_aops.c
77
xfs_trans_log_inode(tp, ip, XFS_ILOG_CORE);
fs/xfs/xfs_aops.c
794
xfs_inodegc_flush(ip->i_mount);
fs/xfs/xfs_aops.c
800
sis->bdev = xfs_inode_buftarg(ip)->bt_bdev;
fs/xfs/xfs_aops.h
12
int xfs_setfilesize(struct xfs_inode *ip, xfs_off_t offset, size_t size);
fs/xfs/xfs_attr_item.c
612
struct xfs_inode *ip)
fs/xfs/xfs_attr_item.c
617
tp = xfs_trans_alloc_empty(ip->i_mount);
fs/xfs/xfs_attr_item.c
618
xfs_ilock(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_attr_item.c
619
error = xfs_iread_extents(tp, ip, XFS_ATTR_FORK);
fs/xfs/xfs_attr_item.c
620
xfs_iunlock(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_attr_item.c
636
struct xfs_inode *ip;
fs/xfs/xfs_attr_item.c
649
attrp->alfi_igen, &ip);
fs/xfs/xfs_attr_item.c
652
error = xlog_recover_iget(mp, attrp->alfi_ino, &ip);
fs/xfs/xfs_attr_item.c
661
if (xfs_inode_has_attr_fork(ip)) {
fs/xfs/xfs_attr_item.c
662
error = xfs_attri_iread_extents(ip);
fs/xfs/xfs_attr_item.c
664
xfs_irele(ip);
fs/xfs/xfs_attr_item.c
684
args->dp = ip;
fs/xfs/xfs_attr_item.c
719
*ipp = ip;
fs/xfs/xfs_attr_item.c
736
struct xfs_inode *ip = NULL;
fs/xfs/xfs_attr_item.c
755
attr = xfs_attri_recover_work(mp, dfp, attrp, &ip, nv);
fs/xfs/xfs_attr_item.c
780
xfs_ilock(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_attr_item.c
781
xfs_trans_ijoin(tp, ip, 0);
fs/xfs/xfs_attr_item.c
793
xfs_iunlock(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_attr_item.c
794
xfs_irele(ip);
fs/xfs/xfs_bmap_item.c
501
struct xfs_inode *ip = NULL;
fs/xfs/xfs_bmap_item.c
515
work = xfs_bui_recover_work(mp, dfp, &ip, map);
fs/xfs/xfs_bmap_item.c
526
xfs_ilock(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_bmap_item.c
527
xfs_trans_ijoin(tp, ip, 0);
fs/xfs/xfs_bmap_item.c
530
xfs_ifork_is_realtime(ip, work->bi_whichfork)) {
fs/xfs/xfs_bmap_item.c
540
error = xfs_iext_count_extend(tp, ip, work->bi_whichfork, iext_delta);
fs/xfs/xfs_bmap_item.c
559
xfs_iunlock(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_bmap_item.c
560
xfs_irele(ip);
fs/xfs/xfs_bmap_item.c
566
xfs_iunlock(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_bmap_item.c
568
xfs_irele(ip);
fs/xfs/xfs_bmap_util.c
1000
trace_xfs_collapse_file_space(ip);
fs/xfs/xfs_bmap_util.c
1002
error = xfs_free_file_space(ip, offset, len, ac);
fs/xfs/xfs_bmap_util.c
1006
error = xfs_prepare_shift(ip, offset);
fs/xfs/xfs_bmap_util.c
1014
xfs_ilock(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_bmap_util.c
1015
xfs_trans_ijoin(tp, ip, 0);
fs/xfs/xfs_bmap_util.c
1018
error = xfs_bmap_collapse_extents(tp, ip, &next_fsb, shift_fsb,
fs/xfs/xfs_bmap_util.c
1032
xfs_iunlock(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_bmap_util.c
1037
xfs_iunlock(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_bmap_util.c
104
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_bmap_util.c
105
struct xfs_ifork *ifp = xfs_ifork_ptr(ip, whichfork);
fs/xfs/xfs_bmap_util.c
1055
struct xfs_inode *ip,
fs/xfs/xfs_bmap_util.c
1059
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_bmap_util.c
1067
xfs_assert_ilocked(ip, XFS_IOLOCK_EXCL | XFS_MMAPLOCK_EXCL);
fs/xfs/xfs_bmap_util.c
1069
trace_xfs_insert_file_space(ip);
fs/xfs/xfs_bmap_util.c
1071
error = xfs_bmap_can_insert_extents(ip, stop_fsb, shift_fsb);
fs/xfs/xfs_bmap_util.c
1075
error = xfs_prepare_shift(ip, offset);
fs/xfs/xfs_bmap_util.c
1084
xfs_ilock(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_bmap_util.c
1085
xfs_trans_ijoin(tp, ip, 0);
fs/xfs/xfs_bmap_util.c
1087
error = xfs_iext_count_extend(tp, ip, XFS_DATA_FORK,
fs/xfs/xfs_bmap_util.c
1097
error = xfs_bmap_split_extent(tp, ip, stop_fsb);
fs/xfs/xfs_bmap_util.c
1106
error = xfs_bmap_insert_extents(tp, ip, &next_fsb, shift_fsb,
fs/xfs/xfs_bmap_util.c
1113
xfs_iunlock(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_bmap_util.c
1118
xfs_iunlock(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_bmap_util.c
1144
struct xfs_inode *ip, /* target inode */
fs/xfs/xfs_bmap_util.c
1147
struct xfs_ifork *ifp = &ip->i_df;
fs/xfs/xfs_bmap_util.c
1151
if (XFS_IS_QUOTA_ON(ip->i_mount) &&
fs/xfs/xfs_bmap_util.c
1152
(!uid_eq(VFS_I(ip)->i_uid, VFS_I(tip)->i_uid) ||
fs/xfs/xfs_bmap_util.c
1153
!gid_eq(VFS_I(ip)->i_gid, VFS_I(tip)->i_gid) ||
fs/xfs/xfs_bmap_util.c
1154
ip->i_projid != tip->i_projid))
fs/xfs/xfs_bmap_util.c
1173
if (xfs_has_rmapbt(ip->i_mount))
fs/xfs/xfs_bmap_util.c
118
error = xfs_iread_extents(tp, ip, whichfork);
fs/xfs/xfs_bmap_util.c
1187
tifp->if_nextents > XFS_IFORK_MAXEXT(ip, XFS_DATA_FORK))
fs/xfs/xfs_bmap_util.c
1205
if (xfs_inode_has_attr_fork(ip) &&
fs/xfs/xfs_bmap_util.c
1206
xfs_bmap_bmdr_space(tifp->if_broot) > xfs_inode_fork_boff(ip))
fs/xfs/xfs_bmap_util.c
1208
if (tifp->if_nextents <= XFS_IFORK_MAXEXT(ip, XFS_DATA_FORK))
fs/xfs/xfs_bmap_util.c
1215
xfs_bmap_bmdr_space(ip->i_df.if_broot) > xfs_inode_fork_boff(tip))
fs/xfs/xfs_bmap_util.c
122
cur = xfs_bmbt_init_cursor(mp, tp, ip, whichfork);
fs/xfs/xfs_bmap_util.c
1226
struct xfs_inode *ip)
fs/xfs/xfs_bmap_util.c
1230
error = filemap_write_and_wait(VFS_I(ip)->i_mapping);
fs/xfs/xfs_bmap_util.c
1233
truncate_pagecache_range(VFS_I(ip), 0, -1);
fs/xfs/xfs_bmap_util.c
1236
if (VFS_I(ip)->i_mapping->nrpages)
fs/xfs/xfs_bmap_util.c
1247
struct xfs_inode *ip,
fs/xfs/xfs_bmap_util.c
1270
if (ip->i_diflags2 & XFS_DIFLAG2_REFLINK)
fs/xfs/xfs_bmap_util.c
1274
end_fsb = XFS_B_TO_FSB(ip->i_mount, i_size_read(VFS_I(ip)));
fs/xfs/xfs_bmap_util.c
1297
error = xfs_bmapi_read(ip, tirec.br_startoff,
fs/xfs/xfs_bmap_util.c
1304
trace_xfs_swap_extent_rmap_remap_piece(ip, &irec);
fs/xfs/xfs_bmap_util.c
1314
error = xfs_iext_count_extend(tp, ip,
fs/xfs/xfs_bmap_util.c
1333
xfs_bmap_unmap_extent(tp, ip, XFS_DATA_FORK, &irec);
fs/xfs/xfs_bmap_util.c
1336
xfs_bmap_map_extent(tp, ip, XFS_DATA_FORK, &uirec);
fs/xfs/xfs_bmap_util.c
1362
trace_xfs_swap_extent_rmap_error(ip, error, _RET_IP_);
fs/xfs/xfs_bmap_util.c
1371
struct xfs_inode *ip,
fs/xfs/xfs_bmap_util.c
1385
if (xfs_inode_has_attr_fork(ip) && ip->i_af.if_nextents > 0 &&
fs/xfs/xfs_bmap_util.c
1386
ip->i_af.if_format != XFS_DINODE_FMT_LOCAL) {
fs/xfs/xfs_bmap_util.c
1387
error = xfs_bmap_count_blocks(tp, ip, XFS_ATTR_FORK, &junk,
fs/xfs/xfs_bmap_util.c
1407
if (xfs_has_v3inodes(ip->i_mount)) {
fs/xfs/xfs_bmap_util.c
1408
if (ip->i_df.if_format == XFS_DINODE_FMT_BTREE)
fs/xfs/xfs_bmap_util.c
1417
swap(ip->i_df, tip->i_df);
fs/xfs/xfs_bmap_util.c
1422
tmp = (uint64_t)ip->i_nblocks;
fs/xfs/xfs_bmap_util.c
1423
ip->i_nblocks = tip->i_nblocks - taforkblks + aforkblks;
fs/xfs/xfs_bmap_util.c
1436
tip->i_delayed_blks = ip->i_delayed_blks;
fs/xfs/xfs_bmap_util.c
1437
ip->i_delayed_blks = 0;
fs/xfs/xfs_bmap_util.c
1439
switch (ip->i_df.if_format) {
fs/xfs/xfs_bmap_util.c
1444
ASSERT(!xfs_has_v3inodes(ip->i_mount) ||
fs/xfs/xfs_bmap_util.c
1456
ASSERT(!xfs_has_v3inodes(ip->i_mount) ||
fs/xfs/xfs_bmap_util.c
146
struct xfs_inode *ip,
fs/xfs/xfs_bmap_util.c
1475
struct xfs_inode *ip,
fs/xfs/xfs_bmap_util.c
1482
error = xfs_bmbt_change_owner(tp, ip, XFS_DATA_FORK, ip->i_ino,
fs/xfs/xfs_bmap_util.c
1497
xfs_trans_ijoin(tp, ip, 0);
fs/xfs/xfs_bmap_util.c
1499
xfs_trans_log_inode(tp, ip, XFS_ILOG_CORE);
fs/xfs/xfs_bmap_util.c
1508
struct xfs_inode *ip, /* target inode */
fs/xfs/xfs_bmap_util.c
1512
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_bmap_util.c
1528
lock_two_nondirectories(VFS_I(ip), VFS_I(tip));
fs/xfs/xfs_bmap_util.c
1529
filemap_invalidate_lock_two(VFS_I(ip)->i_mapping,
fs/xfs/xfs_bmap_util.c
1533
if ((VFS_I(ip)->i_mode & S_IFMT) != (VFS_I(tip)->i_mode & S_IFMT)) {
fs/xfs/xfs_bmap_util.c
1539
if (XFS_IS_REALTIME_INODE(ip) != XFS_IS_REALTIME_INODE(tip)) {
fs/xfs/xfs_bmap_util.c
1551
if (XFS_IS_REALTIME_INODE(ip) && xfs_has_rtgroups(ip->i_mount)) {
fs/xfs/xfs_bmap_util.c
1556
error = xfs_qm_dqattach(ip);
fs/xfs/xfs_bmap_util.c
156
error = xfs_reflink_trim_around_shared(ip, got, &shared);
fs/xfs/xfs_bmap_util.c
1564
error = xfs_swap_extent_flush(ip);
fs/xfs/xfs_bmap_util.c
1584
uint32_t ipnext = ip->i_df.if_nextents;
fs/xfs/xfs_bmap_util.c
1613
xfs_lock_two_inodes(ip, XFS_ILOCK_EXCL, tip, XFS_ILOCK_EXCL);
fs/xfs/xfs_bmap_util.c
1614
xfs_trans_ijoin(tp, ip, 0);
fs/xfs/xfs_bmap_util.c
1620
sxp->sx_length != ip->i_disk_size ||
fs/xfs/xfs_bmap_util.c
1626
trace_xfs_swap_extent_before(ip, 0);
fs/xfs/xfs_bmap_util.c
1630
error = xfs_swap_extents_check_format(ip, tip);
fs/xfs/xfs_bmap_util.c
1634
__func__, ip->i_ino);
fs/xfs/xfs_bmap_util.c
1645
ctime = inode_get_ctime(VFS_I(ip));
fs/xfs/xfs_bmap_util.c
1646
mtime = inode_get_mtime(VFS_I(ip));
fs/xfs/xfs_bmap_util.c
1666
error = xfs_swap_extent_rmap(&tp, ip, tip);
fs/xfs/xfs_bmap_util.c
1668
error = xfs_swap_extent_forks(tp, ip, tip, &src_log_flags,
fs/xfs/xfs_bmap_util.c
1674
if ((ip->i_diflags2 & XFS_DIFLAG2_REFLINK) ^
fs/xfs/xfs_bmap_util.c
1676
f = ip->i_diflags2 & XFS_DIFLAG2_REFLINK;
fs/xfs/xfs_bmap_util.c
1677
ip->i_diflags2 &= ~XFS_DIFLAG2_REFLINK;
fs/xfs/xfs_bmap_util.c
1678
ip->i_diflags2 |= tip->i_diflags2 & XFS_DIFLAG2_REFLINK;
fs/xfs/xfs_bmap_util.c
1685
ASSERT(!ip->i_cowfp ||
fs/xfs/xfs_bmap_util.c
1686
ip->i_cowfp->if_format == XFS_DINODE_FMT_EXTENTS);
fs/xfs/xfs_bmap_util.c
1690
swap(ip->i_cowfp, tip->i_cowfp);
fs/xfs/xfs_bmap_util.c
1692
if (ip->i_cowfp && ip->i_cowfp->if_bytes)
fs/xfs/xfs_bmap_util.c
1693
xfs_inode_set_cowblocks_tag(ip);
fs/xfs/xfs_bmap_util.c
1695
xfs_inode_clear_cowblocks_tag(ip);
fs/xfs/xfs_bmap_util.c
1702
xfs_trans_log_inode(tp, ip, src_log_flags);
fs/xfs/xfs_bmap_util.c
1712
error = xfs_swap_change_owner(&tp, ip, tip);
fs/xfs/xfs_bmap_util.c
1717
error = xfs_swap_change_owner(&tp, tip, ip);
fs/xfs/xfs_bmap_util.c
1731
trace_xfs_swap_extent_after(ip, 0);
fs/xfs/xfs_bmap_util.c
1735
xfs_iunlock(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_bmap_util.c
1738
filemap_invalidate_unlock_two(VFS_I(ip)->i_mapping,
fs/xfs/xfs_bmap_util.c
174
p->bmv_block = xfs_fsb_to_db(ip, got->br_startblock);
fs/xfs/xfs_bmap_util.c
1740
unlock_two_nondirectories(VFS_I(ip), VFS_I(tip));
fs/xfs/xfs_bmap_util.c
184
p->bmv_offset = XFS_FSB_TO_BB(ip->i_mount, got->br_startoff);
fs/xfs/xfs_bmap_util.c
185
p->bmv_length = XFS_FSB_TO_BB(ip->i_mount, got->br_blockcount);
fs/xfs/xfs_bmap_util.c
195
struct xfs_inode *ip,
fs/xfs/xfs_bmap_util.c
208
p->bmv_offset = XFS_FSB_TO_BB(ip->i_mount, bno);
fs/xfs/xfs_bmap_util.c
209
p->bmv_length = XFS_FSB_TO_BB(ip->i_mount, end - bno);
fs/xfs/xfs_bmap_util.c
250
struct xfs_inode *ip,
fs/xfs/xfs_bmap_util.c
254
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_bmap_util.c
287
xfs_ilock(ip, XFS_IOLOCK_SHARED);
fs/xfs/xfs_bmap_util.c
290
lock = xfs_ilock_attr_map_shared(ip);
fs/xfs/xfs_bmap_util.c
291
if (!xfs_inode_has_attr_fork(ip))
fs/xfs/xfs_bmap_util.c
298
xfs_ilock(ip, lock);
fs/xfs/xfs_bmap_util.c
301
if (!xfs_ifork_ptr(ip, whichfork))
fs/xfs/xfs_bmap_util.c
304
if (xfs_get_cowextsz_hint(ip))
fs/xfs/xfs_bmap_util.c
307
max_len = XFS_ISIZE(ip);
fs/xfs/xfs_bmap_util.c
311
(ip->i_delayed_blks || XFS_ISIZE(ip) > ip->i_disk_size)) {
fs/xfs/xfs_bmap_util.c
312
error = filemap_write_and_wait(VFS_I(ip)->i_mapping);
fs/xfs/xfs_bmap_util.c
326
if (xfs_get_extsz_hint(ip) ||
fs/xfs/xfs_bmap_util.c
327
(ip->i_diflags & XFS_DIFLAG_PREALLOC))
fs/xfs/xfs_bmap_util.c
330
max_len = XFS_ISIZE(ip);
fs/xfs/xfs_bmap_util.c
332
lock = xfs_ilock_data_map_shared(ip);
fs/xfs/xfs_bmap_util.c
336
ifp = xfs_ifork_ptr(ip, whichfork);
fs/xfs/xfs_bmap_util.c
360
error = xfs_iread_extents(NULL, ip, whichfork);
fs/xfs/xfs_bmap_util.c
364
if (!xfs_iext_lookup_extent(ip, ifp, bno, &icur, &got)) {
fs/xfs/xfs_bmap_util.c
370
xfs_getbmap_report_hole(ip, bmv, out, bmv_end, bno,
fs/xfs/xfs_bmap_util.c
371
XFS_B_TO_FSB(mp, XFS_ISIZE(ip)));
fs/xfs/xfs_bmap_util.c
383
xfs_getbmap_report_hole(ip, bmv, out, bmv_end, bno,
fs/xfs/xfs_bmap_util.c
397
error = xfs_getbmap_report_one(ip, bmv, out, bmv_end,
fs/xfs/xfs_bmap_util.c
404
xfs_fileoff_t end = XFS_B_TO_FSB(mp, XFS_ISIZE(ip));
fs/xfs/xfs_bmap_util.c
412
xfs_getbmap_report_hole(ip, bmv, out, bmv_end,
fs/xfs/xfs_bmap_util.c
423
xfs_iunlock(ip, lock);
fs/xfs/xfs_bmap_util.c
425
xfs_iunlock(ip, XFS_IOLOCK_SHARED);
fs/xfs/xfs_bmap_util.c
43
xfs_fsb_to_db(struct xfs_inode *ip, xfs_fsblock_t fsb)
fs/xfs/xfs_bmap_util.c
437
struct xfs_inode *ip,
fs/xfs/xfs_bmap_util.c
443
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_bmap_util.c
444
struct xfs_ifork *ifp = xfs_ifork_ptr(ip, whichfork);
fs/xfs/xfs_bmap_util.c
45
if (XFS_IS_REALTIME_INODE(ip))
fs/xfs/xfs_bmap_util.c
452
xfs_ilock(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_bmap_util.c
453
if (!xfs_iext_lookup_extent_before(ip, ifp, &end_fsb, &icur, &got))
fs/xfs/xfs_bmap_util.c
46
return xfs_rtb_to_daddr(ip->i_mount, fsb);
fs/xfs/xfs_bmap_util.c
47
return XFS_FSB_TO_DADDR(ip->i_mount, fsb);
fs/xfs/xfs_bmap_util.c
472
if (xfs_is_zoned_inode(ip) && ac) {
fs/xfs/xfs_bmap_util.c
479
xfs_bmap_del_extent_delay(ip, whichfork, &icur, &got,
fs/xfs/xfs_bmap_util.c
483
xfs_bmap_del_extent_delay(ip, whichfork, &icur, &got,
fs/xfs/xfs_bmap_util.c
492
xfs_inode_clear_cowblocks_tag(ip);
fs/xfs/xfs_bmap_util.c
495
xfs_iunlock(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_bmap_util.c
504
struct xfs_inode *ip)
fs/xfs/xfs_bmap_util.c
506
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_bmap_util.c
517
if (!(inode_state_read_once(VFS_I(ip)) & I_FREEING))
fs/xfs/xfs_bmap_util.c
518
xfs_assert_ilocked(ip, XFS_IOLOCK_EXCL);
fs/xfs/xfs_bmap_util.c
521
if (!S_ISREG(VFS_I(ip)->i_mode))
fs/xfs/xfs_bmap_util.c
528
if (VFS_I(ip)->i_size == 0 &&
fs/xfs/xfs_bmap_util.c
529
VFS_I(ip)->i_mapping->nrpages == 0 &&
fs/xfs/xfs_bmap_util.c
530
ip->i_delayed_blks == 0)
fs/xfs/xfs_bmap_util.c
534
if (xfs_need_iread_extents(&ip->i_df))
fs/xfs/xfs_bmap_util.c
541
if ((ip->i_diflags & XFS_DIFLAG_PREALLOC) && !ip->i_delayed_blks)
fs/xfs/xfs_bmap_util.c
549
end_fsb = XFS_B_TO_FSB(mp, (xfs_ufsize_t)XFS_ISIZE(ip));
fs/xfs/xfs_bmap_util.c
55
struct xfs_inode *ip,
fs/xfs/xfs_bmap_util.c
550
if (xfs_inode_has_bigrtalloc(ip))
fs/xfs/xfs_bmap_util.c
562
xfs_ilock(ip, XFS_ILOCK_SHARED);
fs/xfs/xfs_bmap_util.c
563
if (ip->i_delayed_blks ||
fs/xfs/xfs_bmap_util.c
564
xfs_iext_lookup_extent(ip, &ip->i_df, end_fsb, &icur, &imap))
fs/xfs/xfs_bmap_util.c
566
xfs_iunlock(ip, XFS_ILOCK_SHARED);
fs/xfs/xfs_bmap_util.c
577
struct xfs_inode *ip)
fs/xfs/xfs_bmap_util.c
580
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_bmap_util.c
584
error = xfs_qm_dqattach(ip);
fs/xfs/xfs_bmap_util.c
589
inode_dio_wait(VFS_I(ip));
fs/xfs/xfs_bmap_util.c
59
return blkdev_issue_zeroout(xfs_inode_buftarg(ip)->bt_bdev,
fs/xfs/xfs_bmap_util.c
597
if (ip->i_diflags & (XFS_DIFLAG_PREALLOC | XFS_DIFLAG_APPEND)) {
fs/xfs/xfs_bmap_util.c
598
if (ip->i_delayed_blks) {
fs/xfs/xfs_bmap_util.c
599
xfs_bmap_punch_delalloc_range(ip, XFS_DATA_FORK,
fs/xfs/xfs_bmap_util.c
60
xfs_fsb_to_db(ip, start_fsb),
fs/xfs/xfs_bmap_util.c
600
round_up(XFS_ISIZE(ip), mp->m_sb.sb_blocksize),
fs/xfs/xfs_bmap_util.c
603
xfs_inode_clear_eofblocks_tag(ip);
fs/xfs/xfs_bmap_util.c
61
XFS_FSB_TO_BB(ip->i_mount, count_fsb),
fs/xfs/xfs_bmap_util.c
613
xfs_ilock(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_bmap_util.c
614
xfs_trans_ijoin(tp, ip, 0);
fs/xfs/xfs_bmap_util.c
622
error = xfs_itruncate_extents_flags(&tp, ip, XFS_DATA_FORK,
fs/xfs/xfs_bmap_util.c
623
XFS_ISIZE(ip), XFS_BMAPI_NODISCARD);
fs/xfs/xfs_bmap_util.c
631
xfs_inode_clear_eofblocks_tag(ip);
fs/xfs/xfs_bmap_util.c
641
xfs_iunlock(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_bmap_util.c
647
struct xfs_inode *ip,
fs/xfs/xfs_bmap_util.c
651
xfs_mount_t *mp = ip->i_mount;
fs/xfs/xfs_bmap_util.c
662
if (xfs_is_always_cow_inode(ip))
fs/xfs/xfs_bmap_util.c
665
trace_xfs_alloc_file_space(ip);
fs/xfs/xfs_bmap_util.c
670
error = xfs_qm_dqattach(ip);
fs/xfs/xfs_bmap_util.c
677
rt = XFS_IS_REALTIME_INODE(ip);
fs/xfs/xfs_bmap_util.c
678
extsz = xfs_get_extsz_hint(ip);
fs/xfs/xfs_bmap_util.c
731
error = xfs_trans_alloc_inode(ip, &M_RES(mp)->tr_write,
fs/xfs/xfs_bmap_util.c
736
error = xfs_iext_count_extend(tp, ip, XFS_DATA_FORK,
fs/xfs/xfs_bmap_util.c
750
error = xfs_bmapi_write(tp, ip, startoffset_fsb,
fs/xfs/xfs_bmap_util.c
762
ip->i_diflags |= XFS_DIFLAG_PREALLOC;
fs/xfs/xfs_bmap_util.c
763
xfs_trans_log_inode(tp, ip, XFS_ILOG_CORE);
fs/xfs/xfs_bmap_util.c
766
xfs_iunlock(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_bmap_util.c
773
xfs_iunlock(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_bmap_util.c
779
struct xfs_inode *ip,
fs/xfs/xfs_bmap_util.c
784
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_bmap_util.c
789
error = xfs_trans_alloc_inode(ip, &M_RES(mp)->tr_write, resblks, 0,
fs/xfs/xfs_bmap_util.c
794
error = xfs_iext_count_extend(tp, ip, XFS_DATA_FORK,
fs/xfs/xfs_bmap_util.c
799
error = xfs_bunmapi(tp, ip, startoffset_fsb, len_fsb, 0, 2, done);
fs/xfs/xfs_bmap_util.c
805
xfs_iunlock(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_bmap_util.c
816
struct xfs_inode *ip,
fs/xfs/xfs_bmap_util.c
820
struct inode *inode = VFS_I(ip);
fs/xfs/xfs_bmap_util.c
829
rounding = max_t(xfs_off_t, xfs_inode_alloc_unitsize(ip), PAGE_SIZE);
fs/xfs/xfs_bmap_util.c
842
struct xfs_inode *ip,
fs/xfs/xfs_bmap_util.c
847
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_bmap_util.c
852
trace_xfs_free_file_space(ip);
fs/xfs/xfs_bmap_util.c
854
error = xfs_qm_dqattach(ip);
fs/xfs/xfs_bmap_util.c
865
error = xfs_flush_unmap_range(ip, offset, len);
fs/xfs/xfs_bmap_util.c
873
if (xfs_inode_has_bigrtalloc(ip)) {
fs/xfs/xfs_bmap_util.c
883
error = xfs_unmap_extent(ip, startoffset_fsb,
fs/xfs/xfs_bmap_util.c
896
if (offset >= XFS_ISIZE(ip))
fs/xfs/xfs_bmap_util.c
898
if (offset + len > XFS_ISIZE(ip))
fs/xfs/xfs_bmap_util.c
899
len = XFS_ISIZE(ip) - offset;
fs/xfs/xfs_bmap_util.c
900
error = xfs_zero_range(ip, offset, len, ac, NULL);
fs/xfs/xfs_bmap_util.c
910
if (offset + len >= XFS_ISIZE(ip) && offset_in_page(offset + len) > 0) {
fs/xfs/xfs_bmap_util.c
911
error = filemap_write_and_wait_range(VFS_I(ip)->i_mapping,
fs/xfs/xfs_bmap_util.c
920
struct xfs_inode *ip,
fs/xfs/xfs_bmap_util.c
930
if (xfs_can_free_eofblocks(ip)) {
fs/xfs/xfs_bmap_util.c
931
error = xfs_free_eofblocks(ip);
fs/xfs/xfs_bmap_util.c
944
rounding = xfs_inode_alloc_unitsize(ip);
fs/xfs/xfs_bmap_util.c
953
error = xfs_flush_unmap_range(ip, offset, XFS_ISIZE(ip));
fs/xfs/xfs_bmap_util.c
962
if (xfs_inode_has_cow_data(ip)) {
fs/xfs/xfs_bmap_util.c
963
error = xfs_reflink_cancel_cow_range(ip, offset, NULLFILEOFF,
fs/xfs/xfs_bmap_util.c
986
struct xfs_inode *ip,
fs/xfs/xfs_bmap_util.c
99
struct xfs_inode *ip,
fs/xfs/xfs_bmap_util.c
991
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_bmap_util.c
998
xfs_assert_ilocked(ip, XFS_IOLOCK_EXCL | XFS_MMAPLOCK_EXCL);
fs/xfs/xfs_bmap_util.h
34
void xfs_bmap_punch_delalloc_range(struct xfs_inode *ip, int whichfork,
fs/xfs/xfs_bmap_util.h
44
int xfs_getbmap(struct xfs_inode *ip, struct getbmapx *bmv,
fs/xfs/xfs_bmap_util.h
53
int xfs_bmap_last_extent(struct xfs_trans *tp, struct xfs_inode *ip,
fs/xfs/xfs_bmap_util.h
58
int xfs_alloc_file_space(struct xfs_inode *ip, xfs_off_t offset,
fs/xfs/xfs_bmap_util.h
60
int xfs_free_file_space(struct xfs_inode *ip, xfs_off_t offset,
fs/xfs/xfs_bmap_util.h
68
bool xfs_can_free_eofblocks(struct xfs_inode *ip);
fs/xfs/xfs_bmap_util.h
69
int xfs_free_eofblocks(struct xfs_inode *ip);
fs/xfs/xfs_bmap_util.h
71
int xfs_swap_extents(struct xfs_inode *ip, struct xfs_inode *tip,
fs/xfs/xfs_bmap_util.h
74
xfs_daddr_t xfs_fsb_to_db(struct xfs_inode *ip, xfs_fsblock_t fsb);
fs/xfs/xfs_bmap_util.h
77
int xfs_bmap_count_blocks(struct xfs_trans *tp, struct xfs_inode *ip,
fs/xfs/xfs_bmap_util.h
81
int xfs_flush_unmap_range(struct xfs_inode *ip, xfs_off_t offset,
fs/xfs/xfs_dquot.c
1004
xfs_assert_ilocked(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_dquot.c
1010
xfs_assert_ilocked(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_dquot.c
1011
ASSERT(xfs_inode_dquot(ip, type) == NULL);
fs/xfs/xfs_dquot.c
1012
ASSERT(!xfs_is_metadir_inode(ip));
fs/xfs/xfs_dquot.c
1014
id = xfs_qm_id_for_quotatype(ip, type);
fs/xfs/xfs_dquot.c
1028
xfs_iunlock(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_dquot.c
1030
xfs_ilock(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_dquot.c
1041
dqp1 = xfs_inode_dquot(ip, type);
fs/xfs/xfs_dquot.c
1068
xfs_assert_ilocked(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_dquot.c
971
struct xfs_inode *ip,
fs/xfs/xfs_dquot.c
976
return i_uid_read(VFS_I(ip));
fs/xfs/xfs_dquot.c
978
return i_gid_read(VFS_I(ip));
fs/xfs/xfs_dquot.c
980
return ip->i_projid;
fs/xfs/xfs_dquot.c
993
struct xfs_inode *ip,
fs/xfs/xfs_dquot.c
998
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_dquot.h
145
struct xfs_inode *ip,
fs/xfs/xfs_dquot.h
148
if (xfs_is_metadir_inode(ip))
fs/xfs/xfs_dquot.h
153
return ip->i_udquot;
fs/xfs/xfs_dquot.h
155
return ip->i_gdquot;
fs/xfs/xfs_dquot.h
157
return ip->i_pdquot;
fs/xfs/xfs_dquot.h
209
xfs_dqid_t xfs_qm_id_for_quotatype(struct xfs_inode *ip,
fs/xfs/xfs_dquot.h
214
int xfs_qm_dqget_inode(struct xfs_inode *ip, xfs_dqtype_t type,
fs/xfs/xfs_error.c
357
struct xfs_inode *ip,
fs/xfs/xfs_error.c
364
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_error.c
372
fa, ip->i_ino, name);
fs/xfs/xfs_error.h
20
void xfs_inode_verifier_error(struct xfs_inode *ip, int error, const char *name,
fs/xfs/xfs_export.c
112
xfs_inode_t *ip;
fs/xfs/xfs_export.c
126
error = xfs_iget(mp, NULL, ino, XFS_IGET_UNTRUSTED, 0, &ip);
fs/xfs/xfs_export.c
154
if (xfs_inode_unlinked_incomplete(ip)) {
fs/xfs/xfs_export.c
155
error = xfs_inode_reload_unlinked(ip);
fs/xfs/xfs_export.c
158
xfs_irele(ip);
fs/xfs/xfs_export.c
163
if (VFS_I(ip)->i_generation != generation || IS_PRIVATE(VFS_I(ip))) {
fs/xfs/xfs_export.c
164
xfs_irele(ip);
fs/xfs/xfs_export.c
168
return VFS_I(ip);
fs/xfs/xfs_file.c
1028
xfs_iunlock(ip, iolock);
fs/xfs/xfs_file.c
1029
xfs_blockgc_free_quota(ip, XFS_ICWALK_FLAG_SYNC);
fs/xfs/xfs_file.c
103
struct xfs_inode *ip,
fs/xfs/xfs_file.c
1036
xfs_flush_inodes(ip->i_mount);
fs/xfs/xfs_file.c
1038
xfs_iunlock(ip, iolock);
fs/xfs/xfs_file.c
1040
xfs_blockgc_free_space(ip->i_mount, &icw);
fs/xfs/xfs_file.c
1046
xfs_iunlock(ip, iolock);
fs/xfs/xfs_file.c
1049
XFS_STATS_ADD(ip->i_mount, xs_write_bytes, ret);
fs/xfs/xfs_file.c
1061
struct xfs_inode *ip = XFS_I(iocb->ki_filp->f_mapping->host);
fs/xfs/xfs_file.c
1062
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_file.c
107
struct xfs_inode_log_item *iip = ip->i_itemp;
fs/xfs/xfs_file.c
1108
xfs_iunlock(ip, iolock);
fs/xfs/xfs_file.c
1110
xfs_zoned_space_unreserve(ip->i_mount, &ac);
fs/xfs/xfs_file.c
1124
struct xfs_inode *ip = XFS_I(inode);
fs/xfs/xfs_file.c
1128
XFS_STATS_INC(ip->i_mount, xs_write_calls);
fs/xfs/xfs_file.c
1133
if (xfs_is_shutdown(ip->i_mount))
fs/xfs/xfs_file.c
1137
if (ocount < xfs_get_atomic_write_min(ip))
fs/xfs/xfs_file.c
1140
if (ocount > xfs_get_atomic_write_max(ip))
fs/xfs/xfs_file.c
1163
if (xfs_is_zoned_inode(ip))
fs/xfs/xfs_file.c
1171
struct xfs_inode *ip = XFS_I(file_inode(filp));
fs/xfs/xfs_file.c
1173
if (xfs_has_wsync(ip->i_mount))
fs/xfs/xfs_file.c
120
return xfs_log_force_seq(ip->i_mount, seq, XFS_LOG_SYNC,
fs/xfs/xfs_file.c
1298
struct xfs_inode *ip,
fs/xfs/xfs_file.c
1301
if (xfs_is_zoned_inode(ip)) {
fs/xfs/xfs_file.c
1308
return XFS_TEST_ERROR(ip->i_mount, XFS_ERRTAG_FORCE_ZERO_RANGE);
fs/xfs/xfs_file.c
131
struct xfs_inode *ip = XFS_I(file->f_mapping->host);
fs/xfs/xfs_file.c
132
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_file.c
1328
struct xfs_inode *ip = XFS_I(inode);
fs/xfs/xfs_file.c
1333
trace_xfs_zero_file_space(ip);
fs/xfs/xfs_file.c
1339
if (xfs_falloc_force_zero(ip, ac)) {
fs/xfs/xfs_file.c
1340
error = xfs_zero_range(ip, offset, len, ac, NULL);
fs/xfs/xfs_file.c
1342
error = xfs_free_file_space(ip, offset, len, ac);
fs/xfs/xfs_file.c
1349
error = xfs_alloc_file_space(ip, offset, len);
fs/xfs/xfs_file.c
136
trace_xfs_file_fsync(ip);
fs/xfs/xfs_file.c
1424
struct xfs_inode *ip = XFS_I(inode);
fs/xfs/xfs_file.c
1428
xfs_ilock(ip, iolock);
fs/xfs/xfs_file.c
1448
error = xfs_free_file_space(ip, offset, len, ac);
fs/xfs/xfs_file.c
145
xfs_iflags_clear(ip, XFS_ITRUNCATED);
fs/xfs/xfs_file.c
1471
error = xfs_log_force_inode(ip);
fs/xfs/xfs_file.c
1474
xfs_iunlock(ip, iolock);
fs/xfs/xfs_file.c
1486
struct xfs_inode *ip = XFS_I(file_inode(file));
fs/xfs/xfs_file.c
1487
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_file.c
153
if (XFS_IS_REALTIME_INODE(ip) && mp->m_rtdev_targp != mp->m_ddev_targp)
fs/xfs/xfs_file.c
1546
struct xfs_inode *ip = XFS_I(file_inode(file));
fs/xfs/xfs_file.c
1556
xfs_ilock(ip, lockflags);
fs/xfs/xfs_file.c
1560
xfs_iunlock(ip, lockflags);
fs/xfs/xfs_file.c
162
if (ip->i_itemp) {
fs/xfs/xfs_file.c
163
err2 = xfs_fsync_flush_log(ip, datasync, &log_flushed);
fs/xfs/xfs_file.c
1656
struct xfs_inode *ip = XFS_I(inode);
fs/xfs/xfs_file.c
1660
if (xfs_is_shutdown(ip->i_mount))
fs/xfs/xfs_file.c
1670
mode = xfs_ilock_data_map_shared(ip);
fs/xfs/xfs_file.c
1671
if (ip->i_df.if_nextents > 0)
fs/xfs/xfs_file.c
1672
error = xfs_dir3_data_readahead(ip, 0, 0);
fs/xfs/xfs_file.c
1673
xfs_iunlock(ip, mode);
fs/xfs/xfs_file.c
1686
struct xfs_inode *ip = XFS_I(inode);
fs/xfs/xfs_file.c
1687
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_file.c
1705
if (xfs_iflags_test_and_clear(ip, XFS_ITRUNCATED)) {
fs/xfs/xfs_file.c
1706
xfs_iflags_clear(ip, XFS_EOFBLOCKS_RELEASED);
fs/xfs/xfs_file.c
1707
if (ip->i_delayed_blks > 0)
fs/xfs/xfs_file.c
1737
(ip->i_diflags & XFS_DIFLAG_APPEND) ||
fs/xfs/xfs_file.c
1738
xfs_is_zoned_inode(ip))
fs/xfs/xfs_file.c
1747
if (!xfs_iflags_test(ip, XFS_EOFBLOCKS_RELEASED) &&
fs/xfs/xfs_file.c
1748
xfs_ilock_nowait(ip, XFS_IOLOCK_EXCL)) {
fs/xfs/xfs_file.c
1749
if (xfs_can_free_eofblocks(ip) &&
fs/xfs/xfs_file.c
175
if (!log_flushed && !XFS_IS_REALTIME_INODE(ip) &&
fs/xfs/xfs_file.c
1750
!xfs_iflags_test_and_set(ip, XFS_EOFBLOCKS_RELEASED))
fs/xfs/xfs_file.c
1751
xfs_free_eofblocks(ip);
fs/xfs/xfs_file.c
1752
xfs_iunlock(ip, XFS_IOLOCK_EXCL);
fs/xfs/xfs_file.c
1764
xfs_inode_t *ip = XFS_I(inode);
fs/xfs/xfs_file.c
1779
bufsize = (size_t)min_t(loff_t, XFS_READDIR_BUFSIZE, ip->i_disk_size);
fs/xfs/xfs_file.c
1781
return xfs_readdir(NULL, ip, ctx, bufsize);
fs/xfs/xfs_file.c
1838
struct xfs_inode *ip = XFS_I(file_inode(vmf->vma->vm_file));
fs/xfs/xfs_file.c
1841
trace_xfs_read_fault(ip, order);
fs/xfs/xfs_file.c
1843
xfs_ilock(ip, XFS_MMAPLOCK_SHARED);
fs/xfs/xfs_file.c
1845
xfs_iunlock(ip, XFS_MMAPLOCK_SHARED);
fs/xfs/xfs_file.c
1867
struct xfs_inode *ip = XFS_I(inode);
fs/xfs/xfs_file.c
1871
trace_xfs_write_fault(ip, order);
fs/xfs/xfs_file.c
1881
xfs_ilock(ip, XFS_MMAPLOCK_SHARED);
fs/xfs/xfs_file.c
1882
if (xfs_iflags_test(ip, XFS_IREMAPPING)) {
fs/xfs/xfs_file.c
1883
xfs_iunlock(ip, XFS_MMAPLOCK_SHARED);
fs/xfs/xfs_file.c
1884
xfs_ilock(ip, XFS_MMAPLOCK_EXCL);
fs/xfs/xfs_file.c
1893
xfs_iunlock(ip, lock_mode);
fs/xfs/xfs_file.c
190
struct xfs_inode *ip = XFS_I(file_inode(iocb->ki_filp));
fs/xfs/xfs_file.c
1904
struct xfs_inode *ip = XFS_I(file_inode(vmf->vma->vm_file));
fs/xfs/xfs_file.c
1916
error = xfs_zoned_space_reserve(ip->i_mount,
fs/xfs/xfs_file.c
1917
XFS_B_TO_FSB(ip->i_mount, len), 0, &ac);
fs/xfs/xfs_file.c
1921
xfs_zoned_space_unreserve(ip->i_mount, &ac);
fs/xfs/xfs_file.c
193
if (!xfs_ilock_nowait(ip, lock_mode))
fs/xfs/xfs_file.c
196
xfs_ilock(ip, lock_mode);
fs/xfs/xfs_file.c
208
struct xfs_inode *ip = XFS_I(file_inode(iocb->ki_filp));
fs/xfs/xfs_file.c
219
xfs_iflags_test(ip, XFS_IREMAPPING)) {
fs/xfs/xfs_file.c
220
xfs_iunlock(ip, *lock_mode);
fs/xfs/xfs_file.c
253
struct xfs_inode *ip = XFS_I(file_inode(iocb->ki_filp));
fs/xfs/xfs_file.c
274
xfs_iunlock(ip, XFS_IOLOCK_SHARED);
fs/xfs/xfs_file.c
284
struct xfs_inode *ip = XFS_I(iocb->ki_filp->f_mapping->host);
fs/xfs/xfs_file.c
296
xfs_iunlock(ip, XFS_IOLOCK_SHARED);
fs/xfs/xfs_file.c
307
struct xfs_inode *ip = XFS_I(file_inode(iocb->ki_filp));
fs/xfs/xfs_file.c
316
xfs_iunlock(ip, XFS_IOLOCK_SHARED);
fs/xfs/xfs_file.c
356
struct xfs_inode *ip = XFS_I(inode);
fs/xfs/xfs_file.c
357
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_file.c
365
trace_xfs_file_splice_read(ip, *ppos, len);
fs/xfs/xfs_file.c
367
xfs_ilock(ip, XFS_IOLOCK_SHARED);
fs/xfs/xfs_file.c
369
xfs_iunlock(ip, XFS_IOLOCK_SHARED);
fs/xfs/xfs_file.c
391
struct xfs_inode *ip = XFS_I(iocb->ki_filp->f_mapping->host);
fs/xfs/xfs_file.c
404
spin_lock(&ip->i_flags_lock);
fs/xfs/xfs_file.c
405
isize = i_size_read(VFS_I(ip));
fs/xfs/xfs_file.c
407
spin_unlock(&ip->i_flags_lock);
fs/xfs/xfs_file.c
410
spin_unlock(&ip->i_flags_lock);
fs/xfs/xfs_file.c
422
xfs_iunlock(ip, *iolock);
fs/xfs/xfs_file.c
424
xfs_ilock(ip, *iolock);
fs/xfs/xfs_file.c
435
inode_dio_wait(VFS_I(ip));
fs/xfs/xfs_file.c
440
trace_xfs_zero_eof(ip, isize, iocb->ki_pos - isize);
fs/xfs/xfs_file.c
442
xfs_ilock(ip, XFS_MMAPLOCK_EXCL);
fs/xfs/xfs_file.c
443
error = xfs_zero_range(ip, isize, iocb->ki_pos - isize, ac, NULL);
fs/xfs/xfs_file.c
444
xfs_iunlock(ip, XFS_MMAPLOCK_EXCL);
fs/xfs/xfs_file.c
49
struct xfs_inode *ip,
fs/xfs/xfs_file.c
53
unsigned int alloc_unit = xfs_inode_alloc_unitsize(ip);
fs/xfs/xfs_file.c
571
struct xfs_inode *ip = XFS_I(inode);
fs/xfs/xfs_file.c
575
ASSERT(!xfs_is_zoned_inode(ip) ||
fs/xfs/xfs_file.c
578
trace_xfs_end_io_direct_write(ip, offset, size);
fs/xfs/xfs_file.c
580
if (xfs_is_shutdown(ip->i_mount))
fs/xfs/xfs_file.c
592
XFS_STATS_ADD(ip->i_mount, xs_write_bytes, size);
fs/xfs/xfs_file.c
603
error = xfs_reflink_end_atomic_cow(ip, offset, size);
fs/xfs/xfs_file.c
605
error = xfs_reflink_end_cow(ip, offset, size);
fs/xfs/xfs_file.c
617
error = xfs_iomap_write_unwritten(ip, offset, size, true);
fs/xfs/xfs_file.c
642
spin_lock(&ip->i_flags_lock);
fs/xfs/xfs_file.c
645
spin_unlock(&ip->i_flags_lock);
fs/xfs/xfs_file.c
646
error = xfs_setfilesize(ip, offset, size);
fs/xfs/xfs_file.c
648
spin_unlock(&ip->i_flags_lock);
fs/xfs/xfs_file.c
699
struct xfs_inode *ip,
fs/xfs/xfs_file.c
715
if (xfs_is_always_cow_inode(ip))
fs/xfs/xfs_file.c
731
xfs_ilock_demote(ip, XFS_IOLOCK_EXCL);
fs/xfs/xfs_file.c
739
xfs_iunlock(ip, iolock);
fs/xfs/xfs_file.c
748
struct xfs_inode *ip,
fs/xfs/xfs_file.c
75
struct xfs_inode *ip = XFS_I(file->f_mapping->host);
fs/xfs/xfs_file.c
755
ret = xfs_zoned_write_space_reserve(ip->i_mount, iocb, from, 0, &ac);
fs/xfs/xfs_file.c
758
ret = xfs_file_dio_write_aligned(ip, iocb, from,
fs/xfs/xfs_file.c
761
xfs_zoned_space_unreserve(ip->i_mount, &ac);
fs/xfs/xfs_file.c
77
trace_xfs_dir_fsync(ip);
fs/xfs/xfs_file.c
777
struct xfs_inode *ip,
fs/xfs/xfs_file.c
78
return xfs_log_force_inode(ip);
fs/xfs/xfs_file.c
790
if (ocount > xfs_inode_buftarg(ip)->bt_awu_max)
fs/xfs/xfs_file.c
806
xfs_ilock_demote(ip, XFS_IOLOCK_EXCL);
fs/xfs/xfs_file.c
823
xfs_iunlock(ip, iolock);
fs/xfs/xfs_file.c
830
xfs_iunlock(ip, iolock);
fs/xfs/xfs_file.c
853
struct xfs_inode *ip,
fs/xfs/xfs_file.c
857
size_t isize = i_size_read(VFS_I(ip));
fs/xfs/xfs_file.c
884
if (xfs_is_cow_inode(ip)) {
fs/xfs/xfs_file.c
901
inode_dio_wait(VFS_I(ip));
fs/xfs/xfs_file.c
917
xfs_iunlock(ip, iolock);
fs/xfs/xfs_file.c
923
xfs_iunlock(ip, iolock);
fs/xfs/xfs_file.c
932
struct xfs_inode *ip = XFS_I(file_inode(iocb->ki_filp));
fs/xfs/xfs_file.c
933
struct xfs_buftarg *target = xfs_inode_buftarg(ip);
fs/xfs/xfs_file.c
940
if ((iocb->ki_pos | count) & ip->i_mount->m_blockmask)
fs/xfs/xfs_file.c
941
return xfs_file_dio_write_unaligned(ip, iocb, from);
fs/xfs/xfs_file.c
942
if (xfs_is_zoned_inode(ip))
fs/xfs/xfs_file.c
943
return xfs_file_dio_write_zoned(ip, iocb, from);
fs/xfs/xfs_file.c
945
return xfs_file_dio_write_atomic(ip, iocb, from);
fs/xfs/xfs_file.c
946
return xfs_file_dio_write_aligned(ip, iocb, from,
fs/xfs/xfs_file.c
956
struct xfs_inode *ip = XFS_I(inode);
fs/xfs/xfs_file.c
974
error = xfs_setfilesize(ip, pos, ret);
fs/xfs/xfs_file.c
978
xfs_iunlock(ip, iolock);
fs/xfs/xfs_file.c
983
XFS_STATS_ADD(ip->i_mount, xs_write_bytes, ret);
fs/xfs/xfs_file.c
997
struct xfs_inode *ip = XFS_I(inode);
fs/xfs/xfs_file.h
12
bool xfs_is_falloc_aligned(struct xfs_inode *ip, loff_t pos,
fs/xfs/xfs_filestream.c
171
struct xfs_inode *ip)
fs/xfs/xfs_filestream.c
173
struct inode *inode = VFS_I(ip), *dir = NULL;
fs/xfs/xfs_filestream.c
228
trace_xfs_filestream_lookup(pag, ap->ip->i_ino);
fs/xfs/xfs_filestream.c
346
pip = xfs_filestream_get_parent(ap->ip);
fs/xfs/xfs_filestream.c
371
struct xfs_inode *ip)
fs/xfs/xfs_filestream.c
373
xfs_mru_cache_delete(ip->i_mount->m_filestream, ip->i_ino);
fs/xfs/xfs_filestream.h
16
void xfs_filestream_deassociate(struct xfs_inode *ip);
fs/xfs/xfs_filestream.h
22
struct xfs_inode *ip)
fs/xfs/xfs_filestream.h
24
return xfs_has_filestreams(ip->i_mount) ||
fs/xfs/xfs_filestream.h
25
(ip->i_diflags & XFS_DIFLAG_FILESTREAM);
fs/xfs/xfs_fsmap.c
1302
struct xfs_inode *ip,
fs/xfs/xfs_fsmap.c
1342
trace_xfs_getfsmap_low_key(ip->i_mount, &xhead.fmh_keys[0]);
fs/xfs/xfs_fsmap.c
1343
trace_xfs_getfsmap_high_key(ip->i_mount, &xhead.fmh_keys[1]);
fs/xfs/xfs_fsmap.c
1356
error = xfs_getfsmap(ip->i_mount, &xhead, recs);
fs/xfs/xfs_fsmap.c
1399
trace_xfs_getfsmap_low_key(ip->i_mount, &xhead.fmh_keys[0]);
fs/xfs/xfs_fsmap.h
46
int xfs_ioc_getfsmap(struct xfs_inode *ip, struct fsmap_head __user *arg);
fs/xfs/xfs_handle.c
105
ip = XFS_I(inode);
fs/xfs/xfs_handle.c
122
memcpy(&handle.ha_fsid, ip->i_mount->m_fixedfsid, sizeof(xfs_fsid_t));
fs/xfs/xfs_handle.c
125
hsize = xfs_fshandle_init(ip->i_mount, &handle);
fs/xfs/xfs_handle.c
127
hsize = xfs_filehandle_init(ip->i_mount, ip->i_ino,
fs/xfs/xfs_handle.c
185
struct xfs_inode *ip = XFS_I(file_inode(file));
fs/xfs/xfs_handle.c
186
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_handle.c
189
if (!S_ISDIR(VFS_I(ip)->i_mode))
fs/xfs/xfs_handle.c
661
struct xfs_inode *ip;
fs/xfs/xfs_handle.c
691
struct xfs_inode *ip = context->dp;
fs/xfs/xfs_handle.c
692
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_handle.c
707
xfs_inode_mark_sick(ip, XFS_SICK_INO_PARENT);
fs/xfs/xfs_handle.c
728
trace_xfs_getparents_put_listent(ip, gp, context, gpr);
fs/xfs/xfs_handle.c
748
trace_xfs_getparents_expand_lastrec(gpx->ip, gp, &gpx->context, gpr);
fs/xfs/xfs_handle.c
757
struct xfs_inode *ip = gpx->ip;
fs/xfs/xfs_handle.c
758
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_handle.c
782
gpx->context.dp = ip;
fs/xfs/xfs_handle.c
795
trace_xfs_getparents_begin(ip, gp, &gpx->context.cursor);
fs/xfs/xfs_handle.c
811
if (ip->i_ino == mp->m_sb.sb_rootino)
fs/xfs/xfs_handle.c
830
trace_xfs_getparents_end(ip, gp, &gpx->context.cursor);
fs/xfs/xfs_handle.c
852
.ip = XFS_I(file_inode(file)),
fs/xfs/xfs_handle.c
855
struct xfs_mount *mp = gpx.ip->i_mount;
fs/xfs/xfs_handle.c
882
struct xfs_inode *ip = XFS_I(file_inode(file));
fs/xfs/xfs_handle.c
883
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_handle.c
90
struct xfs_inode *ip;
fs/xfs/xfs_handle.c
908
gpx.ip = xfs_khandle_to_inode(file, handle);
fs/xfs/xfs_handle.c
909
if (IS_ERR(gpx.ip))
fs/xfs/xfs_handle.c
910
return PTR_ERR(gpx.ip);
fs/xfs/xfs_handle.c
920
xfs_irele(gpx.ip);
fs/xfs/xfs_health.c
317
static inline void xfs_inode_report_fserror(struct xfs_inode *ip)
fs/xfs/xfs_health.c
323
if (xfs_iflags_test(ip, XFS_INEW | XFS_IRECLAIM) ||
fs/xfs/xfs_health.c
324
xfs_is_internal_inode(ip)) {
fs/xfs/xfs_health.c
325
fserror_report_metadata(ip->i_mount->m_super, -EFSCORRUPTED,
fs/xfs/xfs_health.c
330
fserror_report_file_metadata(VFS_I(ip), -EFSCORRUPTED, GFP_NOFS);
fs/xfs/xfs_health.c
336
struct xfs_inode *ip,
fs/xfs/xfs_health.c
342
trace_xfs_inode_mark_sick(ip, mask);
fs/xfs/xfs_health.c
344
spin_lock(&ip->i_flags_lock);
fs/xfs/xfs_health.c
345
old_mask = ip->i_sick;
fs/xfs/xfs_health.c
346
ip->i_sick |= mask;
fs/xfs/xfs_health.c
347
spin_unlock(&ip->i_flags_lock);
fs/xfs/xfs_health.c
354
spin_lock(&VFS_I(ip)->i_lock);
fs/xfs/xfs_health.c
355
inode_state_clear(VFS_I(ip), I_DONTCACHE);
fs/xfs/xfs_health.c
356
spin_unlock(&VFS_I(ip)->i_lock);
fs/xfs/xfs_health.c
358
xfs_inode_report_fserror(ip);
fs/xfs/xfs_health.c
360
xfs_healthmon_report_inode(ip, XFS_HEALTHMON_SICK, old_mask,
fs/xfs/xfs_health.c
367
struct xfs_inode *ip,
fs/xfs/xfs_health.c
373
trace_xfs_inode_mark_corrupt(ip, mask);
fs/xfs/xfs_health.c
375
spin_lock(&ip->i_flags_lock);
fs/xfs/xfs_health.c
376
old_mask = ip->i_sick;
fs/xfs/xfs_health.c
377
ip->i_sick |= mask;
fs/xfs/xfs_health.c
378
ip->i_checked |= mask;
fs/xfs/xfs_health.c
379
spin_unlock(&ip->i_flags_lock);
fs/xfs/xfs_health.c
386
spin_lock(&VFS_I(ip)->i_lock);
fs/xfs/xfs_health.c
387
inode_state_clear(VFS_I(ip), I_DONTCACHE);
fs/xfs/xfs_health.c
388
spin_unlock(&VFS_I(ip)->i_lock);
fs/xfs/xfs_health.c
390
xfs_inode_report_fserror(ip);
fs/xfs/xfs_health.c
392
xfs_healthmon_report_inode(ip, XFS_HEALTHMON_CORRUPT, old_mask,
fs/xfs/xfs_health.c
399
struct xfs_inode *ip,
fs/xfs/xfs_health.c
405
trace_xfs_inode_mark_healthy(ip, mask);
fs/xfs/xfs_health.c
407
spin_lock(&ip->i_flags_lock);
fs/xfs/xfs_health.c
408
old_mask = ip->i_sick;
fs/xfs/xfs_health.c
409
ip->i_sick &= ~mask;
fs/xfs/xfs_health.c
410
if (!(ip->i_sick & XFS_SICK_INO_PRIMARY))
fs/xfs/xfs_health.c
411
ip->i_sick &= ~XFS_SICK_INO_SECONDARY;
fs/xfs/xfs_health.c
412
ip->i_checked |= mask;
fs/xfs/xfs_health.c
413
spin_unlock(&ip->i_flags_lock);
fs/xfs/xfs_health.c
416
xfs_healthmon_report_inode(ip, XFS_HEALTHMON_HEALTHY, old_mask,
fs/xfs/xfs_health.c
423
struct xfs_inode *ip,
fs/xfs/xfs_health.c
427
spin_lock(&ip->i_flags_lock);
fs/xfs/xfs_health.c
428
*sick = ip->i_sick;
fs/xfs/xfs_health.c
429
*checked = ip->i_checked;
fs/xfs/xfs_health.c
430
spin_unlock(&ip->i_flags_lock);
fs/xfs/xfs_health.c
633
struct xfs_inode *ip,
fs/xfs/xfs_health.c
643
xfs_inode_measure_sickness(ip, &sick, &checked);
fs/xfs/xfs_health.c
671
struct xfs_inode *ip,
fs/xfs/xfs_health.c
691
xfs_inode_mark_sick(ip, mask);
fs/xfs/xfs_health.c
700
xfs_bmap_mark_sick(cur->bc_ino.ip, cur->bc_ino.whichfork);
fs/xfs/xfs_health.c
715
struct xfs_inode *ip,
fs/xfs/xfs_health.c
732
xfs_inode_mark_sick(ip, mask);
fs/xfs/xfs_healthmon.c
1173
struct xfs_inode *ip = XFS_I(file_inode(file));
fs/xfs/xfs_healthmon.c
1174
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_healthmon.c
1185
if (ip->i_ino != mp->m_sb.sb_rootino)
fs/xfs/xfs_healthmon.c
528
struct xfs_inode *ip,
fs/xfs/xfs_healthmon.c
536
.ino = ip->i_ino,
fs/xfs/xfs_healthmon.c
537
.gen = VFS_I(ip)->i_generation,
fs/xfs/xfs_healthmon.c
539
struct xfs_healthmon *hm = xfs_healthmon_get(ip->i_mount);
fs/xfs/xfs_healthmon.c
643
struct xfs_inode *ip,
fs/xfs/xfs_healthmon.c
649
.fino = ip->i_ino,
fs/xfs/xfs_healthmon.c
650
.fgen = VFS_I(ip)->i_generation,
fs/xfs/xfs_healthmon.c
656
struct xfs_healthmon *hm = xfs_healthmon_get(ip->i_mount);
fs/xfs/xfs_healthmon.h
169
void xfs_healthmon_report_inode(struct xfs_inode *ip,
fs/xfs/xfs_healthmon.h
178
void xfs_healthmon_report_file_ioerror(struct xfs_inode *ip,
fs/xfs/xfs_icache.c
100
kmem_cache_free(xfs_inode_cache, ip);
fs/xfs/xfs_icache.c
1011
spin_lock(&ip->i_flags_lock);
fs/xfs/xfs_icache.c
1012
ip->i_flags = XFS_IRECLAIM;
fs/xfs/xfs_icache.c
1013
ip->i_ino = 0;
fs/xfs/xfs_icache.c
1014
ip->i_sick = 0;
fs/xfs/xfs_icache.c
1015
ip->i_checked = 0;
fs/xfs/xfs_icache.c
1016
spin_unlock(&ip->i_flags_lock);
fs/xfs/xfs_icache.c
1018
ASSERT(!ip->i_itemp || ip->i_itemp->ili_item.li_buf == NULL);
fs/xfs/xfs_icache.c
1019
xfs_iunlock(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_icache.c
1021
XFS_STATS_INC(ip->i_mount, xs_ig_reclaims);
fs/xfs/xfs_icache.c
1031
XFS_INO_TO_AGINO(ip->i_mount, ino)))
fs/xfs/xfs_icache.c
1044
xfs_ilock(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_icache.c
1045
ASSERT(!ip->i_udquot && !ip->i_gdquot && !ip->i_pdquot);
fs/xfs/xfs_icache.c
1046
xfs_iunlock(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_icache.c
1047
ASSERT(xfs_inode_clean(ip));
fs/xfs/xfs_icache.c
1049
__xfs_inode_free(ip);
fs/xfs/xfs_icache.c
105
VFS_I(ip)->i_mode = 0;
fs/xfs/xfs_icache.c
1053
xfs_iflags_clear(ip, XFS_IFLUSHING);
fs/xfs/xfs_icache.c
1055
xfs_iunlock(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_icache.c
1057
xfs_iflags_clear(ip, XFS_IRECLAIM);
fs/xfs/xfs_icache.c
106
mapping_set_folio_min_order(VFS_I(ip)->i_mapping,
fs/xfs/xfs_icache.c
110
ASSERT(atomic_read(&ip->i_pincount) == 0);
fs/xfs/xfs_icache.c
111
ASSERT(ip->i_ino == 0);
fs/xfs/xfs_icache.c
1138
struct xfs_inode *ip,
fs/xfs/xfs_icache.c
114
ip->i_ino = ino;
fs/xfs/xfs_icache.c
1142
!uid_eq(VFS_I(ip)->i_uid, icw->icw_uid))
fs/xfs/xfs_icache.c
1146
!gid_eq(VFS_I(ip)->i_gid, icw->icw_gid))
fs/xfs/xfs_icache.c
115
ip->i_mount = mp;
fs/xfs/xfs_icache.c
1150
ip->i_projid != icw->icw_prid)
fs/xfs/xfs_icache.c
116
memset(&ip->i_imap, 0, sizeof(struct xfs_imap));
fs/xfs/xfs_icache.c
1162
struct xfs_inode *ip,
fs/xfs/xfs_icache.c
1166
uid_eq(VFS_I(ip)->i_uid, icw->icw_uid))
fs/xfs/xfs_icache.c
117
ip->i_cowfp = NULL;
fs/xfs/xfs_icache.c
1170
gid_eq(VFS_I(ip)->i_gid, icw->icw_gid))
fs/xfs/xfs_icache.c
1174
ip->i_projid == icw->icw_prid)
fs/xfs/xfs_icache.c
118
memset(&ip->i_af, 0, sizeof(ip->i_af));
fs/xfs/xfs_icache.c
1187
struct xfs_inode *ip,
fs/xfs/xfs_icache.c
119
ip->i_af.if_format = XFS_DINODE_FMT_EXTENTS;
fs/xfs/xfs_icache.c
1196
match = xfs_icwalk_match_id_union(ip, icw);
fs/xfs/xfs_icache.c
1198
match = xfs_icwalk_match_id(ip, icw);
fs/xfs/xfs_icache.c
120
memset(&ip->i_df, 0, sizeof(ip->i_df));
fs/xfs/xfs_icache.c
1204
XFS_ISIZE(ip) < icw->icw_min_file_size)
fs/xfs/xfs_icache.c
121
ip->i_flags = 0;
fs/xfs/xfs_icache.c
122
ip->i_delayed_blks = 0;
fs/xfs/xfs_icache.c
1229
struct xfs_inode *ip,
fs/xfs/xfs_icache.c
123
ip->i_diflags2 = mp->m_ino_geo.new_diflags2;
fs/xfs/xfs_icache.c
1237
if (!xfs_iflags_test(ip, XFS_IEOFBLOCKS))
fs/xfs/xfs_icache.c
124
ip->i_nblocks = 0;
fs/xfs/xfs_icache.c
1244
if (!wait && mapping_tagged(VFS_I(ip)->i_mapping, PAGECACHE_TAG_DIRTY))
fs/xfs/xfs_icache.c
1247
if (!xfs_icwalk_match(ip, icw))
fs/xfs/xfs_icache.c
125
ip->i_forkoff = 0;
fs/xfs/xfs_icache.c
1254
if (!xfs_ilock_nowait(ip, XFS_IOLOCK_EXCL)) {
fs/xfs/xfs_icache.c
126
ip->i_sick = 0;
fs/xfs/xfs_icache.c
1261
if (xfs_can_free_eofblocks(ip))
fs/xfs/xfs_icache.c
1262
return xfs_free_eofblocks(ip);
fs/xfs/xfs_icache.c
1265
trace_xfs_inode_free_eofblocks_invalid(ip);
fs/xfs/xfs_icache.c
1266
xfs_inode_clear_eofblocks_tag(ip);
fs/xfs/xfs_icache.c
127
ip->i_checked = 0;
fs/xfs/xfs_icache.c
1272
struct xfs_inode *ip,
fs/xfs/xfs_icache.c
1275
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_icache.c
128
INIT_WORK(&ip->i_ioend_work, xfs_end_io);
fs/xfs/xfs_icache.c
1284
if (ip->i_flags & iflag)
fs/xfs/xfs_icache.c
1286
spin_lock(&ip->i_flags_lock);
fs/xfs/xfs_icache.c
1287
ip->i_flags |= iflag;
fs/xfs/xfs_icache.c
1288
spin_unlock(&ip->i_flags_lock);
fs/xfs/xfs_icache.c
129
INIT_LIST_HEAD(&ip->i_ioend_list);
fs/xfs/xfs_icache.c
1290
pag = xfs_perag_get(mp, XFS_INO_TO_AGNO(mp, ip->i_ino));
fs/xfs/xfs_icache.c
1293
xfs_perag_set_inode_tag(pag, XFS_INO_TO_AGINO(mp, ip->i_ino),
fs/xfs/xfs_icache.c
130
spin_lock_init(&ip->i_ioend_lock);
fs/xfs/xfs_icache.c
1302
xfs_inode_t *ip)
fs/xfs/xfs_icache.c
1304
trace_xfs_inode_set_eofblocks_tag(ip);
fs/xfs/xfs_icache.c
1305
return xfs_blockgc_set_iflag(ip, XFS_IEOFBLOCKS);
fs/xfs/xfs_icache.c
131
ip->i_next_unlinked = NULLAGINO;
fs/xfs/xfs_icache.c
1310
struct xfs_inode *ip,
fs/xfs/xfs_icache.c
1313
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_icache.c
1319
spin_lock(&ip->i_flags_lock);
fs/xfs/xfs_icache.c
132
ip->i_prev_unlinked = 0;
fs/xfs/xfs_icache.c
1320
ip->i_flags &= ~iflag;
fs/xfs/xfs_icache.c
1321
clear_tag = (ip->i_flags & (XFS_IEOFBLOCKS | XFS_ICOWBLOCKS)) == 0;
fs/xfs/xfs_icache.c
1322
spin_unlock(&ip->i_flags_lock);
fs/xfs/xfs_icache.c
1327
pag = xfs_perag_get(mp, XFS_INO_TO_AGNO(mp, ip->i_ino));
fs/xfs/xfs_icache.c
1330
xfs_perag_clear_inode_tag(pag, XFS_INO_TO_AGINO(mp, ip->i_ino),
fs/xfs/xfs_icache.c
1339
xfs_inode_t *ip)
fs/xfs/xfs_icache.c
134
return ip;
fs/xfs/xfs_icache.c
1341
trace_xfs_inode_clear_eofblocks_tag(ip);
fs/xfs/xfs_icache.c
1342
return xfs_blockgc_clear_iflag(ip, XFS_IEOFBLOCKS);
fs/xfs/xfs_icache.c
1350
struct xfs_inode *ip,
fs/xfs/xfs_icache.c
1361
if (!xfs_inode_has_cow_data(ip)) {
fs/xfs/xfs_icache.c
1362
trace_xfs_inode_free_cowblocks_invalid(ip);
fs/xfs/xfs_icache.c
1363
xfs_inode_clear_cowblocks_tag(ip);
fs/xfs/xfs_icache.c
1381
if (!sync && inode_is_open_for_write(VFS_I(ip)))
fs/xfs/xfs_icache.c
1383
return xfs_can_free_cowblocks(ip);
fs/xfs/xfs_icache.c
1400
struct xfs_inode *ip,
fs/xfs/xfs_icache.c
1409
if (!xfs_iflags_test(ip, XFS_ICOWBLOCKS))
fs/xfs/xfs_icache.c
1412
if (!xfs_prep_free_cowblocks(ip, icw))
fs/xfs/xfs_icache.c
1415
if (!xfs_icwalk_match(ip, icw))
fs/xfs/xfs_icache.c
142
struct xfs_inode *ip = XFS_I(inode);
fs/xfs/xfs_icache.c
1423
!xfs_ilock_nowait(ip, XFS_IOLOCK_EXCL)) {
fs/xfs/xfs_icache.c
1430
if (!xfs_ilock_nowait(ip, XFS_MMAPLOCK_EXCL)) {
fs/xfs/xfs_icache.c
144
switch (VFS_I(ip)->i_mode & S_IFMT) {
fs/xfs/xfs_icache.c
1441
if (xfs_prep_free_cowblocks(ip, icw))
fs/xfs/xfs_icache.c
1442
ret = xfs_reflink_cancel_cow_range(ip, 0, NULLFILEOFF, false);
fs/xfs/xfs_icache.c
1448
xfs_inode_t *ip)
fs/xfs/xfs_icache.c
1450
trace_xfs_inode_set_cowblocks_tag(ip);
fs/xfs/xfs_icache.c
1451
return xfs_blockgc_set_iflag(ip, XFS_ICOWBLOCKS);
fs/xfs/xfs_icache.c
1456
xfs_inode_t *ip)
fs/xfs/xfs_icache.c
1458
trace_xfs_inode_clear_cowblocks_tag(ip);
fs/xfs/xfs_icache.c
1459
return xfs_blockgc_clear_iflag(ip, XFS_ICOWBLOCKS);
fs/xfs/xfs_icache.c
148
xfs_idestroy_fork(&ip->i_df);
fs/xfs/xfs_icache.c
1505
struct xfs_inode *ip)
fs/xfs/xfs_icache.c
1507
struct inode *inode = VFS_I(ip);
fs/xfs/xfs_icache.c
1512
spin_lock(&ip->i_flags_lock);
fs/xfs/xfs_icache.c
1513
if (!ip->i_ino)
fs/xfs/xfs_icache.c
1516
if (ip->i_flags & XFS_BLOCKGC_NOGRAB_IFLAGS)
fs/xfs/xfs_icache.c
1518
spin_unlock(&ip->i_flags_lock);
fs/xfs/xfs_icache.c
152
xfs_ifork_zap_attr(ip);
fs/xfs/xfs_icache.c
1521
if (xfs_is_shutdown(ip->i_mount))
fs/xfs/xfs_icache.c
1532
spin_unlock(&ip->i_flags_lock);
fs/xfs/xfs_icache.c
1539
struct xfs_inode *ip,
fs/xfs/xfs_icache.c
154
if (ip->i_cowfp) {
fs/xfs/xfs_icache.c
1545
error = xfs_inode_free_eofblocks(ip, icw, &lockflags);
fs/xfs/xfs_icache.c
1549
error = xfs_inode_free_cowblocks(ip, icw, &lockflags);
fs/xfs/xfs_icache.c
155
xfs_idestroy_fork(ip->i_cowfp);
fs/xfs/xfs_icache.c
1552
xfs_iunlock(ip, lockflags);
fs/xfs/xfs_icache.c
1553
xfs_irele(ip);
fs/xfs/xfs_icache.c
156
kmem_cache_free(xfs_ifork_cache, ip->i_cowfp);
fs/xfs/xfs_icache.c
158
if (ip->i_itemp) {
fs/xfs/xfs_icache.c
160
&ip->i_itemp->ili_item.li_flags));
fs/xfs/xfs_icache.c
161
xfs_inode_item_destroy(ip);
fs/xfs/xfs_icache.c
164
kmem_cache_free(xfs_inode_cache, ip);
fs/xfs/xfs_icache.c
1678
struct xfs_inode *ip,
fs/xfs/xfs_icache.c
1681
return xfs_blockgc_free_dquots(ip->i_mount,
fs/xfs/xfs_icache.c
1682
xfs_inode_dquot(ip, XFS_DQTYPE_USER),
fs/xfs/xfs_icache.c
1683
xfs_inode_dquot(ip, XFS_DQTYPE_GROUP),
fs/xfs/xfs_icache.c
1684
xfs_inode_dquot(ip, XFS_DQTYPE_PROJ), iwalk_flags);
fs/xfs/xfs_icache.c
169
struct xfs_inode *ip)
fs/xfs/xfs_icache.c
1705
struct xfs_inode *ip,
fs/xfs/xfs_icache.c
1710
return xfs_blockgc_igrab(ip);
fs/xfs/xfs_icache.c
1712
return xfs_reclaim_igrab(ip, icw);
fs/xfs/xfs_icache.c
172
ASSERT(atomic_read(&ip->i_pincount) == 0);
fs/xfs/xfs_icache.c
1725
struct xfs_inode *ip,
fs/xfs/xfs_icache.c
173
ASSERT(!ip->i_itemp || list_empty(&ip->i_itemp->ili_item.li_bio_list));
fs/xfs/xfs_icache.c
1733
error = xfs_blockgc_scan_inode(ip, icw);
fs/xfs/xfs_icache.c
1736
xfs_reclaim_inode(ip, pag);
fs/xfs/xfs_icache.c
174
if (xfs_is_metadir_inode(ip))
fs/xfs/xfs_icache.c
175
XFS_STATS_DEC(ip->i_mount, xs_inodes_meta);
fs/xfs/xfs_icache.c
177
XFS_STATS_DEC(ip->i_mount, xs_inodes_active);
fs/xfs/xfs_icache.c
1788
struct xfs_inode *ip = batch[i];
fs/xfs/xfs_icache.c
179
call_rcu(&VFS_I(ip)->i_rcu, xfs_inode_free_callback);
fs/xfs/xfs_icache.c
1790
if (done || !xfs_icwalk_igrab(goal, ip, icw))
fs/xfs/xfs_icache.c
1805
if (XFS_INO_TO_AGNO(mp, ip->i_ino) != pag_agno(pag))
fs/xfs/xfs_icache.c
1807
first_index = XFS_INO_TO_AGINO(mp, ip->i_ino + 1);
fs/xfs/xfs_icache.c
1808
if (first_index < XFS_INO_TO_AGINO(mp, ip->i_ino))
fs/xfs/xfs_icache.c
184
struct xfs_inode *ip)
fs/xfs/xfs_icache.c
186
ASSERT(!xfs_iflags_test(ip, XFS_IFLUSHING));
fs/xfs/xfs_icache.c
1882
struct xfs_inode *ip,
fs/xfs/xfs_icache.c
1885
struct xfs_ifork *ifp = xfs_ifork_ptr(ip, whichfork);
fs/xfs/xfs_icache.c
1889
if (!ifp || !xfs_iext_lookup_extent(ip, ifp, 0, &icur, &got))
fs/xfs/xfs_icache.c
1893
xfs_warn(ip->i_mount,
fs/xfs/xfs_icache.c
1895
ip->i_ino,
fs/xfs/xfs_icache.c
1902
#define xfs_check_delalloc(ip, whichfork) do { } while (0)
fs/xfs/xfs_icache.c
1908
struct xfs_inode *ip)
fs/xfs/xfs_icache.c
1910
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_icache.c
1913
if (!xfs_is_shutdown(mp) && ip->i_delayed_blks) {
fs/xfs/xfs_icache.c
1914
xfs_check_delalloc(ip, XFS_DATA_FORK);
fs/xfs/xfs_icache.c
1915
xfs_check_delalloc(ip, XFS_COW_FORK);
fs/xfs/xfs_icache.c
1919
pag = xfs_perag_get(mp, XFS_INO_TO_AGNO(mp, ip->i_ino));
fs/xfs/xfs_icache.c
1921
spin_lock(&ip->i_flags_lock);
fs/xfs/xfs_icache.c
1923
trace_xfs_inode_set_reclaimable(ip);
fs/xfs/xfs_icache.c
1924
ip->i_flags &= ~(XFS_NEED_INACTIVE | XFS_INACTIVATING);
fs/xfs/xfs_icache.c
1925
ip->i_flags |= XFS_IRECLAIMABLE;
fs/xfs/xfs_icache.c
1926
xfs_perag_set_inode_tag(pag, XFS_INO_TO_AGINO(mp, ip->i_ino),
fs/xfs/xfs_icache.c
1929
spin_unlock(&ip->i_flags_lock);
fs/xfs/xfs_icache.c
194
spin_lock(&ip->i_flags_lock);
fs/xfs/xfs_icache.c
1941
struct xfs_inode *ip)
fs/xfs/xfs_icache.c
1945
trace_xfs_inode_inactivating(ip);
fs/xfs/xfs_icache.c
1946
error = xfs_inactive(ip);
fs/xfs/xfs_icache.c
1947
xfs_inodegc_set_reclaimable(ip);
fs/xfs/xfs_icache.c
195
ip->i_flags = XFS_IRECLAIM;
fs/xfs/xfs_icache.c
1959
struct xfs_inode *ip, *n;
fs/xfs/xfs_icache.c
196
ip->i_ino = 0;
fs/xfs/xfs_icache.c
197
spin_unlock(&ip->i_flags_lock);
fs/xfs/xfs_icache.c
1984
ip = llist_entry(node, struct xfs_inode, i_gclist);
fs/xfs/xfs_icache.c
1988
llist_for_each_entry_safe(ip, n, node, i_gclist) {
fs/xfs/xfs_icache.c
199
__xfs_inode_free(ip);
fs/xfs/xfs_icache.c
1991
xfs_iflags_set(ip, XFS_INACTIVATING);
fs/xfs/xfs_icache.c
1992
error = xfs_inodegc_inactivate(ip);
fs/xfs/xfs_icache.c
2079
struct xfs_inode *ip)
fs/xfs/xfs_icache.c
2081
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_icache.c
2083
if (!XFS_IS_REALTIME_INODE(ip) || xfs_has_zoned(mp))
fs/xfs/xfs_icache.c
2094
# define xfs_inodegc_want_queue_rt_file(ip) (false)
fs/xfs/xfs_icache.c
2106
struct xfs_inode *ip,
fs/xfs/xfs_icache.c
2109
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_icache.c
2119
if (xfs_inodegc_want_queue_rt_file(ip))
fs/xfs/xfs_icache.c
2122
if (xfs_inode_near_dquot_enforcement(ip, XFS_DQTYPE_USER))
fs/xfs/xfs_icache.c
2125
if (xfs_inode_near_dquot_enforcement(ip, XFS_DQTYPE_GROUP))
fs/xfs/xfs_icache.c
2128
if (xfs_inode_near_dquot_enforcement(ip, XFS_DQTYPE_PROJ))
fs/xfs/xfs_icache.c
2153
struct xfs_inode *ip,
fs/xfs/xfs_icache.c
2176
struct xfs_inode *ip)
fs/xfs/xfs_icache.c
2178
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_icache.c
2185
trace_xfs_inode_set_need_inactive(ip);
fs/xfs/xfs_icache.c
2186
spin_lock(&ip->i_flags_lock);
fs/xfs/xfs_icache.c
2187
ip->i_flags |= XFS_NEED_INACTIVE;
fs/xfs/xfs_icache.c
2188
spin_unlock(&ip->i_flags_lock);
fs/xfs/xfs_icache.c
2192
llist_add(&ip->i_gclist, &gc->list);
fs/xfs/xfs_icache.c
2215
if (xfs_inodegc_want_queue_work(ip, items))
fs/xfs/xfs_icache.c
2223
if (xfs_inodegc_want_flush_work(ip, items, shrinker_hits)) {
fs/xfs/xfs_icache.c
2241
struct xfs_inode *ip)
fs/xfs/xfs_icache.c
2243
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_icache.c
2251
ASSERT_ALWAYS(!xfs_iflags_test(ip, XFS_ALL_IRECLAIM_FLAGS));
fs/xfs/xfs_icache.c
2253
need_inactive = xfs_inode_needs_inactive(ip);
fs/xfs/xfs_icache.c
2255
xfs_inodegc_queue(ip);
fs/xfs/xfs_icache.c
2260
xfs_qm_dqdetach(ip);
fs/xfs/xfs_icache.c
2261
xfs_inodegc_set_reclaimable(ip);
fs/xfs/xfs_icache.c
363
struct xfs_inode *ip)
fs/xfs/xfs_icache.c
365
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_icache.c
366
struct inode *inode = VFS_I(ip);
fs/xfs/xfs_icache.c
369
trace_xfs_iget_recycle(ip);
fs/xfs/xfs_icache.c
373
xfs_iunlock(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_icache.c
380
spin_lock(&ip->i_flags_lock);
fs/xfs/xfs_icache.c
381
ip->i_flags &= ~(XFS_INEW | XFS_IRECLAIM);
fs/xfs/xfs_icache.c
382
ASSERT(ip->i_flags & XFS_IRECLAIMABLE);
fs/xfs/xfs_icache.c
383
spin_unlock(&ip->i_flags_lock);
fs/xfs/xfs_icache.c
386
trace_xfs_iget_recycle_fail(ip);
fs/xfs/xfs_icache.c
391
spin_lock(&ip->i_flags_lock);
fs/xfs/xfs_icache.c
398
ip->i_flags &= ~XFS_IRECLAIM_RESET_FLAGS;
fs/xfs/xfs_icache.c
399
ip->i_flags |= XFS_INEW;
fs/xfs/xfs_icache.c
400
xfs_perag_clear_inode_tag(pag, XFS_INO_TO_AGINO(mp, ip->i_ino),
fs/xfs/xfs_icache.c
403
spin_unlock(&ip->i_flags_lock);
fs/xfs/xfs_icache.c
421
struct xfs_inode *ip,
fs/xfs/xfs_icache.c
426
if (VFS_I(ip)->i_mode != 0) {
fs/xfs/xfs_icache.c
427
xfs_warn(ip->i_mount,
fs/xfs/xfs_icache.c
429
ip->i_ino, VFS_I(ip)->i_mode);
fs/xfs/xfs_icache.c
430
xfs_agno_mark_sick(ip->i_mount,
fs/xfs/xfs_icache.c
431
XFS_INO_TO_AGNO(ip->i_mount, ip->i_ino),
fs/xfs/xfs_icache.c
436
if (ip->i_nblocks != 0) {
fs/xfs/xfs_icache.c
437
xfs_warn(ip->i_mount,
fs/xfs/xfs_icache.c
439
ip->i_ino);
fs/xfs/xfs_icache.c
440
xfs_agno_mark_sick(ip->i_mount,
fs/xfs/xfs_icache.c
441
XFS_INO_TO_AGNO(ip->i_mount, ip->i_ino),
fs/xfs/xfs_icache.c
449
if (VFS_I(ip)->i_mode == 0)
fs/xfs/xfs_icache.c
502
struct xfs_inode *ip,
fs/xfs/xfs_icache.c
507
struct inode *inode = VFS_I(ip);
fs/xfs/xfs_icache.c
508
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_icache.c
518
spin_lock(&ip->i_flags_lock);
fs/xfs/xfs_icache.c
519
if (ip->i_ino != ino)
fs/xfs/xfs_icache.c
540
if (ip->i_flags & (XFS_INEW | XFS_IRECLAIM | XFS_INACTIVATING))
fs/xfs/xfs_icache.c
543
if (ip->i_flags & XFS_NEED_INACTIVE) {
fs/xfs/xfs_icache.c
545
if (VFS_I(ip)->i_nlink == 0) {
fs/xfs/xfs_icache.c
556
error = xfs_iget_check_free_state(ip, flags);
fs/xfs/xfs_icache.c
562
(ip->i_flags & XFS_IRECLAIMABLE))
fs/xfs/xfs_icache.c
566
if (ip->i_flags & XFS_IRECLAIMABLE) {
fs/xfs/xfs_icache.c
573
if (!xfs_ilock_nowait(ip, XFS_ILOCK_EXCL))
fs/xfs/xfs_icache.c
575
ip->i_flags |= XFS_IRECLAIM;
fs/xfs/xfs_icache.c
576
spin_unlock(&ip->i_flags_lock);
fs/xfs/xfs_icache.c
579
error = xfs_iget_recycle(pag, ip);
fs/xfs/xfs_icache.c
588
spin_unlock(&ip->i_flags_lock);
fs/xfs/xfs_icache.c
590
trace_xfs_iget_hit(ip);
fs/xfs/xfs_icache.c
594
xfs_ilock(ip, lock_flags);
fs/xfs/xfs_icache.c
597
xfs_iflags_clear(ip, XFS_ISTALE);
fs/xfs/xfs_icache.c
603
trace_xfs_iget_skip(ip);
fs/xfs/xfs_icache.c
607
spin_unlock(&ip->i_flags_lock);
fs/xfs/xfs_icache.c
612
spin_unlock(&ip->i_flags_lock);
fs/xfs/xfs_icache.c
633
struct xfs_inode *ip;
fs/xfs/xfs_icache.c
637
ip = xfs_inode_alloc(mp, ino);
fs/xfs/xfs_icache.c
638
if (!ip)
fs/xfs/xfs_icache.c
647
xfs_iflags_set(ip, XFS_INEW);
fs/xfs/xfs_icache.c
649
error = xfs_imap(pag, tp, ip->i_ino, &ip->i_imap, flags);
fs/xfs/xfs_icache.c
663
VFS_I(ip)->i_generation = get_random_u32();
fs/xfs/xfs_icache.c
667
error = xfs_imap_to_bp(mp, tp, &ip->i_imap, &bp);
fs/xfs/xfs_icache.c
671
error = xfs_inode_from_disk(ip,
fs/xfs/xfs_icache.c
672
xfs_buf_offset(bp, ip->i_imap.im_boffset));
fs/xfs/xfs_icache.c
676
xfs_inode_mark_sick(ip, XFS_SICK_INO_CORE);
fs/xfs/xfs_icache.c
683
trace_xfs_iget_miss(ip);
fs/xfs/xfs_icache.c
689
error = xfs_iget_check_free_state(ip, flags);
fs/xfs/xfs_icache.c
708
if (!xfs_ilock_nowait(ip, lock_flags))
fs/xfs/xfs_icache.c
722
d_mark_dontcache(VFS_I(ip));
fs/xfs/xfs_icache.c
723
ip->i_udquot = NULL;
fs/xfs/xfs_icache.c
724
ip->i_gdquot = NULL;
fs/xfs/xfs_icache.c
725
ip->i_pdquot = NULL;
fs/xfs/xfs_icache.c
729
error = radix_tree_insert(&pag->pag_ici_root, agino, ip);
fs/xfs/xfs_icache.c
739
*ipp = ip;
fs/xfs/xfs_icache.c
746
xfs_iunlock(ip, lock_flags);
fs/xfs/xfs_icache.c
748
__destroy_inode(VFS_I(ip));
fs/xfs/xfs_icache.c
749
xfs_inode_free(ip);
fs/xfs/xfs_icache.c
774
struct xfs_inode *ip;
fs/xfs/xfs_icache.c
794
ip = radix_tree_lookup(&pag->pag_ici_root, agino);
fs/xfs/xfs_icache.c
796
if (ip) {
fs/xfs/xfs_icache.c
797
error = xfs_iget_cache_hit(pag, ip, ino, flags, lock_flags);
fs/xfs/xfs_icache.c
808
error = xfs_iget_cache_miss(mp, pag, tp, ino, &ip,
fs/xfs/xfs_icache.c
815
*ipp = ip;
fs/xfs/xfs_icache.c
822
if (xfs_iflags_test(ip, XFS_INEW) && VFS_I(ip)->i_mode != 0)
fs/xfs/xfs_icache.c
823
xfs_setup_existing_inode(ip);
fs/xfs/xfs_icache.c
850
struct xfs_inode *ip;
fs/xfs/xfs_icache.c
854
error = xfs_iget(mp, tp, ino, 0, 0, &ip);
fs/xfs/xfs_icache.c
860
if (VFS_I(ip)->i_nlink == 0)
fs/xfs/xfs_icache.c
867
if (inode_wrong_type(VFS_I(ip), mode))
fs/xfs/xfs_icache.c
870
if (!xfs_is_metadir_inode(ip))
fs/xfs/xfs_icache.c
872
if (metafile_type != ip->i_metatype)
fs/xfs/xfs_icache.c
876
*ipp = ip;
fs/xfs/xfs_icache.c
879
xfs_irele(ip);
fs/xfs/xfs_icache.c
91
struct xfs_inode *ip;
fs/xfs/xfs_icache.c
923
struct xfs_inode *ip,
fs/xfs/xfs_icache.c
928
spin_lock(&ip->i_flags_lock);
fs/xfs/xfs_icache.c
929
if (!__xfs_iflags_test(ip, XFS_IRECLAIMABLE) ||
fs/xfs/xfs_icache.c
930
__xfs_iflags_test(ip, XFS_IRECLAIM)) {
fs/xfs/xfs_icache.c
932
spin_unlock(&ip->i_flags_lock);
fs/xfs/xfs_icache.c
937
if (ip->i_sick &&
fs/xfs/xfs_icache.c
939
spin_unlock(&ip->i_flags_lock);
fs/xfs/xfs_icache.c
943
__xfs_iflags_set(ip, XFS_IRECLAIM);
fs/xfs/xfs_icache.c
944
spin_unlock(&ip->i_flags_lock);
fs/xfs/xfs_icache.c
962
struct xfs_inode *ip,
fs/xfs/xfs_icache.c
965
xfs_ino_t ino = ip->i_ino; /* for radix_tree_delete */
fs/xfs/xfs_icache.c
967
if (!xfs_ilock_nowait(ip, XFS_ILOCK_EXCL))
fs/xfs/xfs_icache.c
969
if (xfs_iflags_test_and_set(ip, XFS_IFLUSHING))
fs/xfs/xfs_icache.c
97
ip = alloc_inode_sb(mp->m_super, xfs_inode_cache, GFP_KERNEL | __GFP_NOFAIL);
fs/xfs/xfs_icache.c
979
if (xlog_is_shutdown(ip->i_mount->m_log)) {
fs/xfs/xfs_icache.c
980
xfs_iunpin_wait(ip);
fs/xfs/xfs_icache.c
987
xfs_iunlock(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_icache.c
988
xfs_iflush_shutdown_abort(ip);
fs/xfs/xfs_icache.c
989
xfs_ilock(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_icache.c
99
if (inode_init_always(mp->m_super, VFS_I(ip))) {
fs/xfs/xfs_icache.c
992
if (xfs_ipincount(ip))
fs/xfs/xfs_icache.c
994
if (!xfs_inode_clean(ip))
fs/xfs/xfs_icache.c
997
xfs_iflags_clear(ip, XFS_IFLUSHING);
fs/xfs/xfs_icache.c
999
trace_xfs_inode_reclaiming(ip);
fs/xfs/xfs_icache.h
50
void xfs_inode_free(struct xfs_inode *ip);
fs/xfs/xfs_icache.h
58
void xfs_inode_mark_reclaimable(struct xfs_inode *ip);
fs/xfs/xfs_icache.h
63
int xfs_blockgc_free_quota(struct xfs_inode *ip, unsigned int iwalk_flags);
fs/xfs/xfs_icache.h
67
void xfs_inode_set_eofblocks_tag(struct xfs_inode *ip);
fs/xfs/xfs_icache.h
68
void xfs_inode_clear_eofblocks_tag(struct xfs_inode *ip);
fs/xfs/xfs_icache.h
70
void xfs_inode_set_cowblocks_tag(struct xfs_inode *ip);
fs/xfs/xfs_icache.h
71
void xfs_inode_clear_cowblocks_tag(struct xfs_inode *ip);
fs/xfs/xfs_inode.c
1004
if (!xfs_is_reflink_inode(ip))
fs/xfs/xfs_inode.c
1006
dfork = xfs_ifork_ptr(ip, XFS_DATA_FORK);
fs/xfs/xfs_inode.c
1007
cfork = xfs_ifork_ptr(ip, XFS_COW_FORK);
fs/xfs/xfs_inode.c
1009
ip->i_diflags2 &= ~XFS_DIFLAG2_REFLINK;
fs/xfs/xfs_inode.c
1011
xfs_inode_clear_cowblocks_tag(ip);
fs/xfs/xfs_inode.c
1038
struct xfs_inode *ip,
fs/xfs/xfs_inode.c
1043
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_inode.c
1048
xfs_assert_ilocked(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_inode.c
1049
if (icount_read(VFS_I(ip)))
fs/xfs/xfs_inode.c
1050
xfs_assert_ilocked(ip, XFS_IOLOCK_EXCL);
fs/xfs/xfs_inode.c
1052
ASSERT(new_size <= XFS_ISIZE(ip));
fs/xfs/xfs_inode.c
1054
ASSERT(ip->i_itemp != NULL);
fs/xfs/xfs_inode.c
1055
ASSERT(ip->i_itemp->ili_lock_flags == 0);
fs/xfs/xfs_inode.c
1056
ASSERT(!XFS_NOT_DQATTACHED(mp, ip));
fs/xfs/xfs_inode.c
1058
trace_xfs_itruncate_extents_start(ip, new_size);
fs/xfs/xfs_inode.c
1078
error = xfs_bunmapi_range(&tp, ip, flags, first_unmap_block,
fs/xfs/xfs_inode.c
1085
error = xfs_reflink_cancel_cow_blocks(ip, &tp,
fs/xfs/xfs_inode.c
1090
xfs_itruncate_clear_reflink_flags(ip);
fs/xfs/xfs_inode.c
1097
xfs_trans_log_inode(tp, ip, XFS_ILOG_CORE);
fs/xfs/xfs_inode.c
1099
trace_xfs_itruncate_extents_end(ip, new_size);
fs/xfs/xfs_inode.c
1158
struct xfs_inode *ip)
fs/xfs/xfs_inode.c
1160
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_inode.c
1169
xfs_ilock(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_inode.c
1170
xfs_trans_ijoin(tp, ip, 0);
fs/xfs/xfs_inode.c
1177
ip->i_disk_size = 0;
fs/xfs/xfs_inode.c
1178
xfs_trans_log_inode(tp, ip, XFS_ILOG_CORE);
fs/xfs/xfs_inode.c
1180
error = xfs_itruncate_extents(&tp, ip, XFS_DATA_FORK, 0);
fs/xfs/xfs_inode.c
1184
ASSERT(ip->i_df.if_nextents == 0);
fs/xfs/xfs_inode.c
1190
xfs_iunlock(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_inode.c
1196
xfs_iunlock(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_inode.c
1207
struct xfs_inode *ip)
fs/xfs/xfs_inode.c
1209
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_inode.c
1262
xfs_ilock(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_inode.c
1263
xfs_trans_ijoin(tp, ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_inode.c
1265
error = xfs_ifree(tp, ip);
fs/xfs/xfs_inode.c
1266
xfs_assert_ilocked(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_inode.c
1285
xfs_trans_mod_dquot_byino(tp, ip, XFS_TRANS_DQ_ICOUNT, -1);
fs/xfs/xfs_inode.c
1298
struct xfs_inode *ip)
fs/xfs/xfs_inode.c
1300
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_inode.c
1301
struct xfs_ifork *cow_ifp = xfs_ifork_ptr(ip, XFS_COW_FORK);
fs/xfs/xfs_inode.c
1307
if (VFS_I(ip)->i_mode == 0)
fs/xfs/xfs_inode.c
1322
if (xfs_is_internal_inode(ip))
fs/xfs/xfs_inode.c
1330
if (VFS_I(ip)->i_nlink == 0)
fs/xfs/xfs_inode.c
1341
return xfs_can_free_eofblocks(ip);
fs/xfs/xfs_inode.c
1350
struct xfs_inode *ip)
fs/xfs/xfs_inode.c
1352
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_inode.c
1357
xfs_inode_measure_sickness(ip, &sick, &checked);
fs/xfs/xfs_inode.c
1361
trace_xfs_inode_unfixed_corruption(ip, sick);
fs/xfs/xfs_inode.c
1366
pag = xfs_perag_get(mp, XFS_INO_TO_AGNO(mp, ip->i_ino));
fs/xfs/xfs_inode.c
1387
xfs_inode_t *ip)
fs/xfs/xfs_inode.c
1397
if (VFS_I(ip)->i_mode == 0) {
fs/xfs/xfs_inode.c
1398
ASSERT(ip->i_df.if_broot_bytes == 0);
fs/xfs/xfs_inode.c
1402
mp = ip->i_mount;
fs/xfs/xfs_inode.c
1403
ASSERT(!xfs_iflags_test(ip, XFS_IRECOVERY));
fs/xfs/xfs_inode.c
1405
xfs_inactive_health(ip);
fs/xfs/xfs_inode.c
141
xfs_inode_t *ip,
fs/xfs/xfs_inode.c
1415
if (xfs_is_internal_inode(ip))
fs/xfs/xfs_inode.c
1419
if (xfs_inode_has_cow_data(ip)) {
fs/xfs/xfs_inode.c
1420
error = xfs_reflink_cancel_cow_range(ip, 0, NULLFILEOFF, true);
fs/xfs/xfs_inode.c
1425
if (VFS_I(ip)->i_nlink != 0) {
fs/xfs/xfs_inode.c
1431
if (xfs_can_free_eofblocks(ip))
fs/xfs/xfs_inode.c
1432
error = xfs_free_eofblocks(ip);
fs/xfs/xfs_inode.c
1437
if (S_ISREG(VFS_I(ip)->i_mode) &&
fs/xfs/xfs_inode.c
1438
(ip->i_disk_size != 0 || XFS_ISIZE(ip) != 0 ||
fs/xfs/xfs_inode.c
1439
xfs_inode_has_filedata(ip)))
fs/xfs/xfs_inode.c
144
trace_xfs_ilock(ip, lock_flags, _RET_IP_);
fs/xfs/xfs_inode.c
1442
if (xfs_iflags_test(ip, XFS_IQUOTAUNCHECKED)) {
fs/xfs/xfs_inode.c
1451
xfs_qm_dqdetach(ip);
fs/xfs/xfs_inode.c
1453
error = xfs_qm_dqattach(ip);
fs/xfs/xfs_inode.c
1458
if (S_ISDIR(VFS_I(ip)->i_mode) && ip->i_df.if_nextents > 0) {
fs/xfs/xfs_inode.c
1459
xfs_inactive_dir(ip);
fs/xfs/xfs_inode.c
1463
if (S_ISLNK(VFS_I(ip)->i_mode))
fs/xfs/xfs_inode.c
1464
error = xfs_inactive_symlink(ip);
fs/xfs/xfs_inode.c
1466
error = xfs_inactive_truncate(ip);
fs/xfs/xfs_inode.c
1475
if (xfs_inode_has_attr_fork(ip)) {
fs/xfs/xfs_inode.c
1476
error = xfs_attr_inactive(ip);
fs/xfs/xfs_inode.c
1481
ASSERT(ip->i_forkoff == 0);
fs/xfs/xfs_inode.c
1486
error = xfs_inactive_ifree(ip);
fs/xfs/xfs_inode.c
149
down_write_nested(&VFS_I(ip)->i_rwsem,
fs/xfs/xfs_inode.c
1493
xfs_qm_dqdetach(ip);
fs/xfs/xfs_inode.c
1508
struct xfs_inode *ip;
fs/xfs/xfs_inode.c
1511
ip = radix_tree_lookup(&pag->pag_ici_root, agino);
fs/xfs/xfs_inode.c
1512
if (!ip) {
fs/xfs/xfs_inode.c
152
down_read_nested(&VFS_I(ip)->i_rwsem,
fs/xfs/xfs_inode.c
1522
if (WARN_ON_ONCE(!ip->i_ino)) {
fs/xfs/xfs_inode.c
1526
ASSERT(!xfs_iflags_test(ip, XFS_IRECLAIMABLE | XFS_IRECLAIM));
fs/xfs/xfs_inode.c
1528
return ip;
fs/xfs/xfs_inode.c
157
down_write_nested(&VFS_I(ip)->i_mapping->invalidate_lock,
fs/xfs/xfs_inode.c
160
down_read_nested(&VFS_I(ip)->i_mapping->invalidate_lock,
fs/xfs/xfs_inode.c
1604
struct xfs_inode *ip;
fs/xfs/xfs_inode.c
1608
ip = radix_tree_lookup(&pag->pag_ici_root, XFS_INO_TO_AGINO(mp, inum));
fs/xfs/xfs_inode.c
1611
if (!ip) {
fs/xfs/xfs_inode.c
1622
spin_lock(&ip->i_flags_lock);
fs/xfs/xfs_inode.c
1623
if (ip->i_ino != inum || __xfs_iflags_test(ip, XFS_ISTALE))
fs/xfs/xfs_inode.c
1632
if (ip != free_ip) {
fs/xfs/xfs_inode.c
1633
if (!xfs_ilock_nowait(ip, XFS_ILOCK_EXCL)) {
fs/xfs/xfs_inode.c
1634
spin_unlock(&ip->i_flags_lock);
fs/xfs/xfs_inode.c
1640
ip->i_flags |= XFS_ISTALE;
fs/xfs/xfs_inode.c
1647
iip = ip->i_itemp;
fs/xfs/xfs_inode.c
1648
if (__xfs_iflags_test(ip, XFS_IFLUSHING)) {
fs/xfs/xfs_inode.c
165
down_write_nested(&ip->i_lock, XFS_ILOCK_DEP(lock_flags));
fs/xfs/xfs_inode.c
1663
__xfs_iflags_set(ip, XFS_IFLUSHING);
fs/xfs/xfs_inode.c
1664
spin_unlock(&ip->i_flags_lock);
fs/xfs/xfs_inode.c
167
down_read_nested(&ip->i_lock, XFS_ILOCK_DEP(lock_flags));
fs/xfs/xfs_inode.c
1674
if (ip != free_ip)
fs/xfs/xfs_inode.c
1675
xfs_iunlock(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_inode.c
1679
if (ip != free_ip)
fs/xfs/xfs_inode.c
1680
xfs_iunlock(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_inode.c
1682
spin_unlock(&ip->i_flags_lock);
fs/xfs/xfs_inode.c
1791
struct xfs_inode *ip)
fs/xfs/xfs_inode.c
1793
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_inode.c
1796
struct xfs_inode_log_item *iip = ip->i_itemp;
fs/xfs/xfs_inode.c
1799
xfs_assert_ilocked(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_inode.c
1800
ASSERT(VFS_I(ip)->i_nlink == 0);
fs/xfs/xfs_inode.c
1801
ASSERT(ip->i_df.if_nextents == 0);
fs/xfs/xfs_inode.c
1802
ASSERT(ip->i_disk_size == 0 || !S_ISREG(VFS_I(ip)->i_mode));
fs/xfs/xfs_inode.c
1803
ASSERT(ip->i_nblocks == 0);
fs/xfs/xfs_inode.c
1805
pag = xfs_perag_get(mp, XFS_INO_TO_AGNO(mp, ip->i_ino));
fs/xfs/xfs_inode.c
1807
error = xfs_inode_uninit(tp, pag, ip, &xic);
fs/xfs/xfs_inode.c
1811
if (xfs_iflags_test(ip, XFS_IPRESERVE_DM_FIELDS))
fs/xfs/xfs_inode.c
1812
xfs_iflags_clear(ip, XFS_IPRESERVE_DM_FIELDS);
fs/xfs/xfs_inode.c
1820
error = xfs_ifree_cluster(tp, pag, ip, &xic);
fs/xfs/xfs_inode.c
1833
struct xfs_inode *ip)
fs/xfs/xfs_inode.c
1835
struct xfs_inode_log_item *iip = ip->i_itemp;
fs/xfs/xfs_inode.c
1838
trace_xfs_inode_unpin_nowait(ip, _RET_IP_);
fs/xfs/xfs_inode.c
1839
xfs_assert_ilocked(ip, XFS_ILOCK_EXCL | XFS_ILOCK_SHARED);
fs/xfs/xfs_inode.c
184
xfs_inode_t *ip,
fs/xfs/xfs_inode.c
1848
xfs_log_force_seq(ip->i_mount, seq, 0, NULL);
fs/xfs/xfs_inode.c
1854
struct xfs_inode *ip)
fs/xfs/xfs_inode.c
1856
wait_queue_head_t *wq = bit_waitqueue(&ip->i_flags, __XFS_IPINNED_BIT);
fs/xfs/xfs_inode.c
1857
DEFINE_WAIT_BIT(wait, &ip->i_flags, __XFS_IPINNED_BIT);
fs/xfs/xfs_inode.c
1859
xfs_iunpin(ip);
fs/xfs/xfs_inode.c
1863
if (xfs_ipincount(ip))
fs/xfs/xfs_inode.c
1865
} while (xfs_ipincount(ip));
fs/xfs/xfs_inode.c
187
trace_xfs_ilock_nowait(ip, lock_flags, _RET_IP_);
fs/xfs/xfs_inode.c
1871
struct xfs_inode *ip)
fs/xfs/xfs_inode.c
1873
if (xfs_ipincount(ip))
fs/xfs/xfs_inode.c
1874
__xfs_iunpin_wait(ip);
fs/xfs/xfs_inode.c
1908
struct xfs_inode *ip)
fs/xfs/xfs_inode.c
1913
.ip = ip,
fs/xfs/xfs_inode.c
1917
int is_dir = S_ISDIR(VFS_I(ip)->i_mode);
fs/xfs/xfs_inode.c
192
if (!down_write_trylock(&VFS_I(ip)->i_rwsem))
fs/xfs/xfs_inode.c
1933
error = xfs_qm_dqattach(ip);
fs/xfs/xfs_inode.c
195
if (!down_read_trylock(&VFS_I(ip)->i_rwsem))
fs/xfs/xfs_inode.c
1953
error = xfs_trans_alloc_dir(dp, &M_RES(mp)->tr_remove, ip, &resblks,
fs/xfs/xfs_inode.c
1976
if (is_dir && xfs_inode_is_filestream(ip))
fs/xfs/xfs_inode.c
1977
xfs_filestream_deassociate(ip);
fs/xfs/xfs_inode.c
1979
xfs_iunlock(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_inode.c
1987
xfs_iunlock(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_inode.c
200
if (!down_write_trylock(&VFS_I(ip)->i_mapping->invalidate_lock))
fs/xfs/xfs_inode.c
203
if (!down_read_trylock(&VFS_I(ip)->i_mapping->invalidate_lock))
fs/xfs/xfs_inode.c
208
if (!down_write_trylock(&ip->i_lock))
fs/xfs/xfs_inode.c
211
if (!down_read_trylock(&ip->i_lock))
fs/xfs/xfs_inode.c
2138
.ip = src_ip,
fs/xfs/xfs_inode.c
2143
.ip = target_ip,
fs/xfs/xfs_inode.c
2169
&du_wip.ip);
fs/xfs/xfs_inode.c
2177
xfs_sort_for_rename(src_dp, target_dp, src_ip, target_ip, du_wip.ip,
fs/xfs/xfs_inode.c
218
up_write(&VFS_I(ip)->i_mapping->invalidate_lock);
fs/xfs/xfs_inode.c
2184
if (du_wip.ip) {
fs/xfs/xfs_inode.c
2199
target_name->len, du_wip.ip != NULL);
fs/xfs/xfs_inode.c
220
up_read(&VFS_I(ip)->i_mapping->invalidate_lock);
fs/xfs/xfs_inode.c
223
up_write(&VFS_I(ip)->i_rwsem);
fs/xfs/xfs_inode.c
2246
if (du_wip.ip)
fs/xfs/xfs_inode.c
2247
xfs_trans_ijoin(tp, du_wip.ip, 0);
fs/xfs/xfs_inode.c
225
up_read(&VFS_I(ip)->i_rwsem);
fs/xfs/xfs_inode.c
2308
if (inodes[i] == du_wip.ip ||
fs/xfs/xfs_inode.c
2328
if (du_wip.ip) {
fs/xfs/xfs_inode.c
2334
inode_state_clear_raw(VFS_I(du_wip.ip), I_LINKABLE);
fs/xfs/xfs_inode.c
2360
if (du_wip.ip)
fs/xfs/xfs_inode.c
2361
xfs_irele(du_wip.ip);
fs/xfs/xfs_inode.c
2369
struct xfs_inode *ip,
fs/xfs/xfs_inode.c
2372
struct xfs_inode_log_item *iip = ip->i_itemp;
fs/xfs/xfs_inode.c
2374
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_inode.c
2377
xfs_assert_ilocked(ip, XFS_ILOCK_EXCL | XFS_ILOCK_SHARED);
fs/xfs/xfs_inode.c
2378
ASSERT(xfs_iflags_test(ip, XFS_IFLUSHING));
fs/xfs/xfs_inode.c
2379
ASSERT(ip->i_df.if_format != XFS_DINODE_FMT_BTREE ||
fs/xfs/xfs_inode.c
2380
ip->i_df.if_nextents > XFS_IFORK_MAXEXT(ip, XFS_DATA_FORK));
fs/xfs/xfs_inode.c
2383
dip = xfs_buf_offset(bp, ip->i_imap.im_boffset);
fs/xfs/xfs_inode.c
2396
__func__, ip->i_ino, be16_to_cpu(dip->di_magic), dip);
fs/xfs/xfs_inode.c
2399
if (ip->i_df.if_format == XFS_DINODE_FMT_META_BTREE) {
fs/xfs/xfs_inode.c
2400
if (!S_ISREG(VFS_I(ip)->i_mode) ||
fs/xfs/xfs_inode.c
2401
!(ip->i_diflags2 & XFS_DIFLAG2_METADATA)) {
fs/xfs/xfs_inode.c
2404
__func__, xfs_metafile_type_str(ip->i_metatype),
fs/xfs/xfs_inode.c
2405
ip->i_ino, ip);
fs/xfs/xfs_inode.c
2408
} else if (S_ISREG(VFS_I(ip)->i_mode)) {
fs/xfs/xfs_inode.c
2409
if ((ip->i_df.if_format != XFS_DINODE_FMT_EXTENTS &&
fs/xfs/xfs_inode.c
2410
ip->i_df.if_format != XFS_DINODE_FMT_BTREE) ||
fs/xfs/xfs_inode.c
2414
__func__, ip->i_ino, ip);
fs/xfs/xfs_inode.c
2417
} else if (S_ISDIR(VFS_I(ip)->i_mode)) {
fs/xfs/xfs_inode.c
2418
if ((ip->i_df.if_format != XFS_DINODE_FMT_EXTENTS &&
fs/xfs/xfs_inode.c
2419
ip->i_df.if_format != XFS_DINODE_FMT_BTREE &&
fs/xfs/xfs_inode.c
2420
ip->i_df.if_format != XFS_DINODE_FMT_LOCAL) ||
fs/xfs/xfs_inode.c
2424
__func__, ip->i_ino, ip);
fs/xfs/xfs_inode.c
2428
if (ip->i_df.if_nextents + xfs_ifork_nextents(&ip->i_af) >
fs/xfs/xfs_inode.c
2429
ip->i_nblocks || XFS_TEST_ERROR(mp, XFS_ERRTAG_IFLUSH_5)) {
fs/xfs/xfs_inode.c
2433
__func__, ip->i_ino,
fs/xfs/xfs_inode.c
2434
ip->i_df.if_nextents + xfs_ifork_nextents(&ip->i_af),
fs/xfs/xfs_inode.c
2435
ip->i_nblocks, ip);
fs/xfs/xfs_inode.c
2438
if (ip->i_forkoff > mp->m_sb.sb_inodesize ||
fs/xfs/xfs_inode.c
244
xfs_inode_t *ip,
fs/xfs/xfs_inode.c
2442
__func__, ip->i_ino, ip->i_forkoff, ip);
fs/xfs/xfs_inode.c
2446
if (xfs_inode_has_attr_fork(ip) &&
fs/xfs/xfs_inode.c
2447
ip->i_af.if_format == XFS_DINODE_FMT_META_BTREE) {
fs/xfs/xfs_inode.c
2450
__func__, ip->i_ino, ip);
fs/xfs/xfs_inode.c
2463
ip->i_flushiter++;
fs/xfs/xfs_inode.c
2469
if (ip->i_df.if_format == XFS_DINODE_FMT_LOCAL &&
fs/xfs/xfs_inode.c
2470
xfs_ifork_verify_local_data(ip))
fs/xfs/xfs_inode.c
2472
if (xfs_inode_has_attr_fork(ip) &&
fs/xfs/xfs_inode.c
2473
ip->i_af.if_format == XFS_DINODE_FMT_LOCAL &&
fs/xfs/xfs_inode.c
2474
xfs_ifork_verify_local_attr(ip))
fs/xfs/xfs_inode.c
2482
xfs_inode_to_disk(ip, dip, iip->ili_item.li_lsn);
fs/xfs/xfs_inode.c
2486
if (ip->i_flushiter == DI_MAX_FLUSH)
fs/xfs/xfs_inode.c
2487
ip->i_flushiter = 0;
fs/xfs/xfs_inode.c
2490
xfs_iflush_fork(ip, dip, iip, XFS_DATA_FORK);
fs/xfs/xfs_inode.c
2491
if (xfs_inode_has_attr_fork(ip))
fs/xfs/xfs_inode.c
2492
xfs_iflush_fork(ip, dip, iip, XFS_ATTR_FORK);
fs/xfs/xfs_inode.c
250
up_write(&VFS_I(ip)->i_rwsem);
fs/xfs/xfs_inode.c
252
up_read(&VFS_I(ip)->i_rwsem);
fs/xfs/xfs_inode.c
2528
xfs_inode_mark_sick(ip, XFS_SICK_INO_CORE);
fs/xfs/xfs_inode.c
255
up_write(&VFS_I(ip)->i_mapping->invalidate_lock);
fs/xfs/xfs_inode.c
2551
struct xfs_inode *ip;
fs/xfs/xfs_inode.c
2562
ip = iip->ili_inode;
fs/xfs/xfs_inode.c
2567
if (__xfs_iflags_test(ip, XFS_IRECLAIM | XFS_IFLUSHING))
fs/xfs/xfs_inode.c
2569
if (xfs_ipincount(ip))
fs/xfs/xfs_inode.c
257
up_read(&VFS_I(ip)->i_mapping->invalidate_lock);
fs/xfs/xfs_inode.c
2579
spin_lock(&ip->i_flags_lock);
fs/xfs/xfs_inode.c
2580
ASSERT(!__xfs_iflags_test(ip, XFS_ISTALE));
fs/xfs/xfs_inode.c
2581
if (__xfs_iflags_test(ip, XFS_IRECLAIM | XFS_IFLUSHING)) {
fs/xfs/xfs_inode.c
2582
spin_unlock(&ip->i_flags_lock);
fs/xfs/xfs_inode.c
2592
if (!xfs_ilock_nowait(ip, XFS_ILOCK_SHARED)) {
fs/xfs/xfs_inode.c
2593
spin_unlock(&ip->i_flags_lock);
fs/xfs/xfs_inode.c
2596
__xfs_iflags_set(ip, XFS_IFLUSHING);
fs/xfs/xfs_inode.c
2597
spin_unlock(&ip->i_flags_lock);
fs/xfs/xfs_inode.c
260
up_write(&ip->i_lock);
fs/xfs/xfs_inode.c
2607
xfs_iunpin_wait(ip);
fs/xfs/xfs_inode.c
2608
xfs_iflush_abort(ip);
fs/xfs/xfs_inode.c
2609
xfs_iunlock(ip, XFS_ILOCK_SHARED);
fs/xfs/xfs_inode.c
2615
if (xfs_ipincount(ip)) {
fs/xfs/xfs_inode.c
2616
xfs_iflags_clear(ip, XFS_IFLUSHING);
fs/xfs/xfs_inode.c
2617
xfs_iunlock(ip, XFS_ILOCK_SHARED);
fs/xfs/xfs_inode.c
262
up_read(&ip->i_lock);
fs/xfs/xfs_inode.c
2621
if (!xfs_inode_clean(ip))
fs/xfs/xfs_inode.c
2622
error = xfs_iflush(ip, bp);
fs/xfs/xfs_inode.c
2624
xfs_iflags_clear(ip, XFS_IFLUSHING);
fs/xfs/xfs_inode.c
2625
xfs_iunlock(ip, XFS_ILOCK_SHARED);
fs/xfs/xfs_inode.c
264
trace_xfs_iunlock(ip, lock_flags, _RET_IP_);
fs/xfs/xfs_inode.c
2660
struct xfs_inode *ip)
fs/xfs/xfs_inode.c
2662
trace_xfs_irele(ip, _RET_IP_);
fs/xfs/xfs_inode.c
2663
iput(VFS_I(ip));
fs/xfs/xfs_inode.c
2671
struct xfs_inode *ip)
fs/xfs/xfs_inode.c
2673
struct xfs_inode_log_item *iip = ip->i_itemp;
fs/xfs/xfs_inode.c
2685
return xfs_log_force_seq(ip->i_mount, seq, XFS_LOG_SYNC, NULL);
fs/xfs/xfs_inode.c
273
xfs_inode_t *ip,
fs/xfs/xfs_inode.c
281
downgrade_write(&ip->i_lock);
fs/xfs/xfs_inode.c
283
downgrade_write(&VFS_I(ip)->i_mapping->invalidate_lock);
fs/xfs/xfs_inode.c
285
downgrade_write(&VFS_I(ip)->i_rwsem);
fs/xfs/xfs_inode.c
2856
struct xfs_inode *ip)
fs/xfs/xfs_inode.c
2862
xfs_agnumber_t agno = XFS_INO_TO_AGNO(mp, ip->i_ino);
fs/xfs/xfs_inode.c
2863
xfs_agino_t agino = XFS_INO_TO_AGINO(mp, ip->i_ino);
fs/xfs/xfs_inode.c
287
trace_xfs_ilock_demote(ip, lock_flags, _RET_IP_);
fs/xfs/xfs_inode.c
2881
if (!xfs_inode_unlinked_incomplete(ip)) {
fs/xfs/xfs_inode.c
2889
trace_xfs_inode_reload_unlinked_bucket(ip);
fs/xfs/xfs_inode.c
2902
next_ip = ip;
fs/xfs/xfs_inode.c
292
struct xfs_inode *ip,
fs/xfs/xfs_inode.c
2944
struct xfs_inode *ip)
fs/xfs/xfs_inode.c
2949
tp = xfs_trans_alloc_empty(ip->i_mount);
fs/xfs/xfs_inode.c
2950
xfs_ilock(ip, XFS_ILOCK_SHARED);
fs/xfs/xfs_inode.c
2951
if (xfs_inode_unlinked_incomplete(ip))
fs/xfs/xfs_inode.c
2952
error = xfs_inode_reload_unlinked_bucket(tp, ip);
fs/xfs/xfs_inode.c
2953
xfs_iunlock(ip, XFS_ILOCK_SHARED);
fs/xfs/xfs_inode.c
2962
const struct xfs_inode *ip,
fs/xfs/xfs_inode.c
2969
switch (ip->i_vnode.i_mode & S_IFMT) {
fs/xfs/xfs_inode.c
2977
return ip->i_sick & (XFS_SICK_INO_BMBTD_ZAPPED | datamask);
fs/xfs/xfs_inode.c
2979
return ip->i_sick & XFS_SICK_INO_BMBTA_ZAPPED;
fs/xfs/xfs_inode.c
2989
struct xfs_inode *ip,
fs/xfs/xfs_inode.c
2993
struct xfs_ifork *ifp = xfs_ifork_ptr(ip, XFS_DATA_FORK);
fs/xfs/xfs_inode.c
2996
if (XFS_IS_REALTIME_INODE(ip))
fs/xfs/xfs_inode.c
2998
*dblocks = ip->i_nblocks - *rblocks;
fs/xfs/xfs_inode.c
300
rwsem_assert_held(&ip->i_lock);
fs/xfs/xfs_inode.c
3005
struct xfs_inode *ip = XFS_I(inode);
fs/xfs/xfs_inode.c
3007
xfs_iunlock(ip, XFS_MMAPLOCK_EXCL);
fs/xfs/xfs_inode.c
3009
xfs_ilock(ip, XFS_MMAPLOCK_EXCL);
fs/xfs/xfs_inode.c
302
rwsem_assert_held_write_nolockdep(&ip->i_lock);
fs/xfs/xfs_inode.c
305
rwsem_assert_held(&VFS_I(ip)->i_mapping->invalidate_lock);
fs/xfs/xfs_inode.c
3055
struct xfs_inode *ip)
fs/xfs/xfs_inode.c
3059
if (XFS_IS_REALTIME_INODE(ip))
fs/xfs/xfs_inode.c
3060
blocks = ip->i_mount->m_sb.sb_rextsize;
fs/xfs/xfs_inode.c
3062
return XFS_FSB_TO_B(ip->i_mount, blocks);
fs/xfs/xfs_inode.c
3068
const struct xfs_inode *ip)
fs/xfs/xfs_inode.c
307
rwsem_assert_held_write(&VFS_I(ip)->i_mapping->invalidate_lock);
fs/xfs/xfs_inode.c
3070
return xfs_is_zoned_inode(ip) ||
fs/xfs/xfs_inode.c
3071
(ip->i_mount->m_always_cow && xfs_has_reflink(ip->i_mount));
fs/xfs/xfs_inode.c
310
rwsem_assert_held(&VFS_I(ip)->i_rwsem);
fs/xfs/xfs_inode.c
312
rwsem_assert_held_write(&VFS_I(ip)->i_rwsem);
fs/xfs/xfs_inode.c
593
struct xfs_inode *ip = NULL;
fs/xfs/xfs_inode.c
600
error = xfs_iget(mp, tp, ino, XFS_IGET_CREATE, XFS_ILOCK_EXCL, &ip);
fs/xfs/xfs_inode.c
604
ASSERT(ip != NULL);
fs/xfs/xfs_inode.c
605
xfs_trans_ijoin(tp, ip, 0);
fs/xfs/xfs_inode.c
606
xfs_inode_init(tp, args, ip);
fs/xfs/xfs_inode.c
609
xfs_setup_inode(ip);
fs/xfs/xfs_inode.c
611
*ipp = ip;
fs/xfs/xfs_inode.c
67
struct xfs_inode *ip)
fs/xfs/xfs_inode.c
71
if (xfs_need_iread_extents(&ip->i_df))
fs/xfs/xfs_inode.c
721
error = xfs_icreate(tp, ino, args, &du.ip);
fs/xfs/xfs_inode.c
73
xfs_ilock(ip, lock_mode);
fs/xfs/xfs_inode.c
751
xfs_qm_vop_create_dqattach(tp, du.ip, udqp, gdqp, pdqp);
fs/xfs/xfs_inode.c
761
*ipp = du.ip;
fs/xfs/xfs_inode.c
762
xfs_iunlock(du.ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_inode.c
775
if (du.ip) {
fs/xfs/xfs_inode.c
776
xfs_iunlock(du.ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_inode.c
777
xfs_finish_inode_setup(du.ip);
fs/xfs/xfs_inode.c
778
xfs_irele(du.ip);
fs/xfs/xfs_inode.c
79
struct xfs_inode *ip)
fs/xfs/xfs_inode.c
799
struct xfs_inode *ip = NULL;
fs/xfs/xfs_inode.c
829
error = xfs_icreate(tp, ino, args, &ip);
fs/xfs/xfs_inode.c
83
if (xfs_inode_has_attr_fork(ip) && xfs_need_iread_extents(&ip->i_af))
fs/xfs/xfs_inode.c
841
xfs_qm_vop_create_dqattach(tp, ip, udqp, gdqp, pdqp);
fs/xfs/xfs_inode.c
843
error = xfs_iunlink(tp, ip);
fs/xfs/xfs_inode.c
85
xfs_ilock(ip, lock_mode);
fs/xfs/xfs_inode.c
855
*ipp = ip;
fs/xfs/xfs_inode.c
856
xfs_iunlock(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_inode.c
867
if (ip) {
fs/xfs/xfs_inode.c
868
xfs_iunlock(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_inode.c
869
xfs_finish_inode_setup(ip);
fs/xfs/xfs_inode.c
870
xfs_irele(ip);
fs/xfs/xfs_inode.c
918
.ip = sip,
fs/xfs/xfs_inode.c
999
struct xfs_inode *ip)
fs/xfs/xfs_inode.h
104
static inline bool xfs_inode_on_unlinked_list(const struct xfs_inode *ip)
fs/xfs/xfs_inode.h
106
return ip->i_prev_unlinked != 0;
fs/xfs/xfs_inode.h
109
static inline bool xfs_inode_has_attr_fork(const struct xfs_inode *ip)
fs/xfs/xfs_inode.h
111
return ip->i_forkoff > 0;
fs/xfs/xfs_inode.h
116
struct xfs_inode *ip,
fs/xfs/xfs_inode.h
121
return &ip->i_df;
fs/xfs/xfs_inode.h
123
if (!xfs_inode_has_attr_fork(ip))
fs/xfs/xfs_inode.h
125
return &ip->i_af;
fs/xfs/xfs_inode.h
127
return ip->i_cowfp;
fs/xfs/xfs_inode.h
134
static inline unsigned int xfs_inode_fork_boff(struct xfs_inode *ip)
fs/xfs/xfs_inode.h
136
return ip->i_forkoff << 3;
fs/xfs/xfs_inode.h
139
static inline unsigned int xfs_inode_data_fork_size(struct xfs_inode *ip)
fs/xfs/xfs_inode.h
141
if (xfs_inode_has_attr_fork(ip))
fs/xfs/xfs_inode.h
142
return xfs_inode_fork_boff(ip);
fs/xfs/xfs_inode.h
144
return XFS_LITINO(ip->i_mount);
fs/xfs/xfs_inode.h
147
static inline unsigned int xfs_inode_attr_fork_size(struct xfs_inode *ip)
fs/xfs/xfs_inode.h
149
if (xfs_inode_has_attr_fork(ip))
fs/xfs/xfs_inode.h
150
return XFS_LITINO(ip->i_mount) - xfs_inode_fork_boff(ip);
fs/xfs/xfs_inode.h
156
struct xfs_inode *ip,
fs/xfs/xfs_inode.h
161
return xfs_inode_data_fork_size(ip);
fs/xfs/xfs_inode.h
163
return xfs_inode_attr_fork_size(ip);
fs/xfs/xfs_inode.h
176
static inline struct inode *VFS_I(struct xfs_inode *ip)
fs/xfs/xfs_inode.h
178
return &ip->i_vnode;
fs/xfs/xfs_inode.h
182
static inline const struct inode *VFS_IC(const struct xfs_inode *ip)
fs/xfs/xfs_inode.h
184
return &ip->i_vnode;
fs/xfs/xfs_inode.h
192
static inline xfs_fsize_t XFS_ISIZE(struct xfs_inode *ip)
fs/xfs/xfs_inode.h
194
if (S_ISREG(VFS_I(ip)->i_mode))
fs/xfs/xfs_inode.h
195
return i_size_read(VFS_I(ip));
fs/xfs/xfs_inode.h
196
return ip->i_disk_size;
fs/xfs/xfs_inode.h
204
xfs_new_eof(struct xfs_inode *ip, xfs_fsize_t new_size)
fs/xfs/xfs_inode.h
206
xfs_fsize_t i_size = i_size_read(VFS_I(ip));
fs/xfs/xfs_inode.h
210
return new_size > ip->i_disk_size ? new_size : 0;
fs/xfs/xfs_inode.h
217
__xfs_iflags_set(xfs_inode_t *ip, unsigned long flags)
fs/xfs/xfs_inode.h
219
ip->i_flags |= flags;
fs/xfs/xfs_inode.h
223
xfs_iflags_set(xfs_inode_t *ip, unsigned long flags)
fs/xfs/xfs_inode.h
225
spin_lock(&ip->i_flags_lock);
fs/xfs/xfs_inode.h
226
__xfs_iflags_set(ip, flags);
fs/xfs/xfs_inode.h
227
spin_unlock(&ip->i_flags_lock);
fs/xfs/xfs_inode.h
231
xfs_iflags_clear(xfs_inode_t *ip, unsigned long flags)
fs/xfs/xfs_inode.h
233
spin_lock(&ip->i_flags_lock);
fs/xfs/xfs_inode.h
234
ip->i_flags &= ~flags;
fs/xfs/xfs_inode.h
235
spin_unlock(&ip->i_flags_lock);
fs/xfs/xfs_inode.h
239
__xfs_iflags_test(const struct xfs_inode *ip, unsigned long flags)
fs/xfs/xfs_inode.h
241
return (ip->i_flags & flags);
fs/xfs/xfs_inode.h
245
xfs_iflags_test(xfs_inode_t *ip, unsigned long flags)
fs/xfs/xfs_inode.h
248
spin_lock(&ip->i_flags_lock);
fs/xfs/xfs_inode.h
249
ret = __xfs_iflags_test(ip, flags);
fs/xfs/xfs_inode.h
250
spin_unlock(&ip->i_flags_lock);
fs/xfs/xfs_inode.h
255
xfs_iflags_test_and_clear(xfs_inode_t *ip, unsigned long flags)
fs/xfs/xfs_inode.h
259
spin_lock(&ip->i_flags_lock);
fs/xfs/xfs_inode.h
260
ret = ip->i_flags & flags;
fs/xfs/xfs_inode.h
262
ip->i_flags &= ~flags;
fs/xfs/xfs_inode.h
263
spin_unlock(&ip->i_flags_lock);
fs/xfs/xfs_inode.h
268
xfs_iflags_test_and_set(xfs_inode_t *ip, unsigned long flags)
fs/xfs/xfs_inode.h
272
spin_lock(&ip->i_flags_lock);
fs/xfs/xfs_inode.h
273
ret = ip->i_flags & flags;
fs/xfs/xfs_inode.h
275
ip->i_flags |= flags;
fs/xfs/xfs_inode.h
276
spin_unlock(&ip->i_flags_lock);
fs/xfs/xfs_inode.h
280
static inline bool xfs_is_reflink_inode(const struct xfs_inode *ip)
fs/xfs/xfs_inode.h
282
return ip->i_diflags2 & XFS_DIFLAG2_REFLINK;
fs/xfs/xfs_inode.h
285
static inline bool xfs_is_metadir_inode(const struct xfs_inode *ip)
fs/xfs/xfs_inode.h
287
return ip->i_diflags2 & XFS_DIFLAG2_METADATA;
fs/xfs/xfs_inode.h
290
static inline bool xfs_is_internal_inode(const struct xfs_inode *ip)
fs/xfs/xfs_inode.h
292
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_inode.h
296
return xfs_is_metadir_inode(ip);
fs/xfs/xfs_inode.h
302
return ip->i_ino == mp->m_sb.sb_rbmino ||
fs/xfs/xfs_inode.h
303
ip->i_ino == mp->m_sb.sb_rsumino ||
fs/xfs/xfs_inode.h
304
xfs_is_quota_inode(&mp->m_sb, ip->i_ino);
fs/xfs/xfs_inode.h
307
static inline bool xfs_is_zoned_inode(const struct xfs_inode *ip)
fs/xfs/xfs_inode.h
309
return xfs_has_zoned(ip->i_mount) && XFS_IS_REALTIME_INODE(ip);
fs/xfs/xfs_inode.h
312
bool xfs_is_always_cow_inode(const struct xfs_inode *ip);
fs/xfs/xfs_inode.h
314
static inline bool xfs_is_cow_inode(const struct xfs_inode *ip)
fs/xfs/xfs_inode.h
316
return xfs_is_reflink_inode(ip) || xfs_is_always_cow_inode(ip);
fs/xfs/xfs_inode.h
319
static inline bool xfs_inode_has_filedata(const struct xfs_inode *ip)
fs/xfs/xfs_inode.h
321
return ip->i_df.if_nextents > 0 || ip->i_delayed_blks > 0;
fs/xfs/xfs_inode.h
328
static inline bool xfs_inode_has_cow_data(const struct xfs_inode *ip)
fs/xfs/xfs_inode.h
330
return ip->i_cowfp && ip->i_cowfp->if_bytes;
fs/xfs/xfs_inode.h
333
static inline bool xfs_inode_has_bigtime(const struct xfs_inode *ip)
fs/xfs/xfs_inode.h
335
return ip->i_diflags2 & XFS_DIFLAG2_BIGTIME;
fs/xfs/xfs_inode.h
338
static inline bool xfs_inode_has_large_extent_counts(const struct xfs_inode *ip)
fs/xfs/xfs_inode.h
340
return ip->i_diflags2 & XFS_DIFLAG2_NREXT64;
fs/xfs/xfs_inode.h
347
static inline bool xfs_inode_has_bigrtalloc(const struct xfs_inode *ip)
fs/xfs/xfs_inode.h
349
return XFS_IS_REALTIME_INODE(ip) && ip->i_mount->m_sb.sb_rextsize > 1;
fs/xfs/xfs_inode.h
355
#define xfs_inode_buftarg(ip) \
fs/xfs/xfs_inode.h
356
(XFS_IS_REALTIME_INODE(ip) ? \
fs/xfs/xfs_inode.h
357
(ip)->i_mount->m_rtdev_targp : (ip)->i_mount->m_ddev_targp)
fs/xfs/xfs_inode.h
359
static inline bool xfs_inode_can_hw_atomic_write(const struct xfs_inode *ip)
fs/xfs/xfs_inode.h
361
if (IS_DAX(VFS_IC(ip)))
fs/xfs/xfs_inode.h
364
return xfs_inode_buftarg(ip)->bt_awu_max > 0;
fs/xfs/xfs_inode.h
367
static inline bool xfs_inode_can_sw_atomic_write(const struct xfs_inode *ip)
fs/xfs/xfs_inode.h
369
if (IS_DAX(VFS_IC(ip)))
fs/xfs/xfs_inode.h
372
return xfs_can_sw_atomic_write(ip->i_mount);
fs/xfs/xfs_inode.h
556
int xfs_inactive(struct xfs_inode *ip);
fs/xfs/xfs_inode.h
564
struct xfs_inode *ip);
fs/xfs/xfs_inode.h
586
int xfs_log_force_inode(struct xfs_inode *ip);
fs/xfs/xfs_inode.h
588
#define xfs_ipincount(ip) ((unsigned int) atomic_read(&ip->i_pincount))
fs/xfs/xfs_inode.h
600
struct xfs_inode *ip,
fs/xfs/xfs_inode.h
604
return xfs_itruncate_extents_flags(tpp, ip, whichfork, new_size, 0);
fs/xfs/xfs_inode.h
611
static inline void xfs_update_stable_writes(struct xfs_inode *ip)
fs/xfs/xfs_inode.h
613
if (bdev_stable_writes(xfs_inode_buftarg(ip)->bt_bdev))
fs/xfs/xfs_inode.h
614
mapping_set_stable_writes(VFS_I(ip)->i_mapping);
fs/xfs/xfs_inode.h
616
mapping_clear_stable_writes(VFS_I(ip)->i_mapping);
fs/xfs/xfs_inode.h
626
static inline void xfs_finish_inode_setup(struct xfs_inode *ip)
fs/xfs/xfs_inode.h
628
xfs_iflags_clear(ip, XFS_INEW);
fs/xfs/xfs_inode.h
630
unlock_new_inode(VFS_I(ip));
fs/xfs/xfs_inode.h
633
static inline void xfs_setup_existing_inode(struct xfs_inode *ip)
fs/xfs/xfs_inode.h
635
xfs_setup_inode(ip);
fs/xfs/xfs_inode.h
636
xfs_setup_iops(ip);
fs/xfs/xfs_inode.h
637
xfs_finish_inode_setup(ip);
fs/xfs/xfs_inode.h
640
void xfs_irele(struct xfs_inode *ip);
fs/xfs/xfs_inode.h
647
bool xfs_inode_needs_inactive(struct xfs_inode *ip);
fs/xfs/xfs_inode.h
663
const struct xfs_inode *ip)
fs/xfs/xfs_inode.h
665
return VFS_IC(ip)->i_nlink == 0 && !xfs_inode_on_unlinked_list(ip);
fs/xfs/xfs_inode.h
667
int xfs_inode_reload_unlinked_bucket(struct xfs_trans *tp, struct xfs_inode *ip);
fs/xfs/xfs_inode.h
668
int xfs_inode_reload_unlinked(struct xfs_inode *ip);
fs/xfs/xfs_inode.h
670
bool xfs_ifork_zapped(const struct xfs_inode *ip, int whichfork);
fs/xfs/xfs_inode.h
671
void xfs_inode_count_blocks(struct xfs_trans *tp, struct xfs_inode *ip,
fs/xfs/xfs_inode.h
673
unsigned int xfs_inode_alloc_unitsize(struct xfs_inode *ip);
fs/xfs/xfs_inode_item.c
106
struct xfs_inode *ip = iip->ili_inode;
fs/xfs/xfs_inode_item.c
107
struct inode *inode = VFS_I(ip);
fs/xfs/xfs_inode_item.c
1114
struct xfs_inode *ip)
fs/xfs/xfs_inode_item.c
1116
struct xfs_inode_log_item *iip = ip->i_itemp;
fs/xfs/xfs_inode_item.c
1121
xfs_iflags_clear(ip, XFS_IFLUSHING);
fs/xfs/xfs_inode_item.c
1142
xfs_iflags_clear(ip, XFS_IFLUSHING);
fs/xfs/xfs_inode_item.c
1155
struct xfs_inode *ip)
fs/xfs/xfs_inode_item.c
1157
struct xfs_inode_log_item *iip = ip->i_itemp;
fs/xfs/xfs_inode_item.c
1162
xfs_iflags_clear(ip, XFS_IFLUSHING);
fs/xfs/xfs_inode_item.c
1170
xfs_iflush_abort(ip);
fs/xfs/xfs_inode_item.c
1195
xfs_iflags_clear(ip, XFS_IFLUSHING);
fs/xfs/xfs_inode_item.c
1209
xfs_iflush_abort(ip);
fs/xfs/xfs_inode_item.c
127
xfs_has_bigtime(ip->i_mount) &&
fs/xfs/xfs_inode_item.c
128
!xfs_inode_has_bigtime(ip)) {
fs/xfs/xfs_inode_item.c
129
ip->i_diflags2 |= XFS_DIFLAG2_BIGTIME;
fs/xfs/xfs_inode_item.c
139
if (ip->i_diflags & XFS_DIFLAG_RTINHERIT) {
fs/xfs/xfs_inode_item.c
140
if ((ip->i_diflags & XFS_DIFLAG_EXTSZINHERIT) &&
fs/xfs/xfs_inode_item.c
141
xfs_extlen_to_rtxmod(ip->i_mount, ip->i_extsize) > 0) {
fs/xfs/xfs_inode_item.c
142
ip->i_diflags &= ~(XFS_DIFLAG_EXTSIZE |
fs/xfs/xfs_inode_item.c
144
ip->i_extsize = 0;
fs/xfs/xfs_inode_item.c
147
if ((ip->i_diflags2 & XFS_DIFLAG2_COWEXTSIZE) &&
fs/xfs/xfs_inode_item.c
148
xfs_extlen_to_rtxmod(ip->i_mount, ip->i_cowextsize) > 0) {
fs/xfs/xfs_inode_item.c
149
ip->i_diflags2 &= ~XFS_DIFLAG2_COWEXTSIZE;
fs/xfs/xfs_inode_item.c
150
ip->i_cowextsize = 0;
fs/xfs/xfs_inode_item.c
170
error = xfs_imap_to_bp(ip->i_mount, tp, &ip->i_imap, &bp);
fs/xfs/xfs_inode_item.c
212
xfs_inode_item_precommit_check(ip);
fs/xfs/xfs_inode_item.c
236
struct xfs_inode *ip = iip->ili_inode;
fs/xfs/xfs_inode_item.c
238
switch (ip->i_df.if_format) {
fs/xfs/xfs_inode_item.c
241
ip->i_df.if_nextents > 0 &&
fs/xfs/xfs_inode_item.c
242
ip->i_df.if_bytes > 0) {
fs/xfs/xfs_inode_item.c
244
*nbytes += xfs_inode_data_fork_size(ip);
fs/xfs/xfs_inode_item.c
251
ip->i_df.if_broot_bytes > 0) {
fs/xfs/xfs_inode_item.c
252
*nbytes += ip->i_df.if_broot_bytes;
fs/xfs/xfs_inode_item.c
258
ip->i_df.if_bytes > 0) {
fs/xfs/xfs_inode_item.c
259
*nbytes += xlog_calc_iovec_len(ip->i_df.if_bytes);
fs/xfs/xfs_inode_item.c
278
struct xfs_inode *ip = iip->ili_inode;
fs/xfs/xfs_inode_item.c
280
switch (ip->i_af.if_format) {
fs/xfs/xfs_inode_item.c
283
ip->i_af.if_nextents > 0 &&
fs/xfs/xfs_inode_item.c
284
ip->i_af.if_bytes > 0) {
fs/xfs/xfs_inode_item.c
286
*nbytes += xfs_inode_attr_fork_size(ip);
fs/xfs/xfs_inode_item.c
292
ip->i_af.if_broot_bytes > 0) {
fs/xfs/xfs_inode_item.c
293
*nbytes += ip->i_af.if_broot_bytes;
fs/xfs/xfs_inode_item.c
299
ip->i_af.if_bytes > 0) {
fs/xfs/xfs_inode_item.c
300
*nbytes += xlog_calc_iovec_len(ip->i_af.if_bytes);
fs/xfs/xfs_inode_item.c
324
struct xfs_inode *ip = iip->ili_inode;
fs/xfs/xfs_inode_item.c
328
xfs_log_dinode_size(ip->i_mount);
fs/xfs/xfs_inode_item.c
331
if (xfs_inode_has_attr_fork(ip))
fs/xfs/xfs_inode_item.c
341
struct xfs_inode *ip = iip->ili_inode;
fs/xfs/xfs_inode_item.c
344
switch (ip->i_df.if_format) {
fs/xfs/xfs_inode_item.c
350
ip->i_df.if_nextents > 0 &&
fs/xfs/xfs_inode_item.c
351
ip->i_df.if_bytes > 0) {
fs/xfs/xfs_inode_item.c
354
ASSERT(xfs_iext_count(&ip->i_df) > 0);
fs/xfs/xfs_inode_item.c
357
data_bytes = xfs_iextents_copy(ip, p, XFS_DATA_FORK);
fs/xfs/xfs_inode_item.c
360
ASSERT(data_bytes <= ip->i_df.if_bytes);
fs/xfs/xfs_inode_item.c
374
ip->i_df.if_broot_bytes > 0) {
fs/xfs/xfs_inode_item.c
375
ASSERT(ip->i_df.if_broot != NULL);
fs/xfs/xfs_inode_item.c
377
ip->i_df.if_broot,
fs/xfs/xfs_inode_item.c
378
ip->i_df.if_broot_bytes);
fs/xfs/xfs_inode_item.c
379
ilf->ilf_dsize = ip->i_df.if_broot_bytes;
fs/xfs/xfs_inode_item.c
391
ip->i_df.if_bytes > 0) {
fs/xfs/xfs_inode_item.c
392
ASSERT(ip->i_df.if_data != NULL);
fs/xfs/xfs_inode_item.c
393
ASSERT(ip->i_disk_size > 0);
fs/xfs/xfs_inode_item.c
395
ip->i_df.if_data,
fs/xfs/xfs_inode_item.c
396
ip->i_df.if_bytes);
fs/xfs/xfs_inode_item.c
397
ilf->ilf_dsize = (unsigned)ip->i_df.if_bytes;
fs/xfs/xfs_inode_item.c
407
ilf->ilf_u.ilfu_rdev = sysv_encode_dev(VFS_I(ip)->i_rdev);
fs/xfs/xfs_inode_item.c
421
struct xfs_inode *ip = iip->ili_inode;
fs/xfs/xfs_inode_item.c
424
switch (ip->i_af.if_format) {
fs/xfs/xfs_inode_item.c
43
struct xfs_inode *ip)
fs/xfs/xfs_inode_item.c
430
ip->i_af.if_nextents > 0 &&
fs/xfs/xfs_inode_item.c
431
ip->i_af.if_bytes > 0) {
fs/xfs/xfs_inode_item.c
434
ASSERT(xfs_iext_count(&ip->i_af) ==
fs/xfs/xfs_inode_item.c
435
ip->i_af.if_nextents);
fs/xfs/xfs_inode_item.c
438
data_bytes = xfs_iextents_copy(ip, p, XFS_ATTR_FORK);
fs/xfs/xfs_inode_item.c
45
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_inode_item.c
452
ip->i_af.if_broot_bytes > 0) {
fs/xfs/xfs_inode_item.c
453
ASSERT(ip->i_af.if_broot != NULL);
fs/xfs/xfs_inode_item.c
456
ip->i_af.if_broot,
fs/xfs/xfs_inode_item.c
457
ip->i_af.if_broot_bytes);
fs/xfs/xfs_inode_item.c
458
ilf->ilf_asize = ip->i_af.if_broot_bytes;
fs/xfs/xfs_inode_item.c
469
ip->i_af.if_bytes > 0) {
fs/xfs/xfs_inode_item.c
470
ASSERT(ip->i_af.if_data != NULL);
fs/xfs/xfs_inode_item.c
472
ip->i_af.if_data,
fs/xfs/xfs_inode_item.c
473
ip->i_af.if_bytes);
fs/xfs/xfs_inode_item.c
474
ilf->ilf_asize = (unsigned)ip->i_af.if_bytes;
fs/xfs/xfs_inode_item.c
492
struct xfs_inode *ip,
fs/xfs/xfs_inode_item.c
498
if (xfs_inode_has_bigtime(ip))
fs/xfs/xfs_inode_item.c
518
struct xfs_inode *ip,
fs/xfs/xfs_inode_item.c
523
dip = xfs_buf_offset(ip->i_itemp->ili_item.li_buf,
fs/xfs/xfs_inode_item.c
524
ip->i_imap.im_boffset);
fs/xfs/xfs_inode_item.c
526
if (xfs_iflags_test(ip, XFS_IPRESERVE_DM_FIELDS)) {
fs/xfs/xfs_inode_item.c
537
struct xfs_inode *ip,
fs/xfs/xfs_inode_item.c
540
if (xfs_inode_has_large_extent_counts(ip)) {
fs/xfs/xfs_inode_item.c
541
to->di_big_nextents = xfs_ifork_nextents(&ip->i_df);
fs/xfs/xfs_inode_item.c
542
to->di_big_anextents = xfs_ifork_nextents(&ip->i_af);
fs/xfs/xfs_inode_item.c
545
to->di_nextents = xfs_ifork_nextents(&ip->i_df);
fs/xfs/xfs_inode_item.c
546
to->di_anextents = xfs_ifork_nextents(&ip->i_af);
fs/xfs/xfs_inode_item.c
55
xfs_inode_to_disk(ip, dip, 0);
fs/xfs/xfs_inode_item.c
552
struct xfs_inode *ip,
fs/xfs/xfs_inode_item.c
556
struct inode *inode = VFS_I(ip);
fs/xfs/xfs_inode_item.c
559
to->di_format = xfs_ifork_format(&ip->i_df);
fs/xfs/xfs_inode_item.c
562
to->di_projid_lo = ip->i_projid & 0xffff;
fs/xfs/xfs_inode_item.c
563
to->di_projid_hi = ip->i_projid >> 16;
fs/xfs/xfs_inode_item.c
565
to->di_atime = xfs_inode_to_log_dinode_ts(ip, inode_get_atime(inode));
fs/xfs/xfs_inode_item.c
566
to->di_mtime = xfs_inode_to_log_dinode_ts(ip, inode_get_mtime(inode));
fs/xfs/xfs_inode_item.c
567
to->di_ctime = xfs_inode_to_log_dinode_ts(ip, inode_get_ctime(inode));
fs/xfs/xfs_inode_item.c
57
fa = xfs_dinode_verify(mp, ip->i_ino, dip);
fs/xfs/xfs_inode_item.c
572
to->di_size = ip->i_disk_size;
fs/xfs/xfs_inode_item.c
573
to->di_nblocks = ip->i_nblocks;
fs/xfs/xfs_inode_item.c
574
to->di_extsize = ip->i_extsize;
fs/xfs/xfs_inode_item.c
575
to->di_forkoff = ip->i_forkoff;
fs/xfs/xfs_inode_item.c
576
to->di_aformat = xfs_ifork_format(&ip->i_af);
fs/xfs/xfs_inode_item.c
577
to->di_flags = ip->i_diflags;
fs/xfs/xfs_inode_item.c
579
xfs_copy_dm_fields_to_log_dinode(ip, to);
fs/xfs/xfs_inode_item.c
584
if (xfs_has_v3inodes(ip->i_mount)) {
fs/xfs/xfs_inode_item.c
587
to->di_crtime = xfs_inode_to_log_dinode_ts(ip, ip->i_crtime);
fs/xfs/xfs_inode_item.c
588
to->di_flags2 = ip->i_diflags2;
fs/xfs/xfs_inode_item.c
59
xfs_inode_verifier_error(ip, -EFSCORRUPTED, __func__, dip,
fs/xfs/xfs_inode_item.c
590
to->di_cowextsize = ip->i_cowextsize;
fs/xfs/xfs_inode_item.c
591
to->di_ino = ip->i_ino;
fs/xfs/xfs_inode_item.c
594
uuid_copy(&to->di_uuid, &ip->i_mount->m_sb.sb_meta_uuid);
fs/xfs/xfs_inode_item.c
600
if (xfs_is_metadir_inode(ip))
fs/xfs/xfs_inode_item.c
601
to->di_metatype = ip->i_metatype;
fs/xfs/xfs_inode_item.c
606
to->di_flushiter = ip->i_flushiter;
fs/xfs/xfs_inode_item.c
611
xfs_inode_to_log_dinode_iext_counters(ip, to);
fs/xfs/xfs_inode_item.c
621
struct xfs_inode *ip,
fs/xfs/xfs_inode_item.c
627
xfs_inode_to_log_dinode(ip, dic, ip->i_itemp->ili_item.li_lsn);
fs/xfs/xfs_inode_item.c
628
xlog_format_commit(lfb, xfs_log_dinode_size(ip->i_mount));
fs/xfs/xfs_inode_item.c
649
struct xfs_inode *ip = iip->ili_inode;
fs/xfs/xfs_inode_item.c
654
ilf->ilf_ino = ip->i_ino;
fs/xfs/xfs_inode_item.c
655
ilf->ilf_blkno = ip->i_imap.im_blkno;
fs/xfs/xfs_inode_item.c
656
ilf->ilf_len = ip->i_imap.im_len;
fs/xfs/xfs_inode_item.c
657
ilf->ilf_boffset = ip->i_imap.im_boffset;
fs/xfs/xfs_inode_item.c
67
# define xfs_inode_item_precommit_check(ip) ((void)0)
fs/xfs/xfs_inode_item.c
671
xfs_inode_item_format_core(ip, lfb);
fs/xfs/xfs_inode_item.c
673
if (xfs_inode_has_attr_fork(ip)) {
fs/xfs/xfs_inode_item.c
692
struct xfs_inode *ip = INODE_ITEM(lip)->ili_inode;
fs/xfs/xfs_inode_item.c
694
xfs_assert_ilocked(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_inode_item.c
697
trace_xfs_inode_pin(ip, _RET_IP_);
fs/xfs/xfs_inode_item.c
698
atomic_inc(&ip->i_pincount);
fs/xfs/xfs_inode_item.c
720
struct xfs_inode *ip = iip->ili_inode;
fs/xfs/xfs_inode_item.c
722
trace_xfs_inode_unpin(ip, _RET_IP_);
fs/xfs/xfs_inode_item.c
723
ASSERT(lip->li_buf || xfs_iflags_test(ip, XFS_ISTALE));
fs/xfs/xfs_inode_item.c
724
ASSERT(atomic_read(&ip->i_pincount) > 0);
fs/xfs/xfs_inode_item.c
731
if (atomic_dec_and_lock(&ip->i_pincount, &iip->ili_lock)) {
fs/xfs/xfs_inode_item.c
735
wake_up_bit(&ip->i_flags, __XFS_IPINNED_BIT);
fs/xfs/xfs_inode_item.c
747
struct xfs_inode *ip = iip->ili_inode;
fs/xfs/xfs_inode_item.c
753
if (!bp || (ip->i_flags & XFS_ISTALE)) {
fs/xfs/xfs_inode_item.c
760
trace_xfs_inode_push_stale(ip, _RET_IP_);
fs/xfs/xfs_inode_item.c
764
if (xfs_ipincount(ip) > 0 || xfs_buf_ispinned(bp)) {
fs/xfs/xfs_inode_item.c
765
trace_xfs_inode_push_pinned(ip, _RET_IP_);
fs/xfs/xfs_inode_item.c
769
if (xfs_iflags_test(ip, XFS_IFLUSHING))
fs/xfs/xfs_inode_item.c
815
struct xfs_inode *ip = iip->ili_inode;
fs/xfs/xfs_inode_item.c
818
ASSERT(ip->i_itemp != NULL);
fs/xfs/xfs_inode_item.c
819
xfs_assert_ilocked(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_inode_item.c
824
xfs_iunlock(ip, lock_flags);
fs/xfs/xfs_inode_item.c
855
struct xfs_inode *ip = iip->ili_inode;
fs/xfs/xfs_inode_item.c
857
if (xfs_iflags_test(ip, XFS_ISTALE)) {
fs/xfs/xfs_inode_item.c
925
struct xfs_inode *ip,
fs/xfs/xfs_inode_item.c
930
ASSERT(ip->i_itemp == NULL);
fs/xfs/xfs_inode_item.c
931
iip = ip->i_itemp = kmem_cache_zalloc(xfs_ili_cache,
fs/xfs/xfs_inode_item.c
934
iip->ili_inode = ip;
fs/xfs/xfs_inode_item.c
945
struct xfs_inode *ip)
fs/xfs/xfs_inode_item.c
947
struct xfs_inode_log_item *iip = ip->i_itemp;
fs/xfs/xfs_inode_item.c
951
ip->i_itemp = NULL;
fs/xfs/xfs_inode_item.h
48
static inline int xfs_inode_clean(struct xfs_inode *ip)
fs/xfs/xfs_inode_item.h
50
return !ip->i_itemp || !(ip->i_itemp->ili_fields & XFS_ILOG_ALL);
fs/xfs/xfs_inode_item_recover.c
101
error = xfs_bmbt_change_owner(NULL, ip, XFS_DATA_FORK,
fs/xfs/xfs_inode_item_recover.c
102
ip->i_ino, buffer_list);
fs/xfs/xfs_inode_item_recover.c
109
error = xfs_bmbt_change_owner(NULL, ip, XFS_ATTR_FORK,
fs/xfs/xfs_inode_item_recover.c
110
ip->i_ino, buffer_list);
fs/xfs/xfs_inode_item_recover.c
116
xfs_inode_free(ip);
fs/xfs/xfs_inode_item_recover.c
83
struct xfs_inode *ip;
fs/xfs/xfs_inode_item_recover.c
88
ip = xfs_inode_alloc(mp, in_f->ilf_ino);
fs/xfs/xfs_inode_item_recover.c
89
if (!ip)
fs/xfs/xfs_inode_item_recover.c
95
error = xfs_inode_from_disk(ip, dip);
fs/xfs/xfs_ioctl.c
1185
struct xfs_inode *ip = XFS_I(inode);
fs/xfs/xfs_ioctl.c
1186
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_ioctl.c
1190
trace_xfs_file_ioctl(ip);
fs/xfs/xfs_ioctl.c
1254
return xfs_ioc_fsgetxattra(ip, arg);
fs/xfs/xfs_ioctl.c
1265
return xfs_ioc_getfsmap(ip, arg);
fs/xfs/xfs_ioctl.c
447
struct xfs_inode *ip,
fs/xfs/xfs_ioctl.c
451
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_ioctl.c
452
struct xfs_ifork *ifp = xfs_ifork_ptr(ip, whichfork);
fs/xfs/xfs_ioctl.c
454
fileattr_fill_xflags(fa, xfs_ip2xflags(ip));
fs/xfs/xfs_ioctl.c
456
if (ip->i_diflags & XFS_DIFLAG_EXTSIZE) {
fs/xfs/xfs_ioctl.c
457
fa->fsx_extsize = XFS_FSB_TO_B(mp, ip->i_extsize);
fs/xfs/xfs_ioctl.c
458
} else if (ip->i_diflags & XFS_DIFLAG_EXTSZINHERIT) {
fs/xfs/xfs_ioctl.c
464
if ((ip->i_diflags & XFS_DIFLAG_RTINHERIT) &&
fs/xfs/xfs_ioctl.c
465
xfs_extlen_to_rtxmod(mp, ip->i_extsize) > 0) {
fs/xfs/xfs_ioctl.c
470
fa->fsx_extsize = XFS_FSB_TO_B(mp, ip->i_extsize);
fs/xfs/xfs_ioctl.c
474
if (ip->i_diflags2 & XFS_DIFLAG2_COWEXTSIZE) {
fs/xfs/xfs_ioctl.c
480
if ((ip->i_diflags & XFS_DIFLAG_RTINHERIT) &&
fs/xfs/xfs_ioctl.c
481
ip->i_cowextsize % mp->m_sb.sb_rextsize > 0) {
fs/xfs/xfs_ioctl.c
485
fa->fsx_cowextsize = XFS_FSB_TO_B(mp, ip->i_cowextsize);
fs/xfs/xfs_ioctl.c
489
fa->fsx_projid = ip->i_projid;
fs/xfs/xfs_ioctl.c
498
xfs_inode_t *ip,
fs/xfs/xfs_ioctl.c
503
xfs_ilock(ip, XFS_ILOCK_SHARED);
fs/xfs/xfs_ioctl.c
504
xfs_fill_fsxattr(ip, XFS_ATTR_FORK, &fa);
fs/xfs/xfs_ioctl.c
505
xfs_iunlock(ip, XFS_ILOCK_SHARED);
fs/xfs/xfs_ioctl.c
515
struct xfs_inode *ip = XFS_I(d_inode(dentry));
fs/xfs/xfs_ioctl.c
517
xfs_ilock(ip, XFS_ILOCK_SHARED);
fs/xfs/xfs_ioctl.c
518
xfs_fill_fsxattr(ip, XFS_DATA_FORK, fa);
fs/xfs/xfs_ioctl.c
519
xfs_iunlock(ip, XFS_ILOCK_SHARED);
fs/xfs/xfs_ioctl.c
527
struct xfs_inode *ip,
fs/xfs/xfs_ioctl.c
530
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_ioctl.c
534
if (rtflag != XFS_IS_REALTIME_INODE(ip)) {
fs/xfs/xfs_ioctl.c
536
if (xfs_inode_has_filedata(ip))
fs/xfs/xfs_ioctl.c
544
if (IS_DAX(VFS_I(ip)) &&
fs/xfs/xfs_ioctl.c
554
xfs_extlen_to_rtxmod(mp, ip->i_extsize))
fs/xfs/xfs_ioctl.c
559
i_flags2 = xfs_flags2diflags2(ip, fa->fsx_xflags);
fs/xfs/xfs_ioctl.c
563
ip->i_diflags = xfs_flags2diflags(ip, fa->fsx_xflags);
fs/xfs/xfs_ioctl.c
564
ip->i_diflags2 = i_flags2;
fs/xfs/xfs_ioctl.c
566
xfs_diflags_to_iflags(ip, false);
fs/xfs/xfs_ioctl.c
572
if (rtflag != XFS_IS_REALTIME_INODE(ip) && S_ISREG(VFS_I(ip)->i_mode))
fs/xfs/xfs_ioctl.c
573
xfs_update_stable_writes(ip);
fs/xfs/xfs_ioctl.c
575
xfs_trans_ichgtime(tp, ip, XFS_ICHGTIME_CHG);
fs/xfs/xfs_ioctl.c
576
xfs_trans_log_inode(tp, ip, XFS_ILOG_CORE);
fs/xfs/xfs_ioctl.c
583
struct xfs_inode *ip,
fs/xfs/xfs_ioctl.c
586
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_ioctl.c
587
struct inode *inode = VFS_I(ip);
fs/xfs/xfs_ioctl.c
596
!(ip->i_diflags2 & XFS_DIFLAG2_DAX)) ||
fs/xfs/xfs_ioctl.c
598
(ip->i_diflags2 & XFS_DIFLAG2_DAX)))
fs/xfs/xfs_ioctl.c
610
struct xfs_inode *ip,
fs/xfs/xfs_ioctl.c
613
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_ioctl.c
623
error = xfs_trans_alloc_ichange(ip, NULL, NULL, pdqp,
fs/xfs/xfs_ioctl.c
643
struct xfs_inode *ip,
fs/xfs/xfs_ioctl.c
646
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_ioctl.c
653
if (S_ISREG(VFS_I(ip)->i_mode) && xfs_inode_has_filedata(ip) &&
fs/xfs/xfs_ioctl.c
654
XFS_FSB_TO_B(mp, ip->i_extsize) != fa->fsx_extsize)
fs/xfs/xfs_ioctl.c
660
new_diflags = xfs_flags2diflags(ip, fa->fsx_xflags);
fs/xfs/xfs_ioctl.c
677
failaddr = xfs_inode_validate_extsize(ip->i_mount,
fs/xfs/xfs_ioctl.c
679
VFS_I(ip)->i_mode, new_diflags);
fs/xfs/xfs_ioctl.c
685
struct xfs_inode *ip,
fs/xfs/xfs_ioctl.c
688
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_ioctl.c
699
new_diflags = xfs_flags2diflags(ip, fa->fsx_xflags);
fs/xfs/xfs_ioctl.c
700
new_diflags2 = xfs_flags2diflags2(ip, fa->fsx_xflags);
fs/xfs/xfs_ioctl.c
702
failaddr = xfs_inode_validate_cowextsize(ip->i_mount,
fs/xfs/xfs_ioctl.c
704
VFS_I(ip)->i_mode, new_diflags, new_diflags2);
fs/xfs/xfs_ioctl.c
710
struct xfs_inode *ip,
fs/xfs/xfs_ioctl.c
718
!xfs_has_projid32(ip->i_mount))
fs/xfs/xfs_ioctl.c
729
struct xfs_inode *ip = XFS_I(d_inode(dentry));
fs/xfs/xfs_ioctl.c
730
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_ioctl.c
736
trace_xfs_ioctl_setattr(ip);
fs/xfs/xfs_ioctl.c
745
error = xfs_ioctl_setattr_check_projid(ip, fa);
fs/xfs/xfs_ioctl.c
758
error = xfs_qm_vop_dqalloc(ip, VFS_I(ip)->i_uid,
fs/xfs/xfs_ioctl.c
759
VFS_I(ip)->i_gid, fa->fsx_projid,
fs/xfs/xfs_ioctl.c
765
xfs_ioctl_setattr_prepare_dax(ip, fa);
fs/xfs/xfs_ioctl.c
767
tp = xfs_ioctl_setattr_get_trans(ip, pdqp);
fs/xfs/xfs_ioctl.c
773
error = xfs_ioctl_setattr_check_extsize(ip, fa);
fs/xfs/xfs_ioctl.c
777
error = xfs_ioctl_setattr_check_cowextsize(ip, fa);
fs/xfs/xfs_ioctl.c
781
error = xfs_ioctl_setattr_xflags(tp, ip, fa);
fs/xfs/xfs_ioctl.c
795
if ((VFS_I(ip)->i_mode & (S_ISUID|S_ISGID)) &&
fs/xfs/xfs_ioctl.c
796
!capable_wrt_inode_uidgid(idmap, VFS_I(ip), CAP_FSETID))
fs/xfs/xfs_ioctl.c
797
VFS_I(ip)->i_mode &= ~(S_ISUID|S_ISGID);
fs/xfs/xfs_ioctl.c
800
if (ip->i_projid != fa->fsx_projid) {
fs/xfs/xfs_ioctl.c
802
olddquot = xfs_qm_vop_chown(tp, ip,
fs/xfs/xfs_ioctl.c
803
&ip->i_pdquot, pdqp);
fs/xfs/xfs_ioctl.c
805
ip->i_projid = fa->fsx_projid;
fs/xfs/xfs_ioctl.c
813
if (ip->i_diflags & (XFS_DIFLAG_EXTSIZE | XFS_DIFLAG_EXTSZINHERIT))
fs/xfs/xfs_ioctl.c
814
ip->i_extsize = XFS_B_TO_FSB(mp, fa->fsx_extsize);
fs/xfs/xfs_ioctl.c
816
ip->i_extsize = 0;
fs/xfs/xfs_ioctl.c
819
if (ip->i_diflags2 & XFS_DIFLAG2_COWEXTSIZE)
fs/xfs/xfs_ioctl.c
820
ip->i_cowextsize = XFS_B_TO_FSB(mp, fa->fsx_cowextsize);
fs/xfs/xfs_ioctl.c
822
ip->i_cowextsize = 0;
fs/xfs/xfs_ioctl.c
928
xfs_inode_t *ip, *tip;
fs/xfs/xfs_ioctl.c
962
ip = XFS_I(file_inode(fd_file(f)));
fs/xfs/xfs_ioctl.c
965
if (ip->i_mount != tip->i_mount)
fs/xfs/xfs_ioctl.c
968
if (ip->i_ino == tip->i_ino)
fs/xfs/xfs_ioctl.c
971
if (xfs_is_shutdown(ip->i_mount))
fs/xfs/xfs_ioctl.c
974
return xfs_swap_extents(ip, tip, sxp);
fs/xfs/xfs_ioctl32.c
427
struct xfs_inode *ip = XFS_I(inode);
fs/xfs/xfs_ioctl32.c
431
trace_xfs_file_compat_ioctl(ip);
fs/xfs/xfs_ioctl32.c
436
return xfs_compat_ioc_fsgeometry_v1(ip->i_mount, arg);
fs/xfs/xfs_ioctl32.c
445
error = xfs_growfs_data(ip->i_mount, &in);
fs/xfs/xfs_ioctl32.c
457
error = xfs_growfs_rt(ip->i_mount, &in);
fs/xfs/xfs_iomap.c
1003
if (offset + length > XFS_ISIZE(ip))
fs/xfs/xfs_iomap.c
1004
end_fsb = xfs_iomap_eof_align_last_fsb(ip, end_fsb);
fs/xfs/xfs_iomap.c
1007
xfs_iunlock(ip, lockmode);
fs/xfs/xfs_iomap.c
1009
error = xfs_iomap_write_direct(ip, offset_fsb, end_fsb - offset_fsb,
fs/xfs/xfs_iomap.c
101
struct xfs_inode *ip,
fs/xfs/xfs_iomap.c
1014
trace_xfs_iomap_alloc(ip, offset, length, XFS_DATA_FORK, &imap);
fs/xfs/xfs_iomap.c
1015
return xfs_bmbt_to_iomap(ip, iomap, &imap, flags,
fs/xfs/xfs_iomap.c
1020
trace_xfs_iomap_found(ip, offset, length - offset, XFS_COW_FORK, &cmap);
fs/xfs/xfs_iomap.c
1022
seq = xfs_iomap_inode_sequence(ip, 0);
fs/xfs/xfs_iomap.c
1023
error = xfs_bmbt_to_iomap(ip, srcmap, &imap, flags, 0, seq);
fs/xfs/xfs_iomap.c
1027
seq = xfs_iomap_inode_sequence(ip, IOMAP_F_SHARED);
fs/xfs/xfs_iomap.c
1028
xfs_iunlock(ip, lockmode);
fs/xfs/xfs_iomap.c
1029
return xfs_bmbt_to_iomap(ip, iomap, &cmap, flags, IOMAP_F_SHARED, seq);
fs/xfs/xfs_iomap.c
1033
xfs_iunlock(ip, lockmode);
fs/xfs/xfs_iomap.c
1056
struct xfs_inode *ip = XFS_I(inode);
fs/xfs/xfs_iomap.c
1072
if (xfs_need_iread_extents(&ip->i_df)) {
fs/xfs/xfs_iomap.c
1073
xfs_ilock(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_iomap.c
1074
error = xfs_iread_extents(NULL, ip, XFS_DATA_FORK);
fs/xfs/xfs_iomap.c
1075
xfs_iunlock(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_iomap.c
108
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_iomap.c
1082
iomap->bdev = ip->i_mount->m_rtdev_targp->bt_bdev;
fs/xfs/xfs_iomap.c
109
struct xfs_buftarg *target = xfs_inode_buftarg(ip);
fs/xfs/xfs_iomap.c
1097
struct xfs_inode *ip,
fs/xfs/xfs_iomap.c
1105
if (xfs_iext_lookup_extent(ip, ip->i_cowfp, offset_fsb, &icur, &cmap2))
fs/xfs/xfs_iomap.c
111
if (unlikely(!xfs_valid_startblock(ip, imap->br_startblock))) {
fs/xfs/xfs_iomap.c
112
xfs_bmap_mark_sick(ip, XFS_DATA_FORK);
fs/xfs/xfs_iomap.c
1126
struct xfs_inode *ip = XFS_I(inode);
fs/xfs/xfs_iomap.c
1127
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_iomap.c
113
return xfs_alert_fsblock_zero(ip, imap);
fs/xfs/xfs_iomap.c
1156
trace_xfs_iomap_atomic_write_cow(ip, offset, length);
fs/xfs/xfs_iomap.c
1158
xfs_ilock(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_iomap.c
1160
if (!ip->i_cowfp) {
fs/xfs/xfs_iomap.c
1161
ASSERT(!xfs_is_reflink_inode(ip));
fs/xfs/xfs_iomap.c
1162
xfs_ifork_init_cow(ip);
fs/xfs/xfs_iomap.c
1165
if (!xfs_iext_lookup_extent(ip, ip->i_cowfp, offset_fsb, &icur, &cmap))
fs/xfs/xfs_iomap.c
1184
xfs_get_cowextsz_hint(ip));
fs/xfs/xfs_iomap.c
1185
xfs_iunlock(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_iomap.c
1187
if (XFS_IS_REALTIME_INODE(ip)) {
fs/xfs/xfs_iomap.c
1195
error = xfs_trans_alloc_inode(ip, &M_RES(mp)->tr_write, dblocks,
fs/xfs/xfs_iomap.c
1201
if (!xfs_iext_lookup_extent(ip, ip->i_cowfp, offset_fsb, &icur, &cmap))
fs/xfs/xfs_iomap.c
1219
error = xfs_bmapi_write(tp, ip, offset_fsb, hole_count_fsb,
fs/xfs/xfs_iomap.c
1227
xfs_inode_set_cowblocks_tag(ip);
fs/xfs/xfs_iomap.c
124
xfs_daddr_t daddr = xfs_fsb_to_db(ip, imap->br_startblock);
fs/xfs/xfs_iomap.c
1242
error = xfs_reflink_convert_cow_locked(ip, cmap.br_startoff,
fs/xfs/xfs_iomap.c
1247
xfs_check_atomic_cow_conversion(ip, offset_fsb, count_fsb,
fs/xfs/xfs_iomap.c
1251
trace_xfs_iomap_found(ip, offset, length, XFS_COW_FORK, &cmap);
fs/xfs/xfs_iomap.c
1252
seq = xfs_iomap_inode_sequence(ip, IOMAP_F_SHARED);
fs/xfs/xfs_iomap.c
1253
xfs_iunlock(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_iomap.c
1254
return xfs_bmbt_to_iomap(ip, iomap, &cmap, flags, IOMAP_F_SHARED, seq);
fs/xfs/xfs_iomap.c
1257
xfs_iunlock(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_iomap.c
1258
error = xfs_bmapi_convert_delalloc(ip, XFS_COW_FORK, offset, iomap,
fs/xfs/xfs_iomap.c
1270
xfs_iunlock(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_iomap.c
1287
struct xfs_inode *ip = XFS_I(inode);
fs/xfs/xfs_iomap.c
1289
if (!xfs_is_cow_inode(ip))
fs/xfs/xfs_iomap.c
1293
return xfs_reflink_cancel_cow_range(ip, pos, length, true);
fs/xfs/xfs_iomap.c
1295
return xfs_reflink_end_cow(ip, pos, written);
fs/xfs/xfs_iomap.c
1308
struct xfs_inode *ip, /* incore inode pointer */
fs/xfs/xfs_iomap.c
1321
ifp = xfs_ifork_ptr(ip, whichfork);
fs/xfs/xfs_iomap.c
1376
newlen = XFS_FILBLKS_MIN(xfs_bmap_worst_indlen(ip, temp),
fs/xfs/xfs_iomap.c
1381
xfs_iext_remove(ip, icur, state);
fs/xfs/xfs_iomap.c
1383
xfs_iext_update_extent(ip, state, icur, &left);
fs/xfs/xfs_iomap.c
1396
newlen = XFS_FILBLKS_MIN(xfs_bmap_worst_indlen(ip, temp),
fs/xfs/xfs_iomap.c
140
if (XFS_IS_REALTIME_INODE(ip) && xfs_has_rtgroups(mp) &&
fs/xfs/xfs_iomap.c
1402
xfs_iext_update_extent(ip, state, icur, &left);
fs/xfs/xfs_iomap.c
1414
newlen = XFS_FILBLKS_MIN(xfs_bmap_worst_indlen(ip, temp),
fs/xfs/xfs_iomap.c
1419
xfs_iext_update_extent(ip, state, icur, &right);
fs/xfs/xfs_iomap.c
1429
xfs_iext_insert(ip, icur, new, state);
fs/xfs/xfs_iomap.c
1434
xfs_add_fdblocks(ip->i_mount, oldlen - newlen);
fs/xfs/xfs_iomap.c
1439
xfs_mod_delalloc(ip, 0, (int64_t)newlen - oldlen);
fs/xfs/xfs_iomap.c
1458
struct xfs_inode *ip,
fs/xfs/xfs_iomap.c
1467
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_iomap.c
1468
struct xfs_ifork *ifp = xfs_ifork_ptr(ip, whichfork);
fs/xfs/xfs_iomap.c
1500
xfs_extlen_t extsz = xfs_get_cowextsz_hint(ip);
fs/xfs/xfs_iomap.c
1515
error = xfs_quota_reserve_blkres(ip, alen);
fs/xfs/xfs_iomap.c
1523
indlen = (xfs_extlen_t)xfs_bmap_worst_indlen(ip, alen);
fs/xfs/xfs_iomap.c
1527
if (XFS_IS_REALTIME_INODE(ip)) {
fs/xfs/xfs_iomap.c
1528
ASSERT(!xfs_is_zoned_inode(ip));
fs/xfs/xfs_iomap.c
1540
ip->i_delayed_blks += alen;
fs/xfs/xfs_iomap.c
1541
xfs_mod_delalloc(ip, alen, indlen);
fs/xfs/xfs_iomap.c
1548
xfs_bmap_add_extent_hole_delay(ip, whichfork, icur, got);
fs/xfs/xfs_iomap.c
1556
xfs_inode_set_eofblocks_tag(ip);
fs/xfs/xfs_iomap.c
1558
xfs_inode_set_cowblocks_tag(ip);
fs/xfs/xfs_iomap.c
156
if (ip->i_itemp) {
fs/xfs/xfs_iomap.c
1563
if (XFS_IS_REALTIME_INODE(ip))
fs/xfs/xfs_iomap.c
1567
xfs_quota_unreserve_blkres(ip, alen);
fs/xfs/xfs_iomap.c
157
struct xfs_inode_log_item *iip = ip->i_itemp;
fs/xfs/xfs_iomap.c
1570
trace_xfs_delalloc_enospc(ip, off, len);
fs/xfs/xfs_iomap.c
1594
struct xfs_inode *ip = XFS_I(inode);
fs/xfs/xfs_iomap.c
1595
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_iomap.c
1606
ASSERT(!xfs_get_extsz_hint(ip));
fs/xfs/xfs_iomap.c
1613
error = xfs_qm_dqattach(ip);
fs/xfs/xfs_iomap.c
1617
error = xfs_ilock_for_iomap(ip, flags, &lockmode);
fs/xfs/xfs_iomap.c
1621
if (XFS_IS_CORRUPT(mp, !xfs_ifork_has_extents(&ip->i_df)) ||
fs/xfs/xfs_iomap.c
1623
xfs_bmap_mark_sick(ip, XFS_DATA_FORK);
fs/xfs/xfs_iomap.c
1630
error = xfs_iread_extents(NULL, ip, XFS_DATA_FORK);
fs/xfs/xfs_iomap.c
1650
if (!xfs_iext_lookup_extent(ip, &ip->i_df, offset_fsb, &scur,
fs/xfs/xfs_iomap.c
1658
xfs_hole_to_iomap(ip, iomap, offset_fsb,
fs/xfs/xfs_iomap.c
1668
error = xfs_bmbt_to_iomap(ip, srcmap, &smap, flags, 0,
fs/xfs/xfs_iomap.c
1669
xfs_iomap_inode_sequence(ip, 0));
fs/xfs/xfs_iomap.c
1675
if (!ip->i_cowfp)
fs/xfs/xfs_iomap.c
1676
xfs_ifork_init_cow(ip);
fs/xfs/xfs_iomap.c
1678
if (!xfs_iext_lookup_extent(ip, ip->i_cowfp, offset_fsb, &icur, &got))
fs/xfs/xfs_iomap.c
1681
trace_xfs_reflink_cow_found(ip, &got);
fs/xfs/xfs_iomap.c
171
struct xfs_inode *ip,
fs/xfs/xfs_iomap.c
1715
ip->i_ino, current->comm);
fs/xfs/xfs_iomap.c
1723
error = xfs_quota_reserve_blkres(ip, count_fsb);
fs/xfs/xfs_iomap.c
1727
indlen = xfs_bmap_worst_indlen(ip, count_fsb);
fs/xfs/xfs_iomap.c
1731
ip->i_delayed_blks += count_fsb;
fs/xfs/xfs_iomap.c
1732
xfs_mod_delalloc(ip, count_fsb, indlen);
fs/xfs/xfs_iomap.c
1738
xfs_bmap_add_extent_hole_delay(ip, XFS_COW_FORK, &icur, &got);
fs/xfs/xfs_iomap.c
1742
trace_xfs_iomap_alloc(ip, offset, XFS_FSB_TO_B(mp, count_fsb),
fs/xfs/xfs_iomap.c
1745
error = xfs_bmbt_to_iomap(ip, iomap, &got, flags, iomap_flags,
fs/xfs/xfs_iomap.c
1746
xfs_iomap_inode_sequence(ip, IOMAP_F_SHARED));
fs/xfs/xfs_iomap.c
1748
xfs_iunlock(ip, lockmode);
fs/xfs/xfs_iomap.c
176
struct xfs_buftarg *target = xfs_inode_buftarg(ip);
fs/xfs/xfs_iomap.c
1763
struct xfs_inode *ip = XFS_I(inode);
fs/xfs/xfs_iomap.c
1764
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_iomap.c
1780
if (xfs_is_zoned_inode(ip))
fs/xfs/xfs_iomap.c
1785
if (xfs_get_extsz_hint(ip))
fs/xfs/xfs_iomap.c
1789
error = xfs_qm_dqattach(ip);
fs/xfs/xfs_iomap.c
1793
error = xfs_ilock_for_iomap(ip, flags, &lockmode);
fs/xfs/xfs_iomap.c
1797
if (XFS_IS_CORRUPT(mp, !xfs_ifork_has_extents(&ip->i_df)) ||
fs/xfs/xfs_iomap.c
1799
xfs_bmap_mark_sick(ip, XFS_DATA_FORK);
fs/xfs/xfs_iomap.c
180
iomap->offset = XFS_FSB_TO_B(ip->i_mount, offset_fsb);
fs/xfs/xfs_iomap.c
1806
error = xfs_iread_extents(NULL, ip, XFS_DATA_FORK);
fs/xfs/xfs_iomap.c
181
iomap->length = XFS_FSB_TO_B(ip->i_mount, end_fsb - offset_fsb);
fs/xfs/xfs_iomap.c
1816
eof = !xfs_iext_lookup_extent(ip, &ip->i_df, offset_fsb, &icur, &imap);
fs/xfs/xfs_iomap.c
1823
xfs_hole_to_iomap(ip, iomap, offset_fsb, imap.br_startoff);
fs/xfs/xfs_iomap.c
1833
xfs_fileoff_t eof_fsb = XFS_B_TO_FSB(mp, XFS_ISIZE(ip));
fs/xfs/xfs_iomap.c
1874
if (xfs_is_cow_inode(ip)) {
fs/xfs/xfs_iomap.c
1875
if (!ip->i_cowfp) {
fs/xfs/xfs_iomap.c
1876
ASSERT(!xfs_is_reflink_inode(ip));
fs/xfs/xfs_iomap.c
1877
xfs_ifork_init_cow(ip);
fs/xfs/xfs_iomap.c
1879
cow_eof = !xfs_iext_lookup_extent(ip, ip->i_cowfp, offset_fsb,
fs/xfs/xfs_iomap.c
1882
trace_xfs_reflink_cow_found(ip, &cmap);
fs/xfs/xfs_iomap.c
1893
if (!xfs_is_cow_inode(ip) ||
fs/xfs/xfs_iomap.c
1895
trace_xfs_iomap_found(ip, offset, count, XFS_DATA_FORK,
fs/xfs/xfs_iomap.c
1903
error = xfs_bmap_trim_cow(ip, &imap, &shared);
fs/xfs/xfs_iomap.c
1909
trace_xfs_iomap_found(ip, offset, count, XFS_DATA_FORK,
fs/xfs/xfs_iomap.c
1933
if (xfs_is_always_cow_inode(ip))
fs/xfs/xfs_iomap.c
1937
if (eof && offset + count > XFS_ISIZE(ip)) {
fs/xfs/xfs_iomap.c
1945
prealloc_blocks = xfs_iomap_prealloc_size(ip, allocfork,
fs/xfs/xfs_iomap.c
1948
prealloc_blocks = xfs_iomap_prealloc_size(ip, allocfork,
fs/xfs/xfs_iomap.c
1959
align = xfs_eof_alignment(ip);
fs/xfs/xfs_iomap.c
1976
error = xfs_bmapi_reserve_delalloc(ip, allocfork, offset_fsb,
fs/xfs/xfs_iomap.c
1982
trace_xfs_iomap_alloc(ip, offset, count, allocfork, &cmap);
fs/xfs/xfs_iomap.c
1986
error = xfs_bmapi_reserve_delalloc(ip, allocfork, offset_fsb,
fs/xfs/xfs_iomap.c
199
struct xfs_inode *ip)
fs/xfs/xfs_iomap.c
1992
trace_xfs_iomap_alloc(ip, offset, count, allocfork, &imap);
fs/xfs/xfs_iomap.c
1994
seq = xfs_iomap_inode_sequence(ip, iomap_flags);
fs/xfs/xfs_iomap.c
1995
xfs_iunlock(ip, lockmode);
fs/xfs/xfs_iomap.c
1996
return xfs_bmbt_to_iomap(ip, iomap, &imap, flags, iomap_flags, seq);
fs/xfs/xfs_iomap.c
1999
xfs_iunlock(ip, lockmode);
fs/xfs/xfs_iomap.c
2001
error = xfs_bmapi_convert_delalloc(ip, XFS_DATA_FORK, offset,
fs/xfs/xfs_iomap.c
2006
trace_xfs_iomap_alloc(ip, offset, count, XFS_DATA_FORK, &imap);
fs/xfs/xfs_iomap.c
201
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_iomap.c
2011
error = xfs_bmbt_to_iomap(ip, srcmap, &imap, flags, 0,
fs/xfs/xfs_iomap.c
2012
xfs_iomap_inode_sequence(ip, 0));
fs/xfs/xfs_iomap.c
2021
seq = xfs_iomap_inode_sequence(ip, iomap_flags);
fs/xfs/xfs_iomap.c
2022
xfs_iunlock(ip, lockmode);
fs/xfs/xfs_iomap.c
2023
return xfs_bmbt_to_iomap(ip, iomap, &cmap, flags, iomap_flags, seq);
fs/xfs/xfs_iomap.c
2026
xfs_iunlock(ip, lockmode);
fs/xfs/xfs_iomap.c
204
if (!XFS_IS_REALTIME_INODE(ip)) {
fs/xfs/xfs_iomap.c
2104
struct xfs_inode *ip = XFS_I(inode);
fs/xfs/xfs_iomap.c
2105
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_iomap.c
2119
error = xfs_ilock_for_iomap(ip, flags, &lockmode);
fs/xfs/xfs_iomap.c
2122
error = xfs_bmapi_read(ip, offset_fsb, end_fsb - offset_fsb, &imap,
fs/xfs/xfs_iomap.c
2125
error = xfs_reflink_trim_around_shared(ip, &imap, &shared);
fs/xfs/xfs_iomap.c
2126
seq = xfs_iomap_inode_sequence(ip, shared ? IOMAP_F_SHARED : 0);
fs/xfs/xfs_iomap.c
2127
xfs_iunlock(ip, lockmode);
fs/xfs/xfs_iomap.c
2131
trace_xfs_iomap_found(ip, offset, length, XFS_DATA_FORK, &imap);
fs/xfs/xfs_iomap.c
2132
return xfs_bmbt_to_iomap(ip, iomap, &imap, flags,
fs/xfs/xfs_iomap.c
2149
struct xfs_inode *ip = XFS_I(inode);
fs/xfs/xfs_iomap.c
2150
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_iomap.c
2163
lockmode = xfs_ilock_data_map_shared(ip);
fs/xfs/xfs_iomap.c
2164
error = xfs_iread_extents(NULL, ip, XFS_DATA_FORK);
fs/xfs/xfs_iomap.c
2168
if (xfs_iext_lookup_extent(ip, &ip->i_df, offset_fsb, &icur, &imap)) {
fs/xfs/xfs_iomap.c
218
if (align && XFS_ISIZE(ip) < XFS_FSB_TO_B(mp, align))
fs/xfs/xfs_iomap.c
2186
if (xfs_inode_has_cow_data(ip) &&
fs/xfs/xfs_iomap.c
2187
xfs_iext_lookup_extent(ip, ip->i_cowfp, offset_fsb, &icur, &cmap))
fs/xfs/xfs_iomap.c
2193
seq = xfs_iomap_inode_sequence(ip, IOMAP_F_SHARED);
fs/xfs/xfs_iomap.c
2194
error = xfs_bmbt_to_iomap(ip, iomap, &cmap, flags,
fs/xfs/xfs_iomap.c
2216
seq = xfs_iomap_inode_sequence(ip, 0);
fs/xfs/xfs_iomap.c
2218
error = xfs_bmbt_to_iomap(ip, iomap, &imap, flags, 0, seq);
fs/xfs/xfs_iomap.c
2220
xfs_iunlock(ip, lockmode);
fs/xfs/xfs_iomap.c
2237
struct xfs_inode *ip = XFS_I(inode);
fs/xfs/xfs_iomap.c
2238
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_iomap.c
2249
lockmode = xfs_ilock_attr_map_shared(ip);
fs/xfs/xfs_iomap.c
2252
if (!xfs_inode_has_attr_fork(ip) || !ip->i_af.if_nextents) {
fs/xfs/xfs_iomap.c
2257
ASSERT(ip->i_af.if_format != XFS_DINODE_FMT_LOCAL);
fs/xfs/xfs_iomap.c
2258
error = xfs_bmapi_read(ip, offset_fsb, end_fsb - offset_fsb, &imap,
fs/xfs/xfs_iomap.c
2262
seq = xfs_iomap_inode_sequence(ip, IOMAP_F_XATTR);
fs/xfs/xfs_iomap.c
2263
xfs_iunlock(ip, lockmode);
fs/xfs/xfs_iomap.c
2268
return xfs_bmbt_to_iomap(ip, iomap, &imap, flags, IOMAP_F_XATTR, seq);
fs/xfs/xfs_iomap.c
2277
struct xfs_inode *ip,
fs/xfs/xfs_iomap.c
2283
struct inode *inode = VFS_I(ip);
fs/xfs/xfs_iomap.c
2285
xfs_assert_ilocked(ip, XFS_IOLOCK_EXCL | XFS_MMAPLOCK_EXCL);
fs/xfs/xfs_iomap.c
2297
struct xfs_inode *ip,
fs/xfs/xfs_iomap.c
2302
struct inode *inode = VFS_I(ip);
fs/xfs/xfs_iomap.c
231
struct xfs_inode *ip,
fs/xfs/xfs_iomap.c
234
struct xfs_ifork *ifp = xfs_ifork_ptr(ip, XFS_DATA_FORK);
fs/xfs/xfs_iomap.c
235
xfs_extlen_t extsz = xfs_get_extsz_hint(ip);
fs/xfs/xfs_iomap.c
236
xfs_extlen_t align = xfs_eof_alignment(ip);
fs/xfs/xfs_iomap.c
266
struct xfs_inode *ip,
fs/xfs/xfs_iomap.c
273
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_iomap.c
286
xfs_get_extsz_hint(ip));
fs/xfs/xfs_iomap.c
287
if (unlikely(XFS_IS_REALTIME_INODE(ip))) {
fs/xfs/xfs_iomap.c
295
error = xfs_qm_dqattach(ip);
fs/xfs/xfs_iomap.c
321
error = xfs_trans_alloc_inode(ip, &M_RES(mp)->tr_write, dblocks,
fs/xfs/xfs_iomap.c
326
error = xfs_iext_count_extend(tp, ip, XFS_DATA_FORK, nr_exts);
fs/xfs/xfs_iomap.c
335
error = xfs_bmapi_write(tp, ip, offset_fsb, count_fsb, bmapi_flags, 0,
fs/xfs/xfs_iomap.c
347
if (unlikely(!xfs_valid_startblock(ip, imap->br_startblock))) {
fs/xfs/xfs_iomap.c
348
xfs_bmap_mark_sick(ip, XFS_DATA_FORK);
fs/xfs/xfs_iomap.c
349
error = xfs_alert_fsblock_zero(ip, imap);
fs/xfs/xfs_iomap.c
353
*seq = xfs_iomap_inode_sequence(ip, 0);
fs/xfs/xfs_iomap.c
354
xfs_iunlock(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_iomap.c
364
struct xfs_inode *ip,
fs/xfs/xfs_iomap.c
368
struct xfs_dquot *dq = xfs_inode_dquot(ip, type);
fs/xfs/xfs_iomap.c
372
if (!dq || !xfs_this_quota_on(ip->i_mount, type))
fs/xfs/xfs_iomap.c
375
if (XFS_IS_REALTIME_INODE(ip)) {
fs/xfs/xfs_iomap.c
396
struct xfs_inode *ip,
fs/xfs/xfs_iomap.c
402
struct xfs_dquot *dq = xfs_inode_dquot(ip, type);
fs/xfs/xfs_iomap.c
41
xfs_inode_t *ip,
fs/xfs/xfs_iomap.c
411
} else if (XFS_IS_REALTIME_INODE(ip)) {
fs/xfs/xfs_iomap.c
44
xfs_alert_tag(ip->i_mount, XFS_PTAG_FSBLOCK_ZERO,
fs/xfs/xfs_iomap.c
477
struct xfs_inode *ip,
fs/xfs/xfs_iomap.c
48
(unsigned long long)ip->i_ino,
fs/xfs/xfs_iomap.c
485
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_iomap.c
486
struct xfs_ifork *ifp = xfs_ifork_ptr(ip, whichfork);
fs/xfs/xfs_iomap.c
501
if (XFS_ISIZE(ip) < XFS_FSB_TO_B(mp, mp->m_allocsize_blocks))
fs/xfs/xfs_iomap.c
508
if (XFS_ISIZE(ip) < XFS_FSB_TO_B(mp, mp->m_dalign) ||
fs/xfs/xfs_iomap.c
53
xfs_bmap_mark_sick(ip, XFS_DATA_FORK);
fs/xfs/xfs_iomap.c
551
if (unlikely(XFS_IS_REALTIME_INODE(ip)))
fs/xfs/xfs_iomap.c
563
if (xfs_quota_need_throttle(ip, XFS_DQTYPE_USER, alloc_blocks))
fs/xfs/xfs_iomap.c
564
xfs_quota_calc_throttle(ip, XFS_DQTYPE_USER, &qblocks, &qshift,
fs/xfs/xfs_iomap.c
566
if (xfs_quota_need_throttle(ip, XFS_DQTYPE_GROUP, alloc_blocks))
fs/xfs/xfs_iomap.c
567
xfs_quota_calc_throttle(ip, XFS_DQTYPE_GROUP, &qblocks, &qshift,
fs/xfs/xfs_iomap.c
569
if (xfs_quota_need_throttle(ip, XFS_DQTYPE_PROJ, alloc_blocks))
fs/xfs/xfs_iomap.c
570
xfs_quota_calc_throttle(ip, XFS_DQTYPE_PROJ, &qblocks, &qshift,
fs/xfs/xfs_iomap.c
59
struct xfs_inode *ip,
fs/xfs/xfs_iomap.c
604
trace_xfs_iomap_prealloc_size(ip, alloc_blocks, shift,
fs/xfs/xfs_iomap.c
611
xfs_inode_t *ip,
fs/xfs/xfs_iomap.c
616
xfs_mount_t *mp = ip->i_mount;
fs/xfs/xfs_iomap.c
623
struct inode *inode = VFS_I(ip);
fs/xfs/xfs_iomap.c
628
trace_xfs_unwritten_convert(ip, offset, count);
fs/xfs/xfs_iomap.c
647
error = xfs_qm_dqattach(ip);
fs/xfs/xfs_iomap.c
65
return READ_ONCE(ip->i_af.if_seq);
fs/xfs/xfs_iomap.c
66
if ((iomap_flags & IOMAP_F_SHARED) && ip->i_cowfp)
fs/xfs/xfs_iomap.c
661
error = xfs_trans_alloc_inode(ip, &M_RES(mp)->tr_write, resblks,
fs/xfs/xfs_iomap.c
666
error = xfs_iext_count_extend(tp, ip, XFS_DATA_FORK,
fs/xfs/xfs_iomap.c
67
cookie = (u64)READ_ONCE(ip->i_cowfp->if_seq) << 32;
fs/xfs/xfs_iomap.c
675
error = xfs_bmapi_write(tp, ip, offset_fsb, count_fsb,
fs/xfs/xfs_iomap.c
68
return cookie | READ_ONCE(ip->i_df.if_seq);
fs/xfs/xfs_iomap.c
691
i_size = xfs_new_eof(ip, i_size);
fs/xfs/xfs_iomap.c
693
ip->i_disk_size = i_size;
fs/xfs/xfs_iomap.c
694
xfs_trans_log_inode(tp, ip, XFS_ILOG_CORE);
fs/xfs/xfs_iomap.c
698
xfs_iunlock(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_iomap.c
702
if (unlikely(!xfs_valid_startblock(ip, imap.br_startblock))) {
fs/xfs/xfs_iomap.c
703
xfs_bmap_mark_sick(ip, XFS_DATA_FORK);
fs/xfs/xfs_iomap.c
704
return xfs_alert_fsblock_zero(ip, &imap);
fs/xfs/xfs_iomap.c
723
xfs_iunlock(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_iomap.c
749
struct xfs_inode *ip,
fs/xfs/xfs_iomap.c
754
if (!xfs_is_cow_inode(ip))
fs/xfs/xfs_iomap.c
777
struct xfs_inode *ip,
fs/xfs/xfs_iomap.c
782
if (xfs_need_iread_extents(&ip->i_df))
fs/xfs/xfs_iomap.c
784
if (!xfs_ilock_nowait(ip, *lockmode))
fs/xfs/xfs_iomap.c
787
if (xfs_need_iread_extents(&ip->i_df))
fs/xfs/xfs_iomap.c
789
xfs_ilock(ip, *lockmode);
fs/xfs/xfs_iomap.c
80
struct xfs_inode *ip = XFS_I(inode);
fs/xfs/xfs_iomap.c
814
struct xfs_inode *ip,
fs/xfs/xfs_iomap.c
819
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_iomap.c
841
return len <= xfs_inode_buftarg(ip)->bt_awu_max;
fs/xfs/xfs_iomap.c
853
struct xfs_inode *ip = XFS_I(inode);
fs/xfs/xfs_iomap.c
854
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_iomap.c
86
xfs_iomap_inode_sequence(ip, iomap->flags)) {
fs/xfs/xfs_iomap.c
87
trace_xfs_iomap_invalid(ip, iomap);
fs/xfs/xfs_iomap.c
887
if (xfs_is_cow_inode(ip))
fs/xfs/xfs_iomap.c
893
error = xfs_ilock_for_iomap(ip, flags, &lockmode);
fs/xfs/xfs_iomap.c
901
if (xfs_is_cow_inode(ip) && lockmode == XFS_ILOCK_SHARED) {
fs/xfs/xfs_iomap.c
902
xfs_iunlock(ip, lockmode);
fs/xfs/xfs_iomap.c
907
error = xfs_bmapi_read(ip, offset_fsb, end_fsb - offset_fsb, &imap,
fs/xfs/xfs_iomap.c
91
XFS_ERRORTAG_DELAY(ip->i_mount, XFS_ERRTAG_WRITE_DELAY_MS);
fs/xfs/xfs_iomap.c
912
if (imap_needs_cow(ip, flags, &imap, nimaps)) {
fs/xfs/xfs_iomap.c
918
error = xfs_reflink_allocate_cow(ip, &imap, &cmap, &shared,
fs/xfs/xfs_iomap.c
925
!xfs_bmap_hw_atomic_write_possible(ip, &cmap,
fs/xfs/xfs_iomap.c
948
if (!xfs_bmap_hw_atomic_write_possible(ip, &imap, offset_fsb,
fs/xfs/xfs_iomap.c
981
seq = xfs_iomap_inode_sequence(ip, iomap_flags);
fs/xfs/xfs_iomap.c
982
xfs_iunlock(ip, lockmode);
fs/xfs/xfs_iomap.c
983
trace_xfs_iomap_found(ip, offset, length, XFS_DATA_FORK, &imap);
fs/xfs/xfs_iomap.c
984
return xfs_bmbt_to_iomap(ip, iomap, &imap, flags, iomap_flags, seq);
fs/xfs/xfs_iomap.h
15
int xfs_iomap_write_direct(struct xfs_inode *ip, xfs_fileoff_t offset_fsb,
fs/xfs/xfs_iomap.h
19
xfs_fileoff_t xfs_iomap_eof_align_last_fsb(struct xfs_inode *ip,
fs/xfs/xfs_iomap.h
22
u64 xfs_iomap_inode_sequence(struct xfs_inode *ip, u16 iomap_flags);
fs/xfs/xfs_iomap.h
23
int xfs_bmbt_to_iomap(struct xfs_inode *ip, struct iomap *iomap,
fs/xfs/xfs_iomap.h
27
int xfs_zero_range(struct xfs_inode *ip, loff_t pos, loff_t len,
fs/xfs/xfs_iomap.h
29
int xfs_truncate_page(struct xfs_inode *ip, loff_t pos,
fs/xfs/xfs_iops.c
1037
(newsize > ip->i_disk_size && oldsize != ip->i_disk_size)) {
fs/xfs/xfs_iops.c
1038
error = filemap_write_and_wait_range(VFS_I(ip)->i_mapping,
fs/xfs/xfs_iops.c
1039
ip->i_disk_size, newsize - 1);
fs/xfs/xfs_iops.c
1050
if (xfs_inode_has_bigrtalloc(ip))
fs/xfs/xfs_iops.c
1059
xfs_ilock(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_iops.c
1060
xfs_trans_ijoin(tp, ip, 0);
fs/xfs/xfs_iops.c
1091
ip->i_disk_size = newsize;
fs/xfs/xfs_iops.c
1092
xfs_trans_log_inode(tp, ip, XFS_ILOG_CORE);
fs/xfs/xfs_iops.c
1095
error = xfs_itruncate_extents(&tp, ip, XFS_DATA_FORK, newsize);
fs/xfs/xfs_iops.c
1106
xfs_iflags_set(ip, XFS_ITRUNCATED);
fs/xfs/xfs_iops.c
1109
xfs_inode_clear_eofblocks_tag(ip);
fs/xfs/xfs_iops.c
1114
xfs_trans_log_inode(tp, ip, XFS_ILOG_CORE);
fs/xfs/xfs_iops.c
1124
xfs_iunlock(ip, lock_flags);
fs/xfs/xfs_iops.c
1138
struct xfs_inode *ip = XFS_I(d_inode(dentry));
fs/xfs/xfs_iops.c
1141
trace_xfs_setattr(ip);
fs/xfs/xfs_iops.c
1146
return xfs_setattr_size(idmap, dentry, ip, iattr);
fs/xfs/xfs_iops.c
1156
struct xfs_inode *ip = XFS_I(inode);
fs/xfs/xfs_iops.c
1162
xfs_ilock(ip, XFS_MMAPLOCK_EXCL);
fs/xfs/xfs_iops.c
1167
xfs_iunlock(ip, XFS_MMAPLOCK_EXCL);
fs/xfs/xfs_iops.c
1172
xfs_iunlock(ip, XFS_MMAPLOCK_EXCL);
fs/xfs/xfs_iops.c
1174
trace_xfs_setattr(ip);
fs/xfs/xfs_iops.c
1178
error = xfs_setattr_nonsize(idmap, dentry, ip, iattr);
fs/xfs/xfs_iops.c
1190
struct xfs_inode *ip = XFS_I(inode);
fs/xfs/xfs_iops.c
1191
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_iops.c
1196
trace_xfs_update_time(ip);
fs/xfs/xfs_iops.c
1220
xfs_ilock(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_iops.c
1225
xfs_trans_ijoin(tp, ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_iops.c
1226
xfs_trans_log_inode(tp, ip, log_flags);
fs/xfs/xfs_iops.c
1234
struct xfs_inode *ip = XFS_I(inode);
fs/xfs/xfs_iops.c
1235
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_iops.c
1240
xfs_ilock(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_iops.c
1241
xfs_trans_ijoin(tp, ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_iops.c
1242
xfs_trans_log_inode(tp, ip, XFS_ILOG_TIMESTAMP);
fs/xfs/xfs_iops.c
1364
struct xfs_inode *ip)
fs/xfs/xfs_iops.c
1366
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_iops.c
1369
if (!S_ISREG(VFS_I(ip)->i_mode))
fs/xfs/xfs_iops.c
1377
return xfs_inode_buftarg(ip)->bt_daxdev != NULL;
fs/xfs/xfs_iops.c
1382
struct xfs_inode *ip)
fs/xfs/xfs_iops.c
1386
if (xfs_has_dax_never(ip->i_mount))
fs/xfs/xfs_iops.c
1388
if (!xfs_inode_supports_dax(ip))
fs/xfs/xfs_iops.c
1390
if (xfs_has_dax_always(ip->i_mount))
fs/xfs/xfs_iops.c
1392
if (ip->i_diflags2 & XFS_DIFLAG2_DAX)
fs/xfs/xfs_iops.c
1399
struct xfs_inode *ip,
fs/xfs/xfs_iops.c
1402
struct inode *inode = VFS_I(ip);
fs/xfs/xfs_iops.c
1403
unsigned int xflags = xfs_ip2xflags(ip);
fs/xfs/xfs_iops.c
1416
if (init && xfs_inode_should_enable_dax(ip))
fs/xfs/xfs_iops.c
1437
struct xfs_inode *ip)
fs/xfs/xfs_iops.c
1439
struct inode *inode = &ip->i_vnode;
fs/xfs/xfs_iops.c
1441
bool is_meta = xfs_is_internal_inode(ip);
fs/xfs/xfs_iops.c
1443
inode->i_ino = ip->i_ino;
fs/xfs/xfs_iops.c
1450
i_size_write(inode, ip->i_disk_size);
fs/xfs/xfs_iops.c
1451
xfs_diflags_to_iflags(ip, true);
fs/xfs/xfs_iops.c
1472
lockdep_set_class(&ip->i_lock, &xfs_dir_ilock_class);
fs/xfs/xfs_iops.c
1474
lockdep_set_class(&ip->i_lock, &xfs_nondir_ilock_class);
fs/xfs/xfs_iops.c
1489
if (S_ISREG(inode->i_mode) && XFS_IS_REALTIME_INODE(ip))
fs/xfs/xfs_iops.c
1490
xfs_update_stable_writes(ip);
fs/xfs/xfs_iops.c
1496
if (!xfs_inode_has_attr_fork(ip)) {
fs/xfs/xfs_iops.c
1504
struct xfs_inode *ip)
fs/xfs/xfs_iops.c
1506
struct inode *inode = &ip->i_vnode;
fs/xfs/xfs_iops.c
186
struct xfs_inode *ip = NULL;
fs/xfs/xfs_iops.c
216
error = xfs_create(&args, &name, &ip);
fs/xfs/xfs_iops.c
227
error = xfs_create_tmpfile(&args, &ip);
fs/xfs/xfs_iops.c
232
inode = VFS_I(ip);
fs/xfs/xfs_iops.c
249
xfs_setup_iops(ip);
fs/xfs/xfs_iops.c
265
xfs_finish_inode_setup(ip);
fs/xfs/xfs_iops.c
273
xfs_finish_inode_setup(ip);
fs/xfs/xfs_iops.c
276
xfs_irele(ip);
fs/xfs/xfs_iops.c
343
struct xfs_inode *ip;
fs/xfs/xfs_iops.c
353
error = xfs_lookup(XFS_I(dir), &xname, &ip, &ci_name);
fs/xfs/xfs_iops.c
367
return d_splice_alias(VFS_I(ip), dentry);
fs/xfs/xfs_iops.c
372
dentry = d_add_ci(dentry, VFS_I(ip), &dname);
fs/xfs/xfs_iops.c
540
struct xfs_inode *ip)
fs/xfs/xfs_iops.c
542
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_iops.c
548
if (XFS_IS_REALTIME_INODE(ip))
fs/xfs/xfs_iops.c
549
return XFS_FSB_TO_B(mp, xfs_get_extsz_hint(ip) ? : 1);
fs/xfs/xfs_iops.c
576
struct xfs_inode *ip,
fs/xfs/xfs_iops.c
579
struct xfs_buftarg *target = xfs_inode_buftarg(ip);
fs/xfs/xfs_iops.c
596
if (xfs_is_cow_inode(ip))
fs/xfs/xfs_iops.c
597
stat->dio_offset_align = xfs_inode_alloc_unitsize(ip);
fs/xfs/xfs_iops.c
60
struct xfs_inode *ip = XFS_I(inode);
fs/xfs/xfs_iops.c
604
struct xfs_inode *ip)
fs/xfs/xfs_iops.c
606
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_iops.c
617
if (xfs_inode_can_hw_atomic_write(ip) ||
fs/xfs/xfs_iops.c
618
xfs_inode_can_sw_atomic_write(ip))
fs/xfs/xfs_iops.c
626
struct xfs_inode *ip)
fs/xfs/xfs_iops.c
628
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_iops.c
635
if (!xfs_inode_can_sw_atomic_write(ip)) {
fs/xfs/xfs_iops.c
636
if (xfs_inode_can_hw_atomic_write(ip))
fs/xfs/xfs_iops.c
646
if (XFS_IS_REALTIME_INODE(ip))
fs/xfs/xfs_iops.c
65
.dp = ip,
fs/xfs/xfs_iops.c
653
struct xfs_inode *ip)
fs/xfs/xfs_iops.c
655
unsigned int awu_max = xfs_get_atomic_write_max(ip);
fs/xfs/xfs_iops.c
658
if (awu_max <= ip->i_mount->m_sb.sb_blocksize)
fs/xfs/xfs_iops.c
667
return min(awu_max, xfs_inode_buftarg(ip)->bt_awu_max);
fs/xfs/xfs_iops.c
672
struct xfs_inode *ip,
fs/xfs/xfs_iops.c
676
xfs_get_atomic_write_min(ip),
fs/xfs/xfs_iops.c
677
xfs_get_atomic_write_max(ip),
fs/xfs/xfs_iops.c
678
xfs_get_atomic_write_max_opt(ip));
fs/xfs/xfs_iops.c
690
struct xfs_inode *ip = XFS_I(inode);
fs/xfs/xfs_iops.c
691
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_iops.c
695
trace_xfs_getattr(ip);
fs/xfs/xfs_iops.c
700
stat->size = XFS_ISIZE(ip);
fs/xfs/xfs_iops.c
706
stat->ino = ip->i_ino;
fs/xfs/xfs_iops.c
711
stat->blocks = XFS_FSB_TO_BB(mp, ip->i_nblocks + ip->i_delayed_blks);
fs/xfs/xfs_iops.c
716
stat->btime = ip->i_crtime;
fs/xfs/xfs_iops.c
724
if (ip->i_diflags & XFS_DIFLAG_IMMUTABLE)
fs/xfs/xfs_iops.c
726
if (ip->i_diflags & XFS_DIFLAG_APPEND)
fs/xfs/xfs_iops.c
728
if (ip->i_diflags & XFS_DIFLAG_NODUMP)
fs/xfs/xfs_iops.c
743
xfs_report_dioalign(ip, stat);
fs/xfs/xfs_iops.c
745
xfs_report_atomic_write(ip, stat);
fs/xfs/xfs_iops.c
748
stat->blksize = xfs_stat_blksize(ip);
fs/xfs/xfs_iops.c
783
struct xfs_inode *ip,
fs/xfs/xfs_iops.c
786
xfs_mount_t *mp = ip->i_mount;
fs/xfs/xfs_iops.c
787
struct inode *inode = VFS_I(ip);
fs/xfs/xfs_iops.c
831
error = xfs_qm_vop_dqalloc(ip, uid, gid, ip->i_projid,
fs/xfs/xfs_iops.c
837
error = xfs_trans_alloc_ichange(ip, udqp, gdqp, NULL,
fs/xfs/xfs_iops.c
852
old_udqp = xfs_qm_vop_chown(tp, ip, &ip->i_udquot, udqp);
fs/xfs/xfs_iops.c
858
old_gdqp = xfs_qm_vop_chown(tp, ip, &ip->i_gdquot, gdqp);
fs/xfs/xfs_iops.c
862
xfs_trans_log_inode(tp, ip, XFS_ILOG_CORE);
fs/xfs/xfs_iops.c
912
struct xfs_inode *ip,
fs/xfs/xfs_iops.c
915
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_iops.c
916
struct inode *inode = VFS_I(ip);
fs/xfs/xfs_iops.c
925
xfs_assert_ilocked(ip, XFS_IOLOCK_EXCL | XFS_MMAPLOCK_EXCL);
fs/xfs/xfs_iops.c
936
if (newsize == 0 && oldsize == 0 && ip->i_df.if_nextents == 0) {
fs/xfs/xfs_iops.c
944
return xfs_setattr_nonsize(idmap, dentry, ip, iattr);
fs/xfs/xfs_iops.c
950
error = xfs_qm_dqattach(ip);
fs/xfs/xfs_iops.c
971
if (xfs_is_zoned_inode(ip)) {
fs/xfs/xfs_iops.c
992
trace_xfs_zero_eof(ip, oldsize, newsize - oldsize);
fs/xfs/xfs_iops.c
993
error = xfs_zero_range(ip, oldsize, newsize - oldsize,
fs/xfs/xfs_iops.c
996
error = xfs_truncate_page(ip, newsize, &ac, &did_zeroing);
fs/xfs/xfs_iops.c
999
if (xfs_is_zoned_inode(ip))
fs/xfs/xfs_iops.h
19
extern void xfs_setup_inode(struct xfs_inode *ip);
fs/xfs/xfs_iops.h
20
extern void xfs_setup_iops(struct xfs_inode *ip);
fs/xfs/xfs_iops.h
21
extern void xfs_diflags_to_iflags(struct xfs_inode *ip, bool init);
fs/xfs/xfs_iops.h
22
unsigned int xfs_get_atomic_write_min(struct xfs_inode *ip);
fs/xfs/xfs_iops.h
23
unsigned int xfs_get_atomic_write_max(struct xfs_inode *ip);
fs/xfs/xfs_iops.h
24
unsigned int xfs_get_atomic_write_max_opt(struct xfs_inode *ip);
fs/xfs/xfs_itable.c
100
ASSERT(ip->i_imap.im_blkno != 0);
fs/xfs/xfs_itable.c
101
inode = VFS_I(ip);
fs/xfs/xfs_itable.c
109
if (want_metadir_file(ip, bc->breq)) {
fs/xfs/xfs_itable.c
114
xfs_bulkstat_health(ip, buf);
fs/xfs/xfs_itable.c
116
xfs_iunlock(ip, XFS_ILOCK_SHARED);
fs/xfs/xfs_itable.c
117
xfs_irele(ip);
fs/xfs/xfs_itable.c
127
xfs_iunlock(ip, XFS_ILOCK_SHARED);
fs/xfs/xfs_itable.c
128
xfs_irele(ip);
fs/xfs/xfs_itable.c
136
buf->bs_projectid = ip->i_projid;
fs/xfs/xfs_itable.c
140
buf->bs_size = ip->i_disk_size;
fs/xfs/xfs_itable.c
152
buf->bs_xflags = xfs_ip2xflags(ip);
fs/xfs/xfs_itable.c
153
buf->bs_extsize_blks = ip->i_extsize;
fs/xfs/xfs_itable.c
155
nextents = xfs_ifork_nextents(&ip->i_df);
fs/xfs/xfs_itable.c
161
xfs_bulkstat_health(ip, buf);
fs/xfs/xfs_itable.c
162
buf->bs_aextents = xfs_ifork_nextents(&ip->i_af);
fs/xfs/xfs_itable.c
163
buf->bs_forkoff = xfs_inode_fork_boff(ip);
fs/xfs/xfs_itable.c
167
buf->bs_btime = ip->i_crtime.tv_sec;
fs/xfs/xfs_itable.c
168
buf->bs_btime_nsec = ip->i_crtime.tv_nsec;
fs/xfs/xfs_itable.c
169
if (ip->i_diflags2 & XFS_DIFLAG2_COWEXTSIZE)
fs/xfs/xfs_itable.c
170
buf->bs_cowextsize_blks = ip->i_cowextsize;
fs/xfs/xfs_itable.c
173
switch (ip->i_df.if_format) {
fs/xfs/xfs_itable.c
188
buf->bs_blocks = ip->i_nblocks + ip->i_delayed_blks;
fs/xfs/xfs_itable.c
191
xfs_iunlock(ip, XFS_ILOCK_SHARED);
fs/xfs/xfs_itable.c
192
xfs_irele(ip);
fs/xfs/xfs_itable.c
41
struct xfs_inode *ip,
fs/xfs/xfs_itable.c
44
return xfs_is_metadir_inode(ip) && (breq->flags & XFS_IBULK_METADIR);
fs/xfs/xfs_itable.c
72
struct xfs_inode *ip; /* incore inode pointer */
fs/xfs/xfs_itable.c
82
XFS_ILOCK_SHARED, &ip);
fs/xfs/xfs_itable.c
89
if (xfs_inode_unlinked_incomplete(ip)) {
fs/xfs/xfs_itable.c
90
error = xfs_inode_reload_unlinked_bucket(tp, ip);
fs/xfs/xfs_itable.c
92
xfs_iunlock(ip, XFS_ILOCK_SHARED);
fs/xfs/xfs_itable.c
94
xfs_irele(ip);
fs/xfs/xfs_itable.c
99
ASSERT(ip != NULL);
fs/xfs/xfs_iunlink_item.c
142
struct xfs_inode *ip,
fs/xfs/xfs_iunlink_item.c
150
ASSERT(xfs_verify_agino_or_null(pag, ip->i_next_unlinked));
fs/xfs/xfs_iunlink_item.c
157
if (ip->i_next_unlinked == next_agino) {
fs/xfs/xfs_iunlink_item.c
167
iup->ip = ip;
fs/xfs/xfs_iunlink_item.c
169
iup->old_agino = ip->i_next_unlinked;
fs/xfs/xfs_iunlink_item.c
43
return IUL_ITEM(lip)->ip->i_ino;
fs/xfs/xfs_iunlink_item.c
55
struct xfs_inode *ip = iup->ip;
fs/xfs/xfs_iunlink_item.c
62
error = xfs_imap_to_bp(tp->t_mountp, tp, &ip->i_imap, &ibp);
fs/xfs/xfs_iunlink_item.c
73
dip = xfs_buf_offset(ibp, ip->i_imap.im_boffset);
fs/xfs/xfs_iunlink_item.c
78
xfs_inode_verifier_error(ip, -EFSCORRUPTED, __func__, dip,
fs/xfs/xfs_iunlink_item.c
87
offset = ip->i_imap.im_boffset +
fs/xfs/xfs_iunlink_item.h
16
struct xfs_inode *ip;
fs/xfs/xfs_iunlink_item.h
24
int xfs_iunlink_log_inode(struct xfs_trans *tp, struct xfs_inode *ip,
fs/xfs/xfs_log_recover.c
1781
struct xfs_inode *ip;
fs/xfs/xfs_log_recover.c
1784
error = xlog_recover_iget(mp, ino, &ip);
fs/xfs/xfs_log_recover.c
1788
if (VFS_I(ip)->i_generation != gen) {
fs/xfs/xfs_log_recover.c
1789
xfs_irele(ip);
fs/xfs/xfs_log_recover.c
1793
*ipp = ip;
fs/xfs/xfs_log_recover.c
2728
struct xfs_inode *ip;
fs/xfs/xfs_log_recover.c
2735
&ip);
fs/xfs/xfs_log_recover.c
2739
ASSERT(VFS_I(ip)->i_nlink == 0);
fs/xfs/xfs_log_recover.c
2740
ASSERT(VFS_I(ip)->i_mode != 0);
fs/xfs/xfs_log_recover.c
2741
xfs_iflags_clear(ip, XFS_IRECOVERY);
fs/xfs/xfs_log_recover.c
2742
agino = ip->i_next_unlinked;
fs/xfs/xfs_log_recover.c
2745
ip->i_prev_unlinked = prev_agino;
fs/xfs/xfs_log_recover.c
2763
prev_ip = ip;
fs/xfs/xfs_log_recover.c
2769
ip->i_prev_unlinked = prev_agino;
fs/xfs/xfs_mount.c
1675
struct xfs_inode *ip,
fs/xfs/xfs_mount.c
1679
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_mount.c
1681
if (XFS_IS_REALTIME_INODE(ip)) {
fs/xfs/xfs_mount.h
790
int xfs_zero_extent(struct xfs_inode *ip, xfs_fsblock_t start_fsb,
fs/xfs/xfs_mount.h
798
void xfs_mod_delalloc(struct xfs_inode *ip, int64_t data_delta,
fs/xfs/xfs_notify_failure.c
107
mapping = VFS_I(ip)->i_mapping;
fs/xfs/xfs_notify_failure.c
121
fserror_report_data_lost(VFS_I(ip), (u64)pgoff << PAGE_SHIFT,
fs/xfs/xfs_notify_failure.c
124
xfs_irele(ip);
fs/xfs/xfs_notify_failure.c
80
struct xfs_inode *ip;
fs/xfs/xfs_notify_failure.c
98
0, &ip);
fs/xfs/xfs_platform.h
271
#define XFS_IS_REALTIME_INODE(ip) \
fs/xfs/xfs_platform.h
272
(((ip)->i_diflags & XFS_DIFLAG_REALTIME) && \
fs/xfs/xfs_platform.h
273
(ip)->i_mount->m_rtdev_targp)
fs/xfs/xfs_platform.h
276
#define XFS_IS_REALTIME_INODE(ip) (0)
fs/xfs/xfs_pnfs.c
100
xfs_trans_log_inode(tp, ip, XFS_ILOG_CORE);
fs/xfs/xfs_pnfs.c
116
struct xfs_inode *ip = XFS_I(inode);
fs/xfs/xfs_pnfs.c
117
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_pnfs.c
135
if (XFS_IS_REALTIME_INODE(ip))
fs/xfs/xfs_pnfs.c
142
if (xfs_is_reflink_inode(ip))
fs/xfs/xfs_pnfs.c
152
xfs_ilock(ip, XFS_IOLOCK_EXCL);
fs/xfs/xfs_pnfs.c
174
lock_flags = xfs_ilock_data_map_shared(ip);
fs/xfs/xfs_pnfs.c
175
error = xfs_bmapi_read(ip, offset_fsb, end_fsb - offset_fsb,
fs/xfs/xfs_pnfs.c
177
seq = xfs_iomap_inode_sequence(ip, 0);
fs/xfs/xfs_pnfs.c
183
if (offset + length > XFS_ISIZE(ip))
fs/xfs/xfs_pnfs.c
184
end_fsb = xfs_iomap_eof_align_last_fsb(ip, end_fsb);
fs/xfs/xfs_pnfs.c
188
xfs_iunlock(ip, lock_flags);
fs/xfs/xfs_pnfs.c
190
error = xfs_iomap_write_direct(ip, offset_fsb,
fs/xfs/xfs_pnfs.c
200
error = xfs_fs_map_update_inode(ip);
fs/xfs/xfs_pnfs.c
202
error = xfs_log_force_inode(ip);
fs/xfs/xfs_pnfs.c
207
xfs_iunlock(ip, lock_flags);
fs/xfs/xfs_pnfs.c
209
xfs_iunlock(ip, XFS_IOLOCK_EXCL);
fs/xfs/xfs_pnfs.c
211
error = xfs_bmbt_to_iomap(ip, iomap, &imap, 0, 0, seq);
fs/xfs/xfs_pnfs.c
215
xfs_iunlock(ip, XFS_IOLOCK_EXCL);
fs/xfs/xfs_pnfs.c
224
struct xfs_inode *ip,
fs/xfs/xfs_pnfs.c
231
xfs_ilock(ip, XFS_ILOCK_SHARED);
fs/xfs/xfs_pnfs.c
232
error = xfs_bmapi_read(ip, XFS_B_TO_FSBT(ip->i_mount, isize - 1), 1,
fs/xfs/xfs_pnfs.c
234
xfs_iunlock(ip, XFS_ILOCK_SHARED);
fs/xfs/xfs_pnfs.c
262
struct xfs_inode *ip = XFS_I(inode);
fs/xfs/xfs_pnfs.c
263
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_pnfs.c
271
xfs_ilock(ip, XFS_IOLOCK_EXCL);
fs/xfs/xfs_pnfs.c
302
error = xfs_iomap_write_unwritten(ip, start, length, false);
fs/xfs/xfs_pnfs.c
308
error = xfs_pnfs_validate_isize(ip, size);
fs/xfs/xfs_pnfs.c
317
xfs_ilock(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_pnfs.c
318
xfs_trans_ijoin(tp, ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_pnfs.c
319
xfs_trans_log_inode(tp, ip, XFS_ILOG_CORE);
fs/xfs/xfs_pnfs.c
325
ip->i_disk_size = iattr->ia_size;
fs/xfs/xfs_pnfs.c
33
struct xfs_inode *ip = XFS_I(inode);
fs/xfs/xfs_pnfs.c
332
xfs_iunlock(ip, XFS_IOLOCK_EXCL);
fs/xfs/xfs_pnfs.c
37
xfs_iunlock(ip, *iolock);
fs/xfs/xfs_pnfs.c
42
xfs_ilock(ip, *iolock);
fs/xfs/xfs_pnfs.c
81
struct xfs_inode *ip)
fs/xfs/xfs_pnfs.c
86
error = xfs_trans_alloc(ip->i_mount, &M_RES(ip->i_mount)->tr_writeid,
fs/xfs/xfs_pnfs.c
91
xfs_ilock(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_pnfs.c
92
xfs_trans_ijoin(tp, ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_pnfs.c
94
VFS_I(ip)->i_mode &= ~S_ISUID;
fs/xfs/xfs_pnfs.c
95
if (VFS_I(ip)->i_mode & S_IXGRP)
fs/xfs/xfs_pnfs.c
96
VFS_I(ip)->i_mode &= ~S_ISGID;
fs/xfs/xfs_pnfs.c
97
xfs_trans_ichgtime(tp, ip, XFS_ICHGTIME_MOD | XFS_ICHGTIME_CHG);
fs/xfs/xfs_pnfs.c
98
ip->i_diflags |= XFS_DIFLAG_PREALLOC;
fs/xfs/xfs_qm.c
1250
struct xfs_inode *ip,
fs/xfs/xfs_qm.c
1255
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_qm.c
1260
id = xfs_qm_id_for_quotatype(ip, type);
fs/xfs/xfs_qm.c
1322
struct xfs_inode *ip;
fs/xfs/xfs_qm.c
1340
error = xfs_iget(mp, tp, ino, XFS_IGET_DONTCACHE, 0, &ip);
fs/xfs/xfs_qm.c
1351
if (xfs_inode_unlinked_incomplete(ip)) {
fs/xfs/xfs_qm.c
1352
error = xfs_inode_reload_unlinked(ip);
fs/xfs/xfs_qm.c
1360
if (xfs_is_metadir_inode(ip))
fs/xfs/xfs_qm.c
1363
ASSERT(ip->i_delayed_blks == 0);
fs/xfs/xfs_qm.c
1365
lock_mode = xfs_ilock_data_map_shared(ip);
fs/xfs/xfs_qm.c
1366
if (XFS_IS_REALTIME_INODE(ip)) {
fs/xfs/xfs_qm.c
1367
error = xfs_iread_extents(tp, ip, XFS_DATA_FORK);
fs/xfs/xfs_qm.c
1369
xfs_iunlock(ip, lock_mode);
fs/xfs/xfs_qm.c
1373
xfs_inode_count_blocks(tp, ip, &nblks, &rtblks);
fs/xfs/xfs_qm.c
1374
xfs_iflags_clear(ip, XFS_IQUOTAUNCHECKED);
fs/xfs/xfs_qm.c
1375
xfs_iunlock(ip, lock_mode);
fs/xfs/xfs_qm.c
1390
error = xfs_qm_quotacheck_dqadjust(ip, XFS_DQTYPE_USER, nblks,
fs/xfs/xfs_qm.c
1397
error = xfs_qm_quotacheck_dqadjust(ip, XFS_DQTYPE_GROUP, nblks,
fs/xfs/xfs_qm.c
1404
error = xfs_qm_quotacheck_dqadjust(ip, XFS_DQTYPE_PROJ, nblks,
fs/xfs/xfs_qm.c
1411
xfs_irele(ip);
fs/xfs/xfs_qm.c
1848
struct xfs_inode *ip,
fs/xfs/xfs_qm.c
1857
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_qm.c
1858
struct inode *inode = VFS_I(ip);
fs/xfs/xfs_qm.c
1868
ASSERT(!xfs_is_metadir_inode(ip));
fs/xfs/xfs_qm.c
1870
if ((flags & XFS_QMOPT_INHERIT) && XFS_INHERIT_GID(ip))
fs/xfs/xfs_qm.c
1877
if (XFS_NOT_DQATTACHED(mp, ip)) {
fs/xfs/xfs_qm.c
1878
xfs_ilock(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_qm.c
1879
error = xfs_qm_dqattach_locked(ip, true);
fs/xfs/xfs_qm.c
1880
xfs_iunlock(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_qm.c
1899
ASSERT(ip->i_udquot);
fs/xfs/xfs_qm.c
1900
uq = xfs_qm_dqhold(ip->i_udquot);
fs/xfs/xfs_qm.c
1913
ASSERT(ip->i_gdquot);
fs/xfs/xfs_qm.c
1914
gq = xfs_qm_dqhold(ip->i_gdquot);
fs/xfs/xfs_qm.c
1919
if (ip->i_projid != prid) {
fs/xfs/xfs_qm.c
1927
ASSERT(ip->i_pdquot);
fs/xfs/xfs_qm.c
1928
pq = xfs_qm_dqhold(ip->i_pdquot);
fs/xfs/xfs_qm.c
1931
trace_xfs_dquot_dqalloc(ip);
fs/xfs/xfs_qm.c
1960
struct xfs_inode *ip,
fs/xfs/xfs_qm.c
1966
bool isrt = XFS_IS_REALTIME_INODE(ip);
fs/xfs/xfs_qm.c
1968
xfs_assert_ilocked(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_qm.c
1969
ASSERT(XFS_IS_QUOTA_ON(ip->i_mount));
fs/xfs/xfs_qm.c
1970
ASSERT(!xfs_is_metadir_inode(ip));
fs/xfs/xfs_qm.c
1977
xfs_inode_count_blocks(tp, ip, &dblocks, &rblocks);
fs/xfs/xfs_qm.c
1979
xfs_trans_mod_ino_dquot(tp, ip, prevdq, XFS_TRANS_DQ_BCOUNT,
fs/xfs/xfs_qm.c
1981
xfs_trans_mod_ino_dquot(tp, ip, prevdq, XFS_TRANS_DQ_RTBCOUNT,
fs/xfs/xfs_qm.c
1983
xfs_trans_mod_ino_dquot(tp, ip, prevdq, XFS_TRANS_DQ_ICOUNT, -1);
fs/xfs/xfs_qm.c
1986
xfs_trans_mod_ino_dquot(tp, ip, newdq, XFS_TRANS_DQ_BCOUNT, dblocks);
fs/xfs/xfs_qm.c
1987
xfs_trans_mod_ino_dquot(tp, ip, newdq, XFS_TRANS_DQ_RTBCOUNT, rblocks);
fs/xfs/xfs_qm.c
1988
xfs_trans_mod_ino_dquot(tp, ip, newdq, XFS_TRANS_DQ_ICOUNT, 1);
fs/xfs/xfs_qm.c
1999
-ip->i_delayed_blks);
fs/xfs/xfs_qm.c
2011
ASSERT(prevdq->q_rtb.reserved >= ip->i_delayed_blks);
fs/xfs/xfs_qm.c
2012
prevdq->q_rtb.reserved -= ip->i_delayed_blks;
fs/xfs/xfs_qm.c
2014
ASSERT(prevdq->q_blk.reserved >= ip->i_delayed_blks);
fs/xfs/xfs_qm.c
2015
prevdq->q_blk.reserved -= ip->i_delayed_blks;
fs/xfs/xfs_qm.c
2039
struct xfs_inode *ip = i_tab[i];
fs/xfs/xfs_qm.c
2045
if (i == 0 || ip != i_tab[i-1]) {
fs/xfs/xfs_qm.c
2046
if (XFS_NOT_DQATTACHED(mp, ip)) {
fs/xfs/xfs_qm.c
2047
error = xfs_qm_dqattach(ip);
fs/xfs/xfs_qm.c
2059
struct xfs_inode *ip,
fs/xfs/xfs_qm.c
2069
xfs_assert_ilocked(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_qm.c
2070
ASSERT(!xfs_is_metadir_inode(ip));
fs/xfs/xfs_qm.c
2073
ASSERT(ip->i_udquot == NULL);
fs/xfs/xfs_qm.c
2074
ASSERT(i_uid_read(VFS_I(ip)) == udqp->q_id);
fs/xfs/xfs_qm.c
2076
ip->i_udquot = xfs_qm_dqhold(udqp);
fs/xfs/xfs_qm.c
2079
ASSERT(ip->i_gdquot == NULL);
fs/xfs/xfs_qm.c
2080
ASSERT(i_gid_read(VFS_I(ip)) == gdqp->q_id);
fs/xfs/xfs_qm.c
2082
ip->i_gdquot = xfs_qm_dqhold(gdqp);
fs/xfs/xfs_qm.c
2085
ASSERT(ip->i_pdquot == NULL);
fs/xfs/xfs_qm.c
2086
ASSERT(ip->i_projid == pdqp->q_id);
fs/xfs/xfs_qm.c
2088
ip->i_pdquot = xfs_qm_dqhold(pdqp);
fs/xfs/xfs_qm.c
2091
xfs_trans_mod_dquot_byino(tp, ip, XFS_TRANS_DQ_ICOUNT, 1);
fs/xfs/xfs_qm.c
2097
struct xfs_inode *ip,
fs/xfs/xfs_qm.c
2106
dqp = xfs_inode_dquot(ip, type);
fs/xfs/xfs_qm.c
2115
if (XFS_IS_REALTIME_INODE(ip)) {
fs/xfs/xfs_qm.c
292
struct xfs_inode *ip)
fs/xfs/xfs_qm.c
294
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_qm.c
298
if (!XFS_NOT_DQATTACHED(mp, ip))
fs/xfs/xfs_qm.c
300
if (xfs_is_quota_inode(&mp->m_sb, ip->i_ino))
fs/xfs/xfs_qm.c
302
if (xfs_is_metadir_inode(ip))
fs/xfs/xfs_qm.c
316
xfs_inode_t *ip,
fs/xfs/xfs_qm.c
319
xfs_mount_t *mp = ip->i_mount;
fs/xfs/xfs_qm.c
322
if (!xfs_qm_need_dqattach(ip))
fs/xfs/xfs_qm.c
325
xfs_assert_ilocked(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_qm.c
326
ASSERT(!xfs_is_metadir_inode(ip));
fs/xfs/xfs_qm.c
328
if (XFS_IS_UQUOTA_ON(mp) && !ip->i_udquot) {
fs/xfs/xfs_qm.c
329
error = xfs_qm_dqget_inode(ip, XFS_DQTYPE_USER,
fs/xfs/xfs_qm.c
330
doalloc, &ip->i_udquot);
fs/xfs/xfs_qm.c
333
ASSERT(ip->i_udquot);
fs/xfs/xfs_qm.c
336
if (XFS_IS_GQUOTA_ON(mp) && !ip->i_gdquot) {
fs/xfs/xfs_qm.c
337
error = xfs_qm_dqget_inode(ip, XFS_DQTYPE_GROUP,
fs/xfs/xfs_qm.c
338
doalloc, &ip->i_gdquot);
fs/xfs/xfs_qm.c
341
ASSERT(ip->i_gdquot);
fs/xfs/xfs_qm.c
344
if (XFS_IS_PQUOTA_ON(mp) && !ip->i_pdquot) {
fs/xfs/xfs_qm.c
345
error = xfs_qm_dqget_inode(ip, XFS_DQTYPE_PROJ,
fs/xfs/xfs_qm.c
346
doalloc, &ip->i_pdquot);
fs/xfs/xfs_qm.c
349
ASSERT(ip->i_pdquot);
fs/xfs/xfs_qm.c
357
xfs_assert_ilocked(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_qm.c
363
struct xfs_inode *ip)
fs/xfs/xfs_qm.c
367
if (!xfs_qm_need_dqattach(ip))
fs/xfs/xfs_qm.c
370
xfs_ilock(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_qm.c
371
error = xfs_qm_dqattach_locked(ip, false);
fs/xfs/xfs_qm.c
372
xfs_iunlock(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_qm.c
384
xfs_inode_t *ip)
fs/xfs/xfs_qm.c
386
if (xfs_is_metadir_inode(ip))
fs/xfs/xfs_qm.c
388
if (!(ip->i_udquot || ip->i_gdquot || ip->i_pdquot))
fs/xfs/xfs_qm.c
391
trace_xfs_dquot_dqdetach(ip);
fs/xfs/xfs_qm.c
393
ASSERT(!xfs_is_quota_inode(&ip->i_mount->m_sb, ip->i_ino));
fs/xfs/xfs_qm.c
394
if (ip->i_udquot) {
fs/xfs/xfs_qm.c
395
xfs_qm_dqrele(ip->i_udquot);
fs/xfs/xfs_qm.c
396
ip->i_udquot = NULL;
fs/xfs/xfs_qm.c
398
if (ip->i_gdquot) {
fs/xfs/xfs_qm.c
399
xfs_qm_dqrele(ip->i_gdquot);
fs/xfs/xfs_qm.c
400
ip->i_gdquot = NULL;
fs/xfs/xfs_qm.c
402
if (ip->i_pdquot) {
fs/xfs/xfs_qm.c
403
xfs_qm_dqrele(ip->i_pdquot);
fs/xfs/xfs_qm.c
404
ip->i_pdquot = NULL;
fs/xfs/xfs_qm_bhv.c
22
struct xfs_inode *ip,
fs/xfs/xfs_qm_bhv.c
28
if (XFS_IS_REALTIME_MOUNT(ip->i_mount) &&
fs/xfs/xfs_qm_bhv.c
29
(ip->i_diflags & (XFS_DIFLAG_RTINHERIT | XFS_DIFLAG_REALTIME)))
fs/xfs/xfs_qm_bhv.c
69
struct xfs_inode *ip,
fs/xfs/xfs_qm_bhv.c
72
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_qm_bhv.c
75
if (!xfs_qm_dqget(mp, ip->i_projid, XFS_DQTYPE_PROJ, false, &dqp)) {
fs/xfs/xfs_qm_bhv.c
77
xfs_fill_statvfs_from_dquot(statp, ip, dqp);
fs/xfs/xfs_qm_syscalls.c
58
struct xfs_inode *ip;
fs/xfs/xfs_qm_syscalls.c
62
error = xfs_qm_qino_load(mp, type, &ip);
fs/xfs/xfs_qm_syscalls.c
68
xfs_ilock(ip, XFS_IOLOCK_EXCL);
fs/xfs/xfs_qm_syscalls.c
72
xfs_iunlock(ip, XFS_IOLOCK_EXCL);
fs/xfs/xfs_qm_syscalls.c
76
xfs_ilock(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_qm_syscalls.c
77
xfs_trans_ijoin(tp, ip, 0);
fs/xfs/xfs_qm_syscalls.c
79
ip->i_disk_size = 0;
fs/xfs/xfs_qm_syscalls.c
80
xfs_trans_log_inode(tp, ip, XFS_ILOG_CORE);
fs/xfs/xfs_qm_syscalls.c
82
error = xfs_itruncate_extents(&tp, ip, XFS_DATA_FORK, 0);
fs/xfs/xfs_qm_syscalls.c
88
ASSERT(ip->i_df.if_nextents == 0);
fs/xfs/xfs_qm_syscalls.c
90
xfs_trans_ichgtime(tp, ip, XFS_ICHGTIME_MOD | XFS_ICHGTIME_CHG);
fs/xfs/xfs_qm_syscalls.c
94
xfs_iunlock(ip, XFS_ILOCK_EXCL | XFS_IOLOCK_EXCL);
fs/xfs/xfs_qm_syscalls.c
96
xfs_irele(ip);
fs/xfs/xfs_quota.h
101
int xfs_trans_reserve_quota_nblks(struct xfs_trans *tp, struct xfs_inode *ip,
fs/xfs/xfs_quota.h
119
extern int xfs_qm_dqattach_locked(struct xfs_inode *ip, bool doalloc);
fs/xfs/xfs_quota.h
128
bool xfs_inode_near_dquot_enforcement(struct xfs_inode *ip, xfs_dqtype_t type);
fs/xfs/xfs_quota.h
129
int xfs_quota_reserve_blkres(struct xfs_inode *ip, int64_t blocks);
fs/xfs/xfs_quota.h
132
void xfs_trans_mod_ino_dquot(struct xfs_trans *tp, struct xfs_inode *ip,
fs/xfs/xfs_quota.h
150
# define xfs_trans_mod_ino_dquot(tp, ip, dqp, field, delta) \
fs/xfs/xfs_quota.h
156
xfs_qm_vop_dqalloc(struct xfs_inode *ip, kuid_t kuid, kgid_t kgid,
fs/xfs/xfs_quota.h
168
struct xfs_inode *ip, uint field, int64_t delta)
fs/xfs/xfs_quota.h
174
struct xfs_inode *ip, int64_t dblocks, int64_t rblocks,
fs/xfs/xfs_quota.h
194
#define xfs_qm_vop_create_dqattach(tp, ip, u, g, p)
fs/xfs/xfs_quota.h
196
#define xfs_qm_vop_chown(tp, ip, old, new) (NULL)
fs/xfs/xfs_quota.h
197
#define xfs_qm_dqattach(ip) (0)
fs/xfs/xfs_quota.h
198
#define xfs_qm_dqattach_locked(ip, fl) (0)
fs/xfs/xfs_quota.h
199
#define xfs_qm_dqdetach(ip)
fs/xfs/xfs_quota.h
201
#define xfs_qm_statvfs(ip, s) do { } while(0)
fs/xfs/xfs_quota.h
207
#define xfs_inode_near_dquot_enforcement(ip, type) (false)
fs/xfs/xfs_quota.h
209
static inline int xfs_quota_reserve_blkres(struct xfs_inode *ip, int64_t blocks)
fs/xfs/xfs_quota.h
222
xfs_quota_unreserve_blkres(struct xfs_inode *ip, uint64_t blocks)
fs/xfs/xfs_quota.h
225
xfs_quota_reserve_blkres(ip, -(int64_t)blocks);
fs/xfs/xfs_quota.h
27
#define XFS_NOT_DQATTACHED(mp, ip) \
fs/xfs/xfs_quota.h
28
((XFS_IS_UQUOTA_ON(mp) && (ip)->i_udquot == NULL) || \
fs/xfs/xfs_quota.h
29
(XFS_IS_GQUOTA_ON(mp) && (ip)->i_gdquot == NULL) || \
fs/xfs/xfs_quota.h
30
(XFS_IS_PQUOTA_ON(mp) && (ip)->i_pdquot == NULL))
fs/xfs/xfs_quotaops.c
25
struct xfs_inode *ip;
fs/xfs/xfs_quotaops.c
29
error = xfs_qm_qino_load(mp, type, &ip);
fs/xfs/xfs_quotaops.c
37
tstate->ino = ip->i_ino;
fs/xfs/xfs_quotaops.c
39
tstate->blocks = ip->i_nblocks;
fs/xfs/xfs_quotaops.c
40
tstate->nextents = ip->i_df.if_nextents;
fs/xfs/xfs_quotaops.c
47
xfs_irele(ip);
fs/xfs/xfs_reflink.c
1002
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_reflink.c
1006
trace_xfs_reflink_end_cow(ip, offset, count);
fs/xfs/xfs_reflink.c
1023
xfs_ilock(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_reflink.c
1024
xfs_trans_ijoin(tp, ip, 0);
fs/xfs/xfs_reflink.c
1027
error = xfs_reflink_end_cow_extent_locked(tp, ip, &offset_fsb,
fs/xfs/xfs_reflink.c
1031
trace_xfs_reflink_end_cow_error(ip, error, _RET_IP_);
fs/xfs/xfs_reflink.c
1035
xfs_iunlock(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_reflink.c
1039
xfs_iunlock(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_reflink.c
1284
struct xfs_inode *ip,
fs/xfs/xfs_reflink.c
1293
if (XFS_IS_REALTIME_INODE(ip)) {
fs/xfs/xfs_reflink.c
1314
struct xfs_inode *ip,
fs/xfs/xfs_reflink.c
1319
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_reflink.c
1327
bool isrt = XFS_IS_REALTIME_INODE(ip);
fs/xfs/xfs_reflink.c
1354
if (XFS_IS_REALTIME_INODE(ip)) {
fs/xfs/xfs_reflink.c
1361
error = xfs_trans_alloc_inode(ip, &M_RES(mp)->tr_write,
fs/xfs/xfs_reflink.c
1365
error = xfs_trans_alloc_inode(ip, &M_RES(mp)->tr_write,
fs/xfs/xfs_reflink.c
1377
error = xfs_bmapi_read(ip, dmap->br_startoff, dmap->br_blockcount,
fs/xfs/xfs_reflink.c
1391
trace_xfs_reflink_remap_extent_dest(ip, &smap);
fs/xfs/xfs_reflink.c
1400
xfs_bmap_mark_sick(ip, XFS_DATA_FORK);
fs/xfs/xfs_reflink.c
1413
error = xfs_reflink_ag_has_free_space(mp, ip,
fs/xfs/xfs_reflink.c
1442
if (XFS_IS_REALTIME_INODE(ip)) {
fs/xfs/xfs_reflink.c
1449
error = xfs_trans_reserve_quota_nblks(tp, ip, dblocks, rblocks,
fs/xfs/xfs_reflink.c
1461
error = xfs_iext_count_extend(tp, ip, XFS_DATA_FORK, iext_delta);
fs/xfs/xfs_reflink.c
1470
xfs_bmap_unmap_extent(tp, ip, XFS_DATA_FORK, &smap);
fs/xfs/xfs_reflink.c
1482
error = xfs_bunmapi(NULL, ip, smap.br_startoff,
fs/xfs/xfs_reflink.c
1495
xfs_bmap_map_extent(tp, ip, XFS_DATA_FORK, dmap);
fs/xfs/xfs_reflink.c
1499
xfs_reflink_update_quota(tp, ip, false, qdelta);
fs/xfs/xfs_reflink.c
1504
if (newlen > i_size_read(VFS_I(ip))) {
fs/xfs/xfs_reflink.c
1505
trace_xfs_reflink_update_inode_size(ip, newlen);
fs/xfs/xfs_reflink.c
1506
i_size_write(VFS_I(ip), newlen);
fs/xfs/xfs_reflink.c
1507
ip->i_disk_size = newlen;
fs/xfs/xfs_reflink.c
1508
xfs_trans_log_inode(tp, ip, XFS_ILOG_CORE);
fs/xfs/xfs_reflink.c
1518
xfs_iunlock(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_reflink.c
1521
trace_xfs_reflink_remap_extent_error(ip, error, _RET_IP_);
fs/xfs/xfs_reflink.c
1610
struct xfs_inode *ip,
fs/xfs/xfs_reflink.c
1613
loff_t isize = i_size_read(VFS_I(ip));
fs/xfs/xfs_reflink.c
1618
trace_xfs_zero_eof(ip, isize, pos - isize);
fs/xfs/xfs_reflink.c
1619
return xfs_zero_range(ip, isize, pos - isize, NULL, NULL);
fs/xfs/xfs_reflink.c
1736
struct xfs_inode *ip,
fs/xfs/xfs_reflink.c
1740
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_reflink.c
1746
ifp = xfs_ifork_ptr(ip, XFS_DATA_FORK);
fs/xfs/xfs_reflink.c
1747
error = xfs_iread_extents(tp, ip, XFS_DATA_FORK);
fs/xfs/xfs_reflink.c
1752
found = xfs_iext_lookup_extent(ip, ifp, 0, &icur, &got);
fs/xfs/xfs_reflink.c
1760
if (XFS_IS_REALTIME_INODE(ip))
fs/xfs/xfs_reflink.c
1789
struct xfs_inode *ip,
fs/xfs/xfs_reflink.c
1795
ASSERT(xfs_is_reflink_inode(ip));
fs/xfs/xfs_reflink.c
1797
if (!xfs_can_free_cowblocks(ip))
fs/xfs/xfs_reflink.c
1800
error = xfs_reflink_inode_has_shared_extents(*tpp, ip, &needs_flag);
fs/xfs/xfs_reflink.c
1808
error = xfs_reflink_cancel_cow_blocks(ip, tpp, 0, XFS_MAX_FILEOFF,
fs/xfs/xfs_reflink.c
1814
trace_xfs_reflink_unset_inode_flag(ip);
fs/xfs/xfs_reflink.c
1815
ip->i_diflags2 &= ~XFS_DIFLAG2_REFLINK;
fs/xfs/xfs_reflink.c
1816
xfs_inode_clear_cowblocks_tag(ip);
fs/xfs/xfs_reflink.c
1817
xfs_trans_log_inode(*tpp, ip, XFS_ILOG_CORE);
fs/xfs/xfs_reflink.c
1828
struct xfs_inode *ip)
fs/xfs/xfs_reflink.c
1830
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_reflink.c
1839
xfs_ilock(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_reflink.c
1840
xfs_trans_ijoin(tp, ip, 0);
fs/xfs/xfs_reflink.c
1842
error = xfs_reflink_clear_inode_flag(ip, &tp);
fs/xfs/xfs_reflink.c
1850
xfs_iunlock(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_reflink.c
1855
xfs_iunlock(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_reflink.c
1865
struct xfs_inode *ip,
fs/xfs/xfs_reflink.c
1869
struct inode *inode = VFS_I(ip);
fs/xfs/xfs_reflink.c
1872
if (!xfs_is_reflink_inode(ip))
fs/xfs/xfs_reflink.c
1875
trace_xfs_reflink_unshare(ip, offset, len);
fs/xfs/xfs_reflink.c
1895
error = xfs_reflink_try_clear_inode_flag(ip);
fs/xfs/xfs_reflink.c
1901
trace_xfs_reflink_unshare_error(ip, error, _RET_IP_);
fs/xfs/xfs_reflink.c
229
struct xfs_inode *ip,
fs/xfs/xfs_reflink.c
233
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_reflink.c
238
if (!xfs_is_reflink_inode(ip) || !xfs_bmap_is_written_extent(irec)) {
fs/xfs/xfs_reflink.c
243
trace_xfs_reflink_trim_around_shared(ip, irec);
fs/xfs/xfs_reflink.c
245
if (XFS_IS_REALTIME_INODE(ip))
fs/xfs/xfs_reflink.c
281
struct xfs_inode *ip,
fs/xfs/xfs_reflink.c
286
if (xfs_is_always_cow_inode(ip) &&
fs/xfs/xfs_reflink.c
293
return xfs_reflink_trim_around_shared(ip, imap, shared);
fs/xfs/xfs_reflink.c
298
struct xfs_inode *ip,
fs/xfs/xfs_reflink.c
308
if (!xfs_iext_lookup_extent(ip, ip->i_cowfp, offset_fsb, &icur, &got))
fs/xfs/xfs_reflink.c
324
error = xfs_bmap_add_extent_unwritten_real(NULL, ip,
fs/xfs/xfs_reflink.c
329
} while (xfs_iext_next_extent(ip->i_cowfp, &icur, &got));
fs/xfs/xfs_reflink.c
337
struct xfs_inode *ip,
fs/xfs/xfs_reflink.c
341
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_reflink.c
349
xfs_ilock(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_reflink.c
350
error = xfs_reflink_convert_cow_locked(ip, offset_fsb, count_fsb);
fs/xfs/xfs_reflink.c
351
xfs_iunlock(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_reflink.c
362
struct xfs_inode *ip,
fs/xfs/xfs_reflink.c
378
if (!xfs_iext_lookup_extent(ip, ip->i_cowfp, offset_fsb, &icur, cmap))
fs/xfs/xfs_reflink.c
383
return xfs_bmap_trim_cow(ip, imap, shared);
fs/xfs/xfs_reflink.c
400
struct xfs_inode *ip,
fs/xfs/xfs_reflink.c
422
trace_xfs_reflink_convert_cow(ip, cmap);
fs/xfs/xfs_reflink.c
424
error = xfs_reflink_convert_cow_locked(ip, offset_fsb, count_fsb);
fs/xfs/xfs_reflink.c
433
struct xfs_inode *ip,
fs/xfs/xfs_reflink.c
440
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_reflink.c
449
imap->br_blockcount, xfs_get_cowextsz_hint(ip));
fs/xfs/xfs_reflink.c
450
if (XFS_IS_REALTIME_INODE(ip)) {
fs/xfs/xfs_reflink.c
458
xfs_iunlock(ip, *lockmode);
fs/xfs/xfs_reflink.c
461
error = xfs_trans_alloc_inode(ip, &M_RES(mp)->tr_write, dblocks,
fs/xfs/xfs_reflink.c
468
error = xfs_find_trim_cow_extent(ip, imap, cmap, shared, &found);
fs/xfs/xfs_reflink.c
479
error = xfs_bmapi_write(tp, ip, imap->br_startoff, imap->br_blockcount,
fs/xfs/xfs_reflink.c
485
xfs_inode_set_cowblocks_tag(ip);
fs/xfs/xfs_reflink.c
491
return xfs_reflink_convert_unwritten(ip, imap, cmap, convert_now);
fs/xfs/xfs_reflink.c
500
struct xfs_inode *ip,
fs/xfs/xfs_reflink.c
507
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_reflink.c
514
xfs_iunlock(ip, *lockmode);
fs/xfs/xfs_reflink.c
517
error = xfs_trans_alloc_inode(ip, &M_RES(mp)->tr_write, 0, 0,
fs/xfs/xfs_reflink.c
524
error = xfs_find_trim_cow_extent(ip, imap, cmap, shared,
fs/xfs/xfs_reflink.c
541
error = xfs_bmapi_write(tp, ip, cmap->br_startoff,
fs/xfs/xfs_reflink.c
548
xfs_inode_set_cowblocks_tag(ip);
fs/xfs/xfs_reflink.c
554
return xfs_reflink_convert_unwritten(ip, imap, cmap, convert_now);
fs/xfs/xfs_reflink.c
564
struct xfs_inode *ip,
fs/xfs/xfs_reflink.c
574
xfs_assert_ilocked(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_reflink.c
575
if (!ip->i_cowfp) {
fs/xfs/xfs_reflink.c
576
ASSERT(!xfs_is_reflink_inode(ip));
fs/xfs/xfs_reflink.c
577
xfs_ifork_init_cow(ip);
fs/xfs/xfs_reflink.c
580
error = xfs_find_trim_cow_extent(ip, imap, cmap, shared, &found);
fs/xfs/xfs_reflink.c
586
return xfs_reflink_convert_unwritten(ip, imap, cmap,
fs/xfs/xfs_reflink.c
594
return xfs_reflink_fill_cow_hole(ip, imap, cmap, shared,
fs/xfs/xfs_reflink.c
603
return xfs_reflink_fill_delalloc(ip, imap, cmap, shared,
fs/xfs/xfs_reflink.c
622
struct xfs_inode *ip,
fs/xfs/xfs_reflink.c
628
struct xfs_ifork *ifp = xfs_ifork_ptr(ip, XFS_COW_FORK);
fs/xfs/xfs_reflink.c
631
bool isrt = XFS_IS_REALTIME_INODE(ip);
fs/xfs/xfs_reflink.c
634
if (!xfs_inode_has_cow_data(ip))
fs/xfs/xfs_reflink.c
636
if (!xfs_iext_lookup_extent_before(ip, ifp, &end_fsb, &icur, &got))
fs/xfs/xfs_reflink.c
650
trace_xfs_reflink_cancel_cow(ip, &del);
fs/xfs/xfs_reflink.c
653
xfs_bmap_del_extent_delay(ip, XFS_COW_FORK, &icur, &got,
fs/xfs/xfs_reflink.c
675
xfs_bmap_del_extent_cow(ip, &icur, &got, &del);
fs/xfs/xfs_reflink.c
678
xfs_quota_unreserve_blkres(ip, del.br_blockcount);
fs/xfs/xfs_reflink.c
690
xfs_inode_clear_cowblocks_tag(ip);
fs/xfs/xfs_reflink.c
702
struct xfs_inode *ip,
fs/xfs/xfs_reflink.c
712
trace_xfs_reflink_cancel_cow_range(ip, offset, count);
fs/xfs/xfs_reflink.c
713
ASSERT(ip->i_cowfp);
fs/xfs/xfs_reflink.c
715
offset_fsb = XFS_B_TO_FSBT(ip->i_mount, offset);
fs/xfs/xfs_reflink.c
719
end_fsb = XFS_B_TO_FSB(ip->i_mount, offset + count);
fs/xfs/xfs_reflink.c
722
error = xfs_trans_alloc(ip->i_mount, &M_RES(ip->i_mount)->tr_write,
fs/xfs/xfs_reflink.c
727
xfs_ilock(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_reflink.c
728
xfs_trans_ijoin(tp, ip, 0);
fs/xfs/xfs_reflink.c
731
error = xfs_reflink_cancel_cow_blocks(ip, &tp, offset_fsb, end_fsb,
fs/xfs/xfs_reflink.c
738
xfs_iunlock(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_reflink.c
743
xfs_iunlock(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_reflink.c
745
trace_xfs_reflink_cancel_cow_range_error(ip, error, _RET_IP_);
fs/xfs/xfs_reflink.c
759
struct xfs_inode *ip,
fs/xfs/xfs_reflink.c
765
if (XFS_IS_REALTIME_INODE(ip)) {
fs/xfs/xfs_reflink.c
772
xfs_trans_mod_dquot_byino(tp, ip, qflag, blocks);
fs/xfs/xfs_reflink.c
775
# define xfs_reflink_update_quota(tp, ip, is_cow, blocks) ((void)0)
fs/xfs/xfs_reflink.c
791
struct xfs_inode *ip,
fs/xfs/xfs_reflink.c
797
struct xfs_ifork *ifp = xfs_ifork_ptr(ip, XFS_COW_FORK);
fs/xfs/xfs_reflink.c
799
bool isrt = XFS_IS_REALTIME_INODE(ip);
fs/xfs/xfs_reflink.c
807
if (!xfs_iext_lookup_extent(ip, ifp, *offset_fsb, &icur, &got) ||
fs/xfs/xfs_reflink.c
830
error = xfs_iext_count_extend(tp, ip, XFS_DATA_FORK,
fs/xfs/xfs_reflink.c
837
error = xfs_bmapi_read(ip, del.br_startoff, del.br_blockcount, &data,
fs/xfs/xfs_reflink.c
846
trace_xfs_reflink_cow_remap_from(ip, &del);
fs/xfs/xfs_reflink.c
847
trace_xfs_reflink_cow_remap_to(ip, &data);
fs/xfs/xfs_reflink.c
854
xfs_bmap_unmap_extent(tp, ip, XFS_DATA_FORK, &data);
fs/xfs/xfs_reflink.c
856
xfs_reflink_update_quota(tp, ip, false, -data.br_blockcount);
fs/xfs/xfs_reflink.c
866
error = xfs_bunmapi(NULL, ip, data.br_startoff,
fs/xfs/xfs_reflink.c
878
xfs_bmap_map_extent(tp, ip, XFS_DATA_FORK, &del);
fs/xfs/xfs_reflink.c
881
xfs_reflink_update_quota(tp, ip, true, del.br_blockcount);
fs/xfs/xfs_reflink.c
884
xfs_bmap_del_extent_cow(ip, &icur, &got, &del);
fs/xfs/xfs_reflink.c
903
struct xfs_inode *ip,
fs/xfs/xfs_reflink.c
907
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_reflink.c
917
xfs_ilock(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_reflink.c
918
xfs_trans_ijoin(tp, ip, 0);
fs/xfs/xfs_reflink.c
920
error = xfs_reflink_end_cow_extent_locked(tp, ip, offset_fsb, end_fsb);
fs/xfs/xfs_reflink.c
925
xfs_iunlock(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_reflink.c
934
struct xfs_inode *ip,
fs/xfs/xfs_reflink.c
942
trace_xfs_reflink_end_cow(ip, offset, count);
fs/xfs/xfs_reflink.c
944
offset_fsb = XFS_B_TO_FSBT(ip->i_mount, offset);
fs/xfs/xfs_reflink.c
945
end_fsb = XFS_B_TO_FSB(ip->i_mount, offset + count);
fs/xfs/xfs_reflink.c
980
error = xfs_reflink_end_cow_extent(ip, &offset_fsb, end_fsb);
fs/xfs/xfs_reflink.c
983
trace_xfs_reflink_end_cow_error(ip, error, _RET_IP_);
fs/xfs/xfs_reflink.c
995
struct xfs_inode *ip,
fs/xfs/xfs_reflink.h
16
xfs_can_free_cowblocks(struct xfs_inode *ip)
fs/xfs/xfs_reflink.h
18
struct inode *inode = VFS_I(ip);
fs/xfs/xfs_reflink.h
28
int xfs_reflink_trim_around_shared(struct xfs_inode *ip,
fs/xfs/xfs_reflink.h
30
int xfs_bmap_trim_cow(struct xfs_inode *ip, struct xfs_bmbt_irec *imap,
fs/xfs/xfs_reflink.h
33
int xfs_reflink_allocate_cow(struct xfs_inode *ip, struct xfs_bmbt_irec *imap,
fs/xfs/xfs_reflink.h
36
extern int xfs_reflink_convert_cow(struct xfs_inode *ip, xfs_off_t offset,
fs/xfs/xfs_reflink.h
38
int xfs_reflink_convert_cow_locked(struct xfs_inode *ip,
fs/xfs/xfs_reflink.h
41
extern int xfs_reflink_cancel_cow_blocks(struct xfs_inode *ip,
fs/xfs/xfs_reflink.h
44
extern int xfs_reflink_cancel_cow_range(struct xfs_inode *ip, xfs_off_t offset,
fs/xfs/xfs_reflink.h
46
extern int xfs_reflink_end_cow(struct xfs_inode *ip, xfs_off_t offset,
fs/xfs/xfs_reflink.h
48
int xfs_reflink_end_atomic_cow(struct xfs_inode *ip, xfs_off_t offset,
fs/xfs/xfs_reflink.h
55
struct xfs_inode *ip, bool *has_shared);
fs/xfs/xfs_reflink.h
56
extern int xfs_reflink_clear_inode_flag(struct xfs_inode *ip,
fs/xfs/xfs_reflink.h
58
extern int xfs_reflink_unshare(struct xfs_inode *ip, xfs_off_t offset,
fs/xfs/xfs_rtalloc.c
1646
struct xfs_inode *ip)
fs/xfs/xfs_rtalloc.c
1650
xfs_ilock(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_rtalloc.c
1652
error = xfs_iread_extents(tp, ip, XFS_DATA_FORK);
fs/xfs/xfs_rtalloc.c
1656
if (xfs_inode_has_attr_fork(ip)) {
fs/xfs/xfs_rtalloc.c
1657
error = xfs_iread_extents(tp, ip, XFS_ATTR_FORK);
fs/xfs/xfs_rtalloc.c
1663
xfs_iunlock(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_rtalloc.c
2090
struct xfs_mount *mp = ap->ip->i_mount;
fs/xfs/xfs_rtalloc.c
2101
align = xfs_get_cowextsz_hint(ap->ip);
fs/xfs/xfs_rtalloc.c
2103
align = xfs_get_extsz_hint(ap->ip);
fs/xfs/xfs_rtalloc.c
2181
if (xfs_has_rtgroups(ap->ip->i_mount)) {
fs/xfs/xfs_rtalloc.c
700
struct xfs_inode *ip = mp->m_rootip;
fs/xfs/xfs_rtalloc.c
704
xfs_ilock(ip, XFS_IOLOCK_EXCL);
fs/xfs/xfs_rtalloc.c
705
if (!(ip->i_diflags & XFS_DIFLAG_RTINHERIT) ||
fs/xfs/xfs_rtalloc.c
706
!(ip->i_diflags & XFS_DIFLAG_EXTSZINHERIT))
fs/xfs/xfs_rtalloc.c
709
error = xfs_trans_alloc_inode(ip, &M_RES(mp)->tr_ichange, 0, 0, false,
fs/xfs/xfs_rtalloc.c
714
xfs_trans_log_inode(tp, ip, XFS_ILOG_CORE);
fs/xfs/xfs_rtalloc.c
716
xfs_iunlock(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_rtalloc.c
719
xfs_iunlock(ip, XFS_IOLOCK_EXCL);
fs/xfs/xfs_super.c
710
struct xfs_inode *ip = XFS_I(inode);
fs/xfs/xfs_super.c
712
trace_xfs_destroy_inode(ip);
fs/xfs/xfs_super.c
715
XFS_STATS_INC(ip->i_mount, xs_inode_destroy);
fs/xfs/xfs_super.c
716
XFS_STATS_INC(ip->i_mount, xs_inode_destroy2);
fs/xfs/xfs_super.c
717
xfs_inode_mark_reclaimable(ip);
fs/xfs/xfs_super.c
732
struct xfs_inode *ip = inode;
fs/xfs/xfs_super.c
734
memset(ip, 0, sizeof(struct xfs_inode));
fs/xfs/xfs_super.c
737
inode_init_once(VFS_I(ip));
fs/xfs/xfs_super.c
740
atomic_set(&ip->i_pincount, 0);
fs/xfs/xfs_super.c
741
spin_lock_init(&ip->i_flags_lock);
fs/xfs/xfs_super.c
742
init_rwsem(&ip->i_lock);
fs/xfs/xfs_super.c
756
struct xfs_inode *ip = XFS_I(inode);
fs/xfs/xfs_super.c
763
if (ip->i_flags & XFS_IRECOVERY) {
fs/xfs/xfs_super.c
764
ASSERT(xlog_recovery_needed(ip->i_mount->m_log));
fs/xfs/xfs_super.c
923
struct xfs_inode *ip = XFS_I(d_inode(dentry));
fs/xfs/xfs_super.c
940
(ip->i_diflags & (XFS_DIFLAG_RTINHERIT | XFS_DIFLAG_REALTIME)))
fs/xfs/xfs_super.c
943
if ((ip->i_diflags & XFS_DIFLAG_PROJINHERIT) &&
fs/xfs/xfs_super.c
946
xfs_qm_statvfs(ip, st);
fs/xfs/xfs_symlink.c
170
error = xfs_icreate(tp, ino, &args, &du.ip);
fs/xfs/xfs_symlink.c
186
xfs_qm_vop_create_dqattach(tp, du.ip, udqp, gdqp, pdqp);
fs/xfs/xfs_symlink.c
189
error = xfs_symlink_write_target(tp, du.ip, du.ip->i_ino, target_path,
fs/xfs/xfs_symlink.c
194
i_size_write(VFS_I(du.ip), du.ip->i_disk_size);
fs/xfs/xfs_symlink.c
219
*ipp = du.ip;
fs/xfs/xfs_symlink.c
220
xfs_iunlock(du.ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_symlink.c
233
if (du.ip) {
fs/xfs/xfs_symlink.c
234
xfs_iunlock(du.ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_symlink.c
235
xfs_finish_inode_setup(du.ip);
fs/xfs/xfs_symlink.c
236
xfs_irele(du.ip);
fs/xfs/xfs_symlink.c
261
struct xfs_inode *ip)
fs/xfs/xfs_symlink.c
263
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_symlink.c
267
ASSERT(!xfs_need_iread_extents(&ip->i_df));
fs/xfs/xfs_symlink.c
275
ASSERT(ip->i_df.if_nextents > 0 && ip->i_df.if_nextents <= 2);
fs/xfs/xfs_symlink.c
281
xfs_ilock(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_symlink.c
282
xfs_trans_ijoin(tp, ip, 0);
fs/xfs/xfs_symlink.c
290
ip->i_disk_size = 0;
fs/xfs/xfs_symlink.c
291
VFS_I(ip)->i_mode = (VFS_I(ip)->i_mode & ~S_IFMT) | S_IFREG;
fs/xfs/xfs_symlink.c
292
xfs_trans_log_inode(tp, ip, XFS_ILOG_CORE);
fs/xfs/xfs_symlink.c
294
error = xfs_symlink_remote_truncate(tp, ip);
fs/xfs/xfs_symlink.c
307
if (ip->i_df.if_bytes)
fs/xfs/xfs_symlink.c
308
xfs_idata_realloc(ip, -ip->i_df.if_bytes, XFS_DATA_FORK);
fs/xfs/xfs_symlink.c
309
ASSERT(ip->i_df.if_bytes == 0);
fs/xfs/xfs_symlink.c
311
xfs_iunlock(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_symlink.c
317
xfs_iunlock(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_symlink.c
326
struct xfs_inode *ip)
fs/xfs/xfs_symlink.c
328
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_symlink.c
33
struct xfs_inode *ip,
fs/xfs/xfs_symlink.c
331
trace_xfs_inactive_symlink(ip);
fs/xfs/xfs_symlink.c
336
xfs_ilock(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_symlink.c
337
pathlen = (int)ip->i_disk_size;
fs/xfs/xfs_symlink.c
342
__func__, (unsigned long long)ip->i_ino, pathlen);
fs/xfs/xfs_symlink.c
343
xfs_iunlock(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_symlink.c
345
xfs_inode_mark_sick(ip, XFS_SICK_INO_SYMLINK);
fs/xfs/xfs_symlink.c
353
if (ip->i_df.if_format == XFS_DINODE_FMT_LOCAL) {
fs/xfs/xfs_symlink.c
354
xfs_iunlock(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_symlink.c
358
xfs_iunlock(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_symlink.c
36
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_symlink.c
361
return xfs_inactive_symlink_rmt(ip);
fs/xfs/xfs_symlink.c
40
trace_xfs_readlink(ip);
fs/xfs/xfs_symlink.c
44
if (xfs_ifork_zapped(ip, XFS_DATA_FORK))
fs/xfs/xfs_symlink.c
47
xfs_ilock(ip, XFS_ILOCK_SHARED);
fs/xfs/xfs_symlink.c
49
pathlen = ip->i_disk_size;
fs/xfs/xfs_symlink.c
55
__func__, (unsigned long long) ip->i_ino,
fs/xfs/xfs_symlink.c
61
if (ip->i_df.if_format == XFS_DINODE_FMT_LOCAL) {
fs/xfs/xfs_symlink.c
66
if (XFS_IS_CORRUPT(ip->i_mount, !ip->i_df.if_data))
fs/xfs/xfs_symlink.c
69
memcpy(link, ip->i_df.if_data, pathlen + 1);
fs/xfs/xfs_symlink.c
72
error = xfs_symlink_remote_read(ip, link);
fs/xfs/xfs_symlink.c
75
xfs_iunlock(ip, XFS_ILOCK_SHARED);
fs/xfs/xfs_symlink.c
78
xfs_iunlock(ip, XFS_ILOCK_SHARED);
fs/xfs/xfs_symlink.c
79
xfs_inode_mark_sick(ip, XFS_SICK_INO_SYMLINK);
fs/xfs/xfs_symlink.h
13
int xfs_readlink(struct xfs_inode *ip, char *link);
fs/xfs/xfs_symlink.h
14
int xfs_inactive_symlink(struct xfs_inode *ip);
fs/xfs/xfs_trace.h
1011
TP_PROTO(struct xfs_inode *ip, unsigned lock_flags,
fs/xfs/xfs_trace.h
1013
TP_ARGS(ip, lock_flags, caller_ip),
fs/xfs/xfs_trace.h
1021
__entry->dev = VFS_I(ip)->i_sb->s_dev;
fs/xfs/xfs_trace.h
1022
__entry->ino = ip->i_ino;
fs/xfs/xfs_trace.h
1035
TP_PROTO(struct xfs_inode *ip, unsigned lock_flags, \
fs/xfs/xfs_trace.h
1037
TP_ARGS(ip, lock_flags, caller_ip))
fs/xfs/xfs_trace.h
1044
TP_PROTO(struct xfs_inode *ip),
fs/xfs/xfs_trace.h
1045
TP_ARGS(ip),
fs/xfs/xfs_trace.h
1052
__entry->dev = VFS_I(ip)->i_sb->s_dev;
fs/xfs/xfs_trace.h
1053
__entry->ino = ip->i_ino;
fs/xfs/xfs_trace.h
1054
__entry->iflags = ip->i_flags;
fs/xfs/xfs_trace.h
1064
TP_PROTO(struct xfs_inode *ip), \
fs/xfs/xfs_trace.h
1065
TP_ARGS(ip))
fs/xfs/xfs_trace.h
1120
TP_PROTO(struct xfs_inode *ip, unsigned int order),
fs/xfs/xfs_trace.h
1121
TP_ARGS(ip, order),
fs/xfs/xfs_trace.h
1128
__entry->dev = VFS_I(ip)->i_sb->s_dev;
fs/xfs/xfs_trace.h
1129
__entry->ino = ip->i_ino;
fs/xfs/xfs_trace.h
1140
TP_PROTO(struct xfs_inode *ip, unsigned int order), \
fs/xfs/xfs_trace.h
1141
TP_ARGS(ip, order))
fs/xfs/xfs_trace.h
1146
TP_PROTO(struct xfs_inode *ip, unsigned long caller_ip),
fs/xfs/xfs_trace.h
1147
TP_ARGS(ip, caller_ip),
fs/xfs/xfs_trace.h
1157
__entry->dev = VFS_I(ip)->i_sb->s_dev;
fs/xfs/xfs_trace.h
1158
__entry->ino = ip->i_ino;
fs/xfs/xfs_trace.h
1159
__entry->count = icount_read(VFS_I(ip));
fs/xfs/xfs_trace.h
1160
__entry->pincount = atomic_read(&ip->i_pincount);
fs/xfs/xfs_trace.h
1161
__entry->iflags = ip->i_flags;
fs/xfs/xfs_trace.h
1174
TP_PROTO(struct xfs_inode *ip, xfs_fsblock_t blocks, int shift,
fs/xfs/xfs_trace.h
1176
TP_ARGS(ip, blocks, shift, writeio_blocks),
fs/xfs/xfs_trace.h
1185
__entry->dev = VFS_I(ip)->i_sb->s_dev;
fs/xfs/xfs_trace.h
1186
__entry->ino = ip->i_ino;
fs/xfs/xfs_trace.h
1253
TP_PROTO(struct xfs_inode *ip, unsigned long caller_ip), \
fs/xfs/xfs_trace.h
1254
TP_ARGS(ip, caller_ip))
fs/xfs/xfs_trace.h
1787
TP_PROTO(struct xfs_inode *ip, xfs_off_t offset, ssize_t count),
fs/xfs/xfs_trace.h
1788
TP_ARGS(ip, offset, count),
fs/xfs/xfs_trace.h
1796
__entry->dev = VFS_I(ip)->i_sb->s_dev;
fs/xfs/xfs_trace.h
1797
__entry->ino = ip->i_ino;
fs/xfs/xfs_trace.h
1809
TP_PROTO(struct xfs_inode *ip, xfs_off_t offset, ssize_t count,
fs/xfs/xfs_trace.h
1811
TP_ARGS(ip, offset, count, whichfork, irec),
fs/xfs/xfs_trace.h
1824
__entry->dev = VFS_I(ip)->i_sb->s_dev;
fs/xfs/xfs_trace.h
1825
__entry->ino = ip->i_ino;
fs/xfs/xfs_trace.h
1826
__entry->size = ip->i_disk_size;
fs/xfs/xfs_trace.h
1849
TP_PROTO(struct xfs_inode *ip, xfs_off_t offset, ssize_t count, \
fs/xfs/xfs_trace.h
1851
TP_ARGS(ip, offset, count, whichfork, irec))
fs/xfs/xfs_trace.h
1858
TP_PROTO(struct xfs_inode *ip, xfs_off_t offset, ssize_t count),
fs/xfs/xfs_trace.h
1859
TP_ARGS(ip, offset, count),
fs/xfs/xfs_trace.h
1869
__entry->dev = VFS_I(ip)->i_sb->s_dev;
fs/xfs/xfs_trace.h
1870
__entry->ino = ip->i_ino;
fs/xfs/xfs_trace.h
1871
__entry->isize = VFS_I(ip)->i_size;
fs/xfs/xfs_trace.h
1872
__entry->disize = ip->i_disk_size;
fs/xfs/xfs_trace.h
1888
TP_PROTO(struct xfs_inode *ip, xfs_off_t offset, ssize_t count), \
fs/xfs/xfs_trace.h
1889
TP_ARGS(ip, offset, count))
fs/xfs/xfs_trace.h
1899
TP_PROTO(struct xfs_inode *ip, xfs_fsize_t new_size),
fs/xfs/xfs_trace.h
1900
TP_ARGS(ip, new_size),
fs/xfs/xfs_trace.h
1908
__entry->dev = VFS_I(ip)->i_sb->s_dev;
fs/xfs/xfs_trace.h
1909
__entry->ino = ip->i_ino;
fs/xfs/xfs_trace.h
1910
__entry->size = ip->i_disk_size;
fs/xfs/xfs_trace.h
1922
TP_PROTO(struct xfs_inode *ip, xfs_fsize_t new_size), \
fs/xfs/xfs_trace.h
1923
TP_ARGS(ip, new_size))
fs/xfs/xfs_trace.h
1928
TP_PROTO(struct xfs_inode *ip, xfs_fileoff_t fileoff, xfs_filblks_t len,
fs/xfs/xfs_trace.h
1930
TP_ARGS(ip, fileoff, len, flags, caller_ip),
fs/xfs/xfs_trace.h
1941
__entry->dev = VFS_I(ip)->i_sb->s_dev;
fs/xfs/xfs_trace.h
1942
__entry->ino = ip->i_ino;
fs/xfs/xfs_trace.h
1943
__entry->size = ip->i_disk_size;
fs/xfs/xfs_trace.h
2573
TP_PROTO(struct xfs_inode *ip, int which),
fs/xfs/xfs_trace.h
2574
TP_ARGS(ip, which),
fs/xfs/xfs_trace.h
2585
__entry->dev = VFS_I(ip)->i_sb->s_dev;
fs/xfs/xfs_trace.h
2587
__entry->ino = ip->i_ino;
fs/xfs/xfs_trace.h
2588
__entry->format = ip->i_df.if_format;
fs/xfs/xfs_trace.h
2589
__entry->nex = ip->i_df.if_nextents;
fs/xfs/xfs_trace.h
2590
__entry->broot_size = ip->i_df.if_broot_bytes;
fs/xfs/xfs_trace.h
2591
__entry->fork_off = xfs_inode_fork_boff(ip);
fs/xfs/xfs_trace.h
2606
TP_PROTO(struct xfs_inode *ip, int which), \
fs/xfs/xfs_trace.h
2607
TP_ARGS(ip, which))
fs/xfs/xfs_trace.h
2945
__entry->ino = cur->bc_ino.ip->i_ino;
fs/xfs/xfs_trace.h
2997
__entry->ino = cur->bc_ino.ip->i_ino;
fs/xfs/xfs_trace.h
3252
__entry->ino = cur->bc_ino.ip->i_ino;
fs/xfs/xfs_trace.h
3450
struct xfs_inode *ip = bi->bi_owner;
fs/xfs/xfs_trace.h
3451
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_trace.h
3471
__entry->ino = ip->i_ino;
fs/xfs/xfs_trace.h
3965
TP_PROTO(struct xfs_inode *ip, int error, unsigned long caller_ip),
fs/xfs/xfs_trace.h
3966
TP_ARGS(ip, error, caller_ip),
fs/xfs/xfs_trace.h
3974
__entry->dev = VFS_I(ip)->i_sb->s_dev;
fs/xfs/xfs_trace.h
3975
__entry->ino = ip->i_ino;
fs/xfs/xfs_trace.h
3988
TP_PROTO(struct xfs_inode *ip, int error, \
fs/xfs/xfs_trace.h
3990
TP_ARGS(ip, error, caller_ip))
fs/xfs/xfs_trace.h
4046
TP_PROTO(struct xfs_inode *ip, struct xfs_bmbt_irec *irec),
fs/xfs/xfs_trace.h
4047
TP_ARGS(ip, irec),
fs/xfs/xfs_trace.h
4057
__entry->dev = VFS_I(ip)->i_sb->s_dev;
fs/xfs/xfs_trace.h
4058
__entry->ino = ip->i_ino;
fs/xfs/xfs_trace.h
4074
TP_PROTO(struct xfs_inode *ip, struct xfs_bmbt_irec *irec), \
fs/xfs/xfs_trace.h
4075
TP_ARGS(ip, irec))
fs/xfs/xfs_trace.h
4079
TP_PROTO(struct xfs_inode *ip, const struct iomap *iomap, unsigned int wpcseq, int whichfork),
fs/xfs/xfs_trace.h
4080
TP_ARGS(ip, iomap, wpcseq, whichfork),
fs/xfs/xfs_trace.h
4093
__entry->dev = VFS_I(ip)->i_sb->s_dev;
fs/xfs/xfs_trace.h
4094
__entry->ino = ip->i_ino;
fs/xfs/xfs_trace.h
4101
__entry->forkseq = READ_ONCE(xfs_ifork_ptr(ip, whichfork)->if_seq);
fs/xfs/xfs_trace.h
4116
TP_PROTO(struct xfs_inode *ip, const struct iomap *iomap, unsigned int wpcseq, int whichfork), \
fs/xfs/xfs_trace.h
4117
TP_ARGS(ip, iomap, wpcseq, whichfork))
fs/xfs/xfs_trace.h
4122
TP_PROTO(struct xfs_inode *ip, const struct iomap *iomap),
fs/xfs/xfs_trace.h
4123
TP_ARGS(ip, iomap),
fs/xfs/xfs_trace.h
4136
__entry->dev = VFS_I(ip)->i_sb->s_dev;
fs/xfs/xfs_trace.h
4137
__entry->ino = ip->i_ino;
fs/xfs/xfs_trace.h
4144
__entry->inodeseq = xfs_iomap_inode_sequence(ip, iomap->flags);
fs/xfs/xfs_trace.h
4159
TP_PROTO(struct xfs_inode *ip, const struct iomap *iomap), \
fs/xfs/xfs_trace.h
4160
TP_ARGS(ip, iomap))
fs/xfs/xfs_trace.h
4506
XFS_INO_TO_AGINO(iup->ip->i_mount, iup->ip->i_ino);
fs/xfs/xfs_trace.h
4519
TP_PROTO(struct xfs_inode *ip),
fs/xfs/xfs_trace.h
4520
TP_ARGS(ip),
fs/xfs/xfs_trace.h
4529
__entry->dev = ip->i_mount->m_super->s_dev;
fs/xfs/xfs_trace.h
4530
__entry->agno = XFS_INO_TO_AGNO(ip->i_mount, ip->i_ino);
fs/xfs/xfs_trace.h
4531
__entry->agino = XFS_INO_TO_AGINO(ip->i_mount, ip->i_ino);
fs/xfs/xfs_trace.h
4532
__entry->prev_agino = ip->i_prev_unlinked;
fs/xfs/xfs_trace.h
4533
__entry->next_agino = ip->i_next_unlinked;
fs/xfs/xfs_trace.h
4544
TP_PROTO(struct xfs_inode *ip),
fs/xfs/xfs_trace.h
4545
TP_ARGS(ip),
fs/xfs/xfs_trace.h
4552
__entry->dev = ip->i_mount->m_super->s_dev;
fs/xfs/xfs_trace.h
4553
__entry->agno = XFS_INO_TO_AGNO(ip->i_mount, ip->i_ino);
fs/xfs/xfs_trace.h
4554
__entry->agino = XFS_INO_TO_AGINO(ip->i_mount, ip->i_ino);
fs/xfs/xfs_trace.h
4564
TP_PROTO(struct xfs_inode *ip),
fs/xfs/xfs_trace.h
4565
TP_ARGS(ip),
fs/xfs/xfs_trace.h
4572
__entry->dev = VFS_I(ip)->i_sb->s_dev;
fs/xfs/xfs_trace.h
4573
__entry->agno = XFS_INO_TO_AGNO(ip->i_mount, ip->i_ino);
fs/xfs/xfs_trace.h
4574
__entry->agino = XFS_INO_TO_AGINO(ip->i_mount, ip->i_ino);
fs/xfs/xfs_trace.h
4583
TP_PROTO(struct xfs_inode *ip), \
fs/xfs/xfs_trace.h
4584
TP_ARGS(ip))
fs/xfs/xfs_trace.h
4642
TP_PROTO(struct xfs_inode *ip, unsigned int flags),
fs/xfs/xfs_trace.h
4643
TP_ARGS(ip, flags),
fs/xfs/xfs_trace.h
4650
__entry->dev = ip->i_mount->m_super->s_dev;
fs/xfs/xfs_trace.h
4651
__entry->ino = ip->i_ino;
fs/xfs/xfs_trace.h
4660
TP_PROTO(struct xfs_inode *ip, unsigned int flags), \
fs/xfs/xfs_trace.h
4661
TP_ARGS(ip, flags))
fs/xfs/xfs_trace.h
4771
cur->bc_ino.ip->i_ino);
fs/xfs/xfs_trace.h
4773
cur->bc_ino.ip->i_ino);
fs/xfs/xfs_trace.h
5024
TP_PROTO(int das, struct xfs_inode *ip),
fs/xfs/xfs_trace.h
5025
TP_ARGS(das, ip),
fs/xfs/xfs_trace.h
5032
__entry->ino = ip->i_ino;
fs/xfs/xfs_trace.h
5041
TP_PROTO(int das, struct xfs_inode *ip), \
fs/xfs/xfs_trace.h
5042
TP_ARGS(das, ip))
fs/xfs/xfs_trace.h
5283
TP_PROTO(struct xfs_inode *ip, int whichfile),
fs/xfs/xfs_trace.h
5284
TP_ARGS(ip, whichfile),
fs/xfs/xfs_trace.h
5295
__entry->dev = VFS_I(ip)->i_sb->s_dev;
fs/xfs/xfs_trace.h
5297
__entry->ino = ip->i_ino;
fs/xfs/xfs_trace.h
5298
__entry->format = ip->i_df.if_format;
fs/xfs/xfs_trace.h
5299
__entry->nex = ip->i_df.if_nextents;
fs/xfs/xfs_trace.h
5300
__entry->fork_off = xfs_inode_fork_boff(ip);
fs/xfs/xfs_trace.h
5313
TP_PROTO(struct xfs_inode *ip, int whichfile), \
fs/xfs/xfs_trace.h
5314
TP_ARGS(ip, whichfile))
fs/xfs/xfs_trace.h
5644
TP_PROTO(struct xfs_inode *ip, const struct xfs_getparents *ppi,
fs/xfs/xfs_trace.h
5647
TP_ARGS(ip, ppi, context, pptr),
fs/xfs/xfs_trace.h
5659
__entry->dev = ip->i_mount->m_super->s_dev;
fs/xfs/xfs_trace.h
5660
__entry->ino = ip->i_ino;
fs/xfs/xfs_trace.h
5680
TP_PROTO(struct xfs_inode *ip, const struct xfs_getparents *ppi, \
fs/xfs/xfs_trace.h
5683
TP_ARGS(ip, ppi, context, pptr))
fs/xfs/xfs_trace.h
5688
TP_PROTO(struct xfs_inode *ip, const struct xfs_getparents *ppi,
fs/xfs/xfs_trace.h
5690
TP_ARGS(ip, ppi, cur),
fs/xfs/xfs_trace.h
5703
__entry->dev = ip->i_mount->m_super->s_dev;
fs/xfs/xfs_trace.h
5704
__entry->ino = ip->i_ino;
fs/xfs/xfs_trace.h
5726
TP_PROTO(struct xfs_inode *ip, const struct xfs_getparents *ppi, \
fs/xfs/xfs_trace.h
5728
TP_ARGS(ip, ppi, cur))
fs/xfs/xfs_trace.h
5744
__entry->ino = upd->ip ? upd->ip->i_ino : NULLFSINO;
fs/xfs/xfs_trace.h
5779
__entry->ino = upd->ip ? upd->ip->i_ino : NULLFSINO;
fs/xfs/xfs_trace.h
667
TP_PROTO(struct xfs_inode *ip, struct xfs_iext_cursor *cur, int state,
fs/xfs/xfs_trace.h
669
TP_ARGS(ip, cur, state, caller_ip),
fs/xfs/xfs_trace.h
686
ifp = xfs_iext_state_to_fork(ip, state);
fs/xfs/xfs_trace.h
688
__entry->dev = VFS_I(ip)->i_sb->s_dev;
fs/xfs/xfs_trace.h
689
__entry->ino = ip->i_ino;
fs/xfs/xfs_trace.h
715
TP_PROTO(struct xfs_inode *ip, struct xfs_iext_cursor *cur, int state, \
fs/xfs/xfs_trace.h
717
TP_ARGS(ip, cur, state, caller_ip))
fs/xfs/xfs_trans.c
1067
struct xfs_inode *ip,
fs/xfs/xfs_trans.c
1075
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_trans.c
1086
xfs_ilock(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_trans.c
1087
xfs_trans_ijoin(tp, ip, 0);
fs/xfs/xfs_trans.c
1089
error = xfs_qm_dqattach_locked(ip, false);
fs/xfs/xfs_trans.c
1096
error = xfs_trans_reserve_quota_nblks(tp, ip, dblocks, rblocks, force);
fs/xfs/xfs_trans.c
1099
xfs_iunlock(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_trans.c
1100
xfs_blockgc_free_quota(ip, 0);
fs/xfs/xfs_trans.c
1112
xfs_iunlock(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_trans.c
1152
struct xfs_inode *ip,
fs/xfs/xfs_trans.c
1157
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_trans.c
1161
xfs_assert_ilocked(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_trans.c
1167
if (!XFS_IS_QUOTA_ON(mp) || xfs_is_quota_inode(&mp->m_sb, ip->i_ino))
fs/xfs/xfs_trans.c
1173
error = xfs_trans_reserve_quota_nblks(tp, ip, dblocks, rblocks,
fs/xfs/xfs_trans.c
1236
struct xfs_inode *ip,
fs/xfs/xfs_trans.c
1244
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_trans.c
1256
xfs_ilock(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_trans.c
1257
xfs_trans_ijoin(tp, ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_trans.c
1259
if (xfs_is_metadir_inode(ip))
fs/xfs/xfs_trans.c
1262
error = xfs_qm_dqattach_locked(ip, false);
fs/xfs/xfs_trans.c
1275
udqp = (new_udqp != ip->i_udquot) ? new_udqp : NULL;
fs/xfs/xfs_trans.c
1276
gdqp = (new_gdqp != ip->i_gdquot) ? new_gdqp : NULL;
fs/xfs/xfs_trans.c
1277
pdqp = (new_pdqp != ip->i_pdquot) ? new_pdqp : NULL;
fs/xfs/xfs_trans.c
1281
bool isrt = XFS_IS_REALTIME_INODE(ip);
fs/xfs/xfs_trans.c
1287
error = xfs_iread_extents(tp, ip, XFS_DATA_FORK);
fs/xfs/xfs_trans.c
1292
xfs_inode_count_blocks(tp, ip, &dblocks, &rblocks);
fs/xfs/xfs_trans.c
1295
rblocks += ip->i_delayed_blks;
fs/xfs/xfs_trans.c
1297
dblocks += ip->i_delayed_blks;
fs/xfs/xfs_trans.c
1356
struct xfs_inode *ip,
fs/xfs/xfs_trans.c
1362
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_trans.c
1379
xfs_lock_two_inodes(dp, XFS_ILOCK_EXCL, ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_trans.c
1382
xfs_trans_ijoin(tp, ip, 0);
fs/xfs/xfs_trans.c
1391
error = xfs_qm_dqattach_locked(ip, false);
fs/xfs/xfs_trans.c
1406
if (dp != ip)
fs/xfs/xfs_trans.c
1407
xfs_iunlock(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_trans.c
1428
if (dp != ip)
fs/xfs/xfs_trans.c
1429
xfs_iunlock(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_trans.h
253
int xfs_trans_alloc_inode(struct xfs_inode *ip, struct xfs_trans_res *resv,
fs/xfs/xfs_trans.h
256
int xfs_trans_reserve_more_inode(struct xfs_trans *tp, struct xfs_inode *ip,
fs/xfs/xfs_trans.h
262
int xfs_trans_alloc_ichange(struct xfs_inode *ip, struct xfs_dquot *udqp,
fs/xfs/xfs_trans.h
266
struct xfs_inode *ip, unsigned int *dblocks,
fs/xfs/xfs_trans_dquot.c
1000
error = xfs_trans_reserve_quota_bydquots(tp, mp, ip->i_udquot,
fs/xfs/xfs_trans_dquot.c
1001
ip->i_gdquot, ip->i_pdquot, dblocks, 0,
fs/xfs/xfs_trans_dquot.c
1007
error = xfs_trans_reserve_quota_bydquots(tp, mp, ip->i_udquot,
fs/xfs/xfs_trans_dquot.c
1008
ip->i_gdquot, ip->i_pdquot, rblocks, 0,
fs/xfs/xfs_trans_dquot.c
1011
xfs_trans_reserve_quota_bydquots(tp, mp, ip->i_udquot,
fs/xfs/xfs_trans_dquot.c
1012
ip->i_gdquot, ip->i_pdquot, -dblocks, 0,
fs/xfs/xfs_trans_dquot.c
1058
struct xfs_inode *ip,
fs/xfs/xfs_trans_dquot.c
1061
if (XFS_IS_REALTIME_INODE(ip))
fs/xfs/xfs_trans_dquot.c
1062
return xfs_trans_reserve_quota_nblks(NULL, ip, 0, blocks,
fs/xfs/xfs_trans_dquot.c
1064
return xfs_trans_reserve_quota_nblks(NULL, ip, blocks, 0, false);
fs/xfs/xfs_trans_dquot.c
154
struct xfs_inode *ip,
fs/xfs/xfs_trans_dquot.c
159
if (xfs_is_metadir_inode(ip))
fs/xfs/xfs_trans_dquot.c
167
.ino = ip->i_ino,
fs/xfs/xfs_trans_dquot.c
243
xfs_inode_t *ip,
fs/xfs/xfs_trans_dquot.c
250
xfs_is_quota_inode(&mp->m_sb, ip->i_ino) ||
fs/xfs/xfs_trans_dquot.c
251
xfs_is_metadir_inode(ip))
fs/xfs/xfs_trans_dquot.c
254
if (XFS_IS_UQUOTA_ON(mp) && ip->i_udquot)
fs/xfs/xfs_trans_dquot.c
255
xfs_trans_mod_ino_dquot(tp, ip, ip->i_udquot, field, delta);
fs/xfs/xfs_trans_dquot.c
256
if (XFS_IS_GQUOTA_ON(mp) && ip->i_gdquot)
fs/xfs/xfs_trans_dquot.c
257
xfs_trans_mod_ino_dquot(tp, ip, ip->i_gdquot, field, delta);
fs/xfs/xfs_trans_dquot.c
258
if (XFS_IS_PQUOTA_ON(mp) && ip->i_pdquot)
fs/xfs/xfs_trans_dquot.c
259
xfs_trans_mod_ino_dquot(tp, ip, ip->i_pdquot, field, delta);
fs/xfs/xfs_trans_dquot.c
979
struct xfs_inode *ip,
fs/xfs/xfs_trans_dquot.c
984
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_trans_dquot.c
990
if (xfs_is_metadir_inode(ip))
fs/xfs/xfs_trans_dquot.c
993
ASSERT(!xfs_is_quota_inode(&mp->m_sb, ip->i_ino));
fs/xfs/xfs_trans_dquot.c
994
xfs_assert_ilocked(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_verify_media.c
411
struct xfs_inode *ip = XFS_I(file_inode(file));
fs/xfs/xfs_verify_media.c
412
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_verify_media.c
43
struct xfs_inode *ip;
fs/xfs/xfs_verify_media.c
58
error = xfs_iget(mp, cur->bc_tp, rec->rm_owner, 0, 0, &ip);
fs/xfs/xfs_verify_media.c
63
xfs_bmap_mark_sick(ip, is_attr ? XFS_ATTR_FORK : XFS_DATA_FORK);
fs/xfs/xfs_verify_media.c
68
xfs_inode_mark_sick(ip, XFS_SICK_INO_XATTR);
fs/xfs/xfs_verify_media.c
79
fserror_report_data_lost(VFS_I(ip), XFS_FSB_TO_B(mp, fileoff),
fs/xfs/xfs_verify_media.c
83
xfs_irele(ip);
fs/xfs/xfs_zone_alloc.c
241
struct xfs_inode *ip,
fs/xfs/xfs_zone_alloc.c
251
error = xfs_bmapi_read(ip, new->br_startoff, new->br_blockcount, &data,
fs/xfs/xfs_zone_alloc.c
280
trace_xfs_reflink_cow_remap_from(ip, new);
fs/xfs/xfs_zone_alloc.c
281
trace_xfs_reflink_cow_remap_to(ip, &data);
fs/xfs/xfs_zone_alloc.c
283
error = xfs_iext_count_extend(tp, ip, XFS_DATA_FORK,
fs/xfs/xfs_zone_alloc.c
292
xfs_bmap_unmap_extent(tp, ip, XFS_DATA_FORK, &data);
fs/xfs/xfs_zone_alloc.c
293
if (xfs_is_reflink_inode(ip)) {
fs/xfs/xfs_zone_alloc.c
308
xfs_bmap_map_extent(tp, ip, XFS_DATA_FORK, new);
fs/xfs/xfs_zone_alloc.c
312
trace_xfs_reflink_cow_remap_skip(ip, new);
fs/xfs/xfs_zone_alloc.c
319
struct xfs_inode *ip,
fs/xfs/xfs_zone_alloc.c
326
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_zone_alloc.c
348
xfs_ilock(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_zone_alloc.c
349
xfs_trans_ijoin(tp, ip, 0);
fs/xfs/xfs_zone_alloc.c
351
error = xfs_zoned_map_extent(tp, ip, &new, oz, old_startblock);
fs/xfs/xfs_zone_alloc.c
356
xfs_iunlock(ip, XFS_ILOCK_EXCL);
fs/xfs/xfs_zone_alloc.c
630
static inline enum rw_hint xfs_inode_write_hint(struct xfs_inode *ip)
fs/xfs/xfs_zone_alloc.c
632
if (xfs_has_nolifetime(ip->i_mount))
fs/xfs/xfs_zone_alloc.c
634
return VFS_I(ip)->i_write_hint;
fs/xfs/xfs_zone_alloc.c
644
static inline bool xfs_zoned_pack_tight(struct xfs_inode *ip)
fs/xfs/xfs_zone_alloc.c
646
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_zone_alloc.c
654
if (i_size_read(VFS_I(ip)) >= zone_capacity)
fs/xfs/xfs_zone_alloc.c
657
return !inode_is_open_for_write(VFS_I(ip)) &&
fs/xfs/xfs_zone_alloc.c
658
!(ip->i_diflags & XFS_DIFLAG_APPEND);
fs/xfs/xfs_zone_alloc.c
786
struct xfs_inode *ip)
fs/xfs/xfs_zone_alloc.c
791
oz = VFS_I(ip)->i_private;
fs/xfs/xfs_zone_alloc.c
820
struct xfs_inode *ip,
fs/xfs/xfs_zone_alloc.c
826
old_oz = xchg(&VFS_I(ip)->i_private, oz);
fs/xfs/xfs_zone_alloc.c
856
struct xfs_inode *ip = XFS_I(ioend->io_inode);
fs/xfs/xfs_zone_alloc.c
857
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_zone_alloc.c
858
enum rw_hint write_hint = xfs_inode_write_hint(ip);
fs/xfs/xfs_zone_alloc.c
859
bool pack_tight = xfs_zoned_pack_tight(ip);
fs/xfs/xfs_zone_alloc.c
872
*oz = xfs_get_cached_zone(ip);
fs/xfs/xfs_zone_alloc.c
879
xfs_set_cached_zone(ip, *oz);
fs/xfs/xfs_zone_alloc.h
36
int xfs_zoned_end_io(struct xfs_inode *ip, xfs_off_t offset, xfs_off_t count,
fs/xfs/xfs_zone_gc.c
670
struct xfs_inode *ip;
fs/xfs/xfs_zone_gc.c
679
if (!xfs_zone_gc_iter_next(mp, iter, &irec, &ip))
fs/xfs/xfs_zone_gc.c
684
xfs_irele(ip);
fs/xfs/xfs_zone_gc.c
694
chunk->ip = ip;
fs/xfs/xfs_zone_gc.c
731
xfs_irele(chunk->ip);
fs/xfs/xfs_zone_gc.c
778
ihold(VFS_I(chunk->ip));
fs/xfs/xfs_zone_gc.c
779
split_chunk->ip = chunk->ip;
fs/xfs/xfs_zone_gc.c
807
struct xfs_mount *mp = chunk->ip->i_mount;
fs/xfs/xfs_zone_gc.c
835
struct xfs_inode *ip = chunk->ip;
fs/xfs/xfs_zone_gc.c
836
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_zone_gc.c
858
xfs_ilock(ip, iolock);
fs/xfs/xfs_zone_gc.c
859
error = xfs_break_layouts(VFS_I(ip), &iolock, BREAK_UNMAP);
fs/xfs/xfs_zone_gc.c
861
inode_dio_wait(VFS_I(ip));
fs/xfs/xfs_zone_gc.c
862
xfs_iunlock(ip, iolock);
fs/xfs/xfs_zone_gc.c
868
error = xfs_zoned_end_io(ip, chunk->offset, chunk->len,
fs/xfs/xfs_zone_gc.c
89
struct xfs_inode *ip;
include/linux/bottom_half.h
11
static __always_inline void __local_bh_disable_ip(unsigned long ip, unsigned int cnt)
include/linux/bottom_half.h
24
extern void __local_bh_enable_ip(unsigned long ip, unsigned int cnt);
include/linux/bottom_half.h
26
static inline void local_bh_enable_ip(unsigned long ip)
include/linux/bottom_half.h
28
__local_bh_enable_ip(ip, SOFTIRQ_DISABLE_OFFSET);
include/linux/bottom_half.h
9
extern void __local_bh_disable_ip(unsigned long ip, unsigned int cnt);
include/linux/bpf.h
1343
unsigned long ip;
include/linux/bpf.h
1540
bool bpf_has_frame_pointer(unsigned long ip);
include/linux/bpf.h
3858
int bpf_arch_text_poke(void *ip, enum bpf_text_poke_type old_t,
include/linux/bpf.h
3953
int bpf_prog_get_file_line(struct bpf_prog *prog, unsigned long ip, const char **filep,
include/linux/closure.h
166
unsigned long ip;
include/linux/closure.h
225
cl->ip = _THIS_IP_;
include/linux/closure.h
232
cl->ip = _RET_IP_;
include/linux/cper.h
402
u64 ip;
include/linux/cper.h
420
u64 ip;
include/linux/filter.h
1172
void arch_bpf_stack_walk(bool (*consume_fn)(void *cookie, u64 ip, u64 sp, u64 bp), void *cookie);
include/linux/ftrace.h
1048
extern int skip_trace(unsigned long ip);
include/linux/ftrace.h
1053
static inline int skip_trace(unsigned long ip) { return 0; }
include/linux/ftrace.h
1061
static inline unsigned long ftrace_location(unsigned long ip)
include/linux/ftrace.h
1073
#define ftrace_set_filter_ip(ops, ip, remove, reset) ({ -ENODEV; })
include/linux/ftrace.h
179
#define ftrace_regs_set_instruction_pointer(fregs, ip) do { } while (0)
include/linux/ftrace.h
281
typedef void (*ftrace_func_t)(unsigned long ip, unsigned long parent_ip,
include/linux/ftrace.h
408
typedef int (*ftrace_ops_func_t)(struct ftrace_ops *op, unsigned long ip, enum ftrace_ops_cmd cmd);
include/linux/ftrace.h
417
unsigned long ip, unsigned long direct);
include/linux/ftrace.h
536
unsigned long ip;
include/linux/ftrace.h
541
unsigned long ftrace_find_rec_direct(unsigned long ip);
include/linux/ftrace.h
556
static inline unsigned long ftrace_find_rec_direct(unsigned long ip)
include/linux/ftrace.h
63
void arch_ftrace_ops_list_func(unsigned long ip, unsigned long parent_ip);
include/linux/ftrace.h
66
void arch_ftrace_ops_list_func(unsigned long ip, unsigned long parent_ip,
include/linux/ftrace.h
786
unsigned long ip; /* address of mcount call-site */
include/linux/ftrace.h
791
int ftrace_set_filter_ip(struct ftrace_ops *ops, unsigned long ip,
include/linux/ftrace.h
854
unsigned long ftrace_location(unsigned long ip);
include/linux/irqflags.h
25
extern void lockdep_softirqs_on(unsigned long ip);
include/linux/irqflags.h
26
extern void lockdep_softirqs_off(unsigned long ip);
include/linux/irqflags.h
28
extern void lockdep_hardirqs_on(unsigned long ip);
include/linux/irqflags.h
29
extern void lockdep_hardirqs_off(unsigned long ip);
include/linux/irqflags.h
33
static inline void lockdep_softirqs_on(unsigned long ip) { }
include/linux/irqflags.h
34
static inline void lockdep_softirqs_off(unsigned long ip) { }
include/linux/irqflags.h
36
static inline void lockdep_hardirqs_on(unsigned long ip) { }
include/linux/irqflags.h
37
static inline void lockdep_hardirqs_off(unsigned long ip) { }
include/linux/kallsyms.h
169
static inline void print_ip_sym(const char *loglvl, unsigned long ip)
include/linux/kallsyms.h
171
printk("%s[<%px>] %pS\n", loglvl, (void *) ip, (void *) ip);
include/linux/kasan.h
183
unsigned long ip);
include/linux/kasan.h
240
void __kasan_kfree_large(void *ptr, unsigned long ip);
include/linux/kasan.h
288
unsigned long ip);
include/linux/kasan.h
314
unsigned long ip);
include/linux/kasan.h
334
bool __kasan_mempool_poison_object(void *ptr, unsigned long ip);
include/linux/kasan.h
367
void __kasan_mempool_unpoison_object(void *ptr, size_t size, unsigned long ip);
include/linux/kasan.h
398
bool __kasan_check_byte(const void *addr, unsigned long ip);
include/linux/kasan.h
533
bool is_write, unsigned long ip);
include/linux/kcsan-checks.h
146
unsigned long ip;
include/linux/kprobes.h
369
extern void kprobe_ftrace_handler(unsigned long ip, unsigned long parent_ip,
include/linux/lockdep.h
229
struct lockdep_map *nest_lock, unsigned long ip);
include/linux/lockdep.h
231
extern void lock_release(struct lockdep_map *lock, unsigned long ip);
include/linux/lockdep.h
235
unsigned long ip);
include/linux/lockdep.h
257
unsigned long ip);
include/linux/lockdep.h
263
unsigned int subclass, unsigned long ip)
include/linux/lockdep.h
265
lock_set_class(lock, lock->name, lock->key, subclass, ip);
include/linux/lockdep.h
268
extern void lock_downgrade(struct lockdep_map *lock, unsigned long ip);
include/linux/lockdep.h
438
extern void lock_contended(struct lockdep_map *lock, unsigned long ip);
include/linux/lockdep.h
439
extern void lock_acquired(struct lockdep_map *lock, unsigned long ip);
include/linux/lockdep.h
464
#define lock_contended(lockdep_map, ip) do {} while (0)
include/linux/lockdep.h
465
#define lock_acquired(lockdep_map, ip) do {} while (0)
include/linux/lockdep_types.h
196
unsigned long ip;
include/linux/netfilter/ipset/ip_set.h
349
__be32 ip;
include/linux/netfilter/ipset/ip_set.h
350
int ret = ip_set_get_ipaddr4(nla, &ip);
include/linux/netfilter/ipset/ip_set.h
354
*ipaddr = ntohl(ip);
include/linux/netfilter/ipset/pfxlen.h
16
return ip_set_netmask_map[pfxlen].ip;
include/linux/netfilter/ipset/pfxlen.h
28
return (__force u32) ip_set_hostmask_map[pfxlen].ip;
include/linux/netfilter/ipset/pfxlen.h
46
ip6_netmask(union nf_inet_addr *ip, u8 prefix)
include/linux/netfilter/ipset/pfxlen.h
48
ip->ip6[0] &= ip_set_netmask6(prefix)[0];
include/linux/netfilter/ipset/pfxlen.h
49
ip->ip6[1] &= ip_set_netmask6(prefix)[1];
include/linux/netfilter/ipset/pfxlen.h
50
ip->ip6[2] &= ip_set_netmask6(prefix)[2];
include/linux/netfilter/ipset/pfxlen.h
51
ip->ip6[3] &= ip_set_netmask6(prefix)[3];
include/linux/netfilter/nf_conntrack_h323_types.h
12
unsigned int ip;
include/linux/netfilter/nf_conntrack_h323_types.h
17
unsigned int ip;
include/linux/netfilter_ipv4/ip_tables.h
74
struct ipt_ip ip;
include/linux/netpoll.h
18
__be32 ip;
include/linux/objtool_types.h
14
u32 ip;
include/linux/pe.h
256
uint16_t ip; /* initial %ip register */
include/linux/percpu-rwsem.h
169
unsigned long ip)
include/linux/percpu-rwsem.h
171
lock_release(&sem->dep_map, ip);
include/linux/percpu-rwsem.h
175
bool read, unsigned long ip)
include/linux/percpu-rwsem.h
177
lock_acquire(&sem->dep_map, 0, 1, read, 1, NULL, ip);
include/linux/perf_event.h
1330
u64 ip;
include/linux/perf_event.h
1578
static inline void perf_arch_fetch_caller_regs(struct pt_regs *regs, unsigned long ip) { }
include/linux/perf_event.h
1748
static inline int perf_callchain_store_context(struct perf_callchain_entry_ctx *ctx, u64 ip)
include/linux/perf_event.h
1753
entry->ip[entry->nr++] = ip;
include/linux/perf_event.h
1762
static inline int perf_callchain_store(struct perf_callchain_entry_ctx *ctx, u64 ip)
include/linux/perf_event.h
1767
entry->ip[entry->nr++] = ip;
include/linux/perf_event.h
61
u64 ip[]; /* /proc/sys/kernel/perf_event_max_stack */
include/linux/profile.h
40
void profile_hits(int type, void *ip, unsigned int nr_hits);
include/linux/profile.h
45
static inline void profile_hit(int type, void *ip)
include/linux/profile.h
51
profile_hits(type, ip, 1);
include/linux/profile.h
71
static inline void profile_hits(int type, void *ip, unsigned int nr_hits)
include/linux/profile.h
76
static inline void profile_hit(int type, void *ip)
include/linux/pstore.h
214
unsigned long ip;
include/linux/pstore.h
242
rec->ip |= cpu;
include/linux/pstore.h
248
return rec->ip & PSTORE_CPU_IN_IP;
include/linux/qed/qed_iscsi_if.h
50
u32 ip[4];
include/linux/qed/qed_nvmetcp_if.h
34
u32 ip[4];
include/linux/rseq_entry.h
272
unsigned long ip = instruction_pointer(regs);
include/linux/rseq_entry.h
285
if (ip < start_ip)
include/linux/rseq_entry.h
295
if (ip >= cs_end)
include/linux/rseq_entry.h
338
rseq_trace_ip_fixup(ip, start_ip, offset, abort_ip);
include/linux/rseq_entry.h
395
unsigned long ip = instruction_pointer(regs);
include/linux/rseq_entry.h
421
if (ip - start_ip >= offset)
include/linux/rseq_entry.h
461
rseq_trace_ip_fixup(ip, start_ip, offset, abort_ip);
include/linux/rseq_entry.h
54
void __rseq_trace_ip_fixup(unsigned long ip, unsigned long start_ip,
include/linux/rseq_entry.h
63
static inline void rseq_trace_ip_fixup(unsigned long ip, unsigned long start_ip,
include/linux/rseq_entry.h
67
__rseq_trace_ip_fixup(ip, start_ip, offset, abort_ip);
include/linux/rseq_entry.h
72
static inline void rseq_trace_ip_fixup(unsigned long ip, unsigned long start_ip,
include/linux/sched/mm.h
270
extern void __fs_reclaim_acquire(unsigned long ip);
include/linux/sched/mm.h
271
extern void __fs_reclaim_release(unsigned long ip);
include/linux/sched/mm.h
275
static inline void __fs_reclaim_acquire(unsigned long ip) { }
include/linux/sched/mm.h
276
static inline void __fs_reclaim_release(unsigned long ip) { }
include/linux/trace.h
45
int __trace_array_puts(struct trace_array *tr, unsigned long ip,
include/linux/trace.h
50
int trace_array_printk(struct trace_array *tr, unsigned long ip,
include/linux/trace.h
76
int trace_array_printk(struct trace_array *tr, unsigned long ip, const char *fmt, ...)
include/linux/trace_printk.h
111
int __trace_bprintk(unsigned long ip, const char *fmt, ...);
include/linux/trace_printk.h
114
int __trace_printk(unsigned long ip, const char *fmt, ...);
include/linux/trace_printk.h
151
extern int __trace_bputs(unsigned long ip, const char *str);
include/linux/trace_printk.h
152
extern int __trace_puts(unsigned long ip, const char *str);
include/linux/trace_printk.h
174
__ftrace_vbprintk(unsigned long ip, const char *fmt, va_list ap);
include/linux/trace_printk.h
177
__ftrace_vprintk(unsigned long ip, const char *fmt, va_list ap);
include/linux/trace_recursion.h
100
ftrace_record_recursion(ip, pip); \
include/linux/trace_recursion.h
105
# define do_ftrace_record_recursion(ip, pip) do { } while (0)
include/linux/trace_recursion.h
109
# define trace_warn_on_no_rcu(ip) \
include/linux/trace_recursion.h
114
WARN_ONCE(true, "RCU not on for: %pS\n", (void *)ip); \
include/linux/trace_recursion.h
120
# define trace_warn_on_no_rcu(ip) false
include/linux/trace_recursion.h
126
static __always_inline int trace_test_and_set_recursion(unsigned long ip, unsigned long pip,
include/linux/trace_recursion.h
132
if (trace_warn_on_no_rcu(ip))
include/linux/trace_recursion.h
147
do_ftrace_record_recursion(ip, pip);
include/linux/trace_recursion.h
180
static __always_inline int ftrace_test_recursion_trylock(unsigned long ip,
include/linux/trace_recursion.h
183
return trace_test_and_set_recursion(ip, parent_ip, TRACE_FTRACE_START);
include/linux/trace_recursion.h
95
extern void ftrace_record_recursion(unsigned long ip, unsigned long parent_ip);
include/linux/trace_recursion.h
96
# define do_ftrace_record_recursion(ip, pip) \
include/linux/unwind_user_types.h
36
unsigned long ip;
include/net/arp.h
68
int arp_invalidate(struct net_device *dev, __be32 ip, bool force);
include/net/bonding.h
770
static inline int bond_get_targets_ip(__be32 *targets, __be32 ip)
include/net/bonding.h
775
if (targets[i] == ip)
include/net/bonding.h
784
static inline int bond_get_targets_ip6(struct in6_addr *targets, struct in6_addr *ip)
include/net/bonding.h
791
if ((ipv6_addr_equal(&targets[i], ip)) ||
include/net/bonding.h
792
(ipv6_addr_equal(&mcaddr, ip)))
include/net/inetpeer.h
73
static inline void inetpeer_set_addr_v4(struct inetpeer_addr *iaddr, __be32 ip)
include/net/inetpeer.h
75
iaddr->a4.addr = ip;
include/net/ip.h
214
void ip_send_check(struct iphdr *ip);
include/net/ip.h
703
static inline unsigned int ipv4_addr_hash(__be32 ip)
include/net/ip.h
705
return (__force unsigned int) ip;
include/net/ip.h
708
static inline u32 __ipv4_addr_hash(const __be32 ip, const u32 initval)
include/net/ip.h
710
return jhash_1word((__force u32)ip, initval);
include/net/ip_vs.h
107
iphdr->saddr.ip = iph->saddr;
include/net/ip_vs.h
108
iphdr->daddr.ip = iph->daddr;
include/net/ip_vs.h
159
dst->ip = src->ip;
include/net/ip_vs.h
171
dst->ip = src->ip;
include/net/ip_vs.h
184
return a->ip == b->ip;
include/net/ip_vs.h
204
&addr->ip) + 1;
include/net/netfilter/nf_conntrack_tuple.h
90
&t->src.u3.ip, ntohs(t->src.u.all),
include/net/netfilter/nf_conntrack_tuple.h
91
&t->dst.u3.ip, ntohs(t->dst.u.all));
include/net/netfilter/nf_tables_ipv4.h
10
struct iphdr *ip;
include/net/netfilter/nf_tables_ipv4.h
12
ip = ip_hdr(pkt->skb);
include/net/netfilter/nf_tables_ipv4.h
14
pkt->tprot = ip->protocol;
include/net/netfilter/nf_tables_ipv4.h
16
pkt->fragoff = ntohs(ip->frag_off) & IP_OFFSET;
include/net/netfilter/nf_tables_offload.h
51
struct flow_dissector_key_ip ip;
include/rdma/rdmavt_cq.h
56
struct rvt_mmap_info *ip;
include/rdma/rdmavt_qp.h
341
struct rvt_mmap_info *ip;
include/rdma/rdmavt_qp.h
455
struct rvt_mmap_info *ip;
include/rdma/rdmavt_qp.h
927
return ibcq_to_rvtcq(send_cq)->ip ?
include/rdma/rdmavt_qp.h
943
return ibcq_to_rvtcq(send_cq)->ip ?
include/scsi/scsicam.h
17
int scsicam_bios_param(struct gendisk *disk, sector_t capacity, int *ip);
include/soc/mscc/ocelot_vcap.h
540
enum ocelot_vcap_bit ip; /* Protocol address type IP */
include/sound/emu10k1.h
39
#define IP_TO_CP(ip) ((ip == 0) ? 0 : (((0x00001000uL | (ip & 0x00000FFFL)) << (((ip >> 12) & 0x000FL) + 4)) & 0xFFFF0000uL))
include/trace/events/btrfs.h
1491
__field(const void*, ip)
include/trace/events/btrfs.h
1497
__entry->ip = (const void *)IP
include/trace/events/btrfs.h
1501
show_gfp_flags(__entry->mask), __entry->ip)
include/trace/events/btrfs.h
1512
__field(const void*, ip)
include/trace/events/btrfs.h
1517
__entry->ip = (const void *)IP
include/trace/events/btrfs.h
1520
TP_printk("state=%p caller=%pS", __entry->state, __entry->ip)
include/trace/events/exceptions.h
19
__field( unsigned long, ip )
include/trace/events/exceptions.h
25
__entry->ip = instruction_pointer(regs);
include/trace/events/exceptions.h
30
(void *)__entry->address, (void *)__entry->ip,
include/trace/events/ext4.h
1797
__field( unsigned long, ip )
include/trace/events/ext4.h
1806
__entry->ip = IP;
include/trace/events/ext4.h
1816
__entry->revoke_creds, __entry->type, (void *)__entry->ip)
include/trace/events/ext4.h
1828
__field( unsigned long, ip )
include/trace/events/ext4.h
1837
__entry->ip = IP;
include/trace/events/ext4.h
1849
(void *)__entry->ip)
include/trace/events/ext4.h
1859
__field(unsigned long, ip )
include/trace/events/ext4.h
1865
__entry->ip = IP;
include/trace/events/ext4.h
1871
__entry->blocks, (void *)__entry->ip)
include/trace/events/ext4.h
318
__field(unsigned long, ip )
include/trace/events/ext4.h
324
__entry->ip = IP;
include/trace/events/ext4.h
329
(unsigned long) __entry->ino, (void *)__entry->ip)
include/trace/events/lock.h
28
struct lockdep_map *next_lock, unsigned long ip),
include/trace/events/lock.h
30
TP_ARGS(lock, subclass, trylock, read, check, next_lock, ip),
include/trace/events/lock.h
52
TP_PROTO(struct lockdep_map *lock, unsigned long ip),
include/trace/events/lock.h
54
TP_ARGS(lock, ip),
include/trace/events/lock.h
71
TP_PROTO(struct lockdep_map *lock, unsigned long ip),
include/trace/events/lock.h
73
TP_ARGS(lock, ip)
include/trace/events/lock.h
80
TP_PROTO(struct lockdep_map *lock, unsigned long ip),
include/trace/events/lock.h
82
TP_ARGS(lock, ip)
include/trace/events/lock.h
87
TP_PROTO(struct lockdep_map *lock, unsigned long ip),
include/trace/events/lock.h
89
TP_ARGS(lock, ip)
include/trace/events/mce.h
34
__field( u64, ip )
include/trace/events/mce.h
57
__entry->ip = err->m.ip;
include/trace/events/mce.h
80
__entry->cs, __entry->ip,
include/trace/events/module.h
101
TP_PROTO(struct module *mod, unsigned long ip),
include/trace/events/module.h
103
TP_ARGS(mod, ip)
include/trace/events/module.h
109
TP_PROTO(char *name, bool wait, unsigned long ip),
include/trace/events/module.h
111
TP_ARGS(name, wait, ip),
include/trace/events/module.h
114
__field( unsigned long, ip )
include/trace/events/module.h
120
__entry->ip = ip;
include/trace/events/module.h
126
__get_str(name), (int)__entry->wait, (void *)__entry->ip)
include/trace/events/module.h
72
TP_PROTO(struct module *mod, unsigned long ip),
include/trace/events/module.h
74
TP_ARGS(mod, ip),
include/trace/events/module.h
77
__field( unsigned long, ip )
include/trace/events/module.h
83
__entry->ip = ip;
include/trace/events/module.h
89
__get_str(name), (void *)__entry->ip, __entry->refcnt)
include/trace/events/module.h
94
TP_PROTO(struct module *mod, unsigned long ip),
include/trace/events/module.h
96
TP_ARGS(mod, ip)
include/trace/events/preemptirq.h
16
TP_PROTO(unsigned long ip, unsigned long parent_ip),
include/trace/events/preemptirq.h
18
TP_ARGS(ip, parent_ip),
include/trace/events/preemptirq.h
26
__entry->caller_offs = (s32)(ip - (unsigned long)_stext);
include/trace/events/preemptirq.h
37
TP_PROTO(unsigned long ip, unsigned long parent_ip),
include/trace/events/preemptirq.h
38
TP_ARGS(ip, parent_ip));
include/trace/events/preemptirq.h
41
TP_PROTO(unsigned long ip, unsigned long parent_ip),
include/trace/events/preemptirq.h
42
TP_ARGS(ip, parent_ip));
include/trace/events/preemptirq.h
50
TP_PROTO(unsigned long ip, unsigned long parent_ip),
include/trace/events/preemptirq.h
51
TP_ARGS(ip, parent_ip));
include/trace/events/preemptirq.h
54
TP_PROTO(unsigned long ip, unsigned long parent_ip),
include/trace/events/preemptirq.h
55
TP_ARGS(ip, parent_ip));
include/trace/events/rpm.h
100
TP_printk("%pS:%s ret=%d", (void *)__entry->ip, __get_str(name),
include/trace/events/rpm.h
85
TP_PROTO(struct device *dev, unsigned long ip, int ret),
include/trace/events/rpm.h
86
TP_ARGS(dev, ip, ret),
include/trace/events/rpm.h
90
__field( unsigned long, ip )
include/trace/events/rpm.h
96
__entry->ip = ip;
include/trace/events/sock.h
250
__field(unsigned long, ip)
include/trace/events/sock.h
257
__entry->ip = _RET_IP_;
include/trace/events/sock.h
261
__entry->family, __entry->protocol, (void *)__entry->ip)
include/uapi/linux/atmarp.h
39
__be32 ip; /* IP address (act_need only) */
include/uapi/linux/bfs_fs.h
70
#define BFS_NZFILESIZE(ip) \
include/uapi/linux/bfs_fs.h
71
((le32_to_cpu((ip)->i_eoffset) + 1) - le32_to_cpu((ip)->i_sblock) * BFS_BSIZE)
include/uapi/linux/bfs_fs.h
73
#define BFS_FILESIZE(ip) \
include/uapi/linux/bfs_fs.h
74
((ip)->i_sblock == 0 ? 0 : BFS_NZFILESIZE(ip))
include/uapi/linux/bfs_fs.h
76
#define BFS_FILEBLOCKS(ip) \
include/uapi/linux/bfs_fs.h
77
((ip)->i_sblock == 0 ? 0 : (le32_to_cpu((ip)->i_eblock) + 1) - le32_to_cpu((ip)->i_sblock))
include/uapi/linux/bpf.h
1499
__u64 ip;
include/uapi/linux/netfilter.h
74
__be32 ip;
include/uapi/linux/netfilter/nfnetlink_osf.h
71
struct iphdr ip;
include/uapi/linux/netfilter_bridge/ebt_among.h
37
__be32 ip;
include/uapi/linux/netfilter_ipv4/ip_tables.h
107
struct ipt_ip ip;
include/xen/interface/xen-mca.h
345
__u64 ip;
ipc/shm.c
1375
static int put_compat_shm_info(struct shm_info *ip,
ipc/shm.c
1381
info.used_ids = ip->used_ids;
ipc/shm.c
1382
info.shm_tot = ip->shm_tot;
ipc/shm.c
1383
info.shm_rss = ip->shm_rss;
ipc/shm.c
1384
info.shm_swp = ip->shm_swp;
ipc/shm.c
1385
info.swap_attempts = ip->swap_attempts;
ipc/shm.c
1386
info.swap_successes = ip->swap_successes;
kernel/bpf/core.c
3207
int __weak bpf_arch_text_poke(void *ip, enum bpf_text_poke_type old_t,
kernel/bpf/core.c
3234
void __weak arch_bpf_stack_walk(bool (*consume_fn)(void *cookie, u64 ip, u64 sp, u64 bp), void *cookie)
kernel/bpf/core.c
3317
int bpf_prog_get_file_line(struct bpf_prog *prog, unsigned long ip, const char **filep,
kernel/bpf/core.c
3343
if (jited_linfo[i] >= (void *)ip)
kernel/bpf/core.c
3366
static bool find_from_stack_cb(void *cookie, u64 ip, u64 sp, u64 bp)
kernel/bpf/core.c
3377
prog = bpf_prog_ksym_find(ip);
kernel/bpf/core.c
767
bool bpf_has_frame_pointer(unsigned long ip)
kernel/bpf/core.c
774
ksym = bpf_ksym_find(ip);
kernel/bpf/core.c
778
offset = ip - ksym->start;
kernel/bpf/helpers.c
3120
static bool bpf_stack_walker(void *cookie, u64 ip, u64 sp, u64 bp)
kernel/bpf/helpers.c
3131
prog = bpf_prog_ksym_find(ip);
kernel/bpf/stackmap.c
189
u64 ip = READ_ONCE(id_offs[i].ip);
kernel/bpf/stackmap.c
191
if (range_in_vma(prev_vma, ip, ip)) {
kernel/bpf/stackmap.c
196
vma = find_vma(current->mm, ip);
kernel/bpf/stackmap.c
204
id_offs[i].offset = (vma->vm_pgoff << PAGE_SHIFT) + ip - vma->vm_start;
kernel/bpf/stackmap.c
224
entry->nr = stack_trace_save_tsk(task, (unsigned long *)entry->ip,
kernel/bpf/stackmap.c
232
unsigned long *from = (unsigned long *) entry->ip;
kernel/bpf/stackmap.c
233
u64 *to = entry->ip;
kernel/bpf/stackmap.c
267
ips = trace->ip + skip;
kernel/bpf/stackmap.c
288
id_offs[i].ip = ips[i];
kernel/bpf/stackmap.c
361
if (trace->ip[nr_kernel] == PERF_CONTEXT_USER)
kernel/bpf/stackmap.c
485
ips = trace->ip + skip;
kernel/bpf/stackmap.c
491
id_offs[i].ip = ips[i];
kernel/bpf/stream.c
362
static bool dump_stack_cb(void *cookie, u64 ip, u64 sp, u64 bp)
kernel/bpf/stream.c
370
prog = bpf_prog_ksym_find(ip);
kernel/bpf/stream.c
373
ret = bpf_prog_get_file_line(prog, ip, &file, &line, &num);
kernel/bpf/stream.c
377
(void *)(long)ip, line, file, num);
kernel/bpf/stream.c
381
ctxp->err = bpf_stream_stage_printk(ctxp->ss, "%pS\n", (void *)(long)ip);
kernel/bpf/trampoline.c
197
unsigned long ip, addr = (unsigned long) ptr;
kernel/bpf/trampoline.c
200
ip = ftrace_location(tr->ip);
kernel/bpf/trampoline.c
201
if (!ip)
kernel/bpf/trampoline.c
208
if (!add_ftrace_hash_entry_direct(hash, ip, addr)) {
kernel/bpf/trampoline.c
284
ret = ftrace_set_filter_ip(ops, tr->ip, 0, 1);
kernel/bpf/trampoline.c
331
static struct bpf_trampoline *bpf_trampoline_lookup(u64 key, unsigned long ip)
kernel/bpf/trampoline.c
355
tr->ip = ftrace_location(ip);
kernel/bpf/trampoline.c
359
head = &trampoline_ip_table[hash_64(tr->ip, TRAMPOLINE_HASH_BITS)];
kernel/bpf/trampoline.c
37
static struct bpf_trampoline *direct_ops_ip_lookup(struct ftrace_ops *ops, unsigned long ip)
kernel/bpf/trampoline.c
374
void *ip = tr->func.addr;
kernel/bpf/trampoline.c
386
return bpf_arch_text_poke(ip, old_t, new_t, old_addr, new_addr);
kernel/bpf/trampoline.c
420
void *ip = tr->func.addr;
kernel/bpf/trampoline.c
424
faddr = ftrace_location((unsigned long)ip);
kernel/bpf/trampoline.c
43
head_ip = &trampoline_ip_table[hash_64(ip, TRAMPOLINE_HASH_BITS)];
kernel/bpf/trampoline.c
45
if (tr->ip == ip)
kernel/bpf/trampoline.c
54
static struct bpf_trampoline *direct_ops_ip_lookup(struct ftrace_ops *ops, unsigned long ip)
kernel/bpf/trampoline.c
60
static int bpf_tramp_ftrace_ops_func(struct ftrace_ops *ops, unsigned long ip,
kernel/bpf/trampoline.c
66
tr = direct_ops_ip_lookup(ops, ip);
kernel/bpf/verifier.c
19645
u32 ip)
kernel/bpf/verifier.c
19647
u16 live_regs = env->insn_aux_data[ip].live_regs_before;
kernel/bpf/verifier.c
19671
int i, ip;
kernel/bpf/verifier.c
19676
ip = frame_insn_idx(st, i);
kernel/bpf/verifier.c
19677
clean_func_state(env, st->frame[i], ip);
kernel/events/callchain.c
192
cur_ip = &entry->ip[start_entry_idx];
kernel/events/callchain.c
193
last_ip = &entry->ip[entry->nr - 1];
kernel/events/core.c
2056
size += sizeof(data->ip);
kernel/events/core.c
8195
perf_output_put(handle, data->ip);
kernel/events/core.c
8567
data->ip = perf_instruction_pointer(event, regs);
kernel/events/core.c
8697
data->code_page_size = perf_get_page_size(data->ip);
kernel/kcov.c
197
static notrace unsigned long canonicalize_ip(unsigned long ip)
kernel/kcov.c
200
ip -= kaslr_offset();
kernel/kcov.c
202
return ip;
kernel/kcov.c
213
unsigned long ip = canonicalize_ip(_RET_IP_);
kernel/kcov.c
233
area[pos] = ip;
kernel/kcov.c
239
static void notrace write_comp_data(u64 type, u64 arg1, u64 arg2, u64 ip)
kernel/kcov.c
249
ip = canonicalize_ip(ip);
kernel/kcov.c
270
area[start_index + 3] = ip;
kernel/kcsan/core.c
1131
reorder_access->type, reorder_access->ip);
kernel/kcsan/core.c
210
check_access(const volatile void *ptr, size_t size, int type, unsigned long ip);
kernel/kcsan/core.c
224
scoped_access->type, scoped_access->ip);
kernel/kcsan/core.c
402
int type, unsigned long ip)
kernel/kcsan/core.c
414
reorder_access->type == type && reorder_access->ip == ip;
kernel/kcsan/core.c
419
int type, unsigned long ip)
kernel/kcsan/core.c
436
reorder_access->ip = ip;
kernel/kcsan/core.c
458
unsigned long ip,
kernel/kcsan/core.c
483
if (ctx->access_mask && !find_reorder_access(ctx, ptr, size, type, ip))
kernel/kcsan/core.c
507
kcsan_report_set_info(ptr, size, type, ip, watchpoint - watchpoints);
kernel/kcsan/core.c
527
kcsan_setup_watchpoint(const volatile void *ptr, size_t size, int type, unsigned long ip)
kernel/kcsan/core.c
570
is_reorder_access = find_reorder_access(ctx, ptr, size, type, ip);
kernel/kcsan/core.c
678
kcsan_report_known_origin(ptr, size, type, ip,
kernel/kcsan/core.c
689
kcsan_report_unknown_origin(ptr, size, type, ip,
kernel/kcsan/core.c
713
set_reorder_access(ctx, ptr, size, type, ip);
kernel/kcsan/core.c
719
check_access(const volatile void *ptr, size_t size, int type, unsigned long ip)
kernel/kcsan/core.c
748
kcsan_found_watchpoint(ptr, size, type, ip, watchpoint, encoded_watchpoint);
kernel/kcsan/core.c
753
kcsan_setup_watchpoint(ptr, size, type, ip);
kernel/kcsan/core.c
767
ip = reorder_access->ip;
kernel/kcsan/core.c
920
sa->ip = _RET_IP_;
kernel/kcsan/core.c
952
check_access(sa->ptr, sa->size, sa->type, sa->ip);
kernel/kcsan/kcsan.h
124
unsigned long ip, int watchpoint_idx);
kernel/kcsan/kcsan.h
132
unsigned long ip, enum kcsan_value_change value_change,
kernel/kcsan/kcsan.h
140
unsigned long ip, u64 old, u64 new, u64 mask);
kernel/kcsan/report.c
316
replace_stack_entry(unsigned long stack_entries[], int num_entries, unsigned long ip,
kernel/kcsan/report.c
323
if (kallsyms_lookup_size_offset(ip, &symbolsize, &offset))
kernel/kcsan/report.c
324
target_func = ip - offset;
kernel/kcsan/report.c
337
stack_entries[skip] = ip;
kernel/kcsan/report.c
344
WARN_ONCE(1, "Cannot find frame for %pS in stack trace", (void *)ip);
kernel/kcsan/report.c
349
sanitize_stack_entries(unsigned long stack_entries[], int num_entries, unsigned long ip,
kernel/kcsan/report.c
35
unsigned long ip;
kernel/kcsan/report.c
352
return ip ? replace_stack_entry(stack_entries, num_entries, ip, replaced) :
kernel/kcsan/report.c
400
int skipnr = sanitize_stack_entries(stack_entries, num_stack_entries, ai->ip, &reordered_to);
kernel/kcsan/report.c
415
other_info->ai.ip, &other_reordered_to);
kernel/kcsan/report.c
643
int access_type, unsigned long ip)
kernel/kcsan/report.c
652
.ip = (access_type & KCSAN_ACCESS_SCOPED) ? ip : 0,
kernel/kcsan/report.c
657
unsigned long ip, int watchpoint_idx)
kernel/kcsan/report.c
659
const struct access_info ai = prepare_access_info(ptr, size, access_type, ip);
kernel/kcsan/report.c
672
unsigned long ip, enum kcsan_value_change value_change,
kernel/kcsan/report.c
675
const struct access_info ai = prepare_access_info(ptr, size, access_type, ip);
kernel/kcsan/report.c
706
unsigned long ip, u64 old, u64 new, u64 mask)
kernel/kcsan/report.c
708
const struct access_info ai = prepare_access_info(ptr, size, access_type, ip);
kernel/livepatch/patch.c
40
static void notrace klp_ftrace_handler(unsigned long ip,
kernel/livepatch/patch.c
58
bit = ftrace_test_recursion_trylock(ip, parent_ip);
kernel/locking/lockdep.c
257
static int lock_point(unsigned long points[], unsigned long ip)
kernel/locking/lockdep.c
263
points[i] = ip;
kernel/locking/lockdep.c
266
if (points[i] == ip)
kernel/locking/lockdep.c
4415
void noinstr lockdep_hardirqs_on(unsigned long ip)
kernel/locking/lockdep.c
4473
trace->hardirq_enable_ip = ip;
kernel/locking/lockdep.c
4482
void noinstr lockdep_hardirqs_off(unsigned long ip)
kernel/locking/lockdep.c
4512
trace->hardirq_disable_ip = ip;
kernel/locking/lockdep.c
4524
void lockdep_softirqs_on(unsigned long ip)
kernel/locking/lockdep.c
4548
trace->softirq_enable_ip = ip;
kernel/locking/lockdep.c
4564
void lockdep_softirqs_off(unsigned long ip)
kernel/locking/lockdep.c
4582
trace->softirq_disable_ip = ip;
kernel/locking/lockdep.c
5079
struct lockdep_map *nest_lock, unsigned long ip,
kernel/locking/lockdep.c
5171
hlock->acquire_ip = ip;
kernel/locking/lockdep.c
5274
unsigned long ip)
kernel/locking/lockdep.c
5292
print_ip_sym(KERN_WARNING, ip);
kernel/locking/lockdep.c
5410
unsigned long ip)
kernel/locking/lockdep.c
5431
print_unlock_imbalance_bug(curr, lock, ip);
kernel/locking/lockdep.c
5457
static int __lock_downgrade(struct lockdep_map *lock, unsigned long ip)
kernel/locking/lockdep.c
5477
print_unlock_imbalance_bug(curr, lock, ip);
kernel/locking/lockdep.c
5486
hlock->acquire_ip = ip;
kernel/locking/lockdep.c
5511
__lock_release(struct lockdep_map *lock, unsigned long ip)
kernel/locking/lockdep.c
5527
print_unlock_imbalance_bug(curr, lock, ip);
kernel/locking/lockdep.c
5537
print_unlock_imbalance_bug(curr, lock, ip);
kernel/locking/lockdep.c
5735
unsigned long ip)
kernel/locking/lockdep.c
5745
if (__lock_set_class(lock, name, key, subclass, ip))
kernel/locking/lockdep.c
5752
void lock_downgrade(struct lockdep_map *lock, unsigned long ip)
kernel/locking/lockdep.c
5762
if (__lock_downgrade(lock, ip))
kernel/locking/lockdep.c
5827
struct lockdep_map *nest_lock, unsigned long ip)
kernel/locking/lockdep.c
5831
trace_lock_acquire(lock, subclass, trylock, read, check, nest_lock, ip);
kernel/locking/lockdep.c
5849
hlock.acquire_ip = ip;
kernel/locking/lockdep.c
5869
irqs_disabled_flags(flags), nest_lock, ip, 0, 0, 0);
kernel/locking/lockdep.c
5875
void lock_release(struct lockdep_map *lock, unsigned long ip)
kernel/locking/lockdep.c
5879
trace_lock_release(lock, ip);
kernel/locking/lockdep.c
5889
if (__lock_release(lock, ip))
kernel/locking/lockdep.c
5905
int check, struct lockdep_map *nest_lock, unsigned long ip)
kernel/locking/lockdep.c
5917
irqs_disabled_flags(flags), nest_lock, ip, 0, 0, 1);
kernel/locking/lockdep.c
6006
unsigned long ip)
kernel/locking/lockdep.c
6024
print_ip_sym(KERN_WARNING, ip);
kernel/locking/lockdep.c
6036
__lock_contended(struct lockdep_map *lock, unsigned long ip)
kernel/locking/lockdep.c
6057
print_lock_contention_bug(curr, lock, ip);
kernel/locking/lockdep.c
6066
contention_point = lock_point(hlock_class(hlock)->contention_point, ip);
kernel/locking/lockdep.c
6068
lock->ip);
kernel/locking/lockdep.c
6080
__lock_acquired(struct lockdep_map *lock, unsigned long ip)
kernel/locking/lockdep.c
6127
lock->ip = ip;
kernel/locking/lockdep.c
6130
void lock_contended(struct lockdep_map *lock, unsigned long ip)
kernel/locking/lockdep.c
6134
trace_lock_contended(lock, ip);
kernel/locking/lockdep.c
6142
__lock_contended(lock, ip);
kernel/locking/lockdep.c
6148
void lock_acquired(struct lockdep_map *lock, unsigned long ip)
kernel/locking/lockdep.c
6152
trace_lock_acquired(lock, ip);
kernel/locking/lockdep.c
6160
__lock_acquired(lock, ip);
kernel/locking/lockdep_proc.c
536
char ip[32];
kernel/locking/lockdep_proc.c
544
snprintf(ip, sizeof(ip), "[<%p>]",
kernel/locking/lockdep_proc.c
548
ip, (void *)class->contention_point[i]);
kernel/locking/lockdep_proc.c
551
char ip[32];
kernel/locking/lockdep_proc.c
559
snprintf(ip, sizeof(ip), "[<%p>]",
kernel/locking/lockdep_proc.c
563
ip, (void *)class->contending_point[i]);
kernel/locking/mutex.c
528
static noinline void __sched __mutex_unlock_slowpath(struct mutex *lock, unsigned long ip);
kernel/locking/mutex.c
579
struct lockdep_map *nest_lock, unsigned long ip,
kernel/locking/mutex.c
614
mutex_acquire_nest(&lock->dep_map, subclass, 0, nest_lock, ip);
kernel/locking/mutex.c
620
lock_acquired(&lock->dep_map, ip);
kernel/locking/mutex.c
644
lock_contended(&lock->dep_map, ip);
kernel/locking/mutex.c
748
lock_acquired(&lock->dep_map, ip);
kernel/locking/mutex.c
767
mutex_release(&lock->dep_map, ip);
kernel/locking/mutex.c
774
struct lockdep_map *nest_lock, unsigned long ip)
kernel/locking/mutex.c
776
return __mutex_lock_common(lock, state, subclass, nest_lock, ip, NULL, false);
kernel/locking/mutex.c
781
unsigned long ip, struct ww_acquire_ctx *ww_ctx)
kernel/locking/mutex.c
783
return __mutex_lock_common(lock, state, subclass, NULL, ip, ww_ctx, true);
kernel/locking/mutex.c
931
static noinline void __sched __mutex_unlock_slowpath(struct mutex *lock, unsigned long ip)
kernel/locking/mutex.c
938
mutex_release(&lock->dep_map, ip);
kernel/locking/rtmutex_api.c
528
unsigned long ip)
kernel/locking/rtmutex_api.c
533
mutex_acquire_nest(&lock->dep_map, subclass, 0, nest_lock, ip);
kernel/locking/rtmutex_api.c
536
mutex_release(&lock->dep_map, ip);
kernel/locking/rtmutex_api.c
538
lock_acquired(&lock->dep_map, ip);
kernel/locking/ww_rt_mutex.c
39
unsigned int state, unsigned long ip)
kernel/locking/ww_rt_mutex.c
63
mutex_acquire_nest(&rtm->dep_map, 0, 0, nest_lock, ip);
kernel/locking/ww_rt_mutex.c
74
mutex_release(&rtm->dep_map, ip);
kernel/printk/printk.c
318
static int __down_trylock_console_sem(unsigned long ip)
kernel/printk/printk.c
334
mutex_acquire(&console_lock_dep_map, 0, 1, ip);
kernel/printk/printk.c
339
static void __up_console_sem(unsigned long ip)
kernel/printk/printk.c
343
mutex_release(&console_lock_dep_map, ip);
kernel/rseq.c
120
void __rseq_trace_ip_fixup(unsigned long ip, unsigned long start_ip,
kernel/rseq.c
123
trace_rseq_ip_fixup(ip, start_ip, offset, abort_ip);
kernel/sched/core.c
2078
unsigned long ip = 0;
kernel/sched/core.c
2089
ip = __get_wchan(p);
kernel/sched/core.c
2092
return ip;
kernel/sched/core.c
5747
unsigned long ip = get_lock_parent_ip();
kernel/sched/core.c
5749
current->preempt_disable_ip = ip;
kernel/sched/core.c
5751
trace_preempt_off(CALLER_ADDR0, ip);
kernel/sched/core.c
8821
static void print_preempt_disable_ip(int preempt_offset, unsigned long ip)
kernel/sched/core.c
8830
print_ip_sym(KERN_ERR, ip);
kernel/signal.c
1235
pr_info("code at %08lx: ", regs->ip);
kernel/signal.c
1241
if (get_user(insn, (unsigned char *)(regs->ip + i)))
kernel/softirq.c
156
void __local_bh_disable_ip(unsigned long ip, unsigned int cnt)
kernel/softirq.c
194
lockdep_softirqs_off(ip);
kernel/softirq.c
209
lockdep_softirqs_off(ip);
kernel/softirq.c
258
void __local_bh_enable_ip(unsigned long ip, unsigned int cnt)
kernel/softirq.c
371
void __local_bh_disable_ip(unsigned long ip, unsigned int cnt)
kernel/softirq.c
390
lockdep_softirqs_off(ip);
kernel/softirq.c
427
void __local_bh_enable_ip(unsigned long ip, unsigned int cnt)
kernel/softirq.c
438
lockdep_softirqs_on(ip);
kernel/trace/bpf_trace.c
2321
unsigned long ip = ftrace_get_symaddr(fentry_ip);
kernel/trace/bpf_trace.c
2323
return ip ? : fentry_ip;
kernel/trace/fprobe.c
226
static inline int __fprobe_handler(unsigned long ip, unsigned long parent_ip,
kernel/trace/fprobe.c
233
return fp->entry_handler(fp, ip, parent_ip, fregs, data);
kernel/trace/fprobe.c
236
static inline int __fprobe_kprobe_handler(unsigned long ip, unsigned long parent_ip,
kernel/trace/fprobe.c
253
ret = __fprobe_handler(ip, parent_ip, fp, fregs, data);
kernel/trace/fprobe.c
260
static void fprobe_ftrace_entry(unsigned long ip, unsigned long parent_ip,
kernel/trace/fprobe.c
268
bit = ftrace_test_recursion_trylock(ip, parent_ip);
kernel/trace/fprobe.c
278
head = rhltable_lookup(&fprobe_ip_table, &ip, fprobe_rht_params);
kernel/trace/fprobe.c
281
if (node->addr != ip)
kernel/trace/fprobe.c
288
__fprobe_kprobe_handler(ip, parent_ip, fp, fregs, NULL);
kernel/trace/fprobe.c
290
__fprobe_handler(ip, parent_ip, fp, fregs, NULL);
kernel/trace/ftrace.c
1156
ftrace_hash_key(struct ftrace_hash *hash, unsigned long ip)
kernel/trace/ftrace.c
1159
return hash_long(ip, hash->size_bits);
kernel/trace/ftrace.c
1166
__ftrace_lookup_ip(struct ftrace_hash *hash, unsigned long ip)
kernel/trace/ftrace.c
1172
key = ftrace_hash_key(hash, ip);
kernel/trace/ftrace.c
1176
if (entry->ip == ip)
kernel/trace/ftrace.c
1193
ftrace_lookup_ip(struct ftrace_hash *hash, unsigned long ip)
kernel/trace/ftrace.c
1198
return __ftrace_lookup_ip(hash, ip);
kernel/trace/ftrace.c
1207
key = ftrace_hash_key(hash, entry->ip);
kernel/trace/ftrace.c
1214
add_ftrace_hash_entry_direct(struct ftrace_hash *hash, unsigned long ip, unsigned long direct)
kernel/trace/ftrace.c
1222
entry->ip = ip;
kernel/trace/ftrace.c
1230
add_hash_entry(struct ftrace_hash *hash, unsigned long ip)
kernel/trace/ftrace.c
1232
return add_ftrace_hash_entry_direct(hash, ip, 0);
kernel/trace/ftrace.c
129
void ftrace_ops_list_func(unsigned long ip, unsigned long parent_ip,
kernel/trace/ftrace.c
1407
if (add_ftrace_hash_entry_direct(new_hash, entry->ip, entry->direct) == NULL)
kernel/trace/ftrace.c
141
static void ftrace_ops_nop_func(unsigned long ip, unsigned long parent_ip,
kernel/trace/ftrace.c
1540
static bool hash_contains_ip(unsigned long ip,
kernel/trace/ftrace.c
1550
__ftrace_lookup_ip(hash->filter_hash, ip)) &&
kernel/trace/ftrace.c
1552
!__ftrace_lookup_ip(hash->notrace_hash, ip));
kernel/trace/ftrace.c
1568
ftrace_ops_test(struct ftrace_ops *ops, unsigned long ip, void *regs)
kernel/trace/ftrace.c
1586
if (hash_contains_ip(ip, &hash))
kernel/trace/ftrace.c
1614
if (key->flags < rec->ip)
kernel/trace/ftrace.c
1616
if (key->ip >= rec->ip + MCOUNT_INSN_SIZE)
kernel/trace/ftrace.c
1627
key.ip = start;
kernel/trace/ftrace.c
1632
end < pg->records[0].ip ||
kernel/trace/ftrace.c
1633
start >= (pg->records[pg->index - 1].ip + MCOUNT_INSN_SIZE))
kernel/trace/ftrace.c
1659
unsigned long ip = 0;
kernel/trace/ftrace.c
1664
ip = rec->ip;
kernel/trace/ftrace.c
1667
return ip;
kernel/trace/ftrace.c
1679
unsigned long ftrace_location(unsigned long ip)
kernel/trace/ftrace.c
1685
loc = ftrace_location_range(ip, ip);
kernel/trace/ftrace.c
1687
if (!kallsyms_lookup_size_offset(ip, &size, &offset))
kernel/trace/ftrace.c
1692
loc = ftrace_location_range(ip, ip + size - 1);
kernel/trace/ftrace.c
172
static void ftrace_pid_func(unsigned long ip, unsigned long parent_ip,
kernel/trace/ftrace.c
1726
if (ftrace_ops_test(ops, rec->ip, rec)) {
kernel/trace/ftrace.c
1803
if (!notrace_hash || !ftrace_lookup_ip(notrace_hash, rec->ip))
kernel/trace/ftrace.c
1806
in_hash = !!ftrace_lookup_ip(hash, rec->ip);
kernel/trace/ftrace.c
1807
in_notrace_hash = !!ftrace_lookup_ip(notrace_hash, rec->ip);
kernel/trace/ftrace.c
187
op->saved_func(ip, parent_ip, op, fregs);
kernel/trace/ftrace.c
2046
in_old = !!ftrace_lookup_ip(old_hash, rec->ip);
kernel/trace/ftrace.c
2047
in_new = !!ftrace_lookup_ip(new_hash, rec->ip);
kernel/trace/ftrace.c
2078
ret = ops->ops_func(ops, rec->ip, FTRACE_OPS_CMD_ENABLE_SHARE_IPMODIFY_SELF);
kernel/trace/ftrace.c
2103
in_old = !!ftrace_lookup_ip(old_hash, rec->ip);
kernel/trace/ftrace.c
2104
in_new = !!ftrace_lookup_ip(new_hash, rec->ip);
kernel/trace/ftrace.c
2202
unsigned long ip = rec ? rec->ip : 0;
kernel/trace/ftrace.c
2209
print_ip_sym(KERN_INFO, ip);
kernel/trace/ftrace.c
2213
print_ip_sym(KERN_INFO, ip);
kernel/trace/ftrace.c
2214
print_ip_ins(" actual: ", (unsigned char *)ip);
kernel/trace/ftrace.c
2223
print_ip_sym(KERN_INFO, ip);
kernel/trace/ftrace.c
2227
print_ip_sym(KERN_INFO, ip);
kernel/trace/ftrace.c
2250
ip = ftrace_get_addr_curr(rec);
kernel/trace/ftrace.c
2251
pr_cont("\n expected tramp: %lx\n", ip);
kernel/trace/ftrace.c
2457
unsigned long ip = rec->ip;
kernel/trace/ftrace.c
2464
if (hash_contains_ip(ip, op->func_hash))
kernel/trace/ftrace.c
2475
unsigned long ip = rec->ip;
kernel/trace/ftrace.c
2482
if (hash_contains_ip(ip, op->func_hash))
kernel/trace/ftrace.c
2493
unsigned long ip = rec->ip;
kernel/trace/ftrace.c
2500
if (hash_contains_ip(ip, op->func_hash))
kernel/trace/ftrace.c
2511
unsigned long ip = rec->ip;
kernel/trace/ftrace.c
2520
if (hash_contains_ip(ip, &removed_ops->old_hash))
kernel/trace/ftrace.c
2561
hash_contains_ip(ip, &op->old_hash))
kernel/trace/ftrace.c
2569
hash_contains_ip(ip, op->func_hash))
kernel/trace/ftrace.c
2581
unsigned long ip = rec->ip;
kernel/trace/ftrace.c
2585
if (hash_contains_ip(ip, op->func_hash))
kernel/trace/ftrace.c
2596
unsigned long ip = rec->ip;
kernel/trace/ftrace.c
2600
if (hash_contains_ip(ip, op->func_hash)) {
kernel/trace/ftrace.c
2620
unsigned long ftrace_find_rec_direct(unsigned long ip)
kernel/trace/ftrace.c
2624
entry = __ftrace_lookup_ip(direct_functions, ip);
kernel/trace/ftrace.c
2631
static void call_direct_funcs(unsigned long ip, unsigned long pip,
kernel/trace/ftrace.c
2637
addr = ftrace_find_rec_direct(ip);
kernel/trace/ftrace.c
2665
addr = ftrace_find_rec_direct(rec->ip);
kernel/trace/ftrace.c
2676
(void *)rec->ip, (void *)rec->ip, rec->flags);
kernel/trace/ftrace.c
2706
addr = ftrace_find_rec_direct(rec->ip);
kernel/trace/ftrace.c
2717
(void *)rec->ip, (void *)rec->ip);
kernel/trace/ftrace.c
3205
(void *)rec->ip, rec->flags);
kernel/trace/ftrace.c
3298
if (!__ftrace_lookup_ip(*hash, entry->ip) &&
kernel/trace/ftrace.c
3299
add_hash_entry(*hash, entry->ip) == NULL)
kernel/trace/ftrace.c
3323
if (!__ftrace_lookup_ip(notrace_hash, entry->ip))
kernel/trace/ftrace.c
3358
if (__ftrace_lookup_ip(new_hash2, entry->ip) &&
kernel/trace/ftrace.c
3359
add_hash_entry(*hash, entry->ip) == NULL)
kernel/trace/ftrace.c
3389
if (!__ftrace_lookup_ip(B, entry->ip))
kernel/trace/ftrace.c
4090
return probe_ops->print(m, probe_entry->ip, probe_ops, probe->data);
kernel/trace/ftrace.c
4092
seq_printf(m, "%ps:%ps\n", (void *)probe_entry->ip,
kernel/trace/ftrace.c
417
unsigned long ip;
kernel/trace/ftrace.c
4192
!ftrace_lookup_ip(iter->hash, rec->ip)) ||
kernel/trace/ftrace.c
4344
ret = kallsyms_lookup(rec->ip, NULL, &offset, NULL, str);
kernel/trace/ftrace.c
4393
static int print_rec(struct seq_file *m, unsigned long ip)
kernel/trace/ftrace.c
4400
ret = kallsyms_lookup(ip, NULL, &offset, &modname, str);
kernel/trace/ftrace.c
4419
static inline int print_rec(struct seq_file *m, unsigned long ip)
kernel/trace/ftrace.c
4421
seq_printf(m, "%ps", (void *)ip);
kernel/trace/ftrace.c
4433
hash_contains_ip(rec->ip, subops->func_hash)))
kernel/trace/ftrace.c
4487
seq_printf(m, "%lx ", rec->ip);
kernel/trace/ftrace.c
4489
if (print_rec(m, rec->ip)) {
kernel/trace/ftrace.c
4535
direct = ftrace_find_rec_direct(rec->ip);
kernel/trace/ftrace.c
4842
entry = ftrace_lookup_ip(hash, rec->ip);
kernel/trace/ftrace.c
4853
if (add_hash_entry(hash, rec->ip) == NULL)
kernel/trace/ftrace.c
4885
static int lookup_ip(unsigned long ip, char **modname, char *str)
kernel/trace/ftrace.c
4889
kallsyms_lookup(ip, NULL, &offset, modname, str);
kernel/trace/ftrace.c
4895
static int lookup_ip(unsigned long ip, char **modname, char *str)
kernel/trace/ftrace.c
4897
kallsyms_lookup(ip, NULL, NULL, modname, str);
kernel/trace/ftrace.c
4909
if (lookup_ip(rec->ip, &modname, str)) {
kernel/trace/ftrace.c
5244
static void function_trace_probe_call(unsigned long ip, unsigned long parent_ip,
kernel/trace/ftrace.c
5259
probe_ops->func(ip, parent_ip, probe->tr, probe_ops, probe->data);
kernel/trace/ftrace.c
5306
unsigned long ip)
kernel/trace/ftrace.c
5311
entry = ftrace_lookup_ip(&mapper->hash, ip);
kernel/trace/ftrace.c
5328
unsigned long ip, void *data)
kernel/trace/ftrace.c
5333
entry = ftrace_lookup_ip(&mapper->hash, ip);
kernel/trace/ftrace.c
5341
map->entry.ip = ip;
kernel/trace/ftrace.c
5360
unsigned long ip)
kernel/trace/ftrace.c
5366
entry = ftrace_lookup_ip(&mapper->hash, ip);
kernel/trace/ftrace.c
5520
if (ftrace_lookup_ip(old_hash, entry->ip))
kernel/trace/ftrace.c
5529
entry->ip, data,
kernel/trace/ftrace.c
5582
if (ftrace_lookup_ip(old_hash, entry->ip))
kernel/trace/ftrace.c
5584
probe_ops->free(probe_ops, tr, entry->ip, probe->data);
kernel/trace/ftrace.c
561
if (core_kernel_text(rec->ip)) {
kernel/trace/ftrace.c
563
offset = rec->ip - (unsigned long)_text;
kernel/trace/ftrace.c
5665
kallsyms_lookup(entry->ip, NULL, NULL,
kernel/trace/ftrace.c
568
mod = __module_text_address(rec->ip);
kernel/trace/ftrace.c
5703
probe_ops->free(probe_ops, tr, entry->ip, probe->data);
kernel/trace/ftrace.c
572
offset = rec->ip - (unsigned long)mod->mem[MOD_TEXT].base;
kernel/trace/ftrace.c
579
kallsyms_lookup(rec->ip, NULL, NULL, NULL, str);
kernel/trace/ftrace.c
5856
__ftrace_match_addr(struct ftrace_hash *hash, unsigned long ip, int remove)
kernel/trace/ftrace.c
5860
ip = ftrace_location(ip);
kernel/trace/ftrace.c
5861
if (!ip)
kernel/trace/ftrace.c
5865
entry = ftrace_lookup_ip(hash, ip);
kernel/trace/ftrace.c
5870
} else if (__ftrace_lookup_ip(hash, ip) != NULL) {
kernel/trace/ftrace.c
5875
entry = add_hash_entry(hash, ip);
kernel/trace/ftrace.c
5997
del = __ftrace_lookup_ip(direct_functions, entry->ip);
kernel/trace/ftrace.c
6070
if (ftrace_find_rec_direct(entry->ip))
kernel/trace/ftrace.c
6090
new = add_hash_entry(new_hash, entry->ip);
kernel/trace/ftrace.c
6101
new = add_hash_entry(new_hash, entry->ip);
kernel/trace/ftrace.c
6213
entry = __ftrace_lookup_ip(direct_functions, iter->ip);
kernel/trace/ftrace.c
6320
if (add_ftrace_hash_entry_direct(add, entry->ip, entry->direct) == NULL) {
kernel/trace/ftrace.c
6361
if (__ftrace_lookup_ip(direct_functions, entry->ip))
kernel/trace/ftrace.c
6451
del = __ftrace_lookup_ip(sub, entry->ip);
kernel/trace/ftrace.c
6508
del = __ftrace_lookup_ip(direct_functions, entry->ip);
kernel/trace/ftrace.c
6635
tmp = __ftrace_lookup_ip(direct_functions, entry->ip);
kernel/trace/ftrace.c
6670
int ftrace_set_filter_ip(struct ftrace_ops *ops, unsigned long ip,
kernel/trace/ftrace.c
6674
return ftrace_set_addr(ops, &ip, 1, remove, reset, 1);
kernel/trace/ftrace.c
7150
seq_printf(m, "%ps\n", (void *)entry->ip);
kernel/trace/ftrace.c
736
ftrace_find_profiled_func(struct ftrace_profile_stat *stat, unsigned long ip)
kernel/trace/ftrace.c
7372
entry = ftrace_lookup_ip(hash, rec->ip);
kernel/trace/ftrace.c
7379
if (add_hash_entry(hash, rec->ip) == NULL)
kernel/trace/ftrace.c
742
key = hash_long(ip, FTRACE_PROFILE_HASH_BITS);
kernel/trace/ftrace.c
749
if (rec->ip == ip)
kernel/trace/ftrace.c
761
key = hash_long(rec->ip, FTRACE_PROFILE_HASH_BITS);
kernel/trace/ftrace.c
7640
rec->ip = addr;
kernel/trace/ftrace.c
769
ftrace_profile_alloc(struct ftrace_profile_stat *stat, unsigned long ip)
kernel/trace/ftrace.c
7711
unsigned long ip;
kernel/trace/ftrace.c
7755
ops_references_ip(struct ftrace_ops *ops, unsigned long ip)
kernel/trace/ftrace.c
7767
!__ftrace_lookup_ip(ops->func_hash->filter_hash, ip))
kernel/trace/ftrace.c
7771
if (ftrace_lookup_ip(ops->func_hash->notrace_hash, ip))
kernel/trace/ftrace.c
7790
if (ops_references_ip(ops, rec->ip)) {
kernel/trace/ftrace.c
781
rec = ftrace_find_profiled_func(stat, ip);
kernel/trace/ftrace.c
7820
entry = __ftrace_lookup_ip(hash, rec->ip);
kernel/trace/ftrace.c
7827
entry->ip = 0;
kernel/trace/ftrace.c
7897
if (within_module(rec->ip, mod)) {
kernel/trace/ftrace.c
792
rec->ip = ip;
kernel/trace/ftrace.c
7972
if (!within_module(rec->ip, mod))
kernel/trace/ftrace.c
802
function_profile_call(unsigned long ip, unsigned long parent_ip,
kernel/trace/ftrace.c
8042
ret = kallsyms_lookup(rec->ip, &symsize, &offset, &modname, str);
kernel/trace/ftrace.c
8056
mod_func->ip = rec->ip - offset;
kernel/trace/ftrace.c
8098
if (addr >= mod_func->ip &&
kernel/trace/ftrace.c
8099
addr < mod_func->ip + mod_func->size) {
kernel/trace/ftrace.c
8109
*off = addr - found_func->ip;
kernel/trace/ftrace.c
8163
*value = mod_func->ip;
kernel/trace/ftrace.c
817
rec = ftrace_find_profiled_func(stat, ip);
kernel/trace/ftrace.c
819
rec = ftrace_profile_alloc(stat, ip);
kernel/trace/ftrace.c
8205
unsigned long ip;
kernel/trace/ftrace.c
8214
entry = ftrace_lookup_ip(hash, func->ip);
kernel/trace/ftrace.c
8221
entry->ip = 0;
kernel/trace/ftrace.c
8252
func->ip = rec->ip;
kernel/trace/ftrace.c
8269
key.ip = start;
kernel/trace/ftrace.c
8283
if (end < pg->records[0].ip ||
kernel/trace/ftrace.c
8284
start >= (pg->records[pg->index - 1].ip + MCOUNT_INSN_SIZE))
kernel/trace/ftrace.c
8471
__ftrace_ops_list_func(unsigned long ip, unsigned long parent_ip,
kernel/trace/ftrace.c
8483
bit = trace_test_and_set_recursion(ip, parent_ip, TRACE_LIST_START);
kernel/trace/ftrace.c
8499
ftrace_ops_test(op, ip, regs)) {
kernel/trace/ftrace.c
8504
op->func(ip, parent_ip, op, fregs);
kernel/trace/ftrace.c
8528
void arch_ftrace_ops_list_func(unsigned long ip, unsigned long parent_ip,
kernel/trace/ftrace.c
8532
__ftrace_ops_list_func(ip, parent_ip, NULL, fregs);
kernel/trace/ftrace.c
8535
void arch_ftrace_ops_list_func(unsigned long ip, unsigned long parent_ip)
kernel/trace/ftrace.c
8537
__ftrace_ops_list_func(ip, parent_ip, NULL, NULL);
kernel/trace/ftrace.c
8547
static void ftrace_ops_assist_func(unsigned long ip, unsigned long parent_ip,
kernel/trace/ftrace.c
8552
bit = trace_test_and_set_recursion(ip, parent_ip, TRACE_LIST_START);
kernel/trace/ftrace.c
8557
op->func(ip, parent_ip, op, fregs);
kernel/trace/ftrace.c
9061
unsigned long ip = entry->ip;
kernel/trace/ftrace.c
9068
if (ops_references_ip(op, ip)) {
kernel/trace/ftrace.c
9079
ret = op->ops_func(op, ip, FTRACE_OPS_CMD_ENABLE_SHARE_IPMODIFY_PEER);
kernel/trace/ftrace.c
9110
unsigned long ip = entry->ip;
kernel/trace/ftrace.c
9117
if (ops_references_ip(op, ip)) {
kernel/trace/ftrace.c
9126
op->ops_func(op, ip, FTRACE_OPS_CMD_DISABLE_SHARE_IPMODIFY_PEER);
kernel/trace/ftrace_internal.h
17
int ftrace_ops_test(struct ftrace_ops *ops, unsigned long ip, void *regs);
kernel/trace/ftrace_internal.h
39
ftrace_ops_test(struct ftrace_ops *ops, unsigned long ip, void *regs)
kernel/trace/rv/monitors/opid/opid.c
60
static void handle_irq_disable(void *data, unsigned long ip, unsigned long parent_ip)
kernel/trace/rv/monitors/opid/opid.c
65
static void handle_irq_enable(void *data, unsigned long ip, unsigned long parent_ip)
kernel/trace/rv/monitors/opid/opid.c
75
static void handle_preempt_disable(void *data, unsigned long ip, unsigned long parent_ip)
kernel/trace/rv/monitors/opid/opid.c
80
static void handle_preempt_enable(void *data, unsigned long ip, unsigned long parent_ip)
kernel/trace/rv/monitors/scpd/scpd.c
21
static void handle_preempt_disable(void *data, unsigned long ip, unsigned long parent_ip)
kernel/trace/rv/monitors/scpd/scpd.c
26
static void handle_preempt_enable(void *data, unsigned long ip, unsigned long parent_ip)
kernel/trace/rv/monitors/snep/snep.c
21
static void handle_preempt_disable(void *data, unsigned long ip, unsigned long parent_ip)
kernel/trace/rv/monitors/snep/snep.c
26
static void handle_preempt_enable(void *data, unsigned long ip, unsigned long parent_ip)
kernel/trace/rv/monitors/sts/sts.c
60
static void handle_irq_disable(void *data, unsigned long ip, unsigned long parent_ip)
kernel/trace/rv/monitors/sts/sts.c
65
static void handle_irq_enable(void *data, unsigned long ip, unsigned long parent_ip)
kernel/trace/rv/monitors/wip/wip.c
20
static void handle_preempt_disable(void *data, unsigned long ip, unsigned long parent_ip)
kernel/trace/rv/monitors/wip/wip.c
25
static void handle_preempt_enable(void *data, unsigned long ip, unsigned long parent_ip)
kernel/trace/trace.c
2540
trace_function(struct trace_array *tr, unsigned long ip, unsigned long
kernel/trace/trace.c
2555
entry->ip = ip;
kernel/trace/trace.c
2801
entry->ip = last_info->ip;
kernel/trace/trace.c
6491
size_t cnt, unsigned long ip)
kernel/trace/trace.c
6526
entry->ip = ip;
kernel/trace/trace.c
6863
unsigned long ip;
kernel/trace/trace.c
6886
ip = _THIS_IP_;
kernel/trace/trace.c
6892
written = write_marker_to_buffer(tr, buf, cnt, ip);
kernel/trace/trace.c
6897
written = write_marker_to_buffer(tr, buf, cnt, ip);
kernel/trace/trace.c
8415
ftrace_snapshot(unsigned long ip, unsigned long parent_ip,
kernel/trace/trace.c
8423
ftrace_count_snapshot(unsigned long ip, unsigned long parent_ip,
kernel/trace/trace.c
8431
count = (long *)ftrace_func_mapper_find_ip(mapper, ip);
kernel/trace/trace.c
8445
ftrace_snapshot_print(struct seq_file *m, unsigned long ip,
kernel/trace/trace.c
8451
seq_printf(m, "%ps:", (void *)ip);
kernel/trace/trace.c
8456
count = (long *)ftrace_func_mapper_find_ip(mapper, ip);
kernel/trace/trace.c
8468
unsigned long ip, void *init_data, void **data)
kernel/trace/trace.c
8479
return ftrace_func_mapper_add_ip(mapper, ip, init_data);
kernel/trace/trace.c
8484
unsigned long ip, void *data)
kernel/trace/trace.c
8488
if (!ip) {
kernel/trace/trace.c
8495
ftrace_func_mapper_remove_ip(mapper, ip);
kernel/trace/trace.h
1234
void (*func)(unsigned long ip,
kernel/trace/trace.h
1241
unsigned long ip, void *init_data,
kernel/trace/trace.h
1245
unsigned long ip, void *data);
kernel/trace/trace.h
1247
unsigned long ip,
kernel/trace/trace.h
1257
unsigned long ip);
kernel/trace/trace.h
1259
unsigned long ip, void *data);
kernel/trace/trace.h
1261
unsigned long ip);
kernel/trace/trace.h
158
unsigned long ip;
kernel/trace/trace.h
173
unsigned long ip;
kernel/trace/trace.h
315
unsigned long ip;
kernel/trace/trace.h
746
unsigned long ip,
kernel/trace/trace.h
751
unsigned long ip,
kernel/trace/trace.h
907
int trace_vbprintk(unsigned long ip, const char *fmt, va_list args);
kernel/trace/trace.h
909
int trace_vprintk(unsigned long ip, const char *fmt, va_list args);
kernel/trace/trace.h
912
unsigned long ip, const char *fmt, va_list args);
kernel/trace/trace.h
915
unsigned long ip, const char *fmt, ...);
kernel/trace/trace.h
943
ftrace_lookup_ip(struct ftrace_hash *hash, unsigned long ip);
kernel/trace/trace_entries.h
261
__field( unsigned long, ip )
kernel/trace/trace_entries.h
267
(void *)__entry->ip, __entry->fmt)
kernel/trace/trace_entries.h
275
__field( unsigned long, ip )
kernel/trace/trace_entries.h
280
(void *)__entry->ip, __entry->buf),
kernel/trace/trace_entries.h
303
__field( unsigned long, ip )
kernel/trace/trace_entries.h
308
(void *)__entry->ip, __entry->str)
kernel/trace/trace_entries.h
406
__field( unsigned long, ip )
kernel/trace/trace_entries.h
414
(void *)__entry->ip,
kernel/trace/trace_entries.h
64
__field_fn( unsigned long, ip )
kernel/trace/trace_entries.h
70
(void *)__entry->ip, (void *)__entry->parent_ip),
kernel/trace/trace_event_perf.c
435
perf_ftrace_function_call(unsigned long ip, unsigned long parent_ip,
kernel/trace/trace_event_perf.c
448
bit = ftrace_test_recursion_trylock(ip, parent_ip);
kernel/trace/trace_event_perf.c
477
entry->ip = ip;
kernel/trace/trace_events.c
4137
event_enable_probe(unsigned long ip, unsigned long parent_ip,
kernel/trace/trace_events.c
4145
pdata = ftrace_func_mapper_find_ip(mapper, ip);
kernel/trace/trace_events.c
4154
event_enable_count_probe(unsigned long ip, unsigned long parent_ip,
kernel/trace/trace_events.c
4162
pdata = ftrace_func_mapper_find_ip(mapper, ip);
kernel/trace/trace_events.c
4182
event_enable_print(struct seq_file *m, unsigned long ip,
kernel/trace/trace_events.c
4189
pdata = ftrace_func_mapper_find_ip(mapper, ip);
kernel/trace/trace_events.c
4196
seq_printf(m, "%ps:", (void *)ip);
kernel/trace/trace_events.c
4213
unsigned long ip, void *init_data, void **data)
kernel/trace/trace_events.c
4226
ret = ftrace_func_mapper_add_ip(mapper, ip, edata);
kernel/trace/trace_events.c
4251
unsigned long ip, void *data)
kernel/trace/trace_events.c
4256
if (!ip) {
kernel/trace/trace_events.c
4263
edata = ftrace_func_mapper_remove_ip(mapper, ip);
kernel/trace/trace_events.c
5010
function_test_events_call(unsigned long ip, unsigned long parent_ip,
kernel/trace/trace_events.c
5034
entry->ip = ip;
kernel/trace/trace_events_filter.c
1637
unsigned long ip;
kernel/trace/trace_events_filter.c
1751
ret = kstrtoul(num_buf, 0, &ip);
kernel/trace/trace_events_filter.c
1765
ip = kallsyms_lookup_name(name);
kernel/trace/trace_events_filter.c
1767
if (!ip) {
kernel/trace/trace_events_filter.c
1774
if (!kallsyms_lookup_size_offset(ip, &size, &offset)) {
kernel/trace/trace_events_filter.c
1780
pred->val = ip - offset;
kernel/trace/trace_fprobe.c
1128
unsigned long ip = tracepoint_user_ip(tf->tuser);
kernel/trace/trace_fprobe.c
1130
WARN_ON_ONCE(register_fprobe_ips(&tf->fp, &ip, 1));
kernel/trace/trace_fprobe.c
345
entry->ip = entry_ip;
kernel/trace/trace_fprobe.c
476
entry->ip = entry_ip;
kernel/trace/trace_fprobe.c
635
if (!seq_print_ip_sym_offset(s, field->ip, flags))
kernel/trace/trace_fprobe.c
695
DEFINE_FIELD(unsigned long, ip, FIELD_STRING_IP, 0);
kernel/trace/trace_fprobe.c
771
unsigned long ip;
kernel/trace/trace_fprobe.c
787
ip = tracepoint_user_ip(tuser);
kernel/trace/trace_fprobe.c
788
if (WARN_ON_ONCE(!ip))
kernel/trace/trace_fprobe.c
791
ret = register_fprobe_ips(&tf->fp, &ip, 1);
kernel/trace/trace_functions.c
208
function_trace_call(unsigned long ip, unsigned long parent_ip,
kernel/trace/trace_functions.c
218
bit = ftrace_test_recursion_trylock(ip, parent_ip);
kernel/trace/trace_functions.c
226
trace_function(tr, ip, parent_ip, trace_ctx, NULL);
kernel/trace/trace_functions.c
232
function_args_trace_call(unsigned long ip, unsigned long parent_ip,
kernel/trace/trace_functions.c
242
bit = ftrace_test_recursion_trylock(ip, parent_ip);
kernel/trace/trace_functions.c
248
trace_function(tr, ip, parent_ip, trace_ctx, fregs);
kernel/trace/trace_functions.c
25
function_trace_call(unsigned long ip, unsigned long parent_ip,
kernel/trace/trace_functions.c
272
function_stack_trace_call(unsigned long ip, unsigned long parent_ip,
kernel/trace/trace_functions.c
28
function_args_trace_call(unsigned long ip, unsigned long parent_ip,
kernel/trace/trace_functions.c
298
trace_function(tr, ip, parent_ip, trace_ctx, NULL);
kernel/trace/trace_functions.c
31
function_stack_trace_call(unsigned long ip, unsigned long parent_ip,
kernel/trace/trace_functions.c
312
unsigned long ip, unsigned long parent_ip)
kernel/trace/trace_functions.c
314
if (last_info->ip == ip &&
kernel/trace/trace_functions.c
327
unsigned long ip, unsigned long parent_ip,
kernel/trace/trace_functions.c
336
last_info->ip = ip;
kernel/trace/trace_functions.c
34
function_no_repeats_trace_call(unsigned long ip, unsigned long parent_ip,
kernel/trace/trace_functions.c
341
function_no_repeats_trace_call(unsigned long ip, unsigned long parent_ip,
kernel/trace/trace_functions.c
353
bit = ftrace_test_recursion_trylock(ip, parent_ip);
kernel/trace/trace_functions.c
369
if (is_repeat_check(tr, last_info, ip, parent_ip))
kernel/trace/trace_functions.c
37
function_stack_no_repeats_trace_call(unsigned long ip, unsigned long parent_ip,
kernel/trace/trace_functions.c
373
process_repeats(tr, ip, parent_ip, last_info, trace_ctx);
kernel/trace/trace_functions.c
375
trace_function(tr, ip, parent_ip, trace_ctx, NULL);
kernel/trace/trace_functions.c
382
function_stack_no_repeats_trace_call(unsigned long ip, unsigned long parent_ip,
kernel/trace/trace_functions.c
409
if (is_repeat_check(tr, last_info, ip, parent_ip))
kernel/trace/trace_functions.c
413
process_repeats(tr, ip, parent_ip, last_info, trace_ctx);
kernel/trace/trace_functions.c
415
trace_function(tr, ip, parent_ip, trace_ctx, NULL);
kernel/trace/trace_functions.c
504
unsigned long ip,
kernel/trace/trace_functions.c
541
count = (long *)ftrace_func_mapper_find_ip(mapper, ip);
kernel/trace/trace_functions.c
565
ftrace_traceon_count(unsigned long ip, unsigned long parent_ip,
kernel/trace/trace_functions.c
569
update_traceon_count(ops, ip, tr, 1, data);
kernel/trace/trace_functions.c
573
ftrace_traceoff_count(unsigned long ip, unsigned long parent_ip,
kernel/trace/trace_functions.c
577
update_traceon_count(ops, ip, tr, 0, data);
kernel/trace/trace_functions.c
581
ftrace_traceon(unsigned long ip, unsigned long parent_ip,
kernel/trace/trace_functions.c
592
ftrace_traceoff(unsigned long ip, unsigned long parent_ip,
kernel/trace/trace_functions.c
630
ftrace_stacktrace(unsigned long ip, unsigned long parent_ip,
kernel/trace/trace_functions.c
638
ftrace_stacktrace_count(unsigned long ip, unsigned long parent_ip,
kernel/trace/trace_functions.c
656
count = (long *)ftrace_func_mapper_find_ip(mapper, ip);
kernel/trace/trace_functions.c
679
static int update_count(struct ftrace_probe_ops *ops, unsigned long ip,
kernel/trace/trace_functions.c
686
count = (long *)ftrace_func_mapper_find_ip(mapper, ip);
kernel/trace/trace_functions.c
698
ftrace_dump_probe(unsigned long ip, unsigned long parent_ip,
kernel/trace/trace_functions.c
702
if (update_count(ops, ip, data))
kernel/trace/trace_functions.c
708
ftrace_cpudump_probe(unsigned long ip, unsigned long parent_ip,
kernel/trace/trace_functions.c
712
if (update_count(ops, ip, data))
kernel/trace/trace_functions.c
718
unsigned long ip, struct ftrace_probe_ops *ops,
kernel/trace/trace_functions.c
724
seq_printf(m, "%ps:%s", (void *)ip, name);
kernel/trace/trace_functions.c
727
count = (long *)ftrace_func_mapper_find_ip(mapper, ip);
kernel/trace/trace_functions.c
738
ftrace_traceon_print(struct seq_file *m, unsigned long ip,
kernel/trace/trace_functions.c
742
return ftrace_probe_print("traceon", m, ip, ops, data);
kernel/trace/trace_functions.c
746
ftrace_traceoff_print(struct seq_file *m, unsigned long ip,
kernel/trace/trace_functions.c
749
return ftrace_probe_print("traceoff", m, ip, ops, data);
kernel/trace/trace_functions.c
753
ftrace_stacktrace_print(struct seq_file *m, unsigned long ip,
kernel/trace/trace_functions.c
756
return ftrace_probe_print("stacktrace", m, ip, ops, data);
kernel/trace/trace_functions.c
760
ftrace_dump_print(struct seq_file *m, unsigned long ip,
kernel/trace/trace_functions.c
763
return ftrace_probe_print("dump", m, ip, ops, data);
kernel/trace/trace_functions.c
767
ftrace_cpudump_print(struct seq_file *m, unsigned long ip,
kernel/trace/trace_functions.c
770
return ftrace_probe_print("cpudump", m, ip, ops, data);
kernel/trace/trace_functions.c
776
unsigned long ip, void *init_data, void **data)
kernel/trace/trace_functions.c
787
return ftrace_func_mapper_add_ip(mapper, ip, init_data);
kernel/trace/trace_functions.c
792
unsigned long ip, void *data)
kernel/trace/trace_functions.c
796
if (!ip) {
kernel/trace/trace_functions.c
801
ftrace_func_mapper_remove_ip(mapper, ip);
kernel/trace/trace_functions_graph.c
310
unsigned long ip, unsigned int trace_ctx)
kernel/trace/trace_functions_graph.c
314
.func = ip,
kernel/trace/trace_functions_graph.c
318
.func = ip,
kernel/trace/trace_functions_graph.c
328
unsigned long ip, unsigned long parent_ip,
kernel/trace/trace_functions_graph.c
331
__trace_graph_function(tr, ip, trace_ctx);
kernel/trace/trace_irqsoff.c
140
irqsoff_tracer_call(unsigned long ip, unsigned long parent_ip,
kernel/trace/trace_irqsoff.c
153
trace_function(tr, ip, parent_ip, trace_ctx, fregs);
kernel/trace/trace_irqsoff.c
287
unsigned long ip, unsigned long parent_ip,
kernel/trace/trace_irqsoff.c
291
trace_graph_function(tr, ip, parent_ip, trace_ctx);
kernel/trace/trace_irqsoff.c
293
trace_function(tr, ip, parent_ip, trace_ctx, NULL);
kernel/trace/trace_irqsoff.c
299
unsigned long ip, unsigned long parent_ip,
kernel/trace/trace_irqsoff.c
302
return trace_function(tr, ip, parent_ip, trace_ctx, NULL);
kernel/trace/trace_irqsoff.c
392
start_critical_timing(unsigned long ip, unsigned long parent_ip)
kernel/trace/trace_irqsoff.c
417
data->critical_start = parent_ip ? : ip;
kernel/trace/trace_irqsoff.c
419
__trace_function(tr, ip, parent_ip, tracing_gen_ctx());
kernel/trace/trace_irqsoff.c
428
stop_critical_timing(unsigned long ip, unsigned long parent_ip)
kernel/trace/trace_irqsoff.c
456
__trace_function(tr, ip, parent_ip, trace_ctx);
kernel/trace/trace_irqsoff.c
457
check_critical_timing(tr, data, parent_ip ? : ip, cpu);
kernel/trace/trace_kprobe.c
1488
entry->ip = (unsigned long)tk->rp.kp.addr;
kernel/trace/trace_kprobe.c
1588
if (!seq_print_ip_sym_offset(s, field->ip, flags))
kernel/trace/trace_kprobe.c
1649
DEFINE_FIELD(unsigned long, ip, FIELD_STRING_IP, 0);
kernel/trace/trace_kprobe.c
1712
entry->ip = (unsigned long)tk->rp.kp.addr;
kernel/trace/trace_output.c
1079
unsigned long ip;
kernel/trace/trace_output.c
1081
ip = addr + (unsigned long)_stext;
kernel/trace/trace_output.c
1082
ip = trace_adjust_address(tr, ip);
kernel/trace/trace_output.c
1083
trace_seq_printf(&iter->seq, "%pS ", (void *)ip);
kernel/trace/trace_output.c
1166
static void print_fn_trace(struct trace_seq *s, unsigned long ip,
kernel/trace/trace_output.c
1170
ip = trace_adjust_address(tr, ip);
kernel/trace/trace_output.c
1173
seq_print_ip_sym(s, ip, flags);
kernel/trace/trace_output.c
1175
print_function_args(s, args, ip);
kernel/trace/trace_output.c
1200
print_fn_trace(s, field->ip, field->parent_ip, args, iter->tr, flags);
kernel/trace/trace_output.c
1214
field->ip,
kernel/trace/trace_output.c
1228
SEQ_PUT_HEX_FIELD(s, field->ip);
kernel/trace/trace_output.c
1242
SEQ_PUT_FIELD(s, field->ip);
kernel/trace/trace_output.c
1481
unsigned long ip = field->caller[i];
kernel/trace/trace_output.c
1483
if (!ip || trace_seq_has_overflowed(s))
kernel/trace/trace_output.c
1487
seq_print_user_ip(s, mm, ip, flags);
kernel/trace/trace_output.c
1699
seq_print_ip_sym(s, field->ip, flags);
kernel/trace/trace_output.c
1716
trace_seq_printf(s, ": %lx : ", field->ip);
kernel/trace/trace_output.c
1743
seq_print_ip_sym(s, field->ip, flags);
kernel/trace/trace_output.c
1760
trace_seq_printf(s, ": %lx : ", field->ip);
kernel/trace/trace_output.c
1782
unsigned long ip;
kernel/trace/trace_output.c
1786
ip = trace_adjust_address(iter->tr, field->ip);
kernel/trace/trace_output.c
1788
seq_print_ip_sym(s, ip, flags);
kernel/trace/trace_output.c
1801
trace_seq_printf(&iter->seq, "# %lx %s", field->ip, field->buf);
kernel/trace/trace_output.c
1855
field->ip,
kernel/trace/trace_output.c
1872
print_fn_trace(s, field->ip, field->parent_ip, NULL, iter->tr, flags);
kernel/trace/trace_output.c
421
unsigned long ip, unsigned long sym_flags)
kernel/trace/trace_output.c
434
vma = find_vma(mm, ip);
kernel/trace/trace_output.c
443
ip - vmstart);
kernel/trace/trace_output.c
448
trace_seq_printf(s, " <" IP_FMT ">", ip);
kernel/trace/trace_output.c
453
seq_print_ip_sym(struct trace_seq *s, unsigned long ip, unsigned long sym_flags)
kernel/trace/trace_output.c
455
if (!ip) {
kernel/trace/trace_output.c
460
trace_seq_print_sym(s, ip, sym_flags & TRACE_ITER(SYM_OFFSET));
kernel/trace/trace_output.c
463
trace_seq_printf(s, " <" IP_FMT ">", ip);
kernel/trace/trace_output.h
16
seq_print_ip_sym(struct trace_seq *s, unsigned long ip,
kernel/trace/trace_output.h
19
static inline int seq_print_ip_sym_offset(struct trace_seq *s, unsigned long ip,
kernel/trace/trace_output.h
22
return seq_print_ip_sym(s, ip, sym_flags | TRACE_ITER(SYM_OFFSET));
kernel/trace/trace_output.h
24
static inline int seq_print_ip_sym_no_offset(struct trace_seq *s, unsigned long ip,
kernel/trace/trace_output.h
27
return seq_print_ip_sym(s, ip, sym_flags & ~TRACE_ITER(SYM_OFFSET));
kernel/trace/trace_printk.c
200
int __trace_bprintk(unsigned long ip, const char *fmt, ...)
kernel/trace/trace_printk.c
212
ret = trace_vbprintk(ip, fmt, ap);
kernel/trace/trace_printk.c
218
int __ftrace_vbprintk(unsigned long ip, const char *fmt, va_list ap)
kernel/trace/trace_printk.c
226
return trace_vbprintk(ip, fmt, ap);
kernel/trace/trace_printk.c
230
int __trace_printk(unsigned long ip, const char *fmt, ...)
kernel/trace/trace_printk.c
239
ret = trace_vprintk(ip, fmt, ap);
kernel/trace/trace_printk.c
245
int __ftrace_vprintk(unsigned long ip, const char *fmt, va_list ap)
kernel/trace/trace_printk.c
250
return trace_vprintk(ip, fmt, ap);
kernel/trace/trace_printk.c
390
int __trace_array_puts(struct trace_array *tr, unsigned long ip,
kernel/trace/trace_printk.c
420
entry->ip = ip;
kernel/trace/trace_printk.c
442
int __trace_puts(unsigned long ip, const char *str)
kernel/trace/trace_printk.c
444
return __trace_array_puts(printk_trace, ip, str, strlen(str));
kernel/trace/trace_printk.c
453
int __trace_bputs(unsigned long ip, const char *str)
kernel/trace/trace_printk.c
463
return __trace_puts(ip, str);
kernel/trace/trace_printk.c
481
entry->ip = ip;
kernel/trace/trace_printk.c
607
int trace_vbprintk(unsigned long ip, const char *fmt, va_list args)
kernel/trace/trace_printk.c
618
return trace_vprintk(ip, fmt, args);
kernel/trace/trace_printk.c
648
entry->ip = ip;
kernel/trace/trace_printk.c
667
unsigned long ip, const char *fmt, va_list args)
kernel/trace/trace_printk.c
700
entry->ip = ip;
kernel/trace/trace_printk.c
716
unsigned long ip, const char *fmt, va_list args)
kernel/trace/trace_printk.c
721
return __trace_array_vprintk(tr->array_buffer.buffer, ip, fmt, args);
kernel/trace/trace_printk.c
745
unsigned long ip, const char *fmt, ...)
kernel/trace/trace_printk.c
761
ret = trace_array_vprintk(tr, ip, fmt, ap);
kernel/trace/trace_printk.c
789
unsigned long ip, const char *fmt, ...)
kernel/trace/trace_printk.c
798
ret = __trace_array_vprintk(buffer, ip, fmt, ap);
kernel/trace/trace_printk.c
803
int trace_vprintk(unsigned long ip, const char *fmt, va_list args)
kernel/trace/trace_printk.c
805
return trace_array_vprintk(printk_trace, ip, fmt, args);
kernel/trace/trace_recursion_record.c
112
cmpxchg(&recursed_functions[index].ip, ip, 0);
kernel/trace/trace_recursion_record.c
12
unsigned long ip;
kernel/trace/trace_recursion_record.c
166
trace_seq_print_sym(tseq, record->ip, true);
kernel/trace/trace_recursion_record.c
25
void ftrace_record_recursion(unsigned long ip, unsigned long parent_ip)
kernel/trace/trace_recursion_record.c
33
if (ip == cached_function)
kernel/trace/trace_recursion_record.c
64
if (recursed_functions[i].ip == ip) {
kernel/trace/trace_recursion_record.c
65
cached_function = ip;
kernel/trace/trace_recursion_record.c
70
cached_function = ip;
kernel/trace/trace_recursion_record.c
78
old = cmpxchg(&recursed_functions[index].ip, 0, ip);
kernel/trace/trace_recursion_record.c
81
if (old == ip)
kernel/trace/trace_sched_wakeup.c
227
wakeup_tracer_call(unsigned long ip, unsigned long parent_ip,
kernel/trace/trace_sched_wakeup.c
239
trace_function(tr, ip, parent_ip, trace_ctx, fregs);
kernel/trace/trace_sched_wakeup.c
318
unsigned long ip, unsigned long parent_ip,
kernel/trace/trace_sched_wakeup.c
322
trace_graph_function(tr, ip, parent_ip, trace_ctx);
kernel/trace/trace_sched_wakeup.c
324
trace_function(tr, ip, parent_ip, trace_ctx, NULL);
kernel/trace/trace_selftest.c
108
static void trace_selftest_test_probe1_func(unsigned long ip,
kernel/trace/trace_selftest.c
117
static void trace_selftest_test_probe2_func(unsigned long ip,
kernel/trace/trace_selftest.c
126
static void trace_selftest_test_probe3_func(unsigned long ip,
kernel/trace/trace_selftest.c
135
static void trace_selftest_test_global_func(unsigned long ip,
kernel/trace/trace_selftest.c
144
static void trace_selftest_test_dyn_func(unsigned long ip,
kernel/trace/trace_selftest.c
449
static void trace_selftest_test_recursion_func(unsigned long ip,
kernel/trace/trace_selftest.c
464
static void trace_selftest_test_recursion_safe_func(unsigned long ip,
kernel/trace/trace_selftest.c
583
static void trace_selftest_test_regs_func(unsigned long ip,
kernel/trace/trace_stack.c
155
static void check_stack(unsigned long ip, unsigned long *stack)
kernel/trace/trace_stack.c
197
if (stack_dump_trace[i] == ip)
kernel/trace/trace_stack.c
292
stack_trace_call(unsigned long ip, unsigned long parent_ip,
kernel/trace/trace_stack.c
308
ip += MCOUNT_INSN_SIZE;
kernel/trace/trace_stack.c
310
check_stack(ip, &stack);
kernel/unwind/user.c
129
state->ip = instruction_pointer(regs);
kernel/unwind/user.c
155
trace->entries[trace->nr++] = state.ip;
kernel/unwind/user.c
60
state->ip = ra;
lib/closure.c
269
cl, (void *) cl->ip, cl->fn, cl->parent,
lib/lz4/lz4_compress.c
187
const BYTE *ip = (const BYTE *) source;
lib/lz4/lz4_compress.c
190
const BYTE * const lowRefLimit = ip - dictPtr->dictSize;
lib/lz4/lz4_compress.c
195
const BYTE * const iend = ip + inputSize;
lib/lz4/lz4_compress.c
239
LZ4_putPosition(ip, dictPtr->hashTable, tableType, base);
lib/lz4/lz4_compress.c
240
ip++;
lib/lz4/lz4_compress.c
241
forwardH = LZ4_hashPosition(ip, tableType);
lib/lz4/lz4_compress.c
250
const BYTE *forwardIp = ip;
lib/lz4/lz4_compress.c
257
ip = forwardIp;
lib/lz4/lz4_compress.c
280
LZ4_putPositionOnHash(ip, h, dictPtr->hashTable,
lib/lz4/lz4_compress.c
287
: (match + MAX_DISTANCE < ip))
lib/lz4/lz4_compress.c
289
!= LZ4_read32(ip)));
lib/lz4/lz4_compress.c
293
while (((ip > anchor) & (match + refDelta > lowLimit))
lib/lz4/lz4_compress.c
294
&& (unlikely(ip[-1] == match[refDelta - 1]))) {
lib/lz4/lz4_compress.c
295
ip--;
lib/lz4/lz4_compress.c
301
unsigned const int litLength = (unsigned int)(ip - anchor);
lib/lz4/lz4_compress.c
330
LZ4_writeLE16(op, (U16)(ip - match));
lib/lz4/lz4_compress.c
342
limit = ip + (dictEnd - match);
lib/lz4/lz4_compress.c
347
matchCode = LZ4_count(ip + MINMATCH,
lib/lz4/lz4_compress.c
350
ip += MINMATCH + matchCode;
lib/lz4/lz4_compress.c
352
if (ip == limit) {
lib/lz4/lz4_compress.c
353
unsigned const int more = LZ4_count(ip,
lib/lz4/lz4_compress.c
358
ip += more;
lib/lz4/lz4_compress.c
361
matchCode = LZ4_count(ip + MINMATCH,
lib/lz4/lz4_compress.c
363
ip += MINMATCH + matchCode;
lib/lz4/lz4_compress.c
390
anchor = ip;
lib/lz4/lz4_compress.c
393
if (ip > mflimit)
lib/lz4/lz4_compress.c
397
LZ4_putPosition(ip - 2, dictPtr->hashTable, tableType, base);
lib/lz4/lz4_compress.c
400
match = LZ4_getPosition(ip, dictPtr->hashTable,
lib/lz4/lz4_compress.c
413
LZ4_putPosition(ip, dictPtr->hashTable, tableType, base);
lib/lz4/lz4_compress.c
416
&& (match + MAX_DISTANCE >= ip)
lib/lz4/lz4_compress.c
417
&& (LZ4_read32(match + refDelta) == LZ4_read32(ip))) {
lib/lz4/lz4_compress.c
424
forwardH = LZ4_hashPosition(++ip, tableType);
lib/lz4/lz4_compress.c
529
const BYTE *ip = (const BYTE *) src;
lib/lz4/lz4_compress.c
532
const BYTE *anchor = ip;
lib/lz4/lz4_compress.c
533
const BYTE * const iend = ip + *srcSizePtr;
lib/lz4/lz4_compress.c
563
LZ4_putPosition(ip, ctx->hashTable, tableType, base);
lib/lz4/lz4_compress.c
564
ip++; forwardH = LZ4_hashPosition(ip, tableType);
lib/lz4/lz4_compress.c
573
const BYTE *forwardIp = ip;
lib/lz4/lz4_compress.c
580
ip = forwardIp;
lib/lz4/lz4_compress.c
591
LZ4_putPositionOnHash(ip, h,
lib/lz4/lz4_compress.c
597
: (match + MAX_DISTANCE < ip))
lib/lz4/lz4_compress.c
598
|| (LZ4_read32(match) != LZ4_read32(ip)));
lib/lz4/lz4_compress.c
602
while ((ip > anchor)
lib/lz4/lz4_compress.c
604
&& (unlikely(ip[-1] == match[-1]))) {
lib/lz4/lz4_compress.c
605
ip--;
lib/lz4/lz4_compress.c
611
unsigned int litLength = (unsigned int)(ip - anchor);
lib/lz4/lz4_compress.c
636
LZ4_writeLE16(op, (U16)(ip - match)); op += 2;
lib/lz4/lz4_compress.c
640
size_t matchLength = LZ4_count(ip + MINMATCH,
lib/lz4/lz4_compress.c
647
ip += MINMATCH + matchLength;
lib/lz4/lz4_compress.c
661
anchor = ip;
lib/lz4/lz4_compress.c
664
if (ip > mflimit)
lib/lz4/lz4_compress.c
670
LZ4_putPosition(ip - 2, ctx->hashTable, tableType, base);
lib/lz4/lz4_compress.c
673
match = LZ4_getPosition(ip, ctx->hashTable, tableType, base);
lib/lz4/lz4_compress.c
674
LZ4_putPosition(ip, ctx->hashTable, tableType, base);
lib/lz4/lz4_compress.c
676
if ((match + MAX_DISTANCE >= ip)
lib/lz4/lz4_compress.c
677
&& (LZ4_read32(match) == LZ4_read32(ip))) {
lib/lz4/lz4_compress.c
683
forwardH = LZ4_hashPosition(++ip, tableType);
lib/lz4/lz4_compress.c
698
ip = anchor + lastRunSize;
lib/lz4/lz4_compress.c
715
*srcSizePtr = (int) (((const char *)ip) - src);
lib/lz4/lz4_decompress.c
111
return ((srcSize == 1) && (*ip == 0)) ? 0 : -1;
lib/lz4/lz4_decompress.c
114
return (*ip == 0 ? 1 : -1);
lib/lz4/lz4_decompress.c
126
unsigned int const token = *ip++;
lib/lz4/lz4_decompress.c
130
assert(!endOnInput || ip <= iend);
lib/lz4/lz4_decompress.c
152
&& likely((endOnInput ? ip < shortiend : 1) &
lib/lz4/lz4_decompress.c
155
LZ4_memcpy(op, ip, endOnInput ? 16 : 8);
lib/lz4/lz4_decompress.c
156
op += length; ip += length;
lib/lz4/lz4_decompress.c
164
offset = LZ4_readLE16(ip);
lib/lz4/lz4_decompress.c
165
ip += 2;
lib/lz4/lz4_decompress.c
194
if (unlikely(endOnInput ? ip >= iend - RUN_MASK : 0)) {
lib/lz4/lz4_decompress.c
199
s = *ip++;
lib/lz4/lz4_decompress.c
202
? ip < iend - RUN_MASK
lib/lz4/lz4_decompress.c
212
&& unlikely((uptrval)(ip) +
lib/lz4/lz4_decompress.c
213
length < (uptrval)(ip))) {
lib/lz4/lz4_decompress.c
224
|| (ip + length > iend - (2 + 1 + LASTLITERALS))))
lib/lz4/lz4_decompress.c
236
&& (ip + length > iend)) {
lib/lz4/lz4_decompress.c
255
&& ((ip + length != iend)
lib/lz4/lz4_decompress.c
269
LZ4_memmove(op, ip, length);
lib/lz4/lz4_decompress.c
270
ip += length;
lib/lz4/lz4_decompress.c
278
if (!partialDecoding || (cpy == oend) || (ip >= (iend - 2)))
lib/lz4/lz4_decompress.c
282
LZ4_wildCopy(op, ip, cpy);
lib/lz4/lz4_decompress.c
283
ip += length;
lib/lz4/lz4_decompress.c
288
offset = LZ4_readLE16(ip);
lib/lz4/lz4_decompress.c
289
ip += 2;
lib/lz4/lz4_decompress.c
317
s = *ip++;
lib/lz4/lz4_decompress.c
319
if ((endOnInput) && (ip > iend - LASTLITERALS))
lib/lz4/lz4_decompress.c
451
return (int) (((const char *)ip) - src);
lib/lz4/lz4_decompress.c
456
return (int) (-(((const char *)ip) - src)) - 1;
lib/lz4/lz4_decompress.c
82
const BYTE *ip = (const BYTE *) src;
lib/lz4/lz4_decompress.c
83
const BYTE * const iend = ip + srcSize;
lib/lz4/lz4hc_compress.c
100
const BYTE *ip,
lib/lz4/lz4hc_compress.c
110
const U32 lowLimit = (hc4->lowLimit + 64 * KB > (U32)(ip - base))
lib/lz4/lz4hc_compress.c
112
: (U32)(ip - base) - (64 * KB - 1);
lib/lz4/lz4hc_compress.c
118
LZ4HC_Insert(hc4, ip);
lib/lz4/lz4hc_compress.c
119
matchIndex = HashTable[LZ4HC_hashPtr(ip)];
lib/lz4/lz4hc_compress.c
127
if (*(match + ml) == *(ip + ml)
lib/lz4/lz4hc_compress.c
128
&& (LZ4_read32(match) == LZ4_read32(ip))) {
lib/lz4/lz4hc_compress.c
129
size_t const mlt = LZ4_count(ip + MINMATCH,
lib/lz4/lz4hc_compress.c
140
if (LZ4_read32(match) == LZ4_read32(ip)) {
lib/lz4/lz4hc_compress.c
142
const BYTE *vLimit = ip
lib/lz4/lz4hc_compress.c
147
mlt = LZ4_count(ip + MINMATCH,
lib/lz4/lz4hc_compress.c
149
if ((ip + mlt == vLimit)
lib/lz4/lz4hc_compress.c
151
mlt += LZ4_count(ip + mlt,
lib/lz4/lz4hc_compress.c
169
const BYTE * const ip,
lib/lz4/lz4hc_compress.c
182
const U32 lowLimit = (hc4->lowLimit + 64 * KB > (U32)(ip - base))
lib/lz4/lz4hc_compress.c
184
: (U32)(ip - base) - (64 * KB - 1);
lib/lz4/lz4hc_compress.c
188
int delta = (int)(ip - iLowLimit);
lib/lz4/lz4hc_compress.c
191
LZ4HC_Insert(hc4, ip);
lib/lz4/lz4hc_compress.c
192
matchIndex = HashTable[LZ4HC_hashPtr(ip)];
lib/lz4/lz4hc_compress.c
202
if (LZ4_read32(matchPtr) == LZ4_read32(ip)) {
lib/lz4/lz4hc_compress.c
204
ip + MINMATCH,
lib/lz4/lz4hc_compress.c
209
while ((ip + back > iLowLimit)
lib/lz4/lz4hc_compress.c
211
&& (ip[back - 1] == matchPtr[back - 1]))
lib/lz4/lz4hc_compress.c
219
*startpos = ip + back;
lib/lz4/lz4hc_compress.c
226
if (LZ4_read32(matchPtr) == LZ4_read32(ip)) {
lib/lz4/lz4hc_compress.c
229
const BYTE *vLimit = ip + (dictLimit - matchIndex);
lib/lz4/lz4hc_compress.c
234
mlt = LZ4_count(ip + MINMATCH,
lib/lz4/lz4hc_compress.c
237
if ((ip + mlt == vLimit) && (vLimit < iHighLimit))
lib/lz4/lz4hc_compress.c
238
mlt += LZ4_count(ip + mlt, base + dictLimit,
lib/lz4/lz4hc_compress.c
240
while ((ip + back > iLowLimit)
lib/lz4/lz4hc_compress.c
242
&& (ip[back - 1] == matchPtr[back - 1]))
lib/lz4/lz4hc_compress.c
250
*startpos = ip + back;
lib/lz4/lz4hc_compress.c
262
const BYTE **ip,
lib/lz4/lz4hc_compress.c
274
length = (int)(*ip - *anchor);
lib/lz4/lz4hc_compress.c
299
LZ4_writeLE16(*op, (U16)(*ip - match));
lib/lz4/lz4hc_compress.c
331
*ip += matchLength;
lib/lz4/lz4hc_compress.c
332
*anchor = *ip;
lib/lz4/lz4hc_compress.c
347
const BYTE *ip = (const BYTE *) source;
lib/lz4/lz4hc_compress.c
348
const BYTE *anchor = ip;
lib/lz4/lz4hc_compress.c
349
const BYTE * const iend = ip + inputSize;
lib/lz4/lz4hc_compress.c
374
ip++;
lib/lz4/lz4hc_compress.c
377
while (ip < mflimit) {
lib/lz4/lz4hc_compress.c
378
ml = LZ4HC_InsertAndFindBestMatch(ctx, ip,
lib/lz4/lz4hc_compress.c
381
ip++;
lib/lz4/lz4hc_compress.c
386
start0 = ip;
lib/lz4/lz4hc_compress.c
391
if (ip + ml < mflimit)
lib/lz4/lz4hc_compress.c
393
ip + ml - 2, ip + 0,
lib/lz4/lz4hc_compress.c
401
if (LZ4HC_encodeSequence(&ip, &op,
lib/lz4/lz4hc_compress.c
407
if (start0 < ip) {
lib/lz4/lz4hc_compress.c
408
if (start2 < ip + ml0) {
lib/lz4/lz4hc_compress.c
410
ip = start0;
lib/lz4/lz4hc_compress.c
417
if ((start2 - ip) < 3) {
lib/lz4/lz4hc_compress.c
420
ip = start2;
lib/lz4/lz4hc_compress.c
431
if ((start2 - ip) < OPTIMAL_ML) {
lib/lz4/lz4hc_compress.c
437
if (ip + new_ml > start2 + ml2 - MINMATCH)
lib/lz4/lz4hc_compress.c
438
new_ml = (int)(start2 - ip) + ml2 - MINMATCH;
lib/lz4/lz4hc_compress.c
440
correction = new_ml - (int)(start2 - ip);
lib/lz4/lz4hc_compress.c
464
if (start2 < ip + ml)
lib/lz4/lz4hc_compress.c
465
ml = (int)(start2 - ip);
lib/lz4/lz4hc_compress.c
467
if (LZ4HC_encodeSequence(&ip, &op, &anchor,
lib/lz4/lz4hc_compress.c
470
ip = start2;
lib/lz4/lz4hc_compress.c
471
if (LZ4HC_encodeSequence(&ip, &op, &anchor,
lib/lz4/lz4hc_compress.c
477
if (start3 < ip + ml + 3) {
lib/lz4/lz4hc_compress.c
479
if (start3 >= (ip + ml)) {
lib/lz4/lz4hc_compress.c
484
if (start2 < ip + ml) {
lib/lz4/lz4hc_compress.c
485
int correction = (int)(ip + ml - start2);
lib/lz4/lz4hc_compress.c
497
if (LZ4HC_encodeSequence(&ip, &op, &anchor,
lib/lz4/lz4hc_compress.c
500
ip = start3;
lib/lz4/lz4hc_compress.c
521
if (start2 < ip + ml) {
lib/lz4/lz4hc_compress.c
522
if ((start2 - ip) < (int)ML_MASK) {
lib/lz4/lz4hc_compress.c
527
if (ip + ml > start2 + ml2 - MINMATCH)
lib/lz4/lz4hc_compress.c
528
ml = (int)(start2 - ip) + ml2 - MINMATCH;
lib/lz4/lz4hc_compress.c
529
correction = ml - (int)(start2 - ip);
lib/lz4/lz4hc_compress.c
536
ml = (int)(start2 - ip);
lib/lz4/lz4hc_compress.c
538
if (LZ4HC_encodeSequence(&ip, &op, &anchor, ml,
lib/lz4/lz4hc_compress.c
542
ip = start2;
lib/lz4/lz4hc_compress.c
74
const BYTE *ip)
lib/lz4/lz4hc_compress.c
79
U32 const target = (U32)(ip - base);
lib/lzo/lzo1x_compress.c
114
run_length = ir - ip;
lib/lzo/lzo1x_compress.c
120
dict[t] = (lzo_dict_t) (ip - in);
lib/lzo/lzo1x_compress.c
127
t = ip - ii;
lib/lzo/lzo1x_compress.c
171
ip += run_length;
lib/lzo/lzo1x_compress.c
186
v = get_unaligned((const u64 *) (ip + m_len)) ^
lib/lzo/lzo1x_compress.c
191
v = get_unaligned((const u64 *) (ip + m_len)) ^
lib/lzo/lzo1x_compress.c
193
if (unlikely(ip + m_len >= ip_end))
lib/lzo/lzo1x_compress.c
206
v = get_unaligned((const u32 *) (ip + m_len)) ^
lib/lzo/lzo1x_compress.c
211
v = get_unaligned((const u32 *) (ip + m_len)) ^
lib/lzo/lzo1x_compress.c
216
v = get_unaligned((const u32 *) (ip + m_len)) ^
lib/lzo/lzo1x_compress.c
218
if (unlikely(ip + m_len >= ip_end))
lib/lzo/lzo1x_compress.c
230
if (unlikely(ip[m_len] == m_pos[m_len])) {
lib/lzo/lzo1x_compress.c
233
if (ip[m_len] != m_pos[m_len])
lib/lzo/lzo1x_compress.c
236
if (ip[m_len] != m_pos[m_len])
lib/lzo/lzo1x_compress.c
239
if (ip[m_len] != m_pos[m_len])
lib/lzo/lzo1x_compress.c
242
if (ip[m_len] != m_pos[m_len])
lib/lzo/lzo1x_compress.c
245
if (ip[m_len] != m_pos[m_len])
lib/lzo/lzo1x_compress.c
248
if (ip[m_len] != m_pos[m_len])
lib/lzo/lzo1x_compress.c
251
if (ip[m_len] != m_pos[m_len])
lib/lzo/lzo1x_compress.c
254
if (unlikely(ip + m_len >= ip_end))
lib/lzo/lzo1x_compress.c
256
} while (ip[m_len] == m_pos[m_len]);
lib/lzo/lzo1x_compress.c
262
m_off = ip - m_pos;
lib/lzo/lzo1x_compress.c
263
ip += m_len;
lib/lzo/lzo1x_compress.c
305
ip -= m_len - 260;
lib/lzo/lzo1x_compress.c
324
ii = ip;
lib/lzo/lzo1x_compress.c
341
const unsigned char *ip = in;
lib/lzo/lzo1x_compress.c
363
uintptr_t ll_end = (uintptr_t) ip + ll;
lib/lzo/lzo1x_compress.c
371
ip, ll, &op, op_end, &t, wrkmem,
lib/lzo/lzo1x_compress.c
375
ip += ll;
lib/lzo/lzo1x_compress.c
38
const unsigned char *ip;
lib/lzo/lzo1x_compress.c
47
ip = in;
lib/lzo/lzo1x_compress.c
48
ii = ip;
lib/lzo/lzo1x_compress.c
49
ip += ti < 4 ? 4 - ti : 0;
lib/lzo/lzo1x_compress.c
57
ip += 1 + ((ip - ii) >> 5);
lib/lzo/lzo1x_compress.c
59
if (unlikely(ip >= ip_end))
lib/lzo/lzo1x_compress.c
61
dv = get_unaligned_le32(ip);
lib/lzo/lzo1x_compress.c
64
const unsigned char *ir = ip + 4;
lib/lzo/lzo1x_compress.c
65
const unsigned char *limit = min(ip_end, ip + MAX_ZERO_RUN_LENGTH + 1);
lib/lzo/lzo1x_decompress_safe.c
100
COPY8(op, ip);
lib/lzo/lzo1x_decompress_safe.c
102
ip += 8;
lib/lzo/lzo1x_decompress_safe.c
103
COPY8(op, ip);
lib/lzo/lzo1x_decompress_safe.c
105
ip += 8;
lib/lzo/lzo1x_decompress_safe.c
106
} while (ip < ie);
lib/lzo/lzo1x_decompress_safe.c
107
ip = ie;
lib/lzo/lzo1x_decompress_safe.c
115
*op++ = *ip++;
lib/lzo/lzo1x_decompress_safe.c
124
m_pos -= *ip++ << 2;
lib/lzo/lzo1x_decompress_safe.c
135
m_pos -= *ip++ << 2;
lib/lzo/lzo1x_decompress_safe.c
142
m_pos -= *ip++ << 3;
lib/lzo/lzo1x_decompress_safe.c
148
const unsigned char *ip_last = ip;
lib/lzo/lzo1x_decompress_safe.c
150
while (unlikely(*ip == 0)) {
lib/lzo/lzo1x_decompress_safe.c
151
ip++;
lib/lzo/lzo1x_decompress_safe.c
154
offset = ip - ip_last;
lib/lzo/lzo1x_decompress_safe.c
159
t += offset + 31 + *ip++;
lib/lzo/lzo1x_decompress_safe.c
163
next = get_unaligned_le16(ip);
lib/lzo/lzo1x_decompress_safe.c
164
ip += 2;
lib/lzo/lzo1x_decompress_safe.c
169
next = get_unaligned_le16(ip);
lib/lzo/lzo1x_decompress_safe.c
175
t |= ip[2] << 3;
lib/lzo/lzo1x_decompress_safe.c
181
ip += 3;
lib/lzo/lzo1x_decompress_safe.c
189
const unsigned char *ip_last = ip;
lib/lzo/lzo1x_decompress_safe.c
191
while (unlikely(*ip == 0)) {
lib/lzo/lzo1x_decompress_safe.c
192
ip++;
lib/lzo/lzo1x_decompress_safe.c
195
offset = ip - ip_last;
lib/lzo/lzo1x_decompress_safe.c
200
t += offset + 7 + *ip++;
lib/lzo/lzo1x_decompress_safe.c
202
next = get_unaligned_le16(ip);
lib/lzo/lzo1x_decompress_safe.c
204
ip += 2;
lib/lzo/lzo1x_decompress_safe.c
228
COPY4(op, ip);
lib/lzo/lzo1x_decompress_safe.c
23
#define HAVE_IP(x) ((size_t)(ip_end - ip) >= (size_t)(x))
lib/lzo/lzo1x_decompress_safe.c
230
ip += next;
lib/lzo/lzo1x_decompress_safe.c
257
COPY4(op, ip);
lib/lzo/lzo1x_decompress_safe.c
259
ip += t;
lib/lzo/lzo1x_decompress_safe.c
266
*op++ = *ip++;
lib/lzo/lzo1x_decompress_safe.c
275
ip == ip_end ? LZO_E_OK :
lib/lzo/lzo1x_decompress_safe.c
276
ip < ip_end ? LZO_E_INPUT_NOT_CONSUMED : LZO_E_INPUT_OVERRUN);
lib/lzo/lzo1x_decompress_safe.c
43
const unsigned char *ip;
lib/lzo/lzo1x_decompress_safe.c
53
ip = in;
lib/lzo/lzo1x_decompress_safe.c
58
if (likely(in_len >= 5) && likely(*ip == 17)) {
lib/lzo/lzo1x_decompress_safe.c
59
bitstream_version = ip[1];
lib/lzo/lzo1x_decompress_safe.c
60
ip += 2;
lib/lzo/lzo1x_decompress_safe.c
65
if (*ip > 17) {
lib/lzo/lzo1x_decompress_safe.c
66
t = *ip++ - 17;
lib/lzo/lzo1x_decompress_safe.c
75
t = *ip++;
lib/lzo/lzo1x_decompress_safe.c
80
const unsigned char *ip_last = ip;
lib/lzo/lzo1x_decompress_safe.c
82
while (unlikely(*ip == 0)) {
lib/lzo/lzo1x_decompress_safe.c
83
ip++;
lib/lzo/lzo1x_decompress_safe.c
86
offset = ip - ip_last;
lib/lzo/lzo1x_decompress_safe.c
91
t += offset + 15 + *ip++;
lib/lzo/lzo1x_decompress_safe.c
97
const unsigned char *ie = ip + t;
lib/tests/test_fprobe.c
195
static notrace int entry_only_handler(struct fprobe *fp, unsigned long ip,
lib/tests/test_fprobe.c
200
KUNIT_EXPECT_EQ(current_test, ip, target_ip);
lib/tests/test_fprobe.c
208
static notrace int fprobe_entry_multi_handler(struct fprobe *fp, unsigned long ip,
lib/tests/test_fprobe.c
214
KUNIT_EXPECT_EQ(current_test, ip, target_ip);
lib/tests/test_fprobe.c
222
static notrace void fprobe_exit_multi_handler(struct fprobe *fp, unsigned long ip,
lib/tests/test_fprobe.c
230
KUNIT_EXPECT_EQ(current_test, ip, target_ip);
lib/tests/test_fprobe.c
35
static notrace int fp_entry_handler(struct fprobe *fp, unsigned long ip,
lib/tests/test_fprobe.c
41
if (ip != target_ip)
lib/tests/test_fprobe.c
42
KUNIT_EXPECT_EQ(current_test, ip, target2_ip);
lib/tests/test_fprobe.c
54
static notrace void fp_exit_handler(struct fprobe *fp, unsigned long ip,
lib/tests/test_fprobe.c
61
if (ip != target_ip) {
lib/tests/test_fprobe.c
62
KUNIT_EXPECT_EQ(current_test, ip, target2_ip);
lib/zstd/common/entropy_common.c
121
if (LIKELY(ip <= iend-7) || (ip + (bitCount>>3) <= iend-4)) {
lib/zstd/common/entropy_common.c
123
ip += bitCount>>3;
lib/zstd/common/entropy_common.c
126
bitCount -= (int)(8 * (iend - 4 - ip));
lib/zstd/common/entropy_common.c
128
ip = iend - 4;
lib/zstd/common/entropy_common.c
130
bitStream = MEM_readLE32(ip) >> bitCount;
lib/zstd/common/entropy_common.c
170
if (LIKELY(ip <= iend-7) || (ip + (bitCount>>3) <= iend-4)) {
lib/zstd/common/entropy_common.c
171
ip += bitCount>>3;
lib/zstd/common/entropy_common.c
174
bitCount -= (int)(8 * (iend - 4 - ip));
lib/zstd/common/entropy_common.c
176
ip = iend - 4;
lib/zstd/common/entropy_common.c
178
bitStream = MEM_readLE32(ip) >> bitCount;
lib/zstd/common/entropy_common.c
186
ip += (bitCount+7)>>3;
lib/zstd/common/entropy_common.c
187
return ip-istart;
lib/zstd/common/entropy_common.c
251
const BYTE* ip = (const BYTE*) src;
lib/zstd/common/entropy_common.c
256
iSize = ip[0];
lib/zstd/common/entropy_common.c
264
ip += 1;
lib/zstd/common/entropy_common.c
267
huffWeight[n] = ip[n/2] >> 4;
lib/zstd/common/entropy_common.c
268
huffWeight[n+1] = ip[n/2] & 15;
lib/zstd/common/entropy_common.c
273
oSize = FSE_decompress_wksp_bmi2(huffWeight, hwSize-1, ip+1, iSize, 6, workSpace, wkspSize, bmi2);
lib/zstd/common/entropy_common.c
48
const BYTE* ip = istart;
lib/zstd/common/entropy_common.c
72
bitStream = MEM_readLE32(ip);
lib/zstd/common/entropy_common.c
92
if (LIKELY(ip <= iend-7)) {
lib/zstd/common/entropy_common.c
93
ip += 3;
lib/zstd/common/entropy_common.c
95
bitCount -= (int)(8 * (iend - 7 - ip));
lib/zstd/common/entropy_common.c
97
ip = iend - 4;
lib/zstd/common/entropy_common.c
99
bitStream = MEM_readLE32(ip) >> bitCount;
lib/zstd/common/fse_decompress.c
251
const BYTE* ip = istart;
lib/zstd/common/fse_decompress.c
270
ip += NCountLength;
lib/zstd/common/fse_decompress.c
287
if (fastMode) return FSE_decompress_usingDTable_generic(dst, dstCapacity, ip, cSrcSize, dtable, 1);
lib/zstd/common/fse_decompress.c
288
return FSE_decompress_usingDTable_generic(dst, dstCapacity, ip, cSrcSize, dtable, 0);
lib/zstd/common/zstd_internal.h
213
const BYTE* ip = (const BYTE*)src;
lib/zstd/common/zstd_internal.h
220
COPY8(op, ip);
lib/zstd/common/zstd_internal.h
230
ZSTD_copy16(op, ip);
lib/zstd/common/zstd_internal.h
233
ip += 16;
lib/zstd/common/zstd_internal.h
235
COPY16(op, ip);
lib/zstd/common/zstd_internal.h
236
COPY16(op, ip);
lib/zstd/compress/fse_compress.c
558
const BYTE* ip=iend;
lib/zstd/compress/fse_compress.c
571
FSE_initCState2(&CState1, ct, *--ip);
lib/zstd/compress/fse_compress.c
572
FSE_initCState2(&CState2, ct, *--ip);
lib/zstd/compress/fse_compress.c
573
FSE_encodeSymbol(&bitC, &CState1, *--ip);
lib/zstd/compress/fse_compress.c
576
FSE_initCState2(&CState2, ct, *--ip);
lib/zstd/compress/fse_compress.c
577
FSE_initCState2(&CState1, ct, *--ip);
lib/zstd/compress/fse_compress.c
583
FSE_encodeSymbol(&bitC, &CState2, *--ip);
lib/zstd/compress/fse_compress.c
584
FSE_encodeSymbol(&bitC, &CState1, *--ip);
lib/zstd/compress/fse_compress.c
589
while ( ip>istart ) {
lib/zstd/compress/fse_compress.c
591
FSE_encodeSymbol(&bitC, &CState2, *--ip);
lib/zstd/compress/fse_compress.c
596
FSE_encodeSymbol(&bitC, &CState1, *--ip);
lib/zstd/compress/fse_compress.c
599
FSE_encodeSymbol(&bitC, &CState2, *--ip);
lib/zstd/compress/fse_compress.c
600
FSE_encodeSymbol(&bitC, &CState1, *--ip);
lib/zstd/compress/hist.c
102
{ U32 cached = MEM_read32(ip); ip += 4;
lib/zstd/compress/hist.c
103
while (ip < iend-15) {
lib/zstd/compress/hist.c
104
U32 c = cached; cached = MEM_read32(ip); ip += 4;
lib/zstd/compress/hist.c
109
c = cached; cached = MEM_read32(ip); ip += 4;
lib/zstd/compress/hist.c
114
c = cached; cached = MEM_read32(ip); ip += 4;
lib/zstd/compress/hist.c
119
c = cached; cached = MEM_read32(ip); ip += 4;
lib/zstd/compress/hist.c
125
ip-=4;
lib/zstd/compress/hist.c
129
while (ip<iend) Counting1[*ip++]++;
lib/zstd/compress/hist.c
32
const BYTE* ip = (const BYTE*)src;
lib/zstd/compress/hist.c
33
const BYTE* const end = ip + srcSize;
lib/zstd/compress/hist.c
35
while (ip<end) {
lib/zstd/compress/hist.c
36
count[*ip++]++;
lib/zstd/compress/hist.c
43
const BYTE* ip = (const BYTE*)src;
lib/zstd/compress/hist.c
44
const BYTE* const end = ip + srcSize;
lib/zstd/compress/hist.c
51
while (ip<end) {
lib/zstd/compress/hist.c
52
assert(*ip <= maxSymbolValue);
lib/zstd/compress/hist.c
53
count[*ip++]++;
lib/zstd/compress/hist.c
83
const BYTE* ip = (const BYTE*)source;
lib/zstd/compress/hist.c
84
const BYTE* const iend = ip+sourceSize;
lib/zstd/compress/huf_compress.c
1009
HUF_encodeSymbol(bitC, ip[n - u], ct, 0, 1);
lib/zstd/compress/huf_compress.c
1011
HUF_encodeSymbol(bitC, ip[n - kUnroll], ct, 0, kLastFast);
lib/zstd/compress/huf_compress.c
1021
HUF_encodeSymbol(bitC, ip[n - u], ct, /* idx */ 0, /* fast */ 1);
lib/zstd/compress/huf_compress.c
1023
HUF_encodeSymbol(bitC, ip[n - kUnroll], ct, /* idx */ 0, /* fast */ kLastFast);
lib/zstd/compress/huf_compress.c
1031
HUF_encodeSymbol(bitC, ip[n - kUnroll - u], ct, /* idx */ 1, /* fast */ 1);
lib/zstd/compress/huf_compress.c
1033
HUF_encodeSymbol(bitC, ip[n - kUnroll - kUnroll], ct, /* idx */ 1, /* fast */ kLastFast);
lib/zstd/compress/huf_compress.c
1060
const BYTE* ip = (const BYTE*) src;
lib/zstd/compress/huf_compress.c
1072
HUF_compress1X_usingCTable_internal_body_loop(&bitC, ip, srcSize, ct, /* kUnroll */ MEM_32bits() ? 2 : 4, /* kFast */ 0, /* kLastFast */ 0);
lib/zstd/compress/huf_compress.c
1077
HUF_compress1X_usingCTable_internal_body_loop(&bitC, ip, srcSize, ct, /* kUnroll */ 2, /* kFastFlush */ 1, /* kLastFast */ 0);
lib/zstd/compress/huf_compress.c
1082
HUF_compress1X_usingCTable_internal_body_loop(&bitC, ip, srcSize, ct, /* kUnroll */ 2, /* kFastFlush */ 1, /* kLastFast */ 1);
lib/zstd/compress/huf_compress.c
1086
HUF_compress1X_usingCTable_internal_body_loop(&bitC, ip, srcSize, ct, /* kUnroll */ 3, /* kFastFlush */ 1, /* kLastFast */ 1);
lib/zstd/compress/huf_compress.c
1092
HUF_compress1X_usingCTable_internal_body_loop(&bitC, ip, srcSize, ct, /* kUnroll */ 5, /* kFastFlush */ 1, /* kLastFast */ 0);
lib/zstd/compress/huf_compress.c
1095
HUF_compress1X_usingCTable_internal_body_loop(&bitC, ip, srcSize, ct, /* kUnroll */ 5, /* kFastFlush */ 1, /* kLastFast */ 1);
lib/zstd/compress/huf_compress.c
1098
HUF_compress1X_usingCTable_internal_body_loop(&bitC, ip, srcSize, ct, /* kUnroll */ 6, /* kFastFlush */ 1, /* kLastFast */ 0);
lib/zstd/compress/huf_compress.c
1101
HUF_compress1X_usingCTable_internal_body_loop(&bitC, ip, srcSize, ct, /* kUnroll */ 7, /* kFastFlush */ 1, /* kLastFast */ 0);
lib/zstd/compress/huf_compress.c
1104
HUF_compress1X_usingCTable_internal_body_loop(&bitC, ip, srcSize, ct, /* kUnroll */ 8, /* kFastFlush */ 1, /* kLastFast */ 0);
lib/zstd/compress/huf_compress.c
1108
HUF_compress1X_usingCTable_internal_body_loop(&bitC, ip, srcSize, ct, /* kUnroll */ 9, /* kFastFlush */ 1, /* kLastFast */ 1);
lib/zstd/compress/huf_compress.c
1171
const BYTE* ip = (const BYTE*) src;
lib/zstd/compress/huf_compress.c
1172
const BYTE* const iend = ip + srcSize;
lib/zstd/compress/huf_compress.c
1182
{ CHECK_V_F(cSize, HUF_compress1X_usingCTable_internal(op, (size_t)(oend-op), ip, segmentSize, CTable, flags) );
lib/zstd/compress/huf_compress.c
1188
ip += segmentSize;
lib/zstd/compress/huf_compress.c
1190
{ CHECK_V_F(cSize, HUF_compress1X_usingCTable_internal(op, (size_t)(oend-op), ip, segmentSize, CTable, flags) );
lib/zstd/compress/huf_compress.c
1196
ip += segmentSize;
lib/zstd/compress/huf_compress.c
1198
{ CHECK_V_F(cSize, HUF_compress1X_usingCTable_internal(op, (size_t)(oend-op), ip, segmentSize, CTable, flags) );
lib/zstd/compress/huf_compress.c
1204
ip += segmentSize;
lib/zstd/compress/huf_compress.c
1206
assert(ip <= iend);
lib/zstd/compress/huf_compress.c
1207
{ CHECK_V_F(cSize, HUF_compress1X_usingCTable_internal(op, (size_t)(oend-op), ip, (size_t)(iend-ip), CTable, flags) );
lib/zstd/compress/huf_compress.c
990
const BYTE* ip, size_t srcSize,
lib/zstd/compress/huf_compress.c
999
HUF_encodeSymbol(bitC, ip[--n], ct, 0, /* fast */ 0);
lib/zstd/compress/zstd_compress.c
3487
const BYTE* ip = src;
lib/zstd/compress/zstd_compress.c
3488
const BYTE value = ip[0];
lib/zstd/compress/zstd_compress.c
3496
if (prefixLength && ZSTD_count(ip+1, ip, ip+prefixLength) != prefixLength-1) {
lib/zstd/compress/zstd_compress.c
3502
if (MEM_readST(ip + i + u) != valueST) {
lib/zstd/compress/zstd_compress.c
4028
const BYTE* ip = (const BYTE*)src;
lib/zstd/compress/zstd_compress.c
4066
cSize = ZSTD_noCompressBlock(op, dstCapacity, ip, srcSize, lastBlock);
lib/zstd/compress/zstd_compress.c
4071
cSize = ZSTD_rleCompressBlock(op, dstCapacity, *ip, srcSize, lastBlock);
lib/zstd/compress/zstd_compress.c
4181
const BYTE* ip = (const BYTE*)src;
lib/zstd/compress/zstd_compress.c
4219
ip, blockSize,
lib/zstd/compress/zstd_compress.c
4247
ip, srcBytes,
lib/zstd/compress/zstd_compress.c
4253
ip += srcBytes;
lib/zstd/compress/zstd_compress.c
4307
const BYTE* ip = (const BYTE*)src;
lib/zstd/compress/zstd_compress.c
4344
ZSTD_isRLE(ip, srcSize))
lib/zstd/compress/zstd_compress.c
4347
op[0] = ip[0];
lib/zstd/compress/zstd_compress.c
4443
void const* ip,
lib/zstd/compress/zstd_compress.c
4448
if (ZSTD_window_needOverflowCorrection(ms->window, cycleLog, maxDist, ms->loadedDictEnd, ip, iend)) {
lib/zstd/compress/zstd_compress.c
4449
U32 const correction = ZSTD_window_correctOverflow(&ms->window, cycleLog, maxDist, ip);
lib/zstd/compress/zstd_compress.c
4512
const BYTE* ip = (const BYTE*)src;
lib/zstd/compress/zstd_compress.c
4527
ip, remaining,
lib/zstd/compress/zstd_compress.c
4542
ms, &cctx->workspace, &cctx->appliedParams, ip, ip + blockSize);
lib/zstd/compress/zstd_compress.c
4543
ZSTD_checkDictValidity(&ms->window, ip + blockSize, maxDist, &ms->loadedDictEnd, &ms->dictMatchState);
lib/zstd/compress/zstd_compress.c
4544
ZSTD_window_enforceMaxDist(&ms->window, ip, maxDist, &ms->loadedDictEnd, &ms->dictMatchState);
lib/zstd/compress/zstd_compress.c
4551
cSize = ZSTD_compressBlock_targetCBlockSize(cctx, op, dstCapacity, ip, blockSize, lastBlock);
lib/zstd/compress/zstd_compress.c
4556
cSize = ZSTD_compressBlock_splitBlock(cctx, op, dstCapacity, ip, blockSize, lastBlock);
lib/zstd/compress/zstd_compress.c
4562
ip, blockSize, 1 /* frame */);
lib/zstd/compress/zstd_compress.c
4566
cSize = ZSTD_noCompressBlock(op, dstCapacity, ip, blockSize, lastBlock);
lib/zstd/compress/zstd_compress.c
4593
ip += blockSize;
lib/zstd/compress/zstd_compress.c
4824
const BYTE* ip = (const BYTE*) src;
lib/zstd/compress/zstd_compress.c
4825
const BYTE* const iend = ip + srcSize;
lib/zstd/compress/zstd_compress.c
4853
ip = iend - maxDictSize;
lib/zstd/compress/zstd_compress.c
4854
src = ip;
lib/zstd/compress/zstd_compress.c
4872
ZSTD_ldm_fillHashTable(ls, ip, iend, ¶ms->ldmParams);
lib/zstd/compress/zstd_compress.c
4879
ip = iend - maxDictSize;
lib/zstd/compress/zstd_compress.c
4880
src = ip;
lib/zstd/compress/zstd_compress.c
4885
ms->nextToUpdate = (U32)(ip - ms->window.base);
lib/zstd/compress/zstd_compress.c
4891
ZSTD_overflowCorrectIfNeeded(ms, ws, params, ip, iend);
lib/zstd/compress/zstd_compress.c
5996
const char* ip = (istart != NULL) ? istart + input->pos : istart;
lib/zstd/compress/zstd_compress.c
6008
if (ip) ip -= zcs->stableIn_notConsumed;
lib/zstd/compress/zstd_compress.c
6033
&& ( (size_t)(oend-op) >= ZSTD_compressBound((size_t)(iend-ip)) /* Enough output space */
lib/zstd/compress/zstd_compress.c
6039
ip, (size_t)(iend-ip));
lib/zstd/compress/zstd_compress.c
6042
ip = iend;
lib/zstd/compress/zstd_compress.c
6053
ip, (size_t)(iend-ip));
lib/zstd/compress/zstd_compress.c
6055
if (ip) ip += loaded;
lib/zstd/compress/zstd_compress.c
6069
&& ( (size_t)(iend - ip) < zcs->blockSizeMax) ) {
lib/zstd/compress/zstd_compress.c
6071
zcs->stableIn_notConsumed = (size_t)(iend - ip);
lib/zstd/compress/zstd_compress.c
6072
ip = iend; /* pretend to have consumed input */
lib/zstd/compress/zstd_compress.c
6076
&& (ip == iend) ) {
lib/zstd/compress/zstd_compress.c
6088
: MIN((size_t)(iend - ip), zcs->blockSizeMax);
lib/zstd/compress/zstd_compress.c
6094
unsigned const lastBlock = (flushMode == ZSTD_e_end) && (ip==iend);
lib/zstd/compress/zstd_compress.c
6112
unsigned const lastBlock = (flushMode == ZSTD_e_end) && (ip + iSize == iend);
lib/zstd/compress/zstd_compress.c
6114
ZSTD_compressEnd_public(zcs, cDst, oSize, ip, iSize) :
lib/zstd/compress/zstd_compress.c
6115
ZSTD_compressContinue_public(zcs, cDst, oSize, ip, iSize);
lib/zstd/compress/zstd_compress.c
6117
if (ip) ip += iSize;
lib/zstd/compress/zstd_compress.c
6120
if (lastBlock) assert(ip == iend);
lib/zstd/compress/zstd_compress.c
6169
input->pos = (size_t)(ip - istart);
lib/zstd/compress/zstd_compress.c
6444
BYTE const* ip = (BYTE const*)(src);
lib/zstd/compress/zstd_compress.c
6445
const BYTE* const iend = ip + blockSize;
lib/zstd/compress/zstd_compress.c
6483
ZSTD_storeSeq(&cctx->seqStore, litLength, ip, iend, offBase, matchLength);
lib/zstd/compress/zstd_compress.c
6484
ip += matchLength + litLength;
lib/zstd/compress/zstd_compress.c
6515
ZSTD_storeLastLiterals(&cctx->seqStore, ip, inSeqs[idx].litLength);
lib/zstd/compress/zstd_compress.c
6516
ip += inSeqs[idx].litLength;
lib/zstd/compress/zstd_compress.c
6519
RETURN_ERROR_IF(ip != iend, externalSequences_invalid, "Blocksize doesn't agree with block delimiter!");
lib/zstd/compress/zstd_compress.c
6547
const BYTE* ip = istart;
lib/zstd/compress/zstd_compress.c
6637
ZSTD_storeSeq(&cctx->seqStore, litLength, ip, iend, offBase, matchLength);
lib/zstd/compress/zstd_compress.c
6638
ip += matchLength + litLength;
lib/zstd/compress/zstd_compress.c
6649
if (ip != iend) {
lib/zstd/compress/zstd_compress.c
6651
U32 const lastLLSize = (U32)(iend - ip);
lib/zstd/compress/zstd_compress.c
6652
assert(ip <= iend);
lib/zstd/compress/zstd_compress.c
6654
ZSTD_storeLastLiterals(&cctx->seqStore, ip, lastLLSize);
lib/zstd/compress/zstd_compress.c
6745
const BYTE* ip = (BYTE const*)src;
lib/zstd/compress/zstd_compress.c
6773
ip, blockSize,
lib/zstd/compress/zstd_compress.c
6781
cBlockSize = ZSTD_noCompressBlock(op, dstCapacity, ip, blockSize, lastBlock);
lib/zstd/compress/zstd_compress.c
6785
ip += blockSize;
lib/zstd/compress/zstd_compress.c
6805
ZSTD_isRLE(ip, blockSize)) {
lib/zstd/compress/zstd_compress.c
6815
cBlockSize = ZSTD_noCompressBlock(op, dstCapacity, ip, blockSize, lastBlock);
lib/zstd/compress/zstd_compress.c
6819
cBlockSize = ZSTD_rleCompressBlock(op, dstCapacity, *ip, blockSize, lastBlock);
lib/zstd/compress/zstd_compress.c
6841
ip += blockSize;
lib/zstd/compress/zstd_compress_internal.h
1350
BYTE const* const ip = (BYTE const*)src;
lib/zstd/compress/zstd_compress_internal.h
1366
window->base = ip - distanceFromBase;
lib/zstd/compress/zstd_compress_internal.h
1371
window->nextSrc = ip + srcSize;
lib/zstd/compress/zstd_compress_internal.h
1373
if ( (ip+srcSize > window->dictBase + window->lowLimit)
lib/zstd/compress/zstd_compress_internal.h
1374
& (ip < window->dictBase + window->dictLimit)) {
lib/zstd/compress/zstd_compress_internal.h
1375
size_t const highInputIdx = (size_t)((ip + srcSize) - window->dictBase);
lib/zstd/compress/zstd_compress_internal.h
698
ZSTD_safecopyLiterals(BYTE* op, BYTE const* ip, BYTE const* const iend, BYTE const* ilimit_w)
lib/zstd/compress/zstd_compress_internal.h
701
if (ip <= ilimit_w) {
lib/zstd/compress/zstd_compress_internal.h
702
ZSTD_wildcopy(op, ip, ilimit_w - ip, ZSTD_no_overlap);
lib/zstd/compress/zstd_compress_internal.h
703
op += ilimit_w - ip;
lib/zstd/compress/zstd_compress_internal.h
704
ip = ilimit_w;
lib/zstd/compress/zstd_compress_internal.h
706
while (ip < iend) *op++ = *ip++;
lib/zstd/compress/zstd_compress_internal.h
872
ZSTD_count_2segments(const BYTE* ip, const BYTE* match,
lib/zstd/compress/zstd_compress_internal.h
875
const BYTE* const vEnd = MIN( ip + (mEnd - match), iEnd);
lib/zstd/compress/zstd_compress_internal.h
876
size_t const matchLength = ZSTD_count(ip, match, vEnd);
lib/zstd/compress/zstd_compress_internal.h
880
DEBUGLOG(7, "distance from current pos to end buffer = %i", (int)(iEnd - ip));
lib/zstd/compress/zstd_compress_internal.h
881
DEBUGLOG(7, "next byte : ip==%02X, istart==%02X", ip[matchLength], *iStart);
lib/zstd/compress/zstd_compress_internal.h
882
DEBUGLOG(7, "final match length = %zu", matchLength + ZSTD_count(ip+matchLength, iStart, iEnd));
lib/zstd/compress/zstd_compress_internal.h
883
return matchLength + ZSTD_count(ip+matchLength, iStart, iEnd);
lib/zstd/compress/zstd_compress_literals.c
25
const BYTE* const ip = (const BYTE*)src;
lib/zstd/compress/zstd_compress_literals.c
28
RAWLOG(5, " %02X", ip[u]); (void)ip;
lib/zstd/compress/zstd_compress_superblock.c
498
BYTE const* ip = (BYTE const*)src;
lib/zstd/compress/zstd_compress_superblock.c
499
BYTE const* const iend = ip + srcSize;
lib/zstd/compress/zstd_compress_superblock.c
566
assert(ip + decompressedSize <= iend);
lib/zstd/compress/zstd_compress_superblock.c
567
ip += decompressedSize;
lib/zstd/compress/zstd_compress_superblock.c
610
assert(ip + decompressedSize <= iend);
lib/zstd/compress/zstd_compress_superblock.c
611
ip += decompressedSize;
lib/zstd/compress/zstd_compress_superblock.c
641
if (ip < iend) {
lib/zstd/compress/zstd_compress_superblock.c
643
size_t const rSize = (size_t)((iend - ip));
lib/zstd/compress/zstd_compress_superblock.c
644
size_t const cSize = ZSTD_noCompressBlock(op, (size_t)(oend - op), ip, rSize, lastBlock);
lib/zstd/compress/zstd_double_fast.c
148
const BYTE* ip = istart; /* the current position */
lib/zstd/compress/zstd_double_fast.c
158
ip += ((ip - prefixLowest) == 0);
lib/zstd/compress/zstd_double_fast.c
160
U32 const current = (U32)(ip - base);
lib/zstd/compress/zstd_double_fast.c
170
nextStep = ip + kStepIncr;
lib/zstd/compress/zstd_double_fast.c
171
ip1 = ip + step;
lib/zstd/compress/zstd_double_fast.c
177
hl0 = ZSTD_hashPtr(ip, hBitsL, 8);
lib/zstd/compress/zstd_double_fast.c
183
const size_t hs0 = ZSTD_hashPtr(ip, hBitsS, mls);
lib/zstd/compress/zstd_double_fast.c
185
curr = (U32)(ip-base);
lib/zstd/compress/zstd_double_fast.c
191
if ((offset_1 > 0) & (MEM_read32(ip+1-offset_1) == MEM_read32(ip+1))) {
lib/zstd/compress/zstd_double_fast.c
192
mLength = ZSTD_count(ip+1+4, ip+1+4-offset_1, iend) + 4;
lib/zstd/compress/zstd_double_fast.c
193
ip++;
lib/zstd/compress/zstd_double_fast.c
194
ZSTD_storeSeq(seqStore, (size_t)(ip-anchor), anchor, iend, REPCODE1_TO_OFFBASE, mLength);
lib/zstd/compress/zstd_double_fast.c
207
if (MEM_read64(matchl0_safe) == MEM_read64(ip) && matchl0_safe == matchl0) {
lib/zstd/compress/zstd_double_fast.c
208
mLength = ZSTD_count(ip+8, matchl0+8, iend) + 8;
lib/zstd/compress/zstd_double_fast.c
209
offset = (U32)(ip-matchl0);
lib/zstd/compress/zstd_double_fast.c
210
while (((ip>anchor) & (matchl0>prefixLowest)) && (ip[-1] == matchl0[-1])) { ip--; matchl0--; mLength++; } /* catch up */
lib/zstd/compress/zstd_double_fast.c
221
if(MEM_read32(matchs0_safe) == MEM_read32(ip) && matchs0_safe == matchs0) {
lib/zstd/compress/zstd_double_fast.c
231
ip = ip1;
lib/zstd/compress/zstd_double_fast.c
238
PREFETCH_L1(ip+256);
lib/zstd/compress/zstd_double_fast.c
257
mLength = ZSTD_count(ip+4, matchs0+4, iend) + 4;
lib/zstd/compress/zstd_double_fast.c
258
offset = (U32)(ip - matchs0);
lib/zstd/compress/zstd_double_fast.c
265
ip = ip1;
lib/zstd/compress/zstd_double_fast.c
267
offset = (U32)(ip-matchl1);
lib/zstd/compress/zstd_double_fast.c
272
while (((ip>anchor) & (matchs0>prefixLowest)) && (ip[-1] == matchs0[-1])) { ip--; matchs0--; mLength++; } /* complete backward */
lib/zstd/compress/zstd_double_fast.c
29
const BYTE* ip = base + ms->nextToUpdate;
lib/zstd/compress/zstd_double_fast.c
291
ZSTD_storeSeq(seqStore, (size_t)(ip-anchor), anchor, iend, OFFSET_TO_OFFBASE(offset), mLength);
lib/zstd/compress/zstd_double_fast.c
295
ip += mLength;
lib/zstd/compress/zstd_double_fast.c
296
anchor = ip;
lib/zstd/compress/zstd_double_fast.c
298
if (ip <= ilimit) {
lib/zstd/compress/zstd_double_fast.c
303
hashLong[ZSTD_hashPtr(ip-2, hBitsL, 8)] = (U32)(ip-2-base);
lib/zstd/compress/zstd_double_fast.c
305
hashSmall[ZSTD_hashPtr(ip-1, hBitsS, mls)] = (U32)(ip-1-base);
lib/zstd/compress/zstd_double_fast.c
309
while ( (ip <= ilimit)
lib/zstd/compress/zstd_double_fast.c
311
& (MEM_read32(ip) == MEM_read32(ip - offset_2)) )) {
lib/zstd/compress/zstd_double_fast.c
313
size_t const rLength = ZSTD_count(ip+4, ip+4-offset_2, iend) + 4;
lib/zstd/compress/zstd_double_fast.c
315
hashSmall[ZSTD_hashPtr(ip, hBitsS, mls)] = (U32)(ip-base);
lib/zstd/compress/zstd_double_fast.c
316
hashLong[ZSTD_hashPtr(ip, hBitsL, 8)] = (U32)(ip-base);
lib/zstd/compress/zstd_double_fast.c
318
ip += rLength;
lib/zstd/compress/zstd_double_fast.c
319
anchor = ip;
lib/zstd/compress/zstd_double_fast.c
341
const BYTE* ip = istart;
lib/zstd/compress/zstd_double_fast.c
362
const U32 dictAndPrefixLength = (U32)((ip - prefixLowest) + (dictEnd - dictStart));
lib/zstd/compress/zstd_double_fast.c
37
for (; ip + fastHashFillStep - 1 <= iend; ip += fastHashFillStep) {
lib/zstd/compress/zstd_double_fast.c
377
ip += (dictAndPrefixLength == 0);
lib/zstd/compress/zstd_double_fast.c
38
U32 const curr = (U32)(ip - base);
lib/zstd/compress/zstd_double_fast.c
385
while (ip < ilimit) { /* < instead of <=, because repcode check at (ip+1) */
lib/zstd/compress/zstd_double_fast.c
388
size_t const h2 = ZSTD_hashPtr(ip, hBitsL, 8);
lib/zstd/compress/zstd_double_fast.c
389
size_t const h = ZSTD_hashPtr(ip, hBitsS, mls);
lib/zstd/compress/zstd_double_fast.c
390
size_t const dictHashAndTagL = ZSTD_hashPtr(ip, dictHBitsL, 8);
lib/zstd/compress/zstd_double_fast.c
391
size_t const dictHashAndTagS = ZSTD_hashPtr(ip, dictHBitsS, mls);
lib/zstd/compress/zstd_double_fast.c
396
U32 const curr = (U32)(ip-base);
lib/zstd/compress/zstd_double_fast.c
409
&& (MEM_read32(repMatch) == MEM_read32(ip+1)) ) {
lib/zstd/compress/zstd_double_fast.c
41
size_t const smHashAndTag = ZSTD_hashPtr(ip + i, hBitsS, mls);
lib/zstd/compress/zstd_double_fast.c
411
mLength = ZSTD_count_2segments(ip+1+4, repMatch+4, iend, repMatchEnd, prefixLowest) + 4;
lib/zstd/compress/zstd_double_fast.c
412
ip++;
lib/zstd/compress/zstd_double_fast.c
413
ZSTD_storeSeq(seqStore, (size_t)(ip-anchor), anchor, iend, REPCODE1_TO_OFFBASE, mLength);
lib/zstd/compress/zstd_double_fast.c
417
if ((matchIndexL >= prefixLowestIndex) && (MEM_read64(matchLong) == MEM_read64(ip))) {
lib/zstd/compress/zstd_double_fast.c
419
mLength = ZSTD_count(ip+8, matchLong+8, iend) + 8;
lib/zstd/compress/zstd_double_fast.c
42
size_t const lgHashAndTag = ZSTD_hashPtr(ip + i, hBitsL, 8);
lib/zstd/compress/zstd_double_fast.c
420
offset = (U32)(ip-matchLong);
lib/zstd/compress/zstd_double_fast.c
421
while (((ip>anchor) & (matchLong>prefixLowest)) && (ip[-1] == matchLong[-1])) { ip--; matchLong--; mLength++; } /* catch up */
lib/zstd/compress/zstd_double_fast.c
429
if (dictMatchL > dictStart && MEM_read64(dictMatchL) == MEM_read64(ip)) {
lib/zstd/compress/zstd_double_fast.c
430
mLength = ZSTD_count_2segments(ip+8, dictMatchL+8, iend, dictEnd, prefixLowest) + 8;
lib/zstd/compress/zstd_double_fast.c
432
while (((ip>anchor) & (dictMatchL>dictStart)) && (ip[-1] == dictMatchL[-1])) { ip--; dictMatchL--; mLength++; } /* catch up */
lib/zstd/compress/zstd_double_fast.c
438
if (MEM_read32(match) == MEM_read32(ip)) {
lib/zstd/compress/zstd_double_fast.c
447
if (match > dictStart && MEM_read32(match) == MEM_read32(ip)) {
lib/zstd/compress/zstd_double_fast.c
451
ip += ((ip-anchor) >> kSearchStrength) + 1;
lib/zstd/compress/zstd_double_fast.c
453
PREFETCH_L1(ip+256);
lib/zstd/compress/zstd_double_fast.c
458
{ size_t const hl3 = ZSTD_hashPtr(ip+1, hBitsL, 8);
lib/zstd/compress/zstd_double_fast.c
459
size_t const dictHashAndTagL3 = ZSTD_hashPtr(ip+1, dictHBitsL, 8);
lib/zstd/compress/zstd_double_fast.c
467
if ((matchIndexL3 >= prefixLowestIndex) && (MEM_read64(matchL3) == MEM_read64(ip+1))) {
lib/zstd/compress/zstd_double_fast.c
468
mLength = ZSTD_count(ip+9, matchL3+8, iend) + 8;
lib/zstd/compress/zstd_double_fast.c
469
ip++;
lib/zstd/compress/zstd_double_fast.c
470
offset = (U32)(ip-matchL3);
lib/zstd/compress/zstd_double_fast.c
471
while (((ip>anchor) & (matchL3>prefixLowest)) && (ip[-1] == matchL3[-1])) { ip--; matchL3--; mLength++; } /* catch up */
lib/zstd/compress/zstd_double_fast.c
478
if (dictMatchL3 > dictStart && MEM_read64(dictMatchL3) == MEM_read64(ip+1)) {
lib/zstd/compress/zstd_double_fast.c
479
mLength = ZSTD_count_2segments(ip+1+8, dictMatchL3+8, iend, dictEnd, prefixLowest) + 8;
lib/zstd/compress/zstd_double_fast.c
480
ip++;
lib/zstd/compress/zstd_double_fast.c
482
while (((ip>anchor) & (dictMatchL3>dictStart)) && (ip[-1] == dictMatchL3[-1])) { ip--; dictMatchL3--; mLength++; } /* catch up */
lib/zstd/compress/zstd_double_fast.c
488
mLength = ZSTD_count_2segments(ip+4, match+4, iend, dictEnd, prefixLowest) + 4;
lib/zstd/compress/zstd_double_fast.c
490
while (((ip>anchor) & (match>dictStart)) && (ip[-1] == match[-1])) { ip--; match--; mLength++; } /* catch up */
lib/zstd/compress/zstd_double_fast.c
492
mLength = ZSTD_count(ip+4, match+4, iend) + 4;
lib/zstd/compress/zstd_double_fast.c
493
offset = (U32)(ip - match);
lib/zstd/compress/zstd_double_fast.c
494
while (((ip>anchor) & (match>prefixLowest)) && (ip[-1] == match[-1])) { ip--; match--; mLength++; } /* catch up */
lib/zstd/compress/zstd_double_fast.c
501
ZSTD_storeSeq(seqStore, (size_t)(ip-anchor), anchor, iend, OFFSET_TO_OFFBASE(offset), mLength);
lib/zstd/compress/zstd_double_fast.c
505
ip += mLength;
lib/zstd/compress/zstd_double_fast.c
506
anchor = ip;
lib/zstd/compress/zstd_double_fast.c
508
if (ip <= ilimit) {
lib/zstd/compress/zstd_double_fast.c
513
hashLong[ZSTD_hashPtr(ip-2, hBitsL, 8)] = (U32)(ip-2-base);
lib/zstd/compress/zstd_double_fast.c
515
hashSmall[ZSTD_hashPtr(ip-1, hBitsS, mls)] = (U32)(ip-1-base);
lib/zstd/compress/zstd_double_fast.c
519
while (ip <= ilimit) {
lib/zstd/compress/zstd_double_fast.c
520
U32 const current2 = (U32)(ip-base);
lib/zstd/compress/zstd_double_fast.c
526
&& (MEM_read32(repMatch2) == MEM_read32(ip)) ) {
lib/zstd/compress/zstd_double_fast.c
528
size_t const repLength2 = ZSTD_count_2segments(ip+4, repMatch2+4, iend, repEnd2, prefixLowest) + 4;
lib/zstd/compress/zstd_double_fast.c
531
hashSmall[ZSTD_hashPtr(ip, hBitsS, mls)] = current2;
lib/zstd/compress/zstd_double_fast.c
532
hashLong[ZSTD_hashPtr(ip, hBitsL, 8)] = current2;
lib/zstd/compress/zstd_double_fast.c
533
ip += repLength2;
lib/zstd/compress/zstd_double_fast.c
534
anchor = ip;
lib/zstd/compress/zstd_double_fast.c
622
const BYTE* ip = istart;
lib/zstd/compress/zstd_double_fast.c
645
while (ip < ilimit) { /* < instead of <=, because (ip+1) */
lib/zstd/compress/zstd_double_fast.c
646
const size_t hSmall = ZSTD_hashPtr(ip, hBitsS, mls);
lib/zstd/compress/zstd_double_fast.c
651
const size_t hLong = ZSTD_hashPtr(ip, hBitsL, 8);
lib/zstd/compress/zstd_double_fast.c
656
const U32 curr = (U32)(ip-base);
lib/zstd/compress/zstd_double_fast.c
665
&& (MEM_read32(repMatch) == MEM_read32(ip+1)) ) {
lib/zstd/compress/zstd_double_fast.c
667
mLength = ZSTD_count_2segments(ip+1+4, repMatch+4, iend, repMatchEnd, prefixStart) + 4;
lib/zstd/compress/zstd_double_fast.c
668
ip++;
lib/zstd/compress/zstd_double_fast.c
669
ZSTD_storeSeq(seqStore, (size_t)(ip-anchor), anchor, iend, REPCODE1_TO_OFFBASE, mLength);
lib/zstd/compress/zstd_double_fast.c
67
const BYTE* ip = base + ms->nextToUpdate;
lib/zstd/compress/zstd_double_fast.c
671
if ((matchLongIndex > dictStartIndex) && (MEM_read64(matchLong) == MEM_read64(ip))) {
lib/zstd/compress/zstd_double_fast.c
675
mLength = ZSTD_count_2segments(ip+8, matchLong+8, iend, matchEnd, prefixStart) + 8;
lib/zstd/compress/zstd_double_fast.c
677
while (((ip>anchor) & (matchLong>lowMatchPtr)) && (ip[-1] == matchLong[-1])) { ip--; matchLong--; mLength++; } /* catch up */
lib/zstd/compress/zstd_double_fast.c
680
ZSTD_storeSeq(seqStore, (size_t)(ip-anchor), anchor, iend, OFFSET_TO_OFFBASE(offset), mLength);
lib/zstd/compress/zstd_double_fast.c
682
} else if ((matchIndex > dictStartIndex) && (MEM_read32(match) == MEM_read32(ip))) {
lib/zstd/compress/zstd_double_fast.c
683
size_t const h3 = ZSTD_hashPtr(ip+1, hBitsL, 8);
lib/zstd/compress/zstd_double_fast.c
689
if ( (matchIndex3 > dictStartIndex) && (MEM_read64(match3) == MEM_read64(ip+1)) ) {
lib/zstd/compress/zstd_double_fast.c
692
mLength = ZSTD_count_2segments(ip+9, match3+8, iend, matchEnd, prefixStart) + 8;
lib/zstd/compress/zstd_double_fast.c
693
ip++;
lib/zstd/compress/zstd_double_fast.c
695
while (((ip>anchor) & (match3>lowMatchPtr)) && (ip[-1] == match3[-1])) { ip--; match3--; mLength++; } /* catch up */
lib/zstd/compress/zstd_double_fast.c
699
mLength = ZSTD_count_2segments(ip+4, match+4, iend, matchEnd, prefixStart) + 4;
lib/zstd/compress/zstd_double_fast.c
701
while (((ip>anchor) & (match>lowMatchPtr)) && (ip[-1] == match[-1])) { ip--; match--; mLength++; } /* catch up */
lib/zstd/compress/zstd_double_fast.c
705
ZSTD_storeSeq(seqStore, (size_t)(ip-anchor), anchor, iend, OFFSET_TO_OFFBASE(offset), mLength);
lib/zstd/compress/zstd_double_fast.c
708
ip += ((ip-anchor) >> kSearchStrength) + 1;
lib/zstd/compress/zstd_double_fast.c
713
ip += mLength;
lib/zstd/compress/zstd_double_fast.c
714
anchor = ip;
lib/zstd/compress/zstd_double_fast.c
716
if (ip <= ilimit) {
lib/zstd/compress/zstd_double_fast.c
721
hashLong[ZSTD_hashPtr(ip-2, hBitsL, 8)] = (U32)(ip-2-base);
lib/zstd/compress/zstd_double_fast.c
723
hashSmall[ZSTD_hashPtr(ip-1, hBitsS, mls)] = (U32)(ip-1-base);
lib/zstd/compress/zstd_double_fast.c
727
while (ip <= ilimit) {
lib/zstd/compress/zstd_double_fast.c
728
U32 const current2 = (U32)(ip-base);
lib/zstd/compress/zstd_double_fast.c
733
&& (MEM_read32(repMatch2) == MEM_read32(ip)) ) {
lib/zstd/compress/zstd_double_fast.c
735
size_t const repLength2 = ZSTD_count_2segments(ip+4, repMatch2+4, iend, repEnd2, prefixStart) + 4;
lib/zstd/compress/zstd_double_fast.c
738
hashSmall[ZSTD_hashPtr(ip, hBitsS, mls)] = current2;
lib/zstd/compress/zstd_double_fast.c
739
hashLong[ZSTD_hashPtr(ip, hBitsL, 8)] = current2;
lib/zstd/compress/zstd_double_fast.c
740
ip += repLength2;
lib/zstd/compress/zstd_double_fast.c
741
anchor = ip;
lib/zstd/compress/zstd_double_fast.c
75
for (; ip + fastHashFillStep - 1 <= iend; ip += fastHashFillStep) {
lib/zstd/compress/zstd_double_fast.c
76
U32 const curr = (U32)(ip - base);
lib/zstd/compress/zstd_double_fast.c
79
size_t const smHash = ZSTD_hashPtr(ip + i, hBitsS, mls);
lib/zstd/compress/zstd_double_fast.c
80
size_t const lgHash = ZSTD_hashPtr(ip + i, hBitsL, 8);
lib/zstd/compress/zstd_fast.c
26
const BYTE* ip = base + ms->nextToUpdate;
lib/zstd/compress/zstd_fast.c
37
for ( ; ip + fastHashFillStep < iend + 2; ip += fastHashFillStep) {
lib/zstd/compress/zstd_fast.c
38
U32 const curr = (U32)(ip - base);
lib/zstd/compress/zstd_fast.c
39
{ size_t const hashAndTag = ZSTD_hashPtr(ip, hBits, mls);
lib/zstd/compress/zstd_fast.c
46
size_t const hashAndTag = ZSTD_hashPtr(ip + p, hBits, mls);
lib/zstd/compress/zstd_fast.c
63
const BYTE* ip = base + ms->nextToUpdate;
lib/zstd/compress/zstd_fast.c
74
for ( ; ip + fastHashFillStep < iend + 2; ip += fastHashFillStep) {
lib/zstd/compress/zstd_fast.c
75
U32 const curr = (U32)(ip - base);
lib/zstd/compress/zstd_fast.c
76
size_t const hash0 = ZSTD_hashPtr(ip, hBits, mls);
lib/zstd/compress/zstd_fast.c
82
size_t const hash = ZSTD_hashPtr(ip + p, hBits, mls);
lib/zstd/compress/zstd_lazy.c
105
assert(ip < iend); /* condition for ZSTD_count */
lib/zstd/compress/zstd_lazy.c
1144
const BYTE* const ip, const BYTE* const iLimit,
lib/zstd/compress/zstd_lazy.c
1159
const U32 curr = (U32)(ip-base);
lib/zstd/compress/zstd_lazy.c
1187
ddsIdx = ZSTD_hashPtr(ip, ddsHashLog, mls) << ZSTD_LAZY_DDSS_BUCKET_LOG;
lib/zstd/compress/zstd_lazy.c
1197
U32 const dmsHash = (U32)ZSTD_hashPtr(ip, dms->rowHashLog + ZSTD_ROW_HASH_TAG_BITS, mls);
lib/zstd/compress/zstd_lazy.c
1207
ZSTD_row_update_internal(ms, ip, mls, rowLog, rowMask, 1 /* useCache */);
lib/zstd/compress/zstd_lazy.c
1213
hash = (U32)ZSTD_hashPtrSalted(ip, hashLog + ZSTD_ROW_HASH_TAG_BITS, mls, hashSalt);
lib/zstd/compress/zstd_lazy.c
124
matchLength += ZSTD_count(ip+matchLength, match+matchLength, iend);
lib/zstd/compress/zstd_lazy.c
1265
if (MEM_read32(match + ml - 3) == MEM_read32(ip + ml - 3)) /* potentially better */
lib/zstd/compress/zstd_lazy.c
1266
currentMl = ZSTD_count(ip, match, iLimit);
lib/zstd/compress/zstd_lazy.c
127
matchLength += ZSTD_count_2segments(ip+matchLength, match+matchLength, iend, dictEnd, prefixStart);
lib/zstd/compress/zstd_lazy.c
1270
if (MEM_read32(match) == MEM_read32(ip)) /* assumption : matchIndex <= dictLimit-4 (by table construction) */
lib/zstd/compress/zstd_lazy.c
1271
currentMl = ZSTD_count_2segments(ip+4, match+4, iLimit, dictEnd, prefixStart) + 4;
lib/zstd/compress/zstd_lazy.c
1278
if (ip+currentMl == iLimit) break; /* best possible, avoids read overflow on next attempt */
lib/zstd/compress/zstd_lazy.c
1286
ip, iLimit, prefixStart, curr, dictLimit, ddsIdx);
lib/zstd/compress/zstd_lazy.c
1321
if (MEM_read32(match) == MEM_read32(ip))
lib/zstd/compress/zstd_lazy.c
1322
currentMl = ZSTD_count_2segments(ip+4, match+4, iLimit, dmsEnd, prefixStart) + 4;
lib/zstd/compress/zstd_lazy.c
1329
if (ip+currentMl == iLimit) break;
lib/zstd/compress/zstd_lazy.c
135
if (ip+matchLength == iend) { /* equal : no way to know if inf or sup */
lib/zstd/compress/zstd_lazy.c
1369
const BYTE* ip, const BYTE* const iLimit, \
lib/zstd/compress/zstd_lazy.c
1373
return ZSTD_BtFindBestMatch(ms, ip, iLimit, offBasePtr, mls, ZSTD_##dictMode); \
lib/zstd/compress/zstd_lazy.c
1379
const BYTE* ip, const BYTE* const iLimit, \
lib/zstd/compress/zstd_lazy.c
1383
return ZSTD_HcFindBestMatch(ms, ip, iLimit, offsetPtr, mls, ZSTD_##dictMode); \
lib/zstd/compress/zstd_lazy.c
1389
const BYTE* ip, const BYTE* const iLimit, \
lib/zstd/compress/zstd_lazy.c
139
if (match[matchLength] < ip[matchLength]) { /* necessarily within buffer */
lib/zstd/compress/zstd_lazy.c
1394
return ZSTD_RowFindBestMatch(ms, ip, iLimit, offsetPtr, mls, ZSTD_##dictMode, rowLog); \
lib/zstd/compress/zstd_lazy.c
1429
return ZSTD_BT_SEARCH_FN(dictMode, mls)(ms, ip, iend, offsetPtr);
lib/zstd/compress/zstd_lazy.c
1432
return ZSTD_HC_SEARCH_FN(dictMode, mls)(ms, ip, iend, offsetPtr);
lib/zstd/compress/zstd_lazy.c
1435
return ZSTD_ROW_SEARCH_FN(dictMode, mls, rowLog)(ms, ip, iend, offsetPtr);
lib/zstd/compress/zstd_lazy.c
1490
const BYTE* ip,
lib/zstd/compress/zstd_lazy.c
1525
const BYTE* ip = istart;
lib/zstd/compress/zstd_lazy.c
1549
const U32 dictAndPrefixLength = (U32)((ip - prefixLowest) + (dictEnd - dictLowest));
lib/zstd/compress/zstd_lazy.c
1552
ip += (dictAndPrefixLength == 0);
lib/zstd/compress/zstd_lazy.c
1554
U32 const curr = (U32)(ip - base);
lib/zstd/compress/zstd_lazy.c
1581
while (ip < ilimit) {
lib/zstd/compress/zstd_lazy.c
1584
const BYTE* start=ip+1;
lib/zstd/compress/zstd_lazy.c
1589
const U32 repIndex = (U32)(ip - base) + 1 - offset_1;
lib/zstd/compress/zstd_lazy.c
1595
&& (MEM_read32(repMatch) == MEM_read32(ip+1)) ) {
lib/zstd/compress/zstd_lazy.c
1597
matchLength = ZSTD_count_2segments(ip+1+4, repMatch+4, iend, repMatchEnd, prefixLowest) + 4;
lib/zstd/compress/zstd_lazy.c
1602
&& ((offset_1 > 0) & (MEM_read32(ip+1-offset_1) == MEM_read32(ip+1)))) {
lib/zstd/compress/zstd_lazy.c
1603
matchLength = ZSTD_count(ip+1+4, ip+1+4-offset_1, iend) + 4;
lib/zstd/compress/zstd_lazy.c
1609
size_t const ml2 = ZSTD_searchMax(ms, ip, iend, &offbaseFound, mls, rowLog, searchMethod, dictMode);
lib/zstd/compress/zstd_lazy.c
1611
matchLength = ml2, start = ip, offBase = offbaseFound;
lib/zstd/compress/zstd_lazy.c
1615
size_t const step = ((size_t)(ip-anchor) >> kSearchStrength) + 1; /* jump faster over incompressible sections */;
lib/zstd/compress/zstd_lazy.c
1616
ip += step;
lib/zstd/compress/zstd_lazy.c
1630
while (ip<ilimit) {
lib/zstd/compress/zstd_lazy.c
1632
ip ++;
lib/zstd/compress/zstd_lazy.c
1634
&& (offBase) && ((offset_1>0) & (MEM_read32(ip) == MEM_read32(ip - offset_1)))) {
lib/zstd/compress/zstd_lazy.c
1635
size_t const mlRep = ZSTD_count(ip+4, ip+4-offset_1, iend) + 4;
lib/zstd/compress/zstd_lazy.c
1639
matchLength = mlRep, offBase = REPCODE1_TO_OFFBASE, start = ip;
lib/zstd/compress/zstd_lazy.c
1642
const U32 repIndex = (U32)(ip - base) - offset_1;
lib/zstd/compress/zstd_lazy.c
1647
&& (MEM_read32(repMatch) == MEM_read32(ip)) ) {
lib/zstd/compress/zstd_lazy.c
1649
size_t const mlRep = ZSTD_count_2segments(ip+4, repMatch+4, iend, repMatchEnd, prefixLowest) + 4;
lib/zstd/compress/zstd_lazy.c
1653
matchLength = mlRep, offBase = REPCODE1_TO_OFFBASE, start = ip;
lib/zstd/compress/zstd_lazy.c
1657
size_t const ml2 = ZSTD_searchMax(ms, ip, iend, &ofbCandidate, mls, rowLog, searchMethod, dictMode);
lib/zstd/compress/zstd_lazy.c
1661
matchLength = ml2, offBase = ofbCandidate, start = ip;
lib/zstd/compress/zstd_lazy.c
1666
if ((depth==2) && (ip<ilimit)) {
lib/zstd/compress/zstd_lazy.c
1668
ip ++;
lib/zstd/compress/zstd_lazy.c
167
const BYTE* const ip, const BYTE* const iend,
lib/zstd/compress/zstd_lazy.c
1670
&& (offBase) && ((offset_1>0) & (MEM_read32(ip) == MEM_read32(ip - offset_1)))) {
lib/zstd/compress/zstd_lazy.c
1671
size_t const mlRep = ZSTD_count(ip+4, ip+4-offset_1, iend) + 4;
lib/zstd/compress/zstd_lazy.c
1675
matchLength = mlRep, offBase = REPCODE1_TO_OFFBASE, start = ip;
lib/zstd/compress/zstd_lazy.c
1678
const U32 repIndex = (U32)(ip - base) - offset_1;
lib/zstd/compress/zstd_lazy.c
1683
&& (MEM_read32(repMatch) == MEM_read32(ip)) ) {
lib/zstd/compress/zstd_lazy.c
1685
size_t const mlRep = ZSTD_count_2segments(ip+4, repMatch+4, iend, repMatchEnd, prefixLowest) + 4;
lib/zstd/compress/zstd_lazy.c
1689
matchLength = mlRep, offBase = REPCODE1_TO_OFFBASE, start = ip;
lib/zstd/compress/zstd_lazy.c
1693
size_t const ml2 = ZSTD_searchMax(ms, ip, iend, &ofbCandidate, mls, rowLog, searchMethod, dictMode);
lib/zstd/compress/zstd_lazy.c
1697
matchLength = ml2, offBase = ofbCandidate, start = ip;
lib/zstd/compress/zstd_lazy.c
1726
anchor = ip = start + matchLength;
lib/zstd/compress/zstd_lazy.c
1738
while (ip <= ilimit) {
lib/zstd/compress/zstd_lazy.c
1739
U32 const current2 = (U32)(ip-base);
lib/zstd/compress/zstd_lazy.c
1745
&& (MEM_read32(repMatch) == MEM_read32(ip)) ) {
lib/zstd/compress/zstd_lazy.c
1747
matchLength = ZSTD_count_2segments(ip+4, repMatch+4, iend, repEnd2, prefixLowest) + 4;
lib/zstd/compress/zstd_lazy.c
1750
ip += matchLength;
lib/zstd/compress/zstd_lazy.c
1751
anchor = ip;
lib/zstd/compress/zstd_lazy.c
1759
while ( ((ip <= ilimit) & (offset_2>0))
lib/zstd/compress/zstd_lazy.c
1760
&& (MEM_read32(ip) == MEM_read32(ip - offset_2)) ) {
lib/zstd/compress/zstd_lazy.c
1762
matchLength = ZSTD_count(ip+4, ip+4-offset_2, iend) + 4;
lib/zstd/compress/zstd_lazy.c
1765
ip += matchLength;
lib/zstd/compress/zstd_lazy.c
1766
anchor = ip;
lib/zstd/compress/zstd_lazy.c
178
size_t const h = ZSTD_hashPtr(ip, hashLog, mls);
lib/zstd/compress/zstd_lazy.c
183
U32 const curr = (U32)(ip-base);
lib/zstd/compress/zstd_lazy.c
1945
const BYTE* ip = istart;
lib/zstd/compress/zstd_lazy.c
1967
ip += (ip == prefixStart);
lib/zstd/compress/zstd_lazy.c
1979
while (ip < ilimit) {
lib/zstd/compress/zstd_lazy.c
1982
const BYTE* start=ip+1;
lib/zstd/compress/zstd_lazy.c
1983
U32 curr = (U32)(ip-base);
lib/zstd/compress/zstd_lazy.c
1992
if (MEM_read32(ip+1) == MEM_read32(repMatch)) {
lib/zstd/compress/zstd_lazy.c
1995
matchLength = ZSTD_count_2segments(ip+1+4, repMatch+4, iend, repEnd, prefixStart) + 4;
lib/zstd/compress/zstd_lazy.c
2001
size_t const ml2 = ZSTD_searchMax(ms, ip, iend, &ofbCandidate, mls, rowLog, searchMethod, ZSTD_extDict);
lib/zstd/compress/zstd_lazy.c
2003
matchLength = ml2, start = ip, offBase = ofbCandidate;
lib/zstd/compress/zstd_lazy.c
2007
size_t const step = ((size_t)(ip-anchor) >> kSearchStrength);
lib/zstd/compress/zstd_lazy.c
2008
ip += step + 1; /* jump faster over incompressible sections */
lib/zstd/compress/zstd_lazy.c
2022
while (ip<ilimit) {
lib/zstd/compress/zstd_lazy.c
2023
ip ++;
lib/zstd/compress/zstd_lazy.c
2033
if (MEM_read32(ip) == MEM_read32(repMatch)) {
lib/zstd/compress/zstd_lazy.c
2036
size_t const repLength = ZSTD_count_2segments(ip+4, repMatch+4, iend, repEnd, prefixStart) + 4;
lib/zstd/compress/zstd_lazy.c
204
matchLength += ZSTD_count_2segments(ip+matchLength, match+matchLength, iend, dictEnd, prefixStart);
lib/zstd/compress/zstd_lazy.c
2040
matchLength = repLength, offBase = REPCODE1_TO_OFFBASE, start = ip;
lib/zstd/compress/zstd_lazy.c
2045
size_t const ml2 = ZSTD_searchMax(ms, ip, iend, &ofbCandidate, mls, rowLog, searchMethod, ZSTD_extDict);
lib/zstd/compress/zstd_lazy.c
2049
matchLength = ml2, offBase = ofbCandidate, start = ip;
lib/zstd/compress/zstd_lazy.c
2054
if ((depth==2) && (ip<ilimit)) {
lib/zstd/compress/zstd_lazy.c
2055
ip ++;
lib/zstd/compress/zstd_lazy.c
2065
if (MEM_read32(ip) == MEM_read32(repMatch)) {
lib/zstd/compress/zstd_lazy.c
2068
size_t const repLength = ZSTD_count_2segments(ip+4, repMatch+4, iend, repEnd, prefixStart) + 4;
lib/zstd/compress/zstd_lazy.c
2072
matchLength = repLength, offBase = REPCODE1_TO_OFFBASE, start = ip;
lib/zstd/compress/zstd_lazy.c
2077
size_t const ml2 = ZSTD_searchMax(ms, ip, iend, &ofbCandidate, mls, rowLog, searchMethod, ZSTD_extDict);
lib/zstd/compress/zstd_lazy.c
2081
matchLength = ml2, offBase = ofbCandidate, start = ip;
lib/zstd/compress/zstd_lazy.c
2100
anchor = ip = start + matchLength;
lib/zstd/compress/zstd_lazy.c
2111
while (ip <= ilimit) {
lib/zstd/compress/zstd_lazy.c
2112
const U32 repCurrent = (U32)(ip-base);
lib/zstd/compress/zstd_lazy.c
2119
if (MEM_read32(ip) == MEM_read32(repMatch)) {
lib/zstd/compress/zstd_lazy.c
2122
matchLength = ZSTD_count_2segments(ip+4, repMatch+4, iend, repEnd, prefixStart) + 4;
lib/zstd/compress/zstd_lazy.c
2125
ip += matchLength;
lib/zstd/compress/zstd_lazy.c
2126
anchor = ip;
lib/zstd/compress/zstd_lazy.c
215
if (ip+matchLength == iend) { /* reached end of input : ip[matchLength] is not valid, no way to know if it's larger or smaller than match */
lib/zstd/compress/zstd_lazy.c
220
if (match[matchLength] < ip[matchLength]) {
lib/zstd/compress/zstd_lazy.c
245
const BYTE* const ip, const BYTE* const iend,
lib/zstd/compress/zstd_lazy.c
253
size_t const h = ZSTD_hashPtr(ip, hashLog, mls);
lib/zstd/compress/zstd_lazy.c
257
U32 const curr = (U32)(ip-base);
lib/zstd/compress/zstd_lazy.c
273
assert(ip <= iend-8); /* required for h calculation */
lib/zstd/compress/zstd_lazy.c
31
const BYTE* ip, const BYTE* iend,
lib/zstd/compress/zstd_lazy.c
332
matchLength += ZSTD_count(ip+matchLength, match+matchLength, iend);
lib/zstd/compress/zstd_lazy.c
335
matchLength += ZSTD_count_2segments(ip+matchLength, match+matchLength, iend, dictEnd, prefixStart);
lib/zstd/compress/zstd_lazy.c
345
if (ip+matchLength == iend) { /* equal : no way to know if inf or sup */
lib/zstd/compress/zstd_lazy.c
355
if (match[matchLength] < ip[matchLength]) {
lib/zstd/compress/zstd_lazy.c
376
ms, ip, iend,
lib/zstd/compress/zstd_lazy.c
397
const BYTE* const ip, const BYTE* const iLimit,
lib/zstd/compress/zstd_lazy.c
403
if (ip < ms->window.base + ms->nextToUpdate) return 0; /* skipped area */
lib/zstd/compress/zstd_lazy.c
404
ZSTD_updateDUBT(ms, ip, iLimit, mls);
lib/zstd/compress/zstd_lazy.c
405
return ZSTD_DUBT_findBestMatch(ms, ip, iLimit, offBasePtr, mls, dictMode);
lib/zstd/compress/zstd_lazy.c
412
void ZSTD_dedicatedDictSearch_lazy_loadDictionary(ZSTD_MatchState_t* ms, const BYTE* const ip)
lib/zstd/compress/zstd_lazy.c
415
U32 const target = (U32)(ip - base);
lib/zstd/compress/zstd_lazy.c
43
U32 const target = (U32)(ip - base);
lib/zstd/compress/zstd_lazy.c
49
assert(ip + 8 <= iend); /* condition for ZSTD_hashPtr */
lib/zstd/compress/zstd_lazy.c
532
const BYTE* const ip, const BYTE* const iLimit,
lib/zstd/compress/zstd_lazy.c
570
if (MEM_read32(match) == MEM_read32(ip)) {
lib/zstd/compress/zstd_lazy.c
572
currentMl = ZSTD_count_2segments(ip+4, match+4, iLimit, ddsEnd, prefixStart) + 4;
lib/zstd/compress/zstd_lazy.c
579
if (ip+currentMl == iLimit) {
lib/zstd/compress/zstd_lazy.c
607
if (MEM_read32(match) == MEM_read32(ip)) {
lib/zstd/compress/zstd_lazy.c
609
currentMl = ZSTD_count_2segments(ip+4, match+4, iLimit, ddsEnd, prefixStart) + 4;
lib/zstd/compress/zstd_lazy.c
616
if (ip+currentMl == iLimit) break; /* best possible, avoids read overflow on next attempt */
lib/zstd/compress/zstd_lazy.c
636
const BYTE* ip, U32 const mls, U32 const lazySkipping)
lib/zstd/compress/zstd_lazy.c
643
const U32 target = (U32)(ip - base);
lib/zstd/compress/zstd_lazy.c
657
return hashTable[ZSTD_hashPtr(ip, hashLog, mls)];
lib/zstd/compress/zstd_lazy.c
660
U32 ZSTD_insertAndFindFirstIndex(ZSTD_MatchState_t* ms, const BYTE* ip) {
lib/zstd/compress/zstd_lazy.c
662
return ZSTD_insertAndFindFirstIndex_internal(ms, cParams, ip, ms->cParams.minMatch, /* lazySkipping*/ 0);
lib/zstd/compress/zstd_lazy.c
670
const BYTE* const ip, const BYTE* const iLimit,
lib/zstd/compress/zstd_lazy.c
683
const U32 curr = (U32)(ip-base);
lib/zstd/compress/zstd_lazy.c
697
? ZSTD_hashPtr(ip, ddsHashLog, mls) << ZSTD_LAZY_DDSS_BUCKET_LOG : 0;
lib/zstd/compress/zstd_lazy.c
707
matchIndex = ZSTD_insertAndFindFirstIndex_internal(ms, cParams, ip, mls, ms->lazySkipping);
lib/zstd/compress/zstd_lazy.c
715
if (MEM_read32(match + ml - 3) == MEM_read32(ip + ml - 3)) /* potentially better */
lib/zstd/compress/zstd_lazy.c
716
currentMl = ZSTD_count(ip, match, iLimit);
lib/zstd/compress/zstd_lazy.c
720
if (MEM_read32(match) == MEM_read32(ip)) /* assumption : matchIndex <= dictLimit-4 (by table construction) */
lib/zstd/compress/zstd_lazy.c
721
currentMl = ZSTD_count_2segments(ip+4, match+4, iLimit, dictEnd, prefixStart) + 4;
lib/zstd/compress/zstd_lazy.c
728
if (ip+currentMl == iLimit) break; /* best possible, avoids read overflow on next attempt */
lib/zstd/compress/zstd_lazy.c
738
ip, iLimit, prefixStart, curr, dictLimit, ddsIdx);
lib/zstd/compress/zstd_lazy.c
750
matchIndex = dms->hashTable[ZSTD_hashPtr(ip, dms->cParams.hashLog, mls)];
lib/zstd/compress/zstd_lazy.c
756
if (MEM_read32(match) == MEM_read32(ip)) /* assumption : matchIndex <= dictLimit-4 (by table construction) */
lib/zstd/compress/zstd_lazy.c
757
currentMl = ZSTD_count_2segments(ip+4, match+4, iLimit, dmsEnd, prefixStart) + 4;
lib/zstd/compress/zstd_lazy.c
764
if (ip+currentMl == iLimit) break; /* best possible, avoids read overflow on next attempt */
lib/zstd/compress/zstd_lazy.c
88
const BYTE* const ip = (curr>=dictLimit) ? base + curr : dictBase + curr;
lib/zstd/compress/zstd_lazy.c
917
void ZSTD_row_update_internal(ZSTD_MatchState_t* ms, const BYTE* ip,
lib/zstd/compress/zstd_lazy.c
923
const U32 target = (U32)(ip - base);
lib/zstd/compress/zstd_lazy.c
938
ZSTD_row_fillHashCache(ms, base, rowLog, mls, idx, ip+1);
lib/zstd/compress/zstd_lazy.c
950
void ZSTD_row_update(ZSTD_MatchState_t* const ms, const BYTE* ip) {
lib/zstd/compress/zstd_lazy.c
956
ZSTD_row_update_internal(ms, ip, mls, rowLog, rowMask, 0 /* don't use cache */);
lib/zstd/compress/zstd_lazy.h
31
U32 ZSTD_insertAndFindFirstIndex(ZSTD_MatchState_t* ms, const BYTE* ip);
lib/zstd/compress/zstd_lazy.h
32
void ZSTD_row_update(ZSTD_MatchState_t* const ms, const BYTE* ip);
lib/zstd/compress/zstd_lazy.h
34
void ZSTD_dedicatedDictSearch_lazy_loadDictionary(ZSTD_MatchState_t* ms, const BYTE* const ip);
lib/zstd/compress/zstd_ldm.c
287
ldmState_t* ldmState, const BYTE* ip,
lib/zstd/compress/zstd_ldm.c
294
BYTE const* const istart = ip;
lib/zstd/compress/zstd_ldm.c
302
while (ip < iend) {
lib/zstd/compress/zstd_ldm.c
307
hashed = ZSTD_ldm_gear_feed(&hashState, ip, (size_t)(iend - ip), splits, &numSplits);
lib/zstd/compress/zstd_ldm.c
310
if (ip + splits[n] >= istart + minMatchLength) {
lib/zstd/compress/zstd_ldm.c
311
BYTE const* const split = ip + splits[n] - minMatchLength;
lib/zstd/compress/zstd_ldm.c
322
ip += hashed;
lib/zstd/compress/zstd_ldm.c
366
BYTE const* ip = istart;
lib/zstd/compress/zstd_ldm.c
379
ZSTD_ldm_gear_reset(&hashState, ip, minMatchLength);
lib/zstd/compress/zstd_ldm.c
380
ip += minMatchLength;
lib/zstd/compress/zstd_ldm.c
382
while (ip < ilimit) {
lib/zstd/compress/zstd_ldm.c
387
hashed = ZSTD_ldm_gear_feed(&hashState, ip, ilimit - ip,
lib/zstd/compress/zstd_ldm.c
391
BYTE const* const split = ip + splits[n] - minMatchLength;
lib/zstd/compress/zstd_ldm.c
501
if (anchor > ip + hashed) {
lib/zstd/compress/zstd_ldm.c
504
ip = anchor - hashed;
lib/zstd/compress/zstd_ldm.c
509
ip += hashed;
lib/zstd/compress/zstd_ldm.c
695
BYTE const* ip = istart;
lib/zstd/compress/zstd_ldm.c
710
while (rawSeqStore->pos < rawSeqStore->size && ip < iend) {
lib/zstd/compress/zstd_ldm.c
713
(U32)(iend - ip), minMatch);
lib/zstd/compress/zstd_ldm.c
718
assert(ip + sequence.litLength + sequence.matchLength <= iend);
lib/zstd/compress/zstd_ldm.c
721
ZSTD_ldm_limitTableUpdate(ms, ip);
lib/zstd/compress/zstd_ldm.c
722
ZSTD_ldm_fillFastTables(ms, ip);
lib/zstd/compress/zstd_ldm.c
724
DEBUGLOG(5, "pos %u : calling block compressor on segment of size %u", (unsigned)(ip-istart), sequence.litLength);
lib/zstd/compress/zstd_ldm.c
728
blockCompressor(ms, seqStore, rep, ip, sequence.litLength);
lib/zstd/compress/zstd_ldm.c
729
ip += sequence.litLength;
lib/zstd/compress/zstd_ldm.c
735
ZSTD_storeSeq(seqStore, newLitLength, ip - newLitLength, iend,
lib/zstd/compress/zstd_ldm.c
738
ip += sequence.matchLength;
lib/zstd/compress/zstd_ldm.c
742
ZSTD_ldm_limitTableUpdate(ms, ip);
lib/zstd/compress/zstd_ldm.c
743
ZSTD_ldm_fillFastTables(ms, ip);
lib/zstd/compress/zstd_ldm.c
745
return blockCompressor(ms, seqStore, rep, ip, iend - ip);
lib/zstd/compress/zstd_ldm.h
25
ldmState_t* state, const BYTE* ip,
lib/zstd/compress/zstd_opt.c
1087
const BYTE* ip = istart;
lib/zstd/compress/zstd_opt.c
1110
ZSTD_opt_getNextMatchAndUpdateSeqStore(&optLdm, (U32)(ip-istart), (U32)(iend-ip));
lib/zstd/compress/zstd_opt.c
1114
(U32)(ip - base), ms->window.dictLimit, ms->nextToUpdate);
lib/zstd/compress/zstd_opt.c
1117
ip += (ip==prefixStart);
lib/zstd/compress/zstd_opt.c
1120
while (ip < ilimit) {
lib/zstd/compress/zstd_opt.c
1124
{ U32 const litlen = (U32)(ip - anchor);
lib/zstd/compress/zstd_opt.c
1126
U32 nbMatches = getAllMatches(matches, ms, &nextToUpdate3, ip, iend, rep, ll0, minMatch);
lib/zstd/compress/zstd_opt.c
1128
(U32)(ip-istart), (U32)(iend-ip),
lib/zstd/compress/zstd_opt.c
1131
DEBUGLOG(8, "no match found at cPos %u", (unsigned)(ip-istart));
lib/zstd/compress/zstd_opt.c
1132
ip++;
lib/zstd/compress/zstd_opt.c
1159
nbMatches, maxML, maxOffBase, (U32)(ip-prefixStart));
lib/zstd/compress/zstd_opt.c
1202
const BYTE* const inr = ip + cur;
lib/zstd/compress/zstd_opt.c
1209
+ LIT_PRICE(ip+cur-1)
lib/zstd/compress/zstd_opt.c
1223
&& LIKELY(ip + cur < iend)
lib/zstd/compress/zstd_opt.c
1226
int with1literal = prevMatch.price + LIT_PRICE(ip+cur) + LL_INCPRICE(1);
lib/zstd/compress/zstd_opt.c
1227
int withMoreLiterals = price + LIT_PRICE(ip+cur) + LL_INCPRICE(litlen+1);
lib/zstd/compress/zstd_opt.c
1297
|| (ip + cur + longestML >= iend) ) {
lib/zstd/compress/zstd_opt.c
1352
assert(lastStretch.litlen == (ip - anchor) + last_pos);
lib/zstd/compress/zstd_opt.c
1353
ip += last_pos;
lib/zstd/compress/zstd_opt.c
1424
ip = anchor + llen; /* last "sequence" is a bunch of literals => don't progress anchor */
lib/zstd/compress/zstd_opt.c
1432
ip = anchor;
lib/zstd/compress/zstd_opt.c
414
const BYTE* const ip)
lib/zstd/compress/zstd_opt.c
420
U32 const target = (U32)(ip - base);
lib/zstd/compress/zstd_opt.c
421
size_t const hash3 = ZSTD_hash3Ptr(ip, hashLog3);
lib/zstd/compress/zstd_opt.c
445
const BYTE* const ip, const BYTE* const iend,
lib/zstd/compress/zstd_opt.c
452
size_t const h = ZSTD_hashPtr(ip, hashLog, mls);
lib/zstd/compress/zstd_opt.c
464
const U32 curr = (U32)(ip-base);
lib/zstd/compress/zstd_opt.c
486
assert(ip <= iend-8); /* required for h calculation */
lib/zstd/compress/zstd_opt.c
519
matchLength += ZSTD_count(ip+matchLength, match+matchLength, iend);
lib/zstd/compress/zstd_opt.c
522
matchLength += ZSTD_count_2segments(ip+matchLength, match+matchLength, iend, dictEnd, prefixStart);
lib/zstd/compress/zstd_opt.c
533
if (ip+matchLength == iend) { /* equal : no way to know if inf or sup */
lib/zstd/compress/zstd_opt.c
537
if (match[matchLength] < ip[matchLength]) { /* necessarily within buffer */
lib/zstd/compress/zstd_opt.c
565
const BYTE* const ip, const BYTE* const iend,
lib/zstd/compress/zstd_opt.c
569
U32 const target = (U32)(ip - base);
lib/zstd/compress/zstd_opt.c
579
assert((size_t)(ip - base) <= (size_t)(U32)(-1));
lib/zstd/compress/zstd_opt.c
584
void ZSTD_updateTree(ZSTD_MatchState_t* ms, const BYTE* ip, const BYTE* iend) {
lib/zstd/compress/zstd_opt.c
585
ZSTD_updateTree_internal(ms, ip, iend, ms->cParams.minMatch, ZSTD_noDict);
lib/zstd/compress/zstd_opt.c
595
const BYTE* const ip, const BYTE* const iLimit,
lib/zstd/compress/zstd_opt.c
605
U32 const curr = (U32)(ip-base);
lib/zstd/compress/zstd_opt.c
609
size_t const h = ZSTD_hashPtr(ip, hashLog, mls);
lib/zstd/compress/zstd_opt.c
658
if ((repIndex >= windowLow) & (ZSTD_readMINMATCH(ip, minMatch) == ZSTD_readMINMATCH(ip - repOffset, minMatch))) {
lib/zstd/compress/zstd_opt.c
659
repLen = (U32)ZSTD_count(ip+minMatch, ip+minMatch-repOffset, iLimit) + minMatch;
lib/zstd/compress/zstd_opt.c
669
&& (ZSTD_readMINMATCH(ip, minMatch) == ZSTD_readMINMATCH(repMatch, minMatch)) ) {
lib/zstd/compress/zstd_opt.c
670
repLen = (U32)ZSTD_count_2segments(ip+minMatch, repMatch+minMatch, iLimit, dictEnd, prefixStart) + minMatch;
lib/zstd/compress/zstd_opt.c
675
&& (ZSTD_readMINMATCH(ip, minMatch) == ZSTD_readMINMATCH(repMatch, minMatch)) ) {
lib/zstd/compress/zstd_opt.c
676
repLen = (U32)ZSTD_count_2segments(ip+minMatch, repMatch+minMatch, iLimit, dmsEnd, prefixStart) + minMatch;
lib/zstd/compress/zstd_opt.c
687
| (ip+repLen == iLimit) ) { /* best possible */
lib/zstd/compress/zstd_opt.c
693
U32 const matchIndex3 = ZSTD_insertAndFindFirstIndexHash3(ms, nextToUpdate3, ip);
lib/zstd/compress/zstd_opt.c
699
mlen = ZSTD_count(ip, match, iLimit);
lib/zstd/compress/zstd_opt.c
702
mlen = ZSTD_count_2segments(ip, match, iLimit, dictEnd, prefixStart);
lib/zstd/compress/zstd_opt.c
716
(ip+mlen == iLimit) ) { /* best possible length */
lib/zstd/compress/zstd_opt.c
734
if (matchIndex >= dictLimit) assert(memcmp(match, ip, matchLength) == 0); /* ensure early section of match is equal as expected */
lib/zstd/compress/zstd_opt.c
735
matchLength += ZSTD_count(ip+matchLength, match+matchLength, iLimit);
lib/zstd/compress/zstd_opt.c
738
assert(memcmp(match, ip, matchLength) == 0); /* ensure early section of match is equal as expected */
lib/zstd/compress/zstd_opt.c
739
matchLength += ZSTD_count_2segments(ip+matchLength, match+matchLength, iLimit, dictEnd, prefixStart);
lib/zstd/compress/zstd_opt.c
755
| (ip+matchLength == iLimit) /* equal : no way to know if inf or sup */) {
lib/zstd/compress/zstd_opt.c
760
if (match[matchLength] < ip[matchLength]) {
lib/zstd/compress/zstd_opt.c
779
size_t const dmsH = ZSTD_hashPtr(ip, dmsHashLog, mls);
lib/zstd/compress/zstd_opt.c
787
matchLength += ZSTD_count_2segments(ip+matchLength, match+matchLength, iLimit, dmsEnd, prefixStart);
lib/zstd/compress/zstd_opt.c
802
| (ip+matchLength == iLimit) /* equal : no way to know if inf or sup */) {
lib/zstd/compress/zstd_opt.c
807
if (match[matchLength] < ip[matchLength]) {
lib/zstd/compress/zstd_opt.c
837
const BYTE* ip,
lib/zstd/compress/zstd_opt.c
847
if (ip < ms->window.base + ms->nextToUpdate)
lib/zstd/compress/zstd_opt.c
849
ZSTD_updateTree_internal(ms, ip, iHighLimit, mls, dictMode);
lib/zstd/compress/zstd_opt.c
850
return ZSTD_insertBtAndGetAllMatches(matches, ms, nextToUpdate3, ip, iHighLimit, dictMode, rep, ll0, lengthToBeat, mls);
lib/zstd/compress/zstd_opt.c
860
const BYTE* ip, \
lib/zstd/compress/zstd_opt.c
867
matches, ms, nextToUpdate3, ip, iHighLimit, \
lib/zstd/compress/zstd_opt.h
21
void ZSTD_updateTree(ZSTD_MatchState_t* ms, const BYTE* ip, const BYTE* iend);
lib/zstd/decompress/huf_decompress.c
147
static size_t HUF_initFastDStream(BYTE const* ip) {
lib/zstd/decompress/huf_decompress.c
148
BYTE const lastByte = ip[7];
lib/zstd/decompress/huf_decompress.c
150
size_t const value = MEM_readLEST(ip) | 1;
lib/zstd/decompress/huf_decompress.c
1524
BYTE const* ip[4];
lib/zstd/decompress/huf_decompress.c
1532
ZSTD_memcpy((void*)(&ip), &args->ip, sizeof(ip));
lib/zstd/decompress/huf_decompress.c
1551
assert(ip[stream] >= ilowest);
lib/zstd/decompress/huf_decompress.c
1564
size_t iters = (size_t)(ip[0] - ilowest) / 7;
lib/zstd/decompress/huf_decompress.c
1591
if (ip[stream] < ip[stream - 1])
lib/zstd/decompress/huf_decompress.c
1598
assert(ip[stream] >= ip[stream - 1]);
lib/zstd/decompress/huf_decompress.c
1620
ip[(_stream)] -= nbBytes; \
lib/zstd/decompress/huf_decompress.c
1621
bits[(_stream)] = MEM_read64(ip[(_stream)]) | 1; \
lib/zstd/decompress/huf_decompress.c
1658
ZSTD_memcpy((void*)(&args->ip), &ip, sizeof(ip));
lib/zstd/decompress/huf_decompress.c
1680
assert(args.ip[0] >= args.ilowest);
lib/zstd/decompress/huf_decompress.c
1684
assert(args.ip[0] >= ilowest);
lib/zstd/decompress/huf_decompress.c
1685
assert(args.ip[1] >= ilowest);
lib/zstd/decompress/huf_decompress.c
1686
assert(args.ip[2] >= ilowest);
lib/zstd/decompress/huf_decompress.c
1687
assert(args.ip[3] >= ilowest);
lib/zstd/decompress/huf_decompress.c
171
BYTE const* ip[4];
lib/zstd/decompress/huf_decompress.c
1755
const BYTE* ip = (const BYTE*) cSrc;
lib/zstd/decompress/huf_decompress.c
1761
ip += hSize; cSrcSize -= hSize;
lib/zstd/decompress/huf_decompress.c
1763
return HUF_decompress1X2_usingDTable_internal(dst, dstSize, ip, cSrcSize, DCtx, flags);
lib/zstd/decompress/huf_decompress.c
1770
const BYTE* ip = (const BYTE*) cSrc;
lib/zstd/decompress/huf_decompress.c
1776
ip += hSize; cSrcSize -= hSize;
lib/zstd/decompress/huf_decompress.c
1778
return HUF_decompress4X2_usingDTable_internal(dst, dstSize, ip, cSrcSize, dctx, flags);
lib/zstd/decompress/huf_decompress.c
1893
const BYTE* ip = (const BYTE*) cSrc;
lib/zstd/decompress/huf_decompress.c
1898
ip += hSize; cSrcSize -= hSize;
lib/zstd/decompress/huf_decompress.c
1900
return HUF_decompress1X1_usingDTable_internal(dst, dstSize, ip, cSrcSize, dctx, flags);
lib/zstd/decompress/huf_decompress.c
238
args->ip[0] = args->iend[1] - sizeof(U64);
lib/zstd/decompress/huf_decompress.c
239
args->ip[1] = args->iend[2] - sizeof(U64);
lib/zstd/decompress/huf_decompress.c
240
args->ip[2] = args->iend[3] - sizeof(U64);
lib/zstd/decompress/huf_decompress.c
241
args->ip[3] = (BYTE const*)src + srcSize - sizeof(U64);
lib/zstd/decompress/huf_decompress.c
260
args->bits[0] = HUF_initFastDStream(args->ip[0]);
lib/zstd/decompress/huf_decompress.c
261
args->bits[1] = HUF_initFastDStream(args->ip[1]);
lib/zstd/decompress/huf_decompress.c
262
args->bits[2] = HUF_initFastDStream(args->ip[2]);
lib/zstd/decompress/huf_decompress.c
263
args->bits[3] = HUF_initFastDStream(args->ip[3]);
lib/zstd/decompress/huf_decompress.c
288
if (args->ip[stream] < args->iend[stream] - 8)
lib/zstd/decompress/huf_decompress.c
293
bit->bitContainer = MEM_readLEST(args->ip[stream]);
lib/zstd/decompress/huf_decompress.c
297
bit->ptr = (const char*)args->ip[stream];
lib/zstd/decompress/huf_decompress.c
721
BYTE const* ip[4];
lib/zstd/decompress/huf_decompress.c
729
ZSTD_memcpy((void*)(&ip), &args->ip, sizeof(ip));
lib/zstd/decompress/huf_decompress.c
743
assert(ip[stream] >= ilowest);
lib/zstd/decompress/huf_decompress.c
753
size_t const iiters = (size_t)(ip[0] - ilowest) / 7;
lib/zstd/decompress/huf_decompress.c
773
if (ip[stream] < ip[stream - 1])
lib/zstd/decompress/huf_decompress.c
780
assert(ip[stream] >= ip[stream - 1]);
lib/zstd/decompress/huf_decompress.c
798
ip[(_stream)] -= nbBytes; \
lib/zstd/decompress/huf_decompress.c
799
bits[(_stream)] = MEM_read64(ip[(_stream)]) | 1; \
lib/zstd/decompress/huf_decompress.c
826
ZSTD_memcpy((void*)(&args->ip), &ip, sizeof(ip));
lib/zstd/decompress/huf_decompress.c
853
assert(args.ip[0] >= args.ilowest);
lib/zstd/decompress/huf_decompress.c
859
assert(args.ip[0] >= ilowest);
lib/zstd/decompress/huf_decompress.c
860
assert(args.ip[0] >= ilowest);
lib/zstd/decompress/huf_decompress.c
861
assert(args.ip[1] >= ilowest);
lib/zstd/decompress/huf_decompress.c
862
assert(args.ip[2] >= ilowest);
lib/zstd/decompress/huf_decompress.c
863
assert(args.ip[3] >= ilowest);
lib/zstd/decompress/huf_decompress.c
931
const BYTE* ip = (const BYTE*) cSrc;
lib/zstd/decompress/huf_decompress.c
936
ip += hSize; cSrcSize -= hSize;
lib/zstd/decompress/huf_decompress.c
938
return HUF_decompress4X1_usingDTable_internal(dst, dstSize, ip, cSrcSize, dctx, flags);
lib/zstd/decompress/zstd_decompress.c
1010
checkRead = MEM_readLE32(ip);
lib/zstd/decompress/zstd_decompress.c
1013
ip += 4;
lib/zstd/decompress/zstd_decompress.c
1016
ZSTD_DCtx_trace_end(dctx, (U64)(op-ostart), (U64)(ip-istart), /* streaming */ 0);
lib/zstd/decompress/zstd_decompress.c
1018
DEBUGLOG(4, "ZSTD_decompressFrame: decompressed frame of size %i, consuming %i bytes of input", (int)(op-ostart), (int)(ip - (const BYTE*)*srcPtr));
lib/zstd/decompress/zstd_decompress.c
1019
*srcPtr = ip;
lib/zstd/decompress/zstd_decompress.c
2014
const char* ip = istart;
lib/zstd/decompress/zstd_decompress.c
2048
DEBUGLOG(5, "stage zdss_loadHeader (srcSize : %u)", (U32)(iend - ip));
lib/zstd/decompress/zstd_decompress.c
2058
size_t const remainingInput = (size_t)(iend-ip);
lib/zstd/decompress/zstd_decompress.c
2059
assert(iend >= ip);
lib/zstd/decompress/zstd_decompress.c
2062
ZSTD_memcpy(zds->headerBuffer + zds->lhSize, ip, remainingInput);
lib/zstd/decompress/zstd_decompress.c
2073
assert(ip != NULL);
lib/zstd/decompress/zstd_decompress.c
2074
ZSTD_memcpy(zds->headerBuffer + zds->lhSize, ip, toLoad); zds->lhSize = hSize; ip += toLoad;
lib/zstd/decompress/zstd_decompress.c
2089
ip = istart + cSize;
lib/zstd/decompress/zstd_decompress.c
2168
{ size_t const neededInSize = ZSTD_nextSrcSizeToDecompressWithInputSize(zds, (size_t)(iend - ip));
lib/zstd/decompress/zstd_decompress.c
2175
if ((size_t)(iend-ip) >= neededInSize) { /* decode directly from src */
lib/zstd/decompress/zstd_decompress.c
2176
FORWARD_IF_ERROR(ZSTD_decompressContinueStream(zds, &op, oend, ip, neededInSize), "");
lib/zstd/decompress/zstd_decompress.c
2177
assert(ip != NULL);
lib/zstd/decompress/zstd_decompress.c
2178
ip += neededInSize;
lib/zstd/decompress/zstd_decompress.c
2182
if (ip==iend) { someMoreWork = 0; break; } /* no more input */
lib/zstd/decompress/zstd_decompress.c
2192
assert(neededInSize == ZSTD_nextSrcSizeToDecompressWithInputSize(zds, (size_t)(iend - ip)));
lib/zstd/decompress/zstd_decompress.c
2194
loadedSize = MIN(toLoad, (size_t)(iend-ip));
lib/zstd/decompress/zstd_decompress.c
2199
loadedSize = ZSTD_limitCopy(zds->inBuff + zds->inPos, toLoad, ip, (size_t)(iend-ip));
lib/zstd/decompress/zstd_decompress.c
2203
ip += loadedSize;
lib/zstd/decompress/zstd_decompress.c
2243
input->pos = (size_t)(ip - (const char*)(input->src));
lib/zstd/decompress/zstd_decompress.c
2249
if ((ip==istart) && (op==ostart)) { /* no forward progress */
lib/zstd/decompress/zstd_decompress.c
2253
RETURN_ERROR_IF(ip==iend, noForwardProgress_inputEmpty, "");
lib/zstd/decompress/zstd_decompress.c
433
const BYTE* ip = (const BYTE*)src;
lib/zstd/decompress/zstd_decompress.c
486
{ BYTE const fhdByte = ip[minInputSize-1];
lib/zstd/decompress/zstd_decompress.c
499
BYTE const wlByte = ip[pos++];
lib/zstd/decompress/zstd_decompress.c
511
case 1 : dictID = ip[pos]; pos++; break;
lib/zstd/decompress/zstd_decompress.c
512
case 2 : dictID = MEM_readLE16(ip+pos); pos+=2; break;
lib/zstd/decompress/zstd_decompress.c
513
case 3 : dictID = MEM_readLE32(ip+pos); pos+=4; break;
lib/zstd/decompress/zstd_decompress.c
520
case 0 : if (singleSegment) frameContentSize = ip[pos]; break;
lib/zstd/decompress/zstd_decompress.c
521
case 1 : frameContentSize = MEM_readLE16(ip+pos)+256; break;
lib/zstd/decompress/zstd_decompress.c
522
case 2 : frameContentSize = MEM_readLE32(ip+pos); break;
lib/zstd/decompress/zstd_decompress.c
523
case 3 : frameContentSize = MEM_readLE64(ip+pos); break;
lib/zstd/decompress/zstd_decompress.c
725
const BYTE* ip = (const BYTE*)src;
lib/zstd/decompress/zstd_decompress.c
726
const BYTE* const ipstart = ip;
lib/zstd/decompress/zstd_decompress.c
739
ip += zfh.headerSize;
lib/zstd/decompress/zstd_decompress.c
745
size_t const cBlockSize = ZSTD_getcBlockSize(ip, remainingSize, &blockProperties);
lib/zstd/decompress/zstd_decompress.c
752
ip += ZSTD_blockHeaderSize + cBlockSize;
lib/zstd/decompress/zstd_decompress.c
763
ip += 4;
lib/zstd/decompress/zstd_decompress.c
767
frameSizeInfo.compressedSize = (size_t)(ip - ipstart);
lib/zstd/decompress/zstd_decompress.c
914
const BYTE* ip = istart;
lib/zstd/decompress/zstd_decompress.c
929
ip, ZSTD_FRAMEHEADERSIZE_PREFIX(dctx->format), dctx->format);
lib/zstd/decompress/zstd_decompress.c
933
FORWARD_IF_ERROR( ZSTD_decodeFrameHeader(dctx, ip, frameHeaderSize) , "");
lib/zstd/decompress/zstd_decompress.c
934
ip += frameHeaderSize; remainingSrcSize -= frameHeaderSize;
lib/zstd/decompress/zstd_decompress.c
946
size_t const cBlockSize = ZSTD_getcBlockSize(ip, remainingSrcSize, &blockProperties);
lib/zstd/decompress/zstd_decompress.c
949
ip += ZSTD_blockHeaderSize;
lib/zstd/decompress/zstd_decompress.c
953
if (ip >= op && ip < oBlockEnd) {
lib/zstd/decompress/zstd_decompress.c
967
oBlockEnd = op + (ip - op);
lib/zstd/decompress/zstd_decompress.c
974
decodedSize = ZSTD_decompressBlock_internal(dctx, op, (size_t)(oBlockEnd-op), ip, cBlockSize, not_streaming);
lib/zstd/decompress/zstd_decompress.c
978
decodedSize = ZSTD_copyRawBlock(op, (size_t)(oend-op), ip, cBlockSize);
lib/zstd/decompress/zstd_decompress.c
981
decodedSize = ZSTD_setRleBlock(op, (size_t)(oBlockEnd-op), *ip, blockProperties.origSize);
lib/zstd/decompress/zstd_decompress.c
995
assert(ip != NULL);
lib/zstd/decompress/zstd_decompress.c
996
ip += cBlockSize;
lib/zstd/decompress/zstd_decompress_block.c
1409
const BYTE* ip = (const BYTE*)seqStart;
lib/zstd/decompress/zstd_decompress_block.c
1410
const BYTE* const iend = ip + seqSize;
lib/zstd/decompress/zstd_decompress_block.c
1427
ERR_isError(BIT_initDStream(&seqState.DStream, ip, iend-ip)),
lib/zstd/decompress/zstd_decompress_block.c
1621
const BYTE* ip = (const BYTE*)seqStart;
lib/zstd/decompress/zstd_decompress_block.c
1622
const BYTE* const iend = ip + seqSize;
lib/zstd/decompress/zstd_decompress_block.c
1639
ERR_isError(BIT_initDStream(&seqState.DStream, ip, iend - ip)),
lib/zstd/decompress/zstd_decompress_block.c
1740
const BYTE* ip = (const BYTE*)seqStart;
lib/zstd/decompress/zstd_decompress_block.c
1741
const BYTE* const iend = ip + seqSize;
lib/zstd/decompress/zstd_decompress_block.c
1765
assert(iend >= ip);
lib/zstd/decompress/zstd_decompress_block.c
1767
ERR_isError(BIT_initDStream(&seqState.DStream, ip, iend-ip)),
lib/zstd/decompress/zstd_decompress_block.c
2071
const BYTE* ip = (const BYTE*)src;
lib/zstd/decompress/zstd_decompress_block.c
2088
ip += litCSize;
lib/zstd/decompress/zstd_decompress_block.c
2125
size_t const seqHSize = ZSTD_decodeSeqHeaders(dctx, &nbSeq, ip, srcSize);
lib/zstd/decompress/zstd_decompress_block.c
2127
ip += seqHSize;
lib/zstd/decompress/zstd_decompress_block.c
2163
return ZSTD_decompressSequencesLong(dctx, dst, dstCapacity, ip, srcSize, nbSeq, isLongOffset);
lib/zstd/decompress/zstd_decompress_block.c
2170
return ZSTD_decompressSequencesSplitLitBuffer(dctx, dst, dstCapacity, ip, srcSize, nbSeq, isLongOffset);
lib/zstd/decompress/zstd_decompress_block.c
2172
return ZSTD_decompressSequences(dctx, dst, dstCapacity, ip, srcSize, nbSeq, isLongOffset);
lib/zstd/decompress/zstd_decompress_block.c
701
const BYTE* ip = istart;
lib/zstd/decompress/zstd_decompress_block.c
709
nbSeq = *ip++;
lib/zstd/decompress/zstd_decompress_block.c
712
RETURN_ERROR_IF(ip+2 > iend, srcSize_wrong, "");
lib/zstd/decompress/zstd_decompress_block.c
713
nbSeq = MEM_readLE16(ip) + LONGNBSEQ;
lib/zstd/decompress/zstd_decompress_block.c
714
ip+=2;
lib/zstd/decompress/zstd_decompress_block.c
716
RETURN_ERROR_IF(ip >= iend, srcSize_wrong, "");
lib/zstd/decompress/zstd_decompress_block.c
717
nbSeq = ((nbSeq-0x80)<<8) + *ip++;
lib/zstd/decompress/zstd_decompress_block.c
724
RETURN_ERROR_IF(ip != iend, corruption_detected,
lib/zstd/decompress/zstd_decompress_block.c
726
return (size_t)(ip - istart);
lib/zstd/decompress/zstd_decompress_block.c
730
RETURN_ERROR_IF(ip+1 > iend, srcSize_wrong, ""); /* minimum possible size: 1 byte for symbol encoding types */
lib/zstd/decompress/zstd_decompress_block.c
731
RETURN_ERROR_IF(*ip & 3, corruption_detected, ""); /* The last field, Reserved, must be all-zeroes. */
lib/zstd/decompress/zstd_decompress_block.c
732
{ SymbolEncodingType_e const LLtype = (SymbolEncodingType_e)(*ip >> 6);
lib/zstd/decompress/zstd_decompress_block.c
733
SymbolEncodingType_e const OFtype = (SymbolEncodingType_e)((*ip >> 4) & 3);
lib/zstd/decompress/zstd_decompress_block.c
734
SymbolEncodingType_e const MLtype = (SymbolEncodingType_e)((*ip >> 2) & 3);
lib/zstd/decompress/zstd_decompress_block.c
735
ip++;
lib/zstd/decompress/zstd_decompress_block.c
740
ip, iend-ip,
lib/zstd/decompress/zstd_decompress_block.c
747
ip += llhSize;
lib/zstd/decompress/zstd_decompress_block.c
752
ip, iend-ip,
lib/zstd/decompress/zstd_decompress_block.c
759
ip += ofhSize;
lib/zstd/decompress/zstd_decompress_block.c
764
ip, iend-ip,
lib/zstd/decompress/zstd_decompress_block.c
771
ip += mlhSize;
lib/zstd/decompress/zstd_decompress_block.c
775
return ip-istart;
lib/zstd/decompress/zstd_decompress_block.c
805
HINT_INLINE void ZSTD_overlapCopy8(BYTE** op, BYTE const** ip, size_t offset) {
lib/zstd/decompress/zstd_decompress_block.c
806
assert(*ip <= *op);
lib/zstd/decompress/zstd_decompress_block.c
812
(*op)[0] = (*ip)[0];
lib/zstd/decompress/zstd_decompress_block.c
813
(*op)[1] = (*ip)[1];
lib/zstd/decompress/zstd_decompress_block.c
814
(*op)[2] = (*ip)[2];
lib/zstd/decompress/zstd_decompress_block.c
815
(*op)[3] = (*ip)[3];
lib/zstd/decompress/zstd_decompress_block.c
816
*ip += dec32table[offset];
lib/zstd/decompress/zstd_decompress_block.c
817
ZSTD_copy4(*op+4, *ip);
lib/zstd/decompress/zstd_decompress_block.c
818
*ip -= sub2;
lib/zstd/decompress/zstd_decompress_block.c
820
ZSTD_copy8(*op, *ip);
lib/zstd/decompress/zstd_decompress_block.c
822
*ip += 8;
lib/zstd/decompress/zstd_decompress_block.c
824
assert(*op - *ip >= 8);
lib/zstd/decompress/zstd_decompress_block.c
838
static void ZSTD_safecopy(BYTE* op, const BYTE* const oend_w, BYTE const* ip, ptrdiff_t length, ZSTD_overlap_e ovtype) {
lib/zstd/decompress/zstd_decompress_block.c
839
ptrdiff_t const diff = op - ip;
lib/zstd/decompress/zstd_decompress_block.c
847
while (op < oend) *op++ = *ip++;
lib/zstd/decompress/zstd_decompress_block.c
853
ZSTD_overlapCopy8(&op, &ip, diff);
lib/zstd/decompress/zstd_decompress_block.c
855
assert(op - ip >= 8);
lib/zstd/decompress/zstd_decompress_block.c
861
ZSTD_wildcopy(op, ip, length, ovtype);
lib/zstd/decompress/zstd_decompress_block.c
867
ZSTD_wildcopy(op, ip, oend_w - op, ovtype);
lib/zstd/decompress/zstd_decompress_block.c
868
ip += oend_w - op;
lib/zstd/decompress/zstd_decompress_block.c
872
while (op < oend) *op++ = *ip++;
lib/zstd/decompress/zstd_decompress_block.c
878
static void ZSTD_safecopyDstBeforeSrc(BYTE* op, const BYTE* ip, ptrdiff_t length) {
lib/zstd/decompress/zstd_decompress_block.c
879
ptrdiff_t const diff = op - ip;
lib/zstd/decompress/zstd_decompress_block.c
884
while (op < oend) *op++ = *ip++;
lib/zstd/decompress/zstd_decompress_block.c
889
ZSTD_wildcopy(op, ip, oend - WILDCOPY_OVERLENGTH - op, ZSTD_no_overlap);
lib/zstd/decompress/zstd_decompress_block.c
890
ip += oend - WILDCOPY_OVERLENGTH - op;
lib/zstd/decompress/zstd_decompress_block.c
895
while (op < oend) *op++ = *ip++;
mm/kasan/common.c
223
unsigned long ip)
mm/kasan/common.c
230
kasan_report_invalid_free(tagged_object, ip, KASAN_REPORT_INVALID_FREE);
mm/kasan/common.c
235
kasan_report_invalid_free(tagged_object, ip, KASAN_REPORT_DOUBLE_FREE);
mm/kasan/common.c
257
unsigned long ip)
mm/kasan/common.c
261
return check_slab_allocation(cache, object, ip);
mm/kasan/common.c
307
static inline bool check_page_allocation(void *ptr, unsigned long ip)
mm/kasan/common.c
310
kasan_report_invalid_free(ptr, ip, KASAN_REPORT_INVALID_FREE);
mm/kasan/common.c
315
kasan_report_invalid_free(ptr, ip, KASAN_REPORT_DOUBLE_FREE);
mm/kasan/common.c
322
void __kasan_kfree_large(void *ptr, unsigned long ip)
mm/kasan/common.c
324
check_page_allocation(ptr, ip);
mm/kasan/common.c
491
unsigned long ip)
mm/kasan/common.c
505
if (check_page_allocation(ptr, ip))
mm/kasan/common.c
514
unsigned long ip)
mm/kasan/common.c
519
bool __kasan_mempool_poison_object(void *ptr, unsigned long ip)
mm/kasan/common.c
525
if (check_page_allocation(ptr, ip))
mm/kasan/common.c
536
if (check_slab_allocation(slab->slab_cache, ptr, ip))
mm/kasan/common.c
543
void __kasan_mempool_unpoison_object(void *ptr, size_t size, unsigned long ip)
mm/kasan/common.c
571
bool __kasan_check_byte(const void *address, unsigned long ip)
mm/kasan/common.c
574
kasan_report(address, 1, false, ip);
mm/kasan/kasan.h
214
unsigned long ip;
mm/kasan/kasan.h
382
bool is_write, unsigned long ip);
mm/kasan/kasan.h
383
void kasan_report_invalid_free(void *object, unsigned long ip, enum kasan_report_type type);
mm/kasan/report.c
247
pr_err("BUG: KASAN: %s in %pS\n", info->bug_type, (void *)info->ip);
mm/kasan/report.c
529
void kasan_report_invalid_free(void *ptr, unsigned long ip, enum kasan_report_type type)
mm/kasan/report.c
553
info.ip = ip;
mm/kasan/report.c
572
unsigned long ip)
mm/kasan/report.c
591
info.ip = ip;
mm/memory.c
7200
void print_vma_addr(char *prefix, unsigned long ip)
mm/memory.c
7211
vma = vma_lookup(mm, ip);
mm/memory.c
7214
ip -= vma->vm_start;
mm/memory.c
7215
ip += vma->vm_pgoff << PAGE_SHIFT;
mm/memory.c
7216
printk("%s%pD[%lx,%lx+%lx]", prefix, f, ip,
mm/mmap_lock.c
445
unsigned long ip = exception_ip(regs);
mm/mmap_lock.c
446
if (!search_exception_tables(ip))
mm/mmap_lock.c
470
unsigned long ip = exception_ip(regs);
mm/mmap_lock.c
471
if (!search_exception_tables(ip))
mm/page_alloc.c
4346
void __fs_reclaim_acquire(unsigned long ip)
mm/page_alloc.c
4348
lock_acquire_exclusive(&__fs_reclaim_map, 0, 0, NULL, ip);
mm/page_alloc.c
4351
void __fs_reclaim_release(unsigned long ip)
mm/page_alloc.c
4353
lock_release(&__fs_reclaim_map, ip);
net/atm/clip.c
279
__be32 *ip = (__be32 *) neigh->primary_key;
net/atm/clip.c
282
to_atmarpd(act_need, PRIV(neigh->dev)->number, *ip);
net/atm/clip.c
457
static int clip_setentry(struct atm_vcc *vcc, __be32 ip)
net/atm/clip.c
470
if (!ip) {
net/atm/clip.c
479
rt = ip_route_output(&init_net, ip, 0, 0, 0, RT_SCOPE_LINK);
net/atm/clip.c
482
neigh = __neigh_lookup(&arp_tbl, &ip, rt->dst.dev, 1);
net/atm/clip.c
53
static int to_atmarpd(enum atmarp_ctrl_type type, int itf, __be32 ip)
net/atm/clip.c
77
ctrl->ip = ip;
net/atm/mpoa_proc.c
247
unsigned char ip[4];
net/atm/mpoa_proc.c
255
ip, ip+1, ip+2, ip+3) == 4) {
net/atm/mpoa_proc.c
256
ipaddr = *(__be32 *)ip;
net/atm/mpoa_proc.c
261
ip, ip+1, ip+2, ip+3, &tx_pcr, &tx_sdu) == 6) {
net/atm/mpoa_proc.c
265
ip, ip+1, ip+2, ip+3, &tx_pcr, &tx_sdu, &rx_pcr, &rx_sdu) != 8)
net/atm/mpoa_proc.c
268
ipaddr = *(__be32 *)ip;
net/batman-adv/distributed-arp-table.c
1369
dat_entry->mac_addr, &dat_entry->ip);
net/batman-adv/distributed-arp-table.c
293
key = (__force const unsigned char *)&dat->ip;
net/batman-adv/distributed-arp-table.c
294
for (i = 0; i < sizeof(dat->ip); i++) {
net/batman-adv/distributed-arp-table.c
325
batadv_dat_entry_hash_find(struct batadv_priv *bat_priv, __be32 ip,
net/batman-adv/distributed-arp-table.c
336
to_find.ip = ip;
net/batman-adv/distributed-arp-table.c
344
if (dat_entry->ip != ip)
net/batman-adv/distributed-arp-table.c
365
static void batadv_dat_entry_add(struct batadv_priv *bat_priv, __be32 ip,
net/batman-adv/distributed-arp-table.c
371
dat_entry = batadv_dat_entry_hash_find(bat_priv, ip, vid);
net/batman-adv/distributed-arp-table.c
379
&dat_entry->ip, dat_entry->mac_addr,
net/batman-adv/distributed-arp-table.c
388
dat_entry->ip = ip;
net/batman-adv/distributed-arp-table.c
406
&dat_entry->ip, dat_entry->mac_addr, batadv_print_vid(vid));
net/batman-adv/distributed-arp-table.c
642
dat.ip = ip_dst;
net/batman-adv/distributed-arp-table.c
673
struct sk_buff *skb, __be32 ip,
net/batman-adv/distributed-arp-table.c
683
cand = batadv_dat_select_candidates(bat_priv, ip, vid);
net/batman-adv/distributed-arp-table.c
687
batadv_dbg(BATADV_DBG_DAT, bat_priv, "DHT_SEND for %pI4\n", &ip);
net/batman-adv/distributed-arp-table.c
871
dat_entry->ip) ||
net/batman-adv/types.h
2082
__be32 ip;
net/bridge/br_arp_nd_proxy.c
100
addr = inet_confirm_addr(dev_net(dev), in_dev, 0, ip,
net/bridge/br_arp_nd_proxy.c
103
if (addr == ip)
net/bridge/br_arp_nd_proxy.c
109
static bool br_is_local_ip(struct net_device *dev, __be32 ip)
net/bridge/br_arp_nd_proxy.c
112
.data = (void *)&ip,
net/bridge/br_arp_nd_proxy.c
94
__be32 ip = *(__be32 *)priv->data;
net/bridge/br_mdb.c
1365
struct br_ip ip = cfg->group;
net/bridge/br_mdb.c
1369
mp = br_mdb_ip_get(br, &ip);
net/bridge/br_mdb.c
151
static void __mdb_entry_to_br_ip(struct br_mdb_entry *entry, struct br_ip *ip,
net/bridge/br_mdb.c
154
memset(ip, 0, sizeof(struct br_ip));
net/bridge/br_mdb.c
155
ip->vid = entry->vid;
net/bridge/br_mdb.c
156
ip->proto = entry->addr.proto;
net/bridge/br_mdb.c
157
switch (ip->proto) {
net/bridge/br_mdb.c
159
ip->dst.ip4 = entry->addr.u.ip4;
net/bridge/br_mdb.c
161
ip->src.ip4 = nla_get_in_addr(mdb_attrs[MDBE_ATTR_SOURCE]);
net/bridge/br_mdb.c
165
ip->dst.ip6 = entry->addr.u.ip6;
net/bridge/br_mdb.c
167
ip->src.ip6 = nla_get_in6_addr(mdb_attrs[MDBE_ATTR_SOURCE]);
net/bridge/br_mdb.c
171
ether_addr_copy(ip->dst.mac_addr, entry->addr.u.mac_addr);
net/bridge/br_multicast.c
1339
br_multicast_find_group_src(struct net_bridge_port_group *pg, struct br_ip *ip)
net/bridge/br_multicast.c
1343
switch (ip->proto) {
net/bridge/br_multicast.c
1346
if (ip->src.ip4 == ent->addr.src.ip4)
net/bridge/br_multicast.c
1352
if (!ipv6_addr_cmp(&ent->addr.src.ip6, &ip->src.ip6))
net/bridge/br_multicast.c
153
struct br_ip ip;
net/bridge/br_multicast.c
162
memset(&ip, 0, sizeof(ip));
net/bridge/br_multicast.c
163
ip.proto = skb->protocol;
net/bridge/br_multicast.c
164
ip.vid = vid;
net/bridge/br_multicast.c
168
ip.dst.ip4 = ip_hdr(skb)->daddr;
net/bridge/br_multicast.c
172
ip.src.ip4 = ip_hdr(skb)->saddr;
net/bridge/br_multicast.c
173
mdb = br_mdb_ip_get_rcu(br, &ip);
net/bridge/br_multicast.c
176
ip.src.ip4 = 0;
net/bridge/br_multicast.c
1768
struct br_ip *ip,
net/bridge/br_multicast.c
1771
if (ip->proto == htons(ETH_P_IP))
net/bridge/br_multicast.c
181
ip.dst.ip6 = ipv6_hdr(skb)->daddr;
net/bridge/br_multicast.c
185
ip.src.ip6 = ipv6_hdr(skb)->saddr;
net/bridge/br_multicast.c
186
mdb = br_mdb_ip_get_rcu(br, &ip);
net/bridge/br_multicast.c
189
memset(&ip.src.ip6, 0, sizeof(ip.src.ip6));
net/bridge/br_multicast.c
194
ip.proto = 0;
net/bridge/br_multicast.c
195
ether_addr_copy(ip.dst.mac_addr, eth_hdr(skb)->h_dest);
net/bridge/br_multicast.c
198
return br_mdb_ip_get_rcu(br, &ip);
net/bridge/br_private.h
1050
br_multicast_find_group_src(struct net_bridge_port_group *pg, struct br_ip *ip);
net/bridge/br_private.h
1196
static inline bool br_multicast_is_star_g(const struct br_ip *ip)
net/bridge/br_private.h
1198
switch (ip->proto) {
net/bridge/br_private.h
1200
return ipv4_is_zeronet(ip->src.ip4);
net/bridge/br_private.h
1203
return ipv6_addr_any(&ip->src.ip6);
net/bridge/br_switchdev.c
499
struct br_ip ip;
net/bridge/br_switchdev.c
516
mp = br_mdb_ip_get(br, &data->ip);
net/bridge/br_switchdev.c
668
complete_info->ip = mp->addr;
net/bridge/netfilter/ebt_among.c
20
const char *mac, __be32 ip)
net/bridge/netfilter/ebt_among.c
35
if (ip) {
net/bridge/netfilter/ebt_among.c
39
if (p->ip == 0 || p->ip == ip)
net/bridge/netfilter/ebt_among.c
46
if (p->ip == 0)
net/core/flow_dissector.c
463
struct flow_dissector_key_ip *ip;
net/core/flow_dissector.c
465
ip = skb_flow_dissector_target(flow_dissector,
net/core/flow_dissector.c
468
ip->tos = key->tos;
net/core/flow_dissector.c
469
ip->ttl = key->ttl;
net/core/netpoll.c
389
udph->check = csum_tcpudp_magic(np->local_ip.ip,
net/core/netpoll.c
390
np->remote_ip.ip,
net/core/netpoll.c
458
put_unaligned(np->local_ip.ip, &iph->saddr);
net/core/netpoll.c
459
put_unaligned(np->remote_ip.ip, &iph->daddr);
net/core/netpoll.c
701
np->local_ip.ip = ifa->ifa_local;
net/core/netpoll.c
702
np_info(np, "local IP %pI4\n", &np->local_ip.ip);
net/core/netpoll.c
750
if (!np->local_ip.ip) {
net/ethtool/ioctl.c
3609
struct flow_dissector_key_ip ip;
net/ethtool/ioctl.c
3712
match->key.ip.tos = v4_spec->tos;
net/ethtool/ioctl.c
3713
match->mask.ip.tos = v4_m_spec->tos;
net/ethtool/ioctl.c
3717
offsetof(struct ethtool_rx_flow_key, ip);
net/ethtool/ioctl.c
3764
match->key.ip.tos = v6_spec->tclass;
net/ethtool/ioctl.c
3765
match->mask.ip.tos = v6_m_spec->tclass;
net/ethtool/ioctl.c
3769
offsetof(struct ethtool_rx_flow_key, ip);
net/ipv4/arp.c
1040
__be32 ip;
net/ipv4/arp.c
1048
ip = ((struct sockaddr_in *)&r->arp_pa)->sin_addr.s_addr;
net/ipv4/arp.c
1050
rt = ip_route_output(net, ip, 0, 0, 0, RT_SCOPE_LINK);
net/ipv4/arp.c
1092
__be32 ip = ((struct sockaddr_in *)&r->arp_pa)->sin_addr.s_addr;
net/ipv4/arp.c
1094
return pneigh_create(&arp_tbl, net, &ip, dev, 0, 0, false);
net/ipv4/arp.c
1104
__be32 ip;
net/ipv4/arp.c
1135
ip = ((struct sockaddr_in *)&r->arp_pa)->sin_addr.s_addr;
net/ipv4/arp.c
1137
neigh = __neigh_lookup_errno(&arp_tbl, &ip, dev);
net/ipv4/arp.c
1172
__be32 ip = ((struct sockaddr_in *) &r->arp_pa)->sin_addr.s_addr;
net/ipv4/arp.c
1183
neigh = neigh_lookup(&arp_tbl, &ip, dev);
net/ipv4/arp.c
1206
int arp_invalidate(struct net_device *dev, __be32 ip, bool force)
net/ipv4/arp.c
1208
struct neighbour *neigh = neigh_lookup(&arp_tbl, &ip, dev);
net/ipv4/arp.c
1237
__be32 ip = ((struct sockaddr_in *)&r->arp_pa)->sin_addr.s_addr;
net/ipv4/arp.c
1239
return pneigh_delete(&arp_tbl, net, &ip, dev);
net/ipv4/arp.c
1248
__be32 ip;
net/ipv4/arp.c
1257
ip = ((struct sockaddr_in *)&r->arp_pa)->sin_addr.s_addr;
net/ipv4/arp.c
1259
return arp_invalidate(dev, ip, true);
net/ipv4/icmp.c
1044
ip_hdr(skb_in)->saddr = ct->tuplehash[dir].tuple.src.u3.ip;
net/ipv4/igmp.c
1765
struct ip_mc_list __rcu **ip;
net/ipv4/igmp.c
1769
for (ip = &in_dev->mc_list;
net/ipv4/igmp.c
1770
(i = rtnl_dereference(*ip)) != NULL;
net/ipv4/igmp.c
1771
ip = &i->next_rcu) {
net/ipv4/igmp.c
1775
*ip = i->next_rcu;
net/ipv4/ipconfig.c
1718
char *cp, *ip, *dp;
net/ipv4/ipconfig.c
1744
ip = addrs;
net/ipv4/ipconfig.c
1745
while (ip && *ip) {
net/ipv4/ipconfig.c
1746
if ((cp = strchr(ip, ':')))
net/ipv4/ipconfig.c
1748
if (strlen(ip) > 0) {
net/ipv4/ipconfig.c
1749
pr_debug("IP-Config: Parameter #%d: `%s'\n", num, ip);
net/ipv4/ipconfig.c
1752
if ((ic_myaddr = in_aton(ip)) == ANY)
net/ipv4/ipconfig.c
1756
if ((ic_servaddr = in_aton(ip)) == ANY)
net/ipv4/ipconfig.c
1760
if ((ic_gateway = in_aton(ip)) == ANY)
net/ipv4/ipconfig.c
1764
if ((ic_netmask = in_aton(ip)) == ANY)
net/ipv4/ipconfig.c
1768
if ((dp = strchr(ip, '.'))) {
net/ipv4/ipconfig.c
1773
strscpy(utsname()->nodename, ip,
net/ipv4/ipconfig.c
1778
strscpy(user_dev_name, ip, sizeof(user_dev_name));
net/ipv4/ipconfig.c
1781
if (ic_proto_name(ip) == 0 &&
net/ipv4/ipconfig.c
1788
ic_nameservers[0] = in_aton(ip);
net/ipv4/ipconfig.c
1795
ic_nameservers[1] = in_aton(ip);
net/ipv4/ipconfig.c
1802
ic_ntp_servers[0] = in_aton(ip);
net/ipv4/ipconfig.c
1809
ip = cp;
net/ipv4/netfilter/ip_tables.c
113
memcmp(&e->ip, &uncond, sizeof(uncond)) == 0;
net/ipv4/netfilter/ip_tables.c
1258
const struct ipt_ip *ip,
net/ipv4/netfilter/ip_tables.c
1308
if (!ip_checkentry(&e->ip))
net/ipv4/netfilter/ip_tables.c
1320
ret = compat_find_calc_match(ematch, &e->ip, &off);
net/ipv4/netfilter/ip_tables.c
230
const struct iphdr *ip;
net/ipv4/netfilter/ip_tables.c
243
ip = ip_hdr(skb);
net/ipv4/netfilter/ip_tables.c
252
acpar.fragoff = ntohs(ip->frag_off) & IP_OFFSET;
net/ipv4/netfilter/ip_tables.c
283
if (!ip_packet_match(ip, indev, outdev,
net/ipv4/netfilter/ip_tables.c
284
&e->ip, acpar.fragoff)) {
net/ipv4/netfilter/ip_tables.c
330
!(e->ip.flags & IPT_F_GOTO)) {
net/ipv4/netfilter/ip_tables.c
348
ip = ip_hdr(skb);
net/ipv4/netfilter/ip_tables.c
43
ip_packet_match(const struct iphdr *ip,
net/ipv4/netfilter/ip_tables.c
466
const struct ipt_ip *ip = par->entryinfo;
net/ipv4/netfilter/ip_tables.c
472
ip->proto, ip->invflags & IPT_INV_PROTO);
net/ipv4/netfilter/ip_tables.c
511
e->ip.proto, e->ip.invflags & IPT_INV_PROTO);
net/ipv4/netfilter/ip_tables.c
52
(ip->saddr & ipinfo->smsk.s_addr) != ipinfo->src.s_addr) ||
net/ipv4/netfilter/ip_tables.c
533
mtpar.entryinfo = &e->ip;
net/ipv4/netfilter/ip_tables.c
54
(ip->daddr & ipinfo->dmsk.s_addr) != ipinfo->dst.s_addr))
net/ipv4/netfilter/ip_tables.c
607
if (!ip_checkentry(&e->ip))
net/ipv4/netfilter/ip_tables.c
69
NF_INVF(ipinfo, IPT_INV_PROTO, ip->protocol != ipinfo->proto))
net/ipv4/netfilter/ip_tables.c
82
ip_checkentry(const struct ipt_ip *ip)
net/ipv4/netfilter/ip_tables.c
84
if (ip->flags & ~IPT_F_MASK)
net/ipv4/netfilter/ip_tables.c
86
if (ip->invflags & ~IPT_INV_MASK)
net/ipv4/netfilter/ipt_ECN.c
105
(e->ip.proto != IPPROTO_TCP || (e->ip.invflags & XT_INV_PROTO))) {
net/ipv4/netfilter/ipt_REJECT.c
79
if (e->ip.proto != IPPROTO_TCP ||
net/ipv4/netfilter/ipt_REJECT.c
80
(e->ip.invflags & XT_INV_PROTO)) {
net/ipv4/netfilter/ipt_SYNPROXY.c
70
if (e->ip.proto != IPPROTO_TCP ||
net/ipv4/netfilter/ipt_SYNPROXY.c
71
e->ip.invflags & XT_INV_PROTO)
net/ipv4/netfilter/nf_nat_h323.c
105
if (addr.ip == ct->tuplehash[dir].tuple.src.u3.ip &&
net/ipv4/netfilter/nf_nat_h323.c
113
(ntohl(addr.ip) & 0xff000000) == 0x7f000000)
net/ipv4/netfilter/nf_nat_h323.c
117
&addr.ip, port,
net/ipv4/netfilter/nf_nat_h323.c
118
&ct->tuplehash[!dir].tuple.dst.u3.ip,
net/ipv4/netfilter/nf_nat_h323.c
125
} else if (addr.ip == ct->tuplehash[dir].tuple.dst.u3.ip &&
net/ipv4/netfilter/nf_nat_h323.c
129
&addr.ip, port,
net/ipv4/netfilter/nf_nat_h323.c
130
&ct->tuplehash[!dir].tuple.src.u3.ip,
net/ipv4/netfilter/nf_nat_h323.c
157
addr.ip == ct->tuplehash[dir].tuple.src.u3.ip &&
net/ipv4/netfilter/nf_nat_h323.c
160
&addr.ip, ntohs(port),
net/ipv4/netfilter/nf_nat_h323.c
161
&ct->tuplehash[!dir].tuple.dst.u3.ip,
net/ipv4/netfilter/nf_nat_h323.c
25
unsigned int addroff, __be32 ip, __be16 port)
net/ipv4/netfilter/nf_nat_h323.c
266
&rtp_exp->tuple.src.u3.ip,
net/ipv4/netfilter/nf_nat_h323.c
268
&rtp_exp->tuple.dst.u3.ip,
net/ipv4/netfilter/nf_nat_h323.c
271
&rtcp_exp->tuple.src.u3.ip,
net/ipv4/netfilter/nf_nat_h323.c
273
&rtcp_exp->tuple.dst.u3.ip,
net/ipv4/netfilter/nf_nat_h323.c
30
__be32 ip;
net/ipv4/netfilter/nf_nat_h323.c
309
&exp->tuple.src.u3.ip,
net/ipv4/netfilter/nf_nat_h323.c
311
&exp->tuple.dst.u3.ip,
net/ipv4/netfilter/nf_nat_h323.c
356
&exp->tuple.src.u3.ip,
net/ipv4/netfilter/nf_nat_h323.c
358
&exp->tuple.dst.u3.ip,
net/ipv4/netfilter/nf_nat_h323.c
36
buf.ip = ip;
net/ipv4/netfilter/nf_nat_h323.c
373
if (this->tuple.src.u3.ip != 0) { /* Only accept calls from GK */
net/ipv4/netfilter/nf_nat_h323.c
437
(ntohl(addr.ip) & 0xff000000) == 0x7f000000) {
net/ipv4/netfilter/nf_nat_h323.c
448
&exp->tuple.src.u3.ip,
net/ipv4/netfilter/nf_nat_h323.c
450
&exp->tuple.dst.u3.ip,
net/ipv4/netfilter/nf_nat_h323.c
491
exp->tuple.dst.u3.ip = ct->tuplehash[!dir].tuple.dst.u3.ip;
net/ipv4/netfilter/nf_nat_h323.c
512
&exp->tuple.src.u3.ip,
net/ipv4/netfilter/nf_nat_h323.c
514
&exp->tuple.dst.u3.ip,
net/ipv4/netfilter/nf_nat_h323.c
76
return set_addr(skb, protoff, data, dataoff, taddr->ipAddress.ip,
net/ipv4/netfilter/nf_nat_h323.c
77
addr->ip, port);
net/ipv4/netfilter/nf_nat_h323.c
88
addr->ip, port);
net/ipv4/netfilter/nf_nat_pptp.c
67
t.src.u3.ip = master->tuplehash[!exp->dir].tuple.src.u3.ip;
net/ipv4/netfilter/nf_nat_pptp.c
69
t.dst.u3.ip = master->tuplehash[!exp->dir].tuple.dst.u3.ip;
net/ipv4/netfilter/nf_nat_pptp.c
76
t.src.u3.ip = master->tuplehash[!exp->dir].tuple.src.u3.ip;
net/ipv4/netfilter/nf_nat_pptp.c
78
t.dst.u3.ip = master->tuplehash[!exp->dir].tuple.dst.u3.ip;
net/ipv4/netfilter/nf_nat_snmp_basic_main.c
136
ctx.from = ct->tuplehash[dir].tuple.src.u3.ip;
net/ipv4/netfilter/nf_nat_snmp_basic_main.c
137
ctx.to = ct->tuplehash[!dir].tuple.dst.u3.ip;
net/ipv4/netfilter/nf_nat_snmp_basic_main.c
139
ctx.from = ct->tuplehash[!dir].tuple.src.u3.ip;
net/ipv4/netfilter/nf_nat_snmp_basic_main.c
140
ctx.to = ct->tuplehash[dir].tuple.dst.u3.ip;
net/ipv4/netfilter/nf_socket_ipv4.c
138
daddr = ct->tuplehash[IP_CT_DIR_ORIGINAL].tuple.src.u3.ip;
net/ipv4/tcp_ipv4.c
1576
struct tcp4_pseudohdr ip;
net/ipv4/tcp_ipv4.c
1580
h.ip.saddr = saddr;
net/ipv4/tcp_ipv4.c
1581
h.ip.daddr = daddr;
net/ipv4/tcp_ipv4.c
1582
h.ip.pad = 0;
net/ipv4/tcp_ipv4.c
1583
h.ip.protocol = IPPROTO_TCP;
net/ipv4/tcp_ipv4.c
1584
h.ip.len = cpu_to_be16(nbytes);
net/ipv4/tcp_ipv4.c
1587
md5_update(ctx, (const u8 *)&h, sizeof(h.ip) + sizeof(h.tcp));
net/ipv6/ila/ila_xlat.c
101
xp->ip.locator.v64 = (__force __be64)nla_get_u64(
net/ipv6/ila/ila_xlat.c
105
xp->ip.locator_match.v64 = (__force __be64)nla_get_u64(
net/ipv6/ila/ila_xlat.c
108
xp->ip.csum_mode = nla_get_u8_default(info->attrs[ILA_ATTR_CSUM_MODE],
net/ipv6/ila/ila_xlat.c
111
xp->ip.ident_type = nla_get_u8_default(info->attrs[ILA_ATTR_IDENT_TYPE],
net/ipv6/ila/ila_xlat.c
13
struct ila_params ip;
net/ipv6/ila/ila_xlat.c
145
&xp->ip.locator_match,
net/ipv6/ila/ila_xlat.c
204
spinlock_t *lock = ila_get_lock(ilan, xp->ip.locator_match);
net/ipv6/ila/ila_xlat.c
227
ila_init_saved_csum(&xp->ip);
net/ipv6/ila/ila_xlat.c
236
&xp->ip.locator_match,
net/ipv6/ila/ila_xlat.c
287
spinlock_t *lock = ila_get_lock(ilan, xp->ip.locator_match);
net/ipv6/ila/ila_xlat.c
293
&xp->ip.locator_match, rht_params);
net/ipv6/ila/ila_xlat.c
376
return ila_get_lock(ilan, ila->xp.ip.locator_match);
net/ipv6/ila/ila_xlat.c
427
(__force u64)ila->xp.ip.locator.v64,
net/ipv6/ila/ila_xlat.c
430
(__force u64)ila->xp.ip.locator_match.v64,
net/ipv6/ila/ila_xlat.c
433
nla_put_u8(msg, ILA_ATTR_CSUM_MODE, ila->xp.ip.csum_mode) ||
net/ipv6/ila/ila_xlat.c
434
nla_put_u8(msg, ILA_ATTR_IDENT_TYPE, ila->xp.ip.ident_type))
net/ipv6/ila/ila_xlat.c
660
ila_update_ipv6_locator(skb, &ila->xp.ip, sir2ila);
net/ipv6/ila/ila_xlat.c
71
return (ila->xp.ip.locator_match.v64 != *(__be64 *)arg->key);
net/ipv6/ila/ila_xlat.c
87
.key_offset = offsetof(struct ila_map, xp.ip.locator_match),
net/ipv6/tcp_ipv6.c
700
struct tcp6_pseudohdr ip; /* TCP pseudo-header (RFC2460) */
net/ipv6/tcp_ipv6.c
704
h.ip.saddr = *saddr;
net/ipv6/tcp_ipv6.c
705
h.ip.daddr = *daddr;
net/ipv6/tcp_ipv6.c
706
h.ip.protocol = cpu_to_be32(IPPROTO_TCP);
net/ipv6/tcp_ipv6.c
707
h.ip.len = cpu_to_be32(nbytes);
net/ipv6/tcp_ipv6.c
710
md5_update(ctx, (const u8 *)&h, sizeof(h.ip) + sizeof(h.tcp));
net/l2tp/l2tp_ppp.c
1513
u32 ip = 0;
net/l2tp/l2tp_ppp.c
1519
ip = ntohl(inet->inet_saddr);
net/l2tp/l2tp_ppp.c
1534
session->name, ip, port,
net/netfilter/ipset/ip_set_bitmap_ip.c
118
u32 ip;
net/netfilter/ipset/ip_set_bitmap_ip.c
120
ip = ntohl(ip4addr(skb, opt->flags & IPSET_DIM_ONE_SRC));
net/netfilter/ipset/ip_set_bitmap_ip.c
121
if (ip < map->first_ip || ip > map->last_ip)
net/netfilter/ipset/ip_set_bitmap_ip.c
124
e.id = ip_to_id(map, ip);
net/netfilter/ipset/ip_set_bitmap_ip.c
135
u32 ip = 0, ip_to = 0;
net/netfilter/ipset/ip_set_bitmap_ip.c
146
ret = ip_set_get_hostipaddr4(tb[IPSET_ATTR_IP], &ip);
net/netfilter/ipset/ip_set_bitmap_ip.c
154
if (ip < map->first_ip || ip > map->last_ip)
net/netfilter/ipset/ip_set_bitmap_ip.c
158
e.id = ip_to_id(map, ip);
net/netfilter/ipset/ip_set_bitmap_ip.c
166
if (ip > ip_to)
net/netfilter/ipset/ip_set_bitmap_ip.c
167
swap(ip, ip_to);
net/netfilter/ipset/ip_set_bitmap_ip.c
173
ip_set_mask_from_to(ip, ip_to, cidr);
net/netfilter/ipset/ip_set_bitmap_ip.c
175
ip_to = ip;
net/netfilter/ipset/ip_set_bitmap_ip.c
178
if (ip < map->first_ip || ip_to > map->last_ip)
net/netfilter/ipset/ip_set_bitmap_ip.c
181
for (; !before(ip_to, ip); ip += map->hosts) {
net/netfilter/ipset/ip_set_bitmap_ip.c
182
e.id = ip_to_id(map, ip);
net/netfilter/ipset/ip_set_bitmap_ip.c
59
ip_to_id(const struct bitmap_ip *m, u32 ip)
net/netfilter/ipset/ip_set_bitmap_ip.c
61
return ((ip & ip_set_hostmask(m->netmask)) - m->first_ip) / m->hosts;
net/netfilter/ipset/ip_set_bitmap_ipmac.c
216
u32 ip;
net/netfilter/ipset/ip_set_bitmap_ipmac.c
218
ip = ntohl(ip4addr(skb, opt->flags & IPSET_DIM_ONE_SRC));
net/netfilter/ipset/ip_set_bitmap_ipmac.c
219
if (ip < map->first_ip || ip > map->last_ip)
net/netfilter/ipset/ip_set_bitmap_ipmac.c
227
e.id = ip_to_id(map, ip);
net/netfilter/ipset/ip_set_bitmap_ipmac.c
248
u32 ip = 0;
net/netfilter/ipset/ip_set_bitmap_ipmac.c
257
ret = ip_set_get_hostipaddr4(tb[IPSET_ATTR_IP], &ip);
net/netfilter/ipset/ip_set_bitmap_ipmac.c
265
if (ip < map->first_ip || ip > map->last_ip)
net/netfilter/ipset/ip_set_bitmap_ipmac.c
268
e.id = ip_to_id(map, ip);
net/netfilter/ipset/ip_set_bitmap_ipmac.c
69
ip_to_id(const struct bitmap_ipmac *m, u32 ip)
net/netfilter/ipset/ip_set_bitmap_ipmac.c
71
return ip - m->first_ip;
net/netfilter/ipset/ip_set_hash_gen.h
1292
if (nla_put_ipaddr4(skb, IPSET_ATTR_BITMASK, h->bitmask.ip))
net/netfilter/ipset/ip_set_hash_gen.h
1516
bitmask.ip = ip_set_netmask(netmask);
net/netfilter/ipset/ip_set_hash_gen.h
1529
ret = ip_set_get_ipaddr4(tb[IPSET_ATTR_BITMASK], &bitmask.ip);
net/netfilter/ipset/ip_set_hash_gen.h
1530
if (ret || !bitmask.ip)
net/netfilter/ipset/ip_set_hash_ip.c
107
u32 ip = 0, ip_to = 0, hosts, i = 0;
net/netfilter/ipset/ip_set_hash_ip.c
116
ret = ip_set_get_hostipaddr4(tb[IPSET_ATTR_IP], &ip);
net/netfilter/ipset/ip_set_hash_ip.c
124
ip &= ntohl(h->bitmask.ip);
net/netfilter/ipset/ip_set_hash_ip.c
125
e.ip = htonl(ip);
net/netfilter/ipset/ip_set_hash_ip.c
126
if (e.ip == 0)
net/netfilter/ipset/ip_set_hash_ip.c
132
ip_to = ip;
net/netfilter/ipset/ip_set_hash_ip.c
137
if (ip > ip_to) {
net/netfilter/ipset/ip_set_hash_ip.c
140
swap(ip, ip_to);
net/netfilter/ipset/ip_set_hash_ip.c
147
ip_set_mask_from_to(ip, ip_to, cidr);
net/netfilter/ipset/ip_set_hash_ip.c
153
ip = ntohl(h->next.ip);
net/netfilter/ipset/ip_set_hash_ip.c
154
for (; ip <= ip_to; i++) {
net/netfilter/ipset/ip_set_hash_ip.c
155
e.ip = htonl(ip);
net/netfilter/ipset/ip_set_hash_ip.c
164
ip += hosts;
net/netfilter/ipset/ip_set_hash_ip.c
165
if (ip == 0)
net/netfilter/ipset/ip_set_hash_ip.c
177
union nf_inet_addr ip;
net/netfilter/ipset/ip_set_hash_ip.c
187
return ipv6_addr_equal(&ip1->ip.in6, &ip2->ip.in6);
net/netfilter/ipset/ip_set_hash_ip.c
193
if (nla_put_ipaddr6(skb, IPSET_ATTR_IP, &e->ip.in6))
net/netfilter/ipset/ip_set_hash_ip.c
225
ip6addrptr(skb, opt->flags & IPSET_DIM_ONE_SRC, &e.ip.in6);
net/netfilter/ipset/ip_set_hash_ip.c
226
nf_inet_addr_mask_inplace(&e.ip, &h->bitmask);
net/netfilter/ipset/ip_set_hash_ip.c
227
if (ipv6_addr_any(&e.ip.in6))
net/netfilter/ipset/ip_set_hash_ip.c
257
ret = ip_set_get_ipaddr6(tb[IPSET_ATTR_IP], &e.ip);
net/netfilter/ipset/ip_set_hash_ip.c
265
nf_inet_addr_mask_inplace(&e.ip, &h->bitmask);
net/netfilter/ipset/ip_set_hash_ip.c
266
if (ipv6_addr_any(&e.ip.in6))
net/netfilter/ipset/ip_set_hash_ip.c
45
__be32 ip;
net/netfilter/ipset/ip_set_hash_ip.c
55
return e1->ip == e2->ip;
net/netfilter/ipset/ip_set_hash_ip.c
61
if (nla_put_ipaddr4(skb, IPSET_ATTR_IP, e->ip))
net/netfilter/ipset/ip_set_hash_ip.c
72
next->ip = e->ip;
net/netfilter/ipset/ip_set_hash_ip.c
88
__be32 ip;
net/netfilter/ipset/ip_set_hash_ip.c
90
ip4addrptr(skb, opt->flags & IPSET_DIM_ONE_SRC, &ip);
net/netfilter/ipset/ip_set_hash_ip.c
91
ip &= h->bitmask.ip;
net/netfilter/ipset/ip_set_hash_ip.c
92
if (ip == 0)
net/netfilter/ipset/ip_set_hash_ip.c
95
e.ip = ip;
net/netfilter/ipset/ip_set_hash_ipmac.c
104
ip4addrptr(skb, opt->flags & IPSET_DIM_ONE_SRC, &e.ip);
net/netfilter/ipset/ip_set_hash_ipmac.c
114
struct hash_ipmac4_elem e = { .ip = 0, { .foo[0] = 0, .foo[1] = 0 } };
net/netfilter/ipset/ip_set_hash_ipmac.c
132
ret = ip_set_get_ipaddr4(tb[IPSET_ATTR_IP], &e.ip) ||
net/netfilter/ipset/ip_set_hash_ipmac.c
148
union nf_inet_addr ip;
net/netfilter/ipset/ip_set_hash_ipmac.c
162
return ipv6_addr_equal(&e1->ip.in6, &e2->ip.in6) &&
net/netfilter/ipset/ip_set_hash_ipmac.c
169
if (nla_put_ipaddr6(skb, IPSET_ATTR_IP, &e->ip.in6) ||
net/netfilter/ipset/ip_set_hash_ipmac.c
220
ip6addrptr(skb, opt->flags & IPSET_DIM_ONE_SRC, &e.ip.in6);
net/netfilter/ipset/ip_set_hash_ipmac.c
251
ret = ip_set_get_ipaddr6(tb[IPSET_ATTR_IP], &e.ip) ||
net/netfilter/ipset/ip_set_hash_ipmac.c
41
__be32 ip;
net/netfilter/ipset/ip_set_hash_ipmac.c
55
return e1->ip == e2->ip && ether_addr_equal(e1->ether, e2->ether);
net/netfilter/ipset/ip_set_hash_ipmac.c
61
if (nla_put_ipaddr4(skb, IPSET_ATTR_IP, e->ip) ||
net/netfilter/ipset/ip_set_hash_ipmac.c
74
next->ip = e->ip;
net/netfilter/ipset/ip_set_hash_ipmac.c
89
struct hash_ipmac4_elem e = { .ip = 0, { .foo[0] = 0, .foo[1] = 0 } };
net/netfilter/ipset/ip_set_hash_ipmark.c
104
u32 ip, ip_to = 0, i = 0;
net/netfilter/ipset/ip_set_hash_ipmark.c
114
ret = ip_set_get_ipaddr4(tb[IPSET_ATTR_IP], &e.ip);
net/netfilter/ipset/ip_set_hash_ipmark.c
124
if (e.mark == 0 && e.ip == 0)
net/netfilter/ipset/ip_set_hash_ipmark.c
133
ip_to = ip = ntohl(e.ip);
net/netfilter/ipset/ip_set_hash_ipmark.c
138
if (ip > ip_to) {
net/netfilter/ipset/ip_set_hash_ipmark.c
141
swap(ip, ip_to);
net/netfilter/ipset/ip_set_hash_ipmark.c
148
ip_set_mask_from_to(ip, ip_to, cidr);
net/netfilter/ipset/ip_set_hash_ipmark.c
152
ip = ntohl(h->next.ip);
net/netfilter/ipset/ip_set_hash_ipmark.c
153
for (; ip <= ip_to; ip++, i++) {
net/netfilter/ipset/ip_set_hash_ipmark.c
154
e.ip = htonl(ip);
net/netfilter/ipset/ip_set_hash_ipmark.c
172
union nf_inet_addr ip;
net/netfilter/ipset/ip_set_hash_ipmark.c
183
return ipv6_addr_equal(&ip1->ip.in6, &ip2->ip.in6) &&
net/netfilter/ipset/ip_set_hash_ipmark.c
191
if (nla_put_ipaddr6(skb, IPSET_ATTR_IP, &data->ip.in6) ||
net/netfilter/ipset/ip_set_hash_ipmark.c
227
ip6addrptr(skb, opt->flags & IPSET_DIM_ONE_SRC, &e.ip.in6);
net/netfilter/ipset/ip_set_hash_ipmark.c
256
ret = ip_set_get_ipaddr6(tb[IPSET_ATTR_IP], &e.ip);
net/netfilter/ipset/ip_set_hash_ipmark.c
40
__be32 ip;
net/netfilter/ipset/ip_set_hash_ipmark.c
51
return ip1->ip == ip2->ip &&
net/netfilter/ipset/ip_set_hash_ipmark.c
59
if (nla_put_ipaddr4(skb, IPSET_ATTR_IP, data->ip) ||
net/netfilter/ipset/ip_set_hash_ipmark.c
72
next->ip = d->ip;
net/netfilter/ipset/ip_set_hash_ipmark.c
92
ip4addrptr(skb, opt->flags & IPSET_DIM_ONE_SRC, &e.ip);
net/netfilter/ipset/ip_set_hash_ipport.c
104
ip4addrptr(skb, opt->flags & IPSET_DIM_ONE_SRC, &e.ip);
net/netfilter/ipset/ip_set_hash_ipport.c
105
e.ip &= h->bitmask.ip;
net/netfilter/ipset/ip_set_hash_ipport.c
106
if (e.ip == 0)
net/netfilter/ipset/ip_set_hash_ipport.c
117
struct hash_ipport4_elem e = { .ip = 0 };
net/netfilter/ipset/ip_set_hash_ipport.c
119
u32 ip, ip_to = 0, p = 0, port, port_to, i = 0;
net/netfilter/ipset/ip_set_hash_ipport.c
131
ret = ip_set_get_ipaddr4(tb[IPSET_ATTR_IP], &e.ip);
net/netfilter/ipset/ip_set_hash_ipport.c
139
e.ip &= h->bitmask.ip;
net/netfilter/ipset/ip_set_hash_ipport.c
140
if (e.ip == 0)
net/netfilter/ipset/ip_set_hash_ipport.c
165
ip_to = ip = ntohl(e.ip);
net/netfilter/ipset/ip_set_hash_ipport.c
170
if (ip > ip_to)
net/netfilter/ipset/ip_set_hash_ipport.c
171
swap(ip, ip_to);
net/netfilter/ipset/ip_set_hash_ipport.c
177
ip_set_mask_from_to(ip, ip_to, cidr);
net/netfilter/ipset/ip_set_hash_ipport.c
188
ip = ntohl(h->next.ip);
net/netfilter/ipset/ip_set_hash_ipport.c
189
for (; ip <= ip_to; ip++) {
net/netfilter/ipset/ip_set_hash_ipport.c
190
p = retried && ip == ntohl(h->next.ip) ? ntohs(h->next.port)
net/netfilter/ipset/ip_set_hash_ipport.c
193
e.ip = htonl(ip);
net/netfilter/ipset/ip_set_hash_ipport.c
213
union nf_inet_addr ip;
net/netfilter/ipset/ip_set_hash_ipport.c
226
return ipv6_addr_equal(&ip1->ip.in6, &ip2->ip.in6) &&
net/netfilter/ipset/ip_set_hash_ipport.c
235
if (nla_put_ipaddr6(skb, IPSET_ATTR_IP, &data->ip.in6) ||
net/netfilter/ipset/ip_set_hash_ipport.c
266
struct hash_ipport6_elem e = { .ip = { .all = { 0 } } };
net/netfilter/ipset/ip_set_hash_ipport.c
274
ip6addrptr(skb, opt->flags & IPSET_DIM_ONE_SRC, &e.ip.in6);
net/netfilter/ipset/ip_set_hash_ipport.c
275
nf_inet_addr_mask_inplace(&e.ip, &h->bitmask);
net/netfilter/ipset/ip_set_hash_ipport.c
276
if (ipv6_addr_any(&e.ip.in6))
net/netfilter/ipset/ip_set_hash_ipport.c
288
struct hash_ipport6_elem e = { .ip = { .all = { 0 } } };
net/netfilter/ipset/ip_set_hash_ipport.c
310
ret = ip_set_get_ipaddr6(tb[IPSET_ATTR_IP], &e.ip);
net/netfilter/ipset/ip_set_hash_ipport.c
318
nf_inet_addr_mask_inplace(&e.ip, &h->bitmask);
net/netfilter/ipset/ip_set_hash_ipport.c
319
if (ipv6_addr_any(&e.ip.in6))
net/netfilter/ipset/ip_set_hash_ipport.c
46
__be32 ip;
net/netfilter/ipset/ip_set_hash_ipport.c
59
return ip1->ip == ip2->ip &&
net/netfilter/ipset/ip_set_hash_ipport.c
68
if (nla_put_ipaddr4(skb, IPSET_ATTR_IP, data->ip) ||
net/netfilter/ipset/ip_set_hash_ipport.c
82
next->ip = d->ip;
net/netfilter/ipset/ip_set_hash_ipport.c
96
struct hash_ipport4_elem e = { .ip = 0 };
net/netfilter/ipset/ip_set_hash_ipportip.c
102
ip4addrptr(skb, opt->flags & IPSET_DIM_ONE_SRC, &e.ip);
net/netfilter/ipset/ip_set_hash_ipportip.c
113
struct hash_ipportip4_elem e = { .ip = 0 };
net/netfilter/ipset/ip_set_hash_ipportip.c
115
u32 ip, ip_to = 0, p = 0, port, port_to, i = 0;
net/netfilter/ipset/ip_set_hash_ipportip.c
127
ret = ip_set_get_ipaddr4(tb[IPSET_ATTR_IP], &e.ip);
net/netfilter/ipset/ip_set_hash_ipportip.c
161
ip_to = ip = ntohl(e.ip);
net/netfilter/ipset/ip_set_hash_ipportip.c
166
if (ip > ip_to)
net/netfilter/ipset/ip_set_hash_ipportip.c
167
swap(ip, ip_to);
net/netfilter/ipset/ip_set_hash_ipportip.c
173
ip_set_mask_from_to(ip, ip_to, cidr);
net/netfilter/ipset/ip_set_hash_ipportip.c
184
ip = ntohl(h->next.ip);
net/netfilter/ipset/ip_set_hash_ipportip.c
185
for (; ip <= ip_to; ip++) {
net/netfilter/ipset/ip_set_hash_ipportip.c
186
p = retried && ip == ntohl(h->next.ip) ? ntohs(h->next.port)
net/netfilter/ipset/ip_set_hash_ipportip.c
189
e.ip = htonl(ip);
net/netfilter/ipset/ip_set_hash_ipportip.c
209
union nf_inet_addr ip;
net/netfilter/ipset/ip_set_hash_ipportip.c
223
return ipv6_addr_equal(&ip1->ip.in6, &ip2->ip.in6) &&
net/netfilter/ipset/ip_set_hash_ipportip.c
233
if (nla_put_ipaddr6(skb, IPSET_ATTR_IP, &data->ip.in6) ||
net/netfilter/ipset/ip_set_hash_ipportip.c
265
struct hash_ipportip6_elem e = { .ip = { .all = { 0 } } };
net/netfilter/ipset/ip_set_hash_ipportip.c
272
ip6addrptr(skb, opt->flags & IPSET_DIM_ONE_SRC, &e.ip.in6);
net/netfilter/ipset/ip_set_hash_ipportip.c
283
struct hash_ipportip6_elem e = { .ip = { .all = { 0 } } };
net/netfilter/ipset/ip_set_hash_ipportip.c
305
ret = ip_set_get_ipaddr6(tb[IPSET_ATTR_IP], &e.ip);
net/netfilter/ipset/ip_set_hash_ipportip.c
43
__be32 ip;
net/netfilter/ipset/ip_set_hash_ipportip.c
55
return ip1->ip == ip2->ip &&
net/netfilter/ipset/ip_set_hash_ipportip.c
65
if (nla_put_ipaddr4(skb, IPSET_ATTR_IP, data->ip) ||
net/netfilter/ipset/ip_set_hash_ipportip.c
80
next->ip = d->ip;
net/netfilter/ipset/ip_set_hash_ipportip.c
95
struct hash_ipportip4_elem e = { .ip = 0 };
net/netfilter/ipset/ip_set_hash_ipportnet.c
106
if (nla_put_ipaddr4(skb, IPSET_ATTR_IP, data->ip) ||
net/netfilter/ipset/ip_set_hash_ipportnet.c
124
next->ip = d->ip;
net/netfilter/ipset/ip_set_hash_ipportnet.c
152
ip4addrptr(skb, opt->flags & IPSET_DIM_ONE_SRC, &e.ip);
net/netfilter/ipset/ip_set_hash_ipportnet.c
167
u32 ip = 0, ip_to = 0, p = 0, port, port_to;
net/netfilter/ipset/ip_set_hash_ipportnet.c
182
ret = ip_set_get_hostipaddr4(tb[IPSET_ATTR_IP], &ip);
net/netfilter/ipset/ip_set_hash_ipportnet.c
227
e.ip = htonl(ip);
net/netfilter/ipset/ip_set_hash_ipportnet.c
234
ip_to = ip;
net/netfilter/ipset/ip_set_hash_ipportnet.c
239
if (ip > ip_to)
net/netfilter/ipset/ip_set_hash_ipportnet.c
240
swap(ip, ip_to);
net/netfilter/ipset/ip_set_hash_ipportnet.c
246
ip_set_mask_from_to(ip, ip_to, cidr);
net/netfilter/ipset/ip_set_hash_ipportnet.c
270
ip = ntohl(h->next.ip);
net/netfilter/ipset/ip_set_hash_ipportnet.c
277
for (; ip <= ip_to; ip++) {
net/netfilter/ipset/ip_set_hash_ipportnet.c
278
e.ip = htonl(ip);
net/netfilter/ipset/ip_set_hash_ipportnet.c
308
union nf_inet_addr ip;
net/netfilter/ipset/ip_set_hash_ipportnet.c
323
return ipv6_addr_equal(&ip1->ip.in6, &ip2->ip.in6) &&
net/netfilter/ipset/ip_set_hash_ipportnet.c
361
if (nla_put_ipaddr6(skb, IPSET_ATTR_IP, &data->ip.in6) ||
net/netfilter/ipset/ip_set_hash_ipportnet.c
409
ip6addrptr(skb, opt->flags & IPSET_DIM_ONE_SRC, &e.ip.in6);
net/netfilter/ipset/ip_set_hash_ipportnet.c
446
ret = ip_set_get_ipaddr6(tb[IPSET_ATTR_IP], &e.ip);
net/netfilter/ipset/ip_set_hash_ipportnet.c
53
__be32 ip;
net/netfilter/ipset/ip_set_hash_ipportnet.c
68
return ip1->ip == ip2->ip &&
net/netfilter/ipset/ip_set_hash_net.c
105
next->ip = d->ip;
net/netfilter/ipset/ip_set_hash_net.c
129
ip4addrptr(skb, opt->flags & IPSET_DIM_ONE_SRC, &e.ip);
net/netfilter/ipset/ip_set_hash_net.c
130
e.ip &= ip_set_netmask(e.cidr);
net/netfilter/ipset/ip_set_hash_net.c
143
u32 ip = 0, ip_to = 0, i = 0;
net/netfilter/ipset/ip_set_hash_net.c
153
ret = ip_set_get_hostipaddr4(tb[IPSET_ATTR_IP], &ip);
net/netfilter/ipset/ip_set_hash_net.c
175
e.ip = htonl(ip & ip_set_hostmask(e.cidr));
net/netfilter/ipset/ip_set_hash_net.c
181
ip_to = ip;
net/netfilter/ipset/ip_set_hash_net.c
186
if (ip_to < ip)
net/netfilter/ipset/ip_set_hash_net.c
187
swap(ip, ip_to);
net/netfilter/ipset/ip_set_hash_net.c
188
if (ip + UINT_MAX == ip_to)
net/netfilter/ipset/ip_set_hash_net.c
193
ip = ntohl(h->next.ip);
net/netfilter/ipset/ip_set_hash_net.c
196
e.ip = htonl(ip);
net/netfilter/ipset/ip_set_hash_net.c
201
ip = ip_set_range_to_cidr(ip, ip_to, &e.cidr);
net/netfilter/ipset/ip_set_hash_net.c
207
} while (ip++ < ip_to);
net/netfilter/ipset/ip_set_hash_net.c
214
union nf_inet_addr ip;
net/netfilter/ipset/ip_set_hash_net.c
227
return ipv6_addr_equal(&ip1->ip.in6, &ip2->ip.in6) &&
net/netfilter/ipset/ip_set_hash_net.c
252
ip6_netmask(&elem->ip, cidr);
net/netfilter/ipset/ip_set_hash_net.c
261
if (nla_put_ipaddr6(skb, IPSET_ATTR_IP, &data->ip.in6) ||
net/netfilter/ipset/ip_set_hash_net.c
303
ip6addrptr(skb, opt->flags & IPSET_DIM_ONE_SRC, &e.ip.in6);
net/netfilter/ipset/ip_set_hash_net.c
304
ip6_netmask(&e.ip, e.cidr);
net/netfilter/ipset/ip_set_hash_net.c
327
ret = ip_set_get_ipaddr6(tb[IPSET_ATTR_IP], &e.ip);
net/netfilter/ipset/ip_set_hash_net.c
341
ip6_netmask(&e.ip, e.cidr);
net/netfilter/ipset/ip_set_hash_net.c
43
__be32 ip;
net/netfilter/ipset/ip_set_hash_net.c
56
return ip1->ip == ip2->ip &&
net/netfilter/ipset/ip_set_hash_net.c
81
elem->ip &= ip_set_netmask(cidr);
net/netfilter/ipset/ip_set_hash_net.c
90
if (nla_put_ipaddr4(skb, IPSET_ATTR_IP, data->ip) ||
net/netfilter/ipset/ip_set_hash_netiface.c
103
elem->ip &= ip_set_netmask(cidr);
net/netfilter/ipset/ip_set_hash_netiface.c
116
if (nla_put_ipaddr4(skb, IPSET_ATTR_IP, data->ip) ||
net/netfilter/ipset/ip_set_hash_netiface.c
132
next->ip = d->ip;
net/netfilter/ipset/ip_set_hash_netiface.c
172
ip4addrptr(skb, opt->flags & IPSET_DIM_ONE_SRC, &e.ip);
net/netfilter/ipset/ip_set_hash_netiface.c
173
e.ip &= ip_set_netmask(e.cidr);
net/netfilter/ipset/ip_set_hash_netiface.c
205
u32 ip = 0, ip_to = 0, i = 0;
net/netfilter/ipset/ip_set_hash_netiface.c
216
ret = ip_set_get_hostipaddr4(tb[IPSET_ATTR_IP], &ip);
net/netfilter/ipset/ip_set_hash_netiface.c
242
e.ip = htonl(ip & ip_set_hostmask(e.cidr));
net/netfilter/ipset/ip_set_hash_netiface.c
252
if (ip_to < ip)
net/netfilter/ipset/ip_set_hash_netiface.c
253
swap(ip, ip_to);
net/netfilter/ipset/ip_set_hash_netiface.c
254
if (ip + UINT_MAX == ip_to)
net/netfilter/ipset/ip_set_hash_netiface.c
257
ip_set_mask_from_to(ip, ip_to, e.cidr);
net/netfilter/ipset/ip_set_hash_netiface.c
261
ip = ntohl(h->next.ip);
net/netfilter/ipset/ip_set_hash_netiface.c
264
e.ip = htonl(ip);
net/netfilter/ipset/ip_set_hash_netiface.c
269
ip = ip_set_range_to_cidr(ip, ip_to, &e.cidr);
net/netfilter/ipset/ip_set_hash_netiface.c
276
} while (ip++ < ip_to);
net/netfilter/ipset/ip_set_hash_netiface.c
283
union nf_inet_addr ip;
net/netfilter/ipset/ip_set_hash_netiface.c
291
union nf_inet_addr ip;
net/netfilter/ipset/ip_set_hash_netiface.c
307
return ipv6_addr_equal(&ip1->ip.in6, &ip2->ip.in6) &&
net/netfilter/ipset/ip_set_hash_netiface.c
337
ip6_netmask(&elem->ip, cidr);
net/netfilter/ipset/ip_set_hash_netiface.c
350
if (nla_put_ipaddr6(skb, IPSET_ATTR_IP, &data->ip.in6) ||
net/netfilter/ipset/ip_set_hash_netiface.c
393
ip6addrptr(skb, opt->flags & IPSET_DIM_ONE_SRC, &e.ip.in6);
net/netfilter/ipset/ip_set_hash_netiface.c
394
ip6_netmask(&e.ip, e.cidr);
net/netfilter/ipset/ip_set_hash_netiface.c
435
ret = ip_set_get_ipaddr6(tb[IPSET_ATTR_IP], &e.ip);
net/netfilter/ipset/ip_set_hash_netiface.c
449
ip6_netmask(&e.ip, e.cidr);
net/netfilter/ipset/ip_set_hash_netiface.c
48
__be32 ip;
net/netfilter/ipset/ip_set_hash_netiface.c
57
__be32 ip;
net/netfilter/ipset/ip_set_hash_netiface.c
73
return ip1->ip == ip2->ip &&
net/netfilter/ipset/ip_set_hash_netnet.c
100
elem->ip[0] &= ip_set_netmask(cidr);
net/netfilter/ipset/ip_set_hash_netnet.c
111
if (nla_put_ipaddr4(skb, IPSET_ATTR_IP, data->ip[0]) ||
net/netfilter/ipset/ip_set_hash_netnet.c
112
nla_put_ipaddr4(skb, IPSET_ATTR_IP2, data->ip[1]) ||
net/netfilter/ipset/ip_set_hash_netnet.c
157
ip4addrptr(skb, opt->flags & IPSET_DIM_ONE_SRC, &e.ip[0]);
net/netfilter/ipset/ip_set_hash_netnet.c
158
ip4addrptr(skb, opt->flags & IPSET_DIM_TWO_SRC, &e.ip[1]);
net/netfilter/ipset/ip_set_hash_netnet.c
159
e.ip[0] &= (ip_set_netmask(e.cidr[0]) & h->bitmask.ip);
net/netfilter/ipset/ip_set_hash_netnet.c
160
e.ip[1] &= (ip_set_netmask(e.cidr[1]) & h->bitmask.ip);
net/netfilter/ipset/ip_set_hash_netnet.c
173
u32 ip = 0, ip_to = 0;
net/netfilter/ipset/ip_set_hash_netnet.c
185
ret = ip_set_get_hostipaddr4(tb[IPSET_ATTR_IP], &ip);
net/netfilter/ipset/ip_set_hash_netnet.c
218
e.ip[0] = htonl(ip & ntohl(h->bitmask.ip) & ip_set_hostmask(e.cidr[0]));
net/netfilter/ipset/ip_set_hash_netnet.c
219
e.ip[1] = htonl(ip2_from & ntohl(h->bitmask.ip) & ip_set_hostmask(e.cidr[1]));
net/netfilter/ipset/ip_set_hash_netnet.c
225
ip_to = ip;
net/netfilter/ipset/ip_set_hash_netnet.c
230
if (ip_to < ip)
net/netfilter/ipset/ip_set_hash_netnet.c
231
swap(ip, ip_to);
net/netfilter/ipset/ip_set_hash_netnet.c
232
if (unlikely(ip + UINT_MAX == ip_to))
net/netfilter/ipset/ip_set_hash_netnet.c
235
ip_set_mask_from_to(ip, ip_to, e.cidr[0]);
net/netfilter/ipset/ip_set_hash_netnet.c
252
ip = ntohl(h->next.ip[0]);
net/netfilter/ipset/ip_set_hash_netnet.c
253
ip2 = ntohl(h->next.ip[1]);
net/netfilter/ipset/ip_set_hash_netnet.c
259
e.ip[0] = htonl(ip);
net/netfilter/ipset/ip_set_hash_netnet.c
260
ip = ip_set_range_to_cidr(ip, ip_to, &e.cidr[0]);
net/netfilter/ipset/ip_set_hash_netnet.c
263
e.ip[1] = htonl(ip2);
net/netfilter/ipset/ip_set_hash_netnet.c
276
} while (ip++ < ip_to);
net/netfilter/ipset/ip_set_hash_netnet.c
283
union nf_inet_addr ip[2];
net/netfilter/ipset/ip_set_hash_netnet.c
299
return ipv6_addr_equal(&ip1->ip[0].in6, &ip2->ip[0].in6) &&
net/netfilter/ipset/ip_set_hash_netnet.c
300
ipv6_addr_equal(&ip1->ip[1].in6, &ip2->ip[1].in6) &&
net/netfilter/ipset/ip_set_hash_netnet.c
326
elem->ip[1] = orig->ip[1];
net/netfilter/ipset/ip_set_hash_netnet.c
333
ip6_netmask(&elem->ip[1], cidr);
net/netfilter/ipset/ip_set_hash_netnet.c
336
ip6_netmask(&elem->ip[0], cidr);
net/netfilter/ipset/ip_set_hash_netnet.c
347
if (nla_put_ipaddr6(skb, IPSET_ATTR_IP, &data->ip[0].in6) ||
net/netfilter/ipset/ip_set_hash_netnet.c
348
nla_put_ipaddr6(skb, IPSET_ATTR_IP2, &data->ip[1].in6) ||
net/netfilter/ipset/ip_set_hash_netnet.c
396
ip6addrptr(skb, opt->flags & IPSET_DIM_ONE_SRC, &e.ip[0].in6);
net/netfilter/ipset/ip_set_hash_netnet.c
397
ip6addrptr(skb, opt->flags & IPSET_DIM_TWO_SRC, &e.ip[1].in6);
net/netfilter/ipset/ip_set_hash_netnet.c
398
ip6_netmask(&e.ip[0], e.cidr[0]);
net/netfilter/ipset/ip_set_hash_netnet.c
399
ip6_netmask(&e.ip[1], e.cidr[1]);
net/netfilter/ipset/ip_set_hash_netnet.c
401
nf_inet_addr_mask_inplace(&e.ip[0], &h->bitmask);
net/netfilter/ipset/ip_set_hash_netnet.c
402
nf_inet_addr_mask_inplace(&e.ip[1], &h->bitmask);
net/netfilter/ipset/ip_set_hash_netnet.c
403
if (e.cidr[0] == HOST_MASK && ipv6_addr_any(&e.ip[0].in6))
net/netfilter/ipset/ip_set_hash_netnet.c
429
ret = ip_set_get_ipaddr6(tb[IPSET_ATTR_IP], &e.ip[0]);
net/netfilter/ipset/ip_set_hash_netnet.c
433
ret = ip_set_get_ipaddr6(tb[IPSET_ATTR_IP2], &e.ip[1]);
net/netfilter/ipset/ip_set_hash_netnet.c
453
ip6_netmask(&e.ip[0], e.cidr[0]);
net/netfilter/ipset/ip_set_hash_netnet.c
454
ip6_netmask(&e.ip[1], e.cidr[1]);
net/netfilter/ipset/ip_set_hash_netnet.c
456
nf_inet_addr_mask_inplace(&e.ip[0], &h->bitmask);
net/netfilter/ipset/ip_set_hash_netnet.c
457
nf_inet_addr_mask_inplace(&e.ip[1], &h->bitmask);
net/netfilter/ipset/ip_set_hash_netnet.c
458
if (e.cidr[0] == HOST_MASK && ipv6_addr_any(&e.ip[0].in6))
net/netfilter/ipset/ip_set_hash_netnet.c
46
__be32 ip[2];
net/netfilter/ipset/ip_set_hash_netnet.c
90
elem->ip[1] = orig->ip[1];
net/netfilter/ipset/ip_set_hash_netnet.c
97
elem->ip[1] &= ip_set_netmask(cidr);
net/netfilter/ipset/ip_set_hash_netport.c
103
if (nla_put_ipaddr4(skb, IPSET_ATTR_IP, data->ip) ||
net/netfilter/ipset/ip_set_hash_netport.c
120
next->ip = d->ip;
net/netfilter/ipset/ip_set_hash_netport.c
147
ip4addrptr(skb, opt->flags & IPSET_DIM_ONE_SRC, &e.ip);
net/netfilter/ipset/ip_set_hash_netport.c
148
e.ip &= ip_set_netmask(e.cidr + 1);
net/netfilter/ipset/ip_set_hash_netport.c
161
u32 port, port_to, p = 0, ip = 0, ip_to = 0, i = 0;
net/netfilter/ipset/ip_set_hash_netport.c
175
ret = ip_set_get_hostipaddr4(tb[IPSET_ATTR_IP], &ip);
net/netfilter/ipset/ip_set_hash_netport.c
215
e.ip = htonl(ip & ip_set_hostmask(e.cidr + 1));
net/netfilter/ipset/ip_set_hash_netport.c
231
if (ip_to < ip)
net/netfilter/ipset/ip_set_hash_netport.c
232
swap(ip, ip_to);
net/netfilter/ipset/ip_set_hash_netport.c
233
if (ip + UINT_MAX == ip_to)
net/netfilter/ipset/ip_set_hash_netport.c
236
ip_set_mask_from_to(ip, ip_to, e.cidr + 1);
net/netfilter/ipset/ip_set_hash_netport.c
240
ip = ntohl(h->next.ip);
net/netfilter/ipset/ip_set_hash_netport.c
246
e.ip = htonl(ip);
net/netfilter/ipset/ip_set_hash_netport.c
247
ip = ip_set_range_to_cidr(ip, ip_to, &cidr);
net/netfilter/ipset/ip_set_hash_netport.c
262
} while (ip++ < ip_to);
net/netfilter/ipset/ip_set_hash_netport.c
269
union nf_inet_addr ip;
net/netfilter/ipset/ip_set_hash_netport.c
283
return ipv6_addr_equal(&ip1->ip.in6, &ip2->ip.in6) &&
net/netfilter/ipset/ip_set_hash_netport.c
310
ip6_netmask(&elem->ip, cidr);
net/netfilter/ipset/ip_set_hash_netport.c
320
if (nla_put_ipaddr6(skb, IPSET_ATTR_IP, &data->ip.in6) ||
net/netfilter/ipset/ip_set_hash_netport.c
367
ip6addrptr(skb, opt->flags & IPSET_DIM_ONE_SRC, &e.ip.in6);
net/netfilter/ipset/ip_set_hash_netport.c
368
ip6_netmask(&e.ip, e.cidr + 1);
net/netfilter/ipset/ip_set_hash_netport.c
397
ret = ip_set_get_ipaddr6(tb[IPSET_ATTR_IP], &e.ip);
net/netfilter/ipset/ip_set_hash_netport.c
411
ip6_netmask(&e.ip, e.cidr + 1);
net/netfilter/ipset/ip_set_hash_netport.c
52
__be32 ip;
net/netfilter/ipset/ip_set_hash_netport.c
66
return ip1->ip == ip2->ip &&
net/netfilter/ipset/ip_set_hash_netport.c
93
elem->ip &= ip_set_netmask(cidr);
net/netfilter/ipset/ip_set_hash_netportnet.c
102
elem->ip[1] &= ip_set_netmask(cidr);
net/netfilter/ipset/ip_set_hash_netportnet.c
105
elem->ip[0] &= ip_set_netmask(cidr);
net/netfilter/ipset/ip_set_hash_netportnet.c
116
if (nla_put_ipaddr4(skb, IPSET_ATTR_IP, data->ip[0]) ||
net/netfilter/ipset/ip_set_hash_netportnet.c
117
nla_put_ipaddr4(skb, IPSET_ATTR_IP2, data->ip[1]) ||
net/netfilter/ipset/ip_set_hash_netportnet.c
169
ip4addrptr(skb, opt->flags & IPSET_DIM_ONE_SRC, &e.ip[0]);
net/netfilter/ipset/ip_set_hash_netportnet.c
170
ip4addrptr(skb, opt->flags & IPSET_DIM_THREE_SRC, &e.ip[1]);
net/netfilter/ipset/ip_set_hash_netportnet.c
171
e.ip[0] &= ip_set_netmask(e.cidr[0]);
net/netfilter/ipset/ip_set_hash_netportnet.c
172
e.ip[1] &= ip_set_netmask(e.cidr[1]);
net/netfilter/ipset/ip_set_hash_netportnet.c
195
u32 ip = 0, ip_to = 0, p = 0, port, port_to;
net/netfilter/ipset/ip_set_hash_netportnet.c
210
ret = ip_set_get_hostipaddr4(tb[IPSET_ATTR_IP], &ip);
net/netfilter/ipset/ip_set_hash_netportnet.c
259
e.ip[0] = htonl(ip & ip_set_hostmask(e.cidr[0]));
net/netfilter/ipset/ip_set_hash_netportnet.c
260
e.ip[1] = htonl(ip2_from & ip_set_hostmask(e.cidr[1]));
net/netfilter/ipset/ip_set_hash_netportnet.c
266
ip_to = ip;
net/netfilter/ipset/ip_set_hash_netportnet.c
271
if (ip > ip_to)
net/netfilter/ipset/ip_set_hash_netportnet.c
272
swap(ip, ip_to);
net/netfilter/ipset/ip_set_hash_netportnet.c
273
if (unlikely(ip + UINT_MAX == ip_to))
net/netfilter/ipset/ip_set_hash_netportnet.c
276
ip_set_mask_from_to(ip, ip_to, e.cidr[0]);
net/netfilter/ipset/ip_set_hash_netportnet.c
300
ip = ntohl(h->next.ip[0]);
net/netfilter/ipset/ip_set_hash_netportnet.c
302
ip2 = ntohl(h->next.ip[1]);
net/netfilter/ipset/ip_set_hash_netportnet.c
309
e.ip[0] = htonl(ip);
net/netfilter/ipset/ip_set_hash_netportnet.c
310
ip = hash_netportnet4_range_to_cidr(ip, ip_to, &e.cidr[0]);
net/netfilter/ipset/ip_set_hash_netportnet.c
315
e.ip[1] = htonl(ip2);
net/netfilter/ipset/ip_set_hash_netportnet.c
332
} while (ip++ < ip_to);
net/netfilter/ipset/ip_set_hash_netportnet.c
339
union nf_inet_addr ip[2];
net/netfilter/ipset/ip_set_hash_netportnet.c
357
return ipv6_addr_equal(&ip1->ip[0].in6, &ip2->ip[0].in6) &&
net/netfilter/ipset/ip_set_hash_netportnet.c
358
ipv6_addr_equal(&ip1->ip[1].in6, &ip2->ip[1].in6) &&
net/netfilter/ipset/ip_set_hash_netportnet.c
386
elem->ip[1] = orig->ip[1];
net/netfilter/ipset/ip_set_hash_netportnet.c
394
ip6_netmask(&elem->ip[1], cidr);
net/netfilter/ipset/ip_set_hash_netportnet.c
397
ip6_netmask(&elem->ip[0], cidr);
net/netfilter/ipset/ip_set_hash_netportnet.c
408
if (nla_put_ipaddr6(skb, IPSET_ATTR_IP, &data->ip[0].in6) ||
net/netfilter/ipset/ip_set_hash_netportnet.c
409
nla_put_ipaddr6(skb, IPSET_ATTR_IP2, &data->ip[1].in6) ||
net/netfilter/ipset/ip_set_hash_netportnet.c
46
__be32 ip[2];
net/netfilter/ipset/ip_set_hash_netportnet.c
464
ip6addrptr(skb, opt->flags & IPSET_DIM_ONE_SRC, &e.ip[0].in6);
net/netfilter/ipset/ip_set_hash_netportnet.c
465
ip6addrptr(skb, opt->flags & IPSET_DIM_THREE_SRC, &e.ip[1].in6);
net/netfilter/ipset/ip_set_hash_netportnet.c
466
ip6_netmask(&e.ip[0], e.cidr[0]);
net/netfilter/ipset/ip_set_hash_netportnet.c
467
ip6_netmask(&e.ip[1], e.cidr[1]);
net/netfilter/ipset/ip_set_hash_netportnet.c
496
ret = ip_set_get_ipaddr6(tb[IPSET_ATTR_IP], &e.ip[0]);
net/netfilter/ipset/ip_set_hash_netportnet.c
500
ret = ip_set_get_ipaddr6(tb[IPSET_ATTR_IP2], &e.ip[1]);
net/netfilter/ipset/ip_set_hash_netportnet.c
520
ip6_netmask(&e.ip[0], e.cidr[0]);
net/netfilter/ipset/ip_set_hash_netportnet.c
521
ip6_netmask(&e.ip[1], e.cidr[1]);
net/netfilter/ipset/ip_set_hash_netportnet.c
94
elem->ip[1] = orig->ip[1];
net/netfilter/ipvs/ip_vs_conn.c
1155
ntohl(cp->daddr.ip));
net/netfilter/ipvs/ip_vs_conn.c
116
return (jhash_3words((__force u32)addr->ip, (__force u32)port, proto,
net/netfilter/ipvs/ip_vs_conn.c
1175
ntohl(cp->caddr.ip), ntohs(cp->cport),
net/netfilter/ipvs/ip_vs_conn.c
1176
ntohl(cp->vaddr.ip), ntohs(cp->vport),
net/netfilter/ipvs/ip_vs_conn.c
1221
ntohl(cp->daddr.ip));
net/netfilter/ipvs/ip_vs_conn.c
1241
ntohl(cp->caddr.ip), ntohs(cp->cport),
net/netfilter/ipvs/ip_vs_conn.c
1242
ntohl(cp->vaddr.ip), ntohs(cp->vport),
net/netfilter/ipvs/ip_vs_core.c
1162
snet.ip = caddr->ip & svc->netmask;
net/netfilter/ipvs/ip_vs_core.c
1282
ip_hdr(skb)->saddr = cp->vaddr.ip;
net/netfilter/ipvs/ip_vs_core.c
262
const union nf_inet_addr fwmark = { .ip = htonl(svc->fwmark) };
net/netfilter/ipvs/ip_vs_core.c
283
snet.ip = src_addr->ip & svc->netmask;
net/netfilter/ipvs/ip_vs_core.c
576
return (inet_addr_type(net, addr->ip) == RTN_UNICAST);
net/netfilter/ipvs/ip_vs_core.c
757
iph->saddr = cp->vaddr.ip;
net/netfilter/ipvs/ip_vs_core.c
759
ciph->daddr = cp->vaddr.ip;
net/netfilter/ipvs/ip_vs_core.c
762
iph->daddr = cp->daddr.ip;
net/netfilter/ipvs/ip_vs_core.c
764
ciph->saddr = cp->daddr.ip;
net/netfilter/ipvs/ip_vs_core.c
983
snet.ip = iph->saddr;
net/netfilter/ipvs/ip_vs_ctl.c
1077
atype = inet_addr_type(svc->ipvs->net, udest->addr.ip);
net/netfilter/ipvs/ip_vs_ctl.c
2429
ntohl(svc->addr.ip),
net/netfilter/ipvs/ip_vs_ctl.c
2463
ntohl(dest->addr.ip),
net/netfilter/ipvs/ip_vs_ctl.c
2663
usvc->addr.ip = usvc_compat->addr;
net/netfilter/ipvs/ip_vs_ctl.c
2680
udest->addr.ip = udest_compat->addr;
net/netfilter/ipvs/ip_vs_ctl.c
2764
if (!usvc.fwmark && !usvc.addr.ip && !usvc.port) {
net/netfilter/ipvs/ip_vs_ctl.c
2781
usvc.protocol, &usvc.addr.ip,
net/netfilter/ipvs/ip_vs_ctl.c
2851
dst->addr = src->addr.ip;
net/netfilter/ipvs/ip_vs_ctl.c
2919
union nf_inet_addr addr = { .ip = get->addr };
net/netfilter/ipvs/ip_vs_ctl.c
2947
entry.addr = dest->addr.ip;
net/netfilter/ipvs/ip_vs_ctl.c
3114
addr.ip = entry->addr;
net/netfilter/ipvs/ip_vs_ctl.c
317
__be32 addr_fold = addr->ip;
net/netfilter/ipvs/ip_vs_ctl.c
3728
c->mcast_group.ip))
net/netfilter/ipvs/ip_vs_ctl.c
3812
c.mcast_group.ip = nla_get_in_addr(a);
net/netfilter/ipvs/ip_vs_ctl.c
3813
if (!ipv4_is_multicast(c.mcast_group.ip))
net/netfilter/ipvs/ip_vs_ctl.c
545
__be32 addr_fold = addr->ip;
net/netfilter/ipvs/ip_vs_dh.c
72
__be32 addr_fold = addr->ip;
net/netfilter/ipvs/ip_vs_ftp.c
162
addr->ip = get_unaligned((__be32 *) p);
net/netfilter/ipvs/ip_vs_ftp.c
289
&from.ip, ntohs(port), &cp->caddr.ip, 0);
net/netfilter/ipvs/ip_vs_ftp.c
344
from.ip = n_cp->vaddr.ip;
net/netfilter/ipvs/ip_vs_ftp.c
347
((unsigned char *)&from.ip)[0],
net/netfilter/ipvs/ip_vs_ftp.c
348
((unsigned char *)&from.ip)[1],
net/netfilter/ipvs/ip_vs_ftp.c
349
((unsigned char *)&from.ip)[2],
net/netfilter/ipvs/ip_vs_ftp.c
350
((unsigned char *)&from.ip)[3],
net/netfilter/ipvs/ip_vs_ftp.c
499
IP_VS_DBG(7, "PORT %pI4:%u detected\n", &to.ip, ntohs(port));
net/netfilter/ipvs/ip_vs_ftp.c
504
&to.ip, ntohs(port), &cp->vaddr.ip,
net/netfilter/ipvs/ip_vs_lblc.c
150
__be32 addr_fold = addr->ip;
net/netfilter/ipvs/ip_vs_lblcr.c
313
__be32 addr_fold = addr->ip;
net/netfilter/ipvs/ip_vs_mh.c
92
__be32 addr_fold = addr->ip;
net/netfilter/ipvs/ip_vs_proto_tcp.c
119
csum_fold(ip_vs_check_diff4(oldip->ip, newip->ip,
net/netfilter/ipvs/ip_vs_proto_tcp.c
140
~csum_fold(ip_vs_check_diff4(oldip->ip, newip->ip,
net/netfilter/ipvs/ip_vs_proto_tcp.c
209
tcph->check = csum_tcpudp_magic(cp->vaddr.ip,
net/netfilter/ipvs/ip_vs_proto_tcp.c
210
cp->caddr.ip,
net/netfilter/ipvs/ip_vs_proto_tcp.c
292
tcph->check = csum_tcpudp_magic(cp->caddr.ip,
net/netfilter/ipvs/ip_vs_proto_tcp.c
293
cp->daddr.ip,
net/netfilter/ipvs/ip_vs_proto_udp.c
107
csum_fold(ip_vs_check_diff4(oldip->ip, newip->ip,
net/netfilter/ipvs/ip_vs_proto_udp.c
129
~csum_fold(ip_vs_check_diff4(oldip->ip, newip->ip,
net/netfilter/ipvs/ip_vs_proto_udp.c
202
udph->check = csum_tcpudp_magic(cp->vaddr.ip,
net/netfilter/ipvs/ip_vs_proto_udp.c
203
cp->caddr.ip,
net/netfilter/ipvs/ip_vs_proto_udp.c
286
udph->check = csum_tcpudp_magic(cp->caddr.ip,
net/netfilter/ipvs/ip_vs_proto_udp.c
287
cp->daddr.ip,
net/netfilter/ipvs/ip_vs_sched.c
159
&svc->addr.ip, ntohs(svc->port), msg);
net/netfilter/ipvs/ip_vs_sh.c
86
__be32 addr_fold = addr->ip;
net/netfilter/ipvs/ip_vs_sync.c
1778
c->mcast_group.ip = cpu_to_be32(IP_VS_SYNC_GROUP);
net/netfilter/ipvs/ip_vs_sync.c
597
s->caddr = cp->caddr.ip;
net/netfilter/ipvs/ip_vs_sync.c
598
s->vaddr = cp->vaddr.ip;
net/netfilter/ipvs/ip_vs_sync.c
599
s->daddr = cp->daddr.ip;
net/netfilter/ipvs/ip_vs_sync.c
743
s->v4.caddr = cp->caddr.ip;
net/netfilter/ipvs/ip_vs_sync.c
744
s->v4.vaddr = cp->vaddr.ip;
net/netfilter/ipvs/ip_vs_sync.c
745
s->v4.daddr = cp->daddr.ip;
net/netfilter/ipvs/ip_vs_xmit.c
1153
local = __ip_vs_get_out_rt(ipvs, cp->af, skb, cp->dest, cp->daddr.ip,
net/netfilter/ipvs/ip_vs_xmit.c
1239
udp_set_csum(!check, skb, saddr, cp->daddr.ip, skb->len);
net/netfilter/ipvs/ip_vs_xmit.c
1256
iph->daddr = cp->daddr.ip;
net/netfilter/ipvs/ip_vs_xmit.c
1431
local = __ip_vs_get_out_rt(cp->ipvs, cp->af, skb, cp->dest, cp->daddr.ip,
net/netfilter/ipvs/ip_vs_xmit.c
1522
local = __ip_vs_get_out_rt(cp->ipvs, cp->af, skb, cp->dest, cp->daddr.ip, rt_mode,
net/netfilter/ipvs/ip_vs_xmit.c
1540
__func__, &cp->daddr.ip);
net/netfilter/ipvs/ip_vs_xmit.c
1547
if (local && ipv4_is_loopback(cp->daddr.ip) && was_input) {
net/netfilter/ipvs/ip_vs_xmit.c
1550
__func__, &cp->daddr.ip);
net/netfilter/ipvs/ip_vs_xmit.c
321
*ret_saddr = dest_dst->dst_saddr.ip;
net/netfilter/ipvs/ip_vs_xmit.c
330
rt = do_output_route4(net, dest->addr.ip, rt_mode,
net/netfilter/ipvs/ip_vs_xmit.c
331
&dest_dst->dst_saddr.ip);
net/netfilter/ipvs/ip_vs_xmit.c
347
&dest->addr.ip, &dest_dst->dst_saddr.ip,
net/netfilter/ipvs/ip_vs_xmit.c
350
*ret_saddr = dest_dst->dst_saddr.ip;
net/netfilter/ipvs/ip_vs_xmit.c
792
local = __ip_vs_get_out_rt(cp->ipvs, cp->af, skb, cp->dest, cp->daddr.ip,
net/netfilter/ipvs/ip_vs_xmit.c
818
if (local && ipv4_is_loopback(cp->daddr.ip) && was_input) {
net/netfilter/ipvs/ip_vs_xmit.c
835
ip_hdr(skb)->daddr = cp->daddr.ip;
net/netfilter/nf_conntrack_bpf.c
87
src->ip = bpf_tuple->ipv4.saddr;
net/netfilter/nf_conntrack_bpf.c
89
dst->ip = bpf_tuple->ipv4.daddr;
net/netfilter/nf_conntrack_broadcast.c
68
exp->mask.src.u3.ip = mask;
net/netfilter/nf_conntrack_core.c
301
tuple->src.u3.ip = ap[0];
net/netfilter/nf_conntrack_core.c
302
tuple->dst.u3.ip = ap[1];
net/netfilter/nf_conntrack_core.c
441
inverse->src.u3.ip = orig->dst.u3.ip;
net/netfilter/nf_conntrack_core.c
442
inverse->dst.u3.ip = orig->src.u3.ip;
net/netfilter/nf_conntrack_ftp.c
161
cmd->u3.ip = htonl((array[0] << 24) | (array[1] << 16) |
net/netfilter/nf_conntrack_ftp.c
253
cmd->u3.ip = htonl((array[0] << 24) | (array[1] << 16)
net/netfilter/nf_conntrack_ftp.c
496
&cmd.u3.ip,
net/netfilter/nf_conntrack_ftp.c
497
&ct->tuplehash[dir].tuple.src.u3.ip);
net/netfilter/nf_conntrack_h323_main.c
600
p = data + taddr->ipAddress.ip;
net/netfilter/nf_conntrack_h323_main.c
606
p = data + taddr->ip6Address.ip;
net/netfilter/nf_conntrack_h323_main.c
687
fl1.daddr = src->ip;
net/netfilter/nf_conntrack_h323_main.c
690
fl2.daddr = dst->ip;
net/netfilter/nf_conntrack_h323_types.c
39
offsetof(TransportAddress_ip6Address, ip), NULL},
net/netfilter/nf_conntrack_h323_types.c
9
offsetof(TransportAddress_ipAddress, ip), NULL},
net/netfilter/nf_conntrack_irc.c
220
if ((tuple->src.u3.ip != dcc_ip &&
net/netfilter/nf_conntrack_irc.c
221
ct->tuplehash[!dir].tuple.dst.u3.ip != dcc_ip) ||
net/netfilter/nf_conntrack_irc.c
224
&tuple->src.u3.ip,
net/netfilter/nf_conntrack_irc.c
74
static int parse_dcc(char *data, const char *data_end, __be32 *ip,
net/netfilter/nf_conntrack_irc.c
93
*ip = cpu_to_be32(simple_strtoul(data, &data, 10));
net/netfilter/nf_conntrack_netlink.c
1091
filter_tuple->src.u3.ip != ct_tuple->src.u3.ip)
net/netfilter/nf_conntrack_netlink.c
1095
filter_tuple->dst.u3.ip != ct_tuple->dst.u3.ip)
net/netfilter/nf_conntrack_netlink.c
1310
t->src.u3.ip = nla_get_in_addr(tb[CTA_IP_V4_SRC]);
net/netfilter/nf_conntrack_netlink.c
1317
t->dst.u3.ip = nla_get_in_addr(tb[CTA_IP_V4_DST]);
net/netfilter/nf_conntrack_netlink.c
96
if (nla_put_in_addr(skb, CTA_IP_V4_SRC, tuple->src.u3.ip) ||
net/netfilter/nf_conntrack_netlink.c
97
nla_put_in_addr(skb, CTA_IP_V4_DST, tuple->dst.u3.ip))
net/netfilter/nf_conntrack_proto.c
274
tuple.src.u3.ip = inet->inet_rcv_saddr;
net/netfilter/nf_conntrack_proto.c
276
tuple.dst.u3.ip = inet->inet_daddr;
net/netfilter/nf_conntrack_proto.c
299
.tuple.dst.u3.ip;
net/netfilter/nf_conntrack_proto_icmp.c
176
&outer_daddr->ip, &ct_daddr->ip);
net/netfilter/nf_conntrack_proto_icmp.c
243
outer_daddr.ip = ip_hdr(skb)->daddr;
net/netfilter/nf_conntrack_sip.c
156
ret = in4_pton(cp, limit - cp, (u8 *)&addr->ip, -1, &end);
net/netfilter/nf_conntrack_sip.c
656
ret = in4_pton(cp, limit - cp, (u8 *)&addr->ip, -1, &end);
net/netfilter/nf_conntrack_sip.c
880
fl.u.ip4.daddr = daddr->ip;
net/netfilter/nf_conntrack_standalone.c
42
&tuple->src.u3.ip, &tuple->dst.u3.ip);
net/netfilter/nf_flow_table_path.c
295
fl.u.ip4.daddr = ct->tuplehash[dir].tuple.src.u3.ip;
net/netfilter/nf_flow_table_path.c
296
fl.u.ip4.saddr = ct->tuplehash[!dir].tuple.src.u3.ip;
net/netfilter/nf_nat_core.c
1077
range->min_addr.ip = nla_get_be32(tb[CTA_NAT_V4_MINIP]);
net/netfilter/nf_nat_core.c
1081
range->max_addr.ip = nla_get_be32_default(tb[CTA_NAT_V4_MAXIP],
net/netfilter/nf_nat_core.c
1082
range->min_addr.ip);
net/netfilter/nf_nat_core.c
395
return ntohl(t->src.u3.ip) >= ntohl(range->min_addr.ip) &&
net/netfilter/nf_nat_core.c
396
ntohl(t->src.u3.ip) <= ntohl(range->max_addr.ip);
net/netfilter/nf_nat_core.c
526
max = sizeof(var_ipp->ip) / sizeof(u32) - 1;
net/netfilter/nf_nat_core.c
68
fl4->daddr = t->dst.u3.ip;
net/netfilter/nf_nat_core.c
79
fl4->saddr = t->src.u3.ip;
net/netfilter/nf_nat_core.c
846
union nf_inet_addr ip =
net/netfilter/nf_nat_core.c
852
.min_addr = ip,
net/netfilter/nf_nat_core.c
853
.max_addr = ip,
net/netfilter/nf_nat_ftp.c
41
((unsigned char *)&addr->ip)[0],
net/netfilter/nf_nat_ftp.c
42
((unsigned char *)&addr->ip)[1],
net/netfilter/nf_nat_ftp.c
43
((unsigned char *)&addr->ip)[2],
net/netfilter/nf_nat_ftp.c
44
((unsigned char *)&addr->ip)[3],
net/netfilter/nf_nat_ftp.c
50
&addr->ip, port);
net/netfilter/nf_nat_irc.c
72
snprintf(buffer, sizeof(buffer), "%u %u", ntohl(newaddr.ip), port);
net/netfilter/nf_nat_irc.c
74
buffer, &newaddr.ip, port);
net/netfilter/nf_nat_masquerade.c
205
addr.ip = ifa->ifa_address;
net/netfilter/nf_nat_masquerade.c
49
if (ct->tuplehash[IP_CT_DIR_ORIGINAL].tuple.src.u3.ip == 0)
net/netfilter/nf_nat_masquerade.c
68
newrange.min_addr.ip = newsrc;
net/netfilter/nf_nat_masquerade.c
69
newrange.max_addr.ip = newsrc;
net/netfilter/nf_nat_proto.c
330
csum_replace4(&iph->check, iph->saddr, target->src.u3.ip);
net/netfilter/nf_nat_proto.c
331
iph->saddr = target->src.u3.ip;
net/netfilter/nf_nat_proto.c
333
csum_replace4(&iph->check, iph->daddr, target->dst.u3.ip);
net/netfilter/nf_nat_proto.c
334
iph->daddr = target->dst.u3.ip;
net/netfilter/nf_nat_proto.c
413
newip = t->src.u3.ip;
net/netfilter/nf_nat_proto.c
416
newip = t->dst.u3.ip;
net/netfilter/nf_nat_proto.c
524
struct iphdr ip;
net/netfilter/nf_nat_proto.c
735
if (ct->tuplehash[dir].tuple.src.u3.ip !=
net/netfilter/nf_nat_proto.c
736
ct->tuplehash[!dir].tuple.dst.u3.ip ||
net/netfilter/nf_nat_proto.c
766
if (ct->tuplehash[dir].tuple.dst.u3.ip !=
net/netfilter/nf_nat_proto.c
767
ct->tuplehash[!dir].tuple.src.u3.ip) {
net/netfilter/nf_nat_redirect.c
60
newdst.ip = htonl(INADDR_LOOPBACK);
net/netfilter/nf_nat_redirect.c
70
newdst.ip = ifa->ifa_local;
net/netfilter/nf_nat_redirect.c
73
if (!newdst.ip)
net/netfilter/nf_nat_sip.c
74
return sprintf(buffer, "%pI4", &addr->ip);
net/netfilter/nf_nat_sip.c
87
return sprintf(buffer, "%pI4:%u", &addr->ip, port);
net/netfilter/nfnetlink_osf.c
167
const struct iphdr *ip,
net/netfilter/nfnetlink_osf.c
180
ctx->totlen = ntohs(ip->tot_len);
net/netfilter/nfnetlink_osf.c
181
ctx->df = ntohs(ip->frag_off) & IP_DF;
net/netfilter/nfnetlink_osf.c
202
const struct iphdr *ip = ip_hdr(skb);
net/netfilter/nfnetlink_osf.c
214
tcp = nf_osf_hdr_ctx_init(&ctx, skb, ip, opts, &_tcph);
net/netfilter/nfnetlink_osf.c
239
&ip->saddr, ntohs(tcp->source),
net/netfilter/nfnetlink_osf.c
240
&ip->daddr, ntohs(tcp->dest),
net/netfilter/nfnetlink_osf.c
241
f->ttl - ip->ttl);
net/netfilter/nfnetlink_osf.c
251
&ip->saddr, ntohs(tcp->source),
net/netfilter/nfnetlink_osf.c
252
&ip->daddr, ntohs(tcp->dest));
net/netfilter/nfnetlink_osf.c
265
const struct iphdr *ip = ip_hdr(skb);
net/netfilter/nfnetlink_osf.c
276
tcp = nf_osf_hdr_ctx_init(&ctx, skb, ip, opts, &_tcph);
net/netfilter/nfnetlink_osf.c
35
const struct iphdr *ip = ip_hdr(skb);
net/netfilter/nfnetlink_osf.c
40
return ip->ttl == f_ttl;
net/netfilter/nfnetlink_osf.c
43
else if (ip->ttl <= f_ttl)
net/netfilter/nfnetlink_osf.c
47
if (inet_ifa_match(ip->saddr, ifa)) {
net/netfilter/nfnetlink_osf.c
48
ret = (ip->ttl == f_ttl);
net/netfilter/nft_compat.c
153
entry->e4.ip.proto = proto;
net/netfilter/nft_compat.c
154
entry->e4.ip.invflags = inv ? IPT_INV_PROTO : 0;
net/netfilter/nft_compat.c
454
entry->e4.ip.proto = proto;
net/netfilter/nft_compat.c
455
entry->e4.ip.invflags = inv ? IPT_INV_PROTO : 0;
net/netfilter/nft_ct.c
199
*dest = (__force __u32)tuple->src.u3.ip;
net/netfilter/nft_ct.c
204
*dest = (__force __u32)tuple->dst.u3.ip;
net/netfilter/nft_ct.c
453
src.u3.ip);
net/netfilter/nft_ct.c
469
len = sizeof_field(struct nf_conntrack_tuple, src.u3.ip);
net/netfilter/nft_nat.c
203
alen = sizeof_field(struct nf_nat_range, min_addr.ip);
net/netfilter/nft_nat.c
39
range->min_addr.ip = (__force __be32)
net/netfilter/nft_nat.c
41
range->max_addr.ip = (__force __be32)
net/netfilter/nft_nat.c
75
new_addr.ip = ip_hdr(skb)->saddr;
net/netfilter/nft_nat.c
84
new_addr.ip = ip_hdr(skb)->daddr;
net/netfilter/xt_CT.c
60
if (e->ip.invflags & IPT_INV_PROTO)
net/netfilter/xt_CT.c
62
return e->ip.proto;
net/netfilter/xt_HMARK.c
253
struct iphdr *ip, _ip;
net/netfilter/xt_HMARK.c
256
ip = (struct iphdr *) (skb->data + nhoff);
net/netfilter/xt_HMARK.c
257
if (ip->protocol == IPPROTO_ICMP) {
net/netfilter/xt_HMARK.c
259
if (get_inner_hdr(skb, ip->ihl * 4, &nhoff)) {
net/netfilter/xt_HMARK.c
260
ip = skb_header_pointer(skb, nhoff, sizeof(_ip), &_ip);
net/netfilter/xt_HMARK.c
261
if (ip == NULL)
net/netfilter/xt_HMARK.c
266
t->src = ip->saddr & info->src_mask.ip;
net/netfilter/xt_HMARK.c
267
t->dst = ip->daddr & info->dst_mask.ip;
net/netfilter/xt_HMARK.c
272
t->proto = ip->protocol;
net/netfilter/xt_HMARK.c
279
if (ip_is_fragment(ip))
net/netfilter/xt_HMARK.c
282
hmark_set_tuple_ports(skb, (ip->ihl * 4) + nhoff, t, info);
net/netfilter/xt_NETMAP.c
95
newrange.min_addr.ip = new_ip;
net/netfilter/xt_NETMAP.c
96
newrange.max_addr.ip = new_ip;
net/netfilter/xt_TPROXY.c
98
return tproxy_tg4(xt_net(par), skb, tgi->laddr.ip, tgi->lport,
net/netfilter/xt_cluster.c
18
return (__force u32)ct->tuplehash[IP_CT_DIR_ORIGINAL].tuple.src.u3.ip;
net/netfilter/xt_cluster.c
27
xt_cluster_hash_ipv4(u_int32_t ip, const struct xt_cluster_match_info *info)
net/netfilter/xt_cluster.c
29
return jhash_1word(ip, info->hash_seed);
net/netfilter/xt_cluster.c
33
xt_cluster_hash_ipv6(const void *ip, const struct xt_cluster_match_info *info)
net/netfilter/xt_cluster.c
35
return jhash2(ip, NF_CT_TUPLE_L3SIZE / sizeof(__u32), info->hash_seed);
net/netfilter/xt_connlimit.c
61
key[0] &= (__force __u32)info->mask.ip;
net/netfilter/xt_conntrack.c
31
return ((kaddr->ip ^ uaddr->ip) & umask->ip) == 0;
net/netfilter/xt_ecn.c
129
const struct ip6t_ip6 *ip = par->entryinfo;
net/netfilter/xt_ecn.c
138
(ip->proto != IPPROTO_TCP || ip->invflags & IP6T_INV_PROTO)) {
net/netfilter/xt_ecn.c
87
const struct ipt_ip *ip = par->entryinfo;
net/netfilter/xt_ecn.c
96
(ip->proto != IPPROTO_TCP || ip->invflags & IPT_INV_PROTO)) {
net/netfilter/xt_hashlimit.c
1103
&ent->dst.ip.src,
net/netfilter/xt_hashlimit.c
1105
&ent->dst.ip.dst,
net/netfilter/xt_hashlimit.c
652
dst->ip.dst = maskl(ip_hdr(skb)->daddr,
net/netfilter/xt_hashlimit.c
655
dst->ip.src = maskl(ip_hdr(skb)->saddr,
net/netfilter/xt_hashlimit.c
76
} ip;
net/netfilter/xt_iprange.c
24
m = ntohl(iph->saddr) < ntohl(info->src_min.ip);
net/netfilter/xt_iprange.c
25
m |= ntohl(iph->saddr) > ntohl(info->src_max.ip);
net/netfilter/xt_iprange.c
31
&info->src_min.ip,
net/netfilter/xt_iprange.c
32
&info->src_max.ip);
net/netfilter/xt_iprange.c
37
m = ntohl(iph->daddr) < ntohl(info->dst_min.ip);
net/netfilter/xt_iprange.c
38
m |= ntohl(iph->daddr) > ntohl(info->dst_max.ip);
net/netfilter/xt_iprange.c
44
&info->dst_min.ip,
net/netfilter/xt_iprange.c
45
&info->dst_max.ip);
net/netfilter/xt_ipvs.c
38
return ((kaddr->ip ^ uaddr->ip) & umask->ip) == 0;
net/netfilter/xt_l2tp.c
263
const struct ipt_ip *ip = &e->ip;
net/netfilter/xt_l2tp.c
270
if ((ip->proto != IPPROTO_UDP) &&
net/netfilter/xt_l2tp.c
271
(ip->proto != IPPROTO_L2TP)) {
net/netfilter/xt_l2tp.c
276
if ((ip->proto == IPPROTO_L2TP) &&
net/netfilter/xt_l2tp.c
290
const struct ip6t_ip6 *ip = &e->ipv6;
net/netfilter/xt_l2tp.c
297
if ((ip->proto != IPPROTO_UDP) &&
net/netfilter/xt_l2tp.c
298
(ip->proto != IPPROTO_L2TP)) {
net/netfilter/xt_l2tp.c
303
if ((ip->proto == IPPROTO_L2TP) &&
net/netfilter/xt_multiport.c
127
const struct ipt_ip *ip = par->entryinfo;
net/netfilter/xt_multiport.c
130
return check(ip->proto, ip->invflags, multiinfo->flags,
net/netfilter/xt_multiport.c
136
const struct ip6t_ip6 *ip = par->entryinfo;
net/netfilter/xt_multiport.c
139
return check(ip->proto, ip->invflags, multiinfo->flags,
net/netfilter/xt_nat.c
45
dst->min_addr.ip = src->min_ip;
net/netfilter/xt_nat.c
46
dst->max_addr.ip = src->max_ip;
net/netfilter/xt_policy.c
27
return ((a1->ip ^ a2->ip) & m->ip) == 0;
net/netfilter/xt_recent.c
113
return jhash_1word((__force u32)addr->ip, hash_rnd) &
net/netfilter/xt_recent.c
255
addr.ip = iph->daddr;
net/netfilter/xt_recent.c
257
addr.ip = iph->saddr;
net/netfilter/xt_recent.c
529
&e->addr.ip, e->ttl, e->stamps[i], e->index);
net/openvswitch/actions.c
1189
key->ip.ttl = --nh->hop_limit;
net/openvswitch/actions.c
1206
key->ip.ttl = nh->ttl;
net/openvswitch/actions.c
490
flow_key->ip.tos = nh->tos;
net/openvswitch/actions.c
494
flow_key->ip.ttl = nh->ttl;
net/openvswitch/actions.c
530
set_ipv6_addr(skb, flow_key->ip.proto, saddr, masked,
net/openvswitch/actions.c
552
set_ipv6_addr(skb, flow_key->ip.proto, daddr, masked,
net/openvswitch/actions.c
560
flow_key->ip.tos = ipv6_get_dsfield(nh);
net/openvswitch/actions.c
570
flow_key->ip.ttl = nh->hop_limit;
net/openvswitch/conntrack.c
1053
if (key->ip.frag != OVS_FRAG_TYPE_NONE) {
net/openvswitch/conntrack.c
1413
if (nf_ct_set_timeout(net, ct_info.ct, family, key->ip.proto,
net/openvswitch/conntrack.c
1426
key->ip.proto, ct_info.nat, &ct_info.helper);
net/openvswitch/conntrack.c
1470
info->range.min_addr.ip) ||
net/openvswitch/conntrack.c
1471
(info->range.max_addr.ip
net/openvswitch/conntrack.c
1472
!= info->range.min_addr.ip &&
net/openvswitch/conntrack.c
1474
info->range.max_addr.ip))))
net/openvswitch/conntrack.c
218
key->ipv4.ct_orig.src = orig->src.u3.ip;
net/openvswitch/conntrack.c
219
key->ipv4.ct_orig.dst = orig->dst.u3.ip;
net/openvswitch/conntrack.c
450
err = nf_ct_handle_fragments(net, skb, zone, family, &key->ip.proto, &ovs_cb.mru);
net/openvswitch/conntrack.c
458
key->ip.frag = OVS_FRAG_TYPE_NONE;
net/openvswitch/conntrack.c
631
if (key->ip.proto == IPPROTO_UDP)
net/openvswitch/conntrack.c
633
else if (key->ip.proto == IPPROTO_TCP)
net/openvswitch/conntrack.c
635
else if (key->ip.proto == IPPROTO_SCTP)
net/openvswitch/conntrack.c
653
if (key->ip.proto == IPPROTO_UDP)
net/openvswitch/conntrack.c
655
else if (key->ip.proto == IPPROTO_TCP)
net/openvswitch/conntrack.c
657
else if (key->ip.proto == IPPROTO_SCTP)
net/openvswitch/datapath.c
389
later_key.ip.frag = OVS_FRAG_TYPE_LATER;
net/openvswitch/flow.c
398
key->ip.proto = NEXTHDR_NONE;
net/openvswitch/flow.c
399
key->ip.tos = ipv6_get_dsfield(nh);
net/openvswitch/flow.c
400
key->ip.ttl = nh->hop_limit;
net/openvswitch/flow.c
408
key->ip.frag = OVS_FRAG_TYPE_LATER;
net/openvswitch/flow.c
409
key->ip.proto = NEXTHDR_FRAGMENT;
net/openvswitch/flow.c
412
key->ip.frag = OVS_FRAG_TYPE_FIRST;
net/openvswitch/flow.c
414
key->ip.frag = OVS_FRAG_TYPE_NONE;
net/openvswitch/flow.c
426
key->ip.proto = nexthdr;
net/openvswitch/flow.c
690
memset(&key->ip, 0, sizeof(key->ip));
net/openvswitch/flow.c
703
key->ip.proto = nh->protocol;
net/openvswitch/flow.c
704
key->ip.tos = nh->tos;
net/openvswitch/flow.c
705
key->ip.ttl = nh->ttl;
net/openvswitch/flow.c
709
key->ip.frag = OVS_FRAG_TYPE_LATER;
net/openvswitch/flow.c
715
key->ip.frag = OVS_FRAG_TYPE_FIRST;
net/openvswitch/flow.c
717
key->ip.frag = OVS_FRAG_TYPE_NONE;
net/openvswitch/flow.c
720
if (key->ip.proto == IPPROTO_TCP) {
net/openvswitch/flow.c
730
} else if (key->ip.proto == IPPROTO_UDP) {
net/openvswitch/flow.c
738
} else if (key->ip.proto == IPPROTO_SCTP) {
net/openvswitch/flow.c
746
} else if (key->ip.proto == IPPROTO_ICMP) {
net/openvswitch/flow.c
774
key->ip.proto = ntohs(arp->ar_op);
net/openvswitch/flow.c
776
key->ip.proto = 0;
net/openvswitch/flow.c
783
memset(&key->ip, 0, sizeof(key->ip));
net/openvswitch/flow.c
823
memset(&key->ip, 0, sizeof(key->ip));
net/openvswitch/flow.c
836
if (key->ip.frag == OVS_FRAG_TYPE_LATER) {
net/openvswitch/flow.c
841
key->ip.frag = OVS_FRAG_TYPE_FIRST;
net/openvswitch/flow.c
844
if (key->ip.proto == NEXTHDR_TCP) {
net/openvswitch/flow.c
853
} else if (key->ip.proto == NEXTHDR_UDP) {
net/openvswitch/flow.c
861
} else if (key->ip.proto == NEXTHDR_SCTP) {
net/openvswitch/flow.c
869
} else if (key->ip.proto == NEXTHDR_ICMP) {
net/openvswitch/flow.h
106
} ip;
net/openvswitch/flow.h
172
key->ip.proto == NEXTHDR_ICMP &&
net/openvswitch/flow_netlink.c
1536
SW_FLOW_KEY_PUT(match, ip.proto,
net/openvswitch/flow_netlink.c
1538
SW_FLOW_KEY_PUT(match, ip.tos,
net/openvswitch/flow_netlink.c
1540
SW_FLOW_KEY_PUT(match, ip.ttl,
net/openvswitch/flow_netlink.c
1542
SW_FLOW_KEY_PUT(match, ip.frag,
net/openvswitch/flow_netlink.c
1569
SW_FLOW_KEY_PUT(match, ip.proto,
net/openvswitch/flow_netlink.c
1571
SW_FLOW_KEY_PUT(match, ip.tos,
net/openvswitch/flow_netlink.c
1573
SW_FLOW_KEY_PUT(match, ip.ttl,
net/openvswitch/flow_netlink.c
1575
SW_FLOW_KEY_PUT(match, ip.frag,
net/openvswitch/flow_netlink.c
1614
SW_FLOW_KEY_PUT(match, ip.proto,
net/openvswitch/flow_netlink.c
2096
ipv4_key->ipv4_proto = output->ip.proto;
net/openvswitch/flow_netlink.c
2097
ipv4_key->ipv4_tos = output->ip.tos;
net/openvswitch/flow_netlink.c
2098
ipv4_key->ipv4_ttl = output->ip.ttl;
net/openvswitch/flow_netlink.c
2099
ipv4_key->ipv4_frag = output->ip.frag;
net/openvswitch/flow_netlink.c
2113
ipv6_key->ipv6_proto = output->ip.proto;
net/openvswitch/flow_netlink.c
2114
ipv6_key->ipv6_tclass = output->ip.tos;
net/openvswitch/flow_netlink.c
2115
ipv6_key->ipv6_hlimit = output->ip.ttl;
net/openvswitch/flow_netlink.c
2116
ipv6_key->ipv6_frag = output->ip.frag;
net/openvswitch/flow_netlink.c
212
if (match->key->ip.frag != OVS_FRAG_TYPE_LATER) {
net/openvswitch/flow_netlink.c
213
if (match->key->ip.proto == IPPROTO_UDP) {
net/openvswitch/flow_netlink.c
2138
arp_key->arp_op = htons(output->ip.proto);
net/openvswitch/flow_netlink.c
215
if (match->mask && (match->mask->key.ip.proto == 0xff))
net/openvswitch/flow_netlink.c
2158
swkey->ip.frag != OVS_FRAG_TYPE_LATER) {
net/openvswitch/flow_netlink.c
2160
if (swkey->ip.proto == IPPROTO_TCP) {
net/openvswitch/flow_netlink.c
2172
} else if (swkey->ip.proto == IPPROTO_UDP) {
net/openvswitch/flow_netlink.c
2181
} else if (swkey->ip.proto == IPPROTO_SCTP) {
net/openvswitch/flow_netlink.c
219
if (match->key->ip.proto == IPPROTO_SCTP) {
net/openvswitch/flow_netlink.c
2191
swkey->ip.proto == IPPROTO_ICMP) {
net/openvswitch/flow_netlink.c
2201
swkey->ip.proto == IPPROTO_ICMPV6) {
net/openvswitch/flow_netlink.c
221
if (match->mask && (match->mask->key.ip.proto == 0xff))
net/openvswitch/flow_netlink.c
225
if (match->key->ip.proto == IPPROTO_TCP) {
net/openvswitch/flow_netlink.c
228
if (match->mask && (match->mask->key.ip.proto == 0xff)) {
net/openvswitch/flow_netlink.c
234
if (match->key->ip.proto == IPPROTO_ICMP) {
net/openvswitch/flow_netlink.c
236
if (match->mask && (match->mask->key.ip.proto == 0xff))
net/openvswitch/flow_netlink.c
249
if (match->key->ip.frag != OVS_FRAG_TYPE_LATER) {
net/openvswitch/flow_netlink.c
250
if (match->key->ip.proto == IPPROTO_UDP) {
net/openvswitch/flow_netlink.c
252
if (match->mask && (match->mask->key.ip.proto == 0xff))
net/openvswitch/flow_netlink.c
256
if (match->key->ip.proto == IPPROTO_SCTP) {
net/openvswitch/flow_netlink.c
258
if (match->mask && (match->mask->key.ip.proto == 0xff))
net/openvswitch/flow_netlink.c
262
if (match->key->ip.proto == IPPROTO_TCP) {
net/openvswitch/flow_netlink.c
265
if (match->mask && (match->mask->key.ip.proto == 0xff)) {
net/openvswitch/flow_netlink.c
271
if (match->key->ip.proto == IPPROTO_ICMPV6) {
net/openvswitch/flow_netlink.c
273
if (match->mask && (match->mask->key.ip.proto == 0xff))
net/openvswitch/flow_netlink.c
2899
if (ipv4_key->ipv4_proto != flow_key->ip.proto)
net/openvswitch/flow_netlink.c
2902
if (ipv4_key->ipv4_frag != flow_key->ip.frag)
net/openvswitch/flow_netlink.c
2926
if (ipv6_key->ipv6_proto != flow_key->ip.proto)
net/openvswitch/flow_netlink.c
2929
if (ipv6_key->ipv6_frag != flow_key->ip.frag)
net/openvswitch/flow_netlink.c
2940
flow_key->ip.proto != IPPROTO_TCP)
net/openvswitch/flow_netlink.c
2948
flow_key->ip.proto != IPPROTO_UDP)
net/openvswitch/flow_netlink.c
2963
flow_key->ip.proto != IPPROTO_SCTP)
net/sched/act_ct.c
106
be32_to_cpu(target.src.u3.ip));
net/sched/act_ct.c
111
be32_to_cpu(target.dst.u3.ip));
net/sched/act_ct.c
1185
range->min_addr.ip =
net/sched/act_ct.c
1188
range->max_addr.ip =
net/sched/act_ct.c
1189
nla_get_in_addr_default(max_attr, range->min_addr.ip);
net/sched/act_ct.c
1489
range->min_addr.ip))
net/sched/act_ct.c
1492
range->max_addr.ip))
net/sched/cls_flow.c
157
return ntohl(CTTUPLE(skb, src.u3.ip));
net/sched/cls_flow.c
170
return ntohl(CTTUPLE(skb, dst.u3.ip));
net/sched/cls_flower.c
1930
fl_set_key_ip(tb, false, &key->ip, &mask->ip);
net/sched/cls_flower.c
2186
FLOW_DISSECTOR_KEY_IP, ip);
net/sched/cls_flower.c
3503
fl_dump_key_ip(skb, false, &key->ip, &mask->ip)))
net/sched/cls_flower.c
77
struct flow_dissector_key_ip ip;
net/sched/sch_cake.c
593
__be32 ip;
net/sched/sch_cake.c
601
ip = rev ? tuple.dst.u3.ip : tuple.src.u3.ip;
net/sched/sch_cake.c
602
if (ip != keys->addrs.v4addrs.src) {
net/sched/sch_cake.c
603
keys->addrs.v4addrs.src = ip;
net/sched/sch_cake.c
606
ip = rev ? tuple.src.u3.ip : tuple.dst.u3.ip;
net/sched/sch_cake.c
607
if (ip != keys->addrs.v4addrs.dst) {
net/sched/sch_cake.c
608
keys->addrs.v4addrs.dst = ip;
net/sunrpc/svcauth_unix.c
117
static inline int hash_ip6(const struct in6_addr *ip)
net/sunrpc/svcauth_unix.c
119
return hash_32(ipv6_addr_hash(ip), IP_HASHBITS);
net/sunrpc/svcauth_unix.c
293
struct ip_map ip;
net/sunrpc/svcauth_unix.c
296
strcpy(ip.m_class, class);
net/sunrpc/svcauth_unix.c
297
ip.m_addr = *addr;
net/sunrpc/svcauth_unix.c
298
ch = sunrpc_cache_lookup_rcu(cd, &ip.h,
net/sunrpc/svcauth_unix.c
311
struct ip_map ip;
net/sunrpc/svcauth_unix.c
314
ip.m_client = udom;
net/sunrpc/svcauth_unix.c
315
ip.h.flags = 0;
net/sunrpc/svcauth_unix.c
317
set_bit(CACHE_NEGATIVE, &ip.h.flags);
net/sunrpc/svcauth_unix.c
318
ip.h.expiry_time = expiry;
net/sunrpc/svcauth_unix.c
319
ch = sunrpc_cache_update(cd, &ip.h, &ipm->h,
net/xfrm/xfrm_policy.c
165
struct flow_dissector_key_ip ip;
net/xfrm/xfrm_policy.c
3465
fl4->flowi4_dscp = inet_dsfield_to_dscp(flkeys->ip.tos);
net/xfrm/xfrm_policy.c
4399
.offset = offsetof(struct xfrm_flow_keys, ip),
samples/bpf/offwaketime_user.c
49
__u64 ip[PERF_MAX_STACK_DEPTH] = {};
samples/bpf/offwaketime_user.c
54
if (bpf_map_lookup_elem(map_fd[1], &key->tret, ip) != 0) {
samples/bpf/offwaketime_user.c
58
print_ksym(ip[i]);
samples/bpf/offwaketime_user.c
61
if (bpf_map_lookup_elem(map_fd[1], &key->wret, ip) != 0) {
samples/bpf/offwaketime_user.c
65
print_ksym(ip[i]);
samples/bpf/sampleip_kern.c
25
u64 ip;
samples/bpf/sampleip_kern.c
28
ip = PT_REGS_IP(&ctx->regs);
samples/bpf/sampleip_kern.c
29
value = bpf_map_lookup_elem(&ip_map, &ip);
samples/bpf/sampleip_kern.c
34
bpf_map_update_elem(&ip_map, &ip, &init_val, BPF_NOEXIST);
samples/bpf/sampleip_user.c
102
counts[i].ip = next_key;
samples/bpf/sampleip_user.c
111
if (counts[i].ip > _text_addr) {
samples/bpf/sampleip_user.c
112
sym = ksym_search(counts[i].ip);
samples/bpf/sampleip_user.c
118
printf("0x%-17llx %-32s %u\n", counts[i].ip, sym->name,
samples/bpf/sampleip_user.c
121
printf("0x%-17llx %-32s %u\n", counts[i].ip, "(user)",
samples/bpf/sampleip_user.c
77
__u64 ip;
samples/bpf/trace_event_user.c
63
__u64 ip[PERF_MAX_STACK_DEPTH] = {};
samples/bpf/trace_event_user.c
68
if (bpf_map_lookup_elem(map_fd[1], &key->kernstack, ip) != 0) {
samples/bpf/trace_event_user.c
72
print_ksym(ip[i]);
samples/bpf/trace_event_user.c
75
if (bpf_map_lookup_elem(map_fd[1], &key->userstack, ip) != 0) {
samples/bpf/trace_event_user.c
79
print_addr(ip[i]);
samples/bpf/tracex4.bpf.c
14
u64 ip;
samples/bpf/tracex4.bpf.c
40
long ip = 0;
samples/bpf/tracex4.bpf.c
43
BPF_KRETPROBE_READ_RET_IP(ip, ctx);
samples/bpf/tracex4.bpf.c
47
.ip = ip,
samples/bpf/tracex4_user.c
17
__u64 ip;
samples/bpf/tracex4_user.c
44
next_key, (val - v.val) / 1000000000ll, v.ip);
samples/fprobe/fprobe_example.c
51
static int sample_entry_handler(struct fprobe *fp, unsigned long ip,
samples/fprobe/fprobe_example.c
60
trace_printk("Enter <%pS> ip = 0x%p\n", (void *)ip, (void *)ip);
samples/fprobe/fprobe_example.c
62
pr_info("Enter <%pS> ip = 0x%p\n", (void *)ip, (void *)ip);
samples/fprobe/fprobe_example.c
69
static void sample_exit_handler(struct fprobe *fp, unsigned long ip,
samples/fprobe/fprobe_example.c
81
(void *)ip, (void *)ip, (void *)rip, (void *)rip);
samples/fprobe/fprobe_example.c
84
(void *)ip, (void *)ip, (void *)rip, (void *)rip);
samples/ftrace/ftrace-direct-multi-modify.c
10
extern void my_direct_func2(unsigned long ip);
samples/ftrace/ftrace-direct-multi-modify.c
12
void my_direct_func1(unsigned long ip)
samples/ftrace/ftrace-direct-multi-modify.c
14
trace_printk("my direct func1 ip %lx\n", ip);
samples/ftrace/ftrace-direct-multi-modify.c
17
void my_direct_func2(unsigned long ip)
samples/ftrace/ftrace-direct-multi-modify.c
19
trace_printk("my direct func2 ip %lx\n", ip);
samples/ftrace/ftrace-direct-multi-modify.c
9
extern void my_direct_func1(unsigned long ip);
samples/ftrace/ftrace-direct-multi.c
11
extern void my_direct_func(unsigned long ip);
samples/ftrace/ftrace-direct-multi.c
13
void my_direct_func(unsigned long ip)
samples/ftrace/ftrace-direct-multi.c
15
trace_printk("ip %lx\n", ip);
samples/ftrace/ftrace-ops.c
100
static void ops_func_nop(unsigned long ip, unsigned long parent_ip,
samples/ftrace/ftrace-ops.c
107
static void ops_func_count(unsigned long ip, unsigned long parent_ip,
samples/kprobes/kprobe_example.c
32
p->symbol_name, p->addr, regs->ip, regs->flags);
scripts/kconfig/qconf.cc
248
ConfigItem** ip = (ConfigItem**)&menu->data;
scripts/kconfig/qconf.cc
249
for (; *ip; ip = &(*ip)->nextItem) {
scripts/kconfig/qconf.cc
250
if (*ip == this) {
scripts/kconfig/qconf.cc
251
*ip = nextItem;
scripts/sorttable.c
133
static inline unsigned long orc_ip(const int *ip)
scripts/sorttable.c
135
return (unsigned long)ip + *ip;
security/smack/smack_lsm.c
1123
struct inode *ip = d_backing_inode(dentry);
security/smack/smack_lsm.c
1133
rc = smk_curacc(smk_of_inode(ip), MAY_WRITE, &ad);
security/smack/smack_lsm.c
1134
rc = smk_bu_inode(ip, MAY_WRITE, rc);
security/smack/smack_lsm.c
1613
struct inode *ip = inode;
security/smack/smack_lsm.c
1631
sbp = ip->i_sb;
security/smack/smack_lsm.c
1635
sock = SOCKET_I(ip);
security/smack/smack_lsm.c
275
static struct smack_known *smk_fetch(const char *name, struct inode *ip,
security/smack/smack_lsm.c
282
if (!(ip->i_opflags & IOP_XATTR))
security/smack/smack_lsm.c
289
rc = __vfs_getxattr(dp, ip, name, buffer, SMK_LONGLABEL);
security/tomoyo/common.h
1268
return !memcmp(a->ip, b->ip, sizeof(a->ip)) && a->group == b->group &&
security/tomoyo/common.h
535
struct in6_addr ip[2]; /* Big endian. */
security/tomoyo/group.c
202
if (memcmp(&member->address.ip[0], address, size) > 0 ||
security/tomoyo/group.c
203
memcmp(address, &member->address.ip[1], size) > 0)
security/tomoyo/network.c
128
tomoyo_print_ipv6(buf, size, &ptr->ip[0], &ptr->ip[1]);
security/tomoyo/network.c
130
tomoyo_print_ipv4(buf, size, &ptr->ip[0].s6_addr32[0],
security/tomoyo/network.c
131
&ptr->ip[1].s6_addr32[0]);
security/tomoyo/network.c
438
memcmp(&acl->address.ip[0],
security/tomoyo/network.c
441
&acl->address.ip[1], size) <= 0;
security/tomoyo/network.c
53
u8 * const min = ptr->ip[0].in6_u.u6_addr8;
security/tomoyo/network.c
54
u8 * const max = ptr->ip[1].in6_u.u6_addr8;
security/tomoyo/network.c
62
ptr->ip[1].s6_addr32[0] = ptr->ip[0].s6_addr32[0];
sound/core/control.c
1905
int __user *ip = argp;
sound/core/control.c
1914
return put_user(SNDRV_CTL_VERSION, ip) ? -EFAULT : 0;
sound/core/control.c
1936
return snd_ctl_subscribe_events(ctl, ip);
sound/core/control.c
1964
return put_user(SNDRV_CTL_POWER_D0, ip) ? -EFAULT : 0;
sound/soc/renesas/rcar/rsnd.h
417
#define rsnd_mod_get(ip) (&(ip)->mod)
sound/soc/sti/uniperif.h
1000
UNIPERIF_BIT_CONTROL_OFFSET(ip), \
sound/soc/sti/uniperif.h
1001
UNIPERIF_BIT_CONTROL_CLR_UNDERFLOW_DURATION_SHIFT(ip), \
sound/soc/sti/uniperif.h
1002
UNIPERIF_BIT_CONTROL_CLR_UNDERFLOW_DURATION_MASK(ip))
sound/soc/sti/uniperif.h
1003
#define SET_UNIPERIF_BIT_CONTROL_CLR_UNDERFLOW_DURATION(ip) \
sound/soc/sti/uniperif.h
1004
SET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
1005
UNIPERIF_BIT_CONTROL_OFFSET(ip), \
sound/soc/sti/uniperif.h
1006
UNIPERIF_BIT_CONTROL_CLR_UNDERFLOW_DURATION_SHIFT(ip), \
sound/soc/sti/uniperif.h
1007
UNIPERIF_BIT_CONTROL_CLR_UNDERFLOW_DURATION_MASK(ip), 1)
sound/soc/sti/uniperif.h
101
#define UNIPERIF_CHANNEL_STA_REG5_OFFSET(ip) 0x0074
sound/soc/sti/uniperif.h
1010
#define UNIPERIF_BIT_CONTROL_CHL_STS_UPDATE_SHIFT(ip) 1
sound/soc/sti/uniperif.h
1011
#define UNIPERIF_BIT_CONTROL_CHL_STS_UPDATE_MASK(ip) 0x1
sound/soc/sti/uniperif.h
1012
#define GET_UNIPERIF_BIT_CONTROL_CHL_STS_UPDATE(ip) \
sound/soc/sti/uniperif.h
1013
GET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
1014
UNIPERIF_BIT_CONTROL_OFFSET(ip), \
sound/soc/sti/uniperif.h
1015
UNIPERIF_BIT_CONTROL_CHL_STS_UPDATE_SHIFT(ip), \
sound/soc/sti/uniperif.h
1016
UNIPERIF_BIT_CONTROL_CHL_STS_UPDATE_MASK(ip))
sound/soc/sti/uniperif.h
1017
#define SET_UNIPERIF_BIT_CONTROL_CHL_STS_UPDATE(ip) \
sound/soc/sti/uniperif.h
1018
SET_UNIPERIF_BIT_REG(ip, \
sound/soc/sti/uniperif.h
1019
UNIPERIF_BIT_CONTROL_OFFSET(ip), \
sound/soc/sti/uniperif.h
102
#define GET_UNIPERIF_CHANNEL_STA_REG5(ip) \
sound/soc/sti/uniperif.h
1020
UNIPERIF_BIT_CONTROL_CHL_STS_UPDATE_SHIFT(ip), \
sound/soc/sti/uniperif.h
1021
UNIPERIF_BIT_CONTROL_CHL_STS_UPDATE_MASK(ip), 1)
sound/soc/sti/uniperif.h
1027
#define UNIPERIF_STATUS_1_OFFSET(ip) 0x0050
sound/soc/sti/uniperif.h
1028
#define GET_UNIPERIF_STATUS_1(ip) \
sound/soc/sti/uniperif.h
1029
readl_relaxed(ip->base + UNIPERIF_STATUS_1_OFFSET(ip))
sound/soc/sti/uniperif.h
103
readl_relaxed(ip->base + UNIPERIF_CHANNEL_STA_REG5_OFFSET(ip))
sound/soc/sti/uniperif.h
1030
#define SET_UNIPERIF_STATUS_1(ip, value) \
sound/soc/sti/uniperif.h
1031
writel_relaxed(value, ip->base + UNIPERIF_STATUS_1_OFFSET(ip))
sound/soc/sti/uniperif.h
1034
#define UNIPERIF_STATUS_1_UNDERFLOW_DURATION_SHIFT(ip) \
sound/soc/sti/uniperif.h
1035
((ip)->ver < SND_ST_UNIPERIF_VERSION_UNI_PLR_TOP_1_0 ? -1 : 0)
sound/soc/sti/uniperif.h
1036
#define UNIPERIF_STATUS_1_UNDERFLOW_DURATION_MASK(ip) 0xff
sound/soc/sti/uniperif.h
1037
#define GET_UNIPERIF_STATUS_1_UNDERFLOW_DURATION(ip) \
sound/soc/sti/uniperif.h
1038
GET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
1039
UNIPERIF_STATUS_1_OFFSET(ip), \
sound/soc/sti/uniperif.h
104
#define SET_UNIPERIF_CHANNEL_STA_REG5(ip, value) \
sound/soc/sti/uniperif.h
1040
UNIPERIF_STATUS_1_UNDERFLOW_DURATION_SHIFT(ip), \
sound/soc/sti/uniperif.h
1041
UNIPERIF_STATUS_1_UNDERFLOW_DURATION_MASK(ip))
sound/soc/sti/uniperif.h
1042
#define SET_UNIPERIF_STATUS_1_UNDERFLOW_DURATION(ip, value) \
sound/soc/sti/uniperif.h
1043
SET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
1044
UNIPERIF_STATUS_1_OFFSET(ip), \
sound/soc/sti/uniperif.h
1045
UNIPERIF_STATUS_1_UNDERFLOW_DURATION_SHIFT(ip), \
sound/soc/sti/uniperif.h
1046
UNIPERIF_STATUS_1_UNDERFLOW_DURATION_MASK(ip), value)
sound/soc/sti/uniperif.h
105
writel_relaxed(value, ip->base + UNIPERIF_CHANNEL_STA_REG5_OFFSET(ip))
sound/soc/sti/uniperif.h
1052
#define UNIPERIF_CHANNEL_STA_REGN(ip, n) (0x0060 + (4 * n))
sound/soc/sti/uniperif.h
1053
#define GET_UNIPERIF_CHANNEL_STA_REGN(ip) \
sound/soc/sti/uniperif.h
1054
readl_relaxed(ip->base + UNIPERIF_CHANNEL_STA_REGN(ip, n))
sound/soc/sti/uniperif.h
1055
#define SET_UNIPERIF_CHANNEL_STA_REGN(ip, n, value) \
sound/soc/sti/uniperif.h
1056
writel_relaxed(value, ip->base + \
sound/soc/sti/uniperif.h
1057
UNIPERIF_CHANNEL_STA_REGN(ip, n))
sound/soc/sti/uniperif.h
1063
#define UNIPERIF_USER_VALIDITY_OFFSET(ip) 0x0090
sound/soc/sti/uniperif.h
1064
#define GET_UNIPERIF_USER_VALIDITY(ip) \
sound/soc/sti/uniperif.h
1065
readl_relaxed(ip->base + UNIPERIF_USER_VALIDITY_OFFSET(ip))
sound/soc/sti/uniperif.h
1066
#define SET_UNIPERIF_USER_VALIDITY(ip, value) \
sound/soc/sti/uniperif.h
1067
writel_relaxed(value, ip->base + UNIPERIF_USER_VALIDITY_OFFSET(ip))
sound/soc/sti/uniperif.h
1070
#define UNIPERIF_USER_VALIDITY_VALIDITY_LR_SHIFT(ip) 0
sound/soc/sti/uniperif.h
1071
#define UNIPERIF_USER_VALIDITY_VALIDITY_LR_MASK(ip) 0x3
sound/soc/sti/uniperif.h
1072
#define GET_UNIPERIF_USER_VALIDITY_VALIDITY_LR(ip) \
sound/soc/sti/uniperif.h
1073
GET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
1074
UNIPERIF_USER_VALIDITY_OFFSET(ip), \
sound/soc/sti/uniperif.h
1075
UNIPERIF_USER_VALIDITY_VALIDITY_LR_SHIFT(ip), \
sound/soc/sti/uniperif.h
1076
UNIPERIF_USER_VALIDITY_VALIDITY_LR_MASK(ip))
sound/soc/sti/uniperif.h
1077
#define SET_UNIPERIF_USER_VALIDITY_VALIDITY_LR(ip, value) \
sound/soc/sti/uniperif.h
1078
SET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
1079
UNIPERIF_USER_VALIDITY_OFFSET(ip), \
sound/soc/sti/uniperif.h
1080
UNIPERIF_USER_VALIDITY_VALIDITY_LR_SHIFT(ip), \
sound/soc/sti/uniperif.h
1081
UNIPERIF_USER_VALIDITY_VALIDITY_LR_MASK(ip), \
sound/soc/sti/uniperif.h
1087
#define UNIPERIF_DBG_STANDBY_LEFT_SP_OFFSET(ip) 0x0150
sound/soc/sti/uniperif.h
1088
#define UNIPERIF_DBG_STANDBY_LEFT_SP_SHIFT(ip) \
sound/soc/sti/uniperif.h
1089
((ip)->ver < SND_ST_UNIPERIF_VERSION_UNI_PLR_TOP_1_0 ? -1 : 0)
sound/soc/sti/uniperif.h
1090
#define UNIPERIF_DBG_STANDBY_LEFT_SP_MASK(ip) \
sound/soc/sti/uniperif.h
1091
((ip)->ver < SND_ST_UNIPERIF_VERSION_UNI_PLR_TOP_1_0 ? 0 : 0xFFFFFF)
sound/soc/sti/uniperif.h
1092
#define GET_UNIPERIF_DBG_STANDBY_LEFT_SP(ip) \
sound/soc/sti/uniperif.h
1093
GET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
1094
UNIPERIF_DBG_STANDBY_LEFT_SP_OFFSET(ip), \
sound/soc/sti/uniperif.h
1095
UNIPERIF_DBG_STANDBY_LEFT_SP_SHIFT(ip), \
sound/soc/sti/uniperif.h
1096
UNIPERIF_DBG_STANDBY_LEFT_SP_MASK(ip))
sound/soc/sti/uniperif.h
1097
#define SET_UNIPERIF_DBG_STANDBY_LEFT_SP(ip, value) \
sound/soc/sti/uniperif.h
1098
SET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
1099
UNIPERIF_DBG_STANDBY_LEFT_SP_OFFSET(ip), \
sound/soc/sti/uniperif.h
1100
UNIPERIF_DBG_STANDBY_LEFT_SP_SHIFT(ip), \
sound/soc/sti/uniperif.h
1101
UNIPERIF_DBG_STANDBY_LEFT_SP_MASK(ip), value)
sound/soc/sti/uniperif.h
1106
#define UNIPERIF_TDM_ENABLE_OFFSET(ip) 0x0118
sound/soc/sti/uniperif.h
1107
#define GET_UNIPERIF_TDM_ENABLE(ip) \
sound/soc/sti/uniperif.h
1108
readl_relaxed(ip->base + UNIPERIF_TDM_ENABLE_OFFSET(ip))
sound/soc/sti/uniperif.h
1109
#define SET_UNIPERIF_TDM_ENABLE(ip, value) \
sound/soc/sti/uniperif.h
111
#define UNIPERIF_ITS_OFFSET(ip) 0x000C
sound/soc/sti/uniperif.h
1110
writel_relaxed(value, ip->base + UNIPERIF_TDM_ENABLE_OFFSET(ip))
sound/soc/sti/uniperif.h
1113
#define UNIPERIF_TDM_ENABLE_EN_TDM_SHIFT(ip) 0x0
sound/soc/sti/uniperif.h
1114
#define UNIPERIF_TDM_ENABLE_EN_TDM_MASK(ip) 0x1
sound/soc/sti/uniperif.h
1115
#define GET_UNIPERIF_TDM_ENABLE_EN_TDM(ip) \
sound/soc/sti/uniperif.h
1116
GET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
1117
UNIPERIF_TDM_ENABLE_OFFSET(ip), \
sound/soc/sti/uniperif.h
1118
UNIPERIF_TDM_ENABLE_EN_TDM_SHIFT(ip), \
sound/soc/sti/uniperif.h
1119
UNIPERIF_TDM_ENABLE_EN_TDM_MASK(ip))
sound/soc/sti/uniperif.h
112
#define GET_UNIPERIF_ITS(ip) \
sound/soc/sti/uniperif.h
1120
#define SET_UNIPERIF_TDM_ENABLE_TDM_ENABLE(ip) \
sound/soc/sti/uniperif.h
1121
SET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
1122
UNIPERIF_TDM_ENABLE_OFFSET(ip), \
sound/soc/sti/uniperif.h
1123
UNIPERIF_TDM_ENABLE_EN_TDM_SHIFT(ip), \
sound/soc/sti/uniperif.h
1124
UNIPERIF_TDM_ENABLE_EN_TDM_MASK(ip), 1)
sound/soc/sti/uniperif.h
1125
#define SET_UNIPERIF_TDM_ENABLE_TDM_DISABLE(ip) \
sound/soc/sti/uniperif.h
1126
SET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
1127
UNIPERIF_TDM_ENABLE_OFFSET(ip), \
sound/soc/sti/uniperif.h
1128
UNIPERIF_TDM_ENABLE_EN_TDM_SHIFT(ip), \
sound/soc/sti/uniperif.h
1129
UNIPERIF_TDM_ENABLE_EN_TDM_MASK(ip), 0)
sound/soc/sti/uniperif.h
113
readl_relaxed(ip->base + UNIPERIF_ITS_OFFSET(ip))
sound/soc/sti/uniperif.h
1134
#define UNIPERIF_TDM_FS_REF_FREQ_OFFSET(ip) 0x011c
sound/soc/sti/uniperif.h
1135
#define GET_UNIPERIF_TDM_FS_REF_FREQ(ip) \
sound/soc/sti/uniperif.h
1136
readl_relaxed(ip->base + UNIPERIF_TDM_FS_REF_FREQ_OFFSET(ip))
sound/soc/sti/uniperif.h
1137
#define SET_UNIPERIF_TDM_FS_REF_FREQ(ip, value) \
sound/soc/sti/uniperif.h
1138
writel_relaxed(value, ip->base + \
sound/soc/sti/uniperif.h
1139
UNIPERIF_TDM_FS_REF_FREQ_OFFSET(ip))
sound/soc/sti/uniperif.h
1142
#define UNIPERIF_TDM_FS_REF_FREQ_REF_FREQ_SHIFT(ip) 0x0
sound/soc/sti/uniperif.h
1143
#define VALUE_UNIPERIF_TDM_FS_REF_FREQ_8KHZ(ip) 0
sound/soc/sti/uniperif.h
1144
#define VALUE_UNIPERIF_TDM_FS_REF_FREQ_16KHZ(ip) 1
sound/soc/sti/uniperif.h
1145
#define VALUE_UNIPERIF_TDM_FS_REF_FREQ_32KHZ(ip) 2
sound/soc/sti/uniperif.h
1146
#define VALUE_UNIPERIF_TDM_FS_REF_FREQ_48KHZ(ip) 3
sound/soc/sti/uniperif.h
1147
#define UNIPERIF_TDM_FS_REF_FREQ_REF_FREQ_MASK(ip) 0x3
sound/soc/sti/uniperif.h
1148
#define GET_UNIPERIF_TDM_FS_REF_FREQ_REF_FREQ(ip) \
sound/soc/sti/uniperif.h
1149
GET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
1150
UNIPERIF_TDM_FS_REF_FREQ_OFFSET(ip), \
sound/soc/sti/uniperif.h
1151
UNIPERIF_TDM_FS_REF_FREQ_REF_FREQ_SHIFT(ip), \
sound/soc/sti/uniperif.h
1152
UNIPERIF_TDM_FS_REF_FREQ_REF_FREQ_MASK(ip))
sound/soc/sti/uniperif.h
1153
#define SET_UNIPERIF_TDM_FS_REF_FREQ_8KHZ(ip) \
sound/soc/sti/uniperif.h
1154
SET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
1155
UNIPERIF_TDM_FS_REF_FREQ_OFFSET(ip), \
sound/soc/sti/uniperif.h
1156
UNIPERIF_TDM_FS_REF_FREQ_REF_FREQ_SHIFT(ip), \
sound/soc/sti/uniperif.h
1157
UNIPERIF_TDM_FS_REF_FREQ_REF_FREQ_MASK(ip), \
sound/soc/sti/uniperif.h
1158
VALUE_UNIPERIF_TDM_FS_REF_FREQ_8KHZ(ip))
sound/soc/sti/uniperif.h
1159
#define SET_UNIPERIF_TDM_FS_REF_FREQ_16KHZ(ip) \
sound/soc/sti/uniperif.h
116
#define UNIPERIF_ITS_MEM_BLK_READ_SHIFT(ip) 5
sound/soc/sti/uniperif.h
1160
SET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
1161
UNIPERIF_TDM_FS_REF_FREQ_OFFSET(ip), \
sound/soc/sti/uniperif.h
1162
UNIPERIF_TDM_FS_REF_FREQ_REF_FREQ_SHIFT(ip), \
sound/soc/sti/uniperif.h
1163
UNIPERIF_TDM_FS_REF_FREQ_REF_FREQ_MASK(ip), \
sound/soc/sti/uniperif.h
1164
VALUE_UNIPERIF_TDM_FS_REF_FREQ_16KHZ(ip))
sound/soc/sti/uniperif.h
1165
#define SET_UNIPERIF_TDM_FS_REF_FREQ_32KHZ(ip) \
sound/soc/sti/uniperif.h
1166
SET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
1167
UNIPERIF_TDM_FS_REF_FREQ_OFFSET(ip), \
sound/soc/sti/uniperif.h
1168
UNIPERIF_TDM_FS_REF_FREQ_REF_FREQ_SHIFT(ip), \
sound/soc/sti/uniperif.h
1169
UNIPERIF_TDM_FS_REF_FREQ_REF_FREQ_MASK(ip), \
sound/soc/sti/uniperif.h
117
#define UNIPERIF_ITS_MEM_BLK_READ_MASK(ip) \
sound/soc/sti/uniperif.h
1170
VALUE_UNIPERIF_TDM_FS_REF_FREQ_32KHZ(ip))
sound/soc/sti/uniperif.h
1171
#define SET_UNIPERIF_TDM_FS_REF_FREQ_48KHZ(ip) \
sound/soc/sti/uniperif.h
1172
SET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
1173
UNIPERIF_TDM_FS_REF_FREQ_OFFSET(ip), \
sound/soc/sti/uniperif.h
1174
UNIPERIF_TDM_FS_REF_FREQ_REF_FREQ_SHIFT(ip), \
sound/soc/sti/uniperif.h
1175
UNIPERIF_TDM_FS_REF_FREQ_REF_FREQ_MASK(ip), \
sound/soc/sti/uniperif.h
1176
VALUE_UNIPERIF_TDM_FS_REF_FREQ_48KHZ(ip))
sound/soc/sti/uniperif.h
118
(BIT(UNIPERIF_ITS_MEM_BLK_READ_SHIFT(ip)))
sound/soc/sti/uniperif.h
1181
#define UNIPERIF_TDM_FS_REF_DIV_OFFSET(ip) 0x0120
sound/soc/sti/uniperif.h
1182
#define GET_UNIPERIF_TDM_FS_REF_DIV(ip) \
sound/soc/sti/uniperif.h
1183
readl_relaxed(ip->base + UNIPERIF_TDM_FS_REF_DIV_OFFSET(ip))
sound/soc/sti/uniperif.h
1184
#define SET_UNIPERIF_TDM_FS_REF_DIV(ip, value) \
sound/soc/sti/uniperif.h
1185
writel_relaxed(value, ip->base + \
sound/soc/sti/uniperif.h
1186
UNIPERIF_TDM_FS_REF_DIV_OFFSET(ip))
sound/soc/sti/uniperif.h
1189
#define UNIPERIF_TDM_FS_REF_DIV_NUM_TIMESLOT_SHIFT(ip) 0x0
sound/soc/sti/uniperif.h
1190
#define UNIPERIF_TDM_FS_REF_DIV_NUM_TIMESLOT_MASK(ip) 0xff
sound/soc/sti/uniperif.h
1191
#define GET_UNIPERIF_TDM_FS_REF_DIV_NUM_TIMESLOT(ip) \
sound/soc/sti/uniperif.h
1192
GET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
1193
UNIPERIF_TDM_FS_REF_DIV_OFFSET(ip), \
sound/soc/sti/uniperif.h
1194
UNIPERIF_TDM_FS_REF_DIV_NUM_TIMESLOT_SHIFT(ip), \
sound/soc/sti/uniperif.h
1195
UNIPERIF_TDM_FS_REF_DIV_NUM_TIMESLOT_MASK(ip))
sound/soc/sti/uniperif.h
1196
#define SET_UNIPERIF_TDM_FS_REF_DIV_NUM_TIMESLOT(ip, value) \
sound/soc/sti/uniperif.h
1197
SET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
1198
UNIPERIF_TDM_FS_REF_DIV_OFFSET(ip), \
sound/soc/sti/uniperif.h
1199
UNIPERIF_TDM_FS_REF_DIV_NUM_TIMESLOT_SHIFT(ip), \
sound/soc/sti/uniperif.h
1200
UNIPERIF_TDM_FS_REF_DIV_NUM_TIMESLOT_MASK(ip), value)
sound/soc/sti/uniperif.h
1206
#define UNIPERIF_TDM_WORD_POS_1_2_OFFSET(ip) 0x013c
sound/soc/sti/uniperif.h
1207
#define UNIPERIF_TDM_WORD_POS_3_4_OFFSET(ip) 0x0140
sound/soc/sti/uniperif.h
1208
#define UNIPERIF_TDM_WORD_POS_5_6_OFFSET(ip) 0x0144
sound/soc/sti/uniperif.h
1209
#define UNIPERIF_TDM_WORD_POS_7_8_OFFSET(ip) 0x0148
sound/soc/sti/uniperif.h
121
#define UNIPERIF_ITS_FIFO_ERROR_SHIFT(ip) \
sound/soc/sti/uniperif.h
1210
#define GET_UNIPERIF_TDM_WORD_POS(ip, words) \
sound/soc/sti/uniperif.h
1211
readl_relaxed(ip->base + UNIPERIF_TDM_WORD_POS_##words##_OFFSET(ip))
sound/soc/sti/uniperif.h
1212
#define SET_UNIPERIF_TDM_WORD_POS(ip, words, value) \
sound/soc/sti/uniperif.h
1213
writel_relaxed(value, ip->base + \
sound/soc/sti/uniperif.h
1214
UNIPERIF_TDM_WORD_POS_##words##_OFFSET(ip))
sound/soc/sti/uniperif.h
122
((ip)->ver < SND_ST_UNIPERIF_VERSION_UNI_PLR_TOP_1_0 ? 0 : 8)
sound/soc/sti/uniperif.h
123
#define UNIPERIF_ITS_FIFO_ERROR_MASK(ip) \
sound/soc/sti/uniperif.h
124
(BIT(UNIPERIF_ITS_FIFO_ERROR_SHIFT(ip)))
sound/soc/sti/uniperif.h
127
#define UNIPERIF_ITS_DMA_ERROR_SHIFT(ip) 9
sound/soc/sti/uniperif.h
128
#define UNIPERIF_ITS_DMA_ERROR_MASK(ip) \
sound/soc/sti/uniperif.h
129
(BIT(UNIPERIF_ITS_DMA_ERROR_SHIFT(ip)))
sound/soc/sti/uniperif.h
132
#define UNIPERIF_ITS_UNDERFLOW_REC_DONE_SHIFT(ip) \
sound/soc/sti/uniperif.h
133
((ip)->ver < SND_ST_UNIPERIF_VERSION_UNI_PLR_TOP_1_0 ? -1 : 12)
sound/soc/sti/uniperif.h
134
#define UNIPERIF_ITS_UNDERFLOW_REC_DONE_MASK(ip) \
sound/soc/sti/uniperif.h
135
((ip)->ver < SND_ST_UNIPERIF_VERSION_UNI_PLR_TOP_1_0 ? \
sound/soc/sti/uniperif.h
136
0 : (BIT(UNIPERIF_ITS_UNDERFLOW_REC_DONE_SHIFT(ip))))
sound/soc/sti/uniperif.h
139
#define UNIPERIF_ITS_UNDERFLOW_REC_FAILED_SHIFT(ip) \
sound/soc/sti/uniperif.h
140
((ip)->ver < SND_ST_UNIPERIF_VERSION_UNI_PLR_TOP_1_0 ? -1 : 13)
sound/soc/sti/uniperif.h
141
#define UNIPERIF_ITS_UNDERFLOW_REC_FAILED_MASK(ip) \
sound/soc/sti/uniperif.h
142
((ip)->ver < SND_ST_UNIPERIF_VERSION_UNI_PLR_TOP_1_0 ? \
sound/soc/sti/uniperif.h
143
0 : (BIT(UNIPERIF_ITS_UNDERFLOW_REC_FAILED_SHIFT(ip))))
sound/soc/sti/uniperif.h
150
#define UNIPERIF_ITS_BCLR_FIFO_ERROR_SHIFT(ip) \
sound/soc/sti/uniperif.h
151
((ip)->ver < SND_ST_UNIPERIF_VERSION_UNI_PLR_TOP_1_0 ? 0 : 8)
sound/soc/sti/uniperif.h
152
#define UNIPERIF_ITS_BCLR_FIFO_ERROR_MASK(ip) \
sound/soc/sti/uniperif.h
153
(BIT(UNIPERIF_ITS_BCLR_FIFO_ERROR_SHIFT(ip)))
sound/soc/sti/uniperif.h
154
#define SET_UNIPERIF_ITS_BCLR_FIFO_ERROR(ip) \
sound/soc/sti/uniperif.h
155
SET_UNIPERIF_ITS_BCLR(ip, \
sound/soc/sti/uniperif.h
156
UNIPERIF_ITS_BCLR_FIFO_ERROR_MASK(ip))
sound/soc/sti/uniperif.h
158
#define UNIPERIF_ITS_BCLR_OFFSET(ip) 0x0010
sound/soc/sti/uniperif.h
159
#define SET_UNIPERIF_ITS_BCLR(ip, value) \
sound/soc/sti/uniperif.h
160
writel_relaxed(value, ip->base + UNIPERIF_ITS_BCLR_OFFSET(ip))
sound/soc/sti/uniperif.h
166
#define UNIPERIF_ITM_OFFSET(ip) 0x0018
sound/soc/sti/uniperif.h
167
#define GET_UNIPERIF_ITM(ip) \
sound/soc/sti/uniperif.h
168
readl_relaxed(ip->base + UNIPERIF_ITM_OFFSET(ip))
sound/soc/sti/uniperif.h
171
#define UNIPERIF_ITM_FIFO_ERROR_SHIFT(ip) \
sound/soc/sti/uniperif.h
172
((ip)->ver < SND_ST_UNIPERIF_VERSION_UNI_PLR_TOP_1_0 ? 0 : 8)
sound/soc/sti/uniperif.h
173
#define UNIPERIF_ITM_FIFO_ERROR_MASK(ip) \
sound/soc/sti/uniperif.h
174
(BIT(UNIPERIF_ITM_FIFO_ERROR_SHIFT(ip)))
sound/soc/sti/uniperif.h
177
#define UNIPERIF_ITM_UNDERFLOW_REC_DONE_SHIFT(ip) \
sound/soc/sti/uniperif.h
178
((ip)->ver < SND_ST_UNIPERIF_VERSION_UNI_PLR_TOP_1_0 ? -1 : 12)
sound/soc/sti/uniperif.h
179
#define UNIPERIF_ITM_UNDERFLOW_REC_DONE_MASK(ip) \
sound/soc/sti/uniperif.h
180
((ip)->ver < SND_ST_UNIPERIF_VERSION_UNI_PLR_TOP_1_0 ? \
sound/soc/sti/uniperif.h
181
0 : (BIT(UNIPERIF_ITM_UNDERFLOW_REC_DONE_SHIFT(ip))))
sound/soc/sti/uniperif.h
184
#define UNIPERIF_ITM_UNDERFLOW_REC_FAILED_SHIFT(ip) \
sound/soc/sti/uniperif.h
185
((ip)->ver < SND_ST_UNIPERIF_VERSION_UNI_PLR_TOP_1_0 ? -1 : 13)
sound/soc/sti/uniperif.h
186
#define UNIPERIF_ITM_UNDERFLOW_REC_FAILED_MASK(ip) \
sound/soc/sti/uniperif.h
187
((ip)->ver < SND_ST_UNIPERIF_VERSION_UNI_PLR_TOP_1_0 ? \
sound/soc/sti/uniperif.h
188
0 : (BIT(UNIPERIF_ITM_UNDERFLOW_REC_FAILED_SHIFT(ip))))
sound/soc/sti/uniperif.h
19
#define GET_UNIPERIF_REG(ip, offset, shift, mask) \
sound/soc/sti/uniperif.h
194
#define UNIPERIF_ITM_BCLR_OFFSET(ip) 0x001c
sound/soc/sti/uniperif.h
195
#define SET_UNIPERIF_ITM_BCLR(ip, value) \
sound/soc/sti/uniperif.h
196
writel_relaxed(value, ip->base + UNIPERIF_ITM_BCLR_OFFSET(ip))
sound/soc/sti/uniperif.h
199
#define UNIPERIF_ITM_BCLR_FIFO_ERROR_SHIFT(ip) \
sound/soc/sti/uniperif.h
20
((readl_relaxed(ip->base + offset) >> shift) & mask)
sound/soc/sti/uniperif.h
200
((ip)->ver < SND_ST_UNIPERIF_VERSION_UNI_PLR_TOP_1_0 ? 0 : 8)
sound/soc/sti/uniperif.h
201
#define UNIPERIF_ITM_BCLR_FIFO_ERROR_MASK(ip) \
sound/soc/sti/uniperif.h
202
(BIT(UNIPERIF_ITM_BCLR_FIFO_ERROR_SHIFT(ip)))
sound/soc/sti/uniperif.h
203
#define SET_UNIPERIF_ITM_BCLR_FIFO_ERROR(ip) \
sound/soc/sti/uniperif.h
204
SET_UNIPERIF_ITM_BCLR(ip, \
sound/soc/sti/uniperif.h
205
UNIPERIF_ITM_BCLR_FIFO_ERROR_MASK(ip))
sound/soc/sti/uniperif.h
208
#define UNIPERIF_ITM_BCLR_DMA_ERROR_SHIFT(ip) 9
sound/soc/sti/uniperif.h
209
#define UNIPERIF_ITM_BCLR_DMA_ERROR_MASK(ip) \
sound/soc/sti/uniperif.h
21
#define SET_UNIPERIF_REG(ip, offset, shift, mask, value) \
sound/soc/sti/uniperif.h
210
(BIT(UNIPERIF_ITM_BCLR_DMA_ERROR_SHIFT(ip)))
sound/soc/sti/uniperif.h
211
#define SET_UNIPERIF_ITM_BCLR_DMA_ERROR(ip) \
sound/soc/sti/uniperif.h
212
SET_UNIPERIF_ITM_BCLR(ip, \
sound/soc/sti/uniperif.h
213
UNIPERIF_ITM_BCLR_DMA_ERROR_MASK(ip))
sound/soc/sti/uniperif.h
219
#define UNIPERIF_ITM_BSET_OFFSET(ip) 0x0020
sound/soc/sti/uniperif.h
22
writel_relaxed(((readl_relaxed(ip->base + offset) & \
sound/soc/sti/uniperif.h
220
#define SET_UNIPERIF_ITM_BSET(ip, value) \
sound/soc/sti/uniperif.h
221
writel_relaxed(value, ip->base + UNIPERIF_ITM_BSET_OFFSET(ip))
sound/soc/sti/uniperif.h
224
#define UNIPERIF_ITM_BSET_FIFO_ERROR_SHIFT(ip) \
sound/soc/sti/uniperif.h
225
((ip)->ver < SND_ST_UNIPERIF_VERSION_UNI_PLR_TOP_1_0 ? 0 : 8)
sound/soc/sti/uniperif.h
226
#define UNIPERIF_ITM_BSET_FIFO_ERROR_MASK(ip) \
sound/soc/sti/uniperif.h
227
(BIT(UNIPERIF_ITM_BSET_FIFO_ERROR_SHIFT(ip)))
sound/soc/sti/uniperif.h
228
#define SET_UNIPERIF_ITM_BSET_FIFO_ERROR(ip) \
sound/soc/sti/uniperif.h
229
SET_UNIPERIF_ITM_BSET(ip, \
sound/soc/sti/uniperif.h
23
~(mask << shift)) | (((value) & mask) << shift)), ip->base + offset)
sound/soc/sti/uniperif.h
230
UNIPERIF_ITM_BSET_FIFO_ERROR_MASK(ip))
sound/soc/sti/uniperif.h
233
#define UNIPERIF_ITM_BSET_MEM_BLK_READ_SHIFT(ip) 5
sound/soc/sti/uniperif.h
234
#define UNIPERIF_ITM_BSET_MEM_BLK_READ_MASK(ip) \
sound/soc/sti/uniperif.h
235
(BIT(UNIPERIF_ITM_BSET_MEM_BLK_READ_SHIFT(ip)))
sound/soc/sti/uniperif.h
236
#define SET_UNIPERIF_ITM_BSET_MEM_BLK_READ(ip) \
sound/soc/sti/uniperif.h
237
SET_UNIPERIF_ITM_BSET(ip, \
sound/soc/sti/uniperif.h
238
UNIPERIF_ITM_BSET_MEM_BLK_READ_MASK(ip))
sound/soc/sti/uniperif.h
24
#define SET_UNIPERIF_BIT_REG(ip, offset, shift, mask, value) \
sound/soc/sti/uniperif.h
241
#define UNIPERIF_ITM_BSET_DMA_ERROR_SHIFT(ip) 9
sound/soc/sti/uniperif.h
242
#define UNIPERIF_ITM_BSET_DMA_ERROR_MASK(ip) \
sound/soc/sti/uniperif.h
243
(BIT(UNIPERIF_ITM_BSET_DMA_ERROR_SHIFT(ip)))
sound/soc/sti/uniperif.h
244
#define SET_UNIPERIF_ITM_BSET_DMA_ERROR(ip) \
sound/soc/sti/uniperif.h
245
SET_UNIPERIF_ITM_BSET(ip, \
sound/soc/sti/uniperif.h
246
UNIPERIF_ITM_BSET_DMA_ERROR_MASK(ip))
sound/soc/sti/uniperif.h
249
#define UNIPERIF_ITM_BSET_UNDERFLOW_REC_DONE_SHIFT(ip) \
sound/soc/sti/uniperif.h
25
writel_relaxed((((value) & mask) << shift), ip->base + offset)
sound/soc/sti/uniperif.h
250
((ip)->ver < SND_ST_UNIPERIF_VERSION_UNI_PLR_TOP_1_0 ? -1 : 12)
sound/soc/sti/uniperif.h
251
#define UNIPERIF_ITM_BSET_UNDERFLOW_REC_DONE_MASK(ip) \
sound/soc/sti/uniperif.h
252
((ip)->ver < SND_ST_UNIPERIF_VERSION_UNI_PLR_TOP_1_0 ? \
sound/soc/sti/uniperif.h
253
0 : (BIT(UNIPERIF_ITM_BSET_UNDERFLOW_REC_DONE_SHIFT(ip))))
sound/soc/sti/uniperif.h
254
#define SET_UNIPERIF_ITM_BSET_UNDERFLOW_REC_DONE(ip) \
sound/soc/sti/uniperif.h
255
SET_UNIPERIF_ITM_BSET(ip, \
sound/soc/sti/uniperif.h
256
UNIPERIF_ITM_BSET_UNDERFLOW_REC_DONE_MASK(ip))
sound/soc/sti/uniperif.h
259
#define UNIPERIF_ITM_BSET_UNDERFLOW_REC_FAILED_SHIFT(ip) \
sound/soc/sti/uniperif.h
260
((ip)->ver < SND_ST_UNIPERIF_VERSION_UNI_PLR_TOP_1_0 ? -1 : 13)
sound/soc/sti/uniperif.h
261
#define UNIPERIF_ITM_BSET_UNDERFLOW_REC_FAILED_MASK(ip) \
sound/soc/sti/uniperif.h
262
((ip)->ver < SND_ST_UNIPERIF_VERSION_UNI_PLR_TOP_1_0 ? \
sound/soc/sti/uniperif.h
263
0 : (BIT(UNIPERIF_ITM_BSET_UNDERFLOW_REC_FAILED_SHIFT(ip))))
sound/soc/sti/uniperif.h
264
#define SET_UNIPERIF_ITM_BSET_UNDERFLOW_REC_FAILED(ip) \
sound/soc/sti/uniperif.h
265
SET_UNIPERIF_ITM_BSET(ip, \
sound/soc/sti/uniperif.h
266
UNIPERIF_ITM_BSET_UNDERFLOW_REC_FAILED_MASK(ip))
sound/soc/sti/uniperif.h
272
#define UNIPERIF_CONFIG_OFFSET(ip) 0x0040
sound/soc/sti/uniperif.h
273
#define GET_UNIPERIF_CONFIG(ip) \
sound/soc/sti/uniperif.h
274
readl_relaxed(ip->base + UNIPERIF_CONFIG_OFFSET(ip))
sound/soc/sti/uniperif.h
275
#define SET_UNIPERIF_CONFIG(ip, value) \
sound/soc/sti/uniperif.h
276
writel_relaxed(value, ip->base + UNIPERIF_CONFIG_OFFSET(ip))
sound/soc/sti/uniperif.h
279
#define UNIPERIF_CONFIG_PARITY_CNTR_SHIFT(ip) 0
sound/soc/sti/uniperif.h
280
#define UNIPERIF_CONFIG_PARITY_CNTR_MASK(ip) 0x1
sound/soc/sti/uniperif.h
281
#define GET_UNIPERIF_CONFIG_PARITY_CNTR(ip) \
sound/soc/sti/uniperif.h
282
GET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
283
UNIPERIF_CONFIG_OFFSET(ip), \
sound/soc/sti/uniperif.h
284
UNIPERIF_CONFIG_PARITY_CNTR_SHIFT(ip), \
sound/soc/sti/uniperif.h
285
UNIPERIF_CONFIG_PARITY_CNTR_MASK(ip))
sound/soc/sti/uniperif.h
286
#define SET_UNIPERIF_CONFIG_PARITY_CNTR_BY_HW(ip) \
sound/soc/sti/uniperif.h
287
SET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
288
UNIPERIF_CONFIG_OFFSET(ip), \
sound/soc/sti/uniperif.h
289
UNIPERIF_CONFIG_PARITY_CNTR_SHIFT(ip), \
sound/soc/sti/uniperif.h
290
UNIPERIF_CONFIG_PARITY_CNTR_MASK(ip), 0)
sound/soc/sti/uniperif.h
291
#define SET_UNIPERIF_CONFIG_PARITY_CNTR_BY_SW(ip) \
sound/soc/sti/uniperif.h
292
SET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
293
UNIPERIF_CONFIG_OFFSET(ip), \
sound/soc/sti/uniperif.h
294
UNIPERIF_CONFIG_PARITY_CNTR_SHIFT(ip), \
sound/soc/sti/uniperif.h
295
UNIPERIF_CONFIG_PARITY_CNTR_MASK(ip), 1)
sound/soc/sti/uniperif.h
298
#define UNIPERIF_CONFIG_CHANNEL_STA_CNTR_SHIFT(ip) 1
sound/soc/sti/uniperif.h
299
#define UNIPERIF_CONFIG_CHANNEL_STA_CNTR_MASK(ip) 0x1
sound/soc/sti/uniperif.h
300
#define GET_UNIPERIF_CONFIG_CHANNEL_STA_CNTR(ip) \
sound/soc/sti/uniperif.h
301
GET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
302
UNIPERIF_CONFIG_OFFSET(ip), \
sound/soc/sti/uniperif.h
303
UNIPERIF_CONFIG_CHANNEL_STA_CNTR_SHIFT(ip), \
sound/soc/sti/uniperif.h
304
UNIPERIF_CONFIG_CHANNEL_STA_CNTR_MASK(ip))
sound/soc/sti/uniperif.h
305
#define SET_UNIPERIF_CONFIG_CHANNEL_STA_CNTR_BY_SW(ip) \
sound/soc/sti/uniperif.h
306
SET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
307
UNIPERIF_CONFIG_OFFSET(ip), \
sound/soc/sti/uniperif.h
308
UNIPERIF_CONFIG_CHANNEL_STA_CNTR_SHIFT(ip), \
sound/soc/sti/uniperif.h
309
UNIPERIF_CONFIG_CHANNEL_STA_CNTR_MASK(ip), 0)
sound/soc/sti/uniperif.h
31
#define UNIPERIF_SOFT_RST_OFFSET(ip) 0x0000
sound/soc/sti/uniperif.h
310
#define SET_UNIPERIF_CONFIG_CHANNEL_STA_CNTR_BY_HW(ip) \
sound/soc/sti/uniperif.h
311
SET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
312
UNIPERIF_CONFIG_OFFSET(ip), \
sound/soc/sti/uniperif.h
313
UNIPERIF_CONFIG_CHANNEL_STA_CNTR_SHIFT(ip), \
sound/soc/sti/uniperif.h
314
UNIPERIF_CONFIG_CHANNEL_STA_CNTR_MASK(ip), 1)
sound/soc/sti/uniperif.h
317
#define UNIPERIF_CONFIG_USER_DAT_CNTR_SHIFT(ip) 2
sound/soc/sti/uniperif.h
318
#define UNIPERIF_CONFIG_USER_DAT_CNTR_MASK(ip) 0x1
sound/soc/sti/uniperif.h
319
#define GET_UNIPERIF_CONFIG_USER_DAT_CNTR(ip) \
sound/soc/sti/uniperif.h
32
#define GET_UNIPERIF_SOFT_RST(ip) \
sound/soc/sti/uniperif.h
320
GET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
321
UNIPERIF_CONFIG_OFFSET(ip), \
sound/soc/sti/uniperif.h
322
UNIPERIF_CONFIG_USER_DAT_CNTR_SHIFT(ip), \
sound/soc/sti/uniperif.h
323
UNIPERIF_CONFIG_USER_DAT_CNTR_MASK(ip))
sound/soc/sti/uniperif.h
324
#define SET_UNIPERIF_CONFIG_USER_DAT_CNTR_BY_HW(ip) \
sound/soc/sti/uniperif.h
325
SET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
326
UNIPERIF_CONFIG_OFFSET(ip), \
sound/soc/sti/uniperif.h
327
UNIPERIF_CONFIG_USER_DAT_CNTR_SHIFT(ip), \
sound/soc/sti/uniperif.h
328
UNIPERIF_CONFIG_USER_DAT_CNTR_MASK(ip), 1)
sound/soc/sti/uniperif.h
329
#define SET_UNIPERIF_CONFIG_USER_DAT_CNTR_BY_SW(ip) \
sound/soc/sti/uniperif.h
33
((ip)->ver < SND_ST_UNIPERIF_VERSION_UNI_PLR_TOP_1_0 ? \
sound/soc/sti/uniperif.h
330
SET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
331
UNIPERIF_CONFIG_OFFSET(ip), \
sound/soc/sti/uniperif.h
332
UNIPERIF_CONFIG_USER_DAT_CNTR_SHIFT(ip), \
sound/soc/sti/uniperif.h
333
UNIPERIF_CONFIG_USER_DAT_CNTR_MASK(ip), 0)
sound/soc/sti/uniperif.h
336
#define UNIPERIF_CONFIG_VALIDITY_DAT_CNTR_SHIFT(ip) 3
sound/soc/sti/uniperif.h
337
#define UNIPERIF_CONFIG_VALIDITY_DAT_CNTR_MASK(ip) 0x1
sound/soc/sti/uniperif.h
338
#define GET_UNIPERIF_CONFIG_VALIDITY_DAT_CNTR(ip) \
sound/soc/sti/uniperif.h
339
GET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
34
readl_relaxed(ip->base + UNIPERIF_SOFT_RST_OFFSET(ip)) : 0)
sound/soc/sti/uniperif.h
340
UNIPERIF_CONFIG_OFFSET(ip), \
sound/soc/sti/uniperif.h
341
UNIPERIF_CONFIG_VALIDITY_DAT_CNTR_SHIFT(ip), \
sound/soc/sti/uniperif.h
342
UNIPERIF_CONFIG_VALIDITY_DAT_CNTR_MASK(ip))
sound/soc/sti/uniperif.h
343
#define SET_UNIPERIF_CONFIG_VALIDITY_DAT_CNTR_BY_SW(ip) \
sound/soc/sti/uniperif.h
344
SET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
345
UNIPERIF_CONFIG_OFFSET(ip), \
sound/soc/sti/uniperif.h
346
UNIPERIF_CONFIG_VALIDITY_DAT_CNTR_SHIFT(ip), \
sound/soc/sti/uniperif.h
347
UNIPERIF_CONFIG_VALIDITY_DAT_CNTR_MASK(ip), 0)
sound/soc/sti/uniperif.h
348
#define SET_UNIPERIF_CONFIG_VALIDITY_DAT_CNTR_BY_HW(ip) \
sound/soc/sti/uniperif.h
349
SET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
35
#define SET_UNIPERIF_SOFT_RST(ip, value) \
sound/soc/sti/uniperif.h
350
UNIPERIF_CONFIG_OFFSET(ip), \
sound/soc/sti/uniperif.h
351
UNIPERIF_CONFIG_VALIDITY_DAT_CNTR_SHIFT(ip), \
sound/soc/sti/uniperif.h
352
UNIPERIF_CONFIG_VALIDITY_DAT_CNTR_MASK(ip), 1)
sound/soc/sti/uniperif.h
355
#define UNIPERIF_CONFIG_ONE_BIT_AUD_SHIFT(ip) 4
sound/soc/sti/uniperif.h
356
#define UNIPERIF_CONFIG_ONE_BIT_AUD_MASK(ip) 0x1
sound/soc/sti/uniperif.h
357
#define GET_UNIPERIF_CONFIG_ONE_BIT_AUD(ip) \
sound/soc/sti/uniperif.h
358
GET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
359
UNIPERIF_CONFIG_OFFSET(ip), \
sound/soc/sti/uniperif.h
36
writel_relaxed(value, ip->base + UNIPERIF_SOFT_RST_OFFSET(ip))
sound/soc/sti/uniperif.h
360
UNIPERIF_CONFIG_ONE_BIT_AUD_SHIFT(ip), \
sound/soc/sti/uniperif.h
361
UNIPERIF_CONFIG_ONE_BIT_AUD_MASK(ip))
sound/soc/sti/uniperif.h
362
#define SET_UNIPERIF_CONFIG_ONE_BIT_AUD_DISABLE(ip) \
sound/soc/sti/uniperif.h
363
SET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
364
UNIPERIF_CONFIG_OFFSET(ip), \
sound/soc/sti/uniperif.h
365
UNIPERIF_CONFIG_ONE_BIT_AUD_SHIFT(ip), \
sound/soc/sti/uniperif.h
366
UNIPERIF_CONFIG_ONE_BIT_AUD_MASK(ip), 0)
sound/soc/sti/uniperif.h
367
#define SET_UNIPERIF_CONFIG_ONE_BIT_AUD_ENABLE(ip) \
sound/soc/sti/uniperif.h
368
SET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
369
UNIPERIF_CONFIG_OFFSET(ip), \
sound/soc/sti/uniperif.h
370
UNIPERIF_CONFIG_ONE_BIT_AUD_SHIFT(ip), \
sound/soc/sti/uniperif.h
371
UNIPERIF_CONFIG_ONE_BIT_AUD_MASK(ip), 1)
sound/soc/sti/uniperif.h
374
#define UNIPERIF_CONFIG_MEM_FMT_SHIFT(ip) 5
sound/soc/sti/uniperif.h
375
#define UNIPERIF_CONFIG_MEM_FMT_MASK(ip) 0x1
sound/soc/sti/uniperif.h
376
#define VALUE_UNIPERIF_CONFIG_MEM_FMT_16_0(ip) 0
sound/soc/sti/uniperif.h
377
#define VALUE_UNIPERIF_CONFIG_MEM_FMT_16_16(ip) 1
sound/soc/sti/uniperif.h
378
#define GET_UNIPERIF_CONFIG_MEM_FMT(ip) \
sound/soc/sti/uniperif.h
379
GET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
380
UNIPERIF_CONFIG_OFFSET(ip), \
sound/soc/sti/uniperif.h
381
UNIPERIF_CONFIG_MEM_FMT_SHIFT(ip), \
sound/soc/sti/uniperif.h
382
UNIPERIF_CONFIG_MEM_FMT_MASK(ip))
sound/soc/sti/uniperif.h
383
#define SET_UNIPERIF_CONFIG_MEM_FMT(ip, value) \
sound/soc/sti/uniperif.h
384
SET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
385
UNIPERIF_CONFIG_OFFSET(ip), \
sound/soc/sti/uniperif.h
386
UNIPERIF_CONFIG_MEM_FMT_SHIFT(ip), \
sound/soc/sti/uniperif.h
387
UNIPERIF_CONFIG_MEM_FMT_MASK(ip), value)
sound/soc/sti/uniperif.h
388
#define SET_UNIPERIF_CONFIG_MEM_FMT_16_0(ip) \
sound/soc/sti/uniperif.h
389
SET_UNIPERIF_CONFIG_MEM_FMT(ip, \
sound/soc/sti/uniperif.h
39
#define UNIPERIF_SOFT_RST_SOFT_RST_SHIFT(ip) 0x0
sound/soc/sti/uniperif.h
390
VALUE_UNIPERIF_CONFIG_MEM_FMT_16_0(ip))
sound/soc/sti/uniperif.h
391
#define SET_UNIPERIF_CONFIG_MEM_FMT_16_16(ip) \
sound/soc/sti/uniperif.h
392
SET_UNIPERIF_CONFIG_MEM_FMT(ip, \
sound/soc/sti/uniperif.h
393
VALUE_UNIPERIF_CONFIG_MEM_FMT_16_16(ip))
sound/soc/sti/uniperif.h
396
#define UNIPERIF_CONFIG_REPEAT_CHL_STS_SHIFT(ip) 6
sound/soc/sti/uniperif.h
397
#define UNIPERIF_CONFIG_REPEAT_CHL_STS_MASK(ip) 0x1
sound/soc/sti/uniperif.h
398
#define GET_UNIPERIF_CONFIG_REPEAT_CHL_STS(ip) \
sound/soc/sti/uniperif.h
399
GET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
40
#define UNIPERIF_SOFT_RST_SOFT_RST_MASK(ip) 0x1
sound/soc/sti/uniperif.h
400
UNIPERIF_CONFIG_OFFSET(ip), \
sound/soc/sti/uniperif.h
401
UNIPERIF_CONFIG_REPEAT_CHL_STS_SHIFT(ip), \
sound/soc/sti/uniperif.h
402
UNIPERIF_CONFIG_REPEAT_CHL_STS_MASK(ip))
sound/soc/sti/uniperif.h
403
#define SET_UNIPERIF_CONFIG_REPEAT_CHL_STS_ENABLE(ip) \
sound/soc/sti/uniperif.h
404
SET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
405
UNIPERIF_CONFIG_OFFSET(ip), \
sound/soc/sti/uniperif.h
406
UNIPERIF_CONFIG_REPEAT_CHL_STS_SHIFT(ip), \
sound/soc/sti/uniperif.h
407
UNIPERIF_CONFIG_REPEAT_CHL_STS_MASK(ip), 0)
sound/soc/sti/uniperif.h
408
#define SET_UNIPERIF_CONFIG_REPEAT_CHL_STS_DISABLE(ip) \
sound/soc/sti/uniperif.h
409
SET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
41
#define SET_UNIPERIF_SOFT_RST_SOFT_RST(ip) \
sound/soc/sti/uniperif.h
410
UNIPERIF_CONFIG_OFFSET(ip), \
sound/soc/sti/uniperif.h
411
UNIPERIF_CONFIG_REPEAT_CHL_STS_SHIFT(ip), \
sound/soc/sti/uniperif.h
412
UNIPERIF_CONFIG_REPEAT_CHL_STS_MASK(ip), 1)
sound/soc/sti/uniperif.h
415
#define UNIPERIF_CONFIG_BACK_STALL_REQ_SHIFT(ip) \
sound/soc/sti/uniperif.h
416
((ip)->ver < SND_ST_UNIPERIF_VERSION_UNI_PLR_TOP_1_0 ? 7 : -1)
sound/soc/sti/uniperif.h
417
#define UNIPERIF_CONFIG_BACK_STALL_REQ_MASK(ip) 0x1
sound/soc/sti/uniperif.h
418
#define GET_UNIPERIF_CONFIG_BACK_STALL_REQ(ip) \
sound/soc/sti/uniperif.h
419
GET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
42
SET_UNIPERIF_BIT_REG(ip, \
sound/soc/sti/uniperif.h
420
UNIPERIF_CONFIG_OFFSET(ip), \
sound/soc/sti/uniperif.h
421
UNIPERIF_CONFIG_BACK_STALL_REQ_SHIFT(ip), \
sound/soc/sti/uniperif.h
422
UNIPERIF_CONFIG_BACK_STALL_REQ_MASK(ip))
sound/soc/sti/uniperif.h
423
#define SET_UNIPERIF_CONFIG_BACK_STALL_REQ_DISABLE(ip) \
sound/soc/sti/uniperif.h
424
SET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
425
UNIPERIF_CONFIG_OFFSET(ip), \
sound/soc/sti/uniperif.h
426
UNIPERIF_CONFIG_BACK_STALL_REQ_SHIFT(ip), \
sound/soc/sti/uniperif.h
427
UNIPERIF_CONFIG_BACK_STALL_REQ_MASK(ip), 0)
sound/soc/sti/uniperif.h
428
#define SET_UNIPERIF_CONFIG_BACK_STALL_REQ_ENABLE(ip) \
sound/soc/sti/uniperif.h
429
SET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
43
UNIPERIF_SOFT_RST_OFFSET(ip), \
sound/soc/sti/uniperif.h
430
UNIPERIF_CONFIG_OFFSET(ip), \
sound/soc/sti/uniperif.h
431
UNIPERIF_CONFIG_BACK_STALL_REQ_SHIFT(ip), \
sound/soc/sti/uniperif.h
432
UNIPERIF_CONFIG_BACK_STALL_REQ_MASK(ip), 1)
sound/soc/sti/uniperif.h
435
#define UNIPERIF_CONFIG_DMA_TRIG_LIMIT_SHIFT(ip) 8
sound/soc/sti/uniperif.h
436
#define UNIPERIF_CONFIG_DMA_TRIG_LIMIT_MASK(ip) 0x7F
sound/soc/sti/uniperif.h
437
#define GET_UNIPERIF_CONFIG_DMA_TRIG_LIMIT(ip) \
sound/soc/sti/uniperif.h
438
GET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
439
UNIPERIF_CONFIG_OFFSET(ip), \
sound/soc/sti/uniperif.h
44
UNIPERIF_SOFT_RST_SOFT_RST_SHIFT(ip), \
sound/soc/sti/uniperif.h
440
UNIPERIF_CONFIG_DMA_TRIG_LIMIT_SHIFT(ip), \
sound/soc/sti/uniperif.h
441
UNIPERIF_CONFIG_DMA_TRIG_LIMIT_MASK(ip))
sound/soc/sti/uniperif.h
442
#define SET_UNIPERIF_CONFIG_DMA_TRIG_LIMIT(ip, value) \
sound/soc/sti/uniperif.h
443
SET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
444
UNIPERIF_CONFIG_OFFSET(ip), \
sound/soc/sti/uniperif.h
445
UNIPERIF_CONFIG_DMA_TRIG_LIMIT_SHIFT(ip), \
sound/soc/sti/uniperif.h
446
UNIPERIF_CONFIG_DMA_TRIG_LIMIT_MASK(ip), value)
sound/soc/sti/uniperif.h
449
#define UNIPERIF_CONFIG_CHL_STS_UPDATE_SHIFT(ip) \
sound/soc/sti/uniperif.h
45
UNIPERIF_SOFT_RST_SOFT_RST_MASK(ip), 1)
sound/soc/sti/uniperif.h
450
((ip)->ver < SND_ST_UNIPERIF_VERSION_UNI_PLR_TOP_1_0 ? 16 : -1)
sound/soc/sti/uniperif.h
451
#define UNIPERIF_CONFIG_CHL_STS_UPDATE_MASK(ip) 0x1
sound/soc/sti/uniperif.h
452
#define GET_UNIPERIF_CONFIG_CHL_STS_UPDATE(ip) \
sound/soc/sti/uniperif.h
453
GET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
454
UNIPERIF_CONFIG_OFFSET(ip), \
sound/soc/sti/uniperif.h
455
UNIPERIF_CONFIG_CHL_STS_UPDATE_SHIFT(ip), \
sound/soc/sti/uniperif.h
456
UNIPERIF_CONFIG_CHL_STS_UPDATE_MASK(ip))
sound/soc/sti/uniperif.h
457
#define SET_UNIPERIF_CONFIG_CHL_STS_UPDATE(ip) \
sound/soc/sti/uniperif.h
458
SET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
459
UNIPERIF_CONFIG_OFFSET(ip), \
sound/soc/sti/uniperif.h
46
#define GET_UNIPERIF_SOFT_RST_SOFT_RST(ip) \
sound/soc/sti/uniperif.h
460
UNIPERIF_CONFIG_CHL_STS_UPDATE_SHIFT(ip), \
sound/soc/sti/uniperif.h
461
UNIPERIF_CONFIG_CHL_STS_UPDATE_MASK(ip), 1)
sound/soc/sti/uniperif.h
464
#define UNIPERIF_CONFIG_IDLE_MOD_SHIFT(ip) 18
sound/soc/sti/uniperif.h
465
#define UNIPERIF_CONFIG_IDLE_MOD_MASK(ip) 0x1
sound/soc/sti/uniperif.h
466
#define GET_UNIPERIF_CONFIG_IDLE_MOD(ip) \
sound/soc/sti/uniperif.h
467
GET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
468
UNIPERIF_CONFIG_OFFSET(ip), \
sound/soc/sti/uniperif.h
469
UNIPERIF_CONFIG_IDLE_MOD_SHIFT(ip), \
sound/soc/sti/uniperif.h
47
GET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
470
UNIPERIF_CONFIG_IDLE_MOD_MASK(ip))
sound/soc/sti/uniperif.h
471
#define SET_UNIPERIF_CONFIG_IDLE_MOD_DISABLE(ip) \
sound/soc/sti/uniperif.h
472
SET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
473
UNIPERIF_CONFIG_OFFSET(ip), \
sound/soc/sti/uniperif.h
474
UNIPERIF_CONFIG_IDLE_MOD_SHIFT(ip), \
sound/soc/sti/uniperif.h
475
UNIPERIF_CONFIG_IDLE_MOD_MASK(ip), 0)
sound/soc/sti/uniperif.h
476
#define SET_UNIPERIF_CONFIG_IDLE_MOD_ENABLE(ip) \
sound/soc/sti/uniperif.h
477
SET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
478
UNIPERIF_CONFIG_OFFSET(ip), \
sound/soc/sti/uniperif.h
479
UNIPERIF_CONFIG_IDLE_MOD_SHIFT(ip), \
sound/soc/sti/uniperif.h
48
UNIPERIF_SOFT_RST_OFFSET(ip), \
sound/soc/sti/uniperif.h
480
UNIPERIF_CONFIG_IDLE_MOD_MASK(ip), 1)
sound/soc/sti/uniperif.h
483
#define UNIPERIF_CONFIG_SUBFRAME_SEL_SHIFT(ip) 19
sound/soc/sti/uniperif.h
484
#define UNIPERIF_CONFIG_SUBFRAME_SEL_MASK(ip) 0x1
sound/soc/sti/uniperif.h
485
#define GET_UNIPERIF_CONFIG_SUBFRAME_SEL(ip) \
sound/soc/sti/uniperif.h
486
GET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
487
UNIPERIF_CONFIG_OFFSET(ip), \
sound/soc/sti/uniperif.h
488
UNIPERIF_CONFIG_SUBFRAME_SEL_SHIFT(ip), \
sound/soc/sti/uniperif.h
489
UNIPERIF_CONFIG_SUBFRAME_SEL_MASK(ip))
sound/soc/sti/uniperif.h
49
UNIPERIF_SOFT_RST_SOFT_RST_SHIFT(ip), \
sound/soc/sti/uniperif.h
490
#define SET_UNIPERIF_CONFIG_SUBFRAME_SEL_SUBF1_SUBF0(ip) \
sound/soc/sti/uniperif.h
491
SET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
492
UNIPERIF_CONFIG_OFFSET(ip), \
sound/soc/sti/uniperif.h
493
UNIPERIF_CONFIG_SUBFRAME_SEL_SHIFT(ip), \
sound/soc/sti/uniperif.h
494
UNIPERIF_CONFIG_SUBFRAME_SEL_MASK(ip), 1)
sound/soc/sti/uniperif.h
495
#define SET_UNIPERIF_CONFIG_SUBFRAME_SEL_SUBF0_SUBF1(ip) \
sound/soc/sti/uniperif.h
496
SET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
497
UNIPERIF_CONFIG_OFFSET(ip), \
sound/soc/sti/uniperif.h
498
UNIPERIF_CONFIG_SUBFRAME_SEL_SHIFT(ip), \
sound/soc/sti/uniperif.h
499
UNIPERIF_CONFIG_SUBFRAME_SEL_MASK(ip), 0)
sound/soc/sti/uniperif.h
50
UNIPERIF_SOFT_RST_SOFT_RST_MASK(ip))
sound/soc/sti/uniperif.h
502
#define UNIPERIF_CONFIG_SPDIF_SW_CTRL_SHIFT(ip) 20
sound/soc/sti/uniperif.h
503
#define UNIPERIF_CONFIG_SPDIF_SW_CTRL_MASK(ip) 0x1
sound/soc/sti/uniperif.h
504
#define GET_UNIPERIF_CONFIG_SPDIF_SW_CTRL(ip) \
sound/soc/sti/uniperif.h
505
GET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
506
UNIPERIF_CONFIG_OFFSET(ip), \
sound/soc/sti/uniperif.h
507
UNIPERIF_CONFIG_SPDIF_SW_CTRL_SHIFT(ip), \
sound/soc/sti/uniperif.h
508
UNIPERIF_CONFIG_SPDIF_SW_CTRL_MASK(ip))
sound/soc/sti/uniperif.h
509
#define SET_UNIPERIF_CONFIG_SPDIF_SW_CTRL_ENABLE(ip) \
sound/soc/sti/uniperif.h
510
SET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
511
UNIPERIF_CONFIG_OFFSET(ip), \
sound/soc/sti/uniperif.h
512
UNIPERIF_CONFIG_SPDIF_SW_CTRL_SHIFT(ip), \
sound/soc/sti/uniperif.h
513
UNIPERIF_CONFIG_SPDIF_SW_CTRL_MASK(ip), 1)
sound/soc/sti/uniperif.h
514
#define SET_UNIPERIF_CONFIG_SPDIF_SW_CTRL_DISABLE(ip) \
sound/soc/sti/uniperif.h
515
SET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
516
UNIPERIF_CONFIG_OFFSET(ip), \
sound/soc/sti/uniperif.h
517
UNIPERIF_CONFIG_SPDIF_SW_CTRL_SHIFT(ip), \
sound/soc/sti/uniperif.h
518
UNIPERIF_CONFIG_SPDIF_SW_CTRL_MASK(ip), 0)
sound/soc/sti/uniperif.h
521
#define UNIPERIF_CONFIG_MSTR_CLKEDGE_SHIFT(ip) \
sound/soc/sti/uniperif.h
522
((ip)->ver < SND_ST_UNIPERIF_VERSION_UNI_PLR_TOP_1_0 ? 24 : -1)
sound/soc/sti/uniperif.h
523
#define UNIPERIF_CONFIG_MSTR_CLKEDGE_MASK(ip) 0x1
sound/soc/sti/uniperif.h
524
#define GET_UNIPERIF_CONFIG_MSTR_CLKEDGE(ip) \
sound/soc/sti/uniperif.h
525
GET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
526
UNIPERIF_CONFIG_OFFSET(ip), \
sound/soc/sti/uniperif.h
527
UNIPERIF_CONFIG_MSTR_CLKEDGE_SHIFT(ip), \
sound/soc/sti/uniperif.h
528
UNIPERIF_CONFIG_MSTR_CLKEDGE_MASK(ip))
sound/soc/sti/uniperif.h
529
#define SET_UNIPERIF_CONFIG_MSTR_CLKEDGE_FALLING(ip) \
sound/soc/sti/uniperif.h
530
SET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
531
UNIPERIF_CONFIG_OFFSET(ip), \
sound/soc/sti/uniperif.h
532
UNIPERIF_CONFIG_MSTR_CLKEDGE_SHIFT(ip), \
sound/soc/sti/uniperif.h
533
UNIPERIF_CONFIG_MSTR_CLKEDGE_MASK(ip), 1)
sound/soc/sti/uniperif.h
534
#define SET_UNIPERIF_CONFIG_MSTR_CLKEDGE_RISING(ip) \
sound/soc/sti/uniperif.h
535
SET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
536
UNIPERIF_CONFIG_OFFSET(ip), \
sound/soc/sti/uniperif.h
537
UNIPERIF_CONFIG_MSTR_CLKEDGE_SHIFT(ip), \
sound/soc/sti/uniperif.h
538
UNIPERIF_CONFIG_MSTR_CLKEDGE_MASK(ip), 0)
sound/soc/sti/uniperif.h
544
#define UNIPERIF_CTRL_OFFSET(ip) 0x0044
sound/soc/sti/uniperif.h
545
#define GET_UNIPERIF_CTRL(ip) \
sound/soc/sti/uniperif.h
546
readl_relaxed(ip->base + UNIPERIF_CTRL_OFFSET(ip))
sound/soc/sti/uniperif.h
547
#define SET_UNIPERIF_CTRL(ip, value) \
sound/soc/sti/uniperif.h
548
writel_relaxed(value, ip->base + UNIPERIF_CTRL_OFFSET(ip))
sound/soc/sti/uniperif.h
551
#define UNIPERIF_CTRL_OPERATION_SHIFT(ip) 0
sound/soc/sti/uniperif.h
552
#define UNIPERIF_CTRL_OPERATION_MASK(ip) 0x7
sound/soc/sti/uniperif.h
553
#define GET_UNIPERIF_CTRL_OPERATION(ip) \
sound/soc/sti/uniperif.h
554
GET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
555
UNIPERIF_CTRL_OFFSET(ip), \
sound/soc/sti/uniperif.h
556
UNIPERIF_CTRL_OPERATION_SHIFT(ip), \
sound/soc/sti/uniperif.h
557
UNIPERIF_CTRL_OPERATION_MASK(ip))
sound/soc/sti/uniperif.h
558
#define VALUE_UNIPERIF_CTRL_OPERATION_OFF(ip) 0
sound/soc/sti/uniperif.h
559
#define SET_UNIPERIF_CTRL_OPERATION_OFF(ip) \
sound/soc/sti/uniperif.h
56
#define UNIPERIF_FIFO_DATA_OFFSET(ip) 0x0004
sound/soc/sti/uniperif.h
560
SET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
561
UNIPERIF_CTRL_OFFSET(ip), \
sound/soc/sti/uniperif.h
562
UNIPERIF_CTRL_OPERATION_SHIFT(ip), \
sound/soc/sti/uniperif.h
563
UNIPERIF_CTRL_OPERATION_MASK(ip), \
sound/soc/sti/uniperif.h
564
VALUE_UNIPERIF_CTRL_OPERATION_OFF(ip))
sound/soc/sti/uniperif.h
565
#define VALUE_UNIPERIF_CTRL_OPERATION_MUTE_PCM_NULL(ip) \
sound/soc/sti/uniperif.h
566
((ip)->ver < SND_ST_UNIPERIF_VERSION_UNI_PLR_TOP_1_0 ? 1 : -1)
sound/soc/sti/uniperif.h
567
#define SET_UNIPERIF_CTRL_OPERATION_MUTE_PCM_NULL(ip) \
sound/soc/sti/uniperif.h
568
SET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
569
UNIPERIF_CTRL_OFFSET(ip), \
sound/soc/sti/uniperif.h
57
#define SET_UNIPERIF_DATA(ip, value) \
sound/soc/sti/uniperif.h
570
UNIPERIF_CTRL_OPERATION_SHIFT(ip), \
sound/soc/sti/uniperif.h
571
UNIPERIF_CTRL_OPERATION_MASK(ip), \
sound/soc/sti/uniperif.h
572
VALUE_UNIPERIF_CTRL_OPERATION_MUTE_PCM_NULL(ip))
sound/soc/sti/uniperif.h
573
#define VALUE_UNIPERIF_CTRL_OPERATION_MUTE_PAUSE_BURST(ip) \
sound/soc/sti/uniperif.h
574
((ip)->ver < SND_ST_UNIPERIF_VERSION_UNI_PLR_TOP_1_0 ? 2 : -1)
sound/soc/sti/uniperif.h
575
#define SET_UNIPERIF_CTRL_OPERATION_MUTE_PAUSE_BURST(ip) \
sound/soc/sti/uniperif.h
576
SET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
577
UNIPERIF_CTRL_OFFSET(ip), \
sound/soc/sti/uniperif.h
578
UNIPERIF_CTRL_OPERATION_SHIFT(ip), \
sound/soc/sti/uniperif.h
579
UNIPERIF_CTRL_OPERATION_MASK(ip), \
sound/soc/sti/uniperif.h
58
writel_relaxed(value, ip->base + UNIPERIF_FIFO_DATA_OFFSET(ip))
sound/soc/sti/uniperif.h
580
VALUE_UNIPERIF_CTRL_OPERATION_MUTE_PAUSE_BURST(ip))
sound/soc/sti/uniperif.h
581
#define VALUE_UNIPERIF_CTRL_OPERATION_PCM_DATA(ip) 3
sound/soc/sti/uniperif.h
582
#define SET_UNIPERIF_CTRL_OPERATION_PCM_DATA(ip) \
sound/soc/sti/uniperif.h
583
SET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
584
UNIPERIF_CTRL_OFFSET(ip), \
sound/soc/sti/uniperif.h
585
UNIPERIF_CTRL_OPERATION_SHIFT(ip), \
sound/soc/sti/uniperif.h
586
UNIPERIF_CTRL_OPERATION_MASK(ip), \
sound/soc/sti/uniperif.h
587
VALUE_UNIPERIF_CTRL_OPERATION_PCM_DATA(ip))
sound/soc/sti/uniperif.h
589
#define VALUE_UNIPERIF_CTRL_OPERATION_AUDIO_DATA(ip) 3
sound/soc/sti/uniperif.h
590
#define SET_UNIPERIF_CTRL_OPERATION_AUDIO_DATA(ip) \
sound/soc/sti/uniperif.h
591
SET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
592
UNIPERIF_CTRL_OFFSET(ip), \
sound/soc/sti/uniperif.h
593
UNIPERIF_CTRL_OPERATION_SHIFT(ip), \
sound/soc/sti/uniperif.h
594
UNIPERIF_CTRL_OPERATION_MASK(ip), \
sound/soc/sti/uniperif.h
595
VALUE_UNIPERIF_CTRL_OPERATION_AUDIO_DATA(ip))
sound/soc/sti/uniperif.h
596
#define VALUE_UNIPERIF_CTRL_OPERATION_ENC_DATA(ip) 4
sound/soc/sti/uniperif.h
597
#define SET_UNIPERIF_CTRL_OPERATION_ENC_DATA(ip) \
sound/soc/sti/uniperif.h
598
SET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
599
UNIPERIF_CTRL_OFFSET(ip), \
sound/soc/sti/uniperif.h
600
UNIPERIF_CTRL_OPERATION_SHIFT(ip), \
sound/soc/sti/uniperif.h
601
UNIPERIF_CTRL_OPERATION_MASK(ip), \
sound/soc/sti/uniperif.h
602
VALUE_UNIPERIF_CTRL_OPERATION_ENC_DATA(ip))
sound/soc/sti/uniperif.h
603
#define VALUE_UNIPERIF_CTRL_OPERATION_CD_DATA(ip) \
sound/soc/sti/uniperif.h
604
((ip)->ver < SND_ST_UNIPERIF_VERSION_UNI_PLR_TOP_1_0 ? 5 : -1)
sound/soc/sti/uniperif.h
605
#define SET_UNIPERIF_CTRL_OPERATION_CD_DATA(ip) \
sound/soc/sti/uniperif.h
606
SET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
607
UNIPERIF_CTRL_OFFSET(ip), \
sound/soc/sti/uniperif.h
608
UNIPERIF_CTRL_OPERATION_SHIFT(ip), \
sound/soc/sti/uniperif.h
609
UNIPERIF_CTRL_OPERATION_MASK(ip), \
sound/soc/sti/uniperif.h
610
VALUE_UNIPERIF_CTRL_OPERATION_CD_DATA(ip))
sound/soc/sti/uniperif.h
611
#define VALUE_UNIPERIF_CTRL_OPERATION_STANDBY(ip) \
sound/soc/sti/uniperif.h
612
((ip)->ver < SND_ST_UNIPERIF_VERSION_UNI_PLR_TOP_1_0 ? -1 : 7)
sound/soc/sti/uniperif.h
613
#define SET_UNIPERIF_CTRL_OPERATION_STANDBY(ip) \
sound/soc/sti/uniperif.h
614
SET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
615
UNIPERIF_CTRL_OFFSET(ip), \
sound/soc/sti/uniperif.h
616
UNIPERIF_CTRL_OPERATION_SHIFT(ip), \
sound/soc/sti/uniperif.h
617
UNIPERIF_CTRL_OPERATION_MASK(ip), \
sound/soc/sti/uniperif.h
618
VALUE_UNIPERIF_CTRL_OPERATION_STANDBY(ip))
sound/soc/sti/uniperif.h
621
#define UNIPERIF_CTRL_EXIT_STBY_ON_EOBLOCK_SHIFT(ip) \
sound/soc/sti/uniperif.h
622
((ip)->ver < SND_ST_UNIPERIF_VERSION_UNI_PLR_TOP_1_0 ? -1 : 3)
sound/soc/sti/uniperif.h
623
#define UNIPERIF_CTRL_EXIT_STBY_ON_EOBLOCK_MASK(ip) 0x1
sound/soc/sti/uniperif.h
624
#define GET_UNIPERIF_CTRL_EXIT_STBY_ON_EOBLOCK(ip) \
sound/soc/sti/uniperif.h
625
GET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
626
UNIPERIF_CTRL_OFFSET(ip), \
sound/soc/sti/uniperif.h
627
UNIPERIF_CTRL_EXIT_STBY_ON_EOBLOCK_SHIFT(ip), \
sound/soc/sti/uniperif.h
628
UNIPERIF_CTRL_EXIT_STBY_ON_EOBLOCK_MASK(ip))
sound/soc/sti/uniperif.h
629
#define SET_UNIPERIF_CTRL_EXIT_STBY_ON_EOBLOCK_OFF(ip) \
sound/soc/sti/uniperif.h
630
SET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
631
UNIPERIF_CTRL_OFFSET(ip), \
sound/soc/sti/uniperif.h
632
UNIPERIF_CTRL_EXIT_STBY_ON_EOBLOCK_SHIFT(ip), \
sound/soc/sti/uniperif.h
633
UNIPERIF_CTRL_EXIT_STBY_ON_EOBLOCK_MASK(ip), 0)
sound/soc/sti/uniperif.h
634
#define SET_UNIPERIF_CTRL_EXIT_STBY_ON_EOBLOCK_ON(ip) \
sound/soc/sti/uniperif.h
635
SET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
636
UNIPERIF_CTRL_OFFSET(ip), \
sound/soc/sti/uniperif.h
637
UNIPERIF_CTRL_EXIT_STBY_ON_EOBLOCK_SHIFT(ip), \
sound/soc/sti/uniperif.h
638
UNIPERIF_CTRL_EXIT_STBY_ON_EOBLOCK_MASK(ip), 1)
sound/soc/sti/uniperif.h
64
#define UNIPERIF_CHANNEL_STA_REGN(ip, n) (0x0060 + (4 * n))
sound/soc/sti/uniperif.h
641
#define UNIPERIF_CTRL_ROUNDING_SHIFT(ip) 4
sound/soc/sti/uniperif.h
642
#define UNIPERIF_CTRL_ROUNDING_MASK(ip) 0x1
sound/soc/sti/uniperif.h
643
#define GET_UNIPERIF_CTRL_ROUNDING(ip) \
sound/soc/sti/uniperif.h
644
GET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
645
UNIPERIF_CTRL_OFFSET(ip), \
sound/soc/sti/uniperif.h
646
UNIPERIF_CTRL_ROUNDING_SHIFT(ip), \
sound/soc/sti/uniperif.h
647
UNIPERIF_CTRL_ROUNDING_MASK(ip))
sound/soc/sti/uniperif.h
648
#define SET_UNIPERIF_CTRL_ROUNDING_OFF(ip) \
sound/soc/sti/uniperif.h
649
SET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
65
#define GET_UNIPERIF_CHANNEL_STA_REGN(ip) \
sound/soc/sti/uniperif.h
650
UNIPERIF_CTRL_OFFSET(ip), \
sound/soc/sti/uniperif.h
651
UNIPERIF_CTRL_ROUNDING_SHIFT(ip), \
sound/soc/sti/uniperif.h
652
UNIPERIF_CTRL_ROUNDING_MASK(ip), 0)
sound/soc/sti/uniperif.h
653
#define SET_UNIPERIF_CTRL_ROUNDING_ON(ip) \
sound/soc/sti/uniperif.h
654
SET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
655
UNIPERIF_CTRL_OFFSET(ip), \
sound/soc/sti/uniperif.h
656
UNIPERIF_CTRL_ROUNDING_SHIFT(ip), \
sound/soc/sti/uniperif.h
657
UNIPERIF_CTRL_ROUNDING_MASK(ip), 1)
sound/soc/sti/uniperif.h
66
readl_relaxed(ip->base + UNIPERIF_CHANNEL_STA_REGN(ip, n))
sound/soc/sti/uniperif.h
660
#define UNIPERIF_CTRL_DIVIDER_SHIFT(ip) 5
sound/soc/sti/uniperif.h
661
#define UNIPERIF_CTRL_DIVIDER_MASK(ip) 0xff
sound/soc/sti/uniperif.h
662
#define GET_UNIPERIF_CTRL_DIVIDER(ip) \
sound/soc/sti/uniperif.h
663
GET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
664
UNIPERIF_CTRL_OFFSET(ip), \
sound/soc/sti/uniperif.h
665
UNIPERIF_CTRL_DIVIDER_SHIFT(ip), \
sound/soc/sti/uniperif.h
666
UNIPERIF_CTRL_DIVIDER_MASK(ip))
sound/soc/sti/uniperif.h
667
#define SET_UNIPERIF_CTRL_DIVIDER(ip, value) \
sound/soc/sti/uniperif.h
668
SET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
669
UNIPERIF_CTRL_OFFSET(ip), \
sound/soc/sti/uniperif.h
67
#define SET_UNIPERIF_CHANNEL_STA_REGN(ip, n, value) \
sound/soc/sti/uniperif.h
670
UNIPERIF_CTRL_DIVIDER_SHIFT(ip), \
sound/soc/sti/uniperif.h
671
UNIPERIF_CTRL_DIVIDER_MASK(ip), value)
sound/soc/sti/uniperif.h
674
#define UNIPERIF_CTRL_BYTE_SWP_SHIFT(ip) \
sound/soc/sti/uniperif.h
675
((ip)->ver < SND_ST_UNIPERIF_VERSION_UNI_PLR_TOP_1_0 ? 13 : -1)
sound/soc/sti/uniperif.h
676
#define UNIPERIF_CTRL_BYTE_SWP_MASK(ip) 0x1
sound/soc/sti/uniperif.h
677
#define GET_UNIPERIF_CTRL_BYTE_SWP(ip) \
sound/soc/sti/uniperif.h
678
GET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
679
UNIPERIF_CTRL_OFFSET(ip), \
sound/soc/sti/uniperif.h
68
writel_relaxed(value, ip->base + \
sound/soc/sti/uniperif.h
680
UNIPERIF_CTRL_BYTE_SWP_SHIFT(ip), \
sound/soc/sti/uniperif.h
681
UNIPERIF_CTRL_BYTE_SWP_MASK(ip))
sound/soc/sti/uniperif.h
682
#define SET_UNIPERIF_CTRL_BYTE_SWP_OFF(ip) \
sound/soc/sti/uniperif.h
683
SET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
684
UNIPERIF_CTRL_OFFSET(ip), \
sound/soc/sti/uniperif.h
685
UNIPERIF_CTRL_BYTE_SWP_SHIFT(ip), \
sound/soc/sti/uniperif.h
686
UNIPERIF_CTRL_BYTE_SWP_MASK(ip), 0)
sound/soc/sti/uniperif.h
687
#define SET_UNIPERIF_CTRL_BYTE_SWP_ON(ip) \
sound/soc/sti/uniperif.h
688
SET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
689
UNIPERIF_CTRL_OFFSET(ip), \
sound/soc/sti/uniperif.h
69
UNIPERIF_CHANNEL_STA_REGN(ip, n))
sound/soc/sti/uniperif.h
690
UNIPERIF_CTRL_BYTE_SWP_SHIFT(ip), \
sound/soc/sti/uniperif.h
691
UNIPERIF_CTRL_BYTE_SWP_MASK(ip), 1)
sound/soc/sti/uniperif.h
694
#define UNIPERIF_CTRL_ZERO_STUFF_SHIFT(ip) \
sound/soc/sti/uniperif.h
695
((ip)->ver < SND_ST_UNIPERIF_VERSION_UNI_PLR_TOP_1_0 ? 14 : -1)
sound/soc/sti/uniperif.h
696
#define UNIPERIF_CTRL_ZERO_STUFF_MASK(ip) 0x1
sound/soc/sti/uniperif.h
697
#define GET_UNIPERIF_CTRL_ZERO_STUFF(ip) \
sound/soc/sti/uniperif.h
698
GET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
699
UNIPERIF_CTRL_OFFSET(ip), \
sound/soc/sti/uniperif.h
700
UNIPERIF_CTRL_ZERO_STUFF_SHIFT(ip), \
sound/soc/sti/uniperif.h
701
UNIPERIF_CTRL_ZERO_STUFF_MASK(ip))
sound/soc/sti/uniperif.h
702
#define SET_UNIPERIF_CTRL_ZERO_STUFF_HW(ip) \
sound/soc/sti/uniperif.h
703
SET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
704
UNIPERIF_CTRL_OFFSET(ip), \
sound/soc/sti/uniperif.h
705
UNIPERIF_CTRL_ZERO_STUFF_SHIFT(ip), \
sound/soc/sti/uniperif.h
706
UNIPERIF_CTRL_ZERO_STUFF_MASK(ip), 1)
sound/soc/sti/uniperif.h
707
#define SET_UNIPERIF_CTRL_ZERO_STUFF_SW(ip) \
sound/soc/sti/uniperif.h
708
SET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
709
UNIPERIF_CTRL_OFFSET(ip), \
sound/soc/sti/uniperif.h
71
#define UNIPERIF_CHANNEL_STA_REG0_OFFSET(ip) 0x0060
sound/soc/sti/uniperif.h
710
UNIPERIF_CTRL_ZERO_STUFF_SHIFT(ip), \
sound/soc/sti/uniperif.h
711
UNIPERIF_CTRL_ZERO_STUFF_MASK(ip), 0)
sound/soc/sti/uniperif.h
714
#define UNIPERIF_CTRL_SPDIF_LAT_SHIFT(ip) \
sound/soc/sti/uniperif.h
715
((ip)->ver < SND_ST_UNIPERIF_VERSION_UNI_PLR_TOP_1_0 ? 16 : -1)
sound/soc/sti/uniperif.h
716
#define UNIPERIF_CTRL_SPDIF_LAT_MASK(ip) 0x1
sound/soc/sti/uniperif.h
717
#define GET_UNIPERIF_CTRL_SPDIF_LAT(ip) \
sound/soc/sti/uniperif.h
718
GET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
719
UNIPERIF_CTRL_OFFSET(ip), \
sound/soc/sti/uniperif.h
72
#define GET_UNIPERIF_CHANNEL_STA_REG0(ip) \
sound/soc/sti/uniperif.h
720
UNIPERIF_CTRL_SPDIF_LAT_SHIFT(ip), \
sound/soc/sti/uniperif.h
721
UNIPERIF_CTRL_SPDIF_LAT_MASK(ip))
sound/soc/sti/uniperif.h
722
#define SET_UNIPERIF_CTRL_SPDIF_LAT_ON(ip) \
sound/soc/sti/uniperif.h
723
SET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
724
UNIPERIF_CTRL_OFFSET(ip), \
sound/soc/sti/uniperif.h
725
UNIPERIF_CTRL_SPDIF_LAT_SHIFT(ip), \
sound/soc/sti/uniperif.h
726
UNIPERIF_CTRL_SPDIF_LAT_MASK(ip), 1)
sound/soc/sti/uniperif.h
727
#define SET_UNIPERIF_CTRL_SPDIF_LAT_OFF(ip) \
sound/soc/sti/uniperif.h
728
SET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
729
UNIPERIF_CTRL_OFFSET(ip), \
sound/soc/sti/uniperif.h
73
readl_relaxed(ip->base + UNIPERIF_CHANNEL_STA_REG0_OFFSET(ip))
sound/soc/sti/uniperif.h
730
UNIPERIF_CTRL_SPDIF_LAT_SHIFT(ip), \
sound/soc/sti/uniperif.h
731
UNIPERIF_CTRL_SPDIF_LAT_MASK(ip), 0)
sound/soc/sti/uniperif.h
734
#define UNIPERIF_CTRL_SPDIF_FMT_SHIFT(ip) 17
sound/soc/sti/uniperif.h
735
#define UNIPERIF_CTRL_SPDIF_FMT_MASK(ip) 0x1
sound/soc/sti/uniperif.h
736
#define GET_UNIPERIF_CTRL_SPDIF_FMT(ip) \
sound/soc/sti/uniperif.h
737
GET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
738
UNIPERIF_CTRL_OFFSET(ip), \
sound/soc/sti/uniperif.h
739
UNIPERIF_CTRL_SPDIF_FMT_SHIFT(ip), \
sound/soc/sti/uniperif.h
74
#define SET_UNIPERIF_CHANNEL_STA_REG0(ip, value) \
sound/soc/sti/uniperif.h
740
UNIPERIF_CTRL_SPDIF_FMT_MASK(ip))
sound/soc/sti/uniperif.h
741
#define SET_UNIPERIF_CTRL_SPDIF_FMT_ON(ip) \
sound/soc/sti/uniperif.h
742
SET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
743
UNIPERIF_CTRL_OFFSET(ip), \
sound/soc/sti/uniperif.h
744
UNIPERIF_CTRL_SPDIF_FMT_SHIFT(ip), \
sound/soc/sti/uniperif.h
745
UNIPERIF_CTRL_SPDIF_FMT_MASK(ip), 1)
sound/soc/sti/uniperif.h
746
#define SET_UNIPERIF_CTRL_SPDIF_FMT_OFF(ip) \
sound/soc/sti/uniperif.h
747
SET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
748
UNIPERIF_CTRL_OFFSET(ip), \
sound/soc/sti/uniperif.h
749
UNIPERIF_CTRL_SPDIF_FMT_SHIFT(ip), \
sound/soc/sti/uniperif.h
75
writel_relaxed(value, ip->base + UNIPERIF_CHANNEL_STA_REG0_OFFSET(ip))
sound/soc/sti/uniperif.h
750
UNIPERIF_CTRL_SPDIF_FMT_MASK(ip), 0)
sound/soc/sti/uniperif.h
753
#define UNIPERIF_CTRL_READER_OUT_SEL_SHIFT(ip) \
sound/soc/sti/uniperif.h
754
((ip)->ver < SND_ST_UNIPERIF_VERSION_UNI_PLR_TOP_1_0 ? 18 : -1)
sound/soc/sti/uniperif.h
755
#define UNIPERIF_CTRL_READER_OUT_SEL_MASK(ip) 0x1
sound/soc/sti/uniperif.h
756
#define GET_UNIPERIF_CTRL_READER_OUT_SEL(ip) \
sound/soc/sti/uniperif.h
757
GET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
758
UNIPERIF_CTRL_OFFSET(ip), \
sound/soc/sti/uniperif.h
759
UNIPERIF_CTRL_READER_OUT_SEL_SHIFT(ip), \
sound/soc/sti/uniperif.h
760
UNIPERIF_CTRL_READER_OUT_SEL_MASK(ip))
sound/soc/sti/uniperif.h
761
#define SET_UNIPERIF_CTRL_READER_OUT_SEL_IN_MEM(ip) \
sound/soc/sti/uniperif.h
762
SET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
763
UNIPERIF_CTRL_OFFSET(ip), \
sound/soc/sti/uniperif.h
764
UNIPERIF_CTRL_READER_OUT_SEL_SHIFT(ip), \
sound/soc/sti/uniperif.h
765
UNIPERIF_CTRL_READER_OUT_SEL_MASK(ip), 0)
sound/soc/sti/uniperif.h
766
#define SET_UNIPERIF_CTRL_READER_OUT_SEL_ON_I2S_LINE(ip) \
sound/soc/sti/uniperif.h
767
SET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
768
UNIPERIF_CTRL_OFFSET(ip), \
sound/soc/sti/uniperif.h
769
UNIPERIF_CTRL_READER_OUT_SEL_SHIFT(ip), \
sound/soc/sti/uniperif.h
77
#define UNIPERIF_CHANNEL_STA_REG1_OFFSET(ip) 0x0064
sound/soc/sti/uniperif.h
770
UNIPERIF_CTRL_READER_OUT_SEL_MASK(ip), 1)
sound/soc/sti/uniperif.h
773
#define UNIPERIF_CTRL_UNDERFLOW_REC_WINDOW_SHIFT(ip) 20
sound/soc/sti/uniperif.h
774
#define UNIPERIF_CTRL_UNDERFLOW_REC_WINDOW_MASK(ip) 0xff
sound/soc/sti/uniperif.h
775
#define GET_UNIPERIF_CTRL_UNDERFLOW_REC_WINDOW(ip) \
sound/soc/sti/uniperif.h
776
GET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
777
UNIPERIF_CTRL_OFFSET(ip), \
sound/soc/sti/uniperif.h
778
UNIPERIF_CTRL_UNDERFLOW_REC_WINDOW_SHIFT(ip), \
sound/soc/sti/uniperif.h
779
UNIPERIF_CTRL_UNDERFLOW_REC_WINDOW_MASK(ip))
sound/soc/sti/uniperif.h
78
#define GET_UNIPERIF_CHANNEL_STA_REG1(ip) \
sound/soc/sti/uniperif.h
780
#define SET_UNIPERIF_CTRL_UNDERFLOW_REC_WINDOW(ip, value) \
sound/soc/sti/uniperif.h
781
SET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
782
UNIPERIF_CTRL_OFFSET(ip), \
sound/soc/sti/uniperif.h
783
UNIPERIF_CTRL_UNDERFLOW_REC_WINDOW_SHIFT(ip), \
sound/soc/sti/uniperif.h
784
UNIPERIF_CTRL_UNDERFLOW_REC_WINDOW_MASK(ip), value)
sound/soc/sti/uniperif.h
79
readl_relaxed(ip->base + UNIPERIF_CHANNEL_STA_REG1_OFFSET(ip))
sound/soc/sti/uniperif.h
790
#define UNIPERIF_I2S_FMT_OFFSET(ip) 0x0048
sound/soc/sti/uniperif.h
791
#define GET_UNIPERIF_I2S_FMT(ip) \
sound/soc/sti/uniperif.h
792
readl_relaxed(ip->base + UNIPERIF_I2S_FMT_OFFSET(ip))
sound/soc/sti/uniperif.h
793
#define SET_UNIPERIF_I2S_FMT(ip, value) \
sound/soc/sti/uniperif.h
794
writel_relaxed(value, ip->base + UNIPERIF_I2S_FMT_OFFSET(ip))
sound/soc/sti/uniperif.h
797
#define UNIPERIF_I2S_FMT_NBIT_SHIFT(ip) 0
sound/soc/sti/uniperif.h
798
#define UNIPERIF_I2S_FMT_NBIT_MASK(ip) 0x1
sound/soc/sti/uniperif.h
799
#define GET_UNIPERIF_I2S_FMT_NBIT(ip) \
sound/soc/sti/uniperif.h
80
#define SET_UNIPERIF_CHANNEL_STA_REG1(ip, value) \
sound/soc/sti/uniperif.h
800
GET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
801
UNIPERIF_I2S_FMT_OFFSET(ip), \
sound/soc/sti/uniperif.h
802
UNIPERIF_I2S_FMT_NBIT_SHIFT(ip), \
sound/soc/sti/uniperif.h
803
UNIPERIF_I2S_FMT_NBIT_MASK(ip))
sound/soc/sti/uniperif.h
804
#define SET_UNIPERIF_I2S_FMT_NBIT_32(ip) \
sound/soc/sti/uniperif.h
805
SET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
806
UNIPERIF_I2S_FMT_OFFSET(ip), \
sound/soc/sti/uniperif.h
807
UNIPERIF_I2S_FMT_NBIT_SHIFT(ip), \
sound/soc/sti/uniperif.h
808
UNIPERIF_I2S_FMT_NBIT_MASK(ip), 0)
sound/soc/sti/uniperif.h
809
#define SET_UNIPERIF_I2S_FMT_NBIT_16(ip) \
sound/soc/sti/uniperif.h
81
writel_relaxed(value, ip->base + UNIPERIF_CHANNEL_STA_REG1_OFFSET(ip))
sound/soc/sti/uniperif.h
810
SET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
811
UNIPERIF_I2S_FMT_OFFSET(ip), \
sound/soc/sti/uniperif.h
812
UNIPERIF_I2S_FMT_NBIT_SHIFT(ip), \
sound/soc/sti/uniperif.h
813
UNIPERIF_I2S_FMT_NBIT_MASK(ip), 1)
sound/soc/sti/uniperif.h
816
#define UNIPERIF_I2S_FMT_DATA_SIZE_SHIFT(ip) 1
sound/soc/sti/uniperif.h
817
#define UNIPERIF_I2S_FMT_DATA_SIZE_MASK(ip) 0x7
sound/soc/sti/uniperif.h
818
#define GET_UNIPERIF_I2S_FMT_DATA_SIZE(ip) \
sound/soc/sti/uniperif.h
819
GET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
820
UNIPERIF_I2S_FMT_OFFSET(ip), \
sound/soc/sti/uniperif.h
821
UNIPERIF_I2S_FMT_DATA_SIZE_SHIFT(ip), \
sound/soc/sti/uniperif.h
822
UNIPERIF_I2S_FMT_DATA_SIZE_MASK(ip))
sound/soc/sti/uniperif.h
823
#define SET_UNIPERIF_I2S_FMT_DATA_SIZE_16(ip) \
sound/soc/sti/uniperif.h
824
SET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
825
UNIPERIF_I2S_FMT_OFFSET(ip), \
sound/soc/sti/uniperif.h
826
UNIPERIF_I2S_FMT_DATA_SIZE_SHIFT(ip), \
sound/soc/sti/uniperif.h
827
UNIPERIF_I2S_FMT_DATA_SIZE_MASK(ip), 0)
sound/soc/sti/uniperif.h
828
#define SET_UNIPERIF_I2S_FMT_DATA_SIZE_18(ip) \
sound/soc/sti/uniperif.h
829
SET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
83
#define UNIPERIF_CHANNEL_STA_REG2_OFFSET(ip) 0x0068
sound/soc/sti/uniperif.h
830
UNIPERIF_I2S_FMT_OFFSET(ip), \
sound/soc/sti/uniperif.h
831
UNIPERIF_I2S_FMT_DATA_SIZE_SHIFT(ip), \
sound/soc/sti/uniperif.h
832
UNIPERIF_I2S_FMT_DATA_SIZE_MASK(ip), 1)
sound/soc/sti/uniperif.h
833
#define SET_UNIPERIF_I2S_FMT_DATA_SIZE_20(ip) \
sound/soc/sti/uniperif.h
834
SET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
835
UNIPERIF_I2S_FMT_OFFSET(ip), \
sound/soc/sti/uniperif.h
836
UNIPERIF_I2S_FMT_DATA_SIZE_SHIFT(ip), \
sound/soc/sti/uniperif.h
837
UNIPERIF_I2S_FMT_DATA_SIZE_MASK(ip), 2)
sound/soc/sti/uniperif.h
838
#define SET_UNIPERIF_I2S_FMT_DATA_SIZE_24(ip) \
sound/soc/sti/uniperif.h
839
SET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
84
#define GET_UNIPERIF_CHANNEL_STA_REG2(ip) \
sound/soc/sti/uniperif.h
840
UNIPERIF_I2S_FMT_OFFSET(ip), \
sound/soc/sti/uniperif.h
841
UNIPERIF_I2S_FMT_DATA_SIZE_SHIFT(ip), \
sound/soc/sti/uniperif.h
842
UNIPERIF_I2S_FMT_DATA_SIZE_MASK(ip), 3)
sound/soc/sti/uniperif.h
843
#define SET_UNIPERIF_I2S_FMTL_DATA_SIZE_28(ip) \
sound/soc/sti/uniperif.h
844
SET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
845
UNIPERIF_I2S_FMT_OFFSET(ip), \
sound/soc/sti/uniperif.h
846
UNIPERIF_I2S_FMT_DATA_SIZE_SHIFT(ip), \
sound/soc/sti/uniperif.h
847
UNIPERIF_I2S_FMT_DATA_SIZE_MASK(ip), 4)
sound/soc/sti/uniperif.h
848
#define SET_UNIPERIF_I2S_FMT_DATA_SIZE_32(ip) \
sound/soc/sti/uniperif.h
849
SET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
85
readl_relaxed(ip->base + UNIPERIF_CHANNEL_STA_REG2_OFFSET(ip))
sound/soc/sti/uniperif.h
850
UNIPERIF_I2S_FMT_OFFSET(ip), \
sound/soc/sti/uniperif.h
851
UNIPERIF_I2S_FMT_DATA_SIZE_SHIFT(ip), \
sound/soc/sti/uniperif.h
852
UNIPERIF_I2S_FMT_DATA_SIZE_MASK(ip), 5)
sound/soc/sti/uniperif.h
855
#define UNIPERIF_I2S_FMT_LR_POL_SHIFT(ip) 4
sound/soc/sti/uniperif.h
856
#define UNIPERIF_I2S_FMT_LR_POL_MASK(ip) 0x1
sound/soc/sti/uniperif.h
857
#define VALUE_UNIPERIF_I2S_FMT_LR_POL_LOW(ip) 0x0
sound/soc/sti/uniperif.h
858
#define VALUE_UNIPERIF_I2S_FMT_LR_POL_HIG(ip) 0x1
sound/soc/sti/uniperif.h
859
#define GET_UNIPERIF_I2S_FMT_LR_POL(ip) \
sound/soc/sti/uniperif.h
86
#define SET_UNIPERIF_CHANNEL_STA_REG2(ip, value) \
sound/soc/sti/uniperif.h
860
GET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
861
UNIPERIF_I2S_FMT_OFFSET(ip), \
sound/soc/sti/uniperif.h
862
UNIPERIF_I2S_FMT_LR_POL_SHIFT(ip), \
sound/soc/sti/uniperif.h
863
UNIPERIF_I2S_FMT_LR_POL_MASK(ip))
sound/soc/sti/uniperif.h
864
#define SET_UNIPERIF_I2S_FMT_LR_POL(ip, value) \
sound/soc/sti/uniperif.h
865
SET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
866
UNIPERIF_I2S_FMT_OFFSET(ip), \
sound/soc/sti/uniperif.h
867
UNIPERIF_I2S_FMT_LR_POL_SHIFT(ip), \
sound/soc/sti/uniperif.h
868
UNIPERIF_I2S_FMT_LR_POL_MASK(ip), value)
sound/soc/sti/uniperif.h
869
#define SET_UNIPERIF_I2S_FMT_LR_POL_LOW(ip) \
sound/soc/sti/uniperif.h
87
writel_relaxed(value, ip->base + UNIPERIF_CHANNEL_STA_REG2_OFFSET(ip))
sound/soc/sti/uniperif.h
870
SET_UNIPERIF_I2S_FMT_LR_POL(ip, \
sound/soc/sti/uniperif.h
871
VALUE_UNIPERIF_I2S_FMT_LR_POL_LOW(ip))
sound/soc/sti/uniperif.h
872
#define SET_UNIPERIF_I2S_FMT_LR_POL_HIG(ip) \
sound/soc/sti/uniperif.h
873
SET_UNIPERIF_I2S_FMT_LR_POL(ip, \
sound/soc/sti/uniperif.h
874
VALUE_UNIPERIF_I2S_FMT_LR_POL_HIG(ip))
sound/soc/sti/uniperif.h
877
#define UNIPERIF_I2S_FMT_SCLK_EDGE_SHIFT(ip) 5
sound/soc/sti/uniperif.h
878
#define UNIPERIF_I2S_FMT_SCLK_EDGE_MASK(ip) 0x1
sound/soc/sti/uniperif.h
879
#define GET_UNIPERIF_I2S_FMT_SCLK_EDGE(ip) \
sound/soc/sti/uniperif.h
880
GET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
881
UNIPERIF_I2S_FMT_OFFSET(ip), \
sound/soc/sti/uniperif.h
882
UNIPERIF_I2S_FMT_SCLK_EDGE_SHIFT(ip), \
sound/soc/sti/uniperif.h
883
UNIPERIF_I2S_FMT_SCLK_EDGE_MASK(ip))
sound/soc/sti/uniperif.h
884
#define SET_UNIPERIF_I2S_FMT_SCLK_EDGE_RISING(ip) \
sound/soc/sti/uniperif.h
885
SET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
886
UNIPERIF_I2S_FMT_OFFSET(ip), \
sound/soc/sti/uniperif.h
887
UNIPERIF_I2S_FMT_SCLK_EDGE_SHIFT(ip), \
sound/soc/sti/uniperif.h
888
UNIPERIF_I2S_FMT_SCLK_EDGE_MASK(ip), 0)
sound/soc/sti/uniperif.h
889
#define SET_UNIPERIF_I2S_FMT_SCLK_EDGE_FALLING(ip) \
sound/soc/sti/uniperif.h
89
#define UNIPERIF_CHANNEL_STA_REG3_OFFSET(ip) 0x006C
sound/soc/sti/uniperif.h
890
SET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
891
UNIPERIF_I2S_FMT_OFFSET(ip), \
sound/soc/sti/uniperif.h
892
UNIPERIF_I2S_FMT_SCLK_EDGE_SHIFT(ip), \
sound/soc/sti/uniperif.h
893
UNIPERIF_I2S_FMT_SCLK_EDGE_MASK(ip), 1)
sound/soc/sti/uniperif.h
896
#define UNIPERIF_I2S_FMT_PADDING_SHIFT(ip) 6
sound/soc/sti/uniperif.h
897
#define UNIPERIF_I2S_FMT_PADDING_MASK(ip) 0x1
sound/soc/sti/uniperif.h
898
#define UNIPERIF_I2S_FMT_PADDING_MASK(ip) 0x1
sound/soc/sti/uniperif.h
899
#define VALUE_UNIPERIF_I2S_FMT_PADDING_I2S_MODE(ip) 0x0
sound/soc/sti/uniperif.h
90
#define GET_UNIPERIF_CHANNEL_STA_REG3(ip) \
sound/soc/sti/uniperif.h
900
#define VALUE_UNIPERIF_I2S_FMT_PADDING_SONY_MODE(ip) 0x1
sound/soc/sti/uniperif.h
901
#define GET_UNIPERIF_I2S_FMT_PADDING(ip) \
sound/soc/sti/uniperif.h
902
GET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
903
UNIPERIF_I2S_FMT_OFFSET(ip), \
sound/soc/sti/uniperif.h
904
UNIPERIF_I2S_FMT_PADDING_SHIFT(ip), \
sound/soc/sti/uniperif.h
905
UNIPERIF_I2S_FMT_PADDING_MASK(ip))
sound/soc/sti/uniperif.h
906
#define SET_UNIPERIF_I2S_FMT_PADDING(ip, value) \
sound/soc/sti/uniperif.h
907
SET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
908
UNIPERIF_I2S_FMT_OFFSET(ip), \
sound/soc/sti/uniperif.h
909
UNIPERIF_I2S_FMT_PADDING_SHIFT(ip), \
sound/soc/sti/uniperif.h
91
readl_relaxed(ip->base + UNIPERIF_CHANNEL_STA_REG3_OFFSET(ip))
sound/soc/sti/uniperif.h
910
UNIPERIF_I2S_FMT_PADDING_MASK(ip), value)
sound/soc/sti/uniperif.h
911
#define SET_UNIPERIF_I2S_FMT_PADDING_I2S_MODE(ip) \
sound/soc/sti/uniperif.h
912
SET_UNIPERIF_I2S_FMT_PADDING(ip, \
sound/soc/sti/uniperif.h
913
VALUE_UNIPERIF_I2S_FMT_PADDING_I2S_MODE(ip))
sound/soc/sti/uniperif.h
914
#define SET_UNIPERIF_I2S_FMT_PADDING_SONY_MODE(ip) \
sound/soc/sti/uniperif.h
915
SET_UNIPERIF_I2S_FMT_PADDING(ip, \
sound/soc/sti/uniperif.h
916
VALUE_UNIPERIF_I2S_FMT_PADDING_SONY_MODE(ip))
sound/soc/sti/uniperif.h
919
#define UNIPERIF_I2S_FMT_ALIGN_SHIFT(ip) 7
sound/soc/sti/uniperif.h
92
#define SET_UNIPERIF_CHANNEL_STA_REG3(ip, value) \
sound/soc/sti/uniperif.h
920
#define UNIPERIF_I2S_FMT_ALIGN_MASK(ip) 0x1
sound/soc/sti/uniperif.h
921
#define GET_UNIPERIF_I2S_FMT_ALIGN(ip) \
sound/soc/sti/uniperif.h
922
GET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
923
UNIPERIF_I2S_FMT_OFFSET(ip), \
sound/soc/sti/uniperif.h
924
UNIPERIF_I2S_FMT_ALIGN_SHIFT(ip), \
sound/soc/sti/uniperif.h
925
UNIPERIF_I2S_FMT_ALIGN_MASK(ip))
sound/soc/sti/uniperif.h
926
#define SET_UNIPERIF_I2S_FMT_ALIGN_LEFT(ip) \
sound/soc/sti/uniperif.h
927
SET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
928
UNIPERIF_I2S_FMT_OFFSET(ip), \
sound/soc/sti/uniperif.h
929
UNIPERIF_I2S_FMT_ALIGN_SHIFT(ip), \
sound/soc/sti/uniperif.h
93
writel_relaxed(value, ip->base + UNIPERIF_CHANNEL_STA_REG3_OFFSET(ip))
sound/soc/sti/uniperif.h
930
UNIPERIF_I2S_FMT_ALIGN_MASK(ip), 0)
sound/soc/sti/uniperif.h
931
#define SET_UNIPERIF_I2S_FMT_ALIGN_RIGHT(ip) \
sound/soc/sti/uniperif.h
932
SET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
933
UNIPERIF_I2S_FMT_OFFSET(ip), \
sound/soc/sti/uniperif.h
934
UNIPERIF_I2S_FMT_ALIGN_SHIFT(ip), \
sound/soc/sti/uniperif.h
935
UNIPERIF_I2S_FMT_ALIGN_MASK(ip), 1)
sound/soc/sti/uniperif.h
938
#define UNIPERIF_I2S_FMT_ORDER_SHIFT(ip) 8
sound/soc/sti/uniperif.h
939
#define UNIPERIF_I2S_FMT_ORDER_MASK(ip) 0x1
sound/soc/sti/uniperif.h
940
#define GET_UNIPERIF_I2S_FMT_ORDER(ip) \
sound/soc/sti/uniperif.h
941
GET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
942
UNIPERIF_I2S_FMT_OFFSET(ip), \
sound/soc/sti/uniperif.h
943
UNIPERIF_I2S_FMT_ORDER_SHIFT(ip), \
sound/soc/sti/uniperif.h
944
UNIPERIF_I2S_FMT_ORDER_MASK(ip))
sound/soc/sti/uniperif.h
945
#define SET_UNIPERIF_I2S_FMT_ORDER_LSB(ip) \
sound/soc/sti/uniperif.h
946
SET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
947
UNIPERIF_I2S_FMT_OFFSET(ip), \
sound/soc/sti/uniperif.h
948
UNIPERIF_I2S_FMT_ORDER_SHIFT(ip), \
sound/soc/sti/uniperif.h
949
UNIPERIF_I2S_FMT_ORDER_MASK(ip), 0)
sound/soc/sti/uniperif.h
95
#define UNIPERIF_CHANNEL_STA_REG4_OFFSET(ip) 0x0070
sound/soc/sti/uniperif.h
950
#define SET_UNIPERIF_I2S_FMT_ORDER_MSB(ip) \
sound/soc/sti/uniperif.h
951
SET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
952
UNIPERIF_I2S_FMT_OFFSET(ip), \
sound/soc/sti/uniperif.h
953
UNIPERIF_I2S_FMT_ORDER_SHIFT(ip), \
sound/soc/sti/uniperif.h
954
UNIPERIF_I2S_FMT_ORDER_MASK(ip), 1)
sound/soc/sti/uniperif.h
957
#define UNIPERIF_I2S_FMT_NUM_CH_SHIFT(ip) 9
sound/soc/sti/uniperif.h
958
#define UNIPERIF_I2S_FMT_NUM_CH_MASK(ip) 0x7
sound/soc/sti/uniperif.h
959
#define GET_UNIPERIF_I2S_FMT_NUM_CH(ip) \
sound/soc/sti/uniperif.h
96
#define GET_UNIPERIF_CHANNEL_STA_REG4(ip) \
sound/soc/sti/uniperif.h
960
GET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
961
UNIPERIF_I2S_FMT_OFFSET(ip), \
sound/soc/sti/uniperif.h
962
UNIPERIF_I2S_FMT_NUM_CH_SHIFT(ip), \
sound/soc/sti/uniperif.h
963
UNIPERIF_I2S_FMT_NUM_CH_MASK(ip))
sound/soc/sti/uniperif.h
964
#define SET_UNIPERIF_I2S_FMT_NUM_CH(ip, value) \
sound/soc/sti/uniperif.h
965
SET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
966
UNIPERIF_I2S_FMT_OFFSET(ip), \
sound/soc/sti/uniperif.h
967
UNIPERIF_I2S_FMT_NUM_CH_SHIFT(ip), \
sound/soc/sti/uniperif.h
968
UNIPERIF_I2S_FMT_NUM_CH_MASK(ip), value)
sound/soc/sti/uniperif.h
97
readl_relaxed(ip->base + UNIPERIF_CHANNEL_STA_REG4_OFFSET(ip))
sound/soc/sti/uniperif.h
971
#define UNIPERIF_I2S_FMT_NO_OF_SAMPLES_TO_READ_SHIFT(ip) 12
sound/soc/sti/uniperif.h
972
#define UNIPERIF_I2S_FMT_NO_OF_SAMPLES_TO_READ_MASK(ip) 0xfffff
sound/soc/sti/uniperif.h
973
#define GET_UNIPERIF_I2S_FMT_NO_OF_SAMPLES_TO_READ(ip) \
sound/soc/sti/uniperif.h
974
GET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
975
UNIPERIF_I2S_FMT_OFFSET(ip), \
sound/soc/sti/uniperif.h
976
UNIPERIF_I2S_FMT_NO_OF_SAMPLES_TO_READ_SHIFT(ip), \
sound/soc/sti/uniperif.h
977
UNIPERIF_I2S_FMT_NO_OF_SAMPLES_TO_READ_MASK(ip))
sound/soc/sti/uniperif.h
978
#define SET_UNIPERIF_I2S_FMT_NO_OF_SAMPLES_TO_READ(ip, value) \
sound/soc/sti/uniperif.h
979
SET_UNIPERIF_REG(ip, \
sound/soc/sti/uniperif.h
98
#define SET_UNIPERIF_CHANNEL_STA_REG4(ip, value) \
sound/soc/sti/uniperif.h
980
UNIPERIF_I2S_FMT_OFFSET(ip), \
sound/soc/sti/uniperif.h
981
UNIPERIF_I2S_FMT_NO_OF_SAMPLES_TO_READ_SHIFT(ip), \
sound/soc/sti/uniperif.h
982
UNIPERIF_I2S_FMT_NO_OF_SAMPLES_TO_READ_MASK(ip), value)
sound/soc/sti/uniperif.h
988
#define UNIPERIF_BIT_CONTROL_OFFSET(ip) \
sound/soc/sti/uniperif.h
989
((ip)->ver < SND_ST_UNIPERIF_VERSION_UNI_PLR_TOP_1_0 ? -1 : 0x004c)
sound/soc/sti/uniperif.h
99
writel_relaxed(value, ip->base + UNIPERIF_CHANNEL_STA_REG4_OFFSET(ip))
sound/soc/sti/uniperif.h
990
#define GET_UNIPERIF_BIT_CONTROL(ip) \
sound/soc/sti/uniperif.h
991
readl_relaxed(ip->base + UNIPERIF_BIT_CONTROL_OFFSET(ip))
sound/soc/sti/uniperif.h
992
#define SET_UNIPERIF_BIT_CONTROL(ip, value) \
sound/soc/sti/uniperif.h
993
writel_relaxed(value, ip->base + UNIPERIF_BIT_CONTROL_OFFSET(ip))
sound/soc/sti/uniperif.h
996
#define UNIPERIF_BIT_CONTROL_CLR_UNDERFLOW_DURATION_SHIFT(ip) 0
sound/soc/sti/uniperif.h
997
#define UNIPERIF_BIT_CONTROL_CLR_UNDERFLOW_DURATION_MASK(ip) 0x1
sound/soc/sti/uniperif.h
998
#define GET_UNIPERIF_BIT_CONTROL_CLR_UNDERFLOW_DURATION(ip) \
sound/soc/sti/uniperif.h
999
GET_UNIPERIF_REG(ip, \
sound/soc/stm/stm32_sai.h
257
#define STM_SAI_IS_F4(ip) ((ip)->conf.version == STM_SAI_STM32F4)
sound/soc/stm/stm32_sai.h
258
#define STM_SAI_HAS_SPDIF_PDM(ip)\
sound/soc/stm/stm32_sai.h
259
((ip)->pdata->conf.has_spdif_pdm)
sound/soc/stm/stm32_sai_sub.c
41
#define STM_SAI_IS_PLAYBACK(ip) ((ip)->dir == SNDRV_PCM_STREAM_PLAYBACK)
sound/soc/stm/stm32_sai_sub.c
42
#define STM_SAI_IS_CAPTURE(ip) ((ip)->dir == SNDRV_PCM_STREAM_CAPTURE)
sound/soc/stm/stm32_sai_sub.c
53
#define STM_SAI_PROTOCOL_IS_SPDIF(ip) ((ip)->spdif)
sound/usb/endpoint.c
1726
struct snd_usb_iface_ref *ip, *in;
sound/usb/endpoint.c
1732
list_for_each_entry_safe(ip, in, &chip->iface_ref_list, list)
sound/usb/endpoint.c
1733
kfree(ip);
sound/usb/endpoint.c
619
struct snd_usb_iface_ref *ip;
sound/usb/endpoint.c
621
list_for_each_entry(ip, &chip->iface_ref_list, list)
sound/usb/endpoint.c
622
if (ip->iface == iface)
sound/usb/endpoint.c
623
return ip;
sound/usb/endpoint.c
625
ip = kzalloc_obj(*ip);
sound/usb/endpoint.c
626
if (!ip)
sound/usb/endpoint.c
628
ip->iface = iface;
sound/usb/endpoint.c
629
list_add_tail(&ip->list, &chip->iface_ref_list);
sound/usb/endpoint.c
630
return ip;
tools/build/feature/test-libunwind-aarch64.c
6
unw_word_t ip,
tools/build/feature/test-libunwind-arm.c
6
unw_word_t ip,
tools/build/feature/test-libunwind-debug-frame-aarch64.c
7
unw_word_t ip, unw_word_t segbase,
tools/build/feature/test-libunwind-debug-frame-arm.c
7
unw_word_t ip, unw_word_t segbase,
tools/build/feature/test-libunwind-debug-frame.c
7
unw_word_t ip, unw_word_t segbase,
tools/build/feature/test-libunwind-x86.c
6
unw_word_t ip,
tools/build/feature/test-libunwind-x86_64.c
6
unw_word_t ip,
tools/build/feature/test-libunwind.c
6
unw_word_t ip,
tools/include/linux/kallsyms.h
24
static inline void print_ip_sym(const char *loglvl, unsigned long ip)
tools/include/linux/kallsyms.h
28
name = backtrace_symbols((void **)&ip, 1);
tools/include/linux/kallsyms.h
35
static inline void print_ip_sym(const char *loglvl, unsigned long ip) {}
tools/include/linux/objtool_types.h
14
u32 ip;
tools/include/uapi/linux/bpf.h
1499
__u64 ip;
tools/include/uapi/linux/netfilter.h
74
__be32 ip;
tools/lib/bpf/bpf_tracing.h
108
#define __PT_IP_REG ip
tools/lib/bpf/bpf_tracing.h
522
#define BPF_KPROBE_READ_RET_IP(ip, ctx) ({ (ip) = (ctx)->link; })
tools/lib/bpf/bpf_tracing.h
527
#define BPF_KPROBE_READ_RET_IP(ip, ctx) ({ (ip) = PT_REGS_RET(ctx); })
tools/lib/bpf/bpf_tracing.h
532
#define BPF_KPROBE_READ_RET_IP(ip, ctx) \
tools/lib/bpf/bpf_tracing.h
533
({ bpf_probe_read_kernel(&(ip), sizeof(ip), (void *)PT_REGS_RET(ctx)); })
tools/lib/bpf/bpf_tracing.h
534
#define BPF_KRETPROBE_READ_RET_IP(ip, ctx) \
tools/lib/bpf/bpf_tracing.h
535
({ bpf_probe_read_kernel(&(ip), sizeof(ip), (void *)(PT_REGS_FP(ctx) + sizeof(ip))); })
tools/lib/bpf/bpf_tracing.h
598
#define BPF_KPROBE_READ_RET_IP(ip, ctx) ({ _Pragma(__BPF_TARGET_MISSING); 0l; })
tools/lib/bpf/bpf_tracing.h
599
#define BPF_KRETPROBE_READ_RET_IP(ip, ctx) ({ _Pragma(__BPF_TARGET_MISSING); 0l; })
tools/lib/bpf/usdt.bpf.h
102
long ip = PT_REGS_IP(ctx);
tools/lib/bpf/usdt.bpf.h
105
spec_id_ptr = bpf_map_lookup_elem(&__bpf_usdt_ip_to_spec_id, &ip);
tools/lib/perf/include/perf/event.h
372
__u64 ip;
tools/perf/arch/powerpc/util/skip-callchain-idx.c
217
u64 ip;
tools/perf/arch/powerpc/util/skip-callchain-idx.c
224
ip = chain->ips[1];
tools/perf/arch/powerpc/util/skip-callchain-idx.c
226
thread__find_symbol(thread, PERF_RECORD_MISC_USER, ip, &al);
tools/perf/arch/powerpc/util/skip-callchain-idx.c
232
pr_debug("%" PRIx64 " dso is NULL\n", ip);
tools/perf/arch/powerpc/util/skip-callchain-idx.c
237
rc = check_return_addr(dso, map__map_ip(al.map, ip));
tools/perf/arch/powerpc/util/skip-callchain-idx.c
240
dso__long_name(dso), al.sym->name, ip, rc);
tools/perf/bench/inject-buildid.c
225
.ip = dso_map_addr(dso),
tools/perf/builtin-inject.c
1019
if (thread__find_map(thread, sample->cpumode, sample->ip, &al)) {
tools/perf/builtin-inject.c
442
if (machine__kernel_ip(machine, node->ip))
tools/perf/builtin-inject.c
447
inject->raw_callchain->ips[i++] = node->ip;
tools/perf/builtin-kmem.c
403
u64 result = sample->ip;
tools/perf/builtin-kmem.c
428
key.start = key.end = node->ip;
tools/perf/builtin-kmem.c
434
addr = map__dso_unmap_ip(node->ms.map, node->ip);
tools/perf/builtin-kmem.c
436
addr = node->ip;
tools/perf/builtin-kmem.c
446
pr_debug2("unknown callsite: %"PRIx64 "\n", sample->ip);
tools/perf/builtin-lock.c
1663
u64 ip;
tools/perf/builtin-lock.c
1669
ip = st->callstack[i];
tools/perf/builtin-lock.c
1670
sym = machine__find_kernel_symbol(con->machine, ip, &kmap);
tools/perf/builtin-lock.c
1671
get_symbol_name_offset(kmap, sym, ip, buf, sizeof(buf));
tools/perf/builtin-lock.c
1672
fprintf(lock_output, "\t\t\t%#lx %s\n", (unsigned long)ip, buf);
tools/perf/builtin-lock.c
1716
u64 ip;
tools/perf/builtin-lock.c
1722
ip = st->callstack[i];
tools/perf/builtin-lock.c
1723
sym = machine__find_kernel_symbol(con->machine, ip, &kmap);
tools/perf/builtin-lock.c
1724
get_symbol_name_offset(kmap, sym, ip, buf, sizeof(buf));
tools/perf/builtin-lock.c
1725
fprintf(lock_output, "%s %#lx %s", i ? ":" : sep, (unsigned long) ip, buf);
tools/perf/builtin-lock.c
827
static int get_symbol_name_offset(struct map *map, struct symbol *sym, u64 ip,
tools/perf/builtin-lock.c
837
offset = map__map_ip(map, ip) - sym->start;
tools/perf/builtin-lock.c
887
if (sym && !machine__is_lock_function(machine, node->ip)) {
tools/perf/builtin-lock.c
888
get_symbol_name_offset(node->ms.map, sym, node->ip,
tools/perf/builtin-lock.c
934
if (node->ms.sym && machine__is_lock_function(machine, node->ip))
tools/perf/builtin-lock.c
937
hash ^= hash_long((unsigned long)node->ip, 64);
tools/perf/builtin-lock.c
956
u64 ip = sample->callchain->ips[i];
tools/perf/builtin-lock.c
958
if (ip >= PERF_CONTEXT_MAX)
tools/perf/builtin-lock.c
961
callstack[c++] = ip;
tools/perf/builtin-mem.c
221
sample->ip,
tools/perf/builtin-sched.c
3036
if (chain->ip >= PERF_CONTEXT_MAX)
tools/perf/builtin-script.c
1224
struct perf_insn *x, uint64_t ip,
tools/perf/builtin-script.c
1230
(uint8_t *)inbuf, inlen, ip, lenp,
tools/perf/builtin-script.c
1236
return fprintf(fp, "%s", dump_insn(x, ip, inbuf, inlen, lenp));
tools/perf/builtin-script.c
1246
static int ip__fprintf_jump(uint64_t ip, struct branch_entry *en,
tools/perf/builtin-script.c
1254
int printed = fprintf(fp, "\t%016" PRIx64 "\t", ip);
tools/perf/builtin-script.c
1256
printed += add_padding(fp, any_dump_insn(evsel, x, ip, inbuf, len, &ilen, fp), 30);
tools/perf/builtin-script.c
1266
thread__find_map(thread, x->cpumode, ip, &al);
tools/perf/builtin-script.c
1417
uint64_t ip = start + off;
tools/perf/builtin-script.c
1419
printed += ip__fprintf_sym(ip, thread, x.cpumode, x.cpu, &lastsym, evsel, fp);
tools/perf/builtin-script.c
1420
if (ip == end) {
tools/perf/builtin-script.c
1423
printed += ip__fprintf_jump(ip, &entries[i], &x, buffer + off, len - off, ++insn, fp,
tools/perf/builtin-script.c
1426
printed += print_srccode(thread, x.cpumode, ip);
tools/perf/builtin-script.c
1430
printed += fprintf(fp, "\t%016" PRIx64 "\t", ip);
tools/perf/builtin-script.c
1431
printed += any_dump_insn(evsel, &x, ip, buffer + off, len - off, &ilen, fp);
tools/perf/builtin-script.c
1438
print_srccode(thread, x.cpumode, ip);
tools/perf/builtin-script.c
1450
if (entries[0].from == sample->ip)
tools/perf/builtin-script.c
1464
end = sample->ip;
tools/perf/builtin-script.c
1473
len = grab_bb(buffer, sample->ip, sample->ip,
tools/perf/builtin-script.c
1478
printed += fprintf(fp, "\t%016" PRIx64 "\t", sample->ip);
tools/perf/builtin-script.c
1479
printed += any_dump_insn(evsel, &x, sample->ip, buffer, len, &ilen, fp);
tools/perf/builtin-script.c
1484
print_srccode(thread, x.cpumode, sample->ip);
tools/perf/builtin-script.c
1498
&& start + off != sample->ip) {
tools/perf/builtin-script.c
1545
u64 *ip)
tools/perf/builtin-script.c
1556
*ip = sample->addr;
tools/perf/builtin-script.c
1558
*ip = sample->addr;
tools/perf/builtin-script.c
1564
*ip = sample->ip;
tools/perf/builtin-script.c
1581
u64 ip = 0;
tools/perf/builtin-script.c
1590
name = resolve_branch_sym(sample, evsel, thread, al, addr_al, &ip);
tools/perf/builtin-script.c
1600
else if (ip)
tools/perf/builtin-script.c
1601
len = fprintf(fp, "%*s%16" PRIx64, (int)depth * 4, "", ip);
tools/perf/builtin-script.c
1869
data->ip, val, str);
tools/perf/builtin-script.c
1907
len = fprintf(fp, " IP: %u ", data->ip);
tools/perf/builtin-script.c
1972
s, data->ip, data->vector);
tools/perf/builtin-script.c
1975
data->type, data->ip, data->vector);
tools/perf/builtin-script.c
2393
u64 ip;
tools/perf/builtin-script.c
2395
&ip);
tools/perf/builtin-timechart.c
520
u64 ip;
tools/perf/builtin-timechart.c
524
ip = chain->ips[i];
tools/perf/builtin-timechart.c
526
ip = chain->ips[chain->nr - i - 1];
tools/perf/builtin-timechart.c
528
if (ip >= PERF_CONTEXT_MAX) {
tools/perf/builtin-timechart.c
529
switch (ip) {
tools/perf/builtin-timechart.c
541
"%"PRId64"\n", (s64) ip);
tools/perf/builtin-timechart.c
555
if (thread__find_symbol(al.thread, cpumode, ip, &tal))
tools/perf/builtin-timechart.c
556
fprintf(f, "..... %016" PRIx64 " %s\n", ip, tal.sym->name);
tools/perf/builtin-timechart.c
558
fprintf(f, "..... %016" PRIx64 "\n", ip);
tools/perf/builtin-top.c
170
static void ui__warn_map_erange(struct map *map, struct symbol *sym, u64 ip)
tools/perf/builtin-top.c
186
ip, dso__long_name(dso), dso__symtab_origin(dso),
tools/perf/builtin-top.c
201
struct evsel *evsel, u64 ip)
tools/perf/builtin-top.c
218
err = hist_entry__inc_addr_samples(he, sample, evsel, ip);
tools/perf/builtin-top.c
230
ui__warn_map_erange(he->ms.map, sym, ip);
tools/perf/builtin-trace.c
3403
thread__find_symbol(thread, sample->cpumode, sample->ip, &al);
tools/perf/dlfilters/dlfilter-test-api-v0.c
112
__s32 (*object_code)(void *ctx, __u64 ip, void *buf, __u32 len);
tools/perf/dlfilters/dlfilter-test-api-v0.c
138
__u64 ip;
tools/perf/dlfilters/dlfilter-test-api-v0.c
170
d->ip = strtoull(dlargv[2], NULL, 0);
tools/perf/dlfilters/dlfilter-test-api-v0.c
188
.ip = d->ip,
tools/perf/dlfilters/dlfilter-test-api-v0.c
204
CHECK_SAMPLE(ip);
tools/perf/dlfilters/dlfilter-test-api-v0.c
267
if (perf_dlfilter_fns.resolve_address(ctx, sample->ip, &address_al))
tools/perf/dlfilters/dlfilter-test-api-v0.c
296
CHECK(perf_dlfilter_fns.object_code(ctx, sample->ip, buf, sizeof(buf)) > 0);
tools/perf/dlfilters/dlfilter-test-api-v0.c
41
__u64 ip;
tools/perf/dlfilters/dlfilter-test-api-v2.c
122
__s32 (*object_code)(void *ctx, __u64 ip, void *buf, __u32 len);
tools/perf/dlfilters/dlfilter-test-api-v2.c
153
__u64 ip;
tools/perf/dlfilters/dlfilter-test-api-v2.c
185
d->ip = strtoull(dlargv[2], NULL, 0);
tools/perf/dlfilters/dlfilter-test-api-v2.c
203
.ip = d->ip,
tools/perf/dlfilters/dlfilter-test-api-v2.c
219
CHECK_SAMPLE(ip);
tools/perf/dlfilters/dlfilter-test-api-v2.c
282
if (perf_dlfilter_fns.resolve_address(ctx, sample->ip, &address_al))
tools/perf/dlfilters/dlfilter-test-api-v2.c
315
CHECK(perf_dlfilter_fns.object_code(ctx, sample->ip, buf, sizeof(buf)) > 0);
tools/perf/dlfilters/dlfilter-test-api-v2.c
47
__u64 ip;
tools/perf/include/perf/perf_dlfilter.h
117
__s32 (*object_code)(void *ctx, __u64 ip, void *buf, __u32 len);
tools/perf/include/perf/perf_dlfilter.h
42
__u64 ip;
tools/perf/scripts/python/Perf-Trace-Util/Context.c
93
if (c->sample->ip && !c->sample->insn_len && thread__maps(c->al->thread)) {
tools/perf/tests/code-reading.c
552
ret = read_object_code(sample.ip, READLEN, sample.cpumode, thread,
tools/perf/tests/dlfilter-test.c
177
.ip = td->ip,
tools/perf/tests/dlfilter-test.c
281
td->ip = MAP_START + td->foo;
tools/perf/tests/dlfilter-test.c
298
td->name, verbose, td->ip, td->addr, do_early);
tools/perf/tests/dlfilter-test.c
47
u64 ip;
tools/perf/tests/dwarf-unwind.c
84
entry->ip);
tools/perf/tests/dwarf-unwind.c
90
symbol, entry->ip, funcs[idx]);
tools/perf/tests/hists_cumulate.c
103
sample.ip = fake_samples[i].ip;
tools/perf/tests/hists_cumulate.c
19
u64 ip;
tools/perf/tests/hists_cumulate.c
28
{ .pid = FAKE_PID_PERF1, .ip = FAKE_IP_KERNEL_SCHEDULE, },
tools/perf/tests/hists_cumulate.c
30
{ .pid = FAKE_PID_PERF1, .ip = FAKE_IP_PERF_MAIN, },
tools/perf/tests/hists_cumulate.c
32
{ .pid = FAKE_PID_PERF1, .ip = FAKE_IP_PERF_CMD_RECORD, },
tools/perf/tests/hists_cumulate.c
34
{ .pid = FAKE_PID_PERF1, .ip = FAKE_IP_LIBC_MALLOC, },
tools/perf/tests/hists_cumulate.c
36
{ .pid = FAKE_PID_PERF1, .ip = FAKE_IP_LIBC_FREE, },
tools/perf/tests/hists_cumulate.c
38
{ .pid = FAKE_PID_PERF2, .ip = FAKE_IP_PERF_MAIN, },
tools/perf/tests/hists_cumulate.c
40
{ .pid = FAKE_PID_PERF2, .ip = FAKE_IP_KERNEL_PAGE_FAULT, },
tools/perf/tests/hists_cumulate.c
42
{ .pid = FAKE_PID_BASH, .ip = FAKE_IP_BASH_MAIN, },
tools/perf/tests/hists_cumulate.c
44
{ .pid = FAKE_PID_BASH, .ip = FAKE_IP_BASH_XMALLOC, },
tools/perf/tests/hists_cumulate.c
46
{ .pid = FAKE_PID_BASH, .ip = FAKE_IP_KERNEL_PAGE_FAULT, },
tools/perf/tests/hists_filter.c
18
u64 ip;
tools/perf/tests/hists_filter.c
28
{ .pid = FAKE_PID_PERF1, .ip = FAKE_IP_KERNEL_SCHEDULE, .socket = 0 },
tools/perf/tests/hists_filter.c
30
{ .pid = FAKE_PID_PERF1, .ip = FAKE_IP_PERF_MAIN, .socket = 0 },
tools/perf/tests/hists_filter.c
32
{ .pid = FAKE_PID_PERF1, .ip = FAKE_IP_LIBC_MALLOC, .socket = 0 },
tools/perf/tests/hists_filter.c
34
{ .pid = FAKE_PID_PERF2, .ip = FAKE_IP_PERF_MAIN, .socket = 0 }, /* will be merged */
tools/perf/tests/hists_filter.c
36
{ .pid = FAKE_PID_PERF2, .ip = FAKE_IP_PERF_CMD_RECORD, .socket = 1 },
tools/perf/tests/hists_filter.c
38
{ .pid = FAKE_PID_PERF2, .ip = FAKE_IP_KERNEL_PAGE_FAULT, .socket = 1 },
tools/perf/tests/hists_filter.c
40
{ .pid = FAKE_PID_BASH, .ip = FAKE_IP_BASH_MAIN, .socket = 2 },
tools/perf/tests/hists_filter.c
42
{ .pid = FAKE_PID_BASH, .ip = FAKE_IP_BASH_XMALLOC, .socket = 2 },
tools/perf/tests/hists_filter.c
44
{ .pid = FAKE_PID_BASH, .ip = FAKE_IP_LIBC_MALLOC, .socket = 3 },
tools/perf/tests/hists_filter.c
46
{ .pid = FAKE_PID_BASH, .ip = FAKE_IP_KERNEL_PAGE_FAULT, .socket = 3 },
tools/perf/tests/hists_filter.c
81
sample.ip = fake_samples[i].ip;
tools/perf/tests/hists_link.c
108
sample.ip = fake_samples[i][k].ip;
tools/perf/tests/hists_link.c
19
u64 ip;
tools/perf/tests/hists_link.c
28
{ .pid = FAKE_PID_PERF1, .ip = FAKE_IP_KERNEL_SCHEDULE, },
tools/perf/tests/hists_link.c
30
{ .pid = FAKE_PID_PERF2, .ip = FAKE_IP_PERF_MAIN, },
tools/perf/tests/hists_link.c
32
{ .pid = FAKE_PID_PERF2, .ip = FAKE_IP_PERF_CMD_RECORD, },
tools/perf/tests/hists_link.c
34
{ .pid = FAKE_PID_BASH, .ip = FAKE_IP_BASH_XMALLOC, },
tools/perf/tests/hists_link.c
36
{ .pid = FAKE_PID_BASH, .ip = FAKE_IP_LIBC_MALLOC, },
tools/perf/tests/hists_link.c
42
{ .pid = FAKE_PID_PERF1, .ip = FAKE_IP_PERF_RUN_COMMAND, },
tools/perf/tests/hists_link.c
44
{ .pid = FAKE_PID_PERF1, .ip = FAKE_IP_LIBC_MALLOC, },
tools/perf/tests/hists_link.c
46
{ .pid = FAKE_PID_PERF1, .ip = FAKE_IP_KERNEL_PAGE_FAULT, },
tools/perf/tests/hists_link.c
48
{ .pid = FAKE_PID_PERF2, .ip = FAKE_IP_KERNEL_SYS_PERF_EVENT_OPEN, },
tools/perf/tests/hists_link.c
50
{ .pid = FAKE_PID_BASH, .ip = FAKE_IP_LIBC_FREE, },
tools/perf/tests/hists_link.c
54
{ .pid = FAKE_PID_PERF2, .ip = FAKE_IP_LIBC_FREE, },
tools/perf/tests/hists_link.c
56
{ .pid = FAKE_PID_BASH, .ip = FAKE_IP_LIBC_MALLOC, }, /* will be merged */
tools/perf/tests/hists_link.c
58
{ .pid = FAKE_PID_BASH, .ip = FAKE_IP_BASH_XFREE, },
tools/perf/tests/hists_link.c
60
{ .pid = FAKE_PID_BASH, .ip = FAKE_IP_LIBC_REALLOC, },
tools/perf/tests/hists_link.c
62
{ .pid = FAKE_PID_BASH, .ip = FAKE_IP_KERNEL_PAGE_FAULT, },
tools/perf/tests/hists_link.c
87
sample.ip = fake_common_samples[k].ip;
tools/perf/tests/hists_output.c
20
u64 ip;
tools/perf/tests/hists_output.c
29
{ .cpu = 0, .pid = FAKE_PID_PERF1, .ip = FAKE_IP_KERNEL_SCHEDULE, },
tools/perf/tests/hists_output.c
31
{ .cpu = 1, .pid = FAKE_PID_PERF1, .ip = FAKE_IP_PERF_MAIN, },
tools/perf/tests/hists_output.c
33
{ .cpu = 1, .pid = FAKE_PID_PERF1, .ip = FAKE_IP_PERF_CMD_RECORD, },
tools/perf/tests/hists_output.c
35
{ .cpu = 1, .pid = FAKE_PID_PERF1, .ip = FAKE_IP_LIBC_MALLOC, },
tools/perf/tests/hists_output.c
37
{ .cpu = 2, .pid = FAKE_PID_PERF1, .ip = FAKE_IP_LIBC_FREE, },
tools/perf/tests/hists_output.c
39
{ .cpu = 2, .pid = FAKE_PID_PERF2, .ip = FAKE_IP_PERF_MAIN, },
tools/perf/tests/hists_output.c
41
{ .cpu = 2, .pid = FAKE_PID_PERF2, .ip = FAKE_IP_KERNEL_PAGE_FAULT, },
tools/perf/tests/hists_output.c
43
{ .cpu = 3, .pid = FAKE_PID_BASH, .ip = FAKE_IP_BASH_MAIN, },
tools/perf/tests/hists_output.c
45
{ .cpu = 0, .pid = FAKE_PID_BASH, .ip = FAKE_IP_BASH_XMALLOC, },
tools/perf/tests/hists_output.c
47
{ .cpu = 1, .pid = FAKE_PID_BASH, .ip = FAKE_IP_KERNEL_PAGE_FAULT, },
tools/perf/tests/hists_output.c
70
sample.ip = fake_samples[i].ip;
tools/perf/tests/sample-parsing.c
247
.ip = 101,
tools/perf/tests/sample-parsing.c
53
COMP(ip);
tools/perf/ui/browsers/hists.c
3160
actions->ms.sym = symbol__new_unresolved(browser->he_selection->ip,
tools/perf/ui/browsers/hists.c
3365
browser->he_selection->ip);
tools/perf/ui/stdio/hist.c
300
if (chain->ip >= PERF_CONTEXT_MAX)
tools/perf/ui/stdio/hist.c
349
if (chain->ip >= PERF_CONTEXT_MAX)
tools/perf/util/annotate-arch/annotate-x86.c
261
u64 ip = ms->sym->start + dl->al.offset;
tools/perf/util/annotate-arch/annotate-x86.c
275
u64 var_addr = annotate_calc_pcrel(dloc->ms, ip,
tools/perf/util/annotate-arch/annotate-x86.c
307
if (get_global_var_type(cu_die, dloc, ip, imm_value, &offset,
tools/perf/util/annotate-arch/annotate-x86.c
458
u64 ip = dloc->ms->sym->start + dl->al.offset;
tools/perf/util/annotate-arch/annotate-x86.c
459
u64 pc = map__rip_2objdump(dloc->ms->map, ip);
tools/perf/util/annotate-arch/annotate-x86.c
475
u64 ip = dloc->ms->sym->start + dl->al.offset;
tools/perf/util/annotate-arch/annotate-x86.c
496
if (!get_global_var_type(cu_die, dloc, ip, var_addr,
tools/perf/util/annotate-arch/annotate-x86.c
622
u64 ip = ms->sym->start + dl->al.offset;
tools/perf/util/annotate-arch/annotate-x86.c
626
addr = annotate_calc_pcrel(ms, ip, src->offset, dl);
tools/perf/util/annotate-arch/annotate-x86.c
628
if (!get_global_var_type(cu_die, dloc, ip, addr, &offset,
tools/perf/util/annotate-arch/annotate-x86.c
647
u64 ip = dloc->ms->sym->start + dl->al.offset;
tools/perf/util/annotate-arch/annotate-x86.c
664
if (get_global_var_type(cu_die, dloc, ip, var_addr,
tools/perf/util/annotate-data.c
1224
if (get_global_var_type(cu_die, dloc, dloc->ip, var_addr,
tools/perf/util/annotate-data.c
1269
u64 pc = map__rip_2objdump(dloc->ms->map, dloc->ip);
tools/perf/util/annotate-data.c
1319
if (get_global_var_type(cu_die, dloc, dloc->ip, addr,
tools/perf/util/annotate-data.c
1331
if (get_global_var_type(cu_die, dloc, dloc->ip, addr,
tools/perf/util/annotate-data.c
1379
if (this_ip == dloc->ip) {
tools/perf/util/annotate-data.c
1424
prev_dst_ip = dst_ip = dloc->ip;
tools/perf/util/annotate-data.c
1528
dloc->ms->sym->name, dloc->ip - dloc->ms->sym->start);
tools/perf/util/annotate-data.c
1535
pc = map__rip_2objdump(dloc->ms->map, dloc->ip);
tools/perf/util/annotate-data.c
1551
if (get_global_var_type(&cu_die, dloc, dloc->ip, dloc->var_addr,
tools/perf/util/annotate-data.c
796
u64 ip, u64 var_addr, int *var_offset,
tools/perf/util/annotate-data.c
830
pc = map__rip_2objdump(dloc->ms->map, ip);
tools/perf/util/annotate-data.h
123
u64 ip;
tools/perf/util/annotate-data.h
259
u64 ip, u64 var_addr, int *var_offset,
tools/perf/util/annotate.c
2625
static struct disasm_line *find_disasm_line(struct symbol *sym, u64 ip,
tools/perf/util/annotate.c
2637
if (sym->start + dl->al.offset == ip) {
tools/perf/util/annotate.c
2644
ip++;
tools/perf/util/annotate.c
2756
u64 annotate_calc_pcrel(struct map_symbol *ms, u64 ip, int offset,
tools/perf/util/annotate.c
2775
addr = ip + (next->al.offset - dl->al.offset) + offset;
tools/perf/util/annotate.c
2835
.ip = ms->sym->start + dl->al.offset,
tools/perf/util/annotate.c
2846
dloc.var_addr = annotate_calc_pcrel(ms, dloc.ip,
tools/perf/util/annotate.c
2903
u64 ip = he->ip;
tools/perf/util/annotate.c
2948
dl = find_disasm_line(ms->sym, ip, /*allow_update=*/true);
tools/perf/util/annotate.c
592
struct evsel *evsel, u64 ip)
tools/perf/util/annotate.c
594
return symbol__inc_addr_samples(&he->ms, evsel, ip, sample);
tools/perf/util/annotate.h
564
u64 annotate_calc_pcrel(struct map_symbol *ms, u64 ip, int offset,
tools/perf/util/arm-spe-decoder/arm-spe-decoder.c
156
u64 payload, ip;
tools/perf/util/arm-spe-decoder/arm-spe-decoder.c
176
ip = arm_spe_calc_ip(idx, payload);
tools/perf/util/arm-spe-decoder/arm-spe-decoder.c
178
decoder->record.from_ip = ip;
tools/perf/util/arm-spe-decoder/arm-spe-decoder.c
180
decoder->record.to_ip = ip;
tools/perf/util/arm-spe-decoder/arm-spe-decoder.c
182
decoder->record.virt_addr = ip;
tools/perf/util/arm-spe-decoder/arm-spe-decoder.c
184
decoder->record.phys_addr = ip;
tools/perf/util/arm-spe-decoder/arm-spe-decoder.c
186
decoder->record.prev_br_tgt = ip;
tools/perf/util/arm-spe.c
278
static inline u8 arm_spe_cpumode(struct arm_spe *spe, u64 ip)
tools/perf/util/arm-spe.c
280
return ip >= spe->kernel_start ?
tools/perf/util/arm-spe.c
376
sample->ip = record->from_ip;
tools/perf/util/arm-spe.c
377
sample->cpumode = arm_spe_cpumode(spe, sample->ip);
tools/perf/util/arm64-frame-pointer-unwind-support.c
34
entries->stack[entries->length++] = entry->ip;
tools/perf/util/auxtrace.c
1278
int code, int cpu, pid_t pid, pid_t tid, u64 ip,
tools/perf/util/auxtrace.c
1293
auxtrace_error->ip = ip;
tools/perf/util/auxtrace.c
1309
int code, int cpu, pid_t pid, pid_t tid, u64 ip,
tools/perf/util/auxtrace.c
1313
ip, msg, timestamp, 0, -1);
tools/perf/util/auxtrace.c
1799
e->cpu, e->pid, e->tid, e->ip, e->code, msg);
tools/perf/util/auxtrace.h
604
int code, int cpu, pid_t pid, pid_t tid, u64 ip,
tools/perf/util/auxtrace.h
608
int code, int cpu, pid_t pid, pid_t tid, u64 ip,
tools/perf/util/bpf_skel/sample_filter.bpf.c
109
return kctx->data->ip;
tools/perf/util/bpf_skel/vmlinux/vmlinux.h
163
u64 ip;
tools/perf/util/build-id.c
72
if (thread__find_map(thread, sample->cpumode, sample->ip, &al))
tools/perf/util/call-path.c
100
if (sym < cp->sym || (sym == cp->sym && ip < cp->ip))
tools/perf/util/call-path.c
106
cp = call_path__new(cpr, parent, sym, ip, in_kernel);
tools/perf/util/call-path.c
15
struct symbol *sym, u64 ip, bool in_kernel)
tools/perf/util/call-path.c
19
cp->ip = sym ? 0 : ip;
tools/perf/util/call-path.c
51
struct symbol *sym, u64 ip,
tools/perf/util/call-path.c
72
call_path__init(cp, parent, sym, ip, in_kernel);
tools/perf/util/call-path.c
79
struct symbol *sym, u64 ip, u64 ks)
tools/perf/util/call-path.c
84
bool in_kernel = ip >= ks;
tools/perf/util/call-path.c
87
ip = 0;
tools/perf/util/call-path.c
90
return call_path__new(cpr, parent, sym, ip, in_kernel);
tools/perf/util/call-path.c
97
if (cp->sym == sym && cp->ip == ip)
tools/perf/util/call-path.h
31
u64 ip;
tools/perf/util/call-path.h
66
struct symbol *sym, u64 ip, u64 ks);
tools/perf/util/callchain.c
1049
callchain_cursor_append(cursor, list->ip, &ms, false, NULL, 0, 0, 0, list->srcline);
tools/perf/util/callchain.c
1089
u64 ip, struct map_symbol *ms,
tools/perf/util/callchain.c
1104
node->ip = ip;
tools/perf/util/callchain.c
1160
al->addr = node->ip;
tools/perf/util/callchain.c
1211
printed = scnprintf(bf, bfsize, "%#" PRIx64, cl->ip);
tools/perf/util/callchain.c
1655
rc = callchain_cursor_append(dst, node->ip, &node->ms,
tools/perf/util/callchain.c
1721
base_chain->ip,
tools/perf/util/callchain.c
1723
pair_chain->ip);
tools/perf/util/callchain.c
602
call->ip = cursor_node->ip;
tools/perf/util/callchain.c
633
cursor_node->ip);
tools/perf/util/callchain.c
772
match = match_chain_dso_addresses(cnode->ms.map, cnode->ip, node->ms.map, node->ip);
tools/perf/util/callchain.c
801
node->ip);
tools/perf/util/callchain.h
121
u64 ip;
tools/perf/util/callchain.h
145
u64 ip;
tools/perf/util/callchain.h
204
int callchain_cursor_append(struct callchain_cursor *cursor, u64 ip,
tools/perf/util/capstone.c
220
uint64_t ip __maybe_unused, int *lenp __maybe_unused,
tools/perf/util/capstone.c
234
count = perf_cs_disasm(cs_handle, code, code_size, ip, 1, &insn);
tools/perf/util/capstone.c
309
u64 ip;
tools/perf/util/capstone.c
318
if (start <= data->ip && data->ip < start + len) {
tools/perf/util/capstone.c
319
data->offset = pgoff + data->ip - start;
tools/perf/util/capstone.c
462
.ip = start,
tools/perf/util/capstone.h
20
uint64_t ip, int *lenp, int print_opts, FILE *fp);
tools/perf/util/capstone.h
33
uint64_t ip __maybe_unused,
tools/perf/util/cs-etm.c
1531
sample->ip);
tools/perf/util/cs-etm.c
1536
cs_etm__mem_access(etmq, trace_chan_id, sample->ip, sample->insn_len,
tools/perf/util/cs-etm.c
1579
sample.ip = addr;
tools/perf/util/cs-etm.c
1621
struct perf_sample sample = {.ip = 0,};
tools/perf/util/cs-etm.c
1628
u64 ip;
tools/perf/util/cs-etm.c
1630
ip = cs_etm__last_executed_instr(tidq->prev_packet);
tools/perf/util/cs-etm.c
1633
event->sample.header.misc = cs_etm__cpu_mode(etmq, ip,
tools/perf/util/cs-etm.c
1640
sample.ip = ip;
tools/perf/util/cs-etm.c
1662
.from = sample.ip,
tools/perf/util/data-convert-bt.c
623
ret = value_set_u64_hex(cw, event, "perf_ip", sample->ip);
tools/perf/util/data-convert-json.c
132
u64 ip, struct addr_location *al)
tools/perf/util/data-convert-json.c
138
output_json_key_format(out, false, 5, "ip", "\"0x%" PRIx64 "\"", ip);
tools/perf/util/data-convert-json.c
209
u64 ip = sample->callchain->ips[i];
tools/perf/util/data-convert-json.c
212
if (ip >= PERF_CONTEXT_MAX) {
tools/perf/util/data-convert-json.c
213
switch (ip) {
tools/perf/util/data-convert-json.c
225
PRId64 "\n", (s64) ip);
tools/perf/util/data-convert-json.c
237
ok = thread__find_symbol(al.thread, cpumode, ip, &tal);
tools/perf/util/data-convert-json.c
238
output_sample_callchain_entry(tool, ip, ok ? &tal : NULL);
tools/perf/util/data-convert-json.c
242
output_sample_callchain_entry(tool, sample->ip, &al);
tools/perf/util/db-export.c
257
al.addr = node->ip;
tools/perf/util/db-export.c
267
al.sym, node->ip,
tools/perf/util/dlfilter.c
108
d_al->is_kernel_ip = machine__kernel_ip(d->machine, d->sample->ip);
tools/perf/util/dlfilter.c
227
if (d->sample->ip && !d->sample->insn_len) {
tools/perf/util/dlfilter.c
287
static __s32 code_read(__u64 ip, struct map *map, struct machine *machine, void *buf, __u32 len)
tools/perf/util/dlfilter.c
289
u64 offset = map__map_ip(map, ip);
tools/perf/util/dlfilter.c
291
if (ip + len >= map__end(map))
tools/perf/util/dlfilter.c
292
len = map__end(map) - ip;
tools/perf/util/dlfilter.c
297
static __s32 dlfilter__object_code(void *ctx, __u64 ip, void *buf, __u32 len)
tools/perf/util/dlfilter.c
311
if (al->map && ip >= map__start(al->map) && ip < map__end(al->map) &&
tools/perf/util/dlfilter.c
312
machine__kernel_ip(d->machine, ip) == machine__kernel_ip(d->machine, d->sample->ip))
tools/perf/util/dlfilter.c
313
return code_read(ip, al->map, d->machine, buf, len);
tools/perf/util/dlfilter.c
317
thread__find_map_fb(al->thread, d->sample->cpumode, ip, &a);
tools/perf/util/dlfilter.c
318
ret = a.map ? code_read(ip, a.map, d->machine, buf, len) : -1;
tools/perf/util/dlfilter.c
519
ASSIGN(ip);
tools/perf/util/dso.c
1877
u64 ip;
tools/perf/util/dso.c
1886
if (start <= data->ip && data->ip < start + len) {
tools/perf/util/dso.c
1887
data->offset = pgoff + data->ip - start;
tools/perf/util/dso.c
1901
.ip = start,
tools/perf/util/dump-insn.c
9
u64 ip __maybe_unused, u8 *inbuf __maybe_unused,
tools/perf/util/dump-insn.h
22
const char *dump_insn(struct perf_insn *x, u64 ip,
tools/perf/util/event.c
819
thread__find_map(thread, sample->cpumode, sample->ip, al);
tools/perf/util/event.h
136
u32 ip : 1,
tools/perf/util/event.h
177
u32 ip : 1,
tools/perf/util/event.h
237
ip : 1,
tools/perf/util/evsel.c
3195
data->ip = data->callchain->ips[1];
tools/perf/util/evsel.c
3260
data->ip = *array;
tools/perf/util/evsel_fprintf.c
158
addr = map__map_ip(map, node->ip);
tools/perf/util/evsel_fprintf.c
161
printed += fprintf(fp, "%c%16" PRIx64, s, node->ip);
tools/perf/util/evsel_fprintf.c
172
sample->deferred_cookie == node->ip) {
tools/perf/util/evsel_fprintf.c
236
printed += fprintf(fp, "%16" PRIx64, sample->ip);
tools/perf/util/hist.c
1214
return left->ip != right->ip;
tools/perf/util/hist.c
1233
.ip = al->addr,
tools/perf/util/hist.c
821
.ip = al->addr,
tools/perf/util/hist.h
253
u64 ip;
tools/perf/util/intel-bts.c
265
static inline u8 intel_bts_cpumode(struct intel_bts *bts, uint64_t ip)
tools/perf/util/intel-bts.c
267
return machine__kernel_ip(bts->machine, ip) ?
tools/perf/util/intel-bts.c
285
sample.ip = le64_to_cpu(branch->from);
tools/perf/util/intel-bts.c
286
sample.cpumode = intel_bts_cpumode(bts, sample.ip);
tools/perf/util/intel-bts.c
320
static int intel_bts_get_next_insn(struct intel_bts_queue *btsq, u64 ip)
tools/perf/util/intel-bts.c
333
len = thread__memcpy(thread, machine, buf, ip, INTEL_PT_INSN_BUF_SZ, &x86_64);
tools/perf/util/intel-bts.c
347
pid_t tid, u64 ip)
tools/perf/util/intel-bts.c
353
INTEL_BTS_ERR_NOINSN, cpu, pid, tid, ip,
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
1126
decoder->ip != decoder->last_ip)
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
1136
err = decoder->walk_insn(&intel_pt_insn, &insn_cnt, &decoder->ip,
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
1143
intel_pt_log_at("ERROR: Unexpected branch at FUP instruction", decoder->ip);
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
1155
decoder->ip += intel_pt_insn.length;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
1159
struct intel_pt_insn *intel_pt_insn, uint64_t ip)
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
1169
err = decoder->walk_insn(intel_pt_insn, &insn_cnt, &decoder->ip, ip,
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
1181
decoder->ip);
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
1187
if (ip && decoder->ip == ip) {
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
1197
decoder->state.from_ip = decoder->ip;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
1199
decoder->ip += intel_pt_insn->length;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
120
uint64_t *insn_cnt_ptr, uint64_t *ip, uint64_t to_ip,
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
1208
err = intel_pt_push(&decoder->stack, decoder->ip +
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
122
bool (*pgd_ip)(uint64_t ip, void *data);
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
1220
decoder->state.from_ip = decoder->ip;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
1221
decoder->ip += intel_pt_insn->length +
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
1223
decoder->state.to_ip = decoder->ip;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
1326
bool ip = decoder->set_fup_cfe_ip;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
1331
if (!ip && decoder->pge)
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
1338
if (ip || decoder->pge)
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
1397
decoder->state.to_ip = decoder->ip;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
1402
decoder->state.from_ip = decoder->ip;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
1414
uint64_t ip, int err)
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
1418
ip == decoder->ip + intel_pt_insn->length;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
1424
uint64_t ip;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
1427
ip = decoder->last_ip;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
1430
err = intel_pt_walk_insn(decoder, &intel_pt_insn, ip);
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
1434
intel_pt_fup_with_nlip(decoder, &intel_pt_insn, ip, err)) {
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
1448
decoder->ip);
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
1455
decoder->ip);
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
1497
decoder->state.from_ip = decoder->ip;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
1502
decoder->ip = decoder->last_ip;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
1507
decoder->state.from_ip = decoder->ip;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
1512
decoder->ip = decoder->last_ip;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
1519
uint64_t to_ip = decoder->ip + intel_pt_insn.length +
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
1529
decoder->ip = to_ip;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
153
uint64_t ip;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
1530
decoder->state.from_ip = decoder->ip;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
1536
decoder->ip);
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
1623
decoder->state.from_ip = decoder->ip;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
1650
decoder->ip);
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
1656
decoder->ip);
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
1662
decoder->ip);
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
1672
decoder->state.from_ip = decoder->ip;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
1673
decoder->ip = decoder->ret_addr;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
1674
decoder->state.to_ip = decoder->ip;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
1686
decoder->ip);
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
1692
decoder->state.from_ip = decoder->ip;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
1694
decoder->ip = decoder->last_ip;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
1708
decoder->state.from_ip = decoder->ip;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
1709
decoder->ip += intel_pt_insn.length +
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
1711
decoder->state.to_ip = decoder->ip;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
1718
decoder->state.from_ip = decoder->ip;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
1720
decoder->ip += intel_pt_insn.length;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
1724
decoder->ip += intel_pt_insn.length;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
1863
decoder->state.from_ip = decoder->ip;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
1864
decoder->ip = 0;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
2237
decoder->state.from_ip = decoder->ip;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
2242
decoder->state.to_ip = decoder->ip;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
2253
decoder->ip);
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
2259
decoder->state.to_ip = decoder->ip;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
2267
decoder->state.from_ip = decoder->ip;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
2272
decoder->state.to_ip = decoder->ip;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
2311
decoder->state.from_ip = decoder->ip;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
2961
decoder->state.to_ip = decoder->ip;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
2972
decoder->state.from_ip = decoder->ip;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
2988
decoder->state.from_ip = decoder->ip;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
3013
decoder->ip = decoder->psb_ip;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
3163
if (decoder->ip != decoder->last_ip) {
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
3260
decoder->state.to_ip = decoder->ip;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
3289
decoder->ip = decoder->last_ip;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
3346
decoder->state.from_ip = decoder->ip;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
3418
decoder->state.from_ip = decoder->ip;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
3436
decoder->state.from_ip = decoder->ip;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
3456
decoder->state.from_ip = decoder->ip;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
3462
decoder->state.from_ip = decoder->ip;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
3477
decoder->state.from_ip = decoder->ip;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
3575
uint64_t current_ip = decoder->ip;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
3578
decoder->psb_ip = decoder->ip;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
3581
decoder->ip);
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
3626
if (decoder->ip)
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
3674
if (!decoder->ip)
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
3684
if (!decoder->ip)
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
3693
if (!decoder->ip)
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
3700
if (decoder->ip)
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
3803
if (err || ((decoder->state.type & INTEL_PT_PSB_EVT) && !decoder->ip))
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
3813
decoder->state.to_ip = decoder->ip;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
3814
intel_pt_log_to("Setting IP", decoder->ip);
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
3915
decoder->ip = 0;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
3942
if (decoder->ip) {
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
3984
decoder->ip = 0;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
4030
decoder->state.from_ip = decoder->ip;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
4039
decoder->state.from_ip = decoder->ip;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
408
return stack->blk->ip[--stack->pos];
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
430
static int intel_pt_push(struct intel_pt_stack *stack, uint64_t ip)
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
440
stack->blk->ip[stack->pos++] = ip;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
517
uint64_t ip;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
52
uint64_t ip[INTEL_PT_BLK_SIZE];
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
521
ip = (last_ip & (uint64_t)0xffffffffffff0000ULL) |
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
525
ip = (last_ip & (uint64_t)0xffffffff00000000ULL) |
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
529
ip = packet->payload;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
531
if (ip & (uint64_t)0x800000000000ULL)
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
532
ip |= (uint64_t)0xffff000000000000ULL;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
535
ip = (last_ip & (uint64_t)0xffff000000000000ULL) |
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
539
ip = packet->payload;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
545
return ip;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
557
decoder->ip = decoder->last_ip;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.c
628
decoder->ip = 0;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.h
187
uint64_t ip;
tools/perf/util/intel-pt-decoder/intel-pt-decoder.h
276
uint64_t *insn_cnt_ptr, uint64_t *ip, uint64_t to_ip,
tools/perf/util/intel-pt-decoder/intel-pt-decoder.h
278
bool (*pgd_ip)(uint64_t ip, void *data);
tools/perf/util/intel-pt-decoder/intel-pt-insn-decoder.c
218
const char *dump_insn(struct perf_insn *x, uint64_t ip __maybe_unused,
tools/perf/util/intel-pt-decoder/intel-pt-log.c
225
void __intel_pt_log_insn(struct intel_pt_insn *intel_pt_insn, uint64_t ip)
tools/perf/util/intel-pt-decoder/intel-pt-log.c
235
intel_pt_print_data(intel_pt_insn->buf, len, ip, 8);
tools/perf/util/intel-pt-decoder/intel-pt-log.c
243
uint64_t ip)
tools/perf/util/intel-pt-decoder/intel-pt-log.c
250
intel_pt_print_no_data(ip, 8);
tools/perf/util/intel-pt-decoder/intel-pt-log.h
27
void __intel_pt_log_insn(struct intel_pt_insn *intel_pt_insn, uint64_t ip);
tools/perf/util/intel-pt-decoder/intel-pt-log.h
29
uint64_t ip);
tools/perf/util/intel-pt.c
1000
return intel_pt_match_pgd_ip(ptq->pt, ip, ip, NULL);
tools/perf/util/intel-pt.c
1010
if (!thread__find_map(thread, cpumode, ip, &al) || !map__dso(al.map))
tools/perf/util/intel-pt.c
1013
offset = map__map_ip(al.map, ip);
tools/perf/util/intel-pt.c
1015
res = intel_pt_match_pgd_ip(ptq->pt, ip, offset, dso__long_name(map__dso(al.map)));
tools/perf/util/intel-pt.c
1020
static bool intel_pt_pgd_ip(uint64_t ip, void *data)
tools/perf/util/intel-pt.c
1022
return __intel_pt_pgd_ip(ip, data) > 0;
tools/perf/util/intel-pt.c
1245
pt->synth_opts.callchain_sz + 1, sample->ip,
tools/perf/util/intel-pt.c
1283
pt->br_stack_sz, sample->ip,
tools/perf/util/intel-pt.c
1722
sample->ip = ptq->state->from_ip;
tools/perf/util/intel-pt.c
1724
sample->cpumode = intel_pt_cpumode(ptq, sample->ip, sample->addr);
tools/perf/util/intel-pt.c
1798
.from = sample.ip,
tools/perf/util/intel-pt.c
1829
sample->ip, pt->kernel_start);
tools/perf/util/intel-pt.c
1940
if (!sample->ip)
tools/perf/util/intel-pt.c
1948
struct perf_sample sample = { .ip = 0, };
tools/perf/util/intel-pt.c
1960
raw.ip = !!(ptq->state->flags & INTEL_PT_FUP_IP);
tools/perf/util/intel-pt.c
2110
raw.ip = !!(ptq->state->flags & INTEL_PT_FUP_IP);
tools/perf/util/intel-pt.c
2441
sample.ip = items->ip;
tools/perf/util/intel-pt.c
2443
sample.ip = items->rip;
tools/perf/util/intel-pt.c
2445
sample.ip = ptq->state->from_ip;
tools/perf/util/intel-pt.c
2447
cpumode = intel_pt_cpumode(ptq, sample.ip, 0);
tools/perf/util/intel-pt.c
2467
pt->synth_opts.callchain_sz, sample.ip,
tools/perf/util/intel-pt.c
2627
raw.cfe.ip = !!(ptq->state->flags & INTEL_PT_FUP_IP);
tools/perf/util/intel-pt.c
2685
pid_t pid, pid_t tid, u64 ip, u64 timestamp,
tools/perf/util/intel-pt.c
2706
code, cpu, pid, tid, ip, msg, timestamp,
tools/perf/util/intel-pt.c
2771
static inline bool intel_pt_is_switch_ip(struct intel_pt_queue *ptq, u64 ip)
tools/perf/util/intel-pt.c
2775
return ip == pt->switch_ip &&
tools/perf/util/intel-pt.c
2966
u64 ip, switch_ip = 0;
tools/perf/util/intel-pt.c
2984
ip = map__unmap_ip(map, sym->start);
tools/perf/util/intel-pt.c
2985
if (ip >= map__start(map) && ip < map__end(map)) {
tools/perf/util/intel-pt.c
2986
switch_ip = ip;
tools/perf/util/intel-pt.c
3002
ip = map__unmap_ip(map, sym->start);
tools/perf/util/intel-pt.c
3003
if (ip >= map__start(map) && ip < map__end(map)) {
tools/perf/util/intel-pt.c
3004
*ptss_ip = ip;
tools/perf/util/intel-pt.c
669
static inline bool intel_pt_guest_kernel_ip(uint64_t ip)
tools/perf/util/intel-pt.c
672
return ip & (1ULL << 63);
tools/perf/util/intel-pt.c
675
static inline u8 intel_pt_nr_cpumode(struct intel_pt_queue *ptq, uint64_t ip, bool nr)
tools/perf/util/intel-pt.c
678
return intel_pt_guest_kernel_ip(ip) ?
tools/perf/util/intel-pt.c
683
return ip >= ptq->pt->kernel_start ?
tools/perf/util/intel-pt.c
749
uint64_t *insn_cnt_ptr, uint64_t *ip,
tools/perf/util/intel-pt.c
771
if (to_ip && *ip == to_ip)
tools/perf/util/intel-pt.c
775
cpumode = intel_pt_nr_cpumode(ptq, *ip, nr);
tools/perf/util/intel-pt.c
815
if (!thread__find_map(thread, cpumode, *ip, &al) || !map__dso(al.map)) {
tools/perf/util/intel-pt.c
817
intel_pt_log("ERROR: thread has no dso for %#" PRIx64 "\n", *ip);
tools/perf/util/intel-pt.c
819
intel_pt_log("ERROR: thread has no map for %#" PRIx64 "\n", *ip);
tools/perf/util/intel-pt.c
832
offset = map__map_ip(al.map, *ip);
tools/perf/util/intel-pt.c
841
*ip += e->byte_cnt;
tools/perf/util/intel-pt.c
848
intel_pt_log_insn_no_data(intel_pt_insn, *ip);
tools/perf/util/intel-pt.c
855
start_ip = *ip;
tools/perf/util/intel-pt.c
880
intel_pt_log_insn(intel_pt_insn, *ip);
tools/perf/util/intel-pt.c
900
*ip += intel_pt_insn->length;
tools/perf/util/intel-pt.c
902
if (to_ip && *ip == to_ip) {
tools/perf/util/intel-pt.c
908
if (*ip >= map__end(al.map))
tools/perf/util/intel-pt.c
935
*ip - start_ip, intel_pt_insn);
tools/perf/util/intel-pt.c
947
static bool intel_pt_match_pgd_ip(struct intel_pt *pt, uint64_t ip,
tools/perf/util/intel-pt.c
968
ip, offset, filename ? filename : "[kernel]",
tools/perf/util/intel-pt.c
980
ip, offset, filename ? filename : "[kernel]");
tools/perf/util/intel-pt.c
985
static int __intel_pt_pgd_ip(uint64_t ip, void *data)
tools/perf/util/intel-pt.c
995
if (intel_pt_guest_kernel_ip(ip))
tools/perf/util/intel-pt.c
996
return intel_pt_match_pgd_ip(ptq->pt, ip, ip, NULL);
tools/perf/util/intel-pt.c
999
} else if (ip >= ptq->pt->kernel_start) {
tools/perf/util/jitdump.c
526
sample.ip = addr;
tools/perf/util/jitdump.c
637
sample.ip = jr->move.new_code_addr;
tools/perf/util/lock-contention.c
106
u64 ip;
tools/perf/util/lock-contention.c
132
ip = callstack[i];
tools/perf/util/lock-contention.c
133
sym = machine__find_kernel_symbol(machine, ip, &kmap);
tools/perf/util/machine.c
2002
u64 ip)
tools/perf/util/machine.c
2014
thread__find_cpumode_addr_location(thread, ip, /*symbols=*/true, &al);
tools/perf/util/machine.c
2016
ams->addr = ip;
tools/perf/util/machine.c
2056
ip__resolve_ams(al->thread, mem_info__iaddr(mi), sample->ip);
tools/perf/util/machine.c
2065
static char *callchain_srcline(struct map_symbol *ms, u64 ip)
tools/perf/util/machine.c
2075
srcline = srcline__tree_find(dso__srclines(dso), ip);
tools/perf/util/machine.c
2080
srcline = get_srcline(dso, map__rip_2objdump(map, ip),
tools/perf/util/machine.c
2081
ms->sym, show_sym, show_addr, ip);
tools/perf/util/machine.c
2082
srcline__tree_insert(dso__srclines(dso), ip, srcline);
tools/perf/util/machine.c
2093
static int append_inlines(struct callchain_cursor *cursor, struct map_symbol *ms, u64 ip,
tools/perf/util/machine.c
2110
addr = map__dso_map_ip(map, ip);
tools/perf/util/machine.c
2129
ret = callchain_cursor_append(cursor, ip, &ilist_ms,
tools/perf/util/machine.c
2133
ret = callchain_cursor_append(cursor, ip, &ilist_ms, false,
tools/perf/util/machine.c
2151
u64 ip,
tools/perf/util/machine.c
2169
thread__find_cpumode_addr_location(thread, ip, symbols, &al);
tools/perf/util/machine.c
2171
if (ip >= PERF_CONTEXT_MAX) {
tools/perf/util/machine.c
2172
switch (ip) {
tools/perf/util/machine.c
2185
"%"PRId64"\n", (s64) ip);
tools/perf/util/machine.c
2197
thread__find_symbol(thread, *cpumode, ip, &al);
tools/perf/util/machine.c
2199
thread__find_map(thread, *cpumode, ip, &al);
tools/perf/util/machine.c
2227
if (append_inlines(cursor, &ms, ip, branch, flags, nr_loop_iter,
tools/perf/util/machine.c
2232
err = callchain_cursor_append(cursor, ip, &ms,
tools/perf/util/machine.c
2408
u64 ip;
tools/perf/util/machine.c
2430
if (sample->ip)
tools/perf/util/machine.c
2431
ip = sample->ip;
tools/perf/util/machine.c
2433
ip = entries[0].to;
tools/perf/util/machine.c
2437
root_al, &cpumode, ip,
tools/perf/util/machine.c
2459
ip = entries[i].from;
tools/perf/util/machine.c
2462
root_al, &cpumode, ip,
tools/perf/util/machine.c
2474
ip = entries[i].from;
tools/perf/util/machine.c
2477
root_al, &cpumode, ip,
tools/perf/util/machine.c
2490
if (sample->ip)
tools/perf/util/machine.c
2491
ip = sample->ip;
tools/perf/util/machine.c
2493
ip = entries[0].to;
tools/perf/util/machine.c
2497
root_al, &cpumode, ip,
tools/perf/util/machine.c
2518
err = callchain_cursor_append(cursor, cnode->ip,
tools/perf/util/machine.c
2758
u64 ip = chain->ips[ent];
tools/perf/util/machine.c
2760
if (ip >= PERF_CONTEXT_MAX) {
tools/perf/util/machine.c
2762
root_al, cpumode, ip,
tools/perf/util/machine.c
2901
u64 ip;
tools/perf/util/machine.c
2912
ip = chain->ips[j];
tools/perf/util/machine.c
2913
if (ip < PERF_CONTEXT_MAX)
tools/perf/util/machine.c
2941
if (leaf_frame_caller && leaf_frame_caller != ip) {
tools/perf/util/machine.c
2952
root_al, &cpumode, ip,
tools/perf/util/machine.c
2966
u64 addr = entry->ip;
tools/perf/util/machine.c
2971
if (append_inlines(cursor, &entry->ms, entry->ip, /*branch=*/false, /*branch_flags=*/NULL,
tools/perf/util/machine.c
2980
addr = map__dso_map_ip(entry->ms.map, entry->ip);
tools/perf/util/machine.c
2983
return callchain_cursor_append(cursor, entry->ip, &entry->ms,
tools/perf/util/machine.h
100
return ip >= kernel_start;
tools/perf/util/machine.h
96
static inline bool machine__kernel_ip(struct machine *machine, u64 ip)
tools/perf/util/map.c
574
u64 map__objdump_2mem(const struct map *map, u64 ip)
tools/perf/util/map.c
579
return map__unmap_ip(map, ip);
tools/perf/util/map.c
582
return map__unmap_ip(map, ip + map__pgoff(map));
tools/perf/util/map.c
585
return map__unmap_ip(map, ip - dso__text_offset(dso));
tools/perf/util/map.c
587
return ip + map__reloc(map);
tools/perf/util/map.c
591
u64 map__objdump_2rip(const struct map *map, u64 ip)
tools/perf/util/map.c
596
return ip;
tools/perf/util/map.c
599
return ip + map__pgoff(map);
tools/perf/util/map.c
602
return ip - dso__text_offset(dso);
tools/perf/util/map.c
604
return map__map_ip(map, ip + map__reloc(map));
tools/perf/util/map.c
609
u64 ip = map__unmap_ip(map, sym->start);
tools/perf/util/map.c
611
return ip >= map__start(map) && ip < map__end(map);
tools/perf/util/map.h
108
static inline u64 map__dso_map_ip(const struct map *map, u64 ip)
tools/perf/util/map.h
110
return ip - map__start(map) + map__pgoff(map);
tools/perf/util/map.h
139
u64 map__objdump_2mem(const struct map *map, u64 ip);
tools/perf/util/map.h
142
u64 map__objdump_2rip(const struct map *map, u64 ip);
tools/perf/util/maps.c
1118
const u64 ip = *(const u64 *)key;
tools/perf/util/maps.c
1121
if (ip < map__start(map))
tools/perf/util/maps.c
1123
if (ip >= map__end(map))
tools/perf/util/maps.c
1128
struct map *maps__find(struct maps *maps, u64 ip)
tools/perf/util/maps.c
1142
mapp = bsearch(&ip, maps_by_address, nr_maps, sizeof(*mapp),
tools/perf/util/powerpc-vpadtl.c
186
sample.ip = be64_to_cpu(record->srr0);
tools/perf/util/print_insn.c
47
uint64_t ip, int *lenp, int print_opts, FILE *fp)
tools/perf/util/print_insn.c
50
ip, lenp, print_opts, fp);
tools/perf/util/print_insn.c
62
sample->ip, NULL, 0, fp);
tools/perf/util/print_insn.h
20
uint64_t ip, int *lenp, int print_opts, FILE *fp);
tools/perf/util/python.c
49
sample_member_def(sample_ip, ip, T_ULONGLONG, "event ip"), \
tools/perf/util/s390-cpumsf.c
508
.ip = basic->ia,
tools/perf/util/s390-cpumsf.c
538
__func__, pos, sample.ip, basic->P, basic->CL, sample.pid,
tools/perf/util/s390-cpumsf.c
889
pid_t pid, pid_t tid, u64 ip, u64 timestamp)
tools/perf/util/s390-cpumsf.c
897
code, cpu, pid, tid, ip, msg, timestamp);
tools/perf/util/sample.c
104
sample->ip, len,
tools/perf/util/sample.c
98
if (!sample->ip || sample->insn_len != 0)
tools/perf/util/sample.h
85
u64 ip;
tools/perf/util/scripting-engines/trace-event-perl.c
294
if (!hv_stores(elem, "ip", newSVuv(node->ip))) {
tools/perf/util/scripting-engines/trace-event-python.c
1322
tuple_set_d64(t, 8, es->sample->ip);
tools/perf/util/scripting-engines/trace-event-python.c
1391
tuple_set_d64(t, 3, cp->ip);
tools/perf/util/scripting-engines/trace-event-python.c
1635
tuple_set_u64(t, 5, e->ip);
tools/perf/util/scripting-engines/trace-event-python.c
429
PyLong_FromUnsignedLongLong(node->ip));
tools/perf/util/scripting-engines/trace-event-python.c
452
node_al.addr = map__map_ip(map, node->ip);
tools/perf/util/scripting-engines/trace-event-python.c
865
PyLong_FromUnsignedLongLong(sample->ip));
tools/perf/util/session.c
1133
event->header.misc, sample->pid, sample->tid, sample->ip,
tools/perf/util/session.c
599
event->auxtrace_error.ip = bswap_64(event->auxtrace_error.ip);
tools/perf/util/session.c
861
static const char *callchain_context_str(u64 ip)
tools/perf/util/session.c
863
switch (ip) {
tools/perf/util/sort.c
1352
u64 ip, char level, char *bf, size_t size,
tools/perf/util/sort.c
1364
ip - map__unmap_ip(map, sym->start));
tools/perf/util/sort.c
1369
offs = ip - sym->start;
tools/perf/util/sort.c
1376
len, ip);
tools/perf/util/sort.c
2322
u64 left_ip = left->ip;
tools/perf/util/sort.c
2323
u64 right_ip = right->ip;
tools/perf/util/sort.c
2338
u64 ip = he->ip;
tools/perf/util/sort.c
2342
ip = map__unmap_ip(map, ip);
tools/perf/util/sort.c
2344
return repsep_snprintf(bf, size, "%-#*llx", width, ip);
tools/perf/util/sort.c
368
return _sort__addr_cmp(left->ip, right->ip);
tools/perf/util/sort.c
393
u64 ip, char level, char *bf, size_t size,
tools/perf/util/sort.c
403
u64 rip = ip;
tools/perf/util/sort.c
406
rip = map__unmap_ip(map, ip);
tools/perf/util/sort.c
417
ip - map__unmap_ip(map, sym->start));
tools/perf/util/sort.c
429
len, ip);
tools/perf/util/sort.c
437
return _hist_entry__sym_snprintf(&he->ms, he->ip,
tools/perf/util/sort.c
471
return left->ip - right->ip;
tools/perf/util/sort.c
483
return left->ip - right->ip;
tools/perf/util/sort.c
492
return repsep_snprintf(bf, size, "[%c] %-#.*llx", he->level, width - 4, he->ip);
tools/perf/util/sort.c
494
return repsep_snprintf(bf, size, "[%c] %s+0x%llx", he->level, sym->name, he->ip - sym->start);
tools/perf/util/sort.c
510
return map__srcline(he->ms.map, he->ip, he->ms.sym);
tools/perf/util/sort.c
518
ret = _sort__addr_cmp(left->ip, right->ip);
tools/perf/util/sort.c
825
sf = __get_srcline(map__dso(map), map__rip_2objdump(map, e->ip),
tools/perf/util/sort.c
826
e->ms.sym, false, true, true, e->ip);
tools/perf/util/srcline.c
262
u64 ip)
tools/perf/util/srcline.c
303
ip - sym->start) < 0)
tools/perf/util/srcline.c
352
bool show_sym, bool show_addr, u64 ip)
tools/perf/util/srcline.c
354
return __get_srcline(dso, addr, sym, show_sym, show_addr, false, ip);
tools/perf/util/srcline.h
14
bool show_sym, bool show_addr, u64 ip);
tools/perf/util/srcline.h
17
u64 ip);
tools/perf/util/symbol.c
365
const u64 ip = sym->start;
tools/perf/util/symbol.c
383
if (ip < s->start)
tools/perf/util/symbol.c
399
static struct symbol *symbols__find(struct rb_root_cached *symbols, u64 ip)
tools/perf/util/symbol.c
411
if (ip < s->start)
tools/perf/util/symbol.c
413
else if (ip > s->end || (ip == s->end && ip != s->start))
tools/perf/util/synthetic-events.c
1626
*array = sample->ip;
tools/perf/util/thread-stack.c
1033
ret_addr = sample->ip + sample->insn_len;
tools/perf/util/thread-stack.c
1154
if (!sample->ip || !sample->addr)
tools/perf/util/thread-stack.c
1157
ret_addr = sample->ip + sample->insn_len;
tools/perf/util/thread-stack.c
1188
if (!sample->ip)
tools/perf/util/thread-stack.c
516
static inline u64 callchain_context(u64 ip, u64 kernel_start)
tools/perf/util/thread-stack.c
518
return ip < kernel_start ? PERF_CONTEXT_USER : PERF_CONTEXT_KERNEL;
tools/perf/util/thread-stack.c
523
size_t sz, u64 ip, u64 kernel_start)
tools/perf/util/thread-stack.c
526
u64 context = callchain_context(ip, kernel_start);
tools/perf/util/thread-stack.c
536
chain->ips[1] = ip;
tools/perf/util/thread-stack.c
546
ip = ts->stack[ts->cnt - j].ret_addr;
tools/perf/util/thread-stack.c
547
context = callchain_context(ip, kernel_start);
tools/perf/util/thread-stack.c
554
chain->ips[i] = ip;
tools/perf/util/thread-stack.c
570
u64 last_context, context, ip;
tools/perf/util/thread-stack.c
587
ip = ts->stack[ts->cnt - j].ret_addr;
tools/perf/util/thread-stack.c
588
context = callchain_context(ip, kernel_start);
tools/perf/util/thread-stack.c
590
(context == sample_context && ip == sample_ip))
tools/perf/util/thread-stack.c
597
ip = ts->stack[ts->cnt - j].ret_addr;
tools/perf/util/thread-stack.c
598
context = callchain_context(ip, kernel_start);
tools/perf/util/thread-stack.c
605
chain->ips[nr] = ip;
tools/perf/util/thread-stack.c
680
u64 ip, u64 kernel_start)
tools/perf/util/thread-stack.c
703
if (ip < kernel_start) {
tools/perf/util/thread-stack.c
732
if (ks_start(s, ip, kernel_start, &start, nb)) {
tools/perf/util/thread-stack.c
741
if (ks_start(s, ip, kernel_start, &start, nb)) {
tools/perf/util/thread-stack.c
867
u64 ip;
tools/perf/util/thread-stack.c
869
if (sample->ip) {
tools/perf/util/thread-stack.c
870
ip = sample->ip;
tools/perf/util/thread-stack.c
873
ip = sample->addr;
tools/perf/util/thread-stack.c
879
cp = call_path__findnew(cpr, &cpr->call_path, sym, ip,
tools/perf/util/thread-stack.c
882
return thread_stack__push_cp(ts, ip, sample->time, ref, cp,
tools/perf/util/thread-stack.c
917
u64 ip = sample->ip;
tools/perf/util/thread-stack.c
920
if (ip >= ks && addr < ks) {
tools/perf/util/thread-stack.c
932
} else if (thread_stack__in_kernel(ts) && ip < ks) {
tools/perf/util/thread-stack.c
988
cp = call_path__findnew(cpr, cp, fsym, ip, ks);
tools/perf/util/thread-stack.c
990
err = thread_stack__push_cp(ts, ip, tm, ref, cp, true, false);
tools/perf/util/thread-stack.h
87
size_t sz, u64 ip, u64 kernel_start);
tools/perf/util/thread-stack.h
89
struct ip_callchain *chain, size_t sz, u64 ip,
tools/perf/util/thread.c
566
void *buf, u64 ip, int len, bool *is64bit)
tools/perf/util/thread.c
573
if (machine__kernel_ip(machine, ip))
tools/perf/util/thread.c
577
if (!thread__find_map(thread, cpumode, ip, &al)) {
tools/perf/util/thread.c
589
offset = map__map_ip(al.map, ip);
tools/perf/util/thread.h
136
void *buf, u64 ip, int len, bool *is64bit);
tools/perf/util/unwind-libdw.c
133
return mod && dwfl_addrmodule(ui->dwfl, ip) == mod ? 0 : -1;
tools/perf/util/unwind-libdw.c
136
static int report_module(u64 ip, struct unwind_info *ui)
tools/perf/util/unwind-libdw.c
142
res = __report_module(&al, ip, ui);
tools/perf/util/unwind-libdw.c
151
static int entry(u64 ip, struct unwind_info *ui)
tools/perf/util/unwind-libdw.c
158
if (__report_module(&al, ip, ui)) {
tools/perf/util/unwind-libdw.c
163
e->ip = ip;
tools/perf/util/unwind-libdw.c
170
ip,
tools/perf/util/unwind-libdw.c
171
al.map ? map__map_ip(al.map, ip) : (u64) 0);
tools/perf/util/unwind-libdw.c
355
Dwarf_Word ip;
tools/perf/util/unwind-libdw.c
395
err = perf_reg_value(&ip, data->user_regs, perf_arch_reg_ip(e_machine));
tools/perf/util/unwind-libdw.c
399
err = report_module(ip, ui);
tools/perf/util/unwind-libdw.c
422
err = ui->entries[j].ip ? ui->cb(&ui->entries[j], ui->arg) : 0;
tools/perf/util/unwind-libdw.c
69
static int __report_module(struct addr_location *al, u64 ip,
tools/perf/util/unwind-libdw.c
79
thread__find_symbol(ui->thread, PERF_RECORD_MISC_USER, ip, al);
tools/perf/util/unwind-libdw.c
99
mod = dwfl_addrmodule(ui->dwfl, ip);
tools/perf/util/unwind-libunwind-local.c
433
static struct map *find_map(unw_word_t ip, struct unwind_info *ui)
tools/perf/util/unwind-libunwind-local.c
439
thread__find_map(ui->thread, PERF_RECORD_MISC_USER, ip, &al);
tools/perf/util/unwind-libunwind-local.c
446
find_proc_info(unw_addr_space_t as, unw_word_t ip, unw_proc_info_t *pi,
tools/perf/util/unwind-libunwind-local.c
456
map = find_map(ip, ui);
tools/perf/util/unwind-libunwind-local.c
46
unw_word_t ip,
tools/perf/util/unwind-libunwind-local.c
478
ret = dwarf_search_unwind_table(as, ip, &di, pi,
tools/perf/util/unwind-libunwind-local.c
500
if (dwarf_find_debug_frame(0, &di, ip, base, symfile, start, map__end(map)))
tools/perf/util/unwind-libunwind-local.c
501
ret = dwarf_search_unwind_table(as, ip, &di, pi,
tools/perf/util/unwind-libunwind-local.c
55
unw_word_t ip,
tools/perf/util/unwind-libunwind-local.c
660
static int entry(u64 ip, struct thread *thread,
tools/perf/util/unwind-libunwind-local.c
668
e.ms.sym = thread__find_symbol(thread, PERF_RECORD_MISC_USER, ip, &al);
tools/perf/util/unwind-libunwind-local.c
669
e.ip = ip;
tools/perf/util/unwind-libunwind-local.c
675
ip,
tools/perf/util/unwind-libunwind-local.c
676
al.map ? map__map_ip(al.map, ip) : (u64) 0);
tools/perf/util/unwind.h
15
u64 ip;
tools/testing/selftests/bpf/prog_tests/assign_reuse.c
116
fd_sv = start_reuseport_server(family, sotype, ip, port, 100, 1);
tools/testing/selftests/bpf/prog_tests/assign_reuse.c
82
void run_assign_reuse(int family, int sotype, const char *ip, __u16 port)
tools/testing/selftests/bpf/prog_tests/build_id.c
27
printf("IP = %llx", (unsigned long long)stack[i].ip);
tools/testing/selftests/bpf/prog_tests/cls_redirect.c
181
.ip = {
tools/testing/selftests/bpf/prog_tests/cls_redirect.c
212
struct iphdr ip;
tools/testing/selftests/bpf/prog_tests/cls_redirect.c
241
ip = (struct iphdr){
tools/testing/selftests/bpf/prog_tests/cls_redirect.c
249
p = mempcpy(p, &ip, sizeof(ip));
tools/testing/selftests/bpf/prog_tests/lwt_helpers.h
44
struct iphdr *ip = (struct iphdr *)buf;
tools/testing/selftests/bpf/prog_tests/lwt_helpers.h
45
struct icmphdr *icmp = (struct icmphdr *)(ip + 1);
tools/testing/selftests/bpf/prog_tests/lwt_helpers.h
46
ssize_t min_header_len = sizeof(*ip) + sizeof(*icmp);
tools/testing/selftests/bpf/prog_tests/lwt_helpers.h
51
if (ip->protocol != IPPROTO_ICMP)
tools/testing/selftests/bpf/prog_tests/lwt_redirect.c
77
char ip[256];
tools/testing/selftests/bpf/prog_tests/lwt_redirect.c
83
snprintf(ip, sizeof(ip), "10.0.0.%d", link_index);
tools/testing/selftests/bpf/prog_tests/lwt_redirect.c
85
snprintf(ip, sizeof(ip), "20.0.0.%d", link_index);
tools/testing/selftests/bpf/prog_tests/lwt_redirect.c
89
ip, ICMP_PAYLOAD_SIZE);
tools/testing/selftests/bpf/prog_tests/lwt_reroute.c
191
char ip[256];
tools/testing/selftests/bpf/prog_tests/lwt_reroute.c
205
snprintf(ip, 256, "10.0.0.%d", ifindex);
tools/testing/selftests/bpf/prog_tests/lwt_reroute.c
208
ping_once(ip);
tools/testing/selftests/bpf/prog_tests/lwt_reroute.c
224
char ip[256];
tools/testing/selftests/bpf/prog_tests/lwt_reroute.c
236
snprintf(ip, 256, "10.0.0.%d", ifindex);
tools/testing/selftests/bpf/prog_tests/lwt_reroute.c
237
ASSERT_EQ(overflow_fq(10, ip), 0, "overflow_fq");
tools/testing/selftests/bpf/prog_tests/lwt_reroute.c
65
static void ping_once(const char *ip)
tools/testing/selftests/bpf/prog_tests/lwt_reroute.c
69
ip, ICMP_PAYLOAD_SIZE);
tools/testing/selftests/bpf/prog_tests/perf_skip.c
107
skel->bss->ip = (uintptr_t)test_function;
tools/testing/selftests/bpf/prog_tests/perf_skip.c
114
skel->bss->ip = 0;
tools/testing/selftests/bpf/prog_tests/sk_lookup.c
1188
server_fd = make_server(SOCK_STREAM, t->listen_at.ip,
tools/testing/selftests/bpf/prog_tests/sk_lookup.c
155
static int make_server(int sotype, const char *ip, int port,
tools/testing/selftests/bpf/prog_tests/sk_lookup.c
159
.family = is_ipv6(ip) ? AF_INET6 : AF_INET,
tools/testing/selftests/bpf/prog_tests/sk_lookup.c
170
fd = start_server_str(cb_opts.family, sotype, ip, port, &opts);
tools/testing/selftests/bpf/prog_tests/sk_lookup.c
550
server_fds[i] = make_server(t->sotype, t->listen_at.ip,
tools/testing/selftests/bpf/prog_tests/sk_lookup.c
573
reuse_conn_fd = make_server(t->sotype, t->listen_at.ip,
tools/testing/selftests/bpf/prog_tests/sk_lookup.c
585
client_fd = connect_to_addr_str(is_ipv6(t->connect_to.ip) ? AF_INET6 : AF_INET,
tools/testing/selftests/bpf/prog_tests/sk_lookup.c
586
t->sotype, t->connect_to.ip, t->connect_to.port, NULL);
tools/testing/selftests/bpf/prog_tests/sk_lookup.c
61
const char *ip;
tools/testing/selftests/bpf/prog_tests/sk_lookup.c
802
int family = is_ipv6(t->connect_to.ip) ? AF_INET6 : AF_INET;
tools/testing/selftests/bpf/prog_tests/sk_lookup.c
813
server_fd = make_server(t->sotype, t->listen_at.ip, t->listen_at.port,
tools/testing/selftests/bpf/prog_tests/sk_lookup.c
822
err = make_sockaddr(family, t->connect_to.ip, t->connect_to.port, &dst, &len);
tools/testing/selftests/bpf/prog_tests/sk_lookup.c
85
static bool is_ipv6(const char *ip)
tools/testing/selftests/bpf/prog_tests/sk_lookup.c
87
return !!strchr(ip, ':');
tools/testing/selftests/bpf/prog_tests/sk_lookup.c
920
int family = is_ipv6(t->connect_to.ip) ? AF_INET6 : AF_INET;
tools/testing/selftests/bpf/prog_tests/sk_lookup.c
931
server1 = make_server(t->sotype, t->listen_at.ip, t->listen_at.port,
tools/testing/selftests/bpf/prog_tests/sk_lookup.c
941
server2 = make_server(t->sotype, t->connect_to.ip, t->connect_to.port,
tools/testing/selftests/bpf/prog_tests/sk_lookup.c
950
err = make_sockaddr(family, t->connect_to.ip, t->connect_to.port, &dst, &len);
tools/testing/selftests/bpf/prog_tests/sock_post_bind.c
125
.ip = "::1",
tools/testing/selftests/bpf/prog_tests/sock_post_bind.c
155
.ip = "127.0.0.1",
tools/testing/selftests/bpf/prog_tests/sock_post_bind.c
185
.ip = "127.0.0.1",
tools/testing/selftests/bpf/prog_tests/sock_post_bind.c
20
const char *ip;
tools/testing/selftests/bpf/prog_tests/sock_post_bind.c
216
.ip = "127.0.0.1",
tools/testing/selftests/bpf/prog_tests/sock_post_bind.c
247
.ip = "::1",
tools/testing/selftests/bpf/prog_tests/sock_post_bind.c
262
.ip = "0.0.0.0",
tools/testing/selftests/bpf/prog_tests/sock_post_bind.c
275
.ip = "::",
tools/testing/selftests/bpf/prog_tests/sock_post_bind.c
305
static int bind_sock(int domain, int type, const char *ip,
tools/testing/selftests/bpf/prog_tests/sock_post_bind.c
326
if (inet_pton(domain, ip, (void *)&addr4->sin_addr) != 1)
tools/testing/selftests/bpf/prog_tests/sock_post_bind.c
333
if (inet_pton(domain, ip, (void *)&addr6->sin6_addr) != 1)
tools/testing/selftests/bpf/prog_tests/sock_post_bind.c
388
res = bind_sock(test->domain, test->type, test->ip, test->port,
tools/testing/selftests/bpf/prog_tests/sock_post_bind.c
83
.ip = "0.0.0.0",
tools/testing/selftests/bpf/prog_tests/sock_post_bind.c
96
.ip = "::",
tools/testing/selftests/bpf/prog_tests/uprobe.c
214
skel->bss->ip = (unsigned long) uprobe_regs_change_ip_2;
tools/testing/selftests/bpf/prog_tests/vrf_socket_lookup.c
202
const char *ip, int port, bool test_xdp, bool tcp_skc,
tools/testing/selftests/bpf/prog_tests/vrf_socket_lookup.c
209
fd = make_socket(sotype, ip, port, &addr);
tools/testing/selftests/bpf/prog_tests/vrf_socket_lookup.c
62
static int make_socket(int sotype, const char *ip, int port,
tools/testing/selftests/bpf/prog_tests/vrf_socket_lookup.c
67
err = make_sockaddr(AF_INET, ip, port, addr, NULL);
tools/testing/selftests/bpf/prog_tests/vrf_socket_lookup.c
84
static int make_server(int sotype, const char *ip, int port, const char *ifname)
tools/testing/selftests/bpf/prog_tests/vrf_socket_lookup.c
88
fd = start_server(AF_INET, sotype, ip, port, TIMEOUT_MS);
tools/testing/selftests/bpf/progs/skb_pkt_end.c
17
struct iphdr *ip = NULL;
tools/testing/selftests/bpf/progs/skb_pkt_end.c
24
ip = (void *)(eth + 1);
tools/testing/selftests/bpf/progs/skb_pkt_end.c
27
return ip;
tools/testing/selftests/bpf/progs/skb_pkt_end.c
33
struct iphdr *ip = NULL;
tools/testing/selftests/bpf/progs/skb_pkt_end.c
39
if (!(ip = get_iphdr(skb)))
tools/testing/selftests/bpf/progs/skb_pkt_end.c
42
proto = ip->protocol;
tools/testing/selftests/bpf/progs/skb_pkt_end.c
47
tcp = (void*)(ip + 1);
tools/testing/selftests/bpf/progs/test_bpf_nf.c
177
saddr.ip = bpf_get_prandom_u32();
tools/testing/selftests/bpf/progs/test_bpf_nf.c
180
daddr.ip = bpf_get_prandom_u32();
tools/testing/selftests/bpf/progs/test_bpf_nf.c
194
if (tuple->dst.u3.ip == saddr.ip &&
tools/testing/selftests/bpf/progs/test_bpf_nf.c
197
if (tuple->src.u3.ip == daddr.ip &&
tools/testing/selftests/bpf/progs/test_bpf_nf.c
279
saddr.ip = bpf_get_prandom_u32();
tools/testing/selftests/bpf/progs/test_bpf_nf.c
282
daddr.ip = bpf_get_prandom_u32();
tools/testing/selftests/bpf/progs/test_cls_redirect.c
208
static INLINING bool ipv4_is_fragment(const struct iphdr *ip)
tools/testing/selftests/bpf/progs/test_cls_redirect.c
210
uint16_t frag_off = ip->frag_off & bpf_htons(IP_OFFSET_MASK);
tools/testing/selftests/bpf/progs/test_cls_redirect.c
211
return (ip->frag_off & bpf_htons(IP_MF)) != 0 || frag_off > 0;
tools/testing/selftests/bpf/progs/test_cls_redirect.c
515
encap_gre->ip.protocol = IPPROTO_GRE;
tools/testing/selftests/bpf/progs/test_cls_redirect.c
516
encap_gre->ip.daddr = next_hop->s_addr;
tools/testing/selftests/bpf/progs/test_cls_redirect.c
517
encap_gre->ip.saddr = ENCAPSULATION_IP;
tools/testing/selftests/bpf/progs/test_cls_redirect.c
518
encap_gre->ip.tot_len =
tools/testing/selftests/bpf/progs/test_cls_redirect.c
519
bpf_htons(bpf_ntohs(encap_gre->ip.tot_len) + delta);
tools/testing/selftests/bpf/progs/test_cls_redirect.c
522
pkt_ipv4_checksum((void *)&encap_gre->ip);
tools/testing/selftests/bpf/progs/test_cls_redirect.c
547
uint32_t old_saddr = encap->ip.saddr;
tools/testing/selftests/bpf/progs/test_cls_redirect.c
548
encap->ip.saddr = encap->ip.daddr;
tools/testing/selftests/bpf/progs/test_cls_redirect.c
549
encap->ip.daddr = next_hop->s_addr;
tools/testing/selftests/bpf/progs/test_cls_redirect.c
555
const uint64_t off = offsetof(typeof(*encap), ip.check);
tools/testing/selftests/bpf/progs/test_cls_redirect.c
969
if (encap->ip.ihl != 5) {
tools/testing/selftests/bpf/progs/test_cls_redirect.c
974
if (encap->ip.daddr != ENCAPSULATION_IP ||
tools/testing/selftests/bpf/progs/test_cls_redirect.c
975
encap->ip.protocol != IPPROTO_UDP) {
tools/testing/selftests/bpf/progs/test_cls_redirect.c
987
if (ipv4_is_fragment((void *)&encap->ip)) {
tools/testing/selftests/bpf/progs/test_cls_redirect.h
53
struct iphdr ip;
tools/testing/selftests/bpf/progs/test_cls_redirect.h
59
struct iphdr ip;
tools/testing/selftests/bpf/progs/test_cls_redirect_dynptr.c
119
static bool ipv4_is_fragment(const struct iphdr *ip)
tools/testing/selftests/bpf/progs/test_cls_redirect_dynptr.c
121
uint16_t frag_off = ip->frag_off & bpf_htons(IP_OFFSET_MASK);
tools/testing/selftests/bpf/progs/test_cls_redirect_dynptr.c
122
return (ip->frag_off & bpf_htons(IP_MF)) != 0 || frag_off > 0;
tools/testing/selftests/bpf/progs/test_cls_redirect_dynptr.c
405
encap_gre->ip.protocol = IPPROTO_GRE;
tools/testing/selftests/bpf/progs/test_cls_redirect_dynptr.c
406
encap_gre->ip.daddr = next_hop->s_addr;
tools/testing/selftests/bpf/progs/test_cls_redirect_dynptr.c
407
encap_gre->ip.saddr = ENCAPSULATION_IP;
tools/testing/selftests/bpf/progs/test_cls_redirect_dynptr.c
408
encap_gre->ip.tot_len =
tools/testing/selftests/bpf/progs/test_cls_redirect_dynptr.c
409
bpf_htons(bpf_ntohs(encap_gre->ip.tot_len) + delta);
tools/testing/selftests/bpf/progs/test_cls_redirect_dynptr.c
412
pkt_ipv4_checksum((void *)&encap_gre->ip);
tools/testing/selftests/bpf/progs/test_cls_redirect_dynptr.c
441
uint32_t old_saddr = encap->ip.saddr;
tools/testing/selftests/bpf/progs/test_cls_redirect_dynptr.c
442
encap->ip.saddr = encap->ip.daddr;
tools/testing/selftests/bpf/progs/test_cls_redirect_dynptr.c
443
encap->ip.daddr = next_hop->s_addr;
tools/testing/selftests/bpf/progs/test_cls_redirect_dynptr.c
449
const uint64_t off = offsetof(typeof(*encap), ip.check);
tools/testing/selftests/bpf/progs/test_cls_redirect_dynptr.c
874
if (encap->ip.ihl != 5)
tools/testing/selftests/bpf/progs/test_cls_redirect_dynptr.c
878
if (encap->ip.daddr != ENCAPSULATION_IP ||
tools/testing/selftests/bpf/progs/test_cls_redirect_dynptr.c
879
encap->ip.protocol != IPPROTO_UDP)
tools/testing/selftests/bpf/progs/test_cls_redirect_dynptr.c
889
if (ipv4_is_fragment((void *)&encap->ip)) {
tools/testing/selftests/bpf/progs/test_lwt_seg6local.c
59
struct ip6_t *ip;
tools/testing/selftests/bpf/progs/test_lwt_seg6local.c
72
ip = cursor_advance(cursor, sizeof(*ip));
tools/testing/selftests/bpf/progs/test_lwt_seg6local.c
73
if ((void *)ip + sizeof(*ip) > data_end)
tools/testing/selftests/bpf/progs/test_lwt_seg6local.c
76
if (ip->next_header != 43)
tools/testing/selftests/bpf/progs/test_migrate_reuseport.c
57
struct iphdr *ip = (struct iphdr *)(eth + 1);
tools/testing/selftests/bpf/progs/test_migrate_reuseport.c
59
if (ip + 1 > data_end)
tools/testing/selftests/bpf/progs/test_migrate_reuseport.c
62
if (ip->protocol != IPPROTO_TCP)
tools/testing/selftests/bpf/progs/test_migrate_reuseport.c
65
tcp = (struct tcphdr *)((void *)ip + ip->ihl * 4);
tools/testing/selftests/bpf/progs/test_perf_skip.c
12
return ip != PT_REGS_IP(&data->regs);
tools/testing/selftests/bpf/progs/test_perf_skip.c
6
uintptr_t ip;
tools/testing/selftests/bpf/progs/test_seg6_loop.c
59
struct ip6_t *ip;
tools/testing/selftests/bpf/progs/test_seg6_loop.c
72
ip = cursor_advance(cursor, sizeof(*ip));
tools/testing/selftests/bpf/progs/test_seg6_loop.c
73
if ((void *)ip + sizeof(*ip) > data_end)
tools/testing/selftests/bpf/progs/test_seg6_loop.c
76
if (ip->next_header != 43)
tools/testing/selftests/bpf/progs/test_tc_tunnel.c
148
olen = sizeof(h_outer.ip);
tools/testing/selftests/bpf/progs/test_tc_tunnel.c
223
h_outer.ip = iph_inner;
tools/testing/selftests/bpf/progs/test_tc_tunnel.c
224
h_outer.ip.tot_len = bpf_htons(olen +
tools/testing/selftests/bpf/progs/test_tc_tunnel.c
225
bpf_ntohs(h_outer.ip.tot_len));
tools/testing/selftests/bpf/progs/test_tc_tunnel.c
226
h_outer.ip.protocol = encap_proto;
tools/testing/selftests/bpf/progs/test_tc_tunnel.c
228
set_ipv4_csum((void *)&h_outer.ip);
tools/testing/selftests/bpf/progs/test_tc_tunnel.c
279
olen = sizeof(h_outer.ip);
tools/testing/selftests/bpf/progs/test_tc_tunnel.c
352
h_outer.ip = iph_inner;
tools/testing/selftests/bpf/progs/test_tc_tunnel.c
353
h_outer.ip.payload_len = bpf_htons(olen +
tools/testing/selftests/bpf/progs/test_tc_tunnel.c
354
bpf_ntohs(h_outer.ip.payload_len));
tools/testing/selftests/bpf/progs/test_tc_tunnel.c
356
h_outer.ip.nexthdr = encap_proto;
tools/testing/selftests/bpf/progs/test_tc_tunnel.c
387
olen = sizeof(h_outer.ip);
tools/testing/selftests/bpf/progs/test_tc_tunnel.c
396
h_outer.ip.version = 6;
tools/testing/selftests/bpf/progs/test_tc_tunnel.c
397
h_outer.ip.hop_limit = iph_inner.ttl;
tools/testing/selftests/bpf/progs/test_tc_tunnel.c
398
h_outer.ip.saddr.in6_u.u6_addr8[1] = 0xfd;
tools/testing/selftests/bpf/progs/test_tc_tunnel.c
399
h_outer.ip.saddr.in6_u.u6_addr8[15] = 1;
tools/testing/selftests/bpf/progs/test_tc_tunnel.c
400
h_outer.ip.daddr.in6_u.u6_addr8[1] = 0xfd;
tools/testing/selftests/bpf/progs/test_tc_tunnel.c
401
h_outer.ip.daddr.in6_u.u6_addr8[15] = 2;
tools/testing/selftests/bpf/progs/test_tc_tunnel.c
402
h_outer.ip.payload_len = iph_inner.tot_len;
tools/testing/selftests/bpf/progs/test_tc_tunnel.c
403
h_outer.ip.nexthdr = IPPROTO_IPIP;
tools/testing/selftests/bpf/progs/test_tc_tunnel.c
62
struct iphdr ip;
tools/testing/selftests/bpf/progs/test_tc_tunnel.c
68
struct ipv6hdr ip;
tools/testing/selftests/bpf/progs/test_uprobe.c
86
unsigned long ip;
tools/testing/selftests/bpf/progs/test_uprobe.c
96
ctx->ip = ip;
tools/testing/selftests/bpf/progs/xdpwall.c
153
filter_ipv6_addr(&info->ip.ipv6->saddr);
tools/testing/selftests/bpf/progs/xdpwall.c
156
filter_ipv4_addr(info->ip.ipv4->saddr);
tools/testing/selftests/bpf/progs/xdpwall.c
158
filter_ipv4_lpm(info->ip.ipv4->saddr);
tools/testing/selftests/bpf/progs/xdpwall.c
160
filter_ipv4_lpm(info->ip.ipv4->daddr);
tools/testing/selftests/bpf/progs/xdpwall.c
281
info->ip.ipv6 = inner_ip6h;
tools/testing/selftests/bpf/progs/xdpwall.c
291
info->ip.ipv4 = inner_ip4h;
tools/testing/selftests/bpf/progs/xdpwall.c
307
info->ip.ipv6 = ip6h;
tools/testing/selftests/bpf/progs/xdpwall.c
98
} ip;
tools/testing/selftests/kvm/lib/guest_sprintf.c
249
long *ip = va_arg(args, long *);
tools/testing/selftests/kvm/lib/guest_sprintf.c
250
*ip = (str - buf);
tools/testing/selftests/kvm/lib/guest_sprintf.c
252
int *ip = va_arg(args, int *);
tools/testing/selftests/kvm/lib/guest_sprintf.c
253
*ip = (str - buf);
tools/testing/selftests/mm/protection_keys.c
316
unsigned long ip;
tools/testing/selftests/mm/protection_keys.c
334
ip = MCONTEXT_IP(uctxt->uc_mcontext);
tools/testing/selftests/mm/protection_keys.c
340
__func__, trapno, ip, si_code_str(si->si_code),
tools/testing/selftests/net/ip_defrag.c
101
static uint16_t udp_checksum(struct ip *iphdr, struct udphdr *udphdr)
tools/testing/selftests/net/ip_defrag.c
152
udphdr.check = udp_checksum((struct ip *)ip_frame, &udphdr);
tools/testing/selftests/net/ip_defrag.c
176
struct ip *iphdr = (struct ip *)ip_frame;
tools/testing/selftests/net/ip_defrag.c
206
struct ip *iphdr = (struct ip *)ip_frame;
tools/testing/selftests/net/nat6to4.bpf.c
111
for (int i = 0; i < sizeof(ip) / sizeof(__u16); ++i)
tools/testing/selftests/net/nat6to4.bpf.c
112
sum4 += ((__u16 *)&ip)[i];
tools/testing/selftests/net/nat6to4.bpf.c
117
ip.check = (__u16)~sum4; // sum4 cannot be zero, so this is never 0xFFFF
tools/testing/selftests/net/nat6to4.bpf.c
145
*(struct iphdr *)(new_eth + 1) = ip;
tools/testing/selftests/net/nat6to4.bpf.c
94
struct iphdr ip = {
tools/testing/selftests/net/ovpn/ovpn-cli.c
797
void *ip = pattrs[OVPN_A_PEER_REMOTE_IPV6];
tools/testing/selftests/net/ovpn/ovpn-cli.c
807
inet_ntop(AF_INET6, nla_data(ip), buf, sizeof(buf));
tools/testing/selftests/net/ovpn/ovpn-cli.c
812
void *ip = pattrs[OVPN_A_PEER_LOCAL_IPV6];
tools/testing/selftests/net/ovpn/ovpn-cli.c
814
inet_ntop(AF_INET6, nla_data(ip), buf, sizeof(buf));
tools/testing/selftests/net/ovpn/ovpn-cli.c
820
void *ip = pattrs[OVPN_A_PEER_REMOTE_IPV4];
tools/testing/selftests/net/ovpn/ovpn-cli.c
823
inet_ntop(AF_INET, nla_data(ip), buf, sizeof(buf));
tools/testing/selftests/net/psock_tpacket.c
145
struct iphdr *ip = pay + sizeof(*eth);
tools/testing/selftests/net/psock_tpacket.c
156
for (i = 0; i < sizeof(*ip); ++i)
tools/testing/selftests/net/psock_tpacket.c
159
ip->ihl = 5;
tools/testing/selftests/net/psock_tpacket.c
160
ip->version = 4;
tools/testing/selftests/net/psock_tpacket.c
161
ip->protocol = 0x11;
tools/testing/selftests/net/psock_tpacket.c
162
ip->frag_off = 0;
tools/testing/selftests/net/psock_tpacket.c
163
ip->ttl = 64;
tools/testing/selftests/net/psock_tpacket.c
164
ip->tot_len = htons((uint16_t) *len - sizeof(*eth));
tools/testing/selftests/net/psock_tpacket.c
166
ip->saddr = htonl(INADDR_LOOPBACK);
tools/testing/selftests/net/psock_tpacket.c
167
ip->daddr = htonl(INADDR_LOOPBACK);
tools/testing/selftests/net/psock_tpacket.c
169
memset(pay + sizeof(*eth) + sizeof(*ip),
tools/testing/selftests/net/sctp_hello.c
11
static void set_addr(struct sockaddr_storage *ss, char *ip, char *port, int *len)
tools/testing/selftests/net/sctp_hello.c
16
a->sin_addr.s_addr = inet_addr(ip);
tools/testing/selftests/net/sctp_hello.c
23
inet_pton(AF_INET6, ip, &a->sin6_addr);
tools/testing/selftests/net/tcp_ao/unsigned-md5.c
500
static void client_add_ip(union tcp_addr *client, const char *ip)
tools/testing/selftests/net/tcp_ao/unsigned-md5.c
504
if (inet_pton(family, ip, client) != 1)
tools/testing/selftests/net/tcp_ao/unsigned-md5.c
505
test_error("Can't convert ip address %s", ip);
tools/testing/selftests/net/tcp_port_share.c
57
static void make_inet_addr(int af, const char *ip, __u16 port,
tools/testing/selftests/net/tcp_port_share.c
69
inet_pton(af, ip, &addr->v4.sin_addr);
tools/testing/selftests/net/tcp_port_share.c
76
inet_pton(af, ip, &addr->v6.sin6_addr);
tools/testing/selftests/net/tcp_port_share.c
81
snprintf(addr->str, sizeof(addr->str), fmt, ip, port);
tools/testing/selftests/powerpc/copyloops/exc_validate.c
25
unsigned long *ip = &UCONTEXT_NIA(uc);
tools/testing/selftests/powerpc/copyloops/exc_validate.c
34
if (insn == *ip) {
tools/testing/selftests/powerpc/copyloops/exc_validate.c
35
*ip = fixup;
tools/testing/selftests/powerpc/copyloops/exc_validate.c
40
printf("No exception table match for NIA %lx ADDR %lx\n", *ip, addr);
tools/testing/selftests/powerpc/mm/wild_bctr.c
111
unsigned long ip;
tools/testing/selftests/powerpc/mm/wild_bctr.c
116
.ip = BAD_NIP,
tools/testing/selftests/powerpc/primitives/load_unaligned_zeropad.c
73
unsigned long *ip = &UCONTEXT_NIA(uc);
tools/testing/selftests/powerpc/primitives/load_unaligned_zeropad.c
82
if (insn == *ip) {
tools/testing/selftests/powerpc/primitives/load_unaligned_zeropad.c
83
*ip = fixup;
tools/testing/selftests/powerpc/primitives/load_unaligned_zeropad.c
88
printf("No exception table match for NIA %lx ADDR %lx\n", *ip, addr);
tools/testing/selftests/x86/syscall_arg_fault.c
60
unsigned short *ip = (unsigned short *)ctx->uc_mcontext.gregs[REG_IP];
tools/testing/selftests/x86/syscall_arg_fault.c
62
if (*ip == 0x340f || *ip == 0x050f) {
tools/testing/selftests/x86/syscall_arg_fault.c
77
unsigned short *ip = (unsigned short *)ctx->uc_mcontext.gregs[REG_IP];
tools/testing/selftests/x86/syscall_arg_fault.c
79
if (*ip == 0x0b0f) {
tools/testing/selftests/x86/sysret_rip.c
102
printf("[RUN]\tTrying a SYSCALL that falls through to 0x%lx\n", ip);
tools/testing/selftests/x86/sysret_rip.c
107
if (ip <= (1UL << 47) - PAGE_SIZE) {
tools/testing/selftests/x86/sysret_rip.c
120
rip = ip;
tools/testing/selftests/x86/sysret_rip.c
124
[syscall_insn] "rm" (ip - 2));
tools/testing/selftests/x86/sysret_rip.c
74
static void test_sigreturn_to(unsigned long ip)
tools/testing/selftests/x86/sysret_rip.c
76
rip = ip;
tools/testing/selftests/x86/sysret_rip.c
77
printf("[RUN]\tsigreturn to 0x%lx\n", ip);
tools/testing/selftests/x86/sysret_rip.c
97
static void test_syscall_fallthrough_to(unsigned long ip)
tools/testing/selftests/x86/sysret_rip.c
99
void *new_address = (void *)(ip - 4096);
tools/testing/selftests/x86/test_vsyscall.c
472
unsigned long ip = ctx->uc_mcontext.gregs[REG_RIP];
tools/testing/selftests/x86/test_vsyscall.c
474
if (((ip ^ 0xffffffffff600000UL) & ~0xfffUL) == 0)
tools/testing/selftests/x86/unwind_vdso.c
101
if (!got_sysinfo && ip == sysinfo) {
tools/testing/selftests/x86/unwind_vdso.c
108
ip, return_address);
tools/testing/selftests/x86/unwind_vdso.c
114
if (ip == return_address) {
tools/testing/selftests/x86/unwind_vdso.c
120
printf("\tSIGTRAP at 0x%lx\n", ip);
tools/testing/selftests/x86/unwind_vdso.c
122
state.ip = ip;
tools/testing/selftests/x86/unwind_vdso.c
52
unsigned long ip; /* trap source */
tools/testing/selftests/x86/unwind_vdso.c
59
unsigned long ip = _Unwind_GetIP(ctx);
tools/testing/selftests/x86/unwind_vdso.c
62
if (ip == state->ip)
tools/testing/selftests/x86/unwind_vdso.c
67
printf("\t 0x%lx\n", ip);
tools/testing/selftests/x86/unwind_vdso.c
69
if (ip == return_address) {
tools/testing/selftests/x86/unwind_vdso.c
99
unsigned long ip = ctx->uc_mcontext.gregs[REG_EIP];